


What is hindering the progress of artificial intelligence? Or is it a data problem?
A new survey conducted by Forrester Consulting on behalf of Capital One shows that a lack of solid data foundations and solid data workflows is hindering enterprises from achieving better results in machine learning and artificial intelligence. Big progress.
According to a new report recently released by Capital One, "Achieving Key Business Outcomes with Actionable Machine Learning," although companies are increasingly integrating machine learning (ML) and artificial intelligence There has been some success getting (AI) into production, but they would have made greater progress if data management issues didn't get in the way.
The report is based in part on a Forrester survey of 150 data management decision-makers in North America in July this year, which found that 73% of decision-makers believe that the transparency, traceability and explainability of data flows are barriers to machine learning and AI. Key issues in application operationalization. The survey also found that 57% of respondents said internal silos between their data scientists and business operators hindered the deployment of machine learning.
David Kang, senior vice president and head of data analytics at Capital One, said: "We are still at a stage where machine learning algorithms themselves are not a barrier to people's success." "The key is data!"
When Capital One commissioned this survey, they thought the biggest challenge would focus on the actionability of machine learning. With the development of machine learning and artificial intelligence applications, MLOps (machine learning operations) has become an independent discipline, and it is also an area in which Capital One is investing.
But when this report came out, data decision-makers were most concerned about the lack of progress in building a solid data foundation, including data engineering and data infrastructure, Kang said.
"In some ways, this is disappointing. But in other ways, it is not surprising. Because leveraging data at scale requires a sustained focus on thinking and rethinking the data ecosystem Every capability in the system – how it is produced and consumed, how it is monitored, how it manages data in different ways. The transformation journey of the data ecosystem is still ongoing. It’s not something you do once and forget about. It’s Continued attention is needed."
Capital One's survey is similar to the findings of other recent studies. These studies found that data management issues slowed the pace and extent of adoption of machine learning and artificial intelligence. These include an MIT Technology Review report commissioned by Databricks in September that highlighted the dangers of improper data management on artificial intelligence; and an IDC study commissioned by Collibra in August that found , there is a correlation between companies with “data-intelligent” characteristics such as data cataloging, inheritance, quality management and governance, and market success.
If there is a common theme among these studies, it is that while the sophistication of existing machine learning and artificial intelligence technologies is growing rapidly, enterprises are finding that they have not done some core data management work well. , and these tasks are necessary to achieve these technological advances.
Businesses may find that ML or AI applications have a positive impact on a limited proof-of-concept (POC), but fail to take the necessary steps to ensure a smooth rollout into wider real-world production.
It may take a while before the technology you want to scale starts to make an impact in the market. The temptation is always there for these concepts to start seeing results and then suddenly find themselves somewhere with a bunch of data silos and a bunch of other data engineering infrastructure challenges.
Data science is still a fairly new discipline, and many companies are struggling to fill job openings. Capital One’s report found that 57% of respondents said they intend to use partnerships to fill gaps among data science practitioners. Kang said the lack of in-house expertise also makes it more critical for enterprises to establish core data infrastructure, making it easier for more advanced ML and AI use cases to be built on top of it and easier to repeat.
Capital One’s investigation also uncovered other issues slowing the adoption of machine learning and artificial intelligence. The company found that 36% of respondents cited "large, diverse, and confusing data sets" as a major obstacle, and 38% cited AI risks as the top challenge. 38% cited data silos across the organization and external data partners as a challenge to machine learning maturity.
The “hiccups” of data management don’t seem to be slowing down investment in artificial intelligence and machine learning (at least not yet). Capital One's survey found that 61% of decision makers plan to add new machine learning capabilities and applications in the next three years. More than half (53%) of respondents are currently prioritizing leveraging machine learning to improve business efficiency.
So, what are companies using machine learning for? Another interesting tidbit from the survey is that automated anomaly detection is the top use case for machine learning, with 40% of respondents reporting it as their top use case. This resonated with Kang, who helped Capital One build a machine learning-based anomaly detection system.
Other top use cases for ML and AI include: automated application and infrastructure updates (39%), and meeting new regulatory and privacy requirements for responsible and ethical AI (39%).
The above is the detailed content of What is hindering the progress of artificial intelligence? Or is it a data problem?. For more information, please follow other related articles on the PHP Chinese website!

Hot AI Tools

Undresser.AI Undress
AI-powered app for creating realistic nude photos

AI Clothes Remover
Online AI tool for removing clothes from photos.

Undress AI Tool
Undress images for free

Clothoff.io
AI clothes remover

Video Face Swap
Swap faces in any video effortlessly with our completely free AI face swap tool!

Hot Article

Hot Tools

Notepad++7.3.1
Easy-to-use and free code editor

SublimeText3 Chinese version
Chinese version, very easy to use

Zend Studio 13.0.1
Powerful PHP integrated development environment

Dreamweaver CS6
Visual web development tools

SublimeText3 Mac version
God-level code editing software (SublimeText3)

Hot Topics



This site reported on June 27 that Jianying is a video editing software developed by FaceMeng Technology, a subsidiary of ByteDance. It relies on the Douyin platform and basically produces short video content for users of the platform. It is compatible with iOS, Android, and Windows. , MacOS and other operating systems. Jianying officially announced the upgrade of its membership system and launched a new SVIP, which includes a variety of AI black technologies, such as intelligent translation, intelligent highlighting, intelligent packaging, digital human synthesis, etc. In terms of price, the monthly fee for clipping SVIP is 79 yuan, the annual fee is 599 yuan (note on this site: equivalent to 49.9 yuan per month), the continuous monthly subscription is 59 yuan per month, and the continuous annual subscription is 499 yuan per year (equivalent to 41.6 yuan per month) . In addition, the cut official also stated that in order to improve the user experience, those who have subscribed to the original VIP

Improve developer productivity, efficiency, and accuracy by incorporating retrieval-enhanced generation and semantic memory into AI coding assistants. Translated from EnhancingAICodingAssistantswithContextUsingRAGandSEM-RAG, author JanakiramMSV. While basic AI programming assistants are naturally helpful, they often fail to provide the most relevant and correct code suggestions because they rely on a general understanding of the software language and the most common patterns of writing software. The code generated by these coding assistants is suitable for solving the problems they are responsible for solving, but often does not conform to the coding standards, conventions and styles of the individual teams. This often results in suggestions that need to be modified or refined in order for the code to be accepted into the application

Large Language Models (LLMs) are trained on huge text databases, where they acquire large amounts of real-world knowledge. This knowledge is embedded into their parameters and can then be used when needed. The knowledge of these models is "reified" at the end of training. At the end of pre-training, the model actually stops learning. Align or fine-tune the model to learn how to leverage this knowledge and respond more naturally to user questions. But sometimes model knowledge is not enough, and although the model can access external content through RAG, it is considered beneficial to adapt the model to new domains through fine-tuning. This fine-tuning is performed using input from human annotators or other LLM creations, where the model encounters additional real-world knowledge and integrates it

To learn more about AIGC, please visit: 51CTOAI.x Community https://www.51cto.com/aigc/Translator|Jingyan Reviewer|Chonglou is different from the traditional question bank that can be seen everywhere on the Internet. These questions It requires thinking outside the box. Large Language Models (LLMs) are increasingly important in the fields of data science, generative artificial intelligence (GenAI), and artificial intelligence. These complex algorithms enhance human skills and drive efficiency and innovation in many industries, becoming the key for companies to remain competitive. LLM has a wide range of applications. It can be used in fields such as natural language processing, text generation, speech recognition and recommendation systems. By learning from large amounts of data, LLM is able to generate text

Machine learning is an important branch of artificial intelligence that gives computers the ability to learn from data and improve their capabilities without being explicitly programmed. Machine learning has a wide range of applications in various fields, from image recognition and natural language processing to recommendation systems and fraud detection, and it is changing the way we live. There are many different methods and theories in the field of machine learning, among which the five most influential methods are called the "Five Schools of Machine Learning". The five major schools are the symbolic school, the connectionist school, the evolutionary school, the Bayesian school and the analogy school. 1. Symbolism, also known as symbolism, emphasizes the use of symbols for logical reasoning and expression of knowledge. This school of thought believes that learning is a process of reverse deduction, through existing

Last week, amid the internal wave of resignations and external criticism, OpenAI was plagued by internal and external troubles: - The infringement of the widow sister sparked global heated discussions - Employees signing "overlord clauses" were exposed one after another - Netizens listed Ultraman's "seven deadly sins" Rumors refuting: According to leaked information and documents obtained by Vox, OpenAI’s senior leadership, including Altman, was well aware of these equity recovery provisions and signed off on them. In addition, there is a serious and urgent issue facing OpenAI - AI safety. The recent departures of five security-related employees, including two of its most prominent employees, and the dissolution of the "Super Alignment" team have once again put OpenAI's security issues in the spotlight. Fortune magazine reported that OpenA

Editor |ScienceAI Question Answering (QA) data set plays a vital role in promoting natural language processing (NLP) research. High-quality QA data sets can not only be used to fine-tune models, but also effectively evaluate the capabilities of large language models (LLM), especially the ability to understand and reason about scientific knowledge. Although there are currently many scientific QA data sets covering medicine, chemistry, biology and other fields, these data sets still have some shortcomings. First, the data form is relatively simple, most of which are multiple-choice questions. They are easy to evaluate, but limit the model's answer selection range and cannot fully test the model's ability to answer scientific questions. In contrast, open-ended Q&A

Editor | KX In the field of drug research and development, accurately and effectively predicting the binding affinity of proteins and ligands is crucial for drug screening and optimization. However, current studies do not take into account the important role of molecular surface information in protein-ligand interactions. Based on this, researchers from Xiamen University proposed a novel multi-modal feature extraction (MFE) framework, which for the first time combines information on protein surface, 3D structure and sequence, and uses a cross-attention mechanism to compare different modalities. feature alignment. Experimental results demonstrate that this method achieves state-of-the-art performance in predicting protein-ligand binding affinities. Furthermore, ablation studies demonstrate the effectiveness and necessity of protein surface information and multimodal feature alignment within this framework. Related research begins with "S
