AI Adoption Fails at the Data Layer, Not the Model Layer

Most AI projects underinvest in data infrastructure and overinvest in models. The model is never the limiting factor.

AI Adoption Fails at the Data Layer, Not the Model Layer

The conversation about AI adoption in most organisations starts with the model. Which foundation model to use. Whether to fine-tune or use retrieval-augmented generation. What the benchmark scores say. These are legitimate questions, but they are the wrong questions to lead with. By the time model selection is the relevant decision, all the harder decisions should already be made. Most organisations are nowhere near that point.

The pattern across AI engagements is consistent: organisations invest heavily in the visible, exciting layer — the model — and underinvest in the invisible, unglamorous layer — the data. The result is not a bad model. The result is a model that cannot be used because the data it needs does not exist in a usable form.

The Model Rarely Breaks

When an AI project fails, the post-mortem almost never blames the model. Models are reliable. They do what they are designed to do, given adequate, well-structured data to work with. The failure is almost always upstream: the data does not exist, is in the wrong format, lives in systems that cannot talk to each other, is owned by teams that have no incentive to share it, or cannot be accessed in a way that makes the AI output usable in context.

Gartner's research is unambiguous here: 63% of organisations either do not have or cannot confirm they have the data management practices required for AI. The same research predicts that through 2026, organisations will abandon 60% of AI projects specifically because they are unsupported by AI-ready data. A 2024 Forrester study of 500 enterprise data leaders found that 73% identified data quality and completeness as the primary barrier to AI success — ranking it above model accuracy, computing costs, and talent shortages. These are not fringe results. They are consistent findings across independent research, and they point to the same conclusion: the model is almost never the limiting factor.

What "Not Having the Data" Actually Means

When organisations discover they do not have the data their AI project requires, the failure often looks like a single problem — a missing dataset, a system that does not log what it should. The reality is usually more structural than that.

Data problems in organisations are the accumulated result of years of decisions about what to capture, how to store it, and which systems to invest in. A company that has been running for ten years on a collection of ERP systems, spreadsheets, and email threads has not been negligent — it has been operating rationally, storing information in whatever format served the immediate need. The problem is that AI requires data that is consistent, accessible, and governed in ways that normal business operations do not demand.

The same customer might appear as three different entities across a CRM, an invoicing system, and a contract database. Supplier data captured in a procurement system might be structured differently from supplier data captured in a logistics system, with no link between them. Operational data that would be valuable for an AI application might exist only in the heads of the people who run the process, captured nowhere. None of this is a data team failure. It is an organisational design consequence — the result of building systems for operational purposes rather than analytical ones.

This is why the honest answer to "how long will the data preparation take?" is often longer than the organisation wants to hear. Healthcare organisations attempting AI implementations routinely find that the data integration work alone takes twelve to eighteen months and consumes sixty to seventy per cent of the project budget. The same pattern appears in manufacturing, financial services, and logistics — any domain where operational data has been collected over years in formats that were never designed for the analytical uses now being proposed.

The Organisational Problem Inside the Data Problem

Data is not just a technical problem. It is a governance problem, and governance problems are organisational problems.

In most organisations, data is owned by the function that collected it. The sales team owns the CRM. The operations team owns the ERP. The finance team owns the ledger. These ownership structures made sense when the data served only its originating function. They create serious obstacles when an AI application needs to combine data across functions — which most useful AI applications do.

An AI system designed to predict customer churn needs sales, support, and product usage data together. An AI system designed to optimise procurement needs supplier, inventory, and demand data together. Getting these datasets into a unified, accessible, governed form requires decisions that no individual function can make alone. It requires someone with the authority and the mandate to resolve ownership questions, establish data standards, and fund the integration work. In organisations without that authority clearly assigned, the data preparation stalls — not because of technical complexity, but because of politics.

The organisations that succeed at AI adoption have almost always done this governance work first, or are willing to do it concurrently. They treat data as a shared organisational asset rather than a functional by-product. The technology to do the integration is available and mature. The organisational will to do it is the constraint.

Why Winning AI Programmes Invest Differently

The spending ratio in successful AI programmes looks counterintuitive to organisations that have approached the problem from the model side. McKinsey's research on organisations generating significant financial returns from AI finds they are twice as likely to have redesigned end-to-end workflows before selecting modelling techniques. The Informatica CDO survey found that winning programmes earmark fifty to seventy per cent of timeline and budget for data readiness — extraction, normalisation, governance, quality management — before any model work begins.

This is not slowness. It is sequencing. An organisation that spends three months on data infrastructure before building a model will have a working AI system faster than an organisation that builds the model first and then discovers the data does not support it. The second path involves rebuilding — and rebuilding is always more expensive than building well the first time.

The practical implication of this is that the first question in any AI engagement should not be "which model?" It should be: what data do we have, where does it live, who owns it, and what would it take to make it usable? The answers to these questions determine whether a project is a six-month engagement or a two-year commitment. They determine whether the organisation's AI ambition is achievable on any timeline, or whether the prerequisite investment has not yet been made.

What an AI-Ready Data Audit Looks Like

Before any AI investment is committed, a practical data audit is required. Not a maturity assessment that produces a score and a set of recommendations. A practical inventory: what data exists, what questions it could answer, what format it is currently in, who owns it, what the governance model is, and what integration work would be required to make it accessible to an AI system.

This audit almost always produces surprises. Data that was assumed to exist does not. Data that was assumed to be clean is not. Ownership that was assumed to sit in one place is actually contested. Systems that were assumed to be compatible are not. These surprises are not failures — they are the output of an audit that is doing its job, surfacing the actual situation rather than the assumed one.

An organisation that conducts this audit before committing to an AI programme is an organisation that can plan realistically. It knows what the data preparation work actually involves. It can make an informed decision about whether to proceed, at what timeline, with what resources. An organisation that skips the audit and goes directly to model selection is an organisation that will encounter all the same surprises later — at higher cost, under greater pressure, after having already committed to a programme that the data cannot support.

The model question is interesting. The data question is prior.

Talk to us about this

If this article touches something you are dealing with, we would be glad to have a conversation.