All blog articles
You wouldn't build a skyscraper on sand. Why build AI on bad data?

You wouldn't build a skyscraper on sand. Why build AI on bad data?

Supply chains are racing to adopt AI, but one issue keeps blocking progress: data quality. In a recent session with Intent Group’s Ed Lawson, we explored why AI success depends less on models—and more on clean, contextual data. The biggest risk? Skipping foundational work in the rush to scale.

Tomaz Suklje
March 31, 2025
TIME TO READ:
MINUTES

In a recent conversation with Ed Lawson, Managing Director at Intent Group, we explored AI’s ROI and the Killer App for Supply Chains. We also ran a live poll with supply chain professionals to understand their real-world blockers to AI adoption. The #1 result? Data quality. 57% of respondents flagged it as the primary obstacle - above ROI, talent, or scalability concerns.

The poll results confirmed something we're seeing in the market: the disconnect between AI ambition and foundational readiness. While companies chase AI capabilities, they often skip the first step - ensuring their data is usable, reliable, and integrated.

Data is still the elephant in the room

Despite the AI hype cycle, one truth holds: AI is only as good as the data it’s fed and trained on. If the data is messy, siloed, or outdated, even the most advanced models will underperform - or mislead entirely. If the models ingest good operational data, they will be able to train better and deliver good insights.

We’ve been talking about data quality for decades, and AI brings it into sharp relief. This legacy issue has now become a frontline problem, with AI amplifying the risks of bad data rather than masking them.

This isn't just theoretical. IDC’s research shows that 90% of enterprise data is unstructured, and yet less than half of it is ever analyzed to extract value. IDC calls it what it is: a "gold mine" that’s routinely wasted.

It's not just quality. Context is critical

But quality alone isn’t enough. As anyone in supply chain knows, the right number at the wrong time - or without context -is still the wrong decision. It’s not just about clean data. It’s about context-rich data.

Supply chain leaders aren't just looking for systems that can analyze, they need AI that understands. Missing context leads to misleading outputs, especially in forecasting.

Humans still outperform AI in decision-making when critical nuances aren’t captured by data alone. Ultimately, humans must provide the context. But that’s where AI agents can help by jumping between silos and integrating fragmented information.

AI agents are really good at jumping between silos. But that still requires context to be codified and fed in – and that’s where LLMs come in, with their capacity for structuring, normalizing, mapping, and validating data, as well as reasoning and prompting to action based on trained data.

Why this stops AI in its tracks

Without both clean and contextualized data, AI pilots hit walls fast. We’ve seen this repeatedly with organizations eager to prove value but stuck in data chaos. Typical symptoms?

  • Siloed systems and departmental disconnects
  • Unstructured, incompatible formats
  • Outdated, incomplete records that make predictions unreliable

This erodes trust both in the AI systems and the teams that champion them.

IDC backs this up with hard data: organizations with fragmented, siloed unstructured data face double the cost of data breaches - averaging $4.5M annually, compared to $2.2M for companies that manage their data through centralized, secure platforms.

The productivity and risk equation

The cost of failure isn’t just technological - it’s operational, financial, and regulatory. Fragmented data slows AI adoption and bleeds productivity:

  • 51% of organizations faced non-compliance penalties in the past 12 months due to mismanaged unstructured data
  • Only 26% of unstructured data is analyzed in an automated way; the rest requires human effort, slowing scale

Meanwhile, productivity loss from data hunting, duplicated effort, and version mismatches hurts operational agility - especially in supply chain contexts where timing is everything.

Strategic imperatives

The good news is all these are solvable problems. But they require a strategic shift.

#1 Treat data as a capital asset

Data readiness must precede AI readiness. Without putting the work in structuring processes, transferring context, and cleaning data, AI won’t just under-deliver - it will mislead.

#2 Embed domain knowledge into AI Agents

Vertical-specific AI agents beat generic platforms because they understand nuances in processes, logic, and decision-making.

#3 Move past isolated wins - fix the foundation to scale AI

We’re at a stage where we only see AI succeeding when it solves real, costly inefficiencies. However, chasing one-off ROI creates more silos. Poor data quality is a silent cost center and a clear place to start. But clean data and structured processes are the healthy foundation as well as the steady path to scalable AI and survival over the next decade.

Data quality isn’t an IT hygiene issue. It’s the #1 success factor for AI. AI transformation begins with data transformation, which relies heavily on restructured processes and clean data. No shortcuts. No exceptions.

ABOUT THE AUTHOR
Tomaz Suklje

Enjoyed this read?

Subscribe to our newsletter, and we will send AI automation insights like this straight to your inbox on a regular basis.

Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.