AI & Digital

AI Readiness Isn't an IT Problem. It's a Data Problem.

AI tools amplify whatever structure exists in your data. If your procurement and asset data is messy, AI makes it expensively wrong. Here is what readiness actually requires.

AI & Digital18 March 20268 min read

Every organisation we speak to wants to use AI. Most of them cannot – not because the technology is unavailable, but because their data is not ready for it. AI tools do not create structure from chaos. They amplify whatever structure already exists. If your input data is inconsistent, fragmented, and poorly classified, the AI output will be confidently, expensively wrong.

What AI tools actually require from your data

Modern AI platforms – including Microsoft Copilot, Power BI with AI features, Microsoft Fabric, and third-party procurement intelligence tools – share a common assumption: the data they operate on is structured, consistent, and trustworthy.

Structured means fields contain what they are supposed to contain. Asset type is an asset type, not a free-text field that reads "pump (check with Brian)". Consistent means the same concept is described the same way across the dataset. Trustworthy means the data has been validated against a known standard, not just accumulated without quality checks.

When these conditions are not met, AI tools do not fail cleanly. They produce outputs that look authoritative – well-formatted dashboards, confident predictions – but are based on inputs that do not accurately represent the business.

The Microsoft Fabric reality

Many organisations in hard services, construction, and infrastructure are currently planning or executing a migration to Microsoft Fabric. This is a reasonable strategic direction. Fabric is powerful and well-integrated with the Microsoft ecosystem.

The problem is timing. Fabric migrations surface data quality issues in an uncomfortable way – at exactly the moment when the organisation is most invested in making the platform work. The pattern is consistent: the migration begins, the data is loaded, and the dashboards reveal inconsistencies that were not visible before. The first months of a Fabric deployment get consumed by data remediation rather than value creation.

The organisations that get the most from Fabric are the ones that resolve their data quality issues before migration. The data that flows into the platform is clean, classified, and consistently structured. The deployment delivers value from day one.

AI readiness: a practical checklist

Before your organisation is ready to extract value from AI tools, five data conditions need to be in place:

  • Consistent spend classification – Are all procurement lines classified to a recognised standard such as UNSPSC or eClass? Or described in free text that varies by user, site, and system?
  • Deduplicated supplier records – Does each supplier appear as one record? Siemens Building Technologies Ltd, Siemens BT, and Siemens invoice are three records that should be one.
  • Accurate asset data – Do asset records contain manufacturer name, model number, and installation date – or just a label and a location?
  • Validated data entry – When new data enters the system, is it checked against your existing data structure automatically, or does it arrive in whatever format the sender used?
  • Historical data quality baseline – Do you know what percentage of your current data is correctly classified, complete, and consistent – or is that an open question?

If two or more of these conditions are not met, your AI readiness is lower than the IT roadmap suggests.

What good data preparation looks like

Phase 1 – Baseline assessment. A structured evaluation of the current data: what is missing, what is inconsistent, and where the biggest quality gaps are. Output: a cleaned sample dataset and a clear action plan.

Phase 2 – Automated cleaning and enrichment. Machine learning handles the high-volume corrections. Cases requiring judgement are flagged for human review. This phase typically handles 85—95% of the remediation work without manual intervention.

Phase 3 – Ongoing quality control. A continuous layer that validates new data as it enters the system, maintaining the quality standard rather than allowing it to degrade over time.

If you want to understand how ready your data actually is, a data quality baseline is the right starting point.

"Pearstop built a system that automatically pulls the right items from our visiting reports into a clean proposal document. It saves our team a lot of time by eliminating the repetitive tasks of combining the correct items."

– Vince Out, Lemtech · Manufacturing & Air Filtration

Stephanie Wiechers

Stephanie Wiechers

CEO & Co-founder, Pearstop

Stephanie leads Pearstop's go-to-market and strategic direction. She works directly with procurement and FM leaders across Europe to understand how data quality affects margins, contracts, and AI readiness.

LinkedIn →