Standards

06

As organisations deploy AI systems to support operational decisions, customer interactions, and strategic planning, the reliability and governance of the underlying data ecosystem becomes a safety-critical concern. AI models do not fail loudly — they produce outputs that appear plausible while being systematically biased, outdated, or incomplete. When those outputs drive hiring decisions, credit assessments, medical triage, or operational risk management, the consequences of poor data governance are not merely technical: they manifest as discriminatory outcomes, regulatory breaches, or decisions made on a false picture of reality. This standard establishes that data ecosystems must be governed with the same rigour applied to production software — with defined ownership, lineage tracking, quality SLAs, access controls, and retention policies that ensure data is trustworthy at the point of use.

Good data governance is frequently mischaracterised as a compliance burden — a set of controls imposed on innovation. In practice, well-governed data ecosystems accelerate innovation by making data discoverable, trustworthy, and reusable, reducing the time teams spend questioning data provenance or remediating quality issues before they can act. Federated governance models, informed by data mesh principles, allow domain teams to own and operate their data products within a framework of organisational standards, balancing autonomy with accountability. This standard supports that balance — enabling teams to move quickly while ensuring that the data underpinning AI-driven decisions is reliable, auditable, and fit for purpose across its full lifecycle.

03

Artificial intelligence and machine learning capabilities are only as good as the data they are trained on, fine-tuned with, and operate against at runtime. Organisations frequently invest in AI tooling and model capability while underestimating the foundational requirement: that internal data must be clean, well-catalogued, semantically described, and accessible through reliable interfaces before AI can deliver consistent value. Without this foundation, AI initiatives stall during the data preparation phase, produce unreliable outputs due to inconsistent inputs, or fail to reach production at all. This standard establishes the expectation that data quality, structure, and accessibility are treated as first-class engineering concerns — prerequisites to AI investment rather than afterthoughts.

The organisations that extract the most value from AI are those that treat their internal data as a strategic product. This means building data catalogues that make assets discoverable, defining data contracts between systems so that schemas and quality guarantees are explicit, creating API-accessible data products that AI agents and analytical pipelines can consume reliably, and establishing semantic layers that allow AI models to reason about business concepts rather than raw technical fields. By meeting this standard, engineering and data teams create a compounding asset — a data foundation that not only enables current AI use cases but accelerates future ones, reduces the cost of onboarding new models, and prevents the accumulation of data debt that eventually makes AI initiatives unviable.