Chief AI Officer 2026: Real Role or Just Another C-Level Title?
Tobias Massow
⏳ 9 min read The Chief AI Officer is the most frequently announced-and least understood-C-level ...
For years, the central data warehouse was the gold standard for enterprise analytics. Then came the data lake – designed to store everything, yet too often devolving into a “data swamp.” Now Data Mesh promises to solve the foundational problem: Who owns the data – and who is accountable for its quality?
Data Mesh’s answer is radical: Not IT – but the business domains themselves. Sales owns sales data, production owns production data, HR owns HR data. The central platform provides only infrastructure. For Germany’s Mittelstand (mid-sized enterprises), this approach is especially promising – provided it’s implemented pragmatically.
The pattern is identical across every large enterprise: A business unit needs a new report. It submits a request to the centralized data team. The data team’s backlog stretches six weeks. By the time the report is finally delivered, the original question has changed. The bottleneck is structural.
Centralized data teams cannot build up domain expertise quickly enough to deliver high-quality data products. They don’t understand accounting data as well as the finance team does – or production data as well as the manufacturing team does. The result? Misunderstandings, rework, and reports that miss the mark entirely.
Data Mesh solves this problem by shifting ownership to where the knowledge resides – in the business domains themselves.
1. Domain-Oriented Ownership: Each business domain owns and is accountable for its own data. This does not mean every domain builds its own data warehouse, but rather that it ensures the quality, documentation, and availability of its data.
2. Data as a Product: Data is treated like an internal product – with defined SLAs, documentation, versioning, and user feedback. Every data product has a product owner responsible for its quality and ongoing development.
3. Self-Service Data Platform: A central platform provides tools enabling business domains to independently build, test, and deploy their data products. The platform abstracts away technical complexity.
4. Federated Governance: Cross-domain standards for interoperability, security, and compliance are defined centrally – but implemented decentralised. The governance team sets the framework; domains fill in the details.
The textbook version of Data Mesh assumes a certain level of organisational maturity: a strong data culture, experienced business domains, and a high-performing platform. Reality in Germany’s SME sector looks different. The pragmatic entry point works like this:
Step 1: Identify two to three domains with high data competence and a clear need for better data products – typically Sales, Production, and Finance.
Step 2: Appoint a Data Product Owner for each pilot domain: a business-domain employee with strong data affinity who dedicates 20-30% of their time to data products.
Step 3: Define and deliver one initial data product per domain. Start simple: a well-documented, reliable dataset usable by other departments.
Step 4: After six months, evaluate: Has data quality improved? Are the products being used? Where are the friction points? Then scale.
Data Mesh is not a technology decision – but it does require robust technical foundations:
Data Catalog: A central catalog where all data products are discoverable, documented, and assessable. Tools such as DataHub, Atlan, or Databricks’ Unity Catalog fulfil this function.
Data Contracts: Formal agreements between data producers and consumers covering format, quality, and service-level agreements (SLAs). These prevent changes at the source from silently breaking downstream systems.
Compute and Storage: Cloud platforms such as Snowflake, Databricks, or BigQuery are well suited – because they natively support multi-tenancy and self-service access. On-premises deployment is possible but significantly more complex.
Important: Technology investment for Data Mesh is not higher than for a centralized data warehouse. It’s simply distributed differently – less centrally, and more across platform tooling and domain enablement.
No. The core principle – moving data ownership to where the domain expertise resides – works effectively starting at around 100 employees and three to four clearly defined business domains. The scope of implementation scales with company size.
Data Mesh and data warehouses are not mutually exclusive. Many successful implementations use an existing warehouse as the platform layer on which domains publish their data products. Data Mesh is an organisational shift – not a technical replacement.
It transforms into a platform team. Rather than building reports and pipelines itself, it delivers self-service tools, defines standards, and supports domains in building their data capabilities. Its role doesn’t become less important – just different.
Through three mechanisms: Data contracts formally define quality expectations; automated data quality checks in the platform validate every data delivery; and transparent quality metrics in the data catalog create incentives for high-performing data products – no one wants to be known for delivering the lowest-rated product.
First pilot domains can go live in three to six months. Enterprise-wide scaling typically takes 18 to 24 months. The most critical success factor isn’t technology – it’s the organisation’s willingness to genuinely decentralise accountability.
Source of header image: Unsplash / JJ Ying