Skip to main content

Data Mesh and the State of the Data Lakehouse

Alex Merced
Dremio

Data mesh, an increasingly important decentralized approach to data architecture and organizational design, focuses on treating data as a product, emphasizing domain-oriented data ownership, self-service tools and federated governance. The 2024 State of the Data Lakehouse report from Dremio presents evidence of the growing adoption of data mesh architectures in enterprises. This innovative approach has seen a significant uptake, with 84% of respondents reporting full or partial implementation of data mesh strategies within their organizations. Moreover, 97% expect the implementation of data mesh to continue expanding in the next year.

The report highlights that the drive towards data mesh is increasingly becoming a business strategy to enhance agility and speed in problem-solving and innovation. Interestingly, the initiative for data mesh is more frequently driven by line-of-business units and business leaders (52%) rather than central IT teams. This shift indicates a more integrated approach to data management, where business units are directly involved in the governance and utilization of data, promoting a more agile and responsive data culture.

Objectives for implementing data mesh strategies are varied but focus on improving data quality (64%) and governance (58%), with significant emphasis on enhancing data access, decision-making capabilities, scalability, and agility. These objectives reflect the core benefits of adopting a data mesh approach: a more accessible, reliable, and scalable data infrastructure that can adapt to the fast-paced changes in business requirements and technological advancements.

The synergy between data mesh and data lakehouses is particularly noteworthy. The data lakehouse architecture, which combines the best features of data lakes and data warehouses, provides an ideal environment for implementing data mesh principles. Data lakehouses offer the scalability and flexibility of data lakes, with the added governance, performance, and reliability of data warehouses, making them a perfect match for the decentralized, domain-driven approach of data mesh.

Moreover, adopting data lakehouses is critical in the AI era, as highlighted in the report. Data lakehouses enable self-service and ease of access to data, which are key for AI development and innovation. With 81% of respondents using a data lakehouse to support data scientists in building and improving AI models and applications, it's clear that the data lakehouse architecture is not just a trend, but a foundational element in the future of data management and analytics.

The report also sheds light on the driving forces behind data mesh and lakehouse adoption: improved data quality, governance, and enabling AI and machine learning applications were most cited. This aligns with the broader digital transformation trend, where businesses seek to leverage data more effectively to gain insights, innovate, and maintain competitive advantage.

The report underscores the significant impact of data mesh and lakehouse architectures on the enterprise data landscape. As businesses continue to navigate the complexities of managing vast amounts of data, the principles of data mesh — decentralization, domain-oriented data ownership, and product thinking — coupled with the technological foundation provided by data lakehouses, offer a promising path forward. Together, they enable enterprises to harness the full potential of their data, driving innovation, agility, and growth in the digital age.

Alex Merced is a Developer Advocate at Dremio

Hot Topics

The Latest

In APMdigest's 2026 Observability Predictions Series, industry experts offer predictions on how Observability and related technologies will evolve and impact business in 2025. Part 6 covers OpenTelemetry ...

In APMdigest's 2026 Observability Predictions Series, industry experts offer predictions on how Observability and related technologies will evolve and impact business in 2025. Part 5 covers APM and infrastructure monitoring ...

AI continues to be the top story across the industry, but a big test is coming up as retailers make the final preparations before the holiday season starts. Will new AI powered features help load up Santa's sleigh this year? Or are early adopters in for unpleasant surprises in the form of unexpected high costs, poor performance, or even service outages? ...

In APMdigest's 2026 Observability Predictions Series, industry experts offer predictions on how Observability and related technologies will evolve and impact business in 2025. Part 4 covers user experience, digital performance, website performance and ITSM ...

In APMdigest's 2026 Observability Predictions Series, industry experts offer predictions on how Observability and related technologies will evolve and impact business in 2025. Part 3 covers more predictions about Observability ...

In APMdigest's 2026 Observability Predictions Series, industry experts offer predictions on how Observability and related technologies will evolve and impact business in 2025. Part 2 covers predictions about Observability and AIOps ...

The Holiday Season means it is time for APMdigest's annual list of predictions, covering Observability and other IT performance topics. Industry experts — from analysts and consultants to the top vendors — offer thoughtful, insightful, and often controversial predictions on how Observability, AIOps, APM and related technologies will evolve and impact business in 2026 ...

IT organizations are preparing for 2026 with increased expectations around modernization, cloud maturity, and data readiness. At the same time, many teams continue to operate with limited staffing and are trying to maintain complex environments with small internal groups. These conditions are creating a distinct set of priorities for the year ahead. The DataStrike 2026 Data Infrastructure Survey Report, based on responses from nearly 280 IT leaders across industries, points to five trends that are shaping data infrastructure planning for 2026 ...

Developers building AI applications are not just looking for fault patterns after deployment; they must detect issues quickly during development and have the ability to prevent issues after going live. Unfortunately, traditional observability tools can no longer meet the needs of AI-driven enterprise application development. AI-powered detection and auto-remediation tools designed to keep pace with rapid development are now emerging to proactively manage performance and prevent downtime ...

Every few years, the cybersecurity industry adopts a new buzzword. "Zero Trust" has endured longer than most — and for good reason. Its promise is simple: trust nothing by default, verify everything continuously. Yet many organizations still hesitate to implement Zero Trust Network Access (ZTNA). The problem isn't that ZTNA doesn't work. It's that it's often misunderstood ...

Data Mesh and the State of the Data Lakehouse

Alex Merced
Dremio

Data mesh, an increasingly important decentralized approach to data architecture and organizational design, focuses on treating data as a product, emphasizing domain-oriented data ownership, self-service tools and federated governance. The 2024 State of the Data Lakehouse report from Dremio presents evidence of the growing adoption of data mesh architectures in enterprises. This innovative approach has seen a significant uptake, with 84% of respondents reporting full or partial implementation of data mesh strategies within their organizations. Moreover, 97% expect the implementation of data mesh to continue expanding in the next year.

The report highlights that the drive towards data mesh is increasingly becoming a business strategy to enhance agility and speed in problem-solving and innovation. Interestingly, the initiative for data mesh is more frequently driven by line-of-business units and business leaders (52%) rather than central IT teams. This shift indicates a more integrated approach to data management, where business units are directly involved in the governance and utilization of data, promoting a more agile and responsive data culture.

Objectives for implementing data mesh strategies are varied but focus on improving data quality (64%) and governance (58%), with significant emphasis on enhancing data access, decision-making capabilities, scalability, and agility. These objectives reflect the core benefits of adopting a data mesh approach: a more accessible, reliable, and scalable data infrastructure that can adapt to the fast-paced changes in business requirements and technological advancements.

The synergy between data mesh and data lakehouses is particularly noteworthy. The data lakehouse architecture, which combines the best features of data lakes and data warehouses, provides an ideal environment for implementing data mesh principles. Data lakehouses offer the scalability and flexibility of data lakes, with the added governance, performance, and reliability of data warehouses, making them a perfect match for the decentralized, domain-driven approach of data mesh.

Moreover, adopting data lakehouses is critical in the AI era, as highlighted in the report. Data lakehouses enable self-service and ease of access to data, which are key for AI development and innovation. With 81% of respondents using a data lakehouse to support data scientists in building and improving AI models and applications, it's clear that the data lakehouse architecture is not just a trend, but a foundational element in the future of data management and analytics.

The report also sheds light on the driving forces behind data mesh and lakehouse adoption: improved data quality, governance, and enabling AI and machine learning applications were most cited. This aligns with the broader digital transformation trend, where businesses seek to leverage data more effectively to gain insights, innovate, and maintain competitive advantage.

The report underscores the significant impact of data mesh and lakehouse architectures on the enterprise data landscape. As businesses continue to navigate the complexities of managing vast amounts of data, the principles of data mesh — decentralization, domain-oriented data ownership, and product thinking — coupled with the technological foundation provided by data lakehouses, offer a promising path forward. Together, they enable enterprises to harness the full potential of their data, driving innovation, agility, and growth in the digital age.

Alex Merced is a Developer Advocate at Dremio

Hot Topics

The Latest

In APMdigest's 2026 Observability Predictions Series, industry experts offer predictions on how Observability and related technologies will evolve and impact business in 2025. Part 6 covers OpenTelemetry ...

In APMdigest's 2026 Observability Predictions Series, industry experts offer predictions on how Observability and related technologies will evolve and impact business in 2025. Part 5 covers APM and infrastructure monitoring ...

AI continues to be the top story across the industry, but a big test is coming up as retailers make the final preparations before the holiday season starts. Will new AI powered features help load up Santa's sleigh this year? Or are early adopters in for unpleasant surprises in the form of unexpected high costs, poor performance, or even service outages? ...

In APMdigest's 2026 Observability Predictions Series, industry experts offer predictions on how Observability and related technologies will evolve and impact business in 2025. Part 4 covers user experience, digital performance, website performance and ITSM ...

In APMdigest's 2026 Observability Predictions Series, industry experts offer predictions on how Observability and related technologies will evolve and impact business in 2025. Part 3 covers more predictions about Observability ...

In APMdigest's 2026 Observability Predictions Series, industry experts offer predictions on how Observability and related technologies will evolve and impact business in 2025. Part 2 covers predictions about Observability and AIOps ...

The Holiday Season means it is time for APMdigest's annual list of predictions, covering Observability and other IT performance topics. Industry experts — from analysts and consultants to the top vendors — offer thoughtful, insightful, and often controversial predictions on how Observability, AIOps, APM and related technologies will evolve and impact business in 2026 ...

IT organizations are preparing for 2026 with increased expectations around modernization, cloud maturity, and data readiness. At the same time, many teams continue to operate with limited staffing and are trying to maintain complex environments with small internal groups. These conditions are creating a distinct set of priorities for the year ahead. The DataStrike 2026 Data Infrastructure Survey Report, based on responses from nearly 280 IT leaders across industries, points to five trends that are shaping data infrastructure planning for 2026 ...

Developers building AI applications are not just looking for fault patterns after deployment; they must detect issues quickly during development and have the ability to prevent issues after going live. Unfortunately, traditional observability tools can no longer meet the needs of AI-driven enterprise application development. AI-powered detection and auto-remediation tools designed to keep pace with rapid development are now emerging to proactively manage performance and prevent downtime ...

Every few years, the cybersecurity industry adopts a new buzzword. "Zero Trust" has endured longer than most — and for good reason. Its promise is simple: trust nothing by default, verify everything continuously. Yet many organizations still hesitate to implement Zero Trust Network Access (ZTNA). The problem isn't that ZTNA doesn't work. It's that it's often misunderstood ...