Skip to main content

2026 DataOps Predictions - Part 1

APMdigest's Predictions Series continues with 2026 DataOps Predictions — industry experts offer predictions on how DataOps and related technologies will evolve and impact business in 2026.

THE RISE OF DATAOPS

In 2026, DataOps will become one of the most strategic functions in every enterprise. It will move from a behind-the-scenes enabler to the operational heart of AI success. DataOps teams will ensure that data entering AI systems is clean, governed, cost-optimized, and compliant, functioning as the "nutrition label" for enterprise intelligence. I expect advanced automation will play a role here as well, detecting duplication, drift, and lineage breaks, while AI itself assists in cleansing and transforming data in real time. As the volume and velocity of data continue to explode, organizations will depend on DataOps to deliver trust, speed, and efficiency. Simply put, DataOps will be the make-or-break discipline for AI and analytics maturity in 2026. 
Arturo Oliver
Senior Director, Market Strategy & Analyst Relations, ScienceLogic

AI-POWERED DATAOPS

AI-powered  DataOps  becomes table stakes: In 2026, data observability, lineage and governance will themselves be AI-augmented. Small and specialized models will continuously monitor pipelines for drift, anomalies, quality degradation and policy violations. Natural-language interfaces will allow business users to query data estates conversationally, while generative agents propose transformations, joins and mappings. 
Sunil Senan
Global Head of Data, Analytics and AI, Infosys

AI-DRIVEN AUTOMATION

By 2026, DataOps will become deeply intertwined with AI-driven automation. Manual orchestration and brittle pipelines will give way to autonomous data agents that monitor lineage, repair models, and optimize cost-performance in real time. The boundary between DataOps and MLOps will blur as inference becomes a native part of every data workflow. Teams will manage mixed workloads, i.e., structured, unstructured, and semantic through unified, declarative interfaces. The next frontier of DataOps isn't faster ETL, it's intelligent, self-healing data systems that learn from usage and continuously adapt to deliver trustworthy, production-ready data pipelines.
Yoni Michael
Cofounder and CTO, Typedef

AI READINESS

DataOps  becomes the "front door" to enterprise AI:  In 2026, AI readiness will be defined by the quality and agility of  DataOps, not by model sophistication alone. Enterprises will treat  DataOps  as the core discipline that ensures data is discoverable, trusted, governed and ready for consumption by both traditional analytics and generative AI. 
Sunil Senan
Global Head of Data, Analytics and AI, Infosys

REAL-TIME DATAOPS

Real-time  DataOps  becomes critical for operational AI: 
Edge and operational technology workloads move to cloud, streaming and event-driven data will become as important as batch pipelines.  In  2026,  DataOps  will be deeply embedded in OT environments,  supporting digital twins, predictive maintenance, autonomous  stores  and real-time risk models,  with stringent requirements on latency,  reliability  and explainability.
Sunil Senan
Global Head of Data, Analytics and AI, Infosys

HUMAN-AND-MACHINE-CENTERED DATA ECOSYSTEMS

The Rise of Human-and-Machine-Centered Data Ecosystems: We're moving toward a world where data platforms won't primarily serve people anymore; they'll serve machines. The new consumers of data are AI agents, which will increasingly drive decisions, generate insights, and automate processes at speeds humans can't match. These AI agents will require direct, governed, real-time access to all enterprise data to reason, generate, and act effectively. As AI agents become the primary consumers, enterprises must decide whether their data governance models empower or constrain them. This shift fundamentally changes everything about how we build and operate data infrastructure, from architecture and pipelines to governance and security, demanding a new approach that prioritizes machine-first accessibility without sacrificing trust or compliance.
Justin Borgman
CEO and Cofounder, Starburst

DATA INFRASTRUCTURE COSTS

Data infrastructure costs, including cloud data warehouses, lakehouses, and analytics platforms, will rise by 30-50% as companies continue to invest in AI in 2026, unless they fundamentally change their approach to data platform management. The exponential growth in data volumes, the proliferation of generative AI across all aspects of the business, and the lack of visibility into what's actually driving data consumption will all contribute to this increase. Because data platforms bill based on consumption, costs can be highly variable and difficult to predict, especially for organizations that don't have insights and controls over their company's usage. Data observability will be key for managing and mitigating these cost increases. Understanding which workloads, users, and teams are driving costs and identifying inefficiencies in queries and pipelines is the only way organizations can reduce inefficiencies and ensure maximum ROI. Frankly, 30% to 50% of their data infrastructure spend is likely waste — but the most innovative companies think about data observability from a cost-optimization vs. cost-cutting perspective, uncovering opportunities to reinvest in AI/ML innovation without increasing their overall data platform budget.
Kunal Agarwal
Co-Founder and CEO, Unravel Data

DATAOPS ROLE: CONTEXT ENGINEERING

Data Engineering Will Evolve Into "Context Engineering" as AI Agents Become Primary Data Consumers: A fundamental shift is happening in how we think about data engineering. For decades, data engineers prepared data for human consumption — analysts, data scientists, and business users. In 2026, AI agents will emerge as primary data consumers, and this changes everything. "Context engineering" isn't just a rebrand — it's a recognition that agents have different requirements than humans: they need fresh, streaming context delivered in milliseconds, not batch updates delivered overnight. The best data infrastructure companies will embrace this evolution, using their deep expertise in streaming, storage, and processing to solve genuinely new problems around agent-facing analytics and real-time context delivery. While the underlying principles of good data engineering remain constant, the application layer is transforming.
Sijie Guo
CEO, StreamNative

DATAOPS ROLE: STRATEGIST

2026 Will Lay the Foundation for Data Engineers to Hand Work Over to AI Agents: The coming year will be the inflection point where data engineers transition from builders to strategists, preparing to hand off key tasks to AI agents. In 2026, AI won't just be a tool for data engineers — it will be a co-pilot, laying the groundwork for a new era of autonomous data pipelines. While 2025 was about preparing data for AI, this year is about a fundamental shift: data engineers are moving beyond writing SQL to becoming strategic architects who supervise and validate AI-generated code. As data volume and pipeline complexity continue to outpace team growth, the only way forward is to embrace intelligent automation. This will pave the way for a third phase, where autonomous AI agents seamlessly manage and orchestrate pipelines, freeing data engineers to focus on high-value business outcomes and innovation. 2026 will be a critical year for data engineers, as they lay the foundation for this agentic AI and unlock significant productivity gains.
Chris Child
VP of Product, Data Engineering, Snowflake

DATAOPS ROLE: DATA CHOREOGRAPHER

In 2026, DataOps will shift from managing data pipelines to orchestrating intelligent data ecosystems where AI agents autonomously handle data quality, validation, and routing decisions. The traditional DataOps engineer role will evolve into a "data choreographer" who sets policies and handles exceptions while AI manages the routine operations that currently consume 80% of their time. Organizations that embrace this human-AI collaboration model in their DataOps practices will see 10x improvements in data processing speed and accuracy, fundamentally changing how enterprises think about data operations, from reactive maintenance to proactive optimization.
Deepak Singh
Chief Innovation Officer, Adeptia

DATAOPS ROLE: DATA CURATOR

Data teams will evolve from builders to curators: In 2026, the composition of data teams will change. With AI handling much of the pipeline and analytics work, the emphasis will shift towards skills in governance, curation, and communication. The manual plumbing of data, including integrating, cleaning, and stitching across systems, will increasingly be automated by AI. This will free data teams to focus on higher-order concerns. Data scientists and engineers will increasingly partner with data product managers, stewards, and digital librarians who ensure that systems act on the right context with the right guardrails. The center of gravity for data work will move up the stack — from infrastructure to insight, from data to knowledge.

Recruiting will follow suit. CIOs will seek out information scientists, behavioral researchers, and policy experts who can bridge the gap between human reasoning and machine execution. The old jobs of moving data are giving way to new roles focused on meaning, context, and oversight. The new frontier of data work will be less about moving bits and more about maintaining meaning to keep AI grounded in the nuances that make decisions explainable, compliant, and aligned with business purpose. 
Juan Sequeda
Principal Researcher, ServiceNow

DATAOPS ROLE: BUSINESS DECISION-MAKING PARTNER

Data Engineers Will Increasingly Become Business Decision-Making Partners to Enterprise Leaders in 2026: It's well known that AI models are only as good as the data they are trained on, and because of this, data has become every businesses' most valuable asset. Enterprises need real-time access to high-quality data to be successful, and they're increasingly leaning on their data engineers to deliver that. This shift has elevated the role of the data engineer to a strategic position, where we'll see more and more business-defining conversations now including data engineers' perspectives. Likewise, it will become increasingly important for data engineers to understand the business context behind the problems they're solving — including the broader business impact and the needs of the customer. The organizations that will win are those that recognize data engineers as essential business partners, integrating their expertise into strategic conversations to ensure data drives success.
Chris Child
VP of Product, Data Engineering, Snowflake

Go to: 2026 DataOps Predictions - Part 2

Hot Topics

The Latest

Outages aren't new. What's new is how quickly they spread across systems, vendors, regions and customer workflows. The moment that performance degrades, expectations escalate fast. In today's always-on environment, an outage isn't just a technical event. It's a trust event ...

Most organizations approach OpenTelemetry as a collection of individual tools they need to assemble from scratch. This view misses the bigger picture. OpenTelemetry is a complete telemetry framework with composable components that address specific problems at different stages of organizational maturity. You start with what you need today and adopt additional pieces as your observability practices evolve ...

One of the earliest lessons I learned from architecting throughput-heavy services is that simplicity wins repeatedly: fewer moving parts, loosely coupled execution (fewer synchronous calls), and precise timing metering. You want data and decisions to travel the shortest possible path. The goal is to build a system where every strategy and each line of code (contention is the key metric) complements the decision trees ...

As discussions around AI "autonomous coworkers" accelerate, many industry projections assume that agents will soon operate alongside human staff in making decisions, taking actions, and managing tasks with minimal oversight. But a growing number of critics (including some of the developers building these systems) argue that the industry still has a long way to go to be able to treat AI agents like fully trusted teammates ...

Enterprise AI has entered a transformational phase where, according to Digitate's recently released survey, Agentic AI and the Future of Enterprise IT, companies are moving beyond traditional automation toward Agentic AI systems designed to reason, adapt, and collaborate alongside human teams ...

The numbers back this urgency up. A recent Zapier survey shows that 92% of enterprises now treat AI as a top priority. Leaders want it, and teams are clamoring for it. But if you look closer at the operations of these companies, you see a different picture. The rollout is slow. The results are often delayed. There's a disconnect between what leaders want and what their technical infrastructure can handle ...

Kyndryl's 2025 Readiness Report revealed that 61% of global business and technology leaders report increasing pressure from boards and regulators to prove AI's ROI. As the technology evolves and expectations continue to rise, leaders are compelled to generate and prove impact before scaling further. This will lead to a decisive turning point in 2026 ...

Cloudflare's disruption illustrates how quickly a single provider's issue cascades into widespread exposure. Many organizations don't fully realize how tightly their systems are coupled to thirdparty services, or how quickly availability and security concerns align when those services falter ... You can't avoid these dependencies, but you can understand them ...

If you work with AI, you know this story. A model performs during testing, looks great in early reviews, works perfectly in production and then slowly loses relevance after operating for a while. Everything on the surface looks perfect — pipelines are running, predictions or recommendations are error-free, data quality checks show green; yet outcomes don't meet the ground reality. This pattern often repeats across enterprise AI programs. Take for example, a mid-sized retail banking and wealth-management firm with heavy investments in AI-powered risk analytics, fraud detection and personalized credit-decisioning systems. The model worked well for a while, but transactions increased, so did false positives by 18% ...

Basic uptime is no longer the gold standard. By 2026, network monitoring must do more than report status, it must explain performance in a hybrid-first world. Networks are no longer just static support systems; they are agile, distributed architectures that sit at the very heart of the customer experience and the business outcomes ... The following five trends represent the new standard for network health, providing a blueprint for teams to move from reactive troubleshooting to a proactive, integrated future ...

2026 DataOps Predictions - Part 1

APMdigest's Predictions Series continues with 2026 DataOps Predictions — industry experts offer predictions on how DataOps and related technologies will evolve and impact business in 2026.

THE RISE OF DATAOPS

In 2026, DataOps will become one of the most strategic functions in every enterprise. It will move from a behind-the-scenes enabler to the operational heart of AI success. DataOps teams will ensure that data entering AI systems is clean, governed, cost-optimized, and compliant, functioning as the "nutrition label" for enterprise intelligence. I expect advanced automation will play a role here as well, detecting duplication, drift, and lineage breaks, while AI itself assists in cleansing and transforming data in real time. As the volume and velocity of data continue to explode, organizations will depend on DataOps to deliver trust, speed, and efficiency. Simply put, DataOps will be the make-or-break discipline for AI and analytics maturity in 2026. 
Arturo Oliver
Senior Director, Market Strategy & Analyst Relations, ScienceLogic

AI-POWERED DATAOPS

AI-powered  DataOps  becomes table stakes: In 2026, data observability, lineage and governance will themselves be AI-augmented. Small and specialized models will continuously monitor pipelines for drift, anomalies, quality degradation and policy violations. Natural-language interfaces will allow business users to query data estates conversationally, while generative agents propose transformations, joins and mappings. 
Sunil Senan
Global Head of Data, Analytics and AI, Infosys

AI-DRIVEN AUTOMATION

By 2026, DataOps will become deeply intertwined with AI-driven automation. Manual orchestration and brittle pipelines will give way to autonomous data agents that monitor lineage, repair models, and optimize cost-performance in real time. The boundary between DataOps and MLOps will blur as inference becomes a native part of every data workflow. Teams will manage mixed workloads, i.e., structured, unstructured, and semantic through unified, declarative interfaces. The next frontier of DataOps isn't faster ETL, it's intelligent, self-healing data systems that learn from usage and continuously adapt to deliver trustworthy, production-ready data pipelines.
Yoni Michael
Cofounder and CTO, Typedef

AI READINESS

DataOps  becomes the "front door" to enterprise AI:  In 2026, AI readiness will be defined by the quality and agility of  DataOps, not by model sophistication alone. Enterprises will treat  DataOps  as the core discipline that ensures data is discoverable, trusted, governed and ready for consumption by both traditional analytics and generative AI. 
Sunil Senan
Global Head of Data, Analytics and AI, Infosys

REAL-TIME DATAOPS

Real-time  DataOps  becomes critical for operational AI: 
Edge and operational technology workloads move to cloud, streaming and event-driven data will become as important as batch pipelines.  In  2026,  DataOps  will be deeply embedded in OT environments,  supporting digital twins, predictive maintenance, autonomous  stores  and real-time risk models,  with stringent requirements on latency,  reliability  and explainability.
Sunil Senan
Global Head of Data, Analytics and AI, Infosys

HUMAN-AND-MACHINE-CENTERED DATA ECOSYSTEMS

The Rise of Human-and-Machine-Centered Data Ecosystems: We're moving toward a world where data platforms won't primarily serve people anymore; they'll serve machines. The new consumers of data are AI agents, which will increasingly drive decisions, generate insights, and automate processes at speeds humans can't match. These AI agents will require direct, governed, real-time access to all enterprise data to reason, generate, and act effectively. As AI agents become the primary consumers, enterprises must decide whether their data governance models empower or constrain them. This shift fundamentally changes everything about how we build and operate data infrastructure, from architecture and pipelines to governance and security, demanding a new approach that prioritizes machine-first accessibility without sacrificing trust or compliance.
Justin Borgman
CEO and Cofounder, Starburst

DATA INFRASTRUCTURE COSTS

Data infrastructure costs, including cloud data warehouses, lakehouses, and analytics platforms, will rise by 30-50% as companies continue to invest in AI in 2026, unless they fundamentally change their approach to data platform management. The exponential growth in data volumes, the proliferation of generative AI across all aspects of the business, and the lack of visibility into what's actually driving data consumption will all contribute to this increase. Because data platforms bill based on consumption, costs can be highly variable and difficult to predict, especially for organizations that don't have insights and controls over their company's usage. Data observability will be key for managing and mitigating these cost increases. Understanding which workloads, users, and teams are driving costs and identifying inefficiencies in queries and pipelines is the only way organizations can reduce inefficiencies and ensure maximum ROI. Frankly, 30% to 50% of their data infrastructure spend is likely waste — but the most innovative companies think about data observability from a cost-optimization vs. cost-cutting perspective, uncovering opportunities to reinvest in AI/ML innovation without increasing their overall data platform budget.
Kunal Agarwal
Co-Founder and CEO, Unravel Data

DATAOPS ROLE: CONTEXT ENGINEERING

Data Engineering Will Evolve Into "Context Engineering" as AI Agents Become Primary Data Consumers: A fundamental shift is happening in how we think about data engineering. For decades, data engineers prepared data for human consumption — analysts, data scientists, and business users. In 2026, AI agents will emerge as primary data consumers, and this changes everything. "Context engineering" isn't just a rebrand — it's a recognition that agents have different requirements than humans: they need fresh, streaming context delivered in milliseconds, not batch updates delivered overnight. The best data infrastructure companies will embrace this evolution, using their deep expertise in streaming, storage, and processing to solve genuinely new problems around agent-facing analytics and real-time context delivery. While the underlying principles of good data engineering remain constant, the application layer is transforming.
Sijie Guo
CEO, StreamNative

DATAOPS ROLE: STRATEGIST

2026 Will Lay the Foundation for Data Engineers to Hand Work Over to AI Agents: The coming year will be the inflection point where data engineers transition from builders to strategists, preparing to hand off key tasks to AI agents. In 2026, AI won't just be a tool for data engineers — it will be a co-pilot, laying the groundwork for a new era of autonomous data pipelines. While 2025 was about preparing data for AI, this year is about a fundamental shift: data engineers are moving beyond writing SQL to becoming strategic architects who supervise and validate AI-generated code. As data volume and pipeline complexity continue to outpace team growth, the only way forward is to embrace intelligent automation. This will pave the way for a third phase, where autonomous AI agents seamlessly manage and orchestrate pipelines, freeing data engineers to focus on high-value business outcomes and innovation. 2026 will be a critical year for data engineers, as they lay the foundation for this agentic AI and unlock significant productivity gains.
Chris Child
VP of Product, Data Engineering, Snowflake

DATAOPS ROLE: DATA CHOREOGRAPHER

In 2026, DataOps will shift from managing data pipelines to orchestrating intelligent data ecosystems where AI agents autonomously handle data quality, validation, and routing decisions. The traditional DataOps engineer role will evolve into a "data choreographer" who sets policies and handles exceptions while AI manages the routine operations that currently consume 80% of their time. Organizations that embrace this human-AI collaboration model in their DataOps practices will see 10x improvements in data processing speed and accuracy, fundamentally changing how enterprises think about data operations, from reactive maintenance to proactive optimization.
Deepak Singh
Chief Innovation Officer, Adeptia

DATAOPS ROLE: DATA CURATOR

Data teams will evolve from builders to curators: In 2026, the composition of data teams will change. With AI handling much of the pipeline and analytics work, the emphasis will shift towards skills in governance, curation, and communication. The manual plumbing of data, including integrating, cleaning, and stitching across systems, will increasingly be automated by AI. This will free data teams to focus on higher-order concerns. Data scientists and engineers will increasingly partner with data product managers, stewards, and digital librarians who ensure that systems act on the right context with the right guardrails. The center of gravity for data work will move up the stack — from infrastructure to insight, from data to knowledge.

Recruiting will follow suit. CIOs will seek out information scientists, behavioral researchers, and policy experts who can bridge the gap between human reasoning and machine execution. The old jobs of moving data are giving way to new roles focused on meaning, context, and oversight. The new frontier of data work will be less about moving bits and more about maintaining meaning to keep AI grounded in the nuances that make decisions explainable, compliant, and aligned with business purpose. 
Juan Sequeda
Principal Researcher, ServiceNow

DATAOPS ROLE: BUSINESS DECISION-MAKING PARTNER

Data Engineers Will Increasingly Become Business Decision-Making Partners to Enterprise Leaders in 2026: It's well known that AI models are only as good as the data they are trained on, and because of this, data has become every businesses' most valuable asset. Enterprises need real-time access to high-quality data to be successful, and they're increasingly leaning on their data engineers to deliver that. This shift has elevated the role of the data engineer to a strategic position, where we'll see more and more business-defining conversations now including data engineers' perspectives. Likewise, it will become increasingly important for data engineers to understand the business context behind the problems they're solving — including the broader business impact and the needs of the customer. The organizations that will win are those that recognize data engineers as essential business partners, integrating their expertise into strategic conversations to ensure data drives success.
Chris Child
VP of Product, Data Engineering, Snowflake

Go to: 2026 DataOps Predictions - Part 2

Hot Topics

The Latest

Outages aren't new. What's new is how quickly they spread across systems, vendors, regions and customer workflows. The moment that performance degrades, expectations escalate fast. In today's always-on environment, an outage isn't just a technical event. It's a trust event ...

Most organizations approach OpenTelemetry as a collection of individual tools they need to assemble from scratch. This view misses the bigger picture. OpenTelemetry is a complete telemetry framework with composable components that address specific problems at different stages of organizational maturity. You start with what you need today and adopt additional pieces as your observability practices evolve ...

One of the earliest lessons I learned from architecting throughput-heavy services is that simplicity wins repeatedly: fewer moving parts, loosely coupled execution (fewer synchronous calls), and precise timing metering. You want data and decisions to travel the shortest possible path. The goal is to build a system where every strategy and each line of code (contention is the key metric) complements the decision trees ...

As discussions around AI "autonomous coworkers" accelerate, many industry projections assume that agents will soon operate alongside human staff in making decisions, taking actions, and managing tasks with minimal oversight. But a growing number of critics (including some of the developers building these systems) argue that the industry still has a long way to go to be able to treat AI agents like fully trusted teammates ...

Enterprise AI has entered a transformational phase where, according to Digitate's recently released survey, Agentic AI and the Future of Enterprise IT, companies are moving beyond traditional automation toward Agentic AI systems designed to reason, adapt, and collaborate alongside human teams ...

The numbers back this urgency up. A recent Zapier survey shows that 92% of enterprises now treat AI as a top priority. Leaders want it, and teams are clamoring for it. But if you look closer at the operations of these companies, you see a different picture. The rollout is slow. The results are often delayed. There's a disconnect between what leaders want and what their technical infrastructure can handle ...

Kyndryl's 2025 Readiness Report revealed that 61% of global business and technology leaders report increasing pressure from boards and regulators to prove AI's ROI. As the technology evolves and expectations continue to rise, leaders are compelled to generate and prove impact before scaling further. This will lead to a decisive turning point in 2026 ...

Cloudflare's disruption illustrates how quickly a single provider's issue cascades into widespread exposure. Many organizations don't fully realize how tightly their systems are coupled to thirdparty services, or how quickly availability and security concerns align when those services falter ... You can't avoid these dependencies, but you can understand them ...

If you work with AI, you know this story. A model performs during testing, looks great in early reviews, works perfectly in production and then slowly loses relevance after operating for a while. Everything on the surface looks perfect — pipelines are running, predictions or recommendations are error-free, data quality checks show green; yet outcomes don't meet the ground reality. This pattern often repeats across enterprise AI programs. Take for example, a mid-sized retail banking and wealth-management firm with heavy investments in AI-powered risk analytics, fraud detection and personalized credit-decisioning systems. The model worked well for a while, but transactions increased, so did false positives by 18% ...

Basic uptime is no longer the gold standard. By 2026, network monitoring must do more than report status, it must explain performance in a hybrid-first world. Networks are no longer just static support systems; they are agile, distributed architectures that sit at the very heart of the customer experience and the business outcomes ... The following five trends represent the new standard for network health, providing a blueprint for teams to move from reactive troubleshooting to a proactive, integrated future ...