Skip to main content

Why Observability Is the Missing Piece in Your Business Growth

Mimi Shalash
Splunk

For years, "observability" has been a backstage function. A quiet force that keeps digital systems running. What once lived deep in the data center is now at the center of every digital strategy. Even back in 2020, Gartner foreshadowed this shift, defining observability as "the evolution of monitoring into a process that offers insight into digital business applications, innovation, and customer experience."

That prediction has become even more relevant in an AI-driven world.

Every digital customer interaction, every cloud deployment, and every AI model depends on the same foundation: the ability to see, understand, and act on data in real time.

Digital moments like a mobile purchase, a supply chain handoff, or an AI inference run through complex layers of infrastructure. When that infrastructure falters, so does the business. Recent data from Splunk confirms that 74% of the business leaders believe observability is essential to monitoring critical business processes, and 66% feel it's key to understanding user journeys.

Because while the unknown is inevitable, observability makes it manageable. Let's explore why.

The Symbiotic Relationship Between AI and Observability

Image
Splunk

AI and observability are now inseparable. Data is the common language between them, and when used together, they amplify each other's strengths. AI helps observability teams detect patterns faster and gives engineers time back to focus on what truly moves the business forward: building better products and improving customer experiences.

Still, most teams aren't there yet. Many ITOps and engineering groups struggle with too many disconnected tools and an overload of false alerts, keeping them in a constant state of reactivity. This is the structural challenge that has persisted across enterprises: fragmented telemetry, inconsistent context, and decentralized standards.

It's no surprise, then, that organizations are turning to AI to correlate signals, reduce noise, and surface what matters most. Because prediction without observability is just speculation, and no business can afford to guess.

Splunk's research shows that 76% of practitioners now use AI regularly in daily operations, and 78% say it gives them more time to focus on innovation instead of maintenance. Yet every advancement brings new complexity. Humans in the loop are now responsible for ensuring model performance, and 47% of observability professionals say monitoring AI workloads has made their jobs more challenging, with 40% citing a lack of expertise as a barrier to AI readiness.

This gap represents a strategic opportunity. Organizations that upskill observability teams to measure AI performance and manage data quality will build a foundation of clean, governed, and trusted data that spans the entire enterprise. That means going beyond traditional IT telemetry to include operational technology (OT), IoT, sensor, and other machine data that power critical business systems.

The convergence of these once disparate data domains represents one of the most transformative opportunities in modern observability. Whether it's connecting insights from the factory floor, ERP systems, or even turbine sensors, organizations can finally uncover cross-functional intelligence that drives predictive action and measurable business outcomes.

Unlocking the Business Catalyst in Your Observability Practice

Realizing that vision requires strengthening the foundations of observability. Let's discuss four:

Minimize War Rooms and Reactivity: Many organizations still default to large, cross-functional escalations that duplicate effort and prolong mean time to resolution (MTTR). In fact, 1 in 5 respondents said they "often" or "always" start a war room that includes various departments. A more effective model emphasizes coordinated isolation and parallel response. When ITOps, engineering, and security teams visualize data through a common lens, they can trace the source of an issue faster and determine ownership. For example, when performance degradation in a key app is detected, shared telemetry allows engineering to see that the latency originates in an overloaded API gateway, not the database or underlying infrastructure. It's a simple example, but one that highlights how service mapping can triage, so people can focus on resolution, not reaction.

Get Alerting Under Control: False alerts drain engineering focus and erode trust in monitoring systems. Mature organizations address this by implementing adaptive thresholding, which dynamically adjusts alert parameters based on historical trends, system baselines, and seasonality. For example, instead of triggering dozens of CPU utilization alerts during routine batch processing every night, adaptive thresholding automatically adjusts expectations based on historical behavior. Managing alert suppression without removing early indicators of degradation is as much about data discipline as it is about process. When thresholds, alerts, and suppression logic are governed transparently and evolve with the environment, organizations build the foundation of data needed for higher levels of maturity and ultimately, AI readiness.

Lay the Foundation for Good Data That Reaps AI Benefits: Nearly half of respondents (48%) cite poor data quality as a barrier to achieving AI readiness. When engineering teams align on common data models, standardized collection practices, and comprehensive data coverage that reflects the full complexity of their environments, they establish consistent, reliable inputs for AI systems. The future belongs to organizations that can aggregate and contextualize all machine data from traditional homegrown applications, commercial off-the-shelf applications to environmental signals like temperature, vibration, and motion. When every data source speaks a common language, AI systems will be the catalyst for a new era of operational intelligence grounded in the full reality of the enterprise.

Embrace Forward Looking Architectures: The next evolution of observability is about building architectures that can adapt as fast as the systems they monitor. Organizations are investing in open and extensible technologies such as OpenTelemetry, code profiling, and observability as code to future-proof their data strategy. These approaches establish portability across environments, reduce vendor dependency, and embed observability into the software delivery lifecycle itself. OpenTelemetry, for example, is quickly becoming the industry standard for collecting, normalizing, and enriching telemetry data across hybrid and multicloud ecosystems. By adopting it early, teams can ensure consistency in how data is defined and exchanged, which sets the stage for complementary frameworks like Machine Communication Protocol (MCP). Together, these standards will be the future of advanced analytics, AI workflows, and autonomous operational systems.

Image
Splunk

Realizing Tangible Business Growth

When organizations take an innovative and responsible approach to observability, they create a foundation of agility and resilience that enables them to thrive through disruption and change. While the pace of innovation accelerates, the anchors of business success remain constant: building exceptional products, elevating customer experiences, and delivering measurable ROI that strengthens the bottom line.

In a world defined by data and driven by AI, observability is no longer just about visibility. It's now about vision.

Mimi Shalash is Observability Advisor at Splunk, a Cisco company

The Latest

Most organizations approach OpenTelemetry as a collection of individual tools they need to assemble from scratch. This view misses the bigger picture. OpenTelemetry is a complete telemetry framework with composable components that address specific problems at different stages of organizational maturity. You start with what you need today and adopt additional pieces as your observability practices evolve ...

One of the earliest lessons I learned from architecting throughput-heavy services is that simplicity wins repeatedly: fewer moving parts, loosely coupled execution (fewer synchronous calls), and precise timing metering. You want data and decisions to travel the shortest possible path. The goal is to build a system where every strategy and each line of code (contention is the key metric) complements the decision trees ...

As discussions around AI "autonomous coworkers" accelerate, many industry projections assume that agents will soon operate alongside human staff in making decisions, taking actions, and managing tasks with minimal oversight. But a growing number of critics (including some of the developers building these systems) argue that the industry still has a long way to go to be able to treat AI agents like fully trusted teammates ...

Enterprise AI has entered a transformational phase where, according to Digitate's recently released survey, Agentic AI and the Future of Enterprise IT, companies are moving beyond traditional automation toward Agentic AI systems designed to reason, adapt, and collaborate alongside human teams ...

The numbers back this urgency up. A recent Zapier survey shows that 92% of enterprises now treat AI as a top priority. Leaders want it, and teams are clamoring for it. But if you look closer at the operations of these companies, you see a different picture. The rollout is slow. The results are often delayed. There's a disconnect between what leaders want and what their technical infrastructure can handle ...

Kyndryl's 2025 Readiness Report revealed that 61% of global business and technology leaders report increasing pressure from boards and regulators to prove AI's ROI. As the technology evolves and expectations continue to rise, leaders are compelled to generate and prove impact before scaling further. This will lead to a decisive turning point in 2026 ...

Cloudflare's disruption illustrates how quickly a single provider's issue cascades into widespread exposure. Many organizations don't fully realize how tightly their systems are coupled to thirdparty services, or how quickly availability and security concerns align when those services falter ... You can't avoid these dependencies, but you can understand them ...

If you work with AI, you know this story. A model performs during testing, looks great in early reviews, works perfectly in production and then slowly loses relevance after operating for a while. Everything on the surface looks perfect — pipelines are running, predictions or recommendations are error-free, data quality checks show green; yet outcomes don't meet the ground reality. This pattern often repeats across enterprise AI programs. Take for example, a mid-sized retail banking and wealth-management firm with heavy investments in AI-powered risk analytics, fraud detection and personalized credit-decisioning systems. The model worked well for a while, but transactions increased, so did false positives by 18% ...

Basic uptime is no longer the gold standard. By 2026, network monitoring must do more than report status, it must explain performance in a hybrid-first world. Networks are no longer just static support systems; they are agile, distributed architectures that sit at the very heart of the customer experience and the business outcomes ... The following five trends represent the new standard for network health, providing a blueprint for teams to move from reactive troubleshooting to a proactive, integrated future ...

APMdigest's Predictions Series concludes with 2026 AI Predictions — industry experts offer predictions on how AI and related technologies will evolve and impact business in 2026. Part 5, the final installment, covers AI's impacts on IT teams ...

Why Observability Is the Missing Piece in Your Business Growth

Mimi Shalash
Splunk

For years, "observability" has been a backstage function. A quiet force that keeps digital systems running. What once lived deep in the data center is now at the center of every digital strategy. Even back in 2020, Gartner foreshadowed this shift, defining observability as "the evolution of monitoring into a process that offers insight into digital business applications, innovation, and customer experience."

That prediction has become even more relevant in an AI-driven world.

Every digital customer interaction, every cloud deployment, and every AI model depends on the same foundation: the ability to see, understand, and act on data in real time.

Digital moments like a mobile purchase, a supply chain handoff, or an AI inference run through complex layers of infrastructure. When that infrastructure falters, so does the business. Recent data from Splunk confirms that 74% of the business leaders believe observability is essential to monitoring critical business processes, and 66% feel it's key to understanding user journeys.

Because while the unknown is inevitable, observability makes it manageable. Let's explore why.

The Symbiotic Relationship Between AI and Observability

Image
Splunk

AI and observability are now inseparable. Data is the common language between them, and when used together, they amplify each other's strengths. AI helps observability teams detect patterns faster and gives engineers time back to focus on what truly moves the business forward: building better products and improving customer experiences.

Still, most teams aren't there yet. Many ITOps and engineering groups struggle with too many disconnected tools and an overload of false alerts, keeping them in a constant state of reactivity. This is the structural challenge that has persisted across enterprises: fragmented telemetry, inconsistent context, and decentralized standards.

It's no surprise, then, that organizations are turning to AI to correlate signals, reduce noise, and surface what matters most. Because prediction without observability is just speculation, and no business can afford to guess.

Splunk's research shows that 76% of practitioners now use AI regularly in daily operations, and 78% say it gives them more time to focus on innovation instead of maintenance. Yet every advancement brings new complexity. Humans in the loop are now responsible for ensuring model performance, and 47% of observability professionals say monitoring AI workloads has made their jobs more challenging, with 40% citing a lack of expertise as a barrier to AI readiness.

This gap represents a strategic opportunity. Organizations that upskill observability teams to measure AI performance and manage data quality will build a foundation of clean, governed, and trusted data that spans the entire enterprise. That means going beyond traditional IT telemetry to include operational technology (OT), IoT, sensor, and other machine data that power critical business systems.

The convergence of these once disparate data domains represents one of the most transformative opportunities in modern observability. Whether it's connecting insights from the factory floor, ERP systems, or even turbine sensors, organizations can finally uncover cross-functional intelligence that drives predictive action and measurable business outcomes.

Unlocking the Business Catalyst in Your Observability Practice

Realizing that vision requires strengthening the foundations of observability. Let's discuss four:

Minimize War Rooms and Reactivity: Many organizations still default to large, cross-functional escalations that duplicate effort and prolong mean time to resolution (MTTR). In fact, 1 in 5 respondents said they "often" or "always" start a war room that includes various departments. A more effective model emphasizes coordinated isolation and parallel response. When ITOps, engineering, and security teams visualize data through a common lens, they can trace the source of an issue faster and determine ownership. For example, when performance degradation in a key app is detected, shared telemetry allows engineering to see that the latency originates in an overloaded API gateway, not the database or underlying infrastructure. It's a simple example, but one that highlights how service mapping can triage, so people can focus on resolution, not reaction.

Get Alerting Under Control: False alerts drain engineering focus and erode trust in monitoring systems. Mature organizations address this by implementing adaptive thresholding, which dynamically adjusts alert parameters based on historical trends, system baselines, and seasonality. For example, instead of triggering dozens of CPU utilization alerts during routine batch processing every night, adaptive thresholding automatically adjusts expectations based on historical behavior. Managing alert suppression without removing early indicators of degradation is as much about data discipline as it is about process. When thresholds, alerts, and suppression logic are governed transparently and evolve with the environment, organizations build the foundation of data needed for higher levels of maturity and ultimately, AI readiness.

Lay the Foundation for Good Data That Reaps AI Benefits: Nearly half of respondents (48%) cite poor data quality as a barrier to achieving AI readiness. When engineering teams align on common data models, standardized collection practices, and comprehensive data coverage that reflects the full complexity of their environments, they establish consistent, reliable inputs for AI systems. The future belongs to organizations that can aggregate and contextualize all machine data from traditional homegrown applications, commercial off-the-shelf applications to environmental signals like temperature, vibration, and motion. When every data source speaks a common language, AI systems will be the catalyst for a new era of operational intelligence grounded in the full reality of the enterprise.

Embrace Forward Looking Architectures: The next evolution of observability is about building architectures that can adapt as fast as the systems they monitor. Organizations are investing in open and extensible technologies such as OpenTelemetry, code profiling, and observability as code to future-proof their data strategy. These approaches establish portability across environments, reduce vendor dependency, and embed observability into the software delivery lifecycle itself. OpenTelemetry, for example, is quickly becoming the industry standard for collecting, normalizing, and enriching telemetry data across hybrid and multicloud ecosystems. By adopting it early, teams can ensure consistency in how data is defined and exchanged, which sets the stage for complementary frameworks like Machine Communication Protocol (MCP). Together, these standards will be the future of advanced analytics, AI workflows, and autonomous operational systems.

Image
Splunk

Realizing Tangible Business Growth

When organizations take an innovative and responsible approach to observability, they create a foundation of agility and resilience that enables them to thrive through disruption and change. While the pace of innovation accelerates, the anchors of business success remain constant: building exceptional products, elevating customer experiences, and delivering measurable ROI that strengthens the bottom line.

In a world defined by data and driven by AI, observability is no longer just about visibility. It's now about vision.

Mimi Shalash is Observability Advisor at Splunk, a Cisco company

The Latest

Most organizations approach OpenTelemetry as a collection of individual tools they need to assemble from scratch. This view misses the bigger picture. OpenTelemetry is a complete telemetry framework with composable components that address specific problems at different stages of organizational maturity. You start with what you need today and adopt additional pieces as your observability practices evolve ...

One of the earliest lessons I learned from architecting throughput-heavy services is that simplicity wins repeatedly: fewer moving parts, loosely coupled execution (fewer synchronous calls), and precise timing metering. You want data and decisions to travel the shortest possible path. The goal is to build a system where every strategy and each line of code (contention is the key metric) complements the decision trees ...

As discussions around AI "autonomous coworkers" accelerate, many industry projections assume that agents will soon operate alongside human staff in making decisions, taking actions, and managing tasks with minimal oversight. But a growing number of critics (including some of the developers building these systems) argue that the industry still has a long way to go to be able to treat AI agents like fully trusted teammates ...

Enterprise AI has entered a transformational phase where, according to Digitate's recently released survey, Agentic AI and the Future of Enterprise IT, companies are moving beyond traditional automation toward Agentic AI systems designed to reason, adapt, and collaborate alongside human teams ...

The numbers back this urgency up. A recent Zapier survey shows that 92% of enterprises now treat AI as a top priority. Leaders want it, and teams are clamoring for it. But if you look closer at the operations of these companies, you see a different picture. The rollout is slow. The results are often delayed. There's a disconnect between what leaders want and what their technical infrastructure can handle ...

Kyndryl's 2025 Readiness Report revealed that 61% of global business and technology leaders report increasing pressure from boards and regulators to prove AI's ROI. As the technology evolves and expectations continue to rise, leaders are compelled to generate and prove impact before scaling further. This will lead to a decisive turning point in 2026 ...

Cloudflare's disruption illustrates how quickly a single provider's issue cascades into widespread exposure. Many organizations don't fully realize how tightly their systems are coupled to thirdparty services, or how quickly availability and security concerns align when those services falter ... You can't avoid these dependencies, but you can understand them ...

If you work with AI, you know this story. A model performs during testing, looks great in early reviews, works perfectly in production and then slowly loses relevance after operating for a while. Everything on the surface looks perfect — pipelines are running, predictions or recommendations are error-free, data quality checks show green; yet outcomes don't meet the ground reality. This pattern often repeats across enterprise AI programs. Take for example, a mid-sized retail banking and wealth-management firm with heavy investments in AI-powered risk analytics, fraud detection and personalized credit-decisioning systems. The model worked well for a while, but transactions increased, so did false positives by 18% ...

Basic uptime is no longer the gold standard. By 2026, network monitoring must do more than report status, it must explain performance in a hybrid-first world. Networks are no longer just static support systems; they are agile, distributed architectures that sit at the very heart of the customer experience and the business outcomes ... The following five trends represent the new standard for network health, providing a blueprint for teams to move from reactive troubleshooting to a proactive, integrated future ...

APMdigest's Predictions Series concludes with 2026 AI Predictions — industry experts offer predictions on how AI and related technologies will evolve and impact business in 2026. Part 5, the final installment, covers AI's impacts on IT teams ...