Skip to main content

Grafana Labs Launches Mimir 3.0

Grafana Labs announced the launch of Grafana Mimir 3.0, the latest evolution of its open-source, horizontally scalable metrics backend. 

Mimir 3.0 marks an architectural milestone, delivering new levels of reliability, performance, and cost efficiency for Prometheus-compatible monitoring at enterprise scale.

Built on open source, open standards, and open ecosystems, Grafana Labs helps organizations innovate without lock-in and move fast without compromise. At KubeCon, the company also announced updates across its open-source ecosystem, including Grafana Tempo 2.9 with AI-assisted tracing, continued Kubernetes Monitoring enhancements, and deeper Prometheus and OpenTelemetry support to help teams simplify observability and gain more value from their data.

“From open source and open standards to open ecosystems and open minds – building in the open is core to our philosophy at Grafana Labs,” said Myrle Krantz, Senior Director of Engineering, Grafana Labs. “That’s why we’re continuing to invest in open source, like adding AI-assisted tracing in Tempo and making it easier to get the most out of OpenTelemetry and Prometheus. We’re continuously improving Kubernetes Monitoring. And with the new Mimir 3.0 release, we’re helping teams scale even more reliably, expanding what’s possible for open observability in 2026 and beyond.”

Three years in development, Grafana Mimir 3.0 introduces a new decoupled architecture that separates the read and write paths for more reliable, large-scale metrics operations:

  • Reliability: By decoupling reads and writes through an asynchronous Kafka-based ingest layer, cross-path dependencies are eliminated, keeping queries fast and stable even under heavy ingestion loads.
  • Performance: The new Mimir Query Engine (MQE) streams query results instead of loading entire datasets into memory, improving execution speed and reducing memory usage by up to 92%.
  • Cost efficiency: Early testing reports up to 15% lower resource usage while achieving higher throughput and consistency across large clusters.

Together, these innovations make Mimir 3.0 the most resilient, high-performing, and cost-efficient metrics backend for Prometheus and OpenTelemetry data – now available on Grafana Cloud and for self-managed users via open source. 

The latest release of Grafana Tempo, the open source distributed tracing backend, introduces new capabilities to speed up trace analysis and bring AI into the observability workflow.

  • MCP server support: An experimental Model Context Protocol (MCP) server allows AI assistants like Claude Code and Cursor to query distributed tracing data with TraceQL, enabling natural-language debugging and faster root cause analysis.
  • TraceQL metrics sampling: New probabilistic query hints accelerate analysis in high-volume environments, returning approximate results faster without losing visibility.
  • Multi-tenant and operational improvements: New metrics for query I/O, span timing, and usage tracking improve observability and performance visibility at scale.

Tempo 2.9 also deepens OpenTelemetry support by aligning with newer OpenTelemetry semantic conventions, reaffirming Grafana Labs’ commitment to open, composable observability.

Building on the success of Grafana Cloud Kubernetes Monitoring, Grafana Labs has introduced powerful new capabilities that simplify observability across even the most complex Kubernetes environments. This is especially timely as a recent survey by the CNCF found that 80% of respondents work for IT organizations that have deployed Kubernetes in a production environment.

Kubernetes Monitoring in Grafana Cloud has evolved into an observability solution that doesn’t just visualize telemetry but interprets it, automates insights, and guides teams to action. New updates include:

  • Grafana AI Assistant integration: Teams can now interact with Kubernetes Monitoring using Grafana Assistant (now generally available), an AI-powered agent built into Grafana Cloud that can read dashboards, drill into panels, and summarize results in real time. Using natural language, users can ask how a workload is behaving, what’s impacting performance, or where costs are trending.
  • GPU monitoring: Available at both the Node and Cluster level, new GPU utilization panels help detect overheating, power drain, or underuse before they impact performance, ensuring AI workloads remain stable and efficient.
  • Automated root cause analysis: Now integrated with the generally available Grafana Knowledge Graph, Kubernetes Monitoring gives you automatic RCA and Insight Rings.
  • Expanded workload support: Kubernetes Monitoring now provides full visibility into CronJobs, Argo Rollouts, Bare Pods, Static Pods, Strimzi Pod Sets, and other nonstandard workloads, ensuring comprehensive coverage across diverse infrastructure types.
  • Monitor cron jobs and other job types: Get full visibility into all cron and manual jobs across clusters. Instantly see status, distribution, and missed runs to ensure automation reliability and quick issue detection.
  • CPU and memory panels: New CPU and Memory tabs provide clear, layered views of compute usage – from cluster to container – with efficiency graphs and CPU distribution analysis that help optimize capacity, cost, and performance.
  • Cloud provider nodes: One-click correlation between AWS EC2 instances and Kubernetes workloads enables unified troubleshooting across cloud and container layers, reducing context-switching and mean time to resolution. And for teams on AWS, CloudWatch metric streams in Grafana Cloud can cut metric pipeline costs, including storage and agent infrastructure, by up to 10x while delivering near-real-time metrics.

Together, these updates make Kubernetes Monitoring in Grafana Cloud an intelligent, automated, and AI-capable solution for today’s dynamic, large-scale environments.

Grafana Labs continues to invest in open standards and community-led innovation across its ecosystem:

  • Beyla donation complete: Earlier in 2025, Grafana Labs donated Grafana Beyla, its eBPF-based, zero-code auto-instrumentation agent, to OpenTelemetry. Renamed OpenTelemetry eBPF Instrumentation, the project just marked its first official release under the OpenTelemetry umbrella. The donation reinforces Grafana Labs’ long-standing commitment to advancing open, vendor-neutral observability.
  • Grafana Alloy: Grafana Labs’ distribution of the OpenTelemetry Collector, Grafana Alloy is now the default data pipeline layer across Grafana Cloud and open source deployments. Alloy unifies metrics, logs, and traces collection while supporting both Prometheus and OpenTelemetry pipelines.
  • Prometheus 3.0 and OpenTelemetry interoperability: Grafana engineers contributed to the introduction of profiling signal support, new semantic conventions, and Prometheus 3.0 compatibility, strengthening cross-project interoperability.

The Latest

Enterprises are under pressure to scale AI quickly. Yet despite considerable investment, adoption continues to stall. One of the most overlooked reasons is vendor sprawl ... In reality, no organization deliberately sets out to create sprawling vendor ecosystems. More often, complexity accumulates over time through well-intentioned initiatives, such as enterprise-wide digital transformation efforts, point solutions, or decentralized sourcing strategies ...

Nearly every conversation about AI eventually circles back to compute. GPUs dominate the headlines while cloud platforms compete for workloads and model benchmarks drive investment decisions. But underneath that noise, a quieter infrastructure challenge is taking shape. The real bottleneck in enterprise AI is not processing power, it is the ability to store, manage and retrieve the relentless volumes of data that AI systems generate, consume and multiply ...

The 2026 Observability Survey from Grafana Labs paints a vivid picture of an industry maturing fast, where AI is welcomed with careful conditions, SaaS economics are reshaping spending decisions, complexity remains a defining challenge, and open standards continue to underpin it all ...

The observability industry has an evolving relationship with AI. We're not skeptics, but it's clear that trust in AI must be earned ... In Grafana Labs' annual Observability Survey, 92% said they see real value in AI surfacing anomalies before they cause downtime. Another 91% endorsed AI for forecasting and root cause analysis. So while the demand is there, customers need it to be trustworthy, as the survey also found that the practitioners most enthusiastic about AI are also the most insistent on explainability ...

In the modern enterprise, the conversation around AI has moved past skepticism toward a stage of active adoption. According to our 2026 State of IT Trends Report: The Human Side of Autonomous AI, nearly 90% of IT professionals view AI as a net positive, and this optimism is well-founded. We are seeing agentic AI move beyond simple automation to actively streamlining complex data insights and eliminating the manual toil that has long hindered innovation. However, as we integrate these autonomous agents into our ecosystems, the fundamental DNA of the IT role is evolving ...

AI workloads require an enormous amount of computing power ... What's also becoming abundantly clear is just how quickly AI's computing needs are leading to enterprise systems failure. According to Cockroach Labs' State of AI Infrastructure 2026 report, enterprise systems are much closer to failure than their organizations realize. The report ... suggests AI scale could cause widespread failures in as little as one year — making it a clear risk for business performance and reliability.

The quietest week your engineering team has ever had might also be its best. No alarms going off. No escalations. No frantic Teams or Slack threads at 2 a.m. Everything humming along exactly as it should. And somewhere in a leadership meeting, someone looks at the metrics dashboard, sees a flat line of incidents and says: "Seems like things are pretty calm over there. Do we really need all those people?" ... I've spent many years in engineering, and this pattern keeps repeating ...

The gap is widening between what teams spend on observability tools and the value they receive amid surging data volumes and budget pressures, according to The Breaking Point for Observability Leaders, a report from Imply ...

Seamless shopping is a basic demand of today's boundaryless consumer — one with little patience for friction, limited tolerance for disconnected experiences and minimal hesitation in switching brands. Customers expect intuitive, highly personalized experiences and the ability to move effortlessly across physical and digital channels within the same journey. Failure to deliver can cost dearly ...

If your best engineers spend their days sorting tickets and resetting access, you are wasting talent. New global data shows that employees in the IT sector rank among the least motivated across industries. They're under a lot of pressure from many angles. Pressure to upskill and uncertainty around what agentic AI means for job security is creating anxiety. Meanwhile, these roles often function like an on-call job and require many repetitive tasks ...

Grafana Labs Launches Mimir 3.0

Grafana Labs announced the launch of Grafana Mimir 3.0, the latest evolution of its open-source, horizontally scalable metrics backend. 

Mimir 3.0 marks an architectural milestone, delivering new levels of reliability, performance, and cost efficiency for Prometheus-compatible monitoring at enterprise scale.

Built on open source, open standards, and open ecosystems, Grafana Labs helps organizations innovate without lock-in and move fast without compromise. At KubeCon, the company also announced updates across its open-source ecosystem, including Grafana Tempo 2.9 with AI-assisted tracing, continued Kubernetes Monitoring enhancements, and deeper Prometheus and OpenTelemetry support to help teams simplify observability and gain more value from their data.

“From open source and open standards to open ecosystems and open minds – building in the open is core to our philosophy at Grafana Labs,” said Myrle Krantz, Senior Director of Engineering, Grafana Labs. “That’s why we’re continuing to invest in open source, like adding AI-assisted tracing in Tempo and making it easier to get the most out of OpenTelemetry and Prometheus. We’re continuously improving Kubernetes Monitoring. And with the new Mimir 3.0 release, we’re helping teams scale even more reliably, expanding what’s possible for open observability in 2026 and beyond.”

Three years in development, Grafana Mimir 3.0 introduces a new decoupled architecture that separates the read and write paths for more reliable, large-scale metrics operations:

  • Reliability: By decoupling reads and writes through an asynchronous Kafka-based ingest layer, cross-path dependencies are eliminated, keeping queries fast and stable even under heavy ingestion loads.
  • Performance: The new Mimir Query Engine (MQE) streams query results instead of loading entire datasets into memory, improving execution speed and reducing memory usage by up to 92%.
  • Cost efficiency: Early testing reports up to 15% lower resource usage while achieving higher throughput and consistency across large clusters.

Together, these innovations make Mimir 3.0 the most resilient, high-performing, and cost-efficient metrics backend for Prometheus and OpenTelemetry data – now available on Grafana Cloud and for self-managed users via open source. 

The latest release of Grafana Tempo, the open source distributed tracing backend, introduces new capabilities to speed up trace analysis and bring AI into the observability workflow.

  • MCP server support: An experimental Model Context Protocol (MCP) server allows AI assistants like Claude Code and Cursor to query distributed tracing data with TraceQL, enabling natural-language debugging and faster root cause analysis.
  • TraceQL metrics sampling: New probabilistic query hints accelerate analysis in high-volume environments, returning approximate results faster without losing visibility.
  • Multi-tenant and operational improvements: New metrics for query I/O, span timing, and usage tracking improve observability and performance visibility at scale.

Tempo 2.9 also deepens OpenTelemetry support by aligning with newer OpenTelemetry semantic conventions, reaffirming Grafana Labs’ commitment to open, composable observability.

Building on the success of Grafana Cloud Kubernetes Monitoring, Grafana Labs has introduced powerful new capabilities that simplify observability across even the most complex Kubernetes environments. This is especially timely as a recent survey by the CNCF found that 80% of respondents work for IT organizations that have deployed Kubernetes in a production environment.

Kubernetes Monitoring in Grafana Cloud has evolved into an observability solution that doesn’t just visualize telemetry but interprets it, automates insights, and guides teams to action. New updates include:

  • Grafana AI Assistant integration: Teams can now interact with Kubernetes Monitoring using Grafana Assistant (now generally available), an AI-powered agent built into Grafana Cloud that can read dashboards, drill into panels, and summarize results in real time. Using natural language, users can ask how a workload is behaving, what’s impacting performance, or where costs are trending.
  • GPU monitoring: Available at both the Node and Cluster level, new GPU utilization panels help detect overheating, power drain, or underuse before they impact performance, ensuring AI workloads remain stable and efficient.
  • Automated root cause analysis: Now integrated with the generally available Grafana Knowledge Graph, Kubernetes Monitoring gives you automatic RCA and Insight Rings.
  • Expanded workload support: Kubernetes Monitoring now provides full visibility into CronJobs, Argo Rollouts, Bare Pods, Static Pods, Strimzi Pod Sets, and other nonstandard workloads, ensuring comprehensive coverage across diverse infrastructure types.
  • Monitor cron jobs and other job types: Get full visibility into all cron and manual jobs across clusters. Instantly see status, distribution, and missed runs to ensure automation reliability and quick issue detection.
  • CPU and memory panels: New CPU and Memory tabs provide clear, layered views of compute usage – from cluster to container – with efficiency graphs and CPU distribution analysis that help optimize capacity, cost, and performance.
  • Cloud provider nodes: One-click correlation between AWS EC2 instances and Kubernetes workloads enables unified troubleshooting across cloud and container layers, reducing context-switching and mean time to resolution. And for teams on AWS, CloudWatch metric streams in Grafana Cloud can cut metric pipeline costs, including storage and agent infrastructure, by up to 10x while delivering near-real-time metrics.

Together, these updates make Kubernetes Monitoring in Grafana Cloud an intelligent, automated, and AI-capable solution for today’s dynamic, large-scale environments.

Grafana Labs continues to invest in open standards and community-led innovation across its ecosystem:

  • Beyla donation complete: Earlier in 2025, Grafana Labs donated Grafana Beyla, its eBPF-based, zero-code auto-instrumentation agent, to OpenTelemetry. Renamed OpenTelemetry eBPF Instrumentation, the project just marked its first official release under the OpenTelemetry umbrella. The donation reinforces Grafana Labs’ long-standing commitment to advancing open, vendor-neutral observability.
  • Grafana Alloy: Grafana Labs’ distribution of the OpenTelemetry Collector, Grafana Alloy is now the default data pipeline layer across Grafana Cloud and open source deployments. Alloy unifies metrics, logs, and traces collection while supporting both Prometheus and OpenTelemetry pipelines.
  • Prometheus 3.0 and OpenTelemetry interoperability: Grafana engineers contributed to the introduction of profiling signal support, new semantic conventions, and Prometheus 3.0 compatibility, strengthening cross-project interoperability.

The Latest

Enterprises are under pressure to scale AI quickly. Yet despite considerable investment, adoption continues to stall. One of the most overlooked reasons is vendor sprawl ... In reality, no organization deliberately sets out to create sprawling vendor ecosystems. More often, complexity accumulates over time through well-intentioned initiatives, such as enterprise-wide digital transformation efforts, point solutions, or decentralized sourcing strategies ...

Nearly every conversation about AI eventually circles back to compute. GPUs dominate the headlines while cloud platforms compete for workloads and model benchmarks drive investment decisions. But underneath that noise, a quieter infrastructure challenge is taking shape. The real bottleneck in enterprise AI is not processing power, it is the ability to store, manage and retrieve the relentless volumes of data that AI systems generate, consume and multiply ...

The 2026 Observability Survey from Grafana Labs paints a vivid picture of an industry maturing fast, where AI is welcomed with careful conditions, SaaS economics are reshaping spending decisions, complexity remains a defining challenge, and open standards continue to underpin it all ...

The observability industry has an evolving relationship with AI. We're not skeptics, but it's clear that trust in AI must be earned ... In Grafana Labs' annual Observability Survey, 92% said they see real value in AI surfacing anomalies before they cause downtime. Another 91% endorsed AI for forecasting and root cause analysis. So while the demand is there, customers need it to be trustworthy, as the survey also found that the practitioners most enthusiastic about AI are also the most insistent on explainability ...

In the modern enterprise, the conversation around AI has moved past skepticism toward a stage of active adoption. According to our 2026 State of IT Trends Report: The Human Side of Autonomous AI, nearly 90% of IT professionals view AI as a net positive, and this optimism is well-founded. We are seeing agentic AI move beyond simple automation to actively streamlining complex data insights and eliminating the manual toil that has long hindered innovation. However, as we integrate these autonomous agents into our ecosystems, the fundamental DNA of the IT role is evolving ...

AI workloads require an enormous amount of computing power ... What's also becoming abundantly clear is just how quickly AI's computing needs are leading to enterprise systems failure. According to Cockroach Labs' State of AI Infrastructure 2026 report, enterprise systems are much closer to failure than their organizations realize. The report ... suggests AI scale could cause widespread failures in as little as one year — making it a clear risk for business performance and reliability.

The quietest week your engineering team has ever had might also be its best. No alarms going off. No escalations. No frantic Teams or Slack threads at 2 a.m. Everything humming along exactly as it should. And somewhere in a leadership meeting, someone looks at the metrics dashboard, sees a flat line of incidents and says: "Seems like things are pretty calm over there. Do we really need all those people?" ... I've spent many years in engineering, and this pattern keeps repeating ...

The gap is widening between what teams spend on observability tools and the value they receive amid surging data volumes and budget pressures, according to The Breaking Point for Observability Leaders, a report from Imply ...

Seamless shopping is a basic demand of today's boundaryless consumer — one with little patience for friction, limited tolerance for disconnected experiences and minimal hesitation in switching brands. Customers expect intuitive, highly personalized experiences and the ability to move effortlessly across physical and digital channels within the same journey. Failure to deliver can cost dearly ...

If your best engineers spend their days sorting tickets and resetting access, you are wasting talent. New global data shows that employees in the IT sector rank among the least motivated across industries. They're under a lot of pressure from many angles. Pressure to upskill and uncertainty around what agentic AI means for job security is creating anxiety. Meanwhile, these roles often function like an on-call job and require many repetitive tasks ...