PagerDuty Announces Intelligent Triage and Intelligent Dashboards
September 24, 2019
Share this

PagerDuty announced two new capabilities - Intelligent Triage and Intelligent Dashboards - for teams responsible for reducing the cost and business impact of issues such as slow downs and outages.

Today, it takes companies an average of 80 minutes to coordinate response teams to solve a customer-impacting issue, such as a failed shopping cart or broken web page. PagerDuty’s new solutions are expected to help reduce that to as little as five minutes by providing automation to get the right people working together, with the right information, to more quickly triage issues when seconds count. By adopting real-time digital operations management practices, large companies can gain upwards of $2.5 million in IT staff productivity savings.

Intelligent Triage is a new feature set within PagerDuty’s Event Intelligence product, which uses machine learning to group alerts together so teams don’t receive multiple alerts coming from related issues. Triage provides additional context into the issue; e.g., whether it has happened before, how it was resolved, how widespread it is, what teams and services are affected, who is working on it and how they can be reached. By immediately arming teams with this knowledge, PagerDuty helps organizations pull together the right people, with the right information, to solve problems faster, minimizing the cost of downtime and preventing poor customer experiences.

Intelligent Dashboards — new to PagerDuty’s Analytics product — leverages machine learning to provide teams with recommendations for how to resolve issues, as well as benchmarks against performance metrics from other teams in their organization or vertical industry so they can continually improve. Its Spotlights recommendation engine leverages 10 years of machine and human response data to give teams context for improvements, such as stopping unactionable alerts and recognizing repeat issues.

“Nearly half of companies experience a major technology issue at least monthly,” said PagerDuty’s SVP Product, Jonathan Rende. “In today’s always-on world, slow responses damage a company’s brand, impact employees and erode the bottom line. Companies urgently need insights into how they are handling these issues so they know how to improve. With Spotlights, we are automating the provision of knowledge that is crucial to both solving problems in the moment and continually improving performance.”

Intelligent Triage:

- Provides context into an issue e.g., whether it has happened before, how it was resolved, how widespread it is, what services and teams are affected, who is working on it and how they can be reached.

- Provides automation to ensure teams have the knowledge required to effectively triage issues in real-time (e.g. is this a major incident? Who is needed to help?).

- Reduces the impact of unplanned work by giving adjacent teams visibility so they don’t duplicate efforts or interfere with each other.

- Creates significant time and cost savings — the majority of tech employees will lose 100-plus hours of productivity due to unplanned work this year4.

- Now available for Event Intelligence customers.

Intelligent Dashboards:

- Leverages 10 years of machine data and human response patterns, applied through Spotlight, PagerDuty’s recommendation engine that learns from past issues to make suggestions that teams can use for future improvements, such as stopping unactionable alerts, fixing repeat issues and improving escalation practices.

- Includes interactive charts and graphs that, unlike static status reports, let customers drill into details by team to show incident volume, response effort, interruption volume and more.

- Provides managers with built-in benchmarks to see how their teams compare to peers in the organization and their vertical industry when it comes to spotting issues, mobilizing teams and achieving resolutions.

- Translates the impact of issues into business outcomes, such as total cost of incidents or response team fatigue where other solutions only have basic metrics, such as mean time to response (MTTR).

Share this

The Latest

October 17, 2019

As the data generated by organizations grows, APM tools are now required to do a lot more than basic monitoring of metrics. Modern data is often raw and unstructured and requires more advanced methods of analysis. The tools must help dig deep into this data for both forensic analysis and predictive analysis. To extract more accurate and cheaper insights, modern APM tools use Big Data techniques to store, access, and analyze the multi-dimensional data ...

October 16, 2019

Modern enterprises are generating data at an unprecedented rate but aren't taking advantage of all the data available to them in order to drive real-time, actionable insights. According to a recent study commissioned by Actian, more than half of enterprises today are unable to efficiently manage nor effectively use data to drive decision-making ...

October 15, 2019

According to a study by Forrester Research, an enhanced UX design can increase the conversion rate by 400%. If UX has become the ultimate arbiter in determining the success or failure of a product or service, let us first understand what UX is all about ...

October 10, 2019

The requirements of an APM tool are now much more complex than they've ever been. Not only do they need to trace a user transaction across numerous microservices on the same system, but they also need to happen pretty fast ...

October 09, 2019

Performance monitoring is an old problem. As technology has advanced, we've had to evolve how we monitor applications. Initially, performance monitoring largely involved sending ICMP messages to start troubleshooting a down or slow application. Applications have gotten much more complex, so this is no longer enough. Now we need to know not just whether an application is broken, but why it broke. So APM has had to evolve over the years for us to get there. But how did this evolution take place, and what happens next? Let's find out ...

October 08, 2019

There are some IT organizations that are using DevOps methodology but are wary of getting bogged down in ITSM procedures. But without at least some ITSM controls in place, organizations lose their focus on systematic customer engagement, making it harder for them to scale ...

October 07, 2019
OK, I admit it. "Service modeling" is an awkward term, especially when you're trying to frame three rather controversial acronyms in the same overall place: CMDB, CMS and DDM. Nevertheless, that's exactly what we did in EMA's most recent research: <span style="font-style: italic;">Service Modeling in the Age of Cloud and Containers</span>. The goal was to establish a more holistic context for looking at the synergies and differences across all these areas ...
October 03, 2019

If you have deployed a Java application in production, you've probably encountered a situation where the application suddenly starts to take up a large amount of CPU. When this happens, application response becomes sluggish and users begin to complain about slow response. Often the solution to this problem is to restart the application and, lo and behold, the problem goes away — only to reappear a few days later. A key question then is: how to troubleshoot high CPU usage of a Java application? ...

October 02, 2019

Operations are no longer tethered tightly to a main office, as the headquarters-centric model has been retired in favor of a more decentralized enterprise structure. Rather than focus the business around a single location, enterprises are now comprised of a web of remote offices and individuals, where network connectivity has broken down the geographic barriers that in the past limited the availability of talent and resources. Key to the success of the decentralized enterprise model is a new generation of collaboration and communication tools ...

October 01, 2019

To better understand the AI maturity of businesses, Dotscience conducted a survey of 500 industry professionals. Research findings indicate that although enterprises are dedicating significant time and resources towards their AI deployments, many data science and ML teams don't have the adequate tools needed to properly collaborate on, build and deploy AI models efficiently ...