One point to reinforce is that the digital war room — physical, virtual or hybrid — is not in retreat but in fact is growing in scope to include greater participation from development and security. It's also becoming more proactive, with on average more than 30% of "major incidents" before they impacted business service performance.
The reasons for this added (not diminished) level of relevance will be examined more in depth in my webinar on April 11th (and yes, there will be replays), but generally the answer lies in the fact that improved levels of team efficiency are critical to the future of IT, and the digital war room shines a spotlight on this evolving reality.
In this blog I'm providing a few additional highlights from the insights we got on digital war room organization and processes.
A Few Organizational Insights
One of the questions we asked was directed at finding out whether war rooms, as they evolve, were becoming more organizationally defined, or more sporadic and ad-hoc. The answer was solidly in the "more formalized" category (47%) versus the group with "more ad-hoc teams and processes" (28%). Another 22% indicated that their teams were already solidly formalized and established.
Then, when we evaluated success rates to this mix, we saw that those digital war rooms becoming "more formalized and established" were far more likely to align with digital war room effectiveness than the other groups.
Well defined teams that can be brought together across all domains provide a unique advantage over fragmented, technically isolated teams
If you think about this, it does suggest a contradiction to some of the trendier thinking endorsing multiple teams and more completely decentralized ways of working. But the logic for core consistency is clear. Well defined teams that can be brought together across all domains provide a unique advantage over fragmented, more technically isolated teams when confronting the full gamut of "major incident" possibilities.
And BTW, the average head count for these teams across small, medium and large was about 15. The implication being not that all 15 stakeholders are being activated for every single incident, but there are 15 individuals assigned and available for digital war room decision making on an on-going, as-needed basis. The trend, BTW, is toward growing not shrinking levels of involvement — in large part because of the accelerating need to include development and security professionals. The overall data also showed a significant role in digital war room decision making for non-IT, or business stakeholders.
Having a single organizational owner, also helps to drive war-room efficiencies. Interestingly, "Security/compliance" was in third place for war-room ownership after "ITSM" and the "executive suite." Having senior executive involvement helped, as well. The most prevalent was ongoing "director-level" involvement, but the most effective turned out to be "CIO-level" involvement.
Processes (or Lack Thereof)
In last week's blog, we enumerated the following critical processes that help to define war-room performance:
■ Initial awareness, which is usually driven by events or some other type of automated intelligence, or complaints to the service desk.
■ Response team engagement and coordination, bringing relevant stakeholders together and providing a context for them to work together.
■ Triage and diagnostics, where problems are understood in context and then detailed requirements for remediation can be defined.
■ Remediation, where active fixes to major incidents are made, often through change and configuration management procedures.
■ Validation, in which testing is done to ensure that actions for remediation were successful, ideally from a business impact as well as a purely technical perspective.
In the non-progressive category, we discovered that, based on our data, the average response indicated only a little more than half (2.57) of these processes were defined — a surprising revelation in a rather negative way. When we mapped "success rates" to the number of processes mapped out, however, we did get a reasonable correlation:
■ 3 for the extremely successful
■ 2.5 for the successful
■ 2 for the only marginally successful
The most prevalently defined process was response team coordination — which also turned out to be the most problematic or delay-causing process. In fact, identifying process with delay or problems mapped well to the processes that were most likely to be identified, suggesting that clarifying the reality of what's going on opens the door to realizing what's wrong and how improvements can be made.
Timing is Everything
Time to assemble an effective team, on average, was about 1.5 hours, which could be damaging when a serious outage occurred
We also asked about times associated with these processes. When we asked about the time to assemble an effective team, the average was about 1.5 hours, which could, of course, be meaningfully damaging when a serious outage occurred. When asked about total time to resolution, the average was about six hours, but 20 percent took more than eleven hours. Once again, as an average, this can be concerning for incidents with major business impacts.
These are again, just a few of many highlights from our research.
Don't forget to watch the webinar for a great many more insights.
Read my third and final blog on the digital war room: The Digital War Room in Changing Times: The Impacts of DevOps, Cloud and SecOps
We all know artificial intelligence (AI) is a hot topic — but beyond the buzzword, have you ever wondered how IT departments are actually adopting AI technologies to improve on their operations? ...
How can IT teams focus on the critical events that can impact their business instead of wading through false positives? The emerging discipline of AIOps is a much-needed panacea for detecting patterns, identifying anomalies, and making sense of alerts across hybrid infrastructure ...
In a recent webinar AIOps and IT Analytics at the Crossroads, I was asked several times about the borderline between AIOps and monitoring tools — most particularly application performance monitoring (APM) capabilities. The general direction of the questions was — how are they different? Do you need AIOps if you have APM already? Why should I invest in both? ...
There's no place like the web and smartphones for the holidays. With the biggest shopping season of the year quickly approaching, retailers are gearing up to experience the most traffic their online platforms (web, mobile, IoT) have ever seen. To avoid missing out on millions this holiday season, below are the top five ways developers can keep their apps and websites up and running without a hitch ...
Usage data is multifaceted, with many diverse benefits. Harvesting usage-driven insights effectively requires both good foundational technology and a nimbleness of mind to unify insights across IT's many silos of domains and disciplines. Because of this, leveraging usage-driven insights can in itself become a catalyst for helping IT as a whole transform toward improved efficiencies and enhanced levels of business alignment ...
The requirements to maintain the complete availability and superior performance of your mission-critical workloads is a dynamic process that has never been more challenging. Here are five ways IT teams can measure and guarantee performance-based SLAs in order to increase the value of the infrastructure to the business, and ensure optimal digital performance levels ...
APMdigest asked experts from across the IT industry for their opinions on what IT departments should be monitoring to ensure digital performance. Part 5, the final installment, offers some recommendations you may not have thought about ...
APMdigest asked experts from across the IT industry for their opinions on what IT departments should be monitoring to ensure digital performance. Part 4 covers the infrastructure, including the cloud and the network ...
APMdigest asked experts from across the IT industry for their opinions on what IT departments should be monitoring to ensure digital performance. Part 3 covers the development side ...
APMdigest asked experts from across the IT industry for their opinions on what IT departments should be monitoring to ensure digital performance. Part 2 covers key performance metrics like availability and response time ...