Skip to main content

Keeping Digital Business Running

Network Performance Management for Digital Operations
Jim Frey

The importance of digital business operations is now a given, and for good reason. Recently, Pandora announced that it was launching a subscription service and lowering monthly fees, which means that the already huge percentage of its revenues driven by advertising is going to have to increase in order to maintain the top line. It goes without saying that streaming music, like many other ad-driven business models, relies critically on user experience, and user experience relies critically on network performance. So much so that streaming media, gaming and many other such digital service providers have built private CDNs to guarantee that app and ad bits make it to user eyes and ears in a very timely and reliable fashion.

Network performance monitoring (NPM) has been around a long time. Unlike APM, NPM is still in the process of catching up to cloud realities. In May of this year, Gartner analyst Sanjit Ganguli published a research note entitled Network Performance Monitoring Tools Leave Gaps in Cloud Monitoring. It's a fairly biting critique of the NPM space that says, essentially, that the vast majority of current NPM approaches were largely built for a pre-cloud era, and are unable to adapt because of the new complexities brought by decentralization and full stack virtualization. As a result, network managers are left in the lurch when trying to adapt to the realities of digital operations.

NPM had its origins in open-source manual tools such as MRTG, Nagios, and Wireshark, which are still widely available and useful. However, on a commercial basis, traditional NPM approaches came about during the rise of centralized, private enterprise data centers connected by networks that were built to reach campuses and branch offices across an outsourced, yet essentially private IP/MPLS WAN. Applications of this era were developed in a relatively monolithic fashion. This overall architecture meant that there a few, well defined traffic aggregation points, such as the juncture between LAN and WAN at major datacenters and campuses. Enterprise switches and routers deployed in these environments offered span ports, and thus a generation of NPM packet capture (PCAP) appliances were born that could attach to these span ports directly or via a convenient tap or packet broker device. Appliances weren't the exclusive domain of NPM offerings – they were used for many network management and security products and still are – but the majority of packet-centric NPM solutions leverage appliances to achieve scale and PCAP storage objectives.

A funny thing happened though – the cloud. The rise of IaaS, PaaS, and SaaS meant that there was a new breed of alternative for just about every IT infrastructure and application component. Applications becoming more and more distributed and, increasingly, components started living not just in separate containers, VMs and infrastructure clusters, but in separate datacenters, spread out across networks and the Internet. This cloud way of developing, distributing, hosting and communicating established a dramatically altered set of network traffic patterns.

Unfortunately, NPM appliances aren't nearly as helpful in this new reality. In many clouds you don't have a network interface to tap into for sniffing or capturing packets. The proliferation of application components multiplies the communication endpoints.

In addition, digital business means that users aren't necessarily reached across a private WAN, but rather across the Internet.

Finally, appliances are bedeviled by limited storage and compute power, so they can't offer very much depth of analysis without extreme cost impact. With digital business and DevOps practices being so data-driven, being limited to summary reports and a small window of details isn't acceptable anymore, especially when scale-out computing and storage is so readily available.

This change in how the network and Internet interacts with and influences application performance requires a new approach to NPM. NPM for the digital operations era needs to offer a level of flexibility in deployment and cost-effectiveness to allow for broad, comprehensive instrumentation to collect network performance metric data. In addition, the volume of network performance data ingest, depth of storage, and analytical sophistication needs to scale based on today's cloud economics. Fortunately, there are plenty of technology options available to build these capabilities. So while Gartner has rightly identified a gap in NPM, the good news is that the gap can be readily filled.

Hot Topics

The Latest

While 87% of manufacturing leaders and technical specialists report that ROI from their AIOps initiatives has met or exceeded expectations, only 37% say they are fully prepared to operationalize AI at scale, according to The Future of IT Operations in the AI Era, a report from Riverbed ...

Many organizations rely on cloud-first architectures to aggregate, analyze, and act on their operational data ... However, not all environments are conducive to cloud-first architectures ... There are limitations to cloud-first architectures that render them ineffective in mission-critical situations where responsiveness, cost control, and data sovereignty are non-negotiable; these limitations include ...

For years, cybersecurity was built around a simple assumption: protect the physical network and trust everything inside it. That model made sense when employees worked in offices, applications lived in data centers, and devices rarely left the building. Today's reality is fluid: people work from everywhere, applications run across multiple clouds, and AI-driven agents are beginning to act on behalf of users. But while the old perimeter dissolved, a new one quietly emerged ...

For years, infrastructure teams have treated compute as a relatively stable input. Capacity was provisioned, costs were forecasted, and performance expectations were set based on the assumption that identical resources behaved identically. That mental model is starting to break down. AI infrastructure is no longer behaving like static cloud capacity. It is increasingly behaving like a market ...

Resilience can no longer be defined by how quickly an organization recovers from an incident or disruption. The effectiveness of any resilience strategy is dependent on its ability to anticipate change, operate under continuous stress, and adapt confidently amid uncertainty ...

Mobile users are less tolerant of app instability than ever before. According to a new report from Luciq, No Margin for Error: What Mobile Users Expect and What Mobile Leaders Must Deliver in 2026, even minor performance issues now result in immediate abandonment, lost purchases, and long-term brand impact ...

Artificial intelligence (AI) has become the dominant force shaping enterprise data strategies. Boards expect progress. Executives expect returns. And data leaders are under pressure to prove that their organizations are "AI-ready" ...

Agentic AI is a major buzzword for 2026. Many tech companies are making bold promises about this technology, but many aren't grounded in reality, at least not yet. This coming year will likely be shaped by reality checks for IT teams, and progress will only come from a focus on strong foundations and disciplined execution ...

AI systems are still prone to hallucinations and misjudgments ... To build the trust needed for adoption, AI must be paired with human-in-the-loop (HITL) oversight, or checkpoints where humans verify, guide, and decide what actions are taken. The balance between autonomy and accountability is what will allow AI to deliver on its promise without sacrificing human trust ...

More data center leaders are reducing their reliance on utility grids by investing in onsite power for rapidly scaling data centers, according to the Data Center Power Report from Bloom Energy ...

Keeping Digital Business Running

Network Performance Management for Digital Operations
Jim Frey

The importance of digital business operations is now a given, and for good reason. Recently, Pandora announced that it was launching a subscription service and lowering monthly fees, which means that the already huge percentage of its revenues driven by advertising is going to have to increase in order to maintain the top line. It goes without saying that streaming music, like many other ad-driven business models, relies critically on user experience, and user experience relies critically on network performance. So much so that streaming media, gaming and many other such digital service providers have built private CDNs to guarantee that app and ad bits make it to user eyes and ears in a very timely and reliable fashion.

Network performance monitoring (NPM) has been around a long time. Unlike APM, NPM is still in the process of catching up to cloud realities. In May of this year, Gartner analyst Sanjit Ganguli published a research note entitled Network Performance Monitoring Tools Leave Gaps in Cloud Monitoring. It's a fairly biting critique of the NPM space that says, essentially, that the vast majority of current NPM approaches were largely built for a pre-cloud era, and are unable to adapt because of the new complexities brought by decentralization and full stack virtualization. As a result, network managers are left in the lurch when trying to adapt to the realities of digital operations.

NPM had its origins in open-source manual tools such as MRTG, Nagios, and Wireshark, which are still widely available and useful. However, on a commercial basis, traditional NPM approaches came about during the rise of centralized, private enterprise data centers connected by networks that were built to reach campuses and branch offices across an outsourced, yet essentially private IP/MPLS WAN. Applications of this era were developed in a relatively monolithic fashion. This overall architecture meant that there a few, well defined traffic aggregation points, such as the juncture between LAN and WAN at major datacenters and campuses. Enterprise switches and routers deployed in these environments offered span ports, and thus a generation of NPM packet capture (PCAP) appliances were born that could attach to these span ports directly or via a convenient tap or packet broker device. Appliances weren't the exclusive domain of NPM offerings – they were used for many network management and security products and still are – but the majority of packet-centric NPM solutions leverage appliances to achieve scale and PCAP storage objectives.

A funny thing happened though – the cloud. The rise of IaaS, PaaS, and SaaS meant that there was a new breed of alternative for just about every IT infrastructure and application component. Applications becoming more and more distributed and, increasingly, components started living not just in separate containers, VMs and infrastructure clusters, but in separate datacenters, spread out across networks and the Internet. This cloud way of developing, distributing, hosting and communicating established a dramatically altered set of network traffic patterns.

Unfortunately, NPM appliances aren't nearly as helpful in this new reality. In many clouds you don't have a network interface to tap into for sniffing or capturing packets. The proliferation of application components multiplies the communication endpoints.

In addition, digital business means that users aren't necessarily reached across a private WAN, but rather across the Internet.

Finally, appliances are bedeviled by limited storage and compute power, so they can't offer very much depth of analysis without extreme cost impact. With digital business and DevOps practices being so data-driven, being limited to summary reports and a small window of details isn't acceptable anymore, especially when scale-out computing and storage is so readily available.

This change in how the network and Internet interacts with and influences application performance requires a new approach to NPM. NPM for the digital operations era needs to offer a level of flexibility in deployment and cost-effectiveness to allow for broad, comprehensive instrumentation to collect network performance metric data. In addition, the volume of network performance data ingest, depth of storage, and analytical sophistication needs to scale based on today's cloud economics. Fortunately, there are plenty of technology options available to build these capabilities. So while Gartner has rightly identified a gap in NPM, the good news is that the gap can be readily filled.

Hot Topics

The Latest

While 87% of manufacturing leaders and technical specialists report that ROI from their AIOps initiatives has met or exceeded expectations, only 37% say they are fully prepared to operationalize AI at scale, according to The Future of IT Operations in the AI Era, a report from Riverbed ...

Many organizations rely on cloud-first architectures to aggregate, analyze, and act on their operational data ... However, not all environments are conducive to cloud-first architectures ... There are limitations to cloud-first architectures that render them ineffective in mission-critical situations where responsiveness, cost control, and data sovereignty are non-negotiable; these limitations include ...

For years, cybersecurity was built around a simple assumption: protect the physical network and trust everything inside it. That model made sense when employees worked in offices, applications lived in data centers, and devices rarely left the building. Today's reality is fluid: people work from everywhere, applications run across multiple clouds, and AI-driven agents are beginning to act on behalf of users. But while the old perimeter dissolved, a new one quietly emerged ...

For years, infrastructure teams have treated compute as a relatively stable input. Capacity was provisioned, costs were forecasted, and performance expectations were set based on the assumption that identical resources behaved identically. That mental model is starting to break down. AI infrastructure is no longer behaving like static cloud capacity. It is increasingly behaving like a market ...

Resilience can no longer be defined by how quickly an organization recovers from an incident or disruption. The effectiveness of any resilience strategy is dependent on its ability to anticipate change, operate under continuous stress, and adapt confidently amid uncertainty ...

Mobile users are less tolerant of app instability than ever before. According to a new report from Luciq, No Margin for Error: What Mobile Users Expect and What Mobile Leaders Must Deliver in 2026, even minor performance issues now result in immediate abandonment, lost purchases, and long-term brand impact ...

Artificial intelligence (AI) has become the dominant force shaping enterprise data strategies. Boards expect progress. Executives expect returns. And data leaders are under pressure to prove that their organizations are "AI-ready" ...

Agentic AI is a major buzzword for 2026. Many tech companies are making bold promises about this technology, but many aren't grounded in reality, at least not yet. This coming year will likely be shaped by reality checks for IT teams, and progress will only come from a focus on strong foundations and disciplined execution ...

AI systems are still prone to hallucinations and misjudgments ... To build the trust needed for adoption, AI must be paired with human-in-the-loop (HITL) oversight, or checkpoints where humans verify, guide, and decide what actions are taken. The balance between autonomy and accountability is what will allow AI to deliver on its promise without sacrificing human trust ...

More data center leaders are reducing their reliance on utility grids by investing in onsite power for rapidly scaling data centers, according to the Data Center Power Report from Bloom Energy ...