Keeping Digital Business Running
Network Performance Management for Digital Operations
September 22, 2016

Jim Frey
Kentik

Share this

The importance of digital business operations is now a given, and for good reason. Recently, Pandora announced that it was launching a subscription service and lowering monthly fees, which means that the already huge percentage of its revenues driven by advertising is going to have to increase in order to maintain the top line. It goes without saying that streaming music, like many other ad-driven business models, relies critically on user experience, and user experience relies critically on network performance. So much so that streaming media, gaming and many other such digital service providers have built private CDNs to guarantee that app and ad bits make it to user eyes and ears in a very timely and reliable fashion.

Network performance monitoring (NPM) has been around a long time. Unlike APM, NPM is still in the process of catching up to cloud realities. In May of this year, Gartner analyst Sanjit Ganguli published a research note entitled Network Performance Monitoring Tools Leave Gaps in Cloud Monitoring. It's a fairly biting critique of the NPM space that says, essentially, that the vast majority of current NPM approaches were largely built for a pre-cloud era, and are unable to adapt because of the new complexities brought by decentralization and full stack virtualization. As a result, network managers are left in the lurch when trying to adapt to the realities of digital operations.

NPM had its origins in open-source manual tools such as MRTG, Nagios, and Wireshark, which are still widely available and useful. However, on a commercial basis, traditional NPM approaches came about during the rise of centralized, private enterprise data centers connected by networks that were built to reach campuses and branch offices across an outsourced, yet essentially private IP/MPLS WAN. Applications of this era were developed in a relatively monolithic fashion. This overall architecture meant that there a few, well defined traffic aggregation points, such as the juncture between LAN and WAN at major datacenters and campuses. Enterprise switches and routers deployed in these environments offered span ports, and thus a generation of NPM packet capture (PCAP) appliances were born that could attach to these span ports directly or via a convenient tap or packet broker device. Appliances weren't the exclusive domain of NPM offerings – they were used for many network management and security products and still are – but the majority of packet-centric NPM solutions leverage appliances to achieve scale and PCAP storage objectives.

A funny thing happened though – the cloud. The rise of IaaS, PaaS, and SaaS meant that there was a new breed of alternative for just about every IT infrastructure and application component. Applications becoming more and more distributed and, increasingly, components started living not just in separate containers, VMs and infrastructure clusters, but in separate datacenters, spread out across networks and the Internet. This cloud way of developing, distributing, hosting and communicating established a dramatically altered set of network traffic patterns.

Unfortunately, NPM appliances aren't nearly as helpful in this new reality. In many clouds you don't have a network interface to tap into for sniffing or capturing packets. The proliferation of application components multiplies the communication endpoints.

In addition, digital business means that users aren't necessarily reached across a private WAN, but rather across the Internet.

Finally, appliances are bedeviled by limited storage and compute power, so they can't offer very much depth of analysis without extreme cost impact. With digital business and DevOps practices being so data-driven, being limited to summary reports and a small window of details isn't acceptable anymore, especially when scale-out computing and storage is so readily available.

This change in how the network and Internet interacts with and influences application performance requires a new approach to NPM. NPM for the digital operations era needs to offer a level of flexibility in deployment and cost-effectiveness to allow for broad, comprehensive instrumentation to collect network performance metric data. In addition, the volume of network performance data ingest, depth of storage, and analytical sophistication needs to scale based on today's cloud economics. Fortunately, there are plenty of technology options available to build these capabilities. So while Gartner has rightly identified a gap in NPM, the good news is that the gap can be readily filled.

Jim Frey is VP of Strategic Alliances at Kentik
Share this

The Latest

November 19, 2019

Unexpected and unintentional drops in network quality, so-called network brownouts, cause serious financial damage and frustrate employees. A recent survey sponsored by Netrounds reveals that more than 60% of network brownouts are first discovered by IT’s internal and external customers, or never even reported, instead of being proactively detected by IT organizations ...

November 18, 2019

Digital transformation reaches into every aspect of our work and personal lives, to the point that there is an automatic expectation of 24/7, anywhere availability regarding any organization with an online presence. This environment is ripe for artificial intelligence, so it's no surprise that IT Operations has been an early adopter of AI ...

November 14, 2019

A brief introduction to Applications Performance Monitoring (APM), breaking it down to a few key points, followed by a few important lessons which I have learned over the years ...

November 13, 2019

Research conducted by ServiceNow shows that Gen Zs, now entering the workforce, recognize the promise of technology to improve work experiences, are eager to learn from other generations, and believe they can help older generations be more open‑minded ...

November 12, 2019

We're in the middle of a technology and connectivity revolution, giving us access to infinite digital tools and technologies. Is this multitude of technology solutions empowering us to do our best work, or getting in our way? ...

November 07, 2019

Microservices have become the go-to architectural standard in modern distributed systems. While there are plenty of tools and techniques to architect, manage, and automate the deployment of such distributed systems, issues during troubleshooting still happen at the individual service level, thereby prolonging the time taken to resolve an outage ...

November 06, 2019

A recent APMdigest blog by Jean Tunis provided an excellent background on Application Performance Monitoring (APM) and what it does. A further topic that I wanted to touch on though is the need for good quality data. If you are to get the most out of your APM solution possible, you will need to feed it with the best quality data ...

November 05, 2019

Humans and manual processes can no longer keep pace with network innovation, evolution, complexity, and change. That's why we're hearing more about self-driving networks, self-healing networks, intent-based networking, and other concepts. These approaches collectively belong to a growing focus area called AIOps, which aims to apply automation, AI and ML to support modern network operations ...

November 04, 2019

IT outages happen to companies across the globe, regardless of location, annual revenue or size. Even the most mammoth companies are at risk of downtime. Increasingly over the past few years, high-profile IT outages — defined as when the services or systems a business provides suddenly become unavailable — have ended up splashed across national news headlines ...

October 31, 2019

APM tools are ideal for an application owner or a line of business owner to track the performance of their key applications. But these tools have broader applicability to different stakeholders in an organization. In this blog, we will review the teams and functional departments that can make use of an APM tool and how they could put it to work ...