Application Performance Problems? It's Not Always the Network!
A primer on how to win the application versus network argument
August 15, 2014

Don Thomas Jacob
SolarWinds

Share this

“It must be the network!” Network administrators hear this phrase all too often when an application is slow, data transfer is not fast enough or VoIP calls drop. Now, of course, the network is the underlying infrastructure all of these services run on, so if something does not work as expected it’s understandable that users more often than not place the blame on the network.

And sometimes that blame is rightfully placed on the network. It may indeed be that there isn’t enough bandwidth provisioned for the WAN, non-business traffic is hogging bandwidth, there are issues with high latency or there is incorrect or no QoS priority. Route flaps, the health of network devices or configuration mistakes can all also lead to application performance problems and are related to the network. Despite these potential problem areas, it is certainly not always the network that is to blame. The database, hardware and operating system are also common culprits. And believe it or not, a major cause of poor application performance can be the application itself.

Application performance issues stemming from the application can be caused by a number of different factors related to the design of the application and otherwise. For example, there could be too many elements or too much content in the application; it could be too chatty, making multiple connections for each user request; or it could be slow and long-running queries. Not to mention memory leak, thread lock or a bad database schema that is slowing down data retrieval. As a network administrator, though, try telling this to the application developer or systems administrator and more often than not you’ll find yourself engaged in an epic battle.

Sure, there are half as many reasons why the source of the issue could be the network, but that argument won’t fly. You’re going to have to prove it. Here a few of the common accusations developers and SysAdmins make and how you can be prepared to refute them:

“Hey, the network is just too slow”

Response: You should power up your network monitoring tool and check the health and status of your network devices. SNMP tools can provide a lot of useful information. For example, when monitoring your routers and switches with SNMP, you can see if there were route flaps, packet loss, an increase in RTT and latency, and if the device CPU or memory utilization is high.

“Maybe your WAN link can’t handle my app”

Response: Cisco IPSLA can send synthetic packets and report on the capability or the readiness of the network link to handle IP traffic with TCP and UDP protocols or report specifically about VoIP performance, RTT, etc. If the synthetic packets generated by Cisco IPSLA that match the application protocol can be handled, they should also be able to handle the actual application traffic.

“There’s just not enough bandwidth”

Response: There’s a tool for that too! NetFlow data from routing and switching devices can report on bandwidth usage telling you how much of your WAN link is being utilized, which applications are using it, what end-points are involved and even report on the ToS priority of each IP conversation.

“It’s got to be something to do with your QoS priorities”

Response: Using a monitoring tool that supports Cisco CBQoS reporting, you can validate the performance of your QoS policies — pre and post policy statistics, too much buffer and how much traffic is being dropped for each QoS policy and class.

If your QoS policies are working as expected, it’s time to tell your foe, “Nope, try again!”

“Well, it might not be any of those things, but it’s still definitely the network”

Response: When all else fails, the answer is deep packet inspection (DPI). The visibility that DPI provides is virtually unlimited throughput information, out of order segment details, handshake details, re-transmissions and almost any other information you will need to prove once and for all that it’s not the network, and also find out the actual cause for poor application performance so you can really rub it in.

In conclusion, with the right technology and tools, network administrators can prove that the network is not at fault, but equally important they can be proactive and ensure small, routine network issues don’t become major headaches to begin with.

Don Thomas Jacob is a Head Geek at SolarWinds.

Share this

The Latest

July 25, 2024

The 2024 State of the Data Center Report from CoreSite shows that although C-suite confidence in the economy remains high, a VUCA (volatile, uncertain, complex, ambiguous) environment has many business leaders proceeding with caution when it comes to their IT and data ecosystems, with an emphasis on cost control and predictability, flexibility and risk management ...

July 24, 2024

In June, New Relic published the State of Observability for Energy and Utilities Report to share insights, analysis, and data on the impact of full-stack observability software in energy and utilities organizations' service capabilities. Here are eight key takeaways from the report ...

July 23, 2024

The rapid rise of generative AI (GenAI) has caught everyone's attention, leaving many to wonder if the technology's impact will live up to the immense hype. A recent survey by Alteryx provides valuable insights into the current state of GenAI adoption, revealing a shift from inflated expectations to tangible value realization across enterprises ... Here are five key takeaways that underscore GenAI's progression from hype to real-world impact ...

July 22, 2024
A defective software update caused what some experts are calling the largest IT outage in history on Friday, July 19. The impact reverberated through multiple industries around the world ...
July 18, 2024

As software development grows more intricate, the challenge for observability engineers tasked with ensuring optimal system performance becomes more daunting. Current methodologies are struggling to keep pace, with the annual Observability Pulse surveys indicating a rise in Mean Time to Remediation (MTTR). According to this survey, only a small fraction of organizations, around 10%, achieve full observability today. Generative AI, however, promises to significantly move the needle ...

July 17, 2024

While nearly all data leaders surveyed are building generative AI applications, most don't believe their data estate is actually prepared to support them, according to the State of Reliable AI report from Monte Carlo Data ...

July 16, 2024

Enterprises are putting a lot of effort into improving the digital employee experience (DEX), which has become essential to both improving organizational performance and attracting and retaining talented workers. But to date, most efforts to deliver outstanding DEX have focused on people working with laptops, PCs, or thin clients. Employees on the frontlines, using mobile devices to handle logistics ... have been largely overlooked ...

July 15, 2024

The average customer-facing incident takes nearly three hours to resolve (175 minutes) while the estimated cost of downtime is $4,537 per minute, meaning each incident can cost nearly $794,000, according to new research from PagerDuty ...

July 12, 2024

In MEAN TIME TO INSIGHT Episode 8, Shamus McGillicuddy, VP of Research, Network Infrastructure and Operations, at EMA discusses AutoCon with the conference founders Scott Robohn and Chris Grundemann ...

July 11, 2024

Numerous vendors and service providers have recently embraced the NaaS concept, yet there is still no industry consensus on its definition or the types of networks it involves. Furthermore, providers have varied in how they define the NaaS service delivery model. I conducted research for a new report, Network as a Service: Understanding the Cloud Consumption Model in Networking, to refine the concept of NaaS and reduce buyer confusion over what it is and how it can offer value ...