Skip to main content

Top Factors That Impact Application Performance 2016 - Part 2

In 2013, APMdigest published a list called 15 Top Factors That Impact Application Performance. Even today, this is one of the most popular pieces of content on the site. And for good reason – the whole concept of Application Performance Management (APM) starts with identifying the factors that impact application performance, and then doing something about it. However, in the fast moving world of IT, many aspects of application performance have changed in the 3 years since the list was published. And many new experts have come on the scene. So APMdigest is updating the list for 2016, and you will be surprised how much it has changed.

Start with Top Factors That Impact Application Performance 2016 - Part 1

Part 2 of this list covers more challenges in the environment, including containers, microservices and issues with the network.

6. VIRTUALIZATION AND CONTAINERIZATION

Applications today are disaggregated into multiple components that could be deployed as highly virtualized or containerized workloads. As a result, gaining visibility into the flows to understand the interaction between different components is paramount for IT operations to provide the best user experience for the applications.
Ananda Rajagopal
VP of Product Management, Gigamon

Over the last decade, we have seen the commoditization of the Cloud, and the trend towards running applications on virtualized hardware continues and evolves to an even higher level of modularization and compartmentalization: containers, micro-services, software-defined networks, virtual storage, and more. There is a trend towards small, self-contained, independent components that act as recyclable, multi-purpose building blocks. All that might allow for faster and cheaper development and operation of complex systems, but complexity increases when it comes to APM, tuning, monitoring, logging, debugging. The respective tools need to be able to see and analyze all physical and virtual components and how they interact, and allow developers and ops teams to make sense of all those data points.
Sven Dummer
Senior Director of Product Marketing, Loggly

7. MICROSERVICES

As enterprises are adopting microservices and continuous delivery methodologies, the number of independent applications and web services are growing exponentially. Isolating any application performance issues in an application environment with hundreds and thousands of interdependent services can be challenging if not instrumented and monitored in real-time. Manually instrumenting these microservices and setting static thresholds can be a very difficult task, if not impossible. Enterprises need to automatically discover these large numbers of microservices, dynamically baseline their performance, collect deep diagnostics and alerts when the performance deviates from the normal baseline.
Anand Akela
Director, Product Marketing, AppDynamics

Modern applications increasingly relying on stateless microservices, are often paired with stateful data services (like NoSQL, Kafka, Hadoop etc.), and are being deployed on containers or leverage serverless architectures. As the application substrate is changing, so are the factors that impact performance of these applications. These factors include how the various microservices are interacting with each other, their availability, and correlating issues like errors, latency and throughput across these services. The individual service performance is not paramount in itself. Also as orchestration systems like Mesos, Kubernetes and Docker Swarm become more critical, application performance will increasingly rely on how effectively these orchestration systems can manage resources, both applications as well as underlying infrastructure.
Sachin Agarwal
VP Marketing, OpsClarity

8. SERVICE DESIGN

In today's Everything-As-A-Service hybrid application world the biggest impact on performance is ignoring core principles of proximity and context of the services to the users. Gone are the days of tightly coupled applications, data and infrastructure. Whether born in the cloud or connected to it – the biggest impact I have seen with today's modern applications is ignoring solid user centered design principles. Services that integrate multiple applications, micro-services and clouds require more finesse in balancing the last mile connectivity from the user to the service but also the services, micro-services and data to each other. Experience teaches us that applying same assumptions one would of traditional client/server applications will not work with today's modern applications. In lieu of just looking at a specific micro-service, service or component – performance testing must take into account impact to overall performance or risk production issues that are difficult to identify or pinpoint because of the multiple faceted nature of these solutions. Improper design of a service from a user location/experience is a recipe for disaster.
Jeanne Morain
Strategist and Author, iSpeak Cloud

9. SERVER SIDE CODE

Although we are seeing less tolerance for low applications by users, the primary reason for poor performance continues to be inefficient server side code. As components are increasingly interconnected determining the cause of slowness or faults continues to take longer, creating the need for end-to-end APM.
Jonah Kowall
VP of Market Development and Insights, AppDynamics

10. NETWORK LATENCY

Latency is the top factor that impacts application performance. The most well-developed application will be terribly slow if latency between users and servers is high. And the most poorly-developed application can garner all kinds of praises when everything is local. I believe low latency is the single most important asset that IT managers can have on their networks. You should focus on any ways to reduce end-to-end delay. This includes reducing the various contributors to latency, such as processing delay, queuing delay, serialization delay, and last, but certainly not least, propagation delay.
Jean Tunis
Senior Consultant and Founder of RootPerformance

Poorly peered Internet relationships and congestion remain the top contributor to latency in web and mobile apps, even when using a Content Delivery Network. Seventy-five percent of an application's page load time comes from the latency of the network. Even after solving the poorly constructed HTML, and ensuring your app has no blocking calls, you still have the Internet to deal with.
Josh Gray
Chief Architect, Cedexis

11. LOAD

Applications can become overloaded based on changes in the business environment. More jobs, workloads, or users can in turn negatively affect performance. For example, if your Microsoft Exchange send and receive queue lengths grow for mailbox databases, or if users are experiencing logon latency, you must look at not only the processing and memory resources for Exchange servers, but also the availability of flash storage.
Goran Garevski
VP of Engineering, Comtrade Software

12. RESOURCE AVAILABILITY

Application performance is most impacted by resource availability. With applications migrating to digital media, it can be difficult for businesses to gauge the appropriate resources necessary to deliver a consistent performance metric. It is critical to leverage solutions that can dynamically detect and scale resource requirements through the use of virtualization, orchestration, and automation to provide the cloud-like elasticity and agility that organizations require for successful and consistent application performance.
Frank Yue
Director of Application Delivery Solutions, Radware

13. CACHING BOTTLENECKS

Application speed and scalability are forever intertwined with each other. With load triggered elastic scaling in the cloud, finding every application bottleneck and applying strategic caching architectures are more important than ever. At times, code or database results in a given framework or architecture cannot be refactored quickly enough. Failure to cache wherever possible, relates directly to a higher number of host instances, which means a higher cost of doing business in the cloud. Bottlenecks tend move around within an evolving application with a high release rate and with a lot of developers. A solid APM technology can quickly and automatically identify a deviation from application baseline performance at any tier, essentially safeguarding user experience, brand reputation and digital trust. Hosting an application in the cloud without an APM tool watching for bottlenecks is throwing money away.
Matt Hintze
CTO, HITS Inc.

Read Top Factors That Impact Application Performance 2016 - Part 3 covering how the application interacts with the backend and the front end.

Hot Topics

The Latest

According to Auvik's 2025 IT Trends Report, 60% of IT professionals feel at least moderately burned out on the job, with 43% stating that their workload is contributing to work stress. At the same time, many IT professionals are naming AI and machine learning as key areas they'd most like to upskill ...

Businesses that face downtime or outages risk financial and reputational damage, as well as reducing partner, shareholder, and customer trust. One of the major challenges that enterprises face is implementing a robust business continuity plan. What's the solution? The answer may lie in disaster recovery tactics such as truly immutable storage and regular disaster recovery testing ...

IT spending is expected to jump nearly 10% in 2025, and organizations are now facing pressure to manage costs without slowing down critical functions like observability. To meet the challenge, leaders are turning to smarter, more cost effective business strategies. Enter stage right: OpenTelemetry, the missing piece of the puzzle that is no longer just an option but rather a strategic advantage ...

Amidst the threat of cyberhacks and data breaches, companies install several security measures to keep their business safely afloat. These measures aim to protect businesses, employees, and crucial data. Yet, employees perceive them as burdensome. Frustrated with complex logins, slow access, and constant security checks, workers decide to completely bypass all security set-ups ...

Image
Cloudbrink's Personal SASE services provide last-mile acceleration and reduction in latency

In MEAN TIME TO INSIGHT Episode 13, Shamus McGillicuddy, VP of Research, Network Infrastructure and Operations, at EMA discusses hybrid multi-cloud networking strategy ... 

In high-traffic environments, the sheer volume and unpredictable nature of network incidents can quickly overwhelm even the most skilled teams, hindering their ability to react swiftly and effectively, potentially impacting service availability and overall business performance. This is where closed-loop remediation comes into the picture: an IT management concept designed to address the escalating complexity of modern networks ...

In 2025, enterprise workflows are undergoing a seismic shift. Propelled by breakthroughs in generative AI (GenAI), large language models (LLMs), and natural language processing (NLP), a new paradigm is emerging — agentic AI. This technology is not just automating tasks; it's reimagining how organizations make decisions, engage customers, and operate at scale ...

In the early days of the cloud revolution, business leaders perceived cloud services as a means of sidelining IT organizations. IT was too slow, too expensive, or incapable of supporting new technologies. With a team of developers, line of business managers could deploy new applications and services in the cloud. IT has been fighting to retake control ever since. Today, IT is back in the driver's seat, according to new research by Enterprise Management Associates (EMA) ...

In today's fast-paced and increasingly complex network environments, Network Operations Centers (NOCs) are the backbone of ensuring continuous uptime, smooth service delivery, and rapid issue resolution. However, the challenges faced by NOC teams are only growing. In a recent study, 78% state network complexity has grown significantly over the last few years while 84% regularly learn about network issues from users. It is imperative we adopt a new approach to managing today's network experiences ...

Image
Broadcom

From growing reliance on FinOps teams to the increasing attention on artificial intelligence (AI), and software licensing, the Flexera 2025 State of the Cloud Report digs into how organizations are improving cloud spend efficiency, while tackling the complexities of emerging technologies ...

Top Factors That Impact Application Performance 2016 - Part 2

In 2013, APMdigest published a list called 15 Top Factors That Impact Application Performance. Even today, this is one of the most popular pieces of content on the site. And for good reason – the whole concept of Application Performance Management (APM) starts with identifying the factors that impact application performance, and then doing something about it. However, in the fast moving world of IT, many aspects of application performance have changed in the 3 years since the list was published. And many new experts have come on the scene. So APMdigest is updating the list for 2016, and you will be surprised how much it has changed.

Start with Top Factors That Impact Application Performance 2016 - Part 1

Part 2 of this list covers more challenges in the environment, including containers, microservices and issues with the network.

6. VIRTUALIZATION AND CONTAINERIZATION

Applications today are disaggregated into multiple components that could be deployed as highly virtualized or containerized workloads. As a result, gaining visibility into the flows to understand the interaction between different components is paramount for IT operations to provide the best user experience for the applications.
Ananda Rajagopal
VP of Product Management, Gigamon

Over the last decade, we have seen the commoditization of the Cloud, and the trend towards running applications on virtualized hardware continues and evolves to an even higher level of modularization and compartmentalization: containers, micro-services, software-defined networks, virtual storage, and more. There is a trend towards small, self-contained, independent components that act as recyclable, multi-purpose building blocks. All that might allow for faster and cheaper development and operation of complex systems, but complexity increases when it comes to APM, tuning, monitoring, logging, debugging. The respective tools need to be able to see and analyze all physical and virtual components and how they interact, and allow developers and ops teams to make sense of all those data points.
Sven Dummer
Senior Director of Product Marketing, Loggly

7. MICROSERVICES

As enterprises are adopting microservices and continuous delivery methodologies, the number of independent applications and web services are growing exponentially. Isolating any application performance issues in an application environment with hundreds and thousands of interdependent services can be challenging if not instrumented and monitored in real-time. Manually instrumenting these microservices and setting static thresholds can be a very difficult task, if not impossible. Enterprises need to automatically discover these large numbers of microservices, dynamically baseline their performance, collect deep diagnostics and alerts when the performance deviates from the normal baseline.
Anand Akela
Director, Product Marketing, AppDynamics

Modern applications increasingly relying on stateless microservices, are often paired with stateful data services (like NoSQL, Kafka, Hadoop etc.), and are being deployed on containers or leverage serverless architectures. As the application substrate is changing, so are the factors that impact performance of these applications. These factors include how the various microservices are interacting with each other, their availability, and correlating issues like errors, latency and throughput across these services. The individual service performance is not paramount in itself. Also as orchestration systems like Mesos, Kubernetes and Docker Swarm become more critical, application performance will increasingly rely on how effectively these orchestration systems can manage resources, both applications as well as underlying infrastructure.
Sachin Agarwal
VP Marketing, OpsClarity

8. SERVICE DESIGN

In today's Everything-As-A-Service hybrid application world the biggest impact on performance is ignoring core principles of proximity and context of the services to the users. Gone are the days of tightly coupled applications, data and infrastructure. Whether born in the cloud or connected to it – the biggest impact I have seen with today's modern applications is ignoring solid user centered design principles. Services that integrate multiple applications, micro-services and clouds require more finesse in balancing the last mile connectivity from the user to the service but also the services, micro-services and data to each other. Experience teaches us that applying same assumptions one would of traditional client/server applications will not work with today's modern applications. In lieu of just looking at a specific micro-service, service or component – performance testing must take into account impact to overall performance or risk production issues that are difficult to identify or pinpoint because of the multiple faceted nature of these solutions. Improper design of a service from a user location/experience is a recipe for disaster.
Jeanne Morain
Strategist and Author, iSpeak Cloud

9. SERVER SIDE CODE

Although we are seeing less tolerance for low applications by users, the primary reason for poor performance continues to be inefficient server side code. As components are increasingly interconnected determining the cause of slowness or faults continues to take longer, creating the need for end-to-end APM.
Jonah Kowall
VP of Market Development and Insights, AppDynamics

10. NETWORK LATENCY

Latency is the top factor that impacts application performance. The most well-developed application will be terribly slow if latency between users and servers is high. And the most poorly-developed application can garner all kinds of praises when everything is local. I believe low latency is the single most important asset that IT managers can have on their networks. You should focus on any ways to reduce end-to-end delay. This includes reducing the various contributors to latency, such as processing delay, queuing delay, serialization delay, and last, but certainly not least, propagation delay.
Jean Tunis
Senior Consultant and Founder of RootPerformance

Poorly peered Internet relationships and congestion remain the top contributor to latency in web and mobile apps, even when using a Content Delivery Network. Seventy-five percent of an application's page load time comes from the latency of the network. Even after solving the poorly constructed HTML, and ensuring your app has no blocking calls, you still have the Internet to deal with.
Josh Gray
Chief Architect, Cedexis

11. LOAD

Applications can become overloaded based on changes in the business environment. More jobs, workloads, or users can in turn negatively affect performance. For example, if your Microsoft Exchange send and receive queue lengths grow for mailbox databases, or if users are experiencing logon latency, you must look at not only the processing and memory resources for Exchange servers, but also the availability of flash storage.
Goran Garevski
VP of Engineering, Comtrade Software

12. RESOURCE AVAILABILITY

Application performance is most impacted by resource availability. With applications migrating to digital media, it can be difficult for businesses to gauge the appropriate resources necessary to deliver a consistent performance metric. It is critical to leverage solutions that can dynamically detect and scale resource requirements through the use of virtualization, orchestration, and automation to provide the cloud-like elasticity and agility that organizations require for successful and consistent application performance.
Frank Yue
Director of Application Delivery Solutions, Radware

13. CACHING BOTTLENECKS

Application speed and scalability are forever intertwined with each other. With load triggered elastic scaling in the cloud, finding every application bottleneck and applying strategic caching architectures are more important than ever. At times, code or database results in a given framework or architecture cannot be refactored quickly enough. Failure to cache wherever possible, relates directly to a higher number of host instances, which means a higher cost of doing business in the cloud. Bottlenecks tend move around within an evolving application with a high release rate and with a lot of developers. A solid APM technology can quickly and automatically identify a deviation from application baseline performance at any tier, essentially safeguarding user experience, brand reputation and digital trust. Hosting an application in the cloud without an APM tool watching for bottlenecks is throwing money away.
Matt Hintze
CTO, HITS Inc.

Read Top Factors That Impact Application Performance 2016 - Part 3 covering how the application interacts with the backend and the front end.

Hot Topics

The Latest

According to Auvik's 2025 IT Trends Report, 60% of IT professionals feel at least moderately burned out on the job, with 43% stating that their workload is contributing to work stress. At the same time, many IT professionals are naming AI and machine learning as key areas they'd most like to upskill ...

Businesses that face downtime or outages risk financial and reputational damage, as well as reducing partner, shareholder, and customer trust. One of the major challenges that enterprises face is implementing a robust business continuity plan. What's the solution? The answer may lie in disaster recovery tactics such as truly immutable storage and regular disaster recovery testing ...

IT spending is expected to jump nearly 10% in 2025, and organizations are now facing pressure to manage costs without slowing down critical functions like observability. To meet the challenge, leaders are turning to smarter, more cost effective business strategies. Enter stage right: OpenTelemetry, the missing piece of the puzzle that is no longer just an option but rather a strategic advantage ...

Amidst the threat of cyberhacks and data breaches, companies install several security measures to keep their business safely afloat. These measures aim to protect businesses, employees, and crucial data. Yet, employees perceive them as burdensome. Frustrated with complex logins, slow access, and constant security checks, workers decide to completely bypass all security set-ups ...

Image
Cloudbrink's Personal SASE services provide last-mile acceleration and reduction in latency

In MEAN TIME TO INSIGHT Episode 13, Shamus McGillicuddy, VP of Research, Network Infrastructure and Operations, at EMA discusses hybrid multi-cloud networking strategy ... 

In high-traffic environments, the sheer volume and unpredictable nature of network incidents can quickly overwhelm even the most skilled teams, hindering their ability to react swiftly and effectively, potentially impacting service availability and overall business performance. This is where closed-loop remediation comes into the picture: an IT management concept designed to address the escalating complexity of modern networks ...

In 2025, enterprise workflows are undergoing a seismic shift. Propelled by breakthroughs in generative AI (GenAI), large language models (LLMs), and natural language processing (NLP), a new paradigm is emerging — agentic AI. This technology is not just automating tasks; it's reimagining how organizations make decisions, engage customers, and operate at scale ...

In the early days of the cloud revolution, business leaders perceived cloud services as a means of sidelining IT organizations. IT was too slow, too expensive, or incapable of supporting new technologies. With a team of developers, line of business managers could deploy new applications and services in the cloud. IT has been fighting to retake control ever since. Today, IT is back in the driver's seat, according to new research by Enterprise Management Associates (EMA) ...

In today's fast-paced and increasingly complex network environments, Network Operations Centers (NOCs) are the backbone of ensuring continuous uptime, smooth service delivery, and rapid issue resolution. However, the challenges faced by NOC teams are only growing. In a recent study, 78% state network complexity has grown significantly over the last few years while 84% regularly learn about network issues from users. It is imperative we adopt a new approach to managing today's network experiences ...

Image
Broadcom

From growing reliance on FinOps teams to the increasing attention on artificial intelligence (AI), and software licensing, the Flexera 2025 State of the Cloud Report digs into how organizations are improving cloud spend efficiency, while tackling the complexities of emerging technologies ...