APMdigest asked industry experts – from analysts and consultants to the top vendors – to outline the most important factors that impact website response time. The last installment of the list, featuring factors 16–20, presents various factors you may not have considered.
16. IT INFRASTRUCTURE CHANGES
As new servers are powered on, database configurations changed, shared storage reconfigured, VMs reallocated, along with a whole host of other everyday infrastructure changes, little does the IT admin know how the upstream effects of these changes may be impacting web response times and the company’s bottom line. Consider this example: A storage (disk) change is made. The change slows a group of VMs. One of those VMs supports a database and therefore slows its queries. Let’s say those queries support the e-commerce application servers. As a result 60% of the users of this application experience slower responses. Or let’s say an overzealous VM admin observes that certain hosts are underutilized and adds an additional application to the underutilized servers. Now when an unexpected spike in user load occurs there will be insufficient compute resources to cover for it. Within each of those technologies, and the transitions between them, lies the potential for problems in end user transactions.
VP & GM, Dell Performance Monitoring
17. ALTERED CODE
One of the top factors impacting website response time is altered code, which doesn't trigger traditional monitoring alarms, as those are usually based on existing or known thresholds.
VP of Security Products, Prelert
18. DISTRIBUTED DENIAL OF SERVICE ATTACKS (DDoS)
Quocirca researched the concerns Europe organizations have about the security and performance of their online domains and the action taken to mitigate these in 2014. The survey found, that by a small margin the biggest overall concern was denial of services attacks, which have now become so wide spread that they can effect just about any online resource. This is backed by other non-Quocirca surveys that show the number and scale of attacks has continually increased in the last few years. However, whilst it is the biggest attacks that hit the headlines, it is huge number of smaller, largely unreported attacks, that should be of most concern. These are launched as diversionary measures to mask other more targeted attacks or even as demos. DDoS was followed by user end point issues, poor network performance poor website server performance and DNS performance in that order.
Analyst and Director, Quocirca
19. INFORMATION ARCHITECTURE
The top factor that impacts website response time is the user’s experience. And that is dependent on information architecture. Just responding fast to an http request is insufficient as that would be a technical answer to a business problem. The website exists, presumably to provide information and to answer a user’s question or questions without human intervention; thus, providing information availability 24x7. Is the information on the website structured appropriately so that the user finds the information appropriate to their need instead of waiting and then getting unhelpful information? Information architecture structures the information available for each role and may be grouped by industry. In addition good information architecture structures information in the best form to answer a questions or better yet, solve a problem. Putting the right information in front of the user without excessive navigation and false starts is the best way to improve response. The information provided should build in complexity as the user’s engagement continues. We should be timing how long it takes to get helpful information to the user and not just how long a request/response took.
VP Product Management and Marketing, Nastel Technologies
20. THE IT TEAM
One factor that is often overlooked is finger pointing. Website response time can be impacted by a number of different factors, which can cause internal finger pointing as folks try to pinpoint the problem. Could be that the client’s network is slow; maybe there’s an issue with the WAN link out to the ISP; perhaps the firewall is slowing or denying traffic – you get the picture. Without an overarching performance monitoring platform to keep an eye on all of these disparate areas, that internal struggle can slow things down considerably.
Director of Product Marketing, SevOne
In today's software-defined economy where every business runs on apps, the top factor that impacts performance and response time is inattention to early warning signs like increased load time for key pages, long-running database queries, or unresolved user complaints. Often, apps provide clear indications via monitoring alerts when any of these occurs but restoring and maintaining performance first requires a culture of service quality that associates uptime with customer value. Tools and metrics are useful but only if people and process are aligned to deliver exceptional user experiences. A lack of service culture often leads to early warning signs being ignored. One way to ensure that app teams value site performance is to make it easier for them to focus on solving problems that most directly impact customer experience. Too many irrelevant alerts means valuable resources spend time figuring out what problem to solve or solving the wrong problem - either of which are inefficient and demotivating.
VP Product, Big Panda
This year's Super Bowl drew in viewership of nearly 124 million viewers and made history as the most-watched live broadcast event since the 1969 moon landing. To support this spike in viewership, streaming companies like YouTube TV, Hulu and Paramount+ began preparing their IT infrastructure months in advance to ensure an exceptional viewer experience without outages or major interruptions. New Relic conducted a survey to understand the importance of a seamless viewing experience and the impact of outages during major streaming events such as the Super Bowl ...
As organizations continue to navigate the complexities of the digital era, which has been marked by exponential advancements in AI and technology, the strategic deployment of modern, practical applications has become indispensable for sustaining competitive advantage and realizing business goals. The Info-Tech Research Group report, Applications Priorities 2024, explores the following five initiatives for emerging and leading-edge technologies and practices that can enable IT and applications leaders to optimize their application portfolio and improve on capabilities needed to meet the ambitions of their organizations ...
Despite the growth in popularity of artificial intelligence (AI) and ML across a number of industries, there is still a huge amount of unrealized potential, with many businesses playing catch-up and still planning how ML solutions can best facilitate processes. Further progression could be limited without investment in specialized technical teams to drive development and integration ...
With over 200 streaming services to choose from, including multiple platforms featuring similar types of entertainment, users have little incentive to remain loyal to any given platform if it exhibits performance issues. Big names in streaming like Hulu, Amazon Prime and HBO Max invest thousands of hours into engineering observability and closed-loop monitoring to combat infrastructure and application issues, but smaller platforms struggle to remain competitive without access to the same resources ...
Generative AI has recently experienced unprecedented dramatic growth, making it one of the most exciting transformations the tech industry has seen in some time. However, this growth also poses a challenge for tech leaders who will be expected to deliver on the promise of new technology. In 2024, delivering tangible outcomes that meet the potential of AI, and setting up incubator projects for the future will be key tasks ...
SAP is a tool for automating business processes. Managing SAP solutions, especially with the shift to the cloud-based S/4HANA platform, can be intricate. To explore the concerns of SAP users during operational transformations and automation, a survey was conducted in mid-2023 by Digitate and Americas' SAP Users' Group ...
Some companies are just starting to dip their toes into developing AI capabilities, while (few) others can claim they have built a truly AI-first product. Regardless of where a company is on the AI journey, leaders must understand what it means to build every aspect of their product with AI in mind ...
Generative AI will usher in advantages within various industries. However, the technology is still nascent, and according to the recent Dynatrace survey there are many challenges and risks that organizations need to overcome to use this technology effectively ...
In today's digital era, monitoring and observability are indispensable in software and application development. Their efficacy lies in empowering developers to swiftly identify and address issues, enhance performance, and deliver flawless user experiences. Achieving these objectives requires meticulous planning, strategic implementation, and consistent ongoing maintenance. In this blog, we're sharing our five best practices to fortify your approach to application performance monitoring (APM) and observability ...
In MEAN TIME TO INSIGHT Episode 3, Shamus McGillicuddy, VP of Research, Network Infrastructure and Operations, at Enterprise Management Associates (EMA) discusses network security with Chris Steffen, VP of Research Covering Information Security, Risk, and Compliance Management at EMA ...