Skip to main content

Can the Internet Handle the Expected 2014 World Cup Traffic Records?

Alon Girmonsky

The 2010 FIFA World Cup fever tested the Internet’s limits more than ever before. News site traffic reached a blistering 12.1 million visitors per minute – a record that far exceeds the 8.5 million set by Barack Obama’s presidential election win back in 2008.

And, this year, the Internet is taking it one step further as the BBC plans to host a 24/7 World Cup feed, across all television, radio and digital platforms. That’s 50 percent more coverage than 2010. So, with more than 160 hours of programming, including highlights and match replays across all of their online channels, you have to wonder, how are they going to pull it off?

DevOps will be conducting some pretty rigorous testing to ensure their channels can hold up under what could be another record-breaking moment of traffic in Internet history. But, will this be enough?

Simulating Traffic

A key to performance testing is being able to simulate peak traffic to ensure your website will hold up under load. But, it’s important to avoid the all-too-common mistake of only testing within your corporate local area network (LAN).

Viewers of this year’s World Cup will span continents, so only testing traffic capacity within your own network will not suffice. It’s great if your site is able to sustain one million concurrent connections on your LAN, but when those connections are coming from other regions, putting more strain on your bandwidth, performance becomes uncertain.

Simulating a load scenario where the traffic only originates from within the corporate LAN can be compared to training for the Tour de France … on a stationary bike. Sure, you may be able to tackle the 3,500 kilometers over 23 days of training, but that doesn’t account for friction on the road, cyclist traffic or natural elements like wind, heat and rain.

That kind of training is only testing your body’s ability to perform under the most ideal conditions, which is the same as testing website performance from within the corporate LAN. On the LAN, you don’t have to go through the firewall, cache, load balancer, network equipment, modem or routers, thereby avoiding any kind of packet collisions or re-transmits. Ideal? Yes. Realistic? Not a chance.

Cloud-Based Performance Testing

Cloud-based performance testing enables broadcasters to simulate the millions of real users coming directly from the Internet – just as they will be on June 12 when the World Cup kicks off.

The cloud is extremely well-suited to generating the peak demands required for website performance testing. Not only can you ensure that sufficient compute power is available to scale from 100,000 to 1,000,000 virtual users and beyond, but you can also do it on demand with automatic resource provisioning.

Gone are the performance-testing delays of deploying and verifying internally managed hardware. With the cloud, concerns over the number of available servers on hand and whether idle servers are wasting valuable resources are something of the past. Performance testing can be run from anywhere with an Internet connection and a browser without the risk of costly over provisioning.

If broadcasters like ESPN, the BBC and ITV that are expecting to handle an increase in traffic from the World Cup were to solely use an on-premise testing model, they would have to acquire enough resources to support the tremendous capacity planning for that event. But, those resources could potentially go unused for the rest of the year.

Matters are complicated further when you consider that viewers will expect to watch seamless coverage of the games on TV, tablets and smartphones, so traffic simulations should take multiple devices into account.

The elasticity and agility of cloud resources means they can be easily scaled up or down as needed while only paying for what you use thanks to pay-as-you-go or utility-style pricing. This makes it an extremely efficient and cost-effective solution for performance testing needs.

Handling Global Load

Performance tests for something as big as the World Cup need to go even further to test global demand from most countries around the world. After all, soccer is one of the most widely watched sports there is, with a footballer fan base extending far beyond this year’s host country, Brazil. The global nature of the cloud serves this requirement well. Load tests can easily be carried out across different geographies since the cloud allows virtual users to be replicated in a variety of locations to test international performance. Cloud providers and test solutions can evaluate website global readiness, all without requiring you to stand up an expensive data center of your own in each location.

All in all, it would appear that technology is saving the day once more. The ability to broadcast live international coverage over the Internet enables an increasing number of fans to get connected and stay connected. With that, broadcasters let themselves in to a bottomless pit of demand for live viewing - which, in turn, leads to increased revenue from advertisers. Without cloud-based performance simulations, chances are, broadcasters would be getting yellow cards of dissatisfaction all around.

Alon Girmonsky is CEO of BlazeMeter.

The Latest

AI is the catalyst for significant investment in data teams as enterprises require higher-quality data to power their AI applications, according to the State of Analytics Engineering Report from dbt Labs ...

Misaligned architecture can lead to business consequences, with 93% of respondents reporting negative outcomes such as service disruptions, high operational costs and security challenges ...

A Gartner analyst recently suggested that GenAI tools could create 25% time savings for network operational teams. Where might these time savings come from? How are GenAI tools helping NetOps teams today, and what other tasks might they take on in the future as models continue improving? In general, these savings come from automating or streamlining manual NetOps tasks ...

IT and line-of-business teams are increasingly aligned in their efforts to close the data gap and drive greater collaboration to alleviate IT bottlenecks and offload growing demands on IT teams, according to The 2025 Automation Benchmark Report: Insights from IT Leaders on Enterprise Automation & the Future of AI-Driven Businesses from Jitterbit ...

A large majority (86%) of data management and AI decision makers cite protecting data privacy as a top concern, with 76% of respondents citing ROI on data privacy and AI initiatives across their organization, according to a new Harris Poll from Collibra ...

According to Gartner, Inc. the following six trends will shape the future of cloud over the next four years, ultimately resulting in new ways of working that are digital in nature and transformative in impact ...

2020 was the equivalent of a wedding with a top-shelf open bar. As businesses scrambled to adjust to remote work, digital transformation accelerated at breakneck speed. New software categories emerged overnight. Tech stacks ballooned with all sorts of SaaS apps solving ALL the problems — often with little oversight or long-term integration planning, and yes frequently a lot of duplicated functionality ... But now the music's faded. The lights are on. Everyone from the CIO to the CFO is checking the bill. Welcome to the Great SaaS Hangover ...

Regardless of OpenShift being a scalable and flexible software, it can be a pain to monitor since complete visibility into the underlying operations is not guaranteed ... To effectively monitor an OpenShift environment, IT administrators should focus on these five key elements and their associated metrics ...

An overwhelming majority of IT leaders (95%) believe the upcoming wave of AI-powered digital transformation is set to be the most impactful and intensive seen thus far, according to The Science of Productivity: AI, Adoption, And Employee Experience, a new report from Nexthink ...

Overall outage frequency and the general level of reported severity continue to decline, according to the Outage Analysis 2025 from Uptime Institute. However, cyber security incidents are on the rise and often have severe, lasting impacts ...

Can the Internet Handle the Expected 2014 World Cup Traffic Records?

Alon Girmonsky

The 2010 FIFA World Cup fever tested the Internet’s limits more than ever before. News site traffic reached a blistering 12.1 million visitors per minute – a record that far exceeds the 8.5 million set by Barack Obama’s presidential election win back in 2008.

And, this year, the Internet is taking it one step further as the BBC plans to host a 24/7 World Cup feed, across all television, radio and digital platforms. That’s 50 percent more coverage than 2010. So, with more than 160 hours of programming, including highlights and match replays across all of their online channels, you have to wonder, how are they going to pull it off?

DevOps will be conducting some pretty rigorous testing to ensure their channels can hold up under what could be another record-breaking moment of traffic in Internet history. But, will this be enough?

Simulating Traffic

A key to performance testing is being able to simulate peak traffic to ensure your website will hold up under load. But, it’s important to avoid the all-too-common mistake of only testing within your corporate local area network (LAN).

Viewers of this year’s World Cup will span continents, so only testing traffic capacity within your own network will not suffice. It’s great if your site is able to sustain one million concurrent connections on your LAN, but when those connections are coming from other regions, putting more strain on your bandwidth, performance becomes uncertain.

Simulating a load scenario where the traffic only originates from within the corporate LAN can be compared to training for the Tour de France … on a stationary bike. Sure, you may be able to tackle the 3,500 kilometers over 23 days of training, but that doesn’t account for friction on the road, cyclist traffic or natural elements like wind, heat and rain.

That kind of training is only testing your body’s ability to perform under the most ideal conditions, which is the same as testing website performance from within the corporate LAN. On the LAN, you don’t have to go through the firewall, cache, load balancer, network equipment, modem or routers, thereby avoiding any kind of packet collisions or re-transmits. Ideal? Yes. Realistic? Not a chance.

Cloud-Based Performance Testing

Cloud-based performance testing enables broadcasters to simulate the millions of real users coming directly from the Internet – just as they will be on June 12 when the World Cup kicks off.

The cloud is extremely well-suited to generating the peak demands required for website performance testing. Not only can you ensure that sufficient compute power is available to scale from 100,000 to 1,000,000 virtual users and beyond, but you can also do it on demand with automatic resource provisioning.

Gone are the performance-testing delays of deploying and verifying internally managed hardware. With the cloud, concerns over the number of available servers on hand and whether idle servers are wasting valuable resources are something of the past. Performance testing can be run from anywhere with an Internet connection and a browser without the risk of costly over provisioning.

If broadcasters like ESPN, the BBC and ITV that are expecting to handle an increase in traffic from the World Cup were to solely use an on-premise testing model, they would have to acquire enough resources to support the tremendous capacity planning for that event. But, those resources could potentially go unused for the rest of the year.

Matters are complicated further when you consider that viewers will expect to watch seamless coverage of the games on TV, tablets and smartphones, so traffic simulations should take multiple devices into account.

The elasticity and agility of cloud resources means they can be easily scaled up or down as needed while only paying for what you use thanks to pay-as-you-go or utility-style pricing. This makes it an extremely efficient and cost-effective solution for performance testing needs.

Handling Global Load

Performance tests for something as big as the World Cup need to go even further to test global demand from most countries around the world. After all, soccer is one of the most widely watched sports there is, with a footballer fan base extending far beyond this year’s host country, Brazil. The global nature of the cloud serves this requirement well. Load tests can easily be carried out across different geographies since the cloud allows virtual users to be replicated in a variety of locations to test international performance. Cloud providers and test solutions can evaluate website global readiness, all without requiring you to stand up an expensive data center of your own in each location.

All in all, it would appear that technology is saving the day once more. The ability to broadcast live international coverage over the Internet enables an increasing number of fans to get connected and stay connected. With that, broadcasters let themselves in to a bottomless pit of demand for live viewing - which, in turn, leads to increased revenue from advertisers. Without cloud-based performance simulations, chances are, broadcasters would be getting yellow cards of dissatisfaction all around.

Alon Girmonsky is CEO of BlazeMeter.

The Latest

AI is the catalyst for significant investment in data teams as enterprises require higher-quality data to power their AI applications, according to the State of Analytics Engineering Report from dbt Labs ...

Misaligned architecture can lead to business consequences, with 93% of respondents reporting negative outcomes such as service disruptions, high operational costs and security challenges ...

A Gartner analyst recently suggested that GenAI tools could create 25% time savings for network operational teams. Where might these time savings come from? How are GenAI tools helping NetOps teams today, and what other tasks might they take on in the future as models continue improving? In general, these savings come from automating or streamlining manual NetOps tasks ...

IT and line-of-business teams are increasingly aligned in their efforts to close the data gap and drive greater collaboration to alleviate IT bottlenecks and offload growing demands on IT teams, according to The 2025 Automation Benchmark Report: Insights from IT Leaders on Enterprise Automation & the Future of AI-Driven Businesses from Jitterbit ...

A large majority (86%) of data management and AI decision makers cite protecting data privacy as a top concern, with 76% of respondents citing ROI on data privacy and AI initiatives across their organization, according to a new Harris Poll from Collibra ...

According to Gartner, Inc. the following six trends will shape the future of cloud over the next four years, ultimately resulting in new ways of working that are digital in nature and transformative in impact ...

2020 was the equivalent of a wedding with a top-shelf open bar. As businesses scrambled to adjust to remote work, digital transformation accelerated at breakneck speed. New software categories emerged overnight. Tech stacks ballooned with all sorts of SaaS apps solving ALL the problems — often with little oversight or long-term integration planning, and yes frequently a lot of duplicated functionality ... But now the music's faded. The lights are on. Everyone from the CIO to the CFO is checking the bill. Welcome to the Great SaaS Hangover ...

Regardless of OpenShift being a scalable and flexible software, it can be a pain to monitor since complete visibility into the underlying operations is not guaranteed ... To effectively monitor an OpenShift environment, IT administrators should focus on these five key elements and their associated metrics ...

An overwhelming majority of IT leaders (95%) believe the upcoming wave of AI-powered digital transformation is set to be the most impactful and intensive seen thus far, according to The Science of Productivity: AI, Adoption, And Employee Experience, a new report from Nexthink ...

Overall outage frequency and the general level of reported severity continue to decline, according to the Outage Analysis 2025 from Uptime Institute. However, cyber security incidents are on the rise and often have severe, lasting impacts ...