It's Time to Modernize Pre-Deployment Testing
August 31, 2021

Jeff Atkins
Spirent

Share this

Here's how it happens: You're deploying a new technology, thinking everything's going smoothly, when the alerts start coming in. Your rollout has hit a snag. Whole groups of users are complaining about poor performance on their devices. Some can't access applications at all. You've now blown your service-level agreement (SLA). You might have just introduced a new security vulnerability. In the worst case, your big expensive product launch has missed the mark altogether.

"How did this happen?" you're asking yourself. "Didn't we test everything before we deployed?"

Yes, you did. But you made a critical though common mistake: your tests assumed ideal network conditions. And as you just learned firsthand, the idealized environment in your testing models and the way things work in the real world are two very different things.

Hopefully, this hypothetical doesn't sound too familiar. But if you're relying on traditional testing workflows and you've managed to avoid these kinds of outcomes so far, count your blessings. Because you're taking a big risk with every new launch.

There's a better way to test new enterprise technologies so they get deployed on time, under budget, with the performance you expect. To do it though, you need to get better at predicting the future. That starts with painting a more accurate picture of the present.

Navigating Complexity

Modern IT organizations already deal with more devices, more connections, and complexity than ever before. But even if you get a handle on today's technology landscape, new innovations emerge all the time. Next-generation Ethernet technologies, 5G networks, SD-WAN, Wi-Fi 6, and others can all bring important benefits to your users — benefits your competitors may already be realizing, that you can't afford to ignore. Yet, each new deployment carries significant unpredictability and risk.

All of this means it's more critical than ever to thoroughly test and validate new technology before you deploy. But all the testing in the world can't help you if you're not testing the right things. And the fact is, next-generation enterprise technologies are evolving too quickly for legacy testing approaches to keep up.

In too many cases, enterprises still test new applications and infrastructure by connecting devices directly to datacenters or clouds, with little or no traffic on the network. That kind of testing can tell you how the technology works under ideal conditions, but how often can you expect ideal conditions in the real world?

How will the technology perform on a congested or impaired network?

What kinds of problems will have the biggest impact on user experience?

Too often, those questions get answered only after deployment, when users complain. At which point customer satisfaction has already taken a hit, you may have missed an SLA, and you're looking at a time-consuming, expensive repair process.

Even more concerning, security often gets less attention than performance in pre-deployment validation. Many enterprises still rely on basic tools and firmware checks, or even just assurances from vendors, that software is safe to deploy. Which means there's a good chance you'll only learn about a vulnerability after it's been exploited, and your systems are already compromised.

A Smarter Approach

Fortunately, it's possible to predict and avoid most of these issues. To do it though, we need to recognize that testing models that worked a decade ago won't cut it anymore. We need to reimagine pre-deployment testing for today's more complex, dynamic, and distributed world.

Whatever your updated testing methodology looks like, it should include the following core practices:

Performance validation: Your vendors aren't lying when they claim to hit certain benchmarks, but you can't assume you'll achieve comparable performance in your own environment—especially if you'll be operating under an SLA. You should be measuring everything from voice quality to packet jitter. By validating real-world performance across more granular metrics, you can better evaluate any new solutions you're considering. At the same time, you identify everything you'll need to understand the user experience and troubleshoot problems post-deployment.

Network emulation: If you're going to deploy with confidence, you want to get your test beds as close as possible to real-world conditions. That includes mimicking networks, devices, and users under heavy traffic loads.

Network impairment: Network faults and service degradations are an unavoidable (if hopefully infrequent) reality. So, wouldn't you prefer to know how a new technology will respond under those conditions ahead of time? By running controlled network impairment scenarios alongside emulation, you'll know exactly how problems will affect your users, so you can better prepare. Even more important, you can set realistic expectations with customers and achievable SLAs.

Security assessments: Don't bet your security on third-party assurances or basic firmware checks. Take the time to thoroughly test for vulnerabilities, simulate known attacks, and evaluate weaknesses in the end-to-end network.

Testbed automation: To keep pace with rapidly changing networks and clouds, you should look to automate as much of the testing process as possible. The less you rely on slow, manual testing methodologies, the more quickly and cost-effectively you'll be able to simulate new scenarios as your environment evolves.

Proactive Testing Makes All the Difference

So, what happens when you put these principles into practice — when you modernize your testing to reflect a more realistic picture of your technology landscape?

First, you save time and money by identifying problems before deploying instead of after. It's a lot harder and more expensive to fix issues with a new technology when diverse users and systems already rely on it, and SLAs are already violated.

Second, you protect your users and your business by detecting and mitigating security vulnerabilities before malicious actors can exploit them. Finally, you improve your organization's ability to take advantage of new technology. By automating the testing process, you can continually bring in new testing practices and collect more valuable insights without slowing down innovation.

By overhauling your testing strategy based on realism and automation, you can put your organization in the best position to capitalize on new technologies when they emerge. You can reduce the risk of disruptive (and expensive) problems cropping up out of the blue. And, you can make ongoing innovation a core strength of your IT organization — and a key competitive advantage for your business.

Jeff Atkins is Director of Solutions Marketing at Spirent
Share this

The Latest

July 25, 2024

The 2024 State of the Data Center Report from CoreSite shows that although C-suite confidence in the economy remains high, a VUCA (volatile, uncertain, complex, ambiguous) environment has many business leaders proceeding with caution when it comes to their IT and data ecosystems, with an emphasis on cost control and predictability, flexibility and risk management ...

July 24, 2024

In June, New Relic published the State of Observability for Energy and Utilities Report to share insights, analysis, and data on the impact of full-stack observability software in energy and utilities organizations' service capabilities. Here are eight key takeaways from the report ...

July 23, 2024

The rapid rise of generative AI (GenAI) has caught everyone's attention, leaving many to wonder if the technology's impact will live up to the immense hype. A recent survey by Alteryx provides valuable insights into the current state of GenAI adoption, revealing a shift from inflated expectations to tangible value realization across enterprises ... Here are five key takeaways that underscore GenAI's progression from hype to real-world impact ...

July 22, 2024
A defective software update caused what some experts are calling the largest IT outage in history on Friday, July 19. The impact reverberated through multiple industries around the world ...
July 18, 2024

As software development grows more intricate, the challenge for observability engineers tasked with ensuring optimal system performance becomes more daunting. Current methodologies are struggling to keep pace, with the annual Observability Pulse surveys indicating a rise in Mean Time to Remediation (MTTR). According to this survey, only a small fraction of organizations, around 10%, achieve full observability today. Generative AI, however, promises to significantly move the needle ...

July 17, 2024

While nearly all data leaders surveyed are building generative AI applications, most don't believe their data estate is actually prepared to support them, according to the State of Reliable AI report from Monte Carlo Data ...

July 16, 2024

Enterprises are putting a lot of effort into improving the digital employee experience (DEX), which has become essential to both improving organizational performance and attracting and retaining talented workers. But to date, most efforts to deliver outstanding DEX have focused on people working with laptops, PCs, or thin clients. Employees on the frontlines, using mobile devices to handle logistics ... have been largely overlooked ...

July 15, 2024

The average customer-facing incident takes nearly three hours to resolve (175 minutes) while the estimated cost of downtime is $4,537 per minute, meaning each incident can cost nearly $794,000, according to new research from PagerDuty ...

July 12, 2024

In MEAN TIME TO INSIGHT Episode 8, Shamus McGillicuddy, VP of Research, Network Infrastructure and Operations, at EMA discusses AutoCon with the conference founders Scott Robohn and Chris Grundemann ...

July 11, 2024

Numerous vendors and service providers have recently embraced the NaaS concept, yet there is still no industry consensus on its definition or the types of networks it involves. Furthermore, providers have varied in how they define the NaaS service delivery model. I conducted research for a new report, Network as a Service: Understanding the Cloud Consumption Model in Networking, to refine the concept of NaaS and reduce buyer confusion over what it is and how it can offer value ...