It's Time to Modernize Pre-Deployment Testing
August 31, 2021

Jeff Atkins
Spirent

Share this

Here's how it happens: You're deploying a new technology, thinking everything's going smoothly, when the alerts start coming in. Your rollout has hit a snag. Whole groups of users are complaining about poor performance on their devices. Some can't access applications at all. You've now blown your service-level agreement (SLA). You might have just introduced a new security vulnerability. In the worst case, your big expensive product launch has missed the mark altogether.

"How did this happen?" you're asking yourself. "Didn't we test everything before we deployed?"

Yes, you did. But you made a critical though common mistake: your tests assumed ideal network conditions. And as you just learned firsthand, the idealized environment in your testing models and the way things work in the real world are two very different things.

Hopefully, this hypothetical doesn't sound too familiar. But if you're relying on traditional testing workflows and you've managed to avoid these kinds of outcomes so far, count your blessings. Because you're taking a big risk with every new launch.

There's a better way to test new enterprise technologies so they get deployed on time, under budget, with the performance you expect. To do it though, you need to get better at predicting the future. That starts with painting a more accurate picture of the present.

Navigating Complexity

Modern IT organizations already deal with more devices, more connections, and complexity than ever before. But even if you get a handle on today's technology landscape, new innovations emerge all the time. Next-generation Ethernet technologies, 5G networks, SD-WAN, Wi-Fi 6, and others can all bring important benefits to your users — benefits your competitors may already be realizing, that you can't afford to ignore. Yet, each new deployment carries significant unpredictability and risk.

All of this means it's more critical than ever to thoroughly test and validate new technology before you deploy. But all the testing in the world can't help you if you're not testing the right things. And the fact is, next-generation enterprise technologies are evolving too quickly for legacy testing approaches to keep up.

In too many cases, enterprises still test new applications and infrastructure by connecting devices directly to datacenters or clouds, with little or no traffic on the network. That kind of testing can tell you how the technology works under ideal conditions, but how often can you expect ideal conditions in the real world?

How will the technology perform on a congested or impaired network?

What kinds of problems will have the biggest impact on user experience?

Too often, those questions get answered only after deployment, when users complain. At which point customer satisfaction has already taken a hit, you may have missed an SLA, and you're looking at a time-consuming, expensive repair process.

Even more concerning, security often gets less attention than performance in pre-deployment validation. Many enterprises still rely on basic tools and firmware checks, or even just assurances from vendors, that software is safe to deploy. Which means there's a good chance you'll only learn about a vulnerability after it's been exploited, and your systems are already compromised.

A Smarter Approach

Fortunately, it's possible to predict and avoid most of these issues. To do it though, we need to recognize that testing models that worked a decade ago won't cut it anymore. We need to reimagine pre-deployment testing for today's more complex, dynamic, and distributed world.

Whatever your updated testing methodology looks like, it should include the following core practices:

Performance validation: Your vendors aren't lying when they claim to hit certain benchmarks, but you can't assume you'll achieve comparable performance in your own environment—especially if you'll be operating under an SLA. You should be measuring everything from voice quality to packet jitter. By validating real-world performance across more granular metrics, you can better evaluate any new solutions you're considering. At the same time, you identify everything you'll need to understand the user experience and troubleshoot problems post-deployment.

Network emulation: If you're going to deploy with confidence, you want to get your test beds as close as possible to real-world conditions. That includes mimicking networks, devices, and users under heavy traffic loads.

Network impairment: Network faults and service degradations are an unavoidable (if hopefully infrequent) reality. So, wouldn't you prefer to know how a new technology will respond under those conditions ahead of time? By running controlled network impairment scenarios alongside emulation, you'll know exactly how problems will affect your users, so you can better prepare. Even more important, you can set realistic expectations with customers and achievable SLAs.

Security assessments: Don't bet your security on third-party assurances or basic firmware checks. Take the time to thoroughly test for vulnerabilities, simulate known attacks, and evaluate weaknesses in the end-to-end network.

Testbed automation: To keep pace with rapidly changing networks and clouds, you should look to automate as much of the testing process as possible. The less you rely on slow, manual testing methodologies, the more quickly and cost-effectively you'll be able to simulate new scenarios as your environment evolves.

Proactive Testing Makes All the Difference

So, what happens when you put these principles into practice — when you modernize your testing to reflect a more realistic picture of your technology landscape?

First, you save time and money by identifying problems before deploying instead of after. It's a lot harder and more expensive to fix issues with a new technology when diverse users and systems already rely on it, and SLAs are already violated.

Second, you protect your users and your business by detecting and mitigating security vulnerabilities before malicious actors can exploit them. Finally, you improve your organization's ability to take advantage of new technology. By automating the testing process, you can continually bring in new testing practices and collect more valuable insights without slowing down innovation.

By overhauling your testing strategy based on realism and automation, you can put your organization in the best position to capitalize on new technologies when they emerge. You can reduce the risk of disruptive (and expensive) problems cropping up out of the blue. And, you can make ongoing innovation a core strength of your IT organization — and a key competitive advantage for your business.

Jeff Atkins is Director of Solutions Marketing at Spirent
Share this

The Latest

March 27, 2023

To achieve maximum availability, IT leaders must employ domain-agnostic solutions that identify and escalate issues across all telemetry points. These technologies, which we refer to as Artificial Intelligence for IT Operations, create convergence — in other words, they provide IT and DevOps teams with the full picture of event management and downtime ...

March 23, 2023

APMdigest and leading IT research firm Enterprise Management Associates (EMA) are partnering to bring you the EMA-APMdigest Podcast, a new podcast focused on the latest technologies impacting IT Operations. In Episode 2 - Part 1 Pete Goldin, Editor and Publisher of APMdigest, discusses Network Observability with Shamus McGillicuddy, Vice President of Research, Network Infrastructure and Operations, at EMA ...

March 22, 2023

CIOs have stepped into the role of digital leader and strategic advisor, according to the 2023 Global CIO Survey from Logicalis ...

March 21, 2023

Synthetic monitoring is crucial to deploy code with confidence as catching bugs with E2E tests on staging is becoming increasingly difficult. It isn't trivial to provide realistic staging systems, especially because today's apps are intertwined with many third-party APIs ...

March 20, 2023

Recent EMA field research found that ServiceOps is either an active effort or a formal initiative in 78% of the organizations represented by a global panel of 400+ IT leaders. It is relatively early but gaining momentum across industries and organizations of all sizes globally ...

March 16, 2023

Managing availability and performance within SAP environments has long been a challenge for IT teams. But as IT environments grow more complex and dynamic, and the speed of innovation in almost every industry continues to accelerate, this situation is becoming a whole lot worse ...

March 15, 2023

Harnessing the power of network-derived intelligence and insights is critical in detecting today's increasingly sophisticated security threats across hybrid and multi-cloud infrastructure, according to a new research study from IDC ...

March 14, 2023

Recent research suggests that many organizations are paying for more software than they need. If organizations are looking to reduce IT spend, leaders should take a closer look at the tools being offered to employees, as not all software is essential ...

March 13, 2023

Organizations are challenged by tool sprawl and data source overload, according to the Grafana Labs Observability Survey 2023, with 52% of respondents reporting that their companies use 6 or more observability tools, including 11% that use 16 or more.

March 09, 2023

An array of tools purport to maintain availability — the trick is sorting through the noise to find the right one. Let us discuss why availability is so important and then unpack the ROI of deploying Artificial Intelligence for IT Operations (AIOps) during an economic downturn ...