It's Time to Modernize Pre-Deployment Testing
August 31, 2021

Jeff Atkins

Share this

Here's how it happens: You're deploying a new technology, thinking everything's going smoothly, when the alerts start coming in. Your rollout has hit a snag. Whole groups of users are complaining about poor performance on their devices. Some can't access applications at all. You've now blown your service-level agreement (SLA). You might have just introduced a new security vulnerability. In the worst case, your big expensive product launch has missed the mark altogether.

"How did this happen?" you're asking yourself. "Didn't we test everything before we deployed?"

Yes, you did. But you made a critical though common mistake: your tests assumed ideal network conditions. And as you just learned firsthand, the idealized environment in your testing models and the way things work in the real world are two very different things.

Hopefully, this hypothetical doesn't sound too familiar. But if you're relying on traditional testing workflows and you've managed to avoid these kinds of outcomes so far, count your blessings. Because you're taking a big risk with every new launch.

There's a better way to test new enterprise technologies so they get deployed on time, under budget, with the performance you expect. To do it though, you need to get better at predicting the future. That starts with painting a more accurate picture of the present.

Navigating Complexity

Modern IT organizations already deal with more devices, more connections, and complexity than ever before. But even if you get a handle on today's technology landscape, new innovations emerge all the time. Next-generation Ethernet technologies, 5G networks, SD-WAN, Wi-Fi 6, and others can all bring important benefits to your users — benefits your competitors may already be realizing, that you can't afford to ignore. Yet, each new deployment carries significant unpredictability and risk.

All of this means it's more critical than ever to thoroughly test and validate new technology before you deploy. But all the testing in the world can't help you if you're not testing the right things. And the fact is, next-generation enterprise technologies are evolving too quickly for legacy testing approaches to keep up.

In too many cases, enterprises still test new applications and infrastructure by connecting devices directly to datacenters or clouds, with little or no traffic on the network. That kind of testing can tell you how the technology works under ideal conditions, but how often can you expect ideal conditions in the real world?

How will the technology perform on a congested or impaired network?

What kinds of problems will have the biggest impact on user experience?

Too often, those questions get answered only after deployment, when users complain. At which point customer satisfaction has already taken a hit, you may have missed an SLA, and you're looking at a time-consuming, expensive repair process.

Even more concerning, security often gets less attention than performance in pre-deployment validation. Many enterprises still rely on basic tools and firmware checks, or even just assurances from vendors, that software is safe to deploy. Which means there's a good chance you'll only learn about a vulnerability after it's been exploited, and your systems are already compromised.

A Smarter Approach

Fortunately, it's possible to predict and avoid most of these issues. To do it though, we need to recognize that testing models that worked a decade ago won't cut it anymore. We need to reimagine pre-deployment testing for today's more complex, dynamic, and distributed world.

Whatever your updated testing methodology looks like, it should include the following core practices:

Performance validation: Your vendors aren't lying when they claim to hit certain benchmarks, but you can't assume you'll achieve comparable performance in your own environment—especially if you'll be operating under an SLA. You should be measuring everything from voice quality to packet jitter. By validating real-world performance across more granular metrics, you can better evaluate any new solutions you're considering. At the same time, you identify everything you'll need to understand the user experience and troubleshoot problems post-deployment.

Network emulation: If you're going to deploy with confidence, you want to get your test beds as close as possible to real-world conditions. That includes mimicking networks, devices, and users under heavy traffic loads.

Network impairment: Network faults and service degradations are an unavoidable (if hopefully infrequent) reality. So, wouldn't you prefer to know how a new technology will respond under those conditions ahead of time? By running controlled network impairment scenarios alongside emulation, you'll know exactly how problems will affect your users, so you can better prepare. Even more important, you can set realistic expectations with customers and achievable SLAs.

Security assessments: Don't bet your security on third-party assurances or basic firmware checks. Take the time to thoroughly test for vulnerabilities, simulate known attacks, and evaluate weaknesses in the end-to-end network.

Testbed automation: To keep pace with rapidly changing networks and clouds, you should look to automate as much of the testing process as possible. The less you rely on slow, manual testing methodologies, the more quickly and cost-effectively you'll be able to simulate new scenarios as your environment evolves.

Proactive Testing Makes All the Difference

So, what happens when you put these principles into practice — when you modernize your testing to reflect a more realistic picture of your technology landscape?

First, you save time and money by identifying problems before deploying instead of after. It's a lot harder and more expensive to fix issues with a new technology when diverse users and systems already rely on it, and SLAs are already violated.

Second, you protect your users and your business by detecting and mitigating security vulnerabilities before malicious actors can exploit them. Finally, you improve your organization's ability to take advantage of new technology. By automating the testing process, you can continually bring in new testing practices and collect more valuable insights without slowing down innovation.

By overhauling your testing strategy based on realism and automation, you can put your organization in the best position to capitalize on new technologies when they emerge. You can reduce the risk of disruptive (and expensive) problems cropping up out of the blue. And, you can make ongoing innovation a core strength of your IT organization — and a key competitive advantage for your business.

Jeff Atkins is Director of Solutions Marketing at Spirent
Share this

The Latest

November 28, 2022

Many have assumed that the mainframe is a dying entity, but instead, a mainframe renaissance is underway. Despite this notion, we are ushering in a future of more strategic investments, increased capacity, and leading innovations ...

November 22, 2022

Most (85%) consumers shop online or via a mobile app, with 59% using these digital channels as their primary holiday shopping channel, according to the Black Friday Consumer Report from Perforce Software. As brands head into a highly profitable time of year, starting with Black Friday and Cyber Monday, it's imperative development teams prepare for peak traffic, optimal channel performance, and seamless user experiences to retain and attract shoppers ...

November 21, 2022

From staffing issues to ineffective cloud strategies, NetOps teams are looking at how to streamline processes, consolidate tools, and improve network monitoring. What are some best practices that can help achieve this? Let's dive into five ...

November 18, 2022

On November 1, Taylor Swift announced the Eras Tour ... the whole world is now standing in the same virtual queue, and even the most durable cloud architecture can't handle this level of deluge ...

November 17, 2022

OpenTelemetry, a collaborative open source observability project, has introduced a new network protocol that addresses the infrastructure management headache, coupled with collector configuration options to filter and reduce data volume ...

November 16, 2022

A unified view of digital infrastructure is essential for IT teams that must improve the digital user experience while boosting overall organizational productivity, according to a survey of IT managers in the United Arab Emirates (UAE), from Riverbed and market research firm IDC ...

November 15, 2022

Building the visibility infrastructure to make cloud networks observable is a complex technical challenge. But with careful planning and a few strategic decisions, it's possible to appropriately design, set up and manage visibility solutions for the cloud ...

November 14, 2022

According to a recent IT at Work: 2022 and Beyond study, there have been a few silver linings to the pandemic ... The study revealed some intriguing trends, which will be discussed in turn ...

November 09, 2022

The absence of topology can be a key inhibitor for AIOps tools, creating blind spots for AIOps as they only have access to event data. A topology, an IT service model, or a dependency map is a real-time picture of tools and services that are connected and dependent on each other to deliver an IT service ...

November 08, 2022

A modern data stack is a suite of technologies and apps built specifically to funnel data into an organization, transform it into actionable data, build a plan for acting on that data, and then implement that plan. The majority of modern data stacks are built on cloud-based services, composed of low- and no-code tools that enable a variety of groups within an organization to explore and use their data. Read on to learn how to optimize your data stack ...