It's Time to Modernize Pre-Deployment Testing
August 31, 2021

Jeff Atkins

Share this

Here's how it happens: You're deploying a new technology, thinking everything's going smoothly, when the alerts start coming in. Your rollout has hit a snag. Whole groups of users are complaining about poor performance on their devices. Some can't access applications at all. You've now blown your service-level agreement (SLA). You might have just introduced a new security vulnerability. In the worst case, your big expensive product launch has missed the mark altogether.

"How did this happen?" you're asking yourself. "Didn't we test everything before we deployed?"

Yes, you did. But you made a critical though common mistake: your tests assumed ideal network conditions. And as you just learned firsthand, the idealized environment in your testing models and the way things work in the real world are two very different things.

Hopefully, this hypothetical doesn't sound too familiar. But if you're relying on traditional testing workflows and you've managed to avoid these kinds of outcomes so far, count your blessings. Because you're taking a big risk with every new launch.

There's a better way to test new enterprise technologies so they get deployed on time, under budget, with the performance you expect. To do it though, you need to get better at predicting the future. That starts with painting a more accurate picture of the present.

Navigating Complexity

Modern IT organizations already deal with more devices, more connections, and complexity than ever before. But even if you get a handle on today's technology landscape, new innovations emerge all the time. Next-generation Ethernet technologies, 5G networks, SD-WAN, Wi-Fi 6, and others can all bring important benefits to your users — benefits your competitors may already be realizing, that you can't afford to ignore. Yet, each new deployment carries significant unpredictability and risk.

All of this means it's more critical than ever to thoroughly test and validate new technology before you deploy. But all the testing in the world can't help you if you're not testing the right things. And the fact is, next-generation enterprise technologies are evolving too quickly for legacy testing approaches to keep up.

In too many cases, enterprises still test new applications and infrastructure by connecting devices directly to datacenters or clouds, with little or no traffic on the network. That kind of testing can tell you how the technology works under ideal conditions, but how often can you expect ideal conditions in the real world?

How will the technology perform on a congested or impaired network?

What kinds of problems will have the biggest impact on user experience?

Too often, those questions get answered only after deployment, when users complain. At which point customer satisfaction has already taken a hit, you may have missed an SLA, and you're looking at a time-consuming, expensive repair process.

Even more concerning, security often gets less attention than performance in pre-deployment validation. Many enterprises still rely on basic tools and firmware checks, or even just assurances from vendors, that software is safe to deploy. Which means there's a good chance you'll only learn about a vulnerability after it's been exploited, and your systems are already compromised.

A Smarter Approach

Fortunately, it's possible to predict and avoid most of these issues. To do it though, we need to recognize that testing models that worked a decade ago won't cut it anymore. We need to reimagine pre-deployment testing for today's more complex, dynamic, and distributed world.

Whatever your updated testing methodology looks like, it should include the following core practices:

Performance validation: Your vendors aren't lying when they claim to hit certain benchmarks, but you can't assume you'll achieve comparable performance in your own environment—especially if you'll be operating under an SLA. You should be measuring everything from voice quality to packet jitter. By validating real-world performance across more granular metrics, you can better evaluate any new solutions you're considering. At the same time, you identify everything you'll need to understand the user experience and troubleshoot problems post-deployment.

Network emulation: If you're going to deploy with confidence, you want to get your test beds as close as possible to real-world conditions. That includes mimicking networks, devices, and users under heavy traffic loads.

Network impairment: Network faults and service degradations are an unavoidable (if hopefully infrequent) reality. So, wouldn't you prefer to know how a new technology will respond under those conditions ahead of time? By running controlled network impairment scenarios alongside emulation, you'll know exactly how problems will affect your users, so you can better prepare. Even more important, you can set realistic expectations with customers and achievable SLAs.

Security assessments: Don't bet your security on third-party assurances or basic firmware checks. Take the time to thoroughly test for vulnerabilities, simulate known attacks, and evaluate weaknesses in the end-to-end network.

Testbed automation: To keep pace with rapidly changing networks and clouds, you should look to automate as much of the testing process as possible. The less you rely on slow, manual testing methodologies, the more quickly and cost-effectively you'll be able to simulate new scenarios as your environment evolves.

Proactive Testing Makes All the Difference

So, what happens when you put these principles into practice — when you modernize your testing to reflect a more realistic picture of your technology landscape?

First, you save time and money by identifying problems before deploying instead of after. It's a lot harder and more expensive to fix issues with a new technology when diverse users and systems already rely on it, and SLAs are already violated.

Second, you protect your users and your business by detecting and mitigating security vulnerabilities before malicious actors can exploit them. Finally, you improve your organization's ability to take advantage of new technology. By automating the testing process, you can continually bring in new testing practices and collect more valuable insights without slowing down innovation.

By overhauling your testing strategy based on realism and automation, you can put your organization in the best position to capitalize on new technologies when they emerge. You can reduce the risk of disruptive (and expensive) problems cropping up out of the blue. And, you can make ongoing innovation a core strength of your IT organization — and a key competitive advantage for your business.

Jeff Atkins is Director of Solutions Marketing at Spirent
Share this

The Latest

September 22, 2021

The world's appetite for cloud services has increased but now, more than 18 months since the beginning of the pandemic, organizations are assessing their cloud spend and trying to better understand the IT investments that were made under pressure. This is a huge challenge in and of itself, with the added complexity of embracing hybrid work ...

September 21, 2021

After a year of unprecedented challenges and change, tech pros responding to this year’s survey, IT Pro Day 2021 survey: Bring IT On from SolarWinds, report a positive perception of their roles and say they look forward to what lies ahead ...

September 20, 2021

One of the key performance indicators for IT Ops is MTTR (Mean-Time-To-Resolution). MTTR essentially measures the length of your incident management lifecycle: from detection; through assignment, triage and investigation; to remediation and resolution. IT Ops teams strive to shorten their incident management lifecycle and lower their MTTR, to meet their SLAs and maintain healthy infrastructures and services. But that's often easier said than done, with incident triage being a key factor in that challenge ...

September 16, 2021

Achieve more with less. How many of you feel that pressure — or, even worse, hear those words — trickle down from leadership? The reality is that overworked and under-resourced IT departments will only lead to chronic errors, missed deadlines and service assurance failures. After all, we're only human. So what are overburdened IT departments to do? Reduce the human factor. In a word: automate ...

September 15, 2021

On average, data innovators release twice as many products and increase employee productivity at double the rate of organizations with less mature data strategies, according to the State of Data Innovation report from Splunk ...

September 14, 2021

While 90% of respondents believe observability is important and strategic to their business — and 94% believe it to be strategic to their role — just 26% noted mature observability practices within their business, according to the 2021 Observability Forecast ...

September 13, 2021

Let's explore a few of the most prominent app success indicators and how app engineers can shift their development strategy to better meet the needs of today's app users ...

September 09, 2021

Business enterprises aiming at digital transformation or IT companies developing new software applications face challenges in developing eye-catching, robust, fast-loading, mobile-friendly, content-rich, and user-friendly software. However, with increased pressure to reduce costs and save time, business enterprises often give a short shrift to performance testing services ...

September 08, 2021

DevOps, SRE and other operations teams use observability solutions with AIOps to ingest and normalize data to get visibility into tech stacks from a centralized system, reduce noise and understand the data's context for quicker mean time to recovery (MTTR). With AI using these processes to produce actionable insights, teams are free to spend more time innovating and providing superior service assurance. Let's explore AI's role in ingestion and normalization, and then dive into correlation and deduplication too ...

September 07, 2021

As we look into the future direction of observability, we are paying attention to the rise of artificial intelligence, machine learning, security, and more. I asked top industry experts — DevOps Institute Ambassadors — to offer their predictions for the future of observability. The following are 10 predictions ...