Skip to main content

When Is It Too Early to Think About Monitoring?

David Hickman

When is it too early to think about monitoring?

In a word, “NEVER.”

I’ve heard people say that they are super busy with production deployments and fast-moving projects. They haven’t had much time to think about a performance monitoring strategy for their new applications.

But waiting until your applications are already in production is a mistake. However, it’s one that you can avoid.

Develop, test and deploy your applications with monitoring already in place and save yourself a lot of headaches. Let me count the ways:

1. Using the same monitoring tools for both pre- and post-production

Obviously, you will be doing performance testing during development and QA. But if you use different tools before and after, you might get inconsistent results. And you’ll have to retrain the people that did the original testing on the new performance monitoring tools. After they’ve moved on to other projects.

Having a monitoring strategy in place ensures that people are using the same tools throughout the lifecycle. .So that you can expect consistent measurements and more efficient troubleshooting and response if necessary.

2. Monitoring – independent from administration – eases access and alleviates concerns

Many developers need access to performance metrics during testing and often that means providing administrative access to middleware platforms to get those metrics.

When you’ve implemented an independent monitoring solution then developers can self-service their performance monitoring needs without worrying about providing administrative access.

3. Monitoring best practices implemented at design-time means less effort at run-time

Correlation of metrics between related technologies is what allows you to see things in context. For instance, show me all the servers that support a particular application. That's significantly easier than wading through 1000’s of servers trying to figure out which ones support my “banking” application. This is especially helpful when time is of the essence. But how does your monitoring system know which instances are related?

Well, you could map it manually, often using naming conventions for your services and engines. The better you stick to well-defined naming conventions, the easier it is. This is another reason why including monitoring requirements in your application building process will pay dividends down the road. Because “automatically” is so much easier than “manually”.

4. Capacity Planning is easier if you’ve been collecting performance data all along

If your business is successful, your apps will get busier and you’ll need to ask yourself if you have enough capacity for future growth? Of course, it’s hard to answer that question if you have no baseline data.

However, capturing performance data from the start enables you to see the resource usage trends over time. Correlate that to expected traffic growth for your application and you'll be on stable ground as you analyze your resource requirements for the next 6-12 months.

5. You don’t have to worry about running out of steam

For a lot of companies, once an application is written, people celebrate and move on. Things that you promise you’ll get to later – never seem to be a priority anymore. Unless, of course, you are hit with a severity one outage and then all heck breaks loose. Then people start panicking because there is no visibility into what is breaking down and where.

Building monitoring requirements into the application development cycle ensure that these things are not forgotten in the hustle and bustle of the next project. In fact, some of our customers will not sign off on a project unless monitoring is ready to go BEFORE moving into production. They make it a priority from the start so that they don’t have to struggle for resources AFTER everybody thinks they are done and move on to something else.

So, the next time somebody asks you about your monitoring strategy for your new apps, instead of saying “I’m too busy to think about it right now,” you should say, “I’m way ahead of you”.

David Hickman is in Product Marketing at SL Corporation.

APM

Hot Topics

The Latest

A major architectural shift is underway across enterprise networks, according to a new global study from Cisco. As AI assistants, agents, and data-driven workloads reshape how work gets done, they're creating faster, more dynamic, more latency-sensitive, and more complex network traffic. Combined with the ubiquity of connected devices, 24/7 uptime demands, and intensifying security threats, these shifts are driving infrastructure to adapt and evolve ...

Image
Cisco

The development of banking apps was supposed to provide users with convenience, control and piece of mind. However, for thousands of Halifax customers recently, a major mobile outage caused the exact opposite, leaving customers unable to check balances, or pay bills, sparking widespread frustration. This wasn't an isolated incident ... So why are these failures still happening? ...

Cyber threats are growing more sophisticated every day, and at their forefront are zero-day vulnerabilities. These elusive security gaps are exploited before a fix becomes available, making them among the most dangerous threats in today's digital landscape ... This guide will explore what these vulnerabilities are, how they work, why they pose such a significant threat, and how modern organizations can stay protected ...

The prevention of data center outages continues to be a strategic priority for data center owners and operators. Infrastructure equipment has improved, but the complexity of modern architectures and evolving external threats presents new risks that operators must actively manage, according to the Data Center Outage Analysis 2025 from Uptime Institute ...

As observability engineers, we navigate a sea of telemetry daily. We instrument our applications, configure collectors, and build dashboards, all in pursuit of understanding our complex distributed systems. Yet, amidst this flood of data, a critical question often remains unspoken, or at best, answered by gut feeling: "Is our telemetry actually good?" ... We're inviting you to participate in shaping a foundational element for better observability: the Instrumentation Score ...

We're inching ever closer toward a long-held goal: technology infrastructure that is so automated that it can protect itself. But as IT leaders aggressively employ automation across our enterprises, we need to continuously reassess what AI is ready to manage autonomously and what can not yet be trusted to algorithms ...

Much like a traditional factory turns raw materials into finished products, the AI factory turns vast datasets into actionable business outcomes through advanced models, inferences, and automation. From the earliest data inputs to the final token output, this process must be reliable, repeatable, and scalable. That requires industrializing the way AI is developed, deployed, and managed ...

Almost half (48%) of employees admit they resent their jobs but stay anyway, according to research from Ivanti ... This has obvious consequences across the business, but we're overlooking the massive impact of resenteeism and presenteeism on IT. For IT professionals tasked with managing the backbone of modern business operations, these numbers spell big trouble ...

For many B2B and B2C enterprise brands, technology isn't a core strength. Relying on overly complex architectures (like those that follow a pure MACH doctrine) has been flagged by industry leaders as a source of operational slowdown, creating bottlenecks that limit agility in volatile market conditions ...

FinOps champions crucial cross-departmental collaboration, uniting business, finance, technology and engineering leaders to demystify cloud expenses. Yet, too often, critical cost issues are softened into mere "recommendations" or "insights" — easy to ignore. But what if we adopted security's battle-tested strategy and reframed these as the urgent risks they truly are, demanding immediate action? ...

When Is It Too Early to Think About Monitoring?

David Hickman

When is it too early to think about monitoring?

In a word, “NEVER.”

I’ve heard people say that they are super busy with production deployments and fast-moving projects. They haven’t had much time to think about a performance monitoring strategy for their new applications.

But waiting until your applications are already in production is a mistake. However, it’s one that you can avoid.

Develop, test and deploy your applications with monitoring already in place and save yourself a lot of headaches. Let me count the ways:

1. Using the same monitoring tools for both pre- and post-production

Obviously, you will be doing performance testing during development and QA. But if you use different tools before and after, you might get inconsistent results. And you’ll have to retrain the people that did the original testing on the new performance monitoring tools. After they’ve moved on to other projects.

Having a monitoring strategy in place ensures that people are using the same tools throughout the lifecycle. .So that you can expect consistent measurements and more efficient troubleshooting and response if necessary.

2. Monitoring – independent from administration – eases access and alleviates concerns

Many developers need access to performance metrics during testing and often that means providing administrative access to middleware platforms to get those metrics.

When you’ve implemented an independent monitoring solution then developers can self-service their performance monitoring needs without worrying about providing administrative access.

3. Monitoring best practices implemented at design-time means less effort at run-time

Correlation of metrics between related technologies is what allows you to see things in context. For instance, show me all the servers that support a particular application. That's significantly easier than wading through 1000’s of servers trying to figure out which ones support my “banking” application. This is especially helpful when time is of the essence. But how does your monitoring system know which instances are related?

Well, you could map it manually, often using naming conventions for your services and engines. The better you stick to well-defined naming conventions, the easier it is. This is another reason why including monitoring requirements in your application building process will pay dividends down the road. Because “automatically” is so much easier than “manually”.

4. Capacity Planning is easier if you’ve been collecting performance data all along

If your business is successful, your apps will get busier and you’ll need to ask yourself if you have enough capacity for future growth? Of course, it’s hard to answer that question if you have no baseline data.

However, capturing performance data from the start enables you to see the resource usage trends over time. Correlate that to expected traffic growth for your application and you'll be on stable ground as you analyze your resource requirements for the next 6-12 months.

5. You don’t have to worry about running out of steam

For a lot of companies, once an application is written, people celebrate and move on. Things that you promise you’ll get to later – never seem to be a priority anymore. Unless, of course, you are hit with a severity one outage and then all heck breaks loose. Then people start panicking because there is no visibility into what is breaking down and where.

Building monitoring requirements into the application development cycle ensure that these things are not forgotten in the hustle and bustle of the next project. In fact, some of our customers will not sign off on a project unless monitoring is ready to go BEFORE moving into production. They make it a priority from the start so that they don’t have to struggle for resources AFTER everybody thinks they are done and move on to something else.

So, the next time somebody asks you about your monitoring strategy for your new apps, instead of saying “I’m too busy to think about it right now,” you should say, “I’m way ahead of you”.

David Hickman is in Product Marketing at SL Corporation.

APM

Hot Topics

The Latest

A major architectural shift is underway across enterprise networks, according to a new global study from Cisco. As AI assistants, agents, and data-driven workloads reshape how work gets done, they're creating faster, more dynamic, more latency-sensitive, and more complex network traffic. Combined with the ubiquity of connected devices, 24/7 uptime demands, and intensifying security threats, these shifts are driving infrastructure to adapt and evolve ...

Image
Cisco

The development of banking apps was supposed to provide users with convenience, control and piece of mind. However, for thousands of Halifax customers recently, a major mobile outage caused the exact opposite, leaving customers unable to check balances, or pay bills, sparking widespread frustration. This wasn't an isolated incident ... So why are these failures still happening? ...

Cyber threats are growing more sophisticated every day, and at their forefront are zero-day vulnerabilities. These elusive security gaps are exploited before a fix becomes available, making them among the most dangerous threats in today's digital landscape ... This guide will explore what these vulnerabilities are, how they work, why they pose such a significant threat, and how modern organizations can stay protected ...

The prevention of data center outages continues to be a strategic priority for data center owners and operators. Infrastructure equipment has improved, but the complexity of modern architectures and evolving external threats presents new risks that operators must actively manage, according to the Data Center Outage Analysis 2025 from Uptime Institute ...

As observability engineers, we navigate a sea of telemetry daily. We instrument our applications, configure collectors, and build dashboards, all in pursuit of understanding our complex distributed systems. Yet, amidst this flood of data, a critical question often remains unspoken, or at best, answered by gut feeling: "Is our telemetry actually good?" ... We're inviting you to participate in shaping a foundational element for better observability: the Instrumentation Score ...

We're inching ever closer toward a long-held goal: technology infrastructure that is so automated that it can protect itself. But as IT leaders aggressively employ automation across our enterprises, we need to continuously reassess what AI is ready to manage autonomously and what can not yet be trusted to algorithms ...

Much like a traditional factory turns raw materials into finished products, the AI factory turns vast datasets into actionable business outcomes through advanced models, inferences, and automation. From the earliest data inputs to the final token output, this process must be reliable, repeatable, and scalable. That requires industrializing the way AI is developed, deployed, and managed ...

Almost half (48%) of employees admit they resent their jobs but stay anyway, according to research from Ivanti ... This has obvious consequences across the business, but we're overlooking the massive impact of resenteeism and presenteeism on IT. For IT professionals tasked with managing the backbone of modern business operations, these numbers spell big trouble ...

For many B2B and B2C enterprise brands, technology isn't a core strength. Relying on overly complex architectures (like those that follow a pure MACH doctrine) has been flagged by industry leaders as a source of operational slowdown, creating bottlenecks that limit agility in volatile market conditions ...

FinOps champions crucial cross-departmental collaboration, uniting business, finance, technology and engineering leaders to demystify cloud expenses. Yet, too often, critical cost issues are softened into mere "recommendations" or "insights" — easy to ignore. But what if we adopted security's battle-tested strategy and reframed these as the urgent risks they truly are, demanding immediate action? ...