In BSMdigest’s exclusive interview, Al Sargent, Sr. Product Marketing Manager at VMware, discusses the Business Service Management news coming out of VMworld, and the monitoring and management challenges of the cloud.
BSM: Many companies in the Business Service Management space are exhibiting at, making announcements at and attending VMworld. Why is VMworld such an important event for monitoring and management software companies?
AS: The reason is that VMworld has become the de-facto conference for modern datacenter management. It is more than just one vendor’s conference.
Two of the most important trends in datacenter management are virtualization and cloud computing. VMware is at the center of virtualization with Vsphere, but in addition to that, VMware’s new vFabric cloud application platform provides customers with a pragmatic and evolutionary path to cloud computing.
BSM: What announcements did VMware make at VMworld to address the monitoring and management needs of the market?
AS: The main announcement was around the introduction of vFabric, and how Hyperic supports the monitoring of cloud applications. One key goal for Hyperic going forward is to provide best-in-class monitoring of cloud applications, whether running on our vFabric cloud application platform or a platform from another vendor.
Fulfilling this goal entails the following three capabilities: support for dynamic architectures and elastic capacity; extreme scalability to collect all the performance data from all the VMs in the data center; and monitoring a large number of application infrastructure components.
BSM: You mention the massive amount of performance data. Why is there so much more performance data coming out of the cloud?
AS: Think about a data center with a thousand servers, which is actually a pretty small datacenter. If you are collecting 1,000 metrics on each one of those 1,000 virtual machines every minute, that is one million metrics per minute that needs to be processed. Even a midsized firm can hit this level of metrics data.
BSM: So one factor is the number of VMs. Is another factor that there are more metrics coming out of each application because of all the changes that are going on?
AS: Exactly. That is a really good point. Another thing is this inexorable march towards web applications that streamline business processes and therefore need to accommodate surges in the business cycle. Every industry has these cycles, and applications need to be architected to accommodate the surges in demand that accompany them. As the infrastructure scales up and scales down, that is going to mean more changes in your datacenter and that is going to mean that at the peak of those surges you are going to have a lot of VMs throwing off a lot of performance metrics, and your monitoring tool has to be able to accommodate that.
BSM: So is it just a matter of scalability? The ability to handle many more metrics?
There are three elements. One is the peak number of VMs. Second, it is the fact that the number of those VMs varies over time. And third is the fact that the stakes are so much higher. During these surge periods, if your application is slow or unavailable entirely, for every minute of performance problems you are going to have significant lost revenue.
BSM: What are the current monitoring technologies missing that do not allow them be able to handle this new environment?
AS: We are going to see more metrics collected more and more frequently. Believe it or not, there are many monitoring tools that think it’s perfectly acceptable to capture metrics every 10 minutes. That might have worked fine in the late 90s when those tools first came out, but today that will not cut it.
We are seeing a need for monitoring tools that monitor very frequently - as frequently as one minute or less. There are two big drivers for this. One is the fact that consumer software is driving the enterprise software innovation. Think of Twitter: users expect that when you post something to Twitter, it is immediately available to the world. Users today expect software to work in real time, and that expectation weaves its way into the requirements for monitoring tools and reporting metrics.
The second point goes back to surges in web application workloads due to business cycles. To accommodate those surges, you need to figure out how much you need to dynamically scale up your virtualized environment. Doing that confidently requires that you collect performance metrics very frequently.
Here’s an example: Let’s say you only spin up a new app server VM if you have four datapoints indicating that the app is running slowly, because you don’t want to spin up a new VM based on a single, possibly spurious data point. If you collect metrics once every 15 minutes – a common setting among legacy tools – it will be a whole hour before you spawn a new VM. No business can afford an entire hour of sluggishness in its critical apps. You can imagine the conversation that the business would have with IT.
But let’s say you collect metrics once a minute. In four minutes, you’ll have four datapoints, and can confidently spin up that new VM. IT responds quickly, and the business and customers are happy.
BSM: At VMworld, VMware announced that the introduction of the vFabric cloud application platform will drive IT as a service. Do you see vFabric helping users of the cloud move to a more business-centric view of IT service?
Yes, vFabric lets IT move in that direction. IT can start to scale infrastructure more quickly in response to the needs of the business, and that frees them up to understand more about the cycles of the business. For instance, if IT serves a retail business, they can think about the major shopping days during the holidays, and when they’ll need to ramp up their infrastructure during those shopping days.
About Al Sargent
Al Sargent, Sr. Product Marketing Manager at VMware, handles product marketing for Hyperic, VMware's application monitoring product. He has 15+ years of experience in product management and marketing, business development, sales, and engineering at VMware, Oracle, Mercury and startups such as Wily Technology.
Enterprises that halted their cloud migration journey during the current global pandemic are two and a half times more likely than those that continued their move to the cloud to have experienced IT outages that negatively impacted their SLAs, according to Virtana's latest survey report The Current State of Hybrid Cloud and IT ...
Every business has the responsibility to do their part against climate change by reducing their carbon footprint while increasing sustainability and efficiency. Harnessing optimization of IT infrastructure is one method companies can use to reduce carbon footprint, improve sustainability and increase business efficiency, while also keeping costs down ...
While the adoption of continuous integration (CI) is on the rise, software engineering teams are unable to take a zero-tolerance approach to software failures, costing enterprise organizations billions annually, according to a quantitative study conducted by Undo and a Cambridge Judge Business School MBA project ...
I've had the opportunity to work with a number of organizations embarking on their AIOps journey. I always advise them to start by evaluating their needs and the possibilities AIOps can bring to them through five different levels of AIOps maturity. This is a strategic approach that allows enterprises to achieve complete automation for long-term success ...
Sumo Logic recently commissioned an independent market research study to understand the industry momentum behind continuous intelligence — and the necessity for digital organizations to embrace a cloud-native, real-time continuous intelligence platform to support the speed and agility of business for faster decision-making, optimizing security, driving new innovation and delivering world-class customer experiences. Some of the key findings include ...
When it comes to viruses, it's typically those of the computer/digital variety that IT is concerned about. But with the ongoing pandemic, IT operations teams are on the hook to maintain business functions in the midst of rapid and massive change. One of the biggest challenges for businesses is the shift to remote work at scale. Ensuring that they can continue to provide products and services — and satisfy their customers — against this backdrop is challenging for many ...
Teams tasked with developing and delivering software are under pressure to balance the business imperative for speed with high customer expectations for quality. In the course of trying to achieve this balance, engineering organizations rely on a variety of tools, techniques and processes. The 2020 State of Software Quality report provides a snapshot of the key challenges organizations encounter when it comes to delivering quality software at speed, as well as how they are approaching these hurdles. This blog introduces its key findings ...
For IT teams, run-the-business, commodity areas such as employee help desks, device support and communication platforms are regularly placed in the crosshairs for cost takeout, but these areas are also highly visible to employees. Organizations can improve employee satisfaction and business performance by building unified functions that are measured by employee experience rather than price. This approach will ultimately fund transformation, as well as increase productivity and innovation ...
In the agile DevOps framework, there is a vital piece missing; something that previous approaches to application development did well, but has since fallen by the wayside. That is, the post-delivery portion of the toolchain. Without continuous cloud optimization, the CI/CD toolchain still produces massive inefficiencies and overspend ...
The COVID-19 pandemic has exponentially accelerated digital transformation projects. To better understand where IT professionals are turning for help, we analyzed the online behaviors of IT decision-makers. Our research found an increase in demand for resources related to APM, microservices and dependence on cloud services ...