Achieving Optimum Application Performance with Intelligent Availability
May 31, 2018

Don Boxley
DH2i

Share this

The thought of a dramatic decrease in application performance, let alone a system or application outage, sends shivers down even the hardiest IT professional's spine. That's because they know that in today's pace of business slow response time (or G-d forbid, just a few minutes of downtime) can equate to business loss both immediately and into the future, as well as potentially painful legal and/or regulations compliance consequences. There is no way around it. Availability has cemented itself as one of the most essential elements of any successful data center. However today, what many organizations are beginning to realize — sometimes the hard way — is that traditional methodologies and technologies for high availability (HA) have limits.

What's needed instead is a new approach that enables the dynamic transfer of workloads in IT environments based on optimizing the particular job at hand. Accomplishing this objective necessitates an inherent intelligence, flexibility, lack of downtime, and cost-effective methodology. What's required is intelligent availability, which builds upon some of the basic principles of high availability to provide the previously mentioned advantages — and more.

Intelligent availability is the future of availability and arguably the most critical component in the blueprint for creating business value through digital transformation.

Traditional High Availability

Historically, high availability (HA) has been defined quite simply as the continuous operation of applications and system components. Traditionally, this objective was accomplished in a variety of ways, accompanied by an assortment of drawbacks. One of the most common involves failovers, in which data and operations are transferred to those of a secondary system for scheduled downtime or unplanned failures.

Clustering methodologies are often leveraged with this approach to make resources between systems — including databases, servers, processors and others — available to one another. Clustering is applicable to VMs and physical servers and can help enable resilience for OS, host, and guest failures. Failovers involve a degree of redundancy, which entails maintaining HA by involving backups of system components. Redundant networking and storage options may be leveraged with VMs to encompass system components or data copies.

One of the most serious problems with many of these approaches is cost, especially as there are several instances in which HA is unnecessary. These pertain to the actual use and importance of servers, as well as additional factors pertaining to what virtualization techniques are used. Low priority servers that don't affect end users — such as those for testing — don't need HA, nor do those with recovery time objectives (RTO) significantly greater than their restore times.

Certain HA solutions, such as some of the more comprehensive hypervisor-based platforms, are indiscriminate in this regard. Therefore, users may end up paying for HA for components that don't need them. Also, traditional high availability approaches involve constant testing that can drain human and financial resources. Even worse, neglecting this duty can result in unplanned downtime. Also, arbitrarily implementing redundancy for system components broadens organization's data landscapes, resulting in more copies and potential weaknesses for security and data governance.

The Future: Digital Transformation

Many of these virtualization measures for HA are losing relevance because of digital transformation. To truly transform the way your organization conducts business with digitization technologies, you must deploy them strategically. Traditional HA methods simply do not allow for the fine-grained flexibility needed to optimize business value from digitization. Digital transformation means accounting for the varied computing environments of Linux and Windows operating systems alongside containers. It means integrating an assortment of legacy systems with newer ones specifically designed to handle the flood of big data and modern transactions systems.

Perhaps most importantly, it means aligning that infrastructure for business objectives in an adaptive way for changing domain or customer needs. Such flexibility is essential for optimizing IT processes around end user goals. The reality is most conventional methods of HA simply add to the infrastructural complexity of digital transformation, but don't address the primary need of adapting to evolving business requirements. In the face of digital transformation, organizations need to streamline their various IT systems around domain objectives, as opposed to doing the opposite, which simply decreases efficiency while increasing cost.

Enter Intelligent Availability

Intelligent availability is ideal for digital transformation because it enables workloads to always run on the best execution environment (BEV). It combines this advantage with the continuous operations of HA, but takes a fundamentally different approach in doing so. Intelligent availability takes the base idea of HA, to dedicate resources between systems to prevent downtime, and builds on it — extending it to moving them for maximizing competitive advantage. It allows organizations to move workloads between operating systems, servers, and physical and virtual environments with virtually no downtime.

The core of this approach is in the capacity of technologies that provide availability that includes intelligence to move workloads independent of one another, which is a fundamental limitation of traditional physical or virtualized approaches to workload management. By disengaging an array of system components (containers, application workloads, services and share files) without having to standardize on just one database or OS, these technologies transfer them to the environment which fits best from an IT goal and budgetary standpoint.

It's vital to remember that this judgment call is based on how to best achieve a defined business objective. Furthermore, these technologies provide this flexibility for individual instances to ensure negligible downtime and a smooth transition from one environment to another. The use cases for this instantaneous portability are abundant. Organizations can use these techniques for uninterrupted availability, integration with new or legacy systems, or the incorporation of additional data sources. Most importantly, they can do so with the assurance that the intelligent routing of the underlying technologies are selecting the optimal setting to execute workloads (i.e., BEV). Once suitably architected, the process takes no longer than a simple stop and start of a container or an application.

Intelligent Availability – the Intelligent Choice

Intelligent availability is important for a number of reasons. First, it creates all the advantages of HA, at a lower cost, and with a dramatically greater degree of efficiency. Next, it provides the agility required to capitalize on digital transformation, enabling organizations to quickly and easily move systems, applications, and workloads to where they can create the greatest competitive impact; and then when requirements change, move them back, or to someplace else.

As the saying goes, "The only constant is change." And in today's constantly changing business environment, intelligent availability delivers the agility required to not only survive, but to prevail.

Don Boxley is CEO and Co-Founder of DH2i
Share this

The Latest

July 09, 2020

Enterprises that halted their cloud migration journey during the current global pandemic are two and a half times more likely than those that continued their move to the cloud to have experienced IT outages that negatively impacted their SLAs, according to Virtana's latest survey report The Current State of Hybrid Cloud and IT ...

July 08, 2020

Every business has the responsibility to do their part against climate change by reducing their carbon footprint while increasing sustainability and efficiency. Harnessing optimization of IT infrastructure is one method companies can use to reduce carbon footprint, improve sustainability and increase business efficiency, while also keeping costs down ...

July 07, 2020

While the adoption of continuous integration (CI) is on the rise, software engineering teams are unable to take a zero-tolerance approach to software failures, costing enterprise organizations billions annually, according to a quantitative study conducted by Undo and a Cambridge Judge Business School MBA project ...

June 25, 2020

I've had the opportunity to work with a number of organizations embarking on their AIOps journey. I always advise them to start by evaluating their needs and the possibilities AIOps can bring to them through five different levels of AIOps maturity. This is a strategic approach that allows enterprises to achieve complete automation for long-term success ...

June 24, 2020

Sumo Logic recently commissioned an independent market research study to understand the industry momentum behind continuous intelligence — and the necessity for digital organizations to embrace a cloud-native, real-time continuous intelligence platform to support the speed and agility of business for faster decision-making, optimizing security, driving new innovation and delivering world-class customer experiences. Some of the key findings include ...

June 23, 2020

When it comes to viruses, it's typically those of the computer/digital variety that IT is concerned about. But with the ongoing pandemic, IT operations teams are on the hook to maintain business functions in the midst of rapid and massive change. One of the biggest challenges for businesses is the shift to remote work at scale. Ensuring that they can continue to provide products and services — and satisfy their customers — against this backdrop is challenging for many ...

June 22, 2020

Teams tasked with developing and delivering software are under pressure to balance the business imperative for speed with high customer expectations for quality. In the course of trying to achieve this balance, engineering organizations rely on a variety of tools, techniques and processes. The 2020 State of Software Quality report provides a snapshot of the key challenges organizations encounter when it comes to delivering quality software at speed, as well as how they are approaching these hurdles. This blog introduces its key findings ...

June 18, 2020

For IT teams, run-the-business, commodity areas such as employee help desks, device support and communication platforms are regularly placed in the crosshairs for cost takeout, but these areas are also highly visible to employees. Organizations can improve employee satisfaction and business performance by building unified functions that are measured by employee experience rather than price. This approach will ultimately fund transformation, as well as increase productivity and innovation ...

June 17, 2020

In the agile DevOps framework, there is a vital piece missing; something that previous approaches to application development did well, but has since fallen by the wayside. That is, the post-delivery portion of the toolchain. Without continuous cloud optimization, the CI/CD toolchain still produces massive inefficiencies and overspend ...

June 16, 2020

The COVID-19 pandemic has exponentially accelerated digital transformation projects. To better understand where IT professionals are turning for help, we analyzed the online behaviors of IT decision-makers. Our research found an increase in demand for resources related to APM, microservices and dependence on cloud services ...