Achieving Optimum Application Performance with Intelligent Availability
May 31, 2018

Don Boxley
DH2i

Share this

The thought of a dramatic decrease in application performance, let alone a system or application outage, sends shivers down even the hardiest IT professional's spine. That's because they know that in today's pace of business slow response time (or G-d forbid, just a few minutes of downtime) can equate to business loss both immediately and into the future, as well as potentially painful legal and/or regulations compliance consequences. There is no way around it. Availability has cemented itself as one of the most essential elements of any successful data center. However today, what many organizations are beginning to realize — sometimes the hard way — is that traditional methodologies and technologies for high availability (HA) have limits.

What's needed instead is a new approach that enables the dynamic transfer of workloads in IT environments based on optimizing the particular job at hand. Accomplishing this objective necessitates an inherent intelligence, flexibility, lack of downtime, and cost-effective methodology. What's required is intelligent availability, which builds upon some of the basic principles of high availability to provide the previously mentioned advantages — and more.

Intelligent availability is the future of availability and arguably the most critical component in the blueprint for creating business value through digital transformation.

Traditional High Availability

Historically, high availability (HA) has been defined quite simply as the continuous operation of applications and system components. Traditionally, this objective was accomplished in a variety of ways, accompanied by an assortment of drawbacks. One of the most common involves failovers, in which data and operations are transferred to those of a secondary system for scheduled downtime or unplanned failures.

Clustering methodologies are often leveraged with this approach to make resources between systems — including databases, servers, processors and others — available to one another. Clustering is applicable to VMs and physical servers and can help enable resilience for OS, host, and guest failures. Failovers involve a degree of redundancy, which entails maintaining HA by involving backups of system components. Redundant networking and storage options may be leveraged with VMs to encompass system components or data copies.

One of the most serious problems with many of these approaches is cost, especially as there are several instances in which HA is unnecessary. These pertain to the actual use and importance of servers, as well as additional factors pertaining to what virtualization techniques are used. Low priority servers that don't affect end users — such as those for testing — don't need HA, nor do those with recovery time objectives (RTO) significantly greater than their restore times.

Certain HA solutions, such as some of the more comprehensive hypervisor-based platforms, are indiscriminate in this regard. Therefore, users may end up paying for HA for components that don't need them. Also, traditional high availability approaches involve constant testing that can drain human and financial resources. Even worse, neglecting this duty can result in unplanned downtime. Also, arbitrarily implementing redundancy for system components broadens organization's data landscapes, resulting in more copies and potential weaknesses for security and data governance.

The Future: Digital Transformation

Many of these virtualization measures for HA are losing relevance because of digital transformation. To truly transform the way your organization conducts business with digitization technologies, you must deploy them strategically. Traditional HA methods simply do not allow for the fine-grained flexibility needed to optimize business value from digitization. Digital transformation means accounting for the varied computing environments of Linux and Windows operating systems alongside containers. It means integrating an assortment of legacy systems with newer ones specifically designed to handle the flood of big data and modern transactions systems.

Perhaps most importantly, it means aligning that infrastructure for business objectives in an adaptive way for changing domain or customer needs. Such flexibility is essential for optimizing IT processes around end user goals. The reality is most conventional methods of HA simply add to the infrastructural complexity of digital transformation, but don't address the primary need of adapting to evolving business requirements. In the face of digital transformation, organizations need to streamline their various IT systems around domain objectives, as opposed to doing the opposite, which simply decreases efficiency while increasing cost.

Enter Intelligent Availability

Intelligent availability is ideal for digital transformation because it enables workloads to always run on the best execution environment (BEV). It combines this advantage with the continuous operations of HA, but takes a fundamentally different approach in doing so. Intelligent availability takes the base idea of HA, to dedicate resources between systems to prevent downtime, and builds on it — extending it to moving them for maximizing competitive advantage. It allows organizations to move workloads between operating systems, servers, and physical and virtual environments with virtually no downtime.

The core of this approach is in the capacity of technologies that provide availability that includes intelligence to move workloads independent of one another, which is a fundamental limitation of traditional physical or virtualized approaches to workload management. By disengaging an array of system components (containers, application workloads, services and share files) without having to standardize on just one database or OS, these technologies transfer them to the environment which fits best from an IT goal and budgetary standpoint.

It's vital to remember that this judgment call is based on how to best achieve a defined business objective. Furthermore, these technologies provide this flexibility for individual instances to ensure negligible downtime and a smooth transition from one environment to another. The use cases for this instantaneous portability are abundant. Organizations can use these techniques for uninterrupted availability, integration with new or legacy systems, or the incorporation of additional data sources. Most importantly, they can do so with the assurance that the intelligent routing of the underlying technologies are selecting the optimal setting to execute workloads (i.e., BEV). Once suitably architected, the process takes no longer than a simple stop and start of a container or an application.

Intelligent Availability – the Intelligent Choice

Intelligent availability is important for a number of reasons. First, it creates all the advantages of HA, at a lower cost, and with a dramatically greater degree of efficiency. Next, it provides the agility required to capitalize on digital transformation, enabling organizations to quickly and easily move systems, applications, and workloads to where they can create the greatest competitive impact; and then when requirements change, move them back, or to someplace else.

As the saying goes, "The only constant is change." And in today's constantly changing business environment, intelligent availability delivers the agility required to not only survive, but to prevail.

Don Boxley is CEO and Co-Founder of DH2i
Share this

The Latest

October 20, 2020

Although cost control/expense management remains top of mind, organizations are realizing the necessity of technology solutions to enable them to steer the business during these turbulent times, according to IDG's CIO Pandemic Business Impact Study ...

October 19, 2020

The COVID-19 pandemic has compressed six years of modernization projects into 6 months. According to a recent report, IT leaders have accelerated projects aimed at increasing productivity and business agility, improving application performance and end-user experience, and driving additional revenue through existing channels ...

October 15, 2020

There is no doubt that automation has become the key aspect of modern IT management. The end-user computing market is no exception. With a large and complex technology stack and a huge number of applications, EUC specialists need to handle an ever-increasing number of changes at an ever-increasing rate. Many IT organizations are starting to realize that they can no longer control the flow of changes. It is time to think about how to facilitate change ...

October 14, 2020

Starting this September, the lifespan of an SSL/TLS certificate has been limited to 398 days, a reduction from the previous maximum certificate lifetime of 825 days. With this change, everyone needs to more carefully monitor SSL certificate expiration and server characteristics ...

October 13, 2020

Nearly 6 in 10 responding organizations have accelerated their digital transformations due to the COVID-19 pandemic, according to The IBM Institute for Business Value study COVID-19 and the Future of Business ...

October 08, 2020

Two-thirds (67%) of those surveyed expect the sheer quantity of data to grow nearly five times by 2025, according to a new report from Splunk: The Data Age Is Here. Are You Ready? ...

October 07, 2020

Gaming introduced the world to a whole new range of experiences through augmented reality (AR) and virtual reality (VR). And consumers are really catching on. To unlock the potential of these platforms, enterprises must ensure massive amounts of data can be transferred quickly and reliably to ensure an acceptable quality of experience. As such, this means that enterprises will need to turn to a 5G infrastructure powered by an adaptive network ...

October 06, 2020

A distributed, remote workforce is the new business reality. How can businesses keep operations going smoothly and quickly resolve issues when IT staff is in San Jose, employee A is working remotely in Denver at their home and employee B is a salesperson still doing some road traveling? The key is an IT architecture that promotes and supports "self-healing" at the endpoint to take care of issues before they impact employees. The essential element to achieve this is hyper-automation ...

October 05, 2020

In Episode 10, Prem Naraindas, CEO of Katonic.ai, joins the AI+ITOPS Podcast to discuss how emerging technologies can make life better for ITOps ...

October 02, 2020

Sean McDermott on the AI+ITOPS Podcast: "AIOps is really about the processing of vast amounts of data and the ability to move into a more analytical, prescriptive and automated methodology."