Widespread Downtime Found in 99 Percent of Cloud Environments
December 12, 2017
Share this

Downtime and security risks were present in each cloud environment tested, according to 2016 Private Cloud Resiliency Benchmarks, a report from Continuity Software.

The study also found that security and performance risks were found in 99 percent and 97 percent of the environments respectively, with 82 percent of the companies facing data loss risks.

Some of the top risks identified across the private cloud environments include:

■ Configuration drifts between cluster nodes that prevent failover. Examples for such discrepancies range from the most trivial – e.g., a file that is not accessible by all hosts in the cluster – to more complex ones – such as incorrect settings of affinity rules.

■ Virtual networking configuration errors leading to virtual machine isolation and downtime. Examples include incorrect Virtual Machine Port Group configurations and resources misalignment between ESXi cluster hosts leading to a single point of failure.

■ Incorrect storage settings leading to corrupt backups and data store loss. Such risks range from invalid CBT configuration to inconsistent LUN numbering and incorrect UUID settings.

What do these private cloud environments look like?

■ 48 percent of the organizations included in the study run their virtual machines on Windows compared to 7 percent of the organizations that run on Linux. 46 percent of the organizations use a mix of operating systems.

■ Close to three quarters (73 percent) of the organizations use EMC data storage systems. Other storage systems used include NetApp (38 percent), IBM (26 percent), HP (24 percent) and Hitachi (18 percent).

■ 27 percent of the organizations use replication for automated offsite data protection.

■ 12 percent of the organizations utilize active-active failover for continuous availability.

■ Almost all of the organizations (96 percent) use more than one physical path to transfer data between the host and the external storage device.

With a growing level of the complexity, increasing interdependence among infrastructure components, and an escalating pace of change, keeping cloud infrastructure free of risky misconfiguration is becoming a challenge that most organizations fail to meet.

"Sooner or later, every system fails," said Gil Hecht, CEO of Continuity Software. "And when a popular service goes down, it doesn't take long for customers to notice."

Each year enterprises continue to encounter downtime, which currently costs an estimated $740,000 per outage according to Ponemon's most recent report.

"The good news is that most risks lurking in the cloud infrastructure can be identified and corrected before they turn into a service disruption," explained Hecht. "This requires a specialized set of processes and tools, but above all a mindset and strategy focused on early detection and the remediation of risks."

Share this

The Latest

August 10, 2020

In Episode 2, Jonah Kowall, CTO of Logz.io and former Gartner Research VP, joins the AI+ITOPS Podcast to discuss some of the hottest topics in ITOps today, including AIOps, Open Telemetry, Observability, and the challenges of Big Data in AI ...

August 07, 2020

Dennis Drogseth, VP at EMA, on the AI+ITOPS Podcast: "Digital transformation ... and the need for IT to enable digital business outcomes, is greater than ever, and all the tools including AIOps and automation ... are critical in making the difference ..."

August 06, 2020

Most organizations (75% of those surveyed) find the need to upgrade outdated infrastructure and invest in new technologies, according to the State of IT Infrastructure 2020 report from Wipro Limited ...

August 05, 2020

Application or network downtime is expensive, and given the growing numbers and types of high-availability and mission-critical applications, systems and networks — and our increasing reliance on them — ensuring consistent access to mission-critical applications is essential for ensuring customer loyalty and keeping employees productive. Businesses must recognize that applications availability depends on the network and implement a strategy to ensure network-aware application performance monitoring ...

August 04, 2020

Business leaders around the world are comfortable and optimistic about the broad scale shift toward remote work, according to the Riverbed Future of Work Global Survey 2020 that included 700 business decision makers, however findings also indicate the vast majority of organizations were not well prepared when the COVID-19 pandemic began ...

August 03, 2020

APMdigest and The Field CTO joined forces to launch the AI+ITOPS Podcast. The mission of the podcast is to discuss the struggles faced by ITOps — such as digital transformation and the need to keep IT services "always on" — and explore how AI/ML, AIOps, APM and other ITOps and DevOps technologies can help. Episode 1 features guest Dennis Drogseth, VP at Enterprise Management Associates (EMA) ...

July 30, 2020

One of the most frustrating experiences for website visitors is a slow, unresponsive website. Worst-case scenario, a web bounce causes prospects to permanently bounce from your company. In an effort to help companies improve web performance, Google launched the Web Vitals initiative in May and announced three new search engine ranking factors ...

July 29, 2020

Organizations have benefited from the use of modern applications to adapt and maintain agility and reliability during the COVID-19 pandemic, according to new research by VMware. The global study also reveals improved perceptions of alignment across app developers, IT and business decision makers as they collaborate to help their organizations operate amid the pandemic ...

July 28, 2020

According to The State of ITSM in the COVID-19 Pandemic, a survey by ManageEngine, 72% of IT professionals affirm ITSM's continued effectiveness even in remote work scenarios. However, only one in two organizations have a bring your own device (BYOD) policy to support continued productivity in new remote work environments ...

July 27, 2020

Many remote employees must access a corporate private network from home to continue business as usual. Organizations are turning to virtual private networks (VPN) as never before to keep remote workers connected to critical information and tools. To protect sensitive data and network bandwidth, however, companies must secure and control that network access such as by incorporating digital certificates into their cybersecurity strategy. Follow these five VPN best practices for secure remote worker access ...