Welcome!

Microservices Expo Authors: John Rauser, Liz McMillan, Madhavan Krishnan, VP, Cloud Solutions, Virtusa, Jason Bloomberg, Pat Romanski

Related Topics: @CloudExpo, Java IoT, Microservices Expo, Linux Containers, Cloud Security, SDN Journal

@CloudExpo: Article

The Facts About Cloud High Availability and Disaster Recovery

Understanding the facts about HA and DR in the cloud can help you make informed decisions

Enterprises are moving more and more applications to the cloud. Gartner predicts that the bulk of new IT spending by 2016 will be for cloud computing platforms and applications and that nearly half of large enterprises will have cloud deployments by the end of 2017.1

The far-reaching impact of cloud computing is summarized in a recent McKinsey report on disruptive technologies: "Cloud technology has the potential to improve productivity across $3 trillion in global enterprise IT spending, as well as enabling the creation of new online products and services for billions of consumers and millions of businesses alike."2

For many organizations, moving applications that can tolerate brief periods of downtime to the cloud is a straightforward decision with clear benefits. However, concerns about how to provide high availability and disaster protection in the cloud may make this decision more difficult for business-critical applications such as SQL, SAP, and Exchange. Understanding the facts about HA and DR in the cloud can help you make informed decisions about moving applications to the cloud, while ensuring the important business operations that depend on them are protected from downtime and data loss.

Fact #1: You need high availability protection in a cloud.
Do not assume that your cloud environment provides high availability protection, unless you have specifically configured it for HA. In fact, according to a recent study: "The average unavailability of cloud services is 10 hours per year or more, while the average availability is estimated to be 99.9% far less than the expected availability of business critical applications."3 That is the equivalent of more than a day of downtime. In fact, in 2014, Microsoft Windows Azure, Google, and Amazon Web Services all had some measure of service interruptions or downtime ranging from 4 minutes to several hours.4

For business critical applications, the redundancy that you can get with some cloud solutions, such as Windows Azure, is not enough. When you consider the cost of a minute of downtime for applications, such as SQL Server, Oracle, and SAP that may run many of your key business processes, it becomes clear that you need true high availability and disaster recovery protection. You need to ensure that end users have immediate access to data and applications in the event of a local failure, a regional disaster or anything in between.

However, the traditional way of providing high availability protection is to build a cluster using two identical servers - a primary server and a standby server -  with shared (typically SAN) storage. If the primary server fails, the application operation is moved to the standby server, which has immediate access to the same storage. The problem is that SANs are not only expensive to buy, manage, and maintain, they are simply not an option in public cloud offerings. There are, however, high availability solutions that can be used in a cloud that do not require a SAN.

Fact #2: You can build a cluster in a cloud.
Even though you cannot have a SAN in a cloud, you can build a cluster for high availability protection. In a Windows cloud, you simply add SANLess cluster software to your Windows Server Failover Cluster (WSFC). The SANLess software uses real time, block level replication to keep local storage in two geographic regions of the cloud synchronized. If there is an outage, the application operation is automatically moved to the remote instance, which has immediate access to current data. The synchronized storage looks to the WSFC like a traditional shared storage so there is no added complexity or specialized skills needed to build or manage a SANLess cluster. In fact, a SANLess cluster is easy to manage and has the added benefit of eliminating the single point of failure risk of a SAN. SANLess clusters also provide complete configuration flexibility, allowing you to replicate between physical, virtual, cloud, and hybrid cloud environment as well as between SAN and SANLess clusters.

Fact #3: You can have geographically separated nodes for DR in a cloud.
While providing high availability within the cloud will protect you from normal hardware failures and other unexpected outages within an availability zone (Amazon) or fault domain (Azure), you still need to protect against regional disasters. The easiest solution is to configure a multisite (geographically separated) cluster.

One effective method is to build a SANLess cluster within a cloud and extend it for disaster recovery by adding another node(s) in an alternate data center or a different geographic region within the cloud. Unlike traditional clusters that require you to have identical hardware and software in every node, a SANLess cluster allows you to mix physical, cloud and hybrid cloud configurations. The benefits of a DR configuration are clear. For example, simply adding a third, geographically separated node to your SANLess cluster in a Windows Azure cloud can give you a recovery point objective (RPO) of near zero data loss and a recovery time objective (RTO) of just about one minute.

Fact #4: You can create a cluster that mixes cloud and on-premises nodes.
You can use your on-premises data center as your primary location with a failover cluster to provide high availability protection and use the cloud as your hot standby DR site. This is a very cost-effective alternative to building out your own DR site, or renting rack space in a business continuity facility. In this case, the on-premises servers can be your choice of traditional SAN-based clusters, SANLess clusters, or even single servers not currently participating in a cluster.

The objective of having a "hot" standby DR site is to have standby servers up and running as quickly as possible in the DR site with access to a copy of the most recent application data. In the event of a disaster, recovery is automatic and immediate. A multisite cluster is an effective way to implement a hot standby DR site. In this case, the SANLess date. In the event of a forecasted disaster, such as a storm or a flood, applications can be moved to the cloud before potential disaster strikes. In the event of an unexpected disaster, applications can be recovered manually or in some cases automatically, depending upon the quorum configuration. This mix of cloud and on-premises nodes gives you an excellent RTO and RPO with minimal investment in infrastructure.

Fact #5: HA and DR in a cloud can be easy and highly cost-effective.
If you choose a SANLess software that provides an intuitive configuration interface, you can create a standard WSFC in a cloud in minutes without specialized skills. A SANLess cluster can help you realize significant cost savings in several ways. First, in a Microsoft SQL Server environment a SANLess cluster can give you high availability with SQL Server Standard Edition software licenses without requiring you to upgrade to costly SQL Server Enterprise Edition.

Second, you can realize hundreds of thousands of dollars in savings with a SANLess by eliminating the total cost of ownership (TCO) associated with a SAN. The savings in TCO include the SAN hardware acquisition costs; the power, cooling, and data center floor space costs; and the ongoing labor cost of specialized SAN administration.

If you are thinking about moving your important applications to the cloud, you need to consider how you will protect those applications from downtime and data loss. While traditional SAN-based clusters are not possible in these environments, SANLess clusters can provide an easy, cost-efficient alternative. These clusters not only provide high availability protection, but also enable significantly greater configuration flexibility and potentially dramatic savings in both licensing costs and SAN TCO.

Notes

1"Gartner Says Cloud Computing Will Become the Bulk of New IT Spend by 2016."

2 Manyika, James and Michael Chui, et al, "Disruptive technologies: Advances that will transform life, business, and the global economy," McKinsey Global Institute (May 2013) 

3Whittaker, Josh, "Amazon Web Services Suffers Outage, Takes Out Vine, Instagram, Others with it," ZDNet, (August 26, 2013)

4Mackay, Martin, "Downtime Report: Top Ten Outages in 2013," Business2Community.com, (December 2013)

More Stories By Jerry Melnick

Jerry Melnick ([email protected]) is responsible for defining corporate strategy and operations at SIOS Technology Corp. (www.us.sios.com), maker of SIOS SAN and #SANLess cluster software (www.clustersyourway.com). He more than 25 years of experience in the enterprise and high availability software industries. He holds a Bachelor of Science degree from Beloit College with graduate work in Computer Engineering and Computer Science at Boston University.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


@MicroservicesExpo Stories
While we understand Agile as a means to accelerate innovation, manage uncertainty and cope with ambiguity, many are inclined to think that it conflicts with the objectives of traditional engineering projects, such as building a highway, skyscraper or power plant. These are plan-driven and predictive projects that seek to avoid any uncertainty. This type of thinking, however, is short-sighted. Agile approaches are valuable in controlling uncertainty because they constrain the complexity that ste...
Agile has finally jumped the technology shark, expanding outside the software world. Enterprises are now increasingly adopting Agile practices across their organizations in order to successfully navigate the disruptive waters that threaten to drown them. In our quest for establishing change as a core competency in our organizations, this business-centric notion of Agile is an essential component of Agile Digital Transformation. In the years since the publication of the Agile Manifesto, the conn...
"This all sounds great. But it's just not realistic." This is what a group of five senior IT executives told me during a workshop I held not long ago. We were working through an exercise on the organizational characteristics necessary to successfully execute a digital transformation, and the group was doing their ‘readout.' The executives loved everything we discussed and agreed that if such an environment existed, it would make transformation much easier. They just didn't believe it was reali...
The cloud revolution in enterprises has very clearly crossed the phase of proof-of-concepts into a truly mainstream adoption. One of most popular enterprise-wide initiatives currently going on are “cloud migration” programs of some kind or another. Finding business value for these programs is not hard to fathom – they include hyperelasticity in infrastructure consumption, subscription based models, and agility derived from rapid speed of deployment of applications. These factors will continue to...
"Opsani helps the enterprise adopt containers, help them move their infrastructure into this modern world of DevOps, accelerate the delivery of new features into production, and really get them going on the container path," explained Ross Schibler, CEO of Opsani, and Peter Nickolov, CTO of Opsani, in this SYS-CON.tv interview at DevOps Summit at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
"We're developing a software that is based on the cloud environment and we are providing those services to corporations and the general public," explained Seungmin Kim, CEO/CTO of SM Systems Inc., in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
Enterprises are adopting Kubernetes to accelerate the development and the delivery of cloud-native applications. However, sharing a Kubernetes cluster between members of the same team can be challenging. And, sharing clusters across multiple teams is even harder. Kubernetes offers several constructs to help implement segmentation and isolation. However, these primitives can be complex to understand and apply. As a result, it’s becoming common for enterprises to end up with several clusters. Thi...
"Codigm is based on the cloud and we are here to explore marketing opportunities in America. Our mission is to make an ecosystem of the SW environment that anyone can understand, learn, teach, and develop the SW on the cloud," explained Sung Tae Ryu, CEO of Codigm, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
"CA has been doing a lot of things in the area of DevOps. Now we have a complete set of tool sets in order to enable customers to go all the way from planning to development to testing down to release into the operations," explained Aruna Ravichandran, Vice President of Global Marketing and Strategy at CA Technologies, in this SYS-CON.tv interview at DevOps Summit at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
The nature of test environments is inherently temporary—you set up an environment, run through an automated test suite, and then tear down the environment. If you can reduce the cycle time for this process down to hours or minutes, then you may be able to cut your test environment budgets considerably. The impact of cloud adoption on test environments is a valuable advancement in both cost savings and agility. The on-demand model takes advantage of public cloud APIs requiring only payment for t...
Cavirin Systems has just announced C2, a SaaS offering designed to bring continuous security assessment and remediation to hybrid environments, containers, and data centers. Cavirin C2 is deployed within Amazon Web Services (AWS) and features a flexible licensing model for easy scalability and clear pay-as-you-go pricing. Although native to AWS, it also supports assessment and remediation of virtual or container instances within Microsoft Azure, Google Cloud Platform (GCP), or on-premise. By dr...
Let's do a visualization exercise. Imagine it's December 31, 2018, and you're ringing in the New Year with your friends and family. You think back on everything that you accomplished in the last year: your company's revenue is through the roof thanks to the success of your product, and you were promoted to Lead Developer. 2019 is poised to be an even bigger year for your company because you have the tools and insight to scale as quickly as demand requires. You're a happy human, and it's not just...
Many enterprise and government IT organizations are realizing the benefits of cloud computing by extending IT delivery and management processes across private and public cloud services. But they are often challenged with balancing the need for centralized cloud governance without stifling user-driven innovation. This strategy requires an approach that fundamentally reshapes how IT is delivered today, shifting the focus from infrastructure to services aggregation, and mixing and matching the bes...
identify the sources of event storms and performance anomalies will require automated, real-time root-cause analysis. I think Enterprise Management Associates said it well: “The data and metrics collected at instrumentation points across the application ecosystem are essential to performance monitoring and root cause analysis. However, analytics capable of transforming data and metrics into an application-focused report or dashboards are what separates actual application monitoring from relat...
The benefits of automation are well documented; it increases productivity, cuts cost and minimizes errors. It eliminates repetitive manual tasks, freeing us up to be more innovative. By that logic, surely, we should automate everything possible, right? So, is attempting to automate everything a sensible - even feasible - goal? In a word: no. Consider this your short guide as to what to automate and what not to automate.
DevOps teams have more on their plate than ever. As infrastructure needs grow, so does the time required to ensure that everything's running smoothly. This makes automation crucial - especially in the server and network monitoring world. Server monitoring tools can save teams time by automating server management and providing real-time performance updates. As budgets reset for the New Year, there is no better time to implement a new server monitoring tool (or re-evaluate your current solution)....
While some developers care passionately about how data centers and clouds are architected, for most, it is only the end result that matters. To the majority of companies, technology exists to solve a business problem, and only delivers value when it is solving that problem. 2017 brings the mainstream adoption of containers for production workloads. In his session at 21st Cloud Expo, Ben McCormack, VP of Operations at Evernote, discussed how data centers of the future will be managed, how the p...
We just came off of a review of a product that handles both containers and virtual machines in the same interface. Under the covers, implementation of containers defaults to LXC, though recently Docker support was added. When reading online, or searching for information, increasingly we see “Container Management” products listed as competitors to Docker, when in reality things like Rocket, LXC/LXD, and Virtualization are Dockers competitors. After doing some looking around, we have decided tha...
High-velocity engineering teams are applying not only continuous delivery processes, but also lessons in experimentation from established leaders like Amazon, Netflix, and Facebook. These companies have made experimentation a foundation for their release processes, allowing them to try out major feature releases and redesigns within smaller groups before making them broadly available. In his session at 21st Cloud Expo, Brian Lucas, Senior Staff Engineer at Optimizely, discussed how by using ne...
Digital transformation has changed the way users interact with the world, and the traditional healthcare experience no longer meets rising consumer expectations. Enterprise Health Clouds (EHCs) are designed to easily and securely deliver the smart and engaging digital health experience that patients expect today, while ensuring the compliance and data integration that care providers require. Jikku Venkat