Welcome!

Microservices Expo Authors: Elizabeth White, Pat Romanski, Liz McMillan, Derek Weeks, LeanTaaS Blog

Related Topics: Microservices Expo

Microservices Expo: Article

BCP Lessons Learned and New Ideas for IT Infrastructure Continuity

Learn How to Justify the Creation of Disaster Recovery Facilities

Businesses in the southeastern United States have been hit hard with hurricanes in the last few years, and 2008 was no exception. As a project manager and CBCP for over 1600 disaster recovery deployments I can share real examples of how entire data centers were failed over to the DR operations center in preparation for hurricanes, while others (due to poor planning) did not have the same success. Those that were successful were efficient in organizing the RTO of their communication servers which helped them prioritize the recovery efforts as well utilize creative testing procedures in order to not disrupt normal business activity. The first priority of a BCP is to ensure the safety of the employees, but being able to communicate to those needed is also an important step for successfully executing a BCP. Because of this preparedness many businesses I have heard from were able to proactively allow their employees evacuate and still provide them remote access for business operations from almost anywhere. I will review a few of the  examples of architecture, solutions and best practices for exercising controls in those events as well as discuss what future technology may be utilized to better help justify the creation of disaster recovery facilities.

10 Professional Practices for BCP
There are ten professional practices for business continuity planning; all equally important and if followed appropriately will allow you to create a solid foundation to build upon. For the purpose of this article I will summarize the professional practices, but for more information visit the Disaster Recovery International Institute (www.drii.org). DRII is an excellent resource for BCP and is a consortium of business continuity professionals dedicated to setting industry standards and sharing knowledge around the practice of business continuity management.

The first step in building a BCP is Program Initiation and Management. This step is designed to establish executive approval, support and justification for the need of a resiliency program. Start with building a dedicated team that is committed to supporting the BCP initiative and selecting team members that can effectively manage roles and responsibilities for their portion of the plan. Cost justification is often a hurdle in establishing the need for disaster recovery facilities, so one tip would be to utilize your current assets such as other offices or co-location facilities. You can also work with the IT department to help tie in the IT management budget into the BCP so that you are not just providing continuity in the event of a disaster, but also high availability for day-to-day operational maintenance.

The next couple of steps are important in determining the risk (risk evaluation) your organization faces from either a natural or environment disaster perspective and then determine the business impact (BIA) should one of those events occur. This will help you determine the next step in the business continuity strategy you design and implement to meet your defined recovery point (RPO) and time (RTO) objectives. Once those objectives and controls are defined you will need to integrate emergency response and operations in order to define the process in which a disaster is declared and what prompts the initiation of the BCP.

These previous steps are what allow you to design and implement a comprehensive strategy that meets the requirements of your company’s objectives. I have seen companies try to short cut these previous steps and immediately skip to implementing a solution, only to find out that their infrastructure doesn’t have enough power, bandwidth, resources and or executive approval to support the controls implemented. So the lesson learned is, don’t try to take short cuts and jump into something you have never done before. Following the previous steps will allow you to proceed and likely prevent challenges you may face during the deployment and execution of your plan.

The next three steps include designing and implementing the BCP, generating awareness and training your organization on what to do in event of a disaster, then exercising those plans regularly. Exercising BCP is typically recommended to be tied to your change control process which means the plan should be reviewed any time there is a change within the organization that may affect the plan. (That can be anything as small as a software update to some of the business critical servers to a BCP member leaving the company.) Depending on the situation, exercises could take place as frequently as once a month or at very least 2-3 times per year so that there is a consistent awareness of the plan and procedures.

The last two practices, crisis communication and coordinating with external agencies is really the culmination of the previous practices and will ultimately be the success or failure of your plan. In the event of a disaster, communication is critical to coordinating with emergency responders and your own business continuity team to make sure evacuations and safety procedures are implemented effectively.

When Planning and Exercising is Done Right
Planning is your best friend when it comes to rolling out controls for a business continuity solution. Starting with executive buy in though budget, infrastructure, process, procedures, testing and ultimately execution you can’t plan enough. And when it’s done right deployments go smoothly. However, is more than one way to go about this. As the saying goes “Don’t eat the elephant all in one bite”. Breaking down your overall rollout plan into smaller projects will help you better manage details as well as prioritize the order of the overall deployment. Here are some quotes from companies who did it right and were glad they did after Hurricane Ike made landfall:

  • “All is OK and thanks. Our files were mirrored to our Austin facility with no loss of data or applications. Winds tore a 30'x30' hole in the building roof.  The water damage was bad.  The computer servers were spared but alot of workstations were soaked.  Houston operations were running in Austin just before the hurricane hit and the transfer was seamless.”
  • “Thanks, our company is doing just fine. With our replicated data to one of our other locations, we were up and seeing patients once the patients could get to us. We appreciate your concern, and your overall support of our organization. On behalf of our organization, we want to say thank you!”
  • “Yes we did make it out alive; we activated our business contingency plan, and relocated to Dallas. Luckily our solution allowed us to failover and business continued. “

Exercising the business continuity plan on a regular basis helped these companies not only be prepared but assured that they were ready for anything. And with the adaption of new technologies for IT infrastructure, testing those plans are even easier to exercise while minimizing impact to production operations. In previous years testing business continuity plans for the data center usually required shutting down the entire production facility and running through the restoration process. With the adoption of real-time replication software, co-location facilities and virtualization testing can be accomplished with minimal impact to a production environment. If you have a dedicated disaster recovery facility with hot standby servers you could just segment the networks from each other and bring the site online. However, you had to be very careful about making sure those two sites weren’t talking to each other via domains or active directory services.

How Dynamic Infrastructure Is being used to facilitate BCP Exercise

Dynamic Infrastructure is defined by some as ‘the ability to rapidly move and provision workloads with security and inherent protection’. It may be a new idea to you, but it is being adopted within the IT community with great success. Dynamic Infrastructure not only simplifies the disaster recovery procedures for data center managers, but also provides the ability to use those same controls for day-to-day operations to keep your business operations available all the time - not just during disasters. With the adoption of virtualization technologies saving costs on hardware, power and cooling, data center management budgets can be combined with BCP for maximizing infrastructure availability. These technologies also assist BCP exercises by simulating recovery servers and sites without bringing down production servers. Some solutions like VMware® Site Recovery Manager have this feature but also have some inherent issues. For instance, in the event of a real disaster the virtual solution doesn’t have any failback capability. Typically once that process has been started there is no turning back without a complete restoration which could take days depending on the number of systems and or volume of data that needed to be restored. Dynamic Infrastructure provides the functionality that others are missing as well as allowing for rapid failback capabilities for smaller or “little d” disasters, which are more likely to impact a business critical system.

The Next Generation of BCP
With future technology delivering Dynamic Infrastructure, cloud computing and mobile communication devices, learning how they can protect IT infrastructure for Business Continuity Planning has never been more important. Many management services are offering remote or mobile access for initiating some of these data center management functions. Imagine if you could initiate a failover of a server via your iPhone or BlackBerry®. The reality is that it isn’t very far off. It’s  possible that many business- critical services could be run via cloud computing so that services are available anywhere they are needed - even if there was a disaster at the production facility.

However, this begs the question. Who is protecting the cloud and what is their business continuity plan?

More Stories By Brace Rennels

Brace Rennels is a passionate and experienced interactive marketing professional who thrives on building high energy marketing teams to drive global web strategies, SEO, social media and online PR web marketing. Recognized as an early adopter of technology and applying new techniques to innovative creative marketing, drive brand awareness, lead generation and revenue. As a Sr. Manager Global of Website Strategies his responsibilities included developing and launching global social media, SEO and web marketing initiatives and strategy. Recognized for applying innovative solutions to address unique problems and manage business relationships to effectively accomplish enterprise objectives. An accomplished writer, blogger and author for several publications on various marketing, social media and technical subjects such as industry trends, cloud computing, virtualization, website marketing, disaster recovery and business continuity. Publications include CIO.com, Enterprise Storage Journal, TechNewsWorld, Sys-Con, eWeek and Peer to Peer Magazine. Follow more of Brace's writing on his blog: http://bracerennels.com

@MicroservicesExpo Stories
"I focus on what we are calling CAST Highlight, which is our SaaS application portfolio analysis tool. It is an extremely lightweight tool that can integrate with pretty much any build process right now," explained Andrew Siegmund, Application Migration Specialist for CAST, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
SYS-CON Events announced today that Synametrics Technologies will exhibit at SYS-CON's 22nd International Cloud Expo®, which will take place on June 5-7, 2018, at the Javits Center in New York, NY. Synametrics Technologies is a privately held company based in Plainsboro, New Jersey that has been providing solutions for the developer community since 1997. Based on the success of its initial product offerings such as WinSQL, Xeams, SynaMan and Syncrify, Synametrics continues to create and hone inn...
As many know, the first generation of Cloud Management Platform (CMP) solutions were designed for managing virtual infrastructure (IaaS) and traditional applications. But that's no longer enough to satisfy evolving and complex business requirements. In his session at 21st Cloud Expo, Scott Davis, Embotics CTO, explored how next-generation CMPs ensure organizations can manage cloud-native and microservice-based application architectures, while also facilitating agile DevOps methodology. He expla...
DevOps promotes continuous improvement through a culture of collaboration. But in real terms, how do you: Integrate activities across diverse teams and services? Make objective decisions with system-wide visibility? Use feedback loops to enable learning and improvement? With technology insights and real-world examples, in his general session at @DevOpsSummit, at 21st Cloud Expo, Andi Mann, Chief Technology Advocate at Splunk, explored how leading organizations use data-driven DevOps to close th...
"WineSOFT is a software company making proxy server software, which is widely used in the telecommunication industry or the content delivery networks or e-commerce," explained Jonathan Ahn, COO of WineSOFT, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
Some people are directors, managers, and administrators. Others are disrupters. Eddie Webb (@edwardawebb) is an IT Disrupter for Software Development Platforms at Liberty Mutual and was a presenter at the 2016 All Day DevOps conference. His talk, Organically DevOps: Building Quality and Security into the Software Supply Chain at Liberty Mutual, looked at Liberty Mutual's transformation to Continuous Integration, Continuous Delivery, and DevOps. For a large, heavily regulated industry, this task ...
The notion of improving operational efficiency is conspicuously absent from the healthcare debate - neither Obamacare nor the newly proposed GOP plan discusses the impact that a step-function improvement in efficiency could have on access to healthcare (through more capacity), quality of healthcare services (through reduced wait times for patients) or cost (through better utilization of scarce, expensive assets).
The past few years have seen a huge increase in the amount of critical IT services that companies outsource to SaaS/IaaS/PaaS providers, be it security, storage, monitoring, or operations. Of course, along with any outsourcing to a service provider comes a Service Level Agreement (SLA) to ensure that the vendor is held financially responsible for any lapses in their service which affect the customer’s end users, and ultimately, their bottom line. SLAs can be very tricky to manage for a number ...
The dynamic nature of the cloud means that change is a constant when it comes to modern cloud-based infrastructure. Delivering modern applications to end users, therefore, is a constantly shifting challenge. Delivery automation helps IT Ops teams ensure that apps are providing an optimal end user experience over hybrid-cloud and multi-cloud environments, no matter what the current state of the infrastructure is. To employ a delivery automation strategy that reflects your business rules, making r...
Modern software design has fundamentally changed how we manage applications, causing many to turn to containers as the new virtual machine for resource management. As container adoption grows beyond stateless applications to stateful workloads, the need for persistent storage is foundational - something customers routinely cite as a top pain point. In his session at @DevOpsSummit at 21st Cloud Expo, Bill Borsari, Head of Systems Engineering at Datera, explored how organizations can reap the bene...
The past few years have brought a sea change in the way applications are architected, developed, and consumed—increasing both the complexity of testing and the business impact of software failures. How can software testing professionals keep pace with modern application delivery, given the trends that impact both architectures (cloud, microservices, and APIs) and processes (DevOps, agile, and continuous delivery)? This is where continuous testing comes in. D
In a recent post, titled “10 Surprising Facts About Cloud Computing and What It Really Is”, Zac Johnson highlighted some interesting facts about cloud computing in the SMB marketplace: Cloud Computing is up to 40 times more cost-effective for an SMB, compared to running its own IT system. 94% of SMBs have experienced security benefits in the cloud that they didn’t have with their on-premises service
There is a huge demand for responsive, real-time mobile and web experiences, but current architectural patterns do not easily accommodate applications that respond to events in real time. Common solutions using message queues or HTTP long-polling quickly lead to resiliency, scalability and development velocity challenges. In his session at 21st Cloud Expo, Ryland Degnan, a Senior Software Engineer on the Netflix Edge Platform team, will discuss how by leveraging a reactive stream-based protocol,...
Admiral Calcote - also known as Lee Calcote (@lcalcote) or the Ginger Geek to his friends - gave a presentation entitled Characterizing and Contrasting Container Orchestrators at the 2016 All Day DevOps conference. Okay, he isn't really an admiral - nor does anyone call him that - but he used the title admiral to describe what container orchestrators do, relating it to an admiral directing a fleet of container ships. You could also say that they are like the conductor of an orchestra, directing...
Our work, both with clients and with tools, has lead us to wonder how it is that organizations are handling compliance issues in the cloud. The big cloud vendors offer compliance for their infrastructure, but the shared responsibility model requires that you take certain steps to meet compliance requirements. Which lead us to start poking around a little more. We wanted to get a picture of what was available, and how it was being used. There is a lot of fluidity in this space, as in all things c...
The goal of Microservices is to improve software delivery speed and increase system safety as scale increases. Microservices being modular these are faster to change and enables an evolutionary architecture where systems can change, as the business needs change. Microservices can scale elastically and by being service oriented can enable APIs natively. Microservices also reduce implementation and release cycle time and enables continuous delivery. This paper provides a logical overview of the Mi...
Gaining visibility in today’s sprawling cloud infrastructure is complex and laborious, involving drilling down into tools offered by various cloud services providers. Enterprise IT organizations need smarter and effective tools at their disposal in order to address this pertinent problem. Gaining a 360 - degree view of the cloud costs requires collection and analysis of the cost data across all cloud infrastructures used inside an enterprise.
"We started a Master of Science in business analytics - that's the hot topic. We serve the business community around San Francisco so we educate the working professionals and this is where they all want to be," explained Judy Lee, Associate Professor and Department Chair at Golden Gate University, in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
Gone are the days when application development was the daunting task of the highly skilled developers backed with strong IT skills, low code application development has democratized app development and empowered a new generation of citizen developers. There was a time when app development was in the domain of people with complex coding and technical skills. We called these people by various names like programmers, coders, techies, and they usually worked in a world oblivious of the everyday pri...
The “Digital Era” is forcing us to engage with new methods to build, operate and maintain applications. This transformation also implies an evolution to more and more intelligent applications to better engage with the customers, while creating significant market differentiators. In both cases, the cloud has become a key enabler to embrace this digital revolution. So, moving to the cloud is no longer the question; the new questions are HOW and WHEN. To make this equation even more complex, most ...