Welcome!

SOA & WOA Authors: Elizabeth White, Carmen Gonzalez, Liz McMillan, Aria Blog, Kira Makagon

Related Topics: Cloud Expo, SOA & WOA, Virtualization, Web 2.0, Big Data Journal, SDN Journal

Cloud Expo: Article

Top Three Best Practices for Migrating to the Cloud

Planning your migration strategy

As an Infrastructure-as-a-Service provider, Bluelock sees a lot of migration of applications. Migration is occurring from physical servers to cloud, from private cloud to public cloud and back to private cloud from public cloud.

Migration can be tricky and a poor migration strategy can be responsible for costly time delays, data loss and other roadblocks on your way to successfully modernizing your infrastructure.

While each scenario is different, I'd like to identify three key best practices that will help your team create a solid, successful plan for migrating your application.

Even before you begin to move your application, there's a lot of best practice that goes into choosing which application to migrate to the cloud. Regardless of whether you are migrating that app to a public cloud or a private cloud, you should assess the app for data gravity and connectivity of the application.

Best Practice: Understand the Gravity of Your Data
Data Gravity is a concept first discussed by Dave McCrory in 2010. It's the idea that data has weight and the bigger the data is, the harder it is to move. The bigger the data, the more things are going to stick to it.

McCrory states in his original blog post about Data Gravity, "As data accumulates (builds mass) there is a greater likelihood that additional Services and Applications will be attracted to this data."

McCrory goes on to explain that large data can be virtually impossible to move because of latency and throughput issues that develop upon movement. On his website, datagravity.org, McCrory explains that to increase the portability of an application it should have a lower data gravity.

When moving tier one applications from a physical datacenter to a private or public cloud, we have to take data gravity into account because it will impact the migration.

As you are talking about migrating an application, you can think of the full stack of components as a single VM or a group of VMs that are a vApp (see Figure 1).

Think of a VM with an OS. If we were to migrate that entire VM to the public cloud, we're copying anywhere from 8-20 GB of data with that OS for no reason at all as the cloud you're migrating the app to might already have the OS available to it.

Rather than transferring the data for the OS, whenever possible use metadata instead to describe what OS you want and the configurations using a template or an image on the public or private cloud side. The same metadata concept can be applied to middleware instances too.

What we're left with is our actual data and what the app is. The app is static and static info is easy to move because you can copy it once. There's no need to replicate.

The most difficult part of the migration is the data, however. There's no easy way to shrink down the data, so you need to evaluate the weight of the data in the app you're considering migrating.

Especially if you're a high transaction company, or if it's a high transaction application, as that would be a lot of data to replicate. The data of the app constitutes 99% of the data gravity of the application.

Part of the best practice of understanding the gravity of your application is to understand the ramifications of moving a tier one application with a large amount of data and establish where the best home for that application is.

Another aspect that you should evaluate as part of your pre-migration plan is to determine how connected your VM or vApp is to other apps.

If you have a lot of applications tightly coupled to the application you want to migrate, the cloud might not be an option for that application, or at least only that application.

Best Practice: How Connected Is Your App?
Beyond what applications are connected to the app you want to migrate, the important aspect to evaluate is how coupled the application in question is to other applications, and how tight or loose of a couple they are.

Does your application have data that other applications need to access quickly? If so, a move all or nothing philosophy is your best option.

If you have an application that is tightly coupled to two or three others, you may be able to move them all to the cloud together. Because they are still tightly coupled, you won't experience the latency that would occur if your cloud-hosted application needed to access a physical server to get the data it needs to run.

A step beyond identifying how many apps are tied to the application you wish to migrate, work next to identifying which of those applications will be sensitive to latency problems.

How sensitive it can be should be a consideration of whether you should migrate the app or not.

To be able to check this best practice off your list, be very sure you understand everything your application touches so you won't be surprised later, post-migration.

The final part gets down to the nitty gritty... choosing the correct migration strategy.

Best Practice: Pick Your Migration Strategy.
Your best-fit migration strategy will be a function of the features of the application.

Option one is data migration of just the data. This is typically the correct choice for tier 1 and 2 applications.

Let's say you are able to migrate your VM or vApp. But, it's constantly changing and if it's a tier one application, we may not be able to afford a lot of downtime. Typically, we'll have to invoke some sort of replication.

Replication is an entirely separate subject, but when I think of replication, I think of the size of the data, the rate of change and the bandwidth between our source and target.

Without going into too many details of replication, let's assume you use some sort of SQL or MySQL program for database replication. What you've done is set up your new cloud to have this OS provision. You've got a MySQL provision and the two SQLs are talking to each other and replicating the data.

Option two for migrating your application is machine replication. This is best for tier 1 and tier 2 applications that can afford some downtime. It involves stack migration. There is less configuring in this scenario, but there is more data migrating.

Option two is best if you're moving to an internal private cloud. You will be able to replicate the entire stack because you have plenty of bandwidth to move stuff around.

It's important to note the portability of VMware, because VMware allows you to package the entire VM/vApp, the entire stack, into an OVF. The OVF can then be transported anywhere if you're already on a virtualized physical server.

Option three involves cold P2V migration. You typically see this for tier 2 and 3 apps that are not already virtualized.

The concept involves taking a physical app and virtualizing it. VMware has a VMware converter that does P2V, and it's very easy to go from a physical to a private cloud using P2V. It is, however, an entirely different set of best practices.

In option three, there is no replication. Those apps can also be shipped off to a public cloud provider to run in the public cloud after being virtualized.

A final path some companies take is to treat it as a Disaster Recovery (DR) scenario. Setting up something to basically do replication from one machine to another. Replicate the entire stack from point a to point b, and then click the failover button.

Each application, and migration strategy, is unique, so there is no detailed instruction manual that would work for everyone. The best strategy for some applications may be to stay put, especially if you find that steps one and two of the pre-migration evaluation is closely connected or especially weighty. To truly enjoy the benefits of cloud, you want the right application running that you can leverage to the fullest extent.

When planning your migration strategy, ask for help from those who are familiar with similar use cases and plan and evaluate extensively to save yourself a lot of time, money and headaches that come from rushing into a migration without a strategy.

More Stories By Jake Robinson

Jake Robinson is a Solutions Architect at Bluelock. He is a VCP and former CISSP and a VMware vExpert. Jake’s specialties are in infrastructure automation, virtualization, cloud computing, and security

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


@ThingsExpo Stories
The Industrial Internet revolution is now underway, enabled by connected machines and billions of devices that communicate and collaborate. The massive amounts of Big Data requiring real-time analysis is flooding legacy IT systems and giving way to cloud environments that can handle the unpredictable workloads. Yet many barriers remain until we can fully realize the opportunities and benefits from the convergence of machines and devices with Big Data and the cloud, including interoperability, data security and privacy.
Technology is enabling a new approach to collecting and using data. This approach, commonly referred to as the "Internet of Things" (IoT), enables businesses to use real-time data from all sorts of things including machines, devices and sensors to make better decisions, improve customer service, and lower the risk in the creation of new revenue opportunities. In his General Session at Internet of @ThingsExpo, Dave Wagstaff, Vice President and Chief Architect at BSQUARE Corporation, discuss the real benefits to focus on, how to understand the requirements of a successful solution, the flow of ...
Cloud Expo 2014 TV commercials will feature @ThingsExpo, which was launched in June, 2014 at New York City's Javits Center as the largest 'Internet of Things' event in the world.
"People are a lot more knowledgeable about APIs now. There are two types of people who work with APIs - IT people who want to use APIs for something internal and the product managers who want to do something outside APIs for people to connect to them," explained Roberto Medrano, Executive Vice President at SOA Software, in this SYS-CON.tv interview at Cloud Expo, held Nov 4–6, 2014, at the Santa Clara Convention Center in Santa Clara, CA.
Performance is the intersection of power, agility, control, and choice. If you value performance, and more specifically consistent performance, you need to look beyond simple virtualized compute. Many factors need to be considered to create a truly performant environment. In his General Session at 15th Cloud Expo, Harold Hannon, Sr. Software Architect at SoftLayer, discussed how to take advantage of a multitude of compute options and platform features to make cloud the cornerstone of your online presence.
In this Women in Technology Power Panel at 15th Cloud Expo, moderated by Anne Plese, Senior Consultant, Cloud Product Marketing at Verizon Enterprise, Esmeralda Swartz, CMO at MetraTech; Evelyn de Souza, Data Privacy and Compliance Strategy Leader at Cisco Systems; Seema Jethani, Director of Product Management at Basho Technologies; Victoria Livschitz, CEO of Qubell Inc.; Anne Hungate, Senior Director of Software Quality at DIRECTV, discussed what path they took to find their spot within the technology industry and how do they see opportunities for other women in their area of expertise.
DevOps Summit 2015 New York, co-located with the 16th International Cloud Expo - to be held June 9-11, 2015, at the Javits Center in New York City, NY - announces that it is now accepting Keynote Proposals. The widespread success of cloud computing is driving the DevOps revolution in enterprise IT. Now as never before, development teams must communicate and collaborate in a dynamic, 24/7/365 environment. There is no time to wait for long development cycles that produce software that is obsolete at launch. DevOps may be disruptive, but it is essential.
Wearable devices have come of age. The primary applications of wearables so far have been "the Quantified Self" or the tracking of one's fitness and health status. We propose the evolution of wearables into social and emotional communication devices. Our BE(tm) sensor uses light to visualize the skin conductance response. Our sensors are very inexpensive and can be massively distributed to audiences or groups of any size, in order to gauge reactions to performances, video, or any kind of presentation. In her session at @ThingsExpo, Jocelyn Scheirer, CEO & Founder of Bionolux, will discuss ho...
Almost everyone sees the potential of Internet of Things but how can businesses truly unlock that potential. The key will be in the ability to discover business insight in the midst of an ocean of Big Data generated from billions of embedded devices via Systems of Discover. Businesses will also need to ensure that they can sustain that insight by leveraging the cloud for global reach, scale and elasticity.
“With easy-to-use SDKs for Atmel’s platforms, IoT developers can now reap the benefits of realtime communication, and bypass the security pitfalls and configuration complexities that put IoT deployments at risk,” said Todd Greene, founder & CEO of PubNub. PubNub will team with Atmel at CES 2015 to launch full SDK support for Atmel’s MCU, MPU, and Wireless SoC platforms. Atmel developers now have access to PubNub’s secure Publish/Subscribe messaging with guaranteed ¼ second latencies across PubNub’s 14 global points-of-presence. PubNub delivers secure communication through firewalls, proxy ser...
We’re no longer looking to the future for the IoT wave. It’s no longer a distant dream but a reality that has arrived. It’s now time to make sure the industry is in alignment to meet the IoT growing pains – cooperate and collaborate as well as innovate. In his session at @ThingsExpo, Jim Hunter, Chief Scientist & Technology Evangelist at Greenwave Systems, will examine the key ingredients to IoT success and identify solutions to challenges the industry is facing. The deep industry expertise behind this presentation will provide attendees with a leading edge view of rapidly emerging IoT oppor...
The 3rd International Internet of @ThingsExpo, co-located with the 16th International Cloud Expo - to be held June 9-11, 2015, at the Javits Center in New York City, NY - announces that its Call for Papers is now open. The Internet of Things (IoT) is the biggest idea since the creation of the Worldwide Web more than 20 years ago.
Connected devices and the Internet of Things are getting significant momentum in 2014. In his session at Internet of @ThingsExpo, Jim Hunter, Chief Scientist & Technology Evangelist at Greenwave Systems, examined three key elements that together will drive mass adoption of the IoT before the end of 2015. The first element is the recent advent of robust open source protocols (like AllJoyn and WebRTC) that facilitate M2M communication. The second is broad availability of flexible, cost-effective storage designed to handle the massive surge in back-end data in a world where timely analytics is e...
"There is a natural synchronization between the business models, the IoT is there to support ,” explained Brendan O'Brien, Co-founder and Chief Architect of Aria Systems, in this SYS-CON.tv interview at the 15th International Cloud Expo®, held Nov 4–6, 2014, at the Santa Clara Convention Center in Santa Clara, CA.
The Internet of Things will put IT to its ultimate test by creating infinite new opportunities to digitize products and services, generate and analyze new data to improve customer satisfaction, and discover new ways to gain a competitive advantage across nearly every industry. In order to help corporate business units to capitalize on the rapidly evolving IoT opportunities, IT must stand up to a new set of challenges. In his session at @ThingsExpo, Jeff Kaplan, Managing Director of THINKstrategies, will examine why IT must finally fulfill its role in support of its SBUs or face a new round of...
The BPM world is going through some evolution or changes where traditional business process management solutions really have nowhere to go in terms of development of the road map. In this demo at 15th Cloud Expo, Kyle Hansen, Director of Professional Services at AgilePoint, shows AgilePoint’s unique approach to dealing with this market circumstance by developing a rapid application composition or development framework.
The Internet of Things will greatly expand the opportunities for data collection and new business models driven off of that data. In her session at @ThingsExpo, Esmeralda Swartz, CMO of MetraTech, discussed how for this to be effective you not only need to have infrastructure and operational models capable of utilizing this new phenomenon, but increasingly service providers will need to convince a skeptical public to participate. Get ready to show them the money!

ARMONK, N.Y., Nov. 20, 2014 /PRNewswire/ --  IBM (NYSE: IBM) today announced that it is bringing a greater level of control, security and flexibility to cloud-based application development and delivery with a single-tenant version of Bluemix, IBM's platform-as-a-service. The new platform enables developers to build ap...

Building low-cost wearable devices can enhance the quality of our lives. In his session at Internet of @ThingsExpo, Sai Yamanoor, Embedded Software Engineer at Altschool, provided an example of putting together a small keychain within a $50 budget that educates the user about the air quality in their surroundings. He also provided examples such as building a wearable device that provides transit or recreational information. He then reviewed the resources available to build wearable devices at home including open source hardware, the raw materials required and the options available to power s...
The Internet of Things is not new. Historically, smart businesses have used its basic concept of leveraging data to drive better decision making and have capitalized on those insights to realize additional revenue opportunities. So, what has changed to make the Internet of Things one of the hottest topics in tech? In his session at @ThingsExpo, Chris Gray, Director, Embedded and Internet of Things, discussed the underlying factors that are driving the economics of intelligent systems. Discover how hardware commoditization, the ubiquitous nature of connectivity, and the emergence of Big Data a...