Welcome!

Microservices Expo Authors: Pat Romanski, Elizabeth White, Liz McMillan, Yeshim Deniz, Zakia Bouachraoui

Related Topics: Machine Learning , IBM Cloud, @CloudExpo

Machine Learning : Blog Post

Cloud Computing and Virtual Images

A look at several different cloud computing solutions will reveal a technological enabler present in almost each one

Cloud Expo on Ulitzer

A look at several different cloud computing solutions will reveal a technological enabler present in almost each one. The enabler I'm talking about is the use of virtual images. I cannot think of many, if any, cloud computing solutions that provide software elements (i.e. more than just servers, storage, memory, etc.) that do not use virtual images in some form or fashion.

Of course, one of the reasons virtual images form the backbone of many cloud solutions is obvious. Virtual images provide the benefits of server virtualization. We can activate many virtual images on the same physical machine, thus allowing us to achieve multi-tenancy (multiple operating systems and software stacks installed on the same physical machine). Besides driving higher hardware utilization rates, it also provides us the capability to run heterogeneous software environments on the same piece of hardware. This both enables and encourages the creation of a shared pool of compute resources which is a key characteristic of cloud computing environments.

Server virtualization may be the first thing that comes to mind when we think about virtual images, but at least in the context of cloud computing, I do not believe this is the most important benefit. If we are looking at cloud computing as a means to quickly and consistently provision software environments, then I think virtual images provide us a capability and benefit more valuable than server virtualization. In this respect, virtual images provide us a medium through which we can templatize the configuration of our software environments.

Consider the case of a fairly basic application serving environment. In this environment, you are likely to install an operating system, application server, and probably some type of load balancing solution. Each of these typically requires a different piece of software, different installation procedures, and finally integration with the other components. Installing these into a typical environment, without the use of virtual images, means that you either have scripts that account for each different piece of software and finally integration of the different components, or it means that a person manually installs and integrates the pieces each time you need an environment. Either process can be time-consuming and costly to maintain over time.

Enter the use of virtual images. With a virtual image, you can install and integrate all three components ONE time, and then capture the resultant environment as a virtual image. At this point, when an application environment is needed, the virtual image can simply be activated on top of a hypervisor platform. The application environment is typically available in a much more timely fashion than if manually installed or integrated because the installation and configuration have already been captured in the virtual image.

From what I described above, you may have caught on to what would seem like a drawback of using virtual images to templatize software environments. Specifically, it may seem that you need a distinct virtual image for every unique configuration of your application environment. If this were the case, management of your virtual image library would soon become a nightmare and the resulting cost (in both resource and time) would likely outweigh the original benefits. However, thanks to a relatively new standards-based approach to virtual images, this is not necessarily a problem.

The standard I'm talking about is the Open Virtualization Format (OVF) standard which has been published by the Distributed Management Task Force (DMTF). According to the OVF standard, it "describes an open, secure, portable, efficient and extensible format for the packaging and distribution of software to be run in virtual machines." In particular to our discussion here, there is a part of the standard that describes the use of an ovf-env.xml file within the virtual image.

This file is essentially a key-value style XML file that describes desired aspects of the environment. Keys and values can be supplied during image activation, and configurations scripts that run during virtual image activation can read information from the file and react appropriately. Thus, instead of supplying N different virtual images for N different software environments, you can supply 1 virtual image and utilize the ovf-env.xml file in conjunction with configuration scripts within the image to produce N different environments. The use of this mechanism with virtual images delivers the capability to templatize software environments without sacrificing flexibility or encouraging unsustainable virtual image proliferation.

In WebSphere, we utilize the model outlined in the OVF standard when packaging our WebSphere Application Server Hypervisor Edition virtual images. This allows our WebSphere CloudBurst product to provision these images and create many different types of WebSphere Application Server environments from a single virtual image (read this article for more information). I expect the use of this standard and the mechanisms it provides will become pretty prevalent in the near future. Now if we could just get to the point where virtual disk formats are standardized, but that's an entirely different topic.

More Stories By Dustin Amrhein

Dustin Amrhein joined IBM as a member of the development team for WebSphere Application Server. While in that position, he worked on the development of Web services infrastructure and Web services programming models. In his current role, Dustin is a technical specialist for cloud, mobile, and data grid technology in IBM's WebSphere portfolio. He blogs at http://dustinamrhein.ulitzer.com. You can follow him on Twitter at http://twitter.com/damrhein.

Microservices Articles
Lori MacVittie is a subject matter expert on emerging technology responsible for outbound evangelism across F5's entire product suite. MacVittie has extensive development and technical architecture experience in both high-tech and enterprise organizations, in addition to network and systems administration expertise. Prior to joining F5, MacVittie was an award-winning technology editor at Network Computing Magazine where she evaluated and tested application-focused technologies including app secu...
Using new techniques of information modeling, indexing, and processing, new cloud-based systems can support cloud-based workloads previously not possible for high-throughput insurance, banking, and case-based applications. In his session at 18th Cloud Expo, John Newton, CTO, Founder and Chairman of Alfresco, described how to scale cloud-based content management repositories to store, manage, and retrieve billions of documents and related information with fast and linear scalability. He addresse...
Adding public cloud resources to an existing application can be a daunting process. The tools that you currently use to manage the software and hardware outside the cloud aren’t always the best tools to efficiently grow into the cloud. All of the major configuration management tools have cloud orchestration plugins that can be leveraged, but there are also cloud-native tools that can dramatically improve the efficiency of managing your application lifecycle. In his session at 18th Cloud Expo, ...
The now mainstream platform changes stemming from the first Internet boom brought many changes but didn’t really change the basic relationship between servers and the applications running on them. In fact, that was sort of the point. In his session at 18th Cloud Expo, Gordon Haff, senior cloud strategy marketing and evangelism manager at Red Hat, will discuss how today’s workloads require a new model and a new platform for development and execution. The platform must handle a wide range of rec...
Containers and Kubernetes allow for code portability across on-premise VMs, bare metal, or multiple cloud provider environments. Yet, despite this portability promise, developers may include configuration and application definitions that constrain or even eliminate application portability. In this session we'll describe best practices for "configuration as code" in a Kubernetes environment. We will demonstrate how a properly constructed containerized app can be deployed to both Amazon and Azure ...
SYS-CON Events announced today that DatacenterDynamics has been named “Media Sponsor” of SYS-CON's 18th International Cloud Expo, which will take place on June 7–9, 2016, at the Javits Center in New York City, NY. DatacenterDynamics is a brand of DCD Group, a global B2B media and publishing company that develops products to help senior professionals in the world's most ICT dependent organizations make risk-based infrastructure and capacity decisions.
Discussions of cloud computing have evolved in recent years from a focus on specific types of cloud, to a world of hybrid cloud, and to a world dominated by the APIs that make today's multi-cloud environments and hybrid clouds possible. In this Power Panel at 17th Cloud Expo, moderated by Conference Chair Roger Strukhoff, panelists addressed the importance of customers being able to use the specific technologies they need, through environments and ecosystems that expose their APIs to make true ...
In his keynote at 19th Cloud Expo, Sheng Liang, co-founder and CEO of Rancher Labs, discussed the technological advances and new business opportunities created by the rapid adoption of containers. With the success of Amazon Web Services (AWS) and various open source technologies used to build private clouds, cloud computing has become an essential component of IT strategy. However, users continue to face challenges in implementing clouds, as older technologies evolve and newer ones like Docker c...
CloudEXPO New York 2018, colocated with DXWorldEXPO New York 2018 will be held November 11-13, 2018, in New York City and will bring together Cloud Computing, FinTech and Blockchain, Digital Transformation, Big Data, Internet of Things, DevOps, AI, Machine Learning and WebRTC to one location.
DevOpsSummit New York 2018, colocated with CloudEXPO | DXWorldEXPO New York 2018 will be held November 11-13, 2018, in New York City. Digital Transformation (DX) is a major focus with the introduction of DXWorldEXPO within the program. Successful transformation requires a laser focus on being data-driven and on using all the tools available that enable transformation if they plan to survive over the long term.