Welcome!

Microservices Expo Authors: Liz McMillan, AppDynamics Blog, Pat Romanski, Elizabeth White, APM Blog

Related Topics: @CloudExpo, Java IoT, Microservices Expo, Containers Expo Blog, Agile Computing, Cloud Security

@CloudExpo: Article

Goldie Locks and the Three Clouds: The Rise of the Enterprise Cloud

A cloud needs to be more than an infrastructure dispenser – providing small/medium/large chunks of infrastructure for each user

We all know the story of Goldilocks and the three bears, but have you heard the one about Goldie Locks and the three clouds? This tale is playing out throughout the IT marketplace.

Goldie Locks - an IT executive for a state government - has once again found herself in a dilemma. "If only I could choose one of the three options," she sighs. Goldie's dilemma is a result of competing requirements within her enterprise. Regarding infrastructure costs, Goldie has been told to "do more with less."

"If someone says that one more time, they're going to have porridge thrown at them," she huffs. Goldie knows that standardizing infrastructure requirements to serve the business and its processes securely, reliably and quickly is a proven way to reduce capital and operational costs. On the other hand, various business units and their departments have specific requirements for their mission-critical applications. They are resisting giving up control.

While many of her colleagues suggest she use the public cloud, Goldie believes that the security implications would be a deterrent to acceptance within her enterprise. Though she has done a thorough job investigating various cloud computing models, she needs to put together a request for proposal (RFP) to start searching for outside help with her dilemma.

She begins by taking into consideration the three standard deployment models of cloud infrastructure and their hybrid combinations - as defined by the U.S. National Institute of Standards and Technology (NIST) - and determining whether these are a fit for her enterprise:

  • Private cloud: Provisioned for exclusive use by a single organization comprising multiple consumers (e.g., business units).
  • Public cloud: Provisioned for open use by the general public.
  • Community cloud: Provisioned for exclusive use by a specific community of consumers from organizations that have shared concerns (e.g., mission, security requirements, policy, and compliance considerations)
  • Hybrid cloud: A composition of two or more distinct cloud infrastructures (private, community, or public) that remain unique entities, but are bound together by standardized or proprietary technology that enables data and application portability (e.g., cloud bursting for load balancing between clouds)

"Death by Committee" is her thought after analyzing these models. Each is too narrow for her enterprise. These cloud deployment models only provide the "how" without truly understanding the "what" and "why" of her situation. The IT side of the decision is obvious - drive down operational and capital costs to give the IT team time and money to solve strategic issues for the business. The best way to do this is to standardize processes, automate tasks and share infrastructure and administrative resources as much as possible. This is, in general, what a cloud provides.

But IT is serving a set of key stakeholders who have requirements beyond infrastructure. These stakeholders are the application/business owners who rely on IT to support delivery of their revenue-generating services and products. They are not well served with the three deployment models defined above, nor are they served by a hybrid of those models.

A Look into Goldie's Enterprise
In general, an enterprise consists of distinct parts (such as business units) that serve different customers, have different financial results and offer different products and services. They are fairly autonomous, but all operate from a common set of financial resources and processes, a common strategy and common metrics that determine success.

Consider Goldie's enterprise, which is a consolidation of a number of state agencies:

Her enterprise is partitioned into two high level branches (State Police and Transportation), where each branch consists of multiple, semi-autonomous departments. Each of those departments is interested in controlling their own infrastructure. Furthermore, their security and administrative processes differ from one department to another. Different departments must comply with their own levels of privacy, availability and service metrics.

For example, the Department of Public Info Office, within the State Police branch, may require highly predictable, millisecond response times for public users. In this instance, it may make sense to use public cloud infrastructure for the web servers, since there may be a requirement to scale up very quickly to reach high workload demands.

The Department of the Deputy Commissioner, also within the State Police branch, may also require specific infrastructure services, processes, automation and regulations, such as "hardened" OS images and encryption for all transmission of information.

The Department of Highway Administration, within the Transportation branch, must guarantee that their web site is available 99.999% of the time. They may have high availability requirements that demand duplicate resources at a disaster recovery site, as well as requirements for high-availability configurations.

NIST Deployment Models and the Enterprise - Square Peg, Round Hole
Goldie's enterprise cloud must be structured to support these multiple "parts." In turn, these parts can themselves have parts, and so on. This is similar to many of today's enterprises, which are the result of consolidating other businesses and agencies that need to function in a semi-autonomous fashion, but are still members of the larger organization.

A cloud needs to be more than an infrastructure dispenser - providing small/medium/large chunks of infrastructure for each user, without considering the unique requirements for different groups of users. Goldie knows that today's cloud products and services do not meet the needs of her enterprise stakeholders. She would like to deploy a single, centralized enterprise cloud that allows business units and their sub-units to:

  • Share underlying virtual resources as one large collection of cloud resources
  • Allow end users, such as developers, testers, demonstrators and system admins, to use a simple service catalog to manage the lifecycle of all cloud resources in the same manner
  • Set up autonomous administration, with unique policies and processes, as required
  • Allow business units to deploy their entire spectrum of applications, with unique service level objectives for development, test, production, mission-critical and regulated workloads

Now, let's see why the current cloud models cannot address these requirements.

Private Clouds
Goldie has looked at all of the currently available private cloud products. "These are too small-minded," she thinks. Every private cloud offers "multi-tenancy," which allows each business unit to manage its allotted set of cloud resources. But none of them offers any additional structure beneath the first level. Many of Goldie's business units have their own autonomous sub-units that require unique policies, processes and resources. They will want their own cloud, which does not meet her first requirement.

Public Clouds
She then turns to the available public clouds. "They are big and cheap, but my stakeholders do not want to expose their mission-critical or regulated applications." She chuckles thinking about a specific security dink she knows who actually turned pale when she suggested a public cloud for his application. On the other hand, she is painfully aware of some development teams that are slipping under the radar and deploying virtual resources in a public cloud for test and development. It's cheap and cheerful, but it's not handled by the centralized IT department and it exposes the business to risks.

Community Clouds
A community cloud offers cloud resources to a like-minded set of users / administrators. These users have agency-specific requirements, such as service levels, privacy, etc. If individual community clouds are deployed, then Goldie cannot optimize the sharing of all of the cloud resources. "This just isn't right at all," she says.

Hybrid Clouds
The final NIST deployment model does not provide any capabilities over and above the first three models. Instead, it is defined as one or more distinct instantiations of either a private, public or community cloud. Goldie has looked at all of the hybrid cloud management services and products, compared them to her requirements and decided that it doesn't meet her needs.

The Rise of an Enterprise Cloud
Through her analysis of the traditional cloud models, Goldie concludes that none of them are quite right. What she's looking for is a cloud that can address requirements unique to her enterprise. Let's refer to this as an "Enterprise Cloud." An Enterprise Cloud provides the capabilities of private, public and community clouds within a single cloud management platform that can support heterogeneous processes and requirements.

Goldie eventually conceived of the Enterprise Cloud illustrated below. It consists of a blend of internal datacenter resources, as well as resources provided by one or more public clouds. These are the "raw ingredients" that are abstracted into "cloud resources." Each agency can choose the specific cloud resources it needs to meet its requirements, including high availability, speed of deployment, cost, compliance with regulations and low latency response times.

Cloud-wide administrators, as well as specific agency and sub-agency administrators, are responsible for managing cloud resources through one "single pane of glass" interface. Aside from the properties of the cloud resources, their life cycles are all managed in the same manner, independent of where the raw materials came from. The end users of the cloud (e.g., testers, developers, infrastructure administrators) can be isolated from the underlying source of the raw resources. For example, an application could use public cloud for its web-facing tier, a low-cost set of internal cloud resources for its application tier and a highly regulated, encrypted and hardened set of cloud resources for its data layer. Goldie thinks of this as a "Hybrid Enterprise Application."

Goldie concludes that she needs to strike out on her own and develop a unique RFP that reflects her mental image of an Enterprise Cloud. If she settles for the types of clouds that are enumerated in the NIST document, she will never convince the various stakeholders to share a single cloud.

By focusing on key requirements, such as a single management framework across the enterprise, using public clouds and the datacenter to store virtual resources and providing a hierarchical multi-level tenancy structure, Goldie decides that she has finally found an Enterprise Cloud that is "juuuuuust riiiight."

More Stories By Michael A. Salsburg

Dr. Michael Salsburg is a Distinguished Engineer and Chief Cloud Solutions Architect for Unisys Corporation. He holds two international patents in infrastructure performance modeling algorithms and software. In addition, he has published more than 60 papers and has lectured worldwide on real-time infrastructure, cloud computing and infrastructure optimization. In 2010, Dr. Salsburg received the A. A. Michelson Award from the Computer Measurement Group – its highest award for lifetime achievement.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


@MicroservicesExpo Stories
Apache Hadoop is a key technology for gaining business insights from your Big Data, but the penetration into enterprises is shockingly low. In fact, Apache Hadoop and Big Data proponents recognize that this technology has not yet achieved its game-changing business potential. In his session at 19th Cloud Expo, John Mertic, director of program management for ODPi at The Linux Foundation, will explain why this is, how we can work together as an open data community to increase adoption, and the i...
DevOps and microservices are permeating software engineering teams broadly, whether these teams are in pure software shops but happen to run a business, such Uber and Airbnb, or in companies that rely heavily on software to run more traditional business, such as financial firms or high-end manufacturers. Microservices and DevOps have created software development and therefore business speed and agility benefits, but they have also created problems; specifically, they have created software sec...
Much of the value of DevOps comes from a (renewed) focus on measurement, sharing, and continuous feedback loops. In increasingly complex DevOps workflows and environments, and especially in larger, regulated, or more crystallized organizations, these core concepts become even more critical. In his session at @DevOpsSummit at 18th Cloud Expo, Andi Mann, Chief Technology Advocate at Splunk, showed how, by focusing on 'metrics that matter,' you can provide objective, transparent, and meaningful f...
Information technology is an industry that has always experienced change, and the dramatic change sweeping across the industry today could not be truthfully described as the first time we've seen such widespread change impacting customer investments. However, the rate of the change, and the potential outcomes from today's digital transformation has the distinct potential to separate the industry into two camps: Organizations that see the change coming, embrace it, and successful leverage it; and...
With the rise of Docker, Kubernetes, and other container technologies, the growth of microservices has skyrocketed among dev teams looking to innovate on a faster release cycle. This has enabled teams to finally realize their DevOps goals to ship and iterate quickly in a continuous delivery model. Why containers are growing in popularity is no surprise — they’re extremely easy to spin up or down, but come with an unforeseen issue. However, without the right foresight, DevOps and IT teams may lo...
SYS-CON Events announced today that Tintri Inc., a leading producer of VM-aware storage (VAS) for virtualization and cloud environments, will exhibit at the 19th International Cloud Expo, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. Tintri VM-aware storage is the simplest for virtualized applications and cloud. Organizations including GE, Toyota, United Healthcare, NASA and 6 of the Fortune 15 have said “No to LUNs.” With Tintri they mana...
Digitization is driving a fundamental change in society that is transforming the way businesses work with their customers, their supply chains and their people. Digital transformation leverages DevOps best practices, such as Agile Parallel Development, Continuous Delivery and Agile Operations to capitalize on opportunities and create competitive differentiation in the application economy. However, information security has been notably absent from the DevOps movement. Speed doesn’t have to negat...
The Internet of Things will challenge the status quo of how IT and development organizations operate. Or will it? Certainly the fog layer of IoT requires special insights about data ontology, security and transactional integrity. But the developmental challenges are the same: People, Process and Platform and how we integrate our thinking to solve complicated problems. In his session at 19th Cloud Expo, Craig Sproule, CEO of Metavine, will demonstrate how to move beyond today's coding paradigm ...
There is little doubt that Big Data solutions will have an increasing role in the Enterprise IT mainstream over time. Big Data at Cloud Expo - to be held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA - has announced its Call for Papers is open. Cloud computing is being adopted in one form or another by 94% of enterprises today. Tens of billions of new devices are being connected to The Internet of Things. And Big Data is driving this bus. An exponential increase is...
The Jevons Paradox suggests that when technological advances increase efficiency of a resource, it results in an overall increase in consumption. Writing on the increased use of coal as a result of technological improvements, 19th-century economist William Stanley Jevons found that these improvements led to the development of new ways to utilize coal. In his session at 19th Cloud Expo, Mark Thiele, Chief Strategy Officer for Apcera, will compare the Jevons Paradox to modern-day enterprise IT, e...
Your business relies on your applications and your employees to stay in business. Whether you develop apps or manage business critical apps that help fuel your business, what happens when users experience sluggish performance? You and all technical teams across the organization – application, network, operations, among others, as well as, those outside the organization, like ISPs and third-party providers – are called in to solve the problem.
In his general session at 18th Cloud Expo, Lee Atchison, Principal Cloud Architect and Advocate at New Relic, discussed cloud as a ‘better data center’ and how it adds new capacity (faster) and improves application availability (redundancy). The cloud is a ‘Dynamic Tool for Dynamic Apps’ and resource allocation is an integral part of your application architecture, so use only the resources you need and allocate /de-allocate resources on the fly.
SYS-CON Events announced today the Kubernetes and Google Container Engine Workshop, being held November 3, 2016, in conjunction with @DevOpsSummit at 19th Cloud Expo at the Santa Clara Convention Center in Santa Clara, CA. This workshop led by Sebastian Scheele introduces participants to Kubernetes and Google Container Engine (GKE). Through a combination of instructor-led presentations, demonstrations, and hands-on labs, students learn the key concepts and practices for deploying and maintainin...
SYS-CON Events announced today the Enterprise IoT Bootcamp, being held November 1-2, 2016, in conjunction with 19th Cloud Expo | @ThingsExpo at the Santa Clara Convention Center in Santa Clara, CA. Combined with real-world scenarios and use cases, the Enterprise IoT Bootcamp is not just based on presentations but with hands-on demos and detailed walkthroughs. We will introduce you to a variety of real world use cases prototyped using Arduino, Raspberry Pi, BeagleBone, Spark, and Intel Edison. Y...
As applications are promoted from the development environment to the CI or the QA environment and then into the production environment, it is very common for the configuration settings to be changed as the code is promoted. For example, the settings for the database connection pools are typically lower in development environment than the QA/Load Testing environment. The primary reason for the existence of the configuration setting differences is to enhance application performance. However, occas...
If you’re responsible for an application that depends on the data or functionality of various IoT endpoints – either sensors or devices – your brand reputation depends on the security, reliability, and compliance of its many integrated parts. If your application fails to deliver the expected business results, your customers and partners won't care if that failure stems from the code you developed or from a component that you integrated. What can you do to ensure that the endpoints work as expect...
Internet of @ThingsExpo, taking place November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA, is co-located with the 19th International Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world and ThingsExpo Silicon Valley Call for Papers is now open.
While DevOps promises a better and tighter integration among an organization’s development and operation teams and transforms an application life cycle into a continual deployment, Chef and Azure together provides a speedy, cost-effective and highly scalable vehicle for realizing the business values of this transformation. In his session at @DevOpsSummit at 19th Cloud Expo, Yung Chou, a Technology Evangelist at Microsoft, will present a unique opportunity to witness how Chef and Azure work tog...
When scaling agile / Scrum, we invariable run into the alignment vs autonomy problem. In short: you cannot have autonomous self directing teams if they have no clue in what direction they should go, or even shorter: Alignment breeds autonomy. But how do we create alignment? and what tools can we use to quickly evaluate if what we want to do is part of the mission or better left out? Niel Nickolaisen created the Purpose Alignment model and I use it with innovation labs in large enterprises to de...
Using new techniques of information modeling, indexing, and processing, new cloud-based systems can support cloud-based workloads previously not possible for high-throughput insurance, banking, and case-based applications. In his session at 18th Cloud Expo, John Newton, CTO, Founder and Chairman of Alfresco, described how to scale cloud-based content management repositories to store, manage, and retrieve billions of documents and related information with fast and linear scalability. He addres...