Microservices Expo Authors: Carmen Gonzalez, Lori MacVittie, Liz McMillan, Yeshim Deniz, Pat Romanski

Related Topics: Containers Expo Blog, Java IoT, Microservices Expo, Microsoft Cloud, Linux Containers, SDN Journal

Containers Expo Blog: Blog Feed Post

The Re-Emergence of the Operating System

Each of the major infrastructure silos has to operate with some fixed environment in mind

Compute started its major architectural transition several years ago with the introduction of virtualization. If you pay attention to any of the IT noise today, it should be clear that storage and networking are going through their own architectural evolutions as well. But another shift is also underway: applications are fundamentally changing as well.

An interesting dynamic in all of this is that it is near impossible for each of the four major IT areas to undergo simultaneous, coordinated evolution. Change is hard enough on its own, but changing multiple variable at once makes it difficult to anchor to anything substantial. And when change does occur along multiple fronts at the same time, the task of determining causation for new found results is challenging at best.

Understanding that business underlies much of the change, the best that the industry can collectively do is to take some things as fixed and then change around that. And so we evolve each of the silos somewhat independently, trying hard to keep in mind that the environment into which they plug is also changing. At our best, we try to intersect the various changes. But when we miss, we tend to aim towards slotting into current architectural paradigms, because that allows us the best chance to make a meaningful difference in production environments. Indeed, always playing out ahead of the horizon might be great for making future progress, but it makes building a business around the resulting innovation nigh impossible.

Moving forward

And so IT as a whole continues the trudge forward.

Each of the major infrastructure silos—compute, storage, and networking—has to operate with some fixed environment in mind. The most basic thing to attach to is application infrastructure. While virtualization has made compute containers that make applications portable (the reason dynamic and change are so prominent in most marketing materials), the applications themselves have remained largely static.

Except, of course, that they haven’t.

Application architectures

Anyone watching the major web-scale players (think: Facebook, Google, Twitter, and the like) will know that the architecture for their applications is actually significantly different than what most enterprises currently think of. Applications tend to be flatter and more distributed, typically running on bare metal to avoid some of the virtualization overhead that exists in a pure hypervisor environment.

This new breed of scale-out applications would appear to mark the beginning of the application evolution. If this is truly a trend that will only increase, how will it impact the evolution of the other silos?

Software-defined everything

The prevailing chatter across the whole of IT is about software-defined everything. The view is that compute, storage, and networking will all work in cahoots to meet application requirements. The rise of controller-based architectures is prominent in both networking and storage roadmaps, and the compute side of the house embraced central control awhile back.

But what happens if the underlying assumption that applications emerge largely unscathed turns out not to be true?

It could be that the future of datacenter architectures will hinge not on the supporting infrastructure but on the applications themselves. If this is true, we could see a re-emergence of something that we haven’t really talked about in quite awhile: the operating system itself. Sure, there is still talk about Linux and all the server tools that come with it, but the actual operating system hasn’t materially changed in quite some time.

Learning from web-scale applications

If we learn anything from the web-scale companies pushing the boundaries for application performance, it should be that the future is not necessarily about the containers in which applications run. It could be about the underlying OS itself. What if the reason massively scaled companies are embracing bare metal isn’t only about the cost? There is certainly a performance aspect to it as well.

One somewhat uncomfortable conclusion here would be that all the infrastructure work involved in handling application portability across a containerized infrastructure could be somewhat transient. I don’t mean to suggest that it is not useful; there will be a relatively long transition to any kind of new application architecture. And even if there is a transition, the persistence of mainframes should tell us all that no change is absolute or all-encompassing. But a scenario where pockets of new-era applications co-exist in data centers with legacy applications seems likely. We are already seeing this with Hadoop, but I would expect to see more applications built on new architectures.

But this does mean that future-proofing datacenter investments requires a bit more nuance than just buying and planning on scale. Highly-distributed application architectures are even more dependent on east-west traffic. For every 1 byte of traffic going in and out of the datacenter, close to 1 Gigabyte transits the datacenter fabric in some current applications. This ratio likely gets even more aggressive over time.

The bottom line

How all of this plays out is anyone’s guess. We will certainly end up with a hybrid environment supporting all kinds of application architectures making use of various underlying infrastructure architectures. But none of us should be surprised when the industry starts talking a bit more broadly about the role of the operating system going forward. And if you are making plans based on a set of assumptions anchored to current architectures, it might be worth expanding the strategic aperture some to consider how this impacts current plans.

And as a final thought, if the operating system changes, is it still defined by the server, or do we end up with large, distributed operating systems? Put differently, what is the definition of the underlying platform? Are we looking at a new era of platform that includes all of compute, storage, networking, and applications? The implications would be dramatic.

[Today’s fun fact: Potatoes have more chromosomes than humans. I wonder if that means Mr. Potato Head has the combined chromosomes or just a subset of a potato’s.]

The post The re-emergence of the Operating System appeared first on Plexxi.

More Stories By Michael Bushong

The best marketing efforts leverage deep technology understanding with a highly-approachable means of communicating. Plexxi's Vice President of Marketing Michael Bushong has acquired these skills having spent 12 years at Juniper Networks where he led product management, product strategy and product marketing organizations for Juniper's flagship operating system, Junos. Michael spent the last several years at Juniper leading their SDN efforts across both service provider and enterprise markets. Prior to Juniper, Michael spent time at database supplier Sybase, and ASIC design tool companies Synopsis and Magma Design Automation. Michael's undergraduate work at the University of California Berkeley in advanced fluid mechanics and heat transfer lend new meaning to the marketing phrase "This isn't rocket science."

@MicroservicesExpo Stories
When we talk about the impact of BYOD and BYOA and the Internet of Things, we often focus on the impact on data center architectures. That's because there will be an increasing need for authentication, for access control, for security, for application delivery as the number of potential endpoints (clients, devices, things) increases. That means scale in the data center. What we gloss over, what we skip, is that before any of these "things" ever makes a request to access an application it had to...
The best way to leverage your Cloud Expo presence as a sponsor and exhibitor is to plan your news announcements around our events. The press covering Cloud Expo and @ThingsExpo will have access to these releases and will amplify your news announcements. More than two dozen Cloud companies either set deals at our shows or have announced their mergers and acquisitions at Cloud Expo. Product announcements during our show provide your company with the most reach through our targeted audiences.
Virgil consists of an open-source encryption library, which implements Cryptographic Message Syntax (CMS) and Elliptic Curve Integrated Encryption Scheme (ECIES) (including RSA schema), a Key Management API, and a cloud-based Key Management Service (Virgil Keys). The Virgil Keys Service consists of a public key service and a private key escrow service. 

In his general session at 19th Cloud Expo, Manish Dixit, VP of Product and Engineering at Dice, will discuss how Dice leverages data insights and tools to help both tech professionals and recruiters better understand how skills relate to each other and which skills are in high demand using interactive visualizations and salary indicator tools to maximize earning potential. Manish Dixit is VP of Product and Engineering at Dice. As the leader of the Product, Engineering and Data Sciences team a...
Digitization is driving a fundamental change in society that is transforming the way businesses work with their customers, their supply chains and their people. Digital transformation leverages DevOps best practices, such as Agile Parallel Development, Continuous Delivery and Agile Operations to capitalize on opportunities and create competitive differentiation in the application economy. However, information security has been notably absent from the DevOps movement. Speed doesn’t have to negat...
SYS-CON Events announced today that eCube Systems, the leading provider of modern development tools and best practices for Continuous Integration on OpenVMS, will exhibit at SYS-CON's @DevOpsSummit at Cloud Expo New York, which will take place on June 7-9, 2016, at the Javits Center in New York City, NY. eCube Systems offers a family of middleware products and development tools that maximize return on technology investment by leveraging existing technical equity to meet evolving business needs. ...
More and more companies are looking to microservices as an architectural pattern for breaking apart applications into more manageable pieces so that agile teams can deliver new features quicker and more effectively. What this pattern has done more than anything to date is spark organizational transformations, setting the foundation for future application development. In practice, however, there are a number of considerations to make that go beyond simply “build, ship, and run,” which changes ho...
Join Impiger for their featured webinar: ‘Cloud Computing: A Roadmap to Modern Software Delivery’ on November 10, 2016, at 12:00 pm CST. Very few companies have not experienced some impact to their IT delivery due to the evolution of cloud computing. This webinar is not about deciding whether you should entertain moving some or all of your IT to the cloud, but rather, a detailed look under the hood to help IT professionals understand how cloud adoption has evolved and what trends will impact th...
Whether they’re located in a public, private, or hybrid cloud environment, cloud technologies are constantly evolving. While the innovation is exciting, the end mission of delivering business value and rapidly producing incremental product features is paramount. In his session at @DevOpsSummit at 19th Cloud Expo, Kiran Chitturi, CTO Architect at Sungard AS, will discuss DevOps culture, its evolution of frameworks and technologies, and how it is achieving maturity. He will also cover various st...
As we enter the final week before the 19th International Cloud Expo | @ThingsExpo in Santa Clara, CA, it's time for me to reflect on six big topics that will be important during the show. Hybrid Cloud This general-purpose term seems to provide a comfort zone for many enterprise IT managers. It sounds reassuring to be able to work with one of the major public-cloud providers like AWS or Microsoft Azure while still maintaining an on-site presence.
19th Cloud Expo, taking place November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA, will feature technical sessions from a rock star conference faculty and the leading industry players in the world. Cloud computing is now being embraced by a majority of enterprises of all sizes. Yesterday's debate about public vs. private has transformed into the reality of hybrid cloud: a recent survey shows that 74% of enterprises have a hybrid cloud strategy. Meanwhile, 94% of enterpri...
operations aren’t merging to become one discipline. Nor is operations simply going away. Rather, DevOps is leading software development and operations – together with other practices such as security – to collaborate and coexist with less overhead and conflict than in the past. In his session at @DevOpsSummit at 19th Cloud Expo, Gordon Haff, Red Hat Technology Evangelist, will discuss what modern operational practices look like in a world in which applications are more loosely coupled, are deve...
DevOps is speeding towards the IT world like a freight train and the hype around it is deafening. There is no reason to be afraid of this change as it is the natural reaction to the agile movement that revolutionized development just a few years ago. By definition, DevOps is the natural alignment of IT performance to business profitability. The relevance of this has yet to be quantified but it has been suggested that the route to the CEO’s chair will come from the IT leaders that successfully ma...
As the world moves toward more DevOps and Microservices, application deployment to the cloud ought to become a lot simpler. The Microservices architecture, which is the basis of many new age distributed systems such as OpenStack, NetFlix and so on, is at the heart of Cloud Foundry - a complete developer-oriented Platform as a Service (PaaS) that is IaaS agnostic and supports vCloud, OpenStack and AWS. Serverless computing is revolutionizing computing. In his session at 19th Cloud Expo, Raghav...
This is a no-hype, pragmatic post about why I think you should consider architecting your next project the way SOA and/or microservices suggest. No matter if it’s a greenfield approach or if you’re in dire need of refactoring. Please note: considering still keeps open the option of not taking that approach. After reading this, you will have a better idea about whether building multiple small components instead of a single, large component makes sense for your project. This post assumes that you...
DevOps is being widely accepted (if not fully adopted) as essential in enterprise IT. But as Enterprise DevOps gains maturity, expands scope, and increases velocity, the need for data-driven decisions across teams becomes more acute. DevOps teams in any modern business must wrangle the ‘digital exhaust’ from the delivery toolchain, "pervasive" and "cognitive" computing, APIs and services, mobile devices and applications, the Internet of Things, and now even blockchain. In this power panel at @...
As software becomes more and more complex, we, as software developers, have been splitting up our code into smaller and smaller components. This is also true for the environment in which we run our code: going from bare metal, to VMs to the modern-day Cloud Native world of containers, schedulers and microservices. While we have figured out how to run containerized applications in the cloud using schedulers, we've yet to come up with a good solution to bridge the gap between getting your conta...
Enterprise IT has been in the era of Hybrid Cloud for some time now. But it seems most conversations about Hybrid are focused on integrating AWS, Microsoft Azure, or Google ECM into existing on-premises systems. Where is all the Private Cloud? What do technology providers need to do to make their offerings more compelling? How should enterprise IT executives and buyers define their focus, needs, and roadmap, and communicate that clearly to the providers?
Without lifecycle traceability and visibility across the tool chain, stakeholders from Planning-to-Ops have limited insight and answers to who, what, when, why and how across the DevOps lifecycle. This impacts the ability to deliver high quality software at the needed velocity to drive positive business outcomes. In his general session at @DevOpsSummit at 19th Cloud Expo, Eric Robertson, General Manager at CollabNet, will discuss how customers are able to achieve a level of transparency that e...
Today every business relies on software to drive the innovation necessary for a competitive edge in the Application Economy. This is why collaboration between development and operations, or DevOps, has become IT’s number one priority. Whether you are in Dev or Ops, understanding how to implement a DevOps strategy can deliver faster development cycles, improved software quality, reduced deployment times and overall better experiences for your customers.