|By Marten Terpstra||
|April 10, 2014 01:00 PM EDT||
We love analogies. No matter what the topic, analogies are a great way to explain something in a different context to make a specific point with a frame of reference that may be more familiar to those we are making a point to. There is one that seems to come back over and over again in our industry, the one that compares the network to the power grid, network connections to power plugs. I had not heard it for a while but at Interop last week, I heard it used twice in booth demonstrations as part of plug and play pitches. And I really do not like that analogy.
The comparison to power comes from the angle of it just being there and available. Plug something in and it turns on. Its universal and it’s simple, anyone can use it. It’s highly standardized, you can buy bits and pieces that work together in any hardware store. It’s reliable. You just get a bill each month, pay it, and in many places you can even pick your provider different from the owner of the cable that actually plugs into your house. It’s all good, it’s mostly invisible.
Except that all of these come with their own challenges. The overall power grid reliability is well below 99.99%. Across the US, the average outage minutes per year is hovering in the 2 hour range, in the Northeast almost double that. There are only very people I know that do not have a backup generator, just in case, and get to use it at least once a year, ignoring the smaller outages where we just live in the dark for a few minutes. It is very standardized, with plenty of standard variations from country to country, thank goodness for $30 converters. And the cost, well, it’s rising faster than inflation (almost twice as fast in fact).
Now, this article is not meant to be a complaining session about power. When power works and you forget about the bill, it is very much invisible, which is the only piece of the analogy I do like, maybe.
Networks should be invisible, but not in the same sense power should be invisible. When comparing networks to power it always comes down to “it’s just there, plentiful, accessible, and just works”. And (not surprising), I think the network can do much better than “just be there”. Networks need to migrate away from an entity that is architected, designed, implemented, managed and debugged separately from the applications it serves. Networks have to become an integral part of the application infrastructure.
Now, any two (or more) distinctly different entities need to be glued together. No different for applications, servers and networks. Applications are glued to servers by means of the OS. As a consumer of an application, or even an administrator, the amount of work you have to do to put these two together is (usually) fairy minimal. The creator of the OS has provided tons of hooks and tools to give the application what it needs, the application provides to and asks what it wants from the OS.
The glue between the server and the network is much harsher, much less refined. There is an expectation of matching configuration, very little is actually transacted between the two to get them to work together nicely. Yes there is DHCP and perhaps you can even count LLDP in that, but the amount of alignment of information between the two so they can work together seamlessly without human intervention is minimal. And if something is not quite working the way it should, you need heavily certified network engineers and tons of homegrown tools and scripts to figure out why. That is by no means a knock on network certifications (that is for @cloudtoad to comment on), but it should not be that hard. The exchange is extremely focused on data plane traffic for transportation, very little provisioning, debug and other feedback is exchanged and reacted to.
We need to get to a network that is invisible enough to allow you to create coarse provisioning policies created from application templates. There should be no need for me to configure individual ports and protocols on ports. There should be no need for me to configure how network components are tied together and provide optimized connectivity based on the needs of its customers. The network needs to provide debugging tools that interact with applications in feedback loops. It needs to provide enough details for orchestration systems to adjust their placement of apps. Or in Plexxi’s case, take in information from orchestration systems to change the L1, L2 and L3 behavior of the network to the requested demand. And that orchestration system could be a specialized Hadoop controller or a far more generic cloud orchestrator.
The network needs to cease to exist as a separate entity, it needs to become an integrated part of the application infrastructure that uses its services. That’s the kind of invisible we need to achieve.
[Today's Fun Fact: An Iranian man paid $500 to a wizard to become invisible. The man then walked into a bank and grabbed money from people's hands believing he was invisible. He was promptly arrested and convicted. Conclusion: pay more or pick a better wizard next time.]
The IT industry is undergoing a significant evolution to keep up with cloud application demand. We see this happening as a mindset shift, from traditional IT teams to more well-rounded, cloud-focused job roles. The IT industry has become so cloud-minded that Gartner predicts that by 2020, this cloud shift will impact more than $1 trillion of global IT spending. This shift, however, has left some IT professionals feeling a little anxious about what lies ahead. The good news is that cloud computin...
Mar. 26, 2017 10:30 AM EDT Reads: 1,217
SYS-CON Events announced today that HTBase will exhibit at SYS-CON's 20th International Cloud Expo®, which will take place on June 6-8, 2017, at the Javits Center in New York City, NY. HTBase (Gartner 2016 Cool Vendor) delivers a Composable IT infrastructure solution architected for agility and increased efficiency. It turns compute, storage, and fabric into fluid pools of resources that are easily composed and re-composed to meet each application’s needs. With HTBase, companies can quickly prov...
Mar. 26, 2017 08:15 AM EDT Reads: 2,874
An overall theme of Cloud computing and the specific practices within it is fundamentally one of automation. The core value of technology is to continually automate low level procedures to free up people to work on more value add activities, ultimately leading to the utopian goal of full Autonomic Computing. For example a great way to define your plan for DevOps tool chain adoption is through this lens. In this TechTarget article they outline a simple maturity model for planning this.
Mar. 26, 2017 06:00 AM EDT Reads: 4,261
While DevOps most critically and famously fosters collaboration, communication, and integration through cultural change, culture is more of an output than an input. In order to actively drive cultural evolution, organizations must make substantial organizational and process changes, and adopt new technologies, to encourage a DevOps culture. Moderated by Andi Mann, panelists discussed how to balance these three pillars of DevOps, where to focus attention (and resources), where organizations might...
Mar. 26, 2017 05:15 AM EDT Reads: 6,158
The rise of containers and microservices has skyrocketed the rate at which new applications are moved into production environments today. While developers have been deploying containers to speed up the development processes for some time, there still remain challenges with running microservices efficiently. Most existing IT monitoring tools don’t actually maintain visibility into the containers that make up microservices. As those container applications move into production, some IT operations t...
Mar. 26, 2017 01:00 AM EDT Reads: 2,945
DevOps is often described as a combination of technology and culture. Without both, DevOps isn't complete. However, applying the culture to outdated technology is a recipe for disaster; as response times grow and connections between teams are delayed by technology, the culture will die. A Nutanix Enterprise Cloud has many benefits that provide the needed base for a true DevOps paradigm.
Mar. 26, 2017 12:30 AM EDT Reads: 1,933
For organizations that have amassed large sums of software complexity, taking a microservices approach is the first step toward DevOps and continuous improvement / development. Integrating system-level analysis with microservices makes it easier to change and add functionality to applications at any time without the increase of risk. Before you start big transformation projects or a cloud migration, make sure these changes won’t take down your entire organization.
Mar. 25, 2017 09:45 PM EDT Reads: 3,597
DevOps is often described as a combination of technology and culture. Without both, DevOps isn't complete. However, applying the culture to outdated technology is a recipe for disaster; as response times grow and connections between teams are delayed by technology, the culture will die. A Nutanix Enterprise Cloud has many benefits that provide the needed base for a true DevOps paradigm. In his Day 3 Keynote at 20th Cloud Expo, Chris Brown, a Solutions Marketing Manager at Nutanix, will explore t...
Mar. 25, 2017 03:00 PM EDT Reads: 2,777
Software development is a moving target. You have to keep your eye on trends in the tech space that haven’t even happened yet just to stay current. Consider what’s happened with augmented reality (AR) in this year alone. If you said you were working on an AR app in 2015, you might have gotten a lot of blank stares or jokes about Google Glass. Then Pokémon GO happened. Like AR, the trends listed below have been building steam for some time, but they’ll be taking off in surprising new directions b...
Mar. 25, 2017 01:30 PM EDT Reads: 5,816
Everyone wants to use containers, but monitoring containers is hard. New ephemeral architecture introduces new challenges in how monitoring tools need to monitor and visualize containers, so your team can make sense of everything. In his session at @DevOpsSummit, David Gildeh, co-founder and CEO of Outlyer, will go through the challenges and show there is light at the end of the tunnel if you use the right tools and understand what you need to be monitoring to successfully use containers in your...
Mar. 25, 2017 12:45 PM EDT Reads: 1,523
What if you could build a web application that could support true web-scale traffic without having to ever provision or manage a single server? Sounds magical, and it is! In his session at 20th Cloud Expo, Chris Munns, Senior Developer Advocate for Serverless Applications at Amazon Web Services, will show how to build a serverless website that scales automatically using services like AWS Lambda, Amazon API Gateway, and Amazon S3. We will review several frameworks that can help you build serverle...
Mar. 25, 2017 12:30 PM EDT Reads: 1,873
@DevOpsSummit has been named the ‘Top DevOps Influencer' by iTrend. iTrend processes millions of conversations, tweets, interactions, news articles, press releases, blog posts - and extract meaning form them and analyzes mobile and desktop software platforms used to communicate, various metadata (such as geo location), and automation tools. In overall placement, @DevOpsSummit ranked as the number one ‘DevOps Influencer' followed by @CloudExpo at third, and @MicroservicesE at 24th.
Mar. 25, 2017 10:30 AM EDT Reads: 10,318
Culture is the most important ingredient of DevOps. The challenge for most organizations is defining and communicating a vision of beneficial DevOps culture for their organizations, and then facilitating the changes needed to achieve that. Often this comes down to an ability to provide true leadership. As a CIO, are your direct reports IT managers or are they IT leaders? The hard truth is that many IT managers have risen through the ranks based on their technical skills, not their leadership abi...
Mar. 25, 2017 05:00 AM EDT Reads: 10,985
The essence of cloud computing is that all consumable IT resources are delivered as services. In his session at 15th Cloud Expo, Yung Chou, Technology Evangelist at Microsoft, demonstrated the concepts and implementations of two important cloud computing deliveries: Infrastructure as a Service (IaaS) and Platform as a Service (PaaS). He discussed from business and technical viewpoints what exactly they are, why we care, how they are different and in what ways, and the strategies for IT to transi...
Mar. 25, 2017 05:00 AM EDT Reads: 6,079
After more than five years of DevOps, definitions are evolving, boundaries are expanding, ‘unicorns’ are no longer rare, enterprises are on board, and pundits are moving on. Can we now look at an evolution of DevOps? Should we? Is the foundation of DevOps ‘done’, or is there still too much left to do? What is mature, and what is still missing? What does the next 5 years of DevOps look like? In this Power Panel at DevOps Summit, moderated by DevOps Summit Conference Chair Andi Mann, panelists l...
Mar. 25, 2017 05:00 AM EDT Reads: 9,765
Thanks to Docker and the DevOps revolution, microservices have emerged as the new way to build and deploy applications — and there are plenty of great reasons to embrace the microservices trend. If you are going to adopt microservices, you also have to understand that microservice architectures have many moving parts. When it comes to incident management, this presents an important difference between microservices and monolithic architectures. More moving parts mean more complexity to monitor an...
Mar. 25, 2017 04:15 AM EDT Reads: 3,389
All organizations that did not originate this moment have a pre-existing culture as well as legacy technology and processes that can be more or less amenable to DevOps implementation. That organizational culture is influenced by the personalities and management styles of Executive Management, the wider culture in which the organization is situated, and the personalities of key team members at all levels of the organization. This culture and entrenched interests usually throw a wrench in the work...
Mar. 25, 2017 02:00 AM EDT Reads: 2,944
Microservices (μServices) are a fascinating evolution of the Distributed Object Computing (DOC) paradigm. Initial design of DOC attempted to solve the problem of simplifying developing complex distributed applications by applying object-oriented design principles to disparate components operating across networked infrastructure. In this model, DOC “hid” the complexity of making this work from the developer regardless of the deployment architecture through the use of complex frameworks, such as C...
Mar. 25, 2017 12:15 AM EDT Reads: 4,201
TechTarget storage websites are the best online information resource for news, tips and expert advice for the storage, backup and disaster recovery markets. By creating abundant, high-quality editorial content across more than 140 highly targeted technology-specific websites, TechTarget attracts and nurtures communities of technology buyers researching their companies' information technology needs. By understanding these buyers' content consumption behaviors, TechTarget creates the purchase inte...
Mar. 24, 2017 10:15 PM EDT Reads: 4,259
We've all had that feeling before: The feeling that you're missing something that everyone else is in on. For today's IT leaders, that feeling might come up when you hear talk about cloud brokers. Meanwhile, you head back into your office and deal with your ever-growing shadow IT problem. But the cloud-broker whispers and your shadow IT issues are linked. If you're wondering "what the heck is a cloud broker?" we've got you covered.
Mar. 24, 2017 02:45 PM EDT Reads: 580