Microservices Expo Authors: Yeshim Deniz, Pat Romanski, Liz McMillan, Kareen Kircher, Andreas Grabner

Related Topics: SDN Journal, Java IoT, Microservices Expo, Containers Expo Blog, @CloudExpo, @BigDataExpo

SDN Journal: Article

Using OpenFlow to Extend Software Defined Networking

OpenFlow’s Role in One-Stop Shopping for Control of Enterprise Cloud, Storage and Network Resources

Using OpenFlow to extend software-defined networking (SDN) to the optical layer is a compelling prospect for enterprises seeking to achieve joint orchestration of information technology (IT) and network resources for cloud services, to virtualize the network and to more simply manage interconnections of distributed data centers that require synchronization.

Today's fragmented, specialized management and control approaches are fraught with proprietary protocols and management systems, limited scalability and configuration complexities. With an OpenFlow-enabled transport network, an enterprise could instead engage in a kind of "one-stop shopping" for control of cloud computing, storage and networking resources - all via one, unified application programming interface (API). The benefits could include significantly simplified configuration, management and scaling of large-scale enterprise infrastructures through integration and automation.

That's a new role for OpenFlow, demanding strategic tailoring of the protocol for the optical transport domain. Demonstration and development of the capability are closely watched by enterprises that are under incessant pressure to cost-effectively meet ever-increasing demand for bandwidth and services.

Virtualization's New Frontier
Servers and storage have been virtualized in the enterprise; the next great frontier for virtualization is the network.

Because of the substantial cost savings and performance benefits that it can deliver, SDN-based virtualization is of prime interest to enterprises for a wide range of applications. OpenFlow has emerged as one of the most popular SDN protocols. Web 2.0 network operators and national research and education network (NREN) operators, especially, like OpenFlow.

With OpenFlow, an abstraction of the network's packet switches can be generated and flow-forwarding behavior can be specified across an infrastructure via an external controller. Operations can be substantially automated and streamlined by breaking up the monolithically integrated control and forwarding paradigm of today's switches.

Using OpenFlow, could SDN be extended across layers and create a scenario in which - with a single instruction - the controller could jointly create virtual machines and enable enterprise network administrators to reserve computing, networking and storage resources in one stroke?

It is an obviously compelling notion for enterprise network staffs who desperately need to simplify operations. However, the problem is OpenFlow deployment and development has largely been limited to the electrical packet layer, whereas the interconnection beyond the data center is typically comprised of optical transport technology. Furthermore, the optical domain is where things get hazy for many enterprise network administrators. Their comfort zone tends to be packets - not wavelengths and optics.

The result is that cloud computing is currently decoupled from the transport networking control and operation. The network exists as a static, separated entity in today's cloud implementations. There is no interaction between cloud computing processes and the statically configured network. The two are not in any way interoperable; they speak different languages.

Converging cloud computing and networking requires a more dynamic mode of control and operation, but enterprises largely have judged integrating management of the optical network into the data-center environment to be too complex.

To extend OpenFlow from its established role in the electrical packet domain to the optical layer (and, thereby, extend SDN across multiple network layers), a range of optical-specific concerns must be tackled.

Crafting and Experimenting
Within the European Commission's FP7 ICT Work Programme is a collaborative project, "OpenFlow in Europe - Linking Infrastructure and Applications" (OFELIA), that provides researchers with a test bed in which to experiment with SDN applications and virtual multi-layer networks over shared network infrastructure.

Via standardized, secure interfaces through GÉANT, a high-bandwidth interconnection of European R&E networks, researchers develop, run and control experiments using packet switches and application servers at the University of Essex and seven other test-bed facilities throughout Europe.

OFELIA hosts a prototype implementation of dynamic control of wavelength-switched optical networks via OpenFlow. Bandwidth, latency and power consumption can be adjusted to meet the specific requirements of specific applications.

To make it happen, key OpenFlow additions had to be engineered in order for the protocol to effectively control the optical domain. Optical-specific considerations were required to adapt OpenFlow from the packet world. A packet can travel from any ingress to any egress port in an electrical switch or from any time slot in a time-division multiplexing (TDM) device. The optical domain, however, introduces strict switching constraints, with regard to wavelength continuity, optical impairments, optical power leveling on the line side, etc.

Augmenting OpenFlow to address those optical-specific concerns has resulted in an OFELIA prototype that demonstrates a truly transparent, wavelength-switched optical network. The research community is able to experiment with the capability via a flexible, Web-services approach; commercial enterprises, too, are interested in trialing the capability for their specific applications and environments.

OpenFlow is not sufficient in itself to enable the complete transformation that enterprise network administrators envision, to SDN-enable virtualization across all layers of their infrastructures. The additions to OpenFlow that were engineered for the OFELIA test bed provide only the bridge between the optical layer and packet layer and allow integration into a cloud operating system such as OpenStack.

But that is one very important bridge, and the promise for enterprise network administrators is considerable. The OpenFlow innovation could seamlessly integrate the optical transport network under a common management umbrella with an enterprise's routers and switches - all via one familiar interface. Management of the optical domain could become as simple as the management of Ethernet boxes - using an encapsulation of virtual resources that enterprise network administrators could manage via typical and familiar infrastructure. That's a significant breakthrough. With many enterprises already considering usage of an OpenFlow-based control for their packet networks, extending the framework to the wavelength-switched optical layer would be a natural migration.

Virtualization has developed over phases in enterprise networking. First, resource virtualization inside data centers delivered economic savings through enhanced utilization, scalability and redundancy. Data-center virtualization conveyed greater infrastructure flexibility, higher availability and better workload balancing. The next frontier, network virtualization, promises true platform agility and, with it, a host of long-sought-after enterprise capabilities: capacity on-demand, adaptive infrastructure and dynamic service automation, among them. Adapting OpenFlow and extending SDN to the optical transport domain comprise an important step toward that vision.

More Stories By Jörg-Peter Elbers

Jörg-Peter Elbers is VP Advanced Technology in the CTO office at ADVA Optical Networking in Munich, Germany, and is globally responsible for technology strategy, new product concepts, standardization, and research activities.

Prior to joining ADVA in 2007, he was Director of Technology in the Optical Product Unit of Marconi (now Ericsson). From 1999 to 2001, he worked at Siemens AG, last as Director of Network Architecture in Siemens Optical Networks.

More Stories By Achim Autenrieth

Achim Autenrieth is Principle Research Engineer Advanced Technology in the CTO Office at ADVA Optical Networking, where he is working on the design and evaluation of multilayer networks, control plane, and SDN concepts. Achim is a member of IEEE and VDE/ITG, he authored or co-authored more than 70 reviewed and invited scientific publications and he is technical program committee member of ECOC, DRCN, and RNDM.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.

@MicroservicesExpo Stories
SYS-CON Events announced today that Tintri Inc., a leading producer of VM-aware storage (VAS) for virtualization and cloud environments, will present at the 19th International Cloud Expo, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. Tintri VM-aware storage is the simplest for virtualized applications and cloud. Organizations including GE, Toyota, United Healthcare, NASA and 6 of the Fortune 15 have said “No to LUNs.” With Tintri they manag...
SYS-CON Events announced today that eCube Systems, the leading provider of modern development tools and best practices for Continuous Integration on OpenVMS, will exhibit at SYS-CON's @DevOpsSummit at Cloud Expo New York, which will take place on June 7-9, 2016, at the Javits Center in New York City, NY. eCube Systems offers a family of middleware products and development tools that maximize return on technology investment by leveraging existing technical equity to meet evolving business needs. ...
All clouds are not equal. To succeed in a DevOps context, organizations should plan to develop/deploy apps across a choice of on-premise and public clouds simultaneously depending on the business needs. This is where the concept of the Lean Cloud comes in - resting on the idea that you often need to relocate your app modules over their life cycles for both innovation and operational efficiency in the cloud. In his session at @DevOpsSummit at19th Cloud Expo, Valentin (Val) Bercovici, CTO of So...
Apache Hadoop is a key technology for gaining business insights from your Big Data, but the penetration into enterprises is shockingly low. In fact, Apache Hadoop and Big Data proponents recognize that this technology has not yet achieved its game-changing business potential. In his session at 19th Cloud Expo, John Mertic, director of program management for ODPi at The Linux Foundation, will explain why this is, how we can work together as an open data community to increase adoption, and the i...
operations aren’t merging to become one discipline. Nor is operations simply going away. Rather, DevOps is leading software development and operations – together with other practices such as security – to collaborate and coexist with less overhead and conflict than in the past. In his session at @DevOpsSummit at 19th Cloud Expo, Gordon Haff, Red Hat Technology Evangelist, will discuss what modern operational practices look like in a world in which applications are more loosely coupled, are deve...
DevOps is a term that comes full of controversy. A lot of people are on the bandwagon, while others are waiting for the term to jump the shark, and eventually go back to business as usual. Regardless of where you are along the specturm of loving or hating the term DevOps, one thing is certain. More and more people are using it to describe a system administrator who uses scripts, or tools like, Chef, Puppet or Ansible, in order to provision infrastructure. There is also usually an expectation of...
DevOps is speeding towards the IT world like a freight train and the hype around it is deafening. There is no reason to be afraid of this change as it is the natural reaction to the agile movement that revolutionized development just a few years ago. By definition, DevOps is the natural alignment of IT performance to business profitability. The relevance of this has yet to be quantified but it has been suggested that the route to the CEO’s chair will come from the IT leaders that successfully ma...
As we enter the final week before the 19th International Cloud Expo | @ThingsExpo in Santa Clara, CA, it's time for me to reflect on six big topics that will be important during the show. Hybrid Cloud This general-purpose term seems to provide a comfort zone for many enterprise IT managers. It sounds reassuring to be able to work with one of the major public-cloud providers like AWS or Microsoft Azure while still maintaining an on-site presence.
The best way to leverage your Cloud Expo presence as a sponsor and exhibitor is to plan your news announcements around our events. The press covering Cloud Expo and @ThingsExpo will have access to these releases and will amplify your news announcements. More than two dozen Cloud companies either set deals at our shows or have announced their mergers and acquisitions at Cloud Expo. Product announcements during our show provide your company with the most reach through our targeted audiences.
This is a no-hype, pragmatic post about why I think you should consider architecting your next project the way SOA and/or microservices suggest. No matter if it’s a greenfield approach or if you’re in dire need of refactoring. Please note: considering still keeps open the option of not taking that approach. After reading this, you will have a better idea about whether building multiple small components instead of a single, large component makes sense for your project. This post assumes that you...
As software becomes more and more complex, we, as software developers, have been splitting up our code into smaller and smaller components. This is also true for the environment in which we run our code: going from bare metal, to VMs to the modern-day Cloud Native world of containers, schedulers and microservices. While we have figured out how to run containerized applications in the cloud using schedulers, we've yet to come up with a good solution to bridge the gap between getting your conta...
DevOps is being widely accepted (if not fully adopted) as essential in enterprise IT. But as Enterprise DevOps gains maturity, expands scope, and increases velocity, the need for data-driven decisions across teams becomes more acute. DevOps teams in any modern business must wrangle the ‘digital exhaust’ from the delivery toolchain, "pervasive" and "cognitive" computing, APIs and services, mobile devices and applications, the Internet of Things, and now even blockchain. In this power panel at @...
DevOps theory promotes a culture of continuous improvement built on collaboration, empowerment, systems thinking, and feedback loops. But how do you collaborate effectively across the traditional silos? How can you make decisions without system-wide visibility? How can you see the whole system when it is spread across teams and locations? How do you close feedback loops across teams and activities delivering complex multi-tier, cloud, container, serverless, and/or API-based services?
SYS-CON Events announced today that SoftNet Solutions will exhibit at the 19th International Cloud Expo, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. SoftNet Solutions specializes in Enterprise Solutions for Hadoop and Big Data. It offers customers the most open, robust, and value-conscious portfolio of solutions, services, and tools for the shortest route to success with Big Data. The unique differentiator is the ability to architect and ...
Without lifecycle traceability and visibility across the tool chain, stakeholders from Planning-to-Ops have limited insight and answers to who, what, when, why and how across the DevOps lifecycle. This impacts the ability to deliver high quality software at the needed velocity to drive positive business outcomes. In his session at @DevOpsSummit 19th Cloud Expo, Eric Robertson, General Manager at CollabNet, will show how customers are able to achieve a level of transparency that enables everyon...
Today every business relies on software to drive the innovation necessary for a competitive edge in the Application Economy. This is why collaboration between development and operations, or DevOps, has become IT’s number one priority. Whether you are in Dev or Ops, understanding how to implement a DevOps strategy can deliver faster development cycles, improved software quality, reduced deployment times and overall better experiences for your customers.
At its core DevOps is all about collaboration. The lines of communication must be opened and it takes some effort to ensure that they stay that way. It’s easy to pay lip service to trends and talk about implementing new methodologies, but without action, real benefits cannot be realized. Success requires planning, advocates empowered to effect change, and, of course, the right tooling. To bring about a cultural shift it’s important to share challenges. In simple terms, ensuring that everyone k...
What do dependency resolution, situational awareness, and superheroes have in common? Meet Chris Corriere, a DevOps/Software Engineer at Autotrader, speaking on creative ways to maximize usage of all of the above. Mark Miller, Community Advocate and senior storyteller at Sonatype, caught up with Chris to learn more about what his team is up to.
SYS-CON Events announced today that Super Micro Computer, Inc., a global leader in Embedded and IoT solutions, will exhibit at SYS-CON's 20th International Cloud Expo®, which will take place on June 7-9, 2017, at the Javits Center in New York City, NY. Supermicro (NASDAQ: SMCI), the leading innovator in high-performance, high-efficiency server technology, is a premier provider of advanced server Building Block Solutions® for Data Center, Cloud Computing, Enterprise IT, Hadoop/Big Data, HPC and ...
@DevOpsSummit has been named the ‘Top DevOps Influencer' by iTrend. iTrend processes millions of conversations, tweets, interactions, news articles, press releases, blog posts - and extract meaning form them and analyzes mobile and desktop software platforms used to communicate, various metadata (such as geo location), and automation tools. In overall placement, @DevOpsSummit ranked as the number one ‘DevOps Influencer' followed by @CloudExpo at third, and @MicroservicesE at 24th.