|By Jason Bloomberg||
|May 22, 2014 02:13 PM EDT||
A question we commonly get at EnterpriseWeb is whether our platform follows REST or not. Representational State Transfer (REST) is an architectural style for distributed hypermedia systems such as the World Wide Web, and is perhaps best known for providing a lightweight, uniform Web-style application programming interface (API) to server-based resources. On the one hand, EnterpriseWeb can both consume and expose any type of interface, including tightly coupled APIs, Web Services, as well as RESTful APIs, and the platform has no requirement that customers must build distributed hypermedia systems. It would be easy to conclude, therefore, that while EnterpriseWeb supports REST, it is not truly RESTful.
Such a conclusion, however, would neglect the broader architectural context for EnterpriseWeb. The platform builds on top of and extends REST as the foundation for the dynamic, enterprise-class architectural style we call Agent-Oriented Architecture (AOA). EnterpriseWeb’s intelligent agent, SmartAlex, leverages RESTful constraints as part of the core functionality of the EnterpriseWeb platform. The resulting AOA pattern essentially reinvents application functionality and enterprise integration, heralding a new paradigm for distributed computing.
The Limitations of REST
One of the primary challenges to the successful application of REST is understanding how to extend REST to distributed hypermedia systems in general, beyond the straightforward interactions between browsers and Web servers. To help clarify this point, Figure 1 below illustrates a simple RESTful architecture. In this example, the client is a browser, and it sends GETs and PUTs or other RESTful queries to URIs that resolve to resources on a server, which responds by sending the appropriate representation back to the client. In addition, REST allows for a cache intermediating between client and server that might resolve queries on behalf of the server for scalability purposes.
As an architectural style, however, the point of REST isn’t the uniform interface that the HTTP verbs enable. REST is really about hypermedia, where hypermedia are the engine of application state – the HATEOAS constraint essential to building hypermedia systems. In figure 1, we’re representing HATEOAS by the interactions between human users and their browsers as people click links on Web pages, thus advancing the application state. The RESTful client (in other words, the browser) maintains application state for each user by showing them the Web page (or other representation) they requested when they followed a given hyperlink.
However, software clients that do not necessarily have user interfaces may be problematic for REST, but they are a familiar part of the Service-Oriented Architecture (SOA) architectural style, where we call such clients Service consumers. Combining REST and SOA into the combined architectural style we call REST-Based SOA introduces the notion of an intermediary that presents a Service endpoint and resolves interactions with that endpoint into underlying interactions with various legacy systems. The SOA intermediary in this case exposes RESTful endpoints as URIs that accept GETs, PUTs, etc. from Service consumers, which can be any software client. See figure 2 below for an illustration of the REST-Based SOA pattern.
Note that adding SOA to REST augments the role of the intermediary. Pure REST allows for simple caching and proxy behavior, while SOA calls for policy-based routing and transformation operations that provide the Service abstraction. SOA also reinforces the notion that the Service consumer can be any piece of software, regardless of whether it has a user interface.
Even with REST-based SOA, however, we still have problems with implementing HATEOAS: coding our clients so that they are able to gather the metadata they need by following hyperlinks. In other words, how do we apply REST to any hypermedia system, where instead of a browser we have any piece of software as a client? How do we code the software client to know how to follow hyperlinks, where it doesn’t know what the hyperlinks are ahead of time or what representations they’re supposed to interact with? Humans simply click hyperlinks until they get the representation they want, even if they don’t know beforehand how to find it. How do we teach software to automate this process and gather all the metadata it needs by following a sequence of hyperlinks?
Introducing Agent-Oriented Architecture
The answer to these questions is to cast an intelligent agent in the role of SOA intermediary in the REST-Based SOA pattern in Figure 2. Intelligent software agents (or simply intelligent agents when we know we’re talking about software) are autonomous programs that have the authority to determine what action is appropriate based upon the requests made of them. In this new, Agent-Oriented architectural pattern, the agent interacts with any resource as a RESTful client, where the agent must be able to automatically follow hyperlinks to gather all the information it requires in order to respond appropriately to any request from the client.
In other words, when following this newly coined AOA architectural style, software clients do not have to comply with HATEOAS (they may, but such compliance is optional). Instead, the agent alone must follow the HATEOAS constraint as it interacts with resources. To achieve this behavior, we must underspecify the intelligent agent. In other words, the agent can’t know ahead of time what it’s supposed to do to respond to any particular request. Instead, it must be able to process any request on demand by fetching related resources that provide the appropriate metadata, data, or code it needs to properly respond to that request with a custom response, for each interaction in real time. Figure 3 below illustrates the basic AOA pattern.
For each request from any client, regardless of whether it has a user interface, the agent constructs a custom response based on latest and most relevant information available. In fact, requests to the agent can come from anywhere (i.e., they follow an event-driven pattern). The agent’s underspecification means that it doesn’t know ahead of time what behavior it must exhibit, but it does know how to find the information it needs in order to determine that behavior – and it does that by following hyperlinks, as per HATEOAS. In other words, the goal-oriented agent resolves URIs recursively in order to gather and execute the information it needs – a particularly concise example of fully automated HATEOAS in action.
The Benefits of AOA
An earlier Loosely-Coupled newsletter explained that if you follow REST, you’re unable to accept out-of-band metadata or business context outside of the hypermedia. Agent-Oriented Architecture, however, solves these problems, because the agent is free to fetch whatever it needs to complete the request, since it treats all entities – metadata, data, code, etc. – as resources. In other words, the agent serves as a RESTful client, even when the software client does not. What was out-of-band for REST isn’t out-of-band for AOA. Everything is on the table.
The true power of AOA, though, lies in how it resolves the fundamental challenge of static APIs. Whether they be Web Services, RESTful APIs, or some other type of loosely-coupled interface, every approach to software integration today suffers from the fact that interactions tend to break when API contract metadata change.
By adding an intelligent agent to the mix, we’re able to resolve differences in interaction context between disparate software endpoints dynamically and in real time. Far more than a traditional broker, which must rely on static transformation logic to resolve endpoint differences, the agent must be able to interpret metadata, as well as policies, rules, and the underlying data themselves to create real time interactions that maintain the business context – an example of dynamic coupling, a central principle to AOA.
Dynamic coupling, therefore, represents a paradigm shift in how to build and utilize APIs. Up to this point in time, the focus of both SOA and REST has been on building loosely-coupled interfaces: static, contracted interfaces specified by WSDL and various policy metadata when those interfaces are Web Services, or Internet Media Types and related metadata for RESTful interactions. Neither approach deals well with change. AOA, in contrast, relies upon dynamic coupling that responds automatically to change, since the agent interprets current metadata for every interaction in real time.
Icons by http://dryicons.com
SYS-CON Events announced today that Commvault, a global leader in enterprise data protection and information management, has been named “Bronze Sponsor” of SYS-CON's 18th International Cloud Expo, which will take place on June 7–9, 2016, at the Javits Center in New York City, NY, and the 19th International Cloud Expo, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. Commvault is a leading provider of data protection and information management...
Feb. 8, 2016 10:45 AM EST Reads: 377
SYS-CON Events announced today that AppNeta, the leader in performance insight for business-critical web applications, will exhibit and present at SYS-CON's @DevOpsSummit at Cloud Expo New York, which will take place on June 7-9, 2016, at the Javits Center in New York City, NY. AppNeta is the only application performance monitoring (APM) company to provide solutions for all applications – applications you develop internally, business-critical SaaS applications you use and the networks that deli...
Feb. 8, 2016 10:15 AM EST Reads: 354
The (re?)emergence of Microservices was especially prominent in this week’s news. What are they good for? do they make sense for your application? should you take the plunge? and what do Microservices mean for your DevOps and Continuous Delivery efforts? Continue reading for more on Microservices, containers, DevOps culture, and more top news from the past week. As always, stay tuned to all the news coming from@ElectricCloud on DevOps and Continuous Delivery throughout the week and retweet/favo...
Feb. 8, 2016 09:45 AM EST Reads: 174
If we look at slow, traditional IT and jump to the conclusion that just because we found its issues intractable before, that necessarily means we will again, then it’s time for a rethink. As a matter of fact, the world of IT has changed over the last ten years or so. We’ve been experiencing unprecedented innovation across the board – innovation in technology as well as in how people organize and accomplish tasks. Let’s take a look at three differences between today’s modern, digital context...
Feb. 8, 2016 08:15 AM EST Reads: 167
The battle over bimodal IT is heating up. Now that there’s a reasonably broad consensus that Gartner’s advice about bimodal IT is deeply flawed – consensus everywhere except perhaps at Gartner – various ideas are springing up to fill the void. The bimodal problem, of course, is well understood. ‘Traditional’ or ‘slow’ IT uses hidebound, laborious processes that would only get in the way of ‘fast’ or ‘agile’ digital efforts. The result: incoherent IT strategies and shadow IT struggles that lead ...
Feb. 7, 2016 05:00 PM EST Reads: 434
SYS-CON Events announced today that VAI, a leading ERP software provider, will exhibit at SYS-CON's 18th International Cloud Expo®, which will take place on June 7-9, 2016, at the Javits Center in New York City, NY. VAI (Vormittag Associates, Inc.) is a leading independent mid-market ERP software developer renowned for its flexible solutions and ability to automate critical business functions for the distribution, manufacturing, specialty retail and service sectors. An IBM Premier Business Part...
Feb. 7, 2016 02:00 PM EST Reads: 564
SYS-CON Events announced today that Alert Logic, Inc., the leading provider of Security-as-a-Service solutions for the cloud, will exhibit at SYS-CON's 18th International Cloud Expo®, which will take place on June 7-9, 2016, at the Javits Center in New York City, NY. Alert Logic, Inc., provides Security-as-a-Service for on-premises, cloud, and hybrid infrastructures, delivering deep security insight and continuous protection for customers at a lower cost than traditional security solutions. Ful...
Feb. 7, 2016 01:45 PM EST Reads: 368
In most cases, it is convenient to have some human interaction with a web (micro-)service, no matter how small it is. A traditional approach would be to create an HTTP interface, where user requests will be dispatched and HTML/CSS pages must be served. This approach is indeed very traditional for a web site, but not really convenient for a web service, which is not intended to be good looking, 24x7 up and running and UX-optimized. Instead, talking to a web service in a chat-bot mode would be muc...
Feb. 7, 2016 01:15 PM EST Reads: 194
SYS-CON Events announced today that Catchpoint Systems, Inc., a provider of innovative web and infrastructure monitoring solutions, has been named “Silver Sponsor” of SYS-CON's DevOps Summit at 18th Cloud Expo New York, which will take place June 7-9, 2016, at the Javits Center in New York City, NY. Catchpoint is a leading Digital Performance Analytics company that provides unparalleled insight into customer-critical services to help consistently deliver an amazing customer experience. Designed...
Feb. 7, 2016 01:00 PM EST Reads: 335
SYS-CON Events announced today that Interoute, owner-operator of one of Europe's largest networks and a global cloud services platform, has been named “Bronze Sponsor” of SYS-CON's 18th Cloud Expo, which will take place on June 7-9, 2015 at the Javits Center in New York, New York. Interoute is the owner-operator of one of Europe's largest networks and a global cloud services platform which encompasses 12 data centers, 14 virtual data centers and 31 colocation centers, with connections to 195 ad...
Feb. 7, 2016 11:30 AM EST Reads: 351
[session] From Build to Scale: Lifecycle of Microservices By @fortyfivan | @CloudExpo #Microservices
More and more companies are looking to microservices as an architectural pattern for breaking apart applications into more manageable pieces so that agile teams can deliver new features quicker and more effectively. What this pattern has done more than anything to date is spark organizational transformations, setting the foundation for future application development. In practice, however, there are a number of considerations to make that go beyond simply “build, ship, and run,” which changes ho...
Feb. 7, 2016 10:45 AM EST Reads: 161
Continuous Delivery and Release Automation for Microservices By @Anders_Wallgren | @DevOpsSummit #Microservices
As software organizations continue to invest in achieving Continuous Delivery (CD) of their applications, we see increased interest in microservices architectures, which–on the face of it–seem like a natural fit for enabling CD. In microservices (or its predecessor, “SOA”), the business functionality is decomposed into a set of independent, self-contained services that communicate with each other via an API. Each of the services has their own application release cycle, and are developed and depl...
Feb. 6, 2016 02:00 PM EST Reads: 212
At the heart of the Cloud Native model is a microservices application architecture, and applying this to a telco SDN scenario offers enormous opportunity for product innovation and competitive advantage. For example in the ETSI NFV Ecosystem white paper they describe one of the product markets that SDN might address to be the Home sector. Vendors like Alcatel market SDN-based solutions for the home market, offering Home Gateways – A virtual residential gateway (vRGW) where service provider...
Feb. 6, 2016 01:00 PM EST Reads: 148
In the Bimodal model we find two areas of IT - the traditional kind where the main concern is keeping the lights on and the IT focusing on agility and speed, where everything needs to be faster. Today companies are investing in new technologies and processes to emulate their most agile competitors. Gone are the days of waterfall development and releases only every few months. Today's IT and the business it powers demands performance akin to a supercar - everything needs to be faster, every sc...
Feb. 6, 2016 09:00 AM EST Reads: 513
With microservices, SOA and distributed architectures becoming more popular, it is becoming increasingly harder to keep track of where time is spent in a distributed application when trying to diagnose performance problems. Distributed tracing systems attempt to address this problem by following application requests across service boundaries, persisting metadata along the way that provide context for fine-grained performance monitoring.
Feb. 5, 2016 03:45 PM EST Reads: 800
Web performance issues and advances have been gaining a stronger presence in the headlines as people are becoming more aware of its impact on virtually every business, and 2015 was no exception. We saw a myriad of major outages this year hit some of the biggest corporations, as well as some technology integrations and other news that we IT Ops aficionados find very exciting. This past year has offered several opportunities for growth and evolution in the performance realm — even the worst failu...
Feb. 3, 2016 10:00 PM EST Reads: 547
Are you someone who knows that the number one rule in DevOps is “Don’t Panic”? Especially when it comes to making Continuous Delivery changes inside your organization? Are you someone that theorizes that if anyone implements real automation changes, the solution will instantly become antiquated and be replaced by something even more bizarre and inexplicable?
Feb. 3, 2016 06:30 PM EST Reads: 315
Welcome to the first top DevOps news roundup of 2016! At the end of last year, we saw some great predictions for 2016. While we’re excited to kick off the new year, this week’s top posts reminded us to take a second to slow down and really understand the current state of affairs. For example, do you actually know what microservices are – or aren’t? What about DevOps? Does the emphasis still fall mostly on the development side? This week’s top news definitely got the wheels turning and just migh...
Feb. 3, 2016 03:00 PM EST Reads: 287
Test automation is arguably the most important innovation to the process of QA testing in software development. The ability to automate regression testing and other repetitive test cases can significantly reduce the overall production time for even the most complex solutions. As software continues to be developed for new platforms – including mobile devices and the diverse array of endpoints that will be created during the rise of the Internet of Things - automation integration will have a huge ...
Feb. 3, 2016 02:00 PM EST Reads: 641
Providing a full-duplex communication channel over a single TCP connection, WebSocket is the most efficient protocol for real-time responses over the web. If you’re utilizing WebSocket technology, performance testing will boil down to simulating the bi-directional nature of your application. Introduced with HTML5, the WebSocket protocol allows for more interaction between a browser and website, facilitating real-time applications and live content. WebSocket technology creates a persistent conne...
Feb. 3, 2016 07:00 AM EST Reads: 314