Click here to close now.

Welcome!

Microservices Journal Authors: Pat Romanski, Liz McMillan, XebiaLabs Blog, Elizabeth White, Louis Evans

Related Topics: Cloud Expo, Microservices Journal, Linux, Open Source, Web 2.0, Security

Cloud Expo: Article

Open Source PaaS for Parallel Cloud Application Development

Architectural advantages of cross-layer optimized parallel computing PaaS

Improving application program performance will require parallelizing the program execution at ever finer granularity now that the processor clock rates are no longer increasing. However, even in a per-application dedicated computing environment, the parallelization overhead is known to place a limit on how much application on-time throughput performance increase can be achieved via higher levels of parallel processing. The throughput-limiting impact of parallelization overhead will be significantly amplified when executing multiple internally parallelized applications on dynamically shared cloud computing environment, since the allocation of processing resources to instances and tasks of any given application cannot be done in isolation, but instead it needs to be done collectively across all the applications dynamically sharing the given pool of computing resources. There thus is an urgent need to solve this complex challenge of developing internally parallelized programs for dynamic execution on shared cloud computing infrastructure, if we expect to be able scale the performance and capacity of cloud hosted applications going forward.

PaaS Approach for Parallel Cloud Computing Challenges
The Need for Cross-layer Optimized Platform

For effectiveness of a parallel cloud computing platform, what is essential is how well the platform elements perform together, rather than individually. Consequently, the complex, interleaved challenges of parallel cloud application development and execution cannot be solved by any single layer or element of traditional computing architectures alone. Instead, a comprehensive, cross-layer optimized platform architecture is needed.

This new platform architecture will involve a parallel program development environment for producing application executables based on actors that can be efficiently mapped for concurrent execution on processing cores of dynamically shared manycore arrays. In addition, the parallel cloud computing platform will need an execution environment that, besides executing the program instructions on processing cores, takes care of the dynamic parallel execution routines on behalf of the applications, so that the processing cores are used for executing the actual client programs instead of system functions. If the hardware of the manycore processor fabric did not handle the parallel execution routines[1], eventually the system would be just managing itself rather than providing increased user application throughput as the numbers of processing cores and applications and tasks sharing them are scaled up. This difference in enabling application on-time throughput scalability is illustrated at Figures 4 and 5 (in Ch. 3 Architectural Advantages of Cross-layer Optimized Parallel Computing PaaS).

The Platform Architecture
Vision

Figure 1 below illustrates the overall architecture of the envisioned open parallel program development and execution platform as a service (PaaS).

Figure 1: Overview of the parallel cloud computing PaaS

As illustrated in Figure 1, the open parallel computing PaaS has an open-sourced parallel program development environment and a dynamic parallel execution environment. The development environment allows the platform users to produce executables of their application programs that are made of segments (tasks/actors/threads) that can execute concurrently on parallel (incl. pipelined) processing cores. The execution environment provides, besides an array of processing cores for parallel execution of the user program tasks, the capabilities to dynamically map the highest-priority ready application task instances for execution on their assigned cores. While the state (e.g. which application task instance is mapped to any given core of the shared resource pool at any given time) of the execution environment is highly dynamic during the application runtime, in the platform architecture per Figure 1, the execution environment provides for the application programs (and their developers) a virtual static view of it; any given application can assume that each of its task instances is always active and mapped to a virtual static core in an array that is virtually dedicated to that single application. The hardware automated parallel processing runtime routines of the execution environment per Figure 1 hide (to the desired level) the dynamic details of the processing hardware from the applications as well as the development environment software, thereby providing a higher level, simplified abstraction of the execution environment for the software. The raised level of the software-hardware interface per Figure 1 enables greater productivity for both realizing the much needed open, comprehensive parallel programming environment as well as for developing parallelized applications for cloud deployment, as software does not need to be concerned of the dynamic parallel execution details.

Motivation
Importantly, there is the need to coordinate the various development activities concerning parallel programming and cloud computing tools etc. base technologies (e.g. languages, compilers, parallel file systems, data bases, etc.) around a common overall framework so that the individual technology elements work efficiently together, to enable high productivity development of parallelized applications for cloud deployment. ThroughPuter proposes that the elevated level of the interface to the parallel execution environment per the platform architecture of Figure 1 provides a compelling motive for why the software technology development activities for parallel cloud computing should be based on this execution environment model. More specifically, the major reasons for software ecosystem for parallel computing platform to be based on the dynamic parallel execution environment interface standard per figure include:

  1. Greater productivity through less low-level work: The execution environment in the platform architecture per Figure 1 automates dynamic parallel execution routines in (programmable) hardware, providing higher level application development interface for the software of the PaaS.
  2. Higher performance and scalability via eliminating system software overhead by hardware automation of system tasks such as optimally allocating processing core capacity, scheduling and placing application tasks for execution, inter-task communications, billing etc.
  3. Built-in cloud computing security from the hardware level up: unauthorized interactions between different applications simply not enabled in the hardware.
  4. Open source software and open standard interface between development and execution environment: users have full freedom to choose where to host the development environment as well as the parallelized application executables produced by it. Any execution environment implementation complying to the simple, open execution environment interface standard per Figure 1 is a possible hosting option so there is nothing vendor specific about this platform architecture, and users will not suffer from vendor lock-in.

Technical Overview
Application software developers access the envisioned open parallel cloud computing PaaS through its Integrated Development Environment (IDE) to build application software optimized for parallel processing on dynamically shared cloud processors. The IDE provides a web-based program flow chart, code advisor, profiler etc. (GUI) tools to assist and automate parallelizing the users' programs. The IDE also includes the automated back-end development tool chain, incl. compiler, linker and loader programs, for building and executing the user's parallelized application in a processing hardware complying to the discussed execution environment interface standard per Figure 1. The IDE further equips the user application programs with the system software that automates and optimizes the minimal (to none) interactions between the user programs and the hardware operating system of the dynamic parallel execution environment of the parallel cloud computing PaaS[2].

The IDE software can be hosted with the same parallel computing PaaS as the applications it produces. This practice avoids the need for cross-compiling, and enables straightforward and rapid-cycle interactive testing, debugging and optimization of the parallelized application programs, as well as deployment and scaling releases of the user applications.

The software of the parallel cloud computing PaaS is developed and made fully available as an open-source project, and can be integrated with popular open-source IDEs. In essence, this software project is to add the parallel programming development tools to the major open source IDE and PaaS code bases, while utilizing (and further developing) the applicable existing features from them.

The openness of the promoted comprehensive parallel cloud computing PaaS, besides the open-source IDE, is also manifested via the simple, open standard interface between the development and execution environments of the PaaS architecture. This open standard interface enables any user to host the IDE as well as the parallel program executables produced by it anywhere, e.g. at the user premises, or with ThroughPuter or any 3rd party. Alternative I/PaaS providers furthermore are encouraged to support the execution/development environment interface via their respective implementations of either side or both sides of that interface. The customizable, open-source IDE and open-standard interface to the execution environment provides the users and collaborators a flexible and productive way to approach the major, must-solve parallel cloud application development and execution challenge that is facing much of the software industry and its customers.

The efficient dynamic parallel processing features - which will be critically needed as user application throughput requirements begin exceeding what is available from conventional sequential execution models, and as the parallelized applications will be cloud hosted -- of this open-source software based PaaS are delivered by an execution environment that provides the necessary, dynamic parallel execution core to application task instance allocation, task instance to core assignment, and inter-task communication capabilities. These critical parallel execution capabilities are an integrated feature in the ThroughPuter hosted commercial PaaS offering.

For reference on ThroughPuter's implementation of the dynamic parallel execution environment for the open parallel computing PaaS architecture per Figure 1, the below Figure 2 shows ThroughPuter's realtime application load adaptive manycore processor architecture, highlighting its hardware logic automated operating system functionality enabling a number of customer application programs to securely, dynamically and cost-efficiently share the processing capacity of the manycore processor hardware.

Figure 2: Reference diagram for ThroughPuter manycore processors with hardware-automated multi-user parallel processing optimized operating system.

ThroughPuter's hardware operating system, manycore fabric memory and I/O subsystems are largely responsible for the architectural security, performance and cost-efficiency benefits of the ThroughPuter PaaS. However, the IDE hides the actual execution environment features from the user; the user does not need to be aware of the novel hardware-implemented capabilities of this dynamic parallel execution environment in order to realize the performance benefits.

A possible core to application task/instance allocation scenario over a few core allocation periods (CAPs) is illustrated in Figure 3 below, along with associated highlights of the feature benefits.

Figure 3: Dynamic core to application task instance assignment scenario in the execution environment of the PaaS per Figure 1.

In reference to the dynamic core assignment scenario per Figure 3, it shall be remembered that in the PaaS architecture per Figure 1, the software does not need to be aware of the dynamic parallel execution matters, but can instead maintain a virtual static view of the execution environment where each possible application task instance is constantly mapped to its virtual dedicated core, thus simplifying both the development environment as well as the application software while improving the development productivity and runtime performance.

Architectural Advantages of Cross-layer Optimized Parallel Computing PaaS
The platform model presented here is crucial for enabling application on-time throughput performance scalability in the age of parallel cloud computing, as illustrated in Figures 4 and 5.

Figure 4: Scalability problem in parallel cloud computing due to the system software overhead.

As illustrated in Figure 4, when relying on system software for managing the parallel execution, there is a limit to scalability of cloud computing platforms in the parallel processing era, as application performance improvement begins requiring ever finer grades of intra-application parallelism (i.e. more tasks and task instances per application). This is due to that the parallel processing system software overhead (the need to dynamically coordinate and manage concurrent tasks and parallel processing resources) increases with the number of applications as well as their tasks and instances, and the number of processing cores being dynamically shared among them. When relying on system software to handle parallel execution routines, the processing capacity of a given manycore processor (array) is split between processing user applications and system software, with the rate of system software increasing with scale, at the expense of the user applications. This causes that after some point, the system-wide application processing on-time throughput (the product of number of cores and the percentage of the cores' processing capacity available for user applications) will begin decreasing as the system is scaled up (by adding processing cores and parallelized applications and their tasks sharing the cores).

To solve this fundamental challenge affecting the scalability of cloud computing in the parallel processing era, the processing hardware needs to raise up to this challenge and handle the parallel processing routines in the hardware of the manycore processors, so that the processing cores will be optimally used for processing the user applications (rather than for processing the system functions, and/or be locked to low utilization due to non-load-adaptive allocation).

The impact of the hardware automation of the parallel processing system functions in the hardware of manycore processors per Figures 1-3 is illustrated in Figure 5 below.

Figure 5: Scalability solution for parallel cloud computing delivered via automating the parallel execution system functions in hardware.

The enabling of scalability of cloud computing platforms and cloud applications' on-time processing throughput in the emerging era of (inter and intra) application parallel processing by the execution environment model per Figure 1 serves as a further compelling reason for concentrating the efforts to address the popular parallel programming challenge via the herein presented open parallel cloud computing platform model. Parties interested in collaboration to realize this much needed comprehensive, open source parallel cloud computing platform can contact ThroughPuter via [email protected].

Further Reading
Relevant further material on the parallel program development and execution challenges is available at:

References:

  1. These include: monitoring application processing load demands, periodically allocating processing resources (cores) among the applications based on their processing load variations and contractual entitlements, prioritizing and selecting application task instances for execution, mapping selected task instances for execution on their assigned cores and accordingly configuring the IO and memory access subsystems, arranging the inter task communications, plus contract billing based on applications' resource entitlement and usage.
  2. These interactions are mainly limited to the application program, via the PaaS tool-generated system software, providing to the hardware operating system of the execution environment (per Figure 1) a listing of its schedulable tasks/instances in their priority order; there is very little overhead in interacting with such hardware operating system. Where applicable, the hardware operating system of the PaaS is able to deduce the processing core demands and task/instance priority orders of the applications by itself by monitoring the input processing data load levels for the applications. This feature, where employed for a given application (or task group), will effectively eliminate all the parallelization system software overhead for the given application (task group).

More Stories By Mark Sandstrom

Mark Sandstrom is the president of ThroughPuter, Inc., developer of dynamic parallel program execution technologies with a business model of PaaS provider. He is an innovator and strategist with experience in the high technology industry since 1995, including at Optimum Communications Services, Inc., Turin Networks, Inc. (acquired by Force 10, then by Dell), Cyras Systems, Inc. (acquired by CIENA) and Tellabs, Inc.

Sandstrom holds an MSEE degree from Helsinki University of Technology and Executive MBA from Golden Gate University, and has been granted sixteen US and UK patents in fields of networking and computing system throughput optimization and management system streamlining.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


@MicroservicesExpo Stories
How do you securely enable access to your applications in AWS without exposing any attack surfaces? The answer is usually very complicated because application environments morph over time in response to growing requirements from your employee base, your partners and your customers. In his session at 16th Cloud Expo, Haseeb Budhani, CEO and Co-founder of Soha, will share five common approaches that DevOps teams follow to secure access to applications deployed in AWS, Azure, etc., and the frict...
of cloud, colocation, managed services and disaster recovery solutions, will exhibit at SYS-CON's 16th International Cloud Expo®, which will take place on June 9-11, 2015, at the Javits Center in New York City, NY. TierPoint, LLC, is a leading national provider of information technology and data center services, including cloud, colocation, disaster recovery and managed IT services, with corporate headquarters in St. Louis, MO. TierPoint was formed through the strategic combination of some of t...
SYS-CON Events announced today that Soha will exhibit at SYS-CON's DevOps Summit New York, which will take place on June 9-11, 2015, at the Javits Center in New York City, NY. Soha delivers enterprise-grade application security, on any device, as agile as the cloud. This turnkey, cloud-based service enables customers to solve secure application access and delivery challenges that traditional or virtualized network solutions cannot solve because they are too expensive, inflexible and operational...
There's a real buzz about microservices and containers in the application development and DevOps communities, and of course these are topics we've been talking about a great deal lately here at XebiaLabs too. Microservices and containers offer many attractive features, not least the potential for enhanced flexibility, and a robust architecture based on best-fit services. What we at XebiaLabs are really interested in is how organizations can effectively deliver microservices-based apps to bett...
SYS-CON Events announced today that Vicom Computer Services, Inc., a provider of technology and service solutions, will exhibit at SYS-CON's 16th International Cloud Expo®, which will take place on June 9-11, 2015, at the Javits Center in New York City, NY. They are located at booth #427. Vicom Computer Services, Inc. is a progressive leader in the technology industry for over 30 years. Headquartered in the NY Metropolitan area. Vicom provides products and services based on today’s requirements...
SYS-CON Events announced today that Tufin, the market-leading provider of Security Policy Orchestration Solutions, will exhibit at SYS-CON's 16th International Cloud Expo®, which will take place on June 9-11, 2015, at the Javits Center in New York City, NY. As the market leader of Security Policy Orchestration, Tufin automates and accelerates network configuration changes while maintaining security and compliance. Tufin's award-winning Orchestration Suite™ gives IT organizations the power and a...
Today, the demand for new applications is growing at an unprecedented rate throughout lines of business and across industries. Customer expectations for mobile and e-commerce capabilities are transforming software development speed and quality into a competitive differentiator for even the most unlikely businesses. For existing software development shops, the proliferation of platforms, increasing need for total global uptime, and accelerating pace of industry disruption by fast-paced startups h...
SYS-CON Events announced today that MangoApps will exhibit at SYS-CON's 16th International Cloud Expo®, which will take place on June 9-11, 2015, at the Javits Center in New York City, NY., and the 17th International Cloud Expo®, which will take place on November 3–5, 2015, at the Santa Clara Convention Center in Santa Clara, CA. MangoApps provides private all-in-one social intranets allowing workers to securely collaborate from anywhere in the world and from any device. Social, mobile, and eas...
SYS-CON Events announced today that Cloudian, Inc., the leading provider of hybrid cloud storage solutions, will exhibit at SYS-CON's 16th International Cloud Expo®, which will take place on June 9-11, 2015, at the Javits Center in New York City, NY. Cloudian, Inc., is a Foster City, California - based software company specializing in cloud storage software. The main product is Cloudian, an Amazon S3-compliant cloud object storage platform, the bedrock of cloud computing systems, that enables c...
Change becomes the central principle of today’s enterprises, and thus business agility becomes the most important characteristic our organizations must exhibit. Agile Architecture lays out a best practice approach for achieving this agility – and thus drives and coordinates the other revolutions, as both digital and DevOps are about being able to deal with change better as well.
What exactly is a cognitive application? In her session at 16th Cloud Expo, Ashley Hathaway, Product Manager at IBM Watson, will look at the services being offered by the IBM Watson Developer Cloud and what that means for developers and Big Data. She'll explore how IBM Watson and its partnerships will continue to grow and help define what it means to be a cognitive service, as well as take a look at the offerings on Bluemix. She will also check out how Watson and the Alchemy API team up to off...
Thought experiment: let’s say your app gets a message from somewhere, perhaps from another app, but you don’t know from where. The message contains the number 47 but no other information. What should your app do with the message? The answer: nothing. There’s no way for your app to make any sense out of a single datum with no context, no additional information or metadata about the datum itself. Now, let’s scale up this thought experiment to a data lake. There are a few common definitions o...
The 17th International Cloud Expo has announced that its Call for Papers is open. 17th International Cloud Expo, to be held November 3-5, 2015, at the Santa Clara Convention Center in Santa Clara, CA, brings together Cloud Computing, APM, APIs, Microservices, Security, Big Data, Internet of Things, DevOps and WebRTC to one location. With cloud computing driving a higher percentage of enterprise IT budgets every year, it becomes increasingly important to plant your flag in this fast-expanding bu...
The 5th International DevOps Summit, co-located with 17th International Cloud Expo – being held November 3-5, 2015, at the Santa Clara Convention Center in Santa Clara, CA – announces that its Call for Papers is open. Born out of proven success in agile development, cloud computing, and process automation, DevOps is a macro trend you cannot afford to miss. From showcase success stories from early adopters and web-scale businesses, DevOps is expanding to organizations of all sizes, including the...
The concept and subsequent adoption of 'Containerization'' is growing at a rapid speed with the support of almost every other major player in the industry. This concept is much more efficient than the Virtualization which has been a major option for Infrastructure optimization in the past decade. The following factors distinguish a Container from a Virtual Machine. Containers contain Only the Application Specific libraries and binaries. They do not include a guest operating system. Rather ...
As we recently previewed (read more about our London PoP in Jesse's post), Blue Box is opening a new Data Center in London, but hadn't announced the provider. Today we're excited to partner with TelecityGroup, whom we've selected as our data center partner in London. We chose their Powergate location, which is one of the U.K.'s most advanced, flexible and energy efficient carrier-neutral data centres. Why does that matter to you? Well, when customers choose Blue Box, they're trusting us with ...
Cloud Expo New York is happening from June 9 - 11. This event brings together the worlds of Cloud Computing, DevOps, IoT, WebRTC, Big Data and SDDC. We hope to see you there-members of the Blue Box team will exhibit in booth 218 next to the DevOps area. Plus, our Chief Product Officer, Hernan Alvarez, will present his talk "The Cloud Has a Down-and-Dirty Lining" as part of the Operations track in the DevOps Summit portion of the event on June 9 at 11 am. Learn more about his session her...
SYS-CON Events announced today that Column Technologies, a global technology solutions company, will exhibit at SYS-CON's DevOps Summit 2015 New York, which will take place June 9-11, 2015, at the Javits Center in New York City, NY. Established in 1998, Column Technologies is a leader in application performance and infrastructure management for commercial and federal markets. The company is headquartered in the United States, with a diverse and talented team of more than 350 employees around th...
Financial services organizations were among the earliest enterprise adopters of cloud computing. The ability to leverage massive compute, storage and networking resources via RESTful APIs and automated tools like Chef and Puppet made it possible for their high-horsepower IT users to develop a whole new array of applications. Companies like Wells Fargo, Fidelity and BBVA are visible, vocal and engaged supporters of the OpenStack community, running production clouds for applications ranging from d...
Operationalizing the network continues to be a driving force behind DevOps and SDN. The ability to solve real problems using programmability to automate and orchestrate infrastructure provisioning and configuration across the application release process remains the hope for many interested in one or the other - and often times both. A recent Avaya sponsored, Dynamic Markets survey (reg required) dove deep into the demesne of SDN and found that many of the problems companies have - and expect ...