Welcome!

Microservices Expo Authors: Kalyan Ramanathan, Carmen Gonzalez, Pat Romanski, Liz McMillan, Elizabeth White

Related Topics: @CloudExpo, Java IoT, Microservices Expo, Linux Containers, @BigDataExpo, SDN Journal

@CloudExpo: Article

Mainframe: A Resilient Model for the Modern Cloud

The emerging cloud-based model of computing requires systems that can provide fast response times to huge volumes of requests

Technology is moving at a blistering pace. In today's era of data-centric, complex environments where the lines between business and technology are becoming increasingly blurred, organizations are moving beyond virtualization to cloud computing to meet new challenges and keep up with the pace of change. Critical investments are needed to keep companies competitive, and chief among these technologies is cloud computing. In fact, Gartner expects cloud computing to become the bulk of new IT expenditure by 2016. The bottom line is, if you're not already looking at cloud as an essential investment, you're risking your survival into the next era of computing.

The emerging cloud-based model of computing requires systems that can provide very fast response times to huge volumes of requests. And, mission critical services such as healthcare, finance, transportation, public utilities, and other industries require very high levels of availability, security and other industrial-strength capabilities. Those attributes, qualities and requirements make the mainframe the ideal platform for such mission critical cloud-based workloads.

Cloud computing is a modern extension of a concept first developed nearly 50 years ago with the mainframe. The inherent spirit behind mainframe based computing was to serve users in remote locations at the same time, on a pay-as-you-go basis. The mainframe was introduced as the most robust, scalable system ever built, and with continued innovation the system has maintained its leadership status as one of the platforms of choice to handle today's complex workloads including sophisticated public, private and hybrid cloud computing environments. At its core, the mainframe was designed around three key traits - virtualization, standardization and provisioning. Not coincidentally, these are the foundational requirements for true cloud implementation.

Most enterprises today started their cloud journey with low-risk applications and high agility requirements. This approach allows customers to ease into cloud computing, learn and adjust their management of the cloud, and build the confidence to introduce more demanding applications.  The applications tend to use web technologies and architectures that can be scaled on commodity infrastructures, using load balancing and service cloning. Batch workloads that fit with commodity infrastructures are another popular workload on clouds.

For private, public or hybrid clouds, the mainframe can provide the following key requirements:

  • Scalability - users need to scale quickly and efficiently both up and down with complete confidence and zero loss of availability.
  • Reliability - a cloud computing environment that is always accessible with guaranteed application performance, limited to no downtime with provisions for rapid recovery from failure.
  • Multi-Tenancy - allowing multiple users to access software applications on the same system, concurrently and securely, critical for cloud service providers hosting many organizations in a single cloud infrastructure and for enterprises deploying private clouds to manage growth through acquisitions to host multiple companies in the same infrastructure;
  • Cost Efficiency - consolidating a distributed x86 cloud environment onto one mainframe creates a simplified, more efficient environment with reductions in floor space and power requirements, and higher return on investment over the life of the platform;
  • Security - the mainframe has unmatched system security with ensured isolation and protection of each virtual server environment.

Companies across various industries are gaining these advantages and efficiencies by consolidating cloud environments on a mainframe, such as:

By consolidating cloud on a mainframe private cloud solution that replaced thousands of standalone servers for its daily business activities like policy verification, claims processing, and generating customer quotations, Nationwide Insurance has saved 80% in energy and facility costs. The consolidation saved the company roughly $15 million over three years and will only continue to efficiently keep costs down in the future. Additionally, this solution gives them the capacity, processing speeds and reliability to increase the pace of innovation across its products and channels as it continues to grow.

By leveraging the cloud capabilities offered by the mainframe, Marist College was able to extend its business analytics technology to its academic community including researchers and students, while extracting even more value from its IT investments. By providing its analytics technology via cloud, the college has been able to expose analytics tools to a wide variety of programs, including technical disciplines and also business, liberal arts and communications programs so students learn how to apply it to their fields of study. Marist has also realized significant financial benefits, saving roughly $350,000 by using the cloud to support the college's ERP system.

The mainframe, with its shared platform, integration, and secure design attributes combined with continuous innovation, has enabled organizations to stay ahead of changing market dynamics with a solution that embodies efficiency, economics and agility - a resilient solution for today's cloud environment.

More Stories By Jose Castano

Jose Castano is the Director for the System z Growth Initiatives in IBM’s Systems & Technology Group. He has over 25 years of experience within IBM and has held multiple key positions in System z during this tenure.

Jose has worldwide responsibility to drive new workloads on System z. This includes Cloud, Analytics, Mobile, and Security He sets the business and technical strategy and direction for the System z platform. He drives coordination and collaboration of the System z ecosystem, from marketing, sales, business partners, consultants, and most importantly customers; leading the platform through an evolution that maintains leadership and meets customer and industry requirements.

Jose has a team comprised of workload and industry architects (who focus on business trends, market and industry requirements and develop solutions/offerings). Offering managers (who are responsible for the GTM for the solutions/offerings) and ISV managers (who work with our ecosystem to support new and existing workloads). Together, these teams have responsibility for researching, designing, building and maintaining the new workload strategy and its roadmap for IBM System z, driving the plans for the next 3-5 years.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


@MicroservicesExpo Stories
Logs are continuous digital records of events generated by all components of your software stack – and they’re everywhere – your networks, servers, applications, containers and cloud infrastructure just to name a few. The data logs provide are like an X-ray for your IT infrastructure. Without logs, this lack of visibility creates operational challenges for managing modern applications that drive today’s digital businesses.
Financial Technology has become a topic of intense interest throughout the cloud developer and enterprise IT communities. Accordingly, attendees at the upcoming 20th Cloud Expo at the Javits Center in New York, June 6-8, 2017, will find fresh new content in a new track called FinTech.
You have great SaaS business app ideas. You want to turn your idea quickly into a functional and engaging proof of concept. You need to be able to modify it to meet customers' needs, and you need to deliver a complete and secure SaaS application. How could you achieve all the above and yet avoid unforeseen IT requirements that add unnecessary cost and complexity? You also want your app to be responsive in any device at any time. In his session at 19th Cloud Expo, Mark Allen, General Manager of...
@DevOpsSummit taking place June 6-8, 2017 at Javits Center, New York City, is co-located with the 20th International Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world. @DevOpsSummit at Cloud Expo New York Call for Papers is now open.
The 20th International Cloud Expo has announced that its Call for Papers is open. Cloud Expo, to be held June 6-8, 2017, at the Javits Center in New York City, brings together Cloud Computing, Big Data, Internet of Things, DevOps, Containers, Microservices and WebRTC to one location. With cloud computing driving a higher percentage of enterprise IT budgets every year, it becomes increasingly important to plant your flag in this fast-expanding business opportunity. Submit your speaking proposal ...
"Dice has been around for the last 20 years. We have been helping tech professionals find new jobs and career opportunities," explained Manish Dixit, VP of Product and Engineering at Dice, in this SYS-CON.tv interview at 19th Cloud Expo, held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA.
Rapid innovation, changing business landscapes, and new IT demands force businesses to make changes quickly. In the eyes of many, containers are at the brink of becoming a pervasive technology in enterprise IT to accelerate application delivery. In this presentation, attendees learned about the: The transformation of IT to a DevOps, microservices, and container-based architecture What are containers and how DevOps practices can operate in a container-based environment A demonstration of how ...
Cloud Expo, Inc. has announced today that Andi Mann returns to 'DevOps at Cloud Expo 2017' as Conference Chair The @DevOpsSummit at Cloud Expo will take place on June 6-8, 2017, at the Javits Center in New York City, NY. "DevOps is set to be one of the most profound disruptions to hit IT in decades," said Andi Mann. "It is a natural extension of cloud computing, and I have seen both firsthand and in independent research the fantastic results DevOps delivers. So I am excited to help the great t...
Without lifecycle traceability and visibility across the tool chain, stakeholders from Planning-to-Ops have limited insight and answers to who, what, when, why and how across the DevOps lifecycle. This impacts the ability to deliver high quality software at the needed velocity to drive positive business outcomes. In his general session at @DevOpsSummit at 19th Cloud Expo, Phil Hombledal, Solution Architect at CollabNet, discussed how customers are able to achieve a level of transparency that e...
Get deep visibility into the performance of your databases and expert advice for performance optimization and tuning. You can't get application performance without database performance. Give everyone on the team a comprehensive view of how every aspect of the system affects performance across SQL database operations, host server and OS, virtualization resources and storage I/O. Quickly find bottlenecks and troubleshoot complex problems.
SYS-CON Events announced today that Dataloop.IO, an innovator in cloud IT-monitoring whose products help organizations save time and money, has been named “Bronze Sponsor” of SYS-CON's 20th International Cloud Expo®, which will take place on June 6-8, 2017, at the Javits Center in New York City, NY. Dataloop.IO is an emerging software company on the cutting edge of major IT-infrastructure trends including cloud computing and microservices. The company, founded in the UK but now based in San Fran...
DevOps is being widely accepted (if not fully adopted) as essential in enterprise IT. But as Enterprise DevOps gains maturity, expands scope, and increases velocity, the need for data-driven decisions across teams becomes more acute. DevOps teams in any modern business must wrangle the ‘digital exhaust’ from the delivery toolchain, "pervasive" and "cognitive" computing, APIs and services, mobile devices and applications, the Internet of Things, and now even blockchain. In this power panel at @...
@DevOpsSummit at Cloud taking place June 6-8, 2017, at Javits Center, New York City, is co-located with the 20th International Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world. The widespread success of cloud computing is driving the DevOps revolution in enterprise IT. Now as never before, development teams must communicate and collaborate in a dynamic, 24/7/365 environment. There is no time to wait for long developm...
SYS-CON Events has announced today that Roger Strukhoff has been named conference chair of Cloud Expo and @ThingsExpo 2017 New York. The 20th Cloud Expo and 7th @ThingsExpo will take place on June 6-8, 2017, at the Javits Center in New York City, NY. "The Internet of Things brings trillions of dollars of opportunity to developers and enterprise IT, no matter how you measure it," stated Roger Strukhoff. "More importantly, it leverages the power of devices and the Internet to enable us all to im...
Kubernetes is a new and revolutionary open-sourced system for managing containers across multiple hosts in a cluster. Ansible is a simple IT automation tool for just about any requirement for reproducible environments. In his session at @DevOpsSummit at 18th Cloud Expo, Patrick Galbraith, a principal engineer at HPE, discussed how to build a fully functional Kubernetes cluster on a number of virtual machines or bare-metal hosts. Also included will be a brief demonstration of running a Galera MyS...
Keeping pace with advancements in software delivery processes and tooling is taxing even for the most proficient organizations. Point tools, platforms, open source and the increasing adoption of private and public cloud services requires strong engineering rigor – all in the face of developer demands to use the tools of choice. As Agile has settled in as a mainstream practice, now DevOps has emerged as the next wave to improve software delivery speed and output. To make DevOps work, organization...
As we enter the final week before the 19th International Cloud Expo | @ThingsExpo in Santa Clara, CA, it's time for me to reflect on six big topics that will be important during the show. Hybrid Cloud: This general-purpose term seems to provide a comfort zone for many enterprise IT managers. It sounds reassuring to be able to work with one of the major public-cloud providers like AWS or Microsoft Azure while still maintaining an on-site presence.
Between 2005 and 2020, data volumes will grow by a factor of 300 – enough data to stack CDs from the earth to the moon 162 times. This has come to be known as the ‘big data’ phenomenon. Unfortunately, traditional approaches to handling, storing and analyzing data aren’t adequate at this scale: they’re too costly, slow and physically cumbersome to keep up. Fortunately, in response a new breed of technology has emerged that is cheaper, faster and more scalable. Yet, in meeting these new needs they...
In his general session at 19th Cloud Expo, Manish Dixit, VP of Product and Engineering at Dice, discussed how Dice leverages data insights and tools to help both tech professionals and recruiters better understand how skills relate to each other and which skills are in high demand using interactive visualizations and salary indicator tools to maximize earning potential. Manish Dixit is VP of Product and Engineering at Dice. As the leader of the Product, Engineering and Data Sciences team at D...
I’m a huge fan of open source DevOps tools. I’m also a huge fan of scaling open source tools for the enterprise. But having talked with my fair share of companies over the years, one important thing I’ve learned is that you can’t scale your release process using open source tools alone. They simply require too much scripting and maintenance when used that way. Scripting may be fine for smaller organizations, but it’s not ok in an enterprise environment that includes many independent teams and to...