Welcome!

Microservices Expo Authors: Liz McMillan, Elizabeth White, Pat Romanski, Anders Wallgren, AppDynamics Blog

Related Topics: @CloudExpo, Java IoT, Microservices Expo, Linux Containers, Containers Expo Blog, Agile Computing, @BigDataExpo, SDN Journal, @ThingsExpo

@CloudExpo: Article

Predictive Analytics for IT – Filling the Gaps in APM

Predictive analytics solutions for IT can detect, trace and predict performance issues and their root cause

Application Performance Management (APM) grew out of the movement to better align IT with real business concerns. Instead of monitoring a lot of disparate components, such as servers and switches, APM would provide improved visibility into mission-critical application performance and the user experience. Today, APM solutions help IT track end-to-end application response time and troubleshoot coding errors across application components that have an impact on performance.

APM has a rightful place in the arsenal of monitoring tools that IT uses to keep its applications and systems up and running. However, today's APM solutions have some serious gaps and challenges when it comes to providing IT with the entire application performance picture.

Hardware Visibility
Most APM solutions provide minimal information about the hardware and network components underlying application performance, other than showing which components are involved in each part of the transaction. Those that do a better job usually require users to shift to another screen or monitoring system to get more hardware visibility. As with the blind men touching different parts of an elephant, this approach makes it difficult to correlate hardware performance with all the other components driving the application.

The Virtual, Distributed Environment
Most of today's APM solutions were created before virtualization, the cloud, and complex, composite applications took off in the IT environment. With virtual machines migrating back and forth among physical servers at different times of the day or week, and applications dependent on scores of components and cloud services, APM vendors are hard-pressed to provide visibility into the entire scope of a single application.

Predictive Capabilities
As 24 by 7 by 365 uptime becomes increasingly critical to business success, enterprises need to be able to predict and address issues BEFORE they affect the business, rather than after. APM has had mixed success in this area. A recent survey by TRAC Research[1] found that of organizations deploying APM solutions, 60 percent report a success rate of less than half in identifying performance issues before they have an impact on end users.

Enter Predictive Analytics for IT
Filling these APM gaps is how Big Data and predictive analytics for IT can play a significant, highly beneficial role in IT's efforts to maintain application performance. Today, when IT encounters performance issues, it typically has to collect its server, storage, network, and APM folks into a war room to search through mountains of hardware and APM logs, and correlate information manually to isolate the root cause. This resource-intensive process can frequently take hours or even days.

IT has lots of alerts and thresholds to analyze, but those are only as good as the knowledge, experience, and insight of the IT folks who configured them. Just because a server surpassed its CPU utilization threshold doesn't mean that event had anything to do with the root cause of an application issue. Often the real issue is hidden deep in all the delicate interactions among multiple hardware and software components, and may not be reflected in individual thresholds. The same TRAC Research study shows an average of 46.2 hours spent by IT each month in these war rooms searching for root cause. Even more depressing, the root cause is often not found, so IT just reboots everything in the hope that it all works until the same problem rears its ugly head again.

Predictive analytics take over where APM leaves off, harnessing third-generation machine learning and Big Data analysis techniques to efficiently plow through mountains of log data. They discover all the behavior patterns and interrelationships between the IT software and hardware components driving today's mission-critical applications. Over several hours or days, the best solutions baseline the normal behavior of all those components, relationships, and events and use complex algorithms to detect any anomalies that are the early warning signs of developing performance issues. Better yet, because the analytics understand the chain of events involved in the developing anomaly, IT support staff are immediately provided with not only the alert that something is going wrong, but also the behavior of every component involved. This information can shave hours or even days off those war room scenarios. For example, thanks to a predictive analytics for IT solution, a major retailer was able to trace periodic gift card application outages to a misconfigured VLAN. Similarly, a predictive analytics solution reduced - from six hours in the war room to ten minutes - the time it took to diagnose a financial content management performance issue.

Another advantage of predictive analytics solutions is that because they self-learn the normal behavior patterns of underlying components, they drastically reduce the educated guessing that usually goes along with IT staff identifying and setting thresholds against key performance. The inflexibility of these thresholds results in large numbers of false-positive alerts. But with predictive analytics, highly sophisticated algorithms compute the probability of certain behaviors and can therefore generate much more accurate alerts. Some users of predictive analytics solutions have called them the Donald Rumsfelds of IT management tools because they point IT to infrastructure issues they never even knew existed and never looked for. Rumsfeld called these the "unknown unknowns."

However, it is in their ability to be "predictive" that these advanced analytics solutions really shine. By detecting small anomalies early in the game, predictive analytics can alert IT to performance issues and provide enough information to address their root cause before IT or application users even notice them. This can have a dramatic effect on application uptime and performance and a direct impact on user satisfaction and even enterprise revenue. In the case of the document management application, predictive analytics discovered a developing performance issue, and its root cause, the night before it would have affected users placing the application under load on Monday morning.

APM tools have their place in the enterprise, but predictive analytics solutions for IT can kick the effectiveness of those and other IT monitoring tools up a notch by detecting, tracing, and predicting performance issues and their root cause long before any IT war room can.

Resource:

  1. TRAC Research, March 4, 2013: "2013 Application Performance Management Spectrum" report.

More Stories By Rich Collier

Rich Collier is a Principal Solutions Architect with Prelert, a provider of 100% self-learning predictive analytics solutions that augment IT expertise with machine intelligence to dramatically improve IT Operations.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


@MicroservicesExpo Stories
CIOs and those charged with running IT Operations are challenged to deliver secure, audited, and reliable compute environments for the applications and data for the business. Behind the scenes these tasks are often accomplished by following onerous time-consuming processes and often the management of these environments and processes will be outsourced to multiple IT service providers. In addition, the division of work is often siloed into traditional "towers" that are not well integrated for cro...
With an estimated 50 billion devices connected to the Internet by 2020, several industries will begin to expand their capabilities for retaining end point data at the edge to better utilize the range of data types and sheer volume of M2M data generated by the Internet of Things. In his session at @ThingsExpo, Don DeLoach, CEO and President of Infobright, will discuss the infrastructures businesses will need to implement to handle this explosion of data by providing specific use cases for filte...
SYS-CON Events announced today that Avere Systems, a leading provider of enterprise storage for the hybrid cloud, will exhibit at SYS-CON's 18th International Cloud Expo®, which will take place on June 7-9, 2016, at the Javits Center in New York City, NY. Avere delivers a more modern architectural approach to storage that doesn’t require the overprovisioning of storage capacity to achieve performance, overspending on expensive storage media for inactive data or the overbuilding of data centers ...
SYS-CON Events announced today that Commvault, a global leader in enterprise data protection and information management, has been named “Bronze Sponsor” of SYS-CON's 18th International Cloud Expo, which will take place on June 7–9, 2016, at the Javits Center in New York City, NY, and the 19th International Cloud Expo, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. Commvault is a leading provider of data protection and information management...
SYS-CON Events announced today that Alert Logic, Inc., the leading provider of Security-as-a-Service solutions for the cloud, will exhibit at SYS-CON's 18th International Cloud Expo®, which will take place on June 7-9, 2016, at the Javits Center in New York City, NY. Alert Logic, Inc., provides Security-as-a-Service for on-premises, cloud, and hybrid infrastructures, delivering deep security insight and continuous protection for customers at a lower cost than traditional security solutions. Ful...
SYS-CON Events announced today that Interoute, owner-operator of one of Europe's largest networks and a global cloud services platform, has been named “Bronze Sponsor” of SYS-CON's 18th Cloud Expo, which will take place on June 7-9, 2015 at the Javits Center in New York, New York. Interoute is the owner-operator of one of Europe's largest networks and a global cloud services platform which encompasses 12 data centers, 14 virtual data centers and 31 colocation centers, with connections to 195 ad...
In most cases, it is convenient to have some human interaction with a web (micro-)service, no matter how small it is. A traditional approach would be to create an HTTP interface, where user requests will be dispatched and HTML/CSS pages must be served. This approach is indeed very traditional for a web site, but not really convenient for a web service, which is not intended to be good looking, 24x7 up and running and UX-optimized. Instead, talking to a web service in a chat-bot mode would be muc...
SYS-CON Events announced today that Men & Mice, the leading global provider of DNS, DHCP and IP address management overlay solutions, will exhibit at SYS-CON's 18th International Cloud Expo®, which will take place on June 7-9, 2016, at the Javits Center in New York City, NY. The Men & Mice Suite overlay solution is already known for its powerful application in heterogeneous operating environments, enabling enterprises to scale without fuss. Building on a solid range of diverse platform support,...
More and more companies are looking to microservices as an architectural pattern for breaking apart applications into more manageable pieces so that agile teams can deliver new features quicker and more effectively. What this pattern has done more than anything to date is spark organizational transformations, setting the foundation for future application development. In practice, however, there are a number of considerations to make that go beyond simply “build, ship, and run,” which changes ho...
WebSocket is effectively a persistent and fat pipe that is compatible with a standard web infrastructure; a "TCP for the Web." If you think of WebSocket in this light, there are other more hugely interesting applications of WebSocket than just simply sending data to a browser. In his session at 18th Cloud Expo, Frank Greco, Director of Technology for Kaazing Corporation, will compare other modern web connectivity methods such as HTTP/2, HTTP Streaming, Server-Sent Events and new W3C event APIs ...
SYS-CON Events announced today that AppNeta, the leader in performance insight for business-critical web applications, will exhibit and present at SYS-CON's @DevOpsSummit at Cloud Expo New York, which will take place on June 7-9, 2016, at the Javits Center in New York City, NY. AppNeta is the only application performance monitoring (APM) company to provide solutions for all applications – applications you develop internally, business-critical SaaS applications you use and the networks that deli...
Microservices are all the rage right now — and the industry is still learning, experimenting, and developing patterns, for successfully designing, deploying and managing Microservices in the real world. Are you considering jumping on the Microservices-wagon? Do Microservices make sense for your particular use case? What are some of the “gotchas” you should be aware of? This morning on #c9d9 we had experts from popular chat app Kik, SMB SaaS platform Yodle and hosted CI solution Semaphore sha...
Microservices are a type of software architecture where large applications are made up of small, self-contained units working together through APIs that are not dependent on a specific language. Each service has a limited scope, concentrates on a specific task and is highly independent. This setup allows IT managers and developers to build systems in a modular way. In his book, “Building Microservices,” Sam Newman said microservices are small, focused components built to do a single thing very w...
How is your DevOps transformation coming along? How do you measure Agility? Reliability? Efficiency? Quality? Success?! How do you optimize your processes? This morning on #c9d9 we talked about some of the metrics that matter for the different stakeholders throughout the software delivery pipeline. Our panelists shared their best practices.
The (re?)emergence of Microservices was especially prominent in this week’s news. What are they good for? do they make sense for your application? should you take the plunge? and what do Microservices mean for your DevOps and Continuous Delivery efforts? Continue reading for more on Microservices, containers, DevOps culture, and more top news from the past week. As always, stay tuned to all the news coming from@ElectricCloud on DevOps and Continuous Delivery throughout the week and retweet/favo...
The cloud promises new levels of agility and cost-savings for Big Data, data warehousing and analytics. But it’s challenging to understand all the options – from IaaS and PaaS to newer services like HaaS (Hadoop as a Service) and BDaaS (Big Data as a Service). In her session at @BigDataExpo at @ThingsExpo, Hannah Smalltree, a director at Cazena, will provide an educational overview of emerging “as-a-service” options for Big Data in the cloud. This is critical background for IT and data profes...
In a previous article, I demonstrated how to effectively and efficiently install the Dynatrace Application Monitoring solution using Ansible. In this post, I am going to explain how to achieve the same results using Chef with our official dynatrace cookbook available on GitHub and on the Chef Supermarket. In the following hands-on tutorial, we’ll also apply what we see as good practice on working with and extending our deployment automation blueprints to suit your needs.
Father business cycles and digital consumers are forcing enterprises to respond faster to customer needs and competitive demands. Successful integration of DevOps and Agile development will be key for business success in today’s digital economy. In his session at DevOps Summit, Pradeep Prabhu, Co-Founder & CEO of Cloudmunch, covered the critical practices that enterprises should consider to seamlessly integrate Agile and DevOps processes, barriers to implementing this in the enterprise, and pr...
If we look at slow, traditional IT and jump to the conclusion that just because we found its issues intractable before, that necessarily means we will again, then it’s time for a rethink. As a matter of fact, the world of IT has changed over the last ten years or so. We’ve been experiencing unprecedented innovation across the board – innovation in technology as well as in how people organize and accomplish tasks. Let’s take a look at three differences between today’s modern, digital context...
Sensors and effectors of IoT are solving problems in new ways, but small businesses have been slow to join the quantified world. They’ll need information from IoT using applications as varied as the businesses themselves. In his session at @ThingsExpo, Roger Meike, Distinguished Engineer, Director of Technology Innovation at Intuit, showed how IoT manufacturers can use open standards, public APIs and custom apps to enable the Quantified Small Business. He used a Raspberry Pi to connect sensors...