Welcome!

Microservices Expo Authors: Pat Romanski, Carmen Gonzalez, Elizabeth White, Ken Schwaber, Aruna Ravichandran

Related Topics: Microservices Expo, Java IoT, PowerBuilder, Microsoft Cloud, @CloudExpo, Apache

Microservices Expo: Article

From Efficiency to Effectiveness: The Role of Data

The excitement about Big Data is really about being able to take advantage of the data in which we are all awash

Efficiency may be the most commonly used term in enterprise software marketing - that or "ensure." And not without reason - efficiency is one of the key value propositions of most enterprise software, from collaboration tools, to productivity tools, to integration tools and beyond. At a certain point though, the gains to be achieved from efficiency become smaller and smaller and of lesser and lesser business significance.

This is resulting in a shift in focus from efficiency to effectiveness. At times, these goals are twin, but in many cases, they are not - the most effective allocation of resources may not be the most efficient - at least in the short-term. Managing an organization with an eye toward effectiveness can be a challenge, because business metrics are often tied to processes and other types of "discrete" pieces of work, and how quickly/efficiently they are completed. As a result, when an organization makes the shift to managing for effectiveness rather than efficiency, the metrics used to evaluate success typically have to be "leveled-up," that is, taken up to the level that really matters to the business. An example of this leveling up occurred several years back when customer service organizations changed their focus from shortening call times to increasing the rate of first call resolution. Resolving a customer issue on the first call may result in increasing the length of the call, but over the long term it is a more effective approach, because it may result in a shorter overall expenditure of the Customer Service Representatives' aggregated time, and will certainly result in more satisfied customers.

Operationalizing this "leveling-up" is not an easy task. Most of the greatest challenges associated with doing so relate to data. First, organizations must have an idea that their current efficiency-based metrics are not serving them well. The only way to know that your current practices are ill-serving you is to capture data to make that point. In the CSR example above, that means being able to find out that a customer has called multiple times. But the way that calls are typically handled, a case is created for each one, meaning that the data doesn't tell a story of a customer calling multiple times and taking the time of many different CSRs; instead, the data tells of ten individual calls, each of which lasted three minutes. The complexity of the problem is actually greater than this, because what happens more often than not in such cases is that a customer will try to resolve the problem by contacting the organization through multiple different channels - phone, Web, email, chat. Because the data is so often fragmented, organizations will typically find out about such broken practices through a series of irate letters and phone calls, or in the worst case scenario, in a drop-off in customers. Whatever the means of notification, at some point it becomes clear to the organization that they not only have a problem of misaligned incentives, but also a data problem. They then turn to the data to understand what has been going on in their organization and how to manage more effectively.

The story likely can be pieced together from the data, but the organization must still make sure they are asking the right questions - if "number of cold calls made" is not the right metric, what is? Once the right questions have been identified, then it's time to turn to the data. Because in most organizations the data to be captured was not set up with these higher-level goals in mind, getting the right answer from the data requires some work. The data across these various systems must be integrated and federated - all of the necessary data must be extracted from the various systems inside and out of the organization and loosely coupled so that the data is telling the whole story. It also requires cleansing the data and rationalizing it such that data about the same thing being captured in different systems is in sync.

It may be that even after having all of the data rationalized and accessible, the crucial data needed to manage the business more effectively is not currently being captured. This is a relatively small problem, with practically everything digitized and virtualized, there is very likely a way to capture the data an organization seeks. A common scenario is that the data is being captured, but in an off-premise cloud-based application or in a partner's application or it may be that the data is embedded in the activities carried out on social networks. In all of these cases, new technology makes the data accessible and manageable. As a result, so, too, are the answers to the real business questions of how to manage the business more effectively.

Data integration tools make it possible to integrate and federate data from cloud-based applications with on-premise systems, to incorporated data from third parties. The ability to use Hadoop MapReduce to take in and manage unprecedented volumes of data from social networks and other non-traditional sources makes it possible to truly have, manage and analyze all of your data. New social MDM technology means that you can tap into the data embedded in social interactions on social networks and use this to create an even more fully fleshed-out golden record for your customers.

In truth, it is the gains we have made in efficiency, in finding ever-more efficient ways to access, store and analyze data that make this turn towards effectiveness possible. Without being able to do all of the above in a time- and cost-efficient manner, it is not possible to use the data to manage more effectively.

In many ways, this is what the hype about Big Data is all about. The unarticulated and implicit excitement about Big Data is really about being able to take advantage of the data in which we are all awash and use it to manage our organizations more effectively than ever before. Managing for effectiveness looks different in every industry. In retail, managing for effectiveness is understanding customers - catering to them when, where, how and with what they want. In pharma, managing for effectiveness is limiting physician wash out, getting more clinical trial data more quickly, and being able to complete or pull the plug on trials faster based on the results of that data. In every industry, managing for effectiveness means using the power of data to make the best business decisions possible, getting a true return on data.

More Stories By Emily Burns

Emily Burns is responsible for Platform Product Marketing at Informatica. In that capacity, she has two principal roles. First, she evangelizes the benefits to be achieved from managing data as a key corporate asset, especially using the Informatica Platform. Second, she works to identify and communicate best practices and methodologies that demonstrate how to manage data as a corporate asset.

Prior to Informatica, Emily worked at Pegasystems and at TIBCO. While at Pegasystems, she led their case management product initiative. At TIBCO she was responsible for product marketing for the BPM suite. Emily holds a BS with majors in biochemistry and music, with an emphasis on piano performance. She is an avid reader, cook, and triathlete. Emily lives in Boston with her husband and two young sons.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


@MicroservicesExpo Stories
"We provide DevOps solutions. We also partner with some key players in the DevOps space and we use the technology that we partner with to engineer custom solutions for different organizations," stated Himanshu Chhetri, CTO of Addteq, in this SYS-CON.tv interview at DevOps at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
DevOps tends to focus on the relationship between Dev and Ops, putting an emphasis on the ops and application infrastructure. But that’s changing with microservices architectures. In her session at DevOps Summit, Lori MacVittie, Evangelist for F5 Networks, will focus on how microservices are changing the underlying architectures needed to scale, secure and deliver applications based on highly distributed (micro) services and why that means an expansion into “the network” for DevOps.
All organizations that did not originate this moment have a pre-existing culture as well as legacy technology and processes that can be more or less amenable to DevOps implementation. That organizational culture is influenced by the personalities and management styles of Executive Management, the wider culture in which the organization is situated, and the personalities of key team members at all levels of the organization. This culture and entrenched interests usually throw a wrench in the work...
@DevOpsSummit at Cloud taking place June 6-8, 2017, at Javits Center, New York City, is co-located with the 20th International Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world. The widespread success of cloud computing is driving the DevOps revolution in enterprise IT. Now as never before, development teams must communicate and collaborate in a dynamic, 24/7/365 environment. There is no time to wait for long developm...
"We're bringing out a new application monitoring system to the DevOps space. It manages large enterprise applications that are distributed throughout a node in many enterprises and we manage them as one collective," explained Kevin Barnes, President of eCube Systems, in this SYS-CON.tv interview at DevOps at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
I’m told that it has been 21 years since Scrum became public when Jeff Sutherland and I presented it at an Object-Oriented Programming, Systems, Languages & Applications (OOPSLA) workshop in Austin, TX, in October of 1995. Time sure does fly. Things mature. I’m still in the same building and at the same company where I first formulated Scrum.[1] Initially nobody knew of Scrum, yet it is now an open source body of knowledge translated into more than 30 languages[2] People use Scrum worldwide for ...
Thanks to Docker, it becomes very easy to leverage containers to build, ship, and run any Linux application on any kind of infrastructure. Docker is particularly helpful for microservice architectures because their successful implementation relies on a fast, efficient deployment mechanism – which is precisely one of the features of Docker. Microservice architectures are therefore becoming more popular, and are increasingly seen as an interesting option even for smaller projects, instead of being...
As the race for the presidency heats up, IT leaders would do well to recall the famous catchphrase from Bill Clinton’s successful 1992 campaign against George H. W. Bush: “It’s the economy, stupid.” That catchphrase is important, because IT economics are important. Especially when it comes to cloud. Application performance management (APM) for the cloud may turn out to be as much about those economics as it is about customer experience.
When you focus on a journey from up-close, you look at your own technical and cultural history and how you changed it for the benefit of the customer. This was our starting point: too many integration issues, 13 SWP days and very long cycles. It was evident that in this fast-paced industry we could no longer afford this reality. We needed something that would take us beyond reducing the development lifecycles, CI and Agile methodologies. We made a fundamental difference, even changed our culture...
The 20th International Cloud Expo has announced that its Call for Papers is open. Cloud Expo, to be held June 6-8, 2017, at the Javits Center in New York City, brings together Cloud Computing, Big Data, Internet of Things, DevOps, Containers, Microservices and WebRTC to one location. With cloud computing driving a higher percentage of enterprise IT budgets every year, it becomes increasingly important to plant your flag in this fast-expanding business opportunity. Submit your speaking proposal ...
@DevOpsSummit taking place June 6-8, 2017 at Javits Center, New York City, is co-located with the 20th International Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world. @DevOpsSummit at Cloud Expo New York Call for Papers is now open.
SYS-CON Events announced today that Dataloop.IO, an innovator in cloud IT-monitoring whose products help organizations save time and money, has been named “Bronze Sponsor” of SYS-CON's 20th International Cloud Expo®, which will take place on June 6-8, 2017, at the Javits Center in New York City, NY. Dataloop.IO is an emerging software company on the cutting edge of major IT-infrastructure trends including cloud computing and microservices. The company, founded in the UK but now based in San Fran...
Synthetic monitoring is hardly a new technology. It’s been around almost as long as the commercial World Wide Web has. But the importance of monitoring the performance and availability of a web application by simulating users’ interactions with that application, from around the globe, has never been more important. We’ve seen prominent vendors in the broad APM space add this technology with new development or partnerships just in the last 18 months.
Updating DevOps to the latest production data slows down your development cycle. Probably it is due to slow, inefficient conventional storage and associated copy data management practices. In his session at @DevOpsSummit at 20th Cloud Expo, Dhiraj Sehgal, in Product and Solution at Tintri, will talk about DevOps and cloud-focused storage to update hundreds of child VMs (different flavors) with updates from a master VM in minutes, saving hours or even days in each development cycle. He will also...
SYS-CON Events announced today that Linux Academy, the foremost online Linux and cloud training platform and community, will exhibit at SYS-CON's 20th International Cloud Expo®, which will take place on June 6-8, 2017, at the Javits Center in New York City, NY. Linux Academy was founded on the belief that providing high-quality, in-depth training should be available at an affordable price. Industry leaders in quality training, provided services, and student certification passes, its goal is to c...
The unique combination of Amazon Web Services and Cloud Raxak, a Gartner Cool Vendor in IT Automation, provides a seamless and cost-effective way of securely moving on-premise IT workloads to Amazon Web Services. Any enterprise can now leverage the cloud, manage risk, and maintain continuous security compliance. Forrester's analysis shows that enterprises need automated security to lower security risk and decrease IT operational costs. Through the seamless integration into Amazon Web Services, ...
A lot of time, resources and energy has been invested over the past few years on de-siloing development and operations. And with good reason. DevOps is enabling organizations to more aggressively increase their digital agility, while at the same time reducing digital costs and risks. But as 2017 approaches, the hottest trends in DevOps aren’t specifically about dev or ops. They’re about testing, security, and metrics.
Hardware virtualization and cloud computing allowed us to increase resource utilization and increase our flexibility to respond to business demand. Docker Containers are the next quantum leap - Are they?! Databases always represented an additional set of challenges unique to running workloads requiring a maximum of I/O, network, CPU resources combined with data locality.
"We got started as search consultants. On the services side of the business we have help organizations save time and save money when they hit issues that everyone more or less hits when their data grows," noted Otis Gospodnetić, Founder of Sematext, in this SYS-CON.tv interview at @DevOpsSummit, held June 9-11, 2015, at the Javits Center in New York City.
An overall theme of Cloud computing and the specific practices within it is fundamentally one of automation. The core value of technology is to continually automate low level procedures to free up people to work on more value add activities, ultimately leading to the utopian goal of full Autonomic Computing. For example a great way to define your plan for DevOps tool chain adoption is through this lens. In this TechTarget article they outline a simple maturity model for planning this.