|By Adrian Bridgwater||
|November 23, 2012 10:00 AM EST||
Big Data is everywhere. Predictive analytics and real time in-memory computing isn't everywhere.
This truth (if we can accept it to be so) represents something of an imbalance.
As a subset of data mining, predictive analytics driven by in-memory computing efficiencies now has an opportunity to bring real-time analysis and insight to fast-moving live transactional data flows. Or to put it another (rather shorter) way, we can now start to manage and understand Big Data better than ever.
Application use cases here might typically include:
- Climate simulation
- Oil exploration
- Financial analysis and scientific research
- Telecommunications, to name but seven good examples
If we combine contemporary approaches to predictive analytics with the newly arrived Intel Xeon Phi coprocessor that produces what is claimed to be over one teraflop per second in terms of workload computational power for highly-parallel workloads, then CIOs can start to think about what "50-processor core computing" will mean for us in the very near future.
Pushing forward in this space is SAP with its HANA in-memory computing appliance and platform. The firm is openly partnering with HP, Oracle, Cognizant and variety of other big (and smaller specialist) players to form strategic alliances that will help further the uptake of this kind of technology.
Why Is Analytics Working Better?
Part of the reason that in-memory predictive analytics intelligence is now becoming so important and much easier to bring to bear is hardware related and part of it is a software-related development.
On the hardware end, Intel has worked to sharpen data throughput between memory and processor cores. This means that work that goes on right in the heart of the machine that might have traveled at around five Gigabytes per second (Gbps) five years ago, now has a chance to move at 100 Gbps. On the software end, firms like HP and SAP have been working to produce what are typically referred to as "business process solutions" that can produce "context-aware experiences" to enable one sense-and-respond scenarios and, therefore, faster and more personalized interactions with customers.
As a matter of interest, HP also works to provide datacenter services to SAP that support its enterprise-wide hosting solutions for e-business applications - but that's another story.
In terms of actual application solutions running on SAP HANA, new products include the SAP Liquidity Risk Management application, the SAP Accelerated Trade Promotion Planning application and SAP Operational Process Intelligence software.
"These innovations show how SAP is rapidly delivering real-time, data-centric and industry-specific applications on the SAP HANA platform," said Dr Vishal Sikka, member of the SAP executive board for technology and innovation.
To take one example, SAP Liquidity Risk Management aims to provide banks with the ability to perform real-time, high-speed liquidity risk management and reporting on very large volumes of cash flows. In future then, banks will be able to instantly measure key liquidity risk ratios (such as the Basel III liquidity coverage ratio) and cash flow gaps to resolve potential liquidity bottlenecks. The application aims to allow banks to apply different stress scenarios, such as adjusted run-off rates and bond haircuts, to gain a deeper understanding of how market volatility can impact liquidity positions.
Lessons for CIOs
Now you don't have to be a bank CIO or financial analyst to understand the wider importance of this technology, i.e., this is the "harnessing of Big Data" catchline that you've already heard bandied about by countless IT firm's press departments, except now it's really happening.
The lesson for CIOs and the software application developers serving them is that we now have a route to predictive real-time analysis and the power to view billions of stored records and live transactional data at the same time. CIOs should look to their solution architects and business process experts to compose the analytical data models that will drive the next phase of their technology growth.
SAP is not the only firm driving this space, but its work to publicize its competencies in this arena is very prevalent. The fact that SAP pushes many of the interfaces for managing the result of its data analysis to both Apple iPad (and now Windows 8 format) mobile devices may have confounded Steve Jobs at the time, but this is the firm's proof point for showing off its Big Data number crunching applications. One day, quite soon, none of this will be a surprise.
This post first appeared on Enterprise CIO Forum.
In many organizations governance is still practiced by phase or stage gate peer review, and Agile projects are forced to accommodate, which leads to WaterScrumFall or worse. But governance criteria and policies are often very weak anyway, out of date or non-existent. Consequently governance is frequently a matter of opinion and experience, highly dependent upon the experience of individual reviewers. As we all know, a basic principle of Agile methods is delegation of responsibility, and ideally ...
Oct. 27, 2016 06:00 PM EDT Reads: 3,665
Monitoring of Docker environments is challenging. Why? Because each container typically runs a single process, has its own environment, utilizes virtual networks, or has various methods of managing storage. Traditional monitoring solutions take metrics from each server and applications they run. These servers and applications running on them are typically very static, with very long uptimes. Docker deployments are different: a set of containers may run many applications, all sharing the resource...
Oct. 27, 2016 03:00 PM EDT Reads: 3,808
When we talk about the impact of BYOD and BYOA and the Internet of Things, we often focus on the impact on data center architectures. That's because there will be an increasing need for authentication, for access control, for security, for application delivery as the number of potential endpoints (clients, devices, things) increases. That means scale in the data center. What we gloss over, what we skip, is that before any of these "things" ever makes a request to access an application it had to...
Oct. 27, 2016 01:15 PM EDT Reads: 13,994
Virgil consists of an open-source encryption library, which implements Cryptographic Message Syntax (CMS) and Elliptic Curve Integrated Encryption Scheme (ECIES) (including RSA schema), a Key Management API, and a cloud-based Key Management Service (Virgil Keys). The Virgil Keys Service consists of a public key service and a private key escrow service.
Oct. 27, 2016 01:15 PM EDT Reads: 1,209
The best way to leverage your Cloud Expo presence as a sponsor and exhibitor is to plan your news announcements around our events. The press covering Cloud Expo and @ThingsExpo will have access to these releases and will amplify your news announcements. More than two dozen Cloud companies either set deals at our shows or have announced their mergers and acquisitions at Cloud Expo. Product announcements during our show provide your company with the most reach through our targeted audiences.
Oct. 27, 2016 01:15 PM EDT Reads: 5,089
In his general session at 19th Cloud Expo, Manish Dixit, VP of Product and Engineering at Dice, will discuss how Dice leverages data insights and tools to help both tech professionals and recruiters better understand how skills relate to each other and which skills are in high demand using interactive visualizations and salary indicator tools to maximize earning potential. Manish Dixit is VP of Product and Engineering at Dice. As the leader of the Product, Engineering and Data Sciences team a...
Oct. 27, 2016 01:15 PM EDT Reads: 854
SYS-CON Events announced today that eCube Systems, the leading provider of modern development tools and best practices for Continuous Integration on OpenVMS, will exhibit at SYS-CON's @DevOpsSummit at Cloud Expo New York, which will take place on June 7-9, 2016, at the Javits Center in New York City, NY. eCube Systems offers a family of middleware products and development tools that maximize return on technology investment by leveraging existing technical equity to meet evolving business needs. ...
Oct. 27, 2016 01:00 PM EDT Reads: 4,653
Join Impiger for their featured webinar: ‘Cloud Computing: A Roadmap to Modern Software Delivery’ on November 10, 2016, at 12:00 pm CST. Very few companies have not experienced some impact to their IT delivery due to the evolution of cloud computing. This webinar is not about deciding whether you should entertain moving some or all of your IT to the cloud, but rather, a detailed look under the hood to help IT professionals understand how cloud adoption has evolved and what trends will impact th...
Oct. 27, 2016 12:00 PM EDT Reads: 722
As we enter the final week before the 19th International Cloud Expo | @ThingsExpo in Santa Clara, CA, it's time for me to reflect on six big topics that will be important during the show. Hybrid Cloud This general-purpose term seems to provide a comfort zone for many enterprise IT managers. It sounds reassuring to be able to work with one of the major public-cloud providers like AWS or Microsoft Azure while still maintaining an on-site presence.
Oct. 27, 2016 11:45 AM EDT Reads: 1,186
operations aren’t merging to become one discipline. Nor is operations simply going away. Rather, DevOps is leading software development and operations – together with other practices such as security – to collaborate and coexist with less overhead and conflict than in the past. In his session at @DevOpsSummit at 19th Cloud Expo, Gordon Haff, Red Hat Technology Evangelist, will discuss what modern operational practices look like in a world in which applications are more loosely coupled, are deve...
Oct. 27, 2016 11:45 AM EDT Reads: 2,064
All clouds are not equal. To succeed in a DevOps context, organizations should plan to develop/deploy apps across a choice of on-premise and public clouds simultaneously depending on the business needs. This is where the concept of the Lean Cloud comes in - resting on the idea that you often need to relocate your app modules over their life cycles for both innovation and operational efficiency in the cloud. In his session at @DevOpsSummit at19th Cloud Expo, Valentin (Val) Bercovici, CTO of So...
Oct. 27, 2016 11:15 AM EDT Reads: 2,246
DevOps is a term that comes full of controversy. A lot of people are on the bandwagon, while others are waiting for the term to jump the shark, and eventually go back to business as usual. Regardless of where you are along the specturm of loving or hating the term DevOps, one thing is certain. More and more people are using it to describe a system administrator who uses scripts, or tools like, Chef, Puppet or Ansible, in order to provision infrastructure. There is also usually an expectation of...
Oct. 27, 2016 10:30 AM EDT Reads: 1,775
DevOps is speeding towards the IT world like a freight train and the hype around it is deafening. There is no reason to be afraid of this change as it is the natural reaction to the agile movement that revolutionized development just a few years ago. By definition, DevOps is the natural alignment of IT performance to business profitability. The relevance of this has yet to be quantified but it has been suggested that the route to the CEO’s chair will come from the IT leaders that successfully ma...
Oct. 27, 2016 10:30 AM EDT Reads: 16,646
DevOps is being widely accepted (if not fully adopted) as essential in enterprise IT. But as Enterprise DevOps gains maturity, expands scope, and increases velocity, the need for data-driven decisions across teams becomes more acute. DevOps teams in any modern business must wrangle the ‘digital exhaust’ from the delivery toolchain, "pervasive" and "cognitive" computing, APIs and services, mobile devices and applications, the Internet of Things, and now even blockchain. In this power panel at @...
Oct. 27, 2016 09:15 AM EDT Reads: 2,175
As software becomes more and more complex, we, as software developers, have been splitting up our code into smaller and smaller components. This is also true for the environment in which we run our code: going from bare metal, to VMs to the modern-day Cloud Native world of containers, schedulers and microservices. While we have figured out how to run containerized applications in the cloud using schedulers, we've yet to come up with a good solution to bridge the gap between getting your conta...
Oct. 27, 2016 07:30 AM EDT Reads: 1,629
19th Cloud Expo, taking place November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA, will feature technical sessions from a rock star conference faculty and the leading industry players in the world. Cloud computing is now being embraced by a majority of enterprises of all sizes. Yesterday's debate about public vs. private has transformed into the reality of hybrid cloud: a recent survey shows that 74% of enterprises have a hybrid cloud strategy. Meanwhile, 94% of enterpri...
Oct. 27, 2016 07:00 AM EDT Reads: 5,065
This is a no-hype, pragmatic post about why I think you should consider architecting your next project the way SOA and/or microservices suggest. No matter if it’s a greenfield approach or if you’re in dire need of refactoring. Please note: considering still keeps open the option of not taking that approach. After reading this, you will have a better idea about whether building multiple small components instead of a single, large component makes sense for your project. This post assumes that you...
Oct. 27, 2016 06:00 AM EDT Reads: 7,317
Without lifecycle traceability and visibility across the tool chain, stakeholders from Planning-to-Ops have limited insight and answers to who, what, when, why and how across the DevOps lifecycle. This impacts the ability to deliver high quality software at the needed velocity to drive positive business outcomes. In his general session at @DevOpsSummit at 19th Cloud Expo, Eric Robertson, General Manager at CollabNet, will discuss how customers are able to achieve a level of transparency that e...
Oct. 27, 2016 06:00 AM EDT Reads: 1,054
DevOps theory promotes a culture of continuous improvement built on collaboration, empowerment, systems thinking, and feedback loops. But how do you collaborate effectively across the traditional silos? How can you make decisions without system-wide visibility? How can you see the whole system when it is spread across teams and locations? How do you close feedback loops across teams and activities delivering complex multi-tier, cloud, container, serverless, and/or API-based services?
Oct. 27, 2016 05:15 AM EDT Reads: 1,167
Today every business relies on software to drive the innovation necessary for a competitive edge in the Application Economy. This is why collaboration between development and operations, or DevOps, has become IT’s number one priority. Whether you are in Dev or Ops, understanding how to implement a DevOps strategy can deliver faster development cycles, improved software quality, reduced deployment times and overall better experiences for your customers.
Oct. 27, 2016 05:15 AM EDT Reads: 1,031