Welcome!

Microservices Expo Authors: Liz McMillan, Elizabeth White, Yeshim Deniz, Pat Romanski, Zakia Bouachraoui

Related Topics: @CloudExpo

@CloudExpo: Article

Big Data – The State of Affairs

Big Data is here to stay, but do we have the tools to efficiently process it?

Many products are available as open source or proprietary products that can handle Big Data. Which one is best fit for this task?

Today's classic RDBMSs and tools are able to quickly load the data, process it and present results in an easy to understand format.  You can use SQL or programmatic interface to process the data randomly or in batch; RDBMS's keep data safe, protected against hardware and software failures.

Standards tools and products are not able to cope with Big Data requirement, which is not dissimilar to  what is involved in processing today's regular data sets, just on a much bigger scale. Mainstream companies like telcos, financials, web companies as well as government are reaching the limit of what  can be efficiently processed by classic RDBMS techhnologies.

When it comes to picking a proper platform and tools to handle your Big Data there are a couple of possible choices:

  • Oracle Exadata - it doesn't fit economical mandate; Exadata's weak link and bottleneck is its reliance on classic Oracle RDBMS
  • NoSQL databases -  too immature, they offer no SQL or similar random access query language ( you are presently forced to write  programs to access your data ); often achieve scale-out by not implementing all elements of ACID, CAP
  • Hadoop/MapReduce and related open source ecosystem ( Pig, Hive, HBase ) -  useful for cheap data storage on commodity hardware and batch processing; they offer no efficient, non-programmatic random access
  • proprietary MPP databases running on commodity hardware ( Vertica, Aster Data, Greenplum )  - very fast and can provide random, SQL  access to big data; their management features and general feature sets are immature
  • proprietary MPP databases running on specialized hardware ( Teradata ) - fairly expensive ( don't run on commodity hardware )
  • new platforms that will or are trying to emulate Google Percolator, Dremel  ( latest Google technologies dealing with big data ACID compliant transactions and reporting ), similarly to how Hadoop originated from  Google GFS and MapReduce.

We would say that there is no single, generic product or platform available today that can handle this task. Depending on your needs you have to deploy  and combinne quite a few of technologies to bring you closer to achieving end-to-end efficient, comprehensive processing of Big Data. You will quite likely have to custom build solutions that will fit your particular needs as off-the-shelf solutions are still immature, incomplete or not available.

Big Data is an area of growth and innovation, so current picture is bound to change as new products and technologies appear, bringing us closer to the ultimate goal of routine, efficient processing of Big Data.

More Stories By Ranko Mosic

Ranko Mosic, BScEng, is specializing in Big Data/Data Architecture consulting services ( database/data architecture, machine learning ). His clients are in finance, retail, telecommunications industries. Ranko is welcoming inquiries about his availability for consulting engagements and can be reached at 408-757-0053 or [email protected]

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


Microservices Articles
Most DevOps journeys involve several phases of maturity. Research shows that the inflection point where organizations begin to see maximum value is when they implement tight integration deploying their code to their infrastructure. Success at this level is the last barrier to at-will deployment. Storage, for instance, is more capable than where we read and write data. In his session at @DevOpsSummit at 20th Cloud Expo, Josh Atwell, a Developer Advocate for NetApp, will discuss the role and value...
SYS-CON Events announced today that DatacenterDynamics has been named “Media Sponsor” of SYS-CON's 18th International Cloud Expo, which will take place on June 7–9, 2016, at the Javits Center in New York City, NY. DatacenterDynamics is a brand of DCD Group, a global B2B media and publishing company that develops products to help senior professionals in the world's most ICT dependent organizations make risk-based infrastructure and capacity decisions.
CloudEXPO New York 2018, colocated with DXWorldEXPO New York 2018 will be held November 11-13, 2018, in New York City and will bring together Cloud Computing, FinTech and Blockchain, Digital Transformation, Big Data, Internet of Things, DevOps, AI, Machine Learning and WebRTC to one location.
DevOpsSummit New York 2018, colocated with CloudEXPO | DXWorldEXPO New York 2018 will be held November 11-13, 2018, in New York City. Digital Transformation (DX) is a major focus with the introduction of DXWorldEXPO within the program. Successful transformation requires a laser focus on being data-driven and on using all the tools available that enable transformation if they plan to survive over the long term.
Consumer-driven contracts are an essential part of a mature microservice testing portfolio enabling independent service deployments. In this presentation we'll provide an overview of the tools, patterns and pain points we've seen when implementing contract testing in large development organizations.
Containers and Kubernetes allow for code portability across on-premise VMs, bare metal, or multiple cloud provider environments. Yet, despite this portability promise, developers may include configuration and application definitions that constrain or even eliminate application portability. In this session we'll describe best practices for "configuration as code" in a Kubernetes environment. We will demonstrate how a properly constructed containerized app can be deployed to both Amazon and Azure ...
In his general session at 19th Cloud Expo, Manish Dixit, VP of Product and Engineering at Dice, discussed how Dice leverages data insights and tools to help both tech professionals and recruiters better understand how skills relate to each other and which skills are in high demand using interactive visualizations and salary indicator tools to maximize earning potential. Manish Dixit is VP of Product and Engineering at Dice. As the leader of the Product, Engineering and Data Sciences team at D...
Modern software design has fundamentally changed how we manage applications, causing many to turn to containers as the new virtual machine for resource management. As container adoption grows beyond stateless applications to stateful workloads, the need for persistent storage is foundational - something customers routinely cite as a top pain point. In his session at @DevOpsSummit at 21st Cloud Expo, Bill Borsari, Head of Systems Engineering at Datera, explored how organizations can reap the bene...
In his session at 20th Cloud Expo, Scott Davis, CTO of Embotics, discussed how automation can provide the dynamic management required to cost-effectively deliver microservices and container solutions at scale. He also discussed how flexible automation is the key to effectively bridging and seamlessly coordinating both IT and developer needs for component orchestration across disparate clouds – an increasingly important requirement at today’s multi-cloud enterprise.
"We do one of the best file systems in the world. We learned how to deal with Big Data many years ago and we implemented this knowledge into our software," explained Jakub Ratajczak, Business Development Manager at MooseFS, in this SYS-CON.tv interview at 20th Cloud Expo, held June 6-8, 2017, at the Javits Center in New York City, NY.