Welcome!

Microservices Expo Authors: Stackify Blog, Aruna Ravichandran, Dalibor Siroky, Kevin Jackson, PagerDuty Blog

Related Topics: Microservices Expo, Java IoT, Industrial IoT

Microservices Expo: Article

SQL Peer-to-Peer Dynamic Structured Data Processing Collaboration

Using automatic metadata maintenance

Unstructured and XML semi-structured data is now used more than structured data. Unstructured data is useful because of its fuzzy processing applied to this more common ubiquitous data.  But fixed structured data still keeps businesses running day in and day out, which requires consistent predictable highly principled processing for correct results. This means structured data cannot be replaced by unstructured or semi-structured data.  For this reason, it would be very useful to have a general purpose peer-to-peer collaboration capability that can utilize highly principled hierarchical data processing and its flexible and advanced structured processing to support dynamically structured data and its dynamic structured processing.  This flexible dynamic structured processing can change the structure of the data as necessary for the required processing while preserving the relational and hierarchical data principles and semantics of the data to derive correct structured data results even after structure transformations.

This processing will perform freely across remote unrelated peer locations anytime and transparently support unpredictable structured data and data type changes automatically for immediate processing.  Such an automatic peer-to-peer dynamic structured data collaboration is depicted below with its dynamic working hierarchical data structure being modified at each peer site that are labeled P1 to P4 in the diagram directly below.  Its operation is described below the diagram.


Diagram Description

In the diagram above, peer locations: P1, P2, P3, and P4 located anywhere need to collaborate and share their structured data in order to produce a needed result. This process will require unpredictably changing the data structure and data types as it becomes necessary to achieve the desired need and result.  Peer 1 starts the collaboration process by inputting three relational tables, A, B, and C, and models them into a hierarchical structure sending it off to Peer 2 for further processing. Concurrently overlapping with peer 2's processing, Peer 1 also inputs a XML linear hierarchical structure, XYZ, and transforms it into a nonlinear multipath hierarchical structure sending it off to Peer 3 for further processing.

Peer 2 and peer 3 are now performing independently and concurrently. Each is: retrieving their structure input from peer 1, inputting additional relational table data from their different peer home locations, and joining this data to their working data structures. On completion, peer 2 and peer 3 both send their modified data structures off to common peer 4 for further processing.

Peer 4 accepts the modified data structures from both peer 2 and peer 3 which operated concurrently. It hierarchically joins them together using a matching data item value between nodes B and X (B.b=X.x). Peer 4 then eliminates unneeded data items from the joined result using SQL's dynamic SELECT operation to select data items for output  from nodes A, B, E, Y and W.  This SQL query looks like: SELECT A.a, B.b, E.e, Y.y, W.w FROM P2View LEFT JOIN P3View ON B.b=X.x.  This slices out all nodes (C, D, Z ,X, V) that were not referenced by the SELECT statement. This automatically aggregates the necessary data nicely as shown in the diagram above. This process is known as projection in relational processing and node promotion in hierarchical processing. The LEFT JOIN operation hierarchically places P2's structure over P3's structure connected by the ON clause specification of: B.b=X.x. This newly combined hierarchical structure in peer 4 is sent back to Peer 1 for immediate review and processing where the hierarchical data can be selectively output in different formats each with different data selections as shown in the above diagram.

During this entire peer-to-peer collaboration process, the changing data structures and data types are automatically maintained and utilized transparently for the user as needed. The user at each receiving peer can also view the current active structure and its data types. But knowledge of the structure is not necessary for the user to specify in the query because the maintained structure is automatically known and used inherently by the query processor.  Different working data structure versions can also be saved and restored at each peer by the user.

Integrating SQL With Peer-to-Peer Structured Data Collaboration

The problem with performing the above type of dynamic processing is that structured data processing has been limited to fixed static structure processing because dynamically generated structured data cannot be handled today.  Sharing structured data today is performed with shared metadata.  The metadata remains the same, so the structure must remain static. But with dynamic structured processing, the data structure can be dynamically modified as needed to support the required structured operation as shown and described above. This requires automatic metadata maintenance which has not previously been supported by the industry for structured data processing.

An advanced ANSI SQL transparent hierarchical processor prototype, SQLfX (www.adatinc.com/demo.html), has been developed that can support the required dynamic and flexible structured data processing necessary for collaboration. In addition, it uses SQL's inherent hierarchical data processing capabilities that naturally support full multipath dynamic hierarchical data structures. This allows the most complex hierarchical operations to be performed in order to always meet the need at the required time and peer locations with the SQLfX processor. It already operates structure-aware because of its dynamic processing which is also necessary for the required automatic metadata maintenance to occur automatically at each peer because of the dynamically changing data structure and data items.

The ANSI SQL SQLfX flexible dynamic hierarchical processing technology can be enhanced to integrate with peer-to-peer structured data processing collaboration that eliminates the user control necessary for the dynamically changing metadata. The automatic metadata maintenance supplies the updated current metadata that accompanies the data when transmitted between peers. This allows amazingly fast on the fly advanced hierarchical structured data processing collaboration. This enables previously unknown structure results delivered to any peer to be immediate processed automatically by a SQLfX SQL processor located at the peer location.

SQLfX SQL controls the peer-to-peer processing sending and receiving of data structures using new SQL InFile and OutFile keywords added by SQLfX for this purpose. Password and data encryption can also be supported for data security. Further dynamic processing and data structure modification can be performed at each peer visited in any order including in parallel as shown in the diagram above. This opens up the new capability of dynamic structured data processing and its automatic and transparent metadata handling.

SQL Hierarchical Processing Capabilities for Structured Data Collaboration
SQLfX is a powerful new ANSI SQL transparent multipath hierarchical processor that dynamically processes heterogeneous logical flat data like relational and physical hierarchical data such as XML initially. This SQLfX full dynamic hierarchical data processing enables logical and physical structures to be hierarchically joined and modeled dynamically. This hierarchical processing significantly increases the power of the data structure and the queries applied to it.  This is extremely flexible and powerful and is automatically performed without user hierarchical navigation. This operation naturally utilizes the hierarchical semantic information between the pathways to process powerful multipath queries. This can freely reference multipath queries that can for example select data from one path based on data in another path. This unlimited dynamic processing requires special automatic processing known as Lowest Common Ancestor (LCA) processing which enables any conceivable valid multipath query to be processed automatically.  This capability is supported in SQLfX naturally and is missing in other data processors such as XQuery.

An additional valuable benefit of using hierarchical structures is that they are great at naturally organizing and reusing data. Their ability to freely create and grow logical hierarchical multipath structures dynamically also has another overlooked powerful benefit.  It continually increases the data value of the data nonlinearly through automatic data reuse and sharing of the data at higher levels with the multiple lower levels in a pyramid fashion. In addition, the dynamic joining of these hierarchical structures can dynamically increase their data value and querying power many times.  Another powerful advantage are logical hierarchical data structures that are assembled on the fly when creating new structures such as when structures are joined and exist only when and while they are being used. These logical structures add flexibility to hierarchical structures and efficiency to their new use.

Hierarchical structures can also be hierarchically data filtered in their entirety following hierarchical semantics using SQL's WHERE clause to filter the data by data value to only the precise desired data result. This is a complex and powerful operation because data filtering applied to any node data item in a hierarchical structure affects all other nodes of the structure. This is because every node in a hierarchical structure is related to every other node in the data structure. This is demonstrated in Diagram 2 below where filtering node E winds up affecting all other nodes sometimes indirectly through a cousin relationship such as node B. In this example, all nodes with a data occurrence related to data item E equal to 25 are filtered out. The data filtering flow is represented by the arrows.  This is a powerful concept giving multipath processing and WHERE clause hierarchical data processing significant power. This global hierarchical structure filtering is particularly useful when combined with transferring these powerful multipath structures between peers in peer-to-peer processing and the entire structure needs to be filtered for some data value condition. This can be a complex condition involving multiple paths.

SQLfX SQL also supports a very advanced dynamic any-to-any data structure transformation and a data structure virtualization capability. This allows all hierarchically data transformations to be performed semantically correct at a high SQL processing level. With multipath hierarchical processing and any-to-any structure transformations, a variety of hypothetical, experimental, research, exploratory, and problem solving queries can be carried out immediately in an unrestricted fashion further enhanced by powerful real time hierarchical processing collaboration.

Conclusion

All of the powerful and flexible capabilities mentioned in this article make multipath hierarchical structures and their hierarchical processing the perfect opportunity for this dynamic structured data processing collaboration. The universally known SQL interface makes it a perfect API and this is backed by this new relational hierarchical processing technology.  Single one-way data transmissions will also always be available to send to anyone any time because a receive-only version of SQLfX peer-to-peer will be freely available to download and use to automatically view and utilize the one-way transmitted data structure. Additional information on SQLfX's advanced hierarchical processing capabilities and operation can be found at www.adatinc.com. Persons and Companies wanting more information or help on SQL  peer-to-peer dynamic structured data processing collaboration can contact [email protected].

More Stories By Michael M David

Michael M. David is founder and CTO of Advanced Data Access Technologies, Inc. He has been a staff scientist and lead XML architect for NCR/Teradata and their representative to the SQLX Group. He has researched, designed and developed commercial query languages for heterogeneous hierarchical and relational databases for over twenty years. He has authored the book "Advanced ANSI SQL Data Modeling and Structure Processing" Published by Artech House Publishers and many papers and articles on database topics. His research and findings have shown that Hierarchical Data Processing is a subset of Relational Processing and how to utilize this advanced inherent capability in ANSI SQL. Additionally, his research has shown that advanced multipath (LCA) processing is also naturally supported and performed automatically in ANSI SQL, and advanced hierarchical processing operations are also possible. These advanced capabilities can be performed and explained in the ANSI SQL Transparent XML Hierarchical Processor at his site at: www.adatinc.com/demo.html.

@MicroservicesExpo Stories
How is DevOps going within your organization? If you need some help measuring just how well it is going, we have prepared a list of some key DevOps metrics to track. These metrics can help you understand how your team is doing over time. The word DevOps means different things to different people. Some say it a culture and every vendor in the industry claims that their tools help with DevOps. Depending on how you define DevOps, some of these metrics may matter more or less to you and your team.
For many of us laboring in the fields of digital transformation, 2017 was a year of high-intensity work and high-reward achievement. So we’re looking forward to a little breather over the end-of-year holiday season. But we’re going to have to get right back on the Continuous Delivery bullet train in 2018. Markets move too fast and customer expectations elevate too precipitously for businesses to rest on their laurels. Here’s a DevOps “to-do list” for 2018 that should be priorities for anyone w...
If testing environments are constantly unavailable and affected by outages, release timelines will be affected. You can use three metrics to measure stability events for specific environments and plan around events that will affect your critical path to release.
In a recent post, titled “10 Surprising Facts About Cloud Computing and What It Really Is”, Zac Johnson highlighted some interesting facts about cloud computing in the SMB marketplace: Cloud Computing is up to 40 times more cost-effective for an SMB, compared to running its own IT system. 94% of SMBs have experienced security benefits in the cloud that they didn’t have with their on-premises service
DevOps failure is a touchy subject with some, because DevOps is typically perceived as a way to avoid failure. As a result, when you fail in a DevOps practice, the situation can seem almost hopeless. However, just as a fail-fast business approach, or the “fail and adjust sooner” methodology of Agile often proves, DevOps failures are actually a step in the right direction. They’re the first step toward learning from failures and turning your DevOps practice into one that will lead you toward even...
DevOps is under attack because developers don’t want to mess with infrastructure. They will happily own their code into production, but want to use platforms instead of raw automation. That’s changing the landscape that we understand as DevOps with both architecture concepts (CloudNative) and process redefinition (SRE). Rob Hirschfeld’s recent work in Kubernetes operations has led to the conclusion that containers and related platforms have changed the way we should be thinking about DevOps and...
While walking around the office I happened upon a relatively new employee dragging emails from his inbox into folders. I asked why and was told, “I’m just answering emails and getting stuff off my desk.” An empty inbox may be emotionally satisfying to look at, but in practice, you should never do it. Here’s why. I recently wrote a piece arguing that from a mathematical perspective, Messy Desks Are Perfectly Optimized. While it validated the genius of my friends with messy desks, it also gener...
The goal of Microservices is to improve software delivery speed and increase system safety as scale increases. Microservices being modular these are faster to change and enables an evolutionary architecture where systems can change, as the business needs change. Microservices can scale elastically and by being service oriented can enable APIs natively. Microservices also reduce implementation and release cycle time and enables continuous delivery. This paper provides a logical overview of the Mi...
The next XaaS is CICDaaS. Why? Because CICD saves developers a huge amount of time. CD is an especially great option for projects that require multiple and frequent contributions to be integrated. But… securing CICD best practices is an emerging, essential, yet little understood practice for DevOps teams and their Cloud Service Providers. The only way to get CICD to work in a highly secure environment takes collaboration, patience and persistence. Building CICD in the cloud requires rigorous ar...
The enterprise data storage marketplace is poised to become a battlefield. No longer the quiet backwater of cloud computing services, the focus of this global transition is now going from compute to storage. An overview of recent storage market history is needed to understand why this transition is important. Before 2007 and the birth of the cloud computing market we are witnessing today, the on-premise model hosted in large local data centers dominated enterprise storage. Key marketplace play...
The cloud revolution in enterprises has very clearly crossed the phase of proof-of-concepts into a truly mainstream adoption. One of most popular enterprise-wide initiatives currently going on are “cloud migration” programs of some kind or another. Finding business value for these programs is not hard to fathom – they include hyperelasticity in infrastructure consumption, subscription based models, and agility derived from rapid speed of deployment of applications. These factors will continue to...
Some people are directors, managers, and administrators. Others are disrupters. Eddie Webb (@edwardawebb) is an IT Disrupter for Software Development Platforms at Liberty Mutual and was a presenter at the 2016 All Day DevOps conference. His talk, Organically DevOps: Building Quality and Security into the Software Supply Chain at Liberty Mutual, looked at Liberty Mutual's transformation to Continuous Integration, Continuous Delivery, and DevOps. For a large, heavily regulated industry, this task ...
Following a tradition dating back to 2002 at ZapThink and continuing at Intellyx since 2014, it’s time for Intellyx’s annual predictions for the coming year. If you’re a long-time fan, you know we have a twist to the typical annual prediction post: we actually critique our predictions from the previous year. To make things even more interesting, Charlie and I switch off, judging the other’s predictions. And now that he’s been with Intellyx for more than a year, this Cortex represents my first ...
"Grape Up leverages Cloud Native technologies and helps companies build software using microservices, and work the DevOps agile way. We've been doing digital innovation for the last 12 years," explained Daniel Heckman, of Grape Up in this SYS-CON.tv interview at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.
The Toyota Production System, a world-renowned production system is based on the "complete elimination of all waste". The "Toyota Way", grounded on continuous improvement dates to the 1860s. The methodology is widely proven to be successful yet there are still industries within and tangential to manufacturing struggling to adopt its core principles: Jidoka: a process should stop when an issue is identified prevents releasing defective products
We seem to run this cycle with every new technology that comes along. A good idea with practical applications is born, then both marketers and over-excited users start to declare it is the solution for all or our problems. Compliments of Gartner, we know it generally as “The Hype Cycle”, but each iteration is a little different. 2018’s flavor will be serverless computing, and by 2018, I mean starting now, but going most of next year, you’ll be sick of it. We are already seeing people write such...
Defining the term ‘monitoring’ is a difficult task considering the performance space has evolved significantly over the years. Lately, there has been a shift in the monitoring world, sparking a healthy debate regarding the definition and purpose of monitoring, through which a new term has emerged: observability. Some of that debate can be found in blogs by Charity Majors and Cindy Sridharan.
It’s “time to move on from DevOps and continuous delivery.” This was the provocative title of a recent article in ZDNet, in which Kelsey Hightower, staff developer advocate at Google Cloud Platform, suggested that “software shops should have put these concepts into action years ago.” Reading articles like this or listening to talks at most DevOps conferences might make you think that we’re entering a post-DevOps world. But vast numbers of organizations still struggle to start and drive transfo...
Let's do a visualization exercise. Imagine it's December 31, 2018, and you're ringing in the New Year with your friends and family. You think back on everything that you accomplished in the last year: your company's revenue is through the roof thanks to the success of your product, and you were promoted to Lead Developer. 2019 is poised to be an even bigger year for your company because you have the tools and insight to scale as quickly as demand requires. You're a happy human, and it's not just...
"Opsani helps the enterprise adopt containers, help them move their infrastructure into this modern world of DevOps, accelerate the delivery of new features into production, and really get them going on the container path," explained Ross Schibler, CEO of Opsani, and Peter Nickolov, CTO of Opsani, in this SYS-CON.tv interview at DevOps Summit at 21st Cloud Expo, held Oct 31 – Nov 2, 2017, at the Santa Clara Convention Center in Santa Clara, CA.