skip to Main Content

Data Fabric

Last Updated:
Analyst Coverage:

Global organisations have long struggled to manage their data assets. The average company has hundreds of applications (different surveys show different figures, but they are all in the hundreds) scattered across on-premise data centres, public clouds like Azure or AWS, private clouds and beyond. Over the years many attempts have been made to wrangle the data herd. Enterprises have tried data warehouses (copy what you need into a central place for analytics), master data management (manage the key data that crosses those apps, such as customer, product and location) and enterprise resource planning (try and replace all those apps with a giant app from a single vendor). It is fair to say that, with the number of enterprise applications steadily increasing in number despite massive investment in these initiatives, none has proved entirely and universally successful. In 2023 just a third of executives trust their data, roughly the same as the level of trust as some years ago, according to several independent surveys.

A data fabric is a data architecture and infrastructure that attempts to organise and supervise the entire enterprise data resource, providing all systems and applications with timely access to all the data and services they may need. This spans both structured and unstructured data and may utilise existing metadata. Incidentally, it is different from a “data mesh”, which is a mostly organisational approach to data management, focusing on key data domains through decentralised data ownership.

If you have managed to deploy a working data fabric solution then you can potentially see many benefits. A commercial enterprise could gain a unified view of its customers, for example, across all the various enterprise touchpoints that exist. Hospitals could gain a single, coherent view of patients, encompassing patient records, treatment history and more. Financial institutions would get a better assessment of trading risk, achieving a better knowledge of who the true counterparties are in each trade.

This would allow all kinds of useful business analytics that have proven elusive to deliver in the past, such as knowing who your most profitable products and customers are, a simple notion that is tough to actually deliver. In principle, an all-encompassing data fabric would grow with your business, adapting to new or changed data sources and business models. If it works as promised then data quality, another thorny data problem that has remained stubbornly hard to eradicate, should also improve.

There are significant challenges in making such an architecture work in practice, especially at enterprise scale. Executing complex queries in real-time against a large single database takes considerable resources. To execute such queries against multiple distributed databases, with data scattered amongst them, is a much tougher problem, requiring a highly efficient distributed query capability. Few currently deployed mainstream databases can seriously claim this ability, though some specialist niche databases are designed with distributed query in mind. Even if that problem is ignored, there is the issue of data quality, which is something that continues to plague operational systems despite decades of attempts to address it with assorted data quality software and data governance processes. The level of trust of executives in their data remains stubbornly low even in 2023 surveys, so just being able to execute a query is one thing; being confident that the answer is correct is another. Previous approaches such as data warehouses and master data management put considerable emphasis on data quality (around a third of a typical master data management project is spent resolving data quality issues). A data fabric spurns the movement of data into such hubs, preferring to resolve queries dynamically, but it is unclear how the very real issue with underlying data quality can be wished away.  Certainly, any data fabric initiative needs to be able to explain how it will successfully deal with such challenges. If you spend some time, as I have, searching for data fabric case studies, then you will find plenty of vendor claims but very few that have named end user companies with customer testimonials and quantified business benefits. Although data fabric is not a new idea, it would appear to still be in the early adopter phase based on the dearth of properly documented case studies. There are plenty that are anonymous, mostly promoted by data fabric vendors and consultants, who clearly have a vested interest in making such claims.

Assuming that this can all be made to work, then a data fabric promises a kind of nirvana: a way to visualise and inquire about data across the enterprise without the need to pre-integrate and copy swathes of data around into new data sources like data warehouses or data lakes.  A central data catalog provides a kind of treasure map of data across the enterprise, hooked into a business glossary that allows users to navigate data using business terms like “customer” and “product” and “asset” without having to know about the physical locations of this data. The data fabric technology should be able to resolve inconsistencies and duplication of such data (the average company has six “master” definitions of “customer”, and nine of “product” according to one survey) in real-time.

It should be noted that data fabric does not necessarily replace all other existing data management structures. For example, a master data management hub would just be another source system in a data fabric architecture. Indeed this might be a very useful one, since it is unclear how a data fabric virtual layer is going to dynamically resolve the inconsistencies between the various different customer or product hierarchies that exist in most companies. An MDM hub has “survivorship rules” that classify the reliability of the various sources, and applies assorted data quality rules and merge/matching algorithms to resolve data inconsistencies. All this considerable amount of work could be utilised by a data fabric virtual layer that plugged into an existing MDM hub. If no such hub exists then the data fabric technology is going to have to deal with the same issues itself.

There are several underlying components to a data fabric. A virtual software layer is established that maps data assets across the source applications and attempts to transform and process that data as needed to satisfy business needs, all while leaving the source data in place. At its heart is a data catalog with a business glossary, some form of a recommendation engine (perhaps with an AI component) and a way to visualise the data landscape (such as a “knowledge graph”, a visual set of linked descriptions of data entities), a data preparation layer to retrieve data as needed, and some data orchestration component to manage the end-to-end workflow of data, including on-premise and cloud.

Dozens of vendors claim to have data fabric solutions today. Some of these are giant vendors with a long history of data management solutions, and some are newer companies that have assorted capabilities, from data catalogues to artificial intelligence to “active metadata”. Few have end-to-end solutions covering all components, though some claim to.

Solutions

  • ataccama logo
  • CINCHY logo
  • Denodo (logo)
  • InterSystems logo
  • K2view logo
  • Progress logo
  • SNAP LOGIC logo
  • SOLIX logo
  • STARBURST logo
  • teradata logo

These organisations are also known to offer solutions:

  • Cloudera
  • CluedIn
  • Fluree
  • HPE
  • IBM
  • Informatica
  • IRI
  • Microsoft
  • Oracle
  • SAP
  • Talend
  • TIBCO

Research

Supply Chain InPerspective (June 2024 cover)

Leveraging Data Management Partnerships for Enhanced Supply Chain Solutions with InterSystems Supply Chain Orchestrator

This paper looks in more detail at the component parts of a good data management solution, and how they can add value to supply chain ISVs and SIs.
DATA FABRIC MU 2024 (cover)

Data Fabric

Bloor’s analysis of the Data Fabric market, assessing the leading vendors and showing you who the innovators, challengers and champions are.
K2view Data Fabric InBrief (April 2024 cover thumbnail)

Using K2View in a Data Fabric

K2View has a core data platform that fits well into a data fabric or data mesh architecture.
TERADATA Data Fabric InBrief (cover thumb)

The Fabric of Teradata

The Teradata QueryGrid product supports data fabric.
SNAPLOGIC Data Fabric InBrief (cover thumbnail)

SnapLogic in the Data Fabric

SnapLogic has data integration capabilities that support data fabric and data mesh.
PROGRESS Data Fabric InBrief (cover thumbnail)

Data Fabric with Progress MarkLogic

The MarkLogic database offers an approach to a data fabric architecture.
STARBURST InBrief (cover thumbnail)

Starburst – A Data Fabric Foundation Technology

Starburst sells a distributed query engine for very fast queries against very large datasets.
INTERSYSTEMS InBrief cover thumbnail

InterSystems and the Data Fabric

How the InterSystems IRIS platform plays within a data fabric architecture.
Back To Top