What are the challenges for building a data fabric?
While the benefits of data fabric are multifold, there are many blockers that prevent organizations from successfully adopting a solid data fabric architecture.
According to LaPlante, “many companies spend a lot of time, effort, and money [attempting] to figure out if they can access their data, as well as define and add value to it. However, they don't always use analytics to acquire insights. And almost all of them fail to apply what they've learned and [then optimize to] shut the loop. When data fabrics become commodities, how well organizations execute in a closed-loop analytics system will determine how they compete in the future”.5
Challenges to data fabric adoption include, but aren’t limited to, the following:
Modernization
Organizations today often own and use their data assets. A data fabric helps to unify all these assets, creating a “consolidated data management environment that extends across an organization’s edge-to-core-to-cloud infrastructure for all platforms and applications”.3
Tittel argues that data fabric implementation usually requires a modernization process of “[taking] existing fragmented and siloed approaches to managing, storing and situating data into the data fabric’s single, consistent, policy-driven, self-service environment, supported by DevOps and DataOps principles”.3
Tittel states that “modernization is a vital job for organizations seeking to deploy a data fabric [and that it] involves nothing less than a review of all data assets relevant to a business case, and all applications and services that consume and produce data [while trying to open] the data fabric’s umbrella to cover everything”.3
The human component
Data fabric, like so many other technology-based elements in business and the world today, has a human factor. Tittel notes that “applications come with stakeholders, users, and developers, all of whom must understand and buy into that fabric [and enterprises] risk fighting “shadow IT” and other end-arounds that favor “quick and dirty” over consistent and coherent”.3
Investment
Tittel notes that following the financial crises and the pandemic, there’s increased “demand that any technology investments provide a faster [ROI which] puts added pressure on existing investments [and] it encourages cautious and critical consideration of new [investments]”.3 He argues that the market conditions “[benefit] use cases at the high end of the volume and variety curve where bigger payoffs prevail”.3
Agile methodology
Tittel notes that for data fabric adoption, modern data fabrics requires the enterprise to leverage DataOps and must use the “CI/CD [continuous integration, continuous delivery] Agile methodology [that] requires data analysts, business stakeholders, and IT personnel to stop working in disparate environments [and] collaborate to ensure [data comprehension and] requires data analysts, business stakeholders, and IT personnel to stop working in disparate environments”.3
Tittel notes that this integration will produce “self-monitoring and self-measurement for the data fabric so that it keeps getting better at managing data through its entire lifecycle”.3 By this logic, the data fabric is in a persistent state of optimization, and is constantly performing better than it had been previously.
How to optimize your data fabric and help it generate value
Data fabric helps you make better decisions through improved compute performance across data channels. Ashutosh Gupta argues that “to deliver business value through data fabric design, D&A leaders should ensure a solid technology base, identify the required core capabilities, and evaluate the existing data management tools”.2
Data fabric collects and analyzes metadata
Gupta notes that “Contextual information lays the foundation of a dynamic data fabric design. There should be a mechanism (like a well-connected pool of metadata) that enables data fabric to identify, connect, and analyze all kinds of metadata such as technical, business, operational, and social”.2
Passive metadata must be converted into active metadata
Gupta argues that for data sharing to work best, enterprises must activate their metadata. For this metadata to become active, their data fabric must do the following:
- Locate and analyze the key metrics’ and statistics’ metadata
- Construct a graph model
- Graphically depict the metadata as it relates to the enterprise’s relationships
- Use decisive metadata metrics to enable AI and machine learning (ML) algorithms
- These AI/ML algorithms will gradually learn and generate predications for data management, integration, and more2
Knowledge graphs help enrich data with semantics
Gupta notes that the reason data fabric needs to make and curate knowledge graphs, is because these graphs “[enrich] data with semantics [and help] data and analytics leaders to derive business value”.2 He states that the “semantic layer of the knowledge graph makes it more [intuitive and] the analysis easy for D&A leaders”2. Lastly, the semantic layer “adds depth and meaning to the data usage and content graph, allowing AI/ML algorithms to use the information for analytics and other operational use cases”.2
Data fabric requires a robust foundation of data integration
One of the benefits of data fabric adoption is how flexible it is in terms of data delivery. It easily works with multiple data delivery methods, which help its enterprise to accommodate a vast array of data consumers, extending from traditional IT through finance and business.