Truly data-driven companies see significantly better business outcomes than those that aren\u2019t. According to a recent IDC whitepaper, leaders saw on average two and a half times better results than other organizations in many business metrics. In particular, companies that were leaders at using data and analytics had three times higher improvement in revenues, were nearly three times more likely to report shorter times to market for new products and services, and were over twice as likely to report improvement in customer satisfaction, profits, and operational efficiency.\n\nBut to get maximum value out of data and analytics, companies need to have a data-driven culture permeating the entire organization, one in which every business unit gets full access to the data it needs in the way it needs it.\n\nThis is called data democratization. Doing it right requires thoughtful data collection, careful selection of a data platform that allows holistic and secure access to the data, and training and empowering employees to have a data-first mindset. Security and compliance risks also loom.\n\nStarting on a solid data foundation\n\nBefore choosing a platform for sharing data, an organization needs to understand what data it already has and strip it of errors and duplicates.\n\nA big part of preparing data to be shared is an exercise in data normalization, says Juan Orlandini, chief architect and distinguished engineer at Insight Enterprises.\n\nData formats and data architectures are often inconsistent, and data might even be incomplete. \u201cAll of a sudden, you\u2019re trying to give this data to somebody who\u2019s not a data person,\u201d he says, \u201cand it\u2019s really easy for them to draw erroneous or misleading insights from that data.\u201d\n\nOrganizations often turn to outside help with data normalization because, if done incorrectly, a business might still be left with data quality issues and can\u2019t get as much use out of their data as intended.\n\nAs more companies use the cloud and cloud-native development, normalizing data has become more complicated.\n\n\u201cIt might be in a NoSQL database, a graph database, or in all these other types of databases now available, and making those consistent becomes really challenging,\u201d Orlandini says.\n\nExercising tactful platform selection\n\nIn many cases, only IT has access to data and data intelligence tools in organizations that don\u2019t practice data democratization. So in order to make data accessible to all, new tools and technologies are required.\n\nOf course, cost is a big consideration, says Orlandini, as well as deciding where to host the data, and having it available in a fiscally responsible way. An organization might also question if the data should be maintained on-premises due to security concerns in the public cloud. But Kevin Young, senior data and analytics consultant at consulting firm SPR, says organizations can first share data by creating a data lake like Amazon S3 or Google Cloud Storage. \u201cMembers across the organization can add their data to the lake for all departments to consume,\u201d says Young. But without proper care, a data lake can end up disorganized and cluttered with unusable data. Most organizations don\u2019t end up with data lakes, says Orlandini. \u201cThey have data swamps,\u201d he says.\n\nBut data lakes aren\u2019t the only option for creating a centralized data repository.\n\nAnother is through a data fabric, an architecture and set of data services that provide a unified view of an organization\u2019s data, and enable integration from various sources on-premises, in the cloud and on edge devices.\n\nA data fabric allows datasets to be combined, without the need to make copies, and can make silos less likely.\n\nThere are many data fabric software vendors, like IBM Cloud Pak for Data and SAP Data Intelligence, which were both named leaders in Forrester\u2019s Enterprise Data Fabric Q2 2022 report. But with many available options, it can be difficult to know which to choose.\n\nThe most important thing is to analyze and monitor data, says Amaresh Tripathy, global analytics leader at professional services firm Genpact.\n\n\u201cMany platforms are out there,\u201d he says. \u201cChoose any platform that works for you, but it should be automated and visible.\u201d Also, the data should be easily accessible from a self-service platform that makes data analysis reporting easy, even for people with no technical experience \u2014 \u201cLike a portal where people can see all the data, what it means, what the metrics are, and where it\u2019s coming from,\u201d says Tripathy.\n\nThere\u2019s no perfect tool, and there\u2019s often a trade-off between how well a tool does data lineage, data cataloging, and maintains data quality. \u201cMost organizations are trying to solve all three problems together,\u201d Tripathy adds. \u201cSometimes you over-index on one and don't get a very good value on another.\u201d So an organization should decide what\u2019s most important, he says. \u201cThey should know why they\u2019re doing it, which tool gives them the best bang for their buck on those three dimensions, and then make the appropriate decision.\u201d\n\nWhen thinking about how to share data, an organization can also consider implementing a data mesh, which takes the opposite approach to data fabric. While data fabric manages multiple data sources from a single virtual centralized system, a data mesh is a form of enterprise data architecture that takes a decentralized approach and creates multiple domain-specific systems.\n\nWith a data mesh, organizations can help ensure data is properly handled by putting it in the hands of those who best understand it, says Chris McLellan, director of operations at Data Collaboration Alliance, a global nonprofit that helps people and organizations get full control of their data. It could be a person, such as the head of finance, or a group of people that are acting as data stewards.\n\n\u201cAt its core, it\u2019s got this concept of data as a product,\u201d he says. \u201cAnd a data product is something that can be owned and curated by someone with domain expertise.\u201d\n\nImplementing a data mesh architecture allows an organization to put specific data sets in the hands of subject matter experts. \u201cThese people are closer to the regulations, the customer, and the end users,\u201d McLellan says. \u201cThey\u2019re closer to everything about that specific domain of information.\u201d\n\nData mesh isn\u2019t linked to any specific tools, so individual teams can choose whichever ones best fit their needs, and there isn\u2019t the bottleneck of everything having to go through a central data team.\n\n\u201cYou\u2019re seeing a decentralization not just of IT or app delivery, but also of data management and data governance,\u201d says McLellan, \u201cwhich are good things because marketers know the laws around consumer protection better than the IT team, and finance knows finance regulations better than IT.\u201d\n\nWhile there are many vendors selling data mesh, it\u2019s still a shiny new object, Forrester warns, and it has its challenges, including conflicts in how it\u2019s defined, the technologies it uses, and its value.\n\nTraining and change management\n\nOnce an architecture for data democratization is established, employees need to understand how to work with the new data processes. People can be given the right data, but even if they\u2019re trained as administrators or accountants, they\u2019re not necessarily going to understand what to do with it, says Insight\u2019s Orlandini. Data access is not sufficient in itself to make an organization data-driven. \u201cYou have to do some training,\u201d he says. \u201cIf you don\u2019t do it properly, you\u2019re going to have mixed success at best, or it might be a failure.\u201d\n\nSome organizations have started their own in-house training programs to ensure employees understand how to interpret and properly handle data.\n\nGenpact, for instance, introduced what it calls its DataBridge initiative last year to increase data literacy across the organization.\n\n\u201cOur intention was not to make 100,000 people citizen data scientists,\u201d says Tripathy. \u201cWe provide the awareness in the context of how they do their work.\u201d For example, an employee doing claims analysis doesn\u2019t need to learn all about anomaly detection \u2014 what they need to understand is what anomaly detection means for them. \u201cYou may or may not have all the skill sets to look at the data yourself, but you should be able to raise a question and seek help \u2014 and being able to ask that question in the right manner is the data-aware aspect of it,\u201d he adds.\n\nLaying the security and compliance groundwork\n\nProper data governance needs to be implemented from the start to maintain the integrity of data and avoid costly penalties.\n\nAlong with IT leaders, security and compliance teams need to be part of the initial conversation, says Insight\u2019s Orlandini. \u201cIt\u2019s a big challenge, and a lot of organizations struggle with this,\u201d he says, adding that it\u2019s a prerequisite a company\u2019s leadership understands exactly what they\u2019re offering to share, and makes sure it\u2019s being offered to the right people.\n\n\u201cWe live in a highly regulated world where we have to be super careful,\u201d he says, \u201cespecially in industries like healthcare and finance where there are laws that have severe consequences if you let the wrong person have access to the wrong data.\u201d\n\nThere are also tools that help organizations with data masking and data obfuscation to avoid revealing personally identifiable information. \u201cYou can start getting insights without revealing PII data, HIPAA records, or any of those regulatory requirements that are out there,\u201d he continues. \u201cThere are also tools with attribute-based access controls where you actually tag data with very specific kinds of attributes \u2014 this has PII or HIPAA, whatever your attributes are \u2014 and then you only have access to the data with the right kind of attributes associated with it.\u201d\n\nIn this way, the data controls itself automatically, and it\u2019s available in a public cloud or hybrid environment with data in multiple locations, or even in private environments with strict compliance controls that can be put in place.\n\nLong-term benefits\n\nNot only can data democratization help an enterprise speed up its data pipelines, it can empower people to find new ways to solve problems through a better awareness of how to analyze and work with data.\n\nGartner says that by adopting data democratization, organizations can solve resource shortages, decrease bottlenecks, and enable business units to handle their own data requests more easily. By democratizing data, organizations can improve their decision-making by allowing more people to contribute to the analysis and interpretation of data; increase collaboration across teams within an organization; and enhance transparency, since more people have access to information, and can see how data-driven decisions are made.