<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Quatra</title>
	<atom:link href="https://www.quatra.ai/feed/" rel="self" type="application/rss+xml" />
	<link>https://www.quatra.ai/</link>
	<description></description>
	<lastBuildDate>Mon, 08 Jul 2024 23:41:21 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.9.4</generator>

<image>
	<url>https://www.quatra.ai/wp-content/uploads/cropped-quatra-arrows-512-2-32x32.png</url>
	<title>Quatra</title>
	<link>https://www.quatra.ai/</link>
	<width>32</width>
	<height>32</height>
</image> 
	<item>
		<title>Transforming Workforce Shift Verification with Artificial Intelligence</title>
		<link>https://www.quatra.ai/blog/transforming-workforce-shift-verification-with-artificial-intelligence/</link>
		
		<dc:creator><![CDATA[Quatra Marketing]]></dc:creator>
		<pubDate>Thu, 04 Jul 2024 17:33:21 +0000</pubDate>
				<category><![CDATA[Case Study]]></category>
		<guid isPermaLink="false">https://www.quatra.ai/?p=2726</guid>

					<description><![CDATA[<p>The post <a href="https://www.quatra.ai/blog/transforming-workforce-shift-verification-with-artificial-intelligence/">Transforming Workforce Shift Verification with Artificial Intelligence</a> appeared first on <a href="https://www.quatra.ai">Quatra</a>.</p>
]]></description>
										<content:encoded><![CDATA[<div class="et_pb_section et_pb_section_0 et_section_regular" >
				
				
				
				
				
				
				<div class="et_pb_row et_pb_row_0">
				<div class="et_pb_column et_pb_column_4_4 et_pb_column_0  et_pb_css_mix_blend_mode_passthrough et-last-child">
				
				
				
				
				<div class="et_pb_module et_pb_text et_pb_text_0  et_pb_text_align_left et_pb_bg_layout_light">
				
				
				
				
				<div class="et_pb_text_inner"><h1>Opportunity for Change</h1>
<p>A leading healthcare provider, faced a significant challenge in managing the shift verification process for its hourly workers. Each week, the payroll team had to manually review shift entries and compare them to timecards, shift terminals, or website logs that workers used to clock in and out. This process was not only tedious and time-consuming but also prone to errors.</p>
<p>On average, the team reviewed 50 shifts per hour, with a 12% error rate. Incorrect approvals led to improper payments, requiring extensive corrections involving HR, payroll, and the workers. Conversely, denying correct shifts delayed worker payments, affecting morale and satisfaction. The payroll team dedicated two full days each week to complete these verifications, straining resources and impacting efficiency.</p>
<h1>Solution</h1>
<p>The healthcare provider turned to Quatra to develop a revolutionary shift entry verification solution. Leveraging advanced artificial intelligence and optical character recognition (OCR) technology, Quatra AI automated the verification process at the time of entry by the worker. Here&#8217;s how the solution worked:</p>
<ol>
<li><strong>Automated Verification:</strong> Quatra AI used OCR and natural language processing (NLP) to read and interpret the shift details from the timecards, shift terminals, or websites.</li>
<li><strong>Real-time Comparison:</strong> The AI compared the interpreted data with the shift entries in the web application.</li>
<li><strong>Immediate Feedback:</strong> If a mismatch was detected, the worker received an immediate notification to correct the entry.</li>
<li><strong>Continuous Learning:</strong> The system collected feedback on corrected mismatches, continuously training the AI to improve accuracy.</li>
</ol>
<p>This proactive approach ensured that errors were addressed instantly when enter by the worker, significantly reducing the burden on the payroll team and enhancing overall accuracy.</p>
<h1>Benefits</h1>
<p><strong>High Return on Investment:</strong> By addressing errors at the point of entry, Sadiant Health minimized costly corrections later in the payroll cycle. This proactive error management saved substantial time and money.</p>
<p><strong>Fast ROI:</strong> The solution was implemented and live within just four weeks from the project kickoff, delivering quick and tangible results.</p>
<p><strong>Low Total Cost of Ownership:</strong> The setup was straightforward, and Quatra AI operated independently without requiring constant oversight. Its continuous learning capability ensured ongoing improvements without additional costs.</p>
<p><strong>Increased Payroll Team Morale:</strong> The elimination of tedious shift verification tasks allowed the payroll team to focus on more strategic activities, boosting their job satisfaction and morale.</p>
<p><strong>Enhanced Worker Retention and Satisfaction:</strong> With accurate and timely payments, worker satisfaction increased, leading to improved retention rates.</p>
<h1>Results</h1>
<p>The implementation of the shift verification solution with the Quatra platform yielded remarkable results:</p>
<ul>
<li><strong>Error Rate Reduction:</strong> The error rate in shift verification dropped from 12% to 2%.</li>
<li><strong>Time Savings:</strong> The automated system eliminated the need for 10,000 manual shift verifications per week, saving the payroll team 200 hours weekly.</li>
</ul>
<p>The healthcare providers collaboration with Quatra AI transformed their payroll process, demonstrating the powerful impact of artificial intelligence in streamlining operations and enhancing accuracy. This case study highlights the significant benefits of adopting innovative technology to solve complex, time-consuming problems, setting a new standard for efficiency in workforce management.</p></div>
			</div>
			</div>
				
				
				
				
			</div>
				
				
			</div>
<p>The post <a href="https://www.quatra.ai/blog/transforming-workforce-shift-verification-with-artificial-intelligence/">Transforming Workforce Shift Verification with Artificial Intelligence</a> appeared first on <a href="https://www.quatra.ai">Quatra</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>The Fourth Step to Data Mesh is Federated Governance</title>
		<link>https://www.quatra.ai/blog/the-fourth-step-to-data-mesh-is-federated-governance/</link>
		
		<dc:creator><![CDATA[Quatra Marketing]]></dc:creator>
		<pubDate>Wed, 03 Jul 2024 20:21:16 +0000</pubDate>
				<category><![CDATA[Data Mesh]]></category>
		<guid isPermaLink="false">https://www.quatra.ai/?p=2711</guid>

					<description><![CDATA[<p>The post <a href="https://www.quatra.ai/blog/the-fourth-step-to-data-mesh-is-federated-governance/">The Fourth Step to Data Mesh is Federated Governance</a> appeared first on <a href="https://www.quatra.ai">Quatra</a>.</p>
]]></description>
										<content:encoded><![CDATA[<div class="et_pb_section et_pb_section_1 et_section_regular" >
				
				
				
				
				
				
				<div class="et_pb_row et_pb_row_1">
				<div class="et_pb_column et_pb_column_4_4 et_pb_column_1  et_pb_css_mix_blend_mode_passthrough et-last-child">
				
				
				
				
				<div class="et_pb_module et_pb_text et_pb_text_1  et_pb_text_align_left et_pb_bg_layout_light">
				
				
				
				
				<div class="et_pb_text_inner"><div class="et_pb_module et_pb_text et_pb_text_0  et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<div class="et_pb_row et_pb_row_0">
<div class="et_pb_column et_pb_column_4_4 et_pb_column_0  et_pb_css_mix_blend_mode_passthrough et-last-child">
<div class="et_pb_module et_pb_text et_pb_text_0  et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<p>Data mesh is a strategic approach to decentralized data management that provides standardized self-serve data products through a governance model. This enables data-driven organizations to achieve agile, comprehensive and secure data management.</p>
<p>Four main principles are fundamental in a data mesh system. The following first three principles have been covered in previous blog posts:</p>
<p>1.<span> </span><a href="/blog/the-first-step-to-data-mesh-is-domain-ownership/" rel="noreferrer">Domain ownership</a></p>
<p>2.<span> </span><a href="/blog/the-second-step-to-data-mesh-is-data-as-a-product/" rel="noreferrer">Data as a product</a></p>
<p>3.<span> </span><a href="/blog/the-third-step-to-data-mesh-is-the-self-serve-platform/" rel="noreferrer">Self-serve platform</a></p>
<p>This post will cover the fourth principle, Federated Governance, which enables organizations to maintain a robust and active mesh ecosystem by ensuring system engagement, mitigating domain isolation, administering product standardization, and addressing platform operational needs.</p>
<p>&nbsp;</p>
<h3 class="wp-block-heading"><strong>What is Federated Governance?</strong></h3>
<p>Federated governance is the operating model of the data mesh architecture and allows for the optimum interoperability of the other mesh principles.</p>
<p>The federation comprises domain and platform representatives and representatives from relevant associates in the organization, such as legal, compliance, and security.</p>
<p>The representatives comprise a team tasked with developing the data mesh, platform policies, product development, and operating standards. These oversight tasks of the federated governance team ensure product and operational consistency and maintain interoperability within the data mesh ecosystem.</p>
<p>The governance team is responsible for developing the guide of operating standards for the data mesh. This guide should include how decisions are made and executed, conflicts are resolved, and global standards on platform process, product development, and quality are created.</p>
<p>For the ecosystem to function, federated governance interprets the mesh as an evolving ecosystem overseen globally but also allows for policy flexibility with domains in so much as global policy and standards are followed. In this way, domains are incentivized to innovate to develop high-value and useful products without the friction of high-level bureaucracy.</p>
<p>&nbsp;</p>
<h3 class="wp-block-heading"><strong>What Problems Does Federated Governance Address?</strong></h3>
<p>Data mesh helps organizations optimize data structure and movement by using independent domain team experts to develop valuable data products that are shared and discoverable on a self-serve platform.</p>
<p>However, for the mesh system to thrive, it needs rules and operational oversight to ensure consistency and value throughout the mesh.</p>
<p>Therefore, assembling a federated governance operating model allows for the proper interoperability of the mesh by addressing the following problems.</p>
<p><strong>Domain stability: </strong>Ensures domains are compatible within the mesh system and operate under a unified operational umbrella.</p>
<p><strong>Data product consistency</strong>: Safeguards against product inconsistency and enforces product quality standards and truthfulness.</p>
<p><strong>Organization rules and compliance:</strong> The governance team upholds organizational compliance through the mesh asserting that the mesh components are secure and trusted.</p>
<p><strong>Operational costs: </strong>The governance reduces costs by automating the operational compliance processes for the mesh, which decreases the need for manual operational intervention.</p>
<p>&nbsp;</p>
<h3 class="wp-block-heading"><strong>Federated Governance Role in Data Mesh</strong></h3>
<p>Think of governance as a steering committee where domains have decision-making autonomy around products and adhere to a global set of rules set forth by the governance team. The rules and policies are automated and embedded into all data products and the self-serve platform.</p>
<p>The organization’s specialized functions, such as legal and security, set global rules. The standardized rules’ overall objective is to ensure trust, security, and interoperability of the mesh. These standards, in turn, lead to a consistent experience that end-users have with organization data.</p>
<p>In the book <em>Data Mesh: Delivering Data-Driven Value at Scale</em>, author Zhamak Dehghani outlines three components for federated governance. These components help guide governance development and thinking.</p>
<p><strong>System thinking</strong>: Accounts for the dynamic interconnectedness of the data mesh. When operational and policy decisions are made, governance considers the data product, domain autonomy, self-serve platform, and end users.</p>
<p><strong>Computational Policies</strong>: The rule set for governance determine quality standards and how the standards are enforced and monitored. Governance influences both local and global policies. Likewise, the governance is incentive-based to ensure local and global policy adherence.</p>
<p><strong>Federated Operating Model</strong>: The governance operating model is designed as a decentralized autonomous system governed by a set of rules and policies to ensure consistent quality and operational standards throughout the mesh. The global and local policies are set by a team of cross-functional representatives from domains and subject matter experts.</p>
<p>&nbsp;</p>
<h3 class="wp-block-heading"><strong>The Challenges of Federated Governance</strong></h3>
<p>As the fourth principle in the data mesh, federated governance addresses many challenges that are presented by previous principles. These benefits cannot be realized without handling one challenge presented by federated governance.</p>
<p>This challenge is the ability of the organization to embrace constant change as the mesh evolves and improves. Therefore, when implementing federated governance, it is essential to consult with change experts to help guide and set up the correct structure for your organization to embrace these changes.</p>
<p>&nbsp;</p>
<h3 class="wp-block-heading"><strong>Completing the Mesh</strong></h3>
<p>We have now covered what a data mesh is and the four core principles of a data mesh. However, to optimize the mesh, you need to start with quality data at the domain level. Any mistake in quality can erode the value of a mesh and the self-serve data products. A small amount of poor data originating from one domain data source can spread throughout your enterprise and wreak havoc. Bad decisions, missed opportunities, tarnished brands, diminished credibility, financial audits, wasted time, and expenses may result. For this reason, Quatra necessitates a fifth principle of <a href="/blog/category/data-quality/">quality data</a>, thus ensuring optimized outputs of the mesh system.</p>
</div>
</div>
</div>
</div>
</div>
</div>
<ul></ul></div>
			</div>
			</div>
				
				
				
				
			</div>
				
				
			</div>
<p>The post <a href="https://www.quatra.ai/blog/the-fourth-step-to-data-mesh-is-federated-governance/">The Fourth Step to Data Mesh is Federated Governance</a> appeared first on <a href="https://www.quatra.ai">Quatra</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>The Third Step to Data Mesh is the Self-Serve Platform</title>
		<link>https://www.quatra.ai/blog/the-third-step-to-data-mesh-is-the-self-serve-platform/</link>
		
		<dc:creator><![CDATA[Quatra Marketing]]></dc:creator>
		<pubDate>Wed, 03 Jul 2024 20:13:34 +0000</pubDate>
				<category><![CDATA[Data Mesh]]></category>
		<guid isPermaLink="false">https://www.quatra.ai/?p=2707</guid>

					<description><![CDATA[<p>The post <a href="https://www.quatra.ai/blog/the-third-step-to-data-mesh-is-the-self-serve-platform/">The Third Step to Data Mesh is the Self-Serve Platform</a> appeared first on <a href="https://www.quatra.ai">Quatra</a>.</p>
]]></description>
										<content:encoded><![CDATA[<div class="et_pb_section et_pb_section_2 et_section_regular" >
				
				
				
				
				
				
				<div class="et_pb_row et_pb_row_2 dbdb_default_mobile_width">
				<div class="et_pb_column et_pb_column_4_4 et_pb_column_2  et_pb_css_mix_blend_mode_passthrough et-last-child">
				
				
				
				
				<div class="et_pb_module et_pb_text et_pb_text_2  et_pb_text_align_left et_pb_bg_layout_light">
				
				
				
				
				<div class="et_pb_text_inner"><div class="et_pb_module et_pb_text et_pb_text_0  et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<div class="et_pb_row et_pb_row_0">
<div class="et_pb_column et_pb_column_4_4 et_pb_column_0  et_pb_css_mix_blend_mode_passthrough et-last-child">
<div class="et_pb_module et_pb_text et_pb_text_0  et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<p>Data mesh is an approach to data management that helps organizations become more agile and prudent with managing, analyzing, and distributing data.</p>
<p>Four main principles are fundamental in a data mesh. The first principle is<span> </span><a href="/blog/the-first-step-to-data-mesh-is-domain-ownership/">domain ownership</a>, and the second is<span> </span><a href="/blog/the-second-step-to-data-mesh-is-data-as-a-product/">data as a product</a>, both of which were covered in previous blog posts.  </p>
<p>This post will cover the third principle, the self-serve data platform. This platform is the organization’s data conduit, streamlining the creation, deployment, and maintenance by domain owners and allowing for the data product’s accessibility and shareability.</p>
<p>Organizations benefit by reducing product development and ownership costs and empowering the domain teams to optimize the utility of data products. Likewise, the platform ensures consistency of data product attributes like quality and compliance through automated governance policies.</p>
<p>&nbsp;</p>
<h3 class="wp-block-heading"><strong>What is a Self-Serve Data Platform?</strong></h3>
<p>A self-serve platform is a set of technologies and the underlying infrastructure that simplify otherwise complex decentralized data management and sharing. This platform allows organizations to attain key objectives like:</p>
<p><strong>Autonomous team productivity:</strong><span> </span>The ability to enable a team to complete their work with a sense of autonomy and without involving another team.</p>
<p><strong>Exchange of trusted &amp; governed data products:</strong><span> </span>Smooth exchange of data products with a certified level of data quality from the provider to the consumer. The platform allows for consistent quality, security, compliance, and process through governance policy structure and enforcement.</p>
<p><strong>Accelerate time to value:</strong><span> </span>The platform should abstract technical complexity and provide elegant APIs to decrease the knowledge ramp-up required for domain teams and the steps required to deploy data products.</p>
<p><strong>Scalable and flexible sharing:</strong><span> </span>Data sharing is enabled across internal and external consumers, such as a network of partners. To reach this breadth of sharing, designing for secure interoperability and integration with other platforms is required.</p>
<p><strong>Innovation Culture</strong>: Free domain teams from data management activities not contributing to innovation, so they can focus on data discovery, exploration, and analysis to uncover valuable insights to be shared.</p>
<p>&nbsp;</p>
<h3 class="wp-block-heading"><strong>What Problems Does a Self-Serve Platform Address?</strong></h3>
<p>Data mesh architecture allows organizations to develop valuable and useful data products. Knowledge experts create these products within business domains across the organization.</p>
<p>However, the actual value and utility of the data products can only be realized through cost-effective and efficient shareability and discoverability across the organization. </p>
<p>For this reason, a self-serve platform is key to unlocking the actual value of the data mesh. Below are three specific problems the platform addresses. </p>
<p><strong>Shareability and discoverability: </strong>The platform allows domain teams to seamlessly share their data products across the organization. Likewise, knowledge workers can easily access and find useful data products on the platform.</p>
<p><strong>Duplication of efforts from domains</strong>: Domain teams are empowered to share and use data products on the platform, which reduces duplication, specifically across cross-functional domain teams.</p>
<p><strong>Costs of operation:</strong> The self-serve platform integrates and shares standard capabilities with existing data platforms. Additionally, it reduces cognitive load across domains and allows for the functioning of a general technologist throughout the system.</p>
<p><strong>Product inconsistency and incompatibility: </strong>The platform provides domain agnostic infrastructure and services, which includes the execution of policies to ensure consistency and compatibility of each data product.</p>
<p>&nbsp;</p>
<h3 class="wp-block-heading"><strong>Self-Serve Data Platform Role in Data Mesh</strong></h3>
<p>Within a data mesh architecture, the self-serve data platform is designed to optimize and integrate with existing technologies within a data architecture. Integration examples include data storage, processing frameworks, query languages, data catalogs, and pipeline workflow management.</p>
<p>Likewise, in the book <em>Data Mesh: Delivering Data-Driven Value at Scale</em>, author Zhamak Dehghani outlines six platform characteristics. These self-serve platform characteristics optimize the existing platforms and allow for the functionality of the data mesh architecture.</p>
<p><strong>Serves autonomous domain-oriented teams</strong>: Enables domain engineering teams in building, sharing, and using data products.</p>
<p><strong>Manages autonomous interoperable data products</strong>: The platform works with data products to make them discoverable, usable, trustworthy, and secure for end users throughout the product life cycle.</p>
<p><strong>Integrated platform of operational and analytical capabilities</strong>: The platform provides a connected experience for domains, bringing together the operational and analytical constituents of the data architecture.</p>
<p><strong>Designed for a generalist majority</strong>: The platform promotes interoperability between different technologies and reduces the need for proprietary languages and experiences designed for specialists. It incentivizes and enables experienced generalist developers. </p>
<p><strong>Favor decentralized technologies</strong>: The platform supports and enables the notion of the decentralized data architecture of a data mesh. Additionally, they provide centralization of tasks that help remove friction from domains and technologies.</p>
<p><strong>Domain agnostic</strong>: The platform is designed to enable all domain teams by balancing domain-agnostic capabilities with domain-specific data modeling, processing, and sharing across the<br />organization.</p>
<p>&nbsp;</p>
<h3 class="wp-block-heading"><strong>The Challenges of a Self-Serve Data Platform</strong></h3>
<p>Below is the primary challenge of the self-serve platform and the mesh principle that addresses the challenge.</p>
<p><strong>Consistent product quality and policy adherence: </strong>As with the domain ownership and data as a product principles, the self-serve platform needs to support the optimization of consistent product quality, security, privacy, and legal compliance. This challenge is addressed through the fourth mesh principle of federated computational governance, which provides an operating model to balance domain autonomy with organization-wide interoperability of a data mesh.</p>
<p>&nbsp;</p>
<h3 class="wp-block-heading"><strong>What’s Next After the Self-Serve Platform?</strong></h3>
<p>The self-serve data platform allows for the shareability and discoverability of data products created by business domains. </p>
<p>The data platform brings the data mesh architecture together through the cost reductions of decentralized data ownership, lessens the complexities of data infrastructures, diminishes the need for technology specialists, and automates quality and data policies through governance.</p>
<p>The fourth principle is the principle of federated governance which ensures system engagement, mitigates domain isolation, administers product standardization, and addresses platform operational needs. We will cover this principle in our next blog post.</p>
</div>
</div>
</div>
</div>
</div>
</div>
<ul></ul></div>
			</div>
			</div>
				
				
				
				
			</div>
				
				
			</div>
<p>The post <a href="https://www.quatra.ai/blog/the-third-step-to-data-mesh-is-the-self-serve-platform/">The Third Step to Data Mesh is the Self-Serve Platform</a> appeared first on <a href="https://www.quatra.ai">Quatra</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>The Second Step to Data Mesh is Data as a Product</title>
		<link>https://www.quatra.ai/blog/the-second-step-to-data-mesh-is-data-as-a-product/</link>
		
		<dc:creator><![CDATA[Quatra Marketing]]></dc:creator>
		<pubDate>Wed, 03 Jul 2024 20:09:29 +0000</pubDate>
				<category><![CDATA[Data Mesh]]></category>
		<guid isPermaLink="false">https://www.quatra.ai/?p=2700</guid>

					<description><![CDATA[<p>The post <a href="https://www.quatra.ai/blog/the-second-step-to-data-mesh-is-data-as-a-product/">The Second Step to Data Mesh is Data as a Product</a> appeared first on <a href="https://www.quatra.ai">Quatra</a>.</p>
]]></description>
										<content:encoded><![CDATA[<div class="et_pb_section et_pb_section_3 et_section_regular" >
				
				
				
				
				
				
				<div class="et_pb_row et_pb_row_3 dbdb_default_mobile_width">
				<div class="et_pb_column et_pb_column_4_4 et_pb_column_3  et_pb_css_mix_blend_mode_passthrough et-last-child">
				
				
				
				
				<div class="et_pb_module et_pb_text et_pb_text_3  et_pb_text_align_left et_pb_bg_layout_light">
				
				
				
				
				<div class="et_pb_text_inner"><div class="et_pb_module et_pb_text et_pb_text_0  et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<div class="et_pb_row et_pb_row_0">
<div class="et_pb_column et_pb_column_4_4 et_pb_column_0  et_pb_css_mix_blend_mode_passthrough et-last-child">
<div class="et_pb_module et_pb_text et_pb_text_0  et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<p>Data mesh is a new approach to organizational data management that helps data-driven organizations become more agile and prudent with managing, analyzing, and distributing data.</p>
<p>Four main principles are fundamental in a data mesh system. The first principle is domain ownership which we discussed in our<span> </span><a href="/blog/the-first-step-to-data-mesh-is-domain-ownership/">previous blog post</a>.  </p>
<p>The second principle is data as a product of which domain data is developed and shared as a product for internal and external customers.</p>
<p>The following post will cover the data as a product principle and the benefits it presents to an organization.</p>
<p>&nbsp;</p>
<h3 class="wp-block-heading"><strong>What is Data as a Product?</strong></h3>
<p>Data as a product is the concept of leveraging analytical domain sourced data to develop a high utility data product to be shared and used across the organization.</p>
<p>In a decentralized data mesh architecture, domain teams ingest, process, and distribute their sourced data as usable data products. Developing data products at the domain source reduces friction and high costs typical to traditional data management methodologies.</p>
<p>Data as a product is an evolution in thinking in which external product management techniques are applied internally. Internal users of the data product are the customers, and the domain teams are responsible for the product and prioritizing the customer experience.</p>
<p>&nbsp;</p>
<h3 class="wp-block-heading"><strong>What Problems Does Data as a Product Address?</strong></h3>
<p>The value of data within a data-driven organization is uncovered through its usability. For data to become usable, it needs to be converted into a product and then shared within the organization.</p>
<p>Likewise, evolving the thinking around data from an asset to a product promotes an organization where data becomes part of the culture when planning, making decisions and implementing strategies.</p>
<p>As such, data as a product addresses the following data problems within an organization.</p>
<p><strong>Data siloing</strong>: Data as a product prevents domains from becoming a funnel for data collection, and instead, domains become a data product team creating products to be shared.</p>
<p><strong>Value:</strong> Data as a product increases the value of data by increasing the utility and shareability of the data.</p>
<p><strong>Utility: </strong>Data as a product extracts more utility from data as teams innovate around data products, increasing the quality and trustfulness of the data.</p>
<p>&nbsp;</p>
<h3 class="wp-block-heading"><strong>Data as a Product Role in Data Mesh</strong></h3>
<p>Data as a product can be defined through its customer experience. To provide a unified customer experience, the domain teams optimize the usability of the data for the organization.</p>
<p>Various attributes define organizational data usability. Zhamak Dehghani best defines these attributes in her book Data Mesh. </p>
<p>The following summarizes the eight attributes Dehghani outlines for data usability within the organization:</p>
<p><strong>Discoverable</strong>: Data products are developed with discoverability as a priority. Data users can easily search and find data products for their needs.</p>
<p><strong>Addressable</strong>: Each data product includes a unique address that allows users to programmatically and consistently access the data product as it changes throughout the product lifecycle.</p>
<p><strong>Understandable</strong>: Data products are developed to be easily understood by the user. Users will understand the data’s meaning, how it is presented, and how it has been used within the organization.</p>
<p><strong>Trustworthy</strong>: The truthfulness of the data is established in each data product. The user has confidence in the product’s reliability and accurate facts. Ensuring data quality in a data mesh presents unique challenges. Learn how to <a href="/blog/how-to-select-the-best-technology-for-data-quality-management/">select the best technology for data quality management</a>.</p>
<p><strong>Natively Accessible</strong>: Data products can be read and accessed by various modes of access by different user types.</p>
<p><strong>Interoperable</strong>: Data products have consistent standards for exchanging and using data.</p>
<p><strong>Valuable</strong>: Data products are developed with stand-alone value for the user. </p>
<p><strong>Secure</strong>: Data products contain standardized security policies.</p>
<p>&nbsp;</p>
<h3 class="wp-block-heading"><strong>The Challenges of Data as a Product and How They are Addressed</strong></h3>
<p>Below are the primary challenges of data products and the mesh principle that addresses the challenge.</p>
<p><strong>Cost of ownership: </strong>Domain ownership of data products can increase the cost of data through increased data product management. This cost is mitigated through the self-serve platform principle, which reduces duplicative efforts and increases workload capacity and productivity.</p>
<p><strong>Consistency of value and utility:</strong> Consistency of product quality across multiple domains can become a challenge and ultimately reduce the value and trustworthiness of the data. The federated governance principle helps address this through the application of data quality technology, global policies, and accountability across domains and subject matter experts.</p>
<p>&nbsp;</p>
<h3 class="wp-block-heading"><strong>What’s Next After Data as a Product?</strong></h3>
<p>Data as a product optimizes enterprise data by turning analytical data into a product to be shared and used across the organization. By turning data into a product, the organization reduces data distribution friction and costs and prevents data siloing from domain ownership.</p>
<p>To distribute the product, a self-serve platform is needed. Developing a self-serve platform is the next step in a data mesh architecture, and we will cover it in our next blog post.</p>
</div>
</div>
</div>
</div>
</div>
</div>
<ul></ul></div>
			</div>
			</div>
				
				
				
				
			</div>
				
				
			</div>
<p>The post <a href="https://www.quatra.ai/blog/the-second-step-to-data-mesh-is-data-as-a-product/">The Second Step to Data Mesh is Data as a Product</a> appeared first on <a href="https://www.quatra.ai">Quatra</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>The First Step to Data Mesh is Domain Ownership</title>
		<link>https://www.quatra.ai/blog/the-first-step-to-data-mesh-is-domain-ownership/</link>
		
		<dc:creator><![CDATA[Quatra Marketing]]></dc:creator>
		<pubDate>Wed, 03 Jul 2024 20:02:09 +0000</pubDate>
				<category><![CDATA[Data Mesh]]></category>
		<guid isPermaLink="false">https://www.quatra.ai/?p=2695</guid>

					<description><![CDATA[<p>The post <a href="https://www.quatra.ai/blog/the-first-step-to-data-mesh-is-domain-ownership/">The First Step to Data Mesh is Domain Ownership</a> appeared first on <a href="https://www.quatra.ai">Quatra</a>.</p>
]]></description>
										<content:encoded><![CDATA[<div class="et_pb_section et_pb_section_4 et_section_regular" >
				
				
				
				
				
				
				<div class="et_pb_row et_pb_row_4 dbdb_default_mobile_width">
				<div class="et_pb_column et_pb_column_4_4 et_pb_column_4  et_pb_css_mix_blend_mode_passthrough et-last-child">
				
				
				
				
				<div class="et_pb_module et_pb_text et_pb_text_4  et_pb_text_align_left et_pb_bg_layout_light">
				
				
				
				
				<div class="et_pb_text_inner"><div class="et_pb_module et_pb_text et_pb_text_0  et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<div class="et_pb_row et_pb_row_0">
<div class="et_pb_column et_pb_column_4_4 et_pb_column_0  et_pb_css_mix_blend_mode_passthrough et-last-child">
<div class="et_pb_module et_pb_text et_pb_text_0  et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<p>In our previous<span> </span><a href="/blog/unleashing-value-with-a-decentralized-data-mesh/" rel="noreferrer">post</a>, we discussed the opportunity of developing a data mesh architecture and how it can benefit data-driven enterprise organizations.</p>
<p>In the next four posts of our data mesh series, we will cover the fundamental principles that make up the organizational functionality of a data mesh. This post will cover the first principle, Domain Ownership, and the evolution of decentralized data management.</p>
<p>&nbsp;</p>
<h2 class="wp-block-heading"><strong>What is Domain Ownership?</strong></h2>
<p>Domains represent natural business units in an organization where data is sourced or consumed.</p>
<p>Therefore, domain ownership decentralizes the data ownership and management responsibilities to the business units or domains within an organization, creating data management at the source.</p>
<p>Domain ownership changes the flow of data within the organization. In a traditional architecture, data flows from the source to a centralized system, like a data lake, to be ingested, processed, and served. This centralized system is commonly where new individuals or groups take responsibility for owning the data. In a data mesh, the domain is the source where data is ingested, processed, and served to the organization.</p>
<p>&nbsp;</p>
<h2 class="wp-block-heading"><strong>What Problems Does Domain Ownership Address?</strong></h2>
<p>As enterprise organizations become increasingly data-driven, there is a higher demand for acute knowledge of the data source to quickly process and share data in a valuable and agile way.</p>
<p>Rapidly understanding and processing data is a challenge with traditional centralized architecture where data is partitioned by technology whose data engineers work with multiple sources.</p>
<p>Domain ownership solves these challenges in the following three primary ways.</p>
<p><strong>Knowledge responsibility</strong>: Domain ownership allows teams closest to and most knowledgeable of the data to own it, improving the value and truthfulness of the data outputs.</p>
<p><strong>Change and agility:</strong> Data modeling is localized to the business units most familiar with the data through domain ownership. This localization allows for more agile modeling and the ability to implement change without centralized coordination.</p>
<p><strong>Scaled data use and sharing: </strong>Increasing the number of sources ingesting and processing data means higher data utility and output, leading to increased sharing and consumption.</p>
<p>&nbsp;</p>
<h2 class="wp-block-heading"><strong>Domain Ownership Role in Data Mesh</strong></h2>
<p><strong>Data ingestion and processing</strong>: In a decentralized architecture, the domains are responsible for ingesting data for immediate use or storage. Likewise, they are responsible for processing or cleaning data for analysis and insights.</p>
<p><strong>Analytics and data product</strong>: Once domains have ingested and processed the source data, they are responsible for analytics and developing a data product to be distributed and shared.</p>
<p>Domain data can be classified into three variations.</p>
<p><em>Source-aligned domain data:</em><strong> </strong>Data facts are generated by business operations. Source-aligned data most closely relates to domain events generated by domain operational systems.</p>
<p><em>Aggregate domain data:</em><strong> </strong>Data bound from multiple domain sources to form a data product.</p>
<p><em>Consumer aligned domain data:</em><strong> </strong>Data transformed for specific use cases across organizational functions.</p>
<p>These variations represent how domains interact with the data and develop subsequent data products.</p>
<p>&nbsp;</p>
<h2 class="wp-block-heading"><strong>The Challenges of Domain Ownership and How They are Addressed</strong></h2>
<p>Domain ownership can have a unique set of challenges that must be addressed. Below are the primary challenges and the mesh principle that addresses the challenge.</p>
<p><strong>Risk of data siloing</strong> to where domains collect data and are not incentivized to share. This risk is addressed through the <em>data as a product</em> principle.</p>
<p><strong>Lack of empowerment</strong> across domains to share data within the organization. This challenge is addressed through the <em>self-serve platform</em> principle.</p>
<p><strong>The risk of poor engagement and isolation</strong> occurs if there is no structure for accountability, domain organization, global operation, and policy. This challenge is addressed through the <em>federated governance</em> principle.</p>
<p>&nbsp;</p>
<h2 class="wp-block-heading"><strong>What’s Next After Taking Domain Ownership?</strong></h2>
<p>A domain ownership approach allows data management at the source, with the source experts overseeing the data. This approach allows for reduced friction and scalability of serving analytical data to organizational data consumers. Domains allow the mesh to work by using domain data products upstream and downstream.</p>
<p>To learn more about the next three principles of data as a product, self-serve platform, and federated governance and how they support the challenges of domain ownership, stay tuned for our upcoming blog post.</p>
</div>
</div>
</div>
</div>
<div class="et_pb_row et_pb_row_4">
<div class="et_pb_column et_pb_column_4_4 et_pb_column_4  et_pb_css_mix_blend_mode_passthrough et-last-child">
<div class="et_pb_module et_pb_text et_pb_text_8  et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner"></div>
</div>
</div>
</div>
</div>
</div>
<ul></ul></div>
			</div>
			</div>
				
				
				
				
			</div>
				
				
			</div>
<p>The post <a href="https://www.quatra.ai/blog/the-first-step-to-data-mesh-is-domain-ownership/">The First Step to Data Mesh is Domain Ownership</a> appeared first on <a href="https://www.quatra.ai">Quatra</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Unleashing Value with a Decentralized Data Mesh</title>
		<link>https://www.quatra.ai/blog/unleashing-value-with-a-decentralized-data-mesh/</link>
		
		<dc:creator><![CDATA[Quatra Marketing]]></dc:creator>
		<pubDate>Wed, 03 Jul 2024 19:57:35 +0000</pubDate>
				<category><![CDATA[Data Mesh]]></category>
		<guid isPermaLink="false">https://www.quatra.ai/?p=2690</guid>

					<description><![CDATA[<p>The post <a href="https://www.quatra.ai/blog/unleashing-value-with-a-decentralized-data-mesh/">Unleashing Value with a Decentralized Data Mesh</a> appeared first on <a href="https://www.quatra.ai">Quatra</a>.</p>
]]></description>
										<content:encoded><![CDATA[<div class="et_pb_section et_pb_section_5 et_section_regular" >
				
				
				
				
				
				
				<div class="et_pb_row et_pb_row_5 dbdb_default_mobile_width">
				<div class="et_pb_column et_pb_column_4_4 et_pb_column_5  et_pb_css_mix_blend_mode_passthrough et-last-child">
				
				
				
				
				<div class="et_pb_module et_pb_text et_pb_text_5  et_pb_text_align_left et_pb_bg_layout_light">
				
				
				
				
				<div class="et_pb_text_inner"><div class="et_pb_module et_pb_text et_pb_text_0  et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<div class="et_pb_row et_pb_row_0">
<div class="et_pb_column et_pb_column_4_4 et_pb_column_0  et_pb_css_mix_blend_mode_passthrough et-last-child">
<div class="et_pb_module et_pb_text et_pb_text_0  et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<p>Organizationally sourced data is increasingly complex and filled with valuable business intelligence. As organizations become increasingly data-driven, they will need to become more agile and prudent with managing, analyzing, and distributing data.</p>
<p>Implementing a data mesh architecture allows an organization to leverage a decentralized data management system that aligns with the current organizational structure. A data mesh distributes the data burden to experts at the domain data source. Keeping domain experts close to the data and analytics throughout most of the domain data lifecycle is an efficient and cost-conscious strategy for optimizing data management.</p>
<p>In the following post, I will explain the data mesh concept, how it addresses current data architecture problems, and the four principles of a data mesh.</p>
</div>
</div>
</div>
</div>
<div class="et_pb_row et_pb_row_1">
<div class="et_pb_column et_pb_column_4_4 et_pb_column_1  et_pb_css_mix_blend_mode_passthrough et-last-child">
<div class="et_pb_module et_pb_text et_pb_text_1  et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<h1>What is a Data Mesh?</h1>
</div>
</div>
<div class="et_pb_module et_pb_text et_pb_text_2  et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<p>Zhamak Dehghani coined the term data mesh in her book <em>Data Mesh: Delivering Data Driven Value at Scale</em>.</p>
<p>Dehghani defines data mesh as a decentralized sociotechnical approach to sharing, accessing, and managing analytical data in complex and large-scale environments within or across organizations. </p>
<p>If we break down this definition, we get four main themes inherent to data mesh architecture.</p>
<p><strong>Decentralized</strong>: Moving from one central management system that serves as the aggregator and distributor of data to multiple domains or business units each managing data they create and source. These decentralized domains serve as the aggregators and distributors.</p>
<p><strong>Sociotechnical:</strong> An evolved approach to data management that encompasses the interactions between people, technical architecture, and solutions.</p>
<p><strong>Share &amp; Access:</strong> A data platform that can consistently extract utility and value out of data and make it available to users across the organization when needed.</p>
<p><strong>Manage analytical data</strong>: Mature management and collaborative governance from multiple stakeholders is needed to provide consistency, security, and quality of business intelligence.</p>
</div>
</div>
</div>
</div>
<div class="et_pb_row et_pb_row_2">
<div class="et_pb_column et_pb_column_4_4 et_pb_column_2  et_pb_css_mix_blend_mode_passthrough et-last-child">
<div class="et_pb_module et_pb_text et_pb_text_3  et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<h1><strong>What problems does a data mesh address?</strong></h1>
</div>
</div>
<div class="et_pb_module et_pb_text et_pb_text_4  et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<p>For data-driven enterprises, traditional centralized data architecture can quickly become a bottleneck for organizations that need to quickly ingest, process, and distribute data to add value and utility for business units.   </p>
<p>Outlined below is an overview of how a data mesh solves the common challenges that data-driven organizations have with traditional architecture.</p>
<p><strong>Centralized and siloed management and ownership</strong>: A data mesh takes a distributed and decentralized approach to data management at the source. Taking the burden off a central location to ingest, process, and serve data.</p>
<p><strong>Speed and agility:</strong> Decentralizing data management to the data source improves efficiency and reduces bottlenecks and bureaucratic red tape by having that data processed and distributed by domain or business unit source experts.</p>
<p><strong>Value and utility</strong>: Domain source experts consistently develop and serve data products to a platform based on organizational needs.</p>
<p>The previously mentioned themes and problems a data mesh addresses manifest into<span> </span><strong>four core principles</strong><span> </span>that interplay to make the mesh structure work within an organization.</p>
</div>
</div>
</div>
</div>
<div class="et_pb_row et_pb_row_3">
<div class="et_pb_column et_pb_column_4_4 et_pb_column_3  et_pb_css_mix_blend_mode_passthrough et-last-child">
<div class="et_pb_module et_pb_text et_pb_text_5  et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<h1>Data Mesh Principles</h1>
</div>
</div>
<div class="et_pb_module et_pb_text et_pb_text_6  et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<p>Data mesh leverages four principles that have a symbiotic relationship within a data-driven enterprise information management strategy.</p>
<p><strong>Domain Ownership:</strong> Domains are business units within an organization; these domains take on ownership and management of data in relation to specific domains. </p>
<p><strong>Data as a Product:</strong> Domain data is developed and shared as a product for internal and external consumers. </p>
<p><strong>Self-Serve Platform:</strong> The platform houses the products developed by the domains, allowing for product shareability and accessibility across domains and functions.</p>
<p><strong>Federated Governance:</strong> Governance of the domains, products, and platforms that are decentralized and managed by various stakeholders such as domain representatives, legal, compliance, and security.</p>
<p>The four principles are what make the data mesh architecture function. Likewise, each principle plays off the functionality of the other to address challenges that each one could create. </p>
<p>Here is how the interplay of the principles ensures optimal functionality within the organization. </p>
<p><em>Domain Ownership</em> depends on:</p>
<ul>
<li>Data as a Product to prevent data siloing</li>
<li>Self-Serve Platform to empower domain teams</li>
<li>Federated Governance to increase engagement and reduce domain isolation</li>
</ul>
<p><em>Data as a Product</em> depends on:</p>
<ul>
<li>Self-Serve Platform to reduce the cost of ownership</li>
<li>Federated Governance to get high-order value by interconnecting data products distributed by one or more domain teams</li>
</ul>
<p><em>Federated Governance</em> depends on the Self-Serve Platform to enforce a consistent and reliable policy.</p>
<p>While the four main principles provide the framework for the data mesh to function, a shift in organizational thinking and perspective is required to ensure the data mesh adheres to the principles.  </p>
<p>Below are<span> </span><strong>six categories</strong><span> </span>that are essential for a change in how the organization thinks about data management.</p>
<p><strong>Organization perspective</strong>: The organization moves from a centralized model to a decentralized data ownership model. In a decentralized model, the business domains are the data owners. </p>
<p><strong>Architectural perspective</strong>: The architecture moves beyond collecting data in lakes and monolithic warehouses. The architecture uses a distributed approach to accessing data products through standardized protocols. </p>
<p><strong>Technological perspective</strong>: Data evolves from being a consequence of code. Data and code become a cohesive and adaptable entity. </p>
<p><strong>Operational perspective</strong>: Data governance takes on a federated approach that relies on a computation system of policies replacing a centralized top-down and often human-administered approach. </p>
<p><strong>Principal perspective</strong>: Data evolves from an asset to a product for internal and external users. </p>
<p><strong>Infrastructure perspective</strong>: Removes the bifurcated approach to data and analytics that contains one system for analytics processing and another for transactional processing of operational applications. The infrastructure evolves to integrate transactional and analytical processing physically or virtually under a unified platform.</p>
</div>
</div>
</div>
</div>
<div class="et_pb_row et_pb_row_4">
<div class="et_pb_column et_pb_column_4_4 et_pb_column_4  et_pb_css_mix_blend_mode_passthrough et-last-child">
<div class="et_pb_module et_pb_text et_pb_text_7  et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<h1><strong>Expected Outcome</strong></h1>
</div>
</div>
<div class="et_pb_module et_pb_text et_pb_text_8  et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<p>Data mesh architecture is a solution for data-driven enterprises that consistently rely on data to evolve, improve, make decisions, and manage lines of business. A data mesh can allow an organization to stay agile, extract value and improve the utility of data, and allow for malleability as the organization evolves.</p>
</div>
</div>
</div>
</div>
</div>
</div>
<ul></ul></div>
			</div>
			</div>
				
				
				
				
			</div>
				
				
			</div>
<p>The post <a href="https://www.quatra.ai/blog/unleashing-value-with-a-decentralized-data-mesh/">Unleashing Value with a Decentralized Data Mesh</a> appeared first on <a href="https://www.quatra.ai">Quatra</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>How to Select the Best Technology for Data Quality Management</title>
		<link>https://www.quatra.ai/blog/how-to-select-the-best-technology-for-data-quality-management/</link>
		
		<dc:creator><![CDATA[Quatra Marketing]]></dc:creator>
		<pubDate>Wed, 03 Jul 2024 19:49:27 +0000</pubDate>
				<category><![CDATA[Data Quality]]></category>
		<guid isPermaLink="false">https://www.quatra.ai/?p=2680</guid>

					<description><![CDATA[<p>The post <a href="https://www.quatra.ai/blog/how-to-select-the-best-technology-for-data-quality-management/">How to Select the Best Technology for Data Quality Management</a> appeared first on <a href="https://www.quatra.ai">Quatra</a>.</p>
]]></description>
										<content:encoded><![CDATA[<div class="et_pb_section et_pb_section_6 et_section_regular" >
				
				
				
				
				
				
				<div class="et_pb_row et_pb_row_6 dbdb_default_mobile_width">
				<div class="et_pb_column et_pb_column_4_4 et_pb_column_6  et_pb_css_mix_blend_mode_passthrough et-last-child">
				
				
				
				
				<div class="et_pb_module et_pb_text et_pb_text_6  et_pb_text_align_left et_pb_bg_layout_light">
				
				
				
				
				<div class="et_pb_text_inner"><div class="et_pb_module et_pb_text et_pb_text_0  et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<h2><strong>Outdated data quality software can put you at a competitive disadvantage and drive up organizational costs.</strong></h2>
<p><span id="more-1601"></span></p>
<p><strong>Modern technology is proactive, avoids more costs, and mitigates more risk.</strong></p>
<p>In the following post, we will outline what to look for when selecting data quality technology as well as how to leverage artificial intelligence.</p>
<p>&nbsp;</p>
<h4>Proactive vs. Reactive</h4>
<p>Reactive data quality tools attempt to address errors after they are persisted to a data store. During transfer from an initial data store to a data lake or warehouse, data quality tools identify errors and attempt to resolve them to maintain a cleansed destination data store. This transfer may occur days or months after the data was originally created. Due to this lead time, the user is unlikely to recall details of a single record out of the thousands entered that month.</p>
<p>As a result, these errors may be handled with an elaborate remediation process that is part of a larger data governance program and council. The remediation workflow for a single error can involve technical support representatives, subject matter experts, data stewards, and data engineers. In a typical scenario, a support rep will document a problem then data stewards and engineers will investigate the cause. When the cause is identified, the data steward will discuss the preferred solution with subject matter experts of the data. The fix must be documented by the steward, presented for approval by the data governance council and then implemented in a data quality rule by a data engineer. The estimated cost of remediation for a single new error is $10,000. After this investment, the rule will provide automated quality enforcement for each recurrence of the same error.</p>
<p>Due to the costliness of reactively remediating errors and the risk of accidentally using bad data that was saved, a proactive solution is preferred. Proactive solutions prompt the creator of the data to fix the error at the time of entry. The cost to resolve an error at the time of entry, known as prevention cost, is estimated to be $1.<a href="https://www.f4.co/how-to-select-the-best-technology-for-data-quality-management/#_ftn1" name="_ftnref1"><span>[1]</span></a><span> </span>When the error is resolved by the creator and at the time of entry, the best resolution can be provided at the lowest cost. The user entering the data is not given the time or chance to forget the context of the entry. Poor data introduced by IoT devices are immediately identified and quarantined. A real-time approach at all points of data entry can avoid first time exposure.</p>
<p><span><em><a href="https://www.f4.co/how-to-select-the-best-technology-for-data-quality-management/#_ftnref1" name="_ftn1">[1]</a> Labovitz, G., Chang, Y.S., and Rosansky, V., 1992. Making Quality Work: A Leadership Guide for the Results-Driven Manager. John Wiley &amp;Sons, Hoboken, NJ.</em></span></p>
<p><span><em></em></span></p>
<p><span></span></p>
<table border="1" style="border-style: solid;">
<tbody>
<tr>
<td style="width: 447.625px;"><strong>REACTIVE</strong></td>
<td style="width: 301.198px;"><strong>PROACTIVE</strong></td>
</tr>
<tr>
<td style="width: 447.625px;">Incur risks and costs of first-time error exposure</td>
<td style="width: 301.198px;">Avoid first-time error exposure</td>
</tr>
<tr>
<td style="width: 447.625px;">$10,000 remediation cost</td>
<td style="width: 301.198px;">$1 remediation cost</td>
</tr>
<tr>
<td style="width: 447.625px;">Lengthy Solution</td>
<td style="width: 301.198px;">Immediate resolution</td>
</tr>
<tr></tr>
<tr>
<td style="width: 447.625px;">Delayed identification and remediation cause subpar solution due to limited information availability. Best case solution could be deleting an entire row of data</td>
<td style="width: 301.198px;">Best resolution possible because the data creator is providing the fix at the time of entry</td>
</tr>
</tbody>
</table>
</div>
</div>
<div class="et_pb_module et_pb_text et_pb_text_1  et_pb_text_align_left et_pb_bg_layout_light">
<div class="et_pb_text_inner">
<h3><strong></strong></h3>
<h4>Putting Artificial Intelligence to work</h4>
<p>Traditional data quality tools require rules to be created for each error that your enterprise has experienced or anticipates. Leveraging artificial intelligence and deep learning enables protection against errors you cannot predict. Preventing first time exposures to errors can save $10,000 or more per instance in remediation costs as well as preventing risk and much larger costs from decisions based on poor data. Unlike traditional data quality tools that require updates to rules when data requirements and validation changes, AI technologies can adapt to changes by learning from data and responses from users. This avoids the cost of maintaining a large set of data quality rules.</p>
</div>
</div>
<ul></ul></div>
			</div>
			</div>
				
				
				
				
			</div>
				
				
			</div>
<p>The post <a href="https://www.quatra.ai/blog/how-to-select-the-best-technology-for-data-quality-management/">How to Select the Best Technology for Data Quality Management</a> appeared first on <a href="https://www.quatra.ai">Quatra</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>How to Develop a Strategy for Data Quality Management</title>
		<link>https://www.quatra.ai/blog/how-to-develop-a-strategy-for-data-quality-management/</link>
		
		<dc:creator><![CDATA[Quatra Marketing]]></dc:creator>
		<pubDate>Wed, 03 Jul 2024 18:40:06 +0000</pubDate>
				<category><![CDATA[Data Quality]]></category>
		<guid isPermaLink="false">https://www.quatra.ai/?p=2644</guid>

					<description><![CDATA[<p>The post <a href="https://www.quatra.ai/blog/how-to-develop-a-strategy-for-data-quality-management/">How to Develop a Strategy for Data Quality Management</a> appeared first on <a href="https://www.quatra.ai">Quatra</a>.</p>
]]></description>
										<content:encoded><![CDATA[<div class="et_pb_section et_pb_section_7 et_section_regular" >
				
				
				
				
				
				
				<div class="et_pb_row et_pb_row_7 dbdb_default_mobile_width">
				<div class="et_pb_column et_pb_column_4_4 et_pb_column_7  et_pb_css_mix_blend_mode_passthrough et-last-child">
				
				
				
				
				<div class="et_pb_module et_pb_text et_pb_text_7  et_pb_text_align_left et_pb_bg_layout_light">
				
				
				
				
				<div class="et_pb_text_inner"><p>Does your organization have a strategy for data quality control? Improving data quality and reducing operational costs requires a solid plan. Establishing a strategy to manage data quality and set organizational standards is essential.</p>
<p>To develop and execute a data quality strategy, you need both a dedicated team and a well-defined process. Here&#8217;s how to structure your team and develop effective processes.</p>
<p><strong>Building Your Team</strong></p>
<p>Start by recruiting a data governance team. This team will set clear data definitions, create comprehensive policies, and oversee the documentation process. They ensure that data is collected, managed, and integrated properly across the organization.</p>
<p>Your team should include experts from various functions within the organization. A cross-functional team fosters a data-driven culture and creates data champions throughout the company.</p>
<p>At the helm of this team is typically the Chief Data Officer, who manages the team&#8217;s focus, communicates procedures, and monitors success. This leader forms an executive committee with leaders from different departments like finance and marketing. The committee is responsible for developing and overseeing data governance policies and processes.</p>
<p>Mid-level managers from different departments join the team to champion the data governance strategy and ensure collaboration across functions. They define processes, establish data quality metrics, and promote best practices.</p>
<p>Finally, the team assigns data owners, stewards, and users. Data owners manage compliance, administration, and access control. Data stewards act as intermediaries, interpreting data and creating reports. Users are responsible for entering and utilizing data daily and reporting any irregularities.</p>
<p><strong>Define Scope</strong></p>
<p>When implementing a data quality strategy, start with business processes that can benefit immediately from improved data quality. Choose projects with clear, identifiable issues. Begin with smaller projects for quick results, which will help garner executive support for larger initiatives. Each project should have a clear cost estimate and timeline.</p>
<p><strong>Map Data to Key Business Processes</strong></p>
<p>Once the initial project&#8217;s scope is defined, map the data flow within the process. Understanding how data moves through your organization and which business processes it impacts is crucial. This mapping helps you see the bigger picture and identify areas for improvement.</p>
<p><strong>Analyze Financial Implications</strong></p>
<p>After mapping the data flow, analyze the financial implications. Poor data quality might affect more areas than initially thought, revealing greater cost-saving opportunities. Collaborate with business management, accounting, and finance to ensure accuracy and gain support for future projects.</p>
<p><strong>Select the Right Technology</strong></p>
<p>Determine the technology needed for data quality evaluation. A diagnostics tool for data discovery and profiling is essential. This tool helps evaluate data set differences over time, quantify outcomes from cleansing, and estimate the project&#8217;s ROI.</p>
<p><strong>Determine Data Quality Metrics</strong></p>
<p>Choose metrics to capture the business impact of data quality initiatives. Metrics can range from simple to complex, depending on the data elements involved. Key indicators might include relevance, completeness, timeliness, accuracy, and consistency. Link these metrics to business initiatives to communicate the value of data quality projects.</p>
<p><strong>Establish Best Practices</strong></p>
<p>As you learn from each data quality project, establish best practices. Consistently using these practices will highlight the importance of data quality and influence a cultural shift within the organization. By demonstrating measurable results, you can advocate for the ongoing importance of data quality.</p>
<p>By following these steps, you can develop a robust strategy for managing data quality that supports your organization&#8217;s long-term success.</p>
<ul></ul></div>
			</div>
			</div>
				
				
				
				
			</div>
				
				
			</div>
<p>The post <a href="https://www.quatra.ai/blog/how-to-develop-a-strategy-for-data-quality-management/">How to Develop a Strategy for Data Quality Management</a> appeared first on <a href="https://www.quatra.ai">Quatra</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>4 Steps to Take Now to Determine the Quality of Your Data</title>
		<link>https://www.quatra.ai/blog/4-steps-to-take-now-to-determine-the-quality-of-your-data/</link>
		
		<dc:creator><![CDATA[Quatra Marketing]]></dc:creator>
		<pubDate>Wed, 03 Jul 2024 16:08:45 +0000</pubDate>
				<category><![CDATA[Data Quality]]></category>
		<guid isPermaLink="false">https://www.quatra.ai/?p=2632</guid>

					<description><![CDATA[<p>The post <a href="https://www.quatra.ai/blog/4-steps-to-take-now-to-determine-the-quality-of-your-data/">4 Steps to Take Now to Determine the Quality of Your Data</a> appeared first on <a href="https://www.quatra.ai">Quatra</a>.</p>
]]></description>
										<content:encoded><![CDATA[<div class="et_pb_section et_pb_section_8 et_section_regular" >
				
				
				
				
				
				
				<div class="et_pb_row et_pb_row_8 dbdb_default_mobile_width">
				<div class="et_pb_column et_pb_column_4_4 et_pb_column_8  et_pb_css_mix_blend_mode_passthrough et-last-child">
				
				
				
				
				<div class="et_pb_module et_pb_text et_pb_text_8  et_pb_text_align_left et_pb_bg_layout_light">
				
				
				
				
				<div class="et_pb_text_inner"><p>With executive confidence dwindling and the high costs associated with poor data, it&#8217;s crucial to ask: How bad is your data quality? More importantly, how can it be quantified? US businesses lose an estimated $611 billion annually due to data quality problems, and less than 33% of companies trust their data&#8217;s quality.</p>
<p>Understanding your data&#8217;s quality is essential. Check out our previous post on the cost of bad data for more insights. This post will guide you through a quick method to measure your data quality using a simple yet effective approach.</p>
<p>We recommend the Friday Afternoon Measurement (FAM) method to assess data quality (DQ). This method provides a clear, actionable score for your data quality. According to the Harvard Business Review, 47% of newly created data records contain at least one critical error, and only 3% of data quality scores were rated as “acceptable,” even by the loosest standards. These poor scores span across all business sectors, both private and public.</p>
<h3>How to Use the FAM Method</h3>
<p>Here’s how you can apply the FAM method in four straightforward steps to get a DQ score.<sup>1</sup></p>
<p><strong>Step 1:</strong> Gather the last 100 data records your team used, such as setting up a customer account or delivering a product.</p>
<p><strong>Step 2:</strong> Invite two or three colleagues who understand the data for a two-hour meeting.</p>
<p><strong>Step 3:</strong> Review each record with your colleagues, marking obvious errors. This process should be quick, usually taking no more than 30 seconds per record. In some cases, you may need to discuss whether an item is incorrect, but typically, errors like misspelled customer names or misplaced information will be immediately apparent.</p>
<p><strong>Step 4:</strong> Summarize the results in a spreadsheet. Add a “record perfect” column, marking “yes” if there are no errors and “no” if there are any.</p>
<p>To interpret the data, extrapolate the errors. For example, if only 40 out of 100 records are error-free, you have a 40% DQ score and a 60% error rate. This error rate can be quantified using the rule of 10, which states it costs ten times more to complete a task with defective data than with perfect data.</p>
<p>For instance, if your team must complete 100 units per day at a cost of $1.00 per unit with perfect data, the daily cost is $100. However, with only 40% perfect data, the total cost would be:<span class="katex"><span class="katex-mathml"></span></span></p>
<p>
<math xmlns="http://www.w3.org/1998/Math/MathML"><semantics><mrow><mtext>Total cost</mtext><mo>=</mo><mo stretchy="false">(</mo><mn>40</mn><mo>×</mo><mi mathvariant="normal">$</mi><mn>1.00</mn><mo stretchy="false">)</mo><mo>+</mo><mo stretchy="false">(</mo><mn>60</mn><mo>×</mo><mi mathvariant="normal">$</mi><mn>1.00</mn><mo>×</mo><mn>10</mn><mo stretchy="false">)</mo><mo>=</mo><mi mathvariant="normal">$</mi><mn>40</mn><mo>+</mo><mi mathvariant="normal">$</mi><mn>600</mn><mo>=</mo><mi mathvariant="normal">$</mi><mn>640</mn></mrow><annotation encoding="application/x-tex">\text{Total cost} = (40 \times \$1.00) + (60 \times \$1.00 \times 10) = \$40 + \$600 = \$640</annotation></semantics></math>
</p>
<p><span class="katex-html" aria-hidden="true"><span class="base"><span class="strut"></span><span class="mord text"><span class="mord"></span></span></span></span><span class="katex-html" aria-hidden="true"><span class="base"><span class="mord"></span></span></span></p>
<p>As shown, the cost increases over six times when the DQ score is not 100%. Reducing errors by 50% in this scenario would result in a 42% reduction in daily costs. Imagine the savings your organization could achieve by improving data quality.</p>
<p>By following these steps, you can gain a clearer understanding of your data quality and take actionable steps to improve it, saving time and resources for your organization.</p>
<p><sup>1</sup> Thomas Redman, Harvard Business Review, Assess Whether You Have a Data Quality Problem</p>
<ul></ul></div>
			</div>
			</div>
				
				
				
				
			</div>
				
				
			</div>
<p>The post <a href="https://www.quatra.ai/blog/4-steps-to-take-now-to-determine-the-quality-of-your-data/">4 Steps to Take Now to Determine the Quality of Your Data</a> appeared first on <a href="https://www.quatra.ai">Quatra</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>How to Determine the Cost of Bad Data and Gain Organizational Trust</title>
		<link>https://www.quatra.ai/blog/how-to-determine-the-cost-of-bad-data-and-gain-organizational-trust/</link>
		
		<dc:creator><![CDATA[Quatra Marketing]]></dc:creator>
		<pubDate>Wed, 03 Jul 2024 16:00:14 +0000</pubDate>
				<category><![CDATA[Data Quality]]></category>
		<guid isPermaLink="false">https://www.quatra.ai/?p=2626</guid>

					<description><![CDATA[<p>The post <a href="https://www.quatra.ai/blog/how-to-determine-the-cost-of-bad-data-and-gain-organizational-trust/">How to Determine the Cost of Bad Data and Gain Organizational Trust</a> appeared first on <a href="https://www.quatra.ai">Quatra</a>.</p>
]]></description>
										<content:encoded><![CDATA[<div class="et_pb_section et_pb_section_9 et_section_regular" >
				
				
				
				
				
				
				<div class="et_pb_row et_pb_row_9 dbdb_default_mobile_width">
				<div class="et_pb_column et_pb_column_4_4 et_pb_column_9  et_pb_css_mix_blend_mode_passthrough et-last-child">
				
				
				
				
				<div class="et_pb_module et_pb_text et_pb_text_9  et_pb_text_align_left et_pb_bg_layout_light">
				
				
				
				
				<div class="et_pb_text_inner"><p>Executives often harbor skepticism toward organizational data. Understanding the financial impact of bad data is a crucial first step in earning their trust.</p>
<p>&nbsp;</p>
<h3>Why Executives Distrust Their Data</h3>
<p>The value of enterprise data is determined by a <a href="/blog/11-key-indicators-to-determine-if-your-data-is-an-asset-or-liability/">variety of factors</a>, including accuracy, clarity, and community input. Any deficiencies in these areas can turn valuable data into a liability. Inaccurate data can distort summaries or bias models, leading to poor decisions, missed opportunities, damaged reputations, customer dissatisfaction, and increased risks and expenses.</p>
<p>Such errors can have a significant impact on business decisions and, ultimately, the bottom line. As data volumes and sources grow, managing quality becomes increasingly vital. Unfortunately, data errors are common, leading to widespread mistrust. According to a Harvard Business Review study, only 16% of managers fully trust their data.</p>
<p>A study by New Vantage Partners highlights more reasons for executive concern, especially among those leading data-driven transformations. It identifies cultural resistance, lack of organizational alignment, and agility as major barriers to adopting new data management technologies. Notably, 95% of surveyed executives cited cultural challenges, stemming from people and processes, as the main hurdles. There is a clear need for tools that can be easily adopted to improve data management processes.</p>
<p>&nbsp;</p>
<h3>The Cost of Poor Data</h3>
<p>Despite low trust in data quality, executives acknowledge its importance. Organizations are beginning to understand the high costs associated with poor data quality. Experian Plc. found that bad data costs companies 23% of revenue globally. IBM estimates the total cost of poor data quality to the U.S. economy at $3.1 trillion per year.</p>
<p>These costs primarily arise from initial errors that trigger costly reactionary responses. According to 451 Research, 44.5% of respondents manage data quality by identifying errors through reports and then taking corrective action. Another 37.5% rely on manual data cleansing processes.</p>
<p>Highly skilled data analysts spend valuable time manually fixing errors. Syncsort reports that 38% of data-driven analysts spend over 30% of their time on data remediation. Similarly, an MIT study found that knowledge workers waste up to 50% of their time on mundane quality issues, and for data scientists, this figure can reach 80%. This time could be better spent uncovering insights, solving complex business challenges, or generating revenue.</p>
<ul></ul></div>
			</div>
			</div>
				
				
				
				
			</div>
				
				
			</div>
<p>The post <a href="https://www.quatra.ai/blog/how-to-determine-the-cost-of-bad-data-and-gain-organizational-trust/">How to Determine the Cost of Bad Data and Gain Organizational Trust</a> appeared first on <a href="https://www.quatra.ai">Quatra</a>.</p>
]]></content:encoded>
					
		
		
			</item>
	</channel>
</rss>
