{
  "version": "https://jsonfeed.org/version/1.1",
  "title": "Blog",
  "language": "en",
  "home_page_url": "https://site-dyvenia.netlify.app/",
  "feed_url": "https://site-dyvenia.netlify.app/feed.json",
  "description": "",
  "authors": [
    {
      "name": "dyvenia"
    }
  ],
  "items": [{
      "id": "https://site-dyvenia.netlify.app/insights/flat-tables-vs-snowflake-semantic-models-the-ultimate-bi-data-debate/",
      "url": "https://site-dyvenia.netlify.app/insights/flat-tables-vs-snowflake-semantic-models-the-ultimate-bi-data-debate/",
      "title": "Flat Tables vs. Snowflake Semantic Models: The Ultimate BI Data Debate",
      "content_html": "<nav id=\"toc\" class=\"table-of-contents prose\"><ol><li class=\"flow\"><a href=\"#flat-tables-one-big-table-obt\">Flat Tables (One Big Table - OBT)</a></li><li class=\"flow\"><a href=\"#semantic-models\">Semantic Models</a></li><li class=\"flow\"><a href=\"#which-option-wins\">Which option wins?</a></li></ol></nav><p><span id=\"toc-skipped\" class=\"visually-hidden\"></span></p><div class=\"flow prose\"><p></p><p>Most data ends up in Business Intelligence (BI) reports. That’s no surprise: <strong>BI transforms raw data into actionable insights</strong>, helping businesses to make informed decisions, spot trends, and drive strategy with confidence.</p><p>But when it comes to structuring data for BI, one big question arises: Should data transformations happen within the BI tool, or should you create a flat table in the database?</p><p>This debate exists for good reason. Both approaches have their strengths, and neither is universally better. Flat tables often win points for being simple and fast to build. On the flip side, they can be inflexible and difficult to maintain at scale. Meanwhile, semantic models are praised for their scalability and reusability but tend to be harder to set up and govern consistently across teams.</p><p>The division often mirrors team roles: data analysts and engineers prefer flat tables for control and simplicity, while BI developers lean towards semantic models for flexibility and user-friendliness in tools like Power BI or Looker.</p><p>In this article, I’ll weigh the pros and cons of both approaches and share my perspective on the best way forward. Let’s dive in.</p><h2 id=\"flat-tables-one-big-table-obt\"><a href=\"#flat-tables-one-big-table-obt\" class=\"heading-anchor\">Flat Tables (One Big Table - OBT)</a></h2><p><is-land on:idle></is-land></p><dialog class=\"flow modal3\"><button autofocus class=\"button\">Close</button><picture><source type=\"image/webp\" srcset=\"/img/uogxHx8P6U-960.webp 960w, /img/uogxHx8P6U-1600.webp 1600w\" sizes=\"auto\"><img loading=\"lazy\" decoding=\"async\" src=\"/img/uogxHx8P6U-960.jpeg\" alt=\"one big table model\" width=\"1600\" height=\"662\" srcset=\"/img/uogxHx8P6U-960.jpeg 960w, /img/uogxHx8P6U-1600.jpeg 1600w\" sizes=\"auto\"></picture></dialog><button data-index=\"3\"><picture><source type=\"image/webp\" srcset=\"/img/uogxHx8P6U-960.webp 960w, /img/uogxHx8P6U-1600.webp 1600w\" sizes=\"auto\"><img loading=\"lazy\" decoding=\"async\" src=\"/img/uogxHx8P6U-960.jpeg\" alt=\"one big table model\" width=\"1600\" height=\"662\" srcset=\"/img/uogxHx8P6U-960.jpeg 960w, /img/uogxHx8P6U-1600.jpeg 1600w\" sizes=\"auto\"></picture></button><p></p><p>The flat table approach consolidates all your necessary data into a single, denormalized table before it reaches the BI tool. That means one table with all the columns - facts and attributes - already joined and ready for use. The simplest visualization of this structure is a table inside one Excel sheet. This method has several business-related advantages:</p><ul class=\"list\"><li><strong>Tool Agnostic</strong> – Since all transformations happen at the database level, the solution isn’t tied to a specific BI tool. Moving from Power BI to Qlik, Tableau, or another tool becomes much easier.</li><li><strong>Consistent Business Logic</strong> – Keeping all business rules and transformations in SQL ensures a single source of truth. This minimizes discrepancies between reports and eliminates the risk of conflicting KPIs.</li><li><strong>Version Control &amp; Auditing</strong> – Storing logic in SQL allows for version control, making it easier to track changes, roll back updates, and maintain data integrity.</li><li><strong>Simpler Data Access</strong> – Some users don’t need fancy visualizations—just clean, structured data they can extract and analyze. A flat table makes querying easier without needing to understand complex relationships between tables.</li><li><strong>Faster Development &amp; Maintenance</strong> – Once the OBT is set up, adding new KPIs and making modifications is often quicker and easier than managing complex BI tool transformations.</li></ul><p>But, there are some downsides to this approach:</p><ul class=\"list\"><li><strong>Storage &amp; Performance Concerns</strong></li></ul><p>One of the most common concerns with flat tables is performance and storage. Since they contain a lot of redundant data by design, they can become large and occasionally slower to query, especially when dealing with massive datasets or frequent refresh cycles. This is a real trade-off, but whether it actually becomes a problem depends on your system architecture and performance expectations. Flat tables may struggle in high-volume environments, when powering many dashboards, or when real-time performance is critical - particularly if you’re using databases that aren’t optimized for analytical workloads.</p><p>However, many of these limitations can be addressed with proper setup. Columnar databases like Redshift, BigQuery, or Snowflake only scan the columns needed, significantly reducing overhead. Loading only the required data into dashboards, applying smart distribution and sort keys, or using incremental refreshes can also improve performance. Most importantly, optimization should be driven by clear performance KPIs - not assumptions. If a dashboard loads within your defined threshold and users are satisfied, chasing “faster” just for the sake of it doesn’t bring real value.</p><ul class=\"list\"><li><strong>Redundant Data</strong></li></ul><p>Without normalization, data duplication is inevitable, increasing storage costs and complicating updates.</p><ul class=\"list\"><li><strong>Harder to Scale for Large Enterprise Systems</strong></li></ul><p>Maintaining a flat table becomes inefficient as datasets grow, leading to costly optimizations.</p><h2 id=\"semantic-models\"><a href=\"#semantic-models\" class=\"heading-anchor\">Semantic Models</a></h2><p><is-land on:idle></is-land></p><dialog class=\"flow modal4\"><button autofocus class=\"button\">Close</button><picture><source type=\"image/webp\" srcset=\"/img/8tkaVUWndM-960.webp 960w, /img/8tkaVUWndM-1600.webp 1600w\" sizes=\"auto\"><img loading=\"lazy\" decoding=\"async\" src=\"/img/8tkaVUWndM-960.jpeg\" alt=\"semantic data model\" width=\"1600\" height=\"720\" srcset=\"/img/8tkaVUWndM-960.jpeg 960w, /img/8tkaVUWndM-1600.jpeg 1600w\" sizes=\"auto\"></picture></dialog><button data-index=\"4\"><picture><source type=\"image/webp\" srcset=\"/img/8tkaVUWndM-960.webp 960w, /img/8tkaVUWndM-1600.webp 1600w\" sizes=\"auto\"><img loading=\"lazy\" decoding=\"async\" src=\"/img/8tkaVUWndM-960.jpeg\" alt=\"semantic data model\" width=\"1600\" height=\"720\" srcset=\"/img/8tkaVUWndM-960.jpeg 960w, /img/8tkaVUWndM-1600.jpeg 1600w\" sizes=\"auto\"></picture></button><p></p><p>A semantic model structures data into fact and dimension tables, optimizing storage and improving efficiency. Instead of consolidating all data into one table, facts (metrics like sales, revenue, or orders) are kept separate from dimensions (descriptive attributes like customers, products, or time periods), resulting in:</p><ul class=\"list\"><li><strong>Optimized Storage &amp; Performance</strong> – Since dimensions are stored separately and referenced via keys, this approach reduces redundancy and makes queries more efficient.</li><li><strong>Better Scalability</strong> – When dealing with large datasets, semantic models allow for efficient partitioning, incremental loading, and aggregation, making them ideal for complex analytical queries.</li><li><strong>More Flexibility in BI Tools</strong> – Many BI tools (like Power BI, Looker, and Tableau) have powerful modeling capabilities that work best when data is structured into facts and dimensions.</li><li><strong>Supports Advanced Analytics &amp; AI</strong> – Having a well-defined semantic layer allows for better integration with AI/ML models and predictive analytics, as structured relationships enhance data analysis capabilities.</li></ul><p>But, of course, there are also some trade-offs:</p><ul class=\"list\"><li><strong>Tool Dependency</strong> – Since modeling is done within the BI tool, switching tools may require rebuilding logic, which can be time-consuming and prone to errors.</li><li><strong>More Complex for End Users</strong> – Users who prefer simple, direct SQL queries may find it harder to navigate multiple tables and relationships.</li><li><strong>Risk of Inconsistent KPIs</strong> – When too many transformations occur within the BI tool, different reports might show different results for the same KPI, leading to confusion and mistrust in the data.</li><li><strong>Longer Development Time</strong> – Setting up a semantic model requires detailed planning, coordination, and maintenance, which can slow down report delivery compared to a flat table approach.</li><li><strong>Lack of Versioning &amp; Testing</strong> – Last but not least, most BI tools don’t offer robust support for version control, testing, or CI/CD pipelines. This makes governance and change management significantly more difficult and slower compared to managing logic in the database. Rolling out changes - especially at the organizational level - can quickly turn into a logistical headache.</li></ul><h2 id=\"which-option-wins\"><a href=\"#which-option-wins\" class=\"heading-anchor\">Which option wins?</a></h2><p>As always, it depends on the specific use case.</p><p>Flat tables will always be better when you care about version control, easier testing, and less dependency on specific BI tools. Semantic models will win when you rely heavily on your BI tool’s built-in modeling features.</p><p>But if I had to choose, I’d lean towards the OBT approach due to its consistency and business advantages. One of the most common issues I see is inconsistent KPIs across reports because too many transformations are done within the BI tool.</p><p>That said, I’m a fan of a hybrid approach. Combining both methodologies could be the best solution:</p><ul class=\"list\"><li><strong>Facts and essential KPIs</strong> should be calculated in SQL at the database level, ensuring a single source of truth.</li><li><strong>Master data (dimensions)</strong> can remain separate and be joined in the BI tool, optimizing storage and performance.</li><li><strong>Hybrid storage strategy</strong> for large datasets, storing historical or less frequently used data in a flat table while keeping active data in a semantic model for real-time reporting needs.</li><li><strong>Metadata &amp; governance layer</strong> ensuring business logic consistency, whether KPIs are defined in SQL or the BI tool.</li></ul><p>This approach leverages the best of both worlds - ensuring data consistency while maintaining flexibility and efficiency. It also allows businesses to scale efficiently while minimizing technical debt.</p></div>",
      "date_published": "2025-04-10T14:07:00Z"
    }
    ,{
      "id": "https://site-dyvenia.netlify.app/insights/driving-sustainability-with-data-improving-co%E2%82%82-emissions-reporting-across-supply-chains/",
      "url": "https://site-dyvenia.netlify.app/insights/driving-sustainability-with-data-improving-co%E2%82%82-emissions-reporting-across-supply-chains/",
      "title": "Driving Sustainability with Data: Improving CO₂ Emissions Reporting Across Supply Chains",
      "content_html": "<nav id=\"toc\" class=\"table-of-contents prose\"><ol><li class=\"flow\"><a href=\"#understanding-the-greenhouse-gas-protocol\">Understanding the Greenhouse Gas Protocol</a></li><li class=\"flow\"><a href=\"#why-sustainability-compliance-matters\">Why Sustainability Compliance Matters</a><ol><li class=\"flow\"><a href=\"#regulatory-requirements-for-emissions-reporting\">Regulatory Requirements for Emissions Reporting</a></li></ol></li><li class=\"flow\"><a href=\"#challenges-of-supply-chain-sustainability-reporting\">Challenges of Supply Chain Sustainability Reporting</a><ol><li class=\"flow\"><a href=\"#poor-data-quality\">Poor data quality</a></li><li class=\"flow\"><a href=\"#lack-of-a-solid-data-foundation\">Lack of a Solid Data Foundation</a></li><li class=\"flow\"><a href=\"#collecting-industry-carbon-factors\">Collecting Industry Carbon Factors</a></li><li class=\"flow\"><a href=\"#shortage-of-experts-who-understand-both-data-and-sustainability\">Shortage of Experts Who Understand Both Data and Sustainability</a></li></ol></li><li class=\"flow\"><a href=\"#how-to-improve-co₂-reporting-with-structured-data\">How to Improve CO₂ Reporting with Structured Data</a></li><li class=\"flow\"><a href=\"#conclusion-the-path-to-better-sustainability-reporting\">Conclusion: The Path to Better Sustainability Reporting</a></li></ol></nav><p><span id=\"toc-skipped\" class=\"visually-hidden\"></span></p><div class=\"flow prose\"><p></p><p>According to <a href=\"https://kpmg.com/dk/en/home/insights/2024/11/survey-of-sustainability-reporting-2024.html\" rel=\"noopener\">KPMG’s 2024 Survey of Sustainability Reporting</a>, <strong>80% of the 5,800 surveyed companies have established carbon reduction targets</strong>. This number highlights the increasing focus on sustainability but also the critical need for accurate emissions tracking. While setting targets is the first step, high-quality, reliable data is essential for consistent and accurate reporting across all areas of a business, particularly in supply chains.</p><h2 id=\"understanding-the-greenhouse-gas-protocol\"><a href=\"#understanding-the-greenhouse-gas-protocol\" class=\"heading-anchor\">Understanding the Greenhouse Gas Protocol</a></h2><p>The Greenhouse Gas (GHG) Protocol categorizes emissions into three scopes, each representing different sources of emissions within a company’s operations:</p><ul class=\"list\"><li><strong>Scope 1:</strong> Direct emissions from company-owned or controlled sources, such as fuel combustion in company vehicles or on-site manufacturing processes.</li><li><strong>Scope 2:</strong> Indirect emissions from purchased electricity, steam, heating, and cooling used by the company.</li><li><strong>Scope 3:</strong> Indirect emissions from the company’s broader value chain, including logistics, procurement, and material sourcing.</li></ul><p>While Scope 1 and 2 emissions are relatively easier to track, <strong>Scope 3</strong> emissions are more complex and often represent the largest share of a company’s total carbon footprint. For many manufacturing companies, the biggest portion of Scope 3 emissions comes from <strong>supply chain operations</strong>, which will be the focus of this article. We’ll explore the challenges of Scope 3 reporting and how improved data can support compliance and better decision-making.</p><h2 id=\"why-sustainability-compliance-matters\"><a href=\"#why-sustainability-compliance-matters\" class=\"heading-anchor\">Why Sustainability Compliance Matters</a></h2><p>Companies track their carbon footprint for various reasons, including <strong>environmental responsibility, corporate social responsibility</strong> (CSR), and good <strong>brand perception</strong>. Additionally, regulatory requirements play a key role in driving sustainability reporting. Accurate emissions data is crucial to maintaining compliance and meeting corporate sustainability goals.</p><h3 id=\"regulatory-requirements-for-emissions-reporting\"><a href=\"#regulatory-requirements-for-emissions-reporting\" class=\"heading-anchor\">Regulatory Requirements for Emissions Reporting</a></h3><p>Regulations surrounding emissions reporting are becoming increasingly strict globally. In the European Union, the <a href=\"https://alignedincentives.com/corporate-sustainability-regulations-a-roadmap-for-2025-and-beyond/\" rel=\"noopener\">Corporate Sustainability Reporting Directive (CSRD)</a> requires large companies to disclose detailed environmental, social, and governance (ESG) data, including greenhouse gas emissions. This mandate took effect in 2024 and requires reports published in 2025 to include this data. As a result, nearly <a href=\"https://normative.io/insight/csrd-explained/\" rel=\"noopener\">50,000 EU companies</a> now need to report their Scope 3 emissions.</p><p>Similar regulations exist in the UK (<a href=\"https://energy.drax.com/insights/streamlined-energy-and-carbon-reporting-framework/\" rel=\"noopener\">SECR</a>) and in over <a href=\"https://www.ecohedge.com/blog/emissions-reporting-navigating-the-essentials/\" rel=\"noopener\">40 other countries</a>, including the US, Canada, and Japan (alongside EU countries), all of which have implemented or are planning to implement corporate emissions disclosure requirements for greenhouse gases (GHG).</p><h2 id=\"challenges-of-supply-chain-sustainability-reporting\"><a href=\"#challenges-of-supply-chain-sustainability-reporting\" class=\"heading-anchor\">Challenges of Supply Chain Sustainability Reporting</a></h2><p>Effective Scope 3 emissions reporting comes with a number of obstacles, including data complexity, the frequent need for updates, and ensuring accuracy across a diverse range of<br>sources. Some of the key challenges are:</p><h3 id=\"poor-data-quality\"><a href=\"#poor-data-quality\" class=\"heading-anchor\"><strong>Poor data quality</strong></a></h3><p>One of the biggest challenges in Scope 3 emissions reporting is poor data quality, which can significantly impact the accuracy of sustainability metrics. Companies often deal with incomplete, inconsistent, or outdated data from multiple sources.</p><p>Examples of Poor Data Quality Issues include:</p><ul class=\"list\"><li>Inconsistent material weights or classifications: If different systems use different units (kg vs. tons), this can lead to miscalculations in material emissions.</li><li>Duplicate records in master data: Multiple entries for the same supplier or product cause redundancy, leading to errors in reporting.</li><li>Data silos and manual entries: Sustainability data is often scattered across ERP systems, spreadsheets, and supplier databases, leading to high risks of human error and inefficiency.</li></ul><h3 id=\"lack-of-a-solid-data-foundation\"><a href=\"#lack-of-a-solid-data-foundation\" class=\"heading-anchor\"><strong>Lack of a Solid Data Foundation</strong></a></h3><p>The <strong>poor data quality, complexity, and diversity</strong> of data required for Scope 3 emissions reporting make it challenging to get a full and accurate picture. To track emissions properly, companies need to collect a wide range of information, including:</p><ul class=\"list\"><li><strong>Vendor master data</strong> (such as LFA1 table in SAP)</li><li><strong>Material master data</strong> (such as MARA table in SAP)</li><li><strong>Plant master data</strong> to be able to track shipping routes</li><li><strong>Purchasing data</strong> on the Purchase Order document level</li></ul><p>Many sources of data make data foundations extremely difficult to establish. Not all the data comes from the organization’s internal sources but also requires collaboration with other parties (e.g., vendors).</p><p>Another major challenge arises from geographical data. Vendor and Plant locations change frequently, leading to potential errors in emissions calculations. For example, if a vendor relocates its plant, this change must be automatically reflected in <strong>route calculations</strong> to maintain accurate CO₂ estimates.</p><p>Solid data foundations aren’t just about collecting the right inputs - they also require clearly defined, well-structured metrics. The way data is organized should support easy and consistent metric implementation later on. Without this, even the best-quality data can become difficult to interpret or apply reliably across reports, especially when calculating Scope 3 emissions that rely on complex, multi-source inputs.</p><p><is-land on:idle></is-land></p><dialog class=\"flow modal1\"><button autofocus class=\"button\">Close</button><picture><source type=\"image/webp\" srcset=\"/img/v_-aSuksEX-960.webp 960w, /img/v_-aSuksEX-1600.webp 1600w\" sizes=\"auto\"><img loading=\"lazy\" decoding=\"async\" src=\"/img/v_-aSuksEX-960.jpeg\" alt width=\"1600\" height=\"673\" srcset=\"/img/v_-aSuksEX-960.jpeg 960w, /img/v_-aSuksEX-1600.jpeg 1600w\" sizes=\"auto\"></picture></dialog><button data-index=\"1\"><picture><source type=\"image/webp\" srcset=\"/img/v_-aSuksEX-960.webp 960w, /img/v_-aSuksEX-1600.webp 1600w\" sizes=\"auto\"><img loading=\"lazy\" decoding=\"async\" src=\"/img/v_-aSuksEX-960.jpeg\" alt width=\"1600\" height=\"673\" srcset=\"/img/v_-aSuksEX-960.jpeg 960w, /img/v_-aSuksEX-1600.jpeg 1600w\" sizes=\"auto\"></picture></button><p></p><h3 id=\"collecting-industry-carbon-factors\"><a href=\"#collecting-industry-carbon-factors\" class=\"heading-anchor\"><strong>Collecting Industry Carbon Factors</strong></a></h3><p>A common approach in logistics is estimating emissions using a simple emissions factor model based on CO₂ emitted per kilometer traveled for different transport modes. However, <strong>material-related</strong> emissions are far more complex. This requires gathering and applying carbon factors from multiple sources, such as:</p><ul class=\"list\"><li><strong>EPD (Environmental Product Declarations):</strong> product-specific, 3rd party-verified environmental data per product category</li><li><strong>CDP (Carbon Disclosure Project):</strong> vendor-specific CO₂ footprints</li><li><strong>EPA (Environmental Protection Agency):</strong> standardized US-based emission factors for industries (e.g., transport, energy, waste)</li><li><strong>Industry average data:</strong> generic emission benchmarks based on sector-wide studies for certain materials that lack specific CO₂ data</li></ul><p>These carbon factors need to be regularly updated, mapped to specific materials, and often collected manually, making the process cumbersome and prone to errors.</p><h3 id=\"shortage-of-experts-who-understand-both-data-and-sustainability\"><a href=\"#shortage-of-experts-who-understand-both-data-and-sustainability\" class=\"heading-anchor\"><strong>Shortage of Experts Who Understand Both Data and Sustainability</strong></a></h3><p>Only a few specialists can bridge the gap between <strong>sustainability regulations</strong> and <strong>data analytics</strong>. Applying the correct emission factors (particularly for specific materials) requires deep domain knowledge and industry-specific expertise. This skills gap makes it difficult for many organizations to ensure accurate emissions reporting. To accurately report emissions, professionals need a blend of sustainability knowledge and data expertise. The challenge is that few people have both skill sets - they either specialize in environmental science or in data analytics, but not both.</p><h2 id=\"how-to-improve-co₂-reporting-with-structured-data\"><a href=\"#how-to-improve-co₂-reporting-with-structured-data\" class=\"heading-anchor\">How to Improve CO₂ Reporting with Structured Data</a></h2><p>The challenges around Scope 3 emissions reporting show just how complicated it can get ‒ different sources, frequent updates, and a lot of moving parts. Without a reliable data foundation, companies risk inaccurate reporting, which can lead to poor decisions and compliance issues.</p><p>This is where a structured system like a <strong>sustainability mart</strong> can help.</p><p>When it comes to structuring such a mart, there are two common approaches:</p><ol class=\"list\"><li><strong>Flat Table Approach:</strong> This method consolidates all necessary data into a single, denormalized table. It ensures consistency, simplifies development, and is tool-agnostic - making it a favorite among data teams. However, it can become harder to manage at scale, with potential performance and storage challenges depending on dataset size and architecture.</li><li><strong>Snowflake Schema Approach in BI Tools</strong> (e.g., Power BI): In this approach, data is organized into fact and dimension tables, which are related through keys. While it’s more efficient for large datasets and complex queries, it introduces tool dependency and leads to inconsistent KPIs if governance isn’t tightly controlled.</li></ol><p>Personally, I would opt for a flat table approach rather than creating a semantic model inside a BI tool. I’ve explained this choice further in <a href=\"https://site-dyvenia.netlify.app/insights/flat-tables-vs-snowflake-semantic-models-the-ultimate-bi-data-debate/\" rel=\"noopener\">another article</a>, but with sustainability data, the choice between these two approaches is even clearer. Master data plays a crucial role in metric calculations, which makes it almost impossible to keep it outside the mart. A sustainability mart integrates critical data sources into one system, enabling accurate calculations and reporting of ESG metrics. Such a mart can include a variety of relevant components, for example:</p><p><is-land on:idle></is-land></p><dialog class=\"flow modal2\"><button autofocus class=\"button\">Close</button><picture><source type=\"image/webp\" srcset=\"/img/4vihiXkBRn-960.webp 960w, /img/4vihiXkBRn-1600.webp 1600w\" sizes=\"auto\"><img loading=\"lazy\" decoding=\"async\" src=\"/img/4vihiXkBRn-960.jpeg\" alt width=\"1600\" height=\"706\" srcset=\"/img/4vihiXkBRn-960.jpeg 960w, /img/4vihiXkBRn-1600.jpeg 1600w\" sizes=\"auto\"></picture></dialog><button data-index=\"2\"><picture><source type=\"image/webp\" srcset=\"/img/4vihiXkBRn-960.webp 960w, /img/4vihiXkBRn-1600.webp 1600w\" sizes=\"auto\"><img loading=\"lazy\" decoding=\"async\" src=\"/img/4vihiXkBRn-960.jpeg\" alt width=\"1600\" height=\"706\" srcset=\"/img/4vihiXkBRn-960.jpeg 960w, /img/4vihiXkBRn-1600.jpeg 1600w\" sizes=\"auto\"></picture></button><p></p><p>With these data points in one integrated system, companies can accurately calculate material and logistics-related emissions. By leveraging a sustainability mart, companies can:</p><ul class=\"list\"><li><strong>Generate standardized reports for Sustainability Management</strong> to ensure consistent and comparable reporting across departments and reporting periods, simplifying compliance with regulatory requirements</li><li><strong>Monitor trends through historical snapshots</strong> to help businesses track emission reduction progress, identify inefficiencies, and make data-driven sustainability decisions</li><li><strong>Automatically track and apply changes in master data</strong> (e.g., updates to vendor or material information) to prevent outdated or incorrect sustainability data</li></ul><h2 id=\"conclusion-the-path-to-better-sustainability-reporting\"><a href=\"#conclusion-the-path-to-better-sustainability-reporting\" class=\"heading-anchor\">Conclusion: The Path to Better Sustainability Reporting</a></h2><p>Accurate sustainability reporting starts with <strong>clean, structured data</strong>. For companies, mistakes in emissions reporting can result in both regulatory risks and growing stakeholder scrutiny. While domain knowledge remains crucial, planning and implementing a data-driven strategy is the key to long-term success. Ultimately, a well-designed sustainability data system can transform reporting from a compliance burden into a strategic asset, providing companies with valuable insights into their operations and sustainability performance.</p></div>",
      "date_published": "2025-04-10T13:20:00Z"
    }
    ,{
      "id": "https://site-dyvenia.netlify.app/insights/a-simple-approach-to-master-data-management-to-unify-metrics-and-insights/",
      "url": "https://site-dyvenia.netlify.app/insights/a-simple-approach-to-master-data-management-to-unify-metrics-and-insights/",
      "title": "A Simple Approach to Master Data Management to Unify Metrics and Insights",
      "content_html": "<nav id=\"toc\" class=\"table-of-contents prose\"><ol><li class=\"flow\"><a href=\"#what-is-master-data-management\">What is Master Data Management?</a></li><li class=\"flow\"><a href=\"#typical-pitfalls-of-master-data-management-projects\">Typical pitfalls of master data management projects</a></li><li class=\"flow\"><a href=\"#avoiding-pitfalls-in-data-management-projects\">Avoiding pitfalls in data management projects</a></li><li class=\"flow\"><a href=\"#the-collaborative-master-data-approach\">The Collaborative Master Data Approach</a></li><li class=\"flow\"><a href=\"#the-consolidated-master-data-approach\">The Consolidated Master Data Approach</a></li><li class=\"flow\"><a href=\"#the-hybrid-approach-combining-both-the-collaborative-and-consolidated-approach\">The Hybrid Approach: Combining both the Collaborative and Consolidated Approach</a></li><li class=\"flow\"><a href=\"#conclusion\">Conclusion</a></li></ol></nav><p><span id=\"toc-skipped\" class=\"visually-hidden\"></span></p><div class=\"flow prose\"><p></p><p>Master data management (MDM) often sparks debate. It’s a frequently proposed budget item that rarely gets approved, yet it’s a fundamental reason business leaders struggle to achieve granular and unified metrics. This article aims to demystify MDM by showing that it doesn’t always have to involve big budgets or complex tools. Instead, we will define master data and explore two straightforward approaches to MDM that can be implemented without lengthy multi-year technology roadmaps. While the complexity of MDM varies by case, we believe most organizations can achieve significant improvements by following a few practical steps.</p><h3 id=\"what-is-master-data-management\"><a href=\"#what-is-master-data-management\" class=\"heading-anchor\"><strong>What is Master Data Management?</strong></a></h3><p>Master data management is the process of creating and maintaining a single source of truth for critical business data. It ensures consistency, accuracy, and accessibility across an organization’s operations. By harmonizing data, MDM eliminates discrepancies, supports better reporting, and improves decision-making processes.</p><p>The most commonly used dimensions in manufacturing company metrics include:</p><ul class=\"list\"><li>Product Hierarchies</li><li>Territory Hierarchies</li><li>Key Account Management Groupings</li><li>Industry Codes</li><li>Legal Entities</li><li>Regions and Countries</li><li>Fiscal Time Dimensions</li></ul><h3 id=\"typical-pitfalls-of-master-data-management-projects\"><a href=\"#typical-pitfalls-of-master-data-management-projects\" class=\"heading-anchor\"><strong>Typical pitfalls of master data management projects</strong></a></h3><p>Many master data management projects fail to achieve their intended goals. In some cases, these projects never even get off the ground because they fail to clearly communicate their value to business leaders. Based on our observations, there are three common pitfalls when companies approach master data:</p><ul class=\"list\"><li><strong>Boiling the Ocean</strong>: Attempting to solve all data problems at once instead of prioritizing critical areas often leads to overwhelming complexity and project delays.</li><li><strong>Overemphasis on Tools</strong>: Focusing too much on selecting and implementing expensive tools can distract from addressing foundational issues like data quality and governance.</li><li><strong>Neglecting Data Ownership</strong>: Failing to establish clear accountability for different aspects of master data leads to fragmented efforts and inconsistent results. For example, defining which teams are responsible for maintaining specific data dimensions is often overlooked but important for success.</li></ul><h3 id=\"avoiding-pitfalls-in-data-management-projects\"><a href=\"#avoiding-pitfalls-in-data-management-projects\" class=\"heading-anchor\"><strong>Avoiding pitfalls in data management projects</strong></a></h3><p>To avoid common pitfalls in data management, it is essential to break down the project into two distinct approaches: a collaborative approach and a consolidated approach. While this framework is not our own invention, it has been a reliable methodology we have employed successfully for many years.</p><p>By dividing master data management work into these two approaches, organizations can tackle tasks more effectively. This structured method allows for a clearer focus on priorities and better resource allocation. Moreover, starting with small, incremental steps—rather than attempting a large, comprehensive project—is often a more practical and achievable way to ensure success in master data management.</p><h3 id=\"the-collaborative-master-data-approach\"><a href=\"#the-collaborative-master-data-approach\" class=\"heading-anchor\"><strong>The Collaborative Master Data Approach</strong></a></h3><p>The collaborative approach to MDM is ideal for managing lookup data and creating custom groupings. This method relies on an application with a user-friendly interface that enables teams to collaborate effectively on master data management tasks. Key features include:</p><ul class=\"list\"><li><strong>Approval Workflows</strong>: Changes to lookups and master data are managed through structured approval processes, ensuring accuracy and accountability.</li><li><strong>Real-Time Collaboration</strong>: Teams can work together to refine and update master data in real-time, promoting consistency across departments and improving alignment.</li></ul><p><strong>Advantages</strong></p><ul class=\"list\"><li><strong>Enhanced Flexibility</strong>: This approach is particularly beneficial for organizations that need adaptable and active user involvement in managing dimensions.</li><li><strong>Simplicity</strong>: Many teams already utilize this approach informally by maintaining lookup tables in spreadsheets. At its core, this method can be as simple as sharing a spreadsheet via tools like SharePoint.</li></ul><p><strong>Disadvantages</strong></p><ul class=\"list\"><li><strong>Increased Manual Effort</strong>: This approach requires manual intervention, including manual approval processes, which means at least two users must be involved in managing the master data.</li><li><strong>Reduced Auditability</strong>: If changes are not properly tracked—such as when data is stored in untracked spreadsheets—it becomes challenging to trace historical changes and reconstruct how master data appeared at earlier points in time.</li></ul><p>This type of setup has proven effective for defining dimensions critical to measuring business unit or sales performance. For instance, teams managing sales territories have achieved significant success by aligning their territory dimension reporting with their sales organization structures, ensuring consistency with sales incentive plan reporting.</p><h3 id=\"the-consolidated-master-data-approach\"><a href=\"#the-consolidated-master-data-approach\" class=\"heading-anchor\"><strong>The Consolidated Master Data Approach</strong></a></h3><p>In contrast, the consolidated approach is tailored for application data that is maintained directly at the source. This method involves aggregating data from multiple applications into a central repository without making any modifications. Its key characteristics include:</p><p><strong>Advantages:</strong></p><ul class=\"list\"><li><strong>Data Integrity Preserved:</strong> Data remains unchanged in its original source, ensuring that users can only modify it within the application of origin. This preserves the accuracy and reliability of the master data.</li><li><strong>Reduced Manual Intervention:</strong> By consolidating data from trusted sources, this approach minimizes the need for manual adjustments, thereby reducing errors and enhancing efficiency.</li></ul><p><strong>Disadvantages:</strong></p><ul class=\"list\"><li><strong>Complex Decentralization:</strong> Because master data is distributed across multiple applications, coordinating updates and maintaining consistency can be challenging. This issue is particularly acute in organizations with multiple ERP systems managed by diverse local IT teams.</li><li><strong>Restricted Modifications:</strong> Business leaders often require minor adjustments or recoding of dimensions. In this approach, such changes must be negotiated with the production system of record, which can slow down responsiveness and adaptability.</li></ul><p>The consolidated approach is most effective when source data from applications is of high quality and does not require frequent adjustments by users. It is ideal for organizations seeking streamlined operations where data reliability and minimal intervention are priorities.</p><h3 id=\"the-hybrid-approach-combining-both-the-collaborative-and-consolidated-approach\"><a href=\"#the-hybrid-approach-combining-both-the-collaborative-and-consolidated-approach\" class=\"heading-anchor\"><strong>The Hybrid Approach: Combining both the Collaborative and Consolidated Approach</strong></a></h3><p>Considering that both collaborative and consolidated master data approaches have their own unique advantages and disadvantages, a hybrid strategy can often be the most effective. By using these approaches together:</p><ul class=\"list\"><li><strong>Maintain Complex Master Data Within Applications</strong>: Organizations can retain more complex master data within applications, potentially avoiding the need for a costly and intricate MDM system altogether.</li><li><strong>Enable Business Flexibility</strong>: Business leaders gain the flexibility to add their own local dimensions, improving the visibility of custom dimensions in reporting and enhancing auditability and change management processes.</li><li><strong>Clarify Ownership Roles</strong>: A hybrid approach facilitates clear ownership of master data. For example, Customer data quality in the ERP can be managed by the Customer Service team, while the Sales organization takes responsibility for Sales Territory structures, ensuring accountability and streamlined operations.</li></ul><p><strong>If You Want Real Change in Master Data Quality, You Need Visibility</strong></p><p>Once, I was in a meeting with the CEO of a Business Unit discussing the sales funnel. The conversation was not going well because the data was, quite simply, poor. However, the CEO said that he would be willing to continue reviewing the sales funnel reports even if the data was flawed, as long as it was clear which data was problematic and why.</p><p>The CEO made a smart move here. Instead of insisting that everything be flawless, he told his leadership team that he was willing to work with imperfect data, provided it was clear why the data was flawed and who was responsible for addressing the issues.</p><p>Data quality metrics might seem like a dull report that no business leader would bother opening. However, imagine a business report on key metrics like orders, sales, or inventory that includes a data quality score. Such a report would undoubtedly capture attention. This kind of visibility not only ensures the report gets reviewed but also drives accountability. It motivates people to address data issues, whether in the source system (consolidated master data) or within the application (collaborative master data).</p><h3 id=\"conclusion\"><a href=\"#conclusion\" class=\"heading-anchor\"><strong>Conclusion</strong></a></h3><p>Master data management plays a pivotal role in unifying metrics and insights for informed decision-making. By understanding the collaborative and consolidated approaches, organizations can select the strategy that best aligns with their needs and data quality standards. Whether fostering collaboration for lookup data or consolidating application data for streamlined efficiency, effective MDM can transform how businesses leverage their data assets.</p></div>",
      "date_published": "2025-01-14T00:00:00Z"
    }
    ,{
      "id": "https://site-dyvenia.netlify.app/insights/what-as-a-data-mart-and-what-challenges-does-it-solve/",
      "url": "https://site-dyvenia.netlify.app/insights/what-as-a-data-mart-and-what-challenges-does-it-solve/",
      "title": "What As a Data Mart? And What Challenges Does It Solve?",
      "content_html": "<nav id=\"toc\" class=\"table-of-contents prose\"><ol><li class=\"flow\"><a href=\"#what-is-a-data-mart\">What Is a Data Mart?</a></li><li class=\"flow\"><a href=\"#this-makes-it-challenging-to-benchmark-typical-business-as-usual-travel-costs-and-determine-how-much-can-be-cut-without-negatively-impacting-operations\">This makes it challenging to benchmark typical business-as-usual travel costs and determine how much can be cut without negatively impacting operations.</a></li><li class=\"flow\"><a href=\"#how-data-marts-enable-business-leaders-with-good-metrics\">How Data Marts Enable Business Leaders with Good Metrics</a></li><li class=\"flow\"><a href=\"#common-pitfalls-in-achieving-reliable-metrics\">Common Pitfalls in Achieving Reliable Metrics</a></li><li class=\"flow\"><a href=\"#conclusion\">Conclusion</a></li></ol></nav><p><span id=\"toc-skipped\" class=\"visually-hidden\"></span></p><div class=\"flow prose\"><p></p><p>Business leaders are always under constant pressure to improve operational efficiency, maintain product quality, and optimize financial performance. Good data and, more importantly, good metrics are important for effectively addressing these challenges.</p><p>A critical yet often underutilized tool for achieving reliable business metrics is the <strong>data mart</strong>. In this article, we will explore what a data mart is and how it functions. We will also define what is a “<strong>good metric</strong>” and examine how data marts enable business leaders with metrics that enhance visibility, control, and predictability—ultimately <strong>improving business performance</strong>.</p><h3 id=\"what-is-a-data-mart\"><a href=\"#what-is-a-data-mart\" class=\"heading-anchor\">What Is a Data Mart?</a></h3><p>A data mart is a specialized data repository typically stored in a SQL database designed to serve the specific needs of a particular department, business unit, or function. It contains data relevant to its intended audience, structured to provide streamlined access to insights.</p><p>Key characteristics of a data mart include:</p><ul class=\"list\"><li><strong>Focus</strong>: Designed to support specific business areas, such as top-line (orders, sales, backlog, standard cost &amp; margins), purchasing (supplier performance, material productivity), inventory (aging, excess &amp; obsolescence), and others.</li><li><strong>Efficiency</strong>: Simplifies access to key data, reducing the time needed to generate insights. This efficiency stems from having all the logic centralized in one place rather than spread across various reports and dashboards.</li><li><strong>Usability</strong>: Usability is high because standard connections allow users to utilize familiar reporting tools like PowerBI, Tableau, or Excel. This ensures accessibility, making it easy to query data and present it in user-friendly formats.</li></ul><p>By organizing data into easily digestible segments, data marts bridge the gap between raw data and actionable metrics, enabling business leaders to make informed decisions faster.</p><p><is-land on:idle></is-land></p><dialog class=\"flow modal8\"><button autofocus class=\"button\">Close</button><picture><source type=\"image/webp\" srcset=\"/img/wUbCTCJWfm-960.webp 960w, /img/wUbCTCJWfm-1600.webp 1600w\" sizes=\"auto\"><img loading=\"lazy\" decoding=\"async\" src=\"/img/wUbCTCJWfm-960.jpeg\" alt=\"what is a data mart\" width=\"1600\" height=\"796\" srcset=\"/img/wUbCTCJWfm-960.jpeg 960w, /img/wUbCTCJWfm-1600.jpeg 1600w\" sizes=\"auto\"></picture></dialog><button data-index=\"8\"><picture><source type=\"image/webp\" srcset=\"/img/wUbCTCJWfm-960.webp 960w, /img/wUbCTCJWfm-1600.webp 1600w\" sizes=\"auto\"><img loading=\"lazy\" decoding=\"async\" src=\"/img/wUbCTCJWfm-960.jpeg\" alt=\"what is a data mart\" width=\"1600\" height=\"796\" srcset=\"/img/wUbCTCJWfm-960.jpeg 960w, /img/wUbCTCJWfm-1600.jpeg 1600w\" sizes=\"auto\"></picture></button><p></p><h3 id=\"this-makes-it-challenging-to-benchmark-typical-business-as-usual-travel-costs-and-determine-how-much-can-be-cut-without-negatively-impacting-operations\"><a href=\"#this-makes-it-challenging-to-benchmark-typical-business-as-usual-travel-costs-and-determine-how-much-can-be-cut-without-negatively-impacting-operations\" class=\"heading-anchor\">This makes it challenging to benchmark typical business-as-usual travel costs and determine how much can be cut without negatively impacting operations.</a></h3><p>Not all metrics are created equal. A good metric stands out by its ability to provide meaningful, actionable insights that align with organizational goals. To evaluate the quality of your metrics, consider the following attributes:</p><ol class=\"list\"><li><strong>Compliant</strong>: The metric adheres to industry standards and internal governance policies, ensuring accuracy and reliability. This is especially important for calculations affecting key financial figures such as profitability, working capital, and reserves, where adherence to internal policies ensures consistency and trustworthiness.</li><li><strong>Usable</strong>: It is accessible, easy to interpret, and designed for practical application by its intended audience.</li><li><strong>Granular</strong>: The metric provides detailed information, enabling deep drill-down analysis to uncover root causes and trends. Achieving granularity is challenging but essential, as it enhances the quality of insights and directly influences decision-making through accurate explanations and narratives.</li><li><strong>Aligned</strong>: It supports strategic objectives and connects to critical financial or operational outcomes like the P&amp;L, balance sheet, or product quality. All stakeholders agree on the metric’s calculation and meaning, ensuring organizational clarity and alignment.</li><li><strong>Comparable</strong>: A good metric is consistent across contexts and systems, enabling meaningful comparisons between departments, time periods, or benchmarks. This involves tracking events like organizational changes, sales territory changes, standard cost updates, and currency fluctuations. When changes occur, metrics should be easily reinstated, and different versions of the same metric should remain accessible for comparison across changes.</li></ol><p>To illustrate the impact of missing metric attributes on business performance, let’s consider a case study. Imagine a scenario where a CEO asks the CFO to temporarily reduce travel expenses to meet profitability targets. Now, suppose the CFO has to work with a Travel Expense metric lacking the five key attributes. Here’s what happens:</p><ol class=\"list\"><li><strong>Non-Compliant</strong>: Travel expenses are incorrectly booked in cost categories, making it difficult to calculate the cost savings that could be generated.</li><li><strong>Unusable</strong>: Details like hotel, flight, and car rental bookings are locked in the travel application, requiring weeks for analysts to compile raw data for a detailed analysis.</li><li><strong>Lacking Granularity</strong>: Without granular insights, such as travel by department or business function, it is impossible to quickly identify cost cuts without risking business operations.</li><li><strong>Misaligned</strong>: With travel expenses not booked in the correct cost categories, business leaders across functions will disagree on the actual figures, wasting time aligning on a baseline.</li><li><strong>Non-Comparable</strong>: Because travel expenses are not easily available, year-over-year analysis becomes difficult. This makes it challenging to benchmark typical business-as-usual travel costs and determine how much can be cut without negatively impacting operations.</li></ol><p>Metrics that embody these attributes not only enhance decision-making but also build trust among stakeholders by providing a clear, auditable trail from raw data to actionable insights.</p><h3 id=\"how-data-marts-enable-business-leaders-with-good-metrics\"><a href=\"#how-data-marts-enable-business-leaders-with-good-metrics\" class=\"heading-anchor\">How Data Marts Enable Business Leaders with Good Metrics</a></h3><p>Manufacturing companies operate in complex environments where good metrics are critical for visibility, control, and success. Data marts make good metrics possible through five essential activities:</p><ol class=\"list\"><li><strong>Centralized flat tables stored in SQL databases:</strong> Data marts are ready to use and provide a simplified format for querying data. These tables streamline access to key metrics by organizing raw data into structured and understandable formats, ensuring faster and more reliable insights for decision-makers.</li><li><strong>Centralized Metric Calculation Logic</strong>: Data marts centralize the logic behind metric calculations, ensuring consistency and reducing errors.</li><li><strong>Automated Processing</strong>: Data pipelines powering data marts fully automate the end-to-end transformation process, turning raw data into trusted and actionable metrics.</li><li><strong>Integration of Data Sources</strong>: Data marts consolidate data from multiple sources, both legacy and modern, into single, auditable metrics.</li><li><strong>Monitored Ingestions and Transformations</strong>: Data pipelines powering data marts are fully monitored to ensure that the metrics business leaders rely on are always available. When issues arise, effective communication is promptly initiated, and resolutions are implemented quickly to minimize disruptions.</li></ol><p><is-land on:idle></is-land></p><dialog class=\"flow modal9\"><button autofocus class=\"button\">Close</button><picture><source type=\"image/webp\" srcset=\"/img/VqybQdyCWL-960.webp 960w, /img/VqybQdyCWL-1600.webp 1600w\" sizes=\"auto\"><img loading=\"lazy\" decoding=\"async\" src=\"/img/VqybQdyCWL-960.jpeg\" alt=\"unified metrics\" width=\"1600\" height=\"796\" srcset=\"/img/VqybQdyCWL-960.jpeg 960w, /img/VqybQdyCWL-1600.jpeg 1600w\" sizes=\"auto\"></picture></dialog><button data-index=\"9\"><picture><source type=\"image/webp\" srcset=\"/img/VqybQdyCWL-960.webp 960w, /img/VqybQdyCWL-1600.webp 1600w\" sizes=\"auto\"><img loading=\"lazy\" decoding=\"async\" src=\"/img/VqybQdyCWL-960.jpeg\" alt=\"unified metrics\" width=\"1600\" height=\"796\" srcset=\"/img/VqybQdyCWL-960.jpeg 960w, /img/VqybQdyCWL-1600.jpeg 1600w\" sizes=\"auto\"></picture></button><p></p><p>These capabilities make data marts indispensable for driving performance improvements and informed decision-making in manufacturing operations.</p><h3 id=\"common-pitfalls-in-achieving-reliable-metrics\"><a href=\"#common-pitfalls-in-achieving-reliable-metrics\" class=\"heading-anchor\">Common Pitfalls in Achieving Reliable Metrics</a></h3><p>Data has become popular in business, placing significant pressure on data marts and the traditional business intelligence profession. Many so-called data professionals promise quick ROI through complex data science, AI, and ML projects or by promoting revolutionary new technologies often portrayed as miraculous solutions.</p><p>To avoid costly missteps and wasted resources on data initiatives that fail to deliver, it is important to recognize how the temptation of quick fixes and over-promised results can mislead companies. Below are common pitfalls to steer clear of when seeking the insights and metrics your business needs:</p><p><strong>Excessive Focus on Tools and Technology</strong>: While focusing on tools and technology is not inherently bad, it becomes problematic when attention shifts to unattainable promises tied to “magical” solutions. Good data and insights come from strong processes and robust stakeholder alignment—essential foundational work. Moreover, tools are increasingly standardized, so as long as the fundamentals are solid, selecting one tool over another usually won’t yield vastly different results (unless you’re Google).</p><p><strong>The One ERP Strategy</strong>: This strategy, often proposed by IT leaders in manufacturing with extensive ERP experience, is no longer effective. A single ERP system does not guarantee centralized or unified metrics. Non-ERP data, such as CRM data, has become equally critical and must be included. Furthermore, mergers and acquisitions often result in multiple ERP systems, meaning “one ERP” rarely remains as singular as promised.</p><p><strong>Putting Complex Data Modeling in the ERP</strong>: Often proposed by IT leaders, this strategy is not always a straightforward decision. In some cases, performing data modeling in the ERP makes sense, but in others, it does not. Customizing and updating data marts is now faster and more cost-effective than modifying ERPs. If your organization operates multiple ERPs, it might be more practical to move certain calculations, including key ones such as accounting reserve calculations, to the data marts layer.</p><p><strong>Expecting ROI from All Data Initiatives</strong>: This expectation often stems from a misunderstanding of the diverse aspects of data work. Not all data projects are about driving revolutionary business changes; many are essential for day-to-day operations. Without solid data marts, business functions may create their own versions of metrics, often leading to bad practices and inefficiencies, and this is also a good way for numbers to never match properly.</p><p>As we have seen, avoiding these pitfalls is important for building a reliable data and metrics ecosystem that supports business leaders. Recognizing these challenges allows organizations to focus on sustainable strategies that drive long-term success.</p><h3 id=\"conclusion\"><a href=\"#conclusion\" class=\"heading-anchor\">Conclusion</a></h3><p>Data marts provide powerful technology and best practices for delivering “good metrics” that are compliant, usable, granular, aligned, and comparable. These metrics not only improve visibility and control but also enhance predictability—key drivers of business performance.</p><p>Integrating data marts into your business performance analytics strategy accelerates access to insights, enabling faster decision-making, more efficient operations, and improved business performance.</p></div>",
      "date_published": "2025-01-02T00:00:00Z"
    }
    ,{
      "id": "https://site-dyvenia.netlify.app/insights/faster-metrics-with-data-marts-overcoming-data-warehouses-challenges/",
      "url": "https://site-dyvenia.netlify.app/insights/faster-metrics-with-data-marts-overcoming-data-warehouses-challenges/",
      "title": "Faster Metrics with Data Marts (overcoming Data Warehouses challenges)",
      "content_html": "<nav id=\"toc\" class=\"table-of-contents prose\"><ol><li class=\"flow\"><a href=\"#data-warehousing-challenges-a-double-edged-sword\">Data Warehousing Challenges: A Double-Edged Sword</a><ol><li class=\"flow\"><a href=\"#rigid-and-slow\">Rigid and Slow</a></li><li class=\"flow\"><a href=\"#poor-domain-understanding\">Poor Domain Understanding</a></li><li class=\"flow\"><a href=\"#black-boxes-the-struggle-with-compliance\">Black Boxes: The Struggle with Compliance</a></li><li class=\"flow\"><a href=\"#multiple-sources-and-the-one-erp-myth\">Multiple Sources and the “One ERP” Myth</a></li></ol></li><li class=\"flow\"><a href=\"#are-data-marts-the-way-forward\">Are Data Marts The Way Forward?</a></li><li class=\"flow\"><a href=\"#ok-but-what-is-a-data-mart\">OK, but what is a data mart?</a></li><li class=\"flow\"><a href=\"#data-marts-how-they-resolve-data-warehouse-challenges\">Data Marts: How They Resolve Data Warehouse Challenges</a><ol><li class=\"flow\"><a href=\"#rigid-and-slow-1\">Rigid and Slow</a></li><li class=\"flow\"><a href=\"#poor-domain-understanding-1\">Poor Domain Understanding</a></li><li class=\"flow\"><a href=\"#black-boxes-the-struggle-with-compliance-1\">Black Boxes: The Struggle with Compliance</a></li><li class=\"flow\"><a href=\"#multiple-sources-and-the-one-erp-myth-1\">Multiple Sources and the “One ERP” Myth</a></li></ol></li><li class=\"flow\"><a href=\"#conclusion\">Conclusion</a></li></ol></nav><p><span id=\"toc-skipped\" class=\"visually-hidden\"></span></p><div class=\"flow prose\"><p></p><p>The growing number of tools and solutions in the data ecosystem can be overwhelming for business leaders. Some data solutions are so advanced that they resemble full-scale “digital transformation” initiatives—a term often used by IT professionals.</p><p>However, one area of data analytics has remained largely unchanged for the past 30 years: business intelligence. Companies once relied on robust data warehouses with well-defined structures paired with straightforward reporting tools that efficiently extracted insights.</p><p>Over time, data warehouses began to fade from prominence. This decline was driven by:</p><ol class=\"list\"><li>advancements in data platforms</li><li>shifting focus among analytics professionals toward various data ROI initiatives</li><li>unresolved internal challenges that persisted throughout their 30-year history.</li></ol><p>But data warehousing played a critical role in organizations by supporting unified and standardized metrics. Today, this function is largely neglected, leaving businesses to grapple with its absence. As a result, leaders are increasingly witnessing the impact on performance, compliance, and control, compounded by challenges in effective monitoring.</p><p><is-land on:idle></is-land></p><dialog class=\"flow modal5\"><button autofocus class=\"button\">Close</button><figure><picture><source type=\"image/webp\" srcset=\"/img/JCeC0bUfwh-960.webp 960w\" sizes=\"auto\"><img loading=\"lazy\" decoding=\"async\" src=\"/img/JCeC0bUfwh-960.jpeg\" alt=\"Reference Data Platform Architecture\" title=\"Typical Modern Data Platform Architecture\" width=\"960\" height=\"540\"></picture><figcaption>Typical Modern Data Platform Architecture</figcaption></figure></dialog><button data-index=\"5\"><figure><picture><source type=\"image/webp\" srcset=\"/img/JCeC0bUfwh-960.webp 960w\" sizes=\"auto\"><img loading=\"lazy\" decoding=\"async\" src=\"/img/JCeC0bUfwh-960.jpeg\" alt=\"Reference Data Platform Architecture\" title=\"Typical Modern Data Platform Architecture\" width=\"960\" height=\"540\"></picture><figcaption>Typical Modern Data Platform Architecture</figcaption></figure></button><p></p><h3 id=\"data-warehousing-challenges-a-double-edged-sword\"><a href=\"#data-warehousing-challenges-a-double-edged-sword\" class=\"heading-anchor\">Data Warehousing Challenges: A Double-Edged Sword</a></h3><h4 id=\"rigid-and-slow\"><a href=\"#rigid-and-slow\" class=\"heading-anchor\">Rigid and Slow</a></h4><p>Data warehouses became <strong>the</strong> definition of slow and rigid IT. Change requests often took months—or even years—to implement, as BI teams resisted any changes that might disrupt production systems. This inflexibility drove business functions to develop their own localized data warehouses, typically ignoring best practices, total cost of ownership (TCO), and accumulating significant technical debt.</p><h4 id=\"poor-domain-understanding\"><a href=\"#poor-domain-understanding\" class=\"heading-anchor\">Poor Domain Understanding</a></h4><p>As data warehouses were centralized within IT departments, they lost connection to business domains. IT teams, by nature, lacked deep business understanding, leading to a disconnect. IT expected well-defined, scoped requirements to execute in a waterfall approach, while business professionals anticipated proactive insights and solutions from the “data people.”</p><h4 id=\"black-boxes-the-struggle-with-compliance\"><a href=\"#black-boxes-the-struggle-with-compliance\" class=\"heading-anchor\">Black Boxes: The Struggle with Compliance</a></h4><p>As IT departments took control of data modeling in data warehouses, business professionals began questioning the accuracy and reliability of the metrics they consumed. Compliance with policies and metric calculations became a significant concern: <em>If I can’t understand how a metric is calculated, how can I trust it?</em></p><h4 id=\"multiple-sources-and-the-one-erp-myth\"><a href=\"#multiple-sources-and-the-one-erp-myth\" class=\"heading-anchor\">Multiple Sources and the “One ERP” Myth</a></h4><p>In the early days of data warehousing, the application landscape was relatively simple, with ERP systems dominating as the core business application. Data warehouses were often designed as extensions of these ERP systems.</p><p>Over time, two major shifts occurred. First, mergers and acquisitions introduced multiple ERPs into the ecosystem. Second, new applications—such as CRMs, HRMs, and MES systems—entered the mix, further complicating the landscape. IT departments attempted to address these challenges by pushing for a single ERP implementation across the organization. Meanwhile, business leaders grew frustrated, questioning how many years it would take to answer even basic questions like, “How many employees do we have?”</p><h3 id=\"are-data-marts-the-way-forward\"><a href=\"#are-data-marts-the-way-forward\" class=\"heading-anchor\">Are Data Marts The Way Forward?</a></h3><p>At first glance, a data mart might seem like a minor evolution from traditional Business Intelligence (BI) run by IT:</p><blockquote><p>A data mart is a subset of a data warehouse, focusing on a specific business area or department. It contains a curated set of data tailored to the needs of that group, making it easier and faster to access insights without navigating the entire data warehouse. This closely mirrors the structure of a typical data warehouse OLAP cube, leading to the impression that introducing data marts is just a rebranding exercise.</p></blockquote><p>However, this view overlooks the fundamental shift brought about by data marts, which are driven by a new set of principles. These principles can be traced back to modern data platform architectures, which themselves evolved from the Big Data movement (remember Hadoop?). The three key principles that distinguish data marts today are:</p><ol class=\"list\"><li><strong>Transparency:</strong> Data models should be open and accessible to analysts outside of IT BI teams.</li><li><strong>Flexibility:</strong> Data models should be easy to enhance, modify, and update.</li><li><strong>Version Control:</strong> Data models should be version-controlled in code, similar to software development practices.</li></ol><p>The challenge today is that these principles are not universally adopted across the industry. Depending on who you speak to, you’ll encounter different perspectives on how data marts are implemented and leveraged.</p><h2 id=\"ok-but-what-is-a-data-mart\"><a href=\"#ok-but-what-is-a-data-mart\" class=\"heading-anchor\">OK, but what is a data mart?</a></h2><p>Technically, a data mart is simply a <strong>flat table</strong> within a SQL database. This table is usually accessed through reporting tools like Power BI, Tableau, or Cognos. Analysts can connect to the table and create typical BI reports, and as the data is updated, the connected reporting tools automatically reflect these changes.</p><p><is-land on:idle></is-land></p><dialog class=\"flow modal6\"><button autofocus class=\"button\">Close</button><figure><picture><source type=\"image/webp\" srcset=\"/img/Awz_nSbEGz-960.webp 960w\" sizes=\"auto\"><img loading=\"lazy\" decoding=\"async\" src=\"/img/Awz_nSbEGz-960.jpeg\" alt=\"Reference Data Platform with Modelling Highlight\" title=\"Models Inside a Data Platform\" width=\"960\" height=\"540\"></picture><figcaption>Models Inside a Data Platform</figcaption></figure></dialog><button data-index=\"6\"><figure><picture><source type=\"image/webp\" srcset=\"/img/Awz_nSbEGz-960.webp 960w\" sizes=\"auto\"><img loading=\"lazy\" decoding=\"async\" src=\"/img/Awz_nSbEGz-960.jpeg\" alt=\"Reference Data Platform with Modelling Highlight\" title=\"Models Inside a Data Platform\" width=\"960\" height=\"540\"></picture><figcaption>Models Inside a Data Platform</figcaption></figure></button><p></p><p>In a modern data platform, the data mart is developed using SQL code within the modeling layer. This code is typically hosted in a code repository, such as GitHub or Bitbucket. By storing the code in a repository, the data mart becomes transparent, easy to contribute to, and aligns with the three principles outlined earlier.</p><p><is-land on:idle></is-land></p><dialog class=\"flow modal7\"><button autofocus class=\"button\">Close</button><figure><picture><source type=\"image/webp\" srcset=\"/img/VqybQdyCWL-960.webp 960w, /img/VqybQdyCWL-1600.webp 1600w\" sizes=\"auto\"><img loading=\"lazy\" decoding=\"async\" src=\"/img/VqybQdyCWL-960.jpeg\" alt=\"data marts architecture\" title=\"From Data Sources, Marts and Insights &amp; Metrics\" width=\"1600\" height=\"796\" srcset=\"/img/VqybQdyCWL-960.jpeg 960w, /img/VqybQdyCWL-1600.jpeg 1600w\" sizes=\"auto\"></picture><figcaption>From Data Sources, Marts and Insights &amp; Metrics</figcaption></figure></dialog><button data-index=\"7\"><figure><picture><source type=\"image/webp\" srcset=\"/img/VqybQdyCWL-960.webp 960w, /img/VqybQdyCWL-1600.webp 1600w\" sizes=\"auto\"><img loading=\"lazy\" decoding=\"async\" src=\"/img/VqybQdyCWL-960.jpeg\" alt=\"data marts architecture\" title=\"From Data Sources, Marts and Insights &amp; Metrics\" width=\"1600\" height=\"796\" srcset=\"/img/VqybQdyCWL-960.jpeg 960w, /img/VqybQdyCWL-1600.jpeg 1600w\" sizes=\"auto\"></picture><figcaption>From Data Sources, Marts and Insights &amp; Metrics</figcaption></figure></button><p></p><h2 id=\"data-marts-how-they-resolve-data-warehouse-challenges\"><a href=\"#data-marts-how-they-resolve-data-warehouse-challenges\" class=\"heading-anchor\">Data Marts: How They Resolve Data Warehouse Challenges</a></h2><p>Now, let’s explore how data marts, with their transparency, flexibility, and version control principles, help address the challenges often encountered in traditional data warehouse systems.</p><h4 id=\"rigid-and-slow-1\"><a href=\"#rigid-and-slow-1\" class=\"heading-anchor\">Rigid and Slow</a></h4><p>Data marts are faster to develop because they allow contributions from multiple teams. By using SQL models stored in a code repository, teams can collaborate and contribute to the code. This eliminates the bottleneck caused by a single centralized team handling all the work.</p><h4 id=\"poor-domain-understanding-1\"><a href=\"#poor-domain-understanding-1\" class=\"heading-anchor\">Poor Domain Understanding</a></h4><p>With data marts, teams outside of IT can actively contribute to model creation and data mart definition, bridging the gap caused by limited business knowledge. When paired with good governance, this approach brings data modeling closer to the business, enhancing domain understanding.</p><h4 id=\"black-boxes-the-struggle-with-compliance-1\"><a href=\"#black-boxes-the-struggle-with-compliance-1\" class=\"heading-anchor\">Black Boxes: The Struggle with Compliance</a></h4><p>Since data mart code is stored in verified and recognized repositories, it becomes fully transparent, addressing the issue of “black boxes” that only a few individuals understand. Additionally, implementing data catalogs can further improve transparency by providing metadata and data lineage, making these insights accessible even to less technical users.</p><h4 id=\"multiple-sources-and-the-one-erp-myth-1\"><a href=\"#multiple-sources-and-the-one-erp-myth-1\" class=\"heading-anchor\">Multiple Sources and the “One ERP” Myth</a></h4><p>As part of modern data platforms, data marts can handle significantly larger datasets. This makes ingesting data from multiple applications cost-effective, allowing integration from various systems into a single, well-defined data mart through proper data modeling.</p><h2 id=\"conclusion\"><a href=\"#conclusion\" class=\"heading-anchor\">Conclusion</a></h2><p>Data marts represent a modern solution to many of the challenges faced by traditional data warehousing systems. By embracing transparency, flexibility, and version control, data marts empower teams across the organization to collaborate and contribute to data modeling, breaking down silos and reducing bottlenecks.</p><p>With their ability to bring data modeling closer to the business, data marts enhance domain understanding and improve compliance through transparency and clear data lineage. Moreover, the cost-effectiveness of integrating multiple data sources in a single data mart aligns with the demands of today’s complex data ecosystems.</p><p>Incorporating data marts into your data strategy can help businesses move beyond the limitations of legacy systems, enabling faster decision-making, better insights, and more agile operations.</p></div>",
      "date_published": "2024-12-23T00:00:00Z"
    }
    
  ]
}