<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Best Practices for Designing Bronze Layer with SQL Server Source in Medallion Architecture in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/best-practices-for-designing-bronze-layer-with-sql-server-source/m-p/127273#M47909</link>
    <description>&lt;P&gt;Create the bronze table using Auto Loader and store it in Delta format. Although it might seem like you'll only read from bronze once to populate the silver layer, in real-world production scenarios, you'll often need to re-read from bronze—for reprocessing, debugging, or new use cases. Using Delta ensures those repeated reads are performant and reliable, compared to reading directly from raw formats like JSON or CSV.&lt;/P&gt;</description>
    <pubDate>Mon, 04 Aug 2025 06:34:08 GMT</pubDate>
    <dc:creator>pgo</dc:creator>
    <dc:date>2025-08-04T06:34:08Z</dc:date>
    <item>
      <title>Best Practices for Designing Bronze Layer with SQL Server Source in Medallion Architecture</title>
      <link>https://community.databricks.com/t5/data-engineering/best-practices-for-designing-bronze-layer-with-sql-server-source/m-p/127188#M47882</link>
      <description>&lt;P&gt;Hi Databricks Experts,&lt;BR /&gt;&lt;BR /&gt;I'm working on a Medallion Architecture implementation in Databricks, where the source data is coming from SQL Server. I would like some advice on how to handle the bronze layer correctly and cost-effectively.&lt;BR /&gt;&lt;BR /&gt;Should I create a bronze schema in Unity Catalog and store the raw ingested tables as-is (in Delta format)?&lt;BR /&gt;&lt;BR /&gt;Or is it better to simply save the raw data (as CSV or Parquet) in an ADLS container without registering Delta tables in Unity Catalog?&lt;BR /&gt;&lt;BR /&gt;Additionally, from a cost perspective:&lt;BR /&gt;&lt;BR /&gt;Is there any cost involved in just creating tables in Unity Catalog?&lt;BR /&gt;&lt;BR /&gt;Will saving the raw data directly in ADLS (as Parquet or CSV) be more cost-efficient?&lt;BR /&gt;&lt;BR /&gt;Which components incur Azure Storage vs. Databricks (DBU/compute) charges?&lt;BR /&gt;&lt;BR /&gt;I'm looking to understand the proper architectural approach that balances cost and best practices for production use cases. Any suggestions or experiences would be greatly appreciated.&lt;BR /&gt;&lt;BR /&gt;Thanks!&lt;BR /&gt;&lt;BR /&gt;&lt;/P&gt;</description>
      <pubDate>Fri, 01 Aug 2025 14:26:37 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/best-practices-for-designing-bronze-layer-with-sql-server-source/m-p/127188#M47882</guid>
      <dc:creator>Mohan_Baabu1</dc:creator>
      <dc:date>2025-08-01T14:26:37Z</dc:date>
    </item>
    <item>
      <title>Re: Best Practices for Designing Bronze Layer with SQL Server Source in Medallion Architecture</title>
      <link>https://community.databricks.com/t5/data-engineering/best-practices-for-designing-bronze-layer-with-sql-server-source/m-p/127232#M47896</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/177656"&gt;@Mohan_Baabu1&lt;/a&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Let me break down the bronze layer approach for your Medallion Architecture with detailed cost and architectural considerations.&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Tables in Unity Catalog:&lt;/STRONG&gt;&lt;BR /&gt;Why This is Better Than Raw Files:&lt;BR /&gt;- Schema Evolution - Handle SQL Server schema changes gracefully&lt;BR /&gt;- ACID Transactions - Consistent data during ingestion&lt;BR /&gt;- Time Travel - Debug data issues and recover from bad loads&lt;BR /&gt;- Performance - Z-ordering and file compaction&lt;BR /&gt;- Governance - Unity Catalog lineage and access control&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Cost Breakdown Analysis:&lt;/STRONG&gt;&lt;BR /&gt;&lt;STRONG&gt;FREE Components:&lt;/STRONG&gt;&lt;BR /&gt;- Creating schemas/catalogs&lt;BR /&gt;- Registering tables (metadata only)&lt;BR /&gt;- Unity Catalog metadata storage&lt;BR /&gt;- Lineage tracking&lt;BR /&gt;- Access control policies&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;&lt;span class="lia-unicode-emoji" title=":money_bag:"&gt;💰&lt;/span&gt; PAID Components:&lt;/STRONG&gt;&lt;BR /&gt;- DBU consumption during queries/operations&lt;BR /&gt;- Underlying storage (ADLS/S3)&lt;BR /&gt;- Compute for optimization operations&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Detailed Cost Analysis&lt;/STRONG&gt;&lt;BR /&gt;&lt;STRONG&gt;ADLS vs Databricks Charges:&lt;/STRONG&gt;&lt;BR /&gt;&lt;STRONG&gt;ADLS Storage Costs:&lt;/STRONG&gt;&lt;BR /&gt;- Standard tier: $0.0184/GB/month&lt;BR /&gt;- Cool tier: $0.01/GB/month (for archival bronze data)&lt;BR /&gt;- Transaction costs: ~$0.0004 per 10k operations&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Databricks Costs:&lt;/STRONG&gt;&lt;BR /&gt;- DBU consumption during:&lt;BR /&gt;* Data ingestion jobs&lt;BR /&gt;* Table optimization (OPTIMIZE, VACUUM)&lt;BR /&gt;* Query execution&lt;BR /&gt;* Cluster startup/shutdown&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Cost Optimization Tips:&lt;/STRONG&gt;&lt;BR /&gt;- Use spot instances: 60-70% savings&lt;BR /&gt;- Auto-terminate idle clusters&lt;BR /&gt;- Schedule optimization during off-peak hours&lt;BR /&gt;- Use smaller clusters for bronze layer&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;Final Recommendation&lt;BR /&gt;&lt;STRONG&gt;Use Unity Catalog Delta Tables for Bronze Layer because:&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;Cost-effective: Storage costs are similar to Parquet, but with better compression:&lt;BR /&gt;&amp;nbsp;- Production-ready: ACID transactions, schema evolution, time travel&lt;BR /&gt;&amp;nbsp;- Governance: Built-in lineage, access control, data discovery&lt;BR /&gt;&amp;nbsp;- Performance: Better query performance for silver/gold transformations&lt;BR /&gt;&amp;nbsp;- Maintenance: Automated optimization reduces long-term costs&lt;/P&gt;&lt;P&gt;The additional UC overhead is minimal compared to the operational benefits, especially in production environments where data quality and governance are crucial.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Sat, 02 Aug 2025 21:52:47 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/best-practices-for-designing-bronze-layer-with-sql-server-source/m-p/127232#M47896</guid>
      <dc:creator>lingareddy_Alva</dc:creator>
      <dc:date>2025-08-02T21:52:47Z</dc:date>
    </item>
    <item>
      <title>Re: Best Practices for Designing Bronze Layer with SQL Server Source in Medallion Architecture</title>
      <link>https://community.databricks.com/t5/data-engineering/best-practices-for-designing-bronze-layer-with-sql-server-source/m-p/127269#M47906</link>
      <description>&lt;P&gt;Thanks&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/24053"&gt;@lingareddy_Alva&lt;/a&gt;&amp;nbsp;for your response, considering the cost effective and the medallion architecture, i need to create bronze tables in the unity catalog than saving the files directly to the ADLS container for the bronze layer.&lt;/P&gt;&lt;P&gt;Also considering the features of delta tables,&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;- Schema Evolution - Handle SQL Server schema changes gracefully&lt;/SPAN&gt;&lt;BR /&gt;&lt;SPAN&gt;- ACID Transactions - Consistent data during ingestion&lt;/SPAN&gt;&lt;BR /&gt;&lt;SPAN&gt;- Time Travel - Debug data issues and recover from bad loads&lt;/SPAN&gt;&lt;BR /&gt;&lt;SPAN&gt;- Performance - Z-ordering and file compaction&lt;/SPAN&gt;&lt;BR /&gt;&lt;SPAN&gt;- Governance - Unity Catalog lineage and access control&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Mon, 04 Aug 2025 05:49:02 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/best-practices-for-designing-bronze-layer-with-sql-server-source/m-p/127269#M47906</guid>
      <dc:creator>Mohan_Baabu1</dc:creator>
      <dc:date>2025-08-04T05:49:02Z</dc:date>
    </item>
    <item>
      <title>Re: Best Practices for Designing Bronze Layer with SQL Server Source in Medallion Architecture</title>
      <link>https://community.databricks.com/t5/data-engineering/best-practices-for-designing-bronze-layer-with-sql-server-source/m-p/127273#M47909</link>
      <description>&lt;P&gt;Create the bronze table using Auto Loader and store it in Delta format. Although it might seem like you'll only read from bronze once to populate the silver layer, in real-world production scenarios, you'll often need to re-read from bronze—for reprocessing, debugging, or new use cases. Using Delta ensures those repeated reads are performant and reliable, compared to reading directly from raw formats like JSON or CSV.&lt;/P&gt;</description>
      <pubDate>Mon, 04 Aug 2025 06:34:08 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/best-practices-for-designing-bronze-layer-with-sql-server-source/m-p/127273#M47909</guid>
      <dc:creator>pgo</dc:creator>
      <dc:date>2025-08-04T06:34:08Z</dc:date>
    </item>
    <item>
      <title>Re: Best Practices for Designing Bronze Layer with SQL Server Source in Medallion Architecture</title>
      <link>https://community.databricks.com/t5/data-engineering/best-practices-for-designing-bronze-layer-with-sql-server-source/m-p/127276#M47911</link>
      <description>&lt;P&gt;thanks&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/28850"&gt;@pgo&lt;/a&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Mon, 04 Aug 2025 06:43:39 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/best-practices-for-designing-bronze-layer-with-sql-server-source/m-p/127276#M47911</guid>
      <dc:creator>Mohan_Baabu1</dc:creator>
      <dc:date>2025-08-04T06:43:39Z</dc:date>
    </item>
  </channel>
</rss>

