<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Delta live tables for large number of tables in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/delta-live-tables-for-large-number-of-tables/m-p/48983#M28443</link>
    <description>&lt;P&gt;The only issue is though hat the tables are largely from one schema &lt;span class="lia-unicode-emoji" title=":slightly_smiling_face:"&gt;🙂&lt;/span&gt; I wonder if there is an upper limit on the number of tables in a dlt pipeline/&lt;/P&gt;</description>
    <pubDate>Wed, 11 Oct 2023 21:46:04 GMT</pubDate>
    <dc:creator>priyanananthram</dc:creator>
    <dc:date>2023-10-11T21:46:04Z</dc:date>
    <item>
      <title>Delta live tables for large number of tables</title>
      <link>https://community.databricks.com/t5/data-engineering/delta-live-tables-for-large-number-of-tables/m-p/48918#M28422</link>
      <description>&lt;P&gt;Hi There&amp;nbsp;&lt;/P&gt;&lt;P&gt;I am hoping for some guidance I have some 850 tables that I need to ingest using&amp;nbsp; a DLT Pipeline. When I do this my event log shows that driver node dies becomes unresponsive likely due to GC.&lt;/P&gt;&lt;P&gt;Can DLT be used to ingest large number of tables&lt;/P&gt;&lt;P&gt;Is there some way for me to batch these tables so that I can create dlt tables 50 odd at a time.My tables will be streaming tables and hte plan is for them to run continuously&lt;/P&gt;&lt;P&gt;What can I do to ameliorate these?&lt;/P&gt;&lt;P&gt;I am on azure cloud is there a particular compute type that would be beneficial to read larger number of tables ?&lt;/P&gt;&lt;P&gt;Kind Regards&lt;/P&gt;&lt;P&gt;Priya&lt;/P&gt;</description>
      <pubDate>Wed, 11 Oct 2023 06:20:10 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/delta-live-tables-for-large-number-of-tables/m-p/48918#M28422</guid>
      <dc:creator>priyanananthram</dc:creator>
      <dc:date>2023-10-11T06:20:10Z</dc:date>
    </item>
    <item>
      <title>Re: Delta live tables for large number of tables</title>
      <link>https://community.databricks.com/t5/data-engineering/delta-live-tables-for-large-number-of-tables/m-p/48941#M28432</link>
      <description>&lt;P&gt;This can be controlled at workflow level, my opinion would be to batch it basis schema&lt;/P&gt;</description>
      <pubDate>Wed, 11 Oct 2023 11:27:42 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/delta-live-tables-for-large-number-of-tables/m-p/48941#M28432</guid>
      <dc:creator>Faisal</dc:creator>
      <dc:date>2023-10-11T11:27:42Z</dc:date>
    </item>
    <item>
      <title>Re: Delta live tables for large number of tables</title>
      <link>https://community.databricks.com/t5/data-engineering/delta-live-tables-for-large-number-of-tables/m-p/48983#M28443</link>
      <description>&lt;P&gt;The only issue is though hat the tables are largely from one schema &lt;span class="lia-unicode-emoji" title=":slightly_smiling_face:"&gt;🙂&lt;/span&gt; I wonder if there is an upper limit on the number of tables in a dlt pipeline/&lt;/P&gt;</description>
      <pubDate>Wed, 11 Oct 2023 21:46:04 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/delta-live-tables-for-large-number-of-tables/m-p/48983#M28443</guid>
      <dc:creator>priyanananthram</dc:creator>
      <dc:date>2023-10-11T21:46:04Z</dc:date>
    </item>
    <item>
      <title>Re: Delta live tables for large number of tables</title>
      <link>https://community.databricks.com/t5/data-engineering/delta-live-tables-for-large-number-of-tables/m-p/50344#M28773</link>
      <description>&lt;P&gt;Delta Live Tables (DLT) can indeed be used to ingest a large number of tables. However, if you're experiencing issues with the driver node becoming unresponsive due to garbage collection (GC), it might be a sign that the resources allocated to the driver are insufficient.To manage the ingestion of a large number of tables, you can consider batching the tables. You can create multiple DLT pipelines, each handling a subset of the tables. This way, you can distribute the load across multiple pipelines, reducing the pressure on a single pipeline and potentially mitigating the GC issue.In terms of compute type on Azure, you might want to consider using larger VM sizes for your Databricks clusters, especially for the driver node, to handle the load of reading a large number of tables. The choice of VM size would depend on the size and complexity of your tables.Also, consider tuning the Spark configurations related to memory management and GC. For instance, you can adjust the Spark driver memory, the fraction of memory dedicated to Spark's storage and execution, and the GC settings.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 02 Nov 2023 10:49:37 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/delta-live-tables-for-large-number-of-tables/m-p/50344#M28773</guid>
      <dc:creator>Sidhant07</dc:creator>
      <dc:date>2023-11-02T10:49:37Z</dc:date>
    </item>
    <item>
      <title>Re: Delta live tables for large number of tables</title>
      <link>https://community.databricks.com/t5/data-engineering/delta-live-tables-for-large-number-of-tables/m-p/91476#M38170</link>
      <description>&lt;P&gt;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/36707"&gt;@Sidhant07&lt;/a&gt;&amp;nbsp;this is useful information however are having similar issue in our pipeline. The pipeline has multiple sub pipelines. We have about 26 streaming tables as part of the pipeline and this pipeline is hogging cpu of the job compute cluster, both driver and worker nodes.&amp;nbsp;&lt;/P&gt;&lt;P&gt;There are 5 worker nodes each with type Standard_D4ads_v5 (4 core, 16GB memory) and a driver node with type Standard_D8ads_v5 (8 core, 32GB memory) . All 5 workers are running and they are either dark orange or turned red which means running very hot on cpu. Driver is worse running at over 95% cpu.&amp;nbsp;&lt;/P&gt;&lt;P&gt;How can we troubleshoot which part of pipeline is hogging the CPU ? e.g. which sub pipeline is the cause of issue or which part of sub pipeline is causing cpu hogging and how to narrow down the issue&lt;/P&gt;</description>
      <pubDate>Mon, 23 Sep 2024 17:11:23 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/delta-live-tables-for-large-number-of-tables/m-p/91476#M38170</guid>
      <dc:creator>PushkarDeole</dc:creator>
      <dc:date>2024-09-23T17:11:23Z</dc:date>
    </item>
  </channel>
</rss>

