<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic External tables in DLT pipelines in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/external-tables-in-dlt-pipelines/m-p/106509#M42503</link>
    <description>&lt;P&gt;Hello community,&lt;BR /&gt;&lt;BR /&gt;I have implemented a DLT pipeline.&lt;BR /&gt;In the "Destination" setting of the pipeline I have specified a unity catalog with target schema of type external referring to an S3 destination.&lt;BR /&gt;My DLT pipeline works well. Yet, I&amp;nbsp;noticed that all streaming tables and materialized views generated from this pipeline are stored in non-readable locations.&lt;BR /&gt;Is it possible in DLT pipeline code to specify s3 path of the table using `@dlt.create_streaming_table`?&lt;/P&gt;</description>
    <pubDate>Tue, 21 Jan 2025 14:39:30 GMT</pubDate>
    <dc:creator>HoussemBL</dc:creator>
    <dc:date>2025-01-21T14:39:30Z</dc:date>
    <item>
      <title>External tables in DLT pipelines</title>
      <link>https://community.databricks.com/t5/data-engineering/external-tables-in-dlt-pipelines/m-p/106509#M42503</link>
      <description>&lt;P&gt;Hello community,&lt;BR /&gt;&lt;BR /&gt;I have implemented a DLT pipeline.&lt;BR /&gt;In the "Destination" setting of the pipeline I have specified a unity catalog with target schema of type external referring to an S3 destination.&lt;BR /&gt;My DLT pipeline works well. Yet, I&amp;nbsp;noticed that all streaming tables and materialized views generated from this pipeline are stored in non-readable locations.&lt;BR /&gt;Is it possible in DLT pipeline code to specify s3 path of the table using `@dlt.create_streaming_table`?&lt;/P&gt;</description>
      <pubDate>Tue, 21 Jan 2025 14:39:30 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/external-tables-in-dlt-pipelines/m-p/106509#M42503</guid>
      <dc:creator>HoussemBL</dc:creator>
      <dc:date>2025-01-21T14:39:30Z</dc:date>
    </item>
    <item>
      <title>Re: External tables in DLT pipelines</title>
      <link>https://community.databricks.com/t5/data-engineering/external-tables-in-dlt-pipelines/m-p/106511#M42504</link>
      <description>&lt;P&gt;Hello&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/115968"&gt;@HoussemBL&lt;/a&gt;,&lt;/P&gt;
&lt;P&gt;You can use below code example:&lt;/P&gt;
&lt;P&gt;import dlt&lt;/P&gt;
&lt;P&gt;@dlt.create_streaming_table(&lt;BR /&gt;name="your_table_name",&lt;BR /&gt;path="s3://your-bucket/your-path/",&lt;BR /&gt;schema="schema-definition"&lt;BR /&gt;)&lt;BR /&gt;def your_table_function():&lt;BR /&gt;return ( spark.readStream&lt;BR /&gt;.format("your_format")&lt;BR /&gt;.option("your_option_key", "your_option_value")&lt;BR /&gt;.load("your_source_path")&lt;BR /&gt;)&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;When using Unity Catalog with DLT pipelines, tables are stored in the storage location specified for the target schema. If a schema storage location is not specified, tables are stored in the catalog storage location. If neither schema nor catalog storage locations are specified, tables are stored in the root storage location of the metastore. This could be why the tables are in non-readable locations if the storage paths are not explicitly defined&lt;/P&gt;</description>
      <pubDate>Tue, 21 Jan 2025 14:45:01 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/external-tables-in-dlt-pipelines/m-p/106511#M42504</guid>
      <dc:creator>Alberto_Umana</dc:creator>
      <dc:date>2025-01-21T14:45:01Z</dc:date>
    </item>
    <item>
      <title>Re: External tables in DLT pipelines</title>
      <link>https://community.databricks.com/t5/data-engineering/external-tables-in-dlt-pipelines/m-p/106678#M42546</link>
      <description>&lt;P&gt;Hello&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/106294"&gt;@Alberto_Umana&lt;/a&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks for your reply.&lt;BR /&gt;I have followed your proposal. However, I got the following error when launching a DLT pipeline with Unity catalog.&lt;BR /&gt;&lt;BR /&gt;&lt;FONT color="#FF0000"&gt;&lt;SPAN&gt;java.lang.IllegalArgumentException: Cannot specify an explicit path for a table when using Unity Catalog. Remove the explicit path:...&lt;/SPAN&gt;&lt;/FONT&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 22 Jan 2025 16:08:48 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/external-tables-in-dlt-pipelines/m-p/106678#M42546</guid>
      <dc:creator>HoussemBL</dc:creator>
      <dc:date>2025-01-22T16:08:48Z</dc:date>
    </item>
    <item>
      <title>Re: External tables in DLT pipelines</title>
      <link>https://community.databricks.com/t5/data-engineering/external-tables-in-dlt-pipelines/m-p/113773#M44627</link>
      <description>&lt;P&gt;This won't work.best approach is create dlt sink to write to delta external table. This pipeline should only be 1 step. Read table and append flow using data sink. It works fine.&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 27 Mar 2025 01:00:36 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/external-tables-in-dlt-pipelines/m-p/113773#M44627</guid>
      <dc:creator>Sushil_saini</dc:creator>
      <dc:date>2025-03-27T01:00:36Z</dc:date>
    </item>
  </channel>
</rss>

