<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Autoingest not working with Unity Catalog in DLT pipeline in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/autoingest-not-working-with-unity-catalog-in-dlt-pipeline/m-p/56321#M30519</link>
    <description>&lt;P&gt;Thanks&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/9"&gt;@Retired_mod&lt;/a&gt;, UC can connect to the S3 bucket and read the data but it fails when trying to set up the bucket notifications.&lt;/P&gt;&lt;P&gt;I'll raise a ticket with support and post back here if I find a resolution.&lt;/P&gt;</description>
    <pubDate>Tue, 02 Jan 2024 23:06:52 GMT</pubDate>
    <dc:creator>Red1</dc:creator>
    <dc:date>2024-01-02T23:06:52Z</dc:date>
    <item>
      <title>Autoingest not working with Unity Catalog in DLT pipeline</title>
      <link>https://community.databricks.com/t5/data-engineering/autoingest-not-working-with-unity-catalog-in-dlt-pipeline/m-p/55960#M30475</link>
      <description>&lt;P&gt;Hey Everyone,&lt;/P&gt;&lt;P&gt;I've built a very simple pipeline with a single DLT using auto ingest, and it works, provided I don't specify the output location. When I build the same pipeline but set UC as the output location, it fails when setting up S3 notifications, which is entirely bizarre. I've looked at the logs on the DBZ side and request logs in AWS and it looks like DBZ isn't using the instance profile I've set for some reason. Further details below, any help would be greatly appreciated!&lt;/P&gt;&lt;P&gt;Context&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;Databricks on AWS&lt;/LI&gt;&lt;LI&gt;Deployed 1 week ago so uses all the latest features (unity catalog metastore is the default)&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;Things I've done&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;The instance profile is set in the pipeline settings and it appears in both clusters in the JSON settings&lt;/LI&gt;&lt;LI&gt;The same instance profile is used when setting up the pipeline without unity and it correctly creates the SNS/SQS resources without issue, so it's not a permissions thing on the role&lt;/LI&gt;&lt;LI&gt;The cluster access settings is set to "Shared"&amp;nbsp;&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;Things I've tried&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;I set up a security credential in the target unity catalog (by copying the working instance profile) for the bucket but that didn't change anything (and it's my understanding this is only for accessing data, not used for setting up File Notification settings)&lt;/LI&gt;&lt;LI&gt;I gave the unity IAM role full access to S3, no difference&lt;/LI&gt;&lt;LI&gt;I rebuilt the pipeline, no effect&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;&lt;!--    notionvc: 523251dd-6fbc-49b5-8ef8-4a3d07c7f1c2    --&gt;&lt;/P&gt;</description>
      <pubDate>Sun, 31 Dec 2023 01:52:05 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/autoingest-not-working-with-unity-catalog-in-dlt-pipeline/m-p/55960#M30475</guid>
      <dc:creator>Red1</dc:creator>
      <dc:date>2023-12-31T01:52:05Z</dc:date>
    </item>
    <item>
      <title>Re: Autoingest not working with Unity Catalog in DLT pipeline</title>
      <link>https://community.databricks.com/t5/data-engineering/autoingest-not-working-with-unity-catalog-in-dlt-pipeline/m-p/56047#M30491</link>
      <description>&lt;P&gt;Hey&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/9"&gt;@Retired_mod&lt;/a&gt;&amp;nbsp;,&lt;/P&gt;&lt;P&gt;Thanks for the response!&lt;/P&gt;&lt;P&gt;I tried the above with no luck unfortunately:&lt;BR /&gt;- I don't have an apply_merge function in my pipeline definition, please find the pipe definition below&lt;BR /&gt;- I'm running DBR 14.2&lt;BR /&gt;- I don't think databricks connect applies here as this was all set up in the databricks UI&lt;BR /&gt;- Thanks for the link to that one, I read it a couple times and I've implemented all the recommendations with no luck.&lt;/P&gt;&lt;P&gt;DLT definition:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;CREATE OR REFRESH STREAMING LIVE TABLE raw_testing
AS SELECT *
  FROM cloud_files(
    "s3://bucket-path",
    "csv",
    map(
      "header", "true",
      "sep", "|",
      "cloudFiles.useNotifications", "true",
      "inferSchema", "true"
    )
  );&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;This pipeline works as expected when using HMR as the output location but doesn't work with UC.&amp;nbsp;&lt;/P&gt;&lt;P&gt;Any other thoughts? Is there some way i can escalate this? At this point it feels like a bug.&lt;/P&gt;</description>
      <pubDate>Tue, 02 Jan 2024 07:59:17 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/autoingest-not-working-with-unity-catalog-in-dlt-pipeline/m-p/56047#M30491</guid>
      <dc:creator>Red1</dc:creator>
      <dc:date>2024-01-02T07:59:17Z</dc:date>
    </item>
    <item>
      <title>Re: Autoingest not working with Unity Catalog in DLT pipeline</title>
      <link>https://community.databricks.com/t5/data-engineering/autoingest-not-working-with-unity-catalog-in-dlt-pipeline/m-p/56321#M30519</link>
      <description>&lt;P&gt;Thanks&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/9"&gt;@Retired_mod&lt;/a&gt;, UC can connect to the S3 bucket and read the data but it fails when trying to set up the bucket notifications.&lt;/P&gt;&lt;P&gt;I'll raise a ticket with support and post back here if I find a resolution.&lt;/P&gt;</description>
      <pubDate>Tue, 02 Jan 2024 23:06:52 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/autoingest-not-working-with-unity-catalog-in-dlt-pipeline/m-p/56321#M30519</guid>
      <dc:creator>Red1</dc:creator>
      <dc:date>2024-01-02T23:06:52Z</dc:date>
    </item>
    <item>
      <title>Re: Autoingest not working with Unity Catalog in DLT pipeline</title>
      <link>https://community.databricks.com/t5/data-engineering/autoingest-not-working-with-unity-catalog-in-dlt-pipeline/m-p/67965#M33494</link>
      <description>&lt;P&gt;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/97238"&gt;@Red1&lt;/a&gt;&amp;nbsp; Were you able resolve this issue, if yes , what was the fix ?&lt;/P&gt;</description>
      <pubDate>Thu, 02 May 2024 17:33:11 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/autoingest-not-working-with-unity-catalog-in-dlt-pipeline/m-p/67965#M33494</guid>
      <dc:creator>Babu_Krishnan</dc:creator>
      <dc:date>2024-05-02T17:33:11Z</dc:date>
    </item>
    <item>
      <title>Re: Autoingest not working with Unity Catalog in DLT pipeline</title>
      <link>https://community.databricks.com/t5/data-engineering/autoingest-not-working-with-unity-catalog-in-dlt-pipeline/m-p/67980#M33504</link>
      <description>&lt;P&gt;Hey&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/100208"&gt;@Babu_Krishnan&lt;/a&gt;&amp;nbsp;I was! I had to reach out to my Databricks support engineer directly and the resolution was to add "cloudfiles.awsAccessKey" and "cloudfiles.awsSecretKey" to the params as in the screenshot below (apologies, i don't know why the scrnsht is so grainy). he also mentioned using Databricks secret store for the credentials themselves.&lt;!--  notionvc: a7c4b55c-bf31-4af5-8bf3-59faa441cd63  --&gt;&lt;/P&gt;</description>
      <pubDate>Thu, 02 May 2024 23:08:03 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/autoingest-not-working-with-unity-catalog-in-dlt-pipeline/m-p/67980#M33504</guid>
      <dc:creator>Red1</dc:creator>
      <dc:date>2024-05-02T23:08:03Z</dc:date>
    </item>
    <item>
      <title>Re: Autoingest not working with Unity Catalog in DLT pipeline</title>
      <link>https://community.databricks.com/t5/data-engineering/autoingest-not-working-with-unity-catalog-in-dlt-pipeline/m-p/67988#M33507</link>
      <description>&lt;P&gt;Thanks a lot&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/97238"&gt;@Red1&lt;/a&gt;. Let me try that.&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;But curious to know what the purpose of roleARN is. Also interested in learning how we can utilize Secret Manager to prevent passing credentials as plain text in a notebook. Thanks in advance.&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Fri, 03 May 2024 01:13:25 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/autoingest-not-working-with-unity-catalog-in-dlt-pipeline/m-p/67988#M33507</guid>
      <dc:creator>Babu_Krishnan</dc:creator>
      <dc:date>2024-05-03T01:13:25Z</dc:date>
    </item>
    <item>
      <title>Re: Autoingest not working with Unity Catalog in DLT pipeline</title>
      <link>https://community.databricks.com/t5/data-engineering/autoingest-not-working-with-unity-catalog-in-dlt-pipeline/m-p/67989#M33508</link>
      <description>&lt;P&gt;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/97238"&gt;@Red1&lt;/a&gt;&amp;nbsp;, It worked . Thanks for the details. Used Databricks secrets to store the credentials.&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 03 May 2024 03:05:37 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/autoingest-not-working-with-unity-catalog-in-dlt-pipeline/m-p/67989#M33508</guid>
      <dc:creator>Babu_Krishnan</dc:creator>
      <dc:date>2024-05-03T03:05:37Z</dc:date>
    </item>
  </channel>
</rss>

