<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: DLT workflow failing to read files from AWS S3 in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/dlt-workflow-failing-to-read-files-from-aws-s3/m-p/12848#M7606</link>
    <description>&lt;P&gt;Hi @SUDHANSHU RAJ​&amp;nbsp;is UC enabled on this workspace? What is the access mode set on the cluster? &lt;/P&gt;&lt;P&gt;Is this coming from the metastore or directly when you read from S3? Is the S3 cross-account?&lt;/P&gt;</description>
    <pubDate>Tue, 10 Jan 2023 14:54:32 GMT</pubDate>
    <dc:creator>Vivian_Wilfred</dc:creator>
    <dc:date>2023-01-10T14:54:32Z</dc:date>
    <item>
      <title>DLT workflow failing to read files from AWS S3</title>
      <link>https://community.databricks.com/t5/data-engineering/dlt-workflow-failing-to-read-files-from-aws-s3/m-p/12847#M7605</link>
      <description>&lt;P&gt;Hi All, I am trying to read streams directly from AWS S3. I set the instance profile , but when i run the workflow it fails with below error&lt;/P&gt;&lt;P&gt;"No AWS Credentials provided by TemporaryAWSCredentialsProvider : shaded.databricks.org.apache.hadoop.fs.s3a.CredentialInitializationException: Access key, secret key or session token is unset: "&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;I added below to my cluster &lt;/P&gt;&lt;P&gt;fs.s3a.aws.credentials.provider org.apache.hadoop.fs.s3a.TemporaryAWSCredentialsProvider&lt;/P&gt;&lt;P&gt;fs.s3a.access.key &amp;lt;AccessKeyId&amp;gt;&lt;/P&gt;&lt;P&gt;fs.s3a.secret.key &amp;lt;SecretAccessKey&amp;gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;It still fails with same error. Could some one please help me how to pass these for DLT workflows&lt;/P&gt;</description>
      <pubDate>Tue, 10 Jan 2023 12:00:49 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/dlt-workflow-failing-to-read-files-from-aws-s3/m-p/12847#M7605</guid>
      <dc:creator>sudhanshu1</dc:creator>
      <dc:date>2023-01-10T12:00:49Z</dc:date>
    </item>
    <item>
      <title>Re: DLT workflow failing to read files from AWS S3</title>
      <link>https://community.databricks.com/t5/data-engineering/dlt-workflow-failing-to-read-files-from-aws-s3/m-p/12848#M7606</link>
      <description>&lt;P&gt;Hi @SUDHANSHU RAJ​&amp;nbsp;is UC enabled on this workspace? What is the access mode set on the cluster? &lt;/P&gt;&lt;P&gt;Is this coming from the metastore or directly when you read from S3? Is the S3 cross-account?&lt;/P&gt;</description>
      <pubDate>Tue, 10 Jan 2023 14:54:32 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/dlt-workflow-failing-to-read-files-from-aws-s3/m-p/12848#M7606</guid>
      <dc:creator>Vivian_Wilfred</dc:creator>
      <dc:date>2023-01-10T14:54:32Z</dc:date>
    </item>
    <item>
      <title>Re: DLT workflow failing to read files from AWS S3</title>
      <link>https://community.databricks.com/t5/data-engineering/dlt-workflow-failing-to-read-files-from-aws-s3/m-p/12849#M7607</link>
      <description>&lt;P&gt;Dear Vivian,&lt;/P&gt;&lt;P&gt;UC is not enabled on this workspace . I am using Instance profile set up as per databricks document .&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;S3 is set up for cross account and as i said , i am able to run dbutils.fs.ls("s3a://zuk-comparis-poc/")&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;But when i run the workflow which envoks a delta notebook ,it gives me this error.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;This is a standard cluster , so i have not enabled IAM passthrough.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Am i missing something . Thanks in advance&lt;/P&gt;</description>
      <pubDate>Tue, 10 Jan 2023 15:42:44 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/dlt-workflow-failing-to-read-files-from-aws-s3/m-p/12849#M7607</guid>
      <dc:creator>sudhanshu1</dc:creator>
      <dc:date>2023-01-10T15:42:44Z</dc:date>
    </item>
    <item>
      <title>Re: DLT workflow failing to read files from AWS S3</title>
      <link>https://community.databricks.com/t5/data-engineering/dlt-workflow-failing-to-read-files-from-aws-s3/m-p/12850#M7608</link>
      <description>&lt;P&gt;@SUDHANSHU RAJ​&amp;nbsp;Can you please share the pipeline settings in JSON and also the cluster policy JSON? If this works on a standard cluster but not from a DLT pipeline, we need to verify the DLT pipeline settings for the cluster.&lt;/P&gt;</description>
      <pubDate>Wed, 11 Jan 2023 12:27:40 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/dlt-workflow-failing-to-read-files-from-aws-s3/m-p/12850#M7608</guid>
      <dc:creator>Vivian_Wilfred</dc:creator>
      <dc:date>2023-01-11T12:27:40Z</dc:date>
    </item>
    <item>
      <title>Re: DLT workflow failing to read files from AWS S3</title>
      <link>https://community.databricks.com/t5/data-engineering/dlt-workflow-failing-to-read-files-from-aws-s3/m-p/12851#M7609</link>
      <description>&lt;P&gt;Hi Vivian,&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Thanks for your help. I am happy to inform that it's working now . I think problem was in assigning proper roles and access to instance profile(in AWS) which i created for this purpose . Once i added few more rules , it started working .&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Thanks again for all your help. &lt;/P&gt;</description>
      <pubDate>Thu, 12 Jan 2023 10:07:09 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/dlt-workflow-failing-to-read-files-from-aws-s3/m-p/12851#M7609</guid>
      <dc:creator>sudhanshu1</dc:creator>
      <dc:date>2023-01-12T10:07:09Z</dc:date>
    </item>
  </channel>
</rss>

