<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Loading CSV from private S3 bucket in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/loading-csv-from-private-s3-bucket/m-p/138683#M50994</link>
    <description>&lt;P&gt;Trying to load a csv file from a private S3 bucket&lt;/P&gt;&lt;P&gt;please clarify requirements to do this- Can I do it in community edition (if yes then how)? How to do it in premium version?&lt;/P&gt;&lt;P&gt;I have IAM role and I also access key &amp;amp; secret&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Wed, 12 Nov 2025 04:58:18 GMT</pubDate>
    <dc:creator>intelliconnectq</dc:creator>
    <dc:date>2025-11-12T04:58:18Z</dc:date>
    <item>
      <title>Loading CSV from private S3 bucket</title>
      <link>https://community.databricks.com/t5/data-engineering/loading-csv-from-private-s3-bucket/m-p/138683#M50994</link>
      <description>&lt;P&gt;Trying to load a csv file from a private S3 bucket&lt;/P&gt;&lt;P&gt;please clarify requirements to do this- Can I do it in community edition (if yes then how)? How to do it in premium version?&lt;/P&gt;&lt;P&gt;I have IAM role and I also access key &amp;amp; secret&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 12 Nov 2025 04:58:18 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/loading-csv-from-private-s3-bucket/m-p/138683#M50994</guid>
      <dc:creator>intelliconnectq</dc:creator>
      <dc:date>2025-11-12T04:58:18Z</dc:date>
    </item>
    <item>
      <title>Re: Loading CSV from private S3 bucket</title>
      <link>https://community.databricks.com/t5/data-engineering/loading-csv-from-private-s3-bucket/m-p/138713#M51000</link>
      <description>&lt;P&gt;Assuming you have these pre-requisites:&amp;nbsp;&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;&lt;P&gt;A &lt;STRONG&gt;private S3 bucket&lt;/STRONG&gt; (e.g., s3://my-private-bucket/data/file.csv)&lt;/P&gt;&lt;/LI&gt;&lt;LI&gt;&lt;P&gt;An &lt;STRONG&gt;IAM user or role&lt;/STRONG&gt; with access (list/get) to that bucket&lt;/P&gt;&lt;/LI&gt;&lt;LI&gt;&lt;P&gt;The &lt;STRONG&gt;AWS Access Key ID&lt;/STRONG&gt; and &lt;STRONG&gt;Secret Access Key&lt;/STRONG&gt; (client and secret)&lt;/P&gt;&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;The most straightforward way, for testing and checking that connection works, could be this one by using a notebook:&lt;/P&gt;&lt;P&gt;Set keys in spark directly:&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;spark._jsc.hadoopConfiguration().set("fs.s3a.access.key", "&amp;lt;YOUR_AWS_ACCESS_KEY_ID&amp;gt;")
spark._jsc.hadoopConfiguration().set("fs.s3a.secret.key", "&amp;lt;YOUR_AWS_SECRET_ACCESS_KEY&amp;gt;")
spark._jsc.hadoopConfiguration().set("fs.s3a.endpoint", "s3.amazonaws.com")&lt;/LI-CODE&gt;&lt;P&gt;Then read your file into a dataframe:&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;df = spark.read.option("header", "true").csv("s3a://my-private-bucket/data/file.csv")
display(df)
​&lt;/LI-CODE&gt;&lt;P&gt;For non-testing scenarios you can store secrets in Databricks secrets or some Key Vault and retrieve from here whenever you need. However, best option would be using&amp;nbsp;AWS IAM policies to attach proper roles to clusters in order to access data directly without specifying credentials.&lt;/P&gt;&lt;P&gt;Take a look at this:&amp;nbsp;&lt;A href="https://docs.databricks.com/aws/en/connect/storage/tutorial-s3-instance-profile" target="_blank" rel="noopener"&gt;https://docs.databricks.com/aws/en/connect/storage/tutorial-s3-instance-profile&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 12 Nov 2025 08:44:31 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/loading-csv-from-private-s3-bucket/m-p/138713#M51000</guid>
      <dc:creator>Coffee77</dc:creator>
      <dc:date>2025-11-12T08:44:31Z</dc:date>
    </item>
    <item>
      <title>Re: Loading CSV from private S3 bucket</title>
      <link>https://community.databricks.com/t5/data-engineering/loading-csv-from-private-s3-bucket/m-p/138996#M51068</link>
      <description>&lt;P&gt;Thank you&lt;/P&gt;</description>
      <pubDate>Fri, 14 Nov 2025 04:30:53 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/loading-csv-from-private-s3-bucket/m-p/138996#M51068</guid>
      <dc:creator>intelliconnectq</dc:creator>
      <dc:date>2025-11-14T04:30:53Z</dc:date>
    </item>
  </channel>
</rss>

