<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic medium.datadriveninvestor.com in Machine Learning</title>
    <link>https://community.databricks.com/t5/machine-learning/medium-datadriveninvestor-com/m-p/23101#M1305</link>
    <description>&lt;P&gt;say, I want to download 2 files from this directory (dbfs:/databricks-datasets/abc-quality/") to my local filesystem, how do I do it?&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;I understand that if those files are inside FileStore directory, it is much straightforward, which someone posts some solution here:&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;A href="https://medium.datadriveninvestor.com/how-to-download-a-file-from-databricks-filestore-to-a-local-machine-ae0c40f164f5" target="test_blank"&gt;https://medium.datadriveninvestor.com/how-to-download-a-file-from-databricks-filestore-to-a-local-machine-ae0c40f164f5&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Hence, now I am trying to see if it is possible to do a copy files from directory dbfs:/databricks-datasets/abc-quality/&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;to&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;directory in FileStore, which I tried below:&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;dbutils.fs.cp ("dbfs:/FileStore/", "dbfs:/databricks-datasets/abc-quality/", recurse = True)&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;but gives me error message:&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;java.rmi.RemoteException: com.amazonaws.services.s3.model.AmazonS3Exception: Access Denied; request: PUT &lt;A href="https://databricks-datasets-oregon.s3.us-west-2.amazonaws.com" target="test_blank"&gt;https://databricks-datasets-oregon.s3.us-west-2.amazonaws.com&lt;/A&gt; quality/Bonus_jan.csv {} aws-sdk-java/1.12.101 Linux/5.4.0-1088-aws OpenJDK_64-Bit_Server_VM/25.342-b07 java/1.8.0_342 scala/2.12.14 kotlin/1.4.0 vendor/Private_Build cfg/retry-mode/legacy com.amazonaws.services.s3.transfer.TransferManager/1.12.101 com.amazonaws.services.s3.model.PutObjectRequest; Request ID: JN9487JSJFXDPAY8, Extended Request ID: zkWIIgPXRte6UWeXv5BzC/IzHXhqOntjdCyGiBw34+mm3qMi1irFm9jfY3/iwEfbt/0Ywz4TsKw=, Cloud Provider: AWS, Instance ID: i-0f553097120e72dd4 (Service: Amazon S3; Status Code: 403; Error Code: AccessDenied; Request ID: JN9487JSJFXDPAY8; S3 Extended Request ID: zkWIIgPXRte6UWeXv5BzC/IzHXhqOntjdCyGiBw34+mm3qMi1irFm9jfY3/iwEfbt/0Ywz4TsKw=; Proxy: null), S3 Extended Request ID: zkWIIgPXRte6UWeXv5BzC/IzHXhqOntjdCyGiBw34+mm3qMi1irFm9jfY3/iwEfbt/0Ywz4TsKw=; nested exception is:&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;</description>
    <pubDate>Wed, 09 Nov 2022 08:59:34 GMT</pubDate>
    <dc:creator>THIAM_HUATTAN</dc:creator>
    <dc:date>2022-11-09T08:59:34Z</dc:date>
    <item>
      <title>medium.datadriveninvestor.com</title>
      <link>https://community.databricks.com/t5/machine-learning/medium-datadriveninvestor-com/m-p/23101#M1305</link>
      <description>&lt;P&gt;say, I want to download 2 files from this directory (dbfs:/databricks-datasets/abc-quality/") to my local filesystem, how do I do it?&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;I understand that if those files are inside FileStore directory, it is much straightforward, which someone posts some solution here:&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;A href="https://medium.datadriveninvestor.com/how-to-download-a-file-from-databricks-filestore-to-a-local-machine-ae0c40f164f5" target="test_blank"&gt;https://medium.datadriveninvestor.com/how-to-download-a-file-from-databricks-filestore-to-a-local-machine-ae0c40f164f5&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Hence, now I am trying to see if it is possible to do a copy files from directory dbfs:/databricks-datasets/abc-quality/&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;to&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;directory in FileStore, which I tried below:&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;dbutils.fs.cp ("dbfs:/FileStore/", "dbfs:/databricks-datasets/abc-quality/", recurse = True)&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;but gives me error message:&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;java.rmi.RemoteException: com.amazonaws.services.s3.model.AmazonS3Exception: Access Denied; request: PUT &lt;A href="https://databricks-datasets-oregon.s3.us-west-2.amazonaws.com" target="test_blank"&gt;https://databricks-datasets-oregon.s3.us-west-2.amazonaws.com&lt;/A&gt; quality/Bonus_jan.csv {} aws-sdk-java/1.12.101 Linux/5.4.0-1088-aws OpenJDK_64-Bit_Server_VM/25.342-b07 java/1.8.0_342 scala/2.12.14 kotlin/1.4.0 vendor/Private_Build cfg/retry-mode/legacy com.amazonaws.services.s3.transfer.TransferManager/1.12.101 com.amazonaws.services.s3.model.PutObjectRequest; Request ID: JN9487JSJFXDPAY8, Extended Request ID: zkWIIgPXRte6UWeXv5BzC/IzHXhqOntjdCyGiBw34+mm3qMi1irFm9jfY3/iwEfbt/0Ywz4TsKw=, Cloud Provider: AWS, Instance ID: i-0f553097120e72dd4 (Service: Amazon S3; Status Code: 403; Error Code: AccessDenied; Request ID: JN9487JSJFXDPAY8; S3 Extended Request ID: zkWIIgPXRte6UWeXv5BzC/IzHXhqOntjdCyGiBw34+mm3qMi1irFm9jfY3/iwEfbt/0Ywz4TsKw=; Proxy: null), S3 Extended Request ID: zkWIIgPXRte6UWeXv5BzC/IzHXhqOntjdCyGiBw34+mm3qMi1irFm9jfY3/iwEfbt/0Ywz4TsKw=; nested exception is:&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 09 Nov 2022 08:59:34 GMT</pubDate>
      <guid>https://community.databricks.com/t5/machine-learning/medium-datadriveninvestor-com/m-p/23101#M1305</guid>
      <dc:creator>THIAM_HUATTAN</dc:creator>
      <dc:date>2022-11-09T08:59:34Z</dc:date>
    </item>
    <item>
      <title>Re: medium.datadriveninvestor.com</title>
      <link>https://community.databricks.com/t5/machine-learning/medium-datadriveninvestor-com/m-p/23102#M1306</link>
      <description>&lt;P&gt;Hi @THIAM HUAT TAN​&amp;nbsp;,&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;isn't this dbfs://databricks-datasets Databricks owned s3:// mounted to the workspace? &lt;/P&gt;&lt;P&gt;You got an error - 403 access denied to PUT files into the s3 bucket:  &lt;A href="https://databricks-datasets-oregon.s3.us-west-2.amazonaws.com" alt="https://databricks-datasets-oregon.s3.us-west-2.amazonaws.com" target="_blank"&gt;https://databricks-datasets-oregon.s3.us-west-2.amazonaws.com&lt;/A&gt; &lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;you should use some s3 bucket that you own and you can add files or if the file is small , you might be able to use dbfs://tmp/&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;thanks,&lt;/P&gt;&lt;P&gt;Pat.&lt;/P&gt;</description>
      <pubDate>Wed, 09 Nov 2022 12:21:21 GMT</pubDate>
      <guid>https://community.databricks.com/t5/machine-learning/medium-datadriveninvestor-com/m-p/23102#M1306</guid>
      <dc:creator>Pat</dc:creator>
      <dc:date>2022-11-09T12:21:21Z</dc:date>
    </item>
    <item>
      <title>Re: medium.datadriveninvestor.com</title>
      <link>https://community.databricks.com/t5/machine-learning/medium-datadriveninvestor-com/m-p/23103#M1307</link>
      <description>&lt;P&gt;I got the order of source and destination reversed. Now, it works, thanks &lt;span class="lia-unicode-emoji" title=":thumbs_up:"&gt;👍&lt;/span&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 09 Nov 2022 12:32:48 GMT</pubDate>
      <guid>https://community.databricks.com/t5/machine-learning/medium-datadriveninvestor-com/m-p/23103#M1307</guid>
      <dc:creator>THIAM_HUATTAN</dc:creator>
      <dc:date>2022-11-09T12:32:48Z</dc:date>
    </item>
  </channel>
</rss>

