<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Unable to open a file in dbfs. Trying to move files from Google Bucket to Azure Blob Storage in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/unable-to-open-a-file-in-dbfs-trying-to-move-files-from-google/m-p/45524#M27915</link>
    <description>&lt;P&gt;Thanks you for the response!&lt;/P&gt;&lt;P&gt;2 Questions:&lt;/P&gt;&lt;P&gt;1. How would you create a cluster with the custom requirements for the google cloud sdk? Is that still possible for a Unity Catalog enabled cluster with Shared Access Mode?&lt;/P&gt;&lt;P&gt;2. Is a script action the same as a cluster init script? I couldn't find any documentation for script actions.&amp;nbsp;&lt;/P&gt;&lt;P&gt;I tried running that script on an existing cluster and it returned an AttributeError with no description. Just points to the line running the dbutils.cluster.submit_run (which I also can't find documenation for this command). I verified the cluster_id and driver_node_type_id were correct.&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks for any help&lt;/P&gt;</description>
    <pubDate>Thu, 21 Sep 2023 13:09:28 GMT</pubDate>
    <dc:creator>editter</dc:creator>
    <dc:date>2023-09-21T13:09:28Z</dc:date>
    <item>
      <title>Unable to open a file in dbfs. Trying to move files from Google Bucket to Azure Blob Storage</title>
      <link>https://community.databricks.com/t5/data-engineering/unable-to-open-a-file-in-dbfs-trying-to-move-files-from-google/m-p/45396#M27864</link>
      <description>&lt;P&gt;Background:&lt;/P&gt;&lt;P&gt;I am attempting to download the google cloud sdk on Databricks. The end goal is to be able to use the sdk to transfer files from a Google Cloud Bucket to Azure Blob Storage using Databricks. (If you have any other ideas for this transfer please feel free to share. I do not want to use Azure Data Factory.)&lt;/P&gt;&lt;P&gt;I also have Unity Catalog enabled if that makes a difference.&lt;/P&gt;&lt;P&gt;Right now, I was first attempting to unzip the google cloud sdk in dbfs after I moved it to the following location. I know the file exists here:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;%fs
ls dbfs:/tmp/google_sdk

Returns:
dbfs:/tmp/google_sdk/google_cloud_sdk_352_0_0_linux_x86_64_tar.gz&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I have tried the following to open the file with tarfile. None have worked:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;tar = tarfile.open('dbfs:/tmp/google_sdk/google_cloud_sdk_352_0_0_linux_x86_64_tar.gz', mode="r|gz")

tar = tarfile.open('/dbfs/tmp/google_sdk/google_cloud_sdk_352_0_0_linux_x86_64_tar.gz', mode="r|gz")

tar = tarfile.open('/tmp/google_sdk/google_cloud_sdk_352_0_0_linux_x86_64_tar.gz', mode="r|gz")

tar = tarfile.open('/dbfs/dbfs/tmp/google_sdk/google_cloud_sdk_352_0_0_linux_x86_64_tar.gz', mode="r|gz")

tar = tarfile.open('dbfs/tmp/google_sdk/google_cloud_sdk_352_0_0_linux_x86_64_tar.gz', mode="r|gz")

tar = tarfile.open('tmp/google_sdk/google_cloud_sdk_352_0_0_linux_x86_64_tar.gz', mode="r|gz")&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;All of them returning that no such file or directory exists, but I know it does. What am I missing here? Why am I not able to open this file?&lt;BR /&gt;&lt;BR /&gt;Thanks for any help!&lt;/P&gt;</description>
      <pubDate>Wed, 20 Sep 2023 01:32:50 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/unable-to-open-a-file-in-dbfs-trying-to-move-files-from-google/m-p/45396#M27864</guid>
      <dc:creator>editter</dc:creator>
      <dc:date>2023-09-20T01:32:50Z</dc:date>
    </item>
    <item>
      <title>Re: Unable to open a file in dbfs. Trying to move files from Google Bucket to Azure Blob Storage</title>
      <link>https://community.databricks.com/t5/data-engineering/unable-to-open-a-file-in-dbfs-trying-to-move-files-from-google/m-p/45524#M27915</link>
      <description>&lt;P&gt;Thanks you for the response!&lt;/P&gt;&lt;P&gt;2 Questions:&lt;/P&gt;&lt;P&gt;1. How would you create a cluster with the custom requirements for the google cloud sdk? Is that still possible for a Unity Catalog enabled cluster with Shared Access Mode?&lt;/P&gt;&lt;P&gt;2. Is a script action the same as a cluster init script? I couldn't find any documentation for script actions.&amp;nbsp;&lt;/P&gt;&lt;P&gt;I tried running that script on an existing cluster and it returned an AttributeError with no description. Just points to the line running the dbutils.cluster.submit_run (which I also can't find documenation for this command). I verified the cluster_id and driver_node_type_id were correct.&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks for any help&lt;/P&gt;</description>
      <pubDate>Thu, 21 Sep 2023 13:09:28 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/unable-to-open-a-file-in-dbfs-trying-to-move-files-from-google/m-p/45524#M27915</guid>
      <dc:creator>editter</dc:creator>
      <dc:date>2023-09-21T13:09:28Z</dc:date>
    </item>
  </channel>
</rss>

