<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Databricks mount bug in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/databricks-mount-bug/m-p/101761#M40817</link>
    <description>&lt;P&gt;Hello,&lt;/P&gt;&lt;P&gt;I have a weird problem in databricks for which I hope you can suggest some solutions.&lt;BR /&gt;I have an azureml blob storage mounted to databricks with some folder structure that can be accessed from a notebook as&lt;BR /&gt;/dbfs/mnt/azuremount/foo/bar/something.txt&lt;BR /&gt;Usually there could be a very large number of subfolders in some of the folders, for example foo could contain more than 1000 subfolders.&lt;/P&gt;&lt;P&gt;My problem is the following:&lt;BR /&gt;If I start a new compute cluster I can access something.txt from any notebook. If I do&lt;BR /&gt;ls /dbfs/mnt/azuremount/foo/bar/something.txt | head -n 2&lt;BR /&gt;in some notebook then databricks throws an error that there is no file uner the path "/dbfs/mnt/azuremount/foo/bar/something.txt".&lt;BR /&gt;But if call the following commands sequentially:&lt;BR /&gt;ls /dbfs/mnt/azuremount/ | head -n 2&lt;BR /&gt;ls /dbfs/mnt/azuremount/foo/ | head -n 2&lt;BR /&gt;ls /dbfs/mnt/azuremount/foo/bar/ | head -n 2&lt;BR /&gt;then I can magically open something.txt from any notebook!&lt;/P&gt;&lt;P&gt;I suspect that there must be some caching issue in the background but when I tried dbutils.fs.refreshMount() it didn't fix my issue. I want to find some solutions to this because it is incredibly annoying if I have to open multiple folders by hand all the time once I start my compute cluster.&lt;/P&gt;&lt;P&gt;Thanks,&lt;BR /&gt;Ben&lt;/P&gt;</description>
    <pubDate>Wed, 11 Dec 2024 13:03:02 GMT</pubDate>
    <dc:creator>BenceCzako</dc:creator>
    <dc:date>2024-12-11T13:03:02Z</dc:date>
    <item>
      <title>Databricks mount bug</title>
      <link>https://community.databricks.com/t5/data-engineering/databricks-mount-bug/m-p/101761#M40817</link>
      <description>&lt;P&gt;Hello,&lt;/P&gt;&lt;P&gt;I have a weird problem in databricks for which I hope you can suggest some solutions.&lt;BR /&gt;I have an azureml blob storage mounted to databricks with some folder structure that can be accessed from a notebook as&lt;BR /&gt;/dbfs/mnt/azuremount/foo/bar/something.txt&lt;BR /&gt;Usually there could be a very large number of subfolders in some of the folders, for example foo could contain more than 1000 subfolders.&lt;/P&gt;&lt;P&gt;My problem is the following:&lt;BR /&gt;If I start a new compute cluster I can access something.txt from any notebook. If I do&lt;BR /&gt;ls /dbfs/mnt/azuremount/foo/bar/something.txt | head -n 2&lt;BR /&gt;in some notebook then databricks throws an error that there is no file uner the path "/dbfs/mnt/azuremount/foo/bar/something.txt".&lt;BR /&gt;But if call the following commands sequentially:&lt;BR /&gt;ls /dbfs/mnt/azuremount/ | head -n 2&lt;BR /&gt;ls /dbfs/mnt/azuremount/foo/ | head -n 2&lt;BR /&gt;ls /dbfs/mnt/azuremount/foo/bar/ | head -n 2&lt;BR /&gt;then I can magically open something.txt from any notebook!&lt;/P&gt;&lt;P&gt;I suspect that there must be some caching issue in the background but when I tried dbutils.fs.refreshMount() it didn't fix my issue. I want to find some solutions to this because it is incredibly annoying if I have to open multiple folders by hand all the time once I start my compute cluster.&lt;/P&gt;&lt;P&gt;Thanks,&lt;BR /&gt;Ben&lt;/P&gt;</description>
      <pubDate>Wed, 11 Dec 2024 13:03:02 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/databricks-mount-bug/m-p/101761#M40817</guid>
      <dc:creator>BenceCzako</dc:creator>
      <dc:date>2024-12-11T13:03:02Z</dc:date>
    </item>
    <item>
      <title>Re: Databricks mount bug</title>
      <link>https://community.databricks.com/t5/data-engineering/databricks-mount-bug/m-p/101769#M40822</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/135950"&gt;@BenceCzako&lt;/a&gt;,&lt;/P&gt;
&lt;P&gt;Have you tried by "Detach and re-attaching" the compute on the notebook? And what DBR version are you using?&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 11 Dec 2024 13:45:31 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/databricks-mount-bug/m-p/101769#M40822</guid>
      <dc:creator>Alberto_Umana</dc:creator>
      <dc:date>2024-12-11T13:45:31Z</dc:date>
    </item>
    <item>
      <title>Re: Databricks mount bug</title>
      <link>https://community.databricks.com/t5/data-engineering/databricks-mount-bug/m-p/101882#M40866</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;&lt;P&gt;That is actually my problem, sorry if I didn't explain properly.&lt;/P&gt;&lt;P&gt;So every time I attach/reattach/restart a cluster, I can't find the files under a given mount path only if I start to list things from the mount entrypoint using 'ls'. So this means that every time I start up my cluster in the beginning of the day, I can't reach my files unless I do what I mentioned in my original post.&lt;/P&gt;&lt;P&gt;I am using DBS 14.3. LTS.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 12 Dec 2024 09:35:06 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/databricks-mount-bug/m-p/101882#M40866</guid>
      <dc:creator>BenceCzako</dc:creator>
      <dc:date>2024-12-12T09:35:06Z</dc:date>
    </item>
    <item>
      <title>Re: Databricks mount bug</title>
      <link>https://community.databricks.com/t5/data-engineering/databricks-mount-bug/m-p/101999#M40928</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/135950"&gt;@BenceCzako&lt;/a&gt;,&lt;/P&gt;
&lt;P&gt;Thanks for the details, let me try to replicate it in my environment and get back.&lt;/P&gt;</description>
      <pubDate>Fri, 13 Dec 2024 02:30:47 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/databricks-mount-bug/m-p/101999#M40928</guid>
      <dc:creator>Alberto_Umana</dc:creator>
      <dc:date>2024-12-13T02:30:47Z</dc:date>
    </item>
    <item>
      <title>Re: Databricks mount bug</title>
      <link>https://community.databricks.com/t5/data-engineering/databricks-mount-bug/m-p/102325#M41070</link>
      <description>&lt;P&gt;Hello,&lt;BR /&gt;Were you able to reproduce the bug? Do you need additional information?&lt;/P&gt;</description>
      <pubDate>Tue, 17 Dec 2024 06:56:19 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/databricks-mount-bug/m-p/102325#M41070</guid>
      <dc:creator>BenceCzako</dc:creator>
      <dc:date>2024-12-17T06:56:19Z</dc:date>
    </item>
    <item>
      <title>Re: Databricks mount bug</title>
      <link>https://community.databricks.com/t5/data-engineering/databricks-mount-bug/m-p/106082#M42376</link>
      <description>&lt;P&gt;Hello,&lt;/P&gt;&lt;P&gt;Can you figure out the issue?&lt;/P&gt;</description>
      <pubDate>Fri, 17 Jan 2025 12:53:53 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/databricks-mount-bug/m-p/106082#M42376</guid>
      <dc:creator>BenceCzako</dc:creator>
      <dc:date>2025-01-17T12:53:53Z</dc:date>
    </item>
  </channel>
</rss>

