cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
cancel
Showing results for 
Search instead for 
Did you mean: 

Unable to save file in DBFS

learnerbricks
New Contributor II

I have took the azure datasets that are available for practice. I got the 10 days data from that dataset and now I want to save this data into DBFS in csv format. I have facing an error :

" No such file or directory: 'No such file or directory: '/dbfs/tmp/myfolder/mytest.csv'

but on the other hand if I am able to access the path directly from DBFS. This path is correct.

My code :

from azureml.opendatasets import NoaaIsdWeather
from datetime import datetime
from dateutil import parser 
from dateutil.relativedelta import relativedelta
 
spark.sql('DROP Table if exists mytest')
dbutils.fs.rm("/dbfs/tmp/myfolder",recurse = True)
 
basepath = "/dbfs/tmp/myfolder" 
 
try:
    dbutils.fs.ls(basepath)
except:
    dbutils.fs.mkdirs(basepath)
else:
   raise Exception("The Folder "+ basepath + " already exist, this notebook will remove in the end")
 
dbutils.fs.mkdirs("/dbfs/tmp/myfolder")
 
start_date = parser.parse('2020-5-1')
end_date = parser.parse('2020-5-10')
 
isd = NoaaIsdWeather(start_date, end_date)
pdf = isd.to_spark_dataframe().toPandas().to_csv("/dbfs/tmp/myfolder/mytest.csv")

What should I do ?

Thanks

4 REPLIES 4

Debayan
Esteemed Contributor III
Esteemed Contributor III

Hi, post running .mkdirs command, could you please confirm where the .csv file has been copied?

pardosa
New Contributor II

Hi @learnerbricks 

is this post already solved ? how you solved it?

youssefmrini
Honored Contributor III
Honored Contributor III

hello, you should really avoid writing data to dfbs.

pardosa
New Contributor II

Hi,

after some exercise you need to aware folder create in dbutils.fs.mkdirs("/dbfs/tmp/myfolder") it's created in /dbfs/dbfs/tmp/myfolder

if you want to access path to_csv("/dbfs/tmp/myfolder/mytest.csv") you should created with this script dbutils.fs.mkdirs('dbfs:/tmp/myfolder)