โ02-08-2016 07:45 AM
I couldn't find in documentation a way to export an RDD as a text file to a local folder by using python. Is it possible?
โ02-08-2016 11:20 AM
Sounds like you're looking for saveAsTextFile().
Refer to the documentation here:
https://spark.apache.org/docs/latest/api/python/pyspark.html#pyspark.RDD.saveAsTextFile
โ02-08-2016 02:33 PM
You can use the FileStore to save a data file to that location and retrieve it by going to your shard and retrieve it there. Look at the Databricks Guide -> Product Overview -> FileStore.
You can access the files saved there by going to:
https://xxxxx.cloud.databricks.com/files/folder/specific_file
โ06-09-2017 08:08 AM
What is the code to save a data object to the file store? I have a list object with JSON elements that I want to save to local disk, but am unable to do so.
โ03-31-2016 07:23 AM
Hello all, hello โ@Miklos_Cโ ,
How do you access the files? I don't understand what is the "xxxx" in the URL "https://xxxxx.cloud.databricks.com/files/folder/specific_file". Do we need to replace it with the 16-numbers sequence that identifies us? With our username/mail adress? I'm using Databricks Community Edition, do I need to put the string "community" somewhere?
Thank you for your answer
Have a nice day
โ10-13-2016 05:28 AM
Do you know what is xxxxx in the URL?, I have the same problem.
Regards
Diego
โ10-13-2016 06:45 AM
Simply replace it with "community"
Thibaultโ10-19-2016 01:22 PM
Thanks a lot. It works !!
โ10-20-2016 09:08 AM
do you know how to acces the filestore with wget?
โ10-20-2016 09:35 AM
Absolutely not... ๐
โ10-20-2016 11:25 AM
This won't work because you'd have to authenticate with Databricks in order to download it. This is suitable for doing things like loading javascript libraries but not for extracting data from Databricks. In order to download data you should connect to amazon S3 or use the DBFS api.
http://docs.databricks.com/spark/latest/data-sources/amazon-s3.html
โ10-24-2016 12:58 PM
Thank you for the answer.
โ05-31-2017 11:51 AM
I use s3 as an intermediary. rdd.saveAsTextFile("s3a://...")
โ02-04-2018 08:51 AM
there should really be a way to simply export a file to your desktop
โ03-25-2019 08:18 AM
To: Export a file to local desktop
Workaround : Basically you have to do a "Create a table in notebook" with DBFS
The steps are:
(Tweak this notebook to download different files)
Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you wonโt want to miss the chance to attend and share knowledge.
If there isnโt a group near you, start one and help create a community that brings people together.
Request a New Group