cancel
Showing results forย 
Search instead forย 
Did you mean:ย 
Get Started Discussions
Start your journey with Databricks by joining discussions on getting started guides, tutorials, and introductory topics. Connect with beginners and experts alike to kickstart your Databricks experience.
cancel
Showing results forย 
Search instead forย 
Did you mean:ย 

Whatโ€™s the easiest way to import my local dataset into Databricks for analysis?

Suheb
New Contributor II

How can I upload a data file (like a CSV, Excel, or JSON file) from my computer into Databricks so I can start analyzing it there?

4 REPLIES 4

szymon_dybczak
Esteemed Contributor III

Hi @Suheb ,

I think the easiest way is to use Data Ingestion tab:

szymon_dybczak_0-1762241568538.png

And here you will be able to upload your local file:

szymon_dybczak_1-1762241600707.png

For larger files, for other file formats, or for uploading files to a non-tabular dataset without creating a table the recommended approach is to use upload to a Volume in Unity Catalog.

ShaneCorn
New Contributor III

The easiest way to import a local dataset into Databricks is by uploading it directly to the workspace. First, navigate to the Databricks workspace, then go to the "Data" tab. Click on "Add Data," select "Upload File," and choose your dataset from your local machine. Once uploaded, Databricks will store it in DBFS (Databricks File System). You can then access the file via Spark DataFrames using spark.read.csv() or other methods, depending on your file format. Alternatively, you can use the Databricks CLI or APIs for automation.

bianca_unifeye
New Contributor II

Hi @Suheb ,

The easiest way to bring your local dataset (CSV, Excel, JSON, etc.) into Databricks is by creating a Volume, uploading your files there, and then creating a table on top of that data.

Hereโ€™s how you can do it step by step:

  1. Create a Volume (recommended approach)
  • In the Catalog Explorer, go to your schema (e.g., catalog.my_schema).

  • Click Create โ†’ Volume and give it a name (e.g., raw_data).

  • A Volume provides a secure, governed storage location for your files, managed by Unity Catalog.

2. Upload your local file

  • Open the Volume you just created.

  • Click Upload โ†’ Choose file, then select your CSV, Excel, or JSON file from your computer.

  • Databricks will store it in that Volume path ( Click and copy path).

3. Create a Table referencing the file

  • Once uploaded, you can create a table directly in SQL:

 

 
CREATE TABLE main.default.my_table AS SELECT * FROM read_files('/Volumes/.../raw_data/', format => 'csv', inferSchema => true, header => true);

4. Keep adding new files (optional)

  • If you upload more files into the same Volume folder, the table will automatically read all of them, perfect for incremental uploads.

5.Alternative option

  • You can also use the โ€œUpload Dataโ€ button from the Workspace UI (Data tab) if you just want a quick, one-off import. But Volumes are best if you plan to keep the data managed and reusable.

Databrick5
New Contributor

For large size relational Databases like from Oracle how can we move to Azure databricks. Non-unity catalog to unity catalog?