cancel
Showing results forย 
Search instead forย 
Did you mean:ย 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results forย 
Search instead forย 
Did you mean:ย 

Cloud_files function

Jefke
New Contributor II

Hi 

I'm fairly new to to Databricks and in some examples, blogs,... I see the cloud_files() function being used. But I'm always unable to find any documentation on it? Is there any reason for this? 

And what is the exact use case for the function? Most examples seem to have to do with DLT.

Thanks

2 ACCEPTED SOLUTIONS

Accepted Solutions

Hi @Jefke ,

You can't find any information because it's a deprecated function. You should use read_files instead. 

https://learn.microsoft.com/en-us/azure/databricks/delta-live-tables/load#load-files-from-cloud-obje...

View solution in original post

JissMathew
New Contributor II

Hi @Jefke ,

The cloud_files() function in Databricks is part of the Databricks Auto Loader, a tool used for incremental data ingestion from cloud storage like Azure Blob Storage, Amazon S3, or Google Cloud Storage. This function is specifically optimized for streaming or continuous loading of files, making it popular in Delta Live Tables (DLT) pipelines and other data engineering workflows.

 

View solution in original post

5 REPLIES 5

JakubSkibicki
New Contributor II

Hi, cloud_files functions is releated to autoloader. You can use it to read from checkpoint folder files that autoloader extracted.

Jefke
New Contributor II

Hi, you are talking about the cloud_files_state function whereas I was referring to the cloud_files function. You see it sometimes being used in examples in the docs like in the example below. If you search for it in the docs, you mostly end up with the cloud_files_state function you did mention. But that's something completely different.

Is this a deprecated function? If so, when was it announced? I was just wondering why you often see it being used in examples but there is no trace of it in the docs...

Youโ€™ve gotten familiar with Delta Live Tables (DLT) via the quickstart and getting started guide. Now itโ€™s time to tackle creating a DLT data pipeline for your cloud storageโ€“with one line of code. Hereโ€™s how itโ€™ll look when you're starting:
CREATE OR REFRESH STREAMING LIVE TABLE <table_name>
AS SELECT * FROM cloud_files('<cloud storage location>', '<format>')

https://community.databricks.com/t5/data-engineering/delta-live-tables-s3-5-tips-for-cloud-storage-w...

Hi @Jefke ,

You can't find any information because it's a deprecated function. You should use read_files instead. 

https://learn.microsoft.com/en-us/azure/databricks/delta-live-tables/load#load-files-from-cloud-obje...

JakubSkibicki
New Contributor II

JissMathew
New Contributor II

Hi @Jefke ,

The cloud_files() function in Databricks is part of the Databricks Auto Loader, a tool used for incremental data ingestion from cloud storage like Azure Blob Storage, Amazon S3, or Google Cloud Storage. This function is specifically optimized for streaming or continuous loading of files, making it popular in Delta Live Tables (DLT) pipelines and other data engineering workflows.

 

Connect with Databricks Users in Your Area

Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you wonโ€™t want to miss the chance to attend and share knowledge.

If there isnโ€™t a group near you, start one and help create a community that brings people together.

Request a New Group