cancel
Showing results forย 
Search instead forย 
Did you mean:ย 
Get Started Discussions
Start your journey with Databricks by joining discussions on getting started guides, tutorials, and introductory topics. Connect with beginners and experts alike to kickstart your Databricks experience.
cancel
Showing results forย 
Search instead forย 
Did you mean:ย 

Error when executing spark.readStream Script

giuseppe_esq
Contributor

Hi all,

When I try to execute the script (as per the screenshot below), in a Notebook cell, I get an error message.  I am using Databricks Free Edition, and am not sure if the error relates to the compute cluster that I am using?

giuseppe_esq_0-1756154296423.png

Any guidance would be greatly appreciated.

Thanks

Giuseppe

6 REPLIES 6

nayan_wylde
Honored Contributor

@giuseppe_esq 

Can you please copy the entire error message and paste here.

Yep, as requested.

giuseppe_esq_1-1756155159376.png

 

Thanks

 

brycejune
New Contributor III

Free Edition clusters often come with restricted resources and certain features may not be supported. You may want to review your cluster settings, check compatibility with your script, and consider whether upgrading to a paid tier is necessary for your use case.

Sidhant07
Databricks Employee
Databricks Employee

Hi @giuseppe_esq ,

Can you please share the DBR version, cluster configurations and log4 file containing the error stact trace to further review?

Hi,

Thanks for your reply.

Apologies if I am sending the incorrect information. I am in the process of learning databricks, and adding content to my personal blog.

DBR Version

giuseppe_esq_0-1756220649552.png

Cluster Configurations

From the Compute menu, this is the only cluster available in the Databricks Free Edition:

giuseppe_esq_1-1756221405050.png

From the Workspace Notebook, I  also tried to attach the General Compute option, though this didn't resolve the issue either:

giuseppe_esq_2-1756221534111.png

Log4j File

I have attached the Log4j File, for the time I executed the following query, that errored:

## Incrementally (or stream) data using Auto Loader
(spark.
readStream
.format("cloudFiles")
.option("cloudFiles.format", "csv")
.option("header", "true")
.option("sep", ",")
.option("inferSchema", "true")
.option("cloudFiles.schemaLocation", f"{checkpoint_file_location}")
.load("/Volumes/workspace/python_auto_loader/csv_staging")
.writeStream
.option("checkpointLocation", f"{checkpoint_file_location}")
.trigger(once=True)
.toTable("workspace.python_auto_loader.python_csv_autoloader")
)

If you require any additional information, please message me again.

Thank you for your help.

Giuseppe

 

Hi Sidhant,

Are there any updates on this please? 

Thanks for your time and help

Giuseppe