cancel
Showing results forย 
Search instead forย 
Did you mean:ย 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results forย 
Search instead forย 
Did you mean:ย 

Is it required to run Lakeflow Connect on Serverless?

yit337
Contributor

As the subject states, my question is:

Is it required to run the Ingestion Pipeline in Lakeflow Connect on Serverless compute? Cause I try to define my own cluster in the DAB, but it raises an error:

`Error: cannot create pipeline: You cannot provide cluster settings when using serverless compute.` 

Even though my serverless flag is set to false.

2 REPLIES 2

Ashwin_DSA
Databricks Employee
Databricks Employee

Hi @yit337,

For Lakeflow Connect managed ingestion pipelines, the behaviour youโ€™re seeing is expected. As soon as you use ingestion_definition in your pipeline (i.e., a managed ingestion pipeline), Databricks forces that pipeline to run on serverless compute. Serverless pipelines do not allow a clusters block. If you include cluster settings, you get the error you are seeing.

You can refer to this documentation page.

 

lakeflowserverless.png

 If this answer resolves your question, could you mark it as โ€œAccept as Solutionโ€? That helps other users quickly find the correct fix.

 

Regards,
Ashwin | Delivery Solution Architect @ Databricks
Helping you build and scale the Data Intelligence Platform.
***Opinions are my own***

saurabh18cs
Honored Contributor III

Yes โ€” Lakeflow Connect ingestion pipelines always run on Serverless compute. 

Databricks overrides your compute config and switches back to serverless,
because the ingestion connector requires it.
 
saurabh18cs_0-1774868648207.png