cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
cancel
Showing results for 
Search instead for 
Did you mean: 

Databricks now supports event-driven workloads, especially for loading cloud files from external locations. This means you can save costs and resource...

Hubert-Dudek
Esteemed Contributor III

Databricks now supports event-driven workloads, especially for loading cloud files from external locations. This means you can save costs and resources by triggering your Databricks jobs only when new files arrive in your cloud storage instead of mounting it as DBFS and polling it periodically. To use this feature, you need to follow these steps:

  • Add an external location for your ADLS2 container,
  • Make sure the storage credentials you use (such as Access Connector, service principal, or managed identity) have Storage Blob Data Contributor permissions for that container,
  • Make sure the account you use to run your workload has at least read files permission for the external location,
  • Write a notebook that loads cloud files from the external location,
  • Set a file arrival trigger for your workflow and specify the exact external location as the source.

With these steps, you can easily create and run event-driven workloads on Databricks.

ezgif-3-946af786d0 

2 REPLIES 2

jose_gonzalez
Moderator
Moderator

Thank you for sharing @Hubert Dudek​ 

Vartika
Moderator
Moderator

Hi @Hubert Dudek​ 

We really appreciate you sharing this bit of information.

Cheers!

Welcome to Databricks Community: Lets learn, network and celebrate together

Join our fast-growing data practitioner and expert community of 80K+ members, ready to discover, help and collaborate together while making meaningful connections. 

Click here to register and join today! 

Engage in exciting technical discussions, join a group with your peers and meet our Featured Members.