โ03-13-2024 12:24 PM
I have started to use DLT in a prototype framework and I now face the below challenge for which any help would be appreciated.
First let me give a brief context:
โ03-17-2024 01:25 PM
Thanks Kaniz to your response. It would have been great to use a similar approach like the widgets to a normal notebook. Specifying these parameters at design time does not allow the flexibility needed for running my DLT pipeline truly metadata-driven.
I was also going towards using the job REST API from a notebook but then I ended up tweaking my configuration tables in a way that I can utilise a hardcoded parameter in the DLT definition and still have it dynamic.
If the REST API call functionality could be integrated into the workflows later on to pass these values as to other tasks, that would be really great!
I accept it as a solution because your third suggestion would work. I still keep hoping a more integrated approach will come in the future ๐
โ08-27-2024 03:10 AM
Hi @MartinIsti , How did you manage tweaking the metadata to handle dynamically. Can you pls brief it out based on what you told is the below.
"I ended up tweaking my configuration tables in a way that I can utilise a hardcoded parameter in the DLT definition and still have it dynamic."
โ08-27-2024 02:50 PM
Sure, and for the record I'm still not fully happy with how parameters need to be set at design time.
As mentioned, I store the metadata in a .json file that I read using a standard notebook. The content of that I then save into DBFS as a delta table overwriting any previous version. Then the DLT notebook reads from that table and I only need to specify the name of the process (e.g. IngestAdventureWorks) and that name matches the name of the DLT pipeline itself (or it can be derived).
Once I determine which table to read from the DLT pipeline can be driven by the metadata in that table.
I still find dealing with DLTs inconsistent with orchestration of standard notebook-driven data handling, it is an odd-one out that mostly needs a slightly different way of handling but so far I have found a workaround for every of these small inconsistencies.
โ08-28-2024 10:42 PM
@MartinIsti thanks for your detailed explanation.
โ03-29-2024 12:00 PM
@Retired_mod Can you please provide some reference to REST API approach? I do not see that available on the docs.
TIA
Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you wonโt want to miss the chance to attend and share knowledge.
If there isnโt a group near you, start one and help create a community that brings people together.
Request a New Group