I massively underestimated Week 3's content 🤣. Just got this one ticked off now, so I'll finish off week 4 throughout this week. Not as much as a "speedrun" as I anticipated. The week 3 content consisted of "Build Data Pipelines with Lakeflow Declarative Pipelines".

I've gotta say, I'm really really excited to use these. As I was going through the content I had loads of ideas and can see there's definitely a lot of mastery needed. An example, I setup a pipeline, but I wanted to alter a column's datatype on a streaming table 😔🤔. Not quite as easy as one would think. Got me thinking about how I'd then design an alternate solution.
Started making me think alot around backfilling tables, once they're in production. Also, what happens with _rescued_data? How would we cater for this etc.
The Expectations were a really cool find & I loved the Pipeline Event Logs. There's alot to consider with joining streaming tables as well. CDC was super cool with the APPLY CHANGES INTO although this seems to be replaced with AUTO CDC when I checked it via the docs: https://docs.databricks.com/aws/en/dlt/cdc . Also saw how everything can be executed through code, i.e. databricks asset bundles or SDK/CLI. We can also choose Python instead of SQL when building these things. So much to conquer 😈🤣.
There's a lot around DLT (declarative pipelines) so I'm excited to build some projects. The labs were a great help for getting me started. So yeah, DLT is not one for a speed run 🤣.
Till next week ☺️👀.
All the best,
BS