Hi @MauricioS
Absolutely not a noob question โ you're touching on a common and important challenge in DLT pipelines,
especially when dealing with large shared Delta tables and incremental ingestion from Unity Catalog sources.
Letโs break it down so itโs simple, scalable, and DLT-native.
Ingest from a shared Delta table (Unity Catalog) into your own catalog, incrementally, with daily updates, using DLT.
Best Practice with DLT Pipelines (Incremental Load)
Step 1: Use STREAMING LIVE TABLE to Enable Incremental Load
DLT supports incremental ingestion natively via streaming reads, even if the source table is not a streaming table.
DLT tracks offsets/checkpoints automatically, so you don't reprocess old data.
Step 2: Optional Watermark for Late Records
If you have late-arriving data, you can use watermarks to prevent reprocessing historical rows:
Step 3: Use DLT Expectations for Quality
Step 4: Materialize to Your Catalog
Make sure your DLT pipeline is writing to your own Unity Catalog schema:
DLT Handles Incrementals for You
You donโt need to manually track last_updated_at or store bookmarks โ DLT uses checkpoints for streaming sources and only reads new data.
However, your source table must support:
-- Delta format
-- Append or CDC-compatible operations (if using change_data_feed = true)
If Source Supports Change Data Feed (CDF):
Enable CDF if the source table supports it (or ask the upstream team to enable):
LR