DLT refresh time for combination of streaming and non streaming tables?
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Sunday
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Sunday
the question is default DLT pipeline refresh time is 5seconds but if I use combination of streaming and non streaming data then will it still be 5 seconds?
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Tuesday
Hello @surajitDE!
When using both streaming and batch data, the pipeline may not always refresh every 5 seconds. While the streaming table (fact_stream) updates every 5 seconds, the batch table (dim_table) fully reloads each time, adding overhead from repeatedly loading the batch data.
The actual refresh time depends on the size of dim_table, larger tables take longer to reload, which can delay updates.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
2 hours ago
In a Delta Live Tables (DLT) continuous pipeline, does it make a difference if df_dim_prev (loaded in cell 1) is only read once at the start?
For example, if df_dim_prev is initialized as:
# Cell 1: Read dim_table once
df_dim_prev = spark.read.table("dim_table")
Then used in a streaming join inside a DLT table:
# Cell 2: DLT table with a streaming source
@Dlt.table def joined_table():
dim_df = df_dim_prev
# Using the preloaded dimension table
fact_df = spark.readStream.table("fact_stream")
return fact_df.join(dim_df, "id", "left")
Would this mean that dim_df remains static until the entire pipeline is refreshed, rather than updating dynamically as dim_table changes?
is there a better way to handle this if we want dim_table to update periodically in a continuous pipeline?

