I created a separate pipeline notebook to generate the table via DLT, and a separate notebook to write the entire output to redshift at the end. The table created via DLT is called spark.read.table("{schema}.{table}").
This way, I can import
[MATERIALIZED_VIEW_OPERATION_NOT_ALLOWED.REQUIRES_SHARED_COMPUTE] The materialized view operation query is not allowed: Cannot query the Materialized View main.voyager.shop_silver from an Assigned or No Isolation Shared cluster, please use a Shared cluster or a Databricks SQL warehouse instead.
I get this error. How can I fix it?