cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Ingest data from SQL Server

Vamsi_S
New Contributor

I've been working on data ingestion from SQL Server to UC using lakeflow connect. Lakeflow connect actually made the work easier when everything is right. I am trying to incorporate this with DAB and this would work fine with schema and table tags for 'object' in 'ingestion_definition'. What if someone wants to clean the table names before ingesting them into UC? For now, 'object' only allows 'report', 'table', 'schema'. If I create a python file or notebook that gets the table names from SQL Server using 'source_schema', and then clean/modify the table names, I would not be able to dynamically ingest tables. I understand that DAB doesn't allow runtime ingestion/modification of files, but I would love to have a way so that I can do something like below.
pipeline_sqlserver:
name: sqlserver-ingestion-pipeline
ingestion_definition:
ingestion_gateway_id: ${resources.pipelines.gateway.id}
objects: [] 
    library:
         file: .yml/.json/any other format

Link to Ingestion documentation: Ingest data from SQL Server | Databricks on AWS

Give objects an empty list and get the names from a yaml file or json file or any other format that tables can be saved into from preprocessed script.
If anyone had already faced the same issue and have a solution to it, I would greatly appreciate if you can share it here. Thanks!

1 ACCEPTED SOLUTION

Accepted Solutions

Khaja_Zaffer
Contributor

Hello @Vamsi_S 

Good day! Did you Preprocessing Table Names in CI/CD and Generate YAML Dynamically (Recommended for Dynamic, Automated Ingestion)

Did you contact your databricks account manager (incase if you working with a company) for feature request?

View solution in original post

2 REPLIES 2

Khaja_Zaffer
Contributor

Hello @Vamsi_S 

Good day! Did you Preprocessing Table Names in CI/CD and Generate YAML Dynamically (Recommended for Dynamic, Automated Ingestion)

Did you contact your databricks account manager (incase if you working with a company) for feature request?

Vamsi_S
New Contributor

Hey @Khaja_Zaffer 

Thanks for sharing your thoughts. Yes, dynamic generation of YAML works in this case, but I wanted to know if there is any other workaround that is officially made available by Databricks.

I did not contact anyone from Databricks. Before reaching out I wanted to try all different possibilities. 
Thank you again!

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now