Migrating to Unity Catalog: Read-Only Connections to SQL Server and Snowflake
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
โ08-23-2024 02:00 AM
We are in the process of migrating to Unity Catalog, establishing connections to SQL Server and Snowflake, and creating foreign catalogs that mirror your SQL Server and Snowflake databases. This allows us to leverage Unity Catalogโs query syntax and data governance tools to manage Databricks user access.
However, these features are read-only. This raises an important question: What solutions are people using when thereโs a need to write back to SQL Server or Snowflake?
So far, weโve used JDBC for writing, but this approach lacks the governance provided by Unity Catalog for reads and still requires using Key Vaults for credential management. And if you have different user groups with different access levels, that creates issues as you need a keyvault for each access group or in some casess even single user.
Is there a better way to write to SQL Server from Databricks that offers the same level of governance and control as Unity Catalog provides for reads or writes to data lakes?
Do you know whether this feature is in the Databricks plans?
Weโd love to hear your thoughts!
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
โ09-20-2024 08:53 AM
We just use SQLAlchemy for connection to Snowflake, which, you're right, does not enable databricks governance.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
โ12-03-2024 06:58 AM
Iโm also trying to figure out if this is a limitation in Unity Catalog. I recently used a JDBC URL to write data to an Amazon Aurora PostgreSQL database, but noticed that no entries appeared in the `system.access.table_lineage` table. Has anyone else encountered this issue? If so, how are you handling lineage tracking for JDBC-based writes?

