04-26-2023 02:01 AM
I have created an Azure data factory pipeline with a copy data function to copy data from adls path to a delta table .In the delta table drop downs i am able to see only the hive metastore database and tables only but the unity catalog tables are not listed .Can anyone please help me on this.
04-26-2023 10:07 PM
@Rama Teja Reddy Damireddy :
To see the Unity Catalog databases and tables in the drop-down in Azure Databricks Delta Lake, you need to configure the integration between Unity Catalog and Databricks. Here are the steps to configure the integration:
If you have already followed these steps and are still not seeing the Unity Catalog databases and tables, you may want to check if the user account you are using has the necessary permissions to access the Unity Catalog.
04-28-2023 06:20 AM
I also have the same question. I don't think you understand what is being asked. Within Azure Data Factory, when creating an Azure Databricks DeltaLake Dataset, there does not appear to be a way to select a database (schema) that is not in the hive_metastore catalog. See attached screenshot. Is UC not currently supported? I've sent the same question to Microsoft.
07-10-2023 04:17 AM
Hi, Were you able to solve the problem? I am also trying to do same
04-27-2023 12:49 AM
Hi @Rama Teja Reddy Damireddy
Thank you for posting your question in our community! We are happy to assist you.
To help us provide you with the most accurate information, could you please take a moment to review the responses and select the one that best answers your question?
This will also help other community members who may have similar questions in the future. Thank you for your participation and let us know if you need any further assistance!
01-05-2024 03:14 AM
Hi,
Exact same problem here, any updates please ?
Thanks
01-10-2024 06:29 AM
Hi,
I have the same issue.
Additional information: the linked service created in azure data factory using azure databricks deltalake connector is using system-managed-identity rather than Token.
Could we have an update?
Thank you in advance.
06-17-2024 03:01 PM
Dear community,
I am facing the same issue here.
My workaround is to just create an azure databricks activity in my ADF pipeline and use code inside a notebook to copy over data from tables inside unity catalog to ADLS gen2, but this is in my opinion less ideal then just using a azure databricks deltalake connector. Another workaround could be a view in the metastore that references tables in unity catalog, also not a very pretty solution.
is there some other solution that you would prescribe as best-practice for this usecase?
Thanks,
Bram
06-18-2024 02:46 AM
Hi,
I need to ingest data into Databricks Unity Catalog using ADF, so a workaround is to use an object storage as a Landingzone (ADLS gen2), that is mounted into Databricks under /mnt. I use ADF to copy data (incrementally) to this landingzone first, then use Databricks, to take this data using autoloader into Bronzezone which is under UnityCatalog. Of course, one would expect to perform this data ingestion directly to UnityCatalog schemas, using ADF, but this seems impossible, so far.
Thanks
07-05-2024 10:44 AM
I'm also experiencing the same issue. This article describes a method for getting access to Unity Catalog through an ODBC linked service that connects through DB SQL Warehouse. The article also indicates that direct Unity Catalog access via a Databricks Delta Lake linked service is limited depending on the security setup on the storage accounts used by Delta/DB. This isn't an optimal setup IMO but an acceptable workaround.
Note from the trenches: Securely read or write data from Databricks Unity Catalog using Azure Data F...
07-11-2024 02:22 PM
Hi all,
Please try following this syntax in the database field
<catalog_name>`.`<database_name>
It is important that you use ` not '
This should force the connect to use a Unity Catalog Database.
07-12-2024 06:43 AM
Woo hoo! Thank you so much for this suggestion. It is working now.
This seems like a bug in Azure Data Factory when it constructs the query to send to Databricks. The linked service connector must have been written pre-Unity catalog. We should be able to specify <catalog_name>, <database_name>, and <table_name> as separate parameters when configuring the linked service.
3 weeks ago
Chrisk, Are you using " system-assigned -managed identity" or "access token"?
07-12-2024 07:57 AM
That's crazy, but it just works !!!! Thanks so much, Microsoft should include this trick in its documentation.
3 weeks ago
@StdyFriend1 I have tried your solution, but no luck. I am using system-assigned -managed identity in the linked service created in Azure data factory using "Azure Databricks delta lake connector".
Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.
If there isn’t a group near you, start one and help create a community that brings people together.
Request a New Group