cancel
Showing results for 
Search instead for 
Did you mean: 
Community Platform Discussions
Connect with fellow community members to discuss general topics related to the Databricks platform, industry trends, and best practices. Share experiences, ask questions, and foster collaboration within the community.
cancel
Showing results for 
Search instead for 
Did you mean: 

Suggest ways to get unity catalog data to Aws s3 or sagemaker

Sa_1234
New Contributor II

Please suggest best ways to get databricks unity catalog data to Aws s3 or sagemaker. Data could be around 1gb in some tables and 20gb in others.

currently sagemaker pipelines use data from s3 as batches in different parquet files. But now we would like to keep the sagemaker pipelines as is but get the data from unity catalog. Please suggest the best possible ways to do this. Thanks in advance.

What we tried earlier: delta sharing (unsure if it could work well with this huge data-kindly suggest on this and also the permissions to setup)

3 REPLIES 3

Walter_C
Databricks Employee
Databricks Employee

You could try by using Delta Sharing with your provider as mentioned in doc https://docs.databricks.com/en/delta-sharing/set-up.html 

 

Sa_1234
New Contributor II

Thanks for your response. It helps. Also please suggest on unity catalog open api as well. As we are trying feasibility analysis because the data is huge, would like to have alternatives at hand to try further. Other options to do this use case are welcome as well. Thanks.

Sa_1234
New Contributor II

Hi team any suggestions on the last comment please?

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now