Engage in discussions on data warehousing, analytics, and BI solutions within the Databricks Community. Share insights, tips, and best practices for leveraging data for informed decision-making.
We are looking for a option to copy tables more than 50 TB to be copied from syanpse to databricks on weekly basis , please suggest of there are any feasible ways for same
we are using connector but it is taking too long to copy
I suppose this is a Dedicated Sql pool you are talking about? If so, I'd use Data Factory to extract the data from synapse and land it in ADLS in parquet format f.e. Then create a UC/Hive table on this.
There is no databricks documentation on this as it is only involved for a very tiny bit: "CREATE TABLE catalog.schema.table USING PARQUET LOCATION 'url_to_the_parquet_files'. All the rest is done in Azure Data Factory, or you can even use the built-in Pipelines of Synapse (in Synapse Studio). There is a standard connector for Synapse and ADLS. The pipeline itself converts the data to parquet and writes is. When done, do the create table on databricks. Copy and transform data in Azure Synapse Analytics - Azure Data Factory & Azure Synapse | Microsoft ...
Connect with Databricks Users in Your Area
Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you wonโt want to miss the chance to attend and share knowledge.
If there isnโt a group near you, start one and help create a community that brings people together.