Hi @Phani1 , great catching you in the community!!
Migrating a Databricks workspace from Azure to GCP requires a significant time investment, especially when the workspace is mature and has multiple integrations, connections, and workflows.
The complexity of migration largely depends on how deeply integrated the workspace is with the Azure ecosystem. The greatest benefit comes when most of your enterprise's cloud services are already hosted on GCP.
Key Considerations:
- Consider Azure or GCP pricing and discount systems carefully.
- Consider the time and effort required for infrastructure deployment.
- All existing Databricks integrations with Azure systems need evaluation, as some connections may not work after migration, and require alternative solutions
Element to migrate:
1. Tables
- Use Storage Transfer Service for lift-and-shift migration or Delta Share for data cloning.
- Some tables need temporary parallel pipelines during migration (e.g., using Fivetran).
2. Notebooks and jobs
- Review logic thoroughly, particularly for external service connections.
- Leverage CLI or REST API for automated migration.
3. Security and governance
- Entra ID, Azure's primary identity service, isn't available in GCPโplan for manual migration of governance rules.
Limitations:
While most Azure services have GCP equivalents, migrating each service and setting up new credentials takes considerable time. Additionally, Azure's more mature Databricks ecosystem means GCP might need extra configuration or third-party tools to match functionality. Databricks deploys new features to Azure before GCP.
If you have any questions or comments, donโt hesitate to reach out. Feel free to message me directly in my inboxโIโd be happy to help!
Best,
Eliana Oviedo
Eliana Oviedo
Technical Partnerships Lead | SunnyData
P: (598) 95-974-524
E: eliana.oviedo@sunnydata.ai