<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Databricks interoperability between cloud environments in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/databricks-interoperability-between-cloud-environments/m-p/33901#M24802</link>
    <description>&lt;P&gt;While Databricks is currently available and integrated into all three major cloud platforms (Azure, AWS, GCP) , following are pertinent questions that comes across in the real-world scenarios,&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;1)  Whether Databricks can be cloud agnostic (i.e.,) In case we develop notebooks for data engineering pipelines with Azure databricks - Can the same be exported and used in other cloud platforms in case the client wants to switch the cloud platform at a later point in time?&lt;/P&gt;&lt;P&gt;2) How about the delta lake / tables support in these environments in the above scenario - Say for example: if the root folder containing the delta table data is copied to a different cloud storage (ex: ADLS to S3) and a delta table is created on top of this data - will it work seamlessly?&lt;/P&gt;&lt;P&gt;3)  In another case of a multi cloud environment - what are the possibilities of using the same notebooks /code in the databricks workspace across the cloud environments&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Are there are any documentation available for the above pointers - will be helpful to get a reference for the same if available.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;</description>
    <pubDate>Thu, 02 Dec 2021 08:51:55 GMT</pubDate>
    <dc:creator>Murugan</dc:creator>
    <dc:date>2021-12-02T08:51:55Z</dc:date>
    <item>
      <title>Databricks interoperability between cloud environments</title>
      <link>https://community.databricks.com/t5/data-engineering/databricks-interoperability-between-cloud-environments/m-p/33901#M24802</link>
      <description>&lt;P&gt;While Databricks is currently available and integrated into all three major cloud platforms (Azure, AWS, GCP) , following are pertinent questions that comes across in the real-world scenarios,&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;1)  Whether Databricks can be cloud agnostic (i.e.,) In case we develop notebooks for data engineering pipelines with Azure databricks - Can the same be exported and used in other cloud platforms in case the client wants to switch the cloud platform at a later point in time?&lt;/P&gt;&lt;P&gt;2) How about the delta lake / tables support in these environments in the above scenario - Say for example: if the root folder containing the delta table data is copied to a different cloud storage (ex: ADLS to S3) and a delta table is created on top of this data - will it work seamlessly?&lt;/P&gt;&lt;P&gt;3)  In another case of a multi cloud environment - what are the possibilities of using the same notebooks /code in the databricks workspace across the cloud environments&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Are there are any documentation available for the above pointers - will be helpful to get a reference for the same if available.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;</description>
      <pubDate>Thu, 02 Dec 2021 08:51:55 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/databricks-interoperability-between-cloud-environments/m-p/33901#M24802</guid>
      <dc:creator>Murugan</dc:creator>
      <dc:date>2021-12-02T08:51:55Z</dc:date>
    </item>
    <item>
      <title>Re: Databricks interoperability between cloud environments</title>
      <link>https://community.databricks.com/t5/data-engineering/databricks-interoperability-between-cloud-environments/m-p/33902#M24803</link>
      <description>&lt;P&gt;I have few thoughts:&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;You can use the same runtime version so spark version for sure will be the same in all clouds,&lt;/LI&gt;&lt;LI&gt;Problem can be sometimes with additional new Databricks features like SQL alarms etc. as they are introduced in different times depending on platform (usually it is not such a problem as something is new usually in the beginning you use it only for development/training purposes),&lt;/LI&gt;&lt;LI&gt;Regarding storage you still can mount S3 on Azure and Azure Data Lake in AWS just connections will be not through dedicated private endpoints and traffic-out charges will apply,&lt;/LI&gt;&lt;LI&gt;Regarding code "Repo" will make your live easier as all notebooks etc. can be hosted on git (github, code commit, Azure devops or somewhere else)&lt;/LI&gt;&lt;LI&gt;The biggest but also solvable issue I see in code to automatically deploy pools and clusters through CLI as server spec differ between platform (VM name, hard drive name) so you could make script for that with some mapping (what VM in Azure = what VM in AWS)&lt;/LI&gt;&lt;/UL&gt;</description>
      <pubDate>Thu, 02 Dec 2021 10:32:33 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/databricks-interoperability-between-cloud-environments/m-p/33902#M24803</guid>
      <dc:creator>Hubert-Dudek</dc:creator>
      <dc:date>2021-12-02T10:32:33Z</dc:date>
    </item>
    <item>
      <title>Re: Databricks interoperability between cloud environments</title>
      <link>https://community.databricks.com/t5/data-engineering/databricks-interoperability-between-cloud-environments/m-p/33903#M24804</link>
      <description>&lt;P&gt;Databricks wants to avoid vendor lock-in, so in theory it is cloud platform agnostic.&lt;/P&gt;&lt;P&gt;However, this does not just work.  You have to think about all the configuration you did on your databricks workspace and do the same configuration on the other cloud platform, well not literally the same but conceptually the same (f.e. ADLS vs S3, firewalls, git, jars, ...).&lt;/P&gt;&lt;P&gt;The code itself will work.  I have no knowledge of certain possibilities not being available on a cloud provider, preview functionalities not included!&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;The cool part is that DBFS, on which databricks works is a semantic layer over your physical storage.&lt;/P&gt;&lt;P&gt;So as long as your DBFS paths are the same over the providers you will be ok.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;But any config should be taken into account.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;</description>
      <pubDate>Thu, 02 Dec 2021 15:39:30 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/databricks-interoperability-between-cloud-environments/m-p/33903#M24804</guid>
      <dc:creator>-werners-</dc:creator>
      <dc:date>2021-12-02T15:39:30Z</dc:date>
    </item>
    <item>
      <title>Re: Databricks interoperability between cloud environments</title>
      <link>https://community.databricks.com/t5/data-engineering/databricks-interoperability-between-cloud-environments/m-p/33904#M24805</link>
      <description>&lt;P&gt;" certain possibilities not being available on a cloud provider" roadmaps are different for every platform. I think today situation is that all platforms are the same but probably it can be a bit different after Christmas &lt;span class="lia-unicode-emoji" title=":slightly_smiling_face:"&gt;🙂&lt;/span&gt;&lt;/P&gt;</description>
      <pubDate>Thu, 02 Dec 2021 19:37:13 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/databricks-interoperability-between-cloud-environments/m-p/33904#M24805</guid>
      <dc:creator>Hubert-Dudek</dc:creator>
      <dc:date>2021-12-02T19:37:13Z</dc:date>
    </item>
    <item>
      <title>Re: Databricks interoperability between cloud environments</title>
      <link>https://community.databricks.com/t5/data-engineering/databricks-interoperability-between-cloud-environments/m-p/33905#M24806</link>
      <description>&lt;P&gt;You'll be interested in the &lt;A href="https://databricks.com/product/unity-catalog" alt="https://databricks.com/product/unity-catalog" target="_blank"&gt;Unity Catalog&lt;/A&gt;.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;The notebooks should be the same across all the clouds and there are no syntax differences.  The key things are going to be just changing paths from S3 to ADL2 and having different usernames/logins across the different accounts as werners mentioned above.&lt;/P&gt;</description>
      <pubDate>Fri, 03 Dec 2021 12:49:07 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/databricks-interoperability-between-cloud-environments/m-p/33905#M24806</guid>
      <dc:creator>Anonymous</dc:creator>
      <dc:date>2021-12-03T12:49:07Z</dc:date>
    </item>
  </channel>
</rss>

