<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Delta Live Tables with Common Data Model as source in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/delta-live-tables-with-common-data-model-as-source/m-p/71827#M34413</link>
    <description>&lt;P&gt;I've been looking into this myself for a while, depending on the requirements this new technical blog and code from&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/71426"&gt;@SergioSchena&lt;/a&gt;&amp;nbsp;might help.&lt;/P&gt;&lt;P&gt;&lt;A href="https://community.databricks.com/t5/technical-blog/how-to-efficiently-ingest-dataverse-common-data-model-cdm-tables/ba-p/66671" target="_blank"&gt;How to efficiently ingest Dataverse Common Data Mo... - Databricks Community - 66671&lt;/A&gt;&lt;/P&gt;</description>
    <pubDate>Thu, 06 Jun 2024 06:01:26 GMT</pubDate>
    <dc:creator>dalebidwell</dc:creator>
    <dc:date>2024-06-06T06:01:26Z</dc:date>
    <item>
      <title>Delta Live Tables with Common Data Model as source</title>
      <link>https://community.databricks.com/t5/data-engineering/delta-live-tables-with-common-data-model-as-source/m-p/39423#M26973</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;&lt;P&gt;I'm investigating whether it's possible to use Common Data Model CDM (in particular the Dynamics 365 exported csv and cdm files) as a Delta Live Tables data source? Can someone point me in the right direction?&lt;/P&gt;&lt;P&gt;Thanks!&lt;/P&gt;</description>
      <pubDate>Wed, 09 Aug 2023 08:30:15 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/delta-live-tables-with-common-data-model-as-source/m-p/39423#M26973</guid>
      <dc:creator>BobBubble2000</dc:creator>
      <dc:date>2023-08-09T08:30:15Z</dc:date>
    </item>
    <item>
      <title>Re: Delta Live Tables with Common Data Model as source</title>
      <link>https://community.databricks.com/t5/data-engineering/delta-live-tables-with-common-data-model-as-source/m-p/45607#M27948</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/86148"&gt;@BobBubble2000&lt;/a&gt;&amp;nbsp;&lt;BR /&gt;The MSFT Azure document briefly mentioned that Spark Common Data Model connector natively supports Azure Synapse Analytics. On the other hand,&amp;nbsp;&lt;SPAN&gt;Azure Databricks as a data consumer to read data from the&amp;nbsp;&lt;/SPAN&gt;&lt;SPAN&gt;Common Data Model folders in Data Lake Storage Gen2, this should be achievable&amp;nbsp;as well. You can firstly look for a way to install&amp;nbsp;Spark CDM connector on Azure Databricks, and then you can read CMD data into a dataframe. If you can achieve this step, you can also import dlt module as well as the Spark CMD connector in your DLT notebook and turn the CMD data into a live table in your DLT pipeline.&amp;nbsp;&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;Reference:&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;&lt;A href="https://github.com/Azure/spark-cdm-connector/blob/spark3.1/documentation/overview.md" target="_blank"&gt;https://github.com/Azure/spark-cdm-connector/blob/spark3.1/documentation/overview.md&lt;/A&gt;&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;&lt;A href="https://learn.microsoft.com/en-us/common-data-model/data-lake" target="_blank"&gt;https://learn.microsoft.com/en-us/common-data-model/data-lake&lt;/A&gt;&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 22 Sep 2023 03:30:32 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/delta-live-tables-with-common-data-model-as-source/m-p/45607#M27948</guid>
      <dc:creator>SparkJun</dc:creator>
      <dc:date>2023-09-22T03:30:32Z</dc:date>
    </item>
    <item>
      <title>Re: Delta Live Tables with Common Data Model as source</title>
      <link>https://community.databricks.com/t5/data-engineering/delta-live-tables-with-common-data-model-as-source/m-p/49523#M28574</link>
      <description>&lt;P&gt;hello, I am&amp;nbsp; currently looking the same task. Have you been successful in installing the connector in databricks?&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 19 Oct 2023 10:38:18 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/delta-live-tables-with-common-data-model-as-source/m-p/49523#M28574</guid>
      <dc:creator>developer_bi</dc:creator>
      <dc:date>2023-10-19T10:38:18Z</dc:date>
    </item>
    <item>
      <title>Re: Delta Live Tables with Common Data Model as source</title>
      <link>https://community.databricks.com/t5/data-engineering/delta-live-tables-with-common-data-model-as-source/m-p/71827#M34413</link>
      <description>&lt;P&gt;I've been looking into this myself for a while, depending on the requirements this new technical blog and code from&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/71426"&gt;@SergioSchena&lt;/a&gt;&amp;nbsp;might help.&lt;/P&gt;&lt;P&gt;&lt;A href="https://community.databricks.com/t5/technical-blog/how-to-efficiently-ingest-dataverse-common-data-model-cdm-tables/ba-p/66671" target="_blank"&gt;How to efficiently ingest Dataverse Common Data Mo... - Databricks Community - 66671&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Thu, 06 Jun 2024 06:01:26 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/delta-live-tables-with-common-data-model-as-source/m-p/71827#M34413</guid>
      <dc:creator>dalebidwell</dc:creator>
      <dc:date>2024-06-06T06:01:26Z</dc:date>
    </item>
    <item>
      <title>Re: Delta Live Tables with Common Data Model as source</title>
      <link>https://community.databricks.com/t5/data-engineering/delta-live-tables-with-common-data-model-as-source/m-p/71829#M34414</link>
      <description>&lt;P&gt;&lt;STRONG&gt;Using Delta Live Tables with Common Data Model (CDM) as a Source in Databricks&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;I'm investigating the use of Delta Live Tables (DLT) to process Common Data Model (CDM) files exported from Dynamics 365, and I found a solution that works well. Here’s a quick guide on how to set it up:&lt;/P&gt;&lt;H3&gt;Prerequisites:&lt;/H3&gt;&lt;OL&gt;&lt;LI&gt;&lt;STRONG&gt;Azure Databricks&lt;/STRONG&gt; workspace&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Azure Data Lake Storage&lt;/STRONG&gt; (ADLS) to store your CDM files&lt;/LI&gt;&lt;/OL&gt;&lt;H3&gt;Steps:&lt;/H3&gt;&lt;OL&gt;&lt;LI&gt;&lt;P&gt;&lt;STRONG&gt;Store Data in ADLS&lt;/STRONG&gt;: Upload your Dynamics 365 exported CSV and CDM files to ADLS.&lt;/P&gt;&lt;/LI&gt;&lt;LI&gt;&lt;P&gt;&lt;STRONG&gt;Mount ADLS in Databricks&lt;/STRONG&gt;: Use the following code to mount your ADLS in Databricks:&lt;/P&gt;&lt;/LI&gt;&lt;/OL&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;&amp;nbsp;&lt;/P&gt;&lt;P&gt;configs = {&lt;BR /&gt;"fs.azure.account.auth.type": "OAuth",&lt;BR /&gt;"fs.azure.account.oauth.provider.type": "org.apache.hadoop.fs.azurebfs.oauth2.ClientCredsTokenProvider",&lt;BR /&gt;"fs.azure.account.oauth2.client.id": "&amp;lt;client-id&amp;gt;",&lt;BR /&gt;"fs.azure.account.oauth2.client.secret": "&amp;lt;client-secret&amp;gt;",&lt;BR /&gt;"fs.azure.account.oauth2.client.endpoint": "&lt;A href="https://login.microsoftonline.com/" target="_blank"&gt;https://login.microsoftonline.com/&lt;/A&gt;&amp;lt;tenant-id&amp;gt;/oauth2/token",&lt;BR /&gt;}&lt;/P&gt;&lt;P&gt;dbutils.fs.mount(&lt;BR /&gt;source = "abfss://&amp;lt;container-name&amp;gt;@&amp;lt;storage-account-name&amp;gt;.dfs.core.windows.net/",&lt;BR /&gt;mount_point = "/mnt/datalake",&lt;BR /&gt;extra_configs = configs&lt;BR /&gt;)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;&lt;P&gt;&lt;STRONG&gt;Create a Delta Live Table Pipeline&lt;/STRONG&gt;: Navigate to the "Jobs" section in Databricks and create a new "Delta Live Tables" pipeline.&lt;/P&gt;&lt;/LI&gt;&lt;LI&gt;&lt;P&gt;&lt;STRONG&gt;Define Pipeline Configuration&lt;/STRONG&gt;: Define the pipeline to read and process the CDM files:&lt;/P&gt;&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;&amp;nbsp;&lt;/P&gt;&lt;P&gt;import dlt&lt;BR /&gt;from pyspark.sql.functions import col&lt;/P&gt;&lt;P&gt;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/97035"&gt;@Dlt&lt;/a&gt;.table(comment="Read Account data from CDM")&lt;BR /&gt;def account():&lt;BR /&gt;return (spark.read.csv("/mnt/datalake/CDMExports/Account.csv", header=True, inferSchema=True)&lt;BR /&gt;.withColumnRenamed("AccountID", "account_id")&lt;BR /&gt;.withColumnRenamed("AccountName", "account_name"))&lt;/P&gt;&lt;P&gt;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/97035"&gt;@Dlt&lt;/a&gt;.table(comment="Read Contact data from CDM")&lt;BR /&gt;def contact():&lt;BR /&gt;return (spark.read.csv("/mnt/datalake/CDMExports/Contact.csv", header=True, inferSchema=True)&lt;BR /&gt;.withColumnRenamed("ContactID", "contact_id")&lt;BR /&gt;.withColumnRenamed("ContactName", "contact_name"))&lt;/P&gt;&lt;P&gt;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/97035"&gt;@Dlt&lt;/a&gt;.table(comment="Join Account and Contact data")&lt;BR /&gt;def account_contact():&lt;BR /&gt;account_df = dlt.read("account")&lt;BR /&gt;contact_df = dlt.read("contact")&lt;BR /&gt;return account_df.join(contact_df, account_df.account_id == contact_df.account_id)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;OL&gt;&lt;LI&gt;&lt;P&gt;&lt;STRONG&gt;Deploy and Run the Pipeline&lt;/STRONG&gt;: Deploy the pipeline from the Databricks UI. This will read the CDM files, apply the transformations, and write the results to Delta tables.&lt;/P&gt;&lt;/LI&gt;&lt;/OL&gt;&lt;P&gt;This setup allows you to leverage Databricks and Delta Lake capabilities to process and transform CDM data efficiently.&lt;/P&gt;&lt;P&gt;Hope this helps anyone looking to integrate Dynamics 365 CDM with Delta Live Tables!&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 06 Jun 2024 06:22:30 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/delta-live-tables-with-common-data-model-as-source/m-p/71829#M34414</guid>
      <dc:creator>Suryanarayan</dc:creator>
      <dc:date>2024-06-06T06:22:30Z</dc:date>
    </item>
  </channel>
</rss>

