<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic How to output data from Databricks? in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/how-to-output-data-from-databricks/m-p/63900#M32392</link>
    <description>&lt;P&gt;Hello,&lt;/P&gt;&lt;P&gt;I am just starting with Databricks in Azure and I need to output the data to an Affinity CRM system.&lt;/P&gt;&lt;P&gt;Affinity has an API and I am wondering is there any sort of automated / data pipeline sort of way to tell databricks to just pump the data into Affinity or do I have to write some sort of Python script and make it run in Databricks?&lt;/P&gt;&lt;P&gt;Alternatively I could create a microservice that gets the data from Databricks and sends it to Affinity but that seems like the worst options to me.&lt;/P&gt;&lt;P&gt;I just don't know how the entire thing works.&lt;/P&gt;</description>
    <pubDate>Sat, 16 Mar 2024 13:23:42 GMT</pubDate>
    <dc:creator>Abdul1</dc:creator>
    <dc:date>2024-03-16T13:23:42Z</dc:date>
    <item>
      <title>How to output data from Databricks?</title>
      <link>https://community.databricks.com/t5/data-engineering/how-to-output-data-from-databricks/m-p/63900#M32392</link>
      <description>&lt;P&gt;Hello,&lt;/P&gt;&lt;P&gt;I am just starting with Databricks in Azure and I need to output the data to an Affinity CRM system.&lt;/P&gt;&lt;P&gt;Affinity has an API and I am wondering is there any sort of automated / data pipeline sort of way to tell databricks to just pump the data into Affinity or do I have to write some sort of Python script and make it run in Databricks?&lt;/P&gt;&lt;P&gt;Alternatively I could create a microservice that gets the data from Databricks and sends it to Affinity but that seems like the worst options to me.&lt;/P&gt;&lt;P&gt;I just don't know how the entire thing works.&lt;/P&gt;</description>
      <pubDate>Sat, 16 Mar 2024 13:23:42 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/how-to-output-data-from-databricks/m-p/63900#M32392</guid>
      <dc:creator>Abdul1</dc:creator>
      <dc:date>2024-03-16T13:23:42Z</dc:date>
    </item>
    <item>
      <title>Re: How to output data from Databricks?</title>
      <link>https://community.databricks.com/t5/data-engineering/how-to-output-data-from-databricks/m-p/63912#M32401</link>
      <description>&lt;P&gt;We need more info on what kind of data, volume and what the called APi can handle. Calling an API for single records in parallel can be achieved using UDF(see &lt;A href="https://community.databricks.com/t5/data-engineering/performance-issue-with-pyspark-udf-function-calling-rest-api/td-p/6890" target="_self"&gt;THIS&lt;/A&gt;). You need to be careful to batch the records so that the target API can handle the parallel load. If you want to send an entire file via API (assuming file size is within the API limits), you can use Synapse pipeline activity(assuming you are using Azure). Databricks does not have any inbuilt feature for this as far as I know.&lt;/P&gt;</description>
      <pubDate>Sun, 17 Mar 2024 01:06:43 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/how-to-output-data-from-databricks/m-p/63912#M32401</guid>
      <dc:creator>Edthehead</dc:creator>
      <dc:date>2024-03-17T01:06:43Z</dc:date>
    </item>
  </channel>
</rss>

