<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Delta Jira data import to Databricks in Administration &amp; Architecture</title>
    <link>https://community.databricks.com/t5/administration-architecture/delta-jira-data-import-to-databricks/m-p/154855#M5163</link>
    <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/200255"&gt;@greengil&lt;/a&gt;&amp;nbsp;&amp;nbsp;good question, I went through this something similar recently, so sharing what I found.&lt;/P&gt;&lt;P class=""&gt;My instinct was also to build it in Python, but once I dug in, the "just write a script" path hides a lot of pain:&lt;/P&gt;&lt;UL class=""&gt;&lt;LI&gt;&lt;STRONG&gt;Deletions are invisible.&lt;/STRONG&gt; Jira's REST API doesn't return deleted issues. Without webhooks, you'll have ghost records in Delta forever.&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Field history isn't free.&lt;/STRONG&gt; The API gives you current state, not change history. Reporting usually needs history, which means building and maintaining it yourself.&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Archived issues&lt;/STRONG&gt; aren't returned in JQL queries, only by ID.&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Rate limits, pagination, schema drift for custom fields,&lt;/STRONG&gt;&amp;nbsp;all real work.&lt;/LI&gt;&lt;/UL&gt;&lt;P class=""&gt;Fivetran's Jira connector handles all of this natively, JQL-based incremental sync, webhook-based deletion capture, auto-populated ISSUE_FIELD_HISTORY&amp;nbsp;tables, schema drift detection, MERGE into Delta, and it's available through Databricks Partner Connect for quick setup. There's also a free dbt package (fivetran/dbt_jira) with pre-built analytics models.&lt;/P&gt;&lt;P class=""&gt;&lt;STRONG&gt;My take:&lt;/STRONG&gt;&amp;nbsp;I would suggest go with Fivetran unless you have a specific reason not to - high volume cost concerns, need for archived issues, or data residency restrictions. Custom Python makes sense for narrow use cases, but it's weeks of build plus ongoing maintenance.&lt;/P&gt;&lt;P class=""&gt;References I did research and came up with solution, please take a look, I think you will find it really helpful:&lt;/P&gt;&lt;UL class=""&gt;&lt;LI&gt;Fivetran Jira:&amp;nbsp;&lt;A href="https://fivetran.com/docs/connectors/applications/jira" rel="noopener" target="_blank"&gt;Jira connector by Fivetran | Fivetran documentation&lt;/A&gt;&lt;/LI&gt;&lt;LI&gt;Fivetran Databricks destination:&amp;nbsp;&lt;A href="https://fivetran.com/docs/connectors/databases/databricks" rel="noopener" target="_blank"&gt;Databricks database connector by Fivetran | Fivetran documentation&lt;/A&gt;&lt;/LI&gt;&lt;LI&gt;dbt Jira package: &lt;A href="https://github.com/fivetran/dbt_jira" rel="noopener" target="_blank"&gt;fivetran/dbt_jira: Data models for Fivetran's Jira connector built using dbt.&lt;/A&gt;&lt;/LI&gt;&lt;/UL&gt;&lt;P class=""&gt;Happy to dig in further if you're leaning one way.&lt;/P&gt;</description>
    <pubDate>Sat, 18 Apr 2026 04:49:58 GMT</pubDate>
    <dc:creator>abhi_dabhi</dc:creator>
    <dc:date>2026-04-18T04:49:58Z</dc:date>
    <item>
      <title>Delta Jira data import to Databricks</title>
      <link>https://community.databricks.com/t5/administration-architecture/delta-jira-data-import-to-databricks/m-p/154394#M5137</link>
      <description>&lt;P&gt;We need to import large amount of Jira data into Databricks, and should import only the delta changes.&amp;nbsp; What's the best approach to do so?&amp;nbsp; Using the Fivetran Jira connector or develop our own Python scripts/pipeline code?&amp;nbsp; Thanks.&lt;/P&gt;</description>
      <pubDate>Mon, 13 Apr 2026 22:00:23 GMT</pubDate>
      <guid>https://community.databricks.com/t5/administration-architecture/delta-jira-data-import-to-databricks/m-p/154394#M5137</guid>
      <dc:creator>greengil</dc:creator>
      <dc:date>2026-04-13T22:00:23Z</dc:date>
    </item>
    <item>
      <title>Re: Delta Jira data import to Databricks</title>
      <link>https://community.databricks.com/t5/administration-architecture/delta-jira-data-import-to-databricks/m-p/154402#M5138</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/200255"&gt;@greengil&lt;/a&gt;,&lt;/P&gt;
&lt;P&gt;Have you considered Lakeflow Connect?&amp;nbsp; Databricks now has a native &lt;A href="https://docs.databricks.com/aws/en/ingestion/lakeflow-connect/jira" target="_blank"&gt;Jira connector&lt;/A&gt; in Lakeflow Connect that can achieve what you are looking for. It's in beta, but something you may want to consider.&amp;nbsp;&lt;/P&gt;
&lt;P&gt;It ingests Jira into Delta with incremental (delta) loads out of the box, supports SCD1/SCD2, handles deletes via audit logs, and runs fully managed on serverless with Unity Catalog governance.&amp;nbsp;This is lower-effort and better integrated than both Fivetran and custom Python, and directly targets your large volume + only changes requirement.&lt;/P&gt;
&lt;DIV data-known-size="675" data-item-index="1" data-index="1"&gt;
&lt;DIV&gt;
&lt;DIV data-workflow-run-id="bf0a5f50520e46a19ade1ec99cf2d6c7"&gt;
&lt;DIV&gt;
&lt;DIV tabindex="0" data-ui-element="chat-assistant-content-bubble"&gt;
&lt;DIV&gt;
&lt;P&gt;If you can’t use the Databricks Jira connector, prefer Fivetran Jira --&amp;gt; Databricks over custom code for a managed, low-maintenance ELT path.&amp;nbsp;Only build custom Python pipelines if you have very specific requirements that neither managed option can meet.&lt;/P&gt;
&lt;P class="p1"&gt;&lt;FONT size="2" color="#FF6600"&gt;&lt;STRONG&gt;&lt;I&gt;If this answer resolves your question, could you mark it as “Accept as Solution”? That helps other users quickly find the correct fix.&lt;/I&gt;&lt;/STRONG&gt;&lt;/FONT&gt;&lt;/P&gt;
&lt;/DIV&gt;
&lt;/DIV&gt;
&lt;/DIV&gt;
&lt;/DIV&gt;
&lt;/DIV&gt;
&lt;/DIV&gt;</description>
      <pubDate>Tue, 14 Apr 2026 05:59:58 GMT</pubDate>
      <guid>https://community.databricks.com/t5/administration-architecture/delta-jira-data-import-to-databricks/m-p/154402#M5138</guid>
      <dc:creator>Ashwin_DSA</dc:creator>
      <dc:date>2026-04-14T05:59:58Z</dc:date>
    </item>
    <item>
      <title>Re: Delta Jira data import to Databricks</title>
      <link>https://community.databricks.com/t5/administration-architecture/delta-jira-data-import-to-databricks/m-p/154838#M5161</link>
      <description>&lt;P&gt;Hi &lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/216690"&gt;@Ashwin_DSA&lt;/a&gt;&amp;nbsp; - Thank you for the information.&amp;nbsp; Appreciate it.&amp;nbsp; Regarding the built-in Lakeflow Connect, I see that it will inject all the Jira tables into Databricks.&amp;nbsp; Is there a way to inject only a subset of data?&amp;nbsp; For example, instead of all issues, I want only a subset.&amp;nbsp; Thanks.&lt;/P&gt;</description>
      <pubDate>Fri, 17 Apr 2026 18:59:59 GMT</pubDate>
      <guid>https://community.databricks.com/t5/administration-architecture/delta-jira-data-import-to-databricks/m-p/154838#M5161</guid>
      <dc:creator>greengil</dc:creator>
      <dc:date>2026-04-17T18:59:59Z</dc:date>
    </item>
    <item>
      <title>Re: Delta Jira data import to Databricks</title>
      <link>https://community.databricks.com/t5/administration-architecture/delta-jira-data-import-to-databricks/m-p/154843#M5162</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/200255"&gt;@greengil&lt;/a&gt;,&lt;/P&gt;
&lt;P&gt;Yes, you can restrict what Lakeflow Connect for Jira ingests, both by table and by rows (partially).&lt;/P&gt;
&lt;P&gt;In the UI, on the Source step, you can select only the tables you care about (for example, just issues, or issues + projects) instead of all source tables. In DABs/API, only list the tables you want under objects.&lt;/P&gt;
&lt;P&gt;The Jira connector supports filtering by Jira project/space via jira_options.include_jira_spaces (list of project keys).&amp;nbsp;In the UI, this is exposed as an option to filter the data by Jira spaces or projects (you enter project keys, not names or IDs).&lt;/P&gt;
&lt;P&gt;If you are looking for anything more granular than project/space (e.g. specific issue types, statuses, labels), then that's ot supported as of now. The connector ingests all matching issues for those projects/spaces, and you then filter downstream in silver/gold tables. More general row-level filtering for Jira is on the backlog but not yet available.&lt;/P&gt;
&lt;P&gt;Refer to these pages&amp;nbsp;&lt;A href="https://docs.databricks.com/aws/en/ingestion/lakeflow-connect/jira-pipeline" target="_blank"&gt;jira pipeline&lt;/A&gt;&amp;nbsp;and&amp;nbsp;&lt;A href="https://docs.databricks.com/aws/en/ingestion/lakeflow-connect/jira-limits" target="_blank"&gt;limitation&lt;/A&gt;.&amp;nbsp;&lt;/P&gt;
&lt;P class="p1"&gt;&lt;FONT size="2" color="#FF6600"&gt;&lt;STRONG&gt;&lt;I&gt;If this answer resolves your question, could you mark it as “Accept as Solution”? That helps other users quickly find the correct fix.&lt;/I&gt;&lt;/STRONG&gt;&lt;/FONT&gt;&lt;I&gt;&lt;/I&gt;&lt;/P&gt;</description>
      <pubDate>Fri, 17 Apr 2026 21:56:21 GMT</pubDate>
      <guid>https://community.databricks.com/t5/administration-architecture/delta-jira-data-import-to-databricks/m-p/154843#M5162</guid>
      <dc:creator>Ashwin_DSA</dc:creator>
      <dc:date>2026-04-17T21:56:21Z</dc:date>
    </item>
    <item>
      <title>Re: Delta Jira data import to Databricks</title>
      <link>https://community.databricks.com/t5/administration-architecture/delta-jira-data-import-to-databricks/m-p/154855#M5163</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/200255"&gt;@greengil&lt;/a&gt;&amp;nbsp;&amp;nbsp;good question, I went through this something similar recently, so sharing what I found.&lt;/P&gt;&lt;P class=""&gt;My instinct was also to build it in Python, but once I dug in, the "just write a script" path hides a lot of pain:&lt;/P&gt;&lt;UL class=""&gt;&lt;LI&gt;&lt;STRONG&gt;Deletions are invisible.&lt;/STRONG&gt; Jira's REST API doesn't return deleted issues. Without webhooks, you'll have ghost records in Delta forever.&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Field history isn't free.&lt;/STRONG&gt; The API gives you current state, not change history. Reporting usually needs history, which means building and maintaining it yourself.&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Archived issues&lt;/STRONG&gt; aren't returned in JQL queries, only by ID.&lt;/LI&gt;&lt;LI&gt;&lt;STRONG&gt;Rate limits, pagination, schema drift for custom fields,&lt;/STRONG&gt;&amp;nbsp;all real work.&lt;/LI&gt;&lt;/UL&gt;&lt;P class=""&gt;Fivetran's Jira connector handles all of this natively, JQL-based incremental sync, webhook-based deletion capture, auto-populated ISSUE_FIELD_HISTORY&amp;nbsp;tables, schema drift detection, MERGE into Delta, and it's available through Databricks Partner Connect for quick setup. There's also a free dbt package (fivetran/dbt_jira) with pre-built analytics models.&lt;/P&gt;&lt;P class=""&gt;&lt;STRONG&gt;My take:&lt;/STRONG&gt;&amp;nbsp;I would suggest go with Fivetran unless you have a specific reason not to - high volume cost concerns, need for archived issues, or data residency restrictions. Custom Python makes sense for narrow use cases, but it's weeks of build plus ongoing maintenance.&lt;/P&gt;&lt;P class=""&gt;References I did research and came up with solution, please take a look, I think you will find it really helpful:&lt;/P&gt;&lt;UL class=""&gt;&lt;LI&gt;Fivetran Jira:&amp;nbsp;&lt;A href="https://fivetran.com/docs/connectors/applications/jira" rel="noopener" target="_blank"&gt;Jira connector by Fivetran | Fivetran documentation&lt;/A&gt;&lt;/LI&gt;&lt;LI&gt;Fivetran Databricks destination:&amp;nbsp;&lt;A href="https://fivetran.com/docs/connectors/databases/databricks" rel="noopener" target="_blank"&gt;Databricks database connector by Fivetran | Fivetran documentation&lt;/A&gt;&lt;/LI&gt;&lt;LI&gt;dbt Jira package: &lt;A href="https://github.com/fivetran/dbt_jira" rel="noopener" target="_blank"&gt;fivetran/dbt_jira: Data models for Fivetran's Jira connector built using dbt.&lt;/A&gt;&lt;/LI&gt;&lt;/UL&gt;&lt;P class=""&gt;Happy to dig in further if you're leaning one way.&lt;/P&gt;</description>
      <pubDate>Sat, 18 Apr 2026 04:49:58 GMT</pubDate>
      <guid>https://community.databricks.com/t5/administration-architecture/delta-jira-data-import-to-databricks/m-p/154855#M5163</guid>
      <dc:creator>abhi_dabhi</dc:creator>
      <dc:date>2026-04-18T04:49:58Z</dc:date>
    </item>
  </channel>
</rss>

