<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Integration patterns with Oracle database in Administration &amp; Architecture</title>
    <link>https://community.databricks.com/t5/administration-architecture/integration-patterns-with-oracle-database/m-p/155179#M5176</link>
    <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/225622"&gt;@HideK253&lt;/a&gt;,&lt;/P&gt;
&lt;P&gt;Your list is essentially correct. The main Databricks-native patterns are export to files + COPY INTO / Auto Loader, Lakehouse Federation and&amp;nbsp;Lakeflow Connect. A couple of other options are listed below.&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;Custom Spark JDBC jobs: spark.read.format("jdbc") from Oracle into Delta, often with your own watermark column for incremental loads.&lt;/LI&gt;
&lt;LI&gt;You can also pair third-party CDC tools (Oracle GoldenGate, Fivetran/HVR, Debezium+Kafka, etc.) with Databricks to land CDC data into cloud storage and then process it into Delta.&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;In terms of your query about&amp;nbsp;Lakehouse Federation and JDBC-based query connectors, the answer is yes for both.&amp;nbsp;Lakehouse Federation pushes queries down via JDBC to the foreign database, including Oracle.&amp;nbsp;Lakeflow Connect query-based connectors use Unity Catalog connections and Lakehouse Federation under the hood. The Oracle connector is part of this JDBC-based family.&amp;nbsp;You don’t manage the JDBC details directly. Databricks does.&lt;/P&gt;
&lt;P&gt;You’re correct. Oracle is not yet in the public GA list of Lakeflow CDC database connectors (those call out MySQL/PostgreSQL/SQL Server).. So, the best options as of today are...&lt;/P&gt;
&lt;P&gt;Lakeflow Connect query-based connector for Oracle that provides incremental loads via a cursor column (and supports SCD1/SCD2 and soft/hard deletes), but it’s not log-based CDC... changes are detected by querying the source, not by reading redo logs.&lt;/P&gt;
&lt;P&gt;(Or)&lt;/P&gt;
&lt;P&gt;You can&amp;nbsp;use Oracle CDC tools (GoldenGate/Fivetran/Debezium/etc) to capture redo logs and land a CDC feed in object storage or Kafka, then use Lakeflow Spark Declarative Pipelines with AUTO CDC / AUTO CDC FROM SNAPSHOT to maintain SCD1/SCD2 Delta tables.&lt;/P&gt;
&lt;P&gt;For true CDC you’d use an external CDC tool plus AUTO CDC, and for simpler incremental loads you can use the Oracle query-based connector.&lt;/P&gt;
&lt;P&gt;Hope this helps.&lt;/P&gt;
&lt;P class="p1"&gt;&lt;FONT size="2" color="#FF6600"&gt;&lt;STRONG&gt;&lt;I&gt;If this answer resolves your question, could you mark it as “Accept as Solution”? That helps other users quickly find the correct fix.&lt;/I&gt;&lt;/STRONG&gt;&lt;/FONT&gt;&lt;I&gt;&lt;/I&gt;&lt;/P&gt;</description>
    <pubDate>Wed, 22 Apr 2026 09:08:20 GMT</pubDate>
    <dc:creator>Ashwin_DSA</dc:creator>
    <dc:date>2026-04-22T09:08:20Z</dc:date>
    <item>
      <title>Integration patterns with Oracle database</title>
      <link>https://community.databricks.com/t5/administration-architecture/integration-patterns-with-oracle-database/m-p/155003#M5168</link>
      <description>&lt;P&gt;Hello everyone&lt;/P&gt;&lt;P&gt;I'd like to try integration of Data Bricks with on-premise Oracle database, and now I have the following&amp;nbsp; fundamental questions as below about the architectures of that integration.&lt;/P&gt;&lt;P&gt;1) I guess the following is the list of integration way with Oracle database, but is there any other options?&lt;/P&gt;&lt;P&gt;- Convert table data to csv/parquet via ETL tool, and ingest csv/parquet via COPY INTO/Auto Loader&lt;/P&gt;&lt;P&gt;- Using Lakehouse Federation to query table data&lt;/P&gt;&lt;P&gt;- Using Query based connectors for LakeFlow to ingest table data&lt;/P&gt;&lt;P&gt;2) Does Lakehouse Federation and Query based connectors for Oracle use JDBC connections?&lt;/P&gt;&lt;P&gt;3) Is there any ways to implement CDC for Oracle? I guess Oracle is not listed in the supported database for CDC LakeFlow.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thank you&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Tue, 21 Apr 2026 04:50:28 GMT</pubDate>
      <guid>https://community.databricks.com/t5/administration-architecture/integration-patterns-with-oracle-database/m-p/155003#M5168</guid>
      <dc:creator>HideK253</dc:creator>
      <dc:date>2026-04-21T04:50:28Z</dc:date>
    </item>
    <item>
      <title>Re: Integration patterns with Oracle database</title>
      <link>https://community.databricks.com/t5/administration-architecture/integration-patterns-with-oracle-database/m-p/155179#M5176</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/225622"&gt;@HideK253&lt;/a&gt;,&lt;/P&gt;
&lt;P&gt;Your list is essentially correct. The main Databricks-native patterns are export to files + COPY INTO / Auto Loader, Lakehouse Federation and&amp;nbsp;Lakeflow Connect. A couple of other options are listed below.&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;Custom Spark JDBC jobs: spark.read.format("jdbc") from Oracle into Delta, often with your own watermark column for incremental loads.&lt;/LI&gt;
&lt;LI&gt;You can also pair third-party CDC tools (Oracle GoldenGate, Fivetran/HVR, Debezium+Kafka, etc.) with Databricks to land CDC data into cloud storage and then process it into Delta.&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;In terms of your query about&amp;nbsp;Lakehouse Federation and JDBC-based query connectors, the answer is yes for both.&amp;nbsp;Lakehouse Federation pushes queries down via JDBC to the foreign database, including Oracle.&amp;nbsp;Lakeflow Connect query-based connectors use Unity Catalog connections and Lakehouse Federation under the hood. The Oracle connector is part of this JDBC-based family.&amp;nbsp;You don’t manage the JDBC details directly. Databricks does.&lt;/P&gt;
&lt;P&gt;You’re correct. Oracle is not yet in the public GA list of Lakeflow CDC database connectors (those call out MySQL/PostgreSQL/SQL Server).. So, the best options as of today are...&lt;/P&gt;
&lt;P&gt;Lakeflow Connect query-based connector for Oracle that provides incremental loads via a cursor column (and supports SCD1/SCD2 and soft/hard deletes), but it’s not log-based CDC... changes are detected by querying the source, not by reading redo logs.&lt;/P&gt;
&lt;P&gt;(Or)&lt;/P&gt;
&lt;P&gt;You can&amp;nbsp;use Oracle CDC tools (GoldenGate/Fivetran/Debezium/etc) to capture redo logs and land a CDC feed in object storage or Kafka, then use Lakeflow Spark Declarative Pipelines with AUTO CDC / AUTO CDC FROM SNAPSHOT to maintain SCD1/SCD2 Delta tables.&lt;/P&gt;
&lt;P&gt;For true CDC you’d use an external CDC tool plus AUTO CDC, and for simpler incremental loads you can use the Oracle query-based connector.&lt;/P&gt;
&lt;P&gt;Hope this helps.&lt;/P&gt;
&lt;P class="p1"&gt;&lt;FONT size="2" color="#FF6600"&gt;&lt;STRONG&gt;&lt;I&gt;If this answer resolves your question, could you mark it as “Accept as Solution”? That helps other users quickly find the correct fix.&lt;/I&gt;&lt;/STRONG&gt;&lt;/FONT&gt;&lt;I&gt;&lt;/I&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 22 Apr 2026 09:08:20 GMT</pubDate>
      <guid>https://community.databricks.com/t5/administration-architecture/integration-patterns-with-oracle-database/m-p/155179#M5176</guid>
      <dc:creator>Ashwin_DSA</dc:creator>
      <dc:date>2026-04-22T09:08:20Z</dc:date>
    </item>
  </channel>
</rss>

