<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Lakeflow Connect Data ingestion from SQL Server and PostgreSQL to Databricks with CDC in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/lakeflow-connect-data-ingestion-from-sql-server-and-postgresql/m-p/156635#M54458</link>
    <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/149095"&gt;@shan-databricks&lt;/a&gt;&amp;nbsp;&lt;BR /&gt;&lt;BR /&gt;You should &lt;A href="https://docs.databricks.com/aws/en/ingestion/lakeflow-connect/postgresql-source-setup" target="_self"&gt;setup postgresql&lt;/A&gt; for ingestion via Lakeflow connect. Once your Postgres logical replication is ready you have to create &lt;A href="https://docs.databricks.com/aws/en/ingestion/lakeflow-connect/postgresql-pipeline" target="_self"&gt;ingestion pipelines&lt;/A&gt; which comprise a gateway and ingestion pipeline. Your gateway pipeline is continuous pipeline to pull&amp;nbsp;&lt;SPAN&gt;changed data from the source Postgres database and stores&amp;nbsp;it into you staging catalog in Databricks. Ingestion gateway pipeline should use a compute which resides in your Databricks VPC and that VPC should be whitelisted in your firewall. Second pipeline use serverless compute to move changed data from stage catalog to target catalog's bronze schema.&lt;BR /&gt;&lt;BR /&gt;Thanks&lt;/SPAN&gt;&lt;/P&gt;</description>
    <pubDate>Tue, 12 May 2026 06:43:28 GMT</pubDate>
    <dc:creator>ziafazal</dc:creator>
    <dc:date>2026-05-12T06:43:28Z</dc:date>
    <item>
      <title>Lakeflow Connect Data ingestion from SQL Server and PostgreSQL to Databricks with CDC</title>
      <link>https://community.databricks.com/t5/data-engineering/lakeflow-connect-data-ingestion-from-sql-server-and-postgresql/m-p/156628#M54457</link>
      <description>&lt;DIV&gt;We have a requirement to use Lakeflow Connect for data ingestion from SQL Server and PostgreSQL into Databricks with CDC and&amp;nbsp;&lt;SPAN&gt;Lakehouse federation.&amp;nbsp;&lt;/SPAN&gt;I would like to understand the pros and cons of Lakeflow Connect in the following areas&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;Firewall/gateway considerations&lt;/DIV&gt;&lt;DIV&gt;CDC capabilities&lt;/DIV&gt;&lt;DIV&gt;Reliability&lt;/DIV&gt;&lt;DIV&gt;Overall success of Lakeflow implementation&lt;/DIV&gt;&lt;DIV&gt;Overall success of&amp;nbsp;&lt;SPAN&gt;Lakehouse federation&lt;/SPAN&gt;&lt;/DIV&gt;</description>
      <pubDate>Tue, 12 May 2026 04:43:34 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/lakeflow-connect-data-ingestion-from-sql-server-and-postgresql/m-p/156628#M54457</guid>
      <dc:creator>shan-databricks</dc:creator>
      <dc:date>2026-05-12T04:43:34Z</dc:date>
    </item>
    <item>
      <title>Re: Lakeflow Connect Data ingestion from SQL Server and PostgreSQL to Databricks with CDC</title>
      <link>https://community.databricks.com/t5/data-engineering/lakeflow-connect-data-ingestion-from-sql-server-and-postgresql/m-p/156635#M54458</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/149095"&gt;@shan-databricks&lt;/a&gt;&amp;nbsp;&lt;BR /&gt;&lt;BR /&gt;You should &lt;A href="https://docs.databricks.com/aws/en/ingestion/lakeflow-connect/postgresql-source-setup" target="_self"&gt;setup postgresql&lt;/A&gt; for ingestion via Lakeflow connect. Once your Postgres logical replication is ready you have to create &lt;A href="https://docs.databricks.com/aws/en/ingestion/lakeflow-connect/postgresql-pipeline" target="_self"&gt;ingestion pipelines&lt;/A&gt; which comprise a gateway and ingestion pipeline. Your gateway pipeline is continuous pipeline to pull&amp;nbsp;&lt;SPAN&gt;changed data from the source Postgres database and stores&amp;nbsp;it into you staging catalog in Databricks. Ingestion gateway pipeline should use a compute which resides in your Databricks VPC and that VPC should be whitelisted in your firewall. Second pipeline use serverless compute to move changed data from stage catalog to target catalog's bronze schema.&lt;BR /&gt;&lt;BR /&gt;Thanks&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Tue, 12 May 2026 06:43:28 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/lakeflow-connect-data-ingestion-from-sql-server-and-postgresql/m-p/156635#M54458</guid>
      <dc:creator>ziafazal</dc:creator>
      <dc:date>2026-05-12T06:43:28Z</dc:date>
    </item>
  </channel>
</rss>

