<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: On-Premise SQL Server Ingestion to Databricks Bronze Layer in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/on-premise-sql-server-ingestion-to-databricks-bronze-layer/m-p/37225#M26291</link>
    <description>&lt;P data-unlink="true"&gt;Sure, first error which popped up was (code from notebook below in screenshot):&lt;BR /&gt;&lt;BR /&gt;AnalysisException: Incompatible format detected. A transaction log for Delta was found at `&lt;A href="https://teststorage.blob.core.windows.net/testtest/dbo.Animal.parquet/_delta_log" target="_blank" rel="noopener"&gt;https://teststorage.blob.core.windows.net/testtest/dbo.Animal.parquet/_delta_log&lt;/A&gt;`, but you are trying to read from `&lt;A href="https://teststorage.blob.core.windows.net/testtest/dbo.Animal.parquet" target="_blank" rel="noopener"&gt;https://teststorage.blob.core.windows.net/testtest/dbo.Animal.parquet&lt;/A&gt;` using format("parquet"). You must use 'format("delta")' when reading and writing to a delta table. To disable this check, SET spark.databricks.delta.formatCheck.enabled=false To learn more about Delta, see https://docs.microsoft.com/azure/databricks/delta/index&amp;nbsp;&lt;BR /&gt;&lt;BR /&gt;I tried to fix it like "&lt;SPAN&gt;delta_df = spark.read.&lt;/SPAN&gt;&lt;SPAN class=""&gt;format&lt;/SPAN&gt;&lt;SPAN&gt;(&lt;/SPAN&gt;&lt;SPAN class=""&gt;"parquet"&lt;/SPAN&gt;&lt;SPAN&gt;) --&amp;gt;&amp;nbsp;delta_df = spark.read.&lt;SPAN class=""&gt;format&lt;/SPAN&gt;(&lt;SPAN class=""&gt;"delta"&lt;/SPAN&gt;)"&lt;BR /&gt;&lt;BR /&gt;Its dropped:&lt;BR /&gt;&lt;BR /&gt;File /databricks/spark/python/pyspark/instrumentation_utils.py:48, in _wrap_function.&amp;lt;locals&amp;gt;.wrapper(*args, **kwargs) 46 start = time.perf_counter() 47 try: ---&amp;gt; 48 res = func(*args, **kwargs) 49 logger.log_success( 50 module_name, class_name, function_name, time.perf_counter() - start, signature 51 )&lt;BR /&gt;&lt;BR /&gt;But I am not sure I do it in right way. Trying to finish my project and I made it through copy data pipeline, but I want to change into autoloader and storage data in delta table.&lt;BR /&gt;&lt;BR /&gt;Thanks in advance.&lt;/SPAN&gt;&lt;/P&gt;</description>
    <pubDate>Sat, 08 Jul 2023 15:25:48 GMT</pubDate>
    <dc:creator>frantisek</dc:creator>
    <dc:date>2023-07-08T15:25:48Z</dc:date>
    <item>
      <title>On-Premise SQL Server Ingestion to Databricks Bronze Layer</title>
      <link>https://community.databricks.com/t5/data-engineering/on-premise-sql-server-ingestion-to-databricks-bronze-layer/m-p/3864#M762</link>
      <description>&lt;P&gt;Hello everyone!&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;So I want to ingest tables with schemas from the on-premise SQL server to Databricks Bronze layer with Delta Live Table and I want to do it using Azure Data Factory and I want the load to be a Snapshot batch load, not an incremental load. What are the activities I will have to use in ADF? &lt;/P&gt;</description>
      <pubDate>Tue, 30 May 2023 19:18:46 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/on-premise-sql-server-ingestion-to-databricks-bronze-layer/m-p/3864#M762</guid>
      <dc:creator>Enzo_Bahrami</dc:creator>
      <dc:date>2023-05-30T19:18:46Z</dc:date>
    </item>
    <item>
      <title>Re: On-Premise SQL Server Ingestion to Databricks Bronze Layer</title>
      <link>https://community.databricks.com/t5/data-engineering/on-premise-sql-server-ingestion-to-databricks-bronze-layer/m-p/3865#M763</link>
      <description>&lt;P&gt;@Parsa Bahraminejad​&lt;/P&gt;&lt;P&gt;You'll need to use ADF Copy Activity to fetch the data from SQL Server to ADLS (Storage) in parquet format. Then you can simply ingest the data from ADLS (Raw Layer) to bronze using autoloader or spark.read.format("parquet").&lt;/P&gt;</description>
      <pubDate>Wed, 31 May 2023 04:35:51 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/on-premise-sql-server-ingestion-to-databricks-bronze-layer/m-p/3865#M763</guid>
      <dc:creator>daniel_sahal</dc:creator>
      <dc:date>2023-05-31T04:35:51Z</dc:date>
    </item>
    <item>
      <title>Re: On-Premise SQL Server Ingestion to Databricks Bronze Layer</title>
      <link>https://community.databricks.com/t5/data-engineering/on-premise-sql-server-ingestion-to-databricks-bronze-layer/m-p/3866#M764</link>
      <description>&lt;P&gt;Hi @Parsa Bahraminejad​&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Thank you for posting your question in our community! We are happy to assist you.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;To help us provide you with the most accurate information, could you please take a moment to review the responses and select the one that best answers your question?&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;This will also help other community members who may have similar questions in the future. Thank you for your participation and let us know if you need any further assistance!&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;</description>
      <pubDate>Thu, 01 Jun 2023 03:18:18 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/on-premise-sql-server-ingestion-to-databricks-bronze-layer/m-p/3866#M764</guid>
      <dc:creator>Anonymous</dc:creator>
      <dc:date>2023-06-01T03:18:18Z</dc:date>
    </item>
    <item>
      <title>Re: On-Premise SQL Server Ingestion to Databricks Bronze Layer</title>
      <link>https://community.databricks.com/t5/data-engineering/on-premise-sql-server-ingestion-to-databricks-bronze-layer/m-p/3867#M765</link>
      <description>&lt;P&gt;Thank you!&lt;/P&gt;</description>
      <pubDate>Thu, 01 Jun 2023 03:30:24 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/on-premise-sql-server-ingestion-to-databricks-bronze-layer/m-p/3867#M765</guid>
      <dc:creator>Enzo_Bahrami</dc:creator>
      <dc:date>2023-06-01T03:30:24Z</dc:date>
    </item>
    <item>
      <title>Re: On-Premise SQL Server Ingestion to Databricks Bronze Layer</title>
      <link>https://community.databricks.com/t5/data-engineering/on-premise-sql-server-ingestion-to-databricks-bronze-layer/m-p/37142#M26275</link>
      <description>&lt;P&gt;Hi, could you give me specific script to ingest the data from ADLS (.parquet) to delta table using autoloader? I am not able to do that. I tried everything I could, but I get error all the time in my notebook when I am trying to set up the script (autoloader).&lt;/P&gt;</description>
      <pubDate>Fri, 07 Jul 2023 08:06:52 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/on-premise-sql-server-ingestion-to-databricks-bronze-layer/m-p/37142#M26275</guid>
      <dc:creator>frantisek</dc:creator>
      <dc:date>2023-07-07T08:06:52Z</dc:date>
    </item>
    <item>
      <title>Re: On-Premise SQL Server Ingestion to Databricks Bronze Layer</title>
      <link>https://community.databricks.com/t5/data-engineering/on-premise-sql-server-ingestion-to-databricks-bronze-layer/m-p/37194#M26285</link>
      <description>&lt;P&gt;can you please share the error you are facing?&amp;nbsp;&lt;/P&gt;&lt;P&gt;Example scripts&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;A href="https://learn.microsoft.com/en-us/azure/databricks/getting-started/etl-quick-start" target="_blank"&gt;https://learn.microsoft.com/en-us/azure/databricks/getting-started/etl-quick-start&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&lt;A href="https://learn.microsoft.com/en-us/azure/databricks/ingestion/#automate-etl-with-delta-live-tables-and-auto-loader" target="_blank"&gt;https://learn.microsoft.com/en-us/azure/databricks/ingestion/#automate-etl-with-delta-live-tables-and-auto-loader&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Fri, 07 Jul 2023 21:55:50 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/on-premise-sql-server-ingestion-to-databricks-bronze-layer/m-p/37194#M26285</guid>
      <dc:creator>shan_chandra</dc:creator>
      <dc:date>2023-07-07T21:55:50Z</dc:date>
    </item>
    <item>
      <title>Re: On-Premise SQL Server Ingestion to Databricks Bronze Layer</title>
      <link>https://community.databricks.com/t5/data-engineering/on-premise-sql-server-ingestion-to-databricks-bronze-layer/m-p/37225#M26291</link>
      <description>&lt;P data-unlink="true"&gt;Sure, first error which popped up was (code from notebook below in screenshot):&lt;BR /&gt;&lt;BR /&gt;AnalysisException: Incompatible format detected. A transaction log for Delta was found at `&lt;A href="https://teststorage.blob.core.windows.net/testtest/dbo.Animal.parquet/_delta_log" target="_blank" rel="noopener"&gt;https://teststorage.blob.core.windows.net/testtest/dbo.Animal.parquet/_delta_log&lt;/A&gt;`, but you are trying to read from `&lt;A href="https://teststorage.blob.core.windows.net/testtest/dbo.Animal.parquet" target="_blank" rel="noopener"&gt;https://teststorage.blob.core.windows.net/testtest/dbo.Animal.parquet&lt;/A&gt;` using format("parquet"). You must use 'format("delta")' when reading and writing to a delta table. To disable this check, SET spark.databricks.delta.formatCheck.enabled=false To learn more about Delta, see https://docs.microsoft.com/azure/databricks/delta/index&amp;nbsp;&lt;BR /&gt;&lt;BR /&gt;I tried to fix it like "&lt;SPAN&gt;delta_df = spark.read.&lt;/SPAN&gt;&lt;SPAN class=""&gt;format&lt;/SPAN&gt;&lt;SPAN&gt;(&lt;/SPAN&gt;&lt;SPAN class=""&gt;"parquet"&lt;/SPAN&gt;&lt;SPAN&gt;) --&amp;gt;&amp;nbsp;delta_df = spark.read.&lt;SPAN class=""&gt;format&lt;/SPAN&gt;(&lt;SPAN class=""&gt;"delta"&lt;/SPAN&gt;)"&lt;BR /&gt;&lt;BR /&gt;Its dropped:&lt;BR /&gt;&lt;BR /&gt;File /databricks/spark/python/pyspark/instrumentation_utils.py:48, in _wrap_function.&amp;lt;locals&amp;gt;.wrapper(*args, **kwargs) 46 start = time.perf_counter() 47 try: ---&amp;gt; 48 res = func(*args, **kwargs) 49 logger.log_success( 50 module_name, class_name, function_name, time.perf_counter() - start, signature 51 )&lt;BR /&gt;&lt;BR /&gt;But I am not sure I do it in right way. Trying to finish my project and I made it through copy data pipeline, but I want to change into autoloader and storage data in delta table.&lt;BR /&gt;&lt;BR /&gt;Thanks in advance.&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Sat, 08 Jul 2023 15:25:48 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/on-premise-sql-server-ingestion-to-databricks-bronze-layer/m-p/37225#M26291</guid>
      <dc:creator>frantisek</dc:creator>
      <dc:date>2023-07-08T15:25:48Z</dc:date>
    </item>
  </channel>
</rss>

