<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Insert data to a CDF-enabled Delta table throwing java.lang.StackOverflowError in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/insert-data-to-a-cdf-enabled-delta-table-throwing-java-lang/m-p/4841#M1437</link>
    <description>&lt;P&gt;Its a run time error , check if there are any recursive calls , if everything is okay run it freshly again &lt;/P&gt;</description>
    <pubDate>Thu, 04 May 2023 03:40:10 GMT</pubDate>
    <dc:creator>Priyag1</dc:creator>
    <dc:date>2023-05-04T03:40:10Z</dc:date>
    <item>
      <title>Insert data to a CDF-enabled Delta table throwing java.lang.StackOverflowError</title>
      <link>https://community.databricks.com/t5/data-engineering/insert-data-to-a-cdf-enabled-delta-table-throwing-java-lang/m-p/4840#M1436</link>
      <description>&lt;P&gt;I am building a bronze table with CDF-enables in these steps:&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;Initially, Reading json file from landing zone and write to table location&lt;/LI&gt;&lt;/UL&gt;&lt;PRE&gt;&lt;CODE&gt;df = spark.readStream.format("cloudFiles") \
  .option("cloudFiles.schemaLocation", &amp;lt;schema_loc&amp;gt;) \
  .option("cloudFiles.format", "json") \
  .option("cloudFiles.inferColumnTypes", "true") \
  .option("cloudFiles.schemaEvolutionMode", "addNewColumns") \
  .option("cloudFiles.includeExistingFiles", "true") \
  .load(&amp;lt;landing_loc_info&amp;gt;)
df.writeStream \
      .format("delta") \
      .trigger(once = True) \
      .option("mergeSchema", "true") \
      .option("checkpointLocation", &amp;lt;bronzechk_loc) \
      .start(&amp;lt;bronzetbl_loc&amp;gt;)&lt;/CODE&gt;&lt;/PRE&gt;&lt;UL&gt;&lt;LI&gt;Create a delta table and enable CDF&lt;/LI&gt;&lt;/UL&gt;&lt;PRE&gt;&lt;CODE&gt;CREATE TABLE bronze.mytable
USING DELTA
LOCATION '&amp;lt;file location&amp;gt;';
&amp;nbsp;
ALTER TABLE bronze.mytable
SET TBLPROPERTIES (delta.enableChangeDataFeed = true);&lt;/CODE&gt;&lt;/PRE&gt;&lt;UL&gt;&lt;LI&gt;Read more data with exactly the same structure from landing zone and insert into bronze table&lt;/LI&gt;&lt;/UL&gt;&lt;PRE&gt;&lt;CODE&gt;df = spark.readStream.format("cloudFiles") \
  .option("cloudFiles.schemaLocation", schema_loc) \
  .option("cloudFiles.format", "json") \
  .option("cloudFiles.inferColumnTypes", "true") \
  .option("cloudFiles.schemaEvolutionMode", "addNewColumns") \
  .option("cloudFiles.includeExistingFiles", "true") \
  .load(&amp;lt;landing_loc&amp;gt;)
&amp;nbsp;
df.createOrReplaceTempView("bronze_company_info_dataset")
sql_query = "INSERT INTO bronze.mytable TABLE bronze_dataset"
spark.sql(sql_query)&lt;/CODE&gt;&lt;/PRE&gt;&lt;P&gt;&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;It throws java.lang.StackOverflowError when executing the sql_query:&lt;/LI&gt;&lt;/UL&gt;&lt;PRE&gt;&lt;CODE&gt;/databricks/spark/python/pyspark/instrumentation_utils.py in wrapper(*args, **kwargs)
     46             start = time.perf_counter()
     47             try:
---&amp;gt; 48                 res = func(*args, **kwargs)
     49                 logger.log_success(
     50                     module_name, class_name, function_name, time.perf_counter() - start, signature
&amp;nbsp;
/databricks/spark/python/pyspark/sql/session.py in sql(self, sqlQuery, **kwargs)
   1117             sqlQuery = formatter.format(sqlQuery, **kwargs)
   1118         try:
-&amp;gt; 1119             return DataFrame(self._jsparkSession.sql(sqlQuery), self)
   1120         finally:
   1121             if len(kwargs) &amp;gt; 0:
&amp;nbsp;
/databricks/spark/python/lib/py4j-0.10.9.5-src.zip/py4j/java_gateway.py in __call__(self, *args)
   1319 
   1320         answer = self.gateway_client.send_command(command)
-&amp;gt; 1321         return_value = get_return_value(
   1322             answer, self.gateway_client, self.target_id, self.name)
   &lt;/CODE&gt;&lt;/PRE&gt;&lt;UL&gt;&lt;LI&gt;I am using Community Edition, with runtime version 11.3 LTS (includes Apache Spark 3.3.0, Scala 2.12)&lt;/LI&gt;&lt;/UL&gt;</description>
      <pubDate>Wed, 03 May 2023 22:06:44 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/insert-data-to-a-cdf-enabled-delta-table-throwing-java-lang/m-p/4840#M1436</guid>
      <dc:creator>Johny</dc:creator>
      <dc:date>2023-05-03T22:06:44Z</dc:date>
    </item>
    <item>
      <title>Re: Insert data to a CDF-enabled Delta table throwing java.lang.StackOverflowError</title>
      <link>https://community.databricks.com/t5/data-engineering/insert-data-to-a-cdf-enabled-delta-table-throwing-java-lang/m-p/4841#M1437</link>
      <description>&lt;P&gt;Its a run time error , check if there are any recursive calls , if everything is okay run it freshly again &lt;/P&gt;</description>
      <pubDate>Thu, 04 May 2023 03:40:10 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/insert-data-to-a-cdf-enabled-delta-table-throwing-java-lang/m-p/4841#M1437</guid>
      <dc:creator>Priyag1</dc:creator>
      <dc:date>2023-05-04T03:40:10Z</dc:date>
    </item>
    <item>
      <title>Re: Insert data to a CDF-enabled Delta table throwing java.lang.StackOverflowError</title>
      <link>https://community.databricks.com/t5/data-engineering/insert-data-to-a-cdf-enabled-delta-table-throwing-java-lang/m-p/4842#M1438</link>
      <description>&lt;P&gt;I tried with a simple csv file that only has one column. I got the same error.&lt;/P&gt;</description>
      <pubDate>Thu, 04 May 2023 22:02:26 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/insert-data-to-a-cdf-enabled-delta-table-throwing-java-lang/m-p/4842#M1438</guid>
      <dc:creator>Johny</dc:creator>
      <dc:date>2023-05-04T22:02:26Z</dc:date>
    </item>
  </channel>
</rss>

