<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Dataframe loses its contents after the write operation to Database. in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/dataframe-loses-its-contents-after-the-write-operation-to/m-p/31727#M23107</link>
    <description>&lt;P&gt;We had working code as below.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;print(f"{file_name}Before insert count", datetime.datetime.now(), scan_df_new.count())&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;print(scan_df_new.show())&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;scan_20220908120005_10Before insert count 2022-09-14 11:37:15.853588 3&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;+-------------------+----------+-------------------+--------------------+----------+&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;|&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;tran_id|t_store_id|&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;scan_datetime|&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;customer_id|updated_by|&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;+-------------------+----------+-------------------+--------------------+----------+&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;|1230000000000000004|&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;4395|2022-09-08 03:00:01|20220816a51cee4264f1|Databricks|&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;|1230000000000000005|&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;4394|2022-09-08 02:58:00|20220816a51cee4264f1|Databricks|&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;|1230000000000000006|&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;4393|2022-09-08 03:00:04|20220816a51cee4264f1|Databricks|&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;+-------------------+----------+-------------------+--------------------+----------+&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;The data frame after the write operation is used for further business logic processing. This was working earlier. But recently we are observing a strange behavior, where in the data in the data frame is getting lost . When wee check the contents, or even the dataframe count its shows empty. &lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;scan_df_new.write.format("jdbc").option("url", jdbcUrl).option("dbtable", scan_table).mode("append").save()&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;print(f"{file_name}After insert count", datetime.datetime.now(), scan_df_new.count())&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;print(scan_df_new.show())&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;None&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;scan_20220908120005_10After insert count 2022-09-14 11:37:18.372147 0&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;+-------+----------+-------------+-----------+----------+&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;|tran_id|t_store_id|scan_datetime|customer_id|updated_by|&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;+-------+----------+-------------+-----------+----------+&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Anything recently changed in the data bricks, which is impacting this?&lt;/P&gt;&lt;P&gt;Any help on this is appreciated. &lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Thanks&lt;/P&gt;&lt;P&gt;Krishna&lt;/P&gt;</description>
    <pubDate>Wed, 14 Sep 2022 11:42:53 GMT</pubDate>
    <dc:creator>Krish-685291</dc:creator>
    <dc:date>2022-09-14T11:42:53Z</dc:date>
    <item>
      <title>Dataframe loses its contents after the write operation to Database.</title>
      <link>https://community.databricks.com/t5/data-engineering/dataframe-loses-its-contents-after-the-write-operation-to/m-p/31727#M23107</link>
      <description>&lt;P&gt;We had working code as below.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;print(f"{file_name}Before insert count", datetime.datetime.now(), scan_df_new.count())&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;print(scan_df_new.show())&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;scan_20220908120005_10Before insert count 2022-09-14 11:37:15.853588 3&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;+-------------------+----------+-------------------+--------------------+----------+&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;|&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;tran_id|t_store_id|&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;scan_datetime|&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;customer_id|updated_by|&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;+-------------------+----------+-------------------+--------------------+----------+&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;|1230000000000000004|&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;4395|2022-09-08 03:00:01|20220816a51cee4264f1|Databricks|&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;|1230000000000000005|&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;4394|2022-09-08 02:58:00|20220816a51cee4264f1|Databricks|&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;|1230000000000000006|&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;4393|2022-09-08 03:00:04|20220816a51cee4264f1|Databricks|&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;+-------------------+----------+-------------------+--------------------+----------+&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;The data frame after the write operation is used for further business logic processing. This was working earlier. But recently we are observing a strange behavior, where in the data in the data frame is getting lost . When wee check the contents, or even the dataframe count its shows empty. &lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;scan_df_new.write.format("jdbc").option("url", jdbcUrl).option("dbtable", scan_table).mode("append").save()&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;print(f"{file_name}After insert count", datetime.datetime.now(), scan_df_new.count())&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;print(scan_df_new.show())&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;None&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;scan_20220908120005_10After insert count 2022-09-14 11:37:18.372147 0&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;+-------+----------+-------------+-----------+----------+&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;|tran_id|t_store_id|scan_datetime|customer_id|updated_by|&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;+-------+----------+-------------+-----------+----------+&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Anything recently changed in the data bricks, which is impacting this?&lt;/P&gt;&lt;P&gt;Any help on this is appreciated. &lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Thanks&lt;/P&gt;&lt;P&gt;Krishna&lt;/P&gt;</description>
      <pubDate>Wed, 14 Sep 2022 11:42:53 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/dataframe-loses-its-contents-after-the-write-operation-to/m-p/31727#M23107</guid>
      <dc:creator>Krish-685291</dc:creator>
      <dc:date>2022-09-14T11:42:53Z</dc:date>
    </item>
  </channel>
</rss>

