<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Delta Live Tables UDFs and Versions in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/delta-live-tables-udfs-and-versions/m-p/60310#M31633</link>
    <description>&lt;P&gt;Thanks &lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/9"&gt;@Retired_mod&lt;/a&gt;,&amp;nbsp;for reference if anybody finds this, the DLT release docs are here: &lt;A href="https://docs.databricks.com/en/release-notes/delta-live-tables/index.html" target="_blank"&gt;https://docs.databricks.com/en/release-notes/delta-live-tables/index.html&lt;/A&gt;&lt;BR /&gt;This shows which versions are running for CURRENT and PREVIEW channels. In this case, was running on CURRENT channel (Spark 3.3.2), so PREVIEW channel (Spark 3.5.0) should work for the latest PySpark functions.&lt;/P&gt;</description>
    <pubDate>Thu, 15 Feb 2024 12:54:40 GMT</pubDate>
    <dc:creator>NotARobot</dc:creator>
    <dc:date>2024-02-15T12:54:40Z</dc:date>
    <item>
      <title>Delta Live Tables UDFs and Versions</title>
      <link>https://community.databricks.com/t5/data-engineering/delta-live-tables-udfs-and-versions/m-p/59989#M31558</link>
      <description>&lt;P&gt;Trying to do a url_decode on a column, which works great in development, but running via DLT fails when trying multiple ways.&lt;/P&gt;&lt;P&gt;1. &lt;A title="URL Decode" href="https://spark.apache.org/docs/latest/api/python/reference/pyspark.sql/api/pyspark.sql.functions.url_decode.html" target="_self"&gt;pyspark.sql.functions.url_decode&lt;/A&gt; - This is new as of 3.5.0, but isn't supported using whatever version running a DLT pipeline provides. I haven't been able to figure out what version of PySpark this is actually running. It says 12.2, but I suspect that might actually be the version of something else:&lt;BR /&gt;&lt;FONT face="courier new,courier"&gt;&lt;SPAN class=""&gt;dlt:12.2-delta-pipelines-dlt-release-2024.04-rc0-commit-24b74&lt;/SPAN&gt;&lt;/FONT&gt;&lt;/P&gt;&lt;P&gt;2. Attempt to use a simple UDF that wraps &lt;A title="unquote_plus" href="https://docs.python.org/3/library/urllib.parse.html#urllib.parse.unquote_plus" target="_self"&gt;urllib.parse.unquote_plus&lt;/A&gt;, however this appears to be unsupported with Unit Catalog. Given the &lt;A title="documentation" href="https://learn.microsoft.com/en-us/azure/databricks/udf/python" target="_self"&gt;documentation&lt;/A&gt; states that this should be supported in versions greater than 13.1, again guessing the version is why I get this error:&lt;BR /&gt;&lt;FONT face="courier new,courier"&gt;&lt;SPAN class=""&gt;&lt;SPAN class=""&gt;pyspark.errors.exceptions.AnalysisException: [UC_COMMAND_NOT_SUPPORTED] UDF/UDAF functions are not supported in Unity Catalog&lt;/SPAN&gt;&lt;/SPAN&gt;&lt;/FONT&gt;&lt;/P&gt;&lt;P&gt;3. Have also tried to use cluster policies to attempt to set the version, however regardless of what version this attempts to force, cluster gets the same version as above. Have tried using regex, explicit version, and auto:latest with no luck.&lt;/P&gt;&lt;P&gt;This leads to two questions:&lt;BR /&gt;1. What version of PySpark is DLT running and how can users consistently find this to know what is available for use?&lt;BR /&gt;2. How do users force versions if cluster policies don't work?&lt;BR /&gt;3. Any other recommendations for doing a URL decode via DLT, since this is where the rest of our ETL pipeline is running, would prefer to not fragment out tables into separate workflows to manage.&lt;/P&gt;</description>
      <pubDate>Tue, 13 Feb 2024 00:13:54 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/delta-live-tables-udfs-and-versions/m-p/59989#M31558</guid>
      <dc:creator>NotARobot</dc:creator>
      <dc:date>2024-02-13T00:13:54Z</dc:date>
    </item>
    <item>
      <title>Re: Delta Live Tables UDFs and Versions</title>
      <link>https://community.databricks.com/t5/data-engineering/delta-live-tables-udfs-and-versions/m-p/60310#M31633</link>
      <description>&lt;P&gt;Thanks &lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/9"&gt;@Retired_mod&lt;/a&gt;,&amp;nbsp;for reference if anybody finds this, the DLT release docs are here: &lt;A href="https://docs.databricks.com/en/release-notes/delta-live-tables/index.html" target="_blank"&gt;https://docs.databricks.com/en/release-notes/delta-live-tables/index.html&lt;/A&gt;&lt;BR /&gt;This shows which versions are running for CURRENT and PREVIEW channels. In this case, was running on CURRENT channel (Spark 3.3.2), so PREVIEW channel (Spark 3.5.0) should work for the latest PySpark functions.&lt;/P&gt;</description>
      <pubDate>Thu, 15 Feb 2024 12:54:40 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/delta-live-tables-udfs-and-versions/m-p/60310#M31633</guid>
      <dc:creator>NotARobot</dc:creator>
      <dc:date>2024-02-15T12:54:40Z</dc:date>
    </item>
  </channel>
</rss>

