<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Facing issues while upgrading DBR version from 9.1 LTS to 15.4 LTS in Get Started Discussions</title>
    <link>https://community.databricks.com/t5/get-started-discussions/facing-issues-while-upgrading-dbr-version-from-9-1-lts-to-15-4/m-p/114901#M9344</link>
    <description>&lt;P&gt;Hi &lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/34815"&gt;@Louis_Frolio&lt;/a&gt;&amp;nbsp;Thanks for replying.&lt;/P&gt;&lt;P&gt;I tried by removing all predefined libraries from init script and after starting the cluster, installed required libraries manually. but the issue is still same. I am sharing complete error details below. could you all please look in detail and share some more insights. your Response will be much helpful to us.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;&lt;DIV class=""&gt;&lt;DIV&gt;Py4JJavaError: An error occurred while calling z:org.apache.spark.api.python.PythonRDD.runJob.&lt;/DIV&gt;&lt;DIV&gt;: org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 1.0 failed 4 times, most recent failure: Lost task 0.3 in stage 1.0 (TID 7) (10.141.5.10 executor 0): org.apache.spark.api.python.PythonException: Traceback (most recent call last):&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 192, in _read_with_length&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; return self.loads(obj)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;^^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 572, in loads&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; return cloudpickle.loads(obj, encoding=encoding)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;&lt;STRONG&gt;ModuleNotFoundError: No module named 'dbruntime'&lt;/STRONG&gt;&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;During handling of the above exception, another exception occurred:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;Traceback (most recent call last):&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/worker.py", line 1980, in main&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; process()&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/worker.py", line 1972, in process&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; serializer.dump_stream(out_iter, outfile)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 356, in dump_stream&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; vs = list(itertools.islice(iterator, batch))&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/core/rdd.py", line 2757, in takeUpToNumLeft&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; yield next(iterator)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; ^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 162, in load_stream&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; yield self._read_with_length(stream)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 196, in _read_with_length&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; raise SerializationError("Caused by " + traceback.format_exc())&lt;/DIV&gt;&lt;DIV&gt;pyspark.serializers.SerializationError: Caused by Traceback (most recent call last):&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 192, in _read_with_length&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; return self.loads(obj)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;^^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 572, in loads&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; return cloudpickle.loads(obj, encoding=encoding)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;ModuleNotFoundError: No module named 'dbruntime'&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.api.python.BasePythonRunner$ReaderIterator.handlePythonException(PythonRunner.scala:560)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.api.python.PythonRunner$$anon$3.read(PythonRunner.scala:968)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.api.python.PythonRunner$$anon$3.read(PythonRunner.scala:953)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.api.python.BasePythonRunner$ReaderIterator.hasNext(PythonRunner.scala:516)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:37)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.Iterator.foreach(Iterator.scala:943)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.Iterator.foreach$(Iterator.scala:943)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.InterruptibleIterator.foreach(InterruptibleIterator.scala:28)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.generic.Growable.$plus$plus$eq(Growable.scala:62)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.generic.Growable.$plus$plus$eq$(Growable.scala:53)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:105)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:49)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableOnce.to(TraversableOnce.scala:366)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableOnce.to$(TraversableOnce.scala:364)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.InterruptibleIterator.to(InterruptibleIterator.scala:28)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableOnce.toBuffer(TraversableOnce.scala:358)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableOnce.toBuffer$(TraversableOnce.scala:358)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.InterruptibleIterator.toBuffer(InterruptibleIterator.scala:28)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableOnce.toArray(TraversableOnce.scala:345)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableOnce.toArray$(TraversableOnce.scala:339)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.InterruptibleIterator.toArray(InterruptibleIterator.scala:28)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.api.python.PythonRDD$.$anonfun$collectPartitions$1(PythonRDD.scala:234)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.SparkContext.$anonfun$runJob$2(SparkContext.scala:3208)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.ResultTask.$anonfun$runTask$3(ResultTask.scala:82)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.ExecutorFrameProfiler$.record(ExecutorFrameProfiler.scala:110)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.ResultTask.$anonfun$runTask$1(ResultTask.scala:82)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.ExecutorFrameProfiler$.record(ExecutorFrameProfiler.scala:110)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:62)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:225)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.Task.doRunTask(Task.scala:199)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.Task.$anonfun$run$5(Task.scala:161)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.unity.EmptyHandle$.runWithAndClose(UCSHandle.scala:134)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.Task.$anonfun$run$1(Task.scala:155)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.ExecutorFrameProfiler$.record(ExecutorFrameProfiler.scala:110)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.Task.run(Task.scala:102)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$10(Executor.scala:1043)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:64)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:61)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:111)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:1046)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.ExecutorFrameProfiler$.record(ExecutorFrameProfiler.scala:110)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:933)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at java.lang.Thread.run(Thread.java:750)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;Driver stacktrace:&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.DAGScheduler.$anonfun$failJobAndIndependentStages$1(DAGScheduler.scala:4043)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.Option.getOrElse(Option.scala:189)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.DAGScheduler.failJobAndIndependentStages(DAGScheduler.scala:4041)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2(DAGScheduler.scala:3954)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2$adapted(DAGScheduler.scala:3941)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:3941)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1(DAGScheduler.scala:1767)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1$adapted(DAGScheduler.scala:1750)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.Option.foreach(Option.scala:407)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:1750)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:4302)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:4204)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:4190)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:55)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.DAGScheduler.$anonfun$runJob$1(DAGScheduler.scala:1413)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.FrameProfiler$.record(FrameProfiler.scala:94)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:1401)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.SparkContext.runJobInternal(SparkContext.scala:3168)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.SparkContext.runJob(SparkContext.scala:3149)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.SparkContext.runJob(SparkContext.scala:3189)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.SparkContext.runJob(SparkContext.scala:3208)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.api.python.PythonRDD$.collectPartitions(PythonRDD.scala:234)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.api.python.PythonRDD$.runJob(PythonRDD.scala:254)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.api.python.PythonRDD.runJob(PythonRDD.scala)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at java.lang.reflect.Method.invoke(Method.java:498)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:397)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at py4j.Gateway.invoke(Gateway.java:306)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at py4j.commands.CallCommand.execute(CallCommand.java:79)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at py4j.ClientServerConnection.waitForCommands(ClientServerConnection.java:199)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at py4j.ClientServerConnection.run(ClientServerConnection.java:119)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at java.lang.Thread.run(Thread.java:750)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;Caused by: org.apache.spark.api.python.PythonException: Traceback (most recent call last):&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 192, in _read_with_length&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; return self.loads(obj)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;^^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 572, in loads&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; return cloudpickle.loads(obj, encoding=encoding)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;ModuleNotFoundError: No module named 'dbruntime'&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;During handling of the above exception, another exception occurred:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;Traceback (most recent call last):&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/worker.py", line 1980, in main&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; process()&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/worker.py", line 1972, in process&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; serializer.dump_stream(out_iter, outfile)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 356, in dump_stream&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; vs = list(itertools.islice(iterator, batch))&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/core/rdd.py", line 2757, in takeUpToNumLeft&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; yield next(iterator)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; ^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 162, in load_stream&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; yield self._read_with_length(stream)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 196, in _read_with_length&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; raise SerializationError("Caused by " + traceback.format_exc())&lt;/DIV&gt;&lt;DIV&gt;pyspark.serializers.SerializationError: Caused by Traceback (most recent call last):&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 192, in _read_with_length&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; return self.loads(obj)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;^^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 572, in loads&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; return cloudpickle.loads(obj, encoding=encoding)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;ModuleNotFoundError: No module named 'dbruntime'&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.api.python.BasePythonRunner$ReaderIterator.handlePythonException(PythonRunner.scala:560)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.api.python.PythonRunner$$anon$3.read(PythonRunner.scala:968)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.api.python.PythonRunner$$anon$3.read(PythonRunner.scala:953)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.api.python.BasePythonRunner$ReaderIterator.hasNext(PythonRunner.scala:516)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:37)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.Iterator.foreach(Iterator.scala:943)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.Iterator.foreach$(Iterator.scala:943)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.InterruptibleIterator.foreach(InterruptibleIterator.scala:28)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.generic.Growable.$plus$plus$eq(Growable.scala:62)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.generic.Growable.$plus$plus$eq$(Growable.scala:53)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:105)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:49)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableOnce.to(TraversableOnce.scala:366)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableOnce.to$(TraversableOnce.scala:364)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.InterruptibleIterator.to(InterruptibleIterator.scala:28)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableOnce.toBuffer(TraversableOnce.scala:358)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableOnce.toBuffer$(TraversableOnce.scala:358)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.InterruptibleIterator.toBuffer(InterruptibleIterator.scala:28)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableOnce.toArray(TraversableOnce.scala:345)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableOnce.toArray$(TraversableOnce.scala:339)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.InterruptibleIterator.toArray(InterruptibleIterator.scala:28)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.api.python.PythonRDD$.$anonfun$collectPartitions$1(PythonRDD.scala:234)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.SparkContext.$anonfun$runJob$2(SparkContext.scala:3208)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.ResultTask.$anonfun$runTask$3(ResultTask.scala:82)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.ExecutorFrameProfiler$.record(ExecutorFrameProfiler.scala:110)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.ResultTask.$anonfun$runTask$1(ResultTask.scala:82)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.ExecutorFrameProfiler$.record(ExecutorFrameProfiler.scala:110)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:62)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:225)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.Task.doRunTask(Task.scala:199)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.Task.$anonfun$run$5(Task.scala:161)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.unity.EmptyHandle$.runWithAndClose(UCSHandle.scala:134)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.Task.$anonfun$run$1(Task.scala:155)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.ExecutorFrameProfiler$.record(ExecutorFrameProfiler.scala:110)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.Task.run(Task.scala:102)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$10(Executor.scala:1043)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:64)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:61)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:111)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:1046)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.ExecutorFrameProfiler$.record(ExecutorFrameProfiler.scala:110)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:933)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;... 1 more&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;File &amp;lt;command-3609477148953765&amp;gt;, line 195&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 193 start = time.time()&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 194 start11 = time.time()&lt;/DIV&gt;&lt;DIV&gt;--&amp;gt; 195 print(common.readZipFromStorage(sc, dbutils, "2025", "04", "07", True).mapPartitions(processdata).count())&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 196 print(common.readZipFromStorage(sc, dbutils, "2025", "04", "08", False).mapPartitions(processdata).count())&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 197 print("End of main...")&lt;/DIV&gt;&lt;DIV&gt;File /local_disk0/spark-8459fa57-12da-4b04-9ebe-30bf615a5380/userFiles-d0b4be2b-cfc5-4eb0-9626-f647444f128c/Lib.zip/Lib/common/__init__.py:132, in readZipFromStorage(sc, dbutils, year, month, date, chk)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 129 rdd = sc.parallelize(final_list_of_files)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 131 # Create data frame&lt;/DIV&gt;&lt;DIV&gt;--&amp;gt; 132 df = rdd.toDF()&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 133 df = df.filter(df.name.endswith(".zip"))&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 135 # Extract only name column from df&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/pyspark/sql/session.py:133, in _monkey_patch_RDD.&amp;lt;locals&amp;gt;.toDF(self, schema, sampleRatio)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;98 &lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/16525"&gt;@No&lt;/a&gt;_type_check&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;99 def toDF(self, schema=None, sampleRatio=None):&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 100&amp;nbsp; &amp;nbsp; &amp;nbsp;"""&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 101&amp;nbsp; &amp;nbsp; &amp;nbsp;Converts current :class:`RDD` into a :class:`DataFrame`&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 102&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;(...)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 131&amp;nbsp; &amp;nbsp; &amp;nbsp;+---+&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 132&amp;nbsp; &amp;nbsp; &amp;nbsp;"""&lt;/DIV&gt;&lt;DIV&gt;--&amp;gt; 133&amp;nbsp; &amp;nbsp; &amp;nbsp;return sparkSession.createDataFrame(self, schema, sampleRatio)&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/pyspark/instrumentation_utils.py:47, in _wrap_function.&amp;lt;locals&amp;gt;.wrapper(*args, **kwargs)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;45 start = time.perf_counter()&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;46 try:&lt;/DIV&gt;&lt;DIV&gt;---&amp;gt; 47&amp;nbsp; &amp;nbsp; &amp;nbsp;res = func(*args, **kwargs)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;48&amp;nbsp; &amp;nbsp; &amp;nbsp;logger.log_success(&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;49&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;module_name, class_name, function_name, time.perf_counter() - start, signature&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;50&amp;nbsp; &amp;nbsp; &amp;nbsp;)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;51&amp;nbsp; &amp;nbsp; &amp;nbsp;return res&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/pyspark/sql/session.py:1610, in SparkSession.createDataFrame(self, data, schema, samplingRatio, verifySchema)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1605 if has_pandas and isinstance(data, pd.DataFrame):&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1606&amp;nbsp; &amp;nbsp; &amp;nbsp;# Create a DataFrame from pandas DataFrame.&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1607&amp;nbsp; &amp;nbsp; &amp;nbsp;return super(SparkSession, self).createDataFrame(&amp;nbsp; # type: ignore[call-overload]&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1608&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;data, schema, samplingRatio, verifySchema&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1609&amp;nbsp; &amp;nbsp; &amp;nbsp;)&lt;/DIV&gt;&lt;DIV&gt;-&amp;gt; 1610 return self._create_dataframe(&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1611&amp;nbsp; &amp;nbsp; &amp;nbsp;data, schema, samplingRatio, verifySchema&amp;nbsp; # type: ignore[arg-type]&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1612 )&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/pyspark/sql/session.py:1665, in SparkSession._create_dataframe(self, data, schema, samplingRatio, verifySchema)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1663 else:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1664&amp;nbsp; &amp;nbsp; &amp;nbsp;if not is_remote_only() and isinstance(data, RDD):&lt;/DIV&gt;&lt;DIV&gt;-&amp;gt; 1665&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;rdd, struct = self._createFromRDD(data.map(prepare), schema, samplingRatio)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1666&amp;nbsp; &amp;nbsp; &amp;nbsp;else:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1667&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;rdd, struct = self._createFromLocal(map(prepare, data), schema)&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/pyspark/sql/session.py:1168, in SparkSession._createFromRDD(self, rdd, schema, samplingRatio)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1164 """&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1165 Create an RDD for DataFrame from an existing RDD, returns the RDD and schema.&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1166 """&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1167 if schema is None or isinstance(schema, (list, tuple)):&lt;/DIV&gt;&lt;DIV&gt;-&amp;gt; 1168&amp;nbsp; &amp;nbsp; &amp;nbsp;struct = self._inferSchema(rdd, samplingRatio, names=schema)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1169&amp;nbsp; &amp;nbsp; &amp;nbsp;converter = _create_converter(struct)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1170&amp;nbsp; &amp;nbsp; &amp;nbsp;tupled_rdd = rdd.map(converter)&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/pyspark/sql/session.py:1108, in SparkSession._inferSchema(self, rdd, samplingRatio, names)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1087 def _inferSchema(&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1088&amp;nbsp; &amp;nbsp; &amp;nbsp;self,&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1089&amp;nbsp; &amp;nbsp; &amp;nbsp;rdd: "RDD[Any]",&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1090&amp;nbsp; &amp;nbsp; &amp;nbsp;samplingRatio: Optional[float] = None,&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1091&amp;nbsp; &amp;nbsp; &amp;nbsp;names: Optional[List[str]] = None,&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1092 ) -&amp;gt; StructType:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1093&amp;nbsp; &amp;nbsp; &amp;nbsp;"""&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1094&amp;nbsp; &amp;nbsp; &amp;nbsp;Infer schema from an RDD of Row, dict, or tuple.&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1095&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;(...)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1106&amp;nbsp; &amp;nbsp; &amp;nbsp;:class:`pyspark.sql.types.StructType`&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1107&amp;nbsp; &amp;nbsp; &amp;nbsp;"""&lt;/DIV&gt;&lt;DIV&gt;-&amp;gt; 1108&amp;nbsp; &amp;nbsp; &amp;nbsp;first = rdd.first()&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1109&amp;nbsp; &amp;nbsp; &amp;nbsp;if isinstance(first, Sized) and len(first) == 0:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1110&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;raise ValueError("The first row in RDD is empty, can not infer schema")&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/pyspark/instrumentation_utils.py:42, in _wrap_function.&amp;lt;locals&amp;gt;.wrapper(*args, **kwargs)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;38 @functools.wraps(func)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;39 def wrapper(*args: Any, **kwargs: Any) -&amp;gt; Any:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;40&amp;nbsp; &amp;nbsp; &amp;nbsp;if hasattr(_local, "logging") and _local.logging:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;41&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;# no need to log since this should be internal call.&lt;/DIV&gt;&lt;DIV&gt;---&amp;gt; 42&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;return func(*args, **kwargs)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;43&amp;nbsp; &amp;nbsp; &amp;nbsp;_local.logging = True&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;44&amp;nbsp; &amp;nbsp; &amp;nbsp;try:&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/pyspark/core/rdd.py:2796, in RDD.first(self)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2770 def first(self: "RDD[T]") -&amp;gt; T:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2771&amp;nbsp; &amp;nbsp; &amp;nbsp;"""&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2772&amp;nbsp; &amp;nbsp; &amp;nbsp;Return the first element in this RDD.&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2773&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;(...)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2794&amp;nbsp; &amp;nbsp; &amp;nbsp;ValueError: RDD is empty&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2795&amp;nbsp; &amp;nbsp; &amp;nbsp;"""&lt;/DIV&gt;&lt;DIV&gt;-&amp;gt; 2796&amp;nbsp; &amp;nbsp; &amp;nbsp;rs = self.take(1)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2797&amp;nbsp; &amp;nbsp; &amp;nbsp;if rs:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2798&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;return rs[0]&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/pyspark/instrumentation_utils.py:42, in _wrap_function.&amp;lt;locals&amp;gt;.wrapper(*args, **kwargs)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;38 @functools.wraps(func)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;39 def wrapper(*args: Any, **kwargs: Any) -&amp;gt; Any:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;40&amp;nbsp; &amp;nbsp; &amp;nbsp;if hasattr(_local, "logging") and _local.logging:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;41&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;# no need to log since this should be internal call.&lt;/DIV&gt;&lt;DIV&gt;---&amp;gt; 42&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;return func(*args, **kwargs)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;43&amp;nbsp; &amp;nbsp; &amp;nbsp;_local.logging = True&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;44&amp;nbsp; &amp;nbsp; &amp;nbsp;try:&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/pyspark/core/rdd.py:2763, in RDD.take(self, num)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2760&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;taken += 1&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2762 p = range(partsScanned, min(partsScanned + numPartsToTry, totalParts))&lt;/DIV&gt;&lt;DIV&gt;-&amp;gt; 2763 res = self.context.runJob(self, takeUpToNumLeft, p)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2765 items += res&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2766 partsScanned += numPartsToTry&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/pyspark/core/context.py:2702, in SparkContext.runJob(self, rdd, partitionFunc, partitions, allowLocal)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2700&amp;nbsp; &amp;nbsp; &amp;nbsp;finally:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2701&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;os.remove(filename)&lt;/DIV&gt;&lt;DIV&gt;-&amp;gt; 2702 sock_info = self._jvm.PythonRDD.runJob(self._jsc.sc(), mappedRDD._jrdd, partitions)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2703 return list(_load_from_socket(sock_info, mappedRDD._jrdd_deserializer))&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/lib/py4j-0.10.9.7-src.zip/py4j/java_gateway.py:1355, in JavaMember.__call__(self, *args)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1349 command = proto.CALL_COMMAND_NAME +\&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1350&amp;nbsp; &amp;nbsp; &amp;nbsp;self.command_header +\&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1351&amp;nbsp; &amp;nbsp; &amp;nbsp;args_command +\&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1352&amp;nbsp; &amp;nbsp; &amp;nbsp;proto.END_COMMAND_PART&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1354 answer = self.gateway_client.send_command(command)&lt;/DIV&gt;&lt;DIV&gt;-&amp;gt; 1355 return_value = get_return_value(&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1356&amp;nbsp; &amp;nbsp; &amp;nbsp;answer, self.gateway_client, self.target_id, self.name)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1358 for temp_arg in temp_args:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1359&amp;nbsp; &amp;nbsp; &amp;nbsp;if hasattr(temp_arg, "_detach"):&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/pyspark/errors/exceptions/captured.py:255, in capture_sql_exception.&amp;lt;locals&amp;gt;.deco(*a, **kw)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 252 from py4j.protocol import Py4JJavaError&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 254 try:&lt;/DIV&gt;&lt;DIV&gt;--&amp;gt; 255&amp;nbsp; &amp;nbsp; &amp;nbsp;return f(*a, **kw)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 256 except Py4JJavaError as e:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 257&amp;nbsp; &amp;nbsp; &amp;nbsp;converted = convert_exception(e.java_exception)&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/lib/py4j-0.10.9.7-src.zip/py4j/protocol.py:326, in get_return_value(answer, gateway_client, target_id, name)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 324 value = OUTPUT_CONVERTER[type](answer[2:], gateway_client)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 325 if answer[1] == REFERENCE_TYPE:&lt;/DIV&gt;&lt;DIV&gt;--&amp;gt; 326&amp;nbsp; &amp;nbsp; &amp;nbsp;raise Py4JJavaError(&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 327&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;"An error occurred while calling {0}{1}{2}.\n".&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 328&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;format(target_id, ".", name), value)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 329 else:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 330&amp;nbsp; &amp;nbsp; &amp;nbsp;raise Py4JError(&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 331&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;"An error occurred while calling {0}{1}{2}. Trace:\n{3}\n".&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 332&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;format(target_id, ".", name, value))&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;Here is my Cluster Configuration with DBR version 15.4 LTS:&lt;/DIV&gt;&lt;DIV&gt;{&lt;BR /&gt;"cluster_id": "0321-120122-h2mf81qg",&lt;BR /&gt;"creator_user_name": "rahul.790578@gmail.com",&lt;BR /&gt;"driver": {&lt;BR /&gt;"private_ip": "10.141.5.11",&lt;BR /&gt;"public_dns": "172.184.184.61",&lt;BR /&gt;"node_id": "9818cc182cfd4b40b2210d69cd531011",&lt;BR /&gt;"instance_id": "0c4d9ca030fd44a9b347acb8cef8d8f9",&lt;BR /&gt;"start_timestamp": 1744177619114,&lt;BR /&gt;"node_attributes": {&lt;BR /&gt;"is_spot": false&lt;BR /&gt;},&lt;BR /&gt;"host_private_ip": "10.141.4.10"&lt;BR /&gt;},&lt;BR /&gt;"executors": [&lt;BR /&gt;{&lt;BR /&gt;"private_ip": "10.141.5.12",&lt;BR /&gt;"public_dns": "13.93.231.102",&lt;BR /&gt;"node_id": "7f4fcc2b934e44199daa7f1d58c94e88",&lt;BR /&gt;"instance_id": "5e0a86855feb48528ef18e9c4e2cad83",&lt;BR /&gt;"start_timestamp": 1744177619187,&lt;BR /&gt;"node_attributes": {&lt;BR /&gt;"is_spot": true&lt;BR /&gt;},&lt;BR /&gt;"host_private_ip": "10.141.4.11"&lt;BR /&gt;},&lt;BR /&gt;{&lt;BR /&gt;"private_ip": "10.141.5.10",&lt;BR /&gt;"public_dns": "20.253.192.255",&lt;BR /&gt;"node_id": "b0f85ccd11b24d23919578216453a496",&lt;BR /&gt;"instance_id": "9b7b71b46ca946b8a87895e9cf7602b2",&lt;BR /&gt;"start_timestamp": 1744177619142,&lt;BR /&gt;"node_attributes": {&lt;BR /&gt;"is_spot": true&lt;BR /&gt;},&lt;BR /&gt;"host_private_ip": "10.141.4.12"&lt;BR /&gt;}&lt;BR /&gt;],&lt;BR /&gt;"spark_context_id": 791941095622325100,&lt;BR /&gt;"driver_healthy": true,&lt;BR /&gt;"jdbc_port": 10000,&lt;BR /&gt;"cluster_name": "Rahul Jaiswa's Cluster",&lt;BR /&gt;"spark_version": "15.4.x-scala2.12",&lt;BR /&gt;"spark_conf": {&lt;BR /&gt;"spark.sql.adaptive.coalescePartitions.initialPartitionNum": "8",&lt;BR /&gt;"spark.sql.files.maxPartitionBytes": "2469606",&lt;BR /&gt;"spark.sql.files.minPartitionNum": "8",&lt;BR /&gt;"spark.sql.adaptive.enabled": "true",&lt;BR /&gt;"spark.driver.maxResultSize": "5g",&lt;BR /&gt;"spark.databricks.delta.preview.enabled": "true",&lt;BR /&gt;"spark.sql.adaptive.advisoryPartitionSizeInBytes": "2469606",&lt;BR /&gt;"spark.driver.memory": "5g",&lt;BR /&gt;"spark.default.parallelism": "8",&lt;BR /&gt;"spark.sql.adaptive.coalescePartitions.enabled": "true",&lt;BR /&gt;"spark.driver.cores": "8",&lt;BR /&gt;"spark.executor.memory": "4g",&lt;BR /&gt;"spark.sql.files.openCostInBytes": "134217",&lt;BR /&gt;"spark.sql.adaptive.coalescePartitions.minPartitionNum": "8",&lt;BR /&gt;"spark.sql.shuffle.partitions": "8"&lt;BR /&gt;},&lt;BR /&gt;"azure_attributes": {&lt;BR /&gt;"first_on_demand": 1,&lt;BR /&gt;"availability": "SPOT_WITH_FALLBACK_AZURE",&lt;BR /&gt;"spot_bid_max_price": -1&lt;BR /&gt;},&lt;BR /&gt;"node_type_id": "Standard_E8s_v3",&lt;BR /&gt;"driver_node_type_id": "Standard_E8s_v3",&lt;BR /&gt;"cluster_log_conf": {&lt;BR /&gt;"dbfs": {&lt;BR /&gt;"destination": "dbfs:/FileStore/manual-logs"&lt;BR /&gt;}&lt;BR /&gt;},&lt;BR /&gt;"spark_env_vars": {&lt;BR /&gt;"PYSPARK_PYTHON": "/databricks/python3/bin/python3"&lt;BR /&gt;},&lt;BR /&gt;"autotermination_minutes": 60,&lt;BR /&gt;"enable_elastic_disk": true,&lt;BR /&gt;"disk_spec": {},&lt;BR /&gt;"cluster_source": "UI",&lt;BR /&gt;"init_scripts": [&lt;BR /&gt;{&lt;BR /&gt;"workspace": {&lt;BR /&gt;"destination": "/startup1/pyodbc-install.sh"&lt;BR /&gt;}&lt;BR /&gt;}&lt;BR /&gt;],&lt;BR /&gt;"single_user_name": "rahul.790578@gmail.com",&lt;BR /&gt;"enable_local_disk_encryption": false,&lt;BR /&gt;"instance_source": {&lt;BR /&gt;"node_type_id": "Standard_E8s_v3"&lt;BR /&gt;},&lt;BR /&gt;"driver_instance_source": {&lt;BR /&gt;"node_type_id": "Standard_E8s_v3"&lt;BR /&gt;},&lt;BR /&gt;"data_security_mode": "LEGACY_SINGLE_USER_STANDARD",&lt;BR /&gt;"runtime_engine": "PHOTON",&lt;BR /&gt;"effective_spark_version": "15.4.x-photon-scala2.12",&lt;BR /&gt;"assigned_principal": "user:rahul.790578@gmail.com",&lt;BR /&gt;"release_version": "15.4.12",&lt;BR /&gt;"state": "RUNNING",&lt;BR /&gt;"state_message": "",&lt;BR /&gt;"start_time": 1742558482247,&lt;BR /&gt;"last_state_loss_time": 1744177737748,&lt;BR /&gt;"last_activity_time": 1744177679642,&lt;BR /&gt;"last_restarted_time": 1744177737824,&lt;BR /&gt;"autoscale": {&lt;BR /&gt;"min_workers": 2,&lt;BR /&gt;"max_workers": 3,&lt;BR /&gt;"target_workers": 2&lt;BR /&gt;},&lt;BR /&gt;"cluster_memory_mb": 196608,&lt;BR /&gt;"cluster_cores": 24,&lt;BR /&gt;"default_tags": {&lt;BR /&gt;"Vendor": "Databricks",&lt;BR /&gt;"Creator": "rahul.790578@gmail.com",&lt;BR /&gt;"ClusterName": "Rahul Jaiswa's Cluster",&lt;BR /&gt;"ClusterId": "0321-120122-h2mf81qg",&lt;BR /&gt;"Environmet": "dvt"&lt;BR /&gt;},&lt;BR /&gt;"cluster_log_status": {&lt;BR /&gt;"last_attempted": 1744178955397&lt;BR /&gt;},&lt;BR /&gt;"init_scripts_safe_mode": false,&lt;BR /&gt;"spec": {&lt;BR /&gt;"cluster_name": "Rahul Jaiswa's Cluster",&lt;BR /&gt;"&lt;STRONG&gt;spark_version": "15.4.x-scala2.12&lt;/STRONG&gt;",&lt;BR /&gt;"spark_conf": {&lt;BR /&gt;"spark.databricks.delta.preview.enabled": "true",&lt;BR /&gt;"spark.sql.files.minPartitionNum": "8",&lt;BR /&gt;"spark.sql.files.openCostInBytes": "134217",&lt;BR /&gt;"spark.sql.adaptive.coalescePartitions.initialPartitionNum": "8",&lt;BR /&gt;"spark.sql.adaptive.coalescePartitions.enabled": "true",&lt;BR /&gt;"spark.executor.memory": "4g",&lt;BR /&gt;"spark.driver.maxResultSize": "5g",&lt;BR /&gt;"spark.sql.shuffle.partitions": "8",&lt;BR /&gt;"spark.driver.memory": "5g",&lt;BR /&gt;"spark.sql.adaptive.coalescePartitions.minPartitionNum": "8",&lt;BR /&gt;"spark.sql.adaptive.enabled": "true",&lt;BR /&gt;"spark.sql.adaptive.advisoryPartitionSizeInBytes": "2469606",&lt;BR /&gt;"spark.sql.files.maxPartitionBytes": "2469606",&lt;BR /&gt;"spark.default.parallelism": "8",&lt;BR /&gt;"spark.driver.cores": "8"&lt;BR /&gt;},&lt;BR /&gt;"azure_attributes": {&lt;BR /&gt;"first_on_demand": 1,&lt;BR /&gt;"availability": "SPOT_WITH_FALLBACK_AZURE",&lt;BR /&gt;"spot_bid_max_price": -1&lt;BR /&gt;},&lt;BR /&gt;"node_type_id": "Standard_E8s_v3",&lt;BR /&gt;"driver_node_type_id": "Standard_E8s_v3",&lt;BR /&gt;"cluster_log_conf": {&lt;BR /&gt;"dbfs": {&lt;BR /&gt;"destination": "dbfs:/FileStore/manual-logs"&lt;BR /&gt;}&lt;BR /&gt;},&lt;BR /&gt;"spark_env_vars": {&lt;BR /&gt;"PYSPARK_PYTHON": "/databricks/python3/bin/python3"&lt;BR /&gt;},&lt;BR /&gt;"autotermination_minutes": 60,&lt;BR /&gt;"enable_elastic_disk": true,&lt;BR /&gt;"init_scripts": [&lt;BR /&gt;{&lt;BR /&gt;"workspace": {&lt;BR /&gt;"destination": "/startup1/pyodbc-install.sh"&lt;BR /&gt;}&lt;BR /&gt;}&lt;BR /&gt;],&lt;BR /&gt;"single_user_name": "rahul.790578@gmail.com",&lt;BR /&gt;"enable_local_disk_encryption": false,&lt;BR /&gt;"data_security_mode": "LEGACY_SINGLE_USER_STANDARD",&lt;BR /&gt;"runtime_engine": "PHOTON",&lt;BR /&gt;"assigned_principal": "user:rrahul.790578@gmail.com",&lt;BR /&gt;"autoscale": {&lt;BR /&gt;"min_workers": 2,&lt;BR /&gt;"max_workers": 3&lt;BR /&gt;},&lt;BR /&gt;"apply_policy_default_values": false&lt;BR /&gt;}&lt;BR /&gt;}&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;</description>
    <pubDate>Wed, 09 Apr 2025 06:22:51 GMT</pubDate>
    <dc:creator>Kuchnhi</dc:creator>
    <dc:date>2025-04-09T06:22:51Z</dc:date>
    <item>
      <title>Facing issues while upgrading DBR version from 9.1 LTS to 15.4 LTS</title>
      <link>https://community.databricks.com/t5/get-started-discussions/facing-issues-while-upgrading-dbr-version-from-9-1-lts-to-15-4/m-p/114800#M9342</link>
      <description>&lt;P&gt;Dear all,&lt;/P&gt;&lt;P&gt;I am upgrading DBR version from 9.1 LTS to 15.4 LTS in Azure Databricks. for that I have created a new cluster with 15.4 DBR attached init script for installing application dependencies. Cluster has started successfully but it takes 30 min. So, when I run my application I am getting below error. what is the reason of issue and how to resolve that?&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; Py4JJavaError: An error occurred while calling z:org.apache.spark.api.python.PythonRDD.runJob.&lt;BR /&gt;: org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 0.0 failed 4 times, most recent failure: Lost task 0.3 in stage 0.0 (TID 3) (10.141.5.11 executor 5): org.apache.spark.api.python.PythonException: Traceback (most recent call last):&lt;BR /&gt;File "/databricks/spark/python/pyspark/serializers.py", line 192, in _read_with_length&lt;BR /&gt;return self.loads(obj)&lt;BR /&gt;^^^^^^^^^^^^^^^&lt;BR /&gt;File "/databricks/spark/python/pyspark/serializers.py", line 572, in loads&lt;BR /&gt;return cloudpickle.loads(obj, encoding=encoding)&lt;BR /&gt;^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^&lt;BR /&gt;ModuleNotFoundError: No module named 'dbruntime'&lt;/P&gt;</description>
      <pubDate>Tue, 08 Apr 2025 08:40:05 GMT</pubDate>
      <guid>https://community.databricks.com/t5/get-started-discussions/facing-issues-while-upgrading-dbr-version-from-9-1-lts-to-15-4/m-p/114800#M9342</guid>
      <dc:creator>Kuchnhi</dc:creator>
      <dc:date>2025-04-08T08:40:05Z</dc:date>
    </item>
    <item>
      <title>Re: Facing issues while upgrading DBR version from 9.1 LTS to 15.4 LTS</title>
      <link>https://community.databricks.com/t5/get-started-discussions/facing-issues-while-upgrading-dbr-version-from-9-1-lts-to-15-4/m-p/114852#M9343</link>
      <description>&lt;P&gt;I suggest troubleshooting the application dependencies and conflicts with libraries installed on 15.4. Seems to me there is a conflict.&lt;/P&gt;</description>
      <pubDate>Tue, 08 Apr 2025 18:25:08 GMT</pubDate>
      <guid>https://community.databricks.com/t5/get-started-discussions/facing-issues-while-upgrading-dbr-version-from-9-1-lts-to-15-4/m-p/114852#M9343</guid>
      <dc:creator>Louis_Frolio</dc:creator>
      <dc:date>2025-04-08T18:25:08Z</dc:date>
    </item>
    <item>
      <title>Re: Facing issues while upgrading DBR version from 9.1 LTS to 15.4 LTS</title>
      <link>https://community.databricks.com/t5/get-started-discussions/facing-issues-while-upgrading-dbr-version-from-9-1-lts-to-15-4/m-p/114901#M9344</link>
      <description>&lt;P&gt;Hi &lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/34815"&gt;@Louis_Frolio&lt;/a&gt;&amp;nbsp;Thanks for replying.&lt;/P&gt;&lt;P&gt;I tried by removing all predefined libraries from init script and after starting the cluster, installed required libraries manually. but the issue is still same. I am sharing complete error details below. could you all please look in detail and share some more insights. your Response will be much helpful to us.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;&lt;DIV class=""&gt;&lt;DIV&gt;Py4JJavaError: An error occurred while calling z:org.apache.spark.api.python.PythonRDD.runJob.&lt;/DIV&gt;&lt;DIV&gt;: org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 1.0 failed 4 times, most recent failure: Lost task 0.3 in stage 1.0 (TID 7) (10.141.5.10 executor 0): org.apache.spark.api.python.PythonException: Traceback (most recent call last):&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 192, in _read_with_length&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; return self.loads(obj)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;^^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 572, in loads&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; return cloudpickle.loads(obj, encoding=encoding)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;&lt;STRONG&gt;ModuleNotFoundError: No module named 'dbruntime'&lt;/STRONG&gt;&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;During handling of the above exception, another exception occurred:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;Traceback (most recent call last):&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/worker.py", line 1980, in main&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; process()&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/worker.py", line 1972, in process&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; serializer.dump_stream(out_iter, outfile)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 356, in dump_stream&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; vs = list(itertools.islice(iterator, batch))&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/core/rdd.py", line 2757, in takeUpToNumLeft&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; yield next(iterator)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; ^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 162, in load_stream&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; yield self._read_with_length(stream)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 196, in _read_with_length&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; raise SerializationError("Caused by " + traceback.format_exc())&lt;/DIV&gt;&lt;DIV&gt;pyspark.serializers.SerializationError: Caused by Traceback (most recent call last):&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 192, in _read_with_length&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; return self.loads(obj)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;^^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 572, in loads&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; return cloudpickle.loads(obj, encoding=encoding)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;ModuleNotFoundError: No module named 'dbruntime'&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.api.python.BasePythonRunner$ReaderIterator.handlePythonException(PythonRunner.scala:560)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.api.python.PythonRunner$$anon$3.read(PythonRunner.scala:968)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.api.python.PythonRunner$$anon$3.read(PythonRunner.scala:953)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.api.python.BasePythonRunner$ReaderIterator.hasNext(PythonRunner.scala:516)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:37)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.Iterator.foreach(Iterator.scala:943)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.Iterator.foreach$(Iterator.scala:943)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.InterruptibleIterator.foreach(InterruptibleIterator.scala:28)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.generic.Growable.$plus$plus$eq(Growable.scala:62)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.generic.Growable.$plus$plus$eq$(Growable.scala:53)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:105)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:49)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableOnce.to(TraversableOnce.scala:366)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableOnce.to$(TraversableOnce.scala:364)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.InterruptibleIterator.to(InterruptibleIterator.scala:28)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableOnce.toBuffer(TraversableOnce.scala:358)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableOnce.toBuffer$(TraversableOnce.scala:358)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.InterruptibleIterator.toBuffer(InterruptibleIterator.scala:28)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableOnce.toArray(TraversableOnce.scala:345)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableOnce.toArray$(TraversableOnce.scala:339)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.InterruptibleIterator.toArray(InterruptibleIterator.scala:28)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.api.python.PythonRDD$.$anonfun$collectPartitions$1(PythonRDD.scala:234)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.SparkContext.$anonfun$runJob$2(SparkContext.scala:3208)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.ResultTask.$anonfun$runTask$3(ResultTask.scala:82)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.ExecutorFrameProfiler$.record(ExecutorFrameProfiler.scala:110)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.ResultTask.$anonfun$runTask$1(ResultTask.scala:82)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.ExecutorFrameProfiler$.record(ExecutorFrameProfiler.scala:110)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:62)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:225)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.Task.doRunTask(Task.scala:199)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.Task.$anonfun$run$5(Task.scala:161)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.unity.EmptyHandle$.runWithAndClose(UCSHandle.scala:134)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.Task.$anonfun$run$1(Task.scala:155)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.ExecutorFrameProfiler$.record(ExecutorFrameProfiler.scala:110)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.Task.run(Task.scala:102)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$10(Executor.scala:1043)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:64)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:61)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:111)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:1046)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.ExecutorFrameProfiler$.record(ExecutorFrameProfiler.scala:110)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:933)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at java.lang.Thread.run(Thread.java:750)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;Driver stacktrace:&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.DAGScheduler.$anonfun$failJobAndIndependentStages$1(DAGScheduler.scala:4043)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.Option.getOrElse(Option.scala:189)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.DAGScheduler.failJobAndIndependentStages(DAGScheduler.scala:4041)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2(DAGScheduler.scala:3954)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.DAGScheduler.$anonfun$abortStage$2$adapted(DAGScheduler.scala:3941)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:3941)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1(DAGScheduler.scala:1767)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.DAGScheduler.$anonfun$handleTaskSetFailed$1$adapted(DAGScheduler.scala:1750)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.Option.foreach(Option.scala:407)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:1750)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:4302)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:4204)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:4190)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:55)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.DAGScheduler.$anonfun$runJob$1(DAGScheduler.scala:1413)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.FrameProfiler$.record(FrameProfiler.scala:94)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:1401)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.SparkContext.runJobInternal(SparkContext.scala:3168)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.SparkContext.runJob(SparkContext.scala:3149)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.SparkContext.runJob(SparkContext.scala:3189)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.SparkContext.runJob(SparkContext.scala:3208)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.api.python.PythonRDD$.collectPartitions(PythonRDD.scala:234)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.api.python.PythonRDD$.runJob(PythonRDD.scala:254)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.api.python.PythonRDD.runJob(PythonRDD.scala)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at java.lang.reflect.Method.invoke(Method.java:498)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:397)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at py4j.Gateway.invoke(Gateway.java:306)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at py4j.commands.CallCommand.execute(CallCommand.java:79)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at py4j.ClientServerConnection.waitForCommands(ClientServerConnection.java:199)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at py4j.ClientServerConnection.run(ClientServerConnection.java:119)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at java.lang.Thread.run(Thread.java:750)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;Caused by: org.apache.spark.api.python.PythonException: Traceback (most recent call last):&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 192, in _read_with_length&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; return self.loads(obj)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;^^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 572, in loads&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; return cloudpickle.loads(obj, encoding=encoding)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;ModuleNotFoundError: No module named 'dbruntime'&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;During handling of the above exception, another exception occurred:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;Traceback (most recent call last):&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/worker.py", line 1980, in main&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; process()&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/worker.py", line 1972, in process&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; serializer.dump_stream(out_iter, outfile)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 356, in dump_stream&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; vs = list(itertools.islice(iterator, batch))&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/core/rdd.py", line 2757, in takeUpToNumLeft&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; yield next(iterator)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; ^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 162, in load_stream&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; yield self._read_with_length(stream)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 196, in _read_with_length&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; raise SerializationError("Caused by " + traceback.format_exc())&lt;/DIV&gt;&lt;DIV&gt;pyspark.serializers.SerializationError: Caused by Traceback (most recent call last):&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 192, in _read_with_length&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; return self.loads(obj)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;^^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; File "/databricks/spark/python/pyspark/serializers.py", line 572, in loads&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; return cloudpickle.loads(obj, encoding=encoding)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^&lt;/DIV&gt;&lt;DIV&gt;ModuleNotFoundError: No module named 'dbruntime'&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.api.python.BasePythonRunner$ReaderIterator.handlePythonException(PythonRunner.scala:560)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.api.python.PythonRunner$$anon$3.read(PythonRunner.scala:968)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.api.python.PythonRunner$$anon$3.read(PythonRunner.scala:953)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.api.python.BasePythonRunner$ReaderIterator.hasNext(PythonRunner.scala:516)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.InterruptibleIterator.hasNext(InterruptibleIterator.scala:37)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.Iterator.foreach(Iterator.scala:943)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.Iterator.foreach$(Iterator.scala:943)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.InterruptibleIterator.foreach(InterruptibleIterator.scala:28)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.generic.Growable.$plus$plus$eq(Growable.scala:62)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.generic.Growable.$plus$plus$eq$(Growable.scala:53)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:105)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:49)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableOnce.to(TraversableOnce.scala:366)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableOnce.to$(TraversableOnce.scala:364)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.InterruptibleIterator.to(InterruptibleIterator.scala:28)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableOnce.toBuffer(TraversableOnce.scala:358)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableOnce.toBuffer$(TraversableOnce.scala:358)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.InterruptibleIterator.toBuffer(InterruptibleIterator.scala:28)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableOnce.toArray(TraversableOnce.scala:345)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableOnce.toArray$(TraversableOnce.scala:339)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.InterruptibleIterator.toArray(InterruptibleIterator.scala:28)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.api.python.PythonRDD$.$anonfun$collectPartitions$1(PythonRDD.scala:234)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.SparkContext.$anonfun$runJob$2(SparkContext.scala:3208)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.ResultTask.$anonfun$runTask$3(ResultTask.scala:82)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.ExecutorFrameProfiler$.record(ExecutorFrameProfiler.scala:110)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.ResultTask.$anonfun$runTask$1(ResultTask.scala:82)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.ExecutorFrameProfiler$.record(ExecutorFrameProfiler.scala:110)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:62)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.TaskContext.runTaskWithListeners(TaskContext.scala:225)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.Task.doRunTask(Task.scala:199)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.Task.$anonfun$run$5(Task.scala:161)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.unity.EmptyHandle$.runWithAndClose(UCSHandle.scala:134)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.Task.$anonfun$run$1(Task.scala:155)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.ExecutorFrameProfiler$.record(ExecutorFrameProfiler.scala:110)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.scheduler.Task.run(Task.scala:102)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$10(Executor.scala:1043)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally(SparkErrorUtils.scala:64)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.util.SparkErrorUtils.tryWithSafeFinally$(SparkErrorUtils.scala:61)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:111)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:1046)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.ExecutorFrameProfiler$.record(ExecutorFrameProfiler.scala:110)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:933)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;... 1 more&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;File &amp;lt;command-3609477148953765&amp;gt;, line 195&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 193 start = time.time()&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 194 start11 = time.time()&lt;/DIV&gt;&lt;DIV&gt;--&amp;gt; 195 print(common.readZipFromStorage(sc, dbutils, "2025", "04", "07", True).mapPartitions(processdata).count())&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 196 print(common.readZipFromStorage(sc, dbutils, "2025", "04", "08", False).mapPartitions(processdata).count())&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 197 print("End of main...")&lt;/DIV&gt;&lt;DIV&gt;File /local_disk0/spark-8459fa57-12da-4b04-9ebe-30bf615a5380/userFiles-d0b4be2b-cfc5-4eb0-9626-f647444f128c/Lib.zip/Lib/common/__init__.py:132, in readZipFromStorage(sc, dbutils, year, month, date, chk)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 129 rdd = sc.parallelize(final_list_of_files)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 131 # Create data frame&lt;/DIV&gt;&lt;DIV&gt;--&amp;gt; 132 df = rdd.toDF()&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 133 df = df.filter(df.name.endswith(".zip"))&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 135 # Extract only name column from df&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/pyspark/sql/session.py:133, in _monkey_patch_RDD.&amp;lt;locals&amp;gt;.toDF(self, schema, sampleRatio)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;98 &lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/16525"&gt;@No&lt;/a&gt;_type_check&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;99 def toDF(self, schema=None, sampleRatio=None):&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 100&amp;nbsp; &amp;nbsp; &amp;nbsp;"""&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 101&amp;nbsp; &amp;nbsp; &amp;nbsp;Converts current :class:`RDD` into a :class:`DataFrame`&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 102&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;(...)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 131&amp;nbsp; &amp;nbsp; &amp;nbsp;+---+&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 132&amp;nbsp; &amp;nbsp; &amp;nbsp;"""&lt;/DIV&gt;&lt;DIV&gt;--&amp;gt; 133&amp;nbsp; &amp;nbsp; &amp;nbsp;return sparkSession.createDataFrame(self, schema, sampleRatio)&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/pyspark/instrumentation_utils.py:47, in _wrap_function.&amp;lt;locals&amp;gt;.wrapper(*args, **kwargs)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;45 start = time.perf_counter()&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;46 try:&lt;/DIV&gt;&lt;DIV&gt;---&amp;gt; 47&amp;nbsp; &amp;nbsp; &amp;nbsp;res = func(*args, **kwargs)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;48&amp;nbsp; &amp;nbsp; &amp;nbsp;logger.log_success(&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;49&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;module_name, class_name, function_name, time.perf_counter() - start, signature&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;50&amp;nbsp; &amp;nbsp; &amp;nbsp;)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;51&amp;nbsp; &amp;nbsp; &amp;nbsp;return res&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/pyspark/sql/session.py:1610, in SparkSession.createDataFrame(self, data, schema, samplingRatio, verifySchema)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1605 if has_pandas and isinstance(data, pd.DataFrame):&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1606&amp;nbsp; &amp;nbsp; &amp;nbsp;# Create a DataFrame from pandas DataFrame.&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1607&amp;nbsp; &amp;nbsp; &amp;nbsp;return super(SparkSession, self).createDataFrame(&amp;nbsp; # type: ignore[call-overload]&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1608&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;data, schema, samplingRatio, verifySchema&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1609&amp;nbsp; &amp;nbsp; &amp;nbsp;)&lt;/DIV&gt;&lt;DIV&gt;-&amp;gt; 1610 return self._create_dataframe(&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1611&amp;nbsp; &amp;nbsp; &amp;nbsp;data, schema, samplingRatio, verifySchema&amp;nbsp; # type: ignore[arg-type]&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1612 )&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/pyspark/sql/session.py:1665, in SparkSession._create_dataframe(self, data, schema, samplingRatio, verifySchema)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1663 else:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1664&amp;nbsp; &amp;nbsp; &amp;nbsp;if not is_remote_only() and isinstance(data, RDD):&lt;/DIV&gt;&lt;DIV&gt;-&amp;gt; 1665&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;rdd, struct = self._createFromRDD(data.map(prepare), schema, samplingRatio)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1666&amp;nbsp; &amp;nbsp; &amp;nbsp;else:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1667&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;rdd, struct = self._createFromLocal(map(prepare, data), schema)&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/pyspark/sql/session.py:1168, in SparkSession._createFromRDD(self, rdd, schema, samplingRatio)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1164 """&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1165 Create an RDD for DataFrame from an existing RDD, returns the RDD and schema.&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1166 """&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1167 if schema is None or isinstance(schema, (list, tuple)):&lt;/DIV&gt;&lt;DIV&gt;-&amp;gt; 1168&amp;nbsp; &amp;nbsp; &amp;nbsp;struct = self._inferSchema(rdd, samplingRatio, names=schema)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1169&amp;nbsp; &amp;nbsp; &amp;nbsp;converter = _create_converter(struct)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1170&amp;nbsp; &amp;nbsp; &amp;nbsp;tupled_rdd = rdd.map(converter)&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/pyspark/sql/session.py:1108, in SparkSession._inferSchema(self, rdd, samplingRatio, names)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1087 def _inferSchema(&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1088&amp;nbsp; &amp;nbsp; &amp;nbsp;self,&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1089&amp;nbsp; &amp;nbsp; &amp;nbsp;rdd: "RDD[Any]",&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1090&amp;nbsp; &amp;nbsp; &amp;nbsp;samplingRatio: Optional[float] = None,&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1091&amp;nbsp; &amp;nbsp; &amp;nbsp;names: Optional[List[str]] = None,&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1092 ) -&amp;gt; StructType:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1093&amp;nbsp; &amp;nbsp; &amp;nbsp;"""&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1094&amp;nbsp; &amp;nbsp; &amp;nbsp;Infer schema from an RDD of Row, dict, or tuple.&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1095&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;(...)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1106&amp;nbsp; &amp;nbsp; &amp;nbsp;:class:`pyspark.sql.types.StructType`&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1107&amp;nbsp; &amp;nbsp; &amp;nbsp;"""&lt;/DIV&gt;&lt;DIV&gt;-&amp;gt; 1108&amp;nbsp; &amp;nbsp; &amp;nbsp;first = rdd.first()&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1109&amp;nbsp; &amp;nbsp; &amp;nbsp;if isinstance(first, Sized) and len(first) == 0:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1110&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;raise ValueError("The first row in RDD is empty, can not infer schema")&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/pyspark/instrumentation_utils.py:42, in _wrap_function.&amp;lt;locals&amp;gt;.wrapper(*args, **kwargs)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;38 @functools.wraps(func)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;39 def wrapper(*args: Any, **kwargs: Any) -&amp;gt; Any:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;40&amp;nbsp; &amp;nbsp; &amp;nbsp;if hasattr(_local, "logging") and _local.logging:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;41&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;# no need to log since this should be internal call.&lt;/DIV&gt;&lt;DIV&gt;---&amp;gt; 42&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;return func(*args, **kwargs)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;43&amp;nbsp; &amp;nbsp; &amp;nbsp;_local.logging = True&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;44&amp;nbsp; &amp;nbsp; &amp;nbsp;try:&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/pyspark/core/rdd.py:2796, in RDD.first(self)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2770 def first(self: "RDD[T]") -&amp;gt; T:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2771&amp;nbsp; &amp;nbsp; &amp;nbsp;"""&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2772&amp;nbsp; &amp;nbsp; &amp;nbsp;Return the first element in this RDD.&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2773&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;(...)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2794&amp;nbsp; &amp;nbsp; &amp;nbsp;ValueError: RDD is empty&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2795&amp;nbsp; &amp;nbsp; &amp;nbsp;"""&lt;/DIV&gt;&lt;DIV&gt;-&amp;gt; 2796&amp;nbsp; &amp;nbsp; &amp;nbsp;rs = self.take(1)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2797&amp;nbsp; &amp;nbsp; &amp;nbsp;if rs:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2798&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;return rs[0]&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/pyspark/instrumentation_utils.py:42, in _wrap_function.&amp;lt;locals&amp;gt;.wrapper(*args, **kwargs)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;38 @functools.wraps(func)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;39 def wrapper(*args: Any, **kwargs: Any) -&amp;gt; Any:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;40&amp;nbsp; &amp;nbsp; &amp;nbsp;if hasattr(_local, "logging") and _local.logging:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;41&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;# no need to log since this should be internal call.&lt;/DIV&gt;&lt;DIV&gt;---&amp;gt; 42&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;return func(*args, **kwargs)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;43&amp;nbsp; &amp;nbsp; &amp;nbsp;_local.logging = True&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp;44&amp;nbsp; &amp;nbsp; &amp;nbsp;try:&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/pyspark/core/rdd.py:2763, in RDD.take(self, num)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2760&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;taken += 1&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2762 p = range(partsScanned, min(partsScanned + numPartsToTry, totalParts))&lt;/DIV&gt;&lt;DIV&gt;-&amp;gt; 2763 res = self.context.runJob(self, takeUpToNumLeft, p)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2765 items += res&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2766 partsScanned += numPartsToTry&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/pyspark/core/context.py:2702, in SparkContext.runJob(self, rdd, partitionFunc, partitions, allowLocal)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2700&amp;nbsp; &amp;nbsp; &amp;nbsp;finally:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2701&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;os.remove(filename)&lt;/DIV&gt;&lt;DIV&gt;-&amp;gt; 2702 sock_info = self._jvm.PythonRDD.runJob(self._jsc.sc(), mappedRDD._jrdd, partitions)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;2703 return list(_load_from_socket(sock_info, mappedRDD._jrdd_deserializer))&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/lib/py4j-0.10.9.7-src.zip/py4j/java_gateway.py:1355, in JavaMember.__call__(self, *args)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1349 command = proto.CALL_COMMAND_NAME +\&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1350&amp;nbsp; &amp;nbsp; &amp;nbsp;self.command_header +\&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1351&amp;nbsp; &amp;nbsp; &amp;nbsp;args_command +\&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1352&amp;nbsp; &amp;nbsp; &amp;nbsp;proto.END_COMMAND_PART&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1354 answer = self.gateway_client.send_command(command)&lt;/DIV&gt;&lt;DIV&gt;-&amp;gt; 1355 return_value = get_return_value(&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1356&amp;nbsp; &amp;nbsp; &amp;nbsp;answer, self.gateway_client, self.target_id, self.name)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1358 for temp_arg in temp_args:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp;1359&amp;nbsp; &amp;nbsp; &amp;nbsp;if hasattr(temp_arg, "_detach"):&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/pyspark/errors/exceptions/captured.py:255, in capture_sql_exception.&amp;lt;locals&amp;gt;.deco(*a, **kw)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 252 from py4j.protocol import Py4JJavaError&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 254 try:&lt;/DIV&gt;&lt;DIV&gt;--&amp;gt; 255&amp;nbsp; &amp;nbsp; &amp;nbsp;return f(*a, **kw)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 256 except Py4JJavaError as e:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 257&amp;nbsp; &amp;nbsp; &amp;nbsp;converted = convert_exception(e.java_exception)&lt;/DIV&gt;&lt;DIV&gt;File /databricks/spark/python/lib/py4j-0.10.9.7-src.zip/py4j/protocol.py:326, in get_return_value(answer, gateway_client, target_id, name)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 324 value = OUTPUT_CONVERTER[type](answer[2:], gateway_client)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 325 if answer[1] == REFERENCE_TYPE:&lt;/DIV&gt;&lt;DIV&gt;--&amp;gt; 326&amp;nbsp; &amp;nbsp; &amp;nbsp;raise Py4JJavaError(&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 327&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;"An error occurred while calling {0}{1}{2}.\n".&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 328&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;format(target_id, ".", name), value)&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 329 else:&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 330&amp;nbsp; &amp;nbsp; &amp;nbsp;raise Py4JError(&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 331&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;"An error occurred while calling {0}{1}{2}. Trace:\n{3}\n".&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp; &amp;nbsp; 332&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;format(target_id, ".", name, value))&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;Here is my Cluster Configuration with DBR version 15.4 LTS:&lt;/DIV&gt;&lt;DIV&gt;{&lt;BR /&gt;"cluster_id": "0321-120122-h2mf81qg",&lt;BR /&gt;"creator_user_name": "rahul.790578@gmail.com",&lt;BR /&gt;"driver": {&lt;BR /&gt;"private_ip": "10.141.5.11",&lt;BR /&gt;"public_dns": "172.184.184.61",&lt;BR /&gt;"node_id": "9818cc182cfd4b40b2210d69cd531011",&lt;BR /&gt;"instance_id": "0c4d9ca030fd44a9b347acb8cef8d8f9",&lt;BR /&gt;"start_timestamp": 1744177619114,&lt;BR /&gt;"node_attributes": {&lt;BR /&gt;"is_spot": false&lt;BR /&gt;},&lt;BR /&gt;"host_private_ip": "10.141.4.10"&lt;BR /&gt;},&lt;BR /&gt;"executors": [&lt;BR /&gt;{&lt;BR /&gt;"private_ip": "10.141.5.12",&lt;BR /&gt;"public_dns": "13.93.231.102",&lt;BR /&gt;"node_id": "7f4fcc2b934e44199daa7f1d58c94e88",&lt;BR /&gt;"instance_id": "5e0a86855feb48528ef18e9c4e2cad83",&lt;BR /&gt;"start_timestamp": 1744177619187,&lt;BR /&gt;"node_attributes": {&lt;BR /&gt;"is_spot": true&lt;BR /&gt;},&lt;BR /&gt;"host_private_ip": "10.141.4.11"&lt;BR /&gt;},&lt;BR /&gt;{&lt;BR /&gt;"private_ip": "10.141.5.10",&lt;BR /&gt;"public_dns": "20.253.192.255",&lt;BR /&gt;"node_id": "b0f85ccd11b24d23919578216453a496",&lt;BR /&gt;"instance_id": "9b7b71b46ca946b8a87895e9cf7602b2",&lt;BR /&gt;"start_timestamp": 1744177619142,&lt;BR /&gt;"node_attributes": {&lt;BR /&gt;"is_spot": true&lt;BR /&gt;},&lt;BR /&gt;"host_private_ip": "10.141.4.12"&lt;BR /&gt;}&lt;BR /&gt;],&lt;BR /&gt;"spark_context_id": 791941095622325100,&lt;BR /&gt;"driver_healthy": true,&lt;BR /&gt;"jdbc_port": 10000,&lt;BR /&gt;"cluster_name": "Rahul Jaiswa's Cluster",&lt;BR /&gt;"spark_version": "15.4.x-scala2.12",&lt;BR /&gt;"spark_conf": {&lt;BR /&gt;"spark.sql.adaptive.coalescePartitions.initialPartitionNum": "8",&lt;BR /&gt;"spark.sql.files.maxPartitionBytes": "2469606",&lt;BR /&gt;"spark.sql.files.minPartitionNum": "8",&lt;BR /&gt;"spark.sql.adaptive.enabled": "true",&lt;BR /&gt;"spark.driver.maxResultSize": "5g",&lt;BR /&gt;"spark.databricks.delta.preview.enabled": "true",&lt;BR /&gt;"spark.sql.adaptive.advisoryPartitionSizeInBytes": "2469606",&lt;BR /&gt;"spark.driver.memory": "5g",&lt;BR /&gt;"spark.default.parallelism": "8",&lt;BR /&gt;"spark.sql.adaptive.coalescePartitions.enabled": "true",&lt;BR /&gt;"spark.driver.cores": "8",&lt;BR /&gt;"spark.executor.memory": "4g",&lt;BR /&gt;"spark.sql.files.openCostInBytes": "134217",&lt;BR /&gt;"spark.sql.adaptive.coalescePartitions.minPartitionNum": "8",&lt;BR /&gt;"spark.sql.shuffle.partitions": "8"&lt;BR /&gt;},&lt;BR /&gt;"azure_attributes": {&lt;BR /&gt;"first_on_demand": 1,&lt;BR /&gt;"availability": "SPOT_WITH_FALLBACK_AZURE",&lt;BR /&gt;"spot_bid_max_price": -1&lt;BR /&gt;},&lt;BR /&gt;"node_type_id": "Standard_E8s_v3",&lt;BR /&gt;"driver_node_type_id": "Standard_E8s_v3",&lt;BR /&gt;"cluster_log_conf": {&lt;BR /&gt;"dbfs": {&lt;BR /&gt;"destination": "dbfs:/FileStore/manual-logs"&lt;BR /&gt;}&lt;BR /&gt;},&lt;BR /&gt;"spark_env_vars": {&lt;BR /&gt;"PYSPARK_PYTHON": "/databricks/python3/bin/python3"&lt;BR /&gt;},&lt;BR /&gt;"autotermination_minutes": 60,&lt;BR /&gt;"enable_elastic_disk": true,&lt;BR /&gt;"disk_spec": {},&lt;BR /&gt;"cluster_source": "UI",&lt;BR /&gt;"init_scripts": [&lt;BR /&gt;{&lt;BR /&gt;"workspace": {&lt;BR /&gt;"destination": "/startup1/pyodbc-install.sh"&lt;BR /&gt;}&lt;BR /&gt;}&lt;BR /&gt;],&lt;BR /&gt;"single_user_name": "rahul.790578@gmail.com",&lt;BR /&gt;"enable_local_disk_encryption": false,&lt;BR /&gt;"instance_source": {&lt;BR /&gt;"node_type_id": "Standard_E8s_v3"&lt;BR /&gt;},&lt;BR /&gt;"driver_instance_source": {&lt;BR /&gt;"node_type_id": "Standard_E8s_v3"&lt;BR /&gt;},&lt;BR /&gt;"data_security_mode": "LEGACY_SINGLE_USER_STANDARD",&lt;BR /&gt;"runtime_engine": "PHOTON",&lt;BR /&gt;"effective_spark_version": "15.4.x-photon-scala2.12",&lt;BR /&gt;"assigned_principal": "user:rahul.790578@gmail.com",&lt;BR /&gt;"release_version": "15.4.12",&lt;BR /&gt;"state": "RUNNING",&lt;BR /&gt;"state_message": "",&lt;BR /&gt;"start_time": 1742558482247,&lt;BR /&gt;"last_state_loss_time": 1744177737748,&lt;BR /&gt;"last_activity_time": 1744177679642,&lt;BR /&gt;"last_restarted_time": 1744177737824,&lt;BR /&gt;"autoscale": {&lt;BR /&gt;"min_workers": 2,&lt;BR /&gt;"max_workers": 3,&lt;BR /&gt;"target_workers": 2&lt;BR /&gt;},&lt;BR /&gt;"cluster_memory_mb": 196608,&lt;BR /&gt;"cluster_cores": 24,&lt;BR /&gt;"default_tags": {&lt;BR /&gt;"Vendor": "Databricks",&lt;BR /&gt;"Creator": "rahul.790578@gmail.com",&lt;BR /&gt;"ClusterName": "Rahul Jaiswa's Cluster",&lt;BR /&gt;"ClusterId": "0321-120122-h2mf81qg",&lt;BR /&gt;"Environmet": "dvt"&lt;BR /&gt;},&lt;BR /&gt;"cluster_log_status": {&lt;BR /&gt;"last_attempted": 1744178955397&lt;BR /&gt;},&lt;BR /&gt;"init_scripts_safe_mode": false,&lt;BR /&gt;"spec": {&lt;BR /&gt;"cluster_name": "Rahul Jaiswa's Cluster",&lt;BR /&gt;"&lt;STRONG&gt;spark_version": "15.4.x-scala2.12&lt;/STRONG&gt;",&lt;BR /&gt;"spark_conf": {&lt;BR /&gt;"spark.databricks.delta.preview.enabled": "true",&lt;BR /&gt;"spark.sql.files.minPartitionNum": "8",&lt;BR /&gt;"spark.sql.files.openCostInBytes": "134217",&lt;BR /&gt;"spark.sql.adaptive.coalescePartitions.initialPartitionNum": "8",&lt;BR /&gt;"spark.sql.adaptive.coalescePartitions.enabled": "true",&lt;BR /&gt;"spark.executor.memory": "4g",&lt;BR /&gt;"spark.driver.maxResultSize": "5g",&lt;BR /&gt;"spark.sql.shuffle.partitions": "8",&lt;BR /&gt;"spark.driver.memory": "5g",&lt;BR /&gt;"spark.sql.adaptive.coalescePartitions.minPartitionNum": "8",&lt;BR /&gt;"spark.sql.adaptive.enabled": "true",&lt;BR /&gt;"spark.sql.adaptive.advisoryPartitionSizeInBytes": "2469606",&lt;BR /&gt;"spark.sql.files.maxPartitionBytes": "2469606",&lt;BR /&gt;"spark.default.parallelism": "8",&lt;BR /&gt;"spark.driver.cores": "8"&lt;BR /&gt;},&lt;BR /&gt;"azure_attributes": {&lt;BR /&gt;"first_on_demand": 1,&lt;BR /&gt;"availability": "SPOT_WITH_FALLBACK_AZURE",&lt;BR /&gt;"spot_bid_max_price": -1&lt;BR /&gt;},&lt;BR /&gt;"node_type_id": "Standard_E8s_v3",&lt;BR /&gt;"driver_node_type_id": "Standard_E8s_v3",&lt;BR /&gt;"cluster_log_conf": {&lt;BR /&gt;"dbfs": {&lt;BR /&gt;"destination": "dbfs:/FileStore/manual-logs"&lt;BR /&gt;}&lt;BR /&gt;},&lt;BR /&gt;"spark_env_vars": {&lt;BR /&gt;"PYSPARK_PYTHON": "/databricks/python3/bin/python3"&lt;BR /&gt;},&lt;BR /&gt;"autotermination_minutes": 60,&lt;BR /&gt;"enable_elastic_disk": true,&lt;BR /&gt;"init_scripts": [&lt;BR /&gt;{&lt;BR /&gt;"workspace": {&lt;BR /&gt;"destination": "/startup1/pyodbc-install.sh"&lt;BR /&gt;}&lt;BR /&gt;}&lt;BR /&gt;],&lt;BR /&gt;"single_user_name": "rahul.790578@gmail.com",&lt;BR /&gt;"enable_local_disk_encryption": false,&lt;BR /&gt;"data_security_mode": "LEGACY_SINGLE_USER_STANDARD",&lt;BR /&gt;"runtime_engine": "PHOTON",&lt;BR /&gt;"assigned_principal": "user:rrahul.790578@gmail.com",&lt;BR /&gt;"autoscale": {&lt;BR /&gt;"min_workers": 2,&lt;BR /&gt;"max_workers": 3&lt;BR /&gt;},&lt;BR /&gt;"apply_policy_default_values": false&lt;BR /&gt;}&lt;BR /&gt;}&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;</description>
      <pubDate>Wed, 09 Apr 2025 06:22:51 GMT</pubDate>
      <guid>https://community.databricks.com/t5/get-started-discussions/facing-issues-while-upgrading-dbr-version-from-9-1-lts-to-15-4/m-p/114901#M9344</guid>
      <dc:creator>Kuchnhi</dc:creator>
      <dc:date>2025-04-09T06:22:51Z</dc:date>
    </item>
    <item>
      <title>Re: Facing issues while upgrading DBR version from 9.1 LTS to 15.4 LTS</title>
      <link>https://community.databricks.com/t5/get-started-discussions/facing-issues-while-upgrading-dbr-version-from-9-1-lts-to-15-4/m-p/114952#M9345</link>
      <description>&lt;P&gt;What type of cluster are you spinning up? Is it "serverless or classic", "jobs cluster", "sql warehouse"?&amp;nbsp;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 09 Apr 2025 11:39:38 GMT</pubDate>
      <guid>https://community.databricks.com/t5/get-started-discussions/facing-issues-while-upgrading-dbr-version-from-9-1-lts-to-15-4/m-p/114952#M9345</guid>
      <dc:creator>Louis_Frolio</dc:creator>
      <dc:date>2025-04-09T11:39:38Z</dc:date>
    </item>
    <item>
      <title>Re: Facing issues while upgrading DBR version from 9.1 LTS to 15.4 LTS</title>
      <link>https://community.databricks.com/t5/get-started-discussions/facing-issues-while-upgrading-dbr-version-from-9-1-lts-to-15-4/m-p/114961#M9346</link>
      <description>&lt;P&gt;Dear&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/34815"&gt;@Louis_Frolio&lt;/a&gt;&amp;nbsp; it's a Jobs cluster... in last post I have shared the cluster configuration as well in bottom.&lt;/P&gt;</description>
      <pubDate>Wed, 09 Apr 2025 12:19:47 GMT</pubDate>
      <guid>https://community.databricks.com/t5/get-started-discussions/facing-issues-while-upgrading-dbr-version-from-9-1-lts-to-15-4/m-p/114961#M9346</guid>
      <dc:creator>Kuchnhi</dc:creator>
      <dc:date>2025-04-09T12:19:47Z</dc:date>
    </item>
    <item>
      <title>Re: Facing issues while upgrading DBR version from 9.1 LTS to 15.4 LTS</title>
      <link>https://community.databricks.com/t5/get-started-discussions/facing-issues-while-upgrading-dbr-version-from-9-1-lts-to-15-4/m-p/114995#M9347</link>
      <description>&lt;P&gt;Is it serverless or classic jobs compute?&amp;nbsp; Also, if it were me I would try with a different runtime (LTS of course) to see if the problem exists there (assuming you are not using serverless). e.g 14.3 LTS, 13.3 LTS, 12.3 LTS, etc.&amp;nbsp; It might help you understand where the conflict is happening.&amp;nbsp; Cheers, Lou.&lt;/P&gt;</description>
      <pubDate>Wed, 09 Apr 2025 14:31:09 GMT</pubDate>
      <guid>https://community.databricks.com/t5/get-started-discussions/facing-issues-while-upgrading-dbr-version-from-9-1-lts-to-15-4/m-p/114995#M9347</guid>
      <dc:creator>Louis_Frolio</dc:creator>
      <dc:date>2025-04-09T14:31:09Z</dc:date>
    </item>
    <item>
      <title>Re: Facing issues while upgrading DBR version from 9.1 LTS to 15.4 LTS</title>
      <link>https://community.databricks.com/t5/get-started-discussions/facing-issues-while-upgrading-dbr-version-from-9-1-lts-to-15-4/m-p/115045#M9348</link>
      <description>&lt;P&gt;Dear&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/34815"&gt;@Louis_Frolio&lt;/a&gt;&amp;nbsp;compute which is running is classic jobs compute. I had already tried with lower DBR versions (&lt;SPAN&gt;14.3 LTS, 13.3 LTS, 12.3 LTS)&lt;/SPAN&gt; but it throws same error which i have shared before, only it is working with 9.1 &amp;amp; 10.4 LTS but support is already expired. so we wanted to upgrade it to latest version. we want to know, if you can tell the possible reasons of this error&lt;/P&gt;</description>
      <pubDate>Wed, 09 Apr 2025 17:57:17 GMT</pubDate>
      <guid>https://community.databricks.com/t5/get-started-discussions/facing-issues-while-upgrading-dbr-version-from-9-1-lts-to-15-4/m-p/115045#M9348</guid>
      <dc:creator>Kuchnhi</dc:creator>
      <dc:date>2025-04-09T17:57:17Z</dc:date>
    </item>
    <item>
      <title>Re: Facing issues while upgrading DBR version from 9.1 LTS to 15.4 LTS</title>
      <link>https://community.databricks.com/t5/get-started-discussions/facing-issues-while-upgrading-dbr-version-from-9-1-lts-to-15-4/m-p/115051#M9349</link>
      <description>&lt;P&gt;Thank you for further explaining your issue. While I appreciate the opportunity to engage in this discussion, I must admit that based on what you have shared that the complexity of the issue you’re describing is beyond my expertise. It seems like this might be better addressed by Databricks Support, as they are equipped to handle advanced technical challenges and provide tailored solutions. The forum is a great place for community-driven discussions and sharing knowledge, but it may not always serve as a replacement for official support channels. I recommend reaching out to Databricks Support directly for assistance—they’ll be able to provide the specialized help you need.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Cheers, Louis.&lt;/P&gt;</description>
      <pubDate>Wed, 09 Apr 2025 18:04:51 GMT</pubDate>
      <guid>https://community.databricks.com/t5/get-started-discussions/facing-issues-while-upgrading-dbr-version-from-9-1-lts-to-15-4/m-p/115051#M9349</guid>
      <dc:creator>Louis_Frolio</dc:creator>
      <dc:date>2025-04-09T18:04:51Z</dc:date>
    </item>
    <item>
      <title>Re: Facing issues while upgrading DBR version from 9.1 LTS to 15.4 LTS</title>
      <link>https://community.databricks.com/t5/get-started-discussions/facing-issues-while-upgrading-dbr-version-from-9-1-lts-to-15-4/m-p/115162#M9350</link>
      <description>&lt;P&gt;I just thought of one more question.&amp;nbsp; Are you still trying to run your init scripts out of DBFS?&amp;nbsp; It could be causing your issues.&amp;nbsp; Here are some notes on the end of life for DBFS init scripts, hope this helps.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;DIV class="paragraph"&gt;As of March 31, 2024, the End-of-Life (EoL) for running init scripts from the Databricks File System (DBFS) has been implemented for all customers that did not request an extension. Databricks no longer supports init scripts stored on DBFS for the latest Databricks Runtime (DBR) versions, and users are required to migrate their scripts to alternative storage options.&lt;/DIV&gt;
&lt;DIV class="paragraph"&gt;Here are the supported alternatives for storing init scripts based on your Databricks Runtime version and setup:&lt;/DIV&gt;
&lt;OL start="1"&gt;
&lt;LI&gt;
&lt;DIV class="paragraph"&gt;&lt;STRONG&gt;Unity Catalog Volumes&lt;/STRONG&gt;:
&lt;UL&gt;
&lt;LI&gt;Recommended if using &lt;STRONG&gt;Databricks Runtime 13.3 LTS or above&lt;/STRONG&gt; with Unity Catalog-enabled clusters.&lt;/LI&gt;
&lt;LI&gt;Use this for secure and governed storage when Unity Catalog is available.&lt;/LI&gt;
&lt;/UL&gt;
&lt;/DIV&gt;
&lt;/LI&gt;
&lt;LI&gt;
&lt;DIV class="paragraph"&gt;&lt;STRONG&gt;Workspace Files&lt;/STRONG&gt;:
&lt;UL&gt;
&lt;LI&gt;Suitable for &lt;STRONG&gt;Databricks Runtime 11.3 LTS and above&lt;/STRONG&gt; without Unity Catalog.&lt;/LI&gt;
&lt;LI&gt;Workspace files provide a Databricks-native storage option with file-level permissions.&lt;/LI&gt;
&lt;/UL&gt;
&lt;/DIV&gt;
&lt;/LI&gt;
&lt;LI&gt;
&lt;DIV class="paragraph"&gt;&lt;STRONG&gt;Cloud Object Storage&lt;/STRONG&gt;:
&lt;UL&gt;
&lt;LI&gt;For &lt;STRONG&gt;Databricks Runtime 10.4 LTS or below&lt;/STRONG&gt;, or for more complex use cases referencing additional files.&lt;/LI&gt;
&lt;LI&gt;Examples: Amazon S3, Azure Data Lake Storage (ADLS), or Google Cloud Storage (GCS).&lt;/LI&gt;
&lt;/UL&gt;
&lt;/DIV&gt;
&lt;/LI&gt;
&lt;/OL&gt;
&lt;DIV class="paragraph"&gt;&lt;STRONG&gt;Summary of Supported Runtimes for Migration and Storage&lt;/STRONG&gt;: - &lt;STRONG&gt;Databricks Runtime 13.3 LTS and above (with Unity Catalog)&lt;/STRONG&gt;: Use Unity Catalog Volumes. - &lt;STRONG&gt;Databricks Runtime 11.3 LTS and above (without Unity Catalog)&lt;/STRONG&gt;: Use Workspace Files. - &lt;STRONG&gt;Databricks Runtime 10.4 LTS and below&lt;/STRONG&gt;: Use external cloud object storage.&lt;/DIV&gt;
&lt;DIV class="paragraph"&gt;It is highly recommended that users on unsupported runtimes upgrade to a supported DBR as soon as possible to avoid potential issues. If needed, migration guidance is available in the public Databricks documentation, which details steps for transitioning scripts to these secure and supported storage formats.&lt;/DIV&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 10 Apr 2025 14:13:48 GMT</pubDate>
      <guid>https://community.databricks.com/t5/get-started-discussions/facing-issues-while-upgrading-dbr-version-from-9-1-lts-to-15-4/m-p/115162#M9350</guid>
      <dc:creator>Louis_Frolio</dc:creator>
      <dc:date>2025-04-10T14:13:48Z</dc:date>
    </item>
    <item>
      <title>Re: Facing issues while upgrading DBR version from 9.1 LTS to 15.4 LTS</title>
      <link>https://community.databricks.com/t5/get-started-discussions/facing-issues-while-upgrading-dbr-version-from-9-1-lts-to-15-4/m-p/115233#M9351</link>
      <description>&lt;P&gt;Currently I am using workspace location to store init scripts. I will try to use unity catalog for init script and install the packages with higher DBR version..&lt;/P&gt;&lt;P&gt;Thankyou for your Suggestion&lt;/P&gt;</description>
      <pubDate>Fri, 11 Apr 2025 03:51:23 GMT</pubDate>
      <guid>https://community.databricks.com/t5/get-started-discussions/facing-issues-while-upgrading-dbr-version-from-9-1-lts-to-15-4/m-p/115233#M9351</guid>
      <dc:creator>Kuchnhi</dc:creator>
      <dc:date>2025-04-11T03:51:23Z</dc:date>
    </item>
    <item>
      <title>Re: Facing issues while upgrading DBR version from 9.1 LTS to 15.4 LTS</title>
      <link>https://community.databricks.com/t5/get-started-discussions/facing-issues-while-upgrading-dbr-version-from-9-1-lts-to-15-4/m-p/115479#M9352</link>
      <description>&lt;P class=""&gt;Hey, this error usually happens when the cluster isn't fully ready before your application starts running&lt;A href="https://castleappp.org/castle-app-old-version" target="_self"&gt;.&lt;/A&gt; Since your init script takes about 30 minutes, it’s likely that your job starts before all dependencies are properly installed. The ModuleNotFoundError: No module named 'dbruntime' often indicates that the environment isn't fully initialized yet.&lt;/P&gt;&lt;P class=""&gt;A good approach is to delay job execution until the init script completes or use a cluster startup script to validate readiness. Also, double-check that your init script is installing everything needed and exiting without errors.&lt;/P&gt;&lt;P class=""&gt;Hope this helps! Let me know how it goes.&lt;/P&gt;</description>
      <pubDate>Tue, 15 Apr 2025 08:16:05 GMT</pubDate>
      <guid>https://community.databricks.com/t5/get-started-discussions/facing-issues-while-upgrading-dbr-version-from-9-1-lts-to-15-4/m-p/115479#M9352</guid>
      <dc:creator>SmithPoll</dc:creator>
      <dc:date>2025-04-15T08:16:05Z</dc:date>
    </item>
    <item>
      <title>Re: Facing issues while upgrading DBR version from 9.1 LTS to 15.4 LTS</title>
      <link>https://community.databricks.com/t5/get-started-discussions/facing-issues-while-upgrading-dbr-version-from-9-1-lts-to-15-4/m-p/120261#M10096</link>
      <description>&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;DIV class=""&gt;&lt;P&gt;ust to add, you might also want to check the cluster logs (driver and init script logs) for any hidden errors or timeouts during startup. Sometimes dependencies silently fail to install&lt;A title="," href="https://supremetaxis.co.uk/" target="_self"&gt;,&lt;/A&gt;even if the cluster appears to be running. If possible, try breaking your init script into smaller steps and logging each one to identify where it might be hanging or failing.&lt;/P&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;DIV class=""&gt;&amp;nbsp;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;/DIV&gt;</description>
      <pubDate>Mon, 26 May 2025 18:56:02 GMT</pubDate>
      <guid>https://community.databricks.com/t5/get-started-discussions/facing-issues-while-upgrading-dbr-version-from-9-1-lts-to-15-4/m-p/120261#M10096</guid>
      <dc:creator>SmithPoll</dc:creator>
      <dc:date>2025-05-26T18:56:02Z</dc:date>
    </item>
  </channel>
</rss>

