<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: (java.util.concurrent.ExecutionException) Boxed Error in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/java-util-concurrent-executionexception-boxed-error/m-p/98323#M39689</link>
    <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/93173"&gt;@Nathant93&lt;/a&gt;&amp;nbsp;,&lt;/P&gt;
&lt;P&gt;Does it come with a "Caused by" in the error stacktrace? If there isn't any in the Spark logs, perhaps you can provide a reproducer code leading to this exception. Stacktrace, DBR version and repro code would help.&lt;/P&gt;
&lt;P&gt;The&amp;nbsp;&lt;SPAN&gt;(java.util.concurrent.ExecutionException) Boxed Error, itself, is just a top-level error, so it could be many things.&lt;/SPAN&gt;&lt;/P&gt;</description>
    <pubDate>Mon, 11 Nov 2024 11:58:10 GMT</pubDate>
    <dc:creator>VZLA</dc:creator>
    <dc:date>2024-11-11T11:58:10Z</dc:date>
    <item>
      <title>(java.util.concurrent.ExecutionException) Boxed Error</title>
      <link>https://community.databricks.com/t5/data-engineering/java-util-concurrent-executionexception-boxed-error/m-p/98312#M39685</link>
      <description>&lt;P&gt;Has anyone ever come across the error above?&lt;/P&gt;&lt;P&gt;I am trying to get two tables from unity catalog and join them, the join is fairly complex as it is imitating a where not exists top 1 sql query.&lt;/P&gt;</description>
      <pubDate>Mon, 11 Nov 2024 09:38:27 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/java-util-concurrent-executionexception-boxed-error/m-p/98312#M39685</guid>
      <dc:creator>Nathant93</dc:creator>
      <dc:date>2024-11-11T09:38:27Z</dc:date>
    </item>
    <item>
      <title>Re: (java.util.concurrent.ExecutionException) Boxed Error</title>
      <link>https://community.databricks.com/t5/data-engineering/java-util-concurrent-executionexception-boxed-error/m-p/98323#M39689</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/93173"&gt;@Nathant93&lt;/a&gt;&amp;nbsp;,&lt;/P&gt;
&lt;P&gt;Does it come with a "Caused by" in the error stacktrace? If there isn't any in the Spark logs, perhaps you can provide a reproducer code leading to this exception. Stacktrace, DBR version and repro code would help.&lt;/P&gt;
&lt;P&gt;The&amp;nbsp;&lt;SPAN&gt;(java.util.concurrent.ExecutionException) Boxed Error, itself, is just a top-level error, so it could be many things.&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Mon, 11 Nov 2024 11:58:10 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/java-util-concurrent-executionexception-boxed-error/m-p/98323#M39689</guid>
      <dc:creator>VZLA</dc:creator>
      <dc:date>2024-11-11T11:58:10Z</dc:date>
    </item>
    <item>
      <title>Re: (java.util.concurrent.ExecutionException) Boxed Error</title>
      <link>https://community.databricks.com/t5/data-engineering/java-util-concurrent-executionexception-boxed-error/m-p/114951#M44999</link>
      <description>&lt;P&gt;Hello&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/34618"&gt;@VZLA&lt;/a&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Recently, I am getting the exact same error.&lt;/P&gt;&lt;P&gt;It has a caused by as below -&lt;BR /&gt;```&lt;BR /&gt;&lt;SPAN&gt;Caused by: kafkashaded.org.apache.kafka.common.errors.UnknownTopicOrPartitionException: This server does not host this topic-partition.&lt;/SPAN&gt;&lt;BR /&gt;```&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Stacktrace -&lt;/STRONG&gt;&lt;/P&gt;&lt;DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;ERROR&lt;/SPAN&gt;&lt;SPAN&gt;: Some streams terminated before this command could finish!&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;java.util.concurrent.ExecutionException&lt;/SPAN&gt;&lt;SPAN&gt;: &lt;/SPAN&gt;&lt;SPAN&gt;kafkashaded.org.apache.kafka.common.errors.UnknownTopicOrPartitionException&lt;/SPAN&gt;&lt;SPAN&gt;: This server does not host this topic-partition.&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at java.util.concurrent.CompletableFuture.reportGet(CompletableFuture.java:357)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at java.util.concurrent.CompletableFuture.get(CompletableFuture.java:1908)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at kafkashaded.org.apache.kafka.common.internals.KafkaFutureImpl.get(KafkaFutureImpl.java:165)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.sql.kafka010.ConsumerStrategy.retrieveAllPartitions(ConsumerStrategy.scala:67)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.sql.kafka010.ConsumerStrategy.retrieveAllPartitions$(ConsumerStrategy.scala:66)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.sql.kafka010.SubscribeStrategy.retrieveAllPartitions(ConsumerStrategy.scala:103)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.sql.kafka010.SubscribeStrategy.assignedTopicPartitions(ConsumerStrategy.scala:114)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.sql.kafka010.KafkaOffsetReaderAdmin.$anonfun$partitionsAssignedToAdmin$1(KafkaOffsetReaderAdmin.scala:525)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.sql.kafka010.KafkaOffsetReaderAdmin.withRetries(KafkaOffsetReaderAdmin.scala:544)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.sql.kafka010.KafkaOffsetReaderAdmin.partitionsAssignedToAdmin(KafkaOffsetReaderAdmin.scala:524)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.sql.kafka010.KafkaOffsetReaderAdmin.fetchEarliestOffsets(KafkaOffsetReaderAdmin.scala:299)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.sql.kafka010.KafkaMicroBatchStream.$anonfun$getOrCreateInitialPartitionOffsets$1(KafkaMicroBatchStream.scala:416)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.Option.getOrElse(Option.scala:189)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.sql.kafka010.KafkaMicroBatchStream.getOrCreateInitialPartitionOffsets(KafkaMicroBatchStream.scala:413)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.sql.kafka010.KafkaMicroBatchStream.prepareForTriggerAvailableNow(KafkaMicroBatchStream.scala:533)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.sql.execution.streaming.MicroBatchExecution.$anonfun$initSources$4(MicroBatchExecution.scala:401)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableLike.$anonfun$map$1(TraversableLike.scala:286)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.mutable.ResizableArray.foreach(ResizableArray.scala:62)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.mutable.ResizableArray.foreach$(ResizableArray.scala:55)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:49)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableLike.map(TraversableLike.scala:286)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.TraversableLike.map$(TraversableLike.scala:279)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.collection.AbstractTraversable.map(Traversable.scala:108)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.sql.execution.streaming.MicroBatchExecution.initSources(MicroBatchExecution.scala:400)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.sql.execution.streaming.StreamExecution.$anonfun$runStream$2(StreamExecution.scala:432)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:1210)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.sql.execution.streaming.StreamExecution.$anonfun$runStream$1(StreamExecution.scala:401)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.logging.AttributionContextTracing.$anonfun$withAttributionContext$1(AttributionContextTracing.scala:48)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.logging.AttributionContext$.$anonfun$withValue$1(AttributionContext.scala:276)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.util.DynamicVariable.withValue(DynamicVariable.scala:62)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.logging.AttributionContext$.withValue(AttributionContext.scala:272)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.logging.AttributionContextTracing.withAttributionContext(AttributionContextTracing.scala:46)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.logging.AttributionContextTracing.withAttributionContext$(AttributionContextTracing.scala:43)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.PublicDBLogging.withAttributionContext(DatabricksSparkUsageLogger.scala:30)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.logging.AttributionContextTracing.withAttributionTags(AttributionContextTracing.scala:95)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.logging.AttributionContextTracing.withAttributionTags$(AttributionContextTracing.scala:76)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.PublicDBLogging.withAttributionTags(DatabricksSparkUsageLogger.scala:30)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.PublicDBLogging.withAttributionTags0(DatabricksSparkUsageLogger.scala:91)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.DatabricksSparkUsageLogger.withAttributionTags(DatabricksSparkUsageLogger.scala:195)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.UsageLogging.$anonfun$withAttributionTags$1(UsageLogger.scala:617)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.UsageLogging$.withAttributionTags(UsageLogger.scala:729)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.UsageLogging$.withAttributionTags(UsageLogger.scala:738)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.UsageLogging.withAttributionTags(UsageLogger.scala:617)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.spark.util.UsageLogging.withAttributionTags$(UsageLogger.scala:615)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.sql.execution.streaming.StreamExecution.withAttributionTags(StreamExecution.scala:86)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.sql.execution.streaming.StreamExecution.org$apache$spark$sql$execution$streaming$StreamExecution$$runStream(StreamExecution.scala:381)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.sql.execution.streaming.StreamExecution$$anon$1.$anonfun$run$3(StreamExecution.scala:284)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.JobArtifactSet$.withActiveJobArtifactState(JobArtifactSet.scala:97)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.sql.execution.streaming.StreamExecution$$anon$1.$anonfun$run$2(StreamExecution.scala:284)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.unity.UCSEphemeralState$Handle.runWith(UCSEphemeralState.scala:51)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.unity.HandleImpl.runWith(UCSHandle.scala:104)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.unity.HandleImpl.$anonfun$runWithAndClose$1(UCSHandle.scala:109)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at scala.util.Using$.resource(Using.scala:269)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at com.databricks.unity.HandleImpl.runWithAndClose(UCSHandle.scala:108)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;at org.apache.spark.sql.execution.streaming.StreamExecution$$anon$1.run(StreamExecution.scala:283)&lt;/SPAN&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;SPAN&gt;Caused by: &lt;/SPAN&gt;&lt;SPAN&gt;kafkashaded.org.apache.kafka.common.errors.UnknownTopicOrPartitionException&lt;/SPAN&gt;&lt;SPAN&gt;: This server does not host this topic-partition.&lt;/SPAN&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;P&gt;This is a valid error and I want to catch it and take some external action based on it. I already have a try-catch surrounding this place, but it is still causing the job to fail.&lt;BR /&gt;Also, I cannot find the origin of this error from the code I have written.&lt;/P&gt;</description>
      <pubDate>Wed, 09 Apr 2025 11:33:36 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/java-util-concurrent-executionexception-boxed-error/m-p/114951#M44999</guid>
      <dc:creator>pk13</dc:creator>
      <dc:date>2025-04-09T11:33:36Z</dc:date>
    </item>
  </channel>
</rss>

