<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: I am getting this error:
com.databricks.backend.common.rpc.DatabricksExceptions$SQLExecutionExce in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/i-am-getting-this-error-com-databricks-backend-common-rpc/m-p/98302#M39678</link>
    <description>&lt;P&gt;i am facing the issue before i trying to execute a code error message:&lt;SPAN&gt;com.databricks.rpc.UnknownRemoteException: Remote exception occurred:&lt;/SPAN&gt;&lt;/P&gt;</description>
    <pubDate>Mon, 11 Nov 2024 01:56:45 GMT</pubDate>
    <dc:creator>aparna123</dc:creator>
    <dc:date>2024-11-11T01:56:45Z</dc:date>
    <item>
      <title>I am getting this error:
com.databricks.backend.common.rpc.DatabricksExceptions$SQLExecutionException: com.databricks.rpc.UnknownRemoteException: Remote exception occurred:</title>
      <link>https://community.databricks.com/t5/data-engineering/i-am-getting-this-error-com-databricks-backend-common-rpc/m-p/24141#M16764</link>
      <description>&lt;P&gt;I am teaching a class for BYU Idaho and every table in every database has been imploded for my class. We keep getting this error:&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;com.databricks.backend.common.rpc.DatabricksExceptions$SQLExecutionException: com.databricks.rpc.UnknownRemoteException: Remote exception occurred:&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Databricks Premium on AWS&lt;/P&gt;</description>
      <pubDate>Wed, 02 Nov 2022 16:54:22 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/i-am-getting-this-error-com-databricks-backend-common-rpc/m-p/24141#M16764</guid>
      <dc:creator>TeachingWithDat</dc:creator>
      <dc:date>2022-11-02T16:54:22Z</dc:date>
    </item>
    <item>
      <title>Re: I am getting this error:
com.databricks.backend.common.rpc.DatabricksExceptions$SQLExecutionException: com.databricks.rpc.UnknownRemoteException: Remote exception occurred:</title>
      <link>https://community.databricks.com/t5/data-engineering/i-am-getting-this-error-com-databricks-backend-common-rpc/m-p/24144#M16767</link>
      <description>&lt;P&gt;com.databricks.backend.common.rpc.DatabricksExceptions$SQLExecutionException: org.apache.spark.sql.AnalysisException:&lt;/P&gt;&lt;P&gt;The query operator `UpdateCommandEdge` contains one or more unsupported&lt;/P&gt;&lt;P&gt;expression types Aggregate, Window or Generate.&lt;/P&gt;&lt;P&gt;Invalid expressions: [avg(spark_catalog.eds_us_lake_cdp.cdp_job_log.Duration) OVER (PARTITION BY spark_catalog.eds_us_lake_cdp.cdp_job_log.job_id ORDER BY spark_catalog.eds_us_lake_cdp.cdp_job_log.Job_Start_Date_Time ASC NULLS FIRST RANGE BETWEEN INTERVAL '29' DAY PRECEDING AND CURRENT ROW), avg(spark_catalog.eds_us_lake_cdp.cdp_job_log.Duration)];&lt;/P&gt;&lt;P&gt;UpdateCommandEdge Delta[version=433, s3://tpc-aws-ted-dev-edpp-lake-cdp-us-east-1/eds_us_lake_cdp/cdp_job_log/delta], [Job_Run_Id#10299, Job_Id#10300, Batch_Run_Id#10301, Tidal_Job_No#10302, Source_Layer#10303, Source_Object_Location#10304, Source_Object_Name#10305, Target_Layer#10306, Target_Object_Location#10307, Target_Object_Name#10308, Status#10309, Status_Source#10310, Step_Control_Log#10311, Job_Scheduled_Date_Time#10312, Job_Start_Date_Time#10313, Job_End_Date_Time#10314, Error_Description#10315, Source_Record_Count#10316, Target_Record_Count#10317, MD5_HASH#10318, User_Id#10319, Created_Date_Time#10320, Duration#10321, round(avg(Duration#10321) windowspecdefinition(job_id#10300, Job_Start_Date_Time#10313 ASC NULLS FIRST, specifiedwindowframe(RangeFrame, -INTERVAL '29' DAY, currentrow$())), 2)]&lt;/P&gt;&lt;P&gt;+- SubqueryAlias spark_catalog.eds_us_lake_cdp.cdp_job_log&lt;/P&gt;&lt;P&gt;+- Relation eds_us_lake_cdp.cdp_job_log[Job_Run_Id#10299,Job_Id#10300,Batch_Run_Id#10301,Tidal_Job_No#10302,Source_Layer#10303,Source_Object_Location#10304,Source_Object_Name#10305,Target_Layer#10306,Target_Object_Location#10307,Target_Object_Name#10308,Status#10309,Status_Source#10310,Step_Control_Log#10311,Job_Scheduled_Date_Time#10312,Job_Start_Date_Time#10313,Job_End_Date_Time#10314,Error_Description#10315,Source_Record_Count#10316,Target_Record_Count#10317,MD5_HASH#10318,User_Id#10319,Created_Date_Time#10320,Duration#10321,Average_Run#10322] parquet&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.catalyst.analysis.CheckAnalysis.failAnalysis(CheckAnalysis.scala:60)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.catalyst.analysis.CheckAnalysis.failAnalysis$(CheckAnalysis.scala:59)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.catalyst.analysis.Analyzer.failAnalysis(Analyzer.scala:221)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.catalyst.analysis.CheckAnalysis.$anonfun$checkAnalysis$2(CheckAnalysis.scala:623)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.catalyst.analysis.CheckAnalysis.$anonfun$checkAnalysis$2$adapted(CheckAnalysis.scala:105)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.catalyst.trees.TreeNode.foreachUp(TreeNode.scala:358)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.catalyst.analysis.CheckAnalysis.$anonfun$checkAnalysis$1(CheckAnalysis.scala:105)&lt;/P&gt;&lt;P&gt;at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)&lt;/P&gt;&lt;P&gt;at com.databricks.spark.util.FrameProfiler$.record(FrameProfiler.scala:80)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.catalyst.analysis.CheckAnalysis.checkAnalysis(CheckAnalysis.scala:100)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.catalyst.analysis.CheckAnalysis.checkAnalysis$(CheckAnalysis.scala:100)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.catalyst.analysis.Analyzer.checkAnalysis(Analyzer.scala:221)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.catalyst.analysis.Analyzer.$anonfun$executeAndCheck$1(Analyzer.scala:275)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.markInAnalyzer(AnalysisHelper.scala:331)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.catalyst.analysis.Analyzer.executeAndCheck(Analyzer.scala:272)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.execution.QueryExecution.$anonfun$analyzed$1(QueryExecution.scala:128)&lt;/P&gt;&lt;P&gt;at com.databricks.spark.util.FrameProfiler$.record(FrameProfiler.scala:80)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.catalyst.QueryPlanningTracker.measurePhase(QueryPlanningTracker.scala:268)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.execution.QueryExecution.$anonfun$executePhase$1(QueryExecution.scala:265)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:968)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.execution.QueryExecution.executePhase(QueryExecution.scala:265)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.execution.QueryExecution.analyzed$lzycompute(QueryExecution.scala:129)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.execution.QueryExecution.analyzed(QueryExecution.scala:126)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.execution.QueryExecution.assertAnalyzed(QueryExecution.scala:118)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.Dataset$.$anonfun$ofRows$2(Dataset.scala:103)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:968)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:101)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.SparkSession.$anonfun$sql$1(SparkSession.scala:803)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:968)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:798)&lt;/P&gt;&lt;P&gt;at org.apache.spark.sql.SQLContext.sql(SQLContext.scala:695)&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.SQLDriverLocal.$anonfun$executeSql$1(SQLDriverLocal.scala:91)&lt;/P&gt;&lt;P&gt;at scala.collection.immutable.List.map(List.scala:297)&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.SQLDriverLocal.executeSql(SQLDriverLocal.scala:37)&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.SQLDriverLocal.repl(SQLDriverLocal.scala:145)&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.DriverLocal.$anonfun$execute$13(DriverLocal.scala:634)&lt;/P&gt;&lt;P&gt;at com.databricks.logging.Log4jUsageLoggingShim$.$anonfun$withAttributionContext$1(Log4jUsageLoggingShim.scala:33)&lt;/P&gt;&lt;P&gt;at scala.util.DynamicVariable.withValue(DynamicVariable.scala:62)&lt;/P&gt;&lt;P&gt;at com.databricks.logging.AttributionContext$.withValue(AttributionContext.scala:94)&lt;/P&gt;&lt;P&gt;at com.databricks.logging.Log4jUsageLoggingShim$.withAttributionContext(Log4jUsageLoggingShim.scala:31)&lt;/P&gt;&lt;P&gt;at com.databricks.logging.UsageLogging.withAttributionContext(UsageLogging.scala:205)&lt;/P&gt;&lt;P&gt;at com.databricks.logging.UsageLogging.withAttributionContext$(UsageLogging.scala:204)&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.DriverLocal.withAttributionContext(DriverLocal.scala:59)&lt;/P&gt;&lt;P&gt;at com.databricks.logging.UsageLogging.withAttributionTags(UsageLogging.scala:240)&lt;/P&gt;&lt;P&gt;at com.databricks.logging.UsageLogging.withAttributionTags$(UsageLogging.scala:225)&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.DriverLocal.withAttributionTags(DriverLocal.scala:59)&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.DriverLocal.execute(DriverLocal.scala:611)&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.DriverWrapper.$anonfun$tryExecutingCommand$1(DriverWrapper.scala:615)&lt;/P&gt;&lt;P&gt;at scala.util.Try$.apply(Try.scala:213)&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.DriverWrapper.tryExecutingCommand(DriverWrapper.scala:607)&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.DriverWrapper.executeCommandAndGetError(DriverWrapper.scala:526)&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.DriverWrapper.executeCommand(DriverWrapper.scala:561)&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.DriverWrapper.runInnerLoop(DriverWrapper.scala:431)&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.DriverWrapper.runInner(DriverWrapper.scala:374)&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.DriverWrapper.run(DriverWrapper.scala:225)&lt;/P&gt;&lt;P&gt;at java.lang.Thread.run(Thread.java:748)&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.SQLDriverLocal.executeSql(SQLDriverLocal.scala:130)&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.SQLDriverLocal.repl(SQLDriverLocal.scala:145)&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.DriverLocal.$anonfun$execute$13(DriverLocal.scala:634)&lt;/P&gt;&lt;P&gt;at com.databricks.logging.Log4jUsageLoggingShim$.$anonfun$withAttributionContext$1(Log4jUsageLoggingShim.scala:33)&lt;/P&gt;&lt;P&gt;at scala.util.DynamicVariable.withValue(DynamicVariable.scala:62)&lt;/P&gt;&lt;P&gt;at com.databricks.logging.AttributionContext$.withValue(AttributionContext.scala:94)&lt;/P&gt;&lt;P&gt;at com.databricks.logging.Log4jUsageLoggingShim$.withAttributionContext(Log4jUsageLoggingShim.scala:31)&lt;/P&gt;&lt;P&gt;at com.databricks.logging.UsageLogging.withAttributionContext(UsageLogging.scala:205)&lt;/P&gt;&lt;P&gt;at com.databricks.logging.UsageLogging.withAttributionContext$(UsageLogging.scala:204)&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.DriverLocal.withAttributionContext(DriverLocal.scala:59)&lt;/P&gt;&lt;P&gt;at com.databricks.logging.UsageLogging.withAttributionTags(UsageLogging.scala:240)&lt;/P&gt;&lt;P&gt;at com.databricks.logging.UsageLogging.withAttributionTags$(UsageLogging.scala:225)&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.DriverLocal.withAttributionTags(DriverLocal.scala:59)&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.DriverLocal.execute(DriverLocal.scala:611)&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.DriverWrapper.$anonfun$tryExecutingCommand$1(DriverWrapper.scala:615)&lt;/P&gt;&lt;P&gt;at scala.util.Try$.apply(Try.scala:213)&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.DriverWrapper.tryExecutingCommand(DriverWrapper.scala:607)&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.DriverWrapper.executeCommandAndGetError(DriverWrapper.scala:526)&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.DriverWrapper.executeCommand(DriverWrapper.scala:561)&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.DriverWrapper.runInnerLoop(DriverWrapper.scala:431)&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.DriverWrapper.runInner(DriverWrapper.scala:374)&lt;/P&gt;&lt;P&gt;at com.databricks.backend.daemon.driver.DriverWrapper.run(DriverWrapper.scala:225)&lt;/P&gt;&lt;P&gt;at java.lang.Thread.run(Thread.java:748)&lt;/P&gt;</description>
      <pubDate>Wed, 01 Feb 2023 14:01:32 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/i-am-getting-this-error-com-databricks-backend-common-rpc/m-p/24144#M16767</guid>
      <dc:creator>pc</dc:creator>
      <dc:date>2023-02-01T14:01:32Z</dc:date>
    </item>
    <item>
      <title>Re: I am getting this error:
com.databricks.backend.common.rpc.DatabricksExceptions$SQLExecutionException: com.databricks.rpc.UnknownRemoteException: Remote exception occurred:</title>
      <link>https://community.databricks.com/t5/data-engineering/i-am-getting-this-error-com-databricks-backend-common-rpc/m-p/24142#M16765</link>
      <description>&lt;P&gt;Hi @Databricks University Alliance​&amp;nbsp;, Could you please paste the whole error snippet here? &lt;/P&gt;</description>
      <pubDate>Thu, 03 Nov 2022 06:34:58 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/i-am-getting-this-error-com-databricks-backend-common-rpc/m-p/24142#M16765</guid>
      <dc:creator>Debayan</dc:creator>
      <dc:date>2022-11-03T06:34:58Z</dc:date>
    </item>
    <item>
      <title>Re: I am getting this error:
com.databricks.backend.common.rpc.DatabricksExceptions$SQLExecutionExce</title>
      <link>https://community.databricks.com/t5/data-engineering/i-am-getting-this-error-com-databricks-backend-common-rpc/m-p/98302#M39678</link>
      <description>&lt;P&gt;i am facing the issue before i trying to execute a code error message:&lt;SPAN&gt;com.databricks.rpc.UnknownRemoteException: Remote exception occurred:&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Mon, 11 Nov 2024 01:56:45 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/i-am-getting-this-error-com-databricks-backend-common-rpc/m-p/98302#M39678</guid>
      <dc:creator>aparna123</dc:creator>
      <dc:date>2024-11-11T01:56:45Z</dc:date>
    </item>
  </channel>
</rss>

