<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Getting databricks-connect com.fasterxml.jackson.databind.exc.MismatchedInputException parse warning in Get Started Discussions</title>
    <link>https://community.databricks.com/t5/get-started-discussions/getting-databricks-connect-com-fasterxml-jackson-databind-exc/m-p/67908#M2965</link>
    <description>&lt;P&gt;Hi community,&amp;nbsp;&lt;/P&gt;&lt;P&gt;I am getting below warning when I try using pyspark code for some of my use-cases using databricks-connect.&amp;nbsp;&lt;/P&gt;&lt;P&gt;Is this a critical warning, and any idea what does it mean?&lt;/P&gt;&lt;P&gt;Logs:&amp;nbsp;&lt;BR /&gt;WARN DatabricksConnectConf: Could not parse /root/.databricks-connect&lt;BR /&gt;INFO - WARN DatabricksConnectConf: Could not parse /root/.databricks-connect&lt;BR /&gt;INFO - com.fasterxml.jackson.databind.exc.MismatchedInputException: No content to map due to end-of-input&lt;BR /&gt;INFO - at [Source: (String)""; line: 1, column: 0]&lt;BR /&gt;INFO - at com.fasterxml.jackson.databind.exc.MismatchedInputException.from(MismatchedInputException.java:59)&lt;BR /&gt;INFO - at com.fasterxml.jackson.databind.ObjectMapper._initForReading(ObjectMapper.java:4765)&lt;BR /&gt;INFO - at com.fasterxml.jackson.databind.ObjectMapper._readMapAndClose(ObjectMapper.java:4667)&lt;BR /&gt;INFO - at com.fasterxml.jackson.databind.ObjectMapper.readValue(ObjectMapper.java:3629)&lt;BR /&gt;INFO - at com.databricks.spark.util.DatabricksConnectConf$.liftedTree1$1(DatabricksConnectConf.scala:68)&lt;BR /&gt;INFO - at com.databricks.spark.util.DatabricksConnectConf$.jsonConfig$lzycompute(DatabricksConnectConf.scala:64)&lt;BR /&gt;INFO - at com.databricks.spark.util.DatabricksConnectConf$.jsonConfig(DatabricksConnectConf.scala:56)&lt;BR /&gt;INFO - at com.databricks.spark.util.DatabricksConnectConf$.getToken(DatabricksConnectConf.scala:83)&lt;BR /&gt;INFO - at com.databricks.sql.DatabricksSQLConf$.$anonfun$SPARK_SERVICE_TOKEN$1(DatabricksSQLConf.scala:2591)&lt;BR /&gt;INFO - at org.apache.spark.internal.config.ConfigEntryWithDefaultFunction.defaultValueString(ConfigEntry.scala:181)&lt;BR /&gt;INFO - at org.apache.spark.sql.internal.SQLConf.$anonfun$getAllDefaultConfs$1(SQLConf.scala:5478)&lt;BR /&gt;INFO - at scala.collection.Iterator.foreach(Iterator.scala:943)&lt;BR /&gt;INFO - at scala.collection.AbstractIterator.foreach(Iterator.scala:1431)&lt;BR /&gt;INFO - at scala.collection.IterableLike.foreach$(IterableLike.scala:73)&lt;BR /&gt;INFO - at scala.collection.AbstractIterable.foreach(Iterable.scala:56)&lt;BR /&gt;INFO - at org.apache.spark.sql.internal.SQLConf.getAllDefaultConfs(SQLConf.scala:5476)&lt;BR /&gt;INFO - at org.apache.spark.sql.internal.SQLConf.recordNonDefaultConfs(SQLConf.scala:5490)&lt;BR /&gt;INFO - at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withCustomExecutionEnv$8(SQLExecution.scala:191)&lt;BR /&gt;INFO - at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:430)&lt;BR /&gt;INFO - at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withCustomExecutionEnv$1(SQLExecution.scala:187)&lt;BR /&gt;INFO - at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:1038)&lt;BR /&gt;INFO - at org.apache.spark.sql.execution.SQLExecution$.withCustomExecutionEnv(SQLExecution.scala:129)&lt;BR /&gt;INFO - at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:380)&lt;BR /&gt;INFO - at org.apache.spark.sql.execution.QueryExecution$$anonfun$$nestedInanonfun$eagerlyExecuteCommands$1$1.$anonfun$applyOrElse$1(QueryExecution.scala:237)&lt;BR /&gt;INFO - at org.apache.spark.sql.execution.QueryExecution.org$apache$spark$sql$execution$QueryExecution$$withMVTagsIfNecessary(QueryExecution.scala:220)&lt;BR /&gt;INFO - at org.apache.spark.sql.execution.QueryExecution$$anonfun$$nestedInanonfun$eagerlyExecuteCommands$1$1.applyOrElse(QueryExecution.scala:233)&lt;BR /&gt;QueryExecution.scala:226)&lt;BR /&gt;INFO - at org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$transformDownWithPruning$1(TreeNode.scala:519)&lt;BR /&gt;INFO - at org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:106)&lt;BR /&gt;INFO - at org.apache.spark.sql.catalyst.trees.TreeNode.transformDownWithPruning(TreeNode.scala:519)&lt;BR /&gt;INFO - at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.org$apache$spark$sql$catalyst$plans$logical$AnalysisHelper$$super$transformDownWithPruning(LogicalPlan.scala:32)&lt;BR /&gt;INFO - at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning(AnalysisHelper.scala:316)&lt;BR /&gt;INFO - at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning$(AnalysisHelper.scala:312)&lt;BR /&gt;INFO - at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.transformDownWithPruning(LogicalPlan.scala:32)&lt;BR /&gt;INFO - at org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:495)&lt;BR /&gt;INFO - at org.apache.spark.sql.execution.QueryExecution.$anonfun$eagerlyExecuteCommands$1(QueryExecution.scala:226)&lt;BR /&gt;INFO - at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.allowInvokingTransformsInAnalyzer(AnalysisHelper.scala:372)&lt;BR /&gt;INFO - at org.apache.spark.sql.execution.QueryExecution.eagerlyExecuteCommands(QueryExecution.scala:226)&lt;BR /&gt;INFO - at org.apache.spark.sql.execution.QueryExecution.commandExecuted$lzycompute(QueryExecution.scala:180)&lt;BR /&gt;INFO - at org.apache.spark.sql.execution.QueryExecution.commandExecuted(QueryExecution.scala:171)&lt;BR /&gt;INFO - at org.apache.spark.sql.Dataset.&amp;lt;init&amp;gt;(Dataset.scala:250)&lt;BR /&gt;INFO - at org.apache.spark.sql.Dataset$.$anonfun$ofRows$1(Dataset.scala:101)&lt;BR /&gt;INFO - at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:1038)&lt;BR /&gt;INFO - at org.apache.spark.sql.SparkSession.$anonfun$withActiveAndFrameProfiler$1(SparkSession.scala:1045)&lt;BR /&gt;INFO - at com.databricks.spark.util.FrameProfiler$.record(FrameProfiler.scala:24)&lt;BR /&gt;INFO - at org.apache.spark.sql.SparkSession.withActiveAndFrameProfiler(SparkSession.scala:1045)&lt;BR /&gt;INFO - at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:98)&lt;BR /&gt;INFO - at org.apache.spark.sql.Dataset.$anonfun$org$apache$spark$sql$Dataset$$withPlan$1(Dataset.scala:4414)&lt;BR /&gt;INFO - at com.databricks.spark.util.FrameProfiler$.record(FrameProfiler.scala:24)&lt;BR /&gt;INFO - at org.apache.spark.sql.Dataset.org$apache$spark$sql$Dataset$$withPlan(Dataset.scala:4414)&lt;BR /&gt;INFO - at org.apache.spark.sql.Dataset.withPlan(Dataset.scala:4408)&lt;BR /&gt;INFO - at org.apache.spark.sql.Dataset.createOrReplaceTempView(Dataset.scala:3858)&lt;BR /&gt;INFO - at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)&lt;BR /&gt;INFO - at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)&lt;BR /&gt;INFO - at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)&lt;BR /&gt;INFO - at java.lang.reflect.Method.invoke(Method.java:498)&lt;BR /&gt;INFO - at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)&lt;BR /&gt;INFO - at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:380)&lt;BR /&gt;INFO - at py4j.Gateway.invoke(Gateway.java:306)&lt;BR /&gt;INFO - at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)&lt;BR /&gt;INFO - at py4j.commands.CallCommand.execute(CallCommand.java:79)&lt;BR /&gt;INFO - at py4j.ClientServerConnection.waitForCommands(ClientServerConnection.java:195)&lt;BR /&gt;INFO - at py4j.ClientServerConnection.run(ClientServerConnection.java:115)&lt;BR /&gt;INFO - at java.lang.Thread.run(Thread.java:750)&lt;/P&gt;</description>
    <pubDate>Thu, 02 May 2024 06:11:17 GMT</pubDate>
    <dc:creator>Surajv</dc:creator>
    <dc:date>2024-05-02T06:11:17Z</dc:date>
    <item>
      <title>Getting databricks-connect com.fasterxml.jackson.databind.exc.MismatchedInputException parse warning</title>
      <link>https://community.databricks.com/t5/get-started-discussions/getting-databricks-connect-com-fasterxml-jackson-databind-exc/m-p/67908#M2965</link>
      <description>&lt;P&gt;Hi community,&amp;nbsp;&lt;/P&gt;&lt;P&gt;I am getting below warning when I try using pyspark code for some of my use-cases using databricks-connect.&amp;nbsp;&lt;/P&gt;&lt;P&gt;Is this a critical warning, and any idea what does it mean?&lt;/P&gt;&lt;P&gt;Logs:&amp;nbsp;&lt;BR /&gt;WARN DatabricksConnectConf: Could not parse /root/.databricks-connect&lt;BR /&gt;INFO - WARN DatabricksConnectConf: Could not parse /root/.databricks-connect&lt;BR /&gt;INFO - com.fasterxml.jackson.databind.exc.MismatchedInputException: No content to map due to end-of-input&lt;BR /&gt;INFO - at [Source: (String)""; line: 1, column: 0]&lt;BR /&gt;INFO - at com.fasterxml.jackson.databind.exc.MismatchedInputException.from(MismatchedInputException.java:59)&lt;BR /&gt;INFO - at com.fasterxml.jackson.databind.ObjectMapper._initForReading(ObjectMapper.java:4765)&lt;BR /&gt;INFO - at com.fasterxml.jackson.databind.ObjectMapper._readMapAndClose(ObjectMapper.java:4667)&lt;BR /&gt;INFO - at com.fasterxml.jackson.databind.ObjectMapper.readValue(ObjectMapper.java:3629)&lt;BR /&gt;INFO - at com.databricks.spark.util.DatabricksConnectConf$.liftedTree1$1(DatabricksConnectConf.scala:68)&lt;BR /&gt;INFO - at com.databricks.spark.util.DatabricksConnectConf$.jsonConfig$lzycompute(DatabricksConnectConf.scala:64)&lt;BR /&gt;INFO - at com.databricks.spark.util.DatabricksConnectConf$.jsonConfig(DatabricksConnectConf.scala:56)&lt;BR /&gt;INFO - at com.databricks.spark.util.DatabricksConnectConf$.getToken(DatabricksConnectConf.scala:83)&lt;BR /&gt;INFO - at com.databricks.sql.DatabricksSQLConf$.$anonfun$SPARK_SERVICE_TOKEN$1(DatabricksSQLConf.scala:2591)&lt;BR /&gt;INFO - at org.apache.spark.internal.config.ConfigEntryWithDefaultFunction.defaultValueString(ConfigEntry.scala:181)&lt;BR /&gt;INFO - at org.apache.spark.sql.internal.SQLConf.$anonfun$getAllDefaultConfs$1(SQLConf.scala:5478)&lt;BR /&gt;INFO - at scala.collection.Iterator.foreach(Iterator.scala:943)&lt;BR /&gt;INFO - at scala.collection.AbstractIterator.foreach(Iterator.scala:1431)&lt;BR /&gt;INFO - at scala.collection.IterableLike.foreach$(IterableLike.scala:73)&lt;BR /&gt;INFO - at scala.collection.AbstractIterable.foreach(Iterable.scala:56)&lt;BR /&gt;INFO - at org.apache.spark.sql.internal.SQLConf.getAllDefaultConfs(SQLConf.scala:5476)&lt;BR /&gt;INFO - at org.apache.spark.sql.internal.SQLConf.recordNonDefaultConfs(SQLConf.scala:5490)&lt;BR /&gt;INFO - at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withCustomExecutionEnv$8(SQLExecution.scala:191)&lt;BR /&gt;INFO - at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:430)&lt;BR /&gt;INFO - at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withCustomExecutionEnv$1(SQLExecution.scala:187)&lt;BR /&gt;INFO - at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:1038)&lt;BR /&gt;INFO - at org.apache.spark.sql.execution.SQLExecution$.withCustomExecutionEnv(SQLExecution.scala:129)&lt;BR /&gt;INFO - at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:380)&lt;BR /&gt;INFO - at org.apache.spark.sql.execution.QueryExecution$$anonfun$$nestedInanonfun$eagerlyExecuteCommands$1$1.$anonfun$applyOrElse$1(QueryExecution.scala:237)&lt;BR /&gt;INFO - at org.apache.spark.sql.execution.QueryExecution.org$apache$spark$sql$execution$QueryExecution$$withMVTagsIfNecessary(QueryExecution.scala:220)&lt;BR /&gt;INFO - at org.apache.spark.sql.execution.QueryExecution$$anonfun$$nestedInanonfun$eagerlyExecuteCommands$1$1.applyOrElse(QueryExecution.scala:233)&lt;BR /&gt;QueryExecution.scala:226)&lt;BR /&gt;INFO - at org.apache.spark.sql.catalyst.trees.TreeNode.$anonfun$transformDownWithPruning$1(TreeNode.scala:519)&lt;BR /&gt;INFO - at org.apache.spark.sql.catalyst.trees.CurrentOrigin$.withOrigin(TreeNode.scala:106)&lt;BR /&gt;INFO - at org.apache.spark.sql.catalyst.trees.TreeNode.transformDownWithPruning(TreeNode.scala:519)&lt;BR /&gt;INFO - at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.org$apache$spark$sql$catalyst$plans$logical$AnalysisHelper$$super$transformDownWithPruning(LogicalPlan.scala:32)&lt;BR /&gt;INFO - at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning(AnalysisHelper.scala:316)&lt;BR /&gt;INFO - at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper.transformDownWithPruning$(AnalysisHelper.scala:312)&lt;BR /&gt;INFO - at org.apache.spark.sql.catalyst.plans.logical.LogicalPlan.transformDownWithPruning(LogicalPlan.scala:32)&lt;BR /&gt;INFO - at org.apache.spark.sql.catalyst.trees.TreeNode.transformDown(TreeNode.scala:495)&lt;BR /&gt;INFO - at org.apache.spark.sql.execution.QueryExecution.$anonfun$eagerlyExecuteCommands$1(QueryExecution.scala:226)&lt;BR /&gt;INFO - at org.apache.spark.sql.catalyst.plans.logical.AnalysisHelper$.allowInvokingTransformsInAnalyzer(AnalysisHelper.scala:372)&lt;BR /&gt;INFO - at org.apache.spark.sql.execution.QueryExecution.eagerlyExecuteCommands(QueryExecution.scala:226)&lt;BR /&gt;INFO - at org.apache.spark.sql.execution.QueryExecution.commandExecuted$lzycompute(QueryExecution.scala:180)&lt;BR /&gt;INFO - at org.apache.spark.sql.execution.QueryExecution.commandExecuted(QueryExecution.scala:171)&lt;BR /&gt;INFO - at org.apache.spark.sql.Dataset.&amp;lt;init&amp;gt;(Dataset.scala:250)&lt;BR /&gt;INFO - at org.apache.spark.sql.Dataset$.$anonfun$ofRows$1(Dataset.scala:101)&lt;BR /&gt;INFO - at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:1038)&lt;BR /&gt;INFO - at org.apache.spark.sql.SparkSession.$anonfun$withActiveAndFrameProfiler$1(SparkSession.scala:1045)&lt;BR /&gt;INFO - at com.databricks.spark.util.FrameProfiler$.record(FrameProfiler.scala:24)&lt;BR /&gt;INFO - at org.apache.spark.sql.SparkSession.withActiveAndFrameProfiler(SparkSession.scala:1045)&lt;BR /&gt;INFO - at org.apache.spark.sql.Dataset$.ofRows(Dataset.scala:98)&lt;BR /&gt;INFO - at org.apache.spark.sql.Dataset.$anonfun$org$apache$spark$sql$Dataset$$withPlan$1(Dataset.scala:4414)&lt;BR /&gt;INFO - at com.databricks.spark.util.FrameProfiler$.record(FrameProfiler.scala:24)&lt;BR /&gt;INFO - at org.apache.spark.sql.Dataset.org$apache$spark$sql$Dataset$$withPlan(Dataset.scala:4414)&lt;BR /&gt;INFO - at org.apache.spark.sql.Dataset.withPlan(Dataset.scala:4408)&lt;BR /&gt;INFO - at org.apache.spark.sql.Dataset.createOrReplaceTempView(Dataset.scala:3858)&lt;BR /&gt;INFO - at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)&lt;BR /&gt;INFO - at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)&lt;BR /&gt;INFO - at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)&lt;BR /&gt;INFO - at java.lang.reflect.Method.invoke(Method.java:498)&lt;BR /&gt;INFO - at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)&lt;BR /&gt;INFO - at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:380)&lt;BR /&gt;INFO - at py4j.Gateway.invoke(Gateway.java:306)&lt;BR /&gt;INFO - at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)&lt;BR /&gt;INFO - at py4j.commands.CallCommand.execute(CallCommand.java:79)&lt;BR /&gt;INFO - at py4j.ClientServerConnection.waitForCommands(ClientServerConnection.java:195)&lt;BR /&gt;INFO - at py4j.ClientServerConnection.run(ClientServerConnection.java:115)&lt;BR /&gt;INFO - at java.lang.Thread.run(Thread.java:750)&lt;/P&gt;</description>
      <pubDate>Thu, 02 May 2024 06:11:17 GMT</pubDate>
      <guid>https://community.databricks.com/t5/get-started-discussions/getting-databricks-connect-com-fasterxml-jackson-databind-exc/m-p/67908#M2965</guid>
      <dc:creator>Surajv</dc:creator>
      <dc:date>2024-05-02T06:11:17Z</dc:date>
    </item>
  </channel>
</rss>

