โ05-08-2023 02:00 AM
I install the newest version "databricks-connect==13.0.0". Now get the issue
Command C:\Users\Y\AppData\Local\pypoetry\Cache\virtualenvs\X-py3.9\Lib\site-packages\pyspark\bin\spark-class2.cmd"" not found
konnte nicht gefunden werden.
Traceback (most recent call last):
File "C:\X\repositories\schema-integration-customer\tmp_run_builder.py", line 37, in <module>
spark = get_spark()
File "C:\X\repositories\data-common\X\da\common\_library\spark.py", line 60, in get_spark
return builder.getOrCreate()
File "C:\Users\Y\AppData\Local\pypoetry\Cache\virtualenvs\X-schema-integration-customer-hjO9aLIy-py3.9\lib\site-packages\pyspark\sql\session.py", line 479, in getOrCreate
else SparkContext.getOrCreate(sparkConf)
File "C:\Users\Y\AppData\Local\pypoetry\Cache\virtualenvs\X-schema-integration-customer-hjO9aLIy-py3.9\lib\site-packages\pyspark\context.py", line 560, in getOrCreate
SparkContext(conf=conf or SparkConf())
File "C:\Users\y\AppData\Local\pypoetry\Cache\virtualenvs\x-schema-integration-customer-hjO9aLIy-py3.9\lib\site-packages\pyspark\context.py", line 202, in __init__
SparkContext._ensure_initialized(self, gateway=gateway, conf=conf)
File "C:\Users\y\AppData\Local\pypoetry\Cache\virtualenvs\x-schema-integration-customer-hjO9aLIy-py3.9\lib\site-packages\pyspark\context.py", line 480, in _ensure_initialized
SparkContext._gateway = gateway or launch_gateway(conf)
File "C:\Users\y\AppData\Local\pypoetry\Cache\virtualenvs\x-schema-integration-customer-hjO9aLIy-py3.9\lib\site-packages\pyspark\java_gateway.py", line 106, in launch_gateway
raise RuntimeError("Java gateway process exited before sending its port number")
RuntimeError: Java gateway process exited before sending its port number
Process finished with exit code 1
I use Windows and for the version "databricks-connect==11.3.10" everything run smooth
โ05-13-2023 08:38 AM
@Lazloo XPโ :
The error message you received indicates that there is a problem with launching the Java gateway process. This is typically caused by a misconfiguration in the environment variables that point to the location of the Spark and Java executables.
To resolve this issue, you can try the following steps:
If you are still encountering issues after verifying these steps, you may want to try reinstalling Databricks Connect 13.0.0 or rolling back to version 11.3.10 if it was working previously.
โ05-16-2023 11:41 PM
Hey @Suteja Kanuriโ ,
I have no issue with version 11.3.10.. Therefore I believe that my environment variables are set correctly.
โ05-17-2023 04:30 AM
with the newest Version the error changed to
```
Traceback (most recent call last):
File "C:\x\repositories\lf_backup_repo\snippets.py", line 4, in <module>
spark = SparkSession.builder.getOrCreate()
File "C:\Users\x\AppData\Local\pypoetry\Cache\virtualenvs\lf-privat-eCptrNhE-py3.8\lib\site-packages\pyspark\sql\session.py", line 469, in getOrCreate
raise RuntimeError(
RuntimeError: Only remote Spark sessions using Databricks Connect are supported. Could not find connection parameters to start a Spark remote session.
```
โ05-30-2023 10:21 PM
I am also facing similar issue with databricks-connect 13.
Getting RuntimeError: Only remote Spark sessions using Databricks Connect are supported. Could not find connection parameters to start a Spark remote session.
Databricks-connect - 13
DBR - 13
Python: 3.10.11
โ10-28-2023 12:09 PM
I get the same error. Please help with any hints.
โ04-30-2024 07:21 AM
Use this code:
from databricks.connect import DatabricksSession
spark = DatabricksSession.builder.getOrCreate()
Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you wonโt want to miss the chance to attend and share knowledge.
If there isnโt a group near you, start one and help create a community that brings people together.
Request a New Group