<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Running `pyspark` with `databricks-connect` in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10323#M5529</link>
    <description>&lt;P&gt;These are the permissions on the cluster. Is that what you wanted?&lt;/P&gt;</description>
    <pubDate>Thu, 09 Feb 2023 14:11:02 GMT</pubDate>
    <dc:creator>agagrins</dc:creator>
    <dc:date>2023-02-09T14:11:02Z</dc:date>
    <item>
      <title>Running `pyspark` with `databricks-connect`</title>
      <link>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10309#M5515</link>
      <description>&lt;P&gt;Hiya,&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;I'm trying to run `pyspark` with `databricks-connect==11.30.b0`, but am failing.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;The trace I see is&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;```&lt;/P&gt;&lt;P&gt;&amp;nbsp;File "/home/agagrins/databricks9/lib/python3.9/site-packages/py4j/java_gateway.py", line 1321, in __call__&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;return_value = get_return_value(&lt;/P&gt;&lt;P&gt;&amp;nbsp;File "/home/agagrins/databricks9/lib/python3.9/site-packages/pyspark/sql/utils.py", line 196, in deco&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;return f(*a, **kw)&lt;/P&gt;&lt;P&gt;&amp;nbsp;File "/home/agagrins/databricks9/lib/python3.9/site-packages/py4j/protocol.py", line 326, in get_return_value&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;raise Py4JJavaError(&lt;/P&gt;&lt;P&gt;py4j.protocol.Py4JJavaError: An error occurred while calling o33.sql.&lt;/P&gt;&lt;P&gt;: org.apache.spark.SparkException: There is no Credential Scope.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;at com.databricks.unity.UCSDriver$Manager.$anonfun$currentScopeId$1(UCSDriver.scala:94)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;at scala.Option.getOrElse(Option.scala:189)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;at com.databricks.unity.UCSDriver$Manager.currentScopeId(UCSDriver.scala:94)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;at com.databricks.unity.UCSDriver$Manager.currentScope(UCSDriver.scala:97)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;at com.databricks.unity.UnityCredentialScope$.currentScope(UnityCredentialScope.scala:100)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;at com.databricks.unity.UnityCredentialScope$.getCredentialManager(UnityCredentialScope.scala:128)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;at com.databricks.unity.CredentialManager$.getUnityApiTokenOpt(CredentialManager.scala:456)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;at com.databricks.unity.UnityCatalogClientHelper$.getToken(UnityCatalogClientHelper.scala:34)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;at com.databricks.managedcatalog.ManagedCatalogClientImpl.$anonfun$getCatalog$1(ManagedCatalogClientImpl.scala:163)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;at com.databricks.spark.util.FrameProfiler$.record(FrameProfiler.scala:80)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;at com.databricks.managedcatalog.ManagedCatalogClientImpl.$anonfun$recordAndWrapException$1(ManagedCatalogClientImpl.scala:2904)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;at com.databricks.managedcatalog.ErrorDetailsHandler.wrapServiceException(ErrorDetailsHandler.scala:25)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;at com.databricks.managedcatalog.ErrorDetailsHandler.wrapServiceException$(ErrorDetailsHandler.scala:23)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;at com.databricks.managedcatalog.ManagedCatalogClientImpl.wrapServiceException(ManagedCatalogClientImpl.scala:77)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;at com.databricks.managedcatalog.ManagedCatalogClientImpl.recordAndWrapException(ManagedCatalogClientImpl.scala:2903)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;at com.databricks.managedcatalog.ManagedCatalogClientImpl.getCatalog(ManagedCatalogClientImpl.scala:156)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;at com.databricks.sql.managedcatalog.ManagedCatalogCommon.catalogExists(ManagedCatalogCommon.scala:94)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;at com.databricks.sql.managedcatalog.PermissionEnforcingManagedCatalog.catalogExists(PermissionEnforcingManagedCatalog.scala:177)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;at com.databricks.sql.managedcatalog.ManagedCatalogSessionCatalog.catalogExists(ManagedCatalogSessionCatalog.scala:384)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;at com.databricks.sql.DatabricksCatalogManager.isCatalogRegistered(DatabricksCatalogManager.scala:104)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;at org.apache.spark.sql.SparkServiceCatalogV2Handler$.catalogOperationV2(SparkServiceCatalogV2Handler.scala:58)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;at com.databricks.service.SparkServiceImpl$.$anonfun$catalogOperationV2$1(SparkServiceImpl.scala:165)&lt;/P&gt;&lt;P&gt;```&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;I've tried to Google "There is no Credential Scope", but to no avail. Anyone have a clue of where to start to look?&lt;/P&gt;</description>
      <pubDate>Wed, 01 Feb 2023 11:24:28 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10309#M5515</guid>
      <dc:creator>agagrins</dc:creator>
      <dc:date>2023-02-01T11:24:28Z</dc:date>
    </item>
    <item>
      <title>Re: Running `pyspark` with `databricks-connect`</title>
      <link>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10310#M5516</link>
      <description>&lt;P&gt;where you are running?&lt;/P&gt;</description>
      <pubDate>Wed, 01 Feb 2023 11:28:04 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10310#M5516</guid>
      <dc:creator>sher</dc:creator>
      <dc:date>2023-02-01T11:28:04Z</dc:date>
    </item>
    <item>
      <title>Re: Running `pyspark` with `databricks-connect`</title>
      <link>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10311#M5517</link>
      <description>&lt;P&gt;I'm starting the run locally, with Python 3.9.1 under WSL, but the idea then is to run the job in Databricks on AWS&lt;/P&gt;</description>
      <pubDate>Wed, 01 Feb 2023 11:32:45 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10311#M5517</guid>
      <dc:creator>agagrins</dc:creator>
      <dc:date>2023-02-01T11:32:45Z</dc:date>
    </item>
    <item>
      <title>Re: Running `pyspark` with `databricks-connect`</title>
      <link>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10312#M5518</link>
      <description>&lt;P&gt;Hello @Aigars Grins​. Can you tell me a bit more about what you are trying to run via Databricks Connect? Generally, we recommend using dbx for local development over Databricks Connect.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Could you also provide more information on what type of compute you are connecting to? Such as runtime and whether it is running on Unity Catalog or the legacy Hive Metastore?&lt;/P&gt;</description>
      <pubDate>Mon, 06 Feb 2023 07:58:23 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10312#M5518</guid>
      <dc:creator>sergiu</dc:creator>
      <dc:date>2023-02-06T07:58:23Z</dc:date>
    </item>
    <item>
      <title>Re: Running `pyspark` with `databricks-connect`</title>
      <link>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10313#M5519</link>
      <description>&lt;P&gt;My understanding is that there are three main ways for me to work with Databricks: `databricks-connect`, `databricks-sql-connector`, and `dbx`. I'm trying out all three, for slightly different purposes, to see what fits our worksflows best where.&lt;/P&gt;</description>
      <pubDate>Mon, 06 Feb 2023 13:23:11 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10313#M5519</guid>
      <dc:creator>agagrins</dc:creator>
      <dc:date>2023-02-06T13:23:11Z</dc:date>
    </item>
    <item>
      <title>Re: Running `pyspark` with `databricks-connect`</title>
      <link>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10314#M5520</link>
      <description>&lt;P&gt;As for the problem above it seems to have gone away. While I'm not sure, it felt a bit like I didn't do anything different. Buy instead I'm faced with a much more mundane situation.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Again, I'm here trying to make `databricks-connect` work.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;I simply do&lt;/P&gt;&lt;P&gt;```&lt;/P&gt;&lt;P&gt;$ python3 -m venv ~/databricks11&lt;/P&gt;&lt;P&gt;$ . ~/databricks11/bin/activate&lt;/P&gt;&lt;P&gt;$ pip install --upgrade pip&lt;/P&gt;&lt;P&gt;$ pip install --upgrade setuptools&lt;/P&gt;&lt;P&gt;$ pip install databricks-connect==11.3.0b0&lt;/P&gt;&lt;P&gt;$ databricks-connect configure&lt;/P&gt;&lt;P&gt;$ databricks-connect test&lt;/P&gt;&lt;P&gt;```&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;My `.databricks-connect` looks like&lt;/P&gt;&lt;P&gt;```&lt;/P&gt;&lt;P&gt;{&lt;/P&gt;&lt;P&gt;  "host": "&lt;A href="https://dbc-34a0347a-0fd1.cloud.databricks.com/" alt="https://dbc-34a0347a-0fd1.cloud.databricks.com/" target="_blank"&gt;https://dbc-****.cloud.databricks.com&lt;/A&gt;",&lt;/P&gt;&lt;P&gt;  "token": "dapi****",&lt;/P&gt;&lt;P&gt;  "cluster_id": "0110-****,&lt;/P&gt;&lt;P&gt;  "port": "15001"&lt;/P&gt;&lt;P&gt;}&lt;/P&gt;&lt;P&gt;```&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;I also have some environment variables, just in case&lt;/P&gt;&lt;P&gt;```&lt;/P&gt;&lt;P&gt;DATABRICKS_ADDRESS=&lt;A href="https://dbc-34a0347a-0fd1.cloud.databricks.com/" alt="https://dbc-34a0347a-0fd1.cloud.databricks.com/" target="_blank"&gt;https://dbc-****.cloud.databricks.com&lt;/A&gt;&lt;/P&gt;&lt;P&gt;DATABRICKS_API_TOKEN=dapi****&lt;/P&gt;&lt;P&gt;DATABRICKS_CLUSTER_ID=0110-****&lt;/P&gt;&lt;P&gt;DATABRICKS_PORT=15001&lt;/P&gt;&lt;P&gt;```&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;But I get an error&lt;/P&gt;&lt;P&gt;```&lt;/P&gt;&lt;P&gt;23/02/03 11:47:17 ERROR SparkClientManager: Fail to get the SparkClient&lt;/P&gt;&lt;P&gt;java.util.concurrent.ExecutionException: com.databricks.service.SparkServiceConnectionException: Invalid token&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;To connect to a Databricks cluster, you must specify an API token.&lt;/P&gt;&lt;P&gt;API Token: The API token used to confirm your identity to Databricks&lt;/P&gt;&lt;P&gt;  - Learn more about API tokens here: &lt;A href="https://docs.databricks.com/api/latest/authentication.html#generate-a-token" alt="https://docs.databricks.com/api/latest/authentication.html#generate-a-token" target="_blank"&gt;https://docs.databricks.com/api/latest/authentication.html#generate-a-token&lt;/A&gt;&lt;/P&gt;&lt;P&gt;  - Get current value: spark.conf.get("spark.databricks.service.token")&lt;/P&gt;&lt;P&gt;  - Set via conf: spark.conf.set("spark.databricks.service.token", &amp;lt;your API token&amp;gt;)&lt;/P&gt;&lt;P&gt;  - Set via environment variable: export DATABRICKS_API_TOKEN=&amp;lt;your API token&amp;gt;&lt;/P&gt;&lt;P&gt;```&lt;/P&gt;</description>
      <pubDate>Mon, 06 Feb 2023 13:27:00 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10314#M5520</guid>
      <dc:creator>agagrins</dc:creator>
      <dc:date>2023-02-06T13:27:00Z</dc:date>
    </item>
    <item>
      <title>Re: Running `pyspark` with `databricks-connect`</title>
      <link>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10315#M5521</link>
      <description>&lt;P&gt;The cluster I'm connecting to runs "11.3 LTS (includes Apache Spark 3.3.0, Scala 2.12)"&lt;/P&gt;</description>
      <pubDate>Mon, 06 Feb 2023 13:28:57 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10315#M5521</guid>
      <dc:creator>agagrins</dc:creator>
      <dc:date>2023-02-06T13:28:57Z</dc:date>
    </item>
    <item>
      <title>Re: Running `pyspark` with `databricks-connect`</title>
      <link>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10316#M5522</link>
      <description>&lt;P&gt;Hmm, the connect info looks good to me. Can you try either of the following, see if you still get the error:&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;Run against a cluster unconnected to Unity Catalog (put it in Access mode - No isolation shared)&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;               OR&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;Try with an earlier runtime, like 10.4 (and appropriate version of the connector, do pip install -U "databricks-connect==10.4.*"&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Lastly, as stated in the &lt;A href="https://docs.databricks.com/dev-tools/databricks-connect.html" alt="https://docs.databricks.com/dev-tools/databricks-connect.html" target="_blank"&gt;documentation&lt;/A&gt;, we recommend running dbx for local development, over databricks-connect. Is there anything specific you believe you can do with databricks-connect, which you cannot achieve with dbx?&lt;/P&gt;</description>
      <pubDate>Mon, 06 Feb 2023 14:13:56 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10316#M5522</guid>
      <dc:creator>sergiu</dc:creator>
      <dc:date>2023-02-06T14:13:56Z</dc:date>
    </item>
    <item>
      <title>Re: Running `pyspark` with `databricks-connect`</title>
      <link>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10317#M5523</link>
      <description>&lt;P&gt;I tried with creating a new cluster, for 10.4, but that didn't get my anywhere either. The steps I followed where:&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;```&lt;/P&gt;&lt;P&gt;$ databricks clusters create --json-file cluster.json&lt;/P&gt;&lt;P&gt;```&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Where `cluster.json` looks like&lt;/P&gt;&lt;P&gt;```&lt;/P&gt;&lt;P&gt;{&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;"cluster_name": "test50",&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;"spark_version": "10.4.x-scala2.12",&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;"spark_conf": {&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;"spark.databricks.service.client.enabled": true,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;"spark.databricks.service.server.enabled": true,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;"spark.speculation": true,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;"spark.sql.session.timeZone": "UTC"&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;},&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;"spark_env_vars": {&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;"PYSPARK_PYTHON": "/databricks/python3/bin/python3"&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;},&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;"node_type_id": "i3.xlarge",&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;"autoscale": {&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;"min_workers": 1,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;"max_workers": 8&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;},&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;"autotermination_minutes": 10,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;"aws_attributes": {&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;"first_on_demand": 0,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;"availability": "SPOT_WITH_FALLBACK",&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;"zone_id": "eu-west-1b",&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&amp;nbsp;&amp;nbsp;"spot_bid_price_percent": 100&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;},&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;"enable_elastic_disk": false,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;"data_security_mode": "SINGLE_USER",&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;"single_user_name": "****"&lt;/P&gt;&lt;P&gt;}&lt;/P&gt;&lt;P&gt;```&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;And then&lt;/P&gt;&lt;P&gt;```&lt;/P&gt;&lt;P&gt;$ python3 -m venv ~/databricks12&lt;/P&gt;&lt;P&gt;$ . ~/databricks12/bin/activate&lt;/P&gt;&lt;P&gt;$ pip install --upgrade pip&lt;/P&gt;&lt;P&gt;$ pip install --upgrade setuptools&lt;/P&gt;&lt;P&gt;$ pip install databricks-connect==10.4.18&lt;/P&gt;&lt;P&gt;$ databricks-connect test&lt;/P&gt;&lt;P&gt;```&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;And the result is as before&lt;/P&gt;&lt;P&gt;```&lt;/P&gt;&lt;P&gt;23/02/09 10:22:14 ERROR SparkServiceRPCClient: Failed to sync with the spark cluster. This could be a intermittent issue, please check your cluster's state and retry.&lt;/P&gt;&lt;P&gt;com.databricks.service.SparkServiceConnectionException: Invalid token&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;To connect to a Databricks cluster, you must specify an API token.&lt;/P&gt;&lt;P&gt;API Token: The API token used to confirm your identity to Databricks&lt;/P&gt;&lt;P&gt;&amp;nbsp;- Learn more about API tokens here: &lt;A href="https://docs.databricks.com/api/latest/authentication.html#generate-a-token" target="test_blank"&gt;https://docs.databricks.com/api/latest/authentication.html#generate-a-token&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;- Get current value: spark.conf.get("spark.databricks.service.token")&lt;/P&gt;&lt;P&gt;&amp;nbsp;- Set via conf: spark.conf.set("spark.databricks.service.token", &amp;lt;your API token&amp;gt;)&lt;/P&gt;&lt;P&gt;&amp;nbsp;- Set via environment variable: export DATABRICKS_API_TOKEN=&amp;lt;your API token&amp;gt;&lt;/P&gt;&lt;P&gt;```&lt;/P&gt;</description>
      <pubDate>Thu, 09 Feb 2023 09:28:49 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10317#M5523</guid>
      <dc:creator>agagrins</dc:creator>
      <dc:date>2023-02-09T09:28:49Z</dc:date>
    </item>
    <item>
      <title>Re: Running `pyspark` with `databricks-connect`</title>
      <link>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10318#M5524</link>
      <description>&lt;P&gt;I'm not sure how to test the "Run against a cluster unconnected to Unity Catalog (put it in Access mode - No isolation shared)" thingy. Could you provide a `cluster.json` with the corresponding settings?&lt;/P&gt;&lt;P&gt;&lt;/P&gt;</description>
      <pubDate>Thu, 09 Feb 2023 09:30:01 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10318#M5524</guid>
      <dc:creator>agagrins</dc:creator>
      <dc:date>2023-02-09T09:30:01Z</dc:date>
    </item>
    <item>
      <title>Re: Running `pyspark` with `databricks-connect`</title>
      <link>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10319#M5525</link>
      <description>&lt;P&gt;Why then `databricks-connect` and not `dbx`? Well, I'm trying to get both to work.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;I posted a related question about `dbx` &lt;A href="https://community.databricks.com/s/feed/0D58Y00009qtFLrSAM" target="test_blank"&gt;https://community.databricks.com/s/feed/0D58Y00009qtFLrSAM&lt;/A&gt;.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;My hope here is that `databricks-connect` can have a much quicker turnaround time, compared to `dbx`, since no new environments have to be set up.&lt;/P&gt;</description>
      <pubDate>Thu, 09 Feb 2023 09:54:11 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10319#M5525</guid>
      <dc:creator>agagrins</dc:creator>
      <dc:date>2023-02-09T09:54:11Z</dc:date>
    </item>
    <item>
      <title>Re: Running `pyspark` with `databricks-connect`</title>
      <link>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10320#M5526</link>
      <description>&lt;P&gt;I tried your exact code on my environment and it worked without issue.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Could it be something about the token you are using and its permissions? Is it the same token you are using for the databricks CLI? What workspace permissions does the principal have?&lt;/P&gt;&lt;P&gt;&lt;/P&gt;</description>
      <pubDate>Thu, 09 Feb 2023 13:33:27 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10320#M5526</guid>
      <dc:creator>sergiu</dc:creator>
      <dc:date>2023-02-09T13:33:27Z</dc:date>
    </item>
    <item>
      <title>Re: Running `pyspark` with `databricks-connect`</title>
      <link>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10321#M5527</link>
      <description>&lt;P&gt;Change the data_security_mode field in the cluster config to NO_ISOLATION. It's unlikely that it is related to the issues you are facing, it's more likely an issue with the configuration.&lt;/P&gt;&lt;P&gt;But it might be worth double checking.&lt;/P&gt;</description>
      <pubDate>Thu, 09 Feb 2023 13:34:58 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10321#M5527</guid>
      <dc:creator>sergiu</dc:creator>
      <dc:date>2023-02-09T13:34:58Z</dc:date>
    </item>
    <item>
      <title>Re: Running `pyspark` with `databricks-connect`</title>
      <link>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10322#M5528</link>
      <description>&lt;P&gt;I use the same token when working with `dbx`, and that works, so I suspect the token itself isn't a problem. I'll check the permissions&lt;/P&gt;</description>
      <pubDate>Thu, 09 Feb 2023 13:47:14 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10322#M5528</guid>
      <dc:creator>agagrins</dc:creator>
      <dc:date>2023-02-09T13:47:14Z</dc:date>
    </item>
    <item>
      <title>Re: Running `pyspark` with `databricks-connect`</title>
      <link>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10323#M5529</link>
      <description>&lt;P&gt;These are the permissions on the cluster. Is that what you wanted?&lt;/P&gt;</description>
      <pubDate>Thu, 09 Feb 2023 14:11:02 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10323#M5529</guid>
      <dc:creator>agagrins</dc:creator>
      <dc:date>2023-02-09T14:11:02Z</dc:date>
    </item>
    <item>
      <title>Re: Running `pyspark` with `databricks-connect`</title>
      <link>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10324#M5530</link>
      <description>&lt;P&gt;Yes, this is what I was looking for. Does the token belong to the censored principal, or to a principal within the admin group? The token needs to belong to a principal which can attach on the cluster.&lt;/P&gt;</description>
      <pubDate>Fri, 10 Feb 2023 15:10:10 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10324#M5530</guid>
      <dc:creator>sergiu</dc:creator>
      <dc:date>2023-02-10T15:10:10Z</dc:date>
    </item>
    <item>
      <title>Re: Running `pyspark` with `databricks-connect`</title>
      <link>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10325#M5531</link>
      <description>&lt;P&gt;Can you put the whole error trace here? Or was the above the full error?&lt;/P&gt;</description>
      <pubDate>Fri, 10 Feb 2023 15:18:41 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10325#M5531</guid>
      <dc:creator>sergiu</dc:creator>
      <dc:date>2023-02-10T15:18:41Z</dc:date>
    </item>
    <item>
      <title>Re: Running `pyspark` with `databricks-connect`</title>
      <link>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10326#M5532</link>
      <description>&lt;P&gt;How to make it work in a cluster with Unity Catalog enabled?&lt;/P&gt;</description>
      <pubDate>Fri, 24 Mar 2023 03:06:33 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/running-pyspark-with-databricks-connect/m-p/10326#M5532</guid>
      <dc:creator>ryojikn</dc:creator>
      <dc:date>2023-03-24T03:06:33Z</dc:date>
    </item>
  </channel>
</rss>

