โ02-22-2023 02:14 PM
I want to run a block of code in a script and not in a notebook on databricks, however I cannot properly instantiate the spark context without some error.
I have tried ` SparkContext.getOrCreate()`, but this does not work.
Is there a simple way to do this I am missing?
โ02-22-2023 08:29 PM
Hi @Andrej Erkelensโ ,
Can you please send error which you are getting while using above code?
โ02-23-2023 09:11 AM
@Kaniz Fatmaโ
Hi, I have tried this but receive an error
`RuntimeError: A master URL must be set in your configuration`
Is there something I am missing to use databricks cluster (AWS backend) in a py script?
Thanks
โ04-14-2023 02:21 PM
I have the same problem, and would be interested in a solution
โ03-12-2024 10:02 PM
Hi,
I tried doing this and get master URL and app name error. I tried setting those and get an error message that asks not to create a spark session in databricks and use sparkcontext.getOrCreate() method instead.
But that leads to the same error. I used the getActiveSession method to verify the python script does not have access to a spark session.
โ01-05-2024 01:06 PM
Did this ever get addressed? I would like to use a Databricks notebook to launch a python-based child process (os.popen) that itself ultimately needs to use pyspark. When I try this, I either get told to supply a Master URL to the Spark context, or if I apply local[*] as master, I get told in an exception message on Spark interaction that Notebooks should use the shared context available via sc. This code is executing in a standalone python library being run by the subprocess (based on python, but not just a python script) launched from Notebook.
Is it simply disallowed to access Spark outside of the shared context sc? If so, how can we access that shared context from a standalone python library as I describe?
โ01-07-2024 02:37 AM - edited โ01-07-2024 02:41 AM
Thanks for information
โ03-13-2024 03:51 AM
Try this pay attention to import
from pyspark.sql import SparkSession
appName = "abc"
# Create a SparkSession
spark = SparkSession.builder \
.appName(appName) \
.getOrCreate()
# Your PySpark code blah blah
# Stop the SparkSession when done
spark.stop()
โ03-13-2024 07:55 AM
Thanks @MichTalebzadeh , but I have tried this to no avail, I get [MASTER_URL_NOT_SET] error, and when I try to set it, I get an error stating I cannot create another spark session. the getActiveSession() returns null from within the script, but returns the session when called from the notebook.
โ03-13-2024 08:03 AM
Thanks. Please send the full detail of error you are getting
โ03-13-2024 10:03 AM
Sadly, I have the errors in my corporate environment and I can't show the exact error details from this account.But it is quite close to that of @testing3
โ03-13-2024 08:31 AM
I came across a similar issue.
Please detail how you are executing the python script. Are you calling it from the web terminal? or from a notebook?
Note: If you are calling it from the web terminal - your spark session wont be passed. You could create a local variable and pass it in if youd like. I have never gotten to that point yet tho
โ03-13-2024 10:04 AM
I am running the script from databricks notebook : !streamlit run MyScript.py
โ03-14-2024 07:47 AM
Hi,
I believe running a Streamlit app directly from a Databricks notebook using !streamlit run <python_code>.py is not the way to do it, because Databricks notebooks are not designed to host Streamlit
OK try below
Create a Databricks Runtime with Streamlit Pre-installed. Configure Cluster, ehen creating a new Databricks cluster, select a runtime that includes Streamlit pre-installed. This eliminates the installation step.
Run the script: Within the notebook cell, simply execute the script directly:
!streamlit run MyScript.py
HTH
apps.
โ03-14-2024 10:57 AM
Hi @MichTalebzadeh ,
Once again, thanks for your replies.
My databricks cluster does come preinstalled with streamlit, and I have been running the script the way you mentioned.
I am going to try using alternatives to spark for the time being, and try with spark session isolation disabled as well.
I appreciate you taking out time to respond to this issue.
Passionate about hosting events and connecting people? Help us grow a vibrant local communityโsign up today to get started!
Sign Up Now