<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Spark submit - not reading one of my --py-files arguments in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/spark-submit-not-reading-one-of-my-py-files-arguments/m-p/62783#M32068</link>
    <description>&lt;DIV&gt;&lt;FONT face="arial, sans-serif"&gt;&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;This below is catered for yarn&amp;nbsp;mode&lt;BR /&gt;&lt;/FONT&gt;&lt;P&gt;&lt;FONT face="arial, sans-serif"&gt;if your application code primarily consists of Python files &lt;STRONG&gt;and does not require a separate virtual environment with specific dependencies&lt;/STRONG&gt;, you can use the --py-files argument in spark-submit&lt;/FONT&gt;&lt;/P&gt;&lt;DIV&gt;&lt;PRE&gt;spark-submit --verbose \
   --master yarn \
  --deploy-mode cluster \
  --name $APPNAME \
  --driver-memory 1g \  # Adjust memory as needed
  --executor-memory 1g \  # Adjust memory as needed
  --num-executors 2 \     # Adjust executors as needed
  -&lt;STRONG&gt;-py-files ${build_directory}/source_code.zip \&lt;/STRONG&gt;
  $CODE_DIRECTORY_CLOUD/my_application_entry_point.py  # Path to your main application script&lt;/PRE&gt;&lt;/DIV&gt;&lt;P&gt;&lt;STRONG&gt;For application code with a separate virtual environment)&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&lt;FONT face="Google Sans, Helvetica Neue, sans-serif"&gt;&lt;SPAN&gt;If y&lt;/SPAN&gt;&lt;/FONT&gt;&lt;FONT face="arial, sans-serif"&gt;our application code has specific dependencies that you manage in a separate virtual environment, you can leverage the&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;&lt;/FONT&gt;&lt;FONT face="arial, sans-serif"&gt;--conf spark.yarn.dist.archive&lt;/FONT&gt;&lt;FONT&gt;s argument&lt;/FONT&gt;&lt;FONT face="arial, sans-serif"&gt;.&lt;/FONT&gt;&lt;/P&gt;&lt;FONT face="monospace"&gt;spark-submit --verbose \&amp;nbsp;&lt;/FONT&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;FONT face="monospace"&gt;-master yarn \&amp;nbsp;&lt;/FONT&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;FONT face="monospace"&gt;-deploy-mode cluster \&amp;nbsp;&lt;/FONT&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;FONT face="monospace"&gt;--name $APPNAME \&lt;/FONT&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;FONT face="monospace"&gt;&amp;nbsp;--driver-memory 1g \ # Adjust memory as needed&amp;nbsp;&lt;/FONT&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;FONT face="monospace"&gt;--executor-memory 1g \ # Adjust memory as needed&amp;nbsp;&lt;/FONT&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;FONT face="monospace"&gt;--num-executors 2 \ # Adjust executors as needed-&lt;/FONT&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;FONT face="monospace"&gt;&lt;STRONG&gt;-conf "spark.yarn.dist.archives"=${pyspark_venv}.tar.gz#pyspark_venv \&lt;/STRONG&gt;&lt;/FONT&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;FONT face="monospace"&gt;$CODE_DIRECTORY_CLOUD/my_application_entry_point.py # Path to your main application script&lt;/FONT&gt;&lt;P&gt;&lt;SPAN&gt;&lt;FONT face="arial, sans-serif"&gt;Explanation:&lt;/FONT&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;&lt;FONT face="arial, sans-serif"&gt;--conf "spark.yarn.dist.archives"=${pyspark_venv}.tar.gz#pyspark_venv&lt;/FONT&gt;&lt;FONT face="Google Sans, Helvetica Neue, sans-serif"&gt;:&lt;/FONT&gt;&lt;FONT face="arial, sans-serif"&gt;&amp;nbsp;This configures Spark to distribute your virtual environment archiv&lt;/FONT&gt;&lt;FONT face="arial, sans-serif"&gt;e&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;&lt;/FONT&gt;&lt;FONT face="Google Sans, Helvetica Neue, sans-serif"&gt;(&lt;/FONT&gt;&lt;FONT face="arial, sans-serif"&gt;pyspark_venv.tar.gz&lt;FONT&gt;)&lt;/FONT&gt;&lt;FONT&gt;&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;to the Yarn cluster nodes.&amp;nbsp;The&amp;nbsp;&lt;/FONT&gt;#pyspark_venv&amp;nbsp; part defines&lt;FONT&gt;&amp;nbsp;a symbolic link name within the container.&lt;/FONT&gt;&lt;/FONT&gt;&lt;/LI&gt;&lt;LI&gt;&lt;FONT face="arial, sans-serif"&gt;You do not need&amp;nbsp;--py-fileshere because the virtual environment archive will contain all necessary dependencies.&lt;/FONT&gt;&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;&lt;SPAN&gt;&lt;FONT face="arial, sans-serif"&gt;Choosing the best approach:&lt;/FONT&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;FONT face="arial, sans-serif"&gt;The choice depends on your project setup:&lt;/FONT&gt;&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;&lt;FONT face="arial, sans-serif"&gt;No Separate Virtual Environment:&amp;nbsp;Use&amp;nbsp; --py-files&amp;nbsp;if your application code consists mainly of Python files and doesn't require a separate virtual environment.&lt;/FONT&gt;&lt;/LI&gt;&lt;LI&gt;&lt;FONT face="arial, sans-serif"&gt;Separate Virtual Environment:&amp;nbsp;Use --conf spark.yarn.dist.archives&amp;nbsp;if you manage dependencies in a separate virtual environment archive.&lt;/FONT&gt;&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;&lt;FONT face="arial, sans-serif"&gt;HTH&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;&lt;/FONT&gt;&lt;/P&gt;&lt;P&gt;Mich Talebzadeh,&lt;/P&gt;&lt;P&gt;Dad | Technologist | Solutions Architect | Engineer&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;London&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;United Kingdom&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;/DIV&gt;</description>
    <pubDate>Wed, 06 Mar 2024 19:36:36 GMT</pubDate>
    <dc:creator>MichTalebzadeh</dc:creator>
    <dc:date>2024-03-06T19:36:36Z</dc:date>
    <item>
      <title>Spark submit - not reading one of my --py-files arguments</title>
      <link>https://community.databricks.com/t5/data-engineering/spark-submit-not-reading-one-of-my-py-files-arguments/m-p/62361#M31953</link>
      <description>&lt;P&gt;Hi. In Databricks workflows, I submit a spark job (Type = "Spark Submit"), and a bunch of parameters, starting with --py-files.&lt;/P&gt;&lt;P&gt;This works where all the files are in the same s3 path, but I get errors when I put a "common" module in a different s3 path:&lt;/P&gt;&lt;P&gt;"--py-files",&lt;BR /&gt;"s3://some_path/appl_src.py",&lt;BR /&gt;"s3://some_path/main.py",&lt;BR /&gt;"s3://a_different_path/common.py",&lt;/P&gt;&lt;P&gt;I get an error saying "common" doesn't exist, when I know in fact the path exists. From Standard output:&lt;/P&gt;&lt;P&gt;Traceback (most recent call last):&lt;BR /&gt;File "/local_disk0/tmp/spark-123/appl_src.py", line 21, in &amp;lt;module&amp;gt;&lt;BR /&gt;from common import my_functions&lt;BR /&gt;ModuleNotFoundError: No module named 'common'&lt;/P&gt;&lt;P&gt;Additionally log4j mentions the first two, but not the third:&lt;/P&gt;&lt;P&gt;24/02/28 21:41:00 INFO Utils: Fetching s3://some_path/appl_src.py to ...&lt;BR /&gt;24/02/28 21:41:00 INFO Utils: Fetching s3://some_path/main.py to ...&lt;/P&gt;&lt;P&gt;Why does Spark ignore the third argument? Or it has to be in the same s3 path?&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 29 Feb 2024 15:02:46 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/spark-submit-not-reading-one-of-my-py-files-arguments/m-p/62361#M31953</guid>
      <dc:creator>397973</dc:creator>
      <dc:date>2024-02-29T15:02:46Z</dc:date>
    </item>
    <item>
      <title>Re: Spark submit - not reading one of my --py-files arguments</title>
      <link>https://community.databricks.com/t5/data-engineering/spark-submit-not-reading-one-of-my-py-files-arguments/m-p/62783#M32068</link>
      <description>&lt;DIV&gt;&lt;FONT face="arial, sans-serif"&gt;&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;This below is catered for yarn&amp;nbsp;mode&lt;BR /&gt;&lt;/FONT&gt;&lt;P&gt;&lt;FONT face="arial, sans-serif"&gt;if your application code primarily consists of Python files &lt;STRONG&gt;and does not require a separate virtual environment with specific dependencies&lt;/STRONG&gt;, you can use the --py-files argument in spark-submit&lt;/FONT&gt;&lt;/P&gt;&lt;DIV&gt;&lt;PRE&gt;spark-submit --verbose \
   --master yarn \
  --deploy-mode cluster \
  --name $APPNAME \
  --driver-memory 1g \  # Adjust memory as needed
  --executor-memory 1g \  # Adjust memory as needed
  --num-executors 2 \     # Adjust executors as needed
  -&lt;STRONG&gt;-py-files ${build_directory}/source_code.zip \&lt;/STRONG&gt;
  $CODE_DIRECTORY_CLOUD/my_application_entry_point.py  # Path to your main application script&lt;/PRE&gt;&lt;/DIV&gt;&lt;P&gt;&lt;STRONG&gt;For application code with a separate virtual environment)&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&lt;FONT face="Google Sans, Helvetica Neue, sans-serif"&gt;&lt;SPAN&gt;If y&lt;/SPAN&gt;&lt;/FONT&gt;&lt;FONT face="arial, sans-serif"&gt;our application code has specific dependencies that you manage in a separate virtual environment, you can leverage the&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;&lt;/FONT&gt;&lt;FONT face="arial, sans-serif"&gt;--conf spark.yarn.dist.archive&lt;/FONT&gt;&lt;FONT&gt;s argument&lt;/FONT&gt;&lt;FONT face="arial, sans-serif"&gt;.&lt;/FONT&gt;&lt;/P&gt;&lt;FONT face="monospace"&gt;spark-submit --verbose \&amp;nbsp;&lt;/FONT&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;FONT face="monospace"&gt;-master yarn \&amp;nbsp;&lt;/FONT&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;FONT face="monospace"&gt;-deploy-mode cluster \&amp;nbsp;&lt;/FONT&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;FONT face="monospace"&gt;--name $APPNAME \&lt;/FONT&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;FONT face="monospace"&gt;&amp;nbsp;--driver-memory 1g \ # Adjust memory as needed&amp;nbsp;&lt;/FONT&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;FONT face="monospace"&gt;--executor-memory 1g \ # Adjust memory as needed&amp;nbsp;&lt;/FONT&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;FONT face="monospace"&gt;--num-executors 2 \ # Adjust executors as needed-&lt;/FONT&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;FONT face="monospace"&gt;&lt;STRONG&gt;-conf "spark.yarn.dist.archives"=${pyspark_venv}.tar.gz#pyspark_venv \&lt;/STRONG&gt;&lt;/FONT&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;FONT face="monospace"&gt;$CODE_DIRECTORY_CLOUD/my_application_entry_point.py # Path to your main application script&lt;/FONT&gt;&lt;P&gt;&lt;SPAN&gt;&lt;FONT face="arial, sans-serif"&gt;Explanation:&lt;/FONT&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;&lt;FONT face="arial, sans-serif"&gt;--conf "spark.yarn.dist.archives"=${pyspark_venv}.tar.gz#pyspark_venv&lt;/FONT&gt;&lt;FONT face="Google Sans, Helvetica Neue, sans-serif"&gt;:&lt;/FONT&gt;&lt;FONT face="arial, sans-serif"&gt;&amp;nbsp;This configures Spark to distribute your virtual environment archiv&lt;/FONT&gt;&lt;FONT face="arial, sans-serif"&gt;e&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;&lt;/FONT&gt;&lt;FONT face="Google Sans, Helvetica Neue, sans-serif"&gt;(&lt;/FONT&gt;&lt;FONT face="arial, sans-serif"&gt;pyspark_venv.tar.gz&lt;FONT&gt;)&lt;/FONT&gt;&lt;FONT&gt;&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;to the Yarn cluster nodes.&amp;nbsp;The&amp;nbsp;&lt;/FONT&gt;#pyspark_venv&amp;nbsp; part defines&lt;FONT&gt;&amp;nbsp;a symbolic link name within the container.&lt;/FONT&gt;&lt;/FONT&gt;&lt;/LI&gt;&lt;LI&gt;&lt;FONT face="arial, sans-serif"&gt;You do not need&amp;nbsp;--py-fileshere because the virtual environment archive will contain all necessary dependencies.&lt;/FONT&gt;&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;&lt;SPAN&gt;&lt;FONT face="arial, sans-serif"&gt;Choosing the best approach:&lt;/FONT&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;FONT face="arial, sans-serif"&gt;The choice depends on your project setup:&lt;/FONT&gt;&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;&lt;FONT face="arial, sans-serif"&gt;No Separate Virtual Environment:&amp;nbsp;Use&amp;nbsp; --py-files&amp;nbsp;if your application code consists mainly of Python files and doesn't require a separate virtual environment.&lt;/FONT&gt;&lt;/LI&gt;&lt;LI&gt;&lt;FONT face="arial, sans-serif"&gt;Separate Virtual Environment:&amp;nbsp;Use --conf spark.yarn.dist.archives&amp;nbsp;if you manage dependencies in a separate virtual environment archive.&lt;/FONT&gt;&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;&lt;FONT face="arial, sans-serif"&gt;HTH&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;&lt;/FONT&gt;&lt;/P&gt;&lt;P&gt;Mich Talebzadeh,&lt;/P&gt;&lt;P&gt;Dad | Technologist | Solutions Architect | Engineer&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;London&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;United Kingdom&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;/DIV&gt;</description>
      <pubDate>Wed, 06 Mar 2024 19:36:36 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/spark-submit-not-reading-one-of-my-py-files-arguments/m-p/62783#M32068</guid>
      <dc:creator>MichTalebzadeh</dc:creator>
      <dc:date>2024-03-06T19:36:36Z</dc:date>
    </item>
    <item>
      <title>Re: Spark submit - not reading one of my --py-files arguments</title>
      <link>https://community.databricks.com/t5/data-engineering/spark-submit-not-reading-one-of-my-py-files-arguments/m-p/62785#M32069</link>
      <description>&lt;P&gt;OK this one is for k8s for Google cloud. However, you can adjust it to any cloud vendor&lt;/P&gt;&lt;P&gt;I use zip file personally and pass the application name (in your case main.py) as the last input line like below&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;APPLICATION is your main.py. It does not need to be called main.py. It could be anything like&amp;nbsp; testpython.py&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;CODE_DIRECTORY_CLOUD="gs://spark-on-k8s/codes"&amp;nbsp; &amp;nbsp;## replace gs with s3&lt;/P&gt;&lt;P&gt;# zip needs to be done at root directory of code&lt;/P&gt;&lt;P&gt;zip -rq ${source_code}.zip ${source_code}&lt;/P&gt;&lt;P&gt;gsutil cp ${source_code}.zip $CODE_DIRECTORY_CLOUD&amp;nbsp; ## replace gsutil with aws s3&lt;/P&gt;&lt;P&gt;gsutil cp /${source_code}/src/${APPLICATION} $CODE_DIRECTORY_CLOUD&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;your spark job&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;spark-submit --verbose \&lt;/P&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;--properties-file ${property_file} \&lt;/P&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;--master k8s://https://$KUBERNETES_MASTER_IP:443 \&lt;/P&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;--deploy-mode cluster \&lt;/P&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;--name $APPNAME \&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;STRONG&gt;&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;--py-files $CODE_DIRECTORY_CLOUD/spark_on_gke.zip \&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;&amp;nbsp;&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;&lt;/STRONG&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;--conf spark.kubernetes.namespace=$NAMESPACE \&lt;/P&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;--conf spark.network.timeout=300 \&lt;/P&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;--conf spark.kubernetes.allocation.batch.size=3 \&lt;/P&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;--conf spark.kubernetes.allocation.batch.delay=1 \&lt;/P&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;--conf spark.kubernetes.driver.container.image=${IMAGEDRIVER} \&lt;/P&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;--conf spark.kubernetes.executor.container.image=${IMAGEDRIVER} \&lt;/P&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;--conf&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;&lt;A href="http://spark.kubernetes.driver.pod.name/" target="_blank" rel="noopener"&gt;spark.kubernetes.driver.pod.name&lt;/A&gt;=$APPNAME \&lt;/P&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;--conf spark.kubernetes.authenticate.driver.serviceAccountName=spark-bq \&lt;/P&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;--conf spark.driver.extraJavaOptions="-Dio.netty.tryReflectionSetAccessible=true" \&lt;/P&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;--conf spark.executor.extraJavaOptions="-Dio.netty.tryReflectionSetAccessible=true" \&lt;/P&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;--conf spark.dynamicAllocation.enabled=true \&lt;/P&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;--conf spark.dynamicAllocation.shuffleTracking.enabled=true \&lt;/P&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;--conf spark.dynamicAllocation.shuffleTracking.timeout=20s \&lt;/P&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;--conf spark.dynamicAllocation.executorIdleTimeout=30s \&lt;/P&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;--conf spark.dynamicAllocation.cachedExecutorIdleTimeout=40s \&lt;/P&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;--conf spark.dynamicAllocation.minExecutors=0 \&lt;/P&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;--conf spark.dynamicAllocation.maxExecutors=20 \&lt;/P&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;--conf spark.driver.cores=3 \&lt;/P&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;--conf spark.executor.cores=3 \&lt;/P&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;--conf spark.driver.memory=1024m \&lt;/P&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;--conf spark.executor.memory=1024m \&lt;/P&gt;&lt;P&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp;&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;&lt;STRONG&gt;&amp;nbsp; &amp;nbsp;$CODE_DIRECTORY_CLOUD/${APPLICATION}&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;HTH&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Mich Talebzadeh,&lt;/P&gt;&lt;P&gt;Dad | Technologist | Solutions Architect | Engineer&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;London&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;United Kingdom&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="MichTalebzadeh_0-1709754798807.jpeg" style="width: 400px;"&gt;&lt;img src="https://community.databricks.com/t5/image/serverpage/image-id/6526i5D1D6A526F3C8EF6/image-size/medium/is-moderation-mode/true?v=v2&amp;amp;px=400" role="button" title="MichTalebzadeh_0-1709754798807.jpeg" alt="MichTalebzadeh_0-1709754798807.jpeg" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 06 Mar 2024 19:57:36 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/spark-submit-not-reading-one-of-my-py-files-arguments/m-p/62785#M32069</guid>
      <dc:creator>MichTalebzadeh</dc:creator>
      <dc:date>2024-03-06T19:57:36Z</dc:date>
    </item>
  </channel>
</rss>

