<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: spark-submit Error &amp;quot;Unrecognized option: --executor-memory 3G&amp;quot; although  --executor-memory is available in Options. in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/spark-submit-error-quot-unrecognized-option-executor-memory-3g/m-p/32743#M23879</link>
    <description>&lt;P&gt;I will highly recommend to run your job with the default values. Then you can have a good reference point in case you would like to optimize further.  Check  your cluster utilization and Spark UI. This will help you to undertand better what is happening as your job is running&lt;/P&gt;</description>
    <pubDate>Mon, 31 Oct 2022 18:33:51 GMT</pubDate>
    <dc:creator>jose_gonzalez</dc:creator>
    <dc:date>2022-10-31T18:33:51Z</dc:date>
    <item>
      <title>spark-submit Error "Unrecognized option: --executor-memory 3G" although  --executor-memory is available in Options.</title>
      <link>https://community.databricks.com/t5/data-engineering/spark-submit-error-quot-unrecognized-option-executor-memory-3g/m-p/32738#M23874</link>
      <description>&lt;P&gt;Executed a spark-submit job through databricks cli with the following job configurations.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;PRE&gt;&lt;CODE&gt;{
  "job_id": 123,
  "creator_user_name": "******",
  "run_as_user_name": "******",
  "run_as_owner": true,
  "settings": {
    "name": "44aa-8447-c123aad310",
    "email_notifications": {},
    "max_concurrent_runs": 1,
    "tasks": [
      {
        "task_key": "4aa-8447-c90aad310",
        "spark_submit_task": {
          "parameters": [
            "--driver-memory 3G",
            "--executor-memory 3G",
            "--conf",
            "spark.hadoop.mapreduce.fileoutputcommitter.algorithm.version=2",
            "--conf",
            "spark.speculation=false",
            "--conf",
            "spark.sql.parquet.fs.optimized.committer.optimization-enabled=true",
            "--conf",
            "spark.executorEnv.JAVA_HOME=/usr/lib/jvm/jdk-11.0.1",
            "--conf",
            "spark.executor.instances=3",
            "--conf",
            "spark.network.timeout=600s",
            "--conf",
            "spark.yarn.appMasterEnv.JAVA_HOME=/usr/lib/jvm/jdk-11.0.1",
            "--conf",
            "spark.driver.maxResultSize=1g",
            "--conf",
            "spark.yarn.maxAppAttempts=1",
            "--jars",
            "/home/hadoop/somejar.jar,/home/hadoop/somejar2.jar",
            "--class",
            "we.databricks.some.path.ER",
            "/home/hadoop/some-jar-SNAPSHOT.jar",
            "'******'"
          ]
        },
        "new_cluster": {
          "spark_version": "10.4.x-scala2.12",
          "spark_conf": {
            "spark.databricks.delta.preview.enabled": "true",
            "spark.hadoop.fs.azure.account.key": "******"
          },
          "node_type_id": "Standard_DS3_v2",
          "custom_tags": {
            "application": "******",
            "name": "******",
            "environment": "******",
            "owner": "******",
            "CURRENT_VERSION": "1.20.0-ab6303d9d"
          },
          "cluster_log_conf": {
            "dbfs": {
              "destination": "******"
            }
          },
          "spark_env_vars": {
            "ENVIRONMENT": "******",
            "AZURE_ACCOUNT_KEY": "******",
            "AZURE_ACCOUNT_NAME": "******",
            "PYSPARK_PYTHON": "/databricks/python3/bin/python3",
            "JNAME": "zulu11-ca-amd64",
            "AZURE_CONTAINER_NAME": "******"
          },
          "enable_elastic_disk": true,
          "init_scripts": [
            {
              "abfss": {
                "destination": "******"
              }
            }
          ],
          "num_workers": 3
        },
        "timeout_seconds": 0
      }
    ],
    "format": "MULTI_TASK"
  },
  "created_time": 1662096418457
}&lt;/CODE&gt;&lt;/PRE&gt;&lt;P&gt;But this gives error in spark submit. Error: Unrecognized option: --executor-memory 3G&lt;/P&gt;</description>
      <pubDate>Fri, 02 Sep 2022 05:48:56 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/spark-submit-error-quot-unrecognized-option-executor-memory-3g/m-p/32738#M23874</guid>
      <dc:creator>talha</dc:creator>
      <dc:date>2022-09-02T05:48:56Z</dc:date>
    </item>
    <item>
      <title>Re: spark-submit Error "Unrecognized option: --executor-memory 3G" although  --executor-memory is available in Options.</title>
      <link>https://community.databricks.com/t5/data-engineering/spark-submit-error-quot-unrecognized-option-executor-memory-3g/m-p/32740#M23876</link>
      <description>&lt;P&gt;Not really sure if running spark on local mode. But have used alternate property &lt;/P&gt;&lt;PRE&gt;&lt;CODE&gt;spark.executor.memory&lt;/CODE&gt;&lt;/PRE&gt;&lt;P&gt;and passed it as --conf and now it works&lt;/P&gt;</description>
      <pubDate>Mon, 05 Sep 2022 10:23:31 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/spark-submit-error-quot-unrecognized-option-executor-memory-3g/m-p/32740#M23876</guid>
      <dc:creator>talha</dc:creator>
      <dc:date>2022-09-05T10:23:31Z</dc:date>
    </item>
    <item>
      <title>Re: spark-submit Error "Unrecognized option: --executor-memory 3G" although  --executor-memory is available in Options.</title>
      <link>https://community.databricks.com/t5/data-engineering/spark-submit-error-quot-unrecognized-option-executor-memory-3g/m-p/32741#M23877</link>
      <description>&lt;P&gt;Hi @Muhammad Talha Jamil​,&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;We don't recommend to change the default settings. I would like to undertand better the reason why you would like to change the default values. Are you trying to defined the Executor memory because you had an error in the past? or what would be the reason? &lt;/P&gt;</description>
      <pubDate>Fri, 09 Sep 2022 23:29:26 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/spark-submit-error-quot-unrecognized-option-executor-memory-3g/m-p/32741#M23877</guid>
      <dc:creator>jose_gonzalez</dc:creator>
      <dc:date>2022-09-09T23:29:26Z</dc:date>
    </item>
    <item>
      <title>Re: spark-submit Error "Unrecognized option: --executor-memory 3G" although  --executor-memory is available in Options.</title>
      <link>https://community.databricks.com/t5/data-engineering/spark-submit-error-quot-unrecognized-option-executor-memory-3g/m-p/32742#M23878</link>
      <description>&lt;P&gt;We are moving from aws emr to azure databricks. In emr we used to change executor memory with respect to job requirements. Wont we require that on databricks?&lt;/P&gt;</description>
      <pubDate>Mon, 12 Sep 2022 09:58:38 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/spark-submit-error-quot-unrecognized-option-executor-memory-3g/m-p/32742#M23878</guid>
      <dc:creator>talha</dc:creator>
      <dc:date>2022-09-12T09:58:38Z</dc:date>
    </item>
    <item>
      <title>Re: spark-submit Error "Unrecognized option: --executor-memory 3G" although  --executor-memory is available in Options.</title>
      <link>https://community.databricks.com/t5/data-engineering/spark-submit-error-quot-unrecognized-option-executor-memory-3g/m-p/32743#M23879</link>
      <description>&lt;P&gt;I will highly recommend to run your job with the default values. Then you can have a good reference point in case you would like to optimize further.  Check  your cluster utilization and Spark UI. This will help you to undertand better what is happening as your job is running&lt;/P&gt;</description>
      <pubDate>Mon, 31 Oct 2022 18:33:51 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/spark-submit-error-quot-unrecognized-option-executor-memory-3g/m-p/32743#M23879</guid>
      <dc:creator>jose_gonzalez</dc:creator>
      <dc:date>2022-10-31T18:33:51Z</dc:date>
    </item>
    <item>
      <title>Re: spark-submit Error "Unrecognized option: --executor-memory 3G" although  --executor-memory is available in Options.</title>
      <link>https://community.databricks.com/t5/data-engineering/spark-submit-error-quot-unrecognized-option-executor-memory-3g/m-p/32739#M23875</link>
      <description>&lt;P&gt;Hi, Thanks for reaching out to community.databricks.com. &lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;Are you running spark in local mode? &lt;/P&gt;&lt;P&gt;Please check &lt;A href="https://stackoverflow.com/questions/26562033/how-to-set-apache-spark-executor-memory" target="test_blank"&gt;https://stackoverflow.com/questions/26562033/how-to-set-apache-spark-executor-memory&lt;/A&gt;, please let us know if this helps, Also please let us know in case if you have further queries on the same. &lt;/P&gt;</description>
      <pubDate>Fri, 02 Sep 2022 21:34:56 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/spark-submit-error-quot-unrecognized-option-executor-memory-3g/m-p/32739#M23875</guid>
      <dc:creator>Debayan</dc:creator>
      <dc:date>2022-09-02T21:34:56Z</dc:date>
    </item>
  </channel>
</rss>

