How to create a databricks job with parameters via CLI?
Options
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
06-24-2021 08:52 AM
I'm creating a new job in databricks using the databricks-cli:
databricks jobs create --json-file ./deploy/databricks/config/job.config.json
With the following json:
{
"name": "Job Name",
"new_cluster": {
"spark_version": "4.1.x-scala2.11",
"node_type_id": "Standard_D3_v2",
"num_workers": 3,
"spark_env_vars": {
"PYSPARK_PYTHON": "/databricks/python3/bin/python3"
}
},
"libraries": [
{
"maven": {
"coordinates": "com.microsoft.sqlserver:mssql-jdbc:6.5.3.jre8-preview"
}
}
],
"timeout_seconds": 3600,
"max_retries": 3,
"schedule": {
"quartz_cron_expression": "0 0 22 ? * *",
"timezone_id": "Israel"
},
"notebook_task": {
"notebook_path": "/notebooks/python_notebook"
}
}
And I want to add parameters that will be accessible in the notebook via:
dbutils.widgets.text("argument1", "<default value>")
dbutils.widgets.get("argument1")
Labels:
- Labels:
-
Cli
-
Databricks Job
-
DatabricksJobs
-
New Job
1 REPLY 1
Options
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
10-12-2023 09:37 AM - edited 10-12-2023 09:37 AM
This is an old post but still relevant for future readers, so will answer how it is done. You need to add base_parameters flag in the notebook_task config, like the following.
"notebook_task": {
"notebook_path": "...",
"base_parameters": {
"argument1": "value"
}
}

