10-26-2022 02:42 AM
Hi,
there is no option to take VMs from a Pool for a new workflow (Azure Cloud)?
default schema for a new cluster:
{
"num_workers": 0,
"spark_version": "10.4.x-scala2.12",
"spark_conf": {
"spark.master": "local[*, 4]",
"spark.databricks.cluster.profile": "singleNode"
},
"azure_attributes": {
"first_on_demand": 1,
"availability": "ON_DEMAND_AZURE",
"spot_bid_max_price": -1
},
"node_type_id": "Standard_DS3_v2",
"ssh_public_keys": [],
"custom_tags": {
"ResourceClass": "SingleNode"
},
"spark_env_vars": {
"PYSPARK_PYTHON": "/databricks/python3/bin/python3"
},
"enable_elastic_disk": true,
"cluster_source": "JOB",
"init_scripts": [],
"data_security_mode": "SINGLE_USER",
"runtime_engine": "STANDARD"
}
10-26-2022 06:52 AM
@Michal Mlaka I just checked on the UI and I could find the pools listing under worker type in a job cluster configuration. It should work.
10-26-2022 04:30 AM
Hi @Michal Mlaka , it should be possible. If you are using APIs to launch the Job, remove "node_type_id" and "driver_node_type_id" from the JSON and pass the "instance_pool_id" and "driver_instance_pool_id" instead to make the cluster pick VMs from the pool.
Check this doc for the structure of cluster definition-
Let me know if this helps. Please mark the comment as "best answer" to resolve the query.
10-26-2022 04:38 AM
I will try. I started with UI and could not find that option there. Thanks.
10-26-2022 06:52 AM
@Michal Mlaka I just checked on the UI and I could find the pools listing under worker type in a job cluster configuration. It should work.
10-26-2022 07:45 AM
ooou. sorry. yea its there 🙈 Thank you.
Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.
If there isn’t a group near you, start one and help create a community that brings people together.
Request a New Group