- 5064 Views
- 3 replies
- 4 kudos
When I run a job enabling using spot instances , I would like to know how many number of workers are using spot and how many number of workers are using on demand instances for a given job run In order to identify the spot instances we got for any...
- 5064 Views
- 3 replies
- 4 kudos
- 5486 Views
- 2 replies
- 1 kudos
I have an always-on job cluster triggering Spark Streaming jobs. I would like to stop this streaming job once a week to run table maintenance. I was looking to leverage the foreachBatch function to check a condition and stop the job accordingly.
- 5486 Views
- 2 replies
- 1 kudos
Latest Reply
You could also use the "Available-now micro-batch" trigger. It only processes one batch at a time, and you can do whatever you want in between batches (sleep, shut down, vacuum, etc.)
1 More Replies
- 2382 Views
- 2 replies
- 0 kudos
Hi All,I am new to databricks need some understanding for my requirement .our requirement:a: we have zip file in azure blob storage and we are bringing that file to dbfs and unzip that file and executing our transformations in multiple steps (3 steps...
- 2382 Views
- 2 replies
- 0 kudos
Latest Reply
Hi @praveen rajak Does @Debayan Mukherjee response answer your question? If yes, would you be happy to mark it as best so that other members can find the solution more quickly?We'd love to hear from you.Thanks!
1 More Replies
by
kjoth
• Contributor II
- 15947 Views
- 8 replies
- 3 kudos
I'm running a scheduled job on Job clusters. I didnt mention the log location for the cluster. Where can we get the stored logs location. Yes, I can see the logs in the runs, but i need the logs location.
- 15947 Views
- 8 replies
- 3 kudos
by
dmayi
• New Contributor
- 4944 Views
- 1 replies
- 0 kudos
Hi i want to set up custom tags on an all-purpose cluster for purposes of cost break down and chargebacks. What: specifically, i want to capture JobName, JobID, UserId who ran jobI can set other custom tags such as Business Unit, Owner... However,...
- 4944 Views
- 1 replies
- 0 kudos
Latest Reply
Hey there @DIEUDONNE MAYI Does @Kaniz Fatma response answer your question? If yes, would you be happy to mark it as best so that other members can find the solution more quickly?We'd love to hear from you.Thanks!
- 9896 Views
- 7 replies
- 5 kudos
I have a job running with no issues in Databricks runtime 7.3 LTS. When I upgraded to 8.3 it fails with error An exception was thrown from a UDF: 'pyspark.serializers.SerializationError'... SparkContext should only be created and accessed on the driv...
- 9896 Views
- 7 replies
- 5 kudos
Latest Reply
Adding to @Sean Owen comments, The only reason this is working is that the optimizer is evaluating this locally rather than creating a context on executors and evaluating it.
6 More Replies
- 6575 Views
- 4 replies
- 0 kudos
Continuing the above case, does that mean if i have several like 5 ADF pipelines scheduled regularly at the same time, its better to use an existing cluster as all of the ADF pipelines would share the same cluster and hence the cost will be lower?
- 6575 Views
- 4 replies
- 0 kudos
Latest Reply
Atanu
Databricks Employee
for adf or job run we always prefer job cluster. but for streaming, you may consider using interactive cluster . but anyway you need to monitor the cluster load, if loads are high there will be chance to job slowness as well as failure. also data siz...
3 More Replies
by
Junee
• New Contributor III
- 6137 Views
- 5 replies
- 3 kudos
I am using Databeicks Job Api 2.1 to trigger and run my jobs. "jobs/runs/submit" this API helps in starting the cluster, as well as create the job and run it. This API works great for normal jobs as it also cleans the cluster once job is finished suc...
- 6137 Views
- 5 replies
- 3 kudos
Latest Reply
@Junee, Anytime! It is crisply mentioned in the doc too. https://docs.databricks.com/clusters/index.html
4 More Replies
- 2214 Views
- 1 replies
- 0 kudos
I would like to make it so users can only create job clusters and not interactive clusters. Is it possible to do this in a workspace?
- 2214 Views
- 1 replies
- 0 kudos
Latest Reply
This can be accomplished with cluster policies. You can use a policy similar to this example to restrict certain users or groups to only have permission to create job clusters.