Hi,
We have a scenario where we need to deploy 15 spark streaming applications on databricks reading from kafka to single Job cluster.
We tried following approach:
1. create job 1 with new job cluster (C1)
2. create job2 pointing to C1
...
3. create job15 pointing to C1
But, the problem here is if the job 1 fails, it is terminating all the other 14 jobs.
One of the options we are thinking is to have a ***** kafka topic with no messages in it and ***** spark streaming job reading from ***** kafka topic (which will never fail 99.99%) which create new job cluster (C1) and rest of the 15 jobs will point to C1. We are assuming Job cluster C1 will never fail 99.99%.
Other solution we have is to create each job cluster for each job (15 Clusters for 15 jobs ) but it is going to kill our operational costs as it is continuous streaming job and some of topics have very less volume.
Could you please advice on how to address this issue.
Thanks
Jin.