Hi, we encountered a problem of timeout every (60 / 80 minutes ) on a long execution of copy json into parquet file in datalake Gen 2.
Options
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
11-11-2022 02:06 AM
First, our process was triggered by the datafactory. First the connexion was set with token access, then with managed service identity.
We prove the untimely time out was not due to the datafactory by running directly the notebook.
Secondly, we tried running the notebook alone and scale up the inactivity time before shutdown.
But none of our tries lead to a success, we have had to scale up to 16 workers and restart every hour our pipeline to success. We did this for our integration plan but we refuse to face it again for our production plan.
Could you please, suggest to us some good advises.
You will find attached the executing code
Kind Regards,
Natacha
Labels:
- Labels:
-
Databricks notebook
-
Timeout
0 REPLIES 0

