When the GVCF combiner is running, I can monitor the job progress in Spark UI.
I am wondering if I can see the job log in the same way after terminating the cluster which Hail was run.
For jobs like this, I’d recommend using hailctl dataproc submit (which wraps gcloud dataproc submit) since jupyter notebooks have problems with reconnecting to see execution status.
Thank you for your reply. I will try to use ‘submit’ instead of ‘notebook’. If I use ‘submit’, can I see the job log using Spark UI, even after terminating the cluster? What I want is to analyze the executors’ info after the function run is done and the cluster is stopped. I would like to know if it is possible or not.
Thank you for your answer. I am trying to do joint-calling for 1000 GVCFs using Hail GVCF combiner. If it can be run successfully in our conditions, we may extend the number of GVCFs. In this work, my first step is to find out the parameters in Hail function, Spark, and Dataproc for optimizing the runtime and cost of Hail GVCF combiner function for a GVCF small set. I would like to keep track of all job logs that depend on parameter changes. When I analyze the results, I want to stop the cluster for saving the cost, because I only need to see the completed job results. I hope it could be the answer of your question. If you can give me some advice related to my work, it would be welcome. Thank you.
You can use hailctl dataproc modify --num-preemptible-workers 0 to shrink the cluster. The minimal cluster is 2 non-preemptible (regular) workers and one leader. That should cost very little money per hour and give you plenty of time to analyze the logs.
I doubt you’ll find much useful information in the worker logs.
Maybe @tpoterba or @chrisvittal can provide some information on recommended worker configurations.