I’m having trouble setting a limit to the memory Hail uses when running it locally on a server. I have set spark.driver.memory, spark.executor.pyspark.memory and spark.executor.memory:
As a direct answer to your question, you cannot limit the memory used by Hail. That said, using 1 terabyte of RAM is probably a memory leak. We’ll continue to investigate.
You have to update to 0.2.47 to get the fix, which just released this morning (it seems you posted your update about the problem not being fixed before the release came out)
The update indeed solves it, thank you very much! Previously I built from source on github (version 0.2.46-3a514a199ccd) after seeing your post - perhaps I still missed out something.
Sorry, I got confused. We made two memory fixes in the span of two days and I forgot which fix we were talking about. It’s true that 0.2.46-3a514a199ccd would have contained the fix I intended for you. However, seems like you also needed fix https://github.com/hail-is/hail/pull/9009.
Again, sorry for confusion, but glad things are working now!