Hi all,

we have a 300-core cluster with a ~150Tb shared directory (GPFS). Our users run some genomic analysis that use huge files and usually cannot fit the 500Gb internal HDD of the nodes. As you can imagine, sometimes things get pretty intense and all the nagios disk alarms start going off (the disk "works" but we got 10+ sec timeouts).

Knowing that I cannot trust our users to request any "disk_intensive" parameter/flag, I was pondering on setting a suspend_threshold in the queues, watching the shared disk status (e.g. timing an ls to the shared disk) and start suspending jobs when the disk has, say, a 3 sec delay. This would be a nice fix for our issue, but it has some problems: When there are both "IO-intensive" and "normal" jobs, and the suspend_threshold kicks in, SGE will start suspending jobs ¿without any particular criteria? (I don't know this part), and lots of innocent "normal" jobs will be suspended through all the nodes before the disk load is stabilized.

Does anyone have any idea/workaround to solve this? Or should I ignore/relax all the disk alarms?

Thanks in advance,

Txema
_______________________________________________
users mailing list
[email protected]
https://gridengine.org/mailman/listinfo/users

Reply via email to