Reuti <[email protected]> writes: > If you have this often, another way to have more flexibility could be > a parallel file system, so that any node can access the scratch > data. Then you could split the jobs without any special setup.
[We've had reasonable success with ~100 nodes just with NFS over 10Gbe to a Sun x4500. It works better than an old Panasas system we had.] > For now: the step A of the job could use a `qalter` to add a "-l h=$HOSTNAME" > to step B of this flow of steps. (All exechosts need to be submission hosts > this way though). Besides a jobnumber, `qalter` can also you a job name which > could make it easier to target the following step when you name the jobs > properly. It occurs to me that this might be a use for something like per-user complexes <https://arc.liv.ac.uk/trac/SGE/ticket/1387>. I'd be interested in suggestions for implementation of something like that or information on what LSF etc. provide. _______________________________________________ users mailing list [email protected] https://gridengine.org/mailman/listinfo/users
