Reuti <[email protected]> writes:

> If you have this often, another way to have more flexibility could be
> a parallel file system, so that any node can access the scratch
> data. Then you could split the jobs without any special setup.

[We've had reasonable success with ~100 nodes just with NFS over 10Gbe
to a Sun x4500.  It works better than an old Panasas system we had.]

> For now: the step A of the job could use a `qalter` to add a "-l h=$HOSTNAME" 
> to step B of this flow of steps. (All exechosts need to be submission hosts 
> this way though). Besides a jobnumber, `qalter` can also you a job name which 
> could make it easier to target the following step when you name the jobs 
> properly.

It occurs to me that this might be a use for something like per-user
complexes <https://arc.liv.ac.uk/trac/SGE/ticket/1387>.  I'd be
interested in suggestions for implementation of something like that or
information on what LSF etc. provide.
_______________________________________________
users mailing list
[email protected]
https://gridengine.org/mailman/listinfo/users

Reply via email to