Am 27.04.2012 um 01:29 schrieb Rayson Ho: > Hi Hung-Sheng, > > I also read that email... But I did not respond to it as I think > others have already answered it: the best way to fix it is to change > the limit - SGE in general does not support interactive batch jobs > that disconnect & re-connect at will.
Neverthless, if the admin increases the wallclock time, one could start the tool `screen` and inside the `screen` session issue a `qlogin`. To this `screen` session you can detach and reattach from whereever and whenever you like (supposing it's issued an always switched on headnode of the cluster). http://linux.die.net/man/1/screen Most important: "CTRL-A d" to detach, and later start it with `screen -xRR` to get the session back. -- Reuti > Rayson > > > On Thu, Apr 26, 2012 at 7:20 PM, "Hung-Sheng Tsao (LaoTsao 老曹) Ph. D." > <[email protected]> wrote: >> if this is SGE related >> >> >> >> Hi, >> >> I've been assigned a debugging task on a Rocks 5.4.3 cluster (I helped >> build built the software, but on OSX). I know only the very basics of >> using the cluster, like making sure I qlogin to run something instead of >> doing it on the head node, and that I can submit tasks using qsub, but >> that's about it. >> >> The app I'm debugging is segfaulting but only on the cluster, not on my >> Mac, and it will take 20+ hours to segfault judging from the current >> rate. I'm currently running it in gdb via a qlogin session. >> >> My question is wheter there's a way to start an interactive session, >> then suspend the qlogin session without ending the interactive job >> itself, and then and reattach to the job later on to debug once the >> segfault has happened. The main reason for doing this is that the >> cluster admin has set a 24-hour limit on qlogin sessions. Also, it'd be >> easier to not have to maintain my terminal connection, but that's not a >> big issue. >> >> Or are there other ways to accomplish this? Many thanks for any help? >> >> Cheers, >> Michael >> >> >> _______________________________________________ >> users mailing list >> [email protected] >> https://gridengine.org/mailman/listinfo/users > > _______________________________________________ > users mailing list > [email protected] > https://gridengine.org/mailman/listinfo/users > _______________________________________________ users mailing list [email protected] https://gridengine.org/mailman/listinfo/users
