No:
# qconf -sq free2 | fgrep terminate
terminate_method NONE
On 10/30/2012 03:07 PM, Reuti wrote:
Mmh, was the terminate method redefined in the queue configuration of the queue
in question?
Am 30.10.2012 um 23:04 schrieb Joseph Farran:
No, still no cigar.
# cat /var/spool/ge/compute-12-22/messages | grep wall
#
Here is what is strange.
Some jobs do get killed just fine. One job that just went over the time limit
on another queue, GE killed it and here is the log:
10/30/2012 14:32:06| main|compute-1-7|I|registered at qmaster host "hpc.local"
10/30/2012 14:32:06| main|compute-1-7|I|Reconnected to qmaster - enabled
delayed job reporting period
10/30/2012 14:42:04| main|compute-1-7|I|Delayed job reporting period finished
10/30/2012 14:57:35| main|compute-1-7|W|job 12730.1 exceeded hard wallclock
time - initiate terminate method
10/30/2012 14:57:36| main|compute-1-7|I|SIGNAL jid: 12730 jatask: 1 signal:
KILL
On 10/30/2012 03:00 PM, Reuti wrote:
Sorry, should be like:
10/30/2012 22:59:50| main|pc15370|W|job 5281.1 exceeded hard wallclock time -
initiate terminate method
Am 30.10.2012 um 22:57 schrieb Joseph Farran:
Did not have loglevel set to log_info, so I updated it, restarted GE on the
master and softstop and start on the compute node.
I got a lot more log information now, but still no cigar:
# cat /var/spool/ge/compute-12-22/messages | fgrep h_rt
#
Checked a few other compute nodes as well for the "h_rt" and nothing either.
On 10/30/2012 01:49 PM, Reuti wrote:
Am 30.10.2012 um 20:18 schrieb Joseph Farran:
Here is one case:
qstat| egrep "12959|12960"
12959 0.50500 dna.pmf_17 amentes r 10/24/2012 18:59:12
[email protected] 1
12960 0.50500 dna.pmf_17 amentes r 10/24/2012 18:59:12
[email protected] 1
On compute-12-22:
compute-12-22 ~]# ps -e f -o ruid,euid,rgid,egid,stat,command --cols=500
0 570 0 201 Sl /data/hpc/ge/bin/lx-amd64/sge_execd
0 0 0 0 S \_ /bin/bash
/data/hpc/ge/load-sensor-cores-in-use.sh
0 570 0 201 S \_ sge_shepherd-12959 -bg
993 993 115 115 Ss | \_ -bash
/var/spool/ge/compute-12-22/job_scripts/12959
993 993 115 115 Rs | \_ ./pcharmm32
0 570 0 201 S \_ sge_shepherd-12960 -bg
993 993 115 115 Ss \_ -bash
/var/spool/ge/compute-12-22/job_scripts/12960
993 993 115 115 Rs \_ ./pcharmm32
Good, then: do you see any remark about the h_rt being exceeded in the messages
file of the host $SGE_ROOT/default/spool/compute-12-22/messages
I.e.:
$ qconf -sconf
...
loglevel log_info
is set?
-- Reuti
On 10/30/2012 12:07 PM, Reuti wrote:
Am 30.10.2012 um 20:02 schrieb Joseph Farran:
Hi Reuti.
Yes, I had that already set:
qconf -sconf|fgrep execd_params
execd_params ENABLE_ADDGRP_KILL=TRUE
What is strange is that 1 out of 10 jobs or so do get killed just fine when
they go past the hard wall time clock.
However, the majority of the jobs are not being killed when they go past their
wall time clock.
How can I investigate this further?
ps -e f -o ruid,euid,rgid,egid,stat,command --cols=500
(f w/o -) and post the relevant lines of the application please.
-- Reuti
On 10/30/2012 11:44 AM, Reuti wrote:
Hi,
Am 30.10.2012 um 19:31 schrieb Joseph Farran:
I google this issue but did not see much help on the subject.
I have several queues with hard wall clock limits like this one:
# qconf -sq queue | grep h_rt
h_rt 96:00:00
I am running Son of Grid engine 8.1.2 and many jobs run past the hard wall
clock limit and continue to run.
Look at GE qmaster logs, I see dozens and dozens of these entries:
10/30/2012 11:23:10|schedu|hpc|W|job 13179.1 should have finished since
42318s
Maybe they jumped out of the process tree (usually jobs are killed by `kill -9
-- -pgrp`. You can kill them by their additional group id, which is attached to
all started processes even if the executed something like `setsid`:
$ qconf -sconf
...
execd_params ENABLE_ADDGRP_KILL=TRUE
If it's still not working, we have to investigate the process tree.
HTH - Reuti
These entries correspond to the running jobs that should have ended 96 hours
ago, but they keep on running.
Why is GE not killing these jobs correctly when they run past the 96 hour limit
but yet complains they should have ended?
_______________________________________________
users mailing list
[email protected]
https://gridengine.org/mailman/listinfo/users
_______________________________________________
users mailing list
[email protected]
https://gridengine.org/mailman/listinfo/users