Looking at one of the other running job (that should have ended by now), I
don't see the notify:
# cat /var/spool/ge/qmaster/job_scripts/12923 | fgrep notify
# qstat| grep 12923
12923 0.50500 dna.pmf_15 amentes r 10/24/2012 18:59:08
[email protected] 1
On 10/30/2012 04:18 PM, Reuti wrote:
Am 31.10.2012 um 00:13 schrieb Joseph Farran:
At first, I only had the hard wall clock "h_rt", but a while ago I also added
the soft one:
Here are all of the related fields:
# qconf -sq free2 | egrep "rt|notify|terminate"
shell_start_mode posix_compliant
starter_method NONE
terminate_method NONE
notify 00:00:60
s_rt 96:00:00
h_rt 96:00:00
Notify is set to 60, but I don't know what this does.
Were they also submitted with -notify? There was (is) an issue if both warnings
by s_rt and -notify are requested. The warning to the job are send every 90
seconds but it's never getting killed.
-- Reuti
On 10/30/2012 04:06 PM, Reuti wrote:
Am 31.10.2012 um 00:03 schrieb Joseph Farran:
The strace shows job running ok: doing work and then writing to a file.
I was able to kill the jobs ( 1-core each ) just fine with "kill -9".
Looking at the qmaster log after a few minutes said:
10/30/2012 15:58:41|worker|hpc|I|removing trigger to terminate job 12960.1
10/30/2012 15:58:41|worker|hpc|I|job 12960.1 finished on host
compute-12-22.local
10/30/2012 15:58:41|worker|hpc|I|removing trigger to terminate job 12959.1
10/30/2012 15:58:41|worker|hpc|I|job 12959.1 finished on host
compute-12-22.local
Did you define s_rt and -notify too?
-- Reuti
So GE cleared out the jobs ok. Not sure why the node sge is not killing
correctly.
Oh well, thanks Reuti. I will keep playing with this...
On 10/30/2012 03:53 PM, Reuti wrote:
Am 30.10.2012 um 23:45 schrieb Joseph Farran:
No:
# qconf -sq free2 | fgrep terminate
terminate_method NONE
Is the process still doing something serious or hanging somewhere in a loop:
$ strace -p 1234
and 1234 is the pid of the process on the node (you have to be root or owner of
the process).
Afterwards: is a kill -9 1234 by hand successful?
-- Reuti
On 10/30/2012 03:07 PM, Reuti wrote:
Mmh, was the terminate method redefined in the queue configuration of the queue
in question?
Am 30.10.2012 um 23:04 schrieb Joseph Farran:
No, still no cigar.
# cat /var/spool/ge/compute-12-22/messages | grep wall
#
Here is what is strange.
Some jobs do get killed just fine. One job that just went over the time limit
on another queue, GE killed it and here is the log:
10/30/2012 14:32:06| main|compute-1-7|I|registered at qmaster host "hpc.local"
10/30/2012 14:32:06| main|compute-1-7|I|Reconnected to qmaster - enabled
delayed job reporting period
10/30/2012 14:42:04| main|compute-1-7|I|Delayed job reporting period finished
10/30/2012 14:57:35| main|compute-1-7|W|job 12730.1 exceeded hard wallclock
time - initiate terminate method
10/30/2012 14:57:36| main|compute-1-7|I|SIGNAL jid: 12730 jatask: 1 signal:
KILL
On 10/30/2012 03:00 PM, Reuti wrote:
Sorry, should be like:
10/30/2012 22:59:50| main|pc15370|W|job 5281.1 exceeded hard wallclock time -
initiate terminate method
Am 30.10.2012 um 22:57 schrieb Joseph Farran:
Did not have loglevel set to log_info, so I updated it, restarted GE on the
master and softstop and start on the compute node.
I got a lot more log information now, but still no cigar:
# cat /var/spool/ge/compute-12-22/messages | fgrep h_rt
#
Checked a few other compute nodes as well for the "h_rt" and nothing either.
On 10/30/2012 01:49 PM, Reuti wrote:
Am 30.10.2012 um 20:18 schrieb Joseph Farran:
Here is one case:
qstat| egrep "12959|12960"
12959 0.50500 dna.pmf_17 amentes r 10/24/2012 18:59:12
[email protected] 1
12960 0.50500 dna.pmf_17 amentes r 10/24/2012 18:59:12
[email protected] 1
On compute-12-22:
compute-12-22 ~]# ps -e f -o ruid,euid,rgid,egid,stat,command --cols=500
0 570 0 201 Sl /data/hpc/ge/bin/lx-amd64/sge_execd
0 0 0 0 S \_ /bin/bash
/data/hpc/ge/load-sensor-cores-in-use.sh
0 570 0 201 S \_ sge_shepherd-12959 -bg
993 993 115 115 Ss | \_ -bash
/var/spool/ge/compute-12-22/job_scripts/12959
993 993 115 115 Rs | \_ ./pcharmm32
0 570 0 201 S \_ sge_shepherd-12960 -bg
993 993 115 115 Ss \_ -bash
/var/spool/ge/compute-12-22/job_scripts/12960
993 993 115 115 Rs \_ ./pcharmm32
Good, then: do you see any remark about the h_rt being exceeded in the messages
file of the host $SGE_ROOT/default/spool/compute-12-22/messages
I.e.:
$ qconf -sconf
...
loglevel log_info
is set?
-- Reuti
On 10/30/2012 12:07 PM, Reuti wrote:
Am 30.10.2012 um 20:02 schrieb Joseph Farran:
Hi Reuti.
Yes, I had that already set:
qconf -sconf|fgrep execd_params
execd_params ENABLE_ADDGRP_KILL=TRUE
What is strange is that 1 out of 10 jobs or so do get killed just fine when
they go past the hard wall time clock.
However, the majority of the jobs are not being killed when they go past their
wall time clock.
How can I investigate this further?
ps -e f -o ruid,euid,rgid,egid,stat,command --cols=500
(f w/o -) and post the relevant lines of the application please.
-- Reuti
On 10/30/2012 11:44 AM, Reuti wrote:
Hi,
Am 30.10.2012 um 19:31 schrieb Joseph Farran:
I google this issue but did not see much help on the subject.
I have several queues with hard wall clock limits like this one:
# qconf -sq queue | grep h_rt
h_rt 96:00:00
I am running Son of Grid engine 8.1.2 and many jobs run past the hard wall
clock limit and continue to run.
Look at GE qmaster logs, I see dozens and dozens of these entries:
10/30/2012 11:23:10|schedu|hpc|W|job 13179.1 should have finished since
42318s
Maybe they jumped out of the process tree (usually jobs are killed by `kill -9
-- -pgrp`. You can kill them by their additional group id, which is attached to
all started processes even if the executed something like `setsid`:
$ qconf -sconf
...
execd_params ENABLE_ADDGRP_KILL=TRUE
If it's still not working, we have to investigate the process tree.
HTH - Reuti
These entries correspond to the running jobs that should have ended 96 hours
ago, but they keep on running.
Why is GE not killing these jobs correctly when they run past the 96 hour limit
but yet complains they should have ended?
_______________________________________________
users mailing list
[email protected]
https://gridengine.org/mailman/listinfo/users
_______________________________________________
users mailing list
[email protected]
https://gridengine.org/mailman/listinfo/users