We will wait then for now. Be we look forward to the new version.
-Paul Edmon-
On 10/17/2013 1:56 PM, Danny Auble wrote:
You could install 13.12 on top of 2.6 today, but it isn't considered
stable, the RPCs may change from now until it is officially tagged.
Currently there is a cutoff on Nov 15 for new code. Until then RPCs
could change which could make job loss possible. We would not
recommend running any production system on 13.12 at this moment.
If you did want to fly by the seat of you pants you should be able to
update 2.6 to 13.12 today with no job loss, but any future version of
13.12 doesn't offer this gaurentee. Only after offical tags would
this be considered safe.
On 10/17/13 10:49, Paul Edmon wrote:
Is there an upgrade path from 2.6 to 13.12?
-Paul Edmon-
On 10/17/2013 1:34 PM, Danny Auble wrote:
There was a modification to the way lists were sorted in 13.12. This
could make a massive difference with the speed of squeue on your
system. You could backport the change to 2.6 but it took a few
revisions and might not be that easy to fully backport.
On 10/17/13 10:30, Paul Edmon wrote:
# TIMERS
#BatchStartTimeout=10
#CompleteWait=0
#EpilogMsgTime=2000
#GetEnvTimeout=2
HealthCheckInterval=500
HealthCheckProgram=/usr/local/bin/node_monitor.sh
InactiveLimit=0
KillWait=30
MessageTimeout=100
#ResvOverRun=0
MinJobAge=300
#OverTimeLimit=0
SlurmctldTimeout=120
SlurmdTimeout=300
#UnkillableStepTimeout=60
#VSizeFactor=0
Waittime=0
We have alot of churn in our system. We are up to job ID 2.3
million and that's after roughly 2 months. So we do about 40,000
jobs per day. It gets particularly bad when there are a ton of
jobs being submitted or exiting at the same time as the ctld is
busy. Plus if it hits a blocking portion of code it can take a
bit. We don't typically see time outs but when we do it is due to
high load on the ctld.
[root@holy-slurm01 slurm]# sdiag
*******************************************************
sdiag output at Thu Oct 17 13:26:07 2013
Data since Thu Oct 17 05:17:09 2013
*******************************************************
Server thread count: 5
Agent queue size: 0
Jobs submitted: 7262
Jobs started: 25457
Jobs completed: 22316
Jobs canceled: 112
Jobs failed: 3
Main schedule statistics (microseconds):
Last cycle: 536974
Max cycle: 9358430
Total cycles: 576
Mean cycle: 3065748
Mean depth cycle: 9151
Cycles per minute: 1
Last queue length: 4642
Backfilling stats (WARNING: data obtained in the middle of
backfilling execution
Total backfilled jobs (since last slurm start): 16607
Total backfilled jobs (since last stats cycle start): 16607
Total cycles: 20
Last cycle when: Thu Oct 17 13:13:38 2013
Last cycle: 140076614
Max cycle: 2722993429
Mean cycle: 162420765
Last depth cycle: 226
Last depth cycle (try sched): 226
Depth Mean: 852
Depth Mean (try depth): 852
Last queue length: 4790
Queue length mean: 5498
On 10/17/2013 1:24 PM, Danny Auble wrote:
I am surprised you are still having timeouts. I would expect the
longest anyone waits is 2 seconds. What is your MessageTimeout
set to?
Danny
On 10/17/13 10:21, Paul Edmon wrote:
True. I was just contemplating ways to make it more responsive.
Multiple copies of the data would do that, I just wasn't sure
whether keeping that in sync would be a head ache.
-Paul Edmon-
On 10/17/2013 1:01 PM, Moe Jette wrote:
Sending old data quickly seems very dangerous, especially if
there are scripts submitting jobs and then running squeue to
look for them.
Quoting Paul Edmon <[email protected]>:
Another way is to use the showq script that we have been
working on:
https://github.com/fasrc/slurm_showq
That gives over all statistics as well. However sdiag is a
great way to see if the system is running properly and get a
good view of it.
I will note that these sorts of queries tend to hang when the
ctld is busy. We were discussing it in our group meeting
yesterday. It might be good to have the ctld set up a thread
dedicated to servicing these requests in a timely manner. When
our users see time out messages or slow response they freak out
and send in help tickets. To me diagnostic checks like this
should respond in a timely manner, even if the data that is
contained there is a little out of date.
-Paul Edmon-
On 10/17/2013 11:57 AM, Moe Jette wrote:
There is no faster way to get job counts, but you might find
the sdiag command helpful.
Quoting Damien François <[email protected]>:
Hello,
what is the most efficient way of finding how many jobs are
currently running, pending, etc in the system ?
At the moment, I use squeue and wc -l but that sometimes gets
slow.
Is there a command/flag I would have missed that would output
that information quickly ?
Thanks
damien=