Perhaps teaching the users about job arrays would help a lot in this
situation. They could submit all 20-30k jobs with only 1 sbatch
command. It is much more efficient for the scheduler and would probably
eliminate almost all the timeout issues you are seeing.
On 07/14/15 08:42, Charles Johnson wrote:
slurm 14.11.7
cgroups implemented
backfill implemented
We have a small cluster -- ~650 nodes and ~6500 processors. We are
looking for ways to lessen the impact of a busy scheduler for users
who submit jobs with an automated submission process. Their job
monitoring will fail with:
squeue: error: slurm_receive_msg: Socket timed out on send/recv operation
slurm_load_jobs error: Socket timed out on send/recv operation
We are using back-fill:
SchedulerParameters=bf_interval=120,bf_continue,bf_resolution=300,bf_max_job_test=2000,bf_max_job_user=100,max_sched_time=2
Our cluster generally has numerous small, single-core; and when a user
submits 20,000 or 30,000 jobs the system can fail to respond to
squeue, or even sbatch.
One user has suggested we write a wrapper for certain commands, like
squeue, which auto re-try when such messages are returned. This
doesn't seem like the appropriate "fix." IMHO, a better approach would
be to "fix" the submission systems that some users have.
Are there other who have faced this issue? I have thought about
caching the output to squeue in a file, refreshing the file in a
timely way, and pointing an squeue wrapper to return that; but again
that doesn't seem like a good approach.
Any suggestions would be great.
Charles