Remi, The obvious questions are:
Have you set up the accounting? Added a cluster, added some users, etc? ie, on the link below, there's a section under "Tools" and "Database Configuration" that might apply? http://slurm.schedmd.com/accounting.html I think that this section is ripe for a how to as well - it's a very dense wall of text and could do with a quick 5-min preview. Cheers L. ------ The most dangerous phrase in the language is, "We've always done it this way." - Grace Hopper On 2 June 2016 at 15:56, remi marchal <[email protected]> wrote: > Dear slurm users, > > I am quite new in the community and I would like to monitor the running > jobs. > > Looking through internet, I found this command: > sacct -j > > However, here is the result of one of my jobs (submission script bellow): > > JobID JobIDRaw JobName Partition MaxVMSize MaxVMSizeNode > MaxVMSizeTask AveVMSize MaxRSS MaxRSSNode MaxRSSTask AveRSS > MaxPages MaxPagesNode MaxPagesTask AvePages MinCPU MinCPUNode > MinCPUTask AveCPU NTasks AllocCPUS Elapsed State ExitCode > AveCPUFreq ReqCPUFreqMin ReqCPUFreqMax ReqCPUFreqGov ReqMem > ConsumedEnergy MaxDiskRead MaxDiskReadNode MaxDiskReadTask AveDiskRead > MaxDiskWrite MaxDiskWriteNode MaxDiskWriteTask AveDiskWrite AllocGRES > ReqGRES ReqTRES AllocTRES > ------------ ------------ ---------- ---------- ---------- -------------- > -------------- ---------- ---------- ---------- ---------- ---------- > -------- ------------ -------------- ---------- ---------- ---------- > ---------- ---------- -------- ---------- ---------- ---------- -------- > ---------- ------------- ------------- ------------- ---------- > -------------- ------------ --------------- --------------- -------------- > ------------ ---------------- ---------------- -------------- ------------ > ------------ ---------- ---------- > 156 156 test debug > > > 36 00:00:15 RUNNING 0:0 > Unknown Unknown Unknown 0n > > cpu=36,no+ > cpu=36,no+ > > Submission script > > #!/bin/bash > # > #SBATCH --job-name=test > #SBATCH --output=res.txt > # > #SBATCH --tasks=36 > > #SBATCH --time=5-00:00 > > SLURM_SUB=$(pwd) > echo $SLURM_SUB > testres > echo $SLURM_JOBID > res2 > mkdir /tmp/job_$SLURM_JOBID > cp *psf /tmp/job_$SLURM_JOBID > cp *fdf /tmp/job_$SLURM_JOBID > ls /tmp/job_$SLURM_JOBID >> res2 > cd /tmp/job_$SLURM_JOBID > > source /opt/intel/bin/ifortvars.sh intel64 > > /cluster_cti/utils/openmpi/openmpi-1.10.2/bin/mpirun -n 18 > /cluster_cti/bin/SIESTA/bin/siesta-so2 < tecfai.fdf > tecfai.log > cp *.log $SLURM_SUB > > Can anyone help me > > Regards, > > Rémi > > > > > >
