Also, if you have slurm installed on a deb based distro, you can try this https://github.com/edf-hpc/slurm-web
I tried to get it running on RPM (Centos) but it is too tightly coupled to deb for my ability to port it. cheers L. ------ The most dangerous phrase in the language is, "We've always done it this way." - Grace Hopper On 19 September 2016 at 12:07, Lachlan Musicman <data...@gmail.com> wrote: > I think you need a couple of things going on: > > 1. you have to have some sort of accounting organised and set up > 2. your sbatch scripts need to use: srun <command> not just <command> > 3. sinfo should then work on the job number. > > When I asked, that was the response iirc. > > > cheers > L. > > ------ > The most dangerous phrase in the language is, "We've always done it this > way." > > - Grace Hopper > > On 19 September 2016 at 10:41, Igor Yakushin <igor.2...@gmail.com> wrote: > >> Hi All, >> >> I'd like to be able to see for a given jobid how much resources are used >> by a job on each node it is running on at this moment. Is there a way to do >> it? >> >> So far it looks like I have to script it: get the list of the involved >> nodes using, for example, squeue or qstat, ssh to each node and find all >> the user processes (not 100% guaranteed that they would be from the job I >> am interested in: is there a way to find UNIX pids corresponding to Slurm >> jobid?). >> >> Another question: is there python API to slurm? I found pyslurm but so >> far it would not build with my version of Slurm. >> >> Thank you, >> Igor >> >> >