Gah, yes. sstat, not sinfo.

The most dangerous phrase in the language is, "We've always done it this

- Grace Hopper

On 19 September 2016 at 13:00, Peter A Ruprecht <
> wrote:

> Igor,
> Would sstat give you what you need?  (
>  It doesn't update instantaneously but at least a few times a minute.
> If you want to get fancy, I believe that xdmod can integrate with
> TACC-stats to provide graphs about what is happening inside a job but I'm
> not sure whether that updates in "real" time.
> One of our summer interns created a custom ganglia interface that checked
> what nodes a job was running on and graphed several relevant variables
> selected from the ganglia RRD files for those nodes.  If you're interested
> in seeing that work, I can look into whether we can share it.
> So there are some existing ways of going at this.
> Pete
> From: Igor Yakushin <>
> Reply-To: slurm-dev <>
> Date: Sunday, September 18, 2016 at 6:42 PM
> To: slurm-dev <>
> Subject: [slurm-dev] how to monitor CPU/RAM usage on each node of a slurm
> job? python API?
> Hi All,
> I'd like to be able to see for a given jobid how much resources are used
> by a job on each node it is running on at this moment. Is there a way to do
> it?
> So far it looks like I have to script it: get the list of the involved
> nodes using, for example, squeue or qstat, ssh to each node and find all
> the user processes (not 100% guaranteed that they would be from the job I
> am interested in: is there a way to find UNIX pids corresponding to Slurm
> jobid?).
> Another question: is there python API to slurm? I found pyslurm but so far
> it would not build with my version of Slurm.
> Thank you,
> Igor

Reply via email to