Am 13.06.2014 um 04:30 schrieb Sangmin Park:

> Yes, both files, 'mpiexec' and 'mpiexec.hydra', are in the "bin" directory 
> inside Intel MPI.
> But, 'mpiexec' file is linked to 'mpiexec.py' file.
> Does it okay if I create a symbolic link 'mpiexec' pointing to 
> 'mpiexec.hydra' instead of 'mpiexec.py' ?

Sure, this should work too.

-- Reuti


> Since there are several running jobs on the cluster, I need to check it again 
> and again.
> Thanks.
> 
> --Sangmin
> 
> 
> On Thu, Jun 12, 2014 at 8:04 PM, Reuti <re...@staff.uni-marburg.de> wrote:
> Am 12.06.2014 um 04:23 schrieb Sangmin Park:
> 
> > I've checked the version of Intel MPI. He uses Intel MPI 4.0.3.008 version.
> > Our system uses rsh to access computing nodes. SGE doses, too.
> >
> > Please let me know how to cehck which one is used 'mpiexec.hydry' or 
> > 'mpiexec'.
> 
> Do you have both files somewhere in a "bin" directory inside the Intel MPI? 
> You could rename "mpiexec" and create a symbolic link "mpiexec" pointing to 
> "mpiexec.hydra". The old startup will need some daemons running on the node 
> (which are outside of SGE's control and accounting*), but "mpiexec.hydra" 
> will startup the child processes on its own as kids of its own and should 
> hence be under SGE's control. And as long as you are staying on one and the 
> same node, this should work already without further setup then. To avoid a 
> later surprise when you compute between nodes, the `rsh`/`ssh` should 
> nevertheless being caught and redirected to `qrsh -inherit...` like outlined 
> in "$SGE_ROOT/mpi".
> 
> -- Reuti
> 
> *) It's even possible to force the daemons to be started under SGE, but it's 
> convoluted and not recommended.
> 
> 
> > Sangmin
> >
> >
> > On Wed, Jun 11, 2014 at 6:46 PM, Reuti <re...@staff.uni-marburg.de> wrote:
> > Hi,
> >
> > Am 11.06.2014 um 02:38 schrieb Sangmin Park:
> >
> > > For the best performance, we recommend users to use 8 cores on a single 
> > > particular node, not distributed with multi node.
> > > Before I said, he uses VASP application compiled with Intel MPI. So he 
> > > uses Intel MPI now.
> >
> > Which version of Intel MPI? Even with the latest one it's not tightly 
> > integrated by default (despite the fact, that MPICH3 [on which it is based] 
> > is tightly integrated by default).
> >
> > Depending on the version it might be necessary to make some adjustments - 
> > IIRC mainly use `mpiexec.hydra` instead of `mpiexec` and supply a wrapper 
> > to catch the `rsh`/`ssh` call (like in the MPI demo in SGE's directory).
> >
> > -- Reuti
> >
> >
> > > --Sangmin
> > >
> > >
> > > On Tue, Jun 10, 2014 at 5:58 PM, Reuti <re...@staff.uni-marburg.de> wrote:
> > > Hi,
> > >
> > > Am 10.06.2014 um 10:21 schrieb Sangmin Park:
> > >
> > > > This user does always parallel job using VASP application.
> > > > Usually, he uses 8 cores per a job. Lots of this kind of job have been 
> > > > submitted by the user.
> > >
> > > 8 cores on a particular node or 8 slots across the cluster? What MPI 
> > > implementation does he use?
> > >
> > > -- Reuti
> > >
> > > NB: Please keep the list posted.
> > >
> > >
> > > > Sangmin
> > > >
> > > >
> > > > On Tue, Jun 10, 2014 at 3:42 PM, Reuti <re...@staff.uni-marburg.de> 
> > > > wrote:
> > > > Am 10.06.2014 um 08:00 schrieb Sangmin Park:
> > > >
> > > > > Hello,
> > > > >
> > > > > I'm very confused about the output of qacct command.
> > > > > I thought CPU column time is the best way to measure resource usage 
> > > > > by users through this web page, 
> > > > > https://wiki.duke.edu/display/SCSC/Checking+SGE+Usage
> > > > >
> > > > > But, I have some situation.
> > > > > One of users in my institution, actually this user is a one of heavy 
> > > > > users, uses lots of HPC resources. To get the resource usage by this 
> > > > > user for requirement of the payment, I commanded qacct and the output 
> > > > > is below, this is just for May.
> > > > >
> > > > > OWNER       WALLCLOCK         UTIME         STIME           CPU       
> > > > >       MEMORY                 IO                IOW
> > > > > ========================================================================================================================
> > > > > p012chm       2980810        28.485        35.012       100.634       
> > > > >        4.277              0.576              0.000
> > > > >
> > > > > CPU time is too much small. Because he is very heavy user of our 
> > > > > institution, I can not accept this result. However, the WALLCLOCK 
> > > > > time is very much.
> > > > >
> > > > > How do I get correct information of usage resources by users via 
> > > > > qacct?
> > > >
> > > > This may happen in case you have parallel jobs which are not tightly 
> > > > integrated into SGE. What types of jobs is the user running?
> > > >
> > > > -- Reuti
> > > >
> > > >
> > > > > ===========================
> > > > > Sangmin Park
> > > > > Supercomputing Center
> > > > > Ulsan National Institute of Science and Technology(UNIST)
> > > > > Ulsan, 689-798, Korea
> > > > >
> > > > > phone : +82-52-217-4201
> > > > > mobile : +82-10-5094-0405
> > > > > fax : +82-52-217-4209
> > > > > ===========================
> > > > > _______________________________________________
> > > > > users mailing list
> > > > > users@gridengine.org
> > > > > https://gridengine.org/mailman/listinfo/users
> > > >
> > > >
> > > >
> > > >
> > > > --
> > > > ===========================
> > > > Sangmin Park
> > > > Supercomputing Center
> > > > Ulsan National Institute of Science and Technology(UNIST)
> > > > Ulsan, 689-798, Korea
> > > >
> > > > phone : +82-52-217-4201
> > > > mobile : +82-10-5094-0405
> > > > fax : +82-52-217-4209
> > > > ===========================
> > >
> > >
> > >
> > >
> > > --
> > > ===========================
> > > Sangmin Park
> > > Supercomputing Center
> > > Ulsan National Institute of Science and Technology(UNIST)
> > > Ulsan, 689-798, Korea
> > >
> > > phone : +82-52-217-4201
> > > mobile : +82-10-5094-0405
> > > fax : +82-52-217-4209
> > > ===========================
> >
> >
> >
> >
> > --
> > ===========================
> > Sangmin Park
> > Supercomputing Center
> > Ulsan National Institute of Science and Technology(UNIST)
> > Ulsan, 689-798, Korea
> >
> > phone : +82-52-217-4201
> > mobile : +82-10-5094-0405
> > fax : +82-52-217-4209
> > ===========================
> 
> 
> 
> 
> -- 
> ===========================
> Sangmin Park 
> Supercomputing Center
> Ulsan National Institute of Science and Technology(UNIST)
> Ulsan, 689-798, Korea 
> 
> phone : +82-52-217-4201
> mobile : +82-10-5094-0405
> fax : +82-52-217-4209
> ===========================


_______________________________________________
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users

Reply via email to