Hi,
I'm no expert for this stuff, but could it be that you generate about 40
of the #my_mol.log.$n# files (probably only 39)?
It could be that the 'mpirun' starts 40 'mdrun'-jobs and each generates
its own out put.
For GROMACS 4.6.x I always used
mdrun -nt X ...
to start a parallel run (where X would be 40 in your case). I think
GROMACS 4.6.x has the MPI stuff build into it and therefore doesn't need
an external 'mpirun' (but i could be wrong - I only know how to use the
stuff, but don't completly understand it...)
Hope this helps a little
Greetings
Thomas
Am 13.02.2014 18:13, schrieb
gromacs.org_gmx-users-requ...@maillist.sys.kth.se:
Dear GROMACS users,
I am facing a strange situation by running Gromacs (v - 4.6.3) in our local
cpu-cluster using MPI/OpenMP parallelization process. I am trying to simulate a
big heterogeneous aquas-polymer system in octahedron box.
I use the following command to run my simulations and use sge queuing system to
submit the job in the cluster.
mpirun -nolocal -np 40 mdrun -ntomp 1
Immediately after launching the job a huge number of backup files are getting
generated in the same directory. Generally naming conventions of these backup
files are #my_mol.log.$n# OR #my_mol.edr.$n# etc, i.e. more than one backup log,
.edr, .gro, etc files are generated, I suppose due to the parallelization. But,
after running some steps the log file and the sge-error file starts complaining
about the disk space although there is no disk space scarcity. Lowering the
frequency of writing to the output did not help neither increasing -cpt from 15
-> 50 helped.
The stranger part is here when I see the backup files are getting updated
regularly individually at the same directory (although the main .log file
stopped updating itself) and run for the entire steps (e.g. 10000 steps)
individually as if $n number of individual simulations are running.
My .mdp file is :
title = MD test run of 2 ns
; using Verlet scheme
cutoff-scheme = Verlet
; Run parameters
integrator = md
nsteps = 1000000
dt = 0.002
; Output control
nstxout = 10000
nstvout = 5000
nstxtcout = 1000
nstenergy = 1000
nstlog = 1000
; Bond parameters
continuation = yes
constraint_algorithm = lincs
constraints = all-bonds
lincs_iter = 1
lincs_order = 4
; Neighborsearching
ns_type = grid
nstlist = 10
rlist = 1.0
rcoulomb = 1.0
vdw-type = cut-off
rvdw = 1.0
; Electrostatics
coulombtype = PME
pme_order = 4
fourierspacing = 0.16
; Temperature coupling is on
tcoupl = V-rescale
tc-grps = polymer SOL_Ion
tau_t = 0.1 0.1
ref_t = 300 300
; Pressure coupling is on
pcoupl = Parrinello-Rahman
pcoupltype = isotropic
tau_p = 2.0
ref_p = 1.0
compressibility = 4.5e-5
; Periodic boundary conditions
pbc = xyz
; Dispersion correction
DispCorr = EnerPres
; Velocity generation
gen_vel = no
The commands I used to run:
J = i-1
### When I changing in .mdp file
grompp -f md.mdp -c test_$j.tpr -o test_$i.tpr -t test_$j.cpt -p test_solv.top -n
my_index.ndx >& log.grompp
mpirun -nolocal -np 40 $GMX_HOME/bin/mdrun -ntomp 1 -s test_$i.tpr -deffnm
test_$i -cpt 30
### When I try to extend the simulations
tpbconv -s test_${j}.tpr -o test_${i}.tpr -extend 1000 >& log_${i}.tpbconv
mpirun -nolocal -np 40 $GMX_HOME/bin/mdrun -s test_${i}.tpr -deffnm test_${i}
-cpi test_${j}.cpt -cpt 30
As this is my first attempt with the Gromacs and no one has ever used
MD/Gromacs in my surrounding I could not verify this strange behavior. I have
tried to search for the error in the web as well as in the community forum but
not found any reference to this issue. I may not have searched with proper
terminology however. I am also in touch with my sys admin who are also little
bit lost in this. If anyone could help me to get out of this situation I would
be highly obliged. Please let me know if I am not very explicit in describing
the issue.
Looking forward to hear from you.
Thanks in advance,
Mousumi
Ontario Institute for Cancer Research
MaRS Centre
Toronto, Ontario
--
Gromacs Users mailing list
* Please search the archive at
http://www.gromacs.org/Support/Mailing_Lists/GMX-Users_List before posting!
* Can't post? Read http://www.gromacs.org/Support/Mailing_Lists
* For (un)subscribe requests visit
https://maillist.sys.kth.se/mailman/listinfo/gromacs.org_gmx-users or send a
mail to gmx-users-requ...@gromacs.org.