Hello,in reply to my previous question even with options in compilation with 
mpi_d I wasn't able to solve my problem. Every time I choose to run gromacs in 
more than one node, after some number of steps I get that error. It is very 
annoying, because I cannot increase the speed of my runs, yet only up to two 
processors (that means that I can at least run in parallel somehow). In order 
to see if the compilation was successful I even restarted the whole cluster,yet 
to no avail. The runs are ok if I run them in only one node though. Any help 
would be greatly appreciated!
Regards,Nikos
--- Claus Valka <[EMAIL PROTECTED]> schrieb am Fr, 19.9.2008:
Von: Claus Valka <[EMAIL PROTECTED]>
Betreff: [gmx-users] ci range checking error even applying the patch / parallel 
runs
An: [email protected]
Datum: Freitag, 19. September 2008, 16:24

Hello,after searching extensively the mailing lists I wasn' able to solve my 
problem. This has to do with running gromacs in parallel (more than one nodes) 
in a rocks cluster. I 'm able to run a simulation both in one or two processors 
in a dual core node, yet every time I try to use more than one nodes, this 
error appears :Program mdrun_mpi, VERSION 3.3.2
Source code file: nsgrid.c, line: 226
    
    Range checking error:
    Explanation: During neighborsearching, we assign each particle to a grid   
based on its coordinates. If your system contains collisions or parameter   
errors that give particles very high velocities you might end up with some   
coordinates being +-Infinity or NaN (not-a-number). Obviously, we cannot   put 
these on a grid, so this is usually where we detect those errors.   Make sure
 your system is properly energy-minimized and that the potential   energy seems 
reasonable before trying again.
    
   Variable ci has value 741. It should have been within [ 0 .. 740 ]

There isn't only me who is facing that problem so searching the forums I came 
accross a patch, that is : [gmx-users] fix for range checking errors in 
parallel double precision mdrun 
Compiling gromacs with this version of xtcio.c doesn't seem to solve the 
problem. Is there anything else I'm missing?I give you my configure options in 
case this is helpful:./configure --prefix=/export/local/gromacs-3.3.2 
--program-suffix= --enable-double --disable-float --disable-fortran --with-x 
--with-fft=fftw3

My commands to run a simulation are for example in six nodes :grompp -f *.mdp 
-c *.gro -p *.top -np 6and in my script the arguments for running are:-pe mpi 
6mpirun -np 6 --hostfile
 $TMPDIR/machines mdrun_mpiThank you in advance,Nikos
__________________________________________________
Do You Yahoo!?
Sie sind Spam leid? Yahoo! Mail verfügt über einen herausragenden Schutz gegen 
Massenmails. 
http://mail.yahoo.com _______________________________________________
gmx-users mailing list    [email protected]
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the 
www interface or send it to [EMAIL PROTECTED]
Can't post? Read http://www.gromacs.org/mailing_lists/users.php


      
_______________________________________________
gmx-users mailing list    [email protected]
http://www.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at http://www.gromacs.org/search before posting!
Please don't post (un)subscribe requests to the list. Use the 
www interface or send it to [EMAIL PROTECTED]
Can't post? Read http://www.gromacs.org/mailing_lists/users.php

Reply via email to