Fwd: [gmx-users] mdrun on several clusternodes using PME gromacs 3.3.1

Joern Lenz JLenz at zbh.uni-hamburg.de
Fri Oct 27 15:20:16 CEST 2006


thank you lars and sorry guys, 
i forgot to let you know about another outputfile concerning the problem
discussed above:
lars: i submitted my jobs with an mdrun_mpi command to our cluster with:
"qsub run_the_simulation.sh" and in the sh script the command is 

grompp -np $NODES -n $1_ion -f pr_mdrun01 -c $1_b4pr_mdrun01 -r 
$1_b4pr_mdrun01 -p $1 -o $1_pr_mdrun01 2>&1 > output.grompp_pr_mdrun01

/opt/mpich/ch-p4/bin/mpirun -machinefile ~/machines.lst -np $NODES 
-v /.../${CONFIGGUESS}/bin/mdrun_mpi -np $NODES -s $1_pr_mdrun01.tpr -o 
$1_pr_mdrun01 -c $1_pr_mdrun01 -v -e $1_pr_mdrun01 2>&1 > 
output.mdrun_pr_mdrun01

where $NODES is 16 and CONFIGGUESS guesses the architecture of the system (32 
or 64 bit)

i dont know if nohup and/or & will change anything about the problem.

SOME FURTHER INFORMATION OF THE PROBLEM:

running /../i686-pc-linux-gnu/bin/mdrun_mpi on 2 LINUX ch_p4 processors
Created /../PI13506
/home/blah
p0_13586: (5921.985420) net_recv failed for fd = 6
p0_13586:  p4_error: net_recv read, errno = : 104


perhaps this gives you further information.
i tried nearly everything and also discussed the problem with our
system-administrators... no ideas to solve ...
the thing only works if I run it local, but this is very time-consuming

please help me with suggestions or i will die unhappy
thank you so much
joern



More information about the gromacs.org_gmx-users mailing list