[gmx-users] Reg Gromacs in cluster problem
Justin Lemkul
jalemkul at vt.edu
Wed Mar 26 15:59:25 CET 2014
On 3/26/14, 10:26 AM, vidhya sankar wrote:
> Dear Justin,
> Thank you for your Previous reply I am Running Gromacs on Cluster with 16 Processor
> I am using the following script to submit Job
>
> #!/bin/bash
>
> #PBS -l nodes=compute-0-5:ppn=16
> #PBS -l walltime=900:10:5
> hostname
> date
>
>
> cd $PBS_O_WORKDIR
> echo "files copied from" $PBS_O_WORKDIR
> echo "to computing directory" $TMPDIR
> cd $TMPDIR
>
>
> cp $PBS_O_WORKDIR/newplumed04RestartHILLS.dat $TMPDIR/
>
> mpirun=/opt/openmpi/bin/mpirun
>
> LD_LIBRARY_PATH=/share/apps/gromacsplu/lib
>
> source="/share/apps/gromacsplu//bin"
> MDRUN="/share/apps/gromacsplu/bin/mdrun_mpi_d"
> $mpirun -np 16 $MDRUN -s CNTPEPRSOLIONSfullplumed.tpr -npme 0 -cpi CNTPEPRSOLIONSfullplumed_prev.cpt -append -nt 1 -plumed newplumed04RestartHILLS.dat -v -deffnm CNTPEPRSOLIONSfullplumed &>/dev/null
>
> cp --force $TMPDIR/* $PBS_O_WORKDIR/out14/
>
> rm -rf $TMPDIR
> date
>
>
> But after 19 hours Running My Node Become down and shows as follows
>
> state = down,job-exclusive
> np = 16
> ntype = cluster
>
>
> When i Try to Access My node using the following Command
>
> ssh compute-0-5 then it shows the follwong error
>
> ssh: connect to host compute-0-5 port 22: No route to host
>
> I would be very garteful if you Give any suggestion
>
This isn't a Gromacs problem. Contact your sysadmin.
-Justin
--
==================================================
Justin A. Lemkul, Ph.D.
Ruth L. Kirschstein NRSA Postdoctoral Fellow
Department of Pharmaceutical Sciences
School of Pharmacy
Health Sciences Facility II, Room 601
University of Maryland, Baltimore
20 Penn St.
Baltimore, MD 21201
jalemkul at outerbanks.umaryland.edu | (410) 706-7441
http://mackerell.umaryland.edu/~jalemkul
==================================================
More information about the gromacs.org_gmx-users
mailing list