HPC/Applications/namd: Difference between revisions
m (→NAMD on Carbon) |
m (Stern moved page HPC/Modules/namd to HPC/Applications/namd) |
||
(10 intermediate revisions by the same user not shown) | |||
Line 1: | Line 1: | ||
== | == MPI version == | ||
=== Modules === | === Modules === | ||
* As of NAMD-2.9, the | * As of NAMD-2.9, the namd module requires [[../fftw3|FFTW-3]]. You can load both modules together: | ||
module load fftw3 namd | module load fftw3 namd | ||
=== | === Using MPI === | ||
* NAMD uses its own MPI wrapper, called [http://charm.cs.uiuc.edu/ Charm++], but the default <code>mpirun</code> from OpenMPI works just fine. | * NAMD uses its own MPI wrapper, called [http://charm.cs.uiuc.edu/ Charm++], but the default <code>mpirun</code> from OpenMPI works just fine. | ||
* Use the [[HPC/Submitting and Managing Jobs/Example Job Script#Generic job scripts|Generic job template]] | * Use the [[HPC/Submitting and Managing Jobs/Example Job Script#Generic job scripts|Generic job template]] | ||
Line 22: | Line 10: | ||
#!/bin/bash | #!/bin/bash | ||
… | … | ||
cd $PBS_O_WORKDIR | cd $PBS_O_WORKDIR | ||
mpirun -machinefile $PBS_NODEFILE -np $PBS_NP \ | mpirun -machinefile $PBS_NODEFILE -np $PBS_NP \ | ||
namd2 | namd2 file.namd | ||
</syntaxhighlight> | </syntaxhighlight> | ||
== | == Using GPUs and MPI == | ||
; Update 2012-11-02: | As of module version namd/2.9plus-MPI-icc-3 the module provides two binaries, a regular one called <code>namd2</code> (as usual) and a GPU-eneabled binary called <code>namd2-cuda</code>. The latter requires the cuda module as additional prerequisite: | ||
* http://www.ks.uiuc.edu/Research/namd/development.html – build instructions | module load fftw3 | ||
* http://www.ks.uiuc.edu/Research/namd/2.7/ug/node72.html – running with CUDA. | module load cuda | ||
module load namd | |||
<!-- | |||
; Update 2012-11-02: This section is in development. --[[User:Stern|stern]]. | |||
* http://www.ks.uiuc.edu/Research/namd/development.html – build instructions | |||
* http://www.ks.uiuc.edu/Research/namd/2.7/ug/node72.html – running with CUDA (namd-2.7) --> | |||
The relevant paragraphs from the [http://www.ks.uiuc.edu/Research/namd/2.9/ug/node88.html NAMD User Guide] are: | |||
<blockquote> | |||
Energy evaluation is slower than calculating forces alone, and the loss is much greater in CUDA-accelerated builds. Therefore you should set outputEnergies to 100 or higher in the simulation config file. Some features are unavailable in CUDA builds, including alchemical free energy perturbation and the Lowe-Andersen thermostat. | |||
<br> | |||
As this is a new feature you are encouraged to test all simulations before beginning production runs. Forces evaluated on the GPU differ slightly from a CPU-only calculation, an effect more visible in reported scalar pressure values than in energies. | |||
<br.>… <br> | |||
Each namd2 thread can use only one GPU. Therefore you will need to run at least one thread for each GPU you want to use. Multiple threads can share a single GPU, usually with an increase in performance. NAMD will automatically distribute threads equally among the GPUs on a node. | |||
</blockquote> | |||
To run on GPU nodes: | |||
#PBS -l nodes=''N'':ppn=''PPN''''':gpus=1''' | |||
… | |||
mpirun -machinfile $PBS_NODEFILE -np $PBS_NP \ | |||
namd2'''-cuda +idlepoll''' ''file.namd'' | |||
* Test and optimize the <code>''N''</code> and <code>''PPN''</code> parameters for your situation. Start with <code>nodes=1:ppn=4</code>, increase up to <code>ppn='''12'''</code>, and/or increase <code>nodes > 1</code>. Speedups from a GPU can be expected to be of order 4–5 ×. | |||
* The <code>gpus=…</code> modifier refers to GPUs ''per node'' and currently must always be 1. | |||
== Documentation == | == Documentation == |
Latest revision as of 15:21, June 25, 2013
MPI version
Modules
- As of NAMD-2.9, the namd module requires FFTW-3. You can load both modules together:
module load fftw3 namd
Using MPI
- NAMD uses its own MPI wrapper, called Charm++, but the default
mpirun
from OpenMPI works just fine. - Use the Generic job template
#!/bin/bash
…
cd $PBS_O_WORKDIR
mpirun -machinefile $PBS_NODEFILE -np $PBS_NP \
namd2 file.namd
Using GPUs and MPI
As of module version namd/2.9plus-MPI-icc-3 the module provides two binaries, a regular one called namd2
(as usual) and a GPU-eneabled binary called namd2-cuda
. The latter requires the cuda module as additional prerequisite:
module load fftw3 module load cuda module load namd
The relevant paragraphs from the NAMD User Guide are:
Energy evaluation is slower than calculating forces alone, and the loss is much greater in CUDA-accelerated builds. Therefore you should set outputEnergies to 100 or higher in the simulation config file. Some features are unavailable in CUDA builds, including alchemical free energy perturbation and the Lowe-Andersen thermostat.
As this is a new feature you are encouraged to test all simulations before beginning production runs. Forces evaluated on the GPU differ slightly from a CPU-only calculation, an effect more visible in reported scalar pressure values than in energies. <br.>…
Each namd2 thread can use only one GPU. Therefore you will need to run at least one thread for each GPU you want to use. Multiple threads can share a single GPU, usually with an increase in performance. NAMD will automatically distribute threads equally among the GPUs on a node.
To run on GPU nodes:
#PBS -l nodes=N:ppn=PPN:gpus=1 … mpirun -machinfile $PBS_NODEFILE -np $PBS_NP \ namd2-cuda +idlepoll file.namd
- Test and optimize the
N
andPPN
parameters for your situation. Start withnodes=1:ppn=4
, increase up toppn=12
, and/or increasenodes > 1
. Speedups from a GPU can be expected to be of order 4–5 ×. - The
gpus=…
modifier refers to GPUs per node and currently must always be 1.
Documentation
- See Documentation on the NAMD home page – contains official User guide
- Local documentation:
$ ls $NAMD_HOME/share/doc README.txt announce.txt license.txt notes.txt ug.pdf
Benchmarks
- Using ApoA1 example from http://www.ks.uiuc.edu/Research/namd/performance.html
grep Benchm *2.7*/*.o* | sort -k4,4n | _ -w namd-2.7b1-gen2-1/namd-gen2-1.o248034:Info: Benchmark time: 1 CPUs 1.81607 s/step 21.0193 days/ns 241.219 MB memory namd-2.7b1-gen2-1/namd-gen2-1.o248034:Info: Benchmark time: 1 CPUs 1.8238 s/step 21.1088 days/ns 240.773 MB memory namd-2.7b1-gen2-4/namd-gen2-4.o248032:Info: Benchmark time: 4 CPUs 0.428028 s/step 4.95403 days/ns 90.7305 MB memory namd-2.7b1-gen2-4/namd-gen2-4.o248032:Info: Benchmark time: 4 CPUs 0.430383 s/step 4.98129 days/ns 90.25 MB memory namd-2.7b1-gen2-8/namd-gen2-8.o248035:Info: Benchmark time: 8 CPUs 0.568039 s/step 6.57453 days/ns 65.707 MB memory namd-2.7b1-gen2-8/namd-gen2-8.o248035:Info: Benchmark time: 8 CPUs 0.887236 s/step 10.2689 days/ns 65.3945 MB memory namd-2.7b1-gen2-8/namd-gen2-8.o248036:Info: Benchmark time: 8 CPUs 0.227086 s/step 2.62831 days/ns 64.2656 MB memory namd-2.7b1-gen2-8/namd-gen2-8.o248036:Info: Benchmark time: 8 CPUs 0.229058 s/step 2.65113 days/ns 64.5469 MB memory namd-2.7b1-gen2-16/namd-gen2-16.o248033:Info: Benchmark time: 16 CPUs 0.105355 s/step 1.21938 days/ns 50.6367 MB memory namd-2.7b1-gen2-16/namd-gen2-16.o248033:Info: Benchmark time: 16 CPUs 0.105658 s/step 1.2229 days/ns 50.3477 MB memory grep Benchm *2.9*/*.o* | sort -k4,4n | _ -w namd-2.9-gen2-1/namd-gen2-1.o248004:Info: Benchmark time: 1 CPUs 1.58065 s/step 18.2946 days/ns 398.93 MB memory namd-2.9-gen2-1/namd-gen2-1.o248004:Info: Benchmark time: 1 CPUs 1.58666 s/step 18.3641 days/ns 398.695 MB memory namd-2.9-gen2-1/namd-gen2-1.o248011:Info: Benchmark time: 1 CPUs 1.79167 s/step 20.737 days/ns 398.926 MB memory namd-2.9-gen2-1/namd-gen2-1.o248011:Info: Benchmark time: 1 CPUs 1.79819 s/step 20.8124 days/ns 398.695 MB memory namd-2.9-gen2-4/namd-gen2-4.o248003:Info: Benchmark time: 4 CPUs 1.05068 s/step 12.1606 days/ns 254.582 MB memory namd-2.9-gen2-4/namd-gen2-4.o248003:Info: Benchmark time: 4 CPUs 1.17972 s/step 13.6542 days/ns 254.582 MB memory namd-2.9-gen2-4/namd-gen2-4.o248006:Info: Benchmark time: 4 CPUs 0.424272 s/step 4.91055 days/ns 269.023 MB memory namd-2.9-gen2-4/namd-gen2-4.o248006:Info: Benchmark time: 4 CPUs 0.42632 s/step 4.93426 days/ns 269.023 MB memory namd-2.9-gen2-8/namd-gen2-8.o247983:Info: Benchmark time: 8 CPUs 0.211136 s/step 2.4437 days/ns 242.109 MB memory namd-2.9-gen2-8/namd-gen2-8.o247983:Info: Benchmark time: 8 CPUs 0.219128 s/step 2.53621 days/ns 242.109 MB memory namd-2.9-gen2-16/namd-gen2-1.o248005:Info: Benchmark time: 16 CPUs 0.103843 s/step 1.20188 days/ns 226.25 MB memory namd-2.9-gen2-16/namd-gen2-1.o248005:Info: Benchmark time: 16 CPUs 0.103924 s/step 1.20282 days/ns 225.961 MB memory