===== NAMD =====
NAMD is a parallel molecular dynamics code designed for high-performance simulation of large biomolecular systems.
[[http://www.ks.uiuc.edu/Research/namd|NAMD]]
==== Job NAMD on BDW ====
=== Job NAMD multicore (single BDW node) ===
Example script ''slurm-namd-bdw.sh'' to run NAMD multicore on 1 node using 16 cores and at most 32 GB of memory:
#!/bin/bash
#SBATCH --job-name=NAMD
#SBATCH --output=%x.o%j
#SBATCH --error=%x.e%j
#SBATCH --nodes=1
#SBATCH --ntasks-per-node=16
#SBATCH --mem=32G
#SBATCH --time=1-00:00:00
#SBATCH --partition=bdw
#
# Edit next line
#SBATCH --account=
module load namd/2.12/bdw
if [ ! -e "${SLURM_JOB_NAME}.conf" ]; then
echo "Error: file '${SLURM_JOB_NAME}.conf' not found." 1>&2
exit 1
fi
namd2 +p ${SLURM_TASKS_PER_NODE} "${SLURM_JOB_NAME}.conf" > "${SLURM_JOB_NAME}.log"
Submission example:
sbatch --job-name=t1r1_min1 slurm-namd-bdw.sh
=== Job NAMD multicore (single BDW node, simplified version) ===
Example script ''slurm-namd-bdw.sh'' to run NAMD multicore on 1 node using 16 cores and at most 32 GB of memory:
#!/bin/bash
#SBATCH --job-name=t1r1_min1
#SBATCH --output=%x.o%j
#SBATCH --error=%x.e%j
#SBATCH --nodes=1
#SBATCH --ntasks-per-node=16
#SBATCH --mem=32G
#SBATCH --time=1-00:00:00
#SBATCH --partition=bdw
#
# Edit next line
#SBATCH --account=
module load namd/2.12/bdw
namd2 +p ${SLURM_TASKS_PER_NODE} t1r1_min1.conf > t1r1_min1.log
Submission example:
sbatch slurm-namd-bdw.sh
=== Job NAMD multicore (single BDW node, exclusive mode) ===
Example script ''slurm-namd.sh'' to run NAMD multicore on 1 node in exclusive mode:
#!/bin/bash
#SBATCH --job-name=NAMD
#SBATCH --output=%x.o%j
#SBATCH --error=%x.e%j
#SBATCH --nodes=1
#SBATCH --mem=0
#SBATCH --exclusive
#SBATCH --time=1-00:00:00
#SBATCH --partition=bdw
#
# Edit next line
#SBATCH --account=
module load namd/2.12/bdw
if [ ! -e "${SLURM_JOB_NAME}.conf" ]; then
echo "Error: file '${SLURM_JOB_NAME}.conf' not found." 1>&2
exit 1
fi
namd2 +p ${SLURM_TASKS_PER_NODE} "${SLURM_JOB_NAME}.conf" > "${SLURM_JOB_NAME}.log"
Submission example:
sbatch --job-name=t1r1_min1 slurm-namd-bdw.sh
=== Job NAMD multicore (single BDW node, exclusive mode, simplified version) ===
Example script ''slurm-namd-bdw.sh'' to run NAMD multicore on 1 node in exclusive mode:
#!/bin/bash
#SBATCH --job-name=t1r1_min1
#SBATCH --output=%x.o%j
#SBATCH --error=%x.e%j
#SBATCH --nodes=1
#SBATCH --mem=0
#SBATCH --exclusive
#SBATCH --time=1-00:00:00
#SBATCH --partition=bdw
#
# Edit next line
#SBATCH --account=
module load namd/2.12/bdw
namd2 +p ${SLURM_TASKS_PER_NODE} t1r1_min1.conf > t1r1_min1.log
Submission example:
sbatch slurm-namd-bdw.sh
==== Job NAMD on KNL ====
=== Job NAMD multicore (single KNL node) ===
Example script ''slurm-namd-knl.sh'' to run NAMD multicore on 1 KNL node in exclusive mode using 64 cores:
#!/bin/bash
#SBATCH --job-name=NAMD
#SBATCH --output=%x.o%j
#SBATCH --error=%x.e%j
#SBATCH --nodes=1
#SBATCH --ntasks-per-node=64
#SBATCH --exclusive
#SBATCH --mem=0
#SBATCH --time=1-00:00:00
#SBATCH --partition=knl
#
# Edit next line
#SBATCH --account=
module load namd/2.12/knl
if [ ! -e "${SLURM_JOB_NAME}.conf" ]; then
echo "Error: file '${SLURM_JOB_NAME}.conf' not found." 1>&2
exit 1
fi
namd2 +p ${SLURM_TASKS_PER_NODE} "${SLURM_JOB_NAME}.conf" > "${SLURM_JOB_NAME}.log"
Submission example:
sbatch --job-name=t1r1_min1 slurm-namd-knl.sh
==== Job NAMD on GPU ====
=== Job NAMD multicore (single GPU node) ===
Example script ''slurm-namd-gpu.sh'' to run NAMD multicore on 1 GPU node using 8 cores, 2 GPUs and at most 32 GB of memory:
#!/bin/bash
#SBATCH --job-name=NAMD
#SBATCH --output=%x.o%j
#SBATCH --error=%x.e%j
#SBATCH --nodes=1
#SBATCH --ntasks-per-node=8
#SBATCH --gres=gpu:2
#SBATCH --mem=32G
#SBATCH --time=1-00:00:00
#SBATCH --partition=gpu
#
# Edit next line
#SBATCH --account=
module load namd/2.12/gpu
if [ ! -e "${SLURM_JOB_NAME}.conf" ]; then
echo "Error: file '${SLURM_JOB_NAME}.conf' not found." 1>&2
exit 1
fi
namd2 +p ${SLURM_TASKS_PER_NODE} "${SLURM_JOB_NAME}.conf" > "${SLURM_JOB_NAME}.log"
Submission example:
sbatch --job-name=t1r1_min1 slurm-namd-gpu.sh
Submission example overriding the options specified in the launch script:
sbatch --job-name=t1r1_min1 --ntasks-per-node=16 --gres=gpu:4 --mem=64G slurm-namd-gpu.sh