How to Run a Converge Job

How to Run a Job for Converge on a Single Node

Follow these steps to run a Converge job on a single node.

1. Log on to the Grid.

2. Copy the required contents using the following command:

cp -R /wsu/el7/scripts/tutorial/WSU-Case .

3. Change to the directory that has the job script by typing the following command: cd WSU-Case

4. Edit the job script: vim singlenode_converge.sh

This is contents of the job script:

#!/bin/bash

#SBATCH --job-name Singlenode_Converge

#SBATCH -q secondary

# Request one node

#SBATCH -N 1

# Request 5 MPI processes per node

#SBATCH --ntasks-per-node=5

# Request 5G of memory

#SBATCH --mem=5G

# Request the node features cpu type intel

#SBATCH --constraint=intel

# Send email alert when job begins, ends, requeues

#SBATCH --mail-type=ALL

# Where to send email alerts

#SBATCH --mail-user=xxyyyy@wayne.edu

# Create an error file that will be error_<jobid>.out

#SBATCH -e errors_%j.err

# Set maximum time limit

#SBATCH -t 1-0:0:0

cd ~/WSU-Case/

module load converge/2.4-serial

converge-2.4.16-serial super > output_singlenode_converge.out

NOTE: Be sure to change '#SBATCH --mail-user=xxyyyy@wayne.edu' with YOUR email!

5. To submit the job type: sbatch singlenode_converge.sh

6. The job is submitted. You can check useful job information with the following command: qme

Image

In this example, the job is running on the secondary QoS, the node wsu197, it has 5 CPU's, and the following features: intel cpu type and 10e network. The output can be found in output_singlenode_converge.out.

How to Run a Job for Converge on Multiple Nodes

Follow these steps to run a Converge job on multiple nodes.

1. Log on to the Grid.

2. Copy the required contents using the following command:

cp -R /wsu/el7/scripts/tutorial/WSU_Case .

3. Change to the directory that has the job script by typing the following command: cd WSU-Case

4. Edit the job script: vim multinode_converge.sh

It contains the following:

#!/bin/bash

#SBATCH --job-name Multinode_Converge

#SBATCH -q secondary

# Number of nodes

#SBATCH --nodes=4

# Number of MPI processes per node

#SBATCH --ntasks-per-node=32

# Send email alert when job begins, ends, requeues

#SBATCH --mail-type=ALL

# Where to send email alerts

#SBATCH --mail-user=xxyyyy@wayne.edu

# Create an error file that will be error_<jobid>.out

#SBATCH -e errors_%j.err

# Set maximum time limit

#SBATCH -t 10:00:00

WORKDIR=${HOME}/WSU-Case

if [[ ! -d ${WORKDIR} ]]

then

echo "WORKDIR ${WORKDIR} does not exist" >&2

exit 1

fi

THREADS=`cat $SLURM_JOB_NODELIST | wc -l`

cd ${WORKDIR}

ml swap openmpi3 mpich/3.2.1

ml converge/2.4-mpich

NPROCS=128

srun -n $NPROCS converge-2.4.16-mpich super > output_multinode_converge.out 2>&1

NOTE: Be sure to change '#SBATCH --mail-user=xxyyyy@wayne.edu' with YOUR email!

5. To submit the job, type: sbatch multinode_converge.sh

6. The job is submitted. You can check to see what node it is running on with the following command: qme

Image

In this example, the job is running on the secondary QoS, on the nodes asx14-17 and with 128 CPU's. The output can be found in the output_multinode_converge.out.