The Epyc cluster consists of 16 compute nodes based on the EPYC 7662 128 core processor.
The cluster has a grand total of 2048 cores (16 * 128 = 2048).
To log into the epyc cluster use the following command:
→ $ ssh epyc
To launch, a job submission script is used. An example script is as follows:
#!/bin/bash # execute in the general partition #SBATCH --partition=general # execute with 40 processes/tasks #SBATCH --ntasks=40 # execute on 4 nodes #SBATCH --cpus-per-task=4 # maximum time is 30 minutes #SBATCH --time=00:30:00 # job name is my_job #SBATCH --job-name=my_job # load environment module load openmpi module load # application execution mpiexec application command line arguments
For non-MPI application users the srun process launcher is available for use.
$ srun <application> <command line arguments>
To submit the script for execution on the compute nodes use the following command:
$ sbatch job_script.sh
Exhaustive description of the “sbatch” command can be found in the official documentation.
To launch, the “srun” command is used. An example interactive job request is as follows:
To run an interactive job using GPUs, a user might execute:
Bash> $ srun --x11 --time=1-00:00:00 --nodelist=epyc08 --partition=gpu --gres=gpu:1 --ntasks=4 --pty /bin/bash -lTo run an interactive job without GPUs, a user might execute:
Bash> $srun --x11 --time=1-00:00:00 --nodelist=epyc09 --partition=general --ntasks=112 --pty /bin/bash -lExhaustive description of “srun” command can be found in the Official Documentation