Sample Script This first script uses the SLURM_JOBID variable to create a directory to store the results of the job #!/bin/bash #SBATCH --job-name=mclz # Job name #SBATCH --partition=batch # Partition (queue) name #SBATCH --nodes=1 # Number of nodes #SBATCH --ntasks-per-node=1 # Number of tasks to call on each node #SBATCH --mem-per-cpu=200mb # Memory per processor #SBATCH --time=24:00:00 # Time limit hrs:min:sec #SBATCH --output=mclz.%j.out # Standard output log #SBATCH --error=mclz.%j.err # Standard error log JOBDIR=${SLURM_JOBID} mkdir $JOBDIR export PATH=./bin:$PATH time srun ./bin/mclz.sh -n -I C -Z 6 -A He -L TA -N "TA" cp c6+he* $JOBDIR Sample OpenMPI Script #!/bin/sh #SBATCH --nodes=5 # Number of nodes #SBATCH --ntasks-per-node=6 # Number of tasks per node #SBATCH --mem=20 # Memory per CPU (MB) #SBATCH -p batch # Partition to use #SBATCH --output=example.%j.out # Standard output log #SBATCH --error=example.%j.err # Standard error log export PMIX_MCA_psec=native export OMPI_MCA_btl="tcp,self" srun centmpi Sample Intel MPI Script #!/bin/bash #SBATCH --job-name=water0 #SBATCH --partition=batch # Partition (queue) name #SBATCH --nodes=1 # Number of nodes #SBATCH --ntasks=2 # Number of MPI ranks #SBATCH --ntasks-per-node=2 # Number of tasks to run on each node #SBATCH --cpus-per-task=1 # Number of cores per MPI rank #SBATCH --ntasks-per-core=1 # Number of tasks to call on each core #SBATCH --output=water0-%j.out # Standard output log #SBATCH --error=water0-%j.err # Standard error log export PMIX_MCA_psec=native # Required for munge to work #export OMPI_MCA_btl="tcp,self" # Required for OpenMPI ## If compiling before submitting, run these three commands before compiling. ## Otherwise, run these three commands before submitting your job, or remove the '###' to uncomment them ###module purge ###module load intel/2022.1.0 ###module load impi/2021.6.0 unset I_MPI_PMI_LIBRARY # When using impi and mpirun mpirun -np 2 ./water0_MPI.exe Sample OpenMP/Intel MPI Hybrid Script #!/bin/bash #SBATCH --job-name=test1 #SBATCH --partition=batch # Partition (queue) name #SBATCH --nodes=1 # Number of nodes #SBATCH --ntasks=4 # Number of MPI ranks #SBATCH --ntasks-per-node=4 # Number of tasks to run on each node #SBATCH --output=test1-%j.out # Standard output log #SBATCH --error=test1-%j.err # Standard error log # Since this code is compiled with intel mpi compilers, swap out gcc/OpenMPI # You can also run these three module commands from the command line before submitting your job # Comment them out here if you run these commands from the command line before job submission. module purge module load intel/2022.1.0 module load impi/2021.6.0