Table of Contents

DARWIN Slurm Job Script Templates

As on Caviness, environment sense and setup code has been shifted out of the job script templates and into external script fragments that are sourced (executed) by the job script. What remains in the job script templates is the setting of variables that influence those external fragments' execute and the sourcing of them. When IT-RCI must change the behavior of job scripts, the external fragments are modified and the change is effected for all job scripts deriving from the templates.

Where Can I Find Them?

IT-RCI staff are maintaining the DARWIN templates via git. The production copy of the repository is checked-out in /opt/shared/slurm/templates with a symbolic link present at /opt/shared/templates/slurm to maintain parity with other HPC systems.

The external script fragments (mentioned above and discussed in detail below) can be found in the /opt/shared/slurm/templates/libexec directory.

The /opt/shared/templates/slurm symbolic link points to the /opt/shared/slurm/templates/job-scripts directory, which is organized into distinct job classes. The top level is split into applications and generic.

Applications

Software packages that have unique runtime requirements will have a single script or a directory of scripts located in the applications directory. TensorFlow is a good example: DARWIN uses Linux containers (created by Google and distributed via Docker) to execute TensorFlow scripts on compute nodes. Gaussian requires extra work to tailor input files and its own expected environment variables to match the resources allocated to the job by Slurm.

Generic

The application-specific job scripts are actually based on the generic scripts present in the generic directory. Serial jobs can make use of the serial.qs script; programs leveraging threaded (e.g. OpenMP) parallelism can use threads.qs as a starting point.

The mpi directory divides that programming paradigm into implementation-specific variants: mpich, openmpi, and generic (a catch-all that uses machine files and should generally NOT be used).

Hierarchical Modularity

Environment setup tasks have been abstracted into each external script fragment file. Examining the fragment directory:

-rw-r--r-- 1 frey sysadmin 1733 Sep 12  2018 common.sh
-rw-r--r-- 1 frey sysadmin 5621 May  6 14:11 gaussian.sh
-rw-r--r-- 1 frey sysadmin 1580 Sep 12  2018 generic-mpi.sh
-rw-r--r-- 1 frey sysadmin 1432 Sep 14  2018 mpich.sh
-rw-r--r-- 1 frey sysadmin 4805 Sep 12  2018 openmpi.sh
-rw-r--r-- 1 frey sysadmin 2209 May  6 14:11 openmp.sh

Signal Handling

One thing added to the job environment by the common.sh fragment for jobs that register a preemption/timeout signal handler is the UD_EXEC function. By default, if a Bash shell is currently executing a command it will not handle any signals until that command has completed executing. Consider a job script like this:

  :
cleanup() {
    echo "Time limit exceeded, scrubbing all junk files now"
    exit 0
}
UD_JOB_EXIT_FN=cleanup
. /opt/shared/slurm/templates/libexec/common.sh

sleep 500000000

When this job is preempted, the SIGTERM signal is delivered to the Bash shell. The shell notes this, but waits for the sleep command to finish executing. Since that sleep will last 15.85 years, the 5 minute grace period given to jobs that are preempted or time out expires before the cleanup function ever gets called and the job is killed.

In order to get signals to work asynchronously in the Bash shell, long-running commands must be run in the background. The UD_EXEC function does just that:

  :
cleanup() {
    echo "Time limit exceeded, scrubbing all junk files now"
    exit 0
}
UD_JOB_EXIT_FN=cleanup
. /opt/shared/slurm/templates/libexec/common.sh

UD_EXEC sleep 500000000

The sleep command is executed in the background and when the SIGTERM preemption/timeout signal is delivered, the shell immediately calls the cleanup function as expected.