Table of Contents

Revisions to Slurm Configuration v1.0.7 on DARWIN

This document summarizes alterations to the Slurm job scheduler configuration on the DARWIN cluster.

Issues

See this document discussing swap limits in Slurm jobs.

Implementation

The following aspects of the system construction will be observed in the configuration of the plugin:

These details produce the following configuration string:

partition(extended-mem)=none,host(r2v[00-02])=1.5%/cpu,host(r0m01)=0.25%/cpu,default()=1%/cpu

Impact

No downtime is expected. The slurmd daemon must be restarted on all compute nodes, but currently-executing jobs/job steps should not be affected (they will reconnect to the new slurmd as necessary to communicate job status, etc.). The slurmctld daemons do not use the SPANK plugin, thus they do not need to be restarted.

Timeline

Date Time Goal/Description
2021-11-19 Authoring of this document
2021-11-2409:00Implementation
2021-12-0114:54Update: appropriate aggregate limits on job cgroup