abstract:caviness:runjobs:queues

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
abstract:caviness:runjobs:queues [2021-04-08 10:48] – [The workgroup-specific partitions] anitaabstract:caviness:runjobs:queues [2023-05-30 13:48] (current) – [The job queues (partitions) on Caviness] anita
Line 4: Line 4:
 The Caviness cluster has several kinds of partition (queue) available in which to run jobs: The Caviness cluster has several kinds of partition (queue) available in which to run jobs:
  
-^Kind^Description^ +^Kind^Description^Nodes
-|standard|The default partition if no ''%%--%%partition'' submission flag is specified; jobs can be preempted (killed)| +|standard|The default partition if no ''%%--%%partition'' submission flag is specified; jobs can be preempted (killed)|''scontrol show partition standard''
-|devel|A partition with very short runtime limits and small resource limits; important to use for any development using compilers| +|devel|A partition with very short runtime limits and small resource limits; important to use for any development using compilers|''scontrol show partition devel''
-|workgroup-specific|Partitions associated with specific kinds of compute equipment in the cluster purchased by a research group <<//investing-entity//>> (workgroup)|+|workgroup-specific|Partitions associated with specific kinds of compute equipment in the cluster purchased by a research group <<//investing-entity//>> (workgroup)|''scontrol show partition''<<//workgroup//>>|
  
 ===== The standard partition ===== ===== The standard partition =====
Line 29: Line 29:
   * Testing correctness of program parallelization   * Testing correctness of program parallelization
   * Interactive sessions   * Interactive sessions
 +  * Removing files especially if cleaning up many files and directories in ''$HOME'', ''$WORKDIR'' and ''/lustre/scratch''
 Because performance is not critical for these use cases, the nodes serviced by the ''devel'' partition have hyperthreads enabled, effectively doubling the number of CPUs available. Because performance is not critical for these use cases, the nodes serviced by the ''devel'' partition have hyperthreads enabled, effectively doubling the number of CPUs available.
  
Line 38: Line 39:
 For example: For example:
 <code bash> <code bash>
 +[traine@login01 ~]$ workgroup -g it_css
 [(it_css:traine)@login00 ~]$ srun --partition=devel --nodes=1 --ntasks=1 --cpus-per-task=4 date [(it_css:traine)@login00 ~]$ srun --partition=devel --nodes=1 --ntasks=1 --cpus-per-task=4 date
 Mon Jul 23 15:25:07 EDT 2018 Mon Jul 23 15:25:07 EDT 2018
 </code> </code>
  
-One copy of the ''date'' command is executed on one node in the ''devel'' partition; the command has four cores (or in this case, hyperthreads) allocated to it.  An interactive shell in the ''devel'' partition with two cores available would be started via:+One copy of the ''date'' command is executed on one node in the ''devel'' partition; the command has four cores (or in this case, hyperthreads) allocated to it.  An interactive shell in the ''devel'' partition with two cores and one hour of time available would be started via:
 <code bash> <code bash>
 [traine@login01 ~]$ workgroup -g it_css [traine@login01 ~]$ workgroup -g it_css
-[(it_css:traine)@login01 ~]$ salloc --partition=devel --cpus-per-task=2+[(it_css:traine)@login01 ~]$ salloc --partition=devel --cpus-per-task=2 --time=1:0:0
 salloc: Granted job allocation 940 salloc: Granted job allocation 940
 salloc: Waiting for resource configuration salloc: Waiting for resource configuration
  • abstract/caviness/runjobs/queues.1617893310.txt.gz
  • Last modified: 2021-04-08 10:48
  • by anita