abstract:darwin:runjobs:runjobs

Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
abstract:darwin:runjobs:runjobs [2021-04-26 12:00] – [Queues] anitaabstract:darwin:runjobs:runjobs [2022-06-03 12:53] (current) – [Runtime environment] anita
Line 7: Line 7:
 Without a job scheduler, a cluster user would need to manually search for the resources required by his or her job, perhaps by randomly logging-in to nodes and checking for other users' programs already executing thereon.  The user would have to "sign-out" the nodes he or she wishes to use in order to notify the other cluster users of resource availability((Historically, this is actually how some clusters were managed!)).  A computer will perform this kind of chore more quickly and efficiently than a human can, and with far greater sophistication. Without a job scheduler, a cluster user would need to manually search for the resources required by his or her job, perhaps by randomly logging-in to nodes and checking for other users' programs already executing thereon.  The user would have to "sign-out" the nodes he or she wishes to use in order to notify the other cluster users of resource availability((Historically, this is actually how some clusters were managed!)).  A computer will perform this kind of chore more quickly and efficiently than a human can, and with far greater sophistication.
  
-Slurm is an open source, fault-tolerant, and highly scalable cluster management and job scheduling system for large and small Linux clusters.  Documentation for the current version of Slurm provide by SchedMD [[https://slurm.schedmd.com/documentation.html|SchedMD Slurm Documentation]].+Slurm is an open source, fault-tolerant, and highly scalable cluster management and job scheduling system for large and small Linux clusters.  Documentation for the current version of Slurm provided by SchedMD [[https://slurm.schedmd.com/documentation.html|SchedMD Slurm Documentation]].
  
-You may find it helpful when migrating from one scheduler to another such as GridEngine to Slurm refer to SchedMD's [[https://slurm.schedmd.com/rosetta.pdf|rosetta]] showing equivalent commands across various schedulers. +You may find it helpful when migrating from one scheduler to another such as GridEngine to Slurm to refer to SchedMD's [[https://slurm.schedmd.com/rosetta.pdf|rosetta]] showing equivalent commands across various schedulers and their [[https://slurm.schedmd.com/pdfs/summary.pdf|command/option summary (two pages)]]
  
-<note tip>It is a good idea to periodically check in ''/opt/templates/slurm/'' for updated or new [[technical:slurm:darwin:templates:start|templates]] to use as job scripts to run generic or specific applicationsdesigned to provide the best performance on DARWIN.</note>+<note tip>It is a good idea to periodically check in ''/opt/shared/templates/slurm/'' for updated or new [[technical:slurm:darwin:templates:start|templates]] to use as job scripts to run generic or specific applications designed to provide the best performance on DARWIN.</note>
  
 Need help? See [[http://www.hpc.udel.edu/presentations/intro_to_slurm/|Introduction to Slurm]] in UD's HPC community cluster environment. Need help? See [[http://www.hpc.udel.edu/presentations/intro_to_slurm/|Introduction to Slurm]] in UD's HPC community cluster environment.
Line 51: Line 51:
 The Slurm workload manager is used to manage and control the computing resources for all jobs submitted to a cluster. This includes load balancing, reconciling requests for memory and processor cores with availability of those resources, suspending and restarting jobs, and managing jobs with different priorities.  The Slurm workload manager is used to manage and control the computing resources for all jobs submitted to a cluster. This includes load balancing, reconciling requests for memory and processor cores with availability of those resources, suspending and restarting jobs, and managing jobs with different priorities. 
  
-In order to schedule any job (interactively or batch) on a cluster, you must set your [[abstract:darwin:app_dev:compute_env#using-workgroup-and-directories|workgroup]] to define your allocation workgroup.+In order to schedule any job (interactively or batch) on a cluster, you must set your [[abstract:darwin:app_dev:compute_env#using-workgroup-and-directories|workgroup]] to define your allocation workgroup **and** explicitly request a single partition.
  
-See [[abstract/darwin/runjobs/schedule_jobs|Scheduling Jobs]] and [[abstract/darwin/runjobs/job_status|Managing Jobs]] on the <html><span style="color:#ffffff;background-color:#2fa4e7;padding:3px 7px !important;border-radius:4px;">sidebar</span></html> for general information about getting started with scheduling and managing jobs on DARWIN using Slurm+See [[abstract/darwin/runjobs/schedule_jobs|Scheduling Jobs]] and [[abstract/darwin/runjobs/job_status|Managing Jobs]] on the <html><span style="color:#ffffff;background-color:#2fa4e7;padding:3px 7px !important;border-radius:4px;">sidebar</span></html> for general information about getting started with Slurm commands for scheduling and managing jobs on DARWIN. 
  
 ===== Runtime environment ===== ===== Runtime environment =====
Line 66: Line 66:
 You do not need this command when you You do not need this command when you
   - type commands, or source the command file,   - type commands, or source the command file,
-  - include lines in the file to be submitted to the sbatch.+  - include lines in the file to be submitted with sbatch.
 </note> </note>
  
Line 76: Line 76:
  
 <code bash> <code bash>
-[traine@darwin ~]$ man squeue+[traine@login00.darwin ~]$ man squeue
 </code> </code>
  
  • abstract/darwin/runjobs/runjobs.1619452858.txt.gz
  • Last modified: 2021-04-26 12:00
  • by anita