Differences
This shows you the differences between two versions of the page.
abstract:darwin:earlyaccess [2021-04-22 18:26] – [Software] anita | abstract:darwin:earlyaccess [2021-04-27 16:21] (current) – external edit 127.0.0.1 | ||
---|---|---|---|
Line 74: | Line 74: | ||
* Each allocation will be assigned a workgroup storage in the Lustre directory (''/ | * Each allocation will be assigned a workgroup storage in the Lustre directory (''/ | ||
- | * Each workgroup storage will have a users directory (''/ | + | * Each workgroup storage will have a users directory (''/ |
* Each workgroup storage will have a software and VALET directory (''/ | * Each workgroup storage will have a software and VALET directory (''/ | ||
* There will be a 1 TiB quota limit for the workgroup storage. | * There will be a 1 TiB quota limit for the workgroup storage. | ||
Line 87: | Line 87: | ||
Each node scratch filesystem disk is only accessible by the node in which it is physically installed. The job scheduling system creates a temporary directory associated with each running job on this filesystem. When your job terminates, the job scheduler automatically erases that directory and its contents. | Each node scratch filesystem disk is only accessible by the node in which it is physically installed. The job scheduling system creates a temporary directory associated with each running job on this filesystem. When your job terminates, the job scheduler automatically erases that directory and its contents. | ||
+ | More detailed information about DARWIN storage and quotas can be found on the < | ||
===== Software ===== | ===== Software ===== | ||
- | A list of installed software that IT builds and maintains for DARWIN users can be found by logging into DARWIN and using the VALET command '' | + | A list of installed software that IT builds and maintains for DARWIN users can be found by [[abstract: |
+ | |||
+ | Documentation for all software is organized in alphabetical order on the < | ||
There will **not** be a full set of software during early access and testing, but we will be continually installing and updating software. | There will **not** be a full set of software during early access and testing, but we will be continually installing and updating software. | ||
Line 104: | Line 106: | ||
===== Scheduler ===== | ===== Scheduler ===== | ||
- | DARWIN will being using the Slurm scheduler like Caviness, and is the most common scheduler among XSEDE resources. Slurm on DARWIN is configured as fairshare with each user being giving equal shares to access the current HPC resources | + | DARWIN will being using the Slurm scheduler like Caviness, and is the most common scheduler among XSEDE resources. Slurm on DARWIN is configured as fairshare with each user being giving equal shares to access the current HPC resources available on DARWIN. |
==== Queues (Partitions) ==== | ==== Queues (Partitions) ==== | ||
- | Initially for early access | + | During Phase 2 early access partitions |
- | (1) '' | + | See [[abstract/ |
- | + | ||
- | * 7 day run time per job (default 30 minutes) | + | |
- | * 576 cores per job (default 1 core) | + | |
- | * 9 nodes per job | + | |
- | * 1152 cores total per user | + | |
- | * maximum 400 job submissions per user (this includes the number of indices specified for an array job) | + | |
- | + | ||
- | Beginning | + | |
- | + | ||
- | * 2 day run time per job (default 30 minutes) | + | |
- | * no core limit per job or user (default 1 core) | + | |
- | * maximum 400 job submissions per user (this includes the number of indices specified | + | |
- | + | ||
- | (2) '' | + | |
- | + | ||
- | * 7 day run time per job (default 30 minutes) | + | |
We fully expect these limits to be changed and adjusted during the early access period. | We fully expect these limits to be changed and adjusted during the early access period. | ||
Line 132: | Line 118: | ||
==== Run Jobs ==== | ==== Run Jobs ==== | ||
- | In order to schedule any job (interactively or batch) on the DARWIN cluster, you must set your workgroup to define your cluster group. For early access, | + | In order to schedule any job (interactively or batch) on the DARWIN cluster, you must set your workgroup to define your cluster group. For Phase 2 early access, |
<code bash> | <code bash> | ||
- | workgroup -g unsponsored | + | workgroup -g it_css |
</ | </ | ||
- | accomplishes this. You will know if you are in your workgroup based on the change in your bash prompt. | + | will enter the workgroup for '' |
<code bash> | <code bash> | ||
- | [traine@login00.darwin ~]$ workgroup -g unsponsored | + | [traine@login00.darwin ~]$ workgroup -g it_css |
[(unsponsored: | [(unsponsored: | ||
traine | traine | ||
/home/1201 | /home/1201 | ||
- | /lustre/unsponsored | + | /lustre/it_css |
- | unsponsored | + | it_css |
- | /lustre/unsponsored/ | + | /lustre/it_css/ |
- | [(unsponsored:anita)@login00.darwin ~]$ | + | [(it_css:traine)@login00.darwin ~]$ |
</ | </ | ||
- | Now we can use '' | + | Now we can use '' |
<note important> | <note important> | ||
Line 157: | Line 143: | ||
</ | </ | ||
- | <note tip>It is a good idea to periodically check in ''/ | + | <note tip>It is a good idea to periodically check in ''/ |
- | + | ||
- | ==== Memory ==== | + | |
- | + | ||
- | The table below provides the usable memory values available for each type of node currently available on the DARWIN. | + | |
- | ^Node type ^Slurm selection options | + | See [[abstract/darwin/runjobs/|Run jobs]] on the < |
- | |Standard/512 GiB |%%--%%constraint=' | + | |
- | |Large Memory/1 TiB |%%--%%constraint=' | + | |
- | |Extra-Large Memory/2 TiB | + | |
- | |nVidia-T4/512 GiB |%%--%%constraint=' | + | |
- | |nVidia-V100/ | + | |
- | |amd-MI50/512 GiB |%%--%%constraint=' | + | |
- | |Extended Memory/3.73 TiB | + | |
- | The **Extended Memory** node is not accessible via Slurm constraint or gres, but instead specifying the partition '' | ||
===== Help ===== | ===== Help ===== | ||