Differences
This shows you the differences between two versions of the page.
abstract:darwin:earlyaccess [2021-04-24 14:38] – [Lustre high-performance filesystem] anita | abstract:darwin:earlyaccess [2021-04-27 16:21] (current) – external edit 127.0.0.1 | ||
---|---|---|---|
Line 87: | Line 87: | ||
Each node scratch filesystem disk is only accessible by the node in which it is physically installed. The job scheduling system creates a temporary directory associated with each running job on this filesystem. When your job terminates, the job scheduler automatically erases that directory and its contents. | Each node scratch filesystem disk is only accessible by the node in which it is physically installed. The job scheduling system creates a temporary directory associated with each running job on this filesystem. When your job terminates, the job scheduler automatically erases that directory and its contents. | ||
+ | More detailed information about DARWIN storage and quotas can be found on the < | ||
===== Software ===== | ===== Software ===== | ||
- | A list of installed software that IT builds and maintains for DARWIN users can be found by logging into DARWIN and using the VALET command '' | + | A list of installed software that IT builds and maintains for DARWIN users can be found by [[abstract: |
+ | |||
+ | Documentation for all software is organized in alphabetical order on the < | ||
There will **not** be a full set of software during early access and testing, but we will be continually installing and updating software. | There will **not** be a full set of software during early access and testing, but we will be continually installing and updating software. | ||
Line 108: | Line 110: | ||
==== Queues (Partitions) ==== | ==== Queues (Partitions) ==== | ||
- | Initially for early access | + | During Phase 2 early access partitions |
- | (1) '' | + | See [[abstract/ |
- | + | ||
- | * 7 day run time per job (default 30 minutes) | + | |
- | * 576 cores per job (default 1 core) | + | |
- | * 9 nodes per job | + | |
- | * 1152 cores total per user | + | |
- | * maximum 400 job submissions per user (this includes the number of indices specified for an array job) | + | |
- | + | ||
- | Beginning | + | |
- | + | ||
- | * 2 day run time per job (default 30 minutes) | + | |
- | * no core limit per job or user (default 1 core) | + | |
- | * maximum 400 job submissions per user (this includes the number of indices specified | + | |
- | + | ||
- | (2) '' | + | |
- | + | ||
- | * 7 day run time per job (default 30 minutes) | + | |
We fully expect these limits to be changed and adjusted during the early access period. | We fully expect these limits to be changed and adjusted during the early access period. | ||
Line 132: | Line 118: | ||
==== Run Jobs ==== | ==== Run Jobs ==== | ||
- | In order to schedule any job (interactively or batch) on the DARWIN cluster, you must set your workgroup to define your cluster group. For early access, | + | In order to schedule any job (interactively or batch) on the DARWIN cluster, you must set your workgroup to define your cluster group. For Phase 2 early access, |
<code bash> | <code bash> | ||
- | workgroup -g unsponsored | + | workgroup -g it_css |
</ | </ | ||
- | accomplishes this. You will know if you are in your workgroup based on the change in your bash prompt. | + | will enter the workgroup for '' |
<code bash> | <code bash> | ||
- | [traine@login00.darwin ~]$ workgroup -g unsponsored | + | [traine@login00.darwin ~]$ workgroup -g it_css |
[(unsponsored: | [(unsponsored: | ||
traine | traine | ||
/home/1201 | /home/1201 | ||
- | /lustre/unsponsored | + | /lustre/it_css |
- | unsponsored | + | it_css |
- | /lustre/unsponsored/ | + | /lustre/it_css/ |
- | [(unsponsored:anita)@login00.darwin ~]$ | + | [(it_css:traine)@login00.darwin ~]$ |
</ | </ | ||
- | Now we can use '' | + | Now we can use '' |
<note important> | <note important> | ||
Line 157: | Line 143: | ||
</ | </ | ||
- | <note tip>It is a good idea to periodically check in ''/ | + | <note tip>It is a good idea to periodically check in ''/ |
- | + | ||
- | ==== Memory ==== | + | |
- | + | ||
- | The table below provides the usable memory values available for each type of node currently available on the DARWIN. | + | |
- | ^Node type ^Slurm selection options | + | See [[abstract/darwin/runjobs/|Run jobs]] on the < |
- | |Standard/512 GiB |%%--%%constraint=' | + | |
- | |Large Memory/1 TiB |%%--%%constraint=' | + | |
- | |Extra-Large Memory/2 TiB | + | |
- | |nVidia-T4/512 GiB |%%--%%constraint=' | + | |
- | |nVidia-V100/ | + | |
- | |amd-MI50/512 GiB |%%--%%constraint=' | + | |
- | |Extended Memory/3.73 TiB | + | |
- | The **Extended Memory** node is not accessible via Slurm constraint or gres, but instead specifying the partition '' | ||
===== Help ===== | ===== Help ===== | ||