Nexus: Difference between revisions

From UMIACS
Jump to navigation Jump to search
No edit summary
 
(190 intermediate revisions by 4 users not shown)
Line 1: Line 1:
The Nexus is the combined scheduler of resources in UMIACS.  Many of our existing computational clusters that have discrete schedulers will be folding into this scheduler in the future.  The resource manager for Nexus (as with our other existing computational clusters) is [[SLURM]].  Resources are arranged into partitions where users are able to schedule computational jobs.  Users are arranged into a number of SLURM accounts based on faculty, lab, or center investments.
The Nexus is the combined scheduler of resources in UMIACS.  The resource manager for Nexus is [[SLURM]].  Resources are arranged into partitions where users are able to schedule computational jobs.  Users are arranged into a number of SLURM accounts based on faculty, lab, or center investments.


= Getting Started =
= Getting Started =
All accounts in UMIACS are sponsored.  If you don't already have a UMIACS account, please see [[Nexus/Accounts]] for information on getting one.
All accounts in UMIACS are sponsored.  If you don't already have a UMIACS account, please see [[Accounts]] for information on getting one.  You need a full UMIACS account (not a [[Accounts/Collaborator | collaborator account]]) in order to access Nexus.


== Access ==
== Access ==
The submission nodes for the Nexus computational resources are determined by department, center, or lab affiliation.  You can log into the [https://intranet.umiacs.umd.edu/directory/cr/ UMIACS Directory CR application] and select the Computational Resource (CR) in the list that has the prefix <code>nexus</code>. The Hosts section lists your available login nodes.
Your access to submission nodes (alternatively called login nodes) for Nexus computational resources is determined by your account sponsor's department, center, or lab affiliation.  You can log into the [https://intranet.umiacs.umd.edu/directory/cr/ UMIACS Directory CR application] and select the Computational Resource (CR) in the list that has the prefix <code>nexus</code>. The Hosts section lists your available submission nodes - generally a pair of nodes of the format <tt>nexus<department, lab, or center abbreviation>[00,01]</tt>, e.g., <tt>nexusgroup00</tt> and <tt>nexusgroup01</tt>.


'''Note''' - UMIACS requires multi-factor authentication through our [[Duo]] instance.  This is completely discrete from both UMD's and CSD's Duo instances.  You will need to enroll one or more devices to access resources in UMIACS, and will be prompted to enroll when you log into the Directory application for the first time.
'''Note''' - UMIACS requires multi-factor authentication through our [[Duo]] instance.  This is completely discrete from both UMD's and CSD's Duo instances.  You will need to enroll one or more factors to access resources in UMIACS, and will be prompted to enroll when you log into the Directory application for the first time.


Once you have identified your submission nodes, you can [[SSH]] directly into them.  From there, you are able to submit to the cluster via our [[SLURM]] workload manager.  You need to make sure that your submitted jobs have the correct account, partition, and qos.
Once you have identified your submission nodes, you can [[SSH]] directly into them.  From there, you are able to submit to the cluster via our [[SLURM]] workload manager.  You need to make sure that your submitted jobs have the correct account, partition, and qos.


== Jobs ==
== Jobs ==
[[SLURM]] jobs are submitted by either <code>srun</code> or <code>sbatch</code> depending if you are doing an interactive job or batch job, respectively.  You need to provide the where/how/who to run the job and specify the resources you need to run with.
[[SLURM]] jobs are [[SLURM/JobSubmission | submitted]] by either <code>srun</code> or <code>sbatch</code> depending if you are doing an interactive job or batch job, respectively.  You need to provide the where/how/who to run the job and specify the resources you need to run with.


For the where/how/who, you may be required to specify <code>--partition</code>, <code>--qos</code>, and/or <code>--account</code> (respectively) to be able to adequately submit jobs to the Nexus.
For the who/where/how, you may be required to specify <code>--account</code>, <code>--partition</code>, and/or <code>--qos</code> (respectively) to be able to adequately submit jobs to the Nexus.


For resources, you may need to specify <code>--time</code> for time, <code>--tasks</code> for CPUs, <code>--mem</code> for RAM, and <code>--gres=gpu</code> for GPUs in your submission arguments to meet your requirements.  There are defaults for all four, so if you don't specify something, you may be scheduled with a very minimal set of time and resources (e.g., by default, NO GPUs are included if you do not specify <code>--gres=gpu</code>).  For more information about submission flags for GPU resources, see [[SLURM/JobSubmission#Requesting_GPUs]].  You can also can run <code>man srun</code> on your submission node for a complete list of available submission arguments.
For resources, you may need to specify <code>--time</code> for time, <code>--ntasks</code> for CPUs, <code>--mem</code> for RAM, and <code>--gres=gpu</code> for GPUs in your submission arguments to meet your requirements.  There are defaults for all four, so if you don't specify something, you may be scheduled with a very minimal set of time and resources (e.g., by default, NO GPUs are included if you do not specify <code>--gres=gpu</code>).  For more information about submission flags for GPU resources, see [[SLURM/JobSubmission#Requesting_GPUs | here]].  You can also can run <code>man srun</code> on your submission node for a complete list of available submission arguments.
 
For a list of available GPU types on Nexus and their specs, please see [[Nexus/GPUs]].
 
For details on how the network for Nexus is architected, please see [[Nexus/Network]]. This can be important if you wish to optimize performance of your jobs.


=== Interactive ===
=== Interactive ===
Line 22: Line 26:


<pre>
<pre>
$ srun --pty --ntasks 4 --mem=2gb --gres=gpu:1 nvidia-smi -L
$ srun --pty --ntasks=4 --mem=2gb --gres=gpu:1 bash
GPU 0: NVIDIA RTX A4000 (UUID: GPU-ae5dc1f5-c266-5b9f-58d5-7976e62b3ca1)
srun: Job account was unset; set to user default of 'nexus'
srun: Job partition was unset; set to cluster default of 'tron'
srun: Job QoS was unset; set to association default of 'default'
srun: Job time limit was unset; set to partition default of 60 minutes
srun: job 1 queued and waiting for resources
srun: job 1 has been allocated resources
$ hostname
tron62.umiacs.umd.edu
$ nvidia-smi -L
GPU 0: NVIDIA GeForce RTX 2080 Ti (UUID: GPU-daad6a04-a2ce-1183-ce53-b267048f750a)
</pre>
</pre>


Line 30: Line 43:


= Partitions =  
= Partitions =  
The SLURM resource manager uses partitions to act as job queues which can restrict size, time and user limits. The Nexus (when fully operational) will have a number of different partitions of resources. Different Centers, Labs, and Faculty will be able to invest in computational resources that will be restricted to approved users through these partitions.
The SLURM resource manager uses partitions to act as job queues which can restrict size, time and user limits. The Nexus has a number of different partitions of resources. Different Centers, Labs, and Faculty are able to invest in computational resources that are restricted to approved users through these partitions.
* [[Nexus/Tron]] - This is the pool of resources available to all UMIACS and CSD faculty and graduate students.  It provides access for undergraduate and graduate teaching resources.
 
* Scavenger - This is a [https://slurm.schedmd.com/preempt.html preemption] partition that supports nodes from multiple other partitions. More resources are available to schedule simultaneously than in other partitions, however jobs are subject to preemption rules. You are responsible for ensuring your jobs handle this preemption correctly, as the SLURM scheduler will simply restart each job with the same submission arguments when preempted jobs are available to run again.
'''Partitions usable by all non-[[ClassAccounts |class account]] users:'''
* [[Nexus/Tron]] - Pool of resources available to all non-class accounts sponsored by either UMIACS or CSD faculty.
* Scavenger - [https://slurm.schedmd.com/preempt.html Preemption] partition that contains [https://en.wikipedia.org/wiki/X86-64 x86_64] architecture nodes from multiple other partitions. More resources are available to schedule simultaneously than in other partitions, however jobs are subject to preemption rules. You are responsible for ensuring your jobs handle this preemption correctly. The SLURM scheduler will simply restart a preempted job with the same submission arguments when it is available to run again. For an overview of things you can check within scripts to determine if your job was preempted/resumed, see [[SLURM/Preemption]].
* Scavenger (aarch64) - Preemption partition identical in design to <tt>scavenger</tt>, but only contains [https://en.wikipedia.org/wiki/AArch64 aarch64] architecture nodes.
 
'''Partitions usable by [[ClassAccounts]]:'''
* [[ClassAccounts | Class]] - Pool of resources available to class accounts sponsored by either UMIACS or CSD faculty.
 
'''Partitions usable by specific lab/center users:'''
* [[Nexus/CBCB]] - CBCB lab pool available for CBCB lab members.
* [[Nexus/CLIP]] - CLIP lab pool available for CLIP lab members.
* [[Nexus/CML]] - CML lab pool available for CML lab members.
* [[Nexus/GAMMA]] - GAMMA lab pool available for GAMMA lab members.
* [[Nexus/MBRC]] - MBRC lab pool available for MBRC lab members.
* [[Nexus/MC2]] - MC2 lab pool available for MC2 lab members.
* [[Nexus/Vulcan]] - Vulcan lab pool available for Vulcan lab members.


= Quality of Service (QoS) =
= Quality of Service (QoS) =
SLURM uses a QoS to provide limits on job sizes to users. Note that you should still try to only allocate the minimum resources for your jobs, as resources that each of your jobs schedules are counted against your [https://slurm.schedmd.com/fair_tree.html FairShare priority] in the future.
SLURM uses Quality of Service (QoS) both to provide limits on job sizes (termed by us as "job QoS") as well as to limit resources used by all jobs running in a partition, either per user or per group (termed by us as "partition QoS").
* default - Default QoS. Limited to 4 cores, 32GB RAM, and 1 GPU per job.  The maximum wall time per job is 3 days.  4 jobs are permitted simultaneously.
 
* medium - Limited to 8 cores, 64GB RAM, and 2 GPUs per job .  The maximum wall time per job is 2 days.  2 jobs are permitted simultaneously.
=== Job QoS ===
* high - Limited to 16 cores, 128GB RAM, and 4 GPUs per job.  The maximum wall time per job is 1 day.  Only 1 job is permitted simultaneously.
Job QoS are used to provide limits on the size of job that you can run. You should try to allocate only the resources your job actually needs, as resources that each of your jobs schedules are counted against your [[SLURM/Priority#Fair-share | fair-share priority]] in the future.
* scavenger - Limited to 64 cores, 256GB RAM, and 8 GPUs per job.  The maximum wall time per job is 2 daysOnly 16 GPUs are permitted simultaneously.  This QoS is only available in the scavenger partition.
* default - Default job QoS. Limited to 4 CPU cores, 1 GPU, and 32GB RAM per job.  The maximum wall time per job is 3 days.
* medium - Limited to 8 CPU cores, 2 GPUs, and 64GB RAM per job.  The maximum wall time per job is 2 days.
* high - Limited to 16 CPU cores, 4 GPUs, and 128GB RAM per job.  The maximum wall time per job is 1 day.
* scavenger - No resource limits per job, only a maximum wall time per job of 3 daysYou are responsible for ensuring your job requests multiple nodes if it requests resources beyond what any one node is capable of.  576 total CPU cores, 72 total GPUs, and 2304GB total RAM are permitted simultaneously across all of your jobs running with this job QoS.  This job QoS is paired 1-1 with the scavenger partition. To use this job QoS, include <code>--partition=scavenger</code> and <code>--account=scavenger</code> in your submission arguments.  Do not include any job QoS argument other than <code>--qos=scavenger</code> (optional) or submission will fail.
* scavenger-aarch64 - No resource limits per job, only a maximum wall time per job of 3 days.  You are responsible for ensuring your job requests multiple nodes if it requests resources beyond what any one node is capable of1600 total CPU cores and 281140MB total RAM are permitted simultaneously across all of your jobs running with this job QoS.  This job QoS is paired 1-1 with the scavenger-aarch64 partition. To use this job QoS, include <code>--partition=scavenger-aarch64</code>, <code>--account=scavenger</code>, and <code>--qos=scavenger-aarch64</code> in your submission arguments.
 
You can display these job QoS from the command line using the <code>show_qos</code> command.  By default, the command will only show job QoS that you can access.  The above four job QoS are the ones that everyone can access.
 
<pre>
$ show_qos
                Name    MaxWall                        MaxTRES MaxJobsPU                      MaxTRESPU
-------------------- ----------- ------------------------------ --------- ------------------------------
            default  3-00:00:00      cpu=4,gres/gpu=1,mem=32G                                         
                high  1-00:00:00    cpu=16,gres/gpu=4,mem=128G                                         
              medium  2-00:00:00      cpu=8,gres/gpu=2,mem=64G                                         
          scavenger  3-00:00:00                                          cpu=576,gres/gpu=72,mem=2304G
  scavenger-aarch64  3-00:00:00                                                    cpu=1600,mem=281140M
</pre>
 
If you want to see all job QoS, including those that you do not have access to, you can use the <code>show_qos --all</code> command.  


You can display these QoSes from the command line using <code>show_qos</code> command.
<pre>
<pre>
# show_qos
$ show_qos --all
      Name    MaxWall MaxJobs                       MaxTRES     MaxTRESPU   Priority
                Name    MaxWall                        MaxTRES MaxJobsPU                      MaxTRESPU
---------- ----------- ------- ------------------------------ ------------- ----------
-------------------- ----------- ------------------------------ --------- ------------------------------
  scavenger 2-00:00:00            cpu=64,gres/gpu=8,mem=256G  gres/gpu=16          0
            cml-cpu 7-00:00:00                                        8
    medium  2-00:00:00       2       cpu=8,gres/gpu=2,mem=64G                       0
        cml-default 7-00:00:00       cpu=4,gres/gpu=1,mem=32G        2
       high  1-00:00:00       1     cpu=16,gres/gpu=4,mem=128G                       0
             cml-high  1-12:00:00    cpu=16,gres/gpu=4,mem=128G        2
   default 3-00:00:00      4       cpu=4,gres/gpu=1,mem=32G                       0
      cml-high_long 14-00:00:00              cpu=32,gres/gpu=8        8                    gres/gpu=8
       tron                                                      gres/gpu=4          0
          cml-medium  3-00:00:00      cpu=8,gres/gpu=2,mem=64G         2
      cml-scavenger  3-00:00:00                                                            gres/gpu=24
      cml-very_high  1-12:00:00    cpu=32,gres/gpu=8,mem=256G        8                    gres/gpu=12
            default  3-00:00:00       cpu=4,gres/gpu=1,mem=32G
                high  1-00:00:00    cpu=16,gres/gpu=4,mem=128G
            highmem 21-00:00:00                cpu=128,mem=2T
          huge-long 10-00:00:00    cpu=32,gres/gpu=8,mem=256G
        interactive    12:00:00                cpu=4,mem=128G
              medium  2-00:00:00      cpu=8,gres/gpu=2,mem=64G
        oasis-exempt 10-00:00:00                                                      cpu=160,mem=28114M
          scavenger  3-00:00:00                                          cpu=576,gres/gpu=72,mem=2304G
   scavenger-aarch64 3-00:00:00                                                   cpu=1600,mem=281140M
          vulcan-cpu  2-00:00:00                cpu=1024,mem=4T        4
       vulcan-default  7-00:00:00       cpu=4,gres/gpu=1,mem=32G         2
      vulcan-exempt  7-00:00:00    cpu=32,gres/gpu=8,mem=256G        2
        vulcan-high  1-12:00:00    cpu=16,gres/gpu=4,mem=128G        2
    vulcan-high_long 14-00:00:00              cpu=32,gres/gpu=8        8                    gres/gpu=8
      vulcan-medium  3-00:00:00       cpu=8,gres/gpu=2,mem=64G        2
      vulcan-sailon  3-00:00:00    cpu=32,gres/gpu=8,mem=256G                              gres/gpu=48
    vulcan-scavenger  3-00:00:00    cpu=32,gres/gpu=8,mem=256G
</pre>
</pre>


Currently in our non-preemption partition, you will be restricted to 4 GPUs at once.
To find out what accounts and partitions you have access to, first use the <code>show_assoc</code> command to show your account/job QoS combinations. Then, use the <code>scontrol show partition</code> command and note the <tt>AllowAccounts</tt> entry for each listed partition. You are able to submit to any partition that allows an account that you have. If you need to use an account other than the default account <tt>nexus</tt>, you will need to specify it via the <code>--account</code> submission argument.


To find out what accounts and partitions you have access to, use the <code>show_assoc</code> command.
=== Partition QoS ===
Partition QoS are used to limit resources used by all jobs running in a partition, either per user (MaxTRESPU) or per group (GrpTRES).
 
To view partition QoS, use the <code>show_partition_qos</code> command.
 
<pre>
$ show_partition_qos
                Name MaxSubmitPU                      MaxTRESPU              GrpTRES
-------------------- ----------- ------------------------------ --------------------
          scavenger        500  cpu=576,gres/gpu=72,mem=2304G
  scavenger-aarch64        500          cpu=1600,mem=281140M
                tron        500    cpu=32,gres/gpu=4,mem=256G
</pre>
 
If you want to see all partition QoS, including those that you do not have access to, you can use the <code>show_partition_qos --all</code> command.
 
<pre>
$ show_partition_qos --all
                Name MaxSubmitPU                      MaxTRESPU              GrpTRES
-------------------- ----------- ------------------------------ --------------------
                cbcb        500                                cpu=1228,mem=48003G
          cbcb-heng        500
    cbcb-interactive        500
              class        500    cpu=32,gres/gpu=4,mem=256G
                clip        500                                  cpu=564,mem=5590G
                cml        500                                cpu=1096,mem=10890G
            cml-cpu        500
        cml-furongh        500
      cml-scavenger        500                    gres/gpu=24
          cml-wriva        500
            cml-zhou        500
              gamma        500                                  cpu=616,mem=5204G
                mbrc        500                                  cpu=240,mem=2345G
                mc2        500                                  cpu=312,mem=3092G
              oasis        500
              quics        500                                  cpu=328,mem=3484G
          scavenger        500  cpu=576,gres/gpu=72,mem=2304G
  scavenger-aarch64        500          cpu=1600,mem=281140M
                tron        500    cpu=32,gres/gpu=4,mem=256G
              vulcan        500                                cpu=1272,mem=11710G
      vulcan-ampere        500
          vulcan-cpu        500
      vulcan-ramani        500
    vulcan-scavenger        500
</pre>
 
'''NOTE''': These QoS cannot be used directly when submitting jobs, with the exception of the scavenger QoS (i.e., they are not in the AllowQos field for their respective partition). Partition QoS limits apply to all jobs running on a given partition, regardless of what job QoS is used.
 
For example, in the default non-preemption partition (<tt>tron</tt>), you are restricted to 32 total CPU cores, 4 total GPUs, and 256GB total RAM at once across all jobs you have running in the partition.
 
Lab/group-specific partitions may also have their own user limits, and/or may also have group limits on the total number of resources consumed simultaneously by all users that are using their partition, codified by the line in the output above that matches their lab/group name. Note that the values listed above in the two "TRES" columns are not fixed and may fluctuate per-partition as more resources are added to or removed from each partition.
 
'''All partitions also only allow a maximum of 500 submitted (running (R) or pending (PD)) jobs per user in the partition simultaneously.''' This is to prevent excess pending jobs causing [https://slurm.schedmd.com/sched_config.html#backfill backfill] issues with the SLURM scheduler.
* If you need to submit more than 500 jobs in batch at once, you can develop and run an "outer submission script" that repeatedly attempts to run an "inner submission script" (your original submission script) to submit jobs in the batch periodically, until all job submissions are successful. The outer submission script should use looping logic to check if you are at the max job limit and should then retry submission after waiting for some time interval.
: An example outer submission script is as follows. In this example, <code>example_inner.sh</code> is your inner submission script and is not an [[SLURM/ArrayJobs | array job]], and you want to run 1000 jobs. If your inner submission script is an array job, adjust the number of jobs accordingly. Array jobs must be of size 500 or less.
<pre>
#!/bin/bash
numjobs=1000
i=0
while [ $i -lt $numjobs ]
do
  while [[ "$(sbatch example_inner.sh 2>&1)" =~ "QOSMaxSubmitJobPerUserLimit" ]]
  do
    echo "Currently at maximum job submissions allowed."
    echo "Waiting for 5 minutes before trying to submit more jobs."
    sleep 300
  done
  i=$(( $i + 1 ))
  echo "Submitted job $i of $numjobs"
done
</pre>
 
It is suggested that you run the outer submission script in a [[Tmux]] session to keep the terminal window executing it from being interrupted.


= Storage =
= Storage =
All storage available in Nexus is currently [[NFS]] based. We will be introducing some changes for Phase 2 to support high performance GPUDirect Storage (GDS).  These storage allocation procedures will be revised and approved by the launch of Phase 2 by a joint UMIACS and CSD faculty committee.
All network storage available in Nexus is currently [[NFS]] based, and comes in a few different flavors. Compute nodes also have local scratch storage that can be used.


== Home Directories ==
== Home Directories ==
Each user account in UMIACS is allocated 20GB of home directory storage in <code>/fs/nfshomes/$username</code>.  This file system has [[Snapshots | snapshots]] and [[NightlyBackups | backups]] available.  The quota is fixed and cannot be increased.
{{Nfshomes}}


In Phase 2, other standalone compute clusters will fold into partitions in Nexus and you will start to have the same home directory across all systems.
== Scratch Directories ==
Scratch data has no data protection including no snapshots and the data is not backed up. There are two types of scratch directories in the Nexus compute infrastructure:
* Network scratch directories
* Local scratch directories


== Scratch Directories ==
Please note that [[ClassAccounts | class accounts]] do not have network scratch directories.
Each user is allocated a 200GB network scratch directory under <code>/fs/nexus-scratch/$username</code>.  If your network scratch directory is completely filled, you may request a permanent increase of up to 400GB total.  '''This space does not have snapshots and is not backed up.'''  Please ensure that any data you have under your network scratch directory is reproducible.
 
=== Network Scratch Directories ===
You are allocated 200GB of scratch space via NFS from <code>/fs/nexus-scratch/<USERNAME></code> where <USERNAME> is your UMIACS username'''It is not backed up or protected in any way.'''  This directory is '''[[Automounter | automounted]]'''; you will need to <code>cd</code> into the directory or request/specify a fully qualified file path to access it.
 
You can view your quota usage by running <code>df -h /fs/nexus-scratch/<USERNAME></code>.
 
You may request a permanent increase of up to 400GB total space without any faculty approval by [[HelpDesk | contacting staff]]If you need space beyond 400GB, you will need faculty approval and/or a [[#Project_Allocations | project allocation]] for this. If you choose to increase your scratch space beyond 400GB, the increased space is also subject to the 270 TB days limit mentioned in the project allocation section before we check back in for renewal. For example, if you request 1.4TB total space, you may have this for 270 days (1TB beyond the 400GB permanent increase). The amount increased beyond 400GB will also count against your faculty member's 20TB total storage limit mentioned below.
 
This file system is available on all submission, data management, and computational nodes within the cluster.


Each computational node that a user can schedule compute jobs on also has one or more local scratch directories.  These are always named <code>/scratch0</code>, <code>/scratch1</code>, etc.  These are almost always more performant than any other storage available to the job.  However, users must stage their data within the confine of their job and stage the data out before the end of their job.
=== Local Scratch Directories ===
Each computational node that you can schedule compute jobs on also has one or more local scratch directories.  These are always named <code>/scratch0</code>, <code>/scratch1</code>, etc. and '''are not backed up or protected in any way.''' These directories are almost always more performant than any other storage available to the job as they are mounted from disks directly attached to the compute node.  However, you must stage your data within the confines of your job and extract the relevant resultant data elsewhere before the end of your job.


These local scratch directories have a tmpwatch job which will '''delete unmodified data after 90 days'''.  Please make sure you secure any data you write to these directories at the end of your job.
These local scratch directories have a tmpwatch job which will '''delete unaccessed data after 90 days''', scheduled via maintenance jobs to run once a month during our [[MonthlyMaintenanceWindow | monthly maintenance windows]].  Please make sure you secure any resultant data you wish to keep from these directories at the end of your job.


== Faculty Allocations ==
== Faculty Allocations ==
Each faculty is allocated 1TB of lab space when their account is installed.  We also can support grouping these individual allocations together into larger center, lab, or research group allocations if desired by the faculty.  Please [[HelpDesk | contact staff]] to inquire.
Each faculty member can be allocated 1TB of permanent lab space upon request.  We can also support grouping these individual allocations together into larger center, lab, or research group allocations if desired by the faculty.  Please [[HelpDesk | contact staff]] to inquire.


This lab space does not have [[Snapshots | snapshots]] by default (but are available if requested), but is [[NightlyBackups | backed up]].
Lab space storage is fully protected.  It has [[Snapshots | snapshots]] enabled and is [[NightlyBackups | backed up nightly]].


== Project Allocations ==
== Project Allocations ==
Project allocations are available per user for 270 TB days; you can have a 1TB allocation for up to 270 days, a 3TB allocation for 90 days, etc.. A single faculty member can not have more than 20 TB of sponsored account project allocations active at any point.  
Project allocations are available per user for 270 TB days; you can have a 1TB allocation for up to 270 days, a 3TB allocation for 90 days, etc..
 
A single faculty member can not have more than 20TB of project allocations across all of their sponsored accounts active simultaneously. Network scratch allocation space increases beyond the 400GB permanent maximum also have the increase count against this limit (i.e., a 1TB network scratch allocation would have 600GB counted towards this limit).
 
Project storage is fully protected.  It has [[Snapshots | snapshots]] enabled and is [[NightlyBackups | backed up nightly]].


The minimum storage space you can request (maximum length) is 500GB (540 days) and the minimum allocation length you can request (maximum storage) is 30 days (9TB).
The maximum allocation length you can request is 540 days (500GB space) and the maximum storage space you can request is 9TB (30 day length).


To request an allocation, please [[HelpDesk | contact staff]] with your account sponsor involved in the conversation.  Please include the following details:
To request an allocation, please [[HelpDesk | contact staff]] with the faculty member(s) that the project is under involved in the conversation.  Please include the following details:
* Project Name (short)
* Project Name (short)
* Description
* Description
* Size (1TB, 2TB, etc.)
* Size (1TB, 2TB, etc.)
* Length in days (270 days, 135 days, etc.)
* Length in days (270 days, 135 days, etc.)
* Other user(s) that need to access the allocation, if any


These allocations will be available via <code>/fs/nexus-projects/$project_name</code>.
These allocations are available via <code>/fs/nexus-projects/<project name></code>.  '''Renewal is not guaranteed to be available due to limits on the amount of total storage.'''  Near the end of the allocation period, staff will contact you and ask if you are still in need of the storage allocation.  If renewal is available, you can renew for up to another 270 TB days with reapproval from the original faculty approver.
* If you are no longer in need of the storage allocation, you will need to relocate all desired data within two weeks of the end of the allocation period.  Staff will then remove the allocation.
* If you do not respond to staff's request by the end of the allocation period, staff will make the allocation temporarily inaccessible.
** If you do respond asking for renewal but the original faculty approver does not respond within two weeks of the end of the allocation period, staff will also make the allocation temporarily inaccessible.
** If one month from the end of the allocation period is reached without both you and the faculty approver responding, staff will remove the allocation.


== Datasets ==
== Datasets ==
Datasets are hosted in <code>/fs/nexus-datasets</code>.  If you want to request a dataset for for consideration, please [[HelpDesk | contact staff]]. We will have a more formal process to approve datasets by phase 2 of Nexus.  Please note that datasets that require accepting a license will need to be reviewed by [https://ora.umd.edu/ UMD's Office of Research Administration (ORA)] which may require some time to process.
We have read-only dataset storage available at <code>/fs/nexus-datasets</code>.  If there are datasets that you would like to see curated and made available, please see [[Datasets | this page]].
 
The list of Nexus datasets we currently host can be viewed [https://info.umiacs.umd.edu/datasets/list/?q=Nexus here].

Latest revision as of 16:36, 2 December 2024

The Nexus is the combined scheduler of resources in UMIACS. The resource manager for Nexus is SLURM. Resources are arranged into partitions where users are able to schedule computational jobs. Users are arranged into a number of SLURM accounts based on faculty, lab, or center investments.

Getting Started

All accounts in UMIACS are sponsored. If you don't already have a UMIACS account, please see Accounts for information on getting one. You need a full UMIACS account (not a collaborator account) in order to access Nexus.

Access

Your access to submission nodes (alternatively called login nodes) for Nexus computational resources is determined by your account sponsor's department, center, or lab affiliation. You can log into the UMIACS Directory CR application and select the Computational Resource (CR) in the list that has the prefix nexus. The Hosts section lists your available submission nodes - generally a pair of nodes of the format nexus<department, lab, or center abbreviation>[00,01], e.g., nexusgroup00 and nexusgroup01.

Note - UMIACS requires multi-factor authentication through our Duo instance. This is completely discrete from both UMD's and CSD's Duo instances. You will need to enroll one or more factors to access resources in UMIACS, and will be prompted to enroll when you log into the Directory application for the first time.

Once you have identified your submission nodes, you can SSH directly into them. From there, you are able to submit to the cluster via our SLURM workload manager. You need to make sure that your submitted jobs have the correct account, partition, and qos.

Jobs

SLURM jobs are submitted by either srun or sbatch depending if you are doing an interactive job or batch job, respectively. You need to provide the where/how/who to run the job and specify the resources you need to run with.

For the who/where/how, you may be required to specify --account, --partition, and/or --qos (respectively) to be able to adequately submit jobs to the Nexus.

For resources, you may need to specify --time for time, --ntasks for CPUs, --mem for RAM, and --gres=gpu for GPUs in your submission arguments to meet your requirements. There are defaults for all four, so if you don't specify something, you may be scheduled with a very minimal set of time and resources (e.g., by default, NO GPUs are included if you do not specify --gres=gpu). For more information about submission flags for GPU resources, see here. You can also can run man srun on your submission node for a complete list of available submission arguments.

For a list of available GPU types on Nexus and their specs, please see Nexus/GPUs.

For details on how the network for Nexus is architected, please see Nexus/Network. This can be important if you wish to optimize performance of your jobs.

Interactive

Once logged into a submission node, you can run simple interactive jobs. If your session is interrupted from the submission node, the job will be killed. As such, we encourage use of a terminal multiplexer such as Tmux.

$ srun --pty --ntasks=4 --mem=2gb --gres=gpu:1 bash
srun: Job account was unset; set to user default of 'nexus'
srun: Job partition was unset; set to cluster default of 'tron'
srun: Job QoS was unset; set to association default of 'default'
srun: Job time limit was unset; set to partition default of 60 minutes
srun: job 1 queued and waiting for resources
srun: job 1 has been allocated resources
$ hostname
tron62.umiacs.umd.edu
$ nvidia-smi -L
GPU 0: NVIDIA GeForce RTX 2080 Ti (UUID: GPU-daad6a04-a2ce-1183-ce53-b267048f750a)

Batch

Batch jobs are scheduled with a script file with an optional ability to embed job scheduling parameters via variables that are defined by #SBATCH lines at the top of the file. You can find some examples in our SLURM/JobSubmission documentation.

Partitions

The SLURM resource manager uses partitions to act as job queues which can restrict size, time and user limits. The Nexus has a number of different partitions of resources. Different Centers, Labs, and Faculty are able to invest in computational resources that are restricted to approved users through these partitions.

Partitions usable by all non-class account users:

  • Nexus/Tron - Pool of resources available to all non-class accounts sponsored by either UMIACS or CSD faculty.
  • Scavenger - Preemption partition that contains x86_64 architecture nodes from multiple other partitions. More resources are available to schedule simultaneously than in other partitions, however jobs are subject to preemption rules. You are responsible for ensuring your jobs handle this preemption correctly. The SLURM scheduler will simply restart a preempted job with the same submission arguments when it is available to run again. For an overview of things you can check within scripts to determine if your job was preempted/resumed, see SLURM/Preemption.
  • Scavenger (aarch64) - Preemption partition identical in design to scavenger, but only contains aarch64 architecture nodes.

Partitions usable by ClassAccounts:

  • Class - Pool of resources available to class accounts sponsored by either UMIACS or CSD faculty.

Partitions usable by specific lab/center users:

  • Nexus/CBCB - CBCB lab pool available for CBCB lab members.
  • Nexus/CLIP - CLIP lab pool available for CLIP lab members.
  • Nexus/CML - CML lab pool available for CML lab members.
  • Nexus/GAMMA - GAMMA lab pool available for GAMMA lab members.
  • Nexus/MBRC - MBRC lab pool available for MBRC lab members.
  • Nexus/MC2 - MC2 lab pool available for MC2 lab members.
  • Nexus/Vulcan - Vulcan lab pool available for Vulcan lab members.

Quality of Service (QoS)

SLURM uses Quality of Service (QoS) both to provide limits on job sizes (termed by us as "job QoS") as well as to limit resources used by all jobs running in a partition, either per user or per group (termed by us as "partition QoS").

Job QoS

Job QoS are used to provide limits on the size of job that you can run. You should try to allocate only the resources your job actually needs, as resources that each of your jobs schedules are counted against your fair-share priority in the future.

  • default - Default job QoS. Limited to 4 CPU cores, 1 GPU, and 32GB RAM per job. The maximum wall time per job is 3 days.
  • medium - Limited to 8 CPU cores, 2 GPUs, and 64GB RAM per job. The maximum wall time per job is 2 days.
  • high - Limited to 16 CPU cores, 4 GPUs, and 128GB RAM per job. The maximum wall time per job is 1 day.
  • scavenger - No resource limits per job, only a maximum wall time per job of 3 days. You are responsible for ensuring your job requests multiple nodes if it requests resources beyond what any one node is capable of. 576 total CPU cores, 72 total GPUs, and 2304GB total RAM are permitted simultaneously across all of your jobs running with this job QoS. This job QoS is paired 1-1 with the scavenger partition. To use this job QoS, include --partition=scavenger and --account=scavenger in your submission arguments. Do not include any job QoS argument other than --qos=scavenger (optional) or submission will fail.
  • scavenger-aarch64 - No resource limits per job, only a maximum wall time per job of 3 days. You are responsible for ensuring your job requests multiple nodes if it requests resources beyond what any one node is capable of. 1600 total CPU cores and 281140MB total RAM are permitted simultaneously across all of your jobs running with this job QoS. This job QoS is paired 1-1 with the scavenger-aarch64 partition. To use this job QoS, include --partition=scavenger-aarch64, --account=scavenger, and --qos=scavenger-aarch64 in your submission arguments.

You can display these job QoS from the command line using the show_qos command. By default, the command will only show job QoS that you can access. The above four job QoS are the ones that everyone can access.

$ show_qos
                Name     MaxWall                        MaxTRES MaxJobsPU                      MaxTRESPU 
-------------------- ----------- ------------------------------ --------- ------------------------------ 
             default  3-00:00:00       cpu=4,gres/gpu=1,mem=32G                                          
                high  1-00:00:00     cpu=16,gres/gpu=4,mem=128G                                          
              medium  2-00:00:00       cpu=8,gres/gpu=2,mem=64G                                          
           scavenger  3-00:00:00                                           cpu=576,gres/gpu=72,mem=2304G 
   scavenger-aarch64  3-00:00:00                                                    cpu=1600,mem=281140M 

If you want to see all job QoS, including those that you do not have access to, you can use the show_qos --all command.

$ show_qos --all
                Name     MaxWall                        MaxTRES MaxJobsPU                      MaxTRESPU
-------------------- ----------- ------------------------------ --------- ------------------------------
             cml-cpu  7-00:00:00                                        8
         cml-default  7-00:00:00       cpu=4,gres/gpu=1,mem=32G         2
            cml-high  1-12:00:00     cpu=16,gres/gpu=4,mem=128G         2
       cml-high_long 14-00:00:00              cpu=32,gres/gpu=8         8                     gres/gpu=8
          cml-medium  3-00:00:00       cpu=8,gres/gpu=2,mem=64G         2
       cml-scavenger  3-00:00:00                                                             gres/gpu=24
       cml-very_high  1-12:00:00     cpu=32,gres/gpu=8,mem=256G         8                    gres/gpu=12
             default  3-00:00:00       cpu=4,gres/gpu=1,mem=32G
                high  1-00:00:00     cpu=16,gres/gpu=4,mem=128G
             highmem 21-00:00:00                 cpu=128,mem=2T
           huge-long 10-00:00:00     cpu=32,gres/gpu=8,mem=256G
         interactive    12:00:00                 cpu=4,mem=128G
              medium  2-00:00:00       cpu=8,gres/gpu=2,mem=64G
        oasis-exempt 10-00:00:00                                                      cpu=160,mem=28114M
           scavenger  3-00:00:00                                           cpu=576,gres/gpu=72,mem=2304G
   scavenger-aarch64  3-00:00:00                                                    cpu=1600,mem=281140M
          vulcan-cpu  2-00:00:00                cpu=1024,mem=4T         4
      vulcan-default  7-00:00:00       cpu=4,gres/gpu=1,mem=32G         2
       vulcan-exempt  7-00:00:00     cpu=32,gres/gpu=8,mem=256G         2
         vulcan-high  1-12:00:00     cpu=16,gres/gpu=4,mem=128G         2
    vulcan-high_long 14-00:00:00              cpu=32,gres/gpu=8         8                     gres/gpu=8
       vulcan-medium  3-00:00:00       cpu=8,gres/gpu=2,mem=64G         2
       vulcan-sailon  3-00:00:00     cpu=32,gres/gpu=8,mem=256G                              gres/gpu=48
    vulcan-scavenger  3-00:00:00     cpu=32,gres/gpu=8,mem=256G

To find out what accounts and partitions you have access to, first use the show_assoc command to show your account/job QoS combinations. Then, use the scontrol show partition command and note the AllowAccounts entry for each listed partition. You are able to submit to any partition that allows an account that you have. If you need to use an account other than the default account nexus, you will need to specify it via the --account submission argument.

Partition QoS

Partition QoS are used to limit resources used by all jobs running in a partition, either per user (MaxTRESPU) or per group (GrpTRES).

To view partition QoS, use the show_partition_qos command.

$ show_partition_qos
                Name MaxSubmitPU                      MaxTRESPU              GrpTRES
-------------------- ----------- ------------------------------ --------------------
           scavenger         500  cpu=576,gres/gpu=72,mem=2304G
   scavenger-aarch64         500           cpu=1600,mem=281140M
                tron         500     cpu=32,gres/gpu=4,mem=256G

If you want to see all partition QoS, including those that you do not have access to, you can use the show_partition_qos --all command.

$ show_partition_qos --all
                Name MaxSubmitPU                      MaxTRESPU              GrpTRES
-------------------- ----------- ------------------------------ --------------------
                cbcb         500                                 cpu=1228,mem=48003G
           cbcb-heng         500
    cbcb-interactive         500
               class         500     cpu=32,gres/gpu=4,mem=256G
                clip         500                                   cpu=564,mem=5590G
                 cml         500                                 cpu=1096,mem=10890G
             cml-cpu         500
         cml-furongh         500
       cml-scavenger         500                    gres/gpu=24
           cml-wriva         500
            cml-zhou         500
               gamma         500                                   cpu=616,mem=5204G
                mbrc         500                                   cpu=240,mem=2345G
                 mc2         500                                   cpu=312,mem=3092G
               oasis         500
               quics         500                                   cpu=328,mem=3484G
           scavenger         500  cpu=576,gres/gpu=72,mem=2304G
   scavenger-aarch64         500           cpu=1600,mem=281140M
                tron         500     cpu=32,gres/gpu=4,mem=256G
              vulcan         500                                 cpu=1272,mem=11710G
       vulcan-ampere         500
          vulcan-cpu         500
       vulcan-ramani         500
    vulcan-scavenger         500

NOTE: These QoS cannot be used directly when submitting jobs, with the exception of the scavenger QoS (i.e., they are not in the AllowQos field for their respective partition). Partition QoS limits apply to all jobs running on a given partition, regardless of what job QoS is used.

For example, in the default non-preemption partition (tron), you are restricted to 32 total CPU cores, 4 total GPUs, and 256GB total RAM at once across all jobs you have running in the partition.

Lab/group-specific partitions may also have their own user limits, and/or may also have group limits on the total number of resources consumed simultaneously by all users that are using their partition, codified by the line in the output above that matches their lab/group name. Note that the values listed above in the two "TRES" columns are not fixed and may fluctuate per-partition as more resources are added to or removed from each partition.

All partitions also only allow a maximum of 500 submitted (running (R) or pending (PD)) jobs per user in the partition simultaneously. This is to prevent excess pending jobs causing backfill issues with the SLURM scheduler.

  • If you need to submit more than 500 jobs in batch at once, you can develop and run an "outer submission script" that repeatedly attempts to run an "inner submission script" (your original submission script) to submit jobs in the batch periodically, until all job submissions are successful. The outer submission script should use looping logic to check if you are at the max job limit and should then retry submission after waiting for some time interval.
An example outer submission script is as follows. In this example, example_inner.sh is your inner submission script and is not an array job, and you want to run 1000 jobs. If your inner submission script is an array job, adjust the number of jobs accordingly. Array jobs must be of size 500 or less.
#!/bin/bash
numjobs=1000
i=0
while [ $i -lt $numjobs ]
do
  while [[ "$(sbatch example_inner.sh 2>&1)" =~ "QOSMaxSubmitJobPerUserLimit" ]]
  do
    echo "Currently at maximum job submissions allowed."
    echo "Waiting for 5 minutes before trying to submit more jobs."
    sleep 300
  done
  i=$(( $i + 1 ))
  echo "Submitted job $i of $numjobs"
done

It is suggested that you run the outer submission script in a Tmux session to keep the terminal window executing it from being interrupted.

Storage

All network storage available in Nexus is currently NFS based, and comes in a few different flavors. Compute nodes also have local scratch storage that can be used.

Home Directories

You have 30GB of home directory storage available at /nfshomes/<username>. It has both Snapshots and Backups enabled.

Home directories are intended to store personal or configuration files only. We encourage you to not share any data in your home directory. You are encouraged to utilize our GitLab infrastructure to host your code repositories.

NOTE: To check your quota on this directory, use the command df -h ~.

Scratch Directories

Scratch data has no data protection including no snapshots and the data is not backed up. There are two types of scratch directories in the Nexus compute infrastructure:

  • Network scratch directories
  • Local scratch directories

Please note that class accounts do not have network scratch directories.

Network Scratch Directories

You are allocated 200GB of scratch space via NFS from /fs/nexus-scratch/<USERNAME> where <USERNAME> is your UMIACS username. It is not backed up or protected in any way. This directory is automounted; you will need to cd into the directory or request/specify a fully qualified file path to access it.

You can view your quota usage by running df -h /fs/nexus-scratch/<USERNAME>.

You may request a permanent increase of up to 400GB total space without any faculty approval by contacting staff. If you need space beyond 400GB, you will need faculty approval and/or a project allocation for this. If you choose to increase your scratch space beyond 400GB, the increased space is also subject to the 270 TB days limit mentioned in the project allocation section before we check back in for renewal. For example, if you request 1.4TB total space, you may have this for 270 days (1TB beyond the 400GB permanent increase). The amount increased beyond 400GB will also count against your faculty member's 20TB total storage limit mentioned below.

This file system is available on all submission, data management, and computational nodes within the cluster.

Local Scratch Directories

Each computational node that you can schedule compute jobs on also has one or more local scratch directories. These are always named /scratch0, /scratch1, etc. and are not backed up or protected in any way. These directories are almost always more performant than any other storage available to the job as they are mounted from disks directly attached to the compute node. However, you must stage your data within the confines of your job and extract the relevant resultant data elsewhere before the end of your job.

These local scratch directories have a tmpwatch job which will delete unaccessed data after 90 days, scheduled via maintenance jobs to run once a month during our monthly maintenance windows. Please make sure you secure any resultant data you wish to keep from these directories at the end of your job.

Faculty Allocations

Each faculty member can be allocated 1TB of permanent lab space upon request. We can also support grouping these individual allocations together into larger center, lab, or research group allocations if desired by the faculty. Please contact staff to inquire.

Lab space storage is fully protected. It has snapshots enabled and is backed up nightly.

Project Allocations

Project allocations are available per user for 270 TB days; you can have a 1TB allocation for up to 270 days, a 3TB allocation for 90 days, etc..

A single faculty member can not have more than 20TB of project allocations across all of their sponsored accounts active simultaneously. Network scratch allocation space increases beyond the 400GB permanent maximum also have the increase count against this limit (i.e., a 1TB network scratch allocation would have 600GB counted towards this limit).

Project storage is fully protected. It has snapshots enabled and is backed up nightly.

The maximum allocation length you can request is 540 days (500GB space) and the maximum storage space you can request is 9TB (30 day length).

To request an allocation, please contact staff with the faculty member(s) that the project is under involved in the conversation. Please include the following details:

  • Project Name (short)
  • Description
  • Size (1TB, 2TB, etc.)
  • Length in days (270 days, 135 days, etc.)
  • Other user(s) that need to access the allocation, if any

These allocations are available via /fs/nexus-projects/<project name>. Renewal is not guaranteed to be available due to limits on the amount of total storage. Near the end of the allocation period, staff will contact you and ask if you are still in need of the storage allocation. If renewal is available, you can renew for up to another 270 TB days with reapproval from the original faculty approver.

  • If you are no longer in need of the storage allocation, you will need to relocate all desired data within two weeks of the end of the allocation period. Staff will then remove the allocation.
  • If you do not respond to staff's request by the end of the allocation period, staff will make the allocation temporarily inaccessible.
    • If you do respond asking for renewal but the original faculty approver does not respond within two weeks of the end of the allocation period, staff will also make the allocation temporarily inaccessible.
    • If one month from the end of the allocation period is reached without both you and the faculty approver responding, staff will remove the allocation.

Datasets

We have read-only dataset storage available at /fs/nexus-datasets. If there are datasets that you would like to see curated and made available, please see this page.

The list of Nexus datasets we currently host can be viewed here.