<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://wiki.umiacs.umd.edu/umiacs/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Jheager2</id>
	<title>UMIACS - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://wiki.umiacs.umd.edu/umiacs/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Jheager2"/>
	<link rel="alternate" type="text/html" href="https://wiki.umiacs.umd.edu/umiacs/index.php/Special:Contributions/Jheager2"/>
	<updated>2026-04-22T02:16:02Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.43.7</generator>
	<entry>
		<id>https://wiki.umiacs.umd.edu/umiacs/index.php?title=Matlab&amp;diff=9398</id>
		<title>Matlab</title>
		<link rel="alternate" type="text/html" href="https://wiki.umiacs.umd.edu/umiacs/index.php?title=Matlab&amp;diff=9398"/>
		<updated>2020-09-24T19:36:50Z</updated>

		<summary type="html">&lt;p&gt;Jheager2: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;UMIACS has institutional Matlab licenses through UMD that cover Linux, Windows and Mac. To access them, you have to be using a University owned computer. If you would like to run Matlab on a machine that is not University owned, you can check the pricing through OIT/SLIC [https://terpware.umd.edu/Windows/Title/1866 here].&lt;br /&gt;
&lt;br /&gt;
If you would like to install the institutional version of Matlab on a self-supported machine, please [[HelpDesk | contact staff]] and we can provide you install media.&lt;br /&gt;
&lt;br /&gt;
==Linux==&lt;br /&gt;
[[Modules]] are the preferred way to interact with Matlab on staff-managed Linux machines. The Modules page will provide information on adding Matlab into your environment and using it.&lt;br /&gt;
&lt;br /&gt;
==Windows==&lt;br /&gt;
All staff-managed Windows machines should come with a pre-installed version of Matlab. Please [[HelpDesk | contact staff]] if you would like a later version installed.&lt;br /&gt;
&lt;br /&gt;
==macOS==&lt;br /&gt;
Please [[HelpDesk | contact staff]] if you would like Matlab installed on a staff-managed macOS machine.&lt;br /&gt;
&lt;br /&gt;
==License Information==&lt;br /&gt;
Campus has a limited number of seats for Matlab itself, as well as the various toolboxes. When calling functions in a toolbox, Matlab will automatically check out a license for that toolbox, making it temporarily unavailable to other users. If while attempting to use Matlab, you run into a license manager error stating that the maximum number of users for Matlab has been reached, this likely indicates that all of campus&#039; available licenses are in use. To check the number of licenses available for a specific toolbox, please see the below sections:&lt;br /&gt;
&lt;br /&gt;
===Windows License Troubleshooting===&lt;br /&gt;
On Windows machines, the necessary command is as follows. Please note that you may need to adjust the path to reflect the available Matlab versions and the installation path. Additionally, this assumes that you are using a 64-bit version. &amp;quot;win64&amp;quot; needs to be adjusted to &amp;quot;win32&amp;quot; on a 32-bit installation:&lt;br /&gt;
 $MATLAB_PATH\etc\win64\lmutil lmstat -c &amp;quot;$MATLAB_PATH\licenses\network.lic&amp;quot; -f $TOOLBOX_SHORTNAME&lt;br /&gt;
&lt;br /&gt;
As an example, here is how to query Matlab r2012b on a 64-bit Windows host for the Aerospace Blockset Toolbox: &lt;br /&gt;
 &amp;quot;C:\Program Files\matlab\r2012b\etc\win64\lmutil.exe&amp;quot; lmstat -c &amp;quot;C:\Program Files\matlab\r2012b\licenses\network.lic&amp;quot; -f Aerospace_Blockset&lt;br /&gt;
&lt;br /&gt;
===Unix License Troubleshooting===&lt;br /&gt;
On Unix machines, the necessary command is as follows. Please note that you may need to adjust the path to reflect the available Matlab versions and the installation path. Additionally, this assumes that you are using a 64-bit version. &amp;quot;glnxa64&amp;quot; needs to adjusted to &amp;quot;glnx86&amp;quot; on a 32-bit installation::&lt;br /&gt;
 $MATLAB_PATH/etc/glnxa64/lmutil lmstat -c $MATLAB_PATH/licenses/network.lic -f $TOOLBOX_SHORTNAME&lt;br /&gt;
&lt;br /&gt;
As an example, here is how to query Matlab r2015a on our 64-bit RHEL hosts for the Parallel Computing Toolbox:&lt;br /&gt;
 /opt/common/matlab-r2015a/etc/glnxa64/lmutil lmstat -c /opt/common/matlab-r2015a/licenses/network.lic -f Video_and_Image_Blockset&lt;br /&gt;
&lt;br /&gt;
===Toolbox Shortnames===&lt;br /&gt;
The toolbox names themselves don&#039;t always work with this command; rather, Matlab has a set of shortnames which can be passed to the lmstat command. The following are all the valid shortnames:&lt;br /&gt;
&lt;br /&gt;
 Aerospace_Blockset   &lt;br /&gt;
 Aerospace_Toolbox&lt;br /&gt;
 Antenna_Toolbox  &lt;br /&gt;
 Bioinformatics_Toolbox  &lt;br /&gt;
 Communication_Blocks  &lt;br /&gt;
 Communication_Toolbox  &lt;br /&gt;
 Compiler  &lt;br /&gt;
 Control_Toolbox  &lt;br /&gt;
 Curve_Fitting_Toolbox  &lt;br /&gt;
 Data_Acq_Toolbox  &lt;br /&gt;
 Database_Toolbox  &lt;br /&gt;
 Datafeed_Toolbox  &lt;br /&gt;
 Distrib_Computing_Toolbox  &lt;br /&gt;
 Econometrics_Toolbox  &lt;br /&gt;
 Excel_Link  &lt;br /&gt;
 Fin_Derivatives_Toolbox&lt;br /&gt;
 Fin_Instruments_Toolbox  &lt;br /&gt;
 Financial_Toolbox  &lt;br /&gt;
 Fixed_Income_Toolbox  &lt;br /&gt;
 Fixed_Point_Toolbox&lt;br /&gt;
 Fuzzy_Toolbox &lt;br /&gt;
 GADS_Toolbox  &lt;br /&gt;
 Identification_Toolbox  &lt;br /&gt;
 Image_Acquisition_Toolbox  &lt;br /&gt;
 Image_Toolbox  &lt;br /&gt;
 Instr_Control_Toolbox  &lt;br /&gt;
 MAP_Toolbox  &lt;br /&gt;
 MATLAB  &lt;br /&gt;
 MATLAB_Builder_for_Java  &lt;br /&gt;
 MATLAB_Coder  &lt;br /&gt;
 MATLAB_Excel_Builder  &lt;br /&gt;
 MATLAB_Report_Gen  &lt;br /&gt;
 MPC_Toolbox  &lt;br /&gt;
 Neural_Network_Toolbox  &lt;br /&gt;
 Optimization_Toolbox&lt;br /&gt;
 OPC_Toolbox  &lt;br /&gt;
 PDE_Toolbox&lt;br /&gt;
 Phased_Array_System_Toolbox  &lt;br /&gt;
 Real-Time_Workshop  &lt;br /&gt;
 Robotics_System_Toolbox &lt;br /&gt;
 Robust_Toolbox&lt;br /&gt;
 RTW_Embedded_Coder&lt;br /&gt;
 Signal_Blocks  &lt;br /&gt;
 Signal_Toolbox&lt;br /&gt;
 SimEvents  &lt;br /&gt;
 SimMechanics  &lt;br /&gt;
 Simscape  &lt;br /&gt;
 SIMULINK  &lt;br /&gt;
 Simulink_Control_Design&lt;br /&gt;
 Simulink_Design_Optim  &lt;br /&gt;
 Stateflow  &lt;br /&gt;
 Stateflow_Coder  &lt;br /&gt;
 Statistics_Toolbox  &lt;br /&gt;
 Symbolic_Toolbox  &lt;br /&gt;
 Video_and_Image_Blockset  &lt;br /&gt;
 Virtual_Reality_Toolbox  &lt;br /&gt;
 Wavelet_Toolbox&lt;br /&gt;
&lt;br /&gt;
If you are having problems, feel free to contact [[HelpDesk | UMIACS Staff]], however please note that the Matlab license servers are hosted by campus&#039; Division of Information Technology.&lt;br /&gt;
&lt;br /&gt;
==Matlab Customizations==&lt;br /&gt;
You can tweak your Matlab experience using a file &amp;quot;startup.m&amp;quot;. If this file is present in the current working directory where you launch Matlab, it will be executed. You can perform a variety of tasks using this file. These include (but are not limited to) customizing the window environment, setting variables for later use, and enabling use of 3rd party toolboxes. Alternatively, you can use the userpath env variable to set a path other than the current working directory for startup and 3rd party toolbox locations. Please see the following links for more information:&lt;br /&gt;
&lt;br /&gt;
* [http://www.mathworks.com/help/matlab/ref/matlabrc.html Mathworks: Startup.m]&lt;br /&gt;
* [https://www.mathworks.com/help/matlab/ref/userpath.html Mathworks: Userpath function]&lt;/div&gt;</summary>
		<author><name>Jheager2</name></author>
	</entry>
	<entry>
		<id>https://wiki.umiacs.umd.edu/umiacs/index.php?title=MBRC&amp;diff=8528</id>
		<title>MBRC</title>
		<link rel="alternate" type="text/html" href="https://wiki.umiacs.umd.edu/umiacs/index.php?title=MBRC&amp;diff=8528"/>
		<updated>2019-08-22T15:26:02Z</updated>

		<summary type="html">&lt;p&gt;Jheager2: /* Project Directories */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The MBRC ([https://mbrc.umd.edu MBRC]) at the University of Maryland is located within the Institute for Advanced Computer Studies.  The MBRC has a cluster of computational (CPU/GPU) resources that are available to be scheduled.&lt;br /&gt;
&lt;br /&gt;
=Compute Infrastructure=&lt;br /&gt;
&lt;br /&gt;
Each of UMIACS cluster computational infrastructures is accessed through the submission node.  Users will need to submit jobs through the [[SLURM]] resource manager once they have logged into the submission node.  Each cluster in UMIACS has different quality of service (QoS) that need to be selected upon submission of a job and many like this one has specific other resources such as GPUs that need to be requested for a job.  &lt;br /&gt;
&lt;br /&gt;
The current submission node(s) for &#039;&#039;MBRC&#039;&#039; is:&lt;br /&gt;
* &amp;lt;code&amp;gt;mbrcsub00.umiacs.umd.edu&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Partition ==&lt;br /&gt;
There are two partitions to the MBRC [[SLURM]] computational infrastructure.  If you do not specify a partition when submitting your job you will receive the &#039;&#039;&#039;dpart&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;dpart&#039;&#039;&#039; - This is the default partition and job allocations are guaranteed.&lt;br /&gt;
* &#039;&#039;&#039;scavenger&#039;&#039;&#039; - This is the alternate partition that allows jobs longer run times and more resources but is preemptable when jobs in the &#039;&#039;&#039;dpart&#039;&#039;&#039; are ready to be scheduled.&lt;br /&gt;
&lt;br /&gt;
== QoS ==&lt;br /&gt;
MBRC currently has 1 QoS for the &#039;&#039;&#039;dpart&#039;&#039;&#039; and 1 QoS for the &#039;&#039;&#039;scavenger&#039;&#039;&#039; partition.  The important parts here is that in different QoS you can have a shorter/longer maximum wall time, a total number of jobs running at once and a maximum number of track-able resources (TRES) for the job.  In the scavenger QoS there is one more constraint that you are restricted by the total number of TRES per user (over multiple jobs). &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
# show_qos&lt;br /&gt;
      Name     MaxWall MaxJobs                        MaxTRES     MaxTRESPU&lt;br /&gt;
---------- ----------- ------- ------------------------------ -------------&lt;br /&gt;
   default  1-00:00:00       1       cpu=4,mem=32G,gres/gpu=2&lt;br /&gt;
 scavenger  2-00:00:00                                          gres/gpu=8&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== GPUs ==&lt;br /&gt;
Jobs that require GPU resources need to explicitly request the resources within their job submission.  This is done through Generic Resource Scheduling (GRES).  Currently all nodes in the cluster are homogeneous however in the future this may not be the case.  Users may use the most generic identifier in this case &#039;&#039;&#039;gpu&#039;&#039;&#039; a colon and a number to select without explicitly naming the type of GPU (ie. &amp;lt;code&amp;gt;--gres gpu:4&amp;lt;/code&amp;gt; for 4 GPUs).  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ sinfo -o &amp;quot;%15n %10c %10m  %25f %25G&amp;quot;&lt;br /&gt;
NODELIST        CPUS       MEMORY      AVAIL_FEATURES            GRES&lt;br /&gt;
mbrc[00]      32         191896          (null)                    gpu:rtx2080ti:8&lt;br /&gt;
mbrc[01]      32         191896          (null)                    gpu:rtx2080ti:8&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Job Submission and Management ==&lt;br /&gt;
Users should review our [[SLURM]] [[SLURM/JobSubmission | job submission]] and [[SLURM/JobStatus | job management]] documentation.  &lt;br /&gt;
&lt;br /&gt;
A very quick start to get an interactive shell is as follows when run on the submission node.  This will allocate 1 GPU with 16GB of memory (system RAM) in the QoS default for 4 hours maximum time.  If the job goes beyond these limits either the memory allocation or the maximum time it will be terminated immediately. &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
srun --pty --gres=gpu:1 --mem=16G --qos=default --time=04:00:00 bash&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
[jheager2@mbrcsub00:~ ] $ srun --pty --gres=gpu:1 --mem=16G --qos=default --time=04:00:00 bash&lt;br /&gt;
[jheager2@mbrc00:~ ] $ nvidia-smi -L&lt;br /&gt;
GPU 0: GeForce RTX 2080 Ti (UUID: GPU-4ad5c018-b9bc-e664-233a-5d9ee8ad05cb)&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Data Storage=&lt;br /&gt;
&lt;br /&gt;
Until the final storage investment arrives we have made available a temporary allocation of storage.  This section is subject to change.  There are 3 types of storage available to users in the MBRC home directories, project directories and scratch directories.&lt;br /&gt;
&lt;br /&gt;
== Home Directories ==&lt;br /&gt;
 &lt;br /&gt;
Home directories in the MBRC computational infrastructure are available from the Institutes [[NFShomes]] as &amp;lt;code&amp;gt;/nfshomes/USERNAME&amp;lt;/code&amp;gt; where USERNAME is your username.  These home directories have very limited storage and are intended for your personal files, configuration and source code.  Your home directory is &#039;&#039;&#039;not&#039;&#039;&#039; intended for data sets or other large scale data holdings.  Users are encouraged to utilize our [[GitLab]] infrastructure to host your code repositories.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;NOTE&#039;&#039;&#039;: To check your quota on this directory you will need to use the &amp;lt;code&amp;gt;quota -s&amp;lt;/code&amp;gt; command.&lt;br /&gt;
&lt;br /&gt;
Your home directory data is fully protected and has both snapshots and is backed up nightly.&lt;br /&gt;
&lt;br /&gt;
== Project Directories ==&lt;br /&gt;
For this cluster we have decided to allocate network storage on a project by project basis. Jonathan Heagerty will be the point of contact as it pertains to allocating the requested/required storage for each project. As a whole, the MBRC Cluster has limited network storage and for this there will be limits to how much and how long network storage can be appropriated.&lt;br /&gt;
&lt;br /&gt;
If the requested storage size is significantly large relative to the total allotted amount, the request will be relayed from Jonathan Heagerty to the MBRC Cluster faculty for approval. Two other situations that would need approval from the MBRC Cluster faculty would be: To request an increase to a projects current storage allotment or To request a time extension for a projects storage.&lt;br /&gt;
&lt;br /&gt;
When making a request for storage please provide the following information to staff@umiacs.umd.edu:&lt;br /&gt;
        - Name of user requesting storage:&lt;br /&gt;
                Example: jheager2&lt;br /&gt;
        - Name of project:&lt;br /&gt;
                Example: Foveated Rendering&lt;br /&gt;
        - Collaborators working on the project:&lt;br /&gt;
                Example: Sida Li&lt;br /&gt;
        - Storage size:&lt;br /&gt;
                Example: 1TB&lt;br /&gt;
        - Length of time for storage:&lt;br /&gt;
                Example: 6-8 months&lt;br /&gt;
&lt;br /&gt;
== Scratch Directories ==&lt;br /&gt;
&lt;br /&gt;
=== Local Scratch Directory===&lt;br /&gt;
Each computational node that a user can schedule compute jobs on has one or more local scratch directories. These are always named /scratch0, /scratch1 ~ 3.5TB. These are almost always more performant than any other storage available to the job. However users must stage their data within the confine of their job and stage the data out before the end of their job.&lt;/div&gt;</summary>
		<author><name>Jheager2</name></author>
	</entry>
	<entry>
		<id>https://wiki.umiacs.umd.edu/umiacs/index.php?title=MBRC&amp;diff=8527</id>
		<title>MBRC</title>
		<link rel="alternate" type="text/html" href="https://wiki.umiacs.umd.edu/umiacs/index.php?title=MBRC&amp;diff=8527"/>
		<updated>2019-08-22T15:22:19Z</updated>

		<summary type="html">&lt;p&gt;Jheager2: /* Local Scratch Directory */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The MBRC ([https://mbrc.umd.edu MBRC]) at the University of Maryland is located within the Institute for Advanced Computer Studies.  The MBRC has a cluster of computational (CPU/GPU) resources that are available to be scheduled.&lt;br /&gt;
&lt;br /&gt;
=Compute Infrastructure=&lt;br /&gt;
&lt;br /&gt;
Each of UMIACS cluster computational infrastructures is accessed through the submission node.  Users will need to submit jobs through the [[SLURM]] resource manager once they have logged into the submission node.  Each cluster in UMIACS has different quality of service (QoS) that need to be selected upon submission of a job and many like this one has specific other resources such as GPUs that need to be requested for a job.  &lt;br /&gt;
&lt;br /&gt;
The current submission node(s) for &#039;&#039;MBRC&#039;&#039; is:&lt;br /&gt;
* &amp;lt;code&amp;gt;mbrcsub00.umiacs.umd.edu&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Partition ==&lt;br /&gt;
There are two partitions to the MBRC [[SLURM]] computational infrastructure.  If you do not specify a partition when submitting your job you will receive the &#039;&#039;&#039;dpart&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;dpart&#039;&#039;&#039; - This is the default partition and job allocations are guaranteed.&lt;br /&gt;
* &#039;&#039;&#039;scavenger&#039;&#039;&#039; - This is the alternate partition that allows jobs longer run times and more resources but is preemptable when jobs in the &#039;&#039;&#039;dpart&#039;&#039;&#039; are ready to be scheduled.&lt;br /&gt;
&lt;br /&gt;
== QoS ==&lt;br /&gt;
MBRC currently has 1 QoS for the &#039;&#039;&#039;dpart&#039;&#039;&#039; and 1 QoS for the &#039;&#039;&#039;scavenger&#039;&#039;&#039; partition.  The important parts here is that in different QoS you can have a shorter/longer maximum wall time, a total number of jobs running at once and a maximum number of track-able resources (TRES) for the job.  In the scavenger QoS there is one more constraint that you are restricted by the total number of TRES per user (over multiple jobs). &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
# show_qos&lt;br /&gt;
      Name     MaxWall MaxJobs                        MaxTRES     MaxTRESPU&lt;br /&gt;
---------- ----------- ------- ------------------------------ -------------&lt;br /&gt;
   default  1-00:00:00       1       cpu=4,mem=32G,gres/gpu=2&lt;br /&gt;
 scavenger  2-00:00:00                                          gres/gpu=8&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== GPUs ==&lt;br /&gt;
Jobs that require GPU resources need to explicitly request the resources within their job submission.  This is done through Generic Resource Scheduling (GRES).  Currently all nodes in the cluster are homogeneous however in the future this may not be the case.  Users may use the most generic identifier in this case &#039;&#039;&#039;gpu&#039;&#039;&#039; a colon and a number to select without explicitly naming the type of GPU (ie. &amp;lt;code&amp;gt;--gres gpu:4&amp;lt;/code&amp;gt; for 4 GPUs).  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ sinfo -o &amp;quot;%15n %10c %10m  %25f %25G&amp;quot;&lt;br /&gt;
NODELIST        CPUS       MEMORY      AVAIL_FEATURES            GRES&lt;br /&gt;
mbrc[00]      32         191896          (null)                    gpu:rtx2080ti:8&lt;br /&gt;
mbrc[01]      32         191896          (null)                    gpu:rtx2080ti:8&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Job Submission and Management ==&lt;br /&gt;
Users should review our [[SLURM]] [[SLURM/JobSubmission | job submission]] and [[SLURM/JobStatus | job management]] documentation.  &lt;br /&gt;
&lt;br /&gt;
A very quick start to get an interactive shell is as follows when run on the submission node.  This will allocate 1 GPU with 16GB of memory (system RAM) in the QoS default for 4 hours maximum time.  If the job goes beyond these limits either the memory allocation or the maximum time it will be terminated immediately. &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
srun --pty --gres=gpu:1 --mem=16G --qos=default --time=04:00:00 bash&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
[jheager2@mbrcsub00:~ ] $ srun --pty --gres=gpu:1 --mem=16G --qos=default --time=04:00:00 bash&lt;br /&gt;
[jheager2@mbrc00:~ ] $ nvidia-smi -L&lt;br /&gt;
GPU 0: GeForce RTX 2080 Ti (UUID: GPU-4ad5c018-b9bc-e664-233a-5d9ee8ad05cb)&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Data Storage=&lt;br /&gt;
&lt;br /&gt;
Until the final storage investment arrives we have made available a temporary allocation of storage.  This section is subject to change.  There are 3 types of storage available to users in the MBRC home directories, project directories and scratch directories.&lt;br /&gt;
&lt;br /&gt;
== Home Directories ==&lt;br /&gt;
 &lt;br /&gt;
Home directories in the MBRC computational infrastructure are available from the Institutes [[NFShomes]] as &amp;lt;code&amp;gt;/nfshomes/USERNAME&amp;lt;/code&amp;gt; where USERNAME is your username.  These home directories have very limited storage and are intended for your personal files, configuration and source code.  Your home directory is &#039;&#039;&#039;not&#039;&#039;&#039; intended for data sets or other large scale data holdings.  Users are encouraged to utilize our [[GitLab]] infrastructure to host your code repositories.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;NOTE&#039;&#039;&#039;: To check your quota on this directory you will need to use the &amp;lt;code&amp;gt;quota -s&amp;lt;/code&amp;gt; command.&lt;br /&gt;
&lt;br /&gt;
Your home directory data is fully protected and has both snapshots and is backed up nightly.&lt;br /&gt;
&lt;br /&gt;
== Project Directories ==&lt;br /&gt;
For this cluster we have decided to allocate network storage on a project by project basis. Jonathan Heagerty will be the point of contact as it pertains to allocating the requested/required storage for each project. As a whole, the MBRC Cluster has limited network storage and for this there will be limits to how much and how long network storage can be appropriated.&lt;br /&gt;
&lt;br /&gt;
If the requested storage size is significantly large relative to the total allotted amount, the request will be relayed from Jonathan Heagerty to the MBRC Cluster faculty for approval. Two other situations that would need approval from the MBRC Cluster faculty would be: To request an increase to a projects current storage allotment or To request a time extension for a projects storage.&lt;br /&gt;
&lt;br /&gt;
When making a request for storage please provide the following information:&lt;br /&gt;
        - Name of user requesting storage:&lt;br /&gt;
                Example: jheager2&lt;br /&gt;
        - Name of project:&lt;br /&gt;
                Example: Foveated Rendering&lt;br /&gt;
        - Collaborators working on the project:&lt;br /&gt;
                Example: Sida Li&lt;br /&gt;
        - Storage size:&lt;br /&gt;
                Example: 1TB&lt;br /&gt;
        - Length of time for storage:&lt;br /&gt;
                Example: 6-8 months&lt;br /&gt;
&lt;br /&gt;
== Scratch Directories ==&lt;br /&gt;
&lt;br /&gt;
=== Local Scratch Directory===&lt;br /&gt;
Each computational node that a user can schedule compute jobs on has one or more local scratch directories. These are always named /scratch0, /scratch1 ~ 3.5TB. These are almost always more performant than any other storage available to the job. However users must stage their data within the confine of their job and stage the data out before the end of their job.&lt;/div&gt;</summary>
		<author><name>Jheager2</name></author>
	</entry>
	<entry>
		<id>https://wiki.umiacs.umd.edu/umiacs/index.php?title=MBRC&amp;diff=8526</id>
		<title>MBRC</title>
		<link rel="alternate" type="text/html" href="https://wiki.umiacs.umd.edu/umiacs/index.php?title=MBRC&amp;diff=8526"/>
		<updated>2019-08-22T15:18:02Z</updated>

		<summary type="html">&lt;p&gt;Jheager2: /* Local Scratch Directory */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The MBRC ([https://mbrc.umd.edu MBRC]) at the University of Maryland is located within the Institute for Advanced Computer Studies.  The MBRC has a cluster of computational (CPU/GPU) resources that are available to be scheduled.&lt;br /&gt;
&lt;br /&gt;
=Compute Infrastructure=&lt;br /&gt;
&lt;br /&gt;
Each of UMIACS cluster computational infrastructures is accessed through the submission node.  Users will need to submit jobs through the [[SLURM]] resource manager once they have logged into the submission node.  Each cluster in UMIACS has different quality of service (QoS) that need to be selected upon submission of a job and many like this one has specific other resources such as GPUs that need to be requested for a job.  &lt;br /&gt;
&lt;br /&gt;
The current submission node(s) for &#039;&#039;MBRC&#039;&#039; is:&lt;br /&gt;
* &amp;lt;code&amp;gt;mbrcsub00.umiacs.umd.edu&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Partition ==&lt;br /&gt;
There are two partitions to the MBRC [[SLURM]] computational infrastructure.  If you do not specify a partition when submitting your job you will receive the &#039;&#039;&#039;dpart&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;dpart&#039;&#039;&#039; - This is the default partition and job allocations are guaranteed.&lt;br /&gt;
* &#039;&#039;&#039;scavenger&#039;&#039;&#039; - This is the alternate partition that allows jobs longer run times and more resources but is preemptable when jobs in the &#039;&#039;&#039;dpart&#039;&#039;&#039; are ready to be scheduled.&lt;br /&gt;
&lt;br /&gt;
== QoS ==&lt;br /&gt;
MBRC currently has 1 QoS for the &#039;&#039;&#039;dpart&#039;&#039;&#039; and 1 QoS for the &#039;&#039;&#039;scavenger&#039;&#039;&#039; partition.  The important parts here is that in different QoS you can have a shorter/longer maximum wall time, a total number of jobs running at once and a maximum number of track-able resources (TRES) for the job.  In the scavenger QoS there is one more constraint that you are restricted by the total number of TRES per user (over multiple jobs). &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
# show_qos&lt;br /&gt;
      Name     MaxWall MaxJobs                        MaxTRES     MaxTRESPU&lt;br /&gt;
---------- ----------- ------- ------------------------------ -------------&lt;br /&gt;
   default  1-00:00:00       1       cpu=4,mem=32G,gres/gpu=2&lt;br /&gt;
 scavenger  2-00:00:00                                          gres/gpu=8&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== GPUs ==&lt;br /&gt;
Jobs that require GPU resources need to explicitly request the resources within their job submission.  This is done through Generic Resource Scheduling (GRES).  Currently all nodes in the cluster are homogeneous however in the future this may not be the case.  Users may use the most generic identifier in this case &#039;&#039;&#039;gpu&#039;&#039;&#039; a colon and a number to select without explicitly naming the type of GPU (ie. &amp;lt;code&amp;gt;--gres gpu:4&amp;lt;/code&amp;gt; for 4 GPUs).  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ sinfo -o &amp;quot;%15n %10c %10m  %25f %25G&amp;quot;&lt;br /&gt;
NODELIST        CPUS       MEMORY      AVAIL_FEATURES            GRES&lt;br /&gt;
mbrc[00]      32         191896          (null)                    gpu:rtx2080ti:8&lt;br /&gt;
mbrc[01]      32         191896          (null)                    gpu:rtx2080ti:8&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Job Submission and Management ==&lt;br /&gt;
Users should review our [[SLURM]] [[SLURM/JobSubmission | job submission]] and [[SLURM/JobStatus | job management]] documentation.  &lt;br /&gt;
&lt;br /&gt;
A very quick start to get an interactive shell is as follows when run on the submission node.  This will allocate 1 GPU with 16GB of memory (system RAM) in the QoS default for 4 hours maximum time.  If the job goes beyond these limits either the memory allocation or the maximum time it will be terminated immediately. &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
srun --pty --gres=gpu:1 --mem=16G --qos=default --time=04:00:00 bash&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
[jheager2@mbrcsub00:~ ] $ srun --pty --gres=gpu:1 --mem=16G --qos=default --time=04:00:00 bash&lt;br /&gt;
[jheager2@mbrc00:~ ] $ nvidia-smi -L&lt;br /&gt;
GPU 0: GeForce RTX 2080 Ti (UUID: GPU-4ad5c018-b9bc-e664-233a-5d9ee8ad05cb)&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Data Storage=&lt;br /&gt;
&lt;br /&gt;
Until the final storage investment arrives we have made available a temporary allocation of storage.  This section is subject to change.  There are 3 types of storage available to users in the MBRC home directories, project directories and scratch directories.&lt;br /&gt;
&lt;br /&gt;
== Home Directories ==&lt;br /&gt;
 &lt;br /&gt;
Home directories in the MBRC computational infrastructure are available from the Institutes [[NFShomes]] as &amp;lt;code&amp;gt;/nfshomes/USERNAME&amp;lt;/code&amp;gt; where USERNAME is your username.  These home directories have very limited storage and are intended for your personal files, configuration and source code.  Your home directory is &#039;&#039;&#039;not&#039;&#039;&#039; intended for data sets or other large scale data holdings.  Users are encouraged to utilize our [[GitLab]] infrastructure to host your code repositories.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;NOTE&#039;&#039;&#039;: To check your quota on this directory you will need to use the &amp;lt;code&amp;gt;quota -s&amp;lt;/code&amp;gt; command.&lt;br /&gt;
&lt;br /&gt;
Your home directory data is fully protected and has both snapshots and is backed up nightly.&lt;br /&gt;
&lt;br /&gt;
== Project Directories ==&lt;br /&gt;
For this cluster we have decided to allocate network storage on a project by project basis. Jonathan Heagerty will be the point of contact as it pertains to allocating the requested/required storage for each project. As a whole, the MBRC Cluster has limited network storage and for this there will be limits to how much and how long network storage can be appropriated.&lt;br /&gt;
&lt;br /&gt;
If the requested storage size is significantly large relative to the total allotted amount, the request will be relayed from Jonathan Heagerty to the MBRC Cluster faculty for approval. Two other situations that would need approval from the MBRC Cluster faculty would be: To request an increase to a projects current storage allotment or To request a time extension for a projects storage.&lt;br /&gt;
&lt;br /&gt;
When making a request for storage please provide the following information:&lt;br /&gt;
        - Name of user requesting storage:&lt;br /&gt;
                Example: jheager2&lt;br /&gt;
        - Name of project:&lt;br /&gt;
                Example: Foveated Rendering&lt;br /&gt;
        - Collaborators working on the project:&lt;br /&gt;
                Example: Sida Li&lt;br /&gt;
        - Storage size:&lt;br /&gt;
                Example: 1TB&lt;br /&gt;
        - Length of time for storage:&lt;br /&gt;
                Example: 6-8 months&lt;br /&gt;
&lt;br /&gt;
== Scratch Directories ==&lt;br /&gt;
&lt;br /&gt;
=== Local Scratch Directory===&lt;br /&gt;
Each computational node that a user can schedule compute jobs on has one or more local scratch directories. These are always named /scratch0 ~ 240GB, /scratch1 ~ 4TB. These are almost always more performant than any other storage available to the job. However users must stage their data within the confine of their job and stage the data out before the end of their job.&lt;/div&gt;</summary>
		<author><name>Jheager2</name></author>
	</entry>
	<entry>
		<id>https://wiki.umiacs.umd.edu/umiacs/index.php?title=MBRC&amp;diff=8525</id>
		<title>MBRC</title>
		<link rel="alternate" type="text/html" href="https://wiki.umiacs.umd.edu/umiacs/index.php?title=MBRC&amp;diff=8525"/>
		<updated>2019-08-22T15:17:51Z</updated>

		<summary type="html">&lt;p&gt;Jheager2: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The MBRC ([https://mbrc.umd.edu MBRC]) at the University of Maryland is located within the Institute for Advanced Computer Studies.  The MBRC has a cluster of computational (CPU/GPU) resources that are available to be scheduled.&lt;br /&gt;
&lt;br /&gt;
=Compute Infrastructure=&lt;br /&gt;
&lt;br /&gt;
Each of UMIACS cluster computational infrastructures is accessed through the submission node.  Users will need to submit jobs through the [[SLURM]] resource manager once they have logged into the submission node.  Each cluster in UMIACS has different quality of service (QoS) that need to be selected upon submission of a job and many like this one has specific other resources such as GPUs that need to be requested for a job.  &lt;br /&gt;
&lt;br /&gt;
The current submission node(s) for &#039;&#039;MBRC&#039;&#039; is:&lt;br /&gt;
* &amp;lt;code&amp;gt;mbrcsub00.umiacs.umd.edu&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Partition ==&lt;br /&gt;
There are two partitions to the MBRC [[SLURM]] computational infrastructure.  If you do not specify a partition when submitting your job you will receive the &#039;&#039;&#039;dpart&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;dpart&#039;&#039;&#039; - This is the default partition and job allocations are guaranteed.&lt;br /&gt;
* &#039;&#039;&#039;scavenger&#039;&#039;&#039; - This is the alternate partition that allows jobs longer run times and more resources but is preemptable when jobs in the &#039;&#039;&#039;dpart&#039;&#039;&#039; are ready to be scheduled.&lt;br /&gt;
&lt;br /&gt;
== QoS ==&lt;br /&gt;
MBRC currently has 1 QoS for the &#039;&#039;&#039;dpart&#039;&#039;&#039; and 1 QoS for the &#039;&#039;&#039;scavenger&#039;&#039;&#039; partition.  The important parts here is that in different QoS you can have a shorter/longer maximum wall time, a total number of jobs running at once and a maximum number of track-able resources (TRES) for the job.  In the scavenger QoS there is one more constraint that you are restricted by the total number of TRES per user (over multiple jobs). &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
# show_qos&lt;br /&gt;
      Name     MaxWall MaxJobs                        MaxTRES     MaxTRESPU&lt;br /&gt;
---------- ----------- ------- ------------------------------ -------------&lt;br /&gt;
   default  1-00:00:00       1       cpu=4,mem=32G,gres/gpu=2&lt;br /&gt;
 scavenger  2-00:00:00                                          gres/gpu=8&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== GPUs ==&lt;br /&gt;
Jobs that require GPU resources need to explicitly request the resources within their job submission.  This is done through Generic Resource Scheduling (GRES).  Currently all nodes in the cluster are homogeneous however in the future this may not be the case.  Users may use the most generic identifier in this case &#039;&#039;&#039;gpu&#039;&#039;&#039; a colon and a number to select without explicitly naming the type of GPU (ie. &amp;lt;code&amp;gt;--gres gpu:4&amp;lt;/code&amp;gt; for 4 GPUs).  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ sinfo -o &amp;quot;%15n %10c %10m  %25f %25G&amp;quot;&lt;br /&gt;
NODELIST        CPUS       MEMORY      AVAIL_FEATURES            GRES&lt;br /&gt;
mbrc[00]      32         191896          (null)                    gpu:rtx2080ti:8&lt;br /&gt;
mbrc[01]      32         191896          (null)                    gpu:rtx2080ti:8&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Job Submission and Management ==&lt;br /&gt;
Users should review our [[SLURM]] [[SLURM/JobSubmission | job submission]] and [[SLURM/JobStatus | job management]] documentation.  &lt;br /&gt;
&lt;br /&gt;
A very quick start to get an interactive shell is as follows when run on the submission node.  This will allocate 1 GPU with 16GB of memory (system RAM) in the QoS default for 4 hours maximum time.  If the job goes beyond these limits either the memory allocation or the maximum time it will be terminated immediately. &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
srun --pty --gres=gpu:1 --mem=16G --qos=default --time=04:00:00 bash&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
[jheager2@mbrcsub00:~ ] $ srun --pty --gres=gpu:1 --mem=16G --qos=default --time=04:00:00 bash&lt;br /&gt;
[jheager2@mbrc00:~ ] $ nvidia-smi -L&lt;br /&gt;
GPU 0: GeForce RTX 2080 Ti (UUID: GPU-4ad5c018-b9bc-e664-233a-5d9ee8ad05cb)&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Data Storage=&lt;br /&gt;
&lt;br /&gt;
Until the final storage investment arrives we have made available a temporary allocation of storage.  This section is subject to change.  There are 3 types of storage available to users in the MBRC home directories, project directories and scratch directories.&lt;br /&gt;
&lt;br /&gt;
== Home Directories ==&lt;br /&gt;
 &lt;br /&gt;
Home directories in the MBRC computational infrastructure are available from the Institutes [[NFShomes]] as &amp;lt;code&amp;gt;/nfshomes/USERNAME&amp;lt;/code&amp;gt; where USERNAME is your username.  These home directories have very limited storage and are intended for your personal files, configuration and source code.  Your home directory is &#039;&#039;&#039;not&#039;&#039;&#039; intended for data sets or other large scale data holdings.  Users are encouraged to utilize our [[GitLab]] infrastructure to host your code repositories.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;NOTE&#039;&#039;&#039;: To check your quota on this directory you will need to use the &amp;lt;code&amp;gt;quota -s&amp;lt;/code&amp;gt; command.&lt;br /&gt;
&lt;br /&gt;
Your home directory data is fully protected and has both snapshots and is backed up nightly.&lt;br /&gt;
&lt;br /&gt;
== Project Directories ==&lt;br /&gt;
For this cluster we have decided to allocate network storage on a project by project basis. Jonathan Heagerty will be the point of contact as it pertains to allocating the requested/required storage for each project. As a whole, the MBRC Cluster has limited network storage and for this there will be limits to how much and how long network storage can be appropriated.&lt;br /&gt;
&lt;br /&gt;
If the requested storage size is significantly large relative to the total allotted amount, the request will be relayed from Jonathan Heagerty to the MBRC Cluster faculty for approval. Two other situations that would need approval from the MBRC Cluster faculty would be: To request an increase to a projects current storage allotment or To request a time extension for a projects storage.&lt;br /&gt;
&lt;br /&gt;
When making a request for storage please provide the following information:&lt;br /&gt;
        - Name of user requesting storage:&lt;br /&gt;
                Example: jheager2&lt;br /&gt;
        - Name of project:&lt;br /&gt;
                Example: Foveated Rendering&lt;br /&gt;
        - Collaborators working on the project:&lt;br /&gt;
                Example: Sida Li&lt;br /&gt;
        - Storage size:&lt;br /&gt;
                Example: 1TB&lt;br /&gt;
        - Length of time for storage:&lt;br /&gt;
                Example: 6-8 months&lt;br /&gt;
&lt;br /&gt;
== Scratch Directories ==&lt;br /&gt;
&lt;br /&gt;
=== Local Scratch Directory===&lt;br /&gt;
Each computational node that a user can schedule compute jobs on has one or more local scratch directories. These are always named /scratch0 ~240GB, /scratch1 ~4TB. These are almost always more performant than any other storage available to the job. However users must stage their data within the confine of their job and stage the data out before the end of their job.&lt;/div&gt;</summary>
		<author><name>Jheager2</name></author>
	</entry>
	<entry>
		<id>https://wiki.umiacs.umd.edu/umiacs/index.php?title=MBRC&amp;diff=8517</id>
		<title>MBRC</title>
		<link rel="alternate" type="text/html" href="https://wiki.umiacs.umd.edu/umiacs/index.php?title=MBRC&amp;diff=8517"/>
		<updated>2019-08-12T19:45:10Z</updated>

		<summary type="html">&lt;p&gt;Jheager2: /* Project Directories */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The MBRC ([https://mbrc.umd.edu MBRC]) at the University of Maryland is located within the Institute for Advanced Computer Studies.  The MBRC has a cluster of computational (CPU/GPU) resources that are available to be scheduled.&lt;br /&gt;
&lt;br /&gt;
=Compute Infrastructure=&lt;br /&gt;
&lt;br /&gt;
Each of UMIACS cluster computational infrastructures is accessed through the submission node.  Users will need to submit jobs through the [[SLURM]] resource manager once they have logged into the submission node.  Each cluster in UMIACS has different quality of service (QoS) that need to be selected upon submission of a job and many like this one has specific other resources such as GPUs that need to be requested for a job.  &lt;br /&gt;
&lt;br /&gt;
The current submission node(s) for &#039;&#039;MBRC&#039;&#039; is:&lt;br /&gt;
* &amp;lt;code&amp;gt;mbrcsub00.umiacs.umd.edu&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Partition ==&lt;br /&gt;
There are two partitions to the MBRC [[SLURM]] computational infrastructure.  If you do not specify a partition when submitting your job you will receive the &#039;&#039;&#039;dpart&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;dpart&#039;&#039;&#039; - This is the default partition and job allocations are guaranteed.&lt;br /&gt;
* &#039;&#039;&#039;scavenger&#039;&#039;&#039; - This is the alternate partition that allows jobs longer run times and more resources but is preemptable when jobs in the &#039;&#039;&#039;dpart&#039;&#039;&#039; are ready to be scheduled.&lt;br /&gt;
&lt;br /&gt;
== QoS ==&lt;br /&gt;
MBRC currently has 1 QoS for the &#039;&#039;&#039;dpart&#039;&#039;&#039; and 1 QoS for the &#039;&#039;&#039;scavenger&#039;&#039;&#039; partition.  The important parts here is that in different QoS you can have a shorter/longer maximum wall time, a total number of jobs running at once and a maximum number of track-able resources (TRES) for the job.  In the scavenger QoS there is one more constraint that you are restricted by the total number of TRES per user (over multiple jobs). &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
# show_qos&lt;br /&gt;
      Name     MaxWall MaxJobs                        MaxTRES     MaxTRESPU&lt;br /&gt;
---------- ----------- ------- ------------------------------ -------------&lt;br /&gt;
   default  1-00:00:00       1       cpu=4,mem=32G,gres/gpu=2&lt;br /&gt;
 scavenger  2-00:00:00                                          gres/gpu=8&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== GPUs ==&lt;br /&gt;
Jobs that require GPU resources need to explicitly request the resources within their job submission.  This is done through Generic Resource Scheduling (GRES).  Currently all nodes in the cluster are homogeneous however in the future this may not be the case.  Users may use the most generic identifier in this case &#039;&#039;&#039;gpu&#039;&#039;&#039; a colon and a number to select without explicitly naming the type of GPU (ie. &amp;lt;code&amp;gt;--gres gpu:4&amp;lt;/code&amp;gt; for 4 GPUs).  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ sinfo -o &amp;quot;%15n %10c %10m  %25f %25G&amp;quot;&lt;br /&gt;
NODELIST        CPUS       MEMORY      AVAIL_FEATURES            GRES&lt;br /&gt;
mbrc[00]      32         191896          (null)                    gpu:rtx2080ti:8&lt;br /&gt;
mbrc[01]      32         191896          (null)                    gpu:rtx2080ti:8&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Job Submission and Management ==&lt;br /&gt;
Users should review our [[SLURM]] [[SLURM/JobSubmission | job submission]] and [[SLURM/JobStatus | job management]] documentation.  &lt;br /&gt;
&lt;br /&gt;
A very quick start to get an interactive shell is as follows when run on the submission node.  This will allocate 1 GPU with 16GB of memory (system RAM) in the QoS default for 4 hours maximum time.  If the job goes beyond these limits either the memory allocation or the maximum time it will be terminated immediately. &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
srun --pty --gres=gpu:1 --mem=16G --qos=default --time=04:00:00 bash&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
[jheager2@mbrcsub00:~ ] $ srun --pty --gres=gpu:1 --mem=16G --qos=default --time=04:00:00 bash&lt;br /&gt;
[jheager2@mbrc00:~ ] $ nvidia-smi -L&lt;br /&gt;
GPU 0: GeForce RTX 2080 Ti (UUID: GPU-4ad5c018-b9bc-e664-233a-5d9ee8ad05cb)&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Data Storage=&lt;br /&gt;
&lt;br /&gt;
Until the final storage investment arrives we have made available a temporary allocation of storage.  This section is subject to change.  There are 3 types of storage available to users in the MBRC home directories, project directories and scratch directories.&lt;br /&gt;
&lt;br /&gt;
== Home Directories ==&lt;br /&gt;
 &lt;br /&gt;
Home directories in the MBRC computational infrastructure are available from the Institutes [[NFShomes]] as &amp;lt;code&amp;gt;/nfshomes/USERNAME&amp;lt;/code&amp;gt; where USERNAME is your username.  These home directories have very limited storage and are intended for your personal files, configuration and source code.  Your home directory is &#039;&#039;&#039;not&#039;&#039;&#039; intended for data sets or other large scale data holdings.  Users are encouraged to utilize our [[GitLab]] infrastructure to host your code repositories.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;NOTE&#039;&#039;&#039;: To check your quota on this directory you will need to use the &amp;lt;code&amp;gt;quota -s&amp;lt;/code&amp;gt; command.&lt;br /&gt;
&lt;br /&gt;
Your home directory data is fully protected and has both snapshots and is backed up nightly.&lt;br /&gt;
&lt;br /&gt;
== Project Directories ==&lt;br /&gt;
For this cluster we have decided to allocate network storage on a project by project basis. Jonathan Heagerty will be the point of contact as it pertains to allocating the requested/required storage for each project. As a whole, the MBRC Cluster has limited network storage and for this there will be limits to how much and how long network storage can be appropriated.&lt;br /&gt;
&lt;br /&gt;
If the requested storage size is significantly large relative to the total allotted amount, the request will be relayed from Jonathan Heagerty to the MBRC Cluster faculty for approval. Two other situations that would need approval from the MBRC Cluster faculty would be: To request an increase to a projects current storage allotment or To request a time extension for a projects storage.&lt;br /&gt;
&lt;br /&gt;
When making a request for storage please provide the following information:&lt;br /&gt;
        - Name of user requesting storage:&lt;br /&gt;
                Example: jheager2&lt;br /&gt;
        - Name of project:&lt;br /&gt;
                Example: Foveated Rendering&lt;br /&gt;
        - Collaborators working on the project:&lt;br /&gt;
                Example: Sida Li&lt;br /&gt;
        - Storage size:&lt;br /&gt;
                Example: 1TB&lt;br /&gt;
        - Length of time for storage:&lt;br /&gt;
                Example: 6-8 months&lt;br /&gt;
&lt;br /&gt;
== Scratch Directories ==&lt;br /&gt;
&lt;br /&gt;
=== Local Scratch Directory===&lt;/div&gt;</summary>
		<author><name>Jheager2</name></author>
	</entry>
	<entry>
		<id>https://wiki.umiacs.umd.edu/umiacs/index.php?title=MBRC&amp;diff=8516</id>
		<title>MBRC</title>
		<link rel="alternate" type="text/html" href="https://wiki.umiacs.umd.edu/umiacs/index.php?title=MBRC&amp;diff=8516"/>
		<updated>2019-08-12T19:44:59Z</updated>

		<summary type="html">&lt;p&gt;Jheager2: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The MBRC ([https://mbrc.umd.edu MBRC]) at the University of Maryland is located within the Institute for Advanced Computer Studies.  The MBRC has a cluster of computational (CPU/GPU) resources that are available to be scheduled.&lt;br /&gt;
&lt;br /&gt;
=Compute Infrastructure=&lt;br /&gt;
&lt;br /&gt;
Each of UMIACS cluster computational infrastructures is accessed through the submission node.  Users will need to submit jobs through the [[SLURM]] resource manager once they have logged into the submission node.  Each cluster in UMIACS has different quality of service (QoS) that need to be selected upon submission of a job and many like this one has specific other resources such as GPUs that need to be requested for a job.  &lt;br /&gt;
&lt;br /&gt;
The current submission node(s) for &#039;&#039;MBRC&#039;&#039; is:&lt;br /&gt;
* &amp;lt;code&amp;gt;mbrcsub00.umiacs.umd.edu&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Partition ==&lt;br /&gt;
There are two partitions to the MBRC [[SLURM]] computational infrastructure.  If you do not specify a partition when submitting your job you will receive the &#039;&#039;&#039;dpart&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;dpart&#039;&#039;&#039; - This is the default partition and job allocations are guaranteed.&lt;br /&gt;
* &#039;&#039;&#039;scavenger&#039;&#039;&#039; - This is the alternate partition that allows jobs longer run times and more resources but is preemptable when jobs in the &#039;&#039;&#039;dpart&#039;&#039;&#039; are ready to be scheduled.&lt;br /&gt;
&lt;br /&gt;
== QoS ==&lt;br /&gt;
MBRC currently has 1 QoS for the &#039;&#039;&#039;dpart&#039;&#039;&#039; and 1 QoS for the &#039;&#039;&#039;scavenger&#039;&#039;&#039; partition.  The important parts here is that in different QoS you can have a shorter/longer maximum wall time, a total number of jobs running at once and a maximum number of track-able resources (TRES) for the job.  In the scavenger QoS there is one more constraint that you are restricted by the total number of TRES per user (over multiple jobs). &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
# show_qos&lt;br /&gt;
      Name     MaxWall MaxJobs                        MaxTRES     MaxTRESPU&lt;br /&gt;
---------- ----------- ------- ------------------------------ -------------&lt;br /&gt;
   default  1-00:00:00       1       cpu=4,mem=32G,gres/gpu=2&lt;br /&gt;
 scavenger  2-00:00:00                                          gres/gpu=8&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== GPUs ==&lt;br /&gt;
Jobs that require GPU resources need to explicitly request the resources within their job submission.  This is done through Generic Resource Scheduling (GRES).  Currently all nodes in the cluster are homogeneous however in the future this may not be the case.  Users may use the most generic identifier in this case &#039;&#039;&#039;gpu&#039;&#039;&#039; a colon and a number to select without explicitly naming the type of GPU (ie. &amp;lt;code&amp;gt;--gres gpu:4&amp;lt;/code&amp;gt; for 4 GPUs).  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ sinfo -o &amp;quot;%15n %10c %10m  %25f %25G&amp;quot;&lt;br /&gt;
NODELIST        CPUS       MEMORY      AVAIL_FEATURES            GRES&lt;br /&gt;
mbrc[00]      32         191896          (null)                    gpu:rtx2080ti:8&lt;br /&gt;
mbrc[01]      32         191896          (null)                    gpu:rtx2080ti:8&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Job Submission and Management ==&lt;br /&gt;
Users should review our [[SLURM]] [[SLURM/JobSubmission | job submission]] and [[SLURM/JobStatus | job management]] documentation.  &lt;br /&gt;
&lt;br /&gt;
A very quick start to get an interactive shell is as follows when run on the submission node.  This will allocate 1 GPU with 16GB of memory (system RAM) in the QoS default for 4 hours maximum time.  If the job goes beyond these limits either the memory allocation or the maximum time it will be terminated immediately. &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
srun --pty --gres=gpu:1 --mem=16G --qos=default --time=04:00:00 bash&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
[jheager2@mbrcsub00:~ ] $ srun --pty --gres=gpu:1 --mem=16G --qos=default --time=04:00:00 bash&lt;br /&gt;
[jheager2@mbrc00:~ ] $ nvidia-smi -L&lt;br /&gt;
GPU 0: GeForce RTX 2080 Ti (UUID: GPU-4ad5c018-b9bc-e664-233a-5d9ee8ad05cb)&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Data Storage=&lt;br /&gt;
&lt;br /&gt;
Until the final storage investment arrives we have made available a temporary allocation of storage.  This section is subject to change.  There are 3 types of storage available to users in the MBRC home directories, project directories and scratch directories.&lt;br /&gt;
&lt;br /&gt;
== Home Directories ==&lt;br /&gt;
 &lt;br /&gt;
Home directories in the MBRC computational infrastructure are available from the Institutes [[NFShomes]] as &amp;lt;code&amp;gt;/nfshomes/USERNAME&amp;lt;/code&amp;gt; where USERNAME is your username.  These home directories have very limited storage and are intended for your personal files, configuration and source code.  Your home directory is &#039;&#039;&#039;not&#039;&#039;&#039; intended for data sets or other large scale data holdings.  Users are encouraged to utilize our [[GitLab]] infrastructure to host your code repositories.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;NOTE&#039;&#039;&#039;: To check your quota on this directory you will need to use the &amp;lt;code&amp;gt;quota -s&amp;lt;/code&amp;gt; command.&lt;br /&gt;
&lt;br /&gt;
Your home directory data is fully protected and has both snapshots and is backed up nightly.&lt;br /&gt;
&lt;br /&gt;
== Project Directories ==&lt;br /&gt;
For this cluster we have decided to allocate network storage on a project by project basis. Jonathan Heagerty will be the point of contact as it pertains to allocating the requested/required storage for each project. As a whole, the MBRC Cluster has limited network storage and for this there will be limits to how much and how long network storage can be appropriated.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
If the requested storage size is significantly large relative to the total allotted amount, the request will be relayed from Jonathan Heagerty to the MBRC Cluster faculty for approval. Two other situations that would need approval from the MBRC Cluster faculty would be: To request an increase to a projects current storage allotment or To request a time extension for a projects storage.&lt;br /&gt;
&lt;br /&gt;
When making a request for storage please provide the following information:&lt;br /&gt;
        - Name of user requesting storage:&lt;br /&gt;
                Example: jheager2&lt;br /&gt;
        - Name of project:&lt;br /&gt;
                Example: Foveated Rendering&lt;br /&gt;
        - Collaborators working on the project:&lt;br /&gt;
                Example: Sida Li&lt;br /&gt;
        - Storage size:&lt;br /&gt;
                Example: 1TB&lt;br /&gt;
        - Length of time for storage:&lt;br /&gt;
                Example: 6-8 months&lt;br /&gt;
&lt;br /&gt;
== Scratch Directories ==&lt;br /&gt;
&lt;br /&gt;
=== Local Scratch Directory===&lt;/div&gt;</summary>
		<author><name>Jheager2</name></author>
	</entry>
	<entry>
		<id>https://wiki.umiacs.umd.edu/umiacs/index.php?title=MBRC&amp;diff=8515</id>
		<title>MBRC</title>
		<link rel="alternate" type="text/html" href="https://wiki.umiacs.umd.edu/umiacs/index.php?title=MBRC&amp;diff=8515"/>
		<updated>2019-08-12T19:08:49Z</updated>

		<summary type="html">&lt;p&gt;Jheager2: /* Project Directories */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The MBRC ([https://mbrc.umd.edu MBRC]) at the University of Maryland is located within the Institute for Advanced Computer Studies.  The MBRC has a cluster of computational (CPU/GPU) resources that are available to be scheduled.&lt;br /&gt;
&lt;br /&gt;
=Compute Infrastructure=&lt;br /&gt;
&lt;br /&gt;
Each of UMIACS cluster computational infrastructures is accessed through the submission node.  Users will need to submit jobs through the [[SLURM]] resource manager once they have logged into the submission node.  Each cluster in UMIACS has different quality of service (QoS) that need to be selected upon submission of a job and many like this one has specific other resources such as GPUs that need to be requested for a job.  &lt;br /&gt;
&lt;br /&gt;
The current submission node(s) for &#039;&#039;MBRC&#039;&#039; is:&lt;br /&gt;
* &amp;lt;code&amp;gt;mbrcsub00.umiacs.umd.edu&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Partition ==&lt;br /&gt;
There are two partitions to the MBRC [[SLURM]] computational infrastructure.  If you do not specify a partition when submitting your job you will receive the &#039;&#039;&#039;dpart&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;dpart&#039;&#039;&#039; - This is the default partition and job allocations are guaranteed.&lt;br /&gt;
* &#039;&#039;&#039;scavenger&#039;&#039;&#039; - This is the alternate partition that allows jobs longer run times and more resources but is preemptable when jobs in the &#039;&#039;&#039;dpart&#039;&#039;&#039; are ready to be scheduled.&lt;br /&gt;
&lt;br /&gt;
== QoS ==&lt;br /&gt;
MBRC currently has 1 QoS for the &#039;&#039;&#039;dpart&#039;&#039;&#039; and 1 QoS for the &#039;&#039;&#039;scavenger&#039;&#039;&#039; partition.  The important parts here is that in different QoS you can have a shorter/longer maximum wall time, a total number of jobs running at once and a maximum number of track-able resources (TRES) for the job.  In the scavenger QoS there is one more constraint that you are restricted by the total number of TRES per user (over multiple jobs). &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
# show_qos&lt;br /&gt;
      Name     MaxWall MaxJobs                        MaxTRES     MaxTRESPU&lt;br /&gt;
---------- ----------- ------- ------------------------------ -------------&lt;br /&gt;
   default  1-00:00:00       1       cpu=4,mem=32G,gres/gpu=2&lt;br /&gt;
 scavenger  2-00:00:00                                          gres/gpu=8&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== GPUs ==&lt;br /&gt;
Jobs that require GPU resources need to explicitly request the resources within their job submission.  This is done through Generic Resource Scheduling (GRES).  Currently all nodes in the cluster are homogeneous however in the future this may not be the case.  Users may use the most generic identifier in this case &#039;&#039;&#039;gpu&#039;&#039;&#039; a colon and a number to select without explicitly naming the type of GPU (ie. &amp;lt;code&amp;gt;--gres gpu:4&amp;lt;/code&amp;gt; for 4 GPUs).  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ sinfo -o &amp;quot;%15n %10c %10m  %25f %25G&amp;quot;&lt;br /&gt;
NODELIST        CPUS       MEMORY      AVAIL_FEATURES            GRES&lt;br /&gt;
mbrc[00]      32         191896          (null)                    gpu:rtx2080ti:8&lt;br /&gt;
mbrc[01]      32         191896          (null)                    gpu:rtx2080ti:8&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Job Submission and Management ==&lt;br /&gt;
Users should review our [[SLURM]] [[SLURM/JobSubmission | job submission]] and [[SLURM/JobStatus | job management]] documentation.  &lt;br /&gt;
&lt;br /&gt;
A very quick start to get an interactive shell is as follows when run on the submission node.  This will allocate 1 GPU with 16GB of memory (system RAM) in the QoS default for 4 hours maximum time.  If the job goes beyond these limits either the memory allocation or the maximum time it will be terminated immediately. &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
srun --pty --gres=gpu:1 --mem=16G --qos=default --time=04:00:00 bash&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
[jheager2@mbrcsub00:~ ] $ srun --pty --gres=gpu:1 --mem=16G --qos=default --time=04:00:00 bash&lt;br /&gt;
[jheager2@mbrc00:~ ] $ nvidia-smi -L&lt;br /&gt;
GPU 0: GeForce RTX 2080 Ti (UUID: GPU-4ad5c018-b9bc-e664-233a-5d9ee8ad05cb)&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Data Storage=&lt;br /&gt;
&lt;br /&gt;
Until the final storage investment arrives we have made available a temporary allocation of storage.  This section is subject to change.  There are 3 types of storage available to users in the MBRC home directories, project directories and scratch directories.&lt;br /&gt;
&lt;br /&gt;
== Home Directories ==&lt;br /&gt;
 &lt;br /&gt;
Home directories in the MBRC computational infrastructure are available from the Institutes [[NFShomes]] as &amp;lt;code&amp;gt;/nfshomes/USERNAME&amp;lt;/code&amp;gt; where USERNAME is your username.  These home directories have very limited storage and are intended for your personal files, configuration and source code.  Your home directory is &#039;&#039;&#039;not&#039;&#039;&#039; intended for data sets or other large scale data holdings.  Users are encouraged to utilize our [[GitLab]] infrastructure to host your code repositories.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;NOTE&#039;&#039;&#039;: To check your quota on this directory you will need to use the &amp;lt;code&amp;gt;quota -s&amp;lt;/code&amp;gt; command.&lt;br /&gt;
&lt;br /&gt;
Your home directory data is fully protected and has both snapshots and is backed up nightly.&lt;br /&gt;
&lt;br /&gt;
== Project Directories ==&lt;br /&gt;
For this cluster we have decided to allocate network storage on a project by project basis. Jonathan Heagerty will be the point of contact as it pertains to allocating the requested/required storage for each project. As a whole, there is an allotment of network storage for the MRBC Cluster and is to why there will be limitations to the storage size and lifecycle for the data. &lt;br /&gt;
&lt;br /&gt;
If the requested storage size is significantly large relative to the total allotted amount, the request will be relayed from Jonathan Heagerty to the MBRC Cluster faculty for approval. Two other situations that would need approval from the MBRC Cluster faculty would be: To request an increase to a projects current storage allotment or To request a time extension for a projects storage.&lt;br /&gt;
&lt;br /&gt;
When making a request for storage please provide the following information:&lt;br /&gt;
        - Name of user requesting storage:&lt;br /&gt;
                Example: jheager2&lt;br /&gt;
        - Name of project:&lt;br /&gt;
                Example: Foveated Rendering&lt;br /&gt;
        - Collaborators working on the project:&lt;br /&gt;
                Example: Sida Li&lt;br /&gt;
        - Storage size:&lt;br /&gt;
                Example: 1TB&lt;br /&gt;
        - Length of time for storage:&lt;br /&gt;
                Example: 6-8 months&lt;br /&gt;
&lt;br /&gt;
== Scratch Directories ==&lt;br /&gt;
&lt;br /&gt;
=== Local Scratch Directory===&lt;/div&gt;</summary>
		<author><name>Jheager2</name></author>
	</entry>
	<entry>
		<id>https://wiki.umiacs.umd.edu/umiacs/index.php?title=MBRC&amp;diff=8513</id>
		<title>MBRC</title>
		<link rel="alternate" type="text/html" href="https://wiki.umiacs.umd.edu/umiacs/index.php?title=MBRC&amp;diff=8513"/>
		<updated>2019-08-12T17:44:13Z</updated>

		<summary type="html">&lt;p&gt;Jheager2: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The MBRC ([https://mbrc.umd.edu MBRC]) at the University of Maryland is located within the Institute for Advanced Computer Studies.  The MBRC has a cluster of computational (CPU/GPU) resources that are available to be scheduled.&lt;br /&gt;
&lt;br /&gt;
=Compute Infrastructure=&lt;br /&gt;
&lt;br /&gt;
Each of UMIACS cluster computational infrastructures is accessed through the submission node.  Users will need to submit jobs through the [[SLURM]] resource manager once they have logged into the submission node.  Each cluster in UMIACS has different quality of service (QoS) that need to be selected upon submission of a job and many like this one has specific other resources such as GPUs that need to be requested for a job.  &lt;br /&gt;
&lt;br /&gt;
The current submission node(s) for &#039;&#039;MBRC&#039;&#039; is:&lt;br /&gt;
* &amp;lt;code&amp;gt;mbrcsub00.umiacs.umd.edu&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Partition ==&lt;br /&gt;
There are two partitions to the MBRC [[SLURM]] computational infrastructure.  If you do not specify a partition when submitting your job you will receive the &#039;&#039;&#039;dpart&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;dpart&#039;&#039;&#039; - This is the default partition and job allocations are guaranteed.&lt;br /&gt;
* &#039;&#039;&#039;scavenger&#039;&#039;&#039; - This is the alternate partition that allows jobs longer run times and more resources but is preemptable when jobs in the &#039;&#039;&#039;dpart&#039;&#039;&#039; are ready to be scheduled.&lt;br /&gt;
&lt;br /&gt;
== QoS ==&lt;br /&gt;
MBRC currently has 1 QoS for the &#039;&#039;&#039;dpart&#039;&#039;&#039; and 1 QoS for the &#039;&#039;&#039;scavenger&#039;&#039;&#039; partition.  The important parts here is that in different QoS you can have a shorter/longer maximum wall time, a total number of jobs running at once and a maximum number of track-able resources (TRES) for the job.  In the scavenger QoS there is one more constraint that you are restricted by the total number of TRES per user (over multiple jobs). &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
# show_qos&lt;br /&gt;
      Name     MaxWall MaxJobs                        MaxTRES     MaxTRESPU&lt;br /&gt;
---------- ----------- ------- ------------------------------ -------------&lt;br /&gt;
   default  1-00:00:00       1       cpu=4,mem=32G,gres/gpu=2&lt;br /&gt;
 scavenger  2-00:00:00                                          gres/gpu=8&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== GPUs ==&lt;br /&gt;
Jobs that require GPU resources need to explicitly request the resources within their job submission.  This is done through Generic Resource Scheduling (GRES).  Currently all nodes in the cluster are homogeneous however in the future this may not be the case.  Users may use the most generic identifier in this case &#039;&#039;&#039;gpu&#039;&#039;&#039; a colon and a number to select without explicitly naming the type of GPU (ie. &amp;lt;code&amp;gt;--gres gpu:4&amp;lt;/code&amp;gt; for 4 GPUs).  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ sinfo -o &amp;quot;%15n %10c %10m  %25f %25G&amp;quot;&lt;br /&gt;
NODELIST        CPUS       MEMORY      AVAIL_FEATURES            GRES&lt;br /&gt;
mbrc[00]      32         191896          (null)                    gpu:rtx2080ti:8&lt;br /&gt;
mbrc[01]      32         191896          (null)                    gpu:rtx2080ti:8&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Job Submission and Management ==&lt;br /&gt;
Users should review our [[SLURM]] [[SLURM/JobSubmission | job submission]] and [[SLURM/JobStatus | job management]] documentation.  &lt;br /&gt;
&lt;br /&gt;
A very quick start to get an interactive shell is as follows when run on the submission node.  This will allocate 1 GPU with 16GB of memory (system RAM) in the QoS default for 4 hours maximum time.  If the job goes beyond these limits either the memory allocation or the maximum time it will be terminated immediately. &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
srun --pty --gres=gpu:1 --mem=16G --qos=default --time=04:00:00 bash&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
[jheager2@mbrcsub00:~ ] $ srun --pty --gres=gpu:1 --mem=16G --qos=default --time=04:00:00 bash&lt;br /&gt;
[jheager2@mbrc00:~ ] $ nvidia-smi -L&lt;br /&gt;
GPU 0: GeForce RTX 2080 Ti (UUID: GPU-4ad5c018-b9bc-e664-233a-5d9ee8ad05cb)&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Data Storage=&lt;br /&gt;
&lt;br /&gt;
Until the final storage investment arrives we have made available a temporary allocation of storage.  This section is subject to change.  There are 3 types of storage available to users in the MBRC home directories, project directories and scratch directories.&lt;br /&gt;
&lt;br /&gt;
== Home Directories ==&lt;br /&gt;
 &lt;br /&gt;
Home directories in the MBRC computational infrastructure are available from the Institutes [[NFShomes]] as &amp;lt;code&amp;gt;/nfshomes/USERNAME&amp;lt;/code&amp;gt; where USERNAME is your username.  These home directories have very limited storage and are intended for your personal files, configuration and source code.  Your home directory is &#039;&#039;&#039;not&#039;&#039;&#039; intended for data sets or other large scale data holdings.  Users are encouraged to utilize our [[GitLab]] infrastructure to host your code repositories.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;NOTE&#039;&#039;&#039;: To check your quota on this directory you will need to use the &amp;lt;code&amp;gt;quota -s&amp;lt;/code&amp;gt; command.&lt;br /&gt;
&lt;br /&gt;
Your home directory data is fully protected and has both snapshots and is backed up nightly.&lt;br /&gt;
&lt;br /&gt;
== Project Directories ==&lt;br /&gt;
For this cluster we have decided to allocate network storage on a project by project basis. Jonathan Heagerty will be the point of contact as it pertains to allocating the requested/required storage for each project. As a whole, there is an allotment of network storage for the MRBC Cluster and is to why there will be limitations to the storage size and lifecycle for the data. &lt;br /&gt;
&lt;br /&gt;
If the requested storage size is significantly large relative to the total allotted amount, the request will be relayed from Jonathan Heagerty to the MBRC Cluster faculty for approval. Two other situations that would need approval from the MBRC Cluster faculty would be: To request an increase to a projects current storage allotment or To request a time extension for a project.&lt;br /&gt;
&lt;br /&gt;
When making a request for storage please provide the following information:&lt;br /&gt;
        - Name of user requesting storage:&lt;br /&gt;
                Example: jheager2&lt;br /&gt;
        - Name of project:&lt;br /&gt;
                Example: Foveated Rendering&lt;br /&gt;
        - Collaborators working on the project:&lt;br /&gt;
                Example: Sida Li&lt;br /&gt;
        - Storage size:&lt;br /&gt;
                Example: 1TB&lt;br /&gt;
        - Length of time for storage:&lt;br /&gt;
                Example: 6-8 months&lt;br /&gt;
&lt;br /&gt;
== Scratch Directories ==&lt;br /&gt;
&lt;br /&gt;
=== Local Scratch Directory===&lt;/div&gt;</summary>
		<author><name>Jheager2</name></author>
	</entry>
	<entry>
		<id>https://wiki.umiacs.umd.edu/umiacs/index.php?title=MBRC&amp;diff=8512</id>
		<title>MBRC</title>
		<link rel="alternate" type="text/html" href="https://wiki.umiacs.umd.edu/umiacs/index.php?title=MBRC&amp;diff=8512"/>
		<updated>2019-08-09T20:29:14Z</updated>

		<summary type="html">&lt;p&gt;Jheager2: /* QoS */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The MBRC ([https://mbrc.umd.edu MBRC]) at the University of Maryland is located within the Institute for Advanced Computer Studies.  The MBRC has a cluster of computational (CPU/GPU) resources that are available to be scheduled.&lt;br /&gt;
&lt;br /&gt;
=Compute Infrastructure=&lt;br /&gt;
&lt;br /&gt;
Each of UMIACS cluster computational infrastructures is accessed through the submission node.  Users will need to submit jobs through the [[SLURM]] resource manager once they have logged into the submission node.  Each cluster in UMIACS has different quality of service (QoS) that need to be selected upon submission of a job and many like this one has specific other resources such as GPUs that need to be requested for a job.  &lt;br /&gt;
&lt;br /&gt;
The current submission node(s) for &#039;&#039;MBRC&#039;&#039; is:&lt;br /&gt;
* &amp;lt;code&amp;gt;mbrcsub00.umiacs.umd.edu&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Partition ==&lt;br /&gt;
There are two partitions to the MBRC [[SLURM]] computational infrastructure.  If you do not specify a partition when submitting your job you will receive the &#039;&#039;&#039;dpart&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;dpart&#039;&#039;&#039; - This is the default partition and job allocations are guaranteed.&lt;br /&gt;
* &#039;&#039;&#039;scavenger&#039;&#039;&#039; - This is the alternate partition that allows jobs longer run times and more resources but is preemptable when jobs in the &#039;&#039;&#039;dpart&#039;&#039;&#039; are ready to be scheduled.&lt;br /&gt;
&lt;br /&gt;
== QoS ==&lt;br /&gt;
MBRC currently has 1 QoS for the &#039;&#039;&#039;dpart&#039;&#039;&#039; and 1 QoS for the &#039;&#039;&#039;scavenger&#039;&#039;&#039; partition.  The important parts here is that in different QoS you can have a shorter/longer maximum wall time, a total number of jobs running at once and a maximum number of track-able resources (TRES) for the job.  In the scavenger QoS there is one more constraint that you are restricted by the total number of TRES per user (over multiple jobs). &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
# show_qos&lt;br /&gt;
      Name     MaxWall MaxJobs                        MaxTRES     MaxTRESPU&lt;br /&gt;
---------- ----------- ------- ------------------------------ -------------&lt;br /&gt;
   default  1-00:00:00       1       cpu=4,mem=32G,gres/gpu=2&lt;br /&gt;
 scavenger  2-00:00:00                                          gres/gpu=8&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== GPUs ==&lt;br /&gt;
Jobs that require GPU resources need to explicitly request the resources within their job submission.  This is done through Generic Resource Scheduling (GRES).  Currently all nodes in the cluster are homogeneous however in the future this may not be the case.  Users may use the most generic identifier in this case &#039;&#039;&#039;gpu&#039;&#039;&#039; a colon and a number to select without explicitly naming the type of GPU (ie. &amp;lt;code&amp;gt;--gres gpu:4&amp;lt;/code&amp;gt; for 4 GPUs).  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ sinfo -o &amp;quot;%15n %10c %10m  %25f %25G&amp;quot;&lt;br /&gt;
NODELIST        CPUS       MEMORY      AVAIL_FEATURES            GRES&lt;br /&gt;
mbrc[00]      32         191896          (null)                    gpu:rtx2080ti:8&lt;br /&gt;
mbrc[01]      32         191896          (null)                    gpu:rtx2080ti:8&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Job Submission and Management ==&lt;br /&gt;
Users should review our [[SLURM]] [[SLURM/JobSubmission | job submission]] and [[SLURM/JobStatus | job management]] documentation.  &lt;br /&gt;
&lt;br /&gt;
A very quick start to get an interactive shell is as follows when run on the submission node.  This will allocate 1 GPU with 16GB of memory (system RAM) in the QoS default for 4 hours maximum time.  If the job goes beyond these limits either the memory allocation or the maximum time it will be terminated immediately. &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
srun --pty --gres=gpu:1 --mem=16G --qos=default --time=04:00:00 bash&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
[jheager2@mbrcsub00:~ ] $ srun --pty --gres=gpu:1 --mem=16G --qos=default --time=04:00:00 bash&lt;br /&gt;
[jheager2@mbrc00:~ ] $ nvidia-smi -L&lt;br /&gt;
GPU 0: GeForce RTX 2080 Ti (UUID: GPU-4ad5c018-b9bc-e664-233a-5d9ee8ad05cb)&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Data Storage=&lt;br /&gt;
&lt;br /&gt;
Until the final storage investment arrives we have made available a temporary allocation of storage.  This section is subject to change.  There are 3 types of storage available to users in the MBRC home directories, project directories and scratch directories.&lt;br /&gt;
&lt;br /&gt;
== Home Directories ==&lt;br /&gt;
 &lt;br /&gt;
Home directories in the MBRC computational infrastructure are available from the Institutes [[NFShomes]] as &amp;lt;code&amp;gt;/nfshomes/USERNAME&amp;lt;/code&amp;gt; where USERNAME is your username.  These home directories have very limited storage and are intended for your personal files, configuration and source code.  Your home directory is &#039;&#039;&#039;not&#039;&#039;&#039; intended for data sets or other large scale data holdings.  Users are encouraged to utilize our [[GitLab]] infrastructure to host your code repositories.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;NOTE&#039;&#039;&#039;: To check your quota on this directory you will need to use the &amp;lt;code&amp;gt;quota -s&amp;lt;/code&amp;gt; command.&lt;br /&gt;
&lt;br /&gt;
Your home directory data is fully protected and has both snapshots and is backed up nightly.&lt;br /&gt;
&lt;br /&gt;
== Project Directories ==&lt;br /&gt;
&lt;br /&gt;
== Scratch Directories ==&lt;br /&gt;
&lt;br /&gt;
=== Local Scratch Directory===&lt;/div&gt;</summary>
		<author><name>Jheager2</name></author>
	</entry>
	<entry>
		<id>https://wiki.umiacs.umd.edu/umiacs/index.php?title=MBRC&amp;diff=8511</id>
		<title>MBRC</title>
		<link rel="alternate" type="text/html" href="https://wiki.umiacs.umd.edu/umiacs/index.php?title=MBRC&amp;diff=8511"/>
		<updated>2019-08-09T18:36:42Z</updated>

		<summary type="html">&lt;p&gt;Jheager2: /* Network Scratch Directory */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The MBRC ([https://mbrc.umd.edu MBRC]) at the University of Maryland is located within the Institute for Advanced Computer Studies.  The MBRC has a cluster of computational (CPU/GPU) resources that are available to be scheduled.&lt;br /&gt;
&lt;br /&gt;
=Compute Infrastructure=&lt;br /&gt;
&lt;br /&gt;
Each of UMIACS cluster computational infrastructures is accessed through the submission node.  Users will need to submit jobs through the [[SLURM]] resource manager once they have logged into the submission node.  Each cluster in UMIACS has different quality of service (QoS) that need to be selected upon submission of a job and many like this one has specific other resources such as GPUs that need to be requested for a job.  &lt;br /&gt;
&lt;br /&gt;
The current submission node(s) for &#039;&#039;MBRC&#039;&#039; is:&lt;br /&gt;
* &amp;lt;code&amp;gt;mbrcsub00.umiacs.umd.edu&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Partition ==&lt;br /&gt;
There are two partitions to the MBRC [[SLURM]] computational infrastructure.  If you do not specify a partition when submitting your job you will receive the &#039;&#039;&#039;dpart&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;dpart&#039;&#039;&#039; - This is the default partition and job allocations are guaranteed.&lt;br /&gt;
* &#039;&#039;&#039;scavenger&#039;&#039;&#039; - This is the alternate partition that allows jobs longer run times and more resources but is preemptable when jobs in the &#039;&#039;&#039;dpart&#039;&#039;&#039; are ready to be scheduled.&lt;br /&gt;
&lt;br /&gt;
== QoS ==&lt;br /&gt;
MBRC currently has 1 QoS for the &#039;&#039;&#039;dpart&#039;&#039;&#039; and 1 QoS for the &#039;&#039;&#039;scavenger&#039;&#039;&#039; partition.  The important parts here is that in different QoS you can have a shorter/longer maximum wall time, a total number of jobs running at once and a maximum number of track-able resources (TRES) for the job.  In the scavenger QoS there is one more constraint that you are restricted by the total number of TRES per user (over multiple jobs). &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
# show_qos&lt;br /&gt;
      Name     MaxWall MaxJobs                        MaxTRES     MaxTRESPU&lt;br /&gt;
---------- ----------- ------- ------------------------------ -------------&lt;br /&gt;
   default  1-00:00:00       1       cpu=4,mem=32G,gres/gpu=1&lt;br /&gt;
 scavenger  2-00:00:00                                          gres/gpu=8&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== GPUs ==&lt;br /&gt;
Jobs that require GPU resources need to explicitly request the resources within their job submission.  This is done through Generic Resource Scheduling (GRES).  Currently all nodes in the cluster are homogeneous however in the future this may not be the case.  Users may use the most generic identifier in this case &#039;&#039;&#039;gpu&#039;&#039;&#039; a colon and a number to select without explicitly naming the type of GPU (ie. &amp;lt;code&amp;gt;--gres gpu:4&amp;lt;/code&amp;gt; for 4 GPUs).  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ sinfo -o &amp;quot;%15n %10c %10m  %25f %25G&amp;quot;&lt;br /&gt;
NODELIST        CPUS       MEMORY      AVAIL_FEATURES            GRES&lt;br /&gt;
mbrc[00]      32         191896          (null)                    gpu:rtx2080ti:8&lt;br /&gt;
mbrc[01]      32         191896          (null)                    gpu:rtx2080ti:8&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Job Submission and Management ==&lt;br /&gt;
Users should review our [[SLURM]] [[SLURM/JobSubmission | job submission]] and [[SLURM/JobStatus | job management]] documentation.  &lt;br /&gt;
&lt;br /&gt;
A very quick start to get an interactive shell is as follows when run on the submission node.  This will allocate 1 GPU with 16GB of memory (system RAM) in the QoS default for 4 hours maximum time.  If the job goes beyond these limits either the memory allocation or the maximum time it will be terminated immediately. &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
srun --pty --gres=gpu:1 --mem=16G --qos=default --time=04:00:00 bash&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
[jheager2@mbrcsub00:~ ] $ srun --pty --gres=gpu:1 --mem=16G --qos=default --time=04:00:00 bash&lt;br /&gt;
[jheager2@mbrc00:~ ] $ nvidia-smi -L&lt;br /&gt;
GPU 0: GeForce RTX 2080 Ti (UUID: GPU-4ad5c018-b9bc-e664-233a-5d9ee8ad05cb)&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Data Storage=&lt;br /&gt;
&lt;br /&gt;
Until the final storage investment arrives we have made available a temporary allocation of storage.  This section is subject to change.  There are 3 types of storage available to users in the MBRC home directories, project directories and scratch directories.&lt;br /&gt;
&lt;br /&gt;
== Home Directories ==&lt;br /&gt;
 &lt;br /&gt;
Home directories in the MBRC computational infrastructure are available from the Institutes [[NFShomes]] as &amp;lt;code&amp;gt;/nfshomes/USERNAME&amp;lt;/code&amp;gt; where USERNAME is your username.  These home directories have very limited storage and are intended for your personal files, configuration and source code.  Your home directory is &#039;&#039;&#039;not&#039;&#039;&#039; intended for data sets or other large scale data holdings.  Users are encouraged to utilize our [[GitLab]] infrastructure to host your code repositories.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;NOTE&#039;&#039;&#039;: To check your quota on this directory you will need to use the &amp;lt;code&amp;gt;quota -s&amp;lt;/code&amp;gt; command.&lt;br /&gt;
&lt;br /&gt;
Your home directory data is fully protected and has both snapshots and is backed up nightly.&lt;br /&gt;
&lt;br /&gt;
== Project Directories ==&lt;br /&gt;
&lt;br /&gt;
== Scratch Directories ==&lt;br /&gt;
&lt;br /&gt;
=== Local Scratch Directory===&lt;/div&gt;</summary>
		<author><name>Jheager2</name></author>
	</entry>
	<entry>
		<id>https://wiki.umiacs.umd.edu/umiacs/index.php?title=MBRC&amp;diff=8494</id>
		<title>MBRC</title>
		<link rel="alternate" type="text/html" href="https://wiki.umiacs.umd.edu/umiacs/index.php?title=MBRC&amp;diff=8494"/>
		<updated>2019-08-08T20:02:38Z</updated>

		<summary type="html">&lt;p&gt;Jheager2: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The MBRC ([https://mbrc.umd.edu MBRC]) at the University of Maryland is located within the Institute for Advanced Computer Studies.  The MBRC has a cluster of computational (CPU/GPU) resources that are available to be scheduled.&lt;br /&gt;
&lt;br /&gt;
=Compute Infrastructure=&lt;br /&gt;
&lt;br /&gt;
Each of UMIACS cluster computational infrastructures is accessed through the submission node.  Users will need to submit jobs through the [[SLURM]] resource manager once they have logged into the submission node.  Each cluster in UMIACS has different quality of service (QoS) that need to be selected upon submission of a job and many like this one has specific other resources such as GPUs that need to be requested for a job.  &lt;br /&gt;
&lt;br /&gt;
The current submission node(s) for &#039;&#039;MBRC&#039;&#039; is:&lt;br /&gt;
* &amp;lt;code&amp;gt;mbrcsub00.umiacs.umd.edu&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Partition ==&lt;br /&gt;
There are two partitions to the MBRC [[SLURM]] computational infrastructure.  If you do not specify a partition when submitting your job you will receive the &#039;&#039;&#039;dpart&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;dpart&#039;&#039;&#039; - This is the default partition and job allocations are guaranteed.&lt;br /&gt;
* &#039;&#039;&#039;scavenger&#039;&#039;&#039; - This is the alternate partition that allows jobs longer run times and more resources but is preemptable when jobs in the &#039;&#039;&#039;dpart&#039;&#039;&#039; are ready to be scheduled.&lt;br /&gt;
&lt;br /&gt;
== QoS ==&lt;br /&gt;
MBRC currently has 1 QoS for the &#039;&#039;&#039;dpart&#039;&#039;&#039; and 1 QoS for the &#039;&#039;&#039;scavenger&#039;&#039;&#039; partition.  The important parts here is that in different QoS you can have a shorter/longer maximum wall time, a total number of jobs running at once and a maximum number of track-able resources (TRES) for the job.  In the scavenger QoS there is one more constraint that you are restricted by the total number of TRES per user (over multiple jobs). &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
# show_qos&lt;br /&gt;
      Name     MaxWall MaxJobs                        MaxTRES     MaxTRESPU&lt;br /&gt;
---------- ----------- ------- ------------------------------ -------------&lt;br /&gt;
   default  1-00:00:00       1       cpu=4,mem=32G,gres/gpu=1&lt;br /&gt;
 scavenger  2-00:00:00                                          gres/gpu=8&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== GPUs ==&lt;br /&gt;
Jobs that require GPU resources need to explicitly request the resources within their job submission.  This is done through Generic Resource Scheduling (GRES).  Currently all nodes in the cluster are homogeneous however in the future this may not be the case.  Users may use the most generic identifier in this case &#039;&#039;&#039;gpu&#039;&#039;&#039; a colon and a number to select without explicitly naming the type of GPU (ie. &amp;lt;code&amp;gt;--gres gpu:4&amp;lt;/code&amp;gt; for 4 GPUs).  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ sinfo -o &amp;quot;%15n %10c %10m  %25f %25G&amp;quot;&lt;br /&gt;
NODELIST        CPUS       MEMORY      AVAIL_FEATURES            GRES&lt;br /&gt;
mbrc[00]      32         191896          (null)                    gpu:rtx2080ti:8&lt;br /&gt;
mbrc[01]      32         191896          (null)                    gpu:rtx2080ti:8&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Job Submission and Management ==&lt;br /&gt;
Users should review our [[SLURM]] [[SLURM/JobSubmission | job submission]] and [[SLURM/JobStatus | job management]] documentation.  &lt;br /&gt;
&lt;br /&gt;
A very quick start to get an interactive shell is as follows when run on the submission node.  This will allocate 1 GPU with 16GB of memory (system RAM) in the QoS default for 4 hours maximum time.  If the job goes beyond these limits either the memory allocation or the maximum time it will be terminated immediately. &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
srun --pty --gres=gpu:1 --mem=16G --qos=default --time=04:00:00 bash&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
[jheager2@mbrcsub00:~ ] $ srun --pty --gres=gpu:1 --mem=16G --qos=default --time=04:00:00 bash&lt;br /&gt;
[jheager2@mbrc00:~ ] $ nvidia-smi -L&lt;br /&gt;
GPU 0: GeForce RTX 2080 Ti (UUID: GPU-4ad5c018-b9bc-e664-233a-5d9ee8ad05cb)&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Data Storage=&lt;br /&gt;
&lt;br /&gt;
Until the final storage investment arrives we have made available a temporary allocation of storage.  This section is subject to change.  There are 3 types of storage available to users in the MBRC home directories, project directories and scratch directories.&lt;br /&gt;
&lt;br /&gt;
== Home Directories ==&lt;br /&gt;
 &lt;br /&gt;
Home directories in the MBRC computational infrastructure are available from the Institutes [[NFShomes]] as &amp;lt;code&amp;gt;/nfshomes/USERNAME&amp;lt;/code&amp;gt; where USERNAME is your username.  These home directories have very limited storage and are intended for your personal files, configuration and source code.  Your home directory is &#039;&#039;&#039;not&#039;&#039;&#039; intended for data sets or other large scale data holdings.  Users are encouraged to utilize our [[GitLab]] infrastructure to host your code repositories.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;NOTE&#039;&#039;&#039;: To check your quota on this directory you will need to use the &amp;lt;code&amp;gt;quota -s&amp;lt;/code&amp;gt; command.&lt;br /&gt;
&lt;br /&gt;
Your home directory data is fully protected and has both snapshots and is backed up nightly.&lt;br /&gt;
&lt;br /&gt;
== Project Directories ==&lt;br /&gt;
&lt;br /&gt;
== Scratch Directories ==&lt;br /&gt;
&lt;br /&gt;
=== Network Scratch Directory===&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Local Scratch Directory===&lt;/div&gt;</summary>
		<author><name>Jheager2</name></author>
	</entry>
	<entry>
		<id>https://wiki.umiacs.umd.edu/umiacs/index.php?title=MBRC&amp;diff=8493</id>
		<title>MBRC</title>
		<link rel="alternate" type="text/html" href="https://wiki.umiacs.umd.edu/umiacs/index.php?title=MBRC&amp;diff=8493"/>
		<updated>2019-08-08T20:00:16Z</updated>

		<summary type="html">&lt;p&gt;Jheager2: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The MBRC ([https://mbrc.umd.edu MBRC]) at the University of Maryland is located within the Institute for Advanced Computer Studies.  The MBRC has a cluster of computational (CPU/GPU) resources that are available to be scheduled.&lt;br /&gt;
&lt;br /&gt;
=Compute Infrastructure=&lt;br /&gt;
&lt;br /&gt;
Each of UMIACS cluster computational infrastructures is accessed through the submission node.  Users will need to submit jobs through the [[SLURM]] resource manager once they have logged into the submission node.  Each cluster in UMIACS has different quality of service (QoS) that need to be selected upon submission of a job and many like this one has specific other resources such as GPUs that need to be requested for a job.  &lt;br /&gt;
&lt;br /&gt;
The current submission node(s) for &#039;&#039;MBRC&#039;&#039; is:&lt;br /&gt;
* &amp;lt;code&amp;gt;mbrcsub00.umiacs.umd.edu&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Partition ==&lt;br /&gt;
There are two partitions to the MBRC [[SLURM]] computational infrastructure.  If you do not specify a partition when submitting your job you will receive the &#039;&#039;&#039;dpart&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;dpart&#039;&#039;&#039; - This is the default partition and job allocations are guaranteed.&lt;br /&gt;
* &#039;&#039;&#039;scavenger&#039;&#039;&#039; - This is the alternate partition that allows jobs longer run times and more resources but is preemptable when jobs in the &#039;&#039;&#039;dpart&#039;&#039;&#039; are ready to be scheduled.&lt;br /&gt;
&lt;br /&gt;
== QoS ==&lt;br /&gt;
MBRC currently has 1 QoS for the &#039;&#039;&#039;dpart&#039;&#039;&#039; and 1 QoS for the &#039;&#039;&#039;scavenger&#039;&#039;&#039; partition.  The important parts here is that in different QoS you can have a shorter/longer maximum wall time, a total number of jobs running at once and a maximum number of track-able resources (TRES) for the job.  In the scavenger QoS there is one more constraint that you are restricted by the total number of TRES per user (over multiple jobs). &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
# show_qos&lt;br /&gt;
      Name     MaxWall MaxJobs                        MaxTRES     MaxTRESPU&lt;br /&gt;
---------- ----------- ------- ------------------------------ -------------&lt;br /&gt;
   default  1-00:00:00       1       cpu=4,mem=32G,gres/gpu=1&lt;br /&gt;
 scavenger  2-00:00:00                                          gres/gpu=8&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== GPUs ==&lt;br /&gt;
Jobs that require GPU resources need to explicitly request the resources within their job submission.  This is done through Generic Resource Scheduling (GRES).  Currently all nodes in the cluster are homogeneous however in the future this may not be the case.  Users may use the most generic identifier in this case &#039;&#039;&#039;gpu&#039;&#039;&#039; a colon and a number to select without explicitly naming the type of GPU (ie. &amp;lt;code&amp;gt;--gres gpu:4&amp;lt;/code&amp;gt; for 4 GPUs).  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ sinfo -o &amp;quot;%15N %10c %10m  %25f %25G&amp;quot;&lt;br /&gt;
NODELIST        CPUS       MEMORY      AVAIL_FEATURES            GRES&lt;br /&gt;
mbrc[00-01]      32         191896          (null)                    gpu:rtx2080ti:8&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
*Each node has the above resource specifications&lt;br /&gt;
&lt;br /&gt;
== Job Submission and Management ==&lt;br /&gt;
Users should review our [[SLURM]] [[SLURM/JobSubmission | job submission]] and [[SLURM/JobStatus | job management]] documentation.  &lt;br /&gt;
&lt;br /&gt;
A very quick start to get an interactive shell is as follows when run on the submission node.  This will allocate 1 GPU with 16GB of memory (system RAM) in the QoS default for 4 hours maximum time.  If the job goes beyond these limits either the memory allocation or the maximum time it will be terminated immediately. &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
srun --pty --gres=gpu:1 --mem=16G --qos=default --time=04:00:00 bash&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
[jheager2@mbrcsub00:~ ] $ srun --pty --gres=gpu:1 --mem=16G --qos=default --time=04:00:00 bash&lt;br /&gt;
[jheager2@mbrc00:~ ] $ nvidia-smi -L&lt;br /&gt;
GPU 0: GeForce RTX 2080 Ti (UUID: GPU-4ad5c018-b9bc-e664-233a-5d9ee8ad05cb)&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Data Storage=&lt;br /&gt;
&lt;br /&gt;
Until the final storage investment arrives we have made available a temporary allocation of storage.  This section is subject to change.  There are 3 types of storage available to users in the MBRC home directories, project directories and scratch directories.&lt;br /&gt;
&lt;br /&gt;
== Home Directories ==&lt;br /&gt;
 &lt;br /&gt;
Home directories in the MBRC computational infrastructure are available from the Institutes [[NFShomes]] as &amp;lt;code&amp;gt;/nfshomes/USERNAME&amp;lt;/code&amp;gt; where USERNAME is your username.  These home directories have very limited storage and are intended for your personal files, configuration and source code.  Your home directory is &#039;&#039;&#039;not&#039;&#039;&#039; intended for data sets or other large scale data holdings.  Users are encouraged to utilize our [[GitLab]] infrastructure to host your code repositories.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;NOTE&#039;&#039;&#039;: To check your quota on this directory you will need to use the &amp;lt;code&amp;gt;quota -s&amp;lt;/code&amp;gt; command.&lt;br /&gt;
&lt;br /&gt;
Your home directory data is fully protected and has both snapshots and is backed up nightly.&lt;br /&gt;
&lt;br /&gt;
== Project Directories ==&lt;br /&gt;
&lt;br /&gt;
== Scratch Directories ==&lt;br /&gt;
&lt;br /&gt;
=== Network Scratch Directory===&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Local Scratch Directory===&lt;/div&gt;</summary>
		<author><name>Jheager2</name></author>
	</entry>
	<entry>
		<id>https://wiki.umiacs.umd.edu/umiacs/index.php?title=MBRC&amp;diff=8492</id>
		<title>MBRC</title>
		<link rel="alternate" type="text/html" href="https://wiki.umiacs.umd.edu/umiacs/index.php?title=MBRC&amp;diff=8492"/>
		<updated>2019-08-08T19:50:19Z</updated>

		<summary type="html">&lt;p&gt;Jheager2: MBRC&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;The MBRC ([https://mbrc.umd.edu MBRC]) at the University of Maryland is located within the Institute for Advanced Computer Studies.  The MBRC has a cluster of computational (CPU/GPU) resources that are available to be scheduled.&lt;br /&gt;
&lt;br /&gt;
=Compute Infrastructure=&lt;br /&gt;
&lt;br /&gt;
Each of UMIACS cluster computational infrastructures is accessed through the submission node.  Users will need to submit jobs through the [[SLURM]] resource manager once they have logged into the submission node.  Each cluster in UMIACS has different quality of service (QoS) that need to be selected upon submission of a job and many like this one has specific other resources such as GPUs that need to be requested for a job.  &lt;br /&gt;
&lt;br /&gt;
The current submission node(s) for &#039;&#039;MBRC&#039;&#039; is:&lt;br /&gt;
* &amp;lt;code&amp;gt;mbrcsub00.umiacs.umd.edu&amp;lt;/code&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Partition ==&lt;br /&gt;
There are two partitions to the MBRC [[SLURM]] computational infrastructure.  If you do not specify a partition when submitting your job you will receive the &#039;&#039;&#039;dpart&#039;&#039;&#039;. &lt;br /&gt;
&lt;br /&gt;
* &#039;&#039;&#039;dpart&#039;&#039;&#039; - This is the default partition and job allocations are guaranteed.&lt;br /&gt;
* &#039;&#039;&#039;scavenger&#039;&#039;&#039; - This is the alternate partition that allows jobs longer run times and more resources but is preemptable when jobs in the &#039;&#039;&#039;dpart&#039;&#039;&#039; are ready to be scheduled.&lt;br /&gt;
&lt;br /&gt;
== QoS ==&lt;br /&gt;
MBRC currently has 1 QoS for the &#039;&#039;&#039;dpart&#039;&#039;&#039; and 1 QoS for the &#039;&#039;&#039;scavenger&#039;&#039;&#039; partition.  The important parts here is that in different QoS you can have a shorter/longer maximum wall time, a total number of jobs running at once and a maximum number of track-able resources (TRES) for the job.  In the scavenger QoS there is one more constraint that you are restricted by the total number of TRES per user (over multiple jobs). &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
# show_qos&lt;br /&gt;
      Name     MaxWall MaxJobs                        MaxTRES     MaxTRESPU&lt;br /&gt;
---------- ----------- ------- ------------------------------ -------------&lt;br /&gt;
   default  1-00:00:00       1       cpu=4,mem=32G,gres/gpu=1&lt;br /&gt;
 scavenger  2-00:00:00                                          gres/gpu=8&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== GPUs ==&lt;br /&gt;
Jobs that require GPU resources need to explicitly request the resources within their job submission.  This is done through Generic Resource Scheduling (GRES).  Currently all nodes in the cluster are homogeneous however in the future this may not be the case.  Users may use the most generic identifier in this case &#039;&#039;&#039;gpu&#039;&#039;&#039; a colon and a number to select without explicitly naming the type of GPU (ie. &amp;lt;code&amp;gt;--gres gpu:4&amp;lt;/code&amp;gt; for 4 GPUs).  &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ sinfo -o &amp;quot;%15N %10c %10m  %25f %25G&amp;quot;&lt;br /&gt;
NODELIST        CPUS       MEMORY      AVAIL_FEATURES            GRES&lt;br /&gt;
mbrc[00-01]      32         191896          (null)                    gpu:rtx2080ti:8&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Job Submission and Management ==&lt;br /&gt;
Users should review our [[SLURM]] [[SLURM/JobSubmission | job submission]] and [[SLURM/JobStatus | job management]] documentation.  &lt;br /&gt;
&lt;br /&gt;
A very quick start to get an interactive shell is as follows when run on the submission node.  This will allocate 1 GPU with 16GB of memory (system RAM) in the QoS default for 4 hours maximum time.  If the job goes beyond these limits either the memory allocation or the maximum time it will be terminated immediately. &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
srun --pty --gres=gpu:1 --mem=16G --qos=default --time=04:00:00 bash&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
[jheager2@mbrcsub00:~ ] $ srun --pty --gres=gpu:1 --mem=16G --qos=default --time=04:00:00 bash&lt;br /&gt;
[jheager2@mbrc00:~ ] $ nvidia-smi -L&lt;br /&gt;
GPU 0: GeForce RTX 2080 Ti (UUID: GPU-4ad5c018-b9bc-e664-233a-5d9ee8ad05cb)&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=Data Storage=&lt;br /&gt;
&lt;br /&gt;
Until the final storage investment arrives we have made available a temporary allocation of storage.  This section is subject to change.  There are 3 types of storage available to users in the MBRC home directories, project directories and scratch directories.&lt;br /&gt;
&lt;br /&gt;
== Home Directories ==&lt;br /&gt;
 &lt;br /&gt;
Home directories in the MBRC computational infrastructure are available from the Institutes [[NFShomes]] as &amp;lt;code&amp;gt;/nfshomes/USERNAME&amp;lt;/code&amp;gt; where USERNAME is your username.  These home directories have very limited storage and are intended for your personal files, configuration and source code.  Your home directory is &#039;&#039;&#039;not&#039;&#039;&#039; intended for data sets or other large scale data holdings.  Users are encouraged to utilize our [[GitLab]] infrastructure to host your code repositories.&lt;br /&gt;
&lt;br /&gt;
&#039;&#039;&#039;NOTE&#039;&#039;&#039;: To check your quota on this directory you will need to use the &amp;lt;code&amp;gt;quota -s&amp;lt;/code&amp;gt; command.&lt;br /&gt;
&lt;br /&gt;
Your home directory data is fully protected and has both snapshots and is backed up nightly.&lt;br /&gt;
&lt;br /&gt;
== Project Directories ==&lt;br /&gt;
&lt;br /&gt;
== Scratch Directories ==&lt;br /&gt;
&lt;br /&gt;
=== Network Scratch Directory===&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Local Scratch Directory===&lt;/div&gt;</summary>
		<author><name>Jheager2</name></author>
	</entry>
</feed>