<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
	<id>https://wiki.umiacs.umd.edu/umiacs/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Ackumar</id>
	<title>UMIACS - User contributions [en]</title>
	<link rel="self" type="application/atom+xml" href="https://wiki.umiacs.umd.edu/umiacs/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Ackumar"/>
	<link rel="alternate" type="text/html" href="https://wiki.umiacs.umd.edu/umiacs/index.php/Special:Contributions/Ackumar"/>
	<updated>2026-05-12T15:26:51Z</updated>
	<subtitle>User contributions</subtitle>
	<generator>MediaWiki 1.43.7</generator>
	<entry>
		<id>https://wiki.umiacs.umd.edu/umiacs/index.php?title=SLURM/ClusterStatus&amp;diff=12994</id>
		<title>SLURM/ClusterStatus</title>
		<link rel="alternate" type="text/html" href="https://wiki.umiacs.umd.edu/umiacs/index.php?title=SLURM/ClusterStatus&amp;diff=12994"/>
		<updated>2026-02-06T17:57:27Z</updated>

		<summary type="html">&lt;p&gt;Ackumar: added slurm-web as a way to view cluster status&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;=Cluster Status=&lt;br /&gt;
SLURM offers a variety of tools to check the general status of nodes/partitions in a cluster.&lt;br /&gt;
&lt;br /&gt;
==sinfo==&lt;br /&gt;
The sinfo command will show you the status of partitions in the cluster, in alphabetical order. Passing the -N flag will show each node individually, also in alphabetical order.&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ sinfo&lt;br /&gt;
PARTITION         AVAIL  TIMELIMIT  NODES  STATE NODELIST&lt;br /&gt;
cbcb                 up   infinite      9    mix cbcb[00,03,22,26-28],legacy[00,10,20]&lt;br /&gt;
cbcb                 up   infinite      3  alloc cbcb[01-02,29]&lt;br /&gt;
cbcb                 up   infinite     49   idle cbcb[04-20,23-25],legacy[01-09,11,13-19,21-28,30-31,34-35]&lt;br /&gt;
...&lt;br /&gt;
vulcan-scavenger     up   infinite     14    mix brigid[16-17],vulcan[23-25,27-30,32-33,36-37,45]&lt;br /&gt;
vulcan-scavenger     up   infinite      1  alloc vulcan35&lt;br /&gt;
vulcan-scavenger     up   infinite     34   idle brigid[18-19],vulcan[00-22,26,34,38-44]&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ sinfo -N&lt;br /&gt;
NODELIST     NODES         PARTITION STATE&lt;br /&gt;
brigid16         1        vulcan-cpu mix&lt;br /&gt;
brigid16         1         scavenger mix&lt;br /&gt;
brigid16         1      vulcan-dpart mix&lt;br /&gt;
brigid16         1  vulcan-scavenger mix&lt;br /&gt;
...&lt;br /&gt;
vulcan45         1     vulcan-ramani mix&lt;br /&gt;
vulcan45         1  vulcan-scavenger mix&lt;br /&gt;
vulcan45         1         scavenger mix&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==scontrol==&lt;br /&gt;
The scontrol command can be used to view the status/configuration of the nodes in the cluster. If passed specific node name(s) only information about those node(s) will be displayed, otherwise all nodes will be listed. To specify multiple nodes, separate each node name by a comma (no spaces).&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ scontrol show nodes tron05,tron13&lt;br /&gt;
NodeName=tron05 Arch=x86_64 CoresPerSocket=16&lt;br /&gt;
   CPUAlloc=28 CPUTot=32 CPULoad=47.32&lt;br /&gt;
   AvailableFeatures=rhel8,x86_64,Zen,EPYC-7302,Ampere&lt;br /&gt;
   ActiveFeatures=rhel8,x86_64,Zen,EPYC-7302,Ampere&lt;br /&gt;
   Gres=gpu:rtxa6000:8&lt;br /&gt;
   NodeAddr=tron05 NodeHostName=tron05 Version=21.08.5&lt;br /&gt;
   OS=Linux 4.18.0-348.20.1.el8_5.x86_64 #1 SMP Tue Mar 8 12:56:54 EST 2022&lt;br /&gt;
   RealMemory=257538 AllocMem=157696 FreeMem=197620 Sockets=2 Boards=1&lt;br /&gt;
   State=MIXED ThreadsPerCore=1 TmpDisk=0 Weight=100 Owner=N/A MCS_label=N/A&lt;br /&gt;
   Partitions=scavenger,tron&lt;br /&gt;
   BootTime=2022-04-21T17:40:51 SlurmdStartTime=2022-04-21T18:00:56&lt;br /&gt;
   LastBusyTime=2022-04-22T11:21:16&lt;br /&gt;
   CfgTRES=cpu=32,mem=257538M,billing=346,gres/gpu=8,gres/gpu:rtxa6000=8&lt;br /&gt;
   AllocTRES=cpu=28,mem=154G,gres/gpu=7,gres/gpu:rtxa6000=7&lt;br /&gt;
   CapWatts=n/a&lt;br /&gt;
   CurrentWatts=0 AveWatts=0&lt;br /&gt;
   ExtSensorsJoules=n/s ExtSensorsWatts=0 ExtSensorsTemp=n/s&lt;br /&gt;
&lt;br /&gt;
NodeName=tron13 Arch=x86_64 CoresPerSocket=16&lt;br /&gt;
   CPUAlloc=1 CPUTot=16 CPULoad=8.41&lt;br /&gt;
   AvailableFeatures=rhel8,x86_64,Zen,EPYC-7302P,Ampere&lt;br /&gt;
   ActiveFeatures=rhel8,x86_64,Zen,EPYC-7302P,Ampere&lt;br /&gt;
   Gres=gpu:rtxa4000:4&lt;br /&gt;
   NodeAddr=tron13 NodeHostName=tron13 Version=21.08.5&lt;br /&gt;
   OS=Linux 4.18.0-348.20.1.el8_5.x86_64 #1 SMP Tue Mar 8 12:56:54 EST 2022&lt;br /&gt;
   RealMemory=128525 AllocMem=65536 FreeMem=33463 Sockets=1 Boards=1&lt;br /&gt;
   State=MIXED ThreadsPerCore=1 TmpDisk=0 Weight=10 Owner=N/A MCS_label=N/A&lt;br /&gt;
   Partitions=scavenger,tron&lt;br /&gt;
   BootTime=2022-04-21T17:40:46 SlurmdStartTime=2022-04-21T17:54:51&lt;br /&gt;
   LastBusyTime=2022-04-22T13:04:57&lt;br /&gt;
   CfgTRES=cpu=16,mem=128525M,billing=173,gres/gpu=4,gres/gpu:rtxa4000=4&lt;br /&gt;
   AllocTRES=cpu=1,mem=64G,gres/gpu=4,gres/gpu:rtxa4000=4&lt;br /&gt;
   CapWatts=n/a&lt;br /&gt;
   CurrentWatts=0 AveWatts=0&lt;br /&gt;
   ExtSensorsJoules=n/s ExtSensorsWatts=0 ExtSensorsTemp=n/s&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==sacctmgr==&lt;br /&gt;
The sacctmgr command shows cluster accounting information.  One of the helpful commands is to list the available QoSes. &lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
$ sacctmgr list qos format=Name,Priority,MaxWall,MaxJobsPU&lt;br /&gt;
      Name   Priority     MaxWall MaxJobsPU&lt;br /&gt;
---------- ---------- ----------- ---------&lt;br /&gt;
    normal          0&lt;br /&gt;
     dpart          0  2-00:00:00         8&lt;br /&gt;
       gpu          0    08:00:00         2&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==slurm-web==&lt;br /&gt;
Slurm-web is a web dashboard for our Slurm HPC cluster. It provides cluster statistics and Slurm information for the Nexus cluster.&lt;br /&gt;
* URL: https://slurm-web.umiacs.umd.edu&lt;/div&gt;</summary>
		<author><name>Ackumar</name></author>
	</entry>
</feed>