ARROW Cluster: Difference between revisions

From TRACC Wiki
Jump to navigation Jump to search
Line 3: Line 3:
<p>ARROW is arranged such that there is a single set of login nodes, a singe file system, and single user home directory that serves all of the nodes in all of the queues.
<p>ARROW is arranged such that there is a single set of login nodes, a singe file system, and single user home directory that serves all of the nodes in all of the queues.
== ARROW Queues==
== ARROW Queues==
There are currently three queues that are available with some restrictions about who can use them as described below.
There are currently five queues that are available with some restrictions about who can use them as described below.
* batch ('''default queue''', with 94 nodes, each node with 16 floating point cores available for general use)
* batch ('''default queue''', with 94 nodes, each node with 16 floating point cores available for general use)
** 92 nodes have 32 GB of RAM
** 92 nodes have 32 GB of RAM
Line 10: Line 10:
* nhtsa (with 12 nodes, each with 28 cores and 64 GB of RAM, only available to the NHTSA project)
* nhtsa (with 12 nodes, each with 28 cores and 64 GB of RAM, only available to the NHTSA project)
* arrow (one new EPYC server with 64 cores, for use for testing by TRACC staff or special permission by the TRACC Director)
* arrow (one new EPYC server with 64 cores, for use for testing by TRACC staff or special permission by the TRACC Director)
* virtual (This queue is only available for testing and is considered under construction. Please do not use for now.)
* test (This queue is only available for testing, and is only available with permission by the TRACC Director. The nodes as currently configure are
not very powerful but have large amounts of RAM.)

Revision as of 19:45, April 13, 2021

Introduction To ARROW

TRACC has now combined the hardware from the Phoenix and Zephyr clusters into the ARROW cluster. This consolidation allows efficient administration of TRACC cluster services with limited staff. To avoid the problems of load balancing, the different types of hardware nodes on the ARROW cluster are partitioned and available in queues. When new hardware is installed to expand cluster resources, it will be made available via a new queue. The documentation at Using the Clusters describes procedures for using ARROW.

ARROW is arranged such that there is a single set of login nodes, a singe file system, and single user home directory that serves all of the nodes in all of the queues.

ARROW Queues

There are currently five queues that are available with some restrictions about who can use them as described below.

  • batch (default queue, with 94 nodes, each node with 16 floating point cores available for general use)
    • 92 nodes have 32 GB of RAM
    • 2 nodes (nodes 1 and 2) with 128GB
    • 2 nodes (nodes 3 and 4) with 64GB
  • nhtsa (with 12 nodes, each with 28 cores and 64 GB of RAM, only available to the NHTSA project)
  • arrow (one new EPYC server with 64 cores, for use for testing by TRACC staff or special permission by the TRACC Director)
  • virtual (This queue is only available for testing and is considered under construction. Please do not use for now.)
  • test (This queue is only available for testing, and is only available with permission by the TRACC Director. The nodes as currently configure are

not very powerful but have large amounts of RAM.)