Available Hardware Resources
NMASC (nmpost)
The NMASC provides a processing cluster called the nmpost cluster to support CASA and AIPS execution, and a Lustre filesystem for a data storage.
Cluster
The nmpost cluster consists of 120 nodes. Nodes 001-060 are available for batch processing or interactive processing. They have no local storage, instead they have a 40Gb/s QDR Infiniband connection to the NMASC Lustre filesystem. These nodes support automatic EVLA pipeline processing, archive retrievals, batch processing requests and interactive processing sessions. Nodes 061-120 are reserved for our VLASS project. They have local NVMe storage, but no connection to the NMASC Lustre filesystem.
The operating system is Red Hat Enterprise Linux 8. The cluster scheduling software is Slurm or HTCondor. Torque/Moab is deprecated.
Nodes | CPU | Sockets | Cores per Socket | Clock | CPU Mark | RAM | OS | Local Storage | GPU(s) | Interconnect | Partitions | Scheduler |
---|---|---|---|---|---|---|---|---|---|---|---|---|
nmpost{001..005} |
Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL8 | NA | NA | 40Gb/s QDR Infiniband | batch | Slurm |
nmpost{006..010} |
E5-2630v3 | 2 | 8 | 2.4GHz | 17,041 | 498GiB | RHEL8 | NA | NA | 40Gb/s QDR Infiniband | interactive, batch | Slurm |
nmpost011 |
Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL8 | NA | NA | 40Gb/s QDR Infiniband | batch | HTCondor |
nmpost{012..015} |
Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL8 | NA | NA | 40Gb/s QDR Infiniband | batch | Torque |
nmpost{016..020} |
Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL8 | NA | NA | 40Gb/s QDR Infiniband | batch | HTCondor |
nmpost{021..036} |
E5-2640v4 | 2 | 10 | 2.4GHz | 20,394 | 498GiB | RHEL8 | NA | NA | 40Gb/s QDR Infiniband | interactive, batch | Slurm |
nmpost037 |
E5-2640v4 | 2 | 10 | 2.4GHz | 20,394 | 498GiB | RHEL8 | NA | Nvidia L4 | 40Gb/s QDR Infiniband | interactive, batch | Slurm |
nmpost{038..040} |
E5-2640v4 | 2 | 10 | 2.4GHz | 20,394 | 498GiB | RHEL8 | NA | NA | 40Gb/s QDR Infiniband | interactive, batch | Slurm |
nmpost{041..045} |
Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL8 | NA | NA | 40Gb/s QDR Infiniband | batch | Slurm |
nmpost{046..060} |
Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL8 | NA | NA | 40Gb/s QDR Infiniband | interactive, batch | Slurm |
nmpost061 | Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 750GiB | RHEL8 | 5.8TB NVMe | Nvidia L4 | 40Gb/s QDR Infiniband | vlass | HTCondor |
nmpost{062..090} | Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 750GiB | RHEL8 | 5.8TB NVMe | NA | 40Gb/s QDR Infiniband | vlass | HTCondor |
nmpost{091..120} | Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 750GiB | RHEL8 | 3.5TB NVMe | NA | NA | vlass | HTCondor |
CPU Marks are relative figures indicating performance. For example a node with a CPU Mark of 4000 can theoretically process roughly two times faster than a node with a CPU Mark of 2000. CPU Marks are not a measure of FLOPS.
RAM is in Gibibytes (Base2) units because both Slurm and HTCondor user Base2 for their memory requests. We also reserve 4GiB of RAM for the operating system. So the RAM column represents the maximum memory you can request for each node.
Lustre Filesystem
Lustre is a distributed parallel filesystem commonly used in HPC environments. Each cluster node sees the same shared filesystem. The NMASC Lustre filesystem is made up of ten storage servers each with four RAID arrays (40 total arrays) which are each 44TB in size. The total storage volume is 1.8PB. Individual cluster nodes can read and write to the Lustre filesystem in excess of 1GByte/sec. The entire filesystem can sustain roughly 15GB/s aggregate I/O.
The Lustre filesystem appears as /lustre/aoc on all Lustre enabled client computers.
Public Workstations
The NMASC has nine workstations for local visitors. The systems vary in processor, memory and storage since work is expected to be done mostly on the compute cluster, but all have 10Gb/s ethernet connections to the Lustre filesystem. Instructions for reserving workstations can be found in the Computing section of the Visiting the DSOC web page.
NAASC (cvpost)
The NAASC provides a processing cluster called the cvpost cluster to support CASA execution, and a Lustre filesystem for a data storage.
Cluster
The cvpost cluster consists of a number of nodes, most of which are available to users for either batch or interactive processing. Some nodes are reserved for large programs. None of the nodes have local storage, instead each node has a 40Gb/s QDR Infiniband connection to the NAASC Lustre filesystem. The cluster supports automatic ALMA pipeline processing, archive retrievals, batch processing requests and interactive processing sessions.
The operating system is being upgraded from Red Hat Enterprise Linux 7 to Red Hat Enterprise Linux 8. The cluster scheduling software is Slurm or HTCondor. Torque/Moab is deprecated.
Nodes | CPU | Sockets | Cores per Socket | Clock | CPU Mark | RAM | OS | Interconnect | Partition | Scheduler |
---|---|---|---|---|---|---|---|---|---|---|
cvpost001 | E5-2670 | 2 | 8 | 2.6GHz | 15,695 | 247GiB | RHEL8 | 40Gb/s QDR Infiniband | batch | Slurm |
cvpost{002..004} | E5-2640v3 | 2 | 8 | 2.6GHz | 18,094 | 247GiB | RHEL8 | 40Gb/s QDR Infiniband | batch | Slurm |
cvpost005 | E5-2670 | 2 | 8 | 2.6GHz | 15,695 | 247GiB | RHEL8 | 40Gb/s QDR Infiniband | batch | Slurm |
cvpost{006..007} | E5-2640v3 | 2 | 8 | 2.6GHz | 18,094 | 247GiB | RHEL8 | 40Gb/s QDR Infiniband | batch | Slurm |
cvpost{011..013} | E5-2670 | 2 | 8 | 2.6GHz | 15,695 | 247GiB | RHEL8 | 40Gb/s QDR Infiniband | batch | Slurm |
cvpost{018..023} | E5-2640v3 | 2 | 8 | 2.6GHz | 18,094 | 247GiB | RHEL8 | 40Gb/s QDR Infiniband | interactive, batch | Slurm |
cvpost{027..030} | E5-2630v3 | 2 | 8 | 2.4GHz | 10,271 | 247GiB | RHEL7 | 40Gb/s QDR Infiniband | NA | NA |
cvpost{101..108} | Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL8 | 40Gb/s QDR Infiniband | interactive2, batch2 | Slurm |
cvpost109 | Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL8 | 40Gb/s QDR Infiniband | NA | NA |
cvpost{110..116} | Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL8 | 40Gb/s QDR Infiniband | interactive2, batch2 | Slurm |
cvpost117 | Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL8 | 40Gb/s QDR Infiniband | NA | NA |
cvpost118 | Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL8 | 40Gb/s QDR Infiniband | interactive2, batch2 | Slurm |
cvpost119 | Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL8 | 40Gb/s QDR Infiniband | CVPOST | HTCondor |
cvpost{120..122} | Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL8 | 40Gb/s QDR Infiniband | NA | NA |
cvpost123 | Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL8 | 40Gb/s QDR Infiniband | interactive2, batch2 | Slurm |
cvpost{124..127} | Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL8 | 40Gb/s QDR Infiniband | plwg, plwg-i | Slurm |
cvpost128 | Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL8 | 40Gb/s QDR Infiniband | NA | NA |
cvpost{129..133} | Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL8 | 40Gb/s QDR Infiniband | plwg, plwg-i | Slurm |
cvpost{134..135} | Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL8 | 40Gb/s QDR Infiniband | NA | NA |
cvpost{136..137} | Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 750GiB | RHEL8 | 40Gb/s QDR Infiniband | plwg, plwg-i | Slurm |
cvpost{138..139} | Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL8 | 40Gb/s QDR Infiniband | batch2 | Torque |
cvpst140 | Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL8 | 40Gb/s QDR Infiniband | NA | NA |
cvpost{141..142} | Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL8 | 40Gb/s QDR Infiniband | batch2 | Slurm |
cvpost143 | Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL8 | 40Gb/s QDR Infiniband | batch2 | Torque |
cvpost144 | Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL8 | 40Gb/s QDR Infiniband | interactive2, batch2 | Slurm |
cvpost145 | Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL7 | 40Gb/s QDR Infiniband | NA | NA |
cvpost{146..147} | Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL8 | 40Gb/s QDR Infiniband | NA | NA |
cvpost{148..149} | Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL8 | 40Gb/s QDR Infiniband | interactive2, batch2 | Slurm |
cvpost150 | Gold 6136 | 2 | 12 | 3.0GHz | 33,555 | 498GiB | RHEL8 | 40Gb/s QDR Infiniband | NA | Slurm |
RAM is in Gibibytes (Base2) units because both Slurm and HTCondor user Base2 for their memory requests. We also reserve 4GiB of RAM for the operating system. So the RAM column represents the maximum memory you can request for each node.
Lustre Filesystem
Lustre is a distributed parallel filesystem commonly used in HPC environments. Each cluster node sees the same shared filesystem. There are two Lustre filesystems available at the NAASC. First, the NAASC Lustre filesystem is scratch space, used for operational and pipeline functions. It is made up of six storage servers each with four RAID arrays (16 total arrays) which are each 64TB in size. An additional two servers each provide four RAID arrays, each with ~190 TB. The total storage volume is 3.0PB. Individual nodes can read and write to the Lustre filesystem in excess of 1GByte/sec, the entire filesystem can sustain roughly 10GB/s aggregate I/O.
The Lustre filesystem appears as /lustre/naasc on all Lustre enabled client computers.
A separate, long term storage system is provided by CV Lustre. This storage, while intended mostly for staff use, is also leveraged for guest ("observer") accounts. It has four servers, each providing four RAID arrays of 128TB apiece, and one additional server with 4 x 192TB arrays. Total capacity is 2.8 PB.
Public Workstations
The NAASC has five workstations for local visitors. The systems have 8 x Intel E5-1660 Xeon 3.0 GHz processors, 32 GB RAM, 3 TB local disk space (accessible as /home/<hostname>_1 and /home/<hostname>_2) and a 10Gb/s ethernet connection to the Lustre filesystem.