MPI Cluster¶
The MPI cluster is optimized to support parallel workloads running across many nodes at once. High-speed networking enables low-latency communication between processes, making the cluster ideal for tightly coupled codes using message passing interfaces or other distributed frameworks.
Jobs on the MPI cluster are allocated a minimum of 2 nodes. Users who regularly run single-node workloads should submit to the SMP cluster instead, which is specifically designed for single node jobs. While running single node jobs on MPI is allowed, resources allocated on the unused node(s) will still be count against the user's consumed service units.
Specifications¶
| Partition | Nodes | GPU | VRAM | GPU/Node | --constraint | CPU | Cores/Node | Mem/Node | Scratch | Network | Node Names |
|---|---|---|---|---|---|---|---|---|---|---|---|
| mpi | 136 | N/A | N/A | N/A | hdr | Intel Xeon Gold 6342 | 48 | 512 GB | 1.6 TB NVMe | HDR200 IB | mpi-n[0-1,10,100-109,11,110-119,12,120-129,13,130-135,14-19,2,20-29,3,30-39,4,40-49,5,50-59,6,60-69,7,70-79,8,80-89,9,90-99] |
| ndr | 18 | N/A | N/A | N/A | ndr | Amd Epyc 9575F | 128 | 1.5 TB | 3.2 TB NVMe | NDR-200 IB | mpi-n[136-153] |