Body
The AI.Panther HPC Cluster at Florida Tech is a Aspen Systems cluster, comprised of 16 compute nodes (Total of 768 processor cores and 6,144 GB RAM), 8 GPU nodes with each node containing 4 Nvidia A100 GPUs , 3 storage nodes, 1 login node, and 1 head node. The AI.Panther cluster was funded by the National Science Foundation major Research Implementation grant.
A HPC Cluster is a set of computers working in parallel performing similar to a supercomputer for a fraction of the price. The HPC Cluster is made up of a cluster of nodes connected by a high-speed network that perform intense computing tasks.
To learn about High Performance Computing or Supercomputers, see Wikipedia.
Hardware
Head / Login Nodes
- 4 x Intel Xeon Cascade Lake Silver 4210R, 2.4 GHz 10-Core CPUs
- 12 x 8 = 96GB RAM
- 4 x 960GB Enterprise SSD
A100 PCIe GPU Nodes
- 4 x AMD EPYC 7402P Rome, 2.8 GHz 24-Core CPUs
- 32 x 32 = 1024GB RAM
- 4 x 960GB Enterprise SSD
- 16 x NVIDIA Tesla Ampere A100 40GB Memory, PCIe
- 24 x A100 NVLink Bridge
A100 SXM4 GPU Nodes
- 8 x AMD EPYC 7402 Rome, 2.8 GHz 24-Core CPUs
- 64 x 32 = 2048GB RAM
- 4 x 960GB Enterprise SSD
- 16 x NVIDIA Tesla Ampere A100 40GB Memory, 4 baseboards with 4 A100s with NVLINK
High Memory Compute Nodes
- 32 x Intel Xeon Cascade Lake Gold 6240R, 2.4GHz, 24-Core CPUs
- 192 x 32 = 6144GB RAM
- 16 x 960GB Enterprise SSD
ZFS User/Home Fileserver (78TB after overhead)
- 2 x Intel Xeon Cascade Lake Silver 4215R, 3.20GHz, 8-Core CPUs
- 12 x 16 = 192GB RAM
- 2 x 240GB Enterprise SSD
- 8 x 14TB SAS HDD configured as RAIDZ2
- 2 x P4800X 375GB Optane SSDs
ZFS Archive Fileserver (420TB after overhead)
- 2 x Intel Xeon Cascade Lake Silver 4215R, 3.20GHz, 8-Core CPUs
- 12 x 32 = 384GB RAM
- 2 x 240GB Enterprise SSD
- 36 x 16TB SATA HDD configured as four RAIDZ2 arrays
- 2 x P4800X 375GB Optane SSDs
Network Switches
- One externally managed HDR IB switch
- One managed 1GbE switch
- One 25Gb (18 port)/100Gb (4 port) Ethernet switch
- Nodes have 25Gb Ethernet connection to the 25GbE switch (Dual Port 1GbE with IPMI, HDR100 IB Connection). 100GbE links from 25GbE switch are available for future expansion.