--- type: docs linkTitle: Details title: Hardware Details weight: 2 aliases: - /hardware.html - /hardware --- ## Storage * Four enterprise class HPC storage systems * Approximately 3 PB production and 3 PB backup storage (total 6 PB or 6,144 TB) * GPFS (NFS and SAMBA via GPFS) * Automatic snapshots and archival backups ## Network * Ethernet * 5 x 1 Gb/s switch * 5 x 1 Gb/s switch 10 Gig uplink * 1 x 10 Gb/s switch for campus high performance research network * Redundant, load balanced, robust mesh topology * Interconnect * 56/200/400 Gb/s InfiniBand (FDR/HDR/NDR) ## Head Nodes All users should access the cluster via SSH through `cluster.hpcc.ucr.edu`. This address will automatically balance traffic to one of the available head nodes. * Jay * Resources: 64 cores, 512 GB memory * Primary function: submitting jobs to the queuing system * Secondary function: development; code editing and running small (under 50 % CPU and under 30 % RAM) sample jobs * Lark (Currently acting as a "hot-spare") * Resources: 64 cores, 512 GB memory * Primary function: submitting jobs to the queuing system * Secondary function: development; code editing and running small (under 50 % CPU and under 30 % RAM) sample jobs ## Worker Nodes * Batch * c01-c48: each with 64 AMD cores and 512 GB memory * Intel * i01-i54: each with 32 Intel Broadwell cores and 256-512 GB memory * Epyc * r21-r38: each with 256 AMD EPYC cores and 1 TB memory * Highmem * h01-h06: each with 32 Intel cores and 1024 GB memory * GPU * gpu01-gpu02: each with 32 (HT) cores Intel Haswell CPUs and 2 x NVIDIA Tesla K80 GPUs (12GB and 2496 CUDA cores per GPU) and 128 GB memory * gpu03-gpu04: each with 48 (HT) cores Intel Broadwell CPUs and 4 x NVIDIA Tesla K80 GPUs (12GB and 2496 CUDA cores per GPU) and 512 GB memory * gpu05: with 64 (HT) cores Intel Broadwell CPUs and 2 x NVIDIA Tesla P100 GPUs (16GB and 3584 CUDA cores per GPU) and 256 GB memory * gpu06-gpu08: with 64-128 (HT) cores AMD CPUs and 8 x NVIDIA A100 GPUs (80GB and 6912 CUDA cores per GPU) and 1,024 GB memory