The current infrastructure comprises:
- Seven nodes with 4x 80GB A100 GPUs, sliced into 4x 20GB MIGs.
- Twenty four (28), 32-core nodes, each with 512 GB of memory and approximately 10 TB
of volatile-scratch disk space. - Twelve (18) NVIDIA Tesla P6 GPUs, with 16 GB of memory (compatible with the CUDA,
OpenGL, OpenCL, and Vulkan APIs). - One AMD FirePro S7150 GPUs, with 8 GB of memory (compatible with the Direct X,
OpenGL, OpenCL, and Vulkan APIs). - One node with six (6) V100 GPUs
Job Management is handled by the Slurm Workload Manager.
The "cluster" mounts multi-TB, NFS-provided storage, which serves both persistent-scratch data (not backed up) and persistent-store data (backed up).