The University of Arizona
    For questions, please open a UAService ticket and assign to the Tools Team.
Page tree
Skip to end of metadata
Go to start of metadata

You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 95 Next »


El Gato

Implemented at the start of 2014 and it has been reprovisioned with CentOS 7 and new compilers and libraries. It also uses PBS as the scheduler consistent with Ocelote.  El Gato is a large GPU cluster, purchased by an NSF MRI grant by researchers in Astronomy and SISTA. Whilst the Nvidia K20 GPUs are more than five years old, they are still valuable for single-precision workload. There are 90 nodes with one or two GPU's.


Implemented in the middle of 2016.  It is designed to support all workloads on the standard nodes except:

  1. Large memory workloads that do not run within the 192GB RAM of each node can also be handled with the large memory node.
  2. GPU's are available as a buy-in option or windfall and are now available standard on Ocelote.

Ocelote now has 46 nodes with Nvidia P100's that are available for "standard" and "windfall" queues. See details at GPU Nodes


Similar to Ocelote, it will have standard CPU nodes (with 96 cores and 512 GB of memory per node), GPU nodes (with Nvidia V100) and a high-memory node (3 TB). Local scratch storage increased to ~1.6 TB. The new cluster is going to run on CentOS 7 which will improve compatibility with the latest software.

Puma is currently different from ElGato and Ocelote in these ways:

  • Puma uses SLURM for job scheduling
  • Puma has many programs installed natively (e.g. Singularity). You won't have to module load Singularity any more!
  • There's a new alias command: interactive. This automatically places you in a single-cpu interactive session for debugging, testing, and development. 
  • Modules are no long available on the login nodes. To view and test modules, an interactive session is necessary (see above).

Free vs Buy-In

The HPC resources at UA are differentiated from many other universities in that there is central funding for a significant portion of the available resources. Each PI receives a standard monthly allocation of hours at no charge.  There is no charge to the allocation for windfall usage and that has proven to be very valuable for researchers with substantial compute requirements.  

Research groups can 'Buy-In' (add resources such as processors, memory, storage, etc.) to the base HPC systems as funding becomes available. Buy-In research groups will have highest priority on the resources they add to the system.  If the expansion resources are not fully utilized by the Buy-In group they will be made available to all users as windfall.

Details on allocations

Details on buy-in

Test Environment

HPC has a test / trial environment as well as the primary clusters detailed below.  This environment is intended to be used for projects that are six months or less in duration and cannot be run on the production systems. Reasons for not being able to be run on the production systems include requiring root access, and hardware or software requirements that cannot be met by one of the production systems. If you have a project in mind that we might be able to support, contact 

CPUXeon Westmere-EP X5650
Dual 6-core
Disk10TB (5 x 2TB)
NetworkGbE and QDR IB

Compute System Details


El Gato




IBM System X iDataPlex dx360 M4

Lenovo NeXtScale nx360 M5Penguin Altus XE2242

Year Purchased


2016 (2018 P100 nodes)2020

Node Count



236 CPU-only
2 High-memory

Total System Memory (TB)




2x Xeon E5-2650v2 8-core (Ivy Bridge)

2x Xeon E5-2695v3 14-core (Haswell)
2x Xeon E5-2695v4 14-core (Broadwell)
4x Xeon E7-4850v2 12-core (Ivy Bridge)

2x AMD EPYC 7642 48-core (Rome)

Cores / Node (schedulable)


28c (48c - High-memory node)94c

Total Cores



Processor Speed


2.3GHz (2.4GHz - Broadwell CPUs)2.4GHz

Memory / Node

256GB (64GB - CPU-only nodes)

192GB (2TB - High-memory node)

512GB (3TB - High-memory nodes)


40 nodes with 2 K20X
42 nodes with 1 K20X

46 NVIDIA P100 (16GB)
15 NVIDIA K80 (buy-in only)


/tmp~840 GB spinning
/tmp is part of root filesystem
~840 GB spinning
/tmp is part of root filesystem
~1640 GB NVMe
/tmp is part of root filesystem

HPL Rmax (TFlop/s)




Centos 7

 CentOS 6CentOS 7


FDR Inifinband

FDR Infiniband for node-node
10 Gb Ethernet node-storage

1x 25Gb/s Ethernet RDMA (RoCEv2)
1x 25Gb/s Ethernet storage

* Includes high-memory and GPU node CPUs

Example Resource Requests

Node TypencpuspcmemMax memSample Request Statement

#PBS -l select=1:ncpus=16:mem=62gb:pcmem=4gb


#PBS -l select=1:ncpus=16:mem=250gb:ngpus=1:pcmem=16gb


#PBS -l select=1:ncpus=28:mem=168gb:pcmem=6gb


#PBS -l select=1:ncpus=28:mem=224gb:np100s=1:os7=True

High Memory4842gb2016gb

#PBS -l select=1:ncpus=48:mem=2016gb:pcmem=42gb

Standard945gb 470gb

#SBATCH --nodes=1
#SBATCH --ntasks=94
#SBATCH --mem=470gb

GPU4945gb 470gb

#SBATCH --nodes=1
#SBATCH --ntasks=94
#SBATCH --mem=470gb
#SBATCH --gres=gpu:1

High Memory9432gb3000gb

#SBATCH --nodes=1
#SBATCH --ntasks=94
#SBATCH --mem=3008gb
#SBATCH --constraint=hi_mem

Two GPUs may be requested on ElGato with ngpus=2
There is a single node available on Ocelote with two GPUs. To request it, use np100s=2
Set os7=False for a CentOS 6 GPU node
Up to four GPUs may be requested on Puma with --gres=gpu=1, 2, 3, or 4

  • No labels