The University of Arizona
    For questions, please open a UAService ticket and assign to the Tools Team.
Page tree
Skip to end of metadata
Go to start of metadata


Today's research generates datasets that are increasingly complex, larger, and distributed. This makes modern research analysis, archiving, and sharing ever more challenging. The support for advanced techniques to transport, store, manipulate, visualize, and interpret large datasets is critical to advancing modern science.

The University’s Research Data Center provides data storage for active analysis on the high-performance computers (HPCs). Using central computing storage services and resources, University researchers, faculty researchers, and post-doctoral researchers are able to:

  • Share research data in a collaborative environment with other UA affiliates on the HPC system
  • Store large-scale computational research data
  • Request additional storage for further data analysis

Your storage is mounted as a filesystem and all the clusters have access to the same filesystems.

Qumulo Storage Array

In 2020 we implemented an all-flash based Qumulo Storage array with capacity of 2.29 PB of raw disk. Our leading requirement was fast, very fast. The old array had plenty of capacity but would make for slow calculations due to contention for the slower disk drives.  So the new array was no longer a bottle-neck for compute, and providing faster compute increases the efficiency of the supercomputers. The trade-off is that it is very expensive and we do not currently have an option for renting storage by the TeraByte.

We have developed a plan for Tier 2 storage which will be where research data is stored that is not in active compute.  

Check Disk Quota - Disk quotas can be checked through by selecting Storage.


/home/groups/PI/xdiskOn node /tmp
Backed up?nononono

Lasts as long as your
HPC account 

duration of job 
Maximum space50GB500GB

Up to 20TB
limit of 150 days
can renew once 

< 800GB to 1.4TB
File count limit
600 files / GB 


  • xdisk has been simplified if you have been using it.  No more tables to guess the size or duration
  • The capacity has been greatly increased - the maximum is now 20TB
  • You can specify less than 150 days, but 150 days is the default and the maximum duration you can select, except that you can renew once for a total of 300 days.
  • The usage is detailed on this page.


  • Purchase disk drives to be added to the storage system for dedicated group storage. The end of support for the current array is at the end of 2024. That is the end of guaranteed support.  Beyond that depends on when we need to buy a replacement array.
  • The cost estimate for adding a unit to the Qumulo storage array is $120,000 for 133 TB. This is more expensive and less granular than our previous clusters due to the speed trade-off mentioned above. 
  • For groups that need less than 133TB: using the free storage allocations or /xdisk option is the best option.
  • This space is NOT backed up
  • Files that need to be kept for 1-3 years can be offloaded to other platforms, like the University of Arizona Google Drive.

We strongly recommend that you do some regular housekeeping of your allocated space.
Millions of files are hard to manage for both the user and systems support. Archiving or using a tool like tar will help keep our disk arrays efficient and will make data migration quicker


Research computing has a contract with Globus.  This service provides for efficient data movement and sharing. Tier 2 in this section has the details.  

  • No labels