Usage policy

From Storrs HPC Wiki
Revision as of 09:22, 27 April 2016 by Eds08006 (talk | contribs) (Storage)
Jump to: navigation, search

To be fair to all users of the cluster, please be aware of these resource limits and usage expectations.


Name Path Size (GB) Persistence Backed up? Purpose
Home ~ 2 Permanent Yes Personal storage, available on every node
Group /shared By request Permanent Yes Group storage for collaborative work
Fast /scratch/scratch2 438,000 (shared) 2 weeks No Fast storage, generally for result files
Local to node /work 100 5 days No Useful for large intermediate files, globally accessible from /misc/cnXX
  • Data deletion of directories inside the scratch2 folder is based on modification time. You will get 3 warnings by email before deletion.
  • If you try to run ls on either the /home, /shared, or /misc/cnXX directories, you might not see them. They are invisible because they are mounted on demand by autofs, when an attempt is made to access a file under the directory, or using cd to enter the directory structure.
  • You can recover files on your own from our backed up directories using our snapshots within 2 weeks. Beyond 2 weeks we may be able to help if you contact us.
  • You can check on your home directory quota.

Scheduled Jobs

Jobs submitted through the slurm scheduler:

Job property Standard QoS Limit Longrun QoS Limit Haswell384 QoS Limit
Run time (hours) 36 72 18
Cores / CPUs 48 384
Jobs 8

Unscheduled programs

Run time (minutes) CPU limit Memory limit
20 5% 5%
  • We strongly discourage programs being run on the login nodes without the SLURM scheduler.
  • Any programs running for longer than 20 minutes and not in the approved list below will be throttled to 5% CPU usage.
  • Programs allowed on the login node are:
    • bzip
    • cp
    • du
    • emacs
    • fort
    • gcc
    • gfortran
    • gunzip
    • gzip
    • icc
    • mv
    • sftp
    • smbclient
    • ssh
    • tar
    • vim
    • wget

Shared Read-Only Datasets

Users who need read-only datasets can contact our administrators ( to request the dataset. For example, people who study bioinformatics often need reference dataset for different organisms. The reference dataset is usually very large so user can only save them in /scratch. But. it is inconvenient to touch the dataset every 15 days to prevent deletion. If you have such kind of dataset, we can store the dataset for you. The dataset must meet the following requirements:

The shared dataset is under path: /scratch/scratch2/shareddata/. The data under this directory will be stored permanently. Now we have 4 reference datasets in genome directory: hg19 hg38 mm9 and mm10.

To make the linking path shorter, you can create a soft link with dataset under your home directory. For example:

$ cd 
$ link -s /scratch/scratch2/shareddata/genome ./genome
Retrieved from ""