To be fair to all users of the cluster, please be aware of these resource limits and usage expectations.
|Name||Path||Size (GB)||Persistence||Backed up?||Purpose|
||2||Permanent||Yes||Personal storage, available on every node|
||By request||Permanent||Yes||Group storage for collaborative work|
||438,000 (shared)||2 weeks||No||Fast RAID-0 storage, generally for result files|
|Local to node||
||100||5 days||No|| Useful for large intermediate files, globally accessible from |
- Data deletion of directories inside the scratch2 folder is based on modification time. You will get 3 warnings before deletion.
- If you try to run
lson either the
/misc/cnXXdirectories, you might not see them. They are invisible because they are mounted on demand by
autofs, when an attempt is made to access a file under the directory, or using
cdto enter the directory structure.
- You can recover files on your own from our backed up directories using our snapshots within 2 weeks. Beyond 2 weeks we may be able to help if you contact us.
- You can check on your home directory quota.
Users who needs read-only dataset can contact our administrators (email@example.com) to request the dataset. For example, people who study bioinformatics often needs reference dataset for different organism. The reference dataset is usually very large so user can only save them in /scratch. But meanwhile the dataset is not writable, it is annoying to touch the dataset every 15 days since it will be deleted if stay untouched. If you have such kind of dataset, we can store the dataset for you. The dataset must meet the following requirements:
- dataset is read-only, cannot be writable or executable
- dataset is public, can be used by other users
The shared dataset is under path: /scratch/scratch2/shareddata/. Now we have 4 reference datasets in genome directory: hg19 hg38 mm9 and mm10. To make the linking path shorter, you can create a soft link with dataset under your home directory. For example:
$ cd $ link -s /scratch/scratch2/shareddata/genome ./genome
Jobs submitted through the slurm scheduler:
|Job property||Standard QoS Limit||Longrun QoS Limit||Haswell384 QoS Limit|
|Run time (hours)||36||72||18|
|Cores / CPUs||48||384|
|Node||Run time (minutes)||CPU limit||Memory limit|
|Compute node||10||Not allowed on compute nodes|
- We strongly discourage programs being run on the cluster without the SLURM scheduler.
- Programs on the compute nodes are not allowed.
- Some programs are allowed on the login node only: