Solid state storage¶
In addition to the Lustre file system, the Cirrus login and compute nodes have access to a shared, high-performance, solid state storage system (also known as RPOOL). This storage system is network mounted and shared across the login nodes and GPU compute nodes in a similar way to the normal, spinning-disk Lustre file system but has different performanc characteristics.
The solid state storage has a maximum usable capacity of 256 TB which is shared between all users.
Backups, quotas and data longevity¶
There are no backups of any data on the solid state storage so you should ensure that you have copies of critical data elsewhere.
In addition, the solid state storage does not currently have any quotas (user or group) enabled so all users are potentially able to access the full 256 TB capacity of the storage system. We ask all users to be considerate in their use of this shared storage system and to delete any data on the solid state storage as soon as it no longer needs to be there.
We monitor the usage of the storage system by users and groups and will potentially remove data that is stopping other users getting fair access to the storage and data that has not been actively used for long periods of time.
Accessing the solid-state storage¶
You access the solid-state storage at /scratch/space1
on both the login nodes and
on the compute nodes.
Everybody has access to be able to create directories and add data so we suggest
that you create a directory for your project and/or user to avoid clashes with files
and data added by other users. For example, if my project is t01
and my
username is auser
then I could create a directory with
mkdir -p /scratch/space1/t01/auser
When these directories are initially created they will be world-readable. If you do not want users from other projects to be able to see your data, you should change the permissions on your new directory. For example, to restrict the directory so that only other users in your project can read the data you would use:
chmod -R o-rwx /scratch/space1/t01
Copying data to/from solid-state storage¶
You can move data to/from the solid-state storage in a number of different ways:
- By copying to/from another Cirrus file system - either interactively on login nodes or as part of a job submission script
- By transferring directly to/from an external host via the login nodes
Local data transfer¶
The most efficient tool for copying to/from the Cirrus file systems (/home, /work) to
the solid state storage is generally the cp
command, e.g.
cp -r /path/to/data-dir /scratch/space1/t01/auser/
where /path/to/data-dir
should be replaced with the path to the data directory you are
wanting to copy and assuming, of course, that you have setup the t01/auser
subdirectories
as described above).
Note
If you are transferring data from your /work
directory, these commands can also
be added to job submission scripts running on the compute nodes to move data as
part of the job. If you do this, remember to include the data transfer time in the
overall walltime for the job.
Data from your /home
directory is not available from the compute nodes and must
therefore be transferred from a login node.
Remote data transfer¶
You can transfer data directly to the solid state storage from external locations using
scp
or rsync
in exactly the same way as you would usually do to transfer data
to Cirrus. Simply substitute the path to the location on the solid state storage for that
you would normally use for Cirrus. For example, if you are on the external location (e.g.
your laptop), you could use something like:
scp -r data_dir user@login.cirrus.ac.uk:/scratch/space1/t01/auser/
You can also use commands such as wget
and curl
to pull data from external
locations directly to the solid state storage.
Note
You cannot transfer data from external locations in job scripts as the Cirrus compute nodes do not have external network access.