HPC/Directories: Difference between revisions
m (→Local RAM disk) |
|||
Line 116: | Line 116: | ||
: In contrast to explicit trailing "cp" commands in the job script, this copy will be executed ''even if a job overruns its walltime''. See the [http://www.clusterresources.com/torquedocs21/commands/qsub.shtml#W qsub manual] for further information. | : In contrast to explicit trailing "cp" commands in the job script, this copy will be executed ''even if a job overruns its walltime''. See the [http://www.clusterresources.com/torquedocs21/commands/qsub.shtml#W qsub manual] for further information. | ||
<!-- | |||
== Local RAM disk == | == Local RAM disk == | ||
(2020-05-21: to be added) | (2020-05-21: to be added) | ||
[https://specifications.freedesktop.org/basedir-spec/basedir-spec-latest.html XDG Base Directory Specification] <code>$XDG_RUNTIME_DIR</code> – intended for small files. | |||
--> | |||
<!-- | <!-- |
Latest revision as of 19:50, July 18, 2022
Overview
Here is a summary of key directories related to Carbon, and the environment variables used to access them:
Environment variable | Typical value | Shared across nodes? | Purge schedule | Purpose |
---|---|---|---|---|
$HOME (same as ~ in shells)
|
/home/joe | yes | See CNM data retention policy | Your main configuration files and data. |
$SANDBOX
|
/sandbox/joe | yes | 6 weeks | Scratch space for transient job data, not backed up |
$TMPDIR (on login nodes)
|
/tmp | no | 6 weeks | General Unix scratch space |
$TMPDIR (during jobs)
|
/tmp/12345.sched1.... | no | at end of job | Job-specific scratch space, provided empty on job start. |
$PBS_O_WORKDIR
|
(directory where qsub was run)
|
yes | (same as parent file system) | typically used as cd $PBS_O_WORKDIR as first line in a job script
|
Home directory
$HOME ~ (tilde character)
Your home directory can be referred to in standard Unix fashion, as shown above, by either the environment variable or the tilde sign in most shells (but generally not application programs, especially not those written in Fortran).
- Files are backed up nightly.
- Your total file volume in $HOME is subject to (soft) quota of generally 0.5 TB.
- You may exceed the soft limit by about 10% during a grace period of one week. You will see an over-quota notice upon login.
- If your usage remains above the soft limit beyond the grace time, the file system will appear (to you) as being full. To recover, delete files.
Your files in $HOME are subject to CNM's Data Retention Policy, which specifies that all your files may be deleted from our servers as early as 30 days after your last active proposal has expired. At that time, your access to Carbon and its SSH gateway will be revoked.
Global scratch space
$SANDBOX
This environment variable points to a user-specific directory, shared across nodes like the home-directory.
Use this directory for short-lived files that need to be shared among multiple nodes, can get large, numerous, or change often. To accommodate this, usage policies are stricter than for /home:
- Files are not backed up.
- Hard quota are 3 TB in volume and 2 million in file count.
- Soft quota are 10 GB and 10,000 files.
- The grace period for overflowing a soft limit is 3 weeks.
- Files will be deleted automatically once they are older than 3 months.
These limits are subject to change.
The limits are aimed at keeping the space available for the intended use, typically for files of unusual size (F.O.U.S.) or for small files of unusual count.
Local scratch space
$TMPDIR
This variable and the directory it refers to is provided by the queueing system for all processes that execute a job. The directory:
- resides on local disk on each node,
- is named the same on each node,
- is not shared across nodes,
- is shared for processes on the same node (as many as given in "ppn=…"), in other words, the name is PBS job-specific, but not Unix PID-specific,
- typically provides about 100 GB of space,
- will be wiped upon job exit on each node.
The environment variable TMPDIR is not shared across nodes either. Either communicate it internal to your program, or have it exported by mpirun/mpiexec:
- OpenMPI
mpirun … \ -x TMPDIR \ [-x OTHERVAR] \ …
- Intel MPI
mpiexec.hydra … \ -genvlist TMPDIR[,OTHERVAR] \ …
Use in job files
You can use $TMPDIR
in one or more of the following ways:
- direct your application to store its temporary files there, which is typically done by command line switches or an environment variable such as:
export FOO_SCRATCH=$TMPDIR
- actually run your application there:
cd $TMPDIR
- In this case, make sure you either copy your input files there or you specify full paths to
$HOME
or$PPBS_O_WORKDIR
.
- copy files back upon job termination:
#PBS -W stageout=$TMPDIR/foo.ext@localhost:$PBS_O_WORKDIR #PBS -W stageout=$TMPDIR/*.bar@localhost:$PBS_O_WORKDIR/subdir
- The string on the left of the
@
character names the source files in the compute node file system. - The string on the right gives the destination host and directory, also as seen from the compute node. This means localhost refers to the primary compute node (rank 0 in MPI parlance). $PBS_O_WORKDIR by default stems from what was qsub's current directory on the submission node, but Carbon's user file systems are mounted on login nodes and compute nodes under the same paths.
- You may give the
stageout=…
directive multiple times, as shown.
- In contrast to explicit trailing "cp" commands in the job script, this copy will be executed even if a job overruns its walltime. See the qsub manual for further information.