... | ... | @@ -3,16 +3,15 @@ |
|
|
|
|
|
Jobs can be run on the cluster through the [srun](https://slurm.schedmd.com/srun.html) or [sbatch](https://slurm.schedmd.com/sbatch.html) commands. Both command supports common options, the main differences being that
|
|
|
* `srun` will block until the needed resources are available (or time out). It will print on the standard output and can run arbitrary commands, including sbatch scripts. Please refer to the [srun manual](https://slurm.schedmd.com/srun.html) for a full description.
|
|
|
* `sbatch` will schedule the job for later execution, when the requested resources are available. It will only work with an executable [sbatch script](). Please refer to the [sbatch manual](https://slurm.schedmd.com/sbatch.html) for a full description.
|
|
|
* `sbatch` will schedule the job for later execution, when the requested resources are available. It will only work with an executable [sbatch script](#running-in-batch). Please refer to the [sbatch manual](https://slurm.schedmd.com/sbatch.html) for a full description.
|
|
|
|
|
|
Here, we will only mention the most common/important options.
|
|
|
|
|
|
## Running a command on a compute node
|
|
|
### Selecting a partition
|
|
|
## Selecting a partition
|
|
|
|
|
|
Before you can run anything through SLURM, you need to select a partition. A partition can identify a job profile.
|
|
|
|
|
|
#### Licallo
|
|
|
### Licallo
|
|
|
All queue have a maximum time limit of 24 hours. Other limitation may apply in term of number of concurrently running jobs or resource usage.
|
|
|
|
|
|
Here are the currently available partitions on Licallo
|
... | ... | @@ -23,10 +22,10 @@ Here are the currently available partitions on Licallo |
|
|
* **x40** parallel job requiring fat node (typically hybrid job) should go in this partition. The nodes in this partition have 2 sockets/processors of 20 core each and 192Go of RAM.
|
|
|
* **1to** the big memory partition, only contains one node, with 1To of RAM and 4 sockets of 8 core each.
|
|
|
|
|
|
#### Lagrange
|
|
|
### Lagrange
|
|
|
TBD
|
|
|
|
|
|
### Running interactively
|
|
|
## Running interactively
|
|
|
|
|
|
Any command can be dispatched on a compute node through `srun`:
|
|
|
|
... | ... | @@ -54,7 +53,7 @@ $ |
|
|
* `--exclusive` wont share the node with another task.
|
|
|
* `make -j` our parallel make task
|
|
|
|
|
|
### Running in batch
|
|
|
## Running in batch
|
|
|
|
|
|
We need to write a slurm/sbatch script.
|
|
|
|
... | ... | |