salloc, sattach, sbcast, sdiag, sprio, sshare, strigger, sview, sh5util, scrun, scrontab: add pages (#11651)

pull/23/head
Juri Dispan 2023-12-18 23:53:21 +01:00 committed by GitHub
parent 5fbca8f752
commit 999326de39
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
10 changed files with 182 additions and 0 deletions

16
pages/linux/salloc.md Normal file
View File

@ -0,0 +1,16 @@
# salloc
> Start an interactive shell session or execute a command by allocating one or more nodes in a cluster.
> More information: <https://slurm.schedmd.com/salloc.html>.
- Start an interactive shell session on a node in the cluster:
`salloc`
- Execute the specified command synchronously on a node in the cluster:
`salloc {{ls -a}}`
- Only allocate nodes fulfilling the specified constraints:
`salloc --constraint={{(amd|intel)&gpu}}`

16
pages/linux/sattach.md Normal file
View File

@ -0,0 +1,16 @@
# sattach
> Attach to a Slurm job step.
> More information: <https://slurm.schedmd.com/sattach.html>.
- Redirect the IO streams (`stdout`, `stderr`, and `stdin`) of a Slurm job step to the current terminal:
`sattach {{jobid}}.{{stepid}}`
- Use the current console's input as `stdin` to the specified task:
`sattach --input-filter {{task_number}}`
- Only redirect `stdin`/`stderr` of the specified task:
`sattach --{{output|error}}-filter {{task_number}}`

13
pages/linux/sbcast.md Normal file
View File

@ -0,0 +1,13 @@
# sbcast
> Send a file to a job's allocated nodes.
> This command should only be used from within a Slurm batch job.
> More information: <https://slurm.schedmd.com/sbcast.html>.
- Send a file to all nodes allocated to the current job:
`sbcast {{path/to/file}} {{path/to/destination}}`
- Autodetect shared libraries the transmitted file depends upon and transmit them as well:
`sbcast --send-libs={{yes}} {{path/to/executable}} {{path/to/destination}}`

24
pages/linux/scrontab.md Normal file
View File

@ -0,0 +1,24 @@
# scrontab
> Manage Slurm crontab files.
> More information: <https://slurm.schedmd.com/scrontab.html>.
- Install a new crontab from the specified file:
`scrontab {{path/to/file}}`
- [e]dit the crontab of the current user:
`scrontab -e`
- [e]dit the crontab of the specified user:
`scrontab --user={{user_id}} -e`
- [r]emove the current crontab:
`scrontab -r`
- Print the crontab of the current user to `stdout`:
`scrontab -l`

20
pages/linux/sdiag.md Normal file
View File

@ -0,0 +1,20 @@
# sdiag
> Show information about the execution of `slurmctld`.
> More information: <https://slurm.schedmd.com/sdiag.html>.
- Show all performance counters related to the execution of `slurmctld`:
`sdiag --all`
- Reset performance counters related to the execution of `slurmctld`:
`sdiag --reset`
- Specify the output format:
`sdiag --all --{{json|yaml}}`
- Specify the cluster to send commands to:
`sdiag --all --cluster={{cluster_name}}`

16
pages/linux/sh5util.md Normal file
View File

@ -0,0 +1,16 @@
# sh5util
> Merge HDF5 files produced by the `sacct_gather_profile` plugin.
> More information: <https://slurm.schedmd.com/sh5util.html>.
- Merge HDF5 files produced on each allocated node for the specified job or step:
`sh5util --jobs={{job_id|job_id.step_id}}`
- Extract one or more data series from a merged job file:
`sh5util --jobs={{job_id|job_id.step_id}} --extract -i {{path/to/file.h5}} --series={{Energy|Filesystem|Network|Task}}`
- Extract one data item from all nodes in a merged job file:
`sh5util --jobs={{job_id|job_id.step_id}} --item-extract --series={{Energy|Filesystem|Network|Task}} --data={{data_item}}`

24
pages/linux/sprio.md Normal file
View File

@ -0,0 +1,24 @@
# sprio
> View the factors determining a job's scheduling priority.
> More information: <https://slurm.schedmd.com/sprio.html>.
- View the factors determining the scheduling priority of all jobs:
`sprio`
- View the factors determining the specified job's scheduling priority:
`sprio --jobs={{job_id_1,job_id_2,...}}`
- Output additional information:
`sprio --long`
- View information for the jobs of specified users:
`sprio --user={{user_name_1,user_name_2,...}}`
- Print the weights for each factor determining job scheduling priority:
`sprio --weights`

20
pages/linux/sshare.md Normal file
View File

@ -0,0 +1,20 @@
# sshare
> List the shares of associations to a cluster.
> More information: <https://slurm.schedmd.com/sshare.html>.
- List Slurm share information:
`sshare`
- Control the output format:
`sshare --{{parsable|parsable2|json|yaml}}`
- Control the fields to display:
`sshare --format={{format_string}}`
- Display information for the specified users only:
`sshare --users={{user_id_1,user_id_2,...}}`

25
pages/linux/strigger.md Normal file
View File

@ -0,0 +1,25 @@
# strigger
> View or modify Slurm trigger information.
> Triggers are actions that are automatically run when a given event occurs on a Slurm cluster.
> More information: <https://slurm.schedmd.com/strigger.html>.
- Register a new trigger. Execute the specified program when the specified event occurs:
`strigger --set --{{primary_database_failure|primary_slurmdbd_failure|primary_slurmctld_acct_buffer_full|primary_slurmctld_failure|...}} --program={{path/to/executable}}`
- Execute the specified program when the specified job terminated:
`strigger --set --jobid={{job_id}} --fini --program="{{path/to/executable}} {{argument1 argument2 ...}}"`
- View active triggers:
`strigger --get`
- View active triggers regarding the specified job:
`strigger --get --jobid={{job_id}}`
- Clear the specified trigger:
`strigger --clear {{trigger_id}}`

8
pages/linux/sview.md Normal file
View File

@ -0,0 +1,8 @@
# sview
> Start a GUI to view and modify the state of Slurm.
> More information: <https://slurm.schedmd.com/sview.html>.
- Start a GUI to view and modify the state of Slurm:
`sview`