Wiki source code of Slurm

Version 12.1 by Thomas Coelho (local) on 2025/03/18 13:16

Show last authors
1 SLURM is the Simple Linux Utility for Resource Management and is an open source, fault-tolerant, and highly scalable cluster management and job scheduling system for large and small Linux clusters.
2
3 Slurm is fully integrated in our system. You do not need set any environment variables.
4
5
6
7 {{toc/}}
8
9 == Partitions ==
10
11 A partition is a subset of the cluster, a bundle of compute nodes with the same characteristics.
12
13 Based on access restrictions our cluster is divided in different partitions. 'sinfo' will only show partitions you are allowed to use. Using 'sinfo -a' shows all partitons.
14
15 A partition is selected by '-p PARTITIONNAME'.
16
17 |=**Partition** |=**No. Nodes** |=**Cores/M** |=**Tot. Cores**|=**RAM/GB/;** |=**CPU** |=**Remark/Restriction**
18 |itp |10|20 |200|64 |Intel(R) Xeon(R) CPU E5-2640 v4 @ 2.40GHz|Common Usage
19 |fplo|2|12|24|256|Intel(R) Xeon(R) CPU E5-2630 v2 @ 2.60GHz|Group Valenti
20 |fplo|4|16|32|256|Intel(R) Xeon(R) CPU E5-2630 v3 @ 2.40GHz|Group Valenti
21 |dfg-xeon|5|16|32|128|Intel(R) Xeon(R) CPU E5-2630 v3 @ 2.40GHz|Group Valenti
22 |dfg-xeon|7|20|140|128|Intel(R) Xeon(R) CPU E5-2630 v4 @ 2.20GHz|Group Valenti
23 |iboga|34|20|880|64|Intel(R) Xeon(R) CPU E5-2640 v4 @ 2.40GHz|Group Rezzolla
24 |dreama|1|40|40|1024|Intel(R) Xeon(R) CPU E7-4820 v3 @ 1.90GHz|Group Rezzolla
25 |barcelona|8|40|320|192|Intel(R) Xeon(R) Gold 6148 CPU @ 2.40GHz|(((
26 Group Valenti
27 )))
28 |barcelona|1|40|40|512|(((
29 Intel(R) Xeon(R) Silver 4316 CPU @ 2.30GHz
30 )))|Group Valenti
31 |mallorca|4|48|192|256|AMD EPYC 7352 24-Core Processor|Group Valenti
32 |calea|36|64|2304|512|Intel(R) Xeon(R) Platinum 8358 CPU @ 2.10GHz|(((
33 Group Valenti
34 )))
35 |bilbao|7|64|448|512|Intel Xeon(R) Gold 6540 @ 2.20GHz
36 |majortom|1|64|64|256|AMD EPYC 7513 32-Core Processor|Group Bleicher
37
38 Most nodes are for exclusive use by their corresponding owners. The itp nodes are for common usage. Except for 'fplo' and '
39 majortom', all machines are connected with Infiniband for all traffic (IP and internode communitcation - MPI)
40
41 == Submitting Jobs ==
42
43 In most cases you want to submit a non interactive job to be executed in our cluster.
44
45 This is very simple for serial (1 CPU) jobs:
46
47 {{{ sbatch -p PARTITION jobscript.sh}}}
48
49 where jobscript.sh is a shell script with your job commands.
50
51 Running **openMPI** jobs is not much more complictated:
52
53 {{{ sbatch -p PARTITION -n X jobscript.sh}}}
54
55 where X is the number of desired MPI processes. Launch the job in the jobscript with:
56
57 {{{ mpirun YOUREXECUTABLE}}}
58
59 You don't have to worry about the number of processes or specific nodes. Both slurm and openmpi know
60 about each other.
61
62 Running **SMP jobs** (multiple threads, not necessary mpi). Running MPI jobs on a single node is recommended for the
63
64 dfg-big nodes. This are big host with up to 64 cpu's per node, but 'slow' gigabit network connection. Launch SMP jobs with
65
66 {{{ sbatch -p PARTITION -N 1 -n X jobscript.sh}}}
67
68
69
70 == Defining Resource limits ==
71
72 By default each job allocates 2 GB memory and a run time of 3 days. More resources can be requested by
73
74 {{{ --mem-per-cpu=<MB>}}}
75
76 where <MB> is the memory in megabytes. The virtual memory limit is 2.5 times of the requested real memory limit.
77
78 The memory limit is not a hard limit. When exceeding the limit, your memory will be swapped out. Only when using more the 110% of the limit your job will be killed. So be conservative, to keep enough room for other jobs. Requested memory is blocked from the use by other jobs.
79
80 {{{ -t or --time=<time>}}}
81
82 where time can be set in the format "days-hours". See man page for more formats.
83
84 == Memory Management ==
85
86 In Slurm you specify only one parameter, which is the limit for your real memory usage and drives the decision where your job is started. The virtual memory of your job maybe 2.5 times of your requested memory. You can exceed your memory limit by 20%. But this will be swap space instead of real memory. This prevents crashing if you memory limit is a little to tight.
87
88 == Inline Arguments ==
89
90 sbatch arguments can be written in the jobfile:
91
92 {{{#! /bin/bash
93 #
94 # Choosing a partition:
95 #SBATCH -p housewives
96
97 YOUR JOB COMMANDS....}}}
98
99
100
101 = Links =
102
103
104
105 * SLURM-Homepage [[url:http://slurm.schedmd.com/slurm.html]]