Differences
This shows you the differences between two versions of the page.
Both sides previous revision Previous revision Next revision | Previous revision Next revisionBoth sides next revision | ||
doku:vsc5quickstart [2022/06/24 09:54] – [Intel MPI] jz | doku:vsc5quickstart [2023/02/17 17:45] – msiegel | ||
---|---|---|---|
Line 1: | Line 1: | ||
====== Quick start guide for VSC-5 ====== | ====== Quick start guide for VSC-5 ====== | ||
- | **Status: | + | **Status: |
This page is under construction. | This page is under construction. | ||
Line 120: | Line 120: | ||
128 physical cores (core-id 0-127) and 256 virtual cores available. | 128 physical cores (core-id 0-127) and 256 virtual cores available. | ||
- | The A100 GPU nodes have 512GB RAM and the NVIDIA A100 cards have 40GB RAM each. | + | The A100 GPU nodes have 512GB RAM and the two NVIDIA A100 cards have 40GB RAM each. |
- | At the moment 40 GPU nodes are installed. | + | 60 A100 nodes are installed. |
+ | The A40 GPU nodes have 256GB RAM and the two NVIDIA A40 cards have 46GB each. | ||
+ | 45 A40 nodes are installed. | ||
< | < | ||
$ nvidia-smi | $ nvidia-smi | ||
Line 155: | Line 157: | ||
===== SLURM ===== | ===== SLURM ===== | ||
+ | **For the exact partition/ | ||
+ | |||
+ | type '' | ||
+ | |||
+ | ^partition ^nodes ^description^ | ||
+ | | zen2_0256_a40x2 | | AMD CPU nodes with 2x AMD Epyc (Milan) and 2x NIVIDA A40 and 256GB RAM | | ||
+ | | jupyter | | reserved for the jupyterhub | | ||
+ | | login5 | | login nodes, not an actual slurm partition | | ||
+ | | zen3_2048 | | AMD CPU nodes with 2x AMD Epyc (Milan) and 2TB RAM | | ||
+ | | zen3_1024 | | AMD CPU nodes with 2x AMD Epyc (Milan) and 1TB RAM | | ||
+ | | zen3_0512* | | The default partition. AMD CPU nodes with 2x AMD Epyc (Milan) and 512GB RAM | | ||
+ | | cascadelake_0384 | | Intel CPU nodes with 2x Intel Cascadelake and 384GB RAM | | ||
+ | | zen3_0512_a100x2 | | AMD CPU nodes with 2x AMD Epyc (Milan) and 2x NIVIDA A100 and 512GB RAM | | ||
- | The following partitions are currently available: | ||
- | < | ||
- | $ sinfo -o %P | ||
- | PARTITION | ||
- | gpu_a100_dual* -> Currently the default partition. AMD CPU nodes with 2x AMD Epyc (Milan) and 2x NIVIDA A100 and 512GB RAM | ||
- | cascadelake_0384 -> Intel CPU nodes with 2x Intel Cascadelake and 384GB RAM | ||
- | zen3_0512 -> AMD CPU nodes with 2x AMD Epyc (Milan) and 512GB RAM | ||
- | zen3_1024 -> AMD CPU nodes with 2x AMD Epyc (Milan) and 1TB RAM | ||
- | zen3_2048 -> AMD CPU nodes with 2x AMD Epyc (Milan) and 2TB RAM | ||
- | </ | ||
==== QoS ==== | ==== QoS ==== | ||
- | During the friendly user test phase the QoS '' | + | The following |
+ | |||
+ | ^QOS name ^ gives access to partition ^description^ | ||
+ | |zen3_0512 | zen3_0512 | default | | ||
+ | |zen3_1024 | zen3_1024 | | ||
+ | |zen3_2048 | zen3_2048 | | ||
+ | |cascadelake_0384 | cascadelake_0384 | | ||
+ | |zen2_0256_a40x2 | zen2_0256_a40x2 | | ||
+ | |zen3_0512_a100x2 | zen3_0512_a100x2 | | ||
+ | |zen3_0512_devel | 5 nodes on zen3_0512 | | ||
==== Submit a Job ==== | ==== Submit a Job ==== | ||
Line 179: | Line 194: | ||
#SBATCH -J < | #SBATCH -J < | ||
#SBATCH -N 1 | #SBATCH -N 1 | ||
- | #SBATCH --gres=gpu: | ||
./ | ./ | ||
</ | </ | ||
- | This will submit a job in the default partition (gpu_a100_dual) using the default QoS (gpu_a100_dual). | + | This will submit a job in the default partition (zen3_0512) using the default QoS (zen3_0512). |
To submit a job to the cascadelake nodes: | To submit a job to the cascadelake nodes: | ||
Line 191: | Line 205: | ||
#SBATCH -N 1 | #SBATCH -N 1 | ||
#SBATCH --partition=cascadelake_0384 | #SBATCH --partition=cascadelake_0384 | ||
- | #SBATCH --qos goodluck | + | #SBATCH --qos cascadelake_0384 |
./ | ./ | ||
</ | </ | ||
Line 202: | Line 216: | ||
#SBATCH -N 1 | #SBATCH -N 1 | ||
#SBATCH --partition=zen3_0512 | #SBATCH --partition=zen3_0512 | ||
- | #SBATCH --qos goodluck | + | #SBATCH --qos zen3_0512 |
./ | ./ | ||
</ | </ | ||
Line 211: | Line 225: | ||
#SBATCH -N 1 | #SBATCH -N 1 | ||
#SBATCH --partition=zen3_1024 | #SBATCH --partition=zen3_1024 | ||
- | #SBATCH --qos goodluck | + | #SBATCH --qos zen3_1024 |
./ | ./ | ||
</ | </ | ||
Line 220: | Line 234: | ||
#SBATCH -N 1 | #SBATCH -N 1 | ||
#SBATCH --partition=zen3_2048 | #SBATCH --partition=zen3_2048 | ||
- | #SBATCH --qos goodluck | + | #SBATCH --qos zen3_2048 |
./ | ./ | ||
</ | </ | ||
Line 230: | Line 244: | ||
#SBATCH -J < | #SBATCH -J < | ||
#SBATCH -N 1 | #SBATCH -N 1 | ||
- | #SBATCH --partition=gpu_a100_dual | + | #SBATCH --partition=zen3_0512_a100x2 |
- | #SBATCH --qos goodluck | + | #SBATCH --qos zen3_0512_a100x2 |
#SBATCH --gres=gpu: | #SBATCH --gres=gpu: | ||
./ | ./ | ||
Line 240: | Line 254: | ||
#!/bin/sh | #!/bin/sh | ||
#SBATCH -J < | #SBATCH -J < | ||
- | #SBATCH --partition=gpu_a100_dual | + | #SBATCH --partition=zen3_0512_a100x2 |
- | #SBATCH --qos goodluck | + | #SBATCH --qos zen3_0512_a100x2 |
#SBATCH --gres=gpu: | #SBATCH --gres=gpu: | ||
./ | ./ |