Differences

This shows you the differences between two versions of the page.

Link to this comparison view

Both sides previous revision Previous revision
Next revision
Previous revision
Next revisionBoth sides next revision
doku:vsc5quickstart [2022/06/24 09:54] – [Intel MPI] jzdoku:vsc5quickstart [2023/02/17 17:45] msiegel
Line 1: Line 1:
 ====== Quick start guide for VSC-5 ====== ====== Quick start guide for VSC-5 ======
  
-**Status: 2022/04**+**Status: 2023/01**
  
 This page is under construction. This page is under construction.
Line 120: Line 120:
 128 physical cores (core-id 0-127) and 256 virtual cores available. 128 physical cores (core-id 0-127) and 256 virtual cores available.
  
-The A100 GPU nodes have 512GB RAM and the NVIDIA A100 cards have 40GB RAM each. +The A100 GPU nodes have 512GB RAM and the two NVIDIA A100 cards have 40GB RAM each. 
-At the moment 40 GPU nodes are installed.+60 A100 nodes are installed.
  
 +The A40 GPU nodes have 256GB RAM and the two NVIDIA A40 cards have 46GB each.
 +45 A40 nodes are installed.
 <code> <code>
 $ nvidia-smi $ nvidia-smi
Line 155: Line 157:
  
 ===== SLURM ===== ===== SLURM =====
 +**For the exact partition/queue setup see [[doku:vsc5_queue|Queue | Partition setup on VSC-5]]**
 +
 +type ''sinfo -o %P'' to see the available partitions:
 +
 +^partition ^nodes ^description^
 +| zen2_0256_a40x2 | | AMD CPU nodes with 2x AMD Epyc (Milan) and 2x NIVIDA A40 and 256GB RAM |
 +| jupyter | | reserved for the jupyterhub |
 +| login5 | | login nodes, not an actual slurm partition |
 +| zen3_2048 | | AMD CPU nodes with 2x AMD Epyc (Milan) and 2TB RAM |
 +| zen3_1024 | | AMD CPU nodes with 2x AMD Epyc (Milan) and 1TB RAM |
 +| zen3_0512* | | The default partition. AMD CPU nodes with 2x AMD Epyc (Milan) and 512GB RAM |
 +| cascadelake_0384 | | Intel CPU nodes with 2x Intel Cascadelake and 384GB RAM |
 +| zen3_0512_a100x2 | | AMD CPU nodes with 2x AMD Epyc (Milan) and 2x NIVIDA A100 and 512GB RAM |
  
-The following partitions are currently available: 
-<code> 
-$ sinfo -o %P 
-PARTITION 
-gpu_a100_dual* -> Currently the default partition. AMD CPU nodes with 2x AMD Epyc (Milan) and 2x NIVIDA A100 and 512GB RAM 
-cascadelake_0384 -> Intel CPU nodes with 2x Intel Cascadelake and 384GB RAM 
-zen3_0512 -> AMD CPU nodes with 2x AMD Epyc (Milan) and 512GB RAM 
-zen3_1024 -> AMD CPU nodes with 2x AMD Epyc (Milan) and 1TB RAM 
-zen3_2048 -> AMD CPU nodes with 2x AMD Epyc (Milan) and 2TB RAM 
-</code> 
  
 ==== QoS ==== ==== QoS ====
  
-During the friendly user test phase the QoS ''goodluck'' can be used for both partitions.+The following QoS are available for normal (=non private) projects: 
 + 
 +^QOS name ^ gives access to partition ^description^ 
 +|zen3_0512 | zen3_0512 | default | 
 +|zen3_1024 | zen3_1024 | 
 +|zen3_2048 | zen3_2048 | 
 +|cascadelake_0384 | cascadelake_0384 | 
 +|zen2_0256_a40x2 | zen2_0256_a40x2 | 
 +|zen3_0512_a100x2 | zen3_0512_a100x2 | 
 +|zen3_0512_devel | 5 nodes on zen3_0512 | 
  
 ==== Submit a Job ==== ==== Submit a Job ====
Line 179: Line 194:
 #SBATCH -J <meaningful name for job> #SBATCH -J <meaningful name for job>
 #SBATCH -N 1 #SBATCH -N 1
-#SBATCH --gres=gpu:2 
 ./my_program ./my_program
 </file> </file>
  
-This will submit a job in the default partition (gpu_a100_dual) using the default QoS (gpu_a100_dual).+This will submit a job in the default partition (zen3_0512) using the default QoS (zen3_0512).
  
 To submit a job to the cascadelake nodes: To submit a job to the cascadelake nodes:
Line 191: Line 205:
 #SBATCH -N 1 #SBATCH -N 1
 #SBATCH --partition=cascadelake_0384 #SBATCH --partition=cascadelake_0384
-#SBATCH --qos goodluck+#SBATCH --qos cascadelake_0384
 ./my_program ./my_program
 </file> </file>
Line 202: Line 216:
 #SBATCH -N 1 #SBATCH -N 1
 #SBATCH --partition=zen3_0512 #SBATCH --partition=zen3_0512
-#SBATCH --qos goodluck+#SBATCH --qos zen3_0512
 ./my_program ./my_program
 </file> </file>
Line 211: Line 225:
 #SBATCH -N 1 #SBATCH -N 1
 #SBATCH --partition=zen3_1024 #SBATCH --partition=zen3_1024
-#SBATCH --qos goodluck+#SBATCH --qos zen3_1024
 ./my_program ./my_program
 </file> </file>
Line 220: Line 234:
 #SBATCH -N 1 #SBATCH -N 1
 #SBATCH --partition=zen3_2048 #SBATCH --partition=zen3_2048
-#SBATCH --qos goodluck+#SBATCH --qos zen3_2048
 ./my_program ./my_program
 </file> </file>
Line 230: Line 244:
 #SBATCH -J <meaningful name for job> #SBATCH -J <meaningful name for job>
 #SBATCH -N 1 #SBATCH -N 1
-#SBATCH --partition=gpu_a100_dual +#SBATCH --partition=zen3_0512_a100x2 
-#SBATCH --qos goodluck+#SBATCH --qos zen3_0512_a100x2
 #SBATCH --gres=gpu:2 #SBATCH --gres=gpu:2
 ./my_program ./my_program
Line 240: Line 254:
 #!/bin/sh #!/bin/sh
 #SBATCH -J <meaningful name for job> #SBATCH -J <meaningful name for job>
-#SBATCH --partition=gpu_a100_dual +#SBATCH --partition=zen3_0512_a100x2 
-#SBATCH --qos goodluck+#SBATCH --qos zen3_0512_a100x2
 #SBATCH --gres=gpu:1 #SBATCH --gres=gpu:1
 ./my_program ./my_program
  • doku/vsc5quickstart.txt
  • Last modified: 2023/05/17 15:28
  • by msiegel