clusters:minerva:first_steps
Table of Contents
First steps on Minerva
Logging in
Minerva can be reached via ssh to one of two login nodes, which are known by DNS as minerva01.aei.mpg.de and minerva02.aei.mpg.de. Internally, their names are login01 and login02.
- Learn how to type “minerva” without mixing the letters.
- Find the private counterpart of the submitted public key.
- Setup a couple of entries in your .ssh/config file, to make typing easier, and use the correct key.
Once logged in, you can access the other login node, and the compute nodes (node[001-594]) by passwordless ssh - a special keypair without passphrase has been created and installed for you. Do not use this key outside of Minerva.
Slurm Basics:
Job submission:
Interactive sessions:
salloc -p devel …
will submit a job to the selected partition- if there are free resources (and you're allowed to use them!), this starts a new shell on the login node and assigns the requested resources
- you may check using
env | grep SLURM
[stefgru@login01 ~]$ env | grep SLURM SLURM_NODELIST=node579 SLURM_JOB_NAME=bash SLURM_NODE_ALIASES=(null) SLURM_NNODES=1 SLURM_JOBID=39809 SLURM_TASKS_PER_NODE=16 SLURM_JOB_ID=39809 SLURM_SUBMIT_DIR=/home/stefgru SLURM_JOB_NODELIST=node579 SLURM_CLUSTER_NAME=minerva SLURM_JOB_CPUS_PER_NODE=16 SLURM_SUBMIT_HOST=login01.cluster SLURM_JOB_PARTITION=devel SLURM_JOB_NUM_NODES=1
- you're still on te login node,
srun
will use the allocated node(s) - (recommendation: start
screen
to have access to the node(s) in parallel) srun …
exit
will close the session, and return the allocated node(s) to the pool
It is recommended to use IntelMPI (IMPI) with Slurm (that's what PIK people say). Slurm and IMPI will do the correct binding for you, with OpenMPI you may see strange effects.
(to be extended)
clusters/minerva/first_steps.txt · Last modified: 2021/08/03 12:42 by Steffen Grunewald