SLURM

From wiki.hpc.mk
Revision as of 12:55, 27 August 2021 by Boris (talk | contribs)

Initiate and manage SLURM tasks

Most used parameters:

Parameters Description
#SBATCH --ntasks-per-node=2 # Number of tasks per phisical CPU core
#SBATCH --time=1:00:00 # Script duration (days-hrs:min:sec)
#SBATCH --job-name=test_job # Job name
#SBATCH --mem=1G # Ram memory for rendering (e.g. 1G, 2G, 4G)
#SBATCH --error=testerror_%j.error # Print the errors that occur when executing the job
#SBATCH --cpus-per-task=1 # Number of processors required for a single task
#SBATCH --output=testoutput_%j.out # Print the results from scripts and the values it returns
#SBATCH --gres=gpu:2 # Number of cards per one nod allocated for the job
#SBATCH --nodelist=cuda4 # Executing on specific nodes, e.g. cuda4 is for executing only on cuda4 host


Example by executing a simple script

  1. !/bin/bash
  1. SBATCH --job-name=test_job
  2. SBATCH --time=1:00:00
  3. SBATCH –-ntasks-per-node=1
  4. SBATCH --error=testerror_%j.error
  5. SBATCH --output=testoutput_%j.out

export PATH="/opt/anaconda3/bin:$PATH" source /opt/anaconda3/etc/profile.d/conda.sh conda create -n virtualenv python=3.8 conda activate virtualenv

echo "FINKI FCC"


The script is executed via sbatch <scriptname>.sh