Using Slurm

From HPCwiki
Revision as of 10:18, 23 November 2013 by Hjmegens (talk | contribs)
Jump to navigation Jump to search

submitting jobs: sbatch

Consider this simple python3 script that should calculate Pi up to 1 million digits: <source lang='python'> from decimal import * D=Decimal getcontext().prec=10000000 p=sum(D(1)/16**k*(D(4)/(8*k+1)-D(2)/(8*k+4)-D(1)/(8*k+5)-D(1)/(8*k+6))for k in range(411)) print(str(p)[:10000002]) </source>

In order for this script to run, the first thing that is needed is that Python3, which is not the default Python version on the cluster, is load into your environment. Availability of (different versions of) software can be checked by the following command:

 module avail

In the list you should note that python3 is indeed available to be loaded, which then can be loaded with the following command:

 module load python/3.3.3

The following shell/slurm script can then be used to schedule the job using the sbatch command: <source lang='bash'>

  1. !/bin/bash
  2. #SBATCH --time=100
  3. SBATCH --ntasks=1
  4. SBATCH --output=output_%j.txt
  5. SBATCH --error=error_output_%j.txt
  6. SBATCH --job-name=calc_pi.py
  7. SBATCH --partition=research

time python3 calc_pi.py </source>

<source lang='bash'> sbatch run_calc_pi.sh </source>

 JOBID PARTITION     NAME     USER  ST       TIME  NODES NODELIST(REASON)
  3347  research calc_pi. megen002   R       0:03      1 node049

allocating resources interactively: sallocate

running MPI jobs on B4F cluster

monitoring submitted jobs: squeue

removing jobs from a list: scancel

other

external links