Using Slurm

From HPCwiki
Revision as of 10:11, 23 November 2013 by Hjmegens (talk | contribs)
Jump to navigation Jump to search

submitting jobs: sbatch

from decimal import * D=Decimal getcontext().prec=10000000 p=sum(D(1)/16**k*(D(4)/(8*k+1)-D(2)/(8*k+4)-D(1)/(8*k+5)-D(1)/(8*k+6))for k in range(411)) print(str(p)[:10000002])

<source lang='python'> from decimal import * D=Decimal getcontext().prec=10000000 p=sum(D(1)/16**k*(D(4)/(8*k+1)-D(2)/(8*k+4)-D(1)/(8*k+5)-D(1)/(8*k+6))for k in range(411)) print(str(p)[:10000002]) </source>

<source lang='bash'>

  1. !/bin/bash
  2. #SBATCH --time=100
  3. SBATCH --ntasks=1
  4. SBATCH --output=output_%j.txt
  5. SBATCH --error=error_output_%j.txt
  6. SBATCH --job-name=calc_pi.py
  7. SBATCH --partition=research

time python3 calc_pi.py </source>

 JOBID PARTITION     NAME     USER  ST       TIME  NODES NODELIST(REASON)
  3347  research calc_pi. megen002   R       0:03      1 node049

allocating resources interactively: sallocate

running MPI jobs on B4F cluster

monitoring submitted jobs: squeue

removing jobs from a list: scancel

other

external links