Herramientas de usuario

Herramientas del sitio


hpc:slurm_commds

Comandos de SLURM

Comandos de usuarios PBS/Torque Slurm
Job submissionqsub <job script> sbatch <job script>
Job submissionqsub -q express -l nodes=2:ppn=16 -l mem=64g <job script> sbatch -p express -N 2 -c 16 –mem=64g <job script>
Job deletionqdel <job_id> scancel <job_id>
Job deletionqdel ALL scancel -u <user>
List jobs qstat [-u user] squeue [-u user] [-l for long format]
Job status qstat -f <job_id>jobinfo <job_id>
Job hold qhold <job_id> scontrol hold <job_id>
Job release qrls <job_id> ​scontrol release <job_id>
Node status pbsnodes -l sinfo -N -l
Environment PBS/Torque Slurm
Job ID $PBS_JOBID $SLURM_JOBID
Node list (entry per core) $PBS_NODEFILE $PBS_NODEFILE (still supported)
Slurm node list $SLURM_JOB_NODELIST(new format)
Submit directory $PBS_O_WORKDIR $SLURM_SUBMIT_DIR
Job Specification PBS/Torque Slurm
Script directive #PBS #SBATCH
Queue -q <queue> -p <partition>
Node count -l nodes=<count> -N <min[-max]>
Cores(cpu) per node -l ppn=<count> -c <count>
Memory size -l mem=16384 –mem=16g OR –mem-per-cpu=2g
Wall clock limit -l walltime=<hh:mm:ss> -t <days-hh:mm:ss>
Standard output file -o <file_name> -o <file_name>
Standard error file -e <file_name> -e <file_name>
Combine stdout/err -j oe (use -o without -e) [standard behaviour]
Direct output to directory -o <directory> -o “directory/slurm-%j.out”
Event notification -m abe –mail-type=[BEGIN, END, FAIL, REQUEUE, or ALL]
Email address -M <address> –mail-user=<address>
Job name -N <name> –job-name=<name>
Job dependency -W depend=afterok:<jobid> –depend=C:<jobid>
Node preference –nodelist=<nodes> AND/OR –exclude=<nodes>
Max jobs pool -A [m16,m32,..,m512] –qos=[max16jobs,max32jobs,..,max512jobs]
Account to charge -W group_list=<account> –account=<account>
hpc/slurm_commds.txt · Última modificación: 2017/07/05 10:06 (editor externo)