beautypg.com

A.7 submitting an hp-mpi job with lsf, View the node state, Show the environment – HP XC System 4.x Software User Manual

Page 122: Run the job, View the running job

background image

View the node state:

$ sinfo
PARTITION AVAIL TIMELIMIT NODES STATE NODELIST
lsf up infinite 4 idle n[13-16]

A.7 Submitting an HP-MPI Job with LSF

This example shows how to run an

MPI

job with the bsub command.

Show the environment:

$ lsid
Platform LSF HPC version, Update n, build date stamp
Copyright 1992-2008 Platform Computing Corporation

My cluster name is penguin
My master name is lsfhost.localdomain

$ sinfo
PARTITION AVAIL TIMELIMIT NODES STATE NODELIST
lsf up infinite 4 alloc n[13-16]

$ lshosts
HOST_NAME type model cpuf ncpus maxmem maxswp server RESOURCES
lsfhost.loc SLINUX6 DEFAULT 1.0 8 1M - Yes (slurm)

$ bhosts
HOST_NAME STATUS JL/U MAX NJOBS RUN SSUSP USUSP RSV
lsfhost.localdomai ok - 8 0 0 0 0 0

Run the job:

$ bsub -I -n6 -ext "SLURM[nodes=3]" mpirun -srun /usr/share/hello
Job <1009> is submitted to default queue .
<>
<>
I'm process 0! from ( n13 pid 27222)
Greetings from process 1! from ( n13 pid 27223)
Greetings from process 2! from ( n14 pid 14011)
Greetings from process 3! from ( n14 pid 14012)
Greetings from process 4! from ( n15 pid 18227)
Greetings from process 5! from ( n15 pid 18228)
mpirun exits with status: 0

View the running job:

$ bjobs -l 1009
Job <1009>, User , Project ,
Status , Queue ,
Interactive mode, Extsched ,
Command

date and time stamp: Submitted from host ,
CWD <$HOME>, 6 Processors Requested;
date and time stamp: Started on 6 Hosts/Processors
<6*lsfhost.localdomain>;
date and time stamp: slurm_id=22;ncpus=6;slurm_alloc=n[13-15];
date and time stamp: Done successfully.
The CPU time used is 0.0 seconds.

SCHEDULING PARAMETERS:
r15s r1m r15m ut pg io ls it tmp swp mem
loadSched - - - - - - - - - - -
loadStop - - - - - - - - - - -

122

Examples