Skip to content

Argonne Leadership Computing Facility

Argonne Leadership
Computing Facility

Job Queueing and Submission


SambaNova uses Slurm for job submission and queueing. Below are some of the important commands for using Slurm. For more information refer to Slurm Documentation.

NOTE: Run the python scripts using srun or sbatch, to ensure that concurrent jobs do not interfere with each other.

NOTE: sm-01 and sm-02 each have a separate scheduler. If you have already ssh'd into one of those nodes, you are all set.


The Slurm command srun can be used to run individual python scripts in parallel with other scripts on a cluster managed by Slurm. Examples of srun usage are shown below.

srun python compile -b=1 --pef-name="lenet" --output-folder="pef"
srun python test --pef="pef/lenet/lenet.pef"
srun python run --pef="pef/lenet/lenet.pef"


Alternatively, these jobs can be submitted to the Slurm workload manager through a batch script by using sbatch command. To do this, create a bash script ( here as an example) with the commands that you want to execute.


python compile -b=1 --pef-name="lenet" --output-folder="pef"
python test --pef="pef/lenet/lenet.pef"
python run --pef="pef/lenet/lenet.pef"

Then pass the bash script as an input to the sbatch command as shown below.

sbatch --output=pef/lenet/output.log

In case of the need to use multiple RDU's (2 in the example shown below), the sbatch command would be altered as:

sbatch --gres=rdu:2 <>


Squeue command provides information about jobs located in the Slurm scheduling queue.



Scancel is used to signal or cancel jobs, job arrays or job steps.

scancel job_id