Running interactively on AccelerateAI

While we recommend using the batch queues for intensive workloads, we also have capacity for some interactive use of GPUs to test workflows and validate models before submitting larger jobs to the batch queues.

The A100 GPU is partitioned using Multi-Instance GPU; each interactive GPU has either 5GB or 10GB of memory, and 1/7 or 2/7 of the compute units of the full GPU respectively.

There are two routes to running interactively on AccelerateAI:

Both are time-limited, as all jobs on AccelerateAI are. If your training is at the point that it is taking longer than the time limit, then we recommend packaging your workload for the batch queue, whose more powerful GPUs should let it complete more quickly, and which will allow you to checkpoint and resume your job more easily.

JupyterHub

JupyterHub is currently in the process of being installed. Please check back later for instructions on how to use it.

For the time being, we recommend trying the command-line interface. If you need urgent access to a Jupyter Notebook interface backed by a GPU, then please get in touch with SA2C Support

Command-line

To allocate and start an interactive session on AccelerateAI, run:

$ srun --pty --account=scwXXXX --gres=gpu:1 --partition=accel_ai_mig /bin/bash

(Replace scwXXXX with your Supercomputing Wales project identifier.)

From here, you can load any needed modules and run software as if you were running directly at a command prompt on the machine.

Once your interactive session is concluded, make sure to end the session with the exit command, so that the GPU is freed to allow others to use it.