MPI and GPU Containers

Message Passing Interface (MPI) for running on multiple nodes

Distributed MPI execution of containers is supported by Singularity.

Since (by default) the network is the same inside and outside the container, the communication between containers usually just works. The more complicated bit is making sure that the container has the right set of MPI libraries to interact with high-speed fabrics. MPI is an open specification, but there are several implementations (OpenMPI, MVAPICH2, and Intel MPI to name three) with some non-overlapping feature sets. There are also different hardware implementations (e.g. Infiniband, Intel Omnipath, Cray Aries) that need to match what is inside the container. If the host and container are running different MPI implementations, or even different versions of the same implementation, MPI may not work.

The general rule is that you want the version of MPI inside the container to be the same version or newer than the host. You may be thinking that this is not good for the portability of your container, and you are right. Containerizing MPI applications is not terribly difficult with Singularity, but it comes at the cost of additional requirements for the host system.

Warning

Many HPC Systems, like Frontera, have high-speed, low latency networks that have special drivers. Infiniband, Aries, and OmniPath are three different specs for these types of networks. When running MPI jobs, if the container doesn’t have the right libraries, it won’t be able to use those special interconnects to communicate between nodes. This means that MPI containers don’t provide as much portability between systems.

Base Docker images

When running at TACC, we have a set of curated Docker images for use in the FROM line of your own containers. You can see a list of available images at https://github.com/TACC/tacc-containers.

Image Frontera Stampede2 Maverick2 Longhorn Local Dev
tacc/tacc-centos7-mvapich2.3-ib X   X   X
tacc/tacc-centos7-ppc64le-mvapich2-ib       X* X
tacc/tacc-centos7-mvapich2.3-psm2   X      
tacc/tacc-centos7-impi19.0.7-common X X     X
tacc/tacc-ubuntu18-mvapich2.3-ib X   X   X
tacc/tacc-ubuntu18-mvapich2.3-psm2   X      
tacc/tacc-ubuntu18-impi19.0.7-common X X     X

Note

The singularity version of these containers should be invoked with singularity run on HPC.

* Must be used with the mvapich2-gdr module.

LS5 is not present because it uses a proprietary fabric library which cannot be built into a container.

In this tutorial, we will be using use the tacc/tacc-ubuntu18-impi19.0.7-common image to satisfy the MPI architecture on Frontera, while also allowing intra-node (single-node, multiple-core) testing on your local development system.

Building a MPI aware container

On your local laptop, go back to the directory where you built the “pi” Docker image and download (or copy and paste) two additional files:

Take a look at both files. pi-mpi.py is an updated Python script that uses MPI for parallelization. Dockerfile.mpi is an updated Dockerfile that uses the TACC base image to satisfy all the MPI requirements on Frontera.

Note

A full MPI stack with mpicc is available in these containers, so you can compile code too.

With these files downloaded, we can build a new MPI-capable container for faster execution.

$ docker build -t USERNAME/pi-estimator:0.1-mpi -f Dockerfile.mpi .

Note

Don’t forget to change USERNAME to your DockerHub username!

To prevent this build from overwriting our previous container (USERNAME/pi-estimator:0.1), the “tag” was changed from 0.1 to 0.1-mpi. We also could have just renamed the “repository” to something like USERNAME/pi-estimator-mpi:0.1. You will see both conventions used on DockerHub. For different versions, or maybe architectures, of the same codebase, it is okay to differentiate them by tag. Independent codes should use different repository names.

Once you have successfully built the image, push it up to DockerHub with the docker push command so that we can pull it back down on Frontera.

$ docker push USERNAME/pi-estimator:0.1-mpi

Running an MPI Container Locally

Before using allocation hours at TACC, it’s always a good idea to test your code locally. Since your local workstation may not have as many resources as a TACC compute node, testing is often done with a toy sized problem to check for correctness.

Our pi-estimator:0.1-mpi container started FROM tacc/tacc-ubuntu18-impi19.0.7-common, which is capable of locally testing the MPI capabilities using shared memory. Launch the pi-mpi.py script with mpirun from inside the container. By default, mpirun will launch as many processes as cores, but this can be controlled with the -n argument.

Lets try computing Pi with 10,000,000 samples using 1 and 2 processors.

Run using 1 processor
$ docker run --rm -it USERNAME/pi-estimator:0.1-mpi \
        mpirun -n 1 pi-mpi.py 10000000

Run using 2 processors
$ docker run --rm -it USERNAME/pi-estimator:0.1-mpi \
        mpirun -n 2 pi-mpi.py 10000000

You should notice that while the estimate stayed roughly the same, the execution time halved as the program scaled from one to two processors.

Note

If the computation time did not decrease, your Docker Desktop may not be configured to use multiple cores.

Now that we validated the container locally, we can take it to a TACC node and scale it up further.

Running an MPI Container on Frontera

To start, lets allocate a single CLX (Cascade Lake) node, which has 56 physical cores per node, but lets only use 8 cores to make the log messages a little more legible.

Running interactively

Please use idev to allocate this 8-task compute node.

$ idev -m 60 -p normal -N 1 -n 8

Once you have your node, pull the container and run it as follows:

Load singularity module
$ module load tacc-singularity

Change to $SCRATCH directory so containers do not go over your $HOME quota
$ cd $SCRATCH

Pull container
$ singularity pull docker://USERNAME/pi-estimator:0.1-mpi

Run container sequentially
$ ibrun -n 1 singularity run pi-estimator_0.1-mpi.sif pi-mpi.py 10000000

Run container distributed
$ ibrun singularity run pi-estimator_0.1-mpi.sif pi-mpi.py 10000000

Run container with fewer tasks
$ ibrun -n 4 singularity run pi-estimator_0.1-mpi.sif pi-mpi.py 10000000

In our local tests, the container mpirun program was used to launch multiple processes, but this does not scale to multiple nodes. When using multiple nodes at TACC, you should always use ibrun to call singularity to launch a container per process across each host.

Note

The *impi* containers must be launched with singularity run on HPC systems.

TACC uses a command called ibrun on all of its systems that configures MPI to use the high-speed, low-latency network, and binds processes to specific cores. If you are familiar with MPI, this is the functional equivalent to mpirun.

Take some time and try running the program with more samples. Just remember that each extra digit will increase the runtime by about 10-times the previous, so hit Ctrl-C to terminate something that’s taking too long.

Running via batch submission

To run a container via non-interactive batch job, the container should first be downloaded to a performant filesystem like $SCRATCH or $HOME.

$ idev -m 60 -p normal -N 1
$ cd $SCRATCH
$ module load tacc-singularity
$ singularity pull docker://USERNAME/pi-estimator:0.1-mpi
$ ls *sif
$ exit

After pulling the container, the image file can be referred to in an sbatch script. Please create pi-mpi.sbatch with the following text:

#!/bin/bash

#SBATCH -J calculate-pi-mpi              # Job name
#SBATCH -o calculate-pi-mpi.%j           # Name of stdout output file (%j expands to jobId)
#SBATCH -p development                   # Queue name
#SBATCH -N 1                             # Total number of nodes requested (56 cores/node)
#SBATCH -n 8                             # Total number of mpi tasks requested
#SBATCH -t 00:10:00                      # Run time (hh:mm:ss)
#SBATCH --reservation ContainerTraining  # a reservation only active during the training

module load tacc-singularity
cd $SCRATCH
ibrun singularity run pi-estimator_0.1-mpi.sif pi-mpi.py 10000000

Then, you can submit the job with sbatch

$ sbatch pi-mpi.sbatch

Check the status of your job with squeue

$ squeue -u USERNAME

When your job is done, the output will be in calculate-pi-mpi.[job number], and can be viewed with cat, less, or your favorite text editor.

Once done, try scaling up the program to two nodes (-N 2) and 16 tasks (-n 16) by changing your batch script or idev session. After that, try increasing the number of samples to see how accurate your estimate can get.

Note

If your batch job is running too long, you can finding the job number with squeue -u [username] and then terminate it with scancel [job number]

Singularity and GPU Computing

Singularity fully supports GPU utilization by exposing devices at runtime with the --nv flag. This is similar to nvidia-docker, so all docker containers with libraries that are compatible with the drivers on our systems can work as expected.

Base Docker images

When running at TACC, we have a set of curated Docker images that include TensorFlow and PyTorch for use in the FROM line of your own containers. You can see a list of available images at Docker Hub and the source at https://github.com/TACC/tacc-ml.

Image Frontera/rtx Maverick2 Longhorn Lonestar6
tacc/tacc-ml:centos7-cuda9-tf1.14-pt1.3   X    
tacc/tacc-ml:centos7-cuda10-tf1.15-pt1.3 X X    
tacc/tacc-ml:centos7-cuda10-tf2.4-pt1.7 X X    
tacc/tacc-ml:centos7-cuda11-tf2.6-pt1.10 X X   X
tacc/tacc-ml:ppc64le-centos7-cuda10-tf1.15-pt1.2     X  
tacc/tacc-ml:ppc64le-centos7-cuda10-tf2.1-pt1.3     X  
tacc/tacc-ml:ppc64le-centos7-cuda11-tf2.7-pt1.10     X  
tacc/tacc-ml:ubuntu16.04-cuda9-tf1.14-pt1.3   X    
tacc/tacc-ml:ubuntu16.04-cuda10-tf1.15-pt1.3 X X    
tacc/tacc-ml:ubuntu16.04-cuda10-tf2.4-pt1.7 X X    
tacc/tacc-ml:ubuntu20.04-cuda11-tf2.6-pt1.10 X X   X
tacc/tacc-ml:ppc64le-ubuntu16.04-cuda10-tf1.15-pt1.2     X  
tacc/tacc-ml:ppc64le-ubuntu16.04-cuda10-tf2.1-pt1.3     X  
tacc/tacc-ml:ppc64le-ubuntu20.04-cuda11-tf2.7-pt1.10     X  

For instance, the latest version of caffe can be used on TACC systems as follows:

Work from a compute node
$ idev -m 60 -p rtx

Load the singularity module
$ module load tacc-singularity

Pull your image
$ singularity pull docker://nvidia/caffe:latest

Test the GPU
$ singularity exec --nv caffe_latest.sif caffe device_query -gpu 0

Note

If this resulted in an error and the GPU was not detected, and you are on a GPU-enabled compute node, you may have excluded the --nv flag.

As previously mentioned, the main requirement for GPU-enabled containers to work is that the version of the host drivers matches the major version of the library inside the container. So, for example, if CUDA 10 is on the host, the container needs to use CUDA 10 internally.

For a more exciting test, the latest version of Tensorflow can be benchmarked as follows:

Change to your $SCRATCH directory
$ cd $SCRATCH

Download the benchmarking code
$ git clone https://github.com/tensorflow/benchmarks.git

Pull the image
$ singularity pull docker://tensorflow/tensorflow:latest-gpu

Run the code
$ singularity exec --nv tensorflow_latest-gpu.sif python \
    benchmarks/scripts/tf_cnn_benchmarks/tf_cnn_benchmarks.py \
    --num_gpus=4 --model resnet50 --batch_size 32 --num_batches 100

Try different numbers of gpus, batch sizes, and total batches to see how the parameters affect the benchmark.

Note

If the benchmark crashes the batch may be too large for GPU memory, or you requested more GPUs than exist on the system.

Building a GPU aware container

In the previous two examples, we have used pre-built containers to test GPU capability. Here we are going to build a GPU aware container to do some NLP/Text classification with the BERT transformer model using PyTorch. We are going to use one of the TACC base images (tacc/tacc-ml:centos7-cuda10-tf2.4-pt1.7) as a starting point.

On your local laptop, create a directory to build the “bert-classifier” Docker image and download (or copy and paste) the following files:

Note

For speed, you can also clone the repository from https://github.com/eriksf/bert-classifier.git.

Take a look at the files. bert_classifier.py is a Python script that uses PyTorch to do the text classification. The Dockerfile is based on the tacc/tacc-ml:centos7-cuda10-tf2.4-pt1.7 base image and installs a couple of needed python libraries in addition to moving the datasets into the image. train.csv, test.csv, and valid.csv are pre-processed CSV files containing the training, test, and validation datasets.

With these files downloaded, we can now build the image.

$ docker build -t USERNAME/bert-classifier:0.0.1 .

Note

Don’t forget to change USERNAME to your DockerHub username!

Once you have successfully built the image, push it up to DockerHub with the docker push command so that we can pull it back down on Frontera.

$ docker push USERNAME/bert-classifier:0.0.1

Testing the Container Locally with CPU

As was mentioned above, before using TACC allocation hours, it’s a good idea to test locally. In this case, we can at least test that the program help works.

$ docker run --rm -it eriksf/bert-classifier:0.0.1 bert_classifier.py -h
usage: bert_classifier.py [-h] [-d DEVICE] [-o OUTPUT] [-s SOURCE]

optional arguments:
  -h, --help            show this help message and exit
  -d DEVICE, --device DEVICE
                        The device to run on: cpu or cuda (DEFAULT: cuda)
  -o OUTPUT, --output OUTPUT
                        The output folder (DEFAULT: current directory)
  -s SOURCE, --source SOURCE
                        The location of the data files (DEFAULT: /code in
                        container)

We could even test the classification (very slowly) using the CPU.

$ docker run --rm -it eriksf/bert-classifier:0.0.1 bert_classifier.py -d cpu

Running the Container on Frontera

To start, lets allocate a single RTX node, which has 4 NVIDIA Quadro RTX 5000 GPUs with 16 GB of Memory each.

$ idev -m 60 -p rtx

Once you have your node, pull the container and run it as follows:

Load singularity module
$ module load tacc-singularity

Change to $SCRATCH directory
$ cd $SCRATCH

Pull container
$ singularity pull docker://USERNAME/bert-classifier:0.0.1

Run container
$ singularity exec --nv bert-classifier_0.0.1.sif bert_classifier.py