ComScribe is a tool that identifies communication among all GPU-GPU and CPU-GPU pairs in a single-node multi-GPU system.
You can directly execute install.sh script
./install.sh
OR
You can install it manually.
You will need the following programs:
- Python: ComScribe is a Python script. It uses several packages listed in
requirements.txt
, which you can install via the command:
pip3 install -r requirements.txt
-
nvprof: ComScribe parses the outputs of NVIDIA's profiler nvprof, which is a light-weight command-line profiler available since CUDA 5.
-
NCCL[Optional]. ComScribe modifies NCCL library to profile collective communication primitives. If your application does not use any collective operations, you don't have to perform this step.
cd nccl && make -j src.build
No further installation is required.
To obtain the communication matrices of your application (app
):
python3 comscribe.py -g <num_gpus> -s log|linear -i <cmd_to_run>
-g
lets our tool know how many GPUs will be used, however note that if the application to be run requires such a parameter too, it must be explicitly specified (see -i
below).
-s
can be log
for log scale or linear
for linear scale for the output figures.
-i
takes the input command as a string such as: -i './app --foo 20 --bar 5'
The communication matrix for a communication type is only generated if it is detected, e.g. if there are no Unified Memory transfers then there will not be any output regarding Unified Memory transfers. For the types of communication detected, the generated figures are saved as PDF files in the directory of the script.
python3 comscribe.py -n -c <collective_type> -g <num_gpus> -s log|linear -i <cmd_to_run>
-n
enables the profiling of collectives.
-c
represents the collective to be profiled (if not specified, all collectives will be profiled by default). Options: broadcast, reduce, allgather, allreduce, reducescatter
We have used our tool in an NVIDIA V100 DGX2 system with up to 16 GPUs using CUDA v10.0.130 for the following benchmarks:
-
NVIDIA Monte Carlo Simluation of 2D Ising-GPU | GitHub
-
NVIDIA Multi-GPU Jacobi Solver | GitHub
-
- Full-Duplex | GitHub
- Full-Duplex with Unified Memory | GitHub
- Half-Duplex with peer access | GitHub
- Half-Duplex without peer access | GitHub
- Zero-copy Memory (both Read and Write benchmarks) | GitHub
Note: In order to run a Comm|Scope benchmark with fixed iterations e.g. 100, in the source code of benchmark, replace it's registration with:
benchmark::RegisterBenchmark(...)->SMALL_ARGS()->Iterations(100);
-
MGBench | Github
python3 comscribe.py -g 4 -i './scope --benchmark_filter="Comm_ZeroCopy_GPUToGPU_Read.*18.*" -n 0' -s log
Gives the bar-chart for Zero-copy memory transfers:
python3 comscribe.py -g 4 -i './scope --benchmark_filter="Comm_Demand_Duplex_GPUGPU.*18.*"' -s linear
Gives two matrices, bytes transferred (left) and number of transfers made (right):
python3 comscribe.py -g 4 -i './fullduplex' -s linear
Gives two matrices, bytes transferred (left) and number of transfers made (right):
python3 comscribe.py -g 4 -i './cuIsing -y 32768 -x 65536 -n 128 -p 16 -d 4 -t 1.5' -s log
Gives two matrices, bytes transferred (left) and number of transfers made (right):
-
Akhtar, P., Tezcan, E., Qararyah, F.M., Unat, D. (2021). ComScribe: Identifying Intra-node GPU Communication. In: Wolf, F., Gao, W. (eds) Benchmarking, Measuring, and Optimizing. Bench 2020. Lecture Notes in Computer Science(), vol 12614. Springer, Cham. https://doi.org/10.1007/978-3-030-71058-3_10
-
Soytürk, M.A., Akhtar, P., Tezcan, E., Unat, D. (2022). Monitoring Collective Communication Among GPUs. In: , et al. Euro-Par 2021: Parallel Processing Workshops. Euro-Par 2021. Lecture Notes in Computer Science, vol 13098. Springer, Cham. https://doi.org/10.1007/978-3-031-06156-1_4