ppkMHD stands for Performance Portable Kokkos for Magneto-HydroDynamics (MHD) solvers.
Here a small list of numerical schemes implementations:
- second order MUSCL-HANCOCK scheme for hydro and MHD
- high-order MOOD (hydro only)
- high-order Spectral Difference Method schemes: hydro only
All scheme are available in 2D and 3D using Kokkos+MPI implementation.
- Kokkos library will be built by ppkMHD using the same flags (architecture, optimization, ...).
- CMake with version >= 3.X (3.X is chosen to meet Kokkos own requirement for CMake; i.e. it might increase in the future)
Current application is configured with kokkos library as a git submodule. So you'll need to run the following git commands right after cloning ppkMHD:
git submodule init
git submodule update
Kokkos is built with the same flags as the main application.
A few example builds, with minimal configuration options.
- Create a build directory, configure and make
mkdir build; cd build
cmake -DUSE_MPI=OFF -DKokkos_ENABLE_OPENMP=ON -DKokkos_ENABLE_HWLOC=ON ..
make -j 4
Add variable CXX on the cmake command line to change the compiler (clang++, icpc, pgcc, ....).
- Create a build directory, configure and make
export CXX=icpc
mkdir build; cd build
cmake -DUSE_MPI=OFF -DKokkos_ARCH_KNL=ON -DKokkos_ENABLE_OPENMP=ON ..
make -j 4
To be able to build with CUDA backend, you need to use nvcc_wrapper located in kokkos source (external/kokkos/bin/nvcc_wrapper).
- Create a build directory, configure and make
mkdir build; cd build
export CXX=/path/to/nvcc_wrapper
cmake -DUSE_MPI=OFF -DKokkos_ENABLE_CUDA=ON -DKokkos_ARCH_MAXWELL50=ON ..
make -j 4
nvcc_wrapper
is a compiler wrapper arroud NVIDIA nvcc
. It is available from Kokkos sources: external/kokkos/bin/nvcc_wrapper
. Any Kokkos application target NVIDIA GPUs must be built with nvcc_wrapper
.
Please make sure to use a CUDA-aware MPI implementation (OpenMPI or MVAPICH2) built with the proper flags for activating CUDA support.
It may happen that eventhough your MPI implementation is actually cuda-aware, cmake find_package macro for MPI does not detect it to be cuda aware. In that case, you can enforce cuda awareness by turning option USE_MPI_CUDA_AWARE_ENFORCED to ON.
You don't need to use mpi compiler wrapper mpicxx, cmake should be able to correctly populate MPI_CXX_INCLUDE_PATH, MPI_CXX_LIBRARIES which are passed to all final targets.
- Create a build directory, configure and make
mkdir build; cd build
export CXX=/path/to/nvcc_wrapper
cmake -DUSE_MPI=ON -DKokkos_ENABLE_CUDA=ON -DKokkos_ARCH_MAXWELL50=ON ..
make -j 4
Example command line to run the application (1 GPU used per MPI task)
mpirun -np 4 ./ppkMHD ./test_implode_2D_mpi.ini
In order to activate building SDM schemes, use Cmake option -DUSE_SDM=ON
The MOOD numerical scheme require some linear algebra (QR decomposition) on the host (not device). This is done using a Blas/Lapack implementation using the C language interface named Lapacke.
Please note that Atlas doesn't provide Lapackage. Currently (March 2017), on Ubuntu 16.04, package libatlas-dev is not compatible with package Lapacke (generate errors at link time). So please either Netlib or OpenBLAS implementation.
If you want to enforce the use of OpenBLAS, just use a recent CMake (>=3.6) and add "-DBLA_VENDOR" on the cmake command line. This will tell the cmake system (through the call to find_package(BLAS) ) to only look for OpenBLAS implementation.
On a recent Ubuntu, if atlas is not installed, but OpenBLAS is, you don't need to have a bleeding edge CMake, current cmake will find OpenBLAS.
Make sure to have CMake variable CMAKE_EXPORT_COMPILE_COMMANDS set to ON, it will generate a file named compile_commands.json. Then you can symlink the generated file in the top level source directory.
Please visit :