Craig tierney
2016-10-27 22:23:19 UTC
Hello,
I am trying to build OpenMPI 1.10.3 with CUDA but I am unable to build the
library that will allow me to use IPC on a node or GDR between nodes. I
have tried with with 1.10.4 and 2.0.1 and have the same problems. Here is
my build script:
---------------------------
#!/bin/bash
export OPENMPI_VERSION=1.10.3
export BASEDIR=/tmp/mpi_testing/
export CUDA_HOME=/usr/local/cuda
export PATH=$CUDA_HOME/bin/:$PATH
export LD_LIBRARY_PATH=$CUDA_HOME/lib64:$LD_LIBRARY_PATH
export MPI_HOME=$BASEDIR/openmpi-$OPENMPI_VERSION
which nvcc
nvcc --version
tar -zxf openmpi-$OPENMPI_VERSION.tar.gz
cd openmpi-$OPENMPI_VERSION
./configure --prefix=$MPI_HOME --with-cuda=$CUDA_HOME/include > config.out
2>&1
make -j > build.out 2>&1
make install >> build.out 2>&1
-----------------------
From the docs, it appears that I should not have to set anything but
--with-cuda since my CUDA is in /usr/local/cuda. However, I appended
/usr/local/cuda/include just in case when the first way didn't work.
From the output in config.log, I see that cuda.h is not found. When the
tests are called there is no extra include flag added to specify the
/usr/local/cuda/include path.
With the resulting build, I test for CUDA and GDR with ompi_info. Results
are:
***@dgx-1:~/temp$ /tmp/mpi_testing/openmpi-1.10.3/bin/ompi_info |
grep cuda
MCA btl: smcuda (MCA v2.0.0, API v2.0.0, Component v1.10.3)
MCA coll: cuda (MCA v2.0.0, API v2.0.0, Component v1.10.3)
***@dgx-1:~/temp$ /tmp/mpi_testing/openmpi-1.10.3/bin/ompi_info |
grep gdr
***@dgx-1:~/temp$
Configure and build logs are attached.
Thanks,
Craig
I am trying to build OpenMPI 1.10.3 with CUDA but I am unable to build the
library that will allow me to use IPC on a node or GDR between nodes. I
have tried with with 1.10.4 and 2.0.1 and have the same problems. Here is
my build script:
---------------------------
#!/bin/bash
export OPENMPI_VERSION=1.10.3
export BASEDIR=/tmp/mpi_testing/
export CUDA_HOME=/usr/local/cuda
export PATH=$CUDA_HOME/bin/:$PATH
export LD_LIBRARY_PATH=$CUDA_HOME/lib64:$LD_LIBRARY_PATH
export MPI_HOME=$BASEDIR/openmpi-$OPENMPI_VERSION
which nvcc
nvcc --version
tar -zxf openmpi-$OPENMPI_VERSION.tar.gz
cd openmpi-$OPENMPI_VERSION
./configure --prefix=$MPI_HOME --with-cuda=$CUDA_HOME/include > config.out
2>&1
make -j > build.out 2>&1
make install >> build.out 2>&1
-----------------------
From the docs, it appears that I should not have to set anything but
--with-cuda since my CUDA is in /usr/local/cuda. However, I appended
/usr/local/cuda/include just in case when the first way didn't work.
From the output in config.log, I see that cuda.h is not found. When the
tests are called there is no extra include flag added to specify the
/usr/local/cuda/include path.
With the resulting build, I test for CUDA and GDR with ompi_info. Results
are:
***@dgx-1:~/temp$ /tmp/mpi_testing/openmpi-1.10.3/bin/ompi_info |
grep cuda
MCA btl: smcuda (MCA v2.0.0, API v2.0.0, Component v1.10.3)
MCA coll: cuda (MCA v2.0.0, API v2.0.0, Component v1.10.3)
***@dgx-1:~/temp$ /tmp/mpi_testing/openmpi-1.10.3/bin/ompi_info |
grep gdr
***@dgx-1:~/temp$
Configure and build logs are attached.
Thanks,
Craig