[mpich-discuss] Announcing the Release of MVAPICH2 1.8a1 and OSU Micro-Benchmarks (OMB) 3.5
Dhabaleswar Panda
panda at cse.ohio-state.edu
Wed Nov 9 23:20:20 CST 2011
These releases might be of interest to some of the MPICH users. Thus, I
am posting it here.
Thanks,
DK
---------- Forwarded message ----------
Date: Wed, 9 Nov 2011 22:36:27 -0500 (EST)
From: Dhabaleswar Panda <panda at cse.ohio-state.edu>
To: mvapich-discuss at cse.ohio-state.edu
Cc: Dhabaleswar Panda <panda at cse.ohio-state.edu>
Subject: [mvapich-discuss] Announcing the Release of MVAPICH2 1.8a1 and OSU
Micro-Benchmarks (OMB) 3.5
The MVAPICH team is pleased to announce the release of MVAPICH2 1.8a1 and
OSU Micro-Benchmarks (OMB) 3.5.
The MVAPICH2 1.8a1 release is targeted for MVAPICH2 users to harness
performance on InfiniBand (Mellanox) clusters with NVIDIA GPU adapters and
CUDA support. The OMB 3.5 release is targeted for MPI users to carry out
benchmarking and performance evaluation of MPI stacks on clusters with
NVIDIA GPU adapters and CUDA support.
Features, Enhancements, and Bug Fixes for MVAPICH2 1.8a1 (since MVAPICH2
1.7GA release) are listed here.
* New Features and Enhancements (since 1.7GA):
- Support for MPI communication from NVIDIA GPU device memory
- High performance RDMA-based inter-node point-to-point
communication (GPU-GPU, GPU-Host and Host-GPU)
- High performance intra-node point-to-point communication
for multi-GPU adapters/node (GPU-GPU, GPU-Host and Host-GPU)
- Communication with contiguous datatype
- Reduced memory footprint of the library
- Enhanced one-sided communication design with reduced memory
requirement
- Enhancements and tuned collectives (Bcast and Alltoallv)
- Update to hwloc v1.3.0
- Flexible HCA selection with Nemesis interface
(Thanks to Grigori Inozemtsev, Queens University)
- Support iWARP interoperability between Intel NE020 and
Chelsio T4 Adapters
- RoCE enable environment variable name is changed from
MV2_USE-RDMAOE to MV2_USE_RoCE
* Bug Fixes (since 1.7GA):
- Fix for a bug in mpirun_rsh while doing process clean-up
in abort and other error scenarios
- Fixes for code compilation warnings
- Fix for memory leaks in RDMA CM code path
Successive 1.8-RC releases with multiple new features (specific to
InfiniBand, iWARP and RoCE) will be available in the near future.
New features and Enhancements of OSU Micro-Benchmarks (OMB) 3.5 (since OMB
3.4 release) are listed here.
- Extension of osu_latency, osu_bw, and osu_bibw benchmarks to
evaluate the performance of MPI_Send/MPI_Recv operation with
NVIDIA GPU device and CUDA support
- This functionality is exposed when configured
with --enable-cuda option
- Flexibility for using buffers in NVIDIA GPU device (D)
and host memory (H)
- Flexibility for selecting data movement between D->D,
D->H and H->D
Sample performance numbers for MPI communication from NVIDIA GPU memory
using MVAPICH2 1.8a1 and OMB 3.5 can be obtained from the following URL:
http://mvapich.cse.ohio-state.edu/performance/gpu.shtml
For downloading MVAPICH2 1.8a1, OMB 3.5, associated user guide, quick
start guide, and accessing the SVN, please visit the following URL:
http://mvapich.cse.ohio-state.edu
All questions, feedbacks, bug reports, hints for performance tuning,
patches and enhancements are welcome. Please post it to the
mvapich-discuss mailing list (mvapich-discuss at cse.ohio-state.edu).
We are also happy to inform that the number of organizations using
MVAPICH/MVAPICH2 (and registered at the MVAPICH site) has crossed 1,800
world-wide (in 65 countries). The MVAPICH team extends thanks to all these
organizations.
Thanks,
The MVAPICH Team
More information about the mpich-discuss
mailing list