[mpich-discuss] Announcing the Release of MVAPICH2 1.8a1 and OSU Micro-Benchmarks (OMB) 3.5

Dhabaleswar Panda panda at cse.ohio-state.edu
Wed Nov 9 23:20:20 CST 2011


These releases might be of interest to some of the MPICH users. Thus, I
am posting it here.

Thanks,

DK

---------- Forwarded message ----------
Date: Wed, 9 Nov 2011 22:36:27 -0500 (EST)
From: Dhabaleswar Panda <panda at cse.ohio-state.edu>
To: mvapich-discuss at cse.ohio-state.edu
Cc: Dhabaleswar Panda <panda at cse.ohio-state.edu>
Subject: [mvapich-discuss] Announcing the Release of MVAPICH2 1.8a1 and OSU
    Micro-Benchmarks (OMB) 3.5

The MVAPICH team is pleased to announce the release of MVAPICH2 1.8a1 and
OSU Micro-Benchmarks (OMB) 3.5.

The MVAPICH2 1.8a1 release is targeted for MVAPICH2 users to harness
performance on InfiniBand (Mellanox) clusters with NVIDIA GPU adapters and
CUDA support. The OMB 3.5 release is targeted for MPI users to carry out
benchmarking and performance evaluation of MPI stacks on clusters with
NVIDIA GPU adapters and CUDA support.

Features, Enhancements, and Bug Fixes for MVAPICH2 1.8a1 (since MVAPICH2
1.7GA release) are listed here.

* New Features and Enhancements (since 1.7GA):

  - Support for MPI communication from NVIDIA GPU device memory
     - High performance RDMA-based inter-node point-to-point
       communication (GPU-GPU, GPU-Host and Host-GPU)
     - High performance intra-node point-to-point communication
       for multi-GPU adapters/node (GPU-GPU, GPU-Host and Host-GPU)
     - Communication with contiguous datatype
  - Reduced memory footprint of the library
  - Enhanced one-sided communication design with reduced memory
    requirement
  - Enhancements and tuned collectives (Bcast and Alltoallv)
  - Update to hwloc v1.3.0
  - Flexible HCA selection with Nemesis interface
    (Thanks to Grigori Inozemtsev, Queens University)
  - Support iWARP interoperability between Intel NE020 and
    Chelsio T4 Adapters
  - RoCE enable environment variable name is changed from
    MV2_USE-RDMAOE to MV2_USE_RoCE

* Bug Fixes (since 1.7GA):

  - Fix for a bug in mpirun_rsh while doing process clean-up
    in abort and other error scenarios
  - Fixes for code compilation warnings
  - Fix for memory leaks in RDMA CM code path

Successive 1.8-RC releases with multiple new features (specific to
InfiniBand, iWARP and RoCE) will be available in the near future.

New features and Enhancements of OSU Micro-Benchmarks (OMB) 3.5 (since OMB
3.4 release) are listed here.

  - Extension of osu_latency, osu_bw, and osu_bibw benchmarks to
    evaluate the performance of MPI_Send/MPI_Recv operation with
    NVIDIA GPU device and CUDA support
        - This functionality is exposed when configured
          with --enable-cuda option
  - Flexibility for using buffers in NVIDIA GPU device (D)
    and host memory (H)
  - Flexibility for selecting data movement between D->D,
    D->H and H->D

Sample performance numbers for MPI communication from NVIDIA GPU memory
using MVAPICH2 1.8a1 and OMB 3.5 can be obtained from the following URL:

http://mvapich.cse.ohio-state.edu/performance/gpu.shtml

For downloading MVAPICH2 1.8a1, OMB 3.5, associated user guide, quick
start guide, and accessing the SVN, please visit the following URL:

http://mvapich.cse.ohio-state.edu

All questions, feedbacks, bug reports, hints for performance tuning,
patches and enhancements are welcome. Please post it to the
mvapich-discuss mailing list (mvapich-discuss at cse.ohio-state.edu).

We are also happy to inform that the number of organizations using
MVAPICH/MVAPICH2 (and registered at the MVAPICH site) has crossed 1,800
world-wide (in 65 countries). The MVAPICH team extends thanks to all these
organizations.

Thanks,

The MVAPICH Team




More information about the mpich-discuss mailing list