[mpich-discuss] [mvapich-discuss] Announcing the Release of MVAPICH2 1.8RC1 and OSU Micro-Benchmarks (OMB) 3.5.2
Dhabaleswar Panda
panda at cse.ohio-state.edu
Thu Mar 22 14:05:57 CDT 2012
These releases might be of interest to some of the MPICH users. Thus, I am
posting it here.
Thanks,
DK
---------- Forwarded message ----------
Date: Thu, 22 Mar 2012 13:21:26 -0400 (EDT)
From: Dhabaleswar Panda <panda at cse.ohio-state.edu>
To: mvapich-discuss at cse.ohio-state.edu
Cc: Dhabaleswar Panda <panda at cse.ohio-state.edu>
Subject: [mvapich-discuss] Announcing the Release of MVAPICH2 1.8RC1 and
OSU Micro-Benchmarks (OMB) 3.5.2
The MVAPICH team is pleased to announce the release of MVAPICH2 1.8RC1
and OSU Micro-Benchmarks (OMB) 3.5.2.
Features, Enhancements, and Bug Fixes for MVAPICH2 1.8RC1 are listed
here.
* New Features and Enhancements (since 1.8a2):
- New design for intra-node communication from GPU Device buffers
using CUDA IPC for better performance and correctness
- Thanks to Joel Scherpelz from NVIDIA for his suggestions
- Enabled shared memory communication for host transfers when CUDA is
enabled
- Optimized and tuned collectives for GPU device buffers
- Enhanced pipelined inter-node device transfers
- Enhanced shared memory design for GPU device transfers for
large messages
- Enhanced support for CPU binding with socket and numanode level
granularity
- Support suspend/resume functionality with mpirun_rsh
- Exporting local rank, local size, global rank and global size
through environment variables (both mpirun_rsh and hydra)
- Update to hwloc v1.4
- Checkpoint-Restart support in OFA-IB-Nemesis interface
- Enabling run-through stabilization support to handle process
failures in OFA-IB-Nemesis interface
- Enhancing OFA-IB-Nemesis interface to handle IB errors gracefully
- Performance tuning on various platforms
- Support for Mellanox IB FDR adapter
* Bug Fixes (since 1.8a2):
- Fix a hang issue on InfiniHost SDR/DDR cards
- Thanks to Nirmal Seenu from Fermilab for the report
- Fix an issue with runtime parameter MV2_USE_COALESCE usage
- Fix an issue with LiMIC2 when CUDA is enabled
- Fix an issue with intra-node communication using datatypes and GPU
device buffers
- Fix an issue with Dynamic Process Management when launching
processes on multiple nodes
- Thanks to Rutger Hofman from VU Amsterdam for the report
- Fix build issue in hwloc source with mcmodel=medium flags
- Thanks to Nirmal Seenu from Fermilab for the report
- Fix a build issue in hwloc with --disable-shared or
--disabled-static options
- Use portable stdout and stderr redirection
- Thanks to Dr. Axel Philipp from MTU Aero Engines for the patch
- Fix a build issue with PGI 12.2
- Thanks to Thomas Rothrock from U.S. Army SMDC for the patch
- Fix an issue with send message queue in OFA-IB-Nemesis interface
- Fix a process cleanup issue in Hydra when MPI_ABORT is called
(upstream MPICH2 patch)
- Fix an issue with non-contiguous datatypes in MPI_Gather
- Fix a few memory leaks and warnings
Bugfixes for OSU Micro-Benchmarks (OMB) 3.5.2 is listed here.
* Bug Fix (since OMB 3.5.1):
- Fix typo which led to use of incorrect buffers
The complete set of features and enhancements for MVAPICH2 1.8RC1 compared
to MVAPICH2 1.7 are as follows:
* Features & Enhancements:
- Support for MPI communication from NVIDIA GPU device memory
- High performance RDMA-based inter-node point-to-point
communication (GPU-GPU, GPU-Host and Host-GPU)
- High performance intra-node point-to-point communication for
multi-GPU adapters/node (GPU-GPU, GPU-Host and Host-GPU)
- Taking advantage of CUDA IPC (available in CUDA 4.1) in
intra-node communication for multiple GPU adapters/node
- Optimized and tuned collectives for GPU device buffers
- MPI datatype support for point-to-point and collective
communication from GPU device buffers
- Support suspend/resume functionality with mpirun_rsh
- Enhanced support for CPU binding with socket and numanode level
granularity
- Exporting local rank, local size, global rank and global size
through environment variables (both mpirun_rsh and hydra)
- Update to hwloc v1.4
- Checkpoint-Restart support in OFA-IB-Nemesis interface
- Enabling run-through stabilization support to handle process
failures in OFA-IB-Nemesis interface
- Enhancing OFA-IB-Nemesis interface to handle IB errors gracefully
- Performance tuning on various architecture clusters
- Support for Mellanox IB FDR adapter
- Adjust shared-memory communication block size at runtime
- Enable XRC by default at configure time
- New shared memory design for enhanced intra-node small message
performance
- Tuned inter-node and intra-node performance on different cluster
architectures
- Support for fallback to R3 rendezvous protocol if RGET fails
- SLURM integration with mpiexec.mpirun_rsh to use SLURM allocated
hosts without specifying a hostfile
- Support added to automatically use PBS_NODEFILE in Torque and PBS
environments
- Enable signal-triggered (SIGUSR2) migration
- Reduced memory footprint of the library
- Enhanced one-sided communication design with reduced memory
requirement
- Enhancements and tuned collectives (Bcast and Alltoallv)
- Flexible HCA selection with Nemesis interface
- Thanks to Grigori Inozemtsev, Queens University
- Support iWARP interoperability between Intel NE020 and
Chelsio T4 Adapters
- RoCE enable environment variable name is changed from MV2_USE_RDMAOE
to MV2_USE_RoCE
Sample performance numbers for MPI communication from NVIDIA GPU memory
using MVAPICH2 1.8RC1 and OMB 3.5.2 can be obtained from the following
URL:
http://mvapich.cse.ohio-state.edu/performance/gpu.shtml
For downloading MVAPICH2 1.8RC1, OMB 3.5.2, associated user guide, quick
start guide, and accessing the SVN, please visit the following URL:
http://mvapich.cse.ohio-state.edu
All questions, feedbacks, bug reports, hints for performance tuning,
patches and enhancements are welcome. Please post it to the
mvapich-discuss mailing list (mvapich-discuss at cse.ohio-state.edu).
We are also happy to inform that the number of downloads from MVAPICH
project site has crossed 100,000. The MVAPICH team extends thanks to all
MVAPICH/MVAPICH2 users and their organizations.
Thanks,
The MVAPICH Team
_______________________________________________
mvapich-discuss mailing list
mvapich-discuss at cse.ohio-state.edu
http://mail.cse.ohio-state.edu/mailman/listinfo/mvapich-discuss
More information about the mpich-discuss
mailing list