[mpich-discuss] Regarding MPICH2-1.1.1p1 testing

李俊丽 limu713 at gmail.com
Fri Mar 19 03:18:26 CDT 2010


*Hello,*

   *When I am trying to test mpich2 by "*mpiexec -n 4  mpitests-IMB-EXT
   Bidir_Get" on rhel5.4 server,
   It has this error message:

   *# Bidir_Get *

   *#--------------------------------------------------- *

   *# **Benchmarking Bidir_Get *

   *# #processes = 2 *

   *# ( 2 additional processes waiting in MPI_Barrier) *

   *#--------------------------------------------------- *

   *# *

   *# MODE: AGGREGATE *

   *# *

   *#bytes #repetitions t[usec] Mbytes/sec *

   *0 1000 0.29 0.00 *

   *4 1000 9.01 0.42 *

   *8 1000 8.51 0.90 *

   *16 1000 10.30 1.48 *

   *32 1000 11.12 2.75 *

   *64 1000 16.89 3.61 *

   *128 1000 17.74 6.88 *

   *256 1000 33.94 7.19 *

   *512 1000 61.93 7.88 *

   *Fatal error in MPI_Win_fence: Other MPI error, error stack: *

   *MPI_Win_fence(123)............: MPI_Win_fence(assert=0, win=0xa0000000)
   failed *

   *MPIDI_Win_fence(272)..........: Detected an error while in progress wait
   for RMA messages *

   *MPIDI_CH3I_Progress(150)......: *

   *MPID_nem_mpich2_test_recv(905): *

   *MPID_nem_tcp_connpoll(1663)...: Communication error *

   *Fatal error in MPI_Win_fence: Other MPI error, error stack: *

   *MPI_Win_fence(123)............: MPI_Win_fence(assert=0, win=0xa0000000)
   failed *

   *MPIDI_Win_fence(272)..........: Detected an error while in progress wait
   for RMA messages *

   *MPIDI_CH3I_Progress(150)......: *

   *MPID_nem_mpich2_test_recv(905): *

   *MPID_nem_tcp_connpoll(1663)...: Communication error *

   *Fatal error in PMPI_Barrier: Other MPI error, error stack: *

   *PMPI_Barrier(476)..............: MPI_Barrier(MPI_COMM_WORLD) failed *

   *MPIR_Barrier(82)...............: *

   *MPIC_Sendrecv(164).............: *

   *MPIC_Wait(405).................: *

   *MPIDI_CH3I_Progress(150).......: *

   *MPID_nem_mpich2_test_recv(905).: *

   *MPID_nem_tcp_connpoll(1667)....: *

   *state_commrdy_handler(1517)....: *

   *MPID_nem_tcp_recv_handler(1413): socket closed *

   *Fatal error in PMPI_Barrier: Other MPI error, error stack: *

   *PMPI_Barrier(476)..............: MPI_Barrier(MPI_COMM_WORLD)failed *

   *MPIR_Barrier(82)...............: *

   *MPIC_Sendrecv(164).............: *

   *MPIC_Wait(405).................: *

   *MPIDI_CH3I_Progress(150).......: *

   *MPID_nem_mpich2_test_recv(905).: *

   *MPID_nem_tcp_connpoll(1667)....: *

   *state_commrdy_handler(1517)....: *

   *MPID_nem_tcp_recv_handler(1413): socket closed *

   *rank 1 in job 3 cu02.hpc.com_53583 caused collective abortof all ranks *

   *exit status of rank 1: return code 1 *

   *rank 0 in job 3 cu02.hpc.com_53583 caused collective abort of all ranks
   *

   *exit status of rank 0: return code 1
   *


   *Is there a fix for this?*


   *Thanks*

   *Lily
   *
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/mpich-discuss/attachments/20100319/dde00573/attachment-0001.htm>


More information about the mpich-discuss mailing list