[mpich-discuss] Regarding MPICH2-1.1.1p1 testing
    李俊丽 
    limu713 at gmail.com
       
    Fri Mar 19 03:18:26 CDT 2010
    
    
  
*Hello,*
   *When I am trying to test mpich2 by "*mpiexec -n 4  mpitests-IMB-EXT
   Bidir_Get" on rhel5.4 server,
   It has this error message:
   *# Bidir_Get *
   *#--------------------------------------------------- *
   *# **Benchmarking Bidir_Get *
   *# #processes = 2 *
   *# ( 2 additional processes waiting in MPI_Barrier) *
   *#--------------------------------------------------- *
   *# *
   *# MODE: AGGREGATE *
   *# *
   *#bytes #repetitions t[usec] Mbytes/sec *
   *0 1000 0.29 0.00 *
   *4 1000 9.01 0.42 *
   *8 1000 8.51 0.90 *
   *16 1000 10.30 1.48 *
   *32 1000 11.12 2.75 *
   *64 1000 16.89 3.61 *
   *128 1000 17.74 6.88 *
   *256 1000 33.94 7.19 *
   *512 1000 61.93 7.88 *
   *Fatal error in MPI_Win_fence: Other MPI error, error stack: *
   *MPI_Win_fence(123)............: MPI_Win_fence(assert=0, win=0xa0000000)
   failed *
   *MPIDI_Win_fence(272)..........: Detected an error while in progress wait
   for RMA messages *
   *MPIDI_CH3I_Progress(150)......: *
   *MPID_nem_mpich2_test_recv(905): *
   *MPID_nem_tcp_connpoll(1663)...: Communication error *
   *Fatal error in MPI_Win_fence: Other MPI error, error stack: *
   *MPI_Win_fence(123)............: MPI_Win_fence(assert=0, win=0xa0000000)
   failed *
   *MPIDI_Win_fence(272)..........: Detected an error while in progress wait
   for RMA messages *
   *MPIDI_CH3I_Progress(150)......: *
   *MPID_nem_mpich2_test_recv(905): *
   *MPID_nem_tcp_connpoll(1663)...: Communication error *
   *Fatal error in PMPI_Barrier: Other MPI error, error stack: *
   *PMPI_Barrier(476)..............: MPI_Barrier(MPI_COMM_WORLD) failed *
   *MPIR_Barrier(82)...............: *
   *MPIC_Sendrecv(164).............: *
   *MPIC_Wait(405).................: *
   *MPIDI_CH3I_Progress(150).......: *
   *MPID_nem_mpich2_test_recv(905).: *
   *MPID_nem_tcp_connpoll(1667)....: *
   *state_commrdy_handler(1517)....: *
   *MPID_nem_tcp_recv_handler(1413): socket closed *
   *Fatal error in PMPI_Barrier: Other MPI error, error stack: *
   *PMPI_Barrier(476)..............: MPI_Barrier(MPI_COMM_WORLD)failed *
   *MPIR_Barrier(82)...............: *
   *MPIC_Sendrecv(164).............: *
   *MPIC_Wait(405).................: *
   *MPIDI_CH3I_Progress(150).......: *
   *MPID_nem_mpich2_test_recv(905).: *
   *MPID_nem_tcp_connpoll(1667)....: *
   *state_commrdy_handler(1517)....: *
   *MPID_nem_tcp_recv_handler(1413): socket closed *
   *rank 1 in job 3 cu02.hpc.com_53583 caused collective abortof all ranks *
   *exit status of rank 1: return code 1 *
   *rank 0 in job 3 cu02.hpc.com_53583 caused collective abort of all ranks
   *
   *exit status of rank 0: return code 1
   *
   *Is there a fix for this?*
   *Thanks*
   *Lily
   *
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/mpich-discuss/attachments/20100319/dde00573/attachment-0001.htm>
    
    
More information about the mpich-discuss
mailing list