[mpich-discuss] Regarding MPICH2-1.1.1p1 testing
李俊丽
limu713 at gmail.com
Fri Mar 19 03:18:26 CDT 2010
*Hello,*
*When I am trying to test mpich2 by "*mpiexec -n 4 mpitests-IMB-EXT
Bidir_Get" on rhel5.4 server,
It has this error message:
*# Bidir_Get *
*#--------------------------------------------------- *
*# **Benchmarking Bidir_Get *
*# #processes = 2 *
*# ( 2 additional processes waiting in MPI_Barrier) *
*#--------------------------------------------------- *
*# *
*# MODE: AGGREGATE *
*# *
*#bytes #repetitions t[usec] Mbytes/sec *
*0 1000 0.29 0.00 *
*4 1000 9.01 0.42 *
*8 1000 8.51 0.90 *
*16 1000 10.30 1.48 *
*32 1000 11.12 2.75 *
*64 1000 16.89 3.61 *
*128 1000 17.74 6.88 *
*256 1000 33.94 7.19 *
*512 1000 61.93 7.88 *
*Fatal error in MPI_Win_fence: Other MPI error, error stack: *
*MPI_Win_fence(123)............: MPI_Win_fence(assert=0, win=0xa0000000)
failed *
*MPIDI_Win_fence(272)..........: Detected an error while in progress wait
for RMA messages *
*MPIDI_CH3I_Progress(150)......: *
*MPID_nem_mpich2_test_recv(905): *
*MPID_nem_tcp_connpoll(1663)...: Communication error *
*Fatal error in MPI_Win_fence: Other MPI error, error stack: *
*MPI_Win_fence(123)............: MPI_Win_fence(assert=0, win=0xa0000000)
failed *
*MPIDI_Win_fence(272)..........: Detected an error while in progress wait
for RMA messages *
*MPIDI_CH3I_Progress(150)......: *
*MPID_nem_mpich2_test_recv(905): *
*MPID_nem_tcp_connpoll(1663)...: Communication error *
*Fatal error in PMPI_Barrier: Other MPI error, error stack: *
*PMPI_Barrier(476)..............: MPI_Barrier(MPI_COMM_WORLD) failed *
*MPIR_Barrier(82)...............: *
*MPIC_Sendrecv(164).............: *
*MPIC_Wait(405).................: *
*MPIDI_CH3I_Progress(150).......: *
*MPID_nem_mpich2_test_recv(905).: *
*MPID_nem_tcp_connpoll(1667)....: *
*state_commrdy_handler(1517)....: *
*MPID_nem_tcp_recv_handler(1413): socket closed *
*Fatal error in PMPI_Barrier: Other MPI error, error stack: *
*PMPI_Barrier(476)..............: MPI_Barrier(MPI_COMM_WORLD)failed *
*MPIR_Barrier(82)...............: *
*MPIC_Sendrecv(164).............: *
*MPIC_Wait(405).................: *
*MPIDI_CH3I_Progress(150).......: *
*MPID_nem_mpich2_test_recv(905).: *
*MPID_nem_tcp_connpoll(1667)....: *
*state_commrdy_handler(1517)....: *
*MPID_nem_tcp_recv_handler(1413): socket closed *
*rank 1 in job 3 cu02.hpc.com_53583 caused collective abortof all ranks *
*exit status of rank 1: return code 1 *
*rank 0 in job 3 cu02.hpc.com_53583 caused collective abort of all ranks
*
*exit status of rank 0: return code 1
*
*Is there a fix for this?*
*Thanks*
*Lily
*
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/mpich-discuss/attachments/20100319/dde00573/attachment-0001.htm>
More information about the mpich-discuss
mailing list