[MPICH] Communication problem on a small heterogeneous ring involving Red Hat Linux - Enterprise Edition
Matthew Chambers
matthew.chambers at vanderbilt.edu
Wed Nov 29 14:07:33 CST 2006
Is he actually talking about a heterogeneous system? All I see are
different operating systems. Unless the Xeon is 64 bit, it seems like byte
ordering and type sizes should all be equal, which as far as I know is the
only kind of heterogeneity that would affect MPI. Am I wrong?
Matt Chambers
Vanderbilt Bioinformatics
_____
From: owner-mpich-discuss at mcs.anl.gov
[mailto:owner-mpich-discuss at mcs.anl.gov] On Behalf Of Rajeev Thakur
Sent: Wednesday, November 29, 2006 1:11 PM
To: 'Salvatore Sorce'; mpich-discuss at mcs.anl.gov
Subject: RE: [MPICH] Communication problem on a small heterogeneous ring
involving Red Hat Linux - Enterprise Edition
MPICH2 does not work on heterogeneous systems yet, although we plan to
support it in the future.
Rajeev
_____
From: owner-mpich-discuss at mcs.anl.gov
[mailto:owner-mpich-discuss at mcs.anl.gov] On Behalf Of Salvatore Sorce
Sent: Wednesday, November 29, 2006 8:51 AM
To: mpich-discuss at mcs.anl.gov
Subject: [MPICH] Communication problem on a small heterogeneous ring
involving Red Hat Linux - Enterprise Edition
Dear all,
I have a small cluster composed by four machines: two double-PIII running
Scientific Linux CERN release 3.0.8 (kernel 2.4.21-37.EL), one double-Xeon
running Red Hat Enterprise Linux WS release 3 (Taroon update 4, kernel
2.4.21-27 EL), and one single-PIV running Red Hat Linux Release 9 (Shrike,
kernel 2.4.20-8). All machines have MPICH2 1.0.4p1 installed.
Tests on whatever kind of ring I set up are OK, and processes are correctly
spawned and started on right hosts.
I experienced problems when processes need to communicate each other, and
one side of the communication is the machine with Red Hat Enterprise Linux
OS. If I do not involve the Enterprise Linux machine in communications, all
runs right.
I am using a simple send-and-receive Fortran test program, where process #1
sends an array of real to process #0. Both processes use blocking
communication functions (mpi_send and mpi_recv).
When process #0 (the receiving one) runs on the Red Hat Enterprise Linux
machine, all hangs up at the mpi_send (maybe because on the Enterprise Linux
side the mpi_recv do not accomplish its task).
When process #1 (the sending one) runs on the Red Hat Enterprise Linux
machine, I obtain the following output:
[cli_0]: aborting job:
Fatal error in MPI_Recv: Other MPI error, error stack:
MPI_Recv(186)................................: MPI_Recv(buf=0xbfff2368,
count=2, MPI_REAL, src=1, tag=17, MPI_COMM_WORLD, status=0xbfff2010) failed
MPIDI_CH3_Progress_wait(217).................: an error occurred while
handling an event returned by MPIDU_Sock_Wait()
MPIDI_CH3I_Progress_handle_sock_event(590)...:
MPIDI_CH3_Sockconn_handle_connopen_event(791): unable to find the process
group structure with id <>
[cli_1]: aborting job:
Fatal error in MPI_Send: Other MPI error, error stack:
MPI_Send(173).............................: MPI_Send(buf=0x7fbffebdf0,
count=2, MPI_REAL, dest=0, tag=17, MPI_COMM_WORLD) failed
MPIDI_CH3_Progress_wait(217)..............: an error occurred while handling
an event returned by MPIDU_Sock_Wait()
MPIDI_CH3I_Progress_handle_sock_event(415):
MPIDU_Socki_handle_read(670)..............: connection failure
(set=0,sock=1,errno=104:Connection reset by peer)
rank 0 in job 1 mpitemp_32877 caused collective abort of all ranks
exit status of rank 0: return code 1
I understand that in both cases mpi_recv causes an error, what is the
problem?
Thank you in advance for your attention.
Regards,
Salvatore.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/mpich-discuss/attachments/20061129/b0a055c3/attachment.htm>
More information about the mpich-discuss
mailing list