[mpich-discuss] read from socket failed (errno 10055) on 1.3.2p1
Kuraisa, Roy J (BOSTON)
roy_kuraisa at merck.com
Wed Apr 20 12:16:02 CDT 2011
I also see the same problem when running just on two hosts and one core on each host. The error message is a bit shorter (which I'll place at the end of this message). When running the executable where mpich2 v1.2.1p1 is installed, there is no error (and the executable built against 1.2.1 or 1.3.2 ran fine).
Hope this helps.
cheers, roy
run (with my own debug info) and error message:
-----------------------------------------------
D:\roy>smpd -version
1.3.2p1
D:\roy>mpiexec -hosts 2 usctap3800 1 usctap3826 1 \\usdata011\MPRI-App\BAT\RoyTe
st\CorrelateMPI\CorrelateMPI.exe correlate_rep_traits.xml gwat_all_attieeric.h5
out.h5 debug
>>> Root process on computer: USCTAP3800
>>> Root process on computer: USCTAP3800
>>> No. of computers: 2
>>> Summary of InData
Cfg file: correlate_rep_traits.xml
Input hdf5 file: gwat_all_attieeric.h5
Name of x data: repData
Name of x ids: repIDs
Name of y data: traitData
Name of y ids: traitIDs
Filter: pvalue
Filter threshold: 1.0001
Dataset name for correlation: correlations
Dataset name for pvalue: pvalues
Metric name: pearson
Compression: 0
>>> Rank 0: reading input file: gwat_all_attieeric.h5
>>> File: out.h5 exists.
###
### Perf: Time to input data: 0 mins 2 secs
###
>>> in x data - Rank: 0 rows/cols/total: 39558/506/20016348
>>> in y data - Rank: 0 rows/cols/total: 347/506/175582
>>> Rank 0: Broadcasting input data to worker nodes
>>> rank: 1 metric: pearson Length: 7
>>> in x data - Rank: 1 rows/cols/total: 39558/506/20016348
>>> in y data - Rank: 1 rows/cols/total: 347/506/175582
Fatal error in PMPI_Bcast: Other MPI error, error stack:
PMPI_Bcast(1430).................................: MPI_Bcast(buf=00000000018C004
0, count=20016348, MPI_FLOAT, root=0, MPI_COMM_WORLD) failed
MPIR_Bcast_impl(1273)............................:
MPIR_Bcast_intra(1107)...........................:
MPIR_Bcast_binomial(143).........................:
MPIC_Recv(110)...................................:
MPIC_Wait(540)...................................:
MPIDI_CH3I_Progress(353).........................:
MPID_nem_mpich2_blocking_recv(905)...............:
MPID_nem_newtcp_module_poll(37)..................:
MPID_nem_newtcp_module_connpoll(2669)............:
MPID_nem_newtcp_module_recv_success_handler(2364):
MPID_nem_newtcp_module_post_readv_ex(330)........:
MPIU_SOCKW_Readv_ex(392).........................: read from socket failed, An o
peration on a socket could not be performed because the system lacked sufficient
buffer space or because a queue was full.
(errno 10055)
job aborted:
rank: node: exit code[: error message]
0: usctap3800: 123
1: usctap3826: 1: process 1 exited without calling finalize
-----Original Message-----
From: Jayesh Krishna [mailto:jayesh at mcs.anl.gov]
Sent: Tuesday, April 19, 2011 4:18 PM
To: mpich-discuss at mcs.anl.gov
Cc: Kuraisa, Roy J (BOSTON)
Subject: Re: [mpich-discuss] read from socket failed (errno 10055) on 1.3.2p1
Hi,
Can you send us the complete error message ?
Regards,
Jayesh
----- Original Message -----
From: "Roy J Kuraisa (BOSTON)" <roy_kuraisa at merck.com>
To: mpich-discuss at mcs.anl.gov
Sent: Tuesday, April 19, 2011 2:32:24 PM
Subject: [mpich-discuss] read from socket failed (errno 10055) on 1.3.2p1
Hi,
Summary:
---------------
On Windows when I execute the following command (working on a fairly large dataset):
mpiexec -hosts 2 usctap3825 15 usctap3488 1 \\fs1\correlatempi.exe cfg.xml in.h5 out.h5 debug
I encounter an MPI gather error (read from socket failed (errno 10055). See error stack at end of this message. If I run on only one computer (with 16 cores):
mpiexec -hosts 1 usctap3825 15 \\fs1\correlatempi.exe cfg.xml in.h5 out.h5 debug
the program runs successfully.
Additionally, both of the above commands run successfully on mpich2 v1.2.1 (although I had to build on mpich2 1.2.1 and used different servers that are configured exactly like the origian servers noted above (e.g., usctap3825, 16-core, 64GB memory, etc).
I noticed that a similar error was fixed in mpich2-1.2 ( http://trac.mcs.anl.gov/projects/mpich2/ticket/895 ). Could this have regressed? tia.
System Configuration:
--------------------------------
Server1 (usctap3825)
-------
a. Windows Server 2003, 64-bit, SP2
b. 16 cores/processors
c. 64GB memory
d. Physical computer
Server2 (usctap3488)
-------
a. Windows Server 2003, 64-bit, SP2
b. 2 cores/processors
c. 8GB memory
d. Virtual Machine
cheers, roy
error stack:
----------------
Fatal error in PMPI_Gatherv: Other MPI error, error stack:
PMPI_Gatherv(398)................................: MPI_Gatherv failed(sbuf=00000
0003AA30040, scount=97787376, MPI_FLOAT, rbuf=0000000180040040, rcnts=000000000D
6515E0, displs=000000000D651630, MPI_FLOAT, root=0, MPI_COMM_WORLD) failed
MPIR_Gatherv_impl(210)...........................:
MPIR_Gatherv(118)................................:
MPIC_Waitall_ft(852).............................:
MPIR_Waitall_impl(121)...........................:
MPIDI_CH3I_Progress(353).........................:
MPID_nem_mpich2_blocking_recv(905)...............:
MPID_nem_newtcp_module_poll(37)..................:
MPID_nem_newtcp_module_connpoll(2669)............:
MPID_nem_newtcp_module_recv_success_handler(2364):
MPID_nem_newtcp_module_post_readv_ex(330)........:
MPIU_SOCKW_Readv_ex(392).........................: read from socket failed, An o
peration on a socket could not be performed because the system lacked sufficient
buffer space or because a queue was full.
(errno 10055)
Notice: This e-mail message, together with any attachments, contains
information of Merck & Co., Inc. (One Merck Drive, Whitehouse Station,
New Jersey, USA 08889), and/or its affiliates Direct contact information
for affiliates is available at
http://www.merck.com/contact/contacts.html) that may be confidential,
proprietary copyrighted and/or legally privileged. It is intended solely
for the use of the individual or entity named on this message. If you are
not the intended recipient, and have received this message in error,
please notify us immediately by reply e-mail and then delete it from
your system.
_______________________________________________
mpich-discuss mailing list
mpich-discuss at mcs.anl.gov
https://lists.mcs.anl.gov/mailman/listinfo/mpich-discuss
Notice: This e-mail message, together with any attachments, contains
information of Merck & Co., Inc. (One Merck Drive, Whitehouse Station,
New Jersey, USA 08889), and/or its affiliates Direct contact information
for affiliates is available at
http://www.merck.com/contact/contacts.html) that may be confidential,
proprietary copyrighted and/or legally privileged. It is intended solely
for the use of the individual or entity named on this message. If you are
not the intended recipient, and have received this message in error,
please notify us immediately by reply e-mail and then delete it from
your system.
More information about the mpich-discuss
mailing list