[mpich-discuss] Problem while running example program Cpi with morethan 1 task

Thejna Tharammal ttharammal at marum.de
Thu Sep 2 16:47:51 CDT 2010


 Hi,
I tried with 1.3b1, 
until 6 tasks it works fine like
=================
-bash-3.2$ mpiexec -n 6 ./cpi
Process 0 of 6 is on k1
Process 3 of 6 is on k4
Process 1 of 6 is on k2
Process 2 of 6 is on k3
Process 4 of 6 is on k5
Process 5 of 6 is on k6
pi is approximately 3.1415926544231239, Error is 0.0000000008333307
wall clock time = 0.001316
=======================================

 but from 7 it shows : (I have attached a detailed error log),


-bash-3.2$ mpiexec -n 7 ./cpi
Process 0 of 7 is on k1
Process 6 of 7 is on k1
Process 1 of 7 is on k2
Process 2 of 7 is on k3
Process 3 of 7 is on k4
Process 4 of 7 is on k5
Process 5 of 7 is on k6
pi is approximately 3.1415926544231234, Error is 0.0000000008333303
wall clock time = 0.002738
Fatal error in MPI_Finalize: Other MPI error, error stack:
MPI_Finalize(302).................: MPI_Finalize failed
MPI_Finalize(210).................:
MPID_Finalize(110)................:
MPIDI_CH3U_VC_WaitForClose(343)...: an error occurred while the device was
waiting for all open connections to close
MPIDI_CH3I_Progress(184)..........:
MPID_nem_mpich2_blocking_recv(895):
MPID_nem_tcp_connpoll(1746).......: Communication error with rank 0:
Fatal error in MPI_Finalize: Other MPI error, error stack:
....

 MPID_nem_tcp_connpoll(1746).......: Communication error with rank 0:
APPLICATION TERMINATED WITH THE EXIT STRING: Segmentation fault (signal 11)
=====================================
Thanks,
Thejna

----------------original message-----------------
From: "Pavan Balaji" balaji at mcs.anl.gov
To: mpich-discuss at mcs.anl.gov
Date: Thu, 02 Sep 2010 11:17:06 -0500
-------------------------------------------------
 
 
> 
> On 09/02/2010 11:16 AM, Pavan Balaji wrote:
>>
>> The difference in the working case and the failing cases is the use of
>> shared memory -- the working case has no shared memory usage since you
>> don't have multiple processes on the same node.
>>
>> Can you try out the 1.3b1 release of Hydra before we go digging into
this?
> 
> 1.3b1 release of MPICH2, not Hydra, sorry.
> 
> -- 
> Pavan Balaji
> http://www.mcs.anl.gov/~balaji
> _______________________________________________
> mpich-discuss mailing list
> mpich-discuss at mcs.anl.gov
> https://lists.mcs.anl.gov/mailman/listinfo/mpich-discuss
> 
-------------- next part --------------
A non-text attachment was scrubbed...
Name: error_cpi.log
Type: text/x-log
Size: 5323 bytes
Desc: not available
URL: <http://lists.mcs.anl.gov/pipermail/mpich-discuss/attachments/20100902/b0f602f7/attachment-0001.bin>


More information about the mpich-discuss mailing list