[mpich-discuss] MPICH 2 on Windows 7 64 bit
Leonardo Garbanzos
leonardo.garbanzos at arup.com
Thu May 10 03:14:15 CDT 2012
Hi Jayesh,
We are having issue running the MPICH 2 on our windows 7 64bit PCs. We have installed the MPICH 2 1.0.8, 1.2, 1.3rc2 and 1.4.1 vers of it, but we are still getting the below error.
"Error"
Microsoft Windows [Version 6.1.7601]
Copyright (c) 2009 Microsoft Corporation. All rights reserved.
C:\Users\sin.cfd>cd \fds_mpi\sr_lobby05
C:\FDS_MPI\sr_lobby05>mpiexec -file config_srlobby05.txt
User credentials needed to launch processes:
account (domain\user) [GLOBAL\sin.cfd]:
password:
Process 0 of 5 is running on SINPCSGH146PRT1.clients.global.arup.com
Process 5 of 5 is running on SINPCSGH146PRT1.clients.global.arup.com
Process 1 of 5 is running on SINPCSGH029T4WQ.clients.global.arup.com
Process 3 of 5 is running on SINPCSGH029T4WQ.clients.global.arup.com
Process 2 of 5 is running on SINPCSGH029T4WQ.clients.global.arup.com
Process 4 of 5 is running on SINPCSGH029T4WQ.clients.global.arup.com
Mesh 1 is assigned to Process 0
Mesh 2 is assigned to Process 1
Mesh 3 is assigned to Process 2
Mesh 4 is assigned to Process 3
Mesh 5 is assigned to Process 4
Mesh 6 is assigned to Process 5
Fire Dynamics Simulator
Compilation Date : Fri, 29 Oct 2010
Version: 5.5.3; MPI Enabled; OpenMP Disabled
SVN Revision No. : 7031
Job TITLE : Base:Medium t2 5MW axis with natural ventilation
Job ID string : sr_lobby05
Fatal error in PMPI_Gatherv: Internal MPI error!, error stack:
PMPI_Gatherv(386).....: MPI_Gatherv failed(sbuf=000000003C553848, scount=1, MPI_
DOUBLE_PRECISION, rbuf=000000003C553848, rcnts=000000003C4D6928, displs=00000000
3C4D69E8, MPI_DOUBLE_PRECISION, root=0, MPI_COMM_WORLD) failed
MPIR_Gatherv_impl(199):
MPIR_Gatherv(103).....:
MPIR_Localcopy(349)...: memcpy arguments alias each other, dst=000000003C553848
src=000000003C553848 len=8
Fatal error in MPI_Allreduce: Other MPI error, error stack:
MPI_Allreduce(824)...................: MPI_Allreduce(sbuf=000000003C376CB8, rbuf
=000000003C376C88, count=5, MPI_LOGICAL, MPI_LXOR, MPI_COMM_WORLD) failed
MPIR_Allreduce_impl(682).............:
MPIR_Allreduce_intra(197)............:
MPIR_Bcast_impl(1150)................:
MPIR_Bcast_intra(1021)...............:
MPIR_Bcast_binomial(157).............:
MPIC_Recv(108).......................:
MPIC_Wait(528).......................:
MPIDI_CH3I_Progress(334).............:
MPID_nem_mpich2_blocking_recv(906)...:
MPID_nem_newtcp_module_poll(37)......:
MPID_nem_newtcp_module_connpoll(2655):
gen_read_fail_handler(1145)..........: read from socket failed - The specified n
etwork name is no longer available.
Fatal error in MPI_Allreduce: Other MPI error, error stack:
MPI_Allreduce(824)...................: MPI_Allreduce(sbuf=000000003C346CB8, rbuf
=000000003C346C88, count=5, MPI_LOGICAL, MPI_LXOR, MPI_COMM_WORLD) failed
MPIR_Allreduce_impl(682).............:
MPIR_Allreduce_intra(191)............:
allreduce_intra_or_coll_fn(103)......:
MPIR_Allreduce_intra(361)............:
MPIC_Sendrecv(189)...................:
MPIC_Wait(528).......................:
MPIDI_CH3I_Progress(334).............:
MPID_nem_mpich2_blocking_recv(906)...:
MPID_nem_newtcp_module_poll(37)......:
MPID_nem_newtcp_module_connpoll(2655):
gen_read_fail_handler(1145)..........: read from socket failed - The specified n
etwork name is no longer available.
Fatal error in MPI_Allreduce: Other MPI error, error stack:
MPI_Allreduce(824)...................: MPI_Allreduce(sbuf=000000003C416CB8, rbuf
=000000003C416C88, count=5, MPI_LOGICAL, MPI_LXOR, MPI_COMM_WORLD) failed
MPIR_Allreduce_impl(682).............:
MPIR_Allreduce_intra(197)............:
MPIR_Bcast_impl(1150)................:
MPIR_Bcast_intra(1021)...............:
MPIR_Bcast_binomial(157).............:
MPIC_Recv(108).......................:
MPIC_Wait(528).......................:
MPIDI_CH3I_Progress(334).............:
MPID_nem_mpich2_blocking_recv(906)...:
MPID_nem_newtcp_module_poll(37)......:
MPID_nem_newtcp_module_connpoll(2655):
gen_read_fail_handler(1145)..........: read from socket failed - The specified n
etwork name is no longer available.
Fatal error in MPI_Allreduce: Other MPI error, error stack:
MPI_Allreduce(824)...................: MPI_Allreduce(sbuf=000000003C466CB8, rbuf
=000000003C466C88, count=5, MPI_LOGICAL, MPI_LXOR, MPI_COMM_WORLD) failed
MPIR_Allreduce_impl(682).............:
MPIR_Allreduce_intra(197)............:
MPIR_Bcast_impl(1150)................:
MPIR_Bcast_intra(1021)...............:
MPIR_Bcast_binomial(157).............:
MPIC_Recv(108).......................:
MPIC_Wait(528).......................:
MPIDI_CH3I_Progress(334).............:
MPID_nem_mpich2_blocking_recv(906)...:
MPID_nem_newtcp_module_poll(37)......:
MPID_nem_newtcp_module_connpoll(2655):
gen_read_fail_handler(1145)..........: read from socket failed - The specified n
etwork name is no longer available.
Fatal error in MPI_Allreduce: Other MPI error, error stack:
MPI_Allreduce(824)...................: MPI_Allreduce(sbuf=000000003C496CB8, rbuf
=000000003C496C88, count=5, MPI_LOGICAL, MPI_LXOR, MPI_COMM_WORLD) failed
MPIR_Allreduce_impl(682).............:
MPIR_Allreduce_intra(197)............:
MPIR_Bcast_impl(1150)................:
MPIR_Bcast_intra(1021)...............:
MPIR_Bcast_binomial(157).............:
MPIC_Recv(108).......................:
MPIC_Wait(528).......................:
MPIDI_CH3I_Progress(334).............:
MPID_nem_mpich2_blocking_recv(906)...:
MPID_nem_newtcp_module_poll(37)......:
MPID_nem_newtcp_module_connpoll(2655):
gen_read_fail_handler(1145)..........: read from socket failed - The specified n
etwork name is no longer available.
job aborted:
rank: node: exit code[: error message]
0: 10.197.240.36: 1: process 0 exited without calling finalize
1: 10.197.240.35: 1: process 1 exited without calling finalize
2: 10.197.240.35: 1: process 2 exited without calling finalize
3: 10.197.240.35: 1: process 3 exited without calling finalize
4: 10.197.240.35: 1: process 4 exited without calling finalize
5: 10.197.240.36: 1: process 5 exited without calling finalize
C:\FDS_MPI\sr_lobby05>
Do you have any recommendation for us to resolve the issue.
Thanks
Leonardo Garbanzos
IT Support Analyst
Arup
10 Hoe Chiang Road #26-01 Keppel Towers Singapore 089315
t +65 6411 2500 d +65 6411 2540
f +65 6411 2501 m +65 9817 3002
www.arup.com<http://www.arup.com>
Arup Singapore Pte Ltd - Reg. No. 200100731M
____________________________________________________________
Electronic mail messages entering and leaving Arup business
systems are scanned for acceptability of content and viruses
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/mpich-discuss/attachments/20120510/dc65d19e/attachment-0001.htm>
More information about the mpich-discuss
mailing list