[mpich-discuss] Not able to run MPI program parallely...
Pavan Balaji
balaji at mcs.anl.gov
Tue May 1 13:51:48 CDT 2012
On 05/01/2012 01:39 PM, Albert Spade wrote:
> [mpiexec at beowulf.master] Launch arguments:
> /opt/mpich2-1.4.1p1/bin/hydra_pmi_proxy --control-port
> beowulf.master:60190 --debug --rmk user --launcher ssh --demux poll
> --pgid 0 --retries 10 --proxy-id 0
> [mpiexec at beowulf.master] Launch arguments:
> /opt/mpich2-1.4.1p1/bin/hydra_pmi_proxy --control-port
> beowulf.master:60190 --debug --rmk user --launcher ssh --demux poll
> --pgid 0 --retries 10 --proxy-id 1
> [mpiexec at beowulf.master] Launch arguments:
> /opt/mpich2-1.4.1p1/bin/hydra_pmi_proxy --control-port
> beowulf.master:60190 --debug --rmk user --launcher ssh --demux poll
> --pgid 0 --retries 10 --proxy-id 2
> [mpiexec at beowulf.master] Launch arguments:
> /opt/mpich2-1.4.1p1/bin/hydra_pmi_proxy --control-port
> beowulf.master:60190 --debug --rmk user --launcher ssh --demux poll
> --pgid 0 --retries 10 --proxy-id 3
Thanks. It looks like mpiexec thinks that all four nodes are local
nodes, which is a bug that was fixed after 1.4.1p1 was released. Can
you try the mpiexec from 1.5b1:
http://www.mcs.anl.gov/research/projects/mpich2/downloads/tarballs/1.5b1/hydra-1.5b1.tar.gz
You don't need to recompile the application. Just run it with this new
mpiexec.
-- Pavan
--
Pavan Balaji
http://www.mcs.anl.gov/~balaji
More information about the mpich-discuss
mailing list