[mpich-discuss] MPI on a cluster of dual-CPU machines

Rajeev Thakur thakur at mcs.anl.gov
Sat Jul 18 09:08:11 CDT 2009


The default channel in 1.0.8 is ch3:sock, which will use TCP for
communication even within a node. If you configure with
--with-device=ch3:nemesis, it will use shared memory for communication
within a node and TCP across. In 1.1, nemesis is the default. The way to
launch processes and MPD configuration hasn't changed between 1.0.8 and
1.1.

Rajeev


> -----Original Message-----
> From: mpich-discuss-bounces at mcs.anl.gov 
> [mailto:mpich-discuss-bounces at mcs.anl.gov] On Behalf Of Pavan Balaji
> Sent: Friday, July 17, 2009 7:09 PM
> To: mpich-discuss at mcs.anl.gov
> Subject: Re: [mpich-discuss] MPI on a cluster of dual-CPU machines
> 
> 
> > Just one more question though. You say that the latest 
> release detects 
> > this configuration. I am using mpich2-1.0.8. Does this version too 
> > detect the presence of dual CPUs automatically?
> 
> Not with the default configuration. You'll need to configure 
> with --with-device=ch3:nemesis for that. But, it's still 
> advisable that you move to 1.1, since ch3:nemesis has had 
> quite a few fixes and major improvements between 1.0.8 and 1.1.
> 
>   -- Pavan
> 
> --
> Pavan Balaji
> http://www.mcs.anl.gov/~balaji
> 



More information about the mpich-discuss mailing list