[mpich-discuss] Scalability of Intel quad core (Harpertown) cluster
Pavan Balaji
balaji at mcs.anl.gov
Fri Mar 28 01:21:56 CDT 2008
Is the application itself scalable? It is possible that the application
behavior itself might have some inherent scalability issue and might
have nothing to do with communication. You should consider profiling
your application to see what exactly is happening before trying out
different communication channels, MPI implementations or networks.
MPE/Jumpshot (that is packaged with MPICH2) is a good tool to profile
the application code. mpiP is another tool to see how much of the
application time is actually spent inside MPI. For example, if the
application is only spending 5% of its time within MPI, irrespective of
what network or communication channel you use, it'll not help you much.
-- Pavan
On 03/28/2008 12:50 AM, Hee Il Kim wrote:
> Hi,
>
> I recently made a Rocks cluster of 10 comuting nodes with dual Intel
> quad core cpus (Harpertown). The nodes are connected with gigabit networks.
>
> The problem is I got bad scalability from the benchmark test of Cactus
> which is well-known program for numerical relativity. It might be
> natural with the gigabit network. However, I suspect I missed something
> important or it's a generic problem of clusters of multicore cpus
> because the benchmark problems does not require massive communication
> between the computing nodes.
>
> I tested openmpi, mpich as well as mpich2 but no significant differences
> between openmpi and mpich2. The mpich2 was configured with ssm. After
> reading some documents on nemesis and multicore cpus, I tested with
> nemesis but got worse results than with ssm.
>
> - Is there any optimized configuration option for multicore cpus like
> Harpertown?
> - Could it be improved only with infiniband, myrinet,...?
> - If the gigabit network was the cause, could it be improved with Open-MX?
>
> I'm a newbie on this field. The questions must be not clear to you. I
> appreciate any helps in advance.
>
> Kim, Hee Il
--
Pavan Balaji
http://www.mcs.anl.gov/~balaji
More information about the mpich-discuss
mailing list