[MPICH] Intel 10.1
Rajeev Thakur
thakur at mcs.anl.gov
Wed Feb 13 22:26:56 CST 2008
You can also run the entire MPICH2 test suite by running "make testing" in
the top-level directory. (Can take an hour or more.)
Rajeev
> -----Original Message-----
> From: owner-mpich-discuss at mcs.anl.gov
> [mailto:owner-mpich-discuss at mcs.anl.gov] On Behalf Of Mike Colonno
> Sent: Wednesday, February 13, 2008 9:38 PM
> To: Pavan Balaji
> Cc: mpich-discuss at mcs.anl.gov
> Subject: RE: [MPICH] Intel 10.1
>
>
> Simple cases, like the test cases that come with MPICH
> distributions, seem to work fine for n = anything. Any more
> sophisticated code works for n = small number (8 is typical)
> but fails for n > this number. This is independent of how the
> processes are distributed (number of processes / server). I
> have tried several different version of MPICH and MPICH2,
> including the one you mentioned, but all have the same
> result. This leads me to believe the issue lies with the
> compiler(s). We're using x64 servers, 2 dual-core Xeons per
> machine, RHEL 4.3 and 4.5 (upgrading OS had no effect).
>
> Thanks,
> ~Mike C.
>
>
> ________________________________
>
> From: Pavan Balaji [mailto:balaji at mcs.anl.gov]
> Sent: Wed 2/13/2008 7:22 PM
> To: Mike Colonno
> Cc: mpich-discuss at mcs.anl.gov
> Subject: Re: [MPICH] Intel 10.1
>
>
>
> Mike,
>
> We internally use Intel 9.1. We are still waiting on a site
> license for Intel 10.1. Are you using mpich2-1.0.6p1? Also,
> what hardware platform are you running these tests on?
>
> Can you try a simple test and see if it works fine:
>
> $ mpiexec hostname
>
> -- Pavan
>
> On 02/13/2008 08:55 PM, Mike Colonno wrote:
> > Hi folks ~
> >
> > Has anyone out there had any luck building MPICH and /
> or MPICH2 (and subsequent MPI applications) with Intel 10.1
> compilers (C++ and Fortran)? If so, please forward me the
> details (flags, versions, OS, etc.). All of my MPI programs
> suffer "collective abort" errors killed by signal 9 or 11
> (with essentially no repeatable pattern between the two)
> which are likely caused by seg. faults behind the scenes.
> These codes worked great on older hardware / compiler
> versions and despite a large number of experiments I have
> been unable to find the secret. Compiled in sequential mode
> (without MPI) they all function well.
> >
> > Thanks,
> > ~Mike C.
> >
> >
>
> --
> Pavan Balaji
> http://www.mcs.anl.gov/~balaji
>
>
>
More information about the mpich-discuss
mailing list