[mpich-discuss] configuration problem

Anthony Chan chan at mcs.anl.gov
Tue Apr 27 14:21:30 CDT 2010


Kate,

If I remember correctly, some old versions of mpich-1 copied the Fortran MPI
header file into user's location because some older Fortran compiler does not
support -I.  If you copied your Fortran code which has been built with
mpich-1, that means you may have copied the Fortran MPI header file as well
(delete the MPI header in your source before building it with MPICH2 or any
other MPI implementations).  AFAIK, most modern Fortran compiler has some way to
specify include directory (Not necessarily -I), so the mechanism of copying
fortran mpi header file is no longer needed.  Modern MPI implementations
provide MPI compiler wrappers which take care of include directory as well
linking with all the related helper libraries for MPI, so you should make
sure you are using mpif77/mpif90 in compiling your code.

A.Chan

----- "Kate Steenhauer" <k.steenhauer at abdn.ac.uk> wrote:

> Thanks for your reply. I will look into it. Note that I am relatively
> new to this, I do have another question related to your answer. The
> code I am trying to compile is successfully run in parallel on another
> cluster (this cluster uses Redhat, various mpich1 versions, e.g.
> mpich-1.2.5..12, and various versions ofIntel FORTRAN compiler
> (ifort), e.g. 7, 9 and 12). How is it that I can have a bug in the MPI
> program as I directly copied the whole code on to my workstation? Is
> this because there are discrepancies between mpich1 and mpich2? If so,
> MPI_Scatter routines seems to be correctly working on one cluster with
> mpich1 and not on my workstation with mpich2? This is why I tried to
> imitate the other cluster as much as possible and tried installing
> mpich1, which gives yet another problem, i.e.  the configuration
> problem, described below.
> 
> Have you got an anwer for this maybe? And did you have a look at my
> mpif.inc file as other people using the other cluster tell me it is
> compatible and in use with different mpi versions.
> 
> Thanks
> 
>  Kate
> 
> 
> 
> ________________________________________
> From: mpich-discuss-bounces at mcs.anl.gov
> [mpich-discuss-bounces at mcs.anl.gov] On Behalf Of Rajeev Thakur
> [thakur at mcs.anl.gov]
> Sent: 27 April 2010 17:32
> To: mpich-discuss at mcs.anl.gov; 'Anthony Chan'
> Subject: Re: [mpich-discuss] configuration problem
> 
> Kate,
>      You have to use the mpif.h file that comes with the MPI
> implementation, not from some other MPI implementation. With MPICH2,
> you
> have to use MPICH2's include file.
> 
> The error with MPI_Scatter indicates there is a bug in your MPI
> program.
> You are using the same buffer as sendbuf and recvbuf, which is not
> allowed in MPI. You can use the MPI_IN_PLACE argument instead, as
> described in the MPI standard.
> 
> I would recommend using MPICH2 instead of trying to get MPICH-1 to
> work.
> 
> Rajeev
> 
> 
> > -----Original Message-----
> > From: mpich-discuss-bounces at mcs.anl.gov
> > [mailto:mpich-discuss-bounces at mcs.anl.gov] On Behalf Of
> > Steenhauer, Kate
> > Sent: Tuesday, April 27, 2010 9:50 AM
> > To: Anthony Chan; mpich-discuss at mcs.anl.gov
> > Subject: Re: [mpich-discuss] configuration problem
> >
> > Initially, I tried to run my code with mpich2. The following
> happens.
> >
> > We have 1 workstation with 8 processors, the following
> > software: Linux Centos, Fortran Intel 11.1.
> >
> > Mpich2, Intel Fortran and Linux is not working on our pc. It
> > does not like the mpif.inc file, it immediately gets into
> > problems at the initialisation of the MPI (subroutine MPI_INI).
> > mpirun -np 8 RUN02
> > > Fatal error in MPI_Comm_size: Invalid communicator, error stack:
> > > MPI_Comm_size(111): MPI_Comm_size(comm=0x5b, size=0x7fffce629784)
> > > failed
> > > MPI_Comm_size(69).: Invalid communicator MPISTART
> > > rank 7 in job 2  cops-021026_40378   caused collective
> > abort of all ranks
> > >  exit status of rank 7: killed by signal 9
> >
> > Then when I change the mpif.inc file (see attachment) and
> > direct it to the mpif.h file that came with the mpich2
> > library, it gets passed this problem but then runs into a
> > next problem further down the line at MPI_SCATTER, where it
> > is trying to distribute data to the different processors. The
> > error message is 'memcpy argument memory ranges overlap,
> > dst_=0xafd74f8 src_=0xafd750c len_=16200, internal ABORT'.
> >
> > There is something in the parametresiation within the mpi
> > that is possibly different then when the code is successfully
> > run on another cluster (this cluster uses Redhat, various
> > mpich versions, e.g. mpich-1.2.5..12, and various versions of
> > Intel FORTRAN compiler (ifort), e.g. 7, 9 and 12).
> >
> > I have attached the mpif.inc file.
> >
> > Please let me know if you have any ideas?
> >
> > Thanks
> >
> > Kate
> >
> > -----Original Message-----
> > From: mpich-discuss-bounces at mcs.anl.gov
> > [mailto:mpich-discuss-bounces at mcs.anl.gov] On Behalf Of
> > chan at mcs.anl.gov
> > Sent: 27 April 2010 15:42
> > To: mpich-discuss at mcs.anl.gov
> > Subject: Re: [mpich-discuss] configuration problem
> >
> >
> > Is there any reason you can't use mpich2 ?  The latest stable
> > release of mpich2 is 1.2.1p1.
> >
> > mpich-1 is no longer officially supported.  Latest fortran
> compilers
> > are much better supported in mpich2.
> >
> > A.Chan
> >
> > ----- "Kate Steenhauer" <k.steenhauer at abdn.ac.uk> wrote:
> >
> > > Hello,
> > >
> > > We are trying to install mpich-1.2.7p1. I downloaded and
> > unzipped this
> > > version from http://www.mcs.anl.gov/research/projects/mpi/mpich1/
> > >
> > > We have 1 workstation with 8 processors, and the following
> software:
> > > Linux Centos and Fortran Intel 11.1.
> > >
> > > When the documentation guidelines with regard to configuring and
> > > making are followed (see the files attached) it all seems ok,
> e.g.
> > > mpif90 is generated. However, when a simple parallel job is
> > tested we
> > > get the following error:
> > >
> > > mpif90 -o testA MPITEST.f90
> > > No Fortran 90 compiler specified when mpif90 was created, or
> > > configuration file does not specify a compiler.
> > >
> > > Is there a specific prefix I need to give with an ifort fortran
> > > compiler when I configure mpich?
> > >
> > > I would like to thank you in advance for your help. Please
> > let me know
> > > if you need any further details.
> > >
> > > Regards
> > >
> > > Kate Steenhauer
> > >
> > > University of Aberdeen
> > >
> > > 01224-272806
> > >
> > >
> > >
> > >
> > >
> > > The University of Aberdeen is a charity registered in Scotland,
> No
> > > SC013683.
> > >
> > > _______________________________________________
> > > mpich-discuss mailing list
> > > mpich-discuss at mcs.anl.gov
> > > https://lists.mcs.anl.gov/mailman/listinfo/mpich-discuss
> > _______________________________________________
> > mpich-discuss mailing list
> > mpich-discuss at mcs.anl.gov
> > https://lists.mcs.anl.gov/mailman/listinfo/mpich-discuss
> >
> >
> > The University of Aberdeen is a charity registered in
> > Scotland, No SC013683.
> >
> 
> _______________________________________________
> mpich-discuss mailing list
> mpich-discuss at mcs.anl.gov
> https://lists.mcs.anl.gov/mailman/listinfo/mpich-discuss
> 
> 
> The University of Aberdeen is a charity registered in Scotland, No
> SC013683.


More information about the mpich-discuss mailing list