[petsc-users] Question about PETSc installs and MPI

Barry Smith bsmith at mcs.anl.gov
Fri Jul 18 20:57:55 CDT 2014


   I ran the program on linux with 1,2, 4 processes under valgrind for both types of boundary conditions and it ran fine. 

   Suggest your colleagues do a test configure of PETSc using —download-mpich and see if they still get the problem or if it runs ok.

   The can also run with valgrind and see what it reports. http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind

   Barry

On Jul 18, 2014, at 8:16 PM, John Yawney <jyawney123 at gmail.com> wrote:

> Hello,
> 
> I had a question about PETSc installations. On my local computer I configured PETSc (v 3.4.2) using the options:
> 
> ./configure --with-cc=mpicc --with-cxx=mpic++ --download-f-blas-lapack --download-mpich --download-hypre
> 
> I wrote a test program that defines a vector using DMDAs, computes a dot product, exchanges halo elements, and computes a low-order FD derivative of the vector. Under my installation of PETSc everything works fine. For some reason, when my colleagues run the program, they get segmentation fault errors. If they change the y and z boundary types to GHOSTED as well, they get the program to run until the end (seg faults at the end though) but they get a local value of the dot product. I've attached the main.cpp file for this script.
> 
> When they installed their versions of PETSc they didn't use the --download-mpich option but instead used either:
> ./configure --download-f-blas-lapack --with-scalar-type=complex
> or with the option: --with-mpi-dir=/home/kim/anaconda/pkgs/mpich2-1.3-py27_0
> 
> Could this be causing a problem with the parallelization under PETSc?
> 
> Thanks for the help and sorry for the long question.
> 
> Best regards,
> John
> <main.cpp>



More information about the petsc-users mailing list