[petsc-users] [beginner question] Different communicators in the two objects: Argument # 1 and 2 flag 3!

Niklas Fischer niklas at niklasfi.de
Tue Apr 22 07:12:35 CDT 2014


Am 22.04.2014 13:57, schrieb Matthew Knepley:
> On Tue, Apr 22, 2014 at 6:48 AM, Niklas Fischer <niklas at niklasfi.de 
> <mailto:niklas at niklasfi.de>> wrote:
>
>     Am 22.04.2014 13:08, schrieb Jed Brown:
>
>         Niklas Fischer <niklas at niklasfi.de
>         <mailto:niklas at niklasfi.de>> writes:
>
>             Hello,
>
>             I have attached a small test case for a problem I am
>             experiencing. What
>             this dummy program does is it reads a vector and a matrix
>             from a text
>             file and then solves Ax=b. The same data is available in
>             two forms:
>               - everything is in one file (matops.s.0 and vops.s.0)
>               - the matrix and vector are split between processes
>             (matops.0,
>             matops.1, vops.0, vops.1)
>
>             The serial version of the program works perfectly fine but
>             unfortunately
>             errors occure, when running the parallel version:
>
>             make && mpirun -n 2 a.out matops vops
>
>             mpic++ -DPETSC_CLANGUAGE_CXX -isystem
>             /home/data/fischer/libs/petsc-3.4.3/arch-linux2-c-debug/include
>             -isystem
>             /home/data/fischer/libs/petsc-3.4.3/include petsctest.cpp
>             -Werror -Wall
>             -Wpedantic -std=c++11 -L
>             /home/data/fischer/libs/petsc-3.4.3/arch-linux2-c-debug/lib -lpetsc
>             /usr/bin/ld: warning: libmpi_cxx.so.0, needed by
>             /home/data/fischer/libs/petsc-3.4.3/arch-linux2-c-debug/lib/libpetsc.so,
>             may conflict with libmpi_cxx.so.1
>             /usr/bin/ld: warning: libmpi.so.0, needed by
>             /home/data/fischer/libs/petsc-3.4.3/arch-linux2-c-debug/lib/libpetsc.so,
>             may conflict with libmpi.so.1
>             librdmacm: couldn't read ABI version.
>             librdmacm: assuming: 4
>             CMA: unable to get RDMA device list
>             --------------------------------------------------------------------------
>             [[43019,1],0]: A high-performance Open MPI point-to-point
>             messaging module
>             was unable to find any relevant network interfaces:
>
>             Module: OpenFabrics (openib)
>                Host: dornroeschen.igpm.rwth-aachen.de
>             <http://dornroeschen.igpm.rwth-aachen.de>
>             CMA: unable to get RDMA device list
>
>         It looks like your MPI is either broken or some of the code
>         linked into
>         your application was compiled with a different MPI or
>         different version.
>         Make sure you can compile and run simple MPI programs in parallel.
>
>     Hello Jed,
>
>     thank you for your inputs. Unfortunately MPI does not seem to be
>     the issue here. The attachment contains a simple MPI hello world
>     program which runs flawlessly (I will append the output to this
>     mail) and I have not encountered any problems with other MPI
>     programs. My question still stands.
>
>
> This is a simple error. You created the matrix A using 
> PETSC_COMM_WORLD, but you try to view it
> using PETSC_VIEWER_STDOUT_SELF. You need to use 
> PETSC_VIEWER_STDOUT_WORLD in
> order to match.
>
>   Thanks,
>
>      Matt
>
>     Greetings,
>     Niklas Fischer
>
>     mpirun -np 2 ./mpitest
>
>     librdmacm: couldn't read ABI version.
>     librdmacm: assuming: 4
>     CMA: unable to get RDMA device list
>     --------------------------------------------------------------------------
>     [[44086,1],0]: A high-performance Open MPI point-to-point
>     messaging module
>     was unable to find any relevant network interfaces:
>
>     Module: OpenFabrics (openib)
>       Host: dornroeschen.igpm.rwth-aachen.de
>     <http://dornroeschen.igpm.rwth-aachen.de>
>
>     Another transport will be used instead, although this may result in
>     lower performance.
>     --------------------------------------------------------------------------
>     librdmacm: couldn't read ABI version.
>     librdmacm: assuming: 4
>     CMA: unable to get RDMA device list
>     Hello world from processor dornroeschen.igpm.rwth-aachen.de
>     <http://dornroeschen.igpm.rwth-aachen.de>, rank 0 out of 2 processors
>     Hello world from processor dornroeschen.igpm.rwth-aachen.de
>     <http://dornroeschen.igpm.rwth-aachen.de>, rank 1 out of 2 processors
>     [dornroeschen.igpm.rwth-aachen.de:128141
>     <http://dornroeschen.igpm.rwth-aachen.de:128141>] 1 more process
>     has sent help message help-mpi-btl-base.txt / btl:no-nics
>     [dornroeschen.igpm.rwth-aachen.de:128141
>     <http://dornroeschen.igpm.rwth-aachen.de:128141>] Set MCA
>     parameter "orte_base_help_aggregate" to 0 to see all help / error
>     messages
>
>
Thank you, Matthew, this solves my viewing problem. Am I doing something 
wrong when initializing the matrices as well? The matrix' viewing output 
starts with "Matrix Object: 1 MPI processes" and the Krylov solver does 
not converge.

Your help is really appreciated,
Niklas Fischer
>
>
> -- 
> What most experimenters take for granted before they begin their 
> experiments is infinitely more interesting than any results to which 
> their experiments lead.
> -- Norbert Wiener

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20140422/f543dafe/attachment.html>


More information about the petsc-users mailing list