[petsc-users] [Supercomputing Lab #847] VecView() error in BlueGene

Aron Ahmadia aron.ahmadia at kaust.edu.sa
Tue Apr 27 05:55:51 CDT 2010


I'll take a look at this and report back Rebecca.  I was seeing
similar bus errors on some PETSc example code calling VecView and we
haven't tracked it down yet.

A

On Tue, Apr 27, 2010 at 5:20 AM, Xuefei YUAN via RT
<shaheen-help at kaust.edu.sa> wrote:
>
> Tue Apr 27 12:20:11 2010: Request 847 was acted upon.
>  Transaction: Ticket created by xy2102 at columbia.edu
>       Queue: Shaheen
>     Subject: VecView() error in BlueGene
>       Owner: Nobody
>  Requestors: xy2102 at columbia.edu
>      Status: new
>  Ticket <URL: http://www.hpc.kaust.edu.sa/rt/Ticket/Display.html?id=847 >
>
>
> Hi,
>
> I tried the example code /petsc-dev/src/snes/examples/tutorials/ex5.c
> on bluegene, but get the following message about the VecView().
>
> [0]PETSC ERROR:
> ------------------------------------------------------------------------
> [0]PETSC ERROR: Caught signal number 7 BUS: Bus Error, possibly
> illegal memory access
> [0]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger
> [0]PETSC ERROR: or see
> http://www.mcs.anl.gov/petsc/petsc-as/documentation/faq.html#valgrind[0]PETSC
> ERROR: or try http://valgrind.org on GNU/linux and Apple Mac OS X to
> find memory corruption errors
> [0]PETSC ERROR: likely location of problem given in stack below
> [0]PETSC ERROR: ---------------------  Stack Frames
> ------------------------------------
> [0]PETSC ERROR: Note: The EXACT line numbers in the stack are not available,
> [0]PETSC ERROR:       INSTEAD the line number of the start of the function
> [0]PETSC ERROR:       is given.
> [0]PETSC ERROR: [0] VecView_MPI line 801 src/vec/vec/impls/mpi/pdvec.c
> [0]PETSC ERROR: [0] VecView line 690 src/vec/vec/interface/vector.c
> [0]PETSC ERROR: [0] DAView_VTK line 129 src/dm/da/src/daview.c
> [0]PETSC ERROR: [0] DAView line 227 src/dm/da/src/daview.c
> [0]PETSC ERROR: --------------------- Error Message
> ------------------------------------
> [0]PETSC ERROR: Signal received!
> [0]PETSC ERROR:
> ------------------------------------------------------------------------
> [0]PETSC ERROR: Petsc Development HG revision: unknown HG Date: unknown
> [0]PETSC ERROR: See docs/changes/index.html for recent updates.
> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting.
> [0]PETSC ERROR: See docs/index.html for manual pages.
> [0]PETSC ERROR:
> ------------------------------------------------------------------------
> [0]PETSC ERROR: ./ex5.exe on a arch-bgp- named ionode37 by Unknown Thu
> Jan  1 03:03:12 1970
> [0]PETSC ERROR: Libraries linked from
> /scratch/rebeccaxyf/petsc-dev/arch-bgp-ibm-dbg/lib
> [0]PETSC ERROR: Configure run at Mon Apr 26 15:37:29 2010
> [0]PETSC ERROR: Configure options --with-cc=mpixlc_r
> --with-cxx=mpixlcxx_r --with-fc=mpixlf90_r --with-clanguage=cxx
> --with-blas-lapack-lib="-L/opt/share/math_libraries/lapack/ppc64/IBM/
> -llapack -lblas" --with-x=0 --with-is-color-value-type=short
> --with-shared=0 -CFLAGS="-qmaxmem=-1 -g" -CXXFLAGS="-qmaxmem=-1 -g"
> -FFLAGS="-qmaxmem=-1 -g" --with-debugging=1 --with-fortran-kernels=1
> --with-batch=1 --known-mpi-shared=0 --known-memcmp-ok
> --known-sizeof-char=1 --known-sizeof-void-p=4 --known-sizeof-short=2
> --known-sizeof-int=4 --known-sizeof-long=4 --known-sizeof-size_t=4
> --known-sizeof-long-long=8 --known-sizeof-float=4
> --known-sizeof-double=8 --known-bits-per-byte=8
> --known-sizeof-MPI_Comm=4 --known-sizeof-MPI_Fint=4
> --known-mpi-long-double=1 --known-level1-dcache-assoc=0
> --known-level1-dcache-linesize=32 --known-level1-dcache-size=32768
> --petsc-arch=bgp-dbg PETSC_ARCH=arch-bgp-ibm-dbg
> [0]PETSC ERROR:
> ------------------------------------------------------------------------
> [0]PETSC ERROR: User provided function() line 0 in unknown directory
> unknown file
> Abort(59) on node 0 (rank 0 in comm 1140850688): application called
> MPI_Abort(MPI_COMM_WORLD, 59) - process 0
>
>
> Also, my code is not working for calling VecView() with error message:
>
> [0]PETSC ERROR:
> ------------------------------------------------------------------------
> [0]PETSC ERROR: Caught signal number 7 BUS: Bus Error, possibly
> illegal memory access
> [0]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger
> [0]PETSC ERROR: or see
> http://www.mcs.anl.gov/petsc/petsc-as/documentation/faq.html#valgrind[0]PETSC
> ERROR: or try http://valgrind.org on GNU/linux and Apple Mac OS X to
> find memory corruption errors
> [0]PETSC ERROR: likely location of problem given in stack below
> [0]PETSC ERROR: ---------------------  Stack Frames
> ------------------------------------
> [0]PETSC ERROR: Note: The EXACT line numbers in the stack are not available,
> [0]PETSC ERROR:       INSTEAD the line number of the start of the function
> [0]PETSC ERROR:       is given.
> [0]PETSC ERROR: [0] VecView_MPI_DA line 508 src/dm/da/src/gr2.c
> [0]PETSC ERROR: [0] VecView line 690 src/vec/vec/interface/vector.c
> [0]PETSC ERROR: [0] DumpSolutionToMatlab line 1052 twqt2ff.c
> [0]PETSC ERROR: [0] Solve line 235 twqt2ff.c
> [0]PETSC ERROR: --------------------- Error Message
> ------------------------------------
> [0]PETSC ERROR: Signal received!
> [0]PETSC ERROR:
> ------------------------------------------------------------------------
> [0]PETSC ERROR: Petsc Development HG revision: unknown HG Date: unknown
> [0]PETSC ERROR: See docs/changes/index.html for recent updates.
> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting.
> [0]PETSC ERROR: See docs/index.html for manual pages.
> [0]PETSC ERROR:
> ------------------------------------------------------------------------
> [0]PETSC ERROR: ./twqt2ff.exe on a arch-bgp- named ionode132 by
> Unknown Fri Jan  2 00:44:26 1970
> [0]PETSC ERROR: Libraries linked from
> /scratch/rebeccaxyf/petsc-dev/arch-bgp-ibm-dbg/lib
> [0]PETSC ERROR: Configure run at Mon Apr 26 15:37:29 2010
> [0]PETSC ERROR: Configure options --with-cc=mpixlc_r
> --with-cxx=mpixlcxx_r --with-fc=mpixlf90_r --with-clanguage=cxx
> --with-blas-lapack-lib="-L/opt/share/math_libraries/lapack/ppc64/IBM/
> -llapack -lblas" --with-x=0 --with-is-color-value-type=short
> --with-shared=0 -CFLAGS="-qmaxmem=-1 -g" -CXXFLAGS="-qmaxmem=-1 -g"
> -FFLAGS="-qmaxmem=-1 -g" --with-debugging=1 --with-fortran-kernels=1
> --with-batch=1 --known-mpi-shared=0 --known-memcmp-ok
> --known-sizeof-char=1 --known-sizeof-void-p=4 --known-sizeof-short=2
> --known-sizeof-int=4 --known-sizeof-long=4 --known-sizeof-size_t=4
> --known-sizeof-long-long=8 --known-sizeof-float=4
> --known-sizeof-double=8 --known-bits-per-byte=8
> --known-sizeof-MPI_Comm=4 --known-sizeof-MPI_Fint=4
> --known-mpi-long-double=1 --known-level1-dcache-assoc=0
> --known-level1-dcache-linesize=32 --known-level1-dcache-size=32768
> --petsc-arch=bgp-dbg PETSC_ARCH=arch-bgp-ibm-dbg
> [0]PETSC ERROR:
> ------------------------------------------------------------------------
> [0]PETSC ERROR: User provided function() line 0 in unknown directory
> unknown file
> Abort(59) on node 0 (rank 0 in comm 1140850688): application called
> MPI_Abort(MPI_COMM_WORLD, 59) - process 0
>
>
> Am I wrong at configure?
>
> Thanks a lot!
>
> Rebecca
>
>
> --
> (Rebecca) Xuefei YUAN
> Department of Applied Physics and Applied Mathematics
> Columbia University
> Tel:917-399-8032
> www.columbia.edu/~xy2102
>
>
>
>


More information about the petsc-users mailing list