[petsc-users] segfault running src/dm/examples/tutorials/ex11f90.F

Panruo Wu pwu at mymail.mines.edu
Fri Jun 8 09:43:52 CDT 2012


The problem still persists with petsc-3.3.

I've send the bug report to peetsc-maint.

Thanks & regards
Panruo Wu

On Thu, Jun 7, 2012 at 10:45 PM, Barry Smith <bsmith at mcs.anl.gov> wrote:

>
>   Please switch to petsc-3.3 (just released)
> http://www.mcs.anl.gov/petsc/download/index.html
>
>   If the problem persists then send configure.log make.log and all the
> output from running the example (with petsc-3.3) to
> petsc-maint at mcs.anl.gov
>
>   Barry
>
> On Jun 5, 2012, at 10:54 PM, Panruo Wu wrote:
>
> > Dear list,
> >
> > I run into a segfault when running src/dm/examples/tutorials/ex11f90.F
> >
> > Intel Fortran Compiler 12.1.3; 1.6 OpenMPI; Ubuntu 12.04
> > other examples compiles and runs well.
> >
> > It compiles without any warnings or errors; however it cannot run.
> > Any ideas?  Error message are attached.
> >
> > Thank you!
> > Panruo Wu
> >
> > $ mpirun -np 1 ex11f90
> > [0]PETSC ERROR:
> ------------------------------------------------------------------------
> > [0]PETSC ERROR: Caught signal number 11 SEGV: Segmentation Violation,
> probably memory access out of range
> > [0]PETSC ERROR: Try option -start_in_debugger or
> -on_error_attach_debugger
> > [0]PETSC ERROR: or see
> http://www.mcs.anl.gov/petsc/petsc-as/documentation/faq.html#valgrind[0]PETSCERROR: or try
> http://valgrind.org on GNU/linux and Apple Mac OS X to find memory
> corruption errors
> > [0]PETSC ERROR: likely location of problem given in stack below
> > [0]PETSC ERROR: ---------------------  Stack Frames
> ------------------------------------
> > [0]PETSC ERROR: Note: The EXACT line numbers in the stack are not
> available,
> > [0]PETSC ERROR:       INSTEAD the line number of the start of the
> function
> > [0]PETSC ERROR:       is given.
> > [0]PETSC ERROR: --------------------- Error Message
> ------------------------------------
> > [0]PETSC ERROR: Signal received!
> > [0]PETSC ERROR:
> ------------------------------------------------------------------------
> > [0]PETSC ERROR: Petsc Release Version 3.2.0, Patch 7, Thu Mar 15
> 09:30:51 CDT 2012
> > [0]PETSC ERROR: See docs/changes/index.html for recent updates.
> > [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting.
> > [0]PETSC ERROR: See docs/index.html for manual pages.
> > [0]PETSC ERROR:
> ------------------------------------------------------------------------
> > [0]PETSC ERROR: ex11f90 on a opmi_inte named avahome by pwu Tue Jun  5
> 21:44:11 2012
> > [0]PETSC ERROR: Libraries linked from
> /home/pwu/Downloads/petsc-3.2-p7/opmi_intel_dbg/lib
> > [0]PETSC ERROR: Configure run at Tue Jun  5 10:02:00 2012
> > [0]PETSC ERROR: Configure options
> --with-blas-lapack-dir=/opt/intel/mkl/lib/intel64 --download-umfpack
> --download-superlu --with-x=1
> > [0]PETSC ERROR:
> ------------------------------------------------------------------------
> > [0]PETSC ERROR: User provided function() line 0 in unknown directory
> unknown file
> >
> --------------------------------------------------------------------------
> > MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD
> > with errorcode 59.
> >
> > NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
> > You may or may not see output from other processes, depending on
> > exactly when Open MPI kills them.
> >
> --------------------------------------------------------------------------
> >
> --------------------------------------------------------------------------
> > mpirun has exited due to process rank 0 with PID 22293 on
> > node avahome exiting improperly. There are two reasons this could occur:
> >
> > 1. this process did not call "init" before exiting, but others in
> > the job did. This can cause a job to hang indefinitely while it waits
> > for all processes to call "init". By rule, if one process calls "init",
> > then ALL processes must call "init" prior to termination.
> >
> > 2. this process called "init", but exited without calling "finalize".
> > By rule, all processes that call "init" MUST call "finalize" prior to
> > exiting or it will be considered an "abnormal termination"
> >
> > This may have caused other processes in the application to be
> > terminated by signals sent by mpirun (as reported here).
> >
> --------------------------------------------------------------------------
> >
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20120608/84c2bff0/attachment.html>


More information about the petsc-users mailing list