The problem still persists with petsc-3.3.<div><br></div><div>I've send the bug report to peetsc-maint.</div><div><br></div><div>Thanks & regards</div><div>Panruo Wu</div><div><br><div class="gmail_quote">On Thu, Jun 7, 2012 at 10:45 PM, Barry Smith <span dir="ltr"><<a href="mailto:bsmith@mcs.anl.gov" target="_blank">bsmith@mcs.anl.gov</a>></span> wrote:<br>
<blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex"><br>
Please switch to petsc-3.3 (just released) <a href="http://www.mcs.anl.gov/petsc/download/index.html" target="_blank">http://www.mcs.anl.gov/petsc/download/index.html</a><br>
<br>
If the problem persists then send configure.log make.log and all the output from running the example (with petsc-3.3) to <a href="mailto:petsc-maint@mcs.anl.gov">petsc-maint@mcs.anl.gov</a><br>
<span class="HOEnZb"><font color="#888888"><br>
Barry<br>
</font></span><div class="HOEnZb"><div class="h5"><br>
On Jun 5, 2012, at 10:54 PM, Panruo Wu wrote:<br>
<br>
> Dear list,<br>
><br>
> I run into a segfault when running src/dm/examples/tutorials/ex11f90.F<br>
><br>
> Intel Fortran Compiler 12.1.3; 1.6 OpenMPI; Ubuntu 12.04<br>
> other examples compiles and runs well.<br>
><br>
> It compiles without any warnings or errors; however it cannot run.<br>
> Any ideas? Error message are attached.<br>
><br>
> Thank you!<br>
> Panruo Wu<br>
><br>
> $ mpirun -np 1 ex11f90<br>
> [0]PETSC ERROR: ------------------------------------------------------------------------<br>
> [0]PETSC ERROR: Caught signal number 11 SEGV: Segmentation Violation, probably memory access out of range<br>
> [0]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger<br>
> [0]PETSC ERROR: or see <a href="http://www.mcs.anl.gov/petsc/petsc-as/documentation/faq.html#valgrind[0]PETSC" target="_blank">http://www.mcs.anl.gov/petsc/petsc-as/documentation/faq.html#valgrind[0]PETSC</a> ERROR: or try <a href="http://valgrind.org" target="_blank">http://valgrind.org</a> on GNU/linux and Apple Mac OS X to find memory corruption errors<br>
> [0]PETSC ERROR: likely location of problem given in stack below<br>
> [0]PETSC ERROR: --------------------- Stack Frames ------------------------------------<br>
> [0]PETSC ERROR: Note: The EXACT line numbers in the stack are not available,<br>
> [0]PETSC ERROR: INSTEAD the line number of the start of the function<br>
> [0]PETSC ERROR: is given.<br>
> [0]PETSC ERROR: --------------------- Error Message ------------------------------------<br>
> [0]PETSC ERROR: Signal received!<br>
> [0]PETSC ERROR: ------------------------------------------------------------------------<br>
> [0]PETSC ERROR: Petsc Release Version 3.2.0, Patch 7, Thu Mar 15 09:30:51 CDT 2012<br>
> [0]PETSC ERROR: See docs/changes/index.html for recent updates.<br>
> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting.<br>
> [0]PETSC ERROR: See docs/index.html for manual pages.<br>
> [0]PETSC ERROR: ------------------------------------------------------------------------<br>
> [0]PETSC ERROR: ex11f90 on a opmi_inte named avahome by pwu Tue Jun 5 21:44:11 2012<br>
> [0]PETSC ERROR: Libraries linked from /home/pwu/Downloads/petsc-3.2-p7/opmi_intel_dbg/lib<br>
> [0]PETSC ERROR: Configure run at Tue Jun 5 10:02:00 2012<br>
> [0]PETSC ERROR: Configure options --with-blas-lapack-dir=/opt/intel/mkl/lib/intel64 --download-umfpack --download-superlu --with-x=1<br>
> [0]PETSC ERROR: ------------------------------------------------------------------------<br>
> [0]PETSC ERROR: User provided function() line 0 in unknown directory unknown file<br>
> --------------------------------------------------------------------------<br>
> MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD<br>
> with errorcode 59.<br>
><br>
> NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.<br>
> You may or may not see output from other processes, depending on<br>
> exactly when Open MPI kills them.<br>
> --------------------------------------------------------------------------<br>
> --------------------------------------------------------------------------<br>
> mpirun has exited due to process rank 0 with PID 22293 on<br>
> node avahome exiting improperly. There are two reasons this could occur:<br>
><br>
> 1. this process did not call "init" before exiting, but others in<br>
> the job did. This can cause a job to hang indefinitely while it waits<br>
> for all processes to call "init". By rule, if one process calls "init",<br>
> then ALL processes must call "init" prior to termination.<br>
><br>
> 2. this process called "init", but exited without calling "finalize".<br>
> By rule, all processes that call "init" MUST call "finalize" prior to<br>
> exiting or it will be considered an "abnormal termination"<br>
><br>
> This may have caused other processes in the application to be<br>
> terminated by signals sent by mpirun (as reported here).<br>
> --------------------------------------------------------------------------<br>
><br>
<br>
</div></div></blockquote></div><br></div>