[petsc-dev] TS ex11 crashing in parallel

Matthew Knepley knepley at gmail.com
Tue May 27 23:38:22 CDT 2014


On Tue, May 27, 2014 at 11:19 PM, Adrian Croucher <a.croucher at auckland.ac.nz
> wrote:

> hi
>
> The TS ex11 example (from 'next' branch) runs OK for me in serial, but
> crashes when I run it in parallel (for any number of processors > 1).
>
> Output from serial and parallel runs is below. Any clues as to what has
> gone wrong?
>

Yes,  I put an overaggressive check in. I will push the fix tonight.

  Thanks,

     Matt


> Cheers, Adrian
>
> ------------------------------------------------------------
> ---------------
>
> 1) Serial:
>
> acro018 at des108:~/software/PETSc/code/src/ts/examples/tutorials$ mpirun
> -np 1 ex11 -f ${PETSC_DIR}/share/petsc/datafiles/meshes/sevensi
> de-quad.exo -advect_sol_type BUMP
>   0  time        0  |x|   0.2425
>   1  time   0.3311  |x|   0.1401
>   2  time   0.6623  |x|  0.07485
>   3  time   0.9934  |x|   0.0455
>   4  time    1.325  |x|   0.0316
>   5  time    1.656  |x|  0.02646
>   6  time    1.987  |x|  0.02069
>   7  time    2.318  |x|  0.01547
> CONVERGED_TIME at time 2.31791 after 7 steps
>
> 2) Same thing in parallel:
>
> acro018 at des108:~/software/PETSc/code/src/ts/examples/tutorials$ mpirun
> -np 2 ex11 -f ${PETSC_DIR}/share/petsc/datafiles/meshes/sevensi
> de-quad.exo -advect_sol_type BUMP
>   0  time        0  |x|   0.2425
> [0]PETSC ERROR: --------------------- Error Message
> --------------------------------------------------------------
> [0]PETSC ERROR: Argument out of range
> [0]PETSC ERROR: Point 0 has 0 constraints > -2 dof
> [0]PETSC ERROR: See http://www.mcs.anl.gov/petsc/documentation/faq.htmlfor trouble shooting.
> [0]PETSC ERROR: Petsc Development GIT revision: v3.4.4-5803-gabde81b  GIT
> Date: 2014-05-26 14:52:58 -0500
> [0]PETSC ERROR: ex11 on a linux-gnu-c-opt named des108 by acro018 Wed May
> 28 16:14:26 2014
> [1]PETSC ERROR: --------------------- Error Message
> --------------------------------------------------------------
> [1]PETSC ERROR: Argument out of range
> [1]PETSC ERROR: Point 3 has 0 constraints > -2 dof
> [1]PETSC ERROR: See http://www.mcs.anl.gov/petsc/documentation/faq.htmlfor trouble shooting.
> [1]PETSC ERROR: Petsc Development GIT revision: v3.4.4-5803-gabde81b  GIT
> Date: 2014-05-26 14:52:58 -0500
> [1]PETSC ERROR: ex11 on a linux-gnu-c-opt named des108 by acro018 Wed May
> 28 16:14:26 2014
> [1]PETSC ERROR: Configure options --download-netcdf --download-exodusii
> --with-hdf5-dir=/usr/lib --with-numpy-dir=/usr/lib/pymodules/p
> ython2.7/numpy --with-scientificpython-dir=/usr/lib/python2.7/dist-packages/scipy
> --with-petsc4py-dir=/usr/local/lib/python2.7/dist-pa
> ckages/petsc4py --download-triangle --with-fortran-interfaces=1
> --download-ptscotch --download-chaco --download-ctetgen
> [1]PETSC ERROR: #1 DMCreateDefaultSF() line 3066 in
> /home/acro018/software/PETSc/code/src/dm/interface/dm.c
> [1]PETSC ERROR: #2 DMGetDefaultSF() line 2986 in
> /home/acro018/software/PETSc/code/src/dm/interface/dm.c
> [1]PETSC ERROR: #3 DMGlobalToLocalBegin() line 1646 in
> /home/acro018/software/PETSc/code/src/dm/interface/dm.c
> [1]PETSC ERROR: #4 VecView_Plex() line 118 in /home/acro018/software/PETSc/
> code/src/dm/impls/plex/plex.c
> [1]PETSC ERROR: #5 VecView() line 601 in /home/acro018/software/PETSc/
> code/src/vec/vec/interface/vector.c
> [1]PETSC ERROR: #6 MonitorVTK() line 1395 in /home/acro018/software/PETSc/
> code/src/ts/examples/tutorials/ex11.c
> [1]PETSC ERROR: [0]PETSC ERROR: Configure options --download-netcdf
> --download-exodusii --with-hdf5-dir=/usr/lib --with-numpy-dir=/usr
> /lib/pymodules/python2.7/numpy --with-scientificpython-dir=/
> usr/lib/python2.7/dist-packages/scipy --with-petsc4py-dir=/usr/local/lib/p
> ython2.7/dist-packages/petsc4py --download-triangle
> --with-fortran-interfaces=1 --download-ptscotch --download-chaco
> --download-ctetge
> n
> [0]PETSC ERROR: #1 DMCreateDefaultSF() line 3066 in
> /home/acro018/software/PETSc/code/src/dm/interface/dm.c
> [0]PETSC ERROR: #2 DMGetDefaultSF() line 2986 in
> /home/acro018/software/PETSc/code/src/dm/interface/dm.c
> [0]PETSC ERROR: #7 TSMonitor() line 2810 in /home/acro018/software/PETSc/
> code/src/ts/interface/ts.c
> [1]PETSC ERROR: #8 TSSolve() line 2752 in /home/acro018/software/PETSc/
> code/src/ts/interface/ts.c
> [1]PETSC ERROR: #9 main() line 1541 in /home/acro018/software/PETSc/
> code/src/ts/examples/tutorials/ex11.c
> [1]PETSC ERROR: ----------------End of Error Message -------send entire
> error message to petsc-maint at mcs.anl.gov----------
> #3 DMGlobalToLocalBegin() line 1646 in /home/acro018/software/PETSc/
> code/src/dm/interface/dm.c
> [0]PETSC ERROR: #4 VecView_Plex() line 118 in /home/acro018/software/PETSc/
> code/src/dm/impls/plex/plex.c
> [0]PETSC ERROR: #5 VecView() line 601 in /home/acro018/software/PETSc/
> code/src/vec/vec/interface/vector.c
> [0]PETSC ERROR: #6 MonitorVTK() line 1395 in /home/acro018/software/PETSc/
> code/src/ts/examples/tutorials/ex11.c
> [0]PETSC ERROR: #7 TSMonitor() line 2810 in /home/acro018/software/PETSc/
> code/src/ts/interface/ts.c
> [0]PETSC ERROR: #8 TSSolve() line 2752 in /home/acro018/software/PETSc/
> code/src/ts/interface/ts.c
> [0]PETSC ERROR: #9 main() line 1541 in /home/acro018/software/PETSc/
> code/src/ts/examples/tutorials/ex11.c
> [0]PETSC ERROR: ----------------End of Error Message -------send entire
> error message to petsc-maint at mcs.anl.gov----------
> --------------------------------------------------------------------------
> MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD
> with errorcode 63.
>
> NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
> You may or may not see output from other processes, depending on
> exactly when Open MPI kills them.
> --------------------------------------------------------------------------
> --------------------------------------------------------------------------
> mpirun has exited due to process rank 1 with PID 606 on
> node des108 exiting without calling "finalize". This may
> have caused other processes in the application to be
> terminated by signals sent by mpirun (as reported here).
> --------------------------------------------------------------------------
> [des108:00604] 1 more process has sent help message help-mpi-api.txt /
> mpi-abort
> [des108:00604] Set MCA parameter "orte_base_help_aggregate" to 0 to see
> all help / error messages
>
> --
> Dr Adrian Croucher
> Senior Research Fellow
> Department of Engineering Science
> University of Auckland, New Zealand
> email: a.croucher at auckland.ac.nz
> tel: +64 (0)9 923 4611
>
>


-- 
What most experimenters take for granted before they begin their
experiments is infinitely more interesting than any results to which their
experiments lead.
-- Norbert Wiener
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-dev/attachments/20140527/93284264/attachment.html>


More information about the petsc-dev mailing list