[petsc-dev] TS ex11 crashing in parallel
Adrian Croucher
a.croucher at auckland.ac.nz
Tue May 27 23:19:10 CDT 2014
hi
The TS ex11 example (from 'next' branch) runs OK for me in serial, but
crashes when I run it in parallel (for any number of processors > 1).
Output from serial and parallel runs is below. Any clues as to what has
gone wrong?
Cheers, Adrian
---------------------------------------------------------------------------
1) Serial:
acro018 at des108:~/software/PETSc/code/src/ts/examples/tutorials$ mpirun
-np 1 ex11 -f ${PETSC_DIR}/share/petsc/datafiles/meshes/sevensi
de-quad.exo -advect_sol_type BUMP
0 time 0 |x| 0.2425
1 time 0.3311 |x| 0.1401
2 time 0.6623 |x| 0.07485
3 time 0.9934 |x| 0.0455
4 time 1.325 |x| 0.0316
5 time 1.656 |x| 0.02646
6 time 1.987 |x| 0.02069
7 time 2.318 |x| 0.01547
CONVERGED_TIME at time 2.31791 after 7 steps
2) Same thing in parallel:
acro018 at des108:~/software/PETSc/code/src/ts/examples/tutorials$ mpirun
-np 2 ex11 -f ${PETSC_DIR}/share/petsc/datafiles/meshes/sevensi
de-quad.exo -advect_sol_type BUMP
0 time 0 |x| 0.2425
[0]PETSC ERROR: --------------------- Error Message
--------------------------------------------------------------
[0]PETSC ERROR: Argument out of range
[0]PETSC ERROR: Point 0 has 0 constraints > -2 dof
[0]PETSC ERROR: See http://www.mcs.anl.gov/petsc/documentation/faq.html
for trouble shooting.
[0]PETSC ERROR: Petsc Development GIT revision: v3.4.4-5803-gabde81b
GIT Date: 2014-05-26 14:52:58 -0500
[0]PETSC ERROR: ex11 on a linux-gnu-c-opt named des108 by acro018 Wed
May 28 16:14:26 2014
[1]PETSC ERROR: --------------------- Error Message
--------------------------------------------------------------
[1]PETSC ERROR: Argument out of range
[1]PETSC ERROR: Point 3 has 0 constraints > -2 dof
[1]PETSC ERROR: See http://www.mcs.anl.gov/petsc/documentation/faq.html
for trouble shooting.
[1]PETSC ERROR: Petsc Development GIT revision: v3.4.4-5803-gabde81b
GIT Date: 2014-05-26 14:52:58 -0500
[1]PETSC ERROR: ex11 on a linux-gnu-c-opt named des108 by acro018 Wed
May 28 16:14:26 2014
[1]PETSC ERROR: Configure options --download-netcdf --download-exodusii
--with-hdf5-dir=/usr/lib --with-numpy-dir=/usr/lib/pymodules/p
ython2.7/numpy
--with-scientificpython-dir=/usr/lib/python2.7/dist-packages/scipy
--with-petsc4py-dir=/usr/local/lib/python2.7/dist-pa
ckages/petsc4py --download-triangle --with-fortran-interfaces=1
--download-ptscotch --download-chaco --download-ctetgen
[1]PETSC ERROR: #1 DMCreateDefaultSF() line 3066 in
/home/acro018/software/PETSc/code/src/dm/interface/dm.c
[1]PETSC ERROR: #2 DMGetDefaultSF() line 2986 in
/home/acro018/software/PETSc/code/src/dm/interface/dm.c
[1]PETSC ERROR: #3 DMGlobalToLocalBegin() line 1646 in
/home/acro018/software/PETSc/code/src/dm/interface/dm.c
[1]PETSC ERROR: #4 VecView_Plex() line 118 in
/home/acro018/software/PETSc/code/src/dm/impls/plex/plex.c
[1]PETSC ERROR: #5 VecView() line 601 in
/home/acro018/software/PETSc/code/src/vec/vec/interface/vector.c
[1]PETSC ERROR: #6 MonitorVTK() line 1395 in
/home/acro018/software/PETSc/code/src/ts/examples/tutorials/ex11.c
[1]PETSC ERROR: [0]PETSC ERROR: Configure options --download-netcdf
--download-exodusii --with-hdf5-dir=/usr/lib --with-numpy-dir=/usr
/lib/pymodules/python2.7/numpy
--with-scientificpython-dir=/usr/lib/python2.7/dist-packages/scipy
--with-petsc4py-dir=/usr/local/lib/p
ython2.7/dist-packages/petsc4py --download-triangle
--with-fortran-interfaces=1 --download-ptscotch --download-chaco
--download-ctetge
n
[0]PETSC ERROR: #1 DMCreateDefaultSF() line 3066 in
/home/acro018/software/PETSc/code/src/dm/interface/dm.c
[0]PETSC ERROR: #2 DMGetDefaultSF() line 2986 in
/home/acro018/software/PETSc/code/src/dm/interface/dm.c
[0]PETSC ERROR: #7 TSMonitor() line 2810 in
/home/acro018/software/PETSc/code/src/ts/interface/ts.c
[1]PETSC ERROR: #8 TSSolve() line 2752 in
/home/acro018/software/PETSc/code/src/ts/interface/ts.c
[1]PETSC ERROR: #9 main() line 1541 in
/home/acro018/software/PETSc/code/src/ts/examples/tutorials/ex11.c
[1]PETSC ERROR: ----------------End of Error Message -------send entire
error message to petsc-maint at mcs.anl.gov----------
#3 DMGlobalToLocalBegin() line 1646 in
/home/acro018/software/PETSc/code/src/dm/interface/dm.c
[0]PETSC ERROR: #4 VecView_Plex() line 118 in
/home/acro018/software/PETSc/code/src/dm/impls/plex/plex.c
[0]PETSC ERROR: #5 VecView() line 601 in
/home/acro018/software/PETSc/code/src/vec/vec/interface/vector.c
[0]PETSC ERROR: #6 MonitorVTK() line 1395 in
/home/acro018/software/PETSc/code/src/ts/examples/tutorials/ex11.c
[0]PETSC ERROR: #7 TSMonitor() line 2810 in
/home/acro018/software/PETSc/code/src/ts/interface/ts.c
[0]PETSC ERROR: #8 TSSolve() line 2752 in
/home/acro018/software/PETSc/code/src/ts/interface/ts.c
[0]PETSC ERROR: #9 main() line 1541 in
/home/acro018/software/PETSc/code/src/ts/examples/tutorials/ex11.c
[0]PETSC ERROR: ----------------End of Error Message -------send entire
error message to petsc-maint at mcs.anl.gov----------
--------------------------------------------------------------------------
MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD
with errorcode 63.
NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
You may or may not see output from other processes, depending on
exactly when Open MPI kills them.
--------------------------------------------------------------------------
--------------------------------------------------------------------------
mpirun has exited due to process rank 1 with PID 606 on
node des108 exiting without calling "finalize". This may
have caused other processes in the application to be
terminated by signals sent by mpirun (as reported here).
--------------------------------------------------------------------------
[des108:00604] 1 more process has sent help message help-mpi-api.txt /
mpi-abort
[des108:00604] Set MCA parameter "orte_base_help_aggregate" to 0 to see
all help / error messages
--
Dr Adrian Croucher
Senior Research Fellow
Department of Engineering Science
University of Auckland, New Zealand
email: a.croucher at auckland.ac.nz
tel: +64 (0)9 923 4611
More information about the petsc-dev
mailing list