[petsc-users] Hang while attempting to run EPSSolve()
Satish Balay
balay at mcs.anl.gov
Fri Feb 13 16:47:51 CST 2015
I'll suggest 1.6
I belive 1.6, 1.8 etc are considered stable releases - and 1.5, 1.7
etc are considered development releases.
Satish
On Fri, 13 Feb 2015, Barry Smith wrote:
>
> I think it was introduced in 1.8.1 so I think 1.8.0 should be ok, but if it hangs then go back to 1.7.4
>
>
> > On Feb 13, 2015, at 4:43 PM, Andrew Spott <andrew at spott.us> wrote:
> >
> > What version of OpenMPI was this introduced in? I appear to be finding it in OpenMPI 1.8.3 and 1.8.1 as well. Should I go back to 1.8.0 or 1.7.4?
> >
> > Thanks,
> >
> > -Andrew
> >
> >
> >
> > On Fri, Feb 13, 2015 at 2:23 PM, Andrew Spott <andrew at spott.us> wrote:
> >
> > Thanks! You just saved me hours of debugging.
> >
> > I’ll look into linking against an earlier implementation of OpenMPI.
> >
> > -Andrew
> >
> >
> >
> > On Fri, Feb 13, 2015 at 2:21 PM, Barry Smith <bsmith at mcs.anl.gov> wrote:
> >
> >
> > Andrew,
> >
> > This is a bug in the 1.8.2 OpenMPI implementation they recently introduced. Can you link against an earlier OpenMPI implementation on the machine? Or do they have MPICH installed you could use?
> >
> > Barry
> >
> >
> >
> > > On Feb 13, 2015, at 3:17 PM, Andrew Spott <andrew at spott.us> wrote:
> > >
> > > Local tests on OS X can’t reproduce, but production tests on our local supercomputer always hang while waiting for a lock.
> > >
> > > The back trace:
> > >
> > > #0 0x00002ba2980df054 in __lll_lock_wait () from /lib64/libpthread.so.0
> > > #1 0x00002ba2980da388 in _L_lock_854 () from /lib64/libpthread.so.0
> > > #2 0x00002ba2980da257 in pthread_mutex_lock () from /lib64/libpthread.so.0
> > > #3 0x00002ba29a1d9e2c in ompi_attr_get_c () from /curc/tools/x_86_64/rh6/openmpi/1.8.2/gcc/4.9.1/lib/libmpi.so.1
> > > #4 0x00002ba29a207f8e in PMPI_Attr_get () from /curc/tools/x_86_64/rh6/openmpi/1.8.2/gcc/4.9.1/lib/libmpi.so.1
> > > #5 0x00002ba294aa111e in Petsc_DelComm_Outer () at /home/ansp6066/local/src/petsc-3.5.3/src/sys/objects/pinit.c:409
> > > #6 0x00002ba29a1dae02 in ompi_attr_delete_all () from /curc/tools/x_86_64/rh6/openmpi/1.8.2/gcc/4.9.1/lib/libmpi.so.1
> > > #7 0x00002ba29a1dcb6c in ompi_comm_free () from /curc/tools/x_86_64/rh6/openmpi/1.8.2/gcc/4.9.1/lib/libmpi.so.1
> > > #8 0x00002ba29a20c713 in PMPI_Comm_free () from /curc/tools/x_86_64/rh6/openmpi/1.8.2/gcc/4.9.1/lib/libmpi.so.1
> > > #9 0x00002ba294aba7cf in PetscSubcommCreate_contiguous(_n_PetscSubcomm*) () from /home/ansp6066/local/petsc-3.5.3-debug/lib/libpetsc.so.3.5
> > > #10 0x00002ba294ab89d5 in PetscSubcommSetType () from /home/ansp6066/local/petsc-3.5.3-debug/lib/libpetsc.so.3.5
> > > #11 0x00002ba2958ce437 in PCSetUp_Redundant(_p_PC*) () from /home/ansp6066/local/petsc-3.5.3-debug/lib/libpetsc.so.3.5
> > > #12 0x00002ba2957a243d in PCSetUp () at /home/ansp6066/local/src/petsc-3.5.3/src/ksp/pc/interface/precon.c:902
> > > #13 0x00002ba2958dea31 in KSPSetUp () at /home/ansp6066/local/src/petsc-3.5.3/src/ksp/ksp/interface/itfunc.c:306
> > > #14 0x00002ba29a7f8e70 in STSetUp_Sinvert(_p_ST*) () at /home/ansp6066/local/src/slepc-3.5.3/src/sys/classes/st/impls/sinvert/sinvert.c:145
> > > #15 0x00002ba29a7e92cf in STSetUp () at /home/ansp6066/local/src/slepc-3.5.3/src/sys/classes/st/interface/stsolve.c:301
> > > #16 0x00002ba29a845ea6 in EPSSetUp () at /home/ansp6066/local/src/slepc-3.5.3/src/eps/interface/epssetup.c:207
> > > #17 0x00002ba29a849f91 in EPSSolve () at /home/ansp6066/local/src/slepc-3.5.3/src/eps/interface/epssolve.c:88
> > > #18 0x0000000000410de5 in petsc::EigenvalueSolver::solve() () at /home/ansp6066/code/petsc_cpp_wrapper/src/petsc_cpp/EigenvalueSolver.cpp:40
> > > #19 0x00000000004065c7 in main () at /home/ansp6066/code/new_work_project/src/main.cpp:165
> > >
> > > This happens for MPI or single process runs. Does anyone have any hints on how I can debug this? I honestly have no idea.
> > >
> > > -Andrew
> > >
> >
> >
> >
>
>
More information about the petsc-users
mailing list