[petsc-users] memory access out of range with MUMPS
Jose E. Roman
jroman at dsic.upv.es
Mon Apr 16 10:49:18 CDT 2018
If you are going to run in parallel then you should not provide the communicator comm=PETSc.COMM_SELF in any object.
Jose
> El 16 abr 2018, a las 17:42, Yongxiang Wu <yongxiang27 at gmail.com> escribió:
>
> Hello everyone,
> I am using slepc4py solving a generalized eigenvalue problem, with a large spase matrix. The following is my python code. The matrix is first created using PETSc.Mat().createAIJ() and saved to file matrix-A.dat .
>
>
> opts = PETSc.Options()
> opts.setValue('-st_pc_factor_mat_solver_package','mumps')
>
> viewer = PETSc.Viewer().createBinary('matrix-A.dat', 'r',comm=PETSc.COMM_SELF)
> A = PETSc.Mat().create(comm=PETSc.COMM_SELF)
> A.load(viewer)
> A.setFromOptions()
> A.setUp()
>
> Print('load petsc binary finished')
>
> t0 = time()
>
> E = SLEPc.EPS().create(comm=PETSc.COMM_SELF)
> E.setOperators(A,None)
> E.setDimensions(noEigs2solv,PETSc.DECIDE) # set number of eigenvalues to compute
> E.setWhichEigenpairs(E.Which.TARGET_MAGNITUDE)
> E.setProblemType(SLEPc.EPS.ProblemType.GNHEP) # generalized non-Hermitian eigenvalue problem
>
> E.setTolerances(arnoldiTol,1000) # set tolerance
> if SIGMA is not None:
> E.setTarget(SIGMA) # set the desired eigenvalue
>
> if options_v0 is not None:
> E.setInitialSpace(options_v0) # set the initial vector
>
> st = E.getST()
> st.setType('sinvert')
>
> E.setFromOptions()
> Print('Start solving ...')
> E.solve()
>
> My problem here is, when I run this code on cluster, I would end up with the following error. Strange thing is that I used the save library on my local PC, and it works without any error. If any one of you could give me hint what is wrong? Thanks
>
>
> ** ERROR RETURN ** FROM ZMUMPS INFO(1)= -1
> ** INFO(2)= 0
> [0]PETSC ERROR: ------------------------------------------------------------------------
> [0]PETSC ERROR: Caught signal number 11 SEGV: Segmentation Violation, probably memory access out of range
> [0]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger
> [0]PETSC ERROR: or see http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind
> [0]PETSC ERROR: or try http://valgrind.org on GNU/linux and Apple Mac OS X to find memory corruption errors
> [0]PETSC ERROR: likely location of problem given in stack below
> [0]PETSC ERROR: --------------------- Stack Frames ------------------------------------
> [0]PETSC ERROR: Note: The EXACT line numbers in the stack are not available,
> [0]PETSC ERROR: INSTEAD the line number of the start of the function
> [0]PETSC ERROR: is given.
> [0]PETSC ERROR: [0] PetscInitializeMUMPS line 1188 /zhome/academic/HLRS/iag/iagyonwu/.local/lib/petsc/src/mat/impls/aij/mpi/mumps/mumps.c
> [0]PETSC ERROR: [0] MatGetFactor_aij_mumps line 2125 /zhome/academic/HLRS/iag/iagyonwu/.local/lib/petsc/src/mat/impls/aij/mpi/mumps/mumps.c
> [0]PETSC ERROR: [0] MatGetFactor line 4341 /zhome/academic/HLRS/iag/iagyonwu/.local/lib/petsc/src/mat/interface/matrix.c
> [0]PETSC ERROR: [0] PCSetUp_LU line 59 /zhome/academic/HLRS/iag/iagyonwu/.local/lib/petsc/src/ksp/pc/impls/factor/lu/lu.c
> [0]PETSC ERROR: [0] PCSetUp line 886 /zhome/academic/HLRS/iag/iagyonwu/.local/lib/petsc/src/ksp/pc/interface/precon.c
> [0]PETSC ERROR: [0] KSPSetUp line 294 /zhome/academic/HLRS/iag/iagyonwu/.local/lib/petsc/src/ksp/ksp/interface/itfunc.c
> [0]PETSC ERROR: [0] STSetUp_Sinvert line 96 /zhome/academic/HLRS/iag/iagyonwu/.local/lib/slepc-3.8.2/src/sys/classes/st/impls/sinvert/sinvert.c
> [0]PETSC ERROR: [0] STSetUp line 233 /zhome/academic/HLRS/iag/iagyonwu/.local/lib/slepc-3.8.2/src/sys/classes/st/interface/stsolve.c
> [0]PETSC ERROR: [0] EPSSetUp line 104 /zhome/academic/HLRS/iag/iagyonwu/.local/lib/slepc-3.8.2/src/eps/interface/epssetup.c
> [0]PETSC ERROR: [0] EPSSolve line 129 /zhome/academic/HLRS/iag/iagyonwu/.local/lib/slepc-3.8.2/src/eps/interface/epssolve.c
> application called MPI_Abort(MPI_COMM_WORLD, 59) - process 0
> [unset]: write_line error; fd=-1 buf=:cmd=abort exitcode=59
>
> with regards
> Yong
More information about the petsc-users
mailing list