[petsc-users] h-FGMRES
Gaetan Kenway
gaetank at gmail.com
Mon May 20 08:51:24 CDT 2013
Hi again
It runs if the mattype is mpiaij instead of mpibaij. I gather this is not
implemented for the blocked matrix types?
Gaetan
On Mon, May 20, 2013 at 9:26 AM, Gaetan Kenway <gaetank at gmail.com> wrote:
> Hi again
>
> I installed petsc3.4.0 and I am still getting the following error when
> running with the following options (on 64 procs)
>
> # Matrix Options
> -matload_block_size 5 -mat_type mpibaij
>
> # KSP solver options
> -ksp_type fgmres -ksp_max_it 1000 -ksp_gmres_restart 200 -ksp_monitor
> -ksp_view -ksp_pc_side right -ksp_rtol 1e-6
>
> # Nested GMRES Options
> -pc_type bjacobi -pc_bjacobi_blocks 4 -sub_ksp_type gmres -sub_ksp_max_it
> 5 -sub_pc_type bjacobi -sub_sub_pc_type ilu
> -sub_sub_pc_factor_mat_ordering_type rcm -sub_sub_pc_factor_levels 1
>
> Any thoughts?
>
> Thank you,
>
> Gaetan
>
> [44]PETSC ERROR:
> ------------------------------------------------------------------------
> [44]PETSC ERROR: Caught signal number 11 SEGV: Segmentation Violation,
> probably memory access out of range
> [44]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger
> [44]PETSC ERROR: or see
> http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind[44]PETSCERROR: or try
> http://valgrind.org on GNU/linux and Apple Mac OS X to find memory
> corruption errors
> [44]PETSC ERROR: likely location of problem given in stack below
> [44]PETSC ERROR: --------------------- Stack Frames
> ------------------------------------
> [44]PETSC ERROR: Note: The EXACT line numbers in the stack are not
> available,
> [44]PETSC ERROR: INSTEAD the line number of the start of the function
> [44]PETSC ERROR: is given.
> [44]PETSC ERROR: [44] PCSetUp_BJacobi_Multiproc line 1197
> /home/j/jmartins/kenway/packages/petsc-3.4.0/src/ksp/pc/impls/bjacobi/bjacobi.c
> [44]PETSC ERROR: [44] PCSetUp_BJacobi line 24
> /home/j/jmartins/kenway/packages/petsc-3.4.0/src/ksp/pc/impls/bjacobi/bjacobi.c
> [44]PETSC ERROR: [44] PCSetUp line 868
> /home/j/jmartins/kenway/packages/petsc-3.4.0/src/ksp/pc/interface/precon.c
> [44]PETSC ERROR: [44] KSPSetUp line 192
> /home/j/jmartins/kenway/packages/petsc-3.4.0/src/ksp/ksp/interface/itfunc.c
> [44]PETSC ERROR: [44] KSPSolve line 356
> /home/j/jmartins/kenway/packages/petsc-3.4.0/src/ksp/ksp/interface/itfunc.c
> [43]PETSC ERROR:
> ------------------------------------------------------------------------
>
>
> On Sun, May 19, 2013 at 11:15 PM, Barry Smith <bsmith at mcs.anl.gov> wrote:
>
>>
>> You should be using PETSc version 3.4 which was recently released and
>> is what the paper is based on.
>>
>> Barry
>>
>> On May 19, 2013, at 10:11 PM, Gaetan Kenway <gaetank at gmail.com> wrote:
>>
>> > Hi Everyone
>> >
>> > I am trying to replicate the type of preconditioner described in
>> "Hierarchical and Nested Krylov Methods for Extreme-Scale Computing".
>> >
>> > I have used the following options: (I'm using fortran so the following
>> is my petsc_options file)
>> >
>> > # Matrix Options
>> > -matload_block_size 5
>> > -mat_type mpibaij
>> >
>> > # KSP solver options
>> > -ksp_type gmres
>> > -ksp_max_it 1000
>> > -ksp_gmres_restart 200
>> > -ksp_monitor
>> > -ksp_view
>> > -ksp_pc_side right
>> > -ksp_rtol 1e-6
>> >
>> > # Nested GMRES Options
>> > -pc_type bjacobi
>> > -pc_bjacobi_blocks 4
>> > -sub_ksp_type gmres
>> > -sub_ksp_max_it 5
>> > -sub_pc_type bjacobi
>> > -sub_sub_pc_type ilu
>> > -sub_sub_pc_factor_mat_ordering_type rcm
>> > -sub_sub_pc_factor_levels 1
>> >
>> > The test is run on 64 processors and the total number of block jacobi
>> blocks is 4 (less than nproc). The error I get is:
>> >
>> > [6]PETSC ERROR: Note: The EXACT line numbers in the stack are not
>> available,
>> > [6]PETSC ERROR: INSTEAD the line number of the start of the
>> function
>> > [6]PETSC ERROR: is given.
>> > [6]PETSC ERROR: [6] PCSetUp_BJacobi_Multiproc line 1269
>> /home/j/jmartins/kenway/packages/petsc-3.3-p5/src/ksp/pc/impls/bjacobi/bjacobi.c
>> > [6]PETSC ERROR: [6] PCSetUp_BJacobi line 24
>> /home/j/jmartins/kenway/packages/petsc-3.3-p5/src/ksp/pc/impls/bjacobi/bjacobi.c
>> > [6]PETSC ERROR: [6] PCSetUp line 810
>> /home/j/jmartins/kenway/packages/petsc-3.3-p5/src/ksp/pc/interface/precon.c
>> > [6]PETSC ERROR: [6] KSPSetUp line 182
>> /home/j/jmartins/kenway/packages/petsc-3.3-p5/src/ksp/ksp/interface/itfunc.c
>> > [6]PETSC ERROR: [6] KSPSolve line 351
>> /home/j/jmartins/kenway/packages/petsc-3.3-p5/src/ksp/ksp/interface/itfunc.c
>> > [6]PETSC ERROR: --------------------- Error Message
>> ------------------------------------
>> > [6]PETSC ERROR: Signal received!
>> > [6]PETSC ERROR:
>> ------------------------------------------------------------------------
>> > [6]PETSC ERROR: Petsc Release Version 3.3.0, Patch 5, Sat Dec 1
>> 15:10:41 CST 2012
>> > [6]PETSC ERROR: See docs/changes/index.html for recent updates.
>> > [6]PETSC ERROR: See docs/faq.html for hints about trouble shooting.
>> > [6]PETSC ERROR:
>> ------------------------------------------------------------------------
>> > [6]PETSC ERROR:
>> ------------------------------------------------------------------------
>> > [6]PETSC ERROR: ./main on a intel-rea named gpc-f109n001 by kenway Sun
>> May 19 23:01:52 2013
>> > [6]PETSC ERROR: Libraries linked from
>> /home/j/jmartins/kenway/packages/petsc-3.3-p5/intel-real-debug/lib
>> > [6]PETSC ERROR: Configure run at Sun Jan 20 15:52:20 2013
>> > [6]PETSC ERROR: Configure options --with-shared-libraries
>> --download-superlu_dist=yes --download-parmetis=yes --download-metis=yes
>> --with-fortran-interfaces=1 --with-debugging=yes --with-scalar-type=real
>> -with-petsc-arch=intel-real-debug --with-blas-lapack-dir= --with-pic
>> > [6]PETSC ERROR:
>> ------------------------------------------------------------------------
>> >
>> > If the number of blocks is greater than or equal to the number of
>> processors it runs fine. I'm using version 3.3-p5.
>> >
>> > The options as listed in the paper are:
>> > -flow_ksp_type fgmres -flow_ksp_pc_side right -flow_pc_type bjacobi
>> -flow_pc_bjacobi_blocks ngp
>> > -flow_sub_ksp_type gmres -flow_sub_ksp_max_it 6 -flow_sub_pc_type
>> bjacobi
>> > -flow_sub_sub_pc_type ilu
>> >
>> > Any suggestions would be greatly appreciated.
>> >
>> > Thank you,
>> >
>> > Gaetan Kenway
>> >
>> >
>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20130520/5ff5a157/attachment.html>
More information about the petsc-users
mailing list