[petsc-users] On unknown ordering
Smith, Barry F.
bsmith at mcs.anl.gov
Thu Nov 15 12:16:15 CST 2018
> On Nov 15, 2018, at 4:48 AM, Appel, Thibaut via petsc-users <petsc-users at mcs.anl.gov> wrote:
>
> Good morning,
>
> I would like to ask about the importance of the initial choice of ordering the unknowns when feeding a matrix to PETSc.
>
> I have a regular grid, using high-order finite differences and I simply divide rows of the matrix with PetscSplitOwnership using vertex major, natural ordering for the parallelism (not using DMDA)
So each process is getting a slice of the domain? To minimize communication it is best to use "square-ish" subdomains instead of slices; this is why the DMDA tries to use "square-ish" subdomains. I don't know the relationship between convergence rate and the shapes of the subdomains, it will depend on the operator and possibly "flow direction" etc.
>
> My understanding is that when using LU-MUMPS, this does not matter because either serial or parallel analysis is performed and all the rows are reordered ‘optimally’ before the LU factorization. Quality of reordering might suffer from parallel analysis.
>
> But if I use the default block Jacobi with ILU with one block per processor, the initial ordering seems to have an influence because some tightly coupled degrees of freedom might lay on different processes and the ILU becomes less powerful. You can change the ordering on each block but this won’t necessarily make things better.
>
> Are my observations accurate? Is there a recommended ordering type for a block Jacobi approach in my case? Could I expect natural improvements in fill-in or better GMRES robustness opting for parallelism offered by DMDA?
You might consider using -pc_type asm (additive Schwarz method) instead of block Jacobi. This "reintroduces" some of the tight coupling that is discarded when slicing up the domain for block Jacobi.
Barry
>
> Thank you,
>
> Thibaut
More information about the petsc-users
mailing list