[petsc-users] MatMPIAIJSetPreallocation: "nnz cannot be greater than row length"

David Knezevic david.knezevic at akselos.com
Sun Feb 22 16:09:26 CST 2015


Hi Barry,

Thanks for your help, much appreciated.

I added a prototype for MatDisAssemble_MPIAIJ:
PETSC_INTERN PetscErrorCode MatDisAssemble_MPIAIJ(Mat);

and I added a call to MatDisAssemble_MPIAIJ before
MatMPIAIJSetPreallocation. However, I get a segfault on the call
to MatDisAssemble_MPIAIJ. The segfault occurs in both serial and parallel.

FYI, I'm using Petsc 3.5.2, and I'm not using a non-debug build (though I
could rebuild PETSc in debug mode if you think that would help figure out
what's happening here).

Thanks,
David



On Sun, Feb 22, 2015 at 1:13 PM, Barry Smith <bsmith at mcs.anl.gov> wrote:


>   David,
>
>    This is an obscure little feature of MatMPIAIJ,   each time you change
> the sparsity pattern before you call the MatMPIAIJSetPreallocation you need
> to call  MatDisAssemble_MPIAIJ(Mat mat).    This is a private PETSc
> function so you need to provide your own prototype for it above the
> function you use it in.
>
>   Let us know if this resolves the problem.
>
>    Barry
>
> We never really intended that people would call
> MatMPIAIJSetPreallocation() AFTER they had already used the matrix.
>
>
> > On Feb 22, 2015, at 6:50 AM, David Knezevic <david.knezevic at akselos.com>
> wrote:
> >
> > Hi all,
> >
> > I've implemented a solver for a contact problem using SNES. The sparsity
> pattern of the jacobian matrix needs to change at each nonlinear iteration
> (because the elements which are in contact can change), so I tried to deal
> with this by calling MatSeqAIJSetPreallocation and
> MatMPIAIJSetPreallocation during each iteration in order to update the
> preallocation.
> >
> > This seems to work fine in serial, but with two or more MPI processes I
> run into the error "nnz cannot be greater than row length", e.g.:
> > nnz cannot be greater than row length: local row 528 value 12 rowlength 0
> >
> > This error is from the call to
> > MatSeqAIJSetPreallocation(b->B,o_nz,o_nnz); in
> MatMPIAIJSetPreallocation_MPIAIJ.
> >
> > Any guidance on what the problem might be would be most appreciated. For
> example, I was wondering if there is a problem with calling
> SetPreallocation on a matrix that has already been preallocated?
> >
> > Some notes:
> > - I'm using PETSc via libMesh
> > - The code that triggers this issue is available as a PR on the libMesh
> github repo, in case anyone is interested:
> https://github.com/libMesh/libmesh/pull/460/
> > - I can try to make a minimal pure-PETSc example that reproduces this
> error, if that would be helpful.
> >
> > Many thanks,
> > David
> >
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20150222/7d2ac01c/attachment.html>


More information about the petsc-users mailing list