[petsc-users] Ordering of preallocation and OwnershipRange

Barry Smith bsmith at mcs.anl.gov
Sat Aug 13 17:23:52 CDT 2011


   The PetscLayout object is used to manage default layout among processes so you can create a PetscLayout object and use that to determine how the Mat and Vec objects will be laid out. From the manual page

    PetscLayoutCreate - Allocates PetscLayout space and sets the map contents to the default.

    Collective on MPI_Comm

   Input Parameters:
+    comm - the MPI communicator
-    map - pointer to the map

   Level: developer

    Notes: Typical calling sequence
       PetscLayoutCreate(MPI_Comm,PetscLayout *);
       PetscLayoutSetBlockSize(PetscLayout,1);
       PetscLayoutSetSize(PetscLayout,n) or PetscLayoutSetLocalSize(PetscLayout,N);
       PetscLayoutSetUp(PetscLayout);
       PetscLayoutGetSize(PetscLayout,PetscInt *); or PetscLayoutGetLocalSize(PetscLayout,PetscInt *;)
       PetscLayoutDestroy(PetscLayout);

      The PetscLayout object and methods are intended to be used in the PETSc Vec and Mat implementions; it is 
      recommended they not be used in user codes unless you really gain something in their use.

    Fortran Notes: 
      Not available from Fortran


On Aug 13, 2011, at 10:28 AM, Matthew Knepley wrote:

> On Sat, Aug 13, 2011 at 2:57 PM, Josh Hykes <jmhykes at ncsu.edu> wrote:
> Hello,
> 
> I'm just starting to experiment with PETSc (v3.1), and I like the Python bindings provided by petsc4py (v1.1.2). So far things seem fairly straightforward, but I'm stumped on a small issue.
> 
> While creating a parallel AIJ matrix, I'd like to preallocate it using arrays d_nnz and o_nnz. As I understand it, these arrays correspond to the processor's local rows.
> 
> Currently I specify the global matrix size, and let PETSc decide on the decomposition of the rows. I'd like to ask PETSc what rows each processor has with the getOwnershipRange() function, and then do the preallocation. However, according to the error message
> 
> > [1] MatAnyAIJSetPreallocation() line 393 in petsc4py-1.1.2/src/include/custom.h
> > [1] Operation done in wrong order
> > [1] matrix is already preallocated
> 
> I'm not allowed to do it in this order.
> 
> Thus, my question is: is it possible to let PETSc figure out the row decomposition while still using d_nnz and o_nnz for the preallocation? I figure that I could resolve the problem by doing my own decomposition, but it'd be nice if I could let those details up to PETSc.
> 
> You are correct. We require that preallocation is done at the same time as decomposition. There
> are tricky dependencies in matrix creation. However, an easy workaround is to create a Vec at
> the same time with the same global size, since it is guaranteed to have the same layout. I will look
> into simplifying this if it is possible.
> 
>   Thanks,
> 
>      Matt
>  
> I'm including an example using petsc4py of what I'd like to do, run with 2 MPI processes.
> 
> I apologize if this is a dumb question. Thank you for your help.
> 
> -Josh
> 
> # -----------------------------------------------
> from petsc4py import PETSc as petsc
> 
> M, N = 4, 6
> 
> global_d_nnz = [2, 1, 1, 2]
> global_o_nnz = [1, 3, 2, 1]
> 
> A = petsc.Mat()
> A.create(petsc.COMM_WORLD)
> A.setSizes([M, N])
> A.setType('aij')
> 
> i_start, i_end = A.getOwnershipRange()
> 
> A.setPreallocationNNZ([global_d_nnz[i_start:i_end],
>                        global_o_nnz[i_start:i_end]]) # error occurs here
> 
> 
> 
> 
> -- 
> What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead.
> -- Norbert Wiener



More information about the petsc-users mailing list