[petsc-dev] [pflotran-dev: 1080] Re: [pflotran-users: 998] Distorted structured grid

Gautam Bisht gbisht at lbl.gov
Sun Nov 3 09:40:32 CST 2013


Hi Paolo,


On Sun, Nov 3, 2013 at 3:53 AM, Paolo Orsini <paolo.orsini at gmail.com> wrote:

> Hi Jed,
>
> I found out what the problem is.
> First of all I printed all the matrices out, to be sure that MatConvert
> was working fine.
> And that was ok: AIJ_mat had the same connectivity table than Adj_mat.
> Then I compared the dual graph (stored in DualMat) computed by the
> Parmetis routine (MatMeshToCellGraph) and with MatMatTransposeMult. And
> they were quite different.
>
> The computation of the dual graph of the mesh is a bit more complicated
> than multiplying the adjacency matrix by its transpose, but not far off.
> With this operation, also the cells that share only one node are connected
> in the dual graph.
> Instead, the minimum number of common nodes is >1 (2 in 2D probelms, 3 in
> 3D problems). In fact, this is an input of MatMeshToCellGraph, I should
> have understood this before.
>
> This can be computed doing the transpose adjacency matrix (Adj_T), then
> doing the multiplication line by line of Adj time Adj_T, and discard the
> non zero entries coming from to elements that share a number of nodes less
> than  the minimum number of common nodes imposed. I have not implement this
> yet, any suggestion is welcome.
>
> I also found out that Schotch has a facility to compute a dual graph from
> a mesh, but not PTScotch.
> Once the graph is computed, PTSchotch can load the central dual graph, and
> distribute it into several processors during the loading.
> Am i right to say that PETSC is interfaced only with PTSchotch and not
> with Scotch?
>
> To check if the PTSchotch partition works (within PFLOTRAN ), I am
> computing a DualMat with parmetis, saving it into a file. Then I recompile
> the code (using a petsc compiled with ptscotch), an load the DualMat from a
> file rather then forming a new one. I did a successful test when running on
> one processor. but I am having trouble when try on more.
>
> I though the the dual graph was computed only once, even during the mpi
> process, instead it seems to be recomputed more than once. Not sure why....
> sure i am missing something ???
>

In PFLOTRAN, MatCreateMPIAdj() is called:
- Once, if unstructured grid is specified in IMPLICIT format [in
unstructured_grid.F90];
- Twice, if unstructured grid is specified in EXPLICIT format [in
unstructured_grid.F90]. The first call creates a Adjacency matrix based on
'CONNECTIONS' data of the explicit grid; while the second call is used to
create a Dual matrix.

In order to better understand the code, you could compile the code with
'ugrid_debug=1' and look at the Adj*.out/Dual*.out for the two runs:
- regression_tests/default/discretization/mixed_explicit.in
- regression_tests/default/discretization/mixed_implicit.in

-Gautam.



>  See below the original call from PFLOTRAN:
>
> xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx
> !!! this is the pass where the adjacency matrix is filled
>   num_cells_local_old = unstructured_grid%nlmax
>   allocate(local_vertices(unstructured_grid%max_nvert_per_cell* &
>                           num_cells_local_old))
>   allocate(local_vertex_offset(num_cells_local_old+1))
>   local_vertices = 0
>   local_vertex_offset = 0
>   count = 0
>   local_vertex_offset(1) = 0
>   do local_id = 1, num_cells_local_old
>     do ivertex = 1, unstructured_grid%max_nvert_per_cell
>       if (unstructured_grid%cell_vertices(ivertex,local_id) == 0) exit
>       count = count + 1
>       ! local vertices must be zero-based for MatCreateMPIAdj; thus
> subtract 1
>       local_vertices(count) = &
>         unstructured_grid%cell_vertices(ivertex,local_id) - 1
>     enddo
>     local_vertex_offset(local_id+1) = count
>   enddo
>
>   select case (unstructured_grid%grid_type)
>     case(TWO_DIM_GRID)
>       num_common_vertices = 2 ! cells must share at least this number of
> vertices
>     case(THREE_DIM_GRID)
>       num_common_vertices = 3 ! cells must share at least this number of
> vertices
>     case default
>         option%io_buffer = 'Grid type not recognized '
>         call printErrMsg(option)
>     end select
>
>   ! determine the global offset from 0 for cells on this rank
>   global_offset_old = 0
>   call MPI_Exscan(num_cells_local_old,global_offset_old, &
>                   ONE_INTEGER_MPI,MPIU_INTEGER,MPI_SUM,option%mycomm,ierr)
>
>   !! the adjacency matrix is computed
>   call MatCreateMPIAdj(option%mycomm,num_cells_local_old, &
>                         unstructured_grid%num_vertices_global, &
>                        local_vertex_offset, &
>                        local_vertices,PETSC_NULL_INTEGER,Adj_mat,ierr)
>
> !! the mesh dual graph is computed
> #if defined(PETSC_HAVE_PARMETIS)
>    call MatMeshToCellGraph(Adj_mat,num_common_vertices,Dual_mat,ierr)
> #endif
>   call MatDestroy(Adj_mat,ierr)
>
>   call UGridPartition(unstructured_grid,option,Dual_mat,is_new, &
>                       num_cells_local_new)
>
>   if (allocated(local_vertices)) deallocate(local_vertices)
>   if (allocated(local_vertex_offset)) deallocate(local_vertex_offset)
>
> xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx
>
> Paolo
>
>
>
>
> On Sat, Nov 2, 2013 at 3:57 AM, Jed Brown <jedbrown at mcs.anl.gov> wrote:
>
>> Paolo Orsini <paolo.orsini at gmail.com> writes:
>>
>> > Hi Jed,
>> >
>> > I believe you are right.
>> > Multiplying the connectivity matrix by its tranpose returns an ncell x
>> > ncell matrix with non-zero entries only for those cell sharing at least
>> one
>> > node. Hence the Dual Graph we need....
>> >
>> > Following your suggestions I change the code as follows:
>> >
>> >
>> !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
>> > call MatCreateMPIAdj(option%mycomm,num_cells_local_old, &
>> >                        unstructured_grid%num_vertices_global, &
>> >                        local_vertex_offset, &
>> >                        local_vertices,PETSC_NULL_INTEGER,Adj_mat,ierr)
>> >
>> > call MatConvert(Adj_mat, MATAIJ,MAT_INITIAL_MATRIX,AIJ_mat,ierr)
>> >
>> > call
>> >
>> MatMatTransposeMult(AIJ_mat,AIJ_mat,MAT_INITIAL_MATRIX,PETSC_DEFAULT_DOUBLE_PRECISION,Dual_mat,ierr)
>> > !!  call MatMeshToCellGraph(Adj_mat,num_common_vertices,Dual_mat,ierr)
>> >
>> >   call MatDestroy(Adj_mat,ierr)
>> >   call MatDestroy(AIJ_mat,ierr)
>> >
>> !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
>> >
>> > When I try to run, the MatConvert takes ages, (I suppose I can form
>> > directly the AIJ_mat),
>>
>> I see MatConvert_Basic does not even try to preallocate, which is
>> certainly why it takes a long time.  That's easy to fix in PETSc (at
>> least for AIJ).
>>
>> PETSc devs, why doesn't it make one pass counting and another pass
>> inserting?
>>
>> > the PTScotch partitioning seems quick, but then I get an error later
>> > on in the code, when attempting to compute the internal
>> > connection.... Something is wrong...
>>
>> Always send the entire error message.
>>
>> > I need to double check... do you see anything wrong so far?
>> >
>> > P
>> >
>> >
>> >
>> >
>> >
>> >
>> > On Fri, Nov 1, 2013 at 10:18 PM, Jed Brown <jedbrown at mcs.anl.gov>
>> wrote:
>> >
>> >> Please keep the discussion on at least one list so it can be archived
>> >> and so others can comment.
>> >>
>> >> Paolo Orsini <paolo.orsini at gmail.com> writes:
>> >>
>> >> > I really don't know how the DualGraph is going to affect the the
>> parallel
>> >> > computation efficiency...
>> >> > I am trying to use this partitioning as a black box, not sure what's
>> the
>> >> > right thing to do..
>> >> >
>> >> > In the code calling the partitioning (PFLOTRAN), i can see that
>> MPIAdj is
>> >> > built, then the Dual graph is computed, and finally the partitioning
>> >> takes
>> >> > place, see below:
>> >> >
>> >> >
>> >>
>> !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
>> >> >   call MatCreateMPIAdj(option%mycomm,num_cells_local_old, &
>> >> >                        unstructured_grid%num_vertices_global, &
>> >> >                        local_vertex_offset, &
>> >> >
>>  local_vertices,PETSC_NULL_INTEGER,Adj_mat,ierr)
>> >> >
>> >> > call MatMeshToCellGraph(Adj_mat,num_common_vertices,Dual_mat,ierr)
>> >>
>> >> You can MatConvert Adj_mat to AIJ or assemble as AIJ to begin with,
>> then
>> >> use MatMatTransposeMult to create Dual_mat (which will be in AIJ
>> >> format).
>> >>
>> >> > .........
>> >> >
>> >> >   call MatPartitioningCreate(option%mycomm,Part,ierr)
>> >> >   call MatPartitioningSetAdjacency(Part,Dual_mat,ierr)
>> >> >   call MatPartitioningSetFromOptions(Part,ierr)
>> >> >   call MatPartitioningApply(Part,is_new,ierr)
>> >> >   call MatPartitioningDestroy(Part,ierr)
>> >> >
>> >> >   allocate(cell_counts(option%mycommsize))
>> >> >   call ISPartitioningCount(is_new,option%mycommsize,cell_counts,ierr)
>> >> >   num_cells_local_new = cell_counts(option%myrank+1)
>> >> >   call
>> >> >
>> MPI_Allreduce(num_cells_local_new,iflag,ONE_INTEGER_MPI,MPIU_INTEGER, &
>> >> >                      MPI_MIN,option%mycomm,ierr)
>> >> >
>> >>
>> !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
>> >> >
>> >> > You are telling me we can build a dual graph simply multiplying
>> Adj_mat
>> >> by
>> >> > its transpose?
>> >>
>> >> Yeah, doesn't it connect all cells that share at least one vertex?
>>  That
>> >> is exactly the nonzero pattern of the matrix product.
>> >>
>> >> > Why using a simple Aij when we are pre-processing a parallel process?
>> >>
>> >> MPIAIJ implements MatMatTransposeMult, but we don't have an
>> >> implementation for MPIAdj (without calling ParMETIS).
>> >>
>> >> > Sorry, may be what I want to do is less black box than thought...
>> >> >
>> >> > And thanks again for your patience and help
>> >> >
>> >> > Paolo
>> >> >
>> >> >
>> >> >
>> >> >
>> >> > On Fri, Nov 1, 2013 at 9:17 PM, Jed Brown <jedbrown at mcs.anl.gov>
>> wrote:
>> >> >
>> >> >> Paolo Orsini <paolo.orsini at gmail.com> writes:
>> >> >>
>> >> >> > Hi Jed,
>> >> >> >
>> >> >> > Thanks for your prompt reply.
>> >> >> >
>> >> >> > Is it possible to do the partitioning with PTSchotch, without
>> >> >> > using MatMeshToCellGraph?
>> >> >> >
>> >> >> > Is there any other function (not using ParMetis) to compute the
>> >> adjacency
>> >> >> > graph needed by MatPartitioningSetAdjacency, for parallel
>> unstructured
>> >> >> > grids?
>> >> >>
>> >> >> You can either compute the dual graph yourself or we can implement
>> that
>> >> >> function without calling ParMETIS.  Actually, I believe the
>> operation is
>> >> >> just symbolic
>> >> >>
>> >> >>   DualGraph = E * E^T
>> >> >>
>> >> >> where E is the element connectivity matrix.  So you could build a
>> normal
>> >> >> (AIJ instead of MPIAdj) matrix with the connectivity, then call
>> >> >> MatMatTransposeMult().
>> >> >>
>> >> >> Would that work for you?
>> >> >>
>> >>
>>
>
>  --
> You received this message because you are subscribed to the Google Groups
> "pflotran-dev" group.
> To view this discussion on the web visit
> https://groups.google.com/d/msgid/pflotran-dev/CACGnotZfbdvEieXFn%2B6RYzj4o4yM_%3DK0NkHcFeK5CESL%2BovcKw%40mail.gmail.com
> .
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-dev/attachments/20131103/c5903b5b/attachment.html>


More information about the petsc-dev mailing list