[petsc-dev] [pflotran-dev: 976] Re: [pflotran-users: 998] Distorted structured grid

Paolo Orsini paolo.orsini at gmail.com
Sun Nov 3 05:53:22 CST 2013


Hi Jed,

I found out what the problem is.
First of all I printed all the matrices out, to be sure that MatConvert was
working fine.
And that was ok: AIJ_mat had the same connectivity table than Adj_mat.
Then I compared the dual graph (stored in DualMat) computed by the Parmetis
routine (MatMeshToCellGraph) and with MatMatTransposeMult. And they were
quite different.

The computation of the dual graph of the mesh is a bit more complicated
than multiplying the adjacency matrix by its transpose, but not far off.
With this operation, also the cells that share only one node are connected
in the dual graph.
Instead, the minimum number of common nodes is >1 (2 in 2D probelms, 3 in
3D problems). In fact, this is an input of MatMeshToCellGraph, I should
have understood this before.

This can be computed doing the transpose adjacency matrix (Adj_T), then
doing the multiplication line by line of Adj time Adj_T, and discard the
non zero entries coming from to elements that share a number of nodes less
than  the minimum number of common nodes imposed. I have not implement this
yet, any suggestion is welcome.

I also found out that Schotch has a facility to compute a dual graph from a
mesh, but not PTScotch.
Once the graph is computed, PTSchotch can load the central dual graph, and
distribute it into several processors during the loading.
Am i right to say that PETSC is interfaced only with PTSchotch and not with
Scotch?

To check if the PTSchotch partition works (within PFLOTRAN ), I am
computing a DualMat with parmetis, saving it into a file. Then I recompile
the code (using a petsc compiled with ptscotch), an load the DualMat from a
file rather then forming a new one. I did a successful test when running on
one processor. but I am having trouble when try on more.

I though the the dual graph was computed only once, even during the mpi
process, instead it seems to be recomputed more than once. Not sure why....
sure i am missing something ???
 See below the original call from PFLOTRAN:

xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx
!!! this is the pass where the adjacency matrix is filled
  num_cells_local_old = unstructured_grid%nlmax
  allocate(local_vertices(unstructured_grid%max_nvert_per_cell* &
                          num_cells_local_old))
  allocate(local_vertex_offset(num_cells_local_old+1))
  local_vertices = 0
  local_vertex_offset = 0
  count = 0
  local_vertex_offset(1) = 0
  do local_id = 1, num_cells_local_old
    do ivertex = 1, unstructured_grid%max_nvert_per_cell
      if (unstructured_grid%cell_vertices(ivertex,local_id) == 0) exit
      count = count + 1
      ! local vertices must be zero-based for MatCreateMPIAdj; thus
subtract 1
      local_vertices(count) = &
        unstructured_grid%cell_vertices(ivertex,local_id) - 1
    enddo
    local_vertex_offset(local_id+1) = count
  enddo

  select case (unstructured_grid%grid_type)
    case(TWO_DIM_GRID)
      num_common_vertices = 2 ! cells must share at least this number of
vertices
    case(THREE_DIM_GRID)
      num_common_vertices = 3 ! cells must share at least this number of
vertices
    case default
        option%io_buffer = 'Grid type not recognized '
        call printErrMsg(option)
    end select

  ! determine the global offset from 0 for cells on this rank
  global_offset_old = 0
  call MPI_Exscan(num_cells_local_old,global_offset_old, &
                  ONE_INTEGER_MPI,MPIU_INTEGER,MPI_SUM,option%mycomm,ierr)

  !! the adjacency matrix is computed
  call MatCreateMPIAdj(option%mycomm,num_cells_local_old, &
                       unstructured_grid%num_vertices_global, &
                       local_vertex_offset, &
                       local_vertices,PETSC_NULL_INTEGER,Adj_mat,ierr)

!! the mesh dual graph is computed
#if defined(PETSC_HAVE_PARMETIS)
   call MatMeshToCellGraph(Adj_mat,num_common_vertices,Dual_mat,ierr)
#endif
  call MatDestroy(Adj_mat,ierr)

  call UGridPartition(unstructured_grid,option,Dual_mat,is_new, &
                      num_cells_local_new)

  if (allocated(local_vertices)) deallocate(local_vertices)
  if (allocated(local_vertex_offset)) deallocate(local_vertex_offset)

xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx

Paolo




On Sat, Nov 2, 2013 at 3:57 AM, Jed Brown <jedbrown at mcs.anl.gov> wrote:

> Paolo Orsini <paolo.orsini at gmail.com> writes:
>
> > Hi Jed,
> >
> > I believe you are right.
> > Multiplying the connectivity matrix by its tranpose returns an ncell x
> > ncell matrix with non-zero entries only for those cell sharing at least
> one
> > node. Hence the Dual Graph we need....
> >
> > Following your suggestions I change the code as follows:
> >
> >
> !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
> > call MatCreateMPIAdj(option%mycomm,num_cells_local_old, &
> >                        unstructured_grid%num_vertices_global, &
> >                        local_vertex_offset, &
> >                        local_vertices,PETSC_NULL_INTEGER,Adj_mat,ierr)
> >
> > call MatConvert(Adj_mat, MATAIJ,MAT_INITIAL_MATRIX,AIJ_mat,ierr)
> >
> > call
> >
> MatMatTransposeMult(AIJ_mat,AIJ_mat,MAT_INITIAL_MATRIX,PETSC_DEFAULT_DOUBLE_PRECISION,Dual_mat,ierr)
> > !!  call MatMeshToCellGraph(Adj_mat,num_common_vertices,Dual_mat,ierr)
> >
> >   call MatDestroy(Adj_mat,ierr)
> >   call MatDestroy(AIJ_mat,ierr)
> >
> !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
> >
> > When I try to run, the MatConvert takes ages, (I suppose I can form
> > directly the AIJ_mat),
>
> I see MatConvert_Basic does not even try to preallocate, which is
> certainly why it takes a long time.  That's easy to fix in PETSc (at
> least for AIJ).
>
> PETSc devs, why doesn't it make one pass counting and another pass
> inserting?
>
> > the PTScotch partitioning seems quick, but then I get an error later
> > on in the code, when attempting to compute the internal
> > connection.... Something is wrong...
>
> Always send the entire error message.
>
> > I need to double check... do you see anything wrong so far?
> >
> > P
> >
> >
> >
> >
> >
> >
> > On Fri, Nov 1, 2013 at 10:18 PM, Jed Brown <jedbrown at mcs.anl.gov> wrote:
> >
> >> Please keep the discussion on at least one list so it can be archived
> >> and so others can comment.
> >>
> >> Paolo Orsini <paolo.orsini at gmail.com> writes:
> >>
> >> > I really don't know how the DualGraph is going to affect the the
> parallel
> >> > computation efficiency...
> >> > I am trying to use this partitioning as a black box, not sure what's
> the
> >> > right thing to do..
> >> >
> >> > In the code calling the partitioning (PFLOTRAN), i can see that
> MPIAdj is
> >> > built, then the Dual graph is computed, and finally the partitioning
> >> takes
> >> > place, see below:
> >> >
> >> >
> >>
> !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
> >> >   call MatCreateMPIAdj(option%mycomm,num_cells_local_old, &
> >> >                        unstructured_grid%num_vertices_global, &
> >> >                        local_vertex_offset, &
> >> >                        local_vertices,PETSC_NULL_INTEGER,Adj_mat,ierr)
> >> >
> >> > call MatMeshToCellGraph(Adj_mat,num_common_vertices,Dual_mat,ierr)
> >>
> >> You can MatConvert Adj_mat to AIJ or assemble as AIJ to begin with, then
> >> use MatMatTransposeMult to create Dual_mat (which will be in AIJ
> >> format).
> >>
> >> > .........
> >> >
> >> >   call MatPartitioningCreate(option%mycomm,Part,ierr)
> >> >   call MatPartitioningSetAdjacency(Part,Dual_mat,ierr)
> >> >   call MatPartitioningSetFromOptions(Part,ierr)
> >> >   call MatPartitioningApply(Part,is_new,ierr)
> >> >   call MatPartitioningDestroy(Part,ierr)
> >> >
> >> >   allocate(cell_counts(option%mycommsize))
> >> >   call ISPartitioningCount(is_new,option%mycommsize,cell_counts,ierr)
> >> >   num_cells_local_new = cell_counts(option%myrank+1)
> >> >   call
> >> > MPI_Allreduce(num_cells_local_new,iflag,ONE_INTEGER_MPI,MPIU_INTEGER,
> &
> >> >                      MPI_MIN,option%mycomm,ierr)
> >> >
> >>
> !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!
> >> >
> >> > You are telling me we can build a dual graph simply multiplying
> Adj_mat
> >> by
> >> > its transpose?
> >>
> >> Yeah, doesn't it connect all cells that share at least one vertex?  That
> >> is exactly the nonzero pattern of the matrix product.
> >>
> >> > Why using a simple Aij when we are pre-processing a parallel process?
> >>
> >> MPIAIJ implements MatMatTransposeMult, but we don't have an
> >> implementation for MPIAdj (without calling ParMETIS).
> >>
> >> > Sorry, may be what I want to do is less black box than thought...
> >> >
> >> > And thanks again for your patience and help
> >> >
> >> > Paolo
> >> >
> >> >
> >> >
> >> >
> >> > On Fri, Nov 1, 2013 at 9:17 PM, Jed Brown <jedbrown at mcs.anl.gov>
> wrote:
> >> >
> >> >> Paolo Orsini <paolo.orsini at gmail.com> writes:
> >> >>
> >> >> > Hi Jed,
> >> >> >
> >> >> > Thanks for your prompt reply.
> >> >> >
> >> >> > Is it possible to do the partitioning with PTSchotch, without
> >> >> > using MatMeshToCellGraph?
> >> >> >
> >> >> > Is there any other function (not using ParMetis) to compute the
> >> adjacency
> >> >> > graph needed by MatPartitioningSetAdjacency, for parallel
> unstructured
> >> >> > grids?
> >> >>
> >> >> You can either compute the dual graph yourself or we can implement
> that
> >> >> function without calling ParMETIS.  Actually, I believe the
> operation is
> >> >> just symbolic
> >> >>
> >> >>   DualGraph = E * E^T
> >> >>
> >> >> where E is the element connectivity matrix.  So you could build a
> normal
> >> >> (AIJ instead of MPIAdj) matrix with the connectivity, then call
> >> >> MatMatTransposeMult().
> >> >>
> >> >> Would that work for you?
> >> >>
> >>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-dev/attachments/20131103/1facc64d/attachment.html>


More information about the petsc-dev mailing list