[petsc-dev] Generality of VecScatter

Mark F. Adams mark.adams at columbia.edu
Fri Nov 25 09:38:49 CST 2011


On Nov 24, 2011, at 2:12 PM, Jed Brown wrote:

> On Thu, Nov 24, 2011 at 12:40, Mark F. Adams <mark.adams at columbia.edu> wrote:
> OK, that makes sense at least abstractly.  I'm skeptical that you are going to be able to do much with all this one sided stuff; it seems to be that you would always have the two sided data at some point and then you might setup one sided maps for repeated operations like for a MatVec that are efficient.  But this is kind of trivial.
> 
> Yeah, that's easy. The update for MatMult() just involves each process calling one MPI_Get() for each remote process that it needs data from. What to get is encoded in an MPI_Datatype so there is no user-visible packing on either side. These calls are all non-blocking and you finish the update with MPI_Win_fence().
> 
> You can test performance of this approach using -vecscatter_window (although it's slightly different because it does user-visible packing).
> 
> Of course we can easily build a two-sided representation, but it's not necessary. In the one-sided model above, the owner of a global point does not know how many local parts accessed it (and nowhere inside the MPI implementation or network stack is there storage proportional to the number of accessors, so there isn't a scalability problem with globally coupled dofs).
>  
> 
> I would suggest writing a non-trivail algorithm with your model.  I have two ideas:
> 
> 1) when a coarse AMG (use aggregation MG to me concrete) grid is constructed and you have a map (bipartite graph) with one edge for each fine grid point to a coarse grid point (not necessarily on your processor).  I have too few vertices per processor so I want to reduce the number of processors used in the coarse grid.
> 
> Do you assume the availability of a partitioner at this stage,

No, lets keep it simple for now.

> or you just want to do something simple like rank-order aggregation?

Thats fine.

> In my model, the owner of a point would somehow (e.g. by calling a partitioner) determine the new owner of that point.

My point is that at some point the rubber has to hit the road and you need two side information.  You can then set up pointers for some fast one-sided stuff, and that is fine but you seem to agree that this is trivial.  Its not clear to me that your model is not simply removing some (redundant) data from the user space, while it is probably in the MPI library.

> This information is communicated to the ghosters of that point (again without knowing how many there are),

But you knew how many there were, and chose to forget it, and the MPI layer must know how many there are so it knows when its done communicating.  I'm not sure how you insure correctness with your model.

> any other node/connectivity data is communicated from old owner to new owner (this is a 1-1 relation, thus trivial with these primitives), and finally the ghosters address the new location when they need updates.
>  
>  I can assume that my fine grid was well partitioned, and to keep the process local assume the fine grid was partitioned with a space filling curve, or something, such that when I aggregate processors (eg, (0-3), (4-7) ... if I reduced the number of active processor by 4) I still have decent data locality.  Eventually I would like a local algorithm to improve load balance at this point, but that is more than you might want to start with.
> 
> If you have a local graph partitioner that can tell you what to move for better load balance, then I think I can move it efficiently and without redundant storage or specification.

How do you load balance with a local partitioner?

Mark

>  
> 
> 2) Any AMR algorithm that you like.  Again, you might want to punt on load balancing.
> 
> I think the primitives are fine for this. There is an algorithmic question of how to modify boundary zones with minimal communication, but if you decide how much you want to communicate when (e.g. on the chalk board), I will figure out how to implement it.
> 
> I think redistribution for load balancing is represented pretty elegantly, but it would be good to work out the details.

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-dev/attachments/20111125/c6caee99/attachment.html>


More information about the petsc-dev mailing list