[petsc-users] partial stencil in DMDA?

Barry Smith bsmith at mcs.anl.gov
Fri Feb 21 15:48:13 CST 2014


On Feb 21, 2014, at 2:42 PM, Fischer, Greg A. <fischega at westinghouse.com> wrote:

> Barry,
> 
> If I'm interpreting correctly, the 2nd case would work fine for me. However, perhaps I was previously over-complicating the problem. Could I not just:
> 
> 1. create a DMDA
> 2. create a global vector with DOF=1
> 3. call VecDuplicateVecs() against the DOF=1 global vector for my other ~100 DOFs
> 
> and then independently manage the duplicated vectors with the array returned by VecDuplicateVecs()?

   You can do this and depending on your needs it can be fine. It really depends on what computations you are doing on the vectors. We refer the two different ways of storing the data: x0 y0 z0 x1 y1 z1 …. xn yn zn (where dof is 3 in this case as interlaced storage and x0 x1 … xn      y0 y1 … yn   z0 z1 … zn  as noninterlaced.
Interlaced can be much faster if you have calculations that involve say x_p y_p z_p together (since they are all stored together in memory you get good use of cache and cache lines)   if you have calculations that involve some x then later some y then later some z then non interlaced is better.  If you have a calculation that involves x_p y_p ….. z_p where there are 100 dof then just loading these values will take 100 cache lines and be inefficient and interlaced is better.
> 
> Also, when you mentioned changes to DMSetUp_DA_2D() in your original reply, you're suggesting that I could modify the PETSc source code, not use some pre-existing interface, correct?

   Yes

> 
> Thanks,
> Greg
> 
>> -----Original Message-----
>> From: Barry Smith [mailto:bsmith at mcs.anl.gov]
>> Sent: Friday, February 21, 2014 3:20 PM
>> To: Fischer, Greg A.
>> Cc: petsc-users at mcs.anl.gov
>> Subject: Re: [petsc-users] partial stencil in DMDA?
>> 
>> 
>>  Greg,
>> 
>>    The general mechanism for moving elements of vector between processes
>> is with VecScatterCreate() and VecScatterBegin/End() with it you can indicate
>> exactly what values are to move and to where, but it is all based on "one
>> dimension" indexing into the vector.
>> 
>>   DMDA provides a subset of communication for "structured meshes"
>> (essentially 1,2, 3 dimensional arrays split among processes with horizontal
>> and vertical cuts). It is pretty much all or nothing in terms of communicating
>> neighbor information.
>> 
>>   VecCreateGhost() uses the VecScatter mechanism to set up a SINGLE
>> communication pattern between a Vec and its ghosted partner Vec.
>> 
>>  Based on your email: "However, when communicating ghost values, not all
>> of those degrees of freedom should be exchanged at once." it sounds like
>> you need several (many) communication patterns requiring different ghost
>> entries in each.
>> 
>> 1) The hard general case: if for each set of ghost points you may need several
>> entries from one grid point and a different number of entries from another
>> grid point VecScatterCreate() (called multiple times, one for each pattern)
>> and VecScatterBegin/End() are intended for this purpose. However if you are
>> working with a 2 dimension grid/array you want to do ghosting for you need
>> to initially map your ghosting patterns from the 2d indexing to the 1d
>> indexing of the VecScatterCreate() which is a bit painful. You can see the
>> routine I mentioned before DMSetUp_DA_2D() but it would need to be
>> heavily modified.
>> 
>> 2) The easy case: if, for example, you need just the first index from each
>> point communicated as a ghost and then next the second and then next the
>> third you can avoid all custom communication patterns and just create 2
>> DMDA, one with a dof of 100 (or what ever it is for your case) and one with
>> dof of 1 then use VecStrideGather to pull out the one component of the
>> global vector (with the dof of 100) into a Vec obtained from
>> DMCreateGlobalVector() from the dof of 1 DMDA then use the
>> DMGlobalToLocalBegin/End on the 1 dof DMDA and now you have your single
>> ghost point at each point vector that you want.
>> 
>>  Barry
>> 
>> 
>> 
>> 
>> On Feb 21, 2014, at 1:12 PM, Fischer, Greg A. <fischega at westinghouse.com>
>> wrote:
>> 
>>> Barry,
>>> 
>>> Thanks!  I have another question. The user manual says:
>>> 
>>>                PETSc currently provides no container for multiple arrays sharing
>> the same distributed array communication; note, however, that the dof
>> parameter handles many cases of interest.
>>> 
>>> In my application, each space location will have on the order of hundreds of
>> values associated with it (which I believe translates to dof=O(100) - I don't
>> see the "degrees of freedom" explicitly defined anywhere). However, when
>> communicating ghost values, not all of those degrees of freedom should be
>> exchanged at once. I need to be able to exchange one at a time.
>>> 
>>> It sounds like what I may want to do is use VecCreateGhost(), which would
>> allow me to define exactly where the ghost points are, and then duplicate
>> that vector using VecDuplicateVecs() for each DOF. I can then scatter the
>> vectors individually as the need arises. Does that sound reasonable?
>>> 
>>> Greg
>>> 
>>>> -----Original Message-----
>>>> From: Barry Smith [mailto:bsmith at mcs.anl.gov]
>>>> Sent: Friday, February 21, 2014 12:21 PM
>>>> To: Fischer, Greg A.
>>>> Cc: petsc-users at mcs.anl.gov
>>>> Subject: Re: [petsc-users] partial stencil in DMDA?
>>>> 
>>>> 
>>>> On Feb 21, 2014, at 11:02 AM, Fischer, Greg A.
>>>> <fischega at westinghouse.com>
>>>> wrote:
>>>> 
>>>>> Hello,
>>>>> 
>>>>> I'm interested in using PETSc to manage distributed arrays. Based
>>>>> on my
>>>> reading about the DMDA objects, I see that ghost points can be
>>>> communicated in box-type stencils or star-type stencils.
>>>>> 
>>>>> For my application, assume that I have a 2D DMDA object with
>>>>> star-type
>>>> stencils. For a typical local calculation, I only need to access
>>>> ghost values from two of the four directions at a time. For example,
>>>> I'd like access to ghost values in the South and East directions, but
>>>> not in the North or West directions. Communicating North and West
>>>> data would seem to be wasting bandwidth. Is there any way to accomplish
>> this?
>>>> 
>>>>  Greg,
>>>> 
>>>>   There is not anything built in. Here is what I suggest:
>>>> 
>>>> 1)  write your application code not worrying about the fact that the
>>>> DMGlobalToLocalBegin/End() is moving values you don't need.
>>>> 
>>>> 2) when your code is running correctly for your problem and giving
>>>> useful results if the communication times are impacting how long it
>>>> takes to run you can provide a custom communication pattern. It would
>>>> involve little additional coding essentially taking DMSetUp_DA_2D()
>>>> which creates the list of ghost points and removing the  unneeded
>>>> ghost points.  But it would be premature to do this optimization
>>>> until you have a full working application code.
>>>> 
>>>>  Barry
>>>> 
>>>>> 
>>>>> Thanks,
>>>>> Greg
>>>> 
>>>> 
>> 
>> 
> 
> 



More information about the petsc-users mailing list