[petsc-dev] boomerAmg scalability
Ravi Kannan
rxk at cfdrc.com
Thu Jan 5 17:13:35 CST 2012
Hi Mark,
Files are attached.
Thanks,
Ravi
From: Ravi Kannan [mailto:rxk at cfdrc.com]
Sent: Thursday, January 05, 2012 5:13 PM
To: 'For users of the development version of PETSc'
Subject: RE: [petsc-dev] boomerAmg scalability
Hi Mark,
From: petsc-dev-bounces at mcs.anl.gov [mailto:petsc-dev-bounces at mcs.anl.gov]
On Behalf Of Mark F. Adams
Sent: Wednesday, January 04, 2012 12:09 PM
To: For users of the development version of PETSc
Subject: Re: [petsc-dev] boomerAmg scalability
Is your problem unsymmetric? I don't think this is the problem, it was
fixed recently.
If not, can you run with -pc_gamg_verbose and -info and send me the output?
I have attached the verbose+info outputs for both the serial and the
parallel (2 partitions). NOTE: the serial output at some location says
PC=Jacobi! Is it implicitly converting the PC to a Jacobi?
You don't need to set the block size if it is 1.
I have a feeling it has to do with not setting the buffer sizes correctly.
There are some heuristics in there and this needs to be improved. What kind
of discretization are you using? What order, how many non-zeros per row?
Maybe try the smallest case that you have with more processors. I have a
feeling it is trashing around mallocing stuff in MatSetValues so it just
"looks" hung.
Mark
Thanks,
Ravi.
On Jan 4, 2012, at 12:14 PM, Ravi Kannan wrote:
Hi Mark, Matt,
We recently downloaded the petsc development version, to test the gamg
package.
This works in serial : we tried for small cases. The parallel case (even
with 2 partitions) just hangs. As of now, we have not set any parameters. So
I guess that the default parameters are being used.
Do we need to explicitly set the block size to MatSetBlockSize(mat,1) for a
PARALLEL run? Our solver solves U,V,W and P separately.
Any input on this would be great.
Thanks,
Ravi.
From: petsc-dev-bounces at mcs.anl.gov [mailto:petsc-dev-bounces at mcs.anl.gov]
On Behalf Of Mark F. Adams
Sent: Thursday, December 15, 2011 1:18 PM
To: For users of the development version of PETSc
Subject: Re: [petsc-dev] boomerAmg scalability
On Dec 15, 2011, at 1:56 PM, Matthew Knepley wrote:
On Thu, Dec 15, 2011 at 10:23 AM, Ravi Kannan <rxk at cfdrc.com> wrote:
Dear All,
This is Ravi Kannan from CFD Research Corporation. Recently, we are
experimenting with the BoomerAMG preconditioner for some "stiff" CFD
problems. In that regard, all the other standard solver-preconditioner
combinations failed for the current CFD problem. The boomer is the only one
which is able to provide with "converged" solutions.
We noticed that the scalability of this boomer preconditioner is really
poor. For instance, even with a cell size of 2 million, we cannot scale to
even 16 partitions (in contrast, the other solver-preconditioner
combinations like the BI-CGS/BJacobi gave good enough scalability).
Are we missing something? Do we need to use a more latest version of boomer?
Have you tried -pc_type gamg in petsc-dev?
For gamg you also want to use MPIAIJ matrices and set the block size
MatSetBlockSize(mat,3), for a 3D velocity field, for instance. You can also
try '-pc_gamg_type pa' or '-pc_gamg_type sa'. "pa", for plain aggregation
might be better for CFD problems.
Mark
Matt
Thanks,
Ravi.
--
What most experimenters take for granted before they begin their experiments
is infinitely more interesting than any results to which their experiments
lead.
-- Norbert Wiener
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-dev/attachments/20120105/24746587/attachment.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: info-verbose-parallel
Type: application/octet-stream
Size: 25200 bytes
Desc: not available
URL: <http://lists.mcs.anl.gov/pipermail/petsc-dev/attachments/20120105/24746587/attachment.obj>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: info-verbose-serial
Type: application/octet-stream
Size: 26378 bytes
Desc: not available
URL: <http://lists.mcs.anl.gov/pipermail/petsc-dev/attachments/20120105/24746587/attachment-0001.obj>
More information about the petsc-dev
mailing list