Parallelization of KSPSolve() in multidimensions
Barry Smith
bsmith at mcs.anl.gov
Wed May 20 12:39:12 CDT 2009
This is not universal behavior. We generally divide in all three
dimensions.
It is possibly a function of your machine and its network. You
need a good network for sparse iterative solvers, no type of ethernet
(no matter how good it claims to be) is suitable.
You can send to petsc-maint at mcs.anl.gov the -log_summary output
from a one process run, a slice in the last direction run and a 3d
decomposition run and we'll take a look at it.
Barry
On May 20, 2009, at 10:28 AM, Rolf Kuiper wrote:
> Hi PETSc-Users,
>
> when solving an implicit equation with KSPSolve() in 3D
> (communication with 7-point-stencil) I experienced the following:
> Parallelization of the e.g. 64 x 64 x 64 domain on n cpus in the
> last direction (every cpu has a 64 x 64 x 64/n subdomain) leads to a
> parallel efficiency of approximately 90%, which is fine for us.
> Parallelization of the e.g. 64 x 64 x 64 domain on n cpus in more
> than one direction (every cpu has e.g. a 64 x 64/sqrt(n) x 64/
> sqrt(n) subdomain) leads to a parallel efficiency of approximately
> 10%, which is absolutely unusable.
>
> Is this behavior generally true for this kind of solver? If so, why?
> If not: What did I do wrong most presumably?
> Has anybody made the same experience and/or could help me?
>
> Thanks in advance,
> Rolf
More information about the petsc-users
mailing list