[petsc-users] LU factorization and solution of independent matrices does not scale, why?
Alexander Grayver
agrayver at gfz-potsdam.de
Fri Dec 21 10:00:10 CST 2012
Thomas,
I'm missing one point...
You run N sequential factorizations (i.e. each has its own matrix to
work with and no need to communicate?) independently within ONE node? Or
there are N factorizations that run on N nodes?
Jed,
> MUMPS uses MPI_Iprobe on MPI_COMM_WORLD (hard-coded).
Any reason they do it that way? Which part of the code is that (i.e.
analysis/factorization/solution.)?
Regards,
Alexander
On 21.12.2012 16:51, Thomas Witkowski wrote:
> I use a modified MPICH version. On the system I use for these
> benchmarks I cannot use another MPI library.
>
> I'm not fixed to MUMPS. Superlu_dist, for example, works also
> perfectly for this. But there is still the following problem I cannot
> solve: When I increase the number of coarse space matrices, there
> seems to be no scaling direct solver for this. Just to summaries:
> - one coarse space matrix is created always by one "cluster"
> consisting of four subdomanins/MPI tasks
> - the four tasks are always local to one node, thus inter-node network
> communication is not required for computing factorization and solve
> - independent of the number of cluster, the coarse space matrices are
> the same, have the same number of rows, nnz structure but possibly
> different values
> - there is NO load unbalancing
> - the matrices must be factorized and there are a lot of solves (>
> 100) with them
>
> It should be pretty clear, that computing LU factorization and solving
> with it should scale perfectly. But at the moment, all direct solver I
> tried (mumps, superlu_dist, pastix) are not able to scale. The loos of
> scale is really worse, as you can see from the numbers I send before.
>
> Any ideas? Suggestions? Without a scaling solver method for these kind
> of systems, my multilevel FETI-DP code is just more or less a joke,
> only some orders of magnitude slower than standard FETI-DP method :)
>
> Thomas
More information about the petsc-users
mailing list