[petsc-users] mumps solve with same nonzero pattern
Wen Jiang
jiangwen84 at gmail.com
Mon Apr 30 12:18:19 CDT 2012
Hi Hong,
Thanks for your reply. I check the attachment sent last time and I miss
some very information. I attached the complete info with this email. Sorry
for this.
I also tried other matrix ordering like mat_mumps_icntl_7 2 and I got the
similar performance. I checked the configuration log file on that cluster,
they use --download-f-blas-lapack=1 instead of using optimal BLAS. Would
this be the problem which cause the poor performance? Actually the mumps
runs quite slow, at14 Gflop/s, which is far from the machine's peak.
Thanks.
Wen
>
> Wen :
>
> > Reply to
> >
> >> This is weird. Try
> >> 1) increase work space with
> >> -mat_mumps_icntl_14 50 (default is 20)
> >> 2) different matrix orderings with
> >> -mat_mumps_icntl_7 2 (or number from 0 to 6)
> >>
> >> Run your code with '-log_summary' and see which routine causes this huge
> >> difference.
> >>
> > Why your '-log_summary' only gives
> KSPSolve 4 1.0 2.2645e+03 1.0 0.00e+00 0.0 3.9e+04 3.6e+02
> 5.4e+01 96 0 27 0 9 96 0 27 0 9 0
> PCSetUp 4 1.0 2.2633e+03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00
> 3.4e+01 96 0 0 0 6 96 0 0 0 6 0
> PCApply 4 1.0 1.1641e+00 1.0 0.00e+00 0.0 3.9e+04 3.6e+02
> 2.0e+01 0 0 27 0 3 0 0 27 0 3 0
>
> I get
> petsc-dev/src/ksp/ksp/examples/tutorials>mpiexec -n 2 ./ex2 -pc_type lu
> -pc_factor_mat_solver_package mumps -log_summary
> MatMult 2 1.0 1.6904e-04 1.0 4.44e+02 1.0 4.0e+00 5.6e+01
> 0.0e+00 0 47 25 13 0 0 47 33 13 0 5
> MatSolve 2 1.0 3.8259e-03 1.0 0.00e+00 0.0 8.0e+00 1.9e+02
> 6.0e+00 10 0 50 84 7 11 0 67 87 9 0
> MatLUFactorSym 1 1.0 2.9058e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00
> 8.0e+00 7 0 0 0 9 8 0 0 0 11 0
> MatLUFactorNum 1 1.0 2.0120e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00
> 2.0e+00 5 0 0 0 2 6 0 0 0 3 0
> ...
>
> I like to check these functions. In addition, have you tried other matrix
> orderings?
> Hong
>
> >
> >> Hong
> >>
> >>
> > I just tested the problem according to what you suggested. I set icntl_14
> > = 50 and icntl_7 = 5 (METIS). The problem still persisted. The first
> solve
> > took 920 second and second solve took 215 second with same nonzero
> pattern
> > pc set up. I also attached the log_summary output file. Do you have any
> > further suggestion? Thanks.
> >
> > Regards,
> > Wen
> >
> >
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20120430/a7d3d921/attachment.htm>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: petsc_summary_info
Type: application/octet-stream
Size: 18496 bytes
Desc: not available
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20120430/a7d3d921/attachment.obj>
More information about the petsc-users
mailing list