[petsc-users] Problem when switching from debug to optimized

Matthew Knepley knepley at gmail.com
Mon Nov 14 07:34:37 CST 2011


On Mon, Nov 14, 2011 at 12:02 PM, Bogdan Dita <bogdan at lmn.pub.ro> wrote:

>
>  Hello,
>
>  Below is my post from a few days ago and this time I've attached the
> output from log_summary.
>

The time increase comes completely from SuperLU_dist during the
factorization
phase. You should use -ksp_view so we can see what solver options are used.

   Matt


> "
>  Until a few days ago I've only be using PETSc in debug mode and when I
> switch to the optimised version(--with-debugging=0) I got a strange
> result regarding the solve time, what I mean is that it was 10-15 %
> higher then in debug mode.
>  I'm trying to solve a linear system in parallel with superlu_dist, and
> I've tested my program on a Beowulf cluster, so far only using a single
> node with 2 quad-core Intel processors.
>  From what I know the "no debug" version should be faster and I know it
> should be faster because on my laptop(dual-core Intel) for the same
> program and even the same matrices the solve time for the optimised
> version is 2 times faster, but when I use the cluster the optimised
> version time is slower then the debug version.
>  Any thoughts?
>
> "
>  Best regards,
>   Bogdan Dita
>
>
>
>


-- 
What most experimenters take for granted before they begin their
experiments is infinitely more interesting than any results to which their
experiments lead.
-- Norbert Wiener
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20111114/56f21019/attachment.htm>


More information about the petsc-users mailing list