<html>
<head>
<style>
.hmmessage P
{
margin:0px;
padding:0px
}
body.hmmessage
{
font-size: 10pt;
font-family:Verdana
}
</style>
</head>
<body class='hmmessage'>
Here is my -log_summary:<br>Something looks a bit strange to me; its the MPI Reductions below.<br>Other than that, i dont see anything relevant.<br>What do you think?<br>Thanks<br><br><br><br>---------------------------------------------- PETSc Performance Summary: ----------------------------------------------<br><br>./diff-conv-opt on a linux-gnu named linux-stchouan with 1 processor, by stephane Tue Feb 24 13:54:35 2009<br>Using Petsc Release Version 2.3.3, Patch 13, Thu May 15 17:29:26 CDT 2008 HG revision: 4466c6289a0922df26e20626fd4a0b4dd03c8124<br><br> Max Max/Min Avg Total <br>Time (sec): 2.937e+03 1.00000 2.937e+03<br>Objects: 3.420e+03 1.00000 3.420e+03<br>Flops: 2.690e+09 1.00000 2.690e+09 2.690e+09<br>Flops/sec: 9.161e+05 1.00000 9.161e+05 9.161e+05<br>MPI Messages: 0.000e+00 0.00000 0.000e+00 0.000e+00<br>MPI Message Lengths: 0.000e+00 0.00000 0.000e+00 0.000e+00<br>MPI Reductions: 1.189e+05 1.00000<br><br>Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract)<br> e.g., VecAXPY() for real vectors of length N --> 2N flops<br> and VecAXPY() for complex vectors of length N --> 8N flops<br><br>Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions --<br> Avg %Total Avg %Total counts %Total Avg %Total counts %Total <br> 0: Main Stage: 2.9367e+03 100.0% 2.6905e+09 100.0% 0.000e+00 0.0% 0.000e+00 0.0% 1.106e+03 0.9% <br><br>------------------------------------------------------------------------------------------------------------------------<br>See the 'Profiling' chapter of the users' manual for details on interpreting output.<br>Phase summary info:<br> Count: number of times phase was executed<br> Time and Flops/sec: Max - maximum over all processors<br> Ratio - ratio of maximum to minimum over all processors<br> Mess: number of messages sent<br> Avg. len: average message length<br> Reduct: number of global reductions<br> Global: entire computation<br> Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop().<br> %T - percent time in this phase %F - percent flops in this phase<br> %M - percent messages in this phase %L - percent message lengths in this phase<br> %R - percent reductions in this phase<br> Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors)<br>------------------------------------------------------------------------------------------------------------------------<br><br>Event Count Time (sec) Flops/sec --- Global --- --- Stage --- Total<br> Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %F %M %L %R %T %F %M %L %R Mflop/s<br>------------------------------------------------------------------------------------------------------------------------<br><br>--- Event Stage 0: Main Stage<br><br>VecDot 202 1.0 3.0360e-02 1.0 3.96e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 396<br>VecMDot 202 1.0 3.0552e-02 1.0 3.94e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 394<br>VecNorm 1110 1.0 1.2257e+00 1.0 5.40e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 2 0 0 0 0 2 0 0 0 54<br>VecScale 404 1.0 3.5342e-02 1.0 3.41e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 341<br>VecCopy 507 1.0 8.4626e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0<br>VecSet 1408 1.0 1.1664e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0<br>VecAXPY 202 1.0 2.6221e-02 1.0 4.59e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 459<br>VecWAXPY 202 1.0 4.4239e-02 1.0 1.36e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 136<br>VecMAXPY 404 1.0 7.3515e-02 1.0 3.27e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 327<br>VecAssemblyBegin 302 1.0 9.2960e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0<br>VecAssemblyEnd 302 1.0 5.5790e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0<br>VecScatterBegin 603 1.0 1.9933e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0<br>VecNormalize 404 1.0 5.5408e-01 1.0 6.52e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 65<br>MatMult 404 1.0 2.6457e+00 1.0 2.26e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 22 0 0 0 0 22 0 0 0 226<br>MatSolve 404 1.0 4.6454e+00 1.0 1.28e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 22 0 0 0 0 22 0 0 0 128<br>MatLUFactorNum 202 1.0 1.5211e+01 1.0 8.85e+07 1.0 0.0e+00 0.0e+00 0.0e+00 1 50 0 0 0 1 50 0 0 0 89<br>MatILUFactorSym 100 1.0 1.9993e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+02 0 0 0 0 0 0 0 0 0 9 0<br>MatAssemblyBegin 404 1.0 9.6217e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0<br>MatAssemblyEnd 404 1.0 1.4601e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0<br>MatGetRowIJ 100 1.0 2.4641e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0<br>MatGetOrdering 100 1.0 7.6755e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+02 0 0 0 0 0 0 0 0 0 18 0<br>MatZeroEntries 99 1.0 3.6160e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0<br>SNESSolve 100 1.0 2.9339e+03 1.0 9.17e+05 1.0 0.0e+00 0.0e+00 1.0e+03100100 0 0 1 100100 0 0 91 1<br>SNESLineSearch 202 1.0 7.9707e+02 1.0 4.35e+05 1.0 0.0e+00 0.0e+00 4.0e+02 27 13 0 0 0 27 13 0 0 37 0<br>SNESFunctionEval 302 1.0 1.1836e+03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 3.0e+02 40 0 0 0 0 40 0 0 0 27 0<br>SNESJacobianEval 202 1.0 1.7238e+03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+02 59 0 0 0 0 59 0 0 0 18 0<br>KSPGMRESOrthog 202 1.0 7.0303e-02 1.0 3.42e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 342<br>KSPSetup 202 1.0 4.6391e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0<br>KSPSolve 202 1.0 2.4101e+01 1.0 9.65e+07 1.0 0.0e+00 0.0e+00 3.0e+02 1 86 0 0 0 1 86 0 0 27 97<br>PCSetUp 202 1.0 1.7296e+01 1.0 7.78e+07 1.0 0.0e+00 0.0e+00 3.0e+02 1 50 0 0 0 1 50 0 0 27 78<br>PCApply 404 1.0 4.6487e+00 1.0 1.28e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 22 0 0 0 0 22 0 0 0 128<br>------------------------------------------------------------------------------------------------------------------------<br><br>Memory usage is given in bytes:<br><br>Object Type Creations Destructions Memory Descendants' Mem.<br><br>--- Event Stage 0: Main Stage<br><br> Index Set 904 901 107564984 0<br> Vec 1511 1497 357441684 0<br> Vec Scatter 604 604 0 0<br> Matrix 101 99 942432084 0<br> SNES 100 99 12276 0<br> Krylov Solver 100 99 1671120 0<br> Preconditioner 100 99 14256 0<br>========================================================================================================================<br>Average time to get PetscTime(): 1.49164e-06<br>OptionTable: -snes_converged_reason<br>OptionTable: -snes_max_it 20<br>OptionTable: -snes_rtol 0.0000001<br>OptionTable: -snes_stol 0.001<br>Compiled without FORTRAN kernels<br>Compiled with full precision matrices (default)<br>sizeof(short) 2 sizeof(int) 4 sizeof(long) 4 sizeof(void*) 4 sizeof(PetscScalar) 8<br>Configure run at: Mon Feb 23 23:01:43 2009<br>Configure options: --with-debugging=no -with-shared --download-mpich=1<br>-----------------------------------------<br><br><br><br><br>> <br>> Message: 7<br>> Date: Tue, 24 Feb 2009 18:05:38 +0100<br>> From: STEPHANE TCHOUANMO <tchouanm@msn.com><br>> Subject: Petsc slowness on a single processor machine?<br>> To: <petsc-users@mcs.anl.gov><br>> Message-ID: <BAY107-W5D076FE1F3A7037240F88C2AF0@phx.gbl><br>> Content-Type: text/plain; charset="windows-1256"<br>> <br>> <br>> Dear all,<br>> <br>> I use petsc to solve a nonlinear convection-diffusion type equation in porous media.<br>> The built linear system's size is 30000 and its resolution lasts about a hour on a single processor machine (my laptop...).<br>> I<br>> run for that a hundred time steps and have in average 4 Newton<br>> iterations per time step. So a hour seems huge to me with the optimized<br>> version of Petsc.<br>> I tryed to profile my code with the options<br>> "-log_summary" and "-info". What i get out of it is that the first time<br>> step lasts about 15min representing 25% of the total time and i think<br>> its not normal.<br>> The thing is, these 15min happen at the second call<br>> of the petsc routine "VecScatterCreate()" right after the first Newton residual.<br>> <br>> Any idea?<br>> Thanks.<br>> <br>> Stephane<br>> <br>> <br>> <br>> _________________________________________________________________<br>> Show them the way! Add maps and directions to your party invites. <br>> http://www.microsoft.com/windows/windowslive/products/events.aspx<br>> -------------- next part --------------<br>> An HTML attachment was scrubbed...<br>> URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20090224/c695a2bf/attachment-0001.htm><br>> <br>> ------------------------------<br>> <br>> Message: 8<br>> Date: Tue, 24 Feb 2009 11:22:18 -0600<br>> From: Matthew Knepley <knepley@gmail.com><br>> Subject: Re: Petsc slowness on a single processor machine?<br>> To: PETSc users list <petsc-users@mcs.anl.gov><br>> Message-ID:<br>>         <a9f269830902240922y6751b40aqf8ea84257c139fd9@mail.gmail.com><br>> Content-Type: text/plain; charset="iso-8859-1"<br>> <br>> 2009/2/24 STEPHANE TCHOUANMO <tchouanm@msn.com><br>> <br>> > Dear all,<br>> ><br>> > I use petsc to solve a nonlinear convection-diffusion type equation in<br>> > porous media.<br>> > The built linear system's size is 30000 and its resolution lasts about a<br>> > hour on a single processor machine (my laptop...).<br>> > I run for that a hundred time steps and have in average 4 Newton iterations<br>> > per time step. So a hour seems huge to me with the optimized version of<br>> > Petsc.<br>> > I tryed to profile my code with the options "-log_summary" and "-info".<br>> > What i get out of it is that the first time step lasts about 15min<br>> > representing 25% of the total time and i think its not normal.<br>> > The thing is, these 15min happen at the second call of the petsc routine<br>> > "VecScatterCreate()" right after the first Newton residual.<br>> ><br>> <br>> We can't say anything without seeing the entire output of -log_summary.<br>> <br>> Matt<br>> <br>> <br>> ><br>> > Any idea?<br>> > Thanks.<br>> ><br>> > Stephane<br>> ><br>> > <http://www.live.com/getstarted.aspx><br>> > <http://www.microsoft.com/windows/windowslive/default.aspx><br>> > ------------------------------<br>> > See all the ways you can stay connected to friends and family<http://www.microsoft.com/windows/windowslive/default.aspx><br>> ><br>> <br>> <br>> <br>> -- <br>> What most experimenters take for granted before they begin their experiments<br>> is infinitely more interesting than any results to which their experiments<br>> lead.<br>> -- Norbert Wiener<br>> -------------- next part --------------<br>> An HTML attachment was scrubbed...<br>> URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20090224/cd1ffccb/attachment-0001.htm><br>> <br>> ------------------------------<br>> <br>> _______________________________________________<br>> petsc-users mailing list<br>> petsc-users@mcs.anl.gov<br>> https://lists.mcs.anl.gov/mailman/listinfo/petsc-users<br>> <br>> <br>> End of petsc-users Digest, Vol 2, Issue 32<br>> ******************************************<br><br /><hr />Get news, entertainment and everything you care about at Live.com. <a href='http://www.live.com/getstarted.aspx ' target='_new'>Check it out!</a></body>
</html>