************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- D:\dsu\ResearchAtUBC\Dropbox\ParallelDevelop\AuxiliaryPrograms\Petsc-Solver-Test\petsc_unsym_f\ksp_inhm_d.exe on a arch-mswin-c-debug named NWMOP with 4 processors, by dsu Sun Sep 22 04:23:11 2013 Using Petsc Release Version 3.4.2, Jul, 02, 2013 Max Max/Min Avg Total Time (sec): 3.814e+000 1.00019 3.813e+000 Objects: 4.300e+001 1.00000 4.300e+001 Flops: 1.191e+007 1.00181 1.190e+007 4.760e+007 Flops/sec: 3.124e+006 1.00200 3.121e+006 1.248e+007 Memory: 9.848e+006 1.01103 3.913e+007 MPI Messages: 2.050e+001 1.95238 1.575e+001 6.300e+001 MPI Message Lengths: 1.841e+005 2.40344 7.296e+003 4.597e+005 MPI Reductions: 1.400e+002 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 3.8131e+000 100.0% 4.7602e+007 100.0% 6.300e+001 100.0% 7.296e+003 100.0% 1.390e+002 99.3% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length (bytes) Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ ########################################################## # # # WARNING!!! # # # # This code was compiled with a debugging option, # # To get timing results run ./configure # # using --with-debugging=no, the performance will # # be generally two or three times faster. # # # ########################################################## Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage ThreadCommRunKer 1 1.0 2.1504e-005 4.7 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 ThreadCommBarrie 1 1.0 2.0480e-006 1.3 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 MatMult 8 1.0 1.2054e-002 1.4 4.17e+006 1.0 4.8e+001 2.7e+003 0.0e+000 0 35 76 28 0 0 35 76 28 0 1381 MatSolve 7 1.0 1.1630e-002 1.4 3.61e+006 1.0 0.0e+000 0.0e+000 0.0e+000 0 30 0 0 0 0 30 0 0 0 1243 MatLUFactorNum 1 1.0 3.6862e-002 1.1 2.34e+006 1.0 0.0e+000 0.0e+000 0.0e+000 1 20 0 0 0 1 20 0 0 0 254 MatILUFactorSym 1 1.0 6.9094e-003 1.2 0.00e+000 0.0 0.0e+000 0.0e+000 1.0e+000 0 0 0 0 1 0 0 0 0 1 0 MatAssemblyBegin 1 1.0 7.8909e-00395.1 0.00e+000 0.0 0.0e+000 0.0e+000 2.0e+000 0 0 0 0 1 0 0 0 0 1 0 MatAssemblyEnd 1 1.0 5.3555e-003 1.0 0.00e+000 0.0 1.2e+001 6.7e+002 2.1e+001 0 0 19 2 15 0 0 19 2 15 0 MatGetRowIJ 1 1.0 4.0960e-006 1.6 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 MatGetOrdering 1 1.0 2.9276e-003 1.1 0.00e+000 0.0 0.0e+000 0.0e+000 4.0e+000 0 0 0 0 3 0 0 0 0 3 0 MatZeroEntries 1 1.0 2.1442e-003 1.1 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 VecView 1 1.0 3.6582e-001 4.2 0.00e+000 0.0 3.0e+000 1.1e+005 0.0e+000 6 0 5 70 0 6 0 5 70 0 0 VecMax 1 1.0 1.4694e-004 1.6 0.00e+000 0.0 0.0e+000 0.0e+000 1.0e+000 0 0 0 0 1 0 0 0 0 1 0 VecMin 1 1.0 8.2431e-005 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 1.0e+000 0 0 0 0 1 0 0 0 0 1 0 VecMDot 6 1.0 9.7509e-003 9.4 5.64e+005 1.0 0.0e+000 0.0e+000 6.0e+000 0 5 0 0 4 0 5 0 0 4 232 VecNorm 8 1.0 1.7582e-003 2.6 2.15e+005 1.0 0.0e+000 0.0e+000 8.0e+000 0 2 0 0 6 0 2 0 0 6 489 VecScale 7 1.0 1.2580e-003 1.2 9.41e+004 1.0 0.0e+000 0.0e+000 0.0e+000 0 1 0 0 0 0 1 0 0 0 299 VecCopy 1 1.0 7.2191e-005 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 VecSet 10 1.0 2.0019e-004 1.2 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 4 1.0 7.6236e-004 1.5 1.08e+005 1.0 0.0e+000 0.0e+000 0.0e+000 0 1 0 0 0 0 1 0 0 0 564 VecMAXPY 7 1.0 1.3762e-003 1.7 7.26e+005 1.0 0.0e+000 0.0e+000 0.0e+000 0 6 0 0 0 0 6 0 0 0 2109 VecAssemblyBegin 2 1.0 1.5894e-001867.1 0.00e+000 0.0 0.0e+000 0.0e+000 6.0e+000 3 0 0 0 4 3 0 0 0 4 0 VecAssemblyEnd 2 1.0 9.7279e-006 1.2 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 VecScatterBegin 8 1.0 1.2237e-004 1.7 0.00e+000 0.0 4.8e+001 2.7e+003 0.0e+000 0 0 76 28 0 0 0 76 28 0 0 VecScatterEnd 8 1.0 1.3261e-004 1.6 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 VecReduceArith 4 1.0 3.3894e-004 1.7 8.06e+004 1.0 0.0e+000 0.0e+000 0.0e+000 0 1 0 0 0 0 1 0 0 0 952 VecReduceComm 2 1.0 2.6829e-00415.4 0.00e+000 0.0 0.0e+000 0.0e+000 2.0e+000 0 0 0 0 1 0 0 0 0 1 0 VecNormalize 7 1.0 2.8759e-003 1.4 2.82e+005 1.0 0.0e+000 0.0e+000 7.0e+000 0 2 0 0 5 0 2 0 0 5 393 KSPGMRESOrthog 6 1.0 1.0624e-002 4.7 1.13e+006 1.0 0.0e+000 0.0e+000 2.7e+001 0 9 0 0 19 0 9 0 0 19 425 KSPSetUp 2 1.0 7.3727e-004 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 1 1.0 9.2897e-002 1.0 1.07e+007 1.0 3.6e+001 2.7e+003 7.3e+001 2 90 57 21 52 2 90 57 21 53 460 PCSetUp 2 1.0 4.7200e-002 1.1 2.34e+006 1.0 0.0e+000 0.0e+000 1.6e+001 1 20 0 0 11 1 20 0 0 12 198 PCSetUpOnBlocks 1 1.0 4.6733e-002 1.1 2.34e+006 1.0 0.0e+000 0.0e+000 8.0e+000 1 20 0 0 6 1 20 0 0 6 200 PCApply 7 1.0 1.7098e-002 1.4 3.61e+006 1.0 0.0e+000 0.0e+000 1.4e+001 0 30 0 0 10 0 30 0 0 10 845 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Matrix 4 4 8749324 0 Vector 24 24 2189952 0 Vector Scatter 1 1 1052 0 Index Set 7 7 127616 0 IS L to G Mapping 1 1 54348 0 Krylov Solver 2 2 19504 0 Preconditioner 2 2 1848 0 Viewer 2 1 728 0 ======================================================================================================================== Average time to get PetscTime(): 1.02399e-007 Average time for MPI_Barrier(): 3.48157e-006 Average time for zero size MPI_Send(): 2.30398e-006 #PETSc Option Table entries: -log_summary log_mpi4.log #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 4 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure run at: Tue Aug 6 10:46:18 2013 Configure options: --with-cc="win32fe cl" --with-fc="win32fe ifort" --with-cxx="win32fe cl" --with-parmetis-include=/cygdrive/c/cygwin/packages/parmetis-4.0.3/include --with-parmetis-lib=/cygdrive/c/cygwin/packages/parmetis-4.0.3/build/libparmetis/Release/parmetis.lib --with-metis-include=/cygdrive/c/cygwin/packages/metis-5.1.0/include --with-metis-lib=/cygdrive/c/cygwin/packages/metis-5.1.0/build/libmetis/Release/metis.lib --download-f-blas-lapack --useThreads=0 ----------------------------------------- Libraries compiled on Tue Aug 6 10:46:18 2013 on nwmop Machine characteristics: CYGWIN_NT-6.1-WOW64-1.7.22-0.268-5-3-i686-32bit Using PETSc directory: /cygdrive/c/cygwin/packages/petsc-3.4.2 Using PETSc arch: arch-mswin-c-debug ----------------------------------------- Using C compiler: /cygdrive/c/cygwin/packages/petsc-3.4.2/bin/win32fe/win32fe cl -MT -wd4996 -Z7 ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: /cygdrive/c/cygwin/packages/petsc-3.4.2/bin/win32fe/win32fe ifort -MT -Z7 -fpp ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/cygdrive/c/cygwin/packages/petsc-3.4.2/arch-mswin-c-debug/include -I/cygdrive/c/cygwin/packages/petsc-3.4.2/include -I/cygdrive/c/cygwin/packages/petsc-3.4.2/include -I/cygdrive/c/cygwin/packages/petsc-3.4.2/arch-mswin-c-debug/include -I/cygdrive/c/cygwin/packages/parmetis-4.0.3/include -I/cygdrive/c/cygwin/packages/metis-5.1.0/include -I/cygdrive/c/Program\ Files/MPICH2/include ----------------------------------------- Using C linker: /cygdrive/c/cygwin/packages/petsc-3.4.2/bin/win32fe/win32fe cl Using Fortran linker: /cygdrive/c/cygwin/packages/petsc-3.4.2/bin/win32fe/win32fe ifort Using libraries: -L/cygdrive/c/cygwin/packages/petsc-3.4.2/arch-mswin-c-debug/lib -L/cygdrive/c/cygwin/packages/petsc-3.4.2/arch-mswin-c-debug/lib -lpetsc -L/cygdrive/c/cygwin/packages/petsc-3.4.2/arch-mswin-c-debug/lib -lflapack -lfblas /cygdrive/c/cygwin/packages/parmetis-4.0.3/build/libparmetis/Release/parmetis.lib /cygdrive/c/cygwin/packages/metis-5.1.0/build/libmetis/Release/metis.lib /cygdrive/c/Program\ Files/MPICH2/lib/fmpich2.lib /cygdrive/c/Program\ Files/MPICH2/lib/fmpich2g.lib /cygdrive/c/Program\ Files/MPICH2/lib/mpi.lib -L/cygdrive/c/cygwin/packages/petsc-3.4.2/PROGRA~2IntelCOMPOS~1inintel64 Gdi32.lib User32.lib Advapi32.lib Kernel32.lib Ws2_32.lib -----------------------------------------