************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- D:\dsu\ResearchAtUBC\Dropbox\ParallelDevelop\AuxiliaryPrograms\Petsc-Solver-Test\petsc_unsym_f\ksp_inhm_d.exe on a arch-mswin-c-debug named NWMOP with 1 processor, by dsu Sun Sep 22 04:31:13 2013 With 4 threads per MPI_Comm Using Petsc Development GIT revision: 2492010b6ebf10eed9da155c21c4e5e815469b33 GIT Date: 2013-09-20 15:48:09 -0500 Max Max/Min Avg Total Time (sec): 3.635e+002 1.00000 3.635e+002 Objects: 3.200e+001 1.00000 3.200e+001 Flops: 4.783e+007 1.00000 4.783e+007 4.783e+007 Flops/sec: 1.316e+005 1.00000 1.316e+005 1.316e+005 Memory: 3.831e+007 1.00000 3.831e+007 MPI Messages: 0.000e+000 0.00000 0.000e+000 0.000e+000 MPI Message Lengths: 0.000e+000 0.00000 0.000e+000 0.000e+000 MPI Reductions: 0.000e+000 0.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 3.6355e+002 100.0% 4.7833e+007 100.0% 0.000e+000 0.0% 0.000e+000 0.0% 0.000e+000 0.0% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length (bytes) Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ ########################################################## # # # WARNING!!! # # # # This code was compiled with a debugging option, # # To get timing results run ./configure # # using --with-debugging=no, the performance will # # be generally two or three times faster. # # # ########################################################## Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage ThreadCommRunKer 1 1.0 1.3824e-005 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 ThreadCommBarrie 1 1.0 1.0240e-006 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 MatMult 8 1.0 3.3872e-002 1.0 1.66e+007 1.0 0.0e+000 0.0e+000 0.0e+000 0 35 0 0 0 0 35 0 0 0 492 MatSolve 7 1.0 4.2626e-002 1.0 1.46e+007 1.0 0.0e+000 0.0e+000 0.0e+000 0 30 0 0 0 0 30 0 0 0 342 MatLUFactorNum 1 1.0 1.2375e-001 1.0 9.47e+006 1.0 0.0e+000 0.0e+000 0.0e+000 0 20 0 0 0 0 20 0 0 0 77 MatILUFactorSym 1 1.0 1.9471e-002 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyBegin 1 1.0 2.5600e-006 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyEnd 1 1.0 1.5922e-002 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 MatGetRowIJ 1 1.0 6.1439e-006 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 MatGetOrdering 1 1.0 1.1437e-002 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 MatZeroEntries 1 1.0 8.7090e-004 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 VecView 1 1.0 3.3479e-001 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 VecMax 1 1.0 2.7341e-004 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 VecMin 1 1.0 2.7187e-004 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 VecMDot 6 1.0 2.2246e-003 1.0 2.26e+006 1.0 0.0e+000 0.0e+000 0.0e+000 0 5 0 0 0 0 5 0 0 0 1015 VecNorm 8 1.0 1.7930e-003 1.0 8.60e+005 1.0 0.0e+000 0.0e+000 0.0e+000 0 2 0 0 0 0 2 0 0 0 480 VecScale 7 1.0 2.4156e-003 1.0 3.76e+005 1.0 0.0e+000 0.0e+000 0.0e+000 0 1 0 0 0 0 1 0 0 0 156 VecCopy 1 1.0 1.4745e-004 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 VecSet 2 1.0 1.6179e-004 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 4 1.0 1.9435e-003 1.0 4.30e+005 1.0 0.0e+000 0.0e+000 0.0e+000 0 1 0 0 0 0 1 0 0 0 221 VecMAXPY 7 1.0 3.8584e-003 1.0 2.90e+006 1.0 0.0e+000 0.0e+000 0.0e+000 0 6 0 0 0 0 6 0 0 0 752 VecAssemblyBegin 2 1.0 6.6559e-005 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 VecAssemblyEnd 2 1.0 1.2288e-005 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 VecReduceArith 4 1.0 8.6169e-004 1.0 3.23e+005 1.0 0.0e+000 0.0e+000 0.0e+000 0 1 0 0 0 0 1 0 0 0 374 VecReduceComm 2 1.0 7.6799e-006 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 VecNormalize 7 1.0 4.0258e-003 1.0 1.13e+006 1.0 0.0e+000 0.0e+000 0.0e+000 0 2 0 0 0 0 2 0 0 0 280 KSPGMRESOrthog 6 1.0 5.2915e-003 1.0 4.52e+006 1.0 0.0e+000 0.0e+000 0.0e+000 0 9 0 0 0 0 9 0 0 0 853 KSPSetUp 1 1.0 9.3132e-004 1.0 0.00e+000 0.0 0.0e+000 0.0e+000 0.0e+000 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 1 1.0 2.9056e-001 1.0 4.29e+007 1.0 0.0e+000 0.0e+000 0.0e+000 0 90 0 0 0 0 90 0 0 0 148 PCSetUp 1 1.0 1.5493e-001 1.0 9.47e+006 1.0 0.0e+000 0.0e+000 0.0e+000 0 20 0 0 0 0 20 0 0 0 61 PCApply 7 1.0 4.2695e-002 1.0 1.46e+007 1.0 0.0e+000 0.0e+000 0.0e+000 0 30 0 0 0 0 30 0 0 0 341 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Matrix 2 2 17760012 0 Vector 20 20 8632640 0 Index Set 5 5 487680 0 IS L to G Mapping 1 1 215636 0 Krylov Solver 1 1 18368 0 Preconditioner 1 1 984 0 Viewer 2 1 736 0 ======================================================================================================================== Average time to get PetscTime(): 1.53599e-007 #PETSc Option Table entries: -log_summary log_openmp_petsc_dev.log -threadcomm_nthreads 4 -threadcomm_type openmp #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 4 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure run at: Sat Sep 21 10:55:27 2013 Configure options: --with-cc="win32fe cl" --with-fc="win32fe ifort" --with-cxx="win32fe cl" --download-f-blas-lapack --with-openmp --with-mpi=0 --useThreads=0 ----------------------------------------- Libraries compiled on Sat Sep 21 10:55:27 2013 on nwmop Machine characteristics: CYGWIN_NT-6.1-WOW64-1.7.22-0.268-5-3-i686-32bit Using PETSc directory: /cygdrive/c/cygwin/packages/petsc-dev Using PETSc arch: arch-mswin-c-debug ----------------------------------------- Using C compiler: /cygdrive/c/cygwin/packages/petsc-dev/bin/win32fe/win32fe cl -MT -wd4996 -Z7 -openmp ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: /cygdrive/c/cygwin/packages/petsc-dev/bin/win32fe/win32fe ifort -MT -Z7 -fpp -openmp ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/cygdrive/c/cygwin/packages/petsc-dev/arch-mswin-c-debug/include -I/cygdrive/c/cygwin/packages/petsc-dev/include -I/cygdrive/c/cygwin/packages/petsc-dev/include -I/cygdrive/c/cygwin/packages/petsc-dev/arch-mswin-c-debug/include -I/cygdrive/c/cygwin/packages/petsc-dev/include/mpiuni ----------------------------------------- Using C linker: /cygdrive/c/cygwin/packages/petsc-dev/bin/win32fe/win32fe cl Using Fortran linker: /cygdrive/c/cygwin/packages/petsc-dev/bin/win32fe/win32fe ifort Using libraries: -L/cygdrive/c/cygwin/packages/petsc-dev/arch-mswin-c-debug/lib -L/cygdrive/c/cygwin/packages/petsc-dev/arch-mswin-c-debug/lib -lpetsc -L/cygdrive/c/cygwin/packages/petsc-dev/arch-mswin-c-debug/lib -lflapack -lfblas -L/cygdrive/c/cygwin/packages/petsc-dev/PROGRA~2IntelCOMPOS~1inintel64 Gdi32.lib User32.lib Advapi32.lib Kernel32.lib Ws2_32.lib -----------------------------------------