************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./linearElasticity on a arch-linu named fdkong-computer with 2 processors, by fdkong Sat Aug 18 11:02:49 2012 Using Petsc Release Version 3.3.0, Patch 1, Fri Jun 15 09:30:49 CDT 2012 Max Max/Min Avg Total Time (sec): 1.699e+03 1.00000 1.699e+03 Objects: 1.550e+02 1.00000 1.550e+02 Flops: 4.421e+09 1.02094 4.376e+09 8.752e+09 Flops/sec: 2.603e+06 1.02095 2.576e+06 5.152e+06 Memory: 2.268e+08 1.08297 4.363e+08 MPI Messages: 7.760e+02 1.00000 7.760e+02 1.552e+03 MPI Message Lengths: 5.569e+06 1.00045 7.175e+03 1.114e+07 MPI Reductions: 6.455e+03 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 1.6488e+03 97.1% 1.5346e+09 17.5% 4.240e+02 27.3% 4.306e+03 60.0% 3.624e+03 56.1% 1: MG Apply: 4.9958e+01 2.9% 7.2176e+09 82.5% 1.128e+03 72.7% 2.869e+03 40.0% 2.830e+03 43.8% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ ########################################################## # # # WARNING!!! # # # # This code was compiled with a debugging option, # # To get timing results run ./configure # # using --with-debugging=no, the performance will # # be generally two or three times faster. # # # ########################################################## Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage VecMDot 188 1.0 1.6736e+00 2.9 1.57e+08 1.0 0.0e+00 0.0e+00 1.9e+02 0 4 0 0 3 0 20 0 0 5 186 VecNorm 197 1.0 5.8391e-01 1.2 1.09e+07 1.0 0.0e+00 0.0e+00 2.0e+02 0 0 0 0 3 0 1 0 0 5 37 VecScale 195 1.0 2.7156e-02 1.0 5.41e+06 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 1 0 0 0 396 VecCopy 1 1.0 4.3702e-04 1.6 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 13 1.0 6.5494e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 8 1.0 9.0983e-03 8.0 4.44e+05 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 97 VecWAXPY 6 1.0 1.1086e-03 1.1 1.66e+05 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 299 VecMAXPY 195 1.0 4.5723e-01 1.2 1.67e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 4 0 0 0 0 22 0 0 0 727 VecAssemblyBegin 1 1.0 1.6635e-0111.3 0.00e+00 0.0 2.0e+00 3.6e+04 3.0e+00 0 0 0 1 0 0 0 0 1 0 0 VecAssemblyEnd 1 1.0 7.1049e-05 1.3 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecScatterBegin 196 1.0 1.0723e-0112.1 0.00e+00 0.0 3.9e+02 3.9e+03 0.0e+00 0 0 25 14 0 0 0 92 23 0 0 VecScatterEnd 196 1.0 2.5406e-01 7.5 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatMult 195 1.0 3.1695e+00 1.3 4.27e+08 1.0 3.9e+02 3.9e+03 0.0e+00 0 10 25 14 0 0 55 92 23 0 268 MatAssemblyBegin 2 1.0 1.0655e-02 1.8 0.00e+00 0.0 3.0e+00 4.3e+05 2.0e+00 0 0 0 12 0 0 0 1 19 0 0 MatAssemblyEnd 2 1.0 1.0435e-01 1.4 0.00e+00 0.0 4.0e+00 2.0e+03 2.0e+01 0 0 0 0 0 0 0 1 0 1 0 MatGetSubMatrice 1 1.0 8.1369e-01 1.0 0.00e+00 0.0 1.0e+01 6.5e+04 1.0e+01 0 0 1 6 0 0 0 2 10 0 0 MatIncreaseOvrlp 1 1.0 8.6808e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatZeroEntries 1 1.0 9.6719e-03 1.6 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 PCSetUp 3 1.0 9.5933e-01 1.0 0.00e+00 0.0 1.4e+01 4.7e+04 4.7e+01 0 0 1 6 1 0 0 3 10 1 0 PCApply 188 1.0 5.0091e+01 1.0 3.66e+09 1.0 1.1e+03 3.9e+03 2.8e+03 3 82 73 40 44 3470266 67 78 144 KSPGMRESOrthog 188 1.0 3.9346e+00 1.4 3.14e+08 1.0 0.0e+00 0.0e+00 3.0e+03 0 7 0 0 47 0 41 0 0 83 158 KSPSetUp 3 1.0 4.8312e-02 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 1.3e+01 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 1 1.0 5.8313e+01 1.0 4.42e+09 1.0 1.5e+03 3.9e+03 6.2e+03 3100 98 54 97 4570358 90172 150 --- Event Stage 1: MG Apply VecMDot 188 1.0 1.0612e+00 1.8 1.04e+07 1.0 0.0e+00 0.0e+00 1.9e+02 0 0 0 0 3 2 0 0 0 7 20 VecNorm 376 1.0 1.6270e+00 1.3 2.09e+07 1.0 0.0e+00 0.0e+00 3.8e+02 0 0 0 0 6 3 1 0 0 13 25 VecScale 376 1.0 5.5158e-02 1.3 1.04e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 376 VecCopy 188 1.0 2.2421e-02 1.3 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 1316 1.0 2.4112e-01 2.6 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 188 1.0 2.8037e-02 1.0 1.04e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 740 VecMAXPY 376 1.0 9.1900e-02 1.3 2.09e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 1 0 0 0 451 VecScatterBegin 940 1.0 3.4100e-01 1.0 0.00e+00 0.0 1.1e+03 3.9e+03 0.0e+00 0 0 73 40 0 1 0100100 0 0 VecScatterEnd 940 1.0 1.9832e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecNormalize 376 1.0 1.7564e+00 1.3 3.13e+07 1.0 0.0e+00 0.0e+00 3.8e+02 0 1 0 0 6 3 1 0 0 13 35 MatMult 188 1.0 2.6880e+00 1.1 4.11e+08 1.0 3.8e+02 3.9e+03 0.0e+00 0 9 24 13 0 5 11 33 33 0 305 MatSolve 376 1.0 1.5973e+01 1.0 2.89e+09 1.0 0.0e+00 0.0e+00 0.0e+00 1 65 0 0 0 32 79 0 0 0 357 MatLUFactorNum 1 1.0 8.3368e+00 1.5 2.88e+08 1.1 0.0e+00 0.0e+00 0.0e+00 0 6 0 0 0 14 8 0 0 0 66 MatILUFactorSym 1 1.0 1.1393e+01 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 5.0e+00 1 0 0 0 0 22 0 0 0 0 0 MatGetRowIJ 1 1.0 4.5624e-02 1.6 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetOrdering 1 1.0 1.5832e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 4.0e+00 0 0 0 0 0 0 0 0 0 0 0 PCSetUp 1 1.0 1.9124e+01 1.1 2.88e+08 1.1 0.0e+00 0.0e+00 1.0e+01 1 6 0 0 0 36 8 0 0 0 29 PCSetUpOnBlocks 188 1.0 1.9133e+01 1.1 2.88e+08 1.1 0.0e+00 0.0e+00 1.0e+01 1 6 0 0 0 36 8 0 0 0 29 PCApply 376 1.0 2.0558e+01 1.0 2.89e+09 1.0 7.5e+02 3.9e+03 1.1e+03 1 65 48 27 17 41 79 67 67 40 277 KSPGMRESOrthog 188 1.0 1.7469e+00 1.4 2.09e+07 1.0 0.0e+00 0.0e+00 3.8e+02 0 0 0 0 6 3 1 0 0 13 24 KSPSetUp 1 1.0 3.0994e-05 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 188 1.0 4.9479e+01 1.0 3.66e+09 1.0 1.1e+03 3.9e+03 2.6e+03 3 82 73 40 41 99100100100 93 146 MGSmooth Level 0 188 1.0 4.9512e+01 1.0 3.66e+09 1.0 1.1e+03 3.9e+03 2.6e+03 3 82 73 40 41 99100100100 93 146 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Container 1 0 0 0 Viewer 2 1 736 0 Bipartite Graph 14 12 8880 0 Index Set 19 19 698544 0 IS L to G Mapping 1 1 885292 0 Vector 93 90 19179120 0 Vector Scatter 3 3 3516 0 Matrix 4 5 98389464 0 Distributed Mesh 2 1 4400 0 Preconditioner 5 5 4824 0 Krylov Solver 5 5 41448 0 --- Event Stage 1: MG Apply Index Set 5 2 152464 0 Matrix 1 0 0 0 ======================================================================================================================== Average time to get PetscTime(): 1.02997e-05 Average time for MPI_Barrier(): 5.25951e-05 Average time for zero size MPI_Send(): 0.000335455 #PETSc Option Table entries: -E 215000000000 -coarseILUlevel 2 -coarseIter 1 -coarseTol 0.1 -coordMove 0,0,0 -coordScale 0.001,0.001,0.001 -fineILUlevel 2 -log_summary -nu 0.29 -overlap 1 -pc_mg_log #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 8 Configure run at: Thu Jul 26 10:40:17 2012 Configure options: --with-clanguage=cxx --with-shared-libraries=1 --with-dynamic-loading=1 --download-f-blas-lapack=1 --with-mpi-dir=/home/fdkong/math/mpich2-install --download-parmetis=1 --download-metis=1 --with-64-bit-indices=1 --with-netcdf-dir=/home/fdkong/math/petsc-3.3-p1/externalpackage/netcdf-4.1.1 --with-exodusii-dir=/home/fdkong/math/petsc-3.3-p1/externalpackage/exodusii-5.14-petsc --with-chaco-dir=/home/fdkong/math/petsc-3.3-p1/externalpackage/chaco ----------------------------------------- Libraries compiled on Thu Jul 26 10:40:17 2012 on fdkong-computer Machine characteristics: Linux-3.2.0-25-generic-x86_64-with-Ubuntu-12.04-precise Using PETSc directory: /home/fdkong/math/petsc-3.3-p1 Using PETSc arch: arch-linux264-cxx-debug ----------------------------------------- Using C compiler: /home/fdkong/math/mpich2-install/bin/mpicxx -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -g -fPIC ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: /home/fdkong/math/mpich2-install/bin/mpif90 -fPIC -Wall -Wno-unused-variable -Wno-unused-dummy-argument -g ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/home/fdkong/math/petsc-3.3-p1/arch-linux264-cxx-debug/include -I/home/fdkong/math/petsc-3.3-p1/include -I/home/fdkong/math/petsc-3.3-p1/include -I/home/fdkong/math/petsc-3.3-p1/arch-linux264-cxx-debug/include -I/home/fdkong/math/petsc-3.3-p1/externalpackage/exodusii-5.14-petsc/include -I/home/fdkong/math/petsc-3.3-p1/externalpackage/exodusii-5.14-petsc/cbind/include -I/home/fdkong/math/petsc-3.3-p1/externalpackage/exodusii-5.14-petsc/forbind/include -I/home/fdkong/math/petsc-3.3-p1/externalpackage/netcdf-4.1.1/include -I/home/fdkong/math/mpich2-install/include ----------------------------------------- Using C linker: /home/fdkong/math/mpich2-install/bin/mpicxx Using Fortran linker: /home/fdkong/math/mpich2-install/bin/mpif90 Using libraries: -Wl,-rpath,/home/fdkong/math/petsc-3.3-p1/arch-linux264-cxx-debug/lib -L/home/fdkong/math/petsc-3.3-p1/arch-linux264-cxx-debug/lib -lpetsc -Wl,-rpath,/home/fdkong/math/petsc-3.3-p1/arch-linux264-cxx-debug/lib -L/home/fdkong/math/petsc-3.3-p1/arch-linux264-cxx-debug/lib -lparmetis -lmetis -lpthread -Wl,-rpath,/home/fdkong/math/petsc-3.3-p1/externalpackage/chaco/lib -L/home/fdkong/math/petsc-3.3-p1/externalpackage/chaco/lib -lchaco -lflapack -lfblas -Wl,-rpath,/home/fdkong/math/petsc-3.3-p1/externalpackage/exodusii-5.14-petsc/lib -L/home/fdkong/math/petsc-3.3-p1/externalpackage/exodusii-5.14-petsc/lib -lexoIIv2for -lexodus -Wl,-rpath,/home/fdkong/math/petsc-3.3-p1/externalpackage/netcdf-4.1.1/lib -L/home/fdkong/math/petsc-3.3-p1/externalpackage/netcdf-4.1.1/lib -lnetcdf_c++ -lnetcdf -lm -Wl,-rpath,/home/fdkong/math/mpich2-install/lib -L/home/fdkong/math/mpich2-install/lib -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.6 -L/usr/lib/gcc/x86_64-linux-gnu/4.6 -Wl,-rpath,/usr/lib/x86_64-linux-gnu -L/usr/lib/x86_64-linux-gnu -Wl,-rpath,/lib/x86_64-linux-gnu -L/lib/x86_64-linux-gnu -lmpichf90 -lgfortran -lm -lgfortran -lm -lquadmath -lm -lmpichcxx -lstdc++ -ldl -lmpich -lopa -lmpl -lrt -lpthread -lgcc_s -ldl -----------------------------------------