************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- perturbation3 on a intel-cxx named r10n11 with 16 processors, by thisch Wed Aug 8 00:50:24 2012 Using Petsc Release Version 3.3.0, Patch 2, Fri Jul 13 15:42:00 CDT 2012 Max Max/Min Avg Total Time (sec): 3.001e+02 1.00000 3.001e+02 Objects: 8.150e+02 1.00000 8.150e+02 Flops: 4.113e+10 1.01420 4.066e+10 6.505e+11 Flops/sec: 1.370e+08 1.01420 1.355e+08 2.168e+09 MPI Messages: 9.176e+03 1.50299 8.471e+03 1.355e+05 MPI Message Lengths: 7.282e+08 1.53921 5.956e+04 8.072e+09 MPI Reductions: 2.189e+03 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 3.0010e+02 100.0% 6.5052e+11 100.0% 1.355e+05 100.0% 5.956e+04 100.0% 2.188e+03 100.0% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage STSetUp 1 1.0 1.8547e+01 1.0 8.00e+09 1.0 1.3e+03 4.9e+05 3.1e+01 6 20 1 8 1 6 20 1 8 1 6900 STApply 300 1.0 5.9841e+01 1.0 2.26e+10 1.0 1.3e+05 5.5e+04 0.0e+00 20 54 98 90 0 20 54 98 90 0 5913 EPSSetUp 1 1.0 1.8555e+01 1.0 8.00e+09 1.0 1.3e+03 4.9e+05 6.4e+02 6 20 1 8 29 6 20 1 8 29 6897 EPSSolve 1 1.0 6.4916e+01 1.0 3.21e+10 1.0 1.3e+05 5.5e+04 6.0e+02 22 78 98 90 27 22 78 98 90 27 7802 IPOrthogonalize 301 1.0 4.2184e+00 1.2 7.59e+09 1.0 0.0e+00 0.0e+00 6.0e+02 1 19 0 0 27 1 19 0 0 27 28770 IPInnerProduct 2384 1.0 2.2295e+00 1.3 3.81e+09 1.0 0.0e+00 0.0e+00 6.0e+02 1 9 0 0 27 1 9 0 0 27 27306 DSSolve 1 1.0 6.0701e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 DSVectors 157 1.0 6.3388e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 DSOther 1 1.0 1.3946e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 UpdateVectors 2 1.0 7.5799e-01 1.0 2.95e+09 1.0 0.0e+00 0.0e+00 0.0e+00 0 7 0 0 0 0 7 0 0 0 62334 VecMAXPBY 596 1.0 1.9052e+00 1.0 3.78e+09 1.0 0.0e+00 0.0e+00 0.0e+00 1 9 0 0 0 1 9 0 0 0 31748 MatMult 300 1.0 1.2366e+00 2.3 7.56e+08 4.1 6.1e+04 2.0e+04 0.0e+00 0 1 45 16 0 0 1 45 16 0 3732 MatSolve 300 1.0 5.7929e+01 1.0 2.18e+10 1.0 0.0e+00 0.0e+00 0.0e+00 19 54 0 0 0 19 54 0 0 0 6028 MatLUFactorSym 1 1.0 1.9560e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 3.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatLUFactorNum 1 1.0 1.8041e+01 1.0 8.00e+09 1.0 0.0e+00 0.0e+00 0.0e+00 6 20 0 0 0 6 20 0 0 0 7093 MatConvert 1 1.0 3.9580e-03 3.1 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyBegin 4 1.0 3.0010e-02 4.4 0.00e+00 0.0 6.1e+02 5.0e+04 6.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyEnd 4 1.0 1.2887e-01 1.1 0.00e+00 0.0 4.1e+02 2.6e+03 1.0e+01 0 0 0 0 0 0 0 0 0 0 0 MatGetRowIJ 1 1.0 8.8749e-03 1.8 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetOrdering 1 1.0 7.0110e-02 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 4.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatLoad 1 1.0 7.5572e-02 1.0 0.00e+00 0.0 4.5e+02 7.6e+04 1.7e+01 0 0 0 0 1 0 0 0 0 1 0 MatGetRedundant 1 1.0 2.5958e-01 1.3 0.00e+00 0.0 7.2e+02 8.4e+05 4.0e+00 0 0 1 7 0 0 0 1 7 0 0 VecScale 301 1.0 3.2115e-03 1.1 6.30e+06 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 31396 VecCopy 156 1.0 2.2573e-03 1.5 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 757 1.0 1.5594e-01 1.6 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecScatterBegin 1055 1.0 6.5677e-01 2.4 0.00e+00 0.0 1.3e+05 5.5e+04 1.6e+02 0 0 98 92 7 0 0 98 92 7 0 VecScatterEnd 900 1.0 1.3745e+00 2.3 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecReduceArith 1193 1.0 1.8012e+00 1.1 3.81e+09 1.0 0.0e+00 0.0e+00 0.0e+00 1 9 0 0 0 1 9 0 0 0 33799 VecReduceComm 597 1.0 6.3025e-01 5.5 0.00e+00 0.0 0.0e+00 0.0e+00 6.0e+02 0 0 0 0 27 0 0 0 0 27 0 KSPSetUp 2 1.0 2.1458e-06 0.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 300 1.0 5.9209e+01 1.0 2.18e+10 1.0 7.2e+04 8.4e+04 0.0e+00 20 54 53 75 0 20 54 53 75 0 5898 PCSetUp 1 1.0 1.8546e+01 1.0 8.00e+09 1.0 1.3e+03 4.9e+05 2.9e+01 6 20 1 8 1 6 20 1 8 1 6900 PCApply 300 1.0 5.9195e+01 1.0 2.18e+10 1.0 7.2e+04 8.4e+04 0.0e+00 20 54 53 75 0 20 54 53 75 0 5899 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Container 3 3 1668 0 Spectral Transform 1 1 760 0 Eigenproblem Solver 1 1 16760 0 Inner product 1 1 632 0 Direct solver 1 1 5809064 0 Matrix 8 8 366808480 0 Viewer 2 1 728 0 Vector 468 468 213992384 0 Vector Scatter 159 159 101580 0 Index Set 166 166 2029224 0 PetscRandom 1 1 632 0 Krylov Solver 2 2 2160 0 Preconditioner 2 2 1816 0 ======================================================================================================================== Average time to get PetscTime(): 9.53674e-08 Average time for MPI_Barrier(): 4.95911e-06 Average time for zero size MPI_Send(): 5.87106e-06 #PETSc Option Table entries: -f parameters_initial.xml -k 30.0 -log_summary #End of PETSc Option Table entries Compiled with FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 16 sizeof(PetscInt) 4 Configure run at: Wed Aug 8 00:34:20 2012 Configure options: --with-c++-support=1 --with-scalar-type=complex --with-x11=0 --with-clanguage=cxx --with-shared-libraries=1 --with-fortran-kernels=1 --download-sowing --with-blas-lapack-dir=/opt/intel/Compiler/11.1/046/mkl/lib CXXOPTFLAGS='-O3 -xHost -ipo' COPTFLAGS='-O3 -xHost -ipo' FOPTFLAGS='-O3 -xHost -ipo' --with-debugging=no ----------------------------------------- Libraries compiled on Wed Aug 8 00:34:20 2012 on l01 Machine characteristics: Linux-2.6.18-238.el5-x86_64-with-redhat-5.7-Final Using PETSc directory: /home/lv70072/thisch/local/src/petsc-3.3-p2 Using PETSc arch: intel-cxx-complex_release ----------------------------------------- Using C compiler: mpicxx -wd1572 -fPIC ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: mpif90 -fPIC ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/home/lv70072/thisch/local/src/petsc-3.3-p2/intel-cxx-complex_release/include -I/home/lv70072/thisch/local/src/petsc-3.3-p2/include -I/home/lv70072/thisch/local/src/petsc-3.3-p2/include -I/home/lv70072/thisch/local/src/petsc-3.3-p2/intel-cxx-complex_release/include -I/usr/mpi/intel/openmpi-1.4.3-qlc/include ----------------------------------------- Using C linker: mpicxx Using Fortran linker: mpif90 Using libraries: -Wl,-rpath,/home/lv70072/thisch/local/src/petsc-3.3-p2/intel-cxx-complex_release/lib -L/home/lv70072/thisch/local/src/petsc-3.3-p2/intel-cxx-complex_release/lib -lpetsc -lX11 -lpthread -Wl,-rpath,/opt/intel/Compiler/11.1/046/mkl/lib -L/opt/intel/Compiler/11.1/046/mkl/lib -lmkl_intel_lp64 -lmkl_intel_thread -lmkl_core -liomp5 -lpthread -Wl,-rpath,/usr/mpi/intel/openmpi-1.4.3-qlc/lib64 -L/usr/mpi/intel/openmpi-1.4.3-qlc/lib64 -Wl,-rpath,/opt/intel/Compiler/11.1/046/lib/intel64 -L/opt/intel/Compiler/11.1/046/lib/intel64 -Wl,-rpath,/opt/intel/Compiler/11.1/046/ipp/em64t/lib -L/opt/intel/Compiler/11.1/046/ipp/em64t/lib -Wl,-rpath,/opt/intel/Compiler/11.1/046/mkl/lib/em64t -L/opt/intel/Compiler/11.1/046/mkl/lib/em64t -Wl,-rpath,/opt/intel/Compiler/11.1/046/tbb/em64t/cc4.1.0_libc2.4_kernel2.6.16.21/lib -L/opt/intel/Compiler/11.1/046/tbb/em64t/cc4.1.0_libc2.4_kernel2.6.16.21/lib -Wl,-rpath,/usr/lib/gcc/x86_64-redhat-linux/4.1.2 -L/usr/lib/gcc/x86_64-redhat-linux/4.1.2 -lmpi_f90 -lmpi_f77 -lifport -lifcore -lm -lm -lmpi_cxx -lstdc++ -lmpi_cxx -lstdc++ -ldl -lmpi -lopen-rte -lopen-pal -lnsl -lutil -limf -lsvml -lipgo -ldecimal -lirc -lgcc_s -lpthread -lirc_s -ldl -----------------------------------------