Number of iterations = 8 Residual norm 0.171671 Setup time: 4.9536540508e+00 Solve time: 1.5524308681e+00 Total: 6.5060849190e+00 ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./ex10 on a intel-opt-precise-O3 named lagrange.tomato with 8 processors, by jfe Fri Aug 17 09:59:41 2012 Using Petsc Development HG revision: f9c6cac2d69c724a2258d4e0dc2f51b0825aa875 HG Date: Thu Aug 16 08:37:18 2012 -0700 Max Max/Min Avg Total Time (sec): 6.895e+00 1.00008 6.894e+00 Objects: 3.400e+02 1.00000 3.400e+02 Flops: 7.396e+08 1.83746 4.931e+08 3.945e+09 Flops/sec: 1.073e+08 1.83731 7.153e+07 5.722e+08 MPI Messages: 2.506e+03 1.18017 2.360e+03 1.888e+04 MPI Message Lengths: 3.567e+08 2.05215 1.135e+05 2.144e+09 MPI Reductions: 7.230e+02 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 3.5155e-04 0.0% 0.0000e+00 0.0% 0.000e+00 0.0% 0.000e+00 0.0% 0.000e+00 0.0% 1: Load system: 3.6675e-01 5.3% 0.0000e+00 0.0% 1.470e+02 0.8% 7.604e+03 6.7% 2.400e+01 3.3% 2: KSPSetUpSolve: 6.5274e+00 94.7% 3.9451e+09 100.0% 1.874e+04 99.2% 1.059e+05 93.3% 6.980e+02 96.5% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage PetscBarrier 1 1.0 1.4067e-05 7.4 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 3 0 0 0 0 0 --- Event Stage 1: Load system MatAssemblyBegin 1 1.0 9.1305e-021153.5 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 1 0 0 0 0 12 0 0 0 8 0 MatAssemblyEnd 1 1.0 1.2818e-01 1.0 0.00e+00 0.0 1.1e+02 5.6e+04 8.0e+00 2 0 1 0 1 34 0 76 4 33 0 MatLoad 1 1.0 3.5524e-01 1.0 0.00e+00 0.0 1.3e+02 1.0e+06 1.7e+01 5 0 1 6 2 96 0 90 93 71 0 VecSet 1 1.0 3.3402e-04 6.8 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAssemblyBegin 2 1.0 7.4410e-03106.5 0.00e+00 0.0 0.0e+00 0.0e+00 6.0e+00 0 0 0 0 1 1 0 0 0 25 0 VecAssemblyEnd 2 1.0 2.9087e-05 5.8 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecLoad 2 1.0 1.5213e-02 1.5 0.00e+00 0.0 1.4e+01 7.3e+05 7.0e+00 0 0 0 0 1 4 0 10 7 29 0 --- Event Stage 2: KSPSetUpSolve MatMult 219 1.0 1.6150e+00 1.3 3.11e+08 1.5 1.1e+04 9.8e+04 0.0e+00 21 49 58 50 0 22 49 59 54 0 1192 MatMultAdd 27 1.0 2.3258e-02 1.6 1.86e+06 1.1 1.0e+03 5.6e+03 0.0e+00 0 0 6 0 0 0 0 6 0 0 590 MatMultTranspose 27 1.0 1.0862e-01 8.5 1.86e+06 1.1 1.0e+03 5.6e+03 0.0e+00 1 0 6 0 0 1 0 6 0 0 126 MatSolve 18 0.0 6.6519e-05 0.0 3.41e+04 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 513 MatLUFactorSym 1 1.0 3.9101e-05 3.9 0.00e+00 0.0 0.0e+00 0.0e+00 3.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatLUFactorNum 1 1.0 3.6001e-05 9.4 1.06e+04 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 294 MatConvert 3 1.0 2.0762e-02 1.9 0.00e+00 0.0 0.0e+00 0.0e+00 3.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatScale 3 1.0 2.0531e-02 1.3 3.92e+06 1.5 1.6e+02 8.5e+04 0.0e+00 0 1 1 1 0 0 1 1 1 0 1191 MatAssemblyBegin 37 1.0 1.1074e+0023.3 0.00e+00 0.0 3.9e+02 2.2e+04 4.0e+01 6 0 2 0 6 7 0 2 0 6 0 MatAssemblyEnd 37 1.0 4.9453e-01 1.1 0.00e+00 0.0 1.1e+03 1.5e+04 1.1e+02 7 0 6 1 15 7 0 6 1 16 0 MatGetRow 207104 1.1 4.4730e-02 1.8 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetRowIJ 1 0.0 1.2875e-05 0.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetOrdering 1 0.0 6.3896e-05 0.0 0.00e+00 0.0 0.0e+00 0.0e+00 2.5e-01 0 0 0 0 0 0 0 0 0 0 0 MatCoarsen 3 1.0 1.8429e-01 1.0 0.00e+00 0.0 1.2e+03 1.6e+05 9.0e+01 3 0 6 9 12 3 0 6 10 13 0 MatPtAP 3 1.0 1.8791e-01 1.0 7.55e+06 1.6 1.2e+03 3.0e+04 8.1e+01 3 1 6 2 11 3 1 6 2 12 251 MatPtAPSymbolic 3 1.0 1.3085e-01 1.0 0.00e+00 0.0 1.1e+03 2.8e+04 7.5e+01 2 0 6 1 10 2 0 6 2 11 0 MatPtAPNumeric 3 1.0 5.7091e-02 1.0 7.55e+06 1.6 1.3e+02 4.8e+04 6.0e+00 1 1 1 0 1 1 1 1 0 1 827 MatTrnMatMult 3 1.0 3.5311e+00 1.0 3.28e+08 3.7 9.6e+02 5.4e+05 8.7e+01 51 31 5 24 12 54 31 5 26 12 350 MatGetLocalMat 9 1.0 3.9333e-02 2.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.2e+01 0 0 0 0 2 0 0 0 0 2 0 MatGetBrAoCol 3 1.0 5.4790e-02 1.3 0.00e+00 0.0 4.7e+02 5.7e+04 6.0e+00 1 0 2 1 1 1 0 2 1 1 0 MatGetSymTrans 6 1.0 1.5941e-03 1.5 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecMDot 47 1.0 1.7084e-0113.7 1.79e+07 1.1 0.0e+00 0.0e+00 4.7e+01 1 3 0 0 7 1 3 0 0 7 797 VecNorm 72 1.0 1.1993e-01 6.4 5.28e+06 1.1 0.0e+00 0.0e+00 7.2e+01 1 1 0 0 10 1 1 0 0 10 340 VecScale 168 1.0 1.3566e-02 4.1 5.69e+06 1.1 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 3122 VecCopy 49 1.0 1.4849e-02 2.2 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 182 1.0 1.6513e-02 1.8 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 238 1.0 3.2534e-02 2.2 1.69e+07 1.1 0.0e+00 0.0e+00 0.0e+00 0 3 0 0 0 0 3 0 0 0 3866 VecAYPX 225 1.0 5.9044e-02 2.7 1.01e+07 1.1 0.0e+00 0.0e+00 0.0e+00 0 2 0 0 0 1 2 0 0 0 1272 VecMAXPY 68 1.0 7.0412e-02 3.8 2.80e+07 1.1 0.0e+00 0.0e+00 0.0e+00 1 5 0 0 0 1 5 0 0 0 3060 VecAssemblyBegin 43 1.0 3.5960e-02 5.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.3e+02 0 0 0 0 17 0 0 0 0 18 0 VecAssemblyEnd 43 1.0 5.7697e-05 1.3 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecPointwiseMult 195 1.0 5.8996e-02 2.4 6.73e+06 1.1 0.0e+00 0.0e+00 0.0e+00 1 1 0 0 0 1 1 0 0 0 839 VecScatterBegin 319 1.0 3.9963e-01 2.6 0.00e+00 0.0 1.6e+04 9.2e+04 0.0e+00 4 0 82 67 0 4 0 83 72 0 0 VecScatterEnd 319 1.0 8.3761e-01 4.5 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 8 0 0 0 0 9 0 0 0 0 0 VecSetRandom 3 1.0 3.0630e-03 2.5 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecNormalize 60 1.0 4.5283e-02 4.0 5.87e+06 1.1 0.0e+00 0.0e+00 6.0e+01 0 1 0 0 8 0 1 0 0 9 989 KSPGMRESOrthog 47 1.0 1.9799e-01 5.4 3.59e+07 1.1 0.0e+00 0.0e+00 4.7e+01 2 7 0 0 7 2 7 0 0 7 1376 KSPSetUp 9 1.0 9.8073e-03 1.3 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+01 0 0 0 0 1 0 0 0 0 1 0 KSPSolve 1 1.0 1.5555e+00 1.0 3.31e+08 1.4 1.1e+04 8.2e+04 5.8e+01 23 55 61 44 8 24 55 61 47 8 1385 PCSetUp 2 1.0 4.9523e+00 1.0 4.05e+08 2.7 7.2e+03 1.4e+05 6.4e+02 72 45 38 49 89 76 45 39 52 92 356 PCSetUpOnBlocks 9 1.0 1.9073e-04 2.4 1.06e+04 0.0 0.0e+00 0.0e+00 5.0e+00 0 0 0 0 1 0 0 0 0 1 56 PCApply 9 1.0 1.1458e+00 1.0 2.41e+08 1.4 1.1e+04 6.9e+04 3.2e+01 16 39 56 34 4 17 39 56 36 5 1356 PCGAMGgraph_AGG 3 1.0 4.7532e-01 1.0 3.92e+06 1.5 4.7e+02 4.1e+04 5.7e+01 7 1 2 1 8 7 1 2 1 8 51 PCGAMGcoarse_AGG 3 1.0 3.8397e+00 1.0 3.28e+08 3.7 2.9e+03 2.7e+05 2.4e+02 55 31 16 38 33 58 31 16 40 34 322 PCGAMGProl_AGG 3 1.0 2.1486e-01 1.3 0.00e+00 0.0 9.8e+02 5.3e+04 8.4e+01 3 0 5 2 12 3 0 5 3 12 0 PCGAMGPOpt_AGG 3 1.0 1.0967e-05 3.5 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Viewer 1 0 0 0 --- Event Stage 1: Load system Viewer 1 1 728 0 Matrix 3 0 0 0 Vector 8 1 1504 0 Vector Scatter 1 0 0 0 Index Set 2 2 814104 0 --- Event Stage 2: KSPSetUpSolve Matrix 69 72 158167476 0 Matrix Coarsen 3 3 1860 0 Vector 166 173 58747624 0 Vector Scatter 18 19 20140 0 Index Set 47 47 2462268 0 Krylov Solver 9 9 131888 0 Preconditioner 9 9 8396 0 PetscRandom 3 3 1848 0 ======================================================================================================================== Average time to get PetscTime(): 9.53674e-08 Average time for MPI_Barrier(): 8.10623e-07 Average time for zero size MPI_Send(): 2.74181e-06 #PETSc Option Table entries: -Pressure_ksp_type preonly -Pressure_pc_factor_mat_solver_package mumps -Pressure_pc_type lu -f0 Pressure__3_19_0.mtx -ksp_type gmres -log_summary -pc_type gamg #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure run at: Fri Aug 17 09:14:50 2012 Configure options: --with-x=0 --download-f-blas-lapack=0 --with-blas-lapack-dir=/opt/intel/Compiler/11.1/072/mkl/lib/em64t --with-mpi=1 --with-mpi-shared=1 --with-mpi=1 --download-mpich=no --with-debugging=0 --with-gnu-compilers=no --with-vendor-compilers=intel --with-cc=/usr/local/encap/platform_mpi-8.01/bin/mpicc --with-cxx=/usr/local/encap/platform_mpi-8.01/bin/mpiCC --with-fc=/usr/local/encap/platform_mpi-8.01/bin/mpif90 --with-shared-libraries=1 --with-c++-support --with-clanguage=C --COPTFLAGS="-fPIC -O3 -xSSE4.2 -fp-model precise -g -debug inline_debug_info" --CXXOPTFLAGS="-fPIC -O3 -xSSE4.2 -fp-model precise -g -debug inline_debug_info" --FOPTFLAGS="-fPIC -O3 -xSSE4.2 -fp-model precise -g -debug inline_debug_info" --download-scalapack=1 --download-blacs=1 --with-blacs=1 --download-umfpack=1 --download-parmetis=1 --download-metis=1 --download-superlu=1 --download-superlu_dist=1 --download-mumps=1 --download-ml=1 --download-hypre=1 ----------------------------------------- Libraries compiled on Fri Aug 17 09:14:50 2012 on lagrange.tomato Machine characteristics: Linux-2.6.32-279.2.1.el6.x86_64-x86_64-with-centos-6.3-Final Using PETSc directory: /home/jfe/local/petsc-dev Using PETSc arch: intel-opt-precise-O3 ----------------------------------------- Using C compiler: /usr/local/encap/platform_mpi-8.01/bin/mpicc -fPIC -wd1572 -fPIC -O3 -xSSE4.2 -fp-model precise -g -debug inline_debug_info ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: /usr/local/encap/platform_mpi-8.01/bin/mpif90 -fPIC -fPIC -O3 -xSSE4.2 -fp-model precise -g -debug inline_debug_info ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/home/jfe/local/petsc-dev/intel-opt-precise-O3/include -I/home/jfe/local/petsc-dev/include -I/home/jfe/local/petsc-dev/include -I/home/jfe/local/petsc-dev/intel-opt-precise-O3/include -I/usr/local/encap/platform_mpi-8.01/include ----------------------------------------- Using C linker: /usr/local/encap/platform_mpi-8.01/bin/mpicc Using Fortran linker: /usr/local/encap/platform_mpi-8.01/bin/mpif90 Using libraries: -Wl,-rpath,/home/jfe/local/petsc-dev/intel-opt-precise-O3/lib -L/home/jfe/local/petsc-dev/intel-opt-precise-O3/lib -lpetsc -Wl,-rpath,/home/jfe/local/petsc-dev/intel-opt-precise-O3/lib -L/home/jfe/local/petsc-dev/intel-opt-precise-O3/lib -lcmumps -ldmumps -lsmumps -lzmumps -lmumps_common -lpord -lscalapack -lblacs -lml -Wl,-rpath,/usr/local/encap/platform_mpi-8.01/lib/linux_amd64 -L/usr/local/encap/platform_mpi-8.01/lib/linux_amd64 -lmpiCC -Wl,-rpath,/opt/intel/Compiler/11.1/072/lib/intel64 -L/opt/intel/Compiler/11.1/072/lib/intel64 -Wl,-rpath,/opt/intel/Compiler/11.1/072/ipp/em64t/lib -L/opt/intel/Compiler/11.1/072/ipp/em64t/lib -Wl,-rpath,/opt/intel/Compiler/11.1/072/mkl/lib/em64t -L/opt/intel/Compiler/11.1/072/mkl/lib/em64t -Wl,-rpath,/opt/intel/Compiler/11.1/072/tbb/intel64/cc4.1.0_libc2.4_kernel2.6.16.21/lib -L/opt/intel/Compiler/11.1/072/tbb/intel64/cc4.1.0_libc2.4_kernel2.6.16.21/lib -Wl,-rpath,/usr/lib/gcc/x86_64-redhat-linux/4.4.6 -L/usr/lib/gcc/x86_64-redhat-linux/4.4.6 -lstdc++ -lsuperlu_dist_3.0 -lparmetis -lmetis -lpthread -lsuperlu_4.3 -lHYPRE -lmpiCC -lstdc++ -lumfpack -lamd -lmkl_intel_lp64 -lmkl_intel_thread -lmkl_core -liomp5 -lpthread -lifport -lifcore -lm -lpthread -lm -lmpiCC -lstdc++ -lmpiCC -lstdc++ -lpcmpio -lpcmpi -ldl -limf -lsvml -lipgo -ldecimal -lgcc_s -lirc -lirc_s -ldl -----------------------------------------