Residual norm 1.67183e-07 ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./ex45 on a arch-linu named kazan with 1 processor, by fpoulin Fri Feb 24 21:13:28 2012 Using Petsc Release Version 3.2.0, Patch 6, Wed Jan 11 09:28:45 CST 2012 Max Max/Min Avg Total Time (sec): 1.634e+02 1.00000 1.634e+02 Objects: 1.900e+02 1.00000 1.900e+02 Flops: 7.226e+09 1.00000 7.226e+09 7.226e+09 Flops/sec: 4.422e+07 1.00000 4.422e+07 4.422e+07 MPI Messages: 0.000e+00 0.00000 0.000e+00 0.000e+00 MPI Message Lengths: 0.000e+00 0.00000 0.000e+00 0.000e+00 MPI Reductions: 2.530e+02 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 1.6342e+02 100.0% 7.2263e+09 100.0% 0.000e+00 0.0% 0.000e+00 0.0% 2.520e+02 99.6% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage KSPGMRESOrthog 3 1.0 8.3762e+00 1.0 4.07e+08 1.0 0.0e+00 0.0e+00 0.0e+00 5 6 0 0 0 5 6 0 0 0 49 KSPSetup 8 1.0 8.0366e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 3.4e+01 0 0 0 0 13 0 0 0 0 13 0 KSPSolve 1 1.0 6.8712e+01 1.0 6.81e+09 1.0 0.0e+00 0.0e+00 2.0e+01 42 94 0 0 8 42 94 0 0 8 99 VecMDot 3 1.0 6.8158e+00 1.0 2.04e+08 1.0 0.0e+00 0.0e+00 0.0e+00 4 3 0 0 0 4 3 0 0 0 30 VecNorm 5 1.0 1.6604e+00 1.0 1.70e+08 1.0 0.0e+00 0.0e+00 0.0e+00 1 2 0 0 0 1 2 0 0 0 102 VecScale 4 1.0 3.2112e-01 1.0 6.79e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 211 VecCopy 1 1.0 8.2686e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 71 1.0 6.5225e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 2 1.0 2.1944e-01 1.0 6.79e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 309 VecAYPX 24 1.0 2.7411e+00 1.0 7.77e+07 1.0 0.0e+00 0.0e+00 0.0e+00 2 1 0 0 0 2 1 0 0 0 28 VecMAXPY 4 1.0 1.9590e+00 1.0 3.06e+08 1.0 0.0e+00 0.0e+00 0.0e+00 1 4 0 0 0 1 4 0 0 0 156 VecNormalize 4 1.0 1.6190e+00 1.0 2.04e+08 1.0 0.0e+00 0.0e+00 0.0e+00 1 3 0 0 0 1 3 0 0 0 126 PCSetUp 1 1.0 2.8204e+01 1.0 1.31e+08 1.0 0.0e+00 0.0e+00 2.0e+02 17 2 0 0 81 17 2 0 0 81 5 PCApply 4 1.0 5.4198e+01 1.0 5.40e+09 1.0 0.0e+00 0.0e+00 0.0e+00 33 75 0 0 0 33 75 0 0 0 100 MatMult 28 1.0 6.4484e+00 1.0 1.89e+09 1.0 0.0e+00 0.0e+00 0.0e+00 4 26 0 0 0 4 26 0 0 0 292 MatMultAdd 24 1.0 2.6531e+00 1.0 5.22e+08 1.0 0.0e+00 0.0e+00 0.0e+00 2 7 0 0 0 2 7 0 0 0 197 MatMultTranspose 30 1.0 7.4326e+00 1.0 6.53e+08 1.0 0.0e+00 0.0e+00 0.0e+00 5 9 0 0 0 5 9 0 0 0 88 MatSolve 4 1.0 1.4782e-04 1.0 2.35e+04 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 159 MatSOR 48 1.0 3.8607e+01 1.0 3.27e+09 1.0 0.0e+00 0.0e+00 0.0e+00 24 45 0 0 0 24 45 0 0 0 85 MatLUFactorSym 1 1.0 2.8801e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 3.0e+00 0 0 0 0 1 0 0 0 0 1 0 MatLUFactorNum 1 1.0 2.0790e-04 1.0 9.23e+03 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 44 MatAssemblyBegin 20 1.0 8.1301e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyEnd 20 1.0 4.4871e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 3 0 0 0 0 3 0 0 0 0 0 MatGetRowIJ 1 1.0 2.2888e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetOrdering 1 1.0 3.6001e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 1 0 0 0 0 1 0 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Krylov Solver 8 8 25888 0 Distributed Mesh 7 7 155534824 0 Vector 93 93 3320496176 0 Vector Scatter 14 14 8680 0 Index Set 38 38 77779660 0 IS L to G Mapping 7 7 77754392 0 Preconditioner 8 8 7368 0 Matrix 14 14 3498612548 0 Viewer 1 0 0 0 ======================================================================================================================== Average time to get PetscTime(): 5.00679e-07 #PETSc Option Table entries: -da_grid_x 5 -da_grid_y 5 -da_grid_z 5 -da_refine 6 -log_summary -mg_levels_ksp_type richardson -mg_levels_pc_type sor -pc_type mg #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 Configure run at: Fri Feb 24 18:10:35 2012 Configure options: --with-blas-lapack-dir=/opt/intel/mkl/10.1.1.019/lib/64/ --with-mpi-dir=/opt/sgi/mpt/mpt-1.26/ --with-cc=icc --with-fc=ifort --with-debugging=0 ----------------------------------------- Libraries compiled on Fri Feb 24 18:10:35 2012 on kazan Machine characteristics: Linux-2.6.16.60-0.54.5-default-ia64-with-SuSE-10-ia64 Using PETSc directory: /home/fpoulin/soft/petsc-3.2-p6 Using PETSc arch: arch-linux2-c-debug ----------------------------------------- Using C compiler: icc -wd1572 -Qoption,cpp,--extended_float_type -O3 ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: ifort -O3 ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/home/fpoulin/soft/petsc-3.2-p6/arch-linux2-c-debug/include -I/home/fpoulin/soft/petsc-3.2-p6/include -I/home/fpoulin/soft/petsc-3.2-p6/include -I/home/fpoulin/soft/petsc-3.2-p6/arch-linux2-c-debug/include -I/opt/sgi/mpt/mpt-1.26/include ----------------------------------------- Using C linker: icc Using Fortran linker: ifort Using libraries: -Wl,-rpath,/home/fpoulin/soft/petsc-3.2-p6/arch-linux2-c-debug/lib -L/home/fpoulin/soft/petsc-3.2-p6/arch-linux2-c-debug/lib -lpetsc -lpthread -Wl,-rpath,/opt/intel/mkl/10.1.1.019/lib/64 -L/opt/intel/mkl/10.1.1.019/lib/64 -lmkl_lapack -lmkl -lguide -lpthread -Wl,-rpath,/opt/sgi/mpt/mpt-1.26/lib -L/opt/sgi/mpt/mpt-1.26/lib -lfmpich2g -lmpi -lPEPCF90 -ldl -L/opt/intel/Compiler/11.0/074/ipp/ia64/lib -L/opt/intel/Compiler/11.0/074/mkl/lib/64 -L/opt/intel/Compiler/11.0/074/tbb/itanium/cc4.1.0_libc2.4_kernel2.6.16.21/lib -L/opt/intel/Compiler/11.0/074/lib/ia64 -L/home/fpoulin/soft/petsc-3.2-p6/opt/intel/Compiler/11.0/074/mkl/lib/64 -L/usr/lib/gcc/ia64-suse-linux/4.1.2 -L/usr/ia64-suse-linux/lib -limf -lipgo -lirc -lipr -lgcc_s -lirc_s -L/opt/intel/fc/10.1.021/lib -lifport -lifcore -lm -lm -ldl -limf -lipgo -lirc -lipr -lgcc_s -lirc_s -ldl -----------------------------------------