************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- test on a linux64_openmpi_pthread_openmp named ubuntu with 2 processors, by albuquerque Thu Feb 14 12:04:59 2013 Using Petsc Development HG revision: 75225b570bbaeb0a1163656ce70dfd4157a66dad HG Date: Tue Jan 15 23:55:44 2013 -0600 Max Max/Min Avg Total Time (sec): 5.406e+00 1.26180 4.845e+00 Objects: 5.600e+01 1.00000 5.600e+01 Flops: 3.114e+09 1.00316 3.110e+09 6.219e+09 Flops/sec: 7.247e+08 1.25783 6.504e+08 1.301e+09 MPI Messages: 4.489e+03 1.00000 4.489e+03 8.978e+03 MPI Message Lengths: 1.277e+08 1.00000 2.844e+04 2.553e+08 MPI Reductions: 8.867e+03 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 4.7853e+00 98.8% 6.2192e+09 100.0% 8.978e+03 100.0% 2.844e+04 100.0% 8.866e+03 100.0% 1: ReadGeometry: 5.9111e-02 1.2% 0.0000e+00 0.0% 0.000e+00 0.0% 0.000e+00 0.0% 0.000e+00 0.0% 2: SaveSolution: 5.1105e-04 0.0% 0.0000e+00 0.0% 0.000e+00 0.0% 0.000e+00 0.0% 0.000e+00 0.0% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length (bytes) Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage ThreadCommRunKer 22888 1.0 3.1603e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 65 0 0 0 0 66 0 0 0 0 0 MatMult 4487 1.0 1.2912e+00 1.0 5.45e+08 1.0 9.0e+03 2.8e+04 0.0e+00 27 17100100 0 27 17100100 0 836 MatAssemblyBegin 1 1.0 3.7193e-05 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyEnd 1 1.0 4.3511e-03 1.0 0.00e+00 0.0 4.0e+00 7.1e+03 8.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatZeroEntries 1 1.0 3.8741e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecMDot 4343 1.0 1.4131e+00 1.0 1.17e+09 1.0 0.0e+00 0.0e+00 4.3e+03 29 38 0 0 49 29 38 0 0 49 1651 VecNorm 4488 1.0 9.7212e-02 1.1 7.79e+07 1.0 0.0e+00 0.0e+00 4.5e+03 2 3 0 0 51 2 3 0 0 51 1602 VecScale 4488 1.0 3.6017e-02 1.0 3.89e+07 1.0 0.0e+00 0.0e+00 0.0e+00 1 1 0 0 0 1 1 0 0 0 2162 VecCopy 145 1.0 5.2550e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 156 1.0 3.0172e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 289 1.0 5.5876e-03 1.0 5.02e+06 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 1795 VecMAXPY 4488 1.0 8.5717e-01 1.0 1.24e+09 1.0 0.0e+00 0.0e+00 0.0e+00 18 40 0 0 0 18 40 0 0 0 2898 VecAssemblyBegin 1 1.0 2.0027e-05 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 3.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAssemblyEnd 1 1.0 9.5367e-07 0.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecPointwiseMult 4488 1.0 1.6081e-01 1.1 3.89e+07 1.0 0.0e+00 0.0e+00 0.0e+00 3 1 0 0 0 3 1 0 0 0 484 VecScatterBegin 4489 1.0 6.5872e-02 1.4 0.00e+00 0.0 9.0e+03 2.8e+04 2.0e+00 1 0100100 0 1 0100100 0 0 VecScatterEnd 4487 1.0 2.5025e-01 2.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 4 0 0 0 0 4 0 0 0 0 0 VecNormalize 4488 1.0 1.3857e-01 1.0 1.17e+08 1.0 0.0e+00 0.0e+00 4.5e+03 3 4 0 0 51 3 4 0 0 51 1686 KSPGMRESOrthog 4343 1.0 2.2028e+00 1.0 2.33e+09 1.0 0.0e+00 0.0e+00 4.3e+03 45 75 0 0 49 46 75 0 0 49 2119 KSPSetUp 1 1.0 1.2207e-04 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 1 1.0 3.8972e+00 1.0 3.11e+09 1.0 9.0e+03 2.8e+04 8.8e+03 80100100100100 81100100100100 1596 PCSetUp 1 1.0 0.0000e+00 0.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 PCApply 4488 1.0 1.6385e-01 1.0 3.89e+07 1.0 0.0e+00 0.0e+00 2.0e+00 3 1 0 0 0 3 1 0 0 0 475 --- Event Stage 1: ReadGeometry --- Event Stage 2: SaveSolution ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Matrix 3 0 0 0 Vector 43 1 1544 0 Vector Scatter 3 2 1304 0 Index Set 4 4 22648 0 Krylov Solver 1 0 0 0 Preconditioner 1 0 0 0 Viewer 1 0 0 0 --- Event Stage 1: ReadGeometry --- Event Stage 2: SaveSolution ======================================================================================================================== Average time to get PetscTime(): 0 Average time for MPI_Barrier(): 1.38283e-06 Average time for zero size MPI_Send(): 7.03335e-06 #PETSc Option Table entries: -log_summary -pc_type jacobi -threadcomm_type nothread #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure run at: Thu Feb 14 10:14:19 2013 Configure options: --PETSC_ARCH=linux64_openmpi_pthread_openmp --with-clanguage=cxx --with-shared-libraries=0 --with-c2html=0 --with-debugging=0 --CC=gcc --CXX=g++ --with-fortran=0 --with-ar=ar --with-ranlib=ranlib --COPTFLAGS=-O2 --CXXOPTFLAGS=-O2 --with-mpi=1 --with-mpi-dir=/usr/lib/openmpi --with-pthreadclasses --with-threadcomm --with-openmp --with-threadcomm --with-x=0 ----------------------------------------- Libraries compiled on Thu Feb 14 10:14:19 2013 on ubuntu Machine characteristics: Linux-3.2.0-29-generic-x86_64-with-Ubuntu-12.04-precise Using PETSc directory: /home/albuquerque/dev/petsc-dev Using PETSc arch: linux64_openmpi_pthread_openmp ----------------------------------------- Using C compiler: g++ -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O2 -fopenmp ${COPTFLAGS} ${CFLAGS} ----------------------------------------- Using include paths: -I/home/albuquerque/dev/petsc-dev/linux64_openmpi_pthread_openmp/include -I/home/albuquerque/dev/petsc-dev/include -I/home/albuquerque/dev/petsc-dev/include -I/home/albuquerque/dev/petsc-dev/linux64_openmpi_pthread_openmp/include -I/usr/lib/openmpi/include ----------------------------------------- Using C linker: g++ Using libraries: -Wl,-rpath,/home/albuquerque/dev/petsc-dev/linux64_openmpi_pthread_openmp/lib -L/home/albuquerque/dev/petsc-dev/linux64_openmpi_pthread_openmp/lib -lpetsc -llapack -lblas -lpthread -Wl,-rpath,/usr/lib/openmpi/lib -L/usr/lib/openmpi/lib -lmpi++ -lmpi -lm -L/usr/lib/gcc/x86_64-linux-gnu/4.6 -L/usr/lib/x86_64-linux-gnu -L/lib/x86_64-linux-gnu -lstdc++ -ldl -lgcc_s -ldl -----------------------------------------