************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- test on a linux64_openmpi_pthread_openmp named ubuntu with 2 processors, by albuquerque Thu Feb 14 12:12:00 2013 Using Petsc Development HG revision: 75225b570bbaeb0a1163656ce70dfd4157a66dad HG Date: Tue Jan 15 23:55:44 2013 -0600 Max Max/Min Avg Total Time (sec): 3.222e+02 1.00637 3.212e+02 Objects: 5.600e+01 1.00000 5.600e+01 Flops: 3.114e+09 1.00316 3.110e+09 6.219e+09 Flops/sec: 9.698e+06 1.00320 9.683e+06 1.937e+07 MPI Messages: 4.489e+03 1.00000 4.489e+03 8.978e+03 MPI Message Lengths: 1.277e+08 1.00000 2.844e+04 2.553e+08 MPI Reductions: 8.867e+03 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 3.2104e+02 100.0% 6.2192e+09 100.0% 8.978e+03 100.0% 2.844e+04 100.0% 8.866e+03 100.0% 1: ReadGeometry: 1.1426e-01 0.0% 0.0000e+00 0.0% 0.000e+00 0.0% 0.000e+00 0.0% 0.000e+00 0.0% 2: SaveSolution: 2.7895e-04 0.0% 0.0000e+00 0.0% 0.000e+00 0.0% 0.000e+00 0.0% 0.000e+00 0.0% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length (bytes) Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage ThreadCommRunKer 22888 1.0 2.3912e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 0 MatMult 4487 1.0 2.1345e+02 1.0 5.45e+08 1.0 9.0e+03 2.8e+04 0.0e+00 66 17100100 0 66 17100100 0 5 MatAssemblyBegin 1 1.0 2.4053e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyEnd 1 1.0 1.4430e-01 1.2 0.00e+00 0.0 4.0e+00 7.1e+03 8.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatZeroEntries 1 1.0 2.5580e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecMDot 4343 1.0 5.2464e+01 1.0 1.17e+09 1.0 0.0e+00 0.0e+00 4.3e+03 16 38 0 0 49 16 38 0 0 49 44 VecNorm 4488 1.0 5.2970e+01 1.0 7.79e+07 1.0 0.0e+00 0.0e+00 4.5e+03 16 3 0 0 51 16 3 0 0 51 3 VecScale 4488 1.0 5.8137e-02 1.0 3.89e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 1340 VecCopy 145 1.0 4.2131e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 156 1.0 2.1751e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 289 1.0 4.9455e-03 1.0 5.02e+06 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 2028 VecMAXPY 4488 1.0 6.0319e-01 1.0 1.24e+09 1.0 0.0e+00 0.0e+00 0.0e+00 0 40 0 0 0 0 40 0 0 0 4118 VecAssemblyBegin 1 1.0 4.7652e-02 2.0 0.00e+00 0.0 0.0e+00 0.0e+00 3.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAssemblyEnd 1 1.0 1.9073e-06 2.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecPointwiseMult 4488 1.0 1.3222e-01 1.0 3.89e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 589 VecScatterBegin 4489 1.0 1.6264e-01 1.2 0.00e+00 0.0 9.0e+03 2.8e+04 2.0e+00 0 0100100 0 0 0100100 0 0 VecScatterEnd 4487 1.0 2.1256e+02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 66 0 0 0 0 66 0 0 0 0 0 VecNormalize 4488 1.0 5.3035e+01 1.0 1.17e+08 1.0 0.0e+00 0.0e+00 4.5e+03 16 4 0 0 51 16 4 0 0 51 4 KSPGMRESOrthog 4343 1.0 5.3036e+01 1.0 2.33e+09 1.0 0.0e+00 0.0e+00 4.3e+03 16 75 0 0 49 16 75 0 0 49 88 KSPSetUp 1 1.0 8.2970e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 1 1.0 3.1901e+02 1.0 3.11e+09 1.0 9.0e+03 2.8e+04 8.8e+03 99100100100100 99100100100100 19 PCSetUp 1 1.0 9.5367e-07 0.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 PCApply 4488 1.0 1.6012e-01 1.0 3.89e+07 1.0 0.0e+00 0.0e+00 2.0e+00 0 1 0 0 0 0 1 0 0 0 486 --- Event Stage 1: ReadGeometry --- Event Stage 2: SaveSolution ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Matrix 3 0 0 0 Vector 43 1 1544 0 Vector Scatter 3 2 1304 0 Index Set 4 4 22648 0 Krylov Solver 1 0 0 0 Preconditioner 1 0 0 0 Viewer 1 0 0 0 --- Event Stage 1: ReadGeometry --- Event Stage 2: SaveSolution ======================================================================================================================== Average time to get PetscTime(): 0 Average time for MPI_Barrier(): 0.00870218 Average time for zero size MPI_Send(): 0.00614798 #PETSc Option Table entries: -log_summary -pc_type jacobi -threadcomm_nthreads 1 -threadcomm_type openmp #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure run at: Thu Feb 14 10:14:19 2013 Configure options: --PETSC_ARCH=linux64_openmpi_pthread_openmp --with-clanguage=cxx --with-shared-libraries=0 --with-c2html=0 --with-debugging=0 --CC=gcc --CXX=g++ --with-fortran=0 --with-ar=ar --with-ranlib=ranlib --COPTFLAGS=-O2 --CXXOPTFLAGS=-O2 --with-mpi=1 --with-mpi-dir=/usr/lib/openmpi --with-pthreadclasses --with-threadcomm --with-openmp --with-threadcomm --with-x=0 ----------------------------------------- Libraries compiled on Thu Feb 14 10:14:19 2013 on ubuntu Machine characteristics: Linux-3.2.0-29-generic-x86_64-with-Ubuntu-12.04-precise Using PETSc directory: /home/albuquerque/dev/petsc-dev Using PETSc arch: linux64_openmpi_pthread_openmp ----------------------------------------- Using C compiler: g++ -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O2 -fopenmp ${COPTFLAGS} ${CFLAGS} ----------------------------------------- Using include paths: -I/home/albuquerque/dev/petsc-dev/linux64_openmpi_pthread_openmp/include -I/home/albuquerque/dev/petsc-dev/include -I/home/albuquerque/dev/petsc-dev/include -I/home/albuquerque/dev/petsc-dev/linux64_openmpi_pthread_openmp/include -I/usr/lib/openmpi/include ----------------------------------------- Using C linker: g++ Using libraries: -Wl,-rpath,/home/albuquerque/dev/petsc-dev/linux64_openmpi_pthread_openmp/lib -L/home/albuquerque/dev/petsc-dev/linux64_openmpi_pthread_openmp/lib -lpetsc -llapack -lblas -lpthread -Wl,-rpath,/usr/lib/openmpi/lib -L/usr/lib/openmpi/lib -lmpi++ -lmpi -lm -L/usr/lib/gcc/x86_64-linux-gnu/4.6 -L/usr/lib/x86_64-linux-gnu -L/lib/x86_64-linux-gnu -lstdc++ -ldl -lgcc_s -ldl -----------------------------------------