Norm of error 4.10214, Iterations 9 Norm of error 17.4413, Iterations 4 ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./ex5 on a linux64_openmpi_pthread_openmp named ubuntu with 2 processors, by albuquerque Thu Feb 14 13:50:23 2013 Using Petsc Development HG revision: 75225b570bbaeb0a1163656ce70dfd4157a66dad HG Date: Tue Jan 15 23:55:44 2013 -0600 Max Max/Min Avg Total Time (sec): 4.760e-02 1.00007 4.760e-02 Objects: 4.000e+01 1.00000 4.000e+01 Flops: 1.120e+07 1.00000 1.120e+07 2.240e+07 Flops/sec: 2.353e+08 1.00007 2.353e+08 4.706e+08 MPI Messages: 2.000e+01 1.00000 2.000e+01 4.000e+01 MPI Message Lengths: 7.205e+05 1.00000 3.602e+04 1.441e+06 MPI Reductions: 6.600e+01 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 2.4366e-04 0.5% 0.0000e+00 0.0% 0.000e+00 0.0% 0.000e+00 0.0% 0.000e+00 0.0% 1: Original Solve: 2.8448e-02 59.8% 1.6199e+07 72.3% 2.400e+01 60.0% 1.700e+01 0.0% 5.000e+01 75.8% 2: Second Solve: 1.8905e-02 39.7% 6.1997e+06 27.7% 1.600e+01 40.0% 3.601e+04 100.0% 1.500e+01 22.7% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length (bytes) Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage --- Event Stage 1: Original Solve ThreadCommRunKer 57 1.0 9.2106e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 19 0 0 0 0 32 0 0 0 0 0 MatMult 10 1.0 4.7765e-03 1.0 1.80e+06 1.0 2.0e+01 3.2e+01 0.0e+00 10 16 50 0 0 17 22 83 94 0 754 MatSolve 10 1.0 3.3300e-03 1.0 1.60e+06 1.0 0.0e+00 0.0e+00 0.0e+00 7 14 0 0 0 12 20 0 0 0 961 MatCholFctrNum 1 1.0 1.6921e-03 1.0 2.00e+04 1.0 0.0e+00 0.0e+00 0.0e+00 4 0 0 0 0 6 0 0 0 0 24 MatICCFactorSym 1 1.0 6.9380e-04 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 1 0 0 0 2 2 0 0 0 2 0 MatAssemblyBegin 1 1.0 1.4019e-04 3.6 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 3 0 0 0 0 4 0 MatAssemblyEnd 1 1.0 1.7409e-03 1.0 0.00e+00 0.0 4.0e+00 1.0e+01 9.0e+00 4 0 10 0 14 6 0 17 6 18 0 MatGetRowIJ 1 1.0 9.5367e-07 0.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetOrdering 1 1.0 1.5903e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 3 1 0 0 0 4 0 VecMDot 9 1.0 2.5678e-03 1.0 1.80e+06 1.0 0.0e+00 0.0e+00 9.0e+00 5 16 0 0 14 9 22 0 0 18 1402 VecNorm 11 1.0 4.2129e-04 1.1 4.40e+05 1.0 0.0e+00 0.0e+00 1.1e+01 1 4 0 0 17 1 5 0 0 22 2089 VecScale 10 1.0 2.4605e-04 1.3 2.00e+05 1.0 0.0e+00 0.0e+00 0.0e+00 0 2 0 0 0 1 2 0 0 0 1626 VecCopy 1 1.0 9.1076e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 13 1.0 4.7469e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 2 0 0 0 0 0 VecAXPY 2 1.0 1.9908e-04 1.5 8.00e+04 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 1 1 0 0 0 804 VecMAXPY 10 1.0 1.3618e-03 1.0 2.16e+06 1.0 0.0e+00 0.0e+00 0.0e+00 3 19 0 0 0 5 27 0 0 0 3172 VecAssemblyBegin 1 1.0 4.3869e-05 2.7 0.00e+00 0.0 0.0e+00 0.0e+00 3.0e+00 0 0 0 0 5 0 0 0 0 6 0 VecAssemblyEnd 1 1.0 9.5367e-07 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecScatterBegin 10 1.0 6.3658e-05 1.1 0.00e+00 0.0 2.0e+01 3.2e+01 0.0e+00 0 0 50 0 0 0 0 83 94 0 0 VecScatterEnd 10 1.0 1.3685e-04 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecNormalize 10 1.0 6.4111e-04 1.1 6.00e+05 1.0 0.0e+00 0.0e+00 1.0e+01 1 5 0 0 15 2 7 0 0 20 1872 KSPGMRESOrthog 9 1.0 3.6473e-03 1.0 3.60e+06 1.0 0.0e+00 0.0e+00 9.0e+00 8 32 0 0 14 13 44 0 0 18 1974 KSPSetUp 2 1.0 5.6934e-04 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 1 0 0 0 3 2 0 0 0 4 0 KSPSolve 1 1.0 1.6333e-02 1.0 7.84e+06 1.0 1.8e+01 3.2e+01 2.4e+01 34 70 45 0 36 57 97 75 85 48 960 PCSetUp 2 1.0 2.6329e-03 1.0 2.00e+04 1.0 0.0e+00 0.0e+00 7.0e+00 6 0 0 0 11 9 0 0 0 14 15 PCSetUpOnBlocks 1 1.0 2.5780e-03 1.0 2.00e+04 1.0 0.0e+00 0.0e+00 5.0e+00 5 0 0 0 8 9 0 0 0 10 16 PCApply 10 1.0 3.9036e-03 1.0 1.60e+06 1.0 0.0e+00 0.0e+00 0.0e+00 8 14 0 0 0 14 20 0 0 0 820 --- Event Stage 2: Second Solve ThreadCommRunKer 31 1.0 3.9654e-03 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 8 0 0 0 0 20 0 0 0 0 0 MatMult 5 1.0 2.9578e-03 1.2 9.00e+05 1.0 1.0e+01 3.2e+01 0.0e+00 6 8 25 0 0 15 29 62 0 0 609 MatSolve 5 1.0 1.7729e-03 1.0 8.00e+05 1.0 0.0e+00 0.0e+00 0.0e+00 4 7 0 0 0 9 26 0 0 0 902 MatCholFctrNum 1 1.0 1.7290e-03 1.3 2.00e+04 1.0 0.0e+00 0.0e+00 0.0e+00 3 0 0 0 0 8 1 0 0 0 23 MatAssemblyBegin 1 1.0 1.3220e-03 1.2 0.00e+00 0.0 6.0e+00 2.4e+05 2.0e+00 3 0 15100 3 6 0 38100 13 0 MatAssemblyEnd 1 1.0 2.7099e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 6 0 0 0 2 14 0 0 0 7 0 MatGetRowIJ 1 1.0 9.5367e-07 0.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetOrdering 1 1.0 2.1195e-04 1.4 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 3 1 0 0 0 13 0 MatZeroEntries 1 1.0 1.7405e-04 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 1 0 0 0 0 0 VecMDot 4 1.0 9.0981e-04 1.4 4.00e+05 1.0 0.0e+00 0.0e+00 4.0e+00 2 4 0 0 6 4 13 0 0 27 879 VecNorm 6 1.0 7.0620e-04 3.7 2.40e+05 1.0 0.0e+00 0.0e+00 6.0e+00 1 2 0 0 9 2 8 0 0 40 680 VecScale 5 1.0 8.9884e-05 1.1 1.00e+05 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 3 0 0 0 2225 VecCopy 1 1.0 7.6056e-05 1.3 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 7 1.0 2.9087e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 2 0 0 0 0 0 VecAXPY 2 1.0 1.0872e-04 1.0 8.00e+04 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 1 3 0 0 0 1472 VecMAXPY 5 1.0 2.5630e-04 1.0 5.60e+05 1.0 0.0e+00 0.0e+00 0.0e+00 1 5 0 0 0 1 18 0 0 0 4370 VecScatterBegin 5 1.0 2.1219e-05 1.1 0.00e+00 0.0 1.0e+01 3.2e+01 0.0e+00 0 0 25 0 0 0 0 62 0 0 0 VecScatterEnd 5 1.0 8.5354e-05 1.3 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecNormalize 5 1.0 7.7367e-04 3.1 3.00e+05 1.0 0.0e+00 0.0e+00 5.0e+00 1 3 0 0 8 3 10 0 0 33 776 KSPGMRESOrthog 4 1.0 1.1010e-03 1.3 8.00e+05 1.0 0.0e+00 0.0e+00 4.0e+00 2 7 0 0 6 5 26 0 0 27 1453 KSPSetUp 2 1.0 2.4080e-05 1.6 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 1 1.0 7.8728e-03 1.0 2.84e+06 1.0 8.0e+00 3.2e+01 1.1e+01 16 25 20 0 17 41 92 50 0 73 721 PCSetUp 2 1.0 1.9600e-03 1.3 2.00e+04 1.0 0.0e+00 0.0e+00 2.0e+00 4 0 0 0 3 9 1 0 0 13 20 PCSetUpOnBlocks 1 1.0 1.9569e-03 1.3 2.00e+04 1.0 0.0e+00 0.0e+00 2.0e+00 4 0 0 0 3 9 1 0 0 13 20 PCApply 5 1.0 2.0721e-03 1.0 8.00e+05 1.0 0.0e+00 0.0e+00 0.0e+00 4 7 0 0 0 11 26 0 0 0 772 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Viewer 1 0 0 0 --- Event Stage 1: Original Solve Matrix 4 0 0 0 Vector 23 2 163088 0 Vector Scatter 1 0 0 0 Index Set 5 3 2336 0 Krylov Solver 2 0 0 0 Preconditioner 2 0 0 0 --- Event Stage 2: Second Solve Matrix 0 4 2990500 0 Vector 0 21 2912264 0 Vector Scatter 0 1 1084 0 Index Set 2 4 163104 0 Krylov Solver 0 2 19472 0 Preconditioner 0 2 1816 0 ======================================================================================================================== Average time to get PetscTime(): 9.53674e-08 Average time for MPI_Barrier(): 1.23978e-06 Average time for zero size MPI_Send(): 5.00679e-06 #PETSc Option Table entries: -log_summary -m 10000 #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure run at: Thu Feb 14 10:14:19 2013 Configure options: --PETSC_ARCH=linux64_openmpi_pthread_openmp --with-clanguage=cxx --with-shared-libraries=0 --with-c2html=0 --with-debugging=0 --CC=gcc --CXX=g++ --with-fortran=0 --with-ar=ar --with-ranlib=ranlib --COPTFLAGS=-O2 --CXXOPTFLAGS=-O2 --with-mpi=1 --with-mpi-dir=/usr/lib/openmpi --with-pthreadclasses --with-threadcomm --with-openmp --with-threadcomm --with-x=0 ----------------------------------------- Libraries compiled on Thu Feb 14 10:14:19 2013 on ubuntu Machine characteristics: Linux-3.2.0-29-generic-x86_64-with-Ubuntu-12.04-precise Using PETSc directory: /home/albuquerque/dev/petsc-dev Using PETSc arch: linux64_openmpi_pthread_openmp ----------------------------------------- Using C compiler: g++ -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O2 -fopenmp ${COPTFLAGS} ${CFLAGS} ----------------------------------------- Using include paths: -I/home/albuquerque/dev/petsc-dev/linux64_openmpi_pthread_openmp/include -I/home/albuquerque/dev/petsc-dev/include -I/home/albuquerque/dev/petsc-dev/include -I/home/albuquerque/dev/petsc-dev/linux64_openmpi_pthread_openmp/include -I/usr/lib/openmpi/include ----------------------------------------- Using C linker: g++ Using libraries: -Wl,-rpath,/home/albuquerque/dev/petsc-dev/linux64_openmpi_pthread_openmp/lib -L/home/albuquerque/dev/petsc-dev/linux64_openmpi_pthread_openmp/lib -lpetsc -llapack -lblas -lpthread -Wl,-rpath,/usr/lib/openmpi/lib -L/usr/lib/openmpi/lib -lmpi++ -lmpi -lm -L/usr/lib/gcc/x86_64-linux-gnu/4.6 -L/usr/lib/x86_64-linux-gnu -L/lib/x86_64-linux-gnu -lstdc++ -ldl -lgcc_s -ldl -----------------------------------------