Norm of error 4.10214, Iterations 9 Norm of error 17.4413, Iterations 4 ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./ex5 on a linux64_openmpi_pthread_openmp named ubuntu with 2 processors, by albuquerque Thu Feb 14 13:48:39 2013 Using Petsc Development HG revision: 75225b570bbaeb0a1163656ce70dfd4157a66dad HG Date: Tue Jan 15 23:55:44 2013 -0600 Max Max/Min Avg Total Time (sec): 1.018e+00 1.01187 1.012e+00 Objects: 4.000e+01 1.00000 4.000e+01 Flops: 1.120e+07 1.00000 1.120e+07 2.240e+07 Flops/sec: 1.114e+07 1.01187 1.107e+07 2.214e+07 MPI Messages: 2.000e+01 1.00000 2.000e+01 4.000e+01 MPI Message Lengths: 7.205e+05 1.00000 3.602e+04 1.441e+06 MPI Reductions: 6.600e+01 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 4.0584e-02 4.0% 0.0000e+00 0.0% 0.000e+00 0.0% 0.000e+00 0.0% 0.000e+00 0.0% 1: Original Solve: 6.2304e-01 61.6% 1.6199e+07 72.3% 2.400e+01 60.0% 1.700e+01 0.0% 5.000e+01 75.8% 2: Second Solve: 3.4808e-01 34.4% 6.1997e+06 27.7% 1.600e+01 40.0% 3.601e+04 100.0% 1.500e+01 22.7% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length (bytes) Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage --- Event Stage 1: Original Solve ThreadCommRunKer 57 1.0 6.4156e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 0 MatMult 10 1.0 1.4055e-01 1.4 1.80e+06 1.0 2.0e+01 3.2e+01 0.0e+00 12 16 50 0 0 19 22 83 94 0 26 MatSolve 10 1.0 2.5086e-03 1.0 1.60e+06 1.0 0.0e+00 0.0e+00 0.0e+00 0 14 0 0 0 0 20 0 0 0 1275 MatCholFctrNum 1 1.0 1.6491e-03 1.1 2.00e+04 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 24 MatICCFactorSym 1 1.0 5.0807e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 0 0 0 0 2 0 0 0 0 2 0 MatAssemblyBegin 1 1.0 1.6595e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 2 0 0 0 3 3 0 0 0 4 0 MatAssemblyEnd 1 1.0 1.2010e-01 1.2 0.00e+00 0.0 4.0e+00 1.0e+01 9.0e+00 11 0 10 0 14 17 0 17 6 18 0 MatGetRowIJ 1 1.0 9.5367e-07 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetOrdering 1 1.0 1.5497e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 3 0 0 0 0 4 0 VecMDot 9 1.0 1.1683e-01 1.2 1.80e+06 1.0 0.0e+00 0.0e+00 9.0e+00 10 16 0 0 14 17 22 0 0 18 31 VecNorm 11 1.0 1.3794e-01 1.2 4.40e+05 1.0 0.0e+00 0.0e+00 1.1e+01 13 4 0 0 17 20 5 0 0 22 6 VecScale 10 1.0 2.8610e-04 1.1 2.00e+05 1.0 0.0e+00 0.0e+00 0.0e+00 0 2 0 0 0 0 2 0 0 0 1398 VecCopy 1 1.0 7.0810e-05 1.3 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 13 1.0 4.1294e-04 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 2 1.0 9.0837e-05 1.0 8.00e+04 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 1761 VecMAXPY 10 1.0 1.2455e-03 1.1 2.16e+06 1.0 0.0e+00 0.0e+00 0.0e+00 0 19 0 0 0 0 27 0 0 0 3468 VecAssemblyBegin 1 1.0 4.7200e-02 2.0 0.00e+00 0.0 0.0e+00 0.0e+00 3.0e+00 3 0 0 0 5 6 0 0 0 6 0 VecAssemblyEnd 1 1.0 1.1921e-06 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecScatterBegin 10 1.0 5.8174e-05 1.1 0.00e+00 0.0 2.0e+01 3.2e+01 0.0e+00 0 0 50 0 0 0 0 83 94 0 0 VecScatterEnd 10 1.0 1.3768e-01 1.5 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 11 0 0 0 0 19 0 0 0 0 0 VecNormalize 10 1.0 1.1729e-01 1.0 6.00e+05 1.0 0.0e+00 0.0e+00 1.0e+01 11 5 0 0 15 19 7 0 0 20 10 KSPGMRESOrthog 9 1.0 1.1788e-01 1.2 3.60e+06 1.0 0.0e+00 0.0e+00 9.0e+00 11 32 0 0 14 17 44 0 0 18 61 KSPSetUp 2 1.0 2.4314e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 2 0 0 0 3 4 0 0 0 4 0 KSPSolve 1 1.0 3.3608e-01 1.0 7.84e+06 1.0 1.8e+01 3.2e+01 2.4e+01 33 70 45 0 36 54 97 75 85 48 47 PCSetUp 2 1.0 2.4459e-03 1.1 2.00e+04 1.0 0.0e+00 0.0e+00 7.0e+00 0 0 0 0 11 0 0 0 0 14 16 PCSetUpOnBlocks 1 1.0 2.3570e-03 1.1 2.00e+04 1.0 0.0e+00 0.0e+00 5.0e+00 0 0 0 0 8 0 0 0 0 10 17 PCApply 10 1.0 3.0060e-03 1.0 1.60e+06 1.0 0.0e+00 0.0e+00 0.0e+00 0 14 0 0 0 0 20 0 0 0 1064 --- Event Stage 2: Second Solve ThreadCommRunKer 31 1.0 2.4369e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 1 0 0 0 0 0 MatMult 5 1.0 7.0151e-02 1.4 9.00e+05 1.0 1.0e+01 3.2e+01 0.0e+00 6 8 25 0 0 17 29 62 0 0 26 MatSolve 5 1.0 1.2379e-03 1.0 8.00e+05 1.0 0.0e+00 0.0e+00 0.0e+00 0 7 0 0 0 0 26 0 0 0 1292 MatCholFctrNum 1 1.0 1.3478e-03 1.1 2.00e+04 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 1 0 0 0 30 MatAssemblyBegin 1 1.0 4.1847e-02 2.3 0.00e+00 0.0 6.0e+00 2.4e+05 2.0e+00 3 0 15100 3 9 0 38100 13 0 MatAssemblyEnd 1 1.0 1.4336e-01 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 13 0 0 0 2 38 0 0 0 7 0 MatGetRowIJ 1 1.0 9.5367e-07 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetOrdering 1 1.0 1.3590e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 3 0 0 0 0 13 0 MatZeroEntries 1 1.0 1.7595e-04 1.3 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecMDot 4 1.0 4.6986e-02 1.0 4.00e+05 1.0 0.0e+00 0.0e+00 4.0e+00 5 4 0 0 6 13 13 0 0 27 17 VecNorm 6 1.0 7.0805e-02 1.0 2.40e+05 1.0 0.0e+00 0.0e+00 6.0e+00 7 2 0 0 9 20 8 0 0 40 7 VecScale 5 1.0 1.3566e-04 1.1 1.00e+05 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 3 0 0 0 1474 VecCopy 1 1.0 5.2929e-05 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 7 1.0 2.4366e-04 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 2 1.0 8.0824e-05 1.1 8.00e+04 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 3 0 0 0 1980 VecMAXPY 5 1.0 3.0398e-04 1.0 5.60e+05 1.0 0.0e+00 0.0e+00 0.0e+00 0 5 0 0 0 0 18 0 0 0 3684 VecScatterBegin 5 1.0 2.0742e-05 1.0 0.00e+00 0.0 1.0e+01 3.2e+01 0.0e+00 0 0 25 0 0 0 0 62 0 0 0 VecScatterEnd 5 1.0 6.8789e-02 1.5 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 6 0 0 0 0 17 0 0 0 0 0 VecNormalize 5 1.0 6.8974e-02 1.5 3.00e+05 1.0 0.0e+00 0.0e+00 5.0e+00 6 3 0 0 8 17 10 0 0 33 9 KSPGMRESOrthog 4 1.0 4.7226e-02 1.0 8.00e+05 1.0 0.0e+00 0.0e+00 4.0e+00 5 7 0 0 6 14 26 0 0 27 34 KSPSetUp 2 1.0 1.7166e-05 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 1 1.0 1.6771e-01 1.2 2.84e+06 1.0 8.0e+00 3.2e+01 1.1e+01 15 25 20 0 17 45 92 50 0 73 34 PCSetUp 2 1.0 1.5001e-03 1.1 2.00e+04 1.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 3 0 1 0 0 13 27 PCSetUpOnBlocks 1 1.0 1.4970e-03 1.1 2.00e+04 1.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 3 0 1 0 0 13 27 PCApply 5 1.0 1.4873e-03 1.0 8.00e+05 1.0 0.0e+00 0.0e+00 0.0e+00 0 7 0 0 0 0 26 0 0 0 1076 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Viewer 1 0 0 0 --- Event Stage 1: Original Solve Matrix 4 0 0 0 Vector 23 2 163088 0 Vector Scatter 1 0 0 0 Index Set 5 3 2336 0 Krylov Solver 2 0 0 0 Preconditioner 2 0 0 0 --- Event Stage 2: Second Solve Matrix 0 4 2990500 0 Vector 0 21 2912264 0 Vector Scatter 0 1 1084 0 Index Set 2 4 163104 0 Krylov Solver 0 2 19472 0 Preconditioner 0 2 1816 0 ======================================================================================================================== Average time to get PetscTime(): 1.19209e-07 Average time for MPI_Barrier(): 0.00960441 Average time for zero size MPI_Send(): 0.00612295 #PETSc Option Table entries: -log_summary -m 10000 -threadcomm_nthreads 1 -threadcomm_type openmp #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure run at: Thu Feb 14 10:14:19 2013 Configure options: --PETSC_ARCH=linux64_openmpi_pthread_openmp --with-clanguage=cxx --with-shared-libraries=0 --with-c2html=0 --with-debugging=0 --CC=gcc --CXX=g++ --with-fortran=0 --with-ar=ar --with-ranlib=ranlib --COPTFLAGS=-O2 --CXXOPTFLAGS=-O2 --with-mpi=1 --with-mpi-dir=/usr/lib/openmpi --with-pthreadclasses --with-threadcomm --with-openmp --with-threadcomm --with-x=0 ----------------------------------------- Libraries compiled on Thu Feb 14 10:14:19 2013 on ubuntu Machine characteristics: Linux-3.2.0-29-generic-x86_64-with-Ubuntu-12.04-precise Using PETSc directory: /home/albuquerque/dev/petsc-dev Using PETSc arch: linux64_openmpi_pthread_openmp ----------------------------------------- Using C compiler: g++ -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O2 -fopenmp ${COPTFLAGS} ${CFLAGS} ----------------------------------------- Using include paths: -I/home/albuquerque/dev/petsc-dev/linux64_openmpi_pthread_openmp/include -I/home/albuquerque/dev/petsc-dev/include -I/home/albuquerque/dev/petsc-dev/include -I/home/albuquerque/dev/petsc-dev/linux64_openmpi_pthread_openmp/include -I/usr/lib/openmpi/include ----------------------------------------- Using C linker: g++ Using libraries: -Wl,-rpath,/home/albuquerque/dev/petsc-dev/linux64_openmpi_pthread_openmp/lib -L/home/albuquerque/dev/petsc-dev/linux64_openmpi_pthread_openmp/lib -lpetsc -llapack -lblas -lpthread -Wl,-rpath,/usr/lib/openmpi/lib -L/usr/lib/openmpi/lib -lmpi++ -lmpi -lm -L/usr/lib/gcc/x86_64-linux-gnu/4.6 -L/usr/lib/x86_64-linux-gnu -L/lib/x86_64-linux-gnu -lstdc++ -ldl -lgcc_s -ldl -----------------------------------------