************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- bin/navier-stokes on a arch-linu named slippy.sen.cwi.nl with 2 processors, by sanderse Wed May 30 09:20:30 2012 Using Petsc Release Version 3.2.0, Patch 6, Wed Jan 11 09:28:45 CST 2012 Max Max/Min Avg Total Time (sec): 2.491e+02 1.00000 2.491e+02 Objects: 5.430e+03 1.00000 5.430e+03 Flops: 1.333e+09 1.00000 1.333e+09 2.666e+09 Flops/sec: 5.351e+06 1.00000 5.351e+06 1.070e+07 Memory: 1.031e+09 1.00001 2.063e+09 MPI Messages: 1.613e+03 1.00000 1.613e+03 3.226e+03 MPI Message Lengths: 2.030e+07 1.00000 1.259e+04 4.060e+07 MPI Reductions: 1.871e+04 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 2.4910e+02 100.0% 2.6656e+09 100.0% 3.226e+03 100.0% 1.259e+04 100.0% 1.871e+04 100.0% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ ########################################################## # # # WARNING!!! # # # # This code was compiled with a debugging option, # # To get timing results run ./configure # # using --with-debugging=no, the performance will # # be generally two or three times faster. # # # ########################################################## Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage VecView 49 1.0 5.0092e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 2 0 0 0 0 2 0 0 0 0 0 VecDot 198 1.0 2.6501e-02 1.0 8.74e+06 1.0 0.0e+00 0.0e+00 6.6e+01 0 1 0 0 0 0 1 0 0 0 660 VecTDot 534 1.0 2.1990e-01 1.0 6.67e+07 1.0 0.0e+00 0.0e+00 5.3e+02 0 5 0 0 3 0 5 0 0 3 607 VecNorm 393 1.0 2.2518e-01 1.0 4.36e+07 1.0 0.0e+00 0.0e+00 3.9e+02 0 3 0 0 2 0 3 0 0 2 387 VecScale 86 1.0 1.2420e-02 1.0 4.12e+06 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 663 VecCopy 345 1.0 9.3375e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 927 1.0 2.7838e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 939 1.0 2.9736e-01 1.0 1.16e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 9 0 0 0 0 9 0 0 0 783 VecAYPX 497 1.0 1.4447e-01 1.0 4.99e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 4 0 0 0 0 4 0 0 0 690 VecAXPBYCZ 230 1.0 1.0469e-01 1.0 5.64e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 4 0 0 0 0 4 0 0 0 1078 VecWAXPY 199 1.0 7.8604e-02 1.0 1.22e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 311 VecMAXPY 120 1.0 7.9428e-02 1.0 3.68e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 3 0 0 0 0 3 0 0 0 925 VecAssemblyBegin 494 1.0 2.0484e-01 1.4 0.00e+00 0.0 1.6e+01 3.7e+05 8.6e+02 0 0 0 15 5 0 0 0 15 5 0 VecAssemblyEnd 494 1.0 1.2151e-02 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecPointwiseMult 975 1.0 3.2390e-01 1.0 5.58e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 4 0 0 0 0 4 0 0 0 345 VecScatterBegin 2700 1.0 1.1761e-01 1.0 0.00e+00 0.0 2.8e+03 1.2e+04 0.0e+00 0 0 86 82 0 0 0 86 82 0 0 VecScatterEnd 2700 1.0 9.1572e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatMult 1126 1.0 4.1493e+01 1.0 3.72e+08 1.0 1.2e+03 1.4e+04 0.0e+00 17 28 38 42 0 17 28 38 42 0 18 MatMultAdd 1580 1.0 6.0113e+01 1.0 4.96e+08 1.0 1.5e+03 1.1e+04 0.0e+00 24 37 48 41 0 24 37 48 41 0 17 MatConvert 73 1.0 5.8954e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.2e+01 0 0 0 0 0 0 0 0 0 0 0 MatScale 9 1.0 1.7318e-02 1.1 2.00e+06 1.0 2.0e+00 1.0e+04 1.2e+01 0 0 0 0 0 0 0 0 0 0 231 MatAssemblyBegin 1435 1.0 1.5572e-01 1.2 0.00e+00 0.0 6.0e+00 6.7e+03 4.3e+02 0 0 0 0 2 0 0 0 0 2 0 MatAssemblyEnd 1435 1.0 2.8024e+00 1.0 0.00e+00 0.0 2.1e+02 2.0e+03 3.5e+03 1 0 7 1 18 1 0 7 1 18 0 MatGetValues 96 1.0 1.6570e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetRow 9524844 1.0 1.1424e+01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 5 0 0 0 0 5 0 0 0 0 0 MatGetRowIJ 2 1.0 4.0531e-06 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetSubMatrice 96 1.0 8.2534e+00 1.0 0.00e+00 0.0 2.3e+02 3.5e+03 6.7e+02 3 0 7 2 4 3 0 7 2 4 0 MatAXPY 56 1.0 3.6502e+00 1.0 0.00e+00 0.0 2.0e+01 4.0e+03 4.3e+02 1 0 1 0 2 1 0 1 0 2 0 MatTranspose 3 1.0 1.0419e-01 1.0 0.00e+00 0.0 1.0e+01 5.0e+03 9.0e+01 0 0 0 0 0 0 0 0 0 0 0 MatMatMult 357 1.0 1.9565e+01 1.0 1.17e+07 1.0 3.0e+02 3.3e+03 3.7e+03 8 1 9 2 20 8 1 9 2 20 1 MatGetLocalMatCondensed 48 1.0 4.2491e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 3.4e+02 2 0 0 0 2 2 0 0 0 2 0 MatGetBrowsOfAcols 48 1.0 4.1129e+00 1.0 0.00e+00 0.0 2.3e+02 3.5e+03 4.3e+02 2 0 7 2 2 2 0 7 2 2 0 KSPSetup 1 1.0 3.2361e-03 1.5 0.00e+00 0.0 0.0e+00 0.0e+00 3.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 41 1.0 8.1559e+01 1.0 4.54e+08 1.0 6.2e+02 2.0e+04 2.8e+03 33 34 19 30 15 33 34 19 30 15 11 PCSetUp 1 1.0 2.3789e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 6.0e+00 1 0 0 0 0 1 0 0 0 0 0 PCApply 349 1.0 6.8372e+01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 3.5e+02 27 0 0 0 2 27 0 0 0 2 0 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Container 48 12 6576 0 Vector 3027 1779 614018416 0 Vector Scatter 222 58 60088 0 Matrix 1546 632 318195076 0 Matrix Null Space 1 1 580 0 Index Set 576 468 346272 0 Krylov Solver 1 1 1136 0 Preconditioner 1 1 1032 0 Viewer 8 7 4760 0 ======================================================================================================================== Average time to get PetscTime(): 1.28746e-06 Average time for MPI_Barrier(): 5.89848e-05 Average time for zero size MPI_Send(): 3.29018e-05 #PETSc Option Table entries: -log_summary #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 Configure run at: Wed Feb 22 18:04:02 2012 Configure options: --download-mpich=1 --with-shared-libraries --download-f-blas-lapack=1 --with-fc=gfortran --with-cxx=g++ --download-hypre --with-hdf5 --download-hdf5 --with-cc=gcc ----------------------------------------- Libraries compiled on Wed Feb 22 18:04:02 2012 on slippy.mas.cwi.nl Machine characteristics: Linux-3.2.2-1.fc16.x86_64-x86_64-with-fedora-16-Verne Using PETSc directory: /export/scratch1/sanderse/software/petsc-3.2-p6-debug/ Using PETSc arch: arch-linux2-c-opt ----------------------------------------- Using C compiler: /export/scratch1/sanderse/software/petsc-3.2-p6-debug/arch-linux2-c-opt/bin/mpicc -fPIC -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -g3 ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: /export/scratch1/sanderse/software/petsc-3.2-p6-debug/arch-linux2-c-opt/bin/mpif90 -fPIC -Wall -Wno-unused-variable -Wno-unused-dummy-argument -g ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/export/scratch1/sanderse/software/petsc-3.2-p6-debug/arch-linux2-c-opt/include -I/export/scratch1/sanderse/software/petsc-3.2-p6-debug/include -I/export/scratch1/sanderse/software/petsc-3.2-p6-debug/include -I/export/scratch1/sanderse/software/petsc-3.2-p6-debug/arch-linux2-c-opt/include ----------------------------------------- Using C linker: /export/scratch1/sanderse/software/petsc-3.2-p6-debug/arch-linux2-c-opt/bin/mpicc Using Fortran linker: /export/scratch1/sanderse/software/petsc-3.2-p6-debug/arch-linux2-c-opt/bin/mpif90 Using libraries: -Wl,-rpath,/export/scratch1/sanderse/software/petsc-3.2-p6-debug/arch-linux2-c-opt/lib -L/export/scratch1/sanderse/software/petsc-3.2-p6-debug/arch-linux2-c-opt/lib -lpetsc -lX11 -Wl,-rpath,/export/scratch1/sanderse/software/petsc-3.2-p6-debug/arch-linux2-c-opt/lib -L/export/scratch1/sanderse/software/petsc-3.2-p6-debug/arch-linux2-c-opt/lib -lHYPRE -Wl,-rpath,/usr/lib/gcc/x86_64-redhat-linux/4.6.2 -lmpichcxx -lstdc++ -lpthread -lflapack -lfblas -lhdf5_fortran -lhdf5 -lz -lm -L/usr/lib/gcc/x86_64-redhat-linux/4.6.2 -ldl -lmpich -lopa -lmpl -lrt -lpthread -lgcc_s -lmpichf90 -lgfortran -lm -lgfortran -lm -lquadmath -lm -lmpichcxx -lstdc++ -ldl -lmpich -lopa -lmpl -lrt -lpthread -lgcc_s -ldl -----------------------------------------