************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ../../src/cooper on a linux-gnu named animal2 with 2 processors, by fateneja Mon May 12 14:46:22 2008 Using Petsc Release Version 2.3.3, Patch 8, Fri Nov 16 17:03:40 CST 2007 HG revision: 414581156e67e55c761739b0deb119f7590d0f4b Max Max/Min Avg Total Time (sec): 3.769e+01 1.00002 3.769e+01 Objects: 3.862e+03 1.00000 3.862e+03 Flops: 2.003e+06 1.09091 1.920e+06 3.839e+06 Flops/sec: 5.314e+04 1.09088 5.093e+04 1.019e+05 Memory: 1.811e+07 1.07228 3.499e+07 MPI Messages: 4.649e+03 1.00000 4.649e+03 9.298e+03 MPI Message Lengths: 3.734e+06 1.00000 8.032e+02 7.468e+06 MPI Reductions: 2.460e+03 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 3.7691e+01 100.0% 3.8392e+06 100.0% 9.298e+03 100.0% 8.032e+02 100.0% 4.919e+03 100.0% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops/sec: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %F - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ ########################################################## # # # WARNING!!! # # # # This code was compiled with a debugging option, # # To get timing results run config/configure.py # # using --with-debugging=no, the performance will # # be generally two or three times faster. # # # ########################################################## ########################################################## # # # WARNING!!! # # # # This code was run without the PreLoadBegin() # # macros. To get timing results we always recommend # # preloading. otherwise timing numbers may be # # meaningless. # ########################################################## Event Count Time (sec) Flops/sec --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %F %M %L %R %T %F %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage VecScale 63 1.0 3.0446e-04 1.1 2.58e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 4 0 0 0 0 4 0 0 0 495 VecCopy 243 1.0 2.3115e-03 1.6 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 91 1.0 3.8838e-04 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 693 1.0 4.2827e-03 1.0 4.13e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 88 0 0 0 0 88 0 0 0 791 VecPointwiseMult 46 1.0 3.8862e-04 1.1 1.57e+08 1.2 0.0e+00 0.0e+00 0.0e+00 0 3 0 0 0 0 3 0 0 0 283 VecScatterBegin 4327 1.0 1.6362e+01 1.0 0.00e+00 0.0 8.7e+03 8.3e+02 0.0e+00 43 0 93 96 0 43 0 93 96 0 0 VecScatterEnd 4327 1.0 1.7747e+01 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 45 0 0 0 0 45 0 0 0 0 0 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. --- Event Stage 0: Main Stage Distributed array 322 1 6824 0 Index Set 1288 1288 3791872 0 Vec 1286 1286 10817568 0 Vec Scatter 644 2 0 0 IS L to G Mapping 322 1 5728 0 ======================================================================================================================== Average time to get PetscTime(): 9.53674e-08 Average time for MPI_Barrier(): 7.82013e-06 Average time for zero size MPI_Send(): 6.91414e-06 OptionTable: -c 0 OptionTable: -g cylinder-grid.h5 OptionTable: -i input.xml OptionTable: -log_summary OptionTable: -s ../cooper.rng Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 Configure run at: Wed Feb 27 15:16:16 2008 Configure options: --prefix=/usr/local/petsc --with-blas-lib=/usr/local/lib/libfblas.a --with-lapack-lib=/usr/local/lib/libflapack.a --with-mpi-dir=/usr/local/mpich --with-shared=0 ----------------------------------------- Libraries compiled on Wed Feb 27 15:18:18 CST 2008 on animal2 Machine characteristics: Linux animal2 2.6.22-14-generic #1 SMP Tue Feb 12 02:46:46 UTC 2008 x86_64 GNU/Linux Using PETSc directory: /home/fateneja/source/petsc-2.3.3-p8 Using PETSc arch: linux-gnu-c-debug ----------------------------------------- Using C compiler: /usr/local/mpich/bin/mpicc -fPIC -Wall -Wwrite-strings -Wno-strict-aliasing -g3 Using Fortran compiler: /usr/local/mpich/bin/mpif90 -I. -fPIC -Wall -Wno-unused-variable -g ----------------------------------------- Using include paths: -I/home/fateneja/source/petsc-2.3.3-p8 -I/home/fateneja/source/petsc-2.3.3-p8/bmake/linux-gnu-c-debug -I/home/fateneja/source/petsc-2.3.3-p8/include -I/usr/local/mpich/include -I. -I/usr/local/mpich/include -I/usr/local/mpich/include ------------------------------------------ Using C linker: /usr/local/mpich/bin/mpicc -fPIC -Wall -Wwrite-strings -Wno-strict-aliasing -g3 Using Fortran linker: /usr/local/mpich/bin/mpif90 -I. -fPIC -Wall -Wno-unused-variable -g Using libraries: -Wl,-rpath,/home/fateneja/source/petsc-2.3.3-p8/lib/linux-gnu-c-debug -L/home/fateneja/source/petsc-2.3.3-p8/lib/linux-gnu-c-debug -lpetscts -lpetscsnes -lpetscksp -lpetscdm -lpetscmat -lpetscvec -lpetsc -Wl,-rpath,/usr/local/lib -L/usr/local/lib -lflapack -Wl,-rpath,/usr/local/lib -L/usr/local/lib -lfblas -Wl,-rpath,/usr/local/mpich/lib -L/usr/local/mpich/lib -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.1.3 -L/usr/lib/gcc/x86_64-linux-gnu/4.1.3 -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.1.3/../../../../lib -L/usr/lib/gcc/x86_64-linux-gnu/4.1.3/../../../../lib -Wl,-rpath,/lib/../lib -L/lib/../lib -Wl,-rpath,/usr/lib/../lib -L/usr/lib/../lib -ldl -lmpich -lpthread -lrt -lgcc_s -lmpichf90 -lgfortranbegin -lgfortran -lm -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.2.1 -L/usr/lib/gcc/x86_64-linux-gnu/4.2.1 -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.2.1/../../../../lib -L/usr/lib/gcc/x86_64-linux-gnu/4.2.1/../../../../lib -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.2.1/../../.. -L/usr/lib/gcc/x86_64-linux-gnu/4.2.1/../../.. -Wl,-rpath,/usr/local/mpich/lib -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.1.3 -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.1.3/../../../../lib -Wl,-rpath,/lib/../lib -Wl,-rpath,/usr/lib/../lib -lm -Wl,-rpath,/usr/local/mpich/lib -L/usr/local/mpich/lib -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.1.3 -L/usr/lib/gcc/x86_64-linux-gnu/4.1.3 -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.1.3/../../../../lib -L/usr/lib/gcc/x86_64-linux-gnu/4.1.3/../../../../lib -Wl,-rpath,/lib/../lib -L/lib/../lib -Wl,-rpath,/usr/lib/../lib -L/usr/lib/../lib -ldl -lmpich -lpthread -lrt -lgcc_s -lmpichf90 -lgfortranbegin -lgfortran -lm -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.2.1 -L/usr/lib/gcc/x86_64-linux-gnu/4.2.1 -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.2.1/../../../../lib -L/usr/lib/gcc/x86_64-linux-gnu/4.2.1/../../../../lib -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.2.1/../../.. -L/usr/lib/gcc/x86_64-linux-gnu/4.2.1/../../.. -Wl,-rpath,/usr/local/mpich/lib -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.1.3 -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.1.3/../../../../lib -Wl,-rpath,/lib/../lib -Wl,-rpath,/usr/lib/../lib -lm -lm -Wl,-rpath,/usr/local/mpich/lib -L/usr/local/mpich/lib -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.1.3 -L/usr/lib/gcc/x86_64-linux-gnu/4.1.3 -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.1.3/../../../../lib -L/usr/lib/gcc/x86_64-linux-gnu/4.1.3/../../../../lib -Wl,-rpath,/lib/../lib -L/lib/../lib -Wl,-rpath,/usr/lib/../lib -L/usr/lib/../lib -ldl -lmpich -lpthread -lrt -lgcc_s -lmpichf90 -lgfortranbegin -lgfortran -lm -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.2.1 -L/usr/lib/gcc/x86_64-linux-gnu/4.2.1 -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.2.1/../../../../lib -L/usr/lib/gcc/x86_64-linux-gnu/4.2.1/../../../../lib -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.2.1/../../.. -L/usr/lib/gcc/x86_64-linux-gnu/4.2.1/../../.. -Wl,-rpath,/usr/local/mpich/lib -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.1.3 -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.1.3/../../../../lib -Wl,-rpath,/lib/../lib -Wl,-rpath,/usr/lib/../lib -lm -Wl,-rpath,/usr/local/mpich/lib -L/usr/local/mpich/lib -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.1.3 -L/usr/lib/gcc/x86_64-linux-gnu/4.1.3 -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.1.3/../../../../lib -L/usr/lib/gcc/x86_64-linux-gnu/4.1.3/../../../../lib -Wl,-rpath,/lib/../lib -L/lib/../lib -Wl,-rpath,/usr/lib/../lib -L/usr/lib/../lib -ldl -lmpich -lpthread -lrt -lgcc_s -ldl ------------------------------------------