************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./fdtd_petsc on a linux-gnu named neptun with 1 processor, by s5152891 Tue Oct 7 17:17:18 2008 Using Petsc Release Version 2.3.2, Patch 8, Tue Jan 2 14:33:59 PST 2007 HG revision: ebeddcedcc065e32fc252af32cf1d01ed4fc7a80 Max Max/Min Avg Total Time (sec): 1.381e+02 1.00000 1.381e+02 Objects: 8.049e+03 1.00000 8.049e+03 Flops: 5.416e+09 1.00000 5.416e+09 5.416e+09 Flops/sec: 3.922e+07 1.00000 3.922e+07 3.922e+07 Memory: 1.888e+06 1.00000 1.888e+06 MPI Messages: 0.000e+00 0.00000 0.000e+00 0.000e+00 MPI Message Lengths: 0.000e+00 0.00000 0.000e+00 0.000e+00 MPI Reductions: 6.708e+04 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 1.3810e+02 100.0% 5.4160e+09 100.0% 0.000e+00 0.0% 0.000e+00 0.0% 6.708e+04 100.0% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops/sec: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %F - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ ########################################################## # # # WARNING!!! # # # # This code was compiled with a debugging option, # # To get timing results run config/configure.py # # using --with-debugging=no, the performance will # # be generally two or three times faster. # # # ########################################################## ########################################################## # # # WARNING!!! # # # # This code was run without the PreLoadBegin() # # macros. To get timing results we always recommend # # preloading. otherwise timing numbers may be # # meaningless. # ########################################################## Event Count Time (sec) Flops/sec --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %F %M %L %R %T %F %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage MatMult 27000 1.0 2.6440e+01 1.0 2.26e+07 1.0 0.0e+00 0.0e+00 0.0e+00 19 11 0 0 0 19 11 0 0 0 23 MatSolve 29000 1.0 2.7497e+01 1.0 2.33e+07 1.0 0.0e+00 0.0e+00 0.0e+00 20 12 0 0 0 20 12 0 0 0 23 MatLUFactorNum 2000 1.0 1.1127e+01 1.0 3.94e+06 1.0 0.0e+00 0.0e+00 0.0e+00 8 1 0 0 0 8 1 0 0 0 4 MatILUFactorSym 2000 1.0 2.1341e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 4.0e+03 2 0 0 0 6 2 0 0 0 6 0 MatAssemblyBegin 2 1.0 3.7193e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyEnd 2 1.0 6.4430e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetOrdering 2000 1.0 1.1456e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 4.0e+03 1 0 0 0 6 1 0 0 0 6 0 VecMDot 27000 1.0 2.9711e+01 1.0 6.36e+07 1.0 0.0e+00 0.0e+00 2.7e+04 22 35 0 0 40 22 35 0 0 40 64 VecNorm 29000 1.0 4.9954e+00 1.0 2.90e+07 1.0 0.0e+00 0.0e+00 2.9e+04 4 3 0 0 43 4 3 0 0 43 29 VecScale 28000 1.0 2.7392e+00 1.0 2.56e+07 1.0 0.0e+00 0.0e+00 0.0e+00 2 1 0 0 0 2 1 0 0 0 26 VecCopy 2000 1.0 1.0494e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 3001 1.0 1.2301e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 1000 1.0 1.4137e-01 1.0 3.54e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 35 VecMAXPY 28000 1.0 3.0303e+01 1.0 6.68e+07 1.0 0.0e+00 0.0e+00 0.0e+00 22 37 0 0 0 22 37 0 0 0 67 VecNormalize 29000 1.0 7.9736e+00 1.0 2.70e+07 1.0 0.0e+00 0.0e+00 2.9e+04 6 4 0 0 43 6 4 0 0 43 27 KSPGMRESOrthog 27000 1.0 5.8338e+01 1.0 6.48e+07 1.0 0.0e+00 0.0e+00 2.7e+04 42 70 0 0 40 42 70 0 0 40 65 KSPSetup 2000 1.0 5.6820e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+01 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 2000 1.0 1.3797e+02 1.0 3.93e+07 1.0 0.0e+00 0.0e+00 6.7e+04100100 0 0100 100100 0 0100 39 PCSetUp 2000 1.0 1.4874e+01 1.0 2.95e+06 1.0 0.0e+00 0.0e+00 8.0e+03 11 1 0 0 12 11 1 0 0 12 3 PCApply 29000 1.0 2.7676e+01 1.0 2.32e+07 1.0 0.0e+00 0.0e+00 0.0e+00 20 12 0 0 0 20 12 0 0 0 23 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. --- Event Stage 0: Main Stage Matrix 2002 2002 20008000 0 Index Set 6000 6000 62064000 0 Vec 43 43 872672 0 Krylov Solver 2 2 34432 0 Preconditioner 2 2 336 0 ======================================================================================================================== Average time to get PetscTime(): 1.28746e-06 OptionTable: -ksp_type gmres OptionTable: -log_summary OptionTable: -m 50 OptionTable: -n 50 Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 Configure run at: Thu Feb 8 10:33:41 2007 Configure options: --with-cc=icc --with-fc=ifort --download-f-blas-lapack=1 --download-mpich=0 --with-mpi=1 --with-mpi-include=/usr/include/ --with-mpi-lib=/usr/lib/libmpi.so --with-shared ----------------------------------------- Libraries compiled on Thu Feb 8 10:34:08 CET 2007 on mars Machine characteristics: Linux mars 2.6.16.21-0.25-default #1 SMP Tue Sep 19 07:26:15 UTC 2006 ia64 ia64 ia64 GNU/Linux Using PETSc directory: /work/licsoft/libraries/petsc/petsc-2.3.2-p8/ Using PETSc arch: linux-gnu-c-debug ----------------------------------------- Using C compiler: icc -fPIC -g Using Fortran compiler: ifort -fPIC -g ----------------------------------------- Using include paths: -I/work/licsoft/libraries/petsc/petsc-2.3.2-p8/ -I/work/licsoft/libraries/petsc/petsc-2.3.2-p8//bmake/linux-gnu-c-debug -I/work/licsoft/libraries/petsc/petsc-2.3.2-p8//include -I/usr/X11R6/include -I/usr/include ------------------------------------------ Using C linker: icc -fPIC -g Using Fortran linker: ifort -fPIC -g Using libraries: -Wl,-rpath,/work/licsoft/libraries/petsc/petsc-2.3.2-p8//lib/linux-gnu-c-debug -L/work/licsoft/libraries/petsc/petsc-2.3.2-p8//lib/linux-gnu-c-debug -lpetscts -lpetscsnes -lpetscksp -lpetscdm -lpetscmat -lpetscvec -lpetsc -L/usr/X11R6/lib -lX11 -Wl,-rpath,/usr/lib -L/usr/lib -lmpi -Wl,-rpath,/work/licsoft/libraries/petsc/petsc-2.3.2-p8/externalpackages/fblaslapack/linux-gnu-c-debug -L/work/licsoft/libraries/petsc/petsc-2.3.2-p8/externalpackages/fblaslapack/linux-gnu-c-debug -lflapack -Wl,-rpath,/work/licsoft/libraries/petsc/petsc-2.3.2-p8/externalpackages/fblaslapack/linux-gnu-c-debug -L/work/licsoft/libraries/petsc/petsc-2.3.2-p8/externalpackages/fblaslapack/linux-gnu-c-debug -lfblas -ldl -Wl,-rpath,/opt/intel/cc/10.0.013/lib -L/opt/intel/cc/10.0.013/lib -Wl,-rpath,/usr/lib/gcc/ia64-suse-linux/4.1.0/ -L/usr/lib/gcc/ia64-suse-linux/4.1.0/ -Wl,-rpath,/usr/lib/gcc/ia64-suse-linux/4.1.0/../../../ -L/usr/lib/gcc/ia64-suse-linux/4.1.0/../../../ -limf -lipgo -lintlc -lipr -lgcc_s -lirc -lirc_s -Wl,-rpath,/opt/intel/cc/10.0.013/lib -Wl,-rpath,/usr/lib/gcc/ia64-suse-linux/4.1.0/ -Wl,-rpath,/usr/lib/gcc/ia64-suse-linux/4.1.0/../../../ -Wl,-rpath,/opt/intel/fc/10.0.013/lib -L/opt/intel/fc/10.0.013/lib -lifport -lifcore -lm -lunwind -Wl,-rpath,/opt/intel/cc/10.0.013/lib -Wl,-rpath,/usr/lib/gcc/ia64-suse-linux/4.1.0/ -Wl,-rpath,/usr/lib/gcc/ia64-suse-linux/4.1.0/../../../ -lm -ldl -Wl,-rpath,/opt/intel/cc/10.0.013/lib -L/opt/intel/cc/10.0.013/lib -Wl,-rpath,/usr/lib/gcc/ia64-suse-linux/4.1.0/ -L/usr/lib/gcc/ia64-suse-linux/4.1.0/ -Wl,-rpath,/usr/lib/gcc/ia64-suse-linux/4.1.0/../../../ -L/usr/lib/gcc/ia64-suse-linux/4.1.0/../../../ -limf -lipgo -lintlc -lipr -lgcc_s -lirc -lirc_s -ldl ------------------------------------------