************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./fdtd_petsc on a altix4700 named neptun with 1 processor, by s5152891 Tue Oct 7 17:12:29 2008 Using Petsc Release Version 2.3.3, Patch 15, Tue Sep 23 10:02:49 CDT 2008 HG revision: 31306062cd1a6f6a2496fccb4878f485c9b91760 Max Max/Min Avg Total Time (sec): 9.955e+00 1.00000 9.955e+00 Objects: 8.049e+03 1.00000 8.049e+03 Flops: 5.484e+09 1.00000 5.484e+09 5.484e+09 Flops/sec: 5.508e+08 1.00000 5.508e+08 5.508e+08 MPI Messages: 0.000e+00 0.00000 0.000e+00 0.000e+00 MPI Message Lengths: 0.000e+00 0.00000 0.000e+00 0.000e+00 MPI Reductions: 6.200e+04 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 9.9550e+00 100.0% 5.4835e+09 100.0% 0.000e+00 0.0% 0.000e+00 0.0% 6.200e+04 100.0% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops/sec: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %F - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ ########################################################## # # # WARNING!!! # # # # This code was run without the PreLoadBegin() # # macros. To get timing results we always recommend # # preloading. otherwise timing numbers may be # # meaningless. # ########################################################## Event Count Time (sec) Flops/sec --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %F %M %L %R %T %F %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage MatMult 27000 1.0 1.8770e+00 1.0 3.54e+08 1.0 0.0e+00 0.0e+00 0.0e+00 19 12 0 0 0 19 12 0 0 0 354 MatSolve 29000 1.0 4.6273e+00 1.0 1.39e+08 1.0 0.0e+00 0.0e+00 0.0e+00 46 12 0 0 0 46 12 0 0 0 139 MatLUFactorNum 2000 1.0 1.4341e+00 1.0 3.05e+07 1.0 0.0e+00 0.0e+00 0.0e+00 14 1 0 0 0 14 1 0 0 0 31 MatILUFactorSym 2000 1.0 1.7379e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+03 2 0 0 0 3 2 0 0 0 3 0 MatAssemblyBegin 2 1.0 2.8610e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyEnd 2 1.0 5.5408e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetRowIJ 2000 1.0 1.2269e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetOrdering 2000 1.0 6.3031e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 4.0e+03 1 0 0 0 6 1 0 0 0 6 0 VecMDot 27000 1.0 6.1105e-01 1.0 3.09e+09 1.0 0.0e+00 0.0e+00 2.7e+04 6 34 0 0 44 6 34 0 0 44 3092 VecNorm 29000 1.0 2.9923e-01 1.0 4.85e+08 1.0 0.0e+00 0.0e+00 2.9e+04 3 3 0 0 47 3 3 0 0 47 485 VecScale 28000 1.0 5.1373e-02 1.0 1.36e+09 1.0 0.0e+00 0.0e+00 0.0e+00 1 1 0 0 0 1 1 0 0 0 1363 VecCopy 2000 1.0 4.7135e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 3001 1.0 4.8151e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 1000 1.0 2.3828e-03 1.0 2.10e+09 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 2098 VecMAXPY 28000 1.0 4.6929e-01 1.0 4.31e+09 1.0 0.0e+00 0.0e+00 0.0e+00 5 37 0 0 0 5 37 0 0 0 4315 VecNormalize 29000 1.0 4.0173e-01 1.0 5.35e+08 1.0 0.0e+00 0.0e+00 2.9e+04 4 4 0 0 47 4 4 0 0 47 535 KSPGMRESOrthog 27000 1.0 1.1081e+00 1.0 3.41e+09 1.0 0.0e+00 0.0e+00 2.7e+04 11 69 0 0 44 11 69 0 0 44 3411 KSPSetup 2000 1.0 1.2789e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 2000 1.0 9.9182e+00 1.0 5.53e+08 1.0 0.0e+00 0.0e+00 6.2e+04100100 0 0100 100100 0 0100 553 PCSetUp 2000 1.0 1.7180e+00 1.0 2.55e+07 1.0 0.0e+00 0.0e+00 6.0e+03 17 1 0 0 10 17 1 0 0 10 25 PCApply 29000 1.0 4.6638e+00 1.0 1.37e+08 1.0 0.0e+00 0.0e+00 0.0e+00 47 12 0 0 0 47 12 0 0 0 137 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. --- Event Stage 0: Main Stage Matrix 2002 2002 375596008 0 Index Set 6000 6000 62064000 0 Vec 43 43 873016 0 Krylov Solver 2 2 34432 0 Preconditioner 2 2 336 0 ======================================================================================================================== Average time to get PetscTime(): 3.8147e-07 OptionTable: -ksp_type gmres OptionTable: -log_summary OptionTable: -m 50 OptionTable: -n 50 Compiled with FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 Configure run at: Tue Oct 7 16:08:36 2008 Configure options: --prefix=/fastfs/radar/modules/petsc/2.3.3-p15-real --with-scalar-type=real --with-debugging=0 --with-blas-lapack-lib="-L/opt/sgi/scsl/1.6.1.0-sgi403r1/lib -lscs" --with-cc=icc --CFLAGS="-fPIC -mtune=itanium2-p9000 -O3 -ftz -unroll" --with-cxx=icpc --CXXFLAGS="-fPIC -mtune=itanium2-p9000 -O3 -ftz" --with-fc=ifort --FFLAGS="-fPIC -O3 -ftz -mtune=itanium2-p9000 -save-cray-ptr -unroll" --with-vendor-compilers=intel --with-fortran-kernels=generic --with-fortran --download-parmetis=1 --download-mpich=0 --with-mpi-compilers=0 --with-mpi-dir=/usr/ --LIBS=-limf --with-gnu-compilers=0 --with-ar=/usr/bin/ar --with-shared=0 ----------------------------------------- Libraries compiled on Tue Oct 7 16:11:20 CEST 2008 on neptun Machine characteristics: Linux neptun 2.6.16.54-0.2.5.PTF.352893.0-default #1 SMP Mon Jan 21 13:29:51 UTC 2008 ia64 ia64 ia64 GNU/Linux Using PETSc directory: /tmp/petsc/petsc-2.3.3-p15 Using PETSc arch: altix4700-itanium-linux-intel-opt ----------------------------------------- Using C compiler: icc -fPIC -mtune=itanium2-p9000 -O3 -ftz -unroll -fPIC Using Fortran compiler: ifort -fPIC -O3 -ftz -mtune=itanium2-p9000 -save-cray-ptr -unroll ----------------------------------------- Using include paths: -I/tmp/petsc/petsc-2.3.3-p15 -I/tmp/petsc/petsc-2.3.3-p15/bmake/altix4700-itanium-linux-intel-opt -I/tmp/petsc/petsc-2.3.3-p15/include -I/usr/X11R6/include -I/tmp/petsc/petsc-2.3.3-p15/externalpackages/ParMetis-dev-p1/altix4700-itanium-linux-intel-opt/include -I/usr/include ------------------------------------------ Using C linker: icc -fPIC -mtune=itanium2-p9000 -O3 -ftz -unroll -fPIC Using Fortran linker: ifort -fPIC -O3 -ftz -mtune=itanium2-p9000 -save-cray-ptr -unroll Using libraries: -Wl,-rpath,/tmp/petsc/petsc-2.3.3-p15/lib/altix4700-itanium-linux-intel-opt -L/tmp/petsc/petsc-2.3.3-p15/lib/altix4700-itanium-linux-intel-opt -lpetscts -lpetscsnes -lpetscksp -lpetscdm -lpetscmat -lpetscvec -lpetsc -L/usr/X11R6/lib -lX11 -Wl,-rpath,/tmp/petsc/petsc-2.3.3-p15/externalpackages/ParMetis-dev-p1/altix4700-itanium-linux-intel-opt/lib -L/tmp/petsc/petsc-2.3.3-p15/externalpackages/ParMetis-dev-p1/altix4700-itanium-linux-intel-opt/lib -lparmetis -lmetis -Wl,-rpath,/usr/lib -L/usr/lib -lmpi -L/opt/sgi/scsl/1.6.1.0-sgi403r1/lib -lscs -ldl -limf -L/opt/sgi/scsl/1.6.1.0-sgi403r1/lib -L/opt/intel/cc/10.0.026/lib -L/usr/lib/gcc/ia64-suse-linux/4.1.2/ -L/usr/lib/gcc/ia64-suse-linux/4.1.2/../../../ -lipgo -lirc -lipr -lgcc_s -lirc_s -Wl,-rpath,/opt/sgi/scsl/1.6.1.0-sgi403r1/lib -Wl,-rpath,/opt/intel/cc/10.0.026/lib -Wl,-rpath,/usr/lib/gcc/ia64-suse-linux/4.1.2/ -Wl,-rpath,/usr/lib/gcc/ia64-suse-linux/4.1.2/../../../ -L/opt/intel/fc/10.0.026/lib -lifport -lifcore -lm -Wl,-rpath,/opt/sgi/scsl/1.6.1.0-sgi403r1/lib -Wl,-rpath,/opt/intel/cc/10.0.026/lib -Wl,-rpath,/usr/lib/gcc/ia64-suse-linux/4.1.2/ -Wl,-rpath,/usr/lib/gcc/ia64-suse-linux/4.1.2/../../../ -lm -ldl -limf -Wl,-rpath,/opt/sgi/scsl/1.6.1.0-sgi403r1/lib -L/opt/sgi/scsl/1.6.1.0-sgi403r1/lib -Wl,-rpath,/opt/intel/cc/10.0.026/lib -L/opt/intel/cc/10.0.026/lib -Wl,-rpath,/usr/lib/gcc/ia64-suse-linux/4.1.2/ -L/usr/lib/gcc/ia64-suse-linux/4.1.2/ -Wl,-rpath,/usr/lib/gcc/ia64-suse-linux/4.1.2/../../../ -L/usr/lib/gcc/ia64-suse-linux/4.1.2/../../../ -lipgo -lirc -lipr -lgcc_s -lirc_s -ldl -limf ------------------------------------------