************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- /home/cfd/sgao/mycodes/fensap-merge-sgs/bin/fensapMPI_LINUX64 on a linux named pendragon.cfdlab.mcgill.ca with 4 processors, by sgao Fri Jan 15 11:20:15 2016 Using Petsc Release Version 3.3.0, Patch 6, Mon Feb 11 12:26:34 CST 2013 Max Max/Min Avg Total Time (sec): 2.027e+02 1.00003 2.027e+02 Objects: 4.668e+03 1.00000 4.668e+03 Flops: 1.082e+10 1.04135 1.057e+10 4.226e+10 Flops/sec: 5.336e+07 1.04138 5.212e+07 2.085e+08 MPI Messages: 3.676e+03 1.02826 3.650e+03 1.460e+04 MPI Message Lengths: 1.431e+08 1.53794 3.266e+04 4.768e+08 MPI Reductions: 7.827e+03 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 2.0270e+02 100.0% 4.2260e+10 100.0% 1.460e+04 100.0% 3.266e+04 100.0% 7.826e+03 100.0% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage compute_rhs 101 1.0 8.7983e+01 1.0 0.00e+00 0.0 7.0e+03 5.7e+04 2.2e+03 43 0 48 83 28 43 0 48 83 28 0 SURFINT 101 1.0 4.7722e+01 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 22 0 0 0 0 22 0 0 0 0 0 VOLINT 101 1.0 3.4820e+01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 17 0 0 0 0 17 0 0 0 0 0 Before_surf 101 1.0 5.1162e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 3 0 0 0 0 3 0 0 0 0 0 after_surf 101 1.0 9.2629e+00 9.7 0.00e+00 0.0 7.0e+03 5.7e+04 2.2e+03 2 0 48 83 28 2 0 48 83 28 0 KSPGMRESOrthog 636 1.0 1.6209e+00 1.1 1.19e+09 1.0 0.0e+00 0.0e+00 6.4e+02 1 11 0 0 8 1 11 0 0 8 2887 KSPSetUp 101 1.0 3.3832e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 101 1.0 9.9190e+01 1.0 1.08e+10 1.0 7.6e+03 1.1e+04 3.3e+03 49100 52 17 42 49100 52 17 42 426 MatMult 636 1.0 9.0361e+00 1.0 9.21e+09 1.0 7.6e+03 1.1e+04 0.0e+00 4 85 52 17 0 4 85 52 17 0 3980 MatAssemblyBegin 101 1.0 6.3162e-0211.9 0.00e+00 0.0 1.8e+03 1.5e+05 3.0e+02 0 0 12 55 4 0 0 12 55 4 0 MatAssemblyEnd 101 1.0 1.5324e-01 1.0 0.00e+00 0.0 2.4e+03 8.8e+02 8.1e+02 0 0 17 0 10 0 0 17 0 10 0 MatView 101 1.0 8.8043e-03 1.8 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+02 0 0 0 0 1 0 0 0 0 1 0 VecMax 505 1.0 4.9066e-02 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 5.0e+02 0 0 0 0 6 0 0 0 0 6 0 VecMDot 636 1.0 8.3194e-01 1.1 5.97e+08 1.0 0.0e+00 0.0e+00 6.4e+02 0 6 0 0 8 0 6 0 0 8 2812 VecNorm 737 1.0 1.1989e-01 1.2 1.65e+08 1.0 0.0e+00 0.0e+00 7.4e+02 0 2 0 0 9 0 2 0 0 9 5397 VecScale 737 1.0 2.8933e-02 1.0 8.25e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 11182 VecCopy 606 1.0 1.2639e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 505 1.0 7.8906e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 101 1.0 4.3704e-02 1.1 2.26e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 2029 VecMAXPY 737 1.0 1.0096e+00 1.0 7.39e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 7 0 0 0 0 7 0 0 0 2870 VecAssemblyBegin 838 1.0 1.1236e-01 2.5 0.00e+00 0.0 2.4e+03 5.4e+04 2.5e+03 0 0 17 27 32 0 0 17 27 32 0 VecAssemblyEnd 838 1.0 1.7009e-02 1.3 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecScatterBegin 636 1.0 3.2463e-02 1.5 0.00e+00 0.0 7.6e+03 1.1e+04 0.0e+00 0 0 52 17 0 0 0 52 17 0 0 VecScatterEnd 636 1.0 1.3705e-01 3.5 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 PCSetUp 101 1.0 6.0081e-05 1.6 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 PCApply 636 1.0 8.7006e+01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.9e+03 43 0 0 0 24 43 0 0 0 24 0 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Krylov Solver 101 101 1923040 0 Matrix 303 303 5901812992 0 Vector 3859 3859 3140059632 0 Vector Scatter 101 101 107060 0 Index Set 202 202 612060 0 Preconditioner 101 101 86456 0 Viewer 1 0 0 0 ======================================================================================================================== Average time to get PetscTime(): 9.53674e-08 Average time for MPI_Barrier(): 1.19209e-06 Average time for zero size MPI_Send(): 2.26498e-06 #PETSc Option Table entries: -ksp_monitor -ksp_type fgmres -ksp_view -log_summary #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure run at: Thu Mar 7 10:32:33 2013 Configure options: --with-mpi-dir=/usr/local.linux64/lib64/MPI-openmpi-1.4.5/ --with-shared-libraries=0 --COPTFLAGS=-O3 --FOPTFLAGS=-O3 --with-debugging=no ----------------------------------------- Libraries compiled on Thu Mar 7 10:32:33 2013 on kraken.cfdlab.mcgill.ca Machine characteristics: Linux-2.6.9-42.ELsmp-x86_64-with-redhat-4.4-Final Using PETSc directory: /home/nti/plagace/WORK/PETSC/petsc-3.3-p6 Using PETSc arch: linux ----------------------------------------- Using C compiler: /usr/local.linux64/lib64/MPI-openmpi-1.4.5/bin/mpicc -wd1572 -O3 ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: /usr/local.linux64/lib64/MPI-openmpi-1.4.5/bin/mpif90 -O3 ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/home/nti/plagace/WORK/PETSC/petsc-3.3-p6/linux/include -I/home/nti/plagace/WORK/PETSC/petsc-3.3-p6/include -I/home/nti/plagace/WORK/PETSC/petsc-3.3-p6/include -I/home/nti/plagace/WORK/PETSC/petsc-3.3-p6/linux/include -I/usr/X11R6/include -I/usr/local.linux64/lib64/MPI-openmpi-1.4.5/include ----------------------------------------- Using C linker: /usr/local.linux64/lib64/MPI-openmpi-1.4.5/bin/mpicc Using Fortran linker: /usr/local.linux64/lib64/MPI-openmpi-1.4.5/bin/mpif90 Using libraries: -Wl,-rpath,/home/nti/plagace/WORK/PETSC/petsc-3.3-p6/linux/lib -L/home/nti/plagace/WORK/PETSC/petsc-3.3-p6/linux/lib -lpetsc -L/usr/X11R6/lib64 -lX11 -lpthread -llapack -lblas -L/usr/local.linux64/lib64/MPI-openmpi-1.4.5/lib -L/opt/intel/cce/10.1.018/lib -L/usr/lib/gcc/x86_64-redhat-linux/3.4.6 -lmpi_f90 -lmpi_f77 -L/opt/intel/fce/10.1.018/lib -lifport -lifcoremt -lm -lm -ldl -lmpi -lopen-rte -lopen-pal -lnsl -lutil -limf -lsvml -lipgo -lirc -lgcc_s -lpthread -lirc_s -ldl -----------------------------------------