************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./PB3D on a debug-complex named toon-XPS-L501X with 4 processors, by toon Thu Mar 30 08:55:33 2017 Using Petsc Release Version 3.6.4, Apr, 12, 2016 Max Max/Min Avg Total Time (sec): 3.756e+01 1.00000 3.756e+01 Objects: 6.280e+02 1.00000 6.280e+02 Flops: 7.262e+07 1.01103 7.222e+07 2.889e+08 Flops/sec: 1.933e+06 1.01103 1.923e+06 7.692e+06 Memory: 5.151e+06 1.01262 2.050e+07 MPI Messages: 3.820e+02 1.64655 3.120e+02 1.248e+03 MPI Message Lengths: 1.590e+06 1.81680 3.468e+03 4.329e+06 MPI Reductions: 9.666e+03 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 3.7558e+01 100.0% 2.8889e+08 100.0% 1.248e+03 100.0% 3.468e+03 100.0% 9.665e+03 100.0% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length (bytes) Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %F - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ ########################################################## # # # WARNING!!! # # # # This code was compiled with a debugging option, # # To get timing results run ./configure # # using --with-debugging=no, the performance will # # be generally two or three times faster. # # # ########################################################## ########################################################## # # # WARNING!!! # # # # The code for various complex numbers numerical # # kernels uses C++, which generally is not well # # optimized. For performance that is about 4-5 times # # faster, specify --with-fortran-kernels=1 # # when running ./configure.py. # # # ########################################################## Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %F %M %L %R %T %F %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage MatMult 99 1.0 4.6523e-01 1.2 3.54e+07 1.0 5.9e+02 3.5e+02 0.0e+00 1 49 48 5 0 1 49 48 5 0 301 MatSolve 85 1.0 1.1909e+00 1.0 0.00e+00 0.0 5.2e+02 6.9e+03 8.9e+01 3 0 42 84 1 3 0 42 84 1 0 MatLUFactorSym 1 1.0 4.0681e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 6.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatLUFactorNum 1 1.0 6.1663e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 2 0 0 0 0 2 0 0 0 0 0 MatCopy 1 1.0 3.2616e-04 1.6 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatConvert 2 1.0 6.7891e-02 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 4.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyBegin 7 1.0 4.9762e-02 1.9 0.00e+00 0.0 3.6e+01 9.6e+03 2.1e+01 0 0 3 8 0 0 0 3 8 0 0 MatAssemblyEnd 7 1.0 3.4189e-03 1.0 0.00e+00 0.0 2.4e+01 1.0e+01 2.4e+01 0 0 2 0 0 0 0 2 0 0 0 MatGetRow 1000 1.0 1.4525e-02 1.5 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatZeroEntries 30 1.0 1.5497e-04 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatView 2 1.0 6.4642e-03 1.7 0.00e+00 0.0 7.2e+01 2.1e+03 2.0e+00 0 0 6 3 0 0 0 6 3 0 0 MatAXPY 1 1.0 3.8889e-02 1.0 0.00e+00 0.0 1.2e+01 1.0e+01 2.3e+01 0 0 1 0 0 0 0 1 0 0 0 VecDot 3 1.0 5.1339e-03 7.7 1.20e+04 1.0 0.0e+00 0.0e+00 3.0e+00 0 0 0 0 0 0 0 0 0 0 9 VecNorm 45 1.0 8.7210e-02 6.6 1.80e+05 1.0 0.0e+00 0.0e+00 4.5e+01 0 0 0 0 0 0 0 0 0 0 8 VecCopy 174 1.0 2.3546e-03 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 5 1.0 5.0783e-05 1.3 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 1 1.0 3.5130e-02 6.8 4.00e+03 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecScatterBegin 269 1.0 4.4491e-02 4.7 0.00e+00 0.0 1.1e+03 3.5e+03 8.5e+01 0 0 88 88 1 0 0 88 88 1 0 VecScatterEnd 184 1.0 4.3146e-02 1.5 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecReduceArith 86 1.0 1.2254e-02 2.0 3.44e+05 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 112 VecReduceComm 43 1.0 7.0889e-01 1.3 0.00e+00 0.0 0.0e+00 0.0e+00 4.3e+01 2 0 0 0 0 2 0 0 0 0 0 EPSSetUp 1 1.0 3.1850e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.6e+02 0 0 0 0 2 0 0 0 0 2 0 EPSSolve 1 1.0 3.7208e+01 1.0 7.08e+07 1.0 1.1e+03 3.5e+03 9.5e+03 99 98 87 88 99 99 98 87 88 99 8 STSetUp 1 1.0 1.9503e-04 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 3.0e+00 0 0 0 0 0 0 0 0 0 0 0 PCSetUp 1 1.0 6.5858e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.5e+01 2 0 0 0 0 2 0 0 0 0 0 PCApply 85 1.0 1.1921e+00 1.0 0.00e+00 0.0 5.2e+02 6.9e+03 8.9e+01 3 0 42 84 1 3 0 42 84 1 0 BVCreate 49 1.0 4.0621e-02 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 6.4e+02 0 0 0 0 7 0 0 0 0 7 0 BVCopy 131 1.0 5.5204e-02 5.7 0.00e+00 0.0 0.0e+00 0.0e+00 1.8e+02 0 0 0 0 2 0 0 0 0 2 0 BVMult 497 1.0 1.0523e+01 1.1 2.17e+07 1.0 0.0e+00 0.0e+00 3.5e+02 27 30 0 0 4 27 30 0 0 4 8 BVDot 474 1.0 3.3794e+00 1.1 1.45e+07 1.0 0.0e+00 0.0e+00 4.7e+02 9 20 0 0 5 9 20 0 0 5 17 BVOrthogonalize 137 1.0 1.7250e+01 1.0 1.38e+07 1.0 0.0e+00 0.0e+00 1.6e+03 46 19 0 0 17 46 19 0 0 17 3 BVScale 136 1.0 1.5892e-02 1.3 2.72e+05 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 68 BVNorm 43 1.0 7.6287e-03 3.9 1.72e+05 1.0 0.0e+00 0.0e+00 4.3e+01 0 0 0 0 0 0 0 0 0 0 90 BVSetRandom 5 1.0 8.3723e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+01 0 0 0 0 0 0 0 0 0 0 0 BVMatProject 129 1.0 2.7438e-01 1.1 7.19e+06 1.0 0.0e+00 0.0e+00 2.1e+02 1 10 0 0 2 1 10 0 0 2 105 BVAXPY 43 1.0 3.7920e-03 1.1 1.88e+05 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 198 DSSolve 49 1.0 5.4828e-02 1.6 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 DSVectors 87 1.0 1.5099e-02 1.4 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 DSOther 187 1.0 3.3645e-02 1.5 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Matrix 342 342 4061592 0 Vector 213 212 2827680 0 Vector Scatter 6 6 6096 0 Index Set 9 9 7208 0 EPS Solver 1 1 3016 0 PetscRandom 1 1 640 0 Spectral Transform 1 1 828 0 Viewer 1 0 0 0 Krylov Solver 1 1 1144 0 Preconditioner 1 1 1024 0 Basis Vectors 50 50 185376 0 Region 1 1 648 0 Direct Solver 1 1 35264 0 ======================================================================================================================== Average time to get PetscTime(): 9.53674e-08 Average time for MPI_Barrier(): 1.81198e-06 Average time for zero size MPI_Send(): 1.2517e-06 #PETSc Option Table entries: -eps_monitor -eps_ncv 16 -eps_type gd -eps_view -ksp_converged_reason -ksp_monitor_true_residual -ksp_view -log_view -st_pc_factor_mat_solver_package mumps -st_pc_type lu #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 16 sizeof(PetscInt) 4 Configure options: PETSC_ARCH=debug-complex --with-scalar-type=complex --with-debugging --download-metis --download-mumps --download-parmetis --download-scalapack --with-valgrind-dir=/usr --with-hdf5-dir=/opt/hdf5-1.8.16/hdf5/ ----------------------------------------- Libraries compiled on Thu Apr 21 21:40:55 2016 on toon-XPS-L501X Machine characteristics: Linux-4.4.0-21-generic-x86_64-with-Ubuntu-16.04-xenial Using PETSc directory: /opt/petsc-3.6.4 Using PETSc arch: debug-complex ----------------------------------------- Using C compiler: mpicc -fPIC -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -g3 -O0 ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: mpif90 -fPIC -Wall -Wno-unused-variable -ffree-line-length-0 -Wno-unused-dummy-argument -g -O0 ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/opt/petsc-3.6.4/debug-complex/include -I/opt/petsc-3.6.4/include -I/opt/petsc-3.6.4/include -I/opt/petsc-3.6.4/debug-complex/include -I/opt/hdf5-1.8.16/hdf5/include -I/usr/include -I/usr/lib/openmpi/include/openmpi/opal/mca/event/libevent2021/libevent -I/usr/lib/openmpi/include/openmpi/opal/mca/event/libevent2021/libevent/include -I/usr/lib/openmpi/include -I/usr/lib/openmpi/include/openmpi ----------------------------------------- Using C linker: mpicc Using Fortran linker: mpif90 Using libraries: -Wl,-rpath,/opt/petsc-3.6.4/debug-complex/lib -L/opt/petsc-3.6.4/debug-complex/lib -lpetsc -Wl,-rpath,/opt/petsc-3.6.4/debug-complex/lib -L/opt/petsc-3.6.4/debug-complex/lib -lcmumps -ldmumps -lsmumps -lzmumps -lmumps_common -lpord -lscalapack -llapack -lblas -lparmetis -lmetis -lhwloc -Wl,-rpath,/opt/hdf5-1.8.16/hdf5/lib -L/opt/hdf5-1.8.16/hdf5/lib -lhdf5hl_fortran -lhdf5_fortran -lhdf5_hl -lhdf5 -lm -Wl,-rpath,/usr/lib/openmpi/lib -L/usr/lib/openmpi/lib -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/5 -L/usr/lib/gcc/x86_64-linux-gnu/5 -Wl,-rpath,/usr/lib/x86_64-linux-gnu -L/usr/lib/x86_64-linux-gnu -Wl,-rpath,/lib/x86_64-linux-gnu -L/lib/x86_64-linux-gnu -lmpi_usempif08 -lmpi_usempi_ignore_tkr -lmpi_mpifh -lgfortran -lm -lgfortran -lm -lquadmath -lm -lmpi_cxx -lstdc++ -Wl,-rpath,/usr/lib/openmpi/lib -L/usr/lib/openmpi/lib -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/5 -L/usr/lib/gcc/x86_64-linux-gnu/5 -Wl,-rpath,/usr/lib/x86_64-linux-gnu -L/usr/lib/x86_64-linux-gnu -Wl,-rpath,/lib/x86_64-linux-gnu -L/lib/x86_64-linux-gnu -Wl,-rpath,/usr/lib/x86_64-linux-gnu -L/usr/lib/x86_64-linux-gnu -ldl -Wl,-rpath,/usr/lib/openmpi/lib -lmpi -lgcc_s -lpthread -ldl ----------------------------------------- WARNING! There are options you set that were not used! WARNING! could be spelling mistake, etc! Option left: name:-ksp_converged_reason (no value) Option left: name:-ksp_monitor_true_residual (no value) Option left: name:-ksp_view (no value)