TIME ELAPSED DURING ASSEMBLY (ASSEMBLY LOCAL + SETVALUES GLOBAL + MATASSEMBLY): 0.133433 KSP Object: 16 MPI processes type: gmres restart=30, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement happy breakdown tolerance 1e-30 maximum iterations=10000, initial guess is zero tolerances: relative=1e-07, absolute=1e-50, divergence=10000. left preconditioning using PRECONDITIONED norm type for convergence test PC Object: 16 MPI processes type: jacobi linear system matrix = precond matrix: Mat Object: 16 MPI processes type: mpiaij rows=159999, cols=159999 total: nonzeros=1435195, allocated nonzeros=2079987 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines TIME ELAPSED DURING SOLVING: 9.04757 KSP Object: 16 MPI processes type: gmres restart=30, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement happy breakdown tolerance 1e-30 maximum iterations=10000, initial guess is zero tolerances: relative=1e-07, absolute=1e-50, divergence=10000. left preconditioning using PRECONDITIONED norm type for convergence test PC Object: 16 MPI processes type: jacobi linear system matrix = precond matrix: Mat Object: 16 MPI processes type: mpiaij rows=159999, cols=159999 total: nonzeros=1435195, allocated nonzeros=2079987 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines Norm of error 30544.5, Iterations 9660 MEMORY USAGE FOR THIS RUN: 2.27615e+07 INFO: MAX OVER RANKS block_size 1., nz_allocated 0, nz_used 0, nz_unneeded 0, memory 130000., assemblies 0, mallocs 2, INFO: SUM OVER RANKS block_size 1., nz_allocated 0, nz_used 0, nz_unneeded 0, memory 2.07999e+06, assemblies 0, mallocs 2, TIME ELAPSED DURING ENTIRE EXECUTION: 9.29877 MEMORY USAGE FOR THIS RUN: 2.29949e+07 ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./CANE on a named iron with 16 processors, by iron Wed Dec 19 17:54:44 2018 Using Petsc Release Version 3.10.2, unknown Max Max/Min Avg Total Time (sec): 9.300e+00 1.000 9.300e+00 Objects: 5.200e+01 1.000 5.200e+01 Flop: 8.288e+09 1.003 8.285e+09 1.326e+11 Flop/sec: 8.913e+08 1.003 8.909e+08 1.425e+10 MPI Messages: 1.999e+04 2.000 1.874e+04 2.998e+05 MPI Message Lengths: 6.506e+07 1.965 3.259e+03 9.770e+08 MPI Reductions: 1.968e+04 1.000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flop and VecAXPY() for complex vectors of length N --> 8N flop Summary of Stages: ----- Time ------ ----- Flop ------ --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total Count %Total Avg %Total Count %Total 0: Main Stage: 9.2997e+00 100.0% 1.3256e+11 100.0% 2.998e+05 100.0% 3.259e+03 100.0% 1.967e+04 100.0% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flop: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent AvgLen: average message length (bytes) Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %F - percent flop in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flop over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flop --- Global --- --- Stage ---- Total Max Ratio Max Ratio Max Ratio Mess AvgLen Reduct %T %F %M %L %R %T %F %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage BuildTwoSided 2 1.0 2.2846e-0256.3 0.00e+00 0.0 1.3e+02 8.0e+00 2.0e+00 0 0 0 0 0 0 0 0 0 0 0 BuildTwoSidedF 2 1.0 2.2932e-0246.7 0.00e+00 0.0 3.2e+02 6.0e+04 2.0e+00 0 0 0 2 0 0 0 0 2 0 0 MatMult 9981 1.0 2.8194e+00 1.5 1.69e+09 1.0 3.0e+05 3.2e+03 0.0e+00 25 20100 98 0 25 20100 98 0 9595 MatAssemblyBegin 1 1.0 4.3322e-02 1.4 0.00e+00 0.0 1.3e+02 1.4e+05 1.0e+00 0 0 0 2 0 0 0 0 2 0 0 MatAssemblyEnd 1 1.0 3.9880e-02 1.0 0.00e+00 0.0 6.0e+01 8.0e+02 8.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatView 2 1.0 1.0889e-03 2.2 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecMDot 9660 1.0 3.7946e+00 1.4 2.99e+09 1.0 0.0e+00 0.0e+00 9.7e+03 36 36 0 0 49 36 36 0 0 49 12626 VecNorm 9983 1.0 1.2429e+00 2.9 2.00e+08 1.0 0.0e+00 0.0e+00 1.0e+04 10 2 0 0 51 10 2 0 0 51 2570 VecScale 9982 1.0 1.1333e-01 1.6 9.98e+07 1.0 0.0e+00 0.0e+00 0.0e+00 1 1 0 0 0 1 1 0 0 0 14092 VecCopy 322 1.0 1.2512e-02 2.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 326 1.0 1.1598e-02 2.4 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 643 1.0 1.5949e-02 1.9 1.29e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 12901 VecMAXPY 9982 1.0 2.3951e+00 1.4 3.19e+09 1.0 0.0e+00 0.0e+00 0.0e+00 21 38 0 0 0 21 38 0 0 0 21295 VecAssemblyBegin 1 1.0 3.2640e-03 1.0 0.00e+00 0.0 1.9e+02 7.8e+03 1.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAssemblyEnd 1 1.0 3.6407e-04 2.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecPointwiseMult 9982 1.0 4.5607e-01 2.2 9.98e+07 1.0 0.0e+00 0.0e+00 0.0e+00 3 1 0 0 0 3 1 0 0 0 3502 VecScatterBegin 9982 1.0 1.1842e-01 2.1 0.00e+00 0.0 3.0e+05 3.2e+03 1.0e+00 1 0100 98 0 1 0100 98 0 0 VecScatterEnd 9981 1.0 1.3057e-01 2.5 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 0 VecNormalize 9982 1.0 1.3233e+00 2.3 2.99e+08 1.0 0.0e+00 0.0e+00 1.0e+04 11 4 0 0 51 11 4 0 0 51 3621 KSPSetUp 1 1.0 1.8692e-04 3.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 1 1.0 9.0469e+00 1.0 8.29e+09 1.0 3.0e+05 3.2e+03 2.0e+04 97100100 98100 97100100 98100 14653 KSPGMRESOrthog 9660 1.0 5.4111e+00 1.1 5.99e+09 1.0 0.0e+00 0.0e+00 9.7e+03 56 72 0 0 49 56 72 0 0 49 17709 PCSetUp 1 1.0 2.1458e-06 0.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 PCApply 9982 1.0 4.6511e-01 2.2 9.98e+07 1.0 0.0e+00 0.0e+00 2.0e+00 3 1 0 0 0 3 1 0 0 0 3434 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Matrix 3 3 1971068 0. Vector 40 40 4309512 0. Index Set 3 3 4000 0. Vec Scatter 2 2 1920 0. Krylov Solver 1 1 18632 0. Preconditioner 1 1 864 0. Viewer 2 1 848 0. ======================================================================================================================== Average time to get PetscTime(): 9.53674e-08 Average time for MPI_Barrier(): 1.2207e-05 Average time for zero size MPI_Send(): 1.07586e-05 #PETSc Option Table entries: -ksp_view -log_view #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure options: --prefix=/home/iron/valerio/petsc-build-3.10.2 --with-debugging=no ----------------------------------------- Libraries compiled on 2018-12-10 11:21:57 on iron Machine characteristics: Linux-3.13.0-32-generic-x86_64-with-Ubuntu-14.04-trusty Using PETSc directory: /home/iron/valerio/petsc-build-3.10.2 Using PETSc arch: ----------------------------------------- Using C compiler: mpicc -fPIC -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -fstack-protector -fvisibility=hidden -g -O Using Fortran compiler: mpif90 -fPIC -Wall -ffree-line-length-0 -Wno-unused-dummy-argument -g -O ----------------------------------------- Using include paths: -I/home/iron/valerio/petsc-build-3.10.2/include ----------------------------------------- Using C linker: mpicc Using Fortran linker: mpif90 Using libraries: -Wl,-rpath,/home/iron/valerio/petsc-build-3.10.2/lib -L/home/iron/valerio/petsc-build-3.10.2/lib -lpetsc -Wl,-rpath,/usr/lib/openmpi/lib -L/usr/lib/openmpi/lib -Wl,-rpath,/usr/lib/gcc/x86_64-linux-gnu/4.8 -L/usr/lib/gcc/x86_64-linux-gnu/4.8 -Wl,-rpath,/usr/lib/x86_64-linux-gnu -L/usr/lib/x86_64-linux-gnu -Wl,-rpath,/lib/x86_64-linux-gnu -L/lib/x86_64-linux-gnu -Wl,-rpath,/home/iron/intel/composer_xe_2015.1.133/compiler/lib/intel64 -L/home/iron/intel/composer_xe_2015.1.133/compiler/lib/intel64 -Wl,-rpath,/home/iron/intel/composer_xe_2015.1.133/ipp/lib/intel64 -L/home/iron/intel/composer_xe_2015.1.133/ipp/lib/intel64 -Wl,-rpath,/home/iron/intel/composer_xe_2015.1.133/mkl/lib/intel64 -L/home/iron/intel/composer_xe_2015.1.133/mkl/lib/intel64 -Wl,-rpath,/home/iron/intel/composer_xe_2015.1.133/tbb/lib/intel64/gcc4.4 -L/home/iron/intel/composer_xe_2015.1.133/tbb/lib/intel64/gcc4.4 -llapack -lblas -lm -lX11 -lstdc++ -ldl -lmpi_f90 -lmpi_f77 -lmpi -lhwloc -lgfortran -lm -lgfortran -lm -lgcc_s -lquadmath -lpthread -lstdc++ -ldl -----------------------------------------