0 KSP Residual norm 6.820168618335e+05 1 KSP Residual norm 6.716770567668e+04 2 KSP Residual norm 5.557965009563e+04 3 KSP Residual norm 4.041779210783e+04 4 KSP Residual norm 2.996229992532e+04 5 KSP Residual norm 2.541300053061e+04 6 KSP Residual norm 1.912337299029e+04 7 KSP Residual norm 1.280773229301e+04 8 KSP Residual norm 8.758027318806e+03 9 KSP Residual norm 5.532581144988e+03 10 KSP Residual norm 4.349711860734e+03 11 KSP Residual norm 2.906070514248e+03 12 KSP Residual norm 2.232965967118e+03 13 KSP Residual norm 1.884439877094e+03 14 KSP Residual norm 1.302888573061e+03 15 KSP Residual norm 9.324264414717e+02 16 KSP Residual norm 7.377397619041e+02 17 KSP Residual norm 5.081290611519e+02 18 KSP Residual norm 3.865239914146e+02 19 KSP Residual norm 2.804241005889e+02 20 KSP Residual norm 2.013507746043e+02 21 KSP Residual norm 1.256750942667e+02 22 KSP Residual norm 1.004817490897e+02 23 KSP Residual norm 7.716326909600e+01 24 KSP Residual norm 5.675788703526e+01 25 KSP Residual norm 4.363941278747e+01 26 KSP Residual norm 3.392815236477e+01 27 KSP Residual norm 2.248290551741e+01 28 KSP Residual norm 1.574252792682e+01 29 KSP Residual norm 1.213470120562e+01 30 KSP Residual norm 9.764382257859e+00 31 KSP Residual norm 9.662422987968e+00 32 KSP Residual norm 9.093036133366e+00 33 KSP Residual norm 8.702140959748e+00 34 KSP Residual norm 7.834276137062e+00 35 KSP Residual norm 5.570273250717e+00 36 KSP Residual norm 4.631822044881e+00 37 KSP Residual norm 3.944000008554e+00 38 KSP Residual norm 3.756468108638e+00 39 KSP Residual norm 2.758231081799e+00 40 KSP Residual norm 2.126510218280e+00 41 KSP Residual norm 1.480343015721e+00 42 KSP Residual norm 1.357536924923e+00 43 KSP Residual norm 1.038759707922e+00 44 KSP Residual norm 9.005321565629e-01 45 KSP Residual norm 6.819540996230e-01 46 KSP Residual norm 5.587570226214e-01 47 KSP Residual norm 5.334482693864e-01 48 KSP Residual norm 4.690887748334e-01 49 KSP Residual norm 4.001938903679e-01 50 KSP Residual norm 3.622102586575e-01 51 KSP Residual norm 3.380418951937e-01 52 KSP Residual norm 2.494779901818e-01 53 KSP Residual norm 2.384213135369e-01 54 KSP Residual norm 2.307578769138e-01 55 KSP Residual norm 1.747042756750e-01 56 KSP Residual norm 1.092937922015e-01 57 KSP Residual norm 9.591401193462e-02 58 KSP Residual norm 8.141750031544e-02 59 KSP Residual norm 5.154975524786e-02 KSP Object: 1 MPI process type: dgmres restart=30, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement happy breakdown tolerance 1e-30 Adaptive strategy is used: TRUE Frequency of extracted eigenvalues = 1 using Ritz values Total number of extracted eigenvalues = 1 Maximum number of eigenvalues set to be extracted = 9 relaxation parameter for the adaptive strategy(smv) = 1. Number of matvecs : 62 maximum iterations=10000, nonzero initial guess tolerances: relative=1e-07, absolute=1e-50, divergence=10000. left preconditioning using PRECONDITIONED norm type for convergence test PC Object: 1 MPI process type: hypre HYPRE BoomerAMG preconditioning Cycle type V Maximum number of levels 25 Maximum number of iterations PER hypre call 1 Convergence tolerance PER hypre call 0. Threshold for strong coupling 0.25 Interpolation truncation factor 0.2 Interpolation: max elements per row 0 Number of levels of aggressive coarsening 4 Number of paths for aggressive coarsening 4 Maximum row sums 0.9 Sweeps down 1 Sweeps up 1 Sweeps on coarse 1 Relax down l1scaled-Jacobi Relax up l1scaled-Jacobi Relax on coarse Gaussian-elimination Relax weight (all) 1. Outer relax weight (all) 1. Not using CF-relaxation Not using more complex smoothers. Measure type local Coarsen type Falgout Interpolation type ext+i SpGEMM type cusparse linear system matrix = precond matrix: Mat Object: 1 MPI process type: seqaij rows=1468928, cols=1468928 total: nonzeros=13176768, allocated nonzeros=0 total number of mallocs used during MatSetValues calls=0 not using I-node routines solver time 51.3648228645325 **************************************************************************************************************************************************************** *** WIDEN YOUR WINDOW TO 160 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** **************************************************************************************************************************************************************** ------------------------------------------------------------------ PETSc Performance Summary: ------------------------------------------------------------------ ########################################################## # # # WARNING!!! # # # # This code was compiled with a debugging option. # # To get timing results run ./configure # # using --with-debugging=no, the performance will # # be generally two or three times faster. # # # ########################################################## /marconi_work/FUA36_GBS2N/nicola/miniapp_solver/./test_miniapp on a named r000u10l05 with 1 processor, by nvarini1 Thu Nov 24 16:51:50 2022 Using Petsc Release Version 3.18.1, unknown Max Max/Min Avg Total Time (sec): 5.172e+01 1.000 5.172e+01 Objects: 6.500e+01 1.000 6.500e+01 Flops: 7.643e+09 1.000 7.643e+09 7.643e+09 Flops/sec: 1.478e+08 1.000 1.478e+08 1.478e+08 Memory (bytes): 6.997e+08 1.000 6.997e+08 6.997e+08 MPI Msg Count: 0.000e+00 0.000 0.000e+00 0.000e+00 MPI Msg Len (bytes): 0.000e+00 0.000 0.000e+00 0.000e+00 MPI Reductions: 0.000e+00 0.000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flop ------ --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total Count %Total Avg %Total Count %Total 0: Main Stage: 5.1722e+01 100.0% 7.6433e+09 100.0% 0.000e+00 0.0% 0.000e+00 0.0% 0.000e+00 0.0% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flop: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent AvgLen: average message length (bytes) Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %F - percent flop in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flop over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ ########################################################## # # # WARNING!!! # # # # This code was compiled with a debugging option. # # To get timing results run ./configure # # using --with-debugging=no, the performance will # # be generally two or three times faster. # # # ########################################################## Event Count Time (sec) Flop --- Global --- --- Stage ---- Total Max Ratio Max Ratio Max Ratio Mess AvgLen Reduct %T %F %M %L %R %T %F %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage MatMult 62 1.0 3.4230e+00 1.0 1.54e+09 1.0 0.0e+00 0.0e+00 0.0e+00 7 20 0 0 0 7 20 0 0 0 451 MatConvert 1 1.0 1.7976e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyBegin 2 1.0 1.3828e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyEnd 2 1.0 2.1923e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetRowIJ 1 1.0 6.1989e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatView 1 1.0 3.6788e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecMDot 90 1.0 1.8570e+00 1.0 2.74e+09 1.0 0.0e+00 0.0e+00 0.0e+00 4 36 0 0 0 4 36 0 0 0 1473 VecNorm 62 1.0 3.9238e-02 1.0 1.82e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 2 0 0 0 0 2 0 0 0 4642 VecScale 61 1.0 3.9972e-02 1.0 8.96e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 2242 VecCopy 5 1.0 1.2081e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 96 1.0 1.0933e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 34 1.0 4.8357e-02 1.0 9.99e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 2066 VecMAXPY 92 1.0 2.2219e+00 1.0 2.99e+09 1.0 0.0e+00 0.0e+00 0.0e+00 4 39 0 0 0 4 39 0 0 0 1347 VecAssemblyBegin 2 1.0 3.8147e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAssemblyEnd 2 1.0 1.9073e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecNormalize 61 1.0 7.9533e-02 1.0 2.69e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 4 0 0 0 0 4 0 0 0 3380 KSPSetUp 1 1.0 2.9897e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 1 1.0 4.4527e+01 1.0 7.64e+09 1.0 0.0e+00 0.0e+00 0.0e+00 86 100 0 0 0 86 100 0 0 0 172 KSPGMRESOrthog 59 1.0 3.6771e+00 1.0 5.29e+09 1.0 0.0e+00 0.0e+00 0.0e+00 7 69 0 0 0 7 69 0 0 0 1438 PCSetUp 1 1.0 6.8044e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 13 0 0 0 0 13 0 0 0 0 0 PCApply 63 1.0 3.6415e+01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 70 0 0 0 0 70 0 0 0 0 0 DGMRESCompDefl 1 1.0 8.8946e-01 1.0 1.16e+08 1.0 0.0e+00 0.0e+00 0.0e+00 2 2 0 0 0 2 2 0 0 0 130 DGMRESApplyDefl 30 1.0 3.3060e-01 1.0 2.64e+08 1.0 0.0e+00 0.0e+00 0.0e+00 1 3 0 0 0 1 3 0 0 0 800 --- Event Stage 1: Unknown ------------------------------------------------------------------------------------------------------------------------ Object Type Creations Destructions. Reports information only for process 0. --- Event Stage 0: Main Stage Matrix 2 2 Vector 59 59 Krylov Solver 1 1 Preconditioner 1 1 Viewer 2 1 --- Event Stage 1: Unknown ======================================================================================================================== Average time to get PetscTime(): 5.96046e-07 #PETSc Option Table entries: -ksp_dgmres_force -ksp_initial_guess_nonzero yes -ksp_monitor -ksp_reuse_preconditioner yes -ksp_rtol 1e-7 -ksp_type dgmres -ksp_view -log_view -pc_hypre_boomeramg_agg_nl 4 -pc_hypre_boomeramg_agg_num_paths 4 -pc_hypre_boomeramg_coarsen_type Falgout -pc_hypre_boomeramg_interp_type ext+i -pc_hypre_boomeramg_no_CF true -pc_hypre_boomeramg_relax_type_all l1scaled-Jacobi -pc_hypre_boomeramg_strong_threshold 0.25 -pc_hypre_boomeramg_truncfactor 0.2 -pc_hypre_type boomeramg -pc_type hypre #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure options: --prefix=/marconi_work/FUA36_GBS2N/petsc-install-3.18.1 --with-blaslapack-dir=/cineca/prod/opt/compilers/intel/pe-xe-2018/binary/mkl --download-hypre --with-cxx=mpiicpc --with-fc=mpiifort --with-cc=mpiicc ----------------------------------------- Libraries compiled on 2022-11-10 13:50:37 on r000u06l01 Machine characteristics: Linux-3.10.0-327.36.3.el7.x86_64-x86_64-with-centos-7.2.1511-Core Using PETSc directory: /marconi_work/FUA36_GBS2N/petsc-install-3.18.1 Using PETSc arch: ----------------------------------------- Using C compiler: mpiicc -fPIC -wd1572 -Wno-unknown-pragmas -g -O0 -std=c99 Using Fortran compiler: mpiifort -fPIC -g -O0 -std=c99 ----------------------------------------- Using include paths: -I/marconi_work/FUA36_GBS2N/petsc-install-3.18.1/include -I/cineca/prod/opt/compilers/intel/pe-xe-2018/binary/mkl/include ----------------------------------------- Using C linker: mpiicc Using Fortran linker: mpiifort Using libraries: -Wl,-rpath,/marconi_work/FUA36_GBS2N/petsc-install-3.18.1/lib -L/marconi_work/FUA36_GBS2N/petsc-install-3.18.1/lib -lpetsc -Wl,-rpath,/marconi_work/FUA36_GBS2N/petsc-install-3.18.1/lib -L/marconi_work/FUA36_GBS2N/petsc-install-3.18.1/lib -Wl,-rpath,/cineca/prod/opt/compilers/intel/pe-xe-2018/binary/mkl/lib/intel64 -L/cineca/prod/opt/compilers/intel/pe-xe-2018/binary/mkl/lib/intel64 -Wl,-rpath,/cineca/prod/opt/compilers/intel/pe-xe-2018/binary/impi/2018.4.274/intel64/lib/debug_mt -L/cineca/prod/opt/compilers/intel/pe-xe-2018/binary/impi/2018.4.274/intel64/lib/debug_mt -Wl,-rpath,/cineca/prod/opt/compilers/intel/pe-xe-2018/binary/impi/2018.4.274/intel64/lib -L/cineca/prod/opt/compilers/intel/pe-xe-2018/binary/impi/2018.4.274/intel64/lib -Wl,-rpath,/marconi/prod/opt/compilers/intel/pe-xe-2018/binary/compilers_and_libraries_2018.5.274/linux/compiler/lib/intel64_lin -L/marconi/prod/opt/compilers/intel/pe-xe-2018/binary/compilers_and_libraries_2018.5.274/linux/compiler/lib/intel64_lin -Wl,-rpath,/usr/lib/gcc/x86_64-redhat-linux/4.8.5 -L/usr/lib/gcc/x86_64-redhat-linux/4.8.5 -Wl,-rpath,/opt/intel/mpi-rt/2017.0.0/intel64/lib/debug_mt -Wl,-rpath,/opt/intel/mpi-rt/2017.0.0/intel64/lib -lHYPRE -lmkl_intel_lp64 -lmkl_core -lmkl_sequential -lpthread -lX11 -lstdc++ -ldl -lmpifort -lmpi -lmpigi -lrt -lpthread -lifport -lifcoremt_pic -limf -lsvml -lm -lipgo -lirc -lgcc_s -lirc_s -lquadmath -lstdc++ -ldl ----------------------------------------- ########################################################## # # # WARNING!!! # # # # This code was compiled with a debugging option. # # To get timing results run ./configure # # using --with-debugging=no, the performance will # # be generally two or three times faster. # # # ##########################################################