Residual norm 7.68921e-06 Summary of Memory Usage in PETSc Maximum (over computational time) process memory: total 4.9671e+10 max 2.4836e+10 min 2.4835e+10 Current process memory: total 4.5088e+09 max 2.3270e+09 min 2.1818e+09 ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./ex45 on a named glados.dl.ac.uk with 2 processors, by kchockalingam Mon Nov 22 08:19:18 2021 Using Petsc Release Version 3.15.3, Aug 06, 2021 Max Max/Min Avg Total Time (sec): 1.740e+02 1.000 1.740e+02 Objects: 7.220e+02 1.000 7.220e+02 Flop: 1.119e+11 1.001 1.119e+11 2.238e+11 Flop/sec: 6.435e+08 1.001 6.432e+08 1.286e+09 MPI Messages: 6.105e+02 1.005 6.090e+02 1.218e+03 MPI Message Lengths: 2.849e+08 1.000 4.678e+05 5.698e+08 MPI Reductions: 6.940e+02 1.000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flop and VecAXPY() for complex vectors of length N --> 8N flop Summary of Stages: ----- Time ------ ----- Flop ------ --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total Count %Total Avg %Total Count %Total 0: Main Stage: 1.7396e+02 100.0% 2.2377e+11 100.0% 1.218e+03 100.0% 4.678e+05 100.0% 6.760e+02 97.4% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flop: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent AvgLen: average message length (bytes) Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %F - percent flop in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flop over all processors)/(max time over all processors) GPU Mflop/s: 10e-6 * (sum of flop on GPU over all processors)/(max GPU time over all processors) CpuToGpu Count: total number of CPU to GPU copies per processor CpuToGpu Size (Mbytes): 10e-6 * (total size of CPU to GPU copies per processor) GpuToCpu Count: total number of GPU to CPU copies per processor GpuToCpu Size (Mbytes): 10e-6 * (total size of GPU to CPU copies per processor) GPU %F: percent flops on GPU in this event ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flop --- Global --- --- Stage ---- Total GPU - CpuToGpu - - GpuToCpu - GPU Max Ratio Max Ratio Max Ratio Mess AvgLen Reduct %T %F %M %L %R %T %F %M %L %R Mflop/s Mflop/s Count Size Count Size %F --------------------------------------------------------------------------------------------------------------------------------------------------------------- --- Event Stage 0: Main Stage BuildTwoSided 76 1.0 7.5865e-0134.8 0.00e+00 0.0 9.0e+01 4.0e+00 7.6e+01 0 0 7 0 11 0 0 7 0 11 0 0 0 0.00e+00 0 0.00e+00 0 BuildTwoSidedF 29 1.0 1.6669e-0115.4 0.00e+00 0.0 1.6e+01 2.0e+06 2.9e+01 0 0 1 6 4 0 0 1 6 4 0 0 0 0.00e+00 0 0.00e+00 0 MatMult 310 1.0 1.2298e+00 1.2 3.61e+10 1.0 6.4e+02 3.6e+05 5.0e+00 1 32 53 40 1 1 32 53 40 1 58589 121425 2 2.19e+03 0 0.00e+00 100 MatMultAdd 50 1.0 8.1880e-02 1.6 2.23e+09 1.0 9.0e+01 7.0e+04 0.0e+00 0 2 7 1 0 0 2 7 1 0 54400 92984 0 0.00e+00 0 0.00e+00 100 MatMultTranspose 50 1.0 5.8773e-02 1.0 1.95e+09 1.0 1.1e+02 6.1e+04 5.0e+00 0 2 9 1 1 0 2 9 1 1 66343 79796 1 8.43e-03 0 0.00e+00 100 MatSolve 10 0.0 5.5659e-04 0.0 9.10e+02 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 2 2 0 0.00e+00 0 0.00e+00 100 MatSOR 255 1.0 4.7986e+01 1.0 2.92e+10 1.0 0.0e+00 0.0e+00 0.0e+00 27 26 0 0 0 27 26 0 0 0 1217 0 0 0.00e+00 444 1.94e+04 0 MatLUFactorSym 1 1.0 1.0550e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 0 0 0.00e+00 0 0.00e+00 0 MatLUFactorNum 1 1.0 5.7214e-04232.7 2.10e+02 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 0 2 3.96e-04 0 0.00e+00 0 MatConvert 15 1.0 1.3423e+00 1.0 0.00e+00 0.0 2.0e+01 8.5e+04 5.0e+00 1 0 2 0 1 1 0 2 0 1 0 0 0 0.00e+00 8 9.05e+02 0 MatScale 15 1.0 1.3180e+00 1.0 7.97e+08 1.0 1.0e+01 3.4e+05 0.0e+00 1 1 1 1 0 1 1 1 1 0 1209 70111 10 8.91e+02 15 8.93e+02 14 MatResidual 50 1.0 1.8463e-01 1.7 6.02e+09 1.0 1.0e+02 3.4e+05 0.0e+00 0 5 8 6 0 0 5 8 6 0 65185 111619 50 4.00e-04 0 0.00e+00 100 MatAssemblyBegin 35 1.0 1.3208e-01 2.6 0.00e+00 0.0 1.6e+01 2.0e+06 1.5e+01 0 0 1 6 2 0 0 1 6 2 0 0 0 0.00e+00 0 0.00e+00 0 MatAssemblyEnd 35 1.0 6.3233e+00 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 5.6e+01 4 0 0 0 8 4 0 0 0 8 0 0 0 0.00e+00 0 0.00e+00 0 MatGetRowIJ 1 0.0 1.5944e-06 0.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 0 0 0.00e+00 0 0.00e+00 0 MatCreateSubMat 2 1.0 4.7421e-04 1.0 0.00e+00 0.0 5.0e+00 7.9e+01 2.8e+01 0 0 0 0 4 0 0 0 0 4 0 0 0 0.00e+00 2 1.96e-04 0 MatGetOrdering 1 0.0 2.5809e-05 0.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 0 0 0.00e+00 0 0.00e+00 0 MatCoarsen 5 1.0 3.7309e+00 1.0 0.00e+00 0.0 6.0e+01 2.3e+05 1.5e+01 2 0 5 2 2 2 0 5 2 2 0 0 0 0.00e+00 0 0.00e+00 0 MatAXPY 5 1.0 3.1612e+00 1.0 2.76e+07 1.0 0.0e+00 0.0e+00 5.0e+00 2 0 0 0 1 2 0 0 0 1 17 0 0 0.00e+00 10 8.91e+02 0 MatMatMultSym 5 1.0 3.8528e+00 1.0 5.75e+08 1.0 3.0e+01 2.3e+05 3.0e+01 2 1 2 1 4 2 1 2 1 4 298 1762 43 3.78e+03 30 1.34e+03 100 MatMatMultNum 5 1.0 1.3230e-02 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 0 0 0.00e+00 0 0.00e+00 0 MatPtAPSymbolic 5 1.0 7.6662e+00 1.0 7.03e+09 1.0 1.6e+02 1.2e+06 4.0e+01 4 6 13 34 6 4 6 13 34 6 1832 4422 47 6.68e+03 40 2.86e+03 100 MatPtAPNumeric 5 1.0 2.0198e+00 1.0 7.00e+09 1.0 1.0e+01 3.2e+06 0.0e+00 1 6 1 6 0 1 6 1 6 0 6924 7248 20 4.64e+02 0 0.00e+00 100 MatTrnMatMultSym 1 1.0 3.4088e+01 1.0 0.00e+00 0.0 1.0e+01 4.0e+06 1.2e+01 20 0 1 7 2 20 0 1 7 2 0 0 0 0.00e+00 0 0.00e+00 0 MatGetLocalMat 6 1.0 2.3597e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 0 0 8 8.90e+02 5 5.56e+02 0 MatGetBrAoCol 10 1.0 9.3712e-02 1.0 0.00e+00 0.0 6.0e+01 6.9e+05 0.0e+00 0 0 5 7 0 0 0 5 7 0 0 0 0 0.00e+00 0 0.00e+00 0 MatCUSPARSCopyTo 68 1.1 1.4593e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 0 0 66 9.68e+03 0 0.00e+00 0 MatCUSPARSCopyFr 30 1.0 4.0342e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 0 0 0.00e+00 30 2.69e+03 0 MatCUSPARSSolAnl 2 0.0 2.5864e-04 0.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 0 0 0.00e+00 0 0.00e+00 0 MatCUSPARSGenT 11 1.0 2.7540e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 0 0 0.00e+00 0 0.00e+00 0 MatSetPreallCOO 10 1.0 2.4171e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 3.0e+01 1 0 0 0 4 1 0 0 0 4 0 0 50 6.46e+03 30 1.15e+03 0 MatSetValuesCOO 10 1.0 2.2698e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 0 0 0.00e+00 0 0.00e+00 0 KSPSetUp 13 1.0 1.3010e+01 1.0 1.94e+10 1.0 1.0e+02 3.4e+05 1.6e+02 7 17 8 6 22 7 17 8 6 23 2977 116420 170 2.43e+03 155 2.21e+03 67 KSPSolve 1 1.0 3.9555e+01 1.0 6.34e+10 1.0 6.2e+02 2.7e+05 3.3e+01 23 57 51 30 5 23 57 51 30 5 3202 107147 830 9.02e+03 422 1.72e+04 64 KSPGMRESOrthog 100 1.0 5.5859e-01 1.1 1.21e+10 1.0 0.0e+00 0.0e+00 1.0e+02 0 11 0 0 14 0 11 0 0 15 43431 127991 150 2.21e+03 100 5.32e-01 100 DMCreateMat 1 1.0 1.7780e+01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 7.0e+00 10 0 0 0 1 10 0 0 0 1 0 0 0 0.00e+00 0 0.00e+00 0 SFSetGraph 56 1.0 1.3704e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 0 0 0.00e+00 0 0.00e+00 0 SFSetUp 47 1.0 7.6317e-01 4.1 0.00e+00 0.0 1.6e+02 3.6e+05 4.7e+01 0 0 13 10 7 0 0 13 10 7 0 0 0 0.00e+00 0 0.00e+00 0 SFBcastBegin 25 1.0 4.3326e-03 1.2 0.00e+00 0.0 5.0e+01 5.4e+05 0.0e+00 0 0 4 5 0 0 0 4 5 0 0 0 0 0.00e+00 0 0.00e+00 0 SFBcastEnd 25 1.0 5.9710e-02 4.5 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 0 0 0.00e+00 0 0.00e+00 0 SFReduceBegin 25 1.0 3.0703e-03 1.0 0.00e+00 0.0 5.0e+01 1.9e+06 0.0e+00 0 0 4 17 0 0 0 4 17 0 0 0 0 0.00e+00 0 0.00e+00 0 SFReduceEnd 25 1.0 2.1760e-01 4.4 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 0 0 0.00e+00 0 0.00e+00 0 SFFetchOpBegin 5 1.0 1.1191e-03 1.5 0.00e+00 0.0 1.0e+01 1.6e+06 0.0e+00 0 0 1 3 0 0 0 1 3 0 0 0 0 0.00e+00 0 0.00e+00 0 SFFetchOpEnd 5 1.0 9.7333e-03 1.0 0.00e+00 0.0 1.0e+01 1.6e+06 0.0e+00 0 0 1 3 0 0 0 1 3 0 0 0 0 0.00e+00 0 0.00e+00 0 SFPack 486 1.0 9.0454e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 0 0 0.00e+00 0 0.00e+00 0 SFUnpack 491 1.0 2.3874e-02 1.2 3.99e+05 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 33 0 0 0.00e+00 0 0.00e+00 100 VecMDot 100 1.0 4.3873e-01 1.2 6.07e+09 1.0 0.0e+00 0.0e+00 1.0e+02 0 5 0 0 14 0 5 0 0 15 27648 172456 50 2.21e+03 100 5.32e-01 100 VecTDot 18 1.0 1.1606e-02 1.0 8.97e+08 1.0 0.0e+00 0.0e+00 1.8e+01 0 1 0 0 3 0 1 0 0 3 154578 156117 0 0.00e+00 18 1.44e-04 100 VecNorm 121 1.0 1.0401e-01 2.4 1.76e+09 1.0 0.0e+00 0.0e+00 1.2e+02 0 2 0 0 17 0 2 0 0 18 33867 196387 5 2.21e+02 121 9.68e-04 100 VecScale 110 1.0 1.7113e-02 1.0 6.07e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 70881 71050 110 8.80e-04 0 0.00e+00 100 VecCopy 162 1.0 9.8063e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 0 0 0.00e+00 0 0.00e+00 0 VecSet 357 1.0 4.8593e-02 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 0 0 0.00e+00 0 0.00e+00 0 VecAXPY 29 1.0 5.3274e-02 1.0 1.06e+09 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 39687 101587 30 1.99e+02 0 0.00e+00 100 VecAYPX 308 1.0 7.4778e-01 1.0 3.71e+09 1.0 0.0e+00 0.0e+00 0.0e+00 0 3 0 0 0 0 3 0 0 0 9914 58799 409 4.61e+03 0 0.00e+00 100 VecAXPBYCZ 100 1.0 6.5374e-01 1.0 5.51e+09 1.0 0.0e+00 0.0e+00 0.0e+00 0 5 0 0 0 0 5 0 0 0 16868 179853 400 4.41e+03 0 0.00e+00 100 VecMAXPY 110 1.0 1.4140e-01 1.0 7.17e+09 1.0 0.0e+00 0.0e+00 0.0e+00 0 6 0 0 0 0 6 0 0 0 101383 101406 110 5.20e-03 0 0.00e+00 100 VecAssemblyBegin 15 1.0 8.2847e-0219.8 0.00e+00 0.0 0.0e+00 0.0e+00 1.4e+01 0 0 0 0 2 0 0 0 0 2 0 0 0 0.00e+00 0 0.00e+00 0 VecAssemblyEnd 15 1.0 2.3514e-05 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 0 0 0.00e+00 0 0.00e+00 0 VecPointwiseMult 55 1.0 4.6009e-02 1.0 3.03e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 13182 48660 5 2.21e+02 0 0.00e+00 100 VecScatterBegin 431 1.0 9.3988e-02 1.0 0.00e+00 0.0 9.0e+02 3.1e+05 1.7e+01 0 0 74 49 2 0 0 74 49 3 0 0 0 0.00e+00 0 0.00e+00 0 VecScatterEnd 431 1.0 3.1769e-01 7.3 3.99e+05 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 2 0 0 0.00e+00 0 0.00e+00 100 VecSetRandom 5 1.0 3.0144e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 0 0 0.00e+00 0 0.00e+00 0 VecNormalize 110 1.0 7.0528e-02 1.3 1.82e+09 1.0 0.0e+00 0.0e+00 1.1e+02 0 2 0 0 16 0 2 0 0 16 51596 118147 115 2.21e+02 110 8.80e-04 100 VecCUDACopyTo 262 1.0 1.5966e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 0 0 262 1.19e+04 0 0.00e+00 0 VecCUDACopyFrom 449 1.0 2.9575e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 2 0 0 0 0 2 0 0 0 0 0 0 0 0.00e+00 449 1.94e+04 0 PCGAMGGraph_AGG 5 1.0 2.1182e+01 1.0 5.75e+08 1.0 3.0e+01 1.7e+05 4.5e+01 12 1 2 1 6 12 1 2 1 7 54 0 0 0.00e+00 13 9.06e+02 0 PCGAMGCoarse_AGG 5 1.0 4.8827e+01 1.0 0.00e+00 0.0 8.8e+01 8.7e+05 3.8e+01 28 0 7 13 5 28 0 7 13 6 0 0 0 0.00e+00 0 0.00e+00 0 PCGAMGProl_AGG 5 1.0 7.0233e+00 1.0 0.00e+00 0.0 4.8e+01 2.7e+05 7.9e+01 4 0 4 2 11 4 0 4 2 12 0 0 0 0.00e+00 0 0.00e+00 0 PCGAMGPOpt_AGG 5 1.0 1.1097e+01 1.0 1.42e+10 1.0 1.5e+02 2.8e+05 1.8e+02 6 13 12 7 27 6 13 12 7 27 2555 31195 175 7.08e+03 155 3.12e+03 99 GAMG: createProl 5 1.0 8.8829e+01 1.0 1.48e+10 1.0 3.2e+02 4.3e+05 3.5e+02 51 13 26 24 50 51 13 26 24 51 332 30620 175 7.08e+03 168 4.02e+03 95 Graph 10 1.0 2.1133e+01 1.0 5.75e+08 1.0 3.0e+01 1.7e+05 4.5e+01 12 1 2 1 6 12 1 2 1 7 54 0 0 0.00e+00 13 9.06e+02 0 MIS/Agg 5 1.0 3.7314e+00 1.0 0.00e+00 0.0 6.0e+01 2.3e+05 1.5e+01 2 0 5 2 2 2 0 5 2 2 0 0 0 0.00e+00 0 0.00e+00 0 SA: col data 5 1.0 1.1503e+00 1.0 0.00e+00 0.0 3.6e+01 3.3e+05 3.4e+01 1 0 3 2 5 1 0 3 2 5 0 0 0 0.00e+00 0 0.00e+00 0 SA: frmProl0 5 1.0 5.6186e+00 1.0 0.00e+00 0.0 1.2e+01 1.2e+05 2.5e+01 3 0 1 0 4 3 0 1 0 4 0 0 0 0.00e+00 0 0.00e+00 0 SA: smooth 5 1.0 8.1271e+00 1.0 8.25e+08 1.0 3.0e+01 2.3e+05 4.5e+01 5 1 2 1 6 5 1 2 1 7 203 2055 53 4.67e+03 50 3.12e+03 83 GAMG: partLevel 5 1.0 9.6874e+00 1.0 1.40e+10 1.0 1.8e+02 1.2e+06 9.3e+01 6 13 15 40 13 6 13 15 40 14 2894 5490 67 7.14e+03 42 2.86e+03 100 repartition 1 1.0 1.4764e-03 1.0 0.00e+00 0.0 1.4e+01 3.7e+01 5.3e+01 0 0 1 0 8 0 0 1 0 8 0 0 0 0.00e+00 2 1.96e-04 0 Invert-Sort 1 1.0 6.1192e-05 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 6.0e+00 0 0 0 0 1 0 0 0 0 1 0 0 0 0.00e+00 0 0.00e+00 0 Move A 1 1.0 7.0602e-04 1.0 0.00e+00 0.0 5.0e+00 7.9e+01 1.5e+01 0 0 0 0 2 0 0 0 0 2 0 0 0 0.00e+00 2 1.96e-04 0 Move P 1 1.0 2.9498e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.6e+01 0 0 0 0 2 0 0 0 0 2 0 0 0 0.00e+00 0 0.00e+00 0 PCGAMG Squ l00 1 1.0 3.4088e+01 1.0 0.00e+00 0.0 1.0e+01 4.0e+06 1.2e+01 20 0 1 7 2 20 0 1 7 2 0 0 0 0.00e+00 0 0.00e+00 0 PCGAMG Gal l00 1 1.0 5.5473e+00 1.0 5.28e+09 1.0 3.4e+01 2.4e+06 8.0e+00 3 5 3 14 1 3 5 3 14 1 1903 4210 14 5.01e+03 8 2.05e+03 100 PCGAMG Opt l00 1 1.0 2.8926e+00 1.0 3.48e+08 1.0 6.0e+00 7.2e+05 6.0e+00 2 0 0 1 1 2 0 0 1 1 241 1610 9 3.02e+03 6 1.09e+03 100 PCGAMG Gal l01 1 1.0 3.0978e+00 1.0 6.19e+09 1.0 3.4e+01 3.4e+06 8.0e+00 2 6 3 21 1 2 6 3 21 1 3993 7014 14 1.94e+03 8 7.30e+02 100 PCGAMG Opt l01 1 1.0 7.3372e-01 1.0 1.59e+08 1.0 6.0e+00 3.0e+05 6.0e+00 0 0 0 0 1 0 0 0 0 1 433 2001 8 6.69e+02 6 2.15e+02 100 PCGAMG Gal l02 1 1.0 9.7982e-01 1.0 2.45e+09 1.0 3.4e+01 8.3e+05 8.0e+00 1 2 3 5 1 1 2 3 5 1 5001 6196 14 1.86e+02 8 8.35e+01 100 PCGAMG Opt l02 1 1.0 2.2427e-01 1.0 6.46e+07 1.0 6.0e+00 1.0e+05 6.0e+00 0 0 0 0 1 0 0 0 0 1 573 1958 8 8.82e+01 6 2.79e+01 100 PCGAMG Gal l03 1 1.0 5.5081e-02 1.0 1.07e+08 1.1 3.4e+01 7.0e+04 8.0e+00 0 0 3 0 1 0 0 3 0 1 3780 4920 14 4.86e+00 8 2.43e+00 100 PCGAMG Opt l03 1 1.0 1.4115e-02 1.0 2.89e+06 1.0 6.0e+00 1.1e+04 6.0e+00 0 0 0 0 1 0 0 0 0 1 404 1032 8 2.64e+00 6 8.67e-01 100 PCGAMG Gal l04 1 1.0 6.0144e-03 1.0 3.85e+05 1.1 3.4e+01 6.3e+02 8.0e+00 0 0 3 0 1 0 0 3 0 1 120 256 12 2.26e-02 8 1.21e-02 100 PCGAMG Opt l04 1 1.0 2.6873e-03 1.0 3.82e+04 1.0 6.0e+00 6.8e+02 6.0e+00 0 0 0 0 1 0 0 0 0 1 28 66 8 3.02e-02 6 1.13e-02 100 PCSetUp 2 1.0 1.1120e+02 1.0 4.81e+10 1.0 6.0e+02 6.6e+05 6.1e+02 64 43 49 70 88 64 43 49 70 90 866 13162 414 1.67e+04 365 9.09e+03 85 PCSetUpOnBlocks 10 1.0 6.6756e-04 7.1 2.10e+02 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 0 2 3.96e-04 0 0.00e+00 0 PCApply 10 1.0 3.9430e+01 1.0 5.78e+10 1.0 6.0e+02 2.5e+05 5.0e+00 23 52 49 26 1 23 52 49 26 1 2928 106327 804 9.02e+03 394 1.72e+04 60 --------------------------------------------------------------------------------------------------------------------------------------------------------------- Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Krylov Solver 18 18 320312 0. DMKSP interface 1 1 664 0. Matrix 150 150 27331245384 0. Matrix Coarsen 5 5 3160 0. Distributed Mesh 15 15 76232 0. Index Set 75 75 114024764 0. IS L to G Mapping 21 21 102164020 0. Star Forest Graph 81 81 95112 0. Discrete System 15 15 13560 0. Weak Form 15 15 12360 0. Vector 296 296 10935097304 0. Preconditioner 18 18 17700 0. PetscRandom 10 10 6740 0. Viewer 2 1 848 0. ======================================================================================================================== Average time to get PetscTime(): 2.5332e-08 Average time for MPI_Barrier(): 7.34627e-07 Average time for zero size MPI_Send(): 3.86685e-06 #PETSc Option Table entries: -da_grid_x 368 -da_grid_y 368 -da_grid_z 368 -dm_mat_type mpiaijcusparse -dm_vec_type mpicuda -ksp_type cg -log_view -malloc_log -memory_view -pc_type gamg #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure options: --package-prefix-hash=/home/kchockalingam/petsc-hash-pkgs --with-make-test-np=2 COPTFLAGS="-g -O3 -fno-omit-frame-pointer" FOPTFLAGS="-g -O3 -fno-omit-frame-pointer" CXXOPTFLAGS="-g -O3 -fno-omit-frame-pointer" --with-cuda=1 --with-cuda-arch=70 --with-blaslapack=1 --with-cuda-dir=/apps/packages/cuda/10.1/ --with-mpi-dir=/apps/packages/gcc/7.3.0/openmpi/3.1.2 --download-hypre=1 --download-hypre-configure-arguments=--enable-gpu-profiling=yes,--enable-cusparse=yes,--enable-cublas=yes,--enable-curand=yes,HYPRE_CUDA_SM=70 --with-debugging=no PETSC_ARCH=arch-ci-linux-cuda11-hypre-double ----------------------------------------- Libraries compiled on 2021-11-18 14:19:41 on glados.dl.ac.uk Machine characteristics: Linux-4.18.0-193.6.3.el8_2.x86_64-x86_64-with-centos-8.2.2004-Core Using PETSc directory: /home/kchockalingam/tools/petsc-3.15.3 Using PETSc arch: ----------------------------------------- Using C compiler: /apps/packages/gcc/7.3.0/openmpi/3.1.2/bin/mpicc -fPIC -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -fstack-protector -fvisibility=hidden -g -O3 -fno-omit-frame-pointer Using Fortran compiler: /apps/packages/gcc/7.3.0/openmpi/3.1.2/bin/mpif90 -fPIC -Wall -ffree-line-length-0 -Wno-unused-dummy-argument -g -O3 -fno-omit-frame-pointer ----------------------------------------- Using include paths: -I/home/kchockalingam/tools/petsc-3.15.3/include -I/home/kchockalingam/tools/petsc-3.15.3/arch-ci-linux-cuda11-hypre-double/include -I/home/kchockalingam/petsc-hash-pkgs/194329/include -I/apps/packages/gcc/7.3.0/openmpi/3.1.2/include -I/apps/packages/cuda/10.1/include ----------------------------------------- Using C linker: /apps/packages/gcc/7.3.0/openmpi/3.1.2/bin/mpicc Using Fortran linker: /apps/packages/gcc/7.3.0/openmpi/3.1.2/bin/mpif90 Using libraries: -Wl,-rpath,/home/kchockalingam/tools/petsc-3.15.3/lib -L/home/kchockalingam/tools/petsc-3.15.3/lib -lpetsc -Wl,-rpath,/home/kchockalingam/petsc-hash-pkgs/194329/lib -L/home/kchockalingam/petsc-hash-pkgs/194329/lib -Wl,-rpath,/apps/packages/cuda/10.1/lib64 -L/apps/packages/cuda/10.1/lib64 -Wl,-rpath,/apps/packages/gcc/7.3.0/openmpi/3.1.2/lib -L/apps/packages/gcc/7.3.0/openmpi/3.1.2/lib -Wl,-rpath,/apps/packages/compilers/gcc/7.3.0/lib/gcc/x86_64-pc-linux-gnu/7.3.0 -L/apps/packages/compilers/gcc/7.3.0/lib/gcc/x86_64-pc-linux-gnu/7.3.0 -Wl,-rpath,/apps/packages/compilers/gcc/7.3.0/lib64 -L/apps/packages/compilers/gcc/7.3.0/lib64 -Wl,-rpath,/apps/packages/compilers/gcc/7.3.0/lib -L/apps/packages/compilers/gcc/7.3.0/lib -lHYPRE -llapack -lblas -lcufft -lcublas -lcudart -lcusparse -lcusolver -lcurand -lX11 -lstdc++ -ldl -lmpi_usempi_ignore_tkr -lmpi_mpifh -lmpi -lgfortran -lm -lutil -lrt -lz -lgfortran -lm -lgfortran -lgcc_s -lquadmath -lpthread -lquadmath -lstdc++ -ldl -----------------------------------------