[petsc-users] petsc 3.7.2 memory usage is much higher when compared to 3.6.1

Mark Adams mfadams at lbl.gov
Thu Jul 7 13:27:46 CDT 2016


On Tue, Jul 5, 2016 at 11:13 PM, Matthew Knepley <knepley at gmail.com> wrote:

> On Tue, Jul 5, 2016 at 3:42 PM, Hassan Raiesi <
> Hassan.Raiesi at aero.bombardier.com> wrote:
>
>> Hi,
>>
>>
>>
>> PETSc 3.7.2 seems to have a much higher memory usage when compared with
>> PETSc- 3.1.1 c, to a point that it crashes our code for large problems that
>> we ran with version 3.6.1 in the past.
>>
>> I have re-compiled the code with same options, and ran the same code
>> linked with the two versions, here are the log-summarie:
>>
>
> According to the log_summary (which you NEED to send in full if we are to
> understand anything), the memory usage is largely the same.
> There are more matrices, which leads me to believe that GAMG is not
> coarsening as quickly. You might consider a non-zero threshold for
> it.
>
>
FYI There are the same number of grids in these two outputs.


> The best way to understand what is happening is to run Massif (from
> valgrind) on both.
>
>   Thanks,
>
>      Matt
>
>
>> -flow_ksp_max_it 20
>>
>> -flow_ksp_monitor_true_residual
>>
>> -flow_ksp_rtol 0.1
>>
>> -flow_ksp_type fgmres
>>
>> -flow_mg_coarse_pc_factor_mat_solver_package mumps
>>
>> -flow_mg_coarse_pc_type lu
>>
>> -flow_mg_levels_ksp_type richardson
>>
>> -flow_mg_levels_pc_type sor
>>
>> -flow_pc_gamg_agg_nsmooths 0
>>
>> -flow_pc_gamg_coarse_eq_limit 2000
>>
>> -flow_pc_gamg_process_eq_limit 2500
>>
>> -flow_pc_gamg_repartition true
>>
>> -flow_pc_gamg_reuse_interpolation true
>>
>> -flow_pc_gamg_square_graph 3
>>
>> -flow_pc_gamg_sym_graph true
>>
>> -flow_pc_gamg_type agg
>>
>> -flow_pc_mg_cycle v
>>
>> -flow_pc_mg_levels 20
>>
>> -flow_pc_mg_type kaskade
>>
>> -flow_pc_type gamg
>>
>> -log_summary
>>
>>
>>
>> Note: it is not specific to PCGAMG, even a bjacobi+fgmres would need more
>> memory (4.5GB/core in version 3.6.1 compared to 6.8GB/core for 3.7.2).
>>
>>
>>
>>
>>
>>
>>
>> Using Petsc Development GIT revision: v3.7.2-812-gc68d048  GIT Date:
>> 2016-07-05 12:04:34 -0400
>>
>>
>>
>>                          Max       Max/Min        Avg      Total
>>
>> Time (sec):           6.760e+02      1.00006   6.760e+02
>>
>> Objects:              1.284e+03      1.00469   1.279e+03
>>
>> Flops:                3.563e+10      1.10884   3.370e+10  1.348e+13
>>
>> Flops/sec:            5.271e+07      1.10884   4.985e+07  1.994e+10
>>
>> MPI Messages:         4.279e+04      7.21359   1.635e+04  6.542e+06
>>
>> MPI Message Lengths:  3.833e+09     17.25274   7.681e+04  5.024e+11
>>
>> MPI Reductions:       4.023e+03      1.00149
>>
>>
>>
>> Flop counting convention: 1 flop = 1 real number operation of type
>> (multiply/divide/add/subtract)
>>
>>                             e.g., VecAXPY() for real vectors of length N
>> --> 2N flops
>>
>>                             and VecAXPY() for complex vectors of length N
>> --> 8N flops
>>
>>
>>
>> Summary of Stages:   ----- Time ------  ----- Flops -----  --- Messages
>> ---  -- Message Lengths --  -- Reductions --
>>
>>                         Avg     %Total     Avg     %Total   counts
>> %Total     Avg         %Total   counts   %Total
>>
>> 0:      Main Stage: 6.7600e+02 100.0%  1.3478e+13 100.0%  6.533e+06
>> 99.9%  7.674e+04       99.9%  4.010e+03  99.7%
>>
>>
>>
>>
>> ------------------------------------------------------------------------------------------------------------------------
>>
>> See the 'Profiling' chapter of the users' manual for details on
>> interpreting output.
>>
>> Phase summary info:
>>
>>    Count: number of times phase was executed
>>
>>    Time and Flops: Max - maximum over all processors
>>
>>                    Ratio - ratio of maximum to minimum over all processors
>>
>>    Mess: number of messages sent
>>
>>    Avg. len: average message length (bytes)
>>
>>    Reduct: number of global reductions
>>
>>    Global: entire computation
>>
>>    Stage: stages of a computation. Set stages with PetscLogStagePush()
>> and PetscLogStagePop().
>>
>>       %T - percent time in this phase         %F - percent flops in this
>> phase
>>
>>       %M - percent messages in this phase     %L - percent message
>> lengths in this phase
>>
>>       %R - percent reductions in this phase
>>
>>    Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time
>> over all processors)
>>
>>
>> ------------------------------------------------------------------------------------------------------------------------
>>
>> Event                Count      Time (sec)     Flops
>>        --- Global ---  --- Stage ---   Total
>>
>>                    Max Ratio  Max     Ratio   Max  Ratio  Mess   Avg len
>> Reduct  %T %F %M %L %R  %T %F %M %L %R Mflop/s
>>
>>
>> ------------------------------------------------------------------------------------------------------------------------
>>
>>
>>
>> --- Event Stage 0: Main Stage
>>
>>
>>
>> MatMult              500 1.0 1.0582e+01 1.2 6.68e+09 1.1 1.9e+06 1.0e+04
>> 0.0e+00  1 19 28  4  0   1 19 29  4  0 237625
>>
>> MatMultTranspose     120 1.0 7.6262e-01 1.3 3.58e+08 1.1 2.4e+05 1.5e+04
>> 0.0e+00  0  1  4  1  0   0  1  4  1  0 180994
>>
>> MatSolve             380 1.0 4.1580e+00 1.1 1.17e+09 1.1 8.6e+03 8.8e+01
>> 6.0e+01  1  3  0  0  1   1  3  0  0  1 105950
>>
>> MatSOR               120 1.0 1.4316e+01 1.2 6.75e+09 1.1 9.5e+05 7.4e+03
>> 0.0e+00  2 19 15  1  0   2 19 15  1  0 177298
>>
>> MatLUFactorSym         2 1.0 2.3449e-01 1.1 0.00e+00 0.0 0.0e+00 0.0e+00
>> 1.0e+01  0  0  0  0  0   0  0  0  0  0     0
>>
>> MatLUFactorNum        60 1.0 8.8820e+00 1.0 1.95e+08 1.2 0.0e+00 0.0e+00
>> 0.0e+00  1  1  0  0  0   1  1  0  0  0  7877
>>
>> MatILUFactorSym        1 1.0 1.9795e-02 1.4 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> MatConvert             6 1.0 2.9893e-02 1.1 0.00e+00 0.0 0.0e+00 0.0e+00
>> 1.8e+01  0  0  0  0  0   0  0  0  0  0     0
>>
>> MatScale               6 1.0 1.8810e-02 1.4 4.52e+06 1.1 2.4e+04 1.5e+03
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0 90171
>>
>> MatAssemblyBegin     782 1.0 1.8294e+01 2.9 0.00e+00 0.0 9.2e+05 4.1e+05
>> 4.2e+02  2  0 14 75 10   2  0 14 75 10     0
>>
>> MatAssemblyEnd       782 1.0 1.4283e+01 3.0 0.00e+00 0.0 4.1e+05 8.7e+02
>> 4.7e+02  1  0  6  0 12   1  0  6  0 12     0
>>
>> MatGetRow        6774900 1.1 9.4289e-01 1.2 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> MatGetRowIJ            3 3.0 6.6261e-036948.0 0.00e+00 0.0 0.0e+00
>> 0.0e+00 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> MatGetSubMatrix       12 1.0 2.6783e+01 1.0 0.00e+00 0.0 1.1e+05 1.3e+05
>> 2.0e+02  4  0  2  3  5   4  0  2  3  5     0
>>
>> MatGetOrdering         3 3.0 7.7400e-03 7.2 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> MatPartitioning        6 1.0 1.8949e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00
>> 1.4e+01  0  0  0  0  0   0  0  0  0  0     0
>>
>> MatCoarsen             6 1.0 9.5692e-02 1.2 0.00e+00 0.0 2.6e+05 1.1e+03
>> 4.1e+01  0  0  4  0  1   0  0  4  0  1     0
>>
>> MatZeroEntries       142 1.0 9.7085e-01 1.1 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> MatTranspose           6 1.0 2.1740e-01 1.0 0.00e+00 0.0 1.9e+05 8.5e+02
>> 7.8e+01  0  0  3  0  2   0  0  3  0  2     0
>>
>> MatPtAP              120 1.0 6.0157e+01 1.0 1.82e+10 1.1 1.5e+06 2.7e+05
>> 4.2e+02  9 51 22 80 10   9 51 22 80 10 114269
>>
>> MatPtAPSymbolic       12 1.0 8.1081e+00 1.0 0.00e+00 0.0 2.2e+05 3.8e+04
>> 8.4e+01  1  0  3  2  2   1  0  3  2  2     0
>>
>> MatPtAPNumeric       120 1.0 5.2205e+01 1.0 1.82e+10 1.1 1.2e+06 3.1e+05
>> 3.4e+02  8 51 19 78  8   8 51 19 78  8 131676
>>
>> MatTrnMatMult          3 1.0 1.8608e+00 1.0 3.23e+07 1.2 8.3e+04 7.9e+03
>> 5.7e+01  0  0  1  0  1   0  0  1  0  1  6275
>>
>> MatTrnMatMultSym       3 1.0 1.3447e+00 1.0 0.00e+00 0.0 6.9e+04 3.8e+03
>> 5.1e+01  0  0  1  0  1   0  0  1  0  1     0
>>
>> MatTrnMatMultNum       3 1.0 5.1695e-01 1.0 3.23e+07 1.2 1.3e+04 3.0e+04
>> 6.0e+00  0  0  0  0  0   0  0  0  0  0 22588
>>
>> MatGetLocalMat       126 1.0 1.0355e+00 1.8 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> MatGetBrAoCol        120 1.0 9.5921e+0019.2 0.00e+00 0.0 5.7e+05 3.3e+04
>> 0.0e+00  1  0  9  4  0   1  0  9  4  0     0
>>
>> VecDot               320 1.0 1.1400e+00 1.6 2.04e+08 1.1 0.0e+00 0.0e+00
>> 3.2e+02  0  1  0  0  8   0  1  0  0  8 68967
>>
>> VecMDot              260 1.0 1.9577e+00 2.8 3.70e+08 1.1 0.0e+00 0.0e+00
>> 2.6e+02  0  1  0  0  6   0  1  0  0  6 72792
>>
>> VecNorm              440 1.0 2.6273e+00 1.9 5.88e+08 1.1 0.0e+00 0.0e+00
>> 4.4e+02  0  2  0  0 11   0  2  0  0 11 86035
>>
>> VecScale             320 1.0 2.1386e-01 1.2 7.91e+07 1.1 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0 141968
>>
>> VecCopy              220 1.0 7.0370e-01 1.2 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> VecSet               862 1.0 7.1000e-01 1.6 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> VecAXPY              440 1.0 8.6790e-01 1.1 3.83e+08 1.1 0.0e+00 0.0e+00
>> 0.0e+00  0  1  0  0  0   0  1  0  0  0 169857
>>
>> VecAYPX              280 1.0 5.7766e-01 1.5 1.92e+08 1.1 0.0e+00 0.0e+00
>> 0.0e+00  0  1  0  0  0   0  1  0  0  0 127599
>>
>> VecMAXPY             300 1.0 9.7396e-01 1.2 4.98e+08 1.1 0.0e+00 0.0e+00
>> 0.0e+00  0  1  0  0  0   0  1  0  0  0 196768
>>
>> VecAssemblyBegin     234 1.0 4.6313e+00 5.6 0.00e+00 0.0 0.0e+00 0.0e+00
>> 6.8e+02  0  0  0  0 17   0  0  0  0 17     0
>>
>> VecAssemblyEnd       234 1.0 5.1503e-0319.5 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> VecScatterBegin     1083 1.0 2.9274e-01 4.5 0.00e+00 0.0 3.8e+06 8.5e+03
>> 2.0e+01  0  0 59  6  0   0  0 59  6  0     0
>>
>> VecScatterEnd       1063 1.0 3.9653e+00 5.6 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> KSPGMRESOrthog        20 1.0 1.7405e+00 3.7 1.28e+08 1.1 0.0e+00 0.0e+00
>> 2.0e+01  0  0  0  0  0   0  0  0  0  0 28232
>>
>> KSPSetUp             222 1.0 6.8469e-02 1.1 0.00e+00 0.0 0.0e+00 0.0e+00
>> 1.2e+01  0  0  0  0  0   0  0  0  0  0     0
>>
>> KSPSolve              60 1.0 1.4767e+02 1.0 3.55e+10 1.1 6.3e+06 7.2e+04
>> 3.2e+03 22100 96 90 79  22100 96 90 79 91007
>>
>> PCGAMGGraph_AGG        6 1.0 6.0792e+00 1.0 4.52e+06 1.1 3.8e+05 9.0e+02
>> 2.5e+02  1  0  6  0  6   1  0  6  0  6   279
>>
>> PCGAMGCoarse_AGG       6 1.0 2.0660e+00 1.0 3.23e+07 1.2 4.2e+05 3.1e+03
>> 1.5e+02  0  0  6  0  4   0  0  6  0  4  5652
>>
>> PCGAMGProl_AGG         6 1.0 1.8842e+00 1.0 0.00e+00 0.0 7.3e+05 3.3e+03
>> 8.6e+02  0  0 11  0 21   0  0 11  0 22     0
>>
>> PCGAMGPOpt_AGG         6 1.0 6.4373e-06 0.0 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> GAMG: createProl       6 1.0 1.0036e+01 1.0 3.68e+07 1.2 1.5e+06 2.7e+03
>> 1.3e+03  1  0 23  1 31   1  0 23  1 31  1332
>>
>>   Graph               12 1.0 6.0783e+00 1.0 4.52e+06 1.1 3.8e+05 9.0e+02
>> 2.5e+02  1  0  6  0  6   1  0  6  0  6   279
>>
>>   MIS/Agg              6 1.0 9.5831e-02 1.2 0.00e+00 0.0 2.6e+05 1.1e+03
>> 4.1e+01  0  0  4  0  1   0  0  4  0  1     0
>>
>>   SA: col data         6 1.0 7.7358e-01 1.0 0.00e+00 0.0 6.7e+05 2.9e+03
>> 7.8e+02  0  0 10  0 19   0  0 10  0 19     0
>>
>>   SA: frmProl0         6 1.0 1.0759e+00 1.0 0.00e+00 0.0 6.2e+04 7.6e+03
>> 6.0e+01  0  0  1  0  1   0  0  1  0  1     0
>>
>> GAMG: partLevel        6 1.0 3.8136e+01 1.0 9.09e+08 1.1 3.8e+05 5.0e+04
>> 5.4e+02  6  3  6  4 13   6  3  6  4 14  9013
>>
>>   repartition          6 1.0 2.7910e+00 1.0 0.00e+00 0.0 4.6e+04 1.3e+02
>> 1.6e+02  0  0  1  0  4   0  0  1  0  4     0
>>
>>   Invert-Sort          6 1.0 2.5045e+00 2.0 0.00e+00 0.0 0.0e+00 0.0e+00
>> 2.4e+01  0  0  0  0  1   0  0  0  0  1     0
>>
>>   Move A               6 1.0 1.4832e+01 1.0 0.00e+00 0.0 8.5e+04 1.7e+05
>> 1.1e+02  2  0  1  3  3   2  0  1  3  3     0
>>
>>   Move P               6 1.0 1.2023e+01 1.0 0.00e+00 0.0 2.4e+04 3.8e+03
>> 1.1e+02  2  0  0  0  3   2  0  0  0  3     0
>>
>> PCSetUp              100 1.0 1.1212e+02 1.0 1.84e+10 1.1 3.2e+06 1.3e+05
>> 2.2e+03 17 52 49 84 54  17 52 49 84 54 62052
>>
>> PCSetUpOnBlocks       40 1.0 1.0386e+00 1.2 1.95e+08 1.2 0.0e+00 0.0e+00
>> 0.0e+00  0  1  0  0  0   0  1  0  0  0 67368
>>
>> PCApply              380 1.0 2.0034e+01 1.1 8.60e+09 1.1 1.5e+06 9.9e+03
>> 6.0e+01  3 24 22  3  1   3 24 22  3  1 161973
>>
>> SFSetGraph            12 1.0 4.9813e-0310.2 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> SFBcastBegin          47 1.0 3.3110e-02 2.6 0.00e+00 0.0 2.6e+05 1.1e+03
>> 6.0e+00  0  0  4  0  0   0  0  4  0  0     0
>>
>> SFBcastEnd            47 1.0 1.3497e-02 3.6 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> SFReduceBegin          6 1.0 1.8593e-02 4.2 0.00e+00 0.0 7.2e+04 4.9e+02
>> 6.0e+00  0  0  1  0  0   0  0  1  0  0     0
>>
>> SFReduceEnd            6 1.0 7.1628e-0318.5 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> BuildTwoSided         12 1.0 3.5771e-02 2.5 0.00e+00 0.0 5.0e+04 4.0e+00
>> 1.2e+01  0  0  1  0  0   0  0  1  0  0     0
>>
>>
>> ------------------------------------------------------------------------------------------------------------------------
>>
>>
>>
>> Memory usage is given in bytes:
>>
>>
>>
>> Object Type          Creations   Destructions     Memory  Descendants'
>> Mem.
>>
>> Reports information only for process 0.
>>
>>
>>
>> --- Event Stage 0: Main Stage
>>
>>
>>
>>               Matrix   302            299   1992700700     0.
>>
>> Matrix Partitioning     6              6         3888     0.
>>
>>       Matrix Coarsen     6              6         3768     0.
>>
>>               Vector   600            600   1582204168     0.
>>
>>       Vector Scatter    87             87      5614432     0.
>>
>>        Krylov Solver    11             11        59472     0.
>>
>>       Preconditioner    11             11        11120     0.
>>
>>          PetscRandom     1              1          638     0.
>>
>>               Viewer     1              0            0     0.
>>
>>            Index Set   247            247      9008420     0.
>>
>> Star Forest Bipartite Graph    12             12        10176     0.
>>
>>
>> ========================================================================================================================
>>
>>
>>
>> And for  petsc 3.6.1:
>>
>>
>>
>> Using Petsc Development GIT revision: v3.6.1-307-g26c82d3  GIT Date:
>> 2015-08-06 11:50:34 -0500
>>
>>
>>
>>                          Max       Max/Min        Avg      Total
>>
>> Time (sec):           5.515e+02      1.00001   5.515e+02
>>
>> Objects:              1.231e+03      1.00490   1.226e+03
>>
>> Flops:                3.431e+10      1.12609   3.253e+10  1.301e+13
>>
>> Flops/sec:            6.222e+07      1.12609   5.899e+07  2.359e+10
>>
>> MPI Messages:         4.432e+04      7.84165   1.504e+04  6.016e+06
>>
>> MPI Message Lengths:  2.236e+09     12.61261   5.027e+04  3.024e+11
>>
>> MPI Reductions:       4.012e+03      1.00150
>>
>>
>>
>> Flop counting convention: 1 flop = 1 real number operation of type
>> (multiply/divide/add/subtract)
>>
>>                             e.g., VecAXPY() for real vectors of length N
>> --> 2N flops
>>
>>                             and VecAXPY() for complex vectors of length N
>> --> 8N flops
>>
>>
>>
>> Summary of Stages:   ----- Time ------  ----- Flops -----  --- Messages
>> ---  -- Message Lengths --  -- Reductions --
>>
>>                         Avg     %Total     Avg     %Total   counts
>> %Total     Avg         %Total   counts   %Total
>>
>> 0:      Main Stage: 5.5145e+02 100.0%  1.3011e+13 100.0%  6.007e+06
>> 99.9%  5.020e+04       99.9%  3.999e+03  99.7%
>>
>>
>>
>>
>> ------------------------------------------------------------------------------------------------------------------------
>>
>> See the 'Profiling' chapter of the users' manual for details on
>> interpreting output.
>>
>> Phase summary info:
>>
>>    Count: number of times phase was executed
>>
>>    Time and Flops: Max - maximum over all processors
>>
>>                    Ratio - ratio of maximum to minimum over all processors
>>
>>    Mess: number of messages sent
>>
>>    Avg. len: average message length (bytes)
>>
>>    Reduct: number of global reductions
>>
>>    Global: entire computation
>>
>>    Stage: stages of a computation. Set stages with PetscLogStagePush()
>> and PetscLogStagePop().
>>
>>       %T - percent time in this phase         %F - percent flops in this
>> phase
>>
>>       %M - percent messages in this phase     %L - percent message
>> lengths in this phase
>>
>>       %R - percent reductions in this phase
>>
>>    Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time
>> over all processors)
>>
>>
>> ------------------------------------------------------------------------------------------------------------------------
>>
>> Event                Count      Time (sec)
>> Flops                             --- Global ---  --- Stage ---   Total
>>
>>                    Max Ratio  Max     Ratio   Max  Ratio  Mess   Avg len
>> Reduct  %T %F %M %L %R  %T %F %M %L %R Mflop/s
>>
>>
>> ------------------------------------------------------------------------------------------------------------------------
>>
>>
>>
>> --- Event Stage 0: Main Stage
>>
>>
>>
>> MatMult              500 1.0 1.0172e+01 1.2 6.68e+09 1.1 1.9e+06 9.9e+03
>> 0.0e+00  2 19 31  6  0   2 19 31  6  0 247182
>>
>> MatMultTranspose     120 1.0 6.9889e-01 1.2 3.56e+08 1.1 2.5e+05 1.4e+04
>> 0.0e+00  0  1  4  1  0   0  1  4  1  0 197492
>>
>> MatSolve             380 1.0 3.9310e+00 1.1 1.17e+09 1.1 1.3e+04 5.7e+01
>> 6.0e+01  1  3  0  0  1   1  3  0  0  2 112069
>>
>> MatSOR               120 1.0 1.3915e+01 1.1 6.73e+09 1.1 9.5e+05 7.4e+03
>> 0.0e+00  2 20 16  2  0   2 20 16  2  0 182405
>>
>> MatLUFactorSym         2 1.0 2.1180e-01 1.8 0.00e+00 0.0 0.0e+00 0.0e+00
>> 1.0e+01  0  0  0  0  0   0  0  0  0  0     0
>>
>> MatLUFactorNum        60 1.0 7.9378e+00 1.0 1.95e+08 1.2 0.0e+00 0.0e+00
>> 0.0e+00  1  1  0  0  0   1  1  0  0  0  8814
>>
>> MatILUFactorSym        1 1.0 2.3076e-02 1.5 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> MatConvert             6 1.0 3.2693e-02 1.5 0.00e+00 0.0 0.0e+00 0.0e+00
>> 1.8e+01  0  0  0  0  0   0  0  0  0  0     0
>>
>> MatScale               6 1.0 2.1923e-02 1.7 4.50e+06 1.1 2.4e+04 1.5e+03
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0 77365
>>
>> MatAssemblyBegin     266 1.0 1.0337e+01 4.4 0.00e+00 0.0 1.8e+05 3.8e+03
>> 4.2e+02  1  0  3  0 10   1  0  3  0 10     0
>>
>> MatAssemblyEnd       266 1.0 3.0336e+00 1.0 0.00e+00 0.0 4.1e+05 8.6e+02
>> 4.7e+02  1  0  7  0 12   1  0  7  0 12     0
>>
>> MatGetRow        6730366 1.1 8.6473e-01 1.2 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> MatGetRowIJ            3 3.0 5.2931e-035550.2 0.00e+00 0.0 0.0e+00
>> 0.0e+00 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> MatGetSubMatrix       12 1.0 2.2689e+01 1.0 0.00e+00 0.0 1.1e+05 1.3e+05
>> 1.9e+02  4  0  2  5  5   4  0  2  5  5     0
>>
>> MatGetOrdering         3 3.0 6.5000e-03 5.4 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> MatPartitioning        6 1.0 2.9801e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00
>> 1.4e+01  1  0  0  0  0   1  0  0  0  0     0
>>
>> MatCoarsen             6 1.0 9.5374e-02 1.1 0.00e+00 0.0 2.5e+05 1.1e+03
>> 3.8e+01  0  0  4  0  1   0  0  4  0  1     0
>>
>> MatZeroEntries        22 1.0 6.1185e-01 1.2 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> MatTranspose           6 1.0 1.9780e-01 1.1 0.00e+00 0.0 1.9e+05 8.6e+02
>> 7.8e+01  0  0  3  0  2   0  0  3  0  2     0
>>
>> MatPtAP              120 1.0 5.2996e+01 1.0 1.70e+10 1.1 9.7e+05 2.1e+05
>> 4.2e+02 10 49 16 67 10  10 49 16 67 11 120900
>>
>> MatPtAPSymbolic       12 1.0 5.8209e+00 1.0 0.00e+00 0.0 2.2e+05 3.7e+04
>> 8.4e+01  1  0  4  3  2   1  0  4  3  2     0
>>
>> MatPtAPNumeric       120 1.0 4.7185e+01 1.0 1.70e+10 1.1 7.6e+05 2.6e+05
>> 3.4e+02  9 49 13 64  8   9 49 13 64  8 135789
>>
>> MatTrnMatMult          3 1.0 1.1679e+00 1.0 3.22e+07 1.2 8.2e+04 8.0e+03
>> 5.7e+01  0  0  1  0  1   0  0  1  0  1  9997
>>
>> MatTrnMatMultSym       3 1.0 6.8366e-01 1.0 0.00e+00 0.0 6.9e+04 3.9e+03
>> 5.1e+01  0  0  1  0  1   0  0  1  0  1     0
>>
>> MatTrnMatMultNum       3 1.0 4.8513e-01 1.0 3.22e+07 1.2 1.3e+04 3.0e+04
>> 6.0e+00  0  0  0  0  0   0  0  0  0  0 24069
>>
>> MatGetLocalMat       126 1.0 1.1939e+00 1.3 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> MatGetBrAoCol        120 1.0 5.9887e-01 2.7 0.00e+00 0.0 5.7e+05 3.3e+04
>> 0.0e+00  0  0  9  6  0   0  0  9  6  0     0
>>
>> MatGetSymTrans        24 1.0 1.4878e-01 1.4 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> VecDot               320 1.0 1.5860e+00 1.5 2.04e+08 1.1 0.0e+00 0.0e+00
>> 3.2e+02  0  1  0  0  8   0  1  0  0  8 49574
>>
>> VecMDot              260 1.0 1.8154e+00 2.5 3.70e+08 1.1 0.0e+00 0.0e+00
>> 2.6e+02  0  1  0  0  6   0  1  0  0  7 78497
>>
>> VecNorm              440 1.0 2.8876e+00 1.8 5.88e+08 1.1 0.0e+00 0.0e+00
>> 4.4e+02  0  2  0  0 11   0  2  0  0 11 78281
>>
>> VecScale             320 1.0 2.2738e-01 1.2 7.88e+07 1.1 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0 133517
>>
>> VecCopy              220 1.0 7.1162e-01 1.2 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> VecSet               862 1.0 7.0683e-01 1.5 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> VecAXPY              440 1.0 9.0657e-01 1.2 3.83e+08 1.1 0.0e+00 0.0e+00
>> 0.0e+00  0  1  0  0  0   0  1  0  0  0 162612
>>
>> VecAYPX              280 1.0 5.8935e-01 1.5 1.92e+08 1.1 0.0e+00 0.0e+00
>> 0.0e+00  0  1  0  0  0   0  1  0  0  0 125070
>>
>> VecMAXPY             300 1.0 9.7644e-01 1.2 4.98e+08 1.1 0.0e+00 0.0e+00
>> 0.0e+00  0  1  0  0  0   0  1  0  0  0 196269
>>
>> VecAssemblyBegin     234 1.0 5.0308e+00 5.5 0.00e+00 0.0 0.0e+00 0.0e+00
>> 6.8e+02  1  0  0  0 17   1  0  0  0 17     0
>>
>> VecAssemblyEnd       234 1.0 1.8253e-03 8.8 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> VecScatterBegin     1083 1.0 2.8195e-01 4.7 0.00e+00 0.0 3.8e+06 8.4e+03
>> 2.0e+01  0  0 64 11  0   0  0 64 11  1     0
>>
>> VecScatterEnd       1063 1.0 3.4924e+00 6.9 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> KSPGMRESOrthog        20 1.0 1.5598e+00 3.2 1.28e+08 1.1 0.0e+00 0.0e+00
>> 2.0e+01  0  0  0  0  0   0  0  0  0  1 31503
>>
>> KSPSetUp             222 1.0 9.7521e-02 1.6 0.00e+00 0.0 0.0e+00 0.0e+00
>> 1.2e+01  0  0  0  0  0   0  0  0  0  0     0
>>
>> KSPSolve              60 1.0 1.3742e+02 1.0 3.42e+10 1.1 5.7e+06 4.4e+04
>> 3.2e+03 25100 95 83 79  25100 95 83 79 94396
>>
>> PCGAMGGraph_AGG        6 1.0 5.7683e+00 1.0 4.50e+06 1.1 3.8e+05 9.1e+02
>> 2.5e+02  1  0  6  0  6   1  0  6  0  6   294
>>
>> PCGAMGCoarse_AGG       6 1.0 1.4101e+00 1.0 3.22e+07 1.2 4.0e+05 3.2e+03
>> 1.4e+02  0  0  7  0  4   0  0  7  0  4  8280
>>
>> PCGAMGProl_AGG         6 1.0 1.8976e+00 1.0 0.00e+00 0.0 7.2e+05 3.4e+03
>> 8.6e+02  0  0 12  1 22   0  0 12  1 22     0
>>
>> PCGAMGPOpt_AGG         6 1.0 5.7220e-06 0.0 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> GAMG: createProl       6 1.0 9.0840e+00 1.0 3.67e+07 1.2 1.5e+06 2.7e+03
>> 1.3e+03  2  0 25  1 31   2  0 25  1 31  1472
>>
>>   Graph               12 1.0 5.7669e+00 1.0 4.50e+06 1.1 3.8e+05 9.1e+02
>> 2.5e+02  1  0  6  0  6   1  0  6  0  6   294
>>
>>   MIS/Agg              6 1.0 9.5481e-02 1.1 0.00e+00 0.0 2.5e+05 1.1e+03
>> 3.8e+01  0  0  4  0  1   0  0  4  0  1     0
>>
>>   SA: col data         6 1.0 8.5414e-01 1.0 0.00e+00 0.0 6.6e+05 3.0e+03
>> 7.8e+02  0  0 11  1 19   0  0 11  1 20     0
>>
>>   SA: frmProl0         6 1.0 1.0123e+00 1.0 0.00e+00 0.0 6.2e+04 7.6e+03
>> 6.0e+01  0  0  1  0  1   0  0  1  0  2     0
>>
>> GAMG: partLevel        6 1.0 3.6150e+01 1.0 8.41e+08 1.1 3.5e+05 5.0e+04
>> 5.3e+02  7  2  6  6 13   7  2  6  6 13  8804
>>
>>   repartition          6 1.0 3.8351e+00 1.0 0.00e+00 0.0 4.7e+04 1.3e+02
>> 1.6e+02  1  0  1  0  4   1  0  1  0  4     0
>>
>>   Invert-Sort          6 1.0 4.4953e+00 1.8 0.00e+00 0.0 0.0e+00 0.0e+00
>> 2.4e+01  1  0  0  0  1   1  0  0  0  1     0
>>
>>   Move A               6 1.0 1.0806e+01 1.0 0.00e+00 0.0 8.5e+04 1.6e+05
>> 1.0e+02  2  0  1  5  3   2  0  1  5  3     0
>>
>>   Move P               6 1.0 1.1953e+01 1.0 0.00e+00 0.0 2.5e+04 3.6e+03
>> 1.0e+02  2  0  0  0  3   2  0  0  0  3     0
>>
>> PCSetUp              100 1.0 1.0166e+02 1.0 1.72e+10 1.1 2.7e+06 8.3e+04
>> 2.2e+03 18 50 44 73 54  18 50 44 73 54 63848
>>
>> PCSetUpOnBlocks       40 1.0 1.0812e+00 1.2 1.95e+08 1.2 0.0e+00 0.0e+00
>> 0.0e+00  0  1  0  0  0   0  1  0  0  0 64711
>>
>> PCApply              380 1.0 1.9359e+01 1.1 8.58e+09 1.1 1.4e+06 9.6e+03
>> 6.0e+01  3 25 24  5  1   3 25 24  5  2 167605
>>
>> SFSetGraph            12 1.0 3.5203e-03 6.8 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> SFBcastBegin          44 1.0 2.4242e-02 3.0 0.00e+00 0.0 2.5e+05 1.1e+03
>> 6.0e+00  0  0  4  0  0   0  0  4  0  0     0
>>
>> SFBcastEnd            44 1.0 3.0994e-02 8.6 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>> SFReduceBegin          6 1.0 1.6784e-02 3.8 0.00e+00 0.0 7.1e+04 5.0e+02
>> 6.0e+00  0  0  1  0  0   0  0  1  0  0     0
>>
>> SFReduceEnd            6 1.0 8.6989e-0332.0 0.00e+00 0.0 0.0e+00 0.0e+00
>> 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
>>
>>
>> ------------------------------------------------------------------------------------------------------------------------
>>
>>
>>
>> Memory usage is given in bytes:
>>
>>
>>
>> Object Type          Creations   Destructions     Memory  Descendants'
>> Mem.
>>
>> Reports information only for process 0.
>>
>>
>>
>> --- Event Stage 0: Main Stage
>>
>>
>>
>>               Matrix   246            243   1730595756     0
>>
>> Matrix Partitioning     6              6         3816     0
>>
>>       Matrix Coarsen     6              6         3720     0
>>
>>               Vector   602            602   1603749672     0
>>
>>       Vector Scatter    87             87      4291136     0
>>
>>        Krylov Solver    12             12        60416     0
>>
>>       Preconditioner    12             12        12040     0
>>
>>               Viewer     1              0            0     0
>>
>>            Index Set   247            247      9018060     0
>>
>> Star Forest Bipartite Graph    12             12        10080     0
>>
>>
>> ========================================================================================================================
>>
>>
>>
>> Any idea why there are more matrix created with version 3.7.2? I only
>> have 2 MatCreate calls and 4 VecCreate calls in my code!, so I assume the
>> others are internally created.
>>
>>
>>
>>
>>
>> Thank you,
>>
>>
>>
>>
>>
>> *Hassan Raiesi, PhD*
>>
>>
>>
>> Advanced Aerodynamics Department
>>
>> Bombardier Aerospace
>>
>>
>>
>> hassan.raiesi at aero.bombardier.com
>>
>>
>>
>> *2351 boul. Alfred-Nobel (BAN1)*
>>
>> *Ville Saint-Laurent, Québec, H4S 2A9*
>>
>>
>>
>>
>>
>>
>>
>> Tél.
>>
>>   514-855-5001    # 62204
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> *CONFIDENTIALITY NOTICE* - This communication may contain privileged or
>> confidential information.
>> If you are not the intended recipient or received this communication by
>> error, please notify the sender
>> and delete the message without copying, forwarding and/or disclosing it.
>>
>>
>>
>>
>>
>>
>>
>
>
>
> --
> What most experimenters take for granted before they begin their
> experiments is infinitely more interesting than any results to which their
> experiments lead.
> -- Norbert Wiener
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20160707/224e56a6/attachment-0001.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: image001.png
Type: image/png
Size: 6402 bytes
Desc: not available
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20160707/224e56a6/attachment-0001.png>


More information about the petsc-users mailing list