[petsc-users] Questions Regarding PETSc and Solving Constrained Structural Mechanics Problems
hexioafeng
hexiaofeng at buaa.edu.cn
Sun Jun 15 20:45:43 CDT 2025
Hello,
Here are the options and outputs:
options:
-ksp_type cg -pc_type gamg -pc_gamg_parallel_coarse_grid_solver -pc_fieldsplit_detect_saddle_point -pc_fieldsplit_type schur -pc_fieldsplit_schur_precondition selfp -fieldsplit_1_mat_schur_complement_ainv_type lump -pc_fieldsplit_schur_fact_type full -fieldsplit_0_ksp_type preonly -fieldsplit_0_pc_type gamg -fieldsplit_0_mg_coarse_pc_type_type svd -fieldsplit_1_ksp_type preonly -fieldsplit_1_pc_type bjacobi -fieldsplit_1_sub_pc_type sor -ksp_view -ksp_monitor_true_residual -ksp_converged_reason -fieldsplit_0_mg_levels_ksp_monitor_true_residual -fieldsplit_0_mg_levels_ksp_converged_reason -fieldsplit_1_ksp_monitor_true_residual -fieldsplit_1_ksp_converged_reason
output:
0 KSP unpreconditioned resid norm 2.777777777778e+01 true resid norm 2.777777777778e+01 ||r(i)||/||b|| 1.000000000000e+00
Linear solve did not converge due to DIVERGED_PC_FAILED iterations 0
PC failed due to SUBPC_ERROR
KSP Object: 1 MPI processes
type: cg
maximum iterations=200, initial guess is zero
tolerances: relative=1e-06, absolute=1e-12, divergence=1e+30
left preconditioning
using UNPRECONDITIONED norm type for convergence test
PC Object: 1 MPI processes
type: gamg
type is MULTIPLICATIVE, levels=2 cycles=v
Cycles per PCApply=1
Using externally compute Galerkin coarse grid matrices
GAMG specific options
Threshold for dropping small values in graph on each level =
Threshold scaling factor for each level not specified = 1.
AGG specific options
Symmetric graph false
Number of levels to square graph 1
Number smoothing steps 1
Complexity: grid = 1.00176
Coarse grid solver -- level -------------------------------
KSP Object: (mg_coarse_) 1 MPI processes
type: preonly
maximum iterations=10000, initial guess is zero
tolerances: relative=1e-05, absolute=1e-50, divergence=10000.
left preconditioning
using NONE norm type for convergence test
PC Object: (mg_coarse_) 1 MPI processes
type: bjacobi
number of blocks = 1
Local solver is the same for all blocks, as in the following KSP and PC objects on rank 0:
KSP Object: (mg_coarse_sub_) 1 MPI processes
type: preonly
maximum iterations=1, initial guess is zero
tolerances: relative=1e-05, absolute=1e-50, divergence=10000.
left preconditioning
using NONE norm type for convergence test
PC Object: (mg_coarse_sub_) 1 MPI processes
type: lu
out-of-place factorization
tolerance for zero pivot 2.22045e-14
using diagonal shift on blocks to prevent zero pivot [INBLOCKS]
matrix ordering: nd
factor fill ratio given 5., needed 1.
Factored matrix follows:
Mat Object: 1 MPI processes
type: seqaij
rows=7, cols=7
package used to perform factorization: petsc
total: nonzeros=45, allocated nonzeros=45
using I-node routines: found 3 nodes, limit used is 5
linear system matrix = precond matrix:
Mat Object: 1 MPI processes
type: seqaij
rows=7, cols=7
total: nonzeros=45, allocated nonzeros=45
total number of mallocs used during MatSetValues calls=0
using I-node routines: found 3 nodes, limit used is 5
linear system matrix = precond matrix:
Mat Object: 1 MPI processes
type: mpiaij
rows=7, cols=7
total: nonzeros=45, allocated nonzeros=45
total number of mallocs used during MatSetValues calls=0
using nonscalable MatPtAP() implementation
using I-node (on process 0) routines: found 3 nodes, limit used is 5
Down solver (pre-smoother) on level 1 -------------------------------
KSP Object: (mg_levels_1_) 1 MPI processes
type: chebyshev
eigenvalue estimates used: min = 0., max = 0.
eigenvalues estimate via gmres min 0., max 0.
eigenvalues estimated using gmres with translations [0. 0.1; 0. 1.1]
KSP Object: (mg_levels_1_esteig_) 1 MPI processes
type: gmres
restart=30, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement
happy breakdown tolerance 1e-30
maximum iterations=10, initial guess is zero
tolerances: relative=1e-12, absolute=1e-50, divergence=10000.
left preconditioning
using PRECONDITIONED norm type for convergence test
PC Object: (mg_levels_1_) 1 MPI processes
type: sor
type = local_symmetric, iterations = 1, local iterations = 1, omega = 1.
linear system matrix = precond matrix:
Mat Object: 1 MPI processes
type: mpiaij
rows=624, cols=624
total: nonzeros=25536, allocated nonzeros=25536
total number of mallocs used during MatSetValues calls=0
using I-node (on process 0) routines: found 336 nodes, limit used is 5
estimating eigenvalues using noisy right hand side
maximum iterations=2, nonzero initial guess
tolerances: relative=1e-05, absolute=1e-50, divergence=10000.
left preconditioning
using NONE norm type for convergence test
PC Object: (mg_levels_1_) 1 MPI processes
type: sor
type = local_symmetric, iterations = 1, local iterations = 1, omega = 1. linear system matrix = precond matrix:
Mat Object: 1 MPI processes
type: mpiaij
rows=624, cols=624
total: nonzeros=25536, allocated nonzeros=25536
total number of mallocs used during MatSetValues calls=0
using I-node (on process 0) routines: found 336 nodes, limit used is 5 Up solver (post-smoother) same as down solver (pre-smoother)
linear system matrix = precond matrix:
Mat Object: 1 MPI processes
type: mpiaij
rows=624, cols=624
total: nonzeros=25536, allocated nonzeros=25536
total number of mallocs used during MatSetValues calls=0
using I-node (on process 0) routines: found 336 nodes, limit used is 5
Best regards,
Xiaofeng
> On Jun 14, 2025, at 07:28, Barry Smith <bsmith at petsc.dev> wrote:
>
>
> Matt,
>
> Perhaps we should add options -ksp_monitor_debug and -snes_monitor_debug that turn on all possible monitoring for the (possibly) nested solvers and all of their converged reasons also? Note this is not completely trivial because each preconditioner will have to supply its list based on the current solver options for it.
>
> Then we won't need to constantly list a big string of problem specific monitor options to ask the user to use.
>
> Barry
>
>
>
>
>> On Jun 13, 2025, at 9:09 AM, Matthew Knepley <knepley at gmail.com> wrote:
>>
>> On Thu, Jun 12, 2025 at 10:55 PM hexioafeng <hexiaofeng at buaa.edu.cn <mailto:hexiaofeng at buaa.edu.cn>> wrote:
>>> Dear authors,
>>>
>>> I tried -pc_type game -pc_gamg_parallel_coarse_grid_solver and -pc_type field split -pc_fieldsplit_detect_saddle_point -fieldsplit_0_ksp_type pronely -fieldsplit_0_pc_type game -fieldsplit_0_mg_coarse_pc_type sad -fieldsplit_1_ksp_type pronely -fieldsplit_1_pc_type Jacobi _fieldsplit_1_sub_pc_type for , both options got the KSP_DIVERGE_PC_FAILED error.
>>
>> With any question about convergence, we need to see the output of
>>
>> -ksp_view -ksp_monitor_true_residual -ksp_converged_reason -fieldsplit_0_mg_levels_ksp_monitor_true_residual -fieldsplit_0_mg_levels_ksp_converged_reason -fieldsplit_1_ksp_monitor_true_residual -fieldsplit_1_ksp_converged_reason
>>
>> and all the error output.
>>
>> Thanks,
>>
>> Matt
>>
>>> Thanks,
>>>
>>> Xiaofeng
>>>
>>>
>>>> On Jun 12, 2025, at 20:50, Mark Adams <mfadams at lbl.gov <mailto:mfadams at lbl.gov>> wrote:
>>>>
>>>>
>>>>
>>>> On Thu, Jun 12, 2025 at 8:44 AM Matthew Knepley <knepley at gmail.com <mailto:knepley at gmail.com>> wrote:
>>>>> On Thu, Jun 12, 2025 at 4:58 AM Mark Adams <mfadams at lbl.gov <mailto:mfadams at lbl.gov>> wrote:
>>>>>> Adding this to the PETSc mailing list,
>>>>>>
>>>>>> On Thu, Jun 12, 2025 at 3:43 AM hexioafeng <hexiaofeng at buaa.edu.cn <mailto:hexiaofeng at buaa.edu.cn>> wrote:
>>>>>>>
>>>>>>> Dear Professor,
>>>>>>>
>>>>>>> I hope this message finds you well.
>>>>>>>
>>>>>>> I am an employee at a CAE company and a heavy user of the PETSc library. I would like to thank you for your contributions to PETSc and express my deep appreciation for your work.
>>>>>>>
>>>>>>> Recently, I encountered some difficulties when using PETSc to solve structural mechanics problems with Lagrange multiplier constraints. After searching extensively online and reviewing several papers, I found your previous paper titled "Algebraic multigrid methods for constrained linear systems with applications to contact problems in solid mechanics" seems to be the most relevant and helpful.
>>>>>>>
>>>>>>> The stiffness matrix I'm working with, K, is a block saddle-point matrix of the form (A00 A01; A10 0), where A00 is singular—just as described in your paper, and different from many other articles . I have a few questions regarding your work and would greatly appreciate your insights:
>>>>>>>
>>>>>>> 1. Is the AMG/KKT method presented in your paper available in PETSc? I tried using CG+GAMG directly but received a KSP_DIVERGED_PC_FAILED error. I also attempted to use CG+PCFIELDSPLIT with the following options:
>>>>>>
>>>>>> No
>>>>>>
>>>>>>>
>>>>>>> -pc_type fieldsplit -pc_fieldsplit_detect_saddle_point -pc_fieldsplit_type schur -pc_fieldsplit_schur_precondition selfp -pc_fieldsplit_schur_fact_type full -fieldsplit_0_ksp_type preonly -fieldsplit_0_pc_type gamg -fieldsplit_1_ksp_type preonly -fieldsplit_1_pc_type bjacobi
>>>>>>>
>>>>>>> Unfortunately, this also resulted in a KSP_DIVERGED_PC_FAILED error. Do you have any suggestions?
>>>>>>>
>>>>>>> 2. In your paper, you compare the method with Uzawa-type approaches. To my understanding, Uzawa methods typically require A00 to be invertible. How did you handle the singularity of A00 to construct an M-matrix that is invertible?
>>>>>>>
>>>>>>
>>>>>> You add a regularization term like A01 * A10 (like springs). See the paper or any reference to augmented lagrange or Uzawa
>>>>>>
>>>>>>
>>>>>>> 3. Can i implement the AMG/KKT method in your paper using existing AMG APIs? Implementing a production-level AMG solver from scratch would be quite challenging for me, so I’m hoping to utilize existing AMG interfaces within PETSc or other packages.
>>>>>>>
>>>>>>
>>>>>> You can do Uzawa and make the regularization matrix with matrix-matrix products. Just use AMG for the A00 block.
>>>>>>
>>>>>>
>>>>>>> 4. For saddle-point systems where A00 is singular, can you recommend any more robust or efficient solutions? Alternatively, are you aware of any open-source software packages that can handle such cases out-of-the-box?
>>>>>>>
>>>>>>
>>>>>> No, and I don't think PETSc can do this out-of-the-box, but others may be able to give you a better idea of what PETSc can do.
>>>>>> I think PETSc can do Uzawa or other similar algorithms but it will not do the regularization automatically (it is a bit more complicated than just A01 * A10)
>>>>>
>>>>> One other trick you can use is to have
>>>>>
>>>>> -fieldsplit_0_mg_coarse_pc_type svd
>>>>>
>>>>> This will use SVD on the coarse grid of GAMG, which can handle the null space in A00 as long as the prolongation does not put it back in. I have used this for the Laplacian with Neumann conditions and for freely floating elastic problems.
>>>>>
>>>>
>>>> Good point.
>>>> You can also use -pc_gamg_parallel_coarse_grid_solver to get GAMG to use a on level iterative solver for the coarse grid.
>>>>
>>>>> Thanks,
>>>>>
>>>>> Matt
>>>>>
>>>>>> Thanks,
>>>>>> Mark
>>>>>>>
>>>>>>> Thank you very much for taking the time to read my email. Looking forward to hearing from you.
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> Sincerely,
>>>>>>>
>>>>>>> Xiaofeng He
>>>>>>> -----------------------------------------------------
>>>>>>>
>>>>>>> Research Engineer
>>>>>>>
>>>>>>> Internet Based Engineering, Beijing, China
>>>>>>>
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead.
>>>>> -- Norbert Wiener
>>>>>
>>>>> https://urldefense.us/v3/__https://www.cse.buffalo.edu/*knepley/__;fg!!G_uCfscf7eWS!e0AMQ70ZmiHfz7acc2y7bj16p9tQITJx5wjrNfKN3d2Iu_q0ghMi5CQvuisbw_fPYd8w-s_2iyDRH5xvxgTZ9JBIm5UKKQ$ <https://urldefense.us/v3/__http://www.cse.buffalo.edu/*knepley/__;fg!!G_uCfscf7eWS!f-YJSzthRa7atIa1xs1GPHW53hGIqSenvp1eO2kDsSyf4jv1_Vp0kL9Lg8pyyPeG8al4Im8XlLqGRHw1FxYh$>
>>
>>
>>
>> --
>> What most experimenters take for granted before they begin their experiments is infinitely more interesting than any results to which their experiments lead.
>> -- Norbert Wiener
>>
>> https://urldefense.us/v3/__https://www.cse.buffalo.edu/*knepley/__;fg!!G_uCfscf7eWS!e0AMQ70ZmiHfz7acc2y7bj16p9tQITJx5wjrNfKN3d2Iu_q0ghMi5CQvuisbw_fPYd8w-s_2iyDRH5xvxgTZ9JBIm5UKKQ$ <https://urldefense.us/v3/__http://www.cse.buffalo.edu/*knepley/__;fg!!G_uCfscf7eWS!f-YJSzthRa7atIa1xs1GPHW53hGIqSenvp1eO2kDsSyf4jv1_Vp0kL9Lg8pyyPeG8al4Im8XlLqGRHw1FxYh$>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20250616/807893c8/attachment-0001.html>
More information about the petsc-users
mailing list