[petsc-users] CUDA error

Stefano Zampini stefano.zampini at gmail.com
Wed Apr 15 07:24:30 CDT 2020


Mark

I have fixed few things in the solver and it is tested with the current master.
Can you write a MWE to reproduce the issue? Which version of CUDA and CUSPARSE are you using?
I was planning to reorganize the factor code in AIJCUSPARSE in the next days.

kl-18967:petsc zampins$ git grep "solver_type cusparse"
src/ksp/ksp/examples/tests/ex43.c:      args: -f ${DATAFILESPATH}/matrices/cfd.2.10 -mat_type seqaijcusparse -pc_factor_mat_solver_type cusparse -mat_cusparse_storage_format ell -vec_type cuda -pc_type ilu
src/ksp/ksp/examples/tests/ex43.c:      args: -f ${DATAFILESPATH}/matrices/shallow_water1 -mat_type seqaijcusparse -pc_factor_mat_solver_type cusparse -mat_cusparse_storage_format hyb -vec_type cuda -ksp_type cg -pc_type icc
src/ksp/ksp/examples/tests/ex43.c:      args: -f ${DATAFILESPATH}/matrices/cfd.2.10 -mat_type seqaijcusparse -pc_factor_mat_solver_type cusparse -mat_cusparse_storage_format csr -vec_type cuda -ksp_type bicg -pc_type ilu
src/ksp/ksp/examples/tests/ex43.c:      args: -f ${DATAFILESPATH}/matrices/cfd.2.10 -mat_type seqaijcusparse -pc_factor_mat_solver_type cusparse -mat_cusparse_storage_format csr -vec_type cuda -ksp_type bicg -pc_type ilu -pc_factor_mat_ordering_type nd
src/ksp/ksp/examples/tutorials/ex46.c:      args: -dm_mat_type aijcusparse -dm_vec_type cuda -random_exact_sol -pc_type ilu -pc_factor_mat_solver_type cusparse
src/ksp/ksp/examples/tutorials/ex59.c:     args: -subdomain_mat_type aijcusparse -physical_pc_bddc_dirichlet_pc_factor_mat_solver_type cusparse
src/ksp/ksp/examples/tutorials/ex7.c:      args: -ksp_monitor_short -mat_type aijcusparse -sub_pc_factor_mat_solver_type cusparse -vec_type cuda -sub_ksp_type preonly -sub_pc_type ilu
src/ksp/ksp/examples/tutorials/ex7.c:      args: -ksp_monitor_short -mat_type aijcusparse -sub_pc_factor_mat_solver_type cusparse -vec_type cuda -sub_ksp_type preonly -sub_pc_type ilu
src/ksp/ksp/examples/tutorials/ex7.c:      args: -ksp_monitor_short -mat_type aijcusparse -sub_pc_factor_mat_solver_type cusparse -vec_type cuda
src/ksp/ksp/examples/tutorials/ex7.c:      args: -ksp_monitor_short -mat_type aijcusparse -sub_pc_factor_mat_solver_type cusparse -vec_type cuda
src/ksp/ksp/examples/tutorials/ex71.c:   args: -pde_type Poisson -cells 7,9,8 -dim 3 -ksp_view -pc_bddc_coarse_redundant_pc_type svd -ksp_error_if_not_converged -pc_bddc_dirichlet_pc_type cholesky -pc_bddc_dirichlet_pc_factor_mat_solver_type cusparse -pc_bddc_dirichlet_pc_factor_mat_ordering_type nd -pc_bddc_neumann_pc_type cholesky -pc_bddc_neumann_pc_factor_mat_solver_type cusparse -pc_bddc_neumann_pc_factor_mat_ordering_type nd -matis_localmat_type aijcusparse
src/ksp/ksp/examples/tutorials/ex72.c:      args: -f0 ${DATAFILESPATH}/matrices/medium -ksp_monitor_short -ksp_view -mat_view ascii::ascii_info -mat_type aijcusparse -pc_factor_mat_solver_type cusparse -pc_type ilu -vec_type cuda
src/snes/examples/tutorials/ex12.c:      args: -matis_localmat_type aijcusparse -pc_bddc_dirichlet_pc_factor_mat_solver_type cusparse -pc_bddc_neumann_pc_factor_mat_solver_type cusparse

> On Apr 15, 2020, at 2:20 PM, Mark Adams <mfadams at lbl.gov> wrote:
> 
> I tried using a serial direct solver in cusparse and got bad numerics:
> 
> -vector_type cuda -mat_type aijcusparse -pc_factor_mat_solver_type cusparse 
> 
> Before I start debugging this I wanted to see if there are any known issues that I should be aware of.
> 
> Thanks,

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20200415/e603f15c/attachment-0001.html>


More information about the petsc-users mailing list