[petsc-users] Using superlu_dist in a direct solve
Sanjay Govindjee
s_g at berkeley.edu
Sun Dec 23 20:15:34 CST 2012
Sorry for the confusion. I thought I was clear. Here is the make line
I was running.
-@${MPIEXEC} -n 2 ./ex6 -ksp_type preonly -pc_type lu
-pc_factor_mat_solver_package superlu_dist -options_left no \
-f arco1 > ex6_1.tmp 2>&1; \
if (${DIFF} output/ex6_1.out ex6_1.tmp) then true; \
else echo ${PWD} ; echo "Possible problem with with ex6_1,
diffs above \n========================================="; fi; \
${RM} -f ex6_1.tmp
If you change superlu_dist to spooles it works just fine as well as any
other iterative methods you care to try. The matrix arcos1 was
downloaded as per the instructions in the makefile.
I will try reproducing the superlu_dist error with
snes/examples/tutorials/ex5 now. (fyi under snes/examples/tests/output
the files ex5_1.out and ex5_2.out are missing one can not run the test
out of the box).
-sanjay
On 12/23/12 6:07 PM, Jed Brown wrote:
> You didn't say what options you were running ex6 with, but with the
> options used for the tests, I see
>
> ~/petsc/src/ksp/ksp/examples/tests$ mpirun.hydra -n 2 ./ex6 -f
> ~/petsc/datafiles/matrices/arco1 -pc_type lu
> -pc_factor_mat_solver_package superlu_dist
> Number of iterations = 1
> Residual norm = 2.23439e-11
>
>
> You need to give precise instructions for how to reproduce the
> behavior you are seeing.
>
> Also, for experimenting with matrices read from files, we prefer
> src/ksp/ksp/examples/tutorials/ex10.c because it is better commented
> and has more features.
>
>
> On Sun, Dec 23, 2012 at 7:08 PM, Sanjay Govindjee <s_g at berkeley.edu
> <mailto:s_g at berkeley.edu>> wrote:
>
> Not sure what you mean by where is your matrix? I am simply
> running ex6 in the ksp/examples/tests directory.
>
> The reason I ran this test is because I was seeing the same
> behavior with my finite element code (on perfectly benign problems).
>
> Is there a built-in test that you use to check that superlu_dist
> is working properly with petsc?
> i.e. something you know that works with with petsc 3.3-p5?
>
> -sanjay
>
>
>
> On 12/23/12 4:56 PM, Jed Brown wrote:
>> Where is your matrix? It might be ending up with a very bad
>> pivot. If the problem can be reproduced, it should be reported to
>> the SuperLU_DIST developers to fix. (Note that we do not see this
>> with other matrices.) You can also try MUMPS.
>>
>>
>> On Sun, Dec 23, 2012 at 6:48 PM, Sanjay Govindjee
>> <s_g at berkeley.edu <mailto:s_g at berkeley.edu>> wrote:
>>
>> I wanted to use SuperLU Dist to perform a direct solve but
>> seem to be encountering
>> a problem. I was wonder if this is a know issue and if there
>> is a solution for it.
>>
>> The problem is easily observed using ex6.c in
>> src/ksp/ksp/examples/tests.
>>
>> Out of the box: make runex6 produces a residual error of
>> O(1e-11), all is well.
>>
>> I then changed the run to run on two processors and add the flag
>> -pc_factor_mat_solver_package spooles this produces a
>> residual error of O(1e-11), all is still well.
>>
>> I then switch over to -pc_factor_mat_solver_package
>> superlu_dist and the
>> residual error comes back as 22.6637! Something seems very wrong.
>>
>> My build is perfectly vanilla:
>>
>> export PETSC_DIR=/Users/sg/petsc-3.3-p5/
>> export PETSC_ARCH=intel
>>
>> ./configure --with-cc=icc --with-fc=ifort \
>> -download-{spooles,parmetis,superlu_dist,prometheus,mpich,ml,hypre,metis}
>>
>> make PETSC_DIR=/Users/sg/petsc-3.3-p5/ PETSC_ARCH=intel all
>> make PETSC_DIR=/Users/sg/petsc-3.3-p5/ PETSC_ARCH=intel test
>>
>> -sanjay
>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20121223/41f58c2b/attachment.html>
More information about the petsc-users
mailing list