mesh ordering and partition
Matthew Knepley
knepley at gmail.com
Fri May 16 12:37:52 CDT 2008
On Fri, May 16, 2008 at 11:47 AM, Gong Ding <gdiso at ustc.edu> wrote:
>
> ----- Original Message ----- From: "Matthew Knepley" <knepley at gmail.com>
> To: <petsc-users at mcs.anl.gov>
> Sent: Saturday, May 17, 2008 12:19 AM
> Subject: Re: mesh ordering and partition
>
>
>>
>> I think, if you are using the serial PETSc ILU, you should just use a
>> MatOrdering,
>> which can be done from the command line:
>>
>> -pc_type ilu -pc_factor_mat_ordering_type rcm
>>
>> which I tested on KSP ex2.
>>
>> Matt
>
>
> I am developing parallel code for 3D semiconductor device simulation.
> From the experience of 2D code, the GMRES solver with ILU works well (the
> matrix is asymmetric.)
> As a result, I'd like to use GMRES+ILU again for 3D, in parallel.
> Does -pc_type ilu -pc_factor_mat_ordering_type rcm still work?
> Since the parallel martrix requires continuous index in subdomain, the
> matrix ordering seems troublesome.
> maybe only a local ordering can be done... Am I right?
Its a local ordering. Remember that Block-Jacobi ILU is a LOT worse than
serial ILU. I would not expect it to scale very well. You can try ASM to fix it
up, but there are no guarantees.
Matt
> Gong Ding
>
>
>
--
What most experimenters take for granted before they begin their
experiments is infinitely more interesting than any results to which
their experiments lead.
-- Norbert Wiener
More information about the petsc-users
mailing list