[petsc-users] [petsc-dev] How to compute RARt with A and R as distributed (MPI) matrices ?

Jed Brown jed at jedbrown.org
Thu Jun 22 11:17:33 CDT 2017


Hong <hzhang at mcs.anl.gov> writes:

> Jed:
>>
>> >> Is it better this way or as a fallback when !A->ops->rart?  MatPtAP
>> >> handles other combinations like MAIJ.
>> >>
>> >
>> > Do you mean
>> > if ( !A->ops->rart) {
>> >     Mat Rt;
>> >     ierr = MatTranspose(R,MAT_INITIAL_MATRIX,&Rt);CHKERRQ(ierr);
>> >     ierr = MatMatMatMult(R,A,Rt,scall,fill,C);CHKERRQ(ierr);
>> >     ierr = MatDestroy(&Rt);CHKERRQ(ierr);
>> > }
>> > This does NOT work for most matrix formats because we do not have
>> fallbacks
>> > for MatTranspose() and MatMatMult().
>>
>> That's fine; they'll trigger an error and we'll be able to see from the
>> stack that it can be made to work by either implementing the appropriate
>> MatRARt or MatTranspose and MatMatMatMult.
>>
>
> You prefer adding this default, even though it gives error in either
>  MatTranspose() or MatMatMatMult() depends on input matrix format?

Yeah, in the sense that it gives more opportunities to succeed.

> If so, we need add this type of 'default' to all mat operations --
> currently, all routines do
> if (!mat->ops-> )
> SETERRQ1(PetscObjectComm((PetscObject)mat),PETSC_ERR_SUP,"Mat type
> %s",((PetscObject)mat)->type_name);

Probably.
-------------- next part --------------
A non-text attachment was scrubbed...
Name: signature.asc
Type: application/pgp-signature
Size: 832 bytes
Desc: not available
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20170622/6ddf4a95/attachment.pgp>


More information about the petsc-users mailing list