[petsc-users] MatTransposeMatMult ends up with an MPI error
Hong Zhang
hzhang at mcs.anl.gov
Wed Oct 17 10:50:51 CDT 2012
Thomas:
Does this occur only for large matrices?
Can you dump your matrices into petsc binary files
(e.g., A.dat, B.dat) and send to us for debugging?
Lately, we added a new implementation of MatTransposeMatMult() in petsc-dev
which is shown much faster than released MatTransposeMatMult().
You might give it a try by
1. install petsc-dev (see http://www.mcs.anl.gov/petsc/developers/index.html
)
2. run your code with option '-mattransposematmult_viamatmatmult 1'
Let us know what you get.
Hong
My code makes use of the function MatTransposeMatMult, and usually it work
> fine! For some larger input data, it now stops with a lot of MPI errors:
>
> fatal error in PMPI_Barrier: Other MPI error, error stack:
> PMPI_Barrier(476)..: MPI_Barrier(comm=0x84000001) failed
> MPIR_Barrier(82)...:
> MPI_Waitall(261): MPI_Waitall(count=9, req_array=0xa787ba0,
> status_array=0xa789240) failed
> MPI_Waitall(113): The supplied request in array element 8 was invalid
> (kind=0)
> Fatal error in PMPI_Barrier: Other MPI error, error stack:
> PMPI_Barrier(476)..: MPI_Barrier(comm=0x84000001) failed
> MPIR_Barrier(82)...:
> mpid_irecv_done(98): read from socket failed - request state:recv(pde)done
>
>
> Here is the stack print from the debugger:
>
> 6, MatTransposeMatMult (matrix.c:8907)
> 6, MatTransposeMatMult_MPIAIJ_**MPIAIJ
> (mpimatmatmult.c:809)
> 6, MatTransposeMatMultSymbolic_**MPIAIJ_MPIAIJ
> (mpimatmatmult.c:1136)
> 6, PetscGatherMessageLengths2 (mpimesg.c:213)
> 6, PMPI_Waitall
> 6, MPIR_Err_return_comm
> 6, MPID_Abort
>
>
> I use PETSc 3.3-p3. Any idea whether this is or could be related to some
> bug in PETSc or whether I make wrong use of the function in some way?
>
> Thomas
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20121017/621c5bee/attachment.html>
More information about the petsc-users
mailing list