[petsc-users] Debugging hints welcome
Matthew Knepley
knepley at gmail.com
Wed Jul 13 21:53:02 CDT 2011
On Wed, Jul 13, 2011 at 4:56 PM, Clemens Domanig <clemens.domanig at uibk.ac.at
> wrote:
> I tried with -mat_no_inode - no effect. Thats the output
>
> [1] MatAssemblyBegin_MPIAIJ(): Stash has 0 entries, uses 0 mallocs.
> [0] MatStashScatterBegin_Private()**: No of messages: 1
> [0] MatStashScatterBegin_Private()**: Mesg_to: 1: size: 704692232
>
^^^^^^^^^^ Do you really mean to set 700M of off-process
values?
I think Barry is correct that it is just taking forever to send this.
Matt
[0] MatAssemblyBegin_MPIAIJ(): Stash has 88086528 entries, uses 13 mallocs.
> [0] MatAssemblyEnd_SeqAIJ(): Matrix size: 553824 X 553824; storage space:
> 24984360 unneeded,19875384 used
> [0] MatAssemblyEnd_SeqAIJ(): Number of mallocs during MatSetValues() is 0
> [0] MatAssemblyEnd_SeqAIJ(): Maximum nonzeros in any row is 42
>
>
> Am 2011-07-13 22:10, schrieb Matthew Knepley:
>
>> On Wed, Jul 13, 2011 at 4:08 PM, Clemens Domanig
>> <clemens.domanig at uibk.ac.at <mailto:clemens.domanig at uibk.**ac.at<clemens.domanig at uibk.ac.at>>>
>> wrote:
>>
>> Hi everyone,
>>
>> maybe some can offer som debugging-hints for my problem.
>>
>>
>> Its possible that there is a bug in the inode routines. Please try
>> running with -mat_no_inode
>>
>> Thanks,
>>
>> Matt
>>
>> My FEM-program uses a shell-element that has depending on the
>> geometry 5 or 6 dof per node.
>>
>> The program uses MPI for parallel solving (LU, mumps).
>> It works fine with all examples that have onyl 5 dof per node and
>> that have a mixture of 5 and 6 dof per node.
>> When doing examples that have 6 dof per node this happens:
>> * when using more than 2 MPI processes everything seems to be fine.
>> * when using 1 or 2 MPI processes MatAssemblyBegin() never finishes
>>
>> This is the last output of -info, -mat_view_info, -vec_view_info
>> (with 2 MPI processes, matrix size 1107648x1107648)
>>
>> [1] MatAssemblyBegin_MPIAIJ(): Stash has 0 entries, uses 0 mallocs.
>> [0] MatStashScatterBegin_Private()**__: No of messages: 1
>> [0] MatStashScatterBegin_Private()**__: Mesg_to: 1: size: 704692232
>> [0] MatAssemblyBegin_MPIAIJ(): Stash has 88086528 entries, uses 13
>> mallocs.
>> [0] MatAssemblyEnd_SeqAIJ(): Matrix size: 553824 X 553824; storage
>> space: 24984360 unneeded,19875384 used
>> [0] MatAssemblyEnd_SeqAIJ(): Number of mallocs during MatSetValues()
>> is 0
>> [0] MatAssemblyEnd_SeqAIJ(): Maximum nonzeros in any row is 42
>> [0] Mat_CheckInode(): Found 184608 nodes of 553824. Limit used: 5.
>> Using Inode routines
>>
>> Thx for your help - respectfully C. Domanig
>>
>>
>>
>>
>> --
>> What most experimenters take for granted before they begin their
>> experiments is infinitely more interesting than any results to which
>> their experiments lead.
>> -- Norbert Wiener
>>
>
>
--
What most experimenters take for granted before they begin their experiments
is infinitely more interesting than any results to which their experiments
lead.
-- Norbert Wiener
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20110713/50abb2c0/attachment.htm>
More information about the petsc-users
mailing list