Question on using MUMPS in PETSC

Barry Smith bsmith at mcs.anl.gov
Thu Jul 31 20:24:33 CDT 2008


    If m and n are the number of rows and columns of the sparse matrix  
(i.e. it is
tiny problem) then please
send us matrix so we can experiment with it to petsc-maint at mcs.anl.log

   You can send us the matrix by simply running with -ksp_view_binary  
and
sending us the file binaryoutput.

    Barry

On Jul 31, 2008, at 5:56 PM, Randall Mackie wrote:

> When m = n = small (like 50), it works fine. When I set m=n=5000, I  
> see
> the same thing, where one process on the localhost is taking >4 G of  
> RAM,
> while all other processes are taking 137 M.
>
> Is this the standard behavior for MUMPS? It seems strange to me.
>
> Randy
>
>
> Matthew Knepley wrote:
>> Does it work on KSP ex2?
>>  Matt
>> On Thu, Jul 31, 2008 at 4:35 PM, Randall Mackie <rlmackie862 at gmail.com 
>> > wrote:
>>> I've compiled PETSc with MUMPS support, and I'm trying to run a  
>>> small test
>>> problem, but I'm having some problems. It seems to begin just  
>>> fine, but
>>> what I notice is that on one process (out of 64), the memory just  
>>> keeps
>>> going up and up and up until it crashes, while on the other  
>>> processes,
>>> the memory usage is reasonable. I'm wondering if anyone might have  
>>> any idea
>>> why? By the way, my command file is like this:
>>>
>>> -ksp_type preonly
>>> -pc_type lu
>>> -mat_type aijmumps
>>> -mat_mumps_cntl_4 3
>>> -mat_mumps_cntl_9 1
>>>
>>>
>>> Randy
>>>
>>> ps. This happens after the analysis stage and in the factorization  
>>> stage.
>>>
>>>
>




More information about the petsc-users mailing list