multiple rhs

David Fuentes fuentesdt at gmail.com
Thu Mar 12 20:11:24 CDT 2009


I'm getting plapack errors in "external library" with

  MatMatMult_MPIDense_MPIDense

with plapack? How is memory handled for a matrix
of type MATMPIDENSE? Are all NxN entries allocated and ready for
use at time of creation? or do I have to MatInsertValues
then Assemble to be ready to use a matrix?





[0]PETSC ERROR: --------------------- Error Message 
----------------------------
--------
[0]PETSC ERROR: Error in external library!
[1]PETSC ERROR: [0]PETSC ERROR: --------------------- Error Message 
------------
------------------------
[1]PETSC ERROR: Error in external library!
Due to aparent bugs in PLAPACK,this is not currently supported!
[1]PETSC ERROR: Due to aparent bugs in PLAPACK,this is not currently 
supported!
[1]PETSC ERROR: 
----------------------------------------------------------------
--------
[1]PETSC ERROR: Petsc Release Version 3.0.0, Patch 4, Fri Mar  6 14:46:08 
CST 20
09
[1]PETSC ERROR: See docs/changes/index.html for recent updates.
[0]PETSC ERROR: [1]PETSC ERROR: See docs/faq.html for hints about trouble 
shooting.





On Thu, 12 Mar 2009, Hong Zhang wrote:

>
>>> 
>>> Is MatCreateMPIDense the recommended matrix type to interface w/ mumps ?
>>> Does it use a sparse direct storage or allocate the full n x n matrix?
>> 
>> 
>> No, MUMPS is "sparse direct" so it uses MPIAIJ.
>
> For mpi dense matrix, you can use plapack
>
> Hong
>> 
>> 
>>> 
>>> df
>>> 
>>> On Thu, 12 Mar 2009, Matthew Knepley wrote:
>>>
>>>  You can try using a sparse direct solver like MUMPS instead of PETSc LU.
>>>>
>>>>  Matt
>>>> 
>>>> On Thu, Mar 12, 2009 at 9:17 AM, David Fuentes <fuentesdt at gmail.com>
>>>> wrote:
>>>>
>>>>  Thanks Hong,
>>>>> 
>>>>> The complete error message is attached. I think I just had too big
>>>>> of a matrix. The matrix i'm trying to factor is 327680 x 327680
>>>>> 
>>>>> 
>>>>> [0]PETSC ERROR: --------------------- Error Message
>>>>> ------------------------------------
>>>>> [0]PETSC ERROR: Out of memory. This could be due to allocating
>>>>> [0]PETSC ERROR: too large an object or bleeding by not properly
>>>>> [0]PETSC ERROR: destroying unneeded objects.
>>>>> [0]PETSC ERROR: Memory allocated 2047323584 Memory used by process
>>>>> 2074058752
>>>>> [0]PETSC ERROR: Try running with -malloc_dump or -malloc_log for info.
>>>>> [0]PETSC ERROR: Memory requested 1258466480!
>>>>> [0]PETSC ERROR:
>>>>> ------------------------------------------------------------------------
>>>>> [0]PETSC ERROR: Petsc Release Version 3.0.0, Patch 2, Wed Jan 14 
>>>>> 22:57:05
>>>>> CST 2009
>>>>> [0]PETSC ERROR: See docs/changes/index.html for recent updates.
>>>>> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting.
>>>>> [0]PETSC ERROR: See docs/index.html for manual pages.
>>>>> [0]PETSC ERROR:
>>>>> ------------------------------------------------------------------------
>>>>> [0]PETSC ERROR: ./RealTimeImaging on a gcc-4.1.2 named DIPWS019 by
>>>>> dfuentes
>>>>> Wed Mar 11 20:30:37 2009
>>>>> [0]PETSC ERROR: Libraries linked from
>>>>> /usr/local/petsc/petsc-3.0.0-p2/gcc-4.1.2-mpich2-1.0.7-dbg/lib
>>>>> [0]PETSC ERROR: Configure run at Sat Jan 31 06:53:09 2009
>>>>> [0]PETSC ERROR: Configure options --download-f-blas-lapack=ifneeded
>>>>> --with-mpi-dir=/usr/local --with-matlab=1 --with-matlab-engine=1
>>>>> --with-matlab-dir=/usr/local/matlab2007a --CFLAGS=-fPIC --with-shared=0
>>>>> [0]PETSC ERROR:
>>>>> ------------------------------------------------------------------------
>>>>> [0]PETSC ERROR: PetscMallocAlign() line 61 in src/sys/memory/mal.c
>>>>> [0]PETSC ERROR: PetscTrMallocDefault() line 194 in src/sys/memory/mtr.c
>>>>> [0]PETSC ERROR: PetscFreeSpaceGet() line 14 in src/mat/utils/freespace.c
>>>>> [0]PETSC ERROR: MatLUFactorSymbolic_SeqAIJ() line 381 in
>>>>> src/mat/impls/aij/seq/aijfact.c
>>>>> [0]PETSC ERROR: MatLUFactorSymbolic() line 2289 in
>>>>> src/mat/interface/matrix.c
>>>>> [0]PETSC ERROR: KalmanFilter::DirectStateUpdate() line 456 in
>>>>> unknowndirectory/src/KalmanFilter.cxx
>>>>> [0]PETSC ERROR: GeneratePRFTmap() line 182 in
>>>>> unknowndirectory/src/MainDriver.cxx
>>>>> [0]PETSC ERROR: main() line 90 in unknowndirectory/src/MainDriver.cxx
>>>>> application called MPI_Abort(MPI_COMM_WORLD, 55) - process 0[unset]:
>>>>> aborting job:
>>>>> application called MPI_Abort(MPI_COMM_WORLD, 55) - process 0
>>>>> 
>>>>> 
>>>>> 
>>>>> 
>>>>> 
>>>>> 
>>>>> 
>>>>> 
>>>>> 
>>>>> 
>>>>> On Thu, 12 Mar 2009, Hong Zhang wrote:
>>>>> 
>>>>>
>>>>>  David,
>>>>>> 
>>>>>> I do not see any problem with the calling sequence.
>>>>>> 
>>>>>> The memory is determined in MatLUFactorSymbolic().
>>>>>> Does your code crashes within MatLUFactorSymbolic()?
>>>>>> Please send us complete error message.
>>>>>> 
>>>>>> Hong
>>>>>> 
>>>>>> On Wed, 11 Mar 2009, David Fuentes wrote:
>>>>>> 
>>>>>>
>>>>>>  Hello,
>>>>>>> 
>>>>>>> I have a sparse matrix, A, with which I want to solve multiple right
>>>>>>> hand
>>>>>>> sides
>>>>>>> with a direct solver.  Is this the correct call sequence ?
>>>>>>> 
>>>>>>>
>>>>>>>  MatGetFactor(A,MAT_SOLVER_PETSC,MAT_FACTOR_LU,&Afact);
>>>>>>>  IS isrow,iscol;
>>>>>>>  MatGetOrdering(A,MATORDERING_ND,&isrow,&iscol);
>>>>>>>  MatLUFactorSymbolic(Afact,A,isrow,iscol,&info);
>>>>>>>  MatLUFactorNumeric(Afact,A,&info);
>>>>>>>  MatMatSolve(Afact,B,X);
>>>>>>> 
>>>>>>> 
>>>>>>> my solve keeps running out of memory
>>>>>>> 
>>>>>>> "[0]PETSC ERROR: Memory requested xxx!"
>>>>>>> 
>>>>>>> 
>>>>>>> is this in bytes? I can't tell if the problem I'm trying to solve
>>>>>>> is too large form my machine or if I just have bug in the call
>>>>>>> sequence.
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>>> thank you,
>>>>>>> David Fuentes
>>>>>>> 
>>>>>>> 
>>>>>>> 
>>>>>> 
>>>> 
>>>> --
>>>> What most experimenters take for granted before they begin their
>>>> experiments
>>>> is infinitely more interesting than any results to which their 
>>>> experiments
>>>> lead.
>>>> -- Norbert Wiener
>>>> 
>>>> 
>> 
>> 
>> -- 
>> What most experimenters take for granted before they begin their 
>> experiments
>> is infinitely more interesting than any results to which their experiments
>> lead.
>> -- Norbert Wiener
>> 
>


More information about the petsc-users mailing list