multiple rhs
Matthew Knepley
knepley at gmail.com
Thu Mar 12 09:19:19 CDT 2009
You can try using a sparse direct solver like MUMPS instead of PETSc LU.
Matt
On Thu, Mar 12, 2009 at 9:17 AM, David Fuentes <fuentesdt at gmail.com> wrote:
> Thanks Hong,
>
> The complete error message is attached. I think I just had too big
> of a matrix. The matrix i'm trying to factor is 327680 x 327680
>
>
> [0]PETSC ERROR: --------------------- Error Message
> ------------------------------------
> [0]PETSC ERROR: Out of memory. This could be due to allocating
> [0]PETSC ERROR: too large an object or bleeding by not properly
> [0]PETSC ERROR: destroying unneeded objects.
> [0]PETSC ERROR: Memory allocated 2047323584 Memory used by process
> 2074058752
> [0]PETSC ERROR: Try running with -malloc_dump or -malloc_log for info.
> [0]PETSC ERROR: Memory requested 1258466480!
> [0]PETSC ERROR:
> ------------------------------------------------------------------------
> [0]PETSC ERROR: Petsc Release Version 3.0.0, Patch 2, Wed Jan 14 22:57:05
> CST 2009
> [0]PETSC ERROR: See docs/changes/index.html for recent updates.
> [0]PETSC ERROR: See docs/faq.html for hints about trouble shooting.
> [0]PETSC ERROR: See docs/index.html for manual pages.
> [0]PETSC ERROR:
> ------------------------------------------------------------------------
> [0]PETSC ERROR: ./RealTimeImaging on a gcc-4.1.2 named DIPWS019 by dfuentes
> Wed Mar 11 20:30:37 2009
> [0]PETSC ERROR: Libraries linked from
> /usr/local/petsc/petsc-3.0.0-p2/gcc-4.1.2-mpich2-1.0.7-dbg/lib
> [0]PETSC ERROR: Configure run at Sat Jan 31 06:53:09 2009
> [0]PETSC ERROR: Configure options --download-f-blas-lapack=ifneeded
> --with-mpi-dir=/usr/local --with-matlab=1 --with-matlab-engine=1
> --with-matlab-dir=/usr/local/matlab2007a --CFLAGS=-fPIC --with-shared=0
> [0]PETSC ERROR:
> ------------------------------------------------------------------------
> [0]PETSC ERROR: PetscMallocAlign() line 61 in src/sys/memory/mal.c
> [0]PETSC ERROR: PetscTrMallocDefault() line 194 in src/sys/memory/mtr.c
> [0]PETSC ERROR: PetscFreeSpaceGet() line 14 in src/mat/utils/freespace.c
> [0]PETSC ERROR: MatLUFactorSymbolic_SeqAIJ() line 381 in
> src/mat/impls/aij/seq/aijfact.c
> [0]PETSC ERROR: MatLUFactorSymbolic() line 2289 in
> src/mat/interface/matrix.c
> [0]PETSC ERROR: KalmanFilter::DirectStateUpdate() line 456 in
> unknowndirectory/src/KalmanFilter.cxx
> [0]PETSC ERROR: GeneratePRFTmap() line 182 in
> unknowndirectory/src/MainDriver.cxx
> [0]PETSC ERROR: main() line 90 in unknowndirectory/src/MainDriver.cxx
> application called MPI_Abort(MPI_COMM_WORLD, 55) - process 0[unset]:
> aborting job:
> application called MPI_Abort(MPI_COMM_WORLD, 55) - process 0
>
>
>
>
>
>
>
>
>
>
> On Thu, 12 Mar 2009, Hong Zhang wrote:
>
>
>> David,
>>
>> I do not see any problem with the calling sequence.
>>
>> The memory is determined in MatLUFactorSymbolic().
>> Does your code crashes within MatLUFactorSymbolic()?
>> Please send us complete error message.
>>
>> Hong
>>
>> On Wed, 11 Mar 2009, David Fuentes wrote:
>>
>>
>>> Hello,
>>>
>>> I have a sparse matrix, A, with which I want to solve multiple right hand
>>> sides
>>> with a direct solver. Is this the correct call sequence ?
>>>
>>>
>>> MatGetFactor(A,MAT_SOLVER_PETSC,MAT_FACTOR_LU,&Afact);
>>> IS isrow,iscol;
>>> MatGetOrdering(A,MATORDERING_ND,&isrow,&iscol);
>>> MatLUFactorSymbolic(Afact,A,isrow,iscol,&info);
>>> MatLUFactorNumeric(Afact,A,&info);
>>> MatMatSolve(Afact,B,X);
>>>
>>>
>>> my solve keeps running out of memory
>>>
>>> "[0]PETSC ERROR: Memory requested xxx!"
>>>
>>>
>>> is this in bytes? I can't tell if the problem I'm trying to solve
>>> is too large form my machine or if I just have bug in the call sequence.
>>>
>>>
>>>
>>>
>>> thank you,
>>> David Fuentes
>>>
>>>
>>
--
What most experimenters take for granted before they begin their experiments
is infinitely more interesting than any results to which their experiments
lead.
-- Norbert Wiener
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20090312/cbf2390e/attachment.htm>
More information about the petsc-users
mailing list