[petsc-users] [petsc-3.2] MatView on dense matrix

Matthew Knepley knepley at gmail.com
Fri Sep 16 12:24:10 CDT 2011


On Fri, Sep 16, 2011 at 12:15 PM, Alexander Grayver <agrayver at gfz-potsdam.de
> wrote:

> **
> Matt,
>
> Can you explain what do you by seeing what format is returned?
>

You can step into the MatView function, into MatView_MPIDense_Binary(),
where it checks
the format, and see what format it thought it had. It just feels like a
memory overwrite since
it should work.

   Matt


> I will check in debugger of course, but still I find it strange.
> I have two versions of petsc (3.1-p7 and 3.2-p1) configured with the same
> options, my code and runs are absolutely identical and I get this error only
> with petsc-3.2.
>
> Regards,
> Alexander
>
>
> On 16.09.2011 15:04, Matthew Knepley wrote:
>
> On Fri, Sep 16, 2011 at 5:38 AM, Alexander Grayver <
> agrayver at gfz-potsdam.de> wrote:
>
>> Hello,
>>
>> With petsc-3.1 I used this code to output MPI dense matrix:
>>
>> PetscViewer          :: dviewer
>> call
>> PetscViewerBinaryOpen(MPI_COMM_WORLD,'A.dat',FILE_MODE_WRITE,dviewer,ierr);
>> CHKERRQ(ierr)
>> call PetscViewerSetFormat(dviewer,PETSC_VIEWER_NATIVE,ierr); CHKERRQ(ierr)
>> call MatView(A,dviewer,ierr); CHKERRQ(ierr)
>> call PetscViewerDestroy(dviewer,ierr); CHKERRQ(ierr)
>>
>> With petsc-3.2 this code fails with error:
>>
>
>  This error just says that the viewer did not return a format of
> PETSC_VIEWER_NATIVE. Either you
> accidentally set it in the wrong viewer, or not on every process, or there
> is memory corruption. I recommend
> using the debugger to see what format is returned.
>
>    Thanks,
>
>       Matt
>
>
>> [7]PETSC ERROR: --------------------- Error Message
>> ------------------------------------
>> [7]PETSC ERROR: No support for this operation for this object type!
>> [7]PETSC ERROR: To store a parallel dense matrix you must first call
>> PetscViewerSetFormat(viewer,PETSC_VIEWER_NATIVE)!
>> [7]PETSC ERROR:
>> ------------------------------------------------------------------------
>> [7]PETSC ERROR: Petsc Release Version 3.2.0, Patch 1, Mon Sep 12 16:01:51
>> CDT 2011
>> [7]PETSC ERROR: See docs/changes/index.html for recent updates.
>> [7]PETSC ERROR: See docs/faq.html for hints about trouble shooting.
>> [7]PETSC ERROR: See docs/index.html for manual pages.
>> [7]PETSC ERROR:
>> ------------------------------------------------------------------------
>> [7]PETSC ERROR: /home/run/test on a openmpi-i named node205 by user Fri
>> Sep 16 09:17:13 2011
>> [7]PETSC ERROR: Libraries linked from
>> /home/lib/petsc-3.2-p1/openmpi-intel-complex-debug-f/lib
>> [7]PETSC ERROR: Configure run at Wed Sep 14 14:32:55 2011
>> [7]PETSC ERROR: Configure options
>> --with-petsc-arch=openmpi-intel-complex-debug-f --with-fortran-interfaces=1
>> --with-mpi-dir=/opt/mpi/intel/openmpi-1.4.2 --with-scalar-type=complex
>> --with-blas-lapack-dir=/opt/intel/Compiler/11.1/072/mkl/lib/em64t
>> --with-precision=double --with-x=0
>> [7]PETSC ERROR:
>> ------------------------------------------------------------------------
>> [7]PETSC ERROR: MatView_MPIDense_Binary() line 667 in
>> /home/lib/petsc-3.2-p1/src/mat/impls/dense/mpi/mpidense.c
>> [7]PETSC ERROR: MatView_MPIDense() line 789 in
>> /home/lib/petsc-3.2-p1/src/mat/impls/dense/mpi/mpidense.c
>> [7]PETSC ERROR: MatView() line 757 in
>> /home/lib/petsc-3.2-p1/src/mat/interface/matrix.c
>> MatView_MPIDense() line 789 in
>> /home/lib/petsc-3.2-p1/src/mat/impls/dense/mpi/mpidense.c
>>
>> See full error attached. Is there something new in petsc-3.2 concerning
>> matrix output?
>>
>> Regards,
>>  Alexander
>>
>
>
>
>  --
> What most experimenters take for granted before they begin their
> experiments is infinitely more interesting than any results to which their
> experiments lead.
> -- Norbert Wiener
>
>
>


-- 
What most experimenters take for granted before they begin their experiments
is infinitely more interesting than any results to which their experiments
lead.
-- Norbert Wiener
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20110916/a4eab4a0/attachment.htm>


More information about the petsc-users mailing list