[petsc-users] MatCreateSeqDense ?

John Chludzinski jchludzinski at gmail.com
Wed Jul 27 14:57:55 CDT 2011


---------------------------------------------- PETSc Performance
Summary: ----------------------------------------------

./ex7 on a arch-cygw named LIZZYB with 1 processor, by John Wed Jul 27
15:47:39 2011
Using Petsc Release Version 3.1.0, Patch 7, Mon Dec 20 14:26:37 CST 2010

                         Max       Max/Min        Avg      Total
Time (sec):           1.312e+03      1.00000   1.312e+03
Objects:              2.003e+04      1.00000   2.003e+04
Flops:                2.564e+11      1.00000   2.564e+11  2.564e+11
Flops/sec:            1.955e+08      1.00000   1.955e+08  1.955e+08
Memory:               1.029e+09      1.00000              1.029e+09
MPI Messages:         0.000e+00      0.00000   0.000e+00  0.000e+00
MPI Message Lengths:  0.000e+00      0.00000   0.000e+00  0.000e+00
MPI Reductions:       2.404e+04      1.00000

Flop counting convention: 1 flop = 1 real number operation of type
(multiply/divide/add/subtract)
                            e.g., VecAXPY() for real vectors of length
N --> 2N flops
                            and VecAXPY() for complex vectors of
length N --> 8N flops

Summary of Stages:   ----- Time ------  ----- Flops -----  ---
Messages ---  -- Message Lengths --  -- Reductions --
                        Avg     %Total     Avg     %Total   counts
%Total     Avg         %Total   counts   %Total
 0:      Main Stage: 1.3119e+03 100.0%  2.5645e+11 100.0%  0.000e+00
0.0%  0.000e+00        0.0%  2.002e+04  83.3%



On Wed, Jul 27, 2011 at 2:21 PM, Jose E. Roman <jroman at dsic.upv.es> wrote:
>
> Don't use "time" to measure performance; instead use -log_summary or PetscGetTime for the interesting part of the computation. In this case, computing the residuals will take a lot of time.
> Jose
>
>
>
> El 27/07/2011, a las 19:28, John Chludzinski escribió:
>
> > $ time ./ex7.exe -f1 k.dat -f2 m.dat -eps_gen_hermitian -eps_type lapack -eps_smallest_real > x.out 2>&1
> >
> > real 19m4.487s
> > user 18m19.650s
> > sys 0m1.762s
> >
> > ---John
> >
> >
> > On Wed, Jul 27, 2011 at 7:03 AM, Jose E. Roman <jroman at dsic.upv.es> wrote:
> > Try running with -eps_gen_hermitian (since ex7 does not assume that the problem is symmetric).
> > Jose
> >
> >
> >
> > El 27/07/2011, a las 12:51, John Chludzinski escribió:
> >
> > > I let the SLEPc code run for ~45 min. when it terminated with the same values I was getting using DSYGV in LAPACK.
> > >
> > > If I write code to directly call LAPACK (i.e., DSYGV), it uses ~3.93 min.  What's up with this?
> > >
> > > ---John
> > >
> > >
> > > On Wed, Jul 27, 2011 at 5:29 AM, John Chludzinski <jchludzinski at gmail.com> wrote:
> > > I'm trying to create a dense matrices from values I'm reading from (binary) files.  I tried the following code:
> > >
> > > Mat A;
> > > int n = SIZE; //4002
> > > double *K = (double *)calloc( sizeof(double), SIZE*SIZE );
> > > ...
> > > MatCreateSeqDense(PETSC_COMM_SELF, n, n, K, &A);
> > > MatView(A,PETSC_VIEWER_BINARY_(PETSC_COMM_WORLD));
> > > ierr = PetscFinalize();CHKERRQ(ierr);
> > >
> > > NOTE:*** I'm converting K to the FORTRAN column major from the C row major order before I call  MatCreateSeqDense(...).
> > >
> > > This appears to work but when I try to use the 2 matrices I thus created with SLEPc ex7 (generalized eigenvalue problem) it never terminates, using:
> > >
> > > ./ex7.exe -f1 k.dat -f2 m.dat -eps_type lapack -eps_smallest_real
> > >
> > > Am I creating the proper PETSc binary (canonical) format for my 2 matrices?
> > >
> > > ---John
> > >
> > >
> > >
> >
> >
>


More information about the petsc-users mailing list