[petsc-users] performance issue

Xavier Garnaud xavier.garnaud at ladhyx.polytechnique.fr
Sat Mar 10 11:05:45 CST 2012


I am using an explicit time stepper. The matrices are assembled only once,
and then I use the linear operator for example to compute the least stable
eigenmode(s). I attached the output of log_summary for performing the same
number of time steps using the linear and nonlinear operators.


On Sat, Mar 10, 2012 at 5:10 PM, Jed Brown <jedbrown at mcs.anl.gov> wrote:

> On Sat, Mar 10, 2012 at 09:59, Xavier Garnaud <
> xavier.garnaud at ladhyx.polytechnique.fr> wrote:
>
>> am solving the compressible Navier--Stokes equations in compressible
>> form, so in order to apply the operator, I
>>
>>    1. apply BCs on the flow field
>>    2. compute the flux
>>    3. take the derivative using finite differences
>>    4. apply BCs on the derivatives of the flux
>>
>>
>> In order to apply the linearized operator, I wish to linearize steps 2
>> and 4 (the other are linear). For this I assemble sparse matrices (MPIAIJ).
>> The matrices should be block diagonal -- with square or rectangular blocks
>> --  so I preallocate the whole diagonal blocks (but I only use MatSetValues
>> for nonzero entries). When I do this, the linearized code runs
>> approximately 50% slower (the computation of derivatives takes more that
>> 70% of the time in the non-linear code), so steps 2 and 4 are much slower
>> for the linear operator although the number of operations is very similar.
>> Is this be due to the poor preallocation? Is there a way to improve the
>> performance?
>>
>
> It's not clear to me from this description if you are even using an
> implicit method. Is the linearization for use in a Newton iteration? How
> often do you have to reassemble? Please always send -log_summary output
> with performance questions.
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20120310/1bd0f59c/attachment-0001.htm>
-------------- next part --------------
  nx  = 256
  ny  = 128

  JET MESH 
     ix     = 74
     iy_in  = 44
     iy_out = 65
     rank: 0, i0-i1xj0-j1 : 0 - 74 x 44 - 63
     rank: 2, i0-i1xj0-j1 : 0 - 74 x 64 - 65

  TIME STEPPING
     Tf  = 1 
     CFL = 2 
State loaded from q0.h5 
32768 40960
32768 40960
32768 40960
32768 40960
Euler - x 
Euler - y 
LODI - x 
LODIq - x 
LODI - y 
LODIq - y 
Stress - x 
Stress - x 
dFv/dq - x 
dFv/dtau - x 
dFv/dq - y 
dFv/dtau - y 
 |MatEulerx       | = 21.7871 
 |MatEulery       | = 10.4999 
 |MatLODIx        | = 13.3652 
 |MatLODIy        | = 15.0075 
 |MatLODIqx       | = 4.58531e+06 
 |MatLODIqy       | = 1.00002 
 |MatViscousx_q   | = 0.00122487 
 |MatViscousy_q   | = 0.00125045 
 |MatViscousx_tau | = 1.99893 
 |MatViscousy_tau | = 1.99893 
    dt = 0.00571429 
    |q(1.000000)|/|q(0)| = 1.84842 
Elapsed time (CPU) = 27.2226 s 
************************************************************************************************************************
***             WIDEN YOUR WINDOW TO 120 CHARACTERS.  Use 'enscript -r -fCourier9' to print this document            ***
************************************************************************************************************************

---------------------------------------------- PETSc Performance Summary: ----------------------------------------------

./ns2d on a real_opt named muzo.polytechnique.fr with 4 processors, by garnaud Sat Mar 10 18:02:03 2012
Using Petsc Release Version 3.2.0, Patch 5, Sat Oct 29 13:45:54 CDT 2011 

                         Max       Max/Min        Avg      Total 
Time (sec):           2.762e+01      1.00000   2.762e+01
Objects:              1.900e+02      1.00000   1.900e+02
Flops:                1.068e+10      1.01222   1.065e+10  4.258e+10
Flops/sec:            3.869e+08      1.01222   3.855e+08  1.542e+09
MPI Messages:         3.260e+04      1.00000   3.260e+04  1.304e+05
MPI Message Lengths:  2.277e+08      1.00000   6.984e+03  9.108e+08
MPI Reductions:       4.280e+02      1.00000

Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract)
                            e.g., VecAXPY() for real vectors of length N --> 2N flops
                            and VecAXPY() for complex vectors of length N --> 8N flops

Summary of Stages:   ----- Time ------  ----- Flops -----  --- Messages ---  -- Message Lengths --  -- Reductions --
                        Avg     %Total     Avg     %Total   counts   %Total     Avg         %Total   counts   %Total 
 0:      Main Stage: 2.7615e+01 100.0%  4.2584e+10 100.0%  1.304e+05 100.0%  6.984e+03      100.0%  4.270e+02  99.8% 

------------------------------------------------------------------------------------------------------------------------
See the 'Profiling' chapter of the users' manual for details on interpreting output.
Phase summary info:
   Count: number of times phase was executed
   Time and Flops: Max - maximum over all processors
                   Ratio - ratio of maximum to minimum over all processors
   Mess: number of messages sent
   Avg. len: average message length
   Reduct: number of global reductions
   Global: entire computation
   Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop().
      %T - percent time in this phase         %f - percent flops in this phase
      %M - percent messages in this phase     %L - percent message lengths in this phase
      %R - percent reductions in this phase
   Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors)
------------------------------------------------------------------------------------------------------------------------
Event                Count      Time (sec)     Flops                             --- Global ---  --- Stage ---   Total
                   Max Ratio  Max     Ratio   Max  Ratio  Mess   Avg len Reduct  %T %f %M %L %R  %T %f %M %L %R Mflop/s
------------------------------------------------------------------------------------------------------------------------

--- Event Stage 0: Main Stage

DERIVATIVES        10508 1.0 1.4299e+01 1.0 8.21e+09 1.0 0.0e+00 0.0e+00 0.0e+00 51 77  0  0  0  51 77  0  0  0  2295
FILTERS              350 1.0 1.9905e-01 1.0 1.26e+08 1.0 0.0e+00 0.0e+00 0.0e+00  1  1  0  0  0   1  1  0  0  0  2535
VECMANIP           21716 1.0 2.8288e+00 1.2 0.00e+00 0.0 1.3e+05 7.0e+03 6.0e+00  9  0100100  1   9  0100100  1     0
VecView                6 1.0 3.7352e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  1  0  0  0  0   1  0  0  0  0     0
VecNorm                2 1.0 5.9009e-04 3.3 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00  0  0  0  0  0   0  0  0  0  0     0
VecScale            1800 1.0 7.1079e-02 1.2 5.90e+07 1.0 0.0e+00 0.0e+00 0.0e+00  0  1  0  0  0   0  1  0  0  0  3323
VecCopy              414 1.0 3.7731e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
VecAXPY             7051 1.0 7.2879e-01 1.1 4.97e+08 1.0 0.0e+00 0.0e+00 0.0e+00  3  5  0  0  0   3  5  0  0  0  2726
VecAXPBYCZ           350 1.0 6.3609e-02 1.0 4.01e+07 1.0 0.0e+00 0.0e+00 0.0e+00  0  0  0  0  0   0  0  0  0  0  2524
VecLoad                1 1.0 1.8210e-03 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
VecScatterBegin    31858 1.0 1.5961e+00 1.1 0.00e+00 0.0 1.3e+05 7.0e+03 0.0e+00  6  0100100  0   6  0100100  0     0
VecScatterEnd      31858 1.0 8.4421e-01 1.6 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  2  0  0  0  0   2  0  0  0  0     0
IMPOSEBC_VISC       5251 1.0 7.2675e-03 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
IMPOSEBC_EULER      1945 1.0 8.8332e-03 4.6 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
FLUXES_VISC           22 1.0 4.4665e-03 1.1 4.33e+06 1.0 0.0e+00 0.0e+00 0.0e+00  0  0  0  0  0   0  0  0  0  0  3874
FLUXES_EULER          14 1.0 2.4092e-03 1.3 2.75e+06 1.0 0.0e+00 0.0e+00 0.0e+00  0  0  0  0  0   0  0  0  0  0  4570
STRESSES              12 1.0 1.9977e-03 1.1 1.67e+06 1.0 0.0e+00 0.0e+00 0.0e+00  0  0  0  0  0   0  0  0  0  0  3346
MatMult            12250 1.0 6.6647e+00 1.0 1.34e+09 1.1 0.0e+00 0.0e+00 0.0e+00 24 12  0  0  0  24 12  0  0  0   784
MatMultAdd          8750 1.0 2.5075e+00 1.0 4.13e+08 1.1 0.0e+00 0.0e+00 0.0e+00  9  4  0  0  0   9  4  0  0  0   642
MatAssemblyBegin      12 1.0 7.1454e-04 3.0 0.00e+00 0.0 0.0e+00 0.0e+00 2.4e+01  0  0  0  0  6   0  0  0  0  6     0
MatAssemblyEnd        12 1.0 2.1005e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.1e+02  0  0  0  0 25   0  0  0  0 25     0
TSStep               175 1.0 2.6759e+01 1.0 1.05e+10 1.0 1.3e+05 7.0e+03 6.0e+00 97 99 97 97  1  97 99 97 97  1  1570
TSFunctionEval      1750 1.0 2.6487e+01 1.0 1.04e+10 1.0 1.3e+05 7.0e+03 0.0e+00 96 97 97 97  0  96 97 97 97  0  1562
------------------------------------------------------------------------------------------------------------------------

Memory usage is given in bytes:

Object Type          Creations   Destructions     Memory  Descendants' Mem.
Reports information only for process 0.

--- Event Stage 0: Main Stage

    Distributed Mesh     5              5       905272     0
              Vector    59             59      6793272     0
      Vector Scatter    22             22        23320     0
           Index Set    49             49       220252     0
   IS L to G Mapping    10             10       703172     0
              Viewer     4              3         2064     0
              Matrix    39             36     33641936     0
                  TS     1              1         1088     0
                SNES     1              1         1200     0
========================================================================================================================
Average time to get PetscTime(): 9.53674e-08
Average time for MPI_Barrier(): 3.57628e-06
Average time for zero size MPI_Send(): 5.24521e-06
#PETSc Option Table entries:
-Tf 1.
-cfl 2
-lints
-log_summary
-nx 256
-ny 128
-save 1.
#End of PETSc Option Table entries
Compiled with FORTRAN kernels
Compiled with full precision matrices (default)
sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8
Configure run at: Thu Feb 16 17:51:17 2012
Configure options: --with-mpi=yes --with-shared-libraries --with-scalar-type=real --with-fortran-interfaces=1 --FFLAGS=-I/usr/include --with-fortran --with-fortran-kernels=1 --with-clanguage=c COPTFLAGS=-O3 FOPTFLAGS=-O3 --download-mumps=MUMPS_4.10.0.tar.gz --download-scalapack=SCALAPACK-1.7.tar.gz --download-blacs=blacs-dev.tar.gz --download-parmetis=ParMetis-3.2.0-p1.tar.gz --download-superlu=superlu_4.2.tar.gz --download-superlu_dist=superlu_dist_2.5.tar.gz --download-spooles=spooles-2.2-dec-2008.tar.gz --download-umfpack=UMFPACK-5.5.1.tar.gz --with-debugging=0 --with-mpi-dir=/home/garnaud/local/openmpi-1.4.4 --download-hdf5 --download-f-blas-lapack
-----------------------------------------
Libraries compiled on Thu Feb 16 17:51:17 2012 on muzo.polytechnique.fr 
Machine characteristics: Linux-2.6.39.4-4.2-desktop-x86_64-with-mandrake-2011.0-Official
Using PETSc directory: /home/garnaud/local/petsc/petsc-3.2-p5
Using PETSc arch: real_opt
-----------------------------------------

Using C compiler: /home/garnaud/local/openmpi-1.4.4/bin/mpicc  -fPIC -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O3  ${COPTFLAGS} ${CFLAGS}
Using Fortran compiler: /home/garnaud/local/openmpi-1.4.4/bin/mpif90 -I/usr/include -fPIC -O3   ${FOPTFLAGS} ${FFLAGS} 
-----------------------------------------

Using include paths: -I/home/garnaud/local/petsc/petsc-3.2-p5/real_opt/include -I/home/garnaud/local/petsc/petsc-3.2-p5/include -I/home/garnaud/local/petsc/petsc-3.2-p5/include -I/home/garnaud/local/petsc/petsc-3.2-p5/real_opt/include -I/home/garnaud/local/openmpi-1.4.4/include
-----------------------------------------

Using C linker: /home/garnaud/local/openmpi-1.4.4/bin/mpicc
Using Fortran linker: /home/garnaud/local/openmpi-1.4.4/bin/mpif90
Using libraries: -Wl,-rpath,/home/garnaud/local/petsc/petsc-3.2-p5/real_opt/lib -L/home/garnaud/local/petsc/petsc-3.2-p5/real_opt/lib -lpetsc -lX11 -lpthread -Wl,-rpath,/home/garnaud/local/petsc/petsc-3.2-p5/real_opt/lib -L/home/garnaud/local/petsc/petsc-3.2-p5/real_opt/lib -lsuperlu_dist_2.5 -lcmumps -ldmumps -lsmumps -lzmumps -lmumps_common -lpord -lparmetis -lmetis -lspooles -lscalapack -lblacs -lsuperlu_4.2 -lumfpack -lamd -lflapack -lfblas -lhdf5_fortran -lhdf5 -lz -Wl,-rpath,/home/garnaud/local/openmpi-1.4.4/lib -L/home/garnaud/local/openmpi-1.4.4/lib -Wl,-rpath,/usr/lib64/gcc/x86_64-mandriva-linux-gnu/4.6.1 -L/usr/lib64/gcc/x86_64-mandriva-linux-gnu/4.6.1 -ldl -lmpi -lopen-rte -lopen-pal -lnsl -lutil -lgcc_s -lpthread -lmpi_f90 -lmpi_f77 -lgfortran -lm -lgfortran -lm -lgfortran -lm -lm -lquadmath -lm -ldl -lmpi -lopen-rte -lopen-pal -lnsl -lutil -lgcc_s -lpthread -ldl 
-----------------------------------------
-------------- next part --------------
  nx  = 256
  ny  = 128

  JET MESH 
     ix     = 74
     iy_in  = 44
     iy_out = 65
     rank: 0, i0-i1xj0-j1 : 0 - 74 x 44 - 63
     rank: 2, i0-i1xj0-j1 : 0 - 74 x 64 - 65

  TIME STEPPING
     Tf  = 1 
     CFL = 2 
    dt = 0.00571429 
    |q(1.000000)|/|q(0)| = 1.0005 
Elapsed time (CPU) = 19.2814 s 
Final state saved in q0.h5 
************************************************************************************************************************
***             WIDEN YOUR WINDOW TO 120 CHARACTERS.  Use 'enscript -r -fCourier9' to print this document            ***
************************************************************************************************************************

---------------------------------------------- PETSc Performance Summary: ----------------------------------------------

./ns2d on a real_opt named muzo.polytechnique.fr with 4 processors, by garnaud Sat Mar 10 18:03:09 2012
Using Petsc Release Version 3.2.0, Patch 5, Sat Oct 29 13:45:54 CDT 2011 

                         Max       Max/Min        Avg      Total 
Time (sec):           1.955e+01      1.00000   1.955e+01
Objects:              8.400e+01      1.00000   8.400e+01
Flops:                1.090e+10      1.00000   1.090e+10  4.358e+10
Flops/sec:            5.574e+08      1.00000   5.574e+08  2.229e+09
MPI Messages:         3.259e+04      1.00000   3.259e+04  1.303e+05
MPI Message Lengths:  2.276e+08      1.00000   6.984e+03  9.103e+08
MPI Reductions:       1.180e+02      1.00000

Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract)
                            e.g., VecAXPY() for real vectors of length N --> 2N flops
                            and VecAXPY() for complex vectors of length N --> 8N flops

Summary of Stages:   ----- Time ------  ----- Flops -----  --- Messages ---  -- Message Lengths --  -- Reductions --
                        Avg     %Total     Avg     %Total   counts   %Total     Avg         %Total   counts   %Total 
 0:      Main Stage: 1.9549e+01 100.0%  4.3584e+10 100.0%  1.303e+05 100.0%  6.984e+03      100.0%  1.170e+02  99.2% 

------------------------------------------------------------------------------------------------------------------------
See the 'Profiling' chapter of the users' manual for details on interpreting output.
Phase summary info:
   Count: number of times phase was executed
   Time and Flops: Max - maximum over all processors
                   Ratio - ratio of maximum to minimum over all processors
   Mess: number of messages sent
   Avg. len: average message length
   Reduct: number of global reductions
   Global: entire computation
   Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop().
      %T - percent time in this phase         %f - percent flops in this phase
      %M - percent messages in this phase     %L - percent message lengths in this phase
      %R - percent reductions in this phase
   Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors)
------------------------------------------------------------------------------------------------------------------------
Event                Count      Time (sec)     Flops                             --- Global ---  --- Stage ---   Total
                   Max Ratio  Max     Ratio   Max  Ratio  Mess   Avg len Reduct  %T %f %M %L %R  %T %f %M %L %R Mflop/s
------------------------------------------------------------------------------------------------------------------------

--- Event Stage 0: Main Stage

DERIVATIVES        10502 1.0 1.4136e+01 1.0 8.20e+09 1.0 0.0e+00 0.0e+00 0.0e+00 72 75  0  0  0  72 75  0  0  0  2321
FILTERS              350 1.0 1.9755e-01 1.0 1.26e+08 1.0 0.0e+00 0.0e+00 0.0e+00  1  1  0  0  0   1  1  0  0  0  2554
VECMANIP           21704 1.0 2.4231e+00 1.2 0.00e+00 0.0 1.3e+05 7.0e+03 6.0e+00 11  0100100  5  11  0100100  5     0
VecView                7 1.0 4.2857e-01 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  2  0  0  0  0   2  0  0  0  0     0
VecNorm                2 1.0 6.0606e-04 3.7 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00  0  0  0  0  2   0  0  0  0  2     0
VecScale            1750 1.0 6.4685e-02 1.1 5.73e+07 1.0 0.0e+00 0.0e+00 0.0e+00  0  1  0  0  0   0  1  0  0  0  3546
VecCopy              352 1.0 3.0717e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
VecAXPY             8750 1.0 8.0684e-01 1.1 6.08e+08 1.0 0.0e+00 0.0e+00 0.0e+00  4  6  0  0  0   4  6  0  0  0  3013
VecAXPBYCZ           350 1.0 6.5956e-02 1.0 4.01e+07 1.0 0.0e+00 0.0e+00 0.0e+00  0  0  0  0  0   0  0  0  0  0  2434
VecScatterBegin    10852 1.0 1.4070e+00 1.1 0.00e+00 0.0 1.3e+05 7.0e+03 0.0e+00  7  0100100  0   7  0100100  0     0
VecScatterEnd      10852 1.0 7.2064e-01 1.5 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  3  0  0  0  0   3  0  0  0  0     0
IMPOSEBC_VISC       5250 1.0 7.5293e-03 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
IMPOSEBC_EULER      5425 1.0 5.2320e-02 2.2 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00  0  0  0  0  0   0  0  0  0  0     0
FLUXES_VISC         3500 1.0 6.5287e-01 1.1 6.88e+08 1.0 0.0e+00 0.0e+00 0.0e+00  3  6  0  0  0   3  6  0  0  0  4216
FLUXES_EULER        3500 1.0 4.5190e-01 1.1 6.88e+08 1.0 0.0e+00 0.0e+00 0.0e+00  2  6  0  0  0   2  6  0  0  0  6091
STRESSES            3500 1.0 4.7757e-01 1.1 4.87e+08 1.0 0.0e+00 0.0e+00 0.0e+00  2  4  0  0  0   2  4  0  0  0  4083
TSStep               175 1.0 1.8815e+01 1.0 1.08e+10 1.0 1.3e+05 7.0e+03 1.0e+01 96 99 97 97  8  96 99 97 97  9  2289
TSFunctionEval      1750 1.0 1.8553e+01 1.0 1.06e+10 1.0 1.3e+05 7.0e+03 4.0e+00 95 97 97 97  3  95 97 97 97  3  2287
------------------------------------------------------------------------------------------------------------------------

Memory usage is given in bytes:

Object Type          Creations   Destructions     Memory  Descendants' Mem.
Reports information only for process 0.

--- Event Stage 0: Main Stage

    Distributed Mesh     5              5       905272     0
              Vector    28             28      4715800     0
      Vector Scatter    10             10        10600     0
           Index Set    25             25       202300     0
   IS L to G Mapping    10             10       703172     0
              Viewer     4              3         2064     0
                  TS     1              1         1088     0
                SNES     1              1         1200     0
========================================================================================================================
Average time to get PetscTime(): 2.14577e-07
Average time for MPI_Barrier(): 3.57628e-06
Average time for zero size MPI_Send(): 5.00679e-06
#PETSc Option Table entries:
-Tf 1.
-cfl 2
-log_summary
-nx 256
-ny 128
-save 1.
-ts
#End of PETSc Option Table entries
Compiled with FORTRAN kernels
Compiled with full precision matrices (default)
sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8
Configure run at: Thu Feb 16 17:51:17 2012
Configure options: --with-mpi=yes --with-shared-libraries --with-scalar-type=real --with-fortran-interfaces=1 --FFLAGS=-I/usr/include --with-fortran --with-fortran-kernels=1 --with-clanguage=c COPTFLAGS=-O3 FOPTFLAGS=-O3 --download-mumps=MUMPS_4.10.0.tar.gz --download-scalapack=SCALAPACK-1.7.tar.gz --download-blacs=blacs-dev.tar.gz --download-parmetis=ParMetis-3.2.0-p1.tar.gz --download-superlu=superlu_4.2.tar.gz --download-superlu_dist=superlu_dist_2.5.tar.gz --download-spooles=spooles-2.2-dec-2008.tar.gz --download-umfpack=UMFPACK-5.5.1.tar.gz --with-debugging=0 --with-mpi-dir=/home/garnaud/local/openmpi-1.4.4 --download-hdf5 --download-f-blas-lapack
-----------------------------------------
Libraries compiled on Thu Feb 16 17:51:17 2012 on muzo.polytechnique.fr 
Machine characteristics: Linux-2.6.39.4-4.2-desktop-x86_64-with-mandrake-2011.0-Official
Using PETSc directory: /home/garnaud/local/petsc/petsc-3.2-p5
Using PETSc arch: real_opt
-----------------------------------------

Using C compiler: /home/garnaud/local/openmpi-1.4.4/bin/mpicc  -fPIC -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O3  ${COPTFLAGS} ${CFLAGS}
Using Fortran compiler: /home/garnaud/local/openmpi-1.4.4/bin/mpif90 -I/usr/include -fPIC -O3   ${FOPTFLAGS} ${FFLAGS} 
-----------------------------------------

Using include paths: -I/home/garnaud/local/petsc/petsc-3.2-p5/real_opt/include -I/home/garnaud/local/petsc/petsc-3.2-p5/include -I/home/garnaud/local/petsc/petsc-3.2-p5/include -I/home/garnaud/local/petsc/petsc-3.2-p5/real_opt/include -I/home/garnaud/local/openmpi-1.4.4/include
-----------------------------------------

Using C linker: /home/garnaud/local/openmpi-1.4.4/bin/mpicc
Using Fortran linker: /home/garnaud/local/openmpi-1.4.4/bin/mpif90
Using libraries: -Wl,-rpath,/home/garnaud/local/petsc/petsc-3.2-p5/real_opt/lib -L/home/garnaud/local/petsc/petsc-3.2-p5/real_opt/lib -lpetsc -lX11 -lpthread -Wl,-rpath,/home/garnaud/local/petsc/petsc-3.2-p5/real_opt/lib -L/home/garnaud/local/petsc/petsc-3.2-p5/real_opt/lib -lsuperlu_dist_2.5 -lcmumps -ldmumps -lsmumps -lzmumps -lmumps_common -lpord -lparmetis -lmetis -lspooles -lscalapack -lblacs -lsuperlu_4.2 -lumfpack -lamd -lflapack -lfblas -lhdf5_fortran -lhdf5 -lz -Wl,-rpath,/home/garnaud/local/openmpi-1.4.4/lib -L/home/garnaud/local/openmpi-1.4.4/lib -Wl,-rpath,/usr/lib64/gcc/x86_64-mandriva-linux-gnu/4.6.1 -L/usr/lib64/gcc/x86_64-mandriva-linux-gnu/4.6.1 -ldl -lmpi -lopen-rte -lopen-pal -lnsl -lutil -lgcc_s -lpthread -lmpi_f90 -lmpi_f77 -lgfortran -lm -lgfortran -lm -lgfortran -lm -lm -lquadmath -lm -ldl -lmpi -lopen-rte -lopen-pal -lnsl -lutil -lgcc_s -lpthread -ldl 
-----------------------------------------


More information about the petsc-users mailing list