[petsc-users] Drastic increase in memory usage between DMDA_STENCIL_BOX and DMDA_STENCIL_STAR
Mani Chandra
mc0710 at gmail.com
Tue Mar 18 20:04:38 CDT 2014
Hi,
I see a 4x increase in the memory usage when I change from
DMDA_STENCIL_STAR to DMDA_STENCIL_BOX. Attached are the outputs of
-log_summary which shows a huge increase in the matrix memory usage. Is
this expected?
On a different note, suppose I am running a serial calculation with no need
to exchange data but I am using corner node information, do I need to use
DMDA_STENCIL_BOX? Would the jacobian when computed using colored finite
differences be correctly represented if I use corner information but still
use DMDA_STENCIL_STAR?
Thanks,
Mani
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20140318/aa720342/attachment-0001.html>
-------------- next part --------------
---------------------------------------------- PETSc Performance Summary: ----------------------------------------------
./grim on a arch-linux2-c-debug named Deathstar with 1 processor, by mc Tue Mar 18 19:56:44 2014
Using Petsc Development GIT revision: v3.4.3-3262-g255453a GIT Date: 2014-02-08 22:41:14 -0600
Max Max/Min Avg Total
Time (sec): 9.532e+01 1.00000 9.532e+01
Objects: 2.650e+02 1.00000 2.650e+02
Flops: 2.300e+10 1.00000 2.300e+10 2.300e+10
Flops/sec: 2.413e+08 1.00000 2.413e+08 2.413e+08
MPI Messages: 0.000e+00 0.00000 0.000e+00 0.000e+00
MPI Message Lengths: 0.000e+00 0.00000 0.000e+00 0.000e+00
MPI Reductions: 2.600e+02 1.00000
Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract)
e.g., VecAXPY() for real vectors of length N --> 2N flops
and VecAXPY() for complex vectors of length N --> 8N flops
Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions --
Avg %Total Avg %Total counts %Total Avg %Total counts %Total
0: Main Stage: 9.5319e+01 100.0% 2.2998e+10 100.0% 0.000e+00 0.0% 0.000e+00 0.0% 2.590e+02 99.6%
------------------------------------------------------------------------------------------------------------------------
See the 'Profiling' chapter of the users' manual for details on interpreting output.
Phase summary info:
Count: number of times phase was executed
Time and Flops: Max - maximum over all processors
Ratio - ratio of maximum to minimum over all processors
Mess: number of messages sent
Avg. len: average message length (bytes)
Reduct: number of global reductions
Global: entire computation
Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop().
%T - percent time in this phase %F - percent flops in this phase
%M - percent messages in this phase %L - percent message lengths in this phase
%R - percent reductions in this phase
Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors)
------------------------------------------------------------------------------------------------------------------------
Event Count Time (sec) Flops --- Global --- --- Stage --- Total
Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %F %M %L %R %T %F %M %L %R Mflop/s
------------------------------------------------------------------------------------------------------------------------
--- Event Stage 0: Main Stage
VecView 3 1.0 4.7265e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
VecDot 11 1.0 1.0761e-02 1.0 1.15e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 1072
VecMDot 22 1.0 2.7508e-02 1.0 3.46e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 1258
VecNorm 58 1.0 4.4957e-02 1.0 6.08e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 1353
VecScale 33 1.0 1.4479e-02 1.0 1.73e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 1195
VecCopy 236 1.0 2.5637e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
VecSet 41 1.0 4.4386e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
VecAXPY 212 1.0 2.0891e-01 1.0 2.22e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 1064
VecAXPBYCZ 214 1.0 3.3057e-01 1.0 3.37e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 1018
VecWAXPY 12 1.0 1.8075e-02 1.0 6.82e+06 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 377
VecMAXPY 33 1.0 4.7938e-02 1.0 5.77e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 1203
VecLoad 1 1.0 1.6348e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
VecReduceArith 22 1.0 1.6276e-02 1.0 2.31e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 1417
VecReduceComm 11 1.0 1.1206e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
VecNormalize 33 1.0 3.8849e-02 1.0 5.19e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 1336
MatMult 33 1.0 6.2562e+00 1.0 6.84e+09 1.0 0.0e+00 0.0e+00 0.0e+00 7 30 0 0 0 7 30 0 0 0 1093
MatSolve 33 1.0 6.2317e+00 1.0 6.84e+09 1.0 0.0e+00 0.0e+00 0.0e+00 7 30 0 0 0 7 30 0 0 0 1097
MatLUFactorNum 1 1.0 3.3280e+01 1.0 8.55e+09 1.0 0.0e+00 0.0e+00 0.0e+00 35 37 0 0 0 35 37 0 0 0 257
MatILUFactorSym 1 1.0 5.1644e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 1 0 0 0 0 1 0 0 0 0 0
MatAssemblyBegin 2 1.0 2.1458e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatAssemblyEnd 2 1.0 2.8725e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatGetRowIJ 1 1.0 1.9073e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatGetOrdering 1 1.0 1.1145e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 4.0e+00 0 0 0 0 2 0 0 0 0 2 0
MatZeroEntries 1 1.0 1.2959e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatFDColorCreate 1 1.0 2.9290e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 3.0e+00 0 0 0 0 1 0 0 0 0 1 0
MatFDColorSetUp 1 1.0 5.2455e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+02 6 0 0 0 79 6 0 0 0 79 0
MatFDColorApply 1 1.0 1.4934e+01 1.0 5.25e+08 1.0 0.0e+00 0.0e+00 2.0e+00 16 2 0 0 1 16 2 0 0 1 35
MatFDColorFunc 200 1.0 1.3355e+01 1.0 3.15e+08 1.0 0.0e+00 0.0e+00 0.0e+00 14 1 0 0 0 14 1 0 0 0 24
TSStep 1 1.0 8.9164e+01 1.0 2.30e+10 1.0 0.0e+00 0.0e+00 2.4e+02 94100 0 0 92 94100 0 0 93 258
TSFunctionEval 213 1.0 1.3918e+01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 15 0 0 0 0 15 0 0 0 0 0
SNESSolve 1 1.0 6.7756e+01 1.0 2.30e+10 1.0 0.0e+00 0.0e+00 2.3e+02 71100 0 0 88 71100 0 0 89 339
SNESFunctionEval 13 1.0 8.9281e-01 1.0 2.04e+07 1.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 23
SNESJacobianEval 1 1.0 2.0316e+01 1.0 5.25e+08 1.0 0.0e+00 0.0e+00 2.1e+02 21 2 0 0 81 21 2 0 0 81 26
SNESLineSearch 11 1.0 3.0041e+00 1.0 2.36e+09 1.0 0.0e+00 0.0e+00 0.0e+00 3 10 0 0 0 3 10 0 0 0 787
KSPGMRESOrthog 22 1.0 5.4503e-02 1.0 6.92e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 1270
KSPSetUp 11 1.0 7.4494e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+01 0 0 0 0 4 0 0 0 0 4 0
KSPSolve 11 1.0 4.4361e+01 1.0 2.01e+10 1.0 0.0e+00 0.0e+00 1.5e+01 47 87 0 0 6 47 87 0 0 6 453
PCSetUp 1 1.0 3.3808e+01 1.0 8.55e+09 1.0 0.0e+00 0.0e+00 5.0e+00 35 37 0 0 2 35 37 0 0 2 253
PCApply 33 1.0 6.2317e+00 1.0 6.84e+09 1.0 0.0e+00 0.0e+00 0.0e+00 7 30 0 0 0 7 30 0 0 0 1097
------------------------------------------------------------------------------------------------------------------------
Memory usage is given in bytes:
Object Type Creations Destructions Memory Descendants' Mem.
Reports information only for process 0.
--- Event Stage 0: Main Stage
Vector 23 23 83920248 0
Vector Scatter 3 3 1956 0
Matrix 2 2 2509827340 0
Matrix FD Coloring 1 1 2289081428 0
Distributed Mesh 2 2 279712 0
Bipartite Graph 4 4 3264 0
Index Set 212 212 7242384 0
IS L to G Mapping 3 3 1788 0
TSAdapt 2 2 2384 0
TS 1 1 1240 0
DMTS 1 1 720 0
SNES 1 1 1348 0
SNESLineSearch 1 1 880 0
DMSNES 1 1 680 0
Krylov Solver 1 1 18376 0
DMKSP interface 1 1 664 0
Preconditioner 1 1 992 0
Viewer 5 4 2848 0
========================================================================================================================
Average time to get PetscTime(): 1.90735e-07
#PETSc Option Table entries:
-log_summary
-snes_atol 1e-5
-snes_converged_reason
-snes_lag_jacobian 100
-snes_lag_jacobian_persists TRUE
-snes_max_it 101
-snes_monitor
-snes_rtol 1e-50
-snes_stol 1e-50
-ts_dt 0.03
-ts_final_time 500.
-ts_max_snes_failures -1
-ts_max_steps 1
-ts_monitor
-ts_type theta
#End of PETSc Option Table entries
Compiled without FORTRAN kernels
Compiled with full precision matrices (default)
sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4
Configure run at: Thu Feb 20 16:27:19 2014
Configure options: --prefix=/home/mc/Downloads/petsc_optimized/ --with-debugging=0 COPTFLAGS="-O3 -march=native" FOPTFLAGS="-O3 -qarch=native" --with-clean=1 --with-hdf5=1 --download-hdf5=yes
-----------------------------------------
Libraries compiled on Thu Feb 20 16:27:19 2014 on Deathstar
Machine characteristics: Linux-3.12.9-2-ARCH-x86_64-with-glibc2.2.5
Using PETSc directory: /home/mc/Downloads/petsc
Using PETSc arch: arch-linux2-c-debug
-----------------------------------------
Using C compiler: mpicc -fPIC -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O3 -march=native ${COPTFLAGS} ${CFLAGS}
Using Fortran compiler: mpif90 -fPIC -Wall -Wno-unused-variable -Wno-unused-dummy-argument ${FOPTFLAGS} ${FFLAGS}
-----------------------------------------
Using include paths: -I/home/mc/Downloads/petsc/arch-linux2-c-debug/include -I/home/mc/Downloads/petsc/include -I/home/mc/Downloads/petsc/include -I/home/mc/Downloads/petsc/arch-linux2-c-debug/include
-----------------------------------------
Using C linker: mpicc
Using Fortran linker: mpif90
Using libraries: -Wl,-rpath,/home/mc/Downloads/petsc/arch-linux2-c-debug/lib -L/home/mc/Downloads/petsc/arch-linux2-c-debug/lib -lpetsc -llapack -lblas -lX11 -lpthread -Wl,-rpath,/home/mc/Downloads/petsc/arch-linux2-c-debug/lib -L/home/mc/Downloads/petsc/arch-linux2-c-debug/lib -lhdf5_fortran -lhdf5_hl -lhdf5 -lm -Wl,-rpath,/usr/lib/openmpi -L/usr/lib/openmpi -Wl,-rpath,/usr/lib/gcc/x86_64-unknown-linux-gnu/4.8.2 -L/usr/lib/gcc/x86_64-unknown-linux-gnu/4.8.2 -Wl,-rpath,/opt/intel/composerxe/compiler/lib/intel64 -L/opt/intel/composerxe/compiler/lib/intel64 -Wl,-rpath,/opt/intel/composerxe/ipp/lib/intel64 -L/opt/intel/composerxe/ipp/lib/intel64 -Wl,-rpath,/opt/intel/composerxe/mkl/lib/intel64 -L/opt/intel/composerxe/mkl/lib/intel64 -Wl,-rpath,/opt/intel/composerxe/tbb/lib/intel64/gcc4.4 -L/opt/intel/composerxe/tbb/lib/intel64/gcc4.4 -lmpi_f90 -lmpi_f77 -lgfortran -lm -lgfortran -lm -lquadmath -lm -lmpi_cxx -lstdc++ -ldl -lmpi -lhwloc -lgcc_s -lpthread -ldl
-----------------------------------------
-------------- next part --------------
---------------------------------------------- PETSc Performance Summary: ----------------------------------------------
./grim on a arch-linux2-c-debug named Deathstar with 1 processor, by mc Tue Mar 18 19:48:47 2014
Using Petsc Development GIT revision: v3.4.3-3262-g255453a GIT Date: 2014-02-08 22:41:14 -0600
Max Max/Min Avg Total
Time (sec): 4.053e+01 1.00000 4.053e+01
Objects: 2.750e+02 1.00000 2.750e+02
Flops: 9.192e+09 1.00000 9.192e+09 9.192e+09
Flops/sec: 2.268e+08 1.00000 2.268e+08 2.268e+08
MPI Messages: 0.000e+00 0.00000 0.000e+00 0.000e+00
MPI Message Lengths: 0.000e+00 0.00000 0.000e+00 0.000e+00
MPI Reductions: 2.800e+02 1.00000
Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract)
e.g., VecAXPY() for real vectors of length N --> 2N flops
and VecAXPY() for complex vectors of length N --> 8N flops
Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions --
Avg %Total Avg %Total counts %Total Avg %Total counts %Total
0: Main Stage: 4.0532e+01 100.0% 9.1921e+09 100.0% 0.000e+00 0.0% 0.000e+00 0.0% 2.790e+02 99.6%
------------------------------------------------------------------------------------------------------------------------
See the 'Profiling' chapter of the users' manual for details on interpreting output.
Phase summary info:
Count: number of times phase was executed
Time and Flops: Max - maximum over all processors
Ratio - ratio of maximum to minimum over all processors
Mess: number of messages sent
Avg. len: average message length (bytes)
Reduct: number of global reductions
Global: entire computation
Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop().
%T - percent time in this phase %F - percent flops in this phase
%M - percent messages in this phase %L - percent message lengths in this phase
%R - percent reductions in this phase
Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors)
------------------------------------------------------------------------------------------------------------------------
Event Count Time (sec) Flops --- Global --- --- Stage --- Total
Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %F %M %L %R %T %F %M %L %R Mflop/s
------------------------------------------------------------------------------------------------------------------------
--- Event Stage 0: Main Stage
VecView 3 1.0 4.3488e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 0
VecDot 11 1.0 1.0955e-02 1.0 1.15e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 1053
VecMDot 36 1.0 5.5024e-02 1.0 8.18e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 1486
VecNorm 72 1.0 5.6935e-02 1.0 7.55e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 1326
VecScale 47 1.0 2.1471e-02 1.0 2.46e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 1148
VecCopy 236 1.0 2.6282e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 0
VecSet 51 1.0 6.2247e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
VecAXPY 212 1.0 2.1605e-01 1.0 2.22e+08 1.0 0.0e+00 0.0e+00 0.0e+00 1 2 0 0 0 1 2 0 0 0 1029
VecAXPBYCZ 214 1.0 3.3499e-01 1.0 3.37e+08 1.0 0.0e+00 0.0e+00 0.0e+00 1 4 0 0 0 1 4 0 0 0 1005
VecWAXPY 12 1.0 1.8339e-02 1.0 6.82e+06 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 372
VecMAXPY 47 1.0 8.2657e-02 1.0 1.20e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 1446
VecLoad 1 1.0 1.6351e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
VecReduceArith 22 1.0 1.6958e-02 1.0 2.31e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 1360
VecReduceComm 11 1.0 1.2875e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
VecNormalize 47 1.0 5.7861e-02 1.0 7.39e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 1278
MatMult 47 1.0 3.3177e+00 1.0 3.51e+09 1.0 0.0e+00 0.0e+00 0.0e+00 8 38 0 0 0 8 38 0 0 0 1057
MatSolve 47 1.0 3.2621e+00 1.0 3.51e+09 1.0 0.0e+00 0.0e+00 0.0e+00 8 38 0 0 0 8 38 0 0 0 1075
MatLUFactorNum 1 1.0 6.2700e+00 1.0 1.28e+09 1.0 0.0e+00 0.0e+00 0.0e+00 15 14 0 0 0 15 14 0 0 0 204
MatILUFactorSym 1 1.0 2.2177e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 1 0 0 0 0 1 0 0 0 0 0
MatAssemblyBegin 2 1.0 9.5367e-07 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatAssemblyEnd 2 1.0 1.1208e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatGetRowIJ 1 1.0 1.9073e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatGetOrdering 1 1.0 1.0976e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 4.0e+00 0 0 0 0 1 0 0 0 0 1 0
MatZeroEntries 1 1.0 4.7051e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0
MatFDColorCreate 1 1.0 2.9862e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 3.0e+00 0 0 0 0 1 0 0 0 0 1 0
MatFDColorSetUp 1 1.0 2.0944e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+02 5 0 0 0 73 5 0 0 0 73 0
MatFDColorApply 1 1.0 1.4396e+01 1.0 5.25e+08 1.0 0.0e+00 0.0e+00 2.0e+00 36 6 0 0 1 36 6 0 0 1 36
MatFDColorFunc 200 1.0 1.3555e+01 1.0 3.15e+08 1.0 0.0e+00 0.0e+00 0.0e+00 33 3 0 0 0 33 3 0 0 0 23
TSStep 1 1.0 3.4390e+01 1.0 9.19e+09 1.0 0.0e+00 0.0e+00 2.6e+02 85100 0 0 93 85100 0 0 93 267
TSFunctionEval 213 1.0 1.4103e+01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 35 0 0 0 0 35 0 0 0 0 0
SNESSolve 1 1.0 3.0864e+01 1.0 9.19e+09 1.0 0.0e+00 0.0e+00 2.5e+02 76100 0 0 89 76100 0 0 90 298
SNESFunctionEval 13 1.0 8.8251e-01 1.0 2.04e+07 1.0 0.0e+00 0.0e+00 0.0e+00 2 0 0 0 0 2 0 0 0 0 23
SNESJacobianEval 1 1.0 1.6544e+01 1.0 5.25e+08 1.0 0.0e+00 0.0e+00 2.1e+02 41 6 0 0 75 41 6 0 0 76 32
SNESLineSearch 11 1.0 1.6754e+00 1.0 9.05e+08 1.0 0.0e+00 0.0e+00 0.0e+00 4 10 0 0 0 4 10 0 0 0 540
KSPGMRESOrthog 36 1.0 1.1178e-01 1.0 1.64e+08 1.0 0.0e+00 0.0e+00 0.0e+00 0 2 0 0 0 0 2 0 0 0 1463
KSPSetUp 11 1.0 7.5173e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+01 0 0 0 0 4 0 0 0 0 4 0
KSPSolve 11 1.0 1.2571e+01 1.0 7.76e+09 1.0 0.0e+00 0.0e+00 3.5e+01 31 84 0 0 12 31 84 0 0 13 617
PCSetUp 1 1.0 6.5028e+00 1.0 1.28e+09 1.0 0.0e+00 0.0e+00 5.0e+00 16 14 0 0 2 16 14 0 0 2 197
PCApply 47 1.0 3.2621e+00 1.0 3.51e+09 1.0 0.0e+00 0.0e+00 0.0e+00 8 38 0 0 0 8 38 0 0 0 1075
------------------------------------------------------------------------------------------------------------------------
Memory usage is given in bytes:
Object Type Creations Destructions Memory Descendants' Mem.
Reports information only for process 0.
--- Event Stage 0: Main Stage
Vector 33 33 125878088 0
Vector Scatter 3 3 1956 0
Matrix 2 2 918033676 0
Matrix FD Coloring 1 1 829426772 0
Distributed Mesh 2 2 279712 0
Bipartite Graph 4 4 3264 0
Index Set 212 212 7242384 0
IS L to G Mapping 3 3 1788 0
TSAdapt 2 2 2384 0
TS 1 1 1240 0
DMTS 1 1 720 0
SNES 1 1 1348 0
SNESLineSearch 1 1 880 0
DMSNES 1 1 680 0
Krylov Solver 1 1 18376 0
DMKSP interface 1 1 664 0
Preconditioner 1 1 992 0
Viewer 5 4 2848 0
========================================================================================================================
Average time to get PetscTime(): 9.53674e-08
#PETSc Option Table entries:
-log_summary
-snes_atol 1e-5
-snes_converged_reason
-snes_lag_jacobian 100
-snes_lag_jacobian_persists TRUE
-snes_max_it 101
-snes_monitor
-snes_rtol 1e-50
-snes_stol 1e-50
-ts_dt 0.03
-ts_final_time 500.
-ts_max_snes_failures -1
-ts_max_steps 1
-ts_monitor
-ts_type theta
#End of PETSc Option Table entries
Compiled without FORTRAN kernels
Compiled with full precision matrices (default)
sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4
Configure run at: Thu Feb 20 16:27:19 2014
Configure options: --prefix=/home/mc/Downloads/petsc_optimized/ --with-debugging=0 COPTFLAGS="-O3 -march=native" FOPTFLAGS="-O3 -qarch=native" --with-clean=1 --with-hdf5=1 --download-hdf5=yes
-----------------------------------------
Libraries compiled on Thu Feb 20 16:27:19 2014 on Deathstar
Machine characteristics: Linux-3.12.9-2-ARCH-x86_64-with-glibc2.2.5
Using PETSc directory: /home/mc/Downloads/petsc
Using PETSc arch: arch-linux2-c-debug
-----------------------------------------
Using C compiler: mpicc -fPIC -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O3 -march=native ${COPTFLAGS} ${CFLAGS}
Using Fortran compiler: mpif90 -fPIC -Wall -Wno-unused-variable -Wno-unused-dummy-argument ${FOPTFLAGS} ${FFLAGS}
-----------------------------------------
Using include paths: -I/home/mc/Downloads/petsc/arch-linux2-c-debug/include -I/home/mc/Downloads/petsc/include -I/home/mc/Downloads/petsc/include -I/home/mc/Downloads/petsc/arch-linux2-c-debug/include
-----------------------------------------
Using C linker: mpicc
Using Fortran linker: mpif90
Using libraries: -Wl,-rpath,/home/mc/Downloads/petsc/arch-linux2-c-debug/lib -L/home/mc/Downloads/petsc/arch-linux2-c-debug/lib -lpetsc -llapack -lblas -lX11 -lpthread -Wl,-rpath,/home/mc/Downloads/petsc/arch-linux2-c-debug/lib -L/home/mc/Downloads/petsc/arch-linux2-c-debug/lib -lhdf5_fortran -lhdf5_hl -lhdf5 -lm -Wl,-rpath,/usr/lib/openmpi -L/usr/lib/openmpi -Wl,-rpath,/usr/lib/gcc/x86_64-unknown-linux-gnu/4.8.2 -L/usr/lib/gcc/x86_64-unknown-linux-gnu/4.8.2 -Wl,-rpath,/opt/intel/composerxe/compiler/lib/intel64 -L/opt/intel/composerxe/compiler/lib/intel64 -Wl,-rpath,/opt/intel/composerxe/ipp/lib/intel64 -L/opt/intel/composerxe/ipp/lib/intel64 -Wl,-rpath,/opt/intel/composerxe/mkl/lib/intel64 -L/opt/intel/composerxe/mkl/lib/intel64 -Wl,-rpath,/opt/intel/composerxe/tbb/lib/intel64/gcc4.4 -L/opt/intel/composerxe/tbb/lib/intel64/gcc4.4 -lmpi_f90 -lmpi_f77 -lgfortran -lm -lgfortran -lm -lquadmath -lm -lmpi_cxx -lstdc++ -ldl -lmpi -lhwloc -lgcc_s -lpthread -ldl
-----------------------------------------
More information about the petsc-users
mailing list