DMMG Object with 5 levels Using global (nonghosted) Jacobian coloring computation Processor [0] M 5 N 5 P 5 m 1 n 2 p 2 w 1 s 1 X range of indices: 0 5, Y range of indices: 0 3, Z range of indices: 0 3 Processor [1] M 5 N 5 P 5 m 1 n 2 p 2 w 1 s 1 X range of indices: 0 5, Y range of indices: 3 5, Z range of indices: 0 3 Processor [2] M 5 N 5 P 5 m 1 n 2 p 2 w 1 s 1 X range of indices: 0 5, Y range of indices: 0 3, Z range of indices: 3 5 Processor [3] M 5 N 5 P 5 m 1 n 2 p 2 w 1 s 1 X range of indices: 0 5, Y range of indices: 3 5, Z range of indices: 3 5 Processor [0] M 9 N 9 P 9 m 1 n 2 p 2 w 1 s 1 X range of indices: 0 9, Y range of indices: 0 5, Z range of indices: 0 5 Processor [1] M 9 N 9 P 9 m 1 n 2 p 2 w 1 s 1 X range of indices: 0 9, Y range of indices: 5 9, Z range of indices: 0 5 Processor [2] M 9 N 9 P 9 m 1 n 2 p 2 w 1 s 1 X range of indices: 0 9, Y range of indices: 0 5, Z range of indices: 5 9 Processor [3] M 9 N 9 P 9 m 1 n 2 p 2 w 1 s 1 X range of indices: 0 9, Y range of indices: 5 9, Z range of indices: 5 9 Processor [0] M 17 N 17 P 17 m 1 n 2 p 2 w 1 s 1 X range of indices: 0 17, Y range of indices: 0 9, Z range of indices: 0 9 Processor [1] M 17 N 17 P 17 m 1 n 2 p 2 w 1 s 1 X range of indices: 0 17, Y range of indices: 9 17, Z range of indices: 0 9 Processor [2] M 17 N 17 P 17 m 1 n 2 p 2 w 1 s 1 X range of indices: 0 17, Y range of indices: 0 9, Z range of indices: 9 17 Processor [3] M 17 N 17 P 17 m 1 n 2 p 2 w 1 s 1 X range of indices: 0 17, Y range of indices: 9 17, Z range of indices: 9 17 Processor [0] M 33 N 33 P 33 m 1 n 2 p 2 w 1 s 1 X range of indices: 0 33, Y range of indices: 0 17, Z range of indices: 0 17 Processor [1] M 33 N 33 P 33 m 1 n 2 p 2 w 1 s 1 X range of indices: 0 33, Y range of indices: 17 33, Z range of indices: 0 17 Processor [2] M 33 N 33 P 33 m 1 n 2 p 2 w 1 s 1 X range of indices: 0 33, Y range of indices: 0 17, Z range of indices: 17 33 Processor [3] M 33 N 33 P 33 m 1 n 2 p 2 w 1 s 1 X range of indices: 0 33, Y range of indices: 17 33, Z range of indices: 17 33 Processor [0] M 65 N 65 P 65 m 1 n 2 p 2 w 1 s 1 X range of indices: 0 65, Y range of indices: 0 33, Z range of indices: 0 33 Processor [1] M 65 N 65 P 65 m 1 n 2 p 2 w 1 s 1 X range of indices: 0 65, Y range of indices: 33 65, Z range of indices: 0 33 Processor [2] M 65 N 65 P 65 m 1 n 2 p 2 w 1 s 1 X range of indices: 0 65, Y range of indices: 0 33, Z range of indices: 33 65 Processor [3] M 65 N 65 P 65 m 1 n 2 p 2 w 1 s 1 X range of indices: 0 65, Y range of indices: 33 65, Z range of indices: 33 65 Using matrix type aij KSP Object: type: fgmres GMRES: restart=30, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement GMRES: happy breakdown tolerance 1e-30 maximum iterations=10000, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000 right preconditioning using UNPRECONDITIONED norm type for convergence test PC Object: type: jacobi linear system matrix = precond matrix: Matrix Object: type=mpiaij, rows=274625, cols=274625 total: nonzeros=1897025, allocated nonzeros=1897025 not using I-node (on process 0) routines Number of Newton iterations = 7 Number of Linear iterations = 3770 Average Linear its / Newton = 5.385714e+02 ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./ex20 on a linux-gnu named ubuntu with 4 processors, by vijay Thu Feb 3 12:04:36 2011 Using Petsc Release Version 3.1.0, Patch 7, Mon Dec 20 14:26:37 CST 2010 Max Max/Min Avg Total Time (sec): 8.803e+01 1.00000 8.803e+01 Objects: 2.730e+02 1.00000 2.730e+02 Flops: 4.318e+10 1.06352 4.188e+10 1.675e+11 Flops/sec: 4.906e+08 1.06352 4.758e+08 1.903e+09 MPI Messages: 1.594e+04 1.00378 1.591e+04 6.363e+04 MPI Message Lengths: 2.690e+08 1.03162 1.665e+04 1.060e+09 MPI Reductions: 1.587e+04 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 1.4857e-01 0.2% 2.2441e+06 0.0% 2.230e+02 0.4% 3.449e+00 0.0% 1.590e+02 1.0% 1: Solve: 4.3284e+01 49.2% 8.3768e+10 50.0% 3.174e+04 49.9% 8.326e+03 50.0% 7.803e+03 49.2% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %F - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %F %M %L %R %T %F %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage PetscBarrier 2 1.0 6.5994e-0410.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 12 1.0 1.4329e-04 1.9 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecScatterBegin 4 1.0 8.6069e-05 9.5 0.00e+00 0.0 2.0e+01 1.2e+03 0.0e+00 0 0 0 0 0 0 0 9 11 0 0 VecScatterEnd 4 1.0 9.8228e-0513.7 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatMultTranspose 4 1.0 2.1212e-03 1.3 5.37e+05 1.0 2.0e+01 1.2e+03 0.0e+00 0 0 0 0 0 1 93 9 11 0 988 MatAssemblyBegin 9 1.0 2.3475e-0222.6 0.00e+00 0.0 0.0e+00 0.0e+00 1.8e+01 0 0 0 0 0 8 0 0 0 11 0 MatAssemblyEnd 9 1.0 1.5119e-02 1.1 0.00e+00 0.0 1.2e+02 8.6e+02 6.3e+01 0 0 0 0 0 10 0 54 47 40 0 --- Event Stage 1: Solve VecDot 7 1.0 4.1199e-03 1.9 9.91e+05 1.1 0.0e+00 0.0e+00 7.0e+00 0 0 0 0 0 0 0 0 0 0 933 VecMDot 3770 1.0 1.2646e+01 1.3 8.19e+09 1.1 0.0e+00 0.0e+00 3.8e+03 13 19 0 0 24 26 38 0 0 48 2514 VecNorm 3912 1.0 5.6999e+00 1.6 5.54e+08 1.1 0.0e+00 0.0e+00 3.9e+03 6 1 0 0 25 11 3 0 0 50 377 VecScale 3898 1.0 5.2526e-01 1.3 2.76e+08 1.1 0.0e+00 0.0e+00 0.0e+00 1 1 0 0 0 1 1 0 0 0 2038 VecCopy 28 1.0 7.2310e-03 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 198 1.0 2.6998e-02 1.4 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 128 1.0 2.9975e-02 1.3 1.81e+07 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 2345 VecWAXPY 128 1.0 3.7670e-02 1.2 9.06e+06 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 933 VecMAXPY 3898 1.0 1.3921e+01 1.2 8.73e+09 1.1 0.0e+00 0.0e+00 0.0e+00 14 20 0 0 0 29 40 0 0 0 2433 VecPointwiseMult 3798 1.0 1.7179e+00 1.4 2.67e+08 1.1 0.0e+00 0.0e+00 0.0e+00 2 1 0 0 0 3 1 0 0 0 603 VecScatterBegin 3969 1.0 2.8862e-01 1.3 0.00e+00 0.0 3.2e+04 1.7e+04 0.0e+00 0 0 50 50 0 1 0100100 0 0 VecScatterEnd 3969 1.0 4.1759e-01 1.5 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 1 0 0 0 0 0 VecReduceArith 2 1.0 4.0770e-04 1.6 2.83e+05 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 2694 VecReduceComm 1 1.0 1.1494e-02359.8 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatMult 3898 1.0 1.2538e+01 1.2 3.54e+09 1.1 3.1e+04 1.7e+04 0.0e+00 12 8 49 50 0 25 16 98 99 0 1094 MatMultTranspose 28 1.0 1.4955e-02 1.1 3.76e+06 1.0 1.4e+02 1.2e+03 0.0e+00 0 0 0 0 0 0 0 0 0 0 981 MatAssemblyBegin 35 1.0 1.4884e-01 3.5 0.00e+00 0.0 0.0e+00 0.0e+00 7.0e+01 0 0 0 0 0 0 0 0 0 1 0 MatAssemblyEnd 35 1.0 1.9204e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 3.5e+01 0 0 0 0 0 0 0 0 0 0 0 MatView 1 1.0 4.0760e-03 1.3 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 0 0 0 0 0 0 0 0 0 0 0 SNESSolve 1 1.0 4.3279e+01 1.0 2.16e+10 1.1 3.2e+04 1.7e+04 7.8e+03 49 50 50 50 49 100100100100100 1936 SNESLineSearch 7 1.0 3.7013e-01 1.0 1.09e+07 1.1 1.1e+02 1.7e+04 2.8e+01 0 0 0 0 0 1 0 0 0 0 115 SNESFunctionEval 8 1.0 3.4312e-01 1.1 7.21e+05 1.0 6.4e+01 1.7e+04 0.0e+00 0 0 0 0 0 1 0 0 0 0 8 SNESJacobianEval 7 1.0 9.5757e-01 1.0 5.49e+06 1.0 4.2e+02 3.4e+03 1.0e+02 1 0 1 0 1 2 0 1 0 1 22 KSPGMRESOrthog 3770 1.0 2.3756e+01 1.1 1.64e+10 1.1 0.0e+00 0.0e+00 3.8e+03 26 38 0 0 24 54 76 0 0 48 2677 KSPSetup 7 1.0 1.9073e-06 0.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 7 1.0 4.1912e+01 1.0 2.16e+10 1.1 3.1e+04 1.7e+04 7.7e+03 48 50 49 50 48 97100 98 99 98 1997 PCSetUp 7 1.0 2.3835e-02 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 PCApply 3770 1.0 1.7224e+00 1.4 2.67e+08 1.1 0.0e+00 0.0e+00 0.0e+00 2 1 0 0 0 3 1 0 0 0 601 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Distributed array 5 5 374504 0 Vec 51 133 45067760 0 Vec Scatter 19 19 16492 0 Index Set 38 38 1073524 0 IS L to G Mapping 5 5 353764 0 Matrix 27 27 15326904 0 SNES 5 5 5160 0 Krylov Solver 20 20 35240 0 Preconditioner 20 20 14000 0 Viewer 0 1 544 0 --- Event Stage 1: Solve Vec 82 0 0 0 Viewer 1 0 0 0 ======================================================================================================================== Average time to get PetscTime(): 0 Average time for MPI_Barrier(): 2.24113e-05 Average time for zero size MPI_Send(): 8.76188e-06 #PETSc Option Table entries: -dmmg_nlevels 5 -dmmg_view -log_summary -pc_type jacobi #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 Configure run at: Wed Feb 2 17:09:40 2011 Configure options: --download-f-blas-lapack=1 --with-mpi-dir=/usr/lib/ --with-mpi-shared=0 --with-shared=0 --download-parmetis=1 --download-superlu_dist=1 --download-hypre=1 --download-blacs=1 --download-scalapack=1 --with-clanguage=C++ --download-plapack=1 --download-mumps=1 --download-umfpack=yes --with-debugging=0 --with-errorchecking=yes --download-mpich=1 ----------------------------------------- Libraries compiled on Wed Feb 2 17:16:46 CST 2011 on ubuntu Machine characteristics: Linux ubuntu 2.6.32-24-generic #39-Ubuntu SMP Wed Jul 28 05:14:15 UTC 2010 x86_64 GNU/Linux Using PETSc directory: /home/vijay/karma/contrib/petsc Using PETSc arch: linux-gnu-cxx-opt ----------------------------------------- Using C compiler: /home/vijay/karma/contrib/petsc/linux-gnu-cxx-opt/bin/mpicxx -Wall -Wwrite-strings -Wno-strict-aliasing -O Using Fortran compiler: /home/vijay/karma/contrib/petsc/linux-gnu-cxx-opt/bin/mpif90 -Wall -Wno-unused-variable -O ----------------------------------------- Using include paths: -I/home/vijay/karma/contrib/petsc/linux-gnu-cxx-opt/include -I/home/vijay/karma/contrib/petsc/include -I/home/vijay/karma/contrib/petsc/linux-gnu-cxx-opt/include ------------------------------------------ Using C linker: /home/vijay/karma/contrib/petsc/linux-gnu-cxx-opt/bin/mpicxx -Wall -Wwrite-strings -Wno-strict-aliasing -O Using Fortran linker: /home/vijay/karma/contrib/petsc/linux-gnu-cxx-opt/bin/mpif90 -Wall -Wno-unused-variable -O Using libraries: -Wl,-rpath,/home/vijay/karma/contrib/petsc/linux-gnu-cxx-opt/lib -L/home/vijay/karma/contrib/petsc/linux-gnu-cxx-opt/lib -lpetsc -Wl,-rpath,/home/vijay/karma/contrib/petsc/linux-gnu-cxx-opt/lib -L/home/vijay/karma/contrib/petsc/linux-gnu-cxx-opt/lib -lcmumps -ldmumps -lsmumps -lzmumps -lmumps_common -lpord -lscalapack -lblacs -lPLAPACK -lsuperlu_dist_2.4 -lparmetis -lmetis -lHYPRE -lumfpack -lamd -lflapack -lfblas -lnsl -lrt -L/home/vijay/karma/contrib/petsc/linux-gnu-cxx-opt/lib -L/usr/lib/gcc/x86_64-linux-gnu/4.4.3 -L/usr/lib/x86_64-linux-gnu -ldl -lmpich -lpthread -lrt -lgcc_s -lmpichf90 -lgfortran -lm -lm -lmpichcxx -lstdc++ -lmpichcxx -lstdc++ -ldl -lmpich -lpthread -lrt -lgcc_s -ldl ------------------------------------------