---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./x.r on a arch-linux2-c-opt named node11.cocoa5 with 128 processors, by pzw2 Sun Nov 10 11:31:45 2013 Using Petsc Release Version 3.3.0, Patch 6, Mon Feb 11 12:26:34 CST 2013 Max Max/Min Avg Total Time (sec): 2.507e+02 1.00188 2.504e+02 Objects: 1.370e+02 1.00000 1.370e+02 Flops: 2.651e+10 1.05073 2.556e+10 3.272e+12 Flops/sec: 1.059e+08 1.05109 1.021e+08 1.307e+10 MPI Messages: 9.515e+03 1.96510 7.573e+03 9.693e+05 MPI Message Lengths: 1.821e+09 8.81374 4.244e+04 4.114e+10 MPI Reductions: 2.143e+03 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 2.5038e+02 100.0% 3.2720e+12 100.0% 9.693e+05 100.0% 4.244e+04 100.0% 2.142e+03 100.0% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage VecView 8 1.0 1.9149e+01 1.0 0.00e+00 0.0 2.5e+04 3.1e+05 1.6e+01 8 0 3 19 1 8 0 3 19 1 0 VecMDot 927 1.0 9.2982e+01 1.3 9.63e+09 1.1 0.0e+00 0.0e+00 9.3e+02 34 36 0 0 43 34 36 0 0 43 12733 VecNorm 976 1.0 3.6295e+01 1.6 5.25e+08 1.1 0.0e+00 0.0e+00 9.8e+02 12 2 0 0 46 12 2 0 0 46 1779 VecScale 1000 1.0 1.5538e+00 2.7 2.69e+08 1.1 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 21287 VecCopy 283 1.0 9.1090e-01 1.9 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 1587 1.0 2.6465e+00 2.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 0 VecAXPY 234 1.0 1.0124e+00 3.4 1.26e+08 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 15290 VecMAXPY 976 1.0 4.6520e+01 1.4 1.05e+10 1.1 0.0e+00 0.0e+00 0.0e+00 16 40 0 0 0 16 40 0 0 0 27827 VecScatterBegin 2308 1.0 6.4895e+00 2.2 0.00e+00 0.0 9.5e+05 3.8e+04 0.0e+00 2 0 98 87 0 2 0 98 87 0 0 VecScatterEnd 2308 1.0 2.2144e+01 4.9 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 5 0 0 0 0 5 0 0 0 0 0 MatMult 766 1.0 3.0886e+01 1.7 2.66e+09 1.1 4.7e+05 3.5e+04 0.0e+00 9 10 48 40 0 9 10 48 40 0 10633 MatSolve 767 1.0 2.1649e+01 1.8 2.77e+09 1.1 0.0e+00 0.0e+00 0.0e+00 7 11 0 0 0 7 11 0 0 0 16223 MatLUFactorNum 1 1.0 1.1517e-01 2.5 6.05e+06 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 6599 MatILUFactorSym 1 1.0 1.0871e-01 3.7 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyBegin 3 1.0 5.6961e-01 3.2 0.00e+00 0.0 0.0e+00 0.0e+00 4.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyEnd 3 1.0 2.4171e-01 1.3 0.00e+00 0.0 1.2e+03 8.8e+03 8.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetRowIJ 1 1.0 1.0014e-05 5.2 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetSubMatrice 1 1.0 1.2215e+00 2.0 0.00e+00 0.0 3.0e+03 8.1e+04 7.0e+00 0 0 0 1 0 0 0 0 1 0 0 MatGetOrdering 1 1.0 1.4469e-02 4.2 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatIncreaseOvrlp 1 1.0 4.1109e-01 1.9 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatView 3 3.0 2.9133e-02153.3 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPGMRESOrthog 927 1.0 1.2676e+02 1.1 1.93e+10 1.1 0.0e+00 0.0e+00 9.3e+02 48 72 0 0 43 48 72 0 0 43 18681 KSPSetUp 2 1.0 9.2526e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 5.2e+01 0 0 0 0 2 0 0 0 0 2 0 KSPSolve 1 1.0 2.1155e+02 1.0 2.65e+10 1.1 9.4e+05 3.5e+04 2.1e+03 84100 97 80 96 84100 97 80 96 15467 PCSetUp 2 1.0 1.8918e+00 1.5 6.05e+06 1.1 4.3e+03 6.0e+04 2.3e+01 1 0 0 1 1 1 0 0 1 1 402 PCSetUpOnBlocks 1 1.0 2.2074e-01 2.7 6.05e+06 1.1 0.0e+00 0.0e+00 3.0e+00 0 0 0 0 0 0 0 0 0 0 3443 PCApply 767 1.0 3.5912e+01 1.6 2.77e+09 1.1 4.7e+05 3.5e+04 0.0e+00 12 11 48 40 0 12 11 48 40 0 9779 Generate Vector 1 1.0 3.4500e+01 1.8 0.00e+00 0.0 2.5e+04 3.1e+05 2.4e+01 8 0 3 19 1 8 0 3 19 1 0 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Vector 86 86 176910552 0 Vector Scatter 8 8 8480 0 Matrix 5 5 86898828 0 Distributed Mesh 2 2 6787792 0 Bipartite Graph 4 4 2800 0 Index Set 22 22 4585380 0 IS L to G Mapping 3 3 4520244 0 Krylov Solver 2 2 31856 0 Preconditioner 2 2 1840 0 Viewer 3 2 1448 0 ======================================================================================================================== Average time to get PetscTime(): 1.90735e-07 Average time for MPI_Barrier(): 0.0152486 Average time for zero size MPI_Send(): 0.00012082 #PETSc Option Table entries: -ksp_gmres_restart 40 -ksp_lgmres_augment 10 -ksp_type lgmres -ksp_view -log_summary -my_ksp_monitor true -pc_type asm #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure run at: Sat Nov 9 12:01:53 2013 Configure options: --download-f-blas-lapack --with-mpi-dir=/usr/local/OpenMPI-1.6.4_Intel --download-hypre=1 --download-hdf5=1 --download-superlu_dist --download-parmetis -download-metis --with-debugging=no ----------------------------------------- Libraries compiled on Sat Nov 9 12:01:53 2013 on cocoa5.aero.psu.edu Machine characteristics: Linux-2.6.32-279.5.1.el6.x86_64-x86_64-with-centos-6.3-Final Using PETSc directory: /home/pzw2/ZSoft/petsc-3.3-p6 Using PETSc arch: arch-linux2-c-opt ----------------------------------------- Using C compiler: /usr/local/OpenMPI-1.6.4_Intel/bin/mpicc -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: /usr/local/OpenMPI-1.6.4_Intel/bin/mpif90 -O3 ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/home/pzw2/ZSoft/petsc-3.3-p6/arch-linux2-c-opt/include -I/home/pzw2/ZSoft/petsc-3.3-p6/include -I/home/pzw2/ZSoft/petsc-3.3-p6/include -I/home/pzw2/ZSoft/petsc-3.3-p6/arch-linux2-c-opt/include -I/usr/local/OpenMPI-1.6.4_Intel/include -I/usr/local/OpenMPI-1.6.4_Intel/include/openmpi/opal/mca/hwloc/hwloc132/hwloc/include -I/usr/local/OpenMPI-1.6.4_Intel/include/openmpi ----------------------------------------- Using C linker: /usr/local/OpenMPI-1.6.4_Intel/bin/mpicc Using Fortran linker: /usr/local/OpenMPI-1.6.4_Intel/bin/mpif90 Using libraries: -Wl,-rpath,/home/pzw2/ZSoft/petsc-3.3-p6/arch-linux2-c-opt/lib -L/home/pzw2/ZSoft/petsc-3.3-p6/arch-linux2-c-opt/lib -lpetsc -lX11 -Wl,-rpath,/home/pzw2/ZSoft/petsc-3.3-p6/arch-linux2-c-opt/lib -L/home/pzw2/ZSoft/petsc-3.3-p6/arch-linux2-c-opt/lib -lsuperlu_dist_3.1 -lparmetis -lmetis -lpthread -lHYPRE -L/usr/local/OpenMPI-1.6.4_Intel/lib -L/usr/lib/gcc/x86_64-redhat-linux/4.4.6 -lmpi_cxx -lstdc++ -lflapack -lfblas -lhdf5_fortran -lhdf5 -lhdf5hl_fortran -lhdf5_hl -lz -lmpi_f90 -lmpi_f77 -lm -lm -L/opt/intel/composer_xe_2011_sp1.10.319/compiler/lib/intel64 -limf -lm -lm -lifport -lifcore -lsvml -lm -lipgo -lirc -lirc_s -lm -lm -lm -lmpi_cxx -lstdc++ -lmpi_cxx -lstdc++ -ldl -lmpi -lrt -lnsl -lutil -lgcc_s -lpthread -ldl -----------------------------------------