---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./x.r on a arch-linux2-c-opt named node31.cocoa5 with 256 processors, by pzw2 Sun Nov 10 11:30:07 2013 Using Petsc Release Version 3.3.0, Patch 6, Mon Feb 11 12:26:34 CST 2013 Max Max/Min Avg Total Time (sec): 1.724e+02 1.00082 1.724e+02 Objects: 1.370e+02 1.00000 1.370e+02 Flops: 1.341e+10 1.06368 1.279e+10 3.275e+12 Flops/sec: 7.782e+07 1.06388 7.422e+07 1.900e+10 MPI Messages: 9.695e+03 1.92820 8.145e+03 2.085e+06 MPI Message Lengths: 1.770e+09 14.32685 2.471e+04 5.151e+10 MPI Reductions: 2.141e+03 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 1.7236e+02 100.0% 3.2749e+12 100.0% 2.085e+06 100.0% 2.471e+04 100.0% 2.140e+03 100.0% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage VecView 8 1.0 2.0219e+01 1.0 0.00e+00 0.0 9.8e+04 8.0e+04 1.6e+01 12 0 5 15 1 12 0 5 15 1 0 VecMDot 926 1.0 6.7355e+01 1.3 4.87e+09 1.1 0.0e+00 0.0e+00 9.3e+02 35 36 0 0 43 35 36 0 0 43 17557 VecNorm 975 1.0 2.8405e+01 1.6 2.66e+08 1.1 0.0e+00 0.0e+00 9.8e+02 14 2 0 0 46 14 2 0 0 46 2271 VecScale 999 1.0 6.5648e-01 3.2 1.36e+08 1.1 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 50334 VecCopy 283 1.0 4.4710e-01 2.8 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 1585 1.0 1.4419e+00 2.8 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 234 1.0 4.2862e-01 3.6 6.37e+07 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 36115 VecMAXPY 975 1.0 2.3674e+01 1.8 5.32e+09 1.1 0.0e+00 0.0e+00 0.0e+00 11 39 0 0 0 11 39 0 0 0 54618 VecScatterBegin 2305 1.0 3.9411e+00 3.2 0.00e+00 0.0 2.0e+06 2.3e+04 0.0e+00 1 0 97 89 0 1 0 97 89 0 0 VecScatterEnd 2305 1.0 1.7004e+01 4.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 6 0 0 0 0 6 0 0 0 0 0 MatMult 765 1.0 1.5970e+01 2.0 1.34e+09 1.1 9.8e+05 2.2e+04 0.0e+00 7 10 47 42 0 7 10 47 42 0 20539 MatSolve 766 1.0 1.4117e+01 3.1 1.41e+09 1.1 0.0e+00 0.0e+00 0.0e+00 5 11 0 0 0 5 11 0 0 0 25326 MatLUFactorNum 1 1.0 6.7332e-02 2.7 3.07e+06 1.1 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 11481 MatILUFactorSym 1 1.0 5.3601e-02 3.5 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyBegin 3 1.0 3.5198e-01 3.2 0.00e+00 0.0 0.0e+00 0.0e+00 4.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyEnd 3 1.0 2.0012e-01 1.4 0.00e+00 0.0 2.6e+03 5.5e+03 8.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetRowIJ 1 1.0 1.2159e-05 6.4 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetSubMatrice 1 1.0 8.8131e-01 2.4 0.00e+00 0.0 6.4e+03 5.0e+04 7.0e+00 0 0 0 1 0 0 0 0 1 0 0 MatGetOrdering 1 1.0 1.2251e-02 6.7 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatIncreaseOvrlp 1 1.0 3.9644e-01 2.2 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatView 3 3.0 4.4928e-02198.8 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPGMRESOrthog 926 1.0 8.4192e+01 1.2 9.74e+09 1.1 0.0e+00 0.0e+00 9.3e+02 45 72 0 0 43 45 72 0 0 43 28091 KSPSetUp 2 1.0 1.1193e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 5.2e+01 1 0 0 0 2 1 0 0 0 2 0 KSPSolve 1 1.0 1.3453e+02 1.0 1.34e+10 1.1 2.0e+06 2.2e+04 2.1e+03 78100 94 84 96 78100 94 84 96 24342 PCSetUp 2 1.0 1.4820e+00 1.5 3.07e+06 1.1 9.0e+03 3.7e+04 2.3e+01 1 0 0 1 1 1 0 0 1 1 522 PCSetUpOnBlocks 1 1.0 1.1506e-01 2.7 3.07e+06 1.1 0.0e+00 0.0e+00 3.0e+00 0 0 0 0 0 0 0 0 0 0 6718 PCApply 766 1.0 2.2657e+01 2.2 1.41e+09 1.1 9.8e+05 2.2e+04 0.0e+00 9 11 47 42 0 9 11 47 42 0 15780 Generate Vector 1 1.0 3.2557e+01 1.6 0.00e+00 0.0 9.8e+04 8.0e+04 2.4e+01 12 0 5 15 1 12 0 5 15 1 0 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Vector 86 86 89650392 0 Vector Scatter 8 8 8480 0 Matrix 5 5 44222508 0 Distributed Mesh 2 2 3481552 0 Bipartite Graph 4 4 2800 0 Index Set 22 22 2362180 0 IS L to G Mapping 3 3 2316084 0 Krylov Solver 2 2 31856 0 Preconditioner 2 2 1840 0 Viewer 3 2 1448 0 ======================================================================================================================== Average time to get PetscTime(): 9.53674e-08 Average time for MPI_Barrier(): 0.0160028 Average time for zero size MPI_Send(): 0.000165327 #PETSc Option Table entries: -ksp_gmres_restart 40 -ksp_lgmres_augment 10 -ksp_type lgmres -ksp_view -log_summary -my_ksp_monitor true -pc_type asm #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure run at: Sat Nov 9 12:01:53 2013 Configure options: --download-f-blas-lapack --with-mpi-dir=/usr/local/OpenMPI-1.6.4_Intel --download-hypre=1 --download-hdf5=1 --download-superlu_dist --download-parmetis -download-metis --with-debugging=no ----------------------------------------- Libraries compiled on Sat Nov 9 12:01:53 2013 on cocoa5.aero.psu.edu Machine characteristics: Linux-2.6.32-279.5.1.el6.x86_64-x86_64-with-centos-6.3-Final Using PETSc directory: /home/pzw2/ZSoft/petsc-3.3-p6 Using PETSc arch: arch-linux2-c-opt ----------------------------------------- Using C compiler: /usr/local/OpenMPI-1.6.4_Intel/bin/mpicc -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: /usr/local/OpenMPI-1.6.4_Intel/bin/mpif90 -O3 ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/home/pzw2/ZSoft/petsc-3.3-p6/arch-linux2-c-opt/include -I/home/pzw2/ZSoft/petsc-3.3-p6/include -I/home/pzw2/ZSoft/petsc-3.3-p6/include -I/home/pzw2/ZSoft/petsc-3.3-p6/arch-linux2-c-opt/include -I/usr/local/OpenMPI-1.6.4_Intel/include -I/usr/local/OpenMPI-1.6.4_Intel/include/openmpi/opal/mca/hwloc/hwloc132/hwloc/include -I/usr/local/OpenMPI-1.6.4_Intel/include/openmpi ----------------------------------------- Using C linker: /usr/local/OpenMPI-1.6.4_Intel/bin/mpicc Using Fortran linker: /usr/local/OpenMPI-1.6.4_Intel/bin/mpif90 Using libraries: -Wl,-rpath,/home/pzw2/ZSoft/petsc-3.3-p6/arch-linux2-c-opt/lib -L/home/pzw2/ZSoft/petsc-3.3-p6/arch-linux2-c-opt/lib -lpetsc -lX11 -Wl,-rpath,/home/pzw2/ZSoft/petsc-3.3-p6/arch-linux2-c-opt/lib -L/home/pzw2/ZSoft/petsc-3.3-p6/arch-linux2-c-opt/lib -lsuperlu_dist_3.1 -lparmetis -lmetis -lpthread -lHYPRE -L/usr/local/OpenMPI-1.6.4_Intel/lib -L/usr/lib/gcc/x86_64-redhat-linux/4.4.6 -lmpi_cxx -lstdc++ -lflapack -lfblas -lhdf5_fortran -lhdf5 -lhdf5hl_fortran -lhdf5_hl -lz -lmpi_f90 -lmpi_f77 -lm -lm -L/opt/intel/composer_xe_2011_sp1.10.319/compiler/lib/intel64 -limf -lm -lm -lifport -lifcore -lsvml -lm -lipgo -lirc -lirc_s -lm -lm -lm -lmpi_cxx -lstdc++ -lmpi_cxx -lstdc++ -ldl -lmpi -lrt -lnsl -lutil -lgcc_s -lpthread -ldl -----------------------------------------