lid velocity = 0.0001, prandtl # = 1, grashof # = 1 Number of SNES iterations = 2 ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./ex19 on a intel-opt-precise-O3 named lagrange.tomato with 1 processor, by jfe Thu Oct 25 13:35:39 2012 Using Petsc Development HG revision: f8bbb9afb3f28a97dd47839f1c4674891dd5c594 HG Date: Wed Oct 24 13:40:55 2012 -0400 Max Max/Min Avg Total Time (sec): 5.081e+00 1.00000 5.081e+00 Objects: 9.500e+01 1.00000 9.500e+01 Flops: 8.620e+09 1.00000 8.620e+09 8.620e+09 Flops/sec: 1.697e+09 1.00000 1.697e+09 1.697e+09 MPI Messages: 0.000e+00 0.00000 0.000e+00 0.000e+00 MPI Message Lengths: 0.000e+00 0.00000 0.000e+00 0.000e+00 MPI Reductions: 1.290e+02 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 5.0808e+00 100.0% 8.6196e+09 100.0% 0.000e+00 0.0% 0.000e+00 0.0% 1.280e+02 99.2% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage ThreadCommRunKernel 12903 1.0 4.8756e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 96 0 0 0 0 96 0 0 0 0 0 ThreadCommBarrier 12903 1.0 2.7401e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 SNESSolve 1 1.0 5.0503e+00 1.0 8.62e+09 1.0 0.0e+00 0.0e+00 1.1e+02 99100 0 0 83 99100 0 0 84 1707 SNESFunctionEval 3 1.0 1.1671e-03 1.0 2.52e+06 1.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 2 0 0 0 0 2 2159 SNESJacobianEval 2 1.0 1.4664e-01 1.0 3.85e+07 1.0 0.0e+00 0.0e+00 3.3e+01 3 0 0 0 26 3 0 0 0 26 262 SNESLineSearch 2 1.0 3.9492e-03 1.0 5.09e+06 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 1290 VecDot 2 1.0 1.8406e-04 1.0 1.60e+05 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 869 VecMDot 2024 1.0 1.4353e+00 1.0 2.50e+09 1.0 0.0e+00 0.0e+00 0.0e+00 28 29 0 0 0 28 29 0 0 0 1742 VecNorm 2095 1.0 1.4679e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 3 0 0 0 0 3 0 0 0 0 0 VecScale 2092 1.0 3.8544e-02 1.0 8.37e+07 1.0 0.0e+00 0.0e+00 0.0e+00 1 1 0 0 0 1 1 0 0 0 2171 VecCopy 2206 1.0 9.1011e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 2 0 0 0 0 2 0 0 0 0 0 VecSet 116 1.0 8.2335e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 174 1.0 1.5033e-02 1.0 1.39e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 926 VecWAXPY 2 1.0 1.7214e-04 1.0 8.00e+04 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 465 VecMAXPY 2092 1.0 8.1948e-01 1.0 2.66e+09 1.0 0.0e+00 0.0e+00 0.0e+00 16 31 0 0 0 16 31 0 0 0 3249 VecScatterBegin 47 1.0 3.6745e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecReduceArith 5 1.0 3.9327e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecReduceComm 3 1.0 4.2915e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecNormalize 2092 1.0 1.8828e-01 1.0 8.37e+07 1.0 0.0e+00 0.0e+00 0.0e+00 4 1 0 0 0 4 1 0 0 0 444 MatMult 2092 1.0 2.3284e+00 1.0 3.32e+09 1.0 0.0e+00 0.0e+00 0.0e+00 46 39 0 0 0 46 39 0 0 0 1426 MatAssemblyBegin 3 1.0 9.5367e-07 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyEnd 3 1.0 2.7530e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatZeroEntries 2 1.0 1.2460e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatFDColorCreate 1 1.0 1.3916e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 2.8e+01 0 0 0 0 22 0 0 0 0 22 0 MatFDColorApply 2 1.0 1.3218e-01 1.0 3.85e+07 1.0 0.0e+00 0.0e+00 2.0e+00 3 0 0 0 2 3 0 0 0 2 291 MatFDColorFunc 42 1.0 1.3001e-02 1.0 3.53e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 2714 KSPGMRESOrthog 2024 1.0 2.2084e+00 1.0 5.00e+09 1.0 0.0e+00 0.0e+00 0.0e+00 43 58 0 0 0 43 58 0 0 0 2265 KSPSetUp 2 1.0 7.1192e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+01 0 0 0 0 8 0 0 0 0 8 0 KSPSolve 2 1.0 4.8950e+00 1.0 8.58e+09 1.0 0.0e+00 0.0e+00 6.8e+01 96 99 0 0 53 96 99 0 0 53 1752 PCSetUp 2 1.0 1.1921e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 PCApply 2092 1.0 8.7008e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 2 0 0 0 0 2 0 0 0 0 0 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Container 2 2 1112 0 SNES 1 1 1292 0 SNESLineSearch 1 1 848 0 Vector 50 50 15112128 0 Vector Scatter 3 3 1884 0 Matrix 1 1 10165812 0 Matrix FD Coloring 1 1 724 0 Distributed Mesh 1 1 205328 0 Bipartite Graph 2 2 1400 0 Index Set 27 27 220288 0 IS L to G Mapping 3 3 161716 0 Krylov Solver 1 1 18304 0 Preconditioner 1 1 768 0 Viewer 1 0 0 0 ======================================================================================================================== Average time to get PetscTime(): 0 #PETSc Option Table entries: -da_grid_x 100 -da_grid_y 100 -log_summary -mat_no_inode -pc_type none -preload off -threadcomm_nthreads 1 -threadcomm_type pthread #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure run at: Wed Oct 24 15:17:55 2012 Configure options: --with-x=0 --download-f-blas-lapack=0 --with-blas-lapack-dir=/opt/intel/composerxe/mkl/lib/intel64 --with-mpi=1 --with-mpi-shared=1 --with-mpi=1 --download-mpich=no --with-openmp=1 --with-pthreadclasses=1 --with-debugging=0 --with-gnu-compilers=no --with-vendor-compilers=intel --with-cc=/usr/local/encap/platform_mpi-8.02.01/bin/mpicc --with-cxx=/usr/local/encap/platform_mpi-8.02.01/bin/mpiCC --with-fc=/usr/local/encap/platform_mpi-8.02.01/bin/mpif90 --with-shared-libraries=1 --with-c++-support --with-clanguage=C --COPTFLAGS="-fPIC -O3 -xSSE4.2 -fp-model precise -g -debug inline_debug_info" --CXXOPTFLAGS="-fPIC -O3 -xSSE4.2 -fp-model precise -g -debug inline_debug_info" --FOPTFLAGS="-fPIC -O3 -xSSE4.2 -fp-model precise -g -debug inline_debug_info" --download-scalapack=1 --download-blacs=1 --with-blacs=1 --download-umfpack=1 --download-parmetis=1 --download-metis=1 --download-superlu=1 --download-superlu_dist=1 --download-mumps=1 --download-ml=1 --download-hypre=1 ----------------------------------------- Libraries compiled on Wed Oct 24 15:17:55 2012 on lagrange.tomato Machine characteristics: Linux-2.6.32-279.11.1.el6.x86_64-x86_64-with-centos-6.3-Final Using PETSc directory: /home/jfe/local/petsc-dev Using PETSc arch: intel-opt-precise-O3 ----------------------------------------- Using C compiler: /usr/local/encap/platform_mpi-8.02.01/bin/mpicc -fPIC -wd1572 -fPIC -O3 -xSSE4.2 -fp-model precise -g -debug inline_debug_info -fopenmp ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: /usr/local/encap/platform_mpi-8.02.01/bin/mpif90 -fPIC -fPIC -O3 -xSSE4.2 -fp-model precise -g -debug inline_debug_info -fopenmp ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/home/jfe/local/petsc-dev/intel-opt-precise-O3/include -I/home/jfe/local/petsc-dev/include -I/home/jfe/local/petsc-dev/include -I/home/jfe/local/petsc-dev/intel-opt-precise-O3/include -I/usr/local/encap/platform_mpi-8.02.01/include ----------------------------------------- Using C linker: /usr/local/encap/platform_mpi-8.02.01/bin/mpicc Using Fortran linker: /usr/local/encap/platform_mpi-8.02.01/bin/mpif90 Using libraries: -Wl,-rpath,/home/jfe/local/petsc-dev/intel-opt-precise-O3/lib -L/home/jfe/local/petsc-dev/intel-opt-precise-O3/lib -lpetsc -Wl,-rpath,/home/jfe/local/petsc-dev/intel-opt-precise-O3/lib -L/home/jfe/local/petsc-dev/intel-opt-precise-O3/lib -lcmumps -ldmumps -lsmumps -lzmumps -lmumps_common -lpord -lscalapack -lblacs -lml -Wl,-rpath,/usr/local/encap/platform_mpi-8.02.01/lib/linux_amd64 -L/usr/local/encap/platform_mpi-8.02.01/lib/linux_amd64 -lmpiCC -Wl,-rpath,/opt/intel/composer_xe_2011_sp1.7.256/compiler/lib/intel64 -L/opt/intel/composer_xe_2011_sp1.7.256/compiler/lib/intel64 -Wl,-rpath,/opt/intel/composer_xe_2011_sp1.7.256/ipp/lib/intel64 -L/opt/intel/composer_xe_2011_sp1.7.256/ipp/lib/intel64 -Wl,-rpath,/opt/intel/composer_xe_2011_sp1.7.256/mkl/lib/intel64 -L/opt/intel/composer_xe_2011_sp1.7.256/mkl/lib/intel64 -Wl,-rpath,/opt/intel/composer_xe_2011_sp1.7.256/tbb/lib/intel64/cc4.1.0_libc2.4_kernel2.6.16.21 -L/opt/intel/composer_xe_2011_sp1.7.256/tbb/lib/intel64/cc4.1.0_libc2.4_kernel2.6.16.21 -Wl,-rpath,/usr/lib/gcc/x86_64-redhat-linux/4.4.6 -L/usr/lib/gcc/x86_64-redhat-linux/4.4.6 -lpthread -lsuperlu_dist_3.1 -lparmetis -lmetis -lsuperlu_4.3 -lHYPRE -lmpiCC -lumfpack -lamd -Wl,-rpath,/opt/intel/composerxe/mkl/lib/intel64 -L/opt/intel/composerxe/mkl/lib/intel64 -lmkl_intel_lp64 -lmkl_intel_thread -lmkl_core -liomp5 -lpthread -lifport -lifcore -lm -lpthread -lm -lmpiCC -lpcmpio -lpcmpi -ldl -limf -lsvml -lipgo -ldecimal -lcilkrts -lstdc++ -lgcc_s -lirc -lirc_s -ldl ----------------------------------------- lid velocity = 0.0001, prandtl # = 1, grashof # = 1 Number of SNES iterations = 2 ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./ex19 on a intel-opt-precise-O3 named lagrange.tomato with 1 processor, by jfe Thu Oct 25 13:40:05 2012 With 2 threads per MPI_Comm Using Petsc Development HG revision: f8bbb9afb3f28a97dd47839f1c4674891dd5c594 HG Date: Wed Oct 24 13:40:55 2012 -0400 Max Max/Min Avg Total Time (sec): 2.654e+02 1.00000 2.654e+02 Objects: 9.500e+01 1.00000 9.500e+01 Flops: 8.620e+09 1.00000 8.620e+09 8.620e+09 Flops/sec: 3.248e+07 1.00000 3.248e+07 3.248e+07 MPI Messages: 0.000e+00 0.00000 0.000e+00 0.000e+00 MPI Message Lengths: 0.000e+00 0.00000 0.000e+00 0.000e+00 MPI Reductions: 1.290e+02 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 2.6535e+02 100.0% 8.6196e+09 100.0% 0.000e+00 0.0% 0.000e+00 0.0% 1.280e+02 99.2% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage ThreadCommRunKernel 12903 1.0 2.6515e+02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00100 0 0 0 0 100 0 0 0 0 0 ThreadCommBarrier 12903 1.0 2.6228e+02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 99 0 0 0 0 99 0 0 0 0 0 SNESSolve 1 1.0 2.6533e+02 1.0 8.62e+09 1.0 0.0e+00 0.0e+00 1.1e+02100100 0 0 83 100100 0 0 84 32 SNESFunctionEval 3 1.0 1.1396e-03 1.0 2.52e+06 1.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 2 0 0 0 0 2 2211 SNESJacobianEval 2 1.0 1.7499e+00 1.0 3.85e+07 1.0 0.0e+00 0.0e+00 3.3e+01 1 0 0 0 26 1 0 0 0 26 22 SNESLineSearch 2 1.0 3.4004e-01 1.0 5.09e+06 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 15 VecDot 2 1.0 4.9996e-02 1.0 1.60e+05 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 3 VecMDot 2024 1.0 4.1522e+01 1.0 2.50e+09 1.0 0.0e+00 0.0e+00 0.0e+00 16 29 0 0 0 16 29 0 0 0 60 VecNorm 2095 1.0 4.4373e+01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 17 0 0 0 0 17 0 0 0 0 0 VecScale 2092 1.0 4.2740e+01 1.0 8.37e+07 1.0 0.0e+00 0.0e+00 0.0e+00 16 1 0 0 0 16 1 0 0 0 2 VecCopy 2206 1.0 4.5021e+01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 17 0 0 0 0 17 0 0 0 0 0 VecSet 116 1.0 2.2061e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 0 VecAXPY 174 1.0 3.4747e+00 1.0 1.39e+07 1.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 4 VecWAXPY 2 1.0 3.9988e-02 1.0 8.00e+04 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 2 VecMAXPY 2092 1.0 4.2766e+01 1.0 2.66e+09 1.0 0.0e+00 0.0e+00 0.0e+00 16 31 0 0 0 16 31 0 0 0 62 VecScatterBegin 47 1.0 3.6824e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecReduceArith 5 1.0 8.6548e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecReduceComm 3 1.0 5.7220e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecNormalize 2092 1.0 8.7027e+01 1.0 8.37e+07 1.0 0.0e+00 0.0e+00 0.0e+00 33 1 0 0 0 33 1 0 0 0 1 MatMult 2092 1.0 4.2840e+01 1.0 3.32e+09 1.0 0.0e+00 0.0e+00 0.0e+00 16 39 0 0 0 16 39 0 0 0 78 MatAssemblyBegin 3 1.0 1.9073e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyEnd 3 1.0 2.7621e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatZeroEntries 2 1.0 3.9911e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatFDColorCreate 1 1.0 4.9851e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 2.8e+01 0 0 0 0 22 0 0 0 0 22 0 MatFDColorApply 2 1.0 1.6472e+00 1.0 3.85e+07 1.0 0.0e+00 0.0e+00 2.0e+00 1 0 0 0 2 1 0 0 0 2 23 MatFDColorFunc 42 1.0 1.3023e-02 1.0 3.53e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 2709 KSPGMRESOrthog 2024 1.0 8.2930e+01 1.0 5.00e+09 1.0 0.0e+00 0.0e+00 0.0e+00 31 58 0 0 0 31 58 0 0 0 60 KSPSetUp 2 1.0 9.9998e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+01 0 0 0 0 8 0 0 0 0 8 0 KSPSolve 2 1.0 2.6318e+02 1.0 8.58e+09 1.0 0.0e+00 0.0e+00 6.8e+01 99 99 0 0 53 99 99 0 0 53 33 PCSetUp 2 1.0 0.0000e+00 0.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 PCApply 2092 1.0 4.2851e+01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 16 0 0 0 0 16 0 0 0 0 0 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Container 2 2 1112 0 SNES 1 1 1292 0 SNESLineSearch 1 1 848 0 Vector 50 50 15112128 0 Vector Scatter 3 3 1884 0 Matrix 1 1 10165812 0 Matrix FD Coloring 1 1 724 0 Distributed Mesh 1 1 205328 0 Bipartite Graph 2 2 1400 0 Index Set 27 27 220288 0 IS L to G Mapping 3 3 161716 0 Krylov Solver 1 1 18304 0 Preconditioner 1 1 768 0 Viewer 1 0 0 0 ======================================================================================================================== Average time to get PetscTime(): 0 #PETSc Option Table entries: -da_grid_x 100 -da_grid_y 100 -log_summary -mat_no_inode -pc_type none -preload off -threadcomm_nthreads 2 -threadcomm_type pthread #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure run at: Wed Oct 24 15:17:55 2012 Configure options: --with-x=0 --download-f-blas-lapack=0 --with-blas-lapack-dir=/opt/intel/composerxe/mkl/lib/intel64 --with-mpi=1 --with-mpi-shared=1 --with-mpi=1 --download-mpich=no --with-openmp=1 --with-pthreadclasses=1 --with-debugging=0 --with-gnu-compilers=no --with-vendor-compilers=intel --with-cc=/usr/local/encap/platform_mpi-8.02.01/bin/mpicc --with-cxx=/usr/local/encap/platform_mpi-8.02.01/bin/mpiCC --with-fc=/usr/local/encap/platform_mpi-8.02.01/bin/mpif90 --with-shared-libraries=1 --with-c++-support --with-clanguage=C --COPTFLAGS="-fPIC -O3 -xSSE4.2 -fp-model precise -g -debug inline_debug_info" --CXXOPTFLAGS="-fPIC -O3 -xSSE4.2 -fp-model precise -g -debug inline_debug_info" --FOPTFLAGS="-fPIC -O3 -xSSE4.2 -fp-model precise -g -debug inline_debug_info" --download-scalapack=1 --download-blacs=1 --with-blacs=1 --download-umfpack=1 --download-parmetis=1 --download-metis=1 --download-superlu=1 --download-superlu_dist=1 --download-mumps=1 --download-ml=1 --download-hypre=1 ----------------------------------------- Libraries compiled on Wed Oct 24 15:17:55 2012 on lagrange.tomato Machine characteristics: Linux-2.6.32-279.11.1.el6.x86_64-x86_64-with-centos-6.3-Final Using PETSc directory: /home/jfe/local/petsc-dev Using PETSc arch: intel-opt-precise-O3 ----------------------------------------- Using C compiler: /usr/local/encap/platform_mpi-8.02.01/bin/mpicc -fPIC -wd1572 -fPIC -O3 -xSSE4.2 -fp-model precise -g -debug inline_debug_info -fopenmp ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: /usr/local/encap/platform_mpi-8.02.01/bin/mpif90 -fPIC -fPIC -O3 -xSSE4.2 -fp-model precise -g -debug inline_debug_info -fopenmp ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/home/jfe/local/petsc-dev/intel-opt-precise-O3/include -I/home/jfe/local/petsc-dev/include -I/home/jfe/local/petsc-dev/include -I/home/jfe/local/petsc-dev/intel-opt-precise-O3/include -I/usr/local/encap/platform_mpi-8.02.01/include ----------------------------------------- Using C linker: /usr/local/encap/platform_mpi-8.02.01/bin/mpicc Using Fortran linker: /usr/local/encap/platform_mpi-8.02.01/bin/mpif90 Using libraries: -Wl,-rpath,/home/jfe/local/petsc-dev/intel-opt-precise-O3/lib -L/home/jfe/local/petsc-dev/intel-opt-precise-O3/lib -lpetsc -Wl,-rpath,/home/jfe/local/petsc-dev/intel-opt-precise-O3/lib -L/home/jfe/local/petsc-dev/intel-opt-precise-O3/lib -lcmumps -ldmumps -lsmumps -lzmumps -lmumps_common -lpord -lscalapack -lblacs -lml -Wl,-rpath,/usr/local/encap/platform_mpi-8.02.01/lib/linux_amd64 -L/usr/local/encap/platform_mpi-8.02.01/lib/linux_amd64 -lmpiCC -Wl,-rpath,/opt/intel/composer_xe_2011_sp1.7.256/compiler/lib/intel64 -L/opt/intel/composer_xe_2011_sp1.7.256/compiler/lib/intel64 -Wl,-rpath,/opt/intel/composer_xe_2011_sp1.7.256/ipp/lib/intel64 -L/opt/intel/composer_xe_2011_sp1.7.256/ipp/lib/intel64 -Wl,-rpath,/opt/intel/composer_xe_2011_sp1.7.256/mkl/lib/intel64 -L/opt/intel/composer_xe_2011_sp1.7.256/mkl/lib/intel64 -Wl,-rpath,/opt/intel/composer_xe_2011_sp1.7.256/tbb/lib/intel64/cc4.1.0_libc2.4_kernel2.6.16.21 -L/opt/intel/composer_xe_2011_sp1.7.256/tbb/lib/intel64/cc4.1.0_libc2.4_kernel2.6.16.21 -Wl,-rpath,/usr/lib/gcc/x86_64-redhat-linux/4.4.6 -L/usr/lib/gcc/x86_64-redhat-linux/4.4.6 -lpthread -lsuperlu_dist_3.1 -lparmetis -lmetis -lsuperlu_4.3 -lHYPRE -lmpiCC -lumfpack -lamd -Wl,-rpath,/opt/intel/composerxe/mkl/lib/intel64 -L/opt/intel/composerxe/mkl/lib/intel64 -lmkl_intel_lp64 -lmkl_intel_thread -lmkl_core -liomp5 -lpthread -lifport -lifcore -lm -lpthread -lm -lmpiCC -lpcmpio -lpcmpi -ldl -limf -lsvml -lipgo -ldecimal -lcilkrts -lstdc++ -lgcc_s -lirc -lirc_s -ldl ----------------------------------------- lid velocity = 0.0001, prandtl # = 1, grashof # = 1 Number of SNES iterations = 2 ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./ex19 on a intel-opt-precise-O3 named lagrange.tomato with 1 processor, by jfe Thu Oct 25 13:44:05 2012 With 4 threads per MPI_Comm Using Petsc Development HG revision: f8bbb9afb3f28a97dd47839f1c4674891dd5c594 HG Date: Wed Oct 24 13:40:55 2012 -0400 Max Max/Min Avg Total Time (sec): 2.384e+02 1.00000 2.384e+02 Objects: 9.500e+01 1.00000 9.500e+01 Flops: 8.620e+09 1.00000 8.620e+09 8.620e+09 Flops/sec: 3.615e+07 1.00000 3.615e+07 3.615e+07 MPI Messages: 0.000e+00 0.00000 0.000e+00 0.000e+00 MPI Message Lengths: 0.000e+00 0.00000 0.000e+00 0.000e+00 MPI Reductions: 1.290e+02 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 2.3843e+02 100.0% 8.6196e+09 100.0% 0.000e+00 0.0% 0.000e+00 0.0% 1.280e+02 99.2% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage ThreadCommRunKernel 12903 1.0 2.3823e+02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00100 0 0 0 0 100 0 0 0 0 0 ThreadCommBarrier 12903 1.0 2.3691e+02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 99 0 0 0 0 99 0 0 0 0 0 SNESSolve 1 1.0 2.3841e+02 1.0 8.62e+09 1.0 0.0e+00 0.0e+00 1.1e+02100100 0 0 83 100100 0 0 84 36 SNESFunctionEval 3 1.0 1.1270e-03 1.0 2.52e+06 1.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 2 0 0 0 0 2 2236 SNESJacobianEval 2 1.0 1.4030e+00 1.0 3.85e+07 1.0 0.0e+00 0.0e+00 3.3e+01 1 0 0 0 26 1 0 0 0 26 27 SNESLineSearch 2 1.0 3.0996e-01 1.0 5.09e+06 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 16 VecDot 2 1.0 3.2903e-02 1.0 1.60e+05 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 5 VecMDot 2024 1.0 3.7882e+01 1.0 2.50e+09 1.0 0.0e+00 0.0e+00 0.0e+00 16 29 0 0 0 16 29 0 0 0 66 VecNorm 2095 1.0 3.8691e+01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 16 0 0 0 0 16 0 0 0 0 0 VecScale 2092 1.0 3.8647e+01 1.0 8.37e+07 1.0 0.0e+00 0.0e+00 0.0e+00 16 1 0 0 0 16 1 0 0 0 2 VecCopy 2206 1.0 3.9792e+01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 17 0 0 0 0 17 0 0 0 0 0 VecSet 116 1.0 1.9698e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 0 VecAXPY 174 1.0 3.0586e+00 1.0 1.39e+07 1.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 5 VecWAXPY 2 1.0 3.2971e-02 1.0 8.00e+04 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 2 VecMAXPY 2092 1.0 3.8803e+01 1.0 2.66e+09 1.0 0.0e+00 0.0e+00 0.0e+00 16 31 0 0 0 16 31 0 0 0 69 VecScatterBegin 47 1.0 3.6211e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecReduceArith 5 1.0 8.4755e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecReduceComm 3 1.0 4.0531e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecNormalize 2092 1.0 7.7253e+01 1.0 8.37e+07 1.0 0.0e+00 0.0e+00 0.0e+00 32 1 0 0 0 32 1 0 0 0 1 MatMult 2092 1.0 3.9209e+01 1.0 3.32e+09 1.0 0.0e+00 0.0e+00 0.0e+00 16 39 0 0 0 16 39 0 0 0 85 MatAssemblyBegin 3 1.0 9.5367e-07 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyEnd 3 1.0 2.8040e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatZeroEntries 2 1.0 4.0168e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatFDColorCreate 1 1.0 3.9890e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 2.8e+01 0 0 0 0 22 0 0 0 0 22 0 MatFDColorApply 2 1.0 1.3173e+00 1.0 3.85e+07 1.0 0.0e+00 0.0e+00 2.0e+00 1 0 0 0 2 1 0 0 0 2 29 MatFDColorFunc 42 1.0 1.3178e-02 1.0 3.53e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 2677 KSPGMRESOrthog 2024 1.0 7.5412e+01 1.0 5.00e+09 1.0 0.0e+00 0.0e+00 0.0e+00 32 58 0 0 0 32 58 0 0 0 66 KSPSetUp 2 1.0 7.3021e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+01 0 0 0 0 8 0 0 0 0 8 0 KSPSolve 2 1.0 2.3664e+02 1.0 8.58e+09 1.0 0.0e+00 0.0e+00 6.8e+01 99 99 0 0 53 99 99 0 0 53 36 PCSetUp 2 1.0 0.0000e+00 0.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 PCApply 2092 1.0 3.8000e+01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 16 0 0 0 0 16 0 0 0 0 0 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Container 2 2 1112 0 SNES 1 1 1292 0 SNESLineSearch 1 1 848 0 Vector 50 50 15112128 0 Vector Scatter 3 3 1884 0 Matrix 1 1 10165812 0 Matrix FD Coloring 1 1 724 0 Distributed Mesh 1 1 205328 0 Bipartite Graph 2 2 1400 0 Index Set 27 27 220288 0 IS L to G Mapping 3 3 161716 0 Krylov Solver 1 1 18304 0 Preconditioner 1 1 768 0 Viewer 1 0 0 0 ======================================================================================================================== Average time to get PetscTime(): 9.53674e-08 #PETSc Option Table entries: -da_grid_x 100 -da_grid_y 100 -log_summary -mat_no_inode -pc_type none -preload off -threadcomm_nthreads 4 -threadcomm_type pthread #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure run at: Wed Oct 24 15:17:55 2012 Configure options: --with-x=0 --download-f-blas-lapack=0 --with-blas-lapack-dir=/opt/intel/composerxe/mkl/lib/intel64 --with-mpi=1 --with-mpi-shared=1 --with-mpi=1 --download-mpich=no --with-openmp=1 --with-pthreadclasses=1 --with-debugging=0 --with-gnu-compilers=no --with-vendor-compilers=intel --with-cc=/usr/local/encap/platform_mpi-8.02.01/bin/mpicc --with-cxx=/usr/local/encap/platform_mpi-8.02.01/bin/mpiCC --with-fc=/usr/local/encap/platform_mpi-8.02.01/bin/mpif90 --with-shared-libraries=1 --with-c++-support --with-clanguage=C --COPTFLAGS="-fPIC -O3 -xSSE4.2 -fp-model precise -g -debug inline_debug_info" --CXXOPTFLAGS="-fPIC -O3 -xSSE4.2 -fp-model precise -g -debug inline_debug_info" --FOPTFLAGS="-fPIC -O3 -xSSE4.2 -fp-model precise -g -debug inline_debug_info" --download-scalapack=1 --download-blacs=1 --with-blacs=1 --download-umfpack=1 --download-parmetis=1 --download-metis=1 --download-superlu=1 --download-superlu_dist=1 --download-mumps=1 --download-ml=1 --download-hypre=1 ----------------------------------------- Libraries compiled on Wed Oct 24 15:17:55 2012 on lagrange.tomato Machine characteristics: Linux-2.6.32-279.11.1.el6.x86_64-x86_64-with-centos-6.3-Final Using PETSc directory: /home/jfe/local/petsc-dev Using PETSc arch: intel-opt-precise-O3 ----------------------------------------- Using C compiler: /usr/local/encap/platform_mpi-8.02.01/bin/mpicc -fPIC -wd1572 -fPIC -O3 -xSSE4.2 -fp-model precise -g -debug inline_debug_info -fopenmp ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: /usr/local/encap/platform_mpi-8.02.01/bin/mpif90 -fPIC -fPIC -O3 -xSSE4.2 -fp-model precise -g -debug inline_debug_info -fopenmp ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/home/jfe/local/petsc-dev/intel-opt-precise-O3/include -I/home/jfe/local/petsc-dev/include -I/home/jfe/local/petsc-dev/include -I/home/jfe/local/petsc-dev/intel-opt-precise-O3/include -I/usr/local/encap/platform_mpi-8.02.01/include ----------------------------------------- Using C linker: /usr/local/encap/platform_mpi-8.02.01/bin/mpicc Using Fortran linker: /usr/local/encap/platform_mpi-8.02.01/bin/mpif90 Using libraries: -Wl,-rpath,/home/jfe/local/petsc-dev/intel-opt-precise-O3/lib -L/home/jfe/local/petsc-dev/intel-opt-precise-O3/lib -lpetsc -Wl,-rpath,/home/jfe/local/petsc-dev/intel-opt-precise-O3/lib -L/home/jfe/local/petsc-dev/intel-opt-precise-O3/lib -lcmumps -ldmumps -lsmumps -lzmumps -lmumps_common -lpord -lscalapack -lblacs -lml -Wl,-rpath,/usr/local/encap/platform_mpi-8.02.01/lib/linux_amd64 -L/usr/local/encap/platform_mpi-8.02.01/lib/linux_amd64 -lmpiCC -Wl,-rpath,/opt/intel/composer_xe_2011_sp1.7.256/compiler/lib/intel64 -L/opt/intel/composer_xe_2011_sp1.7.256/compiler/lib/intel64 -Wl,-rpath,/opt/intel/composer_xe_2011_sp1.7.256/ipp/lib/intel64 -L/opt/intel/composer_xe_2011_sp1.7.256/ipp/lib/intel64 -Wl,-rpath,/opt/intel/composer_xe_2011_sp1.7.256/mkl/lib/intel64 -L/opt/intel/composer_xe_2011_sp1.7.256/mkl/lib/intel64 -Wl,-rpath,/opt/intel/composer_xe_2011_sp1.7.256/tbb/lib/intel64/cc4.1.0_libc2.4_kernel2.6.16.21 -L/opt/intel/composer_xe_2011_sp1.7.256/tbb/lib/intel64/cc4.1.0_libc2.4_kernel2.6.16.21 -Wl,-rpath,/usr/lib/gcc/x86_64-redhat-linux/4.4.6 -L/usr/lib/gcc/x86_64-redhat-linux/4.4.6 -lpthread -lsuperlu_dist_3.1 -lparmetis -lmetis -lsuperlu_4.3 -lHYPRE -lmpiCC -lumfpack -lamd -Wl,-rpath,/opt/intel/composerxe/mkl/lib/intel64 -L/opt/intel/composerxe/mkl/lib/intel64 -lmkl_intel_lp64 -lmkl_intel_thread -lmkl_core -liomp5 -lpthread -lifport -lifcore -lm -lpthread -lm -lmpiCC -lpcmpio -lpcmpi -ldl -limf -lsvml -lipgo -ldecimal -lcilkrts -lstdc++ -lgcc_s -lirc -lirc_s -ldl ----------------------------------------- lid velocity = 0.0001, prandtl # = 1, grashof # = 1 Number of SNES iterations = 2 ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./ex19 on a intel-opt-precise-O3 named lagrange.tomato with 1 processor, by jfe Thu Oct 25 13:48:41 2012 With 6 threads per MPI_Comm Using Petsc Development HG revision: f8bbb9afb3f28a97dd47839f1c4674891dd5c594 HG Date: Wed Oct 24 13:40:55 2012 -0400 Max Max/Min Avg Total Time (sec): 2.755e+02 1.00000 2.755e+02 Objects: 9.500e+01 1.00000 9.500e+01 Flops: 8.620e+09 1.00000 8.620e+09 8.620e+09 Flops/sec: 3.128e+07 1.00000 3.128e+07 3.128e+07 MPI Messages: 0.000e+00 0.00000 0.000e+00 0.000e+00 MPI Message Lengths: 0.000e+00 0.00000 0.000e+00 0.000e+00 MPI Reductions: 1.290e+02 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 2.7553e+02 100.0% 8.6196e+09 100.0% 0.000e+00 0.0% 0.000e+00 0.0% 1.280e+02 99.2% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage ThreadCommRunKernel 12903 1.0 2.7533e+02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00100 0 0 0 0 100 0 0 0 0 0 ThreadCommBarrier 12903 1.0 2.7441e+02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00100 0 0 0 0 100 0 0 0 0 0 SNESSolve 1 1.0 2.7550e+02 1.0 8.62e+09 1.0 0.0e+00 0.0e+00 1.1e+02100100 0 0 83 100100 0 0 84 31 SNESFunctionEval 3 1.0 1.1289e-03 1.0 2.52e+06 1.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 2 0 0 0 0 2 2232 SNESJacobianEval 2 1.0 1.8989e+00 1.0 3.85e+07 1.0 0.0e+00 0.0e+00 3.3e+01 1 0 0 0 26 1 0 0 0 26 20 SNESLineSearch 2 1.0 3.9098e-01 1.0 5.09e+06 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 13 VecDot 2 1.0 3.9653e-02 1.0 1.60e+05 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 4 VecMDot 2024 1.0 4.2843e+01 1.0 2.50e+09 1.0 0.0e+00 0.0e+00 0.0e+00 16 29 0 0 0 16 29 0 0 0 58 VecNorm 2095 1.0 4.5446e+01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 16 0 0 0 0 16 0 0 0 0 0 VecScale 2092 1.0 4.3697e+01 1.0 8.37e+07 1.0 0.0e+00 0.0e+00 0.0e+00 16 1 0 0 0 16 1 0 0 0 2 VecCopy 2206 1.0 4.6990e+01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 17 0 0 0 0 17 0 0 0 0 0 VecSet 116 1.0 2.3252e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 0 VecAXPY 174 1.0 3.6222e+00 1.0 1.39e+07 1.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 4 VecWAXPY 2 1.0 3.9968e-02 1.0 8.00e+04 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 2 VecMAXPY 2092 1.0 4.4858e+01 1.0 2.66e+09 1.0 0.0e+00 0.0e+00 0.0e+00 16 31 0 0 0 16 31 0 0 0 59 VecScatterBegin 47 1.0 3.7684e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecReduceArith 5 1.0 1.4365e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecReduceComm 3 1.0 5.9605e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecNormalize 2092 1.0 8.9040e+01 1.0 8.37e+07 1.0 0.0e+00 0.0e+00 0.0e+00 32 1 0 0 0 32 1 0 0 0 1 MatMult 2092 1.0 4.5294e+01 1.0 3.32e+09 1.0 0.0e+00 0.0e+00 0.0e+00 16 39 0 0 0 16 39 0 0 0 73 MatAssemblyBegin 3 1.0 0.0000e+00 0.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyEnd 3 1.0 2.8088e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatZeroEntries 2 1.0 4.4092e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatFDColorCreate 1 1.0 4.9850e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 2.8e+01 0 0 0 0 22 0 0 0 0 22 0 MatFDColorApply 2 1.0 1.7964e+00 1.0 3.85e+07 1.0 0.0e+00 0.0e+00 2.0e+00 1 0 0 0 2 1 0 0 0 2 21 MatFDColorFunc 42 1.0 1.3114e-02 1.0 3.53e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 2690 KSPGMRESOrthog 2024 1.0 8.6243e+01 1.0 5.00e+09 1.0 0.0e+00 0.0e+00 0.0e+00 31 58 0 0 0 31 58 0 0 0 58 KSPSetUp 2 1.0 1.0001e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+01 0 0 0 0 8 0 0 0 0 8 0 KSPSolve 2 1.0 2.7313e+02 1.0 8.58e+09 1.0 0.0e+00 0.0e+00 6.8e+01 99 99 0 0 53 99 99 0 0 53 31 PCSetUp 2 1.0 0.0000e+00 0.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 PCApply 2092 1.0 4.4713e+01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 16 0 0 0 0 16 0 0 0 0 0 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Container 2 2 1112 0 SNES 1 1 1292 0 SNESLineSearch 1 1 848 0 Vector 50 50 15112128 0 Vector Scatter 3 3 1884 0 Matrix 1 1 10165812 0 Matrix FD Coloring 1 1 724 0 Distributed Mesh 1 1 205328 0 Bipartite Graph 2 2 1400 0 Index Set 27 27 220288 0 IS L to G Mapping 3 3 161716 0 Krylov Solver 1 1 18304 0 Preconditioner 1 1 768 0 Viewer 1 0 0 0 ======================================================================================================================== Average time to get PetscTime(): 1.19209e-07 #PETSc Option Table entries: -da_grid_x 100 -da_grid_y 100 -log_summary -mat_no_inode -pc_type none -preload off -threadcomm_nthreads 6 -threadcomm_type pthread #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure run at: Wed Oct 24 15:17:55 2012 Configure options: --with-x=0 --download-f-blas-lapack=0 --with-blas-lapack-dir=/opt/intel/composerxe/mkl/lib/intel64 --with-mpi=1 --with-mpi-shared=1 --with-mpi=1 --download-mpich=no --with-openmp=1 --with-pthreadclasses=1 --with-debugging=0 --with-gnu-compilers=no --with-vendor-compilers=intel --with-cc=/usr/local/encap/platform_mpi-8.02.01/bin/mpicc --with-cxx=/usr/local/encap/platform_mpi-8.02.01/bin/mpiCC --with-fc=/usr/local/encap/platform_mpi-8.02.01/bin/mpif90 --with-shared-libraries=1 --with-c++-support --with-clanguage=C --COPTFLAGS="-fPIC -O3 -xSSE4.2 -fp-model precise -g -debug inline_debug_info" --CXXOPTFLAGS="-fPIC -O3 -xSSE4.2 -fp-model precise -g -debug inline_debug_info" --FOPTFLAGS="-fPIC -O3 -xSSE4.2 -fp-model precise -g -debug inline_debug_info" --download-scalapack=1 --download-blacs=1 --with-blacs=1 --download-umfpack=1 --download-parmetis=1 --download-metis=1 --download-superlu=1 --download-superlu_dist=1 --download-mumps=1 --download-ml=1 --download-hypre=1 ----------------------------------------- Libraries compiled on Wed Oct 24 15:17:55 2012 on lagrange.tomato Machine characteristics: Linux-2.6.32-279.11.1.el6.x86_64-x86_64-with-centos-6.3-Final Using PETSc directory: /home/jfe/local/petsc-dev Using PETSc arch: intel-opt-precise-O3 ----------------------------------------- Using C compiler: /usr/local/encap/platform_mpi-8.02.01/bin/mpicc -fPIC -wd1572 -fPIC -O3 -xSSE4.2 -fp-model precise -g -debug inline_debug_info -fopenmp ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: /usr/local/encap/platform_mpi-8.02.01/bin/mpif90 -fPIC -fPIC -O3 -xSSE4.2 -fp-model precise -g -debug inline_debug_info -fopenmp ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/home/jfe/local/petsc-dev/intel-opt-precise-O3/include -I/home/jfe/local/petsc-dev/include -I/home/jfe/local/petsc-dev/include -I/home/jfe/local/petsc-dev/intel-opt-precise-O3/include -I/usr/local/encap/platform_mpi-8.02.01/include ----------------------------------------- Using C linker: /usr/local/encap/platform_mpi-8.02.01/bin/mpicc Using Fortran linker: /usr/local/encap/platform_mpi-8.02.01/bin/mpif90 Using libraries: -Wl,-rpath,/home/jfe/local/petsc-dev/intel-opt-precise-O3/lib -L/home/jfe/local/petsc-dev/intel-opt-precise-O3/lib -lpetsc -Wl,-rpath,/home/jfe/local/petsc-dev/intel-opt-precise-O3/lib -L/home/jfe/local/petsc-dev/intel-opt-precise-O3/lib -lcmumps -ldmumps -lsmumps -lzmumps -lmumps_common -lpord -lscalapack -lblacs -lml -Wl,-rpath,/usr/local/encap/platform_mpi-8.02.01/lib/linux_amd64 -L/usr/local/encap/platform_mpi-8.02.01/lib/linux_amd64 -lmpiCC -Wl,-rpath,/opt/intel/composer_xe_2011_sp1.7.256/compiler/lib/intel64 -L/opt/intel/composer_xe_2011_sp1.7.256/compiler/lib/intel64 -Wl,-rpath,/opt/intel/composer_xe_2011_sp1.7.256/ipp/lib/intel64 -L/opt/intel/composer_xe_2011_sp1.7.256/ipp/lib/intel64 -Wl,-rpath,/opt/intel/composer_xe_2011_sp1.7.256/mkl/lib/intel64 -L/opt/intel/composer_xe_2011_sp1.7.256/mkl/lib/intel64 -Wl,-rpath,/opt/intel/composer_xe_2011_sp1.7.256/tbb/lib/intel64/cc4.1.0_libc2.4_kernel2.6.16.21 -L/opt/intel/composer_xe_2011_sp1.7.256/tbb/lib/intel64/cc4.1.0_libc2.4_kernel2.6.16.21 -Wl,-rpath,/usr/lib/gcc/x86_64-redhat-linux/4.4.6 -L/usr/lib/gcc/x86_64-redhat-linux/4.4.6 -lpthread -lsuperlu_dist_3.1 -lparmetis -lmetis -lsuperlu_4.3 -lHYPRE -lmpiCC -lumfpack -lamd -Wl,-rpath,/opt/intel/composerxe/mkl/lib/intel64 -L/opt/intel/composerxe/mkl/lib/intel64 -lmkl_intel_lp64 -lmkl_intel_thread -lmkl_core -liomp5 -lpthread -lifport -lifcore -lm -lpthread -lm -lmpiCC -lpcmpio -lpcmpi -ldl -limf -lsvml -lipgo -ldecimal -lcilkrts -lstdc++ -lgcc_s -lirc -lirc_s -ldl ----------------------------------------- lid velocity = 0.0001, prandtl # = 1, grashof # = 1