rank-1: finished creating parallel vectors and matrices Total Memory requrements for local variables is 17416 bytesWhich is 1.6609e-02 MB name-teb10 fname-./meshData/teb10 # Grid data for a very simple unstructured grid of tetrhedrals Number of grid vertices 44 Number of Surface vertices = 44 Processor 0 assigned 22 vertices Processor 1 assigned 22 vertices Vertices assigned to processor 0 Vertices assigned to processor 1 Number of grid elements 60 Processor 0 assigned 30 elements Processor 1 assigned 30 elements Elements assigned to processor 0 Elements assigned to processor 1 Element neighbors on processor 0 Element neighbors on processor 1 Reading node-to-element mapping for processor 0 Reading node-to-element mapping for processor 1 rank-0: finished creating parallel vectors and matrices Total Memory requrements for local variables is 16616 bytesWhich is 1.5846e-02 MB name-teb10 fname-./meshData/teb10 Convergence for initial acceleration 8 1 time = 9.000e-03, deltaT = 9.0000e-03. Convergence in:: 22 priting results data time for one timestep = 1.1945009E-02 2 time = 1.800e-02, deltaT = 9.0000e-03. Convergence in:: 22 priting results data Convergence for initial acceleration 8 time for one timestep = 9.9990368E-03 3 time = 2.700e-02, deltaT = 9.0000e-03. Convergence in:: 22 priting results data time for one timestep = 6.3920021E-03 4 time = 3.600e-02, deltaT = 9.0000e-03. Convergence in:: 22 priting results data time for one timestep = 6.5050125E-03 5 time = 4.500e-02, deltaT = 9.0000e-03. Convergence in:: 22 priting results data time for one timestep = 6.4020157E-03 6 time = 5.400e-02, deltaT = 9.0000e-03. Convergence in:: 22 priting results data time for one timestep = 6.6390038E-03 7 time = 6.300e-02, deltaT = 9.0000e-03. Convergence in:: 22 priting results data time for one timestep = 6.5300465E-03 8 time = 7.200e-02, deltaT = 9.0000e-03. Convergence in:: 22 priting results data time for one timestep = 6.5319538E-03 9 time = 8.100e-02, deltaT = 9.0000e-03. Convergence in:: 22 priting results data time for one timestep = 6.5290928E-03 10 time = 9.000e-02, deltaT = 9.0000e-03. Convergence in:: 22 priting results data time for one timestep = 6.6599846E-03 11 time = 9.900e-02, deltaT = 9.0000e-03. Convergence in:: 22 priting results data time for one timestep = 6.5460205E-03 12 time = 1.080e-01, deltaT = 9.0000e-03. Convergence in:: 22 priting results data time for one timestep = 6.6230297E-03 13 time = 1.170e-01, deltaT = 9.0000e-03. Convergence in:: 21 priting results data time for one timestep = 6.4430237E-03 14 time = 1.260e-01, deltaT = 9.0000e-03. Convergence in:: 21 priting results data time for one timestep = 6.3991547E-03 15 time = 1.350e-01, deltaT = 9.0000e-03. Convergence in:: 21 priting results data time for one timestep = 9.0551376E-03 16 time = 1.440e-01, deltaT = 9.0000e-03. Convergence in:: 21 priting results data time for one timestep = 1.0615110E-02 17 time = 1.530e-01, deltaT = 9.0000e-03. Convergence in:: 21 priting results data time for one timestep = 1.1027098E-02 18 time = 1.620e-01, deltaT = 9.0000e-03. Convergence in:: 21 priting results data time for one timestep = 1.0687113E-02 19 time = 1.710e-01, deltaT = 9.0000e-03. Convergence in:: 21 priting results data time for one timestep = 1.1082172E-02 total Time taken= 1.8102503E-01 ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./te on a linux-mpi named ptb341 with 2 processors, by irfan Tue Nov 24 19:42:56 2009 Using Petsc Release Version 3.0.0, Patch 7, Mon Jul 6 11:33:34 CDT 2009 Max Max/Min Avg Total Time (sec): 1.814e-01 1.00092 1.813e-01 Objects: 1.300e+02 1.00000 1.300e+02 Flops: 2.840e+06 1.21412 2.590e+06 5.179e+06 Flops/sec: 1.566e+07 1.21301 1.429e+07 2.857e+07 Memory: 3.979e+05 1.03485 7.823e+05 MPI Messages: 6.060e+02 1.00165 6.055e+02 1.211e+03 MPI Message Lengths: 9.962e+04 1.00048 1.645e+02 1.992e+05 MPI Reductions: 2.082e+03 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 1.8126e-01 100.0% 5.1791e+06 100.0% 1.211e+03 100.0% 1.645e+02 100.0% 2.010e+03 96.5% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %F - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ ########################################################## # # # WARNING!!! # # # # This code was compiled with a debugging option, # # To get timing results run config/configure.py # # using --with-debugging=no, the performance will # # be generally two or three times faster. # # # ########################################################## Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %F %M %L %R %T %F %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage Read Data 1 1.0 3.0303e-04 1.0 0.00e+00 0.0 5.0e+00 3.2e+02 0.0e+00 0 0 0 1 0 0 0 0 1 0 0 Partition elemen 1 1.0 1.2269e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 1 0 0 0 0 1 0 0 0 0 0 Move elements 1 1.0 6.2013e-04 1.0 0.00e+00 0.0 6.0e+00 4.8e+02 1.1e+01 0 0 0 1 1 0 0 0 1 1 0 Partition vertic 1 1.0 8.8215e-06 1.1 0.00e+00 0.0 1.0e+00 6.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 Move vertices 1 1.0 6.2895e-04 1.0 0.00e+00 0.0 1.2e+01 1.2e+02 1.8e+01 0 0 1 1 1 0 0 1 1 1 0 set var localdof 1 1.0 6.9141e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatMult 457 1.0 5.4289e-02 1.0 1.28e+06 1.2 8.8e+02 1.7e+02 0.0e+00 30 45 72 74 0 30 45 72 74 0 43 MatSolve 439 1.0 5.1858e-03 1.1 1.17e+06 1.3 0.0e+00 0.0e+00 0.0e+00 3 40 0 0 0 3 40 0 0 0 403 MatLUFactorNum 2 1.0 3.1686e-04 1.2 2.97e+04 1.3 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 165 MatILUFactorSym 2 1.0 2.1791e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 6.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyBegin 11 1.0 6.2680e-03 1.0 0.00e+00 0.0 1.2e+01 1.2e+03 1.8e+01 3 0 1 7 1 3 0 1 7 1 0 MatAssemblyEnd 11 1.0 2.5728e-03 1.0 0.00e+00 0.0 1.2e+01 4.4e+01 5.4e+01 1 0 1 0 3 1 0 1 0 3 0 MatGetRow 330 1.0 3.9530e-04 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetRowIJ 2 1.0 3.8147e-06 1.3 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetOrdering 2 1.0 3.4308e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 8.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatPartitioning 1 1.0 5.8913e-04 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatZeroEntries 2 1.0 2.9087e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecDot 838 1.0 3.8185e-03 1.4 1.10e+05 1.0 0.0e+00 0.0e+00 8.4e+02 2 4 0 0 40 2 4 0 0 42 57 VecNorm 439 1.0 3.2668e-03 1.2 5.79e+04 1.0 0.0e+00 0.0e+00 4.4e+02 2 2 0 0 21 2 2 0 0 22 35 VecCopy 78 1.0 7.3433e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 546 1.0 4.0007e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 1066 1.0 1.0393e-03 1.0 1.39e+05 1.0 0.0e+00 0.0e+00 0.0e+00 1 5 0 0 0 1 5 0 0 0 268 VecAYPX 399 1.0 4.2701e-04 1.1 5.27e+04 1.0 0.0e+00 0.0e+00 0.0e+00 0 2 0 0 0 0 2 0 0 0 247 VecSwap 57 1.0 6.3896e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecScatterBegin 597 1.0 1.2872e-03 1.0 0.00e+00 0.0 1.2e+03 1.6e+02 0.0e+00 1 0 95 91 0 1 0 95 91 0 0 VecScatterEnd 597 1.0 1.3673e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 0 KSPSetup 4 1.0 3.1209e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 20 1.0 7.4134e-02 1.0 2.75e+06 1.2 8.4e+02 1.7e+02 1.8e+03 41 97 69 71 84 41 97 69 71 87 68 PCSetUp 4 1.0 1.9372e-03 1.0 2.97e+04 1.3 0.0e+00 0.0e+00 1.8e+01 1 1 0 0 1 1 1 0 0 1 27 PCSetUpOnBlocks 20 1.0 1.4279e-03 1.0 2.97e+04 1.3 0.0e+00 0.0e+00 1.4e+01 1 1 0 0 1 1 1 0 0 1 37 PCApply 439 1.0 1.3052e-02 1.0 1.17e+06 1.3 0.0e+00 0.0e+00 4.4e+02 7 40 0 0 21 7 40 0 0 22 160 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. --- Event Stage 0: Main Stage Matrix 16 16 153344 0 Matrix Partitioning 1 1 472 0 Index Set 35 35 20912 0 IS L to G Mapping 4 3 2636 0 Vec 54 44 70352 0 Vec Scatter 11 10 8680 0 Application Order 1 1 824 0 Krylov Solver 4 4 3440 0 Preconditioner 4 4 2816 0 ======================================================================================================================== Average time to get PetscTime(): 9.53674e-08 Average time for MPI_Barrier(): 3.00407e-06 Average time for zero size MPI_Send(): 4.05312e-06 #PETSc Option Table entries: -ksp_type cg -log_summary -mat_partitioning_type parmetis #End o PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 Configure run at: Tue Nov 17 12:30:02 2009 Configure options: --download-c-blas-lapack=1 --with-mpi-dir=/usr/local/mpich2.12 --download-parmetis=1 --with-shared=0 ----------------------------------------- Libraries compiled on Tue Nov 17 12:31:07 EST 2009 on ptb341 Machine characteristics: Linux ptb341 2.6.24-25-generic #1 SMP Tue Oct 20 06:49:12 UTC 2009 x86_64 GNU/Linux Using PETSc directory: /home/irfan/downloads/petsc-3.0.0-p7 Using PETSc arch: linux-mpich2-debug ----------------------------------------- Using C compiler: /usr/local/mpich2.12/bin/mpicc -Wall -Wwrite-strings -Wno-strict-aliasing -g3 Using Fortran compiler: ----------------------------------------- Using include paths: -I/home/irfan/downloads/petsc-3.0.0-p7/linux-mpich2-debug/include -I/home/irfan/downloads/petsc-3.0.0-p7/include -I/home/irfan/downloads/petsc-3.0.0-p7/linux-mpich2-debug/include -I/usr/local/mpich2.12/include ------------------------------------------ Using C linker: /usr/local/mpich2.12/bin/mpicc -Wall -Wwrite-strings -Wno-strict-aliasing -g3 Using Fortran linker: Using libraries: -Wl,-rpath,/home/irfan/downloads/petsc-3.0.0-p7/linux-mpich2-debug/lib -L/home/irfan/downloads/petsc-3.0.0-p7/linux-mpich2-debug/lib -lpetscts -lpetscsnes -lpetscksp -lpetscdm -lpetscmat -lpetscvec -lpetsc -Wl,-rpath,/home/irfan/downloads/petsc-3.0.0-p7/linux-mpich2-debug/lib -L/home/irfan/downloads/petsc-3.0.0-p7/linux-mpich2-debug/lib -lf2clapack -lf2cblas -lm -lparmetis -lmetis -lnsl -lrt -lm -ldl ------------------------------------------