proc:1 (i,j,k)=(0,0,44) : (0,1,2,3,4,5,6,7)=(89100,89101,89145,89146,0,1,45,46) proc:0 (i,j,k)=(0,0,0) : (0,1,2,3,4,5,6,7)=(0,1,45,46,2025,2026,2070,2071) proc:1 (i,j,k)=(0,44,44) : (0,1,2,3,4,5,6,7)=(91080,91081,89100,89101,1980,1981,0,1) proc:0 (i,j,k)=(0,44,0) : (0,1,2,3,4,5,6,7)=(1980,1981,0,1,4005,4006,2025,2026) proc:1 (i,j,k)=(44,0,44) : (0,1,2,3,4,5,6,7)=(89144,89100,89189,89145,44,0,89,45) proc:1 (i,j,k)=(44,44,44) : (0,1,2,3,4,5,6,7)=(91124,91080,89144,89100,2024,1980,44,0) proc:0 (i,j,k)=(44,0,0) : (0,1,2,3,4,5,6,7)=(44,0,89,45,2069,2025,2114,2070) proc:0 (i,j,k)=(44,44,0) : (0,1,2,3,4,5,6,7)=(2024,1980,44,0,4049,4005,2069,2025) RAMmonitor: KSP_Converged(): Linear solver has converged. Residual norm 6.219287e-06 is less than absolute tolerance 1.000000e-05 at Iteration 729 KSP Object: type: cgs maximum iterations=10000, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using PRECONDITIONED norm type for convergence test PC Object: type: asm Additive Schwarz: total subdomain blocks = 2, amount of overlap = 1 Additive Schwarz: restriction/interpolation type - RESTRICT Local solve is same for all blocks, in the following KSP and PC objects: KSP Object:(sub_) type: preonly maximum iterations=10000, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using PRECONDITIONED norm type for convergence test PC Object:(sub_) type: jacobi linear system matrix = precond matrix: Matrix Object: type=seqaij, rows=50625, cols=50625 total: nonzeros=1330425, allocated nonzeros=1330425 not using I-node routines linear system matrix = precond matrix: Matrix Object: type=mpiaij, rows=91125, cols=91125 total: nonzeros=2460375, allocated nonzeros=2460375 not using I-node (on process 0) routines Residual Norm: 0.000006219287442364 ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./RAM_Main on a linux-gnu named swetaketo-pc with 2 processors, by swetaketo Mon Aug 22 15:03:45 2011 Using Petsc Release Version 3.1.0, Patch 8, Thu Mar 17 13:37:48 CDT 2011 Max Max/Min Avg Total Time (sec): 3.678e+01 1.00003 3.678e+01 Objects: 1.509e+03 1.00000 1.509e+03 Flops: 6.227e+09 1.04541 6.092e+09 1.218e+10 Flops/sec: 1.693e+08 1.04544 1.656e+08 3.312e+08 Memory: 3.155e+08 1.04494 6.174e+08 MPI Messages: 3.664e+03 1.00000 3.664e+03 7.329e+03 MPI Message Lengths: 1.207e+08 1.00000 3.293e+04 2.413e+08 MPI Reductions: 7.377e+03 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 6.4741e-01 1.8% 9.9326e+06 0.1% 1.900e+01 0.3% 3.099e+02 0.9% 3.900e+01 0.5% 1: Iteration :: 3.6135e+01 98.2% 1.2174e+10 99.9% 7.310e+03 99.7% 3.262e+04 99.1% 6.587e+03 89.3% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %F - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ ########################################################## # # # WARNING!!! # # # # This code was compiled with a debugging option, # # To get timing results run config/configure.py # # using --with-debugging=no, the performance will # # be generally two or three times faster. # # # ########################################################## Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %F %M %L %R %T %F %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage VecNorm 1 1.0 7.7212e-0312.2 9.32e+04 1.0 0.0e+00 0.0e+00 1.0e+00 0 0 0 0 0 1 2 0 0 3 24 VecSet 1 1.0 7.7009e-05 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecWAXPY 1 1.0 2.3198e-04 1.1 4.66e+04 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 1 0 0 0 393 VecScatterBegin 2 1.0 1.0729e-04 1.1 0.00e+00 0.0 4.0e+00 3.2e+04 0.0e+00 0 0 0 0 0 0 0 21 6 0 0 VecScatterEnd 2 1.0 1.0452e-03 7.6 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSetRandom 1 1.0 2.6040e-03 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatMult 2 1.0 1.8194e-02 1.1 4.94e+06 1.0 4.0e+00 3.2e+04 0.0e+00 0 0 0 0 0 3 97 21 6 0 531 MatAssemblyBegin 2 1.0 8.9838e-02 9.7 0.00e+00 0.0 6.0e+00 3.5e+05 4.0e+00 0 0 0 1 0 8 0 32 91 10 0 MatAssemblyEnd 2 1.0 3.8163e-02 1.0 0.00e+00 0.0 4.0e+00 8.1e+03 1.1e+01 0 0 0 0 0 6 0 21 1 28 0 --- Event Stage 1: Iteration : VecDot 1459 1.0 1.7151e+00 1.0 1.36e+08 1.0 0.0e+00 0.0e+00 1.5e+03 5 2 0 0 20 5 2 0 0 22 155 VecNorm 1460 1.0 2.5229e+00 1.1 1.36e+08 1.0 0.0e+00 0.0e+00 1.5e+03 7 2 0 0 20 7 2 0 0 22 105 VecCopy 734 1.0 1.8859e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 0 VecSet 2920 1.0 3.8071e-01 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 0 VecAXPY 2186 1.0 5.0222e-01 1.2 2.04e+08 1.0 0.0e+00 0.0e+00 0.0e+00 1 3 0 0 0 1 3 0 0 0 793 VecAYPX 730 1.0 1.6733e-01 1.0 3.40e+07 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 398 VecWAXPY 2914 1.0 8.5323e-01 1.1 2.37e+08 1.0 0.0e+00 0.0e+00 0.0e+00 2 4 0 0 0 2 4 0 0 0 545 VecPointwiseMult 1459 1.0 4.0172e-01 1.0 7.39e+07 1.0 0.0e+00 0.0e+00 0.0e+00 1 1 0 0 0 1 1 0 0 0 360 VecScatterBegin 5106 1.0 9.2218e-01 1.0 0.00e+00 0.0 7.3e+03 3.2e+04 0.0e+00 2 0100 98 0 3 0100 99 0 0 VecScatterEnd 5106 1.0 9.9554e-01 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 2 0 0 0 0 3 0 0 0 0 0 MatMult 2188 1.0 2.3844e+01 1.0 5.40e+09 1.0 4.4e+03 3.2e+04 0.0e+00 64 87 60 59 0 65 87 60 59 0 443 MatAssemblyBegin 1 1.0 2.1458e-06 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyEnd 1 1.0 7.3359e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetSubMatrice 1 1.0 9.7609e-02 1.0 0.00e+00 0.0 1.0e+01 2.7e+05 5.0e+00 0 0 0 1 0 0 0 0 1 0 0 MatIncreaseOvrlp 1 1.0 4.2769e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatView 2 2.0 7.7009e-05 1.3 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSetup 2 1.0 2.5251e-03 1.6 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 1 1.0 3.6133e+01 1.0 6.22e+09 1.0 7.3e+03 3.3e+04 6.6e+03 98100100 99 89 100100100100100 337 PCSetUp 2 1.0 1.5754e-01 1.0 0.00e+00 0.0 1.4e+01 1.9e+05 1.5e+01 0 0 0 1 0 0 0 0 1 0 0 PCSetUpOnBlocks 1 1.0 9.0599e-06 1.3 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 PCApply 1459 1.0 5.4268e+00 1.1 7.39e+07 1.0 2.9e+03 3.2e+04 2.9e+03 14 1 40 39 40 14 1 40 40 44 27 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Application Order 1 0 0 0 Distributed array 1 0 0 0 Vec 8 18 5369872 0 Vec Scatter 3 2 1736 0 Index Set 8 9 816872 0 IS L to G Mapping 1 0 0 0 Matrix 3 4 33380784 0 Krylov Solver 1 2 1664 0 Preconditioner 1 2 1288 0 PetscRandom 1 1 448 0 Viewer 0 1 544 0 --- Event Stage 1: Iteration : Vec 1472 731 273353064 0 Vec Scatter 1 0 0 0 Index Set 3 2 1024 0 Matrix 1 0 0 0 Krylov Solver 1 0 0 0 Preconditioner 1 0 0 0 Viewer 2 1 544 0 ======================================================================================================================== Average time to get PetscTime(): 9.05991e-07 Average time for MPI_Barrier(): 1.20163e-05 Average time for zero size MPI_Send(): 4.60148e-05 #PETSc Option Table entries: -ksp_type cgs -ksp_view -log_summary -pc_type asm -sub_pc_type jacobi #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 Configure run at: Sun Aug 14 11:19:34 2011 Configure options: --download-mpich --download-c-blas-lapack=1 ----------------------------------------- Libraries compiled on Ð