0 KSP Residual norm 1.246700550132e+02 1 KSP Residual norm 3.447956334763e+00 2 KSP Residual norm 5.030225888925e+00 3 KSP Residual norm 5.469105333771e+00 4 KSP Residual norm 2.029637416966e+00 5 KSP Residual norm 1.724727768016e+00 6 KSP Residual norm 2.116757896001e+00 7 KSP Residual norm 2.697904346220e+00 8 KSP Residual norm 1.281088267855e+00 9 KSP Residual norm 1.155521577832e+00 10 KSP Residual norm 1.901999560252e+00 11 KSP Residual norm 1.145566371753e+00 12 KSP Residual norm 8.409714808901e-01 13 KSP Residual norm 9.865321754104e-01 14 KSP Residual norm 1.023257159042e+00 15 KSP Residual norm 5.794626593386e-01 16 KSP Residual norm 5.090627476551e-01 17 KSP Residual norm 6.911143857594e-01 18 KSP Residual norm 3.174524000919e-01 19 KSP Residual norm 2.244293481680e-01 20 KSP Residual norm 2.924278100963e-01 21 KSP Residual norm 1.356553921418e-01 22 KSP Residual norm 8.081998789437e-02 23 KSP Residual norm 9.268266199144e-02 24 KSP Residual norm 4.904798584637e-02 25 KSP Residual norm 2.811025132428e-02 26 KSP Residual norm 2.878644219434e-02 27 KSP Residual norm 1.941717551490e-02 28 KSP Residual norm 1.294103911484e-02 29 KSP Residual norm 1.266176401640e-02 30 KSP Residual norm 1.064449315404e-02 31 KSP Residual norm 7.033858163978e-03 32 KSP Residual norm 6.214423971050e-03 33 KSP Residual norm 5.232893106821e-03 34 KSP Residual norm 2.618780000076e-03 35 KSP Residual norm 2.017142348799e-03 36 KSP Residual norm 1.662705876404e-03 37 KSP Residual norm 9.417412914936e-04 KSP Object: type: cg maximum iterations=10000, initial guess is zero tolerances: relative=1e-05, absolute=1e-10, divergence=10000 left preconditioning using PRECONDITIONED norm type for convergence test PC Object: type: bjacobi block Jacobi: number of blocks = 4 Local solve is same for all blocks, in the following KSP and PC objects: KSP Object: (sub_) type: preonly maximum iterations=10000, initial guess is zero tolerances: relative=1e-05, absolute=1e-50, divergence=10000 left preconditioning using PRECONDITIONED norm type for convergence test PC Object: (sub_) type: lu LU: out-of-place factorization tolerance for zero pivot 1e-12 matrix ordering: nd factor fill ratio given 5, needed 7.60583 Factored matrix follows: Matrix Object: type: seqaij rows=16384, cols=16384 package used to perform factorization: petsc total: nonzeros=618202, allocated nonzeros=618202 total number of mallocs used during MatSetValues calls =0 not using I-node routines linear system matrix = precond matrix: Matrix Object: type: seqaij rows=16384, cols=16384 total: nonzeros=81280, allocated nonzeros=81920 total number of mallocs used during MatSetValues calls =0 not using I-node routines linear system matrix = precond matrix: Matrix Object: type: mpiaij rows=65536, cols=65536 total: nonzeros=326656, allocated nonzeros=655360 total number of mallocs used during MatSetValues calls =0 not using I-node (on process 0) routines Norm of error 0.0212011 iterations 37 ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ./ex2 on a linux-gnu named c0306 with 4 processors, by liluo Tue Jan 18 12:37:37 2011 Using Petsc Development HG revision: 179fe3d1768f57c49fa44a3a47095b573a99716c HG Date: Wed Dec 08 11:34:50 2010 -0600 Max Max/Min Avg Total Time (sec): 6.152e+00 1.00003 6.152e+00 Objects: 2.600e+01 1.00000 2.600e+01 Flops: 8.988e+07 1.00022 8.987e+07 3.595e+08 Flops/sec: 1.461e+07 1.00025 1.461e+07 5.843e+07 MPI Messages: 8.400e+01 1.90909 6.600e+01 2.640e+02 MPI Message Lengths: 1.577e+05 1.99980 1.792e+03 4.732e+05 MPI Reductions: 1.400e+02 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 6.1425e+00 99.8% 3.5948e+08 100.0% 2.520e+02 95.5% 1.769e+03 98.7% 1.170e+02 83.6% 1: Assembly: 9.2819e-03 0.2% 0.0000e+00 0.0% 1.200e+01 4.5% 2.336e+01 1.3% 9.000e+00 6.4% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %F - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %F %M %L %R %T %F %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage MatMult 38 1.0 1.7430e-02 1.2 5.59e+06 1.0 2.3e+02 2.0e+03 0.0e+00 0 6 86 99 0 0 6 90100 0 1281 MatSolve 38 1.0 8.8661e-02 1.3 4.64e+07 1.0 0.0e+00 0.0e+00 0.0e+00 1 52 0 0 0 1 52 0 0 0 2092 MatLUFactorSym 1 1.0 1.7892e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 0 0 0 0 1 0 0 0 0 1 0 MatLUFactorNum 1 1.0 4.8987e-02 1.0 3.06e+07 1.0 0.0e+00 0.0e+00 0.0e+00 1 34 0 0 0 1 34 0 0 0 2497 MatGetRowIJ 1 1.0 4.0603e-04 1.3 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetOrdering 1 1.0 1.5904e-02 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 1 0 0 0 0 2 0 MatView 3 3.0 8.7261e-05 3.6 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 0 0 0 0 1 0 0 0 0 1 0 VecDot 74 1.0 8.0454e-03 1.3 2.42e+06 1.0 0.0e+00 0.0e+00 7.4e+01 0 3 0 0 53 0 3 0 0 63 1206 VecNorm 39 1.0 3.0617e-02 2.9 1.28e+06 1.0 0.0e+00 0.0e+00 3.9e+01 0 1 0 0 28 0 1 0 0 33 167 VecCopy 154 1.0 4.2679e-03 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 40 1.0 3.6025e-04 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 75 1.0 5.2221e-03 1.0 2.46e+06 1.0 0.0e+00 0.0e+00 0.0e+00 0 3 0 0 0 0 3 0 0 0 1882 VecAYPX 36 1.0 1.9548e-03 1.1 1.18e+06 1.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 2414 VecScatterBegin 38 1.0 3.0875e-04 1.3 0.00e+00 0.0 2.3e+02 2.0e+03 0.0e+00 0 0 86 99 0 0 0 90100 0 0 VecScatterEnd 38 1.0 2.4414e-04 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSetup 2 1.0 1.7595e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 1 1.0 2.1532e-01 1.0 8.97e+07 1.0 2.2e+02 2.0e+03 1.2e+02 3100 84 96 82 4100 88 97 98 1666 PCSetUp 2 1.0 8.5682e-02 1.0 3.06e+07 1.0 0.0e+00 0.0e+00 3.0e+00 1 34 0 0 2 1 34 0 0 3 1428 PCSetUpOnBlocks 1 1.0 8.0713e-02 1.0 3.06e+07 1.0 0.0e+00 0.0e+00 3.0e+00 1 34 0 0 2 1 34 0 0 3 1516 PCApply 38 1.0 9.3607e-02 1.2 4.64e+07 1.0 0.0e+00 0.0e+00 0.0e+00 1 52 0 0 0 1 52 0 0 0 1981 --- Event Stage 1: Assembly MatAssemblyBegin 1 1.0 4.0507e-0413.9 0.00e+00 0.0 0.0e+00 0.0e+00 2.0e+00 0 0 0 0 1 3 0 0 0 22 0 MatAssemblyEnd 1 1.0 1.8611e-03 1.0 0.00e+00 0.0 1.2e+01 5.1e+02 7.0e+00 0 0 5 1 5 20 0100100 78 0 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Matrix 4 4 9919556 0 Vec 8 9 1063896 0 Vec Scatter 0 1 1012 0 Index Set 3 3 133232 0 Krylov Solver 2 2 2032 0 Preconditioner 2 2 1696 0 Viewer 2 2 1360 0 --- Event Stage 1: Assembly Vec 2 1 1496 0 Vec Scatter 1 0 0 0 Index Set 2 2 1432 0 ======================================================================================================================== Average time to get PetscTime(): 1.90735e-07 Average time for MPI_Barrier(): 5.00679e-06 Average time for zero size MPI_Send(): 7.7486e-07 #PETSc Option Table entries: -ksp_atol 1.e-10 -ksp_monitor -ksp_rtol 1.e-5 -ksp_type cg -ksp_view -log_summary -m 256 -mat_type mpiaij -n 256 -options_left -pc_type bjacobi -sub_ksp_type preonly -sub_pc_type lu -vec_type mpi #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 Configure run at: Mon Jan 17 21:18:44 2011 Configure options: --download-f-blas-lapack=1 --with-mpi-dir=/bwfs/software/mpich2-1.2.1p1 --with-shared-libraries=0 --with-debugging=no --with-cuda-dir=/bwfs/home/liluo/cuda3.2_64 --with-thrust-dir=/bwfs/home/liluo/cuda3.2_64/include/thrust --with-cusp-dir=/bwfs/home/liluo/cuda3.2_64/include/cusp-library ----------------------------------------- Libraries compiled on Mon Jan 17 21:18:44 2011 on console Machine characteristics: Linux-2.6.18-128.el5-x86_64-with-redhat-5.3-Tikanga Using PETSc directory: /bwfs/home/liluo/petsc-dev Using PETSc arch: linux-gnu-c-debug ----------------------------------------- Using C compiler: /bwfs/software/mpich2-1.2.1p1/bin/mpicc -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: /bwfs/software/mpich2-1.2.1p1/bin/mpif77 -Wall -Wno-unused-variable -O ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/bwfs/home/liluo/petsc-dev/linux-gnu-c-debug/include -I/bwfs/home/liluo/petsc-dev/include -I/bwfs/home/liluo/petsc-dev/include -I/bwfs/home/liluo/petsc-dev/linux-gnu-c-debug/include -I/bwfs/home/liluo/cuda3.2_64/include -I/bwfs/home/liluo/cuda3.2_64/include/cusp-library/ -I/bwfs/home/liluo/cuda3.2_64/include/thrust/ -I/bwfs/software/mpich2-1.2.1p1/include ----------------------------------------- Using C linker: /bwfs/software/mpich2-1.2.1p1/bin/mpicc Using Fortran linker: /bwfs/software/mpich2-1.2.1p1/bin/mpif77 Using libraries: -Wl,-rpath,/bwfs/home/liluo/petsc-dev/linux-gnu-c-debug/lib -L/bwfs/home/liluo/petsc-dev/linux-gnu-c-debug/lib -lpetscts -lpetscsnes -lpetscksp -lpetscdm -lpetscmat -lpetscvec -lpetscsys -lX11 -Wl,-rpath,/bwfs/home/liluo/cuda3.2_64/lib64 -L/bwfs/home/liluo/cuda3.2_64/lib64 -lcufft -lcublas -lcudart -Wl,-rpath,/bwfs/home/liluo/petsc-dev/linux-gnu-c-debug/lib -L/bwfs/home/liluo/petsc-dev/linux-gnu-c-debug/lib -lflapack -lfblas -L/bwfs/software/mpich2-1.2.1p1/lib -L/usr/lib/gcc/x86_64-redhat-linux/4.1.2 -ldl -lmpich -lopa -lpthread -lrt -lgcc_s -lg2c -lm -L/usr/lib/gcc/x86_64-redhat-linux/3.4.6 -lm -lmpichcxx -lstdc++ -lmpichcxx -lstdc++ -ldl -lmpich -lopa -lpthread -lrt -lgcc_s -ldl ----------------------------------------- #PETSc Option Table entries: -ksp_atol 1.e-10 -ksp_monitor -ksp_rtol 1.e-5 -ksp_type cg -ksp_view -log_summary -m 256 -mat_type mpiaij -n 256 -options_left -pc_type bjacobi -sub_ksp_type preonly -sub_pc_type lu -vec_type mpi #End of PETSc Option Table entries There are no unused options.