KSP Object: 1 MPI processes type: cg maximum iterations=10000, initial guess is zero tolerances: relative=1e-08, absolute=1e-50, divergence=10000 left preconditioning using PRECONDITIONED norm type for convergence test PC Object: 1 MPI processes type: jacobi linear system matrix = precond matrix: Matrix Object: 1 MPI processes type: seqaijcusp rows=10240000, cols=10240000 total: nonzeros=51187200, allocated nonzeros=51200000 total number of mallocs used during MatSetValues calls =0 not using I-node routines Total flops over all processors 1.25016e+12 ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ex2f on a LINUX_OPE named lamyridge.trialphaenergy.com with 1 processor, by wdn Mon Apr 4 21:12:01 2011 Using Petsc Development HG revision: HG Date: Max Max/Min Avg Total Time (sec): 1.244e+02 1.00000 1.244e+02 Objects: 1.100e+01 1.00000 1.100e+01 Flops: 1.250e+12 1.00000 1.250e+12 1.250e+12 Flops/sec: 1.005e+10 1.00000 1.005e+10 1.005e+10 MPI Messages: 0.000e+00 0.00000 0.000e+00 0.000e+00 MPI Message Lengths: 0.000e+00 0.00000 0.000e+00 0.000e+00 MPI Reductions: 1.100e+01 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 1.2441e+02 100.0% 1.2502e+12 100.0% 0.000e+00 0.0% 0.000e+00 0.0% 1.000e+00 9.1% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %F - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %F %M %L %R %T %F %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage MatMult 5309 1.0 6.4881e+01 1.0 5.44e+11 1.0 0.0e+00 0.0e+00 0.0e+00 52 43 0 0 0 52 43 0 0 0 8377 MatAssemblyBegin 1 1.0 9.5367e-07 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyEnd 1 1.0 2.4281e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatView 1 1.0 4.1008e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatCUSPCopyTo 1 1.0 1.1661e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecDot 10616 1.0 1.3077e+01 1.0 2.17e+11 1.0 0.0e+00 0.0e+00 0.0e+00 11 17 0 0 0 11 17 0 0 0 16626 VecNorm 5310 1.0 3.7947e+00 1.0 1.09e+11 1.0 0.0e+00 0.0e+00 0.0e+00 3 9 0 0 0 3 9 0 0 0 28658 VecCopy 2 1.0 4.2889e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 3 1.0 4.2930e-03 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 10617 1.0 1.7962e+01 1.0 2.17e+11 1.0 0.0e+00 0.0e+00 0.0e+00 14 17 0 0 0 14 17 0 0 0 12106 VecAYPX 5307 1.0 8.9743e+00 1.0 1.09e+11 1.0 0.0e+00 0.0e+00 0.0e+00 7 9 0 0 0 7 9 0 0 0 12111 VecPointwiseMult 5309 1.0 9.0009e+00 1.0 5.44e+10 1.0 0.0e+00 0.0e+00 0.0e+00 7 4 0 0 0 7 4 0 0 0 6040 VecCUSPCopyTo 1 1.0 1.4069e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecCUSPCopyFrom 1 1.0 3.3111e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSetup 1 1.0 1.6928e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 1 1.0 1.1774e+02 1.0 1.25e+12 1.0 0.0e+00 0.0e+00 1.0e+00 95100 0 0 9 95100 0 0100 10616 PCSetUp 1 1.0 9.5367e-07 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 PCApply 5309 1.0 9.1642e+00 1.0 5.44e+10 1.0 0.0e+00 0.0e+00 1.0e+00 7 4 0 0 9 7 4 0 0100 5932 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Matrix 1 1 778242492 0 Vec 7 7 163850168 0 Krylov Solver 1 1 1064 0 Preconditioner 1 1 768 0 Viewer 1 0 0 0 ======================================================================================================================== Average time to get PetscTime(): 1.90735e-07 #PETSc Option Table entries: -get_total_flops -ksp_rtol 1.0e-8 -ksp_type cg -ksp_view -log_summary -m 3200 -mat_type aijcusp -n 3200 -options_left -pc_type jacobi -vec_type cusp #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 Configure run at: Mon Apr 4 20:21:47 2011 Configure options: --PETSC_DIR=/home/wdn/Packages/Petsc/petsc_dev_2011_04_04_hg --PETSC_ARCH=LINUX_OPENMPI_GNU_OPTIMIZE --with-mpi-dir=/home/wdn/Packages/OpenMPI/openmpi_1.5.3/LINUX_GNU_OPTIMIZE --download-f-blas-lapack=yes --download-plapack=yes --download-blacs=yes --download-spai=yes --download-scalapack=yes --download-hdf5=yes --download-numpy=yes --download-hypre=yes --download-ml=yes --download-sundials=yes --with-cuda=1 --with-cuda-dir=/home/wdn/Packages/Cuda/Cuda_3.2.16/cuda --with-thrust=1 --with-thrust-dir=/home/wdn/Packages/Cuda/Cuda_3.2.16/cuda --with-cusp=1 --with-cusp-dir=/home/wdn/Packages/Cuda/Cuda_3.2.16/cuda --with-shared-libraries --with-debugging=no --with-64-bit-pointers=1 --with-large-file-io=1 ----------------------------------------- Libraries compiled on Mon Apr 4 20:21:47 2011 on lamyridge.trialphaenergy.com Machine characteristics: Linux-2.6.18-164.el5-x86_64-with-redhat-5.4-Final Using PETSc directory: /home/wdn/Packages/Petsc/petsc_dev_2011_04_04_hg Using PETSc arch: LINUX_OPENMPI_GNU_OPTIMIZE ----------------------------------------- Using C compiler: /home/wdn/Packages/OpenMPI/openmpi_1.5.3/LINUX_GNU_OPTIMIZE/bin/mpicc -fPIC -D_LARGEFILE64_SOURCE -D_FILE_OFFSET_BITS=64 -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: /home/wdn/Packages/OpenMPI/openmpi_1.5.3/LINUX_GNU_OPTIMIZE/bin/mpif90 -fPIC -Wall -Wno-unused-variable -O ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/home/wdn/Packages/Petsc/petsc_dev_2011_04_04_hg/LINUX_OPENMPI_GNU_OPTIMIZE/include -I/home/wdn/Packages/Petsc/petsc_dev_2011_04_04_hg/include -I/home/wdn/Packages/Petsc/petsc_dev_2011_04_04_hg/include -I/home/wdn/Packages/Petsc/petsc_dev_2011_04_04_hg/LINUX_OPENMPI_GNU_OPTIMIZE/include -I/home/wdn/Packages/Cuda/Cuda_3.2.16/cuda/include -I/home/wdn/Packages/Cuda/Cuda_3.2.16/cuda/ -I/home/wdn/Packages/Petsc/petsc_dev_2011_04_04_hg/LINUX_OPENMPI_GNU_OPTIMIZE/lib -I/home/wdn/Packages/OpenMPI/openmpi_1.5.3/LINUX_GNU_OPTIMIZE/include ----------------------------------------- Using C linker: /home/wdn/Packages/OpenMPI/openmpi_1.5.3/LINUX_GNU_OPTIMIZE/bin/mpicc Using Fortran linker: /home/wdn/Packages/OpenMPI/openmpi_1.5.3/LINUX_GNU_OPTIMIZE/bin/mpif90 Using libraries: -Wl,-rpath,/home/wdn/Packages/Petsc/petsc_dev_2011_04_04_hg/LINUX_OPENMPI_GNU_OPTIMIZE/lib -L/home/wdn/Packages/Petsc/petsc_dev_2011_04_04_hg/LINUX_OPENMPI_GNU_OPTIMIZE/lib -lpetscts -lpetscsnes -lpetscksp -lpetscdm -lpetscmat -lpetscvec -lpetscsys -lX11 -Wl,-rpath,/home/wdn/Packages/Petsc/petsc_dev_2011_04_04_hg/LINUX_OPENMPI_GNU_OPTIMIZE/lib -L/home/wdn/Packages/Petsc/petsc_dev_2011_04_04_hg/LINUX_OPENMPI_GNU_OPTIMIZE/lib -lPLAPACK -lscalapack -lblacs -lspai -Wl,-rpath,/home/wdn/Packages/Cuda/Cuda_3.2.16/cuda/lib64 -L/home/wdn/Packages/Cuda/Cuda_3.2.16/cuda/lib64 -lcufft -lcublas -lcudart -lHYPRE -Wl,-rpath,/home/wdn/Packages/OpenMPI/openmpi_1.5.3/LINUX_GNU_OPTIMIZE/lib -Wl,-rpath,/usr/lib/gcc/x86_64-redhat-linux/4.1.2 -lmpi_cxx -lstdc++ -lml -lmpi_cxx -lstdc++ -lsundials_cvode -lsundials_nvecserial -lsundials_nvecparallel -lflapack -lfblas -lhdf5_fortran -lhdf5 -lz -L/home/wdn/Packages/OpenMPI/openmpi_1.5.3/LINUX_GNU_OPTIMIZE/lib -L/usr/lib/gcc/x86_64-redhat-linux/4.1.2 -ldl -lmpi -lnsl -lutil -lgcc_s -lpthread -lmpi_f90 -lmpi_f77 -lgfortran -lm -lm -lm -lm -lmpi_cxx -lstdc++ -lmpi_cxx -lstdc++ -ldl -lmpi -lnsl -lutil -lgcc_s -lpthread -ldl ----------------------------------------- #PETSc Option Table entries: -get_total_flops -ksp_rtol 1.0e-8 -ksp_type cg -ksp_view -log_summary -m 3200 -mat_type aijcusp -n 3200 -options_left -pc_type jacobi -vec_type cusp #End of PETSc Option Table entries There are no unused options. Norm of error 0.5317E-03 iterations 5308 real 124.58 user 119.68 sys 4.83