0 KSP Residual norm 4.681784098361e+07 1 KSP Residual norm 3.237900921572e+05 2 KSP Residual norm 3.755574788135e+04 Linear solve converged due to CONVERGED_RTOL iterations 2 0 KSP Residual norm 3.755574788135e+04 1 KSP Residual norm 6.031582602297e+03 2 KSP Residual norm 9.328682255016e+02 3 KSP Residual norm 3.046116191337e+02 4 KSP Residual norm 6.012611302271e+01 Linear solve converged due to CONVERGED_RTOL iterations 4 0 SNES Function norm 3.762861669733e+07 0 KSP Residual norm 3.762861669733e+07 1 KSP Residual norm 9.829885482769e+06 2 KSP Residual norm 2.425056009630e+06 3 KSP Residual norm 6.482630068187e+05 4 KSP Residual norm 1.700293223428e+05 5 KSP Residual norm 4.484604400044e+04 Linear solve converged due to CONVERGED_RTOL iterations 5 1 SNES Function norm 1.322151394202e+07 0 KSP Residual norm 1.322151394202e+07 1 KSP Residual norm 3.871529369898e+06 2 KSP Residual norm 1.532701540285e+06 3 KSP Residual norm 6.924848040040e+05 4 KSP Residual norm 3.693340216365e+05 5 KSP Residual norm 2.169325470893e+05 6 KSP Residual norm 1.274109917825e+05 7 KSP Residual norm 7.454549464625e+04 8 KSP Residual norm 4.279583060485e+04 9 KSP Residual norm 2.402451056294e+04 Linear solve converged due to CONVERGED_RTOL iterations 9 2 SNES Function norm 1.005394806000e+07 0 KSP Residual norm 1.005394806000e+07 1 KSP Residual norm 4.140687923671e+06 2 KSP Residual norm 1.592405127155e+06 3 KSP Residual norm 6.944797835082e+05 4 KSP Residual norm 3.753725581267e+05 5 KSP Residual norm 2.419057096775e+05 6 KSP Residual norm 1.754496239002e+05 7 KSP Residual norm 1.317480581875e+05 8 KSP Residual norm 9.619142893051e+04 9 KSP Residual norm 6.900420654686e+04 10 KSP Residual norm 4.989367211850e+04 11 KSP Residual norm 3.598531582505e+04 12 KSP Residual norm 2.531362329202e+04 13 KSP Residual norm 1.747931426387e+04 Linear solve converged due to CONVERGED_RTOL iterations 13 3 SNES Function norm 6.133282541910e+06 0 KSP Residual norm 6.133282541910e+06 1 KSP Residual norm 2.493342184545e+06 2 KSP Residual norm 9.199355257469e+05 3 KSP Residual norm 4.081074269254e+05 4 KSP Residual norm 2.032200795695e+05 5 KSP Residual norm 1.049764375299e+05 6 KSP Residual norm 5.603396212532e+04 7 KSP Residual norm 3.232591480800e+04 8 KSP Residual norm 2.240823160757e+04 9 KSP Residual norm 1.766699766606e+04 10 KSP Residual norm 1.424743128888e+04 11 KSP Residual norm 1.089182219437e+04 Linear solve converged due to CONVERGED_RTOL iterations 11 4 SNES Function norm 2.825639461880e+06 0 KSP Residual norm 2.825639461880e+06 1 KSP Residual norm 9.382640730620e+05 2 KSP Residual norm 3.956170845326e+05 3 KSP Residual norm 1.742607300353e+05 4 KSP Residual norm 8.348030585981e+04 5 KSP Residual norm 4.274551622984e+04 6 KSP Residual norm 2.576273290470e+04 7 KSP Residual norm 1.900557963501e+04 8 KSP Residual norm 1.550203010837e+04 9 KSP Residual norm 1.258483435163e+04 10 KSP Residual norm 9.532852196020e+03 11 KSP Residual norm 6.841973420699e+03 12 KSP Residual norm 4.975700693559e+03 Linear solve converged due to CONVERGED_RTOL iterations 12 5 SNES Function norm 2.378016641661e+06 0 KSP Residual norm 2.378016641661e+06 1 KSP Residual norm 8.045370200521e+05 2 KSP Residual norm 3.112608707669e+05 3 KSP Residual norm 1.329115588310e+05 4 KSP Residual norm 6.468257911964e+04 5 KSP Residual norm 3.341437566662e+04 6 KSP Residual norm 1.806970214660e+04 7 KSP Residual norm 1.037888207210e+04 8 KSP Residual norm 6.159036490587e+03 9 KSP Residual norm 3.899873517475e+03 Linear solve converged due to CONVERGED_RTOL iterations 9 6 SNES Function norm 1.147418133547e+06 0 KSP Residual norm 1.147418133547e+06 1 KSP Residual norm 4.603120981921e+05 2 KSP Residual norm 1.655974960879e+05 3 KSP Residual norm 6.595764910690e+04 4 KSP Residual norm 2.887308146143e+04 5 KSP Residual norm 1.344506498325e+04 6 KSP Residual norm 6.555246305912e+03 7 KSP Residual norm 3.343821147366e+03 8 KSP Residual norm 1.851924043409e+03 Linear solve converged due to CONVERGED_RTOL iterations 8 7 SNES Function norm 5.112097941573e+05 0 KSP Residual norm 5.112097941573e+05 1 KSP Residual norm 1.589049664791e+05 2 KSP Residual norm 6.033831823456e+04 3 KSP Residual norm 2.608416891376e+04 4 KSP Residual norm 1.177853755077e+04 5 KSP Residual norm 5.417771836082e+03 6 KSP Residual norm 2.761499160295e+03 7 KSP Residual norm 1.599727284116e+03 8 KSP Residual norm 1.073148850801e+03 9 KSP Residual norm 8.134885269693e+02 Linear solve converged due to CONVERGED_RTOL iterations 9 8 SNES Function norm 2.796443368363e+05 0 KSP Residual norm 2.796443368363e+05 1 KSP Residual norm 6.968386983526e+04 2 KSP Residual norm 2.462240953692e+04 3 KSP Residual norm 1.156398194382e+04 4 KSP Residual norm 6.076152535454e+03 5 KSP Residual norm 3.110704123273e+03 6 KSP Residual norm 1.609993784143e+03 7 KSP Residual norm 8.500755950825e+02 8 KSP Residual norm 5.335192138320e+02 Linear solve converged due to CONVERGED_RTOL iterations 8 9 SNES Function norm 4.819210383260e+04 Nonlinear solve converged due to CONVERGED_SNORM_RELATIVE iterations 9 SNES Object: 16 MPI processes type: newtonls maximum iterations=500, maximum function evaluations=10000 tolerances: relative=0.0001, absolute=1e-50, solution=0.0001 total number of linear solver iterations=84 total number of function evaluations=106 SNESLineSearch Object: 16 MPI processes type: bt interpolation: cubic alpha=1.000000e-04 maxstep=1.000000e+08, minlambda=1.000000e-12 tolerances: relative=1.000000e-08, absolute=1.000000e-15, lambda=1.000000e-08 maximum iterations=40 KSP Object: 16 MPI processes type: gmres GMRES: restart=60, using Classical (unmodified) Gram-Schmidt Orthogonalization with no iterative refinement GMRES: happy breakdown tolerance 1e-30 maximum iterations=500, initial guess is zero tolerances: relative=0.002, absolute=1e-50, divergence=10000 right preconditioning using UNPRECONDITIONED norm type for convergence test PC Object: 16 MPI processes type: hypre HYPRE BoomerAMG preconditioning HYPRE BoomerAMG: Cycle type V HYPRE BoomerAMG: Maximum number of levels 25 HYPRE BoomerAMG: Maximum number of iterations PER hypre call 1 HYPRE BoomerAMG: Convergence tolerance PER hypre call 0 HYPRE BoomerAMG: Threshold for strong coupling 0.25 HYPRE BoomerAMG: Interpolation truncation factor 0 HYPRE BoomerAMG: Interpolation: max elements per row 0 HYPRE BoomerAMG: Number of levels of aggressive coarsening 0 HYPRE BoomerAMG: Number of paths for aggressive coarsening 1 HYPRE BoomerAMG: Maximum row sums 0.9 HYPRE BoomerAMG: Sweeps down 1 HYPRE BoomerAMG: Sweeps up 1 HYPRE BoomerAMG: Sweeps on coarse 4 HYPRE BoomerAMG: Relax down symmetric-SOR/Jacobi HYPRE BoomerAMG: Relax up symmetric-SOR/Jacobi HYPRE BoomerAMG: Relax on coarse Gaussian-elimination HYPRE BoomerAMG: Relax weight (all) 1 HYPRE BoomerAMG: Outer relax weight (all) 1 HYPRE BoomerAMG: Using CF-relaxation HYPRE BoomerAMG: Measure type local HYPRE BoomerAMG: Coarsen type Falgout HYPRE BoomerAMG: Interpolation type classical linear system matrix followed by preconditioner matrix: Matrix Object: 16 MPI processes type: mffd rows=3276800, cols=3276800 Matrix-free approximation: err=1.49012e-08 (relative error in function evaluation) Using wp compute h routine Does not compute normU Matrix Object: 16 MPI processes type: mpiaij rows=3276800, cols=3276800, bs=2 total: nonzeros=42598400, allocated nonzeros=62261240 total number of mallocs used during MatSetValues calls =136 not using I-node (on process 0) routines ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- ../../driver2d.Linux.64.CC.ftn.OPTHIGH.MPI.PETSC.ex on a arch-xe6-opt64 named nid01753 with 16 processors, by Unknown Wed Mar 6 16:18:28 2013 Using Petsc Development HG revision: f37196d89aa62310230dff96ac6fce27c1d0da5e HG Date: Mon Jan 28 15:07:26 2013 -0600 Max Max/Min Avg Total Time (sec): 2.195e+02 1.00032 2.195e+02 Objects: 8.900e+01 1.00000 8.900e+01 Flops: 9.298e+08 2.00000 7.264e+08 1.162e+10 Flops/sec: 4.237e+06 2.00029 3.309e+06 5.295e+07 MPI Messages: 1.200e+02 1.66667 9.150e+01 1.464e+03 MPI Message Lengths: 3.284e+05 1.99796 2.414e+03 3.534e+06 MPI Reductions: 7.630e+02 1.00000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 2.1951e+02 100.0% 1.1623e+10 100.0% 1.464e+03 100.0% 2.414e+03 100.0% 7.620e+02 99.9% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length (bytes) Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage MatMult MF 93 1.0 7.2327e+01 1.0 2.19e+08 2.0 0.0e+00 0.0e+00 3.7e+02 33 24 0 0 49 33 24 0 0 49 38 MatMult 101 1.0 7.2585e+01 1.0 2.72e+08 2.0 9.8e+02 2.9e+03 3.7e+02 33 29 67 80 49 33 29 67 80 49 47 MatConvert 11 1.0 1.5847e+00 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 0 MatAssemblyBegin 30 1.0 4.3773e+0080.2 0.00e+00 0.0 0.0e+00 0.0e+00 4.2e+01 1 0 0 0 6 1 0 0 0 6 0 MatAssemblyEnd 30 1.0 8.2351e-01 1.2 0.00e+00 0.0 4.9e+02 1.4e+03 3.7e+01 0 0 33 20 5 0 0 33 20 5 0 MatGetRowIJ 22 1.0 3.0105e-05 1.4 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatZeroEntries 1 1.0 1.5647e-02 3.4 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatView 2 1.0 3.0701e-04 2.7 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecDot 9 1.0 5.0630e-02 3.6 4.72e+06 2.0 0.0e+00 0.0e+00 9.0e+00 0 1 0 0 1 0 1 0 0 1 1165 VecMDot 90 1.0 1.1321e+00 1.7 2.46e+08 2.0 0.0e+00 0.0e+00 9.0e+01 0 27 0 0 12 0 27 0 0 12 2721 VecNorm 206 1.0 6.3892e-01 3.7 1.08e+08 2.0 0.0e+00 0.0e+00 2.1e+02 0 12 0 0 27 0 12 0 0 27 2113 VecScale 194 1.0 1.9393e-01 5.0 5.09e+07 2.0 0.0e+00 0.0e+00 0.0e+00 0 5 0 0 0 0 5 0 0 0 3278 VecCopy 51 1.0 1.0126e-01 2.8 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 235 1.0 3.2654e-01 3.2 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 212 1.0 5.1615e-01 3.2 1.11e+08 2.0 0.0e+00 0.0e+00 0.0e+00 0 12 0 0 0 0 12 0 0 0 2692 VecWAXPY 105 1.0 3.7886e-01 3.9 5.27e+07 2.0 0.0e+00 0.0e+00 0.0e+00 0 6 0 0 0 0 6 0 0 0 1738 VecMAXPY 101 1.0 8.5163e-01 2.3 2.94e+08 2.0 0.0e+00 0.0e+00 0.0e+00 0 32 0 0 0 0 32 0 0 0 4309 VecAssemblyBegin 112 1.0 4.7448e+00124.8 0.00e+00 0.0 0.0e+00 0.0e+00 3.4e+02 1 0 0 0 44 1 0 0 0 44 0 VecAssemblyEnd 112 1.0 5.6762e-04 1.8 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecScatterBegin 8 1.0 7.7875e-04 2.0 0.00e+00 0.0 9.8e+02 2.9e+03 0.0e+00 0 0 67 80 0 0 0 67 80 0 0 VecScatterEnd 8 1.0 1.0894e-01305.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecReduceArith 19 1.0 1.3939e-02 2.2 9.96e+06 2.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 8933 VecReduceComm 10 1.0 3.7179e-02169.8 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+01 0 0 0 0 1 0 0 0 0 1 0 VecNormalize 101 1.0 5.2865e-01 3.1 7.47e+07 2.0 0.0e+00 0.0e+00 9.2e+01 0 8 0 0 12 0 8 0 0 12 1767 KSPGMRESOrthog 90 1.0 1.4561e+00 1.1 4.93e+08 2.0 0.0e+00 0.0e+00 9.0e+01 1 53 0 0 12 1 53 0 0 12 4231 KSPSetUp 11 1.0 1.7270e-02 2.6 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 11 1.0 1.7257e+02 1.0 8.77e+08 2.0 9.8e+02 2.9e+03 5.4e+02 79 94 67 80 70 79 94 67 80 70 64 PCSetUp 11 1.0 3.2159e+01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 8.0e+00 15 0 0 0 1 15 0 0 0 1 0 PCApply 101 1.0 7.3134e+01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 33 0 0 0 0 33 0 0 0 0 0 SNESSolve 1 1.0 1.8582e+02 1.0 8.52e+08 2.0 0.0e+00 0.0e+00 6.8e+02 85 92 0 0 89 85 92 0 0 89 57 SNESFunctionEval 106 1.0 8.1569e+01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 3.2e+02 37 0 0 0 42 37 0 0 0 42 0 SNESJacobianEval 9 1.0 5.7884e+00 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 5.4e+01 3 0 0 0 7 3 0 0 0 7 0 SNESLineSearch 9 1.0 1.6167e+01 1.0 5.19e+07 2.0 0.0e+00 0.0e+00 1.0e+02 7 6 0 0 13 7 6 0 0 13 40 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage MatMFFD 1 1 856 0 Matrix 7 7 193006672 0 Vector 60 60 117591872 0 Vector Scatter 2 2 2408 0 Index Set 4 4 52672 0 Krylov Solver 2 2 128944 0 DMKSP interface 1 1 712 0 Preconditioner 2 2 2480 0 Viewer 1 0 0 0 SNES 1 1 1456 0 SNESLineSearch 1 1 920 0 DMSNES 1 1 728 0 Distributed Mesh 2 2 8912 0 Bipartite Graph 4 4 3488 0 ======================================================================================================================== Average time to get PetscTime(): 1.93097e-07 Average time for MPI_Barrier(): 9.62102e-06 Average time for zero size MPI_Send(): 2.31426e-06 #PETSc Option Table entries: -ksp_converged_reason -ksp_converged_use_initial_residual_norm -ksp_gmres_restart 60 -ksp_max_it 500 -ksp_monitor -ksp_norm_type unpreconditioned -ksp_rtol 2.e-3 -ksp_type gmres -log_summary -mg_levels_ksp_chebyshev_estimate_eigenvalues 0,0.1,0,1.05 -mg_levels_ksp_max_it 2 -mg_levels_ksp_type chebyshev -mg_levels_pc_type sor -pc_gamg_agg_nsmooths 1 -pc_gamg_reuse_interpolation true -pc_gamg_sym_graph -pc_gamg_threshold .05 -pc_gamg_type agg -pc_gamg_verbose 0 -pc_hypre_boomeramg_grid_sweeps_coarse 4 -pc_hypre_type boomeramg -pc_ml_EnergyMinimization 2 -pc_ml_PrintLevel 1 -pc_ml_Threshold 0.01 -pc_type hypre -snes_converged_reason -snes_max_funcs 10000 -snes_max_it 500 -snes_mf_operator -snes_monitor -snes_rtol 1.e-4 -snes_stol 1.e-4 -snes_view #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 8 Configure run at: Wed Jan 30 07:39:23 2013 Configure options: --COPTFLAGS="-O3 -ffast-math -funroll-loops" --CXXOPTFLAGS="-O3 -ffast-math -funroll-loops" --FOPTFLAGS="-O3 -ffast-math -funroll-loops" --download-parmetis --download-metis --download-hypre --with-cc=cc --with-clib-autodetect=0 --with-cxx=CC --with-cxxlib-autodetect=0 --with-debugging=0 --with-fc=ftn --with-fortranlib-autodetect=0 --with-mpiexec=/usr/common/acts/PETSc/3.1/bin/mpiexec.aprun --with-shared-libraries=0 --with-x=0 --with-64-bit-indices PETSC_ARCH=arch-xe6-opt64 ----------------------------------------- Libraries compiled on Wed Jan 30 07:39:23 2013 on hopper09 Machine characteristics: Linux-2.6.32.36-0.5-default-x86_64-with-SuSE-11-x86_64 Using PETSc directory: /global/homes/m/madams/petsc-dev Using PETSc arch: arch-xe6-opt64 ----------------------------------------- Using C compiler: cc -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O3 -ffast-math -funroll-loops ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: ftn -Wall -Wno-unused-variable -Wno-unused-dummy-argument -O3 -ffast-math -funroll-loops ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/global/homes/m/madams/petsc-dev/arch-xe6-opt64/include -I/global/homes/m/madams/petsc-dev/include -I/global/homes/m/madams/petsc-dev/include -I/global/homes/m/madams/petsc-dev/arch-xe6-opt64/include ----------------------------------------- Using C linker: cc Using Fortran linker: ftn Using libraries: -Wl,-rpath,/global/homes/m/madams/petsc-dev/arch-xe6-opt64/lib -L/global/homes/m/madams/petsc-dev/arch-xe6-opt64/lib -lpetsc -Wl,-rpath,/global/homes/m/madams/petsc-dev/arch-xe6-opt64/lib -L/global/homes/m/madams/petsc-dev/arch-xe6-opt64/lib -lHYPRE -lparmetis -lmetis -lpthread -ldl ----------------------------------------- Application 15484282 resources: utime ~3455s, stime ~17s, Rss ~615780, inblocks ~707380, outblocks ~81420