F I N I T E E L E M E N T A N A L Y S I S P R O G R A M FEAP (C) Regents of the University of California All Rights Reserved. VERSION: Release 8.3.19 DATE: 29 March 2011 Files are set as: Status Filename Input (read ) : Exists Icube_0001 Output (write) : Exists Ocube_0001 Restart (read ) : New Rcube_0001 Restart (write) : New Rcube_0001 Plots (write) : New Pcube_0001 Caution, existing write files will be overwritten. Are filenames correct? ( y or n; s = stop) : R U N N I N G F E A P P R O B L E M N O W --> Please report errors by e-mail to: feap@ce.berkeley.edu 0 KSP Residual norm 1.444126847260e-01 1 KSP Residual norm 5.357525404213e-03 2 KSP Residual norm 1.471040678379e-03 3 KSP Residual norm 3.704652302293e-04 4 KSP Residual norm 9.809180893460e-05 5 KSP Residual norm 3.175497350277e-05 6 KSP Residual norm 8.859979496890e-06 7 KSP Residual norm 2.071384344082e-06 8 KSP Residual norm 5.035483717523e-07 9 KSP Residual norm 1.516500637412e-07 10 KSP Residual norm 5.134577847338e-08 11 KSP Residual norm 1.270806138401e-08 12 KSP Residual norm 3.074793756862e-09 13 KSP Residual norm 1.048189477307e-09 KSP Object: 2 MPI processes type: cg maximum iterations=10000, initial guess is zero tolerances: relative=1e-08, absolute=1e-16, divergence=1e+16 left preconditioning using PRECONDITIONED norm type for convergence test PC Object: 2 MPI processes type: prometheus linear system matrix = precond matrix: Matrix Object: 2 MPI processes type: mpibaij rows=2013, cols=2013 total: nonzeros=100899, allocated nonzeros=100899 total number of mallocs used during MatSetValues calls =0 block size is 3 ************************************************************************************************************************ *** WIDEN YOUR WINDOW TO 120 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** ************************************************************************************************************************ ---------------------------------------------- PETSc Performance Summary: ---------------------------------------------- /usr2/tgross/parFEAP/parFEAP83_mod/ver83/parfeap/feap on a linux-gnu named ilfb35.ilsb.tuwien.ac.at with 2 processors, by tgross Mon Jan 27 17:11:10 2014 Using Petsc Release Version 3.2.0, Patch 7, Thu Mar 15 09:30:51 CDT 2012 Max Max/Min Avg Total Time (sec): 1.379e-01 1.00662 1.374e-01 Objects: 6.800e+01 1.00000 6.800e+01 Flops: 4.104e+07 1.16967 3.807e+07 7.613e+07 Flops/sec: 2.997e+08 1.17742 2.771e+08 5.542e+08 MPI Messages: 1.515e+02 1.00664 1.510e+02 3.020e+02 MPI Message Lengths: 4.012e+05 1.02186 2.628e+03 7.938e+05 MPI Reductions: 2.170e+02 1.01402 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flops ----- --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total counts %Total Avg %Total counts %Total 0: Main Stage: 1.3742e-01 100.0% 7.6132e+07 100.0% 3.020e+02 100.0% 2.628e+03 100.0% 2.145e+02 98.8% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flops: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent Avg. len: average message length Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %f - percent flops in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flops over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flops --- Global --- --- Stage --- Total Max Ratio Max Ratio Max Ratio Mess Avg len Reduct %T %f %M %L %R %T %f %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage MatMult 50 1.0 4.6628e-03 1.6 5.02e+06 1.0 1.0e+02 1.4e+03 0.0e+00 3 13 33 18 0 3 13 33 18 0 2142 MatMultAdd 14 1.0 1.0583e-03 2.3 4.16e+05 1.1 1.4e+01 1.4e+03 0.0e+00 1 1 5 2 0 1 1 5 2 0 762 MatMultTranspose 14 1.0 5.5742e-04 1.2 4.16e+05 1.1 1.4e+01 1.4e+03 0.0e+00 0 1 5 2 0 0 1 5 2 0 1446 MatSolve 14 0.0 3.0231e-04 0.0 9.14e+05 0.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 3023 MatLUFactorSym 1 1.0 4.2915e-05 3.9 0.00e+00 0.0 0.0e+00 0.0e+00 3.0e+00 0 0 0 0 1 0 0 0 0 1 0 MatLUFactorNum 1 1.0 1.9600e-03483.6 3.25e+06 0.0 0.0e+00 0.0e+00 0.0e+00 1 4 0 0 0 1 4 0 0 0 1659 MatAssemblyBegin 9 1.0 3.3593e-04 5.4 0.00e+00 0.0 0.0e+00 0.0e+00 2.1e+01 0 0 0 0 10 0 0 0 0 10 0 MatAssemblyEnd 9 1.0 8.5688e-04 1.0 0.00e+00 0.0 6.0e+00 2.3e+02 2.6e+01 1 0 2 0 12 1 0 2 0 12 0 MatGetRow 336 1.0 1.1373e-04 1.1 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetRowIJ 1 0.0 7.1526e-06 0.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatGetSubMatrice 1 1.0 1.3590e-04 2.0 0.00e+00 0.0 0.0e+00 0.0e+00 6.0e+00 0 0 0 0 3 0 0 0 0 3 0 MatGetOrdering 1 0.0 3.0041e-05 0.0 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatZeroEntries 2 1.0 4.2915e-05 1.3 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatView 1 1.0 4.0054e-05 1.6 0.00e+00 0.0 0.0e+00 0.0e+00 1.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecDot 19 1.0 1.2422e-04 1.8 3.83e+04 1.0 0.0e+00 0.0e+00 1.9e+01 0 0 0 0 9 0 0 0 0 9 616 VecTDot 26 1.0 2.5082e-04 2.7 5.24e+04 1.0 0.0e+00 0.0e+00 2.6e+01 0 0 0 0 12 0 0 0 0 12 417 VecNorm 18 1.0 3.7885e-04 1.3 3.02e+04 1.0 0.0e+00 0.0e+00 1.8e+01 0 0 0 0 8 0 0 0 0 8 159 VecScale 28 1.0 4.0770e-05 1.2 2.82e+04 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 1383 VecCopy 88 1.0 5.5552e-05 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 145 1.0 6.1274e-05 1.3 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 58 1.0 7.9393e-05 1.1 1.17e+05 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 2941 VecAYPX 48 1.0 8.5354e-05 1.0 6.85e+04 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 1604 VecAssemblyBegin 3 1.0 2.4080e-05 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 9.0e+00 0 0 0 0 4 0 0 0 0 4 0 VecAssemblyEnd 3 1.0 2.1458e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecScatterBegin 106 1.0 1.4353e-04 1.3 0.00e+00 0.0 1.3e+02 1.4e+03 0.0e+00 0 0 42 22 0 0 0 42 22 0 0 VecScatterEnd 106 1.0 2.8653e-0320.9 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 1 0 0 0 0 1 0 0 0 0 0 KSPSetup 2 1.0 2.0027e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 1 1.0 4.6679e-02 1.0 4.10e+07 1.2 3.0e+02 2.7e+03 1.9e+02 34100 99100 86 34100 99100 87 1631 PCSetUp 1 1.0 4.1773e-02 1.1 3.50e+07 1.2 1.9e+02 3.4e+03 1.5e+02 30 85 62 81 68 30 85 62 81 69 1554 PCSetUpOnBlocks 1 1.0 2.0559e-0351.3 3.25e+06 0.0 0.0e+00 0.0e+00 5.0e+00 1 4 0 0 2 1 4 0 0 2 1582 PCApply 14 1.0 5.5327e-03 1.5 4.56e+06 1.2 8.4e+01 1.4e+03 0.0e+00 3 11 28 15 0 3 11 28 15 0 1498 FEI: init. str. 1 1.0 1.9560e-03 1.0 0.00e+00 0.0 4.0e+00 4.7e+03 9.0e+00 1 0 1 2 4 1 0 1 2 4 0 FEI: Prom setup 4 1.0 1.1263e-02 1.0 0.00e+00 0.0 1.6e+02 1.8e+03 2.2e+01 8 0 53 36 10 8 0 53 36 10 0 FEI: solv.setup 3 1.0 2.8233e-02 1.1 3.50e+07 1.2 2.3e+01 1.4e+04 1.1e+02 20 85 8 42 51 20 85 8 42 52 2300 FEI: BCs & reg. 1 1.0 6.1989e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 7.0e+00 0 0 0 0 3 0 0 0 0 3 0 FEI: MakeRest. 1 1.0 2.5464e-02 1.0 3.18e+07 1.1 2.3e+01 1.4e+04 6.7e+01 19 81 8 42 31 19 81 8 42 31 2422 FEI: SLESSolve*3 14 1.0 5.5134e-03 1.6 4.56e+06 1.2 8.4e+01 1.4e+03 0.0e+00 3 11 28 15 0 3 11 28 15 0 1503 Fine grid 1 1 1.0 1.5974e-05 1.2 0.00e+00 0.0 0.0e+00 0.0e+00 3.0e+00 0 0 0 0 1 0 0 0 0 1 0 MG++(new grid) 1 1 1.0 3.8149e-03 1.0 0.00e+00 0.0 1.1e+01 1.4e+04 1.0e+00 3 0 4 19 0 3 0 4 19 0 0 RAP 2 1 1.0 2.1625e-02 1.0 2.99e+07 1.1 1.0e+00 1.9e+05 1.6e+01 16 76 0 24 7 16 76 0 24 7 2682 SLES setup *2 1 1.0 2.5468e-02 1.0 3.18e+07 1.1 2.3e+01 1.4e+04 6.7e+01 19 81 8 42 31 19 81 8 42 31 2421 Prometheus *1 5 1.0 1.3446e-02 1.0 0.00e+00 0.0 1.6e+02 1.9e+03 3.2e+01 10 0 54 39 15 10 0 54 39 15 0 Coarse Grid Solv 14 1.0 4.7636e-04 3.7 9.14e+05 0.0 0.0e+00 0.0e+00 0.0e+00 0 1 0 0 0 0 1 0 0 0 1918 Grid coarsen 1 1 1.0 4.2310e-03 1.0 0.00e+00 0.0 1.2e+02 7.5e+02 1.1e+01 3 0 41 12 5 3 0 41 12 5 0 New Coarse 1 1 1.0 3.0611e-03 1.0 0.00e+00 0.0 2.3e+01 1.7e+03 6.5e+00 2 0 8 5 3 2 0 8 5 3 0 SmoothP 2 1 1.0 2.2371e-03 1.0 1.83e+06 1.0 2.0e+01 7.2e+03 2.8e+01 2 5 7 18 13 2 5 7 18 13 1634 CG Est. lamb.1 2 1 1.0 1.0400e-03 1.0 1.00e+06 1.0 1.8e+01 1.4e+03 2.8e+01 1 3 6 3 13 1 3 6 3 13 1915 AP_0 2 1 1.0 9.2292e-04 1.3 7.91e+05 1.0 0.0e+00 0.0e+00 0.0e+00 1 2 0 0 0 1 2 0 0 0 1714 RAP core 2 1 1.0 2.1017e-02 1.2 2.99e+07 1.1 0.0e+00 0.0e+00 0.0e+00 14 76 0 0 0 14 76 0 0 0 2760 RAP my assem. 2 1 1.0 4.2980e-0310.2 0.00e+00 0.0 1.0e+00 1.9e+05 1.0e+00 2 0 0 24 0 2 0 0 24 0 0 ------------------------------------------------------------------------------------------------------------------------ Memory usage is given in bytes: Object Type Creations Destructions Memory Descendants' Mem. Reports information only for process 0. --- Event Stage 0: Main Stage Matrix 11 8 1241380 0 Vector 33 32 228544 0 Vector Scatter 4 3 3180 0 Index Set 14 14 11872 0 Krylov Solver 2 2 2224 0 Preconditioner 3 3 2896 0 Viewer 1 0 0 0 ======================================================================================================================== Average time to get PetscTime(): 0 Average time for MPI_Barrier(): 4.29153e-07 Average time for zero size MPI_Send(): 1.07288e-06 #PETSc Option Table entries: -aggmg_smooths 1 -ksp_monitor -ksp_type cg -ksp_view -log_summary -mat_inode_limit 3 -options_left -pc_type prometheus #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 Configure run at: Tue Feb 12 09:44:26 2013 Configure options: --with-mpi-dir=/usr/local/openmpi/1.5.4/gcc/x86_64 --download-f-blas-lapack=1 --download-parmetis=1 --download-prometheus=1 --with-debugging=0 --with-shared-libraries=0 --download-spooles=1 --download-hypre=1 --download-superlu_dist=1 ----------------------------------------- Libraries compiled on Tue Feb 12 09:44:26 2013 on ilfb46.ilsb.tuwien.ac.at Machine characteristics: Linux-2.6.32-279.5.1.el6.x86_64-x86_64-with-redhat-6.3-Carbon Using PETSc directory: /usr2/pahr/software/feap/petsc-3.2-p7 Using PETSc arch: linux-gnu-c ----------------------------------------- Using C compiler: /usr/local/openmpi/1.5.4/gcc/x86_64/bin/mpicc -Wall -Wwrite-strings -Wno-strict-aliasing -Wno-unknown-pragmas -O ${COPTFLAGS} ${CFLAGS} Using Fortran compiler: /usr/local/openmpi/1.5.4/gcc/x86_64/bin/mpif90 -Wall -Wno-unused-variable -O ${FOPTFLAGS} ${FFLAGS} ----------------------------------------- Using include paths: -I/usr2/pahr/software/feap/petsc-3.2-p7/linux-gnu-c/include -I/usr2/pahr/software/feap/petsc-3.2-p7/include -I/usr2/pahr/software/feap/petsc-3.2-p7/include -I/usr2/pahr/software/feap/petsc-3.2-p7/linux-gnu-c/include -I/usr/local/openmpi/1.5.4/gcc/x86_64/include -I/usr/local/include ----------------------------------------- Using C linker: /usr/local/openmpi/1.5.4/gcc/x86_64/bin/mpicc Using Fortran linker: /usr/local/openmpi/1.5.4/gcc/x86_64/bin/mpif90 Using libraries: -Wl,-rpath,/usr2/pahr/software/feap/petsc-3.2-p7/linux-gnu-c/lib -L/usr2/pahr/software/feap/petsc-3.2-p7/linux-gnu-c/lib -lpetsc -lX11 -lpthread -Wl,-rpath,/usr2/pahr/software/feap/petsc-3.2-p7/linux-gnu-c/lib -L/usr2/pahr/software/feap/petsc-3.2-p7/linux-gnu-c/lib -lpromfei -lprometheus -lmpi_cxx -lstdc++ -lsuperlu_dist_2.5 -lparmetis -lmetis -lHYPRE -lmpi_cxx -lstdc++ -lspooles -lflapack -lfblas -L/usr/local/lib64 -L/usr/local/lib64/openmpi -L/usr/local/openmpi/1.5.4/gcc/x86_64/lib64 -L/usr/lib/gcc/x86_64-redhat-linux/4.4.6 -ldl -lmpi -lnsl -lutil -lgcc_s -lpthread -lmpi_f90 -lmpi_f77 -lgfortran -lm -lm -lm -lm -lmpi_cxx -lstdc++ -lmpi_cxx -lstdc++ -ldl -lmpi -lnsl -lutil -lgcc_s -lpthread -ldl ----------------------------------------- #PETSc Option Table entries: -aggmg_smooths 1 -ksp_monitor -ksp_type cg -ksp_view -log_summary -mat_inode_limit 3 -options_left -pc_type prometheus #End of PETSc Option Table entries There are no unused options.