**************************************************************************************************************************************************************** *** WIDEN YOUR WINDOW TO 160 CHARACTERS. Use 'enscript -r -fCourier9' to print this document *** **************************************************************************************************************************************************************** ------------------------------------------------------------------ PETSc Performance Summary: ------------------------------------------------------------------ Unknown Name on a named Jorges-MacBook-Pro.local with 1 processor, by jorgenin Thu Oct 19 15:40:16 2023 Using Petsc Release Version 3.20.0, Sep 28, 2023 Max Max/Min Avg Total Time (sec): 1.403e+02 1.000 1.403e+02 Objects: 0.000e+00 0.000 0.000e+00 Flops: 6.815e+09 1.000 6.815e+09 6.815e+09 Flops/sec: 4.856e+07 1.000 4.856e+07 4.856e+07 MPI Msg Count: 0.000e+00 0.000 0.000e+00 0.000e+00 MPI Msg Len (bytes): 0.000e+00 0.000 0.000e+00 0.000e+00 MPI Reductions: 0.000e+00 0.000 Flop counting convention: 1 flop = 1 real number operation of type (multiply/divide/add/subtract) e.g., VecAXPY() for real vectors of length N --> 2N flops and VecAXPY() for complex vectors of length N --> 8N flops Summary of Stages: ----- Time ------ ----- Flop ------ --- Messages --- -- Message Lengths -- -- Reductions -- Avg %Total Avg %Total Count %Total Avg %Total Count %Total 0: Main Stage: 1.3815e+02 98.4% 6.8152e+09 100.0% 0.000e+00 0.0% 0.000e+00 0.0% 0.000e+00 0.0% ------------------------------------------------------------------------------------------------------------------------ See the 'Profiling' chapter of the users' manual for details on interpreting output. Phase summary info: Count: number of times phase was executed Time and Flop: Max - maximum over all processors Ratio - ratio of maximum to minimum over all processors Mess: number of messages sent AvgLen: average message length (bytes) Reduct: number of global reductions Global: entire computation Stage: stages of a computation. Set stages with PetscLogStagePush() and PetscLogStagePop(). %T - percent time in this phase %F - percent flop in this phase %M - percent messages in this phase %L - percent message lengths in this phase %R - percent reductions in this phase Total Mflop/s: 10e-6 * (sum of flop over all processors)/(max time over all processors) ------------------------------------------------------------------------------------------------------------------------ Event Count Time (sec) Flop --- Global --- --- Stage ---- Total Max Ratio Max Ratio Max Ratio Mess AvgLen Reduct %T %F %M %L %R %T %F %M %L %R Mflop/s ------------------------------------------------------------------------------------------------------------------------ --- Event Stage 0: Main Stage BuildTwoSided 1 1.0 1.0000e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 SFSetGraph 1 1.0 0.0000e+00 0.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 SFSetUp 1 1.0 3.0000e-06 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 SFPack 240 1.0 3.3000e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 SFUnpack 240 1.0 1.8000e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecNorm 100 1.0 7.3040e-03 1.0 3.79e+06 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 518 VecCopy 100 1.0 5.5500e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecSet 120 1.0 2.3800e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecAXPY 100 1.0 5.8210e-03 1.0 3.79e+06 1.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 650 VecScatterBegin 240 1.0 8.2100e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 VecScatterEnd 240 1.0 4.2800e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatSolve 100 1.0 2.6633e+00 1.0 3.59e+09 1.0 0.0e+00 0.0e+00 0.0e+00 2 53 0 0 0 2 53 0 0 0 1349 MatLUFactorSym 1 1.0 1.0006e-01 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatLUFactorNum 100 1.0 1.2305e+02 1.0 2.79e+09 1.0 0.0e+00 0.0e+00 0.0e+00 88 41 0 0 0 89 41 0 0 0 23 MatAssemblyBegin 200 1.0 4.2000e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatAssemblyEnd 200 1.0 7.5100e-04 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 MatZeroEntries 100 1.0 1.3714e-02 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 PCSetUp 100 1.0 1.2315e+02 1.0 2.79e+09 1.0 0.0e+00 0.0e+00 0.0e+00 88 41 0 0 0 89 41 0 0 0 23 PCApply 100 1.0 2.6634e+00 1.0 3.59e+09 1.0 0.0e+00 0.0e+00 0.0e+00 2 53 0 0 0 2 53 0 0 0 1349 KSPSetUp 20 1.0 1.8000e-05 1.0 0.00e+00 0.0 0.0e+00 0.0e+00 0.0e+00 0 0 0 0 0 0 0 0 0 0 0 KSPSolve 100 1.0 9.3857e+01 1.0 5.83e+09 1.0 0.0e+00 0.0e+00 0.0e+00 67 86 0 0 0 68 86 0 0 0 62 ------------------------------------------------------------------------------------------------------------------------ Object Type Creations Destructions. Reports information only for process 0. --- Event Stage 0: Main Stage Viewer 1 0 Index Set 2 2 IS L to G Mapping 1 0 Star Forest Graph 1 0 Vector 4 1 Matrix 1 0 --- Event Stage 1: PCMPI ======================================================================================================================== Average time to get PetscTime(): 0. #PETSc Option Table entries: -nls_solve_ksp_max_it 30 # (source: code) -nls_solve_ksp_type preonly # (source: code) -nls_solve_log_view log # (source: code) -nls_solve_pc_factor_mat_solver_type mumps # (source: code) -nls_solve_pc_type lu # (source: code) #End of PETSc Option Table entries Compiled without FORTRAN kernels Compiled with full precision matrices (default) sizeof(short) 2 sizeof(int) 4 sizeof(long) 8 sizeof(void*) 8 sizeof(PetscScalar) 8 sizeof(PetscInt) 4 Configure options: AR=arm64-apple-darwin20.0.0-ar CC=mpicc CXX=mpicxx FC=mpifort CFLAGS="-ftree-vectorize -fPIC -fPIE -fstack-protector-strong -O2 -pipe -isystem /Users/jorgenin/Documents/Python/Research/2.077/.conda/include " CPPFLAGS="-D_FORTIFY_SOURCE=2 -isystem /Users/jorgenin/Documents/Python/Research/2.077/.conda/include -mmacosx-version-min=11.0" CXXFLAGS="-ftree-vectorize -fPIC -fPIE -fstack-protector-strong -O2 -pipe -stdlib=libc++ -fvisibility-inlines-hidden -fmessage-length=0 -isystem /Users/jorgenin/Documents/Python/Research/2.077/.conda/include " FFLAGS="-march=armv8.3-a -ftree-vectorize -fPIC -fno-stack-protector -O2 -pipe -isystem /Users/jorgenin/Documents/Python/Research/2.077/.conda/include " LDFLAGS="-Wl,-pie -Wl,-headerpad_max_install_names -Wl,-dead_strip_dylibs -Wl,-rpath,/Users/jorgenin/Documents/Python/Research/2.077/.conda/lib -L/Users/jorgenin/Documents/Python/Research/2.077/.conda/lib" LIBS="-lmpifort -lgfortran" --COPTFLAGS=-O3 --CXXOPTFLAGS=-O3 --FOPTFLAGS=-O3 --with-clib-autodetect=0 --with-cxxlib-autodetect=0 --with-fortranlib-autodetect=0 --with-debugging=0 --with-blas-lib=libblas.dylib --with-lapack-lib=liblapack.dylib --with-yaml=1 --with-hdf5=1 --with-fftw=1 --with-hwloc=0 --with-hypre=1 --with-metis=1 --with-mpi=1 --with-mumps=1 --with-parmetis=1 --with-pthread=1 --with-ptscotch=1 --with-shared-libraries --with-ssl=0 --with-scalapack=1 --with-superlu=1 --with-superlu_dist=1 --with-suitesparse=1 --with-x=0 --with-scalar-type=real --with-batch --prefix=/Users/jorgenin/Documents/Python/Research/2.077/.conda ----------------------------------------- Libraries compiled on 2023-10-10 12:44:13 on Mac-1696940584996.local Machine characteristics: macOS-11.7.10-x86_64-i386-64bit Using PETSc directory: /Users/runner/miniforge3/conda-bld/petsc_1696941332205/_h_env_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_pla Using PETSc arch: ----------------------------------------- Using C compiler: mpicc -ftree-vectorize -fPIC -fPIE -fstack-protector-strong -O2 -pipe -isystem /Users/runner/miniforge3/conda-bld/petsc_1696941332205/_h_env_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_pla/include -O3 -D_FORTIFY_SOURCE=2 -isystem /Users/runner/miniforge3/conda-bld/petsc_1696941332205/_h_env_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_pla/include -mmacosx-version-min=11.0 Using Fortran compiler: mpifort -march=armv8.3-a -ftree-vectorize -fPIC -fno-stack-protector -O2 -pipe -isystem /Users/runner/miniforge3/conda-bld/petsc_1696941332205/_h_env_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_pla/include -O3 -D_FORTIFY_SOURCE=2 -isystem /Users/runner/miniforge3/conda-bld/petsc_1696941332205/_h_env_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_pla/include -mmacosx-version-min=11.0 ----------------------------------------- Using include paths: -I/Users/runner/miniforge3/conda-bld/petsc_1696941332205/_h_env_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_pla/include ----------------------------------------- Using C linker: mpicc Using Fortran linker: mpifort Using libraries: -Wl,-rpath,/Users/runner/miniforge3/conda-bld/petsc_1696941332205/_h_env_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_pla/lib -L/Users/runner/miniforge3/conda-bld/petsc_1696941332205/_h_env_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_placehold_pla/lib -lpetsc -lHYPRE -lspqr -lumfpack -lklu -lcholmod -lbtf -lccolamd -lcolamd -lcamd -lamd -lsuitesparseconfig -ldmumps -lmumps_common -lpord -lpthread -lscalapack -lsuperlu -lsuperlu_dist -lfftw3_mpi -lfftw3 -llapack -lblas -lptesmumps -lptscotchparmetis -lptscotch -lptscotcherr -lesmumps -lscotch -lscotcherr -lparmetis -lmetis -lhdf5_hl -lhdf5 -lyaml -lmpifort -lgfortran -----------------------------------------