[petsc-users] MatView to disk for Elemental

Smith, Barry F. bsmith at mcs.anl.gov
Tue Dec 3 02:14:19 CST 2019


  sorry about this. The numerical values between C and Fortran got out of sync. I've attached a patch file you can apply with

  patch -p1 < format.patch


  or you can use the branch https://gitlab.com/petsc/petsc/merge_requests/2346

   Barry


> On Dec 3, 2019, at 1:10 AM, Marius Buerkle <mbuerkle at web.de> wrote:
>
> Hi,
>
> I try to save a matrix in Elemental format to disk. I am doing, where p_matout is of type MATELEMENTAL,
>
>     call PetscViewerCreate(PETSC_COMM_WORLD,v_file,ierr)
>     call PetscViewerPushFormat(v_file,PETSC_VIEWER_NATIVE,ierr)
>     call PetscViewerSetType(v_file,PETSCVIEWERBINARY,ierr)
>     call PetscViewerFileSetMode(v_file,FILE_MODE_WRITE,ierr)
>     call PetscViewerFileSetName(v_file,trim(filename),ierr)
>     call MatView(p_matout,v_file,ierr)
>     call PetscViewerDestroy(v_file,ierr)
>
> This gives the following error
> [18]PETSC ERROR: --------------------- Error Message --------------------------------------------------------------
> [18]PETSC ERROR: No support for this operation for this object type
> [18]PETSC ERROR: To store a parallel dense matrix you must first call PetscViewerPushFormat(viewer,PETSC_VIEWER_NATIVE)
> [18]PETSC ERROR: See https://www.mcs.anl.gov/petsc/documentation/faq.html for trouble shooting.
> [18]PETSC ERROR: Petsc Development GIT revision: v3.12.1-317-gcc59f4f82c  GIT Date: 2019-11-08 00:07:59 -0600
> [18]PETSC ERROR: /home/marius/prog/ownstuff/fortran/programs/transomat_dev/save_load_fs/transomat/transomat on a  named tono-hpc1 by marius Tue Dec  3 16:08:38 2019
> [18]PETSC ERROR: Configure options --prefix=/home/marius/prog/petsc/petsc_slepc_opt --with-scalar-type=complex --with-fortran-kernels=1 --with-64-bit-indices=0 --CC=mpicc --COPTFLAGS="-g -Ofast -std=c11 -qopenmmp" --CXX=mpicxx --CXXOPTFLAGS="-g -Ofast -std=c++14 -qopenmp" --FC=mpif90 --FOPTFLAGS="-g -Ofast -traceback -qopenmp" --with-mpi=1 --with-x=0 --download-parmetis=1 --download-metis=1 --download-superlu_dist=1 --download-superlu_dist-commit=f8ace664ec4ca10e96e258a764552cbda299ba6e --download-superlu_dist-cmake-arguments=-Denable_openmp:BOOL=TRUE --download-hwloc=1 --download-sowing=1 --with-openmp=1 --with-pthread=1 --download-elemental=1 --download-elemental-commit=6eb15a0da2a4998bf1cf971ae231b78e06d989d9 --download-elemental-cmake-arguments=-DEL_HYBRID:BOOL=TRUE --with-cxx-dialect=c++11 --with-debugging=0 --with-valgrind=0 --with-blaslapack-lib=" /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_scalapack_lp64.a -Wl,--start-group /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_intel_lp64.a /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_intel_thread.a /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_core.a /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_blacs_openmpi_lp64.a -Wl,--end-group -liomp5 -lpthread -lm -ldl -lmpi_wrapper" --with-scalapack-lib=" /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_scalapack_lp64.a -Wl,--start-group /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_intel_lp64.a /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_intel_thread.a /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_core.a /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_blacs_openmpi_lp64.a -Wl,--end-group -liomp5 -lpthread -lm -ldl -lmpi_wrapper" --with-mkl_pardiso-dir=/home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl --with-mkl_cpardiso-dir=/home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl --with-mkl_sparse-dir=/home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl --with-mkl_sparse_optimize-dir=/home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl --download-slepc=1 --download-slepc-commit=658271f --download-make=1 --download-cmake=1
> [18]PETSC ERROR: #1 MatView_MPIDense_Binary() line 682 in /home/marius/prog/petsc/git/petsc/src/mat/impls/dense/mpi/mpidense.c
> [18]PETSC ERROR: #2 MatView_MPIDense() line 786 in /home/marius/prog/petsc/git/petsc/src/mat/impls/dense/mpi/mpidense.c
> [18]PETSC ERROR: #3 MatView() line 1066 in /home/marius/prog/petsc/git/petsc/src/mat/interface/matrix.c
> [18]PETSC ERROR: #4 MatView_Elemental() line 83 in /home/marius/prog/petsc/git/petsc/src/mat/impls/elemental/matelem.cxx
> [18]PETSC ERROR: #5 MatView() line 1066 in /home/marius/prog/petsc/git/petsc/src/mat/interface/matrix.c
> [19]PETSC ERROR: --------------------- Error Message --------------------------------------------------------------
>
>
> and on StdOut
> Elemental matrix (explicit ordering)
>
> Any suggestions?
>
> Best,
> Marius

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20191203/6847cf98/attachment-0001.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: format.patch
Type: application/octet-stream
Size: 1781 bytes
Desc: format.patch
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20191203/6847cf98/attachment-0001.obj>


More information about the petsc-users mailing list