Agreed, the fix for the bug you found is now in maint I will try to do another MR that fixes this; but lots to do today so may take a while.
Barry > On Dec 4, 2019, at 1:02 AM, Marius Buerkle <mbuer...@web.de> wrote: > > thanks for the swift fix it works now. One more question though. It still > outputs "Elemental matrix (explicit ordering)" to StdOut which is kinda > annoying, is there anyway to turn this off? > > > Von: "Smith, Barry F." <bsm...@mcs.anl.gov> > An: "Marius Buerkle" <mbuer...@web.de> > Cc: "petsc-users@mcs.anl.gov" <petsc-users@mcs.anl.gov> > Betreff: Re: [petsc-users] MatView to disk for Elemental > > sorry about this. The numerical values between C and Fortran got out of > sync. I've attached a patch file you can apply with > > patch -p1 < format.patch > > > or you can use the branch https://gitlab.com/petsc/petsc/merge_requests/2346 > > Barry > > > > On Dec 3, 2019, at 1:10 AM, Marius Buerkle <mbuer...@web.de> wrote: > > > > Hi, > > > > I try to save a matrix in Elemental format to disk. I am doing, where > > p_matout is of type MATELEMENTAL, > > > > call PetscViewerCreate(PETSC_COMM_WORLD,v_file,ierr) > > call PetscViewerPushFormat(v_file,PETSC_VIEWER_NATIVE,ierr) > > call PetscViewerSetType(v_file,PETSCVIEWERBINARY,ierr) > > call PetscViewerFileSetMode(v_file,FILE_MODE_WRITE,ierr) > > call PetscViewerFileSetName(v_file,trim(filename),ierr) > > call MatView(p_matout,v_file,ierr) > > call PetscViewerDestroy(v_file,ierr) > > > > This gives the following error > > [18]PETSC ERROR: --------------------- Error Message > > -------------------------------------------------------------- > > [18]PETSC ERROR: No support for this operation for this object type > > [18]PETSC ERROR: To store a parallel dense matrix you must first call > > PetscViewerPushFormat(viewer,PETSC_VIEWER_NATIVE) > > [18]PETSC ERROR: See https://www.mcs.anl.gov/petsc/documentation/faq.html > > for trouble shooting. > > [18]PETSC ERROR: Petsc Development GIT revision: v3.12.1-317-gcc59f4f82c > > GIT Date: 2019-11-08 00:07:59 -0600 > > [18]PETSC ERROR: > > /home/marius/prog/ownstuff/fortran/programs/transomat_dev/save_load_fs/transomat/transomat > > on a named tono-hpc1 by marius Tue Dec 3 16:08:38 2019 > > [18]PETSC ERROR: Configure options > > --prefix=/home/marius/prog/petsc/petsc_slepc_opt --with-scalar-type=complex > > --with-fortran-kernels=1 --with-64-bit-indices=0 --CC=mpicc --COPTFLAGS="-g > > -Ofast -std=c11 -qopenmmp" --CXX=mpicxx --CXXOPTFLAGS="-g -Ofast -std=c++14 > > -qopenmp" --FC=mpif90 --FOPTFLAGS="-g -Ofast -traceback -qopenmp" > > --with-mpi=1 --with-x=0 --download-parmetis=1 --download-metis=1 > > --download-superlu_dist=1 > > --download-superlu_dist-commit=f8ace664ec4ca10e96e258a764552cbda299ba6e > > --download-superlu_dist-cmake-arguments=-Denable_openmp:BOOL=TRUE > > --download-hwloc=1 --download-sowing=1 --with-openmp=1 --with-pthread=1 > > --download-elemental=1 > > --download-elemental-commit=6eb15a0da2a4998bf1cf971ae231b78e06d989d9 > > --download-elemental-cmake-arguments=-DEL_HYBRID:BOOL=TRUE > > --with-cxx-dialect=c++11 --with-debugging=0 --with-valgrind=0 > > --with-blaslapack-lib=" > > /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_scalapack_lp64.a > > -Wl,--start-group > > /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_intel_lp64.a > > > > /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_intel_thread.a > > > > /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_core.a > > > > /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_blacs_openmpi_lp64.a > > -Wl,--end-group -liomp5 -lpthread -lm -ldl -lmpi_wrapper" > > --with-scalapack-lib=" > > /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_scalapack_lp64.a > > -Wl,--start-group > > /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_intel_lp64.a > > > > /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_intel_thread.a > > > > /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_core.a > > > > /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_blacs_openmpi_lp64.a > > -Wl,--end-group -liomp5 -lpthread -lm -ldl -lmpi_wrapper" > > --with-mkl_pardiso-dir=/home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl > > > > --with-mkl_cpardiso-dir=/home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl > > > > --with-mkl_sparse-dir=/home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl > > > > --with-mkl_sparse_optimize-dir=/home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl > > --download-slepc=1 --download-slepc-commit=658271f --download-make=1 > > --download-cmake=1 > > [18]PETSC ERROR: #1 MatView_MPIDense_Binary() line 682 in > > /home/marius/prog/petsc/git/petsc/src/mat/impls/dense/mpi/mpidense.c > > [18]PETSC ERROR: #2 MatView_MPIDense() line 786 in > > /home/marius/prog/petsc/git/petsc/src/mat/impls/dense/mpi/mpidense.c > > [18]PETSC ERROR: #3 MatView() line 1066 in > > /home/marius/prog/petsc/git/petsc/src/mat/interface/matrix.c > > [18]PETSC ERROR: #4 MatView_Elemental() line 83 in > > /home/marius/prog/petsc/git/petsc/src/mat/impls/elemental/matelem.cxx > > [18]PETSC ERROR: #5 MatView() line 1066 in > > /home/marius/prog/petsc/git/petsc/src/mat/interface/matrix.c > > [19]PETSC ERROR: --------------------- Error Message > > -------------------------------------------------------------- > > > > > > and on StdOut > > Elemental matrix (explicit ordering) > > > > Any suggestions? > > > > Best, > > Marius >