Agreed, the fix for the bug you found is now in maint I will try to do 
another MR that fixes this; but lots to do today so may take a while.

   Barry



> On Dec 4, 2019, at 1:02 AM, Marius Buerkle <mbuer...@web.de> wrote:
> 
> thanks for the swift fix it works now. One more question though. It still 
> outputs "Elemental matrix (explicit ordering)" to StdOut which is kinda 
> annoying, is there anyway to turn this off?
>  
> 
> Von: "Smith, Barry F." <bsm...@mcs.anl.gov>
> An: "Marius Buerkle" <mbuer...@web.de>
> Cc: "petsc-users@mcs.anl.gov" <petsc-users@mcs.anl.gov>
> Betreff: Re: [petsc-users] MatView to disk for Elemental
> 
>   sorry about this. The numerical values between C and Fortran got out of 
> sync. I've attached a patch file you can apply with
> 
>   patch -p1 < format.patch
> 
> 
>   or you can use the branch https://gitlab.com/petsc/petsc/merge_requests/2346
> 
>    Barry
> 
> 
> > On Dec 3, 2019, at 1:10 AM, Marius Buerkle <mbuer...@web.de> wrote:
> >
> > Hi,
> > 
> > I try to save a matrix in Elemental format to disk. I am doing, where 
> > p_matout is of type MATELEMENTAL,
> > 
> >     call PetscViewerCreate(PETSC_COMM_WORLD,v_file,ierr)   
> >     call PetscViewerPushFormat(v_file,PETSC_VIEWER_NATIVE,ierr)
> >     call PetscViewerSetType(v_file,PETSCVIEWERBINARY,ierr)   
> >     call PetscViewerFileSetMode(v_file,FILE_MODE_WRITE,ierr)   
> >     call PetscViewerFileSetName(v_file,trim(filename),ierr)       
> >     call MatView(p_matout,v_file,ierr)
> >     call PetscViewerDestroy(v_file,ierr)  
> > 
> > This gives the following error
> > [18]PETSC ERROR: --------------------- Error Message 
> > --------------------------------------------------------------
> > [18]PETSC ERROR: No support for this operation for this object type
> > [18]PETSC ERROR: To store a parallel dense matrix you must first call 
> > PetscViewerPushFormat(viewer,PETSC_VIEWER_NATIVE)
> > [18]PETSC ERROR: See https://www.mcs.anl.gov/petsc/documentation/faq.html 
> > for trouble shooting.
> > [18]PETSC ERROR: Petsc Development GIT revision: v3.12.1-317-gcc59f4f82c  
> > GIT Date: 2019-11-08 00:07:59 -0600
> > [18]PETSC ERROR: 
> > /home/marius/prog/ownstuff/fortran/programs/transomat_dev/save_load_fs/transomat/transomat
> >  on a  named tono-hpc1 by marius Tue Dec  3 16:08:38 2019
> > [18]PETSC ERROR: Configure options 
> > --prefix=/home/marius/prog/petsc/petsc_slepc_opt --with-scalar-type=complex 
> > --with-fortran-kernels=1 --with-64-bit-indices=0 --CC=mpicc --COPTFLAGS="-g 
> > -Ofast -std=c11 -qopenmmp" --CXX=mpicxx --CXXOPTFLAGS="-g -Ofast -std=c++14 
> > -qopenmp" --FC=mpif90 --FOPTFLAGS="-g -Ofast -traceback -qopenmp" 
> > --with-mpi=1 --with-x=0 --download-parmetis=1 --download-metis=1 
> > --download-superlu_dist=1 
> > --download-superlu_dist-commit=f8ace664ec4ca10e96e258a764552cbda299ba6e 
> > --download-superlu_dist-cmake-arguments=-Denable_openmp:BOOL=TRUE 
> > --download-hwloc=1 --download-sowing=1 --with-openmp=1 --with-pthread=1 
> > --download-elemental=1 
> > --download-elemental-commit=6eb15a0da2a4998bf1cf971ae231b78e06d989d9 
> > --download-elemental-cmake-arguments=-DEL_HYBRID:BOOL=TRUE 
> > --with-cxx-dialect=c++11 --with-debugging=0 --with-valgrind=0 
> > --with-blaslapack-lib=" 
> > /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_scalapack_lp64.a
> >  -Wl,--start-group 
> > /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_intel_lp64.a
> >  
> > /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_intel_thread.a
> >  
> > /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_core.a
> >  
> > /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_blacs_openmpi_lp64.a
> >  -Wl,--end-group -liomp5 -lpthread -lm -ldl -lmpi_wrapper" 
> > --with-scalapack-lib=" 
> > /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_scalapack_lp64.a
> >  -Wl,--start-group 
> > /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_intel_lp64.a
> >  
> > /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_intel_thread.a
> >  
> > /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_core.a
> >  
> > /home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl/lib/intel64/libmkl_blacs_openmpi_lp64.a
> >  -Wl,--end-group -liomp5 -lpthread -lm -ldl -lmpi_wrapper" 
> > --with-mkl_pardiso-dir=/home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl
> >  
> > --with-mkl_cpardiso-dir=/home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl
> >  
> > --with-mkl_sparse-dir=/home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl
> >  
> > --with-mkl_sparse_optimize-dir=/home/marius/intel/compilers_and_libraries_2019.5.281/linux/mkl
> >  --download-slepc=1 --download-slepc-commit=658271f --download-make=1 
> > --download-cmake=1
> > [18]PETSC ERROR: #1 MatView_MPIDense_Binary() line 682 in 
> > /home/marius/prog/petsc/git/petsc/src/mat/impls/dense/mpi/mpidense.c
> > [18]PETSC ERROR: #2 MatView_MPIDense() line 786 in 
> > /home/marius/prog/petsc/git/petsc/src/mat/impls/dense/mpi/mpidense.c
> > [18]PETSC ERROR: #3 MatView() line 1066 in 
> > /home/marius/prog/petsc/git/petsc/src/mat/interface/matrix.c
> > [18]PETSC ERROR: #4 MatView_Elemental() line 83 in 
> > /home/marius/prog/petsc/git/petsc/src/mat/impls/elemental/matelem.cxx
> > [18]PETSC ERROR: #5 MatView() line 1066 in 
> > /home/marius/prog/petsc/git/petsc/src/mat/interface/matrix.c
> > [19]PETSC ERROR: --------------------- Error Message 
> > --------------------------------------------------------------
> > 
> > 
> > and on StdOut
> > Elemental matrix (explicit ordering)
> > 
> > Any suggestions?
> > 
> > Best,
> > Marius
>  

Reply via email to