Hi again I installed petsc3.4.0 and I am still getting the following error when running with the following options (on 64 procs)
# Matrix Options -matload_block_size 5 -mat_type mpibaij # KSP solver options -ksp_type fgmres -ksp_max_it 1000 -ksp_gmres_restart 200 -ksp_monitor -ksp_view -ksp_pc_side right -ksp_rtol 1e-6 # Nested GMRES Options -pc_type bjacobi -pc_bjacobi_blocks 4 -sub_ksp_type gmres -sub_ksp_max_it 5 -sub_pc_type bjacobi -sub_sub_pc_type ilu -sub_sub_pc_factor_mat_ordering_type rcm -sub_sub_pc_factor_levels 1 Any thoughts? Thank you, Gaetan [44]PETSC ERROR: ------------------------------------------------------------------------ [44]PETSC ERROR: Caught signal number 11 SEGV: Segmentation Violation, probably memory access out of range [44]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger [44]PETSC ERROR: or see http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind[44]PETSCERROR: or try http://valgrind.org on GNU/linux and Apple Mac OS X to find memory corruption errors [44]PETSC ERROR: likely location of problem given in stack below [44]PETSC ERROR: --------------------- Stack Frames ------------------------------------ [44]PETSC ERROR: Note: The EXACT line numbers in the stack are not available, [44]PETSC ERROR: INSTEAD the line number of the start of the function [44]PETSC ERROR: is given. [44]PETSC ERROR: [44] PCSetUp_BJacobi_Multiproc line 1197 /home/j/jmartins/kenway/packages/petsc-3.4.0/src/ksp/pc/impls/bjacobi/bjacobi.c [44]PETSC ERROR: [44] PCSetUp_BJacobi line 24 /home/j/jmartins/kenway/packages/petsc-3.4.0/src/ksp/pc/impls/bjacobi/bjacobi.c [44]PETSC ERROR: [44] PCSetUp line 868 /home/j/jmartins/kenway/packages/petsc-3.4.0/src/ksp/pc/interface/precon.c [44]PETSC ERROR: [44] KSPSetUp line 192 /home/j/jmartins/kenway/packages/petsc-3.4.0/src/ksp/ksp/interface/itfunc.c [44]PETSC ERROR: [44] KSPSolve line 356 /home/j/jmartins/kenway/packages/petsc-3.4.0/src/ksp/ksp/interface/itfunc.c [43]PETSC ERROR: ------------------------------------------------------------------------ On Sun, May 19, 2013 at 11:15 PM, Barry Smith <[email protected]> wrote: > > You should be using PETSc version 3.4 which was recently released and > is what the paper is based on. > > Barry > > On May 19, 2013, at 10:11 PM, Gaetan Kenway <[email protected]> wrote: > > > Hi Everyone > > > > I am trying to replicate the type of preconditioner described in > "Hierarchical and Nested Krylov Methods for Extreme-Scale Computing". > > > > I have used the following options: (I'm using fortran so the following > is my petsc_options file) > > > > # Matrix Options > > -matload_block_size 5 > > -mat_type mpibaij > > > > # KSP solver options > > -ksp_type gmres > > -ksp_max_it 1000 > > -ksp_gmres_restart 200 > > -ksp_monitor > > -ksp_view > > -ksp_pc_side right > > -ksp_rtol 1e-6 > > > > # Nested GMRES Options > > -pc_type bjacobi > > -pc_bjacobi_blocks 4 > > -sub_ksp_type gmres > > -sub_ksp_max_it 5 > > -sub_pc_type bjacobi > > -sub_sub_pc_type ilu > > -sub_sub_pc_factor_mat_ordering_type rcm > > -sub_sub_pc_factor_levels 1 > > > > The test is run on 64 processors and the total number of block jacobi > blocks is 4 (less than nproc). The error I get is: > > > > [6]PETSC ERROR: Note: The EXACT line numbers in the stack are not > available, > > [6]PETSC ERROR: INSTEAD the line number of the start of the > function > > [6]PETSC ERROR: is given. > > [6]PETSC ERROR: [6] PCSetUp_BJacobi_Multiproc line 1269 > /home/j/jmartins/kenway/packages/petsc-3.3-p5/src/ksp/pc/impls/bjacobi/bjacobi.c > > [6]PETSC ERROR: [6] PCSetUp_BJacobi line 24 > /home/j/jmartins/kenway/packages/petsc-3.3-p5/src/ksp/pc/impls/bjacobi/bjacobi.c > > [6]PETSC ERROR: [6] PCSetUp line 810 > /home/j/jmartins/kenway/packages/petsc-3.3-p5/src/ksp/pc/interface/precon.c > > [6]PETSC ERROR: [6] KSPSetUp line 182 > /home/j/jmartins/kenway/packages/petsc-3.3-p5/src/ksp/ksp/interface/itfunc.c > > [6]PETSC ERROR: [6] KSPSolve line 351 > /home/j/jmartins/kenway/packages/petsc-3.3-p5/src/ksp/ksp/interface/itfunc.c > > [6]PETSC ERROR: --------------------- Error Message > ------------------------------------ > > [6]PETSC ERROR: Signal received! > > [6]PETSC ERROR: > ------------------------------------------------------------------------ > > [6]PETSC ERROR: Petsc Release Version 3.3.0, Patch 5, Sat Dec 1 > 15:10:41 CST 2012 > > [6]PETSC ERROR: See docs/changes/index.html for recent updates. > > [6]PETSC ERROR: See docs/faq.html for hints about trouble shooting. > > [6]PETSC ERROR: > ------------------------------------------------------------------------ > > [6]PETSC ERROR: > ------------------------------------------------------------------------ > > [6]PETSC ERROR: ./main on a intel-rea named gpc-f109n001 by kenway Sun > May 19 23:01:52 2013 > > [6]PETSC ERROR: Libraries linked from > /home/j/jmartins/kenway/packages/petsc-3.3-p5/intel-real-debug/lib > > [6]PETSC ERROR: Configure run at Sun Jan 20 15:52:20 2013 > > [6]PETSC ERROR: Configure options --with-shared-libraries > --download-superlu_dist=yes --download-parmetis=yes --download-metis=yes > --with-fortran-interfaces=1 --with-debugging=yes --with-scalar-type=real > -with-petsc-arch=intel-real-debug --with-blas-lapack-dir= --with-pic > > [6]PETSC ERROR: > ------------------------------------------------------------------------ > > > > If the number of blocks is greater than or equal to the number of > processors it runs fine. I'm using version 3.3-p5. > > > > The options as listed in the paper are: > > -flow_ksp_type fgmres -flow_ksp_pc_side right -flow_pc_type bjacobi > -flow_pc_bjacobi_blocks ngp > > -flow_sub_ksp_type gmres -flow_sub_ksp_max_it 6 -flow_sub_pc_type bjacobi > > -flow_sub_sub_pc_type ilu > > > > Any suggestions would be greatly appreciated. > > > > Thank you, > > > > Gaetan Kenway > > > > > >
