On Thu, Oct 13, 2011 at 23:01, <Kevin.Buckley at ecs.vuw.ac.nz> wrote:
> Hi there, > > Some time ago I built a PETSc (3.0.0-p9) on the BG/P down at NZ's > Univ of Canterbury, for a researcher here at VUW who wanted to > run the PISM ice-sheet modelling code on top of it, and in the > process discovered that the version installed by the sys admins > at the facility was buggy, so the researcher carried on compiling > later PISMs against my 3.0.0-p9. > > UoC are in the process of upgrading to a BG/P and the researcher > has asked me to see if I can coddle things together, is keen to > run with what he had before and is under "time pressure" to get > the remaining results needed for a paper, so is hoping I can > install something ahead of the facility's sys admins install > a system wide version. > > Notwithstanding the obvious upgrade in release of PETSc from our > 3.0 series to the current 3.2, I notice that the petsc-bgl-tools > wrapper package only (not surprisingly I guess, given the bgl bit) > provided for IBM 7.0 and 8.0 compiler suites, so, have you guys > tried out PETSc on a BG/P yet? > BG/P has been out for a while, so of course people have been running PETSc on it for years. You can look at config/examples/arch-bgp-ibm-opt.py in the source tree. Here is a configuration that we used for some benchmarks on Shaheen last year: Configure options: --with-x=0 --with-is-color-value-type=short --with-debugging=1 --with-fortran-kernels=1 --with-mpi-dir=/bgsys/drivers/ppcfloor/comm --with-batch=1 --known-mpi-shared-libraries=1 --known-memcmp-ok --known-sizeof-char=1 --known-sizeof-void-p=4 --known-sizeof-short=2 --known-sizeof-int=4 --known-sizeof-long=4 --known-sizeof-size_t=4 --known-sizeof-long-long=8 --known-sizeof-float=4 --known-sizeof-double=8 --known-bits-per-byte=8 --known-sizeof-MPI_Comm=4 --known-sizeof-MPI_Fint=4 --known-mpi-long-double=1 --known-level1-dcache-assoc=0 --known-level1-dcache-linesize=32 --known-level1-dcache-size=32768 --download-hypre=1 --with-shared=0 --prefix=/opt/share/ksl/petsc/dev-dec9-hypre/ppc450d-bgp_xlc_hypre_fast --with-clanguage=c --COPTFLAGS=" -O3 -qhot" --CXXOPTFLAGS=" -O3 -qhot" --FOPTFLAGS=" -O3 -qhot" --LIBS=" -L/bgsys/ibm_essl/sles10/prod/opt/ibmmath/lib -L/opt/ibmcmp/xlsmp/bg/1.7/lib -L/opt/ibmcmp/xlmass/bg/4.4/bglib -L/opt/ibmcmp/xlf/bg/11.1/bglib -L/bgsys/ibm_essl/sles10/prod/opt/ibmmath/lib -L/opt/ibmcmp/xlsmp/bg/1.7/lib -L/opt/ibmcmp/xlmass/bg/4.4/bglib -L/opt/ibmcmp/xlf/bg/11.1/bglib -lesslbg -lxlf90_r -lxlopt -lxlsmp -lxl -lxlfmath -lesslbg -lxlf90_r -lxlopt -lxlsmp -lxl -lxlfmath -O3" --CC=/bgsys/drivers/ppcfloor/comm/xl/bin/mpixlc_r --CXX=/bgsys/drivers/ppcfloor/comm/xl/bin/mpixlcxx_r --FC=/bgsys/drivers/ppcfloor/comm/xl/bin/mpixlf90_r --with-debugging=0 PETSC_ARCH=bgp-xlc-hypre-fast -------------- next part -------------- An HTML attachment was scrubbed... URL: <http://lists.mcs.anl.gov/pipermail/petsc-dev/attachments/20111013/c2852e5c/attachment.html>
