Re: [petsc-users] unable to build petsc with cuda / superlu-dist enabled

2022-08-26 Thread Barry Smith
Glad it worked. The plan is the next release will not require these hacks. > On Aug 26, 2022, at 12:13 PM, byron wrote: > > It looks like I've (we've) cracked it, at least the petsc stage of what I'm > trying to do. To get the configure to work I used the following > > python3

Re: [petsc-users] unable to build petsc with cuda / superlu-dist enabled

2022-08-26 Thread byron
It looks like I've (we've) cracked it, at least the petsc stage of what I'm trying to do. To get the configure to work I used the following python3 ./configure --download-hypre=1 --with-shared-libraries=1 --download-hdf5=1 --download-hdf5-fortran-bindings=0 --with-debugging=no

Re: [petsc-users] unable to build petsc with cuda / superlu-dist enabled

2022-08-26 Thread Barry Smith
Yes, PETSc sends down a single -O3 to SuperLU_DIST CMake but internally SuperLU_DIST CMake is adding its own set of nvcc compiler flags to use and these that get added to the one's PETSc passes in (they do not replace them). A little further down in the file it displays a compile line with

Re: [petsc-users] unable to build petsc with cuda / superlu-dist enabled

2022-08-26 Thread byron
Hi Jacob I tried with "--with-cxx-dialect=auto" which did look to clean up all the references to c++14 / c++11 from the options used to configure superlu-dist but it still didnt work and looked to generate the same "nvcc fatal" errors. Thanks Byron On Fri, Aug 26, 2022 at 2:47 PM byron wrote:

Re: [petsc-users] DMPlex Interpolation

2022-08-26 Thread Matthew Knepley
On Thu, Aug 25, 2022 at 7:12 PM Mike Michell wrote: > Hi, this is a duplication of > https://lists.mcs.anl.gov/pipermail/petsc-users/2022-August/046746.html > for in-depth question. > > I wrote a short code as attached to test interpolation between two DMPlex > objects. Goal is to map solution

Re: [petsc-users] DMPlex in Fortran

2022-08-26 Thread Matthew Knepley
On Thu, Aug 25, 2022 at 4:25 PM Mike Michell wrote: > Thank you for the quick note. > To use PetscFECreateLagrange(), it is unclear about the meaning of "Nc", > which is explained as 'number of components' on the manual page. > I looked at an example with Nc=1 in >

Re: [petsc-users] Slepc Question, shell-matrix

2022-08-26 Thread feng wang
Hi Jose, Thanks for your reply. Then using finite difference in the shell matrix to compute the matrix-vector product is not a good idea here. I will try to find an alternative way. maybe assemble the matrix directly. Thanks, Feng From: Jose E. Roman Sent: 25

Re: [petsc-users] unable to build petsc with cuda / superlu-dist enabled

2022-08-26 Thread byron
Hi Barry I dont think I see the same thing. When I do a build using the release branch petsc passes the following arguements to superlu_dist (taken from /home/blahblah/Projects/220826-vanilla-petsc-release-branch/petsc/arch-linux-c-opt/externalpackages/git.superlu_dist/superlu_dist.petscconf)

[petsc-users] metis version of petsc - segmentation fault

2022-08-26 Thread Alexandre Halbach
Hello, I am using the metis package provided by the latest petsc version, using --download-metis (Linux, 32 bits). This question is arguably more a metis issue but since metis is used in petsc and mumps this seems relevant to ask here too. I systematically get a segmentation fault on this metis