Re: [petsc-users] Error using Metis with PETSc installed with MUMPS

2023-11-07 Thread Victoria Rolandi
giorno mar 7 nov 2023 alle ore 12:25 Pierre Jolivet ha scritto: > > > On 7 Nov 2023, at 8:47 PM, Victoria Rolandi > wrote: > > Hi Pierre, > > Thanks for your reply. I am now trying to configure PETSc with the same > METIS/ParMETIS of my main code. > > I get the

Re: [petsc-users] Error using Metis with PETSc installed with MUMPS

2023-11-07 Thread Victoria Rolandi
:34 Pierre Jolivet ha scritto: > > > On 3 Nov 2023, at 7:28 PM, Victoria Rolandi > wrote: > > Pierre, > > Sure, I have now installed PETSc with MUMPS and PT-SCHOTCH, I got some > errors at the beginning but then it worked adding > --COPTFLAGS="-D_POSIX_C_SOURCE=1

Re: [petsc-users] Error using Metis with PETSc installed with MUMPS

2023-11-03 Thread Victoria Rolandi
an the one separately installed by PETSC during the configuration? Hence should I configure PETSc linking ParMetis to the same library used by my main code? Thanks, Victoria Il giorno gio 2 nov 2023 alle ore 09:35 Pierre Jolivet ha scritto: > > On 2 Nov 2023, at 5:29 PM, Victoria

Re: [petsc-users] Error using Metis with PETSc installed with MUMPS

2023-11-02 Thread Victoria Rolandi
edges > Ordering based on AMF > WARNING: Largest root node of size26 not selected for parallel > execution > > Leaving analysis phase with ... > INFOG(1) = 0 > INFOG(2) = 0 > […] > > Try parmetis. >

[petsc-users] Error using Metis with PETSc installed with MUMPS

2023-10-31 Thread Victoria Rolandi
Hi, I'm solving a large sparse linear system in parallel and I am using PETSc with MUMPS. I am trying to test different options, like the ordering of the matrix. Everything works if I use the *-mat_mumps_icntl_7 2 *or *-mat_mumps_icntl_7 0 *options (with the first one, AMF, performing better