Hi Atgeirr,

I use AMG as preconditioner  (—use-amg=true). When the ParMETIS errors occur, 
the simulation crashes.
I indeed linked to the ParMETIS library when configuring DUNE.

Do you actually advise to use PTScotch instead of ParMETIS? I could try to 
recompile Dune + OPM with PTScotch to see if the simulation runs with this 
configuration.

Thanks for your answer.

Antoine

> On Oct 20, 2020, at 05:43, Atgeirr Rasmussen <atgeirr.rasmus...@sintef.no> 
> wrote:
> 
> Hi Antoine!
> 
> Our partitioning scheme starts with the whole graph on a single process, so 
> indeed this would be a "bad" starting partition. The partitioning we end up 
> with does not seem any worse though, although for very large process counts, 
> this could become a bottleneck.
> 
> I am a little confused though, as OPM Flow uses Zoltan for partitioning, not 
> ParMETIS. This is because ParMETIS is not open source. However, if you do 
> have access to ParMETIS I believe you can configure the dune-istl parallel 
> linear solvers (that in turn are used by OPM Flow) to use ParMETIS (or the 
> PTScotch workalike library) for redistribution of coarse systems within the 
> algebraic multigrid (AMG) solver. However, that is not the default linear 
> solver for OPM Flow. So I am a bit at a loss, as to where those ParMETIS 
> messages come from! Did you run with the default linear solver or not? I 
> assume that the simulation actually runs?
> 
> Atgeirr
> ________________________________
> From: Opm <opm-boun...@opm-project.org> on behalf of Antoine B Jacquey 
> <ajacq...@mit.edu>
> Sent: 19 October 2020 14:22
> To: opm@opm-project.org <opm@opm-project.org>
> Subject: [Opm] ParMETIS error on HPC
> 
> Hi OPM community,
> 
> I recently compiled OPM Flow on a local cluster in my institute. I linked to 
> the PartMETIS library during configuration to make use of mesh partitioning 
> when using large number of MPI processes.
> When I run a flow simulation, it seems that the mesh is partitioned 
> automatically. Here is part of the output I get for a simulation with 8 MPI 
> processes:
> 
> Load balancing distributes 300000 active cells on 8 processes as follows:
>  rank   owned cells   overlap cells   total cells
> --------------------------------------------------
>     0         36960            2760         39720
>     1         40110            3720         43830
>     2         38100            4110         42210
>     3         38940            2250         41190
>     4         36600            2280         38880
>     5         33660            3690         37350
>     6         37800            3690         41490
>     7         37830            2730         40560
> --------------------------------------------------
>   sum        300000           25230        325230
> 
> The problem occurs when I use a larger number of MPI processes (here for 27 
> MPI processes). The mesh is also partitioned:
> 
> Load balancing distributes 1012500 active cells on 27 processes as follows:
>  rank   owned cells   overlap cells   total cells
> --------------------------------------------------
>     0         40230            6390         46620
>     1         40185            5175         45360
>     2         40635            4050         44685
>     3         40230            6255         46485
>     4         40905            5850         46755
>     5         39825            6030         45855
>     6         37035            2610         39645
>     7         36945            5625         42570
>     8         40680            4185         44865
>     9         35835            5460         41295
>    10         41250            6765         48015
>    11         39825            5310         45135
>    12         36855            2655         39510
>    13         32850            3690         36540
>    14         38790            5400         44190
>    15         36540            5625         42165
>    16         30105            3105         33210
>    17         40320            5400         45720
>    18         35685            4185         39870
>    19         39465            5490         44955
>    20         20160            1800         21960
>    21         39915            4860         44775
>    22         40050            6165         46215
>    23         34020            2475         36495
>    24         39645            6345         45990
>    25         36990            6570         43560
>    26         37530            4005         41535
> --------------------------------------------------
>   sum       1012500          131475       1143975
> 
> But during the first time step calculation, I get the following errors:
> 
> Time step 0, stepsize 1 days, at day 0/7, date = 01-Jan-2015
>    Switching control mode for well INJ from RATE to BHP on rank 20
>    Switching control mode for well INJ from BHP to RATE on rank 20
> PARMETIS ERROR: Poor initial vertex distribution. Processor 2 has no vertices 
> assigned to it!
> PARMETIS ERROR: Poor initial vertex distribution. Processor 4 has no vertices 
> assigned to it!
> PARMETIS ERROR: Poor initial vertex distribution. Processor 6 has no vertices 
> assigned to it!
> PARMETIS ERROR: Poor initial vertex distribution. Processor 8 has no vertices 
> assigned to it!
> PARMETIS ERROR: Poor initial vertex distribution. Processor 12 has no 
> vertices assigned to it!
> PARMETIS ERROR: Poor initial vertex distribution. Processor 14 has no 
> vertices assigned to it!
> PARMETIS ERROR: Poor initial vertex distribution. Processor 16 has no 
> vertices assigned to it!
> PARMETIS ERROR: Poor initial vertex distribution. Processor 18 has no 
> vertices assigned to it!
> PARMETIS ERROR: Poor initial vertex distribution. Processor 20 has no 
> vertices assigned to it!
> PARMETIS ERROR: Poor initial vertex distribution. Processor 0 has no vertices 
> assigned to it!
> PARMETIS ERROR: Poor initial vertex distribution. Processor 10 has no 
> vertices assigned to it!
> PARMETIS ERROR: Poor initial vertex distribution. Processor 22 has no 
> vertices assigned to it!
> PARMETIS ERROR: Poor initial vertex distribution. Processor 26 has no 
> vertices assigned to it!
> PARMETIS ERROR: Poor initial vertex distribution. Processor 24 has no 
> vertices assigned to it!
> 
> Does anyone know what this error means? Is it coming because of a bad mesh 
> partitioning or is it due to something else?
> 
> I would appreciate any advice or tip to solve this issue.
> Thank you in advance.
> 
> Best,
> 
> Antoine
> _______________________________________________
> Opm mailing list
> Opm@opm-project.org
> https://opm-project.org/cgi-bin/mailman/listinfo/opm

Attachment: smime.p7s
Description: S/MIME cryptographic signature

_______________________________________________
Opm mailing list
Opm@opm-project.org
https://opm-project.org/cgi-bin/mailman/listinfo/opm

Reply via email to