Dear Kostas

This page (http://getfem.org/tutorial/install.html) says
- Parallel MUMPS, METIS and MPI4PY packages if you want to use the MPI
parallelized version of GetFEM.

Is there a recommended way to install Parallel Parallel MUMPS, METIS and
MPI4PY ?
I could not find the information in the page.

If you could give me any information I will add it to the following page.
http://getfem.org/install/install_linux.html

BR
Tetsuo

2021年5月19日(水) 10:45 Tetsuo Koyama <tkoyama...@gmail.com>:

> Dear Kostast
>
> No I haven't. I am using libmumps-seq-dev of Ubuntu repository.
> I will use parallel version of mumps again.
>
> BR
> Tetsuo
>
> 2021年5月19日(水) 4:50 Konstantinos Poulios <logar...@googlemail.com>:
>
>> Dear Tetsuo,
>>
>> Have you compiled GetFEM with the parallel version of mumps? In
>> Ubuntu/Debian you must link to dmumps instead of dmumps_seq for example.
>>
>> BR
>> Kostast
>>
>> On Tue, May 18, 2021 at 2:09 PM Tetsuo Koyama <tkoyama...@gmail.com>
>> wrote:
>>
>>> Dear Kostas
>>>
>>> Thank you for your report.
>>> I am happy that it runs well in your system.
>>> I will organize the procedure that can reproduce this error. Please wait.
>>>
>>> Best Regards Tetsuo
>>>
>>> 2021年5月18日(火) 18:10 Konstantinos Poulios <logar...@googlemail.com>:
>>>
>>>> Dear Tetsuo,
>>>> I could not confirm this issue. On my system the example runs well both
>>>> on 1 and 2 processes (it doesn't scale well though)
>>>> BR
>>>> Kostas
>>>>
>>>> [image: image.png]
>>>>
>>>>
>>>>
>>>>
>>>> On Sun, May 16, 2021 at 10:07 AM Tetsuo Koyama <tkoyama...@gmail.com>
>>>> wrote:
>>>>
>>>>> Dear Kostas
>>>>>
>>>>> I am looking inside the source code.
>>>>> > if (generic_expressions.size()) {...}
>>>>> Sorry it looks complex for me.
>>>>>
>>>>> FYI. I found that MPI process 1 and 2 is different in the following
>>>>> line.
>>>>> >    if (iter.finished(crit)) {
>>>>> This is in the "Newton_with_step_control" function in
>>>>> getfem_model_solvers.h.
>>>>>
>>>>> "crit" is calculated by rit = res / approx_eln and res and approx_eln
>>>>> is ...
>>>>>
>>>>> $ mpirun -n 1 python demo_parallel_laplacian.py
>>>>> res=1.31449e-11
>>>>> approx_eln=6.10757
>>>>> crit=2.15222e-12
>>>>>
>>>>> $ mpirun -n 2 python demo_parallel_laplacian.py
>>>>> res=6.02926
>>>>> approx_eln=12.2151
>>>>> crit=0.493588
>>>>>
>>>>> res=0.135744
>>>>> approx_eln=12.2151
>>>>> crit=0.0111128
>>>>>
>>>>> I am now trying to understand what is the correct residual value of
>>>>>  Newton(-Raphson) algorithm.
>>>>> I will be glad if you have an opinion.
>>>>>
>>>>> Best Regards Tetsuo
>>>>> 2021年5月11日(火) 19:28 Tetsuo Koyama <tkoyama...@gmail.com>:
>>>>>
>>>>>> Dear Kostas
>>>>>>
>>>>>> > The relevant code is in the void model::assembly function in
>>>>>> getfem_models.cc. The relevant code assembling the term you add with
>>>>>> md.add_nonlinear_term(..) must be executed inside the if condition
>>>>>> >
>>>>>> > if (generic_expressions.size()) {...}
>>>>>> > You can have a look there and ask for further help if it looks too
>>>>>> complex. You should also check if the test works when you run it with
>>>>>> md.add_nonlinear_term but setting the number of MPI processes to one.
>>>>>>
>>>>>> Thanks. I will check it. And the following command completed
>>>>>> successfully..
>>>>>>
>>>>>> $ mpirun -n 1 python demo_parallel_laplacian.py
>>>>>>
>>>>>> So all we have to check is compare -n 1 with -n2 .
>>>>>>
>>>>>> Best regards Tetsuo
>>>>>>
>>>>>> 2021年5月11日(火) 18:44 Konstantinos Poulios <logar...@googlemail.com>:
>>>>>>
>>>>>>> Dear Tetsuo,
>>>>>>>
>>>>>>> The relevant code is in the void model::assembly function in
>>>>>>> getfem_models.cc. The relevant code assembling the term you add with
>>>>>>> md.add_nonlinear_term(..) must be executed inside the if condition
>>>>>>>
>>>>>>> if (generic_expressions.size()) {...}
>>>>>>>
>>>>>>> You can have a look there and ask for further help if it looks too
>>>>>>> complex. You should also check if the test works when you run it with
>>>>>>> md.add_nonlinear_term but setting the number of MPI processes to one.
>>>>>>>
>>>>>>> BR
>>>>>>> Kostas
>>>>>>>
>>>>>>>
>>>>>>> On Tue, May 11, 2021 at 10:44 AM Tetsuo Koyama <tkoyama...@gmail.com>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Dear Kostas
>>>>>>>>
>>>>>>>> Thank you for your reply.
>>>>>>>>
>>>>>>>> > Interesting. In order to isolate the issue, can you also check
>>>>>>>> with
>>>>>>>> > md.add_linear_term(..)
>>>>>>>> > ?
>>>>>>>> It ends when using md.add_linear_term(..).
>>>>>>>> It seems that it is a problem of md.add_nonlinear_term(..).
>>>>>>>> Is there a point which I can check?
>>>>>>>>
>>>>>>>> Best regards Tetsuo.
>>>>>>>>
>>>>>>>> 2021年5月11日(火) 17:19 Konstantinos Poulios <logar...@googlemail.com>:
>>>>>>>>
>>>>>>>>> Dear Tetsuo,
>>>>>>>>>
>>>>>>>>> Interesting. In order to isolate the issue, can you also check with
>>>>>>>>> md.add_linear_term(..)
>>>>>>>>> ?
>>>>>>>>>
>>>>>>>>> Best regards
>>>>>>>>> Kostas
>>>>>>>>>
>>>>>>>>> On Tue, May 11, 2021 at 12:22 AM Tetsuo Koyama <
>>>>>>>>> tkoyama...@gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> Dear GetFEM community
>>>>>>>>>>
>>>>>>>>>> I am running MPI Parallelization of GetFEM.The running command is
>>>>>>>>>>
>>>>>>>>>> $ git clone https://git.savannah.nongnu.org/git/getfem.git
>>>>>>>>>> $ cd getfem
>>>>>>>>>> $ bash autogen.sh
>>>>>>>>>> $ ./configure --with-pic --enable-paralevel=2
>>>>>>>>>> $ make
>>>>>>>>>> $ make install
>>>>>>>>>> $ mpirun -n 2 python demo_parallel_laplacian.py
>>>>>>>>>>
>>>>>>>>>> The python script ends correctly. But when I changed the
>>>>>>>>>> following linear term to nonlinear term the script did not end.
>>>>>>>>>>
>>>>>>>>>> -md.add_Laplacian_brick(mim, 'u')
>>>>>>>>>> +md.add_nonlinear_term(mim, "Grad_u.Grad_Test_u")
>>>>>>>>>>
>>>>>>>>>> Do you know the reason?
>>>>>>>>>> Best regards Tetsuo
>>>>>>>>>>
>>>>>>>>>

Reply via email to