On 4/6/2010 10:11 AM, Rainer Keller wrote:
> Hello Oliver,
> Hmm, this is really a teaser...
> I haven't seen such a drastic behavior, and haven't read of any on the list.
> 
> One thing however, that might interfere is process binding.
> Could You make sure, that processes are not bound to cores (default in 1.4.1):
> with mpirun --bind-to-none 
> 

I have tried version 1.4.1. Using default settings and watched processes
switching from core to core in "top" (with "f" + "j"). Then I tried
--bind-to-core and explicitly --bind-to-none. All with the same result:
~20% cpu wait and lot longer over-all computation times.

Thanks for the idea ...
Every input is helpful.

Oli


> Just an idea...
> 
> Regards,
> Rainer
> 
> On Tuesday 06 April 2010 10:07:35 am Oliver Geisler wrote:
>> Hello Devel-List,
>>
>> I am a little bit helpless about this matter. I already posted in the
>> user list. In case you don't read the users list, I post in here.
>>
>> This is the original posting:
>>
>> http://www.open-mpi.org/community/lists/users/2010/03/12474.php
>>
>> Short:
>> Switching from kernel 2.6.23 to 2.6.24 (and up), using openmpi 1.2.7-rc2
>> (I know outdated, but in debian stable, and same results with 1.4.1)
>> increases communication times between processes (essentially between one
>> master and several slave processes). This is regardless of whether the
>> processes are local only or communication is over ethernet.
>>
>> Did anybody witness such a behavior?
>>
>> Ideas what should I test for?
>>
>> What additional information should I provide for you?
>>
>> Thanks for your time
>>
>> oli
>>
> 


-- 
This message has been scanned for viruses and
dangerous content by MailScanner, and is
believed to be clean.

Reply via email to