Re: [OMPI devel] segfault on host not found error.

2008-04-01 Thread Lenny Verkhovsky
yes, it seems to be fixed.
thanks.

On Mon, Mar 31, 2008 at 9:17 PM, Ralph H Castain  wrote:

> I am unable to replicate the segfault. However, I was able to get the job
> to
> hang. I fixed that behavior with r18044.
>
> Perhaps you can test this again and let me know what you see. A gdb stack
> trace would be more helpful.
>
> Thanks
> Ralph
>
>
>
> On 3/31/08 5:13 AM, "Lenny Verkhovsky"  wrote:
>
> >
> >
> >
> > I accidently run job on the hostfile where one of hosts was not properly
> > mounted. As a result I got an error and a segfault.
> >
> >
> > /home/USERS/lenny/OMPI_ORTE_TRUNK/bin/mpirun -np 29 -hostfile hostfile
> > ./mpi_p01 -t lt
> > bash: /home/USERS/lenny/OMPI_ORTE_TRUNK/bin/orted: No such file or
> > directory
> > 
> > --
> > A daemon (pid 9753) died unexpectedly with status 127 while attempting
> > to launch so we are aborting.
> >
> > There may be more information reported by the environment (see above).
> >
> > This may be because the daemon was unable to find all the needed shared
> > libraries on the remote node. You may set your LD_LIBRARY_PATH to have
> > the
> > location of the shared libraries on the remote nodes and this will
> > automatically be forwarded to the remote nodes.
> > 
> > --
> > 
> > --
> > mpirun was unable to start the specified application as it encountered
> > an error.
> > More information may be available above.
> > 
> > --
> > [witch1:09745] *** Process received signal ***
> > [witch1:09745] Signal: Segmentation fault (11)
> > [witch1:09745] Signal code: Address not mapped (1)
> > [witch1:09745] Failing at address: 0x3c
> > [witch1:09745] [ 0] /lib64/libpthread.so.0 [0x2aff223ebc10]
> > [witch1:09745] [ 1]
> > /home/USERS/lenny/OMPI_ORTE_TRUNK//lib/libopen-rte.so.0 [0x2aff21cdfe21]
> > [witch1:09745] [ 2]
> > /home/USERS/lenny/OMPI_ORTE_TRUNK//lib/openmpi/mca_rml_oob.so
> > [0x2aff22c398f1]
> > [witch1:09745] [ 3]
> > /home/USERS/lenny/OMPI_ORTE_TRUNK//lib/openmpi/mca_oob_tcp.so
> > [0x2aff22d426ee]
> > [witch1:09745] [ 4]
> > /home/USERS/lenny/OMPI_ORTE_TRUNK//lib/openmpi/mca_oob_tcp.so
> > [0x2aff22d433fb]
> > [witch1:09745] [ 5]
> > /home/USERS/lenny/OMPI_ORTE_TRUNK//lib/openmpi/mca_oob_tcp.so
> > [0x2aff22d4485b]
> > [witch1:09745] [ 6]
> > /home/USERS/lenny/OMPI_ORTE_TRUNK//lib/libopen-pal.so.0 [0x2aff21e1242b]
> > [witch1:09745] [ 7] /home/USERS/lenny/OMPI_ORTE_TRUNK/bin/mpirun
> > [0x403203]
> > [witch1:09745] [ 8]
> > /home/USERS/lenny/OMPI_ORTE_TRUNK//lib/libopen-pal.so.0 [0x2aff21e1242b]
> > [witch1:09745] [ 9]
> > /home/USERS/lenny/OMPI_ORTE_TRUNK//lib/libopen-pal.so.0(opal_progress+0x
> > 8b) [0x2aff21e060cb]
> > [witch1:09745] [10]
> > /home/USERS/lenny/OMPI_ORTE_TRUNK//lib/libopen-rte.so.0(orte_trigger_eve
> > nt+0x20) [0x2aff21cc6940]
> > [witch1:09745] [11]
> > /home/USERS/lenny/OMPI_ORTE_TRUNK//lib/libopen-rte.so.0(orte_wakeup+0x2d
> > ) [0x2aff21cc776d]
> > [witch1:09745] [12]
> > /home/USERS/lenny/OMPI_ORTE_TRUNK//lib/openmpi/mca_plm_rsh.so
> > [0x2aff22b34756]
> > [witch1:09745] [13]
> > /home/USERS/lenny/OMPI_ORTE_TRUNK//lib/libopen-rte.so.0 [0x2aff21cc6ea7]
> > [witch1:09745] [14]
> > /home/USERS/lenny/OMPI_ORTE_TRUNK//lib/libopen-pal.so.0 [0x2aff21e1242b]
> > [witch1:09745] [15]
> > /home/USERS/lenny/OMPI_ORTE_TRUNK//lib/libopen-pal.so.0(opal_progress+0x
> > 8b) [0x2aff21e060cb]
> > [witch1:09745] [16]
> > /home/USERS/lenny/OMPI_ORTE_TRUNK//lib/libopen-rte.so.0(orte_plm_base_da
> > emon_callback+0xad) [0x2aff21ce068d]
> > [witch1:09745] [17]
> > /home/USERS/lenny/OMPI_ORTE_TRUNK//lib/openmpi/mca_plm_rsh.so
> > [0x2aff22b34e5e]
> > [witch1:09745] [18] /home/USERS/lenny/OMPI_ORTE_TRUNK/bin/mpirun
> > [0x402e13]
> > [witch1:09745] [19] /home/USERS/lenny/OMPI_ORTE_TRUNK/bin/mpirun
> > [0x402873]
> > [witch1:09745] [20] /lib64/libc.so.6(__libc_start_main+0xf4)
> > [0x2aff22512154]
> > [witch1:09745] [21] /home/USERS/lenny/OMPI_ORTE_TRUNK/bin/mpirun
> > [0x4027c9]
> > [witch1:09745] *** End of error message ***
> > Segmentation fault (core dumped)
> >
> >
> > Best Regards,
> > Lenny.
> >
> >
> > ___
> > devel mailing list
> > de...@open-mpi.org
> > http://www.open-mpi.org/mailman/listinfo.cgi/devel
>
>
> ___
> devel mailing list
> de...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/devel
>


Re: [OMPI devel] RMAPS rank_file component patch and modifications for review

2008-04-01 Thread Jeff Squyres

Unfortunately, we have no way to "alias" an MCA param.  :-\

This topic has come up a few times over the past few years, but no  
one's actually extended the MCA params infrastructure to support  
aliasing.  I'm guessing that it wouldn't be too hard to do...



On Apr 1, 2008, at 5:22 AM, Lenny Verkhovsky wrote:

Hi,
is there any elegant way to register mpi parameter that will  
actually be pointer or alias to hidden opal parameter ?
I still want to leave opal_paffinity_alone flag untouched but  
instead expose mpi_paffinity_alone for the user.


thanks
Lenny.

On Mon, Mar 31, 2008 at 2:55 PM, Jeff Squyres   
wrote:

On Mar 27, 2008, at 8:02 AM, Lenny Verkhovsky wrote:
>
>> - I don't think we can delete the MCA param ompi_paffinity_alone;  
it

>> exists in the v1.2 series and has historical precedent.
> It will not be deleted,
> It will just use the same infrastructure ( slot_list parameter and
> opal_base functions ). It will be transparent for the user.
>
> User have 3 ways to setup it
> 1.mca opal_paffinity_alone 1
>   This will set paffinity as it did before
> 2.mca opal_paffinity_slot_list "slot_list"
>   Used to define slots that will be used for all ranks on all
> nodes.
> 3.mca rmaps_rank_file_path rankfile
>   Assigning ranks to CPUs according to the file


I don't see the MCA parameter "mpi_paffinity_alone" anymore:

-
[4:54] svbu-mpi:~/svn/ompi2 % ompi_info --param all all | grep
paffinity_alone
MCA opal: parameter "opal_paffinity_alone" (current
value: "0")
[4:54] svbu-mpi:~/svn/ompi2 %
-

My point is that I don't think we should delete this parameter; there
is historical precedence for it (and it has been documented on the web
page for a long, long time).  Perhaps it can now simply be a synonym
for opal_paffinity_alone (registered in the MPI layer, not opal).

--
Jeff Squyres
Cisco Systems

___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel




--
Jeff Squyres
Cisco Systems



Re: [OMPI devel] RMAPS rank_file component patch and modifications for review

2008-04-01 Thread Lenny Verkhovsky
Hi,
is there any elegant way to register mpi parameter that will actually be
pointer or alias to hidden opal parameter ?
I still want to leave opal_paffinity_alone flag untouched but instead expose
mpi_paffinity_alone for the user.

thanks
Lenny.

On Mon, Mar 31, 2008 at 2:55 PM, Jeff Squyres  wrote:

> On Mar 27, 2008, at 8:02 AM, Lenny Verkhovsky wrote:
> >
> >> - I don't think we can delete the MCA param ompi_paffinity_alone; it
> >> exists in the v1.2 series and has historical precedent.
> > It will not be deleted,
> > It will just use the same infrastructure ( slot_list parameter and
> > opal_base functions ). It will be transparent for the user.
> >
> > User have 3 ways to setup it
> > 1.mca opal_paffinity_alone 1
> >   This will set paffinity as it did before
> > 2.mca opal_paffinity_slot_list "slot_list"
> >   Used to define slots that will be used for all ranks on all
> > nodes.
> > 3.mca rmaps_rank_file_path rankfile
> >   Assigning ranks to CPUs according to the file
>
>
> I don't see the MCA parameter "mpi_paffinity_alone" anymore:
>
> -
> [4:54] svbu-mpi:~/svn/ompi2 % ompi_info --param all all | grep
> paffinity_alone
> MCA opal: parameter "opal_paffinity_alone" (current
> value: "0")
> [4:54] svbu-mpi:~/svn/ompi2 %
> -
>
> My point is that I don't think we should delete this parameter; there
> is historical precedence for it (and it has been documented on the web
> page for a long, long time).  Perhaps it can now simply be a synonym
> for opal_paffinity_alone (registered in the MPI layer, not opal).
>
> --
>  Jeff Squyres
> Cisco Systems
>
> ___
> devel mailing list
> de...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/devel
>


Re: [OMPI devel] Session directories in $HOME?

2008-04-01 Thread Ralph H Castain
Ummm...actually, there already is an MCA param that does precisely that:

OMPI_MCA_tmpdir_base

Been there for years...sets the tmpdir for both orteds and procs.

The tmpdir argument for mpirun is there if you want to ONLY set the tmpdir
base for mpirun. It provides a protective mechanism for cases where the head
node is configured differently than the compute nodes.

Ralph



On 3/31/08 2:10 PM, "Aurélien Bouteiller"  wrote:

> I more than agree with Galen.
> 
> Aurelien
> Le 31 mars 08 à 16:00, Shipman, Galen M. a écrit :
>> 
>> Slightly OT but along the same lines..
>> 
>> We currently have an argument to mpirun to set the HNP tmpdir (--
>> tmpdir).
>> Why don't we have an mca param to set the tmpdir for all the orted's
>> and such?
>> 
>> - Galen
>> 
>> On Mar 31, 2008, at 3:51 PM, George Bosilca wrote:
>> 
>>> I looked over the code and I don't see any problems with the
>>> changes. The only think I did is replacing the getenv("HOME") by
>>> opal_home_directory ...
>>> 
>>> Here is the logic for selecting the TMP directory:
>>> 
>>>if( NULL == (str = getenv("TMPDIR")) )
>>>if( NULL == (str = getenv("TEMP")) )
>>>if( NULL == (str = getenv("TMP")) )
>>>if( NULL == (str = opal_home_directory()) )
>>>str = ".";
>>> 
>>> Do you have any of those (TMPDIR, TEMP or TMP) in your environment ?
>>> 
>>>  george.
>>> 
>>> On Mar 31, 2008, at 3:13 PM, Josh Hursey wrote:
 So does anyone know why the session directories are in $HOME instead
 of /tmp?
 
 I'm using r18044 and every time I run the session directories are
 created in $HOME. George does this have anything to do with your
 commits from earlier?
 
 -- Josh
 ___
 devel mailing list
 de...@open-mpi.org
 http://www.open-mpi.org/mailman/listinfo.cgi/devel
>>> 
>>> ___
>>> devel mailing list
>>> de...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel
>> 
>> ___
>> devel mailing list
>> de...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/devel
> 
> 
> ___
> devel mailing list
> de...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/devel





Re: [OMPI devel] Session directories in $HOME?

2008-04-01 Thread Shipman, Galen M.


On Apr 1, 2008, at 1:47 PM, Ralph H Castain wrote:

Ummm...actually, there already is an MCA param that does precisely  
that:


OMPI_MCA_tmpdir_base


Perhaps we can modify this so that it reports in ompi_info?

- galen



Been there for years...sets the tmpdir for both orteds and procs.

The tmpdir argument for mpirun is there if you want to ONLY set the  
tmpdir
base for mpirun. It provides a protective mechanism for cases where  
the head

node is configured differently than the compute nodes.

Ralph



On 3/31/08 2:10 PM, "Aurélien Bouteiller"   
wrote:



I more than agree with Galen.

Aurelien
Le 31 mars 08 à 16:00, Shipman, Galen M. a écrit :


Slightly OT but along the same lines..

We currently have an argument to mpirun to set the HNP tmpdir (--
tmpdir).
Why don't we have an mca param to set the tmpdir for all the orted's
and such?

- Galen

On Mar 31, 2008, at 3:51 PM, George Bosilca wrote:


I looked over the code and I don't see any problems with the
changes. The only think I did is replacing the getenv("HOME") by
opal_home_directory ...

Here is the logic for selecting the TMP directory:

   if( NULL == (str = getenv("TMPDIR")) )
   if( NULL == (str = getenv("TEMP")) )
   if( NULL == (str = getenv("TMP")) )
   if( NULL == (str = opal_home_directory()) )
   str = ".";

Do you have any of those (TMPDIR, TEMP or TMP) in your  
environment ?


 george.

On Mar 31, 2008, at 3:13 PM, Josh Hursey wrote:
So does anyone know why the session directories are in $HOME  
instead

of /tmp?

I'm using r18044 and every time I run the session directories are
created in $HOME. George does this have anything to do with your
commits from earlier?

-- Josh
___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel


___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel


___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel



___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel




___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel





Re: [OMPI devel] Session directories in $HOME?

2008-04-01 Thread Ralph H Castain
Sure - I'll rename it "orte_tmpdir_base" so it shows up.


On 4/1/08 12:05 PM, "Shipman, Galen M."  wrote:

> 
> On Apr 1, 2008, at 1:47 PM, Ralph H Castain wrote:
> 
>> Ummm...actually, there already is an MCA param that does precisely
>> that:
>> 
>> OMPI_MCA_tmpdir_base
> 
> Perhaps we can modify this so that it reports in ompi_info?
> 
> - galen
> 
>> 
>> Been there for years...sets the tmpdir for both orteds and procs.
>> 
>> The tmpdir argument for mpirun is there if you want to ONLY set the
>> tmpdir
>> base for mpirun. It provides a protective mechanism for cases where
>> the head
>> node is configured differently than the compute nodes.
>> 
>> Ralph
>> 
>> 
>> 
>> On 3/31/08 2:10 PM, "Aurélien Bouteiller" 
>> wrote:
>> 
>>> I more than agree with Galen.
>>> 
>>> Aurelien
>>> Le 31 mars 08 à 16:00, Shipman, Galen M. a écrit :
 
 Slightly OT but along the same lines..
 
 We currently have an argument to mpirun to set the HNP tmpdir (--
 tmpdir).
 Why don't we have an mca param to set the tmpdir for all the orted's
 and such?
 
 - Galen
 
 On Mar 31, 2008, at 3:51 PM, George Bosilca wrote:
 
> I looked over the code and I don't see any problems with the
> changes. The only think I did is replacing the getenv("HOME") by
> opal_home_directory ...
> 
> Here is the logic for selecting the TMP directory:
> 
>if( NULL == (str = getenv("TMPDIR")) )
>if( NULL == (str = getenv("TEMP")) )
>if( NULL == (str = getenv("TMP")) )
>if( NULL == (str = opal_home_directory()) )
>str = ".";
> 
> Do you have any of those (TMPDIR, TEMP or TMP) in your
> environment ?
> 
>  george.
> 
> On Mar 31, 2008, at 3:13 PM, Josh Hursey wrote:
>> So does anyone know why the session directories are in $HOME
>> instead
>> of /tmp?
>> 
>> I'm using r18044 and every time I run the session directories are
>> created in $HOME. George does this have anything to do with your
>> commits from earlier?
>> 
>> -- Josh
>> ___
>> devel mailing list
>> de...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/devel
> 
> ___
> devel mailing list
> de...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/devel
 
 ___
 devel mailing list
 de...@open-mpi.org
 http://www.open-mpi.org/mailman/listinfo.cgi/devel
>>> 
>>> 
>>> ___
>>> devel mailing list
>>> de...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/devel
>> 
>> 
>> 
>> ___
>> devel mailing list
>> de...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/devel
> 
> 
> ___
> devel mailing list
> de...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/devel





Re: [OMPI devel] Session directories in $HOME?

2008-04-01 Thread Shipman, Galen M.


On Apr 1, 2008, at 2:12 PM, Ralph H Castain wrote:


Sure - I'll rename it "orte_tmpdir_base" so it shows up.



Perfect, do we also need to carry on support for "OMPI_MCA_tmdir_base"?

- Galen



On 4/1/08 12:05 PM, "Shipman, Galen M."  wrote:



On Apr 1, 2008, at 1:47 PM, Ralph H Castain wrote:


Ummm...actually, there already is an MCA param that does precisely
that:

OMPI_MCA_tmpdir_base


Perhaps we can modify this so that it reports in ompi_info?

- galen



Been there for years...sets the tmpdir for both orteds and procs.

The tmpdir argument for mpirun is there if you want to ONLY set the
tmpdir
base for mpirun. It provides a protective mechanism for cases where
the head
node is configured differently than the compute nodes.

Ralph



On 3/31/08 2:10 PM, "Aurélien Bouteiller" 
wrote:


I more than agree with Galen.

Aurelien
Le 31 mars 08 à 16:00, Shipman, Galen M. a écrit :


Slightly OT but along the same lines..

We currently have an argument to mpirun to set the HNP tmpdir (--
tmpdir).
Why don't we have an mca param to set the tmpdir for all the  
orted's

and such?

- Galen

On Mar 31, 2008, at 3:51 PM, George Bosilca wrote:


I looked over the code and I don't see any problems with the
changes. The only think I did is replacing the getenv("HOME") by
opal_home_directory ...

Here is the logic for selecting the TMP directory:

   if( NULL == (str = getenv("TMPDIR")) )
   if( NULL == (str = getenv("TEMP")) )
   if( NULL == (str = getenv("TMP")) )
   if( NULL == (str = opal_home_directory()) )
   str = ".";

Do you have any of those (TMPDIR, TEMP or TMP) in your
environment ?

 george.

On Mar 31, 2008, at 3:13 PM, Josh Hursey wrote:

So does anyone know why the session directories are in $HOME
instead
of /tmp?

I'm using r18044 and every time I run the session directories  
are

created in $HOME. George does this have anything to do with your
commits from earlier?

-- Josh
___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel


___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel


___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel



___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel




___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel



___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel




___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel





[OMPI devel] Trunk launch scaling

2008-04-01 Thread Ralph H Castain
Per this morning's telecon, I have added the latest scaling test results to
the wiki:

https://svn.open-mpi.org/trac/ompi/wiki/ORTEScalabilityTesting

As you will see upon review, the trunk is scaling about an order of
magnitude better than 1.2.x, both in terms of sheer speed and in the
strength of the non-linear components of the scaling law. Those of us
working on scaling issues expect to make additional improvements over the
next few weeks.

Update results will be posted to the wiki as they become available.

Ralph




Re: [OMPI devel] Session directories in $HOME?

2008-04-01 Thread Jeff Squyres
I'll bet that no one was using it; if ompi_info didn't report it,  
there was no way for users to know about it.



On Apr 1, 2008, at 2:19 PM, Shipman, Galen M. wrote:


On Apr 1, 2008, at 2:12 PM, Ralph H Castain wrote:


Sure - I'll rename it "orte_tmpdir_base" so it shows up.



Perfect, do we also need to carry on support for  
"OMPI_MCA_tmdir_base"?


- Galen



On 4/1/08 12:05 PM, "Shipman, Galen M."  wrote:



On Apr 1, 2008, at 1:47 PM, Ralph H Castain wrote:


Ummm...actually, there already is an MCA param that does precisely
that:

OMPI_MCA_tmpdir_base


Perhaps we can modify this so that it reports in ompi_info?

- galen



Been there for years...sets the tmpdir for both orteds and procs.

The tmpdir argument for mpirun is there if you want to ONLY set the
tmpdir
base for mpirun. It provides a protective mechanism for cases where
the head
node is configured differently than the compute nodes.

Ralph



On 3/31/08 2:10 PM, "Aurélien Bouteiller" 
wrote:


I more than agree with Galen.

Aurelien
Le 31 mars 08 à 16:00, Shipman, Galen M. a écrit :


Slightly OT but along the same lines..

We currently have an argument to mpirun to set the HNP tmpdir (--
tmpdir).
Why don't we have an mca param to set the tmpdir for all the
orted's
and such?

- Galen

On Mar 31, 2008, at 3:51 PM, George Bosilca wrote:


I looked over the code and I don't see any problems with the
changes. The only think I did is replacing the getenv("HOME") by
opal_home_directory ...

Here is the logic for selecting the TMP directory:

  if( NULL == (str = getenv("TMPDIR")) )
  if( NULL == (str = getenv("TEMP")) )
  if( NULL == (str = getenv("TMP")) )
  if( NULL == (str = opal_home_directory()) )
  str = ".";

Do you have any of those (TMPDIR, TEMP or TMP) in your
environment ?

george.

On Mar 31, 2008, at 3:13 PM, Josh Hursey wrote:

So does anyone know why the session directories are in $HOME
instead
of /tmp?

I'm using r18044 and every time I run the session directories
are
created in $HOME. George does this have anything to do with  
your

commits from earlier?

-- Josh
___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel


___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel


___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel



___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel




___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel



___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel




___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel



___
devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel



--
Jeff Squyres
Cisco Systems




Re: [OMPI devel] limit tcp fragment size?

2008-04-01 Thread Muhammad Atif
Thanks for the reply. 
I was not able to achieve the required task with the given pointers. 
I ran the application with following command,
mpirun  -np 2 --mca btl_tcp_frag 9 --mca btl_tcp_max_send_size 8192 -host 
node-00,node-01 /home/atif/blah/aa_l

I still see the messages of size 65226 bytes. I am missing something? One side 
question, I cannot find documentation for the mca parameter btl_tcp_frag.

I am currently using openmpi version 1.2.2. 


Best Regards,
Muhammad Atif

- Original Message 
From: George Bosilca 
To: Open MPI Developers 
Sent: Tuesday, April 1, 2008 1:31:46 AM
Subject: Re: [OMPI devel] limit tcp fragment size?

The btl_tcp_min_send_size is not exactly what you expect it to be. It  
drive only the send protocol (as implemented in Open MPI), and not the  
put protocol the TCP BTL is using.

You can achieve what you want with 2 parameters:
1. btl_tcp_frag set to 9. This will force the send protocol over TCP  
all the time
2. btl_tcp_max_send_size set to 16K which will define the size of a  
fragment in the pipelined send protocol.

   george.

On Mar 31, 2008, at 2:46 AM, Muhammad Atif wrote:
> G'day
> Just a quick basic question. in case of tcp btl, how do I limit  
> the frag size?
> I do not want MPI to send a fragment of size greater than lets say  
> 16K in size.
>
> If I am not mistaken, should not the btl_tcp_min_send_size do the  
> trick?  If it is supposed to do it, why do i see packets of lenght  
> 64K ?
>
> Thanks in advance.
>
> Best Regards,
> Muhammad Atif
>
>
>
> Special deal for Yahoo! users & friends - No Cost. Get a month of  
> Blockbuster Total  
> Accessnow___
> devel mailing list
> de...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/devel







  

You rock. That's why Blockbuster's offering you one month of Blockbuster Total 
Access, No Cost.  
http://tc.deals.yahoo.com/tc/blockbuster/text5.com

Re: [OMPI devel] limit tcp fragment size?

2008-04-01 Thread George Bosilca
The parameters I was talking about only split the message at the MPI  
level, pushing the data in 8k fragments into the network. Once the  
data is pushed into the kernel (via the socket), we don't have any  
control over how not when it is physically send to the remote node.  
The only way I see to limit the fragments at the hardware level, is to  
force the send and receive buffer in the kernel to a specific value.  
Tis can be done for each socket, and Open MPI povide 2 MCA parameters  
for this: btl_tcp_sndbuf and btl_tcp_rcvbuf.


  george.

On Apr 1, 2008, at 9:25 PM, Muhammad Atif wrote:

Thanks for the reply.
I was not able to achieve the required task with the given pointers.
I ran the application with following command,
mpirun  -np 2 --mca btl_tcp_frag 9 --mca btl_tcp_max_send_size 8192 - 
host node-00,node-01 /home/atif/blah/aa_l


I still see the messages of size 65226 bytes. I am missing  
something? One side question, I cannot find documentation for the  
mca parameter btl_tcp_frag.


I am currently using openmpi version 1.2.2.

Best Regards,
Muhammad Atif


- Original Message 
From: George Bosilca 
To: Open MPI Developers 
Sent: Tuesday, April 1, 2008 1:31:46 AM
Subject: Re: [OMPI devel] limit tcp fragment size?

The btl_tcp_min_send_size is not exactly what you expect it to be. It
drive only the send protocol (as implemented in Open MPI), and not the
put protocol the TCP BTL is using.

You can achieve what you want with 2 parameters:
1. btl_tcp_frag set to 9. This will force the send protocol over TCP
all the time
2. btl_tcp_max_send_size set to 16K which will define the size of a
fragment in the pipelined send protocol.

  george.

On Mar 31, 2008, at 2:46 AM, Muhammad Atif wrote:
> G'day
> Just a quick basic question. in case of tcp btl, how do I limit
> the frag size?
> I do not want MPI to send a fragment of size greater than lets say
> 16K in size.
>
> If I am not mistaken, should not the btl_tcp_min_send_size do the
> trick?  If it is supposed to do it, why do i see packets of lenght
> 64K ?
>
> Thanks in advance.
>
> Best Regards,
> Muhammad Atif
>
>
>
> Special deal for Yahoo! users & friends - No Cost. Get a month of
> Blockbuster Total
> Accessnow___
> devel mailing list
> de...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/devel



You rock. That's why Blockbuster's offering you one month of  
Blockbuster Total Access, No  
Cost.___

devel mailing list
de...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/devel




smime.p7s
Description: S/MIME cryptographic signature