On Mon, Feb 2, 2009 at 12:15 PM, Reuti <re...@staff.uni-marburg.de> wrote:
> Am 02.02.2009 um 05:44 schrieb Sangamesh B:
>
>> On Sun, Feb 1, 2009 at 10:37 PM, Reuti <re...@staff.uni-marburg.de> wrote:
>>>
>>> Am 01.02.2009 um 16:00 schrieb Sangamesh B:
>>>
>>>> On Sat, Jan 31, 2009 at 6:27 PM, Reuti <re...@staff.uni-marburg.de>
>>>> wrote:
>>>>>
>>>>> Am 31.01.2009 um 08:49 schrieb Sangamesh B:
>>>>>
>>>>>> On Fri, Jan 30, 2009 at 10:20 PM, Reuti <re...@staff.uni-marburg.de>
>>>>>> wrote:
>>>>>>>
>>>>>>> Am 30.01.2009 um 15:02 schrieb Sangamesh B:
>>>>>>>
>>>>>>>> Dear Open MPI,
>>>>>>>>
>>>>>>>> Do you have a solution for the following problem of Open MPI (1.3)
>>>>>>>> when run through Grid Engine.
>>>>>>>>
>>>>>>>> I changed global execd params with H_MEMORYLOCKED=infinity and
>>>>>>>> restarted the sgeexecd in all nodes.
>>>>>>>>
>>>>>>>> But still the problem persists:
>>>>>>>>
>>>>>>>>  $cat err.77.CPMD-OMPI
>>>>>>>> ssh_exchange_identification: Connection closed by remote host
>>>>>>>
>>>>>>> I think this might already be the reason why it's not working. A
>>>>>>> mpihello
>>>>>>> program is running fine through SGE?
>>>>>>>
>>>>>> No.
>>>>>>
>>>>>> Any Open MPI parallel job thru SGE runs only if its running on a
>>>>>> single node (i.e. 8processes on 8 cores of a single node). If number
>>>>>> of processes is more than 8, then SGE will schedule it on 2 nodes -
>>>>>> the job will fail with the above error.
>>>>>>
>>>>>> Now I did a loose integration of Open MPI 1.3 with SGE. The job runs,
>>>>>> but all 16 processes run on a single node.
>>>>>
>>>>> What are the entries in `qconf -sconf`for:
>>>>>
>>>>> rsh_command
>>>>> rsh_daemon
>>>>>
>>>> $ qconf -sconf
>>>> global:
>>>> execd_spool_dir              /opt/gridengine/default/spool
>>>> ...
>>>> .....
>>>> qrsh_command                 /usr/bin/ssh
>>>> rsh_command                  /usr/bin/ssh
>>>> rlogin_command               /usr/bin/ssh
>>>> rsh_daemon                   /usr/sbin/sshd
>>>> qrsh_daemon                  /usr/sbin/sshd
>>>> reprioritize                 0
>>>
>>> Do you must use ssh? Often in a private cluster the rsh based one is ok,
>>> or
>>> with SGE 6.2 the built-in mechanism of SGE. Otherwise please follow this:
>>>
>>> http://gridengine.sunsource.net/howto/qrsh_qlogin_ssh.html
>>>
>>>
>>>> I think its better to check once with Open MPI 1.2.8
>>>>
>>>>> What is your mpirun command in the jobscript - you are getting there
>>>>> the
>>>>> mpirun from Open MPI? According to the output below, it's not a loose
>>>>> integration, but you prepare alraedy a machinefile, which is
>>>>> superfluous
>>>>> for
>>>>> Open MPI.
>>>>>
>>>> No. I've not prepared the machinefile for Open MPI.
>>>> For Tight integartion job:
>>>>
>>>> /opt/mpi/openmpi/1.3/intel/bin/mpirun -np $NSLOTS
>>>> $CPMDBIN/cpmd311-ompi-mkl.x  wf1.in $PP_LIBRARY >
>>>> wf1.out_OMPI$NSLOTS.$JOB_ID
>>>>
>>>> For loose integration job:
>>>>
>>>> /opt/mpi/openmpi/1.3/intel/bin/mpirun -np $NSLOTS -hostfile
>>>> $TMPDIR/machines  $CPMDBIN/cpmd311-ompi-mkl.x  wf1.in $PP_LIBRARY >
>>>> wf1.out_OMPI_$JOB_ID.$NSLOTS
>>>
>>> a) you compiled Open MPI with "--with-sge"?
>>>
>> Yes. But ompi_info shows only one component of sge
>>
>> $ /opt/mpi/openmpi/1.3/intel/bin/ompi_info | grep gridengine
>>                 MCA ras: gridengine (MCA v2.0, API v2.0, Component v1.3)
>>
>>> b) when the $SGE_ROOT variable is set, Open MPI will use a Tight
>>> Integration
>>> automatically.
>>>
>> In SGE job submit script, I set SGE_ROOT= <nothing>
>
> This will set the variable to an empty string. You need to use:
>
> unset SGE_ROOT
>
Right.
I used 'unset SGE_ROOT' in the job submission script. Its working now.
Hello world jobs are working now. (single & multiple nodes)

Thank you for the help.

What can be the problem with tight integration?

Regards,
Sangamesh
> Despite the mentioned error message on the list, I can run Open MPI 1.3 with
> tight integration into SGE.
>
> -- Reuti
>
>
>> And run a loose integration job. It failed to run with following error:
>> $ cat err.87.Hello-OMPI
>> [node-0-18.local:08252] [[INVALID],INVALID] ORTE_ERROR_LOG: Not found
>> in file ess_hnp_module.c at line 126
>> --------------------------------------------------------------------------
>> It looks like orte_init failed for some reason; your parallel process is
>> likely to abort.  There are many reasons that a parallel process can
>> fail during orte_init; some of which are due to configuration or
>> environment problems.  This failure appears to be an internal failure;
>> here's some additional information (which may only be relevant to an
>> Open MPI developer):
>>
>>  orte_plm_base_select failed
>>  --> Returned value Not found (-13) instead of ORTE_SUCCESS
>> --------------------------------------------------------------------------
>> [node-0-18.local:08252] [[INVALID],INVALID] ORTE_ERROR_LOG: Not found
>> in file runtime/orte_init.c at line 132
>> --------------------------------------------------------------------------
>> It looks like orte_init failed for some reason; your parallel process is
>> likely to abort.  There are many reasons that a parallel process can
>> fail during orte_init; some of which are due to configuration or
>> environment problems.  This failure appears to be an internal failure;
>> here's some additional information (which may only be relevant to an
>> Open MPI developer):
>>
>>  orte_ess_set_name failed
>>  --> Returned value Not found (-13) instead of ORTE_SUCCESS
>> --------------------------------------------------------------------------
>> [node-0-18.local:08252] [[INVALID],INVALID] ORTE_ERROR_LOG: Not found
>> in file orterun.c at line 454
>>
>> $ cat out.87.Hello-OMPI
>> /opt/gridengine/default/spool/node-0-18/active_jobs/87.1/pe_hostfile
>> ibc18
>> ibc18
>> ibc18
>> ibc18
>> ibc18
>> ibc18
>> ibc18
>> ibc18
>> ibc17
>> ibc17
>> ibc17
>> ibc17
>> ibc17
>> ibc17
>> ibc17
>> ibc17
>>
>>
>>> c) The machine file you presented looks like being for MPICH(1), the
>>> syntax
>>> for Open MPI in the machine is different:
>>>
>>> ibc17 slots=8
>>> ibc12 slots=8
>>>
>> I tested a helloworld program with Open MPI with machinefile of style
>> MPICH(1).
>> It works.
>>
>> So in a loose integration job,
>> Open MPI may not be able to find $TMPDIR/machines file
>> Or it might be running in a Tight integration style.
>>>
>>> So you would have to adjust the format of the generated file and reset
>>> SGE_ROOT inside your jobscript, to force Open MPI to do a loose
>>> integration
>>> only.
>>>
>>> -- Reuti
>>>
>>>
>>>> I think I should check with Open MPI 1.2.8. That may work..
>>>>
>>>> Thanks,
>>>> Sangamesh
>>>>>>
>>>>>> $ cat out.83.Hello-OMPI
>>>>>> /opt/gridengine/default/spool/node-0-17/active_jobs/83.1/pe_hostfile
>>>>>> ibc17
>>>>>> ibc17
>>>>>> ibc17
>>>>>> ibc17
>>>>>> ibc17
>>>>>> ibc17
>>>>>> ibc17
>>>>>> ibc17
>>>>>> ibc12
>>>>>> ibc12
>>>>>> ibc12
>>>>>> ibc12
>>>>>> ibc12
>>>>>> ibc12
>>>>>> ibc12
>>>>>> ibc12
>>>>>> Greetings: 1 of 16 from the node node-0-17.local
>>>>>> Greetings: 10 of 16 from the node node-0-17.local
>>>>>> Greetings: 15 of 16 from the node node-0-17.local
>>>>>> Greetings: 9 of 16 from the node node-0-17.local
>>>>>> Greetings: 14 of 16 from the node node-0-17.local
>>>>>> Greetings: 8 of 16 from the node node-0-17.local
>>>>>> Greetings: 11 of 16 from the node node-0-17.local
>>>>>> Greetings: 12 of 16 from the node node-0-17.local
>>>>>> Greetings: 6 of 16 from the node node-0-17.local
>>>>>> Greetings: 0 of 16 from the node node-0-17.local
>>>>>> Greetings: 5 of 16 from the node node-0-17.local
>>>>>> Greetings: 3 of 16 from the node node-0-17.local
>>>>>> Greetings: 13 of 16 from the node node-0-17.local
>>>>>> Greetings: 4 of 16 from the node node-0-17.local
>>>>>> Greetings: 7 of 16 from the node node-0-17.local
>>>>>> Greetings: 2 of 16 from the node node-0-17.local
>>>>>>
>>>>>> But qhost -u <user name> shows that it is scheduled/running on two
>>>>>> nodes.
>>>>>>
>>>>>> Any body successful in running Open MPI 1.3 tightly integrated with
>>>>>> SGE?
>>>>>
>>>>> For a Tight Integration there's a FAQ:
>>>>>
>>>>> http://www.open-mpi.org/faq/?category=running#run-n1ge-or-sge
>>>>>
>>>>> -- Reuti
>>>>>
>>>>>>
>>>>>> Thanks,
>>>>>> Sangamesh
>>>>>>
>>>>>>> -- Reuti
>>>>>>>
>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --------------------------------------------------------------------------
>>>>>>>> A daemon (pid 31947) died unexpectedly with status 129 while
>>>>>>>> attempting
>>>>>>>> to launch so we are aborting.
>>>>>>>>
>>>>>>>> There may be more information reported by the environment (see
>>>>>>>> above).
>>>>>>>>
>>>>>>>> This may be because the daemon was unable to find all the needed
>>>>>>>> shared
>>>>>>>> libraries on the remote node. You may set your LD_LIBRARY_PATH to
>>>>>>>> have
>>>>>>>> the
>>>>>>>> location of the shared libraries on the remote nodes and this will
>>>>>>>> automatically be forwarded to the remote nodes.
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --------------------------------------------------------------------------
>>>>>>>> mpirun noticed that the job aborted, but has no info as to the
>>>>>>>> process
>>>>>>>> that caused that situation.
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --------------------------------------------------------------------------
>>>>>>>> ssh_exchange_identification: Connection closed by remote host
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --------------------------------------------------------------------------
>>>>>>>> mpirun was unable to cleanly terminate the daemons on the nodes
>>>>>>>> shown
>>>>>>>> below. Additional manual cleanup may be required - please refer to
>>>>>>>> the "orte-clean" tool for assistance.
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>    node-0-19.local - daemon did not report back when launched
>>>>>>>>    node-0-20.local - daemon did not report back when launched
>>>>>>>>    node-0-21.local - daemon did not report back when launched
>>>>>>>>    node-0-22.local - daemon did not report back when launched
>>>>>>>>
>>>>>>>> The hostnames for infiniband interfaces are ibc0, ibc1, ibc2 ..
>>>>>>>> ibc23.
>>>>>>>> May be Open MPI is not able to identify hosts as it is using
>>>>>>>> node-0-..
>>>>>>>> . Is this causing open mpi to fail?
>>>>>>>>
>>>>>>>> Thanks,
>>>>>>>> Sangamesh
>>>>>>>>
>>>>>>>>
>>>>>>>> On Mon, Jan 26, 2009 at 5:09 PM, mihlon <vacl...@fel.cvut.cz> wrote:
>>>>>>>>>
>>>>>>>>> Hi,
>>>>>>>>>
>>>>>>>>>> Hello SGE users,
>>>>>>>>>>
>>>>>>>>>> The cluster is installed with Rocks-4.3, SGE 6.0 & Open MPI 1.3.
>>>>>>>>>> Open MPI is configured with "--with-sge".
>>>>>>>>>> ompi_info shows only one component:
>>>>>>>>>> # /opt/mpi/openmpi/1.3/intel/bin/ompi_info | grep gridengine
>>>>>>>>>> MCA ras: gridengine (MCA v2.0, API v2.0, Component v1.3)
>>>>>>>>>>
>>>>>>>>>> Is this acceptable?
>>>>>>>>>
>>>>>>>>> maybe yes
>>>>>>>>>
>>>>>>>>> see: http://www.open-mpi.org/faq/?category=building#build-rte-sge
>>>>>>>>>
>>>>>>>>> shell$ ompi_info | grep gridengine
>>>>>>>>> MCA ras: gridengine (MCA v1.0, API v1.3, Component v1.3)
>>>>>>>>> MCA pls: gridengine (MCA v1.0, API v1.3, Component v1.3)
>>>>>>>>>
>>>>>>>>> (Specific frameworks and version numbers may vary, depending on
>>>>>>>>> your
>>>>>>>>> version of Open MPI.)
>>>>>>>>>
>>>>>>>>>> The Open MPI parallel jobs run successfully through command line,
>>>>>>>>>> but
>>>>>>>>>> fail when run thru SGE(with -pe orte <slots>).
>>>>>>>>>>
>>>>>>>>>> The error is:
>>>>>>>>>>
>>>>>>>>>> $ cat err.26.Helloworld-PRL
>>>>>>>>>> ssh_exchange_identification: Connection closed by remote host
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>> A daemon (pid 8462) died unexpectedly with status 129 while
>>>>>>>>>> attempting
>>>>>>>>>> to launch so we are aborting.
>>>>>>>>>>
>>>>>>>>>> There may be more information reported by the environment (see
>>>>>>>>>> above).
>>>>>>>>>>
>>>>>>>>>> This may be because the daemon was unable to find all the needed
>>>>>>>>>> shared
>>>>>>>>>> libraries on the remote node. You may set your LD_LIBRARY_PATH to
>>>>>>>>>> have
>>>>>>>>>> the
>>>>>>>>>> location of the shared libraries on the remote nodes and this will
>>>>>>>>>> automatically be forwarded to the remote nodes.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>> mpirun noticed that the job aborted, but has no info as to the
>>>>>>>>>> process
>>>>>>>>>> that caused that situation.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>> mpirun: clean termination accomplished
>>>>>>>>>>
>>>>>>>>>> But the same job runs well, if it runs on a single node but with
>>>>>>>>>> an
>>>>>>>>>> error:
>>>>>>>>>>
>>>>>>>>>> $ cat err.23.Helloworld-PRL
>>>>>>>>>> libibverbs: Warning: RLIMIT_MEMLOCK is 32768 bytes.
>>>>>>>>>> This will severely limit memory registrations.
>>>>>>>>>> libibverbs: Warning: RLIMIT_MEMLOCK is 32768 bytes.
>>>>>>>>>> This will severely limit memory registrations.
>>>>>>>>>> libibverbs: Warning: RLIMIT_MEMLOCK is 32768 bytes.
>>>>>>>>>> This will severely limit memory registrations.
>>>>>>>>>> libibverbs: Warning: RLIMIT_MEMLOCK is 32768 bytes.
>>>>>>>>>> This will severely limit memory registrations.
>>>>>>>>>> libibverbs: Warning: RLIMIT_MEMLOCK is 32768 bytes.
>>>>>>>>>> This will severely limit memory registrations.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>> WARNING: There was an error initializing an OpenFabrics device.
>>>>>>>>>>
>>>>>>>>>> Local host: node-0-4.local
>>>>>>>>>> Local device: mthca0
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --------------------------------------------------------------------------
>>>>>>>>>> libibverbs: Warning: RLIMIT_MEMLOCK is 32768 bytes.
>>>>>>>>>> This will severely limit memory registrations.
>>>>>>>>>> libibverbs: Warning: RLIMIT_MEMLOCK is 32768 bytes.
>>>>>>>>>> This will severely limit memory registrations.
>>>>>>>>>> libibverbs: Warning: RLIMIT_MEMLOCK is 32768 bytes.
>>>>>>>>>> This will severely limit memory registrations.
>>>>>>>>>> [node-0-4.local:07869] 7 more processes have sent help message
>>>>>>>>>> help-mpi-btl-openib.txt / error in device init
>>>>>>>>>> [node-0-4.local:07869] Set MCA parameter
>>>>>>>>>> "orte_base_help_aggregate"
>>>>>>>>>> to
>>>>>>>>>> 0 to see all help / error messages
>>>>>>>>>>
>>>>>>>>>> The following link explains the same problem:
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> http://gridengine.sunsource.net/ds/viewMessage.do?dsForumId=38&dsMessageId=72398
>>>>>>>>>>
>>>>>>>>>> With this reference, I put 'ulimit -l unlimited' into
>>>>>>>>>> /etc/init.d/sgeexecd in all nodes. Restarted the services.
>>>>>>>>>
>>>>>>>>> Do not set 'ulimit -l unlimited' in /etc/init.d/sgeexecd
>>>>>>>>> but set it in the SGE:
>>>>>>>>>
>>>>>>>>> Run   qconf -mconf   and set    execd_params
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> frontend$> qconf -sconf
>>>>>>>>> ...
>>>>>>>>> execd_params                 H_MEMORYLOCKED=infinity
>>>>>>>>> ...
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Then restart all your sgeexecd hosts.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Milan
>>>>>>>>>
>>>>>>>>>> But still the problem persists.
>>>>>>>>>>
>>>>>>>>>> What could be the way out for this?
>>>>>>>>>>
>>>>>>>>>> Thanks,
>>>>>>>>>> Sangamesh
>>>>>>>>>>
>>>>>>>>>> ------------------------------------------------------
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> http://gridengine.sunsource.net/ds/viewMessage.do?dsForumId=38&dsMessageId=99133
>>>>>>>>>>
>>>>>>>>>> To unsubscribe from this discussion, e-mail:
>>>>>>>>>> [users-unsubscr...@gridengine.sunsource.net].
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>> ------------------------------------------------------
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> http://gridengine.sunsource.net/ds/viewMessage.do?dsForumId=38&dsMessageId=99461
>>>>>>>>>
>>>>>>>>> To unsubscribe from this discussion, e-mail:
>>>>>>>>> [users-unsubscr...@gridengine.sunsource.net].
>>>>>>>>>
>>>>>>>> _______________________________________________
>>>>>>>> users mailing list
>>>>>>>> us...@open-mpi.org
>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> users mailing list
>>>>>>> us...@open-mpi.org
>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>
>>>>>> _______________________________________________
>>>>>> users mailing list
>>>>>> us...@open-mpi.org
>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>
>>>> _______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org
>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>

Reply via email to