Hello Paras did you solved the problem?
Thanks Emmanuel 2013/11/25 emmanuel segura <emi2f...@gmail.com> > Hello Paras > > Maybe i found the solution, in function validate_all we got > > if [ -z "$OCF_RESKEY_hypervisor" ] || > [ "$OCF_RESKEY_hypervisor" = "auto" ]; then > export OCF_RESKEY_hypervisor="`virsh version | grep > \"Running hypervisor:\" | awk '{print $3}' | tr A-Z a-z`" > if [ -z "$OCF_RESKEY_hypervisor" ]; then > ocf_log err "Could not determine Hypervisor" > return $OCF_ERR_ARGS > fi > echo Hypervisor: $OCF_RESKEY_hypervisor > fi > > # > # Xen hypervisor only for when use_virsh = 0. > # > if [ "$OCF_RESKEY_use_virsh" = "0" ]; then > if [ "$OCF_RESKEY_hypervisor" != "xen" ]; then > ocf_log err "Cannot use $OCF_RESKEY_hypervisor > hypervisor without using virsh" > return $OCF_ERR_ARGS > fi > > with this following enviroment variables, when i tested by hand the angent > uses xm commands > > env | grep OCF > OCF_RESKEY_hypervisor=xen > OCF_RESKEY_path=/vms_c > OCF_RESKEY_name=guest1 > OCF_RESKEY_use_virsh=0 > > [root@client ~]# /usr/share/cluster/vm.sh status > Management tool: xm > <err> Cannot find 'xm'; is it installed? > [vm.sh] Cannot find 'xm'; is it installed? > > > I don't have xen installed to test it > > > if [ -n "$OCF_RESKEY_xmlfile" ]; then > ocf_log err "Cannot use xmlfile if use_virsh is > set to 0" > return $OCF_ERR_ARGS > fi > > > > 2013/11/25 emmanuel segura <emi2f...@gmail.com> > >> Hello paras >> >> missing the export command in front of variables, the correct way is this >> >> export OCF_RESKEY_name="guest1" ; export OCF_RESKEY_path="/vms_c" ; >> export OCF_RESKEY_use_virsh=0 >> [root@client ~]# env | grep OCF >> OCF_RESKEY_path=/vms_c >> OCF_RESKEY_name=guest1 >> OCF_RESKEY_use_virsh=0 >> >> >> >> 2013/11/25 emmanuel segura <emi2f...@gmail.com> >> >>> Hello Paras >>> >>> I have a centos 6, i don't know if it is different on redhat 5, but i >>> saw in the script vm.sh calls do_start function when start parameter is >>> given >>> >>> do_start() >>> { >>> if [ "$OCF_RESKEY_use_virsh" = "1" ]; then >>> do_virsh_start $* >>> return $? >>> fi >>> >>> do_xm_start $* >>> return $? >>> } >>> >>> i don't know why because the vm.sh uses virsh when you launch the script >>> by hand :( >>> >>> >>> 2013/11/25 Paras pradhan <pradhanpa...@gmail.com> >>> >>>> Looks like use_virsh=0 has no effect. >>>> >>>> -- >>>> [root@cvtst3 ~]# export OCF_RESKEY_name="guest1" ; >>>> OCF_RESKEY_path="/vms_c" ; OCF_RESKEY_use_virsh=0 >>>> [root@cvtst3 ~]# set -x >>>> ++ printf '\033]0;%s@%s:%s\007' root cvtst3 '~' >>>> [root@cvtst3 ~]# /usr/share/cluster/vm.sh start >>>> + /usr/share/cluster/vm.sh start >>>> Hypervisor: xen >>>> Management tool: virsh >>>> Hypervisor URI: xen:/// >>>> Migration URI format: xenmigr://target_host/ >>>> Virtual machine guest1 is error: failed to get domain 'guest1' >>>> error: Domain not found: xenUnifiedDomainLookupByName >>>> >>>> <debug> virsh -c xen:/// start guest1 >>>> error: failed to get domain 'guest1' >>>> error: Domain not found: xenUnifiedDomainLookupByName >>>> >>>> ++ printf '\033]0;%s@%s:%s\007' root cvtst3 '~' >>>> [root@cvtst3 ~]# set +x >>>> + set +x >>>> --- >>>> >>>> >>>> -Paras. >>>> >>>> >>>> On Fri, Nov 22, 2013 at 5:22 PM, emmanuel segura <emi2f...@gmail.com>wrote: >>>> >>>>> Hellos Paras >>>>> >>>>> Stop the vm and retry to start the vm with following commands and if >>>>> you got some error show it >>>>> >>>>> export OCF_RESKEY_name="guest1" ; OCF_RESKEY_path="/vms_c" ; >>>>> OCF_RESKEY_use_virsh=0 >>>>> >>>>> >>>>> set -x >>>>> /usr/share/cluster/vm.sh start >>>>> set +x >>>>> >>>>> >>>>> 2013/11/22 Paras pradhan <pradhanpa...@gmail.com> >>>>> >>>>>> I found the workaround to my issue. What i did is: >>>>>> >>>>>> run the vm using xm and then start using cluvscadm. This works for me >>>>>> for the time being but I am not sure what is causing this. This is what I >>>>>> did >>>>>> >>>>>> xm create /vms_c/guest1 >>>>>> clusvcadm -e vm: guest1 ( This detects that guest1 is up and quickly >>>>>> changes its status to success) >>>>>> >>>>>> Although i used virt-install, it also create a xem format >>>>>> configuration file and since use_virsh=0 it should be able to use this >>>>>> xen >>>>>> format config file. This is my vm configuration: >>>>>> >>>>>> --- >>>>>> name = "guest1" >>>>>> maxmem = 2048 >>>>>> memory = 512 >>>>>> vcpus = 1 >>>>>> #cpus="1-2" >>>>>> bootloader = "/usr/bin/pygrub" >>>>>> on_poweroff = "destroy" >>>>>> on_reboot = "restart" >>>>>> on_crash = "restart" >>>>>> vfb = [ ] >>>>>> disk = [ "tap:aio:/vms_c/guest1.img,xvda,w", >>>>>> "tap:aio:/vms_c/guest1-disk.img,xvdb,w" ] >>>>>> vif = [ "rate=10MB/s,mac=00:16:3e:6b:be:71,bridge=xenbr0" ] >>>>>> >>>>>> --- >>>>>> >>>>>> Thanks for you help Emmanuel ! Really appreciate it. >>>>>> >>>>>> -Paras. >>>>>> >>>>>> >>>>>> On Fri, Nov 22, 2013 at 11:10 AM, emmanuel segura <emi2f...@gmail.com >>>>>> > wrote: >>>>>> >>>>>>> ok, but your vm doesn't start on others nodes, i think, for >>>>>>> configuration problems >>>>>>> ================================================================ >>>>>>> Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <notice> start on vm >>>>>>> "guest1" returned 1 (generic error) >>>>>>> Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <warning> #68: Failed to >>>>>>> start vm:guest1; return value: 1 >>>>>>> Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <notice> Stopping service >>>>>>> vm:guest1 >>>>>>> Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <notice> Service vm:guest1 >>>>>>> is recovering >>>>>>> Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <warning> #71: Relocating >>>>>>> failed service vm:guest1 >>>>>>> Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <notice> Service vm:guest1 >>>>>>> is stopped >>>>>>> ================================================================ >>>>>>> in few words, try in every cluster node >>>>>>> >>>>>>> >>>>>>> export OCF_RESKEY_name="guest1" ; OCF_RESKEY_path="/vms_c" >>>>>>> >>>>>>> set -x >>>>>>> /usr/share/cluster/vm.sh start >>>>>>> /usr/share/cluster/vm.sh stop >>>>>>> >>>>>>> after you check if your vm can start and stop on every cluster node, >>>>>>> >>>>>>> /usr/share/cluster/vm.sh start >>>>>>> /usr/share/cluster/vm.sh migrate name_of_a_cluster_node >>>>>>> >>>>>>> can you show me your vm configuration under /vms_c? >>>>>>> >>>>>>> Thanks >>>>>>> Emmanuel >>>>>>> >>>>>>> >>>>>>> 2013/11/22 Paras pradhan <pradhanpa...@gmail.com> >>>>>>> >>>>>>>> And also to test I made use_virsh=1 , same problem. The vm does not >>>>>>>> start up if the FD domains are offline. >>>>>>>> >>>>>>>> -Paras. >>>>>>>> >>>>>>>> >>>>>>>> On Fri, Nov 22, 2013 at 10:37 AM, Paras pradhan < >>>>>>>> pradhanpa...@gmail.com> wrote: >>>>>>>> >>>>>>>>> Well thats seems to theoretically correct. But right now my cluser >>>>>>>>> has use_virsh=0 and I don't have any issue untill my mebmers on the >>>>>>>>> failover domains are offline. So wondering what is it that clusvcadm >>>>>>>>> -e is >>>>>>>>> looking when I don't use virsh . >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> On Fri, Nov 22, 2013 at 10:05 AM, emmanuel segura < >>>>>>>>> emi2f...@gmail.com> wrote: >>>>>>>>> >>>>>>>>>> If you used virt-install, i think you need use virsh, the cluster >>>>>>>>>> uses xm xen command if you got use_virsh=0 and virsh if you got >>>>>>>>>> use_virsh=1 >>>>>>>>>> in your cluster config >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> 2013/11/22 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>>>> >>>>>>>>>>> I use virt-install to create virtual machines. Is there a way to >>>>>>>>>>> debug why clusvcadm -e vm:guest1 is failing? vm.sh seems to use >>>>>>>>>>> virsh and >>>>>>>>>>> my cluster.conf has use_virsh=0 >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Thanks >>>>>>>>>>> >>>>>>>>>>> Paras. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Nov 21, 2013 5:53 PM, "emmanuel segura" <emi2f...@gmail.com> >>>>>>>>>>> wrote: >>>>>>>>>>> >>>>>>>>>>>> but did you configure your vm with xen tools or using >>>>>>>>>>>> virt-manager? >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> 2013/11/22 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>>>>>> >>>>>>>>>>>>> Well no i don't want to use virsh. But as we are debugging >>>>>>>>>>>>> with virsh now i found a strange issue. >>>>>>>>>>>>> >>>>>>>>>>>>> I exported an xml file and imported to all nodes . Ran >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> --- >>>>>>>>>>>>> name="guest1" path="/vms_c" >>>>>>>>>>>>> >>>>>>>>>>>>> export OCF_RESKEY_name="guest1" ; OCF_RESKEY_path="/vms_c" >>>>>>>>>>>>> >>>>>>>>>>>>> set -x >>>>>>>>>>>>> /usr/share/cluster/vm.sh start >>>>>>>>>>>>> set +x >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> vm starts now. BUT from a cluster service : cluvscam -e >>>>>>>>>>>>> vm:guest1 , same error. >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> So if i populate all my domains' config files to all my cluser >>>>>>>>>>>>> nodes and make use_virsh=1, then the issue is resolved. But this >>>>>>>>>>>>> is a lot >>>>>>>>>>>>> of work for those who have hundreds of vm. >>>>>>>>>>>>> >>>>>>>>>>>>> vm.start uses virsh . Is there a way to tell him not use virsh? >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> Thanks >>>>>>>>>>>>> Paras. >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> On Thu, Nov 21, 2013 at 5:19 PM, emmanuel segura < >>>>>>>>>>>>> emi2f...@gmail.com> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> if you are using virsh for manage your vms, change this in >>>>>>>>>>>>>> your cluster.conf >>>>>>>>>>>>>> >>>>>>>>>>>>>> from >>>>>>>>>>>>>> use_virsh="0" >>>>>>>>>>>>>> to >>>>>>>>>>>>>> use_virsh="1" >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> 2013/11/22 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>>>>>>>> >>>>>>>>>>>>>>> I think i found the problem. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> virsh list --all does not show my vm . This is because it >>>>>>>>>>>>>>> was created on another node. and another node has it. Now I >>>>>>>>>>>>>>> want to start >>>>>>>>>>>>>>> the service on a different node in which it was not created or >>>>>>>>>>>>>>> where virsh >>>>>>>>>>>>>>> list --all does not have an entry. Is it possible to create >>>>>>>>>>>>>>> this entry >>>>>>>>>>>>>>> using a xen config file?Looks like this is now a Xen issue >>>>>>>>>>>>>>> rather than a >>>>>>>>>>>>>>> linux-cluster issue . :) >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Paras. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On Thu, Nov 21, 2013 at 4:58 PM, emmanuel segura < >>>>>>>>>>>>>>> emi2f...@gmail.com> wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> 1:did you verify your xen livemigration configuration? >>>>>>>>>>>>>>>> 2: where you vm disk reside? >>>>>>>>>>>>>>>> 3: can you see your vm defined on every cluster node with >>>>>>>>>>>>>>>> xm list? >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> 2013/11/21 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> This is what I get >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Hypervisor: xen >>>>>>>>>>>>>>>>> Management tool: virsh >>>>>>>>>>>>>>>>> Hypervisor URI: xen:/// >>>>>>>>>>>>>>>>> Migration URI format: xenmigr://target_host/ >>>>>>>>>>>>>>>>> Virtual machine guest1 is error: failed to get domain >>>>>>>>>>>>>>>>> 'guest1' >>>>>>>>>>>>>>>>> error: Domain not found: xenUnifiedDomainLookupByName >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> <debug> virsh -c xen:/// start guest1 >>>>>>>>>>>>>>>>> error: failed to get domain 'guest1' >>>>>>>>>>>>>>>>> error: Domain not found: xenUnifiedDomainLookupByName >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> ++ printf '\033]0;%s@%s:%s\007' root vtst3 '~' >>>>>>>>>>>>>>>>> [root@cvtst3 ~]# set +x >>>>>>>>>>>>>>>>> + set +x >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> I am wondering why it failed to get domain . >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> -Paras. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> On Thu, Nov 21, 2013 at 4:43 PM, emmanuel segura < >>>>>>>>>>>>>>>>> emi2f...@gmail.com> wrote: >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> yes >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> 2013/11/21 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Well it is guest1. Isn't it?. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> <vm autostart="1" domain="myfd1" exclusive="0" >>>>>>>>>>>>>>>>>>> max_restarts="0" name="guest1" path="/vms_c" >>>>>>>>>>>>>>>>>>> recovery="restart" >>>>>>>>>>>>>>>>>>> restart_expire_time="0" use_virsh="0"/> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> It is a vm service if it matters. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> -Paras. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> On Thu, Nov 21, 2013 at 4:22 PM, emmanuel segura < >>>>>>>>>>>>>>>>>>> emi2f...@gmail.com> wrote: >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> use the servicename you defined in your cluster.conf >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> 2013/11/21 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Says: >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Running in test mode. >>>>>>>>>>>>>>>>>>>>> No resource guest1 of type service found >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> -Paras. >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> On Thu, Nov 21, 2013 at 4:07 PM, emmanuel segura < >>>>>>>>>>>>>>>>>>>>> emi2f...@gmail.com> wrote: >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> rg_test test /etc/cluster/cluster.conf start service >>>>>>>>>>>>>>>>>>>>>> guest1 >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> 2013/11/21 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> My failover domain looks like this: >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> <failoverdomain name="myfd1" nofailback="1" >>>>>>>>>>>>>>>>>>>>>>> ordered="1" restricted="0"> >>>>>>>>>>>>>>>>>>>>>>> <failoverdomainnode >>>>>>>>>>>>>>>>>>>>>>> name="vtst1" priority="1"/> >>>>>>>>>>>>>>>>>>>>>>> <failoverdomainnode >>>>>>>>>>>>>>>>>>>>>>> name="vtst3" priority="2"/> >>>>>>>>>>>>>>>>>>>>>>> <failoverdomainnode >>>>>>>>>>>>>>>>>>>>>>> name="vtst2" priority="3"/> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> </failoverdomain> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> I have vm service that uses this failover domain. If >>>>>>>>>>>>>>>>>>>>>>> my node vtst1 is offline, the service doesnot start on >>>>>>>>>>>>>>>>>>>>>>> vtst3 which is 2nd >>>>>>>>>>>>>>>>>>>>>>> in the priority. >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> I tried to start it with: clusvcadm -e vm:guest1 >>>>>>>>>>>>>>>>>>>>>>> and even with -F and -m option. >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> All i see is this error: >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <notice> >>>>>>>>>>>>>>>>>>>>>>> start on vm "guest1" returned 1 (generic error) >>>>>>>>>>>>>>>>>>>>>>> Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <warning> >>>>>>>>>>>>>>>>>>>>>>> #68: Failed to start vm:guest1; return value: 1 >>>>>>>>>>>>>>>>>>>>>>> Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <notice> >>>>>>>>>>>>>>>>>>>>>>> Stopping service vm:guest1 >>>>>>>>>>>>>>>>>>>>>>> Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <notice> >>>>>>>>>>>>>>>>>>>>>>> Service vm:guest1 is recovering >>>>>>>>>>>>>>>>>>>>>>> Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <warning> >>>>>>>>>>>>>>>>>>>>>>> #71: Relocating failed service vm:guest1 >>>>>>>>>>>>>>>>>>>>>>> Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <notice> >>>>>>>>>>>>>>>>>>>>>>> Service vm:guest1 is stopped >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> How do I debug? >>>>>>>>>>>>>>>>>>>>>>> Thanks! >>>>>>>>>>>>>>>>>>>>>>> Paras. >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> -- >>>>>>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>>>>>> >>>>>>>>>>>>>> -- >>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> Linux-cluster mailing list >>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> Linux-cluster mailing list >>>>>>>> Linux-cluster@redhat.com >>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>> >>>>>>> -- >>>>>>> Linux-cluster mailing list >>>>>>> Linux-cluster@redhat.com >>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> Linux-cluster mailing list >>>>>> Linux-cluster@redhat.com >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>> >>>>> -- >>>>> Linux-cluster mailing list >>>>> Linux-cluster@redhat.com >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>> >>>> >>>> -- >>>> Linux-cluster mailing list >>>> Linux-cluster@redhat.com >>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>> >>> >>> >>> -- >>> esta es mi vida e me la vivo hasta que dios quiera >>> >> >> >> >> -- >> esta es mi vida e me la vivo hasta que dios quiera >> > > > > -- > esta es mi vida e me la vivo hasta que dios quiera > -- esta es mi vida e me la vivo hasta que dios quiera
-- Linux-cluster mailing list Linux-cluster@redhat.com https://www.redhat.com/mailman/listinfo/linux-cluster