Hello paras missing the export command in front of variables, the correct way is this
export OCF_RESKEY_name="guest1" ; export OCF_RESKEY_path="/vms_c" ; export OCF_RESKEY_use_virsh=0 [root@client ~]# env | grep OCF OCF_RESKEY_path=/vms_c OCF_RESKEY_name=guest1 OCF_RESKEY_use_virsh=0 2013/11/25 emmanuel segura <emi2f...@gmail.com> > Hello Paras > > I have a centos 6, i don't know if it is different on redhat 5, but i saw > in the script vm.sh calls do_start function when start parameter is given > > do_start() > { > if [ "$OCF_RESKEY_use_virsh" = "1" ]; then > do_virsh_start $* > return $? > fi > > do_xm_start $* > return $? > } > > i don't know why because the vm.sh uses virsh when you launch the script > by hand :( > > > 2013/11/25 Paras pradhan <pradhanpa...@gmail.com> > >> Looks like use_virsh=0 has no effect. >> >> -- >> [root@cvtst3 ~]# export OCF_RESKEY_name="guest1" ; >> OCF_RESKEY_path="/vms_c" ; OCF_RESKEY_use_virsh=0 >> [root@cvtst3 ~]# set -x >> ++ printf '\033]0;%s@%s:%s\007' root cvtst3 '~' >> [root@cvtst3 ~]# /usr/share/cluster/vm.sh start >> + /usr/share/cluster/vm.sh start >> Hypervisor: xen >> Management tool: virsh >> Hypervisor URI: xen:/// >> Migration URI format: xenmigr://target_host/ >> Virtual machine guest1 is error: failed to get domain 'guest1' >> error: Domain not found: xenUnifiedDomainLookupByName >> >> <debug> virsh -c xen:/// start guest1 >> error: failed to get domain 'guest1' >> error: Domain not found: xenUnifiedDomainLookupByName >> >> ++ printf '\033]0;%s@%s:%s\007' root cvtst3 '~' >> [root@cvtst3 ~]# set +x >> + set +x >> --- >> >> >> -Paras. >> >> >> On Fri, Nov 22, 2013 at 5:22 PM, emmanuel segura <emi2f...@gmail.com>wrote: >> >>> Hellos Paras >>> >>> Stop the vm and retry to start the vm with following commands and if you >>> got some error show it >>> >>> export OCF_RESKEY_name="guest1" ; OCF_RESKEY_path="/vms_c" ; >>> OCF_RESKEY_use_virsh=0 >>> >>> >>> set -x >>> /usr/share/cluster/vm.sh start >>> set +x >>> >>> >>> 2013/11/22 Paras pradhan <pradhanpa...@gmail.com> >>> >>>> I found the workaround to my issue. What i did is: >>>> >>>> run the vm using xm and then start using cluvscadm. This works for me >>>> for the time being but I am not sure what is causing this. This is what I >>>> did >>>> >>>> xm create /vms_c/guest1 >>>> clusvcadm -e vm: guest1 ( This detects that guest1 is up and quickly >>>> changes its status to success) >>>> >>>> Although i used virt-install, it also create a xem format configuration >>>> file and since use_virsh=0 it should be able to use this xen format config >>>> file. This is my vm configuration: >>>> >>>> --- >>>> name = "guest1" >>>> maxmem = 2048 >>>> memory = 512 >>>> vcpus = 1 >>>> #cpus="1-2" >>>> bootloader = "/usr/bin/pygrub" >>>> on_poweroff = "destroy" >>>> on_reboot = "restart" >>>> on_crash = "restart" >>>> vfb = [ ] >>>> disk = [ "tap:aio:/vms_c/guest1.img,xvda,w", >>>> "tap:aio:/vms_c/guest1-disk.img,xvdb,w" ] >>>> vif = [ "rate=10MB/s,mac=00:16:3e:6b:be:71,bridge=xenbr0" ] >>>> >>>> --- >>>> >>>> Thanks for you help Emmanuel ! Really appreciate it. >>>> >>>> -Paras. >>>> >>>> >>>> On Fri, Nov 22, 2013 at 11:10 AM, emmanuel segura >>>> <emi2f...@gmail.com>wrote: >>>> >>>>> ok, but your vm doesn't start on others nodes, i think, for >>>>> configuration problems >>>>> ================================================================ >>>>> Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <notice> start on vm "guest1" >>>>> returned 1 (generic error) >>>>> Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <warning> #68: Failed to start >>>>> vm:guest1; return value: 1 >>>>> Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <notice> Stopping service >>>>> vm:guest1 >>>>> Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <notice> Service vm:guest1 is >>>>> recovering >>>>> Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <warning> #71: Relocating >>>>> failed service vm:guest1 >>>>> Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <notice> Service vm:guest1 is >>>>> stopped >>>>> ================================================================ >>>>> in few words, try in every cluster node >>>>> >>>>> >>>>> export OCF_RESKEY_name="guest1" ; OCF_RESKEY_path="/vms_c" >>>>> >>>>> set -x >>>>> /usr/share/cluster/vm.sh start >>>>> /usr/share/cluster/vm.sh stop >>>>> >>>>> after you check if your vm can start and stop on every cluster node, >>>>> >>>>> /usr/share/cluster/vm.sh start >>>>> /usr/share/cluster/vm.sh migrate name_of_a_cluster_node >>>>> >>>>> can you show me your vm configuration under /vms_c? >>>>> >>>>> Thanks >>>>> Emmanuel >>>>> >>>>> >>>>> 2013/11/22 Paras pradhan <pradhanpa...@gmail.com> >>>>> >>>>>> And also to test I made use_virsh=1 , same problem. The vm does not >>>>>> start up if the FD domains are offline. >>>>>> >>>>>> -Paras. >>>>>> >>>>>> >>>>>> On Fri, Nov 22, 2013 at 10:37 AM, Paras pradhan < >>>>>> pradhanpa...@gmail.com> wrote: >>>>>> >>>>>>> Well thats seems to theoretically correct. But right now my cluser >>>>>>> has use_virsh=0 and I don't have any issue untill my mebmers on the >>>>>>> failover domains are offline. So wondering what is it that clusvcadm -e >>>>>>> is >>>>>>> looking when I don't use virsh . >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> On Fri, Nov 22, 2013 at 10:05 AM, emmanuel segura < >>>>>>> emi2f...@gmail.com> wrote: >>>>>>> >>>>>>>> If you used virt-install, i think you need use virsh, the cluster >>>>>>>> uses xm xen command if you got use_virsh=0 and virsh if you got >>>>>>>> use_virsh=1 >>>>>>>> in your cluster config >>>>>>>> >>>>>>>> >>>>>>>> 2013/11/22 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>> >>>>>>>>> I use virt-install to create virtual machines. Is there a way to >>>>>>>>> debug why clusvcadm -e vm:guest1 is failing? vm.sh seems to use >>>>>>>>> virsh and >>>>>>>>> my cluster.conf has use_virsh=0 >>>>>>>>> >>>>>>>>> >>>>>>>>> Thanks >>>>>>>>> >>>>>>>>> Paras. >>>>>>>>> >>>>>>>>> >>>>>>>>> On Nov 21, 2013 5:53 PM, "emmanuel segura" <emi2f...@gmail.com> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>>> but did you configure your vm with xen tools or using >>>>>>>>>> virt-manager? >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> 2013/11/22 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>>>> >>>>>>>>>>> Well no i don't want to use virsh. But as we are debugging >>>>>>>>>>> with virsh now i found a strange issue. >>>>>>>>>>> >>>>>>>>>>> I exported an xml file and imported to all nodes . Ran >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> --- >>>>>>>>>>> name="guest1" path="/vms_c" >>>>>>>>>>> >>>>>>>>>>> export OCF_RESKEY_name="guest1" ; OCF_RESKEY_path="/vms_c" >>>>>>>>>>> >>>>>>>>>>> set -x >>>>>>>>>>> /usr/share/cluster/vm.sh start >>>>>>>>>>> set +x >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> vm starts now. BUT from a cluster service : cluvscam -e >>>>>>>>>>> vm:guest1 , same error. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> So if i populate all my domains' config files to all my cluser >>>>>>>>>>> nodes and make use_virsh=1, then the issue is resolved. But this is >>>>>>>>>>> a lot >>>>>>>>>>> of work for those who have hundreds of vm. >>>>>>>>>>> >>>>>>>>>>> vm.start uses virsh . Is there a way to tell him not use virsh? >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Thanks >>>>>>>>>>> Paras. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Thu, Nov 21, 2013 at 5:19 PM, emmanuel segura < >>>>>>>>>>> emi2f...@gmail.com> wrote: >>>>>>>>>>> >>>>>>>>>>>> if you are using virsh for manage your vms, change this in your >>>>>>>>>>>> cluster.conf >>>>>>>>>>>> >>>>>>>>>>>> from >>>>>>>>>>>> use_virsh="0" >>>>>>>>>>>> to >>>>>>>>>>>> use_virsh="1" >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> 2013/11/22 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>>>>>> >>>>>>>>>>>>> I think i found the problem. >>>>>>>>>>>>> >>>>>>>>>>>>> virsh list --all does not show my vm . This is because it was >>>>>>>>>>>>> created on another node. and another node has it. Now I want to >>>>>>>>>>>>> start the >>>>>>>>>>>>> service on a different node in which it was not created or where >>>>>>>>>>>>> virsh list >>>>>>>>>>>>> --all does not have an entry. Is it possible to create this entry >>>>>>>>>>>>> using a >>>>>>>>>>>>> xen config file?Looks like this is now a Xen issue rather than a >>>>>>>>>>>>> linux-cluster issue . :) >>>>>>>>>>>>> >>>>>>>>>>>>> Paras. >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> On Thu, Nov 21, 2013 at 4:58 PM, emmanuel segura < >>>>>>>>>>>>> emi2f...@gmail.com> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> 1:did you verify your xen livemigration configuration? >>>>>>>>>>>>>> 2: where you vm disk reside? >>>>>>>>>>>>>> 3: can you see your vm defined on every cluster node with xm >>>>>>>>>>>>>> list? >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> 2013/11/21 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>>>>>>>> >>>>>>>>>>>>>>> This is what I get >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Hypervisor: xen >>>>>>>>>>>>>>> Management tool: virsh >>>>>>>>>>>>>>> Hypervisor URI: xen:/// >>>>>>>>>>>>>>> Migration URI format: xenmigr://target_host/ >>>>>>>>>>>>>>> Virtual machine guest1 is error: failed to get domain >>>>>>>>>>>>>>> 'guest1' >>>>>>>>>>>>>>> error: Domain not found: xenUnifiedDomainLookupByName >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> <debug> virsh -c xen:/// start guest1 >>>>>>>>>>>>>>> error: failed to get domain 'guest1' >>>>>>>>>>>>>>> error: Domain not found: xenUnifiedDomainLookupByName >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> ++ printf '\033]0;%s@%s:%s\007' root vtst3 '~' >>>>>>>>>>>>>>> [root@cvtst3 ~]# set +x >>>>>>>>>>>>>>> + set +x >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> I am wondering why it failed to get domain . >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -Paras. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On Thu, Nov 21, 2013 at 4:43 PM, emmanuel segura < >>>>>>>>>>>>>>> emi2f...@gmail.com> wrote: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> yes >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> 2013/11/21 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Well it is guest1. Isn't it?. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> <vm autostart="1" domain="myfd1" exclusive="0" >>>>>>>>>>>>>>>>> max_restarts="0" name="guest1" path="/vms_c" >>>>>>>>>>>>>>>>> recovery="restart" >>>>>>>>>>>>>>>>> restart_expire_time="0" use_virsh="0"/> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> It is a vm service if it matters. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> -Paras. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> On Thu, Nov 21, 2013 at 4:22 PM, emmanuel segura < >>>>>>>>>>>>>>>>> emi2f...@gmail.com> wrote: >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> use the servicename you defined in your cluster.conf >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> 2013/11/21 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Says: >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> Running in test mode. >>>>>>>>>>>>>>>>>>> No resource guest1 of type service found >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> -Paras. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> On Thu, Nov 21, 2013 at 4:07 PM, emmanuel segura < >>>>>>>>>>>>>>>>>>> emi2f...@gmail.com> wrote: >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> rg_test test /etc/cluster/cluster.conf start service >>>>>>>>>>>>>>>>>>>> guest1 >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> 2013/11/21 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> My failover domain looks like this: >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> <failoverdomain name="myfd1" nofailback="1" >>>>>>>>>>>>>>>>>>>>> ordered="1" restricted="0"> >>>>>>>>>>>>>>>>>>>>> <failoverdomainnode >>>>>>>>>>>>>>>>>>>>> name="vtst1" priority="1"/> >>>>>>>>>>>>>>>>>>>>> <failoverdomainnode >>>>>>>>>>>>>>>>>>>>> name="vtst3" priority="2"/> >>>>>>>>>>>>>>>>>>>>> <failoverdomainnode >>>>>>>>>>>>>>>>>>>>> name="vtst2" priority="3"/> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> </failoverdomain> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> I have vm service that uses this failover domain. If >>>>>>>>>>>>>>>>>>>>> my node vtst1 is offline, the service doesnot start on >>>>>>>>>>>>>>>>>>>>> vtst3 which is 2nd >>>>>>>>>>>>>>>>>>>>> in the priority. >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> I tried to start it with: clusvcadm -e vm:guest1 and >>>>>>>>>>>>>>>>>>>>> even with -F and -m option. >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> All i see is this error: >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <notice> start >>>>>>>>>>>>>>>>>>>>> on vm "guest1" returned 1 (generic error) >>>>>>>>>>>>>>>>>>>>> Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <warning> #68: >>>>>>>>>>>>>>>>>>>>> Failed to start vm:guest1; return value: 1 >>>>>>>>>>>>>>>>>>>>> Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <notice> >>>>>>>>>>>>>>>>>>>>> Stopping service vm:guest1 >>>>>>>>>>>>>>>>>>>>> Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <notice> >>>>>>>>>>>>>>>>>>>>> Service vm:guest1 is recovering >>>>>>>>>>>>>>>>>>>>> Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <warning> #71: >>>>>>>>>>>>>>>>>>>>> Relocating failed service vm:guest1 >>>>>>>>>>>>>>>>>>>>> Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <notice> >>>>>>>>>>>>>>>>>>>>> Service vm:guest1 is stopped >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> How do I debug? >>>>>>>>>>>>>>>>>>>>> Thanks! >>>>>>>>>>>>>>>>>>>>> Paras. >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> -- >>>>>>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>>>>>> >>>>>>>>>>>>>> -- >>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> Linux-cluster mailing list >>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> Linux-cluster mailing list >>>>>>>>> Linux-cluster@redhat.com >>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>> >>>>>>>> -- >>>>>>>> Linux-cluster mailing list >>>>>>>> Linux-cluster@redhat.com >>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>> >>>>>>> >>>>>>> >>>>>> >>>>>> -- >>>>>> Linux-cluster mailing list >>>>>> Linux-cluster@redhat.com >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>> >>>>> -- >>>>> Linux-cluster mailing list >>>>> Linux-cluster@redhat.com >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>> >>>> >>>> -- >>>> Linux-cluster mailing list >>>> Linux-cluster@redhat.com >>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>> >>> >>> >>> -- >>> esta es mi vida e me la vivo hasta que dios quiera >>> >>> -- >>> Linux-cluster mailing list >>> Linux-cluster@redhat.com >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>> >> >> >> -- >> Linux-cluster mailing list >> Linux-cluster@redhat.com >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > > > -- > esta es mi vida e me la vivo hasta que dios quiera > -- esta es mi vida e me la vivo hasta que dios quiera
-- Linux-cluster mailing list Linux-cluster@redhat.com https://www.redhat.com/mailman/listinfo/linux-cluster