ok, but your vm doesn't start on others nodes, i think, for configuration problems ================================================================ Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <notice> start on vm "guest1" returned 1 (generic error) Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <warning> #68: Failed to start vm:guest1; return value: 1 Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <notice> Stopping service vm:guest1 Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <notice> Service vm:guest1 is recovering Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <warning> #71: Relocating failed service vm:guest1 Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <notice> Service vm:guest1 is stopped ================================================================ in few words, try in every cluster node
export OCF_RESKEY_name="guest1" ; OCF_RESKEY_path="/vms_c" set -x /usr/share/cluster/vm.sh start /usr/share/cluster/vm.sh stop after you check if your vm can start and stop on every cluster node, /usr/share/cluster/vm.sh start /usr/share/cluster/vm.sh migrate name_of_a_cluster_node can you show me your vm configuration under /vms_c? Thanks Emmanuel 2013/11/22 Paras pradhan <pradhanpa...@gmail.com> > And also to test I made use_virsh=1 , same problem. The vm does not start > up if the FD domains are offline. > > -Paras. > > > On Fri, Nov 22, 2013 at 10:37 AM, Paras pradhan <pradhanpa...@gmail.com>wrote: > >> Well thats seems to theoretically correct. But right now my cluser has >> use_virsh=0 and I don't have any issue untill my mebmers on the failover >> domains are offline. So wondering what is it that clusvcadm -e is looking >> when I don't use virsh . >> >> >> >> >> On Fri, Nov 22, 2013 at 10:05 AM, emmanuel segura <emi2f...@gmail.com>wrote: >> >>> If you used virt-install, i think you need use virsh, the cluster uses >>> xm xen command if you got use_virsh=0 and virsh if you got use_virsh=1 in >>> your cluster config >>> >>> >>> 2013/11/22 Paras pradhan <pradhanpa...@gmail.com> >>> >>>> I use virt-install to create virtual machines. Is there a way to debug >>>> why clusvcadm -e vm:guest1 is failing? vm.sh seems to use virsh and my >>>> cluster.conf has use_virsh=0 >>>> >>>> >>>> Thanks >>>> >>>> Paras. >>>> >>>> >>>> On Nov 21, 2013 5:53 PM, "emmanuel segura" <emi2f...@gmail.com> wrote: >>>> >>>>> but did you configure your vm with xen tools or using virt-manager? >>>>> >>>>> >>>>> 2013/11/22 Paras pradhan <pradhanpa...@gmail.com> >>>>> >>>>>> Well no i don't want to use virsh. But as we are debugging with >>>>>> virsh now i found a strange issue. >>>>>> >>>>>> I exported an xml file and imported to all nodes . Ran >>>>>> >>>>>> >>>>>> --- >>>>>> name="guest1" path="/vms_c" >>>>>> >>>>>> export OCF_RESKEY_name="guest1" ; OCF_RESKEY_path="/vms_c" >>>>>> >>>>>> set -x >>>>>> /usr/share/cluster/vm.sh start >>>>>> set +x >>>>>> >>>>>> -- >>>>>> vm starts now. BUT from a cluster service : cluvscam -e vm:guest1 , >>>>>> same error. >>>>>> >>>>>> >>>>>> So if i populate all my domains' config files to all my cluser nodes >>>>>> and make use_virsh=1, then the issue is resolved. But this is a lot of >>>>>> work >>>>>> for those who have hundreds of vm. >>>>>> >>>>>> vm.start uses virsh . Is there a way to tell him not use virsh? >>>>>> >>>>>> >>>>>> Thanks >>>>>> Paras. >>>>>> >>>>>> >>>>>> On Thu, Nov 21, 2013 at 5:19 PM, emmanuel segura >>>>>> <emi2f...@gmail.com>wrote: >>>>>> >>>>>>> if you are using virsh for manage your vms, change this in your >>>>>>> cluster.conf >>>>>>> >>>>>>> from >>>>>>> use_virsh="0" >>>>>>> to >>>>>>> use_virsh="1" >>>>>>> >>>>>>> >>>>>>> 2013/11/22 Paras pradhan <pradhanpa...@gmail.com> >>>>>>> >>>>>>>> I think i found the problem. >>>>>>>> >>>>>>>> virsh list --all does not show my vm . This is because it was >>>>>>>> created on another node. and another node has it. Now I want to start >>>>>>>> the >>>>>>>> service on a different node in which it was not created or where virsh >>>>>>>> list >>>>>>>> --all does not have an entry. Is it possible to create this entry >>>>>>>> using a >>>>>>>> xen config file?Looks like this is now a Xen issue rather than a >>>>>>>> linux-cluster issue . :) >>>>>>>> >>>>>>>> Paras. >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> On Thu, Nov 21, 2013 at 4:58 PM, emmanuel segura < >>>>>>>> emi2f...@gmail.com> wrote: >>>>>>>> >>>>>>>>> 1:did you verify your xen livemigration configuration? >>>>>>>>> 2: where you vm disk reside? >>>>>>>>> 3: can you see your vm defined on every cluster node with xm list? >>>>>>>>> >>>>>>>>> >>>>>>>>> 2013/11/21 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>>> >>>>>>>>>> This is what I get >>>>>>>>>> >>>>>>>>>> Hypervisor: xen >>>>>>>>>> Management tool: virsh >>>>>>>>>> Hypervisor URI: xen:/// >>>>>>>>>> Migration URI format: xenmigr://target_host/ >>>>>>>>>> Virtual machine guest1 is error: failed to get domain 'guest1' >>>>>>>>>> error: Domain not found: xenUnifiedDomainLookupByName >>>>>>>>>> >>>>>>>>>> <debug> virsh -c xen:/// start guest1 >>>>>>>>>> error: failed to get domain 'guest1' >>>>>>>>>> error: Domain not found: xenUnifiedDomainLookupByName >>>>>>>>>> >>>>>>>>>> ++ printf '\033]0;%s@%s:%s\007' root vtst3 '~' >>>>>>>>>> [root@cvtst3 ~]# set +x >>>>>>>>>> + set +x >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> I am wondering why it failed to get domain . >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -Paras. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Thu, Nov 21, 2013 at 4:43 PM, emmanuel segura < >>>>>>>>>> emi2f...@gmail.com> wrote: >>>>>>>>>> >>>>>>>>>>> yes >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> 2013/11/21 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>>>>> >>>>>>>>>>>> Well it is guest1. Isn't it?. >>>>>>>>>>>> >>>>>>>>>>>> <vm autostart="1" domain="myfd1" exclusive="0" max_restarts="0" >>>>>>>>>>>> name="guest1" path="/vms_c" recovery="restart" >>>>>>>>>>>> restart_expire_time="0" >>>>>>>>>>>> use_virsh="0"/> >>>>>>>>>>>> >>>>>>>>>>>> It is a vm service if it matters. >>>>>>>>>>>> >>>>>>>>>>>> -Paras. >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> On Thu, Nov 21, 2013 at 4:22 PM, emmanuel segura < >>>>>>>>>>>> emi2f...@gmail.com> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> use the servicename you defined in your cluster.conf >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> 2013/11/21 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>>>>>>> >>>>>>>>>>>>>> Says: >>>>>>>>>>>>>> >>>>>>>>>>>>>> Running in test mode. >>>>>>>>>>>>>> No resource guest1 of type service found >>>>>>>>>>>>>> >>>>>>>>>>>>>> -Paras. >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Thu, Nov 21, 2013 at 4:07 PM, emmanuel segura < >>>>>>>>>>>>>> emi2f...@gmail.com> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> rg_test test /etc/cluster/cluster.conf start service guest1 >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> 2013/11/21 Paras pradhan <pradhanpa...@gmail.com> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Hi, >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> My failover domain looks like this: >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> <failoverdomain name="myfd1" nofailback="1" ordered="1" >>>>>>>>>>>>>>>> restricted="0"> >>>>>>>>>>>>>>>> <failoverdomainnode >>>>>>>>>>>>>>>> name="vtst1" priority="1"/> >>>>>>>>>>>>>>>> <failoverdomainnode >>>>>>>>>>>>>>>> name="vtst3" priority="2"/> >>>>>>>>>>>>>>>> <failoverdomainnode >>>>>>>>>>>>>>>> name="vtst2" priority="3"/> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> </failoverdomain> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> I have vm service that uses this failover domain. If my >>>>>>>>>>>>>>>> node vtst1 is offline, the service doesnot start on vtst3 >>>>>>>>>>>>>>>> which is 2nd in >>>>>>>>>>>>>>>> the priority. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> I tried to start it with: clusvcadm -e vm:guest1 and even >>>>>>>>>>>>>>>> with -F and -m option. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> All i see is this error: >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <notice> start on >>>>>>>>>>>>>>>> vm "guest1" returned 1 (generic error) >>>>>>>>>>>>>>>> Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <warning> #68: >>>>>>>>>>>>>>>> Failed to start vm:guest1; return value: 1 >>>>>>>>>>>>>>>> Nov 21 15:40:29 vtst3 clurgmgrd[13911]: <notice> Stopping >>>>>>>>>>>>>>>> service vm:guest1 >>>>>>>>>>>>>>>> Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <notice> Service >>>>>>>>>>>>>>>> vm:guest1 is recovering >>>>>>>>>>>>>>>> Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <warning> #71: >>>>>>>>>>>>>>>> Relocating failed service vm:guest1 >>>>>>>>>>>>>>>> Nov 21 15:40:35 vtst3 clurgmgrd[13911]: <notice> Service >>>>>>>>>>>>>>>> vm:guest1 is stopped >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> How do I debug? >>>>>>>>>>>>>>>> Thanks! >>>>>>>>>>>>>>>> Paras. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> -- >>>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> >>>>>>>>>>>>>> -- >>>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>>>>> >>>>>>>>>>>>> -- >>>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>>> -- >>>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>>>> >>>>>>>>>>> -- >>>>>>>>>>> Linux-cluster mailing list >>>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> Linux-cluster mailing list >>>>>>>>>> Linux-cluster@redhat.com >>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>>>> >>>>>>>>> -- >>>>>>>>> Linux-cluster mailing list >>>>>>>>> Linux-cluster@redhat.com >>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> Linux-cluster mailing list >>>>>>>> Linux-cluster@redhat.com >>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> -- >>>>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>>>> >>>>>>> -- >>>>>>> Linux-cluster mailing list >>>>>>> Linux-cluster@redhat.com >>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> Linux-cluster mailing list >>>>>> Linux-cluster@redhat.com >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>> >>>>> -- >>>>> Linux-cluster mailing list >>>>> Linux-cluster@redhat.com >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>>> >>>> >>>> -- >>>> Linux-cluster mailing list >>>> Linux-cluster@redhat.com >>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>> >>> >>> >>> -- >>> esta es mi vida e me la vivo hasta que dios quiera >>> >>> -- >>> Linux-cluster mailing list >>> Linux-cluster@redhat.com >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>> >> >> > > -- > Linux-cluster mailing list > Linux-cluster@redhat.com > https://www.redhat.com/mailman/listinfo/linux-cluster > -- esta es mi vida e me la vivo hasta que dios quiera
-- Linux-cluster mailing list Linux-cluster@redhat.com https://www.redhat.com/mailman/listinfo/linux-cluster