On Tue, Sep 29, 2009 at 3:44 PM, brem belguebli <[email protected]> wrote: > It looks correct, rgmanager seems to start on all nodes > > what gives you clustat ? > > If rgmanager doesn't show, check out the logs something may have gone wrong. > > > 2009/9/29 Paras pradhan <[email protected]>: >> Change to 7 and i got this log >> >> Sep 29 15:33:50 cvtst1 rgmanager: [23295]: <notice> Shutting down >> Cluster Service Manager... >> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutting down >> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutting down >> Sep 29 15:33:50 cvtst1 clurgmgrd[22869]: <notice> Shutdown complete, exiting >> Sep 29 15:33:50 cvtst1 rgmanager: [23295]: <notice> Cluster Service >> Manager is stopped. >> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <notice> Resource Group >> Manager Starting >> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <info> Loading Service Data >> Sep 29 15:33:51 cvtst1 clurgmgrd[23324]: <debug> Loading Resource Rules >> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 21 rules loaded >> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Building Resource Trees >> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 0 resources defined >> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Loading Failover Domains >> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 1 domains defined >> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> 1 events defined >> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> Initializing Services >> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> Services Initialized >> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <debug> Event: Port Opened >> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: Local UP >> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: cvtst2 UP >> Sep 29 15:33:52 cvtst1 clurgmgrd[23324]: <info> State change: cvtst3 UP >> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (1:2:1) Processed >> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (0:1:1) Processed >> Sep 29 15:33:57 cvtst1 clurgmgrd[23324]: <debug> Event (0:3:1) Processed >> Sep 29 15:34:02 cvtst1 clurgmgrd[23324]: <debug> 3 events processed >> >> >> Anything unusual here? >> >> Paras. >> >> On Tue, Sep 29, 2009 at 11:51 AM, brem belguebli >> <[email protected]> wrote: >>> I use log_level=7 to have more debugging info. >>> >>> It seems 4 is not enough. >>> >>> Brem >>> >>> >>> 2009/9/29, Paras pradhan <[email protected]>: >>>> Withe log_level of 3 I got only this >>>> >>>> Sep 29 10:31:31 cvtst1 rgmanager: [7170]: <notice> Shutting down >>>> Cluster Service Manager... >>>> Sep 29 10:31:31 cvtst1 clurgmgrd[6673]: <notice> Shutting down >>>> Sep 29 10:31:41 cvtst1 clurgmgrd[6673]: <notice> Shutdown complete, exiting >>>> Sep 29 10:31:41 cvtst1 rgmanager: [7170]: <notice> Cluster Service >>>> Manager is stopped. >>>> Sep 29 10:31:42 cvtst1 clurgmgrd[7224]: <notice> Resource Group >>>> Manager Starting >>>> Sep 29 10:39:06 cvtst1 rgmanager: [10327]: <notice> Shutting down >>>> Cluster Service Manager... >>>> Sep 29 10:39:16 cvtst1 rgmanager: [10327]: <notice> Cluster Service >>>> Manager is stopped. >>>> Sep 29 10:39:16 cvtst1 clurgmgrd[10380]: <notice> Resource Group >>>> Manager Starting >>>> Sep 29 10:39:52 cvtst1 clurgmgrd[10380]: <notice> Member 1 shutting down >>>> >>>> I do not know what the last line means. >>>> >>>> rgmanager version I am running is: >>>> rgmanager-2.0.52-1.el5.centos >>>> >>>> I don't what has gone wrong. >>>> >>>> Thanks >>>> Paras. >>>> >>>> >>>> On Mon, Sep 28, 2009 at 6:41 PM, brem belguebli >>>> <[email protected]> wrote: >>>> > you mean it stopped successfully on all the nodes but it is failing to >>>> > start only on node cvtst1 ? >>>> > >>>> > look at the following page to make rgmanager more verbose. It 'll >>>> > help debug.... >>>> > >>>> > http://sources.redhat.com/cluster/wiki/RGManager >>>> > >>>> > at Logging Configuration section >>>> > >>>> > >>>> > >>>> > >>>> > 2009/9/29 Paras pradhan <[email protected]>: >>>> >> Brem, >>>> >> >>>> >> When I try to restart rgmanager on all the nodes, this time i do not >>>> >> see rgmanager running on the first node. But I do see on other 2 >>>> >> nodes. >>>> >> >>>> >> Log on the first node: >>>> >> >>>> >> Sep 28 18:13:58 cvtst1 clurgmgrd[24099]: <notice> Resource Group >>>> >> Manager Starting >>>> >> Sep 28 18:17:29 cvtst1 rgmanager: [24627]: <notice> Shutting down >>>> >> Cluster Service Manager... >>>> >> Sep 28 18:17:29 cvtst1 clurgmgrd[24099]: <notice> Shutting down >>>> >> Sep 28 18:17:39 cvtst1 clurgmgrd[24099]: <notice> Shutdown complete, >>>> >> exiting >>>> >> Sep 28 18:17:39 cvtst1 rgmanager: [24627]: <notice> Cluster Service >>>> >> Manager is stopped. >>>> >> Sep 28 18:17:40 cvtst1 clurgmgrd[24679]: <notice> Resource Group >>>> >> Manager Starting >>>> >> >>>> >> - >>>> >> It seems service is running , but I do not see rgmanger running using >>>> >> clustat >>>> >> >>>> >> >>>> >> Don't know what is going on. >>>> >> >>>> >> Thanks >>>> >> Paras. >>>> >> >>>> >> >>>> >> On Mon, Sep 28, 2009 at 5:46 PM, brem belguebli >>>> >> <[email protected]> wrote: >>>> >>> Paras, >>>> >>> >>>> >>> Another thing, it would have been more interesting to have a start >>>> >>> DEBUG not a stop. >>>> >>> >>>> >>> That's why I was asking you to first stop the vm manually on all your >>>> >>> nodes, stop eventually rgmanager on all the nodes to reset the >>>> >>> potential wrong states you may have, restart rgmanager. >>>> >>> >>>> >>> If your VM is configured to autostart, this will make it start. >>>> >>> >>>> >>> It should normally fail (as it does now). Send out your newly created >>>> >>> DEBUG file. >>>> >>> >>>> >>> 2009/9/29 brem belguebli <[email protected]>: >>>> >>>> Hi Paras, >>>> >>>> >>>> >>>> >>>> >>>> I don't know the xen/cluster combination well, but if I do remember >>>> >>>> well, I think I've read somewhere that when using xen you have to >>>> >>>> declare the use_virsh=0 key in the VM definition in the cluster.conf. >>>> >>>> >>>> >>>> This would make rgmanager use xm commands instead of virsh >>>> >>>> The DEBUG output shows clearly that you are using virsh to manage your >>>> >>>> VM instead of xm commands. >>>> >>>> Check out the RH docs about virtualization >>>> >>>> >>>> >>>> I'm not a 100% sure about that, I may be completely wrong. >>>> >>>> >>>> >>>> Brem >>>> >>>> >>>> >>>> 2009/9/28 Paras pradhan <[email protected]>: >>>> >>>>> The only thing I noticed is the message after stopping the vm using >>>> >>>>> xm >>>> >>>>> in all nodes and starting using clusvcadm is >>>> >>>>> >>>> >>>>> "Virtual machine guest1 is blocked" >>>> >>>>> >>>> >>>>> The whole DEBUG file is attached. >>>> >>>>> >>>> >>>>> >>>> >>>>> Thanks >>>> >>>>> Paras. >>>> >>>>> >>>> >>>>> On Fri, Sep 25, 2009 at 5:53 PM, brem belguebli >>>> >>>>> <[email protected]> wrote: >>>> >>>>>> There's a problem with the script that is called by rgmanager to >>>> >>>>>> start >>>> >>>>>> the VM, I don't know what causes it >>>> >>>>>> >>>> >>>>>> May be you should try something like : >>>> >>>>>> >>>> >>>>>> 1) stop the VM on all nodes with xm commands >>>> >>>>>> 2) edit the /usr/share/cluster/vm.sh script and add the following >>>> >>>>>> lines (after the #!/bin/bash ): >>>> >>>>>> exec >/tmp/DEBUG 2>&1 >>>> >>>>>> set -x >>>> >>>>>> 3) start the VM with clusvcadm -e vm:guest1 >>>> >>>>>> >>>> >>>>>> It should fail as it did before. >>>> >>>>>> >>>> >>>>>> edit the the /tmp/DEBUG file and you will be able to see where it >>>> >>>>>> fails (it may generate a lot of debug) >>>> >>>>>> >>>> >>>>>> 4) remove the debug lines from /usr/share/cluster/vm.sh >>>> >>>>>> >>>> >>>>>> Post the DEBUG file if you're not able to see where it fails. >>>> >>>>>> >>>> >>>>>> Brem >>>> >>>>>> >>>> >>>>>> 2009/9/26 Paras pradhan <[email protected]>: >>>> >>>>>>> No I am not manually starting not using automatic init scripts. >>>> >>>>>>> >>>> >>>>>>> I started the vm using: clusvcadm -e vm:guest1 >>>> >>>>>>> >>>> >>>>>>> I have just stopped using clusvcadm -s vm:guest1. For few seconds >>>> >>>>>>> it >>>> >>>>>>> says guest1 started . But after a while I can see the guest1 on all >>>> >>>>>>> three nodes. >>>> >>>>>>> >>>> >>>>>>> clustat says: >>>> >>>>>>> >>>> >>>>>>> Service Name Owner >>>> >>>>>>> (Last) >>>> >>>>>>> State >>>> >>>>>>> ------- ---- ----- >>>> >>>>>>> ------ >>>> >>>>>>> ----- >>>> >>>>>>> vm:guest1 (none) >>>> >>>>>>> stopped >>>> >>>>>>> >>>> >>>>>>> But I can see the vm from xm li. >>>> >>>>>>> >>>> >>>>>>> This is what I can see from the log: >>>> >>>>>>> >>>> >>>>>>> >>>> >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> start on vm >>>> >>>>>>> "guest1" >>>> >>>>>>> returned 1 (generic error) >>>> >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <warning> #68: Failed to >>>> >>>>>>> start >>>> >>>>>>> vm:guest1; return value: 1 >>>> >>>>>>> Sep 25 17:19:01 cvtst1 clurgmgrd[4298]: <notice> Stopping service >>>> >>>>>>> vm:guest1 >>>> >>>>>>> Sep 25 17:19:02 cvtst1 clurgmgrd[4298]: <notice> Service vm:guest1 >>>> >>>>>>> is >>>> >>>>>>> recovering >>>> >>>>>>> Sep 25 17:19:15 cvtst1 clurgmgrd[4298]: <notice> Recovering failed >>>> >>>>>>> service vm:guest1 >>>> >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> start on vm >>>> >>>>>>> "guest1" >>>> >>>>>>> returned 1 (generic error) >>>> >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <warning> #68: Failed to >>>> >>>>>>> start >>>> >>>>>>> vm:guest1; return value: 1 >>>> >>>>>>> Sep 25 17:19:16 cvtst1 clurgmgrd[4298]: <notice> Stopping service >>>> >>>>>>> vm:guest1 >>>> >>>>>>> Sep 25 17:19:17 cvtst1 clurgmgrd[4298]: <notice> Service vm:guest1 >>>> >>>>>>> is >>>> >>>>>>> recovering >>>> >>>>>>> >>>> >>>>>>> >>>> >>>>>>> Paras. >>>> >>>>>>> >>>> >>>>>>> On Fri, Sep 25, 2009 at 5:07 PM, brem belguebli >>>> >>>>>>> <[email protected]> wrote: >>>> >>>>>>>> Have you started your VM via rgmanager (clusvcadm -e vm:guest1) >>>> >>>>>>>> or >>>> >>>>>>>> using xm commands out of cluster control (or maybe a thru an >>>> >>>>>>>> automatic init script ?) >>>> >>>>>>>> >>>> >>>>>>>> When clustered, you should never be starting services (manually or >>>> >>>>>>>> thru automatic init script) out of cluster control >>>> >>>>>>>> >>>> >>>>>>>> The thing would be to stop your vm on all the nodes with the >>>> >>>>>>>> adequate >>>> >>>>>>>> xm command (not using xen myself) and try to start it with >>>> >>>>>>>> clusvcadm. >>>> >>>>>>>> >>>> >>>>>>>> Then see if it is started on all nodes (send clustat output) >>>> >>>>>>>> >>>> >>>>>>>> >>>> >>>>>>>> >>>> >>>>>>>> 2009/9/25 Paras pradhan <[email protected]>: >>>> >>>>>>>>> Ok. Please see below. my vm is running on all nodes though >>>> >>>>>>>>> clustat >>>> >>>>>>>>> says it is stopped. >>>> >>>>>>>>> >>>> >>>>>>>>> -- >>>> >>>>>>>>> [r...@cvtst1 ~]# clustat >>>> >>>>>>>>> Cluster Status for test @ Fri Sep 25 16:52:34 2009 >>>> >>>>>>>>> Member Status: Quorate >>>> >>>>>>>>> >>>> >>>>>>>>> Member Name >>>> >>>>>>>>> ID Status >>>> >>>>>>>>> ------ ---- >>>> >>>>>>>>> ---- ------ >>>> >>>>>>>>> cvtst2 1 >>>> >>>>>>>>> Online, rgmanager >>>> >>>>>>>>> cvtst1 2 >>>> >>>>>>>>> Online, >>>> >>>>>>>>> Local, rgmanager >>>> >>>>>>>>> cvtst3 3 >>>> >>>>>>>>> Online, rgmanager >>>> >>>>>>>>> >>>> >>>>>>>>> Service Name Owner >>>> >>>>>>>>> (Last) >>>> >>>>>>>>> State >>>> >>>>>>>>> ------- ---- ----- >>>> >>>>>>>>> ------ >>>> >>>>>>>>> ----- >>>> >>>>>>>>> vm:guest1 (none) >>>> >>>>>>>>> stopped >>>> >>>>>>>>> [r...@cvtst1 ~]# >>>> >>>>>>>>> >>>> >>>>>>>>> >>>> >>>>>>>>> --- >>>> >>>>>>>>> o/p of xm li on cvtst1 >>>> >>>>>>>>> >>>> >>>>>>>>> -- >>>> >>>>>>>>> [r...@cvtst1 ~]# xm li >>>> >>>>>>>>> Name ID Mem(MiB) VCPUs >>>> >>>>>>>>> State Time(s) >>>> >>>>>>>>> Domain-0 0 3470 2 >>>> >>>>>>>>> r----- 28939.4 >>>> >>>>>>>>> guest1 7 511 1 >>>> >>>>>>>>> -b---- 7727.8 >>>> >>>>>>>>> >>>> >>>>>>>>> o/p of xm li on cvtst2 >>>> >>>>>>>>> >>>> >>>>>>>>> -- >>>> >>>>>>>>> [r...@cvtst2 ~]# xm li >>>> >>>>>>>>> Name ID Mem(MiB) VCPUs >>>> >>>>>>>>> State Time(s) >>>> >>>>>>>>> Domain-0 0 3470 2 >>>> >>>>>>>>> r----- 31558.9 >>>> >>>>>>>>> guest1 21 511 1 >>>> >>>>>>>>> -b---- 7558.2 >>>> >>>>>>>>> --- >>>> >>>>>>>>> >>>> >>>>>>>>> Thanks >>>> >>>>>>>>> Paras. >>>> >>>>>>>>> >>>> >>>>>>>>> >>>> >>>>>>>>> >>>> >>>>>>>>> On Fri, Sep 25, 2009 at 4:22 PM, brem belguebli >>>> >>>>>>>>> <[email protected]> wrote: >>>> >>>>>>>>>> It looks like no. >>>> >>>>>>>>>> >>>> >>>>>>>>>> can you send an output of clustat of when the VM is running on >>>> >>>>>>>>>> multiple nodes at the same time? >>>> >>>>>>>>>> >>>> >>>>>>>>>> And by the way, another one after having stopped (clusvcadm -s >>>> >>>>>>>>>> vm:guest1) ? >>>> >>>>>>>>>> >>>> >>>>>>>>>> >>>> >>>>>>>>>> >>>> >>>>>>>>>> 2009/9/25 Paras pradhan <[email protected]>: >>>> >>>>>>>>>>> Anyone having issue as mine? Virtual machine service is not >>>> >>>>>>>>>>> being >>>> >>>>>>>>>>> properly handled by the cluster. >>>> >>>>>>>>>>> >>>> >>>>>>>>>>> >>>> >>>>>>>>>>> Thanks >>>> >>>>>>>>>>> Paras. >>>> >>>>>>>>>>> >>>> >>>>>>>>>>> On Mon, Sep 21, 2009 at 9:55 AM, Paras pradhan >>>> >>>>>>>>>>> <[email protected]> wrote: >>>> >>>>>>>>>>>> Ok.. here is my cluster.conf file >>>> >>>>>>>>>>>> >>>> >>>>>>>>>>>> -- >>>> >>>>>>>>>>>> [r...@cvtst1 cluster]# more cluster.conf >>>> >>>>>>>>>>>> <?xml version="1.0"?> >>>> >>>>>>>>>>>> <cluster alias="test" config_version="9" name="test"> >>>> >>>>>>>>>>>> <fence_daemon clean_start="0" post_fail_delay="0" >>>> >>>>>>>>>>>> post_join_delay="3"/> >>>> >>>>>>>>>>>> <clusternodes> >>>> >>>>>>>>>>>> <clusternode name="cvtst2" nodeid="1" >>>> >>>>>>>>>>>> votes="1"> >>>> >>>>>>>>>>>> <fence/> >>>> >>>>>>>>>>>> </clusternode> >>>> >>>>>>>>>>>> <clusternode name="cvtst1" nodeid="2" >>>> >>>>>>>>>>>> votes="1"> >>>> >>>>>>>>>>>> <fence/> >>>> >>>>>>>>>>>> </clusternode> >>>> >>>>>>>>>>>> <clusternode name="cvtst3" nodeid="3" >>>> >>>>>>>>>>>> votes="1"> >>>> >>>>>>>>>>>> <fence/> >>>> >>>>>>>>>>>> </clusternode> >>>> >>>>>>>>>>>> </clusternodes> >>>> >>>>>>>>>>>> <cman/> >>>> >>>>>>>>>>>> <fencedevices/> >>>> >>>>>>>>>>>> <rm> >>>> >>>>>>>>>>>> <failoverdomains> >>>> >>>>>>>>>>>> <failoverdomain name="myfd1" >>>> >>>>>>>>>>>> nofailback="0" ordered="1" restricted="0"> >>>> >>>>>>>>>>>> <failoverdomainnode >>>> >>>>>>>>>>>> name="cvtst2" priority="3"/> >>>> >>>>>>>>>>>> <failoverdomainnode >>>> >>>>>>>>>>>> name="cvtst1" priority="1"/> >>>> >>>>>>>>>>>> <failoverdomainnode >>>> >>>>>>>>>>>> name="cvtst3" priority="2"/> >>>> >>>>>>>>>>>> </failoverdomain> >>>> >>>>>>>>>>>> </failoverdomains> >>>> >>>>>>>>>>>> <resources/> >>>> >>>>>>>>>>>> <vm autostart="1" domain="myfd1" exclusive="0" >>>> >>>>>>>>>>>> max_restarts="0" >>>> >>>>>>>>>>>> name="guest1" path="/vms" recovery="r >>>> >>>>>>>>>>>> estart" restart_expire_time="0"/> >>>> >>>>>>>>>>>> </rm> >>>> >>>>>>>>>>>> </cluster> >>>> >>>>>>>>>>>> [r...@cvtst1 cluster]# >>>> >>>>>>>>>>>> ------ >>>> >>>>>>>>>>>> >>>> >>>>>>>>>>>> Thanks! >>>> >>>>>>>>>>>> Paras. >>>> >>>>>>>>>>>> >>>> >>>>>>>>>>>> >>>> >>>>>>>>>>>> On Sun, Sep 20, 2009 at 9:44 AM, Volker Dormeyer >>>> >>>>>>>>>>>> <[email protected]> wrote: >>>> >>>>>>>>>>>>> On Fri, Sep 18, 2009 at 05:08:57PM -0500, >>>> >>>>>>>>>>>>> Paras pradhan <[email protected]> wrote: >>>> >>>>>>>>>>>>>> I am using cluster suite for HA of xen virtual machines. >>>> >>>>>>>>>>>>>> Now I am >>>> >>>>>>>>>>>>>> having another problem. When I start the my xen vm in one >>>> >>>>>>>>>>>>>> node, it >>>> >>>>>>>>>>>>>> also starts on other nodes. Which daemon controls this? >>>> >>>>>>>>>>>>> >>>> >>>>>>>>>>>>> This is usually done bei clurgmgrd (which is part of the >>>> >>>>>>>>>>>>> rgmanager >>>> >>>>>>>>>>>>> package). To me, this sounds like a configuration problem. >>>> >>>>>>>>>>>>> Maybe, >>>> >>>>>>>>>>>>> you can post your cluster.conf? >>>> >>>>>>>>>>>>> >>>> >>>>>>>>>>>>> Regards, >>>> >>>>>>>>>>>>> Volker >>>> >>>>>>>>>>>>> >>>> >>>>>>>>>>>>> -- >>>> >>>>>>>>>>>>> Linux-cluster mailing list >>>> >>>>>>>>>>>>> [email protected] >>>> >>>>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>>>>>>>>>>>> >>>> >>>>>>>>>>>> >>>> >>>>>>>>>>> >>>> >>>>>>>>>>> -- >>>> >>>>>>>>>>> Linux-cluster mailing list >>>> >>>>>>>>>>> [email protected] >>>> >>>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>>>>>>>>>> >>>> >>>>>>>>>> >>>> >>>>>>>>>> -- >>>> >>>>>>>>>> Linux-cluster mailing list >>>> >>>>>>>>>> [email protected] >>>> >>>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>>>>>>>>> >>>> >>>>>>>>> >>>> >>>>>>>>> -- >>>> >>>>>>>>> Linux-cluster mailing list >>>> >>>>>>>>> [email protected] >>>> >>>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>>>>>>>> >>>> >>>>>>>> >>>> >>>>>>>> -- >>>> >>>>>>>> Linux-cluster mailing list >>>> >>>>>>>> [email protected] >>>> >>>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>>>>>>> >>>> >>>>>>> >>>> >>>>>>> -- >>>> >>>>>>> Linux-cluster mailing list >>>> >>>>>>> [email protected] >>>> >>>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>>>>>> >>>> >>>>>> >>>> >>>>>> -- >>>> >>>>>> Linux-cluster mailing list >>>> >>>>>> [email protected] >>>> >>>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>>>>> >>>> >>>>> >>>> >>>>> -- >>>> >>>>> Linux-cluster mailing list >>>> >>>>> [email protected] >>>> >>>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>>>> >>>> >>>> >>>> >>> >>>> >>> -- >>>> >>> Linux-cluster mailing list >>>> >>> [email protected] >>>> >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>> >>>> >> >>>> >> -- >>>> >> Linux-cluster mailing list >>>> >> [email protected] >>>> >> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >> >>>> > >>>> > -- >>>> > Linux-cluster mailing list >>>> > [email protected] >>>> > https://www.redhat.com/mailman/listinfo/linux-cluster >>>> > >>>> >>>> -- >>>> Linux-cluster mailing list >>>> [email protected] >>>> https://www.redhat.com/mailman/listinfo/linux-cluster >>>> >>> >>> -- >>> Linux-cluster mailing list >>> [email protected] >>> https://www.redhat.com/mailman/listinfo/linux-cluster >>> >> >> -- >> Linux-cluster mailing list >> [email protected] >> https://www.redhat.com/mailman/listinfo/linux-cluster >> > > -- > Linux-cluster mailing list > [email protected] > https://www.redhat.com/mailman/listinfo/linux-cluster >
-- Linux-cluster mailing list [email protected] https://www.redhat.com/mailman/listinfo/linux-cluster
