Hello William

i did new you are using drbd and i dont't know what type of configuration
you using

But it's better you try to start clvm with clvmd -d

like thak we can see what it's the problem

Il giorno 14 marzo 2012 14:02, William Seligman <selig...@nevis.columbia.edu
> ha scritto:

> On 3/14/12 6:02 AM, emmanuel segura wrote:
>
>  I think it's better you make clvmd start at boot
>>
>> chkconfig cman on ; chkconfig clvmd on
>>
>
> I've already tried it. It doesn't work. The problem is that my LVM
> information is on the drbd. If I start up clvmd before drbd, it won't find
> the logical volumes.
>
> I also don't see why that would make a difference (although this could be
> part of the confusion): a service is a service. I've tried starting up
> clvmd inside and outside pacemaker control, with the same problem. Why
> would starting clvmd at boot make a difference?
>
>  Il giorno 13 marzo 2012 23:29, William Seligman<seligman@nevis.**
>> columbia.edu <selig...@nevis.columbia.edu>
>>
>>> ha scritto:
>>>
>>
>>  On 3/13/12 5:50 PM, emmanuel segura wrote:
>>>
>>>  So if you using cman why you use lsb::clvmd
>>>>
>>>> I think you are very confused
>>>>
>>>
>>> I don't dispute that I may be very confused!
>>>
>>> However, from what I can tell, I still need to run clvmd even if
>>> I'm running cman (I'm not using rgmanager). If I just run cman,
>>> gfs2 and any other form of mount fails. If I run cman, then clvmd,
>>> then gfs2, everything behaves normally.
>>>
>>> Going by these instructions:
>>>
>>> <https://alteeve.com/w/2-Node_**Red_Hat_KVM_Cluster_Tutorial<https://alteeve.com/w/2-Node_Red_Hat_KVM_Cluster_Tutorial>
>>> >
>>>
>>> the resources he puts under "cluster control" (rgmanager) I have to
>>> put under pacemaker control. Those include drbd, clvmd, and gfs2.
>>>
>>> The difference between what I've got, and what's in "Clusters From
>>> Scratch", is in CFS they assign one DRBD volume to a single
>>> filesystem. I create an LVM physical volume on my DRBD resource,
>>> as in the above tutorial, and so I have to start clvmd or the
>>> logical volumes in the DRBD partition won't be recognized.>> Is
>>> there some way to get logical volumes recognized automatically by
>>> cman without rgmanager that I've missed?
>>>
>>
>>  Il giorno 13 marzo 2012 22:42, William Seligman<
>>>>
>>> selig...@nevis.columbia.edu
>>>
>>>> ha scritto:
>>>>>
>>>>
>>>>  On 3/13/12 12:29 PM, William Seligman wrote:
>>>>>
>>>>>> I'm not sure if this is a "Linux-HA" question; please direct
>>>>>> me to the appropriate list if it's not.
>>>>>>
>>>>>> I'm setting up a two-node cman+pacemaker+gfs2 cluster as
>>>>>> described in "Clusters From Scratch." Fencing is through
>>>>>> forcibly rebooting a node by cutting and restoring its power
>>>>>> via UPS.
>>>>>>
>>>>>> My fencing/failover tests have revealed a problem. If I
>>>>>> gracefully turn off one node ("crm node standby"; "service
>>>>>> pacemaker stop"; "shutdown -r now") all the resources
>>>>>> transfer to the other node with no problems. If I cut power
>>>>>> to one node (as would happen if it were fenced), the
>>>>>> lsb::clvmd resource on the remaining node eventually fails.
>>>>>> Since all the other resources depend on clvmd, all the
>>>>>> resources on the remaining node stop and the cluster is left
>>>>>> with nothing running.
>>>>>>
>>>>>> I've traced why the lsb::clvmd fails: The monitor/status
>>>>>> command includes "vgdisplay", which hangs indefinitely.
>>>>>> Therefore the monitor will always time-out.
>>>>>>
>>>>>> So this isn't a problem with pacemaker, but with clvmd/dlm:
>>>>>> If a node is cut off, the cluster isn't handling it properly.
>>>>>> Has anyone on this list seen this before? Any ideas?
>>>>>>
>>>>> >> Details:
>>>>
>>>>>
>>>>>> versions:
>>>>>> Redhat Linux 6.2 (kernel 2.6.32)
>>>>>> cman-3.0.12.1
>>>>>> corosync-1.4.1
>>>>>> pacemaker-1.1.6
>>>>>> lvm2-2.02.87
>>>>>> lvm2-cluster-2.02.87
>>>>>>
>>>>>
>>>>> This may be a Linux-HA question after all!
>>>>>
>>>>> I ran a few more tests. Here's the output from a typical test of
>>>>>
>>>>> grep -E "(dlm|gfs2}clvmd|fenc|syslogd)**" /var/log/messages
>>>>>
>>>>> <http://pastebin.com/uqC6bc1b>
>>>>>
>>>>> It looks like what's happening is that the fence agent (one I
>>>>> wrote) is not returning the proper error code when a node
>>>>> crashes. According to this page, if a fencing agent fails GFS2
>>>>> will freeze to protect the data:
>>>>>
>>>>> <http://docs.redhat.com/docs/**en-US/Red_Hat_Enterprise_**
>>>>> Linux/6/html/Global_File_**System_2/s1-gfs2hand-allnodes.**html<http://docs.redhat.com/docs/en-US/Red_Hat_Enterprise_Linux/6/html/Global_File_System_2/s1-gfs2hand-allnodes.html>>
>>>>>
>>>>> As a test, I tried to fence my test node via standard means:
>>>>>
>>>>> stonith_admin -F 
>>>>> orestes-corosync.nevis.**columbia.edu<http://orestes-corosync.nevis.columbia.edu>
>>>>>
>>>>> These were the log messages, which show that stonith_admin did
>>>>> its job and CMAN was notified of the
>>>>> fencing:<http://pastebin.com/**jaH820Bv <http://pastebin.com/jaH820Bv>
>>>>> >.
>>>>>
>>>>> Unfortunately, I still got the gfs2 freeze, so this is not the
>>>>> complete story.
>>>>>
>>>>> First things first. I vaguely recall a web page that went over
>>>>> the STONITH return codes, but I can't locate it again. Is there
>>>>> any reference to the return codes expected from a fencing
>>>>> agent, perhaps as function of the state of the fencing device?
>>>>>
>>>>
> --
> Bill Seligman             | 
> mailto://seligman@nevis.**columbia.edu<selig...@nevis.columbia.edu>
> Nevis Labs, Columbia Univ | 
> http://www.nevis.columbia.edu/**~seligman/<http://www.nevis.columbia.edu/%7Eseligman/>
> PO Box 137                |
> Irvington NY 10533  USA   | Phone: (914) 591-2823
>
>
> _______________________________________________
> Linux-HA mailing list
> Linux-HA@lists.linux-ha.org
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems
>



-- 
esta es mi vida e me la vivo hasta que dios quiera
_______________________________________________
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to