[Linux-cluster] Where logged messages when intra cluster down?

2014-05-29 Thread Gianluca Cecchi
Hello, on a two node cluster with CentOS 6.3 I set this in cluster.conf for testing Then I cut the intra cluster lan and only when the configured totem token ti

Re: [Linux-cluster] Info on restart of non critical resources

2013-11-19 Thread Gianluca Cecchi
On Tue, Nov 19, 2013 at 11:06 AM, Gianluca Cecchi wrote: > Is this expected behaviour? Any way to configure to try to restart in > place the resource without manual intervention when a resource is > configured as non critical? > > Thanks in advance, > Gianluca BTW: I did read

[Linux-cluster] Info on restart of non critical resources

2013-11-19 Thread Gianluca Cecchi
Hello, I have a cluster with RH EL 6.3 cman-3.0.12.1-32.el6_3.2.x86_64 rgmanager-3.0.12.1-12.el6.x86_64 I configure ssh in cluster changing the default init script. Then I configure it as a non critical resource in a service section ... ... . . .

Re: [Linux-cluster] Info on clvmd with halvm on rhel 6.3 based clusters

2013-07-05 Thread Gianluca Cecchi
On Fri, Jul 5, 2013 at 2:42 AM, Ryan Mitchell wrote: > You aren't starting rgmanager with the -N option are you? It is not the > default. > # man clurgmgrd >-N Do not perform stop-before-start. Combined with the -Z > flag to clusvcadm, this can be used to allow rgmanager to be up

[Linux-cluster] Info on clvmd with halvm on rhel 6.3 based clusters

2013-07-04 Thread Gianluca Cecchi
Hello, I already read these technotes so that it seems my configuration is coherent with them: https://access.redhat.com/site/documentation/en-US/Red_Hat_Enterprise_Linux/6/html/Cluster_Administration/ap-ha-halvm-CA.html https://access.redhat.com/site/solutions/409813 basically I would like to us

Re: [Linux-cluster] Netapp snapdrive connect doesn't like clustered volumes

2012-12-10 Thread Gianluca Cecchi
On Mon, 10 Dec 2012 23:36:51 +1100 yu song wrote: > ok..firstly confirm whether you are using HA-LVM, not GFS/GFS2 .. correct? > if ha-lvm, your vg should only be imported/mounted on one node, instead of > multiple nodes. In my non-CLVMD configuration I use HA-LVM with these entries in cluster.

Re: [Linux-cluster] Netapp snapdrive connect doesn't like clustered volumes

2012-12-10 Thread Gianluca Cecchi
On Mon, 10 Dec 2012 21:02:19 +1100 Yu wrote: > vgchange -cn vgname > Can disable cluster flag > Regards > Yu Ok, thanks, I thought about this option too and it seems it works. More testing in place. In this cluster I have many services and each one is bundled with its own VGs, so that each node is

[Linux-cluster] Netapp snapdrive connect doesn't like clustered volumes

2012-12-10 Thread Gianluca Cecchi
Hello, I was evaluating migration from a rhel 5.8 two node cluster with services based on ha-lvm to a clvmd one. I'm using netapp storage with snapshot technology. I quiesce db, then make a snapshot and then a third server connects to the flashcopy volume, with "snapdrive connect" command and makes

[Linux-cluster] How to see what node is the master for quorum disk?

2012-08-10 Thread Gianluca Cecchi
Hello, in qdiskd.log I get at cluster startup the node that becomes master for quorum disk. config is in fact something like and in syslog.conf # qdisk logging local4.*/var/log/qdiskd.log The file is rotated so after some time I have only empty qd

Re: [Linux-cluster] clvmd problems with centos 6.3 or normal clvmd behaviour?

2012-08-03 Thread Gianluca Cecchi
further debug. While the node starts and clvmd hangs mount -t debugfs debug /sys/kernel/debug # dlm_tool lockdump clvmd id 024d0001 gr CR rq IV pid 3155 master 0 "7w53faJiHU9ZYs0Kw8CmFcdsSXVqujthdh8uMexWiIBPmEecuGc42XgLwwm2VcXU" id 01710001 gr CR rq IV pid 3155 master 0 "7w53faJiHU9ZYs0Kw8CmFcds

Re: [Linux-cluster] clvmd problems with centos 6.3 or normal clvmd behaviour?

2012-08-02 Thread Gianluca Cecchi
On Thu, 2 Aug 2012 16:12:24 +0200 emmanuel segura wrote: > can you show me your lvm.conf? Here it is. Gianluca lvm.conf Description: Binary data -- Linux-cluster mailing list Linux-cluster@redhat.com https://www.redhat.com/mailman/listinfo/linux-cluster

Re: [Linux-cluster] clvmd problems with centos 6.3 or normal clvmd behaviour?

2012-08-02 Thread Gianluca Cecchi
On Thu, 2 Aug 2012 07:07:25 -0600 Corey Kovacs wrte: > I might be reading this wrong but just in case, I thought I'd point this out. > [snip] > A single node can maintain quorum since 2+3>(9/2). > In a split brain condition where a single node cannot talk to the other > nodes, this could be disast

Re: [Linux-cluster] Update a Cluster from RHEL/CentOS 6.2 to 6.3

2012-08-02 Thread Gianluca Cecchi
On Thu, 02 Aug 2012 09:39:34 +0200 Heiko Nardmann wrote: > If that is a real production system and not just for playing you should setup > a test environment before and also create a plan which usecases should run > with the new cluster. +1 for sure for what Heiko recommended Plus, first place:

Re: [Linux-cluster] clvmd problems with centos 6.3 or normal clvmd behaviour?

2012-08-02 Thread Gianluca Cecchi
On Wed, Aug 1, 2012 at 6:15 PM, Gianluca Cecchi wrote: > On Wed, 1 Aug 2012 16:26:38 +0200 emmanuel segura wrote: >> Why you don't remove expected_votes=3 and let the cluster automatic >> calculate that > > Thanks for your answer Emmanuel, but cman starts correctly,

Re: [Linux-cluster] clvmd problems with centos 6.3 or normal clvmd behaviour?

2012-08-01 Thread Gianluca Cecchi
On Wed, 1 Aug 2012 16:26:38 +0200 emmanuel segura wrote: > Why you don't remove expected_votes=3 and let the cluster automatic calculate > that Thanks for your answer Emmanuel, but cman starts correctly, while the problem seems related to vgchange -aly command hanging. But I tried that option too

[Linux-cluster] clvmd problems with centos 6.3 or normal clvmd behaviour?

2012-08-01 Thread Gianluca Cecchi
Hello, testing a three node cluster + quorum disk and clvmd. I was at CentOS 6.2 and I seem to remember to be able to start a single node. Correct? Then I upgraded to CentOS 6.3 and had a working environment. My config has At the moment two nodes are in another site that is powered down and I nee

Re: [Linux-cluster] Problem with ping as an heuristic with qdiskd

2012-03-12 Thread Gianluca Cecchi
On Fri, 9 Mar 2012 17:29:06 +0100 emmanuel segura wrote: > i'll try to be more clear > i work on redhat cluster from 2 years and i seen this topic so much times Sorry, I didn't want to offend anyone. I have been working on rhcs (and other companions from other OSes) for many years too... > I thin

Re: [Linux-cluster] Problem with ping as an heuristic with qdiskd

2012-03-09 Thread Gianluca Cecchi
On Fri, 9 Mar 2012 15:39:43 +0100, emmanuel segura wrote: > Hello Gianluca > Do you have a cluster private network? > if your answer it's yes i recommend don't use heuristic because if your > cluster public network goes down > your cluster take a fencing loop > > Or you can do something better, u

[Linux-cluster] Problem with ping as an heuristic with qdiskd

2012-03-09 Thread Gianluca Cecchi
Hello, I have a cluster in RH EL 5.7 with quorum disk and an heuristic. Current versions of main cluster packages are: rgmanager-2.0.52-21.el5_7.1 cman-2.0.115-85.el5_7.3 This is the loaded heuristic Heuristic: 'ping -c1 -w1 10.4.5.250' score=1 interval=2 tko=200 Line in cluster.conf: where 10

Re: [Linux-cluster] [Linux-HA] resource agents 3.9.1 final release

2011-06-17 Thread Gianluca Cecchi
On Fri, Jun 17, 2011 at 9:28 AM, Fabio M. Di Nitto wrote: > Lon, what's your opinion on this one? Some other considerations of mine. This of the current "abort" default option (as in RH EL 5 cluster suite base) is indeed a difficulty, in case of planned maintenance, so that a change inside the ag

Re: [Linux-cluster] [Linux-HA] resource agents 3.9.1 final release

2011-06-16 Thread Gianluca Cecchi
On Thu, Jun 16, 2011 at 3:13 PM, Fabio M. Di Nitto wrote: > Highlights for the rgmanager resource agents set: > > - oracledb: use shutdown immediate hello, from oracledb.sh.in I can see this actually is not a configurable parameter, so that I cannot choose between "immediate" and "abort", and I

Re: [Linux-cluster] Using specific physical interface to migrate vms

2011-04-15 Thread Gianluca Cecchi
On Thu, 14 Apr 2011 12:44:39 +0200 carlopmart wrote: > How can I configure cluster.conf file to assign eth1 interface when vm live > migration will be required?? Already replied on rhelv6 list. Next time please don't cross post. Send to one list and eventually only after some time choose another

[Linux-cluster] Considerations about fence_virtd with fence_xvm based guests

2011-03-29 Thread Gianluca Cecchi
Hello, I have 2 x rh el 6.0 hosts (rhev1 and rhev2) where I enabled ha and resilient storage beta channels. I'm testing from the Beta HA Addon channel the checkpoint backend This necessary because I want to test managing clusters of rh el 5 guests (where for example I would keep one guest restricte

Re: [Linux-cluster] rhel6 node start causes power on of the other one

2011-03-25 Thread Gianluca Cecchi
On Fri, 25 Mar 2011 07:19:40 +0530, Rajagopal Swaminathan wrote: > On which CPU should service guard run on? ServiceGuard is an HA solution with similar targets as RHCS or other solutions, older or newer ones (aka Kimberlite, LifeKeeper, heartbeat 2.x, Pacemaker/Corosync, Pacemaker/Heartbeat and

Re: [Linux-cluster] rhel6 node start causes power on of the other one

2011-03-24 Thread Gianluca Cecchi
On Thu, 24 Mar 2011 22:33:00 +0530, Rajagopal Swaminathan wrote: > On 3/24/11, Gianluca Cecchi wrote: > > such as need for direct access to server console or cluster without any > > node running at > > all without manual intervention ... > > > > Pardon my ignora

Re: [Linux-cluster] rhel6 node start causes power on of the other one

2011-03-24 Thread Gianluca Cecchi
On Tue, 22 Mar 2011 11:02:09 -0500 Robert Hayden wrote: > I believe you will want to investigate the "clean_start" property in the > fence_daemon stanza (RHEL 5). > Unsure if it is in RHEL6/Cluster3 code. It is my understanding that the > property can be used to > by-pass the timeout and remote

Re: [Linux-cluster] rhel6 node start causes power on of the other one

2011-03-22 Thread Gianluca Cecchi
On Tue, 22 Mar 2011 11:47:58 +0100, Fabio M. Di Nitto wrote: > For RHEL related questions you should always file a ticket with GSS. yes, it is my usual behaviour, but tipically I prefer to analyze in advance and know if a problem I'm encountering is a bug or only my fault in docs understanding...

[Linux-cluster] rhel6 node start causes power on of the other one

2011-03-22 Thread Gianluca Cecchi
Hello, I'm using latest updates on a 2 nodes rhel 6 based cluster. At the moment no quorum disk defined, so this line inside cluster.conf # rpm -q cman rgmanager fence-agents ricci corosync cman-3.0.12-23.el6_0.6.x86_64 rgmanager-3.0.12-10.el6.x86_64 fence-agents-3.0.12-8.el6_0.3.x86_64 ricci-0.1

Re: [Linux-cluster] unable to live migrate a vm in rh el 6: Migration unexpectedly failed

2011-03-10 Thread Gianluca Cecchi
On Wed, Mar 9, 2011 at 9:47 AM, Gianluca Cecchi wrote: [snip] > Or something related with firewall perhaps. > Can I stop firewall at all and have libvirtd working at the same time > to test ...? > I know libvirtd puts some iptables rules itself.. > > Gianluca > OK. It

Re: [Linux-cluster] unable to live migrate a vm in rh el 6: Migration unexpectedly failed

2011-03-09 Thread Gianluca Cecchi
Here is the output of the command strace -f virsh migrate --live exorapr1 qemu+ssh://intrarhev1/system Note that if I run the same with rhev1 (main host name and not intracluster) instead of intrarhev1, I'm asked for the ssh password (ok because I set ssh equivalence only for intracluster) but at

Re: [Linux-cluster] unable to live migrate a vm in rh el 6: Migration unexpectedly failed

2011-03-09 Thread Gianluca Cecchi
On Mon, Mar 7, 2011 at 11:10 PM, Gianluca Cecchi wrote: > Nothing comes to mind; in my RHEL6 development cluster, I have a > custom SELinux policy: I confirm that SElinux is disabled and [root@rhev1 ~]# chkconfig --list | grep audit auditd 0:off 1:off 2:off 3:off 4:off

Re: [Linux-cluster] unable to live migrate a vm in rh el 6: Migration unexpectedly failed

2011-03-07 Thread Gianluca Cecchi
On Mon, 7 Mar 2011 16:52:00 -0500 Lon Hohberger wrote: > Check /var/log/audit/audit.log for an AVC denial around self:capability > setpcap for xm_t? Uhm, SElinux is disabled on both nodes (I'll cross check tomorrow anyway) and auditd is chkconfig off too (even if I notice in rh el 6 many audit me

Re: [Linux-cluster] Info on vm definitions and options in stable3

2011-03-05 Thread Gianluca Cecchi
On Fri, 4 Mar 2011 13:01:20 -0500 Lon Hohberger wrote: > http://sources.redhat.com/cluster/wiki/ServiceOperationalBehaviors > http://sources.redhat.com/cluster/wiki/ServicePolicies > http://sources.redhat.com/cluster/wiki/FailoverDomains Thanks for the links Some comments: 1) http://sources.redhat

[Linux-cluster] unable to live migrate a vm in rh el 6: Migration unexpectedly failed

2011-03-05 Thread Gianluca Cecchi
I have two rh el 6 systems configured with rhcs and clvmd. General cluster services seems to be ok. As I'm not able to successfully migrate a vm through clusvcadm, I'm now downsizing the problem to direct virsh command that fails when called from clusvcadm. The guest's storage is composed by two di

[Linux-cluster] Info on vm definitions and options in stable3

2011-03-03 Thread Gianluca Cecchi
Hello, in stable 3 I can have this kind of config for a KVM virtual machine to manage live migration: It works ok, but I would like to know the possible parameters I can set. At http://sources.redhat.com/cluster/wiki/VirtualMachineBehaviors I can see this piece "..Most of the behaviors are

Re: [Linux-cluster] ha-lvm

2010-11-09 Thread Gianluca Cecchi
On Mon, 8 Nov 2010 11:50:36 -0300 Marco Andres Dominguez wrote: > The tag could be in the vg or in the lv depending on the configurations, I > usually have it in the lv so try this: > > # lvs -o vg_name,lv_name,lv_tags > I hope it helps. > Regards. > Marco Thanks, Marco. Indeed with the lvs conma

Re: [Linux-cluster] ha-lvm

2010-11-08 Thread Gianluca Cecchi
On Wed, 3 Nov 2010 11:55:12 + Corey Kovacs wrote: > John, [snip] > "vgs -o vg_name,vg_tags" > are a welcome addition to my tool belt, thanks for that. On 2 rh el 5.5 clusters I manage, with slightly different level updates, and where I have HA-LVM configured, I don't get anything in vg_tags co

Re: [Linux-cluster] Indenting resources inside a service in rh el 4.8

2010-10-12 Thread Gianluca Cecchi
On Tue, 12 Oct 2010 15:12:46 +0100, Gordan Bobic wrote: > Are you sure you mean indenting rather than nesting? yes, sorry, I meant nesting as my example's contents outlined -- Linux-cluster mailing list Linux-cluster@redhat.com https://www.redhat.com/mailman/listinfo/linux-cluster

Re: [Linux-cluster] Indenting resources inside a service in rh el 4.8

2010-10-12 Thread Gianluca Cecchi
write something like this: But this seems not to work in rh el 4. Is this supposed to work? Thanks Gainluca On Tue, Oct 12, 2010 at 3:40 PM, Gianluca

[Linux-cluster] Indenting resources inside a service in rh el 4.8

2010-10-12 Thread Gianluca Cecchi
Hello, in RH EL 5.4 (and 5.5) it is possible to indent resources inside a service definition (service = test_service), so that, having this: Where

[Linux-cluster] rhcs upgrade from rh el 5.4 to rh el 5.5 and alb bonding on production lan

2010-07-22 Thread Gianluca Cecchi
Hello, target is upgrade from rh el 5.4 to 5.5 in a two nodes cluster with quorum disk. Part of the upgrade is also replacement of the nodes themselves. Current nodes are using alb bonding for the production lan and the driver for the underlying eth adapters is tg3 Target nodes I'm testing are ag

[Linux-cluster] question about stop service and then shutdown node

2010-07-22 Thread Gianluca Cecchi
Hello, based on cluster version 2 architecture, and in particular on a two nodes cluster with quorum disk and rh el 5.5 + updates operating systems, is this below the expected behavior? baseline service X set in autostart=1 node A + node B up and running with service X started on node A first sce

Re: [Linux-cluster] Possible bug in rhel5 for nested HA-LVM resources?

2010-03-05 Thread Gianluca Cecchi
tting status Eventually I can send my cluster.conf.old and cluster.conf.new is needed for further debug. Cheers and thanks again for support, Gianluca On Wed, Mar 3, 2010 at 11:42 PM, Gianluca Cecchi wrote: > On Wed, 03 Mar 2010 16:53:49 -0500, Lon Hohberger wrote: > > As it happens,

Re: [Linux-cluster] Possible bug in rhel5 for nested HA-LVM resources?

2010-03-03 Thread Gianluca Cecchi
On Wed, 03 Mar 2010 16:53:49 -0500, Lon Hohberger wrote: > As it happens, the 'fs' file system type looks for child 'fs' resources: > > > > ... but it does not have an entry for 'lvm', which would be required to > make it work in the order you specified. With this argument I understand ex

[Linux-cluster] Possible bug in rhel5 for nested HA-LVM resources?

2010-03-03 Thread Gianluca Cecchi
Hello, my problem begins from this need: - having a rh el 5.4 cluster with 2 nodes where I have HA-LVM in place and some lvm/fs pairs resources componing one service I want to add a new lvm/fs to the cluster, without disrupting the running service. My already configured and running lvm/mountpoints

Re: [Linux-cluster] suggestion on freeze-on-node1 and unfreeze-on-node2 approach?

2010-01-08 Thread Gianluca Cecchi
On Fri, 08 Jan 2010 09:06:57 -0500 Lon Hohberger wrote: >You could set 'recovery="relocate"', freeze the service, stop the > database cleanly, then unfreeze the service. Ah, thanks, it should work. The only "limit" would be that any recovery action will imply relocation, correct? (Some problems he

[Linux-cluster] suggestion on freeze-on-node1 and unfreeze-on-node2 approach?

2010-01-08 Thread Gianluca Cecchi
Hello, I have a cluster with an Oracle service and rhel 5.4 nodes. Tipically one sets the "shutdown abort" of the DB as the default mechanism to close the service, to prevent stalling and accelerate switch of service itself in case of problems. The same approach is indeed used by the rhcs provided

[Linux-cluster] actions to be taken when changing fence devices ip address

2009-12-16 Thread Gianluca Cecchi
Hello, I'm using RHEL 5.4 based cluster. I'm using fence_ilo fence device and I'm going to change ip address for the iLO of one node of the cluster. Is this action supposed to be made near-online, in the sense that I have not to shutdown all the cluster nodes? Idea would be: 1) services remains on

Re: [Linux-cluster] share experience migrating cluster suite from centos 5.3 to centos 5.4

2009-11-04 Thread Gianluca Cecchi
On Wed, 4 Nov 2009 15:33:19 +1000 Peter Tiggerdine wrote: > 7. Your going to need to copy this over manually otherwise it > will fail, I've fallen victim of this before. All cluster nodes need to start > on > the current revision of the file before you update it. I think this is a > chicken > and

Re: [Linux-cluster] share experience migrating cluster suite from centos 5.3 to centos 5.4

2009-11-02 Thread Gianluca Cecchi
On Mon, Nov 2, 2009 at 6:25 PM, David Teigland wrote: > > The out-of-memory should be fixed in 5.4: > > https://bugzilla.redhat.com/show_bug.cgi?id=508829 > > The fix for dlm_send spinning is not released yet: > > https://bugzilla.redhat.com/show_bug.cgi?id=521093 > > Dave > > Thank you so much f

[Linux-cluster] share experience migrating cluster suite from centos 5.3 to centos 5.4

2009-11-02 Thread Gianluca Cecchi
Hello, sorry for the long e-mail in advance. trying to do on a test environment what in subject and I think it could be useful for others too, both in RH EL and in CentOS. I have configured two ip+fs services and HA-LVM Starting point is CentOS 5.3 updated at these components: cman-2.0.98-1.el5_3.

Re: Updating cluster.conf (was Re: [Linux-cluster] ccs_config_validate in cluster 3.0.X)

2009-10-28 Thread Gianluca Cecchi
On Wed, 28 Oct 2009 15:27:45 +0100 Fabio M. Di Nitto wrote: > File a bugzilla against them, with setup and so on. > > Fabio Ok For cman error messages: https://bugzilla.redhat.com/show_bug.cgi?id=531489 For modclusterd (posted against cman because it seems modcluster doesn't exist as a component

Re: [Linux-cluster] ccs_config_validate in cluster 3.0.X

2009-10-28 Thread Gianluca Cecchi
On Wed, Oct 28, 2009 at 1:00 PM, Gianluca Cecchi wrote: > [snip] Hello, > updated my F11 today from cman-3.0.3-1.fc11.x86_64 to > cman-3.0.4-1.fc11.x86_64 > > I noticed the messages you referred. See the attached image. > Oops, here is the message image... <>-- Linux-cl

Re: [Linux-cluster] ccs_config_validate in cluster 3.0.X

2009-10-28 Thread Gianluca Cecchi
On Wed, 28 Oct 2009 11:36:30 +0100 Fabio M. Di Nitto wrote: > Hi everybody, > > as briefly mentioned in 3.0.4 release note, a new system to validate the > configuration has been enabled in the code. Hello, updated my F11 today from cman-3.0.3-1.fc11.x86_64 to cman-3.0.4-1.fc11.x86_64 I noticed

[Linux-cluster] Re: f11 3.0.3 cluster not compatible with kernel 2.6.30?

2009-10-21 Thread Gianluca Cecchi
On Wed, Oct 21, 2009 at 3:05 PM, Gianluca Cecchi wrote: > Just updated my F11 x86_64 and now cman doesn't work anymore. > Previous version cman-3.0.2-1.fc11.x86_64, updated with > cman-3.0.3-1.fc11.x86_64 > > But now I get in cman start: > [r...@r]# service cman s

[Linux-cluster] f11 3.0.3 cluster not compatible with kernel 2.6.30?

2009-10-21 Thread Gianluca Cecchi
Just updated my F11 x86_64 and now cman doesn't work anymore. Previous version cman-3.0.2-1.fc11.x86_64, updated with cman-3.0.3-1.fc11.x86_64 But now I get in cman start: [r...@r]# service cman start Starting cluster: Global setup... [ OK ] Loading

[Linux-cluster] Re: Suggestion for backbone network maintenance

2009-10-08 Thread Gianluca Cecchi
On Wed, Oct 7, 2009 at 5:03 PM, Gianluca Cecchi wrote: > Hello, > cluster rh el 5.3 with 2 nodes and a quorum disk with heuristics. The nodes > are in different sites. > At this moment inside cluster.conf I have this: > > log_facility="local4" log

[Linux-cluster] Suggestion for backbone network maintenance

2009-10-07 Thread Gianluca Cecchi
Hello, cluster rh el 5.3 with 2 nodes and a quorum disk with heuristics. The nodes are in different sites. At this moment inside cluster.conf I have this: there is a planning for backbone network maintenance and I'm gong to have interruption on backbone switches

[Linux-cluster] info on "A processor failed" message and fencing when going to single user mode

2009-10-05 Thread Gianluca Cecchi
Hello, 2 nodes cluster (virtfed and virtfedbis their names) with F11 x86_64 up2date as of today and without qdisk cman-3.0.2-1.fc11.x86_64 openais-1.0.1-1.fc11.x86_64 corosync-1.0.0-1.fc11.x86_64 and kernel 2.6.30.8-64.fc11.x86_64 I was in a situation where both nodes up, after virtfedbis hust re

[Linux-cluster] where exactly cluster services are stoppped during shutdown?

2009-09-10 Thread Gianluca Cecchi
Hello, suppose that I have a service srvname defined in chkconfig and I would like to insert it as a resource/service in my cluster.conf (version 3 of cluster as found in f11, but thanks for answer for version 2 as in rhel 5 if different) So my cluster.conf is something like this:

[Linux-cluster] Re: info on modclusterd in cluster 3

2009-08-05 Thread Gianluca Cecchi
On Wed, Aug 5, 2009 at 11:20 AM, Gianluca Cecchi wrote: > Hello, > I have a cluster 3 on 2 x F11 x86_64 nodes. I forgot to mention that also ricci doesn't start: [root ~]# service ricci start Starting ricci: main.cpp:228: exception: ServerSocket(port=1): socket() failed: Address

[Linux-cluster] info on modclusterd in cluster 3

2009-08-05 Thread Gianluca Cecchi
Hello, I have a cluster 3 on 2 x F11 x86_64 nodes. main components are: fence-agents-3.0.0-14.fc11.x86_64 clusterlib-3.0.0-20.fc11.x86_64 resource-agents-3.0.0-12.fc11.noarch lvm2-cluster-2.02.48-2.fc11.x86_64 modcluster-0.16.1-1.fc11.x86_64 corosynclib-1.0.0-1.fc11.x86_64 cman-3.0.0-20.fc11.x86_64

Re: [Linux-cluster] lvm2: cluster request failed: Unknown error 65538

2009-07-30 Thread Gianluca Cecchi
On Thu, 30 Jul 2009 14:15:46 +0200 Fabio M. Di Nitto wrote: > The update packages should be available today or tomorrow depending on > your local mirror. > > Fabio Hi, Is this version with the fix the one that I would pick up enabling updates-testing repo? [r...@tekkafedora ~]# yum --enablerepo up

Re: [Linux-cluster] Cluster 3.0.0 final stable release

2009-07-29 Thread Gianluca Cecchi
On Wed, 08 Jul 2009 23:10:12 +0200 Fabio M. Di Nitto wrote: > The cluster team and its community are proud to announce the 3.0.0 final > release from the STABLE3 branch. hello, where to find docs for 3.0.0 final? Thanks in advance, Gianluca -- Linux-cluster mailing list Linux-cluster@redhat.com h

[Linux-cluster] lvm2: cluster request failed: Unknown error 65538

2009-07-22 Thread Gianluca Cecchi
Hello, by mistake I previously sent this to fedora-list. I resend to the appropriate list I wanted... Excuse in advance for eventual cross-posting effects for anyone... fedora11 x86_64 with lvm2, device-mapper and related packages updated at : lvm2-2.02.48-1.fc11.x86_64 lvm2-cluster-2.02.48-1.fc11

[Linux-cluster] dynamically reconfigure qdiskd heuristic: possible?

2009-04-27 Thread Gianluca Cecchi
Is there any command for this? I'm on rh el 5.3 For example if I want change the line so that it becomes It seems that with 1) ccs_tool update /etc/cluster/cluster.conf 2) cman_tool version -r new_vers_number it doesn't work dynamically thanks, Gianluca -- Linux-cluster mailing list Lin

[Linux-cluster] Re: Info on lvm setup for cluster without clvmd

2009-04-24 Thread Gianluca Cecchi
e for all the other tests inside the script.. You can test this using the touch command on lvm.conf for example and trying a relocation of a service to that node. Then if you touch any .img file inside /boot directory, you are able to relocate again umh... Thanks anyway for the original pointer.

[Linux-cluster] expected behaviour when qdisk heuristic fails for both nodes

2009-04-24 Thread Gianluca Cecchi
I'm testing a situation where the gateway (192.168.1.1) fails for some time (a minute or so) for both nodes of a two-node cluster with qdisk, and qdiskd heuristic is set up with this gateway as the ping device. fencing is provided by iLO cluster version is what provided with rhel 5.3 and with the c

[Linux-cluster] Info on lvm setup for cluster without clvmd

2009-04-08 Thread Gianluca Cecchi
Hello, I would like to setup a two-node cluster where I will have some services relying on filesystems on lvm resources. I'm using rh el 5U3 but I only have entitlements for RHEL Clustering and not for Cluster-Storage, so that I cannot use clvmd as in other clusters I set up previously. I don't nee

Re: [Linux-cluster] Can same cluster name in same subnet?

2009-04-01 Thread Gianluca Cecchi
Conversely, how is it dangerous to have two two-node-clusters with different names sharing the intra-cluster network? In particular if one is in production and the other is for testing? And what about relative multicast-adresses for these two clusters? Can I safely use same multicast if the names a

Re: [Linux-cluster] ccsd problems after update to RHEL

2009-03-12 Thread Gianluca Cecchi
On Thu, 12 Mar 2009 11:04:16 + Gordan Bobic wrote: >>Marc Grimme wrote: >> Hi Gordan, >> more information can be found in this bug (at least you can get the >> information how far this bug goes back). >> https://bugzilla.redhat.com/show_bug.cgi?id=485026 > Yeah, I saw that bug entry. The l

re: [Linux-cluster] Adding a new fence agent

2009-03-06 Thread Gianluca Cecchi
I have a c7000 too with two test blades I'm going to install. I'm available to test it if you like. My planned OS will be RedHat EL 5 U3 x86_64 with its clustersuite Blades will be 2 x BL685c G1 serving Oracle 10gR2 At this moment the fw version of the c7000 is 2.25, while iLo fw is 1.60 One que

Re: [Linux-cluster] fencing loop in a 2-node partitioned cluster

2009-02-27 Thread Gianluca Cecchi
unfortunately, doing this seems to have a problematic side effect. Set up -f 1 on one node and -f 10 on the other. Now if I panic one node, it is fenced by the other one, but when restarting it remains in start fencing till it forms after some minute an own cluster and kills cman on the other n

Re: [Linux-cluster] fencing loop in a 2-node partitioned cluster

2009-02-24 Thread Gianluca Cecchi
thanks, but where do I have to put the timeout? Inside fence seciotn of the nodes: or inside definition of fence devices:

Re: [Linux-cluster] fencing loop in a 2-node partitioned cluster

2009-02-24 Thread Gianluca Cecchi
And these are the logs I see on the wo nodes: the first node: Feb 23 16:26:38 oracs1 openais[6020]: [TOTEM] The token was lost in the OPERATIONAL state. Feb 23 16:26:38 oracs1 openais[6020]: [TOTEM] Receive multicast socket recv buffer size (288000 bytes). Feb 23 16:26:38 oracs1 openais[6020]: [TOT

Re: [Linux-cluster] fencing loop in a 2-node partitioned cluster

2009-02-24 Thread Gianluca Cecchi
Actually my situation is pretty different and worse. two nodes cluster with qdisk and hp ilo based fencing, components rh el 5U3 based. if I panic a node, the other correctly fence it with default action of rebooting it. And also the converse is true. But if for example I get down the intracluster

re: [Linux-cluster] rgmanager inconsistency

2009-02-11 Thread Gianluca Cecchi
On Wed, 11 Feb 2009 10:53:45 -0600 (CST) Chris St. Pierre wrote: > Where do I start for troubleshooting this? rgmanager can log (at least in latest releases afaik) to different places and at different levels putting this inside cluster.conf log_level is a number from 0..7, where 7 is 'debug' an

Re: [Linux-cluster] cman startup after after update to 5.3

2009-02-10 Thread Gianluca Cecchi
On Fri, 30 Jan 2009 14:56:53 +0100 Gunther Schlegel wrote: > Rolling back to openais-0.80.3-15.el5 worked for me as well. Hello, with the same strategy I was only able to partially solve the problem. Two nodes with this in cluster.conf and after updating to U3

Re: [Linux-cluster] Mirrored Clustered File Systems

2009-02-08 Thread Gianluca Cecchi
On Sun, 8 Feb 2009 10:30:57 -0800 (PST) Dave Craska wrote: > Does Red Hat support mirrored file system in a cluster environment? It should be one of the news in rh el 5.3. See at http://www.redhat.com/docs/en-US/Red_Hat_Enterprise_Linux/5/html/Release_Notes/index.html : LVM-based Cluster mirrorin

Re: [Linux-cluster] interdependency between different services possible?

2009-02-06 Thread Gianluca Cecchi
On Thu, 5 Feb 2009 15:15:57 + jumanjiman wrote: > A guiding principle for cluser suite is that an ha service (resource group) > should be able to fail over w/o impacting any other RG. Ok, it does make sense. I would like to dig to dependencies a little more, to understand better the degree of

[Linux-cluster] interdependency between different services possible?

2009-02-05 Thread Gianluca Cecchi
Can I have parent-child relations between different services of a cluster? Can I have placement policies relations between different services? It seems this is not covered inside the manuals, at least for rh el 5.2 An example could be if I have: service 1 with several fs resources on it and a virt

Re: [Linux-cluster] Moving clvm filesystem to lvm

2009-02-05 Thread Gianluca Cecchi
On Wed, 04 Feb 2009 10:36:14 -0500 Danny Wall wrote: > Will I have a problem since the filesystem was > created for a cluster using clvmd, and I want to move it to LVM2 on a > RHEL5 standalone server? Take in mind that on the standalone node, if it has not all the clustersuite infrastructure (lvm