Hello,
on a two node cluster with CentOS 6.3 I set this in cluster.conf for testing
Then I cut the intra cluster lan and only when the configured totem token
ti
On Tue, Nov 19, 2013 at 11:06 AM, Gianluca Cecchi wrote:
> Is this expected behaviour? Any way to configure to try to restart in
> place the resource without manual intervention when a resource is
> configured as non critical?
>
> Thanks in advance,
> Gianluca
BTW: I did read
Hello,
I have a cluster with RH EL 6.3
cman-3.0.12.1-32.el6_3.2.x86_64
rgmanager-3.0.12.1-12.el6.x86_64
I configure ssh in cluster changing the default init script.
Then I configure it as a non critical resource in a service section
...
...
. . .
On Fri, Jul 5, 2013 at 2:42 AM, Ryan Mitchell wrote:
> You aren't starting rgmanager with the -N option are you? It is not the
> default.
> # man clurgmgrd
>-N Do not perform stop-before-start. Combined with the -Z
> flag to clusvcadm, this can be used to allow rgmanager to be up
Hello,
I already read these technotes so that it seems my configuration is
coherent with them:
https://access.redhat.com/site/documentation/en-US/Red_Hat_Enterprise_Linux/6/html/Cluster_Administration/ap-ha-halvm-CA.html
https://access.redhat.com/site/solutions/409813
basically I would like to us
On Mon, 10 Dec 2012 23:36:51 +1100 yu song wrote:
> ok..firstly confirm whether you are using HA-LVM, not GFS/GFS2 .. correct?
> if ha-lvm, your vg should only be imported/mounted on one node, instead of
> multiple nodes.
In my non-CLVMD configuration I use HA-LVM with these entries in cluster.
On Mon, 10 Dec 2012 21:02:19 +1100 Yu wrote:
> vgchange -cn vgname
> Can disable cluster flag
> Regards
> Yu
Ok, thanks, I thought about this option too and it seems it works.
More testing in place.
In this cluster I have many services and each one is bundled with its
own VGs, so that each node is
Hello, I was evaluating migration from a rhel 5.8 two node cluster with
services based on ha-lvm to a clvmd one.
I'm using netapp storage with snapshot technology.
I quiesce db, then make a snapshot and then a third server connects to the
flashcopy volume, with "snapdrive connect" command and makes
Hello,
in qdiskd.log I get at cluster startup the node that becomes master
for quorum disk.
config is in fact something like
and in syslog.conf
# qdisk logging
local4.*/var/log/qdiskd.log
The file is rotated so after some time I have only empty qd
further debug.
While the node starts and clvmd hangs
mount -t debugfs debug /sys/kernel/debug
# dlm_tool lockdump clvmd
id 024d0001 gr CR rq IV pid 3155 master 0
"7w53faJiHU9ZYs0Kw8CmFcdsSXVqujthdh8uMexWiIBPmEecuGc42XgLwwm2VcXU"
id 01710001 gr CR rq IV pid 3155 master 0
"7w53faJiHU9ZYs0Kw8CmFcds
On Thu, 2 Aug 2012 16:12:24 +0200 emmanuel segura wrote:
> can you show me your lvm.conf?
Here it is.
Gianluca
lvm.conf
Description: Binary data
--
Linux-cluster mailing list
Linux-cluster@redhat.com
https://www.redhat.com/mailman/listinfo/linux-cluster
On Thu, 2 Aug 2012 07:07:25 -0600 Corey Kovacs wrte:
> I might be reading this wrong but just in case, I thought I'd point this out.
>
[snip]
> A single node can maintain quorum since 2+3>(9/2).
> In a split brain condition where a single node cannot talk to the other
> nodes, this could be disast
On Thu, 02 Aug 2012 09:39:34 +0200 Heiko Nardmann wrote:
> If that is a real production system and not just for playing you should setup
> a test environment before and also create a plan which usecases should run
> with the new cluster.
+1 for sure for what Heiko recommended
Plus, first place:
On Wed, Aug 1, 2012 at 6:15 PM, Gianluca Cecchi wrote:
> On Wed, 1 Aug 2012 16:26:38 +0200 emmanuel segura wrote:
>> Why you don't remove expected_votes=3 and let the cluster automatic
>> calculate that
>
> Thanks for your answer Emmanuel, but cman starts correctly,
On Wed, 1 Aug 2012 16:26:38 +0200 emmanuel segura wrote:
> Why you don't remove expected_votes=3 and let the cluster automatic calculate
> that
Thanks for your answer Emmanuel, but cman starts correctly, while the
problem seems related to
vgchange -aly
command hanging.
But I tried that option too
Hello,
testing a three node cluster + quorum disk and clvmd.
I was at CentOS 6.2 and I seem to remember to be able to start a
single node. Correct?
Then I upgraded to CentOS 6.3 and had a working environment.
My config has
At the moment two nodes are in another site that is powered down and I
nee
On Fri, 9 Mar 2012 17:29:06 +0100 emmanuel segura wrote:
> i'll try to be more clear
> i work on redhat cluster from 2 years and i seen this topic so much times
Sorry, I didn't want to offend anyone.
I have been working on rhcs (and other companions from other OSes) for
many years too...
> I thin
On Fri, 9 Mar 2012 15:39:43 +0100, emmanuel segura wrote:
> Hello Gianluca
> Do you have a cluster private network?
> if your answer it's yes i recommend don't use heuristic because if your
> cluster public network goes down
> your cluster take a fencing loop
>
> Or you can do something better, u
Hello,
I have a cluster in RH EL 5.7 with quorum disk and an heuristic.
Current versions of main cluster packages are:
rgmanager-2.0.52-21.el5_7.1
cman-2.0.115-85.el5_7.3
This is the loaded heuristic
Heuristic: 'ping -c1 -w1 10.4.5.250' score=1 interval=2 tko=200
Line in cluster.conf:
where 10
On Fri, Jun 17, 2011 at 9:28 AM, Fabio M. Di Nitto wrote:
> Lon, what's your opinion on this one?
Some other considerations of mine.
This of the current "abort" default option (as in RH EL 5 cluster
suite base) is indeed a difficulty, in case of planned maintenance, so
that a change inside the ag
On Thu, Jun 16, 2011 at 3:13 PM, Fabio M. Di Nitto wrote:
> Highlights for the rgmanager resource agents set:
>
> - oracledb: use shutdown immediate
hello,
from oracledb.sh.in I can see this actually is not a configurable
parameter, so that I cannot choose between "immediate" and "abort",
and I
On Thu, 14 Apr 2011 12:44:39 +0200 carlopmart wrote:
> How can I configure cluster.conf file to assign eth1 interface when vm live
> migration will be required??
Already replied on rhelv6 list.
Next time please don't cross post.
Send to one list and eventually only after some time choose another
Hello,
I have 2 x rh el 6.0 hosts (rhev1 and rhev2) where I enabled ha and
resilient storage beta channels.
I'm testing from the Beta HA Addon channel the checkpoint backend
This necessary because I want to test managing clusters of rh el 5
guests (where for example I would keep one guest restricte
On Fri, 25 Mar 2011 07:19:40 +0530, Rajagopal Swaminathan wrote:
> On which CPU should service guard run on?
ServiceGuard is an HA solution with similar targets as RHCS or other
solutions, older or newer ones (aka Kimberlite, LifeKeeper, heartbeat
2.x, Pacemaker/Corosync, Pacemaker/Heartbeat and
On Thu, 24 Mar 2011 22:33:00 +0530, Rajagopal Swaminathan wrote:
> On 3/24/11, Gianluca Cecchi wrote:
> > such as need for direct access to server console or cluster without any
> > node running at
> > all without manual intervention ...
> >
>
> Pardon my ignora
On Tue, 22 Mar 2011 11:02:09 -0500 Robert Hayden wrote:
> I believe you will want to investigate the "clean_start" property in the
> fence_daemon stanza (RHEL 5).
> Unsure if it is in RHEL6/Cluster3 code. It is my understanding that the
> property can be used to
> by-pass the timeout and remote
On Tue, 22 Mar 2011 11:47:58 +0100, Fabio M. Di Nitto wrote:
> For RHEL related questions you should always file a ticket with GSS.
yes, it is my usual behaviour, but tipically I prefer to analyze in
advance and know if a problem I'm encountering is a bug or only my
fault in docs understanding...
Hello,
I'm using latest updates on a 2 nodes rhel 6 based cluster.
At the moment no quorum disk defined, so this line inside cluster.conf
# rpm -q cman rgmanager fence-agents ricci corosync
cman-3.0.12-23.el6_0.6.x86_64
rgmanager-3.0.12-10.el6.x86_64
fence-agents-3.0.12-8.el6_0.3.x86_64
ricci-0.1
On Wed, Mar 9, 2011 at 9:47 AM, Gianluca Cecchi
wrote:
[snip]
> Or something related with firewall perhaps.
> Can I stop firewall at all and have libvirtd working at the same time
> to test ...?
> I know libvirtd puts some iptables rules itself..
>
> Gianluca
>
OK. It
Here is the output of the command
strace -f virsh migrate --live exorapr1 qemu+ssh://intrarhev1/system
Note that if I run the same with rhev1 (main host name and not
intracluster) instead of intrarhev1, I'm asked for the ssh password
(ok because I set ssh equivalence only for intracluster) but at
On Mon, Mar 7, 2011 at 11:10 PM, Gianluca Cecchi
wrote:
> Nothing comes to mind; in my RHEL6 development cluster, I have a
> custom SELinux policy:
I confirm that SElinux is disabled and
[root@rhev1 ~]# chkconfig --list | grep audit
auditd 0:off 1:off 2:off 3:off 4:off
On Mon, 7 Mar 2011 16:52:00 -0500 Lon Hohberger wrote:
> Check /var/log/audit/audit.log for an AVC denial around self:capability
> setpcap for xm_t?
Uhm,
SElinux is disabled on both nodes (I'll cross check tomorrow anyway)
and auditd is chkconfig off too (even if I notice in rh el 6 many
audit me
On Fri, 4 Mar 2011 13:01:20 -0500 Lon Hohberger wrote:
> http://sources.redhat.com/cluster/wiki/ServiceOperationalBehaviors
> http://sources.redhat.com/cluster/wiki/ServicePolicies
> http://sources.redhat.com/cluster/wiki/FailoverDomains
Thanks for the links
Some comments:
1) http://sources.redhat
I have two rh el 6 systems configured with rhcs and clvmd.
General cluster services seems to be ok.
As I'm not able to successfully migrate a vm through clusvcadm, I'm
now downsizing the problem to direct virsh command that fails when
called from clusvcadm.
The guest's storage is composed by two di
Hello,
in stable 3 I can have this kind of config for a KVM virtual machine
to manage live migration:
It works ok, but I would like to know the possible parameters I can set.
At http://sources.redhat.com/cluster/wiki/VirtualMachineBehaviors I
can see this piece
"..Most of the behaviors are
On Mon, 8 Nov 2010 11:50:36 -0300 Marco Andres Dominguez wrote:
> The tag could be in the vg or in the lv depending on the configurations, I
> usually have it in the lv so try this:
>
> # lvs -o vg_name,lv_name,lv_tags
> I hope it helps.
> Regards.
> Marco
Thanks, Marco.
Indeed with the lvs conma
On Wed, 3 Nov 2010 11:55:12 + Corey Kovacs wrote:
> John,
[snip]
> "vgs -o vg_name,vg_tags"
> are a welcome addition to my tool belt, thanks for that.
On 2 rh el 5.5 clusters I manage, with slightly different level
updates, and where I have HA-LVM configured, I don't get anything in
vg_tags co
On Tue, 12 Oct 2010 15:12:46 +0100, Gordan Bobic wrote:
> Are you sure you mean indenting rather than nesting?
yes, sorry, I meant nesting as my example's contents outlined
--
Linux-cluster mailing list
Linux-cluster@redhat.com
https://www.redhat.com/mailman/listinfo/linux-cluster
write something like
this:
But this seems not to work in rh el 4. Is this supposed to work?
Thanks
Gainluca
On Tue, Oct 12, 2010 at 3:40 PM, Gianluca
Hello,
in RH EL 5.4 (and 5.5) it is possible to indent resources inside a
service definition (service = test_service), so that, having this:
Where
Hello,
target is upgrade from rh el 5.4 to 5.5 in a two nodes cluster with quorum
disk.
Part of the upgrade is also replacement of the nodes themselves.
Current nodes are using alb bonding for the production lan and the driver
for the underlying eth adapters is tg3
Target nodes I'm testing are ag
Hello,
based on cluster version 2 architecture, and in particular on a two nodes
cluster with quorum disk and rh el 5.5 + updates operating systems, is this
below the expected behavior?
baseline
service X set in autostart=1
node A + node B up and running with service X started on node A
first sce
tting status
Eventually I can send my cluster.conf.old and cluster.conf.new is needed
for further debug.
Cheers and thanks again for support,
Gianluca
On Wed, Mar 3, 2010 at 11:42 PM, Gianluca Cecchi
wrote:
> On Wed, 03 Mar 2010 16:53:49 -0500, Lon Hohberger wrote:
> > As it happens,
On Wed, 03 Mar 2010 16:53:49 -0500, Lon Hohberger wrote:
> As it happens, the 'fs' file system type looks for child 'fs' resources:
>
>
>
> ... but it does not have an entry for 'lvm', which would be required to
> make it work in the order you specified.
With this argument I understand ex
Hello,
my problem begins from this need:
- having a rh el 5.4 cluster with 2 nodes where I have HA-LVM in place and
some lvm/fs pairs resources componing one service
I want to add a new lvm/fs to the cluster, without disrupting the running
service.
My already configured and running lvm/mountpoints
On Fri, 08 Jan 2010 09:06:57 -0500 Lon Hohberger wrote:
>You could set 'recovery="relocate"', freeze the service, stop the
> database cleanly, then unfreeze the service.
Ah, thanks, it should work.
The only "limit" would be that any recovery action will imply
relocation, correct?
(Some problems he
Hello, I have a cluster with an Oracle service and rhel 5.4 nodes.
Tipically one sets the "shutdown abort" of the DB as the default
mechanism to close the service, to prevent stalling and accelerate
switch of service itself in case of problems.
The same approach is indeed used by the rhcs provided
Hello,
I'm using RHEL 5.4 based cluster.
I'm using fence_ilo fence device and I'm going to change ip address
for the iLO of one node of the cluster.
Is this action supposed to be made near-online, in the sense that I
have not to shutdown all the cluster nodes?
Idea would be:
1) services remains on
On Wed, 4 Nov 2009 15:33:19 +1000 Peter Tiggerdine wrote:
> 7. Your going to need to copy this over manually otherwise it
> will fail, I've fallen victim of this before. All cluster nodes need to start
> on
> the current revision of the file before you update it. I think this is a
> chicken
> and
On Mon, Nov 2, 2009 at 6:25 PM, David Teigland wrote:
>
> The out-of-memory should be fixed in 5.4:
>
> https://bugzilla.redhat.com/show_bug.cgi?id=508829
>
> The fix for dlm_send spinning is not released yet:
>
> https://bugzilla.redhat.com/show_bug.cgi?id=521093
>
> Dave
>
>
Thank you so much f
Hello,
sorry for the long e-mail in advance.
trying to do on a test environment what in subject and I think it could be
useful for others too, both in RH EL and in CentOS.
I have configured two ip+fs services and HA-LVM
Starting point is CentOS 5.3 updated at these components:
cman-2.0.98-1.el5_3.
On Wed, 28 Oct 2009 15:27:45 +0100 Fabio M. Di Nitto wrote:
> File a bugzilla against them, with setup and so on.
>
> Fabio
Ok
For cman error messages:
https://bugzilla.redhat.com/show_bug.cgi?id=531489
For modclusterd (posted against cman because it seems modcluster doesn't
exist as a component
On Wed, Oct 28, 2009 at 1:00 PM, Gianluca Cecchi
wrote:
> [snip]
Hello,
> updated my F11 today from cman-3.0.3-1.fc11.x86_64 to
> cman-3.0.4-1.fc11.x86_64
>
> I noticed the messages you referred. See the attached image.
>
Oops, here is the message image...
<>--
Linux-cl
On Wed, 28 Oct 2009 11:36:30 +0100 Fabio M. Di Nitto wrote:
> Hi everybody,
>
> as briefly mentioned in 3.0.4 release note, a new system to validate the
> configuration has been enabled in the code.
Hello,
updated my F11 today from cman-3.0.3-1.fc11.x86_64 to
cman-3.0.4-1.fc11.x86_64
I noticed
On Wed, Oct 21, 2009 at 3:05 PM, Gianluca Cecchi
wrote:
> Just updated my F11 x86_64 and now cman doesn't work anymore.
> Previous version cman-3.0.2-1.fc11.x86_64, updated with
> cman-3.0.3-1.fc11.x86_64
>
> But now I get in cman start:
> [r...@r]# service cman s
Just updated my F11 x86_64 and now cman doesn't work anymore.
Previous version cman-3.0.2-1.fc11.x86_64, updated with
cman-3.0.3-1.fc11.x86_64
But now I get in cman start:
[r...@r]# service cman start
Starting cluster:
Global setup... [ OK ]
Loading
On Wed, Oct 7, 2009 at 5:03 PM, Gianluca Cecchi
wrote:
> Hello,
> cluster rh el 5.3 with 2 nodes and a quorum disk with heuristics. The nodes
> are in different sites.
> At this moment inside cluster.conf I have this:
>
> log_facility="local4" log
Hello,
cluster rh el 5.3 with 2 nodes and a quorum disk with heuristics. The nodes
are in different sites.
At this moment inside cluster.conf I have this:
there is a planning for backbone network maintenance and I'm gong to have
interruption on backbone switches
Hello,
2 nodes cluster (virtfed and virtfedbis their names) with F11 x86_64
up2date as of today and without qdisk
cman-3.0.2-1.fc11.x86_64
openais-1.0.1-1.fc11.x86_64
corosync-1.0.0-1.fc11.x86_64
and kernel 2.6.30.8-64.fc11.x86_64
I was in a situation where both nodes up, after virtfedbis hust re
Hello,
suppose that I have a service srvname defined in chkconfig and I would like
to insert it as a resource/service in my cluster.conf
(version 3 of cluster as found in f11, but thanks for answer for version 2
as in rhel 5 if different)
So my cluster.conf is something like this:
On Wed, Aug 5, 2009 at 11:20 AM, Gianluca
Cecchi wrote:
> Hello,
> I have a cluster 3 on 2 x F11 x86_64 nodes.
I forgot to mention that also ricci doesn't start:
[root ~]# service ricci start
Starting ricci: main.cpp:228: exception: ServerSocket(port=1):
socket() failed: Address
Hello,
I have a cluster 3 on 2 x F11 x86_64 nodes.
main components are:
fence-agents-3.0.0-14.fc11.x86_64
clusterlib-3.0.0-20.fc11.x86_64
resource-agents-3.0.0-12.fc11.noarch
lvm2-cluster-2.02.48-2.fc11.x86_64
modcluster-0.16.1-1.fc11.x86_64
corosynclib-1.0.0-1.fc11.x86_64
cman-3.0.0-20.fc11.x86_64
On Thu, 30 Jul 2009 14:15:46 +0200 Fabio M. Di Nitto wrote:
> The update packages should be available today or tomorrow depending on
> your local mirror.
>
> Fabio
Hi,
Is this version with the fix the one that I would pick up enabling
updates-testing repo?
[r...@tekkafedora ~]# yum --enablerepo up
On Wed, 08 Jul 2009 23:10:12 +0200 Fabio M. Di Nitto wrote:
> The cluster team and its community are proud to announce the 3.0.0 final
> release from the STABLE3 branch.
hello,
where to find docs for 3.0.0 final?
Thanks in advance,
Gianluca
--
Linux-cluster mailing list
Linux-cluster@redhat.com
h
Hello,
by mistake I previously sent this to fedora-list.
I resend to the appropriate list I wanted...
Excuse in advance for eventual cross-posting effects for anyone...
fedora11 x86_64 with lvm2, device-mapper and related packages updated at :
lvm2-2.02.48-1.fc11.x86_64
lvm2-cluster-2.02.48-1.fc11
Is there any command for this? I'm on rh el 5.3
For example if I want change the line
so that it becomes
It seems that with
1) ccs_tool update /etc/cluster/cluster.conf
2) cman_tool version -r new_vers_number
it doesn't work dynamically
thanks,
Gianluca
--
Linux-cluster mailing list
Lin
e for all the
other tests inside the script..
You can test this using the touch command on lvm.conf for example and
trying a relocation of a service to that node.
Then if you touch any .img file inside /boot directory, you are able
to relocate again
umh...
Thanks anyway for the original pointer.
I'm testing a situation where the gateway (192.168.1.1) fails for some
time (a minute or so) for both nodes of a two-node cluster with qdisk,
and qdiskd heuristic is set up with this gateway as the ping device.
fencing is provided by iLO
cluster version is what provided with rhel 5.3 and with the c
Hello,
I would like to setup a two-node cluster where I will have some
services relying on filesystems on lvm resources.
I'm using rh el 5U3 but I only have entitlements for RHEL Clustering
and not for Cluster-Storage, so that I cannot use clvmd as in other
clusters I set up previously.
I don't nee
Conversely, how is it dangerous to have two two-node-clusters with
different names sharing the intra-cluster network?
In particular if one is in production and the other is for testing?
And what about relative multicast-adresses for these two clusters? Can
I safely use same multicast if the names a
On Thu, 12 Mar 2009 11:04:16 + Gordan Bobic wrote:
>>Marc Grimme wrote:
>> Hi Gordan,
>> more information can be found in this bug (at least you can get the
>> information how far this bug goes back).
>> https://bugzilla.redhat.com/show_bug.cgi?id=485026
> Yeah, I saw that bug entry. The l
I have a c7000 too with two test blades I'm going to install.
I'm available to test it if you like.
My planned OS will be RedHat EL 5 U3 x86_64 with its clustersuite
Blades will be 2 x BL685c G1 serving Oracle 10gR2
At this moment the fw version of the c7000 is 2.25, while iLo fw is 1.60
One que
unfortunately, doing this seems to have a problematic side effect.
Set up -f 1 on one node and -f 10 on the other.
Now if I panic one node, it is fenced by the other one, but when
restarting it remains in
start fencing
till it forms after some minute an own cluster and kills cman on the other n
thanks, but where do I have to put the timeout?
Inside fence seciotn of the nodes:
or inside definition of fence devices:
And these are the logs I see on the wo nodes:
the first node:
Feb 23 16:26:38 oracs1 openais[6020]: [TOTEM] The token was lost in
the OPERATIONAL state.
Feb 23 16:26:38 oracs1 openais[6020]: [TOTEM] Receive multicast socket
recv buffer size (288000 bytes).
Feb 23 16:26:38 oracs1 openais[6020]: [TOT
Actually my situation is pretty different and worse.
two nodes cluster with qdisk and hp ilo based fencing, components rh
el 5U3 based.
if I panic a node, the other correctly fence it with default action of
rebooting it. And also the converse is true.
But if for example I get down the intracluster
On Wed, 11 Feb 2009 10:53:45 -0600 (CST) Chris St. Pierre wrote:
> Where do I start for troubleshooting this?
rgmanager can log (at least in latest releases afaik) to different
places and at different levels putting this inside cluster.conf
log_level is a number from 0..7, where 7 is 'debug' an
On Fri, 30 Jan 2009 14:56:53 +0100 Gunther Schlegel wrote:
> Rolling back to openais-0.80.3-15.el5 worked for me as well.
Hello,
with the same strategy I was only able to partially solve the problem.
Two nodes with this in cluster.conf
and
after updating to U3
On Sun, 8 Feb 2009 10:30:57 -0800 (PST) Dave Craska wrote:
> Does Red Hat support mirrored file system in a cluster environment?
It should be one of the news in rh el 5.3.
See at
http://www.redhat.com/docs/en-US/Red_Hat_Enterprise_Linux/5/html/Release_Notes/index.html
:
LVM-based Cluster mirrorin
On Thu, 5 Feb 2009 15:15:57 + jumanjiman wrote:
> A guiding principle for cluser suite is that an ha service (resource group)
> should be able to fail over w/o impacting any other RG.
Ok, it does make sense.
I would like to dig to dependencies a little more, to understand
better the degree of
Can I have parent-child relations between different services of a cluster?
Can I have placement policies relations between different services?
It seems this is not covered inside the manuals, at least for rh el 5.2
An example could be if I have:
service 1 with several fs resources on it and a virt
On Wed, 04 Feb 2009 10:36:14 -0500 Danny Wall wrote:
> Will I have a problem since the filesystem was
> created for a cluster using clvmd, and I want to move it to LVM2 on a
> RHEL5 standalone server?
Take in mind that on the standalone node, if it has not all the
clustersuite infrastructure (lvm
82 matches
Mail list logo