Re: [ClusterLabs] corosync doesn't start any resource

2018-06-25 Thread Stefan Krueger
Hello Andrei,

> Then you need to set symmetrical="false".
yep, now it seems to work now, thank you very much!

> I assume this would be "pcs constraint order set ...
> symmetrical=false".
yes almost:
pcs constraint order set nfs-server vm_storage ha-ip action=start setoptions 
symmetrical=false


Thank you very very much!

best regards
Stefan

> Gesendet: Samstag, 23. Juni 2018 um 22:13 Uhr
> Von: "Andrei Borzenkov" 
> An: users@clusterlabs.org
> Betreff: Re: [ClusterLabs] corosync doesn't start any resource
>
> 22.06.2018 11:22, Stefan Krueger пишет:
> > Hello Andrei,
> > 
> > thanks for this hint, but I need this "special" order. In an other setup it 
> > works.
> > 
> 
> Then you need to set symmetrical="false". Otherwise pacemaker implicitly
> creates reverse order which leads to deadlock. I am not intimately
> familiar with pcs, I assume this would be "pcs constraint order set ...
> symmetrical=false".
> 
> > best regards
> > Stefan
> > 
> >> Gesendet: Freitag, 22. Juni 2018 um 06:57 Uhr
> >> Von: "Andrei Borzenkov" 
> >> An: users@clusterlabs.org
> >> Betreff: Re: [ClusterLabs] corosync doesn't start any resource
> >>
> >> 21.06.2018 16:04, Stefan Krueger пишет:
> >>> Hi Ken,
> >>>
> >>>> Can you attach the pe-input file listed just above here?
> >>> done ;) 
> >>>
> >>> And thank you for your patience!
> >>>
> >>
> >> You delete all context which makes it hard to answer. This is not web
> >> forum where users can simply scroll up to see previous reply.
> >>
> >> Both your logs and pe-input show that nfs-server and vm-storage wait for
> >> each other.
> >>
> >> My best guess is that you have incorrect ordering for start and stop
> >> which causes loop in pacemaker decision. Your start order is "nfs-server
> >> vm-storage" and your stop order is "nfs-server vm-storage", while it
> >> should normally be symmetrical. Reversing order in one of sets makes it
> >> work as intended (verified).
> >>
> >> I would actually expect that asymmetrical configuration still should
> >> work, so I leave it to pacemaker developers to comment whether this is a
> >> bug or feature :)
> >>
> >> ___
> >> Users mailing list: Users@clusterlabs.org
> >> https://lists.clusterlabs.org/mailman/listinfo/users
> >>
> >> Project Home: http://www.clusterlabs.org
> >> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> >> Bugs: http://bugs.clusterlabs.org
> >>
> > ___
> > Users mailing list: Users@clusterlabs.org
> > https://lists.clusterlabs.org/mailman/listinfo/users
> > 
> > Project Home: http://www.clusterlabs.org
> > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> > Bugs: http://bugs.clusterlabs.org
> > 
> 
> ___
> Users mailing list: Users@clusterlabs.org
> https://lists.clusterlabs.org/mailman/listinfo/users
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>
___
Users mailing list: Users@clusterlabs.org
https://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [ClusterLabs] corosync doesn't start any resource

2018-06-22 Thread Stefan Krueger
Hello Andrei,

thanks for this hint, but I need this "special" order. In an other setup it 
works.

best regards
Stefan

> Gesendet: Freitag, 22. Juni 2018 um 06:57 Uhr
> Von: "Andrei Borzenkov" 
> An: users@clusterlabs.org
> Betreff: Re: [ClusterLabs] corosync doesn't start any resource
>
> 21.06.2018 16:04, Stefan Krueger пишет:
> > Hi Ken,
> > 
> >> Can you attach the pe-input file listed just above here?
> > done ;) 
> > 
> > And thank you for your patience!
> > 
> 
> You delete all context which makes it hard to answer. This is not web
> forum where users can simply scroll up to see previous reply.
> 
> Both your logs and pe-input show that nfs-server and vm-storage wait for
> each other.
> 
> My best guess is that you have incorrect ordering for start and stop
> which causes loop in pacemaker decision. Your start order is "nfs-server
> vm-storage" and your stop order is "nfs-server vm-storage", while it
> should normally be symmetrical. Reversing order in one of sets makes it
> work as intended (verified).
> 
> I would actually expect that asymmetrical configuration still should
> work, so I leave it to pacemaker developers to comment whether this is a
> bug or feature :)
> 
> ___
> Users mailing list: Users@clusterlabs.org
> https://lists.clusterlabs.org/mailman/listinfo/users
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org
>
___
Users mailing list: Users@clusterlabs.org
https://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [ClusterLabs] corosync doesn't start any resource

2018-06-21 Thread Stefan Krueger
Hi Ken,

> Can you attach the pe-input file listed just above here?
done ;) 

And thank you for your patience!

best regards
Stefan

pre-input-228.bz2
Description: application/bzip
___
Users mailing list: Users@clusterlabs.org
https://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [ClusterLabs] corosync doesn't start any resource

2018-06-20 Thread Stefan Krueger
Hi Ken,

I don't see any issues in the logs, periodically this is in the logs:

Jun 20 11:52:19 [5613] zfs-serv3   crmd: info: crm_timer_popped:
PEngine Recheck Timer (I_PE_CALC) just popped (90ms)
Jun 20 11:52:19 [5613] zfs-serv3   crmd:   notice: do_state_transition: 
State transition S_IDLE -> S_POLICY_ENGINE | input=I_PE_CALC 
cause=C_TIMER_POPPED origin=crm_timer_popped
Jun 20 11:52:19 [5613] zfs-serv3   crmd: info: do_state_transition: 
Progressed to state S_POLICY_ENGINE after C_TIMER_POPPED
Jun 20 11:52:19 [5612] zfs-serv3pengine: info: process_pe_message:  
Input has not changed since last time, not saving to disk
Jun 20 11:52:19 [5612] zfs-serv3pengine: info: determine_online_status: 
Node zfs-serv3 is online
Jun 20 11:52:19 [5612] zfs-serv3pengine: info: determine_online_status: 
Node zfs-serv4 is online
Jun 20 11:52:19 [5612] zfs-serv3pengine: info: native_print:
vm_storage  (ocf::heartbeat:ZFS):   Stopped
Jun 20 11:52:19 [5612] zfs-serv3pengine: info: native_print:ha-ip   
(ocf::heartbeat:IPaddr2):   Stopped
Jun 20 11:52:19 [5612] zfs-serv3pengine: info: native_print:
resIPMI-zfs4(stonith:external/ipmi):Started zfs-serv3
Jun 20 11:52:19 [5612] zfs-serv3pengine: info: native_print:
resIPMI-zfs3(stonith:external/ipmi):Started zfs-serv4
Jun 20 11:52:19 [5612] zfs-serv3pengine: info: native_print:
nfs-server  (systemd:nfs-server):   Stopped
Jun 20 11:52:19 [5612] zfs-serv3pengine: info: RecurringOp:  Start 
recurring monitor (5s) for vm_storage on zfs-serv3
Jun 20 11:52:19 [5612] zfs-serv3pengine: info: RecurringOp:  Start 
recurring monitor (10s) for ha-ip on zfs-serv3
Jun 20 11:52:19 [5612] zfs-serv3pengine: info: RecurringOp:  Start 
recurring monitor (60s) for nfs-server on zfs-serv3
Jun 20 11:52:19 [5612] zfs-serv3pengine:   notice: LogActions:  Start   
vm_storage  (zfs-serv3)
Jun 20 11:52:19 [5612] zfs-serv3pengine:   notice: LogActions:  Start   
ha-ip   (zfs-serv3)
Jun 20 11:52:19 [5612] zfs-serv3pengine: info: LogActions:  Leave   
resIPMI-zfs4(Started zfs-serv3)
Jun 20 11:52:19 [5612] zfs-serv3pengine: info: LogActions:  Leave   
resIPMI-zfs3(Started zfs-serv4)
Jun 20 11:52:19 [5612] zfs-serv3pengine:   notice: LogActions:  Start   
nfs-server  (zfs-serv3)
Jun 20 11:52:19 [5612] zfs-serv3pengine:   notice: process_pe_message:  
Calculated transition 80, saving inputs in 
/var/lib/pacemaker/pengine/pe-input-228.bz2
Jun 20 11:52:19 [5613] zfs-serv3   crmd: info: do_state_transition: 
State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE | input=I_PE_SUCCESS 
cause=C_IPC_MESSAGE origin=handle_response
Jun 20 11:52:19 [5613] zfs-serv3   crmd: info: do_te_invoke:
Processing graph 80 (ref=pe_calc-dc-1529488339-113) derived from 
/var/lib/pacemaker/pengine/pe-input-228.bz2
Jun 20 11:52:19 [5613] zfs-serv3   crmd:  warning: run_graph:   
Transition 80 (Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=6, 
Source=/var/lib/pacemaker/pengine/pe-input-228.bz2): Terminated
Jun 20 11:52:19 [5613] zfs-serv3   crmd:  warning: te_graph_trigger:
Transition failed: terminated
Jun 20 11:52:19 [5613] zfs-serv3   crmd:   notice: print_graph: Graph 
80 with 6 actions: batch-limit=0 jobs, network-delay=6ms
Jun 20 11:52:19 [5613] zfs-serv3   crmd:   notice: print_synapse:   [Action 
   5]: Pending rsc op vm_storage_monitor_5000 on zfs-serv3 
(priority: 0, waiting:  4)
Jun 20 11:52:19 [5613] zfs-serv3   crmd:   notice: print_synapse:   [Action 
   4]: Pending rsc op vm_storage_start_0  on zfs-serv3 
(priority: 0, waiting:  12)
Jun 20 11:52:19 [5613] zfs-serv3   crmd:   notice: print_synapse:   [Action 
   7]: Pending rsc op ha-ip_monitor_1 on zfs-serv3 
(priority: 0, waiting:  6)
Jun 20 11:52:19 [5613] zfs-serv3   crmd:   notice: print_synapse:   [Action 
   6]: Pending rsc op ha-ip_start_0   on zfs-serv3 
(priority: 0, waiting:  4 12)
Jun 20 11:52:19 [5613] zfs-serv3   crmd:   notice: print_synapse:   [Action 
  13]: Pending rsc op nfs-server_monitor_6on zfs-serv3 
(priority: 0, waiting:  12)
Jun 20 11:52:19 [5613] zfs-serv3   crmd:   notice: print_synapse:   [Action 
  12]: Pending rsc op nfs-server_start_0  on zfs-serv3 
(priority: 0, waiting:  4)
Jun 20 11:52:19 [5613] zfs-serv3   crmd: info: do_log:  Input 
I_TE_SUCCESS received in state S_TRANSITION_ENGINE from notify_crmd
Jun 20 11:52:19 [5613] zfs-serv3   crmd:   notice: do_state_transition: 
State transition S_TRANSITION_ENGINE -> S_IDLE | input=I_TE_SUCCESS 
cause=C_FSA_INTERNAL origin=notify_crmd
Jun 20 12:07:19 [5613] zfs-serv3   crmd: info: crm_timer_popped:
PEngine 

Re: [ClusterLabs] corosync doesn't start any resource

2018-06-19 Thread Stefan Krueger
Hi Ken,

thanks for help!
I create a stonith-device and delete the no-quorum-policy.

It doesn't change anything, so I delete the orders, (co)locations and one 
ressource (nfs-server). at first it works fine but when I stop a cluster via 
'pcs cluster stop' it takes infinity time, it looks like it has an problem with 
the nfs server so I tried to stop them manuelly via systemctl stop nfs-server, 
but it didn't change anything - the nfs-server won't stop. So I did a reset the 
server, now everything should move to the other node but it also didn't happen 
:(

Manually I can start/stop the nfs-server without any problems (nobody mount the 
nfs-share yet):
systemctl start nfs-server.service ; sleep 5; systemctl status 
nfs-server.service ; sleep 5; systemctl stop nfs-server

so, again my ressources won't start
pcs status
Cluster name: zfs-vmstorage
Stack: corosync
Current DC: zfs-serv3 (version 1.1.16-94ff4df) - partition with quorum
Last updated: Tue Jun 19 16:15:37 2018
Last change: Tue Jun 19 15:41:24 2018 by hacluster via crmd on zfs-serv4

2 nodes configured
5 resources configured

Online: [ zfs-serv3 zfs-serv4 ]

Full list of resources:

 vm_storage (ocf::heartbeat:ZFS):   Stopped
 ha-ip  (ocf::heartbeat:IPaddr2):   Stopped
 resIPMI-zfs4   (stonith:external/ipmi):Started zfs-serv3
 resIPMI-zfs3   (stonith:external/ipmi):Started zfs-serv4
 nfs-server (systemd:nfs-server):   Stopped

Daemon Status:
  corosync: active/enabled
  pacemaker: active/enabled
  pcsd: active/enabled




pcs config
Cluster Name: zfs-vmstorage
Corosync Nodes:
 zfs-serv3 zfs-serv4
Pacemaker Nodes:
 zfs-serv3 zfs-serv4

Resources:
 Resource: vm_storage (class=ocf provider=heartbeat type=ZFS)
  Attributes: pool=vm_storage importargs="-d /dev/disk/by-vdev/"
  Operations: monitor interval=5s timeout=30s (vm_storage-monitor-interval-5s)
  start interval=0s timeout=90 (vm_storage-start-interval-0s)
  stop interval=0s timeout=90 (vm_storage-stop-interval-0s)
 Resource: ha-ip (class=ocf provider=heartbeat type=IPaddr2)
  Attributes: ip=172.16.101.73 cidr_netmask=16
  Operations: start interval=0s timeout=20s (ha-ip-start-interval-0s)
  stop interval=0s timeout=20s (ha-ip-stop-interval-0s)
  monitor interval=10s timeout=20s (ha-ip-monitor-interval-10s)
 Resource: nfs-server (class=systemd type=nfs-server)
  Operations: start interval=0s timeout=100 (nfs-server-start-interval-0s)
  stop interval=0s timeout=100 (nfs-server-stop-interval-0s)
  monitor interval=60 timeout=100 (nfs-server-monitor-interval-60)

Stonith Devices:
 Resource: resIPMI-zfs4 (class=stonith type=external/ipmi)
  Attributes: hostname=ipmi-zfs-serv4 ipaddr=172.xx.xx.17 userid=USER 
passwd=GEHEIM interface=lan
  Operations: monitor interval=60s (resIPMI-zfs4-monitor-interval-60s)
 Resource: resIPMI-zfs3 (class=stonith type=external/ipmi)
  Attributes: hostname=ipmi-zfs-serv3 ipaddr=172.xx.xx.16 userid=USER 
passwd=GEHEIM interface=lan
  Operations: monitor interval=60s (resIPMI-zfs3-monitor-interval-60s)
Fencing Levels:

Location Constraints:
  Resource: resIPMI-zfs3
Disabled on: zfs-serv3 (score:-INFINITY) 
(id:location-resIPMI-zfs3-zfs-serv3--INFINITY)
  Resource: resIPMI-zfs4
Disabled on: zfs-serv4 (score:-INFINITY) 
(id:location-resIPMI-zfs4-zfs-serv4--INFINITY)
Ordering Constraints:
  Resource Sets:
set nfs-server vm_storage ha-ip action=start 
(id:pcs_rsc_set_nfs-server_vm_storage_ha-ip) 
(id:pcs_rsc_order_set_nfs-server_vm_storage_ha-ip)
set ha-ip nfs-server vm_storage action=stop 
(id:pcs_rsc_set_ha-ip_nfs-server_vm_storage) 
(id:pcs_rsc_order_set_ha-ip_nfs-server_vm_storage)
Colocation Constraints:
  Resource Sets:
set ha-ip nfs-server vm_storage (id:colocation-ha-ip-nfs-server-INFINITY-0) 
setoptions score=INFINITY (id:colocation-ha-ip-nfs-server-INFINITY)
Ticket Constraints:

Alerts:
 No alerts defined

Resources Defaults:
 resource-stickiness: 100
Operations Defaults:
 No defaults set

Cluster Properties:
 cluster-infrastructure: corosync
 cluster-name: zfs-vmstorage
 dc-version: 1.1.16-94ff4df
 have-watchdog: false
 last-lrm-refresh: 1528814481
 no-quorum-policy: stop
 stonith-enabled: false

Quorum:
  Options:



thanks for help!
best regards
Stefan
___
Users mailing list: Users@clusterlabs.org
https://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [ClusterLabs] questions about fence_scsi

2018-06-15 Thread Stefan Krueger
Hello Andrei,

thanks for this hint, at the moment I try to solve this with a colocation. (but 
it doesn't work, see mailinglist)

best regards

> Gesendet: Freitag, 15. Juni 2018 um 10:32 Uhr
> Von: "Andrei Borzenkov" 
> An: "Cluster Labs - All topics related to open-source clustering welcomed" 
> 
> Betreff: Re: [ClusterLabs] questions about fence_scsi
>
> On Fri, Jun 15, 2018 at 11:18 AM, Andrei Borzenkov  
> wrote:
> > On Fri, Jun 15, 2018 at 10:14 AM, Stefan Krueger  wrote:
> >> Hello,
> >>
> >> so far as I understand I can use fence_scsi on a two node cluster, if the 
> >> fence running on one cluster the other cluster has no access to this 
> >> devices, correct?
> >
> > If I parse this sentence correctly - no, that's not correct to my best
> > knowledge. All active nodes have access to shared resource - only when
> > node fails will it be fenced (i.e. access to devices revoked), and
> > unfenced again (i.e. access granted) when node comes back.
> >
> 
> If you really want to allow only one node at a time to access device,
> you should look at something like sg_persist RA. See example in SLES
> focumentation (scroll down to the end of page):
> 
> https://www.suse.com/documentation/sle-ha-12/book_sleha/data/sec_ha_storage_protect_fencing.html
> 
> 
> >> I've a 2node cluster with shared JBODs and configure fence_scsi, but I 
> >> still can use/mount all this devices on both nodes. Did I something wrong?
> >> My Config:
> >>
> >> pcs resource
> >>  Resource Group: zfs-storage
> >>  vm_storage (ocf::heartbeat:ZFS):   Started zfs-serv3
> >>  ha-ip  (ocf::heartbeat:IPaddr2):   Started zfs-serv3
> >> root@zfs-serv4:~# pcs config
> >> Cluster Name: zfs-vmstorage
> >> Corosync Nodes:
> >>  zfs-serv3 zfs-serv4
> >> Pacemaker Nodes:
> >>  zfs-serv3 zfs-serv4
> >>
> >> Resources:
> >>  Group: zfs-storage
> >>   Resource: vm_storage (class=ocf provider=heartbeat type=ZFS)
> >>Attributes: pool=vm_storage importargs="-d /dev/disk/by-vdev/"
> >>Operations: monitor interval=5s timeout=30s 
> >> (vm_storage-monitor-interval-5s)
> >>start interval=0s timeout=90 (vm_storage-start-interval-0s)
> >>stop interval=0s timeout=90 (vm_storage-stop-interval-0s)
> >>   Resource: ha-ip (class=ocf provider=heartbeat type=IPaddr2)
> >>Attributes: ip=172.16.101.73 cidr_netmask=16
> >>Operations: start interval=0s timeout=20s (ha-ip-start-interval-0s)
> >>stop interval=0s timeout=20s (ha-ip-stop-interval-0s)
> >>monitor interval=10s timeout=20s 
> >> (ha-ip-monitor-interval-10s)
> >>
> >> Stonith Devices:
> >>  Resource: fence-vm_storage (class=stonith type=fence_scsi)
> >>   Attributes: pcmk_monitor_action=metadata 
> >> pcmk_host_list=172.16.101.74,172.16.101.75 devices=" 
> >> /dev/disk/by-vdev/j3d03-hdd /dev/disk/by-vdev/j4d03-hdd 
> >> /dev/disk/by-vdev/j3d04-hdd /dev/disk/by-vdev/j4d04-hdd 
> >> /dev/disk/by-vdev/j3d05-hdd /dev/disk/by-vdev/j4d05-hdd 
> >> /dev/disk/by-vdev/j3d06-hdd /dev/disk/by-vdev/j4d06-hdd 
> >> /dev/disk/by-vdev/j3d07-hdd /dev/disk/by-vdev/j4d07-hdd 
> >> /dev/disk/by-vdev/j3d08-hdd /dev/disk/by-vdev/j4d08-hdd 
> >> /dev/disk/by-vdev/j3d09-hdd /dev/disk/by-vdev/j4d09-hdd 
> >> /dev/disk/by-vdev/j3d10-hdd /dev/disk/by-vdev/j4d10-hdd 
> >> /dev/disk/by-vdev/j3d11-hdd /dev/disk/by-vdev/j4d11-hdd 
> >> /dev/disk/by-vdev/j3d12-hdd /dev/disk/by-vdev/j4d12-hdd 
> >> /dev/disk/by-vdev/j3d13-hdd /dev/disk/by-vdev/j4d13-hdd 
> >> /dev/disk/by-vdev/j3d14-hdd /dev/disk/by-vdev/j4d14-hdd 
> >> /dev/disk/by-vdev/j3d15-hdd /dev/disk/by-vdev/j4d15-hdd 
> >> /dev/disk/by-vdev/j3d16-hdd /dev/disk/by-vdev/j4d16-hdd 
> >> /dev/disk/by-vdev/j3d17-hdd /dev/disk/by-vdev/j4d17-hdd 
> >> /dev/disk/by-vdev/j3d18-hdd /dev/disk/by-vdev/j4d18-hdd /
 de
>  v/d
> >>  isk/by-vdev/j3d19-hdd /dev/disk/by-vdev/j4d19-hdd log 
> >> /dev/disk/by-vdev/j3d00-ssd /dev/disk/by-vdev/j4d00-ssd cache 
> >> /dev/disk/by-vdev/j3d02-ssd"
> >>   Meta Attrs: provides=unfencing
> >>   Operations: monitor interval=60s (fence-vm_storage-monitor-interval-60s)
> >> Fencing Levels:
> >>
> >> Location Constraints:
> >> Ordering Constraints:
> >> Colocation Constraints:
> >> Ticket Constraints:
> >>
> >> Alerts:
> >>  

[ClusterLabs] corosync doesn't start any resource

2018-06-15 Thread Stefan Krueger
Hello, 

corosync doesn't start any ressource and I don't know why. I tried to 
stop/start the cluster, I also tried to reboot it but it doesn't help. also in 
the logs I dont find nothing that could be useful IMHO.

It would be very nice if someone can help me.

pcs status
Cluster name: zfs-vmstorage
Stack: corosync
Current DC: zfs-serv3 (version 1.1.16-94ff4df) - partition with quorum
Last updated: Fri Jun 15 14:42:32 2018
Last change: Fri Jun 15 14:17:23 2018 by root via cibadmin on zfs-serv3

2 nodes configured
3 resources configured

Online: [ zfs-serv3 zfs-serv4 ]

Full list of resources:

 nfs-server (systemd:nfs-server):   Stopped
 vm_storage (ocf::heartbeat:ZFS):   Stopped
 ha-ip  (ocf::heartbeat:IPaddr2):   Stopped

Daemon Status:
  corosync: active/enabled
  pacemaker: active/enabled
  pcsd: active/enabled




pcs config
Cluster Name: zfs-vmstorage
Corosync Nodes:
 zfs-serv3 zfs-serv4
Pacemaker Nodes:
 zfs-serv3 zfs-serv4

Resources:
 Resource: nfs-server (class=systemd type=nfs-server)
  Operations: start interval=0s timeout=100 (nfs-server-start-interval-0s)
  stop interval=0s timeout=100 (nfs-server-stop-interval-0s)
  monitor interval=60 timeout=100 (nfs-server-monitor-interval-60)
 Resource: vm_storage (class=ocf provider=heartbeat type=ZFS)
  Attributes: pool=vm_storage importargs="-d /dev/disk/by-vdev/"
  Operations: monitor interval=5s timeout=30s (vm_storage-monitor-interval-5s)
  start interval=0s timeout=90 (vm_storage-start-interval-0s)
  stop interval=0s timeout=90 (vm_storage-stop-interval-0s)
 Resource: ha-ip (class=ocf provider=heartbeat type=IPaddr2)
  Attributes: ip=172.16.101.73 cidr_netmask=16
  Operations: start interval=0s timeout=20s (ha-ip-start-interval-0s)
  stop interval=0s timeout=20s (ha-ip-stop-interval-0s)
  monitor interval=10s timeout=20s (ha-ip-monitor-interval-10s)

Stonith Devices:
Fencing Levels:

Location Constraints:
Ordering Constraints:
  Resource Sets:
set nfs-server vm_storage ha-ip action=start 
(id:pcs_rsc_set_nfs-server_vm_storage_ha-ip) 
(id:pcs_rsc_order_set_nfs-server_vm_storage_ha-ip)
set ha-ip nfs-server vm_storage action=stop 
(id:pcs_rsc_set_ha-ip_nfs-server_vm_storage) 
(id:pcs_rsc_order_set_ha-ip_nfs-server_vm_storage)
Colocation Constraints:
  Resource Sets:
set ha-ip nfs-server vm_storage (id:colocation-ha-ip-nfs-server-INFINITY-0) 
setoptions score=INFINITY (id:colocation-ha-ip-nfs-server-INFINITY)
Ticket Constraints:

Alerts:
 No alerts defined

Resources Defaults:
 resource-stickiness: 100
Operations Defaults:
 No defaults set

Cluster Properties:
 cluster-infrastructure: corosync
 cluster-name: zfs-vmstorage
 dc-version: 1.1.16-94ff4df
 have-watchdog: false
 last-lrm-refresh: 1528814481
 no-quorum-policy: ignore
 stonith-enabled: false

Quorum:
  Options:



and here are the Log-files

https://paste.debian.net/hidden/9376add7/

best regards
Stefan
___
Users mailing list: Users@clusterlabs.org
https://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


[ClusterLabs] questions about fence_scsi

2018-06-15 Thread Stefan Krueger
Hello,

so far as I understand I can use fence_scsi on a two node cluster, if the fence 
running on one cluster the other cluster has no access to this devices, 
correct? I've a 2node cluster with shared JBODs and configure fence_scsi, but I 
still can use/mount all this devices on both nodes. Did I something wrong?
My Config:

pcs resource 
 Resource Group: zfs-storage
 vm_storage (ocf::heartbeat:ZFS):   Started zfs-serv3
 ha-ip  (ocf::heartbeat:IPaddr2):   Started zfs-serv3
root@zfs-serv4:~# pcs config
Cluster Name: zfs-vmstorage
Corosync Nodes:
 zfs-serv3 zfs-serv4
Pacemaker Nodes:
 zfs-serv3 zfs-serv4

Resources:
 Group: zfs-storage
  Resource: vm_storage (class=ocf provider=heartbeat type=ZFS)
   Attributes: pool=vm_storage importargs="-d /dev/disk/by-vdev/"
   Operations: monitor interval=5s timeout=30s (vm_storage-monitor-interval-5s)
   start interval=0s timeout=90 (vm_storage-start-interval-0s)
   stop interval=0s timeout=90 (vm_storage-stop-interval-0s)
  Resource: ha-ip (class=ocf provider=heartbeat type=IPaddr2)
   Attributes: ip=172.16.101.73 cidr_netmask=16
   Operations: start interval=0s timeout=20s (ha-ip-start-interval-0s)
   stop interval=0s timeout=20s (ha-ip-stop-interval-0s)
   monitor interval=10s timeout=20s (ha-ip-monitor-interval-10s)

Stonith Devices:
 Resource: fence-vm_storage (class=stonith type=fence_scsi)
  Attributes: pcmk_monitor_action=metadata 
pcmk_host_list=172.16.101.74,172.16.101.75 devices=" 
/dev/disk/by-vdev/j3d03-hdd /dev/disk/by-vdev/j4d03-hdd 
/dev/disk/by-vdev/j3d04-hdd /dev/disk/by-vdev/j4d04-hdd 
/dev/disk/by-vdev/j3d05-hdd /dev/disk/by-vdev/j4d05-hdd 
/dev/disk/by-vdev/j3d06-hdd /dev/disk/by-vdev/j4d06-hdd 
/dev/disk/by-vdev/j3d07-hdd /dev/disk/by-vdev/j4d07-hdd 
/dev/disk/by-vdev/j3d08-hdd /dev/disk/by-vdev/j4d08-hdd 
/dev/disk/by-vdev/j3d09-hdd /dev/disk/by-vdev/j4d09-hdd 
/dev/disk/by-vdev/j3d10-hdd /dev/disk/by-vdev/j4d10-hdd 
/dev/disk/by-vdev/j3d11-hdd /dev/disk/by-vdev/j4d11-hdd 
/dev/disk/by-vdev/j3d12-hdd /dev/disk/by-vdev/j4d12-hdd 
/dev/disk/by-vdev/j3d13-hdd /dev/disk/by-vdev/j4d13-hdd 
/dev/disk/by-vdev/j3d14-hdd /dev/disk/by-vdev/j4d14-hdd 
/dev/disk/by-vdev/j3d15-hdd /dev/disk/by-vdev/j4d15-hdd 
/dev/disk/by-vdev/j3d16-hdd /dev/disk/by-vdev/j4d16-hdd 
/dev/disk/by-vdev/j3d17-hdd /dev/disk/by-vdev/j4d17-hdd 
/dev/disk/by-vdev/j3d18-hdd /dev/disk/by-vdev/j4d18-hdd /dev/d
 isk/by-vdev/j3d19-hdd /dev/disk/by-vdev/j4d19-hdd log 
/dev/disk/by-vdev/j3d00-ssd /dev/disk/by-vdev/j4d00-ssd cache 
/dev/disk/by-vdev/j3d02-ssd"
  Meta Attrs: provides=unfencing 
  Operations: monitor interval=60s (fence-vm_storage-monitor-interval-60s)
Fencing Levels:

Location Constraints:
Ordering Constraints:
Colocation Constraints:
Ticket Constraints:

Alerts:
 No alerts defined

Resources Defaults:
 resource-stickiness: 100
Operations Defaults:
 No defaults set

Cluster Properties:
 cluster-infrastructure: corosync
 cluster-name: zfs-vmstorage
 dc-version: 1.1.16-94ff4df
 have-watchdog: false
 last-lrm-refresh: 1528814481
 no-quorum-policy: ignore

Quorum:
  Options:


Thanks for help!
Best regards
Stefan
___
Users mailing list: Users@clusterlabs.org
https://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


[ClusterLabs] ressource stopped unexpectly

2018-06-13 Thread Stefan Krueger
Hello,

I've a problem with my cluster. When I use 'pcs cluster standby serv3' it moves 
all ressources to serv4 that works fine, but when I restart a node the 
Ressource ha-ip become stopped and I don't know why. Can somebody give me an 
hint why this happen and how to resolve that?

btw: i use this guide: https://github.com/ewwhite/zfs-ha/wiki
the logfile is here (i guess it is too long for the mailinglist) 
https://paste.debian.net/hidden/2e001867/

thanks for help!

best regards
Stefan


pcs status
Cluster name: zfs-vmstorage
Stack: corosync
Current DC: zfs-serv3 (version 1.1.16-94ff4df) - partition with quorum
Last updated: Tue Jun 12 16:56:45 2018
Last change: Tue Jun 12 16:44:52 2018 by hacluster via crm_attribute on 
zfs-serv3

2 nodes configured
3 resources configured

Online: [ zfs-serv3 zfs-serv4 ]

Full list of resources:

 fence-vm_storage   (stonith:fence_scsi):   Started zfs-serv3
 Resource Group: zfs-storage
 vm_storage (ocf::heartbeat:ZFS):   Started zfs-serv3
 ha-ip  (ocf::heartbeat:IPaddr2):   Stopped

Daemon Status:
  corosync: active/enabled
  pacemaker: active/enabled
  pcsd: active/enabled



pcs config
Cluster Name: zfs-vmstorage
Corosync Nodes:
 zfs-serv3 zfs-serv4
Pacemaker Nodes:
 zfs-serv3 zfs-serv4

Resources:
 Group: zfs-storage
  Resource: vm_storage (class=ocf provider=heartbeat type=ZFS)
   Attributes: pool=vm_storage importargs="-d /dev/disk/by-vdev/"
   Operations: monitor interval=5s timeout=30s (vm_storage-monitor-interval-5s)
   start interval=0s timeout=90 (vm_storage-start-interval-0s)
   stop interval=0s timeout=90 (vm_storage-stop-interval-0s)
  Resource: ha-ip (class=ocf provider=heartbeat type=IPaddr2)
   Attributes: ip=172.16.101.73 cidr_netmask=16
   Operations: start interval=0s timeout=20s (ha-ip-start-interval-0s)
   stop interval=0s timeout=20s (ha-ip-stop-interval-0s)
   monitor interval=10s timeout=20s (ha-ip-monitor-interval-10s)

Stonith Devices:
 Resource: fence-vm_storage (class=stonith type=fence_scsi)
  Attributes: pcmk_monitor_action=metadata 
pcmk_host_list=172.16.101.74,172.16.101.75 devices=" 
/dev/disk/by-vdev/j3d03-hdd /dev/disk/by-vdev/j4d03-hdd 
/dev/disk/by-vdev/j3d04-hdd /dev/disk/by-vdev/j4d04-hdd 
/dev/disk/by-vdev/j3d05-hdd /dev/disk/by-vdev/j4d05-hdd 
/dev/disk/by-vdev/j3d06-hdd /dev/disk/by-vdev/j4d06-hdd 
/dev/disk/by-vdev/j3d07-hdd /dev/disk/by-vdev/j4d07-hdd 
/dev/disk/by-vdev/j3d08-hdd /dev/disk/by-vdev/j4d08-hdd 
/dev/disk/by-vdev/j3d09-hdd /dev/disk/by-vdev/j4d09-hdd 
/dev/disk/by-vdev/j3d10-hdd /dev/disk/by-vdev/j4d10-hdd 
/dev/disk/by-vdev/j3d11-hdd /dev/disk/by-vdev/j4d11-hdd 
/dev/disk/by-vdev/j3d12-hdd /dev/disk/by-vdev/j4d12-hdd 
/dev/disk/by-vdev/j3d13-hdd /dev/disk/by-vdev/j4d13-hdd 
/dev/disk/by-vdev/j3d14-hdd /dev/disk/by-vdev/j4d14-hdd 
/dev/disk/by-vdev/j3d15-hdd /dev/disk/by-vdev/j4d15-hdd 
/dev/disk/by-vdev/j3d16-hdd /dev/disk/by-vdev/j4d16-hdd 
/dev/disk/by-vdev/j3d17-hdd /dev/disk/by-vdev/j4d17-hdd 
/dev/disk/by-vdev/j3d18-hdd /dev/disk/by-vdev/j4d18-hdd /dev/d
 isk/by-vdev/j3d19-hdd /dev/disk/by-vdev/j4d19-hdd log 
/dev/disk/by-vdev/j3d00-ssd /dev/disk/by-vdev/j4d00-ssd cache 
/dev/disk/by-vdev/j3d02-ssd"
  Meta Attrs: provides=unfencing 
  Operations: monitor interval=60s (fence-vm_storage-monitor-interval-60s)
Fencing Levels:

Location Constraints:
Ordering Constraints:
Colocation Constraints:
Ticket Constraints:

Alerts:
 No alerts defined

Resources Defaults:
 resource-stickiness: 100
Operations Defaults:
 No defaults set

Cluster Properties:
 cluster-infrastructure: corosync
 cluster-name: zfs-vmstorage
 dc-version: 1.1.16-94ff4df
 have-watchdog: false
 last-lrm-refresh: 1528814481
 no-quorum-policy: ignore

Quorum:
  Options:

___
Users mailing list: Users@clusterlabs.org
https://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


[ClusterLabs] dont (re)start a ressource if there is already running

2017-11-01 Thread Stefan Krueger
Hello,

I've a configuration where 2nfs servers are running on different nodes and if 
one node goes down move a ressources to the other one, but dont try to restart 
the ressource, because it is already running.

My config looks like this:
node 1084815873: zfs-serv1
node 1084815874: zfs-serv2
primitive HA_IP-Serv1 IPaddr2 \
params ip=172.16.101.70 cidr_netmask=16 \
op monitor interval=20 timeout=30 on-fail=restart nic=bond0 \
meta target-role=Started
primitive HA_IP-Serv2 IPaddr2 \
params ip=172.16.101.74 cidr_netmask=16 \
op monitor interval=10s nic=bond0
primitive nc_storage ZFS \
params pool=nc_storage importargs="-d /dev/disk/by-partlabel/"
primitive nfs-serv1 systemd:nfs-server \
op monitor interval=5s
primitive nfs-serv2 systemd:nfs-server \
op monitor interval=5s
primitive vm_storage ZFS \
params pool=vm_storage importargs="-d /dev/disk/by-partlabel/"
colocation compl_zfs-serv1 +inf: HA_IP-Serv1 nc_storage nfs-serv1
colocation compl_zfs-serv2 +inf: HA_IP-Serv2 vm_storage nfs-serv2
order start_serv1 nfs-serv1:start nc_storage:start HA_IP-Serv1:start 
symmetrical=false
order start_serv2 nfs-serv2:start vm_storage:start HA_IP-Serv2:start 
symmetrical=false
order stop_serv1 HA_IP-Serv1:stop nfs-serv1:stop nc_storage:stop 
symmetrical=false
order stop_serv2 HA_IP-Serv2:stop nfs-serv2:stop vm_storage:stop 
symmetrical=false
property cib-bootstrap-options: \
have-watchdog=false \
dc-version=1.1.16-94ff4df \
cluster-infrastructure=corosync \
cluster-name=debian \
no-quorum-policy=ignore \
default-resource-stickiness=100 \
stonith-enabled=false \
last-lrm-refresh=1509546667

So is it possible to check if the ressource nfs is already running on the other 
node?

best regards
stefan

___
Users mailing list: Users@clusterlabs.org
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


Re: [ClusterLabs] different start/stop order

2017-10-27 Thread Stefan Krueger
Hello Ken, thanks for this information!

> You can make ordering constraints asymmetrical,
> so they only apply in the listed direction:
> 
> http://clusterlabs.org/doc/en-US/Pacemaker/1.1-pcs/html-single/Pacemaker_Explained/index.html#s-resource-ordering

Ok but i see only how can i create a start order, but how can i create a 
different stop order?

Best regards
Stefan

___
Users mailing list: Users@clusterlabs.org
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


[ClusterLabs] different start/stop order

2017-10-27 Thread Stefan Krueger
Hello,

is it possible to start a ressource in a different way then they stopped?
eg:
primitive HA_IP-Serv1 IPaddr2 \
params ip=172.16.101.70 cidr_netmask=16 \
op monitor interval=20 timeout=30 on-fail=restart nic=bond0 \
meta target-role=Started
primitive HA_IP-Serv2 IPaddr2 \
params ip=172.16.101.74 cidr_netmask=16 \
op monitor interval=10s nic=bond0
primitive nc_storage ZFS \
params pool=nc_storage importargs="-d /dev/disk/by-partlabel/"
primitive nfs-serv1 systemd:nfs-server
primitive nfs-serv2 systemd:nfs-server
group compl_zfs-serv1 nc_storage nfs-serv1 HA_IP-Serv1

I want that pacemaker stop it in this order:
HA_IP-Serv1 nfs-serv1  nc_storage

but start it in this order:
nfs-serv1 nc_storage HA_IP-Serv1

it that possible and how can I achieve that?

best regards
Stefan

___
Users mailing list: Users@clusterlabs.org
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


[ClusterLabs] can't move/migrate ressource

2017-10-11 Thread Stefan Krueger
Hello,

when i try to migrate a ressource from one server to an other (for example for 
maintenance), it don't work.
a single ressource works fine, after that I create a group with 2 ressources 
and try to move that.

my config is:
crm conf show
node 739272007: zfs-serv1
node 739272008: zfs-serv2
primitive HA_IP-Serv1 IPaddr2 \
params ip=172.16.101.70 cidr_netmask=16 \
op monitor interval=20 timeout=30 on-fail=restart nic=bond0 \
meta target-role=Started
primitive HA_IP-Serv2 IPaddr2 \
params ip=172.16.101.74 cidr_netmask=16 \
op monitor interval=10s nic=bond0
primitive nc_storage ZFS \
params pool=nc_storage importargs="-d /dev/disk/by-partlabel/"
group compl_zfs-serv1 nc_storage HA_IP-Serv1
location cli-prefer-HA_IP-Serv1 compl_zfs-serv1 role=Started inf: zfs-serv1
location cli-prefer-HA_IP-Serv2 HA_IP-Serv2 role=Started inf: zfs-serv2
location cli-prefer-compl_zfs-serv1 compl_zfs-serv1 role=Started inf: zfs-serv2
location cli-prefer-nc_storage compl_zfs-serv1 role=Started inf: zfs-serv1
property cib-bootstrap-options: \
have-watchdog=false \
dc-version=1.1.16-94ff4df \
cluster-infrastructure=corosync \
cluster-name=debian \
no-quorum-policy=ignore \
default-resource-stickiness=100 \
stonith-enabled=false \
last-lrm-refresh=1507702403


command:
crm resource move compl_zfs-serv1 zfs-serv2


pacemakerlog from zfs-serv2:
Oct 11 13:55:58 [3556] zfs-serv2cib: info: cib_perform_op:  Diff: 
--- 0.106.0 2
Oct 11 13:55:58 [3556] zfs-serv2cib: info: cib_perform_op:  Diff: 
+++ 0.107.0 cc224b15d0a796e040b026b7c2965770
Oct 11 13:55:58 [3556] zfs-serv2cib: info: cib_perform_op:  -- 
/cib/configuration/constraints/rsc_location[@id='cli-prefer-compl_zfs-serv1']
Oct 11 13:55:58 [3556] zfs-serv2cib: info: cib_perform_op:  +  
/cib:  @epoch=107
Oct 11 13:55:58 [3556] zfs-serv2cib: info: cib_process_request: 
Completed cib_delete operation for section constraints: OK (rc=0, 
origin=zfs-serv1/crm_resource/3, version=0.107.0)
Oct 11 13:55:58 [3561] zfs-serv2   crmd: info: abort_transition_graph:  
Transition aborted by deletion of 
rsc_location[@id='cli-prefer-compl_zfs-serv1']: Configuration change | 
cib=0.107.0 source=te_update_diff:444 
path=/cib/configuration/constraints/rsc_location[@id='cli-prefer-compl_zfs-serv1']
 complete=true
Oct 11 13:55:58 [3561] zfs-serv2   crmd:   notice: do_state_transition: 
State transition S_IDLE -> S_POLICY_ENGINE | input=I_PE_CALC 
cause=C_FSA_INTERNAL origin=abort_transition_graph
Oct 11 13:55:58 [3556] zfs-serv2cib: info: cib_perform_op:  Diff: 
--- 0.107.0 2
Oct 11 13:55:58 [3556] zfs-serv2cib: info: cib_perform_op:  Diff: 
+++ 0.108.0 (null)
Oct 11 13:55:58 [3556] zfs-serv2cib: info: cib_perform_op:  +  
/cib:  @epoch=108
Oct 11 13:55:58 [3556] zfs-serv2cib: info: cib_perform_op:  ++ 
/cib/configuration/constraints:  
Oct 11 13:55:58 [3556] zfs-serv2cib: info: cib_process_request: 
Completed cib_modify operation for section constraints: OK (rc=0, 
origin=zfs-serv1/crm_resource/4, version=0.108.0)
Oct 11 13:55:58 [3561] zfs-serv2   crmd: info: abort_transition_graph:  
Transition aborted by rsc_location.cli-prefer-compl_zfs-serv1 'create': 
Configuration change | cib=0.108.0 source=te_update_diff:444 
path=/cib/configuration/constraints complete=true
Oct 11 13:55:58 [3560] zfs-serv2pengine:   notice: unpack_config:   On loss 
of CCM Quorum: Ignore
Oct 11 13:55:58 [3560] zfs-serv2pengine: info: determine_online_status: 
Node zfs-serv2 is online
Oct 11 13:55:58 [3560] zfs-serv2pengine: info: determine_online_status: 
Node zfs-serv1 is online
Oct 11 13:55:58 [3560] zfs-serv2pengine: info: determine_op_status: 
Operation monitor found resource nc_storage active on zfs-serv2
Oct 11 13:55:58 [3560] zfs-serv2pengine: info: native_print:
HA_IP-Serv2 (ocf::heartbeat:IPaddr2):   Started zfs-serv2
Oct 11 13:55:58 [3560] zfs-serv2pengine: info: group_print:  
Resource Group: compl_zfs-serv1
Oct 11 13:55:58 [3560] zfs-serv2pengine: info: native_print: 
nc_storage (ocf::heartbeat:ZFS):   Started zfs-serv1
Oct 11 13:55:58 [3560] zfs-serv2pengine: info: native_print: 
HA_IP-Serv1(ocf::heartbeat:IPaddr2):   Started zfs-serv1
Oct 11 13:55:58 [3560] zfs-serv2pengine: info: LogActions:  Leave   
HA_IP-Serv2 (Started zfs-serv2)
Oct 11 13:55:58 [3560] zfs-serv2pengine: info: LogActions:  Leave   
nc_storage  (Started zfs-serv1)
Oct 11 13:55:58 [3560] zfs-serv2pengine: info: LogActions:  Leave   
HA_IP-Serv1 (Started zfs-serv1)
Oct 11 13:55:58 [3560] zfs-serv2pengine:   notice: process_pe_message:  
Calculated transition 8, saving inputs in 
/var/lib/pacemaker/pengine/pe-input-1348.bz2

Re: [ClusterLabs] trouble with IPaddr2

2017-10-11 Thread Stefan Krueger
Hello Valentin,
thanks for you help

> Can you share more info on the network of zfs-serv2, for example: ip a?
ip a s
1: lo:  mtu 65536 qdisc noqueue state UNKNOWN group 
default qlen 1
link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
inet 127.0.0.1/8 scope host lo
   valid_lft forever preferred_lft forever
inet6 ::1/128 scope host
   valid_lft forever preferred_lft forever
2: enp4s0f0:  mtu 1500 qdisc mq state UP group 
default qlen 1000
link/ether ac:1f:6b:xx:xx:xx brd ff:ff:ff:ff:ff:ff
inet 172.16.22.126/16 brd 172.16.255.255 scope global enp4s0f0
   valid_lft forever preferred_lft forever
inet 172.16.101.74/16 brd 172.16.255.255 scope global secondary enp4s0f0
   valid_lft forever preferred_lft forever
inet6 fe80::ae1f:6bff::/64 scope link
   valid_lft forever preferred_lft forever
3: enp4s0f1:  mtu 1500 qdisc noop state DOWN group default 
qlen 1000
link/ether ac:1f:6b:xx:xx:xx brd ff:ff:ff:ff:ff:ff
4: ens2f0:  mtu 1500 qdisc mq master 
bond0 state UP group default qlen 1000
link/ether 3c:fd:fe:xx:xx:xx brd ff:ff:ff:ff:ff:ff
5: ens2f1:  mtu 1500 qdisc mq master 
bond0 state UP group default qlen 1000
link/ether 3c:fd:fe:xx:xx:xx brd ff:ff:ff:ff:ff:ff
6: bond0:  mtu 1500 qdisc noqueue state 
UP group default qlen 1000
link/ether 3c:fd:fe:xx:xx:xx brd ff:ff:ff:ff:ff:ff
inet 172.16.101.72/16 brd 172.16.255.255 scope global bond0
   valid_lft forever preferred_lft forever
inet 172.16.101.74/16 brd 172.16.255.255 scope global secondary bond0
   valid_lft forever preferred_lft forever
inet6 fe80::3efd:feff::/64 scope link
   valid_lft forever preferred_lft forever


ohh damn.. thanks a lot for this hint.. I delete all the IPs on enp4s0f0, and 
than it works..
but could you please explain why it now works? why he has a problem with this 
IPs?

best regards

___
Users mailing list: Users@clusterlabs.org
http://lists.clusterlabs.org/mailman/listinfo/users

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org


[ClusterLabs] trouble with IPaddr2

2017-10-11 Thread Stefan Krueger
Hello,

I've a simple setup with just 3 resources (at the moment), the ZFS resource 
also works fine.BUT my IPaddr2 don't work, and I dont know why and how to 
resolve that.

my config:
conf sh
node 739272007: zfs-serv1
node 739272008: zfs-serv2
primitive HA_IP-Serv1 IPaddr2 \
params ip=172.16.101.70 cidr_netmask=16 \
op monitor interval=20 timeout=30 on-fail=restart nic=bond0 \
meta target-role=Started
primitive HA_IP-Serv2 IPaddr2 \
params ip=172.16.101.74 cidr_netmask=16 \
op monitor interval=10s nic=bond0
primitive nc_storage ZFS \
params pool=nc_storage importargs="-d /dev/disk/by-partlabel/"
location cli-prefer-HA_IP-Serv1 HA_IP-Serv1 role=Started inf: zfs-serv1
location cli-prefer-HA_IP-Serv2 HA_IP-Serv2 role=Started inf: zfs-serv2
location cli-prefer-nc_storage nc_storage role=Started inf: zfs-serv2
property cib-bootstrap-options: \
have-watchdog=false \
dc-version=1.1.16-94ff4df \
cluster-infrastructure=corosync \
cluster-name=debian \
no-quorum-policy=ignore \
default-resource-stickiness=100 \
stonith-enabled=false \
last-lrm-refresh=1507702403


command:
resource move HA_IP-Serv1 zfs-serv2


pacemaker log from zfs-serv2:
Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_process_request:
Completed cib_delete operation for section constraints: OK (rc=0, 
origin=zfs-serv1/crm_resource/3, version=0.82.44)
Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_perform_op: Diff: 
--- 0.82.44 2
Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_perform_op: Diff: 
+++ 0.83.0 (null)
Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_perform_op: +  
/cib:  @epoch=83, @num_updates=0
Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_perform_op: +  
/cib/configuration/constraints/rsc_location[@id='cli-prefer-HA_IP-Serv1']:  
@node=zfs-serv2
Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_process_request:
Completed cib_modify operation for section constraints: OK (rc=0, 
origin=zfs-serv1/crm_resource/4, version=0.83.0)
Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_perform_op: Diff: 
--- 0.83.0 2
Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_perform_op: Diff: 
+++ 0.83.1 (null)
Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_perform_op: +  
/cib:  @num_updates=1
Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_perform_op: +  
/cib/status/node_state[@id='739272007']/lrm[@id='739272007']/lrm_resources/lrm_resource[@id='HA_IP-Serv1']/lrm_rsc_op[@id='HA_IP-Serv1_last_0']:
  @operation_key=HA_IP-Serv1_stop_0, @operation=stop, 
@crm-debug-origin=do_update_resource, 
@transition-key=8:1574:0:d4b03c3c-1a4e-4609-86ca-675fa4a2ec8f, 
@transition-magic=0:0;8:1574:0:d4b03c3c-1a4e-4609-86ca-675fa4a2ec8f, 
@call-id=55, @last-run=1507702772, @last-rc-change=1507702772, @exec
Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_process_request:
Completed cib_modify operation for section status: OK (rc=0, 
origin=zfs-serv1/crmd/1853, version=0.83.1)
Oct 11 08:19:32 [23938] zfs-serv2   crmd: info: do_lrm_rsc_op:  
Performing key=9:1574:0:d4b03c3c-1a4e-4609-86ca-675fa4a2ec8f 
op=HA_IP-Serv1_start_0
Oct 11 08:19:32 [23935] zfs-serv2   lrmd: info: log_execute:
executing - rsc:HA_IP-Serv1 action:start call_id:17
Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_file_backup:
Archived previous version as /var/lib/pacemaker/cib/cib-39.raw
Oct 11 08:19:32 [23935] zfs-serv2   lrmd: info: log_finished:   
finished - rsc:HA_IP-Serv1 action:start call_id:17 pid:27672 exit-code:0 
exec-time:48ms queue-time:0ms
Oct 11 08:19:32 [23938] zfs-serv2   crmd: info: action_synced_wait: 
Managed IPaddr2_meta-data_0 process 27735 exited with rc=0
Oct 11 08:19:32 [23938] zfs-serv2   crmd:   notice: process_lrm_event:  
Result of start operation for HA_IP-Serv1 on zfs-serv2: 0 (ok) | call=17 
key=HA_IP-Serv1_start_0 confirmed=true cib-update=15
Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_process_request:
Forwarding cib_modify operation for section status to all (origin=local/crmd/15)
Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_perform_op: Diff: 
--- 0.83.1 2
Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_perform_op: Diff: 
+++ 0.83.2 (null)
Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_perform_op: +  
/cib:  @num_updates=2
Oct 11 08:19:32 [23933] zfs-serv2cib: info: cib_perform_op: +  
/cib/status/node_state[@id='739272008']/lrm[@id='739272008']/lrm_resources/lrm_resource[@id='HA_IP-Serv1']/lrm_rsc_op[@id='HA_IP-Serv1_last_0']:
  @operation_key=HA_IP-Serv1_start_0, @operation=start, 
@transition-key=9:1574:0:d4b03c3c-1a4e-4609-86ca-675fa4a2ec8f, 
@transition-magic=0:0;9:1574:0:d4b03c3c-1a4e-4609-86ca-675fa4a2ec8f, 
@call-id=17, @rc-code=0, @last-run=1507702772,