Thanks for feedback!

Regarding patch 1/2: I grep'd through the sources and could not find any references to the heading names I changed. A quick look through the GUI also didn't reveal any obvious references.

Some of my own notes inline, I will send v2 today.

On 7/9/19 9:19 AM, Thomas Lamprecht wrote:
On 7/8/19 6:26 PM, Stefan Reiter wrote:
Parts about multicast and RRP have been removed entirely. Instead, a new
section 'Corosync Redundancy' has been added explaining the concept of
links and link priorities.


note bad at all, still some notes inline.

Signed-off-by: Stefan Reiter <s.rei...@proxmox.com>
---
  pvecm.adoc | 372 +++++++++++++++++++++--------------------------------
  1 file changed, 147 insertions(+), 225 deletions(-)

diff --git a/pvecm.adoc b/pvecm.adoc
index 1c0b9e7..1246111 100644
--- a/pvecm.adoc
+++ b/pvecm.adoc
@@ -56,13 +56,8 @@ Grouping nodes into a cluster has the following advantages:
  Requirements
  ------------
-* All nodes must be in the same network as `corosync` uses IP Multicast
- to communicate between nodes (also see
- http://www.corosync.org[Corosync Cluster Engine]). Corosync uses UDP
- ports 5404 and 5405 for cluster communication.
-+
-NOTE: Some switches do not support IP multicast by default and must be
-manually enabled first.
+* All nodes must be able to contact each other via UDP ports 5404 and 5405 for
+ corosync to work.
* Date and time have to be synchronized. @@ -84,6 +79,11 @@ NOTE: While it's possible for {pve} 4.4 and {pve} 5.0 this is not supported as
  production configuration and should only used temporarily during upgrading the
  whole cluster from one to another major version.
+NOTE: Mixing {pve} 6.x and earlier versions is not supported, because of the
+major corosync upgrade. While possible to run corosync 3 on {pve} 5.4, this
+configuration is not supported for production environments and should only be
+used while upgrading a cluster.
+
Preparing Nodes
  ---------------
@@ -96,10 +96,12 @@ Currently the cluster creation can either be done on the 
console (login via
  `ssh`) or the API, which we have a GUI implementation for (__Datacenter ->
  Cluster__).
-While it's often common use to reference all other nodenames in `/etc/hosts`
-with their IP this is not strictly necessary for a cluster, which normally uses
-multicast, to work. It maybe useful as you then can connect from one node to
-the other with SSH through the easier to remember node name.
+While it's common to reference all nodenames and their IPs in `/etc/hosts` (or
+make their names resolveable through other means), this is not strictly
+necessary for a cluster to work. It may be useful however, as you can then
+connect from one node to the other with SSH via the easier to remember node
+name. (see also xref:pvecm_corosync_addresses[Link Address Types])
+
[[pvecm_create_cluster]]
  Create the Cluster
@@ -113,31 +115,12 @@ node names.
   hp1# pvecm create CLUSTERNAME
  ----
-CAUTION: The cluster name is used to compute the default multicast address.
-Please use unique cluster names if you run more than one cluster inside your
-network. To avoid human confusion, it is also recommended to choose different
-names even if clusters do not share the cluster network.

Maybe move this from a "CAUTION" to a "NOTE" and keep the hint that it still
makes sense to use unique cluster names, to avoid human confusion and as I have
a feeling that there are other assumption in corosync which depend on that.
Also, _if_ multicast gets integrated into knet we probably have a similar issue
again, so try to bring people in lane now already, even if not 100% required.


Makes sense. I just wanted to avoid mentioning multicast in the general instructions, to avoid people reading the docs for the first time being confused if they need it or not.

-
  To check the state of your cluster use:
----
   hp1# pvecm status
  ----
-Multiple Clusters In Same Network
-~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
-
-It is possible to create multiple clusters in the same physical or logical
-network. Each cluster must have a unique name, which is used to generate the
-cluster's multicast group address. As long as no duplicate cluster names are
-configured in one network segment, the different clusters won't interfere with
-each other.
-
-If multiple clusters operate in a single network it may be beneficial to setup
-an IGMP querier and enable IGMP Snooping in said network. This may reduce the
-load of the network significantly because multicast packets are only delivered
-to endpoints of the respective member nodes.
-

It's still possible to create multiple clusters in the same network, so I'd keep
above and just adapt to non-multicast for now..

[[pvecm_join_node_to_cluster]]
  Adding Nodes to the Cluster
@@ -150,7 +133,7 @@ Login via `ssh` to the node you want to add.
  ----
For `IP-ADDRESS-CLUSTER` use the IP or hostname of an existing cluster node.
-An IP address is recommended (see xref:pvecm_corosync_addresses[Ring Address 
Types]).
+An IP address is recommended (see xref:pvecm_corosync_addresses[Link Address 
Types]).

Maybe somewhere a note that while the new things are named "Link" the config
still refers to "ringX_addr" for backward compatibility.

CAUTION: A new node cannot hold any VMs, because you would get
  conflicts about identical VM IDs. Also, all existing configuration in
@@ -173,7 +156,7 @@ Date:             Mon Apr 20 12:30:13 2015
  Quorum provider:  corosync_votequorum
  Nodes:            4
  Node ID:          0x00000001
-Ring ID:          1928
+Ring ID:          1/8
  Quorate:          Yes
Votequorum information
@@ -217,15 +200,15 @@ Adding Nodes With Separated Cluster Network
  ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
When adding a node to a cluster with a separated cluster network you need to
-use the 'ringX_addr' parameters to set the nodes address on those networks:
+use the 'link0' parameter to set the nodes address on that network:
[source,bash]
  ----
-pvecm add IP-ADDRESS-CLUSTER -ring0_addr IP-ADDRESS-RING0
+pvecm add IP-ADDRESS-CLUSTER -link0 LOCAL-IP-ADDRESS-LINK0
  ----
-If you want to use the Redundant Ring Protocol you will also want to pass the
-'ring1_addr' parameter.
+If you want to use the built-in xref:pvecm_redundancy[redundancy] of the
+kronosnet transport you can also pass the 'link1' parameter.
Remove a Cluster Node
@@ -283,7 +266,7 @@ Date:             Mon Apr 20 12:44:28 2015
  Quorum provider:  corosync_votequorum
  Nodes:            3
  Node ID:          0x00000001
-Ring ID:          1992
+Ring ID:          1/8
  Quorate:          Yes
Votequorum information
@@ -302,8 +285,8 @@ Membership information
  0x00000003          1 192.168.15.92
  ----
-If, for whatever reason, you want that this server joins the same
-cluster again, you have to
+If, for whatever reason, you want this server to join the same cluster again,
+you have to
* reinstall {pve} on it from scratch @@ -329,14 +312,14 @@ storage with another cluster, as storage locking doesn't work over cluster
  boundary. Further, it may also lead to VMID conflicts.
Its suggested that you create a new storage where only the node which you want
-to separate has access. This can be an new export on your NFS or a new Ceph
+to separate has access. This can be a new export on your NFS or a new Ceph
  pool, to name a few examples. Its just important that the exact same storage
  does not gets accessed by multiple clusters. After setting this storage up 
move
  all data from the node and its VMs to it. Then you are ready to separate the
  node from the cluster.
WARNING: Ensure all shared resources are cleanly separated! You will run into
-conflicts and problems else.
+conflicts and problems otherwise.
First stop the corosync and the pve-cluster services on the node:
  [source,bash]
@@ -400,6 +383,7 @@ the nodes can still connect to each other with public key 
authentication. This
  should be fixed by removing the respective keys from the
  '/etc/pve/priv/authorized_keys' file.
+
  Quorum
  ------
@@ -419,12 +403,13 @@ if it loses quorum. NOTE: {pve} assigns a single vote to each node by default. +
  Cluster Network
  ---------------
The cluster network is the core of a cluster. All messages sent over it have to
-be delivered reliable to all nodes in their respective order. In {pve} this
-part is done by corosync, an implementation of a high performance low overhead
+be delivered reliably to all nodes in their respective order. In {pve} this
+part is done by corosync, an implementation of a high performance, low overhead
  high availability development toolkit. It serves our decentralized
  configuration file system (`pmxcfs`).
@@ -432,75 +417,59 @@ configuration file system (`pmxcfs`).
  Network Requirements
  ~~~~~~~~~~~~~~~~~~~~
  This needs a reliable network with latencies under 2 milliseconds (LAN
-performance) to work properly. While corosync can also use unicast for
-communication between nodes its **highly recommended** to have a multicast
-capable network. The network should not be used heavily by other members,
-ideally corosync runs on its own network.
-*never* share it with network where storage communicates too.
+performance) to work properly. The network should not be used heavily by other
+members, ideally corosync runs on its own network. Do not use a shared network
+for corosync and storage (except as a potential low-priority fallback in a
+xref:pvecm_redundancy[redundant] configuration).
Before setting up a cluster it is good practice to check if the network is fit
-for that purpose.
+for that purpose. With corosync 3, it is enough to ensure all nodes can reach
+each other over the interfaces you are planning to use. Using `ping` is enough
+for a basic test.
-* Ensure that all nodes are in the same subnet. This must only be true for the
-  network interfaces used for cluster communication (corosync).
+If the {pve} firewall is enabled, ACCEPT rules for corosync will automatically
+be generated - no manual action is required.

"will automatically be generated" vs "will be automatically generated"?


or "will be generated automatically" ;)

I feel like all of these are correct though.

-* Ensure all nodes can reach each other over those interfaces, using `ping` is
-  enough for a basic test.
+NOTE: Corosync used Multicast before version 3.0 (introduced in {pve} 6.0).
+Modern versions rely on https://kronosnet.org/[Kronosnet] for cluster
+communication, which uses regular UDP unicast.

"... which, for now, only supports regular UDP unicast."?

-* Ensure that multicast works in general and a high package rates. This can be
-  done with the `omping` tool. The final "%loss" number should be < 1%.
-+
-[source,bash]
-----
-omping -c 10000 -i 0.001 -F -q NODE1-IP NODE2-IP ...
-----
-
-* Ensure that multicast communication works over an extended period of time.
-  This uncovers problems where IGMP snooping is activated on the network but
-  no multicast querier is active. This test has a duration of around 10
-  minutes.
-+
-[source,bash]
-----
-omping -c 600 -i 1 -q NODE1-IP NODE2-IP ...
-----
-
-Your network is not ready for clustering if any of these test fails. Recheck
-your network configuration. Especially switches are notorious for having
-multicast disabled by default or IGMP snooping enabled with no IGMP querier
-active.
-
-In smaller cluster its also an option to use unicast if you really cannot get
-multicast to work.
+CAUTION: You can still enable Multicast or legacy unicast by setting your
+transport to `udp` or `udpu` in your 
xref:pvecm_edit_corosync_conf[corosync.conf],
+but keep in mind that this will disable all cryptography and redundancy 
support.
+This is therefore not recommended.

off-topic: what I general see as a loss are the omping checks, they could be 
used
to get connection and latencies stats from all of the cluster nodes easily, 
that was
nice to get a feeling of the whole network...




Is it still good to leave this out then? Should I mention it somewhere?

Felt like its a bit misleading though, since it requires multicast, and may not work on networks where corosync is just fine.

Separate Cluster Network
  ~~~~~~~~~~~~~~~~~~~~~~~~
When creating a cluster without any parameters the cluster network is generally
-shared with the Web UI and the VMs and its traffic. Depending on your setup
+shared with the Web UI and the VMs and their traffic. Depending on your setup,
  even storage traffic may get sent over the same network. Its recommended to
  change that, as corosync is a time critical real time application.
+NOTE: Setting up corosync links on a different network does not affect other
+cluster communications (e.g. Web UI, default migration network, etc...).
+

This note is a bit confusing, IMO, what to you want to tell here?


Couldn't find a better way to word, but it's something that confused me a lot personally reading the doc for the first time. Basically meaning that if you change the corosync network, it will not affect other cluster communication.

We use the term "cluster network" a lot, and it took me a while to realize, that this "cluster network" is completely unrelated to corosync. It makes sense if you know it, but it confused me just looking at the doc.

I'll find a way to reword it or take it out for v2.

snip


Rest makes perfect sense, will be fixed.

_______________________________________________
pve-devel mailing list
pve-devel@pve.proxmox.com
https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-devel

Reply via email to