On Wed, Aug 30, 2017 at 11:19:08PM +0300, Michael S. Tsirkin wrote: > On Wed, Aug 30, 2017 at 04:25:10PM -0300, Eduardo Habkost wrote: > > On Wed, Aug 30, 2017 at 05:23:39PM +0300, Michael S. Tsirkin wrote: > > > On Wed, Aug 30, 2017 at 10:17:27AM -0300, Eduardo Habkost wrote: > > > > I'm CCing libvir-list and qemu-devel because I would like to get > > > > feedback from libvirt and QEMU developers too. > > > > > > > > On Tue, Aug 08, 2017 at 10:49:21PM +0300, Michael S. Tsirkin wrote: > > > > > On Tue, Jul 18, 2017 at 03:42:08PM +0200, Maxime Coquelin wrote: > > > > > > This is an revival from a thread I initiated earlier this year [0], > > > > > > that > > > > > > I had to postpone due to other priorities. > > > > > > > > > > > > First, I'd like to thanks reviewers of my first proposal, this new > > > > > > version tries to address the comments made: > > > > > > 1.This is Nova's role and not Libvirt's to query hosts supported > > > > > > compatibility modes and to select one, since Nova adds the > > > > > > vhost-user > > > > > > ports and has visibility on other hosts. Hence I remove libvirt ML > > > > > > and > > > > > > add Openstack one in the recipient list. > > > > > > 2. By default, the compatibility version selected is the most > > > > > > recent > > > > > > one, except if the admin selects on older compat version. > > > > > > > > > > > > The goal of this thread is to draft a solution based on the outcomes > > > > > > of discussions with contributors of the different parties (DPDK/OVS > > > > > > /Nova/...). > > > > > > > > > > > > I'm really interested on feedback from OVS & Nova contributors, > > > > > > as my experience with these projects is rather limited. > > > > > > > > > > > > Problem statement: > > > > > > ================== > > > > > > > > > > > > When migrating a VM from one host to another, the interfaces > > > > > > exposed by > > > > > > QEMU must stay unchanged in order to guarantee a successful > > > > > > migration. > > > > > > In the case of vhost-user interface, parameters like supported > > > > > > Virtio > > > > > > feature set, max number of queues, max vring sizes,... must remain > > > > > > compatible. Indeed, the frontend not being re-initialized, no > > > > > > re-negotiation happens at migration time. > > > > > > > > > > > > For example, we have a VM that runs on host A, which has its > > > > > > vhost-user > > > > > > backend advertising VIRTIO_F_RING_INDIRECT_DESC feature. Since the > > > > > > Guest > > > > > > also support this feature, it is successfully negotiated, and guest > > > > > > transmit packets using indirect descriptor tables, that the backend > > > > > > knows to handle. > > > > > > > > > > > > At some point, the VM is being migrated to host B, which runs an > > > > > > older > > > > > > version of the backend not supporting this > > > > > > VIRTIO_F_RING_INDIRECT_DESC > > > > > > feature. The migration would break, because the Guest still have the > > > > > > VIRTIO_F_RING_INDIRECT_DESC bit sets, and the virtqueue contains > > > > > > some > > > > > > decriptors pointing to indirect tables, that backend B doesn't know > > > > > > to > > > > > > handle. > > > > > > This is just an example about Virtio features compatibility, but > > > > > > other > > > > > > backend implementation details could cause other failures. (e.g. > > > > > > configurable queues sizes) > > > > > > > > > > > > What we need is to be able to query the destination host's backend > > > > > > to > > > > > > ensure migration is possible before it is initiated. > > > > > > > > > > This remided me strongly of the issues around the virtual CPU modeling > > > > > in KVM, see > > > > > https://wiki.qemu.org/index.php/Features/CPUModels#Querying_host_capabilities > > > > > > > > > > QEMU recently gained query-cpu-model-expansion to allow capability > > > > > queries. > > > > > > > > > > Cc Eduardo accordingly. Eduardo, could you please take a look - > > > > > how is the problem solved on the KVM/VCPU side? Do the above > > > > > problem and solution for vhost look similar? > > > > > > > > (Sorry for taking so long to reply) > > > > > > > > CPU configuration in QEMU has the additional problem of features > > > > depending on host hardware and kernel capabilities (not just QEMU > > > > software capabilities). Do you have vhost-user features that > > > > depend on the host kernel or hardware too, or all of them just > > > > depend on the vhost-user backend software? > > > > > > vhost-net features depend on host kernel. > > > > > > > If it depends only on software, a solution similar to how > > > > machine-types work in QEMU sound enough. If features depend on > > > > host kernel or host hardware too, it is a bit more complex: it > > > > means you need an interface to find out if each configurable > > > > feature/version is really available on the host. > > > > > > > > (In the case of CPU models, we started with an interface that > > > > reported which CPU models were runnable on the host. But as > > > > libvirt allows enabling/disabling individual CPU features, the > > > > interface had to be extended to report which CPU features were > > > > available/unavailable on the host.) > > > > > > > > * * * > > > > > > > > Now, there's one thing that seems very different here: the > > > > guest-visible interface is not defined only by QEMU, but also by > > > > the vhost-user backend. Is that correct? > > > > > > Not exactly. As long as there are no bugs it's defined by QEMU but > > > depends on backend capabilities. Bugs in a backend could be guest > > > visible - same as kvm really. > > > > I'm a bit confused here. > > > > I will try to enumerate the steps involved in the process, for > > clarity: > > 1) Querying which features are available on a host; > > 2) Choosing a reasonable default based on what's available on the > > relevant host(s), before starting a VM; > > 3) Actually configuring what will be seen by the guest, based on > > (1), (2) (and optionally user input/configuration). > > > > Above you say that (1) on vhost-net depend on host kernel too. > > That's OK. > > > > I also understand that (2) can't be done by libvirt and QEMU > > alone, because they don't have information about the vhost-user > > backend before the VM is configured. That's OK too. > > > > However, I don't see the data flow of the configuration step (3) > > clearly. If the guest ABI is only defined by QEMU, does that > > mean configuring the guest-visible features would always be done > > through libvirt+QEMU? > > > > In other words, would the corresponding > > vhostuser_compat.virtio_features value (or other knobs that > > affect guest ABI) always flow this way: > > OVS -> libvirt -> QEMU -> vhost-user-backend -> guest > > and not directly this way: > > OVS -> vhost-user-backend -> guest > > ? > > Barring bugs, after init it works like this > > libvirt -----------------------+ > | > v > OVS -> vhost-user-backend <-> QEMU -> guest > > On device init qemu queries all features, enables > the subset configured by libvirt and also supplies > them (with possible modifications) to guest.
Good. I thought there was no interface to configure the guest-visible virtio features through libvirt, and OVS would bypass that by configuring vhost-user-backend directly. Which variables affect the set of available features, exactly? Does it depend on QEMU capabilities too on some cases, or just vhost-backend + kernel capabilities? Is there anything libvirt+QEMU need do to help implement (1) (querying host capabilities) so the rest of the stack can implement (2) (choosing what to enable based on host capabilities)? Or this can be done without QEMU and libvirt being involved at all? > > > > > > > > > > > This means QEMU won't fully control the resulting guest ABI > > > > anymore. I would really prefer if we could keep libvirt+QEMU in > > > > control of the guest ABI as usual, making QEMU configure all the > > > > guest-visible vhost-user features. But I understand this would > > > > require additional interfaces between QEMU and libvirt, and > > > > extending the libvirt APIs. > > > > > > > > So, if QEMU is really not going to control the resulting guest > > > > ABI completely, can we at least provide a mechanism which QEMU > > > > can use to ask vhost-user for guest ABI details on migration, and > > > > block migration if vhost-user was misconfigured on the > > > > destination host when migrating? > > > > > > > > > > > > > > > > > > > The below proposal has been drafted based on how Qemu manages > > > > > > machine types: > > > > > > > > > > > > Proposal > > > > > > ======== > > > > > > > > > > > > The idea is to have a table of supported version strings in OVS, > > > > > > associated to key/value pairs. Nova or any other management tool > > > > > > could > > > > > > query OVS for the list of supported versions strings for each hosts. > > > > > > By default, the latest compatibility version will be selected, but > > > > > > the > > > > > > admin can select manually an older compatibility mode in order to > > > > > > ensure > > > > > > successful migration to an older destination host. > > > > > > > > > > > > Then, Nova would add OVS's vhost-user port with adding the selected > > > > > > version (compatibility mode) as an extra parameter. > > > > > > > > > > > > Before starting the VM migration, Nova will ensure both source and > > > > > > destination hosts' vhost-user interfaces run in the same > > > > > > compatibility > > > > > > modes, and will prevent it if this is not the case. > > > > > > > > > > > > For example host A runs OVS-2.7, and host B OVS-2.6. > > > > > > Host A's OVS-2.7 has an OVS-2.6 compatibility mode (e.g. with > > > > > > indirect > > > > > > descriptors disabled), which should be selected at vhost-user port > > > > > > add > > > > > > time to ensure migration will succeed to host B. > > > > > > > > > > > > Advantage of doing so is that Nova does not need any update if new > > > > > > keys > > > > > > are introduced (i.e. it does not need to know how the new keys have > > > > > > to > > > > > > be handled), all these checks remain in OVS's vhost-user > > > > > > implementation. > > > > > > > > > > > > Ideally, we would support per vhost-user interface compatibility > > > > > > mode, > > > > > > which may have an impact also on DPDK API, as the Virtio feature > > > > > > update > > > > > > API is global, and not per port. > > > > > > > > > > > > - Implementation: > > > > > > ----------------- > > > > > > > > > > > > Goal here is just to illustrate this proposal, I'm sure you will > > > > > > have > > > > > > good suggestion to improve it. > > > > > > In OVS vhost-user library, we would introduce a new structure, for > > > > > > example (neither compiled nor tested): > > > > > > > > > > > > struct vhostuser_compat { > > > > > > char *version; > > > > > > uint64_t virtio_features; > > > > > > uint32_t max_rx_queue_sz; > > > > > > uint32_t max_nr_queues; > > > > > > }; > > > > > > > > > > > > *version* field is the compatibility version string. It could be > > > > > > something like: "upstream.ovs-dpdk.v2.6". In case for example Fedora > > > > > > adds some more patches to its package that would break migration to > > > > > > upstream version, it could have a dedicated compatibility string: > > > > > > "fc26.ovs-dpdk.v2.6". In case OVS-v2.7 does not break compatibility > > > > > > with > > > > > > previous OVS-v2.6 version, then no need to create a new entry, just > > > > > > keep > > > > > > v2.6 one. > > > > > > > > > > > > *virtio_features* field is the Virtio features set for a given > > > > > > compatibility version. When an OVS tag is to be created, it would be > > > > > > associated to a DPDK version. The Virtio features for these version > > > > > > would be stored in this field. It would allow to upgrade the DPDK > > > > > > package for example from v16.07 to v16.11 without breaking > > > > > > migration. > > > > > > In case the distribution wants to benefit from latests Virtio > > > > > > features, it would have to create a new entry to ensure migration > > > > > > won't be broken. > > > > > > > > > > > > *max_rx_queue_sz* > > > > > > *max_nr_queues* fields are just here for example, don't think this > > > > > > is > > > > > > needed today. I just want to illustrate that we have to anticipate > > > > > > other parameters than the Virtio feature set, even if not necessary > > > > > > at the moment. > > > > > > > > > > > > We create a table with different compatibility versions in OVS > > > > > > vhost-user lib: > > > > > > > > > > > > static struct vhostuser_compat vu_compat[] = { > > > > > > { > > > > > > .version = "upstream.ovs-dpdk.v2.7", > > > > > > .virtio_features = 0x12045694, > > > > > > .max_rx_queue_sz = 512, > > > > > > }, > > > > > > { > > > > > > .version = "upstream.ovs-dpdk.v2.6", > > > > > > .virtio_features = 0x10045694, > > > > > > .max_rx_queue_sz = 1024, > > > > > > }, > > > > > > }; > > > > > > > > > > > > At some time during installation, or system init, the table would be > > > > > > parsed, and compatibility version strings would be stored into the > > > > > > OVS > > > > > > database, or a new tool would be created to list these strings, or a > > > > > > config file packaged with OVS stores the list of compatibiliy > > > > > > versions. > > > > > > > > > > > > Before launching the VM, Nova will query the version strings for the > > > > > > host so that the admin can select an older compatibility mode. If > > > > > > none > > > > > > selected by the admin, then the most recent one will be used by > > > > > > default, > > > > > > and passed to the OVS's add-port command as parameter. Note that if > > > > > > no > > > > > > compatibility mode is passed to the add-port command, the most > > > > > > recent > > > > > > one is selected by OVS as default. > > > > > > > > > > > > When the vhost-user connection is initiated, OVS would know in which > > > > > > compatibility mode to init the interface, for example by > > > > > > restricting the > > > > > > support Virtio features of the interface. > > > > > > > > > > > > Cheers, > > > > > > Maxime > > > > > > > > > > > > [0]: > > > > > > > > > > > > https://mail.openvswitch.org/pipermail/ovs-dev/2017-February/328257.html > > > > > > <b2a5501c-7df7-ad2a-002f-d731c445a502 at redhat.com> > > > > > > > > -- > > > > Eduardo > > > > -- > > Eduardo -- Eduardo __________________________________________________________________________ OpenStack Development Mailing List (not for usage questions) Unsubscribe: openstack-dev-requ...@lists.openstack.org?subject:unsubscribe http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev