On Fri, Aug 5, 2016 at 7:26 PM, Peter Krempa <pkre...@redhat.com> wrote:
> For hotplug purposes it's necessary to retrieve data using > query-hotpluggable-cpus while the old query-cpus API report thread IDs > and order of hotplug. > > This patch adds code that merges the data using a rather non-trivial > algorithm and fills the data to the qemuMonitorCPUInfo structure for > adding to appropriate place in the domain definition. > --- > src/qemu/qemu_domain.c | 2 +- > src/qemu/qemu_monitor.c | 197 ++++++++++++++++++++++++++++++ > ++++++++++++++++-- > src/qemu/qemu_monitor.h | 23 +++++- > 3 files changed, 212 insertions(+), 10 deletions(-) > > diff --git a/src/qemu/qemu_domain.c b/src/qemu/qemu_domain.c > index 0201a4a..1be3ede 100644 > --- a/src/qemu/qemu_domain.c > +++ b/src/qemu/qemu_domain.c > @@ -5776,7 +5776,7 @@ qemuDomainRefreshVcpuInfo(virQEMUDriverPtr driver, > if (qemuDomainObjEnterMonitorAsync(driver, vm, asyncJob) < 0) > return -1; > > - rc = qemuMonitorGetCPUInfo(qemuDomainGetMonitor(vm), &info, > maxvcpus); > + rc = qemuMonitorGetCPUInfo(qemuDomainGetMonitor(vm), &info, > maxvcpus, false); > > if (qemuDomainObjExitMonitor(driver, vm) < 0) > goto cleanup; > diff --git a/src/qemu/qemu_monitor.c b/src/qemu/qemu_monitor.c > index f87f431..12a7fa6 100644 > --- a/src/qemu/qemu_monitor.c > +++ b/src/qemu/qemu_monitor.c > @@ -1656,13 +1656,36 @@ qemuMonitorSystemReset(qemuMonitorPtr mon) > } > > > +static void > +qemuMonitorCPUInfoClear(qemuMonitorCPUInfoPtr cpus, > + size_t ncpus) > +{ > + size_t i; > + > + for (i = 0; i < ncpus; i++) { > + cpus[i].id = 0; > + cpus[i].socket_id = -1; > + cpus[i].core_id = -1; > + cpus[i].thread_id = -1; > + cpus[i].vcpus = 0; > + cpus[i].tid = 0; > + > + VIR_FREE(cpus[i].qom_path); > + VIR_FREE(cpus[i].alias); > + VIR_FREE(cpus[i].type); > + } > +} > + > + > void > qemuMonitorCPUInfoFree(qemuMonitorCPUInfoPtr cpus, > - size_t ncpus ATTRIBUTE_UNUSED) > + size_t ncpus) > { > if (!cpus) > return; > > + qemuMonitorCPUInfoClear(cpus, ncpus); > + > VIR_FREE(cpus); > } > > @@ -1683,10 +1706,148 @@ qemuMonitorQueryCpusFree(struct > qemuMonitorQueryCpusEntry *entries, > > > /** > + * Legacy approach doesn't allow out of order cpus, thus no complex > matching > + * algorithm is necessary */ > +static void > +qemuMonitorGetCPUInfoLegacy(struct qemuMonitorQueryCpusEntry *cpuentries, > + size_t ncpuentries, > + qemuMonitorCPUInfoPtr vcpus, > + size_t maxvcpus) > +{ > + size_t i; > + > + for (i = 0; i < ncpuentries && i < maxvcpus; i++) { > + if (i < ncpuentries) > + vcpus[i].tid = cpuentries[i].tid; > + > + /* for legacy hotplug to work we need to fake the vcpu count > added by > + * enabling a given vcpu */ > + vcpus[i].vcpus = 1; > + } > +} > + > + > +/** > + * qemuMonitorGetCPUInfoHotplug: > + * > + * This function stitches together data retrieved via > query-hotpluggable-cpus > + * which returns entities on the hotpluggable level (which may describe > more > + * than one guest logical vcpu) with the output of query-cpus, having an > entry > + * per enabled guest logical vcpu. > + * > + * query-hotpluggable-cpus conveys following information: > + * - topology information and number of logical vcpus this entry creates > + * - device type name of the entry that needs to be used when hotplugging > + * - qom path in qemu which can be used to map the entry against > query-cpus > + * > + * query-cpus conveys following information: > + * - thread id of a given guest logical vcpu > + * - order in which the vcpus were inserted > + * - qom path to allow mapping the two together > + * > + * The libvirt's internal structure has an entry for each possible (even > + * disabled) guest vcpu. The purpose is to map the data together so that > we are > + * certain of the thread id mapping and the information required for vcpu > + * hotplug. > + * > + * This function returns 0 on success and -1 on error, but does not report > + * libvirt errors so that fallback approach can be used. > + */ > +static int > +qemuMonitorGetCPUInfoHotplug(struct qemuMonitorQueryHotpluggableCpusEntry > *hotplugvcpus, > + size_t nhotplugvcpus, > + struct qemuMonitorQueryCpusEntry *cpuentries, > + size_t ncpuentries, > + qemuMonitorCPUInfoPtr vcpus, > + size_t maxvcpus) > +{ > + int order = 1; > + size_t totalvcpus = 0; > + size_t i; > + size_t j; > + > + /* ensure that the total vcpu count reported by > query-hotpluggable-cpus equals > + * to the libvirt maximum cpu count */ > + for (i = 0; i < nhotplugvcpus; i++) > + totalvcpus += hotplugvcpus[i].vcpus; > + > + if (totalvcpus != maxvcpus) { > + VIR_DEBUG("expected '%zu' total vcpus got '%zu'", maxvcpus, > totalvcpus); > + return -1; > + } > + > + /* Note the order in which the hotpluggable entities are inserted by > + * matching them to the query-cpus entries */ > + for (i = 0; i < ncpuentries; i++) { > + for (j = 0; j < nhotplugvcpus; j++) { > + if (!cpuentries[i].qom_path || > + !hotplugvcpus[j].qom_path || > + !STRPREFIX(cpuentries[i].qom_path, > hotplugvcpus[j].qom_path)) > + continue; > + > + /* add ordering info for hotpluggable entries */ > + if (hotplugvcpus[j].enable_id == 0) > + hotplugvcpus[j].enable_id = order++; > + > + break; > + } > + } > + > + /* transfer appropriate data from the hotpluggable list to > corresponding > + * entries. the entries returned by qemu may in fact describe multiple > + * logical vcpus in the guest */ > + j = 0; > + for (i = 0; i < nhotplugvcpus; i++) { > + vcpus[j].socket_id = hotplugvcpus[i].socket_id; > + vcpus[j].core_id = hotplugvcpus[i].core_id; > + vcpus[j].thread_id = hotplugvcpus[i].thread_id; > + vcpus[j].vcpus = hotplugvcpus[i].vcpus; > + VIR_STEAL_PTR(vcpus[j].qom_path, hotplugvcpus[i].qom_path); > + VIR_STEAL_PTR(vcpus[j].alias, hotplugvcpus[i].alias); > + VIR_STEAL_PTR(vcpus[j].type, hotplugvcpus[i].type); > + vcpus[j].id = hotplugvcpus[i].enable_id; > + > + /* skip over vcpu entries covered by this hotpluggable entry */ > + j += hotplugvcpus[i].vcpus; > + } > + > + /* match entries from query cpus to the output array taking into > account > + * multi-vcpu objects */ > + for (j = 0; j < ncpuentries; j++) { > + /* find the correct entry or beginning of group of entries */ > + for (i = 0; i < maxvcpus; i++) { > + if (cpuentries[j].qom_path && vcpus[i].qom_path && > + STRPREFIX(cpuentries[j].qom_path, vcpus[i].qom_path)) > + break; > + } > + > + if (i == maxvcpus) { > + VIR_DEBUG("too many query-vcpus entries for a given " > s/query-vcpus/query-cpus + "query-hotpluggable-cpus entry"); > + return -1; > + } > + > + if (vcpus[i].vcpus != 1) { > + /* find a possibly empty vcpu thread for core granularity > systems */ > + for (; i < maxvcpus; i++) { > + if (vcpus[i].tid == 0) > + break; > + } > + } > + > + vcpus[i].tid = cpuentries[j].tid; > + } > + > + return 0; > +} > + > + > +/** > * qemuMonitorGetCPUInfo: > * @mon: monitor > * @cpus: pointer filled by array of qemuMonitorCPUInfo structures > * @maxvcpus: total possible number of vcpus > + * @hotplug: query data relevant for hotplug support > * > * Detects VCPU information. If qemu doesn't support or fails reporting > * information this function will return success as other parts of libvirt > @@ -1698,20 +1859,32 @@ qemuMonitorQueryCpusFree(struct > qemuMonitorQueryCpusEntry *entries, > int > qemuMonitorGetCPUInfo(qemuMonitorPtr mon, > qemuMonitorCPUInfoPtr *vcpus, > - size_t maxvcpus) > + size_t maxvcpus, > + bool hotplug) > { > - qemuMonitorCPUInfoPtr info = NULL; > + struct qemuMonitorQueryHotpluggableCpusEntry *hotplugcpus = NULL; > + size_t nhotplugcpus = 0; > struct qemuMonitorQueryCpusEntry *cpuentries = NULL; > size_t ncpuentries = 0; > - size_t i; > int ret = -1; > int rc; > + qemuMonitorCPUInfoPtr info = NULL; > > - QEMU_CHECK_MONITOR(mon); > + if (hotplug) > + QEMU_CHECK_MONITOR_JSON(mon); > + else > + QEMU_CHECK_MONITOR(mon); > > if (VIR_ALLOC_N(info, maxvcpus) < 0) > return -1; > > + /* initialize a few non-zero defaults */ > + qemuMonitorCPUInfoClear(info, maxvcpus); > + > + if (hotplug && > + (qemuMonitorJSONGetHotpluggableCPUs(mon, &hotplugcpus, > &nhotplugcpus)) < 0) > + goto cleanup; > + > if (mon->json) > rc = qemuMonitorJSONQueryCPUs(mon, &cpuentries, &ncpuentries); > else > @@ -1726,15 +1899,23 @@ qemuMonitorGetCPUInfo(qemuMonitorPtr mon, > goto cleanup; > } > > - for (i = 0; i < ncpuentries; i++) > - info[i].tid = cpuentries[i].tid; > + if (!hotplugcpus || > + qemuMonitorGetCPUInfoHotplug(hotplugcpus, nhotplugcpus, > + cpuentries, ncpuentries, > + info, maxvcpus) < 0) { > + /* Fallback to the legacy algorithm. Hotplug paths will make sure > that > + * the apropriate data is present */ > + qemuMonitorCPUInfoClear(info, maxvcpus); > + qemuMonitorGetCPUInfoLegacy(cpuentries, ncpuentries, info, > maxvcpus); > + } > > VIR_STEAL_PTR(*vcpus, info); > ret = 0; > > cleanup: > - qemuMonitorCPUInfoFree(info, maxvcpus); > + qemuMonitorQueryHotpluggableCpusFree(hotplugcpus, nhotplugcpus); > qemuMonitorQueryCpusFree(cpuentries, ncpuentries); > + qemuMonitorCPUInfoFree(info, maxvcpus); > return ret; > } > > diff --git a/src/qemu/qemu_monitor.h b/src/qemu/qemu_monitor.h > index 58f8327..b838725 100644 > --- a/src/qemu/qemu_monitor.h > +++ b/src/qemu/qemu_monitor.h > @@ -409,6 +409,9 @@ struct qemuMonitorQueryHotpluggableCpusEntry { > int socket_id; > int core_id; > int thread_id; > + > + /* internal data */ > + int enable_id; > }; > void qemuMonitorQueryHotpluggableCpusFree(struct > qemuMonitorQueryHotpluggableCpusEntry *entries, > size_t nentries); > @@ -416,6 +419,23 @@ void qemuMonitorQueryHotpluggableCpusFree(struct > qemuMonitorQueryHotpluggableCpu > > struct _qemuMonitorCPUInfo { > pid_t tid; > + int id; /* order of enabling of the given cpu */ > + > + /* topology info for hotplug purposes. Hotplug of given vcpu > impossible if > + * all entries are -1 */ > + int socket_id; > + int core_id; > + int thread_id; > + unsigned int vcpus; /* number of vcpus added if given entry is > hotplugged */ > + > + /* name of the qemu type to add in case of hotplug */ > + char *type; > + > + /* alias of an hotpluggable entry. Entries with alias can be > hot-unplugged */ > + char *alias; > + > + /* internal for use in the matching code */ > + char *qom_path; > }; > typedef struct _qemuMonitorCPUInfo qemuMonitorCPUInfo; > typedef qemuMonitorCPUInfo *qemuMonitorCPUInfoPtr; > @@ -424,7 +444,8 @@ void qemuMonitorCPUInfoFree(qemuMonitorCPUInfoPtr > list, > size_t nitems); > int qemuMonitorGetCPUInfo(qemuMonitorPtr mon, > qemuMonitorCPUInfoPtr *vcpus, > - size_t maxvcpus); > + size_t maxvcpus, > + bool hotplug); > > int qemuMonitorGetVirtType(qemuMonitorPtr mon, > virDomainVirtType *virtType); > -- > 2.9.2 > > -- > libvir-list mailing list > libvir-list@redhat.com > https://www.redhat.com/mailman/listinfo/libvir-list >
-- libvir-list mailing list libvir-list@redhat.com https://www.redhat.com/mailman/listinfo/libvir-list