On Wed, Mar 15, 2017 at 08:49:57PM +0100, Martin Kletzander wrote: > On Wed, Mar 15, 2017 at 02:11:23PM -0300, Marcelo Tosatti wrote: > >On Wed, Mar 15, 2017 at 03:59:31PM +0100, Martin Kletzander wrote: > >>On Wed, Mar 15, 2017 at 02:23:26PM +0000, Daniel P. Berrange wrote: > >>>On Wed, Mar 15, 2017 at 03:11:26PM +0100, Martin Kletzander wrote: > >>>>On Mon, Mar 06, 2017 at 06:06:32PM +0800, Eli Qiao wrote: > >>>>> This patch adds new xml element to support cache tune as: > >>>>> > >>>>> <cputune> > >>>>> ... > >>>>> <cachetune id='1' host_id='0' type='l3' size='2816' unit='KiB' > >>>> > >>>>Again, this was already discussed, probably, I just can't find the > >>>>source of it. But host_id actually already selects what cache is > >>>>supposed to be used, so instead of type='l3' we only need scope='both' > >>>>(or data/instruction) there. Or am I missing something? What I'm > >>>>concerned about is that the host_id is mostly stable on one host (when > >>>>the previously mentioned problems are fixed), but it will make no sense > >>>>when the VM is migrated to another one. > > > >This is the same conditions as pinning a vcpu to a pcpu. > > > >So yes, it might be that you want to migrate to a host where > >a different "host ID" number is used (which might or might not > >be a different socket). > > > >> Unfortunately, the only > >>>>solution I can think of is using multiple keys to precisely describe the > >>>>bank we want (e.g. host's cpu id, cache level and scope), but that seems > >>>>very unclean. > >>> > >>>I tend to view use of this cachetune setting as being similar to > >>>using host CPU passthrough - you're intentionally trading off > >>>migratability of your guest to get a perf boost. > >>> > >>>Even without the host_id bit, this is still non-portable, as you > >>>might be requesting separate regions for code + data, but the > >>>target host of migration may only support shared regions. > > > >Then migration should fail. > > > > Yes, it will automatically fail when you can't do the same allocations, > that's easy. The difference with host_id is that you cannot foresee > whether keeping the host_id makes sense on the destination as well.
But on the destination, the user can use a different vcpu<->pcpu configuration, right? That is, you do not enforce that if the source has say vcpu0 <-> pcpu2, the destination must have vcpu0 <-> pcpu2 (the configuration for that part is different). So, the same thing can be applied to cache configuration. > >>Sure, but those are things we can check during migration. I'd be OK > >>with disabling migration (or making it --unsafe) when migrating with > >>cachetune. > > > >Migration support is required (for NFV usecase they want to migrate > >VMs around). > > > > > > So we at least need a check for the fact that either the caches are the > somehow same or destination XML is supplied. > > The other option would be what I hinted above, that is to change > host_id= to something like level, scope, and pcpu(s)/socket(s). host_id maps to L3 socket via: commit d57e3ab7e34c51a8badeea1b500bfb738d0af66e Author: Fenghua Yu <fenghua...@intel.com> Date: Sat Oct 22 06:19:50 2016 -0700 x86/intel_cacheinfo: Enable cache id in cache info Cache id is retrieved from APIC ID and CPUID leaf 4 on x86. For more details please see the section on "Cache ID Extraction Parameters" in "Intel 64 Architecture Processor Topology Enumeration". Also the documentation of the CPUID instruction in the "Intel 64 and IA-32 Architectures Software Developer's Manual" So on the destination the user has to configure host_id= of the PCPU which the guest is pinned to. > That is > fairly easy change from the code point of view, but I'm trying to think > more about the user experience. > > Martin -- libvir-list mailing list libvir-list@redhat.com https://www.redhat.com/mailman/listinfo/libvir-list