[+] Adding this email address to the conversation. (sorry for the noise)
> From: Salil Mehta > Sent: Thursday, May 16, 2024 2:36 PM > > > From: Harsh Prateek Bora <hars...@linux.ibm.com> > > Sent: Thursday, May 16, 2024 2:07 PM > > > > Hi Salil, > > > > On 5/16/24 17:42, Salil Mehta wrote: > > > Hi Harsh, > > > > > >> From: Harsh Prateek Bora <hars...@linux.ibm.com> > > >> Sent: Thursday, May 16, 2024 11:15 AM > > >> > > >> Hi Salil, > > >> > > >> Thanks for your email. > > >> Your patch 1/8 is included here based on review comments on my previous > > >> patch from one of the maintainers in the community and therefore I had > > >> kept you in CC to be aware of the desire of having this independent patch to > > >> get merged earlier even if your other patches in the series may go through > > >> further reviews. > > > > > > I really don’t know which discussion you are pointing at? Please > > > understand you are fixing a bug and we are pushing a feature which has got large series. > > > It will break the patch-set which is about to be merged. > > > > > > There will be significant overhead of testing on us for the work we > > > have been carrying forward for large time. This will be disruptive. Please dont! > > > > > > > I was referring to the review discussion on my prev patch here: > > > > https://lore.kernel.org/qemu-devel/d191d2jfar7l.2eh4s445m4...@gmail.com/ > > > Ok, I'm, not sure what this means. > > > > Although your patch was included with this series only to facilitate > > review of the additional patches depending on just one of your patch. > > > Generally you rebase your patch-set over the other and clearly state on the > cover letter that this patch-set is dependent upon such and such patch-set. > Just imagine if everyone starts to unilaterally pick up patches from each > other's patch-set it will create a chaos not only for the feature owners but > also for the maintainers. > > > > > > I am not sure what is appearing disruptive here. It is a common > > practive in the community that maintainer(s) can pick individual > > patches from the series if it has been vetted by siginificant number of reviewers. > > > Don’t you think this patch-set is asking for acceptance for a patch already > part of another patch-set which is about to be accepted and is a bigger > feature? Will it cause maintenance overhead at the last moment? Yes, of course! > > > > However, in this case, since you have mentioned to post next version soon, > > you need not worry about it as that would be the preferred version for both > > of the series. > > > Yes, but please understand we are working for the benefit of overall > community. Please cooperate here. > > > > > > > > >> > > >> I am hoping to see your v9 soon and thereafter maintainer(s) may choose to > > >> pick the latest independent patch if needs to be merged earlier. > > > > > > > > > I don’t think you are understanding what problem it is causing. For > > > your small bug fix you are causing significant delays at our end. > > > > > > > I hope I clarfied above that including your patch here doesnt delay anything. > > Hoping to see your v9 soon! > > > > Thanks > > Harsh > > > > > > Thanks > > > Salil. > > >> > > >> Thanks for your work and let's be hopeful it gets merged soon. > > >> > > >> regards, > > >> Harsh > > >> > > >> On 5/16/24 14:00, Salil Mehta wrote: > > >> > Hi Harsh, > > >> > > > >> > Thanks for your interest in the patch-set but taking away patches like > > >> > this from other series without any discussion can disrupt others work > > >> > and its acceptance on time. This is because we will have to put lot of > > >> > effort in rebasing bigger series and then testing overhead comes along > > >> > with it. > > >> > > > >> > The patch-set (from where this patch has been taken) is part of even > > >> > bigger series and there have been many people and companies toiling to > > >> > fix the bugs collectively in that series and for years. > > >> > > > >> > I'm about float the V9 version of the Arch agnostic series which this > > >> > patch is part of and you can rebase your patch-set from there. I'm > > >> > hopeful that it will get accepted in this cycle. > > >> > > > >> > > > >> > Many thanks > > >> > Salil. > > >> > > > >> >> From: Harsh Prateek Bora <hars...@linux.ibm.com> > > >> >> Sent: Thursday, May 16, 2024 6:32 AM > > >> >> > > >> >> From: Salil Mehta <salil.me...@huawei.com> > > >> >> > > >> >> KVM vCPU creation is done once during the vCPU realization when Qemu > > >> >> vCPU thread is spawned. This is common to all the architectures as of now. > > >> >> > > >> >> Hot-unplug of vCPU results in destruction of the vCPU object in QOM but > > >> >> the corresponding KVM vCPU object in the Host KVM is not destroyed as > > >> >> KVM doesn't support vCPU removal. Therefore, its representative KVM > > >> >> vCPU object/context in Qemu is parked. > > >> >> > > >> >> Refactor architecture common logic so that some APIs could be reused by > > >> >> vCPU Hotplug code of some architectures likes ARM, Loongson etc. Update > > >> >> new/old APIs with trace events instead of DPRINTF. No functional change is > > >> >> intended here. > > >> >> > > >> >> Signed-off-by: Salil Mehta <salil.me...@huawei.com> > > >> >> Reviewed-by: Gavin Shan <gs...@redhat.com> > > >> >> Tested-by: Vishnu Pajjuri <vis...@os.amperecomputing.com> > > >> >> Reviewed-by: Jonathan Cameron <jonathan.came...@huawei.com> > > >> >> Tested-by: Xianglai Li <lixiang...@loongson.cn> > > >> >> Tested-by: Miguel Luis <miguel.l...@oracle.com> > > >> >> Reviewed-by: Shaoqin Huang <shahu...@redhat.com> > > >> >> [harshpb: fixed rebase failures in include/sysemu/kvm.h] > > >> >> Signed-off-by: Harsh Prateek Bora <hars...@linux.ibm.com> > > >> >> --- > > >> >> include/sysemu/kvm.h | 15 ++++++++++ > > >> >> accel/kvm/kvm-all.c | 64 > > ++++++++++++++++++++++++++++++++--- > > >> ----- > > >> >> -- > > >> >> accel/kvm/trace-events | 5 +++- > > >> >> 3 files changed, 68 insertions(+), 16 deletions(-) > > >> >> > > >> >> diff --git a/include/sysemu/kvm.h b/include/sysemu/kvm.h > index > > >> >> eaf801bc93..fa3ec74442 100644 > > >> >> --- a/include/sysemu/kvm.h > > >> >> +++ b/include/sysemu/kvm.h > > >> >> @@ -434,6 +434,21 @@ void kvm_set_sigmask_len(KVMState *s, > > >> unsigned > > >> >> int sigmask_len); > > >> >> > > >> >> int kvm_physical_memory_addr_from_host(KVMState *s, void > > >> >> *ram_addr, > > >> >> hwaddr *phys_addr); > > >> >> +/** > > >> >> + * kvm_create_vcpu - Gets a parked KVM vCPU or creates a KVM > > >> vCPU > > >> >> + * @cpu: QOM CPUState object for which KVM vCPU has to be > > >> >> fetched/created. > > >> >> + * > > >> >> + * @returns: 0 when success, errno (<0) when failed. > > >> >> + */ > > >> >> +int kvm_create_vcpu(CPUState *cpu); > > >> >> + > > >> >> +/** > > >> >> + * kvm_park_vcpu - Park QEMU KVM vCPU context > > >> >> + * @cpu: QOM CPUState object for which QEMU KVM vCPU context has to > > >> >> be parked. > > >> >> + * > > >> >> + * @returns: none > > >> >> + */ > > >> >> +void kvm_park_vcpu(CPUState *cpu); > > >> >> > > >> >> #endif /* COMPILING_PER_TARGET */ > > >> >> > > >> >> diff --git a/accel/kvm/kvm-all.c b/accel/kvm/kvm-all.c index > > >> >> d7281b93f3..30d42847de 100644 > > >> >> --- a/accel/kvm/kvm-all.c > > >> >> +++ b/accel/kvm/kvm-all.c > > >> >> @@ -128,6 +128,7 @@ static QemuMutex kml_slots_lock; > #define > > >> >> kvm_slots_unlock() qemu_mutex_unlock(&kml_slots_lock) > > >> >> > > >> >> static void kvm_slot_init_dirty_bitmap(KVMSlot *mem); > > >> >> +static int kvm_get_vcpu(KVMState *s, unsigned long vcpu_id); > > >> >> > > >> >> static inline void kvm_resample_fd_remove(int gsi) { @@ - > 340,14 > > >> +341,53 > > >> >> @@ err: > > >> >> return ret; > > >> >> } > > >> >> > > >> >> +void kvm_park_vcpu(CPUState *cpu) > > >> >> +{ > > >> >> + struct KVMParkedVcpu *vcpu; > > >> >> + > > >> >> + trace_kvm_park_vcpu(cpu->cpu_index, kvm_arch_vcpu_id(cpu)); > > >> >> + > > >> >> + vcpu = g_malloc0(sizeof(*vcpu)); > > >> >> + vcpu->vcpu_id = kvm_arch_vcpu_id(cpu); > > >> >> + vcpu->kvm_fd = cpu->kvm_fd; > > >> >> + QLIST_INSERT_HEAD(&kvm_state->kvm_parked_vcpus, vcpu, node); } > > >> >> + > > >> >> +int kvm_create_vcpu(CPUState *cpu) > > >> >> +{ > > >> >> + unsigned long vcpu_id = kvm_arch_vcpu_id(cpu); > > >> >> + KVMState *s = kvm_state; > > >> >> + int kvm_fd; > > >> >> + > > >> >> + trace_kvm_create_vcpu(cpu->cpu_index, kvm_arch_vcpu_id(cpu)); > > >> >> + > > >> >> + /* check if the KVM vCPU already exist but is parked */ > > >> >> + kvm_fd = kvm_get_vcpu(s, vcpu_id); > > >> >> + if (kvm_fd < 0) { > > >> >> + /* vCPU not parked: create a new KVM vCPU */ > > >> >> + kvm_fd = kvm_vm_ioctl(s, KVM_CREATE_VCPU, vcpu_id); > > >> >> + if (kvm_fd < 0) { > > >> >> + error_report("KVM_CREATE_VCPU IOCTL failed for vCPU %lu", vcpu_id); > > >> >> + return kvm_fd; > > >> >> + } > > >> >> + } > > >> >> + > > >> >> + cpu->kvm_fd = kvm_fd; > > >> >> + cpu->kvm_state = s; > > >> >> + cpu->vcpu_dirty = true; > > >> >> + cpu->dirty_pages = 0; > > >> >> + cpu->throttle_us_per_full = 0; > > >> >> + > > >> >> + return 0; > > >> >> +} > > >> >> + > > >> >> static int do_kvm_destroy_vcpu(CPUState *cpu) { > > >> >> KVMState *s = kvm_state; > > >> >> long mmap_size; > > >> >> - struct KVMParkedVcpu *vcpu = NULL; > > >> >> int ret = 0; > > >> >> > > >> >> - trace_kvm_destroy_vcpu(); > > >> >> + trace_kvm_destroy_vcpu(cpu->cpu_index, > > >> kvm_arch_vcpu_id(cpu)); > > >> >> > > >> >> ret = kvm_arch_destroy_vcpu(cpu); > > >> >> if (ret < 0) { > > >> >> @@ -373,10 +413,7 @@ static int do_kvm_destroy_vcpu(CPUState *cpu) > > >> >> } > > >> >> } > > >> >> > > >> >> - vcpu = g_malloc0(sizeof(*vcpu)); > > >> >> - vcpu->vcpu_id = kvm_arch_vcpu_id(cpu); > > >> >> - vcpu->kvm_fd = cpu->kvm_fd; > > >> >> - QLIST_INSERT_HEAD(&kvm_state->kvm_parked_vcpus, vcpu, > > >> node); > > >> >> + kvm_park_vcpu(cpu); > > >> >> err: > > >> >> return ret; > > >> >> } > > >> >> @@ -397,6 +434,8 @@ static int kvm_get_vcpu(KVMState *s, > > unsigned > > >> long > > >> >> vcpu_id) > > >> >> if (cpu->vcpu_id == vcpu_id) { > > >> >> int kvm_fd; > > >> >> > > >> >> + trace_kvm_get_vcpu(vcpu_id); > > >> >> + > > >> >> QLIST_REMOVE(cpu, node); > > >> >> kvm_fd = cpu->kvm_fd; > > >> >> g_free(cpu); > > >> >> @@ -404,7 +443,7 @@ static int kvm_get_vcpu(KVMState *s, > > unsigned > > >> long > > >> >> vcpu_id) > > >> >> } > > >> >> } > > >> >> > > >> >> - return kvm_vm_ioctl(s, KVM_CREATE_VCPU, (void > *)vcpu_id); > > >> >> + return -ENOENT; > > >> >> } > > >> >> > > >> >> int kvm_init_vcpu(CPUState *cpu, Error **errp) @@ -415,19 > > +454,14 > > >> @@ > > >> >> int kvm_init_vcpu(CPUState *cpu, Error **errp) > > >> >> > > >> >> trace_kvm_init_vcpu(cpu->cpu_index, > kvm_arch_vcpu_id(cpu)); > > >> >> > > >> >> - ret = kvm_get_vcpu(s, kvm_arch_vcpu_id(cpu)); > > >> >> + ret = kvm_create_vcpu(cpu); > > >> >> if (ret < 0) { > > >> >> - error_setg_errno(errp, -ret, "kvm_init_vcpu: kvm_get_vcpu > > failed > > >> >> (%lu)", > > >> >> + error_setg_errno(errp, -ret, > > >> >> + "kvm_init_vcpu: kvm_create_vcpu failed (%lu)", > > >> >> kvm_arch_vcpu_id(cpu)); > > >> >> goto err; > > >> >> } > > >> >> > > >> >> - cpu->kvm_fd = ret; > > >> >> - cpu->kvm_state = s; > > >> >> - cpu->vcpu_dirty = true; > > >> >> - cpu->dirty_pages = 0; > > >> >> - cpu->throttle_us_per_full = 0; > > >> >> - > > >> >> mmap_size = kvm_ioctl(s, KVM_GET_VCPU_MMAP_SIZE, 0); > > >> >> if (mmap_size < 0) { > > >> >> ret = mmap_size; > > >> >> diff --git a/accel/kvm/trace-events b/accel/kvm/trace-events > index > > >> >> 681ccb667d..75c1724e78 100644 > > >> >> --- a/accel/kvm/trace-events > > >> >> +++ b/accel/kvm/trace-events > > >> >> @@ -9,6 +9,10 @@ kvm_device_ioctl(int fd, int type, void *arg) > > "dev fd > > >> %d, > > >> >> type 0x%x, arg %p" > > >> >> kvm_failed_reg_get(uint64_t id, const char *msg) "Warning: > > Unable to > > >> >> retrieve ONEREG %" PRIu64 " from KVM: %s" > > >> >> kvm_failed_reg_set(uint64_t id, const char *msg) "Warning: > > Unable to > > >> set > > >> >> ONEREG %" PRIu64 " to KVM: %s" > > >> >> kvm_init_vcpu(int cpu_index, unsigned long arch_cpu_id) > "index: > > %d > > >> id: > > >> >> %lu" > > >> >> +kvm_create_vcpu(int cpu_index, unsigned long arch_cpu_id) > > "index: > > >> %d > > >> >> id: %lu" > > >> >> +kvm_get_vcpu(unsigned long arch_cpu_id) "id: %lu" > > >> >> +kvm_destroy_vcpu(int cpu_index, unsigned long arch_cpu_id) > > "index: > > >> %d > > >> >> id: %lu" > > >> >> +kvm_park_vcpu(int cpu_index, unsigned long arch_cpu_id) > > "index: %d > > >> id: > > >> >> %lu" > > >> >> kvm_irqchip_commit_routes(void) "" > > >> >> kvm_irqchip_add_msi_route(char *name, int vector, int virq) "dev > > %s > > >> >> vector %d virq %d" > > >> >> kvm_irqchip_update_msi_route(int virq) "Updating MSI route > > >> virq=%d" > > >> >> @@ -25,7 +29,6 @@ kvm_dirty_ring_reaper(const char *s) "%s" > > >> >> kvm_dirty_ring_reap(uint64_t count, int64_t t) "reaped %"PRIu64" > > >> pages > > >> >> (took %"PRIi64" us)" > > >> >> kvm_dirty_ring_reaper_kick(const char *reason) "%s" > > >> >> kvm_dirty_ring_flush(int finished) "%d" > > >> >> -kvm_destroy_vcpu(void) "" > > >> >> kvm_failed_get_vcpu_mmap_size(void) "" > > >> >> kvm_cpu_exec(void) "" > > >> >> kvm_interrupt_exit_request(void) "" > > >> >> -- > > >> >> 2.39.3 > > >> >