On 30/06/14 16:00, Matthew Rosato wrote: > Add memory information to read SCP info and add handlers for > Read Storage Element Information, Attach Storage Element, > Assign Storage and Unassign > > Signed-off-by: Matthew Rosato <mjros...@linux.vnet.ibm.com> > --- > hw/s390x/sclp.c | 259 > ++++++++++++++++++++++++++++++++++++++++++++++++++-- > target-s390x/cpu.h | 15 +++ > target-s390x/kvm.c | 5 + > 3 files changed, 273 insertions(+), 6 deletions(-) > > diff --git a/hw/s390x/sclp.c b/hw/s390x/sclp.c > index 769d7c3..936b189 100644 > --- a/hw/s390x/sclp.c > +++ b/hw/s390x/sclp.c > @@ -16,7 +16,8 @@ > #include "sysemu/kvm.h" > #include "exec/memory.h" > #include "sysemu/sysemu.h" > - > +#include "exec/address-spaces.h" > +#include "qemu/config-file.h" > #include "hw/s390x/sclp.h" > #include "hw/s390x/event-facility.h" > > @@ -33,10 +34,19 @@ static inline SCLPEventFacility *get_event_facility(void) > static void read_SCP_info(SCCB *sccb) > { > ReadInfo *read_info = (ReadInfo *) sccb; > + sclpMemoryHotplugDev *mhd = get_sclp_memory_hotplug_dev(); > CPUState *cpu; > - int shift = 0; > int cpu_count = 0; > int i = 0; > + int increment_size = 20; > + int rnsize, rnmax; > + QemuOpts *opts = qemu_opts_find(qemu_find_opts("memory"), NULL); > + int slots = qemu_opt_get_number(opts, "slots", 0); > + int max_avail_slots = s390_get_memslot_count(kvm_state); > + > + if (slots > max_avail_slots) { > + slots = max_avail_slots; > + } > > CPU_FOREACH(cpu) { > cpu_count++; > @@ -54,14 +64,235 @@ static void read_SCP_info(SCCB *sccb) > > read_info->facilities = cpu_to_be64(SCLP_HAS_CPU_INFO); > > - while ((ram_size >> (20 + shift)) > 65535) { > - shift++; > + /*
I was about to apply this series, but I think it breaks the non-ccw machine for strange memory sizes (e.g. 40001MB). The s390-virtio.c file still does a shifting via 65535. Maybe just do a change over there as well? You should also start making up your mind about migration and prepare patches. I will send cpu migration patches soon. Christian > + * The storage increment size is a multiple of 1M and is a power of 2. > + * The number of storage increments must be MAX_STORAGE_INCREMENTS or > fewer. > + */ > + while ((ram_size >> increment_size) > MAX_STORAGE_INCREMENTS) { > + increment_size++; > + } > + rnmax = ram_size >> increment_size; > + > + /* Memory Hotplug is only supported for the ccw machine type */ > + if (mhd) { > + while ((mhd->standby_mem_size >> increment_size) > > + MAX_STORAGE_INCREMENTS) { > + increment_size++; > + } > + assert(increment_size == mhd->increment_size); > + > + mhd->standby_subregion_size = MEM_SECTION_SIZE; > + /* Deduct the memory slot already used for core */ > + if (slots > 0) { > + while ((mhd->standby_subregion_size * (slots - 1) > + < mhd->standby_mem_size)) { > + mhd->standby_subregion_size = mhd->standby_subregion_size << > 1; > + } > + } > + /* > + * Initialize mapping of guest standby memory sections indicating > which > + * are and are not online. Assume all standby memory begins offline. > + */ > + if (mhd->standby_state_map == 0) { > + if (mhd->standby_mem_size % mhd->standby_subregion_size) { > + mhd->standby_state_map = g_malloc0((mhd->standby_mem_size / > + mhd->standby_subregion_size + > 1) * > + (mhd->standby_subregion_size / > + MEM_SECTION_SIZE)); > + } else { > + mhd->standby_state_map = g_malloc0(mhd->standby_mem_size / > + MEM_SECTION_SIZE); > + } > + } > + mhd->padded_ram_size = ram_size + mhd->pad_size; > + mhd->rzm = 1 << mhd->increment_size; > + rnmax = ((ram_size + mhd->standby_mem_size + mhd->pad_size) > + >> mhd->increment_size); > + > + read_info->facilities |= > cpu_to_be64(SCLP_FC_ASSIGN_ATTACH_READ_STOR); > + } > + > + rnsize = 1 << (increment_size - 20); > + if (rnsize <= 128) { > + read_info->rnsize = rnsize; > + } else { > + read_info->rnsize = 0; > + read_info->rnsize2 = cpu_to_be32(rnsize); > } > - read_info->rnmax = cpu_to_be16(ram_size >> (20 + shift)); > - read_info->rnsize = 1 << shift; > + > + if (rnmax < 0x10000) { > + read_info->rnmax = cpu_to_be16(rnmax); > + } else { > + read_info->rnmax = cpu_to_be16(0); > + read_info->rnmax2 = cpu_to_be64(rnmax); > + } > + > sccb->h.response_code = cpu_to_be16(SCLP_RC_NORMAL_READ_COMPLETION); > } > > +static void read_storage_element0_info(SCCB *sccb) > +{ > + int i, assigned; > + int subincrement_id = SCLP_STARTING_SUBINCREMENT_ID; > + ReadStorageElementInfo *storage_info = (ReadStorageElementInfo *) sccb; > + sclpMemoryHotplugDev *mhd = get_sclp_memory_hotplug_dev(); > + > + assert(mhd); > + > + if ((ram_size >> mhd->increment_size) >= 0x10000) { > + sccb->h.response_code = cpu_to_be16(SCLP_RC_SCCB_BOUNDARY_VIOLATION); > + return; > + } > + > + /* Return information regarding core memory */ > + storage_info->max_id = cpu_to_be16(mhd->standby_mem_size ? 1 : 0); > + assigned = ram_size >> mhd->increment_size; > + storage_info->assigned = cpu_to_be16(assigned); > + > + for (i = 0; i < assigned; i++) { > + storage_info->entries[i] = cpu_to_be32(subincrement_id); > + subincrement_id += SCLP_INCREMENT_UNIT; > + } > + sccb->h.response_code = cpu_to_be16(SCLP_RC_NORMAL_READ_COMPLETION); > +} > + > +static void read_storage_element1_info(SCCB *sccb) > +{ > + ReadStorageElementInfo *storage_info = (ReadStorageElementInfo *) sccb; > + sclpMemoryHotplugDev *mhd = get_sclp_memory_hotplug_dev(); > + > + assert(mhd); > + > + if ((mhd->standby_mem_size >> mhd->increment_size) >= 0x10000) { > + sccb->h.response_code = cpu_to_be16(SCLP_RC_SCCB_BOUNDARY_VIOLATION); > + return; > + } > + > + /* Return information regarding standby memory */ > + storage_info->max_id = cpu_to_be16(mhd->standby_mem_size ? 1 : 0); > + storage_info->assigned = cpu_to_be16(mhd->standby_mem_size >> > + mhd->increment_size); > + storage_info->standby = cpu_to_be16(mhd->standby_mem_size >> > + mhd->increment_size); > + sccb->h.response_code = cpu_to_be16(SCLP_RC_STANDBY_READ_COMPLETION); > +} > + > +static void attach_storage_element(SCCB *sccb, uint16_t element) > +{ > + int i, assigned, subincrement_id; > + AttachStorageElement *attach_info = (AttachStorageElement *) sccb; > + sclpMemoryHotplugDev *mhd = get_sclp_memory_hotplug_dev(); > + > + assert(mhd); > + > + if (element != 1) { > + sccb->h.response_code = cpu_to_be16(SCLP_RC_INVALID_SCLP_COMMAND); > + return; > + } > + > + assigned = mhd->standby_mem_size >> mhd->increment_size; > + attach_info->assigned = cpu_to_be16(assigned); > + subincrement_id = ((ram_size >> mhd->increment_size) << 16) > + + SCLP_STARTING_SUBINCREMENT_ID; > + for (i = 0; i < assigned; i++) { > + attach_info->entries[i] = cpu_to_be32(subincrement_id); > + subincrement_id += SCLP_INCREMENT_UNIT; > + } > + sccb->h.response_code = cpu_to_be16(SCLP_RC_NORMAL_COMPLETION); > +} > + > +static void assign_storage(SCCB *sccb) > +{ > + MemoryRegion *mr = NULL; > + uint64_t this_subregion_size; > + AssignStorage *assign_info = (AssignStorage *) sccb; > + sclpMemoryHotplugDev *mhd = get_sclp_memory_hotplug_dev(); > + assert(mhd); > + ram_addr_t assign_addr = (assign_info->rn - 1) * mhd->rzm; > + MemoryRegion *sysmem = get_system_memory(); > + > + if ((assign_addr % MEM_SECTION_SIZE == 0) && > + (assign_addr >= mhd->padded_ram_size)) { > + /* Re-use existing memory region if found */ > + mr = memory_region_find(sysmem, assign_addr, 1).mr; > + if (!mr) { > + > + MemoryRegion *standby_ram = g_new(MemoryRegion, 1); > + > + /* offset to align to standby_subregion_size for allocation */ > + ram_addr_t offset = assign_addr - > + (assign_addr - mhd->padded_ram_size) > + % mhd->standby_subregion_size; > + > + /* strlen("standby.ram") + 4 (Max of KVM_MEMORY_SLOTS) + NULL */ > + char id[16]; > + snprintf(id, 16, "standby.ram%d", > + (int)((offset - mhd->padded_ram_size) / > + mhd->standby_subregion_size) + 1); > + > + /* Allocate a subregion of the calculated standby_subregion_size > */ > + if (offset + mhd->standby_subregion_size > > + mhd->padded_ram_size + mhd->standby_mem_size) { > + this_subregion_size = mhd->padded_ram_size + > + mhd->standby_mem_size - offset; > + } else { > + this_subregion_size = mhd->standby_subregion_size; > + } > + > + memory_region_init_ram(standby_ram, NULL, id, > this_subregion_size); > + vmstate_register_ram_global(standby_ram); > + memory_region_add_subregion(sysmem, offset, standby_ram); > + } > + /* The specified subregion is no longer in standby */ > + mhd->standby_state_map[(assign_addr - mhd->padded_ram_size) > + / MEM_SECTION_SIZE] = 1; > + } > + sccb->h.response_code = cpu_to_be16(SCLP_RC_NORMAL_COMPLETION); > +} > + > +static void unassign_storage(SCCB *sccb) > +{ > + MemoryRegion *mr = NULL; > + AssignStorage *assign_info = (AssignStorage *) sccb; > + sclpMemoryHotplugDev *mhd = get_sclp_memory_hotplug_dev(); > + assert(mhd); > + ram_addr_t unassign_addr = (assign_info->rn - 1) * mhd->rzm; > + MemoryRegion *sysmem = get_system_memory(); > + > + /* if the addr is a multiple of 256 MB */ > + if ((unassign_addr % MEM_SECTION_SIZE == 0) && > + (unassign_addr >= mhd->padded_ram_size)) { > + mhd->standby_state_map[(unassign_addr - > + mhd->padded_ram_size) / MEM_SECTION_SIZE] = 0; > + > + /* find the specified memory region and destroy it */ > + mr = memory_region_find(sysmem, unassign_addr, 1).mr; > + if (mr) { > + int i; > + int is_removable = 1; > + ram_addr_t map_offset = (unassign_addr - mhd->padded_ram_size - > + (unassign_addr - mhd->padded_ram_size) > + % mhd->standby_subregion_size); > + /* Mark all affected subregions as 'standby' once again */ > + for (i = 0; > + i < (mhd->standby_subregion_size / MEM_SECTION_SIZE); > + i++) { > + > + if (mhd->standby_state_map[i + map_offset / > MEM_SECTION_SIZE]) { > + is_removable = 0; > + break; > + } > + } > + if (is_removable) { > + memory_region_del_subregion(sysmem, mr); > + memory_region_destroy(mr); > + g_free(mr); > + } > + } > + } > + sccb->h.response_code = cpu_to_be16(SCLP_RC_NORMAL_COMPLETION); > +} > + > /* Provide information about the CPU */ > static void sclp_read_cpu_info(SCCB *sccb) > { > @@ -103,6 +334,22 @@ static void sclp_execute(SCCB *sccb, uint32_t code) > case SCLP_CMDW_READ_CPU_INFO: > sclp_read_cpu_info(sccb); > break; > + case SCLP_READ_STORAGE_ELEMENT_INFO: > + if (code & 0xff00) { > + read_storage_element1_info(sccb); > + } else { > + read_storage_element0_info(sccb); > + } > + break; > + case SCLP_ATTACH_STORAGE_ELEMENT: > + attach_storage_element(sccb, (code & 0xff00) >> 8); > + break; > + case SCLP_ASSIGN_STORAGE: > + assign_storage(sccb); > + break; > + case SCLP_UNASSIGN_STORAGE: > + unassign_storage(sccb); > + break; > default: > efc->command_handler(ef, sccb, code); > break; > diff --git a/target-s390x/cpu.h b/target-s390x/cpu.h > index ba0e4b4..1c1681c 100644 > --- a/target-s390x/cpu.h > +++ b/target-s390x/cpu.h > @@ -1047,6 +1047,7 @@ static inline void cpu_inject_crw_mchk(S390CPU *cpu) > > /* from s390-virtio-ccw */ > #define MEM_SECTION_SIZE 0x10000000UL > +#define MAX_AVAIL_SLOTS 32 > > /* fpu_helper.c */ > uint32_t set_cc_nz_f32(float32 v); > @@ -1070,6 +1071,7 @@ void kvm_s390_enable_css_support(S390CPU *cpu); > int kvm_s390_assign_subch_ioeventfd(EventNotifier *notifier, uint32_t sch, > int vq, bool assign); > int kvm_s390_cpu_restart(S390CPU *cpu); > +int kvm_s390_get_memslot_count(KVMState *s); > void kvm_s390_clear_cmma_callback(void *opaque); > #else > static inline void kvm_s390_io_interrupt(uint16_t subchannel_id, > @@ -1097,6 +1099,10 @@ static inline int kvm_s390_cpu_restart(S390CPU *cpu) > static inline void kvm_s390_clear_cmma_callback(void *opaque) > { > } > +static inline int kvm_s390_get_memslot_count(KVMState *s) > +{ > + return MAX_AVAIL_SLOTS; > +} > #endif > > static inline void cmma_reset(S390CPU *cpu) > @@ -1115,6 +1121,15 @@ static inline int s390_cpu_restart(S390CPU *cpu) > return -ENOSYS; > } > > +static inline int s390_get_memslot_count(KVMState *s) > +{ > + if (kvm_enabled()) { > + return kvm_s390_get_memslot_count(s); > + } else { > + return MAX_AVAIL_SLOTS; > + } > +} > + > void s390_io_interrupt(uint16_t subchannel_id, uint16_t subchannel_nr, > uint32_t io_int_parm, uint32_t io_int_word); > void s390_crw_mchk(void); > diff --git a/target-s390x/kvm.c b/target-s390x/kvm.c > index a6e587b..4b05d48 100644 > --- a/target-s390x/kvm.c > +++ b/target-s390x/kvm.c > @@ -1283,3 +1283,8 @@ int kvm_s390_assign_subch_ioeventfd(EventNotifier > *notifier, uint32_t sch, > } > return kvm_vm_ioctl(kvm_state, KVM_IOEVENTFD, &kick); > } > + > +int kvm_s390_get_memslot_count(KVMState *s) > +{ > + return kvm_check_extension(s, KVM_CAP_NR_MEMSLOTS); > +} >