Message ID | 20180416085512.4011-1-bharata@linux.vnet.ibm.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
On Mon, Apr 16, 2018 at 02:25:12PM +0530, Bharata B Rao wrote: > The new property ibm,dynamic-memory-v2 allows memory to be represented > in a more compact manner in device tree. > > Signed-off-by: Bharata B Rao <bharata@linux.vnet.ibm.com> > --- > v2 - https://lists.nongnu.org/archive/html/qemu-ppc/2018-04/msg00052.html > Changes in v3: > - Addressed David Gibson's review comments: use of CamelCase, separate > helper for populating drconf_cell queue entries, removed the user > configurability of drmem_v2 machine property. > > docs/specs/ppc-spapr-hotplug.txt | 19 ++++ > hw/ppc/spapr.c | 233 +++++++++++++++++++++++++++++++-------- > include/hw/ppc/spapr.h | 1 + > include/hw/ppc/spapr_ovec.h | 1 + > 4 files changed, 207 insertions(+), 47 deletions(-) > > diff --git a/docs/specs/ppc-spapr-hotplug.txt b/docs/specs/ppc-spapr-hotplug.txt > index f57e2a09c6..cc7833108e 100644 > --- a/docs/specs/ppc-spapr-hotplug.txt > +++ b/docs/specs/ppc-spapr-hotplug.txt > @@ -387,4 +387,23 @@ Each LMB list entry consists of the following elements: > - A 32bit flags word. The bit at bit position 0x00000008 defines whether > the LMB is assigned to the the partition as of boot time. > > +ibm,dynamic-memory-v2 > + > +This property describes the dynamically reconfigurable memory. This is > +an alternate and newer way to describe dyanamically reconfigurable memory. > +It is a property encoded array that has an integer N (the number of > +LMB set entries) followed by N LMB set entries. There is an LMB set entry > +for each sequential group of LMBs that share common attributes. > + > +Each LMB set entry consists of the following elements: > + > +- Number of sequential LMBs in the entry represented by a 32bit integer. > +- Logical address of the first LMB in the set encoded as a 64bit integer. > +- DRC index of the first LMB in the set. > +- Associativity list index that is used as an index into > + ibm,associativity-lookup-arrays property described earlier. This > + is used to retrieve the right associativity list to be used for all > + the LMBs in this set. > +- A 32bit flags word that applies to all the LMBs in the set. > + > [1] http://thread.gmane.org/gmane.linux.ports.ppc.embedded/75350/focus=106867 > diff --git a/hw/ppc/spapr.c b/hw/ppc/spapr.c > index 3ffadd6ac7..fe327f3c1b 100644 > --- a/hw/ppc/spapr.c > +++ b/hw/ppc/spapr.c > @@ -669,63 +669,136 @@ static uint32_t spapr_pc_dimm_node(MemoryDeviceInfoList *list, ram_addr_t addr) > return -1; > } > > -/* > - * Adds ibm,dynamic-reconfiguration-memory node. > - * Refer to docs/specs/ppc-spapr-hotplug.txt for the documentation > - * of this device tree node. > - */ > -static int spapr_populate_drconf_memory(sPAPRMachineState *spapr, void *fdt) > +struct sPAPRDrconfCellV2 { > + uint32_t seq_lmbs; > + uint64_t base_addr; > + uint32_t drc_index; > + uint32_t aa_index; > + uint32_t flags; > +} __attribute__((packed)); Should be QEMU_PACKED to at least allow for the possibility of other compilers. > + > +/* 6 uint32_t entries in sPAPRDrconfCellV2 */ > +#define SPAPR_NR_DRCONF_CELL_ENTRIES 6 Please use sizeof, as noted before. > + > +typedef struct DrconfCellQueue { > + struct sPAPRDrconfCellV2 cell; > + QSIMPLEQ_ENTRY(DrconfCellQueue) entry; > +} DrconfCellQueue; > + > +QSIMPLEQ_HEAD(, DrconfCellQueue) drconf_queue > + = QSIMPLEQ_HEAD_INITIALIZER(drconf_queue); > + > +static void spapr_add_drconf_cell(uint32_t seq_lmbs, uint64_t base_addr, > + uint32_t drc_index, uint32_t aa_index, > + uint32_t flags) > +{ > + DrconfCellQueue *elem; > + > + elem = g_malloc0(sizeof(*elem)); > + elem->cell.seq_lmbs = cpu_to_be32(seq_lmbs); > + elem->cell.base_addr = cpu_to_be64(base_addr); > + elem->cell.drc_index = cpu_to_be32(drc_index); > + elem->cell.aa_index = cpu_to_be32(aa_index); > + elem->cell.flags = cpu_to_be32(flags); > + QSIMPLEQ_INSERT_TAIL(&drconf_queue, elem, entry); Don't use a global, pass the queue as a parameter. Or alternatively have the helper just return the allocated and initialized element to add to the queue in the caller. > +} > + > +/* ibm,dynamic-memory-v2 */ > +static int spapr_populate_drmem_v2(sPAPRMachineState *spapr, void *fdt, > + int offset, MemoryDeviceInfoList *dimms) > { > - MachineState *machine = MACHINE(spapr); > - int ret, i, offset; > - uint64_t lmb_size = SPAPR_MEMORY_BLOCK_SIZE; > - uint32_t prop_lmb_size[] = {0, cpu_to_be32(lmb_size)}; > - uint32_t hotplug_lmb_start = spapr->hotplug_memory.base / lmb_size; > - uint32_t nr_lmbs = (spapr->hotplug_memory.base + > - memory_region_size(&spapr->hotplug_memory.mr)) / > - lmb_size; > uint32_t *int_buf, *cur_index, buf_len; > - int nr_nodes = nb_numa_nodes ? nb_numa_nodes : 1; > - MemoryDeviceInfoList *dimms = NULL; > + int ret; > + uint64_t lmb_size = SPAPR_MEMORY_BLOCK_SIZE; > + uint64_t addr, cur_addr, size; > + uint32_t nr_boot_lmbs = (spapr->hotplug_memory.base / lmb_size); > + uint64_t mem_end = spapr->hotplug_memory.base + > + memory_region_size(&spapr->hotplug_memory.mr); > + uint32_t node, nr_entries = 0; > + sPAPRDRConnector *drc; > + DrconfCellQueue *elem, *next; > + MemoryDeviceInfoList *info; > > - /* > - * Don't create the node if there is no hotpluggable memory > - */ > - if (machine->ram_size == machine->maxram_size) { > - return 0; > - } > + /* Entry to cover RAM and the gap area */ > + spapr_add_drconf_cell(nr_boot_lmbs, 0, 0, -1, > + SPAPR_LMB_FLAGS_RESERVED | > + SPAPR_LMB_FLAGS_DRC_INVALID); > + nr_entries++; > > - /* > - * Allocate enough buffer size to fit in ibm,dynamic-memory > - * or ibm,associativity-lookup-arrays > - */ > - buf_len = MAX(nr_lmbs * SPAPR_DR_LMB_LIST_ENTRY_SIZE + 1, nr_nodes * 4 + 2) > - * sizeof(uint32_t); > - cur_index = int_buf = g_malloc0(buf_len); > + cur_addr = spapr->hotplug_memory.base; > + for (info = dimms; info; info = info->next) { > + PCDIMMDeviceInfo *di = info->value->u.dimm.data; > > - offset = fdt_add_subnode(fdt, 0, "ibm,dynamic-reconfiguration-memory"); > + addr = di->addr; > + size = di->size; > + node = di->node; > > - ret = fdt_setprop(fdt, offset, "ibm,lmb-size", prop_lmb_size, > - sizeof(prop_lmb_size)); > - if (ret < 0) { > - goto out; > + /* Entry for hot-pluggable area */ > + if (cur_addr < addr) { > + drc = spapr_drc_by_id(TYPE_SPAPR_DRC_LMB, cur_addr / lmb_size); > + g_assert(drc); > + spapr_add_drconf_cell((addr - cur_addr) / lmb_size, > + cur_addr, spapr_drc_index(drc), > + -1, 0); > + nr_entries++; > + } > + > + /* Entry for DIMM */ > + drc = spapr_drc_by_id(TYPE_SPAPR_DRC_LMB, addr / lmb_size); > + g_assert(drc); > + spapr_add_drconf_cell(size / lmb_size, addr, spapr_drc_index(drc), > + node, SPAPR_LMB_FLAGS_ASSIGNED); > + nr_entries++; > + cur_addr = addr + size; > } > > - ret = fdt_setprop_cell(fdt, offset, "ibm,memory-flags-mask", 0xff); > - if (ret < 0) { > - goto out; > + /* Entry for remaining hotpluggable area */ > + if (cur_addr < mem_end) { > + drc = spapr_drc_by_id(TYPE_SPAPR_DRC_LMB, cur_addr / lmb_size); > + g_assert(drc); > + spapr_add_drconf_cell((mem_end - cur_addr) / lmb_size, cur_addr, > + spapr_drc_index(drc), -1, 0); > + nr_entries++; > } > > - ret = fdt_setprop_cell(fdt, offset, "ibm,memory-preservation-time", 0x0); > - if (ret < 0) { > - goto out; > + buf_len = (nr_entries * SPAPR_NR_DRCONF_CELL_ENTRIES + 1) * > + sizeof(uint32_t); > + int_buf = cur_index = g_malloc0(buf_len); > + int_buf[0] = cpu_to_be32(nr_entries); > + cur_index++; > + QSIMPLEQ_FOREACH_SAFE(elem, &drconf_queue, entry, next) { > + memcpy(cur_index, &elem->cell, > + SPAPR_NR_DRCONF_CELL_ENTRIES * sizeof(uint32_t)); > + cur_index += SPAPR_NR_DRCONF_CELL_ENTRIES; > + QSIMPLEQ_REMOVE(&drconf_queue, elem, DrconfCellQueue, entry); > + g_free(elem); > } > > - if (hotplug_lmb_start) { > - dimms = qmp_pc_dimm_device_list(); > + ret = fdt_setprop(fdt, offset, "ibm,dynamic-memory-v2", int_buf, buf_len); > + g_free(int_buf); > + if (ret < 0) { > + return -1; > } > + return 0; > +} > + > +/* ibm,dynamic-memory */ > +static int spapr_populate_drmem_v1(sPAPRMachineState *spapr, void *fdt, > + int offset, MemoryDeviceInfoList *dimms) > +{ > + int i, ret; > + uint64_t lmb_size = SPAPR_MEMORY_BLOCK_SIZE; > + uint32_t hotplug_lmb_start = spapr->hotplug_memory.base / lmb_size; > + uint32_t nr_lmbs = (spapr->hotplug_memory.base + > + memory_region_size(&spapr->hotplug_memory.mr)) / > + lmb_size; > + uint32_t *int_buf, *cur_index, buf_len; > > - /* ibm,dynamic-memory */ > + /* > + * Allocate enough buffer size to fit in ibm,dynamic-memory > + */ > + buf_len = (nr_lmbs * SPAPR_DR_LMB_LIST_ENTRY_SIZE + 1) * sizeof(uint32_t); > + cur_index = int_buf = g_malloc0(buf_len); > int_buf[0] = cpu_to_be32(nr_lmbs); > cur_index++; > for (i = 0; i < nr_lmbs; i++) { > @@ -765,13 +838,71 @@ static int spapr_populate_drconf_memory(sPAPRMachineState *spapr, void *fdt) > > cur_index += SPAPR_DR_LMB_LIST_ENTRY_SIZE; > } > - qapi_free_MemoryDeviceInfoList(dimms); > ret = fdt_setprop(fdt, offset, "ibm,dynamic-memory", int_buf, buf_len); > + g_free(int_buf); > if (ret < 0) { > - goto out; > + return -1; > + } > + return 0; > +} > + > +/* > + * Adds ibm,dynamic-reconfiguration-memory node. > + * Refer to docs/specs/ppc-spapr-hotplug.txt for the documentation > + * of this device tree node. > + */ > +static int spapr_populate_drconf_memory(sPAPRMachineState *spapr, void *fdt) > +{ > + MachineState *machine = MACHINE(spapr); > + int ret, i, offset; > + uint64_t lmb_size = SPAPR_MEMORY_BLOCK_SIZE; > + uint32_t prop_lmb_size[] = {0, cpu_to_be32(lmb_size)}; > + uint32_t *int_buf, *cur_index, buf_len; > + int nr_nodes = nb_numa_nodes ? nb_numa_nodes : 1; > + MemoryDeviceInfoList *dimms = NULL; > + > + /* > + * Don't create the node if there is no hotpluggable memory > + */ > + if (machine->ram_size == machine->maxram_size) { > + return 0; > + } > + > + offset = fdt_add_subnode(fdt, 0, "ibm,dynamic-reconfiguration-memory"); > + > + ret = fdt_setprop(fdt, offset, "ibm,lmb-size", prop_lmb_size, > + sizeof(prop_lmb_size)); > + if (ret < 0) { > + return ret; > + } > + > + ret = fdt_setprop_cell(fdt, offset, "ibm,memory-flags-mask", 0xff); > + if (ret < 0) { > + return ret; > + } > + > + ret = fdt_setprop_cell(fdt, offset, "ibm,memory-preservation-time", 0x0); > + if (ret < 0) { > + return ret; > + } > + > + /* ibm,dynamic-memory or ibm,dynamic-memory-v2 */ > + dimms = qmp_pc_dimm_device_list(); > + if (spapr_ovec_test(spapr->ov5_cas, OV5_DRMEM_V2)) { > + ret = spapr_populate_drmem_v2(spapr, fdt, offset, dimms); > + } else { > + ret = spapr_populate_drmem_v1(spapr, fdt, offset, dimms); > + } > + qapi_free_MemoryDeviceInfoList(dimms); > + > + if (ret < 0) { > + return ret; > } > > /* ibm,associativity-lookup-arrays */ > + buf_len = (nr_nodes * 4 + 2) * sizeof(uint32_t); > + cur_index = int_buf = g_malloc0(buf_len); > + > cur_index = int_buf; > int_buf[0] = cpu_to_be32(nr_nodes); > int_buf[1] = cpu_to_be32(4); /* Number of entries per associativity list */ > @@ -788,8 +919,8 @@ static int spapr_populate_drconf_memory(sPAPRMachineState *spapr, void *fdt) > } > ret = fdt_setprop(fdt, offset, "ibm,associativity-lookup-arrays", int_buf, > (cur_index - int_buf) * sizeof(uint32_t)); > -out: > g_free(int_buf); > + > return ret; > } > > @@ -2500,6 +2631,11 @@ static void spapr_machine_init(MachineState *machine) > spapr_ovec_set(spapr->ov5, OV5_HPT_RESIZE); > } > > + /* advertise support for ibm,dyamic-memory-v2 */ > + if (spapr->use_ibm_dynamic_memory_v2) { > + spapr_ovec_set(spapr->ov5, OV5_DRMEM_V2); > + } > + > /* init CPUs */ > spapr_init_cpus(spapr); > > @@ -2913,6 +3049,7 @@ static void spapr_instance_init(Object *obj) > > spapr->htab_fd = -1; > spapr->use_hotplug_event_source = true; > + spapr->use_ibm_dynamic_memory_v2 = true; > object_property_add_str(obj, "kvm-type", > spapr_get_kvm_type, spapr_set_kvm_type, NULL); > object_property_set_description(obj, "kvm-type", > @@ -2927,7 +3064,6 @@ static void spapr_instance_init(Object *obj) > " place of standard EPOW events when possible" > " (required for memory hot-unplug support)", > NULL); > - > ppc_compat_add_property(obj, "max-cpu-compat", &spapr->max_compat_pvr, > "Maximum permitted CPU compatibility mode", > &error_fatal); > @@ -4006,7 +4142,10 @@ DEFINE_SPAPR_MACHINE(2_13, "2.13", true); > > static void spapr_machine_2_12_instance_options(MachineState *machine) > { > + sPAPRMachineState *spapr = SPAPR_MACHINE(machine); > + > spapr_machine_2_13_instance_options(machine); > + spapr->use_ibm_dynamic_memory_v2 = false; > } > > static void spapr_machine_2_12_class_options(MachineClass *mc) > diff --git a/include/hw/ppc/spapr.h b/include/hw/ppc/spapr.h > index d60b7c6d7a..5e044c44af 100644 > --- a/include/hw/ppc/spapr.h > +++ b/include/hw/ppc/spapr.h > @@ -149,6 +149,7 @@ struct sPAPRMachineState { > sPAPROptionVector *ov5; /* QEMU-supported option vectors */ > sPAPROptionVector *ov5_cas; /* negotiated (via CAS) option vectors */ > uint32_t max_compat_pvr; > + bool use_ibm_dynamic_memory_v2; TBH, I'm not really sure we even need to adjust this by machine type. > > /* Migration state */ > int htab_save_index; > diff --git a/include/hw/ppc/spapr_ovec.h b/include/hw/ppc/spapr_ovec.h > index bf25e5d954..0f2d8d715d 100644 > --- a/include/hw/ppc/spapr_ovec.h > +++ b/include/hw/ppc/spapr_ovec.h > @@ -51,6 +51,7 @@ typedef struct sPAPROptionVector sPAPROptionVector; > #define OV5_FORM1_AFFINITY OV_BIT(5, 0) > #define OV5_HP_EVT OV_BIT(6, 5) > #define OV5_HPT_RESIZE OV_BIT(6, 7) > +#define OV5_DRMEM_V2 OV_BIT(22, 0) > #define OV5_XIVE_BOTH OV_BIT(23, 0) > #define OV5_XIVE_EXPLOIT OV_BIT(23, 1) /* 1=exploitation 0=legacy */ >
On Tue, Apr 17, 2018 at 11:14:27AM +1000, David Gibson wrote: > > static void spapr_machine_2_12_class_options(MachineClass *mc) > > diff --git a/include/hw/ppc/spapr.h b/include/hw/ppc/spapr.h > > index d60b7c6d7a..5e044c44af 100644 > > --- a/include/hw/ppc/spapr.h > > +++ b/include/hw/ppc/spapr.h > > @@ -149,6 +149,7 @@ struct sPAPRMachineState { > > sPAPROptionVector *ov5; /* QEMU-supported option vectors */ > > sPAPROptionVector *ov5_cas; /* negotiated (via CAS) option vectors */ > > uint32_t max_compat_pvr; > > + bool use_ibm_dynamic_memory_v2; > > TBH, I'm not really sure we even need to adjust this by machine type. There are other similar features controlled by ov5 bits that are also determined by machine type version: Memory hotplug support -- sPAPRMachineClass.dr_lmb_enabled Dedicated HP event support -- sPAPRMachineState.use_hotplug_event_source Are you saying that presence of ibm,dynamic-memory-v2 probably shouldn't be dependent on machine type ? Regards, Bharata.
On Tue, Apr 17, 2018 at 02:39:09PM +0530, Bharata B Rao wrote: > On Tue, Apr 17, 2018 at 11:14:27AM +1000, David Gibson wrote: > > > static void spapr_machine_2_12_class_options(MachineClass *mc) > > > diff --git a/include/hw/ppc/spapr.h b/include/hw/ppc/spapr.h > > > index d60b7c6d7a..5e044c44af 100644 > > > --- a/include/hw/ppc/spapr.h > > > +++ b/include/hw/ppc/spapr.h > > > @@ -149,6 +149,7 @@ struct sPAPRMachineState { > > > sPAPROptionVector *ov5; /* QEMU-supported option vectors */ > > > sPAPROptionVector *ov5_cas; /* negotiated (via CAS) option vectors */ > > > uint32_t max_compat_pvr; > > > + bool use_ibm_dynamic_memory_v2; > > > > TBH, I'm not really sure we even need to adjust this by machine type. > > There are other similar features controlled by ov5 bits that > are also determined by machine type version: > > Memory hotplug support -- sPAPRMachineClass.dr_lmb_enabled > Dedicated HP event support -- sPAPRMachineState.use_hotplug_event_source As for user settability the issue isn't that it's set by ov5, but what the effect of the feature is. Those other features alter runtime hypervisor behaviour and that behaviour has to remain the same across a migration. Therefore we have to keep the behaviour consistent for old machine types. This feature affects only boot time behaviour. It has a similar effect to what a firmware update might, on real hardware. Furthermore the way CAS and the device tree work, this is vanishingly unlikely to break existing guests. > Are you saying that presence of ibm,dynamic-memory-v2 probably shouldn't > be dependent on machine type ? Yes, I am.
On Wed, 18 Apr 2018 13:33:44 +1000 David Gibson <david@gibson.dropbear.id.au> wrote: > On Tue, Apr 17, 2018 at 02:39:09PM +0530, Bharata B Rao wrote: > > On Tue, Apr 17, 2018 at 11:14:27AM +1000, David Gibson wrote: > > > > static void spapr_machine_2_12_class_options(MachineClass *mc) > > > > diff --git a/include/hw/ppc/spapr.h b/include/hw/ppc/spapr.h > > > > index d60b7c6d7a..5e044c44af 100644 > > > > --- a/include/hw/ppc/spapr.h > > > > +++ b/include/hw/ppc/spapr.h > > > > @@ -149,6 +149,7 @@ struct sPAPRMachineState { > > > > sPAPROptionVector *ov5; /* QEMU-supported option vectors */ > > > > sPAPROptionVector *ov5_cas; /* negotiated (via CAS) option vectors */ > > > > uint32_t max_compat_pvr; > > > > + bool use_ibm_dynamic_memory_v2; > > > > > > TBH, I'm not really sure we even need to adjust this by machine type. > > > > There are other similar features controlled by ov5 bits that > > are also determined by machine type version: > > > > Memory hotplug support -- sPAPRMachineClass.dr_lmb_enabled > > Dedicated HP event support -- sPAPRMachineState.use_hotplug_event_source > > As for user settability the issue isn't that it's set by ov5, but what > the effect of the feature is. Those other features alter runtime > hypervisor behaviour and that behaviour has to remain the same across > a migration. Therefore we have to keep the behaviour consistent for > old machine types. > > This feature affects only boot time behaviour. It has a similar > effect to what a firmware update might, on real hardware. Furthermore > the way CAS and the device tree work, this is vanishingly unlikely to > break existing guests. > The logic in spapr_ov5_cas_needed() assumes that pre 2.8 machine types only expose OV5_FORM1_AFFINITY and OV5_DRCONF_MEMORY to guests. Adding OV5_DRMEM_V2 unconditionally breaks this assumption and backward migration to pre 2.8 QEMU versions because they don't expect the "spapr_option_vector_ov5_cas" subsection. This can cause problems in cloud environments that still have systems with older QEMU versions, eg, hosts running ubuntu LTS 16.04.4 (QEMU 2.5) which are likely to stay around until admins could transition to some newer OS. > > Are you saying that presence of ibm,dynamic-memory-v2 probably shouldn't > > be dependent on machine type ? > > Yes, I am. > I agree but we should also not put it in the migration stream then, like we already do for OV5_FORM1_AFFINITY and OV5_DRCONF_MEMORY. I've spotted another backward migration breakage wrt old, but still in use, QEMU versions. I'll send a series for both issues ASAP, so that it has a chance to land in QEMU 2.11.2. Cheers, -- Greg
On Thu, May 03, 2018 at 02:34:21PM +0200, Greg Kurz wrote: > On Wed, 18 Apr 2018 13:33:44 +1000 > David Gibson <david@gibson.dropbear.id.au> wrote: > > > On Tue, Apr 17, 2018 at 02:39:09PM +0530, Bharata B Rao wrote: > > > On Tue, Apr 17, 2018 at 11:14:27AM +1000, David Gibson wrote: > > > > > static void spapr_machine_2_12_class_options(MachineClass *mc) > > > > > diff --git a/include/hw/ppc/spapr.h b/include/hw/ppc/spapr.h > > > > > index d60b7c6d7a..5e044c44af 100644 > > > > > --- a/include/hw/ppc/spapr.h > > > > > +++ b/include/hw/ppc/spapr.h > > > > > @@ -149,6 +149,7 @@ struct sPAPRMachineState { > > > > > sPAPROptionVector *ov5; /* QEMU-supported option vectors */ > > > > > sPAPROptionVector *ov5_cas; /* negotiated (via CAS) option vectors */ > > > > > uint32_t max_compat_pvr; > > > > > + bool use_ibm_dynamic_memory_v2; > > > > > > > > TBH, I'm not really sure we even need to adjust this by machine type. > > > > > > There are other similar features controlled by ov5 bits that > > > are also determined by machine type version: > > > > > > Memory hotplug support -- sPAPRMachineClass.dr_lmb_enabled > > > Dedicated HP event support -- sPAPRMachineState.use_hotplug_event_source > > > > As for user settability the issue isn't that it's set by ov5, but what > > the effect of the feature is. Those other features alter runtime > > hypervisor behaviour and that behaviour has to remain the same across > > a migration. Therefore we have to keep the behaviour consistent for > > old machine types. > > > > This feature affects only boot time behaviour. It has a similar > > effect to what a firmware update might, on real hardware. Furthermore > > the way CAS and the device tree work, this is vanishingly unlikely to > > break existing guests. > > > > The logic in spapr_ov5_cas_needed() assumes that pre 2.8 machine types only > expose OV5_FORM1_AFFINITY and OV5_DRCONF_MEMORY to guests. Adding OV5_DRMEM_V2 > unconditionally breaks this assumption and backward migration to pre 2.8 QEMU > versions because they don't expect the "spapr_option_vector_ov5_cas" subsection. > > This can cause problems in cloud environments that still have systems with > older QEMU versions, eg, hosts running ubuntu LTS 16.04.4 (QEMU 2.5) which are > likely to stay around until admins could transition to some newer OS. Ah, good point. > > > Are you saying that presence of ibm,dynamic-memory-v2 probably shouldn't > > > be dependent on machine type ? > > > > Yes, I am. > > > > I agree but we should also not put it in the migration stream then, like we > already do for OV5_FORM1_AFFINITY and OV5_DRCONF_MEMORY. > > I've spotted another backward migration breakage wrt old, but still > in use, QEMU versions. I'll send a series for both issues ASAP, so > that it has a chance to land in QEMU 2.11.2. Thanks. I've merged your patch for 2.13 now.
diff --git a/docs/specs/ppc-spapr-hotplug.txt b/docs/specs/ppc-spapr-hotplug.txt index f57e2a09c6..cc7833108e 100644 --- a/docs/specs/ppc-spapr-hotplug.txt +++ b/docs/specs/ppc-spapr-hotplug.txt @@ -387,4 +387,23 @@ Each LMB list entry consists of the following elements: - A 32bit flags word. The bit at bit position 0x00000008 defines whether the LMB is assigned to the the partition as of boot time. +ibm,dynamic-memory-v2 + +This property describes the dynamically reconfigurable memory. This is +an alternate and newer way to describe dyanamically reconfigurable memory. +It is a property encoded array that has an integer N (the number of +LMB set entries) followed by N LMB set entries. There is an LMB set entry +for each sequential group of LMBs that share common attributes. + +Each LMB set entry consists of the following elements: + +- Number of sequential LMBs in the entry represented by a 32bit integer. +- Logical address of the first LMB in the set encoded as a 64bit integer. +- DRC index of the first LMB in the set. +- Associativity list index that is used as an index into + ibm,associativity-lookup-arrays property described earlier. This + is used to retrieve the right associativity list to be used for all + the LMBs in this set. +- A 32bit flags word that applies to all the LMBs in the set. + [1] http://thread.gmane.org/gmane.linux.ports.ppc.embedded/75350/focus=106867 diff --git a/hw/ppc/spapr.c b/hw/ppc/spapr.c index 3ffadd6ac7..fe327f3c1b 100644 --- a/hw/ppc/spapr.c +++ b/hw/ppc/spapr.c @@ -669,63 +669,136 @@ static uint32_t spapr_pc_dimm_node(MemoryDeviceInfoList *list, ram_addr_t addr) return -1; } -/* - * Adds ibm,dynamic-reconfiguration-memory node. - * Refer to docs/specs/ppc-spapr-hotplug.txt for the documentation - * of this device tree node. - */ -static int spapr_populate_drconf_memory(sPAPRMachineState *spapr, void *fdt) +struct sPAPRDrconfCellV2 { + uint32_t seq_lmbs; + uint64_t base_addr; + uint32_t drc_index; + uint32_t aa_index; + uint32_t flags; +} __attribute__((packed)); + +/* 6 uint32_t entries in sPAPRDrconfCellV2 */ +#define SPAPR_NR_DRCONF_CELL_ENTRIES 6 + +typedef struct DrconfCellQueue { + struct sPAPRDrconfCellV2 cell; + QSIMPLEQ_ENTRY(DrconfCellQueue) entry; +} DrconfCellQueue; + +QSIMPLEQ_HEAD(, DrconfCellQueue) drconf_queue + = QSIMPLEQ_HEAD_INITIALIZER(drconf_queue); + +static void spapr_add_drconf_cell(uint32_t seq_lmbs, uint64_t base_addr, + uint32_t drc_index, uint32_t aa_index, + uint32_t flags) +{ + DrconfCellQueue *elem; + + elem = g_malloc0(sizeof(*elem)); + elem->cell.seq_lmbs = cpu_to_be32(seq_lmbs); + elem->cell.base_addr = cpu_to_be64(base_addr); + elem->cell.drc_index = cpu_to_be32(drc_index); + elem->cell.aa_index = cpu_to_be32(aa_index); + elem->cell.flags = cpu_to_be32(flags); + QSIMPLEQ_INSERT_TAIL(&drconf_queue, elem, entry); +} + +/* ibm,dynamic-memory-v2 */ +static int spapr_populate_drmem_v2(sPAPRMachineState *spapr, void *fdt, + int offset, MemoryDeviceInfoList *dimms) { - MachineState *machine = MACHINE(spapr); - int ret, i, offset; - uint64_t lmb_size = SPAPR_MEMORY_BLOCK_SIZE; - uint32_t prop_lmb_size[] = {0, cpu_to_be32(lmb_size)}; - uint32_t hotplug_lmb_start = spapr->hotplug_memory.base / lmb_size; - uint32_t nr_lmbs = (spapr->hotplug_memory.base + - memory_region_size(&spapr->hotplug_memory.mr)) / - lmb_size; uint32_t *int_buf, *cur_index, buf_len; - int nr_nodes = nb_numa_nodes ? nb_numa_nodes : 1; - MemoryDeviceInfoList *dimms = NULL; + int ret; + uint64_t lmb_size = SPAPR_MEMORY_BLOCK_SIZE; + uint64_t addr, cur_addr, size; + uint32_t nr_boot_lmbs = (spapr->hotplug_memory.base / lmb_size); + uint64_t mem_end = spapr->hotplug_memory.base + + memory_region_size(&spapr->hotplug_memory.mr); + uint32_t node, nr_entries = 0; + sPAPRDRConnector *drc; + DrconfCellQueue *elem, *next; + MemoryDeviceInfoList *info; - /* - * Don't create the node if there is no hotpluggable memory - */ - if (machine->ram_size == machine->maxram_size) { - return 0; - } + /* Entry to cover RAM and the gap area */ + spapr_add_drconf_cell(nr_boot_lmbs, 0, 0, -1, + SPAPR_LMB_FLAGS_RESERVED | + SPAPR_LMB_FLAGS_DRC_INVALID); + nr_entries++; - /* - * Allocate enough buffer size to fit in ibm,dynamic-memory - * or ibm,associativity-lookup-arrays - */ - buf_len = MAX(nr_lmbs * SPAPR_DR_LMB_LIST_ENTRY_SIZE + 1, nr_nodes * 4 + 2) - * sizeof(uint32_t); - cur_index = int_buf = g_malloc0(buf_len); + cur_addr = spapr->hotplug_memory.base; + for (info = dimms; info; info = info->next) { + PCDIMMDeviceInfo *di = info->value->u.dimm.data; - offset = fdt_add_subnode(fdt, 0, "ibm,dynamic-reconfiguration-memory"); + addr = di->addr; + size = di->size; + node = di->node; - ret = fdt_setprop(fdt, offset, "ibm,lmb-size", prop_lmb_size, - sizeof(prop_lmb_size)); - if (ret < 0) { - goto out; + /* Entry for hot-pluggable area */ + if (cur_addr < addr) { + drc = spapr_drc_by_id(TYPE_SPAPR_DRC_LMB, cur_addr / lmb_size); + g_assert(drc); + spapr_add_drconf_cell((addr - cur_addr) / lmb_size, + cur_addr, spapr_drc_index(drc), + -1, 0); + nr_entries++; + } + + /* Entry for DIMM */ + drc = spapr_drc_by_id(TYPE_SPAPR_DRC_LMB, addr / lmb_size); + g_assert(drc); + spapr_add_drconf_cell(size / lmb_size, addr, spapr_drc_index(drc), + node, SPAPR_LMB_FLAGS_ASSIGNED); + nr_entries++; + cur_addr = addr + size; } - ret = fdt_setprop_cell(fdt, offset, "ibm,memory-flags-mask", 0xff); - if (ret < 0) { - goto out; + /* Entry for remaining hotpluggable area */ + if (cur_addr < mem_end) { + drc = spapr_drc_by_id(TYPE_SPAPR_DRC_LMB, cur_addr / lmb_size); + g_assert(drc); + spapr_add_drconf_cell((mem_end - cur_addr) / lmb_size, cur_addr, + spapr_drc_index(drc), -1, 0); + nr_entries++; } - ret = fdt_setprop_cell(fdt, offset, "ibm,memory-preservation-time", 0x0); - if (ret < 0) { - goto out; + buf_len = (nr_entries * SPAPR_NR_DRCONF_CELL_ENTRIES + 1) * + sizeof(uint32_t); + int_buf = cur_index = g_malloc0(buf_len); + int_buf[0] = cpu_to_be32(nr_entries); + cur_index++; + QSIMPLEQ_FOREACH_SAFE(elem, &drconf_queue, entry, next) { + memcpy(cur_index, &elem->cell, + SPAPR_NR_DRCONF_CELL_ENTRIES * sizeof(uint32_t)); + cur_index += SPAPR_NR_DRCONF_CELL_ENTRIES; + QSIMPLEQ_REMOVE(&drconf_queue, elem, DrconfCellQueue, entry); + g_free(elem); } - if (hotplug_lmb_start) { - dimms = qmp_pc_dimm_device_list(); + ret = fdt_setprop(fdt, offset, "ibm,dynamic-memory-v2", int_buf, buf_len); + g_free(int_buf); + if (ret < 0) { + return -1; } + return 0; +} + +/* ibm,dynamic-memory */ +static int spapr_populate_drmem_v1(sPAPRMachineState *spapr, void *fdt, + int offset, MemoryDeviceInfoList *dimms) +{ + int i, ret; + uint64_t lmb_size = SPAPR_MEMORY_BLOCK_SIZE; + uint32_t hotplug_lmb_start = spapr->hotplug_memory.base / lmb_size; + uint32_t nr_lmbs = (spapr->hotplug_memory.base + + memory_region_size(&spapr->hotplug_memory.mr)) / + lmb_size; + uint32_t *int_buf, *cur_index, buf_len; - /* ibm,dynamic-memory */ + /* + * Allocate enough buffer size to fit in ibm,dynamic-memory + */ + buf_len = (nr_lmbs * SPAPR_DR_LMB_LIST_ENTRY_SIZE + 1) * sizeof(uint32_t); + cur_index = int_buf = g_malloc0(buf_len); int_buf[0] = cpu_to_be32(nr_lmbs); cur_index++; for (i = 0; i < nr_lmbs; i++) { @@ -765,13 +838,71 @@ static int spapr_populate_drconf_memory(sPAPRMachineState *spapr, void *fdt) cur_index += SPAPR_DR_LMB_LIST_ENTRY_SIZE; } - qapi_free_MemoryDeviceInfoList(dimms); ret = fdt_setprop(fdt, offset, "ibm,dynamic-memory", int_buf, buf_len); + g_free(int_buf); if (ret < 0) { - goto out; + return -1; + } + return 0; +} + +/* + * Adds ibm,dynamic-reconfiguration-memory node. + * Refer to docs/specs/ppc-spapr-hotplug.txt for the documentation + * of this device tree node. + */ +static int spapr_populate_drconf_memory(sPAPRMachineState *spapr, void *fdt) +{ + MachineState *machine = MACHINE(spapr); + int ret, i, offset; + uint64_t lmb_size = SPAPR_MEMORY_BLOCK_SIZE; + uint32_t prop_lmb_size[] = {0, cpu_to_be32(lmb_size)}; + uint32_t *int_buf, *cur_index, buf_len; + int nr_nodes = nb_numa_nodes ? nb_numa_nodes : 1; + MemoryDeviceInfoList *dimms = NULL; + + /* + * Don't create the node if there is no hotpluggable memory + */ + if (machine->ram_size == machine->maxram_size) { + return 0; + } + + offset = fdt_add_subnode(fdt, 0, "ibm,dynamic-reconfiguration-memory"); + + ret = fdt_setprop(fdt, offset, "ibm,lmb-size", prop_lmb_size, + sizeof(prop_lmb_size)); + if (ret < 0) { + return ret; + } + + ret = fdt_setprop_cell(fdt, offset, "ibm,memory-flags-mask", 0xff); + if (ret < 0) { + return ret; + } + + ret = fdt_setprop_cell(fdt, offset, "ibm,memory-preservation-time", 0x0); + if (ret < 0) { + return ret; + } + + /* ibm,dynamic-memory or ibm,dynamic-memory-v2 */ + dimms = qmp_pc_dimm_device_list(); + if (spapr_ovec_test(spapr->ov5_cas, OV5_DRMEM_V2)) { + ret = spapr_populate_drmem_v2(spapr, fdt, offset, dimms); + } else { + ret = spapr_populate_drmem_v1(spapr, fdt, offset, dimms); + } + qapi_free_MemoryDeviceInfoList(dimms); + + if (ret < 0) { + return ret; } /* ibm,associativity-lookup-arrays */ + buf_len = (nr_nodes * 4 + 2) * sizeof(uint32_t); + cur_index = int_buf = g_malloc0(buf_len); + cur_index = int_buf; int_buf[0] = cpu_to_be32(nr_nodes); int_buf[1] = cpu_to_be32(4); /* Number of entries per associativity list */ @@ -788,8 +919,8 @@ static int spapr_populate_drconf_memory(sPAPRMachineState *spapr, void *fdt) } ret = fdt_setprop(fdt, offset, "ibm,associativity-lookup-arrays", int_buf, (cur_index - int_buf) * sizeof(uint32_t)); -out: g_free(int_buf); + return ret; } @@ -2500,6 +2631,11 @@ static void spapr_machine_init(MachineState *machine) spapr_ovec_set(spapr->ov5, OV5_HPT_RESIZE); } + /* advertise support for ibm,dyamic-memory-v2 */ + if (spapr->use_ibm_dynamic_memory_v2) { + spapr_ovec_set(spapr->ov5, OV5_DRMEM_V2); + } + /* init CPUs */ spapr_init_cpus(spapr); @@ -2913,6 +3049,7 @@ static void spapr_instance_init(Object *obj) spapr->htab_fd = -1; spapr->use_hotplug_event_source = true; + spapr->use_ibm_dynamic_memory_v2 = true; object_property_add_str(obj, "kvm-type", spapr_get_kvm_type, spapr_set_kvm_type, NULL); object_property_set_description(obj, "kvm-type", @@ -2927,7 +3064,6 @@ static void spapr_instance_init(Object *obj) " place of standard EPOW events when possible" " (required for memory hot-unplug support)", NULL); - ppc_compat_add_property(obj, "max-cpu-compat", &spapr->max_compat_pvr, "Maximum permitted CPU compatibility mode", &error_fatal); @@ -4006,7 +4142,10 @@ DEFINE_SPAPR_MACHINE(2_13, "2.13", true); static void spapr_machine_2_12_instance_options(MachineState *machine) { + sPAPRMachineState *spapr = SPAPR_MACHINE(machine); + spapr_machine_2_13_instance_options(machine); + spapr->use_ibm_dynamic_memory_v2 = false; } static void spapr_machine_2_12_class_options(MachineClass *mc) diff --git a/include/hw/ppc/spapr.h b/include/hw/ppc/spapr.h index d60b7c6d7a..5e044c44af 100644 --- a/include/hw/ppc/spapr.h +++ b/include/hw/ppc/spapr.h @@ -149,6 +149,7 @@ struct sPAPRMachineState { sPAPROptionVector *ov5; /* QEMU-supported option vectors */ sPAPROptionVector *ov5_cas; /* negotiated (via CAS) option vectors */ uint32_t max_compat_pvr; + bool use_ibm_dynamic_memory_v2; /* Migration state */ int htab_save_index; diff --git a/include/hw/ppc/spapr_ovec.h b/include/hw/ppc/spapr_ovec.h index bf25e5d954..0f2d8d715d 100644 --- a/include/hw/ppc/spapr_ovec.h +++ b/include/hw/ppc/spapr_ovec.h @@ -51,6 +51,7 @@ typedef struct sPAPROptionVector sPAPROptionVector; #define OV5_FORM1_AFFINITY OV_BIT(5, 0) #define OV5_HP_EVT OV_BIT(6, 5) #define OV5_HPT_RESIZE OV_BIT(6, 7) +#define OV5_DRMEM_V2 OV_BIT(22, 0) #define OV5_XIVE_BOTH OV_BIT(23, 0) #define OV5_XIVE_EXPLOIT OV_BIT(23, 1) /* 1=exploitation 0=legacy */
The new property ibm,dynamic-memory-v2 allows memory to be represented in a more compact manner in device tree. Signed-off-by: Bharata B Rao <bharata@linux.vnet.ibm.com> --- v2 - https://lists.nongnu.org/archive/html/qemu-ppc/2018-04/msg00052.html Changes in v3: - Addressed David Gibson's review comments: use of CamelCase, separate helper for populating drconf_cell queue entries, removed the user configurability of drmem_v2 machine property. docs/specs/ppc-spapr-hotplug.txt | 19 ++++ hw/ppc/spapr.c | 233 +++++++++++++++++++++++++++++++-------- include/hw/ppc/spapr.h | 1 + include/hw/ppc/spapr_ovec.h | 1 + 4 files changed, 207 insertions(+), 47 deletions(-)