From patchwork Tue Apr 19 11:51:41 2016 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Bharata B Rao X-Patchwork-Id: 8879161 Return-Path: X-Original-To: patchwork-qemu-devel@patchwork.kernel.org Delivered-To: patchwork-parsemail@patchwork1.web.kernel.org Received: from mail.kernel.org (mail.kernel.org [198.145.29.136]) by patchwork1.web.kernel.org (Postfix) with ESMTP id CF6BD9F1C1 for ; Tue, 19 Apr 2016 11:52:29 +0000 (UTC) Received: from mail.kernel.org (localhost [127.0.0.1]) by mail.kernel.org (Postfix) with ESMTP id B6CBD2026D for ; Tue, 19 Apr 2016 11:52:28 +0000 (UTC) Received: from lists.gnu.org (lists.gnu.org [208.118.235.17]) (using TLSv1 with cipher AES256-SHA (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPS id 383D720145 for ; Tue, 19 Apr 2016 11:52:27 +0000 (UTC) Received: from localhost ([::1]:56541 helo=lists.gnu.org) by lists.gnu.org with esmtp (Exim 4.71) (envelope-from ) id 1asUCg-00027i-Dm for patchwork-qemu-devel@patchwork.kernel.org; Tue, 19 Apr 2016 07:52:26 -0400 Received: from eggs.gnu.org ([2001:4830:134:3::10]:60277) by lists.gnu.org with esmtp (Exim 4.71) (envelope-from ) id 1asUCM-0001pk-LA for qemu-devel@nongnu.org; Tue, 19 Apr 2016 07:52:08 -0400 Received: from Debian-exim by eggs.gnu.org with spam-scanned (Exim 4.71) (envelope-from ) id 1asUCJ-0003EK-C8 for qemu-devel@nongnu.org; Tue, 19 Apr 2016 07:52:06 -0400 Received: from e28smtp01.in.ibm.com ([125.16.236.1]:50079) by eggs.gnu.org with esmtp (Exim 4.71) (envelope-from ) id 1asUCI-0003DW-DR for qemu-devel@nongnu.org; Tue, 19 Apr 2016 07:52:03 -0400 Received: from localhost by e28smtp01.in.ibm.com with IBM ESMTP SMTP Gateway: Authorized Use Only! Violators will be prosecuted for from ; Tue, 19 Apr 2016 17:21:57 +0530 Received: from d28relay01.in.ibm.com (9.184.220.58) by e28smtp01.in.ibm.com (192.168.1.131) with IBM ESMTP SMTP Gateway: Authorized Use Only! Violators will be prosecuted; Tue, 19 Apr 2016 17:21:55 +0530 X-IBM-Helo: d28relay01.in.ibm.com X-IBM-MailFrom: bharata@linux.vnet.ibm.com X-IBM-RcptTo: qemu-ppc@nongnu.org;qemu-devel@nongnu.org Received: from d28av02.in.ibm.com (d28av02.in.ibm.com [9.184.220.64]) by d28relay01.in.ibm.com (8.14.9/8.14.9/NCO v10.0) with ESMTP id u3JBpsEA51118104; Tue, 19 Apr 2016 17:21:54 +0530 Received: from d28av02.in.ibm.com (localhost [127.0.0.1]) by d28av02.in.ibm.com (8.14.4/8.14.4/NCO v10.0 AVout) with ESMTP id u3JBpqV5027892; Tue, 19 Apr 2016 17:21:53 +0530 Received: from bharata.in.ibm.com ([9.79.183.128]) by d28av02.in.ibm.com (8.14.4/8.14.4/NCO v10.0 AVin) with ESMTP id u3JBpnCm027723; Tue, 19 Apr 2016 17:21:50 +0530 From: Bharata B Rao To: qemu-devel@nongnu.org Date: Tue, 19 Apr 2016 17:21:41 +0530 Message-Id: <1461066701-23212-1-git-send-email-bharata@linux.vnet.ibm.com> X-Mailer: git-send-email 2.1.0 X-TM-AS-MML: disable x-cbid: 16041911-4790-0000-0000-00000F018B69 X-detected-operating-system: by eggs.gnu.org: GNU/Linux 3.x X-Received-From: 125.16.236.1 Subject: [Qemu-devel] [RFC PATCH v0 for 2.7] spapr: Work around the memory hotplug failure with DDW X-BeenThere: qemu-devel@nongnu.org X-Mailman-Version: 2.1.21 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: aik@au1.ibm.com, Michael Roth , qemu-ppc@nongnu.org, Bharata B Rao , Nathan Fontenot , david@gibson.dropbear.id.au Errors-To: qemu-devel-bounces+patchwork-qemu-devel=patchwork.kernel.org@nongnu.org Sender: "Qemu-devel" X-Spam-Status: No, score=-6.9 required=5.0 tests=BAYES_00, RCVD_IN_DNSWL_HI, UNPARSEABLE_RELAY autolearn=unavailable version=3.3.1 X-Spam-Checker-Version: SpamAssassin 3.3.1 (2010-03-16) on mail.kernel.org X-Virus-Scanned: ClamAV using ClamSMTP Memory hotplug can fail for some combinations of RAM and maxmem when DDW is enabled in the presence of devices like nec-xhci-usb. DDW depends on maximum addressable memory by guest and this value is currently being calculated wrongly by the guest kernel routine memory_hotplug_max(). While there is an attempt to fix the guest kernel(*), this patch works around the problem within QEMU itself. memory_hotplug_max() routine in the guest kernel arrives at max addressable memory by multiplying lmb-size with the lmb-count obtained from ibm,dyanmic-memory property. There are two assumptions here: - All LMBs are part of ibm,dynamic memory: This is not true for PowerKVM where only hot-pluggable LMBs are present in this property. - The memory area comprising of RAM and hotplug region is contiguous: This needn't be true always for PowerKVM as there can be gap between boot time RAM and hotplug region. This work around involves having all the LMBs (RMA, rest of the boot time LMBs and hot-pluggable LMBs) as part of ibm,dynamic-memory so that guest kernel's calculation of max addressable memory comes out correct resulting in correct DDW value which prevents memory hotplug failures. memory@0 is created for RMA, but RMA LMBs are also represented as "reserved" LMBs in ibm,dynamic-memory. Parts of this are essenitally a revert of e8f986fc57a664a74b9f685b466506366a15201b. In addition to this, the alignment of hotplug memory region is reduced from current 1G to 256M (LMB size in PowerKVM) so that we don't end up with any gaps between boot time RAM and hotplug region. This change has a side effect on how the memory nodes in DT are represented before and after this change. As an example consider a guest with the following memory related command line options: -m 4G,slots=32,maxmem=8G -numa node,nodeid=0,mem=2G -numa node,nodeid=1,mem=2G Before this change, the guest would have Scenario 1 ---------- memory@0 for RMA memory@80000000 for rest of the boot time memory ibm,dynamic-reconfiguration-memory for hot-pluggable memory. After this commit, the guest will have Scenario 2 ---------- memory@0 for RMA ibm,dynamic-reconfiguration-memory for the entire memory including RMA, boot time memory as well as hot-pluggable memory. If an existing guest having DT nodes as in Scenario 1 above is migrated to a QEMU which has this change, at the target, it continues to have the DT nodes as in Scenario 1. However after 1st reboot, the DT representation changes over to Scenario 2. I haven't yet looked at Jian Jun's DRC migration patchset to ascertain if this change works well with DRC migration. (*) https://patchwork.ozlabs.org/patch/606912/ Signed-off-by: Bharata B Rao Cc: Nathan Fontenot Cc: Michael Roth --- hw/ppc/spapr.c | 59 +++++++++++++++++++++++++++++++++++--------------- include/hw/ppc/spapr.h | 1 + 2 files changed, 43 insertions(+), 17 deletions(-) diff --git a/hw/ppc/spapr.c b/hw/ppc/spapr.c index 79a70a9..6d8de2e 100644 --- a/hw/ppc/spapr.c +++ b/hw/ppc/spapr.c @@ -566,7 +566,6 @@ static int spapr_populate_memory(sPAPRMachineState *spapr, void *fdt) } if (!mem_start) { /* ppc_spapr_init() checks for rma_size <= node0_size already */ - spapr_populate_memory_node(fdt, i, 0, spapr->rma_size); mem_start += spapr->rma_size; node_size -= spapr->rma_size; } @@ -759,18 +758,13 @@ static int spapr_populate_drconf_memory(sPAPRMachineState *spapr, void *fdt) int ret, i, offset; uint64_t lmb_size = SPAPR_MEMORY_BLOCK_SIZE; uint32_t prop_lmb_size[] = {0, cpu_to_be32(lmb_size)}; - uint32_t nr_lmbs = (machine->maxram_size - machine->ram_size)/lmb_size; + uint32_t nr_rma_lmbs = spapr->rma_size / lmb_size; + uint32_t nr_lmbs = machine->maxram_size / lmb_size; + uint32_t nr_assigned_lmbs = machine->ram_size / lmb_size; uint32_t *int_buf, *cur_index, buf_len; int nr_nodes = nb_numa_nodes ? nb_numa_nodes : 1; /* - * Don't create the node if there are no DR LMBs. - */ - if (!nr_lmbs) { - return 0; - } - - /* * Allocate enough buffer size to fit in ibm,dynamic-memory * or ibm,associativity-lookup-arrays */ @@ -802,9 +796,15 @@ static int spapr_populate_drconf_memory(sPAPRMachineState *spapr, void *fdt) for (i = 0; i < nr_lmbs; i++) { sPAPRDRConnector *drc; sPAPRDRConnectorClass *drck; - uint64_t addr = i * lmb_size + spapr->hotplug_memory.base;; + uint64_t addr; uint32_t *dynamic_memory = cur_index; + if (i < nr_assigned_lmbs) { + addr = i * lmb_size; + } else { + addr = (i - nr_assigned_lmbs) * lmb_size + + spapr->hotplug_memory.base; + } drc = spapr_dr_connector_by_id(SPAPR_DR_CONNECTOR_TYPE_LMB, addr/lmb_size); g_assert(drc); @@ -817,7 +817,11 @@ static int spapr_populate_drconf_memory(sPAPRMachineState *spapr, void *fdt) dynamic_memory[4] = cpu_to_be32(numa_get_node(addr, NULL)); if (addr < machine->ram_size || memory_region_present(get_system_memory(), addr)) { - dynamic_memory[5] = cpu_to_be32(SPAPR_LMB_FLAGS_ASSIGNED); + if (i < nr_rma_lmbs) { + dynamic_memory[5] = cpu_to_be32(SPAPR_LMB_FLAGS_RESERVED); + } else { + dynamic_memory[5] = cpu_to_be32(SPAPR_LMB_FLAGS_ASSIGNED); + } } else { dynamic_memory[5] = cpu_to_be32(0); } @@ -879,6 +883,8 @@ int spapr_h_cas_compose_response(sPAPRMachineState *spapr, /* Generate ibm,dynamic-reconfiguration-memory node if required */ if (memory_update && smc->dr_lmb_enabled) { _FDT((spapr_populate_drconf_memory(spapr, fdt))); + } else { + _FDT((spapr_populate_memory(spapr, fdt))); } /* Pack resulting tree */ @@ -916,10 +922,23 @@ static void spapr_finalize_fdt(sPAPRMachineState *spapr, /* open out the base tree into a temp buffer for the final tweaks */ _FDT((fdt_open_into(spapr->fdt_skel, fdt, FDT_MAX_SIZE))); - ret = spapr_populate_memory(spapr, fdt); - if (ret < 0) { - fprintf(stderr, "couldn't setup memory nodes in fdt\n"); - exit(1); + /* + * Add memory@0 node to represent RMA. Rest of the memory is either + * represented by memory nodes or ibm,dynamic-reconfiguration-memory + * node later during ibm,client-architecture-support call. + * + * If NUMA is configured, ensure that memory@0 ends up in the + * first memory-less node. + */ + if (nb_numa_nodes) { + for (i = 0; i < nb_numa_nodes; ++i) { + if (numa_info[i].node_mem) { + spapr_populate_memory_node(fdt, i, 0, spapr->rma_size); + break; + } + } + } else { + spapr_populate_memory_node(fdt, 0, 0, spapr->rma_size); } ret = spapr_populate_vdevice(spapr->vio_bus, fdt); @@ -1659,14 +1678,20 @@ static void spapr_create_lmb_dr_connectors(sPAPRMachineState *spapr) { MachineState *machine = MACHINE(spapr); uint64_t lmb_size = SPAPR_MEMORY_BLOCK_SIZE; - uint32_t nr_lmbs = (machine->maxram_size - machine->ram_size)/lmb_size; + uint32_t nr_lmbs = machine->maxram_size / lmb_size; + uint32_t nr_assigned_lmbs = machine->ram_size / lmb_size; int i; for (i = 0; i < nr_lmbs; i++) { sPAPRDRConnector *drc; uint64_t addr; - addr = i * lmb_size + spapr->hotplug_memory.base; + if (i < nr_assigned_lmbs) { + addr = i * lmb_size; + } else { + addr = (i - nr_assigned_lmbs) * lmb_size + + spapr->hotplug_memory.base; + } drc = spapr_dr_connector_new(OBJECT(spapr), SPAPR_DR_CONNECTOR_TYPE_LMB, addr/lmb_size); qemu_register_reset(spapr_drc_reset, drc); diff --git a/include/hw/ppc/spapr.h b/include/hw/ppc/spapr.h index 098d85d..9f2050d 100644 --- a/include/hw/ppc/spapr.h +++ b/include/hw/ppc/spapr.h @@ -627,5 +627,6 @@ int spapr_rng_populate_dt(void *fdt); * property under ibm,dynamic-reconfiguration-memory node. */ #define SPAPR_LMB_FLAGS_ASSIGNED 0x00000008 +#define SPAPR_LMB_FLAGS_RESERVED 0x00000080 #endif /* !defined (__HW_SPAPR_H__) */