From patchwork Wed Jun 16 16:29:37 2021 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: David Hildenbrand X-Patchwork-Id: 12325675 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-13.6 required=3.0 tests=BAYES_00,DKIM_INVALID, DKIM_SIGNED,HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_CR_TRAILER,INCLUDES_PATCH, MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 98D62C48BE5 for ; Wed, 16 Jun 2021 16:38:13 +0000 (UTC) Received: from lists.gnu.org (lists.gnu.org [209.51.188.17]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPS id 2435D61359 for ; Wed, 16 Jun 2021 16:38:13 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 2435D61359 Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=redhat.com Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=qemu-devel-bounces+qemu-devel=archiver.kernel.org@nongnu.org Received: from localhost ([::1]:34580 helo=lists1p.gnu.org) by lists.gnu.org with esmtp (Exim 4.90_1) (envelope-from ) id 1ltYYO-0006oQ-Ad for qemu-devel@archiver.kernel.org; Wed, 16 Jun 2021 12:38:12 -0400 Received: from eggs.gnu.org ([2001:470:142:3::10]:39816) by lists.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1ltYR6-0002PN-HB for qemu-devel@nongnu.org; Wed, 16 Jun 2021 12:30:40 -0400 Received: from us-smtp-delivery-124.mimecast.com ([216.205.24.124]:27487) by eggs.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256) (Exim 4.90_1) (envelope-from ) id 1ltYR4-0007SO-GW for qemu-devel@nongnu.org; Wed, 16 Jun 2021 12:30:40 -0400 DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1623861038; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=rAd5KKAy4P1jm+87Iaub23Kp15FKRD1SIs182kYRclA=; b=FiIHbOkdlxZ0OdE0TbMrJq1K7cwtmSIVOkVoO/j+cAmT0e3eFEgd5ywRrNLzM+wlx6sbl4 mTzFe2MwotydsYN5PEuhiFjjAbCx6zCFbN+4SBGdeFSVgf51w0bV0iH0kunIumTxSJ0P6G hDZQfl9e8ecAcTpwHTUk5agllJKDARE= Received: from mimecast-mx01.redhat.com (mimecast-mx01.redhat.com [209.132.183.4]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-361-qDM4slNTNmyScZcoVtI94w-1; Wed, 16 Jun 2021 12:30:31 -0400 X-MC-Unique: qDM4slNTNmyScZcoVtI94w-1 Received: from smtp.corp.redhat.com (int-mx03.intmail.prod.int.phx2.redhat.com [10.5.11.13]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx01.redhat.com (Postfix) with ESMTPS id 64F98818703; Wed, 16 Jun 2021 16:30:30 +0000 (UTC) Received: from t480s.redhat.com (ovpn-114-100.ams2.redhat.com [10.36.114.100]) by smtp.corp.redhat.com (Postfix) with ESMTP id D7C2960C54; Wed, 16 Jun 2021 16:30:20 +0000 (UTC) From: David Hildenbrand To: qemu-devel@nongnu.org Subject: [PATCH v1 3/6] migration/ram: Handle RAMBlocks with a RamDiscardManager on the migration source Date: Wed, 16 Jun 2021 18:29:37 +0200 Message-Id: <20210616162940.28630-4-david@redhat.com> In-Reply-To: <20210616162940.28630-1-david@redhat.com> References: <20210616162940.28630-1-david@redhat.com> MIME-Version: 1.0 X-Scanned-By: MIMEDefang 2.79 on 10.5.11.13 Authentication-Results: relay.mimecast.com; auth=pass smtp.auth=CUSA124A263 smtp.mailfrom=david@redhat.com X-Mimecast-Spam-Score: 0 X-Mimecast-Originator: redhat.com Received-SPF: pass client-ip=216.205.24.124; envelope-from=david@redhat.com; helo=us-smtp-delivery-124.mimecast.com X-Spam_score_int: -29 X-Spam_score: -3.0 X-Spam_bar: --- X-Spam_report: (-3.0 / 5.0 requ) BAYES_00=-1.9, DKIMWL_WL_HIGH=-0.199, DKIM_SIGNED=0.1, DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, DKIM_VALID_EF=-0.1, RCVD_IN_DNSWL_LOW=-0.7, RCVD_IN_MSPIKE_H4=0.001, RCVD_IN_MSPIKE_WL=0.001, SPF_HELO_NONE=0.001, SPF_PASS=-0.001 autolearn=ham autolearn_force=no X-Spam_action: no action X-BeenThere: qemu-devel@nongnu.org X-Mailman-Version: 2.1.23 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: Wei Yang , "Michael S. Tsirkin" , Pankaj Gupta , Juan Quintela , David Hildenbrand , "Dr. David Alan Gilbert" , Peter Xu , Marek Kedzierski , Alex Williamson , teawater , Paolo Bonzini , Andrey Gruzdev Errors-To: qemu-devel-bounces+qemu-devel=archiver.kernel.org@nongnu.org Sender: "Qemu-devel" We never want to migrate memory that corresponds to discarded ranges as managed by a RamDiscardManager responsible for the mapped memory region of the RAMBlock. The content of these pages is essentially stale and without any guarantees for the VM ("logically unplugged"). Depending on the underlying memory type, even reading memory might populate memory on the source, resulting in an undesired memory consumption. Of course, on the destination, even writing a zeropage consumes memory, which we also want to avoid (similar to free page hinting). Currently, virtio-mem tries achieving that goal (not migrating "unplugged" memory that was discarded) by going via qemu_guest_free_page_hint() - but it's hackish and incomplete. For example, background snapshots still end up reading all memory, as they don't do bitmap syncs. Postcopy recovery code will re-add previously cleared bits to the dirty bitmap and migrate them. Let's consult the RamDiscardManager whenever we add bits to the bitmap and exclude all discarded pages. As colo is incompatible with discarding of RAM and inhibits it, we don't have to bother. Signed-off-by: David Hildenbrand --- migration/ram.c | 69 ++++++++++++++++++++++++++++++++++++++++++++++--- 1 file changed, 65 insertions(+), 4 deletions(-) diff --git a/migration/ram.c b/migration/ram.c index 60ea913c54..03867b4971 100644 --- a/migration/ram.c +++ b/migration/ram.c @@ -832,14 +832,67 @@ static inline bool migration_bitmap_clear_dirty(RAMState *rs, return ret; } +static void dirty_bitmap_exclude_section(MemoryRegionSection *section, + void *opaque) +{ + const hwaddr offset = section->offset_within_region; + const hwaddr size = int128_get64(section->size); + const unsigned long start = offset >> TARGET_PAGE_BITS; + const unsigned long npages = size >> TARGET_PAGE_BITS; + RAMBlock *rb = section->mr->ram_block; + uint64_t *cleared_bits = opaque; + + /* Clear the discarded part from our bitmap. */ + *cleared_bits += bitmap_count_one_with_offset(rb->bmap, start, npages); + bitmap_clear(rb->bmap, start, npages); +} + +/* + * Exclude all dirty pages that fall into a discarded range as managed by a + * RamDiscardManager responsible for the mapped memory region of the RAMBlock. + * + * Discarded pages ("logically unplugged") have undefined content and must + * never be migrated, because even reading these pages for migrating might + * result in undesired behavior. + * + * Returns the number of cleared bits in the dirty bitmap. + * + * Note: The result is only stable while migration (precopy/postcopy). + */ +static uint64_t ramblock_dirty_bitmap_exclude_discarded_pages(RAMBlock *rb) +{ + uint64_t cleared_bits = 0; + + if (rb->mr && memory_region_has_ram_discard_manager(rb->mr)) { + RamDiscardManager *rdm = memory_region_get_ram_discard_manager(rb->mr); + MemoryRegionSection section = { + .mr = rb->mr, + .offset_within_region = 0, + .size = int128_make64(qemu_ram_get_used_length(rb)), + }; + + ram_discard_manager_replay_discarded(rdm, §ion, + dirty_bitmap_exclude_section, + &cleared_bits); + } + return cleared_bits; +} + /* Called with RCU critical section */ static void ramblock_sync_dirty_bitmap(RAMState *rs, RAMBlock *rb) { uint64_t new_dirty_pages = cpu_physical_memory_sync_dirty_bitmap(rb, 0, rb->used_length); + uint64_t cleared_pages = 0; - rs->migration_dirty_pages += new_dirty_pages; - rs->num_dirty_pages_period += new_dirty_pages; + if (new_dirty_pages) { + cleared_pages = ramblock_dirty_bitmap_exclude_discarded_pages(rb); + } + + rs->migration_dirty_pages += new_dirty_pages - cleared_pages; + if (new_dirty_pages > cleared_pages) { + rs->num_dirty_pages_period += new_dirty_pages - cleared_pages; + } } /** @@ -2603,7 +2656,7 @@ static int ram_state_init(RAMState **rsp) return 0; } -static void ram_list_init_bitmaps(void) +static void ram_list_init_bitmaps(RAMState *rs) { MigrationState *ms = migrate_get_current(); RAMBlock *block; @@ -2638,6 +2691,10 @@ static void ram_list_init_bitmaps(void) bitmap_set(block->bmap, 0, pages); block->clear_bmap_shift = shift; block->clear_bmap = bitmap_new(clear_bmap_size(pages, shift)); + + /* Exclude all discarded pages we never want to migrate. */ + pages = ramblock_dirty_bitmap_exclude_discarded_pages(block); + rs->migration_dirty_pages -= pages; } } } @@ -2649,7 +2706,7 @@ static void ram_init_bitmaps(RAMState *rs) qemu_mutex_lock_ramlist(); WITH_RCU_READ_LOCK_GUARD() { - ram_list_init_bitmaps(); + ram_list_init_bitmaps(rs); /* We don't use dirty log with background snapshots */ if (!migrate_background_snapshot()) { memory_global_dirty_log_start(); @@ -4068,6 +4125,10 @@ int ram_dirty_bitmap_reload(MigrationState *s, RAMBlock *block) */ bitmap_complement(block->bmap, block->bmap, nbits); + /* Exclude all discarded pages we never want to migrate. */ + ramblock_dirty_bitmap_exclude_discarded_pages(block); + + /* We'll recalculate migration_dirty_pages in ram_state_resume_prepare(). */ trace_ram_dirty_bitmap_reload_complete(block->idstr); /*