From patchwork Wed Dec 11 04:27:33 2019 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 8bit X-Patchwork-Submitter: SeongJae Park X-Patchwork-Id: 11284229 Return-Path: Received: from mail.kernel.org (pdx-korg-mail-1.web.codeaurora.org [172.30.200.123]) by pdx-korg-patchwork-2.web.codeaurora.org (Postfix) with ESMTP id 0037E138D for ; Wed, 11 Dec 2019 04:27:43 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by mail.kernel.org (Postfix) with ESMTP id C73652073B for ; Wed, 11 Dec 2019 04:27:42 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=pass (2048-bit key) header.d=gmail.com header.i=@gmail.com header.b="Lofcsni9" Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726874AbfLKE1m (ORCPT ); Tue, 10 Dec 2019 23:27:42 -0500 Received: from mail-pg1-f194.google.com ([209.85.215.194]:37478 "EHLO mail-pg1-f194.google.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1726783AbfLKE1m (ORCPT ); Tue, 10 Dec 2019 23:27:42 -0500 Received: by mail-pg1-f194.google.com with SMTP id q127so10114090pga.4; Tue, 10 Dec 2019 20:27:41 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=from:to:cc:subject:date:message-id:in-reply-to:references; bh=rEcgwnaRhxf5O9dD75Zxn4HOLh2R66s7XyntOC74d9E=; b=Lofcsni9piHB5mGQqPP+2S3S9s2EWFhep17RT10aY33Ak4WVy8+GtR5AqC/lyTbgrr Gg/7AF1H3tw7Ve+kMhOLmpbKBhsALPRZb3g8pXcEmQqBhuJ1TdI359kUsfjWqHhPuF9a c6YUIfaBtfXhrVNjJljAadm+61nNiWTUpAc8EiURHCKSBZUqniMDjhKcXY1RsM8zqQF7 0+Nbn1ryRgDh6di/c5AcWUXRcFRtaiYCz2hmEmFPNPYADcLKg2SsK9lzyB9qTXjUe591 vOtDOH2suB8pnV0xgEsaHX1zr6nzndzvfkECb7EbFXB1tS7YazCJzuSOKlfTRMoSx08m HPBw== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references; bh=rEcgwnaRhxf5O9dD75Zxn4HOLh2R66s7XyntOC74d9E=; b=IokzFSU6RRVFEItTbz+aP1SRf/G2Y1YU4RhLb4WHPPQeozzRQCh3mnQnc1xYLGJBlp bGGEwK4fOTADjDdD3EB6Lb2oQ6wAAIxUzhKqhDWnubHCILak0+btV/mZ0YvgSbPOfcjb ZoAZZhlbDYI1LQOkKYsiiP5L9z2LsM6rSRKxB7nSsmmTKFeXOWQ800EvYaIjfAuMCFTK o+wgGV9d3ymp9yUWcZhjZZcuFrKy+mTIN4qXmouSQnNcbvNWeMz6YUpoV4WoyNNTpThb bidH6d8R80PR7eLXZapdaCCCL+j5pcqxAD0pn1SvaVjhD4vdL7dNO53FX1fLqZ+vCGgm 7JTw== X-Gm-Message-State: APjAAAWkROAluKB7X/MNql8H1F/ZasRdjfAdXWccYUI70XqunYB7mWWH 6vQrDDt9NC7dkBta5HIDIhU= X-Google-Smtp-Source: APXvYqy24bhJZxnJNM/F0i7fclLKt26wgSxp0KGAPJ6m6mq/sBjQJXDJ7S1yh/UQDwwQYt2vwWmJFQ== X-Received: by 2002:a63:5818:: with SMTP id m24mr1905056pgb.358.1576038460800; Tue, 10 Dec 2019 20:27:40 -0800 (PST) Received: from localhost.localdomain ([12.176.148.120]) by smtp.gmail.com with ESMTPSA id m12sm591259pgr.87.2019.12.10.20.27.39 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Tue, 10 Dec 2019 20:27:40 -0800 (PST) From: SeongJae Park X-Google-Original-From: SeongJae Park To: jgross@suse.com, axboe@kernel.dk, konrad.wilk@oracle.com, roger.pau@citrix.com Cc: SeongJae Park , pdurrant@amazon.com, sjpark@amazon.com, xen-devel@lists.xenproject.org, linux-block@vger.kernel.org, linux-kernel@vger.kernel.org Subject: [PATCH v6 3/3] xen/blkback: Remove unnecessary static variable name prefixes Date: Wed, 11 Dec 2019 04:27:33 +0000 Message-Id: <20191211042733.6143-1-sjpark@amazon.de> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20191211042428.5961-1-sjpark@amazon.de> References: <20191211042428.5961-1-sjpark@amazon.de> Sender: linux-block-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-block@vger.kernel.org A few of static variables in blkback have 'xen_blkif_' prefix, though it is unnecessary for static variables. This commit removes such prefixes. Signed-off-by: SeongJae Park Reviewed-by: Roger Pau Monné --- drivers/block/xen-blkback/blkback.c | 37 +++++++++++++---------------- 1 file changed, 17 insertions(+), 20 deletions(-) diff --git a/drivers/block/xen-blkback/blkback.c b/drivers/block/xen-blkback/blkback.c index b493c306e84f..f690373669b8 100644 --- a/drivers/block/xen-blkback/blkback.c +++ b/drivers/block/xen-blkback/blkback.c @@ -62,8 +62,8 @@ * IO workloads. */ -static int xen_blkif_max_buffer_pages = 1024; -module_param_named(max_buffer_pages, xen_blkif_max_buffer_pages, int, 0644); +static int max_buffer_pages = 1024; +module_param_named(max_buffer_pages, max_buffer_pages, int, 0644); MODULE_PARM_DESC(max_buffer_pages, "Maximum number of free pages to keep in each block backend buffer"); @@ -78,8 +78,8 @@ MODULE_PARM_DESC(max_buffer_pages, * algorithm. */ -static int xen_blkif_max_pgrants = 1056; -module_param_named(max_persistent_grants, xen_blkif_max_pgrants, int, 0644); +static int max_pgrants = 1056; +module_param_named(max_persistent_grants, max_pgrants, int, 0644); MODULE_PARM_DESC(max_persistent_grants, "Maximum number of grants to map persistently"); @@ -88,8 +88,8 @@ MODULE_PARM_DESC(max_persistent_grants, * use. The time is in seconds, 0 means indefinitely long. */ -static unsigned int xen_blkif_pgrant_timeout = 60; -module_param_named(persistent_grant_unused_seconds, xen_blkif_pgrant_timeout, +static unsigned int pgrant_timeout = 60; +module_param_named(persistent_grant_unused_seconds, pgrant_timeout, uint, 0644); MODULE_PARM_DESC(persistent_grant_unused_seconds, "Time in seconds an unused persistent grant is allowed to " @@ -137,9 +137,8 @@ module_param(log_stats, int, 0644); static inline bool persistent_gnt_timeout(struct persistent_gnt *persistent_gnt) { - return xen_blkif_pgrant_timeout && - (jiffies - persistent_gnt->last_used >= - HZ * xen_blkif_pgrant_timeout); + return pgrant_timeout && (jiffies - persistent_gnt->last_used >= + HZ * pgrant_timeout); } /* Once a memory pressure is detected, squeeze free page pools for a while. */ @@ -249,7 +248,7 @@ static int add_persistent_gnt(struct xen_blkif_ring *ring, struct persistent_gnt *this; struct xen_blkif *blkif = ring->blkif; - if (ring->persistent_gnt_c >= xen_blkif_max_pgrants) { + if (ring->persistent_gnt_c >= max_pgrants) { if (!blkif->vbd.overflow_max_grants) blkif->vbd.overflow_max_grants = 1; return -EBUSY; @@ -412,14 +411,13 @@ static void purge_persistent_gnt(struct xen_blkif_ring *ring) goto out; } - if (ring->persistent_gnt_c < xen_blkif_max_pgrants || - (ring->persistent_gnt_c == xen_blkif_max_pgrants && + if (ring->persistent_gnt_c < max_pgrants || + (ring->persistent_gnt_c == max_pgrants && !ring->blkif->vbd.overflow_max_grants)) { num_clean = 0; } else { - num_clean = (xen_blkif_max_pgrants / 100) * LRU_PERCENT_CLEAN; - num_clean = ring->persistent_gnt_c - xen_blkif_max_pgrants + - num_clean; + num_clean = (max_pgrants / 100) * LRU_PERCENT_CLEAN; + num_clean = ring->persistent_gnt_c - max_pgrants + num_clean; num_clean = min(ring->persistent_gnt_c, num_clean); pr_debug("Going to purge at least %u persistent grants\n", num_clean); @@ -614,8 +612,7 @@ static void print_stats(struct xen_blkif_ring *ring) current->comm, ring->st_oo_req, ring->st_rd_req, ring->st_wr_req, ring->st_f_req, ring->st_ds_req, - ring->persistent_gnt_c, - xen_blkif_max_pgrants); + ring->persistent_gnt_c, max_pgrants); ring->st_print = jiffies + msecs_to_jiffies(10 * 1000); ring->st_rd_req = 0; ring->st_wr_req = 0; @@ -675,7 +672,7 @@ int xen_blkif_schedule(void *arg) if (time_before(jiffies, buffer_squeeze_end)) shrink_free_pagepool(ring, 0); else - shrink_free_pagepool(ring, xen_blkif_max_buffer_pages); + shrink_free_pagepool(ring, max_buffer_pages); if (log_stats && time_after(jiffies, ring->st_print)) print_stats(ring); @@ -902,7 +899,7 @@ static int xen_blkbk_map(struct xen_blkif_ring *ring, continue; } if (use_persistent_gnts && - ring->persistent_gnt_c < xen_blkif_max_pgrants) { + ring->persistent_gnt_c < max_pgrants) { /* * We are using persistent grants, the grant is * not mapped but we might have room for it. @@ -929,7 +926,7 @@ static int xen_blkbk_map(struct xen_blkif_ring *ring, pages[seg_idx]->persistent_gnt = persistent_gnt; pr_debug("grant %u added to the tree of persistent grants, using %u/%u\n", persistent_gnt->gnt, ring->persistent_gnt_c, - xen_blkif_max_pgrants); + max_pgrants); goto next; } if (use_persistent_gnts && !blkif->vbd.overflow_max_grants) {