From patchwork Mon Jan 13 15:37:43 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Matthew Wilcox X-Patchwork-Id: 11330467 Return-Path: Received: from mail.kernel.org (pdx-korg-mail-1.web.codeaurora.org [172.30.200.123]) by pdx-korg-patchwork-2.web.codeaurora.org (Postfix) with ESMTP id A6F031398 for ; Mon, 13 Jan 2020 15:38:07 +0000 (UTC) Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by mail.kernel.org (Postfix) with ESMTP id 66E432081E for ; Mon, 13 Jan 2020 15:38:07 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=fail reason="signature verification failed" (2048-bit key) header.d=infradead.org header.i=@infradead.org header.b="idavc8Cc" DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 66E432081E Authentication-Results: mail.kernel.org; dmarc=none (p=none dis=none) header.from=infradead.org Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=owner-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix) id C6E958E0003; Mon, 13 Jan 2020 10:37:55 -0500 (EST) Delivered-To: linux-mm-outgoing@kvack.org Received: by kanga.kvack.org (Postfix, from userid 40) id BB07A8E0014; Mon, 13 Jan 2020 10:37:55 -0500 (EST) X-Original-To: int-list-linux-mm@kvack.org X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 788E98E0012; Mon, 13 Jan 2020 10:37:55 -0500 (EST) X-Original-To: linux-mm@kvack.org X-Delivered-To: linux-mm@kvack.org Received: from forelay.hostedemail.com (smtprelay0215.hostedemail.com [216.40.44.215]) by kanga.kvack.org (Postfix) with ESMTP id 285288E0003 for ; Mon, 13 Jan 2020 10:37:55 -0500 (EST) Received: from smtpin15.hostedemail.com (10.5.19.251.rfc1918.com [10.5.19.251]) by forelay03.hostedemail.com (Postfix) with SMTP id DB038824934B for ; Mon, 13 Jan 2020 15:37:54 +0000 (UTC) X-FDA: 76373016468.15.time84_7f3b60d3fa814 X-Spam-Summary: 2,0,0,6e0ba34d8df48cf0,d41d8cd98f00b204,willy@infradead.org,:linux-xfs@vger.kernel.org:linux-fsdevel@vger.kernel.org::willy@infradead.org:jlayton@kernel.org:hch@infradead.org,RULES_HIT:2:41:69:355:379:541:800:960:973:988:989:1260:1311:1314:1345:1359:1437:1515:1535:1605:1730:1747:1777:1792:2198:2199:2393:2559:2562:3138:3139:3140:3141:3142:3865:3867:3868:3870:3871:3872:3874:4050:4120:4250:4321:4605:5007:6119:6261:6653:7576:7903:8660:9592:10004:11026:11232:11473:11658:11914:12043:12296:12297:12438:12555:12683:12895:12986:13148:13230:13894:14096:14110:14394:21080:21451:21627:21809:21990:30029:30054:30070,0,RBL:198.137.202.133:@infradead.org:.lbl8.mailshell.net-62.8.0.100 64.201.201.201,CacheIP:none,Bayesian:0.5,0.5,0.5,Netcheck:none,DomainCache:0,MSF:not bulk,SPF:fn,MSBL:0,DNSBL:neutral,Custom_rules:0:0:0,LFtime:24,LUA_SUMMARY:none X-HE-Tag: time84_7f3b60d3fa814 X-Filterd-Recvd-Size: 9360 Received: from bombadil.infradead.org (bombadil.infradead.org [198.137.202.133]) by imf03.hostedemail.com (Postfix) with ESMTP for ; Mon, 13 Jan 2020 15:37:54 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; q=dns/txt; c=relaxed/relaxed; d=infradead.org; s=bombadil.20170209; h=Content-Transfer-Encoding: MIME-Version:References:In-Reply-To:Message-Id:Date:Subject:Cc:To:From:Sender :Reply-To:Content-Type:Content-ID:Content-Description:Resent-Date:Resent-From :Resent-Sender:Resent-To:Resent-Cc:Resent-Message-ID:List-Id:List-Help: List-Unsubscribe:List-Subscribe:List-Post:List-Owner:List-Archive; bh=XzLp8VMemTuCcHOHx3h9ZSjkeqJlb8tYYS2nICefEuM=; b=idavc8Cc/pyoNqz+6SqHd54F6q dKO2M2mVSfeeLE708RbcR+IFE5XWFq0q8HyEUhSjwweiy3IIZXX9QNKXKZwMgzIQW/KU2IWavOOfD K14AmlaHnSJm9I5bErummNEc6FJgCxWI63FW3c6O7Ofjwl71bVTXQ6foxzRFfUDgO1sQvc9Q2wikF /dlrNIsYGq01es9DhIPtAHN2dJ2QDEuq+ILiCUuSIcf2ydL97xJ/JCuwj8a541Yuw7XRUqCoG7wej 1bfawDu8CNs2pR3+8WJY4Fs91nn0h1NaMWmTWfkNZikGM/KUQKtPoXXBS+glzTzwpBOHnvs842iTp QP1uoiOg==; Received: from willy by bombadil.infradead.org with local (Exim 4.92.3 #3 (Red Hat Linux)) id 1ir1mr-00076U-7n; Mon, 13 Jan 2020 15:37:53 +0000 From: Matthew Wilcox To: linux-xfs@vger.kernel.org, linux-fsdevel@vger.kernel.org, linux-mm@kvack.org Cc: "Matthew Wilcox (Oracle)" , jlayton@kernel.org, hch@infradead.org Subject: [PATCH 5/8] iomap,xfs: Convert from readpages to readahead Date: Mon, 13 Jan 2020 07:37:43 -0800 Message-Id: <20200113153746.26654-6-willy@infradead.org> X-Mailer: git-send-email 2.21.0 In-Reply-To: <20200113153746.26654-1-willy@infradead.org> References: <20200113153746.26654-1-willy@infradead.org> MIME-Version: 1.0 X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: From: "Matthew Wilcox (Oracle)" Use the new readahead operation in XFS and iomap. Signed-off-by: Matthew Wilcox (Oracle) --- fs/iomap/buffered-io.c | 60 +++++++++--------------------------------- fs/iomap/trace.h | 18 ++++++------- fs/xfs/xfs_aops.c | 12 ++++----- include/linux/iomap.h | 4 +-- 4 files changed, 29 insertions(+), 65 deletions(-) diff --git a/fs/iomap/buffered-io.c b/fs/iomap/buffered-io.c index 828444e14d09..818fa5bbd643 100644 --- a/fs/iomap/buffered-io.c +++ b/fs/iomap/buffered-io.c @@ -8,6 +8,7 @@ #include #include #include +#include #include #include #include @@ -216,7 +217,7 @@ struct iomap_readpage_ctx { bool cur_page_in_bio; bool is_readahead; struct bio *bio; - struct list_head *pages; + struct pagevec *pages; }; static void @@ -337,7 +338,7 @@ iomap_readpage(struct page *page, const struct iomap_ops *ops) unsigned poff; loff_t ret; - trace_iomap_readpage(page->mapping->host, 1); + trace_iomap_readpage(page->mapping->host, (loff_t)PAGE_SIZE); for (poff = 0; poff < PAGE_SIZE; poff += ret) { ret = iomap_apply(inode, page_offset(page) + poff, @@ -367,36 +368,8 @@ iomap_readpage(struct page *page, const struct iomap_ops *ops) } EXPORT_SYMBOL_GPL(iomap_readpage); -static struct page * -iomap_next_page(struct inode *inode, struct list_head *pages, loff_t pos, - loff_t length, loff_t *done) -{ - while (!list_empty(pages)) { - struct page *page = lru_to_page(pages); - - if (page_offset(page) >= (u64)pos + length) - break; - - list_del(&page->lru); - if (!add_to_page_cache_lru(page, inode->i_mapping, page->index, - GFP_NOFS)) - return page; - - /* - * If we already have a page in the page cache at index we are - * done. Upper layers don't care if it is uptodate after the - * readpages call itself as every page gets checked again once - * actually needed. - */ - *done += PAGE_SIZE; - put_page(page); - } - - return NULL; -} - static loff_t -iomap_readpages_actor(struct inode *inode, loff_t pos, loff_t length, +iomap_readahead_actor(struct inode *inode, loff_t pos, loff_t length, void *data, struct iomap *iomap, struct iomap *srcmap) { struct iomap_readpage_ctx *ctx = data; @@ -410,8 +383,7 @@ iomap_readpages_actor(struct inode *inode, loff_t pos, loff_t length, ctx->cur_page = NULL; } if (!ctx->cur_page) { - ctx->cur_page = iomap_next_page(inode, ctx->pages, - pos, length, &done); + ctx->cur_page = pagevec_next(ctx->pages); if (!ctx->cur_page) break; ctx->cur_page_in_bio = false; @@ -423,23 +395,22 @@ iomap_readpages_actor(struct inode *inode, loff_t pos, loff_t length, return done; } -int -iomap_readpages(struct address_space *mapping, struct list_head *pages, - unsigned nr_pages, const struct iomap_ops *ops) +void iomap_readahead(struct address_space *mapping, struct pagevec *pages, + pgoff_t index, const struct iomap_ops *ops) { struct iomap_readpage_ctx ctx = { .pages = pages, .is_readahead = true, }; - loff_t pos = page_offset(list_entry(pages->prev, struct page, lru)); - loff_t last = page_offset(list_entry(pages->next, struct page, lru)); + loff_t pos = (loff_t)index << PAGE_SHIFT; + loff_t last = page_offset(pagevec_last(pages)); loff_t length = last - pos + PAGE_SIZE, ret = 0; - trace_iomap_readpages(mapping->host, nr_pages); + trace_iomap_readahead(mapping->host, length); while (length > 0) { ret = iomap_apply(mapping->host, pos, length, 0, ops, - &ctx, iomap_readpages_actor); + &ctx, iomap_readahead_actor); if (ret <= 0) { WARN_ON_ONCE(ret == 0); goto done; @@ -456,15 +427,8 @@ iomap_readpages(struct address_space *mapping, struct list_head *pages, unlock_page(ctx.cur_page); put_page(ctx.cur_page); } - - /* - * Check that we didn't lose a page due to the arcance calling - * conventions.. - */ - WARN_ON_ONCE(!ret && !list_empty(ctx.pages)); - return ret; } -EXPORT_SYMBOL_GPL(iomap_readpages); +EXPORT_SYMBOL_GPL(iomap_readahead); /* * iomap_is_partially_uptodate checks whether blocks within a page are diff --git a/fs/iomap/trace.h b/fs/iomap/trace.h index 6dc227b8c47e..adbfd9fd4275 100644 --- a/fs/iomap/trace.h +++ b/fs/iomap/trace.h @@ -16,30 +16,30 @@ struct inode; DECLARE_EVENT_CLASS(iomap_readpage_class, - TP_PROTO(struct inode *inode, int nr_pages), - TP_ARGS(inode, nr_pages), + TP_PROTO(struct inode *inode, loff_t length), + TP_ARGS(inode, length), TP_STRUCT__entry( __field(dev_t, dev) __field(u64, ino) - __field(int, nr_pages) + __field(loff_t, length) ), TP_fast_assign( __entry->dev = inode->i_sb->s_dev; __entry->ino = inode->i_ino; - __entry->nr_pages = nr_pages; + __entry->length = length; ), - TP_printk("dev %d:%d ino 0x%llx nr_pages %d", + TP_printk("dev %d:%d ino 0x%llx length %lld", MAJOR(__entry->dev), MINOR(__entry->dev), __entry->ino, - __entry->nr_pages) + __entry->length) ) #define DEFINE_READPAGE_EVENT(name) \ DEFINE_EVENT(iomap_readpage_class, name, \ - TP_PROTO(struct inode *inode, int nr_pages), \ - TP_ARGS(inode, nr_pages)) + TP_PROTO(struct inode *inode, loff_t length), \ + TP_ARGS(inode, length)) DEFINE_READPAGE_EVENT(iomap_readpage); -DEFINE_READPAGE_EVENT(iomap_readpages); +DEFINE_READPAGE_EVENT(iomap_readahead); DECLARE_EVENT_CLASS(iomap_page_class, TP_PROTO(struct inode *inode, struct page *page, unsigned long off, diff --git a/fs/xfs/xfs_aops.c b/fs/xfs/xfs_aops.c index 3a688eb5c5ae..e3db35bcfa34 100644 --- a/fs/xfs/xfs_aops.c +++ b/fs/xfs/xfs_aops.c @@ -621,14 +621,14 @@ xfs_vm_readpage( return iomap_readpage(page, &xfs_read_iomap_ops); } -STATIC int -xfs_vm_readpages( +STATIC void +xfs_vm_readahead( struct file *unused, struct address_space *mapping, - struct list_head *pages, - unsigned nr_pages) + struct pagevec *pages, + pgoff_t index) { - return iomap_readpages(mapping, pages, nr_pages, &xfs_read_iomap_ops); + iomap_readahead(mapping, pages, index, &xfs_read_iomap_ops); } static int @@ -644,7 +644,7 @@ xfs_iomap_swapfile_activate( const struct address_space_operations xfs_address_space_operations = { .readpage = xfs_vm_readpage, - .readpages = xfs_vm_readpages, + .readahead = xfs_vm_readahead, .writepage = xfs_vm_writepage, .writepages = xfs_vm_writepages, .set_page_dirty = iomap_set_page_dirty, diff --git a/include/linux/iomap.h b/include/linux/iomap.h index 8b09463dae0d..1af1ec0920d8 100644 --- a/include/linux/iomap.h +++ b/include/linux/iomap.h @@ -155,8 +155,8 @@ loff_t iomap_apply(struct inode *inode, loff_t pos, loff_t length, ssize_t iomap_file_buffered_write(struct kiocb *iocb, struct iov_iter *from, const struct iomap_ops *ops); int iomap_readpage(struct page *page, const struct iomap_ops *ops); -int iomap_readpages(struct address_space *mapping, struct list_head *pages, - unsigned nr_pages, const struct iomap_ops *ops); +void iomap_readahead(struct address_space *mapping, struct pagevec *pages, + pgoff_t index, const struct iomap_ops *ops); int iomap_set_page_dirty(struct page *page); int iomap_is_partially_uptodate(struct page *page, unsigned long from, unsigned long count);