From patchwork Wed May 9 07:48:27 2018 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Christoph Hellwig X-Patchwork-Id: 10388511 Return-Path: Received: from mail.wl.linuxfoundation.org (pdx-wl-mail.web.codeaurora.org [172.30.200.125]) by pdx-korg-patchwork.web.codeaurora.org (Postfix) with ESMTP id A870660170 for ; Wed, 9 May 2018 07:50:38 +0000 (UTC) Received: from mail.wl.linuxfoundation.org (localhost [127.0.0.1]) by mail.wl.linuxfoundation.org (Postfix) with ESMTP id 9B28F28E65 for ; Wed, 9 May 2018 07:50:38 +0000 (UTC) Received: by mail.wl.linuxfoundation.org (Postfix, from userid 486) id 8FA0228E56; Wed, 9 May 2018 07:50:38 +0000 (UTC) X-Spam-Checker-Version: SpamAssassin 3.3.1 (2010-03-16) on pdx-wl-mail.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-7.8 required=2.0 tests=BAYES_00,DKIM_SIGNED, MAILING_LIST_MULTI, RCVD_IN_DNSWL_HI, T_DKIM_INVALID autolearn=unavailable version=3.3.1 Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by mail.wl.linuxfoundation.org (Postfix) with ESMTP id 08A5F28E65 for ; Wed, 9 May 2018 07:50:38 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S933889AbeEIHuf (ORCPT ); Wed, 9 May 2018 03:50:35 -0400 Received: from bombadil.infradead.org ([198.137.202.133]:53674 "EHLO bombadil.infradead.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S933834AbeEIHud (ORCPT ); Wed, 9 May 2018 03:50:33 -0400 DKIM-Signature: v=1; a=rsa-sha256; q=dns/txt; c=relaxed/relaxed; d=infradead.org; s=bombadil.20170209; h=References:In-Reply-To:Message-Id: Date:Subject:Cc:To:From:Sender:Reply-To:MIME-Version:Content-Type: Content-Transfer-Encoding:Content-ID:Content-Description:Resent-Date: Resent-From:Resent-Sender:Resent-To:Resent-Cc:Resent-Message-ID:List-Id: List-Help:List-Unsubscribe:List-Subscribe:List-Post:List-Owner:List-Archive; bh=VoC7VbrLB/6sODMBkSI0wrQrr3gmK/G4hjH4PUEJn/s=; b=mpihslAD8DHxKMHfqKwwzaddg 4DqvzSdZ8794h7/seSrTQ+Q32q5FTQS63b0czHGqIZqzfyLDT9xPASqUPxLek72SSNOyWYZc8vV5n tbw2Mab3o1vsyyUHi0ItoIep6KzNIuAhu3rhaMsAduzT9p6CN4JWV2ShwFymurxS3kOZVrC+IXQwY qWBueZeEVEApeI4IIETOELC9PHMac3qH4GlgYrFRXlfwBOz6um34Q8rH3REfuKcSBiXich1eVIj+8 lC00kmqCDen1Z2/obZs5jOOLsnL9SdeuAV8oRldZ6oeXGBM50LPG7DA4mKcFEQM4haSzTH2iecfsZ CIjm5artg==; Received: from 213-225-15-246.nat.highway.a1.net ([213.225.15.246] helo=localhost) by bombadil.infradead.org with esmtpsa (Exim 4.90_1 #2 (Red Hat Linux)) id 1fGJrr-0003E7-LM; Wed, 09 May 2018 07:50:32 +0000 From: Christoph Hellwig To: linux-xfs@vger.kernel.org Cc: linux-fsdevel@vger.kernel.org, linux-block@vger.kernel.org, linux-mm@kvack.org Subject: [PATCH 30/33] iomap: add initial support for writes without buffer heads Date: Wed, 9 May 2018 09:48:27 +0200 Message-Id: <20180509074830.16196-31-hch@lst.de> X-Mailer: git-send-email 2.17.0 In-Reply-To: <20180509074830.16196-1-hch@lst.de> References: <20180509074830.16196-1-hch@lst.de> X-SRS-Rewrite: SMTP reverse-path rewritten from by bombadil.infradead.org. See http://www.infradead.org/rpr.html Sender: linux-xfs-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-xfs@vger.kernel.org X-Virus-Scanned: ClamAV using ClamSMTP For now just limited to blocksize == PAGE_SIZE, where we can simply read in the full page in write begin, and just set the whole page dirty after copying data into it. This code is enabled by default and XFS will now be feed pages without buffer heads in ->writepage and ->writepages. Signed-off-by: Christoph Hellwig --- fs/iomap.c | 129 +++++++++++++++++++++++++++++++++++++++++++++++++---- 1 file changed, 120 insertions(+), 9 deletions(-) diff --git a/fs/iomap.c b/fs/iomap.c index 967bd31540fe..a3861945504f 100644 --- a/fs/iomap.c +++ b/fs/iomap.c @@ -308,6 +308,56 @@ iomap_write_failed(struct inode *inode, loff_t pos, unsigned len) truncate_pagecache_range(inode, max(pos, i_size), pos + len); } +static int +iomap_read_page_sync(struct inode *inode, loff_t block_start, struct page *page, + unsigned poff, unsigned plen, struct iomap *iomap) +{ + struct bio_vec bvec; + struct bio bio; + int ret; + + bio_init(&bio, &bvec, 1); + bio.bi_opf = REQ_OP_READ; + bio.bi_iter.bi_sector = iomap_sector(iomap, block_start); + bio_set_dev(&bio, iomap->bdev); + __bio_add_page(&bio, page, plen, poff); + ret = submit_bio_wait(&bio); + if (ret < 0 && iomap_block_needs_zeroing(inode, block_start, iomap)) + zero_user(page, poff, plen); + return ret; +} + +static int +__iomap_write_begin(struct inode *inode, loff_t pos, unsigned len, + struct page *page, struct iomap *iomap) +{ + loff_t block_size = i_blocksize(inode); + loff_t block_start = pos & ~(block_size - 1); + loff_t block_end = (pos + len + block_size - 1) & ~(block_size - 1); + unsigned poff = block_start & (PAGE_SIZE - 1); + unsigned plen = min_t(loff_t, PAGE_SIZE - poff, block_end - block_start); + int status; + + if (PageUptodate(page)) + return 0; + + if (iomap_block_needs_zeroing(inode, block_start, iomap)) { + unsigned from = pos & (PAGE_SIZE - 1), to = from + len; + unsigned pend = poff + plen; + + if (poff < from || pend > to) + zero_user_segments(page, poff, from, to, pend); + } else { + status = iomap_read_page_sync(inode, block_start, page, + poff, plen, iomap); + if (status < 0) + return status; + SetPageUptodate(page); + } + + return 0; +} + static int iomap_write_begin(struct inode *inode, loff_t pos, unsigned len, unsigned flags, struct page **pagep, struct iomap *iomap) @@ -325,7 +375,10 @@ iomap_write_begin(struct inode *inode, loff_t pos, unsigned len, unsigned flags, if (!page) return -ENOMEM; - status = __block_write_begin_int(page, pos, len, NULL, iomap); + if (i_blocksize(inode) == PAGE_SIZE) + status = __iomap_write_begin(inode, pos, len, page, iomap); + else + status = __block_write_begin_int(page, pos, len, NULL, iomap); if (unlikely(status)) { unlock_page(page); put_page(page); @@ -338,12 +391,63 @@ iomap_write_begin(struct inode *inode, loff_t pos, unsigned len, unsigned flags, return status; } +static int +iomap_set_page_dirty(struct page *page) +{ + struct address_space *mapping = page_mapping(page); + int newly_dirty; + + if (unlikely(!mapping)) + return !TestSetPageDirty(page); + + /* + * Lock out page->mem_cgroup migration to keep PageDirty + * synchronized with per-memcg dirty page counters. + */ + lock_page_memcg(page); + newly_dirty = !TestSetPageDirty(page); + if (newly_dirty) + __set_page_dirty(page, mapping, 0); + unlock_page_memcg(page); + + if (newly_dirty) + __mark_inode_dirty(mapping->host, I_DIRTY_PAGES); + return newly_dirty; +} + +static int +__iomap_write_end(struct inode *inode, loff_t pos, unsigned len, + unsigned copied, struct page *page, struct iomap *iomap) +{ + unsigned start = pos & (PAGE_SIZE - 1); + int ret; + + if (unlikely(copied < len)) { + /* see block_write_end() for an explanation */ + if (!PageUptodate(page)) + copied = 0; + if (iomap_block_needs_zeroing(inode, pos, iomap)) + zero_user(page, start + copied, len - copied); + } + + flush_dcache_page(page); + SetPageUptodate(page); + iomap_set_page_dirty(page); + ret = __generic_write_end(inode, pos, copied, page); + if (ret < len) + iomap_write_failed(inode, pos, len); + return ret; +} + static int iomap_write_end(struct inode *inode, loff_t pos, unsigned len, - unsigned copied, struct page *page) + unsigned copied, struct page *page, struct iomap *iomap) { int ret; + if (i_blocksize(inode) == PAGE_SIZE) + return __iomap_write_end(inode, pos, len, copied, page, iomap); + ret = generic_write_end(NULL, inode->i_mapping, pos, len, copied, page, NULL); if (ret < len) @@ -400,7 +504,8 @@ iomap_write_actor(struct inode *inode, loff_t pos, loff_t length, void *data, flush_dcache_page(page); - status = iomap_write_end(inode, pos, bytes, copied, page); + status = iomap_write_end(inode, pos, bytes, copied, page, + iomap); if (unlikely(status < 0)) break; copied = status; @@ -494,7 +599,7 @@ iomap_dirty_actor(struct inode *inode, loff_t pos, loff_t length, void *data, WARN_ON_ONCE(!PageUptodate(page)); - status = iomap_write_end(inode, pos, bytes, bytes, page); + status = iomap_write_end(inode, pos, bytes, bytes, page, iomap); if (unlikely(status <= 0)) { if (WARN_ON_ONCE(status == 0)) return -EIO; @@ -546,7 +651,7 @@ static int iomap_zero(struct inode *inode, loff_t pos, unsigned offset, zero_user(page, offset, bytes); mark_page_accessed(page); - return iomap_write_end(inode, pos, bytes, bytes, page); + return iomap_write_end(inode, pos, bytes, bytes, page, iomap); } static int iomap_dax_zero(loff_t pos, unsigned offset, unsigned bytes, @@ -632,11 +737,14 @@ iomap_page_mkwrite_actor(struct inode *inode, loff_t pos, loff_t length, struct page *page = data; int ret; - ret = __block_write_begin_int(page, pos, length, NULL, iomap); - if (ret) - return ret; + if (i_blocksize(inode) != PAGE_SIZE) { + ret = __block_write_begin_int(page, pos, length, NULL, iomap); + if (ret) + return ret; + + block_commit_write(page, 0, length); + } - block_commit_write(page, 0, length); return length; } @@ -663,6 +771,9 @@ int iomap_page_mkwrite(struct vm_fault *vmf, const struct iomap_ops *ops) else length = PAGE_SIZE; + if (i_blocksize(inode) == PAGE_SIZE) + WARN_ON_ONCE(!PageUptodate(page)); + offset = page_offset(page); while (length > 0) { ret = iomap_apply(inode, offset, length,