Message ID | 20240202073104.2418230-7-dlemoal@kernel.org (mailing list archive) |
---|---|
State | Not Applicable |
Headers | show |
Series | Zone write plugging | expand |
On Fri, Feb 02, 2024 at 04:30:44PM +0900, Damien Le Moal wrote: > Zone write plugging implements a per-zone "plug" for write operations to > tightly control the submission and execution order of writes to > sequential write required zones of a zoned block device. Per-zone > plugging of writes guarantees that at any time at most one write request > per zone is in flight. This mechanism is intended to replace zone write > locking which is controlled at the scheduler level and implemented only > by mq-deadline. > > Unlike zone write locking which operates on requests, zone write > plugging operates on BIOs. A zone write plug is simply a BIO list that > is atomically manipulated using a spinlock and a kblockd submission > work. A write BIO to a zone is "plugged" to delay its execution if a > write BIO for the same zone was already issued, that is, if a write > request for the same zone is being executed. The next plugged BIO is > unplugged and issued once the write request completes. > > This mechanism allows to: > - Untangles zone write ordering from block IO schedulers. This allows > removing the restriction on using only mq-deadline for zoned block > devices. Any block IO scheduler, including "none" can be used. > - Zone write plugging operates on BIOs instead of requests. Plugged > BIOs waiting for execution thus do not hold scheduling tags and thus > are not preventing other BIOs to proceed (reads or writes to other > zones). Depending on the workload, this can significantly improve > the device use and performance. > - Both blk-mq (request) based zoned devices and BIO-based devices (e.g. > device mapper) can use zone write plugging. It is mandatory for the > former but optional for the latter: BIO-based driver can use zone > write plugging to implement write ordering guarantees, or the drivers > can implement their own if needed. > - The code is less invasive in the block layer and is mostly limited to > blk-zoned.c with some small changes in blk-mq.c, blk-merge.c and > bio.c. > > Zone write plugging is implemented using struct blk_zone_wplug. This > structurei includes a spinlock, a BIO list and a work structure to > handle the submission of plugged BIOs. > > Plugging of zone write BIOs is done using the function > blk_zone_write_plug_bio() which returns false if a BIO execution does > not need to be delayed and true otherwise. This function is called > from blk_mq_submit_bio() after a BIO is split to avoid large BIOs > spanning multiple zones which would cause mishandling of zone write > plugging. This enables by default zone write plugging for any mq > request-based block device. BIO-based device drivers can also use zone > write plugging by expliclty calling blk_zone_write_plug_bio() in their > ->submit_bio method. For such devices, the driver must ensure that a > BIO passed to blk_zone_write_plug_bio() is already split and not > straddling zone boundaries. > > Only write and write zeroes BIOs are plugged. Zone write plugging does > not introduce any significant overhead for other operations. A BIO that > is being handled through zone write plugging is flagged using the new > BIO flag BIO_ZONE_WRITE_PLUGGING. A request handling a BIO flagged with > this new flag is flagged with the new RQF_ZONE_WRITE_PLUGGING flag. > The completion processing of BIOs and requests flagged trigger > respectively calls to the functions blk_zone_write_plug_bio_endio() and > blk_zone_write_plug_complete_request(). The latter function is used to > trigger submission of the next plugged BIO using the zone plug work. > blk_zone_write_plug_bio_endio() does the same for BIO-based devices. > This ensures that at any time, at most one request (blk-mq devices) or > one BIO (BIO-based devices) are being executed for any zone. The > handling of zone write plug using a per-zone plug spinlock maximizes > parrallelism and device usage by allowing multiple zones to be writen > simultaneously without lock contention. > > Zone write plugging ignores flush BIOs without data. Hovever, any flush > BIO that has data is always plugged so that the write part of the flush > sequence is serialized with other regular writes. > > Given that any BIO handled through zone write plugging will be the only > BIO in flight for the target zone when it is executed, the unplugging > and submission of a BIO will have no chance of successfully merging with > plugged requests or requests in the scheduler. To overcome this > potential performance loss, blk_mq_submit_bio() calls the function > blk_zone_write_plug_attempt_merge() to try to merge other plugged BIOs > with the one just unplugged. Successful merging is signaled using > blk_zone_write_plug_bio_merged(), called from bio_attempt_back_merge(). > Furthermore, to avoid recalculating the number of segments of plugged > BIOs to attempt merging, the number of segments of a plugged BIO is > saved using the new struct bio field __bi_nr_segments. To avoid growing > the size of struct bio, this field is added as a union with the > bio_cookie field. This is safe to do as polling is always disabled for > plugged BIOs. > > When BIOs are plugged in a zone write plug, the device request queue > usage counter is always incremented. This kept and reused when the > plugged BIO is unplugged and submitted again using > submit_bio_noacct_nocheck(). For this case, the unplugged BIO is already > flagged with BIO_ZONE_WRITE_PLUGGING and blk_mq_submit_bio() proceeds > directly to allocating a new request for the BIO, re-using the usage > reference count taken when the BIO was plugged. This extra reference > count is dropped in blk_zone_write_plug_attempt_merge() for any plugged > BIO that is successfully merged. Given that BIO-based devices will not > take this path, the extra reference is dropped when a plugged BIO is > unplugged and submitted. > > To match the new data structures used for zoned disks, the function > disk_free_zone_bitmaps() is renamed to the more generic > disk_free_zone_resources(). > > This commit contains contributions from Christoph Hellwig <hch@lst.de>. > > Signed-off-by: Damien Le Moal <dlemoal@kernel.org> > --- > block/bio.c | 7 + > block/blk-merge.c | 11 + > block/blk-mq.c | 28 +++ > block/blk-zoned.c | 408 +++++++++++++++++++++++++++++++++++++- > block/blk.h | 32 ++- > block/genhd.c | 2 +- > include/linux/blk-mq.h | 2 + > include/linux/blk_types.h | 8 +- > include/linux/blkdev.h | 8 + > 9 files changed, 496 insertions(+), 10 deletions(-) > > diff --git a/block/bio.c b/block/bio.c > index b9642a41f286..c8b0f7e8c713 100644 > --- a/block/bio.c > +++ b/block/bio.c > @@ -1581,6 +1581,13 @@ void bio_endio(struct bio *bio) > if (!bio_integrity_endio(bio)) > return; > > + /* > + * For BIOs handled through a zone write plugs, signal the end of the > + * BIO to the zone write plug to submit the next plugged BIO. > + */ > + if (bio_zone_write_plugging(bio)) > + blk_zone_write_plug_bio_endio(bio); > + > rq_qos_done_bio(bio); > > if (bio->bi_bdev && bio_flagged(bio, BIO_TRACE_COMPLETION)) { > diff --git a/block/blk-merge.c b/block/blk-merge.c > index a1ef61b03e31..2b5489cd9c65 100644 > --- a/block/blk-merge.c > +++ b/block/blk-merge.c > @@ -377,6 +377,7 @@ struct bio *__bio_split_to_limits(struct bio *bio, > blkcg_bio_issue_init(split); > bio_chain(split, bio); > trace_block_split(split, bio->bi_iter.bi_sector); > + WARN_ON_ONCE(bio_zone_write_plugging(bio)); > submit_bio_noacct(bio); > return split; > } > @@ -980,6 +981,9 @@ enum bio_merge_status bio_attempt_back_merge(struct request *req, > > blk_update_mixed_merge(req, bio, false); > > + if (req->rq_flags & RQF_ZONE_WRITE_PLUGGING) > + blk_zone_write_plug_bio_merged(bio); > + > req->biotail->bi_next = bio; > req->biotail = bio; > req->__data_len += bio->bi_iter.bi_size; > @@ -995,6 +999,13 @@ static enum bio_merge_status bio_attempt_front_merge(struct request *req, > { > const blk_opf_t ff = bio_failfast(bio); > > + /* > + * A front merge for zone writes can happen only if the user submitted > + * writes out of order. Do not attempt this to let the write fail. > + */ > + if (req->rq_flags & RQF_ZONE_WRITE_PLUGGING) > + return BIO_MERGE_FAILED; > + > if (!ll_front_merge_fn(req, bio, nr_segs)) > return BIO_MERGE_FAILED; > > diff --git a/block/blk-mq.c b/block/blk-mq.c > index f02e486a02ae..aa49bebf1199 100644 > --- a/block/blk-mq.c > +++ b/block/blk-mq.c > @@ -830,6 +830,9 @@ static void blk_complete_request(struct request *req) > bio = next; > } while (bio); > > + if (req->rq_flags & RQF_ZONE_WRITE_PLUGGING) > + blk_zone_write_plug_complete_request(req); > + > /* > * Reset counters so that the request stacking driver > * can find how many bytes remain in the request > @@ -943,6 +946,9 @@ bool blk_update_request(struct request *req, blk_status_t error, > * completely done > */ > if (!req->bio) { > + if (req->rq_flags & RQF_ZONE_WRITE_PLUGGING) > + blk_zone_write_plug_complete_request(req); > + > /* > * Reset counters so that the request stacking driver > * can find how many bytes remain in the request > @@ -2975,6 +2981,17 @@ void blk_mq_submit_bio(struct bio *bio) > struct request *rq; > blk_status_t ret; > > + /* > + * A BIO that was released form a zone write plug has already been > + * through the preparation in this function, already holds a reference > + * on the queue usage counter, and is the only write BIO in-flight for > + * the target zone. Go straight to allocating a request for it. > + */ > + if (bio_zone_write_plugging(bio)) { > + nr_segs = bio->__bi_nr_segments; > + goto new_request; > + } > + > bio = blk_queue_bounce(bio, q); > bio_set_ioprio(bio); > > @@ -3001,7 +3018,11 @@ void blk_mq_submit_bio(struct bio *bio) > if (blk_mq_attempt_bio_merge(q, bio, nr_segs)) > goto queue_exit; > > + if (blk_queue_is_zoned(q) && blk_zone_write_plug_bio(bio, nr_segs)) > + goto queue_exit; > + > if (!rq) { > +new_request: > rq = blk_mq_get_new_requests(q, plug, bio, nr_segs); > if (unlikely(!rq)) > goto queue_exit; > @@ -3017,8 +3038,12 @@ void blk_mq_submit_bio(struct bio *bio) > > ret = blk_crypto_rq_get_keyslot(rq); > if (ret != BLK_STS_OK) { > + bool zwplugging = bio_zone_write_plugging(bio); > + > bio->bi_status = ret; > bio_endio(bio); > + if (zwplugging) > + blk_zone_write_plug_complete_request(rq); > blk_mq_free_request(rq); > return; > } > @@ -3026,6 +3051,9 @@ void blk_mq_submit_bio(struct bio *bio) > if (op_is_flush(bio->bi_opf) && blk_insert_flush(rq)) > return; > > + if (bio_zone_write_plugging(bio)) > + blk_zone_write_plug_attempt_merge(rq); > + > if (plug) { > blk_add_rq_to_plug(plug, rq); > return; > diff --git a/block/blk-zoned.c b/block/blk-zoned.c > index d343e5756a9c..f6d4f511b664 100644 > --- a/block/blk-zoned.c > +++ b/block/blk-zoned.c > @@ -7,11 +7,11 @@ > * > * Copyright (c) 2016, Damien Le Moal > * Copyright (c) 2016, Western Digital > + * Copyright (c) 2024, Western Digital Corporation or its affiliates. > */ > > #include <linux/kernel.h> > #include <linux/module.h> > -#include <linux/rbtree.h> > #include <linux/blkdev.h> > #include <linux/blk-mq.h> > #include <linux/mm.h> > @@ -19,6 +19,7 @@ > #include <linux/sched/mm.h> > > #include "blk.h" > +#include "blk-mq-sched.h" > > #define ZONE_COND_NAME(name) [BLK_ZONE_COND_##name] = #name > static const char *const zone_cond_name[] = { > @@ -33,6 +34,27 @@ static const char *const zone_cond_name[] = { > }; > #undef ZONE_COND_NAME > > +/* > + * Per-zone write plug. > + */ > +struct blk_zone_wplug { > + spinlock_t lock; > + unsigned int flags; > + struct bio_list bio_list; > + struct work_struct bio_work; > +}; > + > +/* > + * Zone write plug flags bits: > + * - BLK_ZONE_WPLUG_CONV: Indicate that the zone is a conventional one. Writes > + * to these zones are never plugged. > + * - BLK_ZONE_WPLUG_PLUGGED: Indicate that the zone write plug is plugged, > + * that is, that write BIOs are being throttled due to a write BIO already > + * being executed or the zone write plug bio list is not empty. > + */ > +#define BLK_ZONE_WPLUG_CONV (1U << 0) > +#define BLK_ZONE_WPLUG_PLUGGED (1U << 1) BLK_ZONE_WPLUG_PLUGGED == !bio_list_empty(&zwplug->bio_list), so looks this flag isn't necessary. > + > /** > * blk_zone_cond_str - Return string XXX in BLK_ZONE_COND_XXX. > * @zone_cond: BLK_ZONE_COND_XXX. > @@ -429,12 +451,374 @@ int blkdev_zone_mgmt_ioctl(struct block_device *bdev, blk_mode_t mode, > return ret; > } > > -void disk_free_zone_bitmaps(struct gendisk *disk) > +#define blk_zone_wplug_lock(zwplug, flags) \ > + spin_lock_irqsave(&zwplug->lock, flags) > + > +#define blk_zone_wplug_unlock(zwplug, flags) \ > + spin_unlock_irqrestore(&zwplug->lock, flags) > + > +static inline void blk_zone_wplug_bio_io_error(struct bio *bio) > +{ > + struct request_queue *q = bio->bi_bdev->bd_disk->queue; > + > + bio_clear_flag(bio, BIO_ZONE_WRITE_PLUGGING); > + bio_io_error(bio); > + blk_queue_exit(q); > +} > + > +static int blk_zone_wplug_abort(struct gendisk *disk, > + struct blk_zone_wplug *zwplug) > +{ > + struct bio *bio; > + int nr_aborted = 0; > + > + while ((bio = bio_list_pop(&zwplug->bio_list))) { > + blk_zone_wplug_bio_io_error(bio); > + nr_aborted++; > + } > + > + return nr_aborted; > +} > + > +/* > + * Return the zone write plug for sector in sequential write required zone. > + * Given that conventional zones have no write ordering constraints, NULL is > + * returned for sectors in conventional zones, to indicate that zone write > + * plugging is not needed. > + */ > +static inline struct blk_zone_wplug * > +disk_lookup_zone_wplug(struct gendisk *disk, sector_t sector) > +{ > + struct blk_zone_wplug *zwplug; > + > + if (WARN_ON_ONCE(!disk->zone_wplugs)) > + return NULL; > + > + zwplug = &disk->zone_wplugs[disk_zone_no(disk, sector)]; > + if (zwplug->flags & BLK_ZONE_WPLUG_CONV) > + return NULL; > + return zwplug; > +} > + > +static inline struct blk_zone_wplug *bio_lookup_zone_wplug(struct bio *bio) > +{ > + return disk_lookup_zone_wplug(bio->bi_bdev->bd_disk, > + bio->bi_iter.bi_sector); > +} > + > +static inline void blk_zone_wplug_add_bio(struct blk_zone_wplug *zwplug, > + struct bio *bio, unsigned int nr_segs) > +{ > + /* > + * Keep a reference on the BIO request queue usage. This reference will > + * be dropped either if the BIO is failed or after it is issued and > + * completes. > + */ > + percpu_ref_get(&bio->bi_bdev->bd_disk->queue->q_usage_counter); It is fragile to get nested usage_counter, and same with grabbing/releasing it from different contexts or even functions, and it could be much better to just let block layer maintain it. From patch 23's change: + * Zoned block device information. Reads of this information must be + * protected with blk_queue_enter() / blk_queue_exit(). Modifying this Anytime if there is in-flight bio, the block device is opened, so both gendisk and request_queue are live, so not sure if this .q_usage_counter protection is needed. + * information is only allowed while no requests are being processed. + * See also blk_mq_freeze_queue() and blk_mq_unfreeze_queue(). */ > + > + /* > + * The BIO is being plugged and thus will have to wait for the on-going > + * write and for all other writes already plugged. So polling makes > + * no sense. > + */ > + bio_clear_polled(bio); > + > + /* > + * Reuse the poll cookie field to store the number of segments when > + * split to the hardware limits. > + */ > + bio->__bi_nr_segments = nr_segs; > + > + /* > + * We always receive BIOs after they are split and ready to be issued. > + * The block layer passes the parts of a split BIO in order, and the > + * user must also issue write sequentially. So simply add the new BIO > + * at the tail of the list to preserve the sequential write order. > + */ > + bio_list_add(&zwplug->bio_list, bio); > +} > + > +/* > + * Called from bio_attempt_back_merge() when a BIO was merged with a request. > + */ > +void blk_zone_write_plug_bio_merged(struct bio *bio) > +{ > + bio_set_flag(bio, BIO_ZONE_WRITE_PLUGGING); > +} > + > +/* > + * Attempt to merge plugged BIOs with a newly formed request of a BIO that went > + * through zone write plugging (either a new BIO or one that was unplugged). > + */ > +void blk_zone_write_plug_attempt_merge(struct request *req) > +{ > + struct blk_zone_wplug *zwplug = bio_lookup_zone_wplug(req->bio); > + sector_t req_back_sector = blk_rq_pos(req) + blk_rq_sectors(req); > + struct request_queue *q = req->q; > + unsigned long flags; > + struct bio *bio; > + > + /* > + * Completion of this request needs to be handled with > + * blk_zone_write_complete_request(). > + */ > + req->rq_flags |= RQF_ZONE_WRITE_PLUGGING; > + > + if (blk_queue_nomerges(q)) > + return; > + > + /* > + * Walk through the list of plugged BIOs to check if they can be merged > + * into the back of the request. > + */ > + blk_zone_wplug_lock(zwplug, flags); > + while ((bio = bio_list_peek(&zwplug->bio_list))) { > + if (bio->bi_iter.bi_sector != req_back_sector || > + !blk_rq_merge_ok(req, bio)) > + break; > + > + WARN_ON_ONCE(bio_op(bio) != REQ_OP_WRITE_ZEROES && > + !bio->__bi_nr_segments); > + > + bio_list_pop(&zwplug->bio_list); > + if (bio_attempt_back_merge(req, bio, bio->__bi_nr_segments) != > + BIO_MERGE_OK) { > + bio_list_add_head(&zwplug->bio_list, bio); > + break; > + } > + > + /* > + * Drop the extra reference on the queue usage we got when > + * plugging the BIO. > + */ > + blk_queue_exit(q); > + > + req_back_sector += bio_sectors(bio); > + } > + blk_zone_wplug_unlock(zwplug, flags); > +} > + > +static bool blk_zone_wplug_handle_write(struct bio *bio, unsigned int nr_segs) > +{ > + struct blk_zone_wplug *zwplug; > + unsigned long flags; > + > + /* > + * BIOs must be fully contained within a zone so that we use the correct > + * zone write plug for the entire BIO. For blk-mq devices, the block > + * layer should already have done any splitting required to ensure this > + * and this BIO should thus not be straddling zone boundaries. For > + * BIO-based devices, it is the responsibility of the driver to split > + * the bio before submitting it. > + */ > + if (WARN_ON_ONCE(bio_straddle_zones(bio))) { > + bio_io_error(bio); > + return true; > + } > + > + zwplug = bio_lookup_zone_wplug(bio); > + if (!zwplug) > + return false; > + > + blk_zone_wplug_lock(zwplug, flags); > + > + /* Indicate that this BIO is being handled using zone write plugging. */ > + bio_set_flag(bio, BIO_ZONE_WRITE_PLUGGING); > + > + /* > + * If the zone is already plugged, add the BIO to the plug BIO list. > + * Otherwise, plug and let the BIO execute. > + */ > + if (zwplug->flags & BLK_ZONE_WPLUG_PLUGGED) { > + blk_zone_wplug_add_bio(zwplug, bio, nr_segs); > + blk_zone_wplug_unlock(zwplug, flags); > + return true; > + } > + > + zwplug->flags |= BLK_ZONE_WPLUG_PLUGGED; > + > + blk_zone_wplug_unlock(zwplug, flags); > + > + return false; > +} > + > +/** > + * blk_zone_write_plug_bio - Handle a zone write BIO with zone write plugging > + * @bio: The BIO being submitted > + * > + * Handle write and write zeroes operations using zone write plugging. > + * Return true whenever @bio execution needs to be delayed through the zone > + * write plug. Otherwise, return false to let the submission path process > + * @bio normally. > + */ > +bool blk_zone_write_plug_bio(struct bio *bio, unsigned int nr_segs) > +{ > + if (!bio->bi_bdev->bd_disk->zone_wplugs) > + return false; > + > + /* > + * If the BIO already has the plugging flag set, then it was already > + * handled through this path and this is a submission from the zone > + * plug bio submit work. > + */ > + if (bio_flagged(bio, BIO_ZONE_WRITE_PLUGGING)) > + return false; > + > + /* > + * We do not need to do anything special for empty flush BIOs, e.g > + * BIOs such as issued by blkdev_issue_flush(). The is because it is > + * the responsibility of the user to first wait for the completion of > + * write operations for flush to have any effect on the persistence of > + * the written data. > + */ > + if (op_is_flush(bio->bi_opf) && !bio_sectors(bio)) > + return false; > + > + /* > + * Regular writes and write zeroes need to be handled through the target > + * zone write plug. This includes writes with REQ_FUA | REQ_PREFLUSH > + * which may need to go through the flush machinery depending on the > + * target device capabilities. Plugging such writes is fine as the flush > + * machinery operates at the request level, below the plug, and > + * completion of the flush sequence will go through the regular BIO > + * completion, which will handle zone write plugging. > + */ > + switch (bio_op(bio)) { > + case REQ_OP_WRITE: > + case REQ_OP_WRITE_ZEROES: > + return blk_zone_wplug_handle_write(bio, nr_segs); > + default: > + return false; > + } > + > + return false; > +} > +EXPORT_SYMBOL_GPL(blk_zone_write_plug_bio); > + > +static void blk_zone_write_plug_unplug_bio(struct blk_zone_wplug *zwplug) > +{ > + unsigned long flags; > + > + blk_zone_wplug_lock(zwplug, flags); > + > + /* Schedule submission of the next plugged BIO if we have one. */ > + if (!bio_list_empty(&zwplug->bio_list)) > + kblockd_schedule_work(&zwplug->bio_work); > + else > + zwplug->flags &= ~BLK_ZONE_WPLUG_PLUGGED; > + > + blk_zone_wplug_unlock(zwplug, flags); > +} > + > +void blk_zone_write_plug_bio_endio(struct bio *bio) > +{ > + /* Make sure we do not see this BIO again by clearing the plug flag. */ > + bio_clear_flag(bio, BIO_ZONE_WRITE_PLUGGING); > + > + /* > + * For BIO-based devices, blk_zone_write_plug_complete_request() > + * is not called. So we need to schedule execution of the next > + * plugged BIO here. > + */ > + if (bio->bi_bdev->bd_has_submit_bio) { > + struct blk_zone_wplug *zwplug = bio_lookup_zone_wplug(bio); > + > + blk_zone_write_plug_unplug_bio(zwplug); > + } > +} > + > +void blk_zone_write_plug_complete_request(struct request *req) > +{ > + struct gendisk *disk = req->q->disk; > + struct blk_zone_wplug *zwplug = > + disk_lookup_zone_wplug(disk, req->__sector); > + > + req->rq_flags &= ~RQF_ZONE_WRITE_PLUGGING; > + > + blk_zone_write_plug_unplug_bio(zwplug); > +} > + > +static void blk_zone_wplug_bio_work(struct work_struct *work) > +{ > + struct blk_zone_wplug *zwplug = > + container_of(work, struct blk_zone_wplug, bio_work); > + unsigned long flags; > + struct bio *bio; > + > + /* > + * Unplug and submit the next plugged BIO. If we do not have any, clear > + * the plugged flag. > + */ > + blk_zone_wplug_lock(zwplug, flags); > + > + bio = bio_list_pop(&zwplug->bio_list); > + if (!bio) { > + zwplug->flags &= ~BLK_ZONE_WPLUG_PLUGGED; > + blk_zone_wplug_unlock(zwplug, flags); > + return; > + } > + > + blk_zone_wplug_unlock(zwplug, flags); > + > + /* > + * blk-mq devices will reuse the reference on the request queue usage > + * we took when the BIO was plugged, but the submission path for > + * BIO-based devices will not do that. So drop this reference here. > + */ > + if (bio->bi_bdev->bd_has_submit_bio) > + blk_queue_exit(bio->bi_bdev->bd_disk->queue); But I don't see where this reference is reused for blk-mq in this patch, care to point it out? Thanks, Ming
On 2/2/24 15:30, Damien Le Moal wrote: > Zone write plugging implements a per-zone "plug" for write operations to > tightly control the submission and execution order of writes to > sequential write required zones of a zoned block device. Per-zone > plugging of writes guarantees that at any time at most one write request > per zone is in flight. This mechanism is intended to replace zone write > locking which is controlled at the scheduler level and implemented only > by mq-deadline. > > Unlike zone write locking which operates on requests, zone write > plugging operates on BIOs. A zone write plug is simply a BIO list that > is atomically manipulated using a spinlock and a kblockd submission > work. A write BIO to a zone is "plugged" to delay its execution if a > write BIO for the same zone was already issued, that is, if a write > request for the same zone is being executed. The next plugged BIO is > unplugged and issued once the write request completes. > > This mechanism allows to: > - Untangles zone write ordering from block IO schedulers. This allows > removing the restriction on using only mq-deadline for zoned block > devices. Any block IO scheduler, including "none" can be used. > - Zone write plugging operates on BIOs instead of requests. Plugged > BIOs waiting for execution thus do not hold scheduling tags and thus > are not preventing other BIOs to proceed (reads or writes to other > zones). Depending on the workload, this can significantly improve > the device use and performance. > - Both blk-mq (request) based zoned devices and BIO-based devices (e.g. > device mapper) can use zone write plugging. It is mandatory for the > former but optional for the latter: BIO-based driver can use zone > write plugging to implement write ordering guarantees, or the drivers > can implement their own if needed. > - The code is less invasive in the block layer and is mostly limited to > blk-zoned.c with some small changes in blk-mq.c, blk-merge.c and > bio.c. > > Zone write plugging is implemented using struct blk_zone_wplug. This > structurei includes a spinlock, a BIO list and a work structure to > handle the submission of plugged BIOs. > > Plugging of zone write BIOs is done using the function > blk_zone_write_plug_bio() which returns false if a BIO execution does > not need to be delayed and true otherwise. This function is called > from blk_mq_submit_bio() after a BIO is split to avoid large BIOs > spanning multiple zones which would cause mishandling of zone write > plugging. This enables by default zone write plugging for any mq > request-based block device. BIO-based device drivers can also use zone > write plugging by expliclty calling blk_zone_write_plug_bio() in their > ->submit_bio method. For such devices, the driver must ensure that a > BIO passed to blk_zone_write_plug_bio() is already split and not > straddling zone boundaries. > > Only write and write zeroes BIOs are plugged. Zone write plugging does > not introduce any significant overhead for other operations. A BIO that > is being handled through zone write plugging is flagged using the new > BIO flag BIO_ZONE_WRITE_PLUGGING. A request handling a BIO flagged with > this new flag is flagged with the new RQF_ZONE_WRITE_PLUGGING flag. > The completion processing of BIOs and requests flagged trigger > respectively calls to the functions blk_zone_write_plug_bio_endio() and > blk_zone_write_plug_complete_request(). The latter function is used to > trigger submission of the next plugged BIO using the zone plug work. > blk_zone_write_plug_bio_endio() does the same for BIO-based devices. > This ensures that at any time, at most one request (blk-mq devices) or > one BIO (BIO-based devices) are being executed for any zone. The > handling of zone write plug using a per-zone plug spinlock maximizes > parrallelism and device usage by allowing multiple zones to be writen > simultaneously without lock contention. > > Zone write plugging ignores flush BIOs without data. Hovever, any flush > BIO that has data is always plugged so that the write part of the flush > sequence is serialized with other regular writes. > > Given that any BIO handled through zone write plugging will be the only > BIO in flight for the target zone when it is executed, the unplugging > and submission of a BIO will have no chance of successfully merging with > plugged requests or requests in the scheduler. To overcome this > potential performance loss, blk_mq_submit_bio() calls the function > blk_zone_write_plug_attempt_merge() to try to merge other plugged BIOs > with the one just unplugged. Successful merging is signaled using > blk_zone_write_plug_bio_merged(), called from bio_attempt_back_merge(). > Furthermore, to avoid recalculating the number of segments of plugged > BIOs to attempt merging, the number of segments of a plugged BIO is > saved using the new struct bio field __bi_nr_segments. To avoid growing > the size of struct bio, this field is added as a union with the > bio_cookie field. This is safe to do as polling is always disabled for > plugged BIOs. > > When BIOs are plugged in a zone write plug, the device request queue > usage counter is always incremented. This kept and reused when the > plugged BIO is unplugged and submitted again using > submit_bio_noacct_nocheck(). For this case, the unplugged BIO is already > flagged with BIO_ZONE_WRITE_PLUGGING and blk_mq_submit_bio() proceeds > directly to allocating a new request for the BIO, re-using the usage > reference count taken when the BIO was plugged. This extra reference > count is dropped in blk_zone_write_plug_attempt_merge() for any plugged > BIO that is successfully merged. Given that BIO-based devices will not > take this path, the extra reference is dropped when a plugged BIO is > unplugged and submitted. > > To match the new data structures used for zoned disks, the function > disk_free_zone_bitmaps() is renamed to the more generic > disk_free_zone_resources(). > > This commit contains contributions from Christoph Hellwig <hch@lst.de>. > > Signed-off-by: Damien Le Moal <dlemoal@kernel.org> > --- > block/bio.c | 7 + > block/blk-merge.c | 11 + > block/blk-mq.c | 28 +++ > block/blk-zoned.c | 408 +++++++++++++++++++++++++++++++++++++- > block/blk.h | 32 ++- > block/genhd.c | 2 +- > include/linux/blk-mq.h | 2 + > include/linux/blk_types.h | 8 +- > include/linux/blkdev.h | 8 + > 9 files changed, 496 insertions(+), 10 deletions(-) > > diff --git a/block/bio.c b/block/bio.c > index b9642a41f286..c8b0f7e8c713 100644 > --- a/block/bio.c > +++ b/block/bio.c > @@ -1581,6 +1581,13 @@ void bio_endio(struct bio *bio) > if (!bio_integrity_endio(bio)) > return; > > + /* > + * For BIOs handled through a zone write plugs, signal the end of the > + * BIO to the zone write plug to submit the next plugged BIO. > + */ > + if (bio_zone_write_plugging(bio)) > + blk_zone_write_plug_bio_endio(bio); > + > rq_qos_done_bio(bio); > > if (bio->bi_bdev && bio_flagged(bio, BIO_TRACE_COMPLETION)) { > diff --git a/block/blk-merge.c b/block/blk-merge.c > index a1ef61b03e31..2b5489cd9c65 100644 > --- a/block/blk-merge.c > +++ b/block/blk-merge.c > @@ -377,6 +377,7 @@ struct bio *__bio_split_to_limits(struct bio *bio, > blkcg_bio_issue_init(split); > bio_chain(split, bio); > trace_block_split(split, bio->bi_iter.bi_sector); > + WARN_ON_ONCE(bio_zone_write_plugging(bio)); > submit_bio_noacct(bio); > return split; > } > @@ -980,6 +981,9 @@ enum bio_merge_status bio_attempt_back_merge(struct request *req, > > blk_update_mixed_merge(req, bio, false); > > + if (req->rq_flags & RQF_ZONE_WRITE_PLUGGING) > + blk_zone_write_plug_bio_merged(bio); > + > req->biotail->bi_next = bio; > req->biotail = bio; > req->__data_len += bio->bi_iter.bi_size; > @@ -995,6 +999,13 @@ static enum bio_merge_status bio_attempt_front_merge(struct request *req, > { > const blk_opf_t ff = bio_failfast(bio); > > + /* > + * A front merge for zone writes can happen only if the user submitted > + * writes out of order. Do not attempt this to let the write fail. > + */ > + if (req->rq_flags & RQF_ZONE_WRITE_PLUGGING) > + return BIO_MERGE_FAILED; > + > if (!ll_front_merge_fn(req, bio, nr_segs)) > return BIO_MERGE_FAILED; > > diff --git a/block/blk-mq.c b/block/blk-mq.c > index f02e486a02ae..aa49bebf1199 100644 > --- a/block/blk-mq.c > +++ b/block/blk-mq.c > @@ -830,6 +830,9 @@ static void blk_complete_request(struct request *req) > bio = next; > } while (bio); > > + if (req->rq_flags & RQF_ZONE_WRITE_PLUGGING) > + blk_zone_write_plug_complete_request(req); > + > /* > * Reset counters so that the request stacking driver > * can find how many bytes remain in the request > @@ -943,6 +946,9 @@ bool blk_update_request(struct request *req, blk_status_t error, > * completely done > */ > if (!req->bio) { > + if (req->rq_flags & RQF_ZONE_WRITE_PLUGGING) > + blk_zone_write_plug_complete_request(req); > + > /* > * Reset counters so that the request stacking driver > * can find how many bytes remain in the request > @@ -2975,6 +2981,17 @@ void blk_mq_submit_bio(struct bio *bio) > struct request *rq; > blk_status_t ret; > > + /* > + * A BIO that was released form a zone write plug has already been > + * through the preparation in this function, already holds a reference > + * on the queue usage counter, and is the only write BIO in-flight for > + * the target zone. Go straight to allocating a request for it. > + */ > + if (bio_zone_write_plugging(bio)) { > + nr_segs = bio->__bi_nr_segments; > + goto new_request; > + } > + > bio = blk_queue_bounce(bio, q); > bio_set_ioprio(bio); > > @@ -3001,7 +3018,11 @@ void blk_mq_submit_bio(struct bio *bio) > if (blk_mq_attempt_bio_merge(q, bio, nr_segs)) > goto queue_exit; > > + if (blk_queue_is_zoned(q) && blk_zone_write_plug_bio(bio, nr_segs)) > + goto queue_exit; > + > if (!rq) { > +new_request: > rq = blk_mq_get_new_requests(q, plug, bio, nr_segs); > if (unlikely(!rq)) > goto queue_exit; > @@ -3017,8 +3038,12 @@ void blk_mq_submit_bio(struct bio *bio) > > ret = blk_crypto_rq_get_keyslot(rq); > if (ret != BLK_STS_OK) { > + bool zwplugging = bio_zone_write_plugging(bio); > + > bio->bi_status = ret; > bio_endio(bio); > + if (zwplugging) > + blk_zone_write_plug_complete_request(rq); > blk_mq_free_request(rq); > return; > } > @@ -3026,6 +3051,9 @@ void blk_mq_submit_bio(struct bio *bio) > if (op_is_flush(bio->bi_opf) && blk_insert_flush(rq)) > return; > > + if (bio_zone_write_plugging(bio)) > + blk_zone_write_plug_attempt_merge(rq); > + > if (plug) { > blk_add_rq_to_plug(plug, rq); > return; > diff --git a/block/blk-zoned.c b/block/blk-zoned.c > index d343e5756a9c..f6d4f511b664 100644 > --- a/block/blk-zoned.c > +++ b/block/blk-zoned.c > @@ -7,11 +7,11 @@ > * > * Copyright (c) 2016, Damien Le Moal > * Copyright (c) 2016, Western Digital > + * Copyright (c) 2024, Western Digital Corporation or its affiliates. > */ > > #include <linux/kernel.h> > #include <linux/module.h> > -#include <linux/rbtree.h> > #include <linux/blkdev.h> > #include <linux/blk-mq.h> > #include <linux/mm.h> > @@ -19,6 +19,7 @@ > #include <linux/sched/mm.h> > > #include "blk.h" > +#include "blk-mq-sched.h" > > #define ZONE_COND_NAME(name) [BLK_ZONE_COND_##name] = #name > static const char *const zone_cond_name[] = { > @@ -33,6 +34,27 @@ static const char *const zone_cond_name[] = { > }; > #undef ZONE_COND_NAME > > +/* > + * Per-zone write plug. > + */ > +struct blk_zone_wplug { > + spinlock_t lock; > + unsigned int flags; > + struct bio_list bio_list; > + struct work_struct bio_work; > +}; > + > +/* > + * Zone write plug flags bits: > + * - BLK_ZONE_WPLUG_CONV: Indicate that the zone is a conventional one. Writes > + * to these zones are never plugged. > + * - BLK_ZONE_WPLUG_PLUGGED: Indicate that the zone write plug is plugged, > + * that is, that write BIOs are being throttled due to a write BIO already > + * being executed or the zone write plug bio list is not empty. > + */ > +#define BLK_ZONE_WPLUG_CONV (1U << 0) > +#define BLK_ZONE_WPLUG_PLUGGED (1U << 1) > + > /** > * blk_zone_cond_str - Return string XXX in BLK_ZONE_COND_XXX. > * @zone_cond: BLK_ZONE_COND_XXX. > @@ -429,12 +451,374 @@ int blkdev_zone_mgmt_ioctl(struct block_device *bdev, blk_mode_t mode, > return ret; > } > > -void disk_free_zone_bitmaps(struct gendisk *disk) > +#define blk_zone_wplug_lock(zwplug, flags) \ > + spin_lock_irqsave(&zwplug->lock, flags) > + > +#define blk_zone_wplug_unlock(zwplug, flags) \ > + spin_unlock_irqrestore(&zwplug->lock, flags) > + > +static inline void blk_zone_wplug_bio_io_error(struct bio *bio) > +{ > + struct request_queue *q = bio->bi_bdev->bd_disk->queue; > + > + bio_clear_flag(bio, BIO_ZONE_WRITE_PLUGGING); > + bio_io_error(bio); > + blk_queue_exit(q); > +} > + > +static int blk_zone_wplug_abort(struct gendisk *disk, > + struct blk_zone_wplug *zwplug) > +{ > + struct bio *bio; > + int nr_aborted = 0; > + > + while ((bio = bio_list_pop(&zwplug->bio_list))) { > + blk_zone_wplug_bio_io_error(bio); > + nr_aborted++; > + } > + > + return nr_aborted; > +} > + > +/* > + * Return the zone write plug for sector in sequential write required zone. > + * Given that conventional zones have no write ordering constraints, NULL is > + * returned for sectors in conventional zones, to indicate that zone write > + * plugging is not needed. > + */ > +static inline struct blk_zone_wplug * > +disk_lookup_zone_wplug(struct gendisk *disk, sector_t sector) > +{ > + struct blk_zone_wplug *zwplug; > + > + if (WARN_ON_ONCE(!disk->zone_wplugs)) > + return NULL; > + > + zwplug = &disk->zone_wplugs[disk_zone_no(disk, sector)]; > + if (zwplug->flags & BLK_ZONE_WPLUG_CONV) > + return NULL; > + return zwplug; > +} > + > +static inline struct blk_zone_wplug *bio_lookup_zone_wplug(struct bio *bio) > +{ > + return disk_lookup_zone_wplug(bio->bi_bdev->bd_disk, > + bio->bi_iter.bi_sector); > +} > + > +static inline void blk_zone_wplug_add_bio(struct blk_zone_wplug *zwplug, > + struct bio *bio, unsigned int nr_segs) > +{ > + /* > + * Keep a reference on the BIO request queue usage. This reference will > + * be dropped either if the BIO is failed or after it is issued and > + * completes. > + */ > + percpu_ref_get(&bio->bi_bdev->bd_disk->queue->q_usage_counter); > + As discussed, wouldn't it be sufficient to increase the q_usage_counter only for the plug itself, and not the bios? The bios are already allocated for, and I think it we would to use a separate reference here as the 'plug' has a different lifetime than the bios which are added to the plug. > + /* > + * The BIO is being plugged and thus will have to wait for the on-going > + * write and for all other writes already plugged. So polling makes > + * no sense. > + */ > + bio_clear_polled(bio); > + > + /* > + * Reuse the poll cookie field to store the number of segments when > + * split to the hardware limits. > + */ > + bio->__bi_nr_segments = nr_segs; > + > + /* > + * We always receive BIOs after they are split and ready to be issued. > + * The block layer passes the parts of a split BIO in order, and the > + * user must also issue write sequentially. So simply add the new BIO > + * at the tail of the list to preserve the sequential write order. > + */ > + bio_list_add(&zwplug->bio_list, bio); > +} > + > +/* > + * Called from bio_attempt_back_merge() when a BIO was merged with a request. > + */ > +void blk_zone_write_plug_bio_merged(struct bio *bio) > +{ > + bio_set_flag(bio, BIO_ZONE_WRITE_PLUGGING); > +} > + > +/* > + * Attempt to merge plugged BIOs with a newly formed request of a BIO that went > + * through zone write plugging (either a new BIO or one that was unplugged). > + */ > +void blk_zone_write_plug_attempt_merge(struct request *req) > +{ > + struct blk_zone_wplug *zwplug = bio_lookup_zone_wplug(req->bio); > + sector_t req_back_sector = blk_rq_pos(req) + blk_rq_sectors(req); > + struct request_queue *q = req->q; > + unsigned long flags; > + struct bio *bio; > + > + /* > + * Completion of this request needs to be handled with > + * blk_zone_write_complete_request(). > + */ > + req->rq_flags |= RQF_ZONE_WRITE_PLUGGING; > + > + if (blk_queue_nomerges(q)) > + return; > + > + /* > + * Walk through the list of plugged BIOs to check if they can be merged > + * into the back of the request. > + */ > + blk_zone_wplug_lock(zwplug, flags); > + while ((bio = bio_list_peek(&zwplug->bio_list))) { > + if (bio->bi_iter.bi_sector != req_back_sector || > + !blk_rq_merge_ok(req, bio)) > + break; > + > + WARN_ON_ONCE(bio_op(bio) != REQ_OP_WRITE_ZEROES && > + !bio->__bi_nr_segments); > + > + bio_list_pop(&zwplug->bio_list); > + if (bio_attempt_back_merge(req, bio, bio->__bi_nr_segments) != > + BIO_MERGE_OK) { > + bio_list_add_head(&zwplug->bio_list, bio); > + break; > + } > + > + /* > + * Drop the extra reference on the queue usage we got when > + * plugging the BIO. > + */ > + blk_queue_exit(q); > + > + req_back_sector += bio_sectors(bio); > + } > + blk_zone_wplug_unlock(zwplug, flags); > +} And that's the other thing with which I'm slightly uncomfortable. We're replicating parts of the generic merging code here. It would be far better from a maintenance standpoint if we had only one place where we deal with bio merging. But I do see the challenge, so this is more of a reminder than something which needs to be fixed. Cheers, Hannes
On 2/4/24 12:56, Ming Lei wrote: > On Fri, Feb 02, 2024 at 04:30:44PM +0900, Damien Le Moal wrote: >> +/* >> + * Zone write plug flags bits: >> + * - BLK_ZONE_WPLUG_CONV: Indicate that the zone is a conventional one. Writes >> + * to these zones are never plugged. >> + * - BLK_ZONE_WPLUG_PLUGGED: Indicate that the zone write plug is plugged, >> + * that is, that write BIOs are being throttled due to a write BIO already >> + * being executed or the zone write plug bio list is not empty. >> + */ >> +#define BLK_ZONE_WPLUG_CONV (1U << 0) >> +#define BLK_ZONE_WPLUG_PLUGGED (1U << 1) > > BLK_ZONE_WPLUG_PLUGGED == !bio_list_empty(&zwplug->bio_list), so looks > this flag isn't necessary. No, it is. As the description says, the flag not only indicates that there are plugged BIOs, but it also indicates that there is a write for the zone in-flight. And that can happen even with the BIO list being empty. E.g. for a qd=1 workload of small BIOs, no BIO will ever be added to the BIO list, but the zone still must be marked as "plugged" when a write BIO is issued for it. >> +static inline void blk_zone_wplug_add_bio(struct blk_zone_wplug *zwplug, >> + struct bio *bio, unsigned int nr_segs) >> +{ >> + /* >> + * Keep a reference on the BIO request queue usage. This reference will >> + * be dropped either if the BIO is failed or after it is issued and >> + * completes. >> + */ >> + percpu_ref_get(&bio->bi_bdev->bd_disk->queue->q_usage_counter); > > It is fragile to get nested usage_counter, and same with grabbing/releasing it > from different contexts or even functions, and it could be much better to just > let block layer maintain it. > > From patch 23's change: > > + * Zoned block device information. Reads of this information must be > + * protected with blk_queue_enter() / blk_queue_exit(). Modifying this > > Anytime if there is in-flight bio, the block device is opened, so both gendisk and > request_queue are live, so not sure if this .q_usage_counter protection > is needed. Hannes also commented about this. Let me revisit this. >> + /* >> + * blk-mq devices will reuse the reference on the request queue usage >> + * we took when the BIO was plugged, but the submission path for >> + * BIO-based devices will not do that. So drop this reference here. >> + */ >> + if (bio->bi_bdev->bd_has_submit_bio) >> + blk_queue_exit(bio->bi_bdev->bd_disk->queue); > > But I don't see where this reference is reused for blk-mq in this patch, > care to point it out? This patch modifies blk_mq_submit_bio() to add a "goto new_request" at the top for any BIO flagged with BIO_FLAG_ZONE_WRITE_PLUGGING. So when a plugged BIO is unplugged and submitted again, the reference that was taken in blk_zone_wplug_add_bio() is reused for the new request for that BIO.
On Mon, Feb 05, 2024 at 08:57:00AM +0900, Damien Le Moal wrote: > On 2/4/24 12:56, Ming Lei wrote: > > On Fri, Feb 02, 2024 at 04:30:44PM +0900, Damien Le Moal wrote: > >> +/* > >> + * Zone write plug flags bits: > >> + * - BLK_ZONE_WPLUG_CONV: Indicate that the zone is a conventional one. Writes > >> + * to these zones are never plugged. > >> + * - BLK_ZONE_WPLUG_PLUGGED: Indicate that the zone write plug is plugged, > >> + * that is, that write BIOs are being throttled due to a write BIO already > >> + * being executed or the zone write plug bio list is not empty. > >> + */ > >> +#define BLK_ZONE_WPLUG_CONV (1U << 0) > >> +#define BLK_ZONE_WPLUG_PLUGGED (1U << 1) > > > > BLK_ZONE_WPLUG_PLUGGED == !bio_list_empty(&zwplug->bio_list), so looks > > this flag isn't necessary. > > No, it is. As the description says, the flag not only indicates that there are > plugged BIOs, but it also indicates that there is a write for the zone > in-flight. And that can happen even with the BIO list being empty. E.g. for a > qd=1 workload of small BIOs, no BIO will ever be added to the BIO list, but the > zone still must be marked as "plugged" when a write BIO is issued for it. OK. > > >> +static inline void blk_zone_wplug_add_bio(struct blk_zone_wplug *zwplug, > >> + struct bio *bio, unsigned int nr_segs) > >> +{ > >> + /* > >> + * Keep a reference on the BIO request queue usage. This reference will > >> + * be dropped either if the BIO is failed or after it is issued and > >> + * completes. > >> + */ > >> + percpu_ref_get(&bio->bi_bdev->bd_disk->queue->q_usage_counter); > > > > It is fragile to get nested usage_counter, and same with grabbing/releasing it > > from different contexts or even functions, and it could be much better to just > > let block layer maintain it. > > > > From patch 23's change: > > > > + * Zoned block device information. Reads of this information must be > > + * protected with blk_queue_enter() / blk_queue_exit(). Modifying this > > > > Anytime if there is in-flight bio, the block device is opened, so both gendisk and > > request_queue are live, so not sure if this .q_usage_counter protection > > is needed. > > Hannes also commented about this. Let me revisit this. I think only queue re-configuration(blk_revalidate_zone) requires the queue usage counter. Otherwise, bdev open()/close() should work just fine. > > >> + /* > >> + * blk-mq devices will reuse the reference on the request queue usage > >> + * we took when the BIO was plugged, but the submission path for > >> + * BIO-based devices will not do that. So drop this reference here. > >> + */ > >> + if (bio->bi_bdev->bd_has_submit_bio) > >> + blk_queue_exit(bio->bi_bdev->bd_disk->queue); > > > > But I don't see where this reference is reused for blk-mq in this patch, > > care to point it out? > > This patch modifies blk_mq_submit_bio() to add a "goto new_request" at the top > for any BIO flagged with BIO_FLAG_ZONE_WRITE_PLUGGING. So when a plugged BIO is > unplugged and submitted again, the reference that was taken in > blk_zone_wplug_add_bio() is reused for the new request for that BIO. OK, this reference reuse may be worse, because queue freeze can't prevent new write zoned bio from being submitted any more given only percpu_ref_get() is called for all write zoned bios. Thanks, Ming
On 2/5/24 11:19, Ming Lei wrote: >>>> +static inline void blk_zone_wplug_add_bio(struct blk_zone_wplug *zwplug, >>>> + struct bio *bio, unsigned int nr_segs) >>>> +{ >>>> + /* >>>> + * Keep a reference on the BIO request queue usage. This reference will >>>> + * be dropped either if the BIO is failed or after it is issued and >>>> + * completes. >>>> + */ >>>> + percpu_ref_get(&bio->bi_bdev->bd_disk->queue->q_usage_counter); >>> >>> It is fragile to get nested usage_counter, and same with grabbing/releasing it >>> from different contexts or even functions, and it could be much better to just >>> let block layer maintain it. >>> >>> From patch 23's change: >>> >>> + * Zoned block device information. Reads of this information must be >>> + * protected with blk_queue_enter() / blk_queue_exit(). Modifying this >>> >>> Anytime if there is in-flight bio, the block device is opened, so both gendisk and >>> request_queue are live, so not sure if this .q_usage_counter protection >>> is needed. >> >> Hannes also commented about this. Let me revisit this. > > I think only queue re-configuration(blk_revalidate_zone) requires the > queue usage counter. Otherwise, bdev open()/close() should work just > fine. I want to check FS case though. No clear if mounting FS that supports zone (btrfs) also uses bdev open ? >>>> + /* >>>> + * blk-mq devices will reuse the reference on the request queue usage >>>> + * we took when the BIO was plugged, but the submission path for >>>> + * BIO-based devices will not do that. So drop this reference here. >>>> + */ >>>> + if (bio->bi_bdev->bd_has_submit_bio) >>>> + blk_queue_exit(bio->bi_bdev->bd_disk->queue); >>> >>> But I don't see where this reference is reused for blk-mq in this patch, >>> care to point it out? >> >> This patch modifies blk_mq_submit_bio() to add a "goto new_request" at the top >> for any BIO flagged with BIO_FLAG_ZONE_WRITE_PLUGGING. So when a plugged BIO is >> unplugged and submitted again, the reference that was taken in >> blk_zone_wplug_add_bio() is reused for the new request for that BIO. > > OK, this reference reuse may be worse, because queue freeze can't prevent new > write zoned bio from being submitted any more given only percpu_ref_get() is > called for all write zoned bios. New BIOs (BIOS that have never been plugged yet) will go through the normal blk_queue_enter() in blk_mq_submit_bio(), so they will be stopped there if another context asked for a queue freeze. I do not think there is any issue with how things are currently done (we tested that *a lot* with many different drives and drive configs with DM etc). Reference counting as it is is OK, even though it most likely be simplified. I am looking at that now.
On Mon, Feb 05, 2024 at 11:41:04AM +0900, Damien Le Moal wrote: > On 2/5/24 11:19, Ming Lei wrote: > >>>> +static inline void blk_zone_wplug_add_bio(struct blk_zone_wplug *zwplug, > >>>> + struct bio *bio, unsigned int nr_segs) > >>>> +{ > >>>> + /* > >>>> + * Keep a reference on the BIO request queue usage. This reference will > >>>> + * be dropped either if the BIO is failed or after it is issued and > >>>> + * completes. > >>>> + */ > >>>> + percpu_ref_get(&bio->bi_bdev->bd_disk->queue->q_usage_counter); > >>> > >>> It is fragile to get nested usage_counter, and same with grabbing/releasing it > >>> from different contexts or even functions, and it could be much better to just > >>> let block layer maintain it. > >>> > >>> From patch 23's change: > >>> > >>> + * Zoned block device information. Reads of this information must be > >>> + * protected with blk_queue_enter() / blk_queue_exit(). Modifying this > >>> > >>> Anytime if there is in-flight bio, the block device is opened, so both gendisk and > >>> request_queue are live, so not sure if this .q_usage_counter protection > >>> is needed. > >> > >> Hannes also commented about this. Let me revisit this. > > > > I think only queue re-configuration(blk_revalidate_zone) requires the > > queue usage counter. Otherwise, bdev open()/close() should work just > > fine. > > I want to check FS case though. No clear if mounting FS that supports zone > (btrfs) also uses bdev open ? btrfs '-O zoned' shouldn't be one exception: mount -O zoned /dev/ublkb0 /mnt b'blkdev_get_whole' b'bdev_open_by_dev' b'bdev_open_by_path' b'btrfs_scan_one_device' b'btrfs_get_tree' b'vfs_get_tree' b'fc_mount' b'btrfs_get_tree' b'vfs_get_tree' b'path_mount' b'__x64_sys_mount' b'do_syscall_64' b'entry_SYSCALL_64_after_hwframe' b'[unknown]' 1 > > >>>> + /* > >>>> + * blk-mq devices will reuse the reference on the request queue usage > >>>> + * we took when the BIO was plugged, but the submission path for > >>>> + * BIO-based devices will not do that. So drop this reference here. > >>>> + */ > >>>> + if (bio->bi_bdev->bd_has_submit_bio) > >>>> + blk_queue_exit(bio->bi_bdev->bd_disk->queue); > >>> > >>> But I don't see where this reference is reused for blk-mq in this patch, > >>> care to point it out? > >> > >> This patch modifies blk_mq_submit_bio() to add a "goto new_request" at the top > >> for any BIO flagged with BIO_FLAG_ZONE_WRITE_PLUGGING. So when a plugged BIO is > >> unplugged and submitted again, the reference that was taken in > >> blk_zone_wplug_add_bio() is reused for the new request for that BIO. > > > > OK, this reference reuse may be worse, because queue freeze can't prevent new > > write zoned bio from being submitted any more given only percpu_ref_get() is > > called for all write zoned bios. > > New BIOs (BIOS that have never been plugged yet) will go through the normal > blk_queue_enter() in blk_mq_submit_bio(), so they will be stopped there if > another context asked for a queue freeze. I do not think there is any issue with > how things are currently done (we tested that *a lot* with many different drives > and drive configs with DM etc). Reference counting as it is is OK, even though > it most likely be simplified. I am looking at that now. Indeed, new zoned write bio is still covered by blk-mq's queue reference, and the trick is just played on old plugged bio. Thanks, Ming
On Mon, Feb 05, 2024 at 11:41:04AM +0900, Damien Le Moal wrote: > > I think only queue re-configuration(blk_revalidate_zone) requires the > > queue usage counter. Otherwise, bdev open()/close() should work just > > fine. > > I want to check FS case though. No clear if mounting FS that supports zone > (btrfs) also uses bdev open ? Every file system opens the block device. But we don't just need the block device to be open, but we also need the block limits to not change, and the only way to do that is to hold a q_usage_counter reference.
On 2/5/24 14:11, Christoph Hellwig wrote: > On Mon, Feb 05, 2024 at 11:41:04AM +0900, Damien Le Moal wrote: >>> I think only queue re-configuration(blk_revalidate_zone) requires the >>> queue usage counter. Otherwise, bdev open()/close() should work just >>> fine. >> >> I want to check FS case though. No clear if mounting FS that supports zone >> (btrfs) also uses bdev open ? > > Every file system opens the block device. But we don't just need the > block device to be open, but we also need the block limits to not > change, and the only way to do that is to hold a q_usage_counter > reference. OK. So I think that Hannes'idea to get/put the queue usage counter reference based on a zone BIO plug becoming not empty (get ref) and becoming empty (put ref) may be simpler then. And that would also work in the same way for blk-mq and BIO based drivers.
On Mon, Feb 05, 2024 at 02:37:41PM +0900, Damien Le Moal wrote: > OK. So I think that Hannes'idea to get/put the queue usage counter reference > based on a zone BIO plug becoming not empty (get ref) and becoming empty (put > ref) may be simpler then. And that would also work in the same way for blk-mq > and BIO based drivers. Maybe I'm missing something, but I'm not sure how that would even work. We need a q_usage_counter ref when doing all the submissions checks (limits, bounce, etc) early in blk_mq_sunmit_bio, and that one should be taken using the mormal bio_queue_enter patch to do the right thing on nowait submissions, when the queue is already frozen, etc. What is the benefit of not just keeping that references vs releasing it for all but the first bio just so that we need to grab another new reference at the actual submission time?
On 2/5/24 14:50, Christoph Hellwig wrote: > On Mon, Feb 05, 2024 at 02:37:41PM +0900, Damien Le Moal wrote: >> OK. So I think that Hannes'idea to get/put the queue usage counter reference >> based on a zone BIO plug becoming not empty (get ref) and becoming empty (put >> ref) may be simpler then. And that would also work in the same way for blk-mq >> and BIO based drivers. > > Maybe I'm missing something, but I'm not sure how that would even work. > > We need a q_usage_counter ref when doing all the submissions checks > (limits, bounce, etc) early in blk_mq_sunmit_bio, and that one should be > taken using the mormal bio_queue_enter patch to do the right thing on > nowait submissions, when the queue is already frozen, etc. What > is the benefit of not just keeping that references vs releasing it > for all but the first bio just so that we need to grab another > new reference at the actual submission time? I just tried to make the change, but the code does not become easier/cleaner at all. In fact, the contrary, it is very messy. I think I am going to keep things as is regarding the ref counting. There is one thing I need to check though: I re-ran the perf tests but this time took the average of 10 runs to mitigate differences due to variance between runs of the same test. And doing that, I do see a regression in performance for ZNS 4K qd=16 sequential writes (751 MB/s with rc2 vs 661 MB/s with ZWP). I need to check if that is due to never using the cached request for plugged BIOs in blk_mq_submit_bio(). If that is the case, we'll need to tweak the reference dropping there for the cached request case. I am also seeing a regression in performance with btrfs on SMR HDD (244 MB/s with rc2 and block/for-next vs 233 MB/s with ZWP). I need to check that as well.
On Mon, Feb 05, 2024 at 11:41:04AM +0900, Damien Le Moal wrote: > On 2/5/24 11:19, Ming Lei wrote: > >>>> +static inline void blk_zone_wplug_add_bio(struct blk_zone_wplug *zwplug, > >>>> + struct bio *bio, unsigned int nr_segs) > >>>> +{ > >>>> + /* > >>>> + * Keep a reference on the BIO request queue usage. This reference will > >>>> + * be dropped either if the BIO is failed or after it is issued and > >>>> + * completes. > >>>> + */ > >>>> + percpu_ref_get(&bio->bi_bdev->bd_disk->queue->q_usage_counter); > >>> > >>> It is fragile to get nested usage_counter, and same with grabbing/releasing it > >>> from different contexts or even functions, and it could be much better to just > >>> let block layer maintain it. > >>> > >>> From patch 23's change: > >>> > >>> + * Zoned block device information. Reads of this information must be > >>> + * protected with blk_queue_enter() / blk_queue_exit(). Modifying this > >>> > >>> Anytime if there is in-flight bio, the block device is opened, so both gendisk and > >>> request_queue are live, so not sure if this .q_usage_counter protection > >>> is needed. > >> > >> Hannes also commented about this. Let me revisit this. > > > > I think only queue re-configuration(blk_revalidate_zone) requires the > > queue usage counter. Otherwise, bdev open()/close() should work just > > fine. > > I want to check FS case though. No clear if mounting FS that supports zone > (btrfs) also uses bdev open ? I feel the following delta change might be cleaner and easily documented: - one IO takes single reference for both bio based and blk-mq, - no drop & re-grab - only grab extra reference for bio based - two code paths share same pattern diff --git a/block/blk-core.c b/block/blk-core.c index 9520ccab3050..118dd789beb5 100644 --- a/block/blk-core.c +++ b/block/blk-core.c @@ -597,6 +597,10 @@ static void __submit_bio(struct bio *bio) if (!bio->bi_bdev->bd_has_submit_bio) { blk_mq_submit_bio(bio); + } else if (bio_zone_write_plugging(bio)) { + struct gendisk *disk = bio->bi_bdev->bd_disk; + + disk->fops->submit_bio(bio); } else if (likely(bio_queue_enter(bio) == 0)) { struct gendisk *disk = bio->bi_bdev->bd_disk; diff --git a/block/blk-mq.c b/block/blk-mq.c index f0fc61a3ec81..fc6d792747dc 100644 --- a/block/blk-mq.c +++ b/block/blk-mq.c @@ -3006,8 +3006,12 @@ void blk_mq_submit_bio(struct bio *bio) if (blk_mq_attempt_bio_merge(q, bio, nr_segs)) goto queue_exit; + /* + * Grab one reference for plugged zoned write and it will be reused in + * next real submission + */ if (blk_queue_is_zoned(q) && blk_zone_write_plug_bio(bio, nr_segs)) - goto queue_exit; + return; if (!rq) { new_request: diff --git a/block/blk-zoned.c b/block/blk-zoned.c index f6d4f511b664..87abb3f7ef30 100644 --- a/block/blk-zoned.c +++ b/block/blk-zoned.c @@ -514,7 +514,8 @@ static inline void blk_zone_wplug_add_bio(struct blk_zone_wplug *zwplug, * be dropped either if the BIO is failed or after it is issued and * completes. */ - percpu_ref_get(&bio->bi_bdev->bd_disk->queue->q_usage_counter); + if (bio->bi_bdev->bd_has_submit_bio) + percpu_ref_get(&bio->bi_bdev->bd_disk->queue->q_usage_counter); /* * The BIO is being plugged and thus will have to wait for the on-going @@ -760,15 +761,10 @@ static void blk_zone_wplug_bio_work(struct work_struct *work) blk_zone_wplug_unlock(zwplug, flags); - /* - * blk-mq devices will reuse the reference on the request queue usage - * we took when the BIO was plugged, but the submission path for - * BIO-based devices will not do that. So drop this reference here. - */ + submit_bio_noacct_nocheck(bio); + if (bio->bi_bdev->bd_has_submit_bio) blk_queue_exit(bio->bi_bdev->bd_disk->queue); - - submit_bio_noacct_nocheck(bio); } static struct blk_zone_wplug *blk_zone_alloc_write_plugs(unsigned int nr_zones) Thanks, Ming
On 2/5/24 19:06, Ming Lei wrote: > On Mon, Feb 05, 2024 at 11:41:04AM +0900, Damien Le Moal wrote: >> On 2/5/24 11:19, Ming Lei wrote: >>>>>> +static inline void blk_zone_wplug_add_bio(struct blk_zone_wplug *zwplug, >>>>>> + struct bio *bio, unsigned int nr_segs) >>>>>> +{ >>>>>> + /* >>>>>> + * Keep a reference on the BIO request queue usage. This reference will >>>>>> + * be dropped either if the BIO is failed or after it is issued and >>>>>> + * completes. >>>>>> + */ >>>>>> + percpu_ref_get(&bio->bi_bdev->bd_disk->queue->q_usage_counter); >>>>> >>>>> It is fragile to get nested usage_counter, and same with grabbing/releasing it >>>>> from different contexts or even functions, and it could be much better to just >>>>> let block layer maintain it. >>>>> >>>>> From patch 23's change: >>>>> >>>>> + * Zoned block device information. Reads of this information must be >>>>> + * protected with blk_queue_enter() / blk_queue_exit(). Modifying this >>>>> >>>>> Anytime if there is in-flight bio, the block device is opened, so both gendisk and >>>>> request_queue are live, so not sure if this .q_usage_counter protection >>>>> is needed. >>>> >>>> Hannes also commented about this. Let me revisit this. >>> >>> I think only queue re-configuration(blk_revalidate_zone) requires the >>> queue usage counter. Otherwise, bdev open()/close() should work just >>> fine. >> >> I want to check FS case though. No clear if mounting FS that supports zone >> (btrfs) also uses bdev open ? > > I feel the following delta change might be cleaner and easily documented: > > - one IO takes single reference for both bio based and blk-mq, > - no drop & re-grab > - only grab extra reference for bio based > - two code paths share same pattern > > diff --git a/block/blk-core.c b/block/blk-core.c > index 9520ccab3050..118dd789beb5 100644 > --- a/block/blk-core.c > +++ b/block/blk-core.c > @@ -597,6 +597,10 @@ static void __submit_bio(struct bio *bio) > > if (!bio->bi_bdev->bd_has_submit_bio) { > blk_mq_submit_bio(bio); > + } else if (bio_zone_write_plugging(bio)) { > + struct gendisk *disk = bio->bi_bdev->bd_disk; > + > + disk->fops->submit_bio(bio); > } else if (likely(bio_queue_enter(bio) == 0)) { > struct gendisk *disk = bio->bi_bdev->bd_disk; > > diff --git a/block/blk-mq.c b/block/blk-mq.c > index f0fc61a3ec81..fc6d792747dc 100644 > --- a/block/blk-mq.c > +++ b/block/blk-mq.c > @@ -3006,8 +3006,12 @@ void blk_mq_submit_bio(struct bio *bio) > if (blk_mq_attempt_bio_merge(q, bio, nr_segs)) > goto queue_exit; > > + /* > + * Grab one reference for plugged zoned write and it will be reused in > + * next real submission > + */ > if (blk_queue_is_zoned(q) && blk_zone_write_plug_bio(bio, nr_segs)) > - goto queue_exit; > + return; > > if (!rq) { > new_request: > diff --git a/block/blk-zoned.c b/block/blk-zoned.c > index f6d4f511b664..87abb3f7ef30 100644 > --- a/block/blk-zoned.c > +++ b/block/blk-zoned.c > @@ -514,7 +514,8 @@ static inline void blk_zone_wplug_add_bio(struct blk_zone_wplug *zwplug, > * be dropped either if the BIO is failed or after it is issued and > * completes. > */ > - percpu_ref_get(&bio->bi_bdev->bd_disk->queue->q_usage_counter); > + if (bio->bi_bdev->bd_has_submit_bio) > + percpu_ref_get(&bio->bi_bdev->bd_disk->queue->q_usage_counter); > > /* > * The BIO is being plugged and thus will have to wait for the on-going > @@ -760,15 +761,10 @@ static void blk_zone_wplug_bio_work(struct work_struct *work) > > blk_zone_wplug_unlock(zwplug, flags); > > - /* > - * blk-mq devices will reuse the reference on the request queue usage > - * we took when the BIO was plugged, but the submission path for > - * BIO-based devices will not do that. So drop this reference here. > - */ > + submit_bio_noacct_nocheck(bio); > + > if (bio->bi_bdev->bd_has_submit_bio) > blk_queue_exit(bio->bi_bdev->bd_disk->queue); Hmm... As-is, this is a potential use-after-free of the bio. But I get the idea. This is indeed a little better. I will integrate this. > - > - submit_bio_noacct_nocheck(bio); > } > > static struct blk_zone_wplug *blk_zone_alloc_write_plugs(unsigned int nr_zones) > > Thanks, > Ming > >
On 2/5/24 21:20, Damien Le Moal wrote: > On 2/5/24 19:06, Ming Lei wrote: >> On Mon, Feb 05, 2024 at 11:41:04AM +0900, Damien Le Moal wrote: >>> On 2/5/24 11:19, Ming Lei wrote: >>>>>>> +static inline void blk_zone_wplug_add_bio(struct blk_zone_wplug *zwplug, >>>>>>> + struct bio *bio, unsigned int nr_segs) >>>>>>> +{ >>>>>>> + /* >>>>>>> + * Keep a reference on the BIO request queue usage. This reference will >>>>>>> + * be dropped either if the BIO is failed or after it is issued and >>>>>>> + * completes. >>>>>>> + */ >>>>>>> + percpu_ref_get(&bio->bi_bdev->bd_disk->queue->q_usage_counter); >>>>>> >>>>>> It is fragile to get nested usage_counter, and same with grabbing/releasing it >>>>>> from different contexts or even functions, and it could be much better to just >>>>>> let block layer maintain it. >>>>>> >>>>>> From patch 23's change: >>>>>> >>>>>> + * Zoned block device information. Reads of this information must be >>>>>> + * protected with blk_queue_enter() / blk_queue_exit(). Modifying this >>>>>> >>>>>> Anytime if there is in-flight bio, the block device is opened, so both gendisk and >>>>>> request_queue are live, so not sure if this .q_usage_counter protection >>>>>> is needed. >>>>> >>>>> Hannes also commented about this. Let me revisit this. >>>> >>>> I think only queue re-configuration(blk_revalidate_zone) requires the >>>> queue usage counter. Otherwise, bdev open()/close() should work just >>>> fine. >>> >>> I want to check FS case though. No clear if mounting FS that supports zone >>> (btrfs) also uses bdev open ? >> >> I feel the following delta change might be cleaner and easily documented: >> >> - one IO takes single reference for both bio based and blk-mq, >> - no drop & re-grab >> - only grab extra reference for bio based >> - two code paths share same pattern >> >> diff --git a/block/blk-core.c b/block/blk-core.c >> index 9520ccab3050..118dd789beb5 100644 >> --- a/block/blk-core.c >> +++ b/block/blk-core.c >> @@ -597,6 +597,10 @@ static void __submit_bio(struct bio *bio) >> >> if (!bio->bi_bdev->bd_has_submit_bio) { >> blk_mq_submit_bio(bio); >> + } else if (bio_zone_write_plugging(bio)) { >> + struct gendisk *disk = bio->bi_bdev->bd_disk; >> + >> + disk->fops->submit_bio(bio); Actually, no, that is not correct. This would not stop BIO submission if blk_queue_freeze() was called by another context. So we cannot do that here without calling blk_queue_enter()... >> } else if (likely(bio_queue_enter(bio) == 0)) { >> struct gendisk *disk = bio->bi_bdev->bd_disk; >> >> diff --git a/block/blk-mq.c b/block/blk-mq.c >> index f0fc61a3ec81..fc6d792747dc 100644 >> --- a/block/blk-mq.c >> +++ b/block/blk-mq.c >> @@ -3006,8 +3006,12 @@ void blk_mq_submit_bio(struct bio *bio) >> if (blk_mq_attempt_bio_merge(q, bio, nr_segs)) >> goto queue_exit; >> >> + /* >> + * Grab one reference for plugged zoned write and it will be reused in >> + * next real submission >> + */ >> if (blk_queue_is_zoned(q) && blk_zone_write_plug_bio(bio, nr_segs)) >> - goto queue_exit; >> + return; ...and this one is not correct because of the cached request: if there was a cached request, blk_mq_submit_bio() did not call blk_queue_enter() because the cached request already had a reference. But we cannot reuse that reference as the next BIO may be a read or a write to a zone that is not plugged, and these would use the cached request and so need the usage counter reference. So we would still need to grab an extra reference in such case. So in the end, it feels a lot simpler to keep the reference counting as it was as it makes things a lot less messier in blk_mq_submit_bio(). I will though try to improve the comments to make it clear how this is working. >> >> if (!rq) { >> new_request: >> diff --git a/block/blk-zoned.c b/block/blk-zoned.c >> index f6d4f511b664..87abb3f7ef30 100644 >> --- a/block/blk-zoned.c >> +++ b/block/blk-zoned.c >> @@ -514,7 +514,8 @@ static inline void blk_zone_wplug_add_bio(struct blk_zone_wplug *zwplug, >> * be dropped either if the BIO is failed or after it is issued and >> * completes. >> */ >> - percpu_ref_get(&bio->bi_bdev->bd_disk->queue->q_usage_counter); >> + if (bio->bi_bdev->bd_has_submit_bio) >> + percpu_ref_get(&bio->bi_bdev->bd_disk->queue->q_usage_counter); >> >> /* >> * The BIO is being plugged and thus will have to wait for the on-going >> @@ -760,15 +761,10 @@ static void blk_zone_wplug_bio_work(struct work_struct *work) >> >> blk_zone_wplug_unlock(zwplug, flags); >> >> - /* >> - * blk-mq devices will reuse the reference on the request queue usage >> - * we took when the BIO was plugged, but the submission path for >> - * BIO-based devices will not do that. So drop this reference here. >> - */ >> + submit_bio_noacct_nocheck(bio); >> + >> if (bio->bi_bdev->bd_has_submit_bio) >> blk_queue_exit(bio->bi_bdev->bd_disk->queue); > > Hmm... As-is, this is a potential use-after-free of the bio. But I get the idea. > This is indeed a little better. I will integrate this. > >> - >> - submit_bio_noacct_nocheck(bio); >> } >> >> static struct blk_zone_wplug *blk_zone_alloc_write_plugs(unsigned int nr_zones) >> >> Thanks, >> Ming >> >> >
On 2/1/24 23:30, Damien Le Moal wrote: > The next plugged BIO is unplugged and issued once the write request completes. So this patch series is orthogonal to my patch series that implements zoned write pipelining? > This mechanism allows to: > - Untangles zone write ordering from block IO schedulers. This allows Untangles -> Untangle > Zone write plugging is implemented using struct blk_zone_wplug. This > structurei includes a spinlock, a BIO list and a work structure to structurei -> structure > This ensures that at any time, at most one request (blk-mq devices) or > one BIO (BIO-based devices) are being executed for any zone. The > handling of zone write plug using a per-zone plug spinlock maximizes > parrallelism and device usage by allowing multiple zones to be writen parrallelism -> parallelism > simultaneously without lock contention. This is not correct. Device usage is not maximized since zone write bios are serialized. Pipelining zoned writes results in higher device utilization. > + /* > + * For BIOs handled through a zone write plugs, signal the end of the plugs -> plug > +#define blk_zone_wplug_lock(zwplug, flags) \ > + spin_lock_irqsave(&zwplug->lock, flags) > + > +#define blk_zone_wplug_unlock(zwplug, flags) \ > + spin_unlock_irqrestore(&zwplug->lock, flags) Hmm ... these macros may make code harder to read rather than improve readability of the code. Thanks, Bart.
On 2/6/24 02:48, Bart Van Assche wrote: > On 2/1/24 23:30, Damien Le Moal wrote: >> The next plugged BIO is unplugged and issued once the write request completes. > > So this patch series is orthogonal to my patch series that implements zoned > write pipelining? I don't know. >> simultaneously without lock contention. > > This is not correct. Device usage is not maximized since zone write bios > are serialized. Pipelining zoned writes results in higher device > utilization. I meant to say that the locking scheme does not get in the way of maximizing device utilization/parallelism. If you are only writing to a single zone, then sure, it does not matter since the drive will be used at qd=1 for that case. >> +#define blk_zone_wplug_lock(zwplug, flags) \ >> + spin_lock_irqsave(&zwplug->lock, flags) >> + >> +#define blk_zone_wplug_unlock(zwplug, flags) \ >> + spin_unlock_irqrestore(&zwplug->lock, flags) > > Hmm ... these macros may make code harder to read rather than improve > readability of the code. I do not see how they make the code less readable. The macro names are not clear enough ?
On 2/5/24 15:48, Damien Le Moal wrote: > On 2/6/24 02:48, Bart Van Assche wrote: >> On 2/1/24 23:30, Damien Le Moal wrote: >>> +#define blk_zone_wplug_lock(zwplug, flags) \ >>> + spin_lock_irqsave(&zwplug->lock, flags) >>> + >>> +#define blk_zone_wplug_unlock(zwplug, flags) \ >>> + spin_unlock_irqrestore(&zwplug->lock, flags) >> >> Hmm ... these macros may make code harder to read rather than improve >> readability of the code. > > I do not see how they make the code less readable. The macro names are not clear > enough ? The macro names are clear but the macro names are almost as long as the macro definitions. Usually this is a sign that it's better not to introduce macros and instead inline the macros. Thanks, Bart.
diff --git a/block/bio.c b/block/bio.c index b9642a41f286..c8b0f7e8c713 100644 --- a/block/bio.c +++ b/block/bio.c @@ -1581,6 +1581,13 @@ void bio_endio(struct bio *bio) if (!bio_integrity_endio(bio)) return; + /* + * For BIOs handled through a zone write plugs, signal the end of the + * BIO to the zone write plug to submit the next plugged BIO. + */ + if (bio_zone_write_plugging(bio)) + blk_zone_write_plug_bio_endio(bio); + rq_qos_done_bio(bio); if (bio->bi_bdev && bio_flagged(bio, BIO_TRACE_COMPLETION)) { diff --git a/block/blk-merge.c b/block/blk-merge.c index a1ef61b03e31..2b5489cd9c65 100644 --- a/block/blk-merge.c +++ b/block/blk-merge.c @@ -377,6 +377,7 @@ struct bio *__bio_split_to_limits(struct bio *bio, blkcg_bio_issue_init(split); bio_chain(split, bio); trace_block_split(split, bio->bi_iter.bi_sector); + WARN_ON_ONCE(bio_zone_write_plugging(bio)); submit_bio_noacct(bio); return split; } @@ -980,6 +981,9 @@ enum bio_merge_status bio_attempt_back_merge(struct request *req, blk_update_mixed_merge(req, bio, false); + if (req->rq_flags & RQF_ZONE_WRITE_PLUGGING) + blk_zone_write_plug_bio_merged(bio); + req->biotail->bi_next = bio; req->biotail = bio; req->__data_len += bio->bi_iter.bi_size; @@ -995,6 +999,13 @@ static enum bio_merge_status bio_attempt_front_merge(struct request *req, { const blk_opf_t ff = bio_failfast(bio); + /* + * A front merge for zone writes can happen only if the user submitted + * writes out of order. Do not attempt this to let the write fail. + */ + if (req->rq_flags & RQF_ZONE_WRITE_PLUGGING) + return BIO_MERGE_FAILED; + if (!ll_front_merge_fn(req, bio, nr_segs)) return BIO_MERGE_FAILED; diff --git a/block/blk-mq.c b/block/blk-mq.c index f02e486a02ae..aa49bebf1199 100644 --- a/block/blk-mq.c +++ b/block/blk-mq.c @@ -830,6 +830,9 @@ static void blk_complete_request(struct request *req) bio = next; } while (bio); + if (req->rq_flags & RQF_ZONE_WRITE_PLUGGING) + blk_zone_write_plug_complete_request(req); + /* * Reset counters so that the request stacking driver * can find how many bytes remain in the request @@ -943,6 +946,9 @@ bool blk_update_request(struct request *req, blk_status_t error, * completely done */ if (!req->bio) { + if (req->rq_flags & RQF_ZONE_WRITE_PLUGGING) + blk_zone_write_plug_complete_request(req); + /* * Reset counters so that the request stacking driver * can find how many bytes remain in the request @@ -2975,6 +2981,17 @@ void blk_mq_submit_bio(struct bio *bio) struct request *rq; blk_status_t ret; + /* + * A BIO that was released form a zone write plug has already been + * through the preparation in this function, already holds a reference + * on the queue usage counter, and is the only write BIO in-flight for + * the target zone. Go straight to allocating a request for it. + */ + if (bio_zone_write_plugging(bio)) { + nr_segs = bio->__bi_nr_segments; + goto new_request; + } + bio = blk_queue_bounce(bio, q); bio_set_ioprio(bio); @@ -3001,7 +3018,11 @@ void blk_mq_submit_bio(struct bio *bio) if (blk_mq_attempt_bio_merge(q, bio, nr_segs)) goto queue_exit; + if (blk_queue_is_zoned(q) && blk_zone_write_plug_bio(bio, nr_segs)) + goto queue_exit; + if (!rq) { +new_request: rq = blk_mq_get_new_requests(q, plug, bio, nr_segs); if (unlikely(!rq)) goto queue_exit; @@ -3017,8 +3038,12 @@ void blk_mq_submit_bio(struct bio *bio) ret = blk_crypto_rq_get_keyslot(rq); if (ret != BLK_STS_OK) { + bool zwplugging = bio_zone_write_plugging(bio); + bio->bi_status = ret; bio_endio(bio); + if (zwplugging) + blk_zone_write_plug_complete_request(rq); blk_mq_free_request(rq); return; } @@ -3026,6 +3051,9 @@ void blk_mq_submit_bio(struct bio *bio) if (op_is_flush(bio->bi_opf) && blk_insert_flush(rq)) return; + if (bio_zone_write_plugging(bio)) + blk_zone_write_plug_attempt_merge(rq); + if (plug) { blk_add_rq_to_plug(plug, rq); return; diff --git a/block/blk-zoned.c b/block/blk-zoned.c index d343e5756a9c..f6d4f511b664 100644 --- a/block/blk-zoned.c +++ b/block/blk-zoned.c @@ -7,11 +7,11 @@ * * Copyright (c) 2016, Damien Le Moal * Copyright (c) 2016, Western Digital + * Copyright (c) 2024, Western Digital Corporation or its affiliates. */ #include <linux/kernel.h> #include <linux/module.h> -#include <linux/rbtree.h> #include <linux/blkdev.h> #include <linux/blk-mq.h> #include <linux/mm.h> @@ -19,6 +19,7 @@ #include <linux/sched/mm.h> #include "blk.h" +#include "blk-mq-sched.h" #define ZONE_COND_NAME(name) [BLK_ZONE_COND_##name] = #name static const char *const zone_cond_name[] = { @@ -33,6 +34,27 @@ static const char *const zone_cond_name[] = { }; #undef ZONE_COND_NAME +/* + * Per-zone write plug. + */ +struct blk_zone_wplug { + spinlock_t lock; + unsigned int flags; + struct bio_list bio_list; + struct work_struct bio_work; +}; + +/* + * Zone write plug flags bits: + * - BLK_ZONE_WPLUG_CONV: Indicate that the zone is a conventional one. Writes + * to these zones are never plugged. + * - BLK_ZONE_WPLUG_PLUGGED: Indicate that the zone write plug is plugged, + * that is, that write BIOs are being throttled due to a write BIO already + * being executed or the zone write plug bio list is not empty. + */ +#define BLK_ZONE_WPLUG_CONV (1U << 0) +#define BLK_ZONE_WPLUG_PLUGGED (1U << 1) + /** * blk_zone_cond_str - Return string XXX in BLK_ZONE_COND_XXX. * @zone_cond: BLK_ZONE_COND_XXX. @@ -429,12 +451,374 @@ int blkdev_zone_mgmt_ioctl(struct block_device *bdev, blk_mode_t mode, return ret; } -void disk_free_zone_bitmaps(struct gendisk *disk) +#define blk_zone_wplug_lock(zwplug, flags) \ + spin_lock_irqsave(&zwplug->lock, flags) + +#define blk_zone_wplug_unlock(zwplug, flags) \ + spin_unlock_irqrestore(&zwplug->lock, flags) + +static inline void blk_zone_wplug_bio_io_error(struct bio *bio) +{ + struct request_queue *q = bio->bi_bdev->bd_disk->queue; + + bio_clear_flag(bio, BIO_ZONE_WRITE_PLUGGING); + bio_io_error(bio); + blk_queue_exit(q); +} + +static int blk_zone_wplug_abort(struct gendisk *disk, + struct blk_zone_wplug *zwplug) +{ + struct bio *bio; + int nr_aborted = 0; + + while ((bio = bio_list_pop(&zwplug->bio_list))) { + blk_zone_wplug_bio_io_error(bio); + nr_aborted++; + } + + return nr_aborted; +} + +/* + * Return the zone write plug for sector in sequential write required zone. + * Given that conventional zones have no write ordering constraints, NULL is + * returned for sectors in conventional zones, to indicate that zone write + * plugging is not needed. + */ +static inline struct blk_zone_wplug * +disk_lookup_zone_wplug(struct gendisk *disk, sector_t sector) +{ + struct blk_zone_wplug *zwplug; + + if (WARN_ON_ONCE(!disk->zone_wplugs)) + return NULL; + + zwplug = &disk->zone_wplugs[disk_zone_no(disk, sector)]; + if (zwplug->flags & BLK_ZONE_WPLUG_CONV) + return NULL; + return zwplug; +} + +static inline struct blk_zone_wplug *bio_lookup_zone_wplug(struct bio *bio) +{ + return disk_lookup_zone_wplug(bio->bi_bdev->bd_disk, + bio->bi_iter.bi_sector); +} + +static inline void blk_zone_wplug_add_bio(struct blk_zone_wplug *zwplug, + struct bio *bio, unsigned int nr_segs) +{ + /* + * Keep a reference on the BIO request queue usage. This reference will + * be dropped either if the BIO is failed or after it is issued and + * completes. + */ + percpu_ref_get(&bio->bi_bdev->bd_disk->queue->q_usage_counter); + + /* + * The BIO is being plugged and thus will have to wait for the on-going + * write and for all other writes already plugged. So polling makes + * no sense. + */ + bio_clear_polled(bio); + + /* + * Reuse the poll cookie field to store the number of segments when + * split to the hardware limits. + */ + bio->__bi_nr_segments = nr_segs; + + /* + * We always receive BIOs after they are split and ready to be issued. + * The block layer passes the parts of a split BIO in order, and the + * user must also issue write sequentially. So simply add the new BIO + * at the tail of the list to preserve the sequential write order. + */ + bio_list_add(&zwplug->bio_list, bio); +} + +/* + * Called from bio_attempt_back_merge() when a BIO was merged with a request. + */ +void blk_zone_write_plug_bio_merged(struct bio *bio) +{ + bio_set_flag(bio, BIO_ZONE_WRITE_PLUGGING); +} + +/* + * Attempt to merge plugged BIOs with a newly formed request of a BIO that went + * through zone write plugging (either a new BIO or one that was unplugged). + */ +void blk_zone_write_plug_attempt_merge(struct request *req) +{ + struct blk_zone_wplug *zwplug = bio_lookup_zone_wplug(req->bio); + sector_t req_back_sector = blk_rq_pos(req) + blk_rq_sectors(req); + struct request_queue *q = req->q; + unsigned long flags; + struct bio *bio; + + /* + * Completion of this request needs to be handled with + * blk_zone_write_complete_request(). + */ + req->rq_flags |= RQF_ZONE_WRITE_PLUGGING; + + if (blk_queue_nomerges(q)) + return; + + /* + * Walk through the list of plugged BIOs to check if they can be merged + * into the back of the request. + */ + blk_zone_wplug_lock(zwplug, flags); + while ((bio = bio_list_peek(&zwplug->bio_list))) { + if (bio->bi_iter.bi_sector != req_back_sector || + !blk_rq_merge_ok(req, bio)) + break; + + WARN_ON_ONCE(bio_op(bio) != REQ_OP_WRITE_ZEROES && + !bio->__bi_nr_segments); + + bio_list_pop(&zwplug->bio_list); + if (bio_attempt_back_merge(req, bio, bio->__bi_nr_segments) != + BIO_MERGE_OK) { + bio_list_add_head(&zwplug->bio_list, bio); + break; + } + + /* + * Drop the extra reference on the queue usage we got when + * plugging the BIO. + */ + blk_queue_exit(q); + + req_back_sector += bio_sectors(bio); + } + blk_zone_wplug_unlock(zwplug, flags); +} + +static bool blk_zone_wplug_handle_write(struct bio *bio, unsigned int nr_segs) +{ + struct blk_zone_wplug *zwplug; + unsigned long flags; + + /* + * BIOs must be fully contained within a zone so that we use the correct + * zone write plug for the entire BIO. For blk-mq devices, the block + * layer should already have done any splitting required to ensure this + * and this BIO should thus not be straddling zone boundaries. For + * BIO-based devices, it is the responsibility of the driver to split + * the bio before submitting it. + */ + if (WARN_ON_ONCE(bio_straddle_zones(bio))) { + bio_io_error(bio); + return true; + } + + zwplug = bio_lookup_zone_wplug(bio); + if (!zwplug) + return false; + + blk_zone_wplug_lock(zwplug, flags); + + /* Indicate that this BIO is being handled using zone write plugging. */ + bio_set_flag(bio, BIO_ZONE_WRITE_PLUGGING); + + /* + * If the zone is already plugged, add the BIO to the plug BIO list. + * Otherwise, plug and let the BIO execute. + */ + if (zwplug->flags & BLK_ZONE_WPLUG_PLUGGED) { + blk_zone_wplug_add_bio(zwplug, bio, nr_segs); + blk_zone_wplug_unlock(zwplug, flags); + return true; + } + + zwplug->flags |= BLK_ZONE_WPLUG_PLUGGED; + + blk_zone_wplug_unlock(zwplug, flags); + + return false; +} + +/** + * blk_zone_write_plug_bio - Handle a zone write BIO with zone write plugging + * @bio: The BIO being submitted + * + * Handle write and write zeroes operations using zone write plugging. + * Return true whenever @bio execution needs to be delayed through the zone + * write plug. Otherwise, return false to let the submission path process + * @bio normally. + */ +bool blk_zone_write_plug_bio(struct bio *bio, unsigned int nr_segs) +{ + if (!bio->bi_bdev->bd_disk->zone_wplugs) + return false; + + /* + * If the BIO already has the plugging flag set, then it was already + * handled through this path and this is a submission from the zone + * plug bio submit work. + */ + if (bio_flagged(bio, BIO_ZONE_WRITE_PLUGGING)) + return false; + + /* + * We do not need to do anything special for empty flush BIOs, e.g + * BIOs such as issued by blkdev_issue_flush(). The is because it is + * the responsibility of the user to first wait for the completion of + * write operations for flush to have any effect on the persistence of + * the written data. + */ + if (op_is_flush(bio->bi_opf) && !bio_sectors(bio)) + return false; + + /* + * Regular writes and write zeroes need to be handled through the target + * zone write plug. This includes writes with REQ_FUA | REQ_PREFLUSH + * which may need to go through the flush machinery depending on the + * target device capabilities. Plugging such writes is fine as the flush + * machinery operates at the request level, below the plug, and + * completion of the flush sequence will go through the regular BIO + * completion, which will handle zone write plugging. + */ + switch (bio_op(bio)) { + case REQ_OP_WRITE: + case REQ_OP_WRITE_ZEROES: + return blk_zone_wplug_handle_write(bio, nr_segs); + default: + return false; + } + + return false; +} +EXPORT_SYMBOL_GPL(blk_zone_write_plug_bio); + +static void blk_zone_write_plug_unplug_bio(struct blk_zone_wplug *zwplug) +{ + unsigned long flags; + + blk_zone_wplug_lock(zwplug, flags); + + /* Schedule submission of the next plugged BIO if we have one. */ + if (!bio_list_empty(&zwplug->bio_list)) + kblockd_schedule_work(&zwplug->bio_work); + else + zwplug->flags &= ~BLK_ZONE_WPLUG_PLUGGED; + + blk_zone_wplug_unlock(zwplug, flags); +} + +void blk_zone_write_plug_bio_endio(struct bio *bio) +{ + /* Make sure we do not see this BIO again by clearing the plug flag. */ + bio_clear_flag(bio, BIO_ZONE_WRITE_PLUGGING); + + /* + * For BIO-based devices, blk_zone_write_plug_complete_request() + * is not called. So we need to schedule execution of the next + * plugged BIO here. + */ + if (bio->bi_bdev->bd_has_submit_bio) { + struct blk_zone_wplug *zwplug = bio_lookup_zone_wplug(bio); + + blk_zone_write_plug_unplug_bio(zwplug); + } +} + +void blk_zone_write_plug_complete_request(struct request *req) +{ + struct gendisk *disk = req->q->disk; + struct blk_zone_wplug *zwplug = + disk_lookup_zone_wplug(disk, req->__sector); + + req->rq_flags &= ~RQF_ZONE_WRITE_PLUGGING; + + blk_zone_write_plug_unplug_bio(zwplug); +} + +static void blk_zone_wplug_bio_work(struct work_struct *work) +{ + struct blk_zone_wplug *zwplug = + container_of(work, struct blk_zone_wplug, bio_work); + unsigned long flags; + struct bio *bio; + + /* + * Unplug and submit the next plugged BIO. If we do not have any, clear + * the plugged flag. + */ + blk_zone_wplug_lock(zwplug, flags); + + bio = bio_list_pop(&zwplug->bio_list); + if (!bio) { + zwplug->flags &= ~BLK_ZONE_WPLUG_PLUGGED; + blk_zone_wplug_unlock(zwplug, flags); + return; + } + + blk_zone_wplug_unlock(zwplug, flags); + + /* + * blk-mq devices will reuse the reference on the request queue usage + * we took when the BIO was plugged, but the submission path for + * BIO-based devices will not do that. So drop this reference here. + */ + if (bio->bi_bdev->bd_has_submit_bio) + blk_queue_exit(bio->bi_bdev->bd_disk->queue); + + submit_bio_noacct_nocheck(bio); +} + +static struct blk_zone_wplug *blk_zone_alloc_write_plugs(unsigned int nr_zones) +{ + struct blk_zone_wplug *zwplugs; + unsigned int i; + + zwplugs = kvcalloc(nr_zones, sizeof(struct blk_zone_wplug), GFP_NOIO); + if (!zwplugs) + return NULL; + + for (i = 0; i < nr_zones; i++) { + spin_lock_init(&zwplugs[i].lock); + bio_list_init(&zwplugs[i].bio_list); + INIT_WORK(&zwplugs[i].bio_work, blk_zone_wplug_bio_work); + } + + return zwplugs; +} + +static void blk_zone_free_write_plugs(struct gendisk *disk, + struct blk_zone_wplug *zwplugs, + unsigned int nr_zones) +{ + struct blk_zone_wplug *zwplug = zwplugs; + unsigned int i, n; + + if (!zwplug) + return; + + /* Make sure we do not leak any plugged BIO. */ + for (i = 0; i < nr_zones; i++, zwplug++) { + n = blk_zone_wplug_abort(disk, zwplug); + if (n) + pr_warn_ratelimited("%s: zone %u, %u plugged BIOs aborted\n", + disk->disk_name, i, n); + } + + kvfree(zwplugs); +} + +void disk_free_zone_resources(struct gendisk *disk) { kfree(disk->conv_zones_bitmap); disk->conv_zones_bitmap = NULL; kfree(disk->seq_zones_wlock); disk->seq_zones_wlock = NULL; + + blk_zone_free_write_plugs(disk, disk->zone_wplugs, disk->nr_zones); + disk->zone_wplugs = NULL; } struct blk_revalidate_zone_args { @@ -442,6 +826,7 @@ struct blk_revalidate_zone_args { unsigned long *conv_zones_bitmap; unsigned long *seq_zones_wlock; unsigned int nr_zones; + struct blk_zone_wplug *zone_wplugs; sector_t sector; }; @@ -496,6 +881,7 @@ static int blk_revalidate_zone_cb(struct blk_zone *zone, unsigned int idx, return -ENOMEM; } set_bit(idx, args->conv_zones_bitmap); + args->zone_wplugs[idx].flags |= BLK_ZONE_WPLUG_CONV; break; case BLK_ZONE_TYPE_SEQWRITE_REQ: if (!args->seq_zones_wlock) { @@ -540,7 +926,7 @@ int blk_revalidate_disk_zones(struct gendisk *disk, sector_t capacity = get_capacity(disk); struct blk_revalidate_zone_args args = { }; unsigned int noio_flag; - int ret; + int ret = -ENOMEM; if (WARN_ON_ONCE(!blk_queue_is_zoned(q))) return -EIO; @@ -570,9 +956,14 @@ int blk_revalidate_disk_zones(struct gendisk *disk, * Ensure that all memory allocations in this context are done as if * GFP_NOIO was specified. */ + noio_flag = memalloc_noio_save(); + args.disk = disk; args.nr_zones = (capacity + zone_sectors - 1) >> ilog2(zone_sectors); - noio_flag = memalloc_noio_save(); + args.zone_wplugs = blk_zone_alloc_write_plugs(args.nr_zones); + if (!args.zone_wplugs) + goto out_restore_noio; + ret = disk->fops->report_zones(disk, 0, UINT_MAX, blk_revalidate_zone_cb, &args); if (!ret) { @@ -601,17 +992,24 @@ int blk_revalidate_disk_zones(struct gendisk *disk, disk->nr_zones = args.nr_zones; swap(disk->seq_zones_wlock, args.seq_zones_wlock); swap(disk->conv_zones_bitmap, args.conv_zones_bitmap); + swap(disk->zone_wplugs, args.zone_wplugs); if (update_driver_data) update_driver_data(disk); ret = 0; } else { pr_warn("%s: failed to revalidate zones\n", disk->disk_name); - disk_free_zone_bitmaps(disk); + disk_free_zone_resources(disk); } blk_mq_unfreeze_queue(q); kfree(args.seq_zones_wlock); kfree(args.conv_zones_bitmap); + blk_zone_free_write_plugs(disk, args.zone_wplugs, args.nr_zones); + + return ret; + +out_restore_noio: + memalloc_noio_restore(noio_flag); return ret; } EXPORT_SYMBOL_GPL(blk_revalidate_disk_zones); diff --git a/block/blk.h b/block/blk.h index 5180e103ed9c..d0ecd5a2002c 100644 --- a/block/blk.h +++ b/block/blk.h @@ -403,7 +403,13 @@ static inline struct bio *blk_queue_bounce(struct bio *bio, } #ifdef CONFIG_BLK_DEV_ZONED -void disk_free_zone_bitmaps(struct gendisk *disk); +void disk_free_zone_resources(struct gendisk *disk); +static inline bool bio_zone_write_plugging(struct bio *bio) +{ + return bio_flagged(bio, BIO_ZONE_WRITE_PLUGGING); +} +void blk_zone_write_plug_bio_merged(struct bio *bio); +void blk_zone_write_plug_attempt_merge(struct request *rq); static inline void blk_zone_complete_request_bio(struct request *rq, struct bio *bio) { @@ -411,22 +417,42 @@ static inline void blk_zone_complete_request_bio(struct request *rq, * For zone append requests, the request sector indicates the location * at which the BIO data was written. Return this value to the BIO * issuer through the BIO iter sector. + * For plugged zone writes, we need the original BIO sector so + * that blk_zone_write_plug_bio_endio() can lookup the zone write plug. */ - if (req_op(rq) == REQ_OP_ZONE_APPEND) + if (req_op(rq) == REQ_OP_ZONE_APPEND || bio_zone_write_plugging(bio)) bio->bi_iter.bi_sector = rq->__sector; } +void blk_zone_write_plug_bio_endio(struct bio *bio); +void blk_zone_write_plug_complete_request(struct request *rq); int blkdev_report_zones_ioctl(struct block_device *bdev, unsigned int cmd, unsigned long arg); int blkdev_zone_mgmt_ioctl(struct block_device *bdev, blk_mode_t mode, unsigned int cmd, unsigned long arg); #else /* CONFIG_BLK_DEV_ZONED */ -static inline void disk_free_zone_bitmaps(struct gendisk *disk) +static inline void disk_free_zone_resources(struct gendisk *disk) +{ +} +static inline bool bio_zone_write_plugging(struct bio *bio) +{ + return false; +} +static inline void blk_zone_write_plug_bio_merged(struct bio *bio) +{ +} +static inline void blk_zone_write_plug_attempt_merge(struct request *rq) { } static inline void blk_zone_complete_request_bio(struct request *rq, struct bio *bio) { } +static inline void blk_zone_write_plug_bio_endio(struct bio *bio) +{ +} +static inline void blk_zone_write_plug_complete_request(struct request *rq) +{ +} static inline int blkdev_report_zones_ioctl(struct block_device *bdev, unsigned int cmd, unsigned long arg) { diff --git a/block/genhd.c b/block/genhd.c index d74fb5b4ae68..fe45d4713b28 100644 --- a/block/genhd.c +++ b/block/genhd.c @@ -1182,7 +1182,7 @@ static void disk_release(struct device *dev) disk_release_events(disk); kfree(disk->random); - disk_free_zone_bitmaps(disk); + disk_free_zone_resources(disk); xa_destroy(&disk->part_tbl); disk->queue->disk = NULL; diff --git a/include/linux/blk-mq.h b/include/linux/blk-mq.h index 7a8150a5f051..bc74f904b5a1 100644 --- a/include/linux/blk-mq.h +++ b/include/linux/blk-mq.h @@ -55,6 +55,8 @@ typedef __u32 __bitwise req_flags_t; #define RQF_SPECIAL_PAYLOAD ((__force req_flags_t)(1 << 18)) /* The per-zone write lock is held for this request */ #define RQF_ZONE_WRITE_LOCKED ((__force req_flags_t)(1 << 19)) +/* The request completion needs to be signaled to zone write pluging. */ +#define RQF_ZONE_WRITE_PLUGGING ((__force req_flags_t)(1 << 20)) /* ->timeout has been called, don't expire again */ #define RQF_TIMED_OUT ((__force req_flags_t)(1 << 21)) #define RQF_RESV ((__force req_flags_t)(1 << 23)) diff --git a/include/linux/blk_types.h b/include/linux/blk_types.h index 1c07848dea7e..19839d303289 100644 --- a/include/linux/blk_types.h +++ b/include/linux/blk_types.h @@ -232,7 +232,12 @@ struct bio { struct bvec_iter bi_iter; - blk_qc_t bi_cookie; + union { + /* for polled bios: */ + blk_qc_t bi_cookie; + /* for plugged zoned writes only: */ + unsigned int __bi_nr_segments; + }; bio_end_io_t *bi_end_io; void *bi_private; #ifdef CONFIG_BLK_CGROUP @@ -303,6 +308,7 @@ enum { BIO_QOS_MERGED, /* but went through rq_qos merge path */ BIO_REMAPPED, BIO_ZONE_WRITE_LOCKED, /* Owns a zoned device zone write lock */ + BIO_ZONE_WRITE_PLUGGING, /* bio handled through zone write plugging */ BIO_FLAG_LAST }; diff --git a/include/linux/blkdev.h b/include/linux/blkdev.h index 0bb897f0501c..d58aaed6dc24 100644 --- a/include/linux/blkdev.h +++ b/include/linux/blkdev.h @@ -39,6 +39,7 @@ struct rq_qos; struct blk_queue_stats; struct blk_stat_callback; struct blk_crypto_profile; +struct blk_zone_wplug; extern const struct device_type disk_type; extern const struct device_type part_type; @@ -193,6 +194,7 @@ struct gendisk { unsigned int max_active_zones; unsigned long *conv_zones_bitmap; unsigned long *seq_zones_wlock; + struct blk_zone_wplug *zone_wplugs; #endif /* CONFIG_BLK_DEV_ZONED */ #if IS_ENABLED(CONFIG_CDROM) @@ -658,6 +660,7 @@ static inline unsigned int bdev_max_active_zones(struct block_device *bdev) return bdev->bd_disk->max_active_zones; } +bool blk_zone_write_plug_bio(struct bio *bio, unsigned int nr_segs); #else /* CONFIG_BLK_DEV_ZONED */ static inline unsigned int bdev_nr_zones(struct block_device *bdev) { @@ -685,6 +688,11 @@ static inline unsigned int bdev_max_active_zones(struct block_device *bdev) { return 0; } +static inline bool blk_zone_write_plug_bio(struct bio *bio, + unsigned int nr_segs) +{ + return false; +} #endif /* CONFIG_BLK_DEV_ZONED */ static inline unsigned int blk_queue_depth(struct request_queue *q)
Zone write plugging implements a per-zone "plug" for write operations to tightly control the submission and execution order of writes to sequential write required zones of a zoned block device. Per-zone plugging of writes guarantees that at any time at most one write request per zone is in flight. This mechanism is intended to replace zone write locking which is controlled at the scheduler level and implemented only by mq-deadline. Unlike zone write locking which operates on requests, zone write plugging operates on BIOs. A zone write plug is simply a BIO list that is atomically manipulated using a spinlock and a kblockd submission work. A write BIO to a zone is "plugged" to delay its execution if a write BIO for the same zone was already issued, that is, if a write request for the same zone is being executed. The next plugged BIO is unplugged and issued once the write request completes. This mechanism allows to: - Untangles zone write ordering from block IO schedulers. This allows removing the restriction on using only mq-deadline for zoned block devices. Any block IO scheduler, including "none" can be used. - Zone write plugging operates on BIOs instead of requests. Plugged BIOs waiting for execution thus do not hold scheduling tags and thus are not preventing other BIOs to proceed (reads or writes to other zones). Depending on the workload, this can significantly improve the device use and performance. - Both blk-mq (request) based zoned devices and BIO-based devices (e.g. device mapper) can use zone write plugging. It is mandatory for the former but optional for the latter: BIO-based driver can use zone write plugging to implement write ordering guarantees, or the drivers can implement their own if needed. - The code is less invasive in the block layer and is mostly limited to blk-zoned.c with some small changes in blk-mq.c, blk-merge.c and bio.c. Zone write plugging is implemented using struct blk_zone_wplug. This structurei includes a spinlock, a BIO list and a work structure to handle the submission of plugged BIOs. Plugging of zone write BIOs is done using the function blk_zone_write_plug_bio() which returns false if a BIO execution does not need to be delayed and true otherwise. This function is called from blk_mq_submit_bio() after a BIO is split to avoid large BIOs spanning multiple zones which would cause mishandling of zone write plugging. This enables by default zone write plugging for any mq request-based block device. BIO-based device drivers can also use zone write plugging by expliclty calling blk_zone_write_plug_bio() in their ->submit_bio method. For such devices, the driver must ensure that a BIO passed to blk_zone_write_plug_bio() is already split and not straddling zone boundaries. Only write and write zeroes BIOs are plugged. Zone write plugging does not introduce any significant overhead for other operations. A BIO that is being handled through zone write plugging is flagged using the new BIO flag BIO_ZONE_WRITE_PLUGGING. A request handling a BIO flagged with this new flag is flagged with the new RQF_ZONE_WRITE_PLUGGING flag. The completion processing of BIOs and requests flagged trigger respectively calls to the functions blk_zone_write_plug_bio_endio() and blk_zone_write_plug_complete_request(). The latter function is used to trigger submission of the next plugged BIO using the zone plug work. blk_zone_write_plug_bio_endio() does the same for BIO-based devices. This ensures that at any time, at most one request (blk-mq devices) or one BIO (BIO-based devices) are being executed for any zone. The handling of zone write plug using a per-zone plug spinlock maximizes parrallelism and device usage by allowing multiple zones to be writen simultaneously without lock contention. Zone write plugging ignores flush BIOs without data. Hovever, any flush BIO that has data is always plugged so that the write part of the flush sequence is serialized with other regular writes. Given that any BIO handled through zone write plugging will be the only BIO in flight for the target zone when it is executed, the unplugging and submission of a BIO will have no chance of successfully merging with plugged requests or requests in the scheduler. To overcome this potential performance loss, blk_mq_submit_bio() calls the function blk_zone_write_plug_attempt_merge() to try to merge other plugged BIOs with the one just unplugged. Successful merging is signaled using blk_zone_write_plug_bio_merged(), called from bio_attempt_back_merge(). Furthermore, to avoid recalculating the number of segments of plugged BIOs to attempt merging, the number of segments of a plugged BIO is saved using the new struct bio field __bi_nr_segments. To avoid growing the size of struct bio, this field is added as a union with the bio_cookie field. This is safe to do as polling is always disabled for plugged BIOs. When BIOs are plugged in a zone write plug, the device request queue usage counter is always incremented. This kept and reused when the plugged BIO is unplugged and submitted again using submit_bio_noacct_nocheck(). For this case, the unplugged BIO is already flagged with BIO_ZONE_WRITE_PLUGGING and blk_mq_submit_bio() proceeds directly to allocating a new request for the BIO, re-using the usage reference count taken when the BIO was plugged. This extra reference count is dropped in blk_zone_write_plug_attempt_merge() for any plugged BIO that is successfully merged. Given that BIO-based devices will not take this path, the extra reference is dropped when a plugged BIO is unplugged and submitted. To match the new data structures used for zoned disks, the function disk_free_zone_bitmaps() is renamed to the more generic disk_free_zone_resources(). This commit contains contributions from Christoph Hellwig <hch@lst.de>. Signed-off-by: Damien Le Moal <dlemoal@kernel.org> --- block/bio.c | 7 + block/blk-merge.c | 11 + block/blk-mq.c | 28 +++ block/blk-zoned.c | 408 +++++++++++++++++++++++++++++++++++++- block/blk.h | 32 ++- block/genhd.c | 2 +- include/linux/blk-mq.h | 2 + include/linux/blk_types.h | 8 +- include/linux/blkdev.h | 8 + 9 files changed, 496 insertions(+), 10 deletions(-)