@@ -1203,12 +1203,6 @@ int blkcg_init_queue(struct request_queue *q)
if (ret)
goto err_destroy_all;
- ret = blk_iolatency_init(q);
- if (ret) {
- blk_throtl_exit(q);
- goto err_destroy_all;
- }
-
return 0;
err_destroy_all:
@@ -90,6 +90,12 @@ struct blk_iolatency {
atomic_t enabled;
};
+static struct rq_qos_ops blkcg_iolatency_ops;
+static inline struct rq_qos *blkcg_rq_qos(struct request_queue *q)
+{
+ return rq_qos_by_id(q, blkcg_iolatency_ops.id);
+}
+
static inline struct blk_iolatency *BLKIOLATENCY(struct rq_qos *rqos)
{
return container_of(rqos, struct blk_iolatency, rqos);
@@ -646,13 +652,18 @@ static void blkcg_iolatency_exit(struct rq_qos *rqos)
del_timer_sync(&blkiolat->timer);
blkcg_deactivate_policy(rqos->q, &blkcg_policy_iolatency);
+ rq_qos_deactivate(rqos);
kfree(blkiolat);
}
+static int blk_iolatency_init(struct request_queue *q);
static struct rq_qos_ops blkcg_iolatency_ops = {
+ .name = "iolat",
+ .flags = RQOS_FLAG_CGRP_POL,
.throttle = blkcg_iolatency_throttle,
.done_bio = blkcg_iolatency_done_bio,
.exit = blkcg_iolatency_exit,
+ .init = blk_iolatency_init,
};
static void blkiolatency_timer_fn(struct timer_list *t)
@@ -727,15 +738,10 @@ int blk_iolatency_init(struct request_queue *q)
return -ENOMEM;
rqos = &blkiolat->rqos;
- rqos->id = RQ_QOS_LATENCY;
- rqos->ops = &blkcg_iolatency_ops;
- rqos->q = q;
-
- rq_qos_add(q, rqos);
-
+ rq_qos_activate(q, rqos, &blkcg_iolatency_ops);
ret = blkcg_activate_policy(q, &blkcg_policy_iolatency);
if (ret) {
- rq_qos_del(q, rqos);
+ rq_qos_deactivate(rqos);
kfree(blkiolat);
return ret;
}
@@ -1046,12 +1052,23 @@ static struct blkcg_policy blkcg_policy_iolatency = {
static int __init iolatency_init(void)
{
- return blkcg_policy_register(&blkcg_policy_iolatency);
+ int ret;
+
+ ret = rq_qos_register(&blkcg_iolatency_ops);
+ if (ret)
+ return ret;
+
+ ret = blkcg_policy_register(&blkcg_policy_iolatency);
+ if (ret)
+ rq_qos_unregister(&blkcg_iolatency_ops);
+
+ return ret;
}
static void __exit iolatency_exit(void)
{
blkcg_policy_unregister(&blkcg_policy_iolatency);
+ rq_qos_unregister(&blkcg_iolatency_ops);
}
module_init(iolatency_init);
@@ -823,8 +823,6 @@ void blk_mq_debugfs_unregister_sched(struct request_queue *q)
static const char *rq_qos_id_to_name(enum rq_qos_id id)
{
switch (id) {
- case RQ_QOS_LATENCY:
- return "latency";
case RQ_QOS_COST:
return "cost";
case RQ_QOS_IOPRIO:
@@ -14,7 +14,6 @@
struct blk_mq_debugfs_attr;
enum rq_qos_id {
- RQ_QOS_LATENCY,
RQ_QOS_COST,
RQ_QOS_IOPRIO,
};
@@ -85,11 +84,6 @@ static inline struct rq_qos *rq_qos_by_id(struct request_queue *q, int id)
return rqos;
}
-static inline struct rq_qos *blkcg_rq_qos(struct request_queue *q)
-{
- return rq_qos_by_id(q, RQ_QOS_LATENCY);
-}
-
static inline void rq_wait_init(struct rq_wait *rq_wait)
{
atomic_set(&rq_wait->inflight, 0);
@@ -400,12 +400,6 @@ static inline void blk_queue_bounce(struct request_queue *q, struct bio **bio)
__blk_queue_bounce(q, bio);
}
-#ifdef CONFIG_BLK_CGROUP_IOLATENCY
-extern int blk_iolatency_init(struct request_queue *q);
-#else
-static inline int blk_iolatency_init(struct request_queue *q) { return 0; }
-#endif
-
struct bio *blk_next_bio(struct bio *bio, unsigned int nr_pages, gfp_t gfp);
#ifdef CONFIG_BLK_DEV_ZONED
Make blk-iolatency pluggable. Then we can close or open it through /sys/block/xxx/queue/qos. Signed-off-by: Wang Jianchao (Kuaishou) <jianchao.wan9@gmail.com> --- block/blk-cgroup.c | 6 ------ block/blk-iolatency.c | 33 +++++++++++++++++++++++++-------- block/blk-mq-debugfs.c | 2 -- block/blk-rq-qos.h | 6 ------ block/blk.h | 6 ------ 5 files changed, 25 insertions(+), 28 deletions(-)