Message ID | 20210608183339.70609-1-dwagner@suse.de (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | blk-mq: Do not lookup ctx with invalid index | expand |
On Tue, Jun 08, 2021 at 08:33:39PM +0200, Daniel Wagner wrote: > cpumask_first_and() returns >= nr_cpu_ids if the two provided masks do > not share a common bit. Verify we get a valid value back from > cpumask_first_and(). So I got feedback on this issue (but not on the patch itself yet). The system starts with 16 virtual CPU cores and during the test 4 cores are removed[1] and as soon there is an error on the storage side, the reset code on the host ends up in this path and crashes. I still don't understand why the CPU removal is not updating the CPU mask correctly before we hit the reset path. I'll continue to investigate. [1] echo 0 > /sys/devices/system/cpu/cpu#/online
On Mon, Jun 14, 2021 at 01:37:06PM +0200, Daniel Wagner wrote: > On Tue, Jun 08, 2021 at 08:33:39PM +0200, Daniel Wagner wrote: > > cpumask_first_and() returns >= nr_cpu_ids if the two provided masks do > > not share a common bit. Verify we get a valid value back from > > cpumask_first_and(). > > So I got feedback on this issue (but not on the patch itself yet). The > system starts with 16 virtual CPU cores and during the test 4 cores are > removed[1] and as soon there is an error on the storage side, the reset > code on the host ends up in this path and crashes. I still don't > understand why the CPU removal is not updating the CPU mask correctly > before we hit the reset path. I'll continue to investigate. We don't update hctx->cpumask when CPU is added/removed, and that is assigned against cpu_possible_mask from beginning. It is one long-term issue, which can be triggered when all cpus in hctx->cpumask become offline. The thing is that only nvmf_connect_io_queue() allocates request via specified hctx. thanks, Ming
diff --git a/block/blk-mq.c b/block/blk-mq.c index 4261adee9964..50f97bc97895 100644 --- a/block/blk-mq.c +++ b/block/blk-mq.c @@ -469,6 +469,8 @@ struct request *blk_mq_alloc_request_hctx(struct request_queue *q, if (!blk_mq_hw_queue_mapped(data.hctx)) goto out_queue_exit; cpu = cpumask_first_and(data.hctx->cpumask, cpu_online_mask); + if (cpu >= nr_cpu_ids) + goto out_queue_exit; data.ctx = __blk_mq_get_ctx(q, cpu); if (!q->elevator)
cpumask_first_and() returns >= nr_cpu_ids if the two provided masks do not share a common bit. Verify we get a valid value back from cpumask_first_and(). Signed-off-by: Daniel Wagner <dwagner@suse.de> --- Hi, Supported: Yes, External CPU: 11 PID: 7264 Comm: kworker/u64:1 Kdump: loaded Tainted: G O KX 5.3.18-53-default #1 SLE15-SP3 Workqueue: nvme-wq nvme_fc_connect_ctrl_work [nvme_fc] NIP: c0000000006b6ab0 LR: c0000000006b6ea0 CTR: c0000000006ad740 REGS: c0000003dbdbb570 TRAP: 0380 Tainted: G O KX (5.3.18-53-default) MSR: 800000000280b033 <SF,VEC,VSX,EE,FP,ME,IR,DR,RI,LE> CR: 44002244 XER: 00000001 CFAR: c0000000006b6e9c IRQMASK: 0 GPR00: c0000000006b6ea0 c0000003dbdbb800 c000000001752a00 c0000003f7b8dd40 GPR04: 8009fffc21e76a80 0000000000000000 0000000000000001 0000000000000001 GPR08: 00000003fb950000 0000000000000000 0000000000000000 c008000000b77840 GPR12: 0000000000002000 c00000001ec60200 c0000000001727f8 c0000000631152c0 GPR16: 0000000000000190 0000000000000600 0000000000001f00 0000000000000000 GPR20: 0000000000000004 0700000048000000 c0000000ff0b0470 3000000004000000 GPR24: c0000003e2dc6868 0000000000000001 0000000000000000 c0000003f7b8dd40 GPR28: 0000000000000000 c0000003f7b8dda0 0000000000000000 c0000003dbdbb8e0 NIP [c0000000006b6ab0] blk_mq_put_tag+0x20/0x90 LR [c0000000006b6ea0] blk_mq_get_tag+0x380/0x3b0 Call Trace: [c0000003dbdbb800] [c0000000006b63e4] __blk_mq_get_tag+0x44/0xf0 (unreliable) [c0000003dbdbb820] [c0000000006b6ea0] blk_mq_get_tag+0x380/0x3b0 [c0000003dbdbb8b0] [c0000000006ad8c4] blk_mq_alloc_request_hctx+0x184/0x230 [c0000003dbdbb930] [c008000000b60edc] nvme_alloc_request+0x54/0xe0 [nvme_core] [c0000003dbdbb960] [c008000000b63dbc] __nvme_submit_sync_cmd+0x74/0x290 [nvme_core] [c0000003dbdbb9f0] [c008000000ec18f0] nvmf_connect_io_queue+0x148/0x1f0 [nvme_fabrics] [c0000003dbdbbae0] [c008000000f97254] nvme_fc_connect_io_queues+0x2d8/0x36c [nvme_fc] [c0000003dbdbbbc0] [c008000000f959fc] nvme_fc_connect_ctrl_work+0xa34/0xb90 [nvme_fc] [c0000003dbdbbc80] [c000000000167d54] process_one_work+0x2c4/0x5c0 [c0000003dbdbbd10] [c000000000168330] worker_thread+0x2e0/0x6b0 [c0000003dbdbbdb0] [c00000000017299c] kthread+0x1ac/0x1c0 [c0000003dbdbbe20] [c00000000000cda8] ret_from_kernel_thread+0x5c/0x74 We allocate a tag in blk_mq_get_tag() but then we find out the BLK_MQ_S_INACTIVE is set and call blk_mq_put_tag() found_tag: /* * Give up this allocation if the hctx is inactive. The caller will * retry on an active hctx. */ if (unlikely(test_bit(BLK_MQ_S_INACTIVE, &data->hctx->state))) { blk_mq_put_tag(tags, data->ctx, tag + tag_offset); return BLK_MQ_NO_TAG; } and if my dissasembly works (unfortuantly, I dont have access to the correct binaries, so we are looking at my own compiled version), the load for ctx->cpu crashes. c0000000006b7870 <blk_mq_put_tag>: c0000000006b7870: 0a 01 4c 3c addis r2,r12,266 c0000000006b7874: 90 b5 42 38 addi r2,r2,-19056 c0000000006b7878: a6 02 08 7c mflr r0 c0000000006b787c: ad f0 9b 4b bl c000000000076928 <_mcount> c0000000006b7880: a6 02 08 7c mflr r0 c0000000006b7884: 78 2b a9 7c mr r9,r5 c0000000006b7888: 10 00 01 f8 std r0,16(r1) c0000000006b788c: e1 ff 21 f8 stdu r1,-32(r1) c0000000006b7890: 80 00 a4 80 lwz r5,128(r4) // XXX this is problably crashing c0000000006b7894: 04 00 83 80 lwz r4,4(r3) c0000000006b7898: 40 20 89 7f cmplw cr7,r9,r4 c0000000006b789c: 44 00 9c 41 blt cr7,c0000000006b78e0 <blk_mq_put_tag+0x70> I figured the only way the data->ctx could be invalid is, when cpumask_first_and() returns an invalid index. This seems to be a nasty corner case. The cpumask changes and we hit the queue in inactive state. Not sure this is really happening. The patch has not yet verified if it really helps. I thought the check might be a good thing independent if it really fixes the above crash or not. Thanks, Daniel block/blk-mq.c | 2 ++ 1 file changed, 2 insertions(+)