From patchwork Mon Jul 6 13:22:18 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Stanley Chu X-Patchwork-Id: 11645745 Return-Path: Received: from mail.kernel.org (pdx-korg-mail-1.web.codeaurora.org [172.30.200.123]) by pdx-korg-patchwork-2.web.codeaurora.org (Postfix) with ESMTP id D543314B7 for ; Mon, 6 Jul 2020 13:22:28 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id B94FD2070C for ; Mon, 6 Jul 2020 13:22:28 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=pass (1024-bit key) header.d=mediatek.com header.i=@mediatek.com header.b="qC7hcFS+" Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1729184AbgGFNWZ (ORCPT ); Mon, 6 Jul 2020 09:22:25 -0400 Received: from mailgw02.mediatek.com ([210.61.82.184]:13617 "EHLO mailgw02.mediatek.com" rhost-flags-OK-FAIL-OK-FAIL) by vger.kernel.org with ESMTP id S1729176AbgGFNWZ (ORCPT ); Mon, 6 Jul 2020 09:22:25 -0400 X-UUID: f28342167a0c4f81b76f5811ceb2ca9f-20200706 DKIM-Signature: v=1; a=rsa-sha256; q=dns/txt; c=relaxed/relaxed; d=mediatek.com; s=dk; h=Content-Transfer-Encoding:Content-Type:MIME-Version:Message-ID:Date:Subject:CC:To:From; bh=hH72iy737LTFZ8Jp12sRADIb3rpiyBdOegJpRLwFXpA=; b=qC7hcFS+yAxMi8qqonesIQDxXohAHIbTQQASwkOcxTUCVKY6qN6eU/Ad9hfyywxZoqCIOw5YmVgrgRdE9/69R4yvOjlRnUaAcNdFYKVHJ3iiz4LQJitkeUgDg7/st2ljWhn1Uz/az+OMqRd0f1SZ0d+krlrRSynQOOSLRLXac+E=; X-UUID: f28342167a0c4f81b76f5811ceb2ca9f-20200706 Received: from mtkcas10.mediatek.inc [(172.21.101.39)] by mailgw02.mediatek.com (envelope-from ) (Cellopoint E-mail Firewall v4.1.10 Build 0809 with TLS) with ESMTP id 944152981; Mon, 06 Jul 2020 21:22:20 +0800 Received: from MTKCAS06.mediatek.inc (172.21.101.30) by mtkmbs05n2.mediatek.inc (172.21.101.140) with Microsoft SMTP Server (TLS) id 15.0.1497.2; Mon, 6 Jul 2020 21:22:15 +0800 Received: from mtksdccf07.mediatek.inc (172.21.84.99) by MTKCAS06.mediatek.inc (172.21.101.73) with Microsoft SMTP Server id 15.0.1497.2 via Frontend Transport; Mon, 6 Jul 2020 21:22:15 +0800 From: Stanley Chu To: , , , , , CC: , , , , , , , , , , , , , Stanley Chu Subject: [RFC PATCH v3] scsi: ufs: Quiesce all scsi devices before shutdown Date: Mon, 6 Jul 2020 21:22:18 +0800 Message-ID: <20200706132218.21171-1-stanley.chu@mediatek.com> X-Mailer: git-send-email 2.18.0 MIME-Version: 1.0 X-MTK: N Sender: linux-scsi-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-scsi@vger.kernel.org Currently I/O request could be still submitted to UFS device while UFS is working on shutdown flow. This may lead to racing as below scenarios and finally system may crash due to unclocked register accesses. To fix this kind of issues, specifically quiesce all SCSI devices before UFS shutdown to block all I/O request sending from block layer. Example of racing scenario: While UFS device is runtime-suspended Thread #1: Executing UFS shutdown flow, e.g., ufshcd_suspend(UFS_SHUTDOWN_PM) Thread #2: Executing runtime resume flow triggered by I/O request, e.g., ufshcd_resume(UFS_RUNTIME_PM) This breaks the assumption that UFS PM flows can not be running concurrently and some unexpected racing behavior may happen. Signed-off-by: Stanley Chu --- drivers/scsi/ufs/ufshcd.c | 38 ++++++++++++++++++++++++++++++++++++++ 1 file changed, 38 insertions(+) diff --git a/drivers/scsi/ufs/ufshcd.c b/drivers/scsi/ufs/ufshcd.c index 59358bb75014..104173c03492 100644 --- a/drivers/scsi/ufs/ufshcd.c +++ b/drivers/scsi/ufs/ufshcd.c @@ -158,6 +158,12 @@ struct ufs_pm_lvl_states ufs_pm_lvl_states[] = { {UFS_POWERDOWN_PWR_MODE, UIC_LINK_OFF_STATE}, }; +#define ufshcd_scsi_for_each_sdev(fn) \ + list_for_each_entry(starget, &hba->host->__targets, siblings) { \ + __starget_for_each_device(starget, NULL, \ + fn); \ + } + static inline enum ufs_dev_pwr_mode ufs_get_pm_lvl_to_dev_pwr_mode(enum ufs_pm_level lvl) { @@ -8588,6 +8594,19 @@ int ufshcd_runtime_idle(struct ufs_hba *hba) } EXPORT_SYMBOL(ufshcd_runtime_idle); +static void ufshcd_cleanup_queue(struct scsi_device *sdev, void *data) +{ + if (sdev->request_queue) + blk_cleanup_queue(sdev->request_queue); +} + +static void ufshcd_quiece_sdev(struct scsi_device *sdev, void *data) +{ + /* Suspended devices are already quiecsed and shall be skipped */ + if (!pm_runtime_suspended(&sdev->sdev_gendev)) + scsi_device_quiesce(sdev); +} + /** * ufshcd_shutdown - shutdown routine * @hba: per adapter instance @@ -8599,6 +8618,7 @@ EXPORT_SYMBOL(ufshcd_runtime_idle); int ufshcd_shutdown(struct ufs_hba *hba) { int ret = 0; + struct scsi_target *starget; if (!hba->is_powered) goto out; @@ -8612,7 +8632,25 @@ int ufshcd_shutdown(struct ufs_hba *hba) goto out; } + /* + * Quiesce all SCSI devices to prevent any non-PM requests sending + * from block layer during and after shutdown. + * + * Here we can not use blk_cleanup_queue() since PM requests + * (with BLK_MQ_REQ_PREEMPT flag) are still required to be sent + * through block layer. Therefore SCSI command queued after the + * scsi_target_quiesce() call returned will block until + * blk_cleanup_queue() is called. + * + * Besides, scsi_target_"un"quiesce (e.g., scsi_target_resume) can + * be ignored since shutdown is one-way flow. + */ + ufshcd_scsi_for_each_sdev(ufshcd_quiece_sdev); + ret = ufshcd_suspend(hba, UFS_SHUTDOWN_PM); + + /* Set queue as dying to not block queueing commands */ + ufshcd_scsi_for_each_sdev(ufshcd_cleanup_queue); out: if (ret) dev_err(hba->dev, "%s failed, err %d\n", __func__, ret);