From patchwork Thu Sep 3 21:23:06 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Ben Levinsky X-Patchwork-Id: 11755191 Return-Path: Received: from mail.kernel.org (pdx-korg-mail-1.web.codeaurora.org [172.30.200.123]) by pdx-korg-patchwork-2.web.codeaurora.org (Postfix) with ESMTP id 6DB2791F for ; Thu, 3 Sep 2020 21:23:35 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 4F182206EF for ; Thu, 3 Sep 2020 21:23:35 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=pass (1024-bit key) header.d=xilinx.onmicrosoft.com header.i=@xilinx.onmicrosoft.com header.b="gAKLmrQz" Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726323AbgICVX0 (ORCPT ); Thu, 3 Sep 2020 17:23:26 -0400 Received: from mail-bn7nam10on2076.outbound.protection.outlook.com ([40.107.92.76]:16672 "EHLO NAM10-BN7-obe.outbound.protection.outlook.com" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1728107AbgICVXU (ORCPT ); Thu, 3 Sep 2020 17:23:20 -0400 ARC-Seal: i=1; a=rsa-sha256; s=arcselector9901; d=microsoft.com; cv=none; b=K+HupyKmm5YSSvxfNhsIpPiOn4pgv7M6mrK3Y27E1cLQ1kJdtQZPqEToO+tbCEOSCeM49iTA1NEuPygK4fJ7qcKD4DwiEpcEcB5Rn9s8ZHMlOtG1WbDzvLVxQ6lW3ctpoRAHJh3NHqAiC3OC/KLPXqWpTOvPBzgBASFa1XJrL6mNYyLUaI8eNmNVFAROuE0Wtufu1y96RmCR+Av+GDgeMJHSMwLsrkL82j7/St386Ew55ag74oXgA9awWmvsCWfz8ZEh2tETrAHhv3lWju1zrsEj3T8C44CvETMfC3dsAdD1LmzSIgNyCqa3grnoKUYPKcZHBVbdSjQaBIMRnXdCDw== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=microsoft.com; s=arcselector9901; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-SenderADCheck; bh=p5xh2iMbOUuYykH3VtoEpRrVeANMsj9APMUlrpsyzW0=; b=dPVpk2KTjh6ogzOSleG0AhuXxjSwcbMGLzVnhaPCQPMt4DtUHOcG2d5rIzc0jcN74lP3S87IoXACB03OpmF4brhSkYzQvJ0bBhtD9yBJpheFBKtyT2dZsx5W6iRWtfNm89dYQrdnZoyxsq+VlzdhfmyynfVR+1JUBjuxXtuEFrt1ZAY5jZ630pM7A860GvvcOX6L+ZQ5DS2ipavuXGXAEArV1xUuZQ3bZb6DT2NJ6HFSw/qA5StdWmcSgmoyFtZa6tigbgAMkCzZX59Vj79BTtx7vAZQ3mTdzL61hi2iTbFgKgb0QOfykQSwb46fYo0ZC0VP3bKXpw1UpP+/MZrvlA== ARC-Authentication-Results: i=1; mx.microsoft.com 1; spf=pass (sender ip is 149.199.60.83) smtp.rcpttodomain=lists.infradead.org smtp.mailfrom=xilinx.com; dmarc=bestguesspass action=none header.from=xilinx.com; dkim=none (message not signed); arc=none DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=xilinx.onmicrosoft.com; s=selector2-xilinx-onmicrosoft-com; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-SenderADCheck; bh=p5xh2iMbOUuYykH3VtoEpRrVeANMsj9APMUlrpsyzW0=; b=gAKLmrQzvtq4XoFV3IFLoqjDV13kgpa8QIYaNG3Ogo6LEmRceKo7YIf10Z2skkJm7iEwT9Dm7Ydi8JniFBX1rXGJBXkdijLvD2vcRlQIi9gX+u5OL7odKnryAx3zyW2MQGSZ+amRedagSFYbjIOgz+MkYwth1UaFBieFqa51cb0= Received: from CY4PR22CA0041.namprd22.prod.outlook.com (2603:10b6:903:ed::27) by BY5PR02MB6452.namprd02.prod.outlook.com (2603:10b6:a03:1b6::20) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.3348.15; Thu, 3 Sep 2020 21:23:16 +0000 Received: from CY1NAM02FT052.eop-nam02.prod.protection.outlook.com (2603:10b6:903:ed:cafe::57) by CY4PR22CA0041.outlook.office365.com (2603:10b6:903:ed::27) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.3348.15 via Frontend Transport; Thu, 3 Sep 2020 21:23:16 +0000 X-MS-Exchange-Authentication-Results: spf=pass (sender IP is 149.199.60.83) smtp.mailfrom=xilinx.com; lists.infradead.org; dkim=none (message not signed) header.d=none;lists.infradead.org; dmarc=bestguesspass action=none header.from=xilinx.com; Received-SPF: Pass (protection.outlook.com: domain of xilinx.com designates 149.199.60.83 as permitted sender) receiver=protection.outlook.com; client-ip=149.199.60.83; helo=xsj-pvapsmtpgw01; Received: from xsj-pvapsmtpgw01 (149.199.60.83) by CY1NAM02FT052.mail.protection.outlook.com (10.152.74.123) with Microsoft SMTP Server id 15.20.3326.19 via Frontend Transport; Thu, 3 Sep 2020 21:23:16 +0000 Received: from [149.199.38.66] (port=37610 helo=smtp.xilinx.com) by xsj-pvapsmtpgw01 with esmtp (Exim 4.90) (envelope-from ) id 1kDwhJ-0006tr-Kr; Thu, 03 Sep 2020 14:23:09 -0700 Received: from [127.0.0.1] (helo=localhost) by smtp.xilinx.com with smtp (Exim 4.63) (envelope-from ) id 1kDwhP-0000OF-VM; Thu, 03 Sep 2020 14:23:16 -0700 Received: from xsj-pvapsmtp01 (smtp-fallback.xilinx.com [149.199.38.66] (may be forged)) by xsj-smtp-dlp2.xlnx.xilinx.com (8.13.8/8.13.1) with ESMTP id 083LNAfB023389; Thu, 3 Sep 2020 14:23:10 -0700 Received: from [172.19.2.206] (helo=xsjblevinsk50.xilinx.com) by xsj-pvapsmtp01 with esmtp (Exim 4.63) (envelope-from ) id 1kDwhK-0000M3-5y; Thu, 03 Sep 2020 14:23:10 -0700 From: Ben Levinsky To: punit1.agrawal@toshiba.co.jp, stefanos@xilinx.com, michals@xilinx.com Cc: michael.auchter@ni.com, devicetree@vger.kernel.org, mathieu.poirier@linaro.org, emooring@xilinx.com, linux-remoteproc@vger.kernel.org, linux-kernel@vger.kernel.org, jliang@xilinx.com, robh+dt@kernel.org, linux-arm-kernel@lists.infradead.org Subject: [PATCH v12 1/5] firmware: xilinx: Add ZynqMP firmware ioctl enums for RPU configuration. Date: Thu, 3 Sep 2020 14:23:06 -0700 Message-Id: <20200903212310.17990-2-ben.levinsky@xilinx.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20200903212310.17990-1-ben.levinsky@xilinx.com> References: <20200903212310.17990-1-ben.levinsky@xilinx.com> X-RCIS-Action: ALLOW X-TM-AS-Product-Ver: IMSS-7.1.0.1224-8.2.0.1013-23620.005 X-TM-AS-User-Approved-Sender: Yes;Yes X-EOPAttributedMessage: 0 X-MS-Office365-Filtering-HT: Tenant X-MS-PublicTrafficType: Email MIME-Version: 1.0 X-MS-Office365-Filtering-Correlation-Id: 1b724122-9dbd-4017-10f6-08d8504f926e X-MS-TrafficTypeDiagnostic: BY5PR02MB6452: X-Microsoft-Antispam-PRVS: X-Auto-Response-Suppress: DR, RN, NRN, OOF, AutoReply X-MS-Oob-TLC-OOBClassifiers: OLM:565; X-MS-Exchange-SenderADCheck: 1 X-Microsoft-Antispam: BCL:0; X-Microsoft-Antispam-Message-Info: VnnanBYmtfPex2ByblmS8Ns1oI4Xuddcqqg77S/iyYZ8An38VjduR/rjARmVwIpzNxnAYhGVjVbdAnH4UzModYZgFY7CY36dJmnC8zs0jSS2o137aoDPftJEc8Z9u11IGzV3eRXEDPw5DnwNQif9Yd3Ct1LB19PcntdXXlTMzJotAneeyQktDul+QF7JkZUHttcmNqbu2cnM4PTUXTpQWFYOizZiobvts1vNUAAI4Xc2oxPvJN23v5R8oU8hnprPl5w9DD1BUlOI2KBuTVFk8qX4BRpX1uZ5ZBUcamF/JKVCn5khy5sXcCYSHCJTnC45g3QPGsWOtu7SkuAz2ZulzdZcL4PDSmWTR1ZmRXdbMwed0miJDNu82jJ46eBqs5Lc66pyzjTjrJBx+qIrH8jfiJHwD0SqtZ2PjzRYdPhd7Is= X-Forefront-Antispam-Report: CIP:149.199.60.83;CTRY:US;LANG:en;SCL:1;SRV:;IPV:NLI;SFV:NSPM;H:xsj-pvapsmtpgw01;PTR:unknown-60-83.xilinx.com;CAT:NONE;SFS:(136003)(39860400002)(346002)(376002)(396003)(46966005)(336012)(6666004)(83380400001)(7696005)(47076004)(82310400003)(2906002)(82740400003)(36756003)(81166007)(356005)(5660300002)(26005)(186003)(70586007)(2616005)(1076003)(6636002)(70206006)(44832011)(9786002)(4326008)(426003)(8676002)(8936002)(316002)(478600001)(42866002);DIR:OUT;SFP:1101; X-OriginatorOrg: xilinx.com X-MS-Exchange-CrossTenant-OriginalArrivalTime: 03 Sep 2020 21:23:16.2030 (UTC) X-MS-Exchange-CrossTenant-Network-Message-Id: 1b724122-9dbd-4017-10f6-08d8504f926e X-MS-Exchange-CrossTenant-Id: 657af505-d5df-48d0-8300-c31994686c5c X-MS-Exchange-CrossTenant-OriginalAttributedTenantConnectingIp: TenantId=657af505-d5df-48d0-8300-c31994686c5c;Ip=[149.199.60.83];Helo=[xsj-pvapsmtpgw01] X-MS-Exchange-CrossTenant-AuthSource: CY1NAM02FT052.eop-nam02.prod.protection.outlook.com X-MS-Exchange-CrossTenant-AuthAs: Anonymous X-MS-Exchange-CrossTenant-FromEntityHeader: HybridOnPrem X-MS-Exchange-Transport-CrossTenantHeadersStamped: BY5PR02MB6452 Sender: linux-remoteproc-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-remoteproc@vger.kernel.org Add ZynqMP firmware ioctl enums for RPU configuration. Signed-off-by: Ben Levinsky --- v3: - add xilinx-related platform mgmt fn's instead of wrapping around function pointer in xilinx eemi ops struct v4: - add default values for enums --- include/linux/firmware/xlnx-zynqmp.h | 19 +++++++++++++++++++ 1 file changed, 19 insertions(+) diff --git a/include/linux/firmware/xlnx-zynqmp.h b/include/linux/firmware/xlnx-zynqmp.h index 5968df82b991..bb347dfe4ba4 100644 --- a/include/linux/firmware/xlnx-zynqmp.h +++ b/include/linux/firmware/xlnx-zynqmp.h @@ -104,6 +104,10 @@ enum pm_ret_status { }; enum pm_ioctl_id { + IOCTL_GET_RPU_OPER_MODE = 0, + IOCTL_SET_RPU_OPER_MODE = 1, + IOCTL_RPU_BOOT_ADDR_CONFIG = 2, + IOCTL_TCM_COMB_CONFIG = 3, IOCTL_SD_DLL_RESET = 6, IOCTL_SET_SD_TAPDELAY, IOCTL_SET_PLL_FRAC_MODE, @@ -129,6 +133,21 @@ enum pm_query_id { PM_QID_CLOCK_GET_MAX_DIVISOR, }; +enum rpu_oper_mode { + PM_RPU_MODE_LOCKSTEP = 0, + PM_RPU_MODE_SPLIT = 1, +}; + +enum rpu_boot_mem { + PM_RPU_BOOTMEM_LOVEC = 0, + PM_RPU_BOOTMEM_HIVEC = 1, +}; + +enum rpu_tcm_comb { + PM_RPU_TCM_SPLIT = 0, + PM_RPU_TCM_COMB = 1, +}; + enum zynqmp_pm_reset_action { PM_RESET_ACTION_RELEASE, PM_RESET_ACTION_ASSERT, From patchwork Thu Sep 3 21:23:07 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Ben Levinsky X-Patchwork-Id: 11755195 Return-Path: Received: from mail.kernel.org (pdx-korg-mail-1.web.codeaurora.org [172.30.200.123]) by pdx-korg-patchwork-2.web.codeaurora.org (Postfix) with ESMTP id 55DAC13B1 for ; Thu, 3 Sep 2020 21:23:42 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 35C7A20897 for ; Thu, 3 Sep 2020 21:23:42 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=pass (1024-bit key) header.d=xilinx.onmicrosoft.com header.i=@xilinx.onmicrosoft.com header.b="RbybBsfG" Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1729167AbgICVXl (ORCPT ); Thu, 3 Sep 2020 17:23:41 -0400 Received: from mail-co1nam11on2066.outbound.protection.outlook.com ([40.107.220.66]:32481 "EHLO NAM11-CO1-obe.outbound.protection.outlook.com" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1728113AbgICVXV (ORCPT ); Thu, 3 Sep 2020 17:23:21 -0400 ARC-Seal: i=1; a=rsa-sha256; s=arcselector9901; d=microsoft.com; cv=none; b=Lt6Ce3jmKheqQCzteJVf0maCnyHBYpbmQkx0pFsYdkM5X1QUv6bhsKVdsl7Bg1FL6imyxpGk+KGwXfNBHWTw3QmFB5CL6wV2xT5XqaxrGB4TTJ5MR5MuxiTYxzlN4lC2v+ac5BWUPnzehIt4YQBJJkp76qFusNrF1qt0rGsv8+YH2xb/F8vFASP3F2/mLXPw+oYNB9+NuV6iUul7cBUYqORo8EXjIqjaeharJI5zJGttDGf5aM0Kmw12qqig6/ZobDj0cQKtJ8VfR4kb+r4aSNPQhWXPgPzw5L/MkxO5hN5fqPCudFi3E1n+B1B17VObb6xg81wR2cbqZJmilKCePg== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=microsoft.com; s=arcselector9901; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-SenderADCheck; bh=RXlvvgkgk5NVtQWRNRs7Ng0J70lamUAT04M4qSPl4yk=; b=i3cuGSvaMFO5/7HwjX1b3AIQ5NRBVehqWkNpVLmelPZlgUPU/OsjTk3bduUd65IbRri6HT5HaecGwm+B9s/DUyx2waY9rBz0j+dk3T2w/d19pA72xG8fEg9bkyohwxwnPSyGAj4467dbZ0tiXMAck8hIMLhyLo9lnuq78G0//pKp+UKIL9B+Sbg3VKTzbCPvRwD7PyJAHbaqPjlqOMSdNh6fGinlKWrIZ39/FLg1klvTwSbROPm6m4Ud7rCMmoCdBanrmnjR8OmSxAomlMG8wXosBWB51h66ilxU9ztZUWA9i/Dg8mRpjAFE+s4fg8L+J+X/CdVT1C4pOoRLGvHONg== ARC-Authentication-Results: i=1; mx.microsoft.com 1; spf=pass (sender ip is 149.199.60.83) smtp.rcpttodomain=lists.infradead.org smtp.mailfrom=xilinx.com; dmarc=bestguesspass action=none header.from=xilinx.com; dkim=none (message not signed); arc=none DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=xilinx.onmicrosoft.com; s=selector2-xilinx-onmicrosoft-com; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-SenderADCheck; bh=RXlvvgkgk5NVtQWRNRs7Ng0J70lamUAT04M4qSPl4yk=; b=RbybBsfGPfa1y8AbHU6DYKQ1ijPK1ENTnLZl918n1hSQwGZ1yvE4UPC9OareMbUSHHoxgu34BHjYb0q5yjm8K9VPqzTPkPtV5QFQX5+XIerAgQPI4yXExhjHT8OWKnjsDUlK7ww/8DpeIx5Sho5zgRGtWUkh/Ga3cgedi9MicwM= Received: from BL0PR0102CA0018.prod.exchangelabs.com (2603:10b6:207:18::31) by DM6PR02MB6235.namprd02.prod.outlook.com (2603:10b6:5:1f8::20) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.3348.15; Thu, 3 Sep 2020 21:23:17 +0000 Received: from BL2NAM02FT051.eop-nam02.prod.protection.outlook.com (2603:10b6:207:18:cafe::6) by BL0PR0102CA0018.outlook.office365.com (2603:10b6:207:18::31) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.3326.19 via Frontend Transport; Thu, 3 Sep 2020 21:23:17 +0000 X-MS-Exchange-Authentication-Results: spf=pass (sender IP is 149.199.60.83) smtp.mailfrom=xilinx.com; lists.infradead.org; dkim=none (message not signed) header.d=none;lists.infradead.org; dmarc=bestguesspass action=none header.from=xilinx.com; Received-SPF: Pass (protection.outlook.com: domain of xilinx.com designates 149.199.60.83 as permitted sender) receiver=protection.outlook.com; client-ip=149.199.60.83; helo=xsj-pvapsmtpgw01; Received: from xsj-pvapsmtpgw01 (149.199.60.83) by BL2NAM02FT051.mail.protection.outlook.com (10.152.76.181) with Microsoft SMTP Server id 15.20.3326.19 via Frontend Transport; Thu, 3 Sep 2020 21:23:16 +0000 Received: from [149.199.38.66] (port=37616 helo=smtp.xilinx.com) by xsj-pvapsmtpgw01 with esmtp (Exim 4.90) (envelope-from ) id 1kDwhJ-0006tv-O0; Thu, 03 Sep 2020 14:23:09 -0700 Received: from [127.0.0.1] (helo=localhost) by smtp.xilinx.com with smtp (Exim 4.63) (envelope-from ) id 1kDwhQ-0000OF-29; Thu, 03 Sep 2020 14:23:16 -0700 Received: from xsj-pvapsmtp01 (smtp3.xilinx.com [149.199.38.66]) by xsj-smtp-dlp2.xlnx.xilinx.com (8.13.8/8.13.1) with ESMTP id 083LNAbf023391; Thu, 3 Sep 2020 14:23:10 -0700 Received: from [172.19.2.206] (helo=xsjblevinsk50.xilinx.com) by xsj-pvapsmtp01 with esmtp (Exim 4.63) (envelope-from ) id 1kDwhK-0000M3-7a; Thu, 03 Sep 2020 14:23:10 -0700 From: Ben Levinsky To: punit1.agrawal@toshiba.co.jp, stefanos@xilinx.com, michals@xilinx.com Cc: michael.auchter@ni.com, devicetree@vger.kernel.org, mathieu.poirier@linaro.org, emooring@xilinx.com, linux-remoteproc@vger.kernel.org, linux-kernel@vger.kernel.org, jliang@xilinx.com, robh+dt@kernel.org, linux-arm-kernel@lists.infradead.org Subject: [PATCH v12 2/5] firmware: xilinx: Add shutdown/wakeup APIs Date: Thu, 3 Sep 2020 14:23:07 -0700 Message-Id: <20200903212310.17990-3-ben.levinsky@xilinx.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20200903212310.17990-1-ben.levinsky@xilinx.com> References: <20200903212310.17990-1-ben.levinsky@xilinx.com> X-RCIS-Action: ALLOW X-TM-AS-Product-Ver: IMSS-7.1.0.1224-8.2.0.1013-23620.005 X-TM-AS-User-Approved-Sender: Yes;Yes X-EOPAttributedMessage: 0 X-MS-Office365-Filtering-HT: Tenant X-MS-PublicTrafficType: Email MIME-Version: 1.0 X-MS-Office365-Filtering-Correlation-Id: 7a095b2b-44fb-4dbd-ccc0-08d8504f928a X-MS-TrafficTypeDiagnostic: DM6PR02MB6235: X-Microsoft-Antispam-PRVS: X-Auto-Response-Suppress: DR, RN, NRN, OOF, AutoReply X-MS-Oob-TLC-OOBClassifiers: OLM:4941; X-MS-Exchange-SenderADCheck: 1 X-Microsoft-Antispam: BCL:0; X-Microsoft-Antispam-Message-Info: r4WKJzHI/uNzypY4t8/kO2H/Pwqc4jiN37GBI/n6iPwfstgjmT31drrKRPeHvWfjeEIS97xqJu5jYfW92VvjqnRFRWOCrVwDicnG6sZfvkeHZ4J2sWyNIAhQFKepgJm1+xzutWbsekabCkWivLqPkpUZo8+hIRN1LVJJxdcLbdtm02q9NXQ6Nkjmm64EjoZFzRSqn/AwE1g5ZRO83nYjsPZiC2+3AoGNwnM9MJWK38mAabaKhep1Fhz37fVOzGwlT06Ve85WRS6oz5BBkg+ok6RwXEiboNEWAR9W+l5TpL/sIrFHoO7zIJiy4rHfSJfw0RICrCG0BPgqldhVUlhuqojbcgp+GGHXLpz+ab5LkPReGmaPzFWyK6UPscWrd6lVLU8xaLzx8hLKXZCdnjXh/A== X-Forefront-Antispam-Report: CIP:149.199.60.83;CTRY:US;LANG:en;SCL:1;SRV:;IPV:NLI;SFV:NSPM;H:xsj-pvapsmtpgw01;PTR:unknown-60-83.xilinx.com;CAT:NONE;SFS:(396003)(346002)(376002)(136003)(39860400002)(46966005)(8936002)(81166007)(44832011)(7696005)(6666004)(8676002)(26005)(2616005)(316002)(36756003)(5660300002)(186003)(336012)(82740400003)(6636002)(478600001)(83380400001)(82310400003)(356005)(9786002)(47076004)(4326008)(70586007)(426003)(1076003)(70206006)(2906002);DIR:OUT;SFP:1101; X-OriginatorOrg: xilinx.com X-MS-Exchange-CrossTenant-OriginalArrivalTime: 03 Sep 2020 21:23:16.4023 (UTC) X-MS-Exchange-CrossTenant-Network-Message-Id: 7a095b2b-44fb-4dbd-ccc0-08d8504f928a X-MS-Exchange-CrossTenant-Id: 657af505-d5df-48d0-8300-c31994686c5c X-MS-Exchange-CrossTenant-OriginalAttributedTenantConnectingIp: TenantId=657af505-d5df-48d0-8300-c31994686c5c;Ip=[149.199.60.83];Helo=[xsj-pvapsmtpgw01] X-MS-Exchange-CrossTenant-AuthSource: BL2NAM02FT051.eop-nam02.prod.protection.outlook.com X-MS-Exchange-CrossTenant-AuthAs: Anonymous X-MS-Exchange-CrossTenant-FromEntityHeader: HybridOnPrem X-MS-Exchange-Transport-CrossTenantHeadersStamped: DM6PR02MB6235 Sender: linux-remoteproc-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-remoteproc@vger.kernel.org Add shutdown/wakeup a resource eemi operations to shutdown or bringup a resource. Note alignment of args matches convention of other fn's in this file. The reason being that the long fn name results in aligned args that otherwise go over 80 chars so shift right to avoid this Signed-off-by: Ben Levinsky --- v3: - add xilinx-related platform mgmt fn's instead of wrapping around function pointer in xilinx eemi ops struct - fix formatting v4: - add default values for enumv3: - add xilinx-related platform mgmt fn's instead of wrapping around function pointer in xilinx eemi ops struct - fix formatting v4: - add default values for enum v9: - zynqmp_pm_force_powerdown update arg 'target' to 'node' - zynqmp_pm_request_wakeup fix code style v10: - add types.h to includes for compilation --- drivers/firmware/xilinx/zynqmp.c | 35 ++++++++++++++++++++++++++++ include/linux/firmware/xlnx-zynqmp.h | 23 ++++++++++++++++++ 2 files changed, 58 insertions(+) diff --git a/drivers/firmware/xilinx/zynqmp.c b/drivers/firmware/xilinx/zynqmp.c index 8d1ff2454e2e..a966ee956573 100644 --- a/drivers/firmware/xilinx/zynqmp.c +++ b/drivers/firmware/xilinx/zynqmp.c @@ -846,6 +846,41 @@ int zynqmp_pm_release_node(const u32 node) } EXPORT_SYMBOL_GPL(zynqmp_pm_release_node); +/** + * zynqmp_pm_force_pwrdwn - PM call to request for another PU or subsystem to + * be powered down forcefully + * @node: Node ID of the targeted PU or subsystem + * @ack: Flag to specify whether acknowledge is requested + * + * Return: status, either success or error+reason + */ +int zynqmp_pm_force_pwrdwn(const u32 node, + const enum zynqmp_pm_request_ack ack) +{ + return zynqmp_pm_invoke_fn(PM_FORCE_POWERDOWN, node, ack, 0, 0, NULL); +} +EXPORT_SYMBOL_GPL(zynqmp_pm_force_pwrdwn); + +/** + * zynqmp_pm_request_wake - PM call to wake up selected master or subsystem + * @node: Node ID of the master or subsystem + * @set_addr: Specifies whether the address argument is relevant + * @address: Address from which to resume when woken up + * @ack: Flag to specify whether acknowledge requested + * + * Return: status, either success or error+reason + */ +int zynqmp_pm_request_wake(const u32 node, + const bool set_addr, + const u64 address, + const enum zynqmp_pm_request_ack ack) +{ + /* set_addr flag is encoded into 1st bit of address */ + return zynqmp_pm_invoke_fn(PM_REQUEST_WAKEUP, node, address | set_addr, + address >> 32, ack, NULL); +} +EXPORT_SYMBOL_GPL(zynqmp_pm_request_wake); + /** * zynqmp_pm_set_requirement() - PM call to set requirement for PM slaves * @node: Node ID of the slave diff --git a/include/linux/firmware/xlnx-zynqmp.h b/include/linux/firmware/xlnx-zynqmp.h index bb347dfe4ba4..6241c5ac51b3 100644 --- a/include/linux/firmware/xlnx-zynqmp.h +++ b/include/linux/firmware/xlnx-zynqmp.h @@ -12,6 +12,7 @@ #ifndef __FIRMWARE_ZYNQMP_H__ #define __FIRMWARE_ZYNQMP_H__ +#include #define ZYNQMP_PM_VERSION_MAJOR 1 #define ZYNQMP_PM_VERSION_MINOR 0 @@ -64,6 +65,8 @@ enum pm_api_id { PM_GET_API_VERSION = 1, + PM_FORCE_POWERDOWN = 8, + PM_REQUEST_WAKEUP = 10, PM_SYSTEM_SHUTDOWN = 12, PM_REQUEST_NODE = 13, PM_RELEASE_NODE, @@ -376,6 +379,12 @@ int zynqmp_pm_write_pggs(u32 index, u32 value); int zynqmp_pm_read_pggs(u32 index, u32 *value); int zynqmp_pm_system_shutdown(const u32 type, const u32 subtype); int zynqmp_pm_set_boot_health_status(u32 value); +int zynqmp_pm_force_pwrdwn(const u32 target, + const enum zynqmp_pm_request_ack ack); +int zynqmp_pm_request_wake(const u32 node, + const bool set_addr, + const u64 address, + const enum zynqmp_pm_request_ack ack); #else static inline struct zynqmp_eemi_ops *zynqmp_pm_get_eemi_ops(void) { @@ -526,6 +535,20 @@ static inline int zynqmp_pm_set_boot_health_status(u32 value) { return -ENODEV; } + +static inline int zynqmp_pm_force_pwrdwn(const u32 target, + const enum zynqmp_pm_request_ack ack) +{ + return -ENODEV; +} + +static inline int zynqmp_pm_request_wake(const u32 node, + const bool set_addr, + const u64 address, + const enum zynqmp_pm_request_ack ack) +{ + return -ENODEV; +} #endif #endif /* __FIRMWARE_ZYNQMP_H__ */ From patchwork Thu Sep 3 21:23:08 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Ben Levinsky X-Patchwork-Id: 11755193 Return-Path: Received: from mail.kernel.org (pdx-korg-mail-1.web.codeaurora.org [172.30.200.123]) by pdx-korg-patchwork-2.web.codeaurora.org (Postfix) with ESMTP id C3ACD618 for ; Thu, 3 Sep 2020 21:23:37 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 9D0C3206EF for ; Thu, 3 Sep 2020 21:23:37 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=pass (1024-bit key) header.d=xilinx.onmicrosoft.com header.i=@xilinx.onmicrosoft.com header.b="lqRZLUsL" Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1728382AbgICVX0 (ORCPT ); Thu, 3 Sep 2020 17:23:26 -0400 Received: from mail-bn8nam12on2089.outbound.protection.outlook.com ([40.107.237.89]:36928 "EHLO NAM12-BN8-obe.outbound.protection.outlook.com" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1728037AbgICVXU (ORCPT ); Thu, 3 Sep 2020 17:23:20 -0400 ARC-Seal: i=1; a=rsa-sha256; s=arcselector9901; d=microsoft.com; cv=none; b=Bgm8HEa0GhHjW17rpcQH1MMVitaAbvV3x8SddS1AxG86judhtYPM7OePD51GwKSXLo/leqXz1Z9ssUI49273/n8n9N/uCDoiijGOoAw5Q3gduINqeN+z0m1srFfJmKurup5pZqB4vTU7xmIiW2iCxAjDUADZbkjc7jJwqIeW1MzYIMXM576EtIprw6SuPA/ZaNVTs9kUy2xyXu4s60QYlXyjw4A6v7Db/xEa+O2tK1dlkVqAwdBJw/33Oob42UER6W3jAXWJ7eYNeohiEDa8afKUTUuNwH+eO2pI5moZmONK9Qp+2LYGnU+Gs1+96OFBM8gF0S4FPOh6L3AZgy2N8g== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=microsoft.com; s=arcselector9901; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-SenderADCheck; bh=oylXQrBpZxzeyQv8gsEA43cVwV/ECuvHJODeuIS4DXU=; b=g0Jt7u9PiR3HSfo9PXPOfu20h8a15cOvwVu2XXX4kYh1dYVj5I3q/RhESu3FG1WvO0ZgCdvuZ4zEKvzIMUdkfz00DVFyAOhNoAWema8p510JBZ7ZD/AUFjcvPc3IDKchFbhpNdfleWWzxPPu84Vy0rG683KZiX97pLqlOl+SnK/ogFiO4EuAd8i4JcZln4R28RI3yUPBVFc/n51ZFsuDp3xggRWuzRuuaIgN8vOFqm5MfAgYSgYN+C/ExMBRcfnF1flPjQdG7zY6jgUG9zALtwJAGOmubVznGkkKGuRi7igc1fI/0ok1HZNT3BS8yYr6QjVkyV5su7mjsQCkeTN6VQ== ARC-Authentication-Results: i=1; mx.microsoft.com 1; spf=pass (sender ip is 149.199.60.83) smtp.rcpttodomain=lists.infradead.org smtp.mailfrom=xilinx.com; dmarc=bestguesspass action=none header.from=xilinx.com; dkim=none (message not signed); arc=none DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=xilinx.onmicrosoft.com; s=selector2-xilinx-onmicrosoft-com; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-SenderADCheck; bh=oylXQrBpZxzeyQv8gsEA43cVwV/ECuvHJODeuIS4DXU=; b=lqRZLUsL1+ghTbrRR6z/p3G77DHSWWr+BtKYqdbndXZrisHQUI5UT6QUJnPTZbxQgUc0xXLjwF7my229Dy150zBydE/vcIT/PmC1cW3TG4GGWQFqcxH59z5k5vxR9EfNXvLOo5SQynYJVIfjSal6kZAZCH8jk8EuwVNOUCr3/Uc= Received: from SN4PR0701CA0010.namprd07.prod.outlook.com (2603:10b6:803:28::20) by DM6PR02MB6010.namprd02.prod.outlook.com (2603:10b6:5:157::30) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.3348.15; Thu, 3 Sep 2020 21:23:16 +0000 Received: from SN1NAM02FT012.eop-nam02.prod.protection.outlook.com (2603:10b6:803:28:cafe::a3) by SN4PR0701CA0010.outlook.office365.com (2603:10b6:803:28::20) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.3348.16 via Frontend Transport; Thu, 3 Sep 2020 21:23:16 +0000 X-MS-Exchange-Authentication-Results: spf=pass (sender IP is 149.199.60.83) smtp.mailfrom=xilinx.com; lists.infradead.org; dkim=none (message not signed) header.d=none;lists.infradead.org; dmarc=bestguesspass action=none header.from=xilinx.com; Received-SPF: Pass (protection.outlook.com: domain of xilinx.com designates 149.199.60.83 as permitted sender) receiver=protection.outlook.com; client-ip=149.199.60.83; helo=xsj-pvapsmtpgw01; Received: from xsj-pvapsmtpgw01 (149.199.60.83) by SN1NAM02FT012.mail.protection.outlook.com (10.152.72.95) with Microsoft SMTP Server id 15.20.3326.19 via Frontend Transport; Thu, 3 Sep 2020 21:23:16 +0000 Received: from [149.199.38.66] (port=37608 helo=smtp.xilinx.com) by xsj-pvapsmtpgw01 with esmtp (Exim 4.90) (envelope-from ) id 1kDwhJ-0006to-JQ; Thu, 03 Sep 2020 14:23:09 -0700 Received: from [127.0.0.1] (helo=localhost) by smtp.xilinx.com with smtp (Exim 4.63) (envelope-from ) id 1kDwhP-0000OF-Tr; Thu, 03 Sep 2020 14:23:15 -0700 Received: from xsj-pvapsmtp01 (smtp-fallback.xilinx.com [149.199.38.66] (may be forged)) by xsj-smtp-dlp2.xlnx.xilinx.com (8.13.8/8.13.1) with ESMTP id 083LNAJm023393; Thu, 3 Sep 2020 14:23:10 -0700 Received: from [172.19.2.206] (helo=xsjblevinsk50.xilinx.com) by xsj-pvapsmtp01 with esmtp (Exim 4.63) (envelope-from ) id 1kDwhK-0000M3-9I; Thu, 03 Sep 2020 14:23:10 -0700 From: Ben Levinsky To: punit1.agrawal@toshiba.co.jp, stefanos@xilinx.com, michals@xilinx.com Cc: michael.auchter@ni.com, devicetree@vger.kernel.org, mathieu.poirier@linaro.org, emooring@xilinx.com, linux-remoteproc@vger.kernel.org, linux-kernel@vger.kernel.org, jliang@xilinx.com, robh+dt@kernel.org, linux-arm-kernel@lists.infradead.org Subject: [PATCH v12 3/5] firmware: xilinx: Add RPU configuration APIs Date: Thu, 3 Sep 2020 14:23:08 -0700 Message-Id: <20200903212310.17990-4-ben.levinsky@xilinx.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20200903212310.17990-1-ben.levinsky@xilinx.com> References: <20200903212310.17990-1-ben.levinsky@xilinx.com> X-RCIS-Action: ALLOW X-TM-AS-Product-Ver: IMSS-7.1.0.1224-8.2.0.1013-23620.005 X-TM-AS-User-Approved-Sender: Yes;Yes X-EOPAttributedMessage: 0 X-MS-Office365-Filtering-HT: Tenant X-MS-PublicTrafficType: Email MIME-Version: 1.0 X-MS-Office365-Filtering-Correlation-Id: 8af75dfe-be2e-491b-2981-08d8504f926e X-MS-TrafficTypeDiagnostic: DM6PR02MB6010: X-Microsoft-Antispam-PRVS: X-Auto-Response-Suppress: DR, RN, NRN, OOF, AutoReply X-MS-Oob-TLC-OOBClassifiers: OLM:179; X-MS-Exchange-SenderADCheck: 1 X-Microsoft-Antispam: BCL:0; X-Microsoft-Antispam-Message-Info: zy4EETx+HPb0AecanH5IY6i8M1w1SqJfjABimj6r7t+uL0Dl4c5gMciF9Q7TMfdILu0USjwMB62s+5g68WF1TnIVcenLkE6vsJgqX44JFUS09ID0Prw9xYVk+90B2nD4SqWJOSYhjX0wJI7qCh6rXaWFraAcLavKXH2Yn9LDMfRw//ZRSOC+G2G+6GX7/BQ3btVwbbkz5TMxje+mmqmadK+ZAbFGDfVe7k6nLJvu5zFaubEYA2CaHnI3vt2b6ovrPtQZ5zLwtqySaXOz5Dcr1s+zhvChCQ4eDOzdna4KuSqRofZ37GXRIGCJT5Lx6L4TuD401U7w9hBB9/VZHO3C+b4b2t4ZzuOo6YuJROBA283PZ4fbY6ZtaGHqY2JPElrAEEpHWmaivvFtdz1+MaajnkIoHCvBwekPsWu6n8MPd+c= X-Forefront-Antispam-Report: CIP:149.199.60.83;CTRY:US;LANG:en;SCL:1;SRV:;IPV:NLI;SFV:NSPM;H:xsj-pvapsmtpgw01;PTR:unknown-60-83.xilinx.com;CAT:NONE;SFS:(39860400002)(376002)(346002)(136003)(396003)(46966005)(82740400003)(44832011)(356005)(8676002)(81166007)(4326008)(186003)(478600001)(7696005)(83380400001)(26005)(1076003)(70206006)(2906002)(70586007)(336012)(6636002)(316002)(9786002)(2616005)(8936002)(82310400003)(5660300002)(47076004)(36756003)(426003)(42866002);DIR:OUT;SFP:1101; X-OriginatorOrg: xilinx.com X-MS-Exchange-CrossTenant-OriginalArrivalTime: 03 Sep 2020 21:23:16.2146 (UTC) X-MS-Exchange-CrossTenant-Network-Message-Id: 8af75dfe-be2e-491b-2981-08d8504f926e X-MS-Exchange-CrossTenant-Id: 657af505-d5df-48d0-8300-c31994686c5c X-MS-Exchange-CrossTenant-OriginalAttributedTenantConnectingIp: TenantId=657af505-d5df-48d0-8300-c31994686c5c;Ip=[149.199.60.83];Helo=[xsj-pvapsmtpgw01] X-MS-Exchange-CrossTenant-AuthSource: SN1NAM02FT012.eop-nam02.prod.protection.outlook.com X-MS-Exchange-CrossTenant-AuthAs: Anonymous X-MS-Exchange-CrossTenant-FromEntityHeader: HybridOnPrem X-MS-Exchange-Transport-CrossTenantHeadersStamped: DM6PR02MB6010 Sender: linux-remoteproc-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-remoteproc@vger.kernel.org This patch adds APIs to access to configure RPU and its processor-specific memory. That is query the run-time mode of RPU as either split or lockstep as well as API to set this mode. In addition add APIs to access configuration of the RPUs' tightly coupled memory (TCM). Signed-off-by: Ben Levinsky --- v3: - add xilinx-related platform mgmt fn's instead of wrapping around function pointer in xilinx eemi ops struct v4: - add default values for enums v9: - update commit message - for zynqmp_pm_set_tcm_config and zynqmp_pm_get_rpu_mode update docs for expected output, arguments as well removing unused args - remove unused fn zynqmp_pm_get_node_status v11: - update usage of zynqmp_pm_get_rpu_mode to return rpu mode in enum - update zynqmp_pm_set_tcm_config and zynqmp_pm_set_rpu_mode arguments to remove unused args v12: - in drivers/firmware/zynqmp.c, update zynqmp_pm_set_rpu_mode so rpu_mode is only set if no error - update args for zynqmp_pm_set_rpu_mode, zynqmp_pm_set_tcm_config fn arg's to reflect what is expected in the function and the usage in zynqmp_r5_remoteproc accordingly --- drivers/firmware/xilinx/zynqmp.c | 60 ++++++++++++++++++++++++++++ include/linux/firmware/xlnx-zynqmp.h | 18 +++++++++ 2 files changed, 78 insertions(+) diff --git a/drivers/firmware/xilinx/zynqmp.c b/drivers/firmware/xilinx/zynqmp.c index a966ee956573..916a0b15ab33 100644 --- a/drivers/firmware/xilinx/zynqmp.c +++ b/drivers/firmware/xilinx/zynqmp.c @@ -846,6 +846,66 @@ int zynqmp_pm_release_node(const u32 node) } EXPORT_SYMBOL_GPL(zynqmp_pm_release_node); +/** + * zynqmp_pm_get_rpu_mode() - Get RPU mode + * @node_id: Node ID of the device + * @rpu_mode: return by reference value + * either split or lockstep + * + * Return: return 0 on success or error+reason. + * if success, then rpu_mode will be set + * to current rpu mode. + */ +int zynqmp_pm_get_rpu_mode(u32 node_id, enum rpu_oper_mode *rpu_mode) +{ + u32 ret_payload[PAYLOAD_ARG_CNT]; + int ret; + + ret = zynqmp_pm_invoke_fn(PM_IOCTL, node_id, + IOCTL_GET_RPU_OPER_MODE, 0, 0, ret_payload); + + /* only set rpu_mode if no error */ + *rpu_mode = ret_payload[0]; + + return ret; +} +EXPORT_SYMBOL_GPL(zynqmp_pm_get_rpu_mode); + +/** + * zynqmp_pm_set_rpu_mode() - Set RPU mode + * @node_id: Node ID of the device + * @rpu_mode: Argument 1 to requested IOCTL call. either split or lockstep + * + * This function is used to set RPU mode to split or + * lockstep + * + * Return: Returns status, either success or error+reason + */ +int zynqmp_pm_set_rpu_mode(u32 node_id, enum rpu_oper_mode rpu_mode) +{ + return zynqmp_pm_invoke_fn(PM_IOCTL, node_id, + IOCTL_SET_RPU_OPER_MODE, (u32)rpu_mode, + 0, NULL); +} +EXPORT_SYMBOL_GPL(zynqmp_pm_set_rpu_mode); + +/** + * zynqmp_pm_set_tcm_config - configure TCM + * @tcm_mode: Argument 1 to requested IOCTL call + * either PM_RPU_TCM_COMB or PM_RPU_TCM_SPLIT + * + * This function is used to set RPU mode to split or combined + * + * Return: status: 0 for success, else failure + */ +int zynqmp_pm_set_tcm_config(u32 node_id, enum rpu_tcm_comb tcm_mode) +{ + return zynqmp_pm_invoke_fn(PM_IOCTL, node_id, + IOCTL_TCM_COMB_CONFIG, (u32)tcm_mode, 0, + NULL); +} +EXPORT_SYMBOL_GPL(zynqmp_pm_set_tcm_config); + /** * zynqmp_pm_force_pwrdwn - PM call to request for another PU or subsystem to * be powered down forcefully diff --git a/include/linux/firmware/xlnx-zynqmp.h b/include/linux/firmware/xlnx-zynqmp.h index 6241c5ac51b3..79aa2fcbcd54 100644 --- a/include/linux/firmware/xlnx-zynqmp.h +++ b/include/linux/firmware/xlnx-zynqmp.h @@ -385,6 +385,9 @@ int zynqmp_pm_request_wake(const u32 node, const bool set_addr, const u64 address, const enum zynqmp_pm_request_ack ack); +int zynqmp_pm_get_rpu_mode(u32 node_id, enum rpu_oper_mode *rpu_mode); +int zynqmp_pm_set_rpu_mode(u32 node_id, u32 arg1); +int zynqmp_pm_set_tcm_config(u32 node_id, u32 arg1); #else static inline struct zynqmp_eemi_ops *zynqmp_pm_get_eemi_ops(void) { @@ -549,6 +552,21 @@ static inline int zynqmp_pm_request_wake(const u32 node, { return -ENODEV; } + +static inline int zynqmp_pm_get_rpu_mode(u32 node_id, enum rpu_oper_mode *rpu_mode) +{ + return -ENODEV; +} + +static inline int zynqmp_pm_set_rpu_mode(u32 node_id, u32 arg1) +{ + return -ENODEV; +} + +static inline int zynqmp_pm_set_tcm_config(u32 node_id, u32 arg1) +{ + return -ENODEV; +} #endif #endif /* __FIRMWARE_ZYNQMP_H__ */ From patchwork Thu Sep 3 21:23:09 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Ben Levinsky X-Patchwork-Id: 11755189 Return-Path: Received: from mail.kernel.org (pdx-korg-mail-1.web.codeaurora.org [172.30.200.123]) by pdx-korg-patchwork-2.web.codeaurora.org (Postfix) with ESMTP id 34A4D618 for ; Thu, 3 Sep 2020 21:23:27 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 144C8206D4 for ; Thu, 3 Sep 2020 21:23:27 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=pass (1024-bit key) header.d=xilinx.onmicrosoft.com header.i=@xilinx.onmicrosoft.com header.b="YIkcOBwq" Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1728134AbgICVXZ (ORCPT ); Thu, 3 Sep 2020 17:23:25 -0400 Received: from mail-bn8nam12on2041.outbound.protection.outlook.com ([40.107.237.41]:2912 "EHLO NAM12-BN8-obe.outbound.protection.outlook.com" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1726323AbgICVXU (ORCPT ); Thu, 3 Sep 2020 17:23:20 -0400 ARC-Seal: i=1; a=rsa-sha256; s=arcselector9901; d=microsoft.com; cv=none; b=eWYwWV1NKj2fb6ETynVYYUoV2lQox4G0f9Se/Lr5cVMhp3yM/FHTBbLUX+sDIzM23E4ro1bm1asyRUgGkiIN9knRmxRBNHr4ThmKb5i0MNZ0rTVtjTfAIQNcdOSiLx+uuoSstqHMD3MjDssYVdNu+hEraG34saleyIeEiu22o0T/sAtlN2XAhKgwzRgV01f30u7WT9ZKW4kpSyNnXCSUL8cZ28mpPJ/KvVQcxeDFctgkLX5reUYTmt4B++YJciHMWsORad1mToJgllXs4T1y2TAqbYQWlBprkZQF6BTbWsGr5k4a7U3iCGQjjI9OouEeVDH5zIbkkzPZXRm7Ii80zA== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=microsoft.com; s=arcselector9901; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-SenderADCheck; bh=WBhzK6/+3qzl2TIEmlP/0lBGAMg4z3Yvf69FMHri2gI=; b=MH9chjd3YQbqbytmvC06cI+gVjlBoJdo47Usaqd7vqZJkYcL4bV/mKA8mgt7PIqE1ik9tP8zlfzlWSRiLiyGVJq9xTb8N8ltzn1EevqfSyIChppXZFoyzjlMY2S6qp7PoIHtIOFI7Px5fY5V+fKK8hrHQHixqJcUciRGyV7NbRmf/ZBr/5AgmFn+S0TA08utr1VL9GiJby4u64XF2abCMY+hVGqLbt8EahMGQtZ9QEEugbrveZwCYIRr0OLWXM3zfuNS7T7VR/Yyuy1fx/FyC/+QCqaPuWpYLhBdV00YftcT531kXce3TTNczPdApYX4Cpy4PxS/4x3vI1ZPMbaObw== ARC-Authentication-Results: i=1; mx.microsoft.com 1; spf=pass (sender ip is 149.199.60.83) smtp.rcpttodomain=lists.infradead.org smtp.mailfrom=xilinx.com; dmarc=bestguesspass action=none header.from=xilinx.com; dkim=none (message not signed); arc=none DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=xilinx.onmicrosoft.com; s=selector2-xilinx-onmicrosoft-com; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-SenderADCheck; bh=WBhzK6/+3qzl2TIEmlP/0lBGAMg4z3Yvf69FMHri2gI=; b=YIkcOBwqdeTu01WHVugUuRQNWDsZtGrprIvES/uIAjBUePkPcMvY87jCHBQ53FJ51njraOSoFkz0Oz0cd7KlL/7gQVXVGsR0hDnQcIAPyjkCcqkS4TvfDA3TpzrTuchYWthcRxXyNwH95adwq/HkzteHiZBqSvuUmMKOdtfGaoI= Received: from CY4PR03CA0006.namprd03.prod.outlook.com (2603:10b6:903:33::16) by BL0PR02MB4707.namprd02.prod.outlook.com (2603:10b6:208:55::13) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.3326.19; Thu, 3 Sep 2020 21:23:16 +0000 Received: from CY1NAM02FT011.eop-nam02.prod.protection.outlook.com (2603:10b6:903:33:cafe::eb) by CY4PR03CA0006.outlook.office365.com (2603:10b6:903:33::16) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.3348.15 via Frontend Transport; Thu, 3 Sep 2020 21:23:16 +0000 X-MS-Exchange-Authentication-Results: spf=pass (sender IP is 149.199.60.83) smtp.mailfrom=xilinx.com; lists.infradead.org; dkim=none (message not signed) header.d=none;lists.infradead.org; dmarc=bestguesspass action=none header.from=xilinx.com; Received-SPF: Pass (protection.outlook.com: domain of xilinx.com designates 149.199.60.83 as permitted sender) receiver=protection.outlook.com; client-ip=149.199.60.83; helo=xsj-pvapsmtpgw01; Received: from xsj-pvapsmtpgw01 (149.199.60.83) by CY1NAM02FT011.mail.protection.outlook.com (10.152.75.156) with Microsoft SMTP Server id 15.20.3326.19 via Frontend Transport; Thu, 3 Sep 2020 21:23:15 +0000 Received: from [149.199.38.66] (port=37580 helo=smtp.xilinx.com) by xsj-pvapsmtpgw01 with esmtp (Exim 4.90) (envelope-from ) id 1kDwhJ-0006tf-8t; Thu, 03 Sep 2020 14:23:09 -0700 Received: from [127.0.0.1] (helo=localhost) by smtp.xilinx.com with smtp (Exim 4.63) (envelope-from ) id 1kDwhP-0000OF-JK; Thu, 03 Sep 2020 14:23:15 -0700 Received: from xsj-pvapsmtp01 (xsj-smtp1.xilinx.com [149.199.38.66]) by xsj-smtp-dlp2.xlnx.xilinx.com (8.13.8/8.13.1) with ESMTP id 083LNAeD023395; Thu, 3 Sep 2020 14:23:10 -0700 Received: from [172.19.2.206] (helo=xsjblevinsk50.xilinx.com) by xsj-pvapsmtp01 with esmtp (Exim 4.63) (envelope-from ) id 1kDwhK-0000M3-B4; Thu, 03 Sep 2020 14:23:10 -0700 From: Ben Levinsky To: punit1.agrawal@toshiba.co.jp, stefanos@xilinx.com, michals@xilinx.com Cc: michael.auchter@ni.com, devicetree@vger.kernel.org, mathieu.poirier@linaro.org, emooring@xilinx.com, linux-remoteproc@vger.kernel.org, linux-kernel@vger.kernel.org, jliang@xilinx.com, robh+dt@kernel.org, linux-arm-kernel@lists.infradead.org, Jason Wu , Michal Simek Subject: [PATCH v12 4/5] dt-bindings: remoteproc: Add documentation for ZynqMP R5 rproc bindings Date: Thu, 3 Sep 2020 14:23:09 -0700 Message-Id: <20200903212310.17990-5-ben.levinsky@xilinx.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20200903212310.17990-1-ben.levinsky@xilinx.com> References: <20200903212310.17990-1-ben.levinsky@xilinx.com> X-RCIS-Action: ALLOW X-TM-AS-Product-Ver: IMSS-7.1.0.1224-8.2.0.1013-23620.005 X-TM-AS-User-Approved-Sender: Yes;Yes X-EOPAttributedMessage: 0 X-MS-Office365-Filtering-HT: Tenant X-MS-PublicTrafficType: Email MIME-Version: 1.0 X-MS-Office365-Filtering-Correlation-Id: b69e53d9-bb46-4954-ac1a-08d8504f9233 X-MS-TrafficTypeDiagnostic: BL0PR02MB4707: X-Microsoft-Antispam-PRVS: X-Auto-Response-Suppress: DR, RN, NRN, OOF, AutoReply X-MS-Oob-TLC-OOBClassifiers: OLM:9508; X-MS-Exchange-SenderADCheck: 1 X-Microsoft-Antispam: BCL:0; X-Microsoft-Antispam-Message-Info: B5rXSQ3iAMmPvSBOqgcT0iP3tgwnpu768A1sAIsw/JBvG6K8jq7R6wQQJYJJXM0hz5CinuqpE3BBWigGT+Ekb+FKxHc9JYSSNRjrchntnEuWMWWKmAgJ82pZNc1AsPZfDzr6BD9W2gEGqpU/HUZxIcCmtEG0SkzT4K5CIdEzLHTqd2oTzltwHAUj5c4DHG/nc7HHPzarjW/UtCzakvX1D2wznumihsaVHdVerPo++TrdEu+v15guZJ29kPi/8F6IlLqVsdH991A3Ah2qme88QhQ+XJG8cAOAJqmgplai1Q0yaY5h4q0KkbRr9Zoo+GiW+v66yXr0JWDbSqVXEcMMeE2mgRqKHoZqA2y1vTbqNmW0L1aJX68z6w5xjacvq7QuYXLHbVpEQxIF9ZFUCbw2f8jZwt0mQyu3kDJgPlytmpJVwxZu0bW7r9ND1rK6zeDZ4s7TKoCsN95TZk0npuWxuaXcGnIbN5lnvUL7yMHI23k= X-Forefront-Antispam-Report: CIP:149.199.60.83;CTRY:US;LANG:en;SCL:1;SRV:;IPV:NLI;SFV:NSPM;H:xsj-pvapsmtpgw01;PTR:unknown-60-83.xilinx.com;CAT:NONE;SFS:(136003)(39860400002)(376002)(346002)(396003)(46966005)(44832011)(5660300002)(426003)(26005)(6636002)(336012)(82310400003)(356005)(54906003)(81166007)(316002)(7696005)(47076004)(1076003)(4326008)(8936002)(478600001)(186003)(70206006)(9786002)(83380400001)(36756003)(8676002)(82740400003)(2616005)(70586007)(107886003)(2906002);DIR:OUT;SFP:1101; X-OriginatorOrg: xilinx.com X-MS-Exchange-CrossTenant-OriginalArrivalTime: 03 Sep 2020 21:23:15.8323 (UTC) X-MS-Exchange-CrossTenant-Network-Message-Id: b69e53d9-bb46-4954-ac1a-08d8504f9233 X-MS-Exchange-CrossTenant-Id: 657af505-d5df-48d0-8300-c31994686c5c X-MS-Exchange-CrossTenant-OriginalAttributedTenantConnectingIp: TenantId=657af505-d5df-48d0-8300-c31994686c5c;Ip=[149.199.60.83];Helo=[xsj-pvapsmtpgw01] X-MS-Exchange-CrossTenant-AuthSource: CY1NAM02FT011.eop-nam02.prod.protection.outlook.com X-MS-Exchange-CrossTenant-AuthAs: Anonymous X-MS-Exchange-CrossTenant-FromEntityHeader: HybridOnPrem X-MS-Exchange-Transport-CrossTenantHeadersStamped: BL0PR02MB4707 Sender: linux-remoteproc-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-remoteproc@vger.kernel.org Add binding for ZynqMP R5 OpenAMP. Represent the RPU domain resources in one device node. Each RPU processor is a subnode of the top RPU domain node. Signed-off-by: Jason Wu Signed-off-by: Wendy Liang Signed-off-by: Michal Simek Signed-off-by: Ben Levinsky --- v3: - update zynqmp_r5 yaml parsing to not raise warnings for extra information in children of R5 node. The warning "node has a unit name, but no reg or ranges property" will still be raised though as this particular node is needed to describe the '#address-cells' and '#size-cells' information. v4:: - remove warning '/example-0/rpu@ff9a0000/r5@0: node has a unit name, but no reg or ranges property' by adding reg to r5 node. v5: - update device tree sample and yaml parsing to not raise any warnings - description for memory-region in yaml parsing - compatible string in yaml parsing for TCM v6: - remove coupling TCM nodes with remoteproc - remove mailbox as it is optional not needed v7: - change lockstep-mode to xlnx,cluster-mode v9: - show example IPC nodes and tcm bank nodes v11: - add property meta-memory-regions to illustrate link between r5 and TCM banks - update so no warnings from 'make dt_binding_check' --- .../xilinx,zynqmp-r5-remoteproc.yaml | 123 ++++++++++++++++++ 1 file changed, 123 insertions(+) create mode 100644 Documentation/devicetree/bindings/remoteproc/xilinx,zynqmp-r5-remoteproc.yaml diff --git a/Documentation/devicetree/bindings/remoteproc/xilinx,zynqmp-r5-remoteproc.yaml b/Documentation/devicetree/bindings/remoteproc/xilinx,zynqmp-r5-remoteproc.yaml new file mode 100644 index 000000000000..148374c61908 --- /dev/null +++ b/Documentation/devicetree/bindings/remoteproc/xilinx,zynqmp-r5-remoteproc.yaml @@ -0,0 +1,123 @@ +# SPDX-License-Identifier: (GPL-2.0 OR BSD-2-Clause) +%YAML 1.2 +--- +$id: "http://devicetree.org/schemas/remoteproc/xilinx,zynqmp-r5-remoteproc.yaml#" +$schema: "http://devicetree.org/meta-schemas/core.yaml#" + +title: Xilinx R5 remote processor controller bindings + +description: + This document defines the binding for the remoteproc component that loads and + boots firmwares on the Xilinx Zynqmp and Versal family chipset. + + Note that the Linux has global addressing view of the R5-related memory (TCM) + so the absolute address ranges are provided in TCM reg's. +maintainers: + - Ed Mooring + - Ben Levinsky + +properties: + compatible: + const: "xlnx,zynqmp-r5-remoteproc-1.0" + + lockstep-mode: + description: + R5 core configuration (split is 0 or lock-step and 1) + maxItems: 1 + + interrupts: + description: + Interrupt mapping for remoteproc IPI. It is required if the + user uses the remoteproc driver with the RPMsg kernel driver. + maxItems: 6 + + memory-region: + description: + collection of memory carveouts used for elf-loading and inter-processor + communication. + maxItems: 4 + minItems: 4 + meta-memory-regions: + description: + collection of memories that are not present in the top level memory + nodes' mapping. For example, R5s' TCM banks. These banks are needed + for R5 firmware meta data such as the R5 firmware's heap and stack + pnode-id: + maxItems: 1 + mboxes: + maxItems: 2 + mbox-names: + maxItems: 2 + +examples: + - | + reserved-memory { + #address-cells = <1>; + #size-cells = <1>; + ranges; + elf_load: rproc@3ed000000 { + no-map; + reg = <0x3ed00000 0x40000>; + }; + }; + rpu { + compatible = "xlnx,zynqmp-r5-remoteproc-1.0"; + #address-cells = <1>; + #size-cells = <1>; + ranges; + lockstep-mode = <0>; + r5_0 { + ranges; + #address-cells = <1>; + #size-cells = <1>; + memory-region = <&elf_load>; + meta-memory-regions = <0x40 0x41>; + pnode-id = <0x7>; + }; + }; + + /* + * Below nodes are required if using TCM to load R5 firmware + * if not, then either do not provide nodes are label as disabled in + * status property + */ + tcm_0a@ffe00000 { + reg = <0xffe00000 0x10000>; + pnode-id = <0xf>; + no-map; + status = "okay"; + phandle = <0x40>; + compatible = "xlnx,tcm"; + }; + tcm_1a@ffe20000 { + reg = <0xffe20000 0x10000>; + pnode-id = <0x10>; + no-map; + status = "okay"; + compatible = "xlnx,tcm"; + phandle = <0x41>; + }; + + /* + * Below nodes are required for IPC, as Xilinx Kernel + * can potentially have other applications that use CMA + * and conflict from hardware design. With that in mind, + * explicitly state the address spaces for the IPC remoteproc + * mechanisms. + * + * If IPC not required and only elf-loading then these are not needed. + */ + rpu0vdev0vring0: rpu0vdev0vring0@3ed40000 { + no-map; + reg = <0x3ed40000 0x4000>; + }; + rpu0vdev0vring1: rpu0vdev0vring1@3ed44000 { + no-map; + reg = <0x3ed44000 0x4000>; + }; + rpu0vdev0buffer: rpu0vdev0buffer@3ed48000 { + no-map; + reg = <0x3ed48000 0x100000>; + }; + +... From patchwork Thu Sep 3 21:23:10 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Ben Levinsky X-Patchwork-Id: 11755213 Return-Path: Received: from mail.kernel.org (pdx-korg-mail-1.web.codeaurora.org [172.30.200.123]) by pdx-korg-patchwork-2.web.codeaurora.org (Postfix) with ESMTP id 7670991F for ; Thu, 3 Sep 2020 21:25:02 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by mail.kernel.org (Postfix) with ESMTP id 3F15B206D4 for ; Thu, 3 Sep 2020 21:25:02 +0000 (UTC) Authentication-Results: mail.kernel.org; dkim=pass (1024-bit key) header.d=xilinx.onmicrosoft.com header.i=@xilinx.onmicrosoft.com header.b="e1t3ZJ6D" Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1726397AbgICVZB (ORCPT ); Thu, 3 Sep 2020 17:25:01 -0400 Received: from mail-dm6nam10on2059.outbound.protection.outlook.com ([40.107.93.59]:38881 "EHLO NAM10-DM6-obe.outbound.protection.outlook.com" rhost-flags-OK-OK-OK-FAIL) by vger.kernel.org with ESMTP id S1728127AbgICVZA (ORCPT ); Thu, 3 Sep 2020 17:25:00 -0400 ARC-Seal: i=1; a=rsa-sha256; s=arcselector9901; d=microsoft.com; cv=none; b=KPG6bt9Rx/3P4S5ygU2AhWIb6Q4iPTM7OE32e2Q1ffkBZqmF05dRaev9mJbKE8at1hd0T9sW9q7LZXkiv2w2wWp4U9ZvC1d05K04M0p/H9YAe2r6eAXHeS6sL13Em6yANn4k/mbfMDBaVX2uwjV0JCeiTVbVTsV39PgJ90NmZw4/JJ5L1cZJI/q171KpxtrtxnW36H0QUQqwSnlZOvT13kocJT5fpX0j9UMMSYpmkKv0JuJTGNfVqG9UkhaGb1wiA6kkbzVxS8RoPVPEdI8c0MjGnBSP5fbzzrv2cU84T33VkZHLNUz42g3rjHCE2RGJ3UFiMykZWm3C3W+2ixGidQ== ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=microsoft.com; s=arcselector9901; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-SenderADCheck; bh=wUVQeHHF7zjIM9DrnTbJW+4bmWW974Kp6QV03qZ445U=; b=feRdOHn1wG5w4jZaD3AziQ93nXn6sR/X14wdswe50lzlTDREXmt2ZJ72/H0uV0EpQP0Lf3/QMuWBm4hh0xOvFZOTugg0yTMzjSk55NOYb+2mWXsmWulkBSko0XxW+2u7iRtQH1ZjL0obxXainYp5UWmGOBxQJQMAauBWodoMLTWQaPt6iCHF4B0AKS6G0TVNB2D6K64uzk4a8f7btrT3bhQO501WsoO7bdTw6nDoCaaG2rzndItHwMiuNkXVcAZs6YzlCZaIldWi8nhec7BP03NofKLPfJ72TAMQba+KiaEQrrkkXhDKU27sdAece7yfZJaBr0JuF+y8LPDizt96Zw== ARC-Authentication-Results: i=1; mx.microsoft.com 1; spf=pass (sender ip is 149.199.60.83) smtp.rcpttodomain=lists.infradead.org smtp.mailfrom=xilinx.com; dmarc=bestguesspass action=none header.from=xilinx.com; dkim=none (message not signed); arc=none DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=xilinx.onmicrosoft.com; s=selector2-xilinx-onmicrosoft-com; h=From:Date:Subject:Message-ID:Content-Type:MIME-Version:X-MS-Exchange-SenderADCheck; bh=wUVQeHHF7zjIM9DrnTbJW+4bmWW974Kp6QV03qZ445U=; b=e1t3ZJ6DtxaHel4SEtaNgrYBvtn73wVP+J+tAXG7JD84wK1lg2NWRL3kkGXZDseAVg3Tp86AbnUh1M9u6wDXkkCYH4hS0jS2dEcxuKyiirB+3AlJhYN8LeFwQ+Qa3kZ9H9jZAs+w74FLiNWnBGYLr2uMlX+Jeay6wDS88gR9kFg= Received: from BL0PR0102CA0053.prod.exchangelabs.com (2603:10b6:208:25::30) by DM6PR02MB6156.namprd02.prod.outlook.com (2603:10b6:5:1f9::15) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.3348.15; Thu, 3 Sep 2020 21:23:17 +0000 Received: from BL2NAM02FT042.eop-nam02.prod.protection.outlook.com (2603:10b6:208:25:cafe::5d) by BL0PR0102CA0053.outlook.office365.com (2603:10b6:208:25::30) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.20.3348.15 via Frontend Transport; Thu, 3 Sep 2020 21:23:16 +0000 X-MS-Exchange-Authentication-Results: spf=pass (sender IP is 149.199.60.83) smtp.mailfrom=xilinx.com; lists.infradead.org; dkim=none (message not signed) header.d=none;lists.infradead.org; dmarc=bestguesspass action=none header.from=xilinx.com; Received-SPF: Pass (protection.outlook.com: domain of xilinx.com designates 149.199.60.83 as permitted sender) receiver=protection.outlook.com; client-ip=149.199.60.83; helo=xsj-pvapsmtpgw01; Received: from xsj-pvapsmtpgw01 (149.199.60.83) by BL2NAM02FT042.mail.protection.outlook.com (10.152.76.193) with Microsoft SMTP Server id 15.20.3326.19 via Frontend Transport; Thu, 3 Sep 2020 21:23:15 +0000 Received: from [149.199.38.66] (port=37582 helo=smtp.xilinx.com) by xsj-pvapsmtpgw01 with esmtp (Exim 4.90) (envelope-from ) id 1kDwhJ-0006ti-AZ; Thu, 03 Sep 2020 14:23:09 -0700 Received: from [127.0.0.1] (helo=localhost) by smtp.xilinx.com with smtp (Exim 4.63) (envelope-from ) id 1kDwhP-0000OF-Ks; Thu, 03 Sep 2020 14:23:15 -0700 Received: from xsj-pvapsmtp01 (smtp-fallback.xilinx.com [149.199.38.66] (may be forged)) by xsj-smtp-dlp2.xlnx.xilinx.com (8.13.8/8.13.1) with ESMTP id 083LNAKE023397; Thu, 3 Sep 2020 14:23:10 -0700 Received: from [172.19.2.206] (helo=xsjblevinsk50.xilinx.com) by xsj-pvapsmtp01 with esmtp (Exim 4.63) (envelope-from ) id 1kDwhK-0000M3-Cu; Thu, 03 Sep 2020 14:23:10 -0700 From: Ben Levinsky To: punit1.agrawal@toshiba.co.jp, stefanos@xilinx.com, michals@xilinx.com Cc: michael.auchter@ni.com, devicetree@vger.kernel.org, mathieu.poirier@linaro.org, emooring@xilinx.com, linux-remoteproc@vger.kernel.org, linux-kernel@vger.kernel.org, jliang@xilinx.com, robh+dt@kernel.org, linux-arm-kernel@lists.infradead.org, Wendy Liang , Michal Simek , Ed Mooring , Jason Wu Subject: [PATCH v12 5/5] remoteproc: Add initial zynqmp R5 remoteproc driver Date: Thu, 3 Sep 2020 14:23:10 -0700 Message-Id: <20200903212310.17990-6-ben.levinsky@xilinx.com> X-Mailer: git-send-email 2.17.1 In-Reply-To: <20200903212310.17990-1-ben.levinsky@xilinx.com> References: <20200903212310.17990-1-ben.levinsky@xilinx.com> X-RCIS-Action: ALLOW X-TM-AS-Product-Ver: IMSS-7.1.0.1224-8.2.0.1013-23620.005 X-TM-AS-User-Approved-Sender: Yes;Yes X-EOPAttributedMessage: 0 X-MS-Office365-Filtering-HT: Tenant X-MS-PublicTrafficType: Email MIME-Version: 1.0 X-MS-Office365-Filtering-Correlation-Id: b994f4e6-a026-44d8-0b6f-08d8504f924d X-MS-TrafficTypeDiagnostic: DM6PR02MB6156: X-Microsoft-Antispam-PRVS: X-Auto-Response-Suppress: DR, RN, NRN, OOF, AutoReply X-MS-Oob-TLC-OOBClassifiers: OLM:59; X-MS-Exchange-SenderADCheck: 1 X-Microsoft-Antispam: BCL:0; X-Microsoft-Antispam-Message-Info: Q+HXrYZWp0l0/Gwoz+SFzoQLzVOivpI3498LTKlRmxuD3Cc6UCh6Fu8hrrgeQW5Y/rsZACZR7Dn4Er+5TOuWXvFSONQxJKms+3qZ4Xu3b2d9CAVeiN2h/iERfN8C+fQRDxp+HkvKhNxbomonGIC7X/2ldKcD/74HyaqMhvPO3CkHLZpNtHvykfzxnVUu+j8x+WhtFJAOzqa56wChQxFMiITHKEGmuIT3n/6VK1sOO0mkg3HKdGgUpGOfgwrqNPSPOwYLiG4Ryc62ZD9udHgOlmRzqN0mVFg5lYj+ld8YwAkKWHpaTxqsXy+J8Vs3SF979Alz2J1Fgd9Su/+426W+X0qE7j4AupEqz4dm+5KI3dZFoGe/pwKL1aZ/bAbeip/AW2e4Fo5GQ4xAHYuZREGKYPAbMtIcpLw3oPwsaWFuyP8= X-Forefront-Antispam-Report: CIP:149.199.60.83;CTRY:US;LANG:en;SCL:1;SRV:;IPV:NLI;SFV:NSPM;H:xsj-pvapsmtpgw01;PTR:unknown-60-83.xilinx.com;CAT:NONE;SFS:(396003)(376002)(39860400002)(136003)(346002)(46966005)(1076003)(186003)(2906002)(4326008)(9786002)(5660300002)(30864003)(8676002)(8936002)(70586007)(6636002)(44832011)(26005)(426003)(70206006)(107886003)(316002)(336012)(81166007)(82310400003)(54906003)(356005)(2616005)(47076004)(36756003)(82740400003)(478600001)(83380400001)(7696005)(42866002);DIR:OUT;SFP:1101; X-OriginatorOrg: xilinx.com X-MS-Exchange-CrossTenant-OriginalArrivalTime: 03 Sep 2020 21:23:15.9981 (UTC) X-MS-Exchange-CrossTenant-Network-Message-Id: b994f4e6-a026-44d8-0b6f-08d8504f924d X-MS-Exchange-CrossTenant-Id: 657af505-d5df-48d0-8300-c31994686c5c X-MS-Exchange-CrossTenant-OriginalAttributedTenantConnectingIp: TenantId=657af505-d5df-48d0-8300-c31994686c5c;Ip=[149.199.60.83];Helo=[xsj-pvapsmtpgw01] X-MS-Exchange-CrossTenant-AuthSource: BL2NAM02FT042.eop-nam02.prod.protection.outlook.com X-MS-Exchange-CrossTenant-AuthAs: Anonymous X-MS-Exchange-CrossTenant-FromEntityHeader: HybridOnPrem X-MS-Exchange-Transport-CrossTenantHeadersStamped: DM6PR02MB6156 Sender: linux-remoteproc-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-remoteproc@vger.kernel.org R5 is included in Xilinx Zynq UltraScale MPSoC so by adding this remotproc driver, we can boot the R5 sub-system in different 2 configurations: split or lock-step. The Xilinx R5 Remoteproc Driver boots the R5's via calls to the Xilinx Platform Management Unit that handles the R5 configuration, memory access and R5 lifecycle management. The interface to this manager is done in this driver via zynqmp_pm_* function calls. Signed-off-by: Wendy Liang Signed-off-by: Michal Simek Signed-off-by: Ed Mooring Signed-off-by: Jason Wu Signed-off-by: Ben Levinsky --- v2: - remove domain struct as per review from Mathieu v3: - add xilinx-related platform mgmt fn's instead of wrapping around function pointer in xilinx eemi ops struct v4: - add default values for enums - fix formatting as per checkpatch.pl --strict. Note that 1 warning and 1 check are still raised as each is due to fixing the warning results in that particular line going over 80 characters. v5: - parse_fw change from use of rproc_of_resm_mem_entry_init to rproc_mem_entry_init and use of alloc/release - var's of type zynqmp_r5_pdata all have same local variable name - use dev_dbg instead of dev_info v6: - adding memory carveouts is handled much more similarly. All mem carveouts are now described in reserved memory as needed. That is, TCM nodes are not coupled to remoteproc anymore. This is reflected in the remoteproc R5 driver and the device tree binding. - remove mailbox from device tree binding as it is not necessary for elf loading - use lockstep-mode property for configuring RPU v7: - remove unused headers - change u32 *lockstep_mode -> u32 lockstep_mode; - change device-tree binding "lockstep-mode" to xlnx,cluster-mode - remove zynqmp_r5_mem_probe and loop to Probe R5 memory devices at remoteproc-probe time - remove is_r5_mode_set from zynqmp rpu remote processor private data - do not error out if no mailbox is provided - remove zynqmp_r5_remoteproc_probe call of platform_set_drvdata as pdata is handled in zynqmp_r5_remoteproc_remove v8: - remove old acks, reviewed-by's in commit message v9: - as mboxes are now optional, if pdata->tx_mc_skbs not initialized then do not call skb_queue_empty - update usage for zynqmp_pm_set_rpu_mode, zynqmp_pm_set_tcm_config and zynqmp_pm_get_rpu_mode - update 5/5 patch commit message to document supported configurations and how they are booted by the driver. - remove copyrights other than SPDX from zynqmp_r5_remoteproc.c - compilation warnings no longer raised - remove unused includes from zynqmp_r5_remoteproc.c - remove unused var autoboot from zynqmp_r5_remoteproc.c - reorder zynqmp_r5_pdata fpr small mem savings due to alignment - use of zynqmp_pm_set_tcm_config now does not have output arg - in tcm handling, unconditionally use &= 0x000fffff mask since all nodes in this fn are for tcm - update comments for translating dma field in tcm handling to device address - update calls to rproc_mem_entry_init in parse_mem_regions so that there are only 2 cases for types of carveouts instead of 3 - in parse_mem_regions, check if device tree node is null before using it - add example device tree nodes used in parse_mem_regions and tcm parsing - add comment for vring id node length - add check for string length so that vring id is at least min length - move tcm nodes from reserved mem to instead own device tree nodes and only use them if enabled in device tree - add comment for explaining handling of rproc_elf_load_rsc_table - remove obsolete check for "if (vqid < 0)" in zynqmp_r5_rproc_kick - remove unused field mems in struct zynqmp_r5_pdata - remove call to zynqmp_r5_mem_probe and the fn itself as tcm handling is done by zyqmp_r5_pm_request_tcm - remove obsolete setting of dma_ops and parent device dma_mask - remove obsolete use of of_dma_configure - add comment for call to r5_set_mode fn - make mbox usage optional and gracefully inform user via dev_dbg if not present - change var lockstep_mode from u32* to u32 v11: - use enums instead of u32 where possible in zynqmp_r5_remoteproc - update usage of zynqmp_pm_set/get_rpu_mode and zynqmp_pm_set_tcm_config - update prints to not use carriage return, just newline - look up tcm banks via property in r5 node instead of string name - print device tree nodes with %pOF instead of %s with node name field - update tcm release to unmap VA - handle r5-1 use case v12: - update signed off by so that latest developer name is last - do not cast enums to u32s for zynqmp_pm* functions --- drivers/remoteproc/Kconfig | 10 + drivers/remoteproc/Makefile | 1 + drivers/remoteproc/zynqmp_r5_remoteproc.c | 898 ++++++++++++++++++++++ 3 files changed, 909 insertions(+) create mode 100644 drivers/remoteproc/zynqmp_r5_remoteproc.c diff --git a/drivers/remoteproc/Kconfig b/drivers/remoteproc/Kconfig index c4d1731295eb..dd9ed45654e0 100644 --- a/drivers/remoteproc/Kconfig +++ b/drivers/remoteproc/Kconfig @@ -249,6 +249,16 @@ config STM32_RPROC This can be either built-in or a loadable module. +config ZYNQMP_R5_REMOTEPROC + tristate "ZynqMP_R5 remoteproc support" + depends on ARM64 && PM && ARCH_ZYNQMP + select RPMSG_VIRTIO + select MAILBOX + select ZYNQMP_IPI_MBOX + help + Say y or m here to support ZynqMP R5 remote processors via the remote + processor framework. + endif # REMOTEPROC endmenu diff --git a/drivers/remoteproc/Makefile b/drivers/remoteproc/Makefile index e8b886e511f0..04d1c95d06d7 100644 --- a/drivers/remoteproc/Makefile +++ b/drivers/remoteproc/Makefile @@ -28,5 +28,6 @@ obj-$(CONFIG_QCOM_WCNSS_PIL) += qcom_wcnss_pil.o qcom_wcnss_pil-y += qcom_wcnss.o qcom_wcnss_pil-y += qcom_wcnss_iris.o obj-$(CONFIG_ST_REMOTEPROC) += st_remoteproc.o +obj-$(CONFIG_ZYNQMP_R5_REMOTEPROC) += zynqmp_r5_remoteproc.o obj-$(CONFIG_ST_SLIM_REMOTEPROC) += st_slim_rproc.o obj-$(CONFIG_STM32_RPROC) += stm32_rproc.o diff --git a/drivers/remoteproc/zynqmp_r5_remoteproc.c b/drivers/remoteproc/zynqmp_r5_remoteproc.c new file mode 100644 index 000000000000..4fc4098ae1ea --- /dev/null +++ b/drivers/remoteproc/zynqmp_r5_remoteproc.c @@ -0,0 +1,898 @@ +// SPDX-License-Identifier: GPL-2.0 +/* + * Zynq R5 Remote Processor driver + * + * Based on origin OMAP and Zynq Remote Processor driver + * + */ + +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include +#include + +#include "remoteproc_internal.h" + +#define MAX_RPROCS 2 /* Support up to 2 RPU */ +#define MAX_MEM_PNODES 4 /* Max power nodes for one RPU memory instance */ + +#define DEFAULT_FIRMWARE_NAME "rproc-rpu-fw" +#define BANK_LIST_PROP "meta-memory-regions" + +/* PM proc states */ +#define PM_PROC_STATE_ACTIVE 1U + +/* IPI buffer MAX length */ +#define IPI_BUF_LEN_MAX 32U +/* RX mailbox client buffer max length */ +#define RX_MBOX_CLIENT_BUF_MAX (IPI_BUF_LEN_MAX + \ + sizeof(struct zynqmp_ipi_message)) + +/** + * struct zynqmp_r5_mem - zynqmp rpu memory data + * @pnode_id: TCM power domain ids + * @res: memory resource + * @node: list node + */ +struct zynqmp_r5_mem { + u32 pnode_id[MAX_MEM_PNODES]; + struct resource res; + struct list_head node; +}; + +/** + * struct zynqmp_r5_pdata - zynqmp rpu remote processor private data + * @dev: device of RPU instance + * @rproc: rproc handle + * @pnode_id: RPU CPU power domain id + * @is_r5_mode_set: indicate if r5 operation mode is set + * @tx_mc: tx mailbox client + * @rx_mc: rx mailbox client + * @tx_chan: tx mailbox channel + * @rx_chan: rx mailbox channel + * @mbox_work: mbox_work for the RPU remoteproc + * @tx_mc_skbs: socket buffers for tx mailbox client + * @rx_mc_buf: rx mailbox client buffer to save the rx message + */ +struct zynqmp_r5_pdata { + unsigned char rx_mc_buf[RX_MBOX_CLIENT_BUF_MAX]; + struct mbox_client tx_mc; + struct mbox_client rx_mc; + struct work_struct mbox_work; + struct sk_buff_head tx_mc_skbs; + struct device dev; + struct rproc *rproc; + struct mbox_chan *tx_chan; + struct mbox_chan *rx_chan; + u32 pnode_id; + bool is_r5_mode_set; +}; + +/** + * table of RPUs + */ +struct zynqmp_r5_pdata rpus[MAX_RPROCS]; +/** + * RPU core configuration + */ +enum rpu_oper_mode rpu_mode; + +/* + * r5_set_mode - set RPU operation mode + * @pdata: Remote processor private data + * + * set RPU oepration mode + * + * Return: 0 for success, negative value for failure + */ +static int r5_set_mode(struct zynqmp_r5_pdata *pdata) +{ + struct device *dev = &pdata->dev; + enum rpu_tcm_comb cur_tcm_mode; + enum rpu_oper_mode cur_rpu_mode; + int ret; + + cur_rpu_mode = rpu_mode; + ret = zynqmp_pm_get_rpu_mode(pdata->pnode_id, &cur_rpu_mode); + if (ret < 0) { + dev_err(dev, "failed to get RPU oper mode.\n"); + return ret; + } + if (rpu_mode == cur_rpu_mode) { + dev_dbg(dev, "RPU mode matches: %x\n", rpu_mode); + } else { + ret = zynqmp_pm_set_rpu_mode(pdata->pnode_id, + cur_rpu_mode); + if (ret < 0) { + dev_err(dev, + "failed to set RPU oper mode.\n"); + return ret; + } + } + + cur_tcm_mode = (cur_rpu_mode == PM_RPU_MODE_LOCKSTEP) ? + PM_RPU_TCM_COMB : PM_RPU_TCM_SPLIT; + ret = zynqmp_pm_set_tcm_config(pdata->pnode_id, cur_tcm_mode); + if (ret < 0) { + dev_err(dev, "failed to config TCM to %x.\n", + cur_rpu_mode); + return ret; + } + pdata->is_r5_mode_set = true; + return 0; +} + +/* + * ZynqMP R5 remoteproc memory release function + */ +static int tcm_mem_release(struct rproc *rproc, struct rproc_mem_entry *mem) +{ + u32 pnode_id; + int ret; + struct device *dev = &rproc->dev; + + pnode_id = *(u32 *)mem->priv; + + if (pnode_id <= 0) + dev_err(dev, "rproc_mem_entry has null priv field\n"); + + iounmap(mem->va); + + ret = zynqmp_pm_release_node(pnode_id); + if (ret < 0) + dev_err(dev, "failed to release power node: %u\n", pnode_id); + + return ret; +} + +/* + * ZynqMP R5 remoteproc operations + */ +static int zynqmp_r5_rproc_start(struct rproc *rproc) +{ + struct device *dev = rproc->dev.parent; + struct zynqmp_r5_pdata *pdata = rproc->priv; + enum rpu_boot_mem bootmem; + int ret; + + if ((rproc->bootaddr & 0xF0000000) == 0xF0000000) + bootmem = PM_RPU_BOOTMEM_HIVEC; + else + bootmem = PM_RPU_BOOTMEM_LOVEC; + dev_dbg(dev, "RPU boot from %s.", + bootmem == PM_RPU_BOOTMEM_HIVEC ? "OCM" : "TCM"); + ret = zynqmp_pm_request_wake(pdata->pnode_id, 1, + bootmem, ZYNQMP_PM_REQUEST_ACK_NO); + if (ret < 0) { + dev_err(dev, "failed to boot R5.\n"); + return ret; + } + return 0; +} + +static int zynqmp_r5_rproc_stop(struct rproc *rproc) +{ + struct zynqmp_r5_pdata *pdata = rproc->priv; + int ret; + + ret = zynqmp_pm_force_pwrdwn(pdata->pnode_id, + ZYNQMP_PM_REQUEST_ACK_BLOCKING); + if (ret < 0) { + dev_err(&pdata->dev, "failed to shutdown R5.\n"); + return ret; + } + return 0; +} + +static int zynqmp_r5_rproc_mem_alloc(struct rproc *rproc, + struct rproc_mem_entry *mem) +{ + struct device *dev = rproc->dev.parent; + void *va; + + dev_dbg(rproc->dev.parent, "map memory: %pa\n", &mem->dma); + va = ioremap_wc(mem->dma, mem->len); + if (IS_ERR_OR_NULL(va)) { + dev_err(dev, "Unable to map memory region: %pa+%lx\n", + &mem->dma, mem->len); + return -ENOMEM; + } + + /* Update memory entry va */ + mem->va = va; + + return 0; +} + +static int zynqmp_r5_rproc_mem_release(struct rproc *rproc, + struct rproc_mem_entry *mem) +{ + dev_dbg(rproc->dev.parent, "unmap memory: %pa\n", &mem->dma); + iounmap(mem->va); + + return 0; +} + +static int parse_mem_regions(struct rproc *rproc) +{ + int num_mems, i; + struct zynqmp_r5_pdata *pdata = rproc->priv; + struct device *dev = &pdata->dev; + struct device_node *np = dev->of_node; + struct rproc_mem_entry *mem; + + num_mems = of_count_phandle_with_args(np, "memory-region", NULL); + if (num_mems <= 0) + return 0; + for (i = 0; i < num_mems; i++) { + struct device_node *node; + struct reserved_mem *rmem; + + node = of_parse_phandle(np, "memory-region", i); + if (!node) { + dev_err(dev, "no memory-region specified\n"); + return -EINVAL; + } + + rmem = of_reserved_mem_lookup(node); + if (!rmem) { + dev_err(dev, "unable to acquire memory-region\n"); + return -EINVAL; + } + + if (strstr(node->name, "vdev0vring")) { + int vring_id; + char name[16]; + + /* + * expecting form of "rpuXvdev0vringX as documented + * in xilinx remoteproc device tree binding + */ + if (strlen(node->name) < 14) { + dev_err(dev, "%pOF is less than 14 chars", + node); + return -EINVAL; + } + + /* + * can be 1 of multiple vring IDs per IPC channel + * e.g. 'vdev0vring0' and 'vdev0vring1' + */ + vring_id = node->name[14] - '0'; + snprintf(name, sizeof(name), "vdev0vring%d", vring_id); + /* Register vring */ + mem = rproc_mem_entry_init(dev, NULL, + (dma_addr_t)rmem->base, + rmem->size, rmem->base, + zynqmp_r5_rproc_mem_alloc, + zynqmp_r5_rproc_mem_release, + name); + dev_dbg(dev, "parsed %s at %llx\n", mem->name, + mem->dma); + } else { + /* Register DMA region */ + int (*alloc)(struct rproc *r, + struct rproc_mem_entry *rme); + int (*release)(struct rproc *r, + struct rproc_mem_entry *rme); + char name[20]; + + if (strstr(node->name, "vdev0buffer")) { + alloc = NULL; + release = NULL; + strcpy(name, "vdev0buffer"); + } else { + alloc = zynqmp_r5_rproc_mem_alloc; + release = zynqmp_r5_rproc_mem_release; + strcpy(name, node->name); + } + + mem = rproc_mem_entry_init(dev, NULL, + (dma_addr_t)rmem->base, + rmem->size, rmem->base, + alloc, release, name); + if (!mem) { + dev_err(dev, + "unable to init memory-region %pOF\n", + node); + return -ENOMEM; + } + dev_dbg(dev, "parsed %s at %llx\n", mem->name, + mem->dma); + } + rproc_add_carveout(rproc, mem); + } + + return 0; +} + +/* call Xilix Platform manager to request access to TCM bank */ +static int zyqmp_r5_pm_request_tcm(struct device_node *tcm_node, + struct device *dev, + u32 *pnode_id) +{ + int ret; + + ret = of_property_read_u32(tcm_node, "pnode-id", pnode_id); + if (ret) { + dev_err(dev, "Can't parse pnode-id property\n"); + return ret; + } + + ret = zynqmp_pm_request_node(*pnode_id, ZYNQMP_PM_CAPABILITY_ACCESS, 0, + ZYNQMP_PM_REQUEST_ACK_BLOCKING); + if (ret < 0) { + dev_err(dev, "failed to request power node: %u\n", *pnode_id); + return ret; + } + + return 0; +} + +/* Given tcm bank entry, + * this callback will set device address for R5 running on TCM + * and also setup virtual address for tcm bank remoteproc carveout + */ +static int tcm_mem_alloc(struct rproc *rproc, + struct rproc_mem_entry *mem) +{ + void *va; + struct device *dev = rproc->dev.parent; + + dev_dbg(rproc->dev.parent, "map memory: %pa\n", &mem->dma); + va = ioremap_wc(mem->dma, mem->len); + if (IS_ERR_OR_NULL(va)) { + dev_err(dev, "Unable to map memory region: %pa+%lx\n", + &mem->dma, mem->len); + return -ENOMEM; + } + + /* Update memory entry va */ + mem->va = va; + + va = devm_ioremap_wc(dev, mem->da, mem->len); + if (!va) + return -ENOMEM; + /* As R5 is 32 bit, wipe out extra high bits */ + mem->da &= 0x000fffff; + /* + * handle tcm banks 1 a and b (0xffe90000 and oxffeb0000) + * As both of these the only common bit found not in tcm bank0 a or b + * is at 0x80000 use this mask to suss it out + */ + if (mem->da & 0x80000) + /* + * similar to wiping tcm banks 0a and 0b with just the + * mask of 0x000fffff will translate to device addr's + * at 0x0 and 0x20000, need to do more to further translate + * tcm banks 1a and 1b at 0xffe90000 and oxffeb0000 + * respectively to 0x0 and 0x20000 + */ + mem->da -= 0x90000; + + return 0; +} + +/* + * Given R5 node in remoteproc instance, + * allocate remoteproc carveout for TCM memory + * needed for firmware to be loaded + */ +static int parse_tcm_banks(struct rproc *rproc) +{ + int i, num_banks; + + struct zynqmp_r5_pdata *pdata = rproc->priv; + struct device *dev = &pdata->dev; + struct device_node *r5_node = dev->of_node; + + /* go through tcm banks for r5 node */ + num_banks = of_count_phandle_with_args(r5_node, BANK_LIST_PROP, NULL); + if (num_banks <= 0) { + dev_err(dev, "need to specify TCM banks\n"); + return -EINVAL; + } + for (i = 0; i < num_banks; i++) { + struct resource rsc; + resource_size_t size; + struct device_node *dt_node; + struct rproc_mem_entry *mem; + int ret; + /* + * use u32 as zynqmp_pm* functions use this type + * + * use ptr because it needs be used later on when releasing + * each tcm node in tcm_mem_release + */ + u32 *pnode_id_ptr; + + dt_node = of_parse_phandle(r5_node, BANK_LIST_PROP, i); + if (!dt_node) { + dev_err(dev, "no tcm banks specified\n"); + return -EINVAL; + } + + if (of_device_is_available(dt_node)) { + dev_dbg(dev, "%pOF is available\n", dt_node); + + /* get address */ + ret = of_address_to_resource(dt_node, 0, &rsc); + if (ret < 0) { + dev_err(dev, "fail to get rsc mem %pOF", + dt_node); + return -EINVAL; + } + + pnode_id_ptr = devm_kzalloc(dev, sizeof(*pdata), + GFP_KERNEL); + if (!pnode_id_ptr) + return -ENOMEM; + + /* request access */ + zyqmp_r5_pm_request_tcm(dt_node, dev, pnode_id_ptr); + + /* add carveout */ + size = resource_size(&rsc); + mem = rproc_mem_entry_init(dev, NULL, rsc.start, + (int)size, rsc.start, + tcm_mem_alloc, + tcm_mem_release, + rsc.name); + if (!mem) + return -ENOMEM; + mem->priv = (void *)pnode_id_ptr; + + rproc_add_carveout(rproc, mem); + } else { + dev_info(dev, "node %pOF is not available\n", dt_node); + } + } + + return 0; +} + +static int zynqmp_r5_parse_fw(struct rproc *rproc, const struct firmware *fw) +{ + int ret; + struct zynqmp_r5_pdata *pdata = rproc->priv; + struct device *dev = &pdata->dev; + + ret = parse_tcm_banks(rproc); + if (ret) { + dev_err(dev, "parse_tcm_banks failed %x\n", ret); + return ret; + } + + ret = parse_mem_regions(rproc); + if (ret) { + dev_err(dev, "parse_mem_regions failed %x\n", ret); + return ret; + } + + ret = rproc_elf_load_rsc_table(rproc, fw); + if (ret == -EINVAL) { + /* + * resource table only required for IPC. + * if not present, this is not necessarily an error; + * for example, loading r5 hello world application + * so simply inform user and keep going. + */ + dev_info(dev, "no resource table found.\n"); + ret = 0; + } + return ret; +} + +/* kick a firmware */ +static void zynqmp_r5_rproc_kick(struct rproc *rproc, int vqid) +{ + struct sk_buff *skb; + unsigned int skb_len; + struct zynqmp_ipi_message *mb_msg; + int ret; + + struct device *dev = rproc->dev.parent; + struct zynqmp_r5_pdata *pdata = rproc->priv; + + dev_dbg(dev, "KICK Firmware to start send messages vqid %d\n", vqid); + + skb_len = (unsigned int)(sizeof(vqid) + sizeof(mb_msg)); + skb = alloc_skb(skb_len, GFP_ATOMIC); + if (!skb) { + dev_err(dev, + "Failed to allocate skb to kick remote.\n"); + return; + } + mb_msg = (struct zynqmp_ipi_message *)skb_put(skb, skb_len); + mb_msg->len = sizeof(vqid); + memcpy(mb_msg->data, &vqid, sizeof(vqid)); + skb_queue_tail(&pdata->tx_mc_skbs, skb); + ret = mbox_send_message(pdata->tx_chan, mb_msg); + if (ret < 0) { + dev_warn(dev, "Failed to kick remote.\n"); + skb_dequeue_tail(&pdata->tx_mc_skbs); + kfree_skb(skb); + } +} + +static struct rproc_ops zynqmp_r5_rproc_ops = { + .start = zynqmp_r5_rproc_start, + .stop = zynqmp_r5_rproc_stop, + .load = rproc_elf_load_segments, + .parse_fw = zynqmp_r5_parse_fw, + .find_loaded_rsc_table = rproc_elf_find_loaded_rsc_table, + .sanity_check = rproc_elf_sanity_check, + .get_boot_addr = rproc_elf_get_boot_addr, + .kick = zynqmp_r5_rproc_kick, +}; + +/** + * zynqmp_r5_release() - ZynqMP R5 device release function + * @dev: pointer to the device struct of ZynqMP R5 + * + * Function to release ZynqMP R5 device. + */ +static void zynqmp_r5_release(struct device *dev) +{ + struct zynqmp_r5_pdata *pdata; + struct rproc *rproc; + struct sk_buff *skb; + + pdata = dev_get_drvdata(dev); + rproc = pdata->rproc; + if (rproc) { + rproc_del(rproc); + rproc_free(rproc); + } + if (pdata->tx_chan) + mbox_free_channel(pdata->tx_chan); + if (pdata->rx_chan) + mbox_free_channel(pdata->rx_chan); + + /* Discard all SKBs if tx_mc_skbs is initialized */ + if (&pdata->tx_mc_skbs.prev) { + while (!skb_queue_empty(&pdata->tx_mc_skbs)) { + skb = skb_dequeue(&pdata->tx_mc_skbs); + kfree_skb(skb); + } + } + + put_device(dev->parent); +} + +/** + * event_notified_idr_cb() - event notified idr callback + * @id: idr id + * @ptr: pointer to idr private data + * @data: data passed to idr_for_each callback + * + * Pass notification to remoteproc virtio + * + * Return: 0. having return is to satisfy the idr_for_each() function + * pointer input argument requirement. + **/ +static int event_notified_idr_cb(int id, void *ptr, void *data) +{ + struct rproc *rproc = data; + + (void)rproc_vq_interrupt(rproc, id); + return 0; +} + +/** + * handle_event_notified() - remoteproc notification work funciton + * @work: pointer to the work structure + * + * It checks each registered remoteproc notify IDs. + */ +static void handle_event_notified(struct work_struct *work) +{ + struct rproc *rproc; + struct zynqmp_r5_pdata *pdata; + + pdata = container_of(work, struct zynqmp_r5_pdata, mbox_work); + + (void)mbox_send_message(pdata->rx_chan, NULL); + rproc = pdata->rproc; + /* + * We only use IPI for interrupt. The firmware side may or may + * not write the notifyid when it trigger IPI. + * And thus, we scan through all the registered notifyids. + */ + idr_for_each(&rproc->notifyids, event_notified_idr_cb, rproc); +} + +/** + * zynqmp_r5_mb_rx_cb() - Receive channel mailbox callback + * @cl: mailbox client + * @mssg: message pointer + * + * It will schedule the R5 notification work. + */ +static void zynqmp_r5_mb_rx_cb(struct mbox_client *cl, void *mssg) +{ + struct zynqmp_r5_pdata *pdata; + + pdata = container_of(cl, struct zynqmp_r5_pdata, rx_mc); + if (mssg) { + struct zynqmp_ipi_message *ipi_msg, *buf_msg; + size_t len; + + ipi_msg = (struct zynqmp_ipi_message *)mssg; + buf_msg = (struct zynqmp_ipi_message *)pdata->rx_mc_buf; + len = (ipi_msg->len >= IPI_BUF_LEN_MAX) ? + IPI_BUF_LEN_MAX : ipi_msg->len; + buf_msg->len = len; + memcpy(buf_msg->data, ipi_msg->data, len); + } + schedule_work(&pdata->mbox_work); +} + +/** + * zynqmp_r5_mb_tx_done() - Request has been sent to the remote + * @cl: mailbox client + * @mssg: pointer to the message which has been sent + * @r: status of last TX - OK or error + * + * It will be called by the mailbox framework when the last TX has done. + */ +static void zynqmp_r5_mb_tx_done(struct mbox_client *cl, void *mssg, int r) +{ + struct zynqmp_r5_pdata *pdata; + struct sk_buff *skb; + + if (!mssg) + return; + pdata = container_of(cl, struct zynqmp_r5_pdata, tx_mc); + skb = skb_dequeue(&pdata->tx_mc_skbs); + kfree_skb(skb); +} + +/** + * zynqmp_r5_setup_mbox() - Setup mailboxes + * + * @pdata: pointer to the ZynqMP R5 processor platform data + * @node: pointer of the device node + * + * Function to setup mailboxes to talk to RPU. + * + * Return: 0 for success, negative value for failure. + */ +static int zynqmp_r5_setup_mbox(struct zynqmp_r5_pdata *pdata, + struct device_node *node) +{ + struct device *dev = &pdata->dev; + struct mbox_client *mclient; + + /* Setup TX mailbox channel client */ + mclient = &pdata->tx_mc; + mclient->dev = dev; + mclient->rx_callback = NULL; + mclient->tx_block = false; + mclient->knows_txdone = false; + mclient->tx_done = zynqmp_r5_mb_tx_done; + + /* Setup TX mailbox channel client */ + mclient = &pdata->rx_mc; + mclient->dev = dev; + mclient->rx_callback = zynqmp_r5_mb_rx_cb; + mclient->tx_block = false; + mclient->knows_txdone = false; + + INIT_WORK(&pdata->mbox_work, handle_event_notified); + + /* Request TX and RX channels */ + pdata->tx_chan = mbox_request_channel_byname(&pdata->tx_mc, "tx"); + if (IS_ERR(pdata->tx_chan)) { + dev_err(dev, "failed to request mbox tx channel.\n"); + pdata->tx_chan = NULL; + return -EINVAL; + } + pdata->rx_chan = mbox_request_channel_byname(&pdata->rx_mc, "rx"); + if (IS_ERR(pdata->rx_chan)) { + dev_err(dev, "failed to request mbox rx channel.\n"); + pdata->rx_chan = NULL; + return -EINVAL; + } + skb_queue_head_init(&pdata->tx_mc_skbs); + + return 0; +} + +/** + * zynqmp_r5_probe() - Probes ZynqMP R5 processor device node + * @pdata: pointer to the ZynqMP R5 processor platform data + * @pdev: parent RPU domain platform device + * @node: pointer of the device node + * + * Function to retrieve the information of the ZynqMP R5 device node. + * + * Return: 0 for success, negative value for failure. + */ +static int zynqmp_r5_probe(struct zynqmp_r5_pdata *pdata, + struct platform_device *pdev, + struct device_node *node) +{ + struct device *dev = &pdata->dev; + struct rproc *rproc; + int ret; + + /* Create device for ZynqMP R5 device */ + dev->parent = &pdev->dev; + dev->release = zynqmp_r5_release; + dev->of_node = node; + dev_set_name(dev, "%pOF", node); + dev_set_drvdata(dev, pdata); + ret = device_register(dev); + if (ret) { + dev_err(dev, "failed to register device.\n"); + return ret; + } + get_device(&pdev->dev); + + /* Allocate remoteproc instance */ + rproc = rproc_alloc(dev, dev_name(dev), &zynqmp_r5_rproc_ops, NULL, 0); + if (!rproc) { + dev_err(dev, "rproc allocation failed.\n"); + ret = -ENOMEM; + goto error; + } + pdata->rproc = rproc; + rproc->priv = pdata; + + /* Set up DMA mask */ + ret = dma_set_coherent_mask(dev, DMA_BIT_MASK(32)); + if (ret) { + dev_err(&pdev->dev, "dma_set_coherent_mask: %d\n", ret); + return ret; + } + + /* Get R5 power domain node */ + ret = of_property_read_u32(node, "pnode-id", &pdata->pnode_id); + if (ret) { + dev_err(dev, "failed to get power node id.\n"); + goto error; + } + + /* TODO Check if R5 is running */ + + /* + * Set up R5 if not already setup + * This check is needed as there are cases where + * a user might repeatedly do modprobe + * and modprobe -r. In this case, upon a + * subsequent load, this check is needed. + */ + ret = pdata->is_r5_mode_set ? 0 : r5_set_mode(pdata); + if (ret) { + dev_err(dev, "failed to set R5 operation mode.\n"); + return ret; + } + + if (!of_get_property(dev->of_node, "mboxes", NULL)) { + dev_dbg(dev, "no mailboxes.\n"); + } else { + ret = zynqmp_r5_setup_mbox(pdata, node); + if (ret < 0) + goto error; + } + + /* Add R5 remoteproc */ + ret = rproc_add(rproc); + if (ret) { + dev_err(dev, "rproc registration failed\n"); + goto error; + } + return 0; +error: + if (pdata->rproc) + rproc_free(pdata->rproc); + pdata->rproc = NULL; + device_unregister(dev); + put_device(&pdev->dev); + return ret; +} + +static int zynqmp_r5_remoteproc_probe(struct platform_device *pdev) +{ + int ret, i = 0; + u32 lockstep_mode; + struct device *dev = &pdev->dev; + struct device_node *nc; + struct zynqmp_r5_pdata *pdata; + + pdata = devm_kzalloc(dev, sizeof(*pdata), GFP_KERNEL); + if (!pdata) + return -ENOMEM; + + platform_set_drvdata(pdev, pdata); + + of_property_read_u32(dev->of_node, "lockstep-mode", &lockstep_mode); + + if (!lockstep_mode) { + rpu_mode = PM_RPU_MODE_SPLIT; + } else if (lockstep_mode == 1) { + rpu_mode = PM_RPU_MODE_LOCKSTEP; + } else { + dev_err(dev, + "Invalid lockstep-mode mode %x in %pOF\n", + lockstep_mode, dev->of_node); + return -EINVAL; + } + dev_dbg(dev, "RPU configuration: %s\n", + lockstep_mode ? "lockstep" : "split"); + + for_each_available_child_of_node(dev->of_node, nc) { + ret = zynqmp_r5_probe(&rpus[i], pdev, nc); + if (ret) { + dev_err(dev, "failed to probe rpu %pOF.\n", nc); + return ret; + } + i++; + } + + return 0; +} + +static int zynqmp_r5_remoteproc_remove(struct platform_device *pdev) +{ + int i; + + for (i = 0; i < MAX_RPROCS; i++) { + struct zynqmp_r5_pdata *pdata = &rpus[i]; + struct rproc *rproc; + + /* only do clean up for pdata with active rpu */ + if (pdata->pnode_id == 0) + continue; + + rproc = pdata->rproc; + if (rproc) { + rproc_del(rproc); + rproc_free(rproc); + pdata->rproc = NULL; + } + if (pdata->tx_chan) { + mbox_free_channel(pdata->tx_chan); + pdata->tx_chan = NULL; + } + if (pdata->rx_chan) { + mbox_free_channel(pdata->rx_chan); + pdata->rx_chan = NULL; + } + if (&(&pdata->dev)->dma_pools) + device_unregister(&pdata->dev); + } + + return 0; +} + +/* Match table for OF platform binding */ +static const struct of_device_id zynqmp_r5_remoteproc_match[] = { + { .compatible = "xlnx,zynqmp-r5-remoteproc-1.0", }, + { /* end of list */ }, +}; +MODULE_DEVICE_TABLE(of, zynqmp_r5_remoteproc_match); + +static struct platform_driver zynqmp_r5_remoteproc_driver = { + .probe = zynqmp_r5_remoteproc_probe, + .remove = zynqmp_r5_remoteproc_remove, + .driver = { + .name = "zynqmp_r5_remoteproc", + .of_match_table = zynqmp_r5_remoteproc_match, + }, +}; +module_platform_driver(zynqmp_r5_remoteproc_driver); + +MODULE_AUTHOR("Ben Levinsky "); +MODULE_LICENSE("GPL v2");