From patchwork Mon Apr 4 08:30:07 2016 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Eric Auger X-Patchwork-Id: 8738591 Return-Path: X-Original-To: patchwork-linux-arm@patchwork.kernel.org Delivered-To: patchwork-parsemail@patchwork1.web.kernel.org Received: from mail.kernel.org (mail.kernel.org [198.145.29.136]) by patchwork1.web.kernel.org (Postfix) with ESMTP id A57BA9F36E for ; Mon, 4 Apr 2016 08:32:47 +0000 (UTC) Received: from mail.kernel.org (localhost [127.0.0.1]) by mail.kernel.org (Postfix) with ESMTP id D6505201B4 for ; Mon, 4 Apr 2016 08:32:45 +0000 (UTC) Received: from bombadil.infradead.org (bombadil.infradead.org [198.137.202.9]) (using TLSv1.2 with cipher AES128-GCM-SHA256 (128/128 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPS id 0AAA62013A for ; Mon, 4 Apr 2016 08:32:45 +0000 (UTC) Received: from localhost ([127.0.0.1] helo=bombadil.infradead.org) by bombadil.infradead.org with esmtp (Exim 4.80.1 #2 (Red Hat Linux)) id 1amzuT-0001d6-4T; Mon, 04 Apr 2016 08:30:57 +0000 Received: from mail-lb0-x232.google.com ([2a00:1450:4010:c04::232]) by bombadil.infradead.org with esmtps (Exim 4.80.1 #2 (Red Hat Linux)) id 1amzuG-0001T7-GQ for linux-arm-kernel@lists.infradead.org; Mon, 04 Apr 2016 08:30:45 +0000 Received: by mail-lb0-x232.google.com with SMTP id u8so149273522lbk.0 for ; Mon, 04 Apr 2016 01:30:24 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; h=from:to:cc:subject:date:message-id:in-reply-to:references; bh=CcuIdBHhXFIQLCC7+pmd6lKBNa6lEKI1lJTf2izdlYE=; b=RugYSaro5h28F8XmtpPC+w7TgwvfOkOpEcG1CKdbH8CovZLWBfhV49JEnqExPpRzFo DGqgOgeuCv5uPe9RiB474CSMmu8yicXUbeTugwybA+QEldVcK9TrYjVC9JGCajzKuL4U o+sJiXNSFAxz2cOn5ynX6zv7GXBNiKF8wOrVw= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references; bh=CcuIdBHhXFIQLCC7+pmd6lKBNa6lEKI1lJTf2izdlYE=; b=B8WwzNXY3Qy4rxtCvqqre1oK/DSzepvzA5hgH2TIyLnIC9HEQgQBhnwS8wWHdSof91 f4ejo4yBCpPKB/pywhMZG9LLZRgNeOJKvTeSgiI6JekFLcSrCuBnlAHTdx8vA1bM3u4X w4+EiEFbQv6Gtt8GT3Px+ZdrgHLaQXwM1REtzWCRhX9DDH1hU/nCOo+yqZoJ2zbLETS0 FqyYRc9u/x/0c2o4i2FXzdIIYeEUU/pazAATnLeQbBx0rkMPa+pHIFH89IBMYmVztqNI rduhi/1P3X9GSm+sRL5z5vQMMjJX8iVu9Nar54owMXWO6HOInELnVKxfdfKPWCuqKr6u bnCw== X-Gm-Message-State: AD7BkJJAzKGPYaoaEHCDzdJwvCU322txh79OxLue7gGRbOmFzrB55MQKCENtUw36OpTSQOWK X-Received: by 10.28.16.141 with SMTP id 135mr10838208wmq.67.1459758622418; Mon, 04 Apr 2016 01:30:22 -0700 (PDT) Received: from new-host-2.home (LMontsouris-657-1-37-90.w80-11.abo.wanadoo.fr. [80.11.198.90]) by smtp.gmail.com with ESMTPSA id j71sm1178257wmj.21.2016.04.04.01.30.19 (version=TLS1_2 cipher=ECDHE-RSA-AES128-SHA bits=128/128); Mon, 04 Apr 2016 01:30:20 -0700 (PDT) From: Eric Auger To: eric.auger@st.com, eric.auger@linaro.org, robin.murphy@arm.com, alex.williamson@redhat.com, will.deacon@arm.com, joro@8bytes.org, tglx@linutronix.de, jason@lakedaemon.net, marc.zyngier@arm.com, christoffer.dall@linaro.org, linux-arm-kernel@lists.infradead.org, kvmarm@lists.cs.columbia.edu, kvm@vger.kernel.org Subject: [PATCH v6 1/5] vfio: introduce VFIO_IOVA_RESERVED vfio_dma type Date: Mon, 4 Apr 2016 08:30:07 +0000 Message-Id: <1459758611-2972-2-git-send-email-eric.auger@linaro.org> X-Mailer: git-send-email 1.9.1 In-Reply-To: <1459758611-2972-1-git-send-email-eric.auger@linaro.org> References: <1459758611-2972-1-git-send-email-eric.auger@linaro.org> X-CRM114-Version: 20100106-BlameMichelson ( TRE 0.8.0 (BSD) ) MR-646709E3 X-CRM114-CacheID: sfid-20160404_013044_750024_73DB23A7 X-CRM114-Status: GOOD ( 14.80 ) X-Spam-Score: -2.7 (--) X-BeenThere: linux-arm-kernel@lists.infradead.org X-Mailman-Version: 2.1.20 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , Cc: julien.grall@arm.com, patches@linaro.org, Jean-Philippe.Brucker@arm.com, Manish.Jaggi@caviumnetworks.com, p.fedin@samsung.com, linux-kernel@vger.kernel.org, Bharat.Bhushan@freescale.com, iommu@lists.linux-foundation.org, pranav.sawargaonkar@gmail.com, suravee.suthikulpanit@amd.com MIME-Version: 1.0 Sender: "linux-arm-kernel" Errors-To: linux-arm-kernel-bounces+patchwork-linux-arm=patchwork.kernel.org@lists.infradead.org X-Spam-Status: No, score=-5.1 required=5.0 tests=BAYES_00,DKIM_SIGNED, RCVD_IN_DNSWL_MED,RP_MATCHES_RCVD,T_DKIM_INVALID,UNPARSEABLE_RELAY autolearn=unavailable version=3.3.1 X-Spam-Checker-Version: SpamAssassin 3.3.1 (2010-03-16) on mail.kernel.org X-Virus-Scanned: ClamAV using ClamSMTP We introduce a vfio_dma type since we will need to discriminate legacy vfio_dma's from new reserved ones. Since those latter are not mapped at registration, some treatments need to be reworked: removal, replay. Currently they are unplugged. In subsequent patches they will be reworked. Signed-off-by: Eric Auger --- drivers/vfio/vfio_iommu_type1.c | 17 ++++++++++++++++- 1 file changed, 16 insertions(+), 1 deletion(-) diff --git a/drivers/vfio/vfio_iommu_type1.c b/drivers/vfio/vfio_iommu_type1.c index 75b24e9..c9ddbde 100644 --- a/drivers/vfio/vfio_iommu_type1.c +++ b/drivers/vfio/vfio_iommu_type1.c @@ -53,6 +53,15 @@ module_param_named(disable_hugepages, MODULE_PARM_DESC(disable_hugepages, "Disable VFIO IOMMU support for IOMMU hugepages."); +enum vfio_iova_type { + VFIO_IOVA_USER = 0, /* standard IOVA used to map user vaddr */ + /* + * IOVA reserved to map special host physical addresses, + * MSI frames for instance + */ + VFIO_IOVA_RESERVED, +}; + struct vfio_iommu { struct list_head domain_list; struct mutex lock; @@ -75,6 +84,7 @@ struct vfio_dma { unsigned long vaddr; /* Process virtual addr */ size_t size; /* Map size (bytes) */ int prot; /* IOMMU_READ/WRITE */ + enum vfio_iova_type type; /* type of IOVA */ }; struct vfio_group { @@ -395,7 +405,8 @@ static void vfio_unmap_unpin(struct vfio_iommu *iommu, struct vfio_dma *dma) static void vfio_remove_dma(struct vfio_iommu *iommu, struct vfio_dma *dma) { - vfio_unmap_unpin(iommu, dma); + if (likely(dma->type != VFIO_IOVA_RESERVED)) + vfio_unmap_unpin(iommu, dma); vfio_unlink_dma(iommu, dma); kfree(dma); } @@ -671,6 +682,10 @@ static int vfio_iommu_replay(struct vfio_iommu *iommu, dma_addr_t iova; dma = rb_entry(n, struct vfio_dma, node); + + if (unlikely(dma->type == VFIO_IOVA_RESERVED)) + continue; + iova = dma->iova; while (iova < dma->iova + dma->size) {