From patchwork Mon Apr 4 08:30:08 2016 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Auger Eric X-Patchwork-Id: 64964 Delivered-To: patches@linaro.org Received: by 10.112.199.169 with SMTP id jl9csp1057216lbc; Mon, 4 Apr 2016 01:30:36 -0700 (PDT) X-Received: by 10.25.147.77 with SMTP id v74mr3396184lfd.19.1459758627113; Mon, 04 Apr 2016 01:30:27 -0700 (PDT) Return-Path: Received: from mail-lf0-x233.google.com (mail-lf0-x233.google.com. [2a00:1450:4010:c07::233]) by mx.google.com with ESMTPS id mx9si15291878lbb.105.2016.04.04.01.30.27 for (version=TLS1_2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Mon, 04 Apr 2016 01:30:27 -0700 (PDT) Received-SPF: pass (google.com: domain of eric.auger@linaro.org designates 2a00:1450:4010:c07::233 as permitted sender) client-ip=2a00:1450:4010:c07::233; Authentication-Results: mx.google.com; dkim=pass header.i=@linaro.org; spf=pass (google.com: domain of eric.auger@linaro.org designates 2a00:1450:4010:c07::233 as permitted sender) smtp.mailfrom=eric.auger@linaro.org; dmarc=pass (p=NONE dis=NONE) header.from=linaro.org Received: by mail-lf0-x233.google.com with SMTP id c126so39268681lfb.2 for ; Mon, 04 Apr 2016 01:30:27 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; h=from:to:cc:subject:date:message-id:in-reply-to:references; bh=jV2AanFubwuXd+k6K9mxuBEbC4zFrIsAc4Hnv0SrL3M=; b=EbxkE99gLdqWph9FE/tz+wqTxP3vwuCcTitlRtnWLSm1R4OrjTCgEet+O7ZrIk0DSK OrtoK0mB+x+i4DTVjCARtWBeXimVYzzF56yH7IJt8O/3QhqnLOxuj95ra+xhEBqun0X1 UC30N3BcHEMeJlQTMk1vVxzw2a2H3t5s7U62U= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references; bh=jV2AanFubwuXd+k6K9mxuBEbC4zFrIsAc4Hnv0SrL3M=; b=ev982OLecX9BcgKGroi4JE/D+3vby9nf2xEj59SuZVDq5fh2JhOoViCk9KnfvWrhPV 5fKaxO337ndYZvWSF9OoYIEtGNeWCQERpTAzKGxwblYOvaMfQA6QMPjdQqYsDt8zBG9b GeN3+P+8oVEhAyMaBQR9ixeBllPfwCyUXmktVGxB1+h91HXqeSoEamzw7c35w+m/Ju0w Zm/EJRhyOx63TI1hdhPlO1pDM0p8vIBYsjs9QlS5boWX9gKi+vB0Y7gykdjMvoI4BHol kKGPVk8fCFJp35db9X/ddvnp3JpsJq2ov7WIW5EziymTsGHICaeEXmgtMShujQ1ePk2B uPNA== X-Gm-Message-State: AD7BkJLj+tNqi9qHGo6jfhlvc66a3YHQCkLhZ8THhmHEut45qxIJUA/M0pSRXjdAAN5Wos1WCVU= X-Received: by 10.194.123.35 with SMTP id lx3mr9324275wjb.132.1459758625835; Mon, 04 Apr 2016 01:30:25 -0700 (PDT) Return-Path: Received: from new-host-2.home (LMontsouris-657-1-37-90.w80-11.abo.wanadoo.fr. [80.11.198.90]) by smtp.gmail.com with ESMTPSA id j71sm1178257wmj.21.2016.04.04.01.30.22 (version=TLS1_2 cipher=ECDHE-RSA-AES128-SHA bits=128/128); Mon, 04 Apr 2016 01:30:24 -0700 (PDT) From: Eric Auger To: eric.auger@st.com, eric.auger@linaro.org, robin.murphy@arm.com, alex.williamson@redhat.com, will.deacon@arm.com, joro@8bytes.org, tglx@linutronix.de, jason@lakedaemon.net, marc.zyngier@arm.com, christoffer.dall@linaro.org, linux-arm-kernel@lists.infradead.org, kvmarm@lists.cs.columbia.edu, kvm@vger.kernel.org Cc: suravee.suthikulpanit@amd.com, patches@linaro.org, linux-kernel@vger.kernel.org, Manish.Jaggi@caviumnetworks.com, Bharat.Bhushan@freescale.com, pranav.sawargaonkar@gmail.com, p.fedin@samsung.com, iommu@lists.linux-foundation.org, Jean-Philippe.Brucker@arm.com, julien.grall@arm.com Subject: [PATCH v6 2/5] vfio: allow the user to register reserved iova range for MSI mapping Date: Mon, 4 Apr 2016 08:30:08 +0000 Message-Id: <1459758611-2972-3-git-send-email-eric.auger@linaro.org> X-Mailer: git-send-email 1.9.1 In-Reply-To: <1459758611-2972-1-git-send-email-eric.auger@linaro.org> References: <1459758611-2972-1-git-send-email-eric.auger@linaro.org> The user is allowed to [un]register a reserved IOVA range by using the DMA MAP API and setting the new flag: VFIO_DMA_MAP_FLAG_MSI_RESERVED_IOVA. It provides the base address and the size. This region is stored in the vfio_dma rb tree. At that point the iova range is not mapped to any target address yet. The host kernel will use those iova when needed, typically when the VFIO-PCI device allocates its MSIs. This patch also handles the destruction of the reserved binding RB-tree and domain's iova_domains. Signed-off-by: Eric Auger Signed-off-by: Bharat Bhushan --- v3 -> v4: - use iommu_alloc/free_reserved_iova_domain exported by dma-reserved-iommu - protect vfio_register_reserved_iova_range implementation with CONFIG_IOMMU_DMA_RESERVED - handle unregistration by user-space and on vfio_iommu_type1 release v1 -> v2: - set returned value according to alloc_reserved_iova_domain result - free the iova domains in case any error occurs RFC v1 -> v1: - takes into account Alex comments, based on [RFC PATCH 1/6] vfio: Add interface for add/del reserved iova region: - use the existing dma map/unmap ioctl interface with a flag to register a reserved IOVA range. A single reserved iova region is allowed. Conflicts: drivers/vfio/vfio_iommu_type1.c --- drivers/vfio/vfio_iommu_type1.c | 141 +++++++++++++++++++++++++++++++++++++++- include/uapi/linux/vfio.h | 12 +++- 2 files changed, 150 insertions(+), 3 deletions(-) -- 1.9.1 diff --git a/drivers/vfio/vfio_iommu_type1.c b/drivers/vfio/vfio_iommu_type1.c index c9ddbde..4497b20 100644 --- a/drivers/vfio/vfio_iommu_type1.c +++ b/drivers/vfio/vfio_iommu_type1.c @@ -36,6 +36,7 @@ #include #include #include +#include #define DRIVER_VERSION "0.2" #define DRIVER_AUTHOR "Alex Williamson " @@ -403,10 +404,22 @@ static void vfio_unmap_unpin(struct vfio_iommu *iommu, struct vfio_dma *dma) vfio_lock_acct(-unlocked); } +static void vfio_unmap_reserved(struct vfio_iommu *iommu) +{ +#ifdef CONFIG_IOMMU_DMA_RESERVED + struct vfio_domain *d; + + list_for_each_entry(d, &iommu->domain_list, next) + iommu_unmap_reserved(d->domain); +#endif +} + static void vfio_remove_dma(struct vfio_iommu *iommu, struct vfio_dma *dma) { if (likely(dma->type != VFIO_IOVA_RESERVED)) vfio_unmap_unpin(iommu, dma); + else + vfio_unmap_reserved(iommu); vfio_unlink_dma(iommu, dma); kfree(dma); } @@ -489,7 +502,8 @@ static int vfio_dma_do_unmap(struct vfio_iommu *iommu, */ if (iommu->v2) { dma = vfio_find_dma(iommu, unmap->iova, 0); - if (dma && dma->iova != unmap->iova) { + if (dma && (dma->iova != unmap->iova || + (dma->type == VFIO_IOVA_RESERVED))) { ret = -EINVAL; goto unlock; } @@ -501,6 +515,10 @@ static int vfio_dma_do_unmap(struct vfio_iommu *iommu, } while ((dma = vfio_find_dma(iommu, unmap->iova, unmap->size))) { + if (dma->type == VFIO_IOVA_RESERVED) { + ret = -EINVAL; + goto unlock; + } if (!iommu->v2 && unmap->iova > dma->iova) break; unmapped += dma->size; @@ -650,6 +668,114 @@ static int vfio_dma_do_map(struct vfio_iommu *iommu, return ret; } +static int vfio_register_reserved_iova_range(struct vfio_iommu *iommu, + struct vfio_iommu_type1_dma_map *map) +{ +#ifdef CONFIG_IOMMU_DMA_RESERVED + dma_addr_t iova = map->iova; + size_t size = map->size; + uint64_t mask; + struct vfio_dma *dma; + int ret = 0; + struct vfio_domain *d; + unsigned long order; + + /* Verify that none of our __u64 fields overflow */ + if (map->size != size || map->iova != iova) + return -EINVAL; + + order = __ffs(vfio_pgsize_bitmap(iommu)); + mask = ((uint64_t)1 << order) - 1; + + WARN_ON(mask & PAGE_MASK); + + if (!size || (size | iova) & mask) + return -EINVAL; + + /* Don't allow IOVA address wrap */ + if (iova + size - 1 < iova) + return -EINVAL; + + mutex_lock(&iommu->lock); + + if (vfio_find_dma(iommu, iova, size)) { + ret = -EEXIST; + goto out; + } + + dma = kzalloc(sizeof(*dma), GFP_KERNEL); + if (!dma) { + ret = -ENOMEM; + goto out; + } + + dma->iova = iova; + dma->size = size; + dma->type = VFIO_IOVA_RESERVED; + + list_for_each_entry(d, &iommu->domain_list, next) + ret |= iommu_alloc_reserved_iova_domain(d->domain, iova, + size, order); + + if (ret) { + list_for_each_entry(d, &iommu->domain_list, next) + iommu_free_reserved_iova_domain(d->domain); + goto out; + } + + vfio_link_dma(iommu, dma); + +out: + mutex_unlock(&iommu->lock); + return ret; +#else /* CONFIG_IOMMU_DMA_RESERVED */ + return -ENODEV; +#endif +} + +static void vfio_unregister_reserved_iova_range(struct vfio_iommu *iommu, + struct vfio_iommu_type1_dma_unmap *unmap) +{ +#ifdef CONFIG_IOMMU_DMA_RESERVED + dma_addr_t iova = unmap->iova; + struct vfio_dma *dma; + size_t size = unmap->size; + uint64_t mask; + unsigned long order; + + /* Verify that none of our __u64 fields overflow */ + if (unmap->size != size || unmap->iova != iova) + return; + + order = __ffs(vfio_pgsize_bitmap(iommu)); + mask = ((uint64_t)1 << order) - 1; + + WARN_ON(mask & PAGE_MASK); + + if (!size || (size | iova) & mask) + return; + + /* Don't allow IOVA address wrap */ + if (iova + size - 1 < iova) + return; + + mutex_lock(&iommu->lock); + + dma = vfio_find_dma(iommu, iova, size); + + if (!dma || (dma->type != VFIO_IOVA_RESERVED)) { + unmap->size = 0; + goto out; + } + + unmap->size = dma->size; + vfio_remove_dma(iommu, dma); + +out: + mutex_unlock(&iommu->lock); +#endif +} + static int vfio_bus_type(struct device *dev, void *data) { struct bus_type **bus = data; @@ -946,6 +1072,7 @@ static void vfio_iommu_type1_release(void *iommu_data) struct vfio_group *group, *group_tmp; vfio_iommu_unmap_unpin_all(iommu); + vfio_unmap_reserved(iommu); list_for_each_entry_safe(domain, domain_tmp, &iommu->domain_list, next) { @@ -1020,7 +1147,8 @@ static long vfio_iommu_type1_ioctl(void *iommu_data, } else if (cmd == VFIO_IOMMU_MAP_DMA) { struct vfio_iommu_type1_dma_map map; uint32_t mask = VFIO_DMA_MAP_FLAG_READ | - VFIO_DMA_MAP_FLAG_WRITE; + VFIO_DMA_MAP_FLAG_WRITE | + VFIO_DMA_MAP_FLAG_MSI_RESERVED_IOVA; minsz = offsetofend(struct vfio_iommu_type1_dma_map, size); @@ -1030,6 +1158,9 @@ static long vfio_iommu_type1_ioctl(void *iommu_data, if (map.argsz < minsz || map.flags & ~mask) return -EINVAL; + if (map.flags & VFIO_DMA_MAP_FLAG_MSI_RESERVED_IOVA) + return vfio_register_reserved_iova_range(iommu, &map); + return vfio_dma_do_map(iommu, &map); } else if (cmd == VFIO_IOMMU_UNMAP_DMA) { @@ -1044,10 +1175,16 @@ static long vfio_iommu_type1_ioctl(void *iommu_data, if (unmap.argsz < minsz || unmap.flags) return -EINVAL; + if (unmap.flags & VFIO_DMA_MAP_FLAG_MSI_RESERVED_IOVA) { + vfio_unregister_reserved_iova_range(iommu, &unmap); + goto out; + } + ret = vfio_dma_do_unmap(iommu, &unmap); if (ret) return ret; +out: return copy_to_user((void __user *)arg, &unmap, minsz) ? -EFAULT : 0; } diff --git a/include/uapi/linux/vfio.h b/include/uapi/linux/vfio.h index 255a211..a49be8a 100644 --- a/include/uapi/linux/vfio.h +++ b/include/uapi/linux/vfio.h @@ -498,12 +498,21 @@ struct vfio_iommu_type1_info { * * Map process virtual addresses to IO virtual addresses using the * provided struct vfio_dma_map. Caller sets argsz. READ &/ WRITE required. + * + * In case MSI_RESERVED_IOVA flag is set, the API only aims at registering an + * IOVA region which will be used on some platforms to map the host MSI frame. + * in that specific case, vaddr and prot are ignored. The requirement for + * provisioning such IOVA range can be checked by calling VFIO_IOMMU_GET_INFO + * with the VFIO_IOMMU_INFO_REQUIRE_MSI_MAP attribute. A single + * MSI_RESERVED_IOVA region can be registered */ struct vfio_iommu_type1_dma_map { __u32 argsz; __u32 flags; #define VFIO_DMA_MAP_FLAG_READ (1 << 0) /* readable from device */ #define VFIO_DMA_MAP_FLAG_WRITE (1 << 1) /* writable from device */ +/* reserved iova for MSI vectors*/ +#define VFIO_DMA_MAP_FLAG_MSI_RESERVED_IOVA (1 << 2) __u64 vaddr; /* Process virtual address */ __u64 iova; /* IO virtual address */ __u64 size; /* Size of mapping (bytes) */ @@ -519,7 +528,8 @@ struct vfio_iommu_type1_dma_map { * Caller sets argsz. The actual unmapped size is returned in the size * field. No guarantee is made to the user that arbitrary unmaps of iova * or size different from those used in the original mapping call will - * succeed. + * succeed. A Reserved DMA region must be unmapped with MSI_RESERVED_IOVA + * flag set. */ struct vfio_iommu_type1_dma_unmap { __u32 argsz;