From patchwork Tue Mar 1 18:27:57 2016 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Auger Eric X-Patchwork-Id: 63340 Delivered-To: patches@linaro.org Received: by 10.112.199.169 with SMTP id jl9csp1981793lbc; Tue, 1 Mar 2016 10:29:14 -0800 (PST) X-Received: by 10.28.131.141 with SMTP id f135mr428372wmd.33.1456856945560; Tue, 01 Mar 2016 10:29:05 -0800 (PST) Return-Path: Received: from mail-wm0-x236.google.com (mail-wm0-x236.google.com. [2a00:1450:400c:c09::236]) by mx.google.com with ESMTPS id fc13si38589246wjc.44.2016.03.01.10.29.05 for (version=TLS1_2 cipher=ECDHE-RSA-AES128-GCM-SHA256 bits=128/128); Tue, 01 Mar 2016 10:29:05 -0800 (PST) Received-SPF: pass (google.com: domain of eric.auger@linaro.org designates 2a00:1450:400c:c09::236 as permitted sender) client-ip=2a00:1450:400c:c09::236; Authentication-Results: mx.google.com; spf=pass (google.com: domain of eric.auger@linaro.org designates 2a00:1450:400c:c09::236 as permitted sender) smtp.mailfrom=eric.auger@linaro.org; dkim=pass header.i=@linaro.org Received: by mail-wm0-x236.google.com with SMTP id l68so45927239wml.1 for ; Tue, 01 Mar 2016 10:29:05 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=linaro.org; s=google; h=from:to:cc:subject:date:message-id:in-reply-to:references; bh=0ZYxL7cy2Aoujq7k3+PNjMAThTwBcZYV7tH9vi686FI=; b=foPCQHkW1KHASsKiDUDjIuGOZuDE5hck5FJV0daAog6/XrNcmdlDnlhRP42pFwOHoz SZADossC116Nn89BmPRjGfzXI7Ma2SFfjP6ste6ptFQ5Uk5TlTzMSr12sgz5Lf88Yykb ifjjJ2bxsxkL18PcnsV8JuHWO864k620Mw274= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20130820; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references; bh=0ZYxL7cy2Aoujq7k3+PNjMAThTwBcZYV7tH9vi686FI=; b=lB0eMXYP1WpzVOEPOvd6wyVL/mMX3jBVRV+Z205uXkAY7yDwDr+mUgicllVUghvzVw F0ItrjCtISYtZni8n7wloTU+5vACNXNc05uY0TBk55VzaqxDhdH6Mg+FMpfDT1EvZCTn hw0irUESsf0NqYp9ZbtTud/6mGZ3MTuXjIRTKKeB5xpSr93HillEXXhUQQiEy8+38WxT FJWU5jHUaXVABpStGEz/q/dtVtgWpurRaUYh0aghpkPgaD9SFE92asn+IL9bkFIUVeTD JHSVJQ3S475OZvks8oCwkJaea7Z/SBPddzTE/aAeF6yZwg3v3IsEEKfXYrVGjfdDAPld swLQ== X-Gm-Message-State: AD7BkJJVyOohWJ/rVuimpyAlUiM4Ce4MhZadFD36KkA+ws+j1TDBjUrZLpTutN42g6M6gFfxob8= X-Received: by 10.194.189.231 with SMTP id gl7mr21800842wjc.162.1456856945339; Tue, 01 Mar 2016 10:29:05 -0800 (PST) Return-Path: Received: from new-host-8.home (LMontsouris-657-1-37-90.w80-11.abo.wanadoo.fr. [80.11.198.90]) by smtp.gmail.com with ESMTPSA id k8sm32176385wjr.38.2016.03.01.10.29.02 (version=TLS1_2 cipher=ECDHE-RSA-AES128-SHA bits=128/128); Tue, 01 Mar 2016 10:29:03 -0800 (PST) From: Eric Auger To: eric.auger@st.com, eric.auger@linaro.org, robin.murphy@arm.com, alex.williamson@redhat.com, will.deacon@arm.com, joro@8bytes.org, tglx@linutronix.de, jason@lakedaemon.net, marc.zyngier@arm.com, christoffer.dall@linaro.org, linux-arm-kernel@lists.infradead.org, kvmarm@lists.cs.columbia.edu, kvm@vger.kernel.org Cc: suravee.suthikulpanit@amd.com, patches@linaro.org, linux-kernel@vger.kernel.org, Manish.Jaggi@caviumnetworks.com, Bharat.Bhushan@freescale.com, pranav.sawargaonkar@gmail.com, p.fedin@samsung.com, iommu@lists.linux-foundation.org Subject: [RFC v5 17/17] vfio/type1: return MSI mapping requirements with VFIO_IOMMU_GET_INFO Date: Tue, 1 Mar 2016 18:27:57 +0000 Message-Id: <1456856877-4817-18-git-send-email-eric.auger@linaro.org> X-Mailer: git-send-email 1.9.1 In-Reply-To: <1456856877-4817-1-git-send-email-eric.auger@linaro.org> References: <1456856877-4817-1-git-send-email-eric.auger@linaro.org> This patch allows the user-space to know whether MSI addresses need to be mapped in the IOMMU. The user-space uses VFIO_IOMMU_GET_INFO ioctl and IOMMU_INFO_REQUIRE_MSI_MAP gets set if they need to. Also the number of IOMMU pages requested to map those is returned in msi_iova_pages field. User-space must use this information to allocate an IOVA contiguous region of size msi_iova_pages * ffs(iova_pgsizes) and pass it with VFIO_IOMMU_MAP_DMA iotcl (VFIO_DMA_MAP_FLAG_MSI_RESERVED_IOVA set). Signed-off-by: Eric Auger --- Currently it is assumed a single doorbell page is used per MSI controller. This is the case for known ARM MSI controllers (GICv2M, GICv3 ITS, ...). If an MSI controller were to expose more doorbells it could implement a new callback at irq_chip interface. v4 -> v5: - move msi_info and ret declaration within the conditional code v3 -> v4: - replace former vfio_domains_require_msi_mapping by more complex computation of MSI mapping requirements, especially the number of pages to be provided by the user-space. - reword patch title RFC v1 -> v1: - derived from [RFC PATCH 3/6] vfio: Extend iommu-info to return MSIs automap state - renamed allow_msi_reconfig into require_msi_mapping - fixed VFIO_IOMMU_GET_INFO --- drivers/vfio/vfio_iommu_type1.c | 149 ++++++++++++++++++++++++++++++++++++++++ include/uapi/linux/vfio.h | 2 + 2 files changed, 151 insertions(+) -- 1.9.1 diff --git a/drivers/vfio/vfio_iommu_type1.c b/drivers/vfio/vfio_iommu_type1.c index 88a40f1..17a941c 100644 --- a/drivers/vfio/vfio_iommu_type1.c +++ b/drivers/vfio/vfio_iommu_type1.c @@ -39,6 +39,7 @@ #include #include #include +#include #define DRIVER_VERSION "0.2" #define DRIVER_AUTHOR "Alex Williamson " @@ -95,6 +96,18 @@ struct vfio_group { struct list_head next; }; +struct vfio_irq_chip { + struct list_head next; + struct irq_chip *chip; +}; + +struct vfio_msi_map_info { + bool mapping_required; + size_t page_size; + unsigned int iova_pages; + struct list_head irq_chip_list; +}; + /* * This code handles mapping and unmapping of user data buffers * into DMA'ble space using the IOMMU @@ -267,6 +280,128 @@ static int vaddr_get_pfn(unsigned long vaddr, int prot, unsigned long *pfn) return ret; } +#if defined(CONFIG_GENERIC_MSI_IRQ_DOMAIN) && defined(CONFIG_IOMMU_DMA_RESERVED) +/** + * vfio_dev_compute_msi_map_info: augment MSI mapping info (@data) with + * the @dev device requirements. + * + * @dev: device handle + * @data: opaque pointing to a struct vfio_msi_map_info + * + * returns 0 upon success or -ENOMEM + */ +static int vfio_dev_compute_msi_map_info(struct device *dev, void *data) +{ + struct irq_domain *domain; + struct msi_domain_info *info; + struct vfio_msi_map_info *msi_info = (struct vfio_msi_map_info *)data; + struct irq_chip *chip; + struct vfio_irq_chip *iter, *new; + + domain = dev_get_msi_domain(dev); + if (!domain) + return 0; + + /* Let's compute the needs for the MSI domain */ + info = msi_get_domain_info(domain); + chip = info->chip; + list_for_each_entry(iter, &msi_info->irq_chip_list, next) { + if (iter->chip == chip) + return 0; + } + + new = kzalloc(sizeof(*new), GFP_KERNEL); + if (!new) + return -ENOMEM; + + new->chip = chip; + + list_add(&new->next, &msi_info->irq_chip_list); + + /* + * new irq_chip to be taken into account; we currently assume + * a single iova doorbell by irq chip requesting MSI mapping + */ + msi_info->iova_pages += 1; + return 0; +} + +/** + * vfio_domain_compute_msi_map_info: compute MSI mapping requirements (@data) + * for vfio_domain @d + * + * @d: vfio domain handle + * @data: opaque pointing to a struct vfio_msi_map_info + * + * returns 0 upon success or -ENOMEM + */ +static int vfio_domain_compute_msi_map_info(struct vfio_domain *d, void *data) +{ + int ret = 0; + struct vfio_msi_map_info *msi_info = (struct vfio_msi_map_info *)data; + struct vfio_irq_chip *iter, *tmp; + struct vfio_group *g; + + msi_info->iova_pages = 0; + INIT_LIST_HEAD(&msi_info->irq_chip_list); + + if (iommu_domain_get_attr(d->domain, + DOMAIN_ATTR_MSI_MAPPING, NULL)) + return 0; + msi_info->mapping_required = true; + list_for_each_entry(g, &d->group_list, next) { + ret = iommu_group_for_each_dev(g->iommu_group, msi_info, + vfio_dev_compute_msi_map_info); + if (ret) + goto out; + } +out: + list_for_each_entry_safe(iter, tmp, &msi_info->irq_chip_list, next) { + list_del(&iter->next); + kfree(iter); + } + return ret; +} + +/** + * vfio_compute_msi_map_info: compute MSI mapping requirements + * + * Do some MSI addresses need to be mapped? IOMMU page size? + * Max number of IOVA pages needed by any domain to map MSI + * + * @iommu: iommu handle + * @info: msi map info handle + * + * returns 0 upon success or -ENOMEM + */ +static int vfio_compute_msi_map_info(struct vfio_iommu *iommu, + struct vfio_msi_map_info *msi_info) +{ + int ret = 0; + struct vfio_domain *d; + unsigned long bitmap = ULONG_MAX; + unsigned int iova_pages = 0; + + msi_info->mapping_required = false; + + mutex_lock(&iommu->lock); + list_for_each_entry(d, &iommu->domain_list, next) { + bitmap &= d->domain->ops->pgsize_bitmap; + ret = vfio_domain_compute_msi_map_info(d, msi_info); + if (ret) + goto out; + if (msi_info->iova_pages > iova_pages) + iova_pages = msi_info->iova_pages; + } +out: + msi_info->page_size = 1 << __ffs(bitmap); + msi_info->iova_pages = iova_pages; + mutex_unlock(&iommu->lock); + return ret; +} + +#endif + /* * Attempt to pin pages. We really don't want to track all the pfns and * the iommu can only map chunks of consecutive pfns anyway, so get the @@ -1179,6 +1314,20 @@ static long vfio_iommu_type1_ioctl(void *iommu_data, info.flags = VFIO_IOMMU_INFO_PGSIZES; +#if defined(CONFIG_GENERIC_MSI_IRQ_DOMAIN) && defined(CONFIG_IOMMU_DMA_RESERVED) + { + struct vfio_msi_map_info msi_info; + int ret; + + ret = vfio_compute_msi_map_info(iommu, &msi_info); + if (ret) + return ret; + + if (msi_info.mapping_required) + info.flags |= VFIO_IOMMU_INFO_REQUIRE_MSI_MAP; + info.msi_iova_pages = msi_info.iova_pages; + } +#endif info.iova_pgsizes = vfio_pgsize_bitmap(iommu); return copy_to_user((void __user *)arg, &info, minsz); diff --git a/include/uapi/linux/vfio.h b/include/uapi/linux/vfio.h index d5a48e7..863c68a 100644 --- a/include/uapi/linux/vfio.h +++ b/include/uapi/linux/vfio.h @@ -400,7 +400,9 @@ struct vfio_iommu_type1_info { __u32 argsz; __u32 flags; #define VFIO_IOMMU_INFO_PGSIZES (1 << 0) /* supported page sizes info */ +#define VFIO_IOMMU_INFO_REQUIRE_MSI_MAP (1 << 1)/* MSI must be mapped */ __u64 iova_pgsizes; /* Bitmap of supported page sizes */ + __u32 msi_iova_pages; /* number of IOVA pages needed to map MSIs */ }; #define VFIO_IOMMU_GET_INFO _IO(VFIO_TYPE, VFIO_BASE + 12)