diff mbox series

[v11,01/12] iova: Export alloc_iova_fast() and free_iova_fast()

Message ID 20210818120642.165-2-xieyongji@bytedance.com
State Superseded
Headers show
Series Introduce VDUSE - vDPA Device in Userspace | expand

Commit Message

Yongji Xie Aug. 18, 2021, 12:06 p.m. UTC
Export alloc_iova_fast() and free_iova_fast() so that
some modules can make use of the per-CPU cache to get
rid of rbtree spinlock in alloc_iova() and free_iova()
during IOVA allocation.

Signed-off-by: Xie Yongji <xieyongji@bytedance.com>
---
 drivers/iommu/iova.c | 2 ++
 1 file changed, 2 insertions(+)

Comments

Jason Wang Aug. 23, 2021, 6:24 a.m. UTC | #1
在 2021/8/18 下午8:06, Xie Yongji 写道:
> Export alloc_iova_fast() and free_iova_fast() so that

> some modules can make use of the per-CPU cache to get

> rid of rbtree spinlock in alloc_iova() and free_iova()

> during IOVA allocation.

>

> Signed-off-by: Xie Yongji <xieyongji@bytedance.com>



Acked-by: Jason Wang <jasowang@redhat.com>


(If we need respin, I'd suggest to put the numbers you measured here)

Thanks


> ---

>   drivers/iommu/iova.c | 2 ++

>   1 file changed, 2 insertions(+)

>

> diff --git a/drivers/iommu/iova.c b/drivers/iommu/iova.c

> index b6cf5f16123b..3941ed6bb99b 100644

> --- a/drivers/iommu/iova.c

> +++ b/drivers/iommu/iova.c

> @@ -521,6 +521,7 @@ alloc_iova_fast(struct iova_domain *iovad, unsigned long size,

>   

>   	return new_iova->pfn_lo;

>   }

> +EXPORT_SYMBOL_GPL(alloc_iova_fast);

>   

>   /**

>    * free_iova_fast - free iova pfn range into rcache

> @@ -538,6 +539,7 @@ free_iova_fast(struct iova_domain *iovad, unsigned long pfn, unsigned long size)

>   

>   	free_iova(iovad, pfn);

>   }

> +EXPORT_SYMBOL_GPL(free_iova_fast);

>   

>   #define fq_ring_for_each(i, fq) \

>   	for ((i) = (fq)->head; (i) != (fq)->tail; (i) = ((i) + 1) % IOVA_FQ_SIZE)
Michael S. Tsirkin Aug. 24, 2021, 6:08 p.m. UTC | #2
On Wed, Aug 18, 2021 at 08:06:31PM +0800, Xie Yongji wrote:
> Export alloc_iova_fast() and free_iova_fast() so that

> some modules can make use of the per-CPU cache to get

> rid of rbtree spinlock in alloc_iova() and free_iova()

> during IOVA allocation.

> 

> Signed-off-by: Xie Yongji <xieyongji@bytedance.com>



This needs ack from iommu maintainers. Guys?

> ---

>  drivers/iommu/iova.c | 2 ++

>  1 file changed, 2 insertions(+)

> 

> diff --git a/drivers/iommu/iova.c b/drivers/iommu/iova.c

> index b6cf5f16123b..3941ed6bb99b 100644

> --- a/drivers/iommu/iova.c

> +++ b/drivers/iommu/iova.c

> @@ -521,6 +521,7 @@ alloc_iova_fast(struct iova_domain *iovad, unsigned long size,

>  

>  	return new_iova->pfn_lo;

>  }

> +EXPORT_SYMBOL_GPL(alloc_iova_fast);

>  

>  /**

>   * free_iova_fast - free iova pfn range into rcache

> @@ -538,6 +539,7 @@ free_iova_fast(struct iova_domain *iovad, unsigned long pfn, unsigned long size)

>  

>  	free_iova(iovad, pfn);

>  }

> +EXPORT_SYMBOL_GPL(free_iova_fast);

>  

>  #define fq_ring_for_each(i, fq) \

>  	for ((i) = (fq)->head; (i) != (fq)->tail; (i) = ((i) + 1) % IOVA_FQ_SIZE)

> -- 

> 2.11.0
Will Deacon Aug. 25, 2021, 9:55 a.m. UTC | #3
On Tue, Aug 24, 2021 at 02:08:33PM -0400, Michael S. Tsirkin wrote:
> On Wed, Aug 18, 2021 at 08:06:31PM +0800, Xie Yongji wrote:

> > Export alloc_iova_fast() and free_iova_fast() so that

> > some modules can make use of the per-CPU cache to get

> > rid of rbtree spinlock in alloc_iova() and free_iova()

> > during IOVA allocation.

> > 

> > Signed-off-by: Xie Yongji <xieyongji@bytedance.com>

> 

> 

> This needs ack from iommu maintainers. Guys?


Looks fine to me:

Acked-by: Will Deacon <will@kernel.org>


Will
John Garry Aug. 25, 2021, 10:39 a.m. UTC | #4
On 25/08/2021 10:55, Will Deacon wrote:
> On Tue, Aug 24, 2021 at 02:08:33PM -0400, Michael S. Tsirkin wrote:

>> On Wed, Aug 18, 2021 at 08:06:31PM +0800, Xie Yongji wrote:

>>> Export alloc_iova_fast() and free_iova_fast() so that

>>> some modules can make use of the per-CPU cache to get

>>> rid of rbtree spinlock in alloc_iova() and free_iova()

>>> during IOVA allocation.

>>>

>>> Signed-off-by: Xie Yongji <xieyongji@bytedance.com>

>>

>>

>> This needs ack from iommu maintainers. Guys?

> 

> Looks fine to me:

> 

> Acked-by: Will Deacon <will@kernel.org>

> 

> Will

> _______________________________________________

> iommu mailing list

> iommu@lists.linux-foundation.org

> https://lists.linuxfoundation.org/mailman/listinfo/iommu

> .

> 


JFYI, There was a preliminary discussion to move the iova rcache code 
(which the iova fast alloc and free functions are based on) out of the 
iova code and maybe into dma-iommu (being the only user). There was 
other motivation.

https://lore.kernel.org/linux-iommu/83de3911-145d-77c8-17c1-981e4ff825d3@arm.com/

Having more users complicates that...

Thanks,
John
Yongji Xie Aug. 25, 2021, 12:17 p.m. UTC | #5
On Wed, Aug 25, 2021 at 6:35 PM John Garry <john.garry@huawei.com> wrote:
>

> On 25/08/2021 10:55, Will Deacon wrote:

> > On Tue, Aug 24, 2021 at 02:08:33PM -0400, Michael S. Tsirkin wrote:

> >> On Wed, Aug 18, 2021 at 08:06:31PM +0800, Xie Yongji wrote:

> >>> Export alloc_iova_fast() and free_iova_fast() so that

> >>> some modules can make use of the per-CPU cache to get

> >>> rid of rbtree spinlock in alloc_iova() and free_iova()

> >>> during IOVA allocation.

> >>>

> >>> Signed-off-by: Xie Yongji <xieyongji@bytedance.com>

> >>

> >>

> >> This needs ack from iommu maintainers. Guys?

> >

> > Looks fine to me:

> >

> > Acked-by: Will Deacon <will@kernel.org>

> >

> > Will

> > _______________________________________________

> > iommu mailing list

> > iommu@lists.linux-foundation.org

> > https://lists.linuxfoundation.org/mailman/listinfo/iommu

> > .

> >

>

> JFYI, There was a preliminary discussion to move the iova rcache code

> (which the iova fast alloc and free functions are based on) out of the

> iova code and maybe into dma-iommu (being the only user). There was

> other motivation.

>


Would it be better to move the code into ./lib as a general library?

> https://lore.kernel.org/linux-iommu/83de3911-145d-77c8-17c1-981e4ff825d3@arm.com/

>

> Having more users complicates that...

>


Do we have some plan for this work? From our test [1],
iova_alloc_fast() is much better than iova_alloc(). So I'd like to use
it as much as possible

[1] https://lore.kernel.org/kvm/CACycT3steXFeg7NRbWpo2J59dpYcumzcvM2zcPJAVe40-EvvEg@mail.gmail.com/

Thanks,
Yongji
John Garry Aug. 27, 2021, 9:03 a.m. UTC | #6
On 25/08/2021 13:17, Yongji Xie wrote:
>> JFYI, There was a preliminary discussion to move the iova rcache code

>> (which the iova fast alloc and free functions are based on) out of the

>> iova code and maybe into dma-iommu (being the only user). There was

>> other motivation.

>>

> Would it be better to move the code into ./lib as a general library?


For a start we/I think that the rcache could be removed from the IOVA 
code, but prob should stay in drivers/iommu. I had another IOVA issue to 
solve, which complicates things. No solid plans. Need to talk to Robin more.

> 

>> https://lore.kernel.org/linux-iommu/83de3911-145d-77c8-17c1-981e4ff825d3@arm.com/

>>

>> Having more users complicates that...

>>

> Do we have some plan for this work? From our test [1],

> iova_alloc_fast() is much better than iova_alloc(). So I'd like to use

> it as much as possible

> 

> [1]https://lore.kernel.org/kvm/CACycT3steXFeg7NRbWpo2J59dpYcumzcvM2zcPJAVe40-EvvEg@mail.gmail.com/


Well if you're alloc'ing and free'ing IOVAs a lot then I can imagine it is.

Thanks,
John
diff mbox series

Patch

diff --git a/drivers/iommu/iova.c b/drivers/iommu/iova.c
index b6cf5f16123b..3941ed6bb99b 100644
--- a/drivers/iommu/iova.c
+++ b/drivers/iommu/iova.c
@@ -521,6 +521,7 @@  alloc_iova_fast(struct iova_domain *iovad, unsigned long size,
 
 	return new_iova->pfn_lo;
 }
+EXPORT_SYMBOL_GPL(alloc_iova_fast);
 
 /**
  * free_iova_fast - free iova pfn range into rcache
@@ -538,6 +539,7 @@  free_iova_fast(struct iova_domain *iovad, unsigned long pfn, unsigned long size)
 
 	free_iova(iovad, pfn);
 }
+EXPORT_SYMBOL_GPL(free_iova_fast);
 
 #define fq_ring_for_each(i, fq) \
 	for ((i) = (fq)->head; (i) != (fq)->tail; (i) = ((i) + 1) % IOVA_FQ_SIZE)