mbox series

[0/3] xsk: fix for xsk_poll writeable

Message ID cover.1605686678.git.xuanzhuo@linux.alibaba.com
Headers show
Series xsk: fix for xsk_poll writeable | expand

Message

Xuan Zhuo Nov. 18, 2020, 8:25 a.m. UTC
I tried to combine cq available and tx writeable, but I found it very difficult.
Sometimes we pay attention to the status of "available" for both, but sometimes,
we may only pay attention to one, such as tx writeable, because we can use the
item of fq to write to tx. And this kind of demand may be constantly changing,
and it may be necessary to set it every time before entering xsk_poll, so
setsockopt is not very convenient. I feel even more that using a new event may
be a better solution, such as EPOLLPRI, I think it can be used here, after all,
xsk should not have OOB data ^_^.

However, two other problems were discovered during the test:

* The mask returned by datagram_poll always contains EPOLLOUT
* It is not particularly reasonable to return EPOLLOUT based on tx not full

After fixing these two problems, I found that when the process is awakened by
EPOLLOUT, the process can always get the item from cq.

Because the number of packets that the network card can send at a time is
actually limited, suppose this value is "nic_num". Once the number of
consumed items in the tx queue is greater than nic_num, this means that there
must also be new recycled items in the cq queue from nic.

In this way, as long as the tx configured by the user is larger, we won't have
the situation that tx is already in the writeable state but cannot get the item
from cq.

Xuan Zhuo (3):
  xsk: replace datagram_poll by sock_poll_wait
  xsk: change the tx writeable condition
  xsk: set tx/rx the min entries

 include/uapi/linux/if_xdp.h |  2 ++
 net/xdp/xsk.c               | 26 ++++++++++++++++++++++----
 net/xdp/xsk_queue.h         |  6 ++++++
 3 files changed, 30 insertions(+), 4 deletions(-)

--
1.8.3.1

Comments

Magnus Karlsson Nov. 23, 2020, 2 p.m. UTC | #1
On Wed, Nov 18, 2020 at 9:25 AM Xuan Zhuo <xuanzhuo@linux.alibaba.com> wrote:
>

> I tried to combine cq available and tx writeable, but I found it very difficult.

> Sometimes we pay attention to the status of "available" for both, but sometimes,

> we may only pay attention to one, such as tx writeable, because we can use the

> item of fq to write to tx. And this kind of demand may be constantly changing,

> and it may be necessary to set it every time before entering xsk_poll, so

> setsockopt is not very convenient. I feel even more that using a new event may

> be a better solution, such as EPOLLPRI, I think it can be used here, after all,

> xsk should not have OOB data ^_^.

>

> However, two other problems were discovered during the test:

>

> * The mask returned by datagram_poll always contains EPOLLOUT

> * It is not particularly reasonable to return EPOLLOUT based on tx not full

>

> After fixing these two problems, I found that when the process is awakened by

> EPOLLOUT, the process can always get the item from cq.

>

> Because the number of packets that the network card can send at a time is

> actually limited, suppose this value is "nic_num". Once the number of

> consumed items in the tx queue is greater than nic_num, this means that there

> must also be new recycled items in the cq queue from nic.

>

> In this way, as long as the tx configured by the user is larger, we won't have

> the situation that tx is already in the writeable state but cannot get the item

> from cq.


I think the overall approach of tying this into poll() instead of
setsockopt() is the right way to go. But we need a more robust
solution. Your patch #3 also breaks backwards compatibility and that
is not allowed. Could you please post some simple code example of what
it is you would like to do in user space? So you would like to wake up
when there are entries in the cq that can be retrieved and the reason
you would like to do this is that you then know you can put some more
entries into the Tx ring and they will get sent as there now are free
slots in the cq. Correct me if wrong. Would an event that wakes you up
when there is both space in the Tx ring and space in the cq work? Is
there a case in which we would like to be woken up when only the Tx
ring is non-full? Maybe there are as it might be beneficial to fill
the Tx and while doing that some entries in the cq has been completed
and away the packets go. But it would be great if you could post some
simple example code, does not need to compile or anything. Can be
pseudo code.

It would also be good to know if your goal is max throughput, max
burst size, or something else.

Thanks: Magnus


> Xuan Zhuo (3):

>   xsk: replace datagram_poll by sock_poll_wait

>   xsk: change the tx writeable condition

>   xsk: set tx/rx the min entries

>

>  include/uapi/linux/if_xdp.h |  2 ++

>  net/xdp/xsk.c               | 26 ++++++++++++++++++++++----

>  net/xdp/xsk_queue.h         |  6 ++++++

>  3 files changed, 30 insertions(+), 4 deletions(-)

>

> --

> 1.8.3.1

>
Magnus Karlsson Nov. 23, 2020, 2:11 p.m. UTC | #2
On Wed, Nov 18, 2020 at 9:26 AM Xuan Zhuo <xuanzhuo@linux.alibaba.com> wrote:
>

> datagram_poll will judge the current socket status (EPOLLIN, EPOLLOUT)

> based on the traditional socket information (eg: sk_wmem_alloc), but

> this does not apply to xsk. So this patch uses sock_poll_wait instead of

> datagram_poll, and the mask is calculated by xsk_poll.

>

> Signed-off-by: Xuan Zhuo <xuanzhuo@linux.alibaba.com>

> ---

>  net/xdp/xsk.c | 4 +++-

>  1 file changed, 3 insertions(+), 1 deletion(-)

>

> diff --git a/net/xdp/xsk.c b/net/xdp/xsk.c

> index cfbec39..7f0353e 100644

> --- a/net/xdp/xsk.c

> +++ b/net/xdp/xsk.c

> @@ -477,11 +477,13 @@ static int xsk_sendmsg(struct socket *sock, struct msghdr *m, size_t total_len)

>  static __poll_t xsk_poll(struct file *file, struct socket *sock,

>                              struct poll_table_struct *wait)

>  {

> -       __poll_t mask = datagram_poll(file, sock, wait);

> +       __poll_t mask = 0;


It would indeed be nice to not execute a number of tests in
datagram_poll that will never be triggered. It will speed up things
for sure. But we need to make sure that removing those flags that
datagram_poll sets do not have any bad effects in the code above this.
But let us tentatively keep this patch for the next version of the
patch set. Just need to figure out how to solve your problem in a nice
way first. See discussion in patch 0/3.

>         struct sock *sk = sock->sk;

>         struct xdp_sock *xs = xdp_sk(sk);

>         struct xsk_buff_pool *pool;

>

> +       sock_poll_wait(file, sock, wait);

> +

>         if (unlikely(!xsk_is_bound(xs)))

>                 return mask;

>

> --

> 1.8.3.1

>
Magnus Karlsson Nov. 24, 2020, 9:01 a.m. UTC | #3
On Mon, Nov 23, 2020 at 4:21 PM Xuan Zhuo <xuanzhuo@linux.alibaba.com> wrote:
>

> On Mon, 23 Nov 2020 15:00:48 +0100, Magnus Karlsson <magnus.karlsson@gmail.com> wrote:

> > On Wed, Nov 18, 2020 at 9:25 AM Xuan Zhuo <xuanzhuo@linux.alibaba.com> wrote:

> > >

> > > I tried to combine cq available and tx writeable, but I found it very difficult.

> > > Sometimes we pay attention to the status of "available" for both, but sometimes,

> > > we may only pay attention to one, such as tx writeable, because we can use the

> > > item of fq to write to tx. And this kind of demand may be constantly changing,

> > > and it may be necessary to set it every time before entering xsk_poll, so

> > > setsockopt is not very convenient. I feel even more that using a new event may

> > > be a better solution, such as EPOLLPRI, I think it can be used here, after all,

> > > xsk should not have OOB data ^_^.

> > >

> > > However, two other problems were discovered during the test:

> > >

> > > * The mask returned by datagram_poll always contains EPOLLOUT

> > > * It is not particularly reasonable to return EPOLLOUT based on tx not full

> > >

> > > After fixing these two problems, I found that when the process is awakened by

> > > EPOLLOUT, the process can always get the item from cq.

> > >

> > > Because the number of packets that the network card can send at a time is

> > > actually limited, suppose this value is "nic_num". Once the number of

> > > consumed items in the tx queue is greater than nic_num, this means that there

> > > must also be new recycled items in the cq queue from nic.

> > >

> > > In this way, as long as the tx configured by the user is larger, we won't have

> > > the situation that tx is already in the writeable state but cannot get the item

> > > from cq.

> >

> > I think the overall approach of tying this into poll() instead of

> > setsockopt() is the right way to go. But we need a more robust

> > solution. Your patch #3 also breaks backwards compatibility and that

> > is not allowed. Could you please post some simple code example of what

> > it is you would like to do in user space? So you would like to wake up

> > when there are entries in the cq that can be retrieved and the reason

> > you would like to do this is that you then know you can put some more

> > entries into the Tx ring and they will get sent as there now are free

> > slots in the cq. Correct me if wrong. Would an event that wakes you up

> > when there is both space in the Tx ring and space in the cq work? Is

> > there a case in which we would like to be woken up when only the Tx

> > ring is non-full? Maybe there are as it might be beneficial to fill

> > the Tx and while doing that some entries in the cq has been completed

> > and away the packets go. But it would be great if you could post some

> > simple example code, does not need to compile or anything. Can be

> > pseudo code.

> >

> > It would also be good to know if your goal is max throughput, max

> > burst size, or something else.

> >

> > Thanks: Magnus

> >

>

> My goal is max pps, If possible, increase the size of buf appropriately to

> improve throughput. like pktgen.

>

> The code like this: (tx and umem cq also is 1024, and that works with zero

> copy.)

>

> ```

> void send_handler(xsk)

> {

>     char buf[22];

>

>         while (true) {

>             while (true){

>                 if (send_buf_to_tx_ring(xsk, buf, sizeof(buf)))

>                     break; // break this when no cq or tx is full

>             }

>

>             if (sendto(xsk->fd))

>                 break;

>                 }

>         }

> }

>

>

> static int loop(int efd, xsk)

> {

>         struct epoll_event e[1024];

>         struct epoll_event ee;

>         int n, i;

>

>         ee.events = EPOLLOUT;

>         ee.data.ptr = NULL;

>

>         epoll_ctl(efd, EPOLL_CTL_ADD, xsk->fd, &e);

>

>         while (1) {

>                 n = epoll_wait(efd, e, sizeof(e)/sizeof(e[0]), -1);

>

>                 if (n == 0)

>                         continue;

>

>                 if (n < 0) {

>                         continue;

>                 }

>

>                 for (i = 0; i < n; ++i) {

>             send_handler(xsk);

>                 }

>         }

> }

> ```

>

> 1. Now, since datagram_poll(that determine whether it is write able based on

>    sock_writeable function) will return EPOLLOUT every time, epoll_wait will

>    always return directly(this results in cpu 100%).


We should keep patch #1. Just need to make sure we do not break
anything as I am not familiar with the path after xsk_poll returns.

> 2. After removing datagram_poll, since tx full is a very short moment, so every

>    time tx is not full is always true, epoll_wait will still return directly

> 3. After epoll_wait returns, app will try to get cq and writes it to tx again,

>    but this time basically it will fail when getting cq. My analysis is that

>    cq item has not returned from the network card at this time.

>

>

> Under normal circumstances, the judgment preparation for this event that can be

> written is not whether the queue or buffer is full. The judgment criterion of

> tcp is whether the free space is more than half.

> This is the origin of my #2 patch, and I found that after adding this patch, my

> above problems no longer appear.

> 1. epoll_wait no longer exits directly

> 2. Every time you receive EPOLLOUT, you can always get cq


Got it. Make sense. And good that there is some precedence that you
are not supposed to wake up when there is one free slot. Instead you
should wake up when a lot of them are free so you can insert a batch.
So let us also keep patch #2, though I might have some comments on it,
but I will reply to that patch in that case.

But patch #3 needs to go. How about you instead make the Tx ring
double the size of the completion ring? Let us assume patch #1 and #2
are in place. You will get woken up when at least half the entries in
the Tx ring are available. At this point fill the Tx ring completely
and after that start cleaning the completion ring. Hopefully by this
time, there will be a number of entries in there that can be cleaned
up. Then you call sendto(). It might even be a good idea to do cq, Tx,
cq in that order.

I consider #1 and #2 bug fixes so please base them on the bpf tree and
note this in your mail header like this: "[PATCH bpf 0/3] xsk: fix for
xsk_poll writeable".

>

> In addition:

>     What is the goal of TX_BATCH_SIZE and why this "restriction" should be added,

>     which causes a lot of trouble in programming without using zero copy


You are right, this is likely too low. I never thought of this as
something that would be used as a "fast-path". It was only a generic
fall back. But it need not be. Please produce a patch #3 that sets
this to a higher value. We do need the limit though. How about 512?

If you are interested in improving the performance of the Tx SKB path,
then there might be other avenues to try if you are interested. Here
are some examples:

* Batch dev_direct_xmit. Maybe skb lists can be used.
* Do not unlock and lock for every single packet in dev_direct_xmit().
Can be combined with the above.
* Use fragments instead of copying packets into the skb itself
* Can the bool more in netdev_start_xmit be used to increase performance

>

> Thanks.

>

> >

> > > Xuan Zhuo (3):

> > >   xsk: replace datagram_poll by sock_poll_wait

> > >   xsk: change the tx writeable condition

> > >   xsk: set tx/rx the min entries

> > >

> > >  include/uapi/linux/if_xdp.h |  2 ++

> > >  net/xdp/xsk.c               | 26 ++++++++++++++++++++++----

> > >  net/xdp/xsk_queue.h         |  6 ++++++

> > >  3 files changed, 30 insertions(+), 4 deletions(-)

> > >

> > > --

> > > 1.8.3.1

> > >
Magnus Karlsson Nov. 24, 2020, 10:38 a.m. UTC | #4
On Tue, Nov 24, 2020 at 10:01 AM Magnus Karlsson
<magnus.karlsson@gmail.com> wrote:
>

> On Mon, Nov 23, 2020 at 4:21 PM Xuan Zhuo <xuanzhuo@linux.alibaba.com> wrote:

> >

> > On Mon, 23 Nov 2020 15:00:48 +0100, Magnus Karlsson <magnus.karlsson@gmail.com> wrote:

> > > On Wed, Nov 18, 2020 at 9:25 AM Xuan Zhuo <xuanzhuo@linux.alibaba.com> wrote:

> > > >

> > > > I tried to combine cq available and tx writeable, but I found it very difficult.

> > > > Sometimes we pay attention to the status of "available" for both, but sometimes,

> > > > we may only pay attention to one, such as tx writeable, because we can use the

> > > > item of fq to write to tx. And this kind of demand may be constantly changing,

> > > > and it may be necessary to set it every time before entering xsk_poll, so

> > > > setsockopt is not very convenient. I feel even more that using a new event may

> > > > be a better solution, such as EPOLLPRI, I think it can be used here, after all,

> > > > xsk should not have OOB data ^_^.

> > > >

> > > > However, two other problems were discovered during the test:

> > > >

> > > > * The mask returned by datagram_poll always contains EPOLLOUT

> > > > * It is not particularly reasonable to return EPOLLOUT based on tx not full

> > > >

> > > > After fixing these two problems, I found that when the process is awakened by

> > > > EPOLLOUT, the process can always get the item from cq.

> > > >

> > > > Because the number of packets that the network card can send at a time is

> > > > actually limited, suppose this value is "nic_num". Once the number of

> > > > consumed items in the tx queue is greater than nic_num, this means that there

> > > > must also be new recycled items in the cq queue from nic.

> > > >

> > > > In this way, as long as the tx configured by the user is larger, we won't have

> > > > the situation that tx is already in the writeable state but cannot get the item

> > > > from cq.

> > >

> > > I think the overall approach of tying this into poll() instead of

> > > setsockopt() is the right way to go. But we need a more robust

> > > solution. Your patch #3 also breaks backwards compatibility and that

> > > is not allowed. Could you please post some simple code example of what

> > > it is you would like to do in user space? So you would like to wake up

> > > when there are entries in the cq that can be retrieved and the reason

> > > you would like to do this is that you then know you can put some more

> > > entries into the Tx ring and they will get sent as there now are free

> > > slots in the cq. Correct me if wrong. Would an event that wakes you up

> > > when there is both space in the Tx ring and space in the cq work? Is

> > > there a case in which we would like to be woken up when only the Tx

> > > ring is non-full? Maybe there are as it might be beneficial to fill

> > > the Tx and while doing that some entries in the cq has been completed

> > > and away the packets go. But it would be great if you could post some

> > > simple example code, does not need to compile or anything. Can be

> > > pseudo code.

> > >

> > > It would also be good to know if your goal is max throughput, max

> > > burst size, or something else.

> > >

> > > Thanks: Magnus

> > >

> >

> > My goal is max pps, If possible, increase the size of buf appropriately to

> > improve throughput. like pktgen.

> >

> > The code like this: (tx and umem cq also is 1024, and that works with zero

> > copy.)

> >

> > ```

> > void send_handler(xsk)

> > {

> >     char buf[22];

> >

> >         while (true) {

> >             while (true){

> >                 if (send_buf_to_tx_ring(xsk, buf, sizeof(buf)))

> >                     break; // break this when no cq or tx is full

> >             }

> >

> >             if (sendto(xsk->fd))

> >                 break;

> >                 }

> >         }

> > }

> >

> >

> > static int loop(int efd, xsk)

> > {

> >         struct epoll_event e[1024];

> >         struct epoll_event ee;

> >         int n, i;

> >

> >         ee.events = EPOLLOUT;

> >         ee.data.ptr = NULL;

> >

> >         epoll_ctl(efd, EPOLL_CTL_ADD, xsk->fd, &e);

> >

> >         while (1) {

> >                 n = epoll_wait(efd, e, sizeof(e)/sizeof(e[0]), -1);

> >

> >                 if (n == 0)

> >                         continue;

> >

> >                 if (n < 0) {

> >                         continue;

> >                 }

> >

> >                 for (i = 0; i < n; ++i) {

> >             send_handler(xsk);

> >                 }

> >         }

> > }

> > ```

> >

> > 1. Now, since datagram_poll(that determine whether it is write able based on

> >    sock_writeable function) will return EPOLLOUT every time, epoll_wait will

> >    always return directly(this results in cpu 100%).

>

> We should keep patch #1. Just need to make sure we do not break

> anything as I am not familiar with the path after xsk_poll returns.

>

> > 2. After removing datagram_poll, since tx full is a very short moment, so every

> >    time tx is not full is always true, epoll_wait will still return directly

> > 3. After epoll_wait returns, app will try to get cq and writes it to tx again,

> >    but this time basically it will fail when getting cq. My analysis is that

> >    cq item has not returned from the network card at this time.

> >

> >

> > Under normal circumstances, the judgment preparation for this event that can be

> > written is not whether the queue or buffer is full. The judgment criterion of

> > tcp is whether the free space is more than half.

> > This is the origin of my #2 patch, and I found that after adding this patch, my

> > above problems no longer appear.

> > 1. epoll_wait no longer exits directly

> > 2. Every time you receive EPOLLOUT, you can always get cq

>

> Got it. Make sense. And good that there is some precedence that you

> are not supposed to wake up when there is one free slot. Instead you

> should wake up when a lot of them are free so you can insert a batch.

> So let us also keep patch #2, though I might have some comments on it,

> but I will reply to that patch in that case.

>

> But patch #3 needs to go. How about you instead make the Tx ring

> double the size of the completion ring? Let us assume patch #1 and #2

> are in place. You will get woken up when at least half the entries in

> the Tx ring are available. At this point fill the Tx ring completely

> and after that start cleaning the completion ring. Hopefully by this

> time, there will be a number of entries in there that can be cleaned

> up. Then you call sendto(). It might even be a good idea to do cq, Tx,

> cq in that order.

>

> I consider #1 and #2 bug fixes so please base them on the bpf tree and

> note this in your mail header like this: "[PATCH bpf 0/3] xsk: fix for

> xsk_poll writeable".

>

> >

> > In addition:

> >     What is the goal of TX_BATCH_SIZE and why this "restriction" should be added,

> >     which causes a lot of trouble in programming without using zero copy

>

> You are right, this is likely too low. I never thought of this as

> something that would be used as a "fast-path". It was only a generic

> fall back. But it need not be. Please produce a patch #3 that sets

> this to a higher value. We do need the limit though. How about 512?


Please produce one patch set with #1 and #2 based on the bpf tree and
keep the TX_BATCH_SIZE patch separate. It is only a performance
optimization and should be based on bpf-next and sent as a sole patch
on its own.

Thanks!

> If you are interested in improving the performance of the Tx SKB path,

> then there might be other avenues to try if you are interested. Here

> are some examples:

>

> * Batch dev_direct_xmit. Maybe skb lists can be used.

> * Do not unlock and lock for every single packet in dev_direct_xmit().

> Can be combined with the above.

> * Use fragments instead of copying packets into the skb itself

> * Can the bool more in netdev_start_xmit be used to increase performance

>

> >

> > Thanks.

> >

> > >

> > > > Xuan Zhuo (3):

> > > >   xsk: replace datagram_poll by sock_poll_wait

> > > >   xsk: change the tx writeable condition

> > > >   xsk: set tx/rx the min entries

> > > >

> > > >  include/uapi/linux/if_xdp.h |  2 ++

> > > >  net/xdp/xsk.c               | 26 ++++++++++++++++++++++----

> > > >  net/xdp/xsk_queue.h         |  6 ++++++

> > > >  3 files changed, 30 insertions(+), 4 deletions(-)

> > > >

> > > > --

> > > > 1.8.3.1

> > > >
Magnus Karlsson Nov. 24, 2020, 11:36 a.m. UTC | #5
On Mon, Nov 23, 2020 at 3:11 PM Magnus Karlsson
<magnus.karlsson@gmail.com> wrote:
>

> On Wed, Nov 18, 2020 at 9:26 AM Xuan Zhuo <xuanzhuo@linux.alibaba.com> wrote:

> >

> > datagram_poll will judge the current socket status (EPOLLIN, EPOLLOUT)

> > based on the traditional socket information (eg: sk_wmem_alloc), but

> > this does not apply to xsk. So this patch uses sock_poll_wait instead of

> > datagram_poll, and the mask is calculated by xsk_poll.

> >

> > Signed-off-by: Xuan Zhuo <xuanzhuo@linux.alibaba.com>

> > ---

> >  net/xdp/xsk.c | 4 +++-

> >  1 file changed, 3 insertions(+), 1 deletion(-)

> >

> > diff --git a/net/xdp/xsk.c b/net/xdp/xsk.c

> > index cfbec39..7f0353e 100644

> > --- a/net/xdp/xsk.c

> > +++ b/net/xdp/xsk.c

> > @@ -477,11 +477,13 @@ static int xsk_sendmsg(struct socket *sock, struct msghdr *m, size_t total_len)

> >  static __poll_t xsk_poll(struct file *file, struct socket *sock,

> >                              struct poll_table_struct *wait)

> >  {

> > -       __poll_t mask = datagram_poll(file, sock, wait);

> > +       __poll_t mask = 0;

>

> It would indeed be nice to not execute a number of tests in

> datagram_poll that will never be triggered. It will speed up things

> for sure. But we need to make sure that removing those flags that

> datagram_poll sets do not have any bad effects in the code above this.

> But let us tentatively keep this patch for the next version of the

> patch set. Just need to figure out how to solve your problem in a nice

> way first. See discussion in patch 0/3.

>

> >         struct sock *sk = sock->sk;

> >         struct xdp_sock *xs = xdp_sk(sk);

> >         struct xsk_buff_pool *pool;

> >

> > +       sock_poll_wait(file, sock, wait);

> > +

> >         if (unlikely(!xsk_is_bound(xs)))

> >                 return mask;

> >

> > --

> > 1.8.3.1

> >


The fix looks correct and it will speed things up too as a bonus.
Please include this patch in the v2 as outlined in my answer to 0/3.

Thanks!
Magnus Karlsson Nov. 27, 2020, 8:23 a.m. UTC | #6
On Wed, Nov 25, 2020 at 7:49 AM Xuan Zhuo <xuanzhuo@linux.alibaba.com> wrote:
>

> datagram_poll will judge the current socket status (EPOLLIN, EPOLLOUT)

> based on the traditional socket information (eg: sk_wmem_alloc), but

> this does not apply to xsk. So this patch uses sock_poll_wait instead of

> datagram_poll, and the mask is calculated by xsk_poll.

>

> Signed-off-by: Xuan Zhuo <xuanzhuo@linux.alibaba.com>

> ---

>  net/xdp/xsk.c | 4 +++-

>  1 file changed, 3 insertions(+), 1 deletion(-)


Acked-by: Magnus Karlsson <magnus.karlsson@intel.com>


> diff --git a/net/xdp/xsk.c b/net/xdp/xsk.c

> index b014197..0df8651 100644

> --- a/net/xdp/xsk.c

> +++ b/net/xdp/xsk.c

> @@ -534,11 +534,13 @@ static int xsk_sendmsg(struct socket *sock, struct msghdr *m, size_t total_len)

>  static __poll_t xsk_poll(struct file *file, struct socket *sock,

>                              struct poll_table_struct *wait)

>  {

> -       __poll_t mask = datagram_poll(file, sock, wait);

> +       __poll_t mask = 0;

>         struct sock *sk = sock->sk;

>         struct xdp_sock *xs = xdp_sk(sk);

>         struct xsk_buff_pool *pool;

>

> +       sock_poll_wait(file, sock, wait);

> +

>         if (unlikely(!xsk_is_bound(xs)))

>                 return mask;

>

> --

> 1.8.3.1

>