diff mbox series

[net-next,v5,11/15] virtio-net: move to virtio_net.h

Message ID 20210610082209.91487-12-xuanzhuo@linux.alibaba.com
State New
Headers show
Series virtio-net: support xdp socket zero copy | expand

Commit Message

Xuan Zhuo June 10, 2021, 8:22 a.m. UTC
Move some structure definitions and inline functions into the
virtio_net.h file.

Signed-off-by: Xuan Zhuo <xuanzhuo@linux.alibaba.com>
---
 drivers/net/virtio/virtio_net.c | 225 +------------------------------
 drivers/net/virtio/virtio_net.h | 230 ++++++++++++++++++++++++++++++++
 2 files changed, 232 insertions(+), 223 deletions(-)
 create mode 100644 drivers/net/virtio/virtio_net.h

Comments

Jason Wang June 16, 2021, 7:35 a.m. UTC | #1
在 2021/6/10 下午4:22, Xuan Zhuo 写道:
> Move some structure definitions and inline functions into the

> virtio_net.h file.

>

> Signed-off-by: Xuan Zhuo <xuanzhuo@linux.alibaba.com>



Acked-by: Jason Wang <jasowang@redhat.com>



> ---

>   drivers/net/virtio/virtio_net.c | 225 +------------------------------

>   drivers/net/virtio/virtio_net.h | 230 ++++++++++++++++++++++++++++++++

>   2 files changed, 232 insertions(+), 223 deletions(-)

>   create mode 100644 drivers/net/virtio/virtio_net.h

>

> diff --git a/drivers/net/virtio/virtio_net.c b/drivers/net/virtio/virtio_net.c

> index 953739860563..395ec1f18331 100644

> --- a/drivers/net/virtio/virtio_net.c

> +++ b/drivers/net/virtio/virtio_net.c

> @@ -4,24 +4,8 @@

>    * Copyright 2007 Rusty Russell <rusty@rustcorp.com.au> IBM Corporation

>    */

>   //#define DEBUG

> -#include <linux/netdevice.h>

> -#include <linux/etherdevice.h>

> -#include <linux/ethtool.h>

> -#include <linux/module.h>

> -#include <linux/virtio.h>

> -#include <linux/virtio_net.h>

> -#include <linux/bpf.h>

> -#include <linux/bpf_trace.h>

> -#include <linux/scatterlist.h>

> -#include <linux/if_vlan.h>

> -#include <linux/slab.h>

> -#include <linux/cpu.h>

> -#include <linux/average.h>

> -#include <linux/filter.h>

> -#include <linux/kernel.h>

> -#include <net/route.h>

> -#include <net/xdp.h>

> -#include <net/net_failover.h>

> +

> +#include "virtio_net.h"

>   

>   static int napi_weight = NAPI_POLL_WEIGHT;

>   module_param(napi_weight, int, 0444);

> @@ -44,15 +28,6 @@ module_param(napi_tx, bool, 0644);

>   #define VIRTIO_XDP_TX		BIT(0)

>   #define VIRTIO_XDP_REDIR	BIT(1)

>   

> -#define VIRTIO_XDP_FLAG	BIT(0)

> -

> -/* RX packet size EWMA. The average packet size is used to determine the packet

> - * buffer size when refilling RX rings. As the entire RX ring may be refilled

> - * at once, the weight is chosen so that the EWMA will be insensitive to short-

> - * term, transient changes in packet size.

> - */

> -DECLARE_EWMA(pkt_len, 0, 64)

> -

>   #define VIRTNET_DRIVER_VERSION "1.0.0"

>   

>   static const unsigned long guest_offloads[] = {

> @@ -68,35 +43,6 @@ static const unsigned long guest_offloads[] = {

>   				(1ULL << VIRTIO_NET_F_GUEST_ECN)  | \

>   				(1ULL << VIRTIO_NET_F_GUEST_UFO))

>   

> -struct virtnet_stat_desc {

> -	char desc[ETH_GSTRING_LEN];

> -	size_t offset;

> -};

> -

> -struct virtnet_sq_stats {

> -	struct u64_stats_sync syncp;

> -	u64 packets;

> -	u64 bytes;

> -	u64 xdp_tx;

> -	u64 xdp_tx_drops;

> -	u64 kicks;

> -};

> -

> -struct virtnet_rq_stats {

> -	struct u64_stats_sync syncp;

> -	u64 packets;

> -	u64 bytes;

> -	u64 drops;

> -	u64 xdp_packets;

> -	u64 xdp_tx;

> -	u64 xdp_redirects;

> -	u64 xdp_drops;

> -	u64 kicks;

> -};

> -

> -#define VIRTNET_SQ_STAT(m)	offsetof(struct virtnet_sq_stats, m)

> -#define VIRTNET_RQ_STAT(m)	offsetof(struct virtnet_rq_stats, m)

> -

>   static const struct virtnet_stat_desc virtnet_sq_stats_desc[] = {

>   	{ "packets",		VIRTNET_SQ_STAT(packets) },

>   	{ "bytes",		VIRTNET_SQ_STAT(bytes) },

> @@ -119,54 +65,6 @@ static const struct virtnet_stat_desc virtnet_rq_stats_desc[] = {

>   #define VIRTNET_SQ_STATS_LEN	ARRAY_SIZE(virtnet_sq_stats_desc)

>   #define VIRTNET_RQ_STATS_LEN	ARRAY_SIZE(virtnet_rq_stats_desc)

>   

> -/* Internal representation of a send virtqueue */

> -struct send_queue {

> -	/* Virtqueue associated with this send _queue */

> -	struct virtqueue *vq;

> -

> -	/* TX: fragments + linear part + virtio header */

> -	struct scatterlist sg[MAX_SKB_FRAGS + 2];

> -

> -	/* Name of the send queue: output.$index */

> -	char name[40];

> -

> -	struct virtnet_sq_stats stats;

> -

> -	struct napi_struct napi;

> -};

> -

> -/* Internal representation of a receive virtqueue */

> -struct receive_queue {

> -	/* Virtqueue associated with this receive_queue */

> -	struct virtqueue *vq;

> -

> -	struct napi_struct napi;

> -

> -	struct bpf_prog __rcu *xdp_prog;

> -

> -	struct virtnet_rq_stats stats;

> -

> -	/* Chain pages by the private ptr. */

> -	struct page *pages;

> -

> -	/* Average packet length for mergeable receive buffers. */

> -	struct ewma_pkt_len mrg_avg_pkt_len;

> -

> -	/* Page frag for packet buffer allocation. */

> -	struct page_frag alloc_frag;

> -

> -	/* RX: fragments + linear part + virtio header */

> -	struct scatterlist sg[MAX_SKB_FRAGS + 2];

> -

> -	/* Min single buffer size for mergeable buffers case. */

> -	unsigned int min_buf_len;

> -

> -	/* Name of this receive queue: input.$index */

> -	char name[40];

> -

> -	struct xdp_rxq_info xdp_rxq;

> -};

> -

>   /* Control VQ buffers: protected by the rtnl lock */

>   struct control_buf {

>   	struct virtio_net_ctrl_hdr hdr;

> @@ -178,67 +76,6 @@ struct control_buf {

>   	__virtio64 offloads;

>   };

>   

> -struct virtnet_info {

> -	struct virtio_device *vdev;

> -	struct virtqueue *cvq;

> -	struct net_device *dev;

> -	struct send_queue *sq;

> -	struct receive_queue *rq;

> -	unsigned int status;

> -

> -	/* Max # of queue pairs supported by the device */

> -	u16 max_queue_pairs;

> -

> -	/* # of queue pairs currently used by the driver */

> -	u16 curr_queue_pairs;

> -

> -	/* # of XDP queue pairs currently used by the driver */

> -	u16 xdp_queue_pairs;

> -

> -	/* xdp_queue_pairs may be 0, when xdp is already loaded. So add this. */

> -	bool xdp_enabled;

> -

> -	/* I like... big packets and I cannot lie! */

> -	bool big_packets;

> -

> -	/* Host will merge rx buffers for big packets (shake it! shake it!) */

> -	bool mergeable_rx_bufs;

> -

> -	/* Has control virtqueue */

> -	bool has_cvq;

> -

> -	/* Host can handle any s/g split between our header and packet data */

> -	bool any_header_sg;

> -

> -	/* Packet virtio header size */

> -	u8 hdr_len;

> -

> -	/* Work struct for refilling if we run low on memory. */

> -	struct delayed_work refill;

> -

> -	/* Work struct for config space updates */

> -	struct work_struct config_work;

> -

> -	/* Does the affinity hint is set for virtqueues? */

> -	bool affinity_hint_set;

> -

> -	/* CPU hotplug instances for online & dead */

> -	struct hlist_node node;

> -	struct hlist_node node_dead;

> -

> -	struct control_buf *ctrl;

> -

> -	/* Ethtool settings */

> -	u8 duplex;

> -	u32 speed;

> -

> -	unsigned long guest_offloads;

> -	unsigned long guest_offloads_capable;

> -

> -	/* failover when STANDBY feature enabled */

> -	struct failover *failover;

> -};

> -

>   struct padded_vnet_hdr {

>   	struct virtio_net_hdr_mrg_rxbuf hdr;

>   	/*

> @@ -249,21 +86,6 @@ struct padded_vnet_hdr {

>   	char padding[4];

>   };

>   

> -static bool is_xdp_frame(void *ptr)

> -{

> -	return (unsigned long)ptr & VIRTIO_XDP_FLAG;

> -}

> -

> -static void *xdp_to_ptr(struct xdp_frame *ptr)

> -{

> -	return (void *)((unsigned long)ptr | VIRTIO_XDP_FLAG);

> -}

> -

> -static struct xdp_frame *ptr_to_xdp(void *ptr)

> -{

> -	return (struct xdp_frame *)((unsigned long)ptr & ~VIRTIO_XDP_FLAG);

> -}

> -

>   static char *virtnet_alloc_frag(struct receive_queue *rq, unsigned int len,

>   				int gfp)

>   {

> @@ -280,30 +102,6 @@ static char *virtnet_alloc_frag(struct receive_queue *rq, unsigned int len,

>   	return buf;

>   }

>   

> -static void __free_old_xmit(struct send_queue *sq, bool in_napi,

> -			    struct virtnet_sq_stats *stats)

> -{

> -	unsigned int len;

> -	void *ptr;

> -

> -	while ((ptr = virtqueue_get_buf(sq->vq, &len)) != NULL) {

> -		if (!is_xdp_frame(ptr)) {

> -			struct sk_buff *skb = ptr;

> -

> -			pr_debug("Sent skb %p\n", skb);

> -

> -			stats->bytes += skb->len;

> -			napi_consume_skb(skb, in_napi);

> -		} else {

> -			struct xdp_frame *frame = ptr_to_xdp(ptr);

> -

> -			stats->bytes += frame->len;

> -			xdp_return_frame(frame);

> -		}

> -		stats->packets++;

> -	}

> -}

> -

>   /* Converting between virtqueue no. and kernel tx/rx queue no.

>    * 0:rx0 1:tx0 2:rx1 3:tx1 ... 2N:rxN 2N+1:txN 2N+2:cvq

>    */

> @@ -359,15 +157,6 @@ static struct page *get_a_page(struct receive_queue *rq, gfp_t gfp_mask)

>   	return p;

>   }

>   

> -static void virtqueue_napi_schedule(struct napi_struct *napi,

> -				    struct virtqueue *vq)

> -{

> -	if (napi_schedule_prep(napi)) {

> -		virtqueue_disable_cb(vq);

> -		__napi_schedule(napi);

> -	}

> -}

> -

>   static void virtqueue_napi_complete(struct napi_struct *napi,

>   				    struct virtqueue *vq, int processed)

>   {

> @@ -1537,16 +1326,6 @@ static void free_old_xmit(struct send_queue *sq, bool in_napi)

>   	u64_stats_update_end(&sq->stats.syncp);

>   }

>   

> -static bool is_xdp_raw_buffer_queue(struct virtnet_info *vi, int q)

> -{

> -	if (q < (vi->curr_queue_pairs - vi->xdp_queue_pairs))

> -		return false;

> -	else if (q < vi->curr_queue_pairs)

> -		return true;

> -	else

> -		return false;

> -}

> -

>   static void virtnet_poll_cleantx(struct receive_queue *rq)

>   {

>   	struct virtnet_info *vi = rq->vq->vdev->priv;

> diff --git a/drivers/net/virtio/virtio_net.h b/drivers/net/virtio/virtio_net.h

> new file mode 100644

> index 000000000000..931cc81f92fb

> --- /dev/null

> +++ b/drivers/net/virtio/virtio_net.h

> @@ -0,0 +1,230 @@

> +/* SPDX-License-Identifier: GPL-2.0-or-later */

> +

> +#ifndef __VIRTIO_NET_H__

> +#define __VIRTIO_NET_H__

> +#include <linux/netdevice.h>

> +#include <linux/etherdevice.h>

> +#include <linux/ethtool.h>

> +#include <linux/module.h>

> +#include <linux/virtio.h>

> +#include <linux/virtio_net.h>

> +#include <linux/bpf.h>

> +#include <linux/bpf_trace.h>

> +#include <linux/scatterlist.h>

> +#include <linux/if_vlan.h>

> +#include <linux/slab.h>

> +#include <linux/cpu.h>

> +#include <linux/average.h>

> +#include <linux/filter.h>

> +#include <linux/kernel.h>

> +#include <net/route.h>

> +#include <net/xdp.h>

> +#include <net/net_failover.h>

> +#include <net/xdp_sock_drv.h>

> +

> +#define VIRTIO_XDP_FLAG	BIT(0)

> +

> +struct virtnet_info {

> +	struct virtio_device *vdev;

> +	struct virtqueue *cvq;

> +	struct net_device *dev;

> +	struct send_queue *sq;

> +	struct receive_queue *rq;

> +	unsigned int status;

> +

> +	/* Max # of queue pairs supported by the device */

> +	u16 max_queue_pairs;

> +

> +	/* # of queue pairs currently used by the driver */

> +	u16 curr_queue_pairs;

> +

> +	/* # of XDP queue pairs currently used by the driver */

> +	u16 xdp_queue_pairs;

> +

> +	/* xdp_queue_pairs may be 0, when xdp is already loaded. So add this. */

> +	bool xdp_enabled;

> +

> +	/* I like... big packets and I cannot lie! */

> +	bool big_packets;

> +

> +	/* Host will merge rx buffers for big packets (shake it! shake it!) */

> +	bool mergeable_rx_bufs;

> +

> +	/* Has control virtqueue */

> +	bool has_cvq;

> +

> +	/* Host can handle any s/g split between our header and packet data */

> +	bool any_header_sg;

> +

> +	/* Packet virtio header size */

> +	u8 hdr_len;

> +

> +	/* Work struct for refilling if we run low on memory. */

> +	struct delayed_work refill;

> +

> +	/* Work struct for config space updates */

> +	struct work_struct config_work;

> +

> +	/* Does the affinity hint is set for virtqueues? */

> +	bool affinity_hint_set;

> +

> +	/* CPU hotplug instances for online & dead */

> +	struct hlist_node node;

> +	struct hlist_node node_dead;

> +

> +	struct control_buf *ctrl;

> +

> +	/* Ethtool settings */

> +	u8 duplex;

> +	u32 speed;

> +

> +	unsigned long guest_offloads;

> +	unsigned long guest_offloads_capable;

> +

> +	/* failover when STANDBY feature enabled */

> +	struct failover *failover;

> +};

> +

> +/* RX packet size EWMA. The average packet size is used to determine the packet

> + * buffer size when refilling RX rings. As the entire RX ring may be refilled

> + * at once, the weight is chosen so that the EWMA will be insensitive to short-

> + * term, transient changes in packet size.

> + */

> +DECLARE_EWMA(pkt_len, 0, 64)

> +

> +struct virtnet_stat_desc {

> +	char desc[ETH_GSTRING_LEN];

> +	size_t offset;

> +};

> +

> +struct virtnet_sq_stats {

> +	struct u64_stats_sync syncp;

> +	u64 packets;

> +	u64 bytes;

> +	u64 xdp_tx;

> +	u64 xdp_tx_drops;

> +	u64 kicks;

> +};

> +

> +struct virtnet_rq_stats {

> +	struct u64_stats_sync syncp;

> +	u64 packets;

> +	u64 bytes;

> +	u64 drops;

> +	u64 xdp_packets;

> +	u64 xdp_tx;

> +	u64 xdp_redirects;

> +	u64 xdp_drops;

> +	u64 kicks;

> +};

> +

> +#define VIRTNET_SQ_STAT(m)	offsetof(struct virtnet_sq_stats, m)

> +#define VIRTNET_RQ_STAT(m)	offsetof(struct virtnet_rq_stats, m)

> +

> +/* Internal representation of a send virtqueue */

> +struct send_queue {

> +	/* Virtqueue associated with this send _queue */

> +	struct virtqueue *vq;

> +

> +	/* TX: fragments + linear part + virtio header */

> +	struct scatterlist sg[MAX_SKB_FRAGS + 2];

> +

> +	/* Name of the send queue: output.$index */

> +	char name[40];

> +

> +	struct virtnet_sq_stats stats;

> +

> +	struct napi_struct napi;

> +};

> +

> +/* Internal representation of a receive virtqueue */

> +struct receive_queue {

> +	/* Virtqueue associated with this receive_queue */

> +	struct virtqueue *vq;

> +

> +	struct napi_struct napi;

> +

> +	struct bpf_prog __rcu *xdp_prog;

> +

> +	struct virtnet_rq_stats stats;

> +

> +	/* Chain pages by the private ptr. */

> +	struct page *pages;

> +

> +	/* Average packet length for mergeable receive buffers. */

> +	struct ewma_pkt_len mrg_avg_pkt_len;

> +

> +	/* Page frag for packet buffer allocation. */

> +	struct page_frag alloc_frag;

> +

> +	/* RX: fragments + linear part + virtio header */

> +	struct scatterlist sg[MAX_SKB_FRAGS + 2];

> +

> +	/* Min single buffer size for mergeable buffers case. */

> +	unsigned int min_buf_len;

> +

> +	/* Name of this receive queue: input.$index */

> +	char name[40];

> +

> +	struct xdp_rxq_info xdp_rxq;

> +};

> +

> +static inline bool is_xdp_raw_buffer_queue(struct virtnet_info *vi, int q)

> +{

> +	if (q < (vi->curr_queue_pairs - vi->xdp_queue_pairs))

> +		return false;

> +	else if (q < vi->curr_queue_pairs)

> +		return true;

> +	else

> +		return false;

> +}

> +

> +static inline void virtqueue_napi_schedule(struct napi_struct *napi,

> +					   struct virtqueue *vq)

> +{

> +	if (napi_schedule_prep(napi)) {

> +		virtqueue_disable_cb(vq);

> +		__napi_schedule(napi);

> +	}

> +}

> +

> +static inline bool is_xdp_frame(void *ptr)

> +{

> +	return (unsigned long)ptr & VIRTIO_XDP_FLAG;

> +}

> +

> +static inline void *xdp_to_ptr(struct xdp_frame *ptr)

> +{

> +	return (void *)((unsigned long)ptr | VIRTIO_XDP_FLAG);

> +}

> +

> +static inline struct xdp_frame *ptr_to_xdp(void *ptr)

> +{

> +	return (struct xdp_frame *)((unsigned long)ptr & ~VIRTIO_XDP_FLAG);

> +}

> +

> +static inline void __free_old_xmit(struct send_queue *sq, bool in_napi,

> +				   struct virtnet_sq_stats *stats)

> +{

> +	unsigned int len;

> +	void *ptr;

> +

> +	while ((ptr = virtqueue_get_buf(sq->vq, &len)) != NULL) {

> +		if (!is_xdp_frame(ptr)) {

> +			struct sk_buff *skb = ptr;

> +

> +			pr_debug("Sent skb %p\n", skb);

> +

> +			stats->bytes += skb->len;

> +			napi_consume_skb(skb, in_napi);

> +		} else {

> +			struct xdp_frame *frame = ptr_to_xdp(ptr);

> +

> +			stats->bytes += frame->len;

> +			xdp_return_frame(frame);

> +		}

> +		stats->packets++;

> +	}

> +}

> +

> +#endif
diff mbox series

Patch

diff --git a/drivers/net/virtio/virtio_net.c b/drivers/net/virtio/virtio_net.c
index 953739860563..395ec1f18331 100644
--- a/drivers/net/virtio/virtio_net.c
+++ b/drivers/net/virtio/virtio_net.c
@@ -4,24 +4,8 @@ 
  * Copyright 2007 Rusty Russell <rusty@rustcorp.com.au> IBM Corporation
  */
 //#define DEBUG
-#include <linux/netdevice.h>
-#include <linux/etherdevice.h>
-#include <linux/ethtool.h>
-#include <linux/module.h>
-#include <linux/virtio.h>
-#include <linux/virtio_net.h>
-#include <linux/bpf.h>
-#include <linux/bpf_trace.h>
-#include <linux/scatterlist.h>
-#include <linux/if_vlan.h>
-#include <linux/slab.h>
-#include <linux/cpu.h>
-#include <linux/average.h>
-#include <linux/filter.h>
-#include <linux/kernel.h>
-#include <net/route.h>
-#include <net/xdp.h>
-#include <net/net_failover.h>
+
+#include "virtio_net.h"
 
 static int napi_weight = NAPI_POLL_WEIGHT;
 module_param(napi_weight, int, 0444);
@@ -44,15 +28,6 @@  module_param(napi_tx, bool, 0644);
 #define VIRTIO_XDP_TX		BIT(0)
 #define VIRTIO_XDP_REDIR	BIT(1)
 
-#define VIRTIO_XDP_FLAG	BIT(0)
-
-/* RX packet size EWMA. The average packet size is used to determine the packet
- * buffer size when refilling RX rings. As the entire RX ring may be refilled
- * at once, the weight is chosen so that the EWMA will be insensitive to short-
- * term, transient changes in packet size.
- */
-DECLARE_EWMA(pkt_len, 0, 64)
-
 #define VIRTNET_DRIVER_VERSION "1.0.0"
 
 static const unsigned long guest_offloads[] = {
@@ -68,35 +43,6 @@  static const unsigned long guest_offloads[] = {
 				(1ULL << VIRTIO_NET_F_GUEST_ECN)  | \
 				(1ULL << VIRTIO_NET_F_GUEST_UFO))
 
-struct virtnet_stat_desc {
-	char desc[ETH_GSTRING_LEN];
-	size_t offset;
-};
-
-struct virtnet_sq_stats {
-	struct u64_stats_sync syncp;
-	u64 packets;
-	u64 bytes;
-	u64 xdp_tx;
-	u64 xdp_tx_drops;
-	u64 kicks;
-};
-
-struct virtnet_rq_stats {
-	struct u64_stats_sync syncp;
-	u64 packets;
-	u64 bytes;
-	u64 drops;
-	u64 xdp_packets;
-	u64 xdp_tx;
-	u64 xdp_redirects;
-	u64 xdp_drops;
-	u64 kicks;
-};
-
-#define VIRTNET_SQ_STAT(m)	offsetof(struct virtnet_sq_stats, m)
-#define VIRTNET_RQ_STAT(m)	offsetof(struct virtnet_rq_stats, m)
-
 static const struct virtnet_stat_desc virtnet_sq_stats_desc[] = {
 	{ "packets",		VIRTNET_SQ_STAT(packets) },
 	{ "bytes",		VIRTNET_SQ_STAT(bytes) },
@@ -119,54 +65,6 @@  static const struct virtnet_stat_desc virtnet_rq_stats_desc[] = {
 #define VIRTNET_SQ_STATS_LEN	ARRAY_SIZE(virtnet_sq_stats_desc)
 #define VIRTNET_RQ_STATS_LEN	ARRAY_SIZE(virtnet_rq_stats_desc)
 
-/* Internal representation of a send virtqueue */
-struct send_queue {
-	/* Virtqueue associated with this send _queue */
-	struct virtqueue *vq;
-
-	/* TX: fragments + linear part + virtio header */
-	struct scatterlist sg[MAX_SKB_FRAGS + 2];
-
-	/* Name of the send queue: output.$index */
-	char name[40];
-
-	struct virtnet_sq_stats stats;
-
-	struct napi_struct napi;
-};
-
-/* Internal representation of a receive virtqueue */
-struct receive_queue {
-	/* Virtqueue associated with this receive_queue */
-	struct virtqueue *vq;
-
-	struct napi_struct napi;
-
-	struct bpf_prog __rcu *xdp_prog;
-
-	struct virtnet_rq_stats stats;
-
-	/* Chain pages by the private ptr. */
-	struct page *pages;
-
-	/* Average packet length for mergeable receive buffers. */
-	struct ewma_pkt_len mrg_avg_pkt_len;
-
-	/* Page frag for packet buffer allocation. */
-	struct page_frag alloc_frag;
-
-	/* RX: fragments + linear part + virtio header */
-	struct scatterlist sg[MAX_SKB_FRAGS + 2];
-
-	/* Min single buffer size for mergeable buffers case. */
-	unsigned int min_buf_len;
-
-	/* Name of this receive queue: input.$index */
-	char name[40];
-
-	struct xdp_rxq_info xdp_rxq;
-};
-
 /* Control VQ buffers: protected by the rtnl lock */
 struct control_buf {
 	struct virtio_net_ctrl_hdr hdr;
@@ -178,67 +76,6 @@  struct control_buf {
 	__virtio64 offloads;
 };
 
-struct virtnet_info {
-	struct virtio_device *vdev;
-	struct virtqueue *cvq;
-	struct net_device *dev;
-	struct send_queue *sq;
-	struct receive_queue *rq;
-	unsigned int status;
-
-	/* Max # of queue pairs supported by the device */
-	u16 max_queue_pairs;
-
-	/* # of queue pairs currently used by the driver */
-	u16 curr_queue_pairs;
-
-	/* # of XDP queue pairs currently used by the driver */
-	u16 xdp_queue_pairs;
-
-	/* xdp_queue_pairs may be 0, when xdp is already loaded. So add this. */
-	bool xdp_enabled;
-
-	/* I like... big packets and I cannot lie! */
-	bool big_packets;
-
-	/* Host will merge rx buffers for big packets (shake it! shake it!) */
-	bool mergeable_rx_bufs;
-
-	/* Has control virtqueue */
-	bool has_cvq;
-
-	/* Host can handle any s/g split between our header and packet data */
-	bool any_header_sg;
-
-	/* Packet virtio header size */
-	u8 hdr_len;
-
-	/* Work struct for refilling if we run low on memory. */
-	struct delayed_work refill;
-
-	/* Work struct for config space updates */
-	struct work_struct config_work;
-
-	/* Does the affinity hint is set for virtqueues? */
-	bool affinity_hint_set;
-
-	/* CPU hotplug instances for online & dead */
-	struct hlist_node node;
-	struct hlist_node node_dead;
-
-	struct control_buf *ctrl;
-
-	/* Ethtool settings */
-	u8 duplex;
-	u32 speed;
-
-	unsigned long guest_offloads;
-	unsigned long guest_offloads_capable;
-
-	/* failover when STANDBY feature enabled */
-	struct failover *failover;
-};
-
 struct padded_vnet_hdr {
 	struct virtio_net_hdr_mrg_rxbuf hdr;
 	/*
@@ -249,21 +86,6 @@  struct padded_vnet_hdr {
 	char padding[4];
 };
 
-static bool is_xdp_frame(void *ptr)
-{
-	return (unsigned long)ptr & VIRTIO_XDP_FLAG;
-}
-
-static void *xdp_to_ptr(struct xdp_frame *ptr)
-{
-	return (void *)((unsigned long)ptr | VIRTIO_XDP_FLAG);
-}
-
-static struct xdp_frame *ptr_to_xdp(void *ptr)
-{
-	return (struct xdp_frame *)((unsigned long)ptr & ~VIRTIO_XDP_FLAG);
-}
-
 static char *virtnet_alloc_frag(struct receive_queue *rq, unsigned int len,
 				int gfp)
 {
@@ -280,30 +102,6 @@  static char *virtnet_alloc_frag(struct receive_queue *rq, unsigned int len,
 	return buf;
 }
 
-static void __free_old_xmit(struct send_queue *sq, bool in_napi,
-			    struct virtnet_sq_stats *stats)
-{
-	unsigned int len;
-	void *ptr;
-
-	while ((ptr = virtqueue_get_buf(sq->vq, &len)) != NULL) {
-		if (!is_xdp_frame(ptr)) {
-			struct sk_buff *skb = ptr;
-
-			pr_debug("Sent skb %p\n", skb);
-
-			stats->bytes += skb->len;
-			napi_consume_skb(skb, in_napi);
-		} else {
-			struct xdp_frame *frame = ptr_to_xdp(ptr);
-
-			stats->bytes += frame->len;
-			xdp_return_frame(frame);
-		}
-		stats->packets++;
-	}
-}
-
 /* Converting between virtqueue no. and kernel tx/rx queue no.
  * 0:rx0 1:tx0 2:rx1 3:tx1 ... 2N:rxN 2N+1:txN 2N+2:cvq
  */
@@ -359,15 +157,6 @@  static struct page *get_a_page(struct receive_queue *rq, gfp_t gfp_mask)
 	return p;
 }
 
-static void virtqueue_napi_schedule(struct napi_struct *napi,
-				    struct virtqueue *vq)
-{
-	if (napi_schedule_prep(napi)) {
-		virtqueue_disable_cb(vq);
-		__napi_schedule(napi);
-	}
-}
-
 static void virtqueue_napi_complete(struct napi_struct *napi,
 				    struct virtqueue *vq, int processed)
 {
@@ -1537,16 +1326,6 @@  static void free_old_xmit(struct send_queue *sq, bool in_napi)
 	u64_stats_update_end(&sq->stats.syncp);
 }
 
-static bool is_xdp_raw_buffer_queue(struct virtnet_info *vi, int q)
-{
-	if (q < (vi->curr_queue_pairs - vi->xdp_queue_pairs))
-		return false;
-	else if (q < vi->curr_queue_pairs)
-		return true;
-	else
-		return false;
-}
-
 static void virtnet_poll_cleantx(struct receive_queue *rq)
 {
 	struct virtnet_info *vi = rq->vq->vdev->priv;
diff --git a/drivers/net/virtio/virtio_net.h b/drivers/net/virtio/virtio_net.h
new file mode 100644
index 000000000000..931cc81f92fb
--- /dev/null
+++ b/drivers/net/virtio/virtio_net.h
@@ -0,0 +1,230 @@ 
+/* SPDX-License-Identifier: GPL-2.0-or-later */
+
+#ifndef __VIRTIO_NET_H__
+#define __VIRTIO_NET_H__
+#include <linux/netdevice.h>
+#include <linux/etherdevice.h>
+#include <linux/ethtool.h>
+#include <linux/module.h>
+#include <linux/virtio.h>
+#include <linux/virtio_net.h>
+#include <linux/bpf.h>
+#include <linux/bpf_trace.h>
+#include <linux/scatterlist.h>
+#include <linux/if_vlan.h>
+#include <linux/slab.h>
+#include <linux/cpu.h>
+#include <linux/average.h>
+#include <linux/filter.h>
+#include <linux/kernel.h>
+#include <net/route.h>
+#include <net/xdp.h>
+#include <net/net_failover.h>
+#include <net/xdp_sock_drv.h>
+
+#define VIRTIO_XDP_FLAG	BIT(0)
+
+struct virtnet_info {
+	struct virtio_device *vdev;
+	struct virtqueue *cvq;
+	struct net_device *dev;
+	struct send_queue *sq;
+	struct receive_queue *rq;
+	unsigned int status;
+
+	/* Max # of queue pairs supported by the device */
+	u16 max_queue_pairs;
+
+	/* # of queue pairs currently used by the driver */
+	u16 curr_queue_pairs;
+
+	/* # of XDP queue pairs currently used by the driver */
+	u16 xdp_queue_pairs;
+
+	/* xdp_queue_pairs may be 0, when xdp is already loaded. So add this. */
+	bool xdp_enabled;
+
+	/* I like... big packets and I cannot lie! */
+	bool big_packets;
+
+	/* Host will merge rx buffers for big packets (shake it! shake it!) */
+	bool mergeable_rx_bufs;
+
+	/* Has control virtqueue */
+	bool has_cvq;
+
+	/* Host can handle any s/g split between our header and packet data */
+	bool any_header_sg;
+
+	/* Packet virtio header size */
+	u8 hdr_len;
+
+	/* Work struct for refilling if we run low on memory. */
+	struct delayed_work refill;
+
+	/* Work struct for config space updates */
+	struct work_struct config_work;
+
+	/* Does the affinity hint is set for virtqueues? */
+	bool affinity_hint_set;
+
+	/* CPU hotplug instances for online & dead */
+	struct hlist_node node;
+	struct hlist_node node_dead;
+
+	struct control_buf *ctrl;
+
+	/* Ethtool settings */
+	u8 duplex;
+	u32 speed;
+
+	unsigned long guest_offloads;
+	unsigned long guest_offloads_capable;
+
+	/* failover when STANDBY feature enabled */
+	struct failover *failover;
+};
+
+/* RX packet size EWMA. The average packet size is used to determine the packet
+ * buffer size when refilling RX rings. As the entire RX ring may be refilled
+ * at once, the weight is chosen so that the EWMA will be insensitive to short-
+ * term, transient changes in packet size.
+ */
+DECLARE_EWMA(pkt_len, 0, 64)
+
+struct virtnet_stat_desc {
+	char desc[ETH_GSTRING_LEN];
+	size_t offset;
+};
+
+struct virtnet_sq_stats {
+	struct u64_stats_sync syncp;
+	u64 packets;
+	u64 bytes;
+	u64 xdp_tx;
+	u64 xdp_tx_drops;
+	u64 kicks;
+};
+
+struct virtnet_rq_stats {
+	struct u64_stats_sync syncp;
+	u64 packets;
+	u64 bytes;
+	u64 drops;
+	u64 xdp_packets;
+	u64 xdp_tx;
+	u64 xdp_redirects;
+	u64 xdp_drops;
+	u64 kicks;
+};
+
+#define VIRTNET_SQ_STAT(m)	offsetof(struct virtnet_sq_stats, m)
+#define VIRTNET_RQ_STAT(m)	offsetof(struct virtnet_rq_stats, m)
+
+/* Internal representation of a send virtqueue */
+struct send_queue {
+	/* Virtqueue associated with this send _queue */
+	struct virtqueue *vq;
+
+	/* TX: fragments + linear part + virtio header */
+	struct scatterlist sg[MAX_SKB_FRAGS + 2];
+
+	/* Name of the send queue: output.$index */
+	char name[40];
+
+	struct virtnet_sq_stats stats;
+
+	struct napi_struct napi;
+};
+
+/* Internal representation of a receive virtqueue */
+struct receive_queue {
+	/* Virtqueue associated with this receive_queue */
+	struct virtqueue *vq;
+
+	struct napi_struct napi;
+
+	struct bpf_prog __rcu *xdp_prog;
+
+	struct virtnet_rq_stats stats;
+
+	/* Chain pages by the private ptr. */
+	struct page *pages;
+
+	/* Average packet length for mergeable receive buffers. */
+	struct ewma_pkt_len mrg_avg_pkt_len;
+
+	/* Page frag for packet buffer allocation. */
+	struct page_frag alloc_frag;
+
+	/* RX: fragments + linear part + virtio header */
+	struct scatterlist sg[MAX_SKB_FRAGS + 2];
+
+	/* Min single buffer size for mergeable buffers case. */
+	unsigned int min_buf_len;
+
+	/* Name of this receive queue: input.$index */
+	char name[40];
+
+	struct xdp_rxq_info xdp_rxq;
+};
+
+static inline bool is_xdp_raw_buffer_queue(struct virtnet_info *vi, int q)
+{
+	if (q < (vi->curr_queue_pairs - vi->xdp_queue_pairs))
+		return false;
+	else if (q < vi->curr_queue_pairs)
+		return true;
+	else
+		return false;
+}
+
+static inline void virtqueue_napi_schedule(struct napi_struct *napi,
+					   struct virtqueue *vq)
+{
+	if (napi_schedule_prep(napi)) {
+		virtqueue_disable_cb(vq);
+		__napi_schedule(napi);
+	}
+}
+
+static inline bool is_xdp_frame(void *ptr)
+{
+	return (unsigned long)ptr & VIRTIO_XDP_FLAG;
+}
+
+static inline void *xdp_to_ptr(struct xdp_frame *ptr)
+{
+	return (void *)((unsigned long)ptr | VIRTIO_XDP_FLAG);
+}
+
+static inline struct xdp_frame *ptr_to_xdp(void *ptr)
+{
+	return (struct xdp_frame *)((unsigned long)ptr & ~VIRTIO_XDP_FLAG);
+}
+
+static inline void __free_old_xmit(struct send_queue *sq, bool in_napi,
+				   struct virtnet_sq_stats *stats)
+{
+	unsigned int len;
+	void *ptr;
+
+	while ((ptr = virtqueue_get_buf(sq->vq, &len)) != NULL) {
+		if (!is_xdp_frame(ptr)) {
+			struct sk_buff *skb = ptr;
+
+			pr_debug("Sent skb %p\n", skb);
+
+			stats->bytes += skb->len;
+			napi_consume_skb(skb, in_napi);
+		} else {
+			struct xdp_frame *frame = ptr_to_xdp(ptr);
+
+			stats->bytes += frame->len;
+			xdp_return_frame(frame);
+		}
+		stats->packets++;
+	}
+}
+
+#endif