On Fri, 3 Feb 2023 03:53:12 -0500, "Michael S. Tsirkin" <[email protected]> wrote:
> On Thu, Feb 02, 2023 at 07:00:40PM +0800, Xuan Zhuo wrote:
> > Move some structure definitions and inline functions into the
> > virtio_net.h file.
> >
> > Signed-off-by: Xuan Zhuo <[email protected]>
> > ---
> >  drivers/net/virtio/main.c       | 247 +----------------------------
> >  drivers/net/virtio/virtio_net.h | 265 ++++++++++++++++++++++++++++++++
> >  2 files changed, 267 insertions(+), 245 deletions(-)
> >  create mode 100644 drivers/net/virtio/virtio_net.h
> >
> > diff --git a/drivers/net/virtio/main.c b/drivers/net/virtio/main.c
> > index eb7f00194b5c..5683cb576474 100644
> > --- a/drivers/net/virtio/main.c
> > +++ b/drivers/net/virtio/main.c
> > @@ -4,24 +4,8 @@
> >   * Copyright 2007 Rusty Russell <[email protected]> IBM Corporation
> >   */
> >  //#define DEBUG
> > -#include <linux/netdevice.h>
> > -#include <linux/etherdevice.h>
> > -#include <linux/ethtool.h>
> > -#include <linux/module.h>
> > -#include <linux/virtio.h>
> > -#include <linux/virtio_net.h>
> > -#include <linux/bpf.h>
> > -#include <linux/bpf_trace.h>
> > -#include <linux/scatterlist.h>
> > -#include <linux/if_vlan.h>
> > -#include <linux/slab.h>
> > -#include <linux/cpu.h>
> > -#include <linux/average.h>
> > -#include <linux/filter.h>
> > -#include <linux/kernel.h>
> > -#include <net/route.h>
> > -#include <net/xdp.h>
> > -#include <net/net_failover.h>
> > +
> > +#include "virtio_net.h"
> >
> >  static int napi_weight = NAPI_POLL_WEIGHT;
> >  module_param(napi_weight, int, 0444);
>
>
> You should only move the headers that are actually needed not
> everything.

You mean the "include".

I think it is a simple way to concentrate "Include" into a header file, and
other .c files reference this header file.

Do you agree?

Thanks.

>
>
> > @@ -44,15 +28,6 @@ module_param(napi_tx, bool, 0644);
> >  #define VIRTIO_XDP_TX              BIT(0)
> >  #define VIRTIO_XDP_REDIR   BIT(1)
> >
> > -#define VIRTIO_XDP_FLAG    BIT(0)
> > -
> > -/* RX packet size EWMA. The average packet size is used to determine the 
> > packet
> > - * buffer size when refilling RX rings. As the entire RX ring may be 
> > refilled
> > - * at once, the weight is chosen so that the EWMA will be insensitive to 
> > short-
> > - * term, transient changes in packet size.
> > - */
> > -DECLARE_EWMA(pkt_len, 0, 64)
> > -
> >  #define VIRTNET_DRIVER_VERSION "1.0.0"
> >
> >  static const unsigned long guest_offloads[] = {
> > @@ -72,36 +47,6 @@ static const unsigned long guest_offloads[] = {
> >                             (1ULL << VIRTIO_NET_F_GUEST_USO4) | \
> >                             (1ULL << VIRTIO_NET_F_GUEST_USO6))
> >
> > -struct virtnet_stat_desc {
> > -   char desc[ETH_GSTRING_LEN];
> > -   size_t offset;
> > -};
> > -
> > -struct virtnet_sq_stats {
> > -   struct u64_stats_sync syncp;
> > -   u64 packets;
> > -   u64 bytes;
> > -   u64 xdp_tx;
> > -   u64 xdp_tx_drops;
> > -   u64 kicks;
> > -   u64 tx_timeouts;
> > -};
> > -
> > -struct virtnet_rq_stats {
> > -   struct u64_stats_sync syncp;
> > -   u64 packets;
> > -   u64 bytes;
> > -   u64 drops;
> > -   u64 xdp_packets;
> > -   u64 xdp_tx;
> > -   u64 xdp_redirects;
> > -   u64 xdp_drops;
> > -   u64 kicks;
> > -};
> > -
> > -#define VIRTNET_SQ_STAT(m) offsetof(struct virtnet_sq_stats, m)
> > -#define VIRTNET_RQ_STAT(m) offsetof(struct virtnet_rq_stats, m)
> > -
> >  static const struct virtnet_stat_desc virtnet_sq_stats_desc[] = {
> >     { "packets",            VIRTNET_SQ_STAT(packets) },
> >     { "bytes",              VIRTNET_SQ_STAT(bytes) },
> > @@ -125,57 +70,6 @@ static const struct virtnet_stat_desc 
> > virtnet_rq_stats_desc[] = {
> >  #define VIRTNET_SQ_STATS_LEN       ARRAY_SIZE(virtnet_sq_stats_desc)
> >  #define VIRTNET_RQ_STATS_LEN       ARRAY_SIZE(virtnet_rq_stats_desc)
> >
> > -/* Internal representation of a send virtqueue */
> > -struct send_queue {
> > -   /* Virtqueue associated with this send _queue */
> > -   struct virtqueue *vq;
> > -
> > -   /* TX: fragments + linear part + virtio header */
> > -   struct scatterlist sg[MAX_SKB_FRAGS + 2];
> > -
> > -   /* Name of the send queue: output.$index */
> > -   char name[16];
> > -
> > -   struct virtnet_sq_stats stats;
> > -
> > -   struct napi_struct napi;
> > -
> > -   /* Record whether sq is in reset state. */
> > -   bool reset;
> > -};
> > -
> > -/* Internal representation of a receive virtqueue */
> > -struct receive_queue {
> > -   /* Virtqueue associated with this receive_queue */
> > -   struct virtqueue *vq;
> > -
> > -   struct napi_struct napi;
> > -
> > -   struct bpf_prog __rcu *xdp_prog;
> > -
> > -   struct virtnet_rq_stats stats;
> > -
> > -   /* Chain pages by the private ptr. */
> > -   struct page *pages;
> > -
> > -   /* Average packet length for mergeable receive buffers. */
> > -   struct ewma_pkt_len mrg_avg_pkt_len;
> > -
> > -   /* Page frag for packet buffer allocation. */
> > -   struct page_frag alloc_frag;
> > -
> > -   /* RX: fragments + linear part + virtio header */
> > -   struct scatterlist sg[MAX_SKB_FRAGS + 2];
> > -
> > -   /* Min single buffer size for mergeable buffers case. */
> > -   unsigned int min_buf_len;
> > -
> > -   /* Name of this receive queue: input.$index */
> > -   char name[16];
> > -
> > -   struct xdp_rxq_info xdp_rxq;
> > -};
> > -
> >  /* This structure can contain rss message with maximum settings for 
> > indirection table and keysize
> >   * Note, that default structure that describes RSS configuration 
> > virtio_net_rss_config
> >   * contains same info but can't handle table values.
> > @@ -206,90 +100,6 @@ struct control_buf {
> >     struct virtio_net_ctrl_rss rss;
> >  };
> >
> > -struct virtnet_info {
> > -   struct virtio_device *vdev;
> > -   struct virtqueue *cvq;
> > -   struct net_device *dev;
> > -   struct send_queue *sq;
> > -   struct receive_queue *rq;
> > -   unsigned int status;
> > -
> > -   /* Max # of queue pairs supported by the device */
> > -   u16 max_queue_pairs;
> > -
> > -   /* # of queue pairs currently used by the driver */
> > -   u16 curr_queue_pairs;
> > -
> > -   /* # of XDP queue pairs currently used by the driver */
> > -   u16 xdp_queue_pairs;
> > -
> > -   /* xdp_queue_pairs may be 0, when xdp is already loaded. So add this. */
> > -   bool xdp_enabled;
> > -
> > -   /* I like... big packets and I cannot lie! */
> > -   bool big_packets;
> > -
> > -   /* number of sg entries allocated for big packets */
> > -   unsigned int big_packets_num_skbfrags;
> > -
> > -   /* Host will merge rx buffers for big packets (shake it! shake it!) */
> > -   bool mergeable_rx_bufs;
> > -
> > -   /* Host supports rss and/or hash report */
> > -   bool has_rss;
> > -   bool has_rss_hash_report;
> > -   u8 rss_key_size;
> > -   u16 rss_indir_table_size;
> > -   u32 rss_hash_types_supported;
> > -   u32 rss_hash_types_saved;
> > -
> > -   /* Has control virtqueue */
> > -   bool has_cvq;
> > -
> > -   /* Host can handle any s/g split between our header and packet data */
> > -   bool any_header_sg;
> > -
> > -   /* Packet virtio header size */
> > -   u8 hdr_len;
> > -
> > -   /* Work struct for delayed refilling if we run low on memory. */
> > -   struct delayed_work refill;
> > -
> > -   /* Is delayed refill enabled? */
> > -   bool refill_enabled;
> > -
> > -   /* The lock to synchronize the access to refill_enabled */
> > -   spinlock_t refill_lock;
> > -
> > -   /* Work struct for config space updates */
> > -   struct work_struct config_work;
> > -
> > -   /* Does the affinity hint is set for virtqueues? */
> > -   bool affinity_hint_set;
> > -
> > -   /* CPU hotplug instances for online & dead */
> > -   struct hlist_node node;
> > -   struct hlist_node node_dead;
> > -
> > -   struct control_buf *ctrl;
> > -
> > -   /* Ethtool settings */
> > -   u8 duplex;
> > -   u32 speed;
> > -
> > -   /* Interrupt coalescing settings */
> > -   u32 tx_usecs;
> > -   u32 rx_usecs;
> > -   u32 tx_max_packets;
> > -   u32 rx_max_packets;
> > -
> > -   unsigned long guest_offloads;
> > -   unsigned long guest_offloads_capable;
> > -
> > -   /* failover when STANDBY feature enabled */
> > -   struct failover *failover;
> > -};
> > -
> >  struct padded_vnet_hdr {
> >     struct virtio_net_hdr_v1_hash hdr;
> >     /*
> > @@ -303,45 +113,11 @@ struct padded_vnet_hdr {
> >  static void virtnet_rq_free_unused_buf(struct virtqueue *vq, void *buf);
> >  static void virtnet_sq_free_unused_buf(struct virtqueue *vq, void *buf);
> >
> > -static bool is_xdp_frame(void *ptr)
> > -{
> > -   return (unsigned long)ptr & VIRTIO_XDP_FLAG;
> > -}
> > -
> >  static void *xdp_to_ptr(struct xdp_frame *ptr)
> >  {
> >     return (void *)((unsigned long)ptr | VIRTIO_XDP_FLAG);
> >  }
> >
> > -static struct xdp_frame *ptr_to_xdp(void *ptr)
> > -{
> > -   return (struct xdp_frame *)((unsigned long)ptr & ~VIRTIO_XDP_FLAG);
> > -}
> > -
> > -static void __free_old_xmit(struct send_queue *sq, bool in_napi,
> > -                       struct virtnet_sq_stats *stats)
> > -{
> > -   unsigned int len;
> > -   void *ptr;
> > -
> > -   while ((ptr = virtqueue_get_buf(sq->vq, &len)) != NULL) {
> > -           if (!is_xdp_frame(ptr)) {
> > -                   struct sk_buff *skb = ptr;
> > -
> > -                   pr_debug("Sent skb %p\n", skb);
> > -
> > -                   stats->bytes += skb->len;
> > -                   napi_consume_skb(skb, in_napi);
> > -           } else {
> > -                   struct xdp_frame *frame = ptr_to_xdp(ptr);
> > -
> > -                   stats->bytes += xdp_get_frame_len(frame);
> > -                   xdp_return_frame(frame);
> > -           }
> > -           stats->packets++;
> > -   }
> > -}
> > -
> >  /* Converting between virtqueue no. and kernel tx/rx queue no.
> >   * 0:rx0 1:tx0 2:rx1 3:tx1 ... 2N:rxN 2N+1:txN 2N+2:cvq
> >   */
> > @@ -411,15 +187,6 @@ static void disable_delayed_refill(struct virtnet_info 
> > *vi)
> >     spin_unlock_bh(&vi->refill_lock);
> >  }
> >
> > -static void virtqueue_napi_schedule(struct napi_struct *napi,
> > -                               struct virtqueue *vq)
> > -{
> > -   if (napi_schedule_prep(napi)) {
> > -           virtqueue_disable_cb(vq);
> > -           __napi_schedule(napi);
> > -   }
> > -}
> > -
> >  static void virtqueue_napi_complete(struct napi_struct *napi,
> >                                 struct virtqueue *vq, int processed)
> >  {
> > @@ -1740,16 +1507,6 @@ static void free_old_xmit(struct send_queue *sq, 
> > bool in_napi)
> >     u64_stats_update_end(&sq->stats.syncp);
> >  }
> >
> > -static bool is_xdp_raw_buffer_queue(struct virtnet_info *vi, int q)
> > -{
> > -   if (q < (vi->curr_queue_pairs - vi->xdp_queue_pairs))
> > -           return false;
> > -   else if (q < vi->curr_queue_pairs)
> > -           return true;
> > -   else
> > -           return false;
> > -}
> > -
> >  static void virtnet_poll_cleantx(struct receive_queue *rq)
> >  {
> >     struct virtnet_info *vi = rq->vq->vdev->priv;
> > diff --git a/drivers/net/virtio/virtio_net.h 
> > b/drivers/net/virtio/virtio_net.h
> > new file mode 100644
> > index 000000000000..8bf31429ae28
> > --- /dev/null
> > +++ b/drivers/net/virtio/virtio_net.h
> > @@ -0,0 +1,265 @@
> > +/* SPDX-License-Identifier: GPL-2.0-or-later */
> > +
> > +#ifndef __VIRTIO_NET_H__
> > +#define __VIRTIO_NET_H__
> > +#include <linux/netdevice.h>
> > +#include <linux/etherdevice.h>
> > +#include <linux/ethtool.h>
> > +#include <linux/module.h>
> > +#include <linux/virtio.h>
> > +#include <linux/virtio_net.h>
> > +#include <linux/bpf.h>
> > +#include <linux/bpf_trace.h>
> > +#include <linux/scatterlist.h>
> > +#include <linux/if_vlan.h>
> > +#include <linux/slab.h>
> > +#include <linux/cpu.h>
> > +#include <linux/average.h>
> > +#include <linux/filter.h>
> > +#include <linux/kernel.h>
> > +#include <net/route.h>
> > +#include <net/xdp.h>
> > +#include <net/net_failover.h>
> > +#include <net/xdp_sock_drv.h>
> > +
> > +#define VIRTIO_XDP_FLAG    BIT(0)
> > +
> > +struct virtnet_info {
> > +   struct virtio_device *vdev;
> > +   struct virtqueue *cvq;
> > +   struct net_device *dev;
> > +   struct send_queue *sq;
> > +   struct receive_queue *rq;
> > +   unsigned int status;
> > +
> > +   /* Max # of queue pairs supported by the device */
> > +   u16 max_queue_pairs;
> > +
> > +   /* # of queue pairs currently used by the driver */
> > +   u16 curr_queue_pairs;
> > +
> > +   /* # of XDP queue pairs currently used by the driver */
> > +   u16 xdp_queue_pairs;
> > +
> > +   /* xdp_queue_pairs may be 0, when xdp is already loaded. So add this. */
> > +   bool xdp_enabled;
> > +
> > +   /* I like... big packets and I cannot lie! */
> > +   bool big_packets;
> > +
> > +   /* number of sg entries allocated for big packets */
> > +   unsigned int big_packets_num_skbfrags;
> > +
> > +   /* Host will merge rx buffers for big packets (shake it! shake it!) */
> > +   bool mergeable_rx_bufs;
> > +
> > +   /* Host supports rss and/or hash report */
> > +   bool has_rss;
> > +   bool has_rss_hash_report;
> > +   u8 rss_key_size;
> > +   u16 rss_indir_table_size;
> > +   u32 rss_hash_types_supported;
> > +   u32 rss_hash_types_saved;
> > +
> > +   /* Has control virtqueue */
> > +   bool has_cvq;
> > +
> > +   /* Host can handle any s/g split between our header and packet data */
> > +   bool any_header_sg;
> > +
> > +   /* Packet virtio header size */
> > +   u8 hdr_len;
> > +
> > +   /* Work struct for delayed refilling if we run low on memory. */
> > +   struct delayed_work refill;
> > +
> > +   /* Is delayed refill enabled? */
> > +   bool refill_enabled;
> > +
> > +   /* The lock to synchronize the access to refill_enabled */
> > +   spinlock_t refill_lock;
> > +
> > +   /* Work struct for config space updates */
> > +   struct work_struct config_work;
> > +
> > +   /* Does the affinity hint is set for virtqueues? */
> > +   bool affinity_hint_set;
> > +
> > +   /* CPU hotplug instances for online & dead */
> > +   struct hlist_node node;
> > +   struct hlist_node node_dead;
> > +
> > +   struct control_buf *ctrl;
> > +
> > +   /* Ethtool settings */
> > +   u8 duplex;
> > +   u32 speed;
> > +
> > +   /* Interrupt coalescing settings */
> > +   u32 tx_usecs;
> > +   u32 rx_usecs;
> > +   u32 tx_max_packets;
> > +   u32 rx_max_packets;
> > +
> > +   unsigned long guest_offloads;
> > +   unsigned long guest_offloads_capable;
> > +
> > +   /* failover when STANDBY feature enabled */
> > +   struct failover *failover;
> > +};
> > +
> > +/* RX packet size EWMA. The average packet size is used to determine the 
> > packet
> > + * buffer size when refilling RX rings. As the entire RX ring may be 
> > refilled
> > + * at once, the weight is chosen so that the EWMA will be insensitive to 
> > short-
> > + * term, transient changes in packet size.
> > + */
> > +DECLARE_EWMA(pkt_len, 0, 64)
> > +
> > +struct virtnet_stat_desc {
> > +   char desc[ETH_GSTRING_LEN];
> > +   size_t offset;
> > +};
> > +
> > +struct virtnet_sq_stats {
> > +   struct u64_stats_sync syncp;
> > +   u64 packets;
> > +   u64 bytes;
> > +   u64 xdp_tx;
> > +   u64 xdp_tx_drops;
> > +   u64 kicks;
> > +   u64 tx_timeouts;
> > +};
> > +
> > +struct virtnet_rq_stats {
> > +   struct u64_stats_sync syncp;
> > +   u64 packets;
> > +   u64 bytes;
> > +   u64 drops;
> > +   u64 xdp_packets;
> > +   u64 xdp_tx;
> > +   u64 xdp_redirects;
> > +   u64 xdp_drops;
> > +   u64 kicks;
> > +};
> > +
> > +#define VIRTNET_SQ_STAT(m) offsetof(struct virtnet_sq_stats, m)
> > +#define VIRTNET_RQ_STAT(m) offsetof(struct virtnet_rq_stats, m)
> > +
> > +/* Internal representation of a send virtqueue */
> > +struct send_queue {
> > +   /* Virtqueue associated with this send _queue */
> > +   struct virtqueue *vq;
> > +
> > +   /* TX: fragments + linear part + virtio header */
> > +   struct scatterlist sg[MAX_SKB_FRAGS + 2];
> > +
> > +   /* Name of the send queue: output.$index */
> > +   char name[16];
> > +
> > +   struct virtnet_sq_stats stats;
> > +
> > +   struct napi_struct napi;
> > +
> > +   /* Record whether sq is in reset state. */
> > +   bool reset;
> > +};
> > +
> > +/* Internal representation of a receive virtqueue */
> > +struct receive_queue {
> > +   /* Virtqueue associated with this receive_queue */
> > +   struct virtqueue *vq;
> > +
> > +   struct napi_struct napi;
> > +
> > +   struct bpf_prog __rcu *xdp_prog;
> > +
> > +   struct virtnet_rq_stats stats;
> > +
> > +   /* Chain pages by the private ptr. */
> > +   struct page *pages;
> > +
> > +   /* Average packet length for mergeable receive buffers. */
> > +   struct ewma_pkt_len mrg_avg_pkt_len;
> > +
> > +   /* Page frag for packet buffer allocation. */
> > +   struct page_frag alloc_frag;
> > +
> > +   /* RX: fragments + linear part + virtio header */
> > +   struct scatterlist sg[MAX_SKB_FRAGS + 2];
> > +
> > +   /* Min single buffer size for mergeable buffers case. */
> > +   unsigned int min_buf_len;
> > +
> > +   /* Name of this receive queue: input.$index */
> > +   char name[16];
> > +
> > +   struct xdp_rxq_info xdp_rxq;
> > +};
> > +
> > +static inline bool is_xdp_raw_buffer_queue(struct virtnet_info *vi, int q)
> > +{
> > +   if (q < (vi->curr_queue_pairs - vi->xdp_queue_pairs))
> > +           return false;
> > +   else if (q < vi->curr_queue_pairs)
> > +           return true;
> > +   else
> > +           return false;
> > +}
> > +
> > +static inline void virtnet_return_xdp_frame(struct send_queue *sq,
> > +                                       struct xdp_frame *frame)
> > +{
> > +   struct virtnet_info *vi = sq->vq->vdev->priv;
> > +   dma_addr_t *p_addr, addr;
> > +
> > +   p_addr = frame->data - sizeof(*p_addr);
> > +   addr = *p_addr;
> > +
> > +   virtio_dma_unmap(&vi->vdev->dev, addr, frame->len, DMA_TO_DEVICE);
> > +
> > +   xdp_return_frame(frame);
> > +}
> > +
> > +static inline void virtqueue_napi_schedule(struct napi_struct *napi,
> > +                                      struct virtqueue *vq)
> > +{
> > +   if (napi_schedule_prep(napi)) {
> > +           virtqueue_disable_cb(vq);
> > +           __napi_schedule(napi);
> > +   }
> > +}
> > +
> > +static inline bool is_xdp_frame(void *ptr)
> > +{
> > +   return (unsigned long)ptr & VIRTIO_XDP_FLAG;
> > +}
> > +
> > +static struct xdp_frame *ptr_to_xdp(void *ptr)
> > +{
> > +   return (struct xdp_frame *)((unsigned long)ptr & ~VIRTIO_XDP_FLAG);
> > +}
> > +
> > +static void __free_old_xmit(struct send_queue *sq, bool in_napi,
> > +                       struct virtnet_sq_stats *stats)
> > +{
> > +   unsigned int len;
> > +   void *ptr;
> > +
> > +   while ((ptr = virtqueue_get_buf(sq->vq, &len)) != NULL) {
> > +           if (!is_xdp_frame(ptr)) {
> > +                   struct sk_buff *skb = ptr;
> > +
> > +                   pr_debug("Sent skb %p\n", skb);
> > +
> > +                   stats->bytes += skb->len;
> > +                   napi_consume_skb(skb, in_napi);
> > +           } else {
> > +                   struct xdp_frame *frame = ptr_to_xdp(ptr);
> > +
> > +                   stats->bytes += xdp_get_frame_len(frame);
> > +                   xdp_return_frame(frame);
> > +           }
> > +           stats->packets++;
> > +   }
> > +}
> > +#endif
>
> All these APIs not prefixed with virtnet were ok as internal
> static functions. No longer ok in a header.

I agree. Will fix.

Thanks.

>
>
> > --
> > 2.32.0.3.g01195cf9f
>
_______________________________________________
Virtualization mailing list
[email protected]
https://lists.linuxfoundation.org/mailman/listinfo/virtualization

Reply via email to