Hello Mikhail, Harald, hoping there will be a v6 of this patch soon, a few comments:
On Monday, 8 January 2024 at 14:10:35 Mikhail Golubev-Ciuchea <mikhail.golubev-ciuc...@opensynergy.com> wrote: [...] > + > +/* virtio_can private data structure */ > +struct virtio_can_priv { > + struct can_priv can; /* must be the first member */ > + /* NAPI for RX messages */ > + struct napi_struct napi; > + /* NAPI for TX messages */ > + struct napi_struct napi_tx; > + /* The network device we're associated with */ > + struct net_device *dev; > + /* The virtio device we're associated with */ > + struct virtio_device *vdev; > + /* The virtqueues */ > + struct virtqueue *vqs[VIRTIO_CAN_QUEUE_COUNT]; > + /* I/O callback function pointers for the virtqueues */ > + vq_callback_t *io_callbacks[VIRTIO_CAN_QUEUE_COUNT]; > + /* Lock for TX operations */ > + spinlock_t tx_lock; > + /* Control queue lock. Defensive programming, may be not needed */ > + struct mutex ctrl_lock; > + /* Wait for control queue processing without polling */ > + struct completion ctrl_done; > + /* List of virtio CAN TX message */ > + struct list_head tx_list; > + /* Array of receive queue messages */ > + struct virtio_can_rx rpkt[128]; This array should probably be allocated dynamically at probe - maybe using a module parameter instead of a hardcoded value as length? > + /* Those control queue messages cannot live on the stack! */ > + struct virtio_can_control_out cpkt_out; > + struct virtio_can_control_in cpkt_in; Consider using a container struct as you did for the tx message, e.g.: struct virtio_can_control { struct virtio_can_control_out ctrl_out; struct virtio_can_control_in ctrl_in; }; > + /* Data to get and maintain the putidx for local TX echo */ > + struct ida tx_putidx_ida; > + /* In flight TX messages */ > + atomic_t tx_inflight; > + /* BusOff pending. Reset after successful indication to upper layer */ > + bool busoff_pending; > +}; > + [...] > + > +/* Send a control message with message type either > + * > + * - VIRTIO_CAN_SET_CTRL_MODE_START or > + * - VIRTIO_CAN_SET_CTRL_MODE_STOP. > + * > + * Unlike AUTOSAR CAN Driver Can_SetControllerMode() there is no requirement > + * for this Linux driver to have an asynchronous implementation of the mode > + * setting function so in order to keep things simple the function is > + * implemented as synchronous function. Design pattern is > + * virtio_console.c/__send_control_msg() & > virtio_net.c/virtnet_send_command(). > + */ > +static u8 virtio_can_send_ctrl_msg(struct net_device *ndev, u16 msg_type) > +{ > + struct scatterlist sg_out, sg_in, *sgs[2] = { &sg_out, &sg_in }; > + struct virtio_can_priv *priv = netdev_priv(ndev); > + struct device *dev = &priv->vdev->dev; > + struct virtqueue *vq; > + unsigned int len; > + int err; > + > + vq = priv->vqs[VIRTIO_CAN_QUEUE_CONTROL]; > + > + /* The function may be serialized by rtnl lock. Not sure. > + * Better safe than sorry. > + */ > + mutex_lock(&priv->ctrl_lock); > + > + priv->cpkt_out.msg_type = cpu_to_le16(msg_type); > + sg_init_one(&sg_out, &priv->cpkt_out, sizeof(priv->cpkt_out)); > + sg_init_one(&sg_in, &priv->cpkt_in, sizeof(priv->cpkt_in)); > + > + err = virtqueue_add_sgs(vq, sgs, 1u, 1u, priv, GFP_ATOMIC); > + if (err != 0) { > + /* Not expected to happen */ > + dev_err(dev, "%s(): virtqueue_add_sgs() failed\n", __func__); > + } Here it should return VIRTIO_CAN_RESULT_NOT_OK after unlocking the mutex, or it might wait for completion indefinitley below. > + > + if (!virtqueue_kick(vq)) { > + /* Not expected to happen */ > + dev_err(dev, "%s(): Kick failed\n", __func__); > + } And here too. > + > + while (!virtqueue_get_buf(vq, &len) && !virtqueue_is_broken(vq)) > + wait_for_completion(&priv->ctrl_done); > + > + mutex_unlock(&priv->ctrl_lock); > + > + return priv->cpkt_in.result; > +} > + [...] > +static netdev_tx_t virtio_can_start_xmit(struct sk_buff *skb, > + struct net_device *dev) > +{ > + const unsigned int hdr_size = offsetof(struct virtio_can_tx_out, sdu); > + struct scatterlist sg_out, sg_in, *sgs[2] = { &sg_out, &sg_in }; > + struct canfd_frame *cf = (struct canfd_frame *)skb->data; > + struct virtio_can_priv *priv = netdev_priv(dev); > + netdev_tx_t xmit_ret = NETDEV_TX_OK; > + struct virtio_can_tx *can_tx_msg; > + struct virtqueue *vq; > + unsigned long flags; > + u32 can_flags; > + int putidx; > + int err; > + > + vq = priv->vqs[VIRTIO_CAN_QUEUE_TX]; > + > + if (can_dev_dropped_skb(dev, skb)) > + goto kick; /* No way to return NET_XMIT_DROP here */ > + > + /* No local check for CAN_RTR_FLAG or FD frame against negotiated > + * features. The device will reject those anyway if not supported. > + */ > + > + can_tx_msg = kzalloc(sizeof(*can_tx_msg), GFP_ATOMIC); > + if (!can_tx_msg) { > + dev->stats.tx_dropped++; > + goto kick; /* No way to return NET_XMIT_DROP here */ > + } > + Since we are allocating tx messages dynamically, the sdu[64] array inside struct virtio_can_tx_out can be converted to a flexible array and here the allocation can become: can_tx_msg = kzalloc(sizeof(*can_tx_msg) + cf->len, GFP_ATOMIC); This would save memory in particular on CAN-CC interfaces, where 56 bytes per message would otherwise be lost (not to mention the case if/when CAN-XL will be supported). > + can_tx_msg->tx_out.msg_type = cpu_to_le16(VIRTIO_CAN_TX); > + can_flags = 0; > + > + if (cf->can_id & CAN_EFF_FLAG) { > + can_flags |= VIRTIO_CAN_FLAGS_EXTENDED; > + can_tx_msg->tx_out.can_id = cpu_to_le32(cf->can_id & > CAN_EFF_MASK); > + } else { > + can_tx_msg->tx_out.can_id = cpu_to_le32(cf->can_id & > CAN_SFF_MASK); > + } > + if (cf->can_id & CAN_RTR_FLAG) > + can_flags |= VIRTIO_CAN_FLAGS_RTR; > + else > + memcpy(can_tx_msg->tx_out.sdu, cf->data, cf->len); > + if (can_is_canfd_skb(skb)) > + can_flags |= VIRTIO_CAN_FLAGS_FD; > + > + can_tx_msg->tx_out.flags = cpu_to_le32(can_flags); > + can_tx_msg->tx_out.length = cpu_to_le16(cf->len); > + > + /* Prepare sending of virtio message */ > + sg_init_one(&sg_out, &can_tx_msg->tx_out, hdr_size + cf->len); > + sg_init_one(&sg_in, &can_tx_msg->tx_in, sizeof(can_tx_msg->tx_in)); > + > + putidx = virtio_can_alloc_tx_idx(priv); > + > + if (unlikely(putidx < 0)) { > + /* -ENOMEM or -ENOSPC here. -ENOSPC should not be possible as > + * tx_inflight >= can.echo_skb_max is checked in flow control > + */ > + WARN_ON_ONCE(putidx == -ENOSPC); > + kfree(can_tx_msg); > + dev->stats.tx_dropped++; > + goto kick; /* No way to return NET_XMIT_DROP here */ > + } > + > + can_tx_msg->putidx = (unsigned int)putidx; > + > + /* Protect list operation */ > + spin_lock_irqsave(&priv->tx_lock, flags); > + list_add_tail(&can_tx_msg->list, &priv->tx_list); > + spin_unlock_irqrestore(&priv->tx_lock, flags); > + > + /* Push loopback echo. Will be looped back on TX interrupt/TX NAPI */ > + can_put_echo_skb(skb, dev, can_tx_msg->putidx, 0); > + > + /* Protect queue and list operations */ > + spin_lock_irqsave(&priv->tx_lock, flags); > + err = virtqueue_add_sgs(vq, sgs, 1u, 1u, can_tx_msg, GFP_ATOMIC); > + if (unlikely(err)) { /* checking vq->num_free in flow control */ > + list_del(&can_tx_msg->list); > + can_free_echo_skb(dev, can_tx_msg->putidx, NULL); > + virtio_can_free_tx_idx(priv, can_tx_msg->putidx); > + spin_unlock_irqrestore(&priv->tx_lock, flags); > + netif_stop_queue(dev); > + kfree(can_tx_msg); > + /* Expected never to be seen */ > + netdev_warn(dev, "TX: Stop queue, err = %d\n", err); > + xmit_ret = NETDEV_TX_BUSY; > + goto kick; > + } > + > + /* Normal flow control: stop queue when no transmission slots left */ > + if (atomic_read(&priv->tx_inflight) >= priv->can.echo_skb_max || > + vq->num_free == 0 || (vq->num_free < ARRAY_SIZE(sgs) && > + !virtio_has_feature(vq->vdev, VIRTIO_RING_F_INDIRECT_DESC))) { > + netif_stop_queue(dev); > + netdev_dbg(dev, "TX: Normal stop queue\n"); > + } > + > + spin_unlock_irqrestore(&priv->tx_lock, flags); > + > +kick: > + if (netif_queue_stopped(dev) || !netdev_xmit_more()) { > + if (!virtqueue_kick(vq)) > + netdev_err(dev, "%s(): Kick failed\n", __func__); > + } > + > + return xmit_ret; > +} > + > +static const struct net_device_ops virtio_can_netdev_ops = { > + .ndo_open = virtio_can_open, > + .ndo_stop = virtio_can_close, > + .ndo_start_xmit = virtio_can_start_xmit, > + .ndo_change_mtu = can_change_mtu, > +}; > + > +static int register_virtio_can_dev(struct net_device *dev) > +{ > + dev->flags |= IFF_ECHO; /* we support local echo */ > + dev->netdev_ops = &virtio_can_netdev_ops; > + > + return register_candev(dev); > +} > + > +/* Compare with m_can.c/m_can_echo_tx_event() */ > +static int virtio_can_read_tx_queue(struct virtqueue *vq) > +{ > + struct virtio_can_priv *can_priv = vq->vdev->priv; > + struct net_device *dev = can_priv->dev; > + struct virtio_can_tx *can_tx_msg; > + struct net_device_stats *stats; > + unsigned long flags; > + unsigned int len; > + u8 result; > + > + stats = &dev->stats; > + > + /* Protect list and virtio queue operations */ > + spin_lock_irqsave(&can_priv->tx_lock, flags); > + > + can_tx_msg = virtqueue_get_buf(vq, &len); > + if (!can_tx_msg) { > + spin_unlock_irqrestore(&can_priv->tx_lock, flags); > + return 0; /* No more data */ > + } > + > + if (unlikely(len < sizeof(struct virtio_can_tx_in))) { > + netdev_err(dev, "TX ACK: Device sent no result code\n"); > + result = VIRTIO_CAN_RESULT_NOT_OK; /* Keep things going */ > + } else { > + result = can_tx_msg->tx_in.result; > + } > + > + if (can_priv->can.state < CAN_STATE_BUS_OFF) { > + /* Here also frames with result != VIRTIO_CAN_RESULT_OK are > + * echoed. Intentional to bring a waiting process in an upper > + * layer to an end. > + * TODO: Any better means to indicate a problem here? > + */ > + if (result != VIRTIO_CAN_RESULT_OK) > + netdev_warn(dev, "TX ACK: Result = %u\n", result); Maybe an error frame reporting CAN_ERR_CRTL_UNSPEC would be better? For sure, counting the known errors as valid tx_packets and tx_bytes is misleading. > + > + stats->tx_bytes += can_get_echo_skb(dev, can_tx_msg->putidx, > + NULL); > + stats->tx_packets++; > + } else { > + netdev_dbg(dev, "TX ACK: Controller inactive, drop echo\n"); > + can_free_echo_skb(dev, can_tx_msg->putidx, NULL); > + } > + > + list_del(&can_tx_msg->list); > + virtio_can_free_tx_idx(can_priv, can_tx_msg->putidx); > + > + /* Flow control */ > + if (netif_queue_stopped(dev)) { > + netdev_dbg(dev, "TX ACK: Wake up stopped queue\n"); > + netif_wake_queue(dev); > + } > + > + spin_unlock_irqrestore(&can_priv->tx_lock, flags); > + > + kfree(can_tx_msg); > + > + return 1; /* Queue was not empty so there may be more data */ > +} > + [...] > + > +static int virtio_can_find_vqs(struct virtio_can_priv *priv) > +{ > + /* The order of RX and TX is exactly the opposite as in console and > + * network. Does not play any role but is a bad trap. > + */ > + static const char * const io_names[VIRTIO_CAN_QUEUE_COUNT] = { > + "can-tx", > + "can-rx", > + "can-state-ctrl" > + }; > + > + priv->io_callbacks[VIRTIO_CAN_QUEUE_TX] = virtio_can_tx_intr; > + priv->io_callbacks[VIRTIO_CAN_QUEUE_RX] = virtio_can_rx_intr; > + priv->io_callbacks[VIRTIO_CAN_QUEUE_CONTROL] = virtio_can_control_intr; > + > + /* Find the queues. */ > + return virtio_find_vqs(priv->vdev, VIRTIO_CAN_QUEUE_COUNT, priv->vqs, > + priv->io_callbacks, io_names, NULL); > +} Syntax of virtio_find_vqs changed a bit, here should now be: struct virtqueue_info vqs_info[] = { { "can-tx", virtio_can_tx_intr }, { "can-rx", virtio_can_rx_intr }, { "can-state-ctrl", virtio_can_control_intr }, }; return virtio_find_vqs(priv->vdev, VIRTIO_CAN_QUEUE_COUNT, priv->vqs, vqs_info, NULL); > + > +/* Function must not be called before virtio_can_find_vqs() has been run */ > +static void virtio_can_del_vq(struct virtio_device *vdev) > +{ > + struct virtio_can_priv *priv = vdev->priv; > + struct list_head *cursor, *next; > + struct virtqueue *vq; > + > + /* Reset the device */ > + if (vdev->config->reset) > + vdev->config->reset(vdev); > + > + /* From here we have dead silence from the device side so no locks > + * are needed to protect against device side events. > + */ > + > + vq = priv->vqs[VIRTIO_CAN_QUEUE_CONTROL]; > + while (virtqueue_detach_unused_buf(vq)) > + ; /* Do nothing, content allocated statically */ > + > + vq = priv->vqs[VIRTIO_CAN_QUEUE_RX]; > + while (virtqueue_detach_unused_buf(vq)) > + ; /* Do nothing, content allocated statically */ > + > + vq = priv->vqs[VIRTIO_CAN_QUEUE_TX]; > + while (virtqueue_detach_unused_buf(vq)) > + ; /* Do nothing, content to be de-allocated separately */ > + > + /* Is keeping track of allocated elements by an own linked list > + * really necessary or may this be optimized using only > + * virtqueue_detach_unused_buf()? > + */ > + list_for_each_safe(cursor, next, &priv->tx_list) { > + struct virtio_can_tx *can_tx; > + > + can_tx = list_entry(cursor, struct virtio_can_tx, list); > + list_del(cursor); > + kfree(can_tx); > + } I'd drop the tx_list entirely and rely on virtqueue_detach_unused_buf(); this would allow to remove at least one spinlock save/restore pair at each transmission. > + > + if (vdev->config->del_vqs) > + vdev->config->del_vqs(vdev); > +} > + [...] > diff --git a/include/uapi/linux/virtio_can.h b/include/uapi/linux/virtio_can.h > new file mode 100644 > index 000000000000..7cf613bb3f1a > --- /dev/null > +++ b/include/uapi/linux/virtio_can.h > @@ -0,0 +1,75 @@ > +/* SPDX-License-Identifier: BSD-3-Clause */ > +/* > + * Copyright (C) 2021-2023 OpenSynergy GmbH > + */ > +#ifndef _LINUX_VIRTIO_VIRTIO_CAN_H > +#define _LINUX_VIRTIO_VIRTIO_CAN_H > + > +#include <linux/types.h> > +#include <linux/virtio_types.h> > +#include <linux/virtio_ids.h> > +#include <linux/virtio_config.h> > + > +/* Feature bit numbers */ > +#define VIRTIO_CAN_F_CAN_CLASSIC 0 > +#define VIRTIO_CAN_F_CAN_FD 1 > +#define VIRTIO_CAN_F_LATE_TX_ACK 2 > +#define VIRTIO_CAN_F_RTR_FRAMES 3 > + The values for VIRTIO_CAN_F_LATE_TX_ACK and VIRTIO_CAN_F_RTR_FRAMES are inverted w.r.t. the merged virto-can spec [1]. Note that this is the only deviation from the spec I found. > +/* CAN Result Types */ > +#define VIRTIO_CAN_RESULT_OK 0 > +#define VIRTIO_CAN_RESULT_NOT_OK 1 > + > +/* CAN flags to determine type of CAN Id */ > +#define VIRTIO_CAN_FLAGS_EXTENDED 0x8000 > +#define VIRTIO_CAN_FLAGS_FD 0x4000 > +#define VIRTIO_CAN_FLAGS_RTR 0x2000 > + > +struct virtio_can_config { > +#define VIRTIO_CAN_S_CTRL_BUSOFF (1u << 0) /* Controller BusOff */ > + /* CAN controller status */ > + __le16 status; > +}; > + > +/* TX queue message types */ > +struct virtio_can_tx_out { > +#define VIRTIO_CAN_TX 0x0001 > + __le16 msg_type; > + __le16 length; /* 0..8 CC, 0..64 CAN-FD, 0..2048 CAN-XL, 12 bits */ > + __u8 reserved_classic_dlc; /* If CAN classic length = 8 then DLC can be > 8..15 */ > + __u8 padding; > + __le16 reserved_xl_priority; /* May be needed for CAN XL priority */ > + __le32 flags; > + __le32 can_id; > + __u8 sdu[64]; > +}; > + sdu[] here might be a flexible array, if the driver allocates virtio_can_tx_out structs dyncamically (see above). This would be beneficial in case of CAN-XL frames (if/when they will be supported). > +struct virtio_can_tx_in { > + __u8 result; > +}; > + > +/* RX queue message types */ > +struct virtio_can_rx { > +#define VIRTIO_CAN_RX 0x0101 > + __le16 msg_type; > + __le16 length; /* 0..8 CC, 0..64 CAN-FD, 0..2048 CAN-XL, 12 bits */ > + __u8 reserved_classic_dlc; /* If CAN classic length = 8 then DLC can be > 8..15 */ > + __u8 padding; > + __le16 reserved_xl_priority; /* May be needed for CAN XL priority */ > + __le32 flags; > + __le32 can_id; > + __u8 sdu[64]; > +}; > + Again, sdu[] might be a flexible array. > +/* Control queue message types */ > +struct virtio_can_control_out { > +#define VIRTIO_CAN_SET_CTRL_MODE_START 0x0201 > +#define VIRTIO_CAN_SET_CTRL_MODE_STOP 0x0202 > + __le16 msg_type; > +}; > + > +struct virtio_can_control_in { > + __u8 result; > +}; > + > +#endif /* #ifndef _LINUX_VIRTIO_VIRTIO_CAN_H */ > Thank you for your work! Regards, Francesco [1] https://github.com/oasis-tcs/virtio-spec/blob/virtio-1.4/device-types/can/description.tex#L45