From: xiongweimin <xiongwei...@kylinos.cn> Add initial support for vhost-user-rdma device model for qemu, this includes the interaction(e.g feature negotiation, setting vring baseaddr, setting device status) between guest kernel and host vhost-user-rdma backend device
Signed-off-by: xiongweimin <xiongwei...@kylinos.cn> --- hw/rdma/vhost-user-rdma.c | 466 ++++++++++++++++++++ hw/virtio/meson.build | 1 + hw/virtio/vhost-user-rdma-pci.c | 102 +++++ hw/virtio/virtio.c | 1 + include/hw/pci/pci.h | 1 + include/hw/virtio/vhost-user-rdma.h | 45 ++ include/standard-headers/linux/virtio_ids.h | 1 + include/standard-headers/rdma/virtio_rdma.h | 60 +++ 8 files changed, 677 insertions(+) create mode 100644 hw/rdma/vhost-user-rdma.c create mode 100644 hw/virtio/vhost-user-rdma-pci.c create mode 100644 include/hw/virtio/vhost-user-rdma.h create mode 100644 include/standard-headers/rdma/virtio_rdma.h diff --git a/hw/rdma/vhost-user-rdma.c b/hw/rdma/vhost-user-rdma.c new file mode 100644 index 000000000..b87ba6941 --- /dev/null +++ b/hw/rdma/vhost-user-rdma.c @@ -0,0 +1,466 @@ +/* + * vhost-user-rdma.c + * + * Copyright (c) 2025 Weimin Xiong <xiongwei...@kylinos.cn> + * + * This program is free software; you can redistribute it and/or modify + * it under the terms of the GNU General Public License as published by + * the Free Software Foundation; either version 2 of the License, or + * (at your option) any later version. + * + * This program is distributed in the hope that it will be useful, + * but WITHOUT ANY WARRANTY; without even the implied warranty of + * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the + * GNU General Public License for more details. + * + * You should have received a copy of the GNU General Public License along + * with this program; if not, see <http://www.gnu.org/licenses/>. + * + * SPDX-License-Identifier: GPL-2.0-or-later + * + */ + +#include "qemu/osdep.h" +#include "qapi/error.h" +#include "qemu/error-report.h" +#include "qemu/cutils.h" +#include "hw/qdev-core.h" +#include "hw/qdev-properties.h" +#include "hw/virtio/vhost.h" +#include "hw/virtio/vhost-user-rdma.h" +#include "hw/virtio/virtio.h" +#include "hw/virtio/virtio-bus.h" +#include "hw/virtio/virtio-access.h" +#include "sysemu/sysemu.h" +#include "sysemu/runstate.h" + +#define VHOST_USER_RDMA_NUM_QUEUES 256 +#define VHOST_USER_RDMA_QUEUE_SIZE 512 + +static const int user_feature_bits[] = { + VIRTIO_F_VERSION_1, + VIRTIO_RING_F_INDIRECT_DESC, + VIRTIO_RING_F_EVENT_IDX, + VIRTIO_F_NOTIFY_ON_EMPTY, + VHOST_INVALID_FEATURE_BIT +}; + +static int vhost_user_rdma_start(VirtIODevice *vdev) +{ + VhostUserRdma *r = VHOST_USER_RDMA(vdev); + BusState *qbus = BUS(qdev_get_parent_bus(DEVICE(vdev))); + VirtioBusClass *k = VIRTIO_BUS_GET_CLASS(qbus); + int i, ret; + + if (!k->set_guest_notifiers) { + error_report("binding does not support guest notifiers"); + return -ENOSYS; + } + + ret = vhost_dev_enable_notifiers(&r->dev, vdev); + if (ret < 0) { + error_report("Error enabling host notifiers: %d", -ret); + return ret; + } + + ret = k->set_guest_notifiers(qbus->parent, r->dev.nvqs, true); + if (ret < 0) { + error_report("Error binding guest notifier: %d", -ret); + goto err_host_notifiers; + } + + r->dev.acked_features = vdev->guest_features; + + ret = vhost_dev_start(&r->dev, vdev, true); + if (ret < 0) { + error_report("Error starting vhost: %d", -ret); + goto err_guest_notifiers; + } + r->started_vu = true; + + for (i = 0; i < r->dev.nvqs; i++) { + vhost_virtqueue_mask(&r->dev, vdev, i, false); + } + + return ret; + +err_guest_notifiers: + k->set_guest_notifiers(qbus->parent, r->dev.nvqs, false); +err_host_notifiers: + vhost_dev_disable_notifiers(&r->dev, vdev); + return ret; +} + +static void vhost_user_rdma_stop(VirtIODevice *vdev) +{ + VhostUserRdma *r = VHOST_USER_RDMA(vdev); + BusState *qbus = BUS(qdev_get_parent_bus(DEVICE(vdev))); + VirtioBusClass *k = VIRTIO_BUS_GET_CLASS(qbus); + int ret; + + if (!r->started_vu) { + return; + } + r->started_vu = false; + + if (!k->set_guest_notifiers) { + return; + } + + vhost_dev_stop(&r->dev, vdev, true); + + ret = k->set_guest_notifiers(qbus->parent, r->dev.nvqs, false); + if (ret < 0) { + error_report("vhost guest notifier cleanup failed: %d", ret); + return; + } + + vhost_dev_disable_notifiers(&r->dev, vdev); +} + +static int vhost_user_rdma_handle_config_change(struct vhost_dev *dev) +{ + int ret; + VhostUserRdma *r = VHOST_USER_RDMA(dev->vdev); + Error *local_err = NULL; + + ret = vhost_dev_get_config(dev, (uint8_t *)&r->rdma_backend_cfg, + sizeof(struct virtio_rdma_config), &local_err); + if (ret < 0) { + error_report("get config space failed"); + return -1; + } + + virtio_notify_config(dev->vdev); + return 0; +} + +const VhostDevConfigOps rdma_ops = { + .vhost_dev_config_notifier = vhost_user_rdma_handle_config_change, +}; + +static int vhost_user_rdma_connect(DeviceState *dev) +{ + VirtIODevice *vdev = VIRTIO_DEVICE(dev); + VhostUserRdma *r = VHOST_USER_RDMA(vdev); + int ret = 0; + Error *local_err = NULL; + + if (r->connected) { + return 0; + } + r->connected = true; + + r->dev.nvqs = r->num_queues; + r->dev.vqs = r->vhost_vqs; + r->dev.vq_index = 0; + r->dev.backend_features = 0; + + vhost_dev_set_config_notifier(&r->dev, &rdma_ops); + + ret = vhost_dev_init(&r->dev, &r->vu_state, + VHOST_BACKEND_TYPE_USER, 0, &local_err); + if (ret < 0) { + error_report("vhost-user-rdma: vhost initialization failed: %s", + strerror(-ret)); + return ret; + } + + /* restore vhost state */ + if (virtio_device_started(vdev, vdev->status)) { + info_report("vhost_user_rdma: vhost ss?"); + ret = vhost_user_rdma_start(vdev); + if (ret < 0) { + error_report("vhost-user-rdma: vhost start failed: %s", + strerror(-ret)); + return ret; + } + } + + return 0; +} + +static void vhost_user_rdma_disconnect(DeviceState *dev) +{ + VirtIODevice *vdev = VIRTIO_DEVICE(dev); + VhostUserRdma *s = VHOST_USER_RDMA(vdev); + + if (!s->connected) { + return; + } + s->connected = false; + + vhost_user_rdma_stop(vdev); + + vhost_dev_cleanup(&s->dev); +} + +static void vhost_user_rdma_event(void *opaque, QEMUChrEvent event); + +static void vhost_user_rdma_chr_closed_bh(void *opaque) +{ + DeviceState *dev = opaque; + VirtIODevice *vdev = VIRTIO_DEVICE(dev); + VhostUserRdma *r = VHOST_USER_RDMA(vdev); + + vhost_user_rdma_disconnect(dev); + qemu_chr_fe_set_handlers(&r->chardev, NULL, NULL, vhost_user_rdma_event, + NULL, opaque, NULL, true); +} + +static void vhost_user_rdma_event(void *opaque, QEMUChrEvent event) +{ + DeviceState *dev = opaque; + VirtIODevice *vdev = VIRTIO_DEVICE(dev); + VhostUserRdma *r = VHOST_USER_RDMA(vdev); + + switch (event) { + case CHR_EVENT_OPENED: + if (vhost_user_rdma_connect(dev) < 0) { + qemu_chr_fe_disconnect(&r->chardev); + return; + } + break; + case CHR_EVENT_CLOSED: + if (runstate_is_running()) { + AioContext *ctx = qemu_get_current_aio_context(); + + qemu_chr_fe_set_handlers(&r->chardev, NULL, NULL, NULL, NULL, + NULL, NULL, false); + aio_bh_schedule_oneshot(ctx, vhost_user_rdma_chr_closed_bh, opaque); + } + + r->dev.started = false; + break; + case CHR_EVENT_BREAK: + case CHR_EVENT_MUX_IN: + case CHR_EVENT_MUX_OUT: + /* Ignore */ + break; + } +} + +static void vhost_user_rdma_handle_output(VirtIODevice *vdev, VirtQueue *vq) +{ + VhostUserRdma *r = VHOST_USER_RDMA(vdev); + int i, ret; + + if (!vdev->start_on_kick) { + return; + } + + if (!r->connected) { + return; + } + + if (r->dev.started) { + return; + } + + ret = vhost_user_rdma_start(vdev); + if (ret < 0) { + error_report("vhost-user-rdma: vhost start failed: %s", + strerror(-ret)); + qemu_chr_fe_disconnect(&r->chardev); + return; + } + + for (i = 0; i < r->dev.nvqs; i++) { + VirtQueue *kick_vq = virtio_get_queue(vdev, i); + + if (!virtio_queue_get_desc_addr(vdev, i)) { + continue; + } + event_notifier_set(virtio_queue_get_host_notifier(kick_vq)); + } +} + +static void vhost_user_rdma_update_config(VirtIODevice *vdev, uint8_t *config) +{ + VhostUserRdma *r = VHOST_USER_RDMA(vdev); + + memcpy(config, &r->rdma_backend_cfg, sizeof(struct virtio_rdma_config)); +} + +static uint64_t vhost_user_rdma_get_features(VirtIODevice *vdev, + uint64_t features, + Error **errp) +{ + VhostUserRdma *s = VHOST_USER_RDMA(vdev); + + return vhost_get_features(&s->dev, user_feature_bits, features); +} + +static void vhost_user_rdma_set_status(VirtIODevice *vdev, uint8_t status) +{ + VhostUserRdma *r = VHOST_USER_RDMA(vdev); + bool should_start = virtio_device_started(vdev, status); + int ret; + + if (!vdev->vm_running) { + should_start = false; + } + + if (!r->connected) { + return; + } + + if (r->dev.started == should_start) { + return; + } + + if (should_start) { + ret = vhost_user_rdma_start(vdev); + if (ret < 0) { + error_report("vhost-user-rdma: vhost start failed: %s", + strerror(-ret)); + qemu_chr_fe_disconnect(&r->chardev); + } + } else { + vhost_user_rdma_stop(vdev); + } +} + +static void vhost_user_rdma_device_realize(DeviceState *dev, Error **errp) +{ + VirtIODevice *vdev = VIRTIO_DEVICE(dev); + VhostUserRdma *r = VHOST_USER_RDMA(vdev); + Error *err = NULL; + int i, ret; + + if (!r->chardev.chr) { + error_setg(errp, "vhost-user-rdma: chardev is mandatory"); + return; + } + + r->num_queues = VHOST_USER_RDMA_NUM_QUEUES; + + if (r->num_queues > VIRTIO_QUEUE_MAX) { + error_setg(errp, "vhost-user-rdma: invalid number of IO queues"); + return; + } + + if (!vhost_user_init(&r->vu_state, &r->chardev, errp)) { + return; + } + + virtio_init(vdev, VIRTIO_ID_RDMA, sizeof(struct virtio_rdma_config)); + + r->virtqs = g_new(VirtQueue *, r->num_queues); + + for (i = 0; i < r->num_queues; i++) { + r->virtqs[i] = virtio_add_queue(vdev, VHOST_USER_RDMA_QUEUE_SIZE, + vhost_user_rdma_handle_output); + } + + r->vhost_vqs = g_new0(struct vhost_virtqueue, r->num_queues); + r->connected = false; + + qemu_chr_fe_set_handlers(&r->chardev, NULL, NULL, vhost_user_rdma_event, + NULL, (void *)dev, NULL, true); + +reconnect: + if (qemu_chr_fe_wait_connected(&r->chardev, &err) < 0) { + error_report_err(err); + goto virtio_err; + } + + /* check whether vhost_user_rdma_connect() failed or not */ + if (!r->connected) { + goto reconnect; + } + + ret = vhost_dev_get_config(&r->dev, (uint8_t *)&r->rdma_backend_cfg, + sizeof(struct virtio_rdma_config), &err); + if (ret < 0) { + error_report("vhost-user-rdma: get rdma config failed"); + goto reconnect; + } + + return; + +virtio_err: + g_free(r->vhost_vqs); + r->vhost_vqs = NULL; + for (i = 0; i < r->num_queues; i++) { + virtio_delete_queue(r->virtqs[i]); + } + g_free(r->virtqs); + virtio_cleanup(vdev); + vhost_user_cleanup(&r->vu_state); +} + +static void vhost_user_rdma_device_unrealize(DeviceState *dev) +{ + VirtIODevice *vdev = VIRTIO_DEVICE(dev); + VhostUserRdma *r = VHOST_USER_RDMA(dev); + int i; + + virtio_set_status(vdev, 0); + qemu_chr_fe_set_handlers(&r->chardev, NULL, NULL, NULL, + NULL, NULL, NULL, false); + vhost_dev_cleanup(&r->dev); + g_free(r->vhost_vqs); + r->vhost_vqs = NULL; + for (i = 0; i < r->num_queues; i++) { + virtio_delete_queue(r->virtqs[i]); + } + g_free(r->virtqs); + virtio_cleanup(vdev); + vhost_user_cleanup(&r->vu_state); +} + +static void vhost_user_rdma_instance_init(Object *obj) +{ + VhostUserRdma *r = VHOST_USER_RDMA(obj); + + device_add_bootindex_property(obj, &r->bootindex, "bootindex", + "bootindex", DEVICE(obj)); +} + +static const VMStateDescription vmstate_vhost_user_rdma = { + .name = "vhost-user-rdma", + .minimum_version_id = 1, + .version_id = 1, + .fields = (VMStateField[]) { + VMSTATE_VIRTIO_DEVICE, + VMSTATE_END_OF_LIST() + }, +}; + +static Property vhost_user_rdma_properties[] = { + DEFINE_PROP_CHR("chardev", VhostUserRdma, chardev), + DEFINE_PROP_END_OF_LIST(), +}; + +static void vhost_user_rdma_class_init(ObjectClass *klass, void *data) +{ + DeviceClass *dc = DEVICE_CLASS(klass); + VirtioDeviceClass *vdc = VIRTIO_DEVICE_CLASS(klass); + + device_class_set_props(dc, vhost_user_rdma_properties); + dc->vmsd = &vmstate_vhost_user_rdma; + set_bit(DEVICE_CATEGORY_NETWORK, dc->categories); + + vdc->realize = vhost_user_rdma_device_realize; + vdc->unrealize = vhost_user_rdma_device_unrealize; + vdc->get_config = vhost_user_rdma_update_config; + vdc->set_config = NULL; + vdc->get_features = vhost_user_rdma_get_features; + vdc->set_status = vhost_user_rdma_set_status; +} + +static const TypeInfo vhost_user_rdma_info = { + .name = TYPE_VHOST_USER_RDMA, + .parent = TYPE_VIRTIO_DEVICE, + .instance_size = sizeof(VhostUserRdma), + .instance_init = vhost_user_rdma_instance_init, + .class_init = vhost_user_rdma_class_init, +}; + +static void virtio_register_types(void) +{ + type_register_static(&vhost_user_rdma_info); +} + +type_init(virtio_register_types) diff --git a/hw/virtio/meson.build b/hw/virtio/meson.build index 164f6fd99..84603ac12 100644 --- a/hw/virtio/meson.build +++ b/hw/virtio/meson.build @@ -65,6 +65,7 @@ virtio_pci_ss.add(when: 'CONFIG_VHOST_VSOCK', if_true: files('vhost-vsock-pci.c' virtio_pci_ss.add(when: 'CONFIG_VHOST_USER_VSOCK', if_true: files('vhost-user-vsock-pci.c')) virtio_pci_ss.add(when: 'CONFIG_VHOST_USER_BLK', if_true: files('vhost-user-blk-pci.c')) virtio_pci_ss.add(when: 'CONFIG_VHOST_USER_SCSI', if_true: files('vhost-user-scsi-pci.c')) +virtio_pci_ss.add(when: 'CONFIG_VHOST_USER_RDMA', if_true: files('vhost-user-rdma-pci.c')) virtio_pci_ss.add(when: 'CONFIG_VHOST_SCSI', if_true: files('vhost-scsi-pci.c')) virtio_pci_ss.add(when: 'CONFIG_VHOST_USER_FS', if_true: files('vhost-user-fs-pci.c')) diff --git a/hw/virtio/vhost-user-rdma-pci.c b/hw/virtio/vhost-user-rdma-pci.c new file mode 100644 index 000000000..bbfa2d357 --- /dev/null +++ b/hw/virtio/vhost-user-rdma-pci.c @@ -0,0 +1,102 @@ +/* + * Vhost user rdma PCI Bindings + * + * Copyright(C) 2025 KylinSoft Inc. All rights reserved. + * + * Authors: + * Weimin Xiong <xiongwei...@kylinos.cn> + * + * This work is licensed under the terms of the GNU LGPL, version 2 or later. + * See the COPYING.LIB file in the top-level directory. + * + * SPDX-License-Identifier: GPL-2.0-or-later + * + */ + +#include "qemu/osdep.h" + +#include "standard-headers/rdma/virtio_rdma.h" +#include "hw/virtio/virtio.h" +#include "hw/virtio/vhost-user-rdma.h" +#include "hw/pci/pci.h" +#include "hw/qdev-properties.h" +#include "qapi/error.h" +#include "qemu/error-report.h" +#include "qemu/module.h" +#include "hw/virtio/virtio-pci.h" +#include "qom/object.h" + +typedef struct VhostUserRdmaPCI VhostUserRdmaPCI; + +#define TYPE_VHOST_USER_RDMA_PCI "vhost-user-rdma-pci" +DECLARE_INSTANCE_CHECKER(VhostUserRdmaPCI, VHOST_USER_RDMA_PCI, + TYPE_VHOST_USER_RDMA_PCI) + +struct VhostUserRdmaPCI { + VirtIOPCIProxy parent_obj; + VhostUserRdma vdev; +}; + +static Property vhost_user_rdma_pci_properties[] = { + DEFINE_PROP_UINT32("class", VirtIOPCIProxy, class_code, 0), + DEFINE_PROP_UINT32("vectors", VirtIOPCIProxy, nvectors, + DEV_NVECTORS_UNSPECIFIED), + DEFINE_PROP_END_OF_LIST(), +}; + +static void vhost_user_rdma_pci_realize(VirtIOPCIProxy *vpci_dev, Error **errp) +{ + VhostUserRdmaPCI *dev = VHOST_USER_RDMA_PCI(vpci_dev); + DeviceState *vdev = DEVICE(&dev->vdev); + + if (vpci_dev->nvectors == DEV_NVECTORS_UNSPECIFIED) { + vpci_dev->nvectors = dev->vdev.num_queues + 1; + } + + virtio_pci_force_virtio_1(vpci_dev); + + qdev_realize(vdev, BUS(&vpci_dev->bus), errp); +} + +static void vhost_user_rdma_pci_class_init(ObjectClass *klass, void *data) +{ + DeviceClass *dc = DEVICE_CLASS(klass); + VirtioPCIClass *k = VIRTIO_PCI_CLASS(klass); + PCIDeviceClass *k_pcidev = PCI_DEVICE_CLASS(klass); + + set_bit(DEVICE_CATEGORY_NETWORK, dc->categories); + device_class_set_props(dc, vhost_user_rdma_pci_properties); + k->realize = vhost_user_rdma_pci_realize; + k_pcidev->vendor_id = PCI_VENDOR_ID_REDHAT_QUMRANET; + k_pcidev->device_id = PCI_DEVICE_ID_VIRTIO_RDMA; + k_pcidev->revision = VIRTIO_PCI_ABI_VERSION; + k_pcidev->class_id = PCI_CLASS_NETWORK_OTHER; +} + +static void vhost_user_rdma_pci_instance_init(Object *obj) +{ + VhostUserRdmaPCI *dev = VHOST_USER_RDMA_PCI(obj); + + virtio_instance_init_common(obj, &dev->vdev, sizeof(dev->vdev), + TYPE_VHOST_USER_RDMA); + + object_property_add_alias(obj, "bootindex", OBJECT(&dev->vdev), + "bootindex"); +} + +static const VirtioPCIDeviceTypeInfo vhost_user_rdma_pci_info = { + .base_name = TYPE_VHOST_USER_RDMA_PCI, + .generic_name = "vhost-user-rdma-pci", + .transitional_name = "vhost-user-rdma-pci-transitional", + .non_transitional_name = "vhost-user-rdma-pci-non-transitional", + .instance_size = sizeof(VhostUserRdmaPCI), + .instance_init = vhost_user_rdma_pci_instance_init, + .class_init = vhost_user_rdma_pci_class_init, +}; + +static void vhost_user_rdma_pci_register(void) +{ + virtio_pci_types_register(&vhost_user_rdma_pci_info); +} + +type_init(vhost_user_rdma_pci_register) diff --git a/hw/virtio/virtio.c b/hw/virtio/virtio.c index 85110bce3..277fc9637 100644 --- a/hw/virtio/virtio.c +++ b/hw/virtio/virtio.c @@ -182,6 +182,7 @@ const char *virtio_device_names[] = { [VIRTIO_ID_FS] = "virtio-user-fs", [VIRTIO_ID_PMEM] = "virtio-pmem", [VIRTIO_ID_RPMB] = "virtio-rpmb", + [VIRTIO_ID_RDMA] = "virtio-rdma", [VIRTIO_ID_MAC80211_HWSIM] = "virtio-mac-hwsim", [VIRTIO_ID_VIDEO_ENCODER] = "virtio-vid-encoder", [VIRTIO_ID_VIDEO_DECODER] = "virtio-vid-decoder", diff --git a/include/hw/pci/pci.h b/include/hw/pci/pci.h index 822fbacdf..776fccb65 100644 --- a/include/hw/pci/pci.h +++ b/include/hw/pci/pci.h @@ -90,6 +90,7 @@ extern bool pci_available; #define PCI_DEVICE_ID_VIRTIO_RNG 0x1005 #define PCI_DEVICE_ID_VIRTIO_9P 0x1009 #define PCI_DEVICE_ID_VIRTIO_VSOCK 0x1012 +#define PCI_DEVICE_ID_VIRTIO_RDMA 0x1016 /* * modern virtio-pci devices get their id assigned automatically, diff --git a/include/hw/virtio/vhost-user-rdma.h b/include/hw/virtio/vhost-user-rdma.h new file mode 100644 index 000000000..89a9dddae --- /dev/null +++ b/include/hw/virtio/vhost-user-rdma.h @@ -0,0 +1,45 @@ +/* + * vhost-user-rdma host device + * Copyright(C) 2025 KylinSoft Inc. All rights reserved. + * + * Authors: + * Weimin Xiong <xiongwei...@kylinos.cn> + * + * This work is licensed under the terms of the GNU LGPL, version 2 or later. + * See the COPYING.LIB file in the top-level directory. + * + * SPDX-License-Identifier: GPL-2.0-or-later + * + */ + +#ifndef VHOST_USER_RDMA_H +#define VHOST_USER_RDMA_H + +#include "standard-headers/rdma/virtio_rdma.h" +#include "chardev/char-fe.h" +#include "hw/virtio/vhost.h" +#include "hw/virtio/vhost-user.h" +#include "qom/object.h" + +#define TYPE_VHOST_USER_RDMA "vhost-user-rdma" +OBJECT_DECLARE_SIMPLE_TYPE(VhostUserRdma, VHOST_USER_RDMA) + +struct VhostUserRdma { + VirtIODevice parent_obj; + CharBackend chardev; + int32_t bootindex; + struct virtio_rdma_config rdma_backend_cfg; + struct vhost_dev dev; + VhostUserState vu_state; + struct vhost_virtqueue *vhost_vqs; + VirtQueue **virtqs; + + int num_queues; + + /* vhost_user_rdma_connect/vhost_user_rdma_disconnect */ + bool connected; + /* vhost_user_rdma_start/vhost_user_rdma_stop */ + bool started_vu; +}; + +#endif diff --git a/include/standard-headers/linux/virtio_ids.h b/include/standard-headers/linux/virtio_ids.h index 7aa2eb766..ff2d0b01b 100644 --- a/include/standard-headers/linux/virtio_ids.h +++ b/include/standard-headers/linux/virtio_ids.h @@ -68,6 +68,7 @@ #define VIRTIO_ID_AUDIO_POLICY 39 /* virtio audio policy */ #define VIRTIO_ID_BT 40 /* virtio bluetooth */ #define VIRTIO_ID_GPIO 41 /* virtio gpio */ +#define VIRTIO_ID_RDMA 42 /* virtio rdma */ /* * Virtio Transitional IDs diff --git a/include/standard-headers/rdma/virtio_rdma.h b/include/standard-headers/rdma/virtio_rdma.h new file mode 100644 index 000000000..01823bfa9 --- /dev/null +++ b/include/standard-headers/rdma/virtio_rdma.h @@ -0,0 +1,60 @@ +/* + * Virtio RDMA Device + * + * Copyright (C) 2025 KylinSoft Inc. + * + * Authors: + * Weimin Xiong <xiongwei...@kylinos.cn> + * + * This work is licensed under the terms of the GNU GPL, version 2. See + * the COPYING file in the top-level directory. + * + */ + +#ifndef _LINUX_VIRTIO_RDMA_H +#define _LINUX_VIRTIO_RDMA_H + +#include <linux/types.h> +#include <infiniband/verbs.h> + +#include "standard-headers/linux/virtio_ids.h" +#include "standard-headers/linux/virtio_config.h" +#include "standard-headers/linux/virtio_types.h" + +struct virtio_rdma_config { + __le32 phys_port_cnt; + + __le64 sys_image_guid; + __le32 vendor_id; + __le32 vendor_part_id; + __le32 hw_ver; + __le64 max_mr_size; + __le64 page_size_cap; + __le32 max_qp; + __le32 max_qp_wr; + __le64 device_cap_flags; + __le32 max_send_sge; + __le32 max_recv_sge; + __le32 max_sge_rd; + __le32 max_cq; + __le32 max_cqe; + __le32 max_mr; + __le32 max_pd; + __le32 max_qp_rd_atom; + __le32 max_res_rd_atom; + __le32 max_qp_init_rd_atom; + __le32 atomic_cap; + __le32 max_mw; + __le32 max_mcast_grp; + __le32 max_mcast_qp_attach; + __le32 max_total_mcast_qp_attach; + __le32 max_ah; + __le32 max_fast_reg_page_list_len; + __le32 max_pi_fast_reg_page_list_len; + __le16 max_pkeys; + uint8_t local_ca_ack_delay; + + uint8_t reserved[64]; +} QEMU_PACKED; + +#endif -- 2.41.0