Message ID | 20200508163218.22592-6-lulu@redhat.com |
---|---|
State | New |
Headers | show |
Series | vDPA support in qemu | expand |
On 2020/5/9 上午12:32, Cindy Lu wrote: > From: Tiwei Bie <tiwei.bie@intel.com> > > Currently we have 2 types of vhost backends in QEMU: vhost kernel and > vhost-user. The above patch provides a generic device for vDPA purpose, > this vDPA device exposes to user space a non-vendor-specific configuration > interface for setting up a vhost HW accelerator, this patch set introduces > a third vhost backend called vhost-vdpa based on the vDPA interface. > > Vhost-vdpa usage: > > qemu-system-x86_64 -cpu host -enable-kvm \ > ...... > -netdev type=vhost-vdpa,vhostdev=/dev/vhost-vdpa-id,id=vhost-vdpa0 \ > -device virtio-net-pci,netdev=vhost-vdpa0,page-per-vq=on \ > > Co-Authored-By: Lingshan zhu <lingshan.zhu@intel.com> > Signed-off-by: Cindy Lu <lulu@redhat.com> Signed-off-by: Jason Wang <jasowang@redhat.com> > --- > hw/net/vhost_net.c | 39 ++- > hw/virtio/Makefile.objs | 1 + > hw/virtio/vhost-backend.c | 5 + > hw/virtio/vhost-vdpa.c | 447 ++++++++++++++++++++++++++++++ > hw/virtio/vhost.c | 14 + > include/hw/virtio/vhost-backend.h | 8 +- > include/hw/virtio/vhost-vdpa.h | 25 ++ > include/hw/virtio/vhost.h | 1 + > 8 files changed, 538 insertions(+), 2 deletions(-) > create mode 100644 hw/virtio/vhost-vdpa.c > create mode 100644 include/hw/virtio/vhost-vdpa.h > > diff --git a/hw/net/vhost_net.c b/hw/net/vhost_net.c > index 63b2a85d6e..1af39abaf3 100644 > --- a/hw/net/vhost_net.c > +++ b/hw/net/vhost_net.c > @@ -17,8 +17,10 @@ > #include "net/net.h" > #include "net/tap.h" > #include "net/vhost-user.h" > +#include "net/vhost-vdpa.h" > > #include "standard-headers/linux/vhost_types.h" > +#include "linux-headers/linux/vhost.h" > #include "hw/virtio/virtio-net.h" > #include "net/vhost_net.h" > #include "qemu/error-report.h" > @@ -85,6 +87,29 @@ static const int user_feature_bits[] = { > VHOST_INVALID_FEATURE_BIT > }; > > +static const int vdpa_feature_bits[] = { > + VIRTIO_F_NOTIFY_ON_EMPTY, > + VIRTIO_RING_F_INDIRECT_DESC, > + VIRTIO_RING_F_EVENT_IDX, > + VIRTIO_F_ANY_LAYOUT, > + VIRTIO_F_VERSION_1, > + VIRTIO_NET_F_CSUM, > + VIRTIO_NET_F_GUEST_CSUM, > + VIRTIO_NET_F_GSO, > + VIRTIO_NET_F_GUEST_TSO4, > + VIRTIO_NET_F_GUEST_TSO6, > + VIRTIO_NET_F_GUEST_ECN, > + VIRTIO_NET_F_GUEST_UFO, > + VIRTIO_NET_F_HOST_TSO4, > + VIRTIO_NET_F_HOST_TSO6, > + VIRTIO_NET_F_HOST_ECN, > + VIRTIO_NET_F_HOST_UFO, > + VIRTIO_NET_F_MRG_RXBUF, > + VIRTIO_NET_F_MTU, > + VIRTIO_F_IOMMU_PLATFORM, > + VIRTIO_NET_F_GUEST_ANNOUNCE, > + VHOST_INVALID_FEATURE_BIT > +}; The framework should be ready for packed ring, let's add that. > static const int *vhost_net_get_feature_bits(struct vhost_net *net) > { > const int *feature_bits = 0; > @@ -96,6 +121,9 @@ static const int *vhost_net_get_feature_bits(struct vhost_net *net) > case NET_CLIENT_DRIVER_VHOST_USER: > feature_bits = user_feature_bits; > break; > + case NET_CLIENT_DRIVER_VHOST_VDPA: > + feature_bits = vdpa_feature_bits; > + break; > default: > error_report("Feature bits not defined for this type: %d", > net->nc->info->type); > @@ -110,7 +138,10 @@ uint64_t vhost_net_get_features(struct vhost_net *net, uint64_t features) > return vhost_get_features(&net->dev, vhost_net_get_feature_bits(net), > features); > } > - > +int vhost_net_get_device_id(struct vhost_net *net, uint32_t * device_id) > +{ > + return vhost_dev_get_device_id(&net->dev, device_id); > +} > void vhost_net_ack_features(struct vhost_net *net, uint64_t features) > { > net->dev.acked_features = net->dev.backend_features; > @@ -433,6 +464,12 @@ VHostNetState *get_vhost_net(NetClientState *nc) > vhost_net = vhost_user_get_vhost_net(nc); > assert(vhost_net); > break; > +#endif > +#ifdef CONFIG_VHOST_NET_VDPA > + case NET_CLIENT_DRIVER_VHOST_VDPA: > + vhost_net = vhost_vdpa_get_vhost_net(nc); > + assert(vhost_net); > + break; > #endif > default: > break; > diff --git a/hw/virtio/Makefile.objs b/hw/virtio/Makefile.objs > index e2f70fbb89..e7c5d4a862 100644 > --- a/hw/virtio/Makefile.objs > +++ b/hw/virtio/Makefile.objs > @@ -5,6 +5,7 @@ obj-y += virtio.o > obj-$(call lor,$(CONFIG_VHOST_USER),$(CONFIG_VHOST_KERNEL)) += vhost.o vhost-backend.o > common-obj-$(call lnot,$(call lor,$(CONFIG_VHOST_USER),$(CONFIG_VHOST_KERNEL))) += vhost-stub.o > obj-$(CONFIG_VHOST_USER) += vhost-user.o > +obj-$(CONFIG_VHOST_VDPA) += vhost-vdpa.o > > common-obj-$(CONFIG_VIRTIO_RNG) += virtio-rng.o > common-obj-$(CONFIG_VIRTIO_PCI) += virtio-pci.o > diff --git a/hw/virtio/vhost-backend.c b/hw/virtio/vhost-backend.c > index 48905383f8..069ddb423d 100644 > --- a/hw/virtio/vhost-backend.c > +++ b/hw/virtio/vhost-backend.c > @@ -285,6 +285,11 @@ int vhost_set_backend_type(struct vhost_dev *dev, VhostBackendType backend_type) > case VHOST_BACKEND_TYPE_USER: > dev->vhost_ops = &user_ops; > break; > +#endif > +#ifdef CONFIG_VHOST_VDPA > + case VHOST_BACKEND_TYPE_VDPA: > + dev->vhost_ops = &vdpa_ops; > + break; > #endif > default: > error_report("Unknown vhost backend type"); > diff --git a/hw/virtio/vhost-vdpa.c b/hw/virtio/vhost-vdpa.c > new file mode 100644 > index 0000000000..bac8a8aa2a > --- /dev/null > +++ b/hw/virtio/vhost-vdpa.c > @@ -0,0 +1,447 @@ > +/* > + * vhost-vdpa > + * > + * Copyright(c) 2017-2018 Intel Corporation. > + * Copyright(c) 2020 Red Hat, Inc. > + * > + * This work is licensed under the terms of the GNU GPL, version 2 or later. > + * See the COPYING file in the top-level directory. > + * > + */ > + > +#include "qemu/osdep.h" > +#include <linux/vhost.h> > +#include <linux/vfio.h> > +#include <sys/eventfd.h> > +#include <sys/ioctl.h> > +#include "hw/virtio/vhost.h" > +#include "hw/virtio/vhost-backend.h" > +#include "hw/virtio/virtio-net.h" > +#include "hw/virtio/vhost-vdpa.h" > +#include "qemu/main-loop.h" > +#include <linux/kvm.h> > +#include "sysemu/kvm.h" > + > + > +static bool vhost_vdpa_listener_skipped_section(MemoryRegionSection *section) > +{ > + return (!memory_region_is_ram(section->mr) && > + !memory_region_is_iommu(section->mr)) || > + /* > + * Sizing an enabled 64-bit BAR can cause spurious mappings to > + * addresses in the upper part of the 64-bit address space. These > + * are never accessed by the CPU and beyond the address width of > + * some IOMMU hardware. TODO: VDPA should tell us the IOMMU width. > + */ > + section->offset_within_address_space & (1ULL << 63); > +} > + > +static int vhost_vdpa_dma_map(struct vhost_vdpa *v, hwaddr iova, hwaddr size, > + void *vaddr, bool readonly) > +{ > + struct vhost_msg_v2 msg; > + int fd = v->device_fd; > + int ret = 0; > + > + msg.type = v->msg_type; > + msg.iotlb.iova = iova; > + msg.iotlb.size = size; > + msg.iotlb.uaddr = (uint64_t)vaddr; > + msg.iotlb.perm = readonly ? VHOST_ACCESS_RO : VHOST_ACCESS_RW; > + msg.iotlb.type = VHOST_IOTLB_UPDATE; > + > + if (write(fd, &msg, sizeof(msg)) != sizeof(msg)) { > + error_report("failed to write, fd=%d, errno=%d (%s)", > + fd, errno, strerror(errno)); > + return -EIO ; > + } > + > + return ret; > +} > + > +static int vhost_vdpa_dma_unmap(struct vhost_vdpa *v, hwaddr iova, > + hwaddr size) > +{ > + struct vhost_msg_v2 msg; > + int fd = v->device_fd; > + int ret = 0; > + > + msg.type = v->msg_type; > + msg.iotlb.iova = iova; > + msg.iotlb.size = size; > + msg.iotlb.type = VHOST_IOTLB_INVALIDATE; > + > + if (write(fd, &msg, sizeof(msg)) != sizeof(msg)) { > + error_report("failed to write, fd=%d, errno=%d (%s)", > + fd, errno, strerror(errno)); > + return -EIO ; > + } > + > + return ret; > +} > + > +static void vhost_vdpa_listener_region_add(MemoryListener *listener, > + MemoryRegionSection *section) > +{ > + struct vhost_vdpa *v = container_of(listener, struct vhost_vdpa, listener); > + hwaddr iova; > + Int128 llend, llsize; > + void *vaddr; > + int ret; > + > + if (vhost_vdpa_listener_skipped_section(section)) { > + return; > + } > + > + if (unlikely((section->offset_within_address_space & ~TARGET_PAGE_MASK) != > + (section->offset_within_region & ~TARGET_PAGE_MASK))) { > + error_report("%s received unaligned region", __func__); > + return; > + } > + > + iova = TARGET_PAGE_ALIGN(section->offset_within_address_space); > + llend = int128_make64(section->offset_within_address_space); > + llend = int128_add(llend, section->size); > + llend = int128_and(llend, int128_exts64(TARGET_PAGE_MASK)); > + > + if (int128_ge(int128_make64(iova), llend)) { > + return; > + } > + > + memory_region_ref(section->mr); > + > + /* Here we assume that memory_region_is_ram(section->mr)==true */ > + > + vaddr = memory_region_get_ram_ptr(section->mr) + > + section->offset_within_region + > + (iova - section->offset_within_address_space); > + > + llsize = int128_sub(llend, int128_make64(iova)); > + > + ret = vhost_vdpa_dma_map(v, iova, int128_get64(llsize), > + vaddr, section->readonly); > + if (ret) { > + error_report("vhost vdpa map fail!"); > + if (memory_region_is_ram_device(section->mr)) { > + /* Allow unexpected mappings not to be fatal for RAM devices */ > + error_report("map ram fail!"); > + return ; > + } > + goto fail; > + } > + > + return; > + > +fail: > + if (memory_region_is_ram_device(section->mr)) { > + error_report("failed to vdpa_dma_map. pci p2p may not work"); > + return; > + > + } > + /* > + * On the initfn path, store the first error in the container so we > + * can gracefully fail. Runtime, there's not much we can do other > + * than throw a hardware error. > + */ > + error_report("vhost-vdpa: DMA mapping failed, unable to continue"); > + return; > + > +} > + > +static void vhost_vdpa_listener_region_del(MemoryListener *listener, > + MemoryRegionSection *section) > +{ > + struct vhost_vdpa *v = container_of(listener, struct vhost_vdpa, listener); > + hwaddr iova; > + Int128 llend, llsize; > + int ret; > + bool try_unmap = true; > + > + if (vhost_vdpa_listener_skipped_section(section)) { > + return; > + } > + > + if (unlikely((section->offset_within_address_space & ~TARGET_PAGE_MASK) != > + (section->offset_within_region & ~TARGET_PAGE_MASK))) { > + error_report("%s received unaligned region", __func__); > + return; > + } > + > + iova = TARGET_PAGE_ALIGN(section->offset_within_address_space); > + llend = int128_make64(section->offset_within_address_space); > + llend = int128_add(llend, section->size); > + llend = int128_and(llend, int128_exts64(TARGET_PAGE_MASK)); > + > + if (int128_ge(int128_make64(iova), llend)) { > + return; > + } > + > + llsize = int128_sub(llend, int128_make64(iova)); > + > + if (try_unmap) { > + ret = vhost_vdpa_dma_unmap(v, iova, int128_get64(llsize)); > + if (ret) { > + error_report("vhost_vdpa dma unmap error!"); > + } > + } > + > + memory_region_unref(section->mr); > +} > + > +static const MemoryListener vhost_vdpa_memory_listener = { > + .region_add = vhost_vdpa_listener_region_add, > + .region_del = vhost_vdpa_listener_region_del, > +}; > + Need comment to explain why vhost-vdpa requires a specific memory listener other than just reusing the one that is provided by vhost. > + This newline is unnecessary. > +static int vhost_vdpa_call(struct vhost_dev *dev, unsigned long int request, > + void *arg) > +{ > + struct vhost_vdpa *v = dev->opaque; > + int fd = v->device_fd; > + > + assert(dev->vhost_ops->backend_type == VHOST_BACKEND_TYPE_VDPA); > + > + return ioctl(fd, request, arg); > +} > + > +static int vhost_vdpa_get_backend_features(struct vhost_dev *dev, > + uint64_t *features) > +{ > + return vhost_vdpa_call(dev, VHOST_GET_BACKEND_FEATURES, features); > +} > + > +static int vhost_vdpa_init(struct vhost_dev *dev, void *opaque) > +{ > + struct vhost_vdpa *v; > + uint64_t backend_features = 0; > + int ret = 0; > + > + assert(dev->vhost_ops->backend_type == VHOST_BACKEND_TYPE_VDPA); > + > + v = opaque; > + dev->opaque = opaque ; > + > + v->listener = vhost_vdpa_memory_listener; > + memory_listener_register(&v->listener, &address_space_memory); > + > + ret = vhost_vdpa_get_backend_features(dev, &backend_features); > + /* > + * In order to compatible with older version kernel, > + * If the kernel not support this ioctl, > + * set the msg_type for v2 by defeault > + */ > + if (ret) { > + v->msg_type = VHOST_IOTLB_MSG_V2; > + return 0; > + } > + if (backend_features & (1ULL << VHOST_BACKEND_F_IOTLB_MSG_V2)) { > + v->msg_type = VHOST_IOTLB_MSG_V2; > + } else { > + v->msg_type = VHOST_IOTLB_MSG; > + } But in vhost_vdpa_dma_unmap/vhost_vdpa_dma_map, v2 is still being used. Consider vdpa implies msg_v2, and it doesn't implement get_backend_features. We use v2 unconditionally and leave this check in the future. > + return 0; > +} > + > +static int vhost_vdpa_cleanup(struct vhost_dev *dev) > +{ > + struct vhost_vdpa *v; > + assert(dev->vhost_ops->backend_type == VHOST_BACKEND_TYPE_VDPA); > + > + v = dev->opaque; > + memory_listener_unregister(&v->listener); > + > + dev->opaque = NULL; > + return 0; > +} > + > +static int vhost_vdpa_memslots_limit(struct vhost_dev *dev) > +{ > + /*Using IOTLB API here,NOTE:please set the ulimit before using*/ The "note" seems unrelated to the memslot here. > + return INT_MAX; > +} > + > +static int vhost_vdpa_set_log_base(struct vhost_dev *dev, uint64_t base, > + struct vhost_log *log) > +{ > + return -1; > +} > + > +static int vhost_vdpa_set_mem_table(struct vhost_dev *dev, > + struct vhost_memory *mem) > +{ > + > + if (mem->padding) { > + return -1; > + } > + > + return 0; > +} > + > +static int vhost_vdpa_set_vring_addr(struct vhost_dev *dev, > + struct vhost_vring_addr *addr) > +{ > + return vhost_vdpa_call(dev, VHOST_SET_VRING_ADDR, addr); > +} > + > +static int vhost_vdpa_set_vring_num(struct vhost_dev *dev, > + struct vhost_vring_state *ring) > +{ > + return vhost_vdpa_call(dev, VHOST_SET_VRING_NUM, ring); > +} > + > +static int vhost_vdpa_set_vring_base(struct vhost_dev *dev, > + struct vhost_vring_state *ring) > +{ > + return vhost_vdpa_call(dev, VHOST_GET_VRING_BASE, ring); > +} > + > +static int vhost_vdpa_get_vring_base(struct vhost_dev *dev, > + struct vhost_vring_state *ring) > +{ > + > + return vhost_vdpa_call(dev, VHOST_GET_VRING_BASE, ring); > +} > + > +static int vhost_vdpa_set_vring_kick(struct vhost_dev *dev, > + struct vhost_vring_file *file) > +{ > + return vhost_vdpa_call(dev, VHOST_SET_VRING_KICK, file); > +} > + > +static int vhost_vdpa_set_vring_call(struct vhost_dev *dev, > + struct vhost_vring_file *file) > +{ > + return vhost_vdpa_call(dev, VHOST_SET_VRING_CALL, file); > +} > + > +static int vhost_vdpa_set_features(struct vhost_dev *dev, > + uint64_t features) > +{ > + return vhost_vdpa_call(dev, VHOST_SET_FEATURES, &features); > +} > + > +static int vhost_vdpa_get_features(struct vhost_dev *dev, > + uint64_t *features) > +{ > + return vhost_vdpa_call(dev, VHOST_GET_FEATURES, features); > +} > + > +static int vhost_vdpa_get_device_id(struct vhost_dev *dev, > + uint32_t *device_id) > +{ > + return vhost_vdpa_call(dev, VHOST_VDPA_GET_DEVICE_ID, device_id); > +} > + > +static int vhost_vdpa_set_owner(struct vhost_dev *dev) > +{ > + return vhost_vdpa_call(dev, VHOST_SET_OWNER, NULL); > +} > + > +static int vhost_vdpa_reset_device(struct vhost_dev *dev) > +{ > + return vhost_vdpa_call(dev, VHOST_RESET_OWNER, NULL); > +} I think we may want to reuse the helpers of kenrel backends instead of duplicating functions here. > + > +static int vhost_vdpa_get_vq_index(struct vhost_dev *dev, int idx) > +{ > + assert(idx >= dev->vq_index && idx < dev->vq_index + dev->nvqs); > + > + return idx - dev->vq_index; > +} > + > +static int vhost_vdpa_set_vring_ready(struct vhost_dev *dev) > +{ > + int i; > + for (i = 0; i < dev->nvqs; ++i) { > + struct vhost_vring_state state = { > + .index = dev->vq_index + i, > + .num = 1, > + }; > + vhost_vdpa_call(dev, VHOST_VDPA_SET_VRING_ENABLE, &state); > + } > + return 0; > +} > +static int vhost_vdpa_set_config(struct vhost_dev *dev, const uint8_t *data, One unnecessary space between int and "vhost_vdpa_set_config" > + uint32_t offset, uint32_t size, > + uint32_t flags) > +{ > + > + struct vhost_vdpa_config *config; > + int ret; > + if ((size > VHOST_VDPA_MAX_CONFIG_SIZE) || (data == NULL)) { > + return -1; > + } > + config = g_new0(struct vhost_vdpa_config, 1); It's better to just use a temporary variable instead of allocate it dynamically? > + if (config == NULL) { > + return -EINVAL; > + } > + config->off = 0; > + config->len = size; > + memcpy(config->buf, data, size); > + ret = vhost_vdpa_call(dev, VHOST_VDPA_SET_CONFIG, config); > + g_free(config); > + return ret; > +} > + > + static int vhost_vdpa_get_config(struct vhost_dev *dev, uint8_t *config, > + uint32_t config_len) > +{ > + struct vhost_vdpa_config *v_config = NULL; > + int ret; > + > + if ((config_len > VHOST_VDPA_MAX_CONFIG_SIZE) || (config == NULL)) { > + return -1; > + } > + v_config = g_new0(struct vhost_vdpa_config, 1); Let's use a variable on the stack. > + > + if (v_config == NULL) { > + return -EINVAL; > + } > + ret = vhost_vdpa_call(dev, VHOST_VDPA_GET_CONFIG, v_config); One extra space after '='. > + if ((v_config->len > config_len - v_config->off) || (v_config->len == 0)) { > + g_free(v_config); > + return -EINVAL; > + } Indentation is wired and why need such check? > + > + memcpy(config, v_config->buf + v_config->off, config_len); > + g_free(v_config); > + return ret; > + } Newline is needed here. > +static int vhost_vdpa_set_state(struct vhost_dev *dev, uint8_t state) > +{ > + return vhost_vdpa_call(dev, VHOST_VDPA_SET_STATUS, &state); > +} > + > + This new line is not needed. > +const VhostOps vdpa_ops = { > + .backend_type = VHOST_BACKEND_TYPE_VDPA, > + .vhost_backend_init = vhost_vdpa_init, > + .vhost_backend_cleanup = vhost_vdpa_cleanup, > + .vhost_backend_memslots_limit = vhost_vdpa_memslots_limit, > + .vhost_set_log_base = vhost_vdpa_set_log_base, > + .vhost_set_mem_table = vhost_vdpa_set_mem_table, > + .vhost_set_vring_addr = vhost_vdpa_set_vring_addr, > + .vhost_set_vring_endian = NULL, > + .vhost_set_vring_num = vhost_vdpa_set_vring_num, > + .vhost_set_vring_base = vhost_vdpa_set_vring_base, > + .vhost_get_vring_base = vhost_vdpa_get_vring_base, > + .vhost_set_vring_kick = vhost_vdpa_set_vring_kick, > + .vhost_set_vring_call = vhost_vdpa_set_vring_call, > + .vhost_set_features = vhost_vdpa_set_features, > + .vhost_get_features = vhost_vdpa_get_features, > + .vhost_set_owner = vhost_vdpa_set_owner, > + .vhost_reset_device = vhost_vdpa_reset_device, > + .vhost_get_vq_index = vhost_vdpa_get_vq_index, > + .vhost_set_vring_ready = vhost_vdpa_set_vring_ready, > + .vhost_get_config = vhost_vdpa_get_config, > + .vhost_set_config = vhost_vdpa_set_config, > + .vhost_requires_shm_log = NULL, > + .vhost_migration_done = NULL, > + .vhost_backend_can_merge = NULL, > + .vhost_net_set_mtu = NULL, > + .vhost_set_iotlb_callback = NULL, > + .vhost_send_device_iotlb_msg = NULL, > + .vhost_set_state = vhost_vdpa_set_state, > + .vhost_get_device_id = vhost_vdpa_get_device_id, > +}; As said previously, need to consider to reuse vhost kernel helpers. > diff --git a/hw/virtio/vhost.c b/hw/virtio/vhost.c > index 4da0d5a6c5..6d2aa461d6 100644 > --- a/hw/virtio/vhost.c > +++ b/hw/virtio/vhost.c > @@ -746,6 +746,12 @@ static int vhost_virtqueue_set_addr(struct vhost_dev *dev, > .log_guest_addr = vq->used_phys, > .flags = enable_log ? (1 << VHOST_VRING_F_LOG) : 0, > }; > + /*vDPA need to use the phys address here to set to hardware*/ > + if (dev->vhost_ops->backend_type == VHOST_BACKEND_TYPE_VDPA) { > + addr.desc_user_addr = (uint64_t)(unsigned long)vq->desc_phys; > + addr.avail_user_addr = (uint64_t)(unsigned long)vq->avail_phys; > + addr.used_user_addr = (uint64_t)(unsigned long)vq->used_phys; > + } > int r = dev->vhost_ops->vhost_set_vring_addr(dev, &addr); > if (r < 0) { > VHOST_OPS_DEBUG("vhost_set_vring_addr failed"); > @@ -1493,6 +1499,14 @@ int vhost_dev_set_config(struct vhost_dev *hdev, const uint8_t *data, > return -1; > } > > +int vhost_dev_get_device_id(struct vhost_dev *hdev, uint32_t *device_id) > +{ > + assert(hdev->vhost_ops); > + if (hdev->vhost_ops->vhost_get_device_id) { If you introduce a new vhost_ops, please use a separate patch. > + return hdev->vhost_ops->vhost_get_device_id(hdev, device_id); > + } > + return -1; > +} > void vhost_dev_set_config_notifier(struct vhost_dev *hdev, > const VhostDevConfigOps *ops) > { > diff --git a/include/hw/virtio/vhost-backend.h b/include/hw/virtio/vhost-backend.h > index f823055167..8ac898a987 100644 > --- a/include/hw/virtio/vhost-backend.h > +++ b/include/hw/virtio/vhost-backend.h > @@ -17,7 +17,8 @@ typedef enum VhostBackendType { > VHOST_BACKEND_TYPE_NONE = 0, > VHOST_BACKEND_TYPE_KERNEL = 1, > VHOST_BACKEND_TYPE_USER = 2, > - VHOST_BACKEND_TYPE_MAX = 3, > + VHOST_BACKEND_TYPE_VDPA = 3, > + VHOST_BACKEND_TYPE_MAX = 4, > } VhostBackendType; > > typedef enum VhostSetConfigType { > @@ -77,6 +78,7 @@ typedef int (*vhost_reset_device_op)(struct vhost_dev *dev); > typedef int (*vhost_get_vq_index_op)(struct vhost_dev *dev, int idx); > typedef int (*vhost_set_vring_enable_op)(struct vhost_dev *dev, > int enable); > +typedef int (*vhost_set_vring_ready_op)(struct vhost_dev *dev); A new patch for this patch. Thanks > typedef bool (*vhost_requires_shm_log_op)(struct vhost_dev *dev); > typedef int (*vhost_migration_done_op)(struct vhost_dev *dev, > char *mac_addr); > @@ -113,6 +115,7 @@ typedef int (*vhost_set_inflight_fd_op)(struct vhost_dev *dev, > struct vhost_inflight *inflight); > > typedef int (*vhost_set_state_op)(struct vhost_dev *dev, uint8_t state); > +typedef int (*vhost_get_device_id_op)(struct vhost_dev *dev, uint32_t *dev_id); > typedef struct VhostOps { > VhostBackendType backend_type; > vhost_backend_init vhost_backend_init; > @@ -139,6 +142,7 @@ typedef struct VhostOps { > vhost_reset_device_op vhost_reset_device; > vhost_get_vq_index_op vhost_get_vq_index; > vhost_set_vring_enable_op vhost_set_vring_enable; > + vhost_set_vring_ready_op vhost_set_vring_ready; > vhost_requires_shm_log_op vhost_requires_shm_log; > vhost_migration_done_op vhost_migration_done; > vhost_backend_can_merge_op vhost_backend_can_merge; > @@ -154,9 +158,11 @@ typedef struct VhostOps { > vhost_get_inflight_fd_op vhost_get_inflight_fd; > vhost_set_inflight_fd_op vhost_set_inflight_fd; > vhost_set_state_op vhost_set_state; > + vhost_get_device_id_op vhost_get_device_id; > } VhostOps; > > extern const VhostOps user_ops; > +extern const VhostOps vdpa_ops; > > int vhost_set_backend_type(struct vhost_dev *dev, > VhostBackendType backend_type); > diff --git a/include/hw/virtio/vhost-vdpa.h b/include/hw/virtio/vhost-vdpa.h > new file mode 100644 > index 0000000000..9d64fbfea9 > --- /dev/null > +++ b/include/hw/virtio/vhost-vdpa.h > @@ -0,0 +1,25 @@ > +/* > + * vhost-vdpa.h > + * > + * Copyright(c) 2017-2018 Intel Corporation. > + * Copyright(c) 2020 Red Hat, Inc. > + * > + * This work is licensed under the terms of the GNU GPL, version 2 or later. > + * See the COPYING file in the top-level directory. > + * > + */ > + > +#ifndef HW_VIRTIO_VHOST_VDPA_H > +#define HW_VIRTIO_VHOST_VDPA_H > + > +#include "hw/virtio/virtio.h" > + > +typedef struct vhost_vdpa { > + int device_fd; > + uint32_t msg_type; > + MemoryListener listener; > +} VhostVDPA; > + > +extern AddressSpace address_space_memory; > + > +#endif > diff --git a/include/hw/virtio/vhost.h b/include/hw/virtio/vhost.h > index 085450c6f8..020956d5a9 100644 > --- a/include/hw/virtio/vhost.h > +++ b/include/hw/virtio/vhost.h > @@ -124,6 +124,7 @@ int vhost_dev_get_config(struct vhost_dev *dev, uint8_t *config, > uint32_t config_len); > int vhost_dev_set_config(struct vhost_dev *dev, const uint8_t *data, > uint32_t offset, uint32_t size, uint32_t flags); > +int vhost_dev_get_device_id(struct vhost_dev *hdev, uint32_t *device_id); > /* notifier callback in case vhost device config space changed > */ > void vhost_dev_set_config_notifier(struct vhost_dev *dev,
On 2020/5/9 上午11:00, Jason Wang wrote: > > + .vhost_get_config = vhost_vdpa_get_config, > + .vhost_set_config = vhost_vdpa_set_config, Btw, I don't see the actual user of those two helpers? Thanks
On Sat, May 9, 2020 at 11:00 AM Jason Wang <jasowang@redhat.com> wrote: > > > On 2020/5/9 上午12:32, Cindy Lu wrote: > > From: Tiwei Bie <tiwei.bie@intel.com> > > > > Currently we have 2 types of vhost backends in QEMU: vhost kernel and > > vhost-user. The above patch provides a generic device for vDPA purpose, > > this vDPA device exposes to user space a non-vendor-specific configuration > > interface for setting up a vhost HW accelerator, this patch set introduces > > a third vhost backend called vhost-vdpa based on the vDPA interface. > > > > Vhost-vdpa usage: > > > > qemu-system-x86_64 -cpu host -enable-kvm \ > > ...... > > -netdev type=vhost-vdpa,vhostdev=/dev/vhost-vdpa-id,id=vhost-vdpa0 \ > > -device virtio-net-pci,netdev=vhost-vdpa0,page-per-vq=on \ > > > > Co-Authored-By: Lingshan zhu <lingshan.zhu@intel.com> > > Signed-off-by: Cindy Lu <lulu@redhat.com> > > > Signed-off-by: Jason Wang <jasowang@redhat.com> > > > > --- > > hw/net/vhost_net.c | 39 ++- > > hw/virtio/Makefile.objs | 1 + > > hw/virtio/vhost-backend.c | 5 + > > hw/virtio/vhost-vdpa.c | 447 ++++++++++++++++++++++++++++++ > > hw/virtio/vhost.c | 14 + > > include/hw/virtio/vhost-backend.h | 8 +- > > include/hw/virtio/vhost-vdpa.h | 25 ++ > > include/hw/virtio/vhost.h | 1 + > > 8 files changed, 538 insertions(+), 2 deletions(-) > > create mode 100644 hw/virtio/vhost-vdpa.c > > create mode 100644 include/hw/virtio/vhost-vdpa.h > > > > diff --git a/hw/net/vhost_net.c b/hw/net/vhost_net.c > > index 63b2a85d6e..1af39abaf3 100644 > > --- a/hw/net/vhost_net.c > > +++ b/hw/net/vhost_net.c > > @@ -17,8 +17,10 @@ > > #include "net/net.h" > > #include "net/tap.h" > > #include "net/vhost-user.h" > > +#include "net/vhost-vdpa.h" > > > > #include "standard-headers/linux/vhost_types.h" > > +#include "linux-headers/linux/vhost.h" > > #include "hw/virtio/virtio-net.h" > > #include "net/vhost_net.h" > > #include "qemu/error-report.h" > > @@ -85,6 +87,29 @@ static const int user_feature_bits[] = { > > VHOST_INVALID_FEATURE_BIT > > }; > > > > +static const int vdpa_feature_bits[] = { > > + VIRTIO_F_NOTIFY_ON_EMPTY, > > + VIRTIO_RING_F_INDIRECT_DESC, > > + VIRTIO_RING_F_EVENT_IDX, > > + VIRTIO_F_ANY_LAYOUT, > > + VIRTIO_F_VERSION_1, > > + VIRTIO_NET_F_CSUM, > > + VIRTIO_NET_F_GUEST_CSUM, > > + VIRTIO_NET_F_GSO, > > + VIRTIO_NET_F_GUEST_TSO4, > > + VIRTIO_NET_F_GUEST_TSO6, > > + VIRTIO_NET_F_GUEST_ECN, > > + VIRTIO_NET_F_GUEST_UFO, > > + VIRTIO_NET_F_HOST_TSO4, > > + VIRTIO_NET_F_HOST_TSO6, > > + VIRTIO_NET_F_HOST_ECN, > > + VIRTIO_NET_F_HOST_UFO, > > + VIRTIO_NET_F_MRG_RXBUF, > > + VIRTIO_NET_F_MTU, > > + VIRTIO_F_IOMMU_PLATFORM, > > + VIRTIO_NET_F_GUEST_ANNOUNCE, > > + VHOST_INVALID_FEATURE_BIT > > +}; > > > The framework should be ready for packed ring, let's add that. > Will fix this > > > static const int *vhost_net_get_feature_bits(struct vhost_net *net) > > { > > const int *feature_bits = 0; > > @@ -96,6 +121,9 @@ static const int *vhost_net_get_feature_bits(struct vhost_net *net) > > case NET_CLIENT_DRIVER_VHOST_USER: > > feature_bits = user_feature_bits; > > break; > > + case NET_CLIENT_DRIVER_VHOST_VDPA: > > + feature_bits = vdpa_feature_bits; > > + break; > > default: > > error_report("Feature bits not defined for this type: %d", > > net->nc->info->type); > > @@ -110,7 +138,10 @@ uint64_t vhost_net_get_features(struct vhost_net *net, uint64_t features) > > return vhost_get_features(&net->dev, vhost_net_get_feature_bits(net), > > features); > > } > > - > > +int vhost_net_get_device_id(struct vhost_net *net, uint32_t * device_id) > > +{ > > + return vhost_dev_get_device_id(&net->dev, device_id); > > +} > > void vhost_net_ack_features(struct vhost_net *net, uint64_t features) > > { > > net->dev.acked_features = net->dev.backend_features; > > @@ -433,6 +464,12 @@ VHostNetState *get_vhost_net(NetClientState *nc) > > vhost_net = vhost_user_get_vhost_net(nc); > > assert(vhost_net); > > break; > > +#endif > > +#ifdef CONFIG_VHOST_NET_VDPA > > + case NET_CLIENT_DRIVER_VHOST_VDPA: > > + vhost_net = vhost_vdpa_get_vhost_net(nc); > > + assert(vhost_net); > > + break; > > #endif > > default: > > break; > > diff --git a/hw/virtio/Makefile.objs b/hw/virtio/Makefile.objs > > index e2f70fbb89..e7c5d4a862 100644 > > --- a/hw/virtio/Makefile.objs > > +++ b/hw/virtio/Makefile.objs > > @@ -5,6 +5,7 @@ obj-y += virtio.o > > obj-$(call lor,$(CONFIG_VHOST_USER),$(CONFIG_VHOST_KERNEL)) += vhost.o vhost-backend.o > > common-obj-$(call lnot,$(call lor,$(CONFIG_VHOST_USER),$(CONFIG_VHOST_KERNEL))) += vhost-stub.o > > obj-$(CONFIG_VHOST_USER) += vhost-user.o > > +obj-$(CONFIG_VHOST_VDPA) += vhost-vdpa.o > > > > common-obj-$(CONFIG_VIRTIO_RNG) += virtio-rng.o > > common-obj-$(CONFIG_VIRTIO_PCI) += virtio-pci.o > > diff --git a/hw/virtio/vhost-backend.c b/hw/virtio/vhost-backend.c > > index 48905383f8..069ddb423d 100644 > > --- a/hw/virtio/vhost-backend.c > > +++ b/hw/virtio/vhost-backend.c > > @@ -285,6 +285,11 @@ int vhost_set_backend_type(struct vhost_dev *dev, VhostBackendType backend_type) > > case VHOST_BACKEND_TYPE_USER: > > dev->vhost_ops = &user_ops; > > break; > > +#endif > > +#ifdef CONFIG_VHOST_VDPA > > + case VHOST_BACKEND_TYPE_VDPA: > > + dev->vhost_ops = &vdpa_ops; > > + break; > > #endif > > default: > > error_report("Unknown vhost backend type"); > > diff --git a/hw/virtio/vhost-vdpa.c b/hw/virtio/vhost-vdpa.c > > new file mode 100644 > > index 0000000000..bac8a8aa2a > > --- /dev/null > > +++ b/hw/virtio/vhost-vdpa.c > > @@ -0,0 +1,447 @@ > > +/* > > + * vhost-vdpa > > + * > > + * Copyright(c) 2017-2018 Intel Corporation. > > + * Copyright(c) 2020 Red Hat, Inc. > > + * > > + * This work is licensed under the terms of the GNU GPL, version 2 or later. > > + * See the COPYING file in the top-level directory. > > + * > > + */ > > + > > +#include "qemu/osdep.h" > > +#include <linux/vhost.h> > > +#include <linux/vfio.h> > > +#include <sys/eventfd.h> > > +#include <sys/ioctl.h> > > +#include "hw/virtio/vhost.h" > > +#include "hw/virtio/vhost-backend.h" > > +#include "hw/virtio/virtio-net.h" > > +#include "hw/virtio/vhost-vdpa.h" > > +#include "qemu/main-loop.h" > > +#include <linux/kvm.h> > > +#include "sysemu/kvm.h" > > + > > + > > +static bool vhost_vdpa_listener_skipped_section(MemoryRegionSection *section) > > +{ > > + return (!memory_region_is_ram(section->mr) && > > + !memory_region_is_iommu(section->mr)) || > > + /* > > + * Sizing an enabled 64-bit BAR can cause spurious mappings to > > + * addresses in the upper part of the 64-bit address space. These > > + * are never accessed by the CPU and beyond the address width of > > + * some IOMMU hardware. TODO: VDPA should tell us the IOMMU width. > > + */ > > + section->offset_within_address_space & (1ULL << 63); > > +} > > + > > +static int vhost_vdpa_dma_map(struct vhost_vdpa *v, hwaddr iova, hwaddr size, > > + void *vaddr, bool readonly) > > +{ > > + struct vhost_msg_v2 msg; > > + int fd = v->device_fd; > > + int ret = 0; > > + > > + msg.type = v->msg_type; > > + msg.iotlb.iova = iova; > > + msg.iotlb.size = size; > > + msg.iotlb.uaddr = (uint64_t)vaddr; > > + msg.iotlb.perm = readonly ? VHOST_ACCESS_RO : VHOST_ACCESS_RW; > > + msg.iotlb.type = VHOST_IOTLB_UPDATE; > > + > > + if (write(fd, &msg, sizeof(msg)) != sizeof(msg)) { > > + error_report("failed to write, fd=%d, errno=%d (%s)", > > + fd, errno, strerror(errno)); > > + return -EIO ; > > + } > > + > > + return ret; > > +} > > + > > +static int vhost_vdpa_dma_unmap(struct vhost_vdpa *v, hwaddr iova, > > + hwaddr size) > > +{ > > + struct vhost_msg_v2 msg; > > + int fd = v->device_fd; > > + int ret = 0; > > + > > + msg.type = v->msg_type; > > + msg.iotlb.iova = iova; > > + msg.iotlb.size = size; > > + msg.iotlb.type = VHOST_IOTLB_INVALIDATE; > > + > > + if (write(fd, &msg, sizeof(msg)) != sizeof(msg)) { > > + error_report("failed to write, fd=%d, errno=%d (%s)", > > + fd, errno, strerror(errno)); > > + return -EIO ; > > + } > > + > > + return ret; > > +} > > + > > +static void vhost_vdpa_listener_region_add(MemoryListener *listener, > > + MemoryRegionSection *section) > > +{ > > + struct vhost_vdpa *v = container_of(listener, struct vhost_vdpa, listener); > > + hwaddr iova; > > + Int128 llend, llsize; > > + void *vaddr; > > + int ret; > > + > > + if (vhost_vdpa_listener_skipped_section(section)) { > > + return; > > + } > > + > > + if (unlikely((section->offset_within_address_space & ~TARGET_PAGE_MASK) != > > + (section->offset_within_region & ~TARGET_PAGE_MASK))) { > > + error_report("%s received unaligned region", __func__); > > + return; > > + } > > + > > + iova = TARGET_PAGE_ALIGN(section->offset_within_address_space); > > + llend = int128_make64(section->offset_within_address_space); > > + llend = int128_add(llend, section->size); > > + llend = int128_and(llend, int128_exts64(TARGET_PAGE_MASK)); > > + > > + if (int128_ge(int128_make64(iova), llend)) { > > + return; > > + } > > + > > + memory_region_ref(section->mr); > > + > > + /* Here we assume that memory_region_is_ram(section->mr)==true */ > > + > > + vaddr = memory_region_get_ram_ptr(section->mr) + > > + section->offset_within_region + > > + (iova - section->offset_within_address_space); > > + > > + llsize = int128_sub(llend, int128_make64(iova)); > > + > > + ret = vhost_vdpa_dma_map(v, iova, int128_get64(llsize), > > + vaddr, section->readonly); > > + if (ret) { > > + error_report("vhost vdpa map fail!"); > > + if (memory_region_is_ram_device(section->mr)) { > > + /* Allow unexpected mappings not to be fatal for RAM devices */ > > + error_report("map ram fail!"); > > + return ; > > + } > > + goto fail; > > + } > > + > > + return; > > + > > +fail: > > + if (memory_region_is_ram_device(section->mr)) { > > + error_report("failed to vdpa_dma_map. pci p2p may not work"); > > + return; > > + > > + } > > + /* > > + * On the initfn path, store the first error in the container so we > > + * can gracefully fail. Runtime, there's not much we can do other > > + * than throw a hardware error. > > + */ > > + error_report("vhost-vdpa: DMA mapping failed, unable to continue"); > > + return; > > + > > +} > > + > > +static void vhost_vdpa_listener_region_del(MemoryListener *listener, > > + MemoryRegionSection *section) > > +{ > > + struct vhost_vdpa *v = container_of(listener, struct vhost_vdpa, listener); > > + hwaddr iova; > > + Int128 llend, llsize; > > + int ret; > > + bool try_unmap = true; > > + > > + if (vhost_vdpa_listener_skipped_section(section)) { > > + return; > > + } > > + > > + if (unlikely((section->offset_within_address_space & ~TARGET_PAGE_MASK) != > > + (section->offset_within_region & ~TARGET_PAGE_MASK))) { > > + error_report("%s received unaligned region", __func__); > > + return; > > + } > > + > > + iova = TARGET_PAGE_ALIGN(section->offset_within_address_space); > > + llend = int128_make64(section->offset_within_address_space); > > + llend = int128_add(llend, section->size); > > + llend = int128_and(llend, int128_exts64(TARGET_PAGE_MASK)); > > + > > + if (int128_ge(int128_make64(iova), llend)) { > > + return; > > + } > > + > > + llsize = int128_sub(llend, int128_make64(iova)); > > + > > + if (try_unmap) { > > + ret = vhost_vdpa_dma_unmap(v, iova, int128_get64(llsize)); > > + if (ret) { > > + error_report("vhost_vdpa dma unmap error!"); > > + } > > + } > > + > > + memory_region_unref(section->mr); > > +} > > + > > +static const MemoryListener vhost_vdpa_memory_listener = { > > + .region_add = vhost_vdpa_listener_region_add, > > + .region_del = vhost_vdpa_listener_region_del, > > +}; > > + > > > Need comment to explain why vhost-vdpa requires a specific memory > listener other than just reusing the one that is provided by vhost. > > will fix this > > + > > > This newline is unnecessary. > will fix this > > > +static int vhost_vdpa_call(struct vhost_dev *dev, unsigned long int request, > > + void *arg) > > +{ > > + struct vhost_vdpa *v = dev->opaque; > > + int fd = v->device_fd; > > + > > + assert(dev->vhost_ops->backend_type == VHOST_BACKEND_TYPE_VDPA); > > + > > + return ioctl(fd, request, arg); > > +} > > + > > +static int vhost_vdpa_get_backend_features(struct vhost_dev *dev, > > + uint64_t *features) > > +{ > > + return vhost_vdpa_call(dev, VHOST_GET_BACKEND_FEATURES, features); > > +} > > + > > +static int vhost_vdpa_init(struct vhost_dev *dev, void *opaque) > > +{ > > + struct vhost_vdpa *v; > > + uint64_t backend_features = 0; > > + int ret = 0; > > + > > + assert(dev->vhost_ops->backend_type == VHOST_BACKEND_TYPE_VDPA); > > + > > + v = opaque; > > + dev->opaque = opaque ; > > + > > + v->listener = vhost_vdpa_memory_listener; > > + memory_listener_register(&v->listener, &address_space_memory); > > + > > + ret = vhost_vdpa_get_backend_features(dev, &backend_features); > > + /* > > + * In order to compatible with older version kernel, > > + * If the kernel not support this ioctl, > > + * set the msg_type for v2 by default > > + */ > > + if (ret) { > > + v->msg_type = VHOST_IOTLB_MSG_V2; > > + return 0; > > + } > > + if (backend_features & (1ULL << VHOST_BACKEND_F_IOTLB_MSG_V2)) { > > + v->msg_type = VHOST_IOTLB_MSG_V2; > > + } else { > > + v->msg_type = VHOST_IOTLB_MSG; > > + } > > > But in vhost_vdpa_dma_unmap/vhost_vdpa_dma_map, v2 is still being used. > > Consider vdpa implies msg_v2, and it doesn't implement > get_backend_features. We use v2 unconditionally and leave this check in > the future. > > will fix this > > + return 0; > > +} > > + > > +static int vhost_vdpa_cleanup(struct vhost_dev *dev) > > +{ > > + struct vhost_vdpa *v; > > + assert(dev->vhost_ops->backend_type == VHOST_BACKEND_TYPE_VDPA); > > + > > + v = dev->opaque; > > + memory_listener_unregister(&v->listener); > > + > > + dev->opaque = NULL; > > + return 0; > > +} > > + > > +static int vhost_vdpa_memslots_limit(struct vhost_dev *dev) > > +{ > > + /*Using IOTLB API here,NOTE:please set the ulimit before using*/ > > > The "note" seems unrelated to the memslot here. > > > > + return INT_MAX; > > +} > > + > > +static int vhost_vdpa_set_log_base(struct vhost_dev *dev, uint64_t base, > > + struct vhost_log *log) > > +{ > > + return -1; > > +} > > + > > +static int vhost_vdpa_set_mem_table(struct vhost_dev *dev, > > + struct vhost_memory *mem) > > +{ > > + > > + if (mem->padding) { > > + return -1; > > + } > > + > > + return 0; > > +} > > + > > +static int vhost_vdpa_set_vring_addr(struct vhost_dev *dev, > > + struct vhost_vring_addr *addr) > > +{ > > + return vhost_vdpa_call(dev, VHOST_SET_VRING_ADDR, addr); > > +} > > + > > +static int vhost_vdpa_set_vring_num(struct vhost_dev *dev, > > + struct vhost_vring_state *ring) > > +{ > > + return vhost_vdpa_call(dev, VHOST_SET_VRING_NUM, ring); > > +} > > + > > +static int vhost_vdpa_set_vring_base(struct vhost_dev *dev, > > + struct vhost_vring_state *ring) > > +{ > > + return vhost_vdpa_call(dev, VHOST_GET_VRING_BASE, ring); > > +} > > + > > +static int vhost_vdpa_get_vring_base(struct vhost_dev *dev, > > + struct vhost_vring_state *ring) > > +{ > > + > > + return vhost_vdpa_call(dev, VHOST_GET_VRING_BASE, ring); > > +} > > + > > +static int vhost_vdpa_set_vring_kick(struct vhost_dev *dev, > > + struct vhost_vring_file *file) > > +{ > > + return vhost_vdpa_call(dev, VHOST_SET_VRING_KICK, file); > > +} > > + > > +static int vhost_vdpa_set_vring_call(struct vhost_dev *dev, > > + struct vhost_vring_file *file) > > +{ > > + return vhost_vdpa_call(dev, VHOST_SET_VRING_CALL, file); > > +} > > + > > +static int vhost_vdpa_set_features(struct vhost_dev *dev, > > + uint64_t features) > > +{ > > + return vhost_vdpa_call(dev, VHOST_SET_FEATURES, &features); > > +} > > + > > +static int vhost_vdpa_get_features(struct vhost_dev *dev, > > + uint64_t *features) > > +{ > > + return vhost_vdpa_call(dev, VHOST_GET_FEATURES, features); > > +} > > + > > +static int vhost_vdpa_get_device_id(struct vhost_dev *dev, > > + uint32_t *device_id) > > +{ > > + return vhost_vdpa_call(dev, VHOST_VDPA_GET_DEVICE_ID, device_id); > > +} > > + > > +static int vhost_vdpa_set_owner(struct vhost_dev *dev) > > +{ > > + return vhost_vdpa_call(dev, VHOST_SET_OWNER, NULL); > > +} > > + > > +static int vhost_vdpa_reset_device(struct vhost_dev *dev) > > +{ > > + return vhost_vdpa_call(dev, VHOST_RESET_OWNER, NULL); > > +} > > > I think we may want to reuse the helpers of kenrel backends instead of > duplicating functions here. > will fix this > > > + > > +static int vhost_vdpa_get_vq_index(struct vhost_dev *dev, int idx) > > +{ > > + assert(idx >= dev->vq_index && idx < dev->vq_index + dev->nvqs); > > + > > + return idx - dev->vq_index; > > +} > > + > > +static int vhost_vdpa_set_vring_ready(struct vhost_dev *dev) > > +{ > > + int i; > > + for (i = 0; i < dev->nvqs; ++i) { > > + struct vhost_vring_state state = { > > + .index = dev->vq_index + i, > > + .num = 1, > > + }; > > + vhost_vdpa_call(dev, VHOST_VDPA_SET_VRING_ENABLE, &state); > > + } > > + return 0; > > +} > > +static int vhost_vdpa_set_config(struct vhost_dev *dev, const uint8_t *data, > > > One unnecessary space between int and "vhost_vdpa_set_config" > will fix this > > > + uint32_t offset, uint32_t size, > > + uint32_t flags) > > +{ > > + > > + struct vhost_vdpa_config *config; > > + int ret; > > + if ((size > VHOST_VDPA_MAX_CONFIG_SIZE) || (data == NULL)) { > > + return -1; > > + } > > + config = g_new0(struct vhost_vdpa_config, 1); > > > It's better to just use a temporary variable instead of allocate it > dynamically? > sure, will fix this > > > + if (config == NULL) { > > + return -EINVAL; > > + } > > + config->off = 0; > > + config->len = size; > > + memcpy(config->buf, data, size); > > + ret = vhost_vdpa_call(dev, VHOST_VDPA_SET_CONFIG, config); > > + g_free(config); > > + return ret; > > +} > > + > > + static int vhost_vdpa_get_config(struct vhost_dev *dev, uint8_t *config, > > + uint32_t config_len) > > +{ > > + struct vhost_vdpa_config *v_config = NULL; > > + int ret; > > + > > + if ((config_len > VHOST_VDPA_MAX_CONFIG_SIZE) || (config == NULL)) { > > + return -1; > > + } > > + v_config = g_new0(struct vhost_vdpa_config, 1); > > > Let's use a variable on the stack. > will fix this > > > + > > + if (v_config == NULL) { > > + return -EINVAL; > > + } > > + ret = vhost_vdpa_call(dev, VHOST_VDPA_GET_CONFIG, v_config); > > > One extra space after '='. > > > > + if ((v_config->len > config_len - v_config->off) || (v_config->len == 0)) { > > + g_free(v_config); > > + return -EINVAL; > > + } > > > Indentation is wired and why need such check? > > > > + > > + memcpy(config, v_config->buf + v_config->off, config_len); > > + g_free(v_config); > > + return ret; > > + } > > > Newline is needed here. > > > > +static int vhost_vdpa_set_state(struct vhost_dev *dev, uint8_t state) > > +{ > > + return vhost_vdpa_call(dev, VHOST_VDPA_SET_STATUS, &state); > > +} > > + > > + > > > This new line is not needed. > Will fix these all > > > +const VhostOps vdpa_ops = { > > + .backend_type = VHOST_BACKEND_TYPE_VDPA, > > + .vhost_backend_init = vhost_vdpa_init, > > + .vhost_backend_cleanup = vhost_vdpa_cleanup, > > + .vhost_backend_memslots_limit = vhost_vdpa_memslots_limit, > > + .vhost_set_log_base = vhost_vdpa_set_log_base, > > + .vhost_set_mem_table = vhost_vdpa_set_mem_table, > > + .vhost_set_vring_addr = vhost_vdpa_set_vring_addr, > > + .vhost_set_vring_endian = NULL, > > + .vhost_set_vring_num = vhost_vdpa_set_vring_num, > > + .vhost_set_vring_base = vhost_vdpa_set_vring_base, > > + .vhost_get_vring_base = vhost_vdpa_get_vring_base, > > + .vhost_set_vring_kick = vhost_vdpa_set_vring_kick, > > + .vhost_set_vring_call = vhost_vdpa_set_vring_call, > > + .vhost_set_features = vhost_vdpa_set_features, > > + .vhost_get_features = vhost_vdpa_get_features, > > + .vhost_set_owner = vhost_vdpa_set_owner, > > + .vhost_reset_device = vhost_vdpa_reset_device, > > + .vhost_get_vq_index = vhost_vdpa_get_vq_index, > > + .vhost_set_vring_ready = vhost_vdpa_set_vring_ready, > > + .vhost_get_config = vhost_vdpa_get_config, > > + .vhost_set_config = vhost_vdpa_set_config, > > + .vhost_requires_shm_log = NULL, > > + .vhost_migration_done = NULL, > > + .vhost_backend_can_merge = NULL, > > + .vhost_net_set_mtu = NULL, > > + .vhost_set_iotlb_callback = NULL, > > + .vhost_send_device_iotlb_msg = NULL, > > + .vhost_set_state = vhost_vdpa_set_state, > > + .vhost_get_device_id = vhost_vdpa_get_device_id, > > +}; > > > As said previously, need to consider to reuse vhost kernel helpers. > I will double check this > > > diff --git a/hw/virtio/vhost.c b/hw/virtio/vhost.c > > index 4da0d5a6c5..6d2aa461d6 100644 > > --- a/hw/virtio/vhost.c > > +++ b/hw/virtio/vhost.c > > @@ -746,6 +746,12 @@ static int vhost_virtqueue_set_addr(struct vhost_dev *dev, > > .log_guest_addr = vq->used_phys, > > .flags = enable_log ? (1 << VHOST_VRING_F_LOG) : 0, > > }; > > + /*vDPA need to use the phys address here to set to hardware*/ > > + if (dev->vhost_ops->backend_type == VHOST_BACKEND_TYPE_VDPA) { > > + addr.desc_user_addr = (uint64_t)(unsigned long)vq->desc_phys; > > + addr.avail_user_addr = (uint64_t)(unsigned long)vq->avail_phys; > > + addr.used_user_addr = (uint64_t)(unsigned long)vq->used_phys; > > + } > > int r = dev->vhost_ops->vhost_set_vring_addr(dev, &addr); > > if (r < 0) { > > VHOST_OPS_DEBUG("vhost_set_vring_addr failed"); > > @@ -1493,6 +1499,14 @@ int vhost_dev_set_config(struct vhost_dev *hdev, const uint8_t *data, > > return -1; > > } > > > > +int vhost_dev_get_device_id(struct vhost_dev *hdev, uint32_t *device_id) > > +{ > > + assert(hdev->vhost_ops); > > + if (hdev->vhost_ops->vhost_get_device_id) { > > > If you introduce a new vhost_ops, please use a separate patch. > > > > + return hdev->vhost_ops->vhost_get_device_id(hdev, device_id); > > + } > > + return -1; > > +} > > void vhost_dev_set_config_notifier(struct vhost_dev *hdev, > > const VhostDevConfigOps *ops) > > { > > diff --git a/include/hw/virtio/vhost-backend.h b/include/hw/virtio/vhost-backend.h > > index f823055167..8ac898a987 100644 > > --- a/include/hw/virtio/vhost-backend.h > > +++ b/include/hw/virtio/vhost-backend.h > > @@ -17,7 +17,8 @@ typedef enum VhostBackendType { > > VHOST_BACKEND_TYPE_NONE = 0, > > VHOST_BACKEND_TYPE_KERNEL = 1, > > VHOST_BACKEND_TYPE_USER = 2, > > - VHOST_BACKEND_TYPE_MAX = 3, > > + VHOST_BACKEND_TYPE_VDPA = 3, > > + VHOST_BACKEND_TYPE_MAX = 4, > > } VhostBackendType; > > > > typedef enum VhostSetConfigType { > > @@ -77,6 +78,7 @@ typedef int (*vhost_reset_device_op)(struct vhost_dev *dev); > > typedef int (*vhost_get_vq_index_op)(struct vhost_dev *dev, int idx); > > typedef int (*vhost_set_vring_enable_op)(struct vhost_dev *dev, > > int enable); > > +typedef int (*vhost_set_vring_ready_op)(struct vhost_dev *dev); > > > A new patch for this patch. > > Thanks > Sure, Will fix this > > > typedef bool (*vhost_requires_shm_log_op)(struct vhost_dev *dev); > > typedef int (*vhost_migration_done_op)(struct vhost_dev *dev, > > char *mac_addr); > > @@ -113,6 +115,7 @@ typedef int (*vhost_set_inflight_fd_op)(struct vhost_dev *dev, > > struct vhost_inflight *inflight); > > > > typedef int (*vhost_set_state_op)(struct vhost_dev *dev, uint8_t state); > > +typedef int (*vhost_get_device_id_op)(struct vhost_dev *dev, uint32_t *dev_id); > > typedef struct VhostOps { > > VhostBackendType backend_type; > > vhost_backend_init vhost_backend_init; > > @@ -139,6 +142,7 @@ typedef struct VhostOps { > > vhost_reset_device_op vhost_reset_device; > > vhost_get_vq_index_op vhost_get_vq_index; > > vhost_set_vring_enable_op vhost_set_vring_enable; > > + vhost_set_vring_ready_op vhost_set_vring_ready; > > vhost_requires_shm_log_op vhost_requires_shm_log; > > vhost_migration_done_op vhost_migration_done; > > vhost_backend_can_merge_op vhost_backend_can_merge; > > @@ -154,9 +158,11 @@ typedef struct VhostOps { > > vhost_get_inflight_fd_op vhost_get_inflight_fd; > > vhost_set_inflight_fd_op vhost_set_inflight_fd; > > vhost_set_state_op vhost_set_state; > > + vhost_get_device_id_op vhost_get_device_id; > > } VhostOps; > > > > extern const VhostOps user_ops; > > +extern const VhostOps vdpa_ops; > > > > int vhost_set_backend_type(struct vhost_dev *dev, > > VhostBackendType backend_type); > > diff --git a/include/hw/virtio/vhost-vdpa.h b/include/hw/virtio/vhost-vdpa.h > > new file mode 100644 > > index 0000000000..9d64fbfea9 > > --- /dev/null > > +++ b/include/hw/virtio/vhost-vdpa.h > > @@ -0,0 +1,25 @@ > > +/* > > + * vhost-vdpa.h > > + * > > + * Copyright(c) 2017-2018 Intel Corporation. > > + * Copyright(c) 2020 Red Hat, Inc. > > + * > > + * This work is licensed under the terms of the GNU GPL, version 2 or later. > > + * See the COPYING file in the top-level directory. > > + * > > + */ > > + > > +#ifndef HW_VIRTIO_VHOST_VDPA_H > > +#define HW_VIRTIO_VHOST_VDPA_H > > + > > +#include "hw/virtio/virtio.h" > > + > > +typedef struct vhost_vdpa { > > + int device_fd; > > + uint32_t msg_type; > > + MemoryListener listener; > > +} VhostVDPA; > > + > > +extern AddressSpace address_space_memory; > > + > > +#endif > > diff --git a/include/hw/virtio/vhost.h b/include/hw/virtio/vhost.h > > index 085450c6f8..020956d5a9 100644 > > --- a/include/hw/virtio/vhost.h > > +++ b/include/hw/virtio/vhost.h > > @@ -124,6 +124,7 @@ int vhost_dev_get_config(struct vhost_dev *dev, uint8_t *config, > > uint32_t config_len); > > int vhost_dev_set_config(struct vhost_dev *dev, const uint8_t *data, > > uint32_t offset, uint32_t size, uint32_t flags); > > +int vhost_dev_get_device_id(struct vhost_dev *hdev, uint32_t *device_id); > > /* notifier callback in case vhost device config space changed > > */ > > void vhost_dev_set_config_notifier(struct vhost_dev *dev, >
On Sat, May 09, 2020 at 12:32:14AM +0800, Cindy Lu wrote: > From: Tiwei Bie <tiwei.bie@intel.com> > > Currently we have 2 types of vhost backends in QEMU: vhost kernel and > vhost-user. The above patch provides a generic device for vDPA purpose, > this vDPA device exposes to user space a non-vendor-specific configuration > interface for setting up a vhost HW accelerator, this patch set introduces > a third vhost backend called vhost-vdpa based on the vDPA interface. > > Vhost-vdpa usage: > > qemu-system-x86_64 -cpu host -enable-kvm \ > ...... > -netdev type=vhost-vdpa,vhostdev=/dev/vhost-vdpa-id,id=vhost-vdpa0 \ > -device virtio-net-pci,netdev=vhost-vdpa0,page-per-vq=on \ I haven't looked at vDPA in depth. What is different here compared to the existing vhost-backend.c kernel backend? It seems to be making the same ioctls calls so I wonder if it makes sense to share the vhost-backend.c kernel code? Stefan
On Thu, May 21, 2020 at 8:40 PM Stefan Hajnoczi <stefanha@gmail.com> wrote: > > On Sat, May 09, 2020 at 12:32:14AM +0800, Cindy Lu wrote: > > From: Tiwei Bie <tiwei.bie@intel.com> > > > > Currently we have 2 types of vhost backends in QEMU: vhost kernel and > > vhost-user. The above patch provides a generic device for vDPA purpose, > > this vDPA device exposes to user space a non-vendor-specific configuration > > interface for setting up a vhost HW accelerator, this patch set introduces > > a third vhost backend called vhost-vdpa based on the vDPA interface. > > > > Vhost-vdpa usage: > > > > qemu-system-x86_64 -cpu host -enable-kvm \ > > ...... > > -netdev type=vhost-vdpa,vhostdev=/dev/vhost-vdpa-id,id=vhost-vdpa0 \ > > -device virtio-net-pci,netdev=vhost-vdpa0,page-per-vq=on \ > > I haven't looked at vDPA in depth. What is different here compared to > the existing vhost-backend.c kernel backend? > > It seems to be making the same ioctls calls so I wonder if it makes > sense to share the vhost-backend.c kernel code? > > Stefan Hi Stefan, Sorry for the late reply and Thanks for these suggestions. I think the most difference between vhost kernel and vdpa is vdpa depends on a real hardware. The point is that vDPA devices work as a virtio device, but vhost-vdpa qemu must present a vhost-like device in qemu vhost layer. The ioctl calls are similar with vhost-backend.c now, but after more and more NIC support vdpa. The difference between vhost-backend.c and vpda will become more and more big. It will make the code complicated to share the code with kernel code. Thanks Cindy
On Mon, May 25, 2020 at 4:34 PM Cindy Lu <lulu@redhat.com> wrote: > On Thu, May 21, 2020 at 8:40 PM Stefan Hajnoczi <stefanha@gmail.com> wrote: > > On Sat, May 09, 2020 at 12:32:14AM +0800, Cindy Lu wrote: > > > From: Tiwei Bie <tiwei.bie@intel.com> > > > > > > Currently we have 2 types of vhost backends in QEMU: vhost kernel and > > > vhost-user. The above patch provides a generic device for vDPA purpose, > > > this vDPA device exposes to user space a non-vendor-specific configuration > > > interface for setting up a vhost HW accelerator, this patch set introduces > > > a third vhost backend called vhost-vdpa based on the vDPA interface. > > > > > > Vhost-vdpa usage: > > > > > > qemu-system-x86_64 -cpu host -enable-kvm \ > > > ...... > > > -netdev type=vhost-vdpa,vhostdev=/dev/vhost-vdpa-id,id=vhost-vdpa0 \ > > > -device virtio-net-pci,netdev=vhost-vdpa0,page-per-vq=on \ > > > > I haven't looked at vDPA in depth. What is different here compared to > > the existing vhost-backend.c kernel backend? > > > > It seems to be making the same ioctls calls so I wonder if it makes > > sense to share the vhost-backend.c kernel code? > > > > Stefan > Hi Stefan, > Sorry for the late reply and Thanks for these suggestions. > I think the most difference between vhost kernel and vdpa is vdpa > depends on a real hardware. > The point is that vDPA devices work as a virtio device, but vhost-vdpa > qemu must present a vhost-like device in qemu vhost layer. > The ioctl calls are similar with vhost-backend.c now, but after more > and more NIC support vdpa. The difference between vhost-backend.c and > vpda will become more and more big. It will make the code complicated > to share the code with kernel code. Okay, thanks. Stefan
diff --git a/hw/net/vhost_net.c b/hw/net/vhost_net.c index 63b2a85d6e..1af39abaf3 100644 --- a/hw/net/vhost_net.c +++ b/hw/net/vhost_net.c @@ -17,8 +17,10 @@ #include "net/net.h" #include "net/tap.h" #include "net/vhost-user.h" +#include "net/vhost-vdpa.h" #include "standard-headers/linux/vhost_types.h" +#include "linux-headers/linux/vhost.h" #include "hw/virtio/virtio-net.h" #include "net/vhost_net.h" #include "qemu/error-report.h" @@ -85,6 +87,29 @@ static const int user_feature_bits[] = { VHOST_INVALID_FEATURE_BIT }; +static const int vdpa_feature_bits[] = { + VIRTIO_F_NOTIFY_ON_EMPTY, + VIRTIO_RING_F_INDIRECT_DESC, + VIRTIO_RING_F_EVENT_IDX, + VIRTIO_F_ANY_LAYOUT, + VIRTIO_F_VERSION_1, + VIRTIO_NET_F_CSUM, + VIRTIO_NET_F_GUEST_CSUM, + VIRTIO_NET_F_GSO, + VIRTIO_NET_F_GUEST_TSO4, + VIRTIO_NET_F_GUEST_TSO6, + VIRTIO_NET_F_GUEST_ECN, + VIRTIO_NET_F_GUEST_UFO, + VIRTIO_NET_F_HOST_TSO4, + VIRTIO_NET_F_HOST_TSO6, + VIRTIO_NET_F_HOST_ECN, + VIRTIO_NET_F_HOST_UFO, + VIRTIO_NET_F_MRG_RXBUF, + VIRTIO_NET_F_MTU, + VIRTIO_F_IOMMU_PLATFORM, + VIRTIO_NET_F_GUEST_ANNOUNCE, + VHOST_INVALID_FEATURE_BIT +}; static const int *vhost_net_get_feature_bits(struct vhost_net *net) { const int *feature_bits = 0; @@ -96,6 +121,9 @@ static const int *vhost_net_get_feature_bits(struct vhost_net *net) case NET_CLIENT_DRIVER_VHOST_USER: feature_bits = user_feature_bits; break; + case NET_CLIENT_DRIVER_VHOST_VDPA: + feature_bits = vdpa_feature_bits; + break; default: error_report("Feature bits not defined for this type: %d", net->nc->info->type); @@ -110,7 +138,10 @@ uint64_t vhost_net_get_features(struct vhost_net *net, uint64_t features) return vhost_get_features(&net->dev, vhost_net_get_feature_bits(net), features); } - +int vhost_net_get_device_id(struct vhost_net *net, uint32_t * device_id) +{ + return vhost_dev_get_device_id(&net->dev, device_id); +} void vhost_net_ack_features(struct vhost_net *net, uint64_t features) { net->dev.acked_features = net->dev.backend_features; @@ -433,6 +464,12 @@ VHostNetState *get_vhost_net(NetClientState *nc) vhost_net = vhost_user_get_vhost_net(nc); assert(vhost_net); break; +#endif +#ifdef CONFIG_VHOST_NET_VDPA + case NET_CLIENT_DRIVER_VHOST_VDPA: + vhost_net = vhost_vdpa_get_vhost_net(nc); + assert(vhost_net); + break; #endif default: break; diff --git a/hw/virtio/Makefile.objs b/hw/virtio/Makefile.objs index e2f70fbb89..e7c5d4a862 100644 --- a/hw/virtio/Makefile.objs +++ b/hw/virtio/Makefile.objs @@ -5,6 +5,7 @@ obj-y += virtio.o obj-$(call lor,$(CONFIG_VHOST_USER),$(CONFIG_VHOST_KERNEL)) += vhost.o vhost-backend.o common-obj-$(call lnot,$(call lor,$(CONFIG_VHOST_USER),$(CONFIG_VHOST_KERNEL))) += vhost-stub.o obj-$(CONFIG_VHOST_USER) += vhost-user.o +obj-$(CONFIG_VHOST_VDPA) += vhost-vdpa.o common-obj-$(CONFIG_VIRTIO_RNG) += virtio-rng.o common-obj-$(CONFIG_VIRTIO_PCI) += virtio-pci.o diff --git a/hw/virtio/vhost-backend.c b/hw/virtio/vhost-backend.c index 48905383f8..069ddb423d 100644 --- a/hw/virtio/vhost-backend.c +++ b/hw/virtio/vhost-backend.c @@ -285,6 +285,11 @@ int vhost_set_backend_type(struct vhost_dev *dev, VhostBackendType backend_type) case VHOST_BACKEND_TYPE_USER: dev->vhost_ops = &user_ops; break; +#endif +#ifdef CONFIG_VHOST_VDPA + case VHOST_BACKEND_TYPE_VDPA: + dev->vhost_ops = &vdpa_ops; + break; #endif default: error_report("Unknown vhost backend type"); diff --git a/hw/virtio/vhost-vdpa.c b/hw/virtio/vhost-vdpa.c new file mode 100644 index 0000000000..bac8a8aa2a --- /dev/null +++ b/hw/virtio/vhost-vdpa.c @@ -0,0 +1,447 @@ +/* + * vhost-vdpa + * + * Copyright(c) 2017-2018 Intel Corporation. + * Copyright(c) 2020 Red Hat, Inc. + * + * This work is licensed under the terms of the GNU GPL, version 2 or later. + * See the COPYING file in the top-level directory. + * + */ + +#include "qemu/osdep.h" +#include <linux/vhost.h> +#include <linux/vfio.h> +#include <sys/eventfd.h> +#include <sys/ioctl.h> +#include "hw/virtio/vhost.h" +#include "hw/virtio/vhost-backend.h" +#include "hw/virtio/virtio-net.h" +#include "hw/virtio/vhost-vdpa.h" +#include "qemu/main-loop.h" +#include <linux/kvm.h> +#include "sysemu/kvm.h" + + +static bool vhost_vdpa_listener_skipped_section(MemoryRegionSection *section) +{ + return (!memory_region_is_ram(section->mr) && + !memory_region_is_iommu(section->mr)) || + /* + * Sizing an enabled 64-bit BAR can cause spurious mappings to + * addresses in the upper part of the 64-bit address space. These + * are never accessed by the CPU and beyond the address width of + * some IOMMU hardware. TODO: VDPA should tell us the IOMMU width. + */ + section->offset_within_address_space & (1ULL << 63); +} + +static int vhost_vdpa_dma_map(struct vhost_vdpa *v, hwaddr iova, hwaddr size, + void *vaddr, bool readonly) +{ + struct vhost_msg_v2 msg; + int fd = v->device_fd; + int ret = 0; + + msg.type = v->msg_type; + msg.iotlb.iova = iova; + msg.iotlb.size = size; + msg.iotlb.uaddr = (uint64_t)vaddr; + msg.iotlb.perm = readonly ? VHOST_ACCESS_RO : VHOST_ACCESS_RW; + msg.iotlb.type = VHOST_IOTLB_UPDATE; + + if (write(fd, &msg, sizeof(msg)) != sizeof(msg)) { + error_report("failed to write, fd=%d, errno=%d (%s)", + fd, errno, strerror(errno)); + return -EIO ; + } + + return ret; +} + +static int vhost_vdpa_dma_unmap(struct vhost_vdpa *v, hwaddr iova, + hwaddr size) +{ + struct vhost_msg_v2 msg; + int fd = v->device_fd; + int ret = 0; + + msg.type = v->msg_type; + msg.iotlb.iova = iova; + msg.iotlb.size = size; + msg.iotlb.type = VHOST_IOTLB_INVALIDATE; + + if (write(fd, &msg, sizeof(msg)) != sizeof(msg)) { + error_report("failed to write, fd=%d, errno=%d (%s)", + fd, errno, strerror(errno)); + return -EIO ; + } + + return ret; +} + +static void vhost_vdpa_listener_region_add(MemoryListener *listener, + MemoryRegionSection *section) +{ + struct vhost_vdpa *v = container_of(listener, struct vhost_vdpa, listener); + hwaddr iova; + Int128 llend, llsize; + void *vaddr; + int ret; + + if (vhost_vdpa_listener_skipped_section(section)) { + return; + } + + if (unlikely((section->offset_within_address_space & ~TARGET_PAGE_MASK) != + (section->offset_within_region & ~TARGET_PAGE_MASK))) { + error_report("%s received unaligned region", __func__); + return; + } + + iova = TARGET_PAGE_ALIGN(section->offset_within_address_space); + llend = int128_make64(section->offset_within_address_space); + llend = int128_add(llend, section->size); + llend = int128_and(llend, int128_exts64(TARGET_PAGE_MASK)); + + if (int128_ge(int128_make64(iova), llend)) { + return; + } + + memory_region_ref(section->mr); + + /* Here we assume that memory_region_is_ram(section->mr)==true */ + + vaddr = memory_region_get_ram_ptr(section->mr) + + section->offset_within_region + + (iova - section->offset_within_address_space); + + llsize = int128_sub(llend, int128_make64(iova)); + + ret = vhost_vdpa_dma_map(v, iova, int128_get64(llsize), + vaddr, section->readonly); + if (ret) { + error_report("vhost vdpa map fail!"); + if (memory_region_is_ram_device(section->mr)) { + /* Allow unexpected mappings not to be fatal for RAM devices */ + error_report("map ram fail!"); + return ; + } + goto fail; + } + + return; + +fail: + if (memory_region_is_ram_device(section->mr)) { + error_report("failed to vdpa_dma_map. pci p2p may not work"); + return; + + } + /* + * On the initfn path, store the first error in the container so we + * can gracefully fail. Runtime, there's not much we can do other + * than throw a hardware error. + */ + error_report("vhost-vdpa: DMA mapping failed, unable to continue"); + return; + +} + +static void vhost_vdpa_listener_region_del(MemoryListener *listener, + MemoryRegionSection *section) +{ + struct vhost_vdpa *v = container_of(listener, struct vhost_vdpa, listener); + hwaddr iova; + Int128 llend, llsize; + int ret; + bool try_unmap = true; + + if (vhost_vdpa_listener_skipped_section(section)) { + return; + } + + if (unlikely((section->offset_within_address_space & ~TARGET_PAGE_MASK) != + (section->offset_within_region & ~TARGET_PAGE_MASK))) { + error_report("%s received unaligned region", __func__); + return; + } + + iova = TARGET_PAGE_ALIGN(section->offset_within_address_space); + llend = int128_make64(section->offset_within_address_space); + llend = int128_add(llend, section->size); + llend = int128_and(llend, int128_exts64(TARGET_PAGE_MASK)); + + if (int128_ge(int128_make64(iova), llend)) { + return; + } + + llsize = int128_sub(llend, int128_make64(iova)); + + if (try_unmap) { + ret = vhost_vdpa_dma_unmap(v, iova, int128_get64(llsize)); + if (ret) { + error_report("vhost_vdpa dma unmap error!"); + } + } + + memory_region_unref(section->mr); +} + +static const MemoryListener vhost_vdpa_memory_listener = { + .region_add = vhost_vdpa_listener_region_add, + .region_del = vhost_vdpa_listener_region_del, +}; + + +static int vhost_vdpa_call(struct vhost_dev *dev, unsigned long int request, + void *arg) +{ + struct vhost_vdpa *v = dev->opaque; + int fd = v->device_fd; + + assert(dev->vhost_ops->backend_type == VHOST_BACKEND_TYPE_VDPA); + + return ioctl(fd, request, arg); +} + +static int vhost_vdpa_get_backend_features(struct vhost_dev *dev, + uint64_t *features) +{ + return vhost_vdpa_call(dev, VHOST_GET_BACKEND_FEATURES, features); +} + +static int vhost_vdpa_init(struct vhost_dev *dev, void *opaque) +{ + struct vhost_vdpa *v; + uint64_t backend_features = 0; + int ret = 0; + + assert(dev->vhost_ops->backend_type == VHOST_BACKEND_TYPE_VDPA); + + v = opaque; + dev->opaque = opaque ; + + v->listener = vhost_vdpa_memory_listener; + memory_listener_register(&v->listener, &address_space_memory); + + ret = vhost_vdpa_get_backend_features(dev, &backend_features); + /* + * In order to compatible with older version kernel, + * If the kernel not support this ioctl, + * set the msg_type for v2 by defeault + */ + if (ret) { + v->msg_type = VHOST_IOTLB_MSG_V2; + return 0; + } + if (backend_features & (1ULL << VHOST_BACKEND_F_IOTLB_MSG_V2)) { + v->msg_type = VHOST_IOTLB_MSG_V2; + } else { + v->msg_type = VHOST_IOTLB_MSG; + } + return 0; +} + +static int vhost_vdpa_cleanup(struct vhost_dev *dev) +{ + struct vhost_vdpa *v; + assert(dev->vhost_ops->backend_type == VHOST_BACKEND_TYPE_VDPA); + + v = dev->opaque; + memory_listener_unregister(&v->listener); + + dev->opaque = NULL; + return 0; +} + +static int vhost_vdpa_memslots_limit(struct vhost_dev *dev) +{ + /*Using IOTLB API here,NOTE:please set the ulimit before using*/ + return INT_MAX; +} + +static int vhost_vdpa_set_log_base(struct vhost_dev *dev, uint64_t base, + struct vhost_log *log) +{ + return -1; +} + +static int vhost_vdpa_set_mem_table(struct vhost_dev *dev, + struct vhost_memory *mem) +{ + + if (mem->padding) { + return -1; + } + + return 0; +} + +static int vhost_vdpa_set_vring_addr(struct vhost_dev *dev, + struct vhost_vring_addr *addr) +{ + return vhost_vdpa_call(dev, VHOST_SET_VRING_ADDR, addr); +} + +static int vhost_vdpa_set_vring_num(struct vhost_dev *dev, + struct vhost_vring_state *ring) +{ + return vhost_vdpa_call(dev, VHOST_SET_VRING_NUM, ring); +} + +static int vhost_vdpa_set_vring_base(struct vhost_dev *dev, + struct vhost_vring_state *ring) +{ + return vhost_vdpa_call(dev, VHOST_GET_VRING_BASE, ring); +} + +static int vhost_vdpa_get_vring_base(struct vhost_dev *dev, + struct vhost_vring_state *ring) +{ + + return vhost_vdpa_call(dev, VHOST_GET_VRING_BASE, ring); +} + +static int vhost_vdpa_set_vring_kick(struct vhost_dev *dev, + struct vhost_vring_file *file) +{ + return vhost_vdpa_call(dev, VHOST_SET_VRING_KICK, file); +} + +static int vhost_vdpa_set_vring_call(struct vhost_dev *dev, + struct vhost_vring_file *file) +{ + return vhost_vdpa_call(dev, VHOST_SET_VRING_CALL, file); +} + +static int vhost_vdpa_set_features(struct vhost_dev *dev, + uint64_t features) +{ + return vhost_vdpa_call(dev, VHOST_SET_FEATURES, &features); +} + +static int vhost_vdpa_get_features(struct vhost_dev *dev, + uint64_t *features) +{ + return vhost_vdpa_call(dev, VHOST_GET_FEATURES, features); +} + +static int vhost_vdpa_get_device_id(struct vhost_dev *dev, + uint32_t *device_id) +{ + return vhost_vdpa_call(dev, VHOST_VDPA_GET_DEVICE_ID, device_id); +} + +static int vhost_vdpa_set_owner(struct vhost_dev *dev) +{ + return vhost_vdpa_call(dev, VHOST_SET_OWNER, NULL); +} + +static int vhost_vdpa_reset_device(struct vhost_dev *dev) +{ + return vhost_vdpa_call(dev, VHOST_RESET_OWNER, NULL); +} + +static int vhost_vdpa_get_vq_index(struct vhost_dev *dev, int idx) +{ + assert(idx >= dev->vq_index && idx < dev->vq_index + dev->nvqs); + + return idx - dev->vq_index; +} + +static int vhost_vdpa_set_vring_ready(struct vhost_dev *dev) +{ + int i; + for (i = 0; i < dev->nvqs; ++i) { + struct vhost_vring_state state = { + .index = dev->vq_index + i, + .num = 1, + }; + vhost_vdpa_call(dev, VHOST_VDPA_SET_VRING_ENABLE, &state); + } + return 0; +} +static int vhost_vdpa_set_config(struct vhost_dev *dev, const uint8_t *data, + uint32_t offset, uint32_t size, + uint32_t flags) +{ + + struct vhost_vdpa_config *config; + int ret; + if ((size > VHOST_VDPA_MAX_CONFIG_SIZE) || (data == NULL)) { + return -1; + } + config = g_new0(struct vhost_vdpa_config, 1); + if (config == NULL) { + return -EINVAL; + } + config->off = 0; + config->len = size; + memcpy(config->buf, data, size); + ret = vhost_vdpa_call(dev, VHOST_VDPA_SET_CONFIG, config); + g_free(config); + return ret; +} + + static int vhost_vdpa_get_config(struct vhost_dev *dev, uint8_t *config, + uint32_t config_len) +{ + struct vhost_vdpa_config *v_config = NULL; + int ret; + + if ((config_len > VHOST_VDPA_MAX_CONFIG_SIZE) || (config == NULL)) { + return -1; + } + v_config = g_new0(struct vhost_vdpa_config, 1); + + if (v_config == NULL) { + return -EINVAL; + } + ret = vhost_vdpa_call(dev, VHOST_VDPA_GET_CONFIG, v_config); + if ((v_config->len > config_len - v_config->off) || (v_config->len == 0)) { + g_free(v_config); + return -EINVAL; + } + + memcpy(config, v_config->buf + v_config->off, config_len); + g_free(v_config); + return ret; + } +static int vhost_vdpa_set_state(struct vhost_dev *dev, uint8_t state) +{ + return vhost_vdpa_call(dev, VHOST_VDPA_SET_STATUS, &state); +} + + +const VhostOps vdpa_ops = { + .backend_type = VHOST_BACKEND_TYPE_VDPA, + .vhost_backend_init = vhost_vdpa_init, + .vhost_backend_cleanup = vhost_vdpa_cleanup, + .vhost_backend_memslots_limit = vhost_vdpa_memslots_limit, + .vhost_set_log_base = vhost_vdpa_set_log_base, + .vhost_set_mem_table = vhost_vdpa_set_mem_table, + .vhost_set_vring_addr = vhost_vdpa_set_vring_addr, + .vhost_set_vring_endian = NULL, + .vhost_set_vring_num = vhost_vdpa_set_vring_num, + .vhost_set_vring_base = vhost_vdpa_set_vring_base, + .vhost_get_vring_base = vhost_vdpa_get_vring_base, + .vhost_set_vring_kick = vhost_vdpa_set_vring_kick, + .vhost_set_vring_call = vhost_vdpa_set_vring_call, + .vhost_set_features = vhost_vdpa_set_features, + .vhost_get_features = vhost_vdpa_get_features, + .vhost_set_owner = vhost_vdpa_set_owner, + .vhost_reset_device = vhost_vdpa_reset_device, + .vhost_get_vq_index = vhost_vdpa_get_vq_index, + .vhost_set_vring_ready = vhost_vdpa_set_vring_ready, + .vhost_get_config = vhost_vdpa_get_config, + .vhost_set_config = vhost_vdpa_set_config, + .vhost_requires_shm_log = NULL, + .vhost_migration_done = NULL, + .vhost_backend_can_merge = NULL, + .vhost_net_set_mtu = NULL, + .vhost_set_iotlb_callback = NULL, + .vhost_send_device_iotlb_msg = NULL, + .vhost_set_state = vhost_vdpa_set_state, + .vhost_get_device_id = vhost_vdpa_get_device_id, +}; diff --git a/hw/virtio/vhost.c b/hw/virtio/vhost.c index 4da0d5a6c5..6d2aa461d6 100644 --- a/hw/virtio/vhost.c +++ b/hw/virtio/vhost.c @@ -746,6 +746,12 @@ static int vhost_virtqueue_set_addr(struct vhost_dev *dev, .log_guest_addr = vq->used_phys, .flags = enable_log ? (1 << VHOST_VRING_F_LOG) : 0, }; + /*vDPA need to use the phys address here to set to hardware*/ + if (dev->vhost_ops->backend_type == VHOST_BACKEND_TYPE_VDPA) { + addr.desc_user_addr = (uint64_t)(unsigned long)vq->desc_phys; + addr.avail_user_addr = (uint64_t)(unsigned long)vq->avail_phys; + addr.used_user_addr = (uint64_t)(unsigned long)vq->used_phys; + } int r = dev->vhost_ops->vhost_set_vring_addr(dev, &addr); if (r < 0) { VHOST_OPS_DEBUG("vhost_set_vring_addr failed"); @@ -1493,6 +1499,14 @@ int vhost_dev_set_config(struct vhost_dev *hdev, const uint8_t *data, return -1; } +int vhost_dev_get_device_id(struct vhost_dev *hdev, uint32_t *device_id) +{ + assert(hdev->vhost_ops); + if (hdev->vhost_ops->vhost_get_device_id) { + return hdev->vhost_ops->vhost_get_device_id(hdev, device_id); + } + return -1; +} void vhost_dev_set_config_notifier(struct vhost_dev *hdev, const VhostDevConfigOps *ops) { diff --git a/include/hw/virtio/vhost-backend.h b/include/hw/virtio/vhost-backend.h index f823055167..8ac898a987 100644 --- a/include/hw/virtio/vhost-backend.h +++ b/include/hw/virtio/vhost-backend.h @@ -17,7 +17,8 @@ typedef enum VhostBackendType { VHOST_BACKEND_TYPE_NONE = 0, VHOST_BACKEND_TYPE_KERNEL = 1, VHOST_BACKEND_TYPE_USER = 2, - VHOST_BACKEND_TYPE_MAX = 3, + VHOST_BACKEND_TYPE_VDPA = 3, + VHOST_BACKEND_TYPE_MAX = 4, } VhostBackendType; typedef enum VhostSetConfigType { @@ -77,6 +78,7 @@ typedef int (*vhost_reset_device_op)(struct vhost_dev *dev); typedef int (*vhost_get_vq_index_op)(struct vhost_dev *dev, int idx); typedef int (*vhost_set_vring_enable_op)(struct vhost_dev *dev, int enable); +typedef int (*vhost_set_vring_ready_op)(struct vhost_dev *dev); typedef bool (*vhost_requires_shm_log_op)(struct vhost_dev *dev); typedef int (*vhost_migration_done_op)(struct vhost_dev *dev, char *mac_addr); @@ -113,6 +115,7 @@ typedef int (*vhost_set_inflight_fd_op)(struct vhost_dev *dev, struct vhost_inflight *inflight); typedef int (*vhost_set_state_op)(struct vhost_dev *dev, uint8_t state); +typedef int (*vhost_get_device_id_op)(struct vhost_dev *dev, uint32_t *dev_id); typedef struct VhostOps { VhostBackendType backend_type; vhost_backend_init vhost_backend_init; @@ -139,6 +142,7 @@ typedef struct VhostOps { vhost_reset_device_op vhost_reset_device; vhost_get_vq_index_op vhost_get_vq_index; vhost_set_vring_enable_op vhost_set_vring_enable; + vhost_set_vring_ready_op vhost_set_vring_ready; vhost_requires_shm_log_op vhost_requires_shm_log; vhost_migration_done_op vhost_migration_done; vhost_backend_can_merge_op vhost_backend_can_merge; @@ -154,9 +158,11 @@ typedef struct VhostOps { vhost_get_inflight_fd_op vhost_get_inflight_fd; vhost_set_inflight_fd_op vhost_set_inflight_fd; vhost_set_state_op vhost_set_state; + vhost_get_device_id_op vhost_get_device_id; } VhostOps; extern const VhostOps user_ops; +extern const VhostOps vdpa_ops; int vhost_set_backend_type(struct vhost_dev *dev, VhostBackendType backend_type); diff --git a/include/hw/virtio/vhost-vdpa.h b/include/hw/virtio/vhost-vdpa.h new file mode 100644 index 0000000000..9d64fbfea9 --- /dev/null +++ b/include/hw/virtio/vhost-vdpa.h @@ -0,0 +1,25 @@ +/* + * vhost-vdpa.h + * + * Copyright(c) 2017-2018 Intel Corporation. + * Copyright(c) 2020 Red Hat, Inc. + * + * This work is licensed under the terms of the GNU GPL, version 2 or later. + * See the COPYING file in the top-level directory. + * + */ + +#ifndef HW_VIRTIO_VHOST_VDPA_H +#define HW_VIRTIO_VHOST_VDPA_H + +#include "hw/virtio/virtio.h" + +typedef struct vhost_vdpa { + int device_fd; + uint32_t msg_type; + MemoryListener listener; +} VhostVDPA; + +extern AddressSpace address_space_memory; + +#endif diff --git a/include/hw/virtio/vhost.h b/include/hw/virtio/vhost.h index 085450c6f8..020956d5a9 100644 --- a/include/hw/virtio/vhost.h +++ b/include/hw/virtio/vhost.h @@ -124,6 +124,7 @@ int vhost_dev_get_config(struct vhost_dev *dev, uint8_t *config, uint32_t config_len); int vhost_dev_set_config(struct vhost_dev *dev, const uint8_t *data, uint32_t offset, uint32_t size, uint32_t flags); +int vhost_dev_get_device_id(struct vhost_dev *hdev, uint32_t *device_id); /* notifier callback in case vhost device config space changed */ void vhost_dev_set_config_notifier(struct vhost_dev *dev,