11e0a84eaSCindy Lu /* 21e0a84eaSCindy Lu * vhost-vdpa.c 31e0a84eaSCindy Lu * 41e0a84eaSCindy Lu * Copyright(c) 2017-2018 Intel Corporation. 51e0a84eaSCindy Lu * Copyright(c) 2020 Red Hat, Inc. 61e0a84eaSCindy Lu * 71e0a84eaSCindy Lu * This work is licensed under the terms of the GNU GPL, version 2 or later. 81e0a84eaSCindy Lu * See the COPYING file in the top-level directory. 91e0a84eaSCindy Lu * 101e0a84eaSCindy Lu */ 111e0a84eaSCindy Lu 121e0a84eaSCindy Lu #include "qemu/osdep.h" 131e0a84eaSCindy Lu #include "clients.h" 14bd907ae4SEugenio Pérez #include "hw/virtio/virtio-net.h" 151e0a84eaSCindy Lu #include "net/vhost_net.h" 161e0a84eaSCindy Lu #include "net/vhost-vdpa.h" 171e0a84eaSCindy Lu #include "hw/virtio/vhost-vdpa.h" 181e0a84eaSCindy Lu #include "qemu/config-file.h" 191e0a84eaSCindy Lu #include "qemu/error-report.h" 20bd907ae4SEugenio Pérez #include "qemu/log.h" 21bd907ae4SEugenio Pérez #include "qemu/memalign.h" 221e0a84eaSCindy Lu #include "qemu/option.h" 231e0a84eaSCindy Lu #include "qapi/error.h" 2440237840SJason Wang #include <linux/vhost.h> 251e0a84eaSCindy Lu #include <sys/ioctl.h> 261e0a84eaSCindy Lu #include <err.h> 271e0a84eaSCindy Lu #include "standard-headers/linux/virtio_net.h" 281e0a84eaSCindy Lu #include "monitor/monitor.h" 2969498430SEugenio Pérez #include "migration/migration.h" 3069498430SEugenio Pérez #include "migration/misc.h" 311e0a84eaSCindy Lu #include "hw/virtio/vhost.h" 321e0a84eaSCindy Lu 331e0a84eaSCindy Lu /* Todo:need to add the multiqueue support here */ 341e0a84eaSCindy Lu typedef struct VhostVDPAState { 351e0a84eaSCindy Lu NetClientState nc; 361e0a84eaSCindy Lu struct vhost_vdpa vhost_vdpa; 3769498430SEugenio Pérez Notifier migration_state; 381e0a84eaSCindy Lu VHostNetState *vhost_net; 392df4dd31SEugenio Pérez 402df4dd31SEugenio Pérez /* Control commands shadow buffers */ 4117fb889fSEugenio Pérez void *cvq_cmd_out_buffer; 4217fb889fSEugenio Pérez virtio_net_ctrl_ack *status; 4317fb889fSEugenio Pérez 447f211a28SEugenio Pérez /* The device always have SVQ enabled */ 457f211a28SEugenio Pérez bool always_svq; 46152128d6SEugenio Pérez 47152128d6SEugenio Pérez /* The device can isolate CVQ in its own ASID */ 48152128d6SEugenio Pérez bool cvq_isolated; 49152128d6SEugenio Pérez 501e0a84eaSCindy Lu bool started; 511e0a84eaSCindy Lu } VhostVDPAState; 521e0a84eaSCindy Lu 532875a0caSHawkins Jiawei /* 542875a0caSHawkins Jiawei * The array is sorted alphabetically in ascending order, 552875a0caSHawkins Jiawei * with the exception of VHOST_INVALID_FEATURE_BIT, 562875a0caSHawkins Jiawei * which should always be the last entry. 572875a0caSHawkins Jiawei */ 581e0a84eaSCindy Lu const int vdpa_feature_bits[] = { 591e0a84eaSCindy Lu VIRTIO_F_ANY_LAYOUT, 602875a0caSHawkins Jiawei VIRTIO_F_IOMMU_PLATFORM, 612875a0caSHawkins Jiawei VIRTIO_F_NOTIFY_ON_EMPTY, 622875a0caSHawkins Jiawei VIRTIO_F_RING_PACKED, 632875a0caSHawkins Jiawei VIRTIO_F_RING_RESET, 641e0a84eaSCindy Lu VIRTIO_F_VERSION_1, 651e0a84eaSCindy Lu VIRTIO_NET_F_CSUM, 6651e84244SEugenio Pérez VIRTIO_NET_F_CTRL_GUEST_OFFLOADS, 672875a0caSHawkins Jiawei VIRTIO_NET_F_CTRL_MAC_ADDR, 6840237840SJason Wang VIRTIO_NET_F_CTRL_RX, 6940237840SJason Wang VIRTIO_NET_F_CTRL_RX_EXTRA, 7040237840SJason Wang VIRTIO_NET_F_CTRL_VLAN, 7140237840SJason Wang VIRTIO_NET_F_CTRL_VQ, 722875a0caSHawkins Jiawei VIRTIO_NET_F_GSO, 732875a0caSHawkins Jiawei VIRTIO_NET_F_GUEST_CSUM, 742875a0caSHawkins Jiawei VIRTIO_NET_F_GUEST_ECN, 752875a0caSHawkins Jiawei VIRTIO_NET_F_GUEST_TSO4, 762875a0caSHawkins Jiawei VIRTIO_NET_F_GUEST_TSO6, 772875a0caSHawkins Jiawei VIRTIO_NET_F_GUEST_UFO, 789da16849SAndrew Melnychenko VIRTIO_NET_F_GUEST_USO4, 799da16849SAndrew Melnychenko VIRTIO_NET_F_GUEST_USO6, 800145c393SAndrew Melnychenko VIRTIO_NET_F_HASH_REPORT, 812875a0caSHawkins Jiawei VIRTIO_NET_F_HOST_ECN, 822875a0caSHawkins Jiawei VIRTIO_NET_F_HOST_TSO4, 832875a0caSHawkins Jiawei VIRTIO_NET_F_HOST_TSO6, 842875a0caSHawkins Jiawei VIRTIO_NET_F_HOST_UFO, 859da16849SAndrew Melnychenko VIRTIO_NET_F_HOST_USO, 862875a0caSHawkins Jiawei VIRTIO_NET_F_MQ, 872875a0caSHawkins Jiawei VIRTIO_NET_F_MRG_RXBUF, 882875a0caSHawkins Jiawei VIRTIO_NET_F_MTU, 892875a0caSHawkins Jiawei VIRTIO_NET_F_RSS, 909aa47eddSSi-Wei Liu VIRTIO_NET_F_STATUS, 912875a0caSHawkins Jiawei VIRTIO_RING_F_EVENT_IDX, 922875a0caSHawkins Jiawei VIRTIO_RING_F_INDIRECT_DESC, 932875a0caSHawkins Jiawei 942875a0caSHawkins Jiawei /* VHOST_INVALID_FEATURE_BIT should always be the last entry */ 951e0a84eaSCindy Lu VHOST_INVALID_FEATURE_BIT 961e0a84eaSCindy Lu }; 971e0a84eaSCindy Lu 981576dbb5SEugenio Pérez /** Supported device specific feature bits with SVQ */ 991576dbb5SEugenio Pérez static const uint64_t vdpa_svq_device_features = 1001576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_CSUM) | 1011576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_GUEST_CSUM) | 1024b4a1378SHawkins Jiawei BIT_ULL(VIRTIO_NET_F_CTRL_GUEST_OFFLOADS) | 1031576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_MTU) | 1041576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_MAC) | 1051576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_GUEST_TSO4) | 1061576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_GUEST_TSO6) | 1071576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_GUEST_ECN) | 1081576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_GUEST_UFO) | 1091576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_HOST_TSO4) | 1101576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_HOST_TSO6) | 1111576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_HOST_ECN) | 1121576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_HOST_UFO) | 1131576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_MRG_RXBUF) | 1141576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_STATUS) | 1151576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_CTRL_VQ) | 116ea6eec49SHawkins Jiawei BIT_ULL(VIRTIO_NET_F_CTRL_RX) | 117e213c45aSHawkins Jiawei BIT_ULL(VIRTIO_NET_F_CTRL_VLAN) | 118d669b7bbSHawkins Jiawei BIT_ULL(VIRTIO_NET_F_CTRL_RX_EXTRA) | 11972b99a87SEugenio Pérez BIT_ULL(VIRTIO_NET_F_MQ) | 1201576dbb5SEugenio Pérez BIT_ULL(VIRTIO_F_ANY_LAYOUT) | 1211576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_CTRL_MAC_ADDR) | 122609ab4c3SEugenio Pérez /* VHOST_F_LOG_ALL is exposed by SVQ */ 123609ab4c3SEugenio Pérez BIT_ULL(VHOST_F_LOG_ALL) | 1241576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_RSC_EXT) | 1250d74e2b7SEugenio Pérez BIT_ULL(VIRTIO_NET_F_STANDBY) | 1260d74e2b7SEugenio Pérez BIT_ULL(VIRTIO_NET_F_SPEED_DUPLEX); 1271576dbb5SEugenio Pérez 128c1a10086SEugenio Pérez #define VHOST_VDPA_NET_CVQ_ASID 1 129c1a10086SEugenio Pérez 1301e0a84eaSCindy Lu VHostNetState *vhost_vdpa_get_vhost_net(NetClientState *nc) 1311e0a84eaSCindy Lu { 1321e0a84eaSCindy Lu VhostVDPAState *s = DO_UPCAST(VhostVDPAState, nc, nc); 1331e0a84eaSCindy Lu assert(nc->info->type == NET_CLIENT_DRIVER_VHOST_VDPA); 1341e0a84eaSCindy Lu return s->vhost_net; 1351e0a84eaSCindy Lu } 1361e0a84eaSCindy Lu 137915bf6ccSEugenio Pérez static size_t vhost_vdpa_net_cvq_cmd_len(void) 138915bf6ccSEugenio Pérez { 139915bf6ccSEugenio Pérez /* 140915bf6ccSEugenio Pérez * MAC_TABLE_SET is the ctrl command that produces the longer out buffer. 141915bf6ccSEugenio Pérez * In buffer is always 1 byte, so it should fit here 142915bf6ccSEugenio Pérez */ 143915bf6ccSEugenio Pérez return sizeof(struct virtio_net_ctrl_hdr) + 144915bf6ccSEugenio Pérez 2 * sizeof(struct virtio_net_ctrl_mac) + 145915bf6ccSEugenio Pérez MAC_TABLE_ENTRIES * ETH_ALEN; 146915bf6ccSEugenio Pérez } 147915bf6ccSEugenio Pérez 148915bf6ccSEugenio Pérez static size_t vhost_vdpa_net_cvq_cmd_page_len(void) 149915bf6ccSEugenio Pérez { 150915bf6ccSEugenio Pérez return ROUND_UP(vhost_vdpa_net_cvq_cmd_len(), qemu_real_host_page_size()); 151915bf6ccSEugenio Pérez } 152915bf6ccSEugenio Pérez 15336e46472SEugenio Pérez static bool vhost_vdpa_net_valid_svq_features(uint64_t features, Error **errp) 15436e46472SEugenio Pérez { 15536e46472SEugenio Pérez uint64_t invalid_dev_features = 15636e46472SEugenio Pérez features & ~vdpa_svq_device_features & 15736e46472SEugenio Pérez /* Transport are all accepted at this point */ 15836e46472SEugenio Pérez ~MAKE_64BIT_MASK(VIRTIO_TRANSPORT_F_START, 15936e46472SEugenio Pérez VIRTIO_TRANSPORT_F_END - VIRTIO_TRANSPORT_F_START); 16036e46472SEugenio Pérez 16136e46472SEugenio Pérez if (invalid_dev_features) { 16236e46472SEugenio Pérez error_setg(errp, "vdpa svq does not work with features 0x%" PRIx64, 16336e46472SEugenio Pérez invalid_dev_features); 164258a0394SEugenio Pérez return false; 16536e46472SEugenio Pérez } 16636e46472SEugenio Pérez 167258a0394SEugenio Pérez return vhost_svq_valid_features(features, errp); 16836e46472SEugenio Pérez } 16936e46472SEugenio Pérez 1701e0a84eaSCindy Lu static int vhost_vdpa_net_check_device_id(struct vhost_net *net) 1711e0a84eaSCindy Lu { 1721e0a84eaSCindy Lu uint32_t device_id; 1731e0a84eaSCindy Lu int ret; 1741e0a84eaSCindy Lu struct vhost_dev *hdev; 1751e0a84eaSCindy Lu 1761e0a84eaSCindy Lu hdev = (struct vhost_dev *)&net->dev; 1771e0a84eaSCindy Lu ret = hdev->vhost_ops->vhost_get_device_id(hdev, &device_id); 1781e0a84eaSCindy Lu if (device_id != VIRTIO_ID_NET) { 1791e0a84eaSCindy Lu return -ENOTSUP; 1801e0a84eaSCindy Lu } 1811e0a84eaSCindy Lu return ret; 1821e0a84eaSCindy Lu } 1831e0a84eaSCindy Lu 18440237840SJason Wang static int vhost_vdpa_add(NetClientState *ncs, void *be, 18540237840SJason Wang int queue_pair_index, int nvqs) 1861e0a84eaSCindy Lu { 1871e0a84eaSCindy Lu VhostNetOptions options; 1881e0a84eaSCindy Lu struct vhost_net *net = NULL; 1891e0a84eaSCindy Lu VhostVDPAState *s; 1901e0a84eaSCindy Lu int ret; 1911e0a84eaSCindy Lu 1921e0a84eaSCindy Lu options.backend_type = VHOST_BACKEND_TYPE_VDPA; 1931e0a84eaSCindy Lu assert(ncs->info->type == NET_CLIENT_DRIVER_VHOST_VDPA); 1941e0a84eaSCindy Lu s = DO_UPCAST(VhostVDPAState, nc, ncs); 1951e0a84eaSCindy Lu options.net_backend = ncs; 1961e0a84eaSCindy Lu options.opaque = be; 1971e0a84eaSCindy Lu options.busyloop_timeout = 0; 19840237840SJason Wang options.nvqs = nvqs; 1991e0a84eaSCindy Lu 2001e0a84eaSCindy Lu net = vhost_net_init(&options); 2011e0a84eaSCindy Lu if (!net) { 2021e0a84eaSCindy Lu error_report("failed to init vhost_net for queue"); 203a97ef87aSJason Wang goto err_init; 2041e0a84eaSCindy Lu } 2051e0a84eaSCindy Lu s->vhost_net = net; 2061e0a84eaSCindy Lu ret = vhost_vdpa_net_check_device_id(net); 2071e0a84eaSCindy Lu if (ret) { 208a97ef87aSJason Wang goto err_check; 2091e0a84eaSCindy Lu } 2101e0a84eaSCindy Lu return 0; 211a97ef87aSJason Wang err_check: 2121e0a84eaSCindy Lu vhost_net_cleanup(net); 213ab36edcfSJason Wang g_free(net); 214a97ef87aSJason Wang err_init: 2151e0a84eaSCindy Lu return -1; 2161e0a84eaSCindy Lu } 2171e0a84eaSCindy Lu 2181e0a84eaSCindy Lu static void vhost_vdpa_cleanup(NetClientState *nc) 2191e0a84eaSCindy Lu { 2201e0a84eaSCindy Lu VhostVDPAState *s = DO_UPCAST(VhostVDPAState, nc, nc); 2211e0a84eaSCindy Lu 222a0d7215eSAni Sinha /* 223a0d7215eSAni Sinha * If a peer NIC is attached, do not cleanup anything. 224a0d7215eSAni Sinha * Cleanup will happen as a part of qemu_cleanup() -> net_cleanup() 225a0d7215eSAni Sinha * when the guest is shutting down. 226a0d7215eSAni Sinha */ 227a0d7215eSAni Sinha if (nc->peer && nc->peer->info->type == NET_CLIENT_DRIVER_NIC) { 228a0d7215eSAni Sinha return; 229a0d7215eSAni Sinha } 230babf8b87SEugenio Pérez munmap(s->cvq_cmd_out_buffer, vhost_vdpa_net_cvq_cmd_page_len()); 231babf8b87SEugenio Pérez munmap(s->status, vhost_vdpa_net_cvq_cmd_page_len()); 2321e0a84eaSCindy Lu if (s->vhost_net) { 2331e0a84eaSCindy Lu vhost_net_cleanup(s->vhost_net); 2341e0a84eaSCindy Lu g_free(s->vhost_net); 2351e0a84eaSCindy Lu s->vhost_net = NULL; 2361e0a84eaSCindy Lu } 23757b3a7d8SCindy Lu if (s->vhost_vdpa.device_fd >= 0) { 23857b3a7d8SCindy Lu qemu_close(s->vhost_vdpa.device_fd); 23957b3a7d8SCindy Lu s->vhost_vdpa.device_fd = -1; 24057b3a7d8SCindy Lu } 2411e0a84eaSCindy Lu } 2421e0a84eaSCindy Lu 2431e0a84eaSCindy Lu static bool vhost_vdpa_has_vnet_hdr(NetClientState *nc) 2441e0a84eaSCindy Lu { 2451e0a84eaSCindy Lu assert(nc->info->type == NET_CLIENT_DRIVER_VHOST_VDPA); 2461e0a84eaSCindy Lu 2471e0a84eaSCindy Lu return true; 2481e0a84eaSCindy Lu } 2491e0a84eaSCindy Lu 2501e0a84eaSCindy Lu static bool vhost_vdpa_has_ufo(NetClientState *nc) 2511e0a84eaSCindy Lu { 2521e0a84eaSCindy Lu assert(nc->info->type == NET_CLIENT_DRIVER_VHOST_VDPA); 2531e0a84eaSCindy Lu VhostVDPAState *s = DO_UPCAST(VhostVDPAState, nc, nc); 2541e0a84eaSCindy Lu uint64_t features = 0; 2551e0a84eaSCindy Lu features |= (1ULL << VIRTIO_NET_F_HOST_UFO); 2561e0a84eaSCindy Lu features = vhost_net_get_features(s->vhost_net, features); 2571e0a84eaSCindy Lu return !!(features & (1ULL << VIRTIO_NET_F_HOST_UFO)); 2581e0a84eaSCindy Lu 2591e0a84eaSCindy Lu } 2601e0a84eaSCindy Lu 261ee8a1c63SKevin Wolf static bool vhost_vdpa_check_peer_type(NetClientState *nc, ObjectClass *oc, 262ee8a1c63SKevin Wolf Error **errp) 263ee8a1c63SKevin Wolf { 264ee8a1c63SKevin Wolf const char *driver = object_class_get_name(oc); 265ee8a1c63SKevin Wolf 266ee8a1c63SKevin Wolf if (!g_str_has_prefix(driver, "virtio-net-")) { 267ee8a1c63SKevin Wolf error_setg(errp, "vhost-vdpa requires frontend driver virtio-net-*"); 268ee8a1c63SKevin Wolf return false; 269ee8a1c63SKevin Wolf } 270ee8a1c63SKevin Wolf 271ee8a1c63SKevin Wolf return true; 272ee8a1c63SKevin Wolf } 273ee8a1c63SKevin Wolf 274846a1e85SEugenio Pérez /** Dummy receive in case qemu falls back to userland tap networking */ 275846a1e85SEugenio Pérez static ssize_t vhost_vdpa_receive(NetClientState *nc, const uint8_t *buf, 276846a1e85SEugenio Pérez size_t size) 277846a1e85SEugenio Pérez { 278bc5add1dSSi-Wei Liu return size; 279846a1e85SEugenio Pérez } 280846a1e85SEugenio Pérez 28100ef422eSEugenio Pérez /** From any vdpa net client, get the netclient of the first queue pair */ 28200ef422eSEugenio Pérez static VhostVDPAState *vhost_vdpa_net_first_nc_vdpa(VhostVDPAState *s) 28300ef422eSEugenio Pérez { 28400ef422eSEugenio Pérez NICState *nic = qemu_get_nic(s->nc.peer); 28500ef422eSEugenio Pérez NetClientState *nc0 = qemu_get_peer(nic->ncs, 0); 28600ef422eSEugenio Pérez 28700ef422eSEugenio Pérez return DO_UPCAST(VhostVDPAState, nc, nc0); 28800ef422eSEugenio Pérez } 28900ef422eSEugenio Pérez 29069498430SEugenio Pérez static void vhost_vdpa_net_log_global_enable(VhostVDPAState *s, bool enable) 29169498430SEugenio Pérez { 29269498430SEugenio Pérez struct vhost_vdpa *v = &s->vhost_vdpa; 29369498430SEugenio Pérez VirtIONet *n; 29469498430SEugenio Pérez VirtIODevice *vdev; 29569498430SEugenio Pérez int data_queue_pairs, cvq, r; 29669498430SEugenio Pérez 29769498430SEugenio Pérez /* We are only called on the first data vqs and only if x-svq is not set */ 29869498430SEugenio Pérez if (s->vhost_vdpa.shadow_vqs_enabled == enable) { 29969498430SEugenio Pérez return; 30069498430SEugenio Pérez } 30169498430SEugenio Pérez 30269498430SEugenio Pérez vdev = v->dev->vdev; 30369498430SEugenio Pérez n = VIRTIO_NET(vdev); 30469498430SEugenio Pérez if (!n->vhost_started) { 30569498430SEugenio Pérez return; 30669498430SEugenio Pérez } 30769498430SEugenio Pérez 30869498430SEugenio Pérez data_queue_pairs = n->multiqueue ? n->max_queue_pairs : 1; 30969498430SEugenio Pérez cvq = virtio_vdev_has_feature(vdev, VIRTIO_NET_F_CTRL_VQ) ? 31069498430SEugenio Pérez n->max_ncs - n->max_queue_pairs : 0; 31169498430SEugenio Pérez /* 31269498430SEugenio Pérez * TODO: vhost_net_stop does suspend, get_base and reset. We can be smarter 31369498430SEugenio Pérez * in the future and resume the device if read-only operations between 31469498430SEugenio Pérez * suspend and reset goes wrong. 31569498430SEugenio Pérez */ 31669498430SEugenio Pérez vhost_net_stop(vdev, n->nic->ncs, data_queue_pairs, cvq); 31769498430SEugenio Pérez 31869498430SEugenio Pérez /* Start will check migration setup_or_active to configure or not SVQ */ 31969498430SEugenio Pérez r = vhost_net_start(vdev, n->nic->ncs, data_queue_pairs, cvq); 32069498430SEugenio Pérez if (unlikely(r < 0)) { 32169498430SEugenio Pérez error_report("unable to start vhost net: %s(%d)", g_strerror(-r), -r); 32269498430SEugenio Pérez } 32369498430SEugenio Pérez } 32469498430SEugenio Pérez 32569498430SEugenio Pérez static void vdpa_net_migration_state_notifier(Notifier *notifier, void *data) 32669498430SEugenio Pérez { 32769498430SEugenio Pérez MigrationState *migration = data; 32869498430SEugenio Pérez VhostVDPAState *s = container_of(notifier, VhostVDPAState, 32969498430SEugenio Pérez migration_state); 33069498430SEugenio Pérez 33169498430SEugenio Pérez if (migration_in_setup(migration)) { 33269498430SEugenio Pérez vhost_vdpa_net_log_global_enable(s, true); 33369498430SEugenio Pérez } else if (migration_has_failed(migration)) { 33469498430SEugenio Pérez vhost_vdpa_net_log_global_enable(s, false); 33569498430SEugenio Pérez } 33669498430SEugenio Pérez } 33769498430SEugenio Pérez 33800ef422eSEugenio Pérez static void vhost_vdpa_net_data_start_first(VhostVDPAState *s) 33900ef422eSEugenio Pérez { 34000ef422eSEugenio Pérez struct vhost_vdpa *v = &s->vhost_vdpa; 34100ef422eSEugenio Pérez 342d9cda213SSteve Sistare migration_add_notifier(&s->migration_state, 343d9cda213SSteve Sistare vdpa_net_migration_state_notifier); 34400ef422eSEugenio Pérez if (v->shadow_vqs_enabled) { 34500ef422eSEugenio Pérez v->iova_tree = vhost_iova_tree_new(v->iova_range.first, 34600ef422eSEugenio Pérez v->iova_range.last); 34700ef422eSEugenio Pérez } 34800ef422eSEugenio Pérez } 34900ef422eSEugenio Pérez 35000ef422eSEugenio Pérez static int vhost_vdpa_net_data_start(NetClientState *nc) 35100ef422eSEugenio Pérez { 35200ef422eSEugenio Pérez VhostVDPAState *s = DO_UPCAST(VhostVDPAState, nc, nc); 35300ef422eSEugenio Pérez struct vhost_vdpa *v = &s->vhost_vdpa; 35400ef422eSEugenio Pérez 35500ef422eSEugenio Pérez assert(nc->info->type == NET_CLIENT_DRIVER_VHOST_VDPA); 35600ef422eSEugenio Pérez 35769498430SEugenio Pérez if (s->always_svq || 35869498430SEugenio Pérez migration_is_setup_or_active(migrate_get_current()->state)) { 35969498430SEugenio Pérez v->shadow_vqs_enabled = true; 36069498430SEugenio Pérez v->shadow_data = true; 36169498430SEugenio Pérez } else { 36269498430SEugenio Pérez v->shadow_vqs_enabled = false; 36369498430SEugenio Pérez v->shadow_data = false; 36469498430SEugenio Pérez } 36569498430SEugenio Pérez 36600ef422eSEugenio Pérez if (v->index == 0) { 36700ef422eSEugenio Pérez vhost_vdpa_net_data_start_first(s); 36800ef422eSEugenio Pérez return 0; 36900ef422eSEugenio Pérez } 37000ef422eSEugenio Pérez 37100ef422eSEugenio Pérez if (v->shadow_vqs_enabled) { 37200ef422eSEugenio Pérez VhostVDPAState *s0 = vhost_vdpa_net_first_nc_vdpa(s); 37300ef422eSEugenio Pérez v->iova_tree = s0->vhost_vdpa.iova_tree; 37400ef422eSEugenio Pérez } 37500ef422eSEugenio Pérez 37600ef422eSEugenio Pérez return 0; 37700ef422eSEugenio Pérez } 37800ef422eSEugenio Pérez 3796c482547SEugenio Pérez static int vhost_vdpa_net_data_load(NetClientState *nc) 3806c482547SEugenio Pérez { 3816c482547SEugenio Pérez VhostVDPAState *s = DO_UPCAST(VhostVDPAState, nc, nc); 3826c482547SEugenio Pérez struct vhost_vdpa *v = &s->vhost_vdpa; 3836c482547SEugenio Pérez bool has_cvq = v->dev->vq_index_end % 2; 3846c482547SEugenio Pérez 3856c482547SEugenio Pérez if (has_cvq) { 3866c482547SEugenio Pérez return 0; 3876c482547SEugenio Pérez } 3886c482547SEugenio Pérez 3896c482547SEugenio Pérez for (int i = 0; i < v->dev->nvqs; ++i) { 3906c482547SEugenio Pérez vhost_vdpa_set_vring_ready(v, i + v->dev->vq_index); 3916c482547SEugenio Pérez } 3926c482547SEugenio Pérez return 0; 3936c482547SEugenio Pérez } 3946c482547SEugenio Pérez 39500ef422eSEugenio Pérez static void vhost_vdpa_net_client_stop(NetClientState *nc) 39600ef422eSEugenio Pérez { 39700ef422eSEugenio Pérez VhostVDPAState *s = DO_UPCAST(VhostVDPAState, nc, nc); 39800ef422eSEugenio Pérez struct vhost_dev *dev; 39900ef422eSEugenio Pérez 40000ef422eSEugenio Pérez assert(nc->info->type == NET_CLIENT_DRIVER_VHOST_VDPA); 40100ef422eSEugenio Pérez 40269498430SEugenio Pérez if (s->vhost_vdpa.index == 0) { 403d9cda213SSteve Sistare migration_remove_notifier(&s->migration_state); 40469498430SEugenio Pérez } 40569498430SEugenio Pérez 40600ef422eSEugenio Pérez dev = s->vhost_vdpa.dev; 40700ef422eSEugenio Pérez if (dev->vq_index + dev->nvqs == dev->vq_index_end) { 40800ef422eSEugenio Pérez g_clear_pointer(&s->vhost_vdpa.iova_tree, vhost_iova_tree_delete); 4090a7a164bSEugenio Pérez } else { 4100a7a164bSEugenio Pérez s->vhost_vdpa.iova_tree = NULL; 41100ef422eSEugenio Pérez } 41200ef422eSEugenio Pérez } 41300ef422eSEugenio Pérez 4141e0a84eaSCindy Lu static NetClientInfo net_vhost_vdpa_info = { 4151e0a84eaSCindy Lu .type = NET_CLIENT_DRIVER_VHOST_VDPA, 4161e0a84eaSCindy Lu .size = sizeof(VhostVDPAState), 417846a1e85SEugenio Pérez .receive = vhost_vdpa_receive, 41800ef422eSEugenio Pérez .start = vhost_vdpa_net_data_start, 4196c482547SEugenio Pérez .load = vhost_vdpa_net_data_load, 42000ef422eSEugenio Pérez .stop = vhost_vdpa_net_client_stop, 4211e0a84eaSCindy Lu .cleanup = vhost_vdpa_cleanup, 4221e0a84eaSCindy Lu .has_vnet_hdr = vhost_vdpa_has_vnet_hdr, 4231e0a84eaSCindy Lu .has_ufo = vhost_vdpa_has_ufo, 424ee8a1c63SKevin Wolf .check_peer_type = vhost_vdpa_check_peer_type, 4251e0a84eaSCindy Lu }; 4261e0a84eaSCindy Lu 427152128d6SEugenio Pérez static int64_t vhost_vdpa_get_vring_group(int device_fd, unsigned vq_index, 428152128d6SEugenio Pérez Error **errp) 429c1a10086SEugenio Pérez { 430c1a10086SEugenio Pérez struct vhost_vring_state state = { 431c1a10086SEugenio Pérez .index = vq_index, 432c1a10086SEugenio Pérez }; 433c1a10086SEugenio Pérez int r = ioctl(device_fd, VHOST_VDPA_GET_VRING_GROUP, &state); 434c1a10086SEugenio Pérez 435c1a10086SEugenio Pérez if (unlikely(r < 0)) { 4360f2bb0bfSEugenio Pérez r = -errno; 437152128d6SEugenio Pérez error_setg_errno(errp, errno, "Cannot get VQ %u group", vq_index); 438c1a10086SEugenio Pérez return r; 439c1a10086SEugenio Pérez } 440c1a10086SEugenio Pérez 441c1a10086SEugenio Pérez return state.num; 442c1a10086SEugenio Pérez } 443c1a10086SEugenio Pérez 444c1a10086SEugenio Pérez static int vhost_vdpa_set_address_space_id(struct vhost_vdpa *v, 445c1a10086SEugenio Pérez unsigned vq_group, 446c1a10086SEugenio Pérez unsigned asid_num) 447c1a10086SEugenio Pérez { 448c1a10086SEugenio Pérez struct vhost_vring_state asid = { 449c1a10086SEugenio Pérez .index = vq_group, 450c1a10086SEugenio Pérez .num = asid_num, 451c1a10086SEugenio Pérez }; 452c1a10086SEugenio Pérez int r; 453c1a10086SEugenio Pérez 454c1a10086SEugenio Pérez r = ioctl(v->device_fd, VHOST_VDPA_SET_GROUP_ASID, &asid); 455c1a10086SEugenio Pérez if (unlikely(r < 0)) { 456c1a10086SEugenio Pérez error_report("Can't set vq group %u asid %u, errno=%d (%s)", 457c1a10086SEugenio Pérez asid.index, asid.num, errno, g_strerror(errno)); 458c1a10086SEugenio Pérez } 459c1a10086SEugenio Pérez return r; 460c1a10086SEugenio Pérez } 461c1a10086SEugenio Pérez 4622df4dd31SEugenio Pérez static void vhost_vdpa_cvq_unmap_buf(struct vhost_vdpa *v, void *addr) 4632df4dd31SEugenio Pérez { 4642df4dd31SEugenio Pérez VhostIOVATree *tree = v->iova_tree; 4652df4dd31SEugenio Pérez DMAMap needle = { 4662df4dd31SEugenio Pérez /* 4672df4dd31SEugenio Pérez * No need to specify size or to look for more translations since 4682df4dd31SEugenio Pérez * this contiguous chunk was allocated by us. 4692df4dd31SEugenio Pérez */ 4702df4dd31SEugenio Pérez .translated_addr = (hwaddr)(uintptr_t)addr, 4712df4dd31SEugenio Pérez }; 4722df4dd31SEugenio Pérez const DMAMap *map = vhost_iova_tree_find_iova(tree, &needle); 4732df4dd31SEugenio Pérez int r; 4742df4dd31SEugenio Pérez 4752df4dd31SEugenio Pérez if (unlikely(!map)) { 4762df4dd31SEugenio Pérez error_report("Cannot locate expected map"); 4772df4dd31SEugenio Pérez return; 4782df4dd31SEugenio Pérez } 4792df4dd31SEugenio Pérez 480cd831ed5SEugenio Pérez r = vhost_vdpa_dma_unmap(v, v->address_space_id, map->iova, map->size + 1); 4812df4dd31SEugenio Pérez if (unlikely(r != 0)) { 4822df4dd31SEugenio Pérez error_report("Device cannot unmap: %s(%d)", g_strerror(r), r); 4832df4dd31SEugenio Pérez } 4842df4dd31SEugenio Pérez 48569292a8eSEugenio Pérez vhost_iova_tree_remove(tree, *map); 4862df4dd31SEugenio Pérez } 4872df4dd31SEugenio Pérez 4887a7f87e9SEugenio Pérez /** Map CVQ buffer. */ 4897a7f87e9SEugenio Pérez static int vhost_vdpa_cvq_map_buf(struct vhost_vdpa *v, void *buf, size_t size, 4907a7f87e9SEugenio Pérez bool write) 4912df4dd31SEugenio Pérez { 4922df4dd31SEugenio Pérez DMAMap map = {}; 4932df4dd31SEugenio Pérez int r; 4942df4dd31SEugenio Pérez 4952df4dd31SEugenio Pérez map.translated_addr = (hwaddr)(uintptr_t)buf; 4967a7f87e9SEugenio Pérez map.size = size - 1; 4972df4dd31SEugenio Pérez map.perm = write ? IOMMU_RW : IOMMU_RO, 4982df4dd31SEugenio Pérez r = vhost_iova_tree_map_alloc(v->iova_tree, &map); 4992df4dd31SEugenio Pérez if (unlikely(r != IOVA_OK)) { 5002df4dd31SEugenio Pérez error_report("Cannot map injected element"); 5017a7f87e9SEugenio Pérez return r; 5022df4dd31SEugenio Pérez } 5032df4dd31SEugenio Pérez 504cd831ed5SEugenio Pérez r = vhost_vdpa_dma_map(v, v->address_space_id, map.iova, 505cd831ed5SEugenio Pérez vhost_vdpa_net_cvq_cmd_page_len(), buf, !write); 5062df4dd31SEugenio Pérez if (unlikely(r < 0)) { 5072df4dd31SEugenio Pérez goto dma_map_err; 5082df4dd31SEugenio Pérez } 5092df4dd31SEugenio Pérez 5107a7f87e9SEugenio Pérez return 0; 5112df4dd31SEugenio Pérez 5122df4dd31SEugenio Pérez dma_map_err: 51369292a8eSEugenio Pérez vhost_iova_tree_remove(v->iova_tree, map); 5147a7f87e9SEugenio Pérez return r; 5152df4dd31SEugenio Pérez } 5162df4dd31SEugenio Pérez 5177a7f87e9SEugenio Pérez static int vhost_vdpa_net_cvq_start(NetClientState *nc) 5182df4dd31SEugenio Pérez { 51900ef422eSEugenio Pérez VhostVDPAState *s, *s0; 520c1a10086SEugenio Pérez struct vhost_vdpa *v; 521c1a10086SEugenio Pérez int64_t cvq_group; 522152128d6SEugenio Pérez int r; 523152128d6SEugenio Pérez Error *err = NULL; 5242df4dd31SEugenio Pérez 5257a7f87e9SEugenio Pérez assert(nc->info->type == NET_CLIENT_DRIVER_VHOST_VDPA); 5267a7f87e9SEugenio Pérez 5277a7f87e9SEugenio Pérez s = DO_UPCAST(VhostVDPAState, nc, nc); 528c1a10086SEugenio Pérez v = &s->vhost_vdpa; 529c1a10086SEugenio Pérez 53069498430SEugenio Pérez s0 = vhost_vdpa_net_first_nc_vdpa(s); 53169498430SEugenio Pérez v->shadow_data = s0->vhost_vdpa.shadow_vqs_enabled; 532b40eba9cSEugenio Pérez v->shadow_vqs_enabled = s0->vhost_vdpa.shadow_vqs_enabled; 533c1a10086SEugenio Pérez s->vhost_vdpa.address_space_id = VHOST_VDPA_GUEST_PA_ASID; 534c1a10086SEugenio Pérez 53569498430SEugenio Pérez if (s->vhost_vdpa.shadow_data) { 536c1a10086SEugenio Pérez /* SVQ is already configured for all virtqueues */ 537c1a10086SEugenio Pérez goto out; 538c1a10086SEugenio Pérez } 539c1a10086SEugenio Pérez 540c1a10086SEugenio Pérez /* 541c1a10086SEugenio Pérez * If we early return in these cases SVQ will not be enabled. The migration 542c1a10086SEugenio Pérez * will be blocked as long as vhost-vdpa backends will not offer _F_LOG. 543c1a10086SEugenio Pérez */ 544152128d6SEugenio Pérez if (!vhost_vdpa_net_valid_svq_features(v->dev->features, NULL)) { 545c1a10086SEugenio Pérez return 0; 546c1a10086SEugenio Pérez } 547c1a10086SEugenio Pérez 548152128d6SEugenio Pérez if (!s->cvq_isolated) { 549152128d6SEugenio Pérez return 0; 550152128d6SEugenio Pérez } 551152128d6SEugenio Pérez 552152128d6SEugenio Pérez cvq_group = vhost_vdpa_get_vring_group(v->device_fd, 553152128d6SEugenio Pérez v->dev->vq_index_end - 1, 554152128d6SEugenio Pérez &err); 555c1a10086SEugenio Pérez if (unlikely(cvq_group < 0)) { 556152128d6SEugenio Pérez error_report_err(err); 557c1a10086SEugenio Pérez return cvq_group; 558c1a10086SEugenio Pérez } 559c1a10086SEugenio Pérez 560c1a10086SEugenio Pérez r = vhost_vdpa_set_address_space_id(v, cvq_group, VHOST_VDPA_NET_CVQ_ASID); 561c1a10086SEugenio Pérez if (unlikely(r < 0)) { 562c1a10086SEugenio Pérez return r; 563c1a10086SEugenio Pérez } 564c1a10086SEugenio Pérez 565c1a10086SEugenio Pérez v->shadow_vqs_enabled = true; 566c1a10086SEugenio Pérez s->vhost_vdpa.address_space_id = VHOST_VDPA_NET_CVQ_ASID; 567c1a10086SEugenio Pérez 568c1a10086SEugenio Pérez out: 5697a7f87e9SEugenio Pérez if (!s->vhost_vdpa.shadow_vqs_enabled) { 5707a7f87e9SEugenio Pérez return 0; 5712df4dd31SEugenio Pérez } 5722df4dd31SEugenio Pérez 57300ef422eSEugenio Pérez if (s0->vhost_vdpa.iova_tree) { 57400ef422eSEugenio Pérez /* 57500ef422eSEugenio Pérez * SVQ is already configured for all virtqueues. Reuse IOVA tree for 57600ef422eSEugenio Pérez * simplicity, whether CVQ shares ASID with guest or not, because: 57700ef422eSEugenio Pérez * - Memory listener need access to guest's memory addresses allocated 57800ef422eSEugenio Pérez * in the IOVA tree. 57900ef422eSEugenio Pérez * - There should be plenty of IOVA address space for both ASID not to 58000ef422eSEugenio Pérez * worry about collisions between them. Guest's translations are 58100ef422eSEugenio Pérez * still validated with virtio virtqueue_pop so there is no risk for 58200ef422eSEugenio Pérez * the guest to access memory that it shouldn't. 58300ef422eSEugenio Pérez * 58400ef422eSEugenio Pérez * To allocate a iova tree per ASID is doable but it complicates the 58500ef422eSEugenio Pérez * code and it is not worth it for the moment. 58600ef422eSEugenio Pérez */ 58700ef422eSEugenio Pérez v->iova_tree = s0->vhost_vdpa.iova_tree; 58800ef422eSEugenio Pérez } else { 58900ef422eSEugenio Pérez v->iova_tree = vhost_iova_tree_new(v->iova_range.first, 59000ef422eSEugenio Pérez v->iova_range.last); 59100ef422eSEugenio Pérez } 59200ef422eSEugenio Pérez 5937a7f87e9SEugenio Pérez r = vhost_vdpa_cvq_map_buf(&s->vhost_vdpa, s->cvq_cmd_out_buffer, 5947a7f87e9SEugenio Pérez vhost_vdpa_net_cvq_cmd_page_len(), false); 5957a7f87e9SEugenio Pérez if (unlikely(r < 0)) { 5967a7f87e9SEugenio Pérez return r; 5977a7f87e9SEugenio Pérez } 5987a7f87e9SEugenio Pérez 59917fb889fSEugenio Pérez r = vhost_vdpa_cvq_map_buf(&s->vhost_vdpa, s->status, 6007a7f87e9SEugenio Pérez vhost_vdpa_net_cvq_cmd_page_len(), true); 6017a7f87e9SEugenio Pérez if (unlikely(r < 0)) { 6022df4dd31SEugenio Pérez vhost_vdpa_cvq_unmap_buf(&s->vhost_vdpa, s->cvq_cmd_out_buffer); 6032df4dd31SEugenio Pérez } 6042df4dd31SEugenio Pérez 6057a7f87e9SEugenio Pérez return r; 6067a7f87e9SEugenio Pérez } 6077a7f87e9SEugenio Pérez 6087a7f87e9SEugenio Pérez static void vhost_vdpa_net_cvq_stop(NetClientState *nc) 6097a7f87e9SEugenio Pérez { 6107a7f87e9SEugenio Pérez VhostVDPAState *s = DO_UPCAST(VhostVDPAState, nc, nc); 6117a7f87e9SEugenio Pérez 6127a7f87e9SEugenio Pérez assert(nc->info->type == NET_CLIENT_DRIVER_VHOST_VDPA); 6137a7f87e9SEugenio Pérez 6147a7f87e9SEugenio Pérez if (s->vhost_vdpa.shadow_vqs_enabled) { 6157a7f87e9SEugenio Pérez vhost_vdpa_cvq_unmap_buf(&s->vhost_vdpa, s->cvq_cmd_out_buffer); 61617fb889fSEugenio Pérez vhost_vdpa_cvq_unmap_buf(&s->vhost_vdpa, s->status); 617c1a10086SEugenio Pérez } 61800ef422eSEugenio Pérez 61900ef422eSEugenio Pérez vhost_vdpa_net_client_stop(nc); 6202df4dd31SEugenio Pérez } 6212df4dd31SEugenio Pérez 6220e6bff0dSHawkins Jiawei static ssize_t vhost_vdpa_net_cvq_add(VhostVDPAState *s, 6230e6bff0dSHawkins Jiawei const struct iovec *out_sg, size_t out_num, 6240e6bff0dSHawkins Jiawei const struct iovec *in_sg, size_t in_num) 625be4278b6SEugenio Pérez { 626be4278b6SEugenio Pérez VhostShadowVirtqueue *svq = g_ptr_array_index(s->vhost_vdpa.shadow_vqs, 0); 627be4278b6SEugenio Pérez int r; 628be4278b6SEugenio Pérez 6290e6bff0dSHawkins Jiawei r = vhost_svq_add(svq, out_sg, out_num, in_sg, in_num, NULL); 630be4278b6SEugenio Pérez if (unlikely(r != 0)) { 631be4278b6SEugenio Pérez if (unlikely(r == -ENOSPC)) { 632be4278b6SEugenio Pérez qemu_log_mask(LOG_GUEST_ERROR, "%s: No space on device queue\n", 633be4278b6SEugenio Pérez __func__); 634be4278b6SEugenio Pérez } 635a864a321SHawkins Jiawei } 636a864a321SHawkins Jiawei 637be4278b6SEugenio Pérez return r; 638be4278b6SEugenio Pérez } 639be4278b6SEugenio Pérez 640be4278b6SEugenio Pérez /* 641a864a321SHawkins Jiawei * Convenience wrapper to poll SVQ for multiple control commands. 642a864a321SHawkins Jiawei * 643a864a321SHawkins Jiawei * Caller should hold the BQL when invoking this function, and should take 644a864a321SHawkins Jiawei * the answer before SVQ pulls by itself when BQL is released. 645be4278b6SEugenio Pérez */ 646a864a321SHawkins Jiawei static ssize_t vhost_vdpa_net_svq_poll(VhostVDPAState *s, size_t cmds_in_flight) 647a864a321SHawkins Jiawei { 648a864a321SHawkins Jiawei VhostShadowVirtqueue *svq = g_ptr_array_index(s->vhost_vdpa.shadow_vqs, 0); 649a864a321SHawkins Jiawei return vhost_svq_poll(svq, cmds_in_flight); 650be4278b6SEugenio Pérez } 651be4278b6SEugenio Pérez 6521d7e2a8fSHawkins Jiawei static void vhost_vdpa_net_load_cursor_reset(VhostVDPAState *s, 6531d7e2a8fSHawkins Jiawei struct iovec *out_cursor, 6541d7e2a8fSHawkins Jiawei struct iovec *in_cursor) 6551d7e2a8fSHawkins Jiawei { 6561d7e2a8fSHawkins Jiawei /* reset the cursor of the output buffer for the device */ 6571d7e2a8fSHawkins Jiawei out_cursor->iov_base = s->cvq_cmd_out_buffer; 6581d7e2a8fSHawkins Jiawei out_cursor->iov_len = vhost_vdpa_net_cvq_cmd_page_len(); 6591d7e2a8fSHawkins Jiawei 6601d7e2a8fSHawkins Jiawei /* reset the cursor of the in buffer for the device */ 6611d7e2a8fSHawkins Jiawei in_cursor->iov_base = s->status; 6621d7e2a8fSHawkins Jiawei in_cursor->iov_len = vhost_vdpa_net_cvq_cmd_page_len(); 6631d7e2a8fSHawkins Jiawei } 6641d7e2a8fSHawkins Jiawei 665acec5f68SHawkins Jiawei /* 666acec5f68SHawkins Jiawei * Poll SVQ for multiple pending control commands and check the device's ack. 667acec5f68SHawkins Jiawei * 668acec5f68SHawkins Jiawei * Caller should hold the BQL when invoking this function. 669acec5f68SHawkins Jiawei * 670acec5f68SHawkins Jiawei * @s: The VhostVDPAState 671acec5f68SHawkins Jiawei * @len: The length of the pending status shadow buffer 672acec5f68SHawkins Jiawei */ 673acec5f68SHawkins Jiawei static ssize_t vhost_vdpa_net_svq_flush(VhostVDPAState *s, size_t len) 674acec5f68SHawkins Jiawei { 675acec5f68SHawkins Jiawei /* device uses a one-byte length ack for each control command */ 676acec5f68SHawkins Jiawei ssize_t dev_written = vhost_vdpa_net_svq_poll(s, len); 677acec5f68SHawkins Jiawei if (unlikely(dev_written != len)) { 678acec5f68SHawkins Jiawei return -EIO; 679acec5f68SHawkins Jiawei } 680acec5f68SHawkins Jiawei 681acec5f68SHawkins Jiawei /* check the device's ack */ 682acec5f68SHawkins Jiawei for (int i = 0; i < len; ++i) { 683acec5f68SHawkins Jiawei if (s->status[i] != VIRTIO_NET_OK) { 684acec5f68SHawkins Jiawei return -EIO; 685acec5f68SHawkins Jiawei } 686acec5f68SHawkins Jiawei } 687acec5f68SHawkins Jiawei return 0; 688acec5f68SHawkins Jiawei } 689acec5f68SHawkins Jiawei 6901d7e2a8fSHawkins Jiawei static ssize_t vhost_vdpa_net_load_cmd(VhostVDPAState *s, 6911d7e2a8fSHawkins Jiawei struct iovec *out_cursor, 6921d7e2a8fSHawkins Jiawei struct iovec *in_cursor, uint8_t class, 6932848c6aaSHawkins Jiawei uint8_t cmd, const struct iovec *data_sg, 6942848c6aaSHawkins Jiawei size_t data_num) 695f73c0c43SEugenio Pérez { 696f73c0c43SEugenio Pérez const struct virtio_net_ctrl_hdr ctrl = { 697f73c0c43SEugenio Pérez .class = class, 698f73c0c43SEugenio Pérez .cmd = cmd, 699f73c0c43SEugenio Pérez }; 700acec5f68SHawkins Jiawei size_t data_size = iov_size(data_sg, data_num), cmd_size; 7011d7e2a8fSHawkins Jiawei struct iovec out, in; 702a864a321SHawkins Jiawei ssize_t r; 703acec5f68SHawkins Jiawei unsigned dummy_cursor_iov_cnt; 704acec5f68SHawkins Jiawei VhostShadowVirtqueue *svq = g_ptr_array_index(s->vhost_vdpa.shadow_vqs, 0); 705f73c0c43SEugenio Pérez 706f73c0c43SEugenio Pérez assert(data_size < vhost_vdpa_net_cvq_cmd_page_len() - sizeof(ctrl)); 707acec5f68SHawkins Jiawei cmd_size = sizeof(ctrl) + data_size; 708acec5f68SHawkins Jiawei if (vhost_svq_available_slots(svq) < 2 || 709acec5f68SHawkins Jiawei iov_size(out_cursor, 1) < cmd_size) { 710acec5f68SHawkins Jiawei /* 711acec5f68SHawkins Jiawei * It is time to flush all pending control commands if SVQ is full 712acec5f68SHawkins Jiawei * or control commands shadow buffers are full. 713acec5f68SHawkins Jiawei * 714acec5f68SHawkins Jiawei * We can poll here since we've had BQL from the time 715acec5f68SHawkins Jiawei * we sent the descriptor. 716acec5f68SHawkins Jiawei */ 717acec5f68SHawkins Jiawei r = vhost_vdpa_net_svq_flush(s, in_cursor->iov_base - 718acec5f68SHawkins Jiawei (void *)s->status); 719acec5f68SHawkins Jiawei if (unlikely(r < 0)) { 720acec5f68SHawkins Jiawei return r; 721f73c0c43SEugenio Pérez } 722f73c0c43SEugenio Pérez 723acec5f68SHawkins Jiawei vhost_vdpa_net_load_cursor_reset(s, out_cursor, in_cursor); 724acec5f68SHawkins Jiawei } 725f73c0c43SEugenio Pérez 726f73c0c43SEugenio Pérez /* pack the CVQ command header */ 7271d7e2a8fSHawkins Jiawei iov_from_buf(out_cursor, 1, 0, &ctrl, sizeof(ctrl)); 728f73c0c43SEugenio Pérez /* pack the CVQ command command-specific-data */ 729f73c0c43SEugenio Pérez iov_to_buf(data_sg, data_num, 0, 7301d7e2a8fSHawkins Jiawei out_cursor->iov_base + sizeof(ctrl), data_size); 7311d7e2a8fSHawkins Jiawei 7321d7e2a8fSHawkins Jiawei /* extract the required buffer from the cursor for output */ 733acec5f68SHawkins Jiawei iov_copy(&out, 1, out_cursor, 1, 0, cmd_size); 7341d7e2a8fSHawkins Jiawei /* extract the required buffer from the cursor for input */ 7351d7e2a8fSHawkins Jiawei iov_copy(&in, 1, in_cursor, 1, 0, sizeof(*s->status)); 736f73c0c43SEugenio Pérez 737a864a321SHawkins Jiawei r = vhost_vdpa_net_cvq_add(s, &out, 1, &in, 1); 738a864a321SHawkins Jiawei if (unlikely(r < 0)) { 739a864a321SHawkins Jiawei return r; 740a864a321SHawkins Jiawei } 741a864a321SHawkins Jiawei 742acec5f68SHawkins Jiawei /* iterate the cursors */ 743acec5f68SHawkins Jiawei dummy_cursor_iov_cnt = 1; 744acec5f68SHawkins Jiawei iov_discard_front(&out_cursor, &dummy_cursor_iov_cnt, cmd_size); 745acec5f68SHawkins Jiawei dummy_cursor_iov_cnt = 1; 746acec5f68SHawkins Jiawei iov_discard_front(&in_cursor, &dummy_cursor_iov_cnt, sizeof(*s->status)); 747acec5f68SHawkins Jiawei 748acec5f68SHawkins Jiawei return 0; 749f73c0c43SEugenio Pérez } 750f73c0c43SEugenio Pérez 7511d7e2a8fSHawkins Jiawei static int vhost_vdpa_net_load_mac(VhostVDPAState *s, const VirtIONet *n, 7521d7e2a8fSHawkins Jiawei struct iovec *out_cursor, 7531d7e2a8fSHawkins Jiawei struct iovec *in_cursor) 754f73c0c43SEugenio Pérez { 75502d3bf09SHawkins Jiawei if (virtio_vdev_has_feature(&n->parent_obj, VIRTIO_NET_F_CTRL_MAC_ADDR)) { 7562848c6aaSHawkins Jiawei const struct iovec data = { 7572848c6aaSHawkins Jiawei .iov_base = (void *)n->mac, 7582848c6aaSHawkins Jiawei .iov_len = sizeof(n->mac), 7592848c6aaSHawkins Jiawei }; 760acec5f68SHawkins Jiawei ssize_t r = vhost_vdpa_net_load_cmd(s, out_cursor, in_cursor, 7611d7e2a8fSHawkins Jiawei VIRTIO_NET_CTRL_MAC, 762f73c0c43SEugenio Pérez VIRTIO_NET_CTRL_MAC_ADDR_SET, 7632848c6aaSHawkins Jiawei &data, 1); 764acec5f68SHawkins Jiawei if (unlikely(r < 0)) { 765acec5f68SHawkins Jiawei return r; 766b479bc3cSHawkins Jiawei } 767f73c0c43SEugenio Pérez } 768f73c0c43SEugenio Pérez 7690ddcecb8SHawkins Jiawei /* 7700ddcecb8SHawkins Jiawei * According to VirtIO standard, "The device MUST have an 7710ddcecb8SHawkins Jiawei * empty MAC filtering table on reset.". 7720ddcecb8SHawkins Jiawei * 7730ddcecb8SHawkins Jiawei * Therefore, there is no need to send this CVQ command if the 7740ddcecb8SHawkins Jiawei * driver also sets an empty MAC filter table, which aligns with 7750ddcecb8SHawkins Jiawei * the device's defaults. 7760ddcecb8SHawkins Jiawei * 7770ddcecb8SHawkins Jiawei * Note that the device's defaults can mismatch the driver's 7780ddcecb8SHawkins Jiawei * configuration only at live migration. 7790ddcecb8SHawkins Jiawei */ 7800ddcecb8SHawkins Jiawei if (!virtio_vdev_has_feature(&n->parent_obj, VIRTIO_NET_F_CTRL_RX) || 7810ddcecb8SHawkins Jiawei n->mac_table.in_use == 0) { 7820ddcecb8SHawkins Jiawei return 0; 7830ddcecb8SHawkins Jiawei } 7840ddcecb8SHawkins Jiawei 7850ddcecb8SHawkins Jiawei uint32_t uni_entries = n->mac_table.first_multi, 7860ddcecb8SHawkins Jiawei uni_macs_size = uni_entries * ETH_ALEN, 7870ddcecb8SHawkins Jiawei mul_entries = n->mac_table.in_use - uni_entries, 7880ddcecb8SHawkins Jiawei mul_macs_size = mul_entries * ETH_ALEN; 7890ddcecb8SHawkins Jiawei struct virtio_net_ctrl_mac uni = { 7900ddcecb8SHawkins Jiawei .entries = cpu_to_le32(uni_entries), 7910ddcecb8SHawkins Jiawei }; 7920ddcecb8SHawkins Jiawei struct virtio_net_ctrl_mac mul = { 7930ddcecb8SHawkins Jiawei .entries = cpu_to_le32(mul_entries), 7940ddcecb8SHawkins Jiawei }; 7950ddcecb8SHawkins Jiawei const struct iovec data[] = { 7960ddcecb8SHawkins Jiawei { 7970ddcecb8SHawkins Jiawei .iov_base = &uni, 7980ddcecb8SHawkins Jiawei .iov_len = sizeof(uni), 7990ddcecb8SHawkins Jiawei }, { 8000ddcecb8SHawkins Jiawei .iov_base = n->mac_table.macs, 8010ddcecb8SHawkins Jiawei .iov_len = uni_macs_size, 8020ddcecb8SHawkins Jiawei }, { 8030ddcecb8SHawkins Jiawei .iov_base = &mul, 8040ddcecb8SHawkins Jiawei .iov_len = sizeof(mul), 8050ddcecb8SHawkins Jiawei }, { 8060ddcecb8SHawkins Jiawei .iov_base = &n->mac_table.macs[uni_macs_size], 8070ddcecb8SHawkins Jiawei .iov_len = mul_macs_size, 8080ddcecb8SHawkins Jiawei }, 8090ddcecb8SHawkins Jiawei }; 810acec5f68SHawkins Jiawei ssize_t r = vhost_vdpa_net_load_cmd(s, out_cursor, in_cursor, 8110ddcecb8SHawkins Jiawei VIRTIO_NET_CTRL_MAC, 8120ddcecb8SHawkins Jiawei VIRTIO_NET_CTRL_MAC_TABLE_SET, 8130ddcecb8SHawkins Jiawei data, ARRAY_SIZE(data)); 814acec5f68SHawkins Jiawei if (unlikely(r < 0)) { 815acec5f68SHawkins Jiawei return r; 8160ddcecb8SHawkins Jiawei } 8170ddcecb8SHawkins Jiawei 818f73c0c43SEugenio Pérez return 0; 819f73c0c43SEugenio Pérez } 820f73c0c43SEugenio Pérez 821*8b98c15fSHawkins Jiawei static int vhost_vdpa_net_load_rss(VhostVDPAState *s, const VirtIONet *n, 822*8b98c15fSHawkins Jiawei struct iovec *out_cursor, 823*8b98c15fSHawkins Jiawei struct iovec *in_cursor) 824*8b98c15fSHawkins Jiawei { 825*8b98c15fSHawkins Jiawei struct virtio_net_rss_config cfg = {}; 826*8b98c15fSHawkins Jiawei ssize_t r; 827*8b98c15fSHawkins Jiawei g_autofree uint16_t *table = NULL; 828*8b98c15fSHawkins Jiawei 829*8b98c15fSHawkins Jiawei /* 830*8b98c15fSHawkins Jiawei * According to VirtIO standard, "Initially the device has all hash 831*8b98c15fSHawkins Jiawei * types disabled and reports only VIRTIO_NET_HASH_REPORT_NONE.". 832*8b98c15fSHawkins Jiawei * 833*8b98c15fSHawkins Jiawei * Therefore, there is no need to send this CVQ command if the 834*8b98c15fSHawkins Jiawei * driver disables the all hash types, which aligns with 835*8b98c15fSHawkins Jiawei * the device's defaults. 836*8b98c15fSHawkins Jiawei * 837*8b98c15fSHawkins Jiawei * Note that the device's defaults can mismatch the driver's 838*8b98c15fSHawkins Jiawei * configuration only at live migration. 839*8b98c15fSHawkins Jiawei */ 840*8b98c15fSHawkins Jiawei if (!n->rss_data.enabled || 841*8b98c15fSHawkins Jiawei n->rss_data.hash_types == VIRTIO_NET_HASH_REPORT_NONE) { 842*8b98c15fSHawkins Jiawei return 0; 843*8b98c15fSHawkins Jiawei } 844*8b98c15fSHawkins Jiawei 845*8b98c15fSHawkins Jiawei table = g_malloc_n(n->rss_data.indirections_len, 846*8b98c15fSHawkins Jiawei sizeof(n->rss_data.indirections_table[0])); 847*8b98c15fSHawkins Jiawei cfg.hash_types = cpu_to_le32(n->rss_data.hash_types); 848*8b98c15fSHawkins Jiawei 849*8b98c15fSHawkins Jiawei /* 850*8b98c15fSHawkins Jiawei * According to VirtIO standard, "Field reserved MUST contain zeroes. 851*8b98c15fSHawkins Jiawei * It is defined to make the structure to match the layout of 852*8b98c15fSHawkins Jiawei * virtio_net_rss_config structure, defined in 5.1.6.5.7.". 853*8b98c15fSHawkins Jiawei * 854*8b98c15fSHawkins Jiawei * Therefore, we need to zero the fields in 855*8b98c15fSHawkins Jiawei * struct virtio_net_rss_config, which corresponds to the 856*8b98c15fSHawkins Jiawei * `reserved` field in struct virtio_net_hash_config. 857*8b98c15fSHawkins Jiawei * 858*8b98c15fSHawkins Jiawei * Note that all other fields are zeroed at their definitions, 859*8b98c15fSHawkins Jiawei * except for the `indirection_table` field, where the actual data 860*8b98c15fSHawkins Jiawei * is stored in the `table` variable to ensure compatibility 861*8b98c15fSHawkins Jiawei * with RSS case. Therefore, we need to zero the `table` variable here. 862*8b98c15fSHawkins Jiawei */ 863*8b98c15fSHawkins Jiawei table[0] = 0; 864*8b98c15fSHawkins Jiawei 865*8b98c15fSHawkins Jiawei /* 866*8b98c15fSHawkins Jiawei * Considering that virtio_net_handle_rss() currently does not restore 867*8b98c15fSHawkins Jiawei * the hash key length parsed from the CVQ command sent from the guest 868*8b98c15fSHawkins Jiawei * into n->rss_data and uses the maximum key length in other code, so 869*8b98c15fSHawkins Jiawei * we also employ the maximum key length here. 870*8b98c15fSHawkins Jiawei */ 871*8b98c15fSHawkins Jiawei cfg.hash_key_length = sizeof(n->rss_data.key); 872*8b98c15fSHawkins Jiawei 873*8b98c15fSHawkins Jiawei const struct iovec data[] = { 874*8b98c15fSHawkins Jiawei { 875*8b98c15fSHawkins Jiawei .iov_base = &cfg, 876*8b98c15fSHawkins Jiawei .iov_len = offsetof(struct virtio_net_rss_config, 877*8b98c15fSHawkins Jiawei indirection_table), 878*8b98c15fSHawkins Jiawei }, { 879*8b98c15fSHawkins Jiawei .iov_base = table, 880*8b98c15fSHawkins Jiawei .iov_len = n->rss_data.indirections_len * 881*8b98c15fSHawkins Jiawei sizeof(n->rss_data.indirections_table[0]), 882*8b98c15fSHawkins Jiawei }, { 883*8b98c15fSHawkins Jiawei .iov_base = &cfg.max_tx_vq, 884*8b98c15fSHawkins Jiawei .iov_len = offsetof(struct virtio_net_rss_config, hash_key_data) - 885*8b98c15fSHawkins Jiawei offsetof(struct virtio_net_rss_config, max_tx_vq), 886*8b98c15fSHawkins Jiawei }, { 887*8b98c15fSHawkins Jiawei .iov_base = (void *)n->rss_data.key, 888*8b98c15fSHawkins Jiawei .iov_len = sizeof(n->rss_data.key), 889*8b98c15fSHawkins Jiawei } 890*8b98c15fSHawkins Jiawei }; 891*8b98c15fSHawkins Jiawei 892*8b98c15fSHawkins Jiawei r = vhost_vdpa_net_load_cmd(s, out_cursor, in_cursor, 893*8b98c15fSHawkins Jiawei VIRTIO_NET_CTRL_MQ, 894*8b98c15fSHawkins Jiawei VIRTIO_NET_CTRL_MQ_HASH_CONFIG, 895*8b98c15fSHawkins Jiawei data, ARRAY_SIZE(data)); 896*8b98c15fSHawkins Jiawei if (unlikely(r < 0)) { 897*8b98c15fSHawkins Jiawei return r; 898*8b98c15fSHawkins Jiawei } 899*8b98c15fSHawkins Jiawei 900*8b98c15fSHawkins Jiawei return 0; 901*8b98c15fSHawkins Jiawei } 902*8b98c15fSHawkins Jiawei 903f64c7cdaSEugenio Pérez static int vhost_vdpa_net_load_mq(VhostVDPAState *s, 9041d7e2a8fSHawkins Jiawei const VirtIONet *n, 9051d7e2a8fSHawkins Jiawei struct iovec *out_cursor, 9061d7e2a8fSHawkins Jiawei struct iovec *in_cursor) 907f64c7cdaSEugenio Pérez { 908f64c7cdaSEugenio Pérez struct virtio_net_ctrl_mq mq; 909acec5f68SHawkins Jiawei ssize_t r; 910f64c7cdaSEugenio Pérez 91102d3bf09SHawkins Jiawei if (!virtio_vdev_has_feature(&n->parent_obj, VIRTIO_NET_F_MQ)) { 912f64c7cdaSEugenio Pérez return 0; 913f64c7cdaSEugenio Pérez } 914f64c7cdaSEugenio Pérez 915f64c7cdaSEugenio Pérez mq.virtqueue_pairs = cpu_to_le16(n->curr_queue_pairs); 9162848c6aaSHawkins Jiawei const struct iovec data = { 9172848c6aaSHawkins Jiawei .iov_base = &mq, 9182848c6aaSHawkins Jiawei .iov_len = sizeof(mq), 9192848c6aaSHawkins Jiawei }; 920acec5f68SHawkins Jiawei r = vhost_vdpa_net_load_cmd(s, out_cursor, in_cursor, 9211d7e2a8fSHawkins Jiawei VIRTIO_NET_CTRL_MQ, 9222848c6aaSHawkins Jiawei VIRTIO_NET_CTRL_MQ_VQ_PAIRS_SET, 9232848c6aaSHawkins Jiawei &data, 1); 924acec5f68SHawkins Jiawei if (unlikely(r < 0)) { 925acec5f68SHawkins Jiawei return r; 926f45fd95eSHawkins Jiawei } 927f64c7cdaSEugenio Pérez 928*8b98c15fSHawkins Jiawei if (!virtio_vdev_has_feature(&n->parent_obj, VIRTIO_NET_F_HASH_REPORT)) { 929*8b98c15fSHawkins Jiawei return 0; 930*8b98c15fSHawkins Jiawei } 931*8b98c15fSHawkins Jiawei 932*8b98c15fSHawkins Jiawei r = vhost_vdpa_net_load_rss(s, n, out_cursor, in_cursor); 933*8b98c15fSHawkins Jiawei if (unlikely(r < 0)) { 934*8b98c15fSHawkins Jiawei return r; 935*8b98c15fSHawkins Jiawei } 936*8b98c15fSHawkins Jiawei 937f45fd95eSHawkins Jiawei return 0; 938f64c7cdaSEugenio Pérez } 939f64c7cdaSEugenio Pérez 9400b58d368SHawkins Jiawei static int vhost_vdpa_net_load_offloads(VhostVDPAState *s, 9411d7e2a8fSHawkins Jiawei const VirtIONet *n, 9421d7e2a8fSHawkins Jiawei struct iovec *out_cursor, 9431d7e2a8fSHawkins Jiawei struct iovec *in_cursor) 9440b58d368SHawkins Jiawei { 9450b58d368SHawkins Jiawei uint64_t offloads; 946acec5f68SHawkins Jiawei ssize_t r; 9470b58d368SHawkins Jiawei 9480b58d368SHawkins Jiawei if (!virtio_vdev_has_feature(&n->parent_obj, 9490b58d368SHawkins Jiawei VIRTIO_NET_F_CTRL_GUEST_OFFLOADS)) { 9500b58d368SHawkins Jiawei return 0; 9510b58d368SHawkins Jiawei } 9520b58d368SHawkins Jiawei 9530b58d368SHawkins Jiawei if (n->curr_guest_offloads == virtio_net_supported_guest_offloads(n)) { 9540b58d368SHawkins Jiawei /* 9550b58d368SHawkins Jiawei * According to VirtIO standard, "Upon feature negotiation 9560b58d368SHawkins Jiawei * corresponding offload gets enabled to preserve 9570b58d368SHawkins Jiawei * backward compatibility.". 9580b58d368SHawkins Jiawei * 9590b58d368SHawkins Jiawei * Therefore, there is no need to send this CVQ command if the 9600b58d368SHawkins Jiawei * driver also enables all supported offloads, which aligns with 9610b58d368SHawkins Jiawei * the device's defaults. 9620b58d368SHawkins Jiawei * 9630b58d368SHawkins Jiawei * Note that the device's defaults can mismatch the driver's 9640b58d368SHawkins Jiawei * configuration only at live migration. 9650b58d368SHawkins Jiawei */ 9660b58d368SHawkins Jiawei return 0; 9670b58d368SHawkins Jiawei } 9680b58d368SHawkins Jiawei 9690b58d368SHawkins Jiawei offloads = cpu_to_le64(n->curr_guest_offloads); 9702848c6aaSHawkins Jiawei const struct iovec data = { 9712848c6aaSHawkins Jiawei .iov_base = &offloads, 9722848c6aaSHawkins Jiawei .iov_len = sizeof(offloads), 9732848c6aaSHawkins Jiawei }; 974acec5f68SHawkins Jiawei r = vhost_vdpa_net_load_cmd(s, out_cursor, in_cursor, 9751d7e2a8fSHawkins Jiawei VIRTIO_NET_CTRL_GUEST_OFFLOADS, 9760b58d368SHawkins Jiawei VIRTIO_NET_CTRL_GUEST_OFFLOADS_SET, 9772848c6aaSHawkins Jiawei &data, 1); 978acec5f68SHawkins Jiawei if (unlikely(r < 0)) { 979acec5f68SHawkins Jiawei return r; 9806f348071SHawkins Jiawei } 9810b58d368SHawkins Jiawei 9826f348071SHawkins Jiawei return 0; 9830b58d368SHawkins Jiawei } 9840b58d368SHawkins Jiawei 985b12f907eSHawkins Jiawei static int vhost_vdpa_net_load_rx_mode(VhostVDPAState *s, 9861d7e2a8fSHawkins Jiawei struct iovec *out_cursor, 9871d7e2a8fSHawkins Jiawei struct iovec *in_cursor, 988b12f907eSHawkins Jiawei uint8_t cmd, 989b12f907eSHawkins Jiawei uint8_t on) 990b12f907eSHawkins Jiawei { 991b12f907eSHawkins Jiawei const struct iovec data = { 992b12f907eSHawkins Jiawei .iov_base = &on, 993b12f907eSHawkins Jiawei .iov_len = sizeof(on), 994b12f907eSHawkins Jiawei }; 995acec5f68SHawkins Jiawei ssize_t r; 99624e59cfeSHawkins Jiawei 997acec5f68SHawkins Jiawei r = vhost_vdpa_net_load_cmd(s, out_cursor, in_cursor, 998acec5f68SHawkins Jiawei VIRTIO_NET_CTRL_RX, cmd, &data, 1); 999acec5f68SHawkins Jiawei if (unlikely(r < 0)) { 1000acec5f68SHawkins Jiawei return r; 100124e59cfeSHawkins Jiawei } 100224e59cfeSHawkins Jiawei 100324e59cfeSHawkins Jiawei return 0; 1004b12f907eSHawkins Jiawei } 1005b12f907eSHawkins Jiawei 1006b12f907eSHawkins Jiawei static int vhost_vdpa_net_load_rx(VhostVDPAState *s, 10071d7e2a8fSHawkins Jiawei const VirtIONet *n, 10081d7e2a8fSHawkins Jiawei struct iovec *out_cursor, 10091d7e2a8fSHawkins Jiawei struct iovec *in_cursor) 1010b12f907eSHawkins Jiawei { 101124e59cfeSHawkins Jiawei ssize_t r; 1012b12f907eSHawkins Jiawei 1013b12f907eSHawkins Jiawei if (!virtio_vdev_has_feature(&n->parent_obj, VIRTIO_NET_F_CTRL_RX)) { 1014b12f907eSHawkins Jiawei return 0; 1015b12f907eSHawkins Jiawei } 1016b12f907eSHawkins Jiawei 1017b12f907eSHawkins Jiawei /* 1018b12f907eSHawkins Jiawei * According to virtio_net_reset(), device turns promiscuous mode 1019b12f907eSHawkins Jiawei * on by default. 1020b12f907eSHawkins Jiawei * 10210a19d879SMichael Tokarev * Additionally, according to VirtIO standard, "Since there are 1022b12f907eSHawkins Jiawei * no guarantees, it can use a hash filter or silently switch to 1023b12f907eSHawkins Jiawei * allmulti or promiscuous mode if it is given too many addresses.". 1024b12f907eSHawkins Jiawei * QEMU marks `n->mac_table.uni_overflow` if guest sets too many 1025b12f907eSHawkins Jiawei * non-multicast MAC addresses, indicating that promiscuous mode 1026b12f907eSHawkins Jiawei * should be enabled. 1027b12f907eSHawkins Jiawei * 1028b12f907eSHawkins Jiawei * Therefore, QEMU should only send this CVQ command if the 1029b12f907eSHawkins Jiawei * `n->mac_table.uni_overflow` is not marked and `n->promisc` is off, 1030b12f907eSHawkins Jiawei * which sets promiscuous mode on, different from the device's defaults. 1031b12f907eSHawkins Jiawei * 1032b12f907eSHawkins Jiawei * Note that the device's defaults can mismatch the driver's 1033b12f907eSHawkins Jiawei * configuration only at live migration. 1034b12f907eSHawkins Jiawei */ 1035b12f907eSHawkins Jiawei if (!n->mac_table.uni_overflow && !n->promisc) { 10361d7e2a8fSHawkins Jiawei r = vhost_vdpa_net_load_rx_mode(s, out_cursor, in_cursor, 1037b12f907eSHawkins Jiawei VIRTIO_NET_CTRL_RX_PROMISC, 0); 103824e59cfeSHawkins Jiawei if (unlikely(r < 0)) { 103924e59cfeSHawkins Jiawei return r; 1040b12f907eSHawkins Jiawei } 1041b12f907eSHawkins Jiawei } 1042b12f907eSHawkins Jiawei 1043b12f907eSHawkins Jiawei /* 1044b12f907eSHawkins Jiawei * According to virtio_net_reset(), device turns all-multicast mode 1045b12f907eSHawkins Jiawei * off by default. 1046b12f907eSHawkins Jiawei * 1047b12f907eSHawkins Jiawei * According to VirtIO standard, "Since there are no guarantees, 1048b12f907eSHawkins Jiawei * it can use a hash filter or silently switch to allmulti or 1049b12f907eSHawkins Jiawei * promiscuous mode if it is given too many addresses.". QEMU marks 1050b12f907eSHawkins Jiawei * `n->mac_table.multi_overflow` if guest sets too many 1051b12f907eSHawkins Jiawei * non-multicast MAC addresses. 1052b12f907eSHawkins Jiawei * 1053b12f907eSHawkins Jiawei * Therefore, QEMU should only send this CVQ command if the 1054b12f907eSHawkins Jiawei * `n->mac_table.multi_overflow` is marked or `n->allmulti` is on, 1055b12f907eSHawkins Jiawei * which sets all-multicast mode on, different from the device's defaults. 1056b12f907eSHawkins Jiawei * 1057b12f907eSHawkins Jiawei * Note that the device's defaults can mismatch the driver's 1058b12f907eSHawkins Jiawei * configuration only at live migration. 1059b12f907eSHawkins Jiawei */ 1060b12f907eSHawkins Jiawei if (n->mac_table.multi_overflow || n->allmulti) { 10611d7e2a8fSHawkins Jiawei r = vhost_vdpa_net_load_rx_mode(s, out_cursor, in_cursor, 1062b12f907eSHawkins Jiawei VIRTIO_NET_CTRL_RX_ALLMULTI, 1); 106324e59cfeSHawkins Jiawei if (unlikely(r < 0)) { 106424e59cfeSHawkins Jiawei return r; 1065b12f907eSHawkins Jiawei } 1066b12f907eSHawkins Jiawei } 1067b12f907eSHawkins Jiawei 10684fd180c7SHawkins Jiawei if (!virtio_vdev_has_feature(&n->parent_obj, VIRTIO_NET_F_CTRL_RX_EXTRA)) { 10694fd180c7SHawkins Jiawei return 0; 10704fd180c7SHawkins Jiawei } 10714fd180c7SHawkins Jiawei 10724fd180c7SHawkins Jiawei /* 10734fd180c7SHawkins Jiawei * According to virtio_net_reset(), device turns all-unicast mode 10744fd180c7SHawkins Jiawei * off by default. 10754fd180c7SHawkins Jiawei * 10764fd180c7SHawkins Jiawei * Therefore, QEMU should only send this CVQ command if the driver 10774fd180c7SHawkins Jiawei * sets all-unicast mode on, different from the device's defaults. 10784fd180c7SHawkins Jiawei * 10794fd180c7SHawkins Jiawei * Note that the device's defaults can mismatch the driver's 10804fd180c7SHawkins Jiawei * configuration only at live migration. 10814fd180c7SHawkins Jiawei */ 10824fd180c7SHawkins Jiawei if (n->alluni) { 10831d7e2a8fSHawkins Jiawei r = vhost_vdpa_net_load_rx_mode(s, out_cursor, in_cursor, 10844fd180c7SHawkins Jiawei VIRTIO_NET_CTRL_RX_ALLUNI, 1); 108524e59cfeSHawkins Jiawei if (r < 0) { 108624e59cfeSHawkins Jiawei return r; 10874fd180c7SHawkins Jiawei } 10884fd180c7SHawkins Jiawei } 10894fd180c7SHawkins Jiawei 10904fd180c7SHawkins Jiawei /* 10914fd180c7SHawkins Jiawei * According to virtio_net_reset(), device turns non-multicast mode 10924fd180c7SHawkins Jiawei * off by default. 10934fd180c7SHawkins Jiawei * 10944fd180c7SHawkins Jiawei * Therefore, QEMU should only send this CVQ command if the driver 10954fd180c7SHawkins Jiawei * sets non-multicast mode on, different from the device's defaults. 10964fd180c7SHawkins Jiawei * 10974fd180c7SHawkins Jiawei * Note that the device's defaults can mismatch the driver's 10984fd180c7SHawkins Jiawei * configuration only at live migration. 10994fd180c7SHawkins Jiawei */ 11004fd180c7SHawkins Jiawei if (n->nomulti) { 11011d7e2a8fSHawkins Jiawei r = vhost_vdpa_net_load_rx_mode(s, out_cursor, in_cursor, 11024fd180c7SHawkins Jiawei VIRTIO_NET_CTRL_RX_NOMULTI, 1); 110324e59cfeSHawkins Jiawei if (r < 0) { 110424e59cfeSHawkins Jiawei return r; 11054fd180c7SHawkins Jiawei } 11064fd180c7SHawkins Jiawei } 11074fd180c7SHawkins Jiawei 11084fd180c7SHawkins Jiawei /* 11094fd180c7SHawkins Jiawei * According to virtio_net_reset(), device turns non-unicast mode 11104fd180c7SHawkins Jiawei * off by default. 11114fd180c7SHawkins Jiawei * 11124fd180c7SHawkins Jiawei * Therefore, QEMU should only send this CVQ command if the driver 11134fd180c7SHawkins Jiawei * sets non-unicast mode on, different from the device's defaults. 11144fd180c7SHawkins Jiawei * 11154fd180c7SHawkins Jiawei * Note that the device's defaults can mismatch the driver's 11164fd180c7SHawkins Jiawei * configuration only at live migration. 11174fd180c7SHawkins Jiawei */ 11184fd180c7SHawkins Jiawei if (n->nouni) { 11191d7e2a8fSHawkins Jiawei r = vhost_vdpa_net_load_rx_mode(s, out_cursor, in_cursor, 11204fd180c7SHawkins Jiawei VIRTIO_NET_CTRL_RX_NOUNI, 1); 112124e59cfeSHawkins Jiawei if (r < 0) { 112224e59cfeSHawkins Jiawei return r; 11234fd180c7SHawkins Jiawei } 11244fd180c7SHawkins Jiawei } 11254fd180c7SHawkins Jiawei 11264fd180c7SHawkins Jiawei /* 11274fd180c7SHawkins Jiawei * According to virtio_net_reset(), device turns non-broadcast mode 11284fd180c7SHawkins Jiawei * off by default. 11294fd180c7SHawkins Jiawei * 11304fd180c7SHawkins Jiawei * Therefore, QEMU should only send this CVQ command if the driver 11314fd180c7SHawkins Jiawei * sets non-broadcast mode on, different from the device's defaults. 11324fd180c7SHawkins Jiawei * 11334fd180c7SHawkins Jiawei * Note that the device's defaults can mismatch the driver's 11344fd180c7SHawkins Jiawei * configuration only at live migration. 11354fd180c7SHawkins Jiawei */ 11364fd180c7SHawkins Jiawei if (n->nobcast) { 11371d7e2a8fSHawkins Jiawei r = vhost_vdpa_net_load_rx_mode(s, out_cursor, in_cursor, 11384fd180c7SHawkins Jiawei VIRTIO_NET_CTRL_RX_NOBCAST, 1); 113924e59cfeSHawkins Jiawei if (r < 0) { 114024e59cfeSHawkins Jiawei return r; 11414fd180c7SHawkins Jiawei } 11424fd180c7SHawkins Jiawei } 11434fd180c7SHawkins Jiawei 1144b12f907eSHawkins Jiawei return 0; 1145b12f907eSHawkins Jiawei } 1146b12f907eSHawkins Jiawei 11478f7e9967SHawkins Jiawei static int vhost_vdpa_net_load_single_vlan(VhostVDPAState *s, 11488f7e9967SHawkins Jiawei const VirtIONet *n, 11491d7e2a8fSHawkins Jiawei struct iovec *out_cursor, 11501d7e2a8fSHawkins Jiawei struct iovec *in_cursor, 11518f7e9967SHawkins Jiawei uint16_t vid) 11528f7e9967SHawkins Jiawei { 11538f7e9967SHawkins Jiawei const struct iovec data = { 11548f7e9967SHawkins Jiawei .iov_base = &vid, 11558f7e9967SHawkins Jiawei .iov_len = sizeof(vid), 11568f7e9967SHawkins Jiawei }; 1157acec5f68SHawkins Jiawei ssize_t r = vhost_vdpa_net_load_cmd(s, out_cursor, in_cursor, 11581d7e2a8fSHawkins Jiawei VIRTIO_NET_CTRL_VLAN, 11598f7e9967SHawkins Jiawei VIRTIO_NET_CTRL_VLAN_ADD, 11608f7e9967SHawkins Jiawei &data, 1); 1161acec5f68SHawkins Jiawei if (unlikely(r < 0)) { 1162acec5f68SHawkins Jiawei return r; 11638f7e9967SHawkins Jiawei } 11648f7e9967SHawkins Jiawei 11658f7e9967SHawkins Jiawei return 0; 11668f7e9967SHawkins Jiawei } 11678f7e9967SHawkins Jiawei 11688f7e9967SHawkins Jiawei static int vhost_vdpa_net_load_vlan(VhostVDPAState *s, 11691d7e2a8fSHawkins Jiawei const VirtIONet *n, 11701d7e2a8fSHawkins Jiawei struct iovec *out_cursor, 11711d7e2a8fSHawkins Jiawei struct iovec *in_cursor) 11728f7e9967SHawkins Jiawei { 11738f7e9967SHawkins Jiawei int r; 11748f7e9967SHawkins Jiawei 11758f7e9967SHawkins Jiawei if (!virtio_vdev_has_feature(&n->parent_obj, VIRTIO_NET_F_CTRL_VLAN)) { 11768f7e9967SHawkins Jiawei return 0; 11778f7e9967SHawkins Jiawei } 11788f7e9967SHawkins Jiawei 11798f7e9967SHawkins Jiawei for (int i = 0; i < MAX_VLAN >> 5; i++) { 11808f7e9967SHawkins Jiawei for (int j = 0; n->vlans[i] && j <= 0x1f; j++) { 11818f7e9967SHawkins Jiawei if (n->vlans[i] & (1U << j)) { 11821d7e2a8fSHawkins Jiawei r = vhost_vdpa_net_load_single_vlan(s, n, out_cursor, 11831d7e2a8fSHawkins Jiawei in_cursor, (i << 5) + j); 11848f7e9967SHawkins Jiawei if (unlikely(r != 0)) { 11858f7e9967SHawkins Jiawei return r; 11868f7e9967SHawkins Jiawei } 11878f7e9967SHawkins Jiawei } 11888f7e9967SHawkins Jiawei } 11898f7e9967SHawkins Jiawei } 11908f7e9967SHawkins Jiawei 11918f7e9967SHawkins Jiawei return 0; 11928f7e9967SHawkins Jiawei } 11938f7e9967SHawkins Jiawei 1194f3fada59SEugenio Pérez static int vhost_vdpa_net_cvq_load(NetClientState *nc) 1195dd036d8dSEugenio Pérez { 1196dd036d8dSEugenio Pérez VhostVDPAState *s = DO_UPCAST(VhostVDPAState, nc, nc); 1197f73c0c43SEugenio Pérez struct vhost_vdpa *v = &s->vhost_vdpa; 1198dd036d8dSEugenio Pérez const VirtIONet *n; 1199f73c0c43SEugenio Pérez int r; 12001d7e2a8fSHawkins Jiawei struct iovec out_cursor, in_cursor; 1201dd036d8dSEugenio Pérez 1202dd036d8dSEugenio Pérez assert(nc->info->type == NET_CLIENT_DRIVER_VHOST_VDPA); 1203dd036d8dSEugenio Pérez 12046c482547SEugenio Pérez vhost_vdpa_set_vring_ready(v, v->dev->vq_index); 1205dd036d8dSEugenio Pérez 12066c482547SEugenio Pérez if (v->shadow_vqs_enabled) { 1207dd036d8dSEugenio Pérez n = VIRTIO_NET(v->dev->vdev); 12081d7e2a8fSHawkins Jiawei vhost_vdpa_net_load_cursor_reset(s, &out_cursor, &in_cursor); 12091d7e2a8fSHawkins Jiawei r = vhost_vdpa_net_load_mac(s, n, &out_cursor, &in_cursor); 1210f73c0c43SEugenio Pérez if (unlikely(r < 0)) { 1211f73c0c43SEugenio Pérez return r; 1212dd036d8dSEugenio Pérez } 12131d7e2a8fSHawkins Jiawei r = vhost_vdpa_net_load_mq(s, n, &out_cursor, &in_cursor); 1214f64c7cdaSEugenio Pérez if (unlikely(r)) { 1215f64c7cdaSEugenio Pérez return r; 1216f64c7cdaSEugenio Pérez } 12171d7e2a8fSHawkins Jiawei r = vhost_vdpa_net_load_offloads(s, n, &out_cursor, &in_cursor); 12180b58d368SHawkins Jiawei if (unlikely(r)) { 12190b58d368SHawkins Jiawei return r; 12200b58d368SHawkins Jiawei } 12211d7e2a8fSHawkins Jiawei r = vhost_vdpa_net_load_rx(s, n, &out_cursor, &in_cursor); 1222b12f907eSHawkins Jiawei if (unlikely(r)) { 1223b12f907eSHawkins Jiawei return r; 1224b12f907eSHawkins Jiawei } 12251d7e2a8fSHawkins Jiawei r = vhost_vdpa_net_load_vlan(s, n, &out_cursor, &in_cursor); 12268f7e9967SHawkins Jiawei if (unlikely(r)) { 12278f7e9967SHawkins Jiawei return r; 12288f7e9967SHawkins Jiawei } 1229acec5f68SHawkins Jiawei 1230acec5f68SHawkins Jiawei /* 1231acec5f68SHawkins Jiawei * We need to poll and check all pending device's used buffers. 1232acec5f68SHawkins Jiawei * 1233acec5f68SHawkins Jiawei * We can poll here since we've had BQL from the time 1234acec5f68SHawkins Jiawei * we sent the descriptor. 1235acec5f68SHawkins Jiawei */ 1236acec5f68SHawkins Jiawei r = vhost_vdpa_net_svq_flush(s, in_cursor.iov_base - (void *)s->status); 1237fee364e4SHawkins Jiawei if (unlikely(r)) { 1238fee364e4SHawkins Jiawei return r; 1239fee364e4SHawkins Jiawei } 12406c482547SEugenio Pérez } 12416c482547SEugenio Pérez 12426c482547SEugenio Pérez for (int i = 0; i < v->dev->vq_index; ++i) { 12436c482547SEugenio Pérez vhost_vdpa_set_vring_ready(v, i); 12446c482547SEugenio Pérez } 1245dd036d8dSEugenio Pérez 1246dd036d8dSEugenio Pérez return 0; 1247dd036d8dSEugenio Pérez } 1248dd036d8dSEugenio Pérez 1249f8972b56SEugenio Pérez static NetClientInfo net_vhost_vdpa_cvq_info = { 1250f8972b56SEugenio Pérez .type = NET_CLIENT_DRIVER_VHOST_VDPA, 1251f8972b56SEugenio Pérez .size = sizeof(VhostVDPAState), 1252f8972b56SEugenio Pérez .receive = vhost_vdpa_receive, 12537a7f87e9SEugenio Pérez .start = vhost_vdpa_net_cvq_start, 1254f3fada59SEugenio Pérez .load = vhost_vdpa_net_cvq_load, 12557a7f87e9SEugenio Pérez .stop = vhost_vdpa_net_cvq_stop, 1256f8972b56SEugenio Pérez .cleanup = vhost_vdpa_cleanup, 1257f8972b56SEugenio Pérez .has_vnet_hdr = vhost_vdpa_has_vnet_hdr, 1258f8972b56SEugenio Pérez .has_ufo = vhost_vdpa_has_ufo, 1259f8972b56SEugenio Pérez .check_peer_type = vhost_vdpa_check_peer_type, 1260f8972b56SEugenio Pérez }; 1261f8972b56SEugenio Pérez 1262fee364e4SHawkins Jiawei /* 1263fee364e4SHawkins Jiawei * Forward the excessive VIRTIO_NET_CTRL_MAC_TABLE_SET CVQ command to 1264fee364e4SHawkins Jiawei * vdpa device. 1265fee364e4SHawkins Jiawei * 1266fee364e4SHawkins Jiawei * Considering that QEMU cannot send the entire filter table to the 1267fee364e4SHawkins Jiawei * vdpa device, it should send the VIRTIO_NET_CTRL_RX_PROMISC CVQ 1268fee364e4SHawkins Jiawei * command to enable promiscuous mode to receive all packets, 1269fee364e4SHawkins Jiawei * according to VirtIO standard, "Since there are no guarantees, 1270fee364e4SHawkins Jiawei * it can use a hash filter or silently switch to allmulti or 1271fee364e4SHawkins Jiawei * promiscuous mode if it is given too many addresses.". 1272fee364e4SHawkins Jiawei * 1273fee364e4SHawkins Jiawei * Since QEMU ignores MAC addresses beyond `MAC_TABLE_ENTRIES` and 1274fee364e4SHawkins Jiawei * marks `n->mac_table.x_overflow` accordingly, it should have 1275fee364e4SHawkins Jiawei * the same effect on the device model to receive 1276fee364e4SHawkins Jiawei * (`MAC_TABLE_ENTRIES` + 1) or more non-multicast MAC addresses. 1277fee364e4SHawkins Jiawei * The same applies to multicast MAC addresses. 1278fee364e4SHawkins Jiawei * 1279fee364e4SHawkins Jiawei * Therefore, QEMU can provide the device model with a fake 1280fee364e4SHawkins Jiawei * VIRTIO_NET_CTRL_MAC_TABLE_SET command with (`MAC_TABLE_ENTRIES` + 1) 1281fee364e4SHawkins Jiawei * non-multicast MAC addresses and (`MAC_TABLE_ENTRIES` + 1) multicast 1282fee364e4SHawkins Jiawei * MAC addresses. This ensures that the device model marks 1283fee364e4SHawkins Jiawei * `n->mac_table.uni_overflow` and `n->mac_table.multi_overflow`, 1284fee364e4SHawkins Jiawei * allowing all packets to be received, which aligns with the 1285fee364e4SHawkins Jiawei * state of the vdpa device. 1286fee364e4SHawkins Jiawei */ 1287fee364e4SHawkins Jiawei static int vhost_vdpa_net_excessive_mac_filter_cvq_add(VhostVDPAState *s, 1288fee364e4SHawkins Jiawei VirtQueueElement *elem, 1289327dedb8SHawkins Jiawei struct iovec *out, 1290327dedb8SHawkins Jiawei const struct iovec *in) 1291fee364e4SHawkins Jiawei { 1292fee364e4SHawkins Jiawei struct virtio_net_ctrl_mac mac_data, *mac_ptr; 1293fee364e4SHawkins Jiawei struct virtio_net_ctrl_hdr *hdr_ptr; 1294fee364e4SHawkins Jiawei uint32_t cursor; 1295fee364e4SHawkins Jiawei ssize_t r; 1296327dedb8SHawkins Jiawei uint8_t on = 1; 1297fee364e4SHawkins Jiawei 1298fee364e4SHawkins Jiawei /* parse the non-multicast MAC address entries from CVQ command */ 1299fee364e4SHawkins Jiawei cursor = sizeof(*hdr_ptr); 1300fee364e4SHawkins Jiawei r = iov_to_buf(elem->out_sg, elem->out_num, cursor, 1301fee364e4SHawkins Jiawei &mac_data, sizeof(mac_data)); 1302fee364e4SHawkins Jiawei if (unlikely(r != sizeof(mac_data))) { 1303fee364e4SHawkins Jiawei /* 1304fee364e4SHawkins Jiawei * If the CVQ command is invalid, we should simulate the vdpa device 1305fee364e4SHawkins Jiawei * to reject the VIRTIO_NET_CTRL_MAC_TABLE_SET CVQ command 1306fee364e4SHawkins Jiawei */ 1307fee364e4SHawkins Jiawei *s->status = VIRTIO_NET_ERR; 1308fee364e4SHawkins Jiawei return sizeof(*s->status); 1309fee364e4SHawkins Jiawei } 1310fee364e4SHawkins Jiawei cursor += sizeof(mac_data) + le32_to_cpu(mac_data.entries) * ETH_ALEN; 1311fee364e4SHawkins Jiawei 1312fee364e4SHawkins Jiawei /* parse the multicast MAC address entries from CVQ command */ 1313fee364e4SHawkins Jiawei r = iov_to_buf(elem->out_sg, elem->out_num, cursor, 1314fee364e4SHawkins Jiawei &mac_data, sizeof(mac_data)); 1315fee364e4SHawkins Jiawei if (r != sizeof(mac_data)) { 1316fee364e4SHawkins Jiawei /* 1317fee364e4SHawkins Jiawei * If the CVQ command is invalid, we should simulate the vdpa device 1318fee364e4SHawkins Jiawei * to reject the VIRTIO_NET_CTRL_MAC_TABLE_SET CVQ command 1319fee364e4SHawkins Jiawei */ 1320fee364e4SHawkins Jiawei *s->status = VIRTIO_NET_ERR; 1321fee364e4SHawkins Jiawei return sizeof(*s->status); 1322fee364e4SHawkins Jiawei } 1323fee364e4SHawkins Jiawei cursor += sizeof(mac_data) + le32_to_cpu(mac_data.entries) * ETH_ALEN; 1324fee364e4SHawkins Jiawei 1325fee364e4SHawkins Jiawei /* validate the CVQ command */ 1326fee364e4SHawkins Jiawei if (iov_size(elem->out_sg, elem->out_num) != cursor) { 1327fee364e4SHawkins Jiawei /* 1328fee364e4SHawkins Jiawei * If the CVQ command is invalid, we should simulate the vdpa device 1329fee364e4SHawkins Jiawei * to reject the VIRTIO_NET_CTRL_MAC_TABLE_SET CVQ command 1330fee364e4SHawkins Jiawei */ 1331fee364e4SHawkins Jiawei *s->status = VIRTIO_NET_ERR; 1332fee364e4SHawkins Jiawei return sizeof(*s->status); 1333fee364e4SHawkins Jiawei } 1334fee364e4SHawkins Jiawei 1335fee364e4SHawkins Jiawei /* 1336fee364e4SHawkins Jiawei * According to VirtIO standard, "Since there are no guarantees, 1337fee364e4SHawkins Jiawei * it can use a hash filter or silently switch to allmulti or 1338fee364e4SHawkins Jiawei * promiscuous mode if it is given too many addresses.". 1339fee364e4SHawkins Jiawei * 1340fee364e4SHawkins Jiawei * Therefore, considering that QEMU is unable to send the entire 1341fee364e4SHawkins Jiawei * filter table to the vdpa device, it should send the 1342fee364e4SHawkins Jiawei * VIRTIO_NET_CTRL_RX_PROMISC CVQ command to enable promiscuous mode 1343fee364e4SHawkins Jiawei */ 1344327dedb8SHawkins Jiawei hdr_ptr = out->iov_base; 1345327dedb8SHawkins Jiawei out->iov_len = sizeof(*hdr_ptr) + sizeof(on); 1346327dedb8SHawkins Jiawei 1347327dedb8SHawkins Jiawei hdr_ptr->class = VIRTIO_NET_CTRL_RX; 1348327dedb8SHawkins Jiawei hdr_ptr->cmd = VIRTIO_NET_CTRL_RX_PROMISC; 1349327dedb8SHawkins Jiawei iov_from_buf(out, 1, sizeof(*hdr_ptr), &on, sizeof(on)); 1350327dedb8SHawkins Jiawei r = vhost_vdpa_net_cvq_add(s, out, 1, in, 1); 1351fee364e4SHawkins Jiawei if (unlikely(r < 0)) { 1352fee364e4SHawkins Jiawei return r; 1353fee364e4SHawkins Jiawei } 1354a864a321SHawkins Jiawei 1355a864a321SHawkins Jiawei /* 1356a864a321SHawkins Jiawei * We can poll here since we've had BQL from the time 1357a864a321SHawkins Jiawei * we sent the descriptor. 1358a864a321SHawkins Jiawei */ 1359a864a321SHawkins Jiawei r = vhost_vdpa_net_svq_poll(s, 1); 1360a864a321SHawkins Jiawei if (unlikely(r < sizeof(*s->status))) { 1361a864a321SHawkins Jiawei return r; 1362a864a321SHawkins Jiawei } 1363fee364e4SHawkins Jiawei if (*s->status != VIRTIO_NET_OK) { 1364fee364e4SHawkins Jiawei return sizeof(*s->status); 1365fee364e4SHawkins Jiawei } 1366fee364e4SHawkins Jiawei 1367fee364e4SHawkins Jiawei /* 1368fee364e4SHawkins Jiawei * QEMU should also send a fake VIRTIO_NET_CTRL_MAC_TABLE_SET CVQ 1369fee364e4SHawkins Jiawei * command to the device model, including (`MAC_TABLE_ENTRIES` + 1) 1370fee364e4SHawkins Jiawei * non-multicast MAC addresses and (`MAC_TABLE_ENTRIES` + 1) 1371fee364e4SHawkins Jiawei * multicast MAC addresses. 1372fee364e4SHawkins Jiawei * 1373fee364e4SHawkins Jiawei * By doing so, the device model can mark `n->mac_table.uni_overflow` 1374fee364e4SHawkins Jiawei * and `n->mac_table.multi_overflow`, enabling all packets to be 1375fee364e4SHawkins Jiawei * received, which aligns with the state of the vdpa device. 1376fee364e4SHawkins Jiawei */ 1377fee364e4SHawkins Jiawei cursor = 0; 1378fee364e4SHawkins Jiawei uint32_t fake_uni_entries = MAC_TABLE_ENTRIES + 1, 1379fee364e4SHawkins Jiawei fake_mul_entries = MAC_TABLE_ENTRIES + 1, 1380fee364e4SHawkins Jiawei fake_cvq_size = sizeof(struct virtio_net_ctrl_hdr) + 1381fee364e4SHawkins Jiawei sizeof(mac_data) + fake_uni_entries * ETH_ALEN + 1382fee364e4SHawkins Jiawei sizeof(mac_data) + fake_mul_entries * ETH_ALEN; 1383fee364e4SHawkins Jiawei 1384fee364e4SHawkins Jiawei assert(fake_cvq_size < vhost_vdpa_net_cvq_cmd_page_len()); 1385fee364e4SHawkins Jiawei out->iov_len = fake_cvq_size; 1386fee364e4SHawkins Jiawei 1387fee364e4SHawkins Jiawei /* pack the header for fake CVQ command */ 1388fee364e4SHawkins Jiawei hdr_ptr = out->iov_base + cursor; 1389fee364e4SHawkins Jiawei hdr_ptr->class = VIRTIO_NET_CTRL_MAC; 1390fee364e4SHawkins Jiawei hdr_ptr->cmd = VIRTIO_NET_CTRL_MAC_TABLE_SET; 1391fee364e4SHawkins Jiawei cursor += sizeof(*hdr_ptr); 1392fee364e4SHawkins Jiawei 1393fee364e4SHawkins Jiawei /* 1394fee364e4SHawkins Jiawei * Pack the non-multicast MAC addresses part for fake CVQ command. 1395fee364e4SHawkins Jiawei * 1396fee364e4SHawkins Jiawei * According to virtio_net_handle_mac(), QEMU doesn't verify the MAC 13970a19d879SMichael Tokarev * addresses provided in CVQ command. Therefore, only the entries 1398fee364e4SHawkins Jiawei * field need to be prepared in the CVQ command. 1399fee364e4SHawkins Jiawei */ 1400fee364e4SHawkins Jiawei mac_ptr = out->iov_base + cursor; 1401fee364e4SHawkins Jiawei mac_ptr->entries = cpu_to_le32(fake_uni_entries); 1402fee364e4SHawkins Jiawei cursor += sizeof(*mac_ptr) + fake_uni_entries * ETH_ALEN; 1403fee364e4SHawkins Jiawei 1404fee364e4SHawkins Jiawei /* 1405fee364e4SHawkins Jiawei * Pack the multicast MAC addresses part for fake CVQ command. 1406fee364e4SHawkins Jiawei * 1407fee364e4SHawkins Jiawei * According to virtio_net_handle_mac(), QEMU doesn't verify the MAC 14080a19d879SMichael Tokarev * addresses provided in CVQ command. Therefore, only the entries 1409fee364e4SHawkins Jiawei * field need to be prepared in the CVQ command. 1410fee364e4SHawkins Jiawei */ 1411fee364e4SHawkins Jiawei mac_ptr = out->iov_base + cursor; 1412fee364e4SHawkins Jiawei mac_ptr->entries = cpu_to_le32(fake_mul_entries); 1413fee364e4SHawkins Jiawei 1414fee364e4SHawkins Jiawei /* 1415fee364e4SHawkins Jiawei * Simulating QEMU poll a vdpa device used buffer 1416fee364e4SHawkins Jiawei * for VIRTIO_NET_CTRL_MAC_TABLE_SET CVQ command 1417fee364e4SHawkins Jiawei */ 1418fee364e4SHawkins Jiawei return sizeof(*s->status); 1419fee364e4SHawkins Jiawei } 1420fee364e4SHawkins Jiawei 14212df4dd31SEugenio Pérez /** 14222df4dd31SEugenio Pérez * Validate and copy control virtqueue commands. 14232df4dd31SEugenio Pérez * 14242df4dd31SEugenio Pérez * Following QEMU guidelines, we offer a copy of the buffers to the device to 14252df4dd31SEugenio Pérez * prevent TOCTOU bugs. 1426bd907ae4SEugenio Pérez */ 1427bd907ae4SEugenio Pérez static int vhost_vdpa_net_handle_ctrl_avail(VhostShadowVirtqueue *svq, 1428bd907ae4SEugenio Pérez VirtQueueElement *elem, 1429bd907ae4SEugenio Pérez void *opaque) 1430bd907ae4SEugenio Pérez { 14312df4dd31SEugenio Pérez VhostVDPAState *s = opaque; 1432be4278b6SEugenio Pérez size_t in_len; 143345c41018SHawkins Jiawei const struct virtio_net_ctrl_hdr *ctrl; 1434bd907ae4SEugenio Pérez virtio_net_ctrl_ack status = VIRTIO_NET_ERR; 14357a7f87e9SEugenio Pérez /* Out buffer sent to both the vdpa device and the device model */ 14367a7f87e9SEugenio Pérez struct iovec out = { 14377a7f87e9SEugenio Pérez .iov_base = s->cvq_cmd_out_buffer, 14387a7f87e9SEugenio Pérez }; 14392df4dd31SEugenio Pérez /* in buffer used for device model */ 14400e6bff0dSHawkins Jiawei const struct iovec model_in = { 14412df4dd31SEugenio Pérez .iov_base = &status, 14422df4dd31SEugenio Pérez .iov_len = sizeof(status), 14432df4dd31SEugenio Pérez }; 14440e6bff0dSHawkins Jiawei /* in buffer used for vdpa device */ 14450e6bff0dSHawkins Jiawei const struct iovec vdpa_in = { 14460e6bff0dSHawkins Jiawei .iov_base = s->status, 14470e6bff0dSHawkins Jiawei .iov_len = sizeof(*s->status), 14480e6bff0dSHawkins Jiawei }; 1449be4278b6SEugenio Pérez ssize_t dev_written = -EINVAL; 1450bd907ae4SEugenio Pérez 14517a7f87e9SEugenio Pérez out.iov_len = iov_to_buf(elem->out_sg, elem->out_num, 0, 14527a7f87e9SEugenio Pérez s->cvq_cmd_out_buffer, 1453fee364e4SHawkins Jiawei vhost_vdpa_net_cvq_cmd_page_len()); 145445c41018SHawkins Jiawei 145545c41018SHawkins Jiawei ctrl = s->cvq_cmd_out_buffer; 145645c41018SHawkins Jiawei if (ctrl->class == VIRTIO_NET_CTRL_ANNOUNCE) { 14573f9a3eebSEugenio Pérez /* 14583f9a3eebSEugenio Pérez * Guest announce capability is emulated by qemu, so don't forward to 14593f9a3eebSEugenio Pérez * the device. 14603f9a3eebSEugenio Pérez */ 14613f9a3eebSEugenio Pérez dev_written = sizeof(status); 14623f9a3eebSEugenio Pérez *s->status = VIRTIO_NET_OK; 1463fee364e4SHawkins Jiawei } else if (unlikely(ctrl->class == VIRTIO_NET_CTRL_MAC && 1464fee364e4SHawkins Jiawei ctrl->cmd == VIRTIO_NET_CTRL_MAC_TABLE_SET && 1465fee364e4SHawkins Jiawei iov_size(elem->out_sg, elem->out_num) > out.iov_len)) { 1466fee364e4SHawkins Jiawei /* 1467fee364e4SHawkins Jiawei * Due to the size limitation of the out buffer sent to the vdpa device, 1468fee364e4SHawkins Jiawei * which is determined by vhost_vdpa_net_cvq_cmd_page_len(), excessive 1469fee364e4SHawkins Jiawei * MAC addresses set by the driver for the filter table can cause 1470fee364e4SHawkins Jiawei * truncation of the CVQ command in QEMU. As a result, the vdpa device 1471fee364e4SHawkins Jiawei * rejects the flawed CVQ command. 1472fee364e4SHawkins Jiawei * 1473fee364e4SHawkins Jiawei * Therefore, QEMU must handle this situation instead of sending 14740a19d879SMichael Tokarev * the CVQ command directly. 1475fee364e4SHawkins Jiawei */ 1476fee364e4SHawkins Jiawei dev_written = vhost_vdpa_net_excessive_mac_filter_cvq_add(s, elem, 1477327dedb8SHawkins Jiawei &out, &vdpa_in); 1478fee364e4SHawkins Jiawei if (unlikely(dev_written < 0)) { 1479fee364e4SHawkins Jiawei goto out; 1480fee364e4SHawkins Jiawei } 14813f9a3eebSEugenio Pérez } else { 1482a864a321SHawkins Jiawei ssize_t r; 1483a864a321SHawkins Jiawei r = vhost_vdpa_net_cvq_add(s, &out, 1, &vdpa_in, 1); 1484a864a321SHawkins Jiawei if (unlikely(r < 0)) { 1485a864a321SHawkins Jiawei dev_written = r; 1486bd907ae4SEugenio Pérez goto out; 1487bd907ae4SEugenio Pérez } 1488a864a321SHawkins Jiawei 1489a864a321SHawkins Jiawei /* 1490a864a321SHawkins Jiawei * We can poll here since we've had BQL from the time 1491a864a321SHawkins Jiawei * we sent the descriptor. 1492a864a321SHawkins Jiawei */ 1493a864a321SHawkins Jiawei dev_written = vhost_vdpa_net_svq_poll(s, 1); 14943f9a3eebSEugenio Pérez } 1495bd907ae4SEugenio Pérez 1496bd907ae4SEugenio Pérez if (unlikely(dev_written < sizeof(status))) { 1497bd907ae4SEugenio Pérez error_report("Insufficient written data (%zu)", dev_written); 14982df4dd31SEugenio Pérez goto out; 14992df4dd31SEugenio Pérez } 15002df4dd31SEugenio Pérez 150117fb889fSEugenio Pérez if (*s->status != VIRTIO_NET_OK) { 1502d45243bcSEugenio Pérez goto out; 15032df4dd31SEugenio Pérez } 15042df4dd31SEugenio Pérez 15052df4dd31SEugenio Pérez status = VIRTIO_NET_ERR; 15060e6bff0dSHawkins Jiawei virtio_net_handle_ctrl_iov(svq->vdev, &model_in, 1, &out, 1); 15072df4dd31SEugenio Pérez if (status != VIRTIO_NET_OK) { 15082df4dd31SEugenio Pérez error_report("Bad CVQ processing in model"); 1509bd907ae4SEugenio Pérez } 1510bd907ae4SEugenio Pérez 1511bd907ae4SEugenio Pérez out: 1512bd907ae4SEugenio Pérez in_len = iov_from_buf(elem->in_sg, elem->in_num, 0, &status, 1513bd907ae4SEugenio Pérez sizeof(status)); 1514bd907ae4SEugenio Pérez if (unlikely(in_len < sizeof(status))) { 1515bd907ae4SEugenio Pérez error_report("Bad device CVQ written length"); 1516bd907ae4SEugenio Pérez } 1517bd907ae4SEugenio Pérez vhost_svq_push_elem(svq, elem, MIN(in_len, sizeof(status))); 1518031b1abaSHawkins Jiawei /* 1519031b1abaSHawkins Jiawei * `elem` belongs to vhost_vdpa_net_handle_ctrl_avail() only when 1520031b1abaSHawkins Jiawei * the function successfully forwards the CVQ command, indicated 1521031b1abaSHawkins Jiawei * by a non-negative value of `dev_written`. Otherwise, it still 1522031b1abaSHawkins Jiawei * belongs to SVQ. 1523031b1abaSHawkins Jiawei * This function should only free the `elem` when it owns. 1524031b1abaSHawkins Jiawei */ 1525031b1abaSHawkins Jiawei if (dev_written >= 0) { 1526bd907ae4SEugenio Pérez g_free(elem); 1527031b1abaSHawkins Jiawei } 1528be4278b6SEugenio Pérez return dev_written < 0 ? dev_written : 0; 1529bd907ae4SEugenio Pérez } 1530bd907ae4SEugenio Pérez 1531bd907ae4SEugenio Pérez static const VhostShadowVirtqueueOps vhost_vdpa_net_svq_ops = { 1532bd907ae4SEugenio Pérez .avail_handler = vhost_vdpa_net_handle_ctrl_avail, 1533bd907ae4SEugenio Pérez }; 1534bd907ae4SEugenio Pérez 1535152128d6SEugenio Pérez /** 1536152128d6SEugenio Pérez * Probe if CVQ is isolated 1537152128d6SEugenio Pérez * 1538152128d6SEugenio Pérez * @device_fd The vdpa device fd 1539152128d6SEugenio Pérez * @features Features offered by the device. 1540152128d6SEugenio Pérez * @cvq_index The control vq pair index 1541152128d6SEugenio Pérez * 1542152128d6SEugenio Pérez * Returns <0 in case of failure, 0 if false and 1 if true. 1543152128d6SEugenio Pérez */ 1544152128d6SEugenio Pérez static int vhost_vdpa_probe_cvq_isolation(int device_fd, uint64_t features, 1545152128d6SEugenio Pérez int cvq_index, Error **errp) 1546152128d6SEugenio Pérez { 1547152128d6SEugenio Pérez uint64_t backend_features; 1548152128d6SEugenio Pérez int64_t cvq_group; 1549152128d6SEugenio Pérez uint8_t status = VIRTIO_CONFIG_S_ACKNOWLEDGE | 1550845ec38aSEugenio Pérez VIRTIO_CONFIG_S_DRIVER; 1551152128d6SEugenio Pérez int r; 1552152128d6SEugenio Pérez 1553152128d6SEugenio Pérez ERRP_GUARD(); 1554152128d6SEugenio Pérez 1555152128d6SEugenio Pérez r = ioctl(device_fd, VHOST_GET_BACKEND_FEATURES, &backend_features); 1556152128d6SEugenio Pérez if (unlikely(r < 0)) { 1557152128d6SEugenio Pérez error_setg_errno(errp, errno, "Cannot get vdpa backend_features"); 1558152128d6SEugenio Pérez return r; 1559152128d6SEugenio Pérez } 1560152128d6SEugenio Pérez 1561152128d6SEugenio Pérez if (!(backend_features & BIT_ULL(VHOST_BACKEND_F_IOTLB_ASID))) { 1562152128d6SEugenio Pérez return 0; 1563152128d6SEugenio Pérez } 1564152128d6SEugenio Pérez 1565845ec38aSEugenio Pérez r = ioctl(device_fd, VHOST_VDPA_SET_STATUS, &status); 1566152128d6SEugenio Pérez if (unlikely(r)) { 1567845ec38aSEugenio Pérez error_setg_errno(errp, -r, "Cannot set device status"); 1568f1085882SEugenio Pérez goto out; 1569152128d6SEugenio Pérez } 1570152128d6SEugenio Pérez 1571845ec38aSEugenio Pérez r = ioctl(device_fd, VHOST_SET_FEATURES, &features); 1572845ec38aSEugenio Pérez if (unlikely(r)) { 1573845ec38aSEugenio Pérez error_setg_errno(errp, -r, "Cannot set features"); 1574845ec38aSEugenio Pérez goto out; 1575845ec38aSEugenio Pérez } 1576845ec38aSEugenio Pérez 1577845ec38aSEugenio Pérez status |= VIRTIO_CONFIG_S_FEATURES_OK; 1578152128d6SEugenio Pérez r = ioctl(device_fd, VHOST_VDPA_SET_STATUS, &status); 1579152128d6SEugenio Pérez if (unlikely(r)) { 1580845ec38aSEugenio Pérez error_setg_errno(errp, -r, "Cannot set device status"); 1581152128d6SEugenio Pérez goto out; 1582152128d6SEugenio Pérez } 1583152128d6SEugenio Pérez 1584152128d6SEugenio Pérez cvq_group = vhost_vdpa_get_vring_group(device_fd, cvq_index, errp); 1585152128d6SEugenio Pérez if (unlikely(cvq_group < 0)) { 1586152128d6SEugenio Pérez if (cvq_group != -ENOTSUP) { 1587152128d6SEugenio Pérez r = cvq_group; 1588152128d6SEugenio Pérez goto out; 1589152128d6SEugenio Pérez } 1590152128d6SEugenio Pérez 1591152128d6SEugenio Pérez /* 1592152128d6SEugenio Pérez * The kernel report VHOST_BACKEND_F_IOTLB_ASID if the vdpa frontend 1593152128d6SEugenio Pérez * support ASID even if the parent driver does not. The CVQ cannot be 1594152128d6SEugenio Pérez * isolated in this case. 1595152128d6SEugenio Pérez */ 1596152128d6SEugenio Pérez error_free(*errp); 1597152128d6SEugenio Pérez *errp = NULL; 1598152128d6SEugenio Pérez r = 0; 1599152128d6SEugenio Pérez goto out; 1600152128d6SEugenio Pérez } 1601152128d6SEugenio Pérez 1602152128d6SEugenio Pérez for (int i = 0; i < cvq_index; ++i) { 1603152128d6SEugenio Pérez int64_t group = vhost_vdpa_get_vring_group(device_fd, i, errp); 1604152128d6SEugenio Pérez if (unlikely(group < 0)) { 1605152128d6SEugenio Pérez r = group; 1606152128d6SEugenio Pérez goto out; 1607152128d6SEugenio Pérez } 1608152128d6SEugenio Pérez 1609152128d6SEugenio Pérez if (group == (int64_t)cvq_group) { 1610152128d6SEugenio Pérez r = 0; 1611152128d6SEugenio Pérez goto out; 1612152128d6SEugenio Pérez } 1613152128d6SEugenio Pérez } 1614152128d6SEugenio Pérez 1615152128d6SEugenio Pérez r = 1; 1616152128d6SEugenio Pérez 1617152128d6SEugenio Pérez out: 1618152128d6SEugenio Pérez status = 0; 1619152128d6SEugenio Pérez ioctl(device_fd, VHOST_VDPA_SET_STATUS, &status); 1620152128d6SEugenio Pérez return r; 1621152128d6SEugenio Pérez } 1622152128d6SEugenio Pérez 1623654790b6SJason Wang static NetClientState *net_vhost_vdpa_init(NetClientState *peer, 1624654790b6SJason Wang const char *device, 1625654790b6SJason Wang const char *name, 162640237840SJason Wang int vdpa_device_fd, 162740237840SJason Wang int queue_pair_index, 162840237840SJason Wang int nvqs, 16291576dbb5SEugenio Pérez bool is_datapath, 16301576dbb5SEugenio Pérez bool svq, 16315c1ebd4cSEugenio Pérez struct vhost_vdpa_iova_range iova_range, 1632152128d6SEugenio Pérez uint64_t features, 1633152128d6SEugenio Pérez Error **errp) 16341e0a84eaSCindy Lu { 16351e0a84eaSCindy Lu NetClientState *nc = NULL; 16361e0a84eaSCindy Lu VhostVDPAState *s; 16371e0a84eaSCindy Lu int ret = 0; 16381e0a84eaSCindy Lu assert(name); 1639e77db790SStefan Hajnoczi int cvq_isolated = 0; 1640152128d6SEugenio Pérez 164140237840SJason Wang if (is_datapath) { 164240237840SJason Wang nc = qemu_new_net_client(&net_vhost_vdpa_info, peer, device, 164340237840SJason Wang name); 164440237840SJason Wang } else { 1645152128d6SEugenio Pérez cvq_isolated = vhost_vdpa_probe_cvq_isolation(vdpa_device_fd, features, 1646152128d6SEugenio Pérez queue_pair_index * 2, 1647152128d6SEugenio Pérez errp); 1648152128d6SEugenio Pérez if (unlikely(cvq_isolated < 0)) { 1649152128d6SEugenio Pérez return NULL; 1650152128d6SEugenio Pérez } 1651152128d6SEugenio Pérez 1652f8972b56SEugenio Pérez nc = qemu_new_net_control_client(&net_vhost_vdpa_cvq_info, peer, 165340237840SJason Wang device, name); 165440237840SJason Wang } 165553b85d95SLaurent Vivier qemu_set_info_str(nc, TYPE_VHOST_VDPA); 16561e0a84eaSCindy Lu s = DO_UPCAST(VhostVDPAState, nc, nc); 16577327813dSJason Wang 16581e0a84eaSCindy Lu s->vhost_vdpa.device_fd = vdpa_device_fd; 165940237840SJason Wang s->vhost_vdpa.index = queue_pair_index; 16607f211a28SEugenio Pérez s->always_svq = svq; 1661d9cda213SSteve Sistare s->migration_state.notify = NULL; 16621576dbb5SEugenio Pérez s->vhost_vdpa.shadow_vqs_enabled = svq; 1663a585fad2SEugenio Pérez s->vhost_vdpa.iova_range = iova_range; 16646188d78aSEugenio Pérez s->vhost_vdpa.shadow_data = svq; 16655c1ebd4cSEugenio Pérez if (queue_pair_index == 0) { 16665c1ebd4cSEugenio Pérez vhost_vdpa_net_valid_svq_features(features, 16675c1ebd4cSEugenio Pérez &s->vhost_vdpa.migration_blocker); 16685c1ebd4cSEugenio Pérez } else if (!is_datapath) { 1669babf8b87SEugenio Pérez s->cvq_cmd_out_buffer = mmap(NULL, vhost_vdpa_net_cvq_cmd_page_len(), 1670babf8b87SEugenio Pérez PROT_READ | PROT_WRITE, 1671babf8b87SEugenio Pérez MAP_SHARED | MAP_ANONYMOUS, -1, 0); 1672babf8b87SEugenio Pérez s->status = mmap(NULL, vhost_vdpa_net_cvq_cmd_page_len(), 1673babf8b87SEugenio Pérez PROT_READ | PROT_WRITE, MAP_SHARED | MAP_ANONYMOUS, 1674babf8b87SEugenio Pérez -1, 0); 16752df4dd31SEugenio Pérez 1676bd907ae4SEugenio Pérez s->vhost_vdpa.shadow_vq_ops = &vhost_vdpa_net_svq_ops; 1677bd907ae4SEugenio Pérez s->vhost_vdpa.shadow_vq_ops_opaque = s; 1678152128d6SEugenio Pérez s->cvq_isolated = cvq_isolated; 16798bc0049eSEugenio Pérez } 168040237840SJason Wang ret = vhost_vdpa_add(nc, (void *)&s->vhost_vdpa, queue_pair_index, nvqs); 168174af5eecSJason Wang if (ret) { 168274af5eecSJason Wang qemu_del_net_client(nc); 1683654790b6SJason Wang return NULL; 168474af5eecSJason Wang } 1685654790b6SJason Wang return nc; 16861e0a84eaSCindy Lu } 16871e0a84eaSCindy Lu 16888170ab3fSEugenio Pérez static int vhost_vdpa_get_features(int fd, uint64_t *features, Error **errp) 16898170ab3fSEugenio Pérez { 16908170ab3fSEugenio Pérez int ret = ioctl(fd, VHOST_GET_FEATURES, features); 16918170ab3fSEugenio Pérez if (unlikely(ret < 0)) { 16928170ab3fSEugenio Pérez error_setg_errno(errp, errno, 16938170ab3fSEugenio Pérez "Fail to query features from vhost-vDPA device"); 16948170ab3fSEugenio Pérez } 16958170ab3fSEugenio Pérez return ret; 16968170ab3fSEugenio Pérez } 16978170ab3fSEugenio Pérez 16988170ab3fSEugenio Pérez static int vhost_vdpa_get_max_queue_pairs(int fd, uint64_t features, 16998170ab3fSEugenio Pérez int *has_cvq, Error **errp) 170040237840SJason Wang { 170140237840SJason Wang unsigned long config_size = offsetof(struct vhost_vdpa_config, buf); 1702cd523a41SStefano Garzarella g_autofree struct vhost_vdpa_config *config = NULL; 170340237840SJason Wang __virtio16 *max_queue_pairs; 170440237840SJason Wang int ret; 170540237840SJason Wang 170640237840SJason Wang if (features & (1 << VIRTIO_NET_F_CTRL_VQ)) { 170740237840SJason Wang *has_cvq = 1; 170840237840SJason Wang } else { 170940237840SJason Wang *has_cvq = 0; 171040237840SJason Wang } 171140237840SJason Wang 171240237840SJason Wang if (features & (1 << VIRTIO_NET_F_MQ)) { 171340237840SJason Wang config = g_malloc0(config_size + sizeof(*max_queue_pairs)); 171440237840SJason Wang config->off = offsetof(struct virtio_net_config, max_virtqueue_pairs); 171540237840SJason Wang config->len = sizeof(*max_queue_pairs); 171640237840SJason Wang 171740237840SJason Wang ret = ioctl(fd, VHOST_VDPA_GET_CONFIG, config); 171840237840SJason Wang if (ret) { 171940237840SJason Wang error_setg(errp, "Fail to get config from vhost-vDPA device"); 172040237840SJason Wang return -ret; 172140237840SJason Wang } 172240237840SJason Wang 172340237840SJason Wang max_queue_pairs = (__virtio16 *)&config->buf; 172440237840SJason Wang 172540237840SJason Wang return lduw_le_p(max_queue_pairs); 172640237840SJason Wang } 172740237840SJason Wang 172840237840SJason Wang return 1; 172940237840SJason Wang } 173040237840SJason Wang 17311e0a84eaSCindy Lu int net_init_vhost_vdpa(const Netdev *netdev, const char *name, 17321e0a84eaSCindy Lu NetClientState *peer, Error **errp) 17331e0a84eaSCindy Lu { 17341e0a84eaSCindy Lu const NetdevVhostVDPAOptions *opts; 17358170ab3fSEugenio Pérez uint64_t features; 1736654790b6SJason Wang int vdpa_device_fd; 1737eb3cb751SEugenio Pérez g_autofree NetClientState **ncs = NULL; 1738a585fad2SEugenio Pérez struct vhost_vdpa_iova_range iova_range; 1739eb3cb751SEugenio Pérez NetClientState *nc; 1740aed5da45SEugenio Pérez int queue_pairs, r, i = 0, has_cvq = 0; 17411e0a84eaSCindy Lu 17421e0a84eaSCindy Lu assert(netdev->type == NET_CLIENT_DRIVER_VHOST_VDPA); 17431e0a84eaSCindy Lu opts = &netdev->u.vhost_vdpa; 17447480874aSMarkus Armbruster if (!opts->vhostdev && !opts->vhostfd) { 17458801ccd0SSi-Wei Liu error_setg(errp, 17468801ccd0SSi-Wei Liu "vhost-vdpa: neither vhostdev= nor vhostfd= was specified"); 1747c8295404SEugenio Pérez return -1; 1748c8295404SEugenio Pérez } 17497327813dSJason Wang 17507480874aSMarkus Armbruster if (opts->vhostdev && opts->vhostfd) { 17518801ccd0SSi-Wei Liu error_setg(errp, 17528801ccd0SSi-Wei Liu "vhost-vdpa: vhostdev= and vhostfd= are mutually exclusive"); 17538801ccd0SSi-Wei Liu return -1; 17548801ccd0SSi-Wei Liu } 17558801ccd0SSi-Wei Liu 17567480874aSMarkus Armbruster if (opts->vhostdev) { 17570351152bSEugenio Pérez vdpa_device_fd = qemu_open(opts->vhostdev, O_RDWR, errp); 17587327813dSJason Wang if (vdpa_device_fd == -1) { 17597327813dSJason Wang return -errno; 17607327813dSJason Wang } 17615107fd3eSPeter Maydell } else { 17625107fd3eSPeter Maydell /* has_vhostfd */ 17638801ccd0SSi-Wei Liu vdpa_device_fd = monitor_fd_param(monitor_cur(), opts->vhostfd, errp); 17648801ccd0SSi-Wei Liu if (vdpa_device_fd == -1) { 17658801ccd0SSi-Wei Liu error_prepend(errp, "vhost-vdpa: unable to parse vhostfd: "); 17668801ccd0SSi-Wei Liu return -1; 17678801ccd0SSi-Wei Liu } 17688801ccd0SSi-Wei Liu } 17697327813dSJason Wang 17708170ab3fSEugenio Pérez r = vhost_vdpa_get_features(vdpa_device_fd, &features, errp); 17718170ab3fSEugenio Pérez if (unlikely(r < 0)) { 1772aed5da45SEugenio Pérez goto err; 17738170ab3fSEugenio Pérez } 17748170ab3fSEugenio Pérez 17758170ab3fSEugenio Pérez queue_pairs = vhost_vdpa_get_max_queue_pairs(vdpa_device_fd, features, 177640237840SJason Wang &has_cvq, errp); 177740237840SJason Wang if (queue_pairs < 0) { 17787327813dSJason Wang qemu_close(vdpa_device_fd); 177940237840SJason Wang return queue_pairs; 17807327813dSJason Wang } 17817327813dSJason Wang 1782bf7a2ad8SLongpeng r = vhost_vdpa_get_iova_range(vdpa_device_fd, &iova_range); 1783bf7a2ad8SLongpeng if (unlikely(r < 0)) { 1784bf7a2ad8SLongpeng error_setg(errp, "vhost-vdpa: get iova range failed: %s", 1785bf7a2ad8SLongpeng strerror(-r)); 1786bf7a2ad8SLongpeng goto err; 1787bf7a2ad8SLongpeng } 1788bf7a2ad8SLongpeng 178900ef422eSEugenio Pérez if (opts->x_svq && !vhost_vdpa_net_valid_svq_features(features, errp)) { 179000ef422eSEugenio Pérez goto err; 17911576dbb5SEugenio Pérez } 17921576dbb5SEugenio Pérez 179340237840SJason Wang ncs = g_malloc0(sizeof(*ncs) * queue_pairs); 179440237840SJason Wang 179540237840SJason Wang for (i = 0; i < queue_pairs; i++) { 179640237840SJason Wang ncs[i] = net_vhost_vdpa_init(peer, TYPE_VHOST_VDPA, name, 17971576dbb5SEugenio Pérez vdpa_device_fd, i, 2, true, opts->x_svq, 1798152128d6SEugenio Pérez iova_range, features, errp); 179940237840SJason Wang if (!ncs[i]) 180040237840SJason Wang goto err; 180140237840SJason Wang } 180240237840SJason Wang 180340237840SJason Wang if (has_cvq) { 180440237840SJason Wang nc = net_vhost_vdpa_init(peer, TYPE_VHOST_VDPA, name, 18051576dbb5SEugenio Pérez vdpa_device_fd, i, 1, false, 1806152128d6SEugenio Pérez opts->x_svq, iova_range, features, errp); 180740237840SJason Wang if (!nc) 180840237840SJason Wang goto err; 180940237840SJason Wang } 181040237840SJason Wang 1811654790b6SJason Wang return 0; 181240237840SJason Wang 181340237840SJason Wang err: 181440237840SJason Wang if (i) { 18159bd05507SSi-Wei Liu for (i--; i >= 0; i--) { 18169bd05507SSi-Wei Liu qemu_del_net_client(ncs[i]); 18179bd05507SSi-Wei Liu } 181840237840SJason Wang } 18191576dbb5SEugenio Pérez 182040237840SJason Wang qemu_close(vdpa_device_fd); 182140237840SJason Wang 182240237840SJason Wang return -1; 18231e0a84eaSCindy Lu } 1824