11e0a84eaSCindy Lu /* 21e0a84eaSCindy Lu * vhost-vdpa.c 31e0a84eaSCindy Lu * 41e0a84eaSCindy Lu * Copyright(c) 2017-2018 Intel Corporation. 51e0a84eaSCindy Lu * Copyright(c) 2020 Red Hat, Inc. 61e0a84eaSCindy Lu * 71e0a84eaSCindy Lu * This work is licensed under the terms of the GNU GPL, version 2 or later. 81e0a84eaSCindy Lu * See the COPYING file in the top-level directory. 91e0a84eaSCindy Lu * 101e0a84eaSCindy Lu */ 111e0a84eaSCindy Lu 121e0a84eaSCindy Lu #include "qemu/osdep.h" 131e0a84eaSCindy Lu #include "clients.h" 14bd907ae4SEugenio Pérez #include "hw/virtio/virtio-net.h" 151e0a84eaSCindy Lu #include "net/vhost_net.h" 161e0a84eaSCindy Lu #include "net/vhost-vdpa.h" 171e0a84eaSCindy Lu #include "hw/virtio/vhost-vdpa.h" 181e0a84eaSCindy Lu #include "qemu/config-file.h" 191e0a84eaSCindy Lu #include "qemu/error-report.h" 20bd907ae4SEugenio Pérez #include "qemu/log.h" 21bd907ae4SEugenio Pérez #include "qemu/memalign.h" 221e0a84eaSCindy Lu #include "qemu/option.h" 231e0a84eaSCindy Lu #include "qapi/error.h" 2440237840SJason Wang #include <linux/vhost.h> 251e0a84eaSCindy Lu #include <sys/ioctl.h> 261e0a84eaSCindy Lu #include <err.h> 271e0a84eaSCindy Lu #include "standard-headers/linux/virtio_net.h" 281e0a84eaSCindy Lu #include "monitor/monitor.h" 2969498430SEugenio Pérez #include "migration/migration.h" 3069498430SEugenio Pérez #include "migration/misc.h" 311e0a84eaSCindy Lu #include "hw/virtio/vhost.h" 321e0a84eaSCindy Lu 331e0a84eaSCindy Lu /* Todo:need to add the multiqueue support here */ 341e0a84eaSCindy Lu typedef struct VhostVDPAState { 351e0a84eaSCindy Lu NetClientState nc; 361e0a84eaSCindy Lu struct vhost_vdpa vhost_vdpa; 3769498430SEugenio Pérez Notifier migration_state; 381e0a84eaSCindy Lu VHostNetState *vhost_net; 392df4dd31SEugenio Pérez 402df4dd31SEugenio Pérez /* Control commands shadow buffers */ 4117fb889fSEugenio Pérez void *cvq_cmd_out_buffer; 4217fb889fSEugenio Pérez virtio_net_ctrl_ack *status; 4317fb889fSEugenio Pérez 447f211a28SEugenio Pérez /* The device always have SVQ enabled */ 457f211a28SEugenio Pérez bool always_svq; 46152128d6SEugenio Pérez 47152128d6SEugenio Pérez /* The device can isolate CVQ in its own ASID */ 48152128d6SEugenio Pérez bool cvq_isolated; 49152128d6SEugenio Pérez 501e0a84eaSCindy Lu bool started; 511e0a84eaSCindy Lu } VhostVDPAState; 521e0a84eaSCindy Lu 532875a0caSHawkins Jiawei /* 542875a0caSHawkins Jiawei * The array is sorted alphabetically in ascending order, 552875a0caSHawkins Jiawei * with the exception of VHOST_INVALID_FEATURE_BIT, 562875a0caSHawkins Jiawei * which should always be the last entry. 572875a0caSHawkins Jiawei */ 581e0a84eaSCindy Lu const int vdpa_feature_bits[] = { 591e0a84eaSCindy Lu VIRTIO_F_ANY_LAYOUT, 602875a0caSHawkins Jiawei VIRTIO_F_IOMMU_PLATFORM, 612875a0caSHawkins Jiawei VIRTIO_F_NOTIFY_ON_EMPTY, 622875a0caSHawkins Jiawei VIRTIO_F_RING_PACKED, 632875a0caSHawkins Jiawei VIRTIO_F_RING_RESET, 641e0a84eaSCindy Lu VIRTIO_F_VERSION_1, 651e0a84eaSCindy Lu VIRTIO_NET_F_CSUM, 6651e84244SEugenio Pérez VIRTIO_NET_F_CTRL_GUEST_OFFLOADS, 672875a0caSHawkins Jiawei VIRTIO_NET_F_CTRL_MAC_ADDR, 6840237840SJason Wang VIRTIO_NET_F_CTRL_RX, 6940237840SJason Wang VIRTIO_NET_F_CTRL_RX_EXTRA, 7040237840SJason Wang VIRTIO_NET_F_CTRL_VLAN, 7140237840SJason Wang VIRTIO_NET_F_CTRL_VQ, 722875a0caSHawkins Jiawei VIRTIO_NET_F_GSO, 732875a0caSHawkins Jiawei VIRTIO_NET_F_GUEST_CSUM, 742875a0caSHawkins Jiawei VIRTIO_NET_F_GUEST_ECN, 752875a0caSHawkins Jiawei VIRTIO_NET_F_GUEST_TSO4, 762875a0caSHawkins Jiawei VIRTIO_NET_F_GUEST_TSO6, 772875a0caSHawkins Jiawei VIRTIO_NET_F_GUEST_UFO, 789da16849SAndrew Melnychenko VIRTIO_NET_F_GUEST_USO4, 799da16849SAndrew Melnychenko VIRTIO_NET_F_GUEST_USO6, 800145c393SAndrew Melnychenko VIRTIO_NET_F_HASH_REPORT, 812875a0caSHawkins Jiawei VIRTIO_NET_F_HOST_ECN, 822875a0caSHawkins Jiawei VIRTIO_NET_F_HOST_TSO4, 832875a0caSHawkins Jiawei VIRTIO_NET_F_HOST_TSO6, 842875a0caSHawkins Jiawei VIRTIO_NET_F_HOST_UFO, 859da16849SAndrew Melnychenko VIRTIO_NET_F_HOST_USO, 862875a0caSHawkins Jiawei VIRTIO_NET_F_MQ, 872875a0caSHawkins Jiawei VIRTIO_NET_F_MRG_RXBUF, 882875a0caSHawkins Jiawei VIRTIO_NET_F_MTU, 892875a0caSHawkins Jiawei VIRTIO_NET_F_RSS, 909aa47eddSSi-Wei Liu VIRTIO_NET_F_STATUS, 912875a0caSHawkins Jiawei VIRTIO_RING_F_EVENT_IDX, 922875a0caSHawkins Jiawei VIRTIO_RING_F_INDIRECT_DESC, 932875a0caSHawkins Jiawei 942875a0caSHawkins Jiawei /* VHOST_INVALID_FEATURE_BIT should always be the last entry */ 951e0a84eaSCindy Lu VHOST_INVALID_FEATURE_BIT 961e0a84eaSCindy Lu }; 971e0a84eaSCindy Lu 981576dbb5SEugenio Pérez /** Supported device specific feature bits with SVQ */ 991576dbb5SEugenio Pérez static const uint64_t vdpa_svq_device_features = 1001576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_CSUM) | 1011576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_GUEST_CSUM) | 1024b4a1378SHawkins Jiawei BIT_ULL(VIRTIO_NET_F_CTRL_GUEST_OFFLOADS) | 1031576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_MTU) | 1041576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_MAC) | 1051576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_GUEST_TSO4) | 1061576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_GUEST_TSO6) | 1071576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_GUEST_ECN) | 1081576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_GUEST_UFO) | 1091576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_HOST_TSO4) | 1101576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_HOST_TSO6) | 1111576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_HOST_ECN) | 1121576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_HOST_UFO) | 1131576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_MRG_RXBUF) | 1141576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_STATUS) | 1151576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_CTRL_VQ) | 116ea6eec49SHawkins Jiawei BIT_ULL(VIRTIO_NET_F_CTRL_RX) | 117e213c45aSHawkins Jiawei BIT_ULL(VIRTIO_NET_F_CTRL_VLAN) | 118d669b7bbSHawkins Jiawei BIT_ULL(VIRTIO_NET_F_CTRL_RX_EXTRA) | 11972b99a87SEugenio Pérez BIT_ULL(VIRTIO_NET_F_MQ) | 1201576dbb5SEugenio Pérez BIT_ULL(VIRTIO_F_ANY_LAYOUT) | 1211576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_CTRL_MAC_ADDR) | 122609ab4c3SEugenio Pérez /* VHOST_F_LOG_ALL is exposed by SVQ */ 123609ab4c3SEugenio Pérez BIT_ULL(VHOST_F_LOG_ALL) | 1241576dbb5SEugenio Pérez BIT_ULL(VIRTIO_NET_F_RSC_EXT) | 1250d74e2b7SEugenio Pérez BIT_ULL(VIRTIO_NET_F_STANDBY) | 1260d74e2b7SEugenio Pérez BIT_ULL(VIRTIO_NET_F_SPEED_DUPLEX); 1271576dbb5SEugenio Pérez 128c1a10086SEugenio Pérez #define VHOST_VDPA_NET_CVQ_ASID 1 129c1a10086SEugenio Pérez 1301e0a84eaSCindy Lu VHostNetState *vhost_vdpa_get_vhost_net(NetClientState *nc) 1311e0a84eaSCindy Lu { 1321e0a84eaSCindy Lu VhostVDPAState *s = DO_UPCAST(VhostVDPAState, nc, nc); 1331e0a84eaSCindy Lu assert(nc->info->type == NET_CLIENT_DRIVER_VHOST_VDPA); 1341e0a84eaSCindy Lu return s->vhost_net; 1351e0a84eaSCindy Lu } 1361e0a84eaSCindy Lu 137915bf6ccSEugenio Pérez static size_t vhost_vdpa_net_cvq_cmd_len(void) 138915bf6ccSEugenio Pérez { 139915bf6ccSEugenio Pérez /* 140915bf6ccSEugenio Pérez * MAC_TABLE_SET is the ctrl command that produces the longer out buffer. 141915bf6ccSEugenio Pérez * In buffer is always 1 byte, so it should fit here 142915bf6ccSEugenio Pérez */ 143915bf6ccSEugenio Pérez return sizeof(struct virtio_net_ctrl_hdr) + 144915bf6ccSEugenio Pérez 2 * sizeof(struct virtio_net_ctrl_mac) + 145915bf6ccSEugenio Pérez MAC_TABLE_ENTRIES * ETH_ALEN; 146915bf6ccSEugenio Pérez } 147915bf6ccSEugenio Pérez 148915bf6ccSEugenio Pérez static size_t vhost_vdpa_net_cvq_cmd_page_len(void) 149915bf6ccSEugenio Pérez { 150915bf6ccSEugenio Pérez return ROUND_UP(vhost_vdpa_net_cvq_cmd_len(), qemu_real_host_page_size()); 151915bf6ccSEugenio Pérez } 152915bf6ccSEugenio Pérez 15336e46472SEugenio Pérez static bool vhost_vdpa_net_valid_svq_features(uint64_t features, Error **errp) 15436e46472SEugenio Pérez { 15536e46472SEugenio Pérez uint64_t invalid_dev_features = 15636e46472SEugenio Pérez features & ~vdpa_svq_device_features & 15736e46472SEugenio Pérez /* Transport are all accepted at this point */ 15836e46472SEugenio Pérez ~MAKE_64BIT_MASK(VIRTIO_TRANSPORT_F_START, 15936e46472SEugenio Pérez VIRTIO_TRANSPORT_F_END - VIRTIO_TRANSPORT_F_START); 16036e46472SEugenio Pérez 16136e46472SEugenio Pérez if (invalid_dev_features) { 16236e46472SEugenio Pérez error_setg(errp, "vdpa svq does not work with features 0x%" PRIx64, 16336e46472SEugenio Pérez invalid_dev_features); 164258a0394SEugenio Pérez return false; 16536e46472SEugenio Pérez } 16636e46472SEugenio Pérez 167258a0394SEugenio Pérez return vhost_svq_valid_features(features, errp); 16836e46472SEugenio Pérez } 16936e46472SEugenio Pérez 1701e0a84eaSCindy Lu static int vhost_vdpa_net_check_device_id(struct vhost_net *net) 1711e0a84eaSCindy Lu { 1721e0a84eaSCindy Lu uint32_t device_id; 1731e0a84eaSCindy Lu int ret; 1741e0a84eaSCindy Lu struct vhost_dev *hdev; 1751e0a84eaSCindy Lu 1761e0a84eaSCindy Lu hdev = (struct vhost_dev *)&net->dev; 1771e0a84eaSCindy Lu ret = hdev->vhost_ops->vhost_get_device_id(hdev, &device_id); 1781e0a84eaSCindy Lu if (device_id != VIRTIO_ID_NET) { 1791e0a84eaSCindy Lu return -ENOTSUP; 1801e0a84eaSCindy Lu } 1811e0a84eaSCindy Lu return ret; 1821e0a84eaSCindy Lu } 1831e0a84eaSCindy Lu 18440237840SJason Wang static int vhost_vdpa_add(NetClientState *ncs, void *be, 18540237840SJason Wang int queue_pair_index, int nvqs) 1861e0a84eaSCindy Lu { 1871e0a84eaSCindy Lu VhostNetOptions options; 1881e0a84eaSCindy Lu struct vhost_net *net = NULL; 1891e0a84eaSCindy Lu VhostVDPAState *s; 1901e0a84eaSCindy Lu int ret; 1911e0a84eaSCindy Lu 1921e0a84eaSCindy Lu options.backend_type = VHOST_BACKEND_TYPE_VDPA; 1931e0a84eaSCindy Lu assert(ncs->info->type == NET_CLIENT_DRIVER_VHOST_VDPA); 1941e0a84eaSCindy Lu s = DO_UPCAST(VhostVDPAState, nc, ncs); 1951e0a84eaSCindy Lu options.net_backend = ncs; 1961e0a84eaSCindy Lu options.opaque = be; 1971e0a84eaSCindy Lu options.busyloop_timeout = 0; 19840237840SJason Wang options.nvqs = nvqs; 1991e0a84eaSCindy Lu 2001e0a84eaSCindy Lu net = vhost_net_init(&options); 2011e0a84eaSCindy Lu if (!net) { 2021e0a84eaSCindy Lu error_report("failed to init vhost_net for queue"); 203a97ef87aSJason Wang goto err_init; 2041e0a84eaSCindy Lu } 2051e0a84eaSCindy Lu s->vhost_net = net; 2061e0a84eaSCindy Lu ret = vhost_vdpa_net_check_device_id(net); 2071e0a84eaSCindy Lu if (ret) { 208a97ef87aSJason Wang goto err_check; 2091e0a84eaSCindy Lu } 2101e0a84eaSCindy Lu return 0; 211a97ef87aSJason Wang err_check: 2121e0a84eaSCindy Lu vhost_net_cleanup(net); 213ab36edcfSJason Wang g_free(net); 214a97ef87aSJason Wang err_init: 2151e0a84eaSCindy Lu return -1; 2161e0a84eaSCindy Lu } 2171e0a84eaSCindy Lu 2181e0a84eaSCindy Lu static void vhost_vdpa_cleanup(NetClientState *nc) 2191e0a84eaSCindy Lu { 2201e0a84eaSCindy Lu VhostVDPAState *s = DO_UPCAST(VhostVDPAState, nc, nc); 2211e0a84eaSCindy Lu 222a0d7215eSAni Sinha /* 223a0d7215eSAni Sinha * If a peer NIC is attached, do not cleanup anything. 224a0d7215eSAni Sinha * Cleanup will happen as a part of qemu_cleanup() -> net_cleanup() 225a0d7215eSAni Sinha * when the guest is shutting down. 226a0d7215eSAni Sinha */ 227a0d7215eSAni Sinha if (nc->peer && nc->peer->info->type == NET_CLIENT_DRIVER_NIC) { 228a0d7215eSAni Sinha return; 229a0d7215eSAni Sinha } 230babf8b87SEugenio Pérez munmap(s->cvq_cmd_out_buffer, vhost_vdpa_net_cvq_cmd_page_len()); 231babf8b87SEugenio Pérez munmap(s->status, vhost_vdpa_net_cvq_cmd_page_len()); 2321e0a84eaSCindy Lu if (s->vhost_net) { 2331e0a84eaSCindy Lu vhost_net_cleanup(s->vhost_net); 2341e0a84eaSCindy Lu g_free(s->vhost_net); 2351e0a84eaSCindy Lu s->vhost_net = NULL; 2361e0a84eaSCindy Lu } 23757b3a7d8SCindy Lu if (s->vhost_vdpa.device_fd >= 0) { 23857b3a7d8SCindy Lu qemu_close(s->vhost_vdpa.device_fd); 23957b3a7d8SCindy Lu s->vhost_vdpa.device_fd = -1; 24057b3a7d8SCindy Lu } 2411e0a84eaSCindy Lu } 2421e0a84eaSCindy Lu 2431e0a84eaSCindy Lu static bool vhost_vdpa_has_vnet_hdr(NetClientState *nc) 2441e0a84eaSCindy Lu { 2451e0a84eaSCindy Lu assert(nc->info->type == NET_CLIENT_DRIVER_VHOST_VDPA); 2461e0a84eaSCindy Lu 2471e0a84eaSCindy Lu return true; 2481e0a84eaSCindy Lu } 2491e0a84eaSCindy Lu 2501e0a84eaSCindy Lu static bool vhost_vdpa_has_ufo(NetClientState *nc) 2511e0a84eaSCindy Lu { 2521e0a84eaSCindy Lu assert(nc->info->type == NET_CLIENT_DRIVER_VHOST_VDPA); 2531e0a84eaSCindy Lu VhostVDPAState *s = DO_UPCAST(VhostVDPAState, nc, nc); 2541e0a84eaSCindy Lu uint64_t features = 0; 2551e0a84eaSCindy Lu features |= (1ULL << VIRTIO_NET_F_HOST_UFO); 2561e0a84eaSCindy Lu features = vhost_net_get_features(s->vhost_net, features); 2571e0a84eaSCindy Lu return !!(features & (1ULL << VIRTIO_NET_F_HOST_UFO)); 2581e0a84eaSCindy Lu 2591e0a84eaSCindy Lu } 2601e0a84eaSCindy Lu 261ee8a1c63SKevin Wolf static bool vhost_vdpa_check_peer_type(NetClientState *nc, ObjectClass *oc, 262ee8a1c63SKevin Wolf Error **errp) 263ee8a1c63SKevin Wolf { 264ee8a1c63SKevin Wolf const char *driver = object_class_get_name(oc); 265ee8a1c63SKevin Wolf 266ee8a1c63SKevin Wolf if (!g_str_has_prefix(driver, "virtio-net-")) { 267ee8a1c63SKevin Wolf error_setg(errp, "vhost-vdpa requires frontend driver virtio-net-*"); 268ee8a1c63SKevin Wolf return false; 269ee8a1c63SKevin Wolf } 270ee8a1c63SKevin Wolf 271ee8a1c63SKevin Wolf return true; 272ee8a1c63SKevin Wolf } 273ee8a1c63SKevin Wolf 274846a1e85SEugenio Pérez /** Dummy receive in case qemu falls back to userland tap networking */ 275846a1e85SEugenio Pérez static ssize_t vhost_vdpa_receive(NetClientState *nc, const uint8_t *buf, 276846a1e85SEugenio Pérez size_t size) 277846a1e85SEugenio Pérez { 278bc5add1dSSi-Wei Liu return size; 279846a1e85SEugenio Pérez } 280846a1e85SEugenio Pérez 28100ef422eSEugenio Pérez /** From any vdpa net client, get the netclient of the first queue pair */ 28200ef422eSEugenio Pérez static VhostVDPAState *vhost_vdpa_net_first_nc_vdpa(VhostVDPAState *s) 28300ef422eSEugenio Pérez { 28400ef422eSEugenio Pérez NICState *nic = qemu_get_nic(s->nc.peer); 28500ef422eSEugenio Pérez NetClientState *nc0 = qemu_get_peer(nic->ncs, 0); 28600ef422eSEugenio Pérez 28700ef422eSEugenio Pérez return DO_UPCAST(VhostVDPAState, nc, nc0); 28800ef422eSEugenio Pérez } 28900ef422eSEugenio Pérez 29069498430SEugenio Pérez static void vhost_vdpa_net_log_global_enable(VhostVDPAState *s, bool enable) 29169498430SEugenio Pérez { 29269498430SEugenio Pérez struct vhost_vdpa *v = &s->vhost_vdpa; 29369498430SEugenio Pérez VirtIONet *n; 29469498430SEugenio Pérez VirtIODevice *vdev; 29569498430SEugenio Pérez int data_queue_pairs, cvq, r; 29669498430SEugenio Pérez 29769498430SEugenio Pérez /* We are only called on the first data vqs and only if x-svq is not set */ 29869498430SEugenio Pérez if (s->vhost_vdpa.shadow_vqs_enabled == enable) { 29969498430SEugenio Pérez return; 30069498430SEugenio Pérez } 30169498430SEugenio Pérez 30269498430SEugenio Pérez vdev = v->dev->vdev; 30369498430SEugenio Pérez n = VIRTIO_NET(vdev); 30469498430SEugenio Pérez if (!n->vhost_started) { 30569498430SEugenio Pérez return; 30669498430SEugenio Pérez } 30769498430SEugenio Pérez 30869498430SEugenio Pérez data_queue_pairs = n->multiqueue ? n->max_queue_pairs : 1; 30969498430SEugenio Pérez cvq = virtio_vdev_has_feature(vdev, VIRTIO_NET_F_CTRL_VQ) ? 31069498430SEugenio Pérez n->max_ncs - n->max_queue_pairs : 0; 31169498430SEugenio Pérez /* 31269498430SEugenio Pérez * TODO: vhost_net_stop does suspend, get_base and reset. We can be smarter 31369498430SEugenio Pérez * in the future and resume the device if read-only operations between 31469498430SEugenio Pérez * suspend and reset goes wrong. 31569498430SEugenio Pérez */ 31669498430SEugenio Pérez vhost_net_stop(vdev, n->nic->ncs, data_queue_pairs, cvq); 31769498430SEugenio Pérez 31869498430SEugenio Pérez /* Start will check migration setup_or_active to configure or not SVQ */ 31969498430SEugenio Pérez r = vhost_net_start(vdev, n->nic->ncs, data_queue_pairs, cvq); 32069498430SEugenio Pérez if (unlikely(r < 0)) { 32169498430SEugenio Pérez error_report("unable to start vhost net: %s(%d)", g_strerror(-r), -r); 32269498430SEugenio Pérez } 32369498430SEugenio Pérez } 32469498430SEugenio Pérez 32569498430SEugenio Pérez static void vdpa_net_migration_state_notifier(Notifier *notifier, void *data) 32669498430SEugenio Pérez { 32769498430SEugenio Pérez MigrationState *migration = data; 32869498430SEugenio Pérez VhostVDPAState *s = container_of(notifier, VhostVDPAState, 32969498430SEugenio Pérez migration_state); 33069498430SEugenio Pérez 33169498430SEugenio Pérez if (migration_in_setup(migration)) { 33269498430SEugenio Pérez vhost_vdpa_net_log_global_enable(s, true); 33369498430SEugenio Pérez } else if (migration_has_failed(migration)) { 33469498430SEugenio Pérez vhost_vdpa_net_log_global_enable(s, false); 33569498430SEugenio Pérez } 33669498430SEugenio Pérez } 33769498430SEugenio Pérez 33800ef422eSEugenio Pérez static void vhost_vdpa_net_data_start_first(VhostVDPAState *s) 33900ef422eSEugenio Pérez { 34000ef422eSEugenio Pérez struct vhost_vdpa *v = &s->vhost_vdpa; 34100ef422eSEugenio Pérez 34269498430SEugenio Pérez add_migration_state_change_notifier(&s->migration_state); 34300ef422eSEugenio Pérez if (v->shadow_vqs_enabled) { 34400ef422eSEugenio Pérez v->iova_tree = vhost_iova_tree_new(v->iova_range.first, 34500ef422eSEugenio Pérez v->iova_range.last); 34600ef422eSEugenio Pérez } 34700ef422eSEugenio Pérez } 34800ef422eSEugenio Pérez 34900ef422eSEugenio Pérez static int vhost_vdpa_net_data_start(NetClientState *nc) 35000ef422eSEugenio Pérez { 35100ef422eSEugenio Pérez VhostVDPAState *s = DO_UPCAST(VhostVDPAState, nc, nc); 35200ef422eSEugenio Pérez struct vhost_vdpa *v = &s->vhost_vdpa; 35300ef422eSEugenio Pérez 35400ef422eSEugenio Pérez assert(nc->info->type == NET_CLIENT_DRIVER_VHOST_VDPA); 35500ef422eSEugenio Pérez 35669498430SEugenio Pérez if (s->always_svq || 35769498430SEugenio Pérez migration_is_setup_or_active(migrate_get_current()->state)) { 35869498430SEugenio Pérez v->shadow_vqs_enabled = true; 35969498430SEugenio Pérez v->shadow_data = true; 36069498430SEugenio Pérez } else { 36169498430SEugenio Pérez v->shadow_vqs_enabled = false; 36269498430SEugenio Pérez v->shadow_data = false; 36369498430SEugenio Pérez } 36469498430SEugenio Pérez 36500ef422eSEugenio Pérez if (v->index == 0) { 36600ef422eSEugenio Pérez vhost_vdpa_net_data_start_first(s); 36700ef422eSEugenio Pérez return 0; 36800ef422eSEugenio Pérez } 36900ef422eSEugenio Pérez 37000ef422eSEugenio Pérez if (v->shadow_vqs_enabled) { 37100ef422eSEugenio Pérez VhostVDPAState *s0 = vhost_vdpa_net_first_nc_vdpa(s); 37200ef422eSEugenio Pérez v->iova_tree = s0->vhost_vdpa.iova_tree; 37300ef422eSEugenio Pérez } 37400ef422eSEugenio Pérez 37500ef422eSEugenio Pérez return 0; 37600ef422eSEugenio Pérez } 37700ef422eSEugenio Pérez 3786c482547SEugenio Pérez static int vhost_vdpa_net_data_load(NetClientState *nc) 3796c482547SEugenio Pérez { 3806c482547SEugenio Pérez VhostVDPAState *s = DO_UPCAST(VhostVDPAState, nc, nc); 3816c482547SEugenio Pérez struct vhost_vdpa *v = &s->vhost_vdpa; 3826c482547SEugenio Pérez bool has_cvq = v->dev->vq_index_end % 2; 3836c482547SEugenio Pérez 3846c482547SEugenio Pérez if (has_cvq) { 3856c482547SEugenio Pérez return 0; 3866c482547SEugenio Pérez } 3876c482547SEugenio Pérez 3886c482547SEugenio Pérez for (int i = 0; i < v->dev->nvqs; ++i) { 3896c482547SEugenio Pérez vhost_vdpa_set_vring_ready(v, i + v->dev->vq_index); 3906c482547SEugenio Pérez } 3916c482547SEugenio Pérez return 0; 3926c482547SEugenio Pérez } 3936c482547SEugenio Pérez 39400ef422eSEugenio Pérez static void vhost_vdpa_net_client_stop(NetClientState *nc) 39500ef422eSEugenio Pérez { 39600ef422eSEugenio Pérez VhostVDPAState *s = DO_UPCAST(VhostVDPAState, nc, nc); 39700ef422eSEugenio Pérez struct vhost_dev *dev; 39800ef422eSEugenio Pérez 39900ef422eSEugenio Pérez assert(nc->info->type == NET_CLIENT_DRIVER_VHOST_VDPA); 40000ef422eSEugenio Pérez 40169498430SEugenio Pérez if (s->vhost_vdpa.index == 0) { 40269498430SEugenio Pérez remove_migration_state_change_notifier(&s->migration_state); 40369498430SEugenio Pérez } 40469498430SEugenio Pérez 40500ef422eSEugenio Pérez dev = s->vhost_vdpa.dev; 40600ef422eSEugenio Pérez if (dev->vq_index + dev->nvqs == dev->vq_index_end) { 40700ef422eSEugenio Pérez g_clear_pointer(&s->vhost_vdpa.iova_tree, vhost_iova_tree_delete); 40800ef422eSEugenio Pérez } 40900ef422eSEugenio Pérez } 41000ef422eSEugenio Pérez 4111e0a84eaSCindy Lu static NetClientInfo net_vhost_vdpa_info = { 4121e0a84eaSCindy Lu .type = NET_CLIENT_DRIVER_VHOST_VDPA, 4131e0a84eaSCindy Lu .size = sizeof(VhostVDPAState), 414846a1e85SEugenio Pérez .receive = vhost_vdpa_receive, 41500ef422eSEugenio Pérez .start = vhost_vdpa_net_data_start, 4166c482547SEugenio Pérez .load = vhost_vdpa_net_data_load, 41700ef422eSEugenio Pérez .stop = vhost_vdpa_net_client_stop, 4181e0a84eaSCindy Lu .cleanup = vhost_vdpa_cleanup, 4191e0a84eaSCindy Lu .has_vnet_hdr = vhost_vdpa_has_vnet_hdr, 4201e0a84eaSCindy Lu .has_ufo = vhost_vdpa_has_ufo, 421ee8a1c63SKevin Wolf .check_peer_type = vhost_vdpa_check_peer_type, 4221e0a84eaSCindy Lu }; 4231e0a84eaSCindy Lu 424152128d6SEugenio Pérez static int64_t vhost_vdpa_get_vring_group(int device_fd, unsigned vq_index, 425152128d6SEugenio Pérez Error **errp) 426c1a10086SEugenio Pérez { 427c1a10086SEugenio Pérez struct vhost_vring_state state = { 428c1a10086SEugenio Pérez .index = vq_index, 429c1a10086SEugenio Pérez }; 430c1a10086SEugenio Pérez int r = ioctl(device_fd, VHOST_VDPA_GET_VRING_GROUP, &state); 431c1a10086SEugenio Pérez 432c1a10086SEugenio Pérez if (unlikely(r < 0)) { 4330f2bb0bfSEugenio Pérez r = -errno; 434152128d6SEugenio Pérez error_setg_errno(errp, errno, "Cannot get VQ %u group", vq_index); 435c1a10086SEugenio Pérez return r; 436c1a10086SEugenio Pérez } 437c1a10086SEugenio Pérez 438c1a10086SEugenio Pérez return state.num; 439c1a10086SEugenio Pérez } 440c1a10086SEugenio Pérez 441c1a10086SEugenio Pérez static int vhost_vdpa_set_address_space_id(struct vhost_vdpa *v, 442c1a10086SEugenio Pérez unsigned vq_group, 443c1a10086SEugenio Pérez unsigned asid_num) 444c1a10086SEugenio Pérez { 445c1a10086SEugenio Pérez struct vhost_vring_state asid = { 446c1a10086SEugenio Pérez .index = vq_group, 447c1a10086SEugenio Pérez .num = asid_num, 448c1a10086SEugenio Pérez }; 449c1a10086SEugenio Pérez int r; 450c1a10086SEugenio Pérez 451c1a10086SEugenio Pérez r = ioctl(v->device_fd, VHOST_VDPA_SET_GROUP_ASID, &asid); 452c1a10086SEugenio Pérez if (unlikely(r < 0)) { 453c1a10086SEugenio Pérez error_report("Can't set vq group %u asid %u, errno=%d (%s)", 454c1a10086SEugenio Pérez asid.index, asid.num, errno, g_strerror(errno)); 455c1a10086SEugenio Pérez } 456c1a10086SEugenio Pérez return r; 457c1a10086SEugenio Pérez } 458c1a10086SEugenio Pérez 4592df4dd31SEugenio Pérez static void vhost_vdpa_cvq_unmap_buf(struct vhost_vdpa *v, void *addr) 4602df4dd31SEugenio Pérez { 4612df4dd31SEugenio Pérez VhostIOVATree *tree = v->iova_tree; 4622df4dd31SEugenio Pérez DMAMap needle = { 4632df4dd31SEugenio Pérez /* 4642df4dd31SEugenio Pérez * No need to specify size or to look for more translations since 4652df4dd31SEugenio Pérez * this contiguous chunk was allocated by us. 4662df4dd31SEugenio Pérez */ 4672df4dd31SEugenio Pérez .translated_addr = (hwaddr)(uintptr_t)addr, 4682df4dd31SEugenio Pérez }; 4692df4dd31SEugenio Pérez const DMAMap *map = vhost_iova_tree_find_iova(tree, &needle); 4702df4dd31SEugenio Pérez int r; 4712df4dd31SEugenio Pérez 4722df4dd31SEugenio Pérez if (unlikely(!map)) { 4732df4dd31SEugenio Pérez error_report("Cannot locate expected map"); 4742df4dd31SEugenio Pérez return; 4752df4dd31SEugenio Pérez } 4762df4dd31SEugenio Pérez 477cd831ed5SEugenio Pérez r = vhost_vdpa_dma_unmap(v, v->address_space_id, map->iova, map->size + 1); 4782df4dd31SEugenio Pérez if (unlikely(r != 0)) { 4792df4dd31SEugenio Pérez error_report("Device cannot unmap: %s(%d)", g_strerror(r), r); 4802df4dd31SEugenio Pérez } 4812df4dd31SEugenio Pérez 48269292a8eSEugenio Pérez vhost_iova_tree_remove(tree, *map); 4832df4dd31SEugenio Pérez } 4842df4dd31SEugenio Pérez 4857a7f87e9SEugenio Pérez /** Map CVQ buffer. */ 4867a7f87e9SEugenio Pérez static int vhost_vdpa_cvq_map_buf(struct vhost_vdpa *v, void *buf, size_t size, 4877a7f87e9SEugenio Pérez bool write) 4882df4dd31SEugenio Pérez { 4892df4dd31SEugenio Pérez DMAMap map = {}; 4902df4dd31SEugenio Pérez int r; 4912df4dd31SEugenio Pérez 4922df4dd31SEugenio Pérez map.translated_addr = (hwaddr)(uintptr_t)buf; 4937a7f87e9SEugenio Pérez map.size = size - 1; 4942df4dd31SEugenio Pérez map.perm = write ? IOMMU_RW : IOMMU_RO, 4952df4dd31SEugenio Pérez r = vhost_iova_tree_map_alloc(v->iova_tree, &map); 4962df4dd31SEugenio Pérez if (unlikely(r != IOVA_OK)) { 4972df4dd31SEugenio Pérez error_report("Cannot map injected element"); 4987a7f87e9SEugenio Pérez return r; 4992df4dd31SEugenio Pérez } 5002df4dd31SEugenio Pérez 501cd831ed5SEugenio Pérez r = vhost_vdpa_dma_map(v, v->address_space_id, map.iova, 502cd831ed5SEugenio Pérez vhost_vdpa_net_cvq_cmd_page_len(), buf, !write); 5032df4dd31SEugenio Pérez if (unlikely(r < 0)) { 5042df4dd31SEugenio Pérez goto dma_map_err; 5052df4dd31SEugenio Pérez } 5062df4dd31SEugenio Pérez 5077a7f87e9SEugenio Pérez return 0; 5082df4dd31SEugenio Pérez 5092df4dd31SEugenio Pérez dma_map_err: 51069292a8eSEugenio Pérez vhost_iova_tree_remove(v->iova_tree, map); 5117a7f87e9SEugenio Pérez return r; 5122df4dd31SEugenio Pérez } 5132df4dd31SEugenio Pérez 5147a7f87e9SEugenio Pérez static int vhost_vdpa_net_cvq_start(NetClientState *nc) 5152df4dd31SEugenio Pérez { 51600ef422eSEugenio Pérez VhostVDPAState *s, *s0; 517c1a10086SEugenio Pérez struct vhost_vdpa *v; 518c1a10086SEugenio Pérez int64_t cvq_group; 519152128d6SEugenio Pérez int r; 520152128d6SEugenio Pérez Error *err = NULL; 5212df4dd31SEugenio Pérez 5227a7f87e9SEugenio Pérez assert(nc->info->type == NET_CLIENT_DRIVER_VHOST_VDPA); 5237a7f87e9SEugenio Pérez 5247a7f87e9SEugenio Pérez s = DO_UPCAST(VhostVDPAState, nc, nc); 525c1a10086SEugenio Pérez v = &s->vhost_vdpa; 526c1a10086SEugenio Pérez 52769498430SEugenio Pérez s0 = vhost_vdpa_net_first_nc_vdpa(s); 52869498430SEugenio Pérez v->shadow_data = s0->vhost_vdpa.shadow_vqs_enabled; 529b40eba9cSEugenio Pérez v->shadow_vqs_enabled = s0->vhost_vdpa.shadow_vqs_enabled; 530c1a10086SEugenio Pérez s->vhost_vdpa.address_space_id = VHOST_VDPA_GUEST_PA_ASID; 531c1a10086SEugenio Pérez 53269498430SEugenio Pérez if (s->vhost_vdpa.shadow_data) { 533c1a10086SEugenio Pérez /* SVQ is already configured for all virtqueues */ 534c1a10086SEugenio Pérez goto out; 535c1a10086SEugenio Pérez } 536c1a10086SEugenio Pérez 537c1a10086SEugenio Pérez /* 538c1a10086SEugenio Pérez * If we early return in these cases SVQ will not be enabled. The migration 539c1a10086SEugenio Pérez * will be blocked as long as vhost-vdpa backends will not offer _F_LOG. 540c1a10086SEugenio Pérez */ 541152128d6SEugenio Pérez if (!vhost_vdpa_net_valid_svq_features(v->dev->features, NULL)) { 542c1a10086SEugenio Pérez return 0; 543c1a10086SEugenio Pérez } 544c1a10086SEugenio Pérez 545152128d6SEugenio Pérez if (!s->cvq_isolated) { 546152128d6SEugenio Pérez return 0; 547152128d6SEugenio Pérez } 548152128d6SEugenio Pérez 549152128d6SEugenio Pérez cvq_group = vhost_vdpa_get_vring_group(v->device_fd, 550152128d6SEugenio Pérez v->dev->vq_index_end - 1, 551152128d6SEugenio Pérez &err); 552c1a10086SEugenio Pérez if (unlikely(cvq_group < 0)) { 553152128d6SEugenio Pérez error_report_err(err); 554c1a10086SEugenio Pérez return cvq_group; 555c1a10086SEugenio Pérez } 556c1a10086SEugenio Pérez 557c1a10086SEugenio Pérez r = vhost_vdpa_set_address_space_id(v, cvq_group, VHOST_VDPA_NET_CVQ_ASID); 558c1a10086SEugenio Pérez if (unlikely(r < 0)) { 559c1a10086SEugenio Pérez return r; 560c1a10086SEugenio Pérez } 561c1a10086SEugenio Pérez 562c1a10086SEugenio Pérez v->shadow_vqs_enabled = true; 563c1a10086SEugenio Pérez s->vhost_vdpa.address_space_id = VHOST_VDPA_NET_CVQ_ASID; 564c1a10086SEugenio Pérez 565c1a10086SEugenio Pérez out: 5667a7f87e9SEugenio Pérez if (!s->vhost_vdpa.shadow_vqs_enabled) { 5677a7f87e9SEugenio Pérez return 0; 5682df4dd31SEugenio Pérez } 5692df4dd31SEugenio Pérez 57000ef422eSEugenio Pérez if (s0->vhost_vdpa.iova_tree) { 57100ef422eSEugenio Pérez /* 57200ef422eSEugenio Pérez * SVQ is already configured for all virtqueues. Reuse IOVA tree for 57300ef422eSEugenio Pérez * simplicity, whether CVQ shares ASID with guest or not, because: 57400ef422eSEugenio Pérez * - Memory listener need access to guest's memory addresses allocated 57500ef422eSEugenio Pérez * in the IOVA tree. 57600ef422eSEugenio Pérez * - There should be plenty of IOVA address space for both ASID not to 57700ef422eSEugenio Pérez * worry about collisions between them. Guest's translations are 57800ef422eSEugenio Pérez * still validated with virtio virtqueue_pop so there is no risk for 57900ef422eSEugenio Pérez * the guest to access memory that it shouldn't. 58000ef422eSEugenio Pérez * 58100ef422eSEugenio Pérez * To allocate a iova tree per ASID is doable but it complicates the 58200ef422eSEugenio Pérez * code and it is not worth it for the moment. 58300ef422eSEugenio Pérez */ 58400ef422eSEugenio Pérez v->iova_tree = s0->vhost_vdpa.iova_tree; 58500ef422eSEugenio Pérez } else { 58600ef422eSEugenio Pérez v->iova_tree = vhost_iova_tree_new(v->iova_range.first, 58700ef422eSEugenio Pérez v->iova_range.last); 58800ef422eSEugenio Pérez } 58900ef422eSEugenio Pérez 5907a7f87e9SEugenio Pérez r = vhost_vdpa_cvq_map_buf(&s->vhost_vdpa, s->cvq_cmd_out_buffer, 5917a7f87e9SEugenio Pérez vhost_vdpa_net_cvq_cmd_page_len(), false); 5927a7f87e9SEugenio Pérez if (unlikely(r < 0)) { 5937a7f87e9SEugenio Pérez return r; 5947a7f87e9SEugenio Pérez } 5957a7f87e9SEugenio Pérez 59617fb889fSEugenio Pérez r = vhost_vdpa_cvq_map_buf(&s->vhost_vdpa, s->status, 5977a7f87e9SEugenio Pérez vhost_vdpa_net_cvq_cmd_page_len(), true); 5987a7f87e9SEugenio Pérez if (unlikely(r < 0)) { 5992df4dd31SEugenio Pérez vhost_vdpa_cvq_unmap_buf(&s->vhost_vdpa, s->cvq_cmd_out_buffer); 6002df4dd31SEugenio Pérez } 6012df4dd31SEugenio Pérez 6027a7f87e9SEugenio Pérez return r; 6037a7f87e9SEugenio Pérez } 6047a7f87e9SEugenio Pérez 6057a7f87e9SEugenio Pérez static void vhost_vdpa_net_cvq_stop(NetClientState *nc) 6067a7f87e9SEugenio Pérez { 6077a7f87e9SEugenio Pérez VhostVDPAState *s = DO_UPCAST(VhostVDPAState, nc, nc); 6087a7f87e9SEugenio Pérez 6097a7f87e9SEugenio Pérez assert(nc->info->type == NET_CLIENT_DRIVER_VHOST_VDPA); 6107a7f87e9SEugenio Pérez 6117a7f87e9SEugenio Pérez if (s->vhost_vdpa.shadow_vqs_enabled) { 6127a7f87e9SEugenio Pérez vhost_vdpa_cvq_unmap_buf(&s->vhost_vdpa, s->cvq_cmd_out_buffer); 61317fb889fSEugenio Pérez vhost_vdpa_cvq_unmap_buf(&s->vhost_vdpa, s->status); 614c1a10086SEugenio Pérez } 61500ef422eSEugenio Pérez 61600ef422eSEugenio Pérez vhost_vdpa_net_client_stop(nc); 6172df4dd31SEugenio Pérez } 6182df4dd31SEugenio Pérez 619be4278b6SEugenio Pérez static ssize_t vhost_vdpa_net_cvq_add(VhostVDPAState *s, size_t out_len, 620be4278b6SEugenio Pérez size_t in_len) 621be4278b6SEugenio Pérez { 622be4278b6SEugenio Pérez /* Buffers for the device */ 623be4278b6SEugenio Pérez const struct iovec out = { 624be4278b6SEugenio Pérez .iov_base = s->cvq_cmd_out_buffer, 625be4278b6SEugenio Pérez .iov_len = out_len, 626be4278b6SEugenio Pérez }; 627be4278b6SEugenio Pérez const struct iovec in = { 62817fb889fSEugenio Pérez .iov_base = s->status, 629be4278b6SEugenio Pérez .iov_len = sizeof(virtio_net_ctrl_ack), 630be4278b6SEugenio Pérez }; 631be4278b6SEugenio Pérez VhostShadowVirtqueue *svq = g_ptr_array_index(s->vhost_vdpa.shadow_vqs, 0); 632be4278b6SEugenio Pérez int r; 633be4278b6SEugenio Pérez 634be4278b6SEugenio Pérez r = vhost_svq_add(svq, &out, 1, &in, 1, NULL); 635be4278b6SEugenio Pérez if (unlikely(r != 0)) { 636be4278b6SEugenio Pérez if (unlikely(r == -ENOSPC)) { 637be4278b6SEugenio Pérez qemu_log_mask(LOG_GUEST_ERROR, "%s: No space on device queue\n", 638be4278b6SEugenio Pérez __func__); 639be4278b6SEugenio Pérez } 640be4278b6SEugenio Pérez return r; 641be4278b6SEugenio Pérez } 642be4278b6SEugenio Pérez 643be4278b6SEugenio Pérez /* 644be4278b6SEugenio Pérez * We can poll here since we've had BQL from the time we sent the 645be4278b6SEugenio Pérez * descriptor. Also, we need to take the answer before SVQ pulls by itself, 646be4278b6SEugenio Pérez * when BQL is released 647be4278b6SEugenio Pérez */ 648*b0de17a2SHawkins Jiawei return vhost_svq_poll(svq, 1); 649be4278b6SEugenio Pérez } 650be4278b6SEugenio Pérez 651f73c0c43SEugenio Pérez static ssize_t vhost_vdpa_net_load_cmd(VhostVDPAState *s, uint8_t class, 6522848c6aaSHawkins Jiawei uint8_t cmd, const struct iovec *data_sg, 6532848c6aaSHawkins Jiawei size_t data_num) 654f73c0c43SEugenio Pérez { 655f73c0c43SEugenio Pérez const struct virtio_net_ctrl_hdr ctrl = { 656f73c0c43SEugenio Pérez .class = class, 657f73c0c43SEugenio Pérez .cmd = cmd, 658f73c0c43SEugenio Pérez }; 6592848c6aaSHawkins Jiawei size_t data_size = iov_size(data_sg, data_num); 660f73c0c43SEugenio Pérez 661f73c0c43SEugenio Pérez assert(data_size < vhost_vdpa_net_cvq_cmd_page_len() - sizeof(ctrl)); 662f73c0c43SEugenio Pérez 6632848c6aaSHawkins Jiawei /* pack the CVQ command header */ 664f73c0c43SEugenio Pérez memcpy(s->cvq_cmd_out_buffer, &ctrl, sizeof(ctrl)); 665f73c0c43SEugenio Pérez 6662848c6aaSHawkins Jiawei /* pack the CVQ command command-specific-data */ 6672848c6aaSHawkins Jiawei iov_to_buf(data_sg, data_num, 0, 6682848c6aaSHawkins Jiawei s->cvq_cmd_out_buffer + sizeof(ctrl), data_size); 6692848c6aaSHawkins Jiawei 6702848c6aaSHawkins Jiawei return vhost_vdpa_net_cvq_add(s, data_size + sizeof(ctrl), 671f73c0c43SEugenio Pérez sizeof(virtio_net_ctrl_ack)); 672f73c0c43SEugenio Pérez } 673f73c0c43SEugenio Pérez 674f73c0c43SEugenio Pérez static int vhost_vdpa_net_load_mac(VhostVDPAState *s, const VirtIONet *n) 675f73c0c43SEugenio Pérez { 67602d3bf09SHawkins Jiawei if (virtio_vdev_has_feature(&n->parent_obj, VIRTIO_NET_F_CTRL_MAC_ADDR)) { 6772848c6aaSHawkins Jiawei const struct iovec data = { 6782848c6aaSHawkins Jiawei .iov_base = (void *)n->mac, 6792848c6aaSHawkins Jiawei .iov_len = sizeof(n->mac), 6802848c6aaSHawkins Jiawei }; 681f73c0c43SEugenio Pérez ssize_t dev_written = vhost_vdpa_net_load_cmd(s, VIRTIO_NET_CTRL_MAC, 682f73c0c43SEugenio Pérez VIRTIO_NET_CTRL_MAC_ADDR_SET, 6832848c6aaSHawkins Jiawei &data, 1); 684f73c0c43SEugenio Pérez if (unlikely(dev_written < 0)) { 685f73c0c43SEugenio Pérez return dev_written; 686f73c0c43SEugenio Pérez } 687b479bc3cSHawkins Jiawei if (*s->status != VIRTIO_NET_OK) { 688b479bc3cSHawkins Jiawei return -EIO; 689b479bc3cSHawkins Jiawei } 690f73c0c43SEugenio Pérez } 691f73c0c43SEugenio Pérez 6920ddcecb8SHawkins Jiawei /* 6930ddcecb8SHawkins Jiawei * According to VirtIO standard, "The device MUST have an 6940ddcecb8SHawkins Jiawei * empty MAC filtering table on reset.". 6950ddcecb8SHawkins Jiawei * 6960ddcecb8SHawkins Jiawei * Therefore, there is no need to send this CVQ command if the 6970ddcecb8SHawkins Jiawei * driver also sets an empty MAC filter table, which aligns with 6980ddcecb8SHawkins Jiawei * the device's defaults. 6990ddcecb8SHawkins Jiawei * 7000ddcecb8SHawkins Jiawei * Note that the device's defaults can mismatch the driver's 7010ddcecb8SHawkins Jiawei * configuration only at live migration. 7020ddcecb8SHawkins Jiawei */ 7030ddcecb8SHawkins Jiawei if (!virtio_vdev_has_feature(&n->parent_obj, VIRTIO_NET_F_CTRL_RX) || 7040ddcecb8SHawkins Jiawei n->mac_table.in_use == 0) { 7050ddcecb8SHawkins Jiawei return 0; 7060ddcecb8SHawkins Jiawei } 7070ddcecb8SHawkins Jiawei 7080ddcecb8SHawkins Jiawei uint32_t uni_entries = n->mac_table.first_multi, 7090ddcecb8SHawkins Jiawei uni_macs_size = uni_entries * ETH_ALEN, 7100ddcecb8SHawkins Jiawei mul_entries = n->mac_table.in_use - uni_entries, 7110ddcecb8SHawkins Jiawei mul_macs_size = mul_entries * ETH_ALEN; 7120ddcecb8SHawkins Jiawei struct virtio_net_ctrl_mac uni = { 7130ddcecb8SHawkins Jiawei .entries = cpu_to_le32(uni_entries), 7140ddcecb8SHawkins Jiawei }; 7150ddcecb8SHawkins Jiawei struct virtio_net_ctrl_mac mul = { 7160ddcecb8SHawkins Jiawei .entries = cpu_to_le32(mul_entries), 7170ddcecb8SHawkins Jiawei }; 7180ddcecb8SHawkins Jiawei const struct iovec data[] = { 7190ddcecb8SHawkins Jiawei { 7200ddcecb8SHawkins Jiawei .iov_base = &uni, 7210ddcecb8SHawkins Jiawei .iov_len = sizeof(uni), 7220ddcecb8SHawkins Jiawei }, { 7230ddcecb8SHawkins Jiawei .iov_base = n->mac_table.macs, 7240ddcecb8SHawkins Jiawei .iov_len = uni_macs_size, 7250ddcecb8SHawkins Jiawei }, { 7260ddcecb8SHawkins Jiawei .iov_base = &mul, 7270ddcecb8SHawkins Jiawei .iov_len = sizeof(mul), 7280ddcecb8SHawkins Jiawei }, { 7290ddcecb8SHawkins Jiawei .iov_base = &n->mac_table.macs[uni_macs_size], 7300ddcecb8SHawkins Jiawei .iov_len = mul_macs_size, 7310ddcecb8SHawkins Jiawei }, 7320ddcecb8SHawkins Jiawei }; 7330ddcecb8SHawkins Jiawei ssize_t dev_written = vhost_vdpa_net_load_cmd(s, 7340ddcecb8SHawkins Jiawei VIRTIO_NET_CTRL_MAC, 7350ddcecb8SHawkins Jiawei VIRTIO_NET_CTRL_MAC_TABLE_SET, 7360ddcecb8SHawkins Jiawei data, ARRAY_SIZE(data)); 7370ddcecb8SHawkins Jiawei if (unlikely(dev_written < 0)) { 7380ddcecb8SHawkins Jiawei return dev_written; 7390ddcecb8SHawkins Jiawei } 7400ddcecb8SHawkins Jiawei if (*s->status != VIRTIO_NET_OK) { 7410ddcecb8SHawkins Jiawei return -EIO; 7420ddcecb8SHawkins Jiawei } 7430ddcecb8SHawkins Jiawei 744f73c0c43SEugenio Pérez return 0; 745f73c0c43SEugenio Pérez } 746f73c0c43SEugenio Pérez 747f64c7cdaSEugenio Pérez static int vhost_vdpa_net_load_mq(VhostVDPAState *s, 748f64c7cdaSEugenio Pérez const VirtIONet *n) 749f64c7cdaSEugenio Pérez { 750f64c7cdaSEugenio Pérez struct virtio_net_ctrl_mq mq; 751f64c7cdaSEugenio Pérez ssize_t dev_written; 752f64c7cdaSEugenio Pérez 75302d3bf09SHawkins Jiawei if (!virtio_vdev_has_feature(&n->parent_obj, VIRTIO_NET_F_MQ)) { 754f64c7cdaSEugenio Pérez return 0; 755f64c7cdaSEugenio Pérez } 756f64c7cdaSEugenio Pérez 757f64c7cdaSEugenio Pérez mq.virtqueue_pairs = cpu_to_le16(n->curr_queue_pairs); 7582848c6aaSHawkins Jiawei const struct iovec data = { 7592848c6aaSHawkins Jiawei .iov_base = &mq, 7602848c6aaSHawkins Jiawei .iov_len = sizeof(mq), 7612848c6aaSHawkins Jiawei }; 762f64c7cdaSEugenio Pérez dev_written = vhost_vdpa_net_load_cmd(s, VIRTIO_NET_CTRL_MQ, 7632848c6aaSHawkins Jiawei VIRTIO_NET_CTRL_MQ_VQ_PAIRS_SET, 7642848c6aaSHawkins Jiawei &data, 1); 765f64c7cdaSEugenio Pérez if (unlikely(dev_written < 0)) { 766f64c7cdaSEugenio Pérez return dev_written; 767f64c7cdaSEugenio Pérez } 768f45fd95eSHawkins Jiawei if (*s->status != VIRTIO_NET_OK) { 769f45fd95eSHawkins Jiawei return -EIO; 770f45fd95eSHawkins Jiawei } 771f64c7cdaSEugenio Pérez 772f45fd95eSHawkins Jiawei return 0; 773f64c7cdaSEugenio Pérez } 774f64c7cdaSEugenio Pérez 7750b58d368SHawkins Jiawei static int vhost_vdpa_net_load_offloads(VhostVDPAState *s, 7760b58d368SHawkins Jiawei const VirtIONet *n) 7770b58d368SHawkins Jiawei { 7780b58d368SHawkins Jiawei uint64_t offloads; 7790b58d368SHawkins Jiawei ssize_t dev_written; 7800b58d368SHawkins Jiawei 7810b58d368SHawkins Jiawei if (!virtio_vdev_has_feature(&n->parent_obj, 7820b58d368SHawkins Jiawei VIRTIO_NET_F_CTRL_GUEST_OFFLOADS)) { 7830b58d368SHawkins Jiawei return 0; 7840b58d368SHawkins Jiawei } 7850b58d368SHawkins Jiawei 7860b58d368SHawkins Jiawei if (n->curr_guest_offloads == virtio_net_supported_guest_offloads(n)) { 7870b58d368SHawkins Jiawei /* 7880b58d368SHawkins Jiawei * According to VirtIO standard, "Upon feature negotiation 7890b58d368SHawkins Jiawei * corresponding offload gets enabled to preserve 7900b58d368SHawkins Jiawei * backward compatibility.". 7910b58d368SHawkins Jiawei * 7920b58d368SHawkins Jiawei * Therefore, there is no need to send this CVQ command if the 7930b58d368SHawkins Jiawei * driver also enables all supported offloads, which aligns with 7940b58d368SHawkins Jiawei * the device's defaults. 7950b58d368SHawkins Jiawei * 7960b58d368SHawkins Jiawei * Note that the device's defaults can mismatch the driver's 7970b58d368SHawkins Jiawei * configuration only at live migration. 7980b58d368SHawkins Jiawei */ 7990b58d368SHawkins Jiawei return 0; 8000b58d368SHawkins Jiawei } 8010b58d368SHawkins Jiawei 8020b58d368SHawkins Jiawei offloads = cpu_to_le64(n->curr_guest_offloads); 8032848c6aaSHawkins Jiawei const struct iovec data = { 8042848c6aaSHawkins Jiawei .iov_base = &offloads, 8052848c6aaSHawkins Jiawei .iov_len = sizeof(offloads), 8062848c6aaSHawkins Jiawei }; 8070b58d368SHawkins Jiawei dev_written = vhost_vdpa_net_load_cmd(s, VIRTIO_NET_CTRL_GUEST_OFFLOADS, 8080b58d368SHawkins Jiawei VIRTIO_NET_CTRL_GUEST_OFFLOADS_SET, 8092848c6aaSHawkins Jiawei &data, 1); 8100b58d368SHawkins Jiawei if (unlikely(dev_written < 0)) { 8110b58d368SHawkins Jiawei return dev_written; 8120b58d368SHawkins Jiawei } 8136f348071SHawkins Jiawei if (*s->status != VIRTIO_NET_OK) { 8146f348071SHawkins Jiawei return -EIO; 8156f348071SHawkins Jiawei } 8160b58d368SHawkins Jiawei 8176f348071SHawkins Jiawei return 0; 8180b58d368SHawkins Jiawei } 8190b58d368SHawkins Jiawei 820b12f907eSHawkins Jiawei static int vhost_vdpa_net_load_rx_mode(VhostVDPAState *s, 821b12f907eSHawkins Jiawei uint8_t cmd, 822b12f907eSHawkins Jiawei uint8_t on) 823b12f907eSHawkins Jiawei { 824b12f907eSHawkins Jiawei const struct iovec data = { 825b12f907eSHawkins Jiawei .iov_base = &on, 826b12f907eSHawkins Jiawei .iov_len = sizeof(on), 827b12f907eSHawkins Jiawei }; 828b12f907eSHawkins Jiawei return vhost_vdpa_net_load_cmd(s, VIRTIO_NET_CTRL_RX, 829b12f907eSHawkins Jiawei cmd, &data, 1); 830b12f907eSHawkins Jiawei } 831b12f907eSHawkins Jiawei 832b12f907eSHawkins Jiawei static int vhost_vdpa_net_load_rx(VhostVDPAState *s, 833b12f907eSHawkins Jiawei const VirtIONet *n) 834b12f907eSHawkins Jiawei { 835b12f907eSHawkins Jiawei ssize_t dev_written; 836b12f907eSHawkins Jiawei 837b12f907eSHawkins Jiawei if (!virtio_vdev_has_feature(&n->parent_obj, VIRTIO_NET_F_CTRL_RX)) { 838b12f907eSHawkins Jiawei return 0; 839b12f907eSHawkins Jiawei } 840b12f907eSHawkins Jiawei 841b12f907eSHawkins Jiawei /* 842b12f907eSHawkins Jiawei * According to virtio_net_reset(), device turns promiscuous mode 843b12f907eSHawkins Jiawei * on by default. 844b12f907eSHawkins Jiawei * 8450a19d879SMichael Tokarev * Additionally, according to VirtIO standard, "Since there are 846b12f907eSHawkins Jiawei * no guarantees, it can use a hash filter or silently switch to 847b12f907eSHawkins Jiawei * allmulti or promiscuous mode if it is given too many addresses.". 848b12f907eSHawkins Jiawei * QEMU marks `n->mac_table.uni_overflow` if guest sets too many 849b12f907eSHawkins Jiawei * non-multicast MAC addresses, indicating that promiscuous mode 850b12f907eSHawkins Jiawei * should be enabled. 851b12f907eSHawkins Jiawei * 852b12f907eSHawkins Jiawei * Therefore, QEMU should only send this CVQ command if the 853b12f907eSHawkins Jiawei * `n->mac_table.uni_overflow` is not marked and `n->promisc` is off, 854b12f907eSHawkins Jiawei * which sets promiscuous mode on, different from the device's defaults. 855b12f907eSHawkins Jiawei * 856b12f907eSHawkins Jiawei * Note that the device's defaults can mismatch the driver's 857b12f907eSHawkins Jiawei * configuration only at live migration. 858b12f907eSHawkins Jiawei */ 859b12f907eSHawkins Jiawei if (!n->mac_table.uni_overflow && !n->promisc) { 860b12f907eSHawkins Jiawei dev_written = vhost_vdpa_net_load_rx_mode(s, 861b12f907eSHawkins Jiawei VIRTIO_NET_CTRL_RX_PROMISC, 0); 862b12f907eSHawkins Jiawei if (unlikely(dev_written < 0)) { 863b12f907eSHawkins Jiawei return dev_written; 864b12f907eSHawkins Jiawei } 865b12f907eSHawkins Jiawei if (*s->status != VIRTIO_NET_OK) { 866b12f907eSHawkins Jiawei return -EIO; 867b12f907eSHawkins Jiawei } 868b12f907eSHawkins Jiawei } 869b12f907eSHawkins Jiawei 870b12f907eSHawkins Jiawei /* 871b12f907eSHawkins Jiawei * According to virtio_net_reset(), device turns all-multicast mode 872b12f907eSHawkins Jiawei * off by default. 873b12f907eSHawkins Jiawei * 874b12f907eSHawkins Jiawei * According to VirtIO standard, "Since there are no guarantees, 875b12f907eSHawkins Jiawei * it can use a hash filter or silently switch to allmulti or 876b12f907eSHawkins Jiawei * promiscuous mode if it is given too many addresses.". QEMU marks 877b12f907eSHawkins Jiawei * `n->mac_table.multi_overflow` if guest sets too many 878b12f907eSHawkins Jiawei * non-multicast MAC addresses. 879b12f907eSHawkins Jiawei * 880b12f907eSHawkins Jiawei * Therefore, QEMU should only send this CVQ command if the 881b12f907eSHawkins Jiawei * `n->mac_table.multi_overflow` is marked or `n->allmulti` is on, 882b12f907eSHawkins Jiawei * which sets all-multicast mode on, different from the device's defaults. 883b12f907eSHawkins Jiawei * 884b12f907eSHawkins Jiawei * Note that the device's defaults can mismatch the driver's 885b12f907eSHawkins Jiawei * configuration only at live migration. 886b12f907eSHawkins Jiawei */ 887b12f907eSHawkins Jiawei if (n->mac_table.multi_overflow || n->allmulti) { 888b12f907eSHawkins Jiawei dev_written = vhost_vdpa_net_load_rx_mode(s, 889b12f907eSHawkins Jiawei VIRTIO_NET_CTRL_RX_ALLMULTI, 1); 890b12f907eSHawkins Jiawei if (unlikely(dev_written < 0)) { 891b12f907eSHawkins Jiawei return dev_written; 892b12f907eSHawkins Jiawei } 893b12f907eSHawkins Jiawei if (*s->status != VIRTIO_NET_OK) { 894b12f907eSHawkins Jiawei return -EIO; 895b12f907eSHawkins Jiawei } 896b12f907eSHawkins Jiawei } 897b12f907eSHawkins Jiawei 8984fd180c7SHawkins Jiawei if (!virtio_vdev_has_feature(&n->parent_obj, VIRTIO_NET_F_CTRL_RX_EXTRA)) { 8994fd180c7SHawkins Jiawei return 0; 9004fd180c7SHawkins Jiawei } 9014fd180c7SHawkins Jiawei 9024fd180c7SHawkins Jiawei /* 9034fd180c7SHawkins Jiawei * According to virtio_net_reset(), device turns all-unicast mode 9044fd180c7SHawkins Jiawei * off by default. 9054fd180c7SHawkins Jiawei * 9064fd180c7SHawkins Jiawei * Therefore, QEMU should only send this CVQ command if the driver 9074fd180c7SHawkins Jiawei * sets all-unicast mode on, different from the device's defaults. 9084fd180c7SHawkins Jiawei * 9094fd180c7SHawkins Jiawei * Note that the device's defaults can mismatch the driver's 9104fd180c7SHawkins Jiawei * configuration only at live migration. 9114fd180c7SHawkins Jiawei */ 9124fd180c7SHawkins Jiawei if (n->alluni) { 9134fd180c7SHawkins Jiawei dev_written = vhost_vdpa_net_load_rx_mode(s, 9144fd180c7SHawkins Jiawei VIRTIO_NET_CTRL_RX_ALLUNI, 1); 9154fd180c7SHawkins Jiawei if (dev_written < 0) { 9164fd180c7SHawkins Jiawei return dev_written; 9174fd180c7SHawkins Jiawei } 9184fd180c7SHawkins Jiawei if (*s->status != VIRTIO_NET_OK) { 9194fd180c7SHawkins Jiawei return -EIO; 9204fd180c7SHawkins Jiawei } 9214fd180c7SHawkins Jiawei } 9224fd180c7SHawkins Jiawei 9234fd180c7SHawkins Jiawei /* 9244fd180c7SHawkins Jiawei * According to virtio_net_reset(), device turns non-multicast mode 9254fd180c7SHawkins Jiawei * off by default. 9264fd180c7SHawkins Jiawei * 9274fd180c7SHawkins Jiawei * Therefore, QEMU should only send this CVQ command if the driver 9284fd180c7SHawkins Jiawei * sets non-multicast mode on, different from the device's defaults. 9294fd180c7SHawkins Jiawei * 9304fd180c7SHawkins Jiawei * Note that the device's defaults can mismatch the driver's 9314fd180c7SHawkins Jiawei * configuration only at live migration. 9324fd180c7SHawkins Jiawei */ 9334fd180c7SHawkins Jiawei if (n->nomulti) { 9344fd180c7SHawkins Jiawei dev_written = vhost_vdpa_net_load_rx_mode(s, 9354fd180c7SHawkins Jiawei VIRTIO_NET_CTRL_RX_NOMULTI, 1); 9364fd180c7SHawkins Jiawei if (dev_written < 0) { 9374fd180c7SHawkins Jiawei return dev_written; 9384fd180c7SHawkins Jiawei } 9394fd180c7SHawkins Jiawei if (*s->status != VIRTIO_NET_OK) { 9404fd180c7SHawkins Jiawei return -EIO; 9414fd180c7SHawkins Jiawei } 9424fd180c7SHawkins Jiawei } 9434fd180c7SHawkins Jiawei 9444fd180c7SHawkins Jiawei /* 9454fd180c7SHawkins Jiawei * According to virtio_net_reset(), device turns non-unicast mode 9464fd180c7SHawkins Jiawei * off by default. 9474fd180c7SHawkins Jiawei * 9484fd180c7SHawkins Jiawei * Therefore, QEMU should only send this CVQ command if the driver 9494fd180c7SHawkins Jiawei * sets non-unicast mode on, different from the device's defaults. 9504fd180c7SHawkins Jiawei * 9514fd180c7SHawkins Jiawei * Note that the device's defaults can mismatch the driver's 9524fd180c7SHawkins Jiawei * configuration only at live migration. 9534fd180c7SHawkins Jiawei */ 9544fd180c7SHawkins Jiawei if (n->nouni) { 9554fd180c7SHawkins Jiawei dev_written = vhost_vdpa_net_load_rx_mode(s, 9564fd180c7SHawkins Jiawei VIRTIO_NET_CTRL_RX_NOUNI, 1); 9574fd180c7SHawkins Jiawei if (dev_written < 0) { 9584fd180c7SHawkins Jiawei return dev_written; 9594fd180c7SHawkins Jiawei } 9604fd180c7SHawkins Jiawei if (*s->status != VIRTIO_NET_OK) { 9614fd180c7SHawkins Jiawei return -EIO; 9624fd180c7SHawkins Jiawei } 9634fd180c7SHawkins Jiawei } 9644fd180c7SHawkins Jiawei 9654fd180c7SHawkins Jiawei /* 9664fd180c7SHawkins Jiawei * According to virtio_net_reset(), device turns non-broadcast mode 9674fd180c7SHawkins Jiawei * off by default. 9684fd180c7SHawkins Jiawei * 9694fd180c7SHawkins Jiawei * Therefore, QEMU should only send this CVQ command if the driver 9704fd180c7SHawkins Jiawei * sets non-broadcast mode on, different from the device's defaults. 9714fd180c7SHawkins Jiawei * 9724fd180c7SHawkins Jiawei * Note that the device's defaults can mismatch the driver's 9734fd180c7SHawkins Jiawei * configuration only at live migration. 9744fd180c7SHawkins Jiawei */ 9754fd180c7SHawkins Jiawei if (n->nobcast) { 9764fd180c7SHawkins Jiawei dev_written = vhost_vdpa_net_load_rx_mode(s, 9774fd180c7SHawkins Jiawei VIRTIO_NET_CTRL_RX_NOBCAST, 1); 9784fd180c7SHawkins Jiawei if (dev_written < 0) { 9794fd180c7SHawkins Jiawei return dev_written; 9804fd180c7SHawkins Jiawei } 9814fd180c7SHawkins Jiawei if (*s->status != VIRTIO_NET_OK) { 9824fd180c7SHawkins Jiawei return -EIO; 9834fd180c7SHawkins Jiawei } 9844fd180c7SHawkins Jiawei } 9854fd180c7SHawkins Jiawei 986b12f907eSHawkins Jiawei return 0; 987b12f907eSHawkins Jiawei } 988b12f907eSHawkins Jiawei 9898f7e9967SHawkins Jiawei static int vhost_vdpa_net_load_single_vlan(VhostVDPAState *s, 9908f7e9967SHawkins Jiawei const VirtIONet *n, 9918f7e9967SHawkins Jiawei uint16_t vid) 9928f7e9967SHawkins Jiawei { 9938f7e9967SHawkins Jiawei const struct iovec data = { 9948f7e9967SHawkins Jiawei .iov_base = &vid, 9958f7e9967SHawkins Jiawei .iov_len = sizeof(vid), 9968f7e9967SHawkins Jiawei }; 9978f7e9967SHawkins Jiawei ssize_t dev_written = vhost_vdpa_net_load_cmd(s, VIRTIO_NET_CTRL_VLAN, 9988f7e9967SHawkins Jiawei VIRTIO_NET_CTRL_VLAN_ADD, 9998f7e9967SHawkins Jiawei &data, 1); 10008f7e9967SHawkins Jiawei if (unlikely(dev_written < 0)) { 10018f7e9967SHawkins Jiawei return dev_written; 10028f7e9967SHawkins Jiawei } 10038f7e9967SHawkins Jiawei if (unlikely(*s->status != VIRTIO_NET_OK)) { 10048f7e9967SHawkins Jiawei return -EIO; 10058f7e9967SHawkins Jiawei } 10068f7e9967SHawkins Jiawei 10078f7e9967SHawkins Jiawei return 0; 10088f7e9967SHawkins Jiawei } 10098f7e9967SHawkins Jiawei 10108f7e9967SHawkins Jiawei static int vhost_vdpa_net_load_vlan(VhostVDPAState *s, 10118f7e9967SHawkins Jiawei const VirtIONet *n) 10128f7e9967SHawkins Jiawei { 10138f7e9967SHawkins Jiawei int r; 10148f7e9967SHawkins Jiawei 10158f7e9967SHawkins Jiawei if (!virtio_vdev_has_feature(&n->parent_obj, VIRTIO_NET_F_CTRL_VLAN)) { 10168f7e9967SHawkins Jiawei return 0; 10178f7e9967SHawkins Jiawei } 10188f7e9967SHawkins Jiawei 10198f7e9967SHawkins Jiawei for (int i = 0; i < MAX_VLAN >> 5; i++) { 10208f7e9967SHawkins Jiawei for (int j = 0; n->vlans[i] && j <= 0x1f; j++) { 10218f7e9967SHawkins Jiawei if (n->vlans[i] & (1U << j)) { 10228f7e9967SHawkins Jiawei r = vhost_vdpa_net_load_single_vlan(s, n, (i << 5) + j); 10238f7e9967SHawkins Jiawei if (unlikely(r != 0)) { 10248f7e9967SHawkins Jiawei return r; 10258f7e9967SHawkins Jiawei } 10268f7e9967SHawkins Jiawei } 10278f7e9967SHawkins Jiawei } 10288f7e9967SHawkins Jiawei } 10298f7e9967SHawkins Jiawei 10308f7e9967SHawkins Jiawei return 0; 10318f7e9967SHawkins Jiawei } 10328f7e9967SHawkins Jiawei 1033f3fada59SEugenio Pérez static int vhost_vdpa_net_cvq_load(NetClientState *nc) 1034dd036d8dSEugenio Pérez { 1035dd036d8dSEugenio Pérez VhostVDPAState *s = DO_UPCAST(VhostVDPAState, nc, nc); 1036f73c0c43SEugenio Pérez struct vhost_vdpa *v = &s->vhost_vdpa; 1037dd036d8dSEugenio Pérez const VirtIONet *n; 1038f73c0c43SEugenio Pérez int r; 1039dd036d8dSEugenio Pérez 1040dd036d8dSEugenio Pérez assert(nc->info->type == NET_CLIENT_DRIVER_VHOST_VDPA); 1041dd036d8dSEugenio Pérez 10426c482547SEugenio Pérez vhost_vdpa_set_vring_ready(v, v->dev->vq_index); 1043dd036d8dSEugenio Pérez 10446c482547SEugenio Pérez if (v->shadow_vqs_enabled) { 1045dd036d8dSEugenio Pérez n = VIRTIO_NET(v->dev->vdev); 1046f73c0c43SEugenio Pérez r = vhost_vdpa_net_load_mac(s, n); 1047f73c0c43SEugenio Pérez if (unlikely(r < 0)) { 1048f73c0c43SEugenio Pérez return r; 1049dd036d8dSEugenio Pérez } 1050f64c7cdaSEugenio Pérez r = vhost_vdpa_net_load_mq(s, n); 1051f64c7cdaSEugenio Pérez if (unlikely(r)) { 1052f64c7cdaSEugenio Pérez return r; 1053f64c7cdaSEugenio Pérez } 10540b58d368SHawkins Jiawei r = vhost_vdpa_net_load_offloads(s, n); 10550b58d368SHawkins Jiawei if (unlikely(r)) { 10560b58d368SHawkins Jiawei return r; 10570b58d368SHawkins Jiawei } 1058b12f907eSHawkins Jiawei r = vhost_vdpa_net_load_rx(s, n); 1059b12f907eSHawkins Jiawei if (unlikely(r)) { 1060b12f907eSHawkins Jiawei return r; 1061b12f907eSHawkins Jiawei } 10628f7e9967SHawkins Jiawei r = vhost_vdpa_net_load_vlan(s, n); 10638f7e9967SHawkins Jiawei if (unlikely(r)) { 10648f7e9967SHawkins Jiawei return r; 10658f7e9967SHawkins Jiawei } 10666c482547SEugenio Pérez } 10676c482547SEugenio Pérez 10686c482547SEugenio Pérez for (int i = 0; i < v->dev->vq_index; ++i) { 10696c482547SEugenio Pérez vhost_vdpa_set_vring_ready(v, i); 10706c482547SEugenio Pérez } 1071dd036d8dSEugenio Pérez 1072dd036d8dSEugenio Pérez return 0; 1073dd036d8dSEugenio Pérez } 1074dd036d8dSEugenio Pérez 1075f8972b56SEugenio Pérez static NetClientInfo net_vhost_vdpa_cvq_info = { 1076f8972b56SEugenio Pérez .type = NET_CLIENT_DRIVER_VHOST_VDPA, 1077f8972b56SEugenio Pérez .size = sizeof(VhostVDPAState), 1078f8972b56SEugenio Pérez .receive = vhost_vdpa_receive, 10797a7f87e9SEugenio Pérez .start = vhost_vdpa_net_cvq_start, 1080f3fada59SEugenio Pérez .load = vhost_vdpa_net_cvq_load, 10817a7f87e9SEugenio Pérez .stop = vhost_vdpa_net_cvq_stop, 1082f8972b56SEugenio Pérez .cleanup = vhost_vdpa_cleanup, 1083f8972b56SEugenio Pérez .has_vnet_hdr = vhost_vdpa_has_vnet_hdr, 1084f8972b56SEugenio Pérez .has_ufo = vhost_vdpa_has_ufo, 1085f8972b56SEugenio Pérez .check_peer_type = vhost_vdpa_check_peer_type, 1086f8972b56SEugenio Pérez }; 1087f8972b56SEugenio Pérez 1088fee364e4SHawkins Jiawei /* 1089fee364e4SHawkins Jiawei * Forward the excessive VIRTIO_NET_CTRL_MAC_TABLE_SET CVQ command to 1090fee364e4SHawkins Jiawei * vdpa device. 1091fee364e4SHawkins Jiawei * 1092fee364e4SHawkins Jiawei * Considering that QEMU cannot send the entire filter table to the 1093fee364e4SHawkins Jiawei * vdpa device, it should send the VIRTIO_NET_CTRL_RX_PROMISC CVQ 1094fee364e4SHawkins Jiawei * command to enable promiscuous mode to receive all packets, 1095fee364e4SHawkins Jiawei * according to VirtIO standard, "Since there are no guarantees, 1096fee364e4SHawkins Jiawei * it can use a hash filter or silently switch to allmulti or 1097fee364e4SHawkins Jiawei * promiscuous mode if it is given too many addresses.". 1098fee364e4SHawkins Jiawei * 1099fee364e4SHawkins Jiawei * Since QEMU ignores MAC addresses beyond `MAC_TABLE_ENTRIES` and 1100fee364e4SHawkins Jiawei * marks `n->mac_table.x_overflow` accordingly, it should have 1101fee364e4SHawkins Jiawei * the same effect on the device model to receive 1102fee364e4SHawkins Jiawei * (`MAC_TABLE_ENTRIES` + 1) or more non-multicast MAC addresses. 1103fee364e4SHawkins Jiawei * The same applies to multicast MAC addresses. 1104fee364e4SHawkins Jiawei * 1105fee364e4SHawkins Jiawei * Therefore, QEMU can provide the device model with a fake 1106fee364e4SHawkins Jiawei * VIRTIO_NET_CTRL_MAC_TABLE_SET command with (`MAC_TABLE_ENTRIES` + 1) 1107fee364e4SHawkins Jiawei * non-multicast MAC addresses and (`MAC_TABLE_ENTRIES` + 1) multicast 1108fee364e4SHawkins Jiawei * MAC addresses. This ensures that the device model marks 1109fee364e4SHawkins Jiawei * `n->mac_table.uni_overflow` and `n->mac_table.multi_overflow`, 1110fee364e4SHawkins Jiawei * allowing all packets to be received, which aligns with the 1111fee364e4SHawkins Jiawei * state of the vdpa device. 1112fee364e4SHawkins Jiawei */ 1113fee364e4SHawkins Jiawei static int vhost_vdpa_net_excessive_mac_filter_cvq_add(VhostVDPAState *s, 1114fee364e4SHawkins Jiawei VirtQueueElement *elem, 1115fee364e4SHawkins Jiawei struct iovec *out) 1116fee364e4SHawkins Jiawei { 1117fee364e4SHawkins Jiawei struct virtio_net_ctrl_mac mac_data, *mac_ptr; 1118fee364e4SHawkins Jiawei struct virtio_net_ctrl_hdr *hdr_ptr; 1119fee364e4SHawkins Jiawei uint32_t cursor; 1120fee364e4SHawkins Jiawei ssize_t r; 1121fee364e4SHawkins Jiawei 1122fee364e4SHawkins Jiawei /* parse the non-multicast MAC address entries from CVQ command */ 1123fee364e4SHawkins Jiawei cursor = sizeof(*hdr_ptr); 1124fee364e4SHawkins Jiawei r = iov_to_buf(elem->out_sg, elem->out_num, cursor, 1125fee364e4SHawkins Jiawei &mac_data, sizeof(mac_data)); 1126fee364e4SHawkins Jiawei if (unlikely(r != sizeof(mac_data))) { 1127fee364e4SHawkins Jiawei /* 1128fee364e4SHawkins Jiawei * If the CVQ command is invalid, we should simulate the vdpa device 1129fee364e4SHawkins Jiawei * to reject the VIRTIO_NET_CTRL_MAC_TABLE_SET CVQ command 1130fee364e4SHawkins Jiawei */ 1131fee364e4SHawkins Jiawei *s->status = VIRTIO_NET_ERR; 1132fee364e4SHawkins Jiawei return sizeof(*s->status); 1133fee364e4SHawkins Jiawei } 1134fee364e4SHawkins Jiawei cursor += sizeof(mac_data) + le32_to_cpu(mac_data.entries) * ETH_ALEN; 1135fee364e4SHawkins Jiawei 1136fee364e4SHawkins Jiawei /* parse the multicast MAC address entries from CVQ command */ 1137fee364e4SHawkins Jiawei r = iov_to_buf(elem->out_sg, elem->out_num, cursor, 1138fee364e4SHawkins Jiawei &mac_data, sizeof(mac_data)); 1139fee364e4SHawkins Jiawei if (r != sizeof(mac_data)) { 1140fee364e4SHawkins Jiawei /* 1141fee364e4SHawkins Jiawei * If the CVQ command is invalid, we should simulate the vdpa device 1142fee364e4SHawkins Jiawei * to reject the VIRTIO_NET_CTRL_MAC_TABLE_SET CVQ command 1143fee364e4SHawkins Jiawei */ 1144fee364e4SHawkins Jiawei *s->status = VIRTIO_NET_ERR; 1145fee364e4SHawkins Jiawei return sizeof(*s->status); 1146fee364e4SHawkins Jiawei } 1147fee364e4SHawkins Jiawei cursor += sizeof(mac_data) + le32_to_cpu(mac_data.entries) * ETH_ALEN; 1148fee364e4SHawkins Jiawei 1149fee364e4SHawkins Jiawei /* validate the CVQ command */ 1150fee364e4SHawkins Jiawei if (iov_size(elem->out_sg, elem->out_num) != cursor) { 1151fee364e4SHawkins Jiawei /* 1152fee364e4SHawkins Jiawei * If the CVQ command is invalid, we should simulate the vdpa device 1153fee364e4SHawkins Jiawei * to reject the VIRTIO_NET_CTRL_MAC_TABLE_SET CVQ command 1154fee364e4SHawkins Jiawei */ 1155fee364e4SHawkins Jiawei *s->status = VIRTIO_NET_ERR; 1156fee364e4SHawkins Jiawei return sizeof(*s->status); 1157fee364e4SHawkins Jiawei } 1158fee364e4SHawkins Jiawei 1159fee364e4SHawkins Jiawei /* 1160fee364e4SHawkins Jiawei * According to VirtIO standard, "Since there are no guarantees, 1161fee364e4SHawkins Jiawei * it can use a hash filter or silently switch to allmulti or 1162fee364e4SHawkins Jiawei * promiscuous mode if it is given too many addresses.". 1163fee364e4SHawkins Jiawei * 1164fee364e4SHawkins Jiawei * Therefore, considering that QEMU is unable to send the entire 1165fee364e4SHawkins Jiawei * filter table to the vdpa device, it should send the 1166fee364e4SHawkins Jiawei * VIRTIO_NET_CTRL_RX_PROMISC CVQ command to enable promiscuous mode 1167fee364e4SHawkins Jiawei */ 1168fee364e4SHawkins Jiawei r = vhost_vdpa_net_load_rx_mode(s, VIRTIO_NET_CTRL_RX_PROMISC, 1); 1169fee364e4SHawkins Jiawei if (unlikely(r < 0)) { 1170fee364e4SHawkins Jiawei return r; 1171fee364e4SHawkins Jiawei } 1172fee364e4SHawkins Jiawei if (*s->status != VIRTIO_NET_OK) { 1173fee364e4SHawkins Jiawei return sizeof(*s->status); 1174fee364e4SHawkins Jiawei } 1175fee364e4SHawkins Jiawei 1176fee364e4SHawkins Jiawei /* 1177fee364e4SHawkins Jiawei * QEMU should also send a fake VIRTIO_NET_CTRL_MAC_TABLE_SET CVQ 1178fee364e4SHawkins Jiawei * command to the device model, including (`MAC_TABLE_ENTRIES` + 1) 1179fee364e4SHawkins Jiawei * non-multicast MAC addresses and (`MAC_TABLE_ENTRIES` + 1) 1180fee364e4SHawkins Jiawei * multicast MAC addresses. 1181fee364e4SHawkins Jiawei * 1182fee364e4SHawkins Jiawei * By doing so, the device model can mark `n->mac_table.uni_overflow` 1183fee364e4SHawkins Jiawei * and `n->mac_table.multi_overflow`, enabling all packets to be 1184fee364e4SHawkins Jiawei * received, which aligns with the state of the vdpa device. 1185fee364e4SHawkins Jiawei */ 1186fee364e4SHawkins Jiawei cursor = 0; 1187fee364e4SHawkins Jiawei uint32_t fake_uni_entries = MAC_TABLE_ENTRIES + 1, 1188fee364e4SHawkins Jiawei fake_mul_entries = MAC_TABLE_ENTRIES + 1, 1189fee364e4SHawkins Jiawei fake_cvq_size = sizeof(struct virtio_net_ctrl_hdr) + 1190fee364e4SHawkins Jiawei sizeof(mac_data) + fake_uni_entries * ETH_ALEN + 1191fee364e4SHawkins Jiawei sizeof(mac_data) + fake_mul_entries * ETH_ALEN; 1192fee364e4SHawkins Jiawei 1193fee364e4SHawkins Jiawei assert(fake_cvq_size < vhost_vdpa_net_cvq_cmd_page_len()); 1194fee364e4SHawkins Jiawei out->iov_len = fake_cvq_size; 1195fee364e4SHawkins Jiawei 1196fee364e4SHawkins Jiawei /* pack the header for fake CVQ command */ 1197fee364e4SHawkins Jiawei hdr_ptr = out->iov_base + cursor; 1198fee364e4SHawkins Jiawei hdr_ptr->class = VIRTIO_NET_CTRL_MAC; 1199fee364e4SHawkins Jiawei hdr_ptr->cmd = VIRTIO_NET_CTRL_MAC_TABLE_SET; 1200fee364e4SHawkins Jiawei cursor += sizeof(*hdr_ptr); 1201fee364e4SHawkins Jiawei 1202fee364e4SHawkins Jiawei /* 1203fee364e4SHawkins Jiawei * Pack the non-multicast MAC addresses part for fake CVQ command. 1204fee364e4SHawkins Jiawei * 1205fee364e4SHawkins Jiawei * According to virtio_net_handle_mac(), QEMU doesn't verify the MAC 12060a19d879SMichael Tokarev * addresses provided in CVQ command. Therefore, only the entries 1207fee364e4SHawkins Jiawei * field need to be prepared in the CVQ command. 1208fee364e4SHawkins Jiawei */ 1209fee364e4SHawkins Jiawei mac_ptr = out->iov_base + cursor; 1210fee364e4SHawkins Jiawei mac_ptr->entries = cpu_to_le32(fake_uni_entries); 1211fee364e4SHawkins Jiawei cursor += sizeof(*mac_ptr) + fake_uni_entries * ETH_ALEN; 1212fee364e4SHawkins Jiawei 1213fee364e4SHawkins Jiawei /* 1214fee364e4SHawkins Jiawei * Pack the multicast MAC addresses part for fake CVQ command. 1215fee364e4SHawkins Jiawei * 1216fee364e4SHawkins Jiawei * According to virtio_net_handle_mac(), QEMU doesn't verify the MAC 12170a19d879SMichael Tokarev * addresses provided in CVQ command. Therefore, only the entries 1218fee364e4SHawkins Jiawei * field need to be prepared in the CVQ command. 1219fee364e4SHawkins Jiawei */ 1220fee364e4SHawkins Jiawei mac_ptr = out->iov_base + cursor; 1221fee364e4SHawkins Jiawei mac_ptr->entries = cpu_to_le32(fake_mul_entries); 1222fee364e4SHawkins Jiawei 1223fee364e4SHawkins Jiawei /* 1224fee364e4SHawkins Jiawei * Simulating QEMU poll a vdpa device used buffer 1225fee364e4SHawkins Jiawei * for VIRTIO_NET_CTRL_MAC_TABLE_SET CVQ command 1226fee364e4SHawkins Jiawei */ 1227fee364e4SHawkins Jiawei return sizeof(*s->status); 1228fee364e4SHawkins Jiawei } 1229fee364e4SHawkins Jiawei 12302df4dd31SEugenio Pérez /** 12312df4dd31SEugenio Pérez * Validate and copy control virtqueue commands. 12322df4dd31SEugenio Pérez * 12332df4dd31SEugenio Pérez * Following QEMU guidelines, we offer a copy of the buffers to the device to 12342df4dd31SEugenio Pérez * prevent TOCTOU bugs. 1235bd907ae4SEugenio Pérez */ 1236bd907ae4SEugenio Pérez static int vhost_vdpa_net_handle_ctrl_avail(VhostShadowVirtqueue *svq, 1237bd907ae4SEugenio Pérez VirtQueueElement *elem, 1238bd907ae4SEugenio Pérez void *opaque) 1239bd907ae4SEugenio Pérez { 12402df4dd31SEugenio Pérez VhostVDPAState *s = opaque; 1241be4278b6SEugenio Pérez size_t in_len; 124245c41018SHawkins Jiawei const struct virtio_net_ctrl_hdr *ctrl; 1243bd907ae4SEugenio Pérez virtio_net_ctrl_ack status = VIRTIO_NET_ERR; 12447a7f87e9SEugenio Pérez /* Out buffer sent to both the vdpa device and the device model */ 12457a7f87e9SEugenio Pérez struct iovec out = { 12467a7f87e9SEugenio Pérez .iov_base = s->cvq_cmd_out_buffer, 12477a7f87e9SEugenio Pérez }; 12482df4dd31SEugenio Pérez /* in buffer used for device model */ 12492df4dd31SEugenio Pérez const struct iovec in = { 12502df4dd31SEugenio Pérez .iov_base = &status, 12512df4dd31SEugenio Pérez .iov_len = sizeof(status), 12522df4dd31SEugenio Pérez }; 1253be4278b6SEugenio Pérez ssize_t dev_written = -EINVAL; 1254bd907ae4SEugenio Pérez 12557a7f87e9SEugenio Pérez out.iov_len = iov_to_buf(elem->out_sg, elem->out_num, 0, 12567a7f87e9SEugenio Pérez s->cvq_cmd_out_buffer, 1257fee364e4SHawkins Jiawei vhost_vdpa_net_cvq_cmd_page_len()); 125845c41018SHawkins Jiawei 125945c41018SHawkins Jiawei ctrl = s->cvq_cmd_out_buffer; 126045c41018SHawkins Jiawei if (ctrl->class == VIRTIO_NET_CTRL_ANNOUNCE) { 12613f9a3eebSEugenio Pérez /* 12623f9a3eebSEugenio Pérez * Guest announce capability is emulated by qemu, so don't forward to 12633f9a3eebSEugenio Pérez * the device. 12643f9a3eebSEugenio Pérez */ 12653f9a3eebSEugenio Pérez dev_written = sizeof(status); 12663f9a3eebSEugenio Pérez *s->status = VIRTIO_NET_OK; 1267fee364e4SHawkins Jiawei } else if (unlikely(ctrl->class == VIRTIO_NET_CTRL_MAC && 1268fee364e4SHawkins Jiawei ctrl->cmd == VIRTIO_NET_CTRL_MAC_TABLE_SET && 1269fee364e4SHawkins Jiawei iov_size(elem->out_sg, elem->out_num) > out.iov_len)) { 1270fee364e4SHawkins Jiawei /* 1271fee364e4SHawkins Jiawei * Due to the size limitation of the out buffer sent to the vdpa device, 1272fee364e4SHawkins Jiawei * which is determined by vhost_vdpa_net_cvq_cmd_page_len(), excessive 1273fee364e4SHawkins Jiawei * MAC addresses set by the driver for the filter table can cause 1274fee364e4SHawkins Jiawei * truncation of the CVQ command in QEMU. As a result, the vdpa device 1275fee364e4SHawkins Jiawei * rejects the flawed CVQ command. 1276fee364e4SHawkins Jiawei * 1277fee364e4SHawkins Jiawei * Therefore, QEMU must handle this situation instead of sending 12780a19d879SMichael Tokarev * the CVQ command directly. 1279fee364e4SHawkins Jiawei */ 1280fee364e4SHawkins Jiawei dev_written = vhost_vdpa_net_excessive_mac_filter_cvq_add(s, elem, 1281fee364e4SHawkins Jiawei &out); 1282fee364e4SHawkins Jiawei if (unlikely(dev_written < 0)) { 1283fee364e4SHawkins Jiawei goto out; 1284fee364e4SHawkins Jiawei } 12853f9a3eebSEugenio Pérez } else { 1286be4278b6SEugenio Pérez dev_written = vhost_vdpa_net_cvq_add(s, out.iov_len, sizeof(status)); 1287be4278b6SEugenio Pérez if (unlikely(dev_written < 0)) { 1288bd907ae4SEugenio Pérez goto out; 1289bd907ae4SEugenio Pérez } 12903f9a3eebSEugenio Pérez } 1291bd907ae4SEugenio Pérez 1292bd907ae4SEugenio Pérez if (unlikely(dev_written < sizeof(status))) { 1293bd907ae4SEugenio Pérez error_report("Insufficient written data (%zu)", dev_written); 12942df4dd31SEugenio Pérez goto out; 12952df4dd31SEugenio Pérez } 12962df4dd31SEugenio Pérez 129717fb889fSEugenio Pérez if (*s->status != VIRTIO_NET_OK) { 1298d45243bcSEugenio Pérez goto out; 12992df4dd31SEugenio Pérez } 13002df4dd31SEugenio Pérez 13012df4dd31SEugenio Pérez status = VIRTIO_NET_ERR; 13027a7f87e9SEugenio Pérez virtio_net_handle_ctrl_iov(svq->vdev, &in, 1, &out, 1); 13032df4dd31SEugenio Pérez if (status != VIRTIO_NET_OK) { 13042df4dd31SEugenio Pérez error_report("Bad CVQ processing in model"); 1305bd907ae4SEugenio Pérez } 1306bd907ae4SEugenio Pérez 1307bd907ae4SEugenio Pérez out: 1308bd907ae4SEugenio Pérez in_len = iov_from_buf(elem->in_sg, elem->in_num, 0, &status, 1309bd907ae4SEugenio Pérez sizeof(status)); 1310bd907ae4SEugenio Pérez if (unlikely(in_len < sizeof(status))) { 1311bd907ae4SEugenio Pérez error_report("Bad device CVQ written length"); 1312bd907ae4SEugenio Pérez } 1313bd907ae4SEugenio Pérez vhost_svq_push_elem(svq, elem, MIN(in_len, sizeof(status))); 1314031b1abaSHawkins Jiawei /* 1315031b1abaSHawkins Jiawei * `elem` belongs to vhost_vdpa_net_handle_ctrl_avail() only when 1316031b1abaSHawkins Jiawei * the function successfully forwards the CVQ command, indicated 1317031b1abaSHawkins Jiawei * by a non-negative value of `dev_written`. Otherwise, it still 1318031b1abaSHawkins Jiawei * belongs to SVQ. 1319031b1abaSHawkins Jiawei * This function should only free the `elem` when it owns. 1320031b1abaSHawkins Jiawei */ 1321031b1abaSHawkins Jiawei if (dev_written >= 0) { 1322bd907ae4SEugenio Pérez g_free(elem); 1323031b1abaSHawkins Jiawei } 1324be4278b6SEugenio Pérez return dev_written < 0 ? dev_written : 0; 1325bd907ae4SEugenio Pérez } 1326bd907ae4SEugenio Pérez 1327bd907ae4SEugenio Pérez static const VhostShadowVirtqueueOps vhost_vdpa_net_svq_ops = { 1328bd907ae4SEugenio Pérez .avail_handler = vhost_vdpa_net_handle_ctrl_avail, 1329bd907ae4SEugenio Pérez }; 1330bd907ae4SEugenio Pérez 1331152128d6SEugenio Pérez /** 1332152128d6SEugenio Pérez * Probe if CVQ is isolated 1333152128d6SEugenio Pérez * 1334152128d6SEugenio Pérez * @device_fd The vdpa device fd 1335152128d6SEugenio Pérez * @features Features offered by the device. 1336152128d6SEugenio Pérez * @cvq_index The control vq pair index 1337152128d6SEugenio Pérez * 1338152128d6SEugenio Pérez * Returns <0 in case of failure, 0 if false and 1 if true. 1339152128d6SEugenio Pérez */ 1340152128d6SEugenio Pérez static int vhost_vdpa_probe_cvq_isolation(int device_fd, uint64_t features, 1341152128d6SEugenio Pérez int cvq_index, Error **errp) 1342152128d6SEugenio Pérez { 1343152128d6SEugenio Pérez uint64_t backend_features; 1344152128d6SEugenio Pérez int64_t cvq_group; 1345152128d6SEugenio Pérez uint8_t status = VIRTIO_CONFIG_S_ACKNOWLEDGE | 1346152128d6SEugenio Pérez VIRTIO_CONFIG_S_DRIVER | 1347152128d6SEugenio Pérez VIRTIO_CONFIG_S_FEATURES_OK; 1348152128d6SEugenio Pérez int r; 1349152128d6SEugenio Pérez 1350152128d6SEugenio Pérez ERRP_GUARD(); 1351152128d6SEugenio Pérez 1352152128d6SEugenio Pérez r = ioctl(device_fd, VHOST_GET_BACKEND_FEATURES, &backend_features); 1353152128d6SEugenio Pérez if (unlikely(r < 0)) { 1354152128d6SEugenio Pérez error_setg_errno(errp, errno, "Cannot get vdpa backend_features"); 1355152128d6SEugenio Pérez return r; 1356152128d6SEugenio Pérez } 1357152128d6SEugenio Pérez 1358152128d6SEugenio Pérez if (!(backend_features & BIT_ULL(VHOST_BACKEND_F_IOTLB_ASID))) { 1359152128d6SEugenio Pérez return 0; 1360152128d6SEugenio Pérez } 1361152128d6SEugenio Pérez 1362152128d6SEugenio Pérez r = ioctl(device_fd, VHOST_SET_FEATURES, &features); 1363152128d6SEugenio Pérez if (unlikely(r)) { 1364152128d6SEugenio Pérez error_setg_errno(errp, errno, "Cannot set features"); 1365152128d6SEugenio Pérez } 1366152128d6SEugenio Pérez 1367152128d6SEugenio Pérez r = ioctl(device_fd, VHOST_VDPA_SET_STATUS, &status); 1368152128d6SEugenio Pérez if (unlikely(r)) { 1369152128d6SEugenio Pérez error_setg_errno(errp, -r, "Cannot set device features"); 1370152128d6SEugenio Pérez goto out; 1371152128d6SEugenio Pérez } 1372152128d6SEugenio Pérez 1373152128d6SEugenio Pérez cvq_group = vhost_vdpa_get_vring_group(device_fd, cvq_index, errp); 1374152128d6SEugenio Pérez if (unlikely(cvq_group < 0)) { 1375152128d6SEugenio Pérez if (cvq_group != -ENOTSUP) { 1376152128d6SEugenio Pérez r = cvq_group; 1377152128d6SEugenio Pérez goto out; 1378152128d6SEugenio Pérez } 1379152128d6SEugenio Pérez 1380152128d6SEugenio Pérez /* 1381152128d6SEugenio Pérez * The kernel report VHOST_BACKEND_F_IOTLB_ASID if the vdpa frontend 1382152128d6SEugenio Pérez * support ASID even if the parent driver does not. The CVQ cannot be 1383152128d6SEugenio Pérez * isolated in this case. 1384152128d6SEugenio Pérez */ 1385152128d6SEugenio Pérez error_free(*errp); 1386152128d6SEugenio Pérez *errp = NULL; 1387152128d6SEugenio Pérez r = 0; 1388152128d6SEugenio Pérez goto out; 1389152128d6SEugenio Pérez } 1390152128d6SEugenio Pérez 1391152128d6SEugenio Pérez for (int i = 0; i < cvq_index; ++i) { 1392152128d6SEugenio Pérez int64_t group = vhost_vdpa_get_vring_group(device_fd, i, errp); 1393152128d6SEugenio Pérez if (unlikely(group < 0)) { 1394152128d6SEugenio Pérez r = group; 1395152128d6SEugenio Pérez goto out; 1396152128d6SEugenio Pérez } 1397152128d6SEugenio Pérez 1398152128d6SEugenio Pérez if (group == (int64_t)cvq_group) { 1399152128d6SEugenio Pérez r = 0; 1400152128d6SEugenio Pérez goto out; 1401152128d6SEugenio Pérez } 1402152128d6SEugenio Pérez } 1403152128d6SEugenio Pérez 1404152128d6SEugenio Pérez r = 1; 1405152128d6SEugenio Pérez 1406152128d6SEugenio Pérez out: 1407152128d6SEugenio Pérez status = 0; 1408152128d6SEugenio Pérez ioctl(device_fd, VHOST_VDPA_SET_STATUS, &status); 1409152128d6SEugenio Pérez return r; 1410152128d6SEugenio Pérez } 1411152128d6SEugenio Pérez 1412654790b6SJason Wang static NetClientState *net_vhost_vdpa_init(NetClientState *peer, 1413654790b6SJason Wang const char *device, 1414654790b6SJason Wang const char *name, 141540237840SJason Wang int vdpa_device_fd, 141640237840SJason Wang int queue_pair_index, 141740237840SJason Wang int nvqs, 14181576dbb5SEugenio Pérez bool is_datapath, 14191576dbb5SEugenio Pérez bool svq, 14205c1ebd4cSEugenio Pérez struct vhost_vdpa_iova_range iova_range, 1421152128d6SEugenio Pérez uint64_t features, 1422152128d6SEugenio Pérez Error **errp) 14231e0a84eaSCindy Lu { 14241e0a84eaSCindy Lu NetClientState *nc = NULL; 14251e0a84eaSCindy Lu VhostVDPAState *s; 14261e0a84eaSCindy Lu int ret = 0; 14271e0a84eaSCindy Lu assert(name); 1428152128d6SEugenio Pérez int cvq_isolated; 1429152128d6SEugenio Pérez 143040237840SJason Wang if (is_datapath) { 143140237840SJason Wang nc = qemu_new_net_client(&net_vhost_vdpa_info, peer, device, 143240237840SJason Wang name); 143340237840SJason Wang } else { 1434152128d6SEugenio Pérez cvq_isolated = vhost_vdpa_probe_cvq_isolation(vdpa_device_fd, features, 1435152128d6SEugenio Pérez queue_pair_index * 2, 1436152128d6SEugenio Pérez errp); 1437152128d6SEugenio Pérez if (unlikely(cvq_isolated < 0)) { 1438152128d6SEugenio Pérez return NULL; 1439152128d6SEugenio Pérez } 1440152128d6SEugenio Pérez 1441f8972b56SEugenio Pérez nc = qemu_new_net_control_client(&net_vhost_vdpa_cvq_info, peer, 144240237840SJason Wang device, name); 144340237840SJason Wang } 144453b85d95SLaurent Vivier qemu_set_info_str(nc, TYPE_VHOST_VDPA); 14451e0a84eaSCindy Lu s = DO_UPCAST(VhostVDPAState, nc, nc); 14467327813dSJason Wang 14471e0a84eaSCindy Lu s->vhost_vdpa.device_fd = vdpa_device_fd; 144840237840SJason Wang s->vhost_vdpa.index = queue_pair_index; 14497f211a28SEugenio Pérez s->always_svq = svq; 145069498430SEugenio Pérez s->migration_state.notify = vdpa_net_migration_state_notifier; 14511576dbb5SEugenio Pérez s->vhost_vdpa.shadow_vqs_enabled = svq; 1452a585fad2SEugenio Pérez s->vhost_vdpa.iova_range = iova_range; 14536188d78aSEugenio Pérez s->vhost_vdpa.shadow_data = svq; 14545c1ebd4cSEugenio Pérez if (queue_pair_index == 0) { 14555c1ebd4cSEugenio Pérez vhost_vdpa_net_valid_svq_features(features, 14565c1ebd4cSEugenio Pérez &s->vhost_vdpa.migration_blocker); 14575c1ebd4cSEugenio Pérez } else if (!is_datapath) { 1458babf8b87SEugenio Pérez s->cvq_cmd_out_buffer = mmap(NULL, vhost_vdpa_net_cvq_cmd_page_len(), 1459babf8b87SEugenio Pérez PROT_READ | PROT_WRITE, 1460babf8b87SEugenio Pérez MAP_SHARED | MAP_ANONYMOUS, -1, 0); 1461babf8b87SEugenio Pérez s->status = mmap(NULL, vhost_vdpa_net_cvq_cmd_page_len(), 1462babf8b87SEugenio Pérez PROT_READ | PROT_WRITE, MAP_SHARED | MAP_ANONYMOUS, 1463babf8b87SEugenio Pérez -1, 0); 14642df4dd31SEugenio Pérez 1465bd907ae4SEugenio Pérez s->vhost_vdpa.shadow_vq_ops = &vhost_vdpa_net_svq_ops; 1466bd907ae4SEugenio Pérez s->vhost_vdpa.shadow_vq_ops_opaque = s; 1467152128d6SEugenio Pérez s->cvq_isolated = cvq_isolated; 14688bc0049eSEugenio Pérez } 146940237840SJason Wang ret = vhost_vdpa_add(nc, (void *)&s->vhost_vdpa, queue_pair_index, nvqs); 147074af5eecSJason Wang if (ret) { 147174af5eecSJason Wang qemu_del_net_client(nc); 1472654790b6SJason Wang return NULL; 147374af5eecSJason Wang } 1474654790b6SJason Wang return nc; 14751e0a84eaSCindy Lu } 14761e0a84eaSCindy Lu 14778170ab3fSEugenio Pérez static int vhost_vdpa_get_features(int fd, uint64_t *features, Error **errp) 14788170ab3fSEugenio Pérez { 14798170ab3fSEugenio Pérez int ret = ioctl(fd, VHOST_GET_FEATURES, features); 14808170ab3fSEugenio Pérez if (unlikely(ret < 0)) { 14818170ab3fSEugenio Pérez error_setg_errno(errp, errno, 14828170ab3fSEugenio Pérez "Fail to query features from vhost-vDPA device"); 14838170ab3fSEugenio Pérez } 14848170ab3fSEugenio Pérez return ret; 14858170ab3fSEugenio Pérez } 14868170ab3fSEugenio Pérez 14878170ab3fSEugenio Pérez static int vhost_vdpa_get_max_queue_pairs(int fd, uint64_t features, 14888170ab3fSEugenio Pérez int *has_cvq, Error **errp) 148940237840SJason Wang { 149040237840SJason Wang unsigned long config_size = offsetof(struct vhost_vdpa_config, buf); 1491cd523a41SStefano Garzarella g_autofree struct vhost_vdpa_config *config = NULL; 149240237840SJason Wang __virtio16 *max_queue_pairs; 149340237840SJason Wang int ret; 149440237840SJason Wang 149540237840SJason Wang if (features & (1 << VIRTIO_NET_F_CTRL_VQ)) { 149640237840SJason Wang *has_cvq = 1; 149740237840SJason Wang } else { 149840237840SJason Wang *has_cvq = 0; 149940237840SJason Wang } 150040237840SJason Wang 150140237840SJason Wang if (features & (1 << VIRTIO_NET_F_MQ)) { 150240237840SJason Wang config = g_malloc0(config_size + sizeof(*max_queue_pairs)); 150340237840SJason Wang config->off = offsetof(struct virtio_net_config, max_virtqueue_pairs); 150440237840SJason Wang config->len = sizeof(*max_queue_pairs); 150540237840SJason Wang 150640237840SJason Wang ret = ioctl(fd, VHOST_VDPA_GET_CONFIG, config); 150740237840SJason Wang if (ret) { 150840237840SJason Wang error_setg(errp, "Fail to get config from vhost-vDPA device"); 150940237840SJason Wang return -ret; 151040237840SJason Wang } 151140237840SJason Wang 151240237840SJason Wang max_queue_pairs = (__virtio16 *)&config->buf; 151340237840SJason Wang 151440237840SJason Wang return lduw_le_p(max_queue_pairs); 151540237840SJason Wang } 151640237840SJason Wang 151740237840SJason Wang return 1; 151840237840SJason Wang } 151940237840SJason Wang 15201e0a84eaSCindy Lu int net_init_vhost_vdpa(const Netdev *netdev, const char *name, 15211e0a84eaSCindy Lu NetClientState *peer, Error **errp) 15221e0a84eaSCindy Lu { 15231e0a84eaSCindy Lu const NetdevVhostVDPAOptions *opts; 15248170ab3fSEugenio Pérez uint64_t features; 1525654790b6SJason Wang int vdpa_device_fd; 1526eb3cb751SEugenio Pérez g_autofree NetClientState **ncs = NULL; 1527a585fad2SEugenio Pérez struct vhost_vdpa_iova_range iova_range; 1528eb3cb751SEugenio Pérez NetClientState *nc; 1529aed5da45SEugenio Pérez int queue_pairs, r, i = 0, has_cvq = 0; 15301e0a84eaSCindy Lu 15311e0a84eaSCindy Lu assert(netdev->type == NET_CLIENT_DRIVER_VHOST_VDPA); 15321e0a84eaSCindy Lu opts = &netdev->u.vhost_vdpa; 15337480874aSMarkus Armbruster if (!opts->vhostdev && !opts->vhostfd) { 15348801ccd0SSi-Wei Liu error_setg(errp, 15358801ccd0SSi-Wei Liu "vhost-vdpa: neither vhostdev= nor vhostfd= was specified"); 1536c8295404SEugenio Pérez return -1; 1537c8295404SEugenio Pérez } 15387327813dSJason Wang 15397480874aSMarkus Armbruster if (opts->vhostdev && opts->vhostfd) { 15408801ccd0SSi-Wei Liu error_setg(errp, 15418801ccd0SSi-Wei Liu "vhost-vdpa: vhostdev= and vhostfd= are mutually exclusive"); 15428801ccd0SSi-Wei Liu return -1; 15438801ccd0SSi-Wei Liu } 15448801ccd0SSi-Wei Liu 15457480874aSMarkus Armbruster if (opts->vhostdev) { 15460351152bSEugenio Pérez vdpa_device_fd = qemu_open(opts->vhostdev, O_RDWR, errp); 15477327813dSJason Wang if (vdpa_device_fd == -1) { 15487327813dSJason Wang return -errno; 15497327813dSJason Wang } 15505107fd3eSPeter Maydell } else { 15515107fd3eSPeter Maydell /* has_vhostfd */ 15528801ccd0SSi-Wei Liu vdpa_device_fd = monitor_fd_param(monitor_cur(), opts->vhostfd, errp); 15538801ccd0SSi-Wei Liu if (vdpa_device_fd == -1) { 15548801ccd0SSi-Wei Liu error_prepend(errp, "vhost-vdpa: unable to parse vhostfd: "); 15558801ccd0SSi-Wei Liu return -1; 15568801ccd0SSi-Wei Liu } 15578801ccd0SSi-Wei Liu } 15587327813dSJason Wang 15598170ab3fSEugenio Pérez r = vhost_vdpa_get_features(vdpa_device_fd, &features, errp); 15608170ab3fSEugenio Pérez if (unlikely(r < 0)) { 1561aed5da45SEugenio Pérez goto err; 15628170ab3fSEugenio Pérez } 15638170ab3fSEugenio Pérez 15648170ab3fSEugenio Pérez queue_pairs = vhost_vdpa_get_max_queue_pairs(vdpa_device_fd, features, 156540237840SJason Wang &has_cvq, errp); 156640237840SJason Wang if (queue_pairs < 0) { 15677327813dSJason Wang qemu_close(vdpa_device_fd); 156840237840SJason Wang return queue_pairs; 15697327813dSJason Wang } 15707327813dSJason Wang 1571bf7a2ad8SLongpeng r = vhost_vdpa_get_iova_range(vdpa_device_fd, &iova_range); 1572bf7a2ad8SLongpeng if (unlikely(r < 0)) { 1573bf7a2ad8SLongpeng error_setg(errp, "vhost-vdpa: get iova range failed: %s", 1574bf7a2ad8SLongpeng strerror(-r)); 1575bf7a2ad8SLongpeng goto err; 1576bf7a2ad8SLongpeng } 1577bf7a2ad8SLongpeng 157800ef422eSEugenio Pérez if (opts->x_svq && !vhost_vdpa_net_valid_svq_features(features, errp)) { 157900ef422eSEugenio Pérez goto err; 15801576dbb5SEugenio Pérez } 15811576dbb5SEugenio Pérez 158240237840SJason Wang ncs = g_malloc0(sizeof(*ncs) * queue_pairs); 158340237840SJason Wang 158440237840SJason Wang for (i = 0; i < queue_pairs; i++) { 158540237840SJason Wang ncs[i] = net_vhost_vdpa_init(peer, TYPE_VHOST_VDPA, name, 15861576dbb5SEugenio Pérez vdpa_device_fd, i, 2, true, opts->x_svq, 1587152128d6SEugenio Pérez iova_range, features, errp); 158840237840SJason Wang if (!ncs[i]) 158940237840SJason Wang goto err; 159040237840SJason Wang } 159140237840SJason Wang 159240237840SJason Wang if (has_cvq) { 159340237840SJason Wang nc = net_vhost_vdpa_init(peer, TYPE_VHOST_VDPA, name, 15941576dbb5SEugenio Pérez vdpa_device_fd, i, 1, false, 1595152128d6SEugenio Pérez opts->x_svq, iova_range, features, errp); 159640237840SJason Wang if (!nc) 159740237840SJason Wang goto err; 159840237840SJason Wang } 159940237840SJason Wang 1600654790b6SJason Wang return 0; 160140237840SJason Wang 160240237840SJason Wang err: 160340237840SJason Wang if (i) { 16049bd05507SSi-Wei Liu for (i--; i >= 0; i--) { 16059bd05507SSi-Wei Liu qemu_del_net_client(ncs[i]); 16069bd05507SSi-Wei Liu } 160740237840SJason Wang } 16081576dbb5SEugenio Pérez 160940237840SJason Wang qemu_close(vdpa_device_fd); 161040237840SJason Wang 161140237840SJason Wang return -1; 16121e0a84eaSCindy Lu } 1613