15f94c174SLaurent Vivier /* 25f94c174SLaurent Vivier * KVM coalesced MMIO 35f94c174SLaurent Vivier * 45f94c174SLaurent Vivier * Copyright (c) 2008 Bull S.A.S. 55f94c174SLaurent Vivier * 65f94c174SLaurent Vivier * Author: Laurent Vivier <Laurent.Vivier@bull.net> 75f94c174SLaurent Vivier * 85f94c174SLaurent Vivier */ 95f94c174SLaurent Vivier 105f94c174SLaurent Vivier #include "iodev.h" 115f94c174SLaurent Vivier 125f94c174SLaurent Vivier #include <linux/kvm_host.h> 135f94c174SLaurent Vivier #include <linux/kvm.h> 145f94c174SLaurent Vivier 155f94c174SLaurent Vivier #include "coalesced_mmio.h" 165f94c174SLaurent Vivier 17d76685c4SGregory Haskins static inline struct kvm_coalesced_mmio_dev *to_mmio(struct kvm_io_device *dev) 18d76685c4SGregory Haskins { 19d76685c4SGregory Haskins return container_of(dev, struct kvm_coalesced_mmio_dev, dev); 20d76685c4SGregory Haskins } 21d76685c4SGregory Haskins 22*bda9020eSMichael S. Tsirkin static int coalesced_mmio_in_range(struct kvm_coalesced_mmio_dev *dev, 23*bda9020eSMichael S. Tsirkin gpa_t addr, int len) 245f94c174SLaurent Vivier { 255f94c174SLaurent Vivier struct kvm_coalesced_mmio_zone *zone; 26105f8d40SAvi Kivity struct kvm_coalesced_mmio_ring *ring; 27105f8d40SAvi Kivity unsigned avail; 285f94c174SLaurent Vivier int i; 295f94c174SLaurent Vivier 305f94c174SLaurent Vivier /* Are we able to batch it ? */ 315f94c174SLaurent Vivier 325f94c174SLaurent Vivier /* last is the first free entry 335f94c174SLaurent Vivier * check if we don't meet the first used entry 345f94c174SLaurent Vivier * there is always one unused entry in the buffer 355f94c174SLaurent Vivier */ 36105f8d40SAvi Kivity ring = dev->kvm->coalesced_mmio_ring; 37105f8d40SAvi Kivity avail = (ring->first - ring->last - 1) % KVM_COALESCED_MMIO_MAX; 3864a2268dSMarcelo Tosatti if (avail < KVM_MAX_VCPUS) { 395f94c174SLaurent Vivier /* full */ 405f94c174SLaurent Vivier return 0; 415f94c174SLaurent Vivier } 425f94c174SLaurent Vivier 435f94c174SLaurent Vivier /* is it in a batchable area ? */ 445f94c174SLaurent Vivier 455f94c174SLaurent Vivier for (i = 0; i < dev->nb_zones; i++) { 465f94c174SLaurent Vivier zone = &dev->zone[i]; 475f94c174SLaurent Vivier 485f94c174SLaurent Vivier /* (addr,len) is fully included in 495f94c174SLaurent Vivier * (zone->addr, zone->size) 505f94c174SLaurent Vivier */ 515f94c174SLaurent Vivier 525f94c174SLaurent Vivier if (zone->addr <= addr && 535f94c174SLaurent Vivier addr + len <= zone->addr + zone->size) 545f94c174SLaurent Vivier return 1; 555f94c174SLaurent Vivier } 565f94c174SLaurent Vivier return 0; 575f94c174SLaurent Vivier } 585f94c174SLaurent Vivier 59*bda9020eSMichael S. Tsirkin static int coalesced_mmio_write(struct kvm_io_device *this, 605f94c174SLaurent Vivier gpa_t addr, int len, const void *val) 615f94c174SLaurent Vivier { 62d76685c4SGregory Haskins struct kvm_coalesced_mmio_dev *dev = to_mmio(this); 635f94c174SLaurent Vivier struct kvm_coalesced_mmio_ring *ring = dev->kvm->coalesced_mmio_ring; 64*bda9020eSMichael S. Tsirkin if (!coalesced_mmio_in_range(dev, addr, len)) 65*bda9020eSMichael S. Tsirkin return -EOPNOTSUPP; 665f94c174SLaurent Vivier 6764a2268dSMarcelo Tosatti spin_lock(&dev->lock); 685f94c174SLaurent Vivier 695f94c174SLaurent Vivier /* copy data in first free entry of the ring */ 705f94c174SLaurent Vivier 715f94c174SLaurent Vivier ring->coalesced_mmio[ring->last].phys_addr = addr; 725f94c174SLaurent Vivier ring->coalesced_mmio[ring->last].len = len; 735f94c174SLaurent Vivier memcpy(ring->coalesced_mmio[ring->last].data, val, len); 745f94c174SLaurent Vivier smp_wmb(); 755f94c174SLaurent Vivier ring->last = (ring->last + 1) % KVM_COALESCED_MMIO_MAX; 7664a2268dSMarcelo Tosatti spin_unlock(&dev->lock); 77*bda9020eSMichael S. Tsirkin return 0; 785f94c174SLaurent Vivier } 795f94c174SLaurent Vivier 805f94c174SLaurent Vivier static void coalesced_mmio_destructor(struct kvm_io_device *this) 815f94c174SLaurent Vivier { 82d76685c4SGregory Haskins struct kvm_coalesced_mmio_dev *dev = to_mmio(this); 83787a660aSGregory Haskins 84787a660aSGregory Haskins kfree(dev); 855f94c174SLaurent Vivier } 865f94c174SLaurent Vivier 87d76685c4SGregory Haskins static const struct kvm_io_device_ops coalesced_mmio_ops = { 88d76685c4SGregory Haskins .write = coalesced_mmio_write, 89d76685c4SGregory Haskins .destructor = coalesced_mmio_destructor, 90d76685c4SGregory Haskins }; 91d76685c4SGregory Haskins 925f94c174SLaurent Vivier int kvm_coalesced_mmio_init(struct kvm *kvm) 935f94c174SLaurent Vivier { 945f94c174SLaurent Vivier struct kvm_coalesced_mmio_dev *dev; 955f94c174SLaurent Vivier 965f94c174SLaurent Vivier dev = kzalloc(sizeof(struct kvm_coalesced_mmio_dev), GFP_KERNEL); 975f94c174SLaurent Vivier if (!dev) 985f94c174SLaurent Vivier return -ENOMEM; 9964a2268dSMarcelo Tosatti spin_lock_init(&dev->lock); 100d76685c4SGregory Haskins kvm_iodevice_init(&dev->dev, &coalesced_mmio_ops); 1015f94c174SLaurent Vivier dev->kvm = kvm; 1025f94c174SLaurent Vivier kvm->coalesced_mmio_dev = dev; 1036c474694SMichael S. Tsirkin kvm_io_bus_register_dev(kvm, &kvm->mmio_bus, &dev->dev); 1045f94c174SLaurent Vivier 1055f94c174SLaurent Vivier return 0; 1065f94c174SLaurent Vivier } 1075f94c174SLaurent Vivier 1085f94c174SLaurent Vivier int kvm_vm_ioctl_register_coalesced_mmio(struct kvm *kvm, 1095f94c174SLaurent Vivier struct kvm_coalesced_mmio_zone *zone) 1105f94c174SLaurent Vivier { 1115f94c174SLaurent Vivier struct kvm_coalesced_mmio_dev *dev = kvm->coalesced_mmio_dev; 1125f94c174SLaurent Vivier 1135f94c174SLaurent Vivier if (dev == NULL) 1145f94c174SLaurent Vivier return -EINVAL; 1155f94c174SLaurent Vivier 116d5c2dcc3SMichael S. Tsirkin down_write(&kvm->slots_lock); 1175f94c174SLaurent Vivier if (dev->nb_zones >= KVM_COALESCED_MMIO_ZONE_MAX) { 118d5c2dcc3SMichael S. Tsirkin up_write(&kvm->slots_lock); 1195f94c174SLaurent Vivier return -ENOBUFS; 1205f94c174SLaurent Vivier } 1215f94c174SLaurent Vivier 1225f94c174SLaurent Vivier dev->zone[dev->nb_zones] = *zone; 1235f94c174SLaurent Vivier dev->nb_zones++; 1245f94c174SLaurent Vivier 125d5c2dcc3SMichael S. Tsirkin up_write(&kvm->slots_lock); 1265f94c174SLaurent Vivier return 0; 1275f94c174SLaurent Vivier } 1285f94c174SLaurent Vivier 1295f94c174SLaurent Vivier int kvm_vm_ioctl_unregister_coalesced_mmio(struct kvm *kvm, 1305f94c174SLaurent Vivier struct kvm_coalesced_mmio_zone *zone) 1315f94c174SLaurent Vivier { 1325f94c174SLaurent Vivier int i; 1335f94c174SLaurent Vivier struct kvm_coalesced_mmio_dev *dev = kvm->coalesced_mmio_dev; 1345f94c174SLaurent Vivier struct kvm_coalesced_mmio_zone *z; 1355f94c174SLaurent Vivier 1365f94c174SLaurent Vivier if (dev == NULL) 1375f94c174SLaurent Vivier return -EINVAL; 1385f94c174SLaurent Vivier 139d5c2dcc3SMichael S. Tsirkin down_write(&kvm->slots_lock); 1405f94c174SLaurent Vivier 1415f94c174SLaurent Vivier i = dev->nb_zones; 1425f94c174SLaurent Vivier while(i) { 1435f94c174SLaurent Vivier z = &dev->zone[i - 1]; 1445f94c174SLaurent Vivier 1455f94c174SLaurent Vivier /* unregister all zones 1465f94c174SLaurent Vivier * included in (zone->addr, zone->size) 1475f94c174SLaurent Vivier */ 1485f94c174SLaurent Vivier 1495f94c174SLaurent Vivier if (zone->addr <= z->addr && 1505f94c174SLaurent Vivier z->addr + z->size <= zone->addr + zone->size) { 1515f94c174SLaurent Vivier dev->nb_zones--; 1525f94c174SLaurent Vivier *z = dev->zone[dev->nb_zones]; 1535f94c174SLaurent Vivier } 1545f94c174SLaurent Vivier i--; 1555f94c174SLaurent Vivier } 1565f94c174SLaurent Vivier 157d5c2dcc3SMichael S. Tsirkin up_write(&kvm->slots_lock); 1585f94c174SLaurent Vivier 1595f94c174SLaurent Vivier return 0; 1605f94c174SLaurent Vivier } 161