1f7116284SIan Campbell /****************************************************************************** 2f7116284SIan Campbell * evtchn.c 3f7116284SIan Campbell * 4f7116284SIan Campbell * Driver for receiving and demuxing event-channel signals. 5f7116284SIan Campbell * 6f7116284SIan Campbell * Copyright (c) 2004-2005, K A Fraser 7f7116284SIan Campbell * Multi-process extensions Copyright (c) 2004, Steven Smith 8f7116284SIan Campbell * 9f7116284SIan Campbell * This program is free software; you can redistribute it and/or 10f7116284SIan Campbell * modify it under the terms of the GNU General Public License version 2 11f7116284SIan Campbell * as published by the Free Software Foundation; or, when distributed 12f7116284SIan Campbell * separately from the Linux kernel or incorporated into other 13f7116284SIan Campbell * software packages, subject to the following license: 14f7116284SIan Campbell * 15f7116284SIan Campbell * Permission is hereby granted, free of charge, to any person obtaining a copy 16f7116284SIan Campbell * of this source file (the "Software"), to deal in the Software without 17f7116284SIan Campbell * restriction, including without limitation the rights to use, copy, modify, 18f7116284SIan Campbell * merge, publish, distribute, sublicense, and/or sell copies of the Software, 19f7116284SIan Campbell * and to permit persons to whom the Software is furnished to do so, subject to 20f7116284SIan Campbell * the following conditions: 21f7116284SIan Campbell * 22f7116284SIan Campbell * The above copyright notice and this permission notice shall be included in 23f7116284SIan Campbell * all copies or substantial portions of the Software. 24f7116284SIan Campbell * 25f7116284SIan Campbell * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 26f7116284SIan Campbell * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 27f7116284SIan Campbell * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE 28f7116284SIan Campbell * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER 29f7116284SIan Campbell * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING 30f7116284SIan Campbell * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS 31f7116284SIan Campbell * IN THE SOFTWARE. 32f7116284SIan Campbell */ 33f7116284SIan Campbell 34283c0972SJoe Perches #define pr_fmt(fmt) "xen:" KBUILD_MODNAME ": " fmt 35283c0972SJoe Perches 36f7116284SIan Campbell #include <linux/module.h> 37f7116284SIan Campbell #include <linux/kernel.h> 38f7116284SIan Campbell #include <linux/sched.h> 39f7116284SIan Campbell #include <linux/slab.h> 40f7116284SIan Campbell #include <linux/string.h> 41f7116284SIan Campbell #include <linux/errno.h> 42f7116284SIan Campbell #include <linux/fs.h> 43f7116284SIan Campbell #include <linux/miscdevice.h> 44f7116284SIan Campbell #include <linux/major.h> 45f7116284SIan Campbell #include <linux/proc_fs.h> 46f7116284SIan Campbell #include <linux/stat.h> 47f7116284SIan Campbell #include <linux/poll.h> 48f7116284SIan Campbell #include <linux/irq.h> 49f7116284SIan Campbell #include <linux/init.h> 50f7116284SIan Campbell #include <linux/mutex.h> 51f7116284SIan Campbell #include <linux/cpu.h> 52*86200154SDavid Vrabel #include <linux/mm.h> 53*86200154SDavid Vrabel #include <linux/vmalloc.h> 541ccbf534SJeremy Fitzhardinge 551ccbf534SJeremy Fitzhardinge #include <xen/xen.h> 56f7116284SIan Campbell #include <xen/events.h> 57f7116284SIan Campbell #include <xen/evtchn.h> 58f7116284SIan Campbell #include <asm/xen/hypervisor.h> 59f7116284SIan Campbell 60f7116284SIan Campbell struct per_user_data { 610a4666b5SJeremy Fitzhardinge struct mutex bind_mutex; /* serialize bind/unbind operations */ 6273cc4bb0SDavid Vrabel struct rb_root evtchns; 63*86200154SDavid Vrabel unsigned int nr_evtchns; 640a4666b5SJeremy Fitzhardinge 65f7116284SIan Campbell /* Notification ring, accessed via /dev/xen/evtchn. */ 66*86200154SDavid Vrabel unsigned int ring_size; 67f7116284SIan Campbell evtchn_port_t *ring; 68f7116284SIan Campbell unsigned int ring_cons, ring_prod, ring_overflow; 69f7116284SIan Campbell struct mutex ring_cons_mutex; /* protect against concurrent readers */ 7073cc4bb0SDavid Vrabel spinlock_t ring_prod_lock; /* product against concurrent interrupts */ 71f7116284SIan Campbell 72f7116284SIan Campbell /* Processes wait on this queue when ring is empty. */ 73f7116284SIan Campbell wait_queue_head_t evtchn_wait; 74f7116284SIan Campbell struct fasync_struct *evtchn_async_queue; 75f7116284SIan Campbell const char *name; 76f7116284SIan Campbell }; 77f7116284SIan Campbell 7873cc4bb0SDavid Vrabel struct user_evtchn { 7973cc4bb0SDavid Vrabel struct rb_node node; 8073cc4bb0SDavid Vrabel struct per_user_data *user; 8173cc4bb0SDavid Vrabel unsigned port; 8273cc4bb0SDavid Vrabel bool enabled; 8373cc4bb0SDavid Vrabel }; 84f7116284SIan Campbell 85*86200154SDavid Vrabel static evtchn_port_t *evtchn_alloc_ring(unsigned int size) 86*86200154SDavid Vrabel { 87*86200154SDavid Vrabel evtchn_port_t *ring; 88*86200154SDavid Vrabel size_t s = size * sizeof(*ring); 89*86200154SDavid Vrabel 90*86200154SDavid Vrabel ring = kmalloc(s, GFP_KERNEL); 91*86200154SDavid Vrabel if (!ring) 92*86200154SDavid Vrabel ring = vmalloc(s); 93*86200154SDavid Vrabel 94*86200154SDavid Vrabel return ring; 95*86200154SDavid Vrabel } 96*86200154SDavid Vrabel 97*86200154SDavid Vrabel static void evtchn_free_ring(evtchn_port_t *ring) 98*86200154SDavid Vrabel { 99*86200154SDavid Vrabel kvfree(ring); 100*86200154SDavid Vrabel } 101*86200154SDavid Vrabel 102*86200154SDavid Vrabel static unsigned int evtchn_ring_offset(struct per_user_data *u, 103*86200154SDavid Vrabel unsigned int idx) 104*86200154SDavid Vrabel { 105*86200154SDavid Vrabel return idx & (u->ring_size - 1); 106*86200154SDavid Vrabel } 107*86200154SDavid Vrabel 108*86200154SDavid Vrabel static evtchn_port_t *evtchn_ring_entry(struct per_user_data *u, 109*86200154SDavid Vrabel unsigned int idx) 110*86200154SDavid Vrabel { 111*86200154SDavid Vrabel return u->ring + evtchn_ring_offset(u, idx); 112*86200154SDavid Vrabel } 113*86200154SDavid Vrabel 11473cc4bb0SDavid Vrabel static int add_evtchn(struct per_user_data *u, struct user_evtchn *evtchn) 115e3cc067bSJeremy Fitzhardinge { 11673cc4bb0SDavid Vrabel struct rb_node **new = &(u->evtchns.rb_node), *parent = NULL; 117e3cc067bSJeremy Fitzhardinge 118*86200154SDavid Vrabel u->nr_evtchns++; 119*86200154SDavid Vrabel 12073cc4bb0SDavid Vrabel while (*new) { 12173cc4bb0SDavid Vrabel struct user_evtchn *this; 122e3cc067bSJeremy Fitzhardinge 12373cc4bb0SDavid Vrabel this = container_of(*new, struct user_evtchn, node); 124e3cc067bSJeremy Fitzhardinge 12573cc4bb0SDavid Vrabel parent = *new; 12673cc4bb0SDavid Vrabel if (this->port < evtchn->port) 12773cc4bb0SDavid Vrabel new = &((*new)->rb_left); 12873cc4bb0SDavid Vrabel else if (this->port > evtchn->port) 12973cc4bb0SDavid Vrabel new = &((*new)->rb_right); 130e3cc067bSJeremy Fitzhardinge else 13173cc4bb0SDavid Vrabel return -EEXIST; 13273cc4bb0SDavid Vrabel } 13373cc4bb0SDavid Vrabel 13473cc4bb0SDavid Vrabel /* Add new node and rebalance tree. */ 13573cc4bb0SDavid Vrabel rb_link_node(&evtchn->node, parent, new); 13673cc4bb0SDavid Vrabel rb_insert_color(&evtchn->node, &u->evtchns); 13773cc4bb0SDavid Vrabel 13873cc4bb0SDavid Vrabel return 0; 13973cc4bb0SDavid Vrabel } 14073cc4bb0SDavid Vrabel 14173cc4bb0SDavid Vrabel static void del_evtchn(struct per_user_data *u, struct user_evtchn *evtchn) 14273cc4bb0SDavid Vrabel { 143*86200154SDavid Vrabel u->nr_evtchns--; 14473cc4bb0SDavid Vrabel rb_erase(&evtchn->node, &u->evtchns); 14573cc4bb0SDavid Vrabel kfree(evtchn); 14673cc4bb0SDavid Vrabel } 14773cc4bb0SDavid Vrabel 14873cc4bb0SDavid Vrabel static struct user_evtchn *find_evtchn(struct per_user_data *u, unsigned port) 14973cc4bb0SDavid Vrabel { 15073cc4bb0SDavid Vrabel struct rb_node *node = u->evtchns.rb_node; 15173cc4bb0SDavid Vrabel 15273cc4bb0SDavid Vrabel while (node) { 15373cc4bb0SDavid Vrabel struct user_evtchn *evtchn; 15473cc4bb0SDavid Vrabel 15573cc4bb0SDavid Vrabel evtchn = container_of(node, struct user_evtchn, node); 15673cc4bb0SDavid Vrabel 15773cc4bb0SDavid Vrabel if (evtchn->port < port) 15873cc4bb0SDavid Vrabel node = node->rb_left; 15973cc4bb0SDavid Vrabel else if (evtchn->port > port) 16073cc4bb0SDavid Vrabel node = node->rb_right; 16173cc4bb0SDavid Vrabel else 16273cc4bb0SDavid Vrabel return evtchn; 16373cc4bb0SDavid Vrabel } 16473cc4bb0SDavid Vrabel return NULL; 165e3cc067bSJeremy Fitzhardinge } 166e3cc067bSJeremy Fitzhardinge 16770697d54SJeremy Fitzhardinge static irqreturn_t evtchn_interrupt(int irq, void *data) 168f7116284SIan Campbell { 16973cc4bb0SDavid Vrabel struct user_evtchn *evtchn = data; 17073cc4bb0SDavid Vrabel struct per_user_data *u = evtchn->user; 171f7116284SIan Campbell 17273cc4bb0SDavid Vrabel WARN(!evtchn->enabled, 173e3cc067bSJeremy Fitzhardinge "Interrupt for port %d, but apparently not enabled; per-user %p\n", 17473cc4bb0SDavid Vrabel evtchn->port, u); 175f7116284SIan Campbell 176f7116284SIan Campbell disable_irq_nosync(irq); 17773cc4bb0SDavid Vrabel evtchn->enabled = false; 17873cc4bb0SDavid Vrabel 17973cc4bb0SDavid Vrabel spin_lock(&u->ring_prod_lock); 180f7116284SIan Campbell 181*86200154SDavid Vrabel if ((u->ring_prod - u->ring_cons) < u->ring_size) { 182*86200154SDavid Vrabel *evtchn_ring_entry(u, u->ring_prod) = evtchn->port; 183f7116284SIan Campbell wmb(); /* Ensure ring contents visible */ 184f7116284SIan Campbell if (u->ring_cons == u->ring_prod++) { 185f7116284SIan Campbell wake_up_interruptible(&u->evtchn_wait); 186f7116284SIan Campbell kill_fasync(&u->evtchn_async_queue, 187f7116284SIan Campbell SIGIO, POLL_IN); 188f7116284SIan Campbell } 189e3cc067bSJeremy Fitzhardinge } else 190f7116284SIan Campbell u->ring_overflow = 1; 191f7116284SIan Campbell 19273cc4bb0SDavid Vrabel spin_unlock(&u->ring_prod_lock); 193f7116284SIan Campbell 194f7116284SIan Campbell return IRQ_HANDLED; 195f7116284SIan Campbell } 196f7116284SIan Campbell 197f7116284SIan Campbell static ssize_t evtchn_read(struct file *file, char __user *buf, 198f7116284SIan Campbell size_t count, loff_t *ppos) 199f7116284SIan Campbell { 200f7116284SIan Campbell int rc; 201f7116284SIan Campbell unsigned int c, p, bytes1 = 0, bytes2 = 0; 202f7116284SIan Campbell struct per_user_data *u = file->private_data; 203f7116284SIan Campbell 204f7116284SIan Campbell /* Whole number of ports. */ 205f7116284SIan Campbell count &= ~(sizeof(evtchn_port_t)-1); 206f7116284SIan Campbell 207f7116284SIan Campbell if (count == 0) 208f7116284SIan Campbell return 0; 209f7116284SIan Campbell 210f7116284SIan Campbell if (count > PAGE_SIZE) 211f7116284SIan Campbell count = PAGE_SIZE; 212f7116284SIan Campbell 213f7116284SIan Campbell for (;;) { 214f7116284SIan Campbell mutex_lock(&u->ring_cons_mutex); 215f7116284SIan Campbell 216f7116284SIan Campbell rc = -EFBIG; 217f7116284SIan Campbell if (u->ring_overflow) 218f7116284SIan Campbell goto unlock_out; 219f7116284SIan Campbell 220f7116284SIan Campbell c = u->ring_cons; 221f7116284SIan Campbell p = u->ring_prod; 222f7116284SIan Campbell if (c != p) 223f7116284SIan Campbell break; 224f7116284SIan Campbell 225f7116284SIan Campbell mutex_unlock(&u->ring_cons_mutex); 226f7116284SIan Campbell 227f7116284SIan Campbell if (file->f_flags & O_NONBLOCK) 228f7116284SIan Campbell return -EAGAIN; 229f7116284SIan Campbell 230f7116284SIan Campbell rc = wait_event_interruptible(u->evtchn_wait, 231f7116284SIan Campbell u->ring_cons != u->ring_prod); 232f7116284SIan Campbell if (rc) 233f7116284SIan Campbell return rc; 234f7116284SIan Campbell } 235f7116284SIan Campbell 236f7116284SIan Campbell /* Byte lengths of two chunks. Chunk split (if any) is at ring wrap. */ 237*86200154SDavid Vrabel if (((c ^ p) & u->ring_size) != 0) { 238*86200154SDavid Vrabel bytes1 = (u->ring_size - evtchn_ring_offset(u, c)) * 239f7116284SIan Campbell sizeof(evtchn_port_t); 240*86200154SDavid Vrabel bytes2 = evtchn_ring_offset(u, p) * sizeof(evtchn_port_t); 241f7116284SIan Campbell } else { 242f7116284SIan Campbell bytes1 = (p - c) * sizeof(evtchn_port_t); 243f7116284SIan Campbell bytes2 = 0; 244f7116284SIan Campbell } 245f7116284SIan Campbell 246f7116284SIan Campbell /* Truncate chunks according to caller's maximum byte count. */ 247f7116284SIan Campbell if (bytes1 > count) { 248f7116284SIan Campbell bytes1 = count; 249f7116284SIan Campbell bytes2 = 0; 250f7116284SIan Campbell } else if ((bytes1 + bytes2) > count) { 251f7116284SIan Campbell bytes2 = count - bytes1; 252f7116284SIan Campbell } 253f7116284SIan Campbell 254f7116284SIan Campbell rc = -EFAULT; 255f7116284SIan Campbell rmb(); /* Ensure that we see the port before we copy it. */ 256*86200154SDavid Vrabel if (copy_to_user(buf, evtchn_ring_entry(u, c), bytes1) || 257f7116284SIan Campbell ((bytes2 != 0) && 258f7116284SIan Campbell copy_to_user(&buf[bytes1], &u->ring[0], bytes2))) 259f7116284SIan Campbell goto unlock_out; 260f7116284SIan Campbell 261f7116284SIan Campbell u->ring_cons += (bytes1 + bytes2) / sizeof(evtchn_port_t); 262f7116284SIan Campbell rc = bytes1 + bytes2; 263f7116284SIan Campbell 264f7116284SIan Campbell unlock_out: 265f7116284SIan Campbell mutex_unlock(&u->ring_cons_mutex); 266f7116284SIan Campbell return rc; 267f7116284SIan Campbell } 268f7116284SIan Campbell 269f7116284SIan Campbell static ssize_t evtchn_write(struct file *file, const char __user *buf, 270f7116284SIan Campbell size_t count, loff_t *ppos) 271f7116284SIan Campbell { 272f7116284SIan Campbell int rc, i; 273f7116284SIan Campbell evtchn_port_t *kbuf = (evtchn_port_t *)__get_free_page(GFP_KERNEL); 274f7116284SIan Campbell struct per_user_data *u = file->private_data; 275f7116284SIan Campbell 276f7116284SIan Campbell if (kbuf == NULL) 277f7116284SIan Campbell return -ENOMEM; 278f7116284SIan Campbell 279f7116284SIan Campbell /* Whole number of ports. */ 280f7116284SIan Campbell count &= ~(sizeof(evtchn_port_t)-1); 281f7116284SIan Campbell 282f7116284SIan Campbell rc = 0; 283f7116284SIan Campbell if (count == 0) 284f7116284SIan Campbell goto out; 285f7116284SIan Campbell 286f7116284SIan Campbell if (count > PAGE_SIZE) 287f7116284SIan Campbell count = PAGE_SIZE; 288f7116284SIan Campbell 289f7116284SIan Campbell rc = -EFAULT; 290f7116284SIan Campbell if (copy_from_user(kbuf, buf, count) != 0) 291f7116284SIan Campbell goto out; 292f7116284SIan Campbell 29373cc4bb0SDavid Vrabel mutex_lock(&u->bind_mutex); 294e3cc067bSJeremy Fitzhardinge 295e3cc067bSJeremy Fitzhardinge for (i = 0; i < (count/sizeof(evtchn_port_t)); i++) { 296e3cc067bSJeremy Fitzhardinge unsigned port = kbuf[i]; 29773cc4bb0SDavid Vrabel struct user_evtchn *evtchn; 298e3cc067bSJeremy Fitzhardinge 29973cc4bb0SDavid Vrabel evtchn = find_evtchn(u, port); 30073cc4bb0SDavid Vrabel if (evtchn && !evtchn->enabled) { 30173cc4bb0SDavid Vrabel evtchn->enabled = true; 302e3cc067bSJeremy Fitzhardinge enable_irq(irq_from_evtchn(port)); 303e3cc067bSJeremy Fitzhardinge } 304e3cc067bSJeremy Fitzhardinge } 305e3cc067bSJeremy Fitzhardinge 30673cc4bb0SDavid Vrabel mutex_unlock(&u->bind_mutex); 307f7116284SIan Campbell 308f7116284SIan Campbell rc = count; 309f7116284SIan Campbell 310f7116284SIan Campbell out: 311f7116284SIan Campbell free_page((unsigned long)kbuf); 312f7116284SIan Campbell return rc; 313f7116284SIan Campbell } 314f7116284SIan Campbell 315*86200154SDavid Vrabel static int evtchn_resize_ring(struct per_user_data *u) 316*86200154SDavid Vrabel { 317*86200154SDavid Vrabel unsigned int new_size; 318*86200154SDavid Vrabel evtchn_port_t *new_ring, *old_ring; 319*86200154SDavid Vrabel unsigned int p, c; 320*86200154SDavid Vrabel 321*86200154SDavid Vrabel /* 322*86200154SDavid Vrabel * Ensure the ring is large enough to capture all possible 323*86200154SDavid Vrabel * events. i.e., one free slot for each bound event. 324*86200154SDavid Vrabel */ 325*86200154SDavid Vrabel if (u->nr_evtchns <= u->ring_size) 326*86200154SDavid Vrabel return 0; 327*86200154SDavid Vrabel 328*86200154SDavid Vrabel if (u->ring_size == 0) 329*86200154SDavid Vrabel new_size = 64; 330*86200154SDavid Vrabel else 331*86200154SDavid Vrabel new_size = 2 * u->ring_size; 332*86200154SDavid Vrabel 333*86200154SDavid Vrabel new_ring = evtchn_alloc_ring(new_size); 334*86200154SDavid Vrabel if (!new_ring) 335*86200154SDavid Vrabel return -ENOMEM; 336*86200154SDavid Vrabel 337*86200154SDavid Vrabel old_ring = u->ring; 338*86200154SDavid Vrabel 339*86200154SDavid Vrabel /* 340*86200154SDavid Vrabel * Access to the ring contents is serialized by either the 341*86200154SDavid Vrabel * prod /or/ cons lock so take both when resizing. 342*86200154SDavid Vrabel */ 343*86200154SDavid Vrabel mutex_lock(&u->ring_cons_mutex); 344*86200154SDavid Vrabel spin_lock_irq(&u->ring_prod_lock); 345*86200154SDavid Vrabel 346*86200154SDavid Vrabel /* 347*86200154SDavid Vrabel * Copy the old ring contents to the new ring. 348*86200154SDavid Vrabel * 349*86200154SDavid Vrabel * If the ring contents crosses the end of the current ring, 350*86200154SDavid Vrabel * it needs to be copied in two chunks. 351*86200154SDavid Vrabel * 352*86200154SDavid Vrabel * +---------+ +------------------+ 353*86200154SDavid Vrabel * |34567 12| -> | 1234567 | 354*86200154SDavid Vrabel * +-----p-c-+ +------------------+ 355*86200154SDavid Vrabel */ 356*86200154SDavid Vrabel p = evtchn_ring_offset(u, u->ring_prod); 357*86200154SDavid Vrabel c = evtchn_ring_offset(u, u->ring_cons); 358*86200154SDavid Vrabel if (p < c) { 359*86200154SDavid Vrabel memcpy(new_ring + c, u->ring + c, (u->ring_size - c) * sizeof(*u->ring)); 360*86200154SDavid Vrabel memcpy(new_ring + u->ring_size, u->ring, p * sizeof(*u->ring)); 361*86200154SDavid Vrabel } else 362*86200154SDavid Vrabel memcpy(new_ring + c, u->ring + c, (p - c) * sizeof(*u->ring)); 363*86200154SDavid Vrabel 364*86200154SDavid Vrabel u->ring = new_ring; 365*86200154SDavid Vrabel u->ring_size = new_size; 366*86200154SDavid Vrabel 367*86200154SDavid Vrabel spin_unlock_irq(&u->ring_prod_lock); 368*86200154SDavid Vrabel mutex_unlock(&u->ring_cons_mutex); 369*86200154SDavid Vrabel 370*86200154SDavid Vrabel evtchn_free_ring(old_ring); 371*86200154SDavid Vrabel 372*86200154SDavid Vrabel return 0; 373*86200154SDavid Vrabel } 374*86200154SDavid Vrabel 375f7116284SIan Campbell static int evtchn_bind_to_user(struct per_user_data *u, int port) 376f7116284SIan Campbell { 37773cc4bb0SDavid Vrabel struct user_evtchn *evtchn; 37873cc4bb0SDavid Vrabel struct evtchn_close close; 379f7116284SIan Campbell int rc = 0; 380f7116284SIan Campbell 3810a4666b5SJeremy Fitzhardinge /* 3820a4666b5SJeremy Fitzhardinge * Ports are never reused, so every caller should pass in a 3830a4666b5SJeremy Fitzhardinge * unique port. 3840a4666b5SJeremy Fitzhardinge * 3850a4666b5SJeremy Fitzhardinge * (Locking not necessary because we haven't registered the 3860a4666b5SJeremy Fitzhardinge * interrupt handler yet, and our caller has already 3870a4666b5SJeremy Fitzhardinge * serialized bind operations.) 3880a4666b5SJeremy Fitzhardinge */ 38973cc4bb0SDavid Vrabel 39073cc4bb0SDavid Vrabel evtchn = kzalloc(sizeof(*evtchn), GFP_KERNEL); 39173cc4bb0SDavid Vrabel if (!evtchn) 39273cc4bb0SDavid Vrabel return -ENOMEM; 39373cc4bb0SDavid Vrabel 39473cc4bb0SDavid Vrabel evtchn->user = u; 39573cc4bb0SDavid Vrabel evtchn->port = port; 39673cc4bb0SDavid Vrabel evtchn->enabled = true; /* start enabled */ 39773cc4bb0SDavid Vrabel 39873cc4bb0SDavid Vrabel rc = add_evtchn(u, evtchn); 39973cc4bb0SDavid Vrabel if (rc < 0) 40073cc4bb0SDavid Vrabel goto err; 401f7116284SIan Campbell 402*86200154SDavid Vrabel rc = evtchn_resize_ring(u); 403*86200154SDavid Vrabel if (rc < 0) 404*86200154SDavid Vrabel goto err; 405*86200154SDavid Vrabel 406af09d1a7SMichael Opdenacker rc = bind_evtchn_to_irqhandler(port, evtchn_interrupt, 0, 40773cc4bb0SDavid Vrabel u->name, evtchn); 40873cc4bb0SDavid Vrabel if (rc < 0) 40973cc4bb0SDavid Vrabel goto err; 41073cc4bb0SDavid Vrabel 411420eb554SDaniel De Graaf rc = evtchn_make_refcounted(port); 41273cc4bb0SDavid Vrabel return rc; 41373cc4bb0SDavid Vrabel 41473cc4bb0SDavid Vrabel err: 415e7e44e44SWei Liu /* bind failed, should close the port now */ 416e7e44e44SWei Liu close.port = port; 417e7e44e44SWei Liu if (HYPERVISOR_event_channel_op(EVTCHNOP_close, &close) != 0) 418e7e44e44SWei Liu BUG(); 41973cc4bb0SDavid Vrabel del_evtchn(u, evtchn); 420f7116284SIan Campbell return rc; 421f7116284SIan Campbell } 422f7116284SIan Campbell 42373cc4bb0SDavid Vrabel static void evtchn_unbind_from_user(struct per_user_data *u, 42473cc4bb0SDavid Vrabel struct user_evtchn *evtchn) 425f7116284SIan Campbell { 42673cc4bb0SDavid Vrabel int irq = irq_from_evtchn(evtchn->port); 427f7116284SIan Campbell 428e7e44e44SWei Liu BUG_ON(irq < 0); 429e7e44e44SWei Liu 43073cc4bb0SDavid Vrabel unbind_from_irqhandler(irq, evtchn); 4310a4666b5SJeremy Fitzhardinge 43273cc4bb0SDavid Vrabel del_evtchn(u, evtchn); 433f7116284SIan Campbell } 434f7116284SIan Campbell 435f7116284SIan Campbell static long evtchn_ioctl(struct file *file, 436f7116284SIan Campbell unsigned int cmd, unsigned long arg) 437f7116284SIan Campbell { 438f7116284SIan Campbell int rc; 439f7116284SIan Campbell struct per_user_data *u = file->private_data; 440f7116284SIan Campbell void __user *uarg = (void __user *) arg; 441f7116284SIan Campbell 4420a4666b5SJeremy Fitzhardinge /* Prevent bind from racing with unbind */ 4430a4666b5SJeremy Fitzhardinge mutex_lock(&u->bind_mutex); 4440a4666b5SJeremy Fitzhardinge 445f7116284SIan Campbell switch (cmd) { 446f7116284SIan Campbell case IOCTL_EVTCHN_BIND_VIRQ: { 447f7116284SIan Campbell struct ioctl_evtchn_bind_virq bind; 448f7116284SIan Campbell struct evtchn_bind_virq bind_virq; 449f7116284SIan Campbell 450f7116284SIan Campbell rc = -EFAULT; 451f7116284SIan Campbell if (copy_from_user(&bind, uarg, sizeof(bind))) 452f7116284SIan Campbell break; 453f7116284SIan Campbell 454f7116284SIan Campbell bind_virq.virq = bind.virq; 455f7116284SIan Campbell bind_virq.vcpu = 0; 456f7116284SIan Campbell rc = HYPERVISOR_event_channel_op(EVTCHNOP_bind_virq, 457f7116284SIan Campbell &bind_virq); 458f7116284SIan Campbell if (rc != 0) 459f7116284SIan Campbell break; 460f7116284SIan Campbell 461f7116284SIan Campbell rc = evtchn_bind_to_user(u, bind_virq.port); 462f7116284SIan Campbell if (rc == 0) 463f7116284SIan Campbell rc = bind_virq.port; 464f7116284SIan Campbell break; 465f7116284SIan Campbell } 466f7116284SIan Campbell 467f7116284SIan Campbell case IOCTL_EVTCHN_BIND_INTERDOMAIN: { 468f7116284SIan Campbell struct ioctl_evtchn_bind_interdomain bind; 469f7116284SIan Campbell struct evtchn_bind_interdomain bind_interdomain; 470f7116284SIan Campbell 471f7116284SIan Campbell rc = -EFAULT; 472f7116284SIan Campbell if (copy_from_user(&bind, uarg, sizeof(bind))) 473f7116284SIan Campbell break; 474f7116284SIan Campbell 475f7116284SIan Campbell bind_interdomain.remote_dom = bind.remote_domain; 476f7116284SIan Campbell bind_interdomain.remote_port = bind.remote_port; 477f7116284SIan Campbell rc = HYPERVISOR_event_channel_op(EVTCHNOP_bind_interdomain, 478f7116284SIan Campbell &bind_interdomain); 479f7116284SIan Campbell if (rc != 0) 480f7116284SIan Campbell break; 481f7116284SIan Campbell 482f7116284SIan Campbell rc = evtchn_bind_to_user(u, bind_interdomain.local_port); 483f7116284SIan Campbell if (rc == 0) 484f7116284SIan Campbell rc = bind_interdomain.local_port; 485f7116284SIan Campbell break; 486f7116284SIan Campbell } 487f7116284SIan Campbell 488f7116284SIan Campbell case IOCTL_EVTCHN_BIND_UNBOUND_PORT: { 489f7116284SIan Campbell struct ioctl_evtchn_bind_unbound_port bind; 490f7116284SIan Campbell struct evtchn_alloc_unbound alloc_unbound; 491f7116284SIan Campbell 492f7116284SIan Campbell rc = -EFAULT; 493f7116284SIan Campbell if (copy_from_user(&bind, uarg, sizeof(bind))) 494f7116284SIan Campbell break; 495f7116284SIan Campbell 496f7116284SIan Campbell alloc_unbound.dom = DOMID_SELF; 497f7116284SIan Campbell alloc_unbound.remote_dom = bind.remote_domain; 498f7116284SIan Campbell rc = HYPERVISOR_event_channel_op(EVTCHNOP_alloc_unbound, 499f7116284SIan Campbell &alloc_unbound); 500f7116284SIan Campbell if (rc != 0) 501f7116284SIan Campbell break; 502f7116284SIan Campbell 503f7116284SIan Campbell rc = evtchn_bind_to_user(u, alloc_unbound.port); 504f7116284SIan Campbell if (rc == 0) 505f7116284SIan Campbell rc = alloc_unbound.port; 506f7116284SIan Campbell break; 507f7116284SIan Campbell } 508f7116284SIan Campbell 509f7116284SIan Campbell case IOCTL_EVTCHN_UNBIND: { 510f7116284SIan Campbell struct ioctl_evtchn_unbind unbind; 51173cc4bb0SDavid Vrabel struct user_evtchn *evtchn; 512f7116284SIan Campbell 513f7116284SIan Campbell rc = -EFAULT; 514f7116284SIan Campbell if (copy_from_user(&unbind, uarg, sizeof(unbind))) 515f7116284SIan Campbell break; 516f7116284SIan Campbell 517f7116284SIan Campbell rc = -EINVAL; 5180dc0064aSDavid Vrabel if (unbind.port >= xen_evtchn_nr_channels()) 519f7116284SIan Campbell break; 520f7116284SIan Campbell 521f7116284SIan Campbell rc = -ENOTCONN; 52273cc4bb0SDavid Vrabel evtchn = find_evtchn(u, unbind.port); 52373cc4bb0SDavid Vrabel if (!evtchn) 524f7116284SIan Campbell break; 525f7116284SIan Campbell 5263f5e554fSJeremy Fitzhardinge disable_irq(irq_from_evtchn(unbind.port)); 52773cc4bb0SDavid Vrabel evtchn_unbind_from_user(u, evtchn); 528f7116284SIan Campbell rc = 0; 529f7116284SIan Campbell break; 530f7116284SIan Campbell } 531f7116284SIan Campbell 532f7116284SIan Campbell case IOCTL_EVTCHN_NOTIFY: { 533f7116284SIan Campbell struct ioctl_evtchn_notify notify; 53473cc4bb0SDavid Vrabel struct user_evtchn *evtchn; 535f7116284SIan Campbell 536f7116284SIan Campbell rc = -EFAULT; 537f7116284SIan Campbell if (copy_from_user(¬ify, uarg, sizeof(notify))) 538f7116284SIan Campbell break; 539f7116284SIan Campbell 540f7116284SIan Campbell rc = -ENOTCONN; 54173cc4bb0SDavid Vrabel evtchn = find_evtchn(u, notify.port); 54273cc4bb0SDavid Vrabel if (evtchn) { 543f7116284SIan Campbell notify_remote_via_evtchn(notify.port); 544f7116284SIan Campbell rc = 0; 545f7116284SIan Campbell } 546f7116284SIan Campbell break; 547f7116284SIan Campbell } 548f7116284SIan Campbell 549f7116284SIan Campbell case IOCTL_EVTCHN_RESET: { 550f7116284SIan Campbell /* Initialise the ring to empty. Clear errors. */ 551f7116284SIan Campbell mutex_lock(&u->ring_cons_mutex); 55273cc4bb0SDavid Vrabel spin_lock_irq(&u->ring_prod_lock); 553f7116284SIan Campbell u->ring_cons = u->ring_prod = u->ring_overflow = 0; 55473cc4bb0SDavid Vrabel spin_unlock_irq(&u->ring_prod_lock); 555f7116284SIan Campbell mutex_unlock(&u->ring_cons_mutex); 556f7116284SIan Campbell rc = 0; 557f7116284SIan Campbell break; 558f7116284SIan Campbell } 559f7116284SIan Campbell 560f7116284SIan Campbell default: 561f7116284SIan Campbell rc = -ENOSYS; 562f7116284SIan Campbell break; 563f7116284SIan Campbell } 5640a4666b5SJeremy Fitzhardinge mutex_unlock(&u->bind_mutex); 565f7116284SIan Campbell 566f7116284SIan Campbell return rc; 567f7116284SIan Campbell } 568f7116284SIan Campbell 569f7116284SIan Campbell static unsigned int evtchn_poll(struct file *file, poll_table *wait) 570f7116284SIan Campbell { 571f7116284SIan Campbell unsigned int mask = POLLOUT | POLLWRNORM; 572f7116284SIan Campbell struct per_user_data *u = file->private_data; 573f7116284SIan Campbell 574f7116284SIan Campbell poll_wait(file, &u->evtchn_wait, wait); 575f7116284SIan Campbell if (u->ring_cons != u->ring_prod) 576f7116284SIan Campbell mask |= POLLIN | POLLRDNORM; 577f7116284SIan Campbell if (u->ring_overflow) 578f7116284SIan Campbell mask = POLLERR; 579f7116284SIan Campbell return mask; 580f7116284SIan Campbell } 581f7116284SIan Campbell 582f7116284SIan Campbell static int evtchn_fasync(int fd, struct file *filp, int on) 583f7116284SIan Campbell { 584f7116284SIan Campbell struct per_user_data *u = filp->private_data; 585f7116284SIan Campbell return fasync_helper(fd, filp, on, &u->evtchn_async_queue); 586f7116284SIan Campbell } 587f7116284SIan Campbell 588f7116284SIan Campbell static int evtchn_open(struct inode *inode, struct file *filp) 589f7116284SIan Campbell { 590f7116284SIan Campbell struct per_user_data *u; 591f7116284SIan Campbell 592f7116284SIan Campbell u = kzalloc(sizeof(*u), GFP_KERNEL); 593f7116284SIan Campbell if (u == NULL) 594f7116284SIan Campbell return -ENOMEM; 595f7116284SIan Campbell 596f7116284SIan Campbell u->name = kasprintf(GFP_KERNEL, "evtchn:%s", current->comm); 597f7116284SIan Campbell if (u->name == NULL) { 598f7116284SIan Campbell kfree(u); 599f7116284SIan Campbell return -ENOMEM; 600f7116284SIan Campbell } 601f7116284SIan Campbell 602f7116284SIan Campbell init_waitqueue_head(&u->evtchn_wait); 603f7116284SIan Campbell 6040a4666b5SJeremy Fitzhardinge mutex_init(&u->bind_mutex); 605f7116284SIan Campbell mutex_init(&u->ring_cons_mutex); 60673cc4bb0SDavid Vrabel spin_lock_init(&u->ring_prod_lock); 607f7116284SIan Campbell 608f7116284SIan Campbell filp->private_data = u; 609f7116284SIan Campbell 6106eab04a8SJustin P. Mattock return nonseekable_open(inode, filp); 611f7116284SIan Campbell } 612f7116284SIan Campbell 613f7116284SIan Campbell static int evtchn_release(struct inode *inode, struct file *filp) 614f7116284SIan Campbell { 615f7116284SIan Campbell struct per_user_data *u = filp->private_data; 61673cc4bb0SDavid Vrabel struct rb_node *node; 617f7116284SIan Campbell 61873cc4bb0SDavid Vrabel while ((node = u->evtchns.rb_node)) { 61973cc4bb0SDavid Vrabel struct user_evtchn *evtchn; 620f7116284SIan Campbell 62173cc4bb0SDavid Vrabel evtchn = rb_entry(node, struct user_evtchn, node); 62273cc4bb0SDavid Vrabel disable_irq(irq_from_evtchn(evtchn->port)); 62373cc4bb0SDavid Vrabel evtchn_unbind_from_user(u, evtchn); 6243f5e554fSJeremy Fitzhardinge } 6253f5e554fSJeremy Fitzhardinge 626*86200154SDavid Vrabel evtchn_free_ring(u->ring); 627f7116284SIan Campbell kfree(u->name); 628f7116284SIan Campbell kfree(u); 629f7116284SIan Campbell 630f7116284SIan Campbell return 0; 631f7116284SIan Campbell } 632f7116284SIan Campbell 633f7116284SIan Campbell static const struct file_operations evtchn_fops = { 634f7116284SIan Campbell .owner = THIS_MODULE, 635f7116284SIan Campbell .read = evtchn_read, 636f7116284SIan Campbell .write = evtchn_write, 637f7116284SIan Campbell .unlocked_ioctl = evtchn_ioctl, 638f7116284SIan Campbell .poll = evtchn_poll, 639f7116284SIan Campbell .fasync = evtchn_fasync, 640f7116284SIan Campbell .open = evtchn_open, 641f7116284SIan Campbell .release = evtchn_release, 642bc7fc5e3SJeremy Fitzhardinge .llseek = no_llseek, 643f7116284SIan Campbell }; 644f7116284SIan Campbell 645f7116284SIan Campbell static struct miscdevice evtchn_miscdev = { 646f7116284SIan Campbell .minor = MISC_DYNAMIC_MINOR, 647376d908fSBastian Blank .name = "xen/evtchn", 648f7116284SIan Campbell .fops = &evtchn_fops, 649f7116284SIan Campbell }; 650f7116284SIan Campbell static int __init evtchn_init(void) 651f7116284SIan Campbell { 652f7116284SIan Campbell int err; 653f7116284SIan Campbell 654f7116284SIan Campbell if (!xen_domain()) 655f7116284SIan Campbell return -ENODEV; 656f7116284SIan Campbell 65718283ea7SWei Liu /* Create '/dev/xen/evtchn'. */ 658f7116284SIan Campbell err = misc_register(&evtchn_miscdev); 659f7116284SIan Campbell if (err != 0) { 660283c0972SJoe Perches pr_err("Could not register /dev/xen/evtchn\n"); 661f7116284SIan Campbell return err; 662f7116284SIan Campbell } 663f7116284SIan Campbell 664283c0972SJoe Perches pr_info("Event-channel device installed\n"); 665f7116284SIan Campbell 666f7116284SIan Campbell return 0; 667f7116284SIan Campbell } 668f7116284SIan Campbell 669f7116284SIan Campbell static void __exit evtchn_cleanup(void) 670f7116284SIan Campbell { 671f7116284SIan Campbell misc_deregister(&evtchn_miscdev); 672f7116284SIan Campbell } 673f7116284SIan Campbell 674f7116284SIan Campbell module_init(evtchn_init); 675f7116284SIan Campbell module_exit(evtchn_cleanup); 676f7116284SIan Campbell 677f7116284SIan Campbell MODULE_LICENSE("GPL"); 678