1d8414d3cSBastian Blank /****************************************************************************** 2d8414d3cSBastian Blank * privcmd.c 3d8414d3cSBastian Blank * 4d8414d3cSBastian Blank * Interface to privileged domain-0 commands. 5d8414d3cSBastian Blank * 6d8414d3cSBastian Blank * Copyright (c) 2002-2004, K A Fraser, B Dragovic 7d8414d3cSBastian Blank */ 8d8414d3cSBastian Blank 9d8414d3cSBastian Blank #include <linux/kernel.h> 10d8414d3cSBastian Blank #include <linux/module.h> 11d8414d3cSBastian Blank #include <linux/sched.h> 12d8414d3cSBastian Blank #include <linux/slab.h> 13d8414d3cSBastian Blank #include <linux/string.h> 14d8414d3cSBastian Blank #include <linux/errno.h> 15d8414d3cSBastian Blank #include <linux/mm.h> 16d8414d3cSBastian Blank #include <linux/mman.h> 17d8414d3cSBastian Blank #include <linux/uaccess.h> 18d8414d3cSBastian Blank #include <linux/swap.h> 19d8414d3cSBastian Blank #include <linux/highmem.h> 20d8414d3cSBastian Blank #include <linux/pagemap.h> 21d8414d3cSBastian Blank #include <linux/seq_file.h> 22d8414d3cSBastian Blank #include <linux/miscdevice.h> 23d8414d3cSBastian Blank 24d8414d3cSBastian Blank #include <asm/pgalloc.h> 25d8414d3cSBastian Blank #include <asm/pgtable.h> 26d8414d3cSBastian Blank #include <asm/tlb.h> 27d8414d3cSBastian Blank #include <asm/xen/hypervisor.h> 28d8414d3cSBastian Blank #include <asm/xen/hypercall.h> 29d8414d3cSBastian Blank 30d8414d3cSBastian Blank #include <xen/xen.h> 31d8414d3cSBastian Blank #include <xen/privcmd.h> 32d8414d3cSBastian Blank #include <xen/interface/xen.h> 33d8414d3cSBastian Blank #include <xen/features.h> 34d8414d3cSBastian Blank #include <xen/page.h> 35d8414d3cSBastian Blank #include <xen/xen-ops.h> 36d8414d3cSBastian Blank 37d8414d3cSBastian Blank #include "privcmd.h" 38d8414d3cSBastian Blank 39d8414d3cSBastian Blank MODULE_LICENSE("GPL"); 40d8414d3cSBastian Blank 41d8414d3cSBastian Blank #ifndef HAVE_ARCH_PRIVCMD_MMAP 42d8414d3cSBastian Blank static int privcmd_enforce_singleshot_mapping(struct vm_area_struct *vma); 43d8414d3cSBastian Blank #endif 44d8414d3cSBastian Blank 45d8414d3cSBastian Blank static long privcmd_ioctl_hypercall(void __user *udata) 46d8414d3cSBastian Blank { 47d8414d3cSBastian Blank struct privcmd_hypercall hypercall; 48d8414d3cSBastian Blank long ret; 49d8414d3cSBastian Blank 50d8414d3cSBastian Blank if (copy_from_user(&hypercall, udata, sizeof(hypercall))) 51d8414d3cSBastian Blank return -EFAULT; 52d8414d3cSBastian Blank 53d8414d3cSBastian Blank ret = privcmd_call(hypercall.op, 54d8414d3cSBastian Blank hypercall.arg[0], hypercall.arg[1], 55d8414d3cSBastian Blank hypercall.arg[2], hypercall.arg[3], 56d8414d3cSBastian Blank hypercall.arg[4]); 57d8414d3cSBastian Blank 58d8414d3cSBastian Blank return ret; 59d8414d3cSBastian Blank } 60d8414d3cSBastian Blank 61d8414d3cSBastian Blank static void free_page_list(struct list_head *pages) 62d8414d3cSBastian Blank { 63d8414d3cSBastian Blank struct page *p, *n; 64d8414d3cSBastian Blank 65d8414d3cSBastian Blank list_for_each_entry_safe(p, n, pages, lru) 66d8414d3cSBastian Blank __free_page(p); 67d8414d3cSBastian Blank 68d8414d3cSBastian Blank INIT_LIST_HEAD(pages); 69d8414d3cSBastian Blank } 70d8414d3cSBastian Blank 71d8414d3cSBastian Blank /* 72d8414d3cSBastian Blank * Given an array of items in userspace, return a list of pages 73d8414d3cSBastian Blank * containing the data. If copying fails, either because of memory 74d8414d3cSBastian Blank * allocation failure or a problem reading user memory, return an 75d8414d3cSBastian Blank * error code; its up to the caller to dispose of any partial list. 76d8414d3cSBastian Blank */ 77d8414d3cSBastian Blank static int gather_array(struct list_head *pagelist, 78d8414d3cSBastian Blank unsigned nelem, size_t size, 79ceb90fa0SAndres Lagar-Cavilla const void __user *data) 80d8414d3cSBastian Blank { 81d8414d3cSBastian Blank unsigned pageidx; 82d8414d3cSBastian Blank void *pagedata; 83d8414d3cSBastian Blank int ret; 84d8414d3cSBastian Blank 85d8414d3cSBastian Blank if (size > PAGE_SIZE) 86d8414d3cSBastian Blank return 0; 87d8414d3cSBastian Blank 88d8414d3cSBastian Blank pageidx = PAGE_SIZE; 89d8414d3cSBastian Blank pagedata = NULL; /* quiet, gcc */ 90d8414d3cSBastian Blank while (nelem--) { 91d8414d3cSBastian Blank if (pageidx > PAGE_SIZE-size) { 92d8414d3cSBastian Blank struct page *page = alloc_page(GFP_KERNEL); 93d8414d3cSBastian Blank 94d8414d3cSBastian Blank ret = -ENOMEM; 95d8414d3cSBastian Blank if (page == NULL) 96d8414d3cSBastian Blank goto fail; 97d8414d3cSBastian Blank 98d8414d3cSBastian Blank pagedata = page_address(page); 99d8414d3cSBastian Blank 100d8414d3cSBastian Blank list_add_tail(&page->lru, pagelist); 101d8414d3cSBastian Blank pageidx = 0; 102d8414d3cSBastian Blank } 103d8414d3cSBastian Blank 104d8414d3cSBastian Blank ret = -EFAULT; 105d8414d3cSBastian Blank if (copy_from_user(pagedata + pageidx, data, size)) 106d8414d3cSBastian Blank goto fail; 107d8414d3cSBastian Blank 108d8414d3cSBastian Blank data += size; 109d8414d3cSBastian Blank pageidx += size; 110d8414d3cSBastian Blank } 111d8414d3cSBastian Blank 112d8414d3cSBastian Blank ret = 0; 113d8414d3cSBastian Blank 114d8414d3cSBastian Blank fail: 115d8414d3cSBastian Blank return ret; 116d8414d3cSBastian Blank } 117d8414d3cSBastian Blank 118d8414d3cSBastian Blank /* 119d8414d3cSBastian Blank * Call function "fn" on each element of the array fragmented 120d8414d3cSBastian Blank * over a list of pages. 121d8414d3cSBastian Blank */ 122d8414d3cSBastian Blank static int traverse_pages(unsigned nelem, size_t size, 123d8414d3cSBastian Blank struct list_head *pos, 124d8414d3cSBastian Blank int (*fn)(void *data, void *state), 125d8414d3cSBastian Blank void *state) 126d8414d3cSBastian Blank { 127d8414d3cSBastian Blank void *pagedata; 128d8414d3cSBastian Blank unsigned pageidx; 129d8414d3cSBastian Blank int ret = 0; 130d8414d3cSBastian Blank 131d8414d3cSBastian Blank BUG_ON(size > PAGE_SIZE); 132d8414d3cSBastian Blank 133d8414d3cSBastian Blank pageidx = PAGE_SIZE; 134d8414d3cSBastian Blank pagedata = NULL; /* hush, gcc */ 135d8414d3cSBastian Blank 136d8414d3cSBastian Blank while (nelem--) { 137d8414d3cSBastian Blank if (pageidx > PAGE_SIZE-size) { 138d8414d3cSBastian Blank struct page *page; 139d8414d3cSBastian Blank pos = pos->next; 140d8414d3cSBastian Blank page = list_entry(pos, struct page, lru); 141d8414d3cSBastian Blank pagedata = page_address(page); 142d8414d3cSBastian Blank pageidx = 0; 143d8414d3cSBastian Blank } 144d8414d3cSBastian Blank 145d8414d3cSBastian Blank ret = (*fn)(pagedata + pageidx, state); 146d8414d3cSBastian Blank if (ret) 147d8414d3cSBastian Blank break; 148d8414d3cSBastian Blank pageidx += size; 149d8414d3cSBastian Blank } 150d8414d3cSBastian Blank 151d8414d3cSBastian Blank return ret; 152d8414d3cSBastian Blank } 153d8414d3cSBastian Blank 154d8414d3cSBastian Blank struct mmap_mfn_state { 155d8414d3cSBastian Blank unsigned long va; 156d8414d3cSBastian Blank struct vm_area_struct *vma; 157d8414d3cSBastian Blank domid_t domain; 158d8414d3cSBastian Blank }; 159d8414d3cSBastian Blank 160d8414d3cSBastian Blank static int mmap_mfn_range(void *data, void *state) 161d8414d3cSBastian Blank { 162d8414d3cSBastian Blank struct privcmd_mmap_entry *msg = data; 163d8414d3cSBastian Blank struct mmap_mfn_state *st = state; 164d8414d3cSBastian Blank struct vm_area_struct *vma = st->vma; 165d8414d3cSBastian Blank int rc; 166d8414d3cSBastian Blank 167d8414d3cSBastian Blank /* Do not allow range to wrap the address space. */ 168d8414d3cSBastian Blank if ((msg->npages > (LONG_MAX >> PAGE_SHIFT)) || 169d8414d3cSBastian Blank ((unsigned long)(msg->npages << PAGE_SHIFT) >= -st->va)) 170d8414d3cSBastian Blank return -EINVAL; 171d8414d3cSBastian Blank 172d8414d3cSBastian Blank /* Range chunks must be contiguous in va space. */ 173d8414d3cSBastian Blank if ((msg->va != st->va) || 174d8414d3cSBastian Blank ((msg->va+(msg->npages<<PAGE_SHIFT)) > vma->vm_end)) 175d8414d3cSBastian Blank return -EINVAL; 176d8414d3cSBastian Blank 177d8414d3cSBastian Blank rc = xen_remap_domain_mfn_range(vma, 178d8414d3cSBastian Blank msg->va & PAGE_MASK, 179d8414d3cSBastian Blank msg->mfn, msg->npages, 180d8414d3cSBastian Blank vma->vm_page_prot, 181d8414d3cSBastian Blank st->domain); 182d8414d3cSBastian Blank if (rc < 0) 183d8414d3cSBastian Blank return rc; 184d8414d3cSBastian Blank 185d8414d3cSBastian Blank st->va += msg->npages << PAGE_SHIFT; 186d8414d3cSBastian Blank 187d8414d3cSBastian Blank return 0; 188d8414d3cSBastian Blank } 189d8414d3cSBastian Blank 190d8414d3cSBastian Blank static long privcmd_ioctl_mmap(void __user *udata) 191d8414d3cSBastian Blank { 192d8414d3cSBastian Blank struct privcmd_mmap mmapcmd; 193d8414d3cSBastian Blank struct mm_struct *mm = current->mm; 194d8414d3cSBastian Blank struct vm_area_struct *vma; 195d8414d3cSBastian Blank int rc; 196d8414d3cSBastian Blank LIST_HEAD(pagelist); 197d8414d3cSBastian Blank struct mmap_mfn_state state; 198d8414d3cSBastian Blank 199d8414d3cSBastian Blank if (!xen_initial_domain()) 200d8414d3cSBastian Blank return -EPERM; 201d8414d3cSBastian Blank 202d8414d3cSBastian Blank if (copy_from_user(&mmapcmd, udata, sizeof(mmapcmd))) 203d8414d3cSBastian Blank return -EFAULT; 204d8414d3cSBastian Blank 205d8414d3cSBastian Blank rc = gather_array(&pagelist, 206d8414d3cSBastian Blank mmapcmd.num, sizeof(struct privcmd_mmap_entry), 207d8414d3cSBastian Blank mmapcmd.entry); 208d8414d3cSBastian Blank 209d8414d3cSBastian Blank if (rc || list_empty(&pagelist)) 210d8414d3cSBastian Blank goto out; 211d8414d3cSBastian Blank 212d8414d3cSBastian Blank down_write(&mm->mmap_sem); 213d8414d3cSBastian Blank 214d8414d3cSBastian Blank { 215d8414d3cSBastian Blank struct page *page = list_first_entry(&pagelist, 216d8414d3cSBastian Blank struct page, lru); 217d8414d3cSBastian Blank struct privcmd_mmap_entry *msg = page_address(page); 218d8414d3cSBastian Blank 219d8414d3cSBastian Blank vma = find_vma(mm, msg->va); 220d8414d3cSBastian Blank rc = -EINVAL; 221d8414d3cSBastian Blank 222d8414d3cSBastian Blank if (!vma || (msg->va != vma->vm_start) || 223d8414d3cSBastian Blank !privcmd_enforce_singleshot_mapping(vma)) 224d8414d3cSBastian Blank goto out_up; 225d8414d3cSBastian Blank } 226d8414d3cSBastian Blank 227d8414d3cSBastian Blank state.va = vma->vm_start; 228d8414d3cSBastian Blank state.vma = vma; 229d8414d3cSBastian Blank state.domain = mmapcmd.dom; 230d8414d3cSBastian Blank 231d8414d3cSBastian Blank rc = traverse_pages(mmapcmd.num, sizeof(struct privcmd_mmap_entry), 232d8414d3cSBastian Blank &pagelist, 233d8414d3cSBastian Blank mmap_mfn_range, &state); 234d8414d3cSBastian Blank 235d8414d3cSBastian Blank 236d8414d3cSBastian Blank out_up: 237d8414d3cSBastian Blank up_write(&mm->mmap_sem); 238d8414d3cSBastian Blank 239d8414d3cSBastian Blank out: 240d8414d3cSBastian Blank free_page_list(&pagelist); 241d8414d3cSBastian Blank 242d8414d3cSBastian Blank return rc; 243d8414d3cSBastian Blank } 244d8414d3cSBastian Blank 245d8414d3cSBastian Blank struct mmap_batch_state { 246d8414d3cSBastian Blank domid_t domain; 247d8414d3cSBastian Blank unsigned long va; 248d8414d3cSBastian Blank struct vm_area_struct *vma; 249ceb90fa0SAndres Lagar-Cavilla /* A tristate: 250ceb90fa0SAndres Lagar-Cavilla * 0 for no errors 251ceb90fa0SAndres Lagar-Cavilla * 1 if at least one error has happened (and no 252ceb90fa0SAndres Lagar-Cavilla * -ENOENT errors have happened) 253ceb90fa0SAndres Lagar-Cavilla * -ENOENT if at least 1 -ENOENT has happened. 254ceb90fa0SAndres Lagar-Cavilla */ 255ceb90fa0SAndres Lagar-Cavilla int global_error; 256ceb90fa0SAndres Lagar-Cavilla /* An array for individual errors */ 257ceb90fa0SAndres Lagar-Cavilla int *err; 258d8414d3cSBastian Blank 259ceb90fa0SAndres Lagar-Cavilla /* User-space mfn array to store errors in the second pass for V1. */ 260ceb90fa0SAndres Lagar-Cavilla xen_pfn_t __user *user_mfn; 261d8414d3cSBastian Blank }; 262d8414d3cSBastian Blank 263d8414d3cSBastian Blank static int mmap_batch_fn(void *data, void *state) 264d8414d3cSBastian Blank { 265d8414d3cSBastian Blank xen_pfn_t *mfnp = data; 266d8414d3cSBastian Blank struct mmap_batch_state *st = state; 267ceb90fa0SAndres Lagar-Cavilla int ret; 268d8414d3cSBastian Blank 269ceb90fa0SAndres Lagar-Cavilla ret = xen_remap_domain_mfn_range(st->vma, st->va & PAGE_MASK, *mfnp, 1, 270ceb90fa0SAndres Lagar-Cavilla st->vma->vm_page_prot, st->domain); 271ceb90fa0SAndres Lagar-Cavilla 272ceb90fa0SAndres Lagar-Cavilla /* Store error code for second pass. */ 273ceb90fa0SAndres Lagar-Cavilla *(st->err++) = ret; 274ceb90fa0SAndres Lagar-Cavilla 275ceb90fa0SAndres Lagar-Cavilla /* And see if it affects the global_error. */ 276ceb90fa0SAndres Lagar-Cavilla if (ret < 0) { 277ceb90fa0SAndres Lagar-Cavilla if (ret == -ENOENT) 278ceb90fa0SAndres Lagar-Cavilla st->global_error = -ENOENT; 279ceb90fa0SAndres Lagar-Cavilla else { 280ceb90fa0SAndres Lagar-Cavilla /* Record that at least one error has happened. */ 281ceb90fa0SAndres Lagar-Cavilla if (st->global_error == 0) 282ceb90fa0SAndres Lagar-Cavilla st->global_error = 1; 283ceb90fa0SAndres Lagar-Cavilla } 284d8414d3cSBastian Blank } 285d8414d3cSBastian Blank st->va += PAGE_SIZE; 286d8414d3cSBastian Blank 287d8414d3cSBastian Blank return 0; 288d8414d3cSBastian Blank } 289d8414d3cSBastian Blank 290ceb90fa0SAndres Lagar-Cavilla static int mmap_return_errors_v1(void *data, void *state) 291d8414d3cSBastian Blank { 292d8414d3cSBastian Blank xen_pfn_t *mfnp = data; 293d8414d3cSBastian Blank struct mmap_batch_state *st = state; 294ceb90fa0SAndres Lagar-Cavilla int err = *(st->err++); 295d8414d3cSBastian Blank 296ceb90fa0SAndres Lagar-Cavilla /* 297ceb90fa0SAndres Lagar-Cavilla * V1 encodes the error codes in the 32bit top nibble of the 298ceb90fa0SAndres Lagar-Cavilla * mfn (with its known limitations vis-a-vis 64 bit callers). 299ceb90fa0SAndres Lagar-Cavilla */ 300ceb90fa0SAndres Lagar-Cavilla *mfnp |= (err == -ENOENT) ? 301ceb90fa0SAndres Lagar-Cavilla PRIVCMD_MMAPBATCH_PAGED_ERROR : 302ceb90fa0SAndres Lagar-Cavilla PRIVCMD_MMAPBATCH_MFN_ERROR; 303ceb90fa0SAndres Lagar-Cavilla return __put_user(*mfnp, st->user_mfn++); 304d8414d3cSBastian Blank } 305d8414d3cSBastian Blank 306d8414d3cSBastian Blank static struct vm_operations_struct privcmd_vm_ops; 307d8414d3cSBastian Blank 308ceb90fa0SAndres Lagar-Cavilla static long privcmd_ioctl_mmap_batch(void __user *udata, int version) 309d8414d3cSBastian Blank { 310d8414d3cSBastian Blank int ret; 311ceb90fa0SAndres Lagar-Cavilla struct privcmd_mmapbatch_v2 m; 312d8414d3cSBastian Blank struct mm_struct *mm = current->mm; 313d8414d3cSBastian Blank struct vm_area_struct *vma; 314d8414d3cSBastian Blank unsigned long nr_pages; 315d8414d3cSBastian Blank LIST_HEAD(pagelist); 316ceb90fa0SAndres Lagar-Cavilla int *err_array = NULL; 317d8414d3cSBastian Blank struct mmap_batch_state state; 318d8414d3cSBastian Blank 319d8414d3cSBastian Blank if (!xen_initial_domain()) 320d8414d3cSBastian Blank return -EPERM; 321d8414d3cSBastian Blank 322ceb90fa0SAndres Lagar-Cavilla switch (version) { 323ceb90fa0SAndres Lagar-Cavilla case 1: 324ceb90fa0SAndres Lagar-Cavilla if (copy_from_user(&m, udata, sizeof(struct privcmd_mmapbatch))) 325d8414d3cSBastian Blank return -EFAULT; 326ceb90fa0SAndres Lagar-Cavilla /* Returns per-frame error in m.arr. */ 327ceb90fa0SAndres Lagar-Cavilla m.err = NULL; 328ceb90fa0SAndres Lagar-Cavilla if (!access_ok(VERIFY_WRITE, m.arr, m.num * sizeof(*m.arr))) 329ceb90fa0SAndres Lagar-Cavilla return -EFAULT; 330ceb90fa0SAndres Lagar-Cavilla break; 331ceb90fa0SAndres Lagar-Cavilla case 2: 332ceb90fa0SAndres Lagar-Cavilla if (copy_from_user(&m, udata, sizeof(struct privcmd_mmapbatch_v2))) 333ceb90fa0SAndres Lagar-Cavilla return -EFAULT; 334ceb90fa0SAndres Lagar-Cavilla /* Returns per-frame error code in m.err. */ 335ceb90fa0SAndres Lagar-Cavilla if (!access_ok(VERIFY_WRITE, m.err, m.num * (sizeof(*m.err)))) 336ceb90fa0SAndres Lagar-Cavilla return -EFAULT; 337ceb90fa0SAndres Lagar-Cavilla break; 338ceb90fa0SAndres Lagar-Cavilla default: 339ceb90fa0SAndres Lagar-Cavilla return -EINVAL; 340ceb90fa0SAndres Lagar-Cavilla } 341d8414d3cSBastian Blank 342d8414d3cSBastian Blank nr_pages = m.num; 343d8414d3cSBastian Blank if ((m.num <= 0) || (nr_pages > (LONG_MAX >> PAGE_SHIFT))) 344d8414d3cSBastian Blank return -EINVAL; 345d8414d3cSBastian Blank 346ceb90fa0SAndres Lagar-Cavilla ret = gather_array(&pagelist, m.num, sizeof(xen_pfn_t), m.arr); 347d8414d3cSBastian Blank 348ceb90fa0SAndres Lagar-Cavilla if (ret) 349d8414d3cSBastian Blank goto out; 350ceb90fa0SAndres Lagar-Cavilla if (list_empty(&pagelist)) { 351ceb90fa0SAndres Lagar-Cavilla ret = -EINVAL; 352ceb90fa0SAndres Lagar-Cavilla goto out; 353ceb90fa0SAndres Lagar-Cavilla } 354ceb90fa0SAndres Lagar-Cavilla 355ceb90fa0SAndres Lagar-Cavilla err_array = kcalloc(m.num, sizeof(int), GFP_KERNEL); 356ceb90fa0SAndres Lagar-Cavilla if (err_array == NULL) { 357ceb90fa0SAndres Lagar-Cavilla ret = -ENOMEM; 358ceb90fa0SAndres Lagar-Cavilla goto out; 359ceb90fa0SAndres Lagar-Cavilla } 360d8414d3cSBastian Blank 361d8414d3cSBastian Blank down_write(&mm->mmap_sem); 362d8414d3cSBastian Blank 363d8414d3cSBastian Blank vma = find_vma(mm, m.addr); 364d8414d3cSBastian Blank ret = -EINVAL; 365d8414d3cSBastian Blank if (!vma || 366d8414d3cSBastian Blank vma->vm_ops != &privcmd_vm_ops || 367d8414d3cSBastian Blank (m.addr != vma->vm_start) || 368d8414d3cSBastian Blank ((m.addr + (nr_pages << PAGE_SHIFT)) != vma->vm_end) || 369d8414d3cSBastian Blank !privcmd_enforce_singleshot_mapping(vma)) { 370d8414d3cSBastian Blank up_write(&mm->mmap_sem); 371d8414d3cSBastian Blank goto out; 372d8414d3cSBastian Blank } 373d8414d3cSBastian Blank 374d8414d3cSBastian Blank state.domain = m.dom; 375d8414d3cSBastian Blank state.vma = vma; 376d8414d3cSBastian Blank state.va = m.addr; 377ceb90fa0SAndres Lagar-Cavilla state.global_error = 0; 378ceb90fa0SAndres Lagar-Cavilla state.err = err_array; 379d8414d3cSBastian Blank 380ceb90fa0SAndres Lagar-Cavilla /* mmap_batch_fn guarantees ret == 0 */ 381ceb90fa0SAndres Lagar-Cavilla BUG_ON(traverse_pages(m.num, sizeof(xen_pfn_t), 382ceb90fa0SAndres Lagar-Cavilla &pagelist, mmap_batch_fn, &state)); 383d8414d3cSBastian Blank 384d8414d3cSBastian Blank up_write(&mm->mmap_sem); 385d8414d3cSBastian Blank 386ceb90fa0SAndres Lagar-Cavilla if (state.global_error && (version == 1)) { 387ceb90fa0SAndres Lagar-Cavilla /* Write back errors in second pass. */ 388ceb90fa0SAndres Lagar-Cavilla state.user_mfn = (xen_pfn_t *)m.arr; 389ceb90fa0SAndres Lagar-Cavilla state.err = err_array; 390d8414d3cSBastian Blank ret = traverse_pages(m.num, sizeof(xen_pfn_t), 391ceb90fa0SAndres Lagar-Cavilla &pagelist, mmap_return_errors_v1, &state); 3929d2be928SDan Carpenter } else if (version == 2) { 393ceb90fa0SAndres Lagar-Cavilla ret = __copy_to_user(m.err, err_array, m.num * sizeof(int)); 3949d2be928SDan Carpenter if (ret) 3959d2be928SDan Carpenter ret = -EFAULT; 3969d2be928SDan Carpenter } 397ceb90fa0SAndres Lagar-Cavilla 398ceb90fa0SAndres Lagar-Cavilla /* If we have not had any EFAULT-like global errors then set the global 399ceb90fa0SAndres Lagar-Cavilla * error to -ENOENT if necessary. */ 400ceb90fa0SAndres Lagar-Cavilla if ((ret == 0) && (state.global_error == -ENOENT)) 401ceb90fa0SAndres Lagar-Cavilla ret = -ENOENT; 402d8414d3cSBastian Blank 403d8414d3cSBastian Blank out: 404ceb90fa0SAndres Lagar-Cavilla kfree(err_array); 405d8414d3cSBastian Blank free_page_list(&pagelist); 406d8414d3cSBastian Blank 407d8414d3cSBastian Blank return ret; 408d8414d3cSBastian Blank } 409d8414d3cSBastian Blank 410d8414d3cSBastian Blank static long privcmd_ioctl(struct file *file, 411d8414d3cSBastian Blank unsigned int cmd, unsigned long data) 412d8414d3cSBastian Blank { 413d8414d3cSBastian Blank int ret = -ENOSYS; 414d8414d3cSBastian Blank void __user *udata = (void __user *) data; 415d8414d3cSBastian Blank 416d8414d3cSBastian Blank switch (cmd) { 417d8414d3cSBastian Blank case IOCTL_PRIVCMD_HYPERCALL: 418d8414d3cSBastian Blank ret = privcmd_ioctl_hypercall(udata); 419d8414d3cSBastian Blank break; 420d8414d3cSBastian Blank 421d8414d3cSBastian Blank case IOCTL_PRIVCMD_MMAP: 422d8414d3cSBastian Blank ret = privcmd_ioctl_mmap(udata); 423d8414d3cSBastian Blank break; 424d8414d3cSBastian Blank 425d8414d3cSBastian Blank case IOCTL_PRIVCMD_MMAPBATCH: 426ceb90fa0SAndres Lagar-Cavilla ret = privcmd_ioctl_mmap_batch(udata, 1); 427ceb90fa0SAndres Lagar-Cavilla break; 428ceb90fa0SAndres Lagar-Cavilla 429ceb90fa0SAndres Lagar-Cavilla case IOCTL_PRIVCMD_MMAPBATCH_V2: 430ceb90fa0SAndres Lagar-Cavilla ret = privcmd_ioctl_mmap_batch(udata, 2); 431d8414d3cSBastian Blank break; 432d8414d3cSBastian Blank 433d8414d3cSBastian Blank default: 434d8414d3cSBastian Blank ret = -EINVAL; 435d8414d3cSBastian Blank break; 436d8414d3cSBastian Blank } 437d8414d3cSBastian Blank 438d8414d3cSBastian Blank return ret; 439d8414d3cSBastian Blank } 440d8414d3cSBastian Blank 441d8414d3cSBastian Blank static int privcmd_fault(struct vm_area_struct *vma, struct vm_fault *vmf) 442d8414d3cSBastian Blank { 443d8414d3cSBastian Blank printk(KERN_DEBUG "privcmd_fault: vma=%p %lx-%lx, pgoff=%lx, uv=%p\n", 444d8414d3cSBastian Blank vma, vma->vm_start, vma->vm_end, 445d8414d3cSBastian Blank vmf->pgoff, vmf->virtual_address); 446d8414d3cSBastian Blank 447d8414d3cSBastian Blank return VM_FAULT_SIGBUS; 448d8414d3cSBastian Blank } 449d8414d3cSBastian Blank 450d8414d3cSBastian Blank static struct vm_operations_struct privcmd_vm_ops = { 451d8414d3cSBastian Blank .fault = privcmd_fault 452d8414d3cSBastian Blank }; 453d8414d3cSBastian Blank 454d8414d3cSBastian Blank static int privcmd_mmap(struct file *file, struct vm_area_struct *vma) 455d8414d3cSBastian Blank { 456d8414d3cSBastian Blank /* DONTCOPY is essential for Xen because copy_page_range doesn't know 457d8414d3cSBastian Blank * how to recreate these mappings */ 458*314e51b9SKonstantin Khlebnikov vma->vm_flags |= VM_IO | VM_PFNMAP | VM_DONTCOPY | 459*314e51b9SKonstantin Khlebnikov VM_DONTEXPAND | VM_DONTDUMP; 460d8414d3cSBastian Blank vma->vm_ops = &privcmd_vm_ops; 461d8414d3cSBastian Blank vma->vm_private_data = NULL; 462d8414d3cSBastian Blank 463d8414d3cSBastian Blank return 0; 464d8414d3cSBastian Blank } 465d8414d3cSBastian Blank 466d8414d3cSBastian Blank static int privcmd_enforce_singleshot_mapping(struct vm_area_struct *vma) 467d8414d3cSBastian Blank { 468d8414d3cSBastian Blank return (xchg(&vma->vm_private_data, (void *)1) == NULL); 469d8414d3cSBastian Blank } 470d8414d3cSBastian Blank 471d8414d3cSBastian Blank const struct file_operations xen_privcmd_fops = { 472d8414d3cSBastian Blank .owner = THIS_MODULE, 473d8414d3cSBastian Blank .unlocked_ioctl = privcmd_ioctl, 474d8414d3cSBastian Blank .mmap = privcmd_mmap, 475d8414d3cSBastian Blank }; 476d8414d3cSBastian Blank EXPORT_SYMBOL_GPL(xen_privcmd_fops); 477d8414d3cSBastian Blank 478d8414d3cSBastian Blank static struct miscdevice privcmd_dev = { 479d8414d3cSBastian Blank .minor = MISC_DYNAMIC_MINOR, 480d8414d3cSBastian Blank .name = "xen/privcmd", 481d8414d3cSBastian Blank .fops = &xen_privcmd_fops, 482d8414d3cSBastian Blank }; 483d8414d3cSBastian Blank 484d8414d3cSBastian Blank static int __init privcmd_init(void) 485d8414d3cSBastian Blank { 486d8414d3cSBastian Blank int err; 487d8414d3cSBastian Blank 488d8414d3cSBastian Blank if (!xen_domain()) 489d8414d3cSBastian Blank return -ENODEV; 490d8414d3cSBastian Blank 491d8414d3cSBastian Blank err = misc_register(&privcmd_dev); 492d8414d3cSBastian Blank if (err != 0) { 493d8414d3cSBastian Blank printk(KERN_ERR "Could not register Xen privcmd device\n"); 494d8414d3cSBastian Blank return err; 495d8414d3cSBastian Blank } 496d8414d3cSBastian Blank return 0; 497d8414d3cSBastian Blank } 498d8414d3cSBastian Blank 499d8414d3cSBastian Blank static void __exit privcmd_exit(void) 500d8414d3cSBastian Blank { 501d8414d3cSBastian Blank misc_deregister(&privcmd_dev); 502d8414d3cSBastian Blank } 503d8414d3cSBastian Blank 504d8414d3cSBastian Blank module_init(privcmd_init); 505d8414d3cSBastian Blank module_exit(privcmd_exit); 506