diff options
Diffstat (limited to 'drivers/xen/xenfs/privcmd.c')
| -rw-r--r-- | drivers/xen/xenfs/privcmd.c | 404 |
1 files changed, 404 insertions, 0 deletions
diff --git a/drivers/xen/xenfs/privcmd.c b/drivers/xen/xenfs/privcmd.c new file mode 100644 index 000000000000..f80be7f6eb95 --- /dev/null +++ b/drivers/xen/xenfs/privcmd.c | |||
| @@ -0,0 +1,404 @@ | |||
| 1 | /****************************************************************************** | ||
| 2 | * privcmd.c | ||
| 3 | * | ||
| 4 | * Interface to privileged domain-0 commands. | ||
| 5 | * | ||
| 6 | * Copyright (c) 2002-2004, K A Fraser, B Dragovic | ||
| 7 | */ | ||
| 8 | |||
| 9 | #include <linux/kernel.h> | ||
| 10 | #include <linux/sched.h> | ||
| 11 | #include <linux/slab.h> | ||
| 12 | #include <linux/string.h> | ||
| 13 | #include <linux/errno.h> | ||
| 14 | #include <linux/mm.h> | ||
| 15 | #include <linux/mman.h> | ||
| 16 | #include <linux/uaccess.h> | ||
| 17 | #include <linux/swap.h> | ||
| 18 | #include <linux/smp_lock.h> | ||
| 19 | #include <linux/highmem.h> | ||
| 20 | #include <linux/pagemap.h> | ||
| 21 | #include <linux/seq_file.h> | ||
| 22 | |||
| 23 | #include <asm/pgalloc.h> | ||
| 24 | #include <asm/pgtable.h> | ||
| 25 | #include <asm/tlb.h> | ||
| 26 | #include <asm/xen/hypervisor.h> | ||
| 27 | #include <asm/xen/hypercall.h> | ||
| 28 | |||
| 29 | #include <xen/xen.h> | ||
| 30 | #include <xen/privcmd.h> | ||
| 31 | #include <xen/interface/xen.h> | ||
| 32 | #include <xen/features.h> | ||
| 33 | #include <xen/page.h> | ||
| 34 | #include <xen/xen-ops.h> | ||
| 35 | |||
| 36 | #ifndef HAVE_ARCH_PRIVCMD_MMAP | ||
| 37 | static int privcmd_enforce_singleshot_mapping(struct vm_area_struct *vma); | ||
| 38 | #endif | ||
| 39 | |||
| 40 | static long privcmd_ioctl_hypercall(void __user *udata) | ||
| 41 | { | ||
| 42 | struct privcmd_hypercall hypercall; | ||
| 43 | long ret; | ||
| 44 | |||
| 45 | if (copy_from_user(&hypercall, udata, sizeof(hypercall))) | ||
| 46 | return -EFAULT; | ||
| 47 | |||
| 48 | ret = privcmd_call(hypercall.op, | ||
| 49 | hypercall.arg[0], hypercall.arg[1], | ||
| 50 | hypercall.arg[2], hypercall.arg[3], | ||
| 51 | hypercall.arg[4]); | ||
| 52 | |||
| 53 | return ret; | ||
| 54 | } | ||
| 55 | |||
| 56 | static void free_page_list(struct list_head *pages) | ||
| 57 | { | ||
| 58 | struct page *p, *n; | ||
| 59 | |||
| 60 | list_for_each_entry_safe(p, n, pages, lru) | ||
| 61 | __free_page(p); | ||
| 62 | |||
| 63 | INIT_LIST_HEAD(pages); | ||
| 64 | } | ||
| 65 | |||
| 66 | /* | ||
| 67 | * Given an array of items in userspace, return a list of pages | ||
| 68 | * containing the data. If copying fails, either because of memory | ||
| 69 | * allocation failure or a problem reading user memory, return an | ||
| 70 | * error code; its up to the caller to dispose of any partial list. | ||
| 71 | */ | ||
| 72 | static int gather_array(struct list_head *pagelist, | ||
| 73 | unsigned nelem, size_t size, | ||
| 74 | void __user *data) | ||
| 75 | { | ||
| 76 | unsigned pageidx; | ||
| 77 | void *pagedata; | ||
| 78 | int ret; | ||
| 79 | |||
| 80 | if (size > PAGE_SIZE) | ||
| 81 | return 0; | ||
| 82 | |||
| 83 | pageidx = PAGE_SIZE; | ||
| 84 | pagedata = NULL; /* quiet, gcc */ | ||
| 85 | while (nelem--) { | ||
| 86 | if (pageidx > PAGE_SIZE-size) { | ||
| 87 | struct page *page = alloc_page(GFP_KERNEL); | ||
| 88 | |||
| 89 | ret = -ENOMEM; | ||
| 90 | if (page == NULL) | ||
| 91 | goto fail; | ||
| 92 | |||
| 93 | pagedata = page_address(page); | ||
| 94 | |||
| 95 | list_add_tail(&page->lru, pagelist); | ||
| 96 | pageidx = 0; | ||
| 97 | } | ||
| 98 | |||
| 99 | ret = -EFAULT; | ||
| 100 | if (copy_from_user(pagedata + pageidx, data, size)) | ||
| 101 | goto fail; | ||
| 102 | |||
| 103 | data += size; | ||
| 104 | pageidx += size; | ||
| 105 | } | ||
| 106 | |||
| 107 | ret = 0; | ||
| 108 | |||
| 109 | fail: | ||
| 110 | return ret; | ||
| 111 | } | ||
| 112 | |||
| 113 | /* | ||
| 114 | * Call function "fn" on each element of the array fragmented | ||
| 115 | * over a list of pages. | ||
| 116 | */ | ||
| 117 | static int traverse_pages(unsigned nelem, size_t size, | ||
| 118 | struct list_head *pos, | ||
| 119 | int (*fn)(void *data, void *state), | ||
| 120 | void *state) | ||
| 121 | { | ||
| 122 | void *pagedata; | ||
| 123 | unsigned pageidx; | ||
| 124 | int ret = 0; | ||
| 125 | |||
| 126 | BUG_ON(size > PAGE_SIZE); | ||
| 127 | |||
| 128 | pageidx = PAGE_SIZE; | ||
| 129 | pagedata = NULL; /* hush, gcc */ | ||
| 130 | |||
| 131 | while (nelem--) { | ||
| 132 | if (pageidx > PAGE_SIZE-size) { | ||
| 133 | struct page *page; | ||
| 134 | pos = pos->next; | ||
| 135 | page = list_entry(pos, struct page, lru); | ||
| 136 | pagedata = page_address(page); | ||
| 137 | pageidx = 0; | ||
| 138 | } | ||
| 139 | |||
| 140 | ret = (*fn)(pagedata + pageidx, state); | ||
| 141 | if (ret) | ||
| 142 | break; | ||
| 143 | pageidx += size; | ||
| 144 | } | ||
| 145 | |||
| 146 | return ret; | ||
| 147 | } | ||
| 148 | |||
| 149 | struct mmap_mfn_state { | ||
| 150 | unsigned long va; | ||
| 151 | struct vm_area_struct *vma; | ||
| 152 | domid_t domain; | ||
| 153 | }; | ||
| 154 | |||
| 155 | static int mmap_mfn_range(void *data, void *state) | ||
| 156 | { | ||
| 157 | struct privcmd_mmap_entry *msg = data; | ||
| 158 | struct mmap_mfn_state *st = state; | ||
| 159 | struct vm_area_struct *vma = st->vma; | ||
| 160 | int rc; | ||
| 161 | |||
| 162 | /* Do not allow range to wrap the address space. */ | ||
| 163 | if ((msg->npages > (LONG_MAX >> PAGE_SHIFT)) || | ||
| 164 | ((unsigned long)(msg->npages << PAGE_SHIFT) >= -st->va)) | ||
| 165 | return -EINVAL; | ||
| 166 | |||
| 167 | /* Range chunks must be contiguous in va space. */ | ||
| 168 | if ((msg->va != st->va) || | ||
| 169 | ((msg->va+(msg->npages<<PAGE_SHIFT)) > vma->vm_end)) | ||
| 170 | return -EINVAL; | ||
| 171 | |||
| 172 | rc = xen_remap_domain_mfn_range(vma, | ||
| 173 | msg->va & PAGE_MASK, | ||
| 174 | msg->mfn, msg->npages, | ||
| 175 | vma->vm_page_prot, | ||
| 176 | st->domain); | ||
| 177 | if (rc < 0) | ||
| 178 | return rc; | ||
| 179 | |||
| 180 | st->va += msg->npages << PAGE_SHIFT; | ||
| 181 | |||
| 182 | return 0; | ||
| 183 | } | ||
| 184 | |||
| 185 | static long privcmd_ioctl_mmap(void __user *udata) | ||
| 186 | { | ||
| 187 | struct privcmd_mmap mmapcmd; | ||
| 188 | struct mm_struct *mm = current->mm; | ||
| 189 | struct vm_area_struct *vma; | ||
| 190 | int rc; | ||
| 191 | LIST_HEAD(pagelist); | ||
| 192 | struct mmap_mfn_state state; | ||
| 193 | |||
| 194 | if (!xen_initial_domain()) | ||
| 195 | return -EPERM; | ||
| 196 | |||
| 197 | if (copy_from_user(&mmapcmd, udata, sizeof(mmapcmd))) | ||
| 198 | return -EFAULT; | ||
| 199 | |||
| 200 | rc = gather_array(&pagelist, | ||
| 201 | mmapcmd.num, sizeof(struct privcmd_mmap_entry), | ||
| 202 | mmapcmd.entry); | ||
| 203 | |||
| 204 | if (rc || list_empty(&pagelist)) | ||
| 205 | goto out; | ||
| 206 | |||
| 207 | down_write(&mm->mmap_sem); | ||
| 208 | |||
| 209 | { | ||
| 210 | struct page *page = list_first_entry(&pagelist, | ||
| 211 | struct page, lru); | ||
| 212 | struct privcmd_mmap_entry *msg = page_address(page); | ||
| 213 | |||
| 214 | vma = find_vma(mm, msg->va); | ||
| 215 | rc = -EINVAL; | ||
| 216 | |||
| 217 | if (!vma || (msg->va != vma->vm_start) || | ||
| 218 | !privcmd_enforce_singleshot_mapping(vma)) | ||
| 219 | goto out_up; | ||
| 220 | } | ||
| 221 | |||
| 222 | state.va = vma->vm_start; | ||
| 223 | state.vma = vma; | ||
| 224 | state.domain = mmapcmd.dom; | ||
| 225 | |||
| 226 | rc = traverse_pages(mmapcmd.num, sizeof(struct privcmd_mmap_entry), | ||
| 227 | &pagelist, | ||
| 228 | mmap_mfn_range, &state); | ||
| 229 | |||
| 230 | |||
| 231 | out_up: | ||
| 232 | up_write(&mm->mmap_sem); | ||
| 233 | |||
| 234 | out: | ||
| 235 | free_page_list(&pagelist); | ||
| 236 | |||
| 237 | return rc; | ||
| 238 | } | ||
| 239 | |||
| 240 | struct mmap_batch_state { | ||
| 241 | domid_t domain; | ||
| 242 | unsigned long va; | ||
| 243 | struct vm_area_struct *vma; | ||
| 244 | int err; | ||
| 245 | |||
| 246 | xen_pfn_t __user *user; | ||
| 247 | }; | ||
| 248 | |||
| 249 | static int mmap_batch_fn(void *data, void *state) | ||
| 250 | { | ||
| 251 | xen_pfn_t *mfnp = data; | ||
| 252 | struct mmap_batch_state *st = state; | ||
| 253 | |||
| 254 | if (xen_remap_domain_mfn_range(st->vma, st->va & PAGE_MASK, *mfnp, 1, | ||
| 255 | st->vma->vm_page_prot, st->domain) < 0) { | ||
| 256 | *mfnp |= 0xf0000000U; | ||
| 257 | st->err++; | ||
| 258 | } | ||
| 259 | st->va += PAGE_SIZE; | ||
| 260 | |||
| 261 | return 0; | ||
| 262 | } | ||
| 263 | |||
| 264 | static int mmap_return_errors(void *data, void *state) | ||
| 265 | { | ||
| 266 | xen_pfn_t *mfnp = data; | ||
| 267 | struct mmap_batch_state *st = state; | ||
| 268 | |||
| 269 | put_user(*mfnp, st->user++); | ||
| 270 | |||
| 271 | return 0; | ||
| 272 | } | ||
| 273 | |||
| 274 | static struct vm_operations_struct privcmd_vm_ops; | ||
| 275 | |||
| 276 | static long privcmd_ioctl_mmap_batch(void __user *udata) | ||
| 277 | { | ||
| 278 | int ret; | ||
| 279 | struct privcmd_mmapbatch m; | ||
| 280 | struct mm_struct *mm = current->mm; | ||
| 281 | struct vm_area_struct *vma; | ||
| 282 | unsigned long nr_pages; | ||
| 283 | LIST_HEAD(pagelist); | ||
| 284 | struct mmap_batch_state state; | ||
| 285 | |||
| 286 | if (!xen_initial_domain()) | ||
| 287 | return -EPERM; | ||
| 288 | |||
| 289 | if (copy_from_user(&m, udata, sizeof(m))) | ||
| 290 | return -EFAULT; | ||
| 291 | |||
| 292 | nr_pages = m.num; | ||
| 293 | if ((m.num <= 0) || (nr_pages > (LONG_MAX >> PAGE_SHIFT))) | ||
| 294 | return -EINVAL; | ||
| 295 | |||
| 296 | ret = gather_array(&pagelist, m.num, sizeof(xen_pfn_t), | ||
| 297 | m.arr); | ||
| 298 | |||
| 299 | if (ret || list_empty(&pagelist)) | ||
| 300 | goto out; | ||
| 301 | |||
| 302 | down_write(&mm->mmap_sem); | ||
| 303 | |||
| 304 | vma = find_vma(mm, m.addr); | ||
| 305 | ret = -EINVAL; | ||
| 306 | if (!vma || | ||
| 307 | vma->vm_ops != &privcmd_vm_ops || | ||
| 308 | (m.addr != vma->vm_start) || | ||
| 309 | ((m.addr + (nr_pages << PAGE_SHIFT)) != vma->vm_end) || | ||
| 310 | !privcmd_enforce_singleshot_mapping(vma)) { | ||
| 311 | up_write(&mm->mmap_sem); | ||
| 312 | goto out; | ||
| 313 | } | ||
| 314 | |||
| 315 | state.domain = m.dom; | ||
| 316 | state.vma = vma; | ||
| 317 | state.va = m.addr; | ||
| 318 | state.err = 0; | ||
| 319 | |||
| 320 | ret = traverse_pages(m.num, sizeof(xen_pfn_t), | ||
| 321 | &pagelist, mmap_batch_fn, &state); | ||
| 322 | |||
| 323 | up_write(&mm->mmap_sem); | ||
| 324 | |||
| 325 | if (state.err > 0) { | ||
| 326 | ret = 0; | ||
| 327 | |||
| 328 | state.user = m.arr; | ||
| 329 | traverse_pages(m.num, sizeof(xen_pfn_t), | ||
| 330 | &pagelist, | ||
| 331 | mmap_return_errors, &state); | ||
| 332 | } | ||
| 333 | |||
| 334 | out: | ||
| 335 | free_page_list(&pagelist); | ||
| 336 | |||
| 337 | return ret; | ||
| 338 | } | ||
| 339 | |||
| 340 | static long privcmd_ioctl(struct file *file, | ||
| 341 | unsigned int cmd, unsigned long data) | ||
| 342 | { | ||
| 343 | int ret = -ENOSYS; | ||
| 344 | void __user *udata = (void __user *) data; | ||
| 345 | |||
| 346 | switch (cmd) { | ||
| 347 | case IOCTL_PRIVCMD_HYPERCALL: | ||
| 348 | ret = privcmd_ioctl_hypercall(udata); | ||
| 349 | break; | ||
| 350 | |||
| 351 | case IOCTL_PRIVCMD_MMAP: | ||
| 352 | ret = privcmd_ioctl_mmap(udata); | ||
| 353 | break; | ||
| 354 | |||
| 355 | case IOCTL_PRIVCMD_MMAPBATCH: | ||
| 356 | ret = privcmd_ioctl_mmap_batch(udata); | ||
| 357 | break; | ||
| 358 | |||
| 359 | default: | ||
| 360 | ret = -EINVAL; | ||
| 361 | break; | ||
| 362 | } | ||
| 363 | |||
| 364 | return ret; | ||
| 365 | } | ||
| 366 | |||
| 367 | #ifndef HAVE_ARCH_PRIVCMD_MMAP | ||
| 368 | static int privcmd_fault(struct vm_area_struct *vma, struct vm_fault *vmf) | ||
| 369 | { | ||
| 370 | printk(KERN_DEBUG "privcmd_fault: vma=%p %lx-%lx, pgoff=%lx, uv=%p\n", | ||
| 371 | vma, vma->vm_start, vma->vm_end, | ||
| 372 | vmf->pgoff, vmf->virtual_address); | ||
| 373 | |||
| 374 | return VM_FAULT_SIGBUS; | ||
| 375 | } | ||
| 376 | |||
| 377 | static struct vm_operations_struct privcmd_vm_ops = { | ||
| 378 | .fault = privcmd_fault | ||
| 379 | }; | ||
| 380 | |||
| 381 | static int privcmd_mmap(struct file *file, struct vm_area_struct *vma) | ||
| 382 | { | ||
| 383 | /* Unsupported for auto-translate guests. */ | ||
| 384 | if (xen_feature(XENFEAT_auto_translated_physmap)) | ||
| 385 | return -ENOSYS; | ||
| 386 | |||
| 387 | /* DONTCOPY is essential for Xen as copy_page_range is broken. */ | ||
| 388 | vma->vm_flags |= VM_RESERVED | VM_IO | VM_DONTCOPY; | ||
| 389 | vma->vm_ops = &privcmd_vm_ops; | ||
| 390 | vma->vm_private_data = NULL; | ||
| 391 | |||
| 392 | return 0; | ||
| 393 | } | ||
| 394 | |||
| 395 | static int privcmd_enforce_singleshot_mapping(struct vm_area_struct *vma) | ||
| 396 | { | ||
| 397 | return (xchg(&vma->vm_private_data, (void *)1) == NULL); | ||
| 398 | } | ||
| 399 | #endif | ||
| 400 | |||
| 401 | const struct file_operations privcmd_file_ops = { | ||
| 402 | .unlocked_ioctl = privcmd_ioctl, | ||
| 403 | .mmap = privcmd_mmap, | ||
| 404 | }; | ||
