diff options
Diffstat (limited to 'drivers/xen')
| -rw-r--r-- | drivers/xen/xenfs/privcmd.c | 400 | ||||
| -rw-r--r-- | drivers/xen/xenfs/xenbus.c | 593 |
2 files changed, 993 insertions, 0 deletions
diff --git a/drivers/xen/xenfs/privcmd.c b/drivers/xen/xenfs/privcmd.c new file mode 100644 index 00000000000..dbd3b16fd13 --- /dev/null +++ b/drivers/xen/xenfs/privcmd.c | |||
| @@ -0,0 +1,400 @@ | |||
| 1 | /****************************************************************************** | ||
| 2 | * privcmd.c | ||
| 3 | * | ||
| 4 | * Interface to privileged domain-0 commands. | ||
| 5 | * | ||
| 6 | * Copyright (c) 2002-2004, K A Fraser, B Dragovic | ||
| 7 | */ | ||
| 8 | |||
| 9 | #include <linux/kernel.h> | ||
| 10 | #include <linux/sched.h> | ||
| 11 | #include <linux/slab.h> | ||
| 12 | #include <linux/string.h> | ||
| 13 | #include <linux/errno.h> | ||
| 14 | #include <linux/mm.h> | ||
| 15 | #include <linux/mman.h> | ||
| 16 | #include <linux/uaccess.h> | ||
| 17 | #include <linux/swap.h> | ||
| 18 | #include <linux/highmem.h> | ||
| 19 | #include <linux/pagemap.h> | ||
| 20 | #include <linux/seq_file.h> | ||
| 21 | |||
| 22 | #include <asm/pgalloc.h> | ||
| 23 | #include <asm/pgtable.h> | ||
| 24 | #include <asm/tlb.h> | ||
| 25 | #include <asm/xen/hypervisor.h> | ||
| 26 | #include <asm/xen/hypercall.h> | ||
| 27 | |||
| 28 | #include <xen/xen.h> | ||
| 29 | #include <xen/privcmd.h> | ||
| 30 | #include <xen/interface/xen.h> | ||
| 31 | #include <xen/features.h> | ||
| 32 | #include <xen/page.h> | ||
| 33 | #include <xen/xen-ops.h> | ||
| 34 | |||
| 35 | #ifndef HAVE_ARCH_PRIVCMD_MMAP | ||
| 36 | static int privcmd_enforce_singleshot_mapping(struct vm_area_struct *vma); | ||
| 37 | #endif | ||
| 38 | |||
| 39 | static long privcmd_ioctl_hypercall(void __user *udata) | ||
| 40 | { | ||
| 41 | struct privcmd_hypercall hypercall; | ||
| 42 | long ret; | ||
| 43 | |||
| 44 | if (copy_from_user(&hypercall, udata, sizeof(hypercall))) | ||
| 45 | return -EFAULT; | ||
| 46 | |||
| 47 | ret = privcmd_call(hypercall.op, | ||
| 48 | hypercall.arg[0], hypercall.arg[1], | ||
| 49 | hypercall.arg[2], hypercall.arg[3], | ||
| 50 | hypercall.arg[4]); | ||
| 51 | |||
| 52 | return ret; | ||
| 53 | } | ||
| 54 | |||
| 55 | static void free_page_list(struct list_head *pages) | ||
| 56 | { | ||
| 57 | struct page *p, *n; | ||
| 58 | |||
| 59 | list_for_each_entry_safe(p, n, pages, lru) | ||
| 60 | __free_page(p); | ||
| 61 | |||
| 62 | INIT_LIST_HEAD(pages); | ||
| 63 | } | ||
| 64 | |||
| 65 | /* | ||
| 66 | * Given an array of items in userspace, return a list of pages | ||
| 67 | * containing the data. If copying fails, either because of memory | ||
| 68 | * allocation failure or a problem reading user memory, return an | ||
| 69 | * error code; its up to the caller to dispose of any partial list. | ||
| 70 | */ | ||
| 71 | static int gather_array(struct list_head *pagelist, | ||
| 72 | unsigned nelem, size_t size, | ||
| 73 | void __user *data) | ||
| 74 | { | ||
| 75 | unsigned pageidx; | ||
| 76 | void *pagedata; | ||
| 77 | int ret; | ||
| 78 | |||
| 79 | if (size > PAGE_SIZE) | ||
| 80 | return 0; | ||
| 81 | |||
| 82 | pageidx = PAGE_SIZE; | ||
| 83 | pagedata = NULL; /* quiet, gcc */ | ||
| 84 | while (nelem--) { | ||
| 85 | if (pageidx > PAGE_SIZE-size) { | ||
| 86 | struct page *page = alloc_page(GFP_KERNEL); | ||
| 87 | |||
| 88 | ret = -ENOMEM; | ||
| 89 | if (page == NULL) | ||
| 90 | goto fail; | ||
| 91 | |||
| 92 | pagedata = page_address(page); | ||
| 93 | |||
| 94 | list_add_tail(&page->lru, pagelist); | ||
| 95 | pageidx = 0; | ||
| 96 | } | ||
| 97 | |||
| 98 | ret = -EFAULT; | ||
| 99 | if (copy_from_user(pagedata + pageidx, data, size)) | ||
| 100 | goto fail; | ||
| 101 | |||
| 102 | data += size; | ||
| 103 | pageidx += size; | ||
| 104 | } | ||
| 105 | |||
| 106 | ret = 0; | ||
| 107 | |||
| 108 | fail: | ||
| 109 | return ret; | ||
| 110 | } | ||
| 111 | |||
| 112 | /* | ||
| 113 | * Call function "fn" on each element of the array fragmented | ||
| 114 | * over a list of pages. | ||
| 115 | */ | ||
| 116 | static int traverse_pages(unsigned nelem, size_t size, | ||
| 117 | struct list_head *pos, | ||
| 118 | int (*fn)(void *data, void *state), | ||
| 119 | void *state) | ||
| 120 | { | ||
| 121 | void *pagedata; | ||
| 122 | unsigned pageidx; | ||
| 123 | int ret = 0; | ||
| 124 | |||
| 125 | BUG_ON(size > PAGE_SIZE); | ||
| 126 | |||
| 127 | pageidx = PAGE_SIZE; | ||
| 128 | pagedata = NULL; /* hush, gcc */ | ||
| 129 | |||
| 130 | while (nelem--) { | ||
| 131 | if (pageidx > PAGE_SIZE-size) { | ||
| 132 | struct page *page; | ||
| 133 | pos = pos->next; | ||
| 134 | page = list_entry(pos, struct page, lru); | ||
| 135 | pagedata = page_address(page); | ||
| 136 | pageidx = 0; | ||
| 137 | } | ||
| 138 | |||
| 139 | ret = (*fn)(pagedata + pageidx, state); | ||
| 140 | if (ret) | ||
| 141 | break; | ||
| 142 | pageidx += size; | ||
| 143 | } | ||
| 144 | |||
| 145 | return ret; | ||
| 146 | } | ||
| 147 | |||
| 148 | struct mmap_mfn_state { | ||
| 149 | unsigned long va; | ||
| 150 | struct vm_area_struct *vma; | ||
| 151 | domid_t domain; | ||
| 152 | }; | ||
| 153 | |||
| 154 | static int mmap_mfn_range(void *data, void *state) | ||
| 155 | { | ||
| 156 | struct privcmd_mmap_entry *msg = data; | ||
| 157 | struct mmap_mfn_state *st = state; | ||
| 158 | struct vm_area_struct *vma = st->vma; | ||
| 159 | int rc; | ||
| 160 | |||
| 161 | /* Do not allow range to wrap the address space. */ | ||
| 162 | if ((msg->npages > (LONG_MAX >> PAGE_SHIFT)) || | ||
| 163 | ((unsigned long)(msg->npages << PAGE_SHIFT) >= -st->va)) | ||
| 164 | return -EINVAL; | ||
| 165 | |||
| 166 | /* Range chunks must be contiguous in va space. */ | ||
| 167 | if ((msg->va != st->va) || | ||
| 168 | ((msg->va+(msg->npages<<PAGE_SHIFT)) > vma->vm_end)) | ||
| 169 | return -EINVAL; | ||
| 170 | |||
| 171 | rc = xen_remap_domain_mfn_range(vma, | ||
| 172 | msg->va & PAGE_MASK, | ||
| 173 | msg->mfn, msg->npages, | ||
| 174 | vma->vm_page_prot, | ||
| 175 | st->domain); | ||
| 176 | if (rc < 0) | ||
| 177 | return rc; | ||
| 178 | |||
| 179 | st->va += msg->npages << PAGE_SHIFT; | ||
| 180 | |||
| 181 | return 0; | ||
| 182 | } | ||
| 183 | |||
| 184 | static long privcmd_ioctl_mmap(void __user *udata) | ||
| 185 | { | ||
| 186 | struct privcmd_mmap mmapcmd; | ||
| 187 | struct mm_struct *mm = current->mm; | ||
| 188 | struct vm_area_struct *vma; | ||
| 189 | int rc; | ||
| 190 | LIST_HEAD(pagelist); | ||
| 191 | struct mmap_mfn_state state; | ||
| 192 | |||
| 193 | if (!xen_initial_domain()) | ||
| 194 | return -EPERM; | ||
| 195 | |||
| 196 | if (copy_from_user(&mmapcmd, udata, sizeof(mmapcmd))) | ||
| 197 | return -EFAULT; | ||
| 198 | |||
| 199 | rc = gather_array(&pagelist, | ||
| 200 | mmapcmd.num, sizeof(struct privcmd_mmap_entry), | ||
| 201 | mmapcmd.entry); | ||
| 202 | |||
| 203 | if (rc || list_empty(&pagelist)) | ||
| 204 | goto out; | ||
| 205 | |||
| 206 | down_write(&mm->mmap_sem); | ||
| 207 | |||
| 208 | { | ||
| 209 | struct page *page = list_first_entry(&pagelist, | ||
| 210 | struct page, lru); | ||
| 211 | struct privcmd_mmap_entry *msg = page_address(page); | ||
| 212 | |||
| 213 | vma = find_vma(mm, msg->va); | ||
| 214 | rc = -EINVAL; | ||
| 215 | |||
| 216 | if (!vma || (msg->va != vma->vm_start) || | ||
| 217 | !privcmd_enforce_singleshot_mapping(vma)) | ||
| 218 | goto out_up; | ||
| 219 | } | ||
| 220 | |||
| 221 | state.va = vma->vm_start; | ||
| 222 | state.vma = vma; | ||
| 223 | state.domain = mmapcmd.dom; | ||
| 224 | |||
| 225 | rc = traverse_pages(mmapcmd.num, sizeof(struct privcmd_mmap_entry), | ||
| 226 | &pagelist, | ||
| 227 | mmap_mfn_range, &state); | ||
| 228 | |||
| 229 | |||
| 230 | out_up: | ||
| 231 | up_write(&mm->mmap_sem); | ||
| 232 | |||
| 233 | out: | ||
| 234 | free_page_list(&pagelist); | ||
| 235 | |||
| 236 | return rc; | ||
| 237 | } | ||
| 238 | |||
| 239 | struct mmap_batch_state { | ||
| 240 | domid_t domain; | ||
| 241 | unsigned long va; | ||
| 242 | struct vm_area_struct *vma; | ||
| 243 | int err; | ||
| 244 | |||
| 245 | xen_pfn_t __user *user; | ||
| 246 | }; | ||
| 247 | |||
| 248 | static int mmap_batch_fn(void *data, void *state) | ||
| 249 | { | ||
| 250 | xen_pfn_t *mfnp = data; | ||
| 251 | struct mmap_batch_state *st = state; | ||
| 252 | |||
| 253 | if (xen_remap_domain_mfn_range(st->vma, st->va & PAGE_MASK, *mfnp, 1, | ||
| 254 | st->vma->vm_page_prot, st->domain) < 0) { | ||
| 255 | *mfnp |= 0xf0000000U; | ||
| 256 | st->err++; | ||
| 257 | } | ||
| 258 | st->va += PAGE_SIZE; | ||
| 259 | |||
| 260 | return 0; | ||
| 261 | } | ||
| 262 | |||
| 263 | static int mmap_return_errors(void *data, void *state) | ||
| 264 | { | ||
| 265 | xen_pfn_t *mfnp = data; | ||
| 266 | struct mmap_batch_state *st = state; | ||
| 267 | |||
| 268 | return put_user(*mfnp, st->user++); | ||
| 269 | } | ||
| 270 | |||
| 271 | static struct vm_operations_struct privcmd_vm_ops; | ||
| 272 | |||
| 273 | static long privcmd_ioctl_mmap_batch(void __user *udata) | ||
| 274 | { | ||
| 275 | int ret; | ||
| 276 | struct privcmd_mmapbatch m; | ||
| 277 | struct mm_struct *mm = current->mm; | ||
| 278 | struct vm_area_struct *vma; | ||
| 279 | unsigned long nr_pages; | ||
| 280 | LIST_HEAD(pagelist); | ||
| 281 | struct mmap_batch_state state; | ||
| 282 | |||
| 283 | if (!xen_initial_domain()) | ||
| 284 | return -EPERM; | ||
| 285 | |||
| 286 | if (copy_from_user(&m, udata, sizeof(m))) | ||
| 287 | return -EFAULT; | ||
| 288 | |||
| 289 | nr_pages = m.num; | ||
| 290 | if ((m.num <= 0) || (nr_pages > (LONG_MAX >> PAGE_SHIFT))) | ||
| 291 | return -EINVAL; | ||
| 292 | |||
| 293 | ret = gather_array(&pagelist, m.num, sizeof(xen_pfn_t), | ||
| 294 | m.arr); | ||
| 295 | |||
| 296 | if (ret || list_empty(&pagelist)) | ||
| 297 | goto out; | ||
| 298 | |||
| 299 | down_write(&mm->mmap_sem); | ||
| 300 | |||
| 301 | vma = find_vma(mm, m.addr); | ||
| 302 | ret = -EINVAL; | ||
| 303 | if (!vma || | ||
| 304 | vma->vm_ops != &privcmd_vm_ops || | ||
| 305 | (m.addr != vma->vm_start) || | ||
| 306 | ((m.addr + (nr_pages << PAGE_SHIFT)) != vma->vm_end) || | ||
| 307 | !privcmd_enforce_singleshot_mapping(vma)) { | ||
| 308 | up_write(&mm->mmap_sem); | ||
| 309 | goto out; | ||
| 310 | } | ||
| 311 | |||
| 312 | state.domain = m.dom; | ||
| 313 | state.vma = vma; | ||
| 314 | state.va = m.addr; | ||
| 315 | state.err = 0; | ||
| 316 | |||
| 317 | ret = traverse_pages(m.num, sizeof(xen_pfn_t), | ||
| 318 | &pagelist, mmap_batch_fn, &state); | ||
| 319 | |||
| 320 | up_write(&mm->mmap_sem); | ||
| 321 | |||
| 322 | if (state.err > 0) { | ||
| 323 | state.user = m.arr; | ||
| 324 | ret = traverse_pages(m.num, sizeof(xen_pfn_t), | ||
| 325 | &pagelist, | ||
| 326 | mmap_return_errors, &state); | ||
| 327 | } | ||
| 328 | |||
| 329 | out: | ||
| 330 | free_page_list(&pagelist); | ||
| 331 | |||
| 332 | return ret; | ||
| 333 | } | ||
| 334 | |||
| 335 | static long privcmd_ioctl(struct file *file, | ||
| 336 | unsigned int cmd, unsigned long data) | ||
| 337 | { | ||
| 338 | int ret = -ENOSYS; | ||
| 339 | void __user *udata = (void __user *) data; | ||
| 340 | |||
| 341 | switch (cmd) { | ||
| 342 | case IOCTL_PRIVCMD_HYPERCALL: | ||
| 343 | ret = privcmd_ioctl_hypercall(udata); | ||
| 344 | break; | ||
| 345 | |||
| 346 | case IOCTL_PRIVCMD_MMAP: | ||
| 347 | ret = privcmd_ioctl_mmap(udata); | ||
| 348 | break; | ||
| 349 | |||
| 350 | case IOCTL_PRIVCMD_MMAPBATCH: | ||
| 351 | ret = privcmd_ioctl_mmap_batch(udata); | ||
| 352 | break; | ||
| 353 | |||
| 354 | default: | ||
| 355 | ret = -EINVAL; | ||
| 356 | break; | ||
| 357 | } | ||
| 358 | |||
| 359 | return ret; | ||
| 360 | } | ||
| 361 | |||
| 362 | #ifndef HAVE_ARCH_PRIVCMD_MMAP | ||
| 363 | static int privcmd_fault(struct vm_area_struct *vma, struct vm_fault *vmf) | ||
| 364 | { | ||
| 365 | printk(KERN_DEBUG "privcmd_fault: vma=%p %lx-%lx, pgoff=%lx, uv=%p\n", | ||
| 366 | vma, vma->vm_start, vma->vm_end, | ||
| 367 | vmf->pgoff, vmf->virtual_address); | ||
| 368 | |||
| 369 | return VM_FAULT_SIGBUS; | ||
| 370 | } | ||
| 371 | |||
| 372 | static struct vm_operations_struct privcmd_vm_ops = { | ||
| 373 | .fault = privcmd_fault | ||
| 374 | }; | ||
| 375 | |||
| 376 | static int privcmd_mmap(struct file *file, struct vm_area_struct *vma) | ||
| 377 | { | ||
| 378 | /* Unsupported for auto-translate guests. */ | ||
| 379 | if (xen_feature(XENFEAT_auto_translated_physmap)) | ||
| 380 | return -ENOSYS; | ||
| 381 | |||
| 382 | /* DONTCOPY is essential for Xen because copy_page_range doesn't know | ||
| 383 | * how to recreate these mappings */ | ||
| 384 | vma->vm_flags |= VM_RESERVED | VM_IO | VM_DONTCOPY | VM_PFNMAP; | ||
| 385 | vma->vm_ops = &privcmd_vm_ops; | ||
| 386 | vma->vm_private_data = NULL; | ||
| 387 | |||
| 388 | return 0; | ||
| 389 | } | ||
| 390 | |||
| 391 | static int privcmd_enforce_singleshot_mapping(struct vm_area_struct *vma) | ||
| 392 | { | ||
| 393 | return (xchg(&vma->vm_private_data, (void *)1) == NULL); | ||
| 394 | } | ||
| 395 | #endif | ||
| 396 | |||
| 397 | const struct file_operations privcmd_file_ops = { | ||
| 398 | .unlocked_ioctl = privcmd_ioctl, | ||
| 399 | .mmap = privcmd_mmap, | ||
| 400 | }; | ||
diff --git a/drivers/xen/xenfs/xenbus.c b/drivers/xen/xenfs/xenbus.c new file mode 100644 index 00000000000..bbd000f88af --- /dev/null +++ b/drivers/xen/xenfs/xenbus.c | |||
| @@ -0,0 +1,593 @@ | |||
| 1 | /* | ||
| 2 | * Driver giving user-space access to the kernel's xenbus connection | ||
| 3 | * to xenstore. | ||
| 4 | * | ||
| 5 | * Copyright (c) 2005, Christian Limpach | ||
| 6 | * Copyright (c) 2005, Rusty Russell, IBM Corporation | ||
| 7 | * | ||
| 8 | * This program is free software; you can redistribute it and/or | ||
| 9 | * modify it under the terms of the GNU General Public License version 2 | ||
| 10 | * as published by the Free Software Foundation; or, when distributed | ||
| 11 | * separately from the Linux kernel or incorporated into other | ||
| 12 | * software packages, subject to the following license: | ||
| 13 | * | ||
| 14 | * Permission is hereby granted, free of charge, to any person obtaining a copy | ||
| 15 | * of this source file (the "Software"), to deal in the Software without | ||
| 16 | * restriction, including without limitation the rights to use, copy, modify, | ||
| 17 | * merge, publish, distribute, sublicense, and/or sell copies of the Software, | ||
| 18 | * and to permit persons to whom the Software is furnished to do so, subject to | ||
| 19 | * the following conditions: | ||
| 20 | * | ||
| 21 | * The above copyright notice and this permission notice shall be included in | ||
| 22 | * all copies or substantial portions of the Software. | ||
| 23 | * | ||
| 24 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | ||
| 25 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | ||
| 26 | * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE | ||
| 27 | * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | ||
| 28 | * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING | ||
| 29 | * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS | ||
| 30 | * IN THE SOFTWARE. | ||
| 31 | * | ||
| 32 | * Changes: | ||
| 33 | * 2008-10-07 Alex Zeffertt Replaced /proc/xen/xenbus with xenfs filesystem | ||
| 34 | * and /proc/xen compatibility mount point. | ||
| 35 | * Turned xenfs into a loadable module. | ||
| 36 | */ | ||
| 37 | |||
| 38 | #include <linux/kernel.h> | ||
| 39 | #include <linux/errno.h> | ||
| 40 | #include <linux/uio.h> | ||
| 41 | #include <linux/notifier.h> | ||
| 42 | #include <linux/wait.h> | ||
| 43 | #include <linux/fs.h> | ||
| 44 | #include <linux/poll.h> | ||
| 45 | #include <linux/mutex.h> | ||
| 46 | #include <linux/sched.h> | ||
| 47 | #include <linux/spinlock.h> | ||
| 48 | #include <linux/mount.h> | ||
| 49 | #include <linux/pagemap.h> | ||
| 50 | #include <linux/uaccess.h> | ||
| 51 | #include <linux/init.h> | ||
| 52 | #include <linux/namei.h> | ||
| 53 | #include <linux/string.h> | ||
| 54 | #include <linux/slab.h> | ||
| 55 | |||
| 56 | #include "xenfs.h" | ||
| 57 | #include "../xenbus/xenbus_comms.h" | ||
| 58 | |||
| 59 | #include <xen/xenbus.h> | ||
| 60 | #include <asm/xen/hypervisor.h> | ||
| 61 | |||
| 62 | /* | ||
| 63 | * An element of a list of outstanding transactions, for which we're | ||
| 64 | * still waiting a reply. | ||
| 65 | */ | ||
| 66 | struct xenbus_transaction_holder { | ||
| 67 | struct list_head list; | ||
| 68 | struct xenbus_transaction handle; | ||
| 69 | }; | ||
| 70 | |||
| 71 | /* | ||
| 72 | * A buffer of data on the queue. | ||
| 73 | */ | ||
| 74 | struct read_buffer { | ||
| 75 | struct list_head list; | ||
| 76 | unsigned int cons; | ||
| 77 | unsigned int len; | ||
| 78 | char msg[]; | ||
| 79 | }; | ||
| 80 | |||
| 81 | struct xenbus_file_priv { | ||
| 82 | /* | ||
| 83 | * msgbuffer_mutex is held while partial requests are built up | ||
| 84 | * and complete requests are acted on. It therefore protects | ||
| 85 | * the "transactions" and "watches" lists, and the partial | ||
| 86 | * request length and buffer. | ||
| 87 | * | ||
| 88 | * reply_mutex protects the reply being built up to return to | ||
| 89 | * usermode. It nests inside msgbuffer_mutex but may be held | ||
| 90 | * alone during a watch callback. | ||
| 91 | */ | ||
| 92 | struct mutex msgbuffer_mutex; | ||
| 93 | |||
| 94 | /* In-progress transactions */ | ||
| 95 | struct list_head transactions; | ||
| 96 | |||
| 97 | /* Active watches. */ | ||
| 98 | struct list_head watches; | ||
| 99 | |||
| 100 | /* Partial request. */ | ||
| 101 | unsigned int len; | ||
| 102 | union { | ||
| 103 | struct xsd_sockmsg msg; | ||
| 104 | char buffer[PAGE_SIZE]; | ||
| 105 | } u; | ||
| 106 | |||
| 107 | /* Response queue. */ | ||
| 108 | struct mutex reply_mutex; | ||
| 109 | struct list_head read_buffers; | ||
| 110 | wait_queue_head_t read_waitq; | ||
| 111 | |||
| 112 | }; | ||
| 113 | |||
| 114 | /* Read out any raw xenbus messages queued up. */ | ||
| 115 | static ssize_t xenbus_file_read(struct file *filp, | ||
| 116 | char __user *ubuf, | ||
| 117 | size_t len, loff_t *ppos) | ||
| 118 | { | ||
| 119 | struct xenbus_file_priv *u = filp->private_data; | ||
| 120 | struct read_buffer *rb; | ||
| 121 | unsigned i; | ||
| 122 | int ret; | ||
| 123 | |||
| 124 | mutex_lock(&u->reply_mutex); | ||
| 125 | again: | ||
| 126 | while (list_empty(&u->read_buffers)) { | ||
| 127 | mutex_unlock(&u->reply_mutex); | ||
| 128 | if (filp->f_flags & O_NONBLOCK) | ||
| 129 | return -EAGAIN; | ||
| 130 | |||
| 131 | ret = wait_event_interruptible(u->read_waitq, | ||
| 132 | !list_empty(&u->read_buffers)); | ||
| 133 | if (ret) | ||
| 134 | return ret; | ||
| 135 | mutex_lock(&u->reply_mutex); | ||
| 136 | } | ||
| 137 | |||
| 138 | rb = list_entry(u->read_buffers.next, struct read_buffer, list); | ||
| 139 | i = 0; | ||
| 140 | while (i < len) { | ||
| 141 | unsigned sz = min((unsigned)len - i, rb->len - rb->cons); | ||
| 142 | |||
| 143 | ret = copy_to_user(ubuf + i, &rb->msg[rb->cons], sz); | ||
| 144 | |||
| 145 | i += sz - ret; | ||
| 146 | rb->cons += sz - ret; | ||
| 147 | |||
| 148 | if (ret != 0) { | ||
| 149 | if (i == 0) | ||
| 150 | i = -EFAULT; | ||
| 151 | goto out; | ||
| 152 | } | ||
| 153 | |||
| 154 | /* Clear out buffer if it has been consumed */ | ||
| 155 | if (rb->cons == rb->len) { | ||
| 156 | list_del(&rb->list); | ||
| 157 | kfree(rb); | ||
| 158 | if (list_empty(&u->read_buffers)) | ||
| 159 | break; | ||
| 160 | rb = list_entry(u->read_buffers.next, | ||
| 161 | struct read_buffer, list); | ||
| 162 | } | ||
| 163 | } | ||
| 164 | if (i == 0) | ||
| 165 | goto again; | ||
| 166 | |||
| 167 | out: | ||
| 168 | mutex_unlock(&u->reply_mutex); | ||
| 169 | return i; | ||
| 170 | } | ||
| 171 | |||
| 172 | /* | ||
| 173 | * Add a buffer to the queue. Caller must hold the appropriate lock | ||
| 174 | * if the queue is not local. (Commonly the caller will build up | ||
| 175 | * multiple queued buffers on a temporary local list, and then add it | ||
| 176 | * to the appropriate list under lock once all the buffers have een | ||
| 177 | * successfully allocated.) | ||
| 178 | */ | ||
| 179 | static int queue_reply(struct list_head *queue, const void *data, size_t len) | ||
| 180 | { | ||
| 181 | struct read_buffer *rb; | ||
| 182 | |||
| 183 | if (len == 0) | ||
| 184 | return 0; | ||
| 185 | |||
| 186 | rb = kmalloc(sizeof(*rb) + len, GFP_KERNEL); | ||
| 187 | if (rb == NULL) | ||
| 188 | return -ENOMEM; | ||
| 189 | |||
| 190 | rb->cons = 0; | ||
| 191 | rb->len = len; | ||
| 192 | |||
| 193 | memcpy(rb->msg, data, len); | ||
| 194 | |||
| 195 | list_add_tail(&rb->list, queue); | ||
| 196 | return 0; | ||
| 197 | } | ||
| 198 | |||
| 199 | /* | ||
| 200 | * Free all the read_buffer s on a list. | ||
| 201 | * Caller must have sole reference to list. | ||
| 202 | */ | ||
| 203 | static void queue_cleanup(struct list_head *list) | ||
| 204 | { | ||
| 205 | struct read_buffer *rb; | ||
| 206 | |||
| 207 | while (!list_empty(list)) { | ||
| 208 | rb = list_entry(list->next, struct read_buffer, list); | ||
| 209 | list_del(list->next); | ||
| 210 | kfree(rb); | ||
| 211 | } | ||
| 212 | } | ||
| 213 | |||
| 214 | struct watch_adapter { | ||
| 215 | struct list_head list; | ||
| 216 | struct xenbus_watch watch; | ||
| 217 | struct xenbus_file_priv *dev_data; | ||
| 218 | char *token; | ||
| 219 | }; | ||
| 220 | |||
| 221 | static void free_watch_adapter(struct watch_adapter *watch) | ||
| 222 | { | ||
| 223 | kfree(watch->watch.node); | ||
| 224 | kfree(watch->token); | ||
| 225 | kfree(watch); | ||
| 226 | } | ||
| 227 | |||
| 228 | static struct watch_adapter *alloc_watch_adapter(const char *path, | ||
| 229 | const char *token) | ||
| 230 | { | ||
| 231 | struct watch_adapter *watch; | ||
| 232 | |||
| 233 | watch = kzalloc(sizeof(*watch), GFP_KERNEL); | ||
| 234 | if (watch == NULL) | ||
| 235 | goto out_fail; | ||
| 236 | |||
| 237 | watch->watch.node = kstrdup(path, GFP_KERNEL); | ||
| 238 | if (watch->watch.node == NULL) | ||
| 239 | goto out_free; | ||
| 240 | |||
| 241 | watch->token = kstrdup(token, GFP_KERNEL); | ||
| 242 | if (watch->token == NULL) | ||
| 243 | goto out_free; | ||
| 244 | |||
| 245 | return watch; | ||
| 246 | |||
| 247 | out_free: | ||
| 248 | free_watch_adapter(watch); | ||
| 249 | |||
| 250 | out_fail: | ||
| 251 | return NULL; | ||
| 252 | } | ||
| 253 | |||
| 254 | static void watch_fired(struct xenbus_watch *watch, | ||
| 255 | const char **vec, | ||
| 256 | unsigned int len) | ||
| 257 | { | ||
| 258 | struct watch_adapter *adap; | ||
| 259 | struct xsd_sockmsg hdr; | ||
| 260 | const char *path, *token; | ||
| 261 | int path_len, tok_len, body_len, data_len = 0; | ||
| 262 | int ret; | ||
| 263 | LIST_HEAD(staging_q); | ||
| 264 | |||
| 265 | adap = container_of(watch, struct watch_adapter, watch); | ||
| 266 | |||
| 267 | path = vec[XS_WATCH_PATH]; | ||
| 268 | token = adap->token; | ||
| 269 | |||
| 270 | path_len = strlen(path) + 1; | ||
| 271 | tok_len = strlen(token) + 1; | ||
| 272 | if (len > 2) | ||
| 273 | data_len = vec[len] - vec[2] + 1; | ||
| 274 | body_len = path_len + tok_len + data_len; | ||
| 275 | |||
| 276 | hdr.type = XS_WATCH_EVENT; | ||
| 277 | hdr.len = body_len; | ||
| 278 | |||
| 279 | mutex_lock(&adap->dev_data->reply_mutex); | ||
| 280 | |||
| 281 | ret = queue_reply(&staging_q, &hdr, sizeof(hdr)); | ||
| 282 | if (!ret) | ||
| 283 | ret = queue_reply(&staging_q, path, path_len); | ||
| 284 | if (!ret) | ||
| 285 | ret = queue_reply(&staging_q, token, tok_len); | ||
| 286 | if (!ret && len > 2) | ||
| 287 | ret = queue_reply(&staging_q, vec[2], data_len); | ||
| 288 | |||
| 289 | if (!ret) { | ||
| 290 | /* success: pass reply list onto watcher */ | ||
| 291 | list_splice_tail(&staging_q, &adap->dev_data->read_buffers); | ||
| 292 | wake_up(&adap->dev_data->read_waitq); | ||
| 293 | } else | ||
| 294 | queue_cleanup(&staging_q); | ||
| 295 | |||
| 296 | mutex_unlock(&adap->dev_data->reply_mutex); | ||
| 297 | } | ||
| 298 | |||
| 299 | static int xenbus_write_transaction(unsigned msg_type, | ||
| 300 | struct xenbus_file_priv *u) | ||
| 301 | { | ||
| 302 | int rc; | ||
| 303 | void *reply; | ||
| 304 | struct xenbus_transaction_holder *trans = NULL; | ||
| 305 | LIST_HEAD(staging_q); | ||
| 306 | |||
| 307 | if (msg_type == XS_TRANSACTION_START) { | ||
| 308 | trans = kmalloc(sizeof(*trans), GFP_KERNEL); | ||
| 309 | if (!trans) { | ||
| 310 | rc = -ENOMEM; | ||
| 311 | goto out; | ||
| 312 | } | ||
| 313 | } | ||
| 314 | |||
| 315 | reply = xenbus_dev_request_and_reply(&u->u.msg); | ||
| 316 | if (IS_ERR(reply)) { | ||
| 317 | kfree(trans); | ||
| 318 | rc = PTR_ERR(reply); | ||
| 319 | goto out; | ||
| 320 | } | ||
| 321 | |||
| 322 | if (msg_type == XS_TRANSACTION_START) { | ||
| 323 | trans->handle.id = simple_strtoul(reply, NULL, 0); | ||
| 324 | |||
| 325 | list_add(&trans->list, &u->transactions); | ||
| 326 | } else if (msg_type == XS_TRANSACTION_END) { | ||
| 327 | list_for_each_entry(trans, &u->transactions, list) | ||
| 328 | if (trans->handle.id == u->u.msg.tx_id) | ||
| 329 | break; | ||
| 330 | BUG_ON(&trans->list == &u->transactions); | ||
| 331 | list_del(&trans->list); | ||
| 332 | |||
| 333 | kfree(trans); | ||
| 334 | } | ||
| 335 | |||
| 336 | mutex_lock(&u->reply_mutex); | ||
| 337 | rc = queue_reply(&staging_q, &u->u.msg, sizeof(u->u.msg)); | ||
| 338 | if (!rc) | ||
| 339 | rc = queue_reply(&staging_q, reply, u->u.msg.len); | ||
| 340 | if (!rc) { | ||
| 341 | list_splice_tail(&staging_q, &u->read_buffers); | ||
| 342 | wake_up(&u->read_waitq); | ||
| 343 | } else { | ||
| 344 | queue_cleanup(&staging_q); | ||
| 345 | } | ||
| 346 | mutex_unlock(&u->reply_mutex); | ||
| 347 | |||
| 348 | kfree(reply); | ||
| 349 | |||
| 350 | out: | ||
| 351 | return rc; | ||
| 352 | } | ||
| 353 | |||
| 354 | static int xenbus_write_watch(unsigned msg_type, struct xenbus_file_priv *u) | ||
| 355 | { | ||
| 356 | struct watch_adapter *watch, *tmp_watch; | ||
| 357 | char *path, *token; | ||
| 358 | int err, rc; | ||
| 359 | LIST_HEAD(staging_q); | ||
| 360 | |||
| 361 | path = u->u.buffer + sizeof(u->u.msg); | ||
| 362 | token = memchr(path, 0, u->u.msg.len); | ||
| 363 | if (token == NULL) { | ||
| 364 | rc = -EILSEQ; | ||
| 365 | goto out; | ||
| 366 | } | ||
| 367 | token++; | ||
| 368 | |||
| 369 | if (msg_type == XS_WATCH) { | ||
| 370 | watch = alloc_watch_adapter(path, token); | ||
| 371 | if (watch == NULL) { | ||
| 372 | rc = -ENOMEM; | ||
| 373 | goto out; | ||
| 374 | } | ||
| 375 | |||
| 376 | watch->watch.callback = watch_fired; | ||
| 377 | watch->dev_data = u; | ||
| 378 | |||
| 379 | err = register_xenbus_watch(&watch->watch); | ||
| 380 | if (err) { | ||
| 381 | free_watch_adapter(watch); | ||
| 382 | rc = err; | ||
| 383 | goto out; | ||
| 384 | } | ||
| 385 | list_add(&watch->list, &u->watches); | ||
| 386 | } else { | ||
| 387 | list_for_each_entry_safe(watch, tmp_watch, &u->watches, list) { | ||
| 388 | if (!strcmp(watch->token, token) && | ||
| 389 | !strcmp(watch->watch.node, path)) { | ||
| 390 | unregister_xenbus_watch(&watch->watch); | ||
| 391 | list_del(&watch->list); | ||
| 392 | free_watch_adapter(watch); | ||
| 393 | break; | ||
| 394 | } | ||
| 395 | } | ||
| 396 | } | ||
| 397 | |||
| 398 | /* Success. Synthesize a reply to say all is OK. */ | ||
| 399 | { | ||
| 400 | struct { | ||
| 401 | struct xsd_sockmsg hdr; | ||
| 402 | char body[3]; | ||
| 403 | } __packed reply = { | ||
| 404 | { | ||
| 405 | .type = msg_type, | ||
| 406 | .len = sizeof(reply.body) | ||
| 407 | }, | ||
| 408 | "OK" | ||
| 409 | }; | ||
| 410 | |||
| 411 | mutex_lock(&u->reply_mutex); | ||
| 412 | rc = queue_reply(&u->read_buffers, &reply, sizeof(reply)); | ||
| 413 | wake_up(&u->read_waitq); | ||
| 414 | mutex_unlock(&u->reply_mutex); | ||
| 415 | } | ||
| 416 | |||
| 417 | out: | ||
| 418 | return rc; | ||
| 419 | } | ||
| 420 | |||
| 421 | static ssize_t xenbus_file_write(struct file *filp, | ||
| 422 | const char __user *ubuf, | ||
| 423 | size_t len, loff_t *ppos) | ||
| 424 | { | ||
| 425 | struct xenbus_file_priv *u = filp->private_data; | ||
| 426 | uint32_t msg_type; | ||
| 427 | int rc = len; | ||
| 428 | int ret; | ||
| 429 | LIST_HEAD(staging_q); | ||
| 430 | |||
| 431 | /* | ||
| 432 | * We're expecting usermode to be writing properly formed | ||
| 433 | * xenbus messages. If they write an incomplete message we | ||
| 434 | * buffer it up. Once it is complete, we act on it. | ||
| 435 | */ | ||
| 436 | |||
| 437 | /* | ||
| 438 | * Make sure concurrent writers can't stomp all over each | ||
| 439 | * other's messages and make a mess of our partial message | ||
| 440 | * buffer. We don't make any attemppt to stop multiple | ||
| 441 | * writers from making a mess of each other's incomplete | ||
| 442 | * messages; we're just trying to guarantee our own internal | ||
| 443 | * consistency and make sure that single writes are handled | ||
| 444 | * atomically. | ||
| 445 | */ | ||
| 446 | mutex_lock(&u->msgbuffer_mutex); | ||
| 447 | |||
| 448 | /* Get this out of the way early to avoid confusion */ | ||
| 449 | if (len == 0) | ||
| 450 | goto out; | ||
| 451 | |||
| 452 | /* Can't write a xenbus message larger we can buffer */ | ||
| 453 | if ((len + u->len) > sizeof(u->u.buffer)) { | ||
| 454 | /* On error, dump existing buffer */ | ||
| 455 | u->len = 0; | ||
| 456 | rc = -EINVAL; | ||
| 457 | goto out; | ||
| 458 | } | ||
| 459 | |||
| 460 | ret = copy_from_user(u->u.buffer + u->len, ubuf, len); | ||
| 461 | |||
| 462 | if (ret != 0) { | ||
| 463 | rc = -EFAULT; | ||
| 464 | goto out; | ||
| 465 | } | ||
| 466 | |||
| 467 | /* Deal with a partial copy. */ | ||
| 468 | len -= ret; | ||
| 469 | rc = len; | ||
| 470 | |||
| 471 | u->len += len; | ||
| 472 | |||
| 473 | /* Return if we haven't got a full message yet */ | ||
| 474 | if (u->len < sizeof(u->u.msg)) | ||
| 475 | goto out; /* not even the header yet */ | ||
| 476 | |||
| 477 | /* If we're expecting a message that's larger than we can | ||
| 478 | possibly send, dump what we have and return an error. */ | ||
| 479 | if ((sizeof(u->u.msg) + u->u.msg.len) > sizeof(u->u.buffer)) { | ||
| 480 | rc = -E2BIG; | ||
| 481 | u->len = 0; | ||
| 482 | goto out; | ||
| 483 | } | ||
| 484 | |||
| 485 | if (u->len < (sizeof(u->u.msg) + u->u.msg.len)) | ||
| 486 | goto out; /* incomplete data portion */ | ||
| 487 | |||
| 488 | /* | ||
| 489 | * OK, now we have a complete message. Do something with it. | ||
| 490 | */ | ||
| 491 | |||
| 492 | msg_type = u->u.msg.type; | ||
| 493 | |||
| 494 | switch (msg_type) { | ||
| 495 | case XS_WATCH: | ||
| 496 | case XS_UNWATCH: | ||
| 497 | /* (Un)Ask for some path to be watched for changes */ | ||
| 498 | ret = xenbus_write_watch(msg_type, u); | ||
| 499 | break; | ||
| 500 | |||
| 501 | default: | ||
| 502 | /* Send out a transaction */ | ||
| 503 | ret = xenbus_write_transaction(msg_type, u); | ||
| 504 | break; | ||
| 505 | } | ||
| 506 | if (ret != 0) | ||
| 507 | rc = ret; | ||
| 508 | |||
| 509 | /* Buffered message consumed */ | ||
| 510 | u->len = 0; | ||
| 511 | |||
| 512 | out: | ||
| 513 | mutex_unlock(&u->msgbuffer_mutex); | ||
| 514 | return rc; | ||
| 515 | } | ||
| 516 | |||
| 517 | static int xenbus_file_open(struct inode *inode, struct file *filp) | ||
| 518 | { | ||
| 519 | struct xenbus_file_priv *u; | ||
| 520 | |||
| 521 | if (xen_store_evtchn == 0) | ||
| 522 | return -ENOENT; | ||
| 523 | |||
| 524 | nonseekable_open(inode, filp); | ||
| 525 | |||
| 526 | u = kzalloc(sizeof(*u), GFP_KERNEL); | ||
| 527 | if (u == NULL) | ||
| 528 | return -ENOMEM; | ||
| 529 | |||
| 530 | INIT_LIST_HEAD(&u->transactions); | ||
| 531 | INIT_LIST_HEAD(&u->watches); | ||
| 532 | INIT_LIST_HEAD(&u->read_buffers); | ||
| 533 | init_waitqueue_head(&u->read_waitq); | ||
| 534 | |||
| 535 | mutex_init(&u->reply_mutex); | ||
| 536 | mutex_init(&u->msgbuffer_mutex); | ||
| 537 | |||
| 538 | filp->private_data = u; | ||
| 539 | |||
| 540 | return 0; | ||
| 541 | } | ||
| 542 | |||
| 543 | static int xenbus_file_release(struct inode *inode, struct file *filp) | ||
| 544 | { | ||
| 545 | struct xenbus_file_priv *u = filp->private_data; | ||
| 546 | struct xenbus_transaction_holder *trans, *tmp; | ||
| 547 | struct watch_adapter *watch, *tmp_watch; | ||
| 548 | struct read_buffer *rb, *tmp_rb; | ||
| 549 | |||
| 550 | /* | ||
| 551 | * No need for locking here because there are no other users, | ||
| 552 | * by definition. | ||
| 553 | */ | ||
| 554 | |||
| 555 | list_for_each_entry_safe(trans, tmp, &u->transactions, list) { | ||
| 556 | xenbus_transaction_end(trans->handle, 1); | ||
| 557 | list_del(&trans->list); | ||
| 558 | kfree(trans); | ||
| 559 | } | ||
| 560 | |||
| 561 | list_for_each_entry_safe(watch, tmp_watch, &u->watches, list) { | ||
| 562 | unregister_xenbus_watch(&watch->watch); | ||
| 563 | list_del(&watch->list); | ||
| 564 | free_watch_adapter(watch); | ||
| 565 | } | ||
| 566 | |||
| 567 | list_for_each_entry_safe(rb, tmp_rb, &u->read_buffers, list) { | ||
| 568 | list_del(&rb->list); | ||
| 569 | kfree(rb); | ||
| 570 | } | ||
| 571 | kfree(u); | ||
| 572 | |||
| 573 | return 0; | ||
| 574 | } | ||
| 575 | |||
| 576 | static unsigned int xenbus_file_poll(struct file *file, poll_table *wait) | ||
| 577 | { | ||
| 578 | struct xenbus_file_priv *u = file->private_data; | ||
| 579 | |||
| 580 | poll_wait(file, &u->read_waitq, wait); | ||
| 581 | if (!list_empty(&u->read_buffers)) | ||
| 582 | return POLLIN | POLLRDNORM; | ||
| 583 | return 0; | ||
| 584 | } | ||
| 585 | |||
| 586 | const struct file_operations xenbus_file_ops = { | ||
| 587 | .read = xenbus_file_read, | ||
| 588 | .write = xenbus_file_write, | ||
| 589 | .open = xenbus_file_open, | ||
| 590 | .release = xenbus_file_release, | ||
| 591 | .poll = xenbus_file_poll, | ||
| 592 | .llseek = no_llseek, | ||
| 593 | }; | ||
