diff options
| -rw-r--r-- | fs/nfs/direct.c | 435 | ||||
| -rw-r--r-- | include/linux/nfs_xdr.h | 2 |
2 files changed, 203 insertions, 234 deletions
diff --git a/fs/nfs/direct.c b/fs/nfs/direct.c index 4cdd1b499e35..fecd3b095deb 100644 --- a/fs/nfs/direct.c +++ b/fs/nfs/direct.c | |||
| @@ -67,25 +67,19 @@ struct nfs_direct_req { | |||
| 67 | struct kref kref; /* release manager */ | 67 | struct kref kref; /* release manager */ |
| 68 | 68 | ||
| 69 | /* I/O parameters */ | 69 | /* I/O parameters */ |
| 70 | struct list_head list, /* nfs_read/write_data structs */ | ||
| 71 | rewrite_list; /* saved nfs_write_data structs */ | ||
| 72 | struct nfs_open_context *ctx; /* file open context info */ | 70 | struct nfs_open_context *ctx; /* file open context info */ |
| 73 | struct kiocb * iocb; /* controlling i/o request */ | 71 | struct kiocb * iocb; /* controlling i/o request */ |
| 74 | struct inode * inode; /* target file of i/o */ | 72 | struct inode * inode; /* target file of i/o */ |
| 75 | unsigned long user_addr; /* location of user's buffer */ | ||
| 76 | size_t user_count; /* total bytes to move */ | ||
| 77 | loff_t pos; /* starting offset in file */ | ||
| 78 | struct page ** pages; /* pages in our buffer */ | ||
| 79 | unsigned int npages; /* count of pages */ | ||
| 80 | 73 | ||
| 81 | /* completion state */ | 74 | /* completion state */ |
| 75 | atomic_t io_count; /* i/os we're waiting for */ | ||
| 82 | spinlock_t lock; /* protect completion state */ | 76 | spinlock_t lock; /* protect completion state */ |
| 83 | int outstanding; /* i/os we're waiting for */ | ||
| 84 | ssize_t count, /* bytes actually processed */ | 77 | ssize_t count, /* bytes actually processed */ |
| 85 | error; /* any reported error */ | 78 | error; /* any reported error */ |
| 86 | struct completion completion; /* wait for i/o completion */ | 79 | struct completion completion; /* wait for i/o completion */ |
| 87 | 80 | ||
| 88 | /* commit state */ | 81 | /* commit state */ |
| 82 | struct list_head rewrite_list; /* saved nfs_write_data structs */ | ||
| 89 | struct nfs_write_data * commit_data; /* special write_data for commits */ | 83 | struct nfs_write_data * commit_data; /* special write_data for commits */ |
| 90 | int flags; | 84 | int flags; |
| 91 | #define NFS_ODIRECT_DO_COMMIT (1) /* an unstable reply was received */ | 85 | #define NFS_ODIRECT_DO_COMMIT (1) /* an unstable reply was received */ |
| @@ -93,8 +87,37 @@ struct nfs_direct_req { | |||
| 93 | struct nfs_writeverf verf; /* unstable write verifier */ | 87 | struct nfs_writeverf verf; /* unstable write verifier */ |
| 94 | }; | 88 | }; |
| 95 | 89 | ||
| 96 | static void nfs_direct_write_schedule(struct nfs_direct_req *dreq, int sync); | ||
| 97 | static void nfs_direct_write_complete(struct nfs_direct_req *dreq, struct inode *inode); | 90 | static void nfs_direct_write_complete(struct nfs_direct_req *dreq, struct inode *inode); |
| 91 | static const struct rpc_call_ops nfs_write_direct_ops; | ||
| 92 | |||
| 93 | static inline void get_dreq(struct nfs_direct_req *dreq) | ||
| 94 | { | ||
| 95 | atomic_inc(&dreq->io_count); | ||
| 96 | } | ||
| 97 | |||
| 98 | static inline int put_dreq(struct nfs_direct_req *dreq) | ||
| 99 | { | ||
| 100 | return atomic_dec_and_test(&dreq->io_count); | ||
| 101 | } | ||
| 102 | |||
| 103 | /* | ||
| 104 | * "size" is never larger than rsize or wsize. | ||
| 105 | */ | ||
| 106 | static inline int nfs_direct_count_pages(unsigned long user_addr, size_t size) | ||
| 107 | { | ||
| 108 | int page_count; | ||
| 109 | |||
| 110 | page_count = (user_addr + size + PAGE_SIZE - 1) >> PAGE_SHIFT; | ||
| 111 | page_count -= user_addr >> PAGE_SHIFT; | ||
| 112 | BUG_ON(page_count < 0); | ||
| 113 | |||
| 114 | return page_count; | ||
| 115 | } | ||
| 116 | |||
| 117 | static inline unsigned int nfs_max_pages(unsigned int size) | ||
| 118 | { | ||
| 119 | return (size + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; | ||
| 120 | } | ||
| 98 | 121 | ||
| 99 | /** | 122 | /** |
| 100 | * nfs_direct_IO - NFS address space operation for direct I/O | 123 | * nfs_direct_IO - NFS address space operation for direct I/O |
| @@ -118,50 +141,21 @@ ssize_t nfs_direct_IO(int rw, struct kiocb *iocb, const struct iovec *iov, loff_ | |||
| 118 | return -EINVAL; | 141 | return -EINVAL; |
| 119 | } | 142 | } |
| 120 | 143 | ||
| 121 | static void nfs_free_user_pages(struct page **pages, int npages, int do_dirty) | 144 | static void nfs_direct_dirty_pages(struct page **pages, int npages) |
| 122 | { | 145 | { |
| 123 | int i; | 146 | int i; |
| 124 | for (i = 0; i < npages; i++) { | 147 | for (i = 0; i < npages; i++) { |
| 125 | struct page *page = pages[i]; | 148 | struct page *page = pages[i]; |
| 126 | if (do_dirty && !PageCompound(page)) | 149 | if (!PageCompound(page)) |
| 127 | set_page_dirty_lock(page); | 150 | set_page_dirty_lock(page); |
| 128 | page_cache_release(page); | ||
| 129 | } | 151 | } |
| 130 | kfree(pages); | ||
| 131 | } | 152 | } |
| 132 | 153 | ||
| 133 | static inline int nfs_get_user_pages(int rw, unsigned long user_addr, size_t size, struct page ***pages) | 154 | static void nfs_direct_release_pages(struct page **pages, int npages) |
| 134 | { | 155 | { |
| 135 | int result = -ENOMEM; | 156 | int i; |
| 136 | unsigned long page_count; | 157 | for (i = 0; i < npages; i++) |
| 137 | size_t array_size; | 158 | page_cache_release(pages[i]); |
| 138 | |||
| 139 | page_count = (user_addr + size + PAGE_SIZE - 1) >> PAGE_SHIFT; | ||
| 140 | page_count -= user_addr >> PAGE_SHIFT; | ||
| 141 | |||
| 142 | array_size = (page_count * sizeof(struct page *)); | ||
| 143 | *pages = kmalloc(array_size, GFP_KERNEL); | ||
| 144 | if (*pages) { | ||
| 145 | down_read(¤t->mm->mmap_sem); | ||
| 146 | result = get_user_pages(current, current->mm, user_addr, | ||
| 147 | page_count, (rw == READ), 0, | ||
| 148 | *pages, NULL); | ||
| 149 | up_read(¤t->mm->mmap_sem); | ||
| 150 | if (result != page_count) { | ||
| 151 | /* | ||
| 152 | * If we got fewer pages than expected from | ||
| 153 | * get_user_pages(), the user buffer runs off the | ||
| 154 | * end of a mapping; return EFAULT. | ||
| 155 | */ | ||
| 156 | if (result >= 0) { | ||
| 157 | nfs_free_user_pages(*pages, result, 0); | ||
| 158 | result = -EFAULT; | ||
| 159 | } else | ||
| 160 | kfree(*pages); | ||
| 161 | *pages = NULL; | ||
| 162 | } | ||
| 163 | } | ||
| 164 | return result; | ||
| 165 | } | 159 | } |
| 166 | 160 | ||
| 167 | static inline struct nfs_direct_req *nfs_direct_req_alloc(void) | 161 | static inline struct nfs_direct_req *nfs_direct_req_alloc(void) |
| @@ -173,13 +167,13 @@ static inline struct nfs_direct_req *nfs_direct_req_alloc(void) | |||
| 173 | return NULL; | 167 | return NULL; |
| 174 | 168 | ||
| 175 | kref_init(&dreq->kref); | 169 | kref_init(&dreq->kref); |
| 170 | kref_get(&dreq->kref); | ||
| 176 | init_completion(&dreq->completion); | 171 | init_completion(&dreq->completion); |
| 177 | INIT_LIST_HEAD(&dreq->list); | ||
| 178 | INIT_LIST_HEAD(&dreq->rewrite_list); | 172 | INIT_LIST_HEAD(&dreq->rewrite_list); |
| 179 | dreq->iocb = NULL; | 173 | dreq->iocb = NULL; |
| 180 | dreq->ctx = NULL; | 174 | dreq->ctx = NULL; |
| 181 | spin_lock_init(&dreq->lock); | 175 | spin_lock_init(&dreq->lock); |
| 182 | dreq->outstanding = 0; | 176 | atomic_set(&dreq->io_count, 0); |
| 183 | dreq->count = 0; | 177 | dreq->count = 0; |
| 184 | dreq->error = 0; | 178 | dreq->error = 0; |
| 185 | dreq->flags = 0; | 179 | dreq->flags = 0; |
| @@ -220,18 +214,11 @@ out: | |||
| 220 | } | 214 | } |
| 221 | 215 | ||
| 222 | /* | 216 | /* |
| 223 | * We must hold a reference to all the pages in this direct read request | 217 | * Synchronous I/O uses a stack-allocated iocb. Thus we can't trust |
| 224 | * until the RPCs complete. This could be long *after* we are woken up in | 218 | * the iocb is still valid here if this is a synchronous request. |
| 225 | * nfs_direct_wait (for instance, if someone hits ^C on a slow server). | ||
| 226 | * | ||
| 227 | * In addition, synchronous I/O uses a stack-allocated iocb. Thus we | ||
| 228 | * can't trust the iocb is still valid here if this is a synchronous | ||
| 229 | * request. If the waiter is woken prematurely, the iocb is long gone. | ||
| 230 | */ | 219 | */ |
| 231 | static void nfs_direct_complete(struct nfs_direct_req *dreq) | 220 | static void nfs_direct_complete(struct nfs_direct_req *dreq) |
| 232 | { | 221 | { |
| 233 | nfs_free_user_pages(dreq->pages, dreq->npages, 1); | ||
| 234 | |||
| 235 | if (dreq->iocb) { | 222 | if (dreq->iocb) { |
| 236 | long res = (long) dreq->error; | 223 | long res = (long) dreq->error; |
| 237 | if (!res) | 224 | if (!res) |
| @@ -244,48 +231,10 @@ static void nfs_direct_complete(struct nfs_direct_req *dreq) | |||
| 244 | } | 231 | } |
| 245 | 232 | ||
| 246 | /* | 233 | /* |
| 247 | * Note we also set the number of requests we have in the dreq when we are | 234 | * We must hold a reference to all the pages in this direct read request |
| 248 | * done. This prevents races with I/O completion so we will always wait | 235 | * until the RPCs complete. This could be long *after* we are woken up in |
| 249 | * until all requests have been dispatched and completed. | 236 | * nfs_direct_wait (for instance, if someone hits ^C on a slow server). |
| 250 | */ | 237 | */ |
| 251 | static struct nfs_direct_req *nfs_direct_read_alloc(size_t nbytes, size_t rsize) | ||
| 252 | { | ||
| 253 | struct list_head *list; | ||
| 254 | struct nfs_direct_req *dreq; | ||
| 255 | unsigned int rpages = (rsize + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; | ||
| 256 | |||
| 257 | dreq = nfs_direct_req_alloc(); | ||
| 258 | if (!dreq) | ||
| 259 | return NULL; | ||
| 260 | |||
| 261 | list = &dreq->list; | ||
| 262 | for(;;) { | ||
| 263 | struct nfs_read_data *data = nfs_readdata_alloc(rpages); | ||
| 264 | |||
| 265 | if (unlikely(!data)) { | ||
| 266 | while (!list_empty(list)) { | ||
| 267 | data = list_entry(list->next, | ||
| 268 | struct nfs_read_data, pages); | ||
| 269 | list_del(&data->pages); | ||
| 270 | nfs_readdata_free(data); | ||
| 271 | } | ||
| 272 | kref_put(&dreq->kref, nfs_direct_req_release); | ||
| 273 | return NULL; | ||
| 274 | } | ||
| 275 | |||
| 276 | INIT_LIST_HEAD(&data->pages); | ||
| 277 | list_add(&data->pages, list); | ||
| 278 | |||
| 279 | data->req = (struct nfs_page *) dreq; | ||
| 280 | dreq->outstanding++; | ||
| 281 | if (nbytes <= rsize) | ||
| 282 | break; | ||
| 283 | nbytes -= rsize; | ||
| 284 | } | ||
| 285 | kref_get(&dreq->kref); | ||
| 286 | return dreq; | ||
| 287 | } | ||
| 288 | |||
| 289 | static void nfs_direct_read_result(struct rpc_task *task, void *calldata) | 238 | static void nfs_direct_read_result(struct rpc_task *task, void *calldata) |
| 290 | { | 239 | { |
| 291 | struct nfs_read_data *data = calldata; | 240 | struct nfs_read_data *data = calldata; |
| @@ -294,6 +243,9 @@ static void nfs_direct_read_result(struct rpc_task *task, void *calldata) | |||
| 294 | if (nfs_readpage_result(task, data) != 0) | 243 | if (nfs_readpage_result(task, data) != 0) |
| 295 | return; | 244 | return; |
| 296 | 245 | ||
| 246 | nfs_direct_dirty_pages(data->pagevec, data->npages); | ||
| 247 | nfs_direct_release_pages(data->pagevec, data->npages); | ||
| 248 | |||
| 297 | spin_lock(&dreq->lock); | 249 | spin_lock(&dreq->lock); |
| 298 | 250 | ||
| 299 | if (likely(task->tk_status >= 0)) | 251 | if (likely(task->tk_status >= 0)) |
| @@ -301,13 +253,10 @@ static void nfs_direct_read_result(struct rpc_task *task, void *calldata) | |||
| 301 | else | 253 | else |
| 302 | dreq->error = task->tk_status; | 254 | dreq->error = task->tk_status; |
| 303 | 255 | ||
| 304 | if (--dreq->outstanding) { | ||
| 305 | spin_unlock(&dreq->lock); | ||
| 306 | return; | ||
| 307 | } | ||
| 308 | |||
| 309 | spin_unlock(&dreq->lock); | 256 | spin_unlock(&dreq->lock); |
| 310 | nfs_direct_complete(dreq); | 257 | |
| 258 | if (put_dreq(dreq)) | ||
| 259 | nfs_direct_complete(dreq); | ||
| 311 | } | 260 | } |
| 312 | 261 | ||
| 313 | static const struct rpc_call_ops nfs_read_direct_ops = { | 262 | static const struct rpc_call_ops nfs_read_direct_ops = { |
| @@ -316,41 +265,60 @@ static const struct rpc_call_ops nfs_read_direct_ops = { | |||
| 316 | }; | 265 | }; |
| 317 | 266 | ||
| 318 | /* | 267 | /* |
| 319 | * For each nfs_read_data struct that was allocated on the list, dispatch | 268 | * For each rsize'd chunk of the user's buffer, dispatch an NFS READ |
| 320 | * an NFS READ operation | 269 | * operation. If nfs_readdata_alloc() or get_user_pages() fails, |
| 270 | * bail and stop sending more reads. Read length accounting is | ||
| 271 | * handled automatically by nfs_direct_read_result(). Otherwise, if | ||
| 272 | * no requests have been sent, just return an error. | ||
| 321 | */ | 273 | */ |
| 322 | static void nfs_direct_read_schedule(struct nfs_direct_req *dreq) | 274 | static ssize_t nfs_direct_read_schedule(struct nfs_direct_req *dreq, unsigned long user_addr, size_t count, loff_t pos) |
| 323 | { | 275 | { |
| 324 | struct nfs_open_context *ctx = dreq->ctx; | 276 | struct nfs_open_context *ctx = dreq->ctx; |
| 325 | struct inode *inode = ctx->dentry->d_inode; | 277 | struct inode *inode = ctx->dentry->d_inode; |
| 326 | struct list_head *list = &dreq->list; | ||
| 327 | struct page **pages = dreq->pages; | ||
| 328 | size_t count = dreq->user_count; | ||
| 329 | loff_t pos = dreq->pos; | ||
| 330 | size_t rsize = NFS_SERVER(inode)->rsize; | 278 | size_t rsize = NFS_SERVER(inode)->rsize; |
| 331 | unsigned int curpage, pgbase; | 279 | unsigned int rpages = nfs_max_pages(rsize); |
| 280 | unsigned int pgbase; | ||
| 281 | int result; | ||
| 282 | ssize_t started = 0; | ||
| 283 | |||
| 284 | get_dreq(dreq); | ||
| 332 | 285 | ||
| 333 | curpage = 0; | 286 | pgbase = user_addr & ~PAGE_MASK; |
| 334 | pgbase = dreq->user_addr & ~PAGE_MASK; | ||
| 335 | do { | 287 | do { |
| 336 | struct nfs_read_data *data; | 288 | struct nfs_read_data *data; |
| 337 | size_t bytes; | 289 | size_t bytes; |
| 338 | 290 | ||
| 291 | result = -ENOMEM; | ||
| 292 | data = nfs_readdata_alloc(rpages); | ||
| 293 | if (unlikely(!data)) | ||
| 294 | break; | ||
| 295 | |||
| 339 | bytes = rsize; | 296 | bytes = rsize; |
| 340 | if (count < rsize) | 297 | if (count < rsize) |
| 341 | bytes = count; | 298 | bytes = count; |
| 342 | 299 | ||
| 343 | BUG_ON(list_empty(list)); | 300 | data->npages = nfs_direct_count_pages(user_addr, bytes); |
| 344 | data = list_entry(list->next, struct nfs_read_data, pages); | 301 | down_read(¤t->mm->mmap_sem); |
| 345 | list_del_init(&data->pages); | 302 | result = get_user_pages(current, current->mm, user_addr, |
| 303 | data->npages, 1, 0, data->pagevec, NULL); | ||
| 304 | up_read(¤t->mm->mmap_sem); | ||
| 305 | if (unlikely(result < data->npages)) { | ||
| 306 | if (result > 0) | ||
| 307 | nfs_direct_release_pages(data->pagevec, result); | ||
| 308 | nfs_readdata_release(data); | ||
| 309 | break; | ||
| 310 | } | ||
| 311 | |||
| 312 | get_dreq(dreq); | ||
| 346 | 313 | ||
| 314 | data->req = (struct nfs_page *) dreq; | ||
| 347 | data->inode = inode; | 315 | data->inode = inode; |
| 348 | data->cred = ctx->cred; | 316 | data->cred = ctx->cred; |
| 349 | data->args.fh = NFS_FH(inode); | 317 | data->args.fh = NFS_FH(inode); |
| 350 | data->args.context = ctx; | 318 | data->args.context = ctx; |
| 351 | data->args.offset = pos; | 319 | data->args.offset = pos; |
| 352 | data->args.pgbase = pgbase; | 320 | data->args.pgbase = pgbase; |
| 353 | data->args.pages = &pages[curpage]; | 321 | data->args.pages = data->pagevec; |
| 354 | data->args.count = bytes; | 322 | data->args.count = bytes; |
| 355 | data->res.fattr = &data->fattr; | 323 | data->res.fattr = &data->fattr; |
| 356 | data->res.eof = 0; | 324 | data->res.eof = 0; |
| @@ -373,33 +341,35 @@ static void nfs_direct_read_schedule(struct nfs_direct_req *dreq) | |||
| 373 | bytes, | 341 | bytes, |
| 374 | (unsigned long long)data->args.offset); | 342 | (unsigned long long)data->args.offset); |
| 375 | 343 | ||
| 344 | started += bytes; | ||
| 345 | user_addr += bytes; | ||
| 376 | pos += bytes; | 346 | pos += bytes; |
| 377 | pgbase += bytes; | 347 | pgbase += bytes; |
| 378 | curpage += pgbase >> PAGE_SHIFT; | ||
| 379 | pgbase &= ~PAGE_MASK; | 348 | pgbase &= ~PAGE_MASK; |
| 380 | 349 | ||
| 381 | count -= bytes; | 350 | count -= bytes; |
| 382 | } while (count != 0); | 351 | } while (count != 0); |
| 383 | BUG_ON(!list_empty(list)); | 352 | |
| 353 | if (put_dreq(dreq)) | ||
| 354 | nfs_direct_complete(dreq); | ||
| 355 | |||
| 356 | if (started) | ||
| 357 | return 0; | ||
| 358 | return result < 0 ? (ssize_t) result : -EFAULT; | ||
| 384 | } | 359 | } |
| 385 | 360 | ||
| 386 | static ssize_t nfs_direct_read(struct kiocb *iocb, unsigned long user_addr, size_t count, loff_t pos, struct page **pages, unsigned int nr_pages) | 361 | static ssize_t nfs_direct_read(struct kiocb *iocb, unsigned long user_addr, size_t count, loff_t pos) |
| 387 | { | 362 | { |
| 388 | ssize_t result; | 363 | ssize_t result = 0; |
| 389 | sigset_t oldset; | 364 | sigset_t oldset; |
| 390 | struct inode *inode = iocb->ki_filp->f_mapping->host; | 365 | struct inode *inode = iocb->ki_filp->f_mapping->host; |
| 391 | struct rpc_clnt *clnt = NFS_CLIENT(inode); | 366 | struct rpc_clnt *clnt = NFS_CLIENT(inode); |
| 392 | struct nfs_direct_req *dreq; | 367 | struct nfs_direct_req *dreq; |
| 393 | 368 | ||
| 394 | dreq = nfs_direct_read_alloc(count, NFS_SERVER(inode)->rsize); | 369 | dreq = nfs_direct_req_alloc(); |
| 395 | if (!dreq) | 370 | if (!dreq) |
| 396 | return -ENOMEM; | 371 | return -ENOMEM; |
| 397 | 372 | ||
| 398 | dreq->user_addr = user_addr; | ||
| 399 | dreq->user_count = count; | ||
| 400 | dreq->pos = pos; | ||
| 401 | dreq->pages = pages; | ||
| 402 | dreq->npages = nr_pages; | ||
| 403 | dreq->inode = inode; | 373 | dreq->inode = inode; |
| 404 | dreq->ctx = get_nfs_open_context((struct nfs_open_context *)iocb->ki_filp->private_data); | 374 | dreq->ctx = get_nfs_open_context((struct nfs_open_context *)iocb->ki_filp->private_data); |
| 405 | if (!is_sync_kiocb(iocb)) | 375 | if (!is_sync_kiocb(iocb)) |
| @@ -407,8 +377,9 @@ static ssize_t nfs_direct_read(struct kiocb *iocb, unsigned long user_addr, size | |||
| 407 | 377 | ||
| 408 | nfs_add_stats(inode, NFSIOS_DIRECTREADBYTES, count); | 378 | nfs_add_stats(inode, NFSIOS_DIRECTREADBYTES, count); |
| 409 | rpc_clnt_sigmask(clnt, &oldset); | 379 | rpc_clnt_sigmask(clnt, &oldset); |
| 410 | nfs_direct_read_schedule(dreq); | 380 | result = nfs_direct_read_schedule(dreq, user_addr, count, pos); |
| 411 | result = nfs_direct_wait(dreq); | 381 | if (!result) |
| 382 | result = nfs_direct_wait(dreq); | ||
| 412 | rpc_clnt_sigunmask(clnt, &oldset); | 383 | rpc_clnt_sigunmask(clnt, &oldset); |
| 413 | 384 | ||
| 414 | return result; | 385 | return result; |
| @@ -416,10 +387,10 @@ static ssize_t nfs_direct_read(struct kiocb *iocb, unsigned long user_addr, size | |||
| 416 | 387 | ||
| 417 | static void nfs_direct_free_writedata(struct nfs_direct_req *dreq) | 388 | static void nfs_direct_free_writedata(struct nfs_direct_req *dreq) |
| 418 | { | 389 | { |
| 419 | list_splice_init(&dreq->rewrite_list, &dreq->list); | 390 | while (!list_empty(&dreq->rewrite_list)) { |
| 420 | while (!list_empty(&dreq->list)) { | 391 | struct nfs_write_data *data = list_entry(dreq->rewrite_list.next, struct nfs_write_data, pages); |
| 421 | struct nfs_write_data *data = list_entry(dreq->list.next, struct nfs_write_data, pages); | ||
| 422 | list_del(&data->pages); | 392 | list_del(&data->pages); |
| 393 | nfs_direct_release_pages(data->pagevec, data->npages); | ||
| 423 | nfs_writedata_release(data); | 394 | nfs_writedata_release(data); |
| 424 | } | 395 | } |
| 425 | } | 396 | } |
| @@ -427,14 +398,51 @@ static void nfs_direct_free_writedata(struct nfs_direct_req *dreq) | |||
| 427 | #if defined(CONFIG_NFS_V3) || defined(CONFIG_NFS_V4) | 398 | #if defined(CONFIG_NFS_V3) || defined(CONFIG_NFS_V4) |
| 428 | static void nfs_direct_write_reschedule(struct nfs_direct_req *dreq) | 399 | static void nfs_direct_write_reschedule(struct nfs_direct_req *dreq) |
| 429 | { | 400 | { |
| 430 | struct list_head *pos; | 401 | struct inode *inode = dreq->inode; |
| 402 | struct list_head *p; | ||
| 403 | struct nfs_write_data *data; | ||
| 431 | 404 | ||
| 432 | list_splice_init(&dreq->rewrite_list, &dreq->list); | ||
| 433 | list_for_each(pos, &dreq->list) | ||
| 434 | dreq->outstanding++; | ||
| 435 | dreq->count = 0; | 405 | dreq->count = 0; |
| 406 | get_dreq(dreq); | ||
| 407 | |||
| 408 | list_for_each(p, &dreq->rewrite_list) { | ||
| 409 | data = list_entry(p, struct nfs_write_data, pages); | ||
| 410 | |||
| 411 | get_dreq(dreq); | ||
| 412 | |||
| 413 | /* | ||
| 414 | * Reset data->res. | ||
| 415 | */ | ||
| 416 | nfs_fattr_init(&data->fattr); | ||
| 417 | data->res.count = data->args.count; | ||
| 418 | memset(&data->verf, 0, sizeof(data->verf)); | ||
| 419 | |||
| 420 | /* | ||
| 421 | * Reuse data->task; data->args should not have changed | ||
| 422 | * since the original request was sent. | ||
| 423 | */ | ||
| 424 | rpc_init_task(&data->task, NFS_CLIENT(inode), RPC_TASK_ASYNC, | ||
| 425 | &nfs_write_direct_ops, data); | ||
| 426 | NFS_PROTO(inode)->write_setup(data, FLUSH_STABLE); | ||
| 427 | |||
| 428 | data->task.tk_priority = RPC_PRIORITY_NORMAL; | ||
| 429 | data->task.tk_cookie = (unsigned long) inode; | ||
| 430 | |||
| 431 | /* | ||
| 432 | * We're called via an RPC callback, so BKL is already held. | ||
| 433 | */ | ||
| 434 | rpc_execute(&data->task); | ||
| 435 | |||
| 436 | dprintk("NFS: %5u rescheduled direct write call (req %s/%Ld, %u bytes @ offset %Lu)\n", | ||
| 437 | data->task.tk_pid, | ||
| 438 | inode->i_sb->s_id, | ||
| 439 | (long long)NFS_FILEID(inode), | ||
| 440 | data->args.count, | ||
| 441 | (unsigned long long)data->args.offset); | ||
| 442 | } | ||
| 436 | 443 | ||
| 437 | nfs_direct_write_schedule(dreq, FLUSH_STABLE); | 444 | if (put_dreq(dreq)) |
| 445 | nfs_direct_write_complete(dreq, inode); | ||
| 438 | } | 446 | } |
| 439 | 447 | ||
| 440 | static void nfs_direct_commit_result(struct rpc_task *task, void *calldata) | 448 | static void nfs_direct_commit_result(struct rpc_task *task, void *calldata) |
| @@ -471,8 +479,8 @@ static void nfs_direct_commit_schedule(struct nfs_direct_req *dreq) | |||
| 471 | data->cred = dreq->ctx->cred; | 479 | data->cred = dreq->ctx->cred; |
| 472 | 480 | ||
| 473 | data->args.fh = NFS_FH(data->inode); | 481 | data->args.fh = NFS_FH(data->inode); |
| 474 | data->args.offset = dreq->pos; | 482 | data->args.offset = 0; |
| 475 | data->args.count = dreq->user_count; | 483 | data->args.count = 0; |
| 476 | data->res.count = 0; | 484 | data->res.count = 0; |
| 477 | data->res.fattr = &data->fattr; | 485 | data->res.fattr = &data->fattr; |
| 478 | data->res.verf = &data->verf; | 486 | data->res.verf = &data->verf; |
| @@ -534,47 +542,6 @@ static void nfs_direct_write_complete(struct nfs_direct_req *dreq, struct inode | |||
| 534 | } | 542 | } |
| 535 | #endif | 543 | #endif |
| 536 | 544 | ||
| 537 | static struct nfs_direct_req *nfs_direct_write_alloc(size_t nbytes, size_t wsize) | ||
| 538 | { | ||
| 539 | struct list_head *list; | ||
| 540 | struct nfs_direct_req *dreq; | ||
| 541 | unsigned int wpages = (wsize + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; | ||
| 542 | |||
| 543 | dreq = nfs_direct_req_alloc(); | ||
| 544 | if (!dreq) | ||
| 545 | return NULL; | ||
| 546 | |||
| 547 | list = &dreq->list; | ||
| 548 | for(;;) { | ||
| 549 | struct nfs_write_data *data = nfs_writedata_alloc(wpages); | ||
| 550 | |||
| 551 | if (unlikely(!data)) { | ||
| 552 | while (!list_empty(list)) { | ||
| 553 | data = list_entry(list->next, | ||
| 554 | struct nfs_write_data, pages); | ||
| 555 | list_del(&data->pages); | ||
| 556 | nfs_writedata_free(data); | ||
| 557 | } | ||
| 558 | kref_put(&dreq->kref, nfs_direct_req_release); | ||
| 559 | return NULL; | ||
| 560 | } | ||
| 561 | |||
| 562 | INIT_LIST_HEAD(&data->pages); | ||
| 563 | list_add(&data->pages, list); | ||
| 564 | |||
| 565 | data->req = (struct nfs_page *) dreq; | ||
| 566 | dreq->outstanding++; | ||
| 567 | if (nbytes <= wsize) | ||
| 568 | break; | ||
| 569 | nbytes -= wsize; | ||
| 570 | } | ||
| 571 | |||
| 572 | nfs_alloc_commit_data(dreq); | ||
| 573 | |||
| 574 | kref_get(&dreq->kref); | ||
| 575 | return dreq; | ||
| 576 | } | ||
| 577 | |||
| 578 | static void nfs_direct_write_result(struct rpc_task *task, void *calldata) | 545 | static void nfs_direct_write_result(struct rpc_task *task, void *calldata) |
| 579 | { | 546 | { |
| 580 | struct nfs_write_data *data = calldata; | 547 | struct nfs_write_data *data = calldata; |
| @@ -604,8 +571,6 @@ static void nfs_direct_write_result(struct rpc_task *task, void *calldata) | |||
| 604 | } | 571 | } |
| 605 | } | 572 | } |
| 606 | } | 573 | } |
| 607 | /* In case we have to resend */ | ||
| 608 | data->args.stable = NFS_FILE_SYNC; | ||
| 609 | 574 | ||
| 610 | spin_unlock(&dreq->lock); | 575 | spin_unlock(&dreq->lock); |
| 611 | } | 576 | } |
| @@ -619,14 +584,8 @@ static void nfs_direct_write_release(void *calldata) | |||
| 619 | struct nfs_write_data *data = calldata; | 584 | struct nfs_write_data *data = calldata; |
| 620 | struct nfs_direct_req *dreq = (struct nfs_direct_req *) data->req; | 585 | struct nfs_direct_req *dreq = (struct nfs_direct_req *) data->req; |
| 621 | 586 | ||
| 622 | spin_lock(&dreq->lock); | 587 | if (put_dreq(dreq)) |
| 623 | if (--dreq->outstanding) { | 588 | nfs_direct_write_complete(dreq, data->inode); |
| 624 | spin_unlock(&dreq->lock); | ||
| 625 | return; | ||
| 626 | } | ||
| 627 | spin_unlock(&dreq->lock); | ||
| 628 | |||
| 629 | nfs_direct_write_complete(dreq, data->inode); | ||
| 630 | } | 589 | } |
| 631 | 590 | ||
| 632 | static const struct rpc_call_ops nfs_write_direct_ops = { | 591 | static const struct rpc_call_ops nfs_write_direct_ops = { |
| @@ -635,41 +594,62 @@ static const struct rpc_call_ops nfs_write_direct_ops = { | |||
| 635 | }; | 594 | }; |
| 636 | 595 | ||
| 637 | /* | 596 | /* |
| 638 | * For each nfs_write_data struct that was allocated on the list, dispatch | 597 | * For each wsize'd chunk of the user's buffer, dispatch an NFS WRITE |
| 639 | * an NFS WRITE operation | 598 | * operation. If nfs_writedata_alloc() or get_user_pages() fails, |
| 599 | * bail and stop sending more writes. Write length accounting is | ||
| 600 | * handled automatically by nfs_direct_write_result(). Otherwise, if | ||
| 601 | * no requests have been sent, just return an error. | ||
| 640 | */ | 602 | */ |
| 641 | static void nfs_direct_write_schedule(struct nfs_direct_req *dreq, int sync) | 603 | static ssize_t nfs_direct_write_schedule(struct nfs_direct_req *dreq, unsigned long user_addr, size_t count, loff_t pos, int sync) |
| 642 | { | 604 | { |
| 643 | struct nfs_open_context *ctx = dreq->ctx; | 605 | struct nfs_open_context *ctx = dreq->ctx; |
| 644 | struct inode *inode = ctx->dentry->d_inode; | 606 | struct inode *inode = ctx->dentry->d_inode; |
| 645 | struct list_head *list = &dreq->list; | ||
| 646 | struct page **pages = dreq->pages; | ||
| 647 | size_t count = dreq->user_count; | ||
| 648 | loff_t pos = dreq->pos; | ||
| 649 | size_t wsize = NFS_SERVER(inode)->wsize; | 607 | size_t wsize = NFS_SERVER(inode)->wsize; |
| 650 | unsigned int curpage, pgbase; | 608 | unsigned int wpages = nfs_max_pages(wsize); |
| 609 | unsigned int pgbase; | ||
| 610 | int result; | ||
| 611 | ssize_t started = 0; | ||
| 651 | 612 | ||
| 652 | curpage = 0; | 613 | get_dreq(dreq); |
| 653 | pgbase = dreq->user_addr & ~PAGE_MASK; | 614 | |
| 615 | pgbase = user_addr & ~PAGE_MASK; | ||
| 654 | do { | 616 | do { |
| 655 | struct nfs_write_data *data; | 617 | struct nfs_write_data *data; |
| 656 | size_t bytes; | 618 | size_t bytes; |
| 657 | 619 | ||
| 620 | result = -ENOMEM; | ||
| 621 | data = nfs_writedata_alloc(wpages); | ||
| 622 | if (unlikely(!data)) | ||
| 623 | break; | ||
| 624 | |||
| 658 | bytes = wsize; | 625 | bytes = wsize; |
| 659 | if (count < wsize) | 626 | if (count < wsize) |
| 660 | bytes = count; | 627 | bytes = count; |
| 661 | 628 | ||
| 662 | BUG_ON(list_empty(list)); | 629 | data->npages = nfs_direct_count_pages(user_addr, bytes); |
| 663 | data = list_entry(list->next, struct nfs_write_data, pages); | 630 | down_read(¤t->mm->mmap_sem); |
| 631 | result = get_user_pages(current, current->mm, user_addr, | ||
| 632 | data->npages, 0, 0, data->pagevec, NULL); | ||
| 633 | up_read(¤t->mm->mmap_sem); | ||
| 634 | if (unlikely(result < data->npages)) { | ||
| 635 | if (result > 0) | ||
| 636 | nfs_direct_release_pages(data->pagevec, result); | ||
| 637 | nfs_writedata_release(data); | ||
| 638 | break; | ||
| 639 | } | ||
| 640 | |||
| 641 | get_dreq(dreq); | ||
| 642 | |||
| 664 | list_move_tail(&data->pages, &dreq->rewrite_list); | 643 | list_move_tail(&data->pages, &dreq->rewrite_list); |
| 665 | 644 | ||
| 645 | data->req = (struct nfs_page *) dreq; | ||
| 666 | data->inode = inode; | 646 | data->inode = inode; |
| 667 | data->cred = ctx->cred; | 647 | data->cred = ctx->cred; |
| 668 | data->args.fh = NFS_FH(inode); | 648 | data->args.fh = NFS_FH(inode); |
| 669 | data->args.context = ctx; | 649 | data->args.context = ctx; |
| 670 | data->args.offset = pos; | 650 | data->args.offset = pos; |
| 671 | data->args.pgbase = pgbase; | 651 | data->args.pgbase = pgbase; |
| 672 | data->args.pages = &pages[curpage]; | 652 | data->args.pages = data->pagevec; |
| 673 | data->args.count = bytes; | 653 | data->args.count = bytes; |
| 674 | data->res.fattr = &data->fattr; | 654 | data->res.fattr = &data->fattr; |
| 675 | data->res.count = bytes; | 655 | data->res.count = bytes; |
| @@ -693,19 +673,26 @@ static void nfs_direct_write_schedule(struct nfs_direct_req *dreq, int sync) | |||
| 693 | bytes, | 673 | bytes, |
| 694 | (unsigned long long)data->args.offset); | 674 | (unsigned long long)data->args.offset); |
| 695 | 675 | ||
| 676 | started += bytes; | ||
| 677 | user_addr += bytes; | ||
| 696 | pos += bytes; | 678 | pos += bytes; |
| 697 | pgbase += bytes; | 679 | pgbase += bytes; |
| 698 | curpage += pgbase >> PAGE_SHIFT; | ||
| 699 | pgbase &= ~PAGE_MASK; | 680 | pgbase &= ~PAGE_MASK; |
| 700 | 681 | ||
| 701 | count -= bytes; | 682 | count -= bytes; |
| 702 | } while (count != 0); | 683 | } while (count != 0); |
| 703 | BUG_ON(!list_empty(list)); | 684 | |
| 685 | if (put_dreq(dreq)) | ||
| 686 | nfs_direct_write_complete(dreq, inode); | ||
| 687 | |||
| 688 | if (started) | ||
| 689 | return 0; | ||
| 690 | return result < 0 ? (ssize_t) result : -EFAULT; | ||
| 704 | } | 691 | } |
| 705 | 692 | ||
| 706 | static ssize_t nfs_direct_write(struct kiocb *iocb, unsigned long user_addr, size_t count, loff_t pos, struct page **pages, int nr_pages) | 693 | static ssize_t nfs_direct_write(struct kiocb *iocb, unsigned long user_addr, size_t count, loff_t pos) |
| 707 | { | 694 | { |
| 708 | ssize_t result; | 695 | ssize_t result = 0; |
| 709 | sigset_t oldset; | 696 | sigset_t oldset; |
| 710 | struct inode *inode = iocb->ki_filp->f_mapping->host; | 697 | struct inode *inode = iocb->ki_filp->f_mapping->host; |
| 711 | struct rpc_clnt *clnt = NFS_CLIENT(inode); | 698 | struct rpc_clnt *clnt = NFS_CLIENT(inode); |
| @@ -713,17 +700,14 @@ static ssize_t nfs_direct_write(struct kiocb *iocb, unsigned long user_addr, siz | |||
| 713 | size_t wsize = NFS_SERVER(inode)->wsize; | 700 | size_t wsize = NFS_SERVER(inode)->wsize; |
| 714 | int sync = 0; | 701 | int sync = 0; |
| 715 | 702 | ||
| 716 | dreq = nfs_direct_write_alloc(count, wsize); | 703 | dreq = nfs_direct_req_alloc(); |
| 717 | if (!dreq) | 704 | if (!dreq) |
| 718 | return -ENOMEM; | 705 | return -ENOMEM; |
| 706 | nfs_alloc_commit_data(dreq); | ||
| 707 | |||
| 719 | if (dreq->commit_data == NULL || count < wsize) | 708 | if (dreq->commit_data == NULL || count < wsize) |
| 720 | sync = FLUSH_STABLE; | 709 | sync = FLUSH_STABLE; |
| 721 | 710 | ||
| 722 | dreq->user_addr = user_addr; | ||
| 723 | dreq->user_count = count; | ||
| 724 | dreq->pos = pos; | ||
| 725 | dreq->pages = pages; | ||
| 726 | dreq->npages = nr_pages; | ||
| 727 | dreq->inode = inode; | 711 | dreq->inode = inode; |
| 728 | dreq->ctx = get_nfs_open_context((struct nfs_open_context *)iocb->ki_filp->private_data); | 712 | dreq->ctx = get_nfs_open_context((struct nfs_open_context *)iocb->ki_filp->private_data); |
| 729 | if (!is_sync_kiocb(iocb)) | 713 | if (!is_sync_kiocb(iocb)) |
| @@ -734,8 +718,9 @@ static ssize_t nfs_direct_write(struct kiocb *iocb, unsigned long user_addr, siz | |||
| 734 | nfs_begin_data_update(inode); | 718 | nfs_begin_data_update(inode); |
| 735 | 719 | ||
| 736 | rpc_clnt_sigmask(clnt, &oldset); | 720 | rpc_clnt_sigmask(clnt, &oldset); |
| 737 | nfs_direct_write_schedule(dreq, sync); | 721 | result = nfs_direct_write_schedule(dreq, user_addr, count, pos, sync); |
| 738 | result = nfs_direct_wait(dreq); | 722 | if (!result) |
| 723 | result = nfs_direct_wait(dreq); | ||
| 739 | rpc_clnt_sigunmask(clnt, &oldset); | 724 | rpc_clnt_sigunmask(clnt, &oldset); |
| 740 | 725 | ||
| 741 | return result; | 726 | return result; |
| @@ -765,8 +750,6 @@ static ssize_t nfs_direct_write(struct kiocb *iocb, unsigned long user_addr, siz | |||
| 765 | ssize_t nfs_file_direct_read(struct kiocb *iocb, char __user *buf, size_t count, loff_t pos) | 750 | ssize_t nfs_file_direct_read(struct kiocb *iocb, char __user *buf, size_t count, loff_t pos) |
| 766 | { | 751 | { |
| 767 | ssize_t retval = -EINVAL; | 752 | ssize_t retval = -EINVAL; |
| 768 | int page_count; | ||
| 769 | struct page **pages; | ||
| 770 | struct file *file = iocb->ki_filp; | 753 | struct file *file = iocb->ki_filp; |
| 771 | struct address_space *mapping = file->f_mapping; | 754 | struct address_space *mapping = file->f_mapping; |
| 772 | 755 | ||
| @@ -788,14 +771,7 @@ ssize_t nfs_file_direct_read(struct kiocb *iocb, char __user *buf, size_t count, | |||
| 788 | if (retval) | 771 | if (retval) |
| 789 | goto out; | 772 | goto out; |
| 790 | 773 | ||
| 791 | retval = nfs_get_user_pages(READ, (unsigned long) buf, | 774 | retval = nfs_direct_read(iocb, (unsigned long) buf, count, pos); |
| 792 | count, &pages); | ||
| 793 | if (retval < 0) | ||
| 794 | goto out; | ||
| 795 | page_count = retval; | ||
| 796 | |||
| 797 | retval = nfs_direct_read(iocb, (unsigned long) buf, count, pos, | ||
| 798 | pages, page_count); | ||
| 799 | if (retval > 0) | 775 | if (retval > 0) |
| 800 | iocb->ki_pos = pos + retval; | 776 | iocb->ki_pos = pos + retval; |
| 801 | 777 | ||
| @@ -831,8 +807,6 @@ out: | |||
| 831 | ssize_t nfs_file_direct_write(struct kiocb *iocb, const char __user *buf, size_t count, loff_t pos) | 807 | ssize_t nfs_file_direct_write(struct kiocb *iocb, const char __user *buf, size_t count, loff_t pos) |
| 832 | { | 808 | { |
| 833 | ssize_t retval; | 809 | ssize_t retval; |
| 834 | int page_count; | ||
| 835 | struct page **pages; | ||
| 836 | struct file *file = iocb->ki_filp; | 810 | struct file *file = iocb->ki_filp; |
| 837 | struct address_space *mapping = file->f_mapping; | 811 | struct address_space *mapping = file->f_mapping; |
| 838 | 812 | ||
| @@ -860,14 +834,7 @@ ssize_t nfs_file_direct_write(struct kiocb *iocb, const char __user *buf, size_t | |||
| 860 | if (retval) | 834 | if (retval) |
| 861 | goto out; | 835 | goto out; |
| 862 | 836 | ||
| 863 | retval = nfs_get_user_pages(WRITE, (unsigned long) buf, | 837 | retval = nfs_direct_write(iocb, (unsigned long) buf, count, pos); |
| 864 | count, &pages); | ||
| 865 | if (retval < 0) | ||
| 866 | goto out; | ||
| 867 | page_count = retval; | ||
| 868 | |||
| 869 | retval = nfs_direct_write(iocb, (unsigned long) buf, count, | ||
| 870 | pos, pages, page_count); | ||
| 871 | 838 | ||
| 872 | /* | 839 | /* |
| 873 | * XXX: nfs_end_data_update() already ensures this file's | 840 | * XXX: nfs_end_data_update() already ensures this file's |
diff --git a/include/linux/nfs_xdr.h b/include/linux/nfs_xdr.h index 7c7320fa51aa..2d3fb6416d91 100644 --- a/include/linux/nfs_xdr.h +++ b/include/linux/nfs_xdr.h | |||
| @@ -729,6 +729,7 @@ struct nfs_read_data { | |||
| 729 | struct list_head pages; /* Coalesced read requests */ | 729 | struct list_head pages; /* Coalesced read requests */ |
| 730 | struct nfs_page *req; /* multi ops per nfs_page */ | 730 | struct nfs_page *req; /* multi ops per nfs_page */ |
| 731 | struct page **pagevec; | 731 | struct page **pagevec; |
| 732 | unsigned int npages; /* active pages in pagevec */ | ||
| 732 | struct nfs_readargs args; | 733 | struct nfs_readargs args; |
| 733 | struct nfs_readres res; | 734 | struct nfs_readres res; |
| 734 | #ifdef CONFIG_NFS_V4 | 735 | #ifdef CONFIG_NFS_V4 |
| @@ -747,6 +748,7 @@ struct nfs_write_data { | |||
| 747 | struct list_head pages; /* Coalesced requests we wish to flush */ | 748 | struct list_head pages; /* Coalesced requests we wish to flush */ |
| 748 | struct nfs_page *req; /* multi ops per nfs_page */ | 749 | struct nfs_page *req; /* multi ops per nfs_page */ |
| 749 | struct page **pagevec; | 750 | struct page **pagevec; |
| 751 | unsigned int npages; /* active pages in pagevec */ | ||
| 750 | struct nfs_writeargs args; /* argument struct */ | 752 | struct nfs_writeargs args; /* argument struct */ |
| 751 | struct nfs_writeres res; /* result struct */ | 753 | struct nfs_writeres res; /* result struct */ |
| 752 | #ifdef CONFIG_NFS_V4 | 754 | #ifdef CONFIG_NFS_V4 |
