diff options
author | Linus Torvalds <torvalds@linux-foundation.org> | 2016-10-07 18:36:58 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2016-10-07 18:36:58 -0400 |
commit | d1f5323370fceaed43a7ee38f4c7bfc7e70f28d0 (patch) | |
tree | cadb1dc22207a4e1838b7af31ac3fc15363e809b /lib | |
parent | 2eee010d092903ee95716b6c2fbd9d3289839aa4 (diff) | |
parent | a949e63992469fed87aef197347960ced31701b8 (diff) |
Merge branch 'work.splice_read' of git://git.kernel.org/pub/scm/linux/kernel/git/viro/vfs
Pull VFS splice updates from Al Viro:
"There's a bunch of branches this cycle, both mine and from other folks
and I'd rather send pull requests separately.
This one is the conversion of ->splice_read() to ITER_PIPE iov_iter
(and introduction of such). Gets rid of a lot of code in fs/splice.c
and elsewhere; there will be followups, but these are for the next
cycle... Some pipe/splice-related cleanups from Miklos in the same
branch as well"
* 'work.splice_read' of git://git.kernel.org/pub/scm/linux/kernel/git/viro/vfs:
pipe: fix comment in pipe_buf_operations
pipe: add pipe_buf_steal() helper
pipe: add pipe_buf_confirm() helper
pipe: add pipe_buf_release() helper
pipe: add pipe_buf_get() helper
relay: simplify relay_file_read()
switch default_file_splice_read() to use of pipe-backed iov_iter
switch generic_file_splice_read() to use of ->read_iter()
new iov_iter flavour: pipe-backed
fuse_dev_splice_read(): switch to add_to_pipe()
skb_splice_bits(): get rid of callback
new helper: add_to_pipe()
splice: lift pipe_lock out of splice_to_pipe()
splice: switch get_iovec_page_array() to iov_iter
splice_to_pipe(): don't open-code wakeup_pipe_readers()
consistent treatment of EFAULT on O_DIRECT read/write
Diffstat (limited to 'lib')
-rw-r--r-- | lib/iov_iter.c | 395 |
1 files changed, 393 insertions, 2 deletions
diff --git a/lib/iov_iter.c b/lib/iov_iter.c index 7e3138cfc8c9..48b8c27acabb 100644 --- a/lib/iov_iter.c +++ b/lib/iov_iter.c | |||
@@ -3,8 +3,11 @@ | |||
3 | #include <linux/pagemap.h> | 3 | #include <linux/pagemap.h> |
4 | #include <linux/slab.h> | 4 | #include <linux/slab.h> |
5 | #include <linux/vmalloc.h> | 5 | #include <linux/vmalloc.h> |
6 | #include <linux/splice.h> | ||
6 | #include <net/checksum.h> | 7 | #include <net/checksum.h> |
7 | 8 | ||
9 | #define PIPE_PARANOIA /* for now */ | ||
10 | |||
8 | #define iterate_iovec(i, n, __v, __p, skip, STEP) { \ | 11 | #define iterate_iovec(i, n, __v, __p, skip, STEP) { \ |
9 | size_t left; \ | 12 | size_t left; \ |
10 | size_t wanted = n; \ | 13 | size_t wanted = n; \ |
@@ -290,6 +293,93 @@ done: | |||
290 | return wanted - bytes; | 293 | return wanted - bytes; |
291 | } | 294 | } |
292 | 295 | ||
296 | #ifdef PIPE_PARANOIA | ||
297 | static bool sanity(const struct iov_iter *i) | ||
298 | { | ||
299 | struct pipe_inode_info *pipe = i->pipe; | ||
300 | int idx = i->idx; | ||
301 | int next = pipe->curbuf + pipe->nrbufs; | ||
302 | if (i->iov_offset) { | ||
303 | struct pipe_buffer *p; | ||
304 | if (unlikely(!pipe->nrbufs)) | ||
305 | goto Bad; // pipe must be non-empty | ||
306 | if (unlikely(idx != ((next - 1) & (pipe->buffers - 1)))) | ||
307 | goto Bad; // must be at the last buffer... | ||
308 | |||
309 | p = &pipe->bufs[idx]; | ||
310 | if (unlikely(p->offset + p->len != i->iov_offset)) | ||
311 | goto Bad; // ... at the end of segment | ||
312 | } else { | ||
313 | if (idx != (next & (pipe->buffers - 1))) | ||
314 | goto Bad; // must be right after the last buffer | ||
315 | } | ||
316 | return true; | ||
317 | Bad: | ||
318 | printk(KERN_ERR "idx = %d, offset = %zd\n", i->idx, i->iov_offset); | ||
319 | printk(KERN_ERR "curbuf = %d, nrbufs = %d, buffers = %d\n", | ||
320 | pipe->curbuf, pipe->nrbufs, pipe->buffers); | ||
321 | for (idx = 0; idx < pipe->buffers; idx++) | ||
322 | printk(KERN_ERR "[%p %p %d %d]\n", | ||
323 | pipe->bufs[idx].ops, | ||
324 | pipe->bufs[idx].page, | ||
325 | pipe->bufs[idx].offset, | ||
326 | pipe->bufs[idx].len); | ||
327 | WARN_ON(1); | ||
328 | return false; | ||
329 | } | ||
330 | #else | ||
331 | #define sanity(i) true | ||
332 | #endif | ||
333 | |||
334 | static inline int next_idx(int idx, struct pipe_inode_info *pipe) | ||
335 | { | ||
336 | return (idx + 1) & (pipe->buffers - 1); | ||
337 | } | ||
338 | |||
339 | static size_t copy_page_to_iter_pipe(struct page *page, size_t offset, size_t bytes, | ||
340 | struct iov_iter *i) | ||
341 | { | ||
342 | struct pipe_inode_info *pipe = i->pipe; | ||
343 | struct pipe_buffer *buf; | ||
344 | size_t off; | ||
345 | int idx; | ||
346 | |||
347 | if (unlikely(bytes > i->count)) | ||
348 | bytes = i->count; | ||
349 | |||
350 | if (unlikely(!bytes)) | ||
351 | return 0; | ||
352 | |||
353 | if (!sanity(i)) | ||
354 | return 0; | ||
355 | |||
356 | off = i->iov_offset; | ||
357 | idx = i->idx; | ||
358 | buf = &pipe->bufs[idx]; | ||
359 | if (off) { | ||
360 | if (offset == off && buf->page == page) { | ||
361 | /* merge with the last one */ | ||
362 | buf->len += bytes; | ||
363 | i->iov_offset += bytes; | ||
364 | goto out; | ||
365 | } | ||
366 | idx = next_idx(idx, pipe); | ||
367 | buf = &pipe->bufs[idx]; | ||
368 | } | ||
369 | if (idx == pipe->curbuf && pipe->nrbufs) | ||
370 | return 0; | ||
371 | pipe->nrbufs++; | ||
372 | buf->ops = &page_cache_pipe_buf_ops; | ||
373 | get_page(buf->page = page); | ||
374 | buf->offset = offset; | ||
375 | buf->len = bytes; | ||
376 | i->iov_offset = offset + bytes; | ||
377 | i->idx = idx; | ||
378 | out: | ||
379 | i->count -= bytes; | ||
380 | return bytes; | ||
381 | } | ||
382 | |||
293 | /* | 383 | /* |
294 | * Fault in one or more iovecs of the given iov_iter, to a maximum length of | 384 | * Fault in one or more iovecs of the given iov_iter, to a maximum length of |
295 | * bytes. For each iovec, fault in each page that constitutes the iovec. | 385 | * bytes. For each iovec, fault in each page that constitutes the iovec. |
@@ -356,9 +446,98 @@ static void memzero_page(struct page *page, size_t offset, size_t len) | |||
356 | kunmap_atomic(addr); | 446 | kunmap_atomic(addr); |
357 | } | 447 | } |
358 | 448 | ||
449 | static inline bool allocated(struct pipe_buffer *buf) | ||
450 | { | ||
451 | return buf->ops == &default_pipe_buf_ops; | ||
452 | } | ||
453 | |||
454 | static inline void data_start(const struct iov_iter *i, int *idxp, size_t *offp) | ||
455 | { | ||
456 | size_t off = i->iov_offset; | ||
457 | int idx = i->idx; | ||
458 | if (off && (!allocated(&i->pipe->bufs[idx]) || off == PAGE_SIZE)) { | ||
459 | idx = next_idx(idx, i->pipe); | ||
460 | off = 0; | ||
461 | } | ||
462 | *idxp = idx; | ||
463 | *offp = off; | ||
464 | } | ||
465 | |||
466 | static size_t push_pipe(struct iov_iter *i, size_t size, | ||
467 | int *idxp, size_t *offp) | ||
468 | { | ||
469 | struct pipe_inode_info *pipe = i->pipe; | ||
470 | size_t off; | ||
471 | int idx; | ||
472 | ssize_t left; | ||
473 | |||
474 | if (unlikely(size > i->count)) | ||
475 | size = i->count; | ||
476 | if (unlikely(!size)) | ||
477 | return 0; | ||
478 | |||
479 | left = size; | ||
480 | data_start(i, &idx, &off); | ||
481 | *idxp = idx; | ||
482 | *offp = off; | ||
483 | if (off) { | ||
484 | left -= PAGE_SIZE - off; | ||
485 | if (left <= 0) { | ||
486 | pipe->bufs[idx].len += size; | ||
487 | return size; | ||
488 | } | ||
489 | pipe->bufs[idx].len = PAGE_SIZE; | ||
490 | idx = next_idx(idx, pipe); | ||
491 | } | ||
492 | while (idx != pipe->curbuf || !pipe->nrbufs) { | ||
493 | struct page *page = alloc_page(GFP_USER); | ||
494 | if (!page) | ||
495 | break; | ||
496 | pipe->nrbufs++; | ||
497 | pipe->bufs[idx].ops = &default_pipe_buf_ops; | ||
498 | pipe->bufs[idx].page = page; | ||
499 | pipe->bufs[idx].offset = 0; | ||
500 | if (left <= PAGE_SIZE) { | ||
501 | pipe->bufs[idx].len = left; | ||
502 | return size; | ||
503 | } | ||
504 | pipe->bufs[idx].len = PAGE_SIZE; | ||
505 | left -= PAGE_SIZE; | ||
506 | idx = next_idx(idx, pipe); | ||
507 | } | ||
508 | return size - left; | ||
509 | } | ||
510 | |||
511 | static size_t copy_pipe_to_iter(const void *addr, size_t bytes, | ||
512 | struct iov_iter *i) | ||
513 | { | ||
514 | struct pipe_inode_info *pipe = i->pipe; | ||
515 | size_t n, off; | ||
516 | int idx; | ||
517 | |||
518 | if (!sanity(i)) | ||
519 | return 0; | ||
520 | |||
521 | bytes = n = push_pipe(i, bytes, &idx, &off); | ||
522 | if (unlikely(!n)) | ||
523 | return 0; | ||
524 | for ( ; n; idx = next_idx(idx, pipe), off = 0) { | ||
525 | size_t chunk = min_t(size_t, n, PAGE_SIZE - off); | ||
526 | memcpy_to_page(pipe->bufs[idx].page, off, addr, chunk); | ||
527 | i->idx = idx; | ||
528 | i->iov_offset = off + chunk; | ||
529 | n -= chunk; | ||
530 | addr += chunk; | ||
531 | } | ||
532 | i->count -= bytes; | ||
533 | return bytes; | ||
534 | } | ||
535 | |||
359 | size_t copy_to_iter(const void *addr, size_t bytes, struct iov_iter *i) | 536 | size_t copy_to_iter(const void *addr, size_t bytes, struct iov_iter *i) |
360 | { | 537 | { |
361 | const char *from = addr; | 538 | const char *from = addr; |
539 | if (unlikely(i->type & ITER_PIPE)) | ||
540 | return copy_pipe_to_iter(addr, bytes, i); | ||
362 | iterate_and_advance(i, bytes, v, | 541 | iterate_and_advance(i, bytes, v, |
363 | __copy_to_user(v.iov_base, (from += v.iov_len) - v.iov_len, | 542 | __copy_to_user(v.iov_base, (from += v.iov_len) - v.iov_len, |
364 | v.iov_len), | 543 | v.iov_len), |
@@ -374,6 +553,10 @@ EXPORT_SYMBOL(copy_to_iter); | |||
374 | size_t copy_from_iter(void *addr, size_t bytes, struct iov_iter *i) | 553 | size_t copy_from_iter(void *addr, size_t bytes, struct iov_iter *i) |
375 | { | 554 | { |
376 | char *to = addr; | 555 | char *to = addr; |
556 | if (unlikely(i->type & ITER_PIPE)) { | ||
557 | WARN_ON(1); | ||
558 | return 0; | ||
559 | } | ||
377 | iterate_and_advance(i, bytes, v, | 560 | iterate_and_advance(i, bytes, v, |
378 | __copy_from_user((to += v.iov_len) - v.iov_len, v.iov_base, | 561 | __copy_from_user((to += v.iov_len) - v.iov_len, v.iov_base, |
379 | v.iov_len), | 562 | v.iov_len), |
@@ -389,6 +572,10 @@ EXPORT_SYMBOL(copy_from_iter); | |||
389 | size_t copy_from_iter_nocache(void *addr, size_t bytes, struct iov_iter *i) | 572 | size_t copy_from_iter_nocache(void *addr, size_t bytes, struct iov_iter *i) |
390 | { | 573 | { |
391 | char *to = addr; | 574 | char *to = addr; |
575 | if (unlikely(i->type & ITER_PIPE)) { | ||
576 | WARN_ON(1); | ||
577 | return 0; | ||
578 | } | ||
392 | iterate_and_advance(i, bytes, v, | 579 | iterate_and_advance(i, bytes, v, |
393 | __copy_from_user_nocache((to += v.iov_len) - v.iov_len, | 580 | __copy_from_user_nocache((to += v.iov_len) - v.iov_len, |
394 | v.iov_base, v.iov_len), | 581 | v.iov_base, v.iov_len), |
@@ -409,14 +596,20 @@ size_t copy_page_to_iter(struct page *page, size_t offset, size_t bytes, | |||
409 | size_t wanted = copy_to_iter(kaddr + offset, bytes, i); | 596 | size_t wanted = copy_to_iter(kaddr + offset, bytes, i); |
410 | kunmap_atomic(kaddr); | 597 | kunmap_atomic(kaddr); |
411 | return wanted; | 598 | return wanted; |
412 | } else | 599 | } else if (likely(!(i->type & ITER_PIPE))) |
413 | return copy_page_to_iter_iovec(page, offset, bytes, i); | 600 | return copy_page_to_iter_iovec(page, offset, bytes, i); |
601 | else | ||
602 | return copy_page_to_iter_pipe(page, offset, bytes, i); | ||
414 | } | 603 | } |
415 | EXPORT_SYMBOL(copy_page_to_iter); | 604 | EXPORT_SYMBOL(copy_page_to_iter); |
416 | 605 | ||
417 | size_t copy_page_from_iter(struct page *page, size_t offset, size_t bytes, | 606 | size_t copy_page_from_iter(struct page *page, size_t offset, size_t bytes, |
418 | struct iov_iter *i) | 607 | struct iov_iter *i) |
419 | { | 608 | { |
609 | if (unlikely(i->type & ITER_PIPE)) { | ||
610 | WARN_ON(1); | ||
611 | return 0; | ||
612 | } | ||
420 | if (i->type & (ITER_BVEC|ITER_KVEC)) { | 613 | if (i->type & (ITER_BVEC|ITER_KVEC)) { |
421 | void *kaddr = kmap_atomic(page); | 614 | void *kaddr = kmap_atomic(page); |
422 | size_t wanted = copy_from_iter(kaddr + offset, bytes, i); | 615 | size_t wanted = copy_from_iter(kaddr + offset, bytes, i); |
@@ -427,8 +620,34 @@ size_t copy_page_from_iter(struct page *page, size_t offset, size_t bytes, | |||
427 | } | 620 | } |
428 | EXPORT_SYMBOL(copy_page_from_iter); | 621 | EXPORT_SYMBOL(copy_page_from_iter); |
429 | 622 | ||
623 | static size_t pipe_zero(size_t bytes, struct iov_iter *i) | ||
624 | { | ||
625 | struct pipe_inode_info *pipe = i->pipe; | ||
626 | size_t n, off; | ||
627 | int idx; | ||
628 | |||
629 | if (!sanity(i)) | ||
630 | return 0; | ||
631 | |||
632 | bytes = n = push_pipe(i, bytes, &idx, &off); | ||
633 | if (unlikely(!n)) | ||
634 | return 0; | ||
635 | |||
636 | for ( ; n; idx = next_idx(idx, pipe), off = 0) { | ||
637 | size_t chunk = min_t(size_t, n, PAGE_SIZE - off); | ||
638 | memzero_page(pipe->bufs[idx].page, off, chunk); | ||
639 | i->idx = idx; | ||
640 | i->iov_offset = off + chunk; | ||
641 | n -= chunk; | ||
642 | } | ||
643 | i->count -= bytes; | ||
644 | return bytes; | ||
645 | } | ||
646 | |||
430 | size_t iov_iter_zero(size_t bytes, struct iov_iter *i) | 647 | size_t iov_iter_zero(size_t bytes, struct iov_iter *i) |
431 | { | 648 | { |
649 | if (unlikely(i->type & ITER_PIPE)) | ||
650 | return pipe_zero(bytes, i); | ||
432 | iterate_and_advance(i, bytes, v, | 651 | iterate_and_advance(i, bytes, v, |
433 | __clear_user(v.iov_base, v.iov_len), | 652 | __clear_user(v.iov_base, v.iov_len), |
434 | memzero_page(v.bv_page, v.bv_offset, v.bv_len), | 653 | memzero_page(v.bv_page, v.bv_offset, v.bv_len), |
@@ -443,6 +662,11 @@ size_t iov_iter_copy_from_user_atomic(struct page *page, | |||
443 | struct iov_iter *i, unsigned long offset, size_t bytes) | 662 | struct iov_iter *i, unsigned long offset, size_t bytes) |
444 | { | 663 | { |
445 | char *kaddr = kmap_atomic(page), *p = kaddr + offset; | 664 | char *kaddr = kmap_atomic(page), *p = kaddr + offset; |
665 | if (unlikely(i->type & ITER_PIPE)) { | ||
666 | kunmap_atomic(kaddr); | ||
667 | WARN_ON(1); | ||
668 | return 0; | ||
669 | } | ||
446 | iterate_all_kinds(i, bytes, v, | 670 | iterate_all_kinds(i, bytes, v, |
447 | __copy_from_user_inatomic((p += v.iov_len) - v.iov_len, | 671 | __copy_from_user_inatomic((p += v.iov_len) - v.iov_len, |
448 | v.iov_base, v.iov_len), | 672 | v.iov_base, v.iov_len), |
@@ -455,8 +679,49 @@ size_t iov_iter_copy_from_user_atomic(struct page *page, | |||
455 | } | 679 | } |
456 | EXPORT_SYMBOL(iov_iter_copy_from_user_atomic); | 680 | EXPORT_SYMBOL(iov_iter_copy_from_user_atomic); |
457 | 681 | ||
682 | static void pipe_advance(struct iov_iter *i, size_t size) | ||
683 | { | ||
684 | struct pipe_inode_info *pipe = i->pipe; | ||
685 | struct pipe_buffer *buf; | ||
686 | int idx = i->idx; | ||
687 | size_t off = i->iov_offset; | ||
688 | |||
689 | if (unlikely(i->count < size)) | ||
690 | size = i->count; | ||
691 | |||
692 | if (size) { | ||
693 | if (off) /* make it relative to the beginning of buffer */ | ||
694 | size += off - pipe->bufs[idx].offset; | ||
695 | while (1) { | ||
696 | buf = &pipe->bufs[idx]; | ||
697 | if (size <= buf->len) | ||
698 | break; | ||
699 | size -= buf->len; | ||
700 | idx = next_idx(idx, pipe); | ||
701 | } | ||
702 | buf->len = size; | ||
703 | i->idx = idx; | ||
704 | off = i->iov_offset = buf->offset + size; | ||
705 | } | ||
706 | if (off) | ||
707 | idx = next_idx(idx, pipe); | ||
708 | if (pipe->nrbufs) { | ||
709 | int unused = (pipe->curbuf + pipe->nrbufs) & (pipe->buffers - 1); | ||
710 | /* [curbuf,unused) is in use. Free [idx,unused) */ | ||
711 | while (idx != unused) { | ||
712 | pipe_buf_release(pipe, &pipe->bufs[idx]); | ||
713 | idx = next_idx(idx, pipe); | ||
714 | pipe->nrbufs--; | ||
715 | } | ||
716 | } | ||
717 | } | ||
718 | |||
458 | void iov_iter_advance(struct iov_iter *i, size_t size) | 719 | void iov_iter_advance(struct iov_iter *i, size_t size) |
459 | { | 720 | { |
721 | if (unlikely(i->type & ITER_PIPE)) { | ||
722 | pipe_advance(i, size); | ||
723 | return; | ||
724 | } | ||
460 | iterate_and_advance(i, size, v, 0, 0, 0) | 725 | iterate_and_advance(i, size, v, 0, 0, 0) |
461 | } | 726 | } |
462 | EXPORT_SYMBOL(iov_iter_advance); | 727 | EXPORT_SYMBOL(iov_iter_advance); |
@@ -466,6 +731,8 @@ EXPORT_SYMBOL(iov_iter_advance); | |||
466 | */ | 731 | */ |
467 | size_t iov_iter_single_seg_count(const struct iov_iter *i) | 732 | size_t iov_iter_single_seg_count(const struct iov_iter *i) |
468 | { | 733 | { |
734 | if (unlikely(i->type & ITER_PIPE)) | ||
735 | return i->count; // it is a silly place, anyway | ||
469 | if (i->nr_segs == 1) | 736 | if (i->nr_segs == 1) |
470 | return i->count; | 737 | return i->count; |
471 | else if (i->type & ITER_BVEC) | 738 | else if (i->type & ITER_BVEC) |
@@ -501,6 +768,19 @@ void iov_iter_bvec(struct iov_iter *i, int direction, | |||
501 | } | 768 | } |
502 | EXPORT_SYMBOL(iov_iter_bvec); | 769 | EXPORT_SYMBOL(iov_iter_bvec); |
503 | 770 | ||
771 | void iov_iter_pipe(struct iov_iter *i, int direction, | ||
772 | struct pipe_inode_info *pipe, | ||
773 | size_t count) | ||
774 | { | ||
775 | BUG_ON(direction != ITER_PIPE); | ||
776 | i->type = direction; | ||
777 | i->pipe = pipe; | ||
778 | i->idx = (pipe->curbuf + pipe->nrbufs) & (pipe->buffers - 1); | ||
779 | i->iov_offset = 0; | ||
780 | i->count = count; | ||
781 | } | ||
782 | EXPORT_SYMBOL(iov_iter_pipe); | ||
783 | |||
504 | unsigned long iov_iter_alignment(const struct iov_iter *i) | 784 | unsigned long iov_iter_alignment(const struct iov_iter *i) |
505 | { | 785 | { |
506 | unsigned long res = 0; | 786 | unsigned long res = 0; |
@@ -509,6 +789,11 @@ unsigned long iov_iter_alignment(const struct iov_iter *i) | |||
509 | if (!size) | 789 | if (!size) |
510 | return 0; | 790 | return 0; |
511 | 791 | ||
792 | if (unlikely(i->type & ITER_PIPE)) { | ||
793 | if (i->iov_offset && allocated(&i->pipe->bufs[i->idx])) | ||
794 | return size | i->iov_offset; | ||
795 | return size; | ||
796 | } | ||
512 | iterate_all_kinds(i, size, v, | 797 | iterate_all_kinds(i, size, v, |
513 | (res |= (unsigned long)v.iov_base | v.iov_len, 0), | 798 | (res |= (unsigned long)v.iov_base | v.iov_len, 0), |
514 | res |= v.bv_offset | v.bv_len, | 799 | res |= v.bv_offset | v.bv_len, |
@@ -525,6 +810,11 @@ unsigned long iov_iter_gap_alignment(const struct iov_iter *i) | |||
525 | if (!size) | 810 | if (!size) |
526 | return 0; | 811 | return 0; |
527 | 812 | ||
813 | if (unlikely(i->type & ITER_PIPE)) { | ||
814 | WARN_ON(1); | ||
815 | return ~0U; | ||
816 | } | ||
817 | |||
528 | iterate_all_kinds(i, size, v, | 818 | iterate_all_kinds(i, size, v, |
529 | (res |= (!res ? 0 : (unsigned long)v.iov_base) | | 819 | (res |= (!res ? 0 : (unsigned long)v.iov_base) | |
530 | (size != v.iov_len ? size : 0), 0), | 820 | (size != v.iov_len ? size : 0), 0), |
@@ -537,6 +827,47 @@ unsigned long iov_iter_gap_alignment(const struct iov_iter *i) | |||
537 | } | 827 | } |
538 | EXPORT_SYMBOL(iov_iter_gap_alignment); | 828 | EXPORT_SYMBOL(iov_iter_gap_alignment); |
539 | 829 | ||
830 | static inline size_t __pipe_get_pages(struct iov_iter *i, | ||
831 | size_t maxsize, | ||
832 | struct page **pages, | ||
833 | int idx, | ||
834 | size_t *start) | ||
835 | { | ||
836 | struct pipe_inode_info *pipe = i->pipe; | ||
837 | size_t n = push_pipe(i, maxsize, &idx, start); | ||
838 | if (!n) | ||
839 | return -EFAULT; | ||
840 | |||
841 | maxsize = n; | ||
842 | n += *start; | ||
843 | while (n >= PAGE_SIZE) { | ||
844 | get_page(*pages++ = pipe->bufs[idx].page); | ||
845 | idx = next_idx(idx, pipe); | ||
846 | n -= PAGE_SIZE; | ||
847 | } | ||
848 | |||
849 | return maxsize; | ||
850 | } | ||
851 | |||
852 | static ssize_t pipe_get_pages(struct iov_iter *i, | ||
853 | struct page **pages, size_t maxsize, unsigned maxpages, | ||
854 | size_t *start) | ||
855 | { | ||
856 | unsigned npages; | ||
857 | size_t capacity; | ||
858 | int idx; | ||
859 | |||
860 | if (!sanity(i)) | ||
861 | return -EFAULT; | ||
862 | |||
863 | data_start(i, &idx, start); | ||
864 | /* some of this one + all after this one */ | ||
865 | npages = ((i->pipe->curbuf - idx - 1) & (i->pipe->buffers - 1)) + 1; | ||
866 | capacity = min(npages,maxpages) * PAGE_SIZE - *start; | ||
867 | |||
868 | return __pipe_get_pages(i, min(maxsize, capacity), pages, idx, start); | ||
869 | } | ||
870 | |||
540 | ssize_t iov_iter_get_pages(struct iov_iter *i, | 871 | ssize_t iov_iter_get_pages(struct iov_iter *i, |
541 | struct page **pages, size_t maxsize, unsigned maxpages, | 872 | struct page **pages, size_t maxsize, unsigned maxpages, |
542 | size_t *start) | 873 | size_t *start) |
@@ -547,6 +878,8 @@ ssize_t iov_iter_get_pages(struct iov_iter *i, | |||
547 | if (!maxsize) | 878 | if (!maxsize) |
548 | return 0; | 879 | return 0; |
549 | 880 | ||
881 | if (unlikely(i->type & ITER_PIPE)) | ||
882 | return pipe_get_pages(i, pages, maxsize, maxpages, start); | ||
550 | iterate_all_kinds(i, maxsize, v, ({ | 883 | iterate_all_kinds(i, maxsize, v, ({ |
551 | unsigned long addr = (unsigned long)v.iov_base; | 884 | unsigned long addr = (unsigned long)v.iov_base; |
552 | size_t len = v.iov_len + (*start = addr & (PAGE_SIZE - 1)); | 885 | size_t len = v.iov_len + (*start = addr & (PAGE_SIZE - 1)); |
@@ -582,6 +915,37 @@ static struct page **get_pages_array(size_t n) | |||
582 | return p; | 915 | return p; |
583 | } | 916 | } |
584 | 917 | ||
918 | static ssize_t pipe_get_pages_alloc(struct iov_iter *i, | ||
919 | struct page ***pages, size_t maxsize, | ||
920 | size_t *start) | ||
921 | { | ||
922 | struct page **p; | ||
923 | size_t n; | ||
924 | int idx; | ||
925 | int npages; | ||
926 | |||
927 | if (!sanity(i)) | ||
928 | return -EFAULT; | ||
929 | |||
930 | data_start(i, &idx, start); | ||
931 | /* some of this one + all after this one */ | ||
932 | npages = ((i->pipe->curbuf - idx - 1) & (i->pipe->buffers - 1)) + 1; | ||
933 | n = npages * PAGE_SIZE - *start; | ||
934 | if (maxsize > n) | ||
935 | maxsize = n; | ||
936 | else | ||
937 | npages = DIV_ROUND_UP(maxsize + *start, PAGE_SIZE); | ||
938 | p = get_pages_array(npages); | ||
939 | if (!p) | ||
940 | return -ENOMEM; | ||
941 | n = __pipe_get_pages(i, maxsize, p, idx, start); | ||
942 | if (n > 0) | ||
943 | *pages = p; | ||
944 | else | ||
945 | kvfree(p); | ||
946 | return n; | ||
947 | } | ||
948 | |||
585 | ssize_t iov_iter_get_pages_alloc(struct iov_iter *i, | 949 | ssize_t iov_iter_get_pages_alloc(struct iov_iter *i, |
586 | struct page ***pages, size_t maxsize, | 950 | struct page ***pages, size_t maxsize, |
587 | size_t *start) | 951 | size_t *start) |
@@ -594,6 +958,8 @@ ssize_t iov_iter_get_pages_alloc(struct iov_iter *i, | |||
594 | if (!maxsize) | 958 | if (!maxsize) |
595 | return 0; | 959 | return 0; |
596 | 960 | ||
961 | if (unlikely(i->type & ITER_PIPE)) | ||
962 | return pipe_get_pages_alloc(i, pages, maxsize, start); | ||
597 | iterate_all_kinds(i, maxsize, v, ({ | 963 | iterate_all_kinds(i, maxsize, v, ({ |
598 | unsigned long addr = (unsigned long)v.iov_base; | 964 | unsigned long addr = (unsigned long)v.iov_base; |
599 | size_t len = v.iov_len + (*start = addr & (PAGE_SIZE - 1)); | 965 | size_t len = v.iov_len + (*start = addr & (PAGE_SIZE - 1)); |
@@ -635,6 +1001,10 @@ size_t csum_and_copy_from_iter(void *addr, size_t bytes, __wsum *csum, | |||
635 | __wsum sum, next; | 1001 | __wsum sum, next; |
636 | size_t off = 0; | 1002 | size_t off = 0; |
637 | sum = *csum; | 1003 | sum = *csum; |
1004 | if (unlikely(i->type & ITER_PIPE)) { | ||
1005 | WARN_ON(1); | ||
1006 | return 0; | ||
1007 | } | ||
638 | iterate_and_advance(i, bytes, v, ({ | 1008 | iterate_and_advance(i, bytes, v, ({ |
639 | int err = 0; | 1009 | int err = 0; |
640 | next = csum_and_copy_from_user(v.iov_base, | 1010 | next = csum_and_copy_from_user(v.iov_base, |
@@ -673,6 +1043,10 @@ size_t csum_and_copy_to_iter(const void *addr, size_t bytes, __wsum *csum, | |||
673 | __wsum sum, next; | 1043 | __wsum sum, next; |
674 | size_t off = 0; | 1044 | size_t off = 0; |
675 | sum = *csum; | 1045 | sum = *csum; |
1046 | if (unlikely(i->type & ITER_PIPE)) { | ||
1047 | WARN_ON(1); /* for now */ | ||
1048 | return 0; | ||
1049 | } | ||
676 | iterate_and_advance(i, bytes, v, ({ | 1050 | iterate_and_advance(i, bytes, v, ({ |
677 | int err = 0; | 1051 | int err = 0; |
678 | next = csum_and_copy_to_user((from += v.iov_len) - v.iov_len, | 1052 | next = csum_and_copy_to_user((from += v.iov_len) - v.iov_len, |
@@ -712,7 +1086,20 @@ int iov_iter_npages(const struct iov_iter *i, int maxpages) | |||
712 | if (!size) | 1086 | if (!size) |
713 | return 0; | 1087 | return 0; |
714 | 1088 | ||
715 | iterate_all_kinds(i, size, v, ({ | 1089 | if (unlikely(i->type & ITER_PIPE)) { |
1090 | struct pipe_inode_info *pipe = i->pipe; | ||
1091 | size_t off; | ||
1092 | int idx; | ||
1093 | |||
1094 | if (!sanity(i)) | ||
1095 | return 0; | ||
1096 | |||
1097 | data_start(i, &idx, &off); | ||
1098 | /* some of this one + all after this one */ | ||
1099 | npages = ((pipe->curbuf - idx - 1) & (pipe->buffers - 1)) + 1; | ||
1100 | if (npages >= maxpages) | ||
1101 | return maxpages; | ||
1102 | } else iterate_all_kinds(i, size, v, ({ | ||
716 | unsigned long p = (unsigned long)v.iov_base; | 1103 | unsigned long p = (unsigned long)v.iov_base; |
717 | npages += DIV_ROUND_UP(p + v.iov_len, PAGE_SIZE) | 1104 | npages += DIV_ROUND_UP(p + v.iov_len, PAGE_SIZE) |
718 | - p / PAGE_SIZE; | 1105 | - p / PAGE_SIZE; |
@@ -737,6 +1124,10 @@ EXPORT_SYMBOL(iov_iter_npages); | |||
737 | const void *dup_iter(struct iov_iter *new, struct iov_iter *old, gfp_t flags) | 1124 | const void *dup_iter(struct iov_iter *new, struct iov_iter *old, gfp_t flags) |
738 | { | 1125 | { |
739 | *new = *old; | 1126 | *new = *old; |
1127 | if (unlikely(new->type & ITER_PIPE)) { | ||
1128 | WARN_ON(1); | ||
1129 | return NULL; | ||
1130 | } | ||
740 | if (new->type & ITER_BVEC) | 1131 | if (new->type & ITER_BVEC) |
741 | return new->bvec = kmemdup(new->bvec, | 1132 | return new->bvec = kmemdup(new->bvec, |
742 | new->nr_segs * sizeof(struct bio_vec), | 1133 | new->nr_segs * sizeof(struct bio_vec), |