diff options
author | Al Viro <viro@zeniv.linux.org.uk> | 2014-04-04 23:12:29 -0400 |
---|---|---|
committer | Al Viro <viro@zeniv.linux.org.uk> | 2014-05-06 17:39:45 -0400 |
commit | 62a8067a7f35dba2de501c9cb00e4cf36da90bc0 (patch) | |
tree | bb008456891c13b9d8a25825a10074efb861cd88 | |
parent | 81055e584f9d743cb13dc7944923d817c20f089d (diff) |
bio_vec-backed iov_iter
New variant of iov_iter - ITER_BVEC in iter->type, backed with
bio_vec array instead of iovec one. Primitives taught to deal
with such beasts, __swap_write() switched to using that kind
of iov_iter.
Note that bio_vec is just a <page, offset, length> triple - there's
nothing block-specific about it. I've left the definition where it
was, but took it from under ifdef CONFIG_BLOCK.
Next target: ->splice_write()...
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
-rw-r--r-- | fs/fuse/file.c | 2 | ||||
-rw-r--r-- | include/linux/blk_types.h | 4 | ||||
-rw-r--r-- | include/linux/uio.h | 14 | ||||
-rw-r--r-- | mm/iov_iter.c | 390 | ||||
-rw-r--r-- | mm/page_io.c | 19 |
5 files changed, 385 insertions, 44 deletions
diff --git a/fs/fuse/file.c b/fs/fuse/file.c index 7fbc803cf51d..b2dae9d1437c 100644 --- a/fs/fuse/file.c +++ b/fs/fuse/file.c | |||
@@ -1288,7 +1288,7 @@ static int fuse_get_user_pages(struct fuse_req *req, struct iov_iter *ii, | |||
1288 | size_t nbytes = 0; /* # bytes already packed in req */ | 1288 | size_t nbytes = 0; /* # bytes already packed in req */ |
1289 | 1289 | ||
1290 | /* Special case for kernel I/O: can copy directly into the buffer */ | 1290 | /* Special case for kernel I/O: can copy directly into the buffer */ |
1291 | if (ii->type & REQ_KERNEL) { | 1291 | if (ii->type & ITER_KVEC) { |
1292 | unsigned long user_addr = fuse_get_user_addr(ii); | 1292 | unsigned long user_addr = fuse_get_user_addr(ii); |
1293 | size_t frag_size = fuse_get_frag_size(ii, *nbytesp); | 1293 | size_t frag_size = fuse_get_frag_size(ii, *nbytesp); |
1294 | 1294 | ||
diff --git a/include/linux/blk_types.h b/include/linux/blk_types.h index aa0eaa2d0bd8..86df13b97160 100644 --- a/include/linux/blk_types.h +++ b/include/linux/blk_types.h | |||
@@ -5,8 +5,6 @@ | |||
5 | #ifndef __LINUX_BLK_TYPES_H | 5 | #ifndef __LINUX_BLK_TYPES_H |
6 | #define __LINUX_BLK_TYPES_H | 6 | #define __LINUX_BLK_TYPES_H |
7 | 7 | ||
8 | #ifdef CONFIG_BLOCK | ||
9 | |||
10 | #include <linux/types.h> | 8 | #include <linux/types.h> |
11 | 9 | ||
12 | struct bio_set; | 10 | struct bio_set; |
@@ -28,6 +26,8 @@ struct bio_vec { | |||
28 | unsigned int bv_offset; | 26 | unsigned int bv_offset; |
29 | }; | 27 | }; |
30 | 28 | ||
29 | #ifdef CONFIG_BLOCK | ||
30 | |||
31 | struct bvec_iter { | 31 | struct bvec_iter { |
32 | sector_t bi_sector; /* device address in 512 byte | 32 | sector_t bi_sector; /* device address in 512 byte |
33 | sectors */ | 33 | sectors */ |
diff --git a/include/linux/uio.h b/include/linux/uio.h index e8a109a75de1..e2231e47cec1 100644 --- a/include/linux/uio.h +++ b/include/linux/uio.h | |||
@@ -19,12 +19,21 @@ struct kvec { | |||
19 | size_t iov_len; | 19 | size_t iov_len; |
20 | }; | 20 | }; |
21 | 21 | ||
22 | enum { | ||
23 | ITER_IOVEC = 0, | ||
24 | ITER_KVEC = 2, | ||
25 | ITER_BVEC = 4, | ||
26 | }; | ||
27 | |||
22 | struct iov_iter { | 28 | struct iov_iter { |
23 | int type; | 29 | int type; |
24 | const struct iovec *iov; | ||
25 | unsigned long nr_segs; | ||
26 | size_t iov_offset; | 30 | size_t iov_offset; |
27 | size_t count; | 31 | size_t count; |
32 | union { | ||
33 | const struct iovec *iov; | ||
34 | const struct bio_vec *bvec; | ||
35 | }; | ||
36 | unsigned long nr_segs; | ||
28 | }; | 37 | }; |
29 | 38 | ||
30 | /* | 39 | /* |
@@ -54,6 +63,7 @@ static inline struct iovec iov_iter_iovec(const struct iov_iter *iter) | |||
54 | } | 63 | } |
55 | 64 | ||
56 | #define iov_for_each(iov, iter, start) \ | 65 | #define iov_for_each(iov, iter, start) \ |
66 | if (!((start).type & ITER_BVEC)) \ | ||
57 | for (iter = (start); \ | 67 | for (iter = (start); \ |
58 | (iter).count && \ | 68 | (iter).count && \ |
59 | ((iov = iov_iter_iovec(&(iter))), 1); \ | 69 | ((iov = iov_iter_iovec(&(iter))), 1); \ |
diff --git a/mm/iov_iter.c b/mm/iov_iter.c index fcdaaab438b6..7b5dbd1517b5 100644 --- a/mm/iov_iter.c +++ b/mm/iov_iter.c | |||
@@ -4,7 +4,7 @@ | |||
4 | #include <linux/slab.h> | 4 | #include <linux/slab.h> |
5 | #include <linux/vmalloc.h> | 5 | #include <linux/vmalloc.h> |
6 | 6 | ||
7 | size_t copy_page_to_iter(struct page *page, size_t offset, size_t bytes, | 7 | static size_t copy_page_to_iter_iovec(struct page *page, size_t offset, size_t bytes, |
8 | struct iov_iter *i) | 8 | struct iov_iter *i) |
9 | { | 9 | { |
10 | size_t skip, copy, left, wanted; | 10 | size_t skip, copy, left, wanted; |
@@ -84,9 +84,8 @@ done: | |||
84 | i->iov_offset = skip; | 84 | i->iov_offset = skip; |
85 | return wanted - bytes; | 85 | return wanted - bytes; |
86 | } | 86 | } |
87 | EXPORT_SYMBOL(copy_page_to_iter); | ||
88 | 87 | ||
89 | size_t copy_page_from_iter(struct page *page, size_t offset, size_t bytes, | 88 | static size_t copy_page_from_iter_iovec(struct page *page, size_t offset, size_t bytes, |
90 | struct iov_iter *i) | 89 | struct iov_iter *i) |
91 | { | 90 | { |
92 | size_t skip, copy, left, wanted; | 91 | size_t skip, copy, left, wanted; |
@@ -166,7 +165,6 @@ done: | |||
166 | i->iov_offset = skip; | 165 | i->iov_offset = skip; |
167 | return wanted - bytes; | 166 | return wanted - bytes; |
168 | } | 167 | } |
169 | EXPORT_SYMBOL(copy_page_from_iter); | ||
170 | 168 | ||
171 | static size_t __iovec_copy_from_user_inatomic(char *vaddr, | 169 | static size_t __iovec_copy_from_user_inatomic(char *vaddr, |
172 | const struct iovec *iov, size_t base, size_t bytes) | 170 | const struct iovec *iov, size_t base, size_t bytes) |
@@ -195,7 +193,7 @@ static size_t __iovec_copy_from_user_inatomic(char *vaddr, | |||
195 | * were successfully copied. If a fault is encountered then return the number of | 193 | * were successfully copied. If a fault is encountered then return the number of |
196 | * bytes which were copied. | 194 | * bytes which were copied. |
197 | */ | 195 | */ |
198 | size_t iov_iter_copy_from_user_atomic(struct page *page, | 196 | static size_t copy_from_user_atomic_iovec(struct page *page, |
199 | struct iov_iter *i, unsigned long offset, size_t bytes) | 197 | struct iov_iter *i, unsigned long offset, size_t bytes) |
200 | { | 198 | { |
201 | char *kaddr; | 199 | char *kaddr; |
@@ -215,9 +213,8 @@ size_t iov_iter_copy_from_user_atomic(struct page *page, | |||
215 | 213 | ||
216 | return copied; | 214 | return copied; |
217 | } | 215 | } |
218 | EXPORT_SYMBOL(iov_iter_copy_from_user_atomic); | ||
219 | 216 | ||
220 | void iov_iter_advance(struct iov_iter *i, size_t bytes) | 217 | static void advance_iovec(struct iov_iter *i, size_t bytes) |
221 | { | 218 | { |
222 | BUG_ON(i->count < bytes); | 219 | BUG_ON(i->count < bytes); |
223 | 220 | ||
@@ -252,7 +249,6 @@ void iov_iter_advance(struct iov_iter *i, size_t bytes) | |||
252 | i->nr_segs = nr_segs; | 249 | i->nr_segs = nr_segs; |
253 | } | 250 | } |
254 | } | 251 | } |
255 | EXPORT_SYMBOL(iov_iter_advance); | ||
256 | 252 | ||
257 | /* | 253 | /* |
258 | * Fault in the first iovec of the given iov_iter, to a maximum length | 254 | * Fault in the first iovec of the given iov_iter, to a maximum length |
@@ -265,26 +261,16 @@ EXPORT_SYMBOL(iov_iter_advance); | |||
265 | */ | 261 | */ |
266 | int iov_iter_fault_in_readable(struct iov_iter *i, size_t bytes) | 262 | int iov_iter_fault_in_readable(struct iov_iter *i, size_t bytes) |
267 | { | 263 | { |
268 | char __user *buf = i->iov->iov_base + i->iov_offset; | 264 | if (!(i->type & ITER_BVEC)) { |
269 | bytes = min(bytes, i->iov->iov_len - i->iov_offset); | 265 | char __user *buf = i->iov->iov_base + i->iov_offset; |
270 | return fault_in_pages_readable(buf, bytes); | 266 | bytes = min(bytes, i->iov->iov_len - i->iov_offset); |
267 | return fault_in_pages_readable(buf, bytes); | ||
268 | } | ||
269 | return 0; | ||
271 | } | 270 | } |
272 | EXPORT_SYMBOL(iov_iter_fault_in_readable); | 271 | EXPORT_SYMBOL(iov_iter_fault_in_readable); |
273 | 272 | ||
274 | /* | 273 | static unsigned long alignment_iovec(const struct iov_iter *i) |
275 | * Return the count of just the current iov_iter segment. | ||
276 | */ | ||
277 | size_t iov_iter_single_seg_count(const struct iov_iter *i) | ||
278 | { | ||
279 | const struct iovec *iov = i->iov; | ||
280 | if (i->nr_segs == 1) | ||
281 | return i->count; | ||
282 | else | ||
283 | return min(i->count, iov->iov_len - i->iov_offset); | ||
284 | } | ||
285 | EXPORT_SYMBOL(iov_iter_single_seg_count); | ||
286 | |||
287 | unsigned long iov_iter_alignment(const struct iov_iter *i) | ||
288 | { | 274 | { |
289 | const struct iovec *iov = i->iov; | 275 | const struct iovec *iov = i->iov; |
290 | unsigned long res; | 276 | unsigned long res; |
@@ -307,7 +293,6 @@ unsigned long iov_iter_alignment(const struct iov_iter *i) | |||
307 | res |= (unsigned long)iov->iov_base | size; | 293 | res |= (unsigned long)iov->iov_base | size; |
308 | return res; | 294 | return res; |
309 | } | 295 | } |
310 | EXPORT_SYMBOL(iov_iter_alignment); | ||
311 | 296 | ||
312 | void iov_iter_init(struct iov_iter *i, int direction, | 297 | void iov_iter_init(struct iov_iter *i, int direction, |
313 | const struct iovec *iov, unsigned long nr_segs, | 298 | const struct iovec *iov, unsigned long nr_segs, |
@@ -315,7 +300,7 @@ void iov_iter_init(struct iov_iter *i, int direction, | |||
315 | { | 300 | { |
316 | /* It will get better. Eventually... */ | 301 | /* It will get better. Eventually... */ |
317 | if (segment_eq(get_fs(), KERNEL_DS)) | 302 | if (segment_eq(get_fs(), KERNEL_DS)) |
318 | direction |= REQ_KERNEL; | 303 | direction |= ITER_KVEC; |
319 | i->type = direction; | 304 | i->type = direction; |
320 | i->iov = iov; | 305 | i->iov = iov; |
321 | i->nr_segs = nr_segs; | 306 | i->nr_segs = nr_segs; |
@@ -324,7 +309,7 @@ void iov_iter_init(struct iov_iter *i, int direction, | |||
324 | } | 309 | } |
325 | EXPORT_SYMBOL(iov_iter_init); | 310 | EXPORT_SYMBOL(iov_iter_init); |
326 | 311 | ||
327 | ssize_t iov_iter_get_pages(struct iov_iter *i, | 312 | static ssize_t get_pages_iovec(struct iov_iter *i, |
328 | struct page **pages, size_t maxsize, | 313 | struct page **pages, size_t maxsize, |
329 | size_t *start) | 314 | size_t *start) |
330 | { | 315 | { |
@@ -349,9 +334,8 @@ ssize_t iov_iter_get_pages(struct iov_iter *i, | |||
349 | return res; | 334 | return res; |
350 | return (res == n ? len : res * PAGE_SIZE) - *start; | 335 | return (res == n ? len : res * PAGE_SIZE) - *start; |
351 | } | 336 | } |
352 | EXPORT_SYMBOL(iov_iter_get_pages); | ||
353 | 337 | ||
354 | ssize_t iov_iter_get_pages_alloc(struct iov_iter *i, | 338 | static ssize_t get_pages_alloc_iovec(struct iov_iter *i, |
355 | struct page ***pages, size_t maxsize, | 339 | struct page ***pages, size_t maxsize, |
356 | size_t *start) | 340 | size_t *start) |
357 | { | 341 | { |
@@ -387,9 +371,8 @@ ssize_t iov_iter_get_pages_alloc(struct iov_iter *i, | |||
387 | *pages = p; | 371 | *pages = p; |
388 | return (res == n ? len : res * PAGE_SIZE) - *start; | 372 | return (res == n ? len : res * PAGE_SIZE) - *start; |
389 | } | 373 | } |
390 | EXPORT_SYMBOL(iov_iter_get_pages_alloc); | ||
391 | 374 | ||
392 | int iov_iter_npages(const struct iov_iter *i, int maxpages) | 375 | static int iov_iter_npages_iovec(const struct iov_iter *i, int maxpages) |
393 | { | 376 | { |
394 | size_t offset = i->iov_offset; | 377 | size_t offset = i->iov_offset; |
395 | size_t size = i->count; | 378 | size_t size = i->count; |
@@ -414,4 +397,347 @@ int iov_iter_npages(const struct iov_iter *i, int maxpages) | |||
414 | } | 397 | } |
415 | return min(npages, maxpages); | 398 | return min(npages, maxpages); |
416 | } | 399 | } |
400 | |||
401 | static void memcpy_from_page(char *to, struct page *page, size_t offset, size_t len) | ||
402 | { | ||
403 | char *from = kmap_atomic(page); | ||
404 | memcpy(to, from + offset, len); | ||
405 | kunmap_atomic(from); | ||
406 | } | ||
407 | |||
408 | static void memcpy_to_page(struct page *page, size_t offset, char *from, size_t len) | ||
409 | { | ||
410 | char *to = kmap_atomic(page); | ||
411 | memcpy(to + offset, from, len); | ||
412 | kunmap_atomic(to); | ||
413 | } | ||
414 | |||
415 | static size_t copy_page_to_iter_bvec(struct page *page, size_t offset, size_t bytes, | ||
416 | struct iov_iter *i) | ||
417 | { | ||
418 | size_t skip, copy, wanted; | ||
419 | const struct bio_vec *bvec; | ||
420 | void *kaddr, *from; | ||
421 | |||
422 | if (unlikely(bytes > i->count)) | ||
423 | bytes = i->count; | ||
424 | |||
425 | if (unlikely(!bytes)) | ||
426 | return 0; | ||
427 | |||
428 | wanted = bytes; | ||
429 | bvec = i->bvec; | ||
430 | skip = i->iov_offset; | ||
431 | copy = min_t(size_t, bytes, bvec->bv_len - skip); | ||
432 | |||
433 | kaddr = kmap_atomic(page); | ||
434 | from = kaddr + offset; | ||
435 | memcpy_to_page(bvec->bv_page, skip + bvec->bv_offset, from, copy); | ||
436 | skip += copy; | ||
437 | from += copy; | ||
438 | bytes -= copy; | ||
439 | while (bytes) { | ||
440 | bvec++; | ||
441 | copy = min(bytes, (size_t)bvec->bv_len); | ||
442 | memcpy_to_page(bvec->bv_page, bvec->bv_offset, from, copy); | ||
443 | skip = copy; | ||
444 | from += copy; | ||
445 | bytes -= copy; | ||
446 | } | ||
447 | kunmap_atomic(kaddr); | ||
448 | if (skip == bvec->bv_len) { | ||
449 | bvec++; | ||
450 | skip = 0; | ||
451 | } | ||
452 | i->count -= wanted - bytes; | ||
453 | i->nr_segs -= bvec - i->bvec; | ||
454 | i->bvec = bvec; | ||
455 | i->iov_offset = skip; | ||
456 | return wanted - bytes; | ||
457 | } | ||
458 | |||
459 | static size_t copy_page_from_iter_bvec(struct page *page, size_t offset, size_t bytes, | ||
460 | struct iov_iter *i) | ||
461 | { | ||
462 | size_t skip, copy, wanted; | ||
463 | const struct bio_vec *bvec; | ||
464 | void *kaddr, *to; | ||
465 | |||
466 | if (unlikely(bytes > i->count)) | ||
467 | bytes = i->count; | ||
468 | |||
469 | if (unlikely(!bytes)) | ||
470 | return 0; | ||
471 | |||
472 | wanted = bytes; | ||
473 | bvec = i->bvec; | ||
474 | skip = i->iov_offset; | ||
475 | |||
476 | kaddr = kmap_atomic(page); | ||
477 | |||
478 | to = kaddr + offset; | ||
479 | |||
480 | copy = min(bytes, bvec->bv_len - skip); | ||
481 | |||
482 | memcpy_from_page(to, bvec->bv_page, bvec->bv_offset + skip, copy); | ||
483 | |||
484 | to += copy; | ||
485 | skip += copy; | ||
486 | bytes -= copy; | ||
487 | |||
488 | while (bytes) { | ||
489 | bvec++; | ||
490 | copy = min(bytes, (size_t)bvec->bv_len); | ||
491 | memcpy_from_page(to, bvec->bv_page, bvec->bv_offset, copy); | ||
492 | skip = copy; | ||
493 | to += copy; | ||
494 | bytes -= copy; | ||
495 | } | ||
496 | kunmap_atomic(kaddr); | ||
497 | if (skip == bvec->bv_len) { | ||
498 | bvec++; | ||
499 | skip = 0; | ||
500 | } | ||
501 | i->count -= wanted; | ||
502 | i->nr_segs -= bvec - i->bvec; | ||
503 | i->bvec = bvec; | ||
504 | i->iov_offset = skip; | ||
505 | return wanted; | ||
506 | } | ||
507 | |||
508 | static size_t copy_from_user_bvec(struct page *page, | ||
509 | struct iov_iter *i, unsigned long offset, size_t bytes) | ||
510 | { | ||
511 | char *kaddr; | ||
512 | size_t left; | ||
513 | const struct bio_vec *bvec; | ||
514 | size_t base = i->iov_offset; | ||
515 | |||
516 | kaddr = kmap_atomic(page); | ||
517 | for (left = bytes, bvec = i->bvec; left; bvec++, base = 0) { | ||
518 | size_t copy = min(left, bvec->bv_len - base); | ||
519 | if (!bvec->bv_len) | ||
520 | continue; | ||
521 | memcpy_from_page(kaddr + offset, bvec->bv_page, | ||
522 | bvec->bv_offset + base, copy); | ||
523 | offset += copy; | ||
524 | left -= copy; | ||
525 | } | ||
526 | kunmap_atomic(kaddr); | ||
527 | return bytes; | ||
528 | } | ||
529 | |||
530 | static void advance_bvec(struct iov_iter *i, size_t bytes) | ||
531 | { | ||
532 | BUG_ON(i->count < bytes); | ||
533 | |||
534 | if (likely(i->nr_segs == 1)) { | ||
535 | i->iov_offset += bytes; | ||
536 | i->count -= bytes; | ||
537 | } else { | ||
538 | const struct bio_vec *bvec = i->bvec; | ||
539 | size_t base = i->iov_offset; | ||
540 | unsigned long nr_segs = i->nr_segs; | ||
541 | |||
542 | /* | ||
543 | * The !iov->iov_len check ensures we skip over unlikely | ||
544 | * zero-length segments (without overruning the iovec). | ||
545 | */ | ||
546 | while (bytes || unlikely(i->count && !bvec->bv_len)) { | ||
547 | int copy; | ||
548 | |||
549 | copy = min(bytes, bvec->bv_len - base); | ||
550 | BUG_ON(!i->count || i->count < copy); | ||
551 | i->count -= copy; | ||
552 | bytes -= copy; | ||
553 | base += copy; | ||
554 | if (bvec->bv_len == base) { | ||
555 | bvec++; | ||
556 | nr_segs--; | ||
557 | base = 0; | ||
558 | } | ||
559 | } | ||
560 | i->bvec = bvec; | ||
561 | i->iov_offset = base; | ||
562 | i->nr_segs = nr_segs; | ||
563 | } | ||
564 | } | ||
565 | |||
566 | static unsigned long alignment_bvec(const struct iov_iter *i) | ||
567 | { | ||
568 | const struct bio_vec *bvec = i->bvec; | ||
569 | unsigned long res; | ||
570 | size_t size = i->count; | ||
571 | size_t n; | ||
572 | |||
573 | if (!size) | ||
574 | return 0; | ||
575 | |||
576 | res = bvec->bv_offset + i->iov_offset; | ||
577 | n = bvec->bv_len - i->iov_offset; | ||
578 | if (n >= size) | ||
579 | return res | size; | ||
580 | size -= n; | ||
581 | res |= n; | ||
582 | while (size > (++bvec)->bv_len) { | ||
583 | res |= bvec->bv_offset | bvec->bv_len; | ||
584 | size -= bvec->bv_len; | ||
585 | } | ||
586 | res |= bvec->bv_offset | size; | ||
587 | return res; | ||
588 | } | ||
589 | |||
590 | static ssize_t get_pages_bvec(struct iov_iter *i, | ||
591 | struct page **pages, size_t maxsize, | ||
592 | size_t *start) | ||
593 | { | ||
594 | const struct bio_vec *bvec = i->bvec; | ||
595 | size_t len = bvec->bv_len - i->iov_offset; | ||
596 | if (len > i->count) | ||
597 | len = i->count; | ||
598 | if (len > maxsize) | ||
599 | len = maxsize; | ||
600 | *start = bvec->bv_offset + i->iov_offset; | ||
601 | |||
602 | get_page(*pages = bvec->bv_page); | ||
603 | |||
604 | return len; | ||
605 | } | ||
606 | |||
607 | static ssize_t get_pages_alloc_bvec(struct iov_iter *i, | ||
608 | struct page ***pages, size_t maxsize, | ||
609 | size_t *start) | ||
610 | { | ||
611 | const struct bio_vec *bvec = i->bvec; | ||
612 | size_t len = bvec->bv_len - i->iov_offset; | ||
613 | if (len > i->count) | ||
614 | len = i->count; | ||
615 | if (len > maxsize) | ||
616 | len = maxsize; | ||
617 | *start = bvec->bv_offset + i->iov_offset; | ||
618 | |||
619 | *pages = kmalloc(sizeof(struct page *), GFP_KERNEL); | ||
620 | if (!*pages) | ||
621 | return -ENOMEM; | ||
622 | |||
623 | get_page(**pages = bvec->bv_page); | ||
624 | |||
625 | return len; | ||
626 | } | ||
627 | |||
628 | static int iov_iter_npages_bvec(const struct iov_iter *i, int maxpages) | ||
629 | { | ||
630 | size_t offset = i->iov_offset; | ||
631 | size_t size = i->count; | ||
632 | const struct bio_vec *bvec = i->bvec; | ||
633 | int npages = 0; | ||
634 | int n; | ||
635 | |||
636 | for (n = 0; size && n < i->nr_segs; n++, bvec++) { | ||
637 | size_t len = bvec->bv_len - offset; | ||
638 | offset = 0; | ||
639 | if (unlikely(!len)) /* empty segment */ | ||
640 | continue; | ||
641 | if (len > size) | ||
642 | len = size; | ||
643 | npages++; | ||
644 | if (npages >= maxpages) /* don't bother going further */ | ||
645 | return maxpages; | ||
646 | size -= len; | ||
647 | offset = 0; | ||
648 | } | ||
649 | return min(npages, maxpages); | ||
650 | } | ||
651 | |||
652 | size_t copy_page_to_iter(struct page *page, size_t offset, size_t bytes, | ||
653 | struct iov_iter *i) | ||
654 | { | ||
655 | if (i->type & ITER_BVEC) | ||
656 | return copy_page_to_iter_bvec(page, offset, bytes, i); | ||
657 | else | ||
658 | return copy_page_to_iter_iovec(page, offset, bytes, i); | ||
659 | } | ||
660 | EXPORT_SYMBOL(copy_page_to_iter); | ||
661 | |||
662 | size_t copy_page_from_iter(struct page *page, size_t offset, size_t bytes, | ||
663 | struct iov_iter *i) | ||
664 | { | ||
665 | if (i->type & ITER_BVEC) | ||
666 | return copy_page_from_iter_bvec(page, offset, bytes, i); | ||
667 | else | ||
668 | return copy_page_from_iter_iovec(page, offset, bytes, i); | ||
669 | } | ||
670 | EXPORT_SYMBOL(copy_page_from_iter); | ||
671 | |||
672 | size_t iov_iter_copy_from_user_atomic(struct page *page, | ||
673 | struct iov_iter *i, unsigned long offset, size_t bytes) | ||
674 | { | ||
675 | if (i->type & ITER_BVEC) | ||
676 | return copy_from_user_bvec(page, i, offset, bytes); | ||
677 | else | ||
678 | return copy_from_user_atomic_iovec(page, i, offset, bytes); | ||
679 | } | ||
680 | EXPORT_SYMBOL(iov_iter_copy_from_user_atomic); | ||
681 | |||
682 | void iov_iter_advance(struct iov_iter *i, size_t size) | ||
683 | { | ||
684 | if (i->type & ITER_BVEC) | ||
685 | advance_bvec(i, size); | ||
686 | else | ||
687 | advance_iovec(i, size); | ||
688 | } | ||
689 | EXPORT_SYMBOL(iov_iter_advance); | ||
690 | |||
691 | /* | ||
692 | * Return the count of just the current iov_iter segment. | ||
693 | */ | ||
694 | size_t iov_iter_single_seg_count(const struct iov_iter *i) | ||
695 | { | ||
696 | if (i->nr_segs == 1) | ||
697 | return i->count; | ||
698 | else if (i->type & ITER_BVEC) | ||
699 | return min(i->count, i->iov->iov_len - i->iov_offset); | ||
700 | else | ||
701 | return min(i->count, i->bvec->bv_len - i->iov_offset); | ||
702 | } | ||
703 | EXPORT_SYMBOL(iov_iter_single_seg_count); | ||
704 | |||
705 | unsigned long iov_iter_alignment(const struct iov_iter *i) | ||
706 | { | ||
707 | if (i->type & ITER_BVEC) | ||
708 | return alignment_bvec(i); | ||
709 | else | ||
710 | return alignment_iovec(i); | ||
711 | } | ||
712 | EXPORT_SYMBOL(iov_iter_alignment); | ||
713 | |||
714 | ssize_t iov_iter_get_pages(struct iov_iter *i, | ||
715 | struct page **pages, size_t maxsize, | ||
716 | size_t *start) | ||
717 | { | ||
718 | if (i->type & ITER_BVEC) | ||
719 | return get_pages_bvec(i, pages, maxsize, start); | ||
720 | else | ||
721 | return get_pages_iovec(i, pages, maxsize, start); | ||
722 | } | ||
723 | EXPORT_SYMBOL(iov_iter_get_pages); | ||
724 | |||
725 | ssize_t iov_iter_get_pages_alloc(struct iov_iter *i, | ||
726 | struct page ***pages, size_t maxsize, | ||
727 | size_t *start) | ||
728 | { | ||
729 | if (i->type & ITER_BVEC) | ||
730 | return get_pages_alloc_bvec(i, pages, maxsize, start); | ||
731 | else | ||
732 | return get_pages_alloc_iovec(i, pages, maxsize, start); | ||
733 | } | ||
734 | EXPORT_SYMBOL(iov_iter_get_pages_alloc); | ||
735 | |||
736 | int iov_iter_npages(const struct iov_iter *i, int maxpages) | ||
737 | { | ||
738 | if (i->type & ITER_BVEC) | ||
739 | return iov_iter_npages_bvec(i, maxpages); | ||
740 | else | ||
741 | return iov_iter_npages_iovec(i, maxpages); | ||
742 | } | ||
417 | EXPORT_SYMBOL(iov_iter_npages); | 743 | EXPORT_SYMBOL(iov_iter_npages); |
diff --git a/mm/page_io.c b/mm/page_io.c index 313bfedb75d1..33bb38c4aad7 100644 --- a/mm/page_io.c +++ b/mm/page_io.c | |||
@@ -259,23 +259,28 @@ int __swap_writepage(struct page *page, struct writeback_control *wbc, | |||
259 | struct kiocb kiocb; | 259 | struct kiocb kiocb; |
260 | struct file *swap_file = sis->swap_file; | 260 | struct file *swap_file = sis->swap_file; |
261 | struct address_space *mapping = swap_file->f_mapping; | 261 | struct address_space *mapping = swap_file->f_mapping; |
262 | struct iovec iov = { | 262 | struct bio_vec bv = { |
263 | .iov_base = kmap(page), | 263 | .bv_page = page, |
264 | .iov_len = PAGE_SIZE, | 264 | .bv_len = PAGE_SIZE, |
265 | .bv_offset = 0 | ||
266 | }; | ||
267 | struct iov_iter from = { | ||
268 | .type = ITER_BVEC | WRITE, | ||
269 | .count = PAGE_SIZE, | ||
270 | .iov_offset = 0, | ||
271 | .nr_segs = 1, | ||
272 | .bvec = &bv | ||
265 | }; | 273 | }; |
266 | struct iov_iter from; | ||
267 | 274 | ||
268 | init_sync_kiocb(&kiocb, swap_file); | 275 | init_sync_kiocb(&kiocb, swap_file); |
269 | kiocb.ki_pos = page_file_offset(page); | 276 | kiocb.ki_pos = page_file_offset(page); |
270 | kiocb.ki_nbytes = PAGE_SIZE; | 277 | kiocb.ki_nbytes = PAGE_SIZE; |
271 | iov_iter_init(&from, KERNEL_WRITE, &iov, 1, PAGE_SIZE); | ||
272 | 278 | ||
273 | set_page_writeback(page); | 279 | set_page_writeback(page); |
274 | unlock_page(page); | 280 | unlock_page(page); |
275 | ret = mapping->a_ops->direct_IO(KERNEL_WRITE, | 281 | ret = mapping->a_ops->direct_IO(ITER_BVEC | WRITE, |
276 | &kiocb, &from, | 282 | &kiocb, &from, |
277 | kiocb.ki_pos); | 283 | kiocb.ki_pos); |
278 | kunmap(page); | ||
279 | if (ret == PAGE_SIZE) { | 284 | if (ret == PAGE_SIZE) { |
280 | count_vm_event(PSWPOUT); | 285 | count_vm_event(PSWPOUT); |
281 | ret = 0; | 286 | ret = 0; |