diff options
| author | Trond Myklebust <Trond.Myklebust@netapp.com> | 2010-03-05 15:46:18 -0500 |
|---|---|---|
| committer | Trond Myklebust <Trond.Myklebust@netapp.com> | 2010-03-05 15:46:18 -0500 |
| commit | 3fa04ecd72780da31ba8b329e148179bc24a9c7d (patch) | |
| tree | f5d462fd4aee086952d18f159f737c450ab46b3b /fs/nfs | |
| parent | 180b62a3d837613fcac3ce89576526423926c3c3 (diff) | |
| parent | 1cda707d52e51a6cafac0aef12d2bd7052d572e6 (diff) | |
Merge branch 'writeback-for-2.6.34' into nfs-for-2.6.34
Diffstat (limited to 'fs/nfs')
| -rw-r--r-- | fs/nfs/Kconfig | 3 | ||||
| -rw-r--r-- | fs/nfs/dir.c | 2 | ||||
| -rw-r--r-- | fs/nfs/direct.c | 3 | ||||
| -rw-r--r-- | fs/nfs/fscache.c | 9 | ||||
| -rw-r--r-- | fs/nfs/inode.c | 95 | ||||
| -rw-r--r-- | fs/nfs/internal.h | 2 | ||||
| -rw-r--r-- | fs/nfs/iostat.h | 4 | ||||
| -rw-r--r-- | fs/nfs/mount_clnt.c | 2 | ||||
| -rw-r--r-- | fs/nfs/nfs2xdr.c | 2 | ||||
| -rw-r--r-- | fs/nfs/nfs4proc.c | 8 | ||||
| -rw-r--r-- | fs/nfs/nfs4xdr.c | 6 | ||||
| -rw-r--r-- | fs/nfs/symlink.c | 2 | ||||
| -rw-r--r-- | fs/nfs/write.c | 250 |
13 files changed, 115 insertions, 273 deletions
diff --git a/fs/nfs/Kconfig b/fs/nfs/Kconfig index 59e5673b4597..a43d07e7b924 100644 --- a/fs/nfs/Kconfig +++ b/fs/nfs/Kconfig | |||
| @@ -95,8 +95,7 @@ config ROOT_NFS | |||
| 95 | Most people say N here. | 95 | Most people say N here. |
| 96 | 96 | ||
| 97 | config NFS_FSCACHE | 97 | config NFS_FSCACHE |
| 98 | bool "Provide NFS client caching support (EXPERIMENTAL)" | 98 | bool "Provide NFS client caching support" |
| 99 | depends on EXPERIMENTAL | ||
| 100 | depends on NFS_FS=m && FSCACHE || NFS_FS=y && FSCACHE=y | 99 | depends on NFS_FS=m && FSCACHE || NFS_FS=y && FSCACHE=y |
| 101 | help | 100 | help |
| 102 | Say Y here if you want NFS data to be cached locally on disc through | 101 | Say Y here if you want NFS data to be cached locally on disc through |
diff --git a/fs/nfs/dir.c b/fs/nfs/dir.c index 3c7f03b669fb..a1f6b4438fb1 100644 --- a/fs/nfs/dir.c +++ b/fs/nfs/dir.c | |||
| @@ -560,7 +560,7 @@ static int nfs_readdir(struct file *filp, void *dirent, filldir_t filldir) | |||
| 560 | desc->entry = &my_entry; | 560 | desc->entry = &my_entry; |
| 561 | 561 | ||
| 562 | nfs_block_sillyrename(dentry); | 562 | nfs_block_sillyrename(dentry); |
| 563 | res = nfs_revalidate_mapping_nolock(inode, filp->f_mapping); | 563 | res = nfs_revalidate_mapping(inode, filp->f_mapping); |
| 564 | if (res < 0) | 564 | if (res < 0) |
| 565 | goto out; | 565 | goto out; |
| 566 | 566 | ||
diff --git a/fs/nfs/direct.c b/fs/nfs/direct.c index e1d415e97849..0d289823e856 100644 --- a/fs/nfs/direct.c +++ b/fs/nfs/direct.c | |||
| @@ -342,6 +342,7 @@ static ssize_t nfs_direct_read_schedule_segment(struct nfs_direct_req *dreq, | |||
| 342 | data->res.fattr = &data->fattr; | 342 | data->res.fattr = &data->fattr; |
| 343 | data->res.eof = 0; | 343 | data->res.eof = 0; |
| 344 | data->res.count = bytes; | 344 | data->res.count = bytes; |
| 345 | nfs_fattr_init(&data->fattr); | ||
| 345 | msg.rpc_argp = &data->args; | 346 | msg.rpc_argp = &data->args; |
| 346 | msg.rpc_resp = &data->res; | 347 | msg.rpc_resp = &data->res; |
| 347 | 348 | ||
| @@ -575,6 +576,7 @@ static void nfs_direct_commit_schedule(struct nfs_direct_req *dreq) | |||
| 575 | data->res.count = 0; | 576 | data->res.count = 0; |
| 576 | data->res.fattr = &data->fattr; | 577 | data->res.fattr = &data->fattr; |
| 577 | data->res.verf = &data->verf; | 578 | data->res.verf = &data->verf; |
| 579 | nfs_fattr_init(&data->fattr); | ||
| 578 | 580 | ||
| 579 | NFS_PROTO(data->inode)->commit_setup(data, &msg); | 581 | NFS_PROTO(data->inode)->commit_setup(data, &msg); |
| 580 | 582 | ||
| @@ -766,6 +768,7 @@ static ssize_t nfs_direct_write_schedule_segment(struct nfs_direct_req *dreq, | |||
| 766 | data->res.fattr = &data->fattr; | 768 | data->res.fattr = &data->fattr; |
| 767 | data->res.count = bytes; | 769 | data->res.count = bytes; |
| 768 | data->res.verf = &data->verf; | 770 | data->res.verf = &data->verf; |
| 771 | nfs_fattr_init(&data->fattr); | ||
| 769 | 772 | ||
| 770 | task_setup_data.task = &data->task; | 773 | task_setup_data.task = &data->task; |
| 771 | task_setup_data.callback_data = data; | 774 | task_setup_data.callback_data = data; |
diff --git a/fs/nfs/fscache.c b/fs/nfs/fscache.c index fa588006588d..237874f1af23 100644 --- a/fs/nfs/fscache.c +++ b/fs/nfs/fscache.c | |||
| @@ -354,12 +354,11 @@ void nfs_fscache_reset_inode_cookie(struct inode *inode) | |||
| 354 | */ | 354 | */ |
| 355 | int nfs_fscache_release_page(struct page *page, gfp_t gfp) | 355 | int nfs_fscache_release_page(struct page *page, gfp_t gfp) |
| 356 | { | 356 | { |
| 357 | struct nfs_inode *nfsi = NFS_I(page->mapping->host); | ||
| 358 | struct fscache_cookie *cookie = nfsi->fscache; | ||
| 359 | |||
| 360 | BUG_ON(!cookie); | ||
| 361 | |||
| 362 | if (PageFsCache(page)) { | 357 | if (PageFsCache(page)) { |
| 358 | struct nfs_inode *nfsi = NFS_I(page->mapping->host); | ||
| 359 | struct fscache_cookie *cookie = nfsi->fscache; | ||
| 360 | |||
| 361 | BUG_ON(!cookie); | ||
| 363 | dfprintk(FSCACHE, "NFS: fscache releasepage (0x%p/0x%p/0x%p)\n", | 362 | dfprintk(FSCACHE, "NFS: fscache releasepage (0x%p/0x%p/0x%p)\n", |
| 364 | cookie, page, nfsi); | 363 | cookie, page, nfsi); |
| 365 | 364 | ||
diff --git a/fs/nfs/inode.c b/fs/nfs/inode.c index 87cca56846d6..657201acda84 100644 --- a/fs/nfs/inode.c +++ b/fs/nfs/inode.c | |||
| @@ -97,22 +97,6 @@ u64 nfs_compat_user_ino64(u64 fileid) | |||
| 97 | return ino; | 97 | return ino; |
| 98 | } | 98 | } |
| 99 | 99 | ||
| 100 | int nfs_write_inode(struct inode *inode, int sync) | ||
| 101 | { | ||
| 102 | int ret; | ||
| 103 | |||
| 104 | if (sync) { | ||
| 105 | ret = filemap_fdatawait(inode->i_mapping); | ||
| 106 | if (ret == 0) | ||
| 107 | ret = nfs_commit_inode(inode, FLUSH_SYNC); | ||
| 108 | } else | ||
| 109 | ret = nfs_commit_inode(inode, 0); | ||
| 110 | if (ret >= 0) | ||
| 111 | return 0; | ||
| 112 | __mark_inode_dirty(inode, I_DIRTY_DATASYNC); | ||
| 113 | return ret; | ||
| 114 | } | ||
| 115 | |||
| 116 | void nfs_clear_inode(struct inode *inode) | 100 | void nfs_clear_inode(struct inode *inode) |
| 117 | { | 101 | { |
| 118 | /* | 102 | /* |
| @@ -130,16 +114,12 @@ void nfs_clear_inode(struct inode *inode) | |||
| 130 | */ | 114 | */ |
| 131 | int nfs_sync_mapping(struct address_space *mapping) | 115 | int nfs_sync_mapping(struct address_space *mapping) |
| 132 | { | 116 | { |
| 133 | int ret; | 117 | int ret = 0; |
| 134 | 118 | ||
| 135 | if (mapping->nrpages == 0) | 119 | if (mapping->nrpages != 0) { |
| 136 | return 0; | 120 | unmap_mapping_range(mapping, 0, 0, 0); |
| 137 | unmap_mapping_range(mapping, 0, 0, 0); | 121 | ret = nfs_wb_all(mapping->host); |
| 138 | ret = filemap_write_and_wait(mapping); | 122 | } |
| 139 | if (ret != 0) | ||
| 140 | goto out; | ||
| 141 | ret = nfs_wb_all(mapping->host); | ||
| 142 | out: | ||
| 143 | return ret; | 123 | return ret; |
| 144 | } | 124 | } |
| 145 | 125 | ||
| @@ -511,17 +491,11 @@ int nfs_getattr(struct vfsmount *mnt, struct dentry *dentry, struct kstat *stat) | |||
| 511 | int need_atime = NFS_I(inode)->cache_validity & NFS_INO_INVALID_ATIME; | 491 | int need_atime = NFS_I(inode)->cache_validity & NFS_INO_INVALID_ATIME; |
| 512 | int err; | 492 | int err; |
| 513 | 493 | ||
| 514 | /* | 494 | /* Flush out writes to the server in order to update c/mtime. */ |
| 515 | * Flush out writes to the server in order to update c/mtime. | ||
| 516 | * | ||
| 517 | * Hold the i_mutex to suspend application writes temporarily; | ||
| 518 | * this prevents long-running writing applications from blocking | ||
| 519 | * nfs_wb_nocommit. | ||
| 520 | */ | ||
| 521 | if (S_ISREG(inode->i_mode)) { | 495 | if (S_ISREG(inode->i_mode)) { |
| 522 | mutex_lock(&inode->i_mutex); | 496 | err = filemap_write_and_wait(inode->i_mapping); |
| 523 | nfs_wb_nocommit(inode); | 497 | if (err) |
| 524 | mutex_unlock(&inode->i_mutex); | 498 | goto out; |
| 525 | } | 499 | } |
| 526 | 500 | ||
| 527 | /* | 501 | /* |
| @@ -545,6 +519,7 @@ int nfs_getattr(struct vfsmount *mnt, struct dentry *dentry, struct kstat *stat) | |||
| 545 | generic_fillattr(inode, stat); | 519 | generic_fillattr(inode, stat); |
| 546 | stat->ino = nfs_compat_user_ino64(NFS_FILEID(inode)); | 520 | stat->ino = nfs_compat_user_ino64(NFS_FILEID(inode)); |
| 547 | } | 521 | } |
| 522 | out: | ||
| 548 | return err; | 523 | return err; |
| 549 | } | 524 | } |
| 550 | 525 | ||
| @@ -574,14 +549,14 @@ void nfs_close_context(struct nfs_open_context *ctx, int is_sync) | |||
| 574 | nfs_revalidate_inode(server, inode); | 549 | nfs_revalidate_inode(server, inode); |
| 575 | } | 550 | } |
| 576 | 551 | ||
| 577 | static struct nfs_open_context *alloc_nfs_open_context(struct vfsmount *mnt, struct dentry *dentry, struct rpc_cred *cred) | 552 | static struct nfs_open_context *alloc_nfs_open_context(struct path *path, struct rpc_cred *cred) |
| 578 | { | 553 | { |
| 579 | struct nfs_open_context *ctx; | 554 | struct nfs_open_context *ctx; |
| 580 | 555 | ||
| 581 | ctx = kmalloc(sizeof(*ctx), GFP_KERNEL); | 556 | ctx = kmalloc(sizeof(*ctx), GFP_KERNEL); |
| 582 | if (ctx != NULL) { | 557 | if (ctx != NULL) { |
| 583 | ctx->path.dentry = dget(dentry); | 558 | ctx->path = *path; |
| 584 | ctx->path.mnt = mntget(mnt); | 559 | path_get(&ctx->path); |
| 585 | ctx->cred = get_rpccred(cred); | 560 | ctx->cred = get_rpccred(cred); |
| 586 | ctx->state = NULL; | 561 | ctx->state = NULL; |
| 587 | ctx->lockowner = current->files; | 562 | ctx->lockowner = current->files; |
| @@ -681,7 +656,7 @@ int nfs_open(struct inode *inode, struct file *filp) | |||
| 681 | cred = rpc_lookup_cred(); | 656 | cred = rpc_lookup_cred(); |
| 682 | if (IS_ERR(cred)) | 657 | if (IS_ERR(cred)) |
| 683 | return PTR_ERR(cred); | 658 | return PTR_ERR(cred); |
| 684 | ctx = alloc_nfs_open_context(filp->f_path.mnt, filp->f_path.dentry, cred); | 659 | ctx = alloc_nfs_open_context(&filp->f_path, cred); |
| 685 | put_rpccred(cred); | 660 | put_rpccred(cred); |
| 686 | if (ctx == NULL) | 661 | if (ctx == NULL) |
| 687 | return -ENOMEM; | 662 | return -ENOMEM; |
| @@ -774,7 +749,7 @@ int nfs_revalidate_inode(struct nfs_server *server, struct inode *inode) | |||
| 774 | return __nfs_revalidate_inode(server, inode); | 749 | return __nfs_revalidate_inode(server, inode); |
| 775 | } | 750 | } |
| 776 | 751 | ||
| 777 | static int nfs_invalidate_mapping_nolock(struct inode *inode, struct address_space *mapping) | 752 | static int nfs_invalidate_mapping(struct inode *inode, struct address_space *mapping) |
| 778 | { | 753 | { |
| 779 | struct nfs_inode *nfsi = NFS_I(inode); | 754 | struct nfs_inode *nfsi = NFS_I(inode); |
| 780 | 755 | ||
| @@ -795,49 +770,10 @@ static int nfs_invalidate_mapping_nolock(struct inode *inode, struct address_spa | |||
| 795 | return 0; | 770 | return 0; |
| 796 | } | 771 | } |
| 797 | 772 | ||
| 798 | static int nfs_invalidate_mapping(struct inode *inode, struct address_space *mapping) | ||
| 799 | { | ||
| 800 | int ret = 0; | ||
| 801 | |||
| 802 | mutex_lock(&inode->i_mutex); | ||
| 803 | if (NFS_I(inode)->cache_validity & NFS_INO_INVALID_DATA) { | ||
| 804 | ret = nfs_sync_mapping(mapping); | ||
| 805 | if (ret == 0) | ||
| 806 | ret = nfs_invalidate_mapping_nolock(inode, mapping); | ||
| 807 | } | ||
| 808 | mutex_unlock(&inode->i_mutex); | ||
| 809 | return ret; | ||
| 810 | } | ||
| 811 | |||
| 812 | /** | ||
| 813 | * nfs_revalidate_mapping_nolock - Revalidate the pagecache | ||
| 814 | * @inode - pointer to host inode | ||
| 815 | * @mapping - pointer to mapping | ||
| 816 | */ | ||
| 817 | int nfs_revalidate_mapping_nolock(struct inode *inode, struct address_space *mapping) | ||
| 818 | { | ||
| 819 | struct nfs_inode *nfsi = NFS_I(inode); | ||
| 820 | int ret = 0; | ||
| 821 | |||
| 822 | if ((nfsi->cache_validity & NFS_INO_REVAL_PAGECACHE) | ||
| 823 | || nfs_attribute_timeout(inode) || NFS_STALE(inode)) { | ||
| 824 | ret = __nfs_revalidate_inode(NFS_SERVER(inode), inode); | ||
| 825 | if (ret < 0) | ||
| 826 | goto out; | ||
| 827 | } | ||
| 828 | if (nfsi->cache_validity & NFS_INO_INVALID_DATA) | ||
| 829 | ret = nfs_invalidate_mapping_nolock(inode, mapping); | ||
| 830 | out: | ||
| 831 | return ret; | ||
| 832 | } | ||
| 833 | |||
| 834 | /** | 773 | /** |
| 835 | * nfs_revalidate_mapping - Revalidate the pagecache | 774 | * nfs_revalidate_mapping - Revalidate the pagecache |
| 836 | * @inode - pointer to host inode | 775 | * @inode - pointer to host inode |
| 837 | * @mapping - pointer to mapping | 776 | * @mapping - pointer to mapping |
| 838 | * | ||
| 839 | * This version of the function will take the inode->i_mutex and attempt to | ||
| 840 | * flush out all dirty data if it needs to invalidate the page cache. | ||
| 841 | */ | 777 | */ |
| 842 | int nfs_revalidate_mapping(struct inode *inode, struct address_space *mapping) | 778 | int nfs_revalidate_mapping(struct inode *inode, struct address_space *mapping) |
| 843 | { | 779 | { |
| @@ -1415,6 +1351,7 @@ static void init_once(void *foo) | |||
| 1415 | INIT_LIST_HEAD(&nfsi->access_cache_inode_lru); | 1351 | INIT_LIST_HEAD(&nfsi->access_cache_inode_lru); |
| 1416 | INIT_RADIX_TREE(&nfsi->nfs_page_tree, GFP_ATOMIC); | 1352 | INIT_RADIX_TREE(&nfsi->nfs_page_tree, GFP_ATOMIC); |
| 1417 | nfsi->npages = 0; | 1353 | nfsi->npages = 0; |
| 1354 | nfsi->ncommit = 0; | ||
| 1418 | atomic_set(&nfsi->silly_count, 1); | 1355 | atomic_set(&nfsi->silly_count, 1); |
| 1419 | INIT_HLIST_HEAD(&nfsi->silly_list); | 1356 | INIT_HLIST_HEAD(&nfsi->silly_list); |
| 1420 | init_waitqueue_head(&nfsi->waitqueue); | 1357 | init_waitqueue_head(&nfsi->waitqueue); |
diff --git a/fs/nfs/internal.h b/fs/nfs/internal.h index 29e464d23b32..11f82f03c5de 100644 --- a/fs/nfs/internal.h +++ b/fs/nfs/internal.h | |||
| @@ -211,7 +211,7 @@ extern int nfs_access_cache_shrinker(int nr_to_scan, gfp_t gfp_mask); | |||
| 211 | extern struct workqueue_struct *nfsiod_workqueue; | 211 | extern struct workqueue_struct *nfsiod_workqueue; |
| 212 | extern struct inode *nfs_alloc_inode(struct super_block *sb); | 212 | extern struct inode *nfs_alloc_inode(struct super_block *sb); |
| 213 | extern void nfs_destroy_inode(struct inode *); | 213 | extern void nfs_destroy_inode(struct inode *); |
| 214 | extern int nfs_write_inode(struct inode *,int); | 214 | extern int nfs_write_inode(struct inode *, struct writeback_control *); |
| 215 | extern void nfs_clear_inode(struct inode *); | 215 | extern void nfs_clear_inode(struct inode *); |
| 216 | #ifdef CONFIG_NFS_V4 | 216 | #ifdef CONFIG_NFS_V4 |
| 217 | extern void nfs4_clear_inode(struct inode *); | 217 | extern void nfs4_clear_inode(struct inode *); |
diff --git a/fs/nfs/iostat.h b/fs/nfs/iostat.h index 46d779abafd3..1d8d5c813b01 100644 --- a/fs/nfs/iostat.h +++ b/fs/nfs/iostat.h | |||
| @@ -57,12 +57,12 @@ static inline void nfs_add_fscache_stats(struct inode *inode, | |||
| 57 | } | 57 | } |
| 58 | #endif | 58 | #endif |
| 59 | 59 | ||
| 60 | static inline struct nfs_iostats *nfs_alloc_iostats(void) | 60 | static inline struct nfs_iostats __percpu *nfs_alloc_iostats(void) |
| 61 | { | 61 | { |
| 62 | return alloc_percpu(struct nfs_iostats); | 62 | return alloc_percpu(struct nfs_iostats); |
| 63 | } | 63 | } |
| 64 | 64 | ||
| 65 | static inline void nfs_free_iostats(struct nfs_iostats *stats) | 65 | static inline void nfs_free_iostats(struct nfs_iostats __percpu *stats) |
| 66 | { | 66 | { |
| 67 | if (stats != NULL) | 67 | if (stats != NULL) |
| 68 | free_percpu(stats); | 68 | free_percpu(stats); |
diff --git a/fs/nfs/mount_clnt.c b/fs/nfs/mount_clnt.c index 0adefc40cc89..59047f8d7d72 100644 --- a/fs/nfs/mount_clnt.c +++ b/fs/nfs/mount_clnt.c | |||
| @@ -120,7 +120,7 @@ static struct { | |||
| 120 | { .status = MNT3ERR_INVAL, .errno = -EINVAL, }, | 120 | { .status = MNT3ERR_INVAL, .errno = -EINVAL, }, |
| 121 | { .status = MNT3ERR_NAMETOOLONG, .errno = -ENAMETOOLONG, }, | 121 | { .status = MNT3ERR_NAMETOOLONG, .errno = -ENAMETOOLONG, }, |
| 122 | { .status = MNT3ERR_NOTSUPP, .errno = -ENOTSUPP, }, | 122 | { .status = MNT3ERR_NOTSUPP, .errno = -ENOTSUPP, }, |
| 123 | { .status = MNT3ERR_SERVERFAULT, .errno = -ESERVERFAULT, }, | 123 | { .status = MNT3ERR_SERVERFAULT, .errno = -EREMOTEIO, }, |
| 124 | }; | 124 | }; |
| 125 | 125 | ||
| 126 | struct mountres { | 126 | struct mountres { |
diff --git a/fs/nfs/nfs2xdr.c b/fs/nfs/nfs2xdr.c index 5e078b222b4e..7bc2da8efd4a 100644 --- a/fs/nfs/nfs2xdr.c +++ b/fs/nfs/nfs2xdr.c | |||
| @@ -699,7 +699,7 @@ static struct { | |||
| 699 | { NFSERR_BAD_COOKIE, -EBADCOOKIE }, | 699 | { NFSERR_BAD_COOKIE, -EBADCOOKIE }, |
| 700 | { NFSERR_NOTSUPP, -ENOTSUPP }, | 700 | { NFSERR_NOTSUPP, -ENOTSUPP }, |
| 701 | { NFSERR_TOOSMALL, -ETOOSMALL }, | 701 | { NFSERR_TOOSMALL, -ETOOSMALL }, |
| 702 | { NFSERR_SERVERFAULT, -ESERVERFAULT }, | 702 | { NFSERR_SERVERFAULT, -EREMOTEIO }, |
| 703 | { NFSERR_BADTYPE, -EBADTYPE }, | 703 | { NFSERR_BADTYPE, -EBADTYPE }, |
| 704 | { NFSERR_JUKEBOX, -EJUKEBOX }, | 704 | { NFSERR_JUKEBOX, -EJUKEBOX }, |
| 705 | { -1, -EIO } | 705 | { -1, -EIO } |
diff --git a/fs/nfs/nfs4proc.c b/fs/nfs/nfs4proc.c index adc116c57e14..eda74c42d552 100644 --- a/fs/nfs/nfs4proc.c +++ b/fs/nfs/nfs4proc.c | |||
| @@ -726,8 +726,8 @@ static struct nfs4_opendata *nfs4_opendata_alloc(struct path *path, | |||
| 726 | p->o_arg.seqid = nfs_alloc_seqid(&sp->so_seqid); | 726 | p->o_arg.seqid = nfs_alloc_seqid(&sp->so_seqid); |
| 727 | if (p->o_arg.seqid == NULL) | 727 | if (p->o_arg.seqid == NULL) |
| 728 | goto err_free; | 728 | goto err_free; |
| 729 | p->path.mnt = mntget(path->mnt); | 729 | path_get(path); |
| 730 | p->path.dentry = dget(path->dentry); | 730 | p->path = *path; |
| 731 | p->dir = parent; | 731 | p->dir = parent; |
| 732 | p->owner = sp; | 732 | p->owner = sp; |
| 733 | atomic_inc(&sp->so_count); | 733 | atomic_inc(&sp->so_count); |
| @@ -1947,8 +1947,8 @@ int nfs4_do_close(struct path *path, struct nfs4_state *state, int wait) | |||
| 1947 | calldata->res.seqid = calldata->arg.seqid; | 1947 | calldata->res.seqid = calldata->arg.seqid; |
| 1948 | calldata->res.server = server; | 1948 | calldata->res.server = server; |
| 1949 | calldata->res.seq_res.sr_slotid = NFS4_MAX_SLOT_TABLE; | 1949 | calldata->res.seq_res.sr_slotid = NFS4_MAX_SLOT_TABLE; |
| 1950 | calldata->path.mnt = mntget(path->mnt); | 1950 | path_get(path); |
| 1951 | calldata->path.dentry = dget(path->dentry); | 1951 | calldata->path = *path; |
| 1952 | 1952 | ||
| 1953 | msg.rpc_argp = &calldata->arg, | 1953 | msg.rpc_argp = &calldata->arg, |
| 1954 | msg.rpc_resp = &calldata->res, | 1954 | msg.rpc_resp = &calldata->res, |
diff --git a/fs/nfs/nfs4xdr.c b/fs/nfs/nfs4xdr.c index 020ebf151184..4d338be492cb 100644 --- a/fs/nfs/nfs4xdr.c +++ b/fs/nfs/nfs4xdr.c | |||
| @@ -4639,7 +4639,7 @@ static int decode_sequence(struct xdr_stream *xdr, | |||
| 4639 | * If the server returns different values for sessionID, slotID or | 4639 | * If the server returns different values for sessionID, slotID or |
| 4640 | * sequence number, the server is looney tunes. | 4640 | * sequence number, the server is looney tunes. |
| 4641 | */ | 4641 | */ |
| 4642 | status = -ESERVERFAULT; | 4642 | status = -EREMOTEIO; |
| 4643 | 4643 | ||
| 4644 | if (memcmp(id.data, res->sr_session->sess_id.data, | 4644 | if (memcmp(id.data, res->sr_session->sess_id.data, |
| 4645 | NFS4_MAX_SESSIONID_LEN)) { | 4645 | NFS4_MAX_SESSIONID_LEN)) { |
| @@ -5782,7 +5782,7 @@ static struct { | |||
| 5782 | { NFS4ERR_BAD_COOKIE, -EBADCOOKIE }, | 5782 | { NFS4ERR_BAD_COOKIE, -EBADCOOKIE }, |
| 5783 | { NFS4ERR_NOTSUPP, -ENOTSUPP }, | 5783 | { NFS4ERR_NOTSUPP, -ENOTSUPP }, |
| 5784 | { NFS4ERR_TOOSMALL, -ETOOSMALL }, | 5784 | { NFS4ERR_TOOSMALL, -ETOOSMALL }, |
| 5785 | { NFS4ERR_SERVERFAULT, -ESERVERFAULT }, | 5785 | { NFS4ERR_SERVERFAULT, -EREMOTEIO }, |
| 5786 | { NFS4ERR_BADTYPE, -EBADTYPE }, | 5786 | { NFS4ERR_BADTYPE, -EBADTYPE }, |
| 5787 | { NFS4ERR_LOCKED, -EAGAIN }, | 5787 | { NFS4ERR_LOCKED, -EAGAIN }, |
| 5788 | { NFS4ERR_SYMLINK, -ELOOP }, | 5788 | { NFS4ERR_SYMLINK, -ELOOP }, |
| @@ -5809,7 +5809,7 @@ nfs4_stat_to_errno(int stat) | |||
| 5809 | } | 5809 | } |
| 5810 | if (stat <= 10000 || stat > 10100) { | 5810 | if (stat <= 10000 || stat > 10100) { |
| 5811 | /* The server is looney tunes. */ | 5811 | /* The server is looney tunes. */ |
| 5812 | return -ESERVERFAULT; | 5812 | return -EREMOTEIO; |
| 5813 | } | 5813 | } |
| 5814 | /* If we cannot translate the error, the recovery routines should | 5814 | /* If we cannot translate the error, the recovery routines should |
| 5815 | * handle it. | 5815 | * handle it. |
diff --git a/fs/nfs/symlink.c b/fs/nfs/symlink.c index 412738dbfbc7..2ea9e5c27e55 100644 --- a/fs/nfs/symlink.c +++ b/fs/nfs/symlink.c | |||
| @@ -50,7 +50,7 @@ static void *nfs_follow_link(struct dentry *dentry, struct nameidata *nd) | |||
| 50 | struct page *page; | 50 | struct page *page; |
| 51 | void *err; | 51 | void *err; |
| 52 | 52 | ||
| 53 | err = ERR_PTR(nfs_revalidate_mapping_nolock(inode, inode->i_mapping)); | 53 | err = ERR_PTR(nfs_revalidate_mapping(inode, inode->i_mapping)); |
| 54 | if (err) | 54 | if (err) |
| 55 | goto read_failed; | 55 | goto read_failed; |
| 56 | page = read_cache_page(&inode->i_data, 0, | 56 | page = read_cache_page(&inode->i_data, 0, |
diff --git a/fs/nfs/write.c b/fs/nfs/write.c index 7b54b8bb101f..53ff70e23993 100644 --- a/fs/nfs/write.c +++ b/fs/nfs/write.c | |||
| @@ -438,6 +438,7 @@ nfs_mark_request_commit(struct nfs_page *req) | |||
| 438 | radix_tree_tag_set(&nfsi->nfs_page_tree, | 438 | radix_tree_tag_set(&nfsi->nfs_page_tree, |
| 439 | req->wb_index, | 439 | req->wb_index, |
| 440 | NFS_PAGE_TAG_COMMIT); | 440 | NFS_PAGE_TAG_COMMIT); |
| 441 | nfsi->ncommit++; | ||
| 441 | spin_unlock(&inode->i_lock); | 442 | spin_unlock(&inode->i_lock); |
| 442 | inc_zone_page_state(req->wb_page, NR_UNSTABLE_NFS); | 443 | inc_zone_page_state(req->wb_page, NR_UNSTABLE_NFS); |
| 443 | inc_bdi_stat(req->wb_page->mapping->backing_dev_info, BDI_RECLAIMABLE); | 444 | inc_bdi_stat(req->wb_page->mapping->backing_dev_info, BDI_RECLAIMABLE); |
| @@ -501,57 +502,6 @@ int nfs_reschedule_unstable_write(struct nfs_page *req) | |||
| 501 | } | 502 | } |
| 502 | #endif | 503 | #endif |
| 503 | 504 | ||
| 504 | /* | ||
| 505 | * Wait for a request to complete. | ||
| 506 | * | ||
| 507 | * Interruptible by fatal signals only. | ||
| 508 | */ | ||
| 509 | static int nfs_wait_on_requests_locked(struct inode *inode, pgoff_t idx_start, unsigned int npages) | ||
| 510 | { | ||
| 511 | struct nfs_inode *nfsi = NFS_I(inode); | ||
| 512 | struct nfs_page *req; | ||
| 513 | pgoff_t idx_end, next; | ||
| 514 | unsigned int res = 0; | ||
| 515 | int error; | ||
| 516 | |||
| 517 | if (npages == 0) | ||
| 518 | idx_end = ~0; | ||
| 519 | else | ||
| 520 | idx_end = idx_start + npages - 1; | ||
| 521 | |||
| 522 | next = idx_start; | ||
| 523 | while (radix_tree_gang_lookup_tag(&nfsi->nfs_page_tree, (void **)&req, next, 1, NFS_PAGE_TAG_LOCKED)) { | ||
| 524 | if (req->wb_index > idx_end) | ||
| 525 | break; | ||
| 526 | |||
| 527 | next = req->wb_index + 1; | ||
| 528 | BUG_ON(!NFS_WBACK_BUSY(req)); | ||
| 529 | |||
| 530 | kref_get(&req->wb_kref); | ||
| 531 | spin_unlock(&inode->i_lock); | ||
| 532 | error = nfs_wait_on_request(req); | ||
| 533 | nfs_release_request(req); | ||
| 534 | spin_lock(&inode->i_lock); | ||
| 535 | if (error < 0) | ||
| 536 | return error; | ||
| 537 | res++; | ||
| 538 | } | ||
| 539 | return res; | ||
| 540 | } | ||
| 541 | |||
| 542 | static void nfs_cancel_commit_list(struct list_head *head) | ||
| 543 | { | ||
| 544 | struct nfs_page *req; | ||
| 545 | |||
| 546 | while(!list_empty(head)) { | ||
| 547 | req = nfs_list_entry(head->next); | ||
| 548 | nfs_list_remove_request(req); | ||
| 549 | nfs_clear_request_commit(req); | ||
| 550 | nfs_inode_remove_request(req); | ||
| 551 | nfs_unlock_request(req); | ||
| 552 | } | ||
| 553 | } | ||
| 554 | |||
| 555 | #if defined(CONFIG_NFS_V3) || defined(CONFIG_NFS_V4) | 505 | #if defined(CONFIG_NFS_V3) || defined(CONFIG_NFS_V4) |
| 556 | static int | 506 | static int |
| 557 | nfs_need_commit(struct nfs_inode *nfsi) | 507 | nfs_need_commit(struct nfs_inode *nfsi) |
| @@ -573,11 +523,17 @@ static int | |||
| 573 | nfs_scan_commit(struct inode *inode, struct list_head *dst, pgoff_t idx_start, unsigned int npages) | 523 | nfs_scan_commit(struct inode *inode, struct list_head *dst, pgoff_t idx_start, unsigned int npages) |
| 574 | { | 524 | { |
| 575 | struct nfs_inode *nfsi = NFS_I(inode); | 525 | struct nfs_inode *nfsi = NFS_I(inode); |
| 526 | int ret; | ||
| 576 | 527 | ||
| 577 | if (!nfs_need_commit(nfsi)) | 528 | if (!nfs_need_commit(nfsi)) |
| 578 | return 0; | 529 | return 0; |
| 579 | 530 | ||
| 580 | return nfs_scan_list(nfsi, dst, idx_start, npages, NFS_PAGE_TAG_COMMIT); | 531 | ret = nfs_scan_list(nfsi, dst, idx_start, npages, NFS_PAGE_TAG_COMMIT); |
| 532 | if (ret > 0) | ||
| 533 | nfsi->ncommit -= ret; | ||
| 534 | if (nfs_need_commit(NFS_I(inode))) | ||
| 535 | __mark_inode_dirty(inode, I_DIRTY_DATASYNC); | ||
| 536 | return ret; | ||
| 581 | } | 537 | } |
| 582 | #else | 538 | #else |
| 583 | static inline int nfs_need_commit(struct nfs_inode *nfsi) | 539 | static inline int nfs_need_commit(struct nfs_inode *nfsi) |
| @@ -642,9 +598,10 @@ static struct nfs_page *nfs_try_to_update_request(struct inode *inode, | |||
| 642 | spin_lock(&inode->i_lock); | 598 | spin_lock(&inode->i_lock); |
| 643 | } | 599 | } |
| 644 | 600 | ||
| 645 | if (nfs_clear_request_commit(req)) | 601 | if (nfs_clear_request_commit(req) && |
| 646 | radix_tree_tag_clear(&NFS_I(inode)->nfs_page_tree, | 602 | radix_tree_tag_clear(&NFS_I(inode)->nfs_page_tree, |
| 647 | req->wb_index, NFS_PAGE_TAG_COMMIT); | 603 | req->wb_index, NFS_PAGE_TAG_COMMIT) != NULL) |
| 604 | NFS_I(inode)->ncommit--; | ||
| 648 | 605 | ||
| 649 | /* Okay, the request matches. Update the region */ | 606 | /* Okay, the request matches. Update the region */ |
| 650 | if (offset < req->wb_offset) { | 607 | if (offset < req->wb_offset) { |
| @@ -1391,7 +1348,7 @@ static const struct rpc_call_ops nfs_commit_ops = { | |||
| 1391 | .rpc_release = nfs_commit_release, | 1348 | .rpc_release = nfs_commit_release, |
| 1392 | }; | 1349 | }; |
| 1393 | 1350 | ||
| 1394 | int nfs_commit_inode(struct inode *inode, int how) | 1351 | static int nfs_commit_inode(struct inode *inode, int how) |
| 1395 | { | 1352 | { |
| 1396 | LIST_HEAD(head); | 1353 | LIST_HEAD(head); |
| 1397 | int res; | 1354 | int res; |
| @@ -1406,92 +1363,51 @@ int nfs_commit_inode(struct inode *inode, int how) | |||
| 1406 | } | 1363 | } |
| 1407 | return res; | 1364 | return res; |
| 1408 | } | 1365 | } |
| 1409 | #else | ||
| 1410 | static inline int nfs_commit_list(struct inode *inode, struct list_head *head, int how) | ||
| 1411 | { | ||
| 1412 | return 0; | ||
| 1413 | } | ||
| 1414 | #endif | ||
| 1415 | 1366 | ||
| 1416 | long nfs_sync_mapping_wait(struct address_space *mapping, struct writeback_control *wbc, int how) | 1367 | static int nfs_commit_unstable_pages(struct inode *inode, struct writeback_control *wbc) |
| 1417 | { | 1368 | { |
| 1418 | struct inode *inode = mapping->host; | 1369 | struct nfs_inode *nfsi = NFS_I(inode); |
| 1419 | pgoff_t idx_start, idx_end; | 1370 | int flags = FLUSH_SYNC; |
| 1420 | unsigned int npages = 0; | 1371 | int ret = 0; |
| 1421 | LIST_HEAD(head); | 1372 | |
| 1422 | int nocommit = how & FLUSH_NOCOMMIT; | 1373 | /* Don't commit yet if this is a non-blocking flush and there are |
| 1423 | long pages, ret; | 1374 | * lots of outstanding writes for this mapping. |
| 1424 | 1375 | */ | |
| 1425 | /* FIXME */ | 1376 | if (wbc->sync_mode == WB_SYNC_NONE && |
| 1426 | if (wbc->range_cyclic) | 1377 | nfsi->ncommit <= (nfsi->npages >> 1)) |
| 1427 | idx_start = 0; | 1378 | goto out_mark_dirty; |
| 1428 | else { | 1379 | |
| 1429 | idx_start = wbc->range_start >> PAGE_CACHE_SHIFT; | 1380 | if (wbc->nonblocking || wbc->for_background) |
| 1430 | idx_end = wbc->range_end >> PAGE_CACHE_SHIFT; | 1381 | flags = 0; |
| 1431 | if (idx_end > idx_start) { | 1382 | ret = nfs_commit_inode(inode, flags); |
| 1432 | pgoff_t l_npages = 1 + idx_end - idx_start; | 1383 | if (ret >= 0) { |
| 1433 | npages = l_npages; | 1384 | if (wbc->sync_mode == WB_SYNC_NONE) { |
| 1434 | if (sizeof(npages) != sizeof(l_npages) && | 1385 | if (ret < wbc->nr_to_write) |
| 1435 | (pgoff_t)npages != l_npages) | 1386 | wbc->nr_to_write -= ret; |
| 1436 | npages = 0; | 1387 | else |
| 1388 | wbc->nr_to_write = 0; | ||
| 1437 | } | 1389 | } |
| 1390 | return 0; | ||
| 1438 | } | 1391 | } |
| 1439 | how &= ~FLUSH_NOCOMMIT; | 1392 | out_mark_dirty: |
| 1440 | spin_lock(&inode->i_lock); | 1393 | __mark_inode_dirty(inode, I_DIRTY_DATASYNC); |
| 1441 | do { | ||
| 1442 | ret = nfs_wait_on_requests_locked(inode, idx_start, npages); | ||
| 1443 | if (ret != 0) | ||
| 1444 | continue; | ||
| 1445 | if (nocommit) | ||
| 1446 | break; | ||
| 1447 | pages = nfs_scan_commit(inode, &head, idx_start, npages); | ||
| 1448 | if (pages == 0) | ||
| 1449 | break; | ||
| 1450 | if (how & FLUSH_INVALIDATE) { | ||
| 1451 | spin_unlock(&inode->i_lock); | ||
| 1452 | nfs_cancel_commit_list(&head); | ||
| 1453 | ret = pages; | ||
| 1454 | spin_lock(&inode->i_lock); | ||
| 1455 | continue; | ||
| 1456 | } | ||
| 1457 | pages += nfs_scan_commit(inode, &head, 0, 0); | ||
| 1458 | spin_unlock(&inode->i_lock); | ||
| 1459 | ret = nfs_commit_list(inode, &head, how); | ||
| 1460 | spin_lock(&inode->i_lock); | ||
| 1461 | |||
| 1462 | } while (ret >= 0); | ||
| 1463 | spin_unlock(&inode->i_lock); | ||
| 1464 | return ret; | 1394 | return ret; |
| 1465 | } | 1395 | } |
| 1466 | 1396 | #else | |
| 1467 | static int __nfs_write_mapping(struct address_space *mapping, struct writeback_control *wbc, int how) | 1397 | static int nfs_commit_inode(struct inode *inode, int how) |
| 1468 | { | 1398 | { |
| 1469 | int ret; | ||
| 1470 | |||
| 1471 | ret = nfs_writepages(mapping, wbc); | ||
| 1472 | if (ret < 0) | ||
| 1473 | goto out; | ||
| 1474 | ret = nfs_sync_mapping_wait(mapping, wbc, how); | ||
| 1475 | if (ret < 0) | ||
| 1476 | goto out; | ||
| 1477 | return 0; | 1399 | return 0; |
| 1478 | out: | ||
| 1479 | __mark_inode_dirty(mapping->host, I_DIRTY_PAGES); | ||
| 1480 | return ret; | ||
| 1481 | } | 1400 | } |
| 1482 | 1401 | ||
| 1483 | /* Two pass sync: first using WB_SYNC_NONE, then WB_SYNC_ALL */ | 1402 | static int nfs_commit_unstable_pages(struct inode *inode, struct writeback_control *wbc) |
| 1484 | static int nfs_write_mapping(struct address_space *mapping, int how) | ||
| 1485 | { | 1403 | { |
| 1486 | struct writeback_control wbc = { | 1404 | return 0; |
| 1487 | .bdi = mapping->backing_dev_info, | 1405 | } |
| 1488 | .sync_mode = WB_SYNC_ALL, | 1406 | #endif |
| 1489 | .nr_to_write = LONG_MAX, | ||
| 1490 | .range_start = 0, | ||
| 1491 | .range_end = LLONG_MAX, | ||
| 1492 | }; | ||
| 1493 | 1407 | ||
| 1494 | return __nfs_write_mapping(mapping, &wbc, how); | 1408 | int nfs_write_inode(struct inode *inode, struct writeback_control *wbc) |
| 1409 | { | ||
| 1410 | return nfs_commit_unstable_pages(inode, wbc); | ||
| 1495 | } | 1411 | } |
| 1496 | 1412 | ||
| 1497 | /* | 1413 | /* |
| @@ -1499,37 +1415,26 @@ static int nfs_write_mapping(struct address_space *mapping, int how) | |||
| 1499 | */ | 1415 | */ |
| 1500 | int nfs_wb_all(struct inode *inode) | 1416 | int nfs_wb_all(struct inode *inode) |
| 1501 | { | 1417 | { |
| 1502 | return nfs_write_mapping(inode->i_mapping, 0); | 1418 | struct writeback_control wbc = { |
| 1503 | } | 1419 | .sync_mode = WB_SYNC_ALL, |
| 1420 | .nr_to_write = LONG_MAX, | ||
| 1421 | .range_start = 0, | ||
| 1422 | .range_end = LLONG_MAX, | ||
| 1423 | }; | ||
| 1504 | 1424 | ||
| 1505 | int nfs_wb_nocommit(struct inode *inode) | 1425 | return sync_inode(inode, &wbc); |
| 1506 | { | ||
| 1507 | return nfs_write_mapping(inode->i_mapping, FLUSH_NOCOMMIT); | ||
| 1508 | } | 1426 | } |
| 1509 | 1427 | ||
| 1510 | int nfs_wb_page_cancel(struct inode *inode, struct page *page) | 1428 | int nfs_wb_page_cancel(struct inode *inode, struct page *page) |
| 1511 | { | 1429 | { |
| 1512 | struct nfs_page *req; | 1430 | struct nfs_page *req; |
| 1513 | loff_t range_start = page_offset(page); | ||
| 1514 | loff_t range_end = range_start + (loff_t)(PAGE_CACHE_SIZE - 1); | ||
| 1515 | struct writeback_control wbc = { | ||
| 1516 | .bdi = page->mapping->backing_dev_info, | ||
| 1517 | .sync_mode = WB_SYNC_ALL, | ||
| 1518 | .nr_to_write = LONG_MAX, | ||
| 1519 | .range_start = range_start, | ||
| 1520 | .range_end = range_end, | ||
| 1521 | }; | ||
| 1522 | int ret = 0; | 1431 | int ret = 0; |
| 1523 | 1432 | ||
| 1524 | BUG_ON(!PageLocked(page)); | 1433 | BUG_ON(!PageLocked(page)); |
| 1525 | for (;;) { | 1434 | for (;;) { |
| 1526 | req = nfs_page_find_request(page); | 1435 | req = nfs_page_find_request(page); |
| 1527 | if (req == NULL) | 1436 | if (req == NULL) |
| 1528 | goto out; | ||
| 1529 | if (test_bit(PG_CLEAN, &req->wb_flags)) { | ||
| 1530 | nfs_release_request(req); | ||
| 1531 | break; | 1437 | break; |
| 1532 | } | ||
| 1533 | if (nfs_lock_request_dontget(req)) { | 1438 | if (nfs_lock_request_dontget(req)) { |
| 1534 | nfs_inode_remove_request(req); | 1439 | nfs_inode_remove_request(req); |
| 1535 | /* | 1440 | /* |
| @@ -1543,54 +1448,54 @@ int nfs_wb_page_cancel(struct inode *inode, struct page *page) | |||
| 1543 | ret = nfs_wait_on_request(req); | 1448 | ret = nfs_wait_on_request(req); |
| 1544 | nfs_release_request(req); | 1449 | nfs_release_request(req); |
| 1545 | if (ret < 0) | 1450 | if (ret < 0) |
| 1546 | goto out; | 1451 | break; |
| 1547 | } | 1452 | } |
| 1548 | if (!PagePrivate(page)) | ||
| 1549 | return 0; | ||
| 1550 | ret = nfs_sync_mapping_wait(page->mapping, &wbc, FLUSH_INVALIDATE); | ||
| 1551 | out: | ||
| 1552 | return ret; | 1453 | return ret; |
| 1553 | } | 1454 | } |
| 1554 | 1455 | ||
| 1555 | static int nfs_wb_page_priority(struct inode *inode, struct page *page, | 1456 | /* |
| 1556 | int how) | 1457 | * Write back all requests on one page - we do this before reading it. |
| 1458 | */ | ||
| 1459 | int nfs_wb_page(struct inode *inode, struct page *page) | ||
| 1557 | { | 1460 | { |
| 1558 | loff_t range_start = page_offset(page); | 1461 | loff_t range_start = page_offset(page); |
| 1559 | loff_t range_end = range_start + (loff_t)(PAGE_CACHE_SIZE - 1); | 1462 | loff_t range_end = range_start + (loff_t)(PAGE_CACHE_SIZE - 1); |
| 1560 | struct writeback_control wbc = { | 1463 | struct writeback_control wbc = { |
| 1561 | .bdi = page->mapping->backing_dev_info, | ||
| 1562 | .sync_mode = WB_SYNC_ALL, | 1464 | .sync_mode = WB_SYNC_ALL, |
| 1563 | .nr_to_write = LONG_MAX, | 1465 | .nr_to_write = 0, |
| 1564 | .range_start = range_start, | 1466 | .range_start = range_start, |
| 1565 | .range_end = range_end, | 1467 | .range_end = range_end, |
| 1566 | }; | 1468 | }; |
| 1469 | struct nfs_page *req; | ||
| 1470 | int need_commit; | ||
| 1567 | int ret; | 1471 | int ret; |
| 1568 | 1472 | ||
| 1569 | do { | 1473 | while(PagePrivate(page)) { |
| 1570 | if (clear_page_dirty_for_io(page)) { | 1474 | if (clear_page_dirty_for_io(page)) { |
| 1571 | ret = nfs_writepage_locked(page, &wbc); | 1475 | ret = nfs_writepage_locked(page, &wbc); |
| 1572 | if (ret < 0) | 1476 | if (ret < 0) |
| 1573 | goto out_error; | 1477 | goto out_error; |
| 1574 | } else if (!PagePrivate(page)) | 1478 | } |
| 1479 | req = nfs_find_and_lock_request(page); | ||
| 1480 | if (!req) | ||
| 1575 | break; | 1481 | break; |
| 1576 | ret = nfs_sync_mapping_wait(page->mapping, &wbc, how); | 1482 | if (IS_ERR(req)) { |
| 1577 | if (ret < 0) | 1483 | ret = PTR_ERR(req); |
| 1578 | goto out_error; | 1484 | goto out_error; |
| 1579 | } while (PagePrivate(page)); | 1485 | } |
| 1486 | need_commit = test_bit(PG_CLEAN, &req->wb_flags); | ||
| 1487 | nfs_clear_page_tag_locked(req); | ||
| 1488 | if (need_commit) { | ||
| 1489 | ret = nfs_commit_inode(inode, FLUSH_SYNC); | ||
| 1490 | if (ret < 0) | ||
| 1491 | goto out_error; | ||
| 1492 | } | ||
| 1493 | } | ||
| 1580 | return 0; | 1494 | return 0; |
| 1581 | out_error: | 1495 | out_error: |
| 1582 | __mark_inode_dirty(inode, I_DIRTY_PAGES); | ||
| 1583 | return ret; | 1496 | return ret; |
| 1584 | } | 1497 | } |
| 1585 | 1498 | ||
| 1586 | /* | ||
| 1587 | * Write back all requests on one page - we do this before reading it. | ||
| 1588 | */ | ||
| 1589 | int nfs_wb_page(struct inode *inode, struct page* page) | ||
| 1590 | { | ||
| 1591 | return nfs_wb_page_priority(inode, page, FLUSH_STABLE); | ||
| 1592 | } | ||
| 1593 | |||
| 1594 | #ifdef CONFIG_MIGRATION | 1499 | #ifdef CONFIG_MIGRATION |
| 1595 | int nfs_migrate_page(struct address_space *mapping, struct page *newpage, | 1500 | int nfs_migrate_page(struct address_space *mapping, struct page *newpage, |
| 1596 | struct page *page) | 1501 | struct page *page) |
| @@ -1598,8 +1503,7 @@ int nfs_migrate_page(struct address_space *mapping, struct page *newpage, | |||
| 1598 | struct nfs_page *req; | 1503 | struct nfs_page *req; |
| 1599 | int ret; | 1504 | int ret; |
| 1600 | 1505 | ||
| 1601 | if (PageFsCache(page)) | 1506 | nfs_fscache_release_page(page, GFP_KERNEL); |
| 1602 | nfs_fscache_release_page(page, GFP_KERNEL); | ||
| 1603 | 1507 | ||
| 1604 | req = nfs_find_and_lock_request(page); | 1508 | req = nfs_find_and_lock_request(page); |
| 1605 | ret = PTR_ERR(req); | 1509 | ret = PTR_ERR(req); |
