diff options
author | Trond Myklebust <Trond.Myklebust@netapp.com> | 2010-03-05 15:46:18 -0500 |
---|---|---|
committer | Trond Myklebust <Trond.Myklebust@netapp.com> | 2010-03-05 15:46:18 -0500 |
commit | 3fa04ecd72780da31ba8b329e148179bc24a9c7d (patch) | |
tree | f5d462fd4aee086952d18f159f737c450ab46b3b /fs/nfs | |
parent | 180b62a3d837613fcac3ce89576526423926c3c3 (diff) | |
parent | 1cda707d52e51a6cafac0aef12d2bd7052d572e6 (diff) |
Merge branch 'writeback-for-2.6.34' into nfs-for-2.6.34
Diffstat (limited to 'fs/nfs')
-rw-r--r-- | fs/nfs/Kconfig | 3 | ||||
-rw-r--r-- | fs/nfs/dir.c | 2 | ||||
-rw-r--r-- | fs/nfs/direct.c | 3 | ||||
-rw-r--r-- | fs/nfs/fscache.c | 9 | ||||
-rw-r--r-- | fs/nfs/inode.c | 95 | ||||
-rw-r--r-- | fs/nfs/internal.h | 2 | ||||
-rw-r--r-- | fs/nfs/iostat.h | 4 | ||||
-rw-r--r-- | fs/nfs/mount_clnt.c | 2 | ||||
-rw-r--r-- | fs/nfs/nfs2xdr.c | 2 | ||||
-rw-r--r-- | fs/nfs/nfs4proc.c | 8 | ||||
-rw-r--r-- | fs/nfs/nfs4xdr.c | 6 | ||||
-rw-r--r-- | fs/nfs/symlink.c | 2 | ||||
-rw-r--r-- | fs/nfs/write.c | 250 |
13 files changed, 115 insertions, 273 deletions
diff --git a/fs/nfs/Kconfig b/fs/nfs/Kconfig index 59e5673b4597..a43d07e7b924 100644 --- a/fs/nfs/Kconfig +++ b/fs/nfs/Kconfig | |||
@@ -95,8 +95,7 @@ config ROOT_NFS | |||
95 | Most people say N here. | 95 | Most people say N here. |
96 | 96 | ||
97 | config NFS_FSCACHE | 97 | config NFS_FSCACHE |
98 | bool "Provide NFS client caching support (EXPERIMENTAL)" | 98 | bool "Provide NFS client caching support" |
99 | depends on EXPERIMENTAL | ||
100 | depends on NFS_FS=m && FSCACHE || NFS_FS=y && FSCACHE=y | 99 | depends on NFS_FS=m && FSCACHE || NFS_FS=y && FSCACHE=y |
101 | help | 100 | help |
102 | Say Y here if you want NFS data to be cached locally on disc through | 101 | Say Y here if you want NFS data to be cached locally on disc through |
diff --git a/fs/nfs/dir.c b/fs/nfs/dir.c index 3c7f03b669fb..a1f6b4438fb1 100644 --- a/fs/nfs/dir.c +++ b/fs/nfs/dir.c | |||
@@ -560,7 +560,7 @@ static int nfs_readdir(struct file *filp, void *dirent, filldir_t filldir) | |||
560 | desc->entry = &my_entry; | 560 | desc->entry = &my_entry; |
561 | 561 | ||
562 | nfs_block_sillyrename(dentry); | 562 | nfs_block_sillyrename(dentry); |
563 | res = nfs_revalidate_mapping_nolock(inode, filp->f_mapping); | 563 | res = nfs_revalidate_mapping(inode, filp->f_mapping); |
564 | if (res < 0) | 564 | if (res < 0) |
565 | goto out; | 565 | goto out; |
566 | 566 | ||
diff --git a/fs/nfs/direct.c b/fs/nfs/direct.c index e1d415e97849..0d289823e856 100644 --- a/fs/nfs/direct.c +++ b/fs/nfs/direct.c | |||
@@ -342,6 +342,7 @@ static ssize_t nfs_direct_read_schedule_segment(struct nfs_direct_req *dreq, | |||
342 | data->res.fattr = &data->fattr; | 342 | data->res.fattr = &data->fattr; |
343 | data->res.eof = 0; | 343 | data->res.eof = 0; |
344 | data->res.count = bytes; | 344 | data->res.count = bytes; |
345 | nfs_fattr_init(&data->fattr); | ||
345 | msg.rpc_argp = &data->args; | 346 | msg.rpc_argp = &data->args; |
346 | msg.rpc_resp = &data->res; | 347 | msg.rpc_resp = &data->res; |
347 | 348 | ||
@@ -575,6 +576,7 @@ static void nfs_direct_commit_schedule(struct nfs_direct_req *dreq) | |||
575 | data->res.count = 0; | 576 | data->res.count = 0; |
576 | data->res.fattr = &data->fattr; | 577 | data->res.fattr = &data->fattr; |
577 | data->res.verf = &data->verf; | 578 | data->res.verf = &data->verf; |
579 | nfs_fattr_init(&data->fattr); | ||
578 | 580 | ||
579 | NFS_PROTO(data->inode)->commit_setup(data, &msg); | 581 | NFS_PROTO(data->inode)->commit_setup(data, &msg); |
580 | 582 | ||
@@ -766,6 +768,7 @@ static ssize_t nfs_direct_write_schedule_segment(struct nfs_direct_req *dreq, | |||
766 | data->res.fattr = &data->fattr; | 768 | data->res.fattr = &data->fattr; |
767 | data->res.count = bytes; | 769 | data->res.count = bytes; |
768 | data->res.verf = &data->verf; | 770 | data->res.verf = &data->verf; |
771 | nfs_fattr_init(&data->fattr); | ||
769 | 772 | ||
770 | task_setup_data.task = &data->task; | 773 | task_setup_data.task = &data->task; |
771 | task_setup_data.callback_data = data; | 774 | task_setup_data.callback_data = data; |
diff --git a/fs/nfs/fscache.c b/fs/nfs/fscache.c index fa588006588d..237874f1af23 100644 --- a/fs/nfs/fscache.c +++ b/fs/nfs/fscache.c | |||
@@ -354,12 +354,11 @@ void nfs_fscache_reset_inode_cookie(struct inode *inode) | |||
354 | */ | 354 | */ |
355 | int nfs_fscache_release_page(struct page *page, gfp_t gfp) | 355 | int nfs_fscache_release_page(struct page *page, gfp_t gfp) |
356 | { | 356 | { |
357 | struct nfs_inode *nfsi = NFS_I(page->mapping->host); | ||
358 | struct fscache_cookie *cookie = nfsi->fscache; | ||
359 | |||
360 | BUG_ON(!cookie); | ||
361 | |||
362 | if (PageFsCache(page)) { | 357 | if (PageFsCache(page)) { |
358 | struct nfs_inode *nfsi = NFS_I(page->mapping->host); | ||
359 | struct fscache_cookie *cookie = nfsi->fscache; | ||
360 | |||
361 | BUG_ON(!cookie); | ||
363 | dfprintk(FSCACHE, "NFS: fscache releasepage (0x%p/0x%p/0x%p)\n", | 362 | dfprintk(FSCACHE, "NFS: fscache releasepage (0x%p/0x%p/0x%p)\n", |
364 | cookie, page, nfsi); | 363 | cookie, page, nfsi); |
365 | 364 | ||
diff --git a/fs/nfs/inode.c b/fs/nfs/inode.c index 87cca56846d6..657201acda84 100644 --- a/fs/nfs/inode.c +++ b/fs/nfs/inode.c | |||
@@ -97,22 +97,6 @@ u64 nfs_compat_user_ino64(u64 fileid) | |||
97 | return ino; | 97 | return ino; |
98 | } | 98 | } |
99 | 99 | ||
100 | int nfs_write_inode(struct inode *inode, int sync) | ||
101 | { | ||
102 | int ret; | ||
103 | |||
104 | if (sync) { | ||
105 | ret = filemap_fdatawait(inode->i_mapping); | ||
106 | if (ret == 0) | ||
107 | ret = nfs_commit_inode(inode, FLUSH_SYNC); | ||
108 | } else | ||
109 | ret = nfs_commit_inode(inode, 0); | ||
110 | if (ret >= 0) | ||
111 | return 0; | ||
112 | __mark_inode_dirty(inode, I_DIRTY_DATASYNC); | ||
113 | return ret; | ||
114 | } | ||
115 | |||
116 | void nfs_clear_inode(struct inode *inode) | 100 | void nfs_clear_inode(struct inode *inode) |
117 | { | 101 | { |
118 | /* | 102 | /* |
@@ -130,16 +114,12 @@ void nfs_clear_inode(struct inode *inode) | |||
130 | */ | 114 | */ |
131 | int nfs_sync_mapping(struct address_space *mapping) | 115 | int nfs_sync_mapping(struct address_space *mapping) |
132 | { | 116 | { |
133 | int ret; | 117 | int ret = 0; |
134 | 118 | ||
135 | if (mapping->nrpages == 0) | 119 | if (mapping->nrpages != 0) { |
136 | return 0; | 120 | unmap_mapping_range(mapping, 0, 0, 0); |
137 | unmap_mapping_range(mapping, 0, 0, 0); | 121 | ret = nfs_wb_all(mapping->host); |
138 | ret = filemap_write_and_wait(mapping); | 122 | } |
139 | if (ret != 0) | ||
140 | goto out; | ||
141 | ret = nfs_wb_all(mapping->host); | ||
142 | out: | ||
143 | return ret; | 123 | return ret; |
144 | } | 124 | } |
145 | 125 | ||
@@ -511,17 +491,11 @@ int nfs_getattr(struct vfsmount *mnt, struct dentry *dentry, struct kstat *stat) | |||
511 | int need_atime = NFS_I(inode)->cache_validity & NFS_INO_INVALID_ATIME; | 491 | int need_atime = NFS_I(inode)->cache_validity & NFS_INO_INVALID_ATIME; |
512 | int err; | 492 | int err; |
513 | 493 | ||
514 | /* | 494 | /* Flush out writes to the server in order to update c/mtime. */ |
515 | * Flush out writes to the server in order to update c/mtime. | ||
516 | * | ||
517 | * Hold the i_mutex to suspend application writes temporarily; | ||
518 | * this prevents long-running writing applications from blocking | ||
519 | * nfs_wb_nocommit. | ||
520 | */ | ||
521 | if (S_ISREG(inode->i_mode)) { | 495 | if (S_ISREG(inode->i_mode)) { |
522 | mutex_lock(&inode->i_mutex); | 496 | err = filemap_write_and_wait(inode->i_mapping); |
523 | nfs_wb_nocommit(inode); | 497 | if (err) |
524 | mutex_unlock(&inode->i_mutex); | 498 | goto out; |
525 | } | 499 | } |
526 | 500 | ||
527 | /* | 501 | /* |
@@ -545,6 +519,7 @@ int nfs_getattr(struct vfsmount *mnt, struct dentry *dentry, struct kstat *stat) | |||
545 | generic_fillattr(inode, stat); | 519 | generic_fillattr(inode, stat); |
546 | stat->ino = nfs_compat_user_ino64(NFS_FILEID(inode)); | 520 | stat->ino = nfs_compat_user_ino64(NFS_FILEID(inode)); |
547 | } | 521 | } |
522 | out: | ||
548 | return err; | 523 | return err; |
549 | } | 524 | } |
550 | 525 | ||
@@ -574,14 +549,14 @@ void nfs_close_context(struct nfs_open_context *ctx, int is_sync) | |||
574 | nfs_revalidate_inode(server, inode); | 549 | nfs_revalidate_inode(server, inode); |
575 | } | 550 | } |
576 | 551 | ||
577 | static struct nfs_open_context *alloc_nfs_open_context(struct vfsmount *mnt, struct dentry *dentry, struct rpc_cred *cred) | 552 | static struct nfs_open_context *alloc_nfs_open_context(struct path *path, struct rpc_cred *cred) |
578 | { | 553 | { |
579 | struct nfs_open_context *ctx; | 554 | struct nfs_open_context *ctx; |
580 | 555 | ||
581 | ctx = kmalloc(sizeof(*ctx), GFP_KERNEL); | 556 | ctx = kmalloc(sizeof(*ctx), GFP_KERNEL); |
582 | if (ctx != NULL) { | 557 | if (ctx != NULL) { |
583 | ctx->path.dentry = dget(dentry); | 558 | ctx->path = *path; |
584 | ctx->path.mnt = mntget(mnt); | 559 | path_get(&ctx->path); |
585 | ctx->cred = get_rpccred(cred); | 560 | ctx->cred = get_rpccred(cred); |
586 | ctx->state = NULL; | 561 | ctx->state = NULL; |
587 | ctx->lockowner = current->files; | 562 | ctx->lockowner = current->files; |
@@ -681,7 +656,7 @@ int nfs_open(struct inode *inode, struct file *filp) | |||
681 | cred = rpc_lookup_cred(); | 656 | cred = rpc_lookup_cred(); |
682 | if (IS_ERR(cred)) | 657 | if (IS_ERR(cred)) |
683 | return PTR_ERR(cred); | 658 | return PTR_ERR(cred); |
684 | ctx = alloc_nfs_open_context(filp->f_path.mnt, filp->f_path.dentry, cred); | 659 | ctx = alloc_nfs_open_context(&filp->f_path, cred); |
685 | put_rpccred(cred); | 660 | put_rpccred(cred); |
686 | if (ctx == NULL) | 661 | if (ctx == NULL) |
687 | return -ENOMEM; | 662 | return -ENOMEM; |
@@ -774,7 +749,7 @@ int nfs_revalidate_inode(struct nfs_server *server, struct inode *inode) | |||
774 | return __nfs_revalidate_inode(server, inode); | 749 | return __nfs_revalidate_inode(server, inode); |
775 | } | 750 | } |
776 | 751 | ||
777 | static int nfs_invalidate_mapping_nolock(struct inode *inode, struct address_space *mapping) | 752 | static int nfs_invalidate_mapping(struct inode *inode, struct address_space *mapping) |
778 | { | 753 | { |
779 | struct nfs_inode *nfsi = NFS_I(inode); | 754 | struct nfs_inode *nfsi = NFS_I(inode); |
780 | 755 | ||
@@ -795,49 +770,10 @@ static int nfs_invalidate_mapping_nolock(struct inode *inode, struct address_spa | |||
795 | return 0; | 770 | return 0; |
796 | } | 771 | } |
797 | 772 | ||
798 | static int nfs_invalidate_mapping(struct inode *inode, struct address_space *mapping) | ||
799 | { | ||
800 | int ret = 0; | ||
801 | |||
802 | mutex_lock(&inode->i_mutex); | ||
803 | if (NFS_I(inode)->cache_validity & NFS_INO_INVALID_DATA) { | ||
804 | ret = nfs_sync_mapping(mapping); | ||
805 | if (ret == 0) | ||
806 | ret = nfs_invalidate_mapping_nolock(inode, mapping); | ||
807 | } | ||
808 | mutex_unlock(&inode->i_mutex); | ||
809 | return ret; | ||
810 | } | ||
811 | |||
812 | /** | ||
813 | * nfs_revalidate_mapping_nolock - Revalidate the pagecache | ||
814 | * @inode - pointer to host inode | ||
815 | * @mapping - pointer to mapping | ||
816 | */ | ||
817 | int nfs_revalidate_mapping_nolock(struct inode *inode, struct address_space *mapping) | ||
818 | { | ||
819 | struct nfs_inode *nfsi = NFS_I(inode); | ||
820 | int ret = 0; | ||
821 | |||
822 | if ((nfsi->cache_validity & NFS_INO_REVAL_PAGECACHE) | ||
823 | || nfs_attribute_timeout(inode) || NFS_STALE(inode)) { | ||
824 | ret = __nfs_revalidate_inode(NFS_SERVER(inode), inode); | ||
825 | if (ret < 0) | ||
826 | goto out; | ||
827 | } | ||
828 | if (nfsi->cache_validity & NFS_INO_INVALID_DATA) | ||
829 | ret = nfs_invalidate_mapping_nolock(inode, mapping); | ||
830 | out: | ||
831 | return ret; | ||
832 | } | ||
833 | |||
834 | /** | 773 | /** |
835 | * nfs_revalidate_mapping - Revalidate the pagecache | 774 | * nfs_revalidate_mapping - Revalidate the pagecache |
836 | * @inode - pointer to host inode | 775 | * @inode - pointer to host inode |
837 | * @mapping - pointer to mapping | 776 | * @mapping - pointer to mapping |
838 | * | ||
839 | * This version of the function will take the inode->i_mutex and attempt to | ||
840 | * flush out all dirty data if it needs to invalidate the page cache. | ||
841 | */ | 777 | */ |
842 | int nfs_revalidate_mapping(struct inode *inode, struct address_space *mapping) | 778 | int nfs_revalidate_mapping(struct inode *inode, struct address_space *mapping) |
843 | { | 779 | { |
@@ -1415,6 +1351,7 @@ static void init_once(void *foo) | |||
1415 | INIT_LIST_HEAD(&nfsi->access_cache_inode_lru); | 1351 | INIT_LIST_HEAD(&nfsi->access_cache_inode_lru); |
1416 | INIT_RADIX_TREE(&nfsi->nfs_page_tree, GFP_ATOMIC); | 1352 | INIT_RADIX_TREE(&nfsi->nfs_page_tree, GFP_ATOMIC); |
1417 | nfsi->npages = 0; | 1353 | nfsi->npages = 0; |
1354 | nfsi->ncommit = 0; | ||
1418 | atomic_set(&nfsi->silly_count, 1); | 1355 | atomic_set(&nfsi->silly_count, 1); |
1419 | INIT_HLIST_HEAD(&nfsi->silly_list); | 1356 | INIT_HLIST_HEAD(&nfsi->silly_list); |
1420 | init_waitqueue_head(&nfsi->waitqueue); | 1357 | init_waitqueue_head(&nfsi->waitqueue); |
diff --git a/fs/nfs/internal.h b/fs/nfs/internal.h index 29e464d23b32..11f82f03c5de 100644 --- a/fs/nfs/internal.h +++ b/fs/nfs/internal.h | |||
@@ -211,7 +211,7 @@ extern int nfs_access_cache_shrinker(int nr_to_scan, gfp_t gfp_mask); | |||
211 | extern struct workqueue_struct *nfsiod_workqueue; | 211 | extern struct workqueue_struct *nfsiod_workqueue; |
212 | extern struct inode *nfs_alloc_inode(struct super_block *sb); | 212 | extern struct inode *nfs_alloc_inode(struct super_block *sb); |
213 | extern void nfs_destroy_inode(struct inode *); | 213 | extern void nfs_destroy_inode(struct inode *); |
214 | extern int nfs_write_inode(struct inode *,int); | 214 | extern int nfs_write_inode(struct inode *, struct writeback_control *); |
215 | extern void nfs_clear_inode(struct inode *); | 215 | extern void nfs_clear_inode(struct inode *); |
216 | #ifdef CONFIG_NFS_V4 | 216 | #ifdef CONFIG_NFS_V4 |
217 | extern void nfs4_clear_inode(struct inode *); | 217 | extern void nfs4_clear_inode(struct inode *); |
diff --git a/fs/nfs/iostat.h b/fs/nfs/iostat.h index 46d779abafd3..1d8d5c813b01 100644 --- a/fs/nfs/iostat.h +++ b/fs/nfs/iostat.h | |||
@@ -57,12 +57,12 @@ static inline void nfs_add_fscache_stats(struct inode *inode, | |||
57 | } | 57 | } |
58 | #endif | 58 | #endif |
59 | 59 | ||
60 | static inline struct nfs_iostats *nfs_alloc_iostats(void) | 60 | static inline struct nfs_iostats __percpu *nfs_alloc_iostats(void) |
61 | { | 61 | { |
62 | return alloc_percpu(struct nfs_iostats); | 62 | return alloc_percpu(struct nfs_iostats); |
63 | } | 63 | } |
64 | 64 | ||
65 | static inline void nfs_free_iostats(struct nfs_iostats *stats) | 65 | static inline void nfs_free_iostats(struct nfs_iostats __percpu *stats) |
66 | { | 66 | { |
67 | if (stats != NULL) | 67 | if (stats != NULL) |
68 | free_percpu(stats); | 68 | free_percpu(stats); |
diff --git a/fs/nfs/mount_clnt.c b/fs/nfs/mount_clnt.c index 0adefc40cc89..59047f8d7d72 100644 --- a/fs/nfs/mount_clnt.c +++ b/fs/nfs/mount_clnt.c | |||
@@ -120,7 +120,7 @@ static struct { | |||
120 | { .status = MNT3ERR_INVAL, .errno = -EINVAL, }, | 120 | { .status = MNT3ERR_INVAL, .errno = -EINVAL, }, |
121 | { .status = MNT3ERR_NAMETOOLONG, .errno = -ENAMETOOLONG, }, | 121 | { .status = MNT3ERR_NAMETOOLONG, .errno = -ENAMETOOLONG, }, |
122 | { .status = MNT3ERR_NOTSUPP, .errno = -ENOTSUPP, }, | 122 | { .status = MNT3ERR_NOTSUPP, .errno = -ENOTSUPP, }, |
123 | { .status = MNT3ERR_SERVERFAULT, .errno = -ESERVERFAULT, }, | 123 | { .status = MNT3ERR_SERVERFAULT, .errno = -EREMOTEIO, }, |
124 | }; | 124 | }; |
125 | 125 | ||
126 | struct mountres { | 126 | struct mountres { |
diff --git a/fs/nfs/nfs2xdr.c b/fs/nfs/nfs2xdr.c index 5e078b222b4e..7bc2da8efd4a 100644 --- a/fs/nfs/nfs2xdr.c +++ b/fs/nfs/nfs2xdr.c | |||
@@ -699,7 +699,7 @@ static struct { | |||
699 | { NFSERR_BAD_COOKIE, -EBADCOOKIE }, | 699 | { NFSERR_BAD_COOKIE, -EBADCOOKIE }, |
700 | { NFSERR_NOTSUPP, -ENOTSUPP }, | 700 | { NFSERR_NOTSUPP, -ENOTSUPP }, |
701 | { NFSERR_TOOSMALL, -ETOOSMALL }, | 701 | { NFSERR_TOOSMALL, -ETOOSMALL }, |
702 | { NFSERR_SERVERFAULT, -ESERVERFAULT }, | 702 | { NFSERR_SERVERFAULT, -EREMOTEIO }, |
703 | { NFSERR_BADTYPE, -EBADTYPE }, | 703 | { NFSERR_BADTYPE, -EBADTYPE }, |
704 | { NFSERR_JUKEBOX, -EJUKEBOX }, | 704 | { NFSERR_JUKEBOX, -EJUKEBOX }, |
705 | { -1, -EIO } | 705 | { -1, -EIO } |
diff --git a/fs/nfs/nfs4proc.c b/fs/nfs/nfs4proc.c index adc116c57e14..eda74c42d552 100644 --- a/fs/nfs/nfs4proc.c +++ b/fs/nfs/nfs4proc.c | |||
@@ -726,8 +726,8 @@ static struct nfs4_opendata *nfs4_opendata_alloc(struct path *path, | |||
726 | p->o_arg.seqid = nfs_alloc_seqid(&sp->so_seqid); | 726 | p->o_arg.seqid = nfs_alloc_seqid(&sp->so_seqid); |
727 | if (p->o_arg.seqid == NULL) | 727 | if (p->o_arg.seqid == NULL) |
728 | goto err_free; | 728 | goto err_free; |
729 | p->path.mnt = mntget(path->mnt); | 729 | path_get(path); |
730 | p->path.dentry = dget(path->dentry); | 730 | p->path = *path; |
731 | p->dir = parent; | 731 | p->dir = parent; |
732 | p->owner = sp; | 732 | p->owner = sp; |
733 | atomic_inc(&sp->so_count); | 733 | atomic_inc(&sp->so_count); |
@@ -1947,8 +1947,8 @@ int nfs4_do_close(struct path *path, struct nfs4_state *state, int wait) | |||
1947 | calldata->res.seqid = calldata->arg.seqid; | 1947 | calldata->res.seqid = calldata->arg.seqid; |
1948 | calldata->res.server = server; | 1948 | calldata->res.server = server; |
1949 | calldata->res.seq_res.sr_slotid = NFS4_MAX_SLOT_TABLE; | 1949 | calldata->res.seq_res.sr_slotid = NFS4_MAX_SLOT_TABLE; |
1950 | calldata->path.mnt = mntget(path->mnt); | 1950 | path_get(path); |
1951 | calldata->path.dentry = dget(path->dentry); | 1951 | calldata->path = *path; |
1952 | 1952 | ||
1953 | msg.rpc_argp = &calldata->arg, | 1953 | msg.rpc_argp = &calldata->arg, |
1954 | msg.rpc_resp = &calldata->res, | 1954 | msg.rpc_resp = &calldata->res, |
diff --git a/fs/nfs/nfs4xdr.c b/fs/nfs/nfs4xdr.c index 020ebf151184..4d338be492cb 100644 --- a/fs/nfs/nfs4xdr.c +++ b/fs/nfs/nfs4xdr.c | |||
@@ -4639,7 +4639,7 @@ static int decode_sequence(struct xdr_stream *xdr, | |||
4639 | * If the server returns different values for sessionID, slotID or | 4639 | * If the server returns different values for sessionID, slotID or |
4640 | * sequence number, the server is looney tunes. | 4640 | * sequence number, the server is looney tunes. |
4641 | */ | 4641 | */ |
4642 | status = -ESERVERFAULT; | 4642 | status = -EREMOTEIO; |
4643 | 4643 | ||
4644 | if (memcmp(id.data, res->sr_session->sess_id.data, | 4644 | if (memcmp(id.data, res->sr_session->sess_id.data, |
4645 | NFS4_MAX_SESSIONID_LEN)) { | 4645 | NFS4_MAX_SESSIONID_LEN)) { |
@@ -5782,7 +5782,7 @@ static struct { | |||
5782 | { NFS4ERR_BAD_COOKIE, -EBADCOOKIE }, | 5782 | { NFS4ERR_BAD_COOKIE, -EBADCOOKIE }, |
5783 | { NFS4ERR_NOTSUPP, -ENOTSUPP }, | 5783 | { NFS4ERR_NOTSUPP, -ENOTSUPP }, |
5784 | { NFS4ERR_TOOSMALL, -ETOOSMALL }, | 5784 | { NFS4ERR_TOOSMALL, -ETOOSMALL }, |
5785 | { NFS4ERR_SERVERFAULT, -ESERVERFAULT }, | 5785 | { NFS4ERR_SERVERFAULT, -EREMOTEIO }, |
5786 | { NFS4ERR_BADTYPE, -EBADTYPE }, | 5786 | { NFS4ERR_BADTYPE, -EBADTYPE }, |
5787 | { NFS4ERR_LOCKED, -EAGAIN }, | 5787 | { NFS4ERR_LOCKED, -EAGAIN }, |
5788 | { NFS4ERR_SYMLINK, -ELOOP }, | 5788 | { NFS4ERR_SYMLINK, -ELOOP }, |
@@ -5809,7 +5809,7 @@ nfs4_stat_to_errno(int stat) | |||
5809 | } | 5809 | } |
5810 | if (stat <= 10000 || stat > 10100) { | 5810 | if (stat <= 10000 || stat > 10100) { |
5811 | /* The server is looney tunes. */ | 5811 | /* The server is looney tunes. */ |
5812 | return -ESERVERFAULT; | 5812 | return -EREMOTEIO; |
5813 | } | 5813 | } |
5814 | /* If we cannot translate the error, the recovery routines should | 5814 | /* If we cannot translate the error, the recovery routines should |
5815 | * handle it. | 5815 | * handle it. |
diff --git a/fs/nfs/symlink.c b/fs/nfs/symlink.c index 412738dbfbc7..2ea9e5c27e55 100644 --- a/fs/nfs/symlink.c +++ b/fs/nfs/symlink.c | |||
@@ -50,7 +50,7 @@ static void *nfs_follow_link(struct dentry *dentry, struct nameidata *nd) | |||
50 | struct page *page; | 50 | struct page *page; |
51 | void *err; | 51 | void *err; |
52 | 52 | ||
53 | err = ERR_PTR(nfs_revalidate_mapping_nolock(inode, inode->i_mapping)); | 53 | err = ERR_PTR(nfs_revalidate_mapping(inode, inode->i_mapping)); |
54 | if (err) | 54 | if (err) |
55 | goto read_failed; | 55 | goto read_failed; |
56 | page = read_cache_page(&inode->i_data, 0, | 56 | page = read_cache_page(&inode->i_data, 0, |
diff --git a/fs/nfs/write.c b/fs/nfs/write.c index 7b54b8bb101f..53ff70e23993 100644 --- a/fs/nfs/write.c +++ b/fs/nfs/write.c | |||
@@ -438,6 +438,7 @@ nfs_mark_request_commit(struct nfs_page *req) | |||
438 | radix_tree_tag_set(&nfsi->nfs_page_tree, | 438 | radix_tree_tag_set(&nfsi->nfs_page_tree, |
439 | req->wb_index, | 439 | req->wb_index, |
440 | NFS_PAGE_TAG_COMMIT); | 440 | NFS_PAGE_TAG_COMMIT); |
441 | nfsi->ncommit++; | ||
441 | spin_unlock(&inode->i_lock); | 442 | spin_unlock(&inode->i_lock); |
442 | inc_zone_page_state(req->wb_page, NR_UNSTABLE_NFS); | 443 | inc_zone_page_state(req->wb_page, NR_UNSTABLE_NFS); |
443 | inc_bdi_stat(req->wb_page->mapping->backing_dev_info, BDI_RECLAIMABLE); | 444 | inc_bdi_stat(req->wb_page->mapping->backing_dev_info, BDI_RECLAIMABLE); |
@@ -501,57 +502,6 @@ int nfs_reschedule_unstable_write(struct nfs_page *req) | |||
501 | } | 502 | } |
502 | #endif | 503 | #endif |
503 | 504 | ||
504 | /* | ||
505 | * Wait for a request to complete. | ||
506 | * | ||
507 | * Interruptible by fatal signals only. | ||
508 | */ | ||
509 | static int nfs_wait_on_requests_locked(struct inode *inode, pgoff_t idx_start, unsigned int npages) | ||
510 | { | ||
511 | struct nfs_inode *nfsi = NFS_I(inode); | ||
512 | struct nfs_page *req; | ||
513 | pgoff_t idx_end, next; | ||
514 | unsigned int res = 0; | ||
515 | int error; | ||
516 | |||
517 | if (npages == 0) | ||
518 | idx_end = ~0; | ||
519 | else | ||
520 | idx_end = idx_start + npages - 1; | ||
521 | |||
522 | next = idx_start; | ||
523 | while (radix_tree_gang_lookup_tag(&nfsi->nfs_page_tree, (void **)&req, next, 1, NFS_PAGE_TAG_LOCKED)) { | ||
524 | if (req->wb_index > idx_end) | ||
525 | break; | ||
526 | |||
527 | next = req->wb_index + 1; | ||
528 | BUG_ON(!NFS_WBACK_BUSY(req)); | ||
529 | |||
530 | kref_get(&req->wb_kref); | ||
531 | spin_unlock(&inode->i_lock); | ||
532 | error = nfs_wait_on_request(req); | ||
533 | nfs_release_request(req); | ||
534 | spin_lock(&inode->i_lock); | ||
535 | if (error < 0) | ||
536 | return error; | ||
537 | res++; | ||
538 | } | ||
539 | return res; | ||
540 | } | ||
541 | |||
542 | static void nfs_cancel_commit_list(struct list_head *head) | ||
543 | { | ||
544 | struct nfs_page *req; | ||
545 | |||
546 | while(!list_empty(head)) { | ||
547 | req = nfs_list_entry(head->next); | ||
548 | nfs_list_remove_request(req); | ||
549 | nfs_clear_request_commit(req); | ||
550 | nfs_inode_remove_request(req); | ||
551 | nfs_unlock_request(req); | ||
552 | } | ||
553 | } | ||
554 | |||
555 | #if defined(CONFIG_NFS_V3) || defined(CONFIG_NFS_V4) | 505 | #if defined(CONFIG_NFS_V3) || defined(CONFIG_NFS_V4) |
556 | static int | 506 | static int |
557 | nfs_need_commit(struct nfs_inode *nfsi) | 507 | nfs_need_commit(struct nfs_inode *nfsi) |
@@ -573,11 +523,17 @@ static int | |||
573 | nfs_scan_commit(struct inode *inode, struct list_head *dst, pgoff_t idx_start, unsigned int npages) | 523 | nfs_scan_commit(struct inode *inode, struct list_head *dst, pgoff_t idx_start, unsigned int npages) |
574 | { | 524 | { |
575 | struct nfs_inode *nfsi = NFS_I(inode); | 525 | struct nfs_inode *nfsi = NFS_I(inode); |
526 | int ret; | ||
576 | 527 | ||
577 | if (!nfs_need_commit(nfsi)) | 528 | if (!nfs_need_commit(nfsi)) |
578 | return 0; | 529 | return 0; |
579 | 530 | ||
580 | return nfs_scan_list(nfsi, dst, idx_start, npages, NFS_PAGE_TAG_COMMIT); | 531 | ret = nfs_scan_list(nfsi, dst, idx_start, npages, NFS_PAGE_TAG_COMMIT); |
532 | if (ret > 0) | ||
533 | nfsi->ncommit -= ret; | ||
534 | if (nfs_need_commit(NFS_I(inode))) | ||
535 | __mark_inode_dirty(inode, I_DIRTY_DATASYNC); | ||
536 | return ret; | ||
581 | } | 537 | } |
582 | #else | 538 | #else |
583 | static inline int nfs_need_commit(struct nfs_inode *nfsi) | 539 | static inline int nfs_need_commit(struct nfs_inode *nfsi) |
@@ -642,9 +598,10 @@ static struct nfs_page *nfs_try_to_update_request(struct inode *inode, | |||
642 | spin_lock(&inode->i_lock); | 598 | spin_lock(&inode->i_lock); |
643 | } | 599 | } |
644 | 600 | ||
645 | if (nfs_clear_request_commit(req)) | 601 | if (nfs_clear_request_commit(req) && |
646 | radix_tree_tag_clear(&NFS_I(inode)->nfs_page_tree, | 602 | radix_tree_tag_clear(&NFS_I(inode)->nfs_page_tree, |
647 | req->wb_index, NFS_PAGE_TAG_COMMIT); | 603 | req->wb_index, NFS_PAGE_TAG_COMMIT) != NULL) |
604 | NFS_I(inode)->ncommit--; | ||
648 | 605 | ||
649 | /* Okay, the request matches. Update the region */ | 606 | /* Okay, the request matches. Update the region */ |
650 | if (offset < req->wb_offset) { | 607 | if (offset < req->wb_offset) { |
@@ -1391,7 +1348,7 @@ static const struct rpc_call_ops nfs_commit_ops = { | |||
1391 | .rpc_release = nfs_commit_release, | 1348 | .rpc_release = nfs_commit_release, |
1392 | }; | 1349 | }; |
1393 | 1350 | ||
1394 | int nfs_commit_inode(struct inode *inode, int how) | 1351 | static int nfs_commit_inode(struct inode *inode, int how) |
1395 | { | 1352 | { |
1396 | LIST_HEAD(head); | 1353 | LIST_HEAD(head); |
1397 | int res; | 1354 | int res; |
@@ -1406,92 +1363,51 @@ int nfs_commit_inode(struct inode *inode, int how) | |||
1406 | } | 1363 | } |
1407 | return res; | 1364 | return res; |
1408 | } | 1365 | } |
1409 | #else | ||
1410 | static inline int nfs_commit_list(struct inode *inode, struct list_head *head, int how) | ||
1411 | { | ||
1412 | return 0; | ||
1413 | } | ||
1414 | #endif | ||
1415 | 1366 | ||
1416 | long nfs_sync_mapping_wait(struct address_space *mapping, struct writeback_control *wbc, int how) | 1367 | static int nfs_commit_unstable_pages(struct inode *inode, struct writeback_control *wbc) |
1417 | { | 1368 | { |
1418 | struct inode *inode = mapping->host; | 1369 | struct nfs_inode *nfsi = NFS_I(inode); |
1419 | pgoff_t idx_start, idx_end; | 1370 | int flags = FLUSH_SYNC; |
1420 | unsigned int npages = 0; | 1371 | int ret = 0; |
1421 | LIST_HEAD(head); | 1372 | |
1422 | int nocommit = how & FLUSH_NOCOMMIT; | 1373 | /* Don't commit yet if this is a non-blocking flush and there are |
1423 | long pages, ret; | 1374 | * lots of outstanding writes for this mapping. |
1424 | 1375 | */ | |
1425 | /* FIXME */ | 1376 | if (wbc->sync_mode == WB_SYNC_NONE && |
1426 | if (wbc->range_cyclic) | 1377 | nfsi->ncommit <= (nfsi->npages >> 1)) |
1427 | idx_start = 0; | 1378 | goto out_mark_dirty; |
1428 | else { | 1379 | |
1429 | idx_start = wbc->range_start >> PAGE_CACHE_SHIFT; | 1380 | if (wbc->nonblocking || wbc->for_background) |
1430 | idx_end = wbc->range_end >> PAGE_CACHE_SHIFT; | 1381 | flags = 0; |
1431 | if (idx_end > idx_start) { | 1382 | ret = nfs_commit_inode(inode, flags); |
1432 | pgoff_t l_npages = 1 + idx_end - idx_start; | 1383 | if (ret >= 0) { |
1433 | npages = l_npages; | 1384 | if (wbc->sync_mode == WB_SYNC_NONE) { |
1434 | if (sizeof(npages) != sizeof(l_npages) && | 1385 | if (ret < wbc->nr_to_write) |
1435 | (pgoff_t)npages != l_npages) | 1386 | wbc->nr_to_write -= ret; |
1436 | npages = 0; | 1387 | else |
1388 | wbc->nr_to_write = 0; | ||
1437 | } | 1389 | } |
1390 | return 0; | ||
1438 | } | 1391 | } |
1439 | how &= ~FLUSH_NOCOMMIT; | 1392 | out_mark_dirty: |
1440 | spin_lock(&inode->i_lock); | 1393 | __mark_inode_dirty(inode, I_DIRTY_DATASYNC); |
1441 | do { | ||
1442 | ret = nfs_wait_on_requests_locked(inode, idx_start, npages); | ||
1443 | if (ret != 0) | ||
1444 | continue; | ||
1445 | if (nocommit) | ||
1446 | break; | ||
1447 | pages = nfs_scan_commit(inode, &head, idx_start, npages); | ||
1448 | if (pages == 0) | ||
1449 | break; | ||
1450 | if (how & FLUSH_INVALIDATE) { | ||
1451 | spin_unlock(&inode->i_lock); | ||
1452 | nfs_cancel_commit_list(&head); | ||
1453 | ret = pages; | ||
1454 | spin_lock(&inode->i_lock); | ||
1455 | continue; | ||
1456 | } | ||
1457 | pages += nfs_scan_commit(inode, &head, 0, 0); | ||
1458 | spin_unlock(&inode->i_lock); | ||
1459 | ret = nfs_commit_list(inode, &head, how); | ||
1460 | spin_lock(&inode->i_lock); | ||
1461 | |||
1462 | } while (ret >= 0); | ||
1463 | spin_unlock(&inode->i_lock); | ||
1464 | return ret; | 1394 | return ret; |
1465 | } | 1395 | } |
1466 | 1396 | #else | |
1467 | static int __nfs_write_mapping(struct address_space *mapping, struct writeback_control *wbc, int how) | 1397 | static int nfs_commit_inode(struct inode *inode, int how) |
1468 | { | 1398 | { |
1469 | int ret; | ||
1470 | |||
1471 | ret = nfs_writepages(mapping, wbc); | ||
1472 | if (ret < 0) | ||
1473 | goto out; | ||
1474 | ret = nfs_sync_mapping_wait(mapping, wbc, how); | ||
1475 | if (ret < 0) | ||
1476 | goto out; | ||
1477 | return 0; | 1399 | return 0; |
1478 | out: | ||
1479 | __mark_inode_dirty(mapping->host, I_DIRTY_PAGES); | ||
1480 | return ret; | ||
1481 | } | 1400 | } |
1482 | 1401 | ||
1483 | /* Two pass sync: first using WB_SYNC_NONE, then WB_SYNC_ALL */ | 1402 | static int nfs_commit_unstable_pages(struct inode *inode, struct writeback_control *wbc) |
1484 | static int nfs_write_mapping(struct address_space *mapping, int how) | ||
1485 | { | 1403 | { |
1486 | struct writeback_control wbc = { | 1404 | return 0; |
1487 | .bdi = mapping->backing_dev_info, | 1405 | } |
1488 | .sync_mode = WB_SYNC_ALL, | 1406 | #endif |
1489 | .nr_to_write = LONG_MAX, | ||
1490 | .range_start = 0, | ||
1491 | .range_end = LLONG_MAX, | ||
1492 | }; | ||
1493 | 1407 | ||
1494 | return __nfs_write_mapping(mapping, &wbc, how); | 1408 | int nfs_write_inode(struct inode *inode, struct writeback_control *wbc) |
1409 | { | ||
1410 | return nfs_commit_unstable_pages(inode, wbc); | ||
1495 | } | 1411 | } |
1496 | 1412 | ||
1497 | /* | 1413 | /* |
@@ -1499,37 +1415,26 @@ static int nfs_write_mapping(struct address_space *mapping, int how) | |||
1499 | */ | 1415 | */ |
1500 | int nfs_wb_all(struct inode *inode) | 1416 | int nfs_wb_all(struct inode *inode) |
1501 | { | 1417 | { |
1502 | return nfs_write_mapping(inode->i_mapping, 0); | 1418 | struct writeback_control wbc = { |
1503 | } | 1419 | .sync_mode = WB_SYNC_ALL, |
1420 | .nr_to_write = LONG_MAX, | ||
1421 | .range_start = 0, | ||
1422 | .range_end = LLONG_MAX, | ||
1423 | }; | ||
1504 | 1424 | ||
1505 | int nfs_wb_nocommit(struct inode *inode) | 1425 | return sync_inode(inode, &wbc); |
1506 | { | ||
1507 | return nfs_write_mapping(inode->i_mapping, FLUSH_NOCOMMIT); | ||
1508 | } | 1426 | } |
1509 | 1427 | ||
1510 | int nfs_wb_page_cancel(struct inode *inode, struct page *page) | 1428 | int nfs_wb_page_cancel(struct inode *inode, struct page *page) |
1511 | { | 1429 | { |
1512 | struct nfs_page *req; | 1430 | struct nfs_page *req; |
1513 | loff_t range_start = page_offset(page); | ||
1514 | loff_t range_end = range_start + (loff_t)(PAGE_CACHE_SIZE - 1); | ||
1515 | struct writeback_control wbc = { | ||
1516 | .bdi = page->mapping->backing_dev_info, | ||
1517 | .sync_mode = WB_SYNC_ALL, | ||
1518 | .nr_to_write = LONG_MAX, | ||
1519 | .range_start = range_start, | ||
1520 | .range_end = range_end, | ||
1521 | }; | ||
1522 | int ret = 0; | 1431 | int ret = 0; |
1523 | 1432 | ||
1524 | BUG_ON(!PageLocked(page)); | 1433 | BUG_ON(!PageLocked(page)); |
1525 | for (;;) { | 1434 | for (;;) { |
1526 | req = nfs_page_find_request(page); | 1435 | req = nfs_page_find_request(page); |
1527 | if (req == NULL) | 1436 | if (req == NULL) |
1528 | goto out; | ||
1529 | if (test_bit(PG_CLEAN, &req->wb_flags)) { | ||
1530 | nfs_release_request(req); | ||
1531 | break; | 1437 | break; |
1532 | } | ||
1533 | if (nfs_lock_request_dontget(req)) { | 1438 | if (nfs_lock_request_dontget(req)) { |
1534 | nfs_inode_remove_request(req); | 1439 | nfs_inode_remove_request(req); |
1535 | /* | 1440 | /* |
@@ -1543,54 +1448,54 @@ int nfs_wb_page_cancel(struct inode *inode, struct page *page) | |||
1543 | ret = nfs_wait_on_request(req); | 1448 | ret = nfs_wait_on_request(req); |
1544 | nfs_release_request(req); | 1449 | nfs_release_request(req); |
1545 | if (ret < 0) | 1450 | if (ret < 0) |
1546 | goto out; | 1451 | break; |
1547 | } | 1452 | } |
1548 | if (!PagePrivate(page)) | ||
1549 | return 0; | ||
1550 | ret = nfs_sync_mapping_wait(page->mapping, &wbc, FLUSH_INVALIDATE); | ||
1551 | out: | ||
1552 | return ret; | 1453 | return ret; |
1553 | } | 1454 | } |
1554 | 1455 | ||
1555 | static int nfs_wb_page_priority(struct inode *inode, struct page *page, | 1456 | /* |
1556 | int how) | 1457 | * Write back all requests on one page - we do this before reading it. |
1458 | */ | ||
1459 | int nfs_wb_page(struct inode *inode, struct page *page) | ||
1557 | { | 1460 | { |
1558 | loff_t range_start = page_offset(page); | 1461 | loff_t range_start = page_offset(page); |
1559 | loff_t range_end = range_start + (loff_t)(PAGE_CACHE_SIZE - 1); | 1462 | loff_t range_end = range_start + (loff_t)(PAGE_CACHE_SIZE - 1); |
1560 | struct writeback_control wbc = { | 1463 | struct writeback_control wbc = { |
1561 | .bdi = page->mapping->backing_dev_info, | ||
1562 | .sync_mode = WB_SYNC_ALL, | 1464 | .sync_mode = WB_SYNC_ALL, |
1563 | .nr_to_write = LONG_MAX, | 1465 | .nr_to_write = 0, |
1564 | .range_start = range_start, | 1466 | .range_start = range_start, |
1565 | .range_end = range_end, | 1467 | .range_end = range_end, |
1566 | }; | 1468 | }; |
1469 | struct nfs_page *req; | ||
1470 | int need_commit; | ||
1567 | int ret; | 1471 | int ret; |
1568 | 1472 | ||
1569 | do { | 1473 | while(PagePrivate(page)) { |
1570 | if (clear_page_dirty_for_io(page)) { | 1474 | if (clear_page_dirty_for_io(page)) { |
1571 | ret = nfs_writepage_locked(page, &wbc); | 1475 | ret = nfs_writepage_locked(page, &wbc); |
1572 | if (ret < 0) | 1476 | if (ret < 0) |
1573 | goto out_error; | 1477 | goto out_error; |
1574 | } else if (!PagePrivate(page)) | 1478 | } |
1479 | req = nfs_find_and_lock_request(page); | ||
1480 | if (!req) | ||
1575 | break; | 1481 | break; |
1576 | ret = nfs_sync_mapping_wait(page->mapping, &wbc, how); | 1482 | if (IS_ERR(req)) { |
1577 | if (ret < 0) | 1483 | ret = PTR_ERR(req); |
1578 | goto out_error; | 1484 | goto out_error; |
1579 | } while (PagePrivate(page)); | 1485 | } |
1486 | need_commit = test_bit(PG_CLEAN, &req->wb_flags); | ||
1487 | nfs_clear_page_tag_locked(req); | ||
1488 | if (need_commit) { | ||
1489 | ret = nfs_commit_inode(inode, FLUSH_SYNC); | ||
1490 | if (ret < 0) | ||
1491 | goto out_error; | ||
1492 | } | ||
1493 | } | ||
1580 | return 0; | 1494 | return 0; |
1581 | out_error: | 1495 | out_error: |
1582 | __mark_inode_dirty(inode, I_DIRTY_PAGES); | ||
1583 | return ret; | 1496 | return ret; |
1584 | } | 1497 | } |
1585 | 1498 | ||
1586 | /* | ||
1587 | * Write back all requests on one page - we do this before reading it. | ||
1588 | */ | ||
1589 | int nfs_wb_page(struct inode *inode, struct page* page) | ||
1590 | { | ||
1591 | return nfs_wb_page_priority(inode, page, FLUSH_STABLE); | ||
1592 | } | ||
1593 | |||
1594 | #ifdef CONFIG_MIGRATION | 1499 | #ifdef CONFIG_MIGRATION |
1595 | int nfs_migrate_page(struct address_space *mapping, struct page *newpage, | 1500 | int nfs_migrate_page(struct address_space *mapping, struct page *newpage, |
1596 | struct page *page) | 1501 | struct page *page) |
@@ -1598,8 +1503,7 @@ int nfs_migrate_page(struct address_space *mapping, struct page *newpage, | |||
1598 | struct nfs_page *req; | 1503 | struct nfs_page *req; |
1599 | int ret; | 1504 | int ret; |
1600 | 1505 | ||
1601 | if (PageFsCache(page)) | 1506 | nfs_fscache_release_page(page, GFP_KERNEL); |
1602 | nfs_fscache_release_page(page, GFP_KERNEL); | ||
1603 | 1507 | ||
1604 | req = nfs_find_and_lock_request(page); | 1508 | req = nfs_find_and_lock_request(page); |
1605 | ret = PTR_ERR(req); | 1509 | ret = PTR_ERR(req); |