diff options
Diffstat (limited to 'fs/nfs/read.c')
-rw-r--r-- | fs/nfs/read.c | 414 |
1 files changed, 78 insertions, 336 deletions
diff --git a/fs/nfs/read.c b/fs/nfs/read.c index 411aedda14bb..e818a475ca64 100644 --- a/fs/nfs/read.c +++ b/fs/nfs/read.c | |||
@@ -24,85 +24,24 @@ | |||
24 | #include "internal.h" | 24 | #include "internal.h" |
25 | #include "iostat.h" | 25 | #include "iostat.h" |
26 | #include "fscache.h" | 26 | #include "fscache.h" |
27 | #include "pnfs.h" | ||
27 | 28 | ||
28 | #define NFSDBG_FACILITY NFSDBG_PAGECACHE | 29 | #define NFSDBG_FACILITY NFSDBG_PAGECACHE |
29 | 30 | ||
30 | static const struct nfs_pageio_ops nfs_pageio_read_ops; | ||
31 | static const struct rpc_call_ops nfs_read_common_ops; | ||
32 | static const struct nfs_pgio_completion_ops nfs_async_read_completion_ops; | 31 | static const struct nfs_pgio_completion_ops nfs_async_read_completion_ops; |
32 | static const struct nfs_rw_ops nfs_rw_read_ops; | ||
33 | 33 | ||
34 | static struct kmem_cache *nfs_rdata_cachep; | 34 | static struct kmem_cache *nfs_rdata_cachep; |
35 | 35 | ||
36 | struct nfs_read_header *nfs_readhdr_alloc(void) | 36 | static struct nfs_rw_header *nfs_readhdr_alloc(void) |
37 | { | 37 | { |
38 | struct nfs_read_header *rhdr; | 38 | return kmem_cache_zalloc(nfs_rdata_cachep, GFP_KERNEL); |
39 | |||
40 | rhdr = kmem_cache_zalloc(nfs_rdata_cachep, GFP_KERNEL); | ||
41 | if (rhdr) { | ||
42 | struct nfs_pgio_header *hdr = &rhdr->header; | ||
43 | |||
44 | INIT_LIST_HEAD(&hdr->pages); | ||
45 | INIT_LIST_HEAD(&hdr->rpc_list); | ||
46 | spin_lock_init(&hdr->lock); | ||
47 | atomic_set(&hdr->refcnt, 0); | ||
48 | } | ||
49 | return rhdr; | ||
50 | } | 39 | } |
51 | EXPORT_SYMBOL_GPL(nfs_readhdr_alloc); | ||
52 | 40 | ||
53 | static struct nfs_read_data *nfs_readdata_alloc(struct nfs_pgio_header *hdr, | 41 | static void nfs_readhdr_free(struct nfs_rw_header *rhdr) |
54 | unsigned int pagecount) | ||
55 | { | 42 | { |
56 | struct nfs_read_data *data, *prealloc; | ||
57 | |||
58 | prealloc = &container_of(hdr, struct nfs_read_header, header)->rpc_data; | ||
59 | if (prealloc->header == NULL) | ||
60 | data = prealloc; | ||
61 | else | ||
62 | data = kzalloc(sizeof(*data), GFP_KERNEL); | ||
63 | if (!data) | ||
64 | goto out; | ||
65 | |||
66 | if (nfs_pgarray_set(&data->pages, pagecount)) { | ||
67 | data->header = hdr; | ||
68 | atomic_inc(&hdr->refcnt); | ||
69 | } else { | ||
70 | if (data != prealloc) | ||
71 | kfree(data); | ||
72 | data = NULL; | ||
73 | } | ||
74 | out: | ||
75 | return data; | ||
76 | } | ||
77 | |||
78 | void nfs_readhdr_free(struct nfs_pgio_header *hdr) | ||
79 | { | ||
80 | struct nfs_read_header *rhdr = container_of(hdr, struct nfs_read_header, header); | ||
81 | |||
82 | kmem_cache_free(nfs_rdata_cachep, rhdr); | 43 | kmem_cache_free(nfs_rdata_cachep, rhdr); |
83 | } | 44 | } |
84 | EXPORT_SYMBOL_GPL(nfs_readhdr_free); | ||
85 | |||
86 | void nfs_readdata_release(struct nfs_read_data *rdata) | ||
87 | { | ||
88 | struct nfs_pgio_header *hdr = rdata->header; | ||
89 | struct nfs_read_header *read_header = container_of(hdr, struct nfs_read_header, header); | ||
90 | |||
91 | put_nfs_open_context(rdata->args.context); | ||
92 | if (rdata->pages.pagevec != rdata->pages.page_array) | ||
93 | kfree(rdata->pages.pagevec); | ||
94 | if (rdata == &read_header->rpc_data) { | ||
95 | rdata->header = NULL; | ||
96 | rdata = NULL; | ||
97 | } | ||
98 | if (atomic_dec_and_test(&hdr->refcnt)) | ||
99 | hdr->completion_ops->completion(hdr); | ||
100 | /* Note: we only free the rpc_task after callbacks are done. | ||
101 | * See the comment in rpc_free_task() for why | ||
102 | */ | ||
103 | kfree(rdata); | ||
104 | } | ||
105 | EXPORT_SYMBOL_GPL(nfs_readdata_release); | ||
106 | 45 | ||
107 | static | 46 | static |
108 | int nfs_return_empty_page(struct page *page) | 47 | int nfs_return_empty_page(struct page *page) |
@@ -114,17 +53,24 @@ int nfs_return_empty_page(struct page *page) | |||
114 | } | 53 | } |
115 | 54 | ||
116 | void nfs_pageio_init_read(struct nfs_pageio_descriptor *pgio, | 55 | void nfs_pageio_init_read(struct nfs_pageio_descriptor *pgio, |
117 | struct inode *inode, | 56 | struct inode *inode, bool force_mds, |
118 | const struct nfs_pgio_completion_ops *compl_ops) | 57 | const struct nfs_pgio_completion_ops *compl_ops) |
119 | { | 58 | { |
120 | nfs_pageio_init(pgio, inode, &nfs_pageio_read_ops, compl_ops, | 59 | struct nfs_server *server = NFS_SERVER(inode); |
121 | NFS_SERVER(inode)->rsize, 0); | 60 | const struct nfs_pageio_ops *pg_ops = &nfs_pgio_rw_ops; |
61 | |||
62 | #ifdef CONFIG_NFS_V4_1 | ||
63 | if (server->pnfs_curr_ld && !force_mds) | ||
64 | pg_ops = server->pnfs_curr_ld->pg_read_ops; | ||
65 | #endif | ||
66 | nfs_pageio_init(pgio, inode, pg_ops, compl_ops, &nfs_rw_read_ops, | ||
67 | server->rsize, 0); | ||
122 | } | 68 | } |
123 | EXPORT_SYMBOL_GPL(nfs_pageio_init_read); | 69 | EXPORT_SYMBOL_GPL(nfs_pageio_init_read); |
124 | 70 | ||
125 | void nfs_pageio_reset_read_mds(struct nfs_pageio_descriptor *pgio) | 71 | void nfs_pageio_reset_read_mds(struct nfs_pageio_descriptor *pgio) |
126 | { | 72 | { |
127 | pgio->pg_ops = &nfs_pageio_read_ops; | 73 | pgio->pg_ops = &nfs_pgio_rw_ops; |
128 | pgio->pg_bsize = NFS_SERVER(pgio->pg_inode)->rsize; | 74 | pgio->pg_bsize = NFS_SERVER(pgio->pg_inode)->rsize; |
129 | } | 75 | } |
130 | EXPORT_SYMBOL_GPL(nfs_pageio_reset_read_mds); | 76 | EXPORT_SYMBOL_GPL(nfs_pageio_reset_read_mds); |
@@ -139,7 +85,7 @@ int nfs_readpage_async(struct nfs_open_context *ctx, struct inode *inode, | |||
139 | len = nfs_page_length(page); | 85 | len = nfs_page_length(page); |
140 | if (len == 0) | 86 | if (len == 0) |
141 | return nfs_return_empty_page(page); | 87 | return nfs_return_empty_page(page); |
142 | new = nfs_create_request(ctx, inode, page, 0, len); | 88 | new = nfs_create_request(ctx, page, NULL, 0, len); |
143 | if (IS_ERR(new)) { | 89 | if (IS_ERR(new)) { |
144 | unlock_page(page); | 90 | unlock_page(page); |
145 | return PTR_ERR(new); | 91 | return PTR_ERR(new); |
@@ -147,7 +93,8 @@ int nfs_readpage_async(struct nfs_open_context *ctx, struct inode *inode, | |||
147 | if (len < PAGE_CACHE_SIZE) | 93 | if (len < PAGE_CACHE_SIZE) |
148 | zero_user_segment(page, len, PAGE_CACHE_SIZE); | 94 | zero_user_segment(page, len, PAGE_CACHE_SIZE); |
149 | 95 | ||
150 | NFS_PROTO(inode)->read_pageio_init(&pgio, inode, &nfs_async_read_completion_ops); | 96 | nfs_pageio_init_read(&pgio, inode, false, |
97 | &nfs_async_read_completion_ops); | ||
151 | nfs_pageio_add_request(&pgio, new); | 98 | nfs_pageio_add_request(&pgio, new); |
152 | nfs_pageio_complete(&pgio); | 99 | nfs_pageio_complete(&pgio); |
153 | NFS_I(inode)->read_io += pgio.pg_bytes_written; | 100 | NFS_I(inode)->read_io += pgio.pg_bytes_written; |
@@ -158,10 +105,16 @@ static void nfs_readpage_release(struct nfs_page *req) | |||
158 | { | 105 | { |
159 | struct inode *d_inode = req->wb_context->dentry->d_inode; | 106 | struct inode *d_inode = req->wb_context->dentry->d_inode; |
160 | 107 | ||
161 | if (PageUptodate(req->wb_page)) | 108 | dprintk("NFS: read done (%s/%llu %d@%lld)\n", d_inode->i_sb->s_id, |
162 | nfs_readpage_to_fscache(d_inode, req->wb_page, 0); | 109 | (unsigned long long)NFS_FILEID(d_inode), req->wb_bytes, |
110 | (long long)req_offset(req)); | ||
163 | 111 | ||
164 | unlock_page(req->wb_page); | 112 | if (nfs_page_group_sync_on_bit(req, PG_UNLOCKPAGE)) { |
113 | if (PageUptodate(req->wb_page)) | ||
114 | nfs_readpage_to_fscache(d_inode, req->wb_page, 0); | ||
115 | |||
116 | unlock_page(req->wb_page); | ||
117 | } | ||
165 | 118 | ||
166 | dprintk("NFS: read done (%s/%Lu %d@%Ld)\n", | 119 | dprintk("NFS: read done (%s/%Lu %d@%Ld)\n", |
167 | req->wb_context->dentry->d_inode->i_sb->s_id, | 120 | req->wb_context->dentry->d_inode->i_sb->s_id, |
@@ -171,7 +124,12 @@ static void nfs_readpage_release(struct nfs_page *req) | |||
171 | nfs_release_request(req); | 124 | nfs_release_request(req); |
172 | } | 125 | } |
173 | 126 | ||
174 | /* Note io was page aligned */ | 127 | static void nfs_page_group_set_uptodate(struct nfs_page *req) |
128 | { | ||
129 | if (nfs_page_group_sync_on_bit(req, PG_UPTODATE)) | ||
130 | SetPageUptodate(req->wb_page); | ||
131 | } | ||
132 | |||
175 | static void nfs_read_completion(struct nfs_pgio_header *hdr) | 133 | static void nfs_read_completion(struct nfs_pgio_header *hdr) |
176 | { | 134 | { |
177 | unsigned long bytes = 0; | 135 | unsigned long bytes = 0; |
@@ -181,21 +139,32 @@ static void nfs_read_completion(struct nfs_pgio_header *hdr) | |||
181 | while (!list_empty(&hdr->pages)) { | 139 | while (!list_empty(&hdr->pages)) { |
182 | struct nfs_page *req = nfs_list_entry(hdr->pages.next); | 140 | struct nfs_page *req = nfs_list_entry(hdr->pages.next); |
183 | struct page *page = req->wb_page; | 141 | struct page *page = req->wb_page; |
142 | unsigned long start = req->wb_pgbase; | ||
143 | unsigned long end = req->wb_pgbase + req->wb_bytes; | ||
184 | 144 | ||
185 | if (test_bit(NFS_IOHDR_EOF, &hdr->flags)) { | 145 | if (test_bit(NFS_IOHDR_EOF, &hdr->flags)) { |
186 | if (bytes > hdr->good_bytes) | 146 | /* note: regions of the page not covered by a |
187 | zero_user(page, 0, PAGE_SIZE); | 147 | * request are zeroed in nfs_readpage_async / |
188 | else if (hdr->good_bytes - bytes < PAGE_SIZE) | 148 | * readpage_async_filler */ |
189 | zero_user_segment(page, | 149 | if (bytes > hdr->good_bytes) { |
190 | hdr->good_bytes & ~PAGE_MASK, | 150 | /* nothing in this request was good, so zero |
191 | PAGE_SIZE); | 151 | * the full extent of the request */ |
152 | zero_user_segment(page, start, end); | ||
153 | |||
154 | } else if (hdr->good_bytes - bytes < req->wb_bytes) { | ||
155 | /* part of this request has good bytes, but | ||
156 | * not all. zero the bad bytes */ | ||
157 | start += hdr->good_bytes - bytes; | ||
158 | WARN_ON(start < req->wb_pgbase); | ||
159 | zero_user_segment(page, start, end); | ||
160 | } | ||
192 | } | 161 | } |
193 | bytes += req->wb_bytes; | 162 | bytes += req->wb_bytes; |
194 | if (test_bit(NFS_IOHDR_ERROR, &hdr->flags)) { | 163 | if (test_bit(NFS_IOHDR_ERROR, &hdr->flags)) { |
195 | if (bytes <= hdr->good_bytes) | 164 | if (bytes <= hdr->good_bytes) |
196 | SetPageUptodate(page); | 165 | nfs_page_group_set_uptodate(req); |
197 | } else | 166 | } else |
198 | SetPageUptodate(page); | 167 | nfs_page_group_set_uptodate(req); |
199 | nfs_list_remove_request(req); | 168 | nfs_list_remove_request(req); |
200 | nfs_readpage_release(req); | 169 | nfs_readpage_release(req); |
201 | } | 170 | } |
@@ -203,95 +172,14 @@ out: | |||
203 | hdr->release(hdr); | 172 | hdr->release(hdr); |
204 | } | 173 | } |
205 | 174 | ||
206 | int nfs_initiate_read(struct rpc_clnt *clnt, | 175 | static void nfs_initiate_read(struct nfs_pgio_data *data, struct rpc_message *msg, |
207 | struct nfs_read_data *data, | 176 | struct rpc_task_setup *task_setup_data, int how) |
208 | const struct rpc_call_ops *call_ops, int flags) | ||
209 | { | 177 | { |
210 | struct inode *inode = data->header->inode; | 178 | struct inode *inode = data->header->inode; |
211 | int swap_flags = IS_SWAPFILE(inode) ? NFS_RPC_SWAPFLAGS : 0; | 179 | int swap_flags = IS_SWAPFILE(inode) ? NFS_RPC_SWAPFLAGS : 0; |
212 | struct rpc_task *task; | ||
213 | struct rpc_message msg = { | ||
214 | .rpc_argp = &data->args, | ||
215 | .rpc_resp = &data->res, | ||
216 | .rpc_cred = data->header->cred, | ||
217 | }; | ||
218 | struct rpc_task_setup task_setup_data = { | ||
219 | .task = &data->task, | ||
220 | .rpc_client = clnt, | ||
221 | .rpc_message = &msg, | ||
222 | .callback_ops = call_ops, | ||
223 | .callback_data = data, | ||
224 | .workqueue = nfsiod_workqueue, | ||
225 | .flags = RPC_TASK_ASYNC | swap_flags | flags, | ||
226 | }; | ||
227 | 180 | ||
228 | /* Set up the initial task struct. */ | 181 | task_setup_data->flags |= swap_flags; |
229 | NFS_PROTO(inode)->read_setup(data, &msg); | 182 | NFS_PROTO(inode)->read_setup(data, msg); |
230 | |||
231 | dprintk("NFS: %5u initiated read call (req %s/%llu, %u bytes @ " | ||
232 | "offset %llu)\n", | ||
233 | data->task.tk_pid, | ||
234 | inode->i_sb->s_id, | ||
235 | (unsigned long long)NFS_FILEID(inode), | ||
236 | data->args.count, | ||
237 | (unsigned long long)data->args.offset); | ||
238 | |||
239 | task = rpc_run_task(&task_setup_data); | ||
240 | if (IS_ERR(task)) | ||
241 | return PTR_ERR(task); | ||
242 | rpc_put_task(task); | ||
243 | return 0; | ||
244 | } | ||
245 | EXPORT_SYMBOL_GPL(nfs_initiate_read); | ||
246 | |||
247 | /* | ||
248 | * Set up the NFS read request struct | ||
249 | */ | ||
250 | static void nfs_read_rpcsetup(struct nfs_read_data *data, | ||
251 | unsigned int count, unsigned int offset) | ||
252 | { | ||
253 | struct nfs_page *req = data->header->req; | ||
254 | |||
255 | data->args.fh = NFS_FH(data->header->inode); | ||
256 | data->args.offset = req_offset(req) + offset; | ||
257 | data->args.pgbase = req->wb_pgbase + offset; | ||
258 | data->args.pages = data->pages.pagevec; | ||
259 | data->args.count = count; | ||
260 | data->args.context = get_nfs_open_context(req->wb_context); | ||
261 | data->args.lock_context = req->wb_lock_context; | ||
262 | |||
263 | data->res.fattr = &data->fattr; | ||
264 | data->res.count = count; | ||
265 | data->res.eof = 0; | ||
266 | nfs_fattr_init(&data->fattr); | ||
267 | } | ||
268 | |||
269 | static int nfs_do_read(struct nfs_read_data *data, | ||
270 | const struct rpc_call_ops *call_ops) | ||
271 | { | ||
272 | struct inode *inode = data->header->inode; | ||
273 | |||
274 | return nfs_initiate_read(NFS_CLIENT(inode), data, call_ops, 0); | ||
275 | } | ||
276 | |||
277 | static int | ||
278 | nfs_do_multiple_reads(struct list_head *head, | ||
279 | const struct rpc_call_ops *call_ops) | ||
280 | { | ||
281 | struct nfs_read_data *data; | ||
282 | int ret = 0; | ||
283 | |||
284 | while (!list_empty(head)) { | ||
285 | int ret2; | ||
286 | |||
287 | data = list_first_entry(head, struct nfs_read_data, list); | ||
288 | list_del_init(&data->list); | ||
289 | |||
290 | ret2 = nfs_do_read(data, call_ops); | ||
291 | if (ret == 0) | ||
292 | ret = ret2; | ||
293 | } | ||
294 | return ret; | ||
295 | } | 183 | } |
296 | 184 | ||
297 | static void | 185 | static void |
@@ -311,143 +199,14 @@ static const struct nfs_pgio_completion_ops nfs_async_read_completion_ops = { | |||
311 | .completion = nfs_read_completion, | 199 | .completion = nfs_read_completion, |
312 | }; | 200 | }; |
313 | 201 | ||
314 | static void nfs_pagein_error(struct nfs_pageio_descriptor *desc, | ||
315 | struct nfs_pgio_header *hdr) | ||
316 | { | ||
317 | set_bit(NFS_IOHDR_REDO, &hdr->flags); | ||
318 | while (!list_empty(&hdr->rpc_list)) { | ||
319 | struct nfs_read_data *data = list_first_entry(&hdr->rpc_list, | ||
320 | struct nfs_read_data, list); | ||
321 | list_del(&data->list); | ||
322 | nfs_readdata_release(data); | ||
323 | } | ||
324 | desc->pg_completion_ops->error_cleanup(&desc->pg_list); | ||
325 | } | ||
326 | |||
327 | /* | ||
328 | * Generate multiple requests to fill a single page. | ||
329 | * | ||
330 | * We optimize to reduce the number of read operations on the wire. If we | ||
331 | * detect that we're reading a page, or an area of a page, that is past the | ||
332 | * end of file, we do not generate NFS read operations but just clear the | ||
333 | * parts of the page that would have come back zero from the server anyway. | ||
334 | * | ||
335 | * We rely on the cached value of i_size to make this determination; another | ||
336 | * client can fill pages on the server past our cached end-of-file, but we | ||
337 | * won't see the new data until our attribute cache is updated. This is more | ||
338 | * or less conventional NFS client behavior. | ||
339 | */ | ||
340 | static int nfs_pagein_multi(struct nfs_pageio_descriptor *desc, | ||
341 | struct nfs_pgio_header *hdr) | ||
342 | { | ||
343 | struct nfs_page *req = hdr->req; | ||
344 | struct page *page = req->wb_page; | ||
345 | struct nfs_read_data *data; | ||
346 | size_t rsize = desc->pg_bsize, nbytes; | ||
347 | unsigned int offset; | ||
348 | |||
349 | offset = 0; | ||
350 | nbytes = desc->pg_count; | ||
351 | do { | ||
352 | size_t len = min(nbytes,rsize); | ||
353 | |||
354 | data = nfs_readdata_alloc(hdr, 1); | ||
355 | if (!data) { | ||
356 | nfs_pagein_error(desc, hdr); | ||
357 | return -ENOMEM; | ||
358 | } | ||
359 | data->pages.pagevec[0] = page; | ||
360 | nfs_read_rpcsetup(data, len, offset); | ||
361 | list_add(&data->list, &hdr->rpc_list); | ||
362 | nbytes -= len; | ||
363 | offset += len; | ||
364 | } while (nbytes != 0); | ||
365 | |||
366 | nfs_list_remove_request(req); | ||
367 | nfs_list_add_request(req, &hdr->pages); | ||
368 | desc->pg_rpc_callops = &nfs_read_common_ops; | ||
369 | return 0; | ||
370 | } | ||
371 | |||
372 | static int nfs_pagein_one(struct nfs_pageio_descriptor *desc, | ||
373 | struct nfs_pgio_header *hdr) | ||
374 | { | ||
375 | struct nfs_page *req; | ||
376 | struct page **pages; | ||
377 | struct nfs_read_data *data; | ||
378 | struct list_head *head = &desc->pg_list; | ||
379 | |||
380 | data = nfs_readdata_alloc(hdr, nfs_page_array_len(desc->pg_base, | ||
381 | desc->pg_count)); | ||
382 | if (!data) { | ||
383 | nfs_pagein_error(desc, hdr); | ||
384 | return -ENOMEM; | ||
385 | } | ||
386 | |||
387 | pages = data->pages.pagevec; | ||
388 | while (!list_empty(head)) { | ||
389 | req = nfs_list_entry(head->next); | ||
390 | nfs_list_remove_request(req); | ||
391 | nfs_list_add_request(req, &hdr->pages); | ||
392 | *pages++ = req->wb_page; | ||
393 | } | ||
394 | |||
395 | nfs_read_rpcsetup(data, desc->pg_count, 0); | ||
396 | list_add(&data->list, &hdr->rpc_list); | ||
397 | desc->pg_rpc_callops = &nfs_read_common_ops; | ||
398 | return 0; | ||
399 | } | ||
400 | |||
401 | int nfs_generic_pagein(struct nfs_pageio_descriptor *desc, | ||
402 | struct nfs_pgio_header *hdr) | ||
403 | { | ||
404 | if (desc->pg_bsize < PAGE_CACHE_SIZE) | ||
405 | return nfs_pagein_multi(desc, hdr); | ||
406 | return nfs_pagein_one(desc, hdr); | ||
407 | } | ||
408 | EXPORT_SYMBOL_GPL(nfs_generic_pagein); | ||
409 | |||
410 | static int nfs_generic_pg_readpages(struct nfs_pageio_descriptor *desc) | ||
411 | { | ||
412 | struct nfs_read_header *rhdr; | ||
413 | struct nfs_pgio_header *hdr; | ||
414 | int ret; | ||
415 | |||
416 | rhdr = nfs_readhdr_alloc(); | ||
417 | if (!rhdr) { | ||
418 | desc->pg_completion_ops->error_cleanup(&desc->pg_list); | ||
419 | return -ENOMEM; | ||
420 | } | ||
421 | hdr = &rhdr->header; | ||
422 | nfs_pgheader_init(desc, hdr, nfs_readhdr_free); | ||
423 | atomic_inc(&hdr->refcnt); | ||
424 | ret = nfs_generic_pagein(desc, hdr); | ||
425 | if (ret == 0) | ||
426 | ret = nfs_do_multiple_reads(&hdr->rpc_list, | ||
427 | desc->pg_rpc_callops); | ||
428 | if (atomic_dec_and_test(&hdr->refcnt)) | ||
429 | hdr->completion_ops->completion(hdr); | ||
430 | return ret; | ||
431 | } | ||
432 | |||
433 | static const struct nfs_pageio_ops nfs_pageio_read_ops = { | ||
434 | .pg_test = nfs_generic_pg_test, | ||
435 | .pg_doio = nfs_generic_pg_readpages, | ||
436 | }; | ||
437 | |||
438 | /* | 202 | /* |
439 | * This is the callback from RPC telling us whether a reply was | 203 | * This is the callback from RPC telling us whether a reply was |
440 | * received or some error occurred (timeout or socket shutdown). | 204 | * received or some error occurred (timeout or socket shutdown). |
441 | */ | 205 | */ |
442 | int nfs_readpage_result(struct rpc_task *task, struct nfs_read_data *data) | 206 | static int nfs_readpage_done(struct rpc_task *task, struct nfs_pgio_data *data, |
207 | struct inode *inode) | ||
443 | { | 208 | { |
444 | struct inode *inode = data->header->inode; | 209 | int status = NFS_PROTO(inode)->read_done(task, data); |
445 | int status; | ||
446 | |||
447 | dprintk("NFS: %s: %5u, (status %d)\n", __func__, task->tk_pid, | ||
448 | task->tk_status); | ||
449 | |||
450 | status = NFS_PROTO(inode)->read_done(task, data); | ||
451 | if (status != 0) | 210 | if (status != 0) |
452 | return status; | 211 | return status; |
453 | 212 | ||
@@ -460,10 +219,10 @@ int nfs_readpage_result(struct rpc_task *task, struct nfs_read_data *data) | |||
460 | return 0; | 219 | return 0; |
461 | } | 220 | } |
462 | 221 | ||
463 | static void nfs_readpage_retry(struct rpc_task *task, struct nfs_read_data *data) | 222 | static void nfs_readpage_retry(struct rpc_task *task, struct nfs_pgio_data *data) |
464 | { | 223 | { |
465 | struct nfs_readargs *argp = &data->args; | 224 | struct nfs_pgio_args *argp = &data->args; |
466 | struct nfs_readres *resp = &data->res; | 225 | struct nfs_pgio_res *resp = &data->res; |
467 | 226 | ||
468 | /* This is a short read! */ | 227 | /* This is a short read! */ |
469 | nfs_inc_stats(data->header->inode, NFSIOS_SHORTREAD); | 228 | nfs_inc_stats(data->header->inode, NFSIOS_SHORTREAD); |
@@ -480,17 +239,11 @@ static void nfs_readpage_retry(struct rpc_task *task, struct nfs_read_data *data | |||
480 | rpc_restart_call_prepare(task); | 239 | rpc_restart_call_prepare(task); |
481 | } | 240 | } |
482 | 241 | ||
483 | static void nfs_readpage_result_common(struct rpc_task *task, void *calldata) | 242 | static void nfs_readpage_result(struct rpc_task *task, struct nfs_pgio_data *data) |
484 | { | 243 | { |
485 | struct nfs_read_data *data = calldata; | ||
486 | struct nfs_pgio_header *hdr = data->header; | 244 | struct nfs_pgio_header *hdr = data->header; |
487 | 245 | ||
488 | /* Note the only returns of nfs_readpage_result are 0 and -EAGAIN */ | 246 | if (data->res.eof) { |
489 | if (nfs_readpage_result(task, data) != 0) | ||
490 | return; | ||
491 | if (task->tk_status < 0) | ||
492 | nfs_set_pgio_error(hdr, task->tk_status, data->args.offset); | ||
493 | else if (data->res.eof) { | ||
494 | loff_t bound; | 247 | loff_t bound; |
495 | 248 | ||
496 | bound = data->args.offset + data->res.count; | 249 | bound = data->args.offset + data->res.count; |
@@ -505,26 +258,6 @@ static void nfs_readpage_result_common(struct rpc_task *task, void *calldata) | |||
505 | nfs_readpage_retry(task, data); | 258 | nfs_readpage_retry(task, data); |
506 | } | 259 | } |
507 | 260 | ||
508 | static void nfs_readpage_release_common(void *calldata) | ||
509 | { | ||
510 | nfs_readdata_release(calldata); | ||
511 | } | ||
512 | |||
513 | void nfs_read_prepare(struct rpc_task *task, void *calldata) | ||
514 | { | ||
515 | struct nfs_read_data *data = calldata; | ||
516 | int err; | ||
517 | err = NFS_PROTO(data->header->inode)->read_rpc_prepare(task, data); | ||
518 | if (err) | ||
519 | rpc_exit(task, err); | ||
520 | } | ||
521 | |||
522 | static const struct rpc_call_ops nfs_read_common_ops = { | ||
523 | .rpc_call_prepare = nfs_read_prepare, | ||
524 | .rpc_call_done = nfs_readpage_result_common, | ||
525 | .rpc_release = nfs_readpage_release_common, | ||
526 | }; | ||
527 | |||
528 | /* | 261 | /* |
529 | * Read a page over NFS. | 262 | * Read a page over NFS. |
530 | * We read the page synchronously in the following case: | 263 | * We read the page synchronously in the following case: |
@@ -592,7 +325,6 @@ static int | |||
592 | readpage_async_filler(void *data, struct page *page) | 325 | readpage_async_filler(void *data, struct page *page) |
593 | { | 326 | { |
594 | struct nfs_readdesc *desc = (struct nfs_readdesc *)data; | 327 | struct nfs_readdesc *desc = (struct nfs_readdesc *)data; |
595 | struct inode *inode = page_file_mapping(page)->host; | ||
596 | struct nfs_page *new; | 328 | struct nfs_page *new; |
597 | unsigned int len; | 329 | unsigned int len; |
598 | int error; | 330 | int error; |
@@ -601,7 +333,7 @@ readpage_async_filler(void *data, struct page *page) | |||
601 | if (len == 0) | 333 | if (len == 0) |
602 | return nfs_return_empty_page(page); | 334 | return nfs_return_empty_page(page); |
603 | 335 | ||
604 | new = nfs_create_request(desc->ctx, inode, page, 0, len); | 336 | new = nfs_create_request(desc->ctx, page, NULL, 0, len); |
605 | if (IS_ERR(new)) | 337 | if (IS_ERR(new)) |
606 | goto out_error; | 338 | goto out_error; |
607 | 339 | ||
@@ -654,7 +386,8 @@ int nfs_readpages(struct file *filp, struct address_space *mapping, | |||
654 | if (ret == 0) | 386 | if (ret == 0) |
655 | goto read_complete; /* all pages were read */ | 387 | goto read_complete; /* all pages were read */ |
656 | 388 | ||
657 | NFS_PROTO(inode)->read_pageio_init(&pgio, inode, &nfs_async_read_completion_ops); | 389 | nfs_pageio_init_read(&pgio, inode, false, |
390 | &nfs_async_read_completion_ops); | ||
658 | 391 | ||
659 | ret = read_cache_pages(mapping, pages, readpage_async_filler, &desc); | 392 | ret = read_cache_pages(mapping, pages, readpage_async_filler, &desc); |
660 | 393 | ||
@@ -671,7 +404,7 @@ out: | |||
671 | int __init nfs_init_readpagecache(void) | 404 | int __init nfs_init_readpagecache(void) |
672 | { | 405 | { |
673 | nfs_rdata_cachep = kmem_cache_create("nfs_read_data", | 406 | nfs_rdata_cachep = kmem_cache_create("nfs_read_data", |
674 | sizeof(struct nfs_read_header), | 407 | sizeof(struct nfs_rw_header), |
675 | 0, SLAB_HWCACHE_ALIGN, | 408 | 0, SLAB_HWCACHE_ALIGN, |
676 | NULL); | 409 | NULL); |
677 | if (nfs_rdata_cachep == NULL) | 410 | if (nfs_rdata_cachep == NULL) |
@@ -684,3 +417,12 @@ void nfs_destroy_readpagecache(void) | |||
684 | { | 417 | { |
685 | kmem_cache_destroy(nfs_rdata_cachep); | 418 | kmem_cache_destroy(nfs_rdata_cachep); |
686 | } | 419 | } |
420 | |||
421 | static const struct nfs_rw_ops nfs_rw_read_ops = { | ||
422 | .rw_mode = FMODE_READ, | ||
423 | .rw_alloc_header = nfs_readhdr_alloc, | ||
424 | .rw_free_header = nfs_readhdr_free, | ||
425 | .rw_done = nfs_readpage_done, | ||
426 | .rw_result = nfs_readpage_result, | ||
427 | .rw_initiate = nfs_initiate_read, | ||
428 | }; | ||