diff options
Diffstat (limited to 'fs/9p/mux.c')
-rw-r--r-- | fs/9p/mux.c | 1145 |
1 files changed, 820 insertions, 325 deletions
diff --git a/fs/9p/mux.c b/fs/9p/mux.c index 8835b576f744..945cb368d451 100644 --- a/fs/9p/mux.c +++ b/fs/9p/mux.c | |||
@@ -4,7 +4,7 @@ | |||
4 | * Protocol Multiplexer | 4 | * Protocol Multiplexer |
5 | * | 5 | * |
6 | * Copyright (C) 2004 by Eric Van Hensbergen <ericvh@gmail.com> | 6 | * Copyright (C) 2004 by Eric Van Hensbergen <ericvh@gmail.com> |
7 | * Copyright (C) 2004 by Latchesar Ionkov <lucho@ionkov.net> | 7 | * Copyright (C) 2004-2005 by Latchesar Ionkov <lucho@ionkov.net> |
8 | * | 8 | * |
9 | * This program is free software; you can redistribute it and/or modify | 9 | * This program is free software; you can redistribute it and/or modify |
10 | * it under the terms of the GNU General Public License as published by | 10 | * it under the terms of the GNU General Public License as published by |
@@ -28,448 +28,943 @@ | |||
28 | #include <linux/module.h> | 28 | #include <linux/module.h> |
29 | #include <linux/errno.h> | 29 | #include <linux/errno.h> |
30 | #include <linux/fs.h> | 30 | #include <linux/fs.h> |
31 | #include <linux/poll.h> | ||
31 | #include <linux/kthread.h> | 32 | #include <linux/kthread.h> |
32 | #include <linux/idr.h> | 33 | #include <linux/idr.h> |
33 | 34 | ||
34 | #include "debug.h" | 35 | #include "debug.h" |
35 | #include "v9fs.h" | 36 | #include "v9fs.h" |
36 | #include "9p.h" | 37 | #include "9p.h" |
37 | #include "transport.h" | ||
38 | #include "conv.h" | 38 | #include "conv.h" |
39 | #include "transport.h" | ||
39 | #include "mux.h" | 40 | #include "mux.h" |
40 | 41 | ||
42 | #define ERREQFLUSH 1 | ||
43 | #define SCHED_TIMEOUT 10 | ||
44 | #define MAXPOLLWADDR 2 | ||
45 | |||
46 | enum { | ||
47 | Rworksched = 1, /* read work scheduled or running */ | ||
48 | Rpending = 2, /* can read */ | ||
49 | Wworksched = 4, /* write work scheduled or running */ | ||
50 | Wpending = 8, /* can write */ | ||
51 | }; | ||
52 | |||
53 | struct v9fs_mux_poll_task; | ||
54 | |||
55 | struct v9fs_req { | ||
56 | int tag; | ||
57 | struct v9fs_fcall *tcall; | ||
58 | struct v9fs_fcall *rcall; | ||
59 | int err; | ||
60 | v9fs_mux_req_callback cb; | ||
61 | void *cba; | ||
62 | struct list_head req_list; | ||
63 | }; | ||
64 | |||
65 | struct v9fs_mux_data { | ||
66 | spinlock_t lock; | ||
67 | struct list_head mux_list; | ||
68 | struct v9fs_mux_poll_task *poll_task; | ||
69 | int msize; | ||
70 | unsigned char *extended; | ||
71 | struct v9fs_transport *trans; | ||
72 | struct v9fs_idpool tidpool; | ||
73 | int err; | ||
74 | wait_queue_head_t equeue; | ||
75 | struct list_head req_list; | ||
76 | struct list_head unsent_req_list; | ||
77 | struct v9fs_fcall *rcall; | ||
78 | int rpos; | ||
79 | char *rbuf; | ||
80 | int wpos; | ||
81 | int wsize; | ||
82 | char *wbuf; | ||
83 | wait_queue_t poll_wait[MAXPOLLWADDR]; | ||
84 | wait_queue_head_t *poll_waddr[MAXPOLLWADDR]; | ||
85 | poll_table pt; | ||
86 | struct work_struct rq; | ||
87 | struct work_struct wq; | ||
88 | unsigned long wsched; | ||
89 | }; | ||
90 | |||
91 | struct v9fs_mux_poll_task { | ||
92 | struct task_struct *task; | ||
93 | struct list_head mux_list; | ||
94 | int muxnum; | ||
95 | }; | ||
96 | |||
97 | struct v9fs_mux_rpc { | ||
98 | struct v9fs_mux_data *m; | ||
99 | struct v9fs_req *req; | ||
100 | int err; | ||
101 | struct v9fs_fcall *rcall; | ||
102 | wait_queue_head_t wqueue; | ||
103 | }; | ||
104 | |||
105 | static int v9fs_poll_proc(void *); | ||
106 | static void v9fs_read_work(void *); | ||
107 | static void v9fs_write_work(void *); | ||
108 | static void v9fs_pollwait(struct file *filp, wait_queue_head_t * wait_address, | ||
109 | poll_table * p); | ||
110 | static u16 v9fs_mux_get_tag(struct v9fs_mux_data *); | ||
111 | static void v9fs_mux_put_tag(struct v9fs_mux_data *, u16); | ||
112 | |||
113 | static DECLARE_MUTEX(v9fs_mux_task_lock); | ||
114 | static struct workqueue_struct *v9fs_mux_wq; | ||
115 | |||
116 | static int v9fs_mux_num; | ||
117 | static int v9fs_mux_poll_task_num; | ||
118 | static struct v9fs_mux_poll_task v9fs_mux_poll_tasks[100]; | ||
119 | |||
120 | int v9fs_mux_global_init(void) | ||
121 | { | ||
122 | int i; | ||
123 | |||
124 | for (i = 0; i < ARRAY_SIZE(v9fs_mux_poll_tasks); i++) | ||
125 | v9fs_mux_poll_tasks[i].task = NULL; | ||
126 | |||
127 | v9fs_mux_wq = create_workqueue("v9fs"); | ||
128 | if (!v9fs_mux_wq) | ||
129 | return -ENOMEM; | ||
130 | |||
131 | return 0; | ||
132 | } | ||
133 | |||
134 | void v9fs_mux_global_exit(void) | ||
135 | { | ||
136 | destroy_workqueue(v9fs_mux_wq); | ||
137 | } | ||
138 | |||
41 | /** | 139 | /** |
42 | * dprintcond - print condition of session info | 140 | * v9fs_mux_calc_poll_procs - calculates the number of polling procs |
43 | * @v9ses: session info structure | 141 | * based on the number of mounted v9fs filesystems. |
44 | * @req: RPC request structure | ||
45 | * | 142 | * |
143 | * The current implementation returns sqrt of the number of mounts. | ||
46 | */ | 144 | */ |
145 | inline int v9fs_mux_calc_poll_procs(int muxnum) | ||
146 | { | ||
147 | int n; | ||
148 | |||
149 | if (v9fs_mux_poll_task_num) | ||
150 | n = muxnum / v9fs_mux_poll_task_num + | ||
151 | (muxnum % v9fs_mux_poll_task_num ? 1 : 0); | ||
152 | else | ||
153 | n = 1; | ||
154 | |||
155 | if (n > ARRAY_SIZE(v9fs_mux_poll_tasks)) | ||
156 | n = ARRAY_SIZE(v9fs_mux_poll_tasks); | ||
157 | |||
158 | return n; | ||
159 | } | ||
47 | 160 | ||
48 | static inline int | 161 | static int v9fs_mux_poll_start(struct v9fs_mux_data *m) |
49 | dprintcond(struct v9fs_session_info *v9ses, struct v9fs_rpcreq *req) | ||
50 | { | 162 | { |
51 | dprintk(DEBUG_MUX, "condition: %d, %p\n", v9ses->transport->status, | 163 | int i, n; |
52 | req->rcall); | 164 | struct v9fs_mux_poll_task *vpt, *vptlast; |
165 | struct task_struct *pproc; | ||
166 | |||
167 | dprintk(DEBUG_MUX, "mux %p muxnum %d procnum %d\n", m, v9fs_mux_num, | ||
168 | v9fs_mux_poll_task_num); | ||
169 | up(&v9fs_mux_task_lock); | ||
170 | |||
171 | n = v9fs_mux_calc_poll_procs(v9fs_mux_num + 1); | ||
172 | if (n > v9fs_mux_poll_task_num) { | ||
173 | for (i = 0; i < ARRAY_SIZE(v9fs_mux_poll_tasks); i++) { | ||
174 | if (v9fs_mux_poll_tasks[i].task == NULL) { | ||
175 | vpt = &v9fs_mux_poll_tasks[i]; | ||
176 | dprintk(DEBUG_MUX, "create proc %p\n", vpt); | ||
177 | pproc = kthread_create(v9fs_poll_proc, vpt, | ||
178 | "v9fs-poll"); | ||
179 | |||
180 | if (!IS_ERR(pproc)) { | ||
181 | vpt->task = pproc; | ||
182 | INIT_LIST_HEAD(&vpt->mux_list); | ||
183 | vpt->muxnum = 0; | ||
184 | v9fs_mux_poll_task_num++; | ||
185 | wake_up_process(vpt->task); | ||
186 | } | ||
187 | break; | ||
188 | } | ||
189 | } | ||
190 | |||
191 | if (i >= ARRAY_SIZE(v9fs_mux_poll_tasks)) | ||
192 | dprintk(DEBUG_ERROR, "warning: no free poll slots\n"); | ||
193 | } | ||
194 | |||
195 | n = (v9fs_mux_num + 1) / v9fs_mux_poll_task_num + | ||
196 | ((v9fs_mux_num + 1) % v9fs_mux_poll_task_num ? 1 : 0); | ||
197 | |||
198 | vptlast = NULL; | ||
199 | for (i = 0; i < ARRAY_SIZE(v9fs_mux_poll_tasks); i++) { | ||
200 | vpt = &v9fs_mux_poll_tasks[i]; | ||
201 | if (vpt->task != NULL) { | ||
202 | vptlast = vpt; | ||
203 | if (vpt->muxnum < n) { | ||
204 | dprintk(DEBUG_MUX, "put in proc %d\n", i); | ||
205 | list_add(&m->mux_list, &vpt->mux_list); | ||
206 | vpt->muxnum++; | ||
207 | m->poll_task = vpt; | ||
208 | memset(&m->poll_waddr, 0, sizeof(m->poll_waddr)); | ||
209 | init_poll_funcptr(&m->pt, v9fs_pollwait); | ||
210 | break; | ||
211 | } | ||
212 | } | ||
213 | } | ||
214 | |||
215 | if (i >= ARRAY_SIZE(v9fs_mux_poll_tasks)) { | ||
216 | if (vptlast == NULL) | ||
217 | return -ENOMEM; | ||
218 | |||
219 | dprintk(DEBUG_MUX, "put in proc %d\n", i); | ||
220 | list_add(&m->mux_list, &vptlast->mux_list); | ||
221 | vptlast->muxnum++; | ||
222 | m->poll_task = vptlast; | ||
223 | memset(&m->poll_waddr, 0, sizeof(m->poll_waddr)); | ||
224 | init_poll_funcptr(&m->pt, v9fs_pollwait); | ||
225 | } | ||
226 | |||
227 | v9fs_mux_num++; | ||
228 | down(&v9fs_mux_task_lock); | ||
229 | |||
53 | return 0; | 230 | return 0; |
54 | } | 231 | } |
55 | 232 | ||
233 | static void v9fs_mux_poll_stop(struct v9fs_mux_data *m) | ||
234 | { | ||
235 | int i; | ||
236 | struct v9fs_mux_poll_task *vpt; | ||
237 | |||
238 | up(&v9fs_mux_task_lock); | ||
239 | vpt = m->poll_task; | ||
240 | list_del(&m->mux_list); | ||
241 | for(i = 0; i < ARRAY_SIZE(m->poll_waddr); i++) { | ||
242 | if (m->poll_waddr[i] != NULL) { | ||
243 | remove_wait_queue(m->poll_waddr[i], &m->poll_wait[i]); | ||
244 | m->poll_waddr[i] = NULL; | ||
245 | } | ||
246 | } | ||
247 | vpt->muxnum--; | ||
248 | if (!vpt->muxnum) { | ||
249 | dprintk(DEBUG_MUX, "destroy proc %p\n", vpt); | ||
250 | send_sig(SIGKILL, vpt->task, 1); | ||
251 | vpt->task = NULL; | ||
252 | v9fs_mux_poll_task_num--; | ||
253 | } | ||
254 | v9fs_mux_num--; | ||
255 | down(&v9fs_mux_task_lock); | ||
256 | } | ||
257 | |||
56 | /** | 258 | /** |
57 | * xread - force read of a certain number of bytes | 259 | * v9fs_mux_init - allocate and initialize the per-session mux data |
58 | * @v9ses: session info structure | 260 | * Creates the polling task if this is the first session. |
59 | * @ptr: pointer to buffer | ||
60 | * @sz: number of bytes to read | ||
61 | * | 261 | * |
62 | * Chuck Cranor CS-533 project1 | 262 | * @trans - transport structure |
263 | * @msize - maximum message size | ||
264 | * @extended - pointer to the extended flag | ||
63 | */ | 265 | */ |
64 | 266 | struct v9fs_mux_data *v9fs_mux_init(struct v9fs_transport *trans, int msize, | |
65 | static int xread(struct v9fs_session_info *v9ses, void *ptr, unsigned long sz) | 267 | unsigned char *extended) |
66 | { | 268 | { |
67 | int rd = 0; | 269 | int i, n; |
68 | int ret = 0; | 270 | struct v9fs_mux_data *m, *mtmp; |
69 | while (rd < sz) { | 271 | |
70 | ret = v9ses->transport->read(v9ses->transport, ptr, sz - rd); | 272 | dprintk(DEBUG_MUX, "transport %p msize %d\n", trans, msize); |
71 | if (ret <= 0) { | 273 | m = kmalloc(sizeof(struct v9fs_mux_data), GFP_KERNEL); |
72 | dprintk(DEBUG_ERROR, "xread errno %d\n", ret); | 274 | if (!m) |
73 | return ret; | 275 | return ERR_PTR(-ENOMEM); |
276 | |||
277 | spin_lock_init(&m->lock); | ||
278 | INIT_LIST_HEAD(&m->mux_list); | ||
279 | m->msize = msize; | ||
280 | m->extended = extended; | ||
281 | m->trans = trans; | ||
282 | idr_init(&m->tidpool.pool); | ||
283 | init_MUTEX(&m->tidpool.lock); | ||
284 | m->err = 0; | ||
285 | init_waitqueue_head(&m->equeue); | ||
286 | INIT_LIST_HEAD(&m->req_list); | ||
287 | INIT_LIST_HEAD(&m->unsent_req_list); | ||
288 | m->rcall = NULL; | ||
289 | m->rpos = 0; | ||
290 | m->rbuf = NULL; | ||
291 | m->wpos = m->wsize = 0; | ||
292 | m->wbuf = NULL; | ||
293 | INIT_WORK(&m->rq, v9fs_read_work, m); | ||
294 | INIT_WORK(&m->wq, v9fs_write_work, m); | ||
295 | m->wsched = 0; | ||
296 | memset(&m->poll_waddr, 0, sizeof(m->poll_waddr)); | ||
297 | m->poll_task = NULL; | ||
298 | n = v9fs_mux_poll_start(m); | ||
299 | if (n) | ||
300 | return ERR_PTR(n); | ||
301 | |||
302 | n = trans->poll(trans, &m->pt); | ||
303 | if (n & POLLIN) { | ||
304 | dprintk(DEBUG_MUX, "mux %p can read\n", m); | ||
305 | set_bit(Rpending, &m->wsched); | ||
306 | } | ||
307 | |||
308 | if (n & POLLOUT) { | ||
309 | dprintk(DEBUG_MUX, "mux %p can write\n", m); | ||
310 | set_bit(Wpending, &m->wsched); | ||
311 | } | ||
312 | |||
313 | for(i = 0; i < ARRAY_SIZE(m->poll_waddr); i++) { | ||
314 | if (IS_ERR(m->poll_waddr[i])) { | ||
315 | v9fs_mux_poll_stop(m); | ||
316 | mtmp = (void *)m->poll_waddr; /* the error code */ | ||
317 | kfree(m); | ||
318 | m = mtmp; | ||
319 | break; | ||
74 | } | 320 | } |
75 | rd += ret; | ||
76 | ptr += ret; | ||
77 | } | 321 | } |
78 | return (rd); | 322 | |
323 | return m; | ||
79 | } | 324 | } |
80 | 325 | ||
81 | /** | 326 | /** |
82 | * read_message - read a full 9P2000 fcall packet | 327 | * v9fs_mux_destroy - cancels all pending requests and frees mux resources |
83 | * @v9ses: session info structure | ||
84 | * @rcall: fcall structure to read into | ||
85 | * @rcalllen: size of fcall buffer | ||
86 | * | ||
87 | */ | 328 | */ |
329 | void v9fs_mux_destroy(struct v9fs_mux_data *m) | ||
330 | { | ||
331 | dprintk(DEBUG_MUX, "mux %p prev %p next %p\n", m, | ||
332 | m->mux_list.prev, m->mux_list.next); | ||
333 | v9fs_mux_cancel(m, -ECONNRESET); | ||
334 | |||
335 | if (!list_empty(&m->req_list)) { | ||
336 | /* wait until all processes waiting on this session exit */ | ||
337 | dprintk(DEBUG_MUX, "mux %p waiting for empty request queue\n", | ||
338 | m); | ||
339 | wait_event_timeout(m->equeue, (list_empty(&m->req_list)), 5000); | ||
340 | dprintk(DEBUG_MUX, "mux %p request queue empty: %d\n", m, | ||
341 | list_empty(&m->req_list)); | ||
342 | } | ||
343 | |||
344 | v9fs_mux_poll_stop(m); | ||
345 | m->trans = NULL; | ||
346 | |||
347 | kfree(m); | ||
348 | } | ||
88 | 349 | ||
89 | static int | 350 | /** |
90 | read_message(struct v9fs_session_info *v9ses, | 351 | * v9fs_pollwait - called by files poll operation to add v9fs-poll task |
91 | struct v9fs_fcall *rcall, int rcalllen) | 352 | * to files wait queue |
353 | */ | ||
354 | static void | ||
355 | v9fs_pollwait(struct file *filp, wait_queue_head_t * wait_address, | ||
356 | poll_table * p) | ||
92 | { | 357 | { |
93 | unsigned char buf[4]; | 358 | int i; |
94 | void *data; | 359 | struct v9fs_mux_data *m; |
95 | int size = 0; | 360 | |
96 | int res = 0; | 361 | m = container_of(p, struct v9fs_mux_data, pt); |
97 | 362 | for(i = 0; i < ARRAY_SIZE(m->poll_waddr); i++) | |
98 | res = xread(v9ses, buf, sizeof(buf)); | 363 | if (m->poll_waddr[i] == NULL) |
99 | if (res < 0) { | 364 | break; |
100 | dprintk(DEBUG_ERROR, | 365 | |
101 | "Reading of count field failed returned: %d\n", res); | 366 | if (i >= ARRAY_SIZE(m->poll_waddr)) { |
102 | return res; | 367 | dprintk(DEBUG_ERROR, "not enough wait_address slots\n"); |
368 | return; | ||
103 | } | 369 | } |
104 | 370 | ||
105 | if (res < 4) { | 371 | m->poll_waddr[i] = wait_address; |
106 | dprintk(DEBUG_ERROR, | 372 | |
107 | "Reading of count field failed returned: %d\n", res); | 373 | if (!wait_address) { |
108 | return -EIO; | 374 | dprintk(DEBUG_ERROR, "no wait_address\n"); |
375 | m->poll_waddr[i] = ERR_PTR(-EIO); | ||
376 | return; | ||
109 | } | 377 | } |
110 | 378 | ||
111 | size = buf[0] | (buf[1] << 8) | (buf[2] << 16) | (buf[3] << 24); | 379 | init_waitqueue_entry(&m->poll_wait[i], m->poll_task->task); |
112 | dprintk(DEBUG_MUX, "got a packet count: %d\n", size); | 380 | add_wait_queue(wait_address, &m->poll_wait[i]); |
381 | } | ||
382 | |||
383 | /** | ||
384 | * v9fs_poll_mux - polls a mux and schedules read or write works if necessary | ||
385 | */ | ||
386 | static inline void v9fs_poll_mux(struct v9fs_mux_data *m) | ||
387 | { | ||
388 | int n; | ||
113 | 389 | ||
114 | /* adjust for the four bytes of size */ | 390 | if (m->err < 0) |
115 | size -= 4; | 391 | return; |
116 | 392 | ||
117 | if (size > v9ses->maxdata) { | 393 | n = m->trans->poll(m->trans, NULL); |
118 | dprintk(DEBUG_ERROR, "packet too big: %d\n", size); | 394 | if (n < 0 || n & (POLLERR | POLLHUP | POLLNVAL)) { |
119 | return -E2BIG; | 395 | dprintk(DEBUG_MUX, "error mux %p err %d\n", m, n); |
396 | if (n >= 0) | ||
397 | n = -ECONNRESET; | ||
398 | v9fs_mux_cancel(m, n); | ||
120 | } | 399 | } |
121 | 400 | ||
122 | data = kmalloc(size, GFP_KERNEL); | 401 | if (n & POLLIN) { |
123 | if (!data) { | 402 | set_bit(Rpending, &m->wsched); |
124 | eprintk(KERN_WARNING, "out of memory\n"); | 403 | dprintk(DEBUG_MUX, "mux %p can read\n", m); |
125 | return -ENOMEM; | 404 | if (!test_and_set_bit(Rworksched, &m->wsched)) { |
405 | dprintk(DEBUG_MUX, "schedule read work mux %p\n", m); | ||
406 | queue_work(v9fs_mux_wq, &m->rq); | ||
407 | } | ||
126 | } | 408 | } |
127 | 409 | ||
128 | res = xread(v9ses, data, size); | 410 | if (n & POLLOUT) { |
129 | if (res < size) { | 411 | set_bit(Wpending, &m->wsched); |
130 | dprintk(DEBUG_ERROR, "Reading of fcall failed returned: %d\n", | 412 | dprintk(DEBUG_MUX, "mux %p can write\n", m); |
131 | res); | 413 | if ((m->wsize || !list_empty(&m->unsent_req_list)) |
132 | kfree(data); | 414 | && !test_and_set_bit(Wworksched, &m->wsched)) { |
133 | return res; | 415 | dprintk(DEBUG_MUX, "schedule write work mux %p\n", m); |
416 | queue_work(v9fs_mux_wq, &m->wq); | ||
417 | } | ||
134 | } | 418 | } |
419 | } | ||
420 | |||
421 | /** | ||
422 | * v9fs_poll_proc - polls all v9fs transports for new events and queues | ||
423 | * the appropriate work to the work queue | ||
424 | */ | ||
425 | static int v9fs_poll_proc(void *a) | ||
426 | { | ||
427 | struct v9fs_mux_data *m, *mtmp; | ||
428 | struct v9fs_mux_poll_task *vpt; | ||
135 | 429 | ||
136 | /* we now have an in-memory string that is the reply. | 430 | vpt = a; |
137 | * deserialize it. There is very little to go wrong at this point | 431 | dprintk(DEBUG_MUX, "start %p %p\n", current, vpt); |
138 | * save for v9fs_alloc errors. | 432 | allow_signal(SIGKILL); |
139 | */ | 433 | while (!kthread_should_stop()) { |
140 | res = v9fs_deserialize_fcall(v9ses, size, data, v9ses->maxdata, | 434 | set_current_state(TASK_INTERRUPTIBLE); |
141 | rcall, rcalllen); | 435 | if (signal_pending(current)) |
436 | break; | ||
142 | 437 | ||
143 | kfree(data); | 438 | list_for_each_entry_safe(m, mtmp, &vpt->mux_list, mux_list) { |
439 | v9fs_poll_mux(m); | ||
440 | } | ||
144 | 441 | ||
145 | if (res < 0) | 442 | dprintk(DEBUG_MUX, "sleeping...\n"); |
146 | return res; | 443 | schedule_timeout(SCHED_TIMEOUT * HZ); |
444 | } | ||
147 | 445 | ||
446 | __set_current_state(TASK_RUNNING); | ||
447 | dprintk(DEBUG_MUX, "finish\n"); | ||
148 | return 0; | 448 | return 0; |
149 | } | 449 | } |
150 | 450 | ||
151 | /** | 451 | /** |
152 | * v9fs_recv - receive an RPC response for a particular tag | 452 | * v9fs_write_work - called when a transport can send some data |
153 | * @v9ses: session info structure | ||
154 | * @req: RPC request structure | ||
155 | * | ||
156 | */ | 453 | */ |
157 | 454 | static void v9fs_write_work(void *a) | |
158 | static int v9fs_recv(struct v9fs_session_info *v9ses, struct v9fs_rpcreq *req) | ||
159 | { | 455 | { |
160 | int ret = 0; | 456 | int n, err; |
457 | struct v9fs_mux_data *m; | ||
458 | struct v9fs_req *req; | ||
161 | 459 | ||
162 | dprintk(DEBUG_MUX, "waiting for response: %d\n", req->tcall->tag); | 460 | m = a; |
163 | ret = wait_event_interruptible(v9ses->read_wait, | ||
164 | ((v9ses->transport->status != Connected) || | ||
165 | (req->rcall != 0) || (req->err < 0) || | ||
166 | dprintcond(v9ses, req))); | ||
167 | 461 | ||
168 | dprintk(DEBUG_MUX, "got it: rcall %p\n", req->rcall); | 462 | if (m->err < 0) { |
463 | clear_bit(Wworksched, &m->wsched); | ||
464 | return; | ||
465 | } | ||
169 | 466 | ||
170 | spin_lock(&v9ses->muxlock); | 467 | if (!m->wsize) { |
171 | list_del(&req->next); | 468 | if (list_empty(&m->unsent_req_list)) { |
172 | spin_unlock(&v9ses->muxlock); | 469 | clear_bit(Wworksched, &m->wsched); |
470 | return; | ||
471 | } | ||
173 | 472 | ||
174 | if (req->err < 0) | 473 | spin_lock(&m->lock); |
175 | return req->err; | 474 | req = |
475 | list_entry(m->unsent_req_list.next, struct v9fs_req, | ||
476 | req_list); | ||
477 | list_move_tail(&req->req_list, &m->req_list); | ||
478 | m->wbuf = req->tcall->sdata; | ||
479 | m->wsize = req->tcall->size; | ||
480 | m->wpos = 0; | ||
481 | dump_data(m->wbuf, m->wsize); | ||
482 | spin_unlock(&m->lock); | ||
483 | } | ||
176 | 484 | ||
177 | if (v9ses->transport->status == Disconnected) | 485 | dprintk(DEBUG_MUX, "mux %p pos %d size %d\n", m, m->wpos, m->wsize); |
178 | return -ECONNRESET; | 486 | clear_bit(Wpending, &m->wsched); |
487 | err = m->trans->write(m->trans, m->wbuf + m->wpos, m->wsize - m->wpos); | ||
488 | dprintk(DEBUG_MUX, "mux %p sent %d bytes\n", m, err); | ||
489 | if (err == -EAGAIN) { | ||
490 | clear_bit(Wworksched, &m->wsched); | ||
491 | return; | ||
492 | } | ||
179 | 493 | ||
180 | return ret; | 494 | if (err <= 0) |
181 | } | 495 | goto error; |
182 | 496 | ||
183 | /** | 497 | m->wpos += err; |
184 | * v9fs_send - send a 9P request | 498 | if (m->wpos == m->wsize) |
185 | * @v9ses: session info structure | 499 | m->wpos = m->wsize = 0; |
186 | * @req: RPC request to send | 500 | |
187 | * | 501 | if (m->wsize == 0 && !list_empty(&m->unsent_req_list)) { |
188 | */ | 502 | if (test_and_clear_bit(Wpending, &m->wsched)) |
503 | n = POLLOUT; | ||
504 | else | ||
505 | n = m->trans->poll(m->trans, NULL); | ||
506 | |||
507 | if (n & POLLOUT) { | ||
508 | dprintk(DEBUG_MUX, "schedule write work mux %p\n", m); | ||
509 | queue_work(v9fs_mux_wq, &m->wq); | ||
510 | } else | ||
511 | clear_bit(Wworksched, &m->wsched); | ||
512 | } else | ||
513 | clear_bit(Wworksched, &m->wsched); | ||
514 | |||
515 | return; | ||
189 | 516 | ||
190 | static int v9fs_send(struct v9fs_session_info *v9ses, struct v9fs_rpcreq *req) | 517 | error: |
518 | v9fs_mux_cancel(m, err); | ||
519 | clear_bit(Wworksched, &m->wsched); | ||
520 | } | ||
521 | |||
522 | static void process_request(struct v9fs_mux_data *m, struct v9fs_req *req) | ||
191 | { | 523 | { |
192 | int ret = -1; | 524 | int ecode, tag; |
193 | void *data = NULL; | 525 | struct v9fs_str *ename; |
194 | struct v9fs_fcall *tcall = req->tcall; | ||
195 | 526 | ||
196 | data = kmalloc(v9ses->maxdata + V9FS_IOHDRSZ, GFP_KERNEL); | 527 | tag = req->tag; |
197 | if (!data) | 528 | if (req->rcall->id == RERROR && !req->err) { |
198 | return -ENOMEM; | 529 | ecode = req->rcall->params.rerror.errno; |
530 | ename = &req->rcall->params.rerror.error; | ||
199 | 531 | ||
200 | tcall->size = 0; /* enforce size recalculation */ | 532 | dprintk(DEBUG_MUX, "Rerror %.*s\n", ename->len, ename->str); |
201 | ret = | ||
202 | v9fs_serialize_fcall(v9ses, tcall, data, | ||
203 | v9ses->maxdata + V9FS_IOHDRSZ); | ||
204 | if (ret < 0) | ||
205 | goto free_data; | ||
206 | 533 | ||
207 | spin_lock(&v9ses->muxlock); | 534 | if (*m->extended) |
208 | list_add(&req->next, &v9ses->mux_fcalls); | 535 | req->err = -ecode; |
209 | spin_unlock(&v9ses->muxlock); | ||
210 | 536 | ||
211 | dprintk(DEBUG_MUX, "sending message: tag %d size %d\n", tcall->tag, | 537 | if (!req->err) { |
212 | tcall->size); | 538 | req->err = v9fs_errstr2errno(ename->str, ename->len); |
213 | ret = v9ses->transport->write(v9ses->transport, data, tcall->size); | ||
214 | 539 | ||
215 | if (ret != tcall->size) { | 540 | if (!req->err) { /* string match failed */ |
216 | spin_lock(&v9ses->muxlock); | 541 | PRINT_FCALL_ERROR("unknown error", req->rcall); |
217 | list_del(&req->next); | 542 | } |
218 | kfree(req->rcall); | 543 | |
544 | if (!req->err) | ||
545 | req->err = -ESERVERFAULT; | ||
546 | } | ||
547 | } else if (req->tcall && req->rcall->id != req->tcall->id + 1) { | ||
548 | dprintk(DEBUG_ERROR, "fcall mismatch: expected %d, got %d\n", | ||
549 | req->tcall->id + 1, req->rcall->id); | ||
550 | if (!req->err) | ||
551 | req->err = -EIO; | ||
552 | } | ||
219 | 553 | ||
220 | spin_unlock(&v9ses->muxlock); | 554 | if (req->cb && req->err != ERREQFLUSH) { |
221 | if (ret >= 0) | 555 | dprintk(DEBUG_MUX, "calling callback tcall %p rcall %p\n", |
222 | ret = -EREMOTEIO; | 556 | req->tcall, req->rcall); |
557 | |||
558 | (*req->cb) (req->cba, req->tcall, req->rcall, req->err); | ||
559 | req->cb = NULL; | ||
223 | } else | 560 | } else |
224 | ret = 0; | 561 | kfree(req->rcall); |
225 | 562 | ||
226 | free_data: | 563 | v9fs_mux_put_tag(m, tag); |
227 | kfree(data); | 564 | |
228 | return ret; | 565 | wake_up(&m->equeue); |
566 | kfree(req); | ||
229 | } | 567 | } |
230 | 568 | ||
231 | /** | 569 | /** |
232 | * v9fs_mux_rpc - send a request, receive a response | 570 | * v9fs_read_work - called when there is some data to be read from a transport |
233 | * @v9ses: session info structure | ||
234 | * @tcall: fcall to send | ||
235 | * @rcall: buffer to place response into | ||
236 | * | ||
237 | */ | 571 | */ |
238 | 572 | static void v9fs_read_work(void *a) | |
239 | long | ||
240 | v9fs_mux_rpc(struct v9fs_session_info *v9ses, struct v9fs_fcall *tcall, | ||
241 | struct v9fs_fcall **rcall) | ||
242 | { | 573 | { |
243 | int tid = -1; | 574 | int n, err; |
244 | struct v9fs_fcall *fcall = NULL; | 575 | struct v9fs_mux_data *m; |
245 | struct v9fs_rpcreq req; | 576 | struct v9fs_req *req, *rptr, *rreq; |
246 | int ret = -1; | 577 | struct v9fs_fcall *rcall; |
247 | 578 | char *rbuf; | |
248 | if (!v9ses) | 579 | |
249 | return -EINVAL; | 580 | m = a; |
250 | 581 | ||
251 | if (!v9ses->transport || v9ses->transport->status != Connected) | 582 | if (m->err < 0) |
252 | return -EIO; | 583 | return; |
584 | |||
585 | rcall = NULL; | ||
586 | dprintk(DEBUG_MUX, "start mux %p pos %d\n", m, m->rpos); | ||
587 | |||
588 | if (!m->rcall) { | ||
589 | m->rcall = | ||
590 | kmalloc(sizeof(struct v9fs_fcall) + m->msize, GFP_KERNEL); | ||
591 | if (!m->rcall) { | ||
592 | err = -ENOMEM; | ||
593 | goto error; | ||
594 | } | ||
253 | 595 | ||
254 | if (rcall) | 596 | m->rbuf = (char *)m->rcall + sizeof(struct v9fs_fcall); |
255 | *rcall = NULL; | 597 | m->rpos = 0; |
598 | } | ||
256 | 599 | ||
257 | if (tcall->id != TVERSION) { | 600 | clear_bit(Rpending, &m->wsched); |
258 | tid = v9fs_get_idpool(&v9ses->tidpool); | 601 | err = m->trans->read(m->trans, m->rbuf + m->rpos, m->msize - m->rpos); |
259 | if (tid < 0) | 602 | dprintk(DEBUG_MUX, "mux %p got %d bytes\n", m, err); |
260 | return -ENOMEM; | 603 | if (err == -EAGAIN) { |
604 | clear_bit(Rworksched, &m->wsched); | ||
605 | return; | ||
261 | } | 606 | } |
262 | 607 | ||
263 | tcall->tag = tid; | 608 | if (err <= 0) |
609 | goto error; | ||
264 | 610 | ||
265 | req.tcall = tcall; | 611 | m->rpos += err; |
266 | req.err = 0; | 612 | while (m->rpos > 4) { |
267 | req.rcall = NULL; | 613 | n = le32_to_cpu(*(__le32 *) m->rbuf); |
614 | if (n >= m->msize) { | ||
615 | dprintk(DEBUG_ERROR, | ||
616 | "requested packet size too big: %d\n", n); | ||
617 | err = -EIO; | ||
618 | goto error; | ||
619 | } | ||
268 | 620 | ||
269 | ret = v9fs_send(v9ses, &req); | 621 | if (m->rpos < n) |
622 | break; | ||
270 | 623 | ||
271 | if (ret < 0) { | 624 | dump_data(m->rbuf, n); |
272 | if (tcall->id != TVERSION) | 625 | err = |
273 | v9fs_put_idpool(tid, &v9ses->tidpool); | 626 | v9fs_deserialize_fcall(m->rbuf, n, m->rcall, *m->extended); |
274 | dprintk(DEBUG_MUX, "error %d\n", ret); | 627 | if (err < 0) { |
275 | return ret; | 628 | goto error; |
276 | } | 629 | } |
630 | |||
631 | rcall = m->rcall; | ||
632 | rbuf = m->rbuf; | ||
633 | if (m->rpos > n) { | ||
634 | m->rcall = kmalloc(sizeof(struct v9fs_fcall) + m->msize, | ||
635 | GFP_KERNEL); | ||
636 | if (!m->rcall) { | ||
637 | err = -ENOMEM; | ||
638 | goto error; | ||
639 | } | ||
277 | 640 | ||
278 | ret = v9fs_recv(v9ses, &req); | 641 | m->rbuf = (char *)m->rcall + sizeof(struct v9fs_fcall); |
279 | 642 | memmove(m->rbuf, rbuf + n, m->rpos - n); | |
280 | fcall = req.rcall; | 643 | m->rpos -= n; |
281 | 644 | } else { | |
282 | dprintk(DEBUG_MUX, "received: tag=%x, ret=%d\n", tcall->tag, ret); | 645 | m->rcall = NULL; |
283 | if (ret == -ERESTARTSYS) { | 646 | m->rbuf = NULL; |
284 | if (v9ses->transport->status != Disconnected | 647 | m->rpos = 0; |
285 | && tcall->id != TFLUSH) { | ||
286 | unsigned long flags; | ||
287 | |||
288 | dprintk(DEBUG_MUX, "flushing the tag: %d\n", | ||
289 | tcall->tag); | ||
290 | clear_thread_flag(TIF_SIGPENDING); | ||
291 | v9fs_t_flush(v9ses, tcall->tag); | ||
292 | spin_lock_irqsave(¤t->sighand->siglock, flags); | ||
293 | recalc_sigpending(); | ||
294 | spin_unlock_irqrestore(¤t->sighand->siglock, | ||
295 | flags); | ||
296 | dprintk(DEBUG_MUX, "flushing done\n"); | ||
297 | } | 648 | } |
298 | 649 | ||
299 | goto release_req; | 650 | dprintk(DEBUG_MUX, "mux %p fcall id %d tag %d\n", m, rcall->id, |
300 | } else if (ret < 0) | 651 | rcall->tag); |
301 | goto release_req; | 652 | |
302 | 653 | req = NULL; | |
303 | if (!fcall) | 654 | spin_lock(&m->lock); |
304 | ret = -EIO; | 655 | list_for_each_entry_safe(rreq, rptr, &m->req_list, req_list) { |
305 | else { | 656 | if (rreq->tag == rcall->tag) { |
306 | if (fcall->id == RERROR) { | 657 | req = rreq; |
307 | ret = v9fs_errstr2errno(fcall->params.rerror.error); | 658 | req->rcall = rcall; |
308 | if (ret == 0) { /* string match failed */ | 659 | list_del(&req->req_list); |
309 | if (fcall->params.rerror.errno) | 660 | spin_unlock(&m->lock); |
310 | ret = -(fcall->params.rerror.errno); | 661 | process_request(m, req); |
311 | else | 662 | break; |
312 | ret = -ESERVERFAULT; | ||
313 | } | 663 | } |
314 | } else if (fcall->id != tcall->id + 1) { | 664 | |
315 | dprintk(DEBUG_ERROR, | 665 | } |
316 | "fcall mismatch: expected %d, got %d\n", | 666 | |
317 | tcall->id + 1, fcall->id); | 667 | if (!req) { |
318 | ret = -EIO; | 668 | spin_unlock(&m->lock); |
669 | if (err >= 0 && rcall->id != RFLUSH) | ||
670 | dprintk(DEBUG_ERROR, | ||
671 | "unexpected response mux %p id %d tag %d\n", | ||
672 | m, rcall->id, rcall->tag); | ||
673 | kfree(rcall); | ||
319 | } | 674 | } |
320 | } | 675 | } |
321 | 676 | ||
322 | release_req: | 677 | if (!list_empty(&m->req_list)) { |
323 | if (tcall->id != TVERSION) | 678 | if (test_and_clear_bit(Rpending, &m->wsched)) |
324 | v9fs_put_idpool(tid, &v9ses->tidpool); | 679 | n = POLLIN; |
325 | if (rcall) | 680 | else |
326 | *rcall = fcall; | 681 | n = m->trans->poll(m->trans, NULL); |
327 | else | 682 | |
328 | kfree(fcall); | 683 | if (n & POLLIN) { |
684 | dprintk(DEBUG_MUX, "schedule read work mux %p\n", m); | ||
685 | queue_work(v9fs_mux_wq, &m->rq); | ||
686 | } else | ||
687 | clear_bit(Rworksched, &m->wsched); | ||
688 | } else | ||
689 | clear_bit(Rworksched, &m->wsched); | ||
690 | |||
691 | return; | ||
329 | 692 | ||
330 | return ret; | 693 | error: |
694 | v9fs_mux_cancel(m, err); | ||
695 | clear_bit(Rworksched, &m->wsched); | ||
331 | } | 696 | } |
332 | 697 | ||
333 | /** | 698 | /** |
334 | * v9fs_mux_cancel_requests - cancels all pending requests | 699 | * v9fs_send_request - send 9P request |
700 | * The function can sleep until the request is scheduled for sending. | ||
701 | * The function can be interrupted. Return from the function is not | ||
702 | * a guarantee that the request is sent succesfully. Can return errors | ||
703 | * that can be retrieved by PTR_ERR macros. | ||
335 | * | 704 | * |
336 | * @v9ses: session info structure | 705 | * @m: mux data |
337 | * @err: error code to return to the requests | 706 | * @tc: request to be sent |
707 | * @cb: callback function to call when response is received | ||
708 | * @cba: parameter to pass to the callback function | ||
338 | */ | 709 | */ |
339 | void v9fs_mux_cancel_requests(struct v9fs_session_info *v9ses, int err) | 710 | static struct v9fs_req *v9fs_send_request(struct v9fs_mux_data *m, |
711 | struct v9fs_fcall *tc, | ||
712 | v9fs_mux_req_callback cb, void *cba) | ||
340 | { | 713 | { |
341 | struct v9fs_rpcreq *rptr; | 714 | int n; |
342 | struct v9fs_rpcreq *rreq; | 715 | struct v9fs_req *req; |
343 | 716 | ||
344 | dprintk(DEBUG_MUX, " %d\n", err); | 717 | dprintk(DEBUG_MUX, "mux %p task %p tcall %p id %d\n", m, current, |
345 | spin_lock(&v9ses->muxlock); | 718 | tc, tc->id); |
346 | list_for_each_entry_safe(rreq, rptr, &v9ses->mux_fcalls, next) { | 719 | if (m->err < 0) |
347 | rreq->err = err; | 720 | return ERR_PTR(m->err); |
348 | } | ||
349 | spin_unlock(&v9ses->muxlock); | ||
350 | wake_up_all(&v9ses->read_wait); | ||
351 | } | ||
352 | 721 | ||
353 | /** | 722 | req = kmalloc(sizeof(struct v9fs_req), GFP_KERNEL); |
354 | * v9fs_recvproc - kproc to handle demultiplexing responses | 723 | if (!req) |
355 | * @data: session info structure | 724 | return ERR_PTR(-ENOMEM); |
356 | * | ||
357 | */ | ||
358 | 725 | ||
359 | static int v9fs_recvproc(void *data) | 726 | if (tc->id == TVERSION) |
360 | { | 727 | n = V9FS_NOTAG; |
361 | struct v9fs_session_info *v9ses = (struct v9fs_session_info *)data; | 728 | else |
362 | struct v9fs_fcall *rcall = NULL; | 729 | n = v9fs_mux_get_tag(m); |
363 | struct v9fs_rpcreq *rptr; | ||
364 | struct v9fs_rpcreq *req; | ||
365 | struct v9fs_rpcreq *rreq; | ||
366 | int err = 0; | ||
367 | 730 | ||
368 | allow_signal(SIGKILL); | 731 | if (n < 0) |
369 | set_current_state(TASK_INTERRUPTIBLE); | 732 | return ERR_PTR(-ENOMEM); |
370 | complete(&v9ses->proccmpl); | ||
371 | while (!kthread_should_stop() && err >= 0) { | ||
372 | req = rptr = rreq = NULL; | ||
373 | |||
374 | rcall = kmalloc(v9ses->maxdata + V9FS_IOHDRSZ, GFP_KERNEL); | ||
375 | if (!rcall) { | ||
376 | eprintk(KERN_ERR, "no memory for buffers\n"); | ||
377 | break; | ||
378 | } | ||
379 | 733 | ||
380 | err = read_message(v9ses, rcall, v9ses->maxdata + V9FS_IOHDRSZ); | 734 | v9fs_set_tag(tc, n); |
381 | spin_lock(&v9ses->muxlock); | ||
382 | if (err < 0) { | ||
383 | list_for_each_entry_safe(rreq, rptr, &v9ses->mux_fcalls, next) { | ||
384 | rreq->err = err; | ||
385 | } | ||
386 | if(err != -ERESTARTSYS) | ||
387 | eprintk(KERN_ERR, | ||
388 | "Transport error while reading message %d\n", err); | ||
389 | } else { | ||
390 | list_for_each_entry_safe(rreq, rptr, &v9ses->mux_fcalls, next) { | ||
391 | if (rreq->tcall->tag == rcall->tag) { | ||
392 | req = rreq; | ||
393 | req->rcall = rcall; | ||
394 | break; | ||
395 | } | ||
396 | } | ||
397 | } | ||
398 | 735 | ||
399 | if (req && (req->tcall->id == TFLUSH)) { | 736 | req->tag = n; |
400 | struct v9fs_rpcreq *treq = NULL; | 737 | req->tcall = tc; |
401 | list_for_each_entry_safe(treq, rptr, &v9ses->mux_fcalls, next) { | 738 | req->rcall = NULL; |
402 | if (treq->tcall->tag == | 739 | req->err = 0; |
403 | req->tcall->params.tflush.oldtag) { | 740 | req->cb = cb; |
404 | list_del(&rptr->next); | 741 | req->cba = cba; |
405 | kfree(treq->rcall); | 742 | |
406 | break; | 743 | spin_lock(&m->lock); |
407 | } | 744 | list_add_tail(&req->req_list, &m->unsent_req_list); |
745 | spin_unlock(&m->lock); | ||
746 | |||
747 | if (test_and_clear_bit(Wpending, &m->wsched)) | ||
748 | n = POLLOUT; | ||
749 | else | ||
750 | n = m->trans->poll(m->trans, NULL); | ||
751 | |||
752 | if (n & POLLOUT && !test_and_set_bit(Wworksched, &m->wsched)) | ||
753 | queue_work(v9fs_mux_wq, &m->wq); | ||
754 | |||
755 | return req; | ||
756 | } | ||
757 | |||
758 | static inline void | ||
759 | v9fs_mux_flush_cb(void *a, struct v9fs_fcall *tc, struct v9fs_fcall *rc, | ||
760 | int err) | ||
761 | { | ||
762 | v9fs_mux_req_callback cb; | ||
763 | int tag; | ||
764 | struct v9fs_mux_data *m; | ||
765 | struct v9fs_req *req, *rptr; | ||
766 | |||
767 | m = a; | ||
768 | dprintk(DEBUG_MUX, "mux %p tc %p rc %p err %d oldtag %d\n", m, tc, | ||
769 | rc, err, tc->params.tflush.oldtag); | ||
770 | |||
771 | spin_lock(&m->lock); | ||
772 | cb = NULL; | ||
773 | tag = tc->params.tflush.oldtag; | ||
774 | list_for_each_entry_safe(req, rptr, &m->req_list, req_list) { | ||
775 | if (req->tag == tag) { | ||
776 | list_del(&req->req_list); | ||
777 | if (req->cb) { | ||
778 | cb = req->cb; | ||
779 | req->cb = NULL; | ||
780 | spin_unlock(&m->lock); | ||
781 | (*cb) (req->cba, req->tcall, req->rcall, | ||
782 | req->err); | ||
408 | } | 783 | } |
784 | kfree(req); | ||
785 | wake_up(&m->equeue); | ||
786 | break; | ||
409 | } | 787 | } |
788 | } | ||
410 | 789 | ||
411 | spin_unlock(&v9ses->muxlock); | 790 | if (!cb) |
791 | spin_unlock(&m->lock); | ||
412 | 792 | ||
413 | if (!req) { | 793 | v9fs_mux_put_tag(m, tag); |
414 | if (err >= 0) | 794 | kfree(tc); |
415 | dprintk(DEBUG_ERROR, | 795 | kfree(rc); |
416 | "unexpected response: id %d tag %d\n", | 796 | } |
417 | rcall->id, rcall->tag); | ||
418 | 797 | ||
419 | kfree(rcall); | 798 | static void |
420 | } | 799 | v9fs_mux_flush_request(struct v9fs_mux_data *m, struct v9fs_req *req) |
800 | { | ||
801 | struct v9fs_fcall *fc; | ||
421 | 802 | ||
422 | wake_up_all(&v9ses->read_wait); | 803 | dprintk(DEBUG_MUX, "mux %p req %p tag %d\n", m, req, req->tag); |
423 | set_current_state(TASK_INTERRUPTIBLE); | 804 | |
805 | fc = v9fs_create_tflush(req->tag); | ||
806 | v9fs_send_request(m, fc, v9fs_mux_flush_cb, m); | ||
807 | } | ||
808 | |||
809 | static void | ||
810 | v9fs_mux_rpc_cb(void *a, struct v9fs_fcall *tc, struct v9fs_fcall *rc, int err) | ||
811 | { | ||
812 | struct v9fs_mux_rpc *r; | ||
813 | |||
814 | if (err == ERREQFLUSH) { | ||
815 | dprintk(DEBUG_MUX, "err req flush\n"); | ||
816 | return; | ||
424 | } | 817 | } |
425 | 818 | ||
426 | v9ses->transport->close(v9ses->transport); | 819 | r = a; |
820 | dprintk(DEBUG_MUX, "mux %p req %p tc %p rc %p err %d\n", r->m, r->req, | ||
821 | tc, rc, err); | ||
822 | r->rcall = rc; | ||
823 | r->err = err; | ||
824 | wake_up(&r->wqueue); | ||
825 | } | ||
427 | 826 | ||
428 | /* Inform all pending processes about the failure */ | 827 | /** |
429 | wake_up_all(&v9ses->read_wait); | 828 | * v9fs_mux_rpc - sends 9P request and waits until a response is available. |
829 | * The function can be interrupted. | ||
830 | * @m: mux data | ||
831 | * @tc: request to be sent | ||
832 | * @rc: pointer where a pointer to the response is stored | ||
833 | */ | ||
834 | int | ||
835 | v9fs_mux_rpc(struct v9fs_mux_data *m, struct v9fs_fcall *tc, | ||
836 | struct v9fs_fcall **rc) | ||
837 | { | ||
838 | int err; | ||
839 | unsigned long flags; | ||
840 | struct v9fs_req *req; | ||
841 | struct v9fs_mux_rpc r; | ||
842 | |||
843 | r.err = 0; | ||
844 | r.rcall = NULL; | ||
845 | r.m = m; | ||
846 | init_waitqueue_head(&r.wqueue); | ||
847 | |||
848 | if (rc) | ||
849 | *rc = NULL; | ||
850 | |||
851 | req = v9fs_send_request(m, tc, v9fs_mux_rpc_cb, &r); | ||
852 | if (IS_ERR(req)) { | ||
853 | err = PTR_ERR(req); | ||
854 | dprintk(DEBUG_MUX, "error %d\n", err); | ||
855 | return PTR_ERR(req); | ||
856 | } | ||
430 | 857 | ||
431 | if (signal_pending(current)) | 858 | r.req = req; |
432 | complete(&v9ses->proccmpl); | 859 | dprintk(DEBUG_MUX, "mux %p tc %p tag %d rpc %p req %p\n", m, tc, |
860 | req->tag, &r, req); | ||
861 | err = wait_event_interruptible(r.wqueue, r.rcall != NULL || r.err < 0); | ||
862 | if (r.err < 0) | ||
863 | err = r.err; | ||
864 | |||
865 | if (err == -ERESTARTSYS && m->trans->status == Connected && m->err == 0) { | ||
866 | spin_lock(&m->lock); | ||
867 | req->tcall = NULL; | ||
868 | req->err = ERREQFLUSH; | ||
869 | spin_unlock(&m->lock); | ||
870 | |||
871 | clear_thread_flag(TIF_SIGPENDING); | ||
872 | v9fs_mux_flush_request(m, req); | ||
873 | spin_lock_irqsave(¤t->sighand->siglock, flags); | ||
874 | recalc_sigpending(); | ||
875 | spin_unlock_irqrestore(¤t->sighand->siglock, flags); | ||
876 | } | ||
433 | 877 | ||
434 | dprintk(DEBUG_MUX, "recvproc: end\n"); | 878 | if (!err) { |
435 | v9ses->recvproc = NULL; | 879 | if (r.rcall) |
880 | dprintk(DEBUG_MUX, "got response id %d tag %d\n", | ||
881 | r.rcall->id, r.rcall->tag); | ||
882 | |||
883 | if (rc) | ||
884 | *rc = r.rcall; | ||
885 | else | ||
886 | kfree(r.rcall); | ||
887 | } else { | ||
888 | kfree(r.rcall); | ||
889 | dprintk(DEBUG_MUX, "got error %d\n", err); | ||
890 | if (err > 0) | ||
891 | err = -EIO; | ||
892 | } | ||
436 | 893 | ||
437 | return err >= 0; | 894 | return err; |
438 | } | 895 | } |
439 | 896 | ||
440 | /** | 897 | /** |
441 | * v9fs_mux_init - initialize multiplexer (spawn kproc) | 898 | * v9fs_mux_rpcnb - sends 9P request without waiting for response. |
442 | * @v9ses: session info structure | 899 | * @m: mux data |
443 | * @dev_name: mount device information (to create unique kproc) | 900 | * @tc: request to be sent |
444 | * | 901 | * @cb: callback function to be called when response arrives |
902 | * @cba: value to pass to the callback function | ||
445 | */ | 903 | */ |
904 | int v9fs_mux_rpcnb(struct v9fs_mux_data *m, struct v9fs_fcall *tc, | ||
905 | v9fs_mux_req_callback cb, void *a) | ||
906 | { | ||
907 | int err; | ||
908 | struct v9fs_req *req; | ||
909 | |||
910 | req = v9fs_send_request(m, tc, cb, a); | ||
911 | if (IS_ERR(req)) { | ||
912 | err = PTR_ERR(req); | ||
913 | dprintk(DEBUG_MUX, "error %d\n", err); | ||
914 | return PTR_ERR(req); | ||
915 | } | ||
916 | |||
917 | dprintk(DEBUG_MUX, "mux %p tc %p tag %d\n", m, tc, req->tag); | ||
918 | return 0; | ||
919 | } | ||
446 | 920 | ||
447 | int v9fs_mux_init(struct v9fs_session_info *v9ses, const char *dev_name) | 921 | /** |
922 | * v9fs_mux_cancel - cancel all pending requests with error | ||
923 | * @m: mux data | ||
924 | * @err: error code | ||
925 | */ | ||
926 | void v9fs_mux_cancel(struct v9fs_mux_data *m, int err) | ||
448 | { | 927 | { |
449 | char procname[60]; | 928 | struct v9fs_req *req, *rtmp; |
450 | 929 | LIST_HEAD(cancel_list); | |
451 | strncpy(procname, dev_name, sizeof(procname)); | 930 | |
452 | procname[sizeof(procname) - 1] = 0; | 931 | dprintk(DEBUG_MUX, "mux %p err %d\n", m, err); |
453 | 932 | m->err = err; | |
454 | init_waitqueue_head(&v9ses->read_wait); | 933 | spin_lock(&m->lock); |
455 | init_completion(&v9ses->fcread); | 934 | list_for_each_entry_safe(req, rtmp, &m->req_list, req_list) { |
456 | init_completion(&v9ses->proccmpl); | 935 | list_move(&req->req_list, &cancel_list); |
457 | spin_lock_init(&v9ses->muxlock); | ||
458 | INIT_LIST_HEAD(&v9ses->mux_fcalls); | ||
459 | v9ses->recvproc = NULL; | ||
460 | v9ses->curfcall = NULL; | ||
461 | |||
462 | v9ses->recvproc = kthread_create(v9fs_recvproc, v9ses, | ||
463 | "v9fs_recvproc %s", procname); | ||
464 | |||
465 | if (IS_ERR(v9ses->recvproc)) { | ||
466 | eprintk(KERN_ERR, "cannot create receiving thread\n"); | ||
467 | v9fs_session_close(v9ses); | ||
468 | return -ECONNABORTED; | ||
469 | } | 936 | } |
937 | spin_unlock(&m->lock); | ||
470 | 938 | ||
471 | wake_up_process(v9ses->recvproc); | 939 | list_for_each_entry_safe(req, rtmp, &cancel_list, req_list) { |
472 | wait_for_completion(&v9ses->proccmpl); | 940 | list_del(&req->req_list); |
941 | if (!req->err) | ||
942 | req->err = err; | ||
473 | 943 | ||
474 | return 0; | 944 | if (req->cb) |
945 | (*req->cb) (req->cba, req->tcall, req->rcall, req->err); | ||
946 | else | ||
947 | kfree(req->rcall); | ||
948 | |||
949 | kfree(req); | ||
950 | } | ||
951 | |||
952 | wake_up(&m->equeue); | ||
953 | } | ||
954 | |||
955 | static u16 v9fs_mux_get_tag(struct v9fs_mux_data *m) | ||
956 | { | ||
957 | int tag; | ||
958 | |||
959 | tag = v9fs_get_idpool(&m->tidpool); | ||
960 | if (tag < 0) | ||
961 | return V9FS_NOTAG; | ||
962 | else | ||
963 | return (u16) tag; | ||
964 | } | ||
965 | |||
966 | static void v9fs_mux_put_tag(struct v9fs_mux_data *m, u16 tag) | ||
967 | { | ||
968 | if (tag != V9FS_NOTAG && v9fs_check_idpool(tag, &m->tidpool)) | ||
969 | v9fs_put_idpool(tag, &m->tidpool); | ||
475 | } | 970 | } |