diff options
author | Alan Stern <stern@rowland.harvard.edu> | 2005-12-17 18:02:38 -0500 |
---|---|---|
committer | Greg Kroah-Hartman <gregkh@suse.de> | 2006-03-20 17:49:57 -0500 |
commit | 0ed8fee1c1d38a62e981025ba40b5eba30c4ce2a (patch) | |
tree | 5f006afb273d5657251c49ae7144f669db9ed05d /drivers | |
parent | af0bb5998abe8ed28ee354dd4c71689cacdc91e9 (diff) |
[PATCH] UHCI: remove main list of URBs
As part of reorienting uhci-hcd away from URBs and toward endpoint
queues, this patch (as625) eliminates the driver's main list of URBs.
The list wsa used mainly in checking for URB completions; now the driver
goes through the list of active endpoints and checks the members of the
queues.
As a side effect, I had to remove the code that looks for FSBR timeouts.
For now, FSBR will remain on so long as any URBs on a full-speed control
or bulk queue request it, even if the queue isn't advancing. A later
patch can add more intelligent handling. This isn't a huge drawback;
it's pretty rare for an URB to get stuck for more than a fraction of a
second. (And it will help the people trying to use those insane HP USB
devices.)
Signed-off-by: Alan Stern <stern@rowland.harvard.edu>
Signed-off-by: Greg Kroah-Hartman <gregkh@suse.de>
Diffstat (limited to 'drivers')
-rw-r--r-- | drivers/usb/host/uhci-debug.c | 1 | ||||
-rw-r--r-- | drivers/usb/host/uhci-hcd.c | 2 | ||||
-rw-r--r-- | drivers/usb/host/uhci-hcd.h | 15 | ||||
-rw-r--r-- | drivers/usb/host/uhci-q.c | 631 |
4 files changed, 279 insertions, 370 deletions
diff --git a/drivers/usb/host/uhci-debug.c b/drivers/usb/host/uhci-debug.c index 6814783adf91..f2f5f8ce1715 100644 --- a/drivers/usb/host/uhci-debug.c +++ b/drivers/usb/host/uhci-debug.c | |||
@@ -114,7 +114,6 @@ static int uhci_show_urbp(struct urb_priv *urbp, char *buf, int len, int space) | |||
114 | } | 114 | } |
115 | 115 | ||
116 | out += sprintf(out, "%s", (urbp->fsbr ? " FSBR" : "")); | 116 | out += sprintf(out, "%s", (urbp->fsbr ? " FSBR" : "")); |
117 | out += sprintf(out, "%s", (urbp->fsbr_timeout ? " FSBR_TO" : "")); | ||
118 | 117 | ||
119 | if (urbp->urb->status != -EINPROGRESS) | 118 | if (urbp->urb->status != -EINPROGRESS) |
120 | out += sprintf(out, " Status=%d", urbp->urb->status); | 119 | out += sprintf(out, " Status=%d", urbp->urb->status); |
diff --git a/drivers/usb/host/uhci-hcd.c b/drivers/usb/host/uhci-hcd.c index 1ff4b8806372..9865f303d3fc 100644 --- a/drivers/usb/host/uhci-hcd.c +++ b/drivers/usb/host/uhci-hcd.c | |||
@@ -491,8 +491,6 @@ static int uhci_start(struct usb_hcd *hcd) | |||
491 | spin_lock_init(&uhci->lock); | 491 | spin_lock_init(&uhci->lock); |
492 | 492 | ||
493 | INIT_LIST_HEAD(&uhci->td_remove_list); | 493 | INIT_LIST_HEAD(&uhci->td_remove_list); |
494 | INIT_LIST_HEAD(&uhci->urb_list); | ||
495 | INIT_LIST_HEAD(&uhci->complete_list); | ||
496 | INIT_LIST_HEAD(&uhci->idle_qh_list); | 494 | INIT_LIST_HEAD(&uhci->idle_qh_list); |
497 | 495 | ||
498 | init_waitqueue_head(&uhci->waitqh); | 496 | init_waitqueue_head(&uhci->waitqh); |
diff --git a/drivers/usb/host/uhci-hcd.h b/drivers/usb/host/uhci-hcd.h index c057956667b5..7e96bef2e88f 100644 --- a/drivers/usb/host/uhci-hcd.h +++ b/drivers/usb/host/uhci-hcd.h | |||
@@ -132,6 +132,10 @@ struct uhci_qh { | |||
132 | 132 | ||
133 | unsigned int unlink_frame; /* When the QH was unlinked */ | 133 | unsigned int unlink_frame; /* When the QH was unlinked */ |
134 | int state; /* QH_STATE_xxx; see above */ | 134 | int state; /* QH_STATE_xxx; see above */ |
135 | |||
136 | unsigned int initial_toggle:1; /* Endpoint's current toggle value */ | ||
137 | unsigned int needs_fixup:1; /* Must fix the TD toggle values */ | ||
138 | unsigned int is_stopped:1; /* Queue was stopped by an error */ | ||
135 | } __attribute__((aligned(16))); | 139 | } __attribute__((aligned(16))); |
136 | 140 | ||
137 | /* | 141 | /* |
@@ -384,6 +388,7 @@ struct uhci_hcd { | |||
384 | 388 | ||
385 | struct uhci_td *term_td; /* Terminating TD, see UHCI bug */ | 389 | struct uhci_td *term_td; /* Terminating TD, see UHCI bug */ |
386 | struct uhci_qh *skelqh[UHCI_NUM_SKELQH]; /* Skeleton QHs */ | 390 | struct uhci_qh *skelqh[UHCI_NUM_SKELQH]; /* Skeleton QHs */ |
391 | struct uhci_qh *next_qh; /* Next QH to scan */ | ||
387 | 392 | ||
388 | spinlock_t lock; | 393 | spinlock_t lock; |
389 | 394 | ||
@@ -413,16 +418,10 @@ struct uhci_hcd { | |||
413 | unsigned long resuming_ports; | 418 | unsigned long resuming_ports; |
414 | unsigned long ports_timeout; /* Time to stop signalling */ | 419 | unsigned long ports_timeout; /* Time to stop signalling */ |
415 | 420 | ||
416 | /* Main list of URBs currently controlled by this HC */ | ||
417 | struct list_head urb_list; | ||
418 | |||
419 | /* List of TDs that are done, but waiting to be freed (race) */ | 421 | /* List of TDs that are done, but waiting to be freed (race) */ |
420 | struct list_head td_remove_list; | 422 | struct list_head td_remove_list; |
421 | unsigned int td_remove_age; /* Age in frames */ | 423 | unsigned int td_remove_age; /* Age in frames */ |
422 | 424 | ||
423 | /* List of URBs awaiting completion callback */ | ||
424 | struct list_head complete_list; | ||
425 | |||
426 | struct list_head idle_qh_list; /* Where the idle QHs live */ | 425 | struct list_head idle_qh_list; /* Where the idle QHs live */ |
427 | 426 | ||
428 | int rh_numports; /* Number of root-hub ports */ | 427 | int rh_numports; /* Number of root-hub ports */ |
@@ -448,7 +447,6 @@ static inline struct usb_hcd *uhci_to_hcd(struct uhci_hcd *uhci) | |||
448 | * Private per-URB data | 447 | * Private per-URB data |
449 | */ | 448 | */ |
450 | struct urb_priv { | 449 | struct urb_priv { |
451 | struct list_head urb_list; | ||
452 | struct list_head node; /* Node in the QH's urbp list */ | 450 | struct list_head node; /* Node in the QH's urbp list */ |
453 | 451 | ||
454 | struct urb *urb; | 452 | struct urb *urb; |
@@ -456,10 +454,7 @@ struct urb_priv { | |||
456 | struct uhci_qh *qh; /* QH for this URB */ | 454 | struct uhci_qh *qh; /* QH for this URB */ |
457 | struct list_head td_list; | 455 | struct list_head td_list; |
458 | 456 | ||
459 | unsigned long fsbrtime; /* In jiffies */ | ||
460 | |||
461 | unsigned fsbr : 1; /* URB turned on FSBR */ | 457 | unsigned fsbr : 1; /* URB turned on FSBR */ |
462 | unsigned fsbr_timeout : 1; /* URB timed out on FSBR */ | ||
463 | unsigned short_transfer : 1; /* URB got a short transfer, no | 458 | unsigned short_transfer : 1; /* URB got a short transfer, no |
464 | * need to rescan */ | 459 | * need to rescan */ |
465 | }; | 460 | }; |
diff --git a/drivers/usb/host/uhci-q.c b/drivers/usb/host/uhci-q.c index c4194182dcc4..44bba9a6d196 100644 --- a/drivers/usb/host/uhci-q.c +++ b/drivers/usb/host/uhci-q.c | |||
@@ -151,53 +151,6 @@ static void uhci_unlink_isochronous_tds(struct uhci_hcd *uhci, struct urb *urb) | |||
151 | wmb(); | 151 | wmb(); |
152 | } | 152 | } |
153 | 153 | ||
154 | /* | ||
155 | * Remove an URB's TDs from the hardware schedule | ||
156 | */ | ||
157 | static void uhci_remove_tds_from_schedule(struct uhci_hcd *uhci, | ||
158 | struct urb *urb, int status) | ||
159 | { | ||
160 | struct urb_priv *urbp = (struct urb_priv *) urb->hcpriv; | ||
161 | |||
162 | /* Isochronous TDs get unlinked directly from the frame list */ | ||
163 | if (usb_pipetype(urb->pipe) == PIPE_ISOCHRONOUS) { | ||
164 | uhci_unlink_isochronous_tds(uhci, urb); | ||
165 | return; | ||
166 | } | ||
167 | |||
168 | /* If the URB isn't first on its queue, adjust the link pointer | ||
169 | * of the last TD in the previous URB. */ | ||
170 | if (urbp->node.prev != &urbp->qh->queue) { | ||
171 | struct urb_priv *purbp; | ||
172 | struct uhci_td *ptd, *ltd; | ||
173 | |||
174 | if (status == -EINPROGRESS) | ||
175 | status = 0; | ||
176 | purbp = list_entry(urbp->node.prev, struct urb_priv, node); | ||
177 | ptd = list_entry(purbp->td_list.prev, struct uhci_td, | ||
178 | list); | ||
179 | ltd = list_entry(urbp->td_list.prev, struct uhci_td, | ||
180 | list); | ||
181 | ptd->link = ltd->link; | ||
182 | } | ||
183 | |||
184 | /* If the URB completed with an error, then the QH element certainly | ||
185 | * points to one of the URB's TDs. If it completed normally then | ||
186 | * the QH element has certainly moved on to the next URB. And if | ||
187 | * the URB is still in progress then it must have been dequeued. | ||
188 | * The QH element either hasn't reached it yet or is somewhere in | ||
189 | * the middle. If the URB wasn't first we can assume that it | ||
190 | * hasn't started yet (see above): Otherwise all the preceding URBs | ||
191 | * would have completed and been removed from the queue, so this one | ||
192 | * _would_ be first. | ||
193 | * | ||
194 | * If the QH element is inside this URB, clear it. It will be | ||
195 | * set properly when the QH is activated. | ||
196 | */ | ||
197 | if (status < 0) | ||
198 | urbp->qh->element = UHCI_PTR_TERM; | ||
199 | } | ||
200 | |||
201 | static struct uhci_qh *uhci_alloc_qh(struct uhci_hcd *uhci, | 154 | static struct uhci_qh *uhci_alloc_qh(struct uhci_hcd *uhci, |
202 | struct usb_device *udev, struct usb_host_endpoint *hep) | 155 | struct usb_device *udev, struct usb_host_endpoint *hep) |
203 | { | 156 | { |
@@ -251,6 +204,90 @@ static void uhci_free_qh(struct uhci_hcd *uhci, struct uhci_qh *qh) | |||
251 | } | 204 | } |
252 | 205 | ||
253 | /* | 206 | /* |
207 | * When the currently executing URB is dequeued, save its current toggle value | ||
208 | */ | ||
209 | static void uhci_save_toggle(struct uhci_qh *qh, struct urb *urb) | ||
210 | { | ||
211 | struct urb_priv *urbp = (struct urb_priv *) urb->hcpriv; | ||
212 | struct uhci_td *td; | ||
213 | |||
214 | /* If the QH element pointer is UHCI_PTR_TERM then then currently | ||
215 | * executing URB has already been unlinked, so this one isn't it. */ | ||
216 | if (qh_element(qh) == UHCI_PTR_TERM || | ||
217 | qh->queue.next != &urbp->node) | ||
218 | return; | ||
219 | qh->element = UHCI_PTR_TERM; | ||
220 | |||
221 | /* Only bulk and interrupt pipes have to worry about toggles */ | ||
222 | if (!(usb_pipetype(urb->pipe) == PIPE_BULK || | ||
223 | usb_pipetype(urb->pipe) == PIPE_INTERRUPT)) | ||
224 | return; | ||
225 | |||
226 | /* Find the first active TD; that's the device's toggle state */ | ||
227 | list_for_each_entry(td, &urbp->td_list, list) { | ||
228 | if (td_status(td) & TD_CTRL_ACTIVE) { | ||
229 | qh->needs_fixup = 1; | ||
230 | qh->initial_toggle = uhci_toggle(td_token(td)); | ||
231 | return; | ||
232 | } | ||
233 | } | ||
234 | |||
235 | WARN_ON(1); | ||
236 | } | ||
237 | |||
238 | /* | ||
239 | * Fix up the data toggles for URBs in a queue, when one of them | ||
240 | * terminates early (short transfer, error, or dequeued). | ||
241 | */ | ||
242 | static void uhci_fixup_toggles(struct uhci_qh *qh, int skip_first) | ||
243 | { | ||
244 | struct urb_priv *urbp = NULL; | ||
245 | struct uhci_td *td; | ||
246 | unsigned int toggle = qh->initial_toggle; | ||
247 | unsigned int pipe; | ||
248 | |||
249 | /* Fixups for a short transfer start with the second URB in the | ||
250 | * queue (the short URB is the first). */ | ||
251 | if (skip_first) | ||
252 | urbp = list_entry(qh->queue.next, struct urb_priv, node); | ||
253 | |||
254 | /* When starting with the first URB, if the QH element pointer is | ||
255 | * still valid then we know the URB's toggles are okay. */ | ||
256 | else if (qh_element(qh) != UHCI_PTR_TERM) | ||
257 | toggle = 2; | ||
258 | |||
259 | /* Fix up the toggle for the URBs in the queue. Normally this | ||
260 | * loop won't run more than once: When an error or short transfer | ||
261 | * occurs, the queue usually gets emptied. */ | ||
262 | list_prepare_entry(urbp, &qh->queue, node); | ||
263 | list_for_each_entry_continue(urbp, &qh->queue, node) { | ||
264 | |||
265 | /* If the first TD has the right toggle value, we don't | ||
266 | * need to change any toggles in this URB */ | ||
267 | td = list_entry(urbp->td_list.next, struct uhci_td, list); | ||
268 | if (toggle > 1 || uhci_toggle(td_token(td)) == toggle) { | ||
269 | td = list_entry(urbp->td_list.next, struct uhci_td, | ||
270 | list); | ||
271 | toggle = uhci_toggle(td_token(td)) ^ 1; | ||
272 | |||
273 | /* Otherwise all the toggles in the URB have to be switched */ | ||
274 | } else { | ||
275 | list_for_each_entry(td, &urbp->td_list, list) { | ||
276 | td->token ^= __constant_cpu_to_le32( | ||
277 | TD_TOKEN_TOGGLE); | ||
278 | toggle ^= 1; | ||
279 | } | ||
280 | } | ||
281 | } | ||
282 | |||
283 | wmb(); | ||
284 | pipe = list_entry(qh->queue.next, struct urb_priv, node)->urb->pipe; | ||
285 | usb_settoggle(qh->udev, usb_pipeendpoint(pipe), | ||
286 | usb_pipeout(pipe), toggle); | ||
287 | qh->needs_fixup = 0; | ||
288 | } | ||
289 | |||
290 | /* | ||
254 | * Put a QH on the schedule in both hardware and software | 291 | * Put a QH on the schedule in both hardware and software |
255 | */ | 292 | */ |
256 | static void uhci_activate_qh(struct uhci_hcd *uhci, struct uhci_qh *qh) | 293 | static void uhci_activate_qh(struct uhci_hcd *uhci, struct uhci_qh *qh) |
@@ -276,6 +313,9 @@ static void uhci_activate_qh(struct uhci_hcd *uhci, struct uhci_qh *qh) | |||
276 | 313 | ||
277 | /* Move the QH from its old list to the end of the appropriate | 314 | /* Move the QH from its old list to the end of the appropriate |
278 | * skeleton's list */ | 315 | * skeleton's list */ |
316 | if (qh == uhci->next_qh) | ||
317 | uhci->next_qh = list_entry(qh->node.next, struct uhci_qh, | ||
318 | node); | ||
279 | list_move_tail(&qh->node, &qh->skel->node); | 319 | list_move_tail(&qh->node, &qh->skel->node); |
280 | 320 | ||
281 | /* Link it into the schedule */ | 321 | /* Link it into the schedule */ |
@@ -310,6 +350,9 @@ static void uhci_unlink_qh(struct uhci_hcd *uhci, struct uhci_qh *qh) | |||
310 | uhci_set_next_interrupt(uhci); | 350 | uhci_set_next_interrupt(uhci); |
311 | 351 | ||
312 | /* Move the QH from its old list to the end of the unlinking list */ | 352 | /* Move the QH from its old list to the end of the unlinking list */ |
353 | if (qh == uhci->next_qh) | ||
354 | uhci->next_qh = list_entry(qh->node.next, struct uhci_qh, | ||
355 | node); | ||
313 | list_move_tail(&qh->node, &uhci->skel_unlink_qh->node); | 356 | list_move_tail(&qh->node, &uhci->skel_unlink_qh->node); |
314 | } | 357 | } |
315 | 358 | ||
@@ -323,6 +366,9 @@ static void uhci_make_qh_idle(struct uhci_hcd *uhci, struct uhci_qh *qh) | |||
323 | { | 366 | { |
324 | WARN_ON(qh->state == QH_STATE_ACTIVE); | 367 | WARN_ON(qh->state == QH_STATE_ACTIVE); |
325 | 368 | ||
369 | if (qh == uhci->next_qh) | ||
370 | uhci->next_qh = list_entry(qh->node.next, struct uhci_qh, | ||
371 | node); | ||
326 | list_move(&qh->node, &uhci->idle_qh_list); | 372 | list_move(&qh->node, &uhci->idle_qh_list); |
327 | qh->state = QH_STATE_IDLE; | 373 | qh->state = QH_STATE_IDLE; |
328 | 374 | ||
@@ -344,11 +390,9 @@ static inline struct urb_priv *uhci_alloc_urb_priv(struct uhci_hcd *uhci, | |||
344 | 390 | ||
345 | urbp->urb = urb; | 391 | urbp->urb = urb; |
346 | urb->hcpriv = urbp; | 392 | urb->hcpriv = urbp; |
347 | urbp->fsbrtime = jiffies; | ||
348 | 393 | ||
349 | INIT_LIST_HEAD(&urbp->node); | 394 | INIT_LIST_HEAD(&urbp->node); |
350 | INIT_LIST_HEAD(&urbp->td_list); | 395 | INIT_LIST_HEAD(&urbp->td_list); |
351 | INIT_LIST_HEAD(&urbp->urb_list); | ||
352 | 396 | ||
353 | return urbp; | 397 | return urbp; |
354 | } | 398 | } |
@@ -373,9 +417,6 @@ static void uhci_free_urb_priv(struct uhci_hcd *uhci, | |||
373 | { | 417 | { |
374 | struct uhci_td *td, *tmp; | 418 | struct uhci_td *td, *tmp; |
375 | 419 | ||
376 | if (!list_empty(&urbp->urb_list)) | ||
377 | dev_warn(uhci_dev(uhci), "urb %p still on uhci->urb_list!\n", | ||
378 | urbp->urb); | ||
379 | if (!list_empty(&urbp->node)) | 420 | if (!list_empty(&urbp->node)) |
380 | dev_warn(uhci_dev(uhci), "urb %p still on QH's list!\n", | 421 | dev_warn(uhci_dev(uhci), "urb %p still on QH's list!\n", |
381 | urbp->urb); | 422 | urbp->urb); |
@@ -453,71 +494,6 @@ static int uhci_map_status(int status, int dir_out) | |||
453 | } | 494 | } |
454 | 495 | ||
455 | /* | 496 | /* |
456 | * Fix up the data toggles for URBs in a queue, when one of them | ||
457 | * terminates early (short transfer, error, or dequeued). | ||
458 | */ | ||
459 | static void uhci_fixup_toggles(struct urb *urb) | ||
460 | { | ||
461 | struct list_head *head; | ||
462 | struct uhci_td *td; | ||
463 | struct urb_priv *urbp = (struct urb_priv *) urb->hcpriv; | ||
464 | int prevactive = 0; | ||
465 | unsigned int toggle = 0; | ||
466 | struct urb_priv *turbp, *list_end; | ||
467 | |||
468 | /* | ||
469 | * We need to find out what the last successful toggle was so | ||
470 | * we can update the data toggles for the following transfers. | ||
471 | * | ||
472 | * There are 2 ways the last successful completed TD is found: | ||
473 | * | ||
474 | * 1) The TD is NOT active and the actual length < expected length | ||
475 | * 2) The TD is NOT active and it's the last TD in the chain | ||
476 | * | ||
477 | * and a third way the first uncompleted TD is found: | ||
478 | * | ||
479 | * 3) The TD is active and the previous TD is NOT active | ||
480 | */ | ||
481 | head = &urbp->td_list; | ||
482 | list_for_each_entry(td, head, list) { | ||
483 | unsigned int ctrlstat = td_status(td); | ||
484 | |||
485 | if (!(ctrlstat & TD_CTRL_ACTIVE) && | ||
486 | (uhci_actual_length(ctrlstat) < | ||
487 | uhci_expected_length(td_token(td)) || | ||
488 | td->list.next == head)) | ||
489 | toggle = uhci_toggle(td_token(td)) ^ 1; | ||
490 | else if ((ctrlstat & TD_CTRL_ACTIVE) && !prevactive) | ||
491 | toggle = uhci_toggle(td_token(td)); | ||
492 | |||
493 | prevactive = ctrlstat & TD_CTRL_ACTIVE; | ||
494 | } | ||
495 | |||
496 | /* | ||
497 | * Fix up the toggle for the following URBs in the queue. | ||
498 | * | ||
499 | * We can stop as soon as we find an URB with toggles set correctly, | ||
500 | * because then all the following URBs will be correct also. | ||
501 | */ | ||
502 | list_end = list_entry(&urbp->qh->queue, struct urb_priv, node); | ||
503 | turbp = urbp; | ||
504 | while ((turbp = list_entry(turbp->node.next, struct urb_priv, node)) | ||
505 | != list_end) { | ||
506 | td = list_entry(turbp->td_list.next, struct uhci_td, list); | ||
507 | if (uhci_toggle(td_token(td)) == toggle) | ||
508 | return; | ||
509 | |||
510 | list_for_each_entry(td, &turbp->td_list, list) { | ||
511 | td->token ^= __constant_cpu_to_le32(TD_TOKEN_TOGGLE); | ||
512 | toggle ^= 1; | ||
513 | } | ||
514 | } | ||
515 | |||
516 | usb_settoggle(urb->dev, usb_pipeendpoint(urb->pipe), | ||
517 | usb_pipeout(urb->pipe), toggle); | ||
518 | } | ||
519 | |||
520 | /* | ||
521 | * Control transfers | 497 | * Control transfers |
522 | */ | 498 | */ |
523 | static int uhci_submit_control(struct uhci_hcd *uhci, struct urb *urb, | 499 | static int uhci_submit_control(struct uhci_hcd *uhci, struct urb *urb, |
@@ -765,6 +741,9 @@ err: | |||
765 | } | 741 | } |
766 | } | 742 | } |
767 | 743 | ||
744 | /* Note that the queue has stopped */ | ||
745 | urbp->qh->element = UHCI_PTR_TERM; | ||
746 | urbp->qh->is_stopped = 1; | ||
768 | return ret; | 747 | return ret; |
769 | } | 748 | } |
770 | 749 | ||
@@ -927,7 +906,10 @@ static int uhci_result_common(struct uhci_hcd *uhci, struct urb *urb) | |||
927 | */ | 906 | */ |
928 | if (!urbp->short_transfer) { | 907 | if (!urbp->short_transfer) { |
929 | urbp->short_transfer = 1; | 908 | urbp->short_transfer = 1; |
930 | uhci_fixup_toggles(urb); | 909 | urbp->qh->initial_toggle = |
910 | uhci_toggle(td_token(td)) ^ 1; | ||
911 | uhci_fixup_toggles(urbp->qh, 1); | ||
912 | |||
931 | td = list_entry(urbp->td_list.prev, | 913 | td = list_entry(urbp->td_list.prev, |
932 | struct uhci_td, list); | 914 | struct uhci_td, list); |
933 | urbp->qh->element = td->link; | 915 | urbp->qh->element = td->link; |
@@ -962,6 +944,13 @@ err: | |||
962 | } | 944 | } |
963 | } | 945 | } |
964 | #endif | 946 | #endif |
947 | |||
948 | /* Note that the queue has stopped and save the next toggle value */ | ||
949 | urbp->qh->element = UHCI_PTR_TERM; | ||
950 | urbp->qh->is_stopped = 1; | ||
951 | urbp->qh->needs_fixup = 1; | ||
952 | urbp->qh->initial_toggle = uhci_toggle(td_token(td)) ^ | ||
953 | (ret == -EREMOTEIO); | ||
965 | return ret; | 954 | return ret; |
966 | } | 955 | } |
967 | 956 | ||
@@ -995,76 +984,39 @@ static inline int uhci_submit_interrupt(struct uhci_hcd *uhci, struct urb *urb, | |||
995 | /* | 984 | /* |
996 | * Isochronous transfers | 985 | * Isochronous transfers |
997 | */ | 986 | */ |
998 | static int isochronous_find_limits(struct uhci_hcd *uhci, struct urb *urb, unsigned int *start, unsigned int *end) | 987 | static int uhci_submit_isochronous(struct uhci_hcd *uhci, struct urb *urb, |
999 | { | 988 | struct uhci_qh *qh) |
1000 | struct urb *last_urb = NULL; | ||
1001 | struct urb_priv *up; | ||
1002 | int ret = 0; | ||
1003 | |||
1004 | list_for_each_entry(up, &uhci->urb_list, urb_list) { | ||
1005 | struct urb *u = up->urb; | ||
1006 | |||
1007 | /* look for pending URBs with identical pipe handle */ | ||
1008 | if ((urb->pipe == u->pipe) && (urb->dev == u->dev) && | ||
1009 | (u->status == -EINPROGRESS) && (u != urb)) { | ||
1010 | if (!last_urb) | ||
1011 | *start = u->start_frame; | ||
1012 | last_urb = u; | ||
1013 | } | ||
1014 | } | ||
1015 | |||
1016 | if (last_urb) { | ||
1017 | *end = (last_urb->start_frame + last_urb->number_of_packets * | ||
1018 | last_urb->interval) & (UHCI_NUMFRAMES-1); | ||
1019 | ret = 0; | ||
1020 | } else | ||
1021 | ret = -1; /* no previous urb found */ | ||
1022 | |||
1023 | return ret; | ||
1024 | } | ||
1025 | |||
1026 | static int isochronous_find_start(struct uhci_hcd *uhci, struct urb *urb) | ||
1027 | { | 989 | { |
1028 | int limits; | 990 | struct uhci_td *td = NULL; /* Since urb->number_of_packets > 0 */ |
1029 | unsigned int start = 0, end = 0; | 991 | int i, frame; |
992 | unsigned long destination, status; | ||
993 | struct urb_priv *urbp = (struct urb_priv *) urb->hcpriv; | ||
1030 | 994 | ||
1031 | if (urb->number_of_packets > 900) /* 900? Why? */ | 995 | if (urb->number_of_packets > 900) /* 900? Why? */ |
1032 | return -EFBIG; | 996 | return -EFBIG; |
1033 | 997 | ||
1034 | limits = isochronous_find_limits(uhci, urb, &start, &end); | 998 | status = TD_CTRL_ACTIVE | TD_CTRL_IOS; |
999 | destination = (urb->pipe & PIPE_DEVEP_MASK) | usb_packetid(urb->pipe); | ||
1035 | 1000 | ||
1001 | /* Figure out the starting frame number */ | ||
1036 | if (urb->transfer_flags & URB_ISO_ASAP) { | 1002 | if (urb->transfer_flags & URB_ISO_ASAP) { |
1037 | if (limits) { | 1003 | if (list_empty(&qh->queue)) { |
1038 | uhci_get_current_frame_number(uhci); | 1004 | uhci_get_current_frame_number(uhci); |
1039 | urb->start_frame = (uhci->frame_number + 10) | 1005 | urb->start_frame = (uhci->frame_number + 10); |
1040 | & (UHCI_NUMFRAMES - 1); | 1006 | |
1041 | } else | 1007 | } else { /* Go right after the last one */ |
1042 | urb->start_frame = end; | 1008 | struct urb *last_urb; |
1009 | |||
1010 | last_urb = list_entry(qh->queue.prev, | ||
1011 | struct urb_priv, node)->urb; | ||
1012 | urb->start_frame = (last_urb->start_frame + | ||
1013 | last_urb->number_of_packets * | ||
1014 | last_urb->interval); | ||
1015 | } | ||
1043 | } else { | 1016 | } else { |
1044 | urb->start_frame &= (UHCI_NUMFRAMES - 1); | ||
1045 | /* FIXME: Sanity check */ | 1017 | /* FIXME: Sanity check */ |
1046 | } | 1018 | } |
1047 | 1019 | urb->start_frame &= (UHCI_NUMFRAMES - 1); | |
1048 | return 0; | ||
1049 | } | ||
1050 | |||
1051 | /* | ||
1052 | * Isochronous transfers | ||
1053 | */ | ||
1054 | static int uhci_submit_isochronous(struct uhci_hcd *uhci, struct urb *urb, | ||
1055 | struct uhci_qh *qh) | ||
1056 | { | ||
1057 | struct uhci_td *td = NULL; /* Since urb->number_of_packets > 0 */ | ||
1058 | int i, ret, frame; | ||
1059 | unsigned long destination, status; | ||
1060 | struct urb_priv *urbp = (struct urb_priv *) urb->hcpriv; | ||
1061 | |||
1062 | status = TD_CTRL_ACTIVE | TD_CTRL_IOS; | ||
1063 | destination = (urb->pipe & PIPE_DEVEP_MASK) | usb_packetid(urb->pipe); | ||
1064 | |||
1065 | ret = isochronous_find_start(uhci, urb); | ||
1066 | if (ret) | ||
1067 | return ret; | ||
1068 | 1020 | ||
1069 | for (i = 0; i < urb->number_of_packets; i++) { | 1021 | for (i = 0; i < urb->number_of_packets; i++) { |
1070 | td = uhci_alloc_td(uhci); | 1022 | td = uhci_alloc_td(uhci); |
@@ -1203,7 +1155,6 @@ static int uhci_urb_enqueue(struct usb_hcd *hcd, | |||
1203 | /* Add this URB to the QH */ | 1155 | /* Add this URB to the QH */ |
1204 | urbp->qh = qh; | 1156 | urbp->qh = qh; |
1205 | list_add_tail(&urbp->node, &qh->queue); | 1157 | list_add_tail(&urbp->node, &qh->queue); |
1206 | list_add_tail(&urbp->urb_list, &uhci->urb_list); | ||
1207 | 1158 | ||
1208 | /* If the new URB is the first and only one on this QH then either | 1159 | /* If the new URB is the first and only one on this QH then either |
1209 | * the QH is new and idle or else it's unlinked and waiting to | 1160 | * the QH is new and idle or else it's unlinked and waiting to |
@@ -1224,49 +1175,66 @@ done: | |||
1224 | return ret; | 1175 | return ret; |
1225 | } | 1176 | } |
1226 | 1177 | ||
1178 | static int uhci_urb_dequeue(struct usb_hcd *hcd, struct urb *urb) | ||
1179 | { | ||
1180 | struct uhci_hcd *uhci = hcd_to_uhci(hcd); | ||
1181 | unsigned long flags; | ||
1182 | struct urb_priv *urbp; | ||
1183 | |||
1184 | spin_lock_irqsave(&uhci->lock, flags); | ||
1185 | urbp = urb->hcpriv; | ||
1186 | if (!urbp) /* URB was never linked! */ | ||
1187 | goto done; | ||
1188 | |||
1189 | /* Remove Isochronous TDs from the frame list ASAP */ | ||
1190 | if (usb_pipetype(urb->pipe) == PIPE_ISOCHRONOUS) | ||
1191 | uhci_unlink_isochronous_tds(uhci, urb); | ||
1192 | uhci_unlink_qh(uhci, urbp->qh); | ||
1193 | |||
1194 | done: | ||
1195 | spin_unlock_irqrestore(&uhci->lock, flags); | ||
1196 | return 0; | ||
1197 | } | ||
1198 | |||
1227 | /* | 1199 | /* |
1228 | * Return the result of a transfer | 1200 | * Finish unlinking an URB and give it back |
1229 | */ | 1201 | */ |
1230 | static void uhci_transfer_result(struct uhci_hcd *uhci, struct urb *urb) | 1202 | static void uhci_giveback_urb(struct uhci_hcd *uhci, struct uhci_qh *qh, |
1203 | struct urb *urb, struct pt_regs *regs) | ||
1204 | __releases(uhci->lock) | ||
1205 | __acquires(uhci->lock) | ||
1231 | { | 1206 | { |
1232 | int status; | ||
1233 | int okay_to_giveback = 0; | ||
1234 | struct urb_priv *urbp = (struct urb_priv *) urb->hcpriv; | 1207 | struct urb_priv *urbp = (struct urb_priv *) urb->hcpriv; |
1235 | 1208 | ||
1236 | switch (usb_pipetype(urb->pipe)) { | 1209 | /* Isochronous TDs get unlinked directly from the frame list */ |
1237 | case PIPE_CONTROL: | 1210 | if (usb_pipetype(urb->pipe) == PIPE_ISOCHRONOUS) |
1238 | status = uhci_result_control(uhci, urb); | 1211 | uhci_unlink_isochronous_tds(uhci, urb); |
1239 | break; | ||
1240 | case PIPE_ISOCHRONOUS: | ||
1241 | status = uhci_result_isochronous(uhci, urb); | ||
1242 | break; | ||
1243 | default: /* PIPE_BULK or PIPE_INTERRUPT */ | ||
1244 | status = uhci_result_common(uhci, urb); | ||
1245 | break; | ||
1246 | } | ||
1247 | 1212 | ||
1248 | spin_lock(&urb->lock); | 1213 | /* If the URB isn't first on its queue, adjust the link pointer |
1249 | if (urb->status == -EINPROGRESS) { /* Not yet dequeued */ | 1214 | * of the last TD in the previous URB. */ |
1250 | if (status != -EINPROGRESS) { /* URB has completed */ | 1215 | else if (qh->queue.next != &urbp->node) { |
1251 | urb->status = status; | 1216 | struct urb_priv *purbp; |
1217 | struct uhci_td *ptd, *ltd; | ||
1252 | 1218 | ||
1253 | /* If the URB got a real error (as opposed to | 1219 | purbp = list_entry(urbp->node.prev, struct urb_priv, node); |
1254 | * simply being dequeued), we don't have to | 1220 | ptd = list_entry(purbp->td_list.prev, struct uhci_td, |
1255 | * unlink the QH. Fix this later... */ | 1221 | list); |
1256 | if (status < 0) | 1222 | ltd = list_entry(urbp->td_list.prev, struct uhci_td, |
1257 | uhci_unlink_qh(uhci, urbp->qh); | 1223 | list); |
1258 | else | 1224 | ptd->link = ltd->link; |
1259 | okay_to_giveback = 1; | ||
1260 | } | ||
1261 | } else { /* Already dequeued */ | ||
1262 | if (urbp->qh->state == QH_STATE_UNLINKING && | ||
1263 | uhci->frame_number + uhci->is_stopped != | ||
1264 | urbp->qh->unlink_frame) | ||
1265 | okay_to_giveback = 1; | ||
1266 | } | 1225 | } |
1267 | spin_unlock(&urb->lock); | 1226 | |
1268 | if (!okay_to_giveback) | 1227 | /* Take the URB off the QH's queue. If the queue is now empty, |
1269 | return; | 1228 | * this is a perfect time for a toggle fixup. */ |
1229 | list_del_init(&urbp->node); | ||
1230 | if (list_empty(&qh->queue) && qh->needs_fixup) { | ||
1231 | usb_settoggle(urb->dev, usb_pipeendpoint(urb->pipe), | ||
1232 | usb_pipeout(urb->pipe), qh->initial_toggle); | ||
1233 | qh->needs_fixup = 0; | ||
1234 | } | ||
1235 | |||
1236 | uhci_dec_fsbr(uhci, urb); /* Safe since it checks */ | ||
1237 | uhci_free_urb_priv(uhci, urbp); | ||
1270 | 1238 | ||
1271 | switch (usb_pipetype(urb->pipe)) { | 1239 | switch (usb_pipetype(urb->pipe)) { |
1272 | case PIPE_ISOCHRONOUS: | 1240 | case PIPE_ISOCHRONOUS: |
@@ -1277,122 +1245,107 @@ static void uhci_transfer_result(struct uhci_hcd *uhci, struct urb *urb) | |||
1277 | case PIPE_INTERRUPT: | 1245 | case PIPE_INTERRUPT: |
1278 | /* Release bandwidth for Interrupt or Isoc. transfers */ | 1246 | /* Release bandwidth for Interrupt or Isoc. transfers */ |
1279 | /* Make sure we don't release if we have a queued URB */ | 1247 | /* Make sure we don't release if we have a queued URB */ |
1280 | if (list_empty(&urbp->qh->queue) && urb->bandwidth) | 1248 | if (list_empty(&qh->queue) && urb->bandwidth) |
1281 | usb_release_bandwidth(urb->dev, urb, 0); | 1249 | usb_release_bandwidth(urb->dev, urb, 0); |
1282 | else | 1250 | else |
1283 | /* bandwidth was passed on to queued URB, */ | 1251 | /* bandwidth was passed on to queued URB, */ |
1284 | /* so don't let usb_unlink_urb() release it */ | 1252 | /* so don't let usb_unlink_urb() release it */ |
1285 | urb->bandwidth = 0; | 1253 | urb->bandwidth = 0; |
1286 | /* Falls through */ | ||
1287 | case PIPE_BULK: | ||
1288 | if (status < 0) | ||
1289 | uhci_fixup_toggles(urb); | ||
1290 | break; | ||
1291 | default: /* PIPE_CONTROL */ | ||
1292 | break; | 1254 | break; |
1293 | } | 1255 | } |
1294 | 1256 | ||
1295 | /* Take the URB's TDs off the hardware schedule */ | 1257 | spin_unlock(&uhci->lock); |
1296 | uhci_remove_tds_from_schedule(uhci, urb, status); | 1258 | usb_hcd_giveback_urb(uhci_to_hcd(uhci), urb, regs); |
1297 | 1259 | spin_lock(&uhci->lock); | |
1298 | /* Take the URB off the QH's queue and see if the QH is now unused */ | ||
1299 | list_del_init(&urbp->node); | ||
1300 | if (list_empty(&urbp->qh->queue)) | ||
1301 | uhci_unlink_qh(uhci, urbp->qh); | ||
1302 | 1260 | ||
1303 | uhci_dec_fsbr(uhci, urb); /* Safe since it checks */ | 1261 | /* If the queue is now empty, we can unlink the QH and give up its |
1262 | * reserved bandwidth. */ | ||
1263 | if (list_empty(&qh->queue)) { | ||
1264 | uhci_unlink_qh(uhci, qh); | ||
1304 | 1265 | ||
1305 | /* Queue it for giving back */ | 1266 | /* Bandwidth stuff not yet implemented */ |
1306 | list_move_tail(&urbp->urb_list, &uhci->complete_list); | 1267 | } |
1307 | } | 1268 | } |
1308 | 1269 | ||
1309 | /* | 1270 | /* |
1310 | * Check out the QHs waiting to be fully unlinked | 1271 | * Scan the URBs in a QH's queue |
1311 | */ | 1272 | */ |
1312 | static void uhci_scan_unlinking_qhs(struct uhci_hcd *uhci) | 1273 | #define QH_FINISHED_UNLINKING(qh) \ |
1313 | { | 1274 | (qh->state == QH_STATE_UNLINKING && \ |
1314 | struct uhci_qh *qh, *tmp; | 1275 | uhci->frame_number + uhci->is_stopped != qh->unlink_frame) |
1315 | 1276 | ||
1316 | list_for_each_entry_safe(qh, tmp, &uhci->skel_unlink_qh->node, node) { | 1277 | static void uhci_scan_qh(struct uhci_hcd *uhci, struct uhci_qh *qh, |
1317 | 1278 | struct pt_regs *regs) | |
1318 | /* If the queue is empty and the QH is fully unlinked then | ||
1319 | * it can become IDLE. */ | ||
1320 | if (list_empty(&qh->queue)) { | ||
1321 | if (uhci->frame_number + uhci->is_stopped != | ||
1322 | qh->unlink_frame) | ||
1323 | uhci_make_qh_idle(uhci, qh); | ||
1324 | |||
1325 | /* If none of the QH's URBs have been dequeued then the QH | ||
1326 | * should be re-activated. */ | ||
1327 | } else { | ||
1328 | struct urb_priv *urbp; | ||
1329 | int any_dequeued = 0; | ||
1330 | |||
1331 | list_for_each_entry(urbp, &qh->queue, node) { | ||
1332 | if (urbp->urb->status != -EINPROGRESS) { | ||
1333 | any_dequeued = 1; | ||
1334 | break; | ||
1335 | } | ||
1336 | } | ||
1337 | if (!any_dequeued) | ||
1338 | uhci_activate_qh(uhci, qh); | ||
1339 | } | ||
1340 | } | ||
1341 | } | ||
1342 | |||
1343 | static int uhci_urb_dequeue(struct usb_hcd *hcd, struct urb *urb) | ||
1344 | { | 1279 | { |
1345 | struct uhci_hcd *uhci = hcd_to_uhci(hcd); | ||
1346 | unsigned long flags; | ||
1347 | struct urb_priv *urbp; | 1280 | struct urb_priv *urbp; |
1281 | struct urb *urb; | ||
1282 | int status; | ||
1348 | 1283 | ||
1349 | spin_lock_irqsave(&uhci->lock, flags); | 1284 | while (!list_empty(&qh->queue)) { |
1350 | urbp = urb->hcpriv; | 1285 | urbp = list_entry(qh->queue.next, struct urb_priv, node); |
1351 | if (!urbp) /* URB was never linked! */ | 1286 | urb = urbp->urb; |
1352 | goto done; | ||
1353 | 1287 | ||
1354 | /* Remove Isochronous TDs from the frame list ASAP */ | 1288 | switch (usb_pipetype(urb->pipe)) { |
1355 | if (usb_pipetype(urb->pipe) == PIPE_ISOCHRONOUS) | 1289 | case PIPE_CONTROL: |
1356 | uhci_unlink_isochronous_tds(uhci, urb); | 1290 | status = uhci_result_control(uhci, urb); |
1357 | uhci_unlink_qh(uhci, urbp->qh); | 1291 | break; |
1292 | case PIPE_ISOCHRONOUS: | ||
1293 | status = uhci_result_isochronous(uhci, urb); | ||
1294 | break; | ||
1295 | default: /* PIPE_BULK or PIPE_INTERRUPT */ | ||
1296 | status = uhci_result_common(uhci, urb); | ||
1297 | break; | ||
1298 | } | ||
1299 | if (status == -EINPROGRESS) | ||
1300 | break; | ||
1358 | 1301 | ||
1359 | done: | 1302 | spin_lock(&urb->lock); |
1360 | spin_unlock_irqrestore(&uhci->lock, flags); | 1303 | if (urb->status == -EINPROGRESS) /* Not dequeued */ |
1361 | return 0; | 1304 | urb->status = status; |
1362 | } | 1305 | else |
1306 | status = -ECONNRESET; | ||
1307 | spin_unlock(&urb->lock); | ||
1363 | 1308 | ||
1364 | static int uhci_fsbr_timeout(struct uhci_hcd *uhci, struct urb *urb) | 1309 | /* Dequeued but completed URBs can't be given back unless |
1365 | { | 1310 | * the QH is stopped or has finished unlinking. */ |
1366 | struct urb_priv *urbp = (struct urb_priv *)urb->hcpriv; | 1311 | if (status == -ECONNRESET && |
1367 | struct list_head *head; | 1312 | !(qh->is_stopped || QH_FINISHED_UNLINKING(qh))) |
1368 | struct uhci_td *td; | 1313 | return; |
1369 | int count = 0; | ||
1370 | 1314 | ||
1371 | uhci_dec_fsbr(uhci, urb); | 1315 | uhci_giveback_urb(uhci, qh, urb, regs); |
1316 | if (qh->is_stopped) | ||
1317 | break; | ||
1318 | } | ||
1372 | 1319 | ||
1373 | urbp->fsbr_timeout = 1; | 1320 | /* If the QH is neither stopped nor finished unlinking (normal case), |
1321 | * our work here is done. */ | ||
1322 | restart: | ||
1323 | if (!(qh->is_stopped || QH_FINISHED_UNLINKING(qh))) | ||
1324 | return; | ||
1374 | 1325 | ||
1375 | /* | 1326 | /* Otherwise give back each of the dequeued URBs */ |
1376 | * Ideally we would want to fix qh->element as well, but it's | 1327 | list_for_each_entry(urbp, &qh->queue, node) { |
1377 | * read/write by the HC, so that can introduce a race. It's not | 1328 | urb = urbp->urb; |
1378 | * really worth the hassle | 1329 | if (urb->status != -EINPROGRESS) { |
1379 | */ | 1330 | uhci_save_toggle(qh, urb); |
1331 | uhci_giveback_urb(uhci, qh, urb, regs); | ||
1332 | goto restart; | ||
1333 | } | ||
1334 | } | ||
1335 | qh->is_stopped = 0; | ||
1380 | 1336 | ||
1381 | head = &urbp->td_list; | 1337 | /* There are no more dequeued URBs. If there are still URBs on the |
1382 | list_for_each_entry(td, head, list) { | 1338 | * queue, the QH can now be re-activated. */ |
1383 | /* | 1339 | if (!list_empty(&qh->queue)) { |
1384 | * Make sure we don't do the last one (since it'll have the | 1340 | if (qh->needs_fixup) |
1385 | * TERM bit set) as well as we skip every so many TDs to | 1341 | uhci_fixup_toggles(qh, 0); |
1386 | * make sure it doesn't hog the bandwidth | 1342 | uhci_activate_qh(uhci, qh); |
1387 | */ | ||
1388 | if (td->list.next != head && (count % DEPTH_INTERVAL) == | ||
1389 | (DEPTH_INTERVAL - 1)) | ||
1390 | td->link |= UHCI_PTR_DEPTH; | ||
1391 | |||
1392 | count++; | ||
1393 | } | 1343 | } |
1394 | 1344 | ||
1395 | return 0; | 1345 | /* The queue is empty. The QH can become idle if it is fully |
1346 | * unlinked. */ | ||
1347 | else if (QH_FINISHED_UNLINKING(qh)) | ||
1348 | uhci_make_qh_idle(uhci, qh); | ||
1396 | } | 1349 | } |
1397 | 1350 | ||
1398 | static void uhci_free_pending_tds(struct uhci_hcd *uhci) | 1351 | static void uhci_free_pending_tds(struct uhci_hcd *uhci) |
@@ -1406,36 +1359,13 @@ static void uhci_free_pending_tds(struct uhci_hcd *uhci) | |||
1406 | } | 1359 | } |
1407 | } | 1360 | } |
1408 | 1361 | ||
1409 | static void | 1362 | /* |
1410 | uhci_finish_urb(struct usb_hcd *hcd, struct urb *urb, struct pt_regs *regs) | 1363 | * Process events in the schedule, but only in one thread at a time |
1411 | __releases(uhci->lock) | 1364 | */ |
1412 | __acquires(uhci->lock) | ||
1413 | { | ||
1414 | struct uhci_hcd *uhci = hcd_to_uhci(hcd); | ||
1415 | |||
1416 | uhci_free_urb_priv(uhci, (struct urb_priv *) (urb->hcpriv)); | ||
1417 | |||
1418 | spin_unlock(&uhci->lock); | ||
1419 | usb_hcd_giveback_urb(hcd, urb, regs); | ||
1420 | spin_lock(&uhci->lock); | ||
1421 | } | ||
1422 | |||
1423 | static void uhci_finish_completion(struct uhci_hcd *uhci, struct pt_regs *regs) | ||
1424 | { | ||
1425 | struct urb_priv *urbp, *tmp; | ||
1426 | |||
1427 | list_for_each_entry_safe(urbp, tmp, &uhci->complete_list, urb_list) { | ||
1428 | struct urb *urb = urbp->urb; | ||
1429 | |||
1430 | list_del_init(&urbp->urb_list); | ||
1431 | uhci_finish_urb(uhci_to_hcd(uhci), urb, regs); | ||
1432 | } | ||
1433 | } | ||
1434 | |||
1435 | /* Process events in the schedule, but only in one thread at a time */ | ||
1436 | static void uhci_scan_schedule(struct uhci_hcd *uhci, struct pt_regs *regs) | 1365 | static void uhci_scan_schedule(struct uhci_hcd *uhci, struct pt_regs *regs) |
1437 | { | 1366 | { |
1438 | struct urb_priv *urbp, *tmp; | 1367 | int i; |
1368 | struct uhci_qh *qh; | ||
1439 | 1369 | ||
1440 | /* Don't allow re-entrant calls */ | 1370 | /* Don't allow re-entrant calls */ |
1441 | if (uhci->scan_in_progress) { | 1371 | if (uhci->scan_in_progress) { |
@@ -1452,26 +1382,24 @@ static void uhci_scan_schedule(struct uhci_hcd *uhci, struct pt_regs *regs) | |||
1452 | if (uhci->frame_number + uhci->is_stopped != uhci->td_remove_age) | 1382 | if (uhci->frame_number + uhci->is_stopped != uhci->td_remove_age) |
1453 | uhci_free_pending_tds(uhci); | 1383 | uhci_free_pending_tds(uhci); |
1454 | 1384 | ||
1455 | /* Walk the list of pending URBs to see which ones completed | 1385 | /* Go through all the QH queues and process the URBs in each one */ |
1456 | * (must be _safe because uhci_transfer_result() dequeues URBs) */ | 1386 | for (i = 0; i < UHCI_NUM_SKELQH - 1; ++i) { |
1457 | list_for_each_entry_safe(urbp, tmp, &uhci->urb_list, urb_list) { | 1387 | uhci->next_qh = list_entry(uhci->skelqh[i]->node.next, |
1458 | struct urb *urb = urbp->urb; | 1388 | struct uhci_qh, node); |
1459 | 1389 | while ((qh = uhci->next_qh) != uhci->skelqh[i]) { | |
1460 | /* Checks the status and does all of the magic necessary */ | 1390 | uhci->next_qh = list_entry(qh->node.next, |
1461 | uhci_transfer_result(uhci, urb); | 1391 | struct uhci_qh, node); |
1392 | uhci_scan_qh(uhci, qh, regs); | ||
1393 | } | ||
1462 | } | 1394 | } |
1463 | uhci_finish_completion(uhci, regs); | ||
1464 | |||
1465 | /* If the controller is stopped, we can finish these off right now */ | ||
1466 | if (uhci->is_stopped) | ||
1467 | uhci_free_pending_tds(uhci); | ||
1468 | 1395 | ||
1469 | if (uhci->need_rescan) | 1396 | if (uhci->need_rescan) |
1470 | goto rescan; | 1397 | goto rescan; |
1471 | uhci->scan_in_progress = 0; | 1398 | uhci->scan_in_progress = 0; |
1472 | 1399 | ||
1473 | /* Check out the QHs waiting for unlinking */ | 1400 | /* If the controller is stopped, we can finish these off right now */ |
1474 | uhci_scan_unlinking_qhs(uhci); | 1401 | if (uhci->is_stopped) |
1402 | uhci_free_pending_tds(uhci); | ||
1475 | 1403 | ||
1476 | if (list_empty(&uhci->td_remove_list) && | 1404 | if (list_empty(&uhci->td_remove_list) && |
1477 | list_empty(&uhci->skel_unlink_qh->node)) | 1405 | list_empty(&uhci->skel_unlink_qh->node)) |
@@ -1482,19 +1410,8 @@ static void uhci_scan_schedule(struct uhci_hcd *uhci, struct pt_regs *regs) | |||
1482 | 1410 | ||
1483 | static void check_fsbr(struct uhci_hcd *uhci) | 1411 | static void check_fsbr(struct uhci_hcd *uhci) |
1484 | { | 1412 | { |
1485 | struct urb_priv *up; | 1413 | /* For now, don't scan URBs for FSBR timeouts. |
1486 | 1414 | * Add it back in later... */ | |
1487 | list_for_each_entry(up, &uhci->urb_list, urb_list) { | ||
1488 | struct urb *u = up->urb; | ||
1489 | |||
1490 | spin_lock(&u->lock); | ||
1491 | |||
1492 | /* Check if the FSBR timed out */ | ||
1493 | if (up->fsbr && !up->fsbr_timeout && time_after_eq(jiffies, up->fsbrtime + IDLE_TIMEOUT)) | ||
1494 | uhci_fsbr_timeout(uhci, u); | ||
1495 | |||
1496 | spin_unlock(&u->lock); | ||
1497 | } | ||
1498 | 1415 | ||
1499 | /* Really disable FSBR */ | 1416 | /* Really disable FSBR */ |
1500 | if (!uhci->fsbr && uhci->fsbrtimeout && time_after_eq(jiffies, uhci->fsbrtimeout)) { | 1417 | if (!uhci->fsbr && uhci->fsbrtimeout && time_after_eq(jiffies, uhci->fsbrtimeout)) { |