diff options
Diffstat (limited to 'drivers/usb/wusbcore/wa-xfer.c')
-rw-r--r-- | drivers/usb/wusbcore/wa-xfer.c | 180 |
1 files changed, 43 insertions, 137 deletions
diff --git a/drivers/usb/wusbcore/wa-xfer.c b/drivers/usb/wusbcore/wa-xfer.c index c038635d1c64..238a96aee3a1 100644 --- a/drivers/usb/wusbcore/wa-xfer.c +++ b/drivers/usb/wusbcore/wa-xfer.c | |||
@@ -82,13 +82,10 @@ | |||
82 | #include <linux/init.h> | 82 | #include <linux/init.h> |
83 | #include <linux/spinlock.h> | 83 | #include <linux/spinlock.h> |
84 | #include <linux/hash.h> | 84 | #include <linux/hash.h> |
85 | |||
85 | #include "wa-hc.h" | 86 | #include "wa-hc.h" |
86 | #include "wusbhc.h" | 87 | #include "wusbhc.h" |
87 | 88 | ||
88 | #undef D_LOCAL | ||
89 | #define D_LOCAL 0 /* 0 disabled, > 0 different levels... */ | ||
90 | #include <linux/uwb/debug.h> | ||
91 | |||
92 | enum { | 89 | enum { |
93 | WA_SEGS_MAX = 255, | 90 | WA_SEGS_MAX = 255, |
94 | }; | 91 | }; |
@@ -180,7 +177,6 @@ static void wa_xfer_destroy(struct kref *_xfer) | |||
180 | } | 177 | } |
181 | } | 178 | } |
182 | kfree(xfer); | 179 | kfree(xfer); |
183 | d_printf(2, NULL, "xfer %p destroyed\n", xfer); | ||
184 | } | 180 | } |
185 | 181 | ||
186 | static void wa_xfer_get(struct wa_xfer *xfer) | 182 | static void wa_xfer_get(struct wa_xfer *xfer) |
@@ -190,10 +186,7 @@ static void wa_xfer_get(struct wa_xfer *xfer) | |||
190 | 186 | ||
191 | static void wa_xfer_put(struct wa_xfer *xfer) | 187 | static void wa_xfer_put(struct wa_xfer *xfer) |
192 | { | 188 | { |
193 | d_fnstart(3, NULL, "(xfer %p) -- ref count bef put %d\n", | ||
194 | xfer, atomic_read(&xfer->refcnt.refcount)); | ||
195 | kref_put(&xfer->refcnt, wa_xfer_destroy); | 189 | kref_put(&xfer->refcnt, wa_xfer_destroy); |
196 | d_fnend(3, NULL, "(xfer %p) = void\n", xfer); | ||
197 | } | 190 | } |
198 | 191 | ||
199 | /* | 192 | /* |
@@ -209,7 +202,7 @@ static void wa_xfer_put(struct wa_xfer *xfer) | |||
209 | static void wa_xfer_giveback(struct wa_xfer *xfer) | 202 | static void wa_xfer_giveback(struct wa_xfer *xfer) |
210 | { | 203 | { |
211 | unsigned long flags; | 204 | unsigned long flags; |
212 | d_fnstart(3, NULL, "(xfer %p)\n", xfer); | 205 | |
213 | spin_lock_irqsave(&xfer->wa->xfer_list_lock, flags); | 206 | spin_lock_irqsave(&xfer->wa->xfer_list_lock, flags); |
214 | list_del_init(&xfer->list_node); | 207 | list_del_init(&xfer->list_node); |
215 | spin_unlock_irqrestore(&xfer->wa->xfer_list_lock, flags); | 208 | spin_unlock_irqrestore(&xfer->wa->xfer_list_lock, flags); |
@@ -217,7 +210,6 @@ static void wa_xfer_giveback(struct wa_xfer *xfer) | |||
217 | wusbhc_giveback_urb(xfer->wa->wusb, xfer->urb, xfer->result); | 210 | wusbhc_giveback_urb(xfer->wa->wusb, xfer->urb, xfer->result); |
218 | wa_put(xfer->wa); | 211 | wa_put(xfer->wa); |
219 | wa_xfer_put(xfer); | 212 | wa_xfer_put(xfer); |
220 | d_fnend(3, NULL, "(xfer %p) = void\n", xfer); | ||
221 | } | 213 | } |
222 | 214 | ||
223 | /* | 215 | /* |
@@ -227,13 +219,10 @@ static void wa_xfer_giveback(struct wa_xfer *xfer) | |||
227 | */ | 219 | */ |
228 | static void wa_xfer_completion(struct wa_xfer *xfer) | 220 | static void wa_xfer_completion(struct wa_xfer *xfer) |
229 | { | 221 | { |
230 | d_fnstart(3, NULL, "(xfer %p)\n", xfer); | ||
231 | if (xfer->wusb_dev) | 222 | if (xfer->wusb_dev) |
232 | wusb_dev_put(xfer->wusb_dev); | 223 | wusb_dev_put(xfer->wusb_dev); |
233 | rpipe_put(xfer->ep->hcpriv); | 224 | rpipe_put(xfer->ep->hcpriv); |
234 | wa_xfer_giveback(xfer); | 225 | wa_xfer_giveback(xfer); |
235 | d_fnend(3, NULL, "(xfer %p) = void\n", xfer); | ||
236 | return; | ||
237 | } | 226 | } |
238 | 227 | ||
239 | /* | 228 | /* |
@@ -243,12 +232,12 @@ static void wa_xfer_completion(struct wa_xfer *xfer) | |||
243 | */ | 232 | */ |
244 | static unsigned __wa_xfer_is_done(struct wa_xfer *xfer) | 233 | static unsigned __wa_xfer_is_done(struct wa_xfer *xfer) |
245 | { | 234 | { |
235 | struct device *dev = &xfer->wa->usb_iface->dev; | ||
246 | unsigned result, cnt; | 236 | unsigned result, cnt; |
247 | struct wa_seg *seg; | 237 | struct wa_seg *seg; |
248 | struct urb *urb = xfer->urb; | 238 | struct urb *urb = xfer->urb; |
249 | unsigned found_short = 0; | 239 | unsigned found_short = 0; |
250 | 240 | ||
251 | d_fnstart(3, NULL, "(xfer %p)\n", xfer); | ||
252 | result = xfer->segs_done == xfer->segs_submitted; | 241 | result = xfer->segs_done == xfer->segs_submitted; |
253 | if (result == 0) | 242 | if (result == 0) |
254 | goto out; | 243 | goto out; |
@@ -258,10 +247,8 @@ static unsigned __wa_xfer_is_done(struct wa_xfer *xfer) | |||
258 | switch (seg->status) { | 247 | switch (seg->status) { |
259 | case WA_SEG_DONE: | 248 | case WA_SEG_DONE: |
260 | if (found_short && seg->result > 0) { | 249 | if (found_short && seg->result > 0) { |
261 | if (printk_ratelimit()) | 250 | dev_dbg(dev, "xfer %p#%u: bad short segments (%zu)\n", |
262 | printk(KERN_ERR "xfer %p#%u: bad short " | 251 | xfer, cnt, seg->result); |
263 | "segments (%zu)\n", xfer, cnt, | ||
264 | seg->result); | ||
265 | urb->status = -EINVAL; | 252 | urb->status = -EINVAL; |
266 | goto out; | 253 | goto out; |
267 | } | 254 | } |
@@ -269,36 +256,30 @@ static unsigned __wa_xfer_is_done(struct wa_xfer *xfer) | |||
269 | if (seg->result < xfer->seg_size | 256 | if (seg->result < xfer->seg_size |
270 | && cnt != xfer->segs-1) | 257 | && cnt != xfer->segs-1) |
271 | found_short = 1; | 258 | found_short = 1; |
272 | d_printf(2, NULL, "xfer %p#%u: DONE short %d " | 259 | dev_dbg(dev, "xfer %p#%u: DONE short %d " |
273 | "result %zu urb->actual_length %d\n", | 260 | "result %zu urb->actual_length %d\n", |
274 | xfer, seg->index, found_short, seg->result, | 261 | xfer, seg->index, found_short, seg->result, |
275 | urb->actual_length); | 262 | urb->actual_length); |
276 | break; | 263 | break; |
277 | case WA_SEG_ERROR: | 264 | case WA_SEG_ERROR: |
278 | xfer->result = seg->result; | 265 | xfer->result = seg->result; |
279 | d_printf(2, NULL, "xfer %p#%u: ERROR result %zu\n", | 266 | dev_dbg(dev, "xfer %p#%u: ERROR result %zu\n", |
280 | xfer, seg->index, seg->result); | 267 | xfer, seg->index, seg->result); |
281 | goto out; | 268 | goto out; |
282 | case WA_SEG_ABORTED: | 269 | case WA_SEG_ABORTED: |
283 | WARN_ON(urb->status != -ECONNRESET | 270 | dev_dbg(dev, "xfer %p#%u ABORTED: result %d\n", |
284 | && urb->status != -ENOENT); | 271 | xfer, seg->index, urb->status); |
285 | d_printf(2, NULL, "xfer %p#%u ABORTED: result %d\n", | ||
286 | xfer, seg->index, urb->status); | ||
287 | xfer->result = urb->status; | 272 | xfer->result = urb->status; |
288 | goto out; | 273 | goto out; |
289 | default: | 274 | default: |
290 | /* if (printk_ratelimit()) */ | 275 | dev_warn(dev, "xfer %p#%u: is_done bad state %d\n", |
291 | printk(KERN_ERR "xfer %p#%u: " | 276 | xfer, cnt, seg->status); |
292 | "is_done bad state %d\n", | ||
293 | xfer, cnt, seg->status); | ||
294 | xfer->result = -EINVAL; | 277 | xfer->result = -EINVAL; |
295 | WARN_ON(1); | ||
296 | goto out; | 278 | goto out; |
297 | } | 279 | } |
298 | } | 280 | } |
299 | xfer->result = 0; | 281 | xfer->result = 0; |
300 | out: | 282 | out: |
301 | d_fnend(3, NULL, "(xfer %p) = void\n", xfer); | ||
302 | return result; | 283 | return result; |
303 | } | 284 | } |
304 | 285 | ||
@@ -424,8 +405,6 @@ static ssize_t __wa_xfer_setup_sizes(struct wa_xfer *xfer, | |||
424 | struct urb *urb = xfer->urb; | 405 | struct urb *urb = xfer->urb; |
425 | struct wa_rpipe *rpipe = xfer->ep->hcpriv; | 406 | struct wa_rpipe *rpipe = xfer->ep->hcpriv; |
426 | 407 | ||
427 | d_fnstart(3, dev, "(xfer %p [rpipe %p] urb %p)\n", | ||
428 | xfer, rpipe, urb); | ||
429 | switch (rpipe->descr.bmAttribute & 0x3) { | 408 | switch (rpipe->descr.bmAttribute & 0x3) { |
430 | case USB_ENDPOINT_XFER_CONTROL: | 409 | case USB_ENDPOINT_XFER_CONTROL: |
431 | *pxfer_type = WA_XFER_TYPE_CTL; | 410 | *pxfer_type = WA_XFER_TYPE_CTL; |
@@ -472,12 +451,10 @@ static ssize_t __wa_xfer_setup_sizes(struct wa_xfer *xfer, | |||
472 | if (xfer->segs == 0 && *pxfer_type == WA_XFER_TYPE_CTL) | 451 | if (xfer->segs == 0 && *pxfer_type == WA_XFER_TYPE_CTL) |
473 | xfer->segs = 1; | 452 | xfer->segs = 1; |
474 | error: | 453 | error: |
475 | d_fnend(3, dev, "(xfer %p [rpipe %p] urb %p) = %d\n", | ||
476 | xfer, rpipe, urb, (int)result); | ||
477 | return result; | 454 | return result; |
478 | } | 455 | } |
479 | 456 | ||
480 | /** Fill in the common request header and xfer-type specific data. */ | 457 | /* Fill in the common request header and xfer-type specific data. */ |
481 | static void __wa_xfer_setup_hdr0(struct wa_xfer *xfer, | 458 | static void __wa_xfer_setup_hdr0(struct wa_xfer *xfer, |
482 | struct wa_xfer_hdr *xfer_hdr0, | 459 | struct wa_xfer_hdr *xfer_hdr0, |
483 | enum wa_xfer_type xfer_type, | 460 | enum wa_xfer_type xfer_type, |
@@ -534,14 +511,13 @@ static void wa_seg_dto_cb(struct urb *urb) | |||
534 | unsigned rpipe_ready = 0; | 511 | unsigned rpipe_ready = 0; |
535 | u8 done = 0; | 512 | u8 done = 0; |
536 | 513 | ||
537 | d_fnstart(3, NULL, "(urb %p [%d])\n", urb, urb->status); | ||
538 | switch (urb->status) { | 514 | switch (urb->status) { |
539 | case 0: | 515 | case 0: |
540 | spin_lock_irqsave(&xfer->lock, flags); | 516 | spin_lock_irqsave(&xfer->lock, flags); |
541 | wa = xfer->wa; | 517 | wa = xfer->wa; |
542 | dev = &wa->usb_iface->dev; | 518 | dev = &wa->usb_iface->dev; |
543 | d_printf(2, dev, "xfer %p#%u: data out done (%d bytes)\n", | 519 | dev_dbg(dev, "xfer %p#%u: data out done (%d bytes)\n", |
544 | xfer, seg->index, urb->actual_length); | 520 | xfer, seg->index, urb->actual_length); |
545 | if (seg->status < WA_SEG_PENDING) | 521 | if (seg->status < WA_SEG_PENDING) |
546 | seg->status = WA_SEG_PENDING; | 522 | seg->status = WA_SEG_PENDING; |
547 | seg->result = urb->actual_length; | 523 | seg->result = urb->actual_length; |
@@ -555,9 +531,8 @@ static void wa_seg_dto_cb(struct urb *urb) | |||
555 | wa = xfer->wa; | 531 | wa = xfer->wa; |
556 | dev = &wa->usb_iface->dev; | 532 | dev = &wa->usb_iface->dev; |
557 | rpipe = xfer->ep->hcpriv; | 533 | rpipe = xfer->ep->hcpriv; |
558 | if (printk_ratelimit()) | 534 | dev_dbg(dev, "xfer %p#%u: data out error %d\n", |
559 | dev_err(dev, "xfer %p#%u: data out error %d\n", | 535 | xfer, seg->index, urb->status); |
560 | xfer, seg->index, urb->status); | ||
561 | if (edc_inc(&wa->nep_edc, EDC_MAX_ERRORS, | 536 | if (edc_inc(&wa->nep_edc, EDC_MAX_ERRORS, |
562 | EDC_ERROR_TIMEFRAME)){ | 537 | EDC_ERROR_TIMEFRAME)){ |
563 | dev_err(dev, "DTO: URB max acceptable errors " | 538 | dev_err(dev, "DTO: URB max acceptable errors " |
@@ -578,7 +553,6 @@ static void wa_seg_dto_cb(struct urb *urb) | |||
578 | if (rpipe_ready) | 553 | if (rpipe_ready) |
579 | wa_xfer_delayed_run(rpipe); | 554 | wa_xfer_delayed_run(rpipe); |
580 | } | 555 | } |
581 | d_fnend(3, NULL, "(urb %p [%d]) = void\n", urb, urb->status); | ||
582 | } | 556 | } |
583 | 557 | ||
584 | /* | 558 | /* |
@@ -610,14 +584,12 @@ static void wa_seg_cb(struct urb *urb) | |||
610 | unsigned rpipe_ready; | 584 | unsigned rpipe_ready; |
611 | u8 done = 0; | 585 | u8 done = 0; |
612 | 586 | ||
613 | d_fnstart(3, NULL, "(urb %p [%d])\n", urb, urb->status); | ||
614 | switch (urb->status) { | 587 | switch (urb->status) { |
615 | case 0: | 588 | case 0: |
616 | spin_lock_irqsave(&xfer->lock, flags); | 589 | spin_lock_irqsave(&xfer->lock, flags); |
617 | wa = xfer->wa; | 590 | wa = xfer->wa; |
618 | dev = &wa->usb_iface->dev; | 591 | dev = &wa->usb_iface->dev; |
619 | d_printf(2, dev, "xfer %p#%u: request done\n", | 592 | dev_dbg(dev, "xfer %p#%u: request done\n", xfer, seg->index); |
620 | xfer, seg->index); | ||
621 | if (xfer->is_inbound && seg->status < WA_SEG_PENDING) | 593 | if (xfer->is_inbound && seg->status < WA_SEG_PENDING) |
622 | seg->status = WA_SEG_PENDING; | 594 | seg->status = WA_SEG_PENDING; |
623 | spin_unlock_irqrestore(&xfer->lock, flags); | 595 | spin_unlock_irqrestore(&xfer->lock, flags); |
@@ -652,7 +624,6 @@ static void wa_seg_cb(struct urb *urb) | |||
652 | if (rpipe_ready) | 624 | if (rpipe_ready) |
653 | wa_xfer_delayed_run(rpipe); | 625 | wa_xfer_delayed_run(rpipe); |
654 | } | 626 | } |
655 | d_fnend(3, NULL, "(urb %p [%d]) = void\n", urb, urb->status); | ||
656 | } | 627 | } |
657 | 628 | ||
658 | /* | 629 | /* |
@@ -750,9 +721,6 @@ static int __wa_xfer_setup(struct wa_xfer *xfer, struct urb *urb) | |||
750 | size_t xfer_hdr_size, cnt, transfer_size; | 721 | size_t xfer_hdr_size, cnt, transfer_size; |
751 | struct wa_xfer_hdr *xfer_hdr0, *xfer_hdr; | 722 | struct wa_xfer_hdr *xfer_hdr0, *xfer_hdr; |
752 | 723 | ||
753 | d_fnstart(3, dev, "(xfer %p [rpipe %p] urb %p)\n", | ||
754 | xfer, xfer->ep->hcpriv, urb); | ||
755 | |||
756 | result = __wa_xfer_setup_sizes(xfer, &xfer_type); | 724 | result = __wa_xfer_setup_sizes(xfer, &xfer_type); |
757 | if (result < 0) | 725 | if (result < 0) |
758 | goto error_setup_sizes; | 726 | goto error_setup_sizes; |
@@ -788,8 +756,6 @@ static int __wa_xfer_setup(struct wa_xfer *xfer, struct urb *urb) | |||
788 | result = 0; | 756 | result = 0; |
789 | error_setup_segs: | 757 | error_setup_segs: |
790 | error_setup_sizes: | 758 | error_setup_sizes: |
791 | d_fnend(3, dev, "(xfer %p [rpipe %p] urb %p) = %d\n", | ||
792 | xfer, xfer->ep->hcpriv, urb, result); | ||
793 | return result; | 759 | return result; |
794 | } | 760 | } |
795 | 761 | ||
@@ -843,9 +809,6 @@ static void wa_xfer_delayed_run(struct wa_rpipe *rpipe) | |||
843 | struct wa_xfer *xfer; | 809 | struct wa_xfer *xfer; |
844 | unsigned long flags; | 810 | unsigned long flags; |
845 | 811 | ||
846 | d_fnstart(1, dev, "(rpipe #%d) %d segments available\n", | ||
847 | le16_to_cpu(rpipe->descr.wRPipeIndex), | ||
848 | atomic_read(&rpipe->segs_available)); | ||
849 | spin_lock_irqsave(&rpipe->seg_lock, flags); | 812 | spin_lock_irqsave(&rpipe->seg_lock, flags); |
850 | while (atomic_read(&rpipe->segs_available) > 0 | 813 | while (atomic_read(&rpipe->segs_available) > 0 |
851 | && !list_empty(&rpipe->seg_list)) { | 814 | && !list_empty(&rpipe->seg_list)) { |
@@ -854,10 +817,8 @@ static void wa_xfer_delayed_run(struct wa_rpipe *rpipe) | |||
854 | list_del(&seg->list_node); | 817 | list_del(&seg->list_node); |
855 | xfer = seg->xfer; | 818 | xfer = seg->xfer; |
856 | result = __wa_seg_submit(rpipe, xfer, seg); | 819 | result = __wa_seg_submit(rpipe, xfer, seg); |
857 | d_printf(1, dev, "xfer %p#%u submitted from delayed " | 820 | dev_dbg(dev, "xfer %p#%u submitted from delayed [%d segments available] %d\n", |
858 | "[%d segments available] %d\n", | 821 | xfer, seg->index, atomic_read(&rpipe->segs_available), result); |
859 | xfer, seg->index, | ||
860 | atomic_read(&rpipe->segs_available), result); | ||
861 | if (unlikely(result < 0)) { | 822 | if (unlikely(result < 0)) { |
862 | spin_unlock_irqrestore(&rpipe->seg_lock, flags); | 823 | spin_unlock_irqrestore(&rpipe->seg_lock, flags); |
863 | spin_lock_irqsave(&xfer->lock, flags); | 824 | spin_lock_irqsave(&xfer->lock, flags); |
@@ -868,10 +829,6 @@ static void wa_xfer_delayed_run(struct wa_rpipe *rpipe) | |||
868 | } | 829 | } |
869 | } | 830 | } |
870 | spin_unlock_irqrestore(&rpipe->seg_lock, flags); | 831 | spin_unlock_irqrestore(&rpipe->seg_lock, flags); |
871 | d_fnend(1, dev, "(rpipe #%d) = void, %d segments available\n", | ||
872 | le16_to_cpu(rpipe->descr.wRPipeIndex), | ||
873 | atomic_read(&rpipe->segs_available)); | ||
874 | |||
875 | } | 832 | } |
876 | 833 | ||
877 | /* | 834 | /* |
@@ -894,9 +851,6 @@ static int __wa_xfer_submit(struct wa_xfer *xfer) | |||
894 | u8 available; | 851 | u8 available; |
895 | u8 empty; | 852 | u8 empty; |
896 | 853 | ||
897 | d_fnstart(3, dev, "(xfer %p [rpipe %p])\n", | ||
898 | xfer, xfer->ep->hcpriv); | ||
899 | |||
900 | spin_lock_irqsave(&wa->xfer_list_lock, flags); | 854 | spin_lock_irqsave(&wa->xfer_list_lock, flags); |
901 | list_add_tail(&xfer->list_node, &wa->xfer_list); | 855 | list_add_tail(&xfer->list_node, &wa->xfer_list); |
902 | spin_unlock_irqrestore(&wa->xfer_list_lock, flags); | 856 | spin_unlock_irqrestore(&wa->xfer_list_lock, flags); |
@@ -908,30 +862,24 @@ static int __wa_xfer_submit(struct wa_xfer *xfer) | |||
908 | available = atomic_read(&rpipe->segs_available); | 862 | available = atomic_read(&rpipe->segs_available); |
909 | empty = list_empty(&rpipe->seg_list); | 863 | empty = list_empty(&rpipe->seg_list); |
910 | seg = xfer->seg[cnt]; | 864 | seg = xfer->seg[cnt]; |
911 | d_printf(2, dev, "xfer %p#%u: available %u empty %u (%s)\n", | 865 | dev_dbg(dev, "xfer %p#%u: available %u empty %u (%s)\n", |
912 | xfer, cnt, available, empty, | 866 | xfer, cnt, available, empty, |
913 | available == 0 || !empty ? "delayed" : "submitted"); | 867 | available == 0 || !empty ? "delayed" : "submitted"); |
914 | if (available == 0 || !empty) { | 868 | if (available == 0 || !empty) { |
915 | d_printf(1, dev, "xfer %p#%u: delayed\n", xfer, cnt); | 869 | dev_dbg(dev, "xfer %p#%u: delayed\n", xfer, cnt); |
916 | seg->status = WA_SEG_DELAYED; | 870 | seg->status = WA_SEG_DELAYED; |
917 | list_add_tail(&seg->list_node, &rpipe->seg_list); | 871 | list_add_tail(&seg->list_node, &rpipe->seg_list); |
918 | } else { | 872 | } else { |
919 | result = __wa_seg_submit(rpipe, xfer, seg); | 873 | result = __wa_seg_submit(rpipe, xfer, seg); |
920 | if (result < 0) | 874 | if (result < 0) { |
875 | __wa_xfer_abort(xfer); | ||
921 | goto error_seg_submit; | 876 | goto error_seg_submit; |
877 | } | ||
922 | } | 878 | } |
923 | xfer->segs_submitted++; | 879 | xfer->segs_submitted++; |
924 | } | 880 | } |
925 | spin_unlock_irqrestore(&rpipe->seg_lock, flags); | ||
926 | d_fnend(3, dev, "(xfer %p [rpipe %p]) = void\n", xfer, | ||
927 | xfer->ep->hcpriv); | ||
928 | return result; | ||
929 | |||
930 | error_seg_submit: | 881 | error_seg_submit: |
931 | __wa_xfer_abort(xfer); | ||
932 | spin_unlock_irqrestore(&rpipe->seg_lock, flags); | 882 | spin_unlock_irqrestore(&rpipe->seg_lock, flags); |
933 | d_fnend(3, dev, "(xfer %p [rpipe %p]) = void\n", xfer, | ||
934 | xfer->ep->hcpriv); | ||
935 | return result; | 883 | return result; |
936 | } | 884 | } |
937 | 885 | ||
@@ -964,11 +912,9 @@ static void wa_urb_enqueue_b(struct wa_xfer *xfer) | |||
964 | struct urb *urb = xfer->urb; | 912 | struct urb *urb = xfer->urb; |
965 | struct wahc *wa = xfer->wa; | 913 | struct wahc *wa = xfer->wa; |
966 | struct wusbhc *wusbhc = wa->wusb; | 914 | struct wusbhc *wusbhc = wa->wusb; |
967 | struct device *dev = &wa->usb_iface->dev; | ||
968 | struct wusb_dev *wusb_dev; | 915 | struct wusb_dev *wusb_dev; |
969 | unsigned done; | 916 | unsigned done; |
970 | 917 | ||
971 | d_fnstart(3, dev, "(wa %p urb %p)\n", wa, urb); | ||
972 | result = rpipe_get_by_ep(wa, xfer->ep, urb, xfer->gfp); | 918 | result = rpipe_get_by_ep(wa, xfer->ep, urb, xfer->gfp); |
973 | if (result < 0) | 919 | if (result < 0) |
974 | goto error_rpipe_get; | 920 | goto error_rpipe_get; |
@@ -997,7 +943,6 @@ static void wa_urb_enqueue_b(struct wa_xfer *xfer) | |||
997 | if (result < 0) | 943 | if (result < 0) |
998 | goto error_xfer_submit; | 944 | goto error_xfer_submit; |
999 | spin_unlock_irqrestore(&xfer->lock, flags); | 945 | spin_unlock_irqrestore(&xfer->lock, flags); |
1000 | d_fnend(3, dev, "(wa %p urb %p) = void\n", wa, urb); | ||
1001 | return; | 946 | return; |
1002 | 947 | ||
1003 | /* this is basically wa_xfer_completion() broken up wa_xfer_giveback() | 948 | /* this is basically wa_xfer_completion() broken up wa_xfer_giveback() |
@@ -1015,7 +960,6 @@ error_dev_gone: | |||
1015 | error_rpipe_get: | 960 | error_rpipe_get: |
1016 | xfer->result = result; | 961 | xfer->result = result; |
1017 | wa_xfer_giveback(xfer); | 962 | wa_xfer_giveback(xfer); |
1018 | d_fnend(3, dev, "(wa %p urb %p) = (void) %d\n", wa, urb, result); | ||
1019 | return; | 963 | return; |
1020 | 964 | ||
1021 | error_xfer_submit: | 965 | error_xfer_submit: |
@@ -1024,8 +968,6 @@ error_xfer_submit: | |||
1024 | spin_unlock_irqrestore(&xfer->lock, flags); | 968 | spin_unlock_irqrestore(&xfer->lock, flags); |
1025 | if (done) | 969 | if (done) |
1026 | wa_xfer_completion(xfer); | 970 | wa_xfer_completion(xfer); |
1027 | d_fnend(3, dev, "(wa %p urb %p) = (void) %d\n", wa, urb, result); | ||
1028 | return; | ||
1029 | } | 971 | } |
1030 | 972 | ||
1031 | /* | 973 | /* |
@@ -1041,11 +983,9 @@ error_xfer_submit: | |||
1041 | void wa_urb_enqueue_run(struct work_struct *ws) | 983 | void wa_urb_enqueue_run(struct work_struct *ws) |
1042 | { | 984 | { |
1043 | struct wahc *wa = container_of(ws, struct wahc, xfer_work); | 985 | struct wahc *wa = container_of(ws, struct wahc, xfer_work); |
1044 | struct device *dev = &wa->usb_iface->dev; | ||
1045 | struct wa_xfer *xfer, *next; | 986 | struct wa_xfer *xfer, *next; |
1046 | struct urb *urb; | 987 | struct urb *urb; |
1047 | 988 | ||
1048 | d_fnstart(3, dev, "(wa %p)\n", wa); | ||
1049 | spin_lock_irq(&wa->xfer_list_lock); | 989 | spin_lock_irq(&wa->xfer_list_lock); |
1050 | list_for_each_entry_safe(xfer, next, &wa->xfer_delayed_list, | 990 | list_for_each_entry_safe(xfer, next, &wa->xfer_delayed_list, |
1051 | list_node) { | 991 | list_node) { |
@@ -1059,7 +999,6 @@ void wa_urb_enqueue_run(struct work_struct *ws) | |||
1059 | spin_lock_irq(&wa->xfer_list_lock); | 999 | spin_lock_irq(&wa->xfer_list_lock); |
1060 | } | 1000 | } |
1061 | spin_unlock_irq(&wa->xfer_list_lock); | 1001 | spin_unlock_irq(&wa->xfer_list_lock); |
1062 | d_fnend(3, dev, "(wa %p) = void\n", wa); | ||
1063 | } | 1002 | } |
1064 | EXPORT_SYMBOL_GPL(wa_urb_enqueue_run); | 1003 | EXPORT_SYMBOL_GPL(wa_urb_enqueue_run); |
1065 | 1004 | ||
@@ -1084,9 +1023,6 @@ int wa_urb_enqueue(struct wahc *wa, struct usb_host_endpoint *ep, | |||
1084 | unsigned long my_flags; | 1023 | unsigned long my_flags; |
1085 | unsigned cant_sleep = irqs_disabled() | in_atomic(); | 1024 | unsigned cant_sleep = irqs_disabled() | in_atomic(); |
1086 | 1025 | ||
1087 | d_fnstart(3, dev, "(wa %p ep %p urb %p [%d] gfp 0x%x)\n", | ||
1088 | wa, ep, urb, urb->transfer_buffer_length, gfp); | ||
1089 | |||
1090 | if (urb->transfer_buffer == NULL | 1026 | if (urb->transfer_buffer == NULL |
1091 | && !(urb->transfer_flags & URB_NO_TRANSFER_DMA_MAP) | 1027 | && !(urb->transfer_flags & URB_NO_TRANSFER_DMA_MAP) |
1092 | && urb->transfer_buffer_length != 0) { | 1028 | && urb->transfer_buffer_length != 0) { |
@@ -1108,11 +1044,13 @@ int wa_urb_enqueue(struct wahc *wa, struct usb_host_endpoint *ep, | |||
1108 | xfer->gfp = gfp; | 1044 | xfer->gfp = gfp; |
1109 | xfer->ep = ep; | 1045 | xfer->ep = ep; |
1110 | urb->hcpriv = xfer; | 1046 | urb->hcpriv = xfer; |
1111 | d_printf(2, dev, "xfer %p urb %p pipe 0x%02x [%d bytes] %s %s %s\n", | 1047 | |
1112 | xfer, urb, urb->pipe, urb->transfer_buffer_length, | 1048 | dev_dbg(dev, "xfer %p urb %p pipe 0x%02x [%d bytes] %s %s %s\n", |
1113 | urb->transfer_flags & URB_NO_TRANSFER_DMA_MAP ? "dma" : "nodma", | 1049 | xfer, urb, urb->pipe, urb->transfer_buffer_length, |
1114 | urb->pipe & USB_DIR_IN ? "inbound" : "outbound", | 1050 | urb->transfer_flags & URB_NO_TRANSFER_DMA_MAP ? "dma" : "nodma", |
1115 | cant_sleep ? "deferred" : "inline"); | 1051 | urb->pipe & USB_DIR_IN ? "inbound" : "outbound", |
1052 | cant_sleep ? "deferred" : "inline"); | ||
1053 | |||
1116 | if (cant_sleep) { | 1054 | if (cant_sleep) { |
1117 | usb_get_urb(urb); | 1055 | usb_get_urb(urb); |
1118 | spin_lock_irqsave(&wa->xfer_list_lock, my_flags); | 1056 | spin_lock_irqsave(&wa->xfer_list_lock, my_flags); |
@@ -1122,15 +1060,11 @@ int wa_urb_enqueue(struct wahc *wa, struct usb_host_endpoint *ep, | |||
1122 | } else { | 1060 | } else { |
1123 | wa_urb_enqueue_b(xfer); | 1061 | wa_urb_enqueue_b(xfer); |
1124 | } | 1062 | } |
1125 | d_fnend(3, dev, "(wa %p ep %p urb %p [%d] gfp 0x%x) = 0\n", | ||
1126 | wa, ep, urb, urb->transfer_buffer_length, gfp); | ||
1127 | return 0; | 1063 | return 0; |
1128 | 1064 | ||
1129 | error_dequeued: | 1065 | error_dequeued: |
1130 | kfree(xfer); | 1066 | kfree(xfer); |
1131 | error_kmalloc: | 1067 | error_kmalloc: |
1132 | d_fnend(3, dev, "(wa %p ep %p urb %p [%d] gfp 0x%x) = %d\n", | ||
1133 | wa, ep, urb, urb->transfer_buffer_length, gfp, result); | ||
1134 | return result; | 1068 | return result; |
1135 | } | 1069 | } |
1136 | EXPORT_SYMBOL_GPL(wa_urb_enqueue); | 1070 | EXPORT_SYMBOL_GPL(wa_urb_enqueue); |
@@ -1155,7 +1089,6 @@ EXPORT_SYMBOL_GPL(wa_urb_enqueue); | |||
1155 | */ | 1089 | */ |
1156 | int wa_urb_dequeue(struct wahc *wa, struct urb *urb) | 1090 | int wa_urb_dequeue(struct wahc *wa, struct urb *urb) |
1157 | { | 1091 | { |
1158 | struct device *dev = &wa->usb_iface->dev; | ||
1159 | unsigned long flags, flags2; | 1092 | unsigned long flags, flags2; |
1160 | struct wa_xfer *xfer; | 1093 | struct wa_xfer *xfer; |
1161 | struct wa_seg *seg; | 1094 | struct wa_seg *seg; |
@@ -1163,9 +1096,6 @@ int wa_urb_dequeue(struct wahc *wa, struct urb *urb) | |||
1163 | unsigned cnt; | 1096 | unsigned cnt; |
1164 | unsigned rpipe_ready = 0; | 1097 | unsigned rpipe_ready = 0; |
1165 | 1098 | ||
1166 | d_fnstart(3, dev, "(wa %p, urb %p)\n", wa, urb); | ||
1167 | |||
1168 | d_printf(1, dev, "xfer %p urb %p: aborting\n", urb->hcpriv, urb); | ||
1169 | xfer = urb->hcpriv; | 1099 | xfer = urb->hcpriv; |
1170 | if (xfer == NULL) { | 1100 | if (xfer == NULL) { |
1171 | /* NOthing setup yet enqueue will see urb->status != | 1101 | /* NOthing setup yet enqueue will see urb->status != |
@@ -1234,13 +1164,11 @@ int wa_urb_dequeue(struct wahc *wa, struct urb *urb) | |||
1234 | wa_xfer_completion(xfer); | 1164 | wa_xfer_completion(xfer); |
1235 | if (rpipe_ready) | 1165 | if (rpipe_ready) |
1236 | wa_xfer_delayed_run(rpipe); | 1166 | wa_xfer_delayed_run(rpipe); |
1237 | d_fnend(3, dev, "(wa %p, urb %p) = 0\n", wa, urb); | ||
1238 | return 0; | 1167 | return 0; |
1239 | 1168 | ||
1240 | out_unlock: | 1169 | out_unlock: |
1241 | spin_unlock_irqrestore(&xfer->lock, flags); | 1170 | spin_unlock_irqrestore(&xfer->lock, flags); |
1242 | out: | 1171 | out: |
1243 | d_fnend(3, dev, "(wa %p, urb %p) = 0\n", wa, urb); | ||
1244 | return 0; | 1172 | return 0; |
1245 | 1173 | ||
1246 | dequeue_delayed: | 1174 | dequeue_delayed: |
@@ -1250,7 +1178,6 @@ dequeue_delayed: | |||
1250 | spin_unlock_irqrestore(&xfer->lock, flags); | 1178 | spin_unlock_irqrestore(&xfer->lock, flags); |
1251 | wa_xfer_giveback(xfer); | 1179 | wa_xfer_giveback(xfer); |
1252 | usb_put_urb(urb); /* we got a ref in enqueue() */ | 1180 | usb_put_urb(urb); /* we got a ref in enqueue() */ |
1253 | d_fnend(3, dev, "(wa %p, urb %p) = 0\n", wa, urb); | ||
1254 | return 0; | 1181 | return 0; |
1255 | } | 1182 | } |
1256 | EXPORT_SYMBOL_GPL(wa_urb_dequeue); | 1183 | EXPORT_SYMBOL_GPL(wa_urb_dequeue); |
@@ -1326,7 +1253,6 @@ static void wa_xfer_result_chew(struct wahc *wa, struct wa_xfer *xfer) | |||
1326 | u8 usb_status; | 1253 | u8 usb_status; |
1327 | unsigned rpipe_ready = 0; | 1254 | unsigned rpipe_ready = 0; |
1328 | 1255 | ||
1329 | d_fnstart(3, dev, "(wa %p xfer %p)\n", wa, xfer); | ||
1330 | spin_lock_irqsave(&xfer->lock, flags); | 1256 | spin_lock_irqsave(&xfer->lock, flags); |
1331 | seg_idx = xfer_result->bTransferSegment & 0x7f; | 1257 | seg_idx = xfer_result->bTransferSegment & 0x7f; |
1332 | if (unlikely(seg_idx >= xfer->segs)) | 1258 | if (unlikely(seg_idx >= xfer->segs)) |
@@ -1334,8 +1260,8 @@ static void wa_xfer_result_chew(struct wahc *wa, struct wa_xfer *xfer) | |||
1334 | seg = xfer->seg[seg_idx]; | 1260 | seg = xfer->seg[seg_idx]; |
1335 | rpipe = xfer->ep->hcpriv; | 1261 | rpipe = xfer->ep->hcpriv; |
1336 | usb_status = xfer_result->bTransferStatus; | 1262 | usb_status = xfer_result->bTransferStatus; |
1337 | d_printf(2, dev, "xfer %p#%u: bTransferStatus 0x%02x (seg %u)\n", | 1263 | dev_dbg(dev, "xfer %p#%u: bTransferStatus 0x%02x (seg %u)\n", |
1338 | xfer, seg_idx, usb_status, seg->status); | 1264 | xfer, seg_idx, usb_status, seg->status); |
1339 | if (seg->status == WA_SEG_ABORTED | 1265 | if (seg->status == WA_SEG_ABORTED |
1340 | || seg->status == WA_SEG_ERROR) /* already handled */ | 1266 | || seg->status == WA_SEG_ERROR) /* already handled */ |
1341 | goto segment_aborted; | 1267 | goto segment_aborted; |
@@ -1391,10 +1317,8 @@ static void wa_xfer_result_chew(struct wahc *wa, struct wa_xfer *xfer) | |||
1391 | wa_xfer_completion(xfer); | 1317 | wa_xfer_completion(xfer); |
1392 | if (rpipe_ready) | 1318 | if (rpipe_ready) |
1393 | wa_xfer_delayed_run(rpipe); | 1319 | wa_xfer_delayed_run(rpipe); |
1394 | d_fnend(3, dev, "(wa %p xfer %p) = void\n", wa, xfer); | ||
1395 | return; | 1320 | return; |
1396 | 1321 | ||
1397 | |||
1398 | error_submit_buf_in: | 1322 | error_submit_buf_in: |
1399 | if (edc_inc(&wa->dti_edc, EDC_MAX_ERRORS, EDC_ERROR_TIMEFRAME)) { | 1323 | if (edc_inc(&wa->dti_edc, EDC_MAX_ERRORS, EDC_ERROR_TIMEFRAME)) { |
1400 | dev_err(dev, "DTI: URB max acceptable errors " | 1324 | dev_err(dev, "DTI: URB max acceptable errors " |
@@ -1416,11 +1340,8 @@ error_complete: | |||
1416 | wa_xfer_completion(xfer); | 1340 | wa_xfer_completion(xfer); |
1417 | if (rpipe_ready) | 1341 | if (rpipe_ready) |
1418 | wa_xfer_delayed_run(rpipe); | 1342 | wa_xfer_delayed_run(rpipe); |
1419 | d_fnend(3, dev, "(wa %p xfer %p) = void [segment/DTI-submit error]\n", | ||
1420 | wa, xfer); | ||
1421 | return; | 1343 | return; |
1422 | 1344 | ||
1423 | |||
1424 | error_bad_seg: | 1345 | error_bad_seg: |
1425 | spin_unlock_irqrestore(&xfer->lock, flags); | 1346 | spin_unlock_irqrestore(&xfer->lock, flags); |
1426 | wa_urb_dequeue(wa, xfer->urb); | 1347 | wa_urb_dequeue(wa, xfer->urb); |
@@ -1431,17 +1352,11 @@ error_bad_seg: | |||
1431 | "exceeded, resetting device\n"); | 1352 | "exceeded, resetting device\n"); |
1432 | wa_reset_all(wa); | 1353 | wa_reset_all(wa); |
1433 | } | 1354 | } |
1434 | d_fnend(3, dev, "(wa %p xfer %p) = void [bad seg]\n", wa, xfer); | ||
1435 | return; | 1355 | return; |
1436 | 1356 | ||
1437 | |||
1438 | segment_aborted: | 1357 | segment_aborted: |
1439 | /* nothing to do, as the aborter did the completion */ | 1358 | /* nothing to do, as the aborter did the completion */ |
1440 | spin_unlock_irqrestore(&xfer->lock, flags); | 1359 | spin_unlock_irqrestore(&xfer->lock, flags); |
1441 | d_fnend(3, dev, "(wa %p xfer %p) = void [segment aborted]\n", | ||
1442 | wa, xfer); | ||
1443 | return; | ||
1444 | |||
1445 | } | 1360 | } |
1446 | 1361 | ||
1447 | /* | 1362 | /* |
@@ -1465,15 +1380,14 @@ static void wa_buf_in_cb(struct urb *urb) | |||
1465 | unsigned long flags; | 1380 | unsigned long flags; |
1466 | u8 done = 0; | 1381 | u8 done = 0; |
1467 | 1382 | ||
1468 | d_fnstart(3, NULL, "(urb %p [%d])\n", urb, urb->status); | ||
1469 | switch (urb->status) { | 1383 | switch (urb->status) { |
1470 | case 0: | 1384 | case 0: |
1471 | spin_lock_irqsave(&xfer->lock, flags); | 1385 | spin_lock_irqsave(&xfer->lock, flags); |
1472 | wa = xfer->wa; | 1386 | wa = xfer->wa; |
1473 | dev = &wa->usb_iface->dev; | 1387 | dev = &wa->usb_iface->dev; |
1474 | rpipe = xfer->ep->hcpriv; | 1388 | rpipe = xfer->ep->hcpriv; |
1475 | d_printf(2, dev, "xfer %p#%u: data in done (%zu bytes)\n", | 1389 | dev_dbg(dev, "xfer %p#%u: data in done (%zu bytes)\n", |
1476 | xfer, seg->index, (size_t)urb->actual_length); | 1390 | xfer, seg->index, (size_t)urb->actual_length); |
1477 | seg->status = WA_SEG_DONE; | 1391 | seg->status = WA_SEG_DONE; |
1478 | seg->result = urb->actual_length; | 1392 | seg->result = urb->actual_length; |
1479 | xfer->segs_done++; | 1393 | xfer->segs_done++; |
@@ -1514,7 +1428,6 @@ static void wa_buf_in_cb(struct urb *urb) | |||
1514 | if (rpipe_ready) | 1428 | if (rpipe_ready) |
1515 | wa_xfer_delayed_run(rpipe); | 1429 | wa_xfer_delayed_run(rpipe); |
1516 | } | 1430 | } |
1517 | d_fnend(3, NULL, "(urb %p [%d]) = void\n", urb, urb->status); | ||
1518 | } | 1431 | } |
1519 | 1432 | ||
1520 | /* | 1433 | /* |
@@ -1553,14 +1466,12 @@ static void wa_xfer_result_cb(struct urb *urb) | |||
1553 | struct wa_xfer *xfer; | 1466 | struct wa_xfer *xfer; |
1554 | u8 usb_status; | 1467 | u8 usb_status; |
1555 | 1468 | ||
1556 | d_fnstart(3, dev, "(%p)\n", wa); | ||
1557 | BUG_ON(wa->dti_urb != urb); | 1469 | BUG_ON(wa->dti_urb != urb); |
1558 | switch (wa->dti_urb->status) { | 1470 | switch (wa->dti_urb->status) { |
1559 | case 0: | 1471 | case 0: |
1560 | /* We have a xfer result buffer; check it */ | 1472 | /* We have a xfer result buffer; check it */ |
1561 | d_printf(2, dev, "DTI: xfer result %d bytes at %p\n", | 1473 | dev_dbg(dev, "DTI: xfer result %d bytes at %p\n", |
1562 | urb->actual_length, urb->transfer_buffer); | 1474 | urb->actual_length, urb->transfer_buffer); |
1563 | d_dump(3, dev, urb->transfer_buffer, urb->actual_length); | ||
1564 | if (wa->dti_urb->actual_length != sizeof(*xfer_result)) { | 1475 | if (wa->dti_urb->actual_length != sizeof(*xfer_result)) { |
1565 | dev_err(dev, "DTI Error: xfer result--bad size " | 1476 | dev_err(dev, "DTI Error: xfer result--bad size " |
1566 | "xfer result (%d bytes vs %zu needed)\n", | 1477 | "xfer result (%d bytes vs %zu needed)\n", |
@@ -1622,7 +1533,6 @@ static void wa_xfer_result_cb(struct urb *urb) | |||
1622 | wa_reset_all(wa); | 1533 | wa_reset_all(wa); |
1623 | } | 1534 | } |
1624 | out: | 1535 | out: |
1625 | d_fnend(3, dev, "(%p) = void\n", wa); | ||
1626 | return; | 1536 | return; |
1627 | } | 1537 | } |
1628 | 1538 | ||
@@ -1653,7 +1563,6 @@ void wa_handle_notif_xfer(struct wahc *wa, struct wa_notif_hdr *notif_hdr) | |||
1653 | struct wa_notif_xfer *notif_xfer; | 1563 | struct wa_notif_xfer *notif_xfer; |
1654 | const struct usb_endpoint_descriptor *dti_epd = wa->dti_epd; | 1564 | const struct usb_endpoint_descriptor *dti_epd = wa->dti_epd; |
1655 | 1565 | ||
1656 | d_fnstart(4, dev, "(%p, %p)\n", wa, notif_hdr); | ||
1657 | notif_xfer = container_of(notif_hdr, struct wa_notif_xfer, hdr); | 1566 | notif_xfer = container_of(notif_hdr, struct wa_notif_xfer, hdr); |
1658 | BUG_ON(notif_hdr->bNotifyType != WA_NOTIF_TRANSFER); | 1567 | BUG_ON(notif_hdr->bNotifyType != WA_NOTIF_TRANSFER); |
1659 | 1568 | ||
@@ -1693,7 +1602,6 @@ void wa_handle_notif_xfer(struct wahc *wa, struct wa_notif_hdr *notif_hdr) | |||
1693 | goto error_dti_urb_submit; | 1602 | goto error_dti_urb_submit; |
1694 | } | 1603 | } |
1695 | out: | 1604 | out: |
1696 | d_fnend(4, dev, "(%p, %p) = void\n", wa, notif_hdr); | ||
1697 | return; | 1605 | return; |
1698 | 1606 | ||
1699 | error_dti_urb_submit: | 1607 | error_dti_urb_submit: |
@@ -1704,6 +1612,4 @@ error_buf_in_urb_alloc: | |||
1704 | error_dti_urb_alloc: | 1612 | error_dti_urb_alloc: |
1705 | error: | 1613 | error: |
1706 | wa_reset_all(wa); | 1614 | wa_reset_all(wa); |
1707 | d_fnend(4, dev, "(%p, %p) = void\n", wa, notif_hdr); | ||
1708 | return; | ||
1709 | } | 1615 | } |