aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorOliver Neukum <oliver@neukum.org>2009-09-09 11:06:53 -0400
committerGreg Kroah-Hartman <gregkh@suse.de>2009-09-23 09:46:40 -0400
commite7389cc9a7ff7c6e760e741c81a751c834f7d145 (patch)
treec32bb55858fd792ac10b90a2a0539d163ad46136
parentb356b7c7696b289dda99022d71e3979c6134af52 (diff)
USB: skel_read really sucks royally
The read code path of the skeleton driver really sucks - skel_read works only for devices which always send data - the timeout comes out of thin air - it blocks signals for the duration of the timeout - it disallows nonblocking IO by design This patch fixes it by using a real urb, a completion and interruptible waits. Signed-off-by: Oliver Neukum <oliver@neukum.org> Signed-off-by: Greg Kroah-Hartman <gregkh@suse.de>
-rw-r--r--drivers/usb/usb-skeleton.c194
1 files changed, 176 insertions, 18 deletions
diff --git a/drivers/usb/usb-skeleton.c b/drivers/usb/usb-skeleton.c
index 60ba631e99c2..5ffa3e246856 100644
--- a/drivers/usb/usb-skeleton.c
+++ b/drivers/usb/usb-skeleton.c
@@ -52,15 +52,21 @@ struct usb_skel {
52 struct usb_interface *interface; /* the interface for this device */ 52 struct usb_interface *interface; /* the interface for this device */
53 struct semaphore limit_sem; /* limiting the number of writes in progress */ 53 struct semaphore limit_sem; /* limiting the number of writes in progress */
54 struct usb_anchor submitted; /* in case we need to retract our submissions */ 54 struct usb_anchor submitted; /* in case we need to retract our submissions */
55 struct urb *bulk_in_urb; /* the urb to read data with */
55 unsigned char *bulk_in_buffer; /* the buffer to receive data */ 56 unsigned char *bulk_in_buffer; /* the buffer to receive data */
56 size_t bulk_in_size; /* the size of the receive buffer */ 57 size_t bulk_in_size; /* the size of the receive buffer */
58 size_t bulk_in_filled; /* number of bytes in the buffer */
59 size_t bulk_in_copied; /* already copied to user space */
57 __u8 bulk_in_endpointAddr; /* the address of the bulk in endpoint */ 60 __u8 bulk_in_endpointAddr; /* the address of the bulk in endpoint */
58 __u8 bulk_out_endpointAddr; /* the address of the bulk out endpoint */ 61 __u8 bulk_out_endpointAddr; /* the address of the bulk out endpoint */
59 int errors; /* the last request tanked */ 62 int errors; /* the last request tanked */
60 int open_count; /* count the number of openers */ 63 int open_count; /* count the number of openers */
64 bool ongoing_read; /* a read is going on */
65 bool processed_urb; /* indicates we haven't processed the urb */
61 spinlock_t err_lock; /* lock for errors */ 66 spinlock_t err_lock; /* lock for errors */
62 struct kref kref; 67 struct kref kref;
63 struct mutex io_mutex; /* synchronize I/O with disconnect */ 68 struct mutex io_mutex; /* synchronize I/O with disconnect */
69 struct completion bulk_in_completion; /* to wait for an ongoing read */
64}; 70};
65#define to_skel_dev(d) container_of(d, struct usb_skel, kref) 71#define to_skel_dev(d) container_of(d, struct usb_skel, kref)
66 72
@@ -71,6 +77,7 @@ static void skel_delete(struct kref *kref)
71{ 77{
72 struct usb_skel *dev = to_skel_dev(kref); 78 struct usb_skel *dev = to_skel_dev(kref);
73 79
80 usb_free_urb(dev->bulk_in_urb);
74 usb_put_dev(dev->udev); 81 usb_put_dev(dev->udev);
75 kfree(dev->bulk_in_buffer); 82 kfree(dev->bulk_in_buffer);
76 kfree(dev); 83 kfree(dev);
@@ -174,38 +181,182 @@ static int skel_flush(struct file *file, fl_owner_t id)
174 return res; 181 return res;
175} 182}
176 183
184static void skel_read_bulk_callback(struct urb *urb)
185{
186 struct usb_skel *dev;
187
188 dev = urb->context;
189
190 spin_lock(&dev->err_lock);
191 /* sync/async unlink faults aren't errors */
192 if (urb->status) {
193 if(!(urb->status == -ENOENT ||
194 urb->status == -ECONNRESET ||
195 urb->status == -ESHUTDOWN))
196 err("%s - nonzero write bulk status received: %d",
197 __func__, urb->status);
198
199 dev->errors = urb->status;
200 } else {
201 dev->bulk_in_filled = urb->actual_length;
202 }
203 dev->ongoing_read = 0;
204 spin_unlock(&dev->err_lock);
205
206 complete(&dev->bulk_in_completion);
207}
208
209static int skel_do_read_io(struct usb_skel *dev, size_t count)
210{
211 int rv;
212
213 /* prepare a read */
214 usb_fill_bulk_urb(dev->bulk_in_urb,
215 dev->udev,
216 usb_rcvbulkpipe(dev->udev,
217 dev->bulk_in_endpointAddr),
218 dev->bulk_in_buffer,
219 min(dev->bulk_in_size, count),
220 skel_read_bulk_callback,
221 dev);
222 /* tell everybody to leave the URB alone */
223 spin_lock_irq(&dev->err_lock);
224 dev->ongoing_read = 1;
225 spin_unlock_irq(&dev->err_lock);
226
227 /* do it */
228 rv = usb_submit_urb(dev->bulk_in_urb, GFP_KERNEL);
229 if (rv < 0) {
230 err("%s - failed submitting read urb, error %d",
231 __func__, rv);
232 dev->bulk_in_filled = 0;
233 rv = (rv == -ENOMEM) ? rv : -EIO;
234 spin_lock_irq(&dev->err_lock);
235 dev->ongoing_read = 0;
236 spin_unlock_irq(&dev->err_lock);
237 }
238
239 return rv;
240}
241
177static ssize_t skel_read(struct file *file, char *buffer, size_t count, loff_t *ppos) 242static ssize_t skel_read(struct file *file, char *buffer, size_t count, loff_t *ppos)
178{ 243{
179 struct usb_skel *dev; 244 struct usb_skel *dev;
180 int retval; 245 int rv;
181 int bytes_read; 246 bool ongoing_io;
182 247
183 dev = (struct usb_skel *)file->private_data; 248 dev = (struct usb_skel *)file->private_data;
184 249
185 mutex_lock(&dev->io_mutex); 250 /* if we cannot read at all, return EOF */
251 if (!dev->bulk_in_urb || !count)
252 return 0;
253
254 /* no concurrent readers */
255 rv = mutex_lock_interruptible(&dev->io_mutex);
256 if (rv < 0)
257 return rv;
258
186 if (!dev->interface) { /* disconnect() was called */ 259 if (!dev->interface) { /* disconnect() was called */
187 retval = -ENODEV; 260 rv = -ENODEV;
188 goto exit; 261 goto exit;
189 } 262 }
190 263
191 /* do a blocking bulk read to get data from the device */ 264 /* if IO is under way, we must not touch things */
192 retval = usb_bulk_msg(dev->udev, 265retry:
193 usb_rcvbulkpipe(dev->udev, dev->bulk_in_endpointAddr), 266 spin_lock_irq(&dev->err_lock);
194 dev->bulk_in_buffer, 267 ongoing_io = dev->ongoing_read;
195 min(dev->bulk_in_size, count), 268 spin_unlock_irq(&dev->err_lock);
196 &bytes_read, 10000); 269
197 270 if (ongoing_io) {
198 /* if the read was successful, copy the data to userspace */ 271 /*
199 if (!retval) { 272 * IO may take forever
200 if (copy_to_user(buffer, dev->bulk_in_buffer, bytes_read)) 273 * hence wait in an interruptible state
201 retval = -EFAULT; 274 */
202 else 275 rv = wait_for_completion_interruptible(&dev->bulk_in_completion);
203 retval = bytes_read; 276 if (rv < 0)
277 goto exit;
278 /*
279 * by waiting we also semiprocessed the urb
280 * we must finish now
281 */
282 dev->bulk_in_copied = 0;
283 dev->processed_urb = 1;
284 }
285
286 if (!dev->processed_urb) {
287 /*
288 * the URB hasn't been processed
289 * do it now
290 */
291 wait_for_completion(&dev->bulk_in_completion);
292 dev->bulk_in_copied = 0;
293 dev->processed_urb = 1;
204 } 294 }
205 295
296 /* errors must be reported */
297 if ((rv = dev->errors) < 0) {
298 /* any error is reported once */
299 dev->errors = 0;
300 /* to preserve notifications about reset */
301 rv = (rv == -EPIPE) ? rv : -EIO;
302 /* no data to deliver */
303 dev->bulk_in_filled = 0;
304 /* report it */
305 goto exit;
306 }
307
308 /*
309 * if the buffer is filled we may satisfy the read
310 * else we need to start IO
311 */
312
313 if (dev->bulk_in_filled) {
314 /* we had read data */
315 size_t available = dev->bulk_in_filled - dev->bulk_in_copied;
316 size_t chunk = min(available, count);
317
318 if (!available) {
319 /*
320 * all data has been used
321 * actual IO needs to be done
322 */
323 rv = skel_do_read_io(dev, count);
324 if (rv < 0)
325 goto exit;
326 else
327 goto retry;
328 }
329 /*
330 * data is available
331 * chunk tells us how much shall be copied
332 */
333
334 if (copy_to_user(buffer,
335 dev->bulk_in_buffer + dev->bulk_in_copied,
336 chunk))
337 rv = -EFAULT;
338 else
339 rv = chunk;
340
341 dev->bulk_in_copied += chunk;
342
343 /*
344 * if we are asked for more than we have,
345 * we start IO but don't wait
346 */
347 if (available < count)
348 skel_do_read_io(dev, count - chunk);
349 } else {
350 /* no data in the buffer */
351 rv = skel_do_read_io(dev, count);
352 if (rv < 0)
353 goto exit;
354 else
355 goto retry;
356 }
206exit: 357exit:
207 mutex_unlock(&dev->io_mutex); 358 mutex_unlock(&dev->io_mutex);
208 return retval; 359 return rv;
209} 360}
210 361
211static void skel_write_bulk_callback(struct urb *urb) 362static void skel_write_bulk_callback(struct urb *urb)
@@ -363,6 +514,7 @@ static int skel_probe(struct usb_interface *interface, const struct usb_device_i
363 mutex_init(&dev->io_mutex); 514 mutex_init(&dev->io_mutex);
364 spin_lock_init(&dev->err_lock); 515 spin_lock_init(&dev->err_lock);
365 init_usb_anchor(&dev->submitted); 516 init_usb_anchor(&dev->submitted);
517 init_completion(&dev->bulk_in_completion);
366 518
367 dev->udev = usb_get_dev(interface_to_usbdev(interface)); 519 dev->udev = usb_get_dev(interface_to_usbdev(interface));
368 dev->interface = interface; 520 dev->interface = interface;
@@ -384,6 +536,11 @@ static int skel_probe(struct usb_interface *interface, const struct usb_device_i
384 err("Could not allocate bulk_in_buffer"); 536 err("Could not allocate bulk_in_buffer");
385 goto error; 537 goto error;
386 } 538 }
539 dev->bulk_in_urb = usb_alloc_urb(0, GFP_KERNEL);
540 if (!dev->bulk_in_urb) {
541 err("Could not allocate bulk_in_urb");
542 goto error;
543 }
387 } 544 }
388 545
389 if (!dev->bulk_out_endpointAddr && 546 if (!dev->bulk_out_endpointAddr &&
@@ -453,6 +610,7 @@ static void skel_draw_down(struct usb_skel *dev)
453 time = usb_wait_anchor_empty_timeout(&dev->submitted, 1000); 610 time = usb_wait_anchor_empty_timeout(&dev->submitted, 1000);
454 if (!time) 611 if (!time)
455 usb_kill_anchored_urbs(&dev->submitted); 612 usb_kill_anchored_urbs(&dev->submitted);
613 usb_kill_urb(dev->bulk_in_urb);
456} 614}
457 615
458static int skel_suspend(struct usb_interface *intf, pm_message_t message) 616static int skel_suspend(struct usb_interface *intf, pm_message_t message)