diff options
Diffstat (limited to 'drivers')
-rw-r--r-- | drivers/block/cciss.c | 72 | ||||
-rw-r--r-- | drivers/ide/ide-io.c | 42 | ||||
-rw-r--r-- | drivers/ide/ide-probe.c | 2 | ||||
-rw-r--r-- | drivers/scsi/scsi.c | 109 | ||||
-rw-r--r-- | drivers/scsi/scsi_lib.c | 36 | ||||
-rw-r--r-- | drivers/scsi/scsi_priv.h | 1 |
6 files changed, 128 insertions, 134 deletions
diff --git a/drivers/block/cciss.c b/drivers/block/cciss.c index 88452c79fb64..e4e9f255bd1f 100644 --- a/drivers/block/cciss.c +++ b/drivers/block/cciss.c | |||
@@ -2178,16 +2178,48 @@ static inline void resend_cciss_cmd( ctlr_info_t *h, CommandList_struct *c) | |||
2178 | 2178 | ||
2179 | start_io(h); | 2179 | start_io(h); |
2180 | } | 2180 | } |
2181 | |||
2182 | static void cciss_softirq_done(struct request *rq) | ||
2183 | { | ||
2184 | CommandList_struct *cmd = rq->completion_data; | ||
2185 | ctlr_info_t *h = hba[cmd->ctlr]; | ||
2186 | u64bit temp64; | ||
2187 | int i, ddir; | ||
2188 | |||
2189 | if (cmd->Request.Type.Direction == XFER_READ) | ||
2190 | ddir = PCI_DMA_FROMDEVICE; | ||
2191 | else | ||
2192 | ddir = PCI_DMA_TODEVICE; | ||
2193 | |||
2194 | /* command did not need to be retried */ | ||
2195 | /* unmap the DMA mapping for all the scatter gather elements */ | ||
2196 | for(i=0; i<cmd->Header.SGList; i++) { | ||
2197 | temp64.val32.lower = cmd->SG[i].Addr.lower; | ||
2198 | temp64.val32.upper = cmd->SG[i].Addr.upper; | ||
2199 | pci_unmap_page(h->pdev, temp64.val, cmd->SG[i].Len, ddir); | ||
2200 | } | ||
2201 | |||
2202 | complete_buffers(rq->bio, rq->errors); | ||
2203 | |||
2204 | #ifdef CCISS_DEBUG | ||
2205 | printk("Done with %p\n", rq); | ||
2206 | #endif /* CCISS_DEBUG */ | ||
2207 | |||
2208 | spin_lock_irq(&h->lock); | ||
2209 | end_that_request_last(rq, rq->errors); | ||
2210 | cmd_free(h, cmd,1); | ||
2211 | spin_unlock_irq(&h->lock); | ||
2212 | } | ||
2213 | |||
2181 | /* checks the status of the job and calls complete buffers to mark all | 2214 | /* checks the status of the job and calls complete buffers to mark all |
2182 | * buffers for the completed job. | 2215 | * buffers for the completed job. Note that this function does not need |
2216 | * to hold the hba/queue lock. | ||
2183 | */ | 2217 | */ |
2184 | static inline void complete_command( ctlr_info_t *h, CommandList_struct *cmd, | 2218 | static inline void complete_command( ctlr_info_t *h, CommandList_struct *cmd, |
2185 | int timeout) | 2219 | int timeout) |
2186 | { | 2220 | { |
2187 | int status = 1; | 2221 | int status = 1; |
2188 | int i; | ||
2189 | int retry_cmd = 0; | 2222 | int retry_cmd = 0; |
2190 | u64bit temp64; | ||
2191 | 2223 | ||
2192 | if (timeout) | 2224 | if (timeout) |
2193 | status = 0; | 2225 | status = 0; |
@@ -2295,24 +2327,10 @@ static inline void complete_command( ctlr_info_t *h, CommandList_struct *cmd, | |||
2295 | resend_cciss_cmd(h,cmd); | 2327 | resend_cciss_cmd(h,cmd); |
2296 | return; | 2328 | return; |
2297 | } | 2329 | } |
2298 | /* command did not need to be retried */ | ||
2299 | /* unmap the DMA mapping for all the scatter gather elements */ | ||
2300 | for(i=0; i<cmd->Header.SGList; i++) { | ||
2301 | temp64.val32.lower = cmd->SG[i].Addr.lower; | ||
2302 | temp64.val32.upper = cmd->SG[i].Addr.upper; | ||
2303 | pci_unmap_page(hba[cmd->ctlr]->pdev, | ||
2304 | temp64.val, cmd->SG[i].Len, | ||
2305 | (cmd->Request.Type.Direction == XFER_READ) ? | ||
2306 | PCI_DMA_FROMDEVICE : PCI_DMA_TODEVICE); | ||
2307 | } | ||
2308 | complete_buffers(cmd->rq->bio, status); | ||
2309 | |||
2310 | #ifdef CCISS_DEBUG | ||
2311 | printk("Done with %p\n", cmd->rq); | ||
2312 | #endif /* CCISS_DEBUG */ | ||
2313 | 2330 | ||
2314 | end_that_request_last(cmd->rq, status ? 1 : -EIO); | 2331 | cmd->rq->completion_data = cmd; |
2315 | cmd_free(h,cmd,1); | 2332 | cmd->rq->errors = status; |
2333 | blk_complete_request(cmd->rq); | ||
2316 | } | 2334 | } |
2317 | 2335 | ||
2318 | /* | 2336 | /* |
@@ -3199,15 +3217,17 @@ static int __devinit cciss_init_one(struct pci_dev *pdev, | |||
3199 | drv->queue = q; | 3217 | drv->queue = q; |
3200 | 3218 | ||
3201 | q->backing_dev_info.ra_pages = READ_AHEAD; | 3219 | q->backing_dev_info.ra_pages = READ_AHEAD; |
3202 | blk_queue_bounce_limit(q, hba[i]->pdev->dma_mask); | 3220 | blk_queue_bounce_limit(q, hba[i]->pdev->dma_mask); |
3221 | |||
3222 | /* This is a hardware imposed limit. */ | ||
3223 | blk_queue_max_hw_segments(q, MAXSGENTRIES); | ||
3203 | 3224 | ||
3204 | /* This is a hardware imposed limit. */ | 3225 | /* This is a limit in the driver and could be eliminated. */ |
3205 | blk_queue_max_hw_segments(q, MAXSGENTRIES); | 3226 | blk_queue_max_phys_segments(q, MAXSGENTRIES); |
3206 | 3227 | ||
3207 | /* This is a limit in the driver and could be eliminated. */ | 3228 | blk_queue_max_sectors(q, 512); |
3208 | blk_queue_max_phys_segments(q, MAXSGENTRIES); | ||
3209 | 3229 | ||
3210 | blk_queue_max_sectors(q, 512); | 3230 | blk_queue_softirq_done(q, cciss_softirq_done); |
3211 | 3231 | ||
3212 | q->queuedata = hba[i]; | 3232 | q->queuedata = hba[i]; |
3213 | sprintf(disk->disk_name, "cciss/c%dd%d", i, j); | 3233 | sprintf(disk->disk_name, "cciss/c%dd%d", i, j); |
diff --git a/drivers/ide/ide-io.c b/drivers/ide/ide-io.c index b5dc6df8e67d..dea2d4dcc698 100644 --- a/drivers/ide/ide-io.c +++ b/drivers/ide/ide-io.c | |||
@@ -55,9 +55,22 @@ | |||
55 | #include <asm/io.h> | 55 | #include <asm/io.h> |
56 | #include <asm/bitops.h> | 56 | #include <asm/bitops.h> |
57 | 57 | ||
58 | void ide_softirq_done(struct request *rq) | ||
59 | { | ||
60 | request_queue_t *q = rq->q; | ||
61 | |||
62 | add_disk_randomness(rq->rq_disk); | ||
63 | end_that_request_chunk(rq, rq->errors, rq->data_len); | ||
64 | |||
65 | spin_lock_irq(q->queue_lock); | ||
66 | end_that_request_last(rq, rq->errors); | ||
67 | spin_unlock_irq(q->queue_lock); | ||
68 | } | ||
69 | |||
58 | int __ide_end_request(ide_drive_t *drive, struct request *rq, int uptodate, | 70 | int __ide_end_request(ide_drive_t *drive, struct request *rq, int uptodate, |
59 | int nr_sectors) | 71 | int nr_sectors) |
60 | { | 72 | { |
73 | unsigned int nbytes; | ||
61 | int ret = 1; | 74 | int ret = 1; |
62 | 75 | ||
63 | BUG_ON(!(rq->flags & REQ_STARTED)); | 76 | BUG_ON(!(rq->flags & REQ_STARTED)); |
@@ -81,17 +94,28 @@ int __ide_end_request(ide_drive_t *drive, struct request *rq, int uptodate, | |||
81 | HWGROUP(drive)->hwif->ide_dma_on(drive); | 94 | HWGROUP(drive)->hwif->ide_dma_on(drive); |
82 | } | 95 | } |
83 | 96 | ||
84 | if (!end_that_request_first(rq, uptodate, nr_sectors)) { | 97 | /* |
85 | add_disk_randomness(rq->rq_disk); | 98 | * For partial completions (or non fs/pc requests), use the regular |
86 | 99 | * direct completion path. | |
87 | if (blk_rq_tagged(rq)) | 100 | */ |
88 | blk_queue_end_tag(drive->queue, rq); | 101 | nbytes = nr_sectors << 9; |
89 | 102 | if (rq_all_done(rq, nbytes)) { | |
103 | rq->errors = uptodate; | ||
104 | rq->data_len = nbytes; | ||
90 | blkdev_dequeue_request(rq); | 105 | blkdev_dequeue_request(rq); |
91 | HWGROUP(drive)->rq = NULL; | 106 | HWGROUP(drive)->rq = NULL; |
92 | end_that_request_last(rq, uptodate); | 107 | blk_complete_request(rq); |
93 | ret = 0; | 108 | ret = 0; |
109 | } else { | ||
110 | if (!end_that_request_first(rq, uptodate, nr_sectors)) { | ||
111 | add_disk_randomness(rq->rq_disk); | ||
112 | blkdev_dequeue_request(rq); | ||
113 | HWGROUP(drive)->rq = NULL; | ||
114 | end_that_request_last(rq, uptodate); | ||
115 | ret = 0; | ||
116 | } | ||
94 | } | 117 | } |
118 | |||
95 | return ret; | 119 | return ret; |
96 | } | 120 | } |
97 | EXPORT_SYMBOL(__ide_end_request); | 121 | EXPORT_SYMBOL(__ide_end_request); |
@@ -113,6 +137,10 @@ int ide_end_request (ide_drive_t *drive, int uptodate, int nr_sectors) | |||
113 | unsigned long flags; | 137 | unsigned long flags; |
114 | int ret = 1; | 138 | int ret = 1; |
115 | 139 | ||
140 | /* | ||
141 | * room for locking improvements here, the calls below don't | ||
142 | * need the queue lock held at all | ||
143 | */ | ||
116 | spin_lock_irqsave(&ide_lock, flags); | 144 | spin_lock_irqsave(&ide_lock, flags); |
117 | rq = HWGROUP(drive)->rq; | 145 | rq = HWGROUP(drive)->rq; |
118 | 146 | ||
diff --git a/drivers/ide/ide-probe.c b/drivers/ide/ide-probe.c index 02167a5b751d..1ddaa71a8f45 100644 --- a/drivers/ide/ide-probe.c +++ b/drivers/ide/ide-probe.c | |||
@@ -1011,6 +1011,8 @@ static int ide_init_queue(ide_drive_t *drive) | |||
1011 | blk_queue_max_hw_segments(q, max_sg_entries); | 1011 | blk_queue_max_hw_segments(q, max_sg_entries); |
1012 | blk_queue_max_phys_segments(q, max_sg_entries); | 1012 | blk_queue_max_phys_segments(q, max_sg_entries); |
1013 | 1013 | ||
1014 | blk_queue_softirq_done(q, ide_softirq_done); | ||
1015 | |||
1014 | /* assign drive queue */ | 1016 | /* assign drive queue */ |
1015 | drive->queue = q; | 1017 | drive->queue = q; |
1016 | 1018 | ||
diff --git a/drivers/scsi/scsi.c b/drivers/scsi/scsi.c index 180676d7115a..ee5f4dfdab14 100644 --- a/drivers/scsi/scsi.c +++ b/drivers/scsi/scsi.c | |||
@@ -69,7 +69,6 @@ | |||
69 | #include "scsi_logging.h" | 69 | #include "scsi_logging.h" |
70 | 70 | ||
71 | static void scsi_done(struct scsi_cmnd *cmd); | 71 | static void scsi_done(struct scsi_cmnd *cmd); |
72 | static int scsi_retry_command(struct scsi_cmnd *cmd); | ||
73 | 72 | ||
74 | /* | 73 | /* |
75 | * Definitions and constants. | 74 | * Definitions and constants. |
@@ -752,7 +751,7 @@ static void scsi_done(struct scsi_cmnd *cmd) | |||
752 | * isn't running --- used by scsi_times_out */ | 751 | * isn't running --- used by scsi_times_out */ |
753 | void __scsi_done(struct scsi_cmnd *cmd) | 752 | void __scsi_done(struct scsi_cmnd *cmd) |
754 | { | 753 | { |
755 | unsigned long flags; | 754 | struct request *rq = cmd->request; |
756 | 755 | ||
757 | /* | 756 | /* |
758 | * Set the serial numbers back to zero | 757 | * Set the serial numbers back to zero |
@@ -763,71 +762,14 @@ void __scsi_done(struct scsi_cmnd *cmd) | |||
763 | if (cmd->result) | 762 | if (cmd->result) |
764 | atomic_inc(&cmd->device->ioerr_cnt); | 763 | atomic_inc(&cmd->device->ioerr_cnt); |
765 | 764 | ||
765 | BUG_ON(!rq); | ||
766 | |||
766 | /* | 767 | /* |
767 | * Next, enqueue the command into the done queue. | 768 | * The uptodate/nbytes values don't matter, as we allow partial |
768 | * It is a per-CPU queue, so we just disable local interrupts | 769 | * completes and thus will check this in the softirq callback |
769 | * and need no spinlock. | ||
770 | */ | 770 | */ |
771 | local_irq_save(flags); | 771 | rq->completion_data = cmd; |
772 | list_add_tail(&cmd->eh_entry, &__get_cpu_var(scsi_done_q)); | 772 | blk_complete_request(rq); |
773 | raise_softirq_irqoff(SCSI_SOFTIRQ); | ||
774 | local_irq_restore(flags); | ||
775 | } | ||
776 | |||
777 | /** | ||
778 | * scsi_softirq - Perform post-interrupt processing of finished SCSI commands. | ||
779 | * | ||
780 | * This is the consumer of the done queue. | ||
781 | * | ||
782 | * This is called with all interrupts enabled. This should reduce | ||
783 | * interrupt latency, stack depth, and reentrancy of the low-level | ||
784 | * drivers. | ||
785 | */ | ||
786 | static void scsi_softirq(struct softirq_action *h) | ||
787 | { | ||
788 | int disposition; | ||
789 | LIST_HEAD(local_q); | ||
790 | |||
791 | local_irq_disable(); | ||
792 | list_splice_init(&__get_cpu_var(scsi_done_q), &local_q); | ||
793 | local_irq_enable(); | ||
794 | |||
795 | while (!list_empty(&local_q)) { | ||
796 | struct scsi_cmnd *cmd = list_entry(local_q.next, | ||
797 | struct scsi_cmnd, eh_entry); | ||
798 | /* The longest time any command should be outstanding is the | ||
799 | * per command timeout multiplied by the number of retries. | ||
800 | * | ||
801 | * For a typical command, this is 2.5 minutes */ | ||
802 | unsigned long wait_for | ||
803 | = cmd->allowed * cmd->timeout_per_command; | ||
804 | list_del_init(&cmd->eh_entry); | ||
805 | |||
806 | disposition = scsi_decide_disposition(cmd); | ||
807 | if (disposition != SUCCESS && | ||
808 | time_before(cmd->jiffies_at_alloc + wait_for, jiffies)) { | ||
809 | sdev_printk(KERN_ERR, cmd->device, | ||
810 | "timing out command, waited %lus\n", | ||
811 | wait_for/HZ); | ||
812 | disposition = SUCCESS; | ||
813 | } | ||
814 | |||
815 | scsi_log_completion(cmd, disposition); | ||
816 | switch (disposition) { | ||
817 | case SUCCESS: | ||
818 | scsi_finish_command(cmd); | ||
819 | break; | ||
820 | case NEEDS_RETRY: | ||
821 | scsi_retry_command(cmd); | ||
822 | break; | ||
823 | case ADD_TO_MLQUEUE: | ||
824 | scsi_queue_insert(cmd, SCSI_MLQUEUE_DEVICE_BUSY); | ||
825 | break; | ||
826 | default: | ||
827 | if (!scsi_eh_scmd_add(cmd, 0)) | ||
828 | scsi_finish_command(cmd); | ||
829 | } | ||
830 | } | ||
831 | } | 773 | } |
832 | 774 | ||
833 | /* | 775 | /* |
@@ -840,7 +782,7 @@ static void scsi_softirq(struct softirq_action *h) | |||
840 | * level drivers should not become re-entrant as a result of | 782 | * level drivers should not become re-entrant as a result of |
841 | * this. | 783 | * this. |
842 | */ | 784 | */ |
843 | static int scsi_retry_command(struct scsi_cmnd *cmd) | 785 | int scsi_retry_command(struct scsi_cmnd *cmd) |
844 | { | 786 | { |
845 | /* | 787 | /* |
846 | * Restore the SCSI command state. | 788 | * Restore the SCSI command state. |
@@ -1273,38 +1215,6 @@ int scsi_device_cancel(struct scsi_device *sdev, int recovery) | |||
1273 | } | 1215 | } |
1274 | EXPORT_SYMBOL(scsi_device_cancel); | 1216 | EXPORT_SYMBOL(scsi_device_cancel); |
1275 | 1217 | ||
1276 | #ifdef CONFIG_HOTPLUG_CPU | ||
1277 | static int scsi_cpu_notify(struct notifier_block *self, | ||
1278 | unsigned long action, void *hcpu) | ||
1279 | { | ||
1280 | int cpu = (unsigned long)hcpu; | ||
1281 | |||
1282 | switch(action) { | ||
1283 | case CPU_DEAD: | ||
1284 | /* Drain scsi_done_q. */ | ||
1285 | local_irq_disable(); | ||
1286 | list_splice_init(&per_cpu(scsi_done_q, cpu), | ||
1287 | &__get_cpu_var(scsi_done_q)); | ||
1288 | raise_softirq_irqoff(SCSI_SOFTIRQ); | ||
1289 | local_irq_enable(); | ||
1290 | break; | ||
1291 | default: | ||
1292 | break; | ||
1293 | } | ||
1294 | return NOTIFY_OK; | ||
1295 | } | ||
1296 | |||
1297 | static struct notifier_block __devinitdata scsi_cpu_nb = { | ||
1298 | .notifier_call = scsi_cpu_notify, | ||
1299 | }; | ||
1300 | |||
1301 | #define register_scsi_cpu() register_cpu_notifier(&scsi_cpu_nb) | ||
1302 | #define unregister_scsi_cpu() unregister_cpu_notifier(&scsi_cpu_nb) | ||
1303 | #else | ||
1304 | #define register_scsi_cpu() | ||
1305 | #define unregister_scsi_cpu() | ||
1306 | #endif /* CONFIG_HOTPLUG_CPU */ | ||
1307 | |||
1308 | MODULE_DESCRIPTION("SCSI core"); | 1218 | MODULE_DESCRIPTION("SCSI core"); |
1309 | MODULE_LICENSE("GPL"); | 1219 | MODULE_LICENSE("GPL"); |
1310 | 1220 | ||
@@ -1338,8 +1248,6 @@ static int __init init_scsi(void) | |||
1338 | INIT_LIST_HEAD(&per_cpu(scsi_done_q, i)); | 1248 | INIT_LIST_HEAD(&per_cpu(scsi_done_q, i)); |
1339 | 1249 | ||
1340 | devfs_mk_dir("scsi"); | 1250 | devfs_mk_dir("scsi"); |
1341 | open_softirq(SCSI_SOFTIRQ, scsi_softirq, NULL); | ||
1342 | register_scsi_cpu(); | ||
1343 | printk(KERN_NOTICE "SCSI subsystem initialized\n"); | 1251 | printk(KERN_NOTICE "SCSI subsystem initialized\n"); |
1344 | return 0; | 1252 | return 0; |
1345 | 1253 | ||
@@ -1367,7 +1275,6 @@ static void __exit exit_scsi(void) | |||
1367 | devfs_remove("scsi"); | 1275 | devfs_remove("scsi"); |
1368 | scsi_exit_procfs(); | 1276 | scsi_exit_procfs(); |
1369 | scsi_exit_queue(); | 1277 | scsi_exit_queue(); |
1370 | unregister_scsi_cpu(); | ||
1371 | } | 1278 | } |
1372 | 1279 | ||
1373 | subsys_initcall(init_scsi); | 1280 | subsys_initcall(init_scsi); |
diff --git a/drivers/scsi/scsi_lib.c b/drivers/scsi/scsi_lib.c index ba93d6e66d48..00c9bf383e23 100644 --- a/drivers/scsi/scsi_lib.c +++ b/drivers/scsi/scsi_lib.c | |||
@@ -1493,6 +1493,41 @@ static void scsi_kill_request(struct request *req, request_queue_t *q) | |||
1493 | __scsi_done(cmd); | 1493 | __scsi_done(cmd); |
1494 | } | 1494 | } |
1495 | 1495 | ||
1496 | static void scsi_softirq_done(struct request *rq) | ||
1497 | { | ||
1498 | struct scsi_cmnd *cmd = rq->completion_data; | ||
1499 | unsigned long wait_for = cmd->allowed * cmd->timeout_per_command; | ||
1500 | int disposition; | ||
1501 | |||
1502 | INIT_LIST_HEAD(&cmd->eh_entry); | ||
1503 | |||
1504 | disposition = scsi_decide_disposition(cmd); | ||
1505 | if (disposition != SUCCESS && | ||
1506 | time_before(cmd->jiffies_at_alloc + wait_for, jiffies)) { | ||
1507 | sdev_printk(KERN_ERR, cmd->device, | ||
1508 | "timing out command, waited %lus\n", | ||
1509 | wait_for/HZ); | ||
1510 | disposition = SUCCESS; | ||
1511 | } | ||
1512 | |||
1513 | scsi_log_completion(cmd, disposition); | ||
1514 | |||
1515 | switch (disposition) { | ||
1516 | case SUCCESS: | ||
1517 | scsi_finish_command(cmd); | ||
1518 | break; | ||
1519 | case NEEDS_RETRY: | ||
1520 | scsi_retry_command(cmd); | ||
1521 | break; | ||
1522 | case ADD_TO_MLQUEUE: | ||
1523 | scsi_queue_insert(cmd, SCSI_MLQUEUE_DEVICE_BUSY); | ||
1524 | break; | ||
1525 | default: | ||
1526 | if (!scsi_eh_scmd_add(cmd, 0)) | ||
1527 | scsi_finish_command(cmd); | ||
1528 | } | ||
1529 | } | ||
1530 | |||
1496 | /* | 1531 | /* |
1497 | * Function: scsi_request_fn() | 1532 | * Function: scsi_request_fn() |
1498 | * | 1533 | * |
@@ -1667,6 +1702,7 @@ struct request_queue *scsi_alloc_queue(struct scsi_device *sdev) | |||
1667 | blk_queue_bounce_limit(q, scsi_calculate_bounce_limit(shost)); | 1702 | blk_queue_bounce_limit(q, scsi_calculate_bounce_limit(shost)); |
1668 | blk_queue_segment_boundary(q, shost->dma_boundary); | 1703 | blk_queue_segment_boundary(q, shost->dma_boundary); |
1669 | blk_queue_issue_flush_fn(q, scsi_issue_flush_fn); | 1704 | blk_queue_issue_flush_fn(q, scsi_issue_flush_fn); |
1705 | blk_queue_softirq_done(q, scsi_softirq_done); | ||
1670 | 1706 | ||
1671 | if (!shost->use_clustering) | 1707 | if (!shost->use_clustering) |
1672 | clear_bit(QUEUE_FLAG_CLUSTER, &q->queue_flags); | 1708 | clear_bit(QUEUE_FLAG_CLUSTER, &q->queue_flags); |
diff --git a/drivers/scsi/scsi_priv.h b/drivers/scsi/scsi_priv.h index f04e7e11f57a..14a6198cb8d2 100644 --- a/drivers/scsi/scsi_priv.h +++ b/drivers/scsi/scsi_priv.h | |||
@@ -44,6 +44,7 @@ extern void scsi_init_cmd_from_req(struct scsi_cmnd *cmd, | |||
44 | struct scsi_request *sreq); | 44 | struct scsi_request *sreq); |
45 | extern void __scsi_release_request(struct scsi_request *sreq); | 45 | extern void __scsi_release_request(struct scsi_request *sreq); |
46 | extern void __scsi_done(struct scsi_cmnd *cmd); | 46 | extern void __scsi_done(struct scsi_cmnd *cmd); |
47 | extern int scsi_retry_command(struct scsi_cmnd *cmd); | ||
47 | #ifdef CONFIG_SCSI_LOGGING | 48 | #ifdef CONFIG_SCSI_LOGGING |
48 | void scsi_log_send(struct scsi_cmnd *cmd); | 49 | void scsi_log_send(struct scsi_cmnd *cmd); |
49 | void scsi_log_completion(struct scsi_cmnd *cmd, int disposition); | 50 | void scsi_log_completion(struct scsi_cmnd *cmd, int disposition); |