diff options
author | Stefan Richter <stefanr@s5r6.in-berlin.de> | 2008-08-14 15:47:21 -0400 |
---|---|---|
committer | Stefan Richter <stefanr@s5r6.in-berlin.de> | 2008-10-15 16:21:10 -0400 |
commit | 09b12dd4e3caff165a0f17a2f3ebd2bbc8544cc6 (patch) | |
tree | 877cbb8887a2b2d0f2b8a512cce3bacf4643e48c | |
parent | 1e119fa9950dfe0e6d97470098db776110ca47a9 (diff) |
firewire: fw-sbp2: enforce s/g segment size limit
1. We don't need to round the SBP-2 segment size limit down to a
multiple of 4 kB (0xffff -> 0xf000). It is only necessary to
ensure quadlet alignment (0xffff -> 0xfffc).
2. Use dma_set_max_seg_size() to tell the DMA mapping infrastructure
and the block IO layer about the restriction. This way we can
remove the size checks and segment splitting in the queuecommand
path.
This assumes that no other code in the firewire stack uses
dma_map_sg() with conflicting requirements. It furthermore assumes
that the controller device's platform actually allows us to set the
segment size to our liking. Assert the latter with a BUG_ON().
3. Also use blk_queue_max_segment_size() to tell the block IO layer
about it. It cannot know it because our scsi_add_host() does not
point to the FireWire controller's device.
Thanks to Grant Grundler and FUJITA Tomonori for advice.
Signed-off-by: Stefan Richter <stefanr@s5r6.in-berlin.de>
-rw-r--r-- | drivers/firewire/fw-sbp2.c | 63 |
1 files changed, 27 insertions, 36 deletions
diff --git a/drivers/firewire/fw-sbp2.c b/drivers/firewire/fw-sbp2.c index 05997cee4f37..5d8411afcedb 100644 --- a/drivers/firewire/fw-sbp2.c +++ b/drivers/firewire/fw-sbp2.c | |||
@@ -29,6 +29,7 @@ | |||
29 | */ | 29 | */ |
30 | 30 | ||
31 | #include <linux/blkdev.h> | 31 | #include <linux/blkdev.h> |
32 | #include <linux/bug.h> | ||
32 | #include <linux/delay.h> | 33 | #include <linux/delay.h> |
33 | #include <linux/device.h> | 34 | #include <linux/device.h> |
34 | #include <linux/dma-mapping.h> | 35 | #include <linux/dma-mapping.h> |
@@ -181,10 +182,16 @@ struct sbp2_target { | |||
181 | #define SBP2_MAX_LOGIN_ORB_TIMEOUT 40000U /* Timeout in ms */ | 182 | #define SBP2_MAX_LOGIN_ORB_TIMEOUT 40000U /* Timeout in ms */ |
182 | #define SBP2_ORB_TIMEOUT 2000U /* Timeout in ms */ | 183 | #define SBP2_ORB_TIMEOUT 2000U /* Timeout in ms */ |
183 | #define SBP2_ORB_NULL 0x80000000 | 184 | #define SBP2_ORB_NULL 0x80000000 |
184 | #define SBP2_MAX_SG_ELEMENT_LENGTH 0xf000 | ||
185 | #define SBP2_RETRY_LIMIT 0xf /* 15 retries */ | 185 | #define SBP2_RETRY_LIMIT 0xf /* 15 retries */ |
186 | #define SBP2_CYCLE_LIMIT (0xc8 << 12) /* 200 125us cycles */ | 186 | #define SBP2_CYCLE_LIMIT (0xc8 << 12) /* 200 125us cycles */ |
187 | 187 | ||
188 | /* | ||
189 | * The default maximum s/g segment size of a FireWire controller is | ||
190 | * usually 0x10000, but SBP-2 only allows 0xffff. Since buffers have to | ||
191 | * be quadlet-aligned, we set the length limit to 0xffff & ~3. | ||
192 | */ | ||
193 | #define SBP2_MAX_SEG_SIZE 0xfffc | ||
194 | |||
188 | /* Unit directory keys */ | 195 | /* Unit directory keys */ |
189 | #define SBP2_CSR_UNIT_CHARACTERISTICS 0x3a | 196 | #define SBP2_CSR_UNIT_CHARACTERISTICS 0x3a |
190 | #define SBP2_CSR_FIRMWARE_REVISION 0x3c | 197 | #define SBP2_CSR_FIRMWARE_REVISION 0x3c |
@@ -1099,6 +1106,10 @@ static int sbp2_probe(struct device *dev) | |||
1099 | struct Scsi_Host *shost; | 1106 | struct Scsi_Host *shost; |
1100 | u32 model, firmware_revision; | 1107 | u32 model, firmware_revision; |
1101 | 1108 | ||
1109 | if (dma_get_max_seg_size(device->card->device) > SBP2_MAX_SEG_SIZE) | ||
1110 | BUG_ON(dma_set_max_seg_size(device->card->device, | ||
1111 | SBP2_MAX_SEG_SIZE)); | ||
1112 | |||
1102 | shost = scsi_host_alloc(&scsi_driver_template, sizeof(*tgt)); | 1113 | shost = scsi_host_alloc(&scsi_driver_template, sizeof(*tgt)); |
1103 | if (shost == NULL) | 1114 | if (shost == NULL) |
1104 | return -ENOMEM; | 1115 | return -ENOMEM; |
@@ -1347,14 +1358,12 @@ static int | |||
1347 | sbp2_map_scatterlist(struct sbp2_command_orb *orb, struct fw_device *device, | 1358 | sbp2_map_scatterlist(struct sbp2_command_orb *orb, struct fw_device *device, |
1348 | struct sbp2_logical_unit *lu) | 1359 | struct sbp2_logical_unit *lu) |
1349 | { | 1360 | { |
1350 | struct scatterlist *sg; | 1361 | struct scatterlist *sg = scsi_sglist(orb->cmd); |
1351 | int sg_len, l, i, j, count; | 1362 | int i, n; |
1352 | dma_addr_t sg_addr; | 1363 | |
1353 | 1364 | n = dma_map_sg(device->card->device, sg, scsi_sg_count(orb->cmd), | |
1354 | sg = scsi_sglist(orb->cmd); | 1365 | orb->cmd->sc_data_direction); |
1355 | count = dma_map_sg(device->card->device, sg, scsi_sg_count(orb->cmd), | 1366 | if (n == 0) |
1356 | orb->cmd->sc_data_direction); | ||
1357 | if (count == 0) | ||
1358 | goto fail; | 1367 | goto fail; |
1359 | 1368 | ||
1360 | /* | 1369 | /* |
@@ -1364,7 +1373,7 @@ sbp2_map_scatterlist(struct sbp2_command_orb *orb, struct fw_device *device, | |||
1364 | * as the second generation iPod which doesn't support page | 1373 | * as the second generation iPod which doesn't support page |
1365 | * tables. | 1374 | * tables. |
1366 | */ | 1375 | */ |
1367 | if (count == 1 && sg_dma_len(sg) < SBP2_MAX_SG_ELEMENT_LENGTH) { | 1376 | if (n == 1) { |
1368 | orb->request.data_descriptor.high = | 1377 | orb->request.data_descriptor.high = |
1369 | cpu_to_be32(lu->tgt->address_high); | 1378 | cpu_to_be32(lu->tgt->address_high); |
1370 | orb->request.data_descriptor.low = | 1379 | orb->request.data_descriptor.low = |
@@ -1374,29 +1383,9 @@ sbp2_map_scatterlist(struct sbp2_command_orb *orb, struct fw_device *device, | |||
1374 | return 0; | 1383 | return 0; |
1375 | } | 1384 | } |
1376 | 1385 | ||
1377 | /* | 1386 | for_each_sg(sg, sg, n, i) { |
1378 | * Convert the scatterlist to an sbp2 page table. If any | 1387 | orb->page_table[i].high = cpu_to_be32(sg_dma_len(sg) << 16); |
1379 | * scatterlist entries are too big for sbp2, we split them as we | 1388 | orb->page_table[i].low = cpu_to_be32(sg_dma_address(sg)); |
1380 | * go. Even if we ask the block I/O layer to not give us sg | ||
1381 | * elements larger than 65535 bytes, some IOMMUs may merge sg elements | ||
1382 | * during DMA mapping, and Linux currently doesn't prevent this. | ||
1383 | */ | ||
1384 | for (i = 0, j = 0; i < count; i++, sg = sg_next(sg)) { | ||
1385 | sg_len = sg_dma_len(sg); | ||
1386 | sg_addr = sg_dma_address(sg); | ||
1387 | while (sg_len) { | ||
1388 | /* FIXME: This won't get us out of the pinch. */ | ||
1389 | if (unlikely(j >= ARRAY_SIZE(orb->page_table))) { | ||
1390 | fw_error("page table overflow\n"); | ||
1391 | goto fail_page_table; | ||
1392 | } | ||
1393 | l = min(sg_len, SBP2_MAX_SG_ELEMENT_LENGTH); | ||
1394 | orb->page_table[j].low = cpu_to_be32(sg_addr); | ||
1395 | orb->page_table[j].high = cpu_to_be32(l << 16); | ||
1396 | sg_addr += l; | ||
1397 | sg_len -= l; | ||
1398 | j++; | ||
1399 | } | ||
1400 | } | 1389 | } |
1401 | 1390 | ||
1402 | orb->page_table_bus = | 1391 | orb->page_table_bus = |
@@ -1415,13 +1404,13 @@ sbp2_map_scatterlist(struct sbp2_command_orb *orb, struct fw_device *device, | |||
1415 | orb->request.data_descriptor.high = cpu_to_be32(lu->tgt->address_high); | 1404 | orb->request.data_descriptor.high = cpu_to_be32(lu->tgt->address_high); |
1416 | orb->request.data_descriptor.low = cpu_to_be32(orb->page_table_bus); | 1405 | orb->request.data_descriptor.low = cpu_to_be32(orb->page_table_bus); |
1417 | orb->request.misc |= cpu_to_be32(COMMAND_ORB_PAGE_TABLE_PRESENT | | 1406 | orb->request.misc |= cpu_to_be32(COMMAND_ORB_PAGE_TABLE_PRESENT | |
1418 | COMMAND_ORB_DATA_SIZE(j)); | 1407 | COMMAND_ORB_DATA_SIZE(n)); |
1419 | 1408 | ||
1420 | return 0; | 1409 | return 0; |
1421 | 1410 | ||
1422 | fail_page_table: | 1411 | fail_page_table: |
1423 | dma_unmap_sg(device->card->device, sg, scsi_sg_count(orb->cmd), | 1412 | dma_unmap_sg(device->card->device, scsi_sglist(orb->cmd), |
1424 | orb->cmd->sc_data_direction); | 1413 | scsi_sg_count(orb->cmd), orb->cmd->sc_data_direction); |
1425 | fail: | 1414 | fail: |
1426 | return -ENOMEM; | 1415 | return -ENOMEM; |
1427 | } | 1416 | } |
@@ -1542,6 +1531,8 @@ static int sbp2_scsi_slave_configure(struct scsi_device *sdev) | |||
1542 | if (lu->tgt->workarounds & SBP2_WORKAROUND_128K_MAX_TRANS) | 1531 | if (lu->tgt->workarounds & SBP2_WORKAROUND_128K_MAX_TRANS) |
1543 | blk_queue_max_sectors(sdev->request_queue, 128 * 1024 / 512); | 1532 | blk_queue_max_sectors(sdev->request_queue, 128 * 1024 / 512); |
1544 | 1533 | ||
1534 | blk_queue_max_segment_size(sdev->request_queue, SBP2_MAX_SEG_SIZE); | ||
1535 | |||
1545 | return 0; | 1536 | return 0; |
1546 | } | 1537 | } |
1547 | 1538 | ||