aboutsummaryrefslogtreecommitdiffstats
path: root/drivers/firewire
diff options
context:
space:
mode:
authorStefan Richter <stefanr@s5r6.in-berlin.de>2008-08-14 15:47:21 -0400
committerStefan Richter <stefanr@s5r6.in-berlin.de>2008-10-15 16:21:10 -0400
commit09b12dd4e3caff165a0f17a2f3ebd2bbc8544cc6 (patch)
tree877cbb8887a2b2d0f2b8a512cce3bacf4643e48c /drivers/firewire
parent1e119fa9950dfe0e6d97470098db776110ca47a9 (diff)
firewire: fw-sbp2: enforce s/g segment size limit
1. We don't need to round the SBP-2 segment size limit down to a multiple of 4 kB (0xffff -> 0xf000). It is only necessary to ensure quadlet alignment (0xffff -> 0xfffc). 2. Use dma_set_max_seg_size() to tell the DMA mapping infrastructure and the block IO layer about the restriction. This way we can remove the size checks and segment splitting in the queuecommand path. This assumes that no other code in the firewire stack uses dma_map_sg() with conflicting requirements. It furthermore assumes that the controller device's platform actually allows us to set the segment size to our liking. Assert the latter with a BUG_ON(). 3. Also use blk_queue_max_segment_size() to tell the block IO layer about it. It cannot know it because our scsi_add_host() does not point to the FireWire controller's device. Thanks to Grant Grundler and FUJITA Tomonori for advice. Signed-off-by: Stefan Richter <stefanr@s5r6.in-berlin.de>
Diffstat (limited to 'drivers/firewire')
-rw-r--r--drivers/firewire/fw-sbp2.c63
1 files changed, 27 insertions, 36 deletions
diff --git a/drivers/firewire/fw-sbp2.c b/drivers/firewire/fw-sbp2.c
index 05997cee4f37..5d8411afcedb 100644
--- a/drivers/firewire/fw-sbp2.c
+++ b/drivers/firewire/fw-sbp2.c
@@ -29,6 +29,7 @@
29 */ 29 */
30 30
31#include <linux/blkdev.h> 31#include <linux/blkdev.h>
32#include <linux/bug.h>
32#include <linux/delay.h> 33#include <linux/delay.h>
33#include <linux/device.h> 34#include <linux/device.h>
34#include <linux/dma-mapping.h> 35#include <linux/dma-mapping.h>
@@ -181,10 +182,16 @@ struct sbp2_target {
181#define SBP2_MAX_LOGIN_ORB_TIMEOUT 40000U /* Timeout in ms */ 182#define SBP2_MAX_LOGIN_ORB_TIMEOUT 40000U /* Timeout in ms */
182#define SBP2_ORB_TIMEOUT 2000U /* Timeout in ms */ 183#define SBP2_ORB_TIMEOUT 2000U /* Timeout in ms */
183#define SBP2_ORB_NULL 0x80000000 184#define SBP2_ORB_NULL 0x80000000
184#define SBP2_MAX_SG_ELEMENT_LENGTH 0xf000
185#define SBP2_RETRY_LIMIT 0xf /* 15 retries */ 185#define SBP2_RETRY_LIMIT 0xf /* 15 retries */
186#define SBP2_CYCLE_LIMIT (0xc8 << 12) /* 200 125us cycles */ 186#define SBP2_CYCLE_LIMIT (0xc8 << 12) /* 200 125us cycles */
187 187
188/*
189 * The default maximum s/g segment size of a FireWire controller is
190 * usually 0x10000, but SBP-2 only allows 0xffff. Since buffers have to
191 * be quadlet-aligned, we set the length limit to 0xffff & ~3.
192 */
193#define SBP2_MAX_SEG_SIZE 0xfffc
194
188/* Unit directory keys */ 195/* Unit directory keys */
189#define SBP2_CSR_UNIT_CHARACTERISTICS 0x3a 196#define SBP2_CSR_UNIT_CHARACTERISTICS 0x3a
190#define SBP2_CSR_FIRMWARE_REVISION 0x3c 197#define SBP2_CSR_FIRMWARE_REVISION 0x3c
@@ -1099,6 +1106,10 @@ static int sbp2_probe(struct device *dev)
1099 struct Scsi_Host *shost; 1106 struct Scsi_Host *shost;
1100 u32 model, firmware_revision; 1107 u32 model, firmware_revision;
1101 1108
1109 if (dma_get_max_seg_size(device->card->device) > SBP2_MAX_SEG_SIZE)
1110 BUG_ON(dma_set_max_seg_size(device->card->device,
1111 SBP2_MAX_SEG_SIZE));
1112
1102 shost = scsi_host_alloc(&scsi_driver_template, sizeof(*tgt)); 1113 shost = scsi_host_alloc(&scsi_driver_template, sizeof(*tgt));
1103 if (shost == NULL) 1114 if (shost == NULL)
1104 return -ENOMEM; 1115 return -ENOMEM;
@@ -1347,14 +1358,12 @@ static int
1347sbp2_map_scatterlist(struct sbp2_command_orb *orb, struct fw_device *device, 1358sbp2_map_scatterlist(struct sbp2_command_orb *orb, struct fw_device *device,
1348 struct sbp2_logical_unit *lu) 1359 struct sbp2_logical_unit *lu)
1349{ 1360{
1350 struct scatterlist *sg; 1361 struct scatterlist *sg = scsi_sglist(orb->cmd);
1351 int sg_len, l, i, j, count; 1362 int i, n;
1352 dma_addr_t sg_addr; 1363
1353 1364 n = dma_map_sg(device->card->device, sg, scsi_sg_count(orb->cmd),
1354 sg = scsi_sglist(orb->cmd); 1365 orb->cmd->sc_data_direction);
1355 count = dma_map_sg(device->card->device, sg, scsi_sg_count(orb->cmd), 1366 if (n == 0)
1356 orb->cmd->sc_data_direction);
1357 if (count == 0)
1358 goto fail; 1367 goto fail;
1359 1368
1360 /* 1369 /*
@@ -1364,7 +1373,7 @@ sbp2_map_scatterlist(struct sbp2_command_orb *orb, struct fw_device *device,
1364 * as the second generation iPod which doesn't support page 1373 * as the second generation iPod which doesn't support page
1365 * tables. 1374 * tables.
1366 */ 1375 */
1367 if (count == 1 && sg_dma_len(sg) < SBP2_MAX_SG_ELEMENT_LENGTH) { 1376 if (n == 1) {
1368 orb->request.data_descriptor.high = 1377 orb->request.data_descriptor.high =
1369 cpu_to_be32(lu->tgt->address_high); 1378 cpu_to_be32(lu->tgt->address_high);
1370 orb->request.data_descriptor.low = 1379 orb->request.data_descriptor.low =
@@ -1374,29 +1383,9 @@ sbp2_map_scatterlist(struct sbp2_command_orb *orb, struct fw_device *device,
1374 return 0; 1383 return 0;
1375 } 1384 }
1376 1385
1377 /* 1386 for_each_sg(sg, sg, n, i) {
1378 * Convert the scatterlist to an sbp2 page table. If any 1387 orb->page_table[i].high = cpu_to_be32(sg_dma_len(sg) << 16);
1379 * scatterlist entries are too big for sbp2, we split them as we 1388 orb->page_table[i].low = cpu_to_be32(sg_dma_address(sg));
1380 * go. Even if we ask the block I/O layer to not give us sg
1381 * elements larger than 65535 bytes, some IOMMUs may merge sg elements
1382 * during DMA mapping, and Linux currently doesn't prevent this.
1383 */
1384 for (i = 0, j = 0; i < count; i++, sg = sg_next(sg)) {
1385 sg_len = sg_dma_len(sg);
1386 sg_addr = sg_dma_address(sg);
1387 while (sg_len) {
1388 /* FIXME: This won't get us out of the pinch. */
1389 if (unlikely(j >= ARRAY_SIZE(orb->page_table))) {
1390 fw_error("page table overflow\n");
1391 goto fail_page_table;
1392 }
1393 l = min(sg_len, SBP2_MAX_SG_ELEMENT_LENGTH);
1394 orb->page_table[j].low = cpu_to_be32(sg_addr);
1395 orb->page_table[j].high = cpu_to_be32(l << 16);
1396 sg_addr += l;
1397 sg_len -= l;
1398 j++;
1399 }
1400 } 1389 }
1401 1390
1402 orb->page_table_bus = 1391 orb->page_table_bus =
@@ -1415,13 +1404,13 @@ sbp2_map_scatterlist(struct sbp2_command_orb *orb, struct fw_device *device,
1415 orb->request.data_descriptor.high = cpu_to_be32(lu->tgt->address_high); 1404 orb->request.data_descriptor.high = cpu_to_be32(lu->tgt->address_high);
1416 orb->request.data_descriptor.low = cpu_to_be32(orb->page_table_bus); 1405 orb->request.data_descriptor.low = cpu_to_be32(orb->page_table_bus);
1417 orb->request.misc |= cpu_to_be32(COMMAND_ORB_PAGE_TABLE_PRESENT | 1406 orb->request.misc |= cpu_to_be32(COMMAND_ORB_PAGE_TABLE_PRESENT |
1418 COMMAND_ORB_DATA_SIZE(j)); 1407 COMMAND_ORB_DATA_SIZE(n));
1419 1408
1420 return 0; 1409 return 0;
1421 1410
1422 fail_page_table: 1411 fail_page_table:
1423 dma_unmap_sg(device->card->device, sg, scsi_sg_count(orb->cmd), 1412 dma_unmap_sg(device->card->device, scsi_sglist(orb->cmd),
1424 orb->cmd->sc_data_direction); 1413 scsi_sg_count(orb->cmd), orb->cmd->sc_data_direction);
1425 fail: 1414 fail:
1426 return -ENOMEM; 1415 return -ENOMEM;
1427} 1416}
@@ -1542,6 +1531,8 @@ static int sbp2_scsi_slave_configure(struct scsi_device *sdev)
1542 if (lu->tgt->workarounds & SBP2_WORKAROUND_128K_MAX_TRANS) 1531 if (lu->tgt->workarounds & SBP2_WORKAROUND_128K_MAX_TRANS)
1543 blk_queue_max_sectors(sdev->request_queue, 128 * 1024 / 512); 1532 blk_queue_max_sectors(sdev->request_queue, 128 * 1024 / 512);
1544 1533
1534 blk_queue_max_segment_size(sdev->request_queue, SBP2_MAX_SEG_SIZE);
1535
1545 return 0; 1536 return 0;
1546} 1537}
1547 1538