diff options
author | Tejun Heo <htejun@gmail.com> | 2006-02-11 05:11:13 -0500 |
---|---|---|
committer | Jeff Garzik <jgarzik@pobox.com> | 2006-02-11 17:51:57 -0500 |
commit | beec7dbc6ff003bbc94de62b3323519c878fb2ac (patch) | |
tree | f1cf8dee4f271356f2025a386eaa746aabafd1fe /drivers/scsi/sata_mv.c | |
parent | a46314744d8fadb91451bf2e5d2fd949c4a752d8 (diff) |
[PATCH] libata: convert assert(xxx)'s in low-level drivers to WARN_ON(!xxx)'s
This patch converts all assert(xxx)'s in low-level drivers to
WARN_ON(!xxx)'s. After this patch, there is no in-kernel user of the
libata assert() macro.
Signed-off-by: Tejun Heo <htejun@gmail.com>
Signed-off-by: Jeff Garzik <jgarzik@pobox.com>
Diffstat (limited to 'drivers/scsi/sata_mv.c')
-rw-r--r-- | drivers/scsi/sata_mv.c | 50 |
1 files changed, 25 insertions, 25 deletions
diff --git a/drivers/scsi/sata_mv.c b/drivers/scsi/sata_mv.c index 6c80527ddd21..732cb64a4d1b 100644 --- a/drivers/scsi/sata_mv.c +++ b/drivers/scsi/sata_mv.c | |||
@@ -632,8 +632,8 @@ static void mv_irq_clear(struct ata_port *ap) | |||
632 | * @base: port base address | 632 | * @base: port base address |
633 | * @pp: port private data | 633 | * @pp: port private data |
634 | * | 634 | * |
635 | * Verify the local cache of the eDMA state is accurate with an | 635 | * Verify the local cache of the eDMA state is accurate with a |
636 | * assert. | 636 | * WARN_ON. |
637 | * | 637 | * |
638 | * LOCKING: | 638 | * LOCKING: |
639 | * Inherited from caller. | 639 | * Inherited from caller. |
@@ -644,15 +644,15 @@ static void mv_start_dma(void __iomem *base, struct mv_port_priv *pp) | |||
644 | writelfl(EDMA_EN, base + EDMA_CMD_OFS); | 644 | writelfl(EDMA_EN, base + EDMA_CMD_OFS); |
645 | pp->pp_flags |= MV_PP_FLAG_EDMA_EN; | 645 | pp->pp_flags |= MV_PP_FLAG_EDMA_EN; |
646 | } | 646 | } |
647 | assert(EDMA_EN & readl(base + EDMA_CMD_OFS)); | 647 | WARN_ON(!(EDMA_EN & readl(base + EDMA_CMD_OFS))); |
648 | } | 648 | } |
649 | 649 | ||
650 | /** | 650 | /** |
651 | * mv_stop_dma - Disable eDMA engine | 651 | * mv_stop_dma - Disable eDMA engine |
652 | * @ap: ATA channel to manipulate | 652 | * @ap: ATA channel to manipulate |
653 | * | 653 | * |
654 | * Verify the local cache of the eDMA state is accurate with an | 654 | * Verify the local cache of the eDMA state is accurate with a |
655 | * assert. | 655 | * WARN_ON. |
656 | * | 656 | * |
657 | * LOCKING: | 657 | * LOCKING: |
658 | * Inherited from caller. | 658 | * Inherited from caller. |
@@ -670,7 +670,7 @@ static void mv_stop_dma(struct ata_port *ap) | |||
670 | writelfl(EDMA_DS, port_mmio + EDMA_CMD_OFS); | 670 | writelfl(EDMA_DS, port_mmio + EDMA_CMD_OFS); |
671 | pp->pp_flags &= ~MV_PP_FLAG_EDMA_EN; | 671 | pp->pp_flags &= ~MV_PP_FLAG_EDMA_EN; |
672 | } else { | 672 | } else { |
673 | assert(!(EDMA_EN & readl(port_mmio + EDMA_CMD_OFS))); | 673 | WARN_ON(EDMA_EN & readl(port_mmio + EDMA_CMD_OFS)); |
674 | } | 674 | } |
675 | 675 | ||
676 | /* now properly wait for the eDMA to stop */ | 676 | /* now properly wait for the eDMA to stop */ |
@@ -1061,15 +1061,15 @@ static void mv_qc_prep(struct ata_queued_cmd *qc) | |||
1061 | return; | 1061 | return; |
1062 | 1062 | ||
1063 | /* the req producer index should be the same as we remember it */ | 1063 | /* the req producer index should be the same as we remember it */ |
1064 | assert(((readl(mv_ap_base(qc->ap) + EDMA_REQ_Q_IN_PTR_OFS) >> | 1064 | WARN_ON(((readl(mv_ap_base(qc->ap) + EDMA_REQ_Q_IN_PTR_OFS) >> |
1065 | EDMA_REQ_Q_PTR_SHIFT) & MV_MAX_Q_DEPTH_MASK) == | 1065 | EDMA_REQ_Q_PTR_SHIFT) & MV_MAX_Q_DEPTH_MASK) != |
1066 | pp->req_producer); | 1066 | pp->req_producer); |
1067 | 1067 | ||
1068 | /* Fill in command request block | 1068 | /* Fill in command request block |
1069 | */ | 1069 | */ |
1070 | if (!(qc->tf.flags & ATA_TFLAG_WRITE)) | 1070 | if (!(qc->tf.flags & ATA_TFLAG_WRITE)) |
1071 | flags |= CRQB_FLAG_READ; | 1071 | flags |= CRQB_FLAG_READ; |
1072 | assert(MV_MAX_Q_DEPTH > qc->tag); | 1072 | WARN_ON(MV_MAX_Q_DEPTH <= qc->tag); |
1073 | flags |= qc->tag << CRQB_TAG_SHIFT; | 1073 | flags |= qc->tag << CRQB_TAG_SHIFT; |
1074 | 1074 | ||
1075 | pp->crqb[pp->req_producer].sg_addr = | 1075 | pp->crqb[pp->req_producer].sg_addr = |
@@ -1152,16 +1152,16 @@ static void mv_qc_prep_iie(struct ata_queued_cmd *qc) | |||
1152 | return; | 1152 | return; |
1153 | 1153 | ||
1154 | /* the req producer index should be the same as we remember it */ | 1154 | /* the req producer index should be the same as we remember it */ |
1155 | assert(((readl(mv_ap_base(qc->ap) + EDMA_REQ_Q_IN_PTR_OFS) >> | 1155 | WARN_ON(((readl(mv_ap_base(qc->ap) + EDMA_REQ_Q_IN_PTR_OFS) >> |
1156 | EDMA_REQ_Q_PTR_SHIFT) & MV_MAX_Q_DEPTH_MASK) == | 1156 | EDMA_REQ_Q_PTR_SHIFT) & MV_MAX_Q_DEPTH_MASK) != |
1157 | pp->req_producer); | 1157 | pp->req_producer); |
1158 | 1158 | ||
1159 | /* Fill in Gen IIE command request block | 1159 | /* Fill in Gen IIE command request block |
1160 | */ | 1160 | */ |
1161 | if (!(qc->tf.flags & ATA_TFLAG_WRITE)) | 1161 | if (!(qc->tf.flags & ATA_TFLAG_WRITE)) |
1162 | flags |= CRQB_FLAG_READ; | 1162 | flags |= CRQB_FLAG_READ; |
1163 | 1163 | ||
1164 | assert(MV_MAX_Q_DEPTH > qc->tag); | 1164 | WARN_ON(MV_MAX_Q_DEPTH <= qc->tag); |
1165 | flags |= qc->tag << CRQB_TAG_SHIFT; | 1165 | flags |= qc->tag << CRQB_TAG_SHIFT; |
1166 | 1166 | ||
1167 | crqb = (struct mv_crqb_iie *) &pp->crqb[pp->req_producer]; | 1167 | crqb = (struct mv_crqb_iie *) &pp->crqb[pp->req_producer]; |
@@ -1226,12 +1226,12 @@ static unsigned int mv_qc_issue(struct ata_queued_cmd *qc) | |||
1226 | in_ptr = readl(port_mmio + EDMA_REQ_Q_IN_PTR_OFS); | 1226 | in_ptr = readl(port_mmio + EDMA_REQ_Q_IN_PTR_OFS); |
1227 | 1227 | ||
1228 | /* the req producer index should be the same as we remember it */ | 1228 | /* the req producer index should be the same as we remember it */ |
1229 | assert(((in_ptr >> EDMA_REQ_Q_PTR_SHIFT) & MV_MAX_Q_DEPTH_MASK) == | 1229 | WARN_ON(((in_ptr >> EDMA_REQ_Q_PTR_SHIFT) & MV_MAX_Q_DEPTH_MASK) != |
1230 | pp->req_producer); | 1230 | pp->req_producer); |
1231 | /* until we do queuing, the queue should be empty at this point */ | 1231 | /* until we do queuing, the queue should be empty at this point */ |
1232 | assert(((in_ptr >> EDMA_REQ_Q_PTR_SHIFT) & MV_MAX_Q_DEPTH_MASK) == | 1232 | WARN_ON(((in_ptr >> EDMA_REQ_Q_PTR_SHIFT) & MV_MAX_Q_DEPTH_MASK) != |
1233 | ((readl(port_mmio + EDMA_REQ_Q_OUT_PTR_OFS) >> | 1233 | ((readl(port_mmio + EDMA_REQ_Q_OUT_PTR_OFS) >> |
1234 | EDMA_REQ_Q_PTR_SHIFT) & MV_MAX_Q_DEPTH_MASK)); | 1234 | EDMA_REQ_Q_PTR_SHIFT) & MV_MAX_Q_DEPTH_MASK)); |
1235 | 1235 | ||
1236 | mv_inc_q_index(&pp->req_producer); /* now incr producer index */ | 1236 | mv_inc_q_index(&pp->req_producer); /* now incr producer index */ |
1237 | 1237 | ||
@@ -1251,7 +1251,7 @@ static unsigned int mv_qc_issue(struct ata_queued_cmd *qc) | |||
1251 | * | 1251 | * |
1252 | * This routine is for use when the port is in DMA mode, when it | 1252 | * This routine is for use when the port is in DMA mode, when it |
1253 | * will be using the CRPB (command response block) method of | 1253 | * will be using the CRPB (command response block) method of |
1254 | * returning command completion information. We assert indices | 1254 | * returning command completion information. We check indices |
1255 | * are good, grab status, and bump the response consumer index to | 1255 | * are good, grab status, and bump the response consumer index to |
1256 | * prove that we're up to date. | 1256 | * prove that we're up to date. |
1257 | * | 1257 | * |
@@ -1267,16 +1267,16 @@ static u8 mv_get_crpb_status(struct ata_port *ap) | |||
1267 | out_ptr = readl(port_mmio + EDMA_RSP_Q_OUT_PTR_OFS); | 1267 | out_ptr = readl(port_mmio + EDMA_RSP_Q_OUT_PTR_OFS); |
1268 | 1268 | ||
1269 | /* the response consumer index should be the same as we remember it */ | 1269 | /* the response consumer index should be the same as we remember it */ |
1270 | assert(((out_ptr >> EDMA_RSP_Q_PTR_SHIFT) & MV_MAX_Q_DEPTH_MASK) == | 1270 | WARN_ON(((out_ptr >> EDMA_RSP_Q_PTR_SHIFT) & MV_MAX_Q_DEPTH_MASK) != |
1271 | pp->rsp_consumer); | 1271 | pp->rsp_consumer); |
1272 | 1272 | ||
1273 | /* increment our consumer index... */ | 1273 | /* increment our consumer index... */ |
1274 | pp->rsp_consumer = mv_inc_q_index(&pp->rsp_consumer); | 1274 | pp->rsp_consumer = mv_inc_q_index(&pp->rsp_consumer); |
1275 | 1275 | ||
1276 | /* and, until we do NCQ, there should only be 1 CRPB waiting */ | 1276 | /* and, until we do NCQ, there should only be 1 CRPB waiting */ |
1277 | assert(((readl(port_mmio + EDMA_RSP_Q_IN_PTR_OFS) >> | 1277 | WARN_ON(((readl(port_mmio + EDMA_RSP_Q_IN_PTR_OFS) >> |
1278 | EDMA_RSP_Q_PTR_SHIFT) & MV_MAX_Q_DEPTH_MASK) == | 1278 | EDMA_RSP_Q_PTR_SHIFT) & MV_MAX_Q_DEPTH_MASK) != |
1279 | pp->rsp_consumer); | 1279 | pp->rsp_consumer); |
1280 | 1280 | ||
1281 | /* write out our inc'd consumer index so EDMA knows we're caught up */ | 1281 | /* write out our inc'd consumer index so EDMA knows we're caught up */ |
1282 | out_ptr &= EDMA_RSP_Q_BASE_LO_MASK; | 1282 | out_ptr &= EDMA_RSP_Q_BASE_LO_MASK; |