aboutsummaryrefslogtreecommitdiffstats
path: root/drivers/net/ethernet/broadcom/cnic.c
diff options
context:
space:
mode:
authorMichael Chan <mchan@broadcom.com>2012-12-06 05:33:09 -0500
committerDavid S. Miller <davem@davemloft.net>2012-12-07 12:44:01 -0500
commit2bc4078e92b28375a762d7236c1c9619eecab315 (patch)
tree420fb018df8b016e9e10927d783140f4b9145964 /drivers/net/ethernet/broadcom/cnic.c
parente503e0662447ce2bd7c0a73c90395c78ebee494c (diff)
bnx2: Add BNX2 prefix to descriptor structures and macros
for namespace consistency. Signed-off-by: Michael Chan <mchan@broadcom.com> Signed-off-by: David S. Miller <davem@davemloft.net>
Diffstat (limited to 'drivers/net/ethernet/broadcom/cnic.c')
-rw-r--r--drivers/net/ethernet/broadcom/cnic.c70
1 files changed, 35 insertions, 35 deletions
diff --git a/drivers/net/ethernet/broadcom/cnic.c b/drivers/net/ethernet/broadcom/cnic.c
index 091c60a9897c..756a2a771291 100644
--- a/drivers/net/ethernet/broadcom/cnic.c
+++ b/drivers/net/ethernet/broadcom/cnic.c
@@ -724,7 +724,7 @@ static void cnic_free_dma(struct cnic_dev *dev, struct cnic_dma *dma)
724 724
725 for (i = 0; i < dma->num_pages; i++) { 725 for (i = 0; i < dma->num_pages; i++) {
726 if (dma->pg_arr[i]) { 726 if (dma->pg_arr[i]) {
727 dma_free_coherent(&dev->pcidev->dev, BCM_PAGE_SIZE, 727 dma_free_coherent(&dev->pcidev->dev, BNX2_PAGE_SIZE,
728 dma->pg_arr[i], dma->pg_map_arr[i]); 728 dma->pg_arr[i], dma->pg_map_arr[i]);
729 dma->pg_arr[i] = NULL; 729 dma->pg_arr[i] = NULL;
730 } 730 }
@@ -783,7 +783,7 @@ static int cnic_alloc_dma(struct cnic_dev *dev, struct cnic_dma *dma,
783 783
784 for (i = 0; i < pages; i++) { 784 for (i = 0; i < pages; i++) {
785 dma->pg_arr[i] = dma_alloc_coherent(&dev->pcidev->dev, 785 dma->pg_arr[i] = dma_alloc_coherent(&dev->pcidev->dev,
786 BCM_PAGE_SIZE, 786 BNX2_PAGE_SIZE,
787 &dma->pg_map_arr[i], 787 &dma->pg_map_arr[i],
788 GFP_ATOMIC); 788 GFP_ATOMIC);
789 if (dma->pg_arr[i] == NULL) 789 if (dma->pg_arr[i] == NULL)
@@ -792,8 +792,8 @@ static int cnic_alloc_dma(struct cnic_dev *dev, struct cnic_dma *dma,
792 if (!use_pg_tbl) 792 if (!use_pg_tbl)
793 return 0; 793 return 0;
794 794
795 dma->pgtbl_size = ((pages * 8) + BCM_PAGE_SIZE - 1) & 795 dma->pgtbl_size = ((pages * 8) + BNX2_PAGE_SIZE - 1) &
796 ~(BCM_PAGE_SIZE - 1); 796 ~(BNX2_PAGE_SIZE - 1);
797 dma->pgtbl = dma_alloc_coherent(&dev->pcidev->dev, dma->pgtbl_size, 797 dma->pgtbl = dma_alloc_coherent(&dev->pcidev->dev, dma->pgtbl_size,
798 &dma->pgtbl_map, GFP_ATOMIC); 798 &dma->pgtbl_map, GFP_ATOMIC);
799 if (dma->pgtbl == NULL) 799 if (dma->pgtbl == NULL)
@@ -898,8 +898,8 @@ static int cnic_alloc_context(struct cnic_dev *dev)
898 if (CHIP_NUM(cp) == CHIP_NUM_5709) { 898 if (CHIP_NUM(cp) == CHIP_NUM_5709) {
899 int i, k, arr_size; 899 int i, k, arr_size;
900 900
901 cp->ctx_blk_size = BCM_PAGE_SIZE; 901 cp->ctx_blk_size = BNX2_PAGE_SIZE;
902 cp->cids_per_blk = BCM_PAGE_SIZE / 128; 902 cp->cids_per_blk = BNX2_PAGE_SIZE / 128;
903 arr_size = BNX2_MAX_CID / cp->cids_per_blk * 903 arr_size = BNX2_MAX_CID / cp->cids_per_blk *
904 sizeof(struct cnic_ctx); 904 sizeof(struct cnic_ctx);
905 cp->ctx_arr = kzalloc(arr_size, GFP_KERNEL); 905 cp->ctx_arr = kzalloc(arr_size, GFP_KERNEL);
@@ -931,7 +931,7 @@ static int cnic_alloc_context(struct cnic_dev *dev)
931 for (i = 0; i < cp->ctx_blks; i++) { 931 for (i = 0; i < cp->ctx_blks; i++) {
932 cp->ctx_arr[i].ctx = 932 cp->ctx_arr[i].ctx =
933 dma_alloc_coherent(&dev->pcidev->dev, 933 dma_alloc_coherent(&dev->pcidev->dev,
934 BCM_PAGE_SIZE, 934 BNX2_PAGE_SIZE,
935 &cp->ctx_arr[i].mapping, 935 &cp->ctx_arr[i].mapping,
936 GFP_KERNEL); 936 GFP_KERNEL);
937 if (cp->ctx_arr[i].ctx == NULL) 937 if (cp->ctx_arr[i].ctx == NULL)
@@ -1011,7 +1011,7 @@ static int __cnic_alloc_uio_rings(struct cnic_uio_dev *udev, int pages)
1011 if (udev->l2_ring) 1011 if (udev->l2_ring)
1012 return 0; 1012 return 0;
1013 1013
1014 udev->l2_ring_size = pages * BCM_PAGE_SIZE; 1014 udev->l2_ring_size = pages * BNX2_PAGE_SIZE;
1015 udev->l2_ring = dma_alloc_coherent(&udev->pdev->dev, udev->l2_ring_size, 1015 udev->l2_ring = dma_alloc_coherent(&udev->pdev->dev, udev->l2_ring_size,
1016 &udev->l2_ring_map, 1016 &udev->l2_ring_map,
1017 GFP_KERNEL | __GFP_COMP); 1017 GFP_KERNEL | __GFP_COMP);
@@ -2898,7 +2898,7 @@ static int cnic_l2_completion(struct cnic_local *cp)
2898 u16 hw_cons, sw_cons; 2898 u16 hw_cons, sw_cons;
2899 struct cnic_uio_dev *udev = cp->udev; 2899 struct cnic_uio_dev *udev = cp->udev;
2900 union eth_rx_cqe *cqe, *cqe_ring = (union eth_rx_cqe *) 2900 union eth_rx_cqe *cqe, *cqe_ring = (union eth_rx_cqe *)
2901 (udev->l2_ring + (2 * BCM_PAGE_SIZE)); 2901 (udev->l2_ring + (2 * BNX2_PAGE_SIZE));
2902 u32 cmd; 2902 u32 cmd;
2903 int comp = 0; 2903 int comp = 0;
2904 2904
@@ -4366,7 +4366,7 @@ static int cnic_setup_5709_context(struct cnic_dev *dev, int valid)
4366 u32 idx = cp->ctx_arr[i].cid / cp->cids_per_blk; 4366 u32 idx = cp->ctx_arr[i].cid / cp->cids_per_blk;
4367 u32 val; 4367 u32 val;
4368 4368
4369 memset(cp->ctx_arr[i].ctx, 0, BCM_PAGE_SIZE); 4369 memset(cp->ctx_arr[i].ctx, 0, BNX2_PAGE_SIZE);
4370 4370
4371 CNIC_WR(dev, BNX2_CTX_HOST_PAGE_TBL_DATA0, 4371 CNIC_WR(dev, BNX2_CTX_HOST_PAGE_TBL_DATA0,
4372 (cp->ctx_arr[i].mapping & 0xffffffff) | valid_bit); 4372 (cp->ctx_arr[i].mapping & 0xffffffff) | valid_bit);
@@ -4508,7 +4508,7 @@ static void cnic_init_bnx2_tx_ring(struct cnic_dev *dev)
4508 u32 cid_addr, tx_cid, sb_id; 4508 u32 cid_addr, tx_cid, sb_id;
4509 u32 val, offset0, offset1, offset2, offset3; 4509 u32 val, offset0, offset1, offset2, offset3;
4510 int i; 4510 int i;
4511 struct tx_bd *txbd; 4511 struct bnx2_tx_bd *txbd;
4512 dma_addr_t buf_map, ring_map = udev->l2_ring_map; 4512 dma_addr_t buf_map, ring_map = udev->l2_ring_map;
4513 struct status_block *s_blk = cp->status_blk.gen; 4513 struct status_block *s_blk = cp->status_blk.gen;
4514 4514
@@ -4554,7 +4554,7 @@ static void cnic_init_bnx2_tx_ring(struct cnic_dev *dev)
4554 txbd = udev->l2_ring; 4554 txbd = udev->l2_ring;
4555 4555
4556 buf_map = udev->l2_buf_map; 4556 buf_map = udev->l2_buf_map;
4557 for (i = 0; i < MAX_TX_DESC_CNT; i++, txbd++) { 4557 for (i = 0; i < BNX2_MAX_TX_DESC_CNT; i++, txbd++) {
4558 txbd->tx_bd_haddr_hi = (u64) buf_map >> 32; 4558 txbd->tx_bd_haddr_hi = (u64) buf_map >> 32;
4559 txbd->tx_bd_haddr_lo = (u64) buf_map & 0xffffffff; 4559 txbd->tx_bd_haddr_lo = (u64) buf_map & 0xffffffff;
4560 } 4560 }
@@ -4574,7 +4574,7 @@ static void cnic_init_bnx2_rx_ring(struct cnic_dev *dev)
4574 struct cnic_uio_dev *udev = cp->udev; 4574 struct cnic_uio_dev *udev = cp->udev;
4575 u32 cid_addr, sb_id, val, coal_reg, coal_val; 4575 u32 cid_addr, sb_id, val, coal_reg, coal_val;
4576 int i; 4576 int i;
4577 struct rx_bd *rxbd; 4577 struct bnx2_rx_bd *rxbd;
4578 struct status_block *s_blk = cp->status_blk.gen; 4578 struct status_block *s_blk = cp->status_blk.gen;
4579 dma_addr_t ring_map = udev->l2_ring_map; 4579 dma_addr_t ring_map = udev->l2_ring_map;
4580 4580
@@ -4610,8 +4610,8 @@ static void cnic_init_bnx2_rx_ring(struct cnic_dev *dev)
4610 val = BNX2_L2CTX_L2_STATUSB_NUM(sb_id); 4610 val = BNX2_L2CTX_L2_STATUSB_NUM(sb_id);
4611 cnic_ctx_wr(dev, cid_addr, BNX2_L2CTX_HOST_BDIDX, val); 4611 cnic_ctx_wr(dev, cid_addr, BNX2_L2CTX_HOST_BDIDX, val);
4612 4612
4613 rxbd = udev->l2_ring + BCM_PAGE_SIZE; 4613 rxbd = udev->l2_ring + BNX2_PAGE_SIZE;
4614 for (i = 0; i < MAX_RX_DESC_CNT; i++, rxbd++) { 4614 for (i = 0; i < BNX2_MAX_RX_DESC_CNT; i++, rxbd++) {
4615 dma_addr_t buf_map; 4615 dma_addr_t buf_map;
4616 int n = (i % cp->l2_rx_ring_size) + 1; 4616 int n = (i % cp->l2_rx_ring_size) + 1;
4617 4617
@@ -4621,11 +4621,11 @@ static void cnic_init_bnx2_rx_ring(struct cnic_dev *dev)
4621 rxbd->rx_bd_haddr_hi = (u64) buf_map >> 32; 4621 rxbd->rx_bd_haddr_hi = (u64) buf_map >> 32;
4622 rxbd->rx_bd_haddr_lo = (u64) buf_map & 0xffffffff; 4622 rxbd->rx_bd_haddr_lo = (u64) buf_map & 0xffffffff;
4623 } 4623 }
4624 val = (u64) (ring_map + BCM_PAGE_SIZE) >> 32; 4624 val = (u64) (ring_map + BNX2_PAGE_SIZE) >> 32;
4625 cnic_ctx_wr(dev, cid_addr, BNX2_L2CTX_NX_BDHADDR_HI, val); 4625 cnic_ctx_wr(dev, cid_addr, BNX2_L2CTX_NX_BDHADDR_HI, val);
4626 rxbd->rx_bd_haddr_hi = val; 4626 rxbd->rx_bd_haddr_hi = val;
4627 4627
4628 val = (u64) (ring_map + BCM_PAGE_SIZE) & 0xffffffff; 4628 val = (u64) (ring_map + BNX2_PAGE_SIZE) & 0xffffffff;
4629 cnic_ctx_wr(dev, cid_addr, BNX2_L2CTX_NX_BDHADDR_LO, val); 4629 cnic_ctx_wr(dev, cid_addr, BNX2_L2CTX_NX_BDHADDR_LO, val);
4630 rxbd->rx_bd_haddr_lo = val; 4630 rxbd->rx_bd_haddr_lo = val;
4631 4631
@@ -4691,10 +4691,10 @@ static int cnic_start_bnx2_hw(struct cnic_dev *dev)
4691 4691
4692 val = CNIC_RD(dev, BNX2_MQ_CONFIG); 4692 val = CNIC_RD(dev, BNX2_MQ_CONFIG);
4693 val &= ~BNX2_MQ_CONFIG_KNL_BYP_BLK_SIZE; 4693 val &= ~BNX2_MQ_CONFIG_KNL_BYP_BLK_SIZE;
4694 if (BCM_PAGE_BITS > 12) 4694 if (BNX2_PAGE_BITS > 12)
4695 val |= (12 - 8) << 4; 4695 val |= (12 - 8) << 4;
4696 else 4696 else
4697 val |= (BCM_PAGE_BITS - 8) << 4; 4697 val |= (BNX2_PAGE_BITS - 8) << 4;
4698 4698
4699 CNIC_WR(dev, BNX2_MQ_CONFIG, val); 4699 CNIC_WR(dev, BNX2_MQ_CONFIG, val);
4700 4700
@@ -4724,13 +4724,13 @@ static int cnic_start_bnx2_hw(struct cnic_dev *dev)
4724 4724
4725 /* Initialize the kernel work queue context. */ 4725 /* Initialize the kernel work queue context. */
4726 val = KRNLQ_TYPE_TYPE_KRNLQ | KRNLQ_SIZE_TYPE_SIZE | 4726 val = KRNLQ_TYPE_TYPE_KRNLQ | KRNLQ_SIZE_TYPE_SIZE |
4727 (BCM_PAGE_BITS - 8) | KRNLQ_FLAGS_QE_SELF_SEQ; 4727 (BNX2_PAGE_BITS - 8) | KRNLQ_FLAGS_QE_SELF_SEQ;
4728 cnic_ctx_wr(dev, kwq_cid_addr, L5_KRNLQ_TYPE, val); 4728 cnic_ctx_wr(dev, kwq_cid_addr, L5_KRNLQ_TYPE, val);
4729 4729
4730 val = (BCM_PAGE_SIZE / sizeof(struct kwqe) - 1) << 16; 4730 val = (BNX2_PAGE_SIZE / sizeof(struct kwqe) - 1) << 16;
4731 cnic_ctx_wr(dev, kwq_cid_addr, L5_KRNLQ_QE_SELF_SEQ_MAX, val); 4731 cnic_ctx_wr(dev, kwq_cid_addr, L5_KRNLQ_QE_SELF_SEQ_MAX, val);
4732 4732
4733 val = ((BCM_PAGE_SIZE / sizeof(struct kwqe)) << 16) | KWQ_PAGE_CNT; 4733 val = ((BNX2_PAGE_SIZE / sizeof(struct kwqe)) << 16) | KWQ_PAGE_CNT;
4734 cnic_ctx_wr(dev, kwq_cid_addr, L5_KRNLQ_PGTBL_NPAGES, val); 4734 cnic_ctx_wr(dev, kwq_cid_addr, L5_KRNLQ_PGTBL_NPAGES, val);
4735 4735
4736 val = (u32) ((u64) cp->kwq_info.pgtbl_map >> 32); 4736 val = (u32) ((u64) cp->kwq_info.pgtbl_map >> 32);
@@ -4750,13 +4750,13 @@ static int cnic_start_bnx2_hw(struct cnic_dev *dev)
4750 4750
4751 /* Initialize the kernel complete queue context. */ 4751 /* Initialize the kernel complete queue context. */
4752 val = KRNLQ_TYPE_TYPE_KRNLQ | KRNLQ_SIZE_TYPE_SIZE | 4752 val = KRNLQ_TYPE_TYPE_KRNLQ | KRNLQ_SIZE_TYPE_SIZE |
4753 (BCM_PAGE_BITS - 8) | KRNLQ_FLAGS_QE_SELF_SEQ; 4753 (BNX2_PAGE_BITS - 8) | KRNLQ_FLAGS_QE_SELF_SEQ;
4754 cnic_ctx_wr(dev, kcq_cid_addr, L5_KRNLQ_TYPE, val); 4754 cnic_ctx_wr(dev, kcq_cid_addr, L5_KRNLQ_TYPE, val);
4755 4755
4756 val = (BCM_PAGE_SIZE / sizeof(struct kcqe) - 1) << 16; 4756 val = (BNX2_PAGE_SIZE / sizeof(struct kcqe) - 1) << 16;
4757 cnic_ctx_wr(dev, kcq_cid_addr, L5_KRNLQ_QE_SELF_SEQ_MAX, val); 4757 cnic_ctx_wr(dev, kcq_cid_addr, L5_KRNLQ_QE_SELF_SEQ_MAX, val);
4758 4758
4759 val = ((BCM_PAGE_SIZE / sizeof(struct kcqe)) << 16) | KCQ_PAGE_CNT; 4759 val = ((BNX2_PAGE_SIZE / sizeof(struct kcqe)) << 16) | KCQ_PAGE_CNT;
4760 cnic_ctx_wr(dev, kcq_cid_addr, L5_KRNLQ_PGTBL_NPAGES, val); 4760 cnic_ctx_wr(dev, kcq_cid_addr, L5_KRNLQ_PGTBL_NPAGES, val);
4761 4761
4762 val = (u32) ((u64) cp->kcq1.dma.pgtbl_map >> 32); 4762 val = (u32) ((u64) cp->kcq1.dma.pgtbl_map >> 32);
@@ -4895,10 +4895,10 @@ static void cnic_init_bnx2x_tx_ring(struct cnic_dev *dev,
4895 u32 cli = cp->ethdev->iscsi_l2_client_id; 4895 u32 cli = cp->ethdev->iscsi_l2_client_id;
4896 u32 val; 4896 u32 val;
4897 4897
4898 memset(txbd, 0, BCM_PAGE_SIZE); 4898 memset(txbd, 0, BNX2_PAGE_SIZE);
4899 4899
4900 buf_map = udev->l2_buf_map; 4900 buf_map = udev->l2_buf_map;
4901 for (i = 0; i < MAX_TX_DESC_CNT; i += 3, txbd += 3) { 4901 for (i = 0; i < BNX2_MAX_TX_DESC_CNT; i += 3, txbd += 3) {
4902 struct eth_tx_start_bd *start_bd = &txbd->start_bd; 4902 struct eth_tx_start_bd *start_bd = &txbd->start_bd;
4903 struct eth_tx_parse_bd_e1x *pbd_e1x = 4903 struct eth_tx_parse_bd_e1x *pbd_e1x =
4904 &((txbd + 1)->parse_bd_e1x); 4904 &((txbd + 1)->parse_bd_e1x);
@@ -4954,9 +4954,9 @@ static void cnic_init_bnx2x_rx_ring(struct cnic_dev *dev,
4954 struct cnic_local *cp = dev->cnic_priv; 4954 struct cnic_local *cp = dev->cnic_priv;
4955 struct cnic_uio_dev *udev = cp->udev; 4955 struct cnic_uio_dev *udev = cp->udev;
4956 struct eth_rx_bd *rxbd = (struct eth_rx_bd *) (udev->l2_ring + 4956 struct eth_rx_bd *rxbd = (struct eth_rx_bd *) (udev->l2_ring +
4957 BCM_PAGE_SIZE); 4957 BNX2_PAGE_SIZE);
4958 struct eth_rx_cqe_next_page *rxcqe = (struct eth_rx_cqe_next_page *) 4958 struct eth_rx_cqe_next_page *rxcqe = (struct eth_rx_cqe_next_page *)
4959 (udev->l2_ring + (2 * BCM_PAGE_SIZE)); 4959 (udev->l2_ring + (2 * BNX2_PAGE_SIZE));
4960 struct host_sp_status_block *sb = cp->bnx2x_def_status_blk; 4960 struct host_sp_status_block *sb = cp->bnx2x_def_status_blk;
4961 int i; 4961 int i;
4962 u32 cli = cp->ethdev->iscsi_l2_client_id; 4962 u32 cli = cp->ethdev->iscsi_l2_client_id;
@@ -4980,20 +4980,20 @@ static void cnic_init_bnx2x_rx_ring(struct cnic_dev *dev,
4980 rxbd->addr_lo = cpu_to_le32(buf_map & 0xffffffff); 4980 rxbd->addr_lo = cpu_to_le32(buf_map & 0xffffffff);
4981 } 4981 }
4982 4982
4983 val = (u64) (ring_map + BCM_PAGE_SIZE) >> 32; 4983 val = (u64) (ring_map + BNX2_PAGE_SIZE) >> 32;
4984 rxbd->addr_hi = cpu_to_le32(val); 4984 rxbd->addr_hi = cpu_to_le32(val);
4985 data->rx.bd_page_base.hi = cpu_to_le32(val); 4985 data->rx.bd_page_base.hi = cpu_to_le32(val);
4986 4986
4987 val = (u64) (ring_map + BCM_PAGE_SIZE) & 0xffffffff; 4987 val = (u64) (ring_map + BNX2_PAGE_SIZE) & 0xffffffff;
4988 rxbd->addr_lo = cpu_to_le32(val); 4988 rxbd->addr_lo = cpu_to_le32(val);
4989 data->rx.bd_page_base.lo = cpu_to_le32(val); 4989 data->rx.bd_page_base.lo = cpu_to_le32(val);
4990 4990
4991 rxcqe += BNX2X_MAX_RCQ_DESC_CNT; 4991 rxcqe += BNX2X_MAX_RCQ_DESC_CNT;
4992 val = (u64) (ring_map + (2 * BCM_PAGE_SIZE)) >> 32; 4992 val = (u64) (ring_map + (2 * BNX2_PAGE_SIZE)) >> 32;
4993 rxcqe->addr_hi = cpu_to_le32(val); 4993 rxcqe->addr_hi = cpu_to_le32(val);
4994 data->rx.cqe_page_base.hi = cpu_to_le32(val); 4994 data->rx.cqe_page_base.hi = cpu_to_le32(val);
4995 4995
4996 val = (u64) (ring_map + (2 * BCM_PAGE_SIZE)) & 0xffffffff; 4996 val = (u64) (ring_map + (2 * BNX2_PAGE_SIZE)) & 0xffffffff;
4997 rxcqe->addr_lo = cpu_to_le32(val); 4997 rxcqe->addr_lo = cpu_to_le32(val);
4998 data->rx.cqe_page_base.lo = cpu_to_le32(val); 4998 data->rx.cqe_page_base.lo = cpu_to_le32(val);
4999 4999
@@ -5258,8 +5258,8 @@ static void cnic_shutdown_rings(struct cnic_dev *dev)
5258 msleep(10); 5258 msleep(10);
5259 } 5259 }
5260 clear_bit(CNIC_LCL_FL_RINGS_INITED, &cp->cnic_local_flags); 5260 clear_bit(CNIC_LCL_FL_RINGS_INITED, &cp->cnic_local_flags);
5261 rx_ring = udev->l2_ring + BCM_PAGE_SIZE; 5261 rx_ring = udev->l2_ring + BNX2_PAGE_SIZE;
5262 memset(rx_ring, 0, BCM_PAGE_SIZE); 5262 memset(rx_ring, 0, BNX2_PAGE_SIZE);
5263} 5263}
5264 5264
5265static int cnic_register_netdev(struct cnic_dev *dev) 5265static int cnic_register_netdev(struct cnic_dev *dev)