aboutsummaryrefslogtreecommitdiffstats
path: root/drivers
diff options
context:
space:
mode:
authorAlan Cox <alan@linux.intel.com>2009-08-19 13:21:50 -0400
committerGreg Kroah-Hartman <gregkh@suse.de>2009-09-15 15:02:26 -0400
commit37628606661a8a1d3abfa5bb898426a38fa62b73 (patch)
tree40d3ba821c49ea87ef319c5f07eadb536ebc2da9 /drivers
parent25ad00bba42d49e37e81e7c449bbcc86a71f2894 (diff)
Staging: et131x: spinlocks
Switch to the more normal "flags" naming. Also fix up the nested use of spin_lock_irqsave Signed-off-by: Alan Cox <alan@linux.intel.com> Signed-off-by: Greg Kroah-Hartman <gregkh@suse.de>
Diffstat (limited to 'drivers')
-rw-r--r--drivers/staging/et131x/et1310_phy.c10
-rw-r--r--drivers/staging/et131x/et1310_pm.c6
-rw-r--r--drivers/staging/et131x/et1310_rx.c28
-rw-r--r--drivers/staging/et131x/et1310_tx.c74
-rw-r--r--drivers/staging/et131x/et131x_initpci.c6
-rw-r--r--drivers/staging/et131x/et131x_netdev.c14
6 files changed, 69 insertions, 69 deletions
diff --git a/drivers/staging/et131x/et1310_phy.c b/drivers/staging/et131x/et1310_phy.c
index 82fcb6678e24..d9c7a4486b41 100644
--- a/drivers/staging/et131x/et1310_phy.c
+++ b/drivers/staging/et131x/et1310_phy.c
@@ -484,7 +484,7 @@ void et131x_Mii_check(struct et131x_adapter *etdev,
484 uint32_t uiMdiMdix; 484 uint32_t uiMdiMdix;
485 uint32_t uiMasterSlave; 485 uint32_t uiMasterSlave;
486 uint32_t uiPolarity; 486 uint32_t uiPolarity;
487 unsigned long lockflags; 487 unsigned long flags;
488 488
489 DBG_ENTER(et131x_dbginfo); 489 DBG_ENTER(et131x_dbginfo);
490 490
@@ -495,12 +495,12 @@ void et131x_Mii_check(struct et131x_adapter *etdev,
495 /* Update our state variables and indicate the 495 /* Update our state variables and indicate the
496 * connected state 496 * connected state
497 */ 497 */
498 spin_lock_irqsave(&etdev->Lock, lockflags); 498 spin_lock_irqsave(&etdev->Lock, flags);
499 499
500 etdev->MediaState = NETIF_STATUS_MEDIA_CONNECT; 500 etdev->MediaState = NETIF_STATUS_MEDIA_CONNECT;
501 MP_CLEAR_FLAG(etdev, fMP_ADAPTER_LINK_DETECTION); 501 MP_CLEAR_FLAG(etdev, fMP_ADAPTER_LINK_DETECTION);
502 502
503 spin_unlock_irqrestore(&etdev->Lock, lockflags); 503 spin_unlock_irqrestore(&etdev->Lock, flags);
504 504
505 /* Don't indicate state if we're in loopback mode */ 505 /* Don't indicate state if we're in loopback mode */
506 if (etdev->RegistryPhyLoopbk == false) 506 if (etdev->RegistryPhyLoopbk == false)
@@ -533,11 +533,11 @@ void et131x_Mii_check(struct et131x_adapter *etdev,
533 (etdev, fMP_ADAPTER_LINK_DETECTION)) 533 (etdev, fMP_ADAPTER_LINK_DETECTION))
534 || (etdev->MediaState == 534 || (etdev->MediaState ==
535 NETIF_STATUS_MEDIA_DISCONNECT)) { 535 NETIF_STATUS_MEDIA_DISCONNECT)) {
536 spin_lock_irqsave(&etdev->Lock, lockflags); 536 spin_lock_irqsave(&etdev->Lock, flags);
537 etdev->MediaState = 537 etdev->MediaState =
538 NETIF_STATUS_MEDIA_DISCONNECT; 538 NETIF_STATUS_MEDIA_DISCONNECT;
539 spin_unlock_irqrestore(&etdev->Lock, 539 spin_unlock_irqrestore(&etdev->Lock,
540 lockflags); 540 flags);
541 541
542 /* Only indicate state if we're in loopback 542 /* Only indicate state if we're in loopback
543 * mode 543 * mode
diff --git a/drivers/staging/et131x/et1310_pm.c b/drivers/staging/et131x/et1310_pm.c
index 864d3ad3fcdd..a50741168f62 100644
--- a/drivers/staging/et131x/et1310_pm.c
+++ b/drivers/staging/et131x/et1310_pm.c
@@ -119,7 +119,7 @@ extern dbg_info_t *et131x_dbginfo;
119 */ 119 */
120void EnablePhyComa(struct et131x_adapter *etdev) 120void EnablePhyComa(struct et131x_adapter *etdev)
121{ 121{
122 unsigned long lockflags; 122 unsigned long flags;
123 PM_CSR_t GlobalPmCSR; 123 PM_CSR_t GlobalPmCSR;
124 int32_t LoopCounter = 10; 124 int32_t LoopCounter = 10;
125 125
@@ -134,9 +134,9 @@ void EnablePhyComa(struct et131x_adapter *etdev)
134 etdev->PoMgmt.PowerDownDuplex = etdev->AiForceDpx; 134 etdev->PoMgmt.PowerDownDuplex = etdev->AiForceDpx;
135 135
136 /* Stop sending packets. */ 136 /* Stop sending packets. */
137 spin_lock_irqsave(&etdev->SendHWLock, lockflags); 137 spin_lock_irqsave(&etdev->SendHWLock, flags);
138 MP_SET_FLAG(etdev, fMP_ADAPTER_LOWER_POWER); 138 MP_SET_FLAG(etdev, fMP_ADAPTER_LOWER_POWER);
139 spin_unlock_irqrestore(&etdev->SendHWLock, lockflags); 139 spin_unlock_irqrestore(&etdev->SendHWLock, flags);
140 140
141 /* Wait for outstanding Receive packets */ 141 /* Wait for outstanding Receive packets */
142 while ((MP_GET_RCV_REF(etdev) != 0) && (LoopCounter-- > 0)) 142 while ((MP_GET_RCV_REF(etdev) != 0) && (LoopCounter-- > 0))
diff --git a/drivers/staging/et131x/et1310_rx.c b/drivers/staging/et131x/et1310_rx.c
index f8cc6a607d2e..9dc08fb4bf27 100644
--- a/drivers/staging/et131x/et1310_rx.c
+++ b/drivers/staging/et131x/et1310_rx.c
@@ -685,7 +685,7 @@ void ConfigRxDmaRegs(struct et131x_adapter *etdev)
685 PFBR_DESC_t pFbrEntry; 685 PFBR_DESC_t pFbrEntry;
686 uint32_t iEntry; 686 uint32_t iEntry;
687 RXDMA_PSR_NUM_DES_t psr_num_des; 687 RXDMA_PSR_NUM_DES_t psr_num_des;
688 unsigned long lockflags; 688 unsigned long flags;
689 689
690 DBG_ENTER(et131x_dbginfo); 690 DBG_ENTER(et131x_dbginfo);
691 691
@@ -718,7 +718,7 @@ void ConfigRxDmaRegs(struct et131x_adapter *etdev)
718 writel((psr_num_des.bits.psr_ndes * LO_MARK_PERCENT_FOR_PSR) / 100, 718 writel((psr_num_des.bits.psr_ndes * LO_MARK_PERCENT_FOR_PSR) / 100,
719 &pRxDma->psr_min_des.value); 719 &pRxDma->psr_min_des.value);
720 720
721 spin_lock_irqsave(&etdev->RcvLock, lockflags); 721 spin_lock_irqsave(&etdev->RcvLock, flags);
722 722
723 /* These local variables track the PSR in the adapter structure */ 723 /* These local variables track the PSR in the adapter structure */
724 pRxLocal->local_psr_full.bits.psr_full = 0; 724 pRxLocal->local_psr_full.bits.psr_full = 0;
@@ -801,7 +801,7 @@ void ConfigRxDmaRegs(struct et131x_adapter *etdev)
801 */ 801 */
802 writel(etdev->RegistryRxTimeInterval, &pRxDma->max_pkt_time.value); 802 writel(etdev->RegistryRxTimeInterval, &pRxDma->max_pkt_time.value);
803 803
804 spin_unlock_irqrestore(&etdev->RcvLock, lockflags); 804 spin_unlock_irqrestore(&etdev->RcvLock, flags);
805 805
806 DBG_LEAVE(et131x_dbginfo); 806 DBG_LEAVE(et131x_dbginfo);
807} 807}
@@ -914,7 +914,7 @@ PMP_RFD nic_rx_pkts(struct et131x_adapter *etdev)
914 PMP_RFD pMpRfd; 914 PMP_RFD pMpRfd;
915 uint32_t nIndex; 915 uint32_t nIndex;
916 uint8_t *pBufVa; 916 uint8_t *pBufVa;
917 unsigned long lockflags; 917 unsigned long flags;
918 struct list_head *element; 918 struct list_head *element;
919 uint8_t ringIndex; 919 uint8_t ringIndex;
920 uint16_t bufferIndex; 920 uint16_t bufferIndex;
@@ -1013,7 +1013,7 @@ PMP_RFD nic_rx_pkts(struct et131x_adapter *etdev)
1013 } 1013 }
1014 1014
1015 /* Get and fill the RFD. */ 1015 /* Get and fill the RFD. */
1016 spin_lock_irqsave(&etdev->RcvLock, lockflags); 1016 spin_lock_irqsave(&etdev->RcvLock, flags);
1017 1017
1018 pMpRfd = NULL; 1018 pMpRfd = NULL;
1019 element = pRxLocal->RecvList.next; 1019 element = pRxLocal->RecvList.next;
@@ -1023,14 +1023,14 @@ PMP_RFD nic_rx_pkts(struct et131x_adapter *etdev)
1023 DBG_RX(et131x_dbginfo, 1023 DBG_RX(et131x_dbginfo,
1024 "NULL RFD returned from RecvList via list_entry()\n"); 1024 "NULL RFD returned from RecvList via list_entry()\n");
1025 DBG_RX_LEAVE(et131x_dbginfo); 1025 DBG_RX_LEAVE(et131x_dbginfo);
1026 spin_unlock_irqrestore(&etdev->RcvLock, lockflags); 1026 spin_unlock_irqrestore(&etdev->RcvLock, flags);
1027 return NULL; 1027 return NULL;
1028 } 1028 }
1029 1029
1030 list_del(&pMpRfd->list_node); 1030 list_del(&pMpRfd->list_node);
1031 pRxLocal->nReadyRecv--; 1031 pRxLocal->nReadyRecv--;
1032 1032
1033 spin_unlock_irqrestore(&etdev->RcvLock, lockflags); 1033 spin_unlock_irqrestore(&etdev->RcvLock, flags);
1034 1034
1035 pMpRfd->iBufferIndex = bufferIndex; 1035 pMpRfd->iBufferIndex = bufferIndex;
1036 pMpRfd->iRingIndex = ringIndex; 1036 pMpRfd->iRingIndex = ringIndex;
@@ -1260,9 +1260,9 @@ void et131x_handle_recv_interrupt(struct et131x_adapter *etdev)
1260 * Besides, we don't really need (at this point) the 1260 * Besides, we don't really need (at this point) the
1261 * pending list anyway. 1261 * pending list anyway.
1262 */ 1262 */
1263 /* spin_lock_irqsave( &etdev->RcvPendLock, lockflags ); 1263 /* spin_lock_irqsave( &etdev->RcvPendLock, flags );
1264 * list_add_tail( &pMpRfd->list_node, &etdev->RxRing.RecvPendingList ); 1264 * list_add_tail( &pMpRfd->list_node, &etdev->RxRing.RecvPendingList );
1265 * spin_unlock_irqrestore( &etdev->RcvPendLock, lockflags ); 1265 * spin_unlock_irqrestore( &etdev->RcvPendLock, flags );
1266 */ 1266 */
1267 1267
1268 /* Update the number of outstanding Recvs */ 1268 /* Update the number of outstanding Recvs */
@@ -1302,7 +1302,7 @@ void nic_return_rfd(struct et131x_adapter *etdev, PMP_RFD pMpRfd)
1302 struct _RXDMA_t __iomem *pRxDma = &etdev->CSRAddress->rxdma; 1302 struct _RXDMA_t __iomem *pRxDma = &etdev->CSRAddress->rxdma;
1303 uint16_t bi = pMpRfd->iBufferIndex; 1303 uint16_t bi = pMpRfd->iBufferIndex;
1304 uint8_t ri = pMpRfd->iRingIndex; 1304 uint8_t ri = pMpRfd->iRingIndex;
1305 unsigned long lockflags; 1305 unsigned long flags;
1306 1306
1307 DBG_RX_ENTER(et131x_dbginfo); 1307 DBG_RX_ENTER(et131x_dbginfo);
1308 1308
@@ -1314,7 +1314,7 @@ void nic_return_rfd(struct et131x_adapter *etdev, PMP_RFD pMpRfd)
1314 (ri == 0 && bi < pRxLocal->Fbr0NumEntries) || 1314 (ri == 0 && bi < pRxLocal->Fbr0NumEntries) ||
1315#endif 1315#endif
1316 (ri == 1 && bi < pRxLocal->Fbr1NumEntries)) { 1316 (ri == 1 && bi < pRxLocal->Fbr1NumEntries)) {
1317 spin_lock_irqsave(&etdev->FbrLock, lockflags); 1317 spin_lock_irqsave(&etdev->FbrLock, flags);
1318 1318
1319 if (ri == 1) { 1319 if (ri == 1) {
1320 PFBR_DESC_t pNextDesc = 1320 PFBR_DESC_t pNextDesc =
@@ -1362,7 +1362,7 @@ void nic_return_rfd(struct et131x_adapter *etdev, PMP_RFD pMpRfd)
1362 &pRxDma->fbr0_full_offset.value); 1362 &pRxDma->fbr0_full_offset.value);
1363 } 1363 }
1364#endif 1364#endif
1365 spin_unlock_irqrestore(&etdev->FbrLock, lockflags); 1365 spin_unlock_irqrestore(&etdev->FbrLock, flags);
1366 } else { 1366 } else {
1367 DBG_ERROR(et131x_dbginfo, 1367 DBG_ERROR(et131x_dbginfo,
1368 "NICReturnRFD illegal Buffer Index returned\n"); 1368 "NICReturnRFD illegal Buffer Index returned\n");
@@ -1371,10 +1371,10 @@ void nic_return_rfd(struct et131x_adapter *etdev, PMP_RFD pMpRfd)
1371 /* The processing on this RFD is done, so put it back on the tail of 1371 /* The processing on this RFD is done, so put it back on the tail of
1372 * our list 1372 * our list
1373 */ 1373 */
1374 spin_lock_irqsave(&etdev->RcvLock, lockflags); 1374 spin_lock_irqsave(&etdev->RcvLock, flags);
1375 list_add_tail(&pMpRfd->list_node, &pRxLocal->RecvList); 1375 list_add_tail(&pMpRfd->list_node, &pRxLocal->RecvList);
1376 pRxLocal->nReadyRecv++; 1376 pRxLocal->nReadyRecv++;
1377 spin_unlock_irqrestore(&etdev->RcvLock, lockflags); 1377 spin_unlock_irqrestore(&etdev->RcvLock, flags);
1378 1378
1379 DBG_ASSERT(pRxLocal->nReadyRecv <= pRxLocal->NumRfd); 1379 DBG_ASSERT(pRxLocal->nReadyRecv <= pRxLocal->NumRfd);
1380 DBG_RX_LEAVE(et131x_dbginfo); 1380 DBG_RX_LEAVE(et131x_dbginfo);
diff --git a/drivers/staging/et131x/et1310_tx.c b/drivers/staging/et131x/et1310_tx.c
index db0f538a335a..a017d271e375 100644
--- a/drivers/staging/et131x/et1310_tx.c
+++ b/drivers/staging/et131x/et1310_tx.c
@@ -461,7 +461,7 @@ static int et131x_send_packet(struct sk_buff *skb,
461 int status = 0; 461 int status = 0;
462 PMP_TCB pMpTcb = NULL; 462 PMP_TCB pMpTcb = NULL;
463 uint16_t *pShBufVa; 463 uint16_t *pShBufVa;
464 unsigned long lockflags; 464 unsigned long flags;
465 465
466 DBG_TX_ENTER(et131x_dbginfo); 466 DBG_TX_ENTER(et131x_dbginfo);
467 467
@@ -482,12 +482,12 @@ static int et131x_send_packet(struct sk_buff *skb,
482 } 482 }
483 483
484 /* Get a TCB for this packet */ 484 /* Get a TCB for this packet */
485 spin_lock_irqsave(&etdev->TCBReadyQLock, lockflags); 485 spin_lock_irqsave(&etdev->TCBReadyQLock, flags);
486 486
487 pMpTcb = etdev->TxRing.TCBReadyQueueHead; 487 pMpTcb = etdev->TxRing.TCBReadyQueueHead;
488 488
489 if (pMpTcb == NULL) { 489 if (pMpTcb == NULL) {
490 spin_unlock_irqrestore(&etdev->TCBReadyQLock, lockflags); 490 spin_unlock_irqrestore(&etdev->TCBReadyQLock, flags);
491 491
492 DBG_WARNING(et131x_dbginfo, "Can't obtain a TCB\n"); 492 DBG_WARNING(et131x_dbginfo, "Can't obtain a TCB\n");
493 DBG_TX_LEAVE(et131x_dbginfo); 493 DBG_TX_LEAVE(et131x_dbginfo);
@@ -499,7 +499,7 @@ static int et131x_send_packet(struct sk_buff *skb,
499 if (etdev->TxRing.TCBReadyQueueHead == NULL) 499 if (etdev->TxRing.TCBReadyQueueHead == NULL)
500 etdev->TxRing.TCBReadyQueueTail = NULL; 500 etdev->TxRing.TCBReadyQueueTail = NULL;
501 501
502 spin_unlock_irqrestore(&etdev->TCBReadyQLock, lockflags); 502 spin_unlock_irqrestore(&etdev->TCBReadyQLock, flags);
503 503
504 pMpTcb->PacketLength = skb->len; 504 pMpTcb->PacketLength = skb->len;
505 pMpTcb->Packet = skb; 505 pMpTcb->Packet = skb;
@@ -522,7 +522,7 @@ static int et131x_send_packet(struct sk_buff *skb,
522 status = nic_send_packet(etdev, pMpTcb); 522 status = nic_send_packet(etdev, pMpTcb);
523 523
524 if (status != 0) { 524 if (status != 0) {
525 spin_lock_irqsave(&etdev->TCBReadyQLock, lockflags); 525 spin_lock_irqsave(&etdev->TCBReadyQLock, flags);
526 526
527 if (etdev->TxRing.TCBReadyQueueTail) { 527 if (etdev->TxRing.TCBReadyQueueTail) {
528 etdev->TxRing.TCBReadyQueueTail->Next = pMpTcb; 528 etdev->TxRing.TCBReadyQueueTail->Next = pMpTcb;
@@ -533,7 +533,7 @@ static int et131x_send_packet(struct sk_buff *skb,
533 533
534 etdev->TxRing.TCBReadyQueueTail = pMpTcb; 534 etdev->TxRing.TCBReadyQueueTail = pMpTcb;
535 535
536 spin_unlock_irqrestore(&etdev->TCBReadyQLock, lockflags); 536 spin_unlock_irqrestore(&etdev->TCBReadyQLock, flags);
537 537
538 DBG_TX_LEAVE(et131x_dbginfo); 538 DBG_TX_LEAVE(et131x_dbginfo);
539 return status; 539 return status;
@@ -561,7 +561,7 @@ static int nic_send_packet(struct et131x_adapter *etdev, PMP_TCB pMpTcb)
561 struct sk_buff *pPacket = pMpTcb->Packet; 561 struct sk_buff *pPacket = pMpTcb->Packet;
562 uint32_t FragListCount = skb_shinfo(pPacket)->nr_frags + 1; 562 uint32_t FragListCount = skb_shinfo(pPacket)->nr_frags + 1;
563 struct skb_frag_struct *pFragList = &skb_shinfo(pPacket)->frags[0]; 563 struct skb_frag_struct *pFragList = &skb_shinfo(pPacket)->frags[0];
564 unsigned long lockflags1, lockflags2; 564 unsigned long flags;
565 565
566 DBG_TX_ENTER(et131x_dbginfo); 566 DBG_TX_ENTER(et131x_dbginfo);
567 567
@@ -726,7 +726,7 @@ static int nic_send_packet(struct et131x_adapter *etdev, PMP_TCB pMpTcb)
726 pMpTcb->WrIndexStart = etdev->TxRing.txDmaReadyToSend; 726 pMpTcb->WrIndexStart = etdev->TxRing.txDmaReadyToSend;
727 pMpTcb->PacketStaleCount = 0; 727 pMpTcb->PacketStaleCount = 0;
728 728
729 spin_lock_irqsave(&etdev->SendHWLock, lockflags1); 729 spin_lock_irqsave(&etdev->SendHWLock, flags);
730 730
731 iThisCopy = 731 iThisCopy =
732 NUM_DESC_PER_RING_TX - etdev->TxRing.txDmaReadyToSend.bits.val; 732 NUM_DESC_PER_RING_TX - etdev->TxRing.txDmaReadyToSend.bits.val;
@@ -771,7 +771,7 @@ static int nic_send_packet(struct et131x_adapter *etdev, PMP_TCB pMpTcb)
771 pMpTcb->WrIndex.value = 771 pMpTcb->WrIndex.value =
772 etdev->TxRing.txDmaReadyToSend.value - 1; 772 etdev->TxRing.txDmaReadyToSend.value - 1;
773 773
774 spin_lock_irqsave(&etdev->TCBSendQLock, lockflags2); 774 spin_lock(&etdev->TCBSendQLock);
775 775
776 if (etdev->TxRing.CurrSendTail) 776 if (etdev->TxRing.CurrSendTail)
777 etdev->TxRing.CurrSendTail->Next = pMpTcb; 777 etdev->TxRing.CurrSendTail->Next = pMpTcb;
@@ -784,7 +784,7 @@ static int nic_send_packet(struct et131x_adapter *etdev, PMP_TCB pMpTcb)
784 784
785 etdev->TxRing.nBusySend++; 785 etdev->TxRing.nBusySend++;
786 786
787 spin_unlock_irqrestore(&etdev->TCBSendQLock, lockflags2); 787 spin_unlock(&etdev->TCBSendQLock);
788 788
789 /* Write the new write pointer back to the device. */ 789 /* Write the new write pointer back to the device. */
790 writel(etdev->TxRing.txDmaReadyToSend.value, 790 writel(etdev->TxRing.txDmaReadyToSend.value,
@@ -798,7 +798,7 @@ static int nic_send_packet(struct et131x_adapter *etdev, PMP_TCB pMpTcb)
798 &etdev->CSRAddress->global.watchdog_timer); 798 &etdev->CSRAddress->global.watchdog_timer);
799 } 799 }
800 800
801 spin_unlock_irqrestore(&etdev->SendHWLock, lockflags1); 801 spin_unlock_irqrestore(&etdev->SendHWLock, flags);
802 802
803 DBG_TX_LEAVE(et131x_dbginfo); 803 DBG_TX_LEAVE(et131x_dbginfo);
804 return 0; 804 return 0;
@@ -829,7 +829,7 @@ static int nic_send_packet(struct et131x_adapter *etdev, PMP_TCB pMpTcb)
829 TX_DESC_ENTRY_t *CurDescPostCopy = NULL; 829 TX_DESC_ENTRY_t *CurDescPostCopy = NULL;
830 uint32_t SlotsAvailable; 830 uint32_t SlotsAvailable;
831 DMA10W_t ServiceComplete; 831 DMA10W_t ServiceComplete;
832 unsigned int lockflags1, lockflags2; 832 unsigned int flags;
833 struct sk_buff *pPacket = pMpTcb->Packet; 833 struct sk_buff *pPacket = pMpTcb->Packet;
834 uint32_t FragListCount = skb_shinfo(pPacket)->nr_frags + 1; 834 uint32_t FragListCount = skb_shinfo(pPacket)->nr_frags + 1;
835 struct skb_frag_struct *pFragList = &skb_shinfo(pPacket)->frags[0]; 835 struct skb_frag_struct *pFragList = &skb_shinfo(pPacket)->frags[0];
@@ -875,7 +875,7 @@ static int nic_send_packet(struct et131x_adapter *etdev, PMP_TCB pMpTcb)
875 SegmentSize = (pPacket->len - pPacket->data_len) / 2; 875 SegmentSize = (pPacket->len - pPacket->data_len) / 2;
876 } 876 }
877 877
878 spin_lock_irqsave(&etdev->SendHWLock, lockflags1); 878 spin_lock_irqsave(&etdev->SendHWLock, flags);
879 879
880 if (etdev->TxRing.txDmaReadyToSend.bits.serv_req_wrap == 880 if (etdev->TxRing.txDmaReadyToSend.bits.serv_req_wrap ==
881 ServiceComplete.bits.serv_cpl_wrap) { 881 ServiceComplete.bits.serv_cpl_wrap) {
@@ -896,7 +896,7 @@ static int nic_send_packet(struct et131x_adapter *etdev, PMP_TCB pMpTcb)
896 if ((FragListCount + iSplitFirstElement) > SlotsAvailable) { 896 if ((FragListCount + iSplitFirstElement) > SlotsAvailable) {
897 DBG_WARNING(et131x_dbginfo, 897 DBG_WARNING(et131x_dbginfo,
898 "Not Enough Space in Tx Desc Ring\n"); 898 "Not Enough Space in Tx Desc Ring\n");
899 spin_unlock_irqrestore(&etdev->SendHWLock, lockflags1); 899 spin_unlock_irqrestore(&etdev->SendHWLock, flags);
900 return -ENOMEM; 900 return -ENOMEM;
901 } 901 }
902 902
@@ -1185,7 +1185,7 @@ static int nic_send_packet(struct et131x_adapter *etdev, PMP_TCB pMpTcb)
1185 NIC_MIN_PACKET_SIZE - pMpTcb->PacketLength); 1185 NIC_MIN_PACKET_SIZE - pMpTcb->PacketLength);
1186 } 1186 }
1187 1187
1188 spin_lock_irqsave(&etdev->TCBSendQLock, lockflags2); 1188 spin_lock(&etdev->TCBSendQLock);
1189 1189
1190 if (etdev->TxRing.CurrSendTail) 1190 if (etdev->TxRing.CurrSendTail)
1191 etdev->TxRing.CurrSendTail->Next = pMpTcb; 1191 etdev->TxRing.CurrSendTail->Next = pMpTcb;
@@ -1198,7 +1198,7 @@ static int nic_send_packet(struct et131x_adapter *etdev, PMP_TCB pMpTcb)
1198 1198
1199 etdev->TxRing.nBusySend++; 1199 etdev->TxRing.nBusySend++;
1200 1200
1201 spin_unlock_irqrestore(&etdev->TCBSendQLock, lockflags2); 1201 spin_unlock(&etdev->TCBSendQLock);
1202 1202
1203 /* Write the new write pointer back to the device. */ 1203 /* Write the new write pointer back to the device. */
1204 writel(etdev->TxRing.txDmaReadyToSend.value, 1204 writel(etdev->TxRing.txDmaReadyToSend.value,
@@ -1216,7 +1216,7 @@ static int nic_send_packet(struct et131x_adapter *etdev, PMP_TCB pMpTcb)
1216 &etdev->CSRAddress->global.watchdog_timer); 1216 &etdev->CSRAddress->global.watchdog_timer);
1217 } 1217 }
1218 1218
1219 spin_unlock_irqrestore(&etdev->SendHWLock, lockflags1); 1219 spin_unlock_irqrestore(&etdev->SendHWLock, flags);
1220 1220
1221 DBG_TX_LEAVE(et131x_dbginfo); 1221 DBG_TX_LEAVE(et131x_dbginfo);
1222 return 0; 1222 return 0;
@@ -1234,7 +1234,7 @@ static int nic_send_packet(struct et131x_adapter *etdev, PMP_TCB pMpTcb)
1234inline void et131x_free_send_packet(struct et131x_adapter *etdev, 1234inline void et131x_free_send_packet(struct et131x_adapter *etdev,
1235 PMP_TCB pMpTcb) 1235 PMP_TCB pMpTcb)
1236{ 1236{
1237 unsigned long lockflags; 1237 unsigned long flags;
1238 TX_DESC_ENTRY_t *desc = NULL; 1238 TX_DESC_ENTRY_t *desc = NULL;
1239 struct net_device_stats *stats = &etdev->net_stats; 1239 struct net_device_stats *stats = &etdev->net_stats;
1240 1240
@@ -1311,7 +1311,7 @@ inline void et131x_free_send_packet(struct et131x_adapter *etdev,
1311 memset(pMpTcb, 0, sizeof(MP_TCB)); 1311 memset(pMpTcb, 0, sizeof(MP_TCB));
1312 1312
1313 /* Add the TCB to the Ready Q */ 1313 /* Add the TCB to the Ready Q */
1314 spin_lock_irqsave(&etdev->TCBReadyQLock, lockflags); 1314 spin_lock_irqsave(&etdev->TCBReadyQLock, flags);
1315 1315
1316 etdev->Stats.opackets++; 1316 etdev->Stats.opackets++;
1317 1317
@@ -1324,7 +1324,7 @@ inline void et131x_free_send_packet(struct et131x_adapter *etdev,
1324 1324
1325 etdev->TxRing.TCBReadyQueueTail = pMpTcb; 1325 etdev->TxRing.TCBReadyQueueTail = pMpTcb;
1326 1326
1327 spin_unlock_irqrestore(&etdev->TCBReadyQLock, lockflags); 1327 spin_unlock_irqrestore(&etdev->TCBReadyQLock, flags);
1328 1328
1329 DBG_ASSERT(etdev->TxRing.nBusySend >= 0); 1329 DBG_ASSERT(etdev->TxRing.nBusySend >= 0);
1330} 1330}
@@ -1339,16 +1339,16 @@ void et131x_free_busy_send_packets(struct et131x_adapter *etdev)
1339{ 1339{
1340 PMP_TCB pMpTcb; 1340 PMP_TCB pMpTcb;
1341 struct list_head *pEntry; 1341 struct list_head *pEntry;
1342 unsigned long lockflags; 1342 unsigned long flags;
1343 uint32_t FreeCounter = 0; 1343 uint32_t FreeCounter = 0;
1344 1344
1345 DBG_ENTER(et131x_dbginfo); 1345 DBG_ENTER(et131x_dbginfo);
1346 1346
1347 while (!list_empty(&etdev->TxRing.SendWaitQueue)) { 1347 while (!list_empty(&etdev->TxRing.SendWaitQueue)) {
1348 spin_lock_irqsave(&etdev->SendWaitLock, lockflags); 1348 spin_lock_irqsave(&etdev->SendWaitLock, flags);
1349 1349
1350 etdev->TxRing.nWaitSend--; 1350 etdev->TxRing.nWaitSend--;
1351 spin_unlock_irqrestore(&etdev->SendWaitLock, lockflags); 1351 spin_unlock_irqrestore(&etdev->SendWaitLock, flags);
1352 1352
1353 pEntry = etdev->TxRing.SendWaitQueue.next; 1353 pEntry = etdev->TxRing.SendWaitQueue.next;
1354 } 1354 }
@@ -1356,7 +1356,7 @@ void et131x_free_busy_send_packets(struct et131x_adapter *etdev)
1356 etdev->TxRing.nWaitSend = 0; 1356 etdev->TxRing.nWaitSend = 0;
1357 1357
1358 /* Any packets being sent? Check the first TCB on the send list */ 1358 /* Any packets being sent? Check the first TCB on the send list */
1359 spin_lock_irqsave(&etdev->TCBSendQLock, lockflags); 1359 spin_lock_irqsave(&etdev->TCBSendQLock, flags);
1360 1360
1361 pMpTcb = etdev->TxRing.CurrSendHead; 1361 pMpTcb = etdev->TxRing.CurrSendHead;
1362 1362
@@ -1370,14 +1370,14 @@ void et131x_free_busy_send_packets(struct et131x_adapter *etdev)
1370 1370
1371 etdev->TxRing.nBusySend--; 1371 etdev->TxRing.nBusySend--;
1372 1372
1373 spin_unlock_irqrestore(&etdev->TCBSendQLock, lockflags); 1373 spin_unlock_irqrestore(&etdev->TCBSendQLock, flags);
1374 1374
1375 DBG_VERBOSE(et131x_dbginfo, "pMpTcb = 0x%p\n", pMpTcb); 1375 DBG_VERBOSE(et131x_dbginfo, "pMpTcb = 0x%p\n", pMpTcb);
1376 1376
1377 FreeCounter++; 1377 FreeCounter++;
1378 MP_FREE_SEND_PACKET_FUN(etdev, pMpTcb); 1378 MP_FREE_SEND_PACKET_FUN(etdev, pMpTcb);
1379 1379
1380 spin_lock_irqsave(&etdev->TCBSendQLock, lockflags); 1380 spin_lock_irqsave(&etdev->TCBSendQLock, flags);
1381 1381
1382 pMpTcb = etdev->TxRing.CurrSendHead; 1382 pMpTcb = etdev->TxRing.CurrSendHead;
1383 } 1383 }
@@ -1388,7 +1388,7 @@ void et131x_free_busy_send_packets(struct et131x_adapter *etdev)
1388 BUG(); 1388 BUG();
1389 } 1389 }
1390 1390
1391 spin_unlock_irqrestore(&etdev->TCBSendQLock, lockflags); 1391 spin_unlock_irqrestore(&etdev->TCBSendQLock, flags);
1392 1392
1393 etdev->TxRing.nBusySend = 0; 1393 etdev->TxRing.nBusySend = 0;
1394 1394
@@ -1429,7 +1429,7 @@ void et131x_handle_send_interrupt(struct et131x_adapter *etdev)
1429 */ 1429 */
1430static void et131x_update_tcb_list(struct et131x_adapter *etdev) 1430static void et131x_update_tcb_list(struct et131x_adapter *etdev)
1431{ 1431{
1432 unsigned long lockflags; 1432 unsigned long flags;
1433 DMA10W_t ServiceComplete; 1433 DMA10W_t ServiceComplete;
1434 PMP_TCB pMpTcb; 1434 PMP_TCB pMpTcb;
1435 1435
@@ -1439,7 +1439,7 @@ static void et131x_update_tcb_list(struct et131x_adapter *etdev)
1439 /* Has the ring wrapped? Process any descriptors that do not have 1439 /* Has the ring wrapped? Process any descriptors that do not have
1440 * the same "wrap" indicator as the current completion indicator 1440 * the same "wrap" indicator as the current completion indicator
1441 */ 1441 */
1442 spin_lock_irqsave(&etdev->TCBSendQLock, lockflags); 1442 spin_lock_irqsave(&etdev->TCBSendQLock, flags);
1443 1443
1444 pMpTcb = etdev->TxRing.CurrSendHead; 1444 pMpTcb = etdev->TxRing.CurrSendHead;
1445 while (pMpTcb && 1445 while (pMpTcb &&
@@ -1450,9 +1450,9 @@ static void et131x_update_tcb_list(struct et131x_adapter *etdev)
1450 if (pMpTcb->Next == NULL) 1450 if (pMpTcb->Next == NULL)
1451 etdev->TxRing.CurrSendTail = NULL; 1451 etdev->TxRing.CurrSendTail = NULL;
1452 1452
1453 spin_unlock_irqrestore(&etdev->TCBSendQLock, lockflags); 1453 spin_unlock_irqrestore(&etdev->TCBSendQLock, flags);
1454 MP_FREE_SEND_PACKET_FUN(etdev, pMpTcb); 1454 MP_FREE_SEND_PACKET_FUN(etdev, pMpTcb);
1455 spin_lock_irqsave(&etdev->TCBSendQLock, lockflags); 1455 spin_lock_irqsave(&etdev->TCBSendQLock, flags);
1456 1456
1457 /* Goto the next packet */ 1457 /* Goto the next packet */
1458 pMpTcb = etdev->TxRing.CurrSendHead; 1458 pMpTcb = etdev->TxRing.CurrSendHead;
@@ -1465,9 +1465,9 @@ static void et131x_update_tcb_list(struct et131x_adapter *etdev)
1465 if (pMpTcb->Next == NULL) 1465 if (pMpTcb->Next == NULL)
1466 etdev->TxRing.CurrSendTail = NULL; 1466 etdev->TxRing.CurrSendTail = NULL;
1467 1467
1468 spin_unlock_irqrestore(&etdev->TCBSendQLock, lockflags); 1468 spin_unlock_irqrestore(&etdev->TCBSendQLock, flags);
1469 MP_FREE_SEND_PACKET_FUN(etdev, pMpTcb); 1469 MP_FREE_SEND_PACKET_FUN(etdev, pMpTcb);
1470 spin_lock_irqsave(&etdev->TCBSendQLock, lockflags); 1470 spin_lock_irqsave(&etdev->TCBSendQLock, flags);
1471 1471
1472 /* Goto the next packet */ 1472 /* Goto the next packet */
1473 pMpTcb = etdev->TxRing.CurrSendHead; 1473 pMpTcb = etdev->TxRing.CurrSendHead;
@@ -1477,7 +1477,7 @@ static void et131x_update_tcb_list(struct et131x_adapter *etdev)
1477 if (etdev->TxRing.nBusySend <= (NUM_TCB / 3)) 1477 if (etdev->TxRing.nBusySend <= (NUM_TCB / 3))
1478 netif_wake_queue(etdev->netdev); 1478 netif_wake_queue(etdev->netdev);
1479 1479
1480 spin_unlock_irqrestore(&etdev->TCBSendQLock, lockflags); 1480 spin_unlock_irqrestore(&etdev->TCBSendQLock, flags);
1481} 1481}
1482 1482
1483/** 1483/**
@@ -1489,9 +1489,9 @@ static void et131x_update_tcb_list(struct et131x_adapter *etdev)
1489 */ 1489 */
1490static void et131x_check_send_wait_list(struct et131x_adapter *etdev) 1490static void et131x_check_send_wait_list(struct et131x_adapter *etdev)
1491{ 1491{
1492 unsigned long lockflags; 1492 unsigned long flags;
1493 1493
1494 spin_lock_irqsave(&etdev->SendWaitLock, lockflags); 1494 spin_lock_irqsave(&etdev->SendWaitLock, flags);
1495 1495
1496 while (!list_empty(&etdev->TxRing.SendWaitQueue) && 1496 while (!list_empty(&etdev->TxRing.SendWaitQueue) &&
1497 MP_TCB_RESOURCES_AVAILABLE(etdev)) { 1497 MP_TCB_RESOURCES_AVAILABLE(etdev)) {
@@ -1508,5 +1508,5 @@ static void et131x_check_send_wait_list(struct et131x_adapter *etdev)
1508 etdev->TxRing.nWaitSend); 1508 etdev->TxRing.nWaitSend);
1509 } 1509 }
1510 1510
1511 spin_unlock_irqrestore(&etdev->SendWaitLock, lockflags); 1511 spin_unlock_irqrestore(&etdev->SendWaitLock, flags);
1512} 1512}
diff --git a/drivers/staging/et131x/et131x_initpci.c b/drivers/staging/et131x/et131x_initpci.c
index 8455e0a8471a..b9018e6f5660 100644
--- a/drivers/staging/et131x/et131x_initpci.c
+++ b/drivers/staging/et131x/et131x_initpci.c
@@ -492,18 +492,18 @@ void et131x_error_timer_handler(unsigned long data)
492void et131x_link_detection_handler(unsigned long data) 492void et131x_link_detection_handler(unsigned long data)
493{ 493{
494 struct et131x_adapter *etdev = (struct et131x_adapter *) data; 494 struct et131x_adapter *etdev = (struct et131x_adapter *) data;
495 unsigned long lockflags; 495 unsigned long flags;
496 496
497 /* Let everyone know that we have run */ 497 /* Let everyone know that we have run */
498 etdev->bLinkTimerActive = false; 498 etdev->bLinkTimerActive = false;
499 499
500 if (etdev->MediaState == 0) { 500 if (etdev->MediaState == 0) {
501 spin_lock_irqsave(&etdev->Lock, lockflags); 501 spin_lock_irqsave(&etdev->Lock, flags);
502 502
503 etdev->MediaState = NETIF_STATUS_MEDIA_DISCONNECT; 503 etdev->MediaState = NETIF_STATUS_MEDIA_DISCONNECT;
504 MP_CLEAR_FLAG(etdev, fMP_ADAPTER_LINK_DETECTION); 504 MP_CLEAR_FLAG(etdev, fMP_ADAPTER_LINK_DETECTION);
505 505
506 spin_unlock_irqrestore(&etdev->Lock, lockflags); 506 spin_unlock_irqrestore(&etdev->Lock, flags);
507 507
508 netif_carrier_off(etdev->netdev); 508 netif_carrier_off(etdev->netdev);
509 509
diff --git a/drivers/staging/et131x/et131x_netdev.c b/drivers/staging/et131x/et131x_netdev.c
index 77fac9592cdb..74ba1774206b 100644
--- a/drivers/staging/et131x/et131x_netdev.c
+++ b/drivers/staging/et131x/et131x_netdev.c
@@ -467,12 +467,12 @@ void et131x_multicast(struct net_device *netdev)
467 struct et131x_adapter *adapter = netdev_priv(netdev); 467 struct et131x_adapter *adapter = netdev_priv(netdev);
468 uint32_t PacketFilter = 0; 468 uint32_t PacketFilter = 0;
469 uint32_t count; 469 uint32_t count;
470 unsigned long lockflags; 470 unsigned long flags;
471 struct dev_mc_list *mclist = netdev->mc_list; 471 struct dev_mc_list *mclist = netdev->mc_list;
472 472
473 DBG_ENTER(et131x_dbginfo); 473 DBG_ENTER(et131x_dbginfo);
474 474
475 spin_lock_irqsave(&adapter->Lock, lockflags); 475 spin_lock_irqsave(&adapter->Lock, flags);
476 476
477 /* Before we modify the platform-independent filter flags, store them 477 /* Before we modify the platform-independent filter flags, store them
478 * locally. This allows us to determine if anything's changed and if 478 * locally. This allows us to determine if anything's changed and if
@@ -552,7 +552,7 @@ void et131x_multicast(struct net_device *netdev)
552 "NO UPDATE REQUIRED, FLAGS didn't change\n"); 552 "NO UPDATE REQUIRED, FLAGS didn't change\n");
553 } 553 }
554 554
555 spin_unlock_irqrestore(&adapter->Lock, lockflags); 555 spin_unlock_irqrestore(&adapter->Lock, flags);
556 556
557 DBG_LEAVE(et131x_dbginfo); 557 DBG_LEAVE(et131x_dbginfo);
558} 558}
@@ -610,7 +610,7 @@ void et131x_tx_timeout(struct net_device *netdev)
610{ 610{
611 struct et131x_adapter *etdev = netdev_priv(netdev); 611 struct et131x_adapter *etdev = netdev_priv(netdev);
612 PMP_TCB pMpTcb; 612 PMP_TCB pMpTcb;
613 unsigned long lockflags; 613 unsigned long flags;
614 614
615 DBG_WARNING(et131x_dbginfo, "TX TIMEOUT\n"); 615 DBG_WARNING(et131x_dbginfo, "TX TIMEOUT\n");
616 616
@@ -635,7 +635,7 @@ void et131x_tx_timeout(struct net_device *netdev)
635 } 635 }
636 636
637 /* Is send stuck? */ 637 /* Is send stuck? */
638 spin_lock_irqsave(&etdev->TCBSendQLock, lockflags); 638 spin_lock_irqsave(&etdev->TCBSendQLock, flags);
639 639
640 pMpTcb = etdev->TxRing.CurrSendHead; 640 pMpTcb = etdev->TxRing.CurrSendHead;
641 641
@@ -660,7 +660,7 @@ void et131x_tx_timeout(struct net_device *netdev)
660 } 660 }
661 661
662 spin_unlock_irqrestore(&etdev->TCBSendQLock, 662 spin_unlock_irqrestore(&etdev->TCBSendQLock,
663 lockflags); 663 flags);
664 664
665 DBG_WARNING(et131x_dbginfo, 665 DBG_WARNING(et131x_dbginfo,
666 "Send stuck - reset. pMpTcb->WrIndex %x, Flags 0x%08x\n", 666 "Send stuck - reset. pMpTcb->WrIndex %x, Flags 0x%08x\n",
@@ -689,7 +689,7 @@ void et131x_tx_timeout(struct net_device *netdev)
689 } 689 }
690 } 690 }
691 691
692 spin_unlock_irqrestore(&etdev->TCBSendQLock, lockflags); 692 spin_unlock_irqrestore(&etdev->TCBSendQLock, flags);
693} 693}
694 694
695/** 695/**