diff options
author | Alan Cox <alan@linux.intel.com> | 2009-08-19 13:21:50 -0400 |
---|---|---|
committer | Greg Kroah-Hartman <gregkh@suse.de> | 2009-09-15 15:02:26 -0400 |
commit | 37628606661a8a1d3abfa5bb898426a38fa62b73 (patch) | |
tree | 40d3ba821c49ea87ef319c5f07eadb536ebc2da9 /drivers | |
parent | 25ad00bba42d49e37e81e7c449bbcc86a71f2894 (diff) |
Staging: et131x: spinlocks
Switch to the more normal "flags" naming. Also fix up the nested use of
spin_lock_irqsave
Signed-off-by: Alan Cox <alan@linux.intel.com>
Signed-off-by: Greg Kroah-Hartman <gregkh@suse.de>
Diffstat (limited to 'drivers')
-rw-r--r-- | drivers/staging/et131x/et1310_phy.c | 10 | ||||
-rw-r--r-- | drivers/staging/et131x/et1310_pm.c | 6 | ||||
-rw-r--r-- | drivers/staging/et131x/et1310_rx.c | 28 | ||||
-rw-r--r-- | drivers/staging/et131x/et1310_tx.c | 74 | ||||
-rw-r--r-- | drivers/staging/et131x/et131x_initpci.c | 6 | ||||
-rw-r--r-- | drivers/staging/et131x/et131x_netdev.c | 14 |
6 files changed, 69 insertions, 69 deletions
diff --git a/drivers/staging/et131x/et1310_phy.c b/drivers/staging/et131x/et1310_phy.c index 82fcb6678e24..d9c7a4486b41 100644 --- a/drivers/staging/et131x/et1310_phy.c +++ b/drivers/staging/et131x/et1310_phy.c | |||
@@ -484,7 +484,7 @@ void et131x_Mii_check(struct et131x_adapter *etdev, | |||
484 | uint32_t uiMdiMdix; | 484 | uint32_t uiMdiMdix; |
485 | uint32_t uiMasterSlave; | 485 | uint32_t uiMasterSlave; |
486 | uint32_t uiPolarity; | 486 | uint32_t uiPolarity; |
487 | unsigned long lockflags; | 487 | unsigned long flags; |
488 | 488 | ||
489 | DBG_ENTER(et131x_dbginfo); | 489 | DBG_ENTER(et131x_dbginfo); |
490 | 490 | ||
@@ -495,12 +495,12 @@ void et131x_Mii_check(struct et131x_adapter *etdev, | |||
495 | /* Update our state variables and indicate the | 495 | /* Update our state variables and indicate the |
496 | * connected state | 496 | * connected state |
497 | */ | 497 | */ |
498 | spin_lock_irqsave(&etdev->Lock, lockflags); | 498 | spin_lock_irqsave(&etdev->Lock, flags); |
499 | 499 | ||
500 | etdev->MediaState = NETIF_STATUS_MEDIA_CONNECT; | 500 | etdev->MediaState = NETIF_STATUS_MEDIA_CONNECT; |
501 | MP_CLEAR_FLAG(etdev, fMP_ADAPTER_LINK_DETECTION); | 501 | MP_CLEAR_FLAG(etdev, fMP_ADAPTER_LINK_DETECTION); |
502 | 502 | ||
503 | spin_unlock_irqrestore(&etdev->Lock, lockflags); | 503 | spin_unlock_irqrestore(&etdev->Lock, flags); |
504 | 504 | ||
505 | /* Don't indicate state if we're in loopback mode */ | 505 | /* Don't indicate state if we're in loopback mode */ |
506 | if (etdev->RegistryPhyLoopbk == false) | 506 | if (etdev->RegistryPhyLoopbk == false) |
@@ -533,11 +533,11 @@ void et131x_Mii_check(struct et131x_adapter *etdev, | |||
533 | (etdev, fMP_ADAPTER_LINK_DETECTION)) | 533 | (etdev, fMP_ADAPTER_LINK_DETECTION)) |
534 | || (etdev->MediaState == | 534 | || (etdev->MediaState == |
535 | NETIF_STATUS_MEDIA_DISCONNECT)) { | 535 | NETIF_STATUS_MEDIA_DISCONNECT)) { |
536 | spin_lock_irqsave(&etdev->Lock, lockflags); | 536 | spin_lock_irqsave(&etdev->Lock, flags); |
537 | etdev->MediaState = | 537 | etdev->MediaState = |
538 | NETIF_STATUS_MEDIA_DISCONNECT; | 538 | NETIF_STATUS_MEDIA_DISCONNECT; |
539 | spin_unlock_irqrestore(&etdev->Lock, | 539 | spin_unlock_irqrestore(&etdev->Lock, |
540 | lockflags); | 540 | flags); |
541 | 541 | ||
542 | /* Only indicate state if we're in loopback | 542 | /* Only indicate state if we're in loopback |
543 | * mode | 543 | * mode |
diff --git a/drivers/staging/et131x/et1310_pm.c b/drivers/staging/et131x/et1310_pm.c index 864d3ad3fcdd..a50741168f62 100644 --- a/drivers/staging/et131x/et1310_pm.c +++ b/drivers/staging/et131x/et1310_pm.c | |||
@@ -119,7 +119,7 @@ extern dbg_info_t *et131x_dbginfo; | |||
119 | */ | 119 | */ |
120 | void EnablePhyComa(struct et131x_adapter *etdev) | 120 | void EnablePhyComa(struct et131x_adapter *etdev) |
121 | { | 121 | { |
122 | unsigned long lockflags; | 122 | unsigned long flags; |
123 | PM_CSR_t GlobalPmCSR; | 123 | PM_CSR_t GlobalPmCSR; |
124 | int32_t LoopCounter = 10; | 124 | int32_t LoopCounter = 10; |
125 | 125 | ||
@@ -134,9 +134,9 @@ void EnablePhyComa(struct et131x_adapter *etdev) | |||
134 | etdev->PoMgmt.PowerDownDuplex = etdev->AiForceDpx; | 134 | etdev->PoMgmt.PowerDownDuplex = etdev->AiForceDpx; |
135 | 135 | ||
136 | /* Stop sending packets. */ | 136 | /* Stop sending packets. */ |
137 | spin_lock_irqsave(&etdev->SendHWLock, lockflags); | 137 | spin_lock_irqsave(&etdev->SendHWLock, flags); |
138 | MP_SET_FLAG(etdev, fMP_ADAPTER_LOWER_POWER); | 138 | MP_SET_FLAG(etdev, fMP_ADAPTER_LOWER_POWER); |
139 | spin_unlock_irqrestore(&etdev->SendHWLock, lockflags); | 139 | spin_unlock_irqrestore(&etdev->SendHWLock, flags); |
140 | 140 | ||
141 | /* Wait for outstanding Receive packets */ | 141 | /* Wait for outstanding Receive packets */ |
142 | while ((MP_GET_RCV_REF(etdev) != 0) && (LoopCounter-- > 0)) | 142 | while ((MP_GET_RCV_REF(etdev) != 0) && (LoopCounter-- > 0)) |
diff --git a/drivers/staging/et131x/et1310_rx.c b/drivers/staging/et131x/et1310_rx.c index f8cc6a607d2e..9dc08fb4bf27 100644 --- a/drivers/staging/et131x/et1310_rx.c +++ b/drivers/staging/et131x/et1310_rx.c | |||
@@ -685,7 +685,7 @@ void ConfigRxDmaRegs(struct et131x_adapter *etdev) | |||
685 | PFBR_DESC_t pFbrEntry; | 685 | PFBR_DESC_t pFbrEntry; |
686 | uint32_t iEntry; | 686 | uint32_t iEntry; |
687 | RXDMA_PSR_NUM_DES_t psr_num_des; | 687 | RXDMA_PSR_NUM_DES_t psr_num_des; |
688 | unsigned long lockflags; | 688 | unsigned long flags; |
689 | 689 | ||
690 | DBG_ENTER(et131x_dbginfo); | 690 | DBG_ENTER(et131x_dbginfo); |
691 | 691 | ||
@@ -718,7 +718,7 @@ void ConfigRxDmaRegs(struct et131x_adapter *etdev) | |||
718 | writel((psr_num_des.bits.psr_ndes * LO_MARK_PERCENT_FOR_PSR) / 100, | 718 | writel((psr_num_des.bits.psr_ndes * LO_MARK_PERCENT_FOR_PSR) / 100, |
719 | &pRxDma->psr_min_des.value); | 719 | &pRxDma->psr_min_des.value); |
720 | 720 | ||
721 | spin_lock_irqsave(&etdev->RcvLock, lockflags); | 721 | spin_lock_irqsave(&etdev->RcvLock, flags); |
722 | 722 | ||
723 | /* These local variables track the PSR in the adapter structure */ | 723 | /* These local variables track the PSR in the adapter structure */ |
724 | pRxLocal->local_psr_full.bits.psr_full = 0; | 724 | pRxLocal->local_psr_full.bits.psr_full = 0; |
@@ -801,7 +801,7 @@ void ConfigRxDmaRegs(struct et131x_adapter *etdev) | |||
801 | */ | 801 | */ |
802 | writel(etdev->RegistryRxTimeInterval, &pRxDma->max_pkt_time.value); | 802 | writel(etdev->RegistryRxTimeInterval, &pRxDma->max_pkt_time.value); |
803 | 803 | ||
804 | spin_unlock_irqrestore(&etdev->RcvLock, lockflags); | 804 | spin_unlock_irqrestore(&etdev->RcvLock, flags); |
805 | 805 | ||
806 | DBG_LEAVE(et131x_dbginfo); | 806 | DBG_LEAVE(et131x_dbginfo); |
807 | } | 807 | } |
@@ -914,7 +914,7 @@ PMP_RFD nic_rx_pkts(struct et131x_adapter *etdev) | |||
914 | PMP_RFD pMpRfd; | 914 | PMP_RFD pMpRfd; |
915 | uint32_t nIndex; | 915 | uint32_t nIndex; |
916 | uint8_t *pBufVa; | 916 | uint8_t *pBufVa; |
917 | unsigned long lockflags; | 917 | unsigned long flags; |
918 | struct list_head *element; | 918 | struct list_head *element; |
919 | uint8_t ringIndex; | 919 | uint8_t ringIndex; |
920 | uint16_t bufferIndex; | 920 | uint16_t bufferIndex; |
@@ -1013,7 +1013,7 @@ PMP_RFD nic_rx_pkts(struct et131x_adapter *etdev) | |||
1013 | } | 1013 | } |
1014 | 1014 | ||
1015 | /* Get and fill the RFD. */ | 1015 | /* Get and fill the RFD. */ |
1016 | spin_lock_irqsave(&etdev->RcvLock, lockflags); | 1016 | spin_lock_irqsave(&etdev->RcvLock, flags); |
1017 | 1017 | ||
1018 | pMpRfd = NULL; | 1018 | pMpRfd = NULL; |
1019 | element = pRxLocal->RecvList.next; | 1019 | element = pRxLocal->RecvList.next; |
@@ -1023,14 +1023,14 @@ PMP_RFD nic_rx_pkts(struct et131x_adapter *etdev) | |||
1023 | DBG_RX(et131x_dbginfo, | 1023 | DBG_RX(et131x_dbginfo, |
1024 | "NULL RFD returned from RecvList via list_entry()\n"); | 1024 | "NULL RFD returned from RecvList via list_entry()\n"); |
1025 | DBG_RX_LEAVE(et131x_dbginfo); | 1025 | DBG_RX_LEAVE(et131x_dbginfo); |
1026 | spin_unlock_irqrestore(&etdev->RcvLock, lockflags); | 1026 | spin_unlock_irqrestore(&etdev->RcvLock, flags); |
1027 | return NULL; | 1027 | return NULL; |
1028 | } | 1028 | } |
1029 | 1029 | ||
1030 | list_del(&pMpRfd->list_node); | 1030 | list_del(&pMpRfd->list_node); |
1031 | pRxLocal->nReadyRecv--; | 1031 | pRxLocal->nReadyRecv--; |
1032 | 1032 | ||
1033 | spin_unlock_irqrestore(&etdev->RcvLock, lockflags); | 1033 | spin_unlock_irqrestore(&etdev->RcvLock, flags); |
1034 | 1034 | ||
1035 | pMpRfd->iBufferIndex = bufferIndex; | 1035 | pMpRfd->iBufferIndex = bufferIndex; |
1036 | pMpRfd->iRingIndex = ringIndex; | 1036 | pMpRfd->iRingIndex = ringIndex; |
@@ -1260,9 +1260,9 @@ void et131x_handle_recv_interrupt(struct et131x_adapter *etdev) | |||
1260 | * Besides, we don't really need (at this point) the | 1260 | * Besides, we don't really need (at this point) the |
1261 | * pending list anyway. | 1261 | * pending list anyway. |
1262 | */ | 1262 | */ |
1263 | /* spin_lock_irqsave( &etdev->RcvPendLock, lockflags ); | 1263 | /* spin_lock_irqsave( &etdev->RcvPendLock, flags ); |
1264 | * list_add_tail( &pMpRfd->list_node, &etdev->RxRing.RecvPendingList ); | 1264 | * list_add_tail( &pMpRfd->list_node, &etdev->RxRing.RecvPendingList ); |
1265 | * spin_unlock_irqrestore( &etdev->RcvPendLock, lockflags ); | 1265 | * spin_unlock_irqrestore( &etdev->RcvPendLock, flags ); |
1266 | */ | 1266 | */ |
1267 | 1267 | ||
1268 | /* Update the number of outstanding Recvs */ | 1268 | /* Update the number of outstanding Recvs */ |
@@ -1302,7 +1302,7 @@ void nic_return_rfd(struct et131x_adapter *etdev, PMP_RFD pMpRfd) | |||
1302 | struct _RXDMA_t __iomem *pRxDma = &etdev->CSRAddress->rxdma; | 1302 | struct _RXDMA_t __iomem *pRxDma = &etdev->CSRAddress->rxdma; |
1303 | uint16_t bi = pMpRfd->iBufferIndex; | 1303 | uint16_t bi = pMpRfd->iBufferIndex; |
1304 | uint8_t ri = pMpRfd->iRingIndex; | 1304 | uint8_t ri = pMpRfd->iRingIndex; |
1305 | unsigned long lockflags; | 1305 | unsigned long flags; |
1306 | 1306 | ||
1307 | DBG_RX_ENTER(et131x_dbginfo); | 1307 | DBG_RX_ENTER(et131x_dbginfo); |
1308 | 1308 | ||
@@ -1314,7 +1314,7 @@ void nic_return_rfd(struct et131x_adapter *etdev, PMP_RFD pMpRfd) | |||
1314 | (ri == 0 && bi < pRxLocal->Fbr0NumEntries) || | 1314 | (ri == 0 && bi < pRxLocal->Fbr0NumEntries) || |
1315 | #endif | 1315 | #endif |
1316 | (ri == 1 && bi < pRxLocal->Fbr1NumEntries)) { | 1316 | (ri == 1 && bi < pRxLocal->Fbr1NumEntries)) { |
1317 | spin_lock_irqsave(&etdev->FbrLock, lockflags); | 1317 | spin_lock_irqsave(&etdev->FbrLock, flags); |
1318 | 1318 | ||
1319 | if (ri == 1) { | 1319 | if (ri == 1) { |
1320 | PFBR_DESC_t pNextDesc = | 1320 | PFBR_DESC_t pNextDesc = |
@@ -1362,7 +1362,7 @@ void nic_return_rfd(struct et131x_adapter *etdev, PMP_RFD pMpRfd) | |||
1362 | &pRxDma->fbr0_full_offset.value); | 1362 | &pRxDma->fbr0_full_offset.value); |
1363 | } | 1363 | } |
1364 | #endif | 1364 | #endif |
1365 | spin_unlock_irqrestore(&etdev->FbrLock, lockflags); | 1365 | spin_unlock_irqrestore(&etdev->FbrLock, flags); |
1366 | } else { | 1366 | } else { |
1367 | DBG_ERROR(et131x_dbginfo, | 1367 | DBG_ERROR(et131x_dbginfo, |
1368 | "NICReturnRFD illegal Buffer Index returned\n"); | 1368 | "NICReturnRFD illegal Buffer Index returned\n"); |
@@ -1371,10 +1371,10 @@ void nic_return_rfd(struct et131x_adapter *etdev, PMP_RFD pMpRfd) | |||
1371 | /* The processing on this RFD is done, so put it back on the tail of | 1371 | /* The processing on this RFD is done, so put it back on the tail of |
1372 | * our list | 1372 | * our list |
1373 | */ | 1373 | */ |
1374 | spin_lock_irqsave(&etdev->RcvLock, lockflags); | 1374 | spin_lock_irqsave(&etdev->RcvLock, flags); |
1375 | list_add_tail(&pMpRfd->list_node, &pRxLocal->RecvList); | 1375 | list_add_tail(&pMpRfd->list_node, &pRxLocal->RecvList); |
1376 | pRxLocal->nReadyRecv++; | 1376 | pRxLocal->nReadyRecv++; |
1377 | spin_unlock_irqrestore(&etdev->RcvLock, lockflags); | 1377 | spin_unlock_irqrestore(&etdev->RcvLock, flags); |
1378 | 1378 | ||
1379 | DBG_ASSERT(pRxLocal->nReadyRecv <= pRxLocal->NumRfd); | 1379 | DBG_ASSERT(pRxLocal->nReadyRecv <= pRxLocal->NumRfd); |
1380 | DBG_RX_LEAVE(et131x_dbginfo); | 1380 | DBG_RX_LEAVE(et131x_dbginfo); |
diff --git a/drivers/staging/et131x/et1310_tx.c b/drivers/staging/et131x/et1310_tx.c index db0f538a335a..a017d271e375 100644 --- a/drivers/staging/et131x/et1310_tx.c +++ b/drivers/staging/et131x/et1310_tx.c | |||
@@ -461,7 +461,7 @@ static int et131x_send_packet(struct sk_buff *skb, | |||
461 | int status = 0; | 461 | int status = 0; |
462 | PMP_TCB pMpTcb = NULL; | 462 | PMP_TCB pMpTcb = NULL; |
463 | uint16_t *pShBufVa; | 463 | uint16_t *pShBufVa; |
464 | unsigned long lockflags; | 464 | unsigned long flags; |
465 | 465 | ||
466 | DBG_TX_ENTER(et131x_dbginfo); | 466 | DBG_TX_ENTER(et131x_dbginfo); |
467 | 467 | ||
@@ -482,12 +482,12 @@ static int et131x_send_packet(struct sk_buff *skb, | |||
482 | } | 482 | } |
483 | 483 | ||
484 | /* Get a TCB for this packet */ | 484 | /* Get a TCB for this packet */ |
485 | spin_lock_irqsave(&etdev->TCBReadyQLock, lockflags); | 485 | spin_lock_irqsave(&etdev->TCBReadyQLock, flags); |
486 | 486 | ||
487 | pMpTcb = etdev->TxRing.TCBReadyQueueHead; | 487 | pMpTcb = etdev->TxRing.TCBReadyQueueHead; |
488 | 488 | ||
489 | if (pMpTcb == NULL) { | 489 | if (pMpTcb == NULL) { |
490 | spin_unlock_irqrestore(&etdev->TCBReadyQLock, lockflags); | 490 | spin_unlock_irqrestore(&etdev->TCBReadyQLock, flags); |
491 | 491 | ||
492 | DBG_WARNING(et131x_dbginfo, "Can't obtain a TCB\n"); | 492 | DBG_WARNING(et131x_dbginfo, "Can't obtain a TCB\n"); |
493 | DBG_TX_LEAVE(et131x_dbginfo); | 493 | DBG_TX_LEAVE(et131x_dbginfo); |
@@ -499,7 +499,7 @@ static int et131x_send_packet(struct sk_buff *skb, | |||
499 | if (etdev->TxRing.TCBReadyQueueHead == NULL) | 499 | if (etdev->TxRing.TCBReadyQueueHead == NULL) |
500 | etdev->TxRing.TCBReadyQueueTail = NULL; | 500 | etdev->TxRing.TCBReadyQueueTail = NULL; |
501 | 501 | ||
502 | spin_unlock_irqrestore(&etdev->TCBReadyQLock, lockflags); | 502 | spin_unlock_irqrestore(&etdev->TCBReadyQLock, flags); |
503 | 503 | ||
504 | pMpTcb->PacketLength = skb->len; | 504 | pMpTcb->PacketLength = skb->len; |
505 | pMpTcb->Packet = skb; | 505 | pMpTcb->Packet = skb; |
@@ -522,7 +522,7 @@ static int et131x_send_packet(struct sk_buff *skb, | |||
522 | status = nic_send_packet(etdev, pMpTcb); | 522 | status = nic_send_packet(etdev, pMpTcb); |
523 | 523 | ||
524 | if (status != 0) { | 524 | if (status != 0) { |
525 | spin_lock_irqsave(&etdev->TCBReadyQLock, lockflags); | 525 | spin_lock_irqsave(&etdev->TCBReadyQLock, flags); |
526 | 526 | ||
527 | if (etdev->TxRing.TCBReadyQueueTail) { | 527 | if (etdev->TxRing.TCBReadyQueueTail) { |
528 | etdev->TxRing.TCBReadyQueueTail->Next = pMpTcb; | 528 | etdev->TxRing.TCBReadyQueueTail->Next = pMpTcb; |
@@ -533,7 +533,7 @@ static int et131x_send_packet(struct sk_buff *skb, | |||
533 | 533 | ||
534 | etdev->TxRing.TCBReadyQueueTail = pMpTcb; | 534 | etdev->TxRing.TCBReadyQueueTail = pMpTcb; |
535 | 535 | ||
536 | spin_unlock_irqrestore(&etdev->TCBReadyQLock, lockflags); | 536 | spin_unlock_irqrestore(&etdev->TCBReadyQLock, flags); |
537 | 537 | ||
538 | DBG_TX_LEAVE(et131x_dbginfo); | 538 | DBG_TX_LEAVE(et131x_dbginfo); |
539 | return status; | 539 | return status; |
@@ -561,7 +561,7 @@ static int nic_send_packet(struct et131x_adapter *etdev, PMP_TCB pMpTcb) | |||
561 | struct sk_buff *pPacket = pMpTcb->Packet; | 561 | struct sk_buff *pPacket = pMpTcb->Packet; |
562 | uint32_t FragListCount = skb_shinfo(pPacket)->nr_frags + 1; | 562 | uint32_t FragListCount = skb_shinfo(pPacket)->nr_frags + 1; |
563 | struct skb_frag_struct *pFragList = &skb_shinfo(pPacket)->frags[0]; | 563 | struct skb_frag_struct *pFragList = &skb_shinfo(pPacket)->frags[0]; |
564 | unsigned long lockflags1, lockflags2; | 564 | unsigned long flags; |
565 | 565 | ||
566 | DBG_TX_ENTER(et131x_dbginfo); | 566 | DBG_TX_ENTER(et131x_dbginfo); |
567 | 567 | ||
@@ -726,7 +726,7 @@ static int nic_send_packet(struct et131x_adapter *etdev, PMP_TCB pMpTcb) | |||
726 | pMpTcb->WrIndexStart = etdev->TxRing.txDmaReadyToSend; | 726 | pMpTcb->WrIndexStart = etdev->TxRing.txDmaReadyToSend; |
727 | pMpTcb->PacketStaleCount = 0; | 727 | pMpTcb->PacketStaleCount = 0; |
728 | 728 | ||
729 | spin_lock_irqsave(&etdev->SendHWLock, lockflags1); | 729 | spin_lock_irqsave(&etdev->SendHWLock, flags); |
730 | 730 | ||
731 | iThisCopy = | 731 | iThisCopy = |
732 | NUM_DESC_PER_RING_TX - etdev->TxRing.txDmaReadyToSend.bits.val; | 732 | NUM_DESC_PER_RING_TX - etdev->TxRing.txDmaReadyToSend.bits.val; |
@@ -771,7 +771,7 @@ static int nic_send_packet(struct et131x_adapter *etdev, PMP_TCB pMpTcb) | |||
771 | pMpTcb->WrIndex.value = | 771 | pMpTcb->WrIndex.value = |
772 | etdev->TxRing.txDmaReadyToSend.value - 1; | 772 | etdev->TxRing.txDmaReadyToSend.value - 1; |
773 | 773 | ||
774 | spin_lock_irqsave(&etdev->TCBSendQLock, lockflags2); | 774 | spin_lock(&etdev->TCBSendQLock); |
775 | 775 | ||
776 | if (etdev->TxRing.CurrSendTail) | 776 | if (etdev->TxRing.CurrSendTail) |
777 | etdev->TxRing.CurrSendTail->Next = pMpTcb; | 777 | etdev->TxRing.CurrSendTail->Next = pMpTcb; |
@@ -784,7 +784,7 @@ static int nic_send_packet(struct et131x_adapter *etdev, PMP_TCB pMpTcb) | |||
784 | 784 | ||
785 | etdev->TxRing.nBusySend++; | 785 | etdev->TxRing.nBusySend++; |
786 | 786 | ||
787 | spin_unlock_irqrestore(&etdev->TCBSendQLock, lockflags2); | 787 | spin_unlock(&etdev->TCBSendQLock); |
788 | 788 | ||
789 | /* Write the new write pointer back to the device. */ | 789 | /* Write the new write pointer back to the device. */ |
790 | writel(etdev->TxRing.txDmaReadyToSend.value, | 790 | writel(etdev->TxRing.txDmaReadyToSend.value, |
@@ -798,7 +798,7 @@ static int nic_send_packet(struct et131x_adapter *etdev, PMP_TCB pMpTcb) | |||
798 | &etdev->CSRAddress->global.watchdog_timer); | 798 | &etdev->CSRAddress->global.watchdog_timer); |
799 | } | 799 | } |
800 | 800 | ||
801 | spin_unlock_irqrestore(&etdev->SendHWLock, lockflags1); | 801 | spin_unlock_irqrestore(&etdev->SendHWLock, flags); |
802 | 802 | ||
803 | DBG_TX_LEAVE(et131x_dbginfo); | 803 | DBG_TX_LEAVE(et131x_dbginfo); |
804 | return 0; | 804 | return 0; |
@@ -829,7 +829,7 @@ static int nic_send_packet(struct et131x_adapter *etdev, PMP_TCB pMpTcb) | |||
829 | TX_DESC_ENTRY_t *CurDescPostCopy = NULL; | 829 | TX_DESC_ENTRY_t *CurDescPostCopy = NULL; |
830 | uint32_t SlotsAvailable; | 830 | uint32_t SlotsAvailable; |
831 | DMA10W_t ServiceComplete; | 831 | DMA10W_t ServiceComplete; |
832 | unsigned int lockflags1, lockflags2; | 832 | unsigned int flags; |
833 | struct sk_buff *pPacket = pMpTcb->Packet; | 833 | struct sk_buff *pPacket = pMpTcb->Packet; |
834 | uint32_t FragListCount = skb_shinfo(pPacket)->nr_frags + 1; | 834 | uint32_t FragListCount = skb_shinfo(pPacket)->nr_frags + 1; |
835 | struct skb_frag_struct *pFragList = &skb_shinfo(pPacket)->frags[0]; | 835 | struct skb_frag_struct *pFragList = &skb_shinfo(pPacket)->frags[0]; |
@@ -875,7 +875,7 @@ static int nic_send_packet(struct et131x_adapter *etdev, PMP_TCB pMpTcb) | |||
875 | SegmentSize = (pPacket->len - pPacket->data_len) / 2; | 875 | SegmentSize = (pPacket->len - pPacket->data_len) / 2; |
876 | } | 876 | } |
877 | 877 | ||
878 | spin_lock_irqsave(&etdev->SendHWLock, lockflags1); | 878 | spin_lock_irqsave(&etdev->SendHWLock, flags); |
879 | 879 | ||
880 | if (etdev->TxRing.txDmaReadyToSend.bits.serv_req_wrap == | 880 | if (etdev->TxRing.txDmaReadyToSend.bits.serv_req_wrap == |
881 | ServiceComplete.bits.serv_cpl_wrap) { | 881 | ServiceComplete.bits.serv_cpl_wrap) { |
@@ -896,7 +896,7 @@ static int nic_send_packet(struct et131x_adapter *etdev, PMP_TCB pMpTcb) | |||
896 | if ((FragListCount + iSplitFirstElement) > SlotsAvailable) { | 896 | if ((FragListCount + iSplitFirstElement) > SlotsAvailable) { |
897 | DBG_WARNING(et131x_dbginfo, | 897 | DBG_WARNING(et131x_dbginfo, |
898 | "Not Enough Space in Tx Desc Ring\n"); | 898 | "Not Enough Space in Tx Desc Ring\n"); |
899 | spin_unlock_irqrestore(&etdev->SendHWLock, lockflags1); | 899 | spin_unlock_irqrestore(&etdev->SendHWLock, flags); |
900 | return -ENOMEM; | 900 | return -ENOMEM; |
901 | } | 901 | } |
902 | 902 | ||
@@ -1185,7 +1185,7 @@ static int nic_send_packet(struct et131x_adapter *etdev, PMP_TCB pMpTcb) | |||
1185 | NIC_MIN_PACKET_SIZE - pMpTcb->PacketLength); | 1185 | NIC_MIN_PACKET_SIZE - pMpTcb->PacketLength); |
1186 | } | 1186 | } |
1187 | 1187 | ||
1188 | spin_lock_irqsave(&etdev->TCBSendQLock, lockflags2); | 1188 | spin_lock(&etdev->TCBSendQLock); |
1189 | 1189 | ||
1190 | if (etdev->TxRing.CurrSendTail) | 1190 | if (etdev->TxRing.CurrSendTail) |
1191 | etdev->TxRing.CurrSendTail->Next = pMpTcb; | 1191 | etdev->TxRing.CurrSendTail->Next = pMpTcb; |
@@ -1198,7 +1198,7 @@ static int nic_send_packet(struct et131x_adapter *etdev, PMP_TCB pMpTcb) | |||
1198 | 1198 | ||
1199 | etdev->TxRing.nBusySend++; | 1199 | etdev->TxRing.nBusySend++; |
1200 | 1200 | ||
1201 | spin_unlock_irqrestore(&etdev->TCBSendQLock, lockflags2); | 1201 | spin_unlock(&etdev->TCBSendQLock); |
1202 | 1202 | ||
1203 | /* Write the new write pointer back to the device. */ | 1203 | /* Write the new write pointer back to the device. */ |
1204 | writel(etdev->TxRing.txDmaReadyToSend.value, | 1204 | writel(etdev->TxRing.txDmaReadyToSend.value, |
@@ -1216,7 +1216,7 @@ static int nic_send_packet(struct et131x_adapter *etdev, PMP_TCB pMpTcb) | |||
1216 | &etdev->CSRAddress->global.watchdog_timer); | 1216 | &etdev->CSRAddress->global.watchdog_timer); |
1217 | } | 1217 | } |
1218 | 1218 | ||
1219 | spin_unlock_irqrestore(&etdev->SendHWLock, lockflags1); | 1219 | spin_unlock_irqrestore(&etdev->SendHWLock, flags); |
1220 | 1220 | ||
1221 | DBG_TX_LEAVE(et131x_dbginfo); | 1221 | DBG_TX_LEAVE(et131x_dbginfo); |
1222 | return 0; | 1222 | return 0; |
@@ -1234,7 +1234,7 @@ static int nic_send_packet(struct et131x_adapter *etdev, PMP_TCB pMpTcb) | |||
1234 | inline void et131x_free_send_packet(struct et131x_adapter *etdev, | 1234 | inline void et131x_free_send_packet(struct et131x_adapter *etdev, |
1235 | PMP_TCB pMpTcb) | 1235 | PMP_TCB pMpTcb) |
1236 | { | 1236 | { |
1237 | unsigned long lockflags; | 1237 | unsigned long flags; |
1238 | TX_DESC_ENTRY_t *desc = NULL; | 1238 | TX_DESC_ENTRY_t *desc = NULL; |
1239 | struct net_device_stats *stats = &etdev->net_stats; | 1239 | struct net_device_stats *stats = &etdev->net_stats; |
1240 | 1240 | ||
@@ -1311,7 +1311,7 @@ inline void et131x_free_send_packet(struct et131x_adapter *etdev, | |||
1311 | memset(pMpTcb, 0, sizeof(MP_TCB)); | 1311 | memset(pMpTcb, 0, sizeof(MP_TCB)); |
1312 | 1312 | ||
1313 | /* Add the TCB to the Ready Q */ | 1313 | /* Add the TCB to the Ready Q */ |
1314 | spin_lock_irqsave(&etdev->TCBReadyQLock, lockflags); | 1314 | spin_lock_irqsave(&etdev->TCBReadyQLock, flags); |
1315 | 1315 | ||
1316 | etdev->Stats.opackets++; | 1316 | etdev->Stats.opackets++; |
1317 | 1317 | ||
@@ -1324,7 +1324,7 @@ inline void et131x_free_send_packet(struct et131x_adapter *etdev, | |||
1324 | 1324 | ||
1325 | etdev->TxRing.TCBReadyQueueTail = pMpTcb; | 1325 | etdev->TxRing.TCBReadyQueueTail = pMpTcb; |
1326 | 1326 | ||
1327 | spin_unlock_irqrestore(&etdev->TCBReadyQLock, lockflags); | 1327 | spin_unlock_irqrestore(&etdev->TCBReadyQLock, flags); |
1328 | 1328 | ||
1329 | DBG_ASSERT(etdev->TxRing.nBusySend >= 0); | 1329 | DBG_ASSERT(etdev->TxRing.nBusySend >= 0); |
1330 | } | 1330 | } |
@@ -1339,16 +1339,16 @@ void et131x_free_busy_send_packets(struct et131x_adapter *etdev) | |||
1339 | { | 1339 | { |
1340 | PMP_TCB pMpTcb; | 1340 | PMP_TCB pMpTcb; |
1341 | struct list_head *pEntry; | 1341 | struct list_head *pEntry; |
1342 | unsigned long lockflags; | 1342 | unsigned long flags; |
1343 | uint32_t FreeCounter = 0; | 1343 | uint32_t FreeCounter = 0; |
1344 | 1344 | ||
1345 | DBG_ENTER(et131x_dbginfo); | 1345 | DBG_ENTER(et131x_dbginfo); |
1346 | 1346 | ||
1347 | while (!list_empty(&etdev->TxRing.SendWaitQueue)) { | 1347 | while (!list_empty(&etdev->TxRing.SendWaitQueue)) { |
1348 | spin_lock_irqsave(&etdev->SendWaitLock, lockflags); | 1348 | spin_lock_irqsave(&etdev->SendWaitLock, flags); |
1349 | 1349 | ||
1350 | etdev->TxRing.nWaitSend--; | 1350 | etdev->TxRing.nWaitSend--; |
1351 | spin_unlock_irqrestore(&etdev->SendWaitLock, lockflags); | 1351 | spin_unlock_irqrestore(&etdev->SendWaitLock, flags); |
1352 | 1352 | ||
1353 | pEntry = etdev->TxRing.SendWaitQueue.next; | 1353 | pEntry = etdev->TxRing.SendWaitQueue.next; |
1354 | } | 1354 | } |
@@ -1356,7 +1356,7 @@ void et131x_free_busy_send_packets(struct et131x_adapter *etdev) | |||
1356 | etdev->TxRing.nWaitSend = 0; | 1356 | etdev->TxRing.nWaitSend = 0; |
1357 | 1357 | ||
1358 | /* Any packets being sent? Check the first TCB on the send list */ | 1358 | /* Any packets being sent? Check the first TCB on the send list */ |
1359 | spin_lock_irqsave(&etdev->TCBSendQLock, lockflags); | 1359 | spin_lock_irqsave(&etdev->TCBSendQLock, flags); |
1360 | 1360 | ||
1361 | pMpTcb = etdev->TxRing.CurrSendHead; | 1361 | pMpTcb = etdev->TxRing.CurrSendHead; |
1362 | 1362 | ||
@@ -1370,14 +1370,14 @@ void et131x_free_busy_send_packets(struct et131x_adapter *etdev) | |||
1370 | 1370 | ||
1371 | etdev->TxRing.nBusySend--; | 1371 | etdev->TxRing.nBusySend--; |
1372 | 1372 | ||
1373 | spin_unlock_irqrestore(&etdev->TCBSendQLock, lockflags); | 1373 | spin_unlock_irqrestore(&etdev->TCBSendQLock, flags); |
1374 | 1374 | ||
1375 | DBG_VERBOSE(et131x_dbginfo, "pMpTcb = 0x%p\n", pMpTcb); | 1375 | DBG_VERBOSE(et131x_dbginfo, "pMpTcb = 0x%p\n", pMpTcb); |
1376 | 1376 | ||
1377 | FreeCounter++; | 1377 | FreeCounter++; |
1378 | MP_FREE_SEND_PACKET_FUN(etdev, pMpTcb); | 1378 | MP_FREE_SEND_PACKET_FUN(etdev, pMpTcb); |
1379 | 1379 | ||
1380 | spin_lock_irqsave(&etdev->TCBSendQLock, lockflags); | 1380 | spin_lock_irqsave(&etdev->TCBSendQLock, flags); |
1381 | 1381 | ||
1382 | pMpTcb = etdev->TxRing.CurrSendHead; | 1382 | pMpTcb = etdev->TxRing.CurrSendHead; |
1383 | } | 1383 | } |
@@ -1388,7 +1388,7 @@ void et131x_free_busy_send_packets(struct et131x_adapter *etdev) | |||
1388 | BUG(); | 1388 | BUG(); |
1389 | } | 1389 | } |
1390 | 1390 | ||
1391 | spin_unlock_irqrestore(&etdev->TCBSendQLock, lockflags); | 1391 | spin_unlock_irqrestore(&etdev->TCBSendQLock, flags); |
1392 | 1392 | ||
1393 | etdev->TxRing.nBusySend = 0; | 1393 | etdev->TxRing.nBusySend = 0; |
1394 | 1394 | ||
@@ -1429,7 +1429,7 @@ void et131x_handle_send_interrupt(struct et131x_adapter *etdev) | |||
1429 | */ | 1429 | */ |
1430 | static void et131x_update_tcb_list(struct et131x_adapter *etdev) | 1430 | static void et131x_update_tcb_list(struct et131x_adapter *etdev) |
1431 | { | 1431 | { |
1432 | unsigned long lockflags; | 1432 | unsigned long flags; |
1433 | DMA10W_t ServiceComplete; | 1433 | DMA10W_t ServiceComplete; |
1434 | PMP_TCB pMpTcb; | 1434 | PMP_TCB pMpTcb; |
1435 | 1435 | ||
@@ -1439,7 +1439,7 @@ static void et131x_update_tcb_list(struct et131x_adapter *etdev) | |||
1439 | /* Has the ring wrapped? Process any descriptors that do not have | 1439 | /* Has the ring wrapped? Process any descriptors that do not have |
1440 | * the same "wrap" indicator as the current completion indicator | 1440 | * the same "wrap" indicator as the current completion indicator |
1441 | */ | 1441 | */ |
1442 | spin_lock_irqsave(&etdev->TCBSendQLock, lockflags); | 1442 | spin_lock_irqsave(&etdev->TCBSendQLock, flags); |
1443 | 1443 | ||
1444 | pMpTcb = etdev->TxRing.CurrSendHead; | 1444 | pMpTcb = etdev->TxRing.CurrSendHead; |
1445 | while (pMpTcb && | 1445 | while (pMpTcb && |
@@ -1450,9 +1450,9 @@ static void et131x_update_tcb_list(struct et131x_adapter *etdev) | |||
1450 | if (pMpTcb->Next == NULL) | 1450 | if (pMpTcb->Next == NULL) |
1451 | etdev->TxRing.CurrSendTail = NULL; | 1451 | etdev->TxRing.CurrSendTail = NULL; |
1452 | 1452 | ||
1453 | spin_unlock_irqrestore(&etdev->TCBSendQLock, lockflags); | 1453 | spin_unlock_irqrestore(&etdev->TCBSendQLock, flags); |
1454 | MP_FREE_SEND_PACKET_FUN(etdev, pMpTcb); | 1454 | MP_FREE_SEND_PACKET_FUN(etdev, pMpTcb); |
1455 | spin_lock_irqsave(&etdev->TCBSendQLock, lockflags); | 1455 | spin_lock_irqsave(&etdev->TCBSendQLock, flags); |
1456 | 1456 | ||
1457 | /* Goto the next packet */ | 1457 | /* Goto the next packet */ |
1458 | pMpTcb = etdev->TxRing.CurrSendHead; | 1458 | pMpTcb = etdev->TxRing.CurrSendHead; |
@@ -1465,9 +1465,9 @@ static void et131x_update_tcb_list(struct et131x_adapter *etdev) | |||
1465 | if (pMpTcb->Next == NULL) | 1465 | if (pMpTcb->Next == NULL) |
1466 | etdev->TxRing.CurrSendTail = NULL; | 1466 | etdev->TxRing.CurrSendTail = NULL; |
1467 | 1467 | ||
1468 | spin_unlock_irqrestore(&etdev->TCBSendQLock, lockflags); | 1468 | spin_unlock_irqrestore(&etdev->TCBSendQLock, flags); |
1469 | MP_FREE_SEND_PACKET_FUN(etdev, pMpTcb); | 1469 | MP_FREE_SEND_PACKET_FUN(etdev, pMpTcb); |
1470 | spin_lock_irqsave(&etdev->TCBSendQLock, lockflags); | 1470 | spin_lock_irqsave(&etdev->TCBSendQLock, flags); |
1471 | 1471 | ||
1472 | /* Goto the next packet */ | 1472 | /* Goto the next packet */ |
1473 | pMpTcb = etdev->TxRing.CurrSendHead; | 1473 | pMpTcb = etdev->TxRing.CurrSendHead; |
@@ -1477,7 +1477,7 @@ static void et131x_update_tcb_list(struct et131x_adapter *etdev) | |||
1477 | if (etdev->TxRing.nBusySend <= (NUM_TCB / 3)) | 1477 | if (etdev->TxRing.nBusySend <= (NUM_TCB / 3)) |
1478 | netif_wake_queue(etdev->netdev); | 1478 | netif_wake_queue(etdev->netdev); |
1479 | 1479 | ||
1480 | spin_unlock_irqrestore(&etdev->TCBSendQLock, lockflags); | 1480 | spin_unlock_irqrestore(&etdev->TCBSendQLock, flags); |
1481 | } | 1481 | } |
1482 | 1482 | ||
1483 | /** | 1483 | /** |
@@ -1489,9 +1489,9 @@ static void et131x_update_tcb_list(struct et131x_adapter *etdev) | |||
1489 | */ | 1489 | */ |
1490 | static void et131x_check_send_wait_list(struct et131x_adapter *etdev) | 1490 | static void et131x_check_send_wait_list(struct et131x_adapter *etdev) |
1491 | { | 1491 | { |
1492 | unsigned long lockflags; | 1492 | unsigned long flags; |
1493 | 1493 | ||
1494 | spin_lock_irqsave(&etdev->SendWaitLock, lockflags); | 1494 | spin_lock_irqsave(&etdev->SendWaitLock, flags); |
1495 | 1495 | ||
1496 | while (!list_empty(&etdev->TxRing.SendWaitQueue) && | 1496 | while (!list_empty(&etdev->TxRing.SendWaitQueue) && |
1497 | MP_TCB_RESOURCES_AVAILABLE(etdev)) { | 1497 | MP_TCB_RESOURCES_AVAILABLE(etdev)) { |
@@ -1508,5 +1508,5 @@ static void et131x_check_send_wait_list(struct et131x_adapter *etdev) | |||
1508 | etdev->TxRing.nWaitSend); | 1508 | etdev->TxRing.nWaitSend); |
1509 | } | 1509 | } |
1510 | 1510 | ||
1511 | spin_unlock_irqrestore(&etdev->SendWaitLock, lockflags); | 1511 | spin_unlock_irqrestore(&etdev->SendWaitLock, flags); |
1512 | } | 1512 | } |
diff --git a/drivers/staging/et131x/et131x_initpci.c b/drivers/staging/et131x/et131x_initpci.c index 8455e0a8471a..b9018e6f5660 100644 --- a/drivers/staging/et131x/et131x_initpci.c +++ b/drivers/staging/et131x/et131x_initpci.c | |||
@@ -492,18 +492,18 @@ void et131x_error_timer_handler(unsigned long data) | |||
492 | void et131x_link_detection_handler(unsigned long data) | 492 | void et131x_link_detection_handler(unsigned long data) |
493 | { | 493 | { |
494 | struct et131x_adapter *etdev = (struct et131x_adapter *) data; | 494 | struct et131x_adapter *etdev = (struct et131x_adapter *) data; |
495 | unsigned long lockflags; | 495 | unsigned long flags; |
496 | 496 | ||
497 | /* Let everyone know that we have run */ | 497 | /* Let everyone know that we have run */ |
498 | etdev->bLinkTimerActive = false; | 498 | etdev->bLinkTimerActive = false; |
499 | 499 | ||
500 | if (etdev->MediaState == 0) { | 500 | if (etdev->MediaState == 0) { |
501 | spin_lock_irqsave(&etdev->Lock, lockflags); | 501 | spin_lock_irqsave(&etdev->Lock, flags); |
502 | 502 | ||
503 | etdev->MediaState = NETIF_STATUS_MEDIA_DISCONNECT; | 503 | etdev->MediaState = NETIF_STATUS_MEDIA_DISCONNECT; |
504 | MP_CLEAR_FLAG(etdev, fMP_ADAPTER_LINK_DETECTION); | 504 | MP_CLEAR_FLAG(etdev, fMP_ADAPTER_LINK_DETECTION); |
505 | 505 | ||
506 | spin_unlock_irqrestore(&etdev->Lock, lockflags); | 506 | spin_unlock_irqrestore(&etdev->Lock, flags); |
507 | 507 | ||
508 | netif_carrier_off(etdev->netdev); | 508 | netif_carrier_off(etdev->netdev); |
509 | 509 | ||
diff --git a/drivers/staging/et131x/et131x_netdev.c b/drivers/staging/et131x/et131x_netdev.c index 77fac9592cdb..74ba1774206b 100644 --- a/drivers/staging/et131x/et131x_netdev.c +++ b/drivers/staging/et131x/et131x_netdev.c | |||
@@ -467,12 +467,12 @@ void et131x_multicast(struct net_device *netdev) | |||
467 | struct et131x_adapter *adapter = netdev_priv(netdev); | 467 | struct et131x_adapter *adapter = netdev_priv(netdev); |
468 | uint32_t PacketFilter = 0; | 468 | uint32_t PacketFilter = 0; |
469 | uint32_t count; | 469 | uint32_t count; |
470 | unsigned long lockflags; | 470 | unsigned long flags; |
471 | struct dev_mc_list *mclist = netdev->mc_list; | 471 | struct dev_mc_list *mclist = netdev->mc_list; |
472 | 472 | ||
473 | DBG_ENTER(et131x_dbginfo); | 473 | DBG_ENTER(et131x_dbginfo); |
474 | 474 | ||
475 | spin_lock_irqsave(&adapter->Lock, lockflags); | 475 | spin_lock_irqsave(&adapter->Lock, flags); |
476 | 476 | ||
477 | /* Before we modify the platform-independent filter flags, store them | 477 | /* Before we modify the platform-independent filter flags, store them |
478 | * locally. This allows us to determine if anything's changed and if | 478 | * locally. This allows us to determine if anything's changed and if |
@@ -552,7 +552,7 @@ void et131x_multicast(struct net_device *netdev) | |||
552 | "NO UPDATE REQUIRED, FLAGS didn't change\n"); | 552 | "NO UPDATE REQUIRED, FLAGS didn't change\n"); |
553 | } | 553 | } |
554 | 554 | ||
555 | spin_unlock_irqrestore(&adapter->Lock, lockflags); | 555 | spin_unlock_irqrestore(&adapter->Lock, flags); |
556 | 556 | ||
557 | DBG_LEAVE(et131x_dbginfo); | 557 | DBG_LEAVE(et131x_dbginfo); |
558 | } | 558 | } |
@@ -610,7 +610,7 @@ void et131x_tx_timeout(struct net_device *netdev) | |||
610 | { | 610 | { |
611 | struct et131x_adapter *etdev = netdev_priv(netdev); | 611 | struct et131x_adapter *etdev = netdev_priv(netdev); |
612 | PMP_TCB pMpTcb; | 612 | PMP_TCB pMpTcb; |
613 | unsigned long lockflags; | 613 | unsigned long flags; |
614 | 614 | ||
615 | DBG_WARNING(et131x_dbginfo, "TX TIMEOUT\n"); | 615 | DBG_WARNING(et131x_dbginfo, "TX TIMEOUT\n"); |
616 | 616 | ||
@@ -635,7 +635,7 @@ void et131x_tx_timeout(struct net_device *netdev) | |||
635 | } | 635 | } |
636 | 636 | ||
637 | /* Is send stuck? */ | 637 | /* Is send stuck? */ |
638 | spin_lock_irqsave(&etdev->TCBSendQLock, lockflags); | 638 | spin_lock_irqsave(&etdev->TCBSendQLock, flags); |
639 | 639 | ||
640 | pMpTcb = etdev->TxRing.CurrSendHead; | 640 | pMpTcb = etdev->TxRing.CurrSendHead; |
641 | 641 | ||
@@ -660,7 +660,7 @@ void et131x_tx_timeout(struct net_device *netdev) | |||
660 | } | 660 | } |
661 | 661 | ||
662 | spin_unlock_irqrestore(&etdev->TCBSendQLock, | 662 | spin_unlock_irqrestore(&etdev->TCBSendQLock, |
663 | lockflags); | 663 | flags); |
664 | 664 | ||
665 | DBG_WARNING(et131x_dbginfo, | 665 | DBG_WARNING(et131x_dbginfo, |
666 | "Send stuck - reset. pMpTcb->WrIndex %x, Flags 0x%08x\n", | 666 | "Send stuck - reset. pMpTcb->WrIndex %x, Flags 0x%08x\n", |
@@ -689,7 +689,7 @@ void et131x_tx_timeout(struct net_device *netdev) | |||
689 | } | 689 | } |
690 | } | 690 | } |
691 | 691 | ||
692 | spin_unlock_irqrestore(&etdev->TCBSendQLock, lockflags); | 692 | spin_unlock_irqrestore(&etdev->TCBSendQLock, flags); |
693 | } | 693 | } |
694 | 694 | ||
695 | /** | 695 | /** |