diff options
author | Segher Boessenkool <segher@kernel.crashing.org> | 2006-03-30 07:49:40 -0500 |
---|---|---|
committer | Paul Mackerras <paulus@samba.org> | 2006-04-01 06:36:57 -0500 |
commit | 706c8c93ba4865a19e981b9770151a7a63c15794 (patch) | |
tree | 63cbae3a74bfe40adcb1efa9ea5a6e7f6898b8ca /drivers | |
parent | 8df83028cf52e3bf68dfd35a4b1a044b326a4a99 (diff) |
[PATCH] powerpc/pseries: Change H_StudlyCaps to H_SHOUTING_CAPS
Also cleans up some nearby whitespace problems.
Signed-off-by: Segher Boessenkool <segher@kernel.crashing.org>
Signed-off-by: Paul Mackerras <paulus@samba.org>
Diffstat (limited to 'drivers')
-rw-r--r-- | drivers/char/hvcs.c | 2 | ||||
-rw-r--r-- | drivers/net/ibmveth.c | 30 | ||||
-rw-r--r-- | drivers/scsi/ibmvscsi/rpa_vscsi.c | 10 |
3 files changed, 21 insertions, 21 deletions
diff --git a/drivers/char/hvcs.c b/drivers/char/hvcs.c index 327b00c3c45e..8d97b3911293 100644 --- a/drivers/char/hvcs.c +++ b/drivers/char/hvcs.c | |||
@@ -904,7 +904,7 @@ static int hvcs_enable_device(struct hvcs_struct *hvcsd, uint32_t unit_address, | |||
904 | * It is possible the vty-server was removed after the irq was | 904 | * It is possible the vty-server was removed after the irq was |
905 | * requested but before we have time to enable interrupts. | 905 | * requested but before we have time to enable interrupts. |
906 | */ | 906 | */ |
907 | if (vio_enable_interrupts(vdev) == H_Success) | 907 | if (vio_enable_interrupts(vdev) == H_SUCCESS) |
908 | return 0; | 908 | return 0; |
909 | else { | 909 | else { |
910 | printk(KERN_ERR "HVCS: int enable failed for" | 910 | printk(KERN_ERR "HVCS: int enable failed for" |
diff --git a/drivers/net/ibmveth.c b/drivers/net/ibmveth.c index ceb98fd398af..52d01027d9e7 100644 --- a/drivers/net/ibmveth.c +++ b/drivers/net/ibmveth.c | |||
@@ -235,7 +235,7 @@ static void ibmveth_replenish_buffer_pool(struct ibmveth_adapter *adapter, struc | |||
235 | 235 | ||
236 | lpar_rc = h_add_logical_lan_buffer(adapter->vdev->unit_address, desc.desc); | 236 | lpar_rc = h_add_logical_lan_buffer(adapter->vdev->unit_address, desc.desc); |
237 | 237 | ||
238 | if(lpar_rc != H_Success) { | 238 | if(lpar_rc != H_SUCCESS) { |
239 | pool->free_map[free_index] = index; | 239 | pool->free_map[free_index] = index; |
240 | pool->skbuff[index] = NULL; | 240 | pool->skbuff[index] = NULL; |
241 | pool->consumer_index--; | 241 | pool->consumer_index--; |
@@ -373,7 +373,7 @@ static void ibmveth_rxq_recycle_buffer(struct ibmveth_adapter *adapter) | |||
373 | 373 | ||
374 | lpar_rc = h_add_logical_lan_buffer(adapter->vdev->unit_address, desc.desc); | 374 | lpar_rc = h_add_logical_lan_buffer(adapter->vdev->unit_address, desc.desc); |
375 | 375 | ||
376 | if(lpar_rc != H_Success) { | 376 | if(lpar_rc != H_SUCCESS) { |
377 | ibmveth_debug_printk("h_add_logical_lan_buffer failed during recycle rc=%ld", lpar_rc); | 377 | ibmveth_debug_printk("h_add_logical_lan_buffer failed during recycle rc=%ld", lpar_rc); |
378 | ibmveth_remove_buffer_from_pool(adapter, adapter->rx_queue.queue_addr[adapter->rx_queue.index].correlator); | 378 | ibmveth_remove_buffer_from_pool(adapter, adapter->rx_queue.queue_addr[adapter->rx_queue.index].correlator); |
379 | } | 379 | } |
@@ -511,7 +511,7 @@ static int ibmveth_open(struct net_device *netdev) | |||
511 | adapter->filter_list_dma, | 511 | adapter->filter_list_dma, |
512 | mac_address); | 512 | mac_address); |
513 | 513 | ||
514 | if(lpar_rc != H_Success) { | 514 | if(lpar_rc != H_SUCCESS) { |
515 | ibmveth_error_printk("h_register_logical_lan failed with %ld\n", lpar_rc); | 515 | ibmveth_error_printk("h_register_logical_lan failed with %ld\n", lpar_rc); |
516 | ibmveth_error_printk("buffer TCE:0x%lx filter TCE:0x%lx rxq desc:0x%lx MAC:0x%lx\n", | 516 | ibmveth_error_printk("buffer TCE:0x%lx filter TCE:0x%lx rxq desc:0x%lx MAC:0x%lx\n", |
517 | adapter->buffer_list_dma, | 517 | adapter->buffer_list_dma, |
@@ -527,7 +527,7 @@ static int ibmveth_open(struct net_device *netdev) | |||
527 | ibmveth_error_printk("unable to request irq 0x%x, rc %d\n", netdev->irq, rc); | 527 | ibmveth_error_printk("unable to request irq 0x%x, rc %d\n", netdev->irq, rc); |
528 | do { | 528 | do { |
529 | rc = h_free_logical_lan(adapter->vdev->unit_address); | 529 | rc = h_free_logical_lan(adapter->vdev->unit_address); |
530 | } while (H_isLongBusy(rc) || (rc == H_Busy)); | 530 | } while (H_IS_LONG_BUSY(rc) || (rc == H_BUSY)); |
531 | 531 | ||
532 | ibmveth_cleanup(adapter); | 532 | ibmveth_cleanup(adapter); |
533 | return rc; | 533 | return rc; |
@@ -556,9 +556,9 @@ static int ibmveth_close(struct net_device *netdev) | |||
556 | 556 | ||
557 | do { | 557 | do { |
558 | lpar_rc = h_free_logical_lan(adapter->vdev->unit_address); | 558 | lpar_rc = h_free_logical_lan(adapter->vdev->unit_address); |
559 | } while (H_isLongBusy(lpar_rc) || (lpar_rc == H_Busy)); | 559 | } while (H_IS_LONG_BUSY(lpar_rc) || (lpar_rc == H_BUSY)); |
560 | 560 | ||
561 | if(lpar_rc != H_Success) | 561 | if(lpar_rc != H_SUCCESS) |
562 | { | 562 | { |
563 | ibmveth_error_printk("h_free_logical_lan failed with %lx, continuing with close\n", | 563 | ibmveth_error_printk("h_free_logical_lan failed with %lx, continuing with close\n", |
564 | lpar_rc); | 564 | lpar_rc); |
@@ -693,9 +693,9 @@ static int ibmveth_start_xmit(struct sk_buff *skb, struct net_device *netdev) | |||
693 | desc[4].desc, | 693 | desc[4].desc, |
694 | desc[5].desc, | 694 | desc[5].desc, |
695 | correlator); | 695 | correlator); |
696 | } while ((lpar_rc == H_Busy) && (retry_count--)); | 696 | } while ((lpar_rc == H_BUSY) && (retry_count--)); |
697 | 697 | ||
698 | if(lpar_rc != H_Success && lpar_rc != H_Dropped) { | 698 | if(lpar_rc != H_SUCCESS && lpar_rc != H_DROPPED) { |
699 | int i; | 699 | int i; |
700 | ibmveth_error_printk("tx: h_send_logical_lan failed with rc=%ld\n", lpar_rc); | 700 | ibmveth_error_printk("tx: h_send_logical_lan failed with rc=%ld\n", lpar_rc); |
701 | for(i = 0; i < 6; i++) { | 701 | for(i = 0; i < 6; i++) { |
@@ -786,14 +786,14 @@ static int ibmveth_poll(struct net_device *netdev, int *budget) | |||
786 | /* we think we are done - reenable interrupts, then check once more to make sure we are done */ | 786 | /* we think we are done - reenable interrupts, then check once more to make sure we are done */ |
787 | lpar_rc = h_vio_signal(adapter->vdev->unit_address, VIO_IRQ_ENABLE); | 787 | lpar_rc = h_vio_signal(adapter->vdev->unit_address, VIO_IRQ_ENABLE); |
788 | 788 | ||
789 | ibmveth_assert(lpar_rc == H_Success); | 789 | ibmveth_assert(lpar_rc == H_SUCCESS); |
790 | 790 | ||
791 | netif_rx_complete(netdev); | 791 | netif_rx_complete(netdev); |
792 | 792 | ||
793 | if(ibmveth_rxq_pending_buffer(adapter) && netif_rx_reschedule(netdev, frames_processed)) | 793 | if(ibmveth_rxq_pending_buffer(adapter) && netif_rx_reschedule(netdev, frames_processed)) |
794 | { | 794 | { |
795 | lpar_rc = h_vio_signal(adapter->vdev->unit_address, VIO_IRQ_DISABLE); | 795 | lpar_rc = h_vio_signal(adapter->vdev->unit_address, VIO_IRQ_DISABLE); |
796 | ibmveth_assert(lpar_rc == H_Success); | 796 | ibmveth_assert(lpar_rc == H_SUCCESS); |
797 | more_work = 1; | 797 | more_work = 1; |
798 | goto restart_poll; | 798 | goto restart_poll; |
799 | } | 799 | } |
@@ -813,7 +813,7 @@ static irqreturn_t ibmveth_interrupt(int irq, void *dev_instance, struct pt_regs | |||
813 | 813 | ||
814 | if(netif_rx_schedule_prep(netdev)) { | 814 | if(netif_rx_schedule_prep(netdev)) { |
815 | lpar_rc = h_vio_signal(adapter->vdev->unit_address, VIO_IRQ_DISABLE); | 815 | lpar_rc = h_vio_signal(adapter->vdev->unit_address, VIO_IRQ_DISABLE); |
816 | ibmveth_assert(lpar_rc == H_Success); | 816 | ibmveth_assert(lpar_rc == H_SUCCESS); |
817 | __netif_rx_schedule(netdev); | 817 | __netif_rx_schedule(netdev); |
818 | } | 818 | } |
819 | return IRQ_HANDLED; | 819 | return IRQ_HANDLED; |
@@ -835,7 +835,7 @@ static void ibmveth_set_multicast_list(struct net_device *netdev) | |||
835 | IbmVethMcastEnableRecv | | 835 | IbmVethMcastEnableRecv | |
836 | IbmVethMcastDisableFiltering, | 836 | IbmVethMcastDisableFiltering, |
837 | 0); | 837 | 0); |
838 | if(lpar_rc != H_Success) { | 838 | if(lpar_rc != H_SUCCESS) { |
839 | ibmveth_error_printk("h_multicast_ctrl rc=%ld when entering promisc mode\n", lpar_rc); | 839 | ibmveth_error_printk("h_multicast_ctrl rc=%ld when entering promisc mode\n", lpar_rc); |
840 | } | 840 | } |
841 | } else { | 841 | } else { |
@@ -847,7 +847,7 @@ static void ibmveth_set_multicast_list(struct net_device *netdev) | |||
847 | IbmVethMcastDisableFiltering | | 847 | IbmVethMcastDisableFiltering | |
848 | IbmVethMcastClearFilterTable, | 848 | IbmVethMcastClearFilterTable, |
849 | 0); | 849 | 0); |
850 | if(lpar_rc != H_Success) { | 850 | if(lpar_rc != H_SUCCESS) { |
851 | ibmveth_error_printk("h_multicast_ctrl rc=%ld when attempting to clear filter table\n", lpar_rc); | 851 | ibmveth_error_printk("h_multicast_ctrl rc=%ld when attempting to clear filter table\n", lpar_rc); |
852 | } | 852 | } |
853 | /* add the addresses to the filter table */ | 853 | /* add the addresses to the filter table */ |
@@ -858,7 +858,7 @@ static void ibmveth_set_multicast_list(struct net_device *netdev) | |||
858 | lpar_rc = h_multicast_ctrl(adapter->vdev->unit_address, | 858 | lpar_rc = h_multicast_ctrl(adapter->vdev->unit_address, |
859 | IbmVethMcastAddFilter, | 859 | IbmVethMcastAddFilter, |
860 | mcast_addr); | 860 | mcast_addr); |
861 | if(lpar_rc != H_Success) { | 861 | if(lpar_rc != H_SUCCESS) { |
862 | ibmveth_error_printk("h_multicast_ctrl rc=%ld when adding an entry to the filter table\n", lpar_rc); | 862 | ibmveth_error_printk("h_multicast_ctrl rc=%ld when adding an entry to the filter table\n", lpar_rc); |
863 | } | 863 | } |
864 | } | 864 | } |
@@ -867,7 +867,7 @@ static void ibmveth_set_multicast_list(struct net_device *netdev) | |||
867 | lpar_rc = h_multicast_ctrl(adapter->vdev->unit_address, | 867 | lpar_rc = h_multicast_ctrl(adapter->vdev->unit_address, |
868 | IbmVethMcastEnableFiltering, | 868 | IbmVethMcastEnableFiltering, |
869 | 0); | 869 | 0); |
870 | if(lpar_rc != H_Success) { | 870 | if(lpar_rc != H_SUCCESS) { |
871 | ibmveth_error_printk("h_multicast_ctrl rc=%ld when enabling filtering\n", lpar_rc); | 871 | ibmveth_error_printk("h_multicast_ctrl rc=%ld when enabling filtering\n", lpar_rc); |
872 | } | 872 | } |
873 | } | 873 | } |
diff --git a/drivers/scsi/ibmvscsi/rpa_vscsi.c b/drivers/scsi/ibmvscsi/rpa_vscsi.c index f47dd87c05e7..892e8ed63091 100644 --- a/drivers/scsi/ibmvscsi/rpa_vscsi.c +++ b/drivers/scsi/ibmvscsi/rpa_vscsi.c | |||
@@ -80,7 +80,7 @@ void ibmvscsi_release_crq_queue(struct crq_queue *queue, | |||
80 | tasklet_kill(&hostdata->srp_task); | 80 | tasklet_kill(&hostdata->srp_task); |
81 | do { | 81 | do { |
82 | rc = plpar_hcall_norets(H_FREE_CRQ, vdev->unit_address); | 82 | rc = plpar_hcall_norets(H_FREE_CRQ, vdev->unit_address); |
83 | } while ((rc == H_Busy) || (H_isLongBusy(rc))); | 83 | } while ((rc == H_BUSY) || (H_IS_LONG_BUSY(rc))); |
84 | dma_unmap_single(hostdata->dev, | 84 | dma_unmap_single(hostdata->dev, |
85 | queue->msg_token, | 85 | queue->msg_token, |
86 | queue->size * sizeof(*queue->msgs), DMA_BIDIRECTIONAL); | 86 | queue->size * sizeof(*queue->msgs), DMA_BIDIRECTIONAL); |
@@ -230,7 +230,7 @@ int ibmvscsi_init_crq_queue(struct crq_queue *queue, | |||
230 | rc = plpar_hcall_norets(H_REG_CRQ, | 230 | rc = plpar_hcall_norets(H_REG_CRQ, |
231 | vdev->unit_address, | 231 | vdev->unit_address, |
232 | queue->msg_token, PAGE_SIZE); | 232 | queue->msg_token, PAGE_SIZE); |
233 | if (rc == H_Resource) | 233 | if (rc == H_RESOURCE) |
234 | /* maybe kexecing and resource is busy. try a reset */ | 234 | /* maybe kexecing and resource is busy. try a reset */ |
235 | rc = ibmvscsi_reset_crq_queue(queue, | 235 | rc = ibmvscsi_reset_crq_queue(queue, |
236 | hostdata); | 236 | hostdata); |
@@ -269,7 +269,7 @@ int ibmvscsi_init_crq_queue(struct crq_queue *queue, | |||
269 | req_irq_failed: | 269 | req_irq_failed: |
270 | do { | 270 | do { |
271 | rc = plpar_hcall_norets(H_FREE_CRQ, vdev->unit_address); | 271 | rc = plpar_hcall_norets(H_FREE_CRQ, vdev->unit_address); |
272 | } while ((rc == H_Busy) || (H_isLongBusy(rc))); | 272 | } while ((rc == H_BUSY) || (H_IS_LONG_BUSY(rc))); |
273 | reg_crq_failed: | 273 | reg_crq_failed: |
274 | dma_unmap_single(hostdata->dev, | 274 | dma_unmap_single(hostdata->dev, |
275 | queue->msg_token, | 275 | queue->msg_token, |
@@ -295,7 +295,7 @@ int ibmvscsi_reenable_crq_queue(struct crq_queue *queue, | |||
295 | /* Re-enable the CRQ */ | 295 | /* Re-enable the CRQ */ |
296 | do { | 296 | do { |
297 | rc = plpar_hcall_norets(H_ENABLE_CRQ, vdev->unit_address); | 297 | rc = plpar_hcall_norets(H_ENABLE_CRQ, vdev->unit_address); |
298 | } while ((rc == H_InProgress) || (rc == H_Busy) || (H_isLongBusy(rc))); | 298 | } while ((rc == H_IN_PROGRESS) || (rc == H_BUSY) || (H_IS_LONG_BUSY(rc))); |
299 | 299 | ||
300 | if (rc) | 300 | if (rc) |
301 | printk(KERN_ERR "ibmvscsi: Error %d enabling adapter\n", rc); | 301 | printk(KERN_ERR "ibmvscsi: Error %d enabling adapter\n", rc); |
@@ -317,7 +317,7 @@ int ibmvscsi_reset_crq_queue(struct crq_queue *queue, | |||
317 | /* Close the CRQ */ | 317 | /* Close the CRQ */ |
318 | do { | 318 | do { |
319 | rc = plpar_hcall_norets(H_FREE_CRQ, vdev->unit_address); | 319 | rc = plpar_hcall_norets(H_FREE_CRQ, vdev->unit_address); |
320 | } while ((rc == H_Busy) || (H_isLongBusy(rc))); | 320 | } while ((rc == H_BUSY) || (H_IS_LONG_BUSY(rc))); |
321 | 321 | ||
322 | /* Clean out the queue */ | 322 | /* Clean out the queue */ |
323 | memset(queue->msgs, 0x00, PAGE_SIZE); | 323 | memset(queue->msgs, 0x00, PAGE_SIZE); |