diff options
author | Harvey Harrison <harvey.harrison@gmail.com> | 2008-04-17 00:01:10 -0400 |
---|---|---|
committer | Roland Dreier <rolandd@cisco.com> | 2008-04-17 00:01:10 -0400 |
commit | 3371836383d63b627b228875f5ac63023cbf11d2 (patch) | |
tree | b997894d9774bdc07f7df76ceca48e6a848760c8 /drivers/infiniband/hw/cxgb3/iwch_cm.c | |
parent | e8e91f6b4dc1179a70b0d21241b769c0ebfaa129 (diff) |
IB: Replace remaining __FUNCTION__ occurrences with __func__
__FUNCTION__ is gcc-specific, use __func__ instead.
Signed-off-by: Harvey Harrison <harvey.harrison@gmail.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Roland Dreier <rolandd@cisco.com>
Diffstat (limited to 'drivers/infiniband/hw/cxgb3/iwch_cm.c')
-rw-r--r-- | drivers/infiniband/hw/cxgb3/iwch_cm.c | 166 |
1 files changed, 83 insertions, 83 deletions
diff --git a/drivers/infiniband/hw/cxgb3/iwch_cm.c b/drivers/infiniband/hw/cxgb3/iwch_cm.c index 99f2f2a46bf..72ca360c3db 100644 --- a/drivers/infiniband/hw/cxgb3/iwch_cm.c +++ b/drivers/infiniband/hw/cxgb3/iwch_cm.c | |||
@@ -110,9 +110,9 @@ static void connect_reply_upcall(struct iwch_ep *ep, int status); | |||
110 | 110 | ||
111 | static void start_ep_timer(struct iwch_ep *ep) | 111 | static void start_ep_timer(struct iwch_ep *ep) |
112 | { | 112 | { |
113 | PDBG("%s ep %p\n", __FUNCTION__, ep); | 113 | PDBG("%s ep %p\n", __func__, ep); |
114 | if (timer_pending(&ep->timer)) { | 114 | if (timer_pending(&ep->timer)) { |
115 | PDBG("%s stopped / restarted timer ep %p\n", __FUNCTION__, ep); | 115 | PDBG("%s stopped / restarted timer ep %p\n", __func__, ep); |
116 | del_timer_sync(&ep->timer); | 116 | del_timer_sync(&ep->timer); |
117 | } else | 117 | } else |
118 | get_ep(&ep->com); | 118 | get_ep(&ep->com); |
@@ -124,7 +124,7 @@ static void start_ep_timer(struct iwch_ep *ep) | |||
124 | 124 | ||
125 | static void stop_ep_timer(struct iwch_ep *ep) | 125 | static void stop_ep_timer(struct iwch_ep *ep) |
126 | { | 126 | { |
127 | PDBG("%s ep %p\n", __FUNCTION__, ep); | 127 | PDBG("%s ep %p\n", __func__, ep); |
128 | del_timer_sync(&ep->timer); | 128 | del_timer_sync(&ep->timer); |
129 | put_ep(&ep->com); | 129 | put_ep(&ep->com); |
130 | } | 130 | } |
@@ -190,7 +190,7 @@ int iwch_resume_tid(struct iwch_ep *ep) | |||
190 | 190 | ||
191 | static void set_emss(struct iwch_ep *ep, u16 opt) | 191 | static void set_emss(struct iwch_ep *ep, u16 opt) |
192 | { | 192 | { |
193 | PDBG("%s ep %p opt %u\n", __FUNCTION__, ep, opt); | 193 | PDBG("%s ep %p opt %u\n", __func__, ep, opt); |
194 | ep->emss = T3C_DATA(ep->com.tdev)->mtus[G_TCPOPT_MSS(opt)] - 40; | 194 | ep->emss = T3C_DATA(ep->com.tdev)->mtus[G_TCPOPT_MSS(opt)] - 40; |
195 | if (G_TCPOPT_TSTAMP(opt)) | 195 | if (G_TCPOPT_TSTAMP(opt)) |
196 | ep->emss -= 12; | 196 | ep->emss -= 12; |
@@ -220,7 +220,7 @@ static void state_set(struct iwch_ep_common *epc, enum iwch_ep_state new) | |||
220 | unsigned long flags; | 220 | unsigned long flags; |
221 | 221 | ||
222 | spin_lock_irqsave(&epc->lock, flags); | 222 | spin_lock_irqsave(&epc->lock, flags); |
223 | PDBG("%s - %s -> %s\n", __FUNCTION__, states[epc->state], states[new]); | 223 | PDBG("%s - %s -> %s\n", __func__, states[epc->state], states[new]); |
224 | __state_set(epc, new); | 224 | __state_set(epc, new); |
225 | spin_unlock_irqrestore(&epc->lock, flags); | 225 | spin_unlock_irqrestore(&epc->lock, flags); |
226 | return; | 226 | return; |
@@ -236,7 +236,7 @@ static void *alloc_ep(int size, gfp_t gfp) | |||
236 | spin_lock_init(&epc->lock); | 236 | spin_lock_init(&epc->lock); |
237 | init_waitqueue_head(&epc->waitq); | 237 | init_waitqueue_head(&epc->waitq); |
238 | } | 238 | } |
239 | PDBG("%s alloc ep %p\n", __FUNCTION__, epc); | 239 | PDBG("%s alloc ep %p\n", __func__, epc); |
240 | return epc; | 240 | return epc; |
241 | } | 241 | } |
242 | 242 | ||
@@ -244,13 +244,13 @@ void __free_ep(struct kref *kref) | |||
244 | { | 244 | { |
245 | struct iwch_ep_common *epc; | 245 | struct iwch_ep_common *epc; |
246 | epc = container_of(kref, struct iwch_ep_common, kref); | 246 | epc = container_of(kref, struct iwch_ep_common, kref); |
247 | PDBG("%s ep %p state %s\n", __FUNCTION__, epc, states[state_read(epc)]); | 247 | PDBG("%s ep %p state %s\n", __func__, epc, states[state_read(epc)]); |
248 | kfree(epc); | 248 | kfree(epc); |
249 | } | 249 | } |
250 | 250 | ||
251 | static void release_ep_resources(struct iwch_ep *ep) | 251 | static void release_ep_resources(struct iwch_ep *ep) |
252 | { | 252 | { |
253 | PDBG("%s ep %p tid %d\n", __FUNCTION__, ep, ep->hwtid); | 253 | PDBG("%s ep %p tid %d\n", __func__, ep, ep->hwtid); |
254 | cxgb3_remove_tid(ep->com.tdev, (void *)ep, ep->hwtid); | 254 | cxgb3_remove_tid(ep->com.tdev, (void *)ep, ep->hwtid); |
255 | dst_release(ep->dst); | 255 | dst_release(ep->dst); |
256 | l2t_release(L2DATA(ep->com.tdev), ep->l2t); | 256 | l2t_release(L2DATA(ep->com.tdev), ep->l2t); |
@@ -349,7 +349,7 @@ static unsigned int find_best_mtu(const struct t3c_data *d, unsigned short mtu) | |||
349 | 349 | ||
350 | static void arp_failure_discard(struct t3cdev *dev, struct sk_buff *skb) | 350 | static void arp_failure_discard(struct t3cdev *dev, struct sk_buff *skb) |
351 | { | 351 | { |
352 | PDBG("%s t3cdev %p\n", __FUNCTION__, dev); | 352 | PDBG("%s t3cdev %p\n", __func__, dev); |
353 | kfree_skb(skb); | 353 | kfree_skb(skb); |
354 | } | 354 | } |
355 | 355 | ||
@@ -370,7 +370,7 @@ static void abort_arp_failure(struct t3cdev *dev, struct sk_buff *skb) | |||
370 | { | 370 | { |
371 | struct cpl_abort_req *req = cplhdr(skb); | 371 | struct cpl_abort_req *req = cplhdr(skb); |
372 | 372 | ||
373 | PDBG("%s t3cdev %p\n", __FUNCTION__, dev); | 373 | PDBG("%s t3cdev %p\n", __func__, dev); |
374 | req->cmd = CPL_ABORT_NO_RST; | 374 | req->cmd = CPL_ABORT_NO_RST; |
375 | cxgb3_ofld_send(dev, skb); | 375 | cxgb3_ofld_send(dev, skb); |
376 | } | 376 | } |
@@ -380,10 +380,10 @@ static int send_halfclose(struct iwch_ep *ep, gfp_t gfp) | |||
380 | struct cpl_close_con_req *req; | 380 | struct cpl_close_con_req *req; |
381 | struct sk_buff *skb; | 381 | struct sk_buff *skb; |
382 | 382 | ||
383 | PDBG("%s ep %p\n", __FUNCTION__, ep); | 383 | PDBG("%s ep %p\n", __func__, ep); |
384 | skb = get_skb(NULL, sizeof(*req), gfp); | 384 | skb = get_skb(NULL, sizeof(*req), gfp); |
385 | if (!skb) { | 385 | if (!skb) { |
386 | printk(KERN_ERR MOD "%s - failed to alloc skb\n", __FUNCTION__); | 386 | printk(KERN_ERR MOD "%s - failed to alloc skb\n", __func__); |
387 | return -ENOMEM; | 387 | return -ENOMEM; |
388 | } | 388 | } |
389 | skb->priority = CPL_PRIORITY_DATA; | 389 | skb->priority = CPL_PRIORITY_DATA; |
@@ -400,11 +400,11 @@ static int send_abort(struct iwch_ep *ep, struct sk_buff *skb, gfp_t gfp) | |||
400 | { | 400 | { |
401 | struct cpl_abort_req *req; | 401 | struct cpl_abort_req *req; |
402 | 402 | ||
403 | PDBG("%s ep %p\n", __FUNCTION__, ep); | 403 | PDBG("%s ep %p\n", __func__, ep); |
404 | skb = get_skb(skb, sizeof(*req), gfp); | 404 | skb = get_skb(skb, sizeof(*req), gfp); |
405 | if (!skb) { | 405 | if (!skb) { |
406 | printk(KERN_ERR MOD "%s - failed to alloc skb.\n", | 406 | printk(KERN_ERR MOD "%s - failed to alloc skb.\n", |
407 | __FUNCTION__); | 407 | __func__); |
408 | return -ENOMEM; | 408 | return -ENOMEM; |
409 | } | 409 | } |
410 | skb->priority = CPL_PRIORITY_DATA; | 410 | skb->priority = CPL_PRIORITY_DATA; |
@@ -426,12 +426,12 @@ static int send_connect(struct iwch_ep *ep) | |||
426 | unsigned int mtu_idx; | 426 | unsigned int mtu_idx; |
427 | int wscale; | 427 | int wscale; |
428 | 428 | ||
429 | PDBG("%s ep %p\n", __FUNCTION__, ep); | 429 | PDBG("%s ep %p\n", __func__, ep); |
430 | 430 | ||
431 | skb = get_skb(NULL, sizeof(*req), GFP_KERNEL); | 431 | skb = get_skb(NULL, sizeof(*req), GFP_KERNEL); |
432 | if (!skb) { | 432 | if (!skb) { |
433 | printk(KERN_ERR MOD "%s - failed to alloc skb.\n", | 433 | printk(KERN_ERR MOD "%s - failed to alloc skb.\n", |
434 | __FUNCTION__); | 434 | __func__); |
435 | return -ENOMEM; | 435 | return -ENOMEM; |
436 | } | 436 | } |
437 | mtu_idx = find_best_mtu(T3C_DATA(ep->com.tdev), dst_mtu(ep->dst)); | 437 | mtu_idx = find_best_mtu(T3C_DATA(ep->com.tdev), dst_mtu(ep->dst)); |
@@ -470,7 +470,7 @@ static void send_mpa_req(struct iwch_ep *ep, struct sk_buff *skb) | |||
470 | struct mpa_message *mpa; | 470 | struct mpa_message *mpa; |
471 | int len; | 471 | int len; |
472 | 472 | ||
473 | PDBG("%s ep %p pd_len %d\n", __FUNCTION__, ep, ep->plen); | 473 | PDBG("%s ep %p pd_len %d\n", __func__, ep, ep->plen); |
474 | 474 | ||
475 | BUG_ON(skb_cloned(skb)); | 475 | BUG_ON(skb_cloned(skb)); |
476 | 476 | ||
@@ -530,13 +530,13 @@ static int send_mpa_reject(struct iwch_ep *ep, const void *pdata, u8 plen) | |||
530 | struct mpa_message *mpa; | 530 | struct mpa_message *mpa; |
531 | struct sk_buff *skb; | 531 | struct sk_buff *skb; |
532 | 532 | ||
533 | PDBG("%s ep %p plen %d\n", __FUNCTION__, ep, plen); | 533 | PDBG("%s ep %p plen %d\n", __func__, ep, plen); |
534 | 534 | ||
535 | mpalen = sizeof(*mpa) + plen; | 535 | mpalen = sizeof(*mpa) + plen; |
536 | 536 | ||
537 | skb = get_skb(NULL, mpalen + sizeof(*req), GFP_KERNEL); | 537 | skb = get_skb(NULL, mpalen + sizeof(*req), GFP_KERNEL); |
538 | if (!skb) { | 538 | if (!skb) { |
539 | printk(KERN_ERR MOD "%s - cannot alloc skb!\n", __FUNCTION__); | 539 | printk(KERN_ERR MOD "%s - cannot alloc skb!\n", __func__); |
540 | return -ENOMEM; | 540 | return -ENOMEM; |
541 | } | 541 | } |
542 | skb_reserve(skb, sizeof(*req)); | 542 | skb_reserve(skb, sizeof(*req)); |
@@ -580,13 +580,13 @@ static int send_mpa_reply(struct iwch_ep *ep, const void *pdata, u8 plen) | |||
580 | int len; | 580 | int len; |
581 | struct sk_buff *skb; | 581 | struct sk_buff *skb; |
582 | 582 | ||
583 | PDBG("%s ep %p plen %d\n", __FUNCTION__, ep, plen); | 583 | PDBG("%s ep %p plen %d\n", __func__, ep, plen); |
584 | 584 | ||
585 | mpalen = sizeof(*mpa) + plen; | 585 | mpalen = sizeof(*mpa) + plen; |
586 | 586 | ||
587 | skb = get_skb(NULL, mpalen + sizeof(*req), GFP_KERNEL); | 587 | skb = get_skb(NULL, mpalen + sizeof(*req), GFP_KERNEL); |
588 | if (!skb) { | 588 | if (!skb) { |
589 | printk(KERN_ERR MOD "%s - cannot alloc skb!\n", __FUNCTION__); | 589 | printk(KERN_ERR MOD "%s - cannot alloc skb!\n", __func__); |
590 | return -ENOMEM; | 590 | return -ENOMEM; |
591 | } | 591 | } |
592 | skb->priority = CPL_PRIORITY_DATA; | 592 | skb->priority = CPL_PRIORITY_DATA; |
@@ -630,7 +630,7 @@ static int act_establish(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) | |||
630 | struct cpl_act_establish *req = cplhdr(skb); | 630 | struct cpl_act_establish *req = cplhdr(skb); |
631 | unsigned int tid = GET_TID(req); | 631 | unsigned int tid = GET_TID(req); |
632 | 632 | ||
633 | PDBG("%s ep %p tid %d\n", __FUNCTION__, ep, tid); | 633 | PDBG("%s ep %p tid %d\n", __func__, ep, tid); |
634 | 634 | ||
635 | dst_confirm(ep->dst); | 635 | dst_confirm(ep->dst); |
636 | 636 | ||
@@ -663,7 +663,7 @@ static void close_complete_upcall(struct iwch_ep *ep) | |||
663 | { | 663 | { |
664 | struct iw_cm_event event; | 664 | struct iw_cm_event event; |
665 | 665 | ||
666 | PDBG("%s ep %p\n", __FUNCTION__, ep); | 666 | PDBG("%s ep %p\n", __func__, ep); |
667 | memset(&event, 0, sizeof(event)); | 667 | memset(&event, 0, sizeof(event)); |
668 | event.event = IW_CM_EVENT_CLOSE; | 668 | event.event = IW_CM_EVENT_CLOSE; |
669 | if (ep->com.cm_id) { | 669 | if (ep->com.cm_id) { |
@@ -680,7 +680,7 @@ static void peer_close_upcall(struct iwch_ep *ep) | |||
680 | { | 680 | { |
681 | struct iw_cm_event event; | 681 | struct iw_cm_event event; |
682 | 682 | ||
683 | PDBG("%s ep %p\n", __FUNCTION__, ep); | 683 | PDBG("%s ep %p\n", __func__, ep); |
684 | memset(&event, 0, sizeof(event)); | 684 | memset(&event, 0, sizeof(event)); |
685 | event.event = IW_CM_EVENT_DISCONNECT; | 685 | event.event = IW_CM_EVENT_DISCONNECT; |
686 | if (ep->com.cm_id) { | 686 | if (ep->com.cm_id) { |
@@ -694,7 +694,7 @@ static void peer_abort_upcall(struct iwch_ep *ep) | |||
694 | { | 694 | { |
695 | struct iw_cm_event event; | 695 | struct iw_cm_event event; |
696 | 696 | ||
697 | PDBG("%s ep %p\n", __FUNCTION__, ep); | 697 | PDBG("%s ep %p\n", __func__, ep); |
698 | memset(&event, 0, sizeof(event)); | 698 | memset(&event, 0, sizeof(event)); |
699 | event.event = IW_CM_EVENT_CLOSE; | 699 | event.event = IW_CM_EVENT_CLOSE; |
700 | event.status = -ECONNRESET; | 700 | event.status = -ECONNRESET; |
@@ -712,7 +712,7 @@ static void connect_reply_upcall(struct iwch_ep *ep, int status) | |||
712 | { | 712 | { |
713 | struct iw_cm_event event; | 713 | struct iw_cm_event event; |
714 | 714 | ||
715 | PDBG("%s ep %p status %d\n", __FUNCTION__, ep, status); | 715 | PDBG("%s ep %p status %d\n", __func__, ep, status); |
716 | memset(&event, 0, sizeof(event)); | 716 | memset(&event, 0, sizeof(event)); |
717 | event.event = IW_CM_EVENT_CONNECT_REPLY; | 717 | event.event = IW_CM_EVENT_CONNECT_REPLY; |
718 | event.status = status; | 718 | event.status = status; |
@@ -724,7 +724,7 @@ static void connect_reply_upcall(struct iwch_ep *ep, int status) | |||
724 | event.private_data = ep->mpa_pkt + sizeof(struct mpa_message); | 724 | event.private_data = ep->mpa_pkt + sizeof(struct mpa_message); |
725 | } | 725 | } |
726 | if (ep->com.cm_id) { | 726 | if (ep->com.cm_id) { |
727 | PDBG("%s ep %p tid %d status %d\n", __FUNCTION__, ep, | 727 | PDBG("%s ep %p tid %d status %d\n", __func__, ep, |
728 | ep->hwtid, status); | 728 | ep->hwtid, status); |
729 | ep->com.cm_id->event_handler(ep->com.cm_id, &event); | 729 | ep->com.cm_id->event_handler(ep->com.cm_id, &event); |
730 | } | 730 | } |
@@ -739,7 +739,7 @@ static void connect_request_upcall(struct iwch_ep *ep) | |||
739 | { | 739 | { |
740 | struct iw_cm_event event; | 740 | struct iw_cm_event event; |
741 | 741 | ||
742 | PDBG("%s ep %p tid %d\n", __FUNCTION__, ep, ep->hwtid); | 742 | PDBG("%s ep %p tid %d\n", __func__, ep, ep->hwtid); |
743 | memset(&event, 0, sizeof(event)); | 743 | memset(&event, 0, sizeof(event)); |
744 | event.event = IW_CM_EVENT_CONNECT_REQUEST; | 744 | event.event = IW_CM_EVENT_CONNECT_REQUEST; |
745 | event.local_addr = ep->com.local_addr; | 745 | event.local_addr = ep->com.local_addr; |
@@ -759,11 +759,11 @@ static void established_upcall(struct iwch_ep *ep) | |||
759 | { | 759 | { |
760 | struct iw_cm_event event; | 760 | struct iw_cm_event event; |
761 | 761 | ||
762 | PDBG("%s ep %p\n", __FUNCTION__, ep); | 762 | PDBG("%s ep %p\n", __func__, ep); |
763 | memset(&event, 0, sizeof(event)); | 763 | memset(&event, 0, sizeof(event)); |
764 | event.event = IW_CM_EVENT_ESTABLISHED; | 764 | event.event = IW_CM_EVENT_ESTABLISHED; |
765 | if (ep->com.cm_id) { | 765 | if (ep->com.cm_id) { |
766 | PDBG("%s ep %p tid %d\n", __FUNCTION__, ep, ep->hwtid); | 766 | PDBG("%s ep %p tid %d\n", __func__, ep, ep->hwtid); |
767 | ep->com.cm_id->event_handler(ep->com.cm_id, &event); | 767 | ep->com.cm_id->event_handler(ep->com.cm_id, &event); |
768 | } | 768 | } |
769 | } | 769 | } |
@@ -773,7 +773,7 @@ static int update_rx_credits(struct iwch_ep *ep, u32 credits) | |||
773 | struct cpl_rx_data_ack *req; | 773 | struct cpl_rx_data_ack *req; |
774 | struct sk_buff *skb; | 774 | struct sk_buff *skb; |
775 | 775 | ||
776 | PDBG("%s ep %p credits %u\n", __FUNCTION__, ep, credits); | 776 | PDBG("%s ep %p credits %u\n", __func__, ep, credits); |
777 | skb = get_skb(NULL, sizeof(*req), GFP_KERNEL); | 777 | skb = get_skb(NULL, sizeof(*req), GFP_KERNEL); |
778 | if (!skb) { | 778 | if (!skb) { |
779 | printk(KERN_ERR MOD "update_rx_credits - cannot alloc skb!\n"); | 779 | printk(KERN_ERR MOD "update_rx_credits - cannot alloc skb!\n"); |
@@ -797,7 +797,7 @@ static void process_mpa_reply(struct iwch_ep *ep, struct sk_buff *skb) | |||
797 | enum iwch_qp_attr_mask mask; | 797 | enum iwch_qp_attr_mask mask; |
798 | int err; | 798 | int err; |
799 | 799 | ||
800 | PDBG("%s ep %p\n", __FUNCTION__, ep); | 800 | PDBG("%s ep %p\n", __func__, ep); |
801 | 801 | ||
802 | /* | 802 | /* |
803 | * Stop mpa timer. If it expired, then the state has | 803 | * Stop mpa timer. If it expired, then the state has |
@@ -884,7 +884,7 @@ static void process_mpa_reply(struct iwch_ep *ep, struct sk_buff *skb) | |||
884 | ep->mpa_attr.xmit_marker_enabled = mpa->flags & MPA_MARKERS ? 1 : 0; | 884 | ep->mpa_attr.xmit_marker_enabled = mpa->flags & MPA_MARKERS ? 1 : 0; |
885 | ep->mpa_attr.version = mpa_rev; | 885 | ep->mpa_attr.version = mpa_rev; |
886 | PDBG("%s - crc_enabled=%d, recv_marker_enabled=%d, " | 886 | PDBG("%s - crc_enabled=%d, recv_marker_enabled=%d, " |
887 | "xmit_marker_enabled=%d, version=%d\n", __FUNCTION__, | 887 | "xmit_marker_enabled=%d, version=%d\n", __func__, |
888 | ep->mpa_attr.crc_enabled, ep->mpa_attr.recv_marker_enabled, | 888 | ep->mpa_attr.crc_enabled, ep->mpa_attr.recv_marker_enabled, |
889 | ep->mpa_attr.xmit_marker_enabled, ep->mpa_attr.version); | 889 | ep->mpa_attr.xmit_marker_enabled, ep->mpa_attr.version); |
890 | 890 | ||
@@ -915,7 +915,7 @@ static void process_mpa_request(struct iwch_ep *ep, struct sk_buff *skb) | |||
915 | struct mpa_message *mpa; | 915 | struct mpa_message *mpa; |
916 | u16 plen; | 916 | u16 plen; |
917 | 917 | ||
918 | PDBG("%s ep %p\n", __FUNCTION__, ep); | 918 | PDBG("%s ep %p\n", __func__, ep); |
919 | 919 | ||
920 | /* | 920 | /* |
921 | * Stop mpa timer. If it expired, then the state has | 921 | * Stop mpa timer. If it expired, then the state has |
@@ -935,7 +935,7 @@ static void process_mpa_request(struct iwch_ep *ep, struct sk_buff *skb) | |||
935 | return; | 935 | return; |
936 | } | 936 | } |
937 | 937 | ||
938 | PDBG("%s enter (%s line %u)\n", __FUNCTION__, __FILE__, __LINE__); | 938 | PDBG("%s enter (%s line %u)\n", __func__, __FILE__, __LINE__); |
939 | 939 | ||
940 | /* | 940 | /* |
941 | * Copy the new data into our accumulation buffer. | 941 | * Copy the new data into our accumulation buffer. |
@@ -950,7 +950,7 @@ static void process_mpa_request(struct iwch_ep *ep, struct sk_buff *skb) | |||
950 | */ | 950 | */ |
951 | if (ep->mpa_pkt_len < sizeof(*mpa)) | 951 | if (ep->mpa_pkt_len < sizeof(*mpa)) |
952 | return; | 952 | return; |
953 | PDBG("%s enter (%s line %u)\n", __FUNCTION__, __FILE__, __LINE__); | 953 | PDBG("%s enter (%s line %u)\n", __func__, __FILE__, __LINE__); |
954 | mpa = (struct mpa_message *) ep->mpa_pkt; | 954 | mpa = (struct mpa_message *) ep->mpa_pkt; |
955 | 955 | ||
956 | /* | 956 | /* |
@@ -1000,7 +1000,7 @@ static void process_mpa_request(struct iwch_ep *ep, struct sk_buff *skb) | |||
1000 | ep->mpa_attr.xmit_marker_enabled = mpa->flags & MPA_MARKERS ? 1 : 0; | 1000 | ep->mpa_attr.xmit_marker_enabled = mpa->flags & MPA_MARKERS ? 1 : 0; |
1001 | ep->mpa_attr.version = mpa_rev; | 1001 | ep->mpa_attr.version = mpa_rev; |
1002 | PDBG("%s - crc_enabled=%d, recv_marker_enabled=%d, " | 1002 | PDBG("%s - crc_enabled=%d, recv_marker_enabled=%d, " |
1003 | "xmit_marker_enabled=%d, version=%d\n", __FUNCTION__, | 1003 | "xmit_marker_enabled=%d, version=%d\n", __func__, |
1004 | ep->mpa_attr.crc_enabled, ep->mpa_attr.recv_marker_enabled, | 1004 | ep->mpa_attr.crc_enabled, ep->mpa_attr.recv_marker_enabled, |
1005 | ep->mpa_attr.xmit_marker_enabled, ep->mpa_attr.version); | 1005 | ep->mpa_attr.xmit_marker_enabled, ep->mpa_attr.version); |
1006 | 1006 | ||
@@ -1017,7 +1017,7 @@ static int rx_data(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) | |||
1017 | struct cpl_rx_data *hdr = cplhdr(skb); | 1017 | struct cpl_rx_data *hdr = cplhdr(skb); |
1018 | unsigned int dlen = ntohs(hdr->len); | 1018 | unsigned int dlen = ntohs(hdr->len); |
1019 | 1019 | ||
1020 | PDBG("%s ep %p dlen %u\n", __FUNCTION__, ep, dlen); | 1020 | PDBG("%s ep %p dlen %u\n", __func__, ep, dlen); |
1021 | 1021 | ||
1022 | skb_pull(skb, sizeof(*hdr)); | 1022 | skb_pull(skb, sizeof(*hdr)); |
1023 | skb_trim(skb, dlen); | 1023 | skb_trim(skb, dlen); |
@@ -1037,7 +1037,7 @@ static int rx_data(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) | |||
1037 | default: | 1037 | default: |
1038 | printk(KERN_ERR MOD "%s Unexpected streaming data." | 1038 | printk(KERN_ERR MOD "%s Unexpected streaming data." |
1039 | " ep %p state %d tid %d\n", | 1039 | " ep %p state %d tid %d\n", |
1040 | __FUNCTION__, ep, state_read(&ep->com), ep->hwtid); | 1040 | __func__, ep, state_read(&ep->com), ep->hwtid); |
1041 | 1041 | ||
1042 | /* | 1042 | /* |
1043 | * The ep will timeout and inform the ULP of the failure. | 1043 | * The ep will timeout and inform the ULP of the failure. |
@@ -1063,7 +1063,7 @@ static int tx_ack(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) | |||
1063 | struct cpl_wr_ack *hdr = cplhdr(skb); | 1063 | struct cpl_wr_ack *hdr = cplhdr(skb); |
1064 | unsigned int credits = ntohs(hdr->credits); | 1064 | unsigned int credits = ntohs(hdr->credits); |
1065 | 1065 | ||
1066 | PDBG("%s ep %p credits %u\n", __FUNCTION__, ep, credits); | 1066 | PDBG("%s ep %p credits %u\n", __func__, ep, credits); |
1067 | 1067 | ||
1068 | if (credits == 0) | 1068 | if (credits == 0) |
1069 | return CPL_RET_BUF_DONE; | 1069 | return CPL_RET_BUF_DONE; |
@@ -1084,7 +1084,7 @@ static int abort_rpl(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) | |||
1084 | { | 1084 | { |
1085 | struct iwch_ep *ep = ctx; | 1085 | struct iwch_ep *ep = ctx; |
1086 | 1086 | ||
1087 | PDBG("%s ep %p\n", __FUNCTION__, ep); | 1087 | PDBG("%s ep %p\n", __func__, ep); |
1088 | 1088 | ||
1089 | /* | 1089 | /* |
1090 | * We get 2 abort replies from the HW. The first one must | 1090 | * We get 2 abort replies from the HW. The first one must |
@@ -1115,7 +1115,7 @@ static int act_open_rpl(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) | |||
1115 | struct iwch_ep *ep = ctx; | 1115 | struct iwch_ep *ep = ctx; |
1116 | struct cpl_act_open_rpl *rpl = cplhdr(skb); | 1116 | struct cpl_act_open_rpl *rpl = cplhdr(skb); |
1117 | 1117 | ||
1118 | PDBG("%s ep %p status %u errno %d\n", __FUNCTION__, ep, rpl->status, | 1118 | PDBG("%s ep %p status %u errno %d\n", __func__, ep, rpl->status, |
1119 | status2errno(rpl->status)); | 1119 | status2errno(rpl->status)); |
1120 | connect_reply_upcall(ep, status2errno(rpl->status)); | 1120 | connect_reply_upcall(ep, status2errno(rpl->status)); |
1121 | state_set(&ep->com, DEAD); | 1121 | state_set(&ep->com, DEAD); |
@@ -1133,7 +1133,7 @@ static int listen_start(struct iwch_listen_ep *ep) | |||
1133 | struct sk_buff *skb; | 1133 | struct sk_buff *skb; |
1134 | struct cpl_pass_open_req *req; | 1134 | struct cpl_pass_open_req *req; |
1135 | 1135 | ||
1136 | PDBG("%s ep %p\n", __FUNCTION__, ep); | 1136 | PDBG("%s ep %p\n", __func__, ep); |
1137 | skb = get_skb(NULL, sizeof(*req), GFP_KERNEL); | 1137 | skb = get_skb(NULL, sizeof(*req), GFP_KERNEL); |
1138 | if (!skb) { | 1138 | if (!skb) { |
1139 | printk(KERN_ERR MOD "t3c_listen_start failed to alloc skb!\n"); | 1139 | printk(KERN_ERR MOD "t3c_listen_start failed to alloc skb!\n"); |
@@ -1162,7 +1162,7 @@ static int pass_open_rpl(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) | |||
1162 | struct iwch_listen_ep *ep = ctx; | 1162 | struct iwch_listen_ep *ep = ctx; |
1163 | struct cpl_pass_open_rpl *rpl = cplhdr(skb); | 1163 | struct cpl_pass_open_rpl *rpl = cplhdr(skb); |
1164 | 1164 | ||
1165 | PDBG("%s ep %p status %d error %d\n", __FUNCTION__, ep, | 1165 | PDBG("%s ep %p status %d error %d\n", __func__, ep, |
1166 | rpl->status, status2errno(rpl->status)); | 1166 | rpl->status, status2errno(rpl->status)); |
1167 | ep->com.rpl_err = status2errno(rpl->status); | 1167 | ep->com.rpl_err = status2errno(rpl->status); |
1168 | ep->com.rpl_done = 1; | 1168 | ep->com.rpl_done = 1; |
@@ -1176,10 +1176,10 @@ static int listen_stop(struct iwch_listen_ep *ep) | |||
1176 | struct sk_buff *skb; | 1176 | struct sk_buff *skb; |
1177 | struct cpl_close_listserv_req *req; | 1177 | struct cpl_close_listserv_req *req; |
1178 | 1178 | ||
1179 | PDBG("%s ep %p\n", __FUNCTION__, ep); | 1179 | PDBG("%s ep %p\n", __func__, ep); |
1180 | skb = get_skb(NULL, sizeof(*req), GFP_KERNEL); | 1180 | skb = get_skb(NULL, sizeof(*req), GFP_KERNEL); |
1181 | if (!skb) { | 1181 | if (!skb) { |
1182 | printk(KERN_ERR MOD "%s - failed to alloc skb\n", __FUNCTION__); | 1182 | printk(KERN_ERR MOD "%s - failed to alloc skb\n", __func__); |
1183 | return -ENOMEM; | 1183 | return -ENOMEM; |
1184 | } | 1184 | } |
1185 | req = (struct cpl_close_listserv_req *) skb_put(skb, sizeof(*req)); | 1185 | req = (struct cpl_close_listserv_req *) skb_put(skb, sizeof(*req)); |
@@ -1197,7 +1197,7 @@ static int close_listsrv_rpl(struct t3cdev *tdev, struct sk_buff *skb, | |||
1197 | struct iwch_listen_ep *ep = ctx; | 1197 | struct iwch_listen_ep *ep = ctx; |
1198 | struct cpl_close_listserv_rpl *rpl = cplhdr(skb); | 1198 | struct cpl_close_listserv_rpl *rpl = cplhdr(skb); |
1199 | 1199 | ||
1200 | PDBG("%s ep %p\n", __FUNCTION__, ep); | 1200 | PDBG("%s ep %p\n", __func__, ep); |
1201 | ep->com.rpl_err = status2errno(rpl->status); | 1201 | ep->com.rpl_err = status2errno(rpl->status); |
1202 | ep->com.rpl_done = 1; | 1202 | ep->com.rpl_done = 1; |
1203 | wake_up(&ep->com.waitq); | 1203 | wake_up(&ep->com.waitq); |
@@ -1211,7 +1211,7 @@ static void accept_cr(struct iwch_ep *ep, __be32 peer_ip, struct sk_buff *skb) | |||
1211 | u32 opt0h, opt0l, opt2; | 1211 | u32 opt0h, opt0l, opt2; |
1212 | int wscale; | 1212 | int wscale; |
1213 | 1213 | ||
1214 | PDBG("%s ep %p\n", __FUNCTION__, ep); | 1214 | PDBG("%s ep %p\n", __func__, ep); |
1215 | BUG_ON(skb_cloned(skb)); | 1215 | BUG_ON(skb_cloned(skb)); |
1216 | skb_trim(skb, sizeof(*rpl)); | 1216 | skb_trim(skb, sizeof(*rpl)); |
1217 | skb_get(skb); | 1217 | skb_get(skb); |
@@ -1244,7 +1244,7 @@ static void accept_cr(struct iwch_ep *ep, __be32 peer_ip, struct sk_buff *skb) | |||
1244 | static void reject_cr(struct t3cdev *tdev, u32 hwtid, __be32 peer_ip, | 1244 | static void reject_cr(struct t3cdev *tdev, u32 hwtid, __be32 peer_ip, |
1245 | struct sk_buff *skb) | 1245 | struct sk_buff *skb) |
1246 | { | 1246 | { |
1247 | PDBG("%s t3cdev %p tid %u peer_ip %x\n", __FUNCTION__, tdev, hwtid, | 1247 | PDBG("%s t3cdev %p tid %u peer_ip %x\n", __func__, tdev, hwtid, |
1248 | peer_ip); | 1248 | peer_ip); |
1249 | BUG_ON(skb_cloned(skb)); | 1249 | BUG_ON(skb_cloned(skb)); |
1250 | skb_trim(skb, sizeof(struct cpl_tid_release)); | 1250 | skb_trim(skb, sizeof(struct cpl_tid_release)); |
@@ -1279,11 +1279,11 @@ static int pass_accept_req(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) | |||
1279 | struct rtable *rt; | 1279 | struct rtable *rt; |
1280 | struct iff_mac tim; | 1280 | struct iff_mac tim; |
1281 | 1281 | ||
1282 | PDBG("%s parent ep %p tid %u\n", __FUNCTION__, parent_ep, hwtid); | 1282 | PDBG("%s parent ep %p tid %u\n", __func__, parent_ep, hwtid); |
1283 | 1283 | ||
1284 | if (state_read(&parent_ep->com) != LISTEN) { | 1284 | if (state_read(&parent_ep->com) != LISTEN) { |
1285 | printk(KERN_ERR "%s - listening ep not in LISTEN\n", | 1285 | printk(KERN_ERR "%s - listening ep not in LISTEN\n", |
1286 | __FUNCTION__); | 1286 | __func__); |
1287 | goto reject; | 1287 | goto reject; |
1288 | } | 1288 | } |
1289 | 1289 | ||
@@ -1295,7 +1295,7 @@ static int pass_accept_req(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) | |||
1295 | if (tdev->ctl(tdev, GET_IFF_FROM_MAC, &tim) < 0 || !tim.dev) { | 1295 | if (tdev->ctl(tdev, GET_IFF_FROM_MAC, &tim) < 0 || !tim.dev) { |
1296 | printk(KERN_ERR | 1296 | printk(KERN_ERR |
1297 | "%s bad dst mac %02x %02x %02x %02x %02x %02x\n", | 1297 | "%s bad dst mac %02x %02x %02x %02x %02x %02x\n", |
1298 | __FUNCTION__, | 1298 | __func__, |
1299 | req->dst_mac[0], | 1299 | req->dst_mac[0], |
1300 | req->dst_mac[1], | 1300 | req->dst_mac[1], |
1301 | req->dst_mac[2], | 1301 | req->dst_mac[2], |
@@ -1313,21 +1313,21 @@ static int pass_accept_req(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) | |||
1313 | req->peer_port, G_PASS_OPEN_TOS(ntohl(req->tos_tid))); | 1313 | req->peer_port, G_PASS_OPEN_TOS(ntohl(req->tos_tid))); |
1314 | if (!rt) { | 1314 | if (!rt) { |
1315 | printk(KERN_ERR MOD "%s - failed to find dst entry!\n", | 1315 | printk(KERN_ERR MOD "%s - failed to find dst entry!\n", |
1316 | __FUNCTION__); | 1316 | __func__); |
1317 | goto reject; | 1317 | goto reject; |
1318 | } | 1318 | } |
1319 | dst = &rt->u.dst; | 1319 | dst = &rt->u.dst; |
1320 | l2t = t3_l2t_get(tdev, dst->neighbour, dst->neighbour->dev); | 1320 | l2t = t3_l2t_get(tdev, dst->neighbour, dst->neighbour->dev); |
1321 | if (!l2t) { | 1321 | if (!l2t) { |
1322 | printk(KERN_ERR MOD "%s - failed to allocate l2t entry!\n", | 1322 | printk(KERN_ERR MOD "%s - failed to allocate l2t entry!\n", |
1323 | __FUNCTION__); | 1323 | __func__); |
1324 | dst_release(dst); | 1324 | dst_release(dst); |
1325 | goto reject; | 1325 | goto reject; |
1326 | } | 1326 | } |
1327 | child_ep = alloc_ep(sizeof(*child_ep), GFP_KERNEL); | 1327 | child_ep = alloc_ep(sizeof(*child_ep), GFP_KERNEL); |
1328 | if (!child_ep) { | 1328 | if (!child_ep) { |
1329 | printk(KERN_ERR MOD "%s - failed to allocate ep entry!\n", | 1329 | printk(KERN_ERR MOD "%s - failed to allocate ep entry!\n", |
1330 | __FUNCTION__); | 1330 | __func__); |
1331 | l2t_release(L2DATA(tdev), l2t); | 1331 | l2t_release(L2DATA(tdev), l2t); |
1332 | dst_release(dst); | 1332 | dst_release(dst); |
1333 | goto reject; | 1333 | goto reject; |
@@ -1362,7 +1362,7 @@ static int pass_establish(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) | |||
1362 | struct iwch_ep *ep = ctx; | 1362 | struct iwch_ep *ep = ctx; |
1363 | struct cpl_pass_establish *req = cplhdr(skb); | 1363 | struct cpl_pass_establish *req = cplhdr(skb); |
1364 | 1364 | ||
1365 | PDBG("%s ep %p\n", __FUNCTION__, ep); | 1365 | PDBG("%s ep %p\n", __func__, ep); |
1366 | ep->snd_seq = ntohl(req->snd_isn); | 1366 | ep->snd_seq = ntohl(req->snd_isn); |
1367 | ep->rcv_seq = ntohl(req->rcv_isn); | 1367 | ep->rcv_seq = ntohl(req->rcv_isn); |
1368 | 1368 | ||
@@ -1383,7 +1383,7 @@ static int peer_close(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) | |||
1383 | int disconnect = 1; | 1383 | int disconnect = 1; |
1384 | int release = 0; | 1384 | int release = 0; |
1385 | 1385 | ||
1386 | PDBG("%s ep %p\n", __FUNCTION__, ep); | 1386 | PDBG("%s ep %p\n", __func__, ep); |
1387 | dst_confirm(ep->dst); | 1387 | dst_confirm(ep->dst); |
1388 | 1388 | ||
1389 | spin_lock_irqsave(&ep->com.lock, flags); | 1389 | spin_lock_irqsave(&ep->com.lock, flags); |
@@ -1473,7 +1473,7 @@ static int peer_abort(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) | |||
1473 | int state; | 1473 | int state; |
1474 | 1474 | ||
1475 | if (is_neg_adv_abort(req->status)) { | 1475 | if (is_neg_adv_abort(req->status)) { |
1476 | PDBG("%s neg_adv_abort ep %p tid %d\n", __FUNCTION__, ep, | 1476 | PDBG("%s neg_adv_abort ep %p tid %d\n", __func__, ep, |
1477 | ep->hwtid); | 1477 | ep->hwtid); |
1478 | t3_l2t_send_event(ep->com.tdev, ep->l2t); | 1478 | t3_l2t_send_event(ep->com.tdev, ep->l2t); |
1479 | return CPL_RET_BUF_DONE; | 1479 | return CPL_RET_BUF_DONE; |
@@ -1489,7 +1489,7 @@ static int peer_abort(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) | |||
1489 | } | 1489 | } |
1490 | 1490 | ||
1491 | state = state_read(&ep->com); | 1491 | state = state_read(&ep->com); |
1492 | PDBG("%s ep %p state %u\n", __FUNCTION__, ep, state); | 1492 | PDBG("%s ep %p state %u\n", __func__, ep, state); |
1493 | switch (state) { | 1493 | switch (state) { |
1494 | case CONNECTING: | 1494 | case CONNECTING: |
1495 | break; | 1495 | break; |
@@ -1528,14 +1528,14 @@ static int peer_abort(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) | |||
1528 | if (ret) | 1528 | if (ret) |
1529 | printk(KERN_ERR MOD | 1529 | printk(KERN_ERR MOD |
1530 | "%s - qp <- error failed!\n", | 1530 | "%s - qp <- error failed!\n", |
1531 | __FUNCTION__); | 1531 | __func__); |
1532 | } | 1532 | } |
1533 | peer_abort_upcall(ep); | 1533 | peer_abort_upcall(ep); |
1534 | break; | 1534 | break; |
1535 | case ABORTING: | 1535 | case ABORTING: |
1536 | break; | 1536 | break; |
1537 | case DEAD: | 1537 | case DEAD: |
1538 | PDBG("%s PEER_ABORT IN DEAD STATE!!!!\n", __FUNCTION__); | 1538 | PDBG("%s PEER_ABORT IN DEAD STATE!!!!\n", __func__); |
1539 | return CPL_RET_BUF_DONE; | 1539 | return CPL_RET_BUF_DONE; |
1540 | default: | 1540 | default: |
1541 | BUG_ON(1); | 1541 | BUG_ON(1); |
@@ -1546,7 +1546,7 @@ static int peer_abort(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) | |||
1546 | rpl_skb = get_skb(skb, sizeof(*rpl), GFP_KERNEL); | 1546 | rpl_skb = get_skb(skb, sizeof(*rpl), GFP_KERNEL); |
1547 | if (!rpl_skb) { | 1547 | if (!rpl_skb) { |
1548 | printk(KERN_ERR MOD "%s - cannot allocate skb!\n", | 1548 | printk(KERN_ERR MOD "%s - cannot allocate skb!\n", |
1549 | __FUNCTION__); | 1549 | __func__); |
1550 | dst_release(ep->dst); | 1550 | dst_release(ep->dst); |
1551 | l2t_release(L2DATA(ep->com.tdev), ep->l2t); | 1551 | l2t_release(L2DATA(ep->com.tdev), ep->l2t); |
1552 | put_ep(&ep->com); | 1552 | put_ep(&ep->com); |
@@ -1573,7 +1573,7 @@ static int close_con_rpl(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) | |||
1573 | unsigned long flags; | 1573 | unsigned long flags; |
1574 | int release = 0; | 1574 | int release = 0; |
1575 | 1575 | ||
1576 | PDBG("%s ep %p\n", __FUNCTION__, ep); | 1576 | PDBG("%s ep %p\n", __func__, ep); |
1577 | BUG_ON(!ep); | 1577 | BUG_ON(!ep); |
1578 | 1578 | ||
1579 | /* The cm_id may be null if we failed to connect */ | 1579 | /* The cm_id may be null if we failed to connect */ |
@@ -1624,9 +1624,9 @@ static int terminate(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) | |||
1624 | { | 1624 | { |
1625 | struct iwch_ep *ep = ctx; | 1625 | struct iwch_ep *ep = ctx; |
1626 | 1626 | ||
1627 | PDBG("%s ep %p\n", __FUNCTION__, ep); | 1627 | PDBG("%s ep %p\n", __func__, ep); |
1628 | skb_pull(skb, sizeof(struct cpl_rdma_terminate)); | 1628 | skb_pull(skb, sizeof(struct cpl_rdma_terminate)); |
1629 | PDBG("%s saving %d bytes of term msg\n", __FUNCTION__, skb->len); | 1629 | PDBG("%s saving %d bytes of term msg\n", __func__, skb->len); |
1630 | skb_copy_from_linear_data(skb, ep->com.qp->attr.terminate_buffer, | 1630 | skb_copy_from_linear_data(skb, ep->com.qp->attr.terminate_buffer, |
1631 | skb->len); | 1631 | skb->len); |
1632 | ep->com.qp->attr.terminate_msg_len = skb->len; | 1632 | ep->com.qp->attr.terminate_msg_len = skb->len; |
@@ -1639,13 +1639,13 @@ static int ec_status(struct t3cdev *tdev, struct sk_buff *skb, void *ctx) | |||
1639 | struct cpl_rdma_ec_status *rep = cplhdr(skb); | 1639 | struct cpl_rdma_ec_status *rep = cplhdr(skb); |
1640 | struct iwch_ep *ep = ctx; | 1640 | struct iwch_ep *ep = ctx; |
1641 | 1641 | ||
1642 | PDBG("%s ep %p tid %u status %d\n", __FUNCTION__, ep, ep->hwtid, | 1642 | PDBG("%s ep %p tid %u status %d\n", __func__, ep, ep->hwtid, |
1643 | rep->status); | 1643 | rep->status); |
1644 | if (rep->status) { | 1644 | if (rep->status) { |
1645 | struct iwch_qp_attributes attrs; | 1645 | struct iwch_qp_attributes attrs; |
1646 | 1646 | ||
1647 | printk(KERN_ERR MOD "%s BAD CLOSE - Aborting tid %u\n", | 1647 | printk(KERN_ERR MOD "%s BAD CLOSE - Aborting tid %u\n", |
1648 | __FUNCTION__, ep->hwtid); | 1648 | __func__, ep->hwtid); |
1649 | stop_ep_timer(ep); | 1649 | stop_ep_timer(ep); |
1650 | attrs.next_state = IWCH_QP_STATE_ERROR; | 1650 | attrs.next_state = IWCH_QP_STATE_ERROR; |
1651 | iwch_modify_qp(ep->com.qp->rhp, | 1651 | iwch_modify_qp(ep->com.qp->rhp, |
@@ -1663,7 +1663,7 @@ static void ep_timeout(unsigned long arg) | |||
1663 | unsigned long flags; | 1663 | unsigned long flags; |
1664 | 1664 | ||
1665 | spin_lock_irqsave(&ep->com.lock, flags); | 1665 | spin_lock_irqsave(&ep->com.lock, flags); |
1666 | PDBG("%s ep %p tid %u state %d\n", __FUNCTION__, ep, ep->hwtid, | 1666 | PDBG("%s ep %p tid %u state %d\n", __func__, ep, ep->hwtid, |
1667 | ep->com.state); | 1667 | ep->com.state); |
1668 | switch (ep->com.state) { | 1668 | switch (ep->com.state) { |
1669 | case MPA_REQ_SENT: | 1669 | case MPA_REQ_SENT: |
@@ -1693,7 +1693,7 @@ int iwch_reject_cr(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len) | |||
1693 | { | 1693 | { |
1694 | int err; | 1694 | int err; |
1695 | struct iwch_ep *ep = to_ep(cm_id); | 1695 | struct iwch_ep *ep = to_ep(cm_id); |
1696 | PDBG("%s ep %p tid %u\n", __FUNCTION__, ep, ep->hwtid); | 1696 | PDBG("%s ep %p tid %u\n", __func__, ep, ep->hwtid); |
1697 | 1697 | ||
1698 | if (state_read(&ep->com) == DEAD) { | 1698 | if (state_read(&ep->com) == DEAD) { |
1699 | put_ep(&ep->com); | 1699 | put_ep(&ep->com); |
@@ -1718,7 +1718,7 @@ int iwch_accept_cr(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) | |||
1718 | struct iwch_dev *h = to_iwch_dev(cm_id->device); | 1718 | struct iwch_dev *h = to_iwch_dev(cm_id->device); |
1719 | struct iwch_qp *qp = get_qhp(h, conn_param->qpn); | 1719 | struct iwch_qp *qp = get_qhp(h, conn_param->qpn); |
1720 | 1720 | ||
1721 | PDBG("%s ep %p tid %u\n", __FUNCTION__, ep, ep->hwtid); | 1721 | PDBG("%s ep %p tid %u\n", __func__, ep, ep->hwtid); |
1722 | if (state_read(&ep->com) == DEAD) | 1722 | if (state_read(&ep->com) == DEAD) |
1723 | return -ECONNRESET; | 1723 | return -ECONNRESET; |
1724 | 1724 | ||
@@ -1739,7 +1739,7 @@ int iwch_accept_cr(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) | |||
1739 | ep->com.rpl_err = 0; | 1739 | ep->com.rpl_err = 0; |
1740 | ep->ird = conn_param->ird; | 1740 | ep->ird = conn_param->ird; |
1741 | ep->ord = conn_param->ord; | 1741 | ep->ord = conn_param->ord; |
1742 | PDBG("%s %d ird %d ord %d\n", __FUNCTION__, __LINE__, ep->ird, ep->ord); | 1742 | PDBG("%s %d ird %d ord %d\n", __func__, __LINE__, ep->ird, ep->ord); |
1743 | 1743 | ||
1744 | get_ep(&ep->com); | 1744 | get_ep(&ep->com); |
1745 | 1745 | ||
@@ -1810,7 +1810,7 @@ int iwch_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) | |||
1810 | 1810 | ||
1811 | ep = alloc_ep(sizeof(*ep), GFP_KERNEL); | 1811 | ep = alloc_ep(sizeof(*ep), GFP_KERNEL); |
1812 | if (!ep) { | 1812 | if (!ep) { |
1813 | printk(KERN_ERR MOD "%s - cannot alloc ep.\n", __FUNCTION__); | 1813 | printk(KERN_ERR MOD "%s - cannot alloc ep.\n", __func__); |
1814 | err = -ENOMEM; | 1814 | err = -ENOMEM; |
1815 | goto out; | 1815 | goto out; |
1816 | } | 1816 | } |
@@ -1827,7 +1827,7 @@ int iwch_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) | |||
1827 | ep->com.cm_id = cm_id; | 1827 | ep->com.cm_id = cm_id; |
1828 | ep->com.qp = get_qhp(h, conn_param->qpn); | 1828 | ep->com.qp = get_qhp(h, conn_param->qpn); |
1829 | BUG_ON(!ep->com.qp); | 1829 | BUG_ON(!ep->com.qp); |
1830 | PDBG("%s qpn 0x%x qp %p cm_id %p\n", __FUNCTION__, conn_param->qpn, | 1830 | PDBG("%s qpn 0x%x qp %p cm_id %p\n", __func__, conn_param->qpn, |
1831 | ep->com.qp, cm_id); | 1831 | ep->com.qp, cm_id); |
1832 | 1832 | ||
1833 | /* | 1833 | /* |
@@ -1835,7 +1835,7 @@ int iwch_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) | |||
1835 | */ | 1835 | */ |
1836 | ep->atid = cxgb3_alloc_atid(h->rdev.t3cdev_p, &t3c_client, ep); | 1836 | ep->atid = cxgb3_alloc_atid(h->rdev.t3cdev_p, &t3c_client, ep); |
1837 | if (ep->atid == -1) { | 1837 | if (ep->atid == -1) { |
1838 | printk(KERN_ERR MOD "%s - cannot alloc atid.\n", __FUNCTION__); | 1838 | printk(KERN_ERR MOD "%s - cannot alloc atid.\n", __func__); |
1839 | err = -ENOMEM; | 1839 | err = -ENOMEM; |
1840 | goto fail2; | 1840 | goto fail2; |
1841 | } | 1841 | } |
@@ -1847,7 +1847,7 @@ int iwch_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) | |||
1847 | cm_id->local_addr.sin_port, | 1847 | cm_id->local_addr.sin_port, |
1848 | cm_id->remote_addr.sin_port, IPTOS_LOWDELAY); | 1848 | cm_id->remote_addr.sin_port, IPTOS_LOWDELAY); |
1849 | if (!rt) { | 1849 | if (!rt) { |
1850 | printk(KERN_ERR MOD "%s - cannot find route.\n", __FUNCTION__); | 1850 | printk(KERN_ERR MOD "%s - cannot find route.\n", __func__); |
1851 | err = -EHOSTUNREACH; | 1851 | err = -EHOSTUNREACH; |
1852 | goto fail3; | 1852 | goto fail3; |
1853 | } | 1853 | } |
@@ -1857,7 +1857,7 @@ int iwch_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param) | |||
1857 | ep->l2t = t3_l2t_get(ep->com.tdev, ep->dst->neighbour, | 1857 | ep->l2t = t3_l2t_get(ep->com.tdev, ep->dst->neighbour, |
1858 | ep->dst->neighbour->dev); | 1858 | ep->dst->neighbour->dev); |
1859 | if (!ep->l2t) { | 1859 | if (!ep->l2t) { |
1860 | printk(KERN_ERR MOD "%s - cannot alloc l2e.\n", __FUNCTION__); | 1860 | printk(KERN_ERR MOD "%s - cannot alloc l2e.\n", __func__); |
1861 | err = -ENOMEM; | 1861 | err = -ENOMEM; |
1862 | goto fail4; | 1862 | goto fail4; |
1863 | } | 1863 | } |
@@ -1894,11 +1894,11 @@ int iwch_create_listen(struct iw_cm_id *cm_id, int backlog) | |||
1894 | 1894 | ||
1895 | ep = alloc_ep(sizeof(*ep), GFP_KERNEL); | 1895 | ep = alloc_ep(sizeof(*ep), GFP_KERNEL); |
1896 | if (!ep) { | 1896 | if (!ep) { |
1897 | printk(KERN_ERR MOD "%s - cannot alloc ep.\n", __FUNCTION__); | 1897 | printk(KERN_ERR MOD "%s - cannot alloc ep.\n", __func__); |
1898 | err = -ENOMEM; | 1898 | err = -ENOMEM; |
1899 | goto fail1; | 1899 | goto fail1; |
1900 | } | 1900 | } |
1901 | PDBG("%s ep %p\n", __FUNCTION__, ep); | 1901 | PDBG("%s ep %p\n", __func__, ep); |
1902 | ep->com.tdev = h->rdev.t3cdev_p; | 1902 | ep->com.tdev = h->rdev.t3cdev_p; |
1903 | cm_id->add_ref(cm_id); | 1903 | cm_id->add_ref(cm_id); |
1904 | ep->com.cm_id = cm_id; | 1904 | ep->com.cm_id = cm_id; |
@@ -1910,7 +1910,7 @@ int iwch_create_listen(struct iw_cm_id *cm_id, int backlog) | |||
1910 | */ | 1910 | */ |
1911 | ep->stid = cxgb3_alloc_stid(h->rdev.t3cdev_p, &t3c_client, ep); | 1911 | ep->stid = cxgb3_alloc_stid(h->rdev.t3cdev_p, &t3c_client, ep); |
1912 | if (ep->stid == -1) { | 1912 | if (ep->stid == -1) { |
1913 | printk(KERN_ERR MOD "%s - cannot alloc atid.\n", __FUNCTION__); | 1913 | printk(KERN_ERR MOD "%s - cannot alloc atid.\n", __func__); |
1914 | err = -ENOMEM; | 1914 | err = -ENOMEM; |
1915 | goto fail2; | 1915 | goto fail2; |
1916 | } | 1916 | } |
@@ -1942,7 +1942,7 @@ int iwch_destroy_listen(struct iw_cm_id *cm_id) | |||
1942 | int err; | 1942 | int err; |
1943 | struct iwch_listen_ep *ep = to_listen_ep(cm_id); | 1943 | struct iwch_listen_ep *ep = to_listen_ep(cm_id); |
1944 | 1944 | ||
1945 | PDBG("%s ep %p\n", __FUNCTION__, ep); | 1945 | PDBG("%s ep %p\n", __func__, ep); |
1946 | 1946 | ||
1947 | might_sleep(); | 1947 | might_sleep(); |
1948 | state_set(&ep->com, DEAD); | 1948 | state_set(&ep->com, DEAD); |
@@ -1965,11 +1965,11 @@ int iwch_ep_disconnect(struct iwch_ep *ep, int abrupt, gfp_t gfp) | |||
1965 | 1965 | ||
1966 | spin_lock_irqsave(&ep->com.lock, flags); | 1966 | spin_lock_irqsave(&ep->com.lock, flags); |
1967 | 1967 | ||
1968 | PDBG("%s ep %p state %s, abrupt %d\n", __FUNCTION__, ep, | 1968 | PDBG("%s ep %p state %s, abrupt %d\n", __func__, ep, |
1969 | states[ep->com.state], abrupt); | 1969 | states[ep->com.state], abrupt); |
1970 | 1970 | ||
1971 | if (ep->com.state == DEAD) { | 1971 | if (ep->com.state == DEAD) { |
1972 | PDBG("%s already dead ep %p\n", __FUNCTION__, ep); | 1972 | PDBG("%s already dead ep %p\n", __func__, ep); |
1973 | goto out; | 1973 | goto out; |
1974 | } | 1974 | } |
1975 | 1975 | ||
@@ -2020,7 +2020,7 @@ int iwch_ep_redirect(void *ctx, struct dst_entry *old, struct dst_entry *new, | |||
2020 | if (ep->dst != old) | 2020 | if (ep->dst != old) |
2021 | return 0; | 2021 | return 0; |
2022 | 2022 | ||
2023 | PDBG("%s ep %p redirect to dst %p l2t %p\n", __FUNCTION__, ep, new, | 2023 | PDBG("%s ep %p redirect to dst %p l2t %p\n", __func__, ep, new, |
2024 | l2t); | 2024 | l2t); |
2025 | dst_hold(new); | 2025 | dst_hold(new); |
2026 | l2t_release(L2DATA(ep->com.tdev), ep->l2t); | 2026 | l2t_release(L2DATA(ep->com.tdev), ep->l2t); |