aboutsummaryrefslogtreecommitdiffstats
path: root/drivers/infiniband/hw/cxgb3/iwch_cm.c
diff options
context:
space:
mode:
authorHarvey Harrison <harvey.harrison@gmail.com>2008-04-17 00:01:10 -0400
committerRoland Dreier <rolandd@cisco.com>2008-04-17 00:01:10 -0400
commit3371836383d63b627b228875f5ac63023cbf11d2 (patch)
treeb997894d9774bdc07f7df76ceca48e6a848760c8 /drivers/infiniband/hw/cxgb3/iwch_cm.c
parente8e91f6b4dc1179a70b0d21241b769c0ebfaa129 (diff)
IB: Replace remaining __FUNCTION__ occurrences with __func__
__FUNCTION__ is gcc-specific, use __func__ instead. Signed-off-by: Harvey Harrison <harvey.harrison@gmail.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Roland Dreier <rolandd@cisco.com>
Diffstat (limited to 'drivers/infiniband/hw/cxgb3/iwch_cm.c')
-rw-r--r--drivers/infiniband/hw/cxgb3/iwch_cm.c166
1 files changed, 83 insertions, 83 deletions
diff --git a/drivers/infiniband/hw/cxgb3/iwch_cm.c b/drivers/infiniband/hw/cxgb3/iwch_cm.c
index 99f2f2a46bf..72ca360c3db 100644
--- a/drivers/infiniband/hw/cxgb3/iwch_cm.c
+++ b/drivers/infiniband/hw/cxgb3/iwch_cm.c
@@ -110,9 +110,9 @@ static void connect_reply_upcall(struct iwch_ep *ep, int status);
110 110
111static void start_ep_timer(struct iwch_ep *ep) 111static void start_ep_timer(struct iwch_ep *ep)
112{ 112{
113 PDBG("%s ep %p\n", __FUNCTION__, ep); 113 PDBG("%s ep %p\n", __func__, ep);
114 if (timer_pending(&ep->timer)) { 114 if (timer_pending(&ep->timer)) {
115 PDBG("%s stopped / restarted timer ep %p\n", __FUNCTION__, ep); 115 PDBG("%s stopped / restarted timer ep %p\n", __func__, ep);
116 del_timer_sync(&ep->timer); 116 del_timer_sync(&ep->timer);
117 } else 117 } else
118 get_ep(&ep->com); 118 get_ep(&ep->com);
@@ -124,7 +124,7 @@ static void start_ep_timer(struct iwch_ep *ep)
124 124
125static void stop_ep_timer(struct iwch_ep *ep) 125static void stop_ep_timer(struct iwch_ep *ep)
126{ 126{
127 PDBG("%s ep %p\n", __FUNCTION__, ep); 127 PDBG("%s ep %p\n", __func__, ep);
128 del_timer_sync(&ep->timer); 128 del_timer_sync(&ep->timer);
129 put_ep(&ep->com); 129 put_ep(&ep->com);
130} 130}
@@ -190,7 +190,7 @@ int iwch_resume_tid(struct iwch_ep *ep)
190 190
191static void set_emss(struct iwch_ep *ep, u16 opt) 191static void set_emss(struct iwch_ep *ep, u16 opt)
192{ 192{
193 PDBG("%s ep %p opt %u\n", __FUNCTION__, ep, opt); 193 PDBG("%s ep %p opt %u\n", __func__, ep, opt);
194 ep->emss = T3C_DATA(ep->com.tdev)->mtus[G_TCPOPT_MSS(opt)] - 40; 194 ep->emss = T3C_DATA(ep->com.tdev)->mtus[G_TCPOPT_MSS(opt)] - 40;
195 if (G_TCPOPT_TSTAMP(opt)) 195 if (G_TCPOPT_TSTAMP(opt))
196 ep->emss -= 12; 196 ep->emss -= 12;
@@ -220,7 +220,7 @@ static void state_set(struct iwch_ep_common *epc, enum iwch_ep_state new)
220 unsigned long flags; 220 unsigned long flags;
221 221
222 spin_lock_irqsave(&epc->lock, flags); 222 spin_lock_irqsave(&epc->lock, flags);
223 PDBG("%s - %s -> %s\n", __FUNCTION__, states[epc->state], states[new]); 223 PDBG("%s - %s -> %s\n", __func__, states[epc->state], states[new]);
224 __state_set(epc, new); 224 __state_set(epc, new);
225 spin_unlock_irqrestore(&epc->lock, flags); 225 spin_unlock_irqrestore(&epc->lock, flags);
226 return; 226 return;
@@ -236,7 +236,7 @@ static void *alloc_ep(int size, gfp_t gfp)
236 spin_lock_init(&epc->lock); 236 spin_lock_init(&epc->lock);
237 init_waitqueue_head(&epc->waitq); 237 init_waitqueue_head(&epc->waitq);
238 } 238 }
239 PDBG("%s alloc ep %p\n", __FUNCTION__, epc); 239 PDBG("%s alloc ep %p\n", __func__, epc);
240 return epc; 240 return epc;
241} 241}
242 242
@@ -244,13 +244,13 @@ void __free_ep(struct kref *kref)
244{ 244{
245 struct iwch_ep_common *epc; 245 struct iwch_ep_common *epc;
246 epc = container_of(kref, struct iwch_ep_common, kref); 246 epc = container_of(kref, struct iwch_ep_common, kref);
247 PDBG("%s ep %p state %s\n", __FUNCTION__, epc, states[state_read(epc)]); 247 PDBG("%s ep %p state %s\n", __func__, epc, states[state_read(epc)]);
248 kfree(epc); 248 kfree(epc);
249} 249}
250 250
251static void release_ep_resources(struct iwch_ep *ep) 251static void release_ep_resources(struct iwch_ep *ep)
252{ 252{
253 PDBG("%s ep %p tid %d\n", __FUNCTION__, ep, ep->hwtid); 253 PDBG("%s ep %p tid %d\n", __func__, ep, ep->hwtid);
254 cxgb3_remove_tid(ep->com.tdev, (void *)ep, ep->hwtid); 254 cxgb3_remove_tid(ep->com.tdev, (void *)ep, ep->hwtid);
255 dst_release(ep->dst); 255 dst_release(ep->dst);
256 l2t_release(L2DATA(ep->com.tdev), ep->l2t); 256 l2t_release(L2DATA(ep->com.tdev), ep->l2t);
@@ -349,7 +349,7 @@ static unsigned int find_best_mtu(const struct t3c_data *d, unsigned short mtu)
349 349
350static void arp_failure_discard(struct t3cdev *dev, struct sk_buff *skb) 350static void arp_failure_discard(struct t3cdev *dev, struct sk_buff *skb)
351{ 351{
352 PDBG("%s t3cdev %p\n", __FUNCTION__, dev); 352 PDBG("%s t3cdev %p\n", __func__, dev);
353 kfree_skb(skb); 353 kfree_skb(skb);
354} 354}
355 355
@@ -370,7 +370,7 @@ static void abort_arp_failure(struct t3cdev *dev, struct sk_buff *skb)
370{ 370{
371 struct cpl_abort_req *req = cplhdr(skb); 371 struct cpl_abort_req *req = cplhdr(skb);
372 372
373 PDBG("%s t3cdev %p\n", __FUNCTION__, dev); 373 PDBG("%s t3cdev %p\n", __func__, dev);
374 req->cmd = CPL_ABORT_NO_RST; 374 req->cmd = CPL_ABORT_NO_RST;
375 cxgb3_ofld_send(dev, skb); 375 cxgb3_ofld_send(dev, skb);
376} 376}
@@ -380,10 +380,10 @@ static int send_halfclose(struct iwch_ep *ep, gfp_t gfp)
380 struct cpl_close_con_req *req; 380 struct cpl_close_con_req *req;
381 struct sk_buff *skb; 381 struct sk_buff *skb;
382 382
383 PDBG("%s ep %p\n", __FUNCTION__, ep); 383 PDBG("%s ep %p\n", __func__, ep);
384 skb = get_skb(NULL, sizeof(*req), gfp); 384 skb = get_skb(NULL, sizeof(*req), gfp);
385 if (!skb) { 385 if (!skb) {
386 printk(KERN_ERR MOD "%s - failed to alloc skb\n", __FUNCTION__); 386 printk(KERN_ERR MOD "%s - failed to alloc skb\n", __func__);
387 return -ENOMEM; 387 return -ENOMEM;
388 } 388 }
389 skb->priority = CPL_PRIORITY_DATA; 389 skb->priority = CPL_PRIORITY_DATA;
@@ -400,11 +400,11 @@ static int send_abort(struct iwch_ep *ep, struct sk_buff *skb, gfp_t gfp)
400{ 400{
401 struct cpl_abort_req *req; 401 struct cpl_abort_req *req;
402 402
403 PDBG("%s ep %p\n", __FUNCTION__, ep); 403 PDBG("%s ep %p\n", __func__, ep);
404 skb = get_skb(skb, sizeof(*req), gfp); 404 skb = get_skb(skb, sizeof(*req), gfp);
405 if (!skb) { 405 if (!skb) {
406 printk(KERN_ERR MOD "%s - failed to alloc skb.\n", 406 printk(KERN_ERR MOD "%s - failed to alloc skb.\n",
407 __FUNCTION__); 407 __func__);
408 return -ENOMEM; 408 return -ENOMEM;
409 } 409 }
410 skb->priority = CPL_PRIORITY_DATA; 410 skb->priority = CPL_PRIORITY_DATA;
@@ -426,12 +426,12 @@ static int send_connect(struct iwch_ep *ep)
426 unsigned int mtu_idx; 426 unsigned int mtu_idx;
427 int wscale; 427 int wscale;
428 428
429 PDBG("%s ep %p\n", __FUNCTION__, ep); 429 PDBG("%s ep %p\n", __func__, ep);
430 430
431 skb = get_skb(NULL, sizeof(*req), GFP_KERNEL); 431 skb = get_skb(NULL, sizeof(*req), GFP_KERNEL);
432 if (!skb) { 432 if (!skb) {
433 printk(KERN_ERR MOD "%s - failed to alloc skb.\n", 433 printk(KERN_ERR MOD "%s - failed to alloc skb.\n",
434 __FUNCTION__); 434 __func__);
435 return -ENOMEM; 435 return -ENOMEM;
436 } 436 }
437 mtu_idx = find_best_mtu(T3C_DATA(ep->com.tdev), dst_mtu(ep->dst)); 437 mtu_idx = find_best_mtu(T3C_DATA(ep->com.tdev), dst_mtu(ep->dst));
@@ -470,7 +470,7 @@ static void send_mpa_req(struct iwch_ep *ep, struct sk_buff *skb)
470 struct mpa_message *mpa; 470 struct mpa_message *mpa;
471 int len; 471 int len;
472 472
473 PDBG("%s ep %p pd_len %d\n", __FUNCTION__, ep, ep->plen); 473 PDBG("%s ep %p pd_len %d\n", __func__, ep, ep->plen);
474 474
475 BUG_ON(skb_cloned(skb)); 475 BUG_ON(skb_cloned(skb));
476 476
@@ -530,13 +530,13 @@ static int send_mpa_reject(struct iwch_ep *ep, const void *pdata, u8 plen)
530 struct mpa_message *mpa; 530 struct mpa_message *mpa;
531 struct sk_buff *skb; 531 struct sk_buff *skb;
532 532
533 PDBG("%s ep %p plen %d\n", __FUNCTION__, ep, plen); 533 PDBG("%s ep %p plen %d\n", __func__, ep, plen);
534 534
535 mpalen = sizeof(*mpa) + plen; 535 mpalen = sizeof(*mpa) + plen;
536 536
537 skb = get_skb(NULL, mpalen + sizeof(*req), GFP_KERNEL); 537 skb = get_skb(NULL, mpalen + sizeof(*req), GFP_KERNEL);
538 if (!skb) { 538 if (!skb) {
539 printk(KERN_ERR MOD "%s - cannot alloc skb!\n", __FUNCTION__); 539 printk(KERN_ERR MOD "%s - cannot alloc skb!\n", __func__);
540 return -ENOMEM; 540 return -ENOMEM;
541 } 541 }
542 skb_reserve(skb, sizeof(*req)); 542 skb_reserve(skb, sizeof(*req));
@@ -580,13 +580,13 @@ static int send_mpa_reply(struct iwch_ep *ep, const void *pdata, u8 plen)
580 int len; 580 int len;
581 struct sk_buff *skb; 581 struct sk_buff *skb;
582 582
583 PDBG("%s ep %p plen %d\n", __FUNCTION__, ep, plen); 583 PDBG("%s ep %p plen %d\n", __func__, ep, plen);
584 584
585 mpalen = sizeof(*mpa) + plen; 585 mpalen = sizeof(*mpa) + plen;
586 586
587 skb = get_skb(NULL, mpalen + sizeof(*req), GFP_KERNEL); 587 skb = get_skb(NULL, mpalen + sizeof(*req), GFP_KERNEL);
588 if (!skb) { 588 if (!skb) {
589 printk(KERN_ERR MOD "%s - cannot alloc skb!\n", __FUNCTION__); 589 printk(KERN_ERR MOD "%s - cannot alloc skb!\n", __func__);
590 return -ENOMEM; 590 return -ENOMEM;
591 } 591 }
592 skb->priority = CPL_PRIORITY_DATA; 592 skb->priority = CPL_PRIORITY_DATA;
@@ -630,7 +630,7 @@ static int act_establish(struct t3cdev *tdev, struct sk_buff *skb, void *ctx)
630 struct cpl_act_establish *req = cplhdr(skb); 630 struct cpl_act_establish *req = cplhdr(skb);
631 unsigned int tid = GET_TID(req); 631 unsigned int tid = GET_TID(req);
632 632
633 PDBG("%s ep %p tid %d\n", __FUNCTION__, ep, tid); 633 PDBG("%s ep %p tid %d\n", __func__, ep, tid);
634 634
635 dst_confirm(ep->dst); 635 dst_confirm(ep->dst);
636 636
@@ -663,7 +663,7 @@ static void close_complete_upcall(struct iwch_ep *ep)
663{ 663{
664 struct iw_cm_event event; 664 struct iw_cm_event event;
665 665
666 PDBG("%s ep %p\n", __FUNCTION__, ep); 666 PDBG("%s ep %p\n", __func__, ep);
667 memset(&event, 0, sizeof(event)); 667 memset(&event, 0, sizeof(event));
668 event.event = IW_CM_EVENT_CLOSE; 668 event.event = IW_CM_EVENT_CLOSE;
669 if (ep->com.cm_id) { 669 if (ep->com.cm_id) {
@@ -680,7 +680,7 @@ static void peer_close_upcall(struct iwch_ep *ep)
680{ 680{
681 struct iw_cm_event event; 681 struct iw_cm_event event;
682 682
683 PDBG("%s ep %p\n", __FUNCTION__, ep); 683 PDBG("%s ep %p\n", __func__, ep);
684 memset(&event, 0, sizeof(event)); 684 memset(&event, 0, sizeof(event));
685 event.event = IW_CM_EVENT_DISCONNECT; 685 event.event = IW_CM_EVENT_DISCONNECT;
686 if (ep->com.cm_id) { 686 if (ep->com.cm_id) {
@@ -694,7 +694,7 @@ static void peer_abort_upcall(struct iwch_ep *ep)
694{ 694{
695 struct iw_cm_event event; 695 struct iw_cm_event event;
696 696
697 PDBG("%s ep %p\n", __FUNCTION__, ep); 697 PDBG("%s ep %p\n", __func__, ep);
698 memset(&event, 0, sizeof(event)); 698 memset(&event, 0, sizeof(event));
699 event.event = IW_CM_EVENT_CLOSE; 699 event.event = IW_CM_EVENT_CLOSE;
700 event.status = -ECONNRESET; 700 event.status = -ECONNRESET;
@@ -712,7 +712,7 @@ static void connect_reply_upcall(struct iwch_ep *ep, int status)
712{ 712{
713 struct iw_cm_event event; 713 struct iw_cm_event event;
714 714
715 PDBG("%s ep %p status %d\n", __FUNCTION__, ep, status); 715 PDBG("%s ep %p status %d\n", __func__, ep, status);
716 memset(&event, 0, sizeof(event)); 716 memset(&event, 0, sizeof(event));
717 event.event = IW_CM_EVENT_CONNECT_REPLY; 717 event.event = IW_CM_EVENT_CONNECT_REPLY;
718 event.status = status; 718 event.status = status;
@@ -724,7 +724,7 @@ static void connect_reply_upcall(struct iwch_ep *ep, int status)
724 event.private_data = ep->mpa_pkt + sizeof(struct mpa_message); 724 event.private_data = ep->mpa_pkt + sizeof(struct mpa_message);
725 } 725 }
726 if (ep->com.cm_id) { 726 if (ep->com.cm_id) {
727 PDBG("%s ep %p tid %d status %d\n", __FUNCTION__, ep, 727 PDBG("%s ep %p tid %d status %d\n", __func__, ep,
728 ep->hwtid, status); 728 ep->hwtid, status);
729 ep->com.cm_id->event_handler(ep->com.cm_id, &event); 729 ep->com.cm_id->event_handler(ep->com.cm_id, &event);
730 } 730 }
@@ -739,7 +739,7 @@ static void connect_request_upcall(struct iwch_ep *ep)
739{ 739{
740 struct iw_cm_event event; 740 struct iw_cm_event event;
741 741
742 PDBG("%s ep %p tid %d\n", __FUNCTION__, ep, ep->hwtid); 742 PDBG("%s ep %p tid %d\n", __func__, ep, ep->hwtid);
743 memset(&event, 0, sizeof(event)); 743 memset(&event, 0, sizeof(event));
744 event.event = IW_CM_EVENT_CONNECT_REQUEST; 744 event.event = IW_CM_EVENT_CONNECT_REQUEST;
745 event.local_addr = ep->com.local_addr; 745 event.local_addr = ep->com.local_addr;
@@ -759,11 +759,11 @@ static void established_upcall(struct iwch_ep *ep)
759{ 759{
760 struct iw_cm_event event; 760 struct iw_cm_event event;
761 761
762 PDBG("%s ep %p\n", __FUNCTION__, ep); 762 PDBG("%s ep %p\n", __func__, ep);
763 memset(&event, 0, sizeof(event)); 763 memset(&event, 0, sizeof(event));
764 event.event = IW_CM_EVENT_ESTABLISHED; 764 event.event = IW_CM_EVENT_ESTABLISHED;
765 if (ep->com.cm_id) { 765 if (ep->com.cm_id) {
766 PDBG("%s ep %p tid %d\n", __FUNCTION__, ep, ep->hwtid); 766 PDBG("%s ep %p tid %d\n", __func__, ep, ep->hwtid);
767 ep->com.cm_id->event_handler(ep->com.cm_id, &event); 767 ep->com.cm_id->event_handler(ep->com.cm_id, &event);
768 } 768 }
769} 769}
@@ -773,7 +773,7 @@ static int update_rx_credits(struct iwch_ep *ep, u32 credits)
773 struct cpl_rx_data_ack *req; 773 struct cpl_rx_data_ack *req;
774 struct sk_buff *skb; 774 struct sk_buff *skb;
775 775
776 PDBG("%s ep %p credits %u\n", __FUNCTION__, ep, credits); 776 PDBG("%s ep %p credits %u\n", __func__, ep, credits);
777 skb = get_skb(NULL, sizeof(*req), GFP_KERNEL); 777 skb = get_skb(NULL, sizeof(*req), GFP_KERNEL);
778 if (!skb) { 778 if (!skb) {
779 printk(KERN_ERR MOD "update_rx_credits - cannot alloc skb!\n"); 779 printk(KERN_ERR MOD "update_rx_credits - cannot alloc skb!\n");
@@ -797,7 +797,7 @@ static void process_mpa_reply(struct iwch_ep *ep, struct sk_buff *skb)
797 enum iwch_qp_attr_mask mask; 797 enum iwch_qp_attr_mask mask;
798 int err; 798 int err;
799 799
800 PDBG("%s ep %p\n", __FUNCTION__, ep); 800 PDBG("%s ep %p\n", __func__, ep);
801 801
802 /* 802 /*
803 * Stop mpa timer. If it expired, then the state has 803 * Stop mpa timer. If it expired, then the state has
@@ -884,7 +884,7 @@ static void process_mpa_reply(struct iwch_ep *ep, struct sk_buff *skb)
884 ep->mpa_attr.xmit_marker_enabled = mpa->flags & MPA_MARKERS ? 1 : 0; 884 ep->mpa_attr.xmit_marker_enabled = mpa->flags & MPA_MARKERS ? 1 : 0;
885 ep->mpa_attr.version = mpa_rev; 885 ep->mpa_attr.version = mpa_rev;
886 PDBG("%s - crc_enabled=%d, recv_marker_enabled=%d, " 886 PDBG("%s - crc_enabled=%d, recv_marker_enabled=%d, "
887 "xmit_marker_enabled=%d, version=%d\n", __FUNCTION__, 887 "xmit_marker_enabled=%d, version=%d\n", __func__,
888 ep->mpa_attr.crc_enabled, ep->mpa_attr.recv_marker_enabled, 888 ep->mpa_attr.crc_enabled, ep->mpa_attr.recv_marker_enabled,
889 ep->mpa_attr.xmit_marker_enabled, ep->mpa_attr.version); 889 ep->mpa_attr.xmit_marker_enabled, ep->mpa_attr.version);
890 890
@@ -915,7 +915,7 @@ static void process_mpa_request(struct iwch_ep *ep, struct sk_buff *skb)
915 struct mpa_message *mpa; 915 struct mpa_message *mpa;
916 u16 plen; 916 u16 plen;
917 917
918 PDBG("%s ep %p\n", __FUNCTION__, ep); 918 PDBG("%s ep %p\n", __func__, ep);
919 919
920 /* 920 /*
921 * Stop mpa timer. If it expired, then the state has 921 * Stop mpa timer. If it expired, then the state has
@@ -935,7 +935,7 @@ static void process_mpa_request(struct iwch_ep *ep, struct sk_buff *skb)
935 return; 935 return;
936 } 936 }
937 937
938 PDBG("%s enter (%s line %u)\n", __FUNCTION__, __FILE__, __LINE__); 938 PDBG("%s enter (%s line %u)\n", __func__, __FILE__, __LINE__);
939 939
940 /* 940 /*
941 * Copy the new data into our accumulation buffer. 941 * Copy the new data into our accumulation buffer.
@@ -950,7 +950,7 @@ static void process_mpa_request(struct iwch_ep *ep, struct sk_buff *skb)
950 */ 950 */
951 if (ep->mpa_pkt_len < sizeof(*mpa)) 951 if (ep->mpa_pkt_len < sizeof(*mpa))
952 return; 952 return;
953 PDBG("%s enter (%s line %u)\n", __FUNCTION__, __FILE__, __LINE__); 953 PDBG("%s enter (%s line %u)\n", __func__, __FILE__, __LINE__);
954 mpa = (struct mpa_message *) ep->mpa_pkt; 954 mpa = (struct mpa_message *) ep->mpa_pkt;
955 955
956 /* 956 /*
@@ -1000,7 +1000,7 @@ static void process_mpa_request(struct iwch_ep *ep, struct sk_buff *skb)
1000 ep->mpa_attr.xmit_marker_enabled = mpa->flags & MPA_MARKERS ? 1 : 0; 1000 ep->mpa_attr.xmit_marker_enabled = mpa->flags & MPA_MARKERS ? 1 : 0;
1001 ep->mpa_attr.version = mpa_rev; 1001 ep->mpa_attr.version = mpa_rev;
1002 PDBG("%s - crc_enabled=%d, recv_marker_enabled=%d, " 1002 PDBG("%s - crc_enabled=%d, recv_marker_enabled=%d, "
1003 "xmit_marker_enabled=%d, version=%d\n", __FUNCTION__, 1003 "xmit_marker_enabled=%d, version=%d\n", __func__,
1004 ep->mpa_attr.crc_enabled, ep->mpa_attr.recv_marker_enabled, 1004 ep->mpa_attr.crc_enabled, ep->mpa_attr.recv_marker_enabled,
1005 ep->mpa_attr.xmit_marker_enabled, ep->mpa_attr.version); 1005 ep->mpa_attr.xmit_marker_enabled, ep->mpa_attr.version);
1006 1006
@@ -1017,7 +1017,7 @@ static int rx_data(struct t3cdev *tdev, struct sk_buff *skb, void *ctx)
1017 struct cpl_rx_data *hdr = cplhdr(skb); 1017 struct cpl_rx_data *hdr = cplhdr(skb);
1018 unsigned int dlen = ntohs(hdr->len); 1018 unsigned int dlen = ntohs(hdr->len);
1019 1019
1020 PDBG("%s ep %p dlen %u\n", __FUNCTION__, ep, dlen); 1020 PDBG("%s ep %p dlen %u\n", __func__, ep, dlen);
1021 1021
1022 skb_pull(skb, sizeof(*hdr)); 1022 skb_pull(skb, sizeof(*hdr));
1023 skb_trim(skb, dlen); 1023 skb_trim(skb, dlen);
@@ -1037,7 +1037,7 @@ static int rx_data(struct t3cdev *tdev, struct sk_buff *skb, void *ctx)
1037 default: 1037 default:
1038 printk(KERN_ERR MOD "%s Unexpected streaming data." 1038 printk(KERN_ERR MOD "%s Unexpected streaming data."
1039 " ep %p state %d tid %d\n", 1039 " ep %p state %d tid %d\n",
1040 __FUNCTION__, ep, state_read(&ep->com), ep->hwtid); 1040 __func__, ep, state_read(&ep->com), ep->hwtid);
1041 1041
1042 /* 1042 /*
1043 * The ep will timeout and inform the ULP of the failure. 1043 * The ep will timeout and inform the ULP of the failure.
@@ -1063,7 +1063,7 @@ static int tx_ack(struct t3cdev *tdev, struct sk_buff *skb, void *ctx)
1063 struct cpl_wr_ack *hdr = cplhdr(skb); 1063 struct cpl_wr_ack *hdr = cplhdr(skb);
1064 unsigned int credits = ntohs(hdr->credits); 1064 unsigned int credits = ntohs(hdr->credits);
1065 1065
1066 PDBG("%s ep %p credits %u\n", __FUNCTION__, ep, credits); 1066 PDBG("%s ep %p credits %u\n", __func__, ep, credits);
1067 1067
1068 if (credits == 0) 1068 if (credits == 0)
1069 return CPL_RET_BUF_DONE; 1069 return CPL_RET_BUF_DONE;
@@ -1084,7 +1084,7 @@ static int abort_rpl(struct t3cdev *tdev, struct sk_buff *skb, void *ctx)
1084{ 1084{
1085 struct iwch_ep *ep = ctx; 1085 struct iwch_ep *ep = ctx;
1086 1086
1087 PDBG("%s ep %p\n", __FUNCTION__, ep); 1087 PDBG("%s ep %p\n", __func__, ep);
1088 1088
1089 /* 1089 /*
1090 * We get 2 abort replies from the HW. The first one must 1090 * We get 2 abort replies from the HW. The first one must
@@ -1115,7 +1115,7 @@ static int act_open_rpl(struct t3cdev *tdev, struct sk_buff *skb, void *ctx)
1115 struct iwch_ep *ep = ctx; 1115 struct iwch_ep *ep = ctx;
1116 struct cpl_act_open_rpl *rpl = cplhdr(skb); 1116 struct cpl_act_open_rpl *rpl = cplhdr(skb);
1117 1117
1118 PDBG("%s ep %p status %u errno %d\n", __FUNCTION__, ep, rpl->status, 1118 PDBG("%s ep %p status %u errno %d\n", __func__, ep, rpl->status,
1119 status2errno(rpl->status)); 1119 status2errno(rpl->status));
1120 connect_reply_upcall(ep, status2errno(rpl->status)); 1120 connect_reply_upcall(ep, status2errno(rpl->status));
1121 state_set(&ep->com, DEAD); 1121 state_set(&ep->com, DEAD);
@@ -1133,7 +1133,7 @@ static int listen_start(struct iwch_listen_ep *ep)
1133 struct sk_buff *skb; 1133 struct sk_buff *skb;
1134 struct cpl_pass_open_req *req; 1134 struct cpl_pass_open_req *req;
1135 1135
1136 PDBG("%s ep %p\n", __FUNCTION__, ep); 1136 PDBG("%s ep %p\n", __func__, ep);
1137 skb = get_skb(NULL, sizeof(*req), GFP_KERNEL); 1137 skb = get_skb(NULL, sizeof(*req), GFP_KERNEL);
1138 if (!skb) { 1138 if (!skb) {
1139 printk(KERN_ERR MOD "t3c_listen_start failed to alloc skb!\n"); 1139 printk(KERN_ERR MOD "t3c_listen_start failed to alloc skb!\n");
@@ -1162,7 +1162,7 @@ static int pass_open_rpl(struct t3cdev *tdev, struct sk_buff *skb, void *ctx)
1162 struct iwch_listen_ep *ep = ctx; 1162 struct iwch_listen_ep *ep = ctx;
1163 struct cpl_pass_open_rpl *rpl = cplhdr(skb); 1163 struct cpl_pass_open_rpl *rpl = cplhdr(skb);
1164 1164
1165 PDBG("%s ep %p status %d error %d\n", __FUNCTION__, ep, 1165 PDBG("%s ep %p status %d error %d\n", __func__, ep,
1166 rpl->status, status2errno(rpl->status)); 1166 rpl->status, status2errno(rpl->status));
1167 ep->com.rpl_err = status2errno(rpl->status); 1167 ep->com.rpl_err = status2errno(rpl->status);
1168 ep->com.rpl_done = 1; 1168 ep->com.rpl_done = 1;
@@ -1176,10 +1176,10 @@ static int listen_stop(struct iwch_listen_ep *ep)
1176 struct sk_buff *skb; 1176 struct sk_buff *skb;
1177 struct cpl_close_listserv_req *req; 1177 struct cpl_close_listserv_req *req;
1178 1178
1179 PDBG("%s ep %p\n", __FUNCTION__, ep); 1179 PDBG("%s ep %p\n", __func__, ep);
1180 skb = get_skb(NULL, sizeof(*req), GFP_KERNEL); 1180 skb = get_skb(NULL, sizeof(*req), GFP_KERNEL);
1181 if (!skb) { 1181 if (!skb) {
1182 printk(KERN_ERR MOD "%s - failed to alloc skb\n", __FUNCTION__); 1182 printk(KERN_ERR MOD "%s - failed to alloc skb\n", __func__);
1183 return -ENOMEM; 1183 return -ENOMEM;
1184 } 1184 }
1185 req = (struct cpl_close_listserv_req *) skb_put(skb, sizeof(*req)); 1185 req = (struct cpl_close_listserv_req *) skb_put(skb, sizeof(*req));
@@ -1197,7 +1197,7 @@ static int close_listsrv_rpl(struct t3cdev *tdev, struct sk_buff *skb,
1197 struct iwch_listen_ep *ep = ctx; 1197 struct iwch_listen_ep *ep = ctx;
1198 struct cpl_close_listserv_rpl *rpl = cplhdr(skb); 1198 struct cpl_close_listserv_rpl *rpl = cplhdr(skb);
1199 1199
1200 PDBG("%s ep %p\n", __FUNCTION__, ep); 1200 PDBG("%s ep %p\n", __func__, ep);
1201 ep->com.rpl_err = status2errno(rpl->status); 1201 ep->com.rpl_err = status2errno(rpl->status);
1202 ep->com.rpl_done = 1; 1202 ep->com.rpl_done = 1;
1203 wake_up(&ep->com.waitq); 1203 wake_up(&ep->com.waitq);
@@ -1211,7 +1211,7 @@ static void accept_cr(struct iwch_ep *ep, __be32 peer_ip, struct sk_buff *skb)
1211 u32 opt0h, opt0l, opt2; 1211 u32 opt0h, opt0l, opt2;
1212 int wscale; 1212 int wscale;
1213 1213
1214 PDBG("%s ep %p\n", __FUNCTION__, ep); 1214 PDBG("%s ep %p\n", __func__, ep);
1215 BUG_ON(skb_cloned(skb)); 1215 BUG_ON(skb_cloned(skb));
1216 skb_trim(skb, sizeof(*rpl)); 1216 skb_trim(skb, sizeof(*rpl));
1217 skb_get(skb); 1217 skb_get(skb);
@@ -1244,7 +1244,7 @@ static void accept_cr(struct iwch_ep *ep, __be32 peer_ip, struct sk_buff *skb)
1244static void reject_cr(struct t3cdev *tdev, u32 hwtid, __be32 peer_ip, 1244static void reject_cr(struct t3cdev *tdev, u32 hwtid, __be32 peer_ip,
1245 struct sk_buff *skb) 1245 struct sk_buff *skb)
1246{ 1246{
1247 PDBG("%s t3cdev %p tid %u peer_ip %x\n", __FUNCTION__, tdev, hwtid, 1247 PDBG("%s t3cdev %p tid %u peer_ip %x\n", __func__, tdev, hwtid,
1248 peer_ip); 1248 peer_ip);
1249 BUG_ON(skb_cloned(skb)); 1249 BUG_ON(skb_cloned(skb));
1250 skb_trim(skb, sizeof(struct cpl_tid_release)); 1250 skb_trim(skb, sizeof(struct cpl_tid_release));
@@ -1279,11 +1279,11 @@ static int pass_accept_req(struct t3cdev *tdev, struct sk_buff *skb, void *ctx)
1279 struct rtable *rt; 1279 struct rtable *rt;
1280 struct iff_mac tim; 1280 struct iff_mac tim;
1281 1281
1282 PDBG("%s parent ep %p tid %u\n", __FUNCTION__, parent_ep, hwtid); 1282 PDBG("%s parent ep %p tid %u\n", __func__, parent_ep, hwtid);
1283 1283
1284 if (state_read(&parent_ep->com) != LISTEN) { 1284 if (state_read(&parent_ep->com) != LISTEN) {
1285 printk(KERN_ERR "%s - listening ep not in LISTEN\n", 1285 printk(KERN_ERR "%s - listening ep not in LISTEN\n",
1286 __FUNCTION__); 1286 __func__);
1287 goto reject; 1287 goto reject;
1288 } 1288 }
1289 1289
@@ -1295,7 +1295,7 @@ static int pass_accept_req(struct t3cdev *tdev, struct sk_buff *skb, void *ctx)
1295 if (tdev->ctl(tdev, GET_IFF_FROM_MAC, &tim) < 0 || !tim.dev) { 1295 if (tdev->ctl(tdev, GET_IFF_FROM_MAC, &tim) < 0 || !tim.dev) {
1296 printk(KERN_ERR 1296 printk(KERN_ERR
1297 "%s bad dst mac %02x %02x %02x %02x %02x %02x\n", 1297 "%s bad dst mac %02x %02x %02x %02x %02x %02x\n",
1298 __FUNCTION__, 1298 __func__,
1299 req->dst_mac[0], 1299 req->dst_mac[0],
1300 req->dst_mac[1], 1300 req->dst_mac[1],
1301 req->dst_mac[2], 1301 req->dst_mac[2],
@@ -1313,21 +1313,21 @@ static int pass_accept_req(struct t3cdev *tdev, struct sk_buff *skb, void *ctx)
1313 req->peer_port, G_PASS_OPEN_TOS(ntohl(req->tos_tid))); 1313 req->peer_port, G_PASS_OPEN_TOS(ntohl(req->tos_tid)));
1314 if (!rt) { 1314 if (!rt) {
1315 printk(KERN_ERR MOD "%s - failed to find dst entry!\n", 1315 printk(KERN_ERR MOD "%s - failed to find dst entry!\n",
1316 __FUNCTION__); 1316 __func__);
1317 goto reject; 1317 goto reject;
1318 } 1318 }
1319 dst = &rt->u.dst; 1319 dst = &rt->u.dst;
1320 l2t = t3_l2t_get(tdev, dst->neighbour, dst->neighbour->dev); 1320 l2t = t3_l2t_get(tdev, dst->neighbour, dst->neighbour->dev);
1321 if (!l2t) { 1321 if (!l2t) {
1322 printk(KERN_ERR MOD "%s - failed to allocate l2t entry!\n", 1322 printk(KERN_ERR MOD "%s - failed to allocate l2t entry!\n",
1323 __FUNCTION__); 1323 __func__);
1324 dst_release(dst); 1324 dst_release(dst);
1325 goto reject; 1325 goto reject;
1326 } 1326 }
1327 child_ep = alloc_ep(sizeof(*child_ep), GFP_KERNEL); 1327 child_ep = alloc_ep(sizeof(*child_ep), GFP_KERNEL);
1328 if (!child_ep) { 1328 if (!child_ep) {
1329 printk(KERN_ERR MOD "%s - failed to allocate ep entry!\n", 1329 printk(KERN_ERR MOD "%s - failed to allocate ep entry!\n",
1330 __FUNCTION__); 1330 __func__);
1331 l2t_release(L2DATA(tdev), l2t); 1331 l2t_release(L2DATA(tdev), l2t);
1332 dst_release(dst); 1332 dst_release(dst);
1333 goto reject; 1333 goto reject;
@@ -1362,7 +1362,7 @@ static int pass_establish(struct t3cdev *tdev, struct sk_buff *skb, void *ctx)
1362 struct iwch_ep *ep = ctx; 1362 struct iwch_ep *ep = ctx;
1363 struct cpl_pass_establish *req = cplhdr(skb); 1363 struct cpl_pass_establish *req = cplhdr(skb);
1364 1364
1365 PDBG("%s ep %p\n", __FUNCTION__, ep); 1365 PDBG("%s ep %p\n", __func__, ep);
1366 ep->snd_seq = ntohl(req->snd_isn); 1366 ep->snd_seq = ntohl(req->snd_isn);
1367 ep->rcv_seq = ntohl(req->rcv_isn); 1367 ep->rcv_seq = ntohl(req->rcv_isn);
1368 1368
@@ -1383,7 +1383,7 @@ static int peer_close(struct t3cdev *tdev, struct sk_buff *skb, void *ctx)
1383 int disconnect = 1; 1383 int disconnect = 1;
1384 int release = 0; 1384 int release = 0;
1385 1385
1386 PDBG("%s ep %p\n", __FUNCTION__, ep); 1386 PDBG("%s ep %p\n", __func__, ep);
1387 dst_confirm(ep->dst); 1387 dst_confirm(ep->dst);
1388 1388
1389 spin_lock_irqsave(&ep->com.lock, flags); 1389 spin_lock_irqsave(&ep->com.lock, flags);
@@ -1473,7 +1473,7 @@ static int peer_abort(struct t3cdev *tdev, struct sk_buff *skb, void *ctx)
1473 int state; 1473 int state;
1474 1474
1475 if (is_neg_adv_abort(req->status)) { 1475 if (is_neg_adv_abort(req->status)) {
1476 PDBG("%s neg_adv_abort ep %p tid %d\n", __FUNCTION__, ep, 1476 PDBG("%s neg_adv_abort ep %p tid %d\n", __func__, ep,
1477 ep->hwtid); 1477 ep->hwtid);
1478 t3_l2t_send_event(ep->com.tdev, ep->l2t); 1478 t3_l2t_send_event(ep->com.tdev, ep->l2t);
1479 return CPL_RET_BUF_DONE; 1479 return CPL_RET_BUF_DONE;
@@ -1489,7 +1489,7 @@ static int peer_abort(struct t3cdev *tdev, struct sk_buff *skb, void *ctx)
1489 } 1489 }
1490 1490
1491 state = state_read(&ep->com); 1491 state = state_read(&ep->com);
1492 PDBG("%s ep %p state %u\n", __FUNCTION__, ep, state); 1492 PDBG("%s ep %p state %u\n", __func__, ep, state);
1493 switch (state) { 1493 switch (state) {
1494 case CONNECTING: 1494 case CONNECTING:
1495 break; 1495 break;
@@ -1528,14 +1528,14 @@ static int peer_abort(struct t3cdev *tdev, struct sk_buff *skb, void *ctx)
1528 if (ret) 1528 if (ret)
1529 printk(KERN_ERR MOD 1529 printk(KERN_ERR MOD
1530 "%s - qp <- error failed!\n", 1530 "%s - qp <- error failed!\n",
1531 __FUNCTION__); 1531 __func__);
1532 } 1532 }
1533 peer_abort_upcall(ep); 1533 peer_abort_upcall(ep);
1534 break; 1534 break;
1535 case ABORTING: 1535 case ABORTING:
1536 break; 1536 break;
1537 case DEAD: 1537 case DEAD:
1538 PDBG("%s PEER_ABORT IN DEAD STATE!!!!\n", __FUNCTION__); 1538 PDBG("%s PEER_ABORT IN DEAD STATE!!!!\n", __func__);
1539 return CPL_RET_BUF_DONE; 1539 return CPL_RET_BUF_DONE;
1540 default: 1540 default:
1541 BUG_ON(1); 1541 BUG_ON(1);
@@ -1546,7 +1546,7 @@ static int peer_abort(struct t3cdev *tdev, struct sk_buff *skb, void *ctx)
1546 rpl_skb = get_skb(skb, sizeof(*rpl), GFP_KERNEL); 1546 rpl_skb = get_skb(skb, sizeof(*rpl), GFP_KERNEL);
1547 if (!rpl_skb) { 1547 if (!rpl_skb) {
1548 printk(KERN_ERR MOD "%s - cannot allocate skb!\n", 1548 printk(KERN_ERR MOD "%s - cannot allocate skb!\n",
1549 __FUNCTION__); 1549 __func__);
1550 dst_release(ep->dst); 1550 dst_release(ep->dst);
1551 l2t_release(L2DATA(ep->com.tdev), ep->l2t); 1551 l2t_release(L2DATA(ep->com.tdev), ep->l2t);
1552 put_ep(&ep->com); 1552 put_ep(&ep->com);
@@ -1573,7 +1573,7 @@ static int close_con_rpl(struct t3cdev *tdev, struct sk_buff *skb, void *ctx)
1573 unsigned long flags; 1573 unsigned long flags;
1574 int release = 0; 1574 int release = 0;
1575 1575
1576 PDBG("%s ep %p\n", __FUNCTION__, ep); 1576 PDBG("%s ep %p\n", __func__, ep);
1577 BUG_ON(!ep); 1577 BUG_ON(!ep);
1578 1578
1579 /* The cm_id may be null if we failed to connect */ 1579 /* The cm_id may be null if we failed to connect */
@@ -1624,9 +1624,9 @@ static int terminate(struct t3cdev *tdev, struct sk_buff *skb, void *ctx)
1624{ 1624{
1625 struct iwch_ep *ep = ctx; 1625 struct iwch_ep *ep = ctx;
1626 1626
1627 PDBG("%s ep %p\n", __FUNCTION__, ep); 1627 PDBG("%s ep %p\n", __func__, ep);
1628 skb_pull(skb, sizeof(struct cpl_rdma_terminate)); 1628 skb_pull(skb, sizeof(struct cpl_rdma_terminate));
1629 PDBG("%s saving %d bytes of term msg\n", __FUNCTION__, skb->len); 1629 PDBG("%s saving %d bytes of term msg\n", __func__, skb->len);
1630 skb_copy_from_linear_data(skb, ep->com.qp->attr.terminate_buffer, 1630 skb_copy_from_linear_data(skb, ep->com.qp->attr.terminate_buffer,
1631 skb->len); 1631 skb->len);
1632 ep->com.qp->attr.terminate_msg_len = skb->len; 1632 ep->com.qp->attr.terminate_msg_len = skb->len;
@@ -1639,13 +1639,13 @@ static int ec_status(struct t3cdev *tdev, struct sk_buff *skb, void *ctx)
1639 struct cpl_rdma_ec_status *rep = cplhdr(skb); 1639 struct cpl_rdma_ec_status *rep = cplhdr(skb);
1640 struct iwch_ep *ep = ctx; 1640 struct iwch_ep *ep = ctx;
1641 1641
1642 PDBG("%s ep %p tid %u status %d\n", __FUNCTION__, ep, ep->hwtid, 1642 PDBG("%s ep %p tid %u status %d\n", __func__, ep, ep->hwtid,
1643 rep->status); 1643 rep->status);
1644 if (rep->status) { 1644 if (rep->status) {
1645 struct iwch_qp_attributes attrs; 1645 struct iwch_qp_attributes attrs;
1646 1646
1647 printk(KERN_ERR MOD "%s BAD CLOSE - Aborting tid %u\n", 1647 printk(KERN_ERR MOD "%s BAD CLOSE - Aborting tid %u\n",
1648 __FUNCTION__, ep->hwtid); 1648 __func__, ep->hwtid);
1649 stop_ep_timer(ep); 1649 stop_ep_timer(ep);
1650 attrs.next_state = IWCH_QP_STATE_ERROR; 1650 attrs.next_state = IWCH_QP_STATE_ERROR;
1651 iwch_modify_qp(ep->com.qp->rhp, 1651 iwch_modify_qp(ep->com.qp->rhp,
@@ -1663,7 +1663,7 @@ static void ep_timeout(unsigned long arg)
1663 unsigned long flags; 1663 unsigned long flags;
1664 1664
1665 spin_lock_irqsave(&ep->com.lock, flags); 1665 spin_lock_irqsave(&ep->com.lock, flags);
1666 PDBG("%s ep %p tid %u state %d\n", __FUNCTION__, ep, ep->hwtid, 1666 PDBG("%s ep %p tid %u state %d\n", __func__, ep, ep->hwtid,
1667 ep->com.state); 1667 ep->com.state);
1668 switch (ep->com.state) { 1668 switch (ep->com.state) {
1669 case MPA_REQ_SENT: 1669 case MPA_REQ_SENT:
@@ -1693,7 +1693,7 @@ int iwch_reject_cr(struct iw_cm_id *cm_id, const void *pdata, u8 pdata_len)
1693{ 1693{
1694 int err; 1694 int err;
1695 struct iwch_ep *ep = to_ep(cm_id); 1695 struct iwch_ep *ep = to_ep(cm_id);
1696 PDBG("%s ep %p tid %u\n", __FUNCTION__, ep, ep->hwtid); 1696 PDBG("%s ep %p tid %u\n", __func__, ep, ep->hwtid);
1697 1697
1698 if (state_read(&ep->com) == DEAD) { 1698 if (state_read(&ep->com) == DEAD) {
1699 put_ep(&ep->com); 1699 put_ep(&ep->com);
@@ -1718,7 +1718,7 @@ int iwch_accept_cr(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param)
1718 struct iwch_dev *h = to_iwch_dev(cm_id->device); 1718 struct iwch_dev *h = to_iwch_dev(cm_id->device);
1719 struct iwch_qp *qp = get_qhp(h, conn_param->qpn); 1719 struct iwch_qp *qp = get_qhp(h, conn_param->qpn);
1720 1720
1721 PDBG("%s ep %p tid %u\n", __FUNCTION__, ep, ep->hwtid); 1721 PDBG("%s ep %p tid %u\n", __func__, ep, ep->hwtid);
1722 if (state_read(&ep->com) == DEAD) 1722 if (state_read(&ep->com) == DEAD)
1723 return -ECONNRESET; 1723 return -ECONNRESET;
1724 1724
@@ -1739,7 +1739,7 @@ int iwch_accept_cr(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param)
1739 ep->com.rpl_err = 0; 1739 ep->com.rpl_err = 0;
1740 ep->ird = conn_param->ird; 1740 ep->ird = conn_param->ird;
1741 ep->ord = conn_param->ord; 1741 ep->ord = conn_param->ord;
1742 PDBG("%s %d ird %d ord %d\n", __FUNCTION__, __LINE__, ep->ird, ep->ord); 1742 PDBG("%s %d ird %d ord %d\n", __func__, __LINE__, ep->ird, ep->ord);
1743 1743
1744 get_ep(&ep->com); 1744 get_ep(&ep->com);
1745 1745
@@ -1810,7 +1810,7 @@ int iwch_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param)
1810 1810
1811 ep = alloc_ep(sizeof(*ep), GFP_KERNEL); 1811 ep = alloc_ep(sizeof(*ep), GFP_KERNEL);
1812 if (!ep) { 1812 if (!ep) {
1813 printk(KERN_ERR MOD "%s - cannot alloc ep.\n", __FUNCTION__); 1813 printk(KERN_ERR MOD "%s - cannot alloc ep.\n", __func__);
1814 err = -ENOMEM; 1814 err = -ENOMEM;
1815 goto out; 1815 goto out;
1816 } 1816 }
@@ -1827,7 +1827,7 @@ int iwch_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param)
1827 ep->com.cm_id = cm_id; 1827 ep->com.cm_id = cm_id;
1828 ep->com.qp = get_qhp(h, conn_param->qpn); 1828 ep->com.qp = get_qhp(h, conn_param->qpn);
1829 BUG_ON(!ep->com.qp); 1829 BUG_ON(!ep->com.qp);
1830 PDBG("%s qpn 0x%x qp %p cm_id %p\n", __FUNCTION__, conn_param->qpn, 1830 PDBG("%s qpn 0x%x qp %p cm_id %p\n", __func__, conn_param->qpn,
1831 ep->com.qp, cm_id); 1831 ep->com.qp, cm_id);
1832 1832
1833 /* 1833 /*
@@ -1835,7 +1835,7 @@ int iwch_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param)
1835 */ 1835 */
1836 ep->atid = cxgb3_alloc_atid(h->rdev.t3cdev_p, &t3c_client, ep); 1836 ep->atid = cxgb3_alloc_atid(h->rdev.t3cdev_p, &t3c_client, ep);
1837 if (ep->atid == -1) { 1837 if (ep->atid == -1) {
1838 printk(KERN_ERR MOD "%s - cannot alloc atid.\n", __FUNCTION__); 1838 printk(KERN_ERR MOD "%s - cannot alloc atid.\n", __func__);
1839 err = -ENOMEM; 1839 err = -ENOMEM;
1840 goto fail2; 1840 goto fail2;
1841 } 1841 }
@@ -1847,7 +1847,7 @@ int iwch_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param)
1847 cm_id->local_addr.sin_port, 1847 cm_id->local_addr.sin_port,
1848 cm_id->remote_addr.sin_port, IPTOS_LOWDELAY); 1848 cm_id->remote_addr.sin_port, IPTOS_LOWDELAY);
1849 if (!rt) { 1849 if (!rt) {
1850 printk(KERN_ERR MOD "%s - cannot find route.\n", __FUNCTION__); 1850 printk(KERN_ERR MOD "%s - cannot find route.\n", __func__);
1851 err = -EHOSTUNREACH; 1851 err = -EHOSTUNREACH;
1852 goto fail3; 1852 goto fail3;
1853 } 1853 }
@@ -1857,7 +1857,7 @@ int iwch_connect(struct iw_cm_id *cm_id, struct iw_cm_conn_param *conn_param)
1857 ep->l2t = t3_l2t_get(ep->com.tdev, ep->dst->neighbour, 1857 ep->l2t = t3_l2t_get(ep->com.tdev, ep->dst->neighbour,
1858 ep->dst->neighbour->dev); 1858 ep->dst->neighbour->dev);
1859 if (!ep->l2t) { 1859 if (!ep->l2t) {
1860 printk(KERN_ERR MOD "%s - cannot alloc l2e.\n", __FUNCTION__); 1860 printk(KERN_ERR MOD "%s - cannot alloc l2e.\n", __func__);
1861 err = -ENOMEM; 1861 err = -ENOMEM;
1862 goto fail4; 1862 goto fail4;
1863 } 1863 }
@@ -1894,11 +1894,11 @@ int iwch_create_listen(struct iw_cm_id *cm_id, int backlog)
1894 1894
1895 ep = alloc_ep(sizeof(*ep), GFP_KERNEL); 1895 ep = alloc_ep(sizeof(*ep), GFP_KERNEL);
1896 if (!ep) { 1896 if (!ep) {
1897 printk(KERN_ERR MOD "%s - cannot alloc ep.\n", __FUNCTION__); 1897 printk(KERN_ERR MOD "%s - cannot alloc ep.\n", __func__);
1898 err = -ENOMEM; 1898 err = -ENOMEM;
1899 goto fail1; 1899 goto fail1;
1900 } 1900 }
1901 PDBG("%s ep %p\n", __FUNCTION__, ep); 1901 PDBG("%s ep %p\n", __func__, ep);
1902 ep->com.tdev = h->rdev.t3cdev_p; 1902 ep->com.tdev = h->rdev.t3cdev_p;
1903 cm_id->add_ref(cm_id); 1903 cm_id->add_ref(cm_id);
1904 ep->com.cm_id = cm_id; 1904 ep->com.cm_id = cm_id;
@@ -1910,7 +1910,7 @@ int iwch_create_listen(struct iw_cm_id *cm_id, int backlog)
1910 */ 1910 */
1911 ep->stid = cxgb3_alloc_stid(h->rdev.t3cdev_p, &t3c_client, ep); 1911 ep->stid = cxgb3_alloc_stid(h->rdev.t3cdev_p, &t3c_client, ep);
1912 if (ep->stid == -1) { 1912 if (ep->stid == -1) {
1913 printk(KERN_ERR MOD "%s - cannot alloc atid.\n", __FUNCTION__); 1913 printk(KERN_ERR MOD "%s - cannot alloc atid.\n", __func__);
1914 err = -ENOMEM; 1914 err = -ENOMEM;
1915 goto fail2; 1915 goto fail2;
1916 } 1916 }
@@ -1942,7 +1942,7 @@ int iwch_destroy_listen(struct iw_cm_id *cm_id)
1942 int err; 1942 int err;
1943 struct iwch_listen_ep *ep = to_listen_ep(cm_id); 1943 struct iwch_listen_ep *ep = to_listen_ep(cm_id);
1944 1944
1945 PDBG("%s ep %p\n", __FUNCTION__, ep); 1945 PDBG("%s ep %p\n", __func__, ep);
1946 1946
1947 might_sleep(); 1947 might_sleep();
1948 state_set(&ep->com, DEAD); 1948 state_set(&ep->com, DEAD);
@@ -1965,11 +1965,11 @@ int iwch_ep_disconnect(struct iwch_ep *ep, int abrupt, gfp_t gfp)
1965 1965
1966 spin_lock_irqsave(&ep->com.lock, flags); 1966 spin_lock_irqsave(&ep->com.lock, flags);
1967 1967
1968 PDBG("%s ep %p state %s, abrupt %d\n", __FUNCTION__, ep, 1968 PDBG("%s ep %p state %s, abrupt %d\n", __func__, ep,
1969 states[ep->com.state], abrupt); 1969 states[ep->com.state], abrupt);
1970 1970
1971 if (ep->com.state == DEAD) { 1971 if (ep->com.state == DEAD) {
1972 PDBG("%s already dead ep %p\n", __FUNCTION__, ep); 1972 PDBG("%s already dead ep %p\n", __func__, ep);
1973 goto out; 1973 goto out;
1974 } 1974 }
1975 1975
@@ -2020,7 +2020,7 @@ int iwch_ep_redirect(void *ctx, struct dst_entry *old, struct dst_entry *new,
2020 if (ep->dst != old) 2020 if (ep->dst != old)
2021 return 0; 2021 return 0;
2022 2022
2023 PDBG("%s ep %p redirect to dst %p l2t %p\n", __FUNCTION__, ep, new, 2023 PDBG("%s ep %p redirect to dst %p l2t %p\n", __func__, ep, new,
2024 l2t); 2024 l2t);
2025 dst_hold(new); 2025 dst_hold(new);
2026 l2t_release(L2DATA(ep->com.tdev), ep->l2t); 2026 l2t_release(L2DATA(ep->com.tdev), ep->l2t);