aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorHannes Eder <heder@google.com>2009-08-02 07:05:41 -0400
committerDavid S. Miller <davem@davemloft.net>2009-08-02 21:29:30 -0400
commit1e3e238e9c4bf9987b19185235cd0cdc21ea038c (patch)
treec58e2f1954ccb86ad1cca93ef269200cb183403c
parentbcd218be5aebed94951a750b1d477aea86fb68ea (diff)
IPVS: use pr_err and friends instead of IP_VS_ERR and friends
Since pr_err and friends are used instead of printk there is no point in keeping IP_VS_ERR and friends. Furthermore make use of '__func__' instead of hard coded function names. Signed-off-by: Hannes Eder <heder@google.com> Acked-by: Simon Horman <horms@verge.net.au> Signed-off-by: David S. Miller <davem@davemloft.net>
-rw-r--r--include/net/ip_vs.h7
-rw-r--r--net/netfilter/ipvs/ip_vs_app.c16
-rw-r--r--net/netfilter/ipvs/ip_vs_conn.c14
-rw-r--r--net/netfilter/ipvs/ip_vs_core.c24
-rw-r--r--net/netfilter/ipvs/ip_vs_ctl.c62
-rw-r--r--net/netfilter/ipvs/ip_vs_dh.c4
-rw-r--r--net/netfilter/ipvs/ip_vs_ftp.c4
-rw-r--r--net/netfilter/ipvs/ip_vs_lblc.c6
-rw-r--r--net/netfilter/ipvs/ip_vs_lblcr.c14
-rw-r--r--net/netfilter/ipvs/ip_vs_lc.c2
-rw-r--r--net/netfilter/ipvs/ip_vs_nq.c2
-rw-r--r--net/netfilter/ipvs/ip_vs_proto.c2
-rw-r--r--net/netfilter/ipvs/ip_vs_proto_tcp.c2
-rw-r--r--net/netfilter/ipvs/ip_vs_proto_udp.c2
-rw-r--r--net/netfilter/ipvs/ip_vs_rr.c2
-rw-r--r--net/netfilter/ipvs/ip_vs_sched.c38
-rw-r--r--net/netfilter/ipvs/ip_vs_sed.c2
-rw-r--r--net/netfilter/ipvs/ip_vs_sh.c2
-rw-r--r--net/netfilter/ipvs/ip_vs_sync.c50
-rw-r--r--net/netfilter/ipvs/ip_vs_wrr.c4
-rw-r--r--net/netfilter/ipvs/ip_vs_xmit.c40
21 files changed, 147 insertions, 152 deletions
diff --git a/include/net/ip_vs.h b/include/net/ip_vs.h
index 910820327bc4..1c8ee1b13651 100644
--- a/include/net/ip_vs.h
+++ b/include/net/ip_vs.h
@@ -150,13 +150,10 @@ static inline const char *ip_vs_dbg_addr(int af, char *buf, size_t buf_len,
150#endif 150#endif
151 151
152#define IP_VS_BUG() BUG() 152#define IP_VS_BUG() BUG()
153#define IP_VS_ERR(msg...) pr_err(msg) 153#define IP_VS_ERR_RL(msg, ...) \
154#define IP_VS_INFO(msg...) pr_info(msg)
155#define IP_VS_WARNING(msg...) pr_warning(msg)
156#define IP_VS_ERR_RL(msg...) \
157 do { \ 154 do { \
158 if (net_ratelimit()) \ 155 if (net_ratelimit()) \
159 pr_err(msg); \ 156 pr_err(msg, ##__VA_ARGS__); \
160 } while (0) 157 } while (0)
161 158
162#ifdef CONFIG_IP_VS_DEBUG 159#ifdef CONFIG_IP_VS_DEBUG
diff --git a/net/netfilter/ipvs/ip_vs_app.c b/net/netfilter/ipvs/ip_vs_app.c
index c1781f80daf2..3c7e42735b60 100644
--- a/net/netfilter/ipvs/ip_vs_app.c
+++ b/net/netfilter/ipvs/ip_vs_app.c
@@ -265,12 +265,12 @@ static inline void vs_fix_seq(const struct ip_vs_seq *vseq, struct tcphdr *th)
265 if (vseq->delta || vseq->previous_delta) { 265 if (vseq->delta || vseq->previous_delta) {
266 if(after(seq, vseq->init_seq)) { 266 if(after(seq, vseq->init_seq)) {
267 th->seq = htonl(seq + vseq->delta); 267 th->seq = htonl(seq + vseq->delta);
268 IP_VS_DBG(9, "vs_fix_seq(): added delta (%d) to seq\n", 268 IP_VS_DBG(9, "%s(): added delta (%d) to seq\n",
269 vseq->delta); 269 __func__, vseq->delta);
270 } else { 270 } else {
271 th->seq = htonl(seq + vseq->previous_delta); 271 th->seq = htonl(seq + vseq->previous_delta);
272 IP_VS_DBG(9, "vs_fix_seq(): added previous_delta " 272 IP_VS_DBG(9, "%s(): added previous_delta (%d) to seq\n",
273 "(%d) to seq\n", vseq->previous_delta); 273 __func__, vseq->previous_delta);
274 } 274 }
275 } 275 }
276} 276}
@@ -294,14 +294,14 @@ vs_fix_ack_seq(const struct ip_vs_seq *vseq, struct tcphdr *th)
294 to receive next, so compare it with init_seq+delta */ 294 to receive next, so compare it with init_seq+delta */
295 if(after(ack_seq, vseq->init_seq+vseq->delta)) { 295 if(after(ack_seq, vseq->init_seq+vseq->delta)) {
296 th->ack_seq = htonl(ack_seq - vseq->delta); 296 th->ack_seq = htonl(ack_seq - vseq->delta);
297 IP_VS_DBG(9, "vs_fix_ack_seq(): subtracted delta " 297 IP_VS_DBG(9, "%s(): subtracted delta "
298 "(%d) from ack_seq\n", vseq->delta); 298 "(%d) from ack_seq\n", __func__, vseq->delta);
299 299
300 } else { 300 } else {
301 th->ack_seq = htonl(ack_seq - vseq->previous_delta); 301 th->ack_seq = htonl(ack_seq - vseq->previous_delta);
302 IP_VS_DBG(9, "vs_fix_ack_seq(): subtracted " 302 IP_VS_DBG(9, "%s(): subtracted "
303 "previous_delta (%d) from ack_seq\n", 303 "previous_delta (%d) from ack_seq\n",
304 vseq->previous_delta); 304 __func__, vseq->previous_delta);
305 } 305 }
306 } 306 }
307} 307}
diff --git a/net/netfilter/ipvs/ip_vs_conn.c b/net/netfilter/ipvs/ip_vs_conn.c
index 4173d7b1d4cc..27c30cf933da 100644
--- a/net/netfilter/ipvs/ip_vs_conn.c
+++ b/net/netfilter/ipvs/ip_vs_conn.c
@@ -153,8 +153,8 @@ static inline int ip_vs_conn_hash(struct ip_vs_conn *cp)
153 atomic_inc(&cp->refcnt); 153 atomic_inc(&cp->refcnt);
154 ret = 1; 154 ret = 1;
155 } else { 155 } else {
156 IP_VS_ERR("ip_vs_conn_hash(): request for already hashed, " 156 pr_err("%s(): request for already hashed, called from %pF\n",
157 "called from %p\n", __builtin_return_address(0)); 157 __func__, __builtin_return_address(0));
158 ret = 0; 158 ret = 0;
159 } 159 }
160 160
@@ -692,7 +692,7 @@ ip_vs_conn_new(int af, int proto, const union nf_inet_addr *caddr, __be16 cport,
692 692
693 cp = kmem_cache_zalloc(ip_vs_conn_cachep, GFP_ATOMIC); 693 cp = kmem_cache_zalloc(ip_vs_conn_cachep, GFP_ATOMIC);
694 if (cp == NULL) { 694 if (cp == NULL) {
695 IP_VS_ERR_RL("ip_vs_conn_new: no memory available.\n"); 695 IP_VS_ERR_RL("%s(): no memory\n", __func__);
696 return NULL; 696 return NULL;
697 } 697 }
698 698
@@ -1076,10 +1076,10 @@ int __init ip_vs_conn_init(void)
1076 return -ENOMEM; 1076 return -ENOMEM;
1077 } 1077 }
1078 1078
1079 IP_VS_INFO("Connection hash table configured " 1079 pr_info("Connection hash table configured "
1080 "(size=%d, memory=%ldKbytes)\n", 1080 "(size=%d, memory=%ldKbytes)\n",
1081 IP_VS_CONN_TAB_SIZE, 1081 IP_VS_CONN_TAB_SIZE,
1082 (long)(IP_VS_CONN_TAB_SIZE*sizeof(struct list_head))/1024); 1082 (long)(IP_VS_CONN_TAB_SIZE*sizeof(struct list_head))/1024);
1083 IP_VS_DBG(0, "Each connection entry needs %Zd bytes at least\n", 1083 IP_VS_DBG(0, "Each connection entry needs %Zd bytes at least\n",
1084 sizeof(struct ip_vs_conn)); 1084 sizeof(struct ip_vs_conn));
1085 1085
diff --git a/net/netfilter/ipvs/ip_vs_core.c b/net/netfilter/ipvs/ip_vs_core.c
index 6811dcaca0f6..b227750af752 100644
--- a/net/netfilter/ipvs/ip_vs_core.c
+++ b/net/netfilter/ipvs/ip_vs_core.c
@@ -391,9 +391,9 @@ ip_vs_schedule(struct ip_vs_service *svc, const struct sk_buff *skb)
391 */ 391 */
392 if (!svc->fwmark && pptr[1] != svc->port) { 392 if (!svc->fwmark && pptr[1] != svc->port) {
393 if (!svc->port) 393 if (!svc->port)
394 IP_VS_ERR("Schedule: port zero only supported " 394 pr_err("Schedule: port zero only supported "
395 "in persistent services, " 395 "in persistent services, "
396 "check your ipvs configuration\n"); 396 "check your ipvs configuration\n");
397 return NULL; 397 return NULL;
398 } 398 }
399 399
@@ -465,7 +465,7 @@ int ip_vs_leave(struct ip_vs_service *svc, struct sk_buff *skb,
465 ip_vs_service_put(svc); 465 ip_vs_service_put(svc);
466 466
467 /* create a new connection entry */ 467 /* create a new connection entry */
468 IP_VS_DBG(6, "ip_vs_leave: create a cache_bypass entry\n"); 468 IP_VS_DBG(6, "%s(): create a cache_bypass entry\n", __func__);
469 cp = ip_vs_conn_new(svc->af, iph.protocol, 469 cp = ip_vs_conn_new(svc->af, iph.protocol,
470 &iph.saddr, pptr[0], 470 &iph.saddr, pptr[0],
471 &iph.daddr, pptr[1], 471 &iph.daddr, pptr[1],
@@ -667,8 +667,8 @@ static int handle_response_icmp(int af, struct sk_buff *skb,
667 unsigned int verdict = NF_DROP; 667 unsigned int verdict = NF_DROP;
668 668
669 if (IP_VS_FWD_METHOD(cp) != 0) { 669 if (IP_VS_FWD_METHOD(cp) != 0) {
670 IP_VS_ERR("shouldn't reach here, because the box is on the " 670 pr_err("shouldn't reach here, because the box is on the "
671 "half connection in the tun/dr module.\n"); 671 "half connection in the tun/dr module.\n");
672 } 672 }
673 673
674 /* Ensure the checksum is correct */ 674 /* Ensure the checksum is correct */
@@ -1490,7 +1490,7 @@ static int __init ip_vs_init(void)
1490 1490
1491 ret = ip_vs_control_init(); 1491 ret = ip_vs_control_init();
1492 if (ret < 0) { 1492 if (ret < 0) {
1493 IP_VS_ERR("can't setup control.\n"); 1493 pr_err("can't setup control.\n");
1494 goto cleanup_estimator; 1494 goto cleanup_estimator;
1495 } 1495 }
1496 1496
@@ -1498,23 +1498,23 @@ static int __init ip_vs_init(void)
1498 1498
1499 ret = ip_vs_app_init(); 1499 ret = ip_vs_app_init();
1500 if (ret < 0) { 1500 if (ret < 0) {
1501 IP_VS_ERR("can't setup application helper.\n"); 1501 pr_err("can't setup application helper.\n");
1502 goto cleanup_protocol; 1502 goto cleanup_protocol;
1503 } 1503 }
1504 1504
1505 ret = ip_vs_conn_init(); 1505 ret = ip_vs_conn_init();
1506 if (ret < 0) { 1506 if (ret < 0) {
1507 IP_VS_ERR("can't setup connection table.\n"); 1507 pr_err("can't setup connection table.\n");
1508 goto cleanup_app; 1508 goto cleanup_app;
1509 } 1509 }
1510 1510
1511 ret = nf_register_hooks(ip_vs_ops, ARRAY_SIZE(ip_vs_ops)); 1511 ret = nf_register_hooks(ip_vs_ops, ARRAY_SIZE(ip_vs_ops));
1512 if (ret < 0) { 1512 if (ret < 0) {
1513 IP_VS_ERR("can't register hooks.\n"); 1513 pr_err("can't register hooks.\n");
1514 goto cleanup_conn; 1514 goto cleanup_conn;
1515 } 1515 }
1516 1516
1517 IP_VS_INFO("ipvs loaded.\n"); 1517 pr_info("ipvs loaded.\n");
1518 return ret; 1518 return ret;
1519 1519
1520 cleanup_conn: 1520 cleanup_conn:
@@ -1537,7 +1537,7 @@ static void __exit ip_vs_cleanup(void)
1537 ip_vs_protocol_cleanup(); 1537 ip_vs_protocol_cleanup();
1538 ip_vs_control_cleanup(); 1538 ip_vs_control_cleanup();
1539 ip_vs_estimator_cleanup(); 1539 ip_vs_estimator_cleanup();
1540 IP_VS_INFO("ipvs unloaded.\n"); 1540 pr_info("ipvs unloaded.\n");
1541} 1541}
1542 1542
1543module_init(ip_vs_init); 1543module_init(ip_vs_init);
diff --git a/net/netfilter/ipvs/ip_vs_ctl.c b/net/netfilter/ipvs/ip_vs_ctl.c
index e6133ea1ea4c..fba2892b99e1 100644
--- a/net/netfilter/ipvs/ip_vs_ctl.c
+++ b/net/netfilter/ipvs/ip_vs_ctl.c
@@ -343,8 +343,8 @@ static int ip_vs_svc_hash(struct ip_vs_service *svc)
343 unsigned hash; 343 unsigned hash;
344 344
345 if (svc->flags & IP_VS_SVC_F_HASHED) { 345 if (svc->flags & IP_VS_SVC_F_HASHED) {
346 IP_VS_ERR("ip_vs_svc_hash(): request for already hashed, " 346 pr_err("%s(): request for already hashed, called from %pF\n",
347 "called from %p\n", __builtin_return_address(0)); 347 __func__, __builtin_return_address(0));
348 return 0; 348 return 0;
349 } 349 }
350 350
@@ -377,8 +377,8 @@ static int ip_vs_svc_hash(struct ip_vs_service *svc)
377static int ip_vs_svc_unhash(struct ip_vs_service *svc) 377static int ip_vs_svc_unhash(struct ip_vs_service *svc)
378{ 378{
379 if (!(svc->flags & IP_VS_SVC_F_HASHED)) { 379 if (!(svc->flags & IP_VS_SVC_F_HASHED)) {
380 IP_VS_ERR("ip_vs_svc_unhash(): request for unhash flagged, " 380 pr_err("%s(): request for unhash flagged, called from %pF\n",
381 "called from %p\n", __builtin_return_address(0)); 381 __func__, __builtin_return_address(0));
382 return 0; 382 return 0;
383 } 383 }
384 384
@@ -844,7 +844,7 @@ ip_vs_new_dest(struct ip_vs_service *svc, struct ip_vs_dest_user_kern *udest,
844 844
845 dest = kzalloc(sizeof(struct ip_vs_dest), GFP_ATOMIC); 845 dest = kzalloc(sizeof(struct ip_vs_dest), GFP_ATOMIC);
846 if (dest == NULL) { 846 if (dest == NULL) {
847 IP_VS_ERR("ip_vs_new_dest: kmalloc failed.\n"); 847 pr_err("%s(): no memory.\n", __func__);
848 return -ENOMEM; 848 return -ENOMEM;
849 } 849 }
850 850
@@ -888,13 +888,13 @@ ip_vs_add_dest(struct ip_vs_service *svc, struct ip_vs_dest_user_kern *udest)
888 EnterFunction(2); 888 EnterFunction(2);
889 889
890 if (udest->weight < 0) { 890 if (udest->weight < 0) {
891 IP_VS_ERR("ip_vs_add_dest(): server weight less than zero\n"); 891 pr_err("%s(): server weight less than zero\n", __func__);
892 return -ERANGE; 892 return -ERANGE;
893 } 893 }
894 894
895 if (udest->l_threshold > udest->u_threshold) { 895 if (udest->l_threshold > udest->u_threshold) {
896 IP_VS_ERR("ip_vs_add_dest(): lower threshold is higher than " 896 pr_err("%s(): lower threshold is higher than upper threshold\n",
897 "upper threshold\n"); 897 __func__);
898 return -ERANGE; 898 return -ERANGE;
899 } 899 }
900 900
@@ -906,7 +906,7 @@ ip_vs_add_dest(struct ip_vs_service *svc, struct ip_vs_dest_user_kern *udest)
906 dest = ip_vs_lookup_dest(svc, &daddr, dport); 906 dest = ip_vs_lookup_dest(svc, &daddr, dport);
907 907
908 if (dest != NULL) { 908 if (dest != NULL) {
909 IP_VS_DBG(1, "ip_vs_add_dest(): dest already exists\n"); 909 IP_VS_DBG(1, "%s(): dest already exists\n", __func__);
910 return -EEXIST; 910 return -EEXIST;
911 } 911 }
912 912
@@ -1000,13 +1000,13 @@ ip_vs_edit_dest(struct ip_vs_service *svc, struct ip_vs_dest_user_kern *udest)
1000 EnterFunction(2); 1000 EnterFunction(2);
1001 1001
1002 if (udest->weight < 0) { 1002 if (udest->weight < 0) {
1003 IP_VS_ERR("ip_vs_edit_dest(): server weight less than zero\n"); 1003 pr_err("%s(): server weight less than zero\n", __func__);
1004 return -ERANGE; 1004 return -ERANGE;
1005 } 1005 }
1006 1006
1007 if (udest->l_threshold > udest->u_threshold) { 1007 if (udest->l_threshold > udest->u_threshold) {
1008 IP_VS_ERR("ip_vs_edit_dest(): lower threshold is higher than " 1008 pr_err("%s(): lower threshold is higher than upper threshold\n",
1009 "upper threshold\n"); 1009 __func__);
1010 return -ERANGE; 1010 return -ERANGE;
1011 } 1011 }
1012 1012
@@ -1018,7 +1018,7 @@ ip_vs_edit_dest(struct ip_vs_service *svc, struct ip_vs_dest_user_kern *udest)
1018 dest = ip_vs_lookup_dest(svc, &daddr, dport); 1018 dest = ip_vs_lookup_dest(svc, &daddr, dport);
1019 1019
1020 if (dest == NULL) { 1020 if (dest == NULL) {
1021 IP_VS_DBG(1, "ip_vs_edit_dest(): dest doesn't exist\n"); 1021 IP_VS_DBG(1, "%s(): dest doesn't exist\n", __func__);
1022 return -ENOENT; 1022 return -ENOENT;
1023 } 1023 }
1024 1024
@@ -1118,7 +1118,7 @@ ip_vs_del_dest(struct ip_vs_service *svc, struct ip_vs_dest_user_kern *udest)
1118 dest = ip_vs_lookup_dest(svc, &udest->addr, dport); 1118 dest = ip_vs_lookup_dest(svc, &udest->addr, dport);
1119 1119
1120 if (dest == NULL) { 1120 if (dest == NULL) {
1121 IP_VS_DBG(1, "ip_vs_del_dest(): destination not found!\n"); 1121 IP_VS_DBG(1, "%s(): destination not found!\n", __func__);
1122 return -ENOENT; 1122 return -ENOENT;
1123 } 1123 }
1124 1124
@@ -1164,8 +1164,7 @@ ip_vs_add_service(struct ip_vs_service_user_kern *u,
1164 /* Lookup the scheduler by 'u->sched_name' */ 1164 /* Lookup the scheduler by 'u->sched_name' */
1165 sched = ip_vs_scheduler_get(u->sched_name); 1165 sched = ip_vs_scheduler_get(u->sched_name);
1166 if (sched == NULL) { 1166 if (sched == NULL) {
1167 IP_VS_INFO("Scheduler module ip_vs_%s not found\n", 1167 pr_info("Scheduler module ip_vs_%s not found\n", u->sched_name);
1168 u->sched_name);
1169 ret = -ENOENT; 1168 ret = -ENOENT;
1170 goto out_mod_dec; 1169 goto out_mod_dec;
1171 } 1170 }
@@ -1179,7 +1178,7 @@ ip_vs_add_service(struct ip_vs_service_user_kern *u,
1179 1178
1180 svc = kzalloc(sizeof(struct ip_vs_service), GFP_ATOMIC); 1179 svc = kzalloc(sizeof(struct ip_vs_service), GFP_ATOMIC);
1181 if (svc == NULL) { 1180 if (svc == NULL) {
1182 IP_VS_DBG(1, "ip_vs_add_service: kmalloc failed.\n"); 1181 IP_VS_DBG(1, "%s(): no memory\n", __func__);
1183 ret = -ENOMEM; 1182 ret = -ENOMEM;
1184 goto out_err; 1183 goto out_err;
1185 } 1184 }
@@ -1262,8 +1261,7 @@ ip_vs_edit_service(struct ip_vs_service *svc, struct ip_vs_service_user_kern *u)
1262 */ 1261 */
1263 sched = ip_vs_scheduler_get(u->sched_name); 1262 sched = ip_vs_scheduler_get(u->sched_name);
1264 if (sched == NULL) { 1263 if (sched == NULL) {
1265 IP_VS_INFO("Scheduler module ip_vs_%s not found\n", 1264 pr_info("Scheduler module ip_vs_%s not found\n", u->sched_name);
1266 u->sched_name);
1267 return -ENOENT; 1265 return -ENOENT;
1268 } 1266 }
1269 old_sched = sched; 1267 old_sched = sched;
@@ -2080,8 +2078,8 @@ do_ip_vs_set_ctl(struct sock *sk, int cmd, void __user *user, unsigned int len)
2080 return -EPERM; 2078 return -EPERM;
2081 2079
2082 if (len != set_arglen[SET_CMDID(cmd)]) { 2080 if (len != set_arglen[SET_CMDID(cmd)]) {
2083 IP_VS_ERR("set_ctl: len %u != %u\n", 2081 pr_err("set_ctl: len %u != %u\n",
2084 len, set_arglen[SET_CMDID(cmd)]); 2082 len, set_arglen[SET_CMDID(cmd)]);
2085 return -EINVAL; 2083 return -EINVAL;
2086 } 2084 }
2087 2085
@@ -2132,9 +2130,9 @@ do_ip_vs_set_ctl(struct sock *sk, int cmd, void __user *user, unsigned int len)
2132 2130
2133 /* Check for valid protocol: TCP or UDP, even for fwmark!=0 */ 2131 /* Check for valid protocol: TCP or UDP, even for fwmark!=0 */
2134 if (usvc.protocol != IPPROTO_TCP && usvc.protocol != IPPROTO_UDP) { 2132 if (usvc.protocol != IPPROTO_TCP && usvc.protocol != IPPROTO_UDP) {
2135 IP_VS_ERR("set_ctl: invalid protocol: %d %pI4:%d %s\n", 2133 pr_err("set_ctl: invalid protocol: %d %pI4:%d %s\n",
2136 usvc.protocol, &usvc.addr.ip, 2134 usvc.protocol, &usvc.addr.ip,
2137 ntohs(usvc.port), usvc.sched_name); 2135 ntohs(usvc.port), usvc.sched_name);
2138 ret = -EFAULT; 2136 ret = -EFAULT;
2139 goto out_unlock; 2137 goto out_unlock;
2140 } 2138 }
@@ -2359,8 +2357,8 @@ do_ip_vs_get_ctl(struct sock *sk, int cmd, void __user *user, int *len)
2359 return -EPERM; 2357 return -EPERM;
2360 2358
2361 if (*len < get_arglen[GET_CMDID(cmd)]) { 2359 if (*len < get_arglen[GET_CMDID(cmd)]) {
2362 IP_VS_ERR("get_ctl: len %u < %u\n", 2360 pr_err("get_ctl: len %u < %u\n",
2363 *len, get_arglen[GET_CMDID(cmd)]); 2361 *len, get_arglen[GET_CMDID(cmd)]);
2364 return -EINVAL; 2362 return -EINVAL;
2365 } 2363 }
2366 2364
@@ -2405,7 +2403,7 @@ do_ip_vs_get_ctl(struct sock *sk, int cmd, void __user *user, int *len)
2405 size = sizeof(*get) + 2403 size = sizeof(*get) +
2406 sizeof(struct ip_vs_service_entry) * get->num_services; 2404 sizeof(struct ip_vs_service_entry) * get->num_services;
2407 if (*len != size) { 2405 if (*len != size) {
2408 IP_VS_ERR("length: %u != %u\n", *len, size); 2406 pr_err("length: %u != %u\n", *len, size);
2409 ret = -EINVAL; 2407 ret = -EINVAL;
2410 goto out; 2408 goto out;
2411 } 2409 }
@@ -2445,7 +2443,7 @@ do_ip_vs_get_ctl(struct sock *sk, int cmd, void __user *user, int *len)
2445 size = sizeof(*get) + 2443 size = sizeof(*get) +
2446 sizeof(struct ip_vs_dest_entry) * get->num_dests; 2444 sizeof(struct ip_vs_dest_entry) * get->num_dests;
2447 if (*len != size) { 2445 if (*len != size) {
2448 IP_VS_ERR("length: %u != %u\n", *len, size); 2446 pr_err("length: %u != %u\n", *len, size);
2449 ret = -EINVAL; 2447 ret = -EINVAL;
2450 goto out; 2448 goto out;
2451 } 2449 }
@@ -3173,7 +3171,7 @@ static int ip_vs_genl_get_cmd(struct sk_buff *skb, struct genl_info *info)
3173 else if (cmd == IPVS_CMD_GET_CONFIG) 3171 else if (cmd == IPVS_CMD_GET_CONFIG)
3174 reply_cmd = IPVS_CMD_SET_CONFIG; 3172 reply_cmd = IPVS_CMD_SET_CONFIG;
3175 else { 3173 else {
3176 IP_VS_ERR("unknown Generic Netlink command\n"); 3174 pr_err("unknown Generic Netlink command\n");
3177 return -EINVAL; 3175 return -EINVAL;
3178 } 3176 }
3179 3177
@@ -3238,7 +3236,7 @@ static int ip_vs_genl_get_cmd(struct sk_buff *skb, struct genl_info *info)
3238 goto out; 3236 goto out;
3239 3237
3240nla_put_failure: 3238nla_put_failure:
3241 IP_VS_ERR("not enough space in Netlink message\n"); 3239 pr_err("not enough space in Netlink message\n");
3242 ret = -EMSGSIZE; 3240 ret = -EMSGSIZE;
3243 3241
3244out_err: 3242out_err:
@@ -3369,13 +3367,13 @@ int __init ip_vs_control_init(void)
3369 3367
3370 ret = nf_register_sockopt(&ip_vs_sockopts); 3368 ret = nf_register_sockopt(&ip_vs_sockopts);
3371 if (ret) { 3369 if (ret) {
3372 IP_VS_ERR("cannot register sockopt.\n"); 3370 pr_err("cannot register sockopt.\n");
3373 return ret; 3371 return ret;
3374 } 3372 }
3375 3373
3376 ret = ip_vs_genl_register(); 3374 ret = ip_vs_genl_register();
3377 if (ret) { 3375 if (ret) {
3378 IP_VS_ERR("cannot register Generic Netlink interface.\n"); 3376 pr_err("cannot register Generic Netlink interface.\n");
3379 nf_unregister_sockopt(&ip_vs_sockopts); 3377 nf_unregister_sockopt(&ip_vs_sockopts);
3380 return ret; 3378 return ret;
3381 } 3379 }
diff --git a/net/netfilter/ipvs/ip_vs_dh.c b/net/netfilter/ipvs/ip_vs_dh.c
index d0c0594d1e2e..fe3e18834b91 100644
--- a/net/netfilter/ipvs/ip_vs_dh.c
+++ b/net/netfilter/ipvs/ip_vs_dh.c
@@ -150,7 +150,7 @@ static int ip_vs_dh_init_svc(struct ip_vs_service *svc)
150 tbl = kmalloc(sizeof(struct ip_vs_dh_bucket)*IP_VS_DH_TAB_SIZE, 150 tbl = kmalloc(sizeof(struct ip_vs_dh_bucket)*IP_VS_DH_TAB_SIZE,
151 GFP_ATOMIC); 151 GFP_ATOMIC);
152 if (tbl == NULL) { 152 if (tbl == NULL) {
153 IP_VS_ERR("ip_vs_dh_init_svc(): no memory\n"); 153 pr_err("%s(): no memory\n", __func__);
154 return -ENOMEM; 154 return -ENOMEM;
155 } 155 }
156 svc->sched_data = tbl; 156 svc->sched_data = tbl;
@@ -217,7 +217,7 @@ ip_vs_dh_schedule(struct ip_vs_service *svc, const struct sk_buff *skb)
217 217
218 ip_vs_fill_iphdr(svc->af, skb_network_header(skb), &iph); 218 ip_vs_fill_iphdr(svc->af, skb_network_header(skb), &iph);
219 219
220 IP_VS_DBG(6, "ip_vs_dh_schedule(): Scheduling...\n"); 220 IP_VS_DBG(6, "%s(): Scheduling...\n", __func__);
221 221
222 tbl = (struct ip_vs_dh_bucket *)svc->sched_data; 222 tbl = (struct ip_vs_dh_bucket *)svc->sched_data;
223 dest = ip_vs_dh_get(svc->af, tbl, &iph.daddr); 223 dest = ip_vs_dh_get(svc->af, tbl, &iph.daddr);
diff --git a/net/netfilter/ipvs/ip_vs_ftp.c b/net/netfilter/ipvs/ip_vs_ftp.c
index 9c16a3f64c1b..33e2c799cba7 100644
--- a/net/netfilter/ipvs/ip_vs_ftp.c
+++ b/net/netfilter/ipvs/ip_vs_ftp.c
@@ -385,8 +385,8 @@ static int __init ip_vs_ftp_init(void)
385 ret = register_ip_vs_app_inc(app, app->protocol, ports[i]); 385 ret = register_ip_vs_app_inc(app, app->protocol, ports[i]);
386 if (ret) 386 if (ret)
387 break; 387 break;
388 IP_VS_INFO("%s: loaded support on port[%d] = %d\n", 388 pr_info("%s: loaded support on port[%d] = %d\n",
389 app->name, i, ports[i]); 389 app->name, i, ports[i]);
390 } 390 }
391 391
392 if (ret) 392 if (ret)
diff --git a/net/netfilter/ipvs/ip_vs_lblc.c b/net/netfilter/ipvs/ip_vs_lblc.c
index 98fb185d890b..c1757f3620cd 100644
--- a/net/netfilter/ipvs/ip_vs_lblc.c
+++ b/net/netfilter/ipvs/ip_vs_lblc.c
@@ -202,7 +202,7 @@ ip_vs_lblc_new(struct ip_vs_lblc_table *tbl, const union nf_inet_addr *daddr,
202 if (!en) { 202 if (!en) {
203 en = kmalloc(sizeof(*en), GFP_ATOMIC); 203 en = kmalloc(sizeof(*en), GFP_ATOMIC);
204 if (!en) { 204 if (!en) {
205 IP_VS_ERR("ip_vs_lblc_new(): no memory\n"); 205 pr_err("%s(): no memory\n", __func__);
206 return NULL; 206 return NULL;
207 } 207 }
208 208
@@ -335,7 +335,7 @@ static int ip_vs_lblc_init_svc(struct ip_vs_service *svc)
335 */ 335 */
336 tbl = kmalloc(sizeof(*tbl), GFP_ATOMIC); 336 tbl = kmalloc(sizeof(*tbl), GFP_ATOMIC);
337 if (tbl == NULL) { 337 if (tbl == NULL) {
338 IP_VS_ERR("ip_vs_lblc_init_svc(): no memory\n"); 338 pr_err("%s(): no memory\n", __func__);
339 return -ENOMEM; 339 return -ENOMEM;
340 } 340 }
341 svc->sched_data = tbl; 341 svc->sched_data = tbl;
@@ -480,7 +480,7 @@ ip_vs_lblc_schedule(struct ip_vs_service *svc, const struct sk_buff *skb)
480 480
481 ip_vs_fill_iphdr(svc->af, skb_network_header(skb), &iph); 481 ip_vs_fill_iphdr(svc->af, skb_network_header(skb), &iph);
482 482
483 IP_VS_DBG(6, "ip_vs_lblc_schedule(): Scheduling...\n"); 483 IP_VS_DBG(6, "%s(): Scheduling...\n", __func__);
484 484
485 /* First look in our cache */ 485 /* First look in our cache */
486 read_lock(&svc->sched_lock); 486 read_lock(&svc->sched_lock);
diff --git a/net/netfilter/ipvs/ip_vs_lblcr.c b/net/netfilter/ipvs/ip_vs_lblcr.c
index 5f5e5f4bad5e..715b57f9540d 100644
--- a/net/netfilter/ipvs/ip_vs_lblcr.c
+++ b/net/netfilter/ipvs/ip_vs_lblcr.c
@@ -111,7 +111,7 @@ ip_vs_dest_set_insert(struct ip_vs_dest_set *set, struct ip_vs_dest *dest)
111 111
112 e = kmalloc(sizeof(*e), GFP_ATOMIC); 112 e = kmalloc(sizeof(*e), GFP_ATOMIC);
113 if (e == NULL) { 113 if (e == NULL) {
114 IP_VS_ERR("ip_vs_dest_set_insert(): no memory\n"); 114 pr_err("%s(): no memory\n", __func__);
115 return NULL; 115 return NULL;
116 } 116 }
117 117
@@ -205,8 +205,9 @@ static inline struct ip_vs_dest *ip_vs_dest_set_min(struct ip_vs_dest_set *set)
205 } 205 }
206 } 206 }
207 207
208 IP_VS_DBG_BUF(6, "ip_vs_dest_set_min: server %s:%d " 208 IP_VS_DBG_BUF(6, "%s(): server %s:%d "
209 "activeconns %d refcnt %d weight %d overhead %d\n", 209 "activeconns %d refcnt %d weight %d overhead %d\n",
210 __func__,
210 IP_VS_DBG_ADDR(least->af, &least->addr), 211 IP_VS_DBG_ADDR(least->af, &least->addr),
211 ntohs(least->port), 212 ntohs(least->port),
212 atomic_read(&least->activeconns), 213 atomic_read(&least->activeconns),
@@ -252,8 +253,9 @@ static inline struct ip_vs_dest *ip_vs_dest_set_max(struct ip_vs_dest_set *set)
252 } 253 }
253 } 254 }
254 255
255 IP_VS_DBG_BUF(6, "ip_vs_dest_set_max: server %s:%d " 256 IP_VS_DBG_BUF(6, "%s(): server %s:%d "
256 "activeconns %d refcnt %d weight %d overhead %d\n", 257 "activeconns %d refcnt %d weight %d overhead %d\n",
258 __func__,
257 IP_VS_DBG_ADDR(most->af, &most->addr), ntohs(most->port), 259 IP_VS_DBG_ADDR(most->af, &most->addr), ntohs(most->port),
258 atomic_read(&most->activeconns), 260 atomic_read(&most->activeconns),
259 atomic_read(&most->refcnt), 261 atomic_read(&most->refcnt),
@@ -377,7 +379,7 @@ ip_vs_lblcr_new(struct ip_vs_lblcr_table *tbl, const union nf_inet_addr *daddr,
377 if (!en) { 379 if (!en) {
378 en = kmalloc(sizeof(*en), GFP_ATOMIC); 380 en = kmalloc(sizeof(*en), GFP_ATOMIC);
379 if (!en) { 381 if (!en) {
380 IP_VS_ERR("ip_vs_lblcr_new(): no memory\n"); 382 pr_err("%s(): no memory\n", __func__);
381 return NULL; 383 return NULL;
382 } 384 }
383 385
@@ -511,7 +513,7 @@ static int ip_vs_lblcr_init_svc(struct ip_vs_service *svc)
511 */ 513 */
512 tbl = kmalloc(sizeof(*tbl), GFP_ATOMIC); 514 tbl = kmalloc(sizeof(*tbl), GFP_ATOMIC);
513 if (tbl == NULL) { 515 if (tbl == NULL) {
514 IP_VS_ERR("ip_vs_lblcr_init_svc(): no memory\n"); 516 pr_err("%s(): no memory\n", __func__);
515 return -ENOMEM; 517 return -ENOMEM;
516 } 518 }
517 svc->sched_data = tbl; 519 svc->sched_data = tbl;
@@ -657,7 +659,7 @@ ip_vs_lblcr_schedule(struct ip_vs_service *svc, const struct sk_buff *skb)
657 659
658 ip_vs_fill_iphdr(svc->af, skb_network_header(skb), &iph); 660 ip_vs_fill_iphdr(svc->af, skb_network_header(skb), &iph);
659 661
660 IP_VS_DBG(6, "ip_vs_lblcr_schedule(): Scheduling...\n"); 662 IP_VS_DBG(6, "%s(): Scheduling...\n", __func__);
661 663
662 /* First look in our cache */ 664 /* First look in our cache */
663 read_lock(&svc->sched_lock); 665 read_lock(&svc->sched_lock);
diff --git a/net/netfilter/ipvs/ip_vs_lc.c b/net/netfilter/ipvs/ip_vs_lc.c
index 4ecd5e19c39a..4f69db1fac56 100644
--- a/net/netfilter/ipvs/ip_vs_lc.c
+++ b/net/netfilter/ipvs/ip_vs_lc.c
@@ -47,7 +47,7 @@ ip_vs_lc_schedule(struct ip_vs_service *svc, const struct sk_buff *skb)
47 struct ip_vs_dest *dest, *least = NULL; 47 struct ip_vs_dest *dest, *least = NULL;
48 unsigned int loh = 0, doh; 48 unsigned int loh = 0, doh;
49 49
50 IP_VS_DBG(6, "ip_vs_lc_schedule(): Scheduling...\n"); 50 IP_VS_DBG(6, "%s(): Scheduling...\n", __func__);
51 51
52 /* 52 /*
53 * Simply select the server with the least number of 53 * Simply select the server with the least number of
diff --git a/net/netfilter/ipvs/ip_vs_nq.c b/net/netfilter/ipvs/ip_vs_nq.c
index 2224478bdea8..c413e1830823 100644
--- a/net/netfilter/ipvs/ip_vs_nq.c
+++ b/net/netfilter/ipvs/ip_vs_nq.c
@@ -60,7 +60,7 @@ ip_vs_nq_schedule(struct ip_vs_service *svc, const struct sk_buff *skb)
60 struct ip_vs_dest *dest, *least = NULL; 60 struct ip_vs_dest *dest, *least = NULL;
61 unsigned int loh = 0, doh; 61 unsigned int loh = 0, doh;
62 62
63 IP_VS_DBG(6, "ip_vs_nq_schedule(): Scheduling...\n"); 63 IP_VS_DBG(6, "%s(): Scheduling...\n", __func__);
64 64
65 /* 65 /*
66 * We calculate the load of each dest server as follows: 66 * We calculate the load of each dest server as follows:
diff --git a/net/netfilter/ipvs/ip_vs_proto.c b/net/netfilter/ipvs/ip_vs_proto.c
index a95bc4021c90..85c8892e1e8b 100644
--- a/net/netfilter/ipvs/ip_vs_proto.c
+++ b/net/netfilter/ipvs/ip_vs_proto.c
@@ -262,7 +262,7 @@ int __init ip_vs_protocol_init(void)
262#ifdef CONFIG_IP_VS_PROTO_ESP 262#ifdef CONFIG_IP_VS_PROTO_ESP
263 REGISTER_PROTOCOL(&ip_vs_protocol_esp); 263 REGISTER_PROTOCOL(&ip_vs_protocol_esp);
264#endif 264#endif
265 IP_VS_INFO("Registered protocols (%s)\n", &protocols[2]); 265 pr_info("Registered protocols (%s)\n", &protocols[2]);
266 266
267 return 0; 267 return 0;
268} 268}
diff --git a/net/netfilter/ipvs/ip_vs_proto_tcp.c b/net/netfilter/ipvs/ip_vs_proto_tcp.c
index c36c80d3a2b4..2278e141489e 100644
--- a/net/netfilter/ipvs/ip_vs_proto_tcp.c
+++ b/net/netfilter/ipvs/ip_vs_proto_tcp.c
@@ -664,7 +664,7 @@ tcp_app_conn_bind(struct ip_vs_conn *cp)
664 break; 664 break;
665 spin_unlock(&tcp_app_lock); 665 spin_unlock(&tcp_app_lock);
666 666
667 IP_VS_DBG_BUF(9, "%s: Binding conn %s:%u->" 667 IP_VS_DBG_BUF(9, "%s(): Binding conn %s:%u->"
668 "%s:%u to app %s on port %u\n", 668 "%s:%u to app %s on port %u\n",
669 __func__, 669 __func__,
670 IP_VS_DBG_ADDR(cp->af, &cp->caddr), 670 IP_VS_DBG_ADDR(cp->af, &cp->caddr),
diff --git a/net/netfilter/ipvs/ip_vs_proto_udp.c b/net/netfilter/ipvs/ip_vs_proto_udp.c
index 96ebe40bc537..33a05d3684d9 100644
--- a/net/netfilter/ipvs/ip_vs_proto_udp.c
+++ b/net/netfilter/ipvs/ip_vs_proto_udp.c
@@ -445,7 +445,7 @@ static int udp_app_conn_bind(struct ip_vs_conn *cp)
445 break; 445 break;
446 spin_unlock(&udp_app_lock); 446 spin_unlock(&udp_app_lock);
447 447
448 IP_VS_DBG_BUF(9, "%s: Binding conn %s:%u->" 448 IP_VS_DBG_BUF(9, "%s(): Binding conn %s:%u->"
449 "%s:%u to app %s on port %u\n", 449 "%s:%u to app %s on port %u\n",
450 __func__, 450 __func__,
451 IP_VS_DBG_ADDR(cp->af, &cp->caddr), 451 IP_VS_DBG_ADDR(cp->af, &cp->caddr),
diff --git a/net/netfilter/ipvs/ip_vs_rr.c b/net/netfilter/ipvs/ip_vs_rr.c
index b01007e1c11e..e210f37d8ea2 100644
--- a/net/netfilter/ipvs/ip_vs_rr.c
+++ b/net/netfilter/ipvs/ip_vs_rr.c
@@ -51,7 +51,7 @@ ip_vs_rr_schedule(struct ip_vs_service *svc, const struct sk_buff *skb)
51 struct list_head *p, *q; 51 struct list_head *p, *q;
52 struct ip_vs_dest *dest; 52 struct ip_vs_dest *dest;
53 53
54 IP_VS_DBG(6, "ip_vs_rr_schedule(): Scheduling...\n"); 54 IP_VS_DBG(6, "%s(): Scheduling...\n", __func__);
55 55
56 write_lock(&svc->sched_lock); 56 write_lock(&svc->sched_lock);
57 p = (struct list_head *)svc->sched_data; 57 p = (struct list_head *)svc->sched_data;
diff --git a/net/netfilter/ipvs/ip_vs_sched.c b/net/netfilter/ipvs/ip_vs_sched.c
index 87bc5ea0ef29..bbc1ac795952 100644
--- a/net/netfilter/ipvs/ip_vs_sched.c
+++ b/net/netfilter/ipvs/ip_vs_sched.c
@@ -47,11 +47,11 @@ int ip_vs_bind_scheduler(struct ip_vs_service *svc,
47 int ret; 47 int ret;
48 48
49 if (svc == NULL) { 49 if (svc == NULL) {
50 IP_VS_ERR("ip_vs_bind_scheduler(): svc arg NULL\n"); 50 pr_err("%s(): svc arg NULL\n", __func__);
51 return -EINVAL; 51 return -EINVAL;
52 } 52 }
53 if (scheduler == NULL) { 53 if (scheduler == NULL) {
54 IP_VS_ERR("ip_vs_bind_scheduler(): scheduler arg NULL\n"); 54 pr_err("%s(): scheduler arg NULL\n", __func__);
55 return -EINVAL; 55 return -EINVAL;
56 } 56 }
57 57
@@ -60,7 +60,7 @@ int ip_vs_bind_scheduler(struct ip_vs_service *svc,
60 if (scheduler->init_service) { 60 if (scheduler->init_service) {
61 ret = scheduler->init_service(svc); 61 ret = scheduler->init_service(svc);
62 if (ret) { 62 if (ret) {
63 IP_VS_ERR("ip_vs_bind_scheduler(): init error\n"); 63 pr_err("%s(): init error\n", __func__);
64 return ret; 64 return ret;
65 } 65 }
66 } 66 }
@@ -77,19 +77,19 @@ int ip_vs_unbind_scheduler(struct ip_vs_service *svc)
77 struct ip_vs_scheduler *sched; 77 struct ip_vs_scheduler *sched;
78 78
79 if (svc == NULL) { 79 if (svc == NULL) {
80 IP_VS_ERR("ip_vs_unbind_scheduler(): svc arg NULL\n"); 80 pr_err("%s(): svc arg NULL\n", __func__);
81 return -EINVAL; 81 return -EINVAL;
82 } 82 }
83 83
84 sched = svc->scheduler; 84 sched = svc->scheduler;
85 if (sched == NULL) { 85 if (sched == NULL) {
86 IP_VS_ERR("ip_vs_unbind_scheduler(): svc isn't bound\n"); 86 pr_err("%s(): svc isn't bound\n", __func__);
87 return -EINVAL; 87 return -EINVAL;
88 } 88 }
89 89
90 if (sched->done_service) { 90 if (sched->done_service) {
91 if (sched->done_service(svc) != 0) { 91 if (sched->done_service(svc) != 0) {
92 IP_VS_ERR("ip_vs_unbind_scheduler(): done error\n"); 92 pr_err("%s(): done error\n", __func__);
93 return -EINVAL; 93 return -EINVAL;
94 } 94 }
95 } 95 }
@@ -106,8 +106,7 @@ static struct ip_vs_scheduler *ip_vs_sched_getbyname(const char *sched_name)
106{ 106{
107 struct ip_vs_scheduler *sched; 107 struct ip_vs_scheduler *sched;
108 108
109 IP_VS_DBG(2, "ip_vs_sched_getbyname(): sched_name \"%s\"\n", 109 IP_VS_DBG(2, "%s(): sched_name \"%s\"\n", __func__, sched_name);
110 sched_name);
111 110
112 read_lock_bh(&__ip_vs_sched_lock); 111 read_lock_bh(&__ip_vs_sched_lock);
113 112
@@ -173,12 +172,12 @@ int register_ip_vs_scheduler(struct ip_vs_scheduler *scheduler)
173 struct ip_vs_scheduler *sched; 172 struct ip_vs_scheduler *sched;
174 173
175 if (!scheduler) { 174 if (!scheduler) {
176 IP_VS_ERR("register_ip_vs_scheduler(): NULL arg\n"); 175 pr_err("%s(): NULL arg\n", __func__);
177 return -EINVAL; 176 return -EINVAL;
178 } 177 }
179 178
180 if (!scheduler->name) { 179 if (!scheduler->name) {
181 IP_VS_ERR("register_ip_vs_scheduler(): NULL scheduler_name\n"); 180 pr_err("%s(): NULL scheduler_name\n", __func__);
182 return -EINVAL; 181 return -EINVAL;
183 } 182 }
184 183
@@ -190,8 +189,8 @@ int register_ip_vs_scheduler(struct ip_vs_scheduler *scheduler)
190 if (!list_empty(&scheduler->n_list)) { 189 if (!list_empty(&scheduler->n_list)) {
191 write_unlock_bh(&__ip_vs_sched_lock); 190 write_unlock_bh(&__ip_vs_sched_lock);
192 ip_vs_use_count_dec(); 191 ip_vs_use_count_dec();
193 IP_VS_ERR("register_ip_vs_scheduler(): [%s] scheduler " 192 pr_err("%s(): [%s] scheduler already linked\n",
194 "already linked\n", scheduler->name); 193 __func__, scheduler->name);
195 return -EINVAL; 194 return -EINVAL;
196 } 195 }
197 196
@@ -203,9 +202,8 @@ int register_ip_vs_scheduler(struct ip_vs_scheduler *scheduler)
203 if (strcmp(scheduler->name, sched->name) == 0) { 202 if (strcmp(scheduler->name, sched->name) == 0) {
204 write_unlock_bh(&__ip_vs_sched_lock); 203 write_unlock_bh(&__ip_vs_sched_lock);
205 ip_vs_use_count_dec(); 204 ip_vs_use_count_dec();
206 IP_VS_ERR("register_ip_vs_scheduler(): [%s] scheduler " 205 pr_err("%s(): [%s] scheduler already existed "
207 "already existed in the system\n", 206 "in the system\n", __func__, scheduler->name);
208 scheduler->name);
209 return -EINVAL; 207 return -EINVAL;
210 } 208 }
211 } 209 }
@@ -215,7 +213,7 @@ int register_ip_vs_scheduler(struct ip_vs_scheduler *scheduler)
215 list_add(&scheduler->n_list, &ip_vs_schedulers); 213 list_add(&scheduler->n_list, &ip_vs_schedulers);
216 write_unlock_bh(&__ip_vs_sched_lock); 214 write_unlock_bh(&__ip_vs_sched_lock);
217 215
218 IP_VS_INFO("[%s] scheduler registered.\n", scheduler->name); 216 pr_info("[%s] scheduler registered.\n", scheduler->name);
219 217
220 return 0; 218 return 0;
221} 219}
@@ -227,15 +225,15 @@ int register_ip_vs_scheduler(struct ip_vs_scheduler *scheduler)
227int unregister_ip_vs_scheduler(struct ip_vs_scheduler *scheduler) 225int unregister_ip_vs_scheduler(struct ip_vs_scheduler *scheduler)
228{ 226{
229 if (!scheduler) { 227 if (!scheduler) {
230 IP_VS_ERR( "unregister_ip_vs_scheduler(): NULL arg\n"); 228 pr_err("%s(): NULL arg\n", __func__);
231 return -EINVAL; 229 return -EINVAL;
232 } 230 }
233 231
234 write_lock_bh(&__ip_vs_sched_lock); 232 write_lock_bh(&__ip_vs_sched_lock);
235 if (list_empty(&scheduler->n_list)) { 233 if (list_empty(&scheduler->n_list)) {
236 write_unlock_bh(&__ip_vs_sched_lock); 234 write_unlock_bh(&__ip_vs_sched_lock);
237 IP_VS_ERR("unregister_ip_vs_scheduler(): [%s] scheduler " 235 pr_err("%s(): [%s] scheduler is not in the list. failed\n",
238 "is not in the list. failed\n", scheduler->name); 236 __func__, scheduler->name);
239 return -EINVAL; 237 return -EINVAL;
240 } 238 }
241 239
@@ -248,7 +246,7 @@ int unregister_ip_vs_scheduler(struct ip_vs_scheduler *scheduler)
248 /* decrease the module use count */ 246 /* decrease the module use count */
249 ip_vs_use_count_dec(); 247 ip_vs_use_count_dec();
250 248
251 IP_VS_INFO("[%s] scheduler unregistered.\n", scheduler->name); 249 pr_info("[%s] scheduler unregistered.\n", scheduler->name);
252 250
253 return 0; 251 return 0;
254} 252}
diff --git a/net/netfilter/ipvs/ip_vs_sed.c b/net/netfilter/ipvs/ip_vs_sed.c
index 4f745dd86dd8..1ab75a9dc400 100644
--- a/net/netfilter/ipvs/ip_vs_sed.c
+++ b/net/netfilter/ipvs/ip_vs_sed.c
@@ -64,7 +64,7 @@ ip_vs_sed_schedule(struct ip_vs_service *svc, const struct sk_buff *skb)
64 struct ip_vs_dest *dest, *least; 64 struct ip_vs_dest *dest, *least;
65 unsigned int loh, doh; 65 unsigned int loh, doh;
66 66
67 IP_VS_DBG(6, "ip_vs_sed_schedule(): Scheduling...\n"); 67 IP_VS_DBG(6, "%s(): Scheduling...\n", __func__);
68 68
69 /* 69 /*
70 * We calculate the load of each dest server as follows: 70 * We calculate the load of each dest server as follows:
diff --git a/net/netfilter/ipvs/ip_vs_sh.c b/net/netfilter/ipvs/ip_vs_sh.c
index fb4d2d23f2fe..8e6cfd36e6f0 100644
--- a/net/netfilter/ipvs/ip_vs_sh.c
+++ b/net/netfilter/ipvs/ip_vs_sh.c
@@ -147,7 +147,7 @@ static int ip_vs_sh_init_svc(struct ip_vs_service *svc)
147 tbl = kmalloc(sizeof(struct ip_vs_sh_bucket)*IP_VS_SH_TAB_SIZE, 147 tbl = kmalloc(sizeof(struct ip_vs_sh_bucket)*IP_VS_SH_TAB_SIZE,
148 GFP_ATOMIC); 148 GFP_ATOMIC);
149 if (tbl == NULL) { 149 if (tbl == NULL) {
150 IP_VS_ERR("ip_vs_sh_init_svc(): no memory\n"); 150 pr_err("%s(): no memory\n", __func__);
151 return -ENOMEM; 151 return -ENOMEM;
152 } 152 }
153 svc->sched_data = tbl; 153 svc->sched_data = tbl;
diff --git a/net/netfilter/ipvs/ip_vs_sync.c b/net/netfilter/ipvs/ip_vs_sync.c
index cc04c99815fd..e177f0dc2084 100644
--- a/net/netfilter/ipvs/ip_vs_sync.c
+++ b/net/netfilter/ipvs/ip_vs_sync.c
@@ -246,7 +246,7 @@ void ip_vs_sync_conn(struct ip_vs_conn *cp)
246 if (!curr_sb) { 246 if (!curr_sb) {
247 if (!(curr_sb=ip_vs_sync_buff_create())) { 247 if (!(curr_sb=ip_vs_sync_buff_create())) {
248 spin_unlock(&curr_sb_lock); 248 spin_unlock(&curr_sb_lock);
249 IP_VS_ERR("ip_vs_sync_buff_create failed.\n"); 249 pr_err("ip_vs_sync_buff_create failed.\n");
250 return; 250 return;
251 } 251 }
252 } 252 }
@@ -412,7 +412,7 @@ static void ip_vs_process_message(const char *buffer, const size_t buflen)
412 if (dest) 412 if (dest)
413 atomic_dec(&dest->refcnt); 413 atomic_dec(&dest->refcnt);
414 if (!cp) { 414 if (!cp) {
415 IP_VS_ERR("ip_vs_conn_new failed\n"); 415 pr_err("ip_vs_conn_new failed\n");
416 return; 416 return;
417 } 417 }
418 } else if (!cp->dest) { 418 } else if (!cp->dest) {
@@ -580,8 +580,8 @@ static int bind_mcastif_addr(struct socket *sock, char *ifname)
580 580
581 addr = inet_select_addr(dev, 0, RT_SCOPE_UNIVERSE); 581 addr = inet_select_addr(dev, 0, RT_SCOPE_UNIVERSE);
582 if (!addr) 582 if (!addr)
583 IP_VS_ERR("You probably need to specify IP address on " 583 pr_err("You probably need to specify IP address on "
584 "multicast interface.\n"); 584 "multicast interface.\n");
585 585
586 IP_VS_DBG(7, "binding socket with (%s) %pI4\n", 586 IP_VS_DBG(7, "binding socket with (%s) %pI4\n",
587 ifname, &addr); 587 ifname, &addr);
@@ -605,13 +605,13 @@ static struct socket * make_send_sock(void)
605 /* First create a socket */ 605 /* First create a socket */
606 result = sock_create_kern(PF_INET, SOCK_DGRAM, IPPROTO_UDP, &sock); 606 result = sock_create_kern(PF_INET, SOCK_DGRAM, IPPROTO_UDP, &sock);
607 if (result < 0) { 607 if (result < 0) {
608 IP_VS_ERR("Error during creation of socket; terminating\n"); 608 pr_err("Error during creation of socket; terminating\n");
609 return ERR_PTR(result); 609 return ERR_PTR(result);
610 } 610 }
611 611
612 result = set_mcast_if(sock->sk, ip_vs_master_mcast_ifn); 612 result = set_mcast_if(sock->sk, ip_vs_master_mcast_ifn);
613 if (result < 0) { 613 if (result < 0) {
614 IP_VS_ERR("Error setting outbound mcast interface\n"); 614 pr_err("Error setting outbound mcast interface\n");
615 goto error; 615 goto error;
616 } 616 }
617 617
@@ -620,14 +620,14 @@ static struct socket * make_send_sock(void)
620 620
621 result = bind_mcastif_addr(sock, ip_vs_master_mcast_ifn); 621 result = bind_mcastif_addr(sock, ip_vs_master_mcast_ifn);
622 if (result < 0) { 622 if (result < 0) {
623 IP_VS_ERR("Error binding address of the mcast interface\n"); 623 pr_err("Error binding address of the mcast interface\n");
624 goto error; 624 goto error;
625 } 625 }
626 626
627 result = sock->ops->connect(sock, (struct sockaddr *) &mcast_addr, 627 result = sock->ops->connect(sock, (struct sockaddr *) &mcast_addr,
628 sizeof(struct sockaddr), 0); 628 sizeof(struct sockaddr), 0);
629 if (result < 0) { 629 if (result < 0) {
630 IP_VS_ERR("Error connecting to the multicast addr\n"); 630 pr_err("Error connecting to the multicast addr\n");
631 goto error; 631 goto error;
632 } 632 }
633 633
@@ -650,7 +650,7 @@ static struct socket * make_receive_sock(void)
650 /* First create a socket */ 650 /* First create a socket */
651 result = sock_create_kern(PF_INET, SOCK_DGRAM, IPPROTO_UDP, &sock); 651 result = sock_create_kern(PF_INET, SOCK_DGRAM, IPPROTO_UDP, &sock);
652 if (result < 0) { 652 if (result < 0) {
653 IP_VS_ERR("Error during creation of socket; terminating\n"); 653 pr_err("Error during creation of socket; terminating\n");
654 return ERR_PTR(result); 654 return ERR_PTR(result);
655 } 655 }
656 656
@@ -660,7 +660,7 @@ static struct socket * make_receive_sock(void)
660 result = sock->ops->bind(sock, (struct sockaddr *) &mcast_addr, 660 result = sock->ops->bind(sock, (struct sockaddr *) &mcast_addr,
661 sizeof(struct sockaddr)); 661 sizeof(struct sockaddr));
662 if (result < 0) { 662 if (result < 0) {
663 IP_VS_ERR("Error binding to the multicast addr\n"); 663 pr_err("Error binding to the multicast addr\n");
664 goto error; 664 goto error;
665 } 665 }
666 666
@@ -669,7 +669,7 @@ static struct socket * make_receive_sock(void)
669 (struct in_addr *) &mcast_addr.sin_addr, 669 (struct in_addr *) &mcast_addr.sin_addr,
670 ip_vs_backup_mcast_ifn); 670 ip_vs_backup_mcast_ifn);
671 if (result < 0) { 671 if (result < 0) {
672 IP_VS_ERR("Error joining to the multicast group\n"); 672 pr_err("Error joining to the multicast group\n");
673 goto error; 673 goto error;
674 } 674 }
675 675
@@ -709,7 +709,7 @@ ip_vs_send_sync_msg(struct socket *sock, struct ip_vs_sync_mesg *msg)
709 msg->size = htons(msg->size); 709 msg->size = htons(msg->size);
710 710
711 if (ip_vs_send_async(sock, (char *)msg, msize) != msize) 711 if (ip_vs_send_async(sock, (char *)msg, msize) != msize)
712 IP_VS_ERR("ip_vs_send_async error\n"); 712 pr_err("ip_vs_send_async error\n");
713} 713}
714 714
715static int 715static int
@@ -740,9 +740,9 @@ static int sync_thread_master(void *data)
740 struct ip_vs_sync_thread_data *tinfo = data; 740 struct ip_vs_sync_thread_data *tinfo = data;
741 struct ip_vs_sync_buff *sb; 741 struct ip_vs_sync_buff *sb;
742 742
743 IP_VS_INFO("sync thread started: state = MASTER, mcast_ifn = %s, " 743 pr_info("sync thread started: state = MASTER, mcast_ifn = %s, "
744 "syncid = %d\n", 744 "syncid = %d\n",
745 ip_vs_master_mcast_ifn, ip_vs_master_syncid); 745 ip_vs_master_mcast_ifn, ip_vs_master_syncid);
746 746
747 while (!kthread_should_stop()) { 747 while (!kthread_should_stop()) {
748 while ((sb = sb_dequeue())) { 748 while ((sb = sb_dequeue())) {
@@ -783,9 +783,9 @@ static int sync_thread_backup(void *data)
783 struct ip_vs_sync_thread_data *tinfo = data; 783 struct ip_vs_sync_thread_data *tinfo = data;
784 int len; 784 int len;
785 785
786 IP_VS_INFO("sync thread started: state = BACKUP, mcast_ifn = %s, " 786 pr_info("sync thread started: state = BACKUP, mcast_ifn = %s, "
787 "syncid = %d\n", 787 "syncid = %d\n",
788 ip_vs_backup_mcast_ifn, ip_vs_backup_syncid); 788 ip_vs_backup_mcast_ifn, ip_vs_backup_syncid);
789 789
790 while (!kthread_should_stop()) { 790 while (!kthread_should_stop()) {
791 wait_event_interruptible(*tinfo->sock->sk->sk_sleep, 791 wait_event_interruptible(*tinfo->sock->sk->sk_sleep,
@@ -797,7 +797,7 @@ static int sync_thread_backup(void *data)
797 len = ip_vs_receive(tinfo->sock, tinfo->buf, 797 len = ip_vs_receive(tinfo->sock, tinfo->buf,
798 sync_recv_mesg_maxlen); 798 sync_recv_mesg_maxlen);
799 if (len <= 0) { 799 if (len <= 0) {
800 IP_VS_ERR("receiving message error\n"); 800 pr_err("receiving message error\n");
801 break; 801 break;
802 } 802 }
803 803
@@ -827,7 +827,7 @@ int start_sync_thread(int state, char *mcast_ifn, __u8 syncid)
827 int (*threadfn)(void *data); 827 int (*threadfn)(void *data);
828 int result = -ENOMEM; 828 int result = -ENOMEM;
829 829
830 IP_VS_DBG(7, "%s: pid %d\n", __func__, task_pid_nr(current)); 830 IP_VS_DBG(7, "%s(): pid %d\n", __func__, task_pid_nr(current));
831 IP_VS_DBG(7, "Each ip_vs_sync_conn entry needs %Zd bytes\n", 831 IP_VS_DBG(7, "Each ip_vs_sync_conn entry needs %Zd bytes\n",
832 sizeof(struct ip_vs_sync_conn)); 832 sizeof(struct ip_vs_sync_conn));
833 833
@@ -904,14 +904,14 @@ out:
904 904
905int stop_sync_thread(int state) 905int stop_sync_thread(int state)
906{ 906{
907 IP_VS_DBG(7, "%s: pid %d\n", __func__, task_pid_nr(current)); 907 IP_VS_DBG(7, "%s(): pid %d\n", __func__, task_pid_nr(current));
908 908
909 if (state == IP_VS_STATE_MASTER) { 909 if (state == IP_VS_STATE_MASTER) {
910 if (!sync_master_thread) 910 if (!sync_master_thread)
911 return -ESRCH; 911 return -ESRCH;
912 912
913 IP_VS_INFO("stopping master sync thread %d ...\n", 913 pr_info("stopping master sync thread %d ...\n",
914 task_pid_nr(sync_master_thread)); 914 task_pid_nr(sync_master_thread));
915 915
916 /* 916 /*
917 * The lock synchronizes with sb_queue_tail(), so that we don't 917 * The lock synchronizes with sb_queue_tail(), so that we don't
@@ -928,8 +928,8 @@ int stop_sync_thread(int state)
928 if (!sync_backup_thread) 928 if (!sync_backup_thread)
929 return -ESRCH; 929 return -ESRCH;
930 930
931 IP_VS_INFO("stopping backup sync thread %d ...\n", 931 pr_info("stopping backup sync thread %d ...\n",
932 task_pid_nr(sync_backup_thread)); 932 task_pid_nr(sync_backup_thread));
933 933
934 ip_vs_sync_state &= ~IP_VS_STATE_BACKUP; 934 ip_vs_sync_state &= ~IP_VS_STATE_BACKUP;
935 kthread_stop(sync_backup_thread); 935 kthread_stop(sync_backup_thread);
diff --git a/net/netfilter/ipvs/ip_vs_wrr.c b/net/netfilter/ipvs/ip_vs_wrr.c
index c39ebb6c5a54..70ff82cda57d 100644
--- a/net/netfilter/ipvs/ip_vs_wrr.c
+++ b/net/netfilter/ipvs/ip_vs_wrr.c
@@ -97,7 +97,7 @@ static int ip_vs_wrr_init_svc(struct ip_vs_service *svc)
97 */ 97 */
98 mark = kmalloc(sizeof(struct ip_vs_wrr_mark), GFP_ATOMIC); 98 mark = kmalloc(sizeof(struct ip_vs_wrr_mark), GFP_ATOMIC);
99 if (mark == NULL) { 99 if (mark == NULL) {
100 IP_VS_ERR("ip_vs_wrr_init_svc(): no memory\n"); 100 pr_err("%s(): no memory\n", __func__);
101 return -ENOMEM; 101 return -ENOMEM;
102 } 102 }
103 mark->cl = &svc->destinations; 103 mark->cl = &svc->destinations;
@@ -144,7 +144,7 @@ ip_vs_wrr_schedule(struct ip_vs_service *svc, const struct sk_buff *skb)
144 struct ip_vs_wrr_mark *mark = svc->sched_data; 144 struct ip_vs_wrr_mark *mark = svc->sched_data;
145 struct list_head *p; 145 struct list_head *p;
146 146
147 IP_VS_DBG(6, "ip_vs_wrr_schedule(): Scheduling...\n"); 147 IP_VS_DBG(6, "%s(): Scheduling...\n", __func__);
148 148
149 /* 149 /*
150 * This loop will always terminate, because mark->cw in (0, max_weight] 150 * This loop will always terminate, because mark->cw in (0, max_weight]
diff --git a/net/netfilter/ipvs/ip_vs_xmit.c b/net/netfilter/ipvs/ip_vs_xmit.c
index 061e76dfdad9..30b3189bd29c 100644
--- a/net/netfilter/ipvs/ip_vs_xmit.c
+++ b/net/netfilter/ipvs/ip_vs_xmit.c
@@ -238,8 +238,8 @@ ip_vs_bypass_xmit(struct sk_buff *skb, struct ip_vs_conn *cp,
238 EnterFunction(10); 238 EnterFunction(10);
239 239
240 if (ip_route_output_key(&init_net, &rt, &fl)) { 240 if (ip_route_output_key(&init_net, &rt, &fl)) {
241 IP_VS_DBG_RL("ip_vs_bypass_xmit(): ip_route_output error, dest: %pI4\n", 241 IP_VS_DBG_RL("%s(): ip_route_output error, dest: %pI4\n",
242 &iph->daddr); 242 __func__, &iph->daddr);
243 goto tx_error_icmp; 243 goto tx_error_icmp;
244 } 244 }
245 245
@@ -248,7 +248,7 @@ ip_vs_bypass_xmit(struct sk_buff *skb, struct ip_vs_conn *cp,
248 if ((skb->len > mtu) && (iph->frag_off & htons(IP_DF))) { 248 if ((skb->len > mtu) && (iph->frag_off & htons(IP_DF))) {
249 ip_rt_put(rt); 249 ip_rt_put(rt);
250 icmp_send(skb, ICMP_DEST_UNREACH,ICMP_FRAG_NEEDED, htonl(mtu)); 250 icmp_send(skb, ICMP_DEST_UNREACH,ICMP_FRAG_NEEDED, htonl(mtu));
251 IP_VS_DBG_RL("ip_vs_bypass_xmit(): frag needed\n"); 251 IP_VS_DBG_RL("%s(): frag needed\n", __func__);
252 goto tx_error; 252 goto tx_error;
253 } 253 }
254 254
@@ -302,8 +302,8 @@ ip_vs_bypass_xmit_v6(struct sk_buff *skb, struct ip_vs_conn *cp,
302 302
303 rt = (struct rt6_info *)ip6_route_output(&init_net, NULL, &fl); 303 rt = (struct rt6_info *)ip6_route_output(&init_net, NULL, &fl);
304 if (!rt) { 304 if (!rt) {
305 IP_VS_DBG_RL("ip_vs_bypass_xmit_v6(): ip6_route_output error, dest: %pI6\n", 305 IP_VS_DBG_RL("%s(): ip6_route_output error, dest: %pI6\n",
306 &iph->daddr); 306 __func__, &iph->daddr);
307 goto tx_error_icmp; 307 goto tx_error_icmp;
308 } 308 }
309 309
@@ -312,7 +312,7 @@ ip_vs_bypass_xmit_v6(struct sk_buff *skb, struct ip_vs_conn *cp,
312 if (skb->len > mtu) { 312 if (skb->len > mtu) {
313 dst_release(&rt->u.dst); 313 dst_release(&rt->u.dst);
314 icmpv6_send(skb, ICMPV6_PKT_TOOBIG, 0, mtu, skb->dev); 314 icmpv6_send(skb, ICMPV6_PKT_TOOBIG, 0, mtu, skb->dev);
315 IP_VS_DBG_RL("ip_vs_bypass_xmit_v6(): frag needed\n"); 315 IP_VS_DBG_RL("%s(): frag needed\n", __func__);
316 goto tx_error; 316 goto tx_error;
317 } 317 }
318 318
@@ -539,9 +539,9 @@ ip_vs_tunnel_xmit(struct sk_buff *skb, struct ip_vs_conn *cp,
539 EnterFunction(10); 539 EnterFunction(10);
540 540
541 if (skb->protocol != htons(ETH_P_IP)) { 541 if (skb->protocol != htons(ETH_P_IP)) {
542 IP_VS_DBG_RL("ip_vs_tunnel_xmit(): protocol error, " 542 IP_VS_DBG_RL("%s(): protocol error, "
543 "ETH_P_IP: %d, skb protocol: %d\n", 543 "ETH_P_IP: %d, skb protocol: %d\n",
544 htons(ETH_P_IP), skb->protocol); 544 __func__, htons(ETH_P_IP), skb->protocol);
545 goto tx_error; 545 goto tx_error;
546 } 546 }
547 547
@@ -553,7 +553,7 @@ ip_vs_tunnel_xmit(struct sk_buff *skb, struct ip_vs_conn *cp,
553 mtu = dst_mtu(&rt->u.dst) - sizeof(struct iphdr); 553 mtu = dst_mtu(&rt->u.dst) - sizeof(struct iphdr);
554 if (mtu < 68) { 554 if (mtu < 68) {
555 ip_rt_put(rt); 555 ip_rt_put(rt);
556 IP_VS_DBG_RL("ip_vs_tunnel_xmit(): mtu less than 68\n"); 556 IP_VS_DBG_RL("%s(): mtu less than 68\n", __func__);
557 goto tx_error; 557 goto tx_error;
558 } 558 }
559 if (skb_dst(skb)) 559 if (skb_dst(skb))
@@ -565,7 +565,7 @@ ip_vs_tunnel_xmit(struct sk_buff *skb, struct ip_vs_conn *cp,
565 && mtu < ntohs(old_iph->tot_len)) { 565 && mtu < ntohs(old_iph->tot_len)) {
566 icmp_send(skb, ICMP_DEST_UNREACH,ICMP_FRAG_NEEDED, htonl(mtu)); 566 icmp_send(skb, ICMP_DEST_UNREACH,ICMP_FRAG_NEEDED, htonl(mtu));
567 ip_rt_put(rt); 567 ip_rt_put(rt);
568 IP_VS_DBG_RL("ip_vs_tunnel_xmit(): frag needed\n"); 568 IP_VS_DBG_RL("%s(): frag needed\n", __func__);
569 goto tx_error; 569 goto tx_error;
570 } 570 }
571 571
@@ -581,7 +581,7 @@ ip_vs_tunnel_xmit(struct sk_buff *skb, struct ip_vs_conn *cp,
581 if (!new_skb) { 581 if (!new_skb) {
582 ip_rt_put(rt); 582 ip_rt_put(rt);
583 kfree_skb(skb); 583 kfree_skb(skb);
584 IP_VS_ERR_RL("ip_vs_tunnel_xmit(): no memory\n"); 584 IP_VS_ERR_RL("%s(): no memory\n", __func__);
585 return NF_STOLEN; 585 return NF_STOLEN;
586 } 586 }
587 kfree_skb(skb); 587 kfree_skb(skb);
@@ -649,9 +649,9 @@ ip_vs_tunnel_xmit_v6(struct sk_buff *skb, struct ip_vs_conn *cp,
649 EnterFunction(10); 649 EnterFunction(10);
650 650
651 if (skb->protocol != htons(ETH_P_IPV6)) { 651 if (skb->protocol != htons(ETH_P_IPV6)) {
652 IP_VS_DBG_RL("ip_vs_tunnel_xmit_v6(): protocol error, " 652 IP_VS_DBG_RL("%s(): protocol error, "
653 "ETH_P_IPV6: %d, skb protocol: %d\n", 653 "ETH_P_IPV6: %d, skb protocol: %d\n",
654 htons(ETH_P_IPV6), skb->protocol); 654 __func__, htons(ETH_P_IPV6), skb->protocol);
655 goto tx_error; 655 goto tx_error;
656 } 656 }
657 657
@@ -665,7 +665,7 @@ ip_vs_tunnel_xmit_v6(struct sk_buff *skb, struct ip_vs_conn *cp,
665 /* TODO IPv6: do we need this check in IPv6? */ 665 /* TODO IPv6: do we need this check in IPv6? */
666 if (mtu < 1280) { 666 if (mtu < 1280) {
667 dst_release(&rt->u.dst); 667 dst_release(&rt->u.dst);
668 IP_VS_DBG_RL("ip_vs_tunnel_xmit_v6(): mtu less than 1280\n"); 668 IP_VS_DBG_RL("%s(): mtu less than 1280\n", __func__);
669 goto tx_error; 669 goto tx_error;
670 } 670 }
671 if (skb_dst(skb)) 671 if (skb_dst(skb))
@@ -674,7 +674,7 @@ ip_vs_tunnel_xmit_v6(struct sk_buff *skb, struct ip_vs_conn *cp,
674 if (mtu < ntohs(old_iph->payload_len) + sizeof(struct ipv6hdr)) { 674 if (mtu < ntohs(old_iph->payload_len) + sizeof(struct ipv6hdr)) {
675 icmpv6_send(skb, ICMPV6_PKT_TOOBIG, 0, mtu, skb->dev); 675 icmpv6_send(skb, ICMPV6_PKT_TOOBIG, 0, mtu, skb->dev);
676 dst_release(&rt->u.dst); 676 dst_release(&rt->u.dst);
677 IP_VS_DBG_RL("ip_vs_tunnel_xmit_v6(): frag needed\n"); 677 IP_VS_DBG_RL("%s(): frag needed\n", __func__);
678 goto tx_error; 678 goto tx_error;
679 } 679 }
680 680
@@ -690,7 +690,7 @@ ip_vs_tunnel_xmit_v6(struct sk_buff *skb, struct ip_vs_conn *cp,
690 if (!new_skb) { 690 if (!new_skb) {
691 dst_release(&rt->u.dst); 691 dst_release(&rt->u.dst);
692 kfree_skb(skb); 692 kfree_skb(skb);
693 IP_VS_ERR_RL("ip_vs_tunnel_xmit_v6(): no memory\n"); 693 IP_VS_ERR_RL("%s(): no memory\n", __func__);
694 return NF_STOLEN; 694 return NF_STOLEN;
695 } 695 }
696 kfree_skb(skb); 696 kfree_skb(skb);
@@ -763,7 +763,7 @@ ip_vs_dr_xmit(struct sk_buff *skb, struct ip_vs_conn *cp,
763 if ((iph->frag_off & htons(IP_DF)) && skb->len > mtu) { 763 if ((iph->frag_off & htons(IP_DF)) && skb->len > mtu) {
764 icmp_send(skb, ICMP_DEST_UNREACH,ICMP_FRAG_NEEDED, htonl(mtu)); 764 icmp_send(skb, ICMP_DEST_UNREACH,ICMP_FRAG_NEEDED, htonl(mtu));
765 ip_rt_put(rt); 765 ip_rt_put(rt);
766 IP_VS_DBG_RL("ip_vs_dr_xmit(): frag needed\n"); 766 IP_VS_DBG_RL("%s(): frag needed\n", __func__);
767 goto tx_error; 767 goto tx_error;
768 } 768 }
769 769
@@ -816,7 +816,7 @@ ip_vs_dr_xmit_v6(struct sk_buff *skb, struct ip_vs_conn *cp,
816 if (skb->len > mtu) { 816 if (skb->len > mtu) {
817 icmpv6_send(skb, ICMPV6_PKT_TOOBIG, 0, mtu, skb->dev); 817 icmpv6_send(skb, ICMPV6_PKT_TOOBIG, 0, mtu, skb->dev);
818 dst_release(&rt->u.dst); 818 dst_release(&rt->u.dst);
819 IP_VS_DBG_RL("ip_vs_dr_xmit_v6(): frag needed\n"); 819 IP_VS_DBG_RL("%s(): frag needed\n", __func__);
820 goto tx_error; 820 goto tx_error;
821 } 821 }
822 822
@@ -891,7 +891,7 @@ ip_vs_icmp_xmit(struct sk_buff *skb, struct ip_vs_conn *cp,
891 if ((skb->len > mtu) && (ip_hdr(skb)->frag_off & htons(IP_DF))) { 891 if ((skb->len > mtu) && (ip_hdr(skb)->frag_off & htons(IP_DF))) {
892 ip_rt_put(rt); 892 ip_rt_put(rt);
893 icmp_send(skb, ICMP_DEST_UNREACH, ICMP_FRAG_NEEDED, htonl(mtu)); 893 icmp_send(skb, ICMP_DEST_UNREACH, ICMP_FRAG_NEEDED, htonl(mtu));
894 IP_VS_DBG_RL("ip_vs_in_icmp(): frag needed\n"); 894 IP_VS_DBG_RL("%s(): frag needed\n", __func__);
895 goto tx_error; 895 goto tx_error;
896 } 896 }
897 897
@@ -966,7 +966,7 @@ ip_vs_icmp_xmit_v6(struct sk_buff *skb, struct ip_vs_conn *cp,
966 if (skb->len > mtu) { 966 if (skb->len > mtu) {
967 dst_release(&rt->u.dst); 967 dst_release(&rt->u.dst);
968 icmpv6_send(skb, ICMPV6_PKT_TOOBIG, 0, mtu, skb->dev); 968 icmpv6_send(skb, ICMPV6_PKT_TOOBIG, 0, mtu, skb->dev);
969 IP_VS_DBG_RL("ip_vs_in_icmp(): frag needed\n"); 969 IP_VS_DBG_RL("%s(): frag needed\n", __func__);
970 goto tx_error; 970 goto tx_error;
971 } 971 }
972 972