diff options
Diffstat (limited to 'drivers/s390/net')
33 files changed, 20316 insertions, 16798 deletions
diff --git a/drivers/s390/net/Kconfig b/drivers/s390/net/Kconfig index eada69dec4fe..a7745c82b4ae 100644 --- a/drivers/s390/net/Kconfig +++ b/drivers/s390/net/Kconfig | |||
@@ -5,22 +5,25 @@ config LCS | |||
5 | tristate "Lan Channel Station Interface" | 5 | tristate "Lan Channel Station Interface" |
6 | depends on CCW && NETDEVICES && (NET_ETHERNET || TR || FDDI) | 6 | depends on CCW && NETDEVICES && (NET_ETHERNET || TR || FDDI) |
7 | help | 7 | help |
8 | Select this option if you want to use LCS networking on IBM S/390 | 8 | Select this option if you want to use LCS networking on IBM System z. |
9 | or zSeries. This device driver supports Token Ring (IEEE 802.5), | 9 | This device driver supports Token Ring (IEEE 802.5), |
10 | FDDI (IEEE 802.7) and Ethernet. | 10 | FDDI (IEEE 802.7) and Ethernet. |
11 | This option is also available as a module which will be | 11 | To compile as a module, choose M. The module name is lcs.ko. |
12 | called lcs.ko. If you do not know what it is, it's safe to say "Y". | 12 | If you do not know what it is, it's safe to choose Y. |
13 | 13 | ||
14 | config CTC | 14 | config CTCM |
15 | tristate "CTC device support" | 15 | tristate "CTC and MPC SNA device support" |
16 | depends on CCW && NETDEVICES | 16 | depends on CCW && NETDEVICES |
17 | help | 17 | help |
18 | Select this option if you want to use channel-to-channel networking | 18 | Select this option if you want to use channel-to-channel |
19 | on IBM S/390 or zSeries. This device driver supports real CTC | 19 | point-to-point networking on IBM System z. |
20 | coupling using ESCON. It also supports virtual CTCs when running | 20 | This device driver supports real CTC coupling using ESCON. |
21 | under VM. It will use the channel device configuration if this is | 21 | It also supports virtual CTCs when running under VM. |
22 | available. This option is also available as a module which will be | 22 | This driver also supports channel-to-channel MPC SNA devices. |
23 | called ctc.ko. If you do not know what it is, it's safe to say "Y". | 23 | MPC is an SNA protocol device used by Communication Server for Linux. |
24 | To compile as a module, choose M. The module name is ctcm.ko. | ||
25 | To compile into the kernel, choose Y. | ||
26 | If you do not need any channel-to-channel connection, choose N. | ||
24 | 27 | ||
25 | config NETIUCV | 28 | config NETIUCV |
26 | tristate "IUCV network device support (VM only)" | 29 | tristate "IUCV network device support (VM only)" |
@@ -29,9 +32,9 @@ config NETIUCV | |||
29 | Select this option if you want to use inter-user communication | 32 | Select this option if you want to use inter-user communication |
30 | vehicle networking under VM or VIF. It enables a fast communication | 33 | vehicle networking under VM or VIF. It enables a fast communication |
31 | link between VM guests. Using ifconfig a point-to-point connection | 34 | link between VM guests. Using ifconfig a point-to-point connection |
32 | can be established to the Linux for zSeries and S7390 system | 35 | can be established to the Linux on IBM System z |
33 | running on the other VM guest. This option is also available | 36 | running on the other VM guest. To compile as a module, choose M. |
34 | as a module which will be called netiucv.ko. If unsure, say "Y". | 37 | The module name is netiucv.ko. If unsure, choose Y. |
35 | 38 | ||
36 | config SMSGIUCV | 39 | config SMSGIUCV |
37 | tristate "IUCV special message support (VM only)" | 40 | tristate "IUCV special message support (VM only)" |
@@ -47,43 +50,46 @@ config CLAW | |||
47 | This driver supports channel attached CLAW devices. | 50 | This driver supports channel attached CLAW devices. |
48 | CLAW is Common Link Access for Workstation. Common devices | 51 | CLAW is Common Link Access for Workstation. Common devices |
49 | that use CLAW are RS/6000s, Cisco Routers (CIP) and 3172 devices. | 52 | that use CLAW are RS/6000s, Cisco Routers (CIP) and 3172 devices. |
50 | To compile as a module choose M here: The module will be called | 53 | To compile as a module, choose M. The module name is claw.ko. |
51 | claw.ko to compile into the kernel choose Y | 54 | To compile into the kernel, choose Y. |
52 | 55 | ||
53 | config QETH | 56 | config QETH |
54 | tristate "Gigabit Ethernet device support" | 57 | tristate "Gigabit Ethernet device support" |
55 | depends on CCW && NETDEVICES && IP_MULTICAST && QDIO | 58 | depends on CCW && NETDEVICES && IP_MULTICAST && QDIO |
56 | help | 59 | help |
57 | This driver supports the IBM S/390 and zSeries OSA Express adapters | 60 | This driver supports the IBM System z OSA Express adapters |
58 | in QDIO mode (all media types), HiperSockets interfaces and VM GuestLAN | 61 | in QDIO mode (all media types), HiperSockets interfaces and VM GuestLAN |
59 | interfaces in QDIO and HIPER mode. | 62 | interfaces in QDIO and HIPER mode. |
60 | 63 | ||
61 | For details please refer to the documentation provided by IBM at | 64 | For details please refer to the documentation provided by IBM at |
62 | <http://www10.software.ibm.com/developerworks/opensource/linux390> | 65 | <http://www.ibm.com/developerworks/linux/linux390> |
63 | 66 | ||
64 | To compile this driver as a module, choose M here: the | 67 | To compile this driver as a module, choose M. |
65 | module will be called qeth.ko. | 68 | The module name is qeth.ko. |
66 | 69 | ||
70 | config QETH_L2 | ||
71 | tristate "qeth layer 2 device support" | ||
72 | depends on QETH | ||
73 | help | ||
74 | Select this option to be able to run qeth devices in layer 2 mode. | ||
75 | To compile as a module, choose M. The module name is qeth_l2.ko. | ||
76 | If unsure, choose y. | ||
67 | 77 | ||
68 | comment "Gigabit Ethernet default settings" | 78 | config QETH_L3 |
69 | depends on QETH | 79 | tristate "qeth layer 3 device support" |
80 | depends on QETH | ||
81 | help | ||
82 | Select this option to be able to run qeth devices in layer 3 mode. | ||
83 | To compile as a module choose M. The module name is qeth_l3.ko. | ||
84 | If unsure, choose Y. | ||
70 | 85 | ||
71 | config QETH_IPV6 | 86 | config QETH_IPV6 |
72 | bool "IPv6 support for gigabit ethernet" | 87 | bool |
73 | depends on (QETH = IPV6) || (QETH && IPV6 = 'y') | 88 | depends on (QETH_L3 = IPV6) || (QETH_L3 && IPV6 = 'y') |
74 | help | 89 | default y |
75 | If CONFIG_QETH is switched on, this option will include IPv6 | ||
76 | support in the qeth device driver. | ||
77 | |||
78 | config QETH_VLAN | ||
79 | bool "VLAN support for gigabit ethernet" | ||
80 | depends on (QETH = VLAN_8021Q) || (QETH && VLAN_8021Q = 'y') | ||
81 | help | ||
82 | If CONFIG_QETH is switched on, this option will include IEEE | ||
83 | 802.1q VLAN support in the qeth device driver. | ||
84 | 90 | ||
85 | config CCWGROUP | 91 | config CCWGROUP |
86 | tristate | 92 | tristate |
87 | default (LCS || CTC || QETH) | 93 | default (LCS || CTCM || QETH) |
88 | 94 | ||
89 | endmenu | 95 | endmenu |
diff --git a/drivers/s390/net/Makefile b/drivers/s390/net/Makefile index bbe3ab2e93d9..6382c04d2bdf 100644 --- a/drivers/s390/net/Makefile +++ b/drivers/s390/net/Makefile | |||
@@ -2,13 +2,15 @@ | |||
2 | # S/390 network devices | 2 | # S/390 network devices |
3 | # | 3 | # |
4 | 4 | ||
5 | ctc-objs := ctcmain.o ctcdbug.o | 5 | ctcm-y += ctcm_main.o ctcm_fsms.o ctcm_mpc.o ctcm_sysfs.o ctcm_dbug.o |
6 | 6 | obj-$(CONFIG_CTCM) += ctcm.o fsm.o cu3088.o | |
7 | obj-$(CONFIG_NETIUCV) += netiucv.o fsm.o | 7 | obj-$(CONFIG_NETIUCV) += netiucv.o fsm.o |
8 | obj-$(CONFIG_SMSGIUCV) += smsgiucv.o | 8 | obj-$(CONFIG_SMSGIUCV) += smsgiucv.o |
9 | obj-$(CONFIG_CTC) += ctc.o fsm.o cu3088.o | ||
10 | obj-$(CONFIG_LCS) += lcs.o cu3088.o | 9 | obj-$(CONFIG_LCS) += lcs.o cu3088.o |
11 | obj-$(CONFIG_CLAW) += claw.o cu3088.o | 10 | obj-$(CONFIG_CLAW) += claw.o cu3088.o |
12 | qeth-y := qeth_main.o qeth_mpc.o qeth_sys.o qeth_eddp.o | 11 | qeth-y += qeth_core_sys.o qeth_core_main.o qeth_core_mpc.o qeth_core_offl.o |
13 | qeth-$(CONFIG_PROC_FS) += qeth_proc.o | ||
14 | obj-$(CONFIG_QETH) += qeth.o | 12 | obj-$(CONFIG_QETH) += qeth.o |
13 | qeth_l2-y += qeth_l2_main.o | ||
14 | obj-$(CONFIG_QETH_L2) += qeth_l2.o | ||
15 | qeth_l3-y += qeth_l3_main.o qeth_l3_sys.o | ||
16 | obj-$(CONFIG_QETH_L3) += qeth_l3.o | ||
diff --git a/drivers/s390/net/ctcdbug.c b/drivers/s390/net/ctcdbug.c deleted file mode 100644 index e6e72deb36b5..000000000000 --- a/drivers/s390/net/ctcdbug.c +++ /dev/null | |||
@@ -1,80 +0,0 @@ | |||
1 | /* | ||
2 | * | ||
3 | * linux/drivers/s390/net/ctcdbug.c | ||
4 | * | ||
5 | * CTC / ESCON network driver - s390 dbf exploit. | ||
6 | * | ||
7 | * Copyright 2000,2003 IBM Corporation | ||
8 | * | ||
9 | * Author(s): Original Code written by | ||
10 | * Peter Tiedemann (ptiedem@de.ibm.com) | ||
11 | * | ||
12 | * This program is free software; you can redistribute it and/or modify | ||
13 | * it under the terms of the GNU General Public License as published by | ||
14 | * the Free Software Foundation; either version 2, or (at your option) | ||
15 | * any later version. | ||
16 | * | ||
17 | * This program is distributed in the hope that it will be useful, | ||
18 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | ||
19 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | ||
20 | * GNU General Public License for more details. | ||
21 | * | ||
22 | * You should have received a copy of the GNU General Public License | ||
23 | * along with this program; if not, write to the Free Software | ||
24 | * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. | ||
25 | */ | ||
26 | |||
27 | #include "ctcdbug.h" | ||
28 | |||
29 | /** | ||
30 | * Debug Facility Stuff | ||
31 | */ | ||
32 | debug_info_t *ctc_dbf_setup = NULL; | ||
33 | debug_info_t *ctc_dbf_data = NULL; | ||
34 | debug_info_t *ctc_dbf_trace = NULL; | ||
35 | |||
36 | DEFINE_PER_CPU(char[256], ctc_dbf_txt_buf); | ||
37 | |||
38 | void | ||
39 | ctc_unregister_dbf_views(void) | ||
40 | { | ||
41 | if (ctc_dbf_setup) | ||
42 | debug_unregister(ctc_dbf_setup); | ||
43 | if (ctc_dbf_data) | ||
44 | debug_unregister(ctc_dbf_data); | ||
45 | if (ctc_dbf_trace) | ||
46 | debug_unregister(ctc_dbf_trace); | ||
47 | } | ||
48 | int | ||
49 | ctc_register_dbf_views(void) | ||
50 | { | ||
51 | ctc_dbf_setup = debug_register(CTC_DBF_SETUP_NAME, | ||
52 | CTC_DBF_SETUP_PAGES, | ||
53 | CTC_DBF_SETUP_NR_AREAS, | ||
54 | CTC_DBF_SETUP_LEN); | ||
55 | ctc_dbf_data = debug_register(CTC_DBF_DATA_NAME, | ||
56 | CTC_DBF_DATA_PAGES, | ||
57 | CTC_DBF_DATA_NR_AREAS, | ||
58 | CTC_DBF_DATA_LEN); | ||
59 | ctc_dbf_trace = debug_register(CTC_DBF_TRACE_NAME, | ||
60 | CTC_DBF_TRACE_PAGES, | ||
61 | CTC_DBF_TRACE_NR_AREAS, | ||
62 | CTC_DBF_TRACE_LEN); | ||
63 | |||
64 | if ((ctc_dbf_setup == NULL) || (ctc_dbf_data == NULL) || | ||
65 | (ctc_dbf_trace == NULL)) { | ||
66 | ctc_unregister_dbf_views(); | ||
67 | return -ENOMEM; | ||
68 | } | ||
69 | debug_register_view(ctc_dbf_setup, &debug_hex_ascii_view); | ||
70 | debug_set_level(ctc_dbf_setup, CTC_DBF_SETUP_LEVEL); | ||
71 | |||
72 | debug_register_view(ctc_dbf_data, &debug_hex_ascii_view); | ||
73 | debug_set_level(ctc_dbf_data, CTC_DBF_DATA_LEVEL); | ||
74 | |||
75 | debug_register_view(ctc_dbf_trace, &debug_hex_ascii_view); | ||
76 | debug_set_level(ctc_dbf_trace, CTC_DBF_TRACE_LEVEL); | ||
77 | |||
78 | return 0; | ||
79 | } | ||
80 | |||
diff --git a/drivers/s390/net/ctcdbug.h b/drivers/s390/net/ctcdbug.h deleted file mode 100644 index 413925ee23d1..000000000000 --- a/drivers/s390/net/ctcdbug.h +++ /dev/null | |||
@@ -1,125 +0,0 @@ | |||
1 | /* | ||
2 | * | ||
3 | * linux/drivers/s390/net/ctcdbug.h | ||
4 | * | ||
5 | * CTC / ESCON network driver - s390 dbf exploit. | ||
6 | * | ||
7 | * Copyright 2000,2003 IBM Corporation | ||
8 | * | ||
9 | * Author(s): Original Code written by | ||
10 | * Peter Tiedemann (ptiedem@de.ibm.com) | ||
11 | * | ||
12 | * This program is free software; you can redistribute it and/or modify | ||
13 | * it under the terms of the GNU General Public License as published by | ||
14 | * the Free Software Foundation; either version 2, or (at your option) | ||
15 | * any later version. | ||
16 | * | ||
17 | * This program is distributed in the hope that it will be useful, | ||
18 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | ||
19 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | ||
20 | * GNU General Public License for more details. | ||
21 | * | ||
22 | * You should have received a copy of the GNU General Public License | ||
23 | * along with this program; if not, write to the Free Software | ||
24 | * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. | ||
25 | */ | ||
26 | #ifndef _CTCDBUG_H_ | ||
27 | #define _CTCDBUG_H_ | ||
28 | |||
29 | #include <asm/debug.h> | ||
30 | #include "ctcmain.h" | ||
31 | /** | ||
32 | * Debug Facility stuff | ||
33 | */ | ||
34 | #define CTC_DBF_SETUP_NAME "ctc_setup" | ||
35 | #define CTC_DBF_SETUP_LEN 16 | ||
36 | #define CTC_DBF_SETUP_PAGES 8 | ||
37 | #define CTC_DBF_SETUP_NR_AREAS 1 | ||
38 | #define CTC_DBF_SETUP_LEVEL 3 | ||
39 | |||
40 | #define CTC_DBF_DATA_NAME "ctc_data" | ||
41 | #define CTC_DBF_DATA_LEN 128 | ||
42 | #define CTC_DBF_DATA_PAGES 8 | ||
43 | #define CTC_DBF_DATA_NR_AREAS 1 | ||
44 | #define CTC_DBF_DATA_LEVEL 3 | ||
45 | |||
46 | #define CTC_DBF_TRACE_NAME "ctc_trace" | ||
47 | #define CTC_DBF_TRACE_LEN 16 | ||
48 | #define CTC_DBF_TRACE_PAGES 4 | ||
49 | #define CTC_DBF_TRACE_NR_AREAS 2 | ||
50 | #define CTC_DBF_TRACE_LEVEL 3 | ||
51 | |||
52 | #define DBF_TEXT(name,level,text) \ | ||
53 | do { \ | ||
54 | debug_text_event(ctc_dbf_##name,level,text); \ | ||
55 | } while (0) | ||
56 | |||
57 | #define DBF_HEX(name,level,addr,len) \ | ||
58 | do { \ | ||
59 | debug_event(ctc_dbf_##name,level,(void*)(addr),len); \ | ||
60 | } while (0) | ||
61 | |||
62 | DECLARE_PER_CPU(char[256], ctc_dbf_txt_buf); | ||
63 | extern debug_info_t *ctc_dbf_setup; | ||
64 | extern debug_info_t *ctc_dbf_data; | ||
65 | extern debug_info_t *ctc_dbf_trace; | ||
66 | |||
67 | |||
68 | #define DBF_TEXT_(name,level,text...) \ | ||
69 | do { \ | ||
70 | char* ctc_dbf_txt_buf = get_cpu_var(ctc_dbf_txt_buf); \ | ||
71 | sprintf(ctc_dbf_txt_buf, text); \ | ||
72 | debug_text_event(ctc_dbf_##name,level,ctc_dbf_txt_buf); \ | ||
73 | put_cpu_var(ctc_dbf_txt_buf); \ | ||
74 | } while (0) | ||
75 | |||
76 | #define DBF_SPRINTF(name,level,text...) \ | ||
77 | do { \ | ||
78 | debug_sprintf_event(ctc_dbf_trace, level, ##text ); \ | ||
79 | debug_sprintf_event(ctc_dbf_trace, level, text ); \ | ||
80 | } while (0) | ||
81 | |||
82 | |||
83 | int ctc_register_dbf_views(void); | ||
84 | |||
85 | void ctc_unregister_dbf_views(void); | ||
86 | |||
87 | /** | ||
88 | * some more debug stuff | ||
89 | */ | ||
90 | |||
91 | #define HEXDUMP16(importance,header,ptr) \ | ||
92 | PRINT_##importance(header "%02x %02x %02x %02x %02x %02x %02x %02x " \ | ||
93 | "%02x %02x %02x %02x %02x %02x %02x %02x\n", \ | ||
94 | *(((char*)ptr)),*(((char*)ptr)+1),*(((char*)ptr)+2), \ | ||
95 | *(((char*)ptr)+3),*(((char*)ptr)+4),*(((char*)ptr)+5), \ | ||
96 | *(((char*)ptr)+6),*(((char*)ptr)+7),*(((char*)ptr)+8), \ | ||
97 | *(((char*)ptr)+9),*(((char*)ptr)+10),*(((char*)ptr)+11), \ | ||
98 | *(((char*)ptr)+12),*(((char*)ptr)+13), \ | ||
99 | *(((char*)ptr)+14),*(((char*)ptr)+15)); \ | ||
100 | PRINT_##importance(header "%02x %02x %02x %02x %02x %02x %02x %02x " \ | ||
101 | "%02x %02x %02x %02x %02x %02x %02x %02x\n", \ | ||
102 | *(((char*)ptr)+16),*(((char*)ptr)+17), \ | ||
103 | *(((char*)ptr)+18),*(((char*)ptr)+19), \ | ||
104 | *(((char*)ptr)+20),*(((char*)ptr)+21), \ | ||
105 | *(((char*)ptr)+22),*(((char*)ptr)+23), \ | ||
106 | *(((char*)ptr)+24),*(((char*)ptr)+25), \ | ||
107 | *(((char*)ptr)+26),*(((char*)ptr)+27), \ | ||
108 | *(((char*)ptr)+28),*(((char*)ptr)+29), \ | ||
109 | *(((char*)ptr)+30),*(((char*)ptr)+31)); | ||
110 | |||
111 | static inline void | ||
112 | hex_dump(unsigned char *buf, size_t len) | ||
113 | { | ||
114 | size_t i; | ||
115 | |||
116 | for (i = 0; i < len; i++) { | ||
117 | if (i && !(i % 16)) | ||
118 | printk("\n"); | ||
119 | printk("%02x ", *(buf + i)); | ||
120 | } | ||
121 | printk("\n"); | ||
122 | } | ||
123 | |||
124 | |||
125 | #endif | ||
diff --git a/drivers/s390/net/ctcm_dbug.c b/drivers/s390/net/ctcm_dbug.c new file mode 100644 index 000000000000..8eb25d00b2e7 --- /dev/null +++ b/drivers/s390/net/ctcm_dbug.c | |||
@@ -0,0 +1,67 @@ | |||
1 | /* | ||
2 | * drivers/s390/net/ctcm_dbug.c | ||
3 | * | ||
4 | * Copyright IBM Corp. 2001, 2007 | ||
5 | * Authors: Peter Tiedemann (ptiedem@de.ibm.com) | ||
6 | * | ||
7 | */ | ||
8 | |||
9 | #include <linux/stddef.h> | ||
10 | #include <linux/kernel.h> | ||
11 | #include <linux/errno.h> | ||
12 | #include <linux/slab.h> | ||
13 | #include <linux/ctype.h> | ||
14 | #include <linux/sysctl.h> | ||
15 | #include <linux/module.h> | ||
16 | #include <linux/init.h> | ||
17 | #include <linux/fs.h> | ||
18 | #include <linux/debugfs.h> | ||
19 | #include "ctcm_dbug.h" | ||
20 | |||
21 | /* | ||
22 | * Debug Facility Stuff | ||
23 | */ | ||
24 | |||
25 | DEFINE_PER_CPU(char[256], ctcm_dbf_txt_buf); | ||
26 | |||
27 | struct ctcm_dbf_info ctcm_dbf[CTCM_DBF_INFOS] = { | ||
28 | [CTCM_DBF_SETUP] = {"ctc_setup", 8, 1, 64, 5, NULL}, | ||
29 | [CTCM_DBF_ERROR] = {"ctc_error", 8, 1, 64, 3, NULL}, | ||
30 | [CTCM_DBF_TRACE] = {"ctc_trace", 8, 1, 64, 3, NULL}, | ||
31 | [CTCM_DBF_MPC_SETUP] = {"mpc_setup", 8, 1, 64, 5, NULL}, | ||
32 | [CTCM_DBF_MPC_ERROR] = {"mpc_error", 8, 1, 64, 3, NULL}, | ||
33 | [CTCM_DBF_MPC_TRACE] = {"mpc_trace", 8, 1, 64, 3, NULL}, | ||
34 | }; | ||
35 | |||
36 | void ctcm_unregister_dbf_views(void) | ||
37 | { | ||
38 | int x; | ||
39 | for (x = 0; x < CTCM_DBF_INFOS; x++) { | ||
40 | debug_unregister(ctcm_dbf[x].id); | ||
41 | ctcm_dbf[x].id = NULL; | ||
42 | } | ||
43 | } | ||
44 | |||
45 | int ctcm_register_dbf_views(void) | ||
46 | { | ||
47 | int x; | ||
48 | for (x = 0; x < CTCM_DBF_INFOS; x++) { | ||
49 | /* register the areas */ | ||
50 | ctcm_dbf[x].id = debug_register(ctcm_dbf[x].name, | ||
51 | ctcm_dbf[x].pages, | ||
52 | ctcm_dbf[x].areas, | ||
53 | ctcm_dbf[x].len); | ||
54 | if (ctcm_dbf[x].id == NULL) { | ||
55 | ctcm_unregister_dbf_views(); | ||
56 | return -ENOMEM; | ||
57 | } | ||
58 | |||
59 | /* register a view */ | ||
60 | debug_register_view(ctcm_dbf[x].id, &debug_hex_ascii_view); | ||
61 | /* set a passing level */ | ||
62 | debug_set_level(ctcm_dbf[x].id, ctcm_dbf[x].level); | ||
63 | } | ||
64 | |||
65 | return 0; | ||
66 | } | ||
67 | |||
diff --git a/drivers/s390/net/ctcm_dbug.h b/drivers/s390/net/ctcm_dbug.h new file mode 100644 index 000000000000..fdff34fe59a2 --- /dev/null +++ b/drivers/s390/net/ctcm_dbug.h | |||
@@ -0,0 +1,158 @@ | |||
1 | /* | ||
2 | * drivers/s390/net/ctcm_dbug.h | ||
3 | * | ||
4 | * Copyright IBM Corp. 2001, 2007 | ||
5 | * Authors: Peter Tiedemann (ptiedem@de.ibm.com) | ||
6 | * | ||
7 | */ | ||
8 | |||
9 | #ifndef _CTCM_DBUG_H_ | ||
10 | #define _CTCM_DBUG_H_ | ||
11 | |||
12 | /* | ||
13 | * Debug Facility stuff | ||
14 | */ | ||
15 | |||
16 | #include <asm/debug.h> | ||
17 | |||
18 | #ifdef DEBUG | ||
19 | #define do_debug 1 | ||
20 | #else | ||
21 | #define do_debug 0 | ||
22 | #endif | ||
23 | #ifdef DEBUGDATA | ||
24 | #define do_debug_data 1 | ||
25 | #else | ||
26 | #define do_debug_data 0 | ||
27 | #endif | ||
28 | #ifdef DEBUGCCW | ||
29 | #define do_debug_ccw 1 | ||
30 | #else | ||
31 | #define do_debug_ccw 0 | ||
32 | #endif | ||
33 | |||
34 | /* define dbf debug levels similar to kernel msg levels */ | ||
35 | #define CTC_DBF_ALWAYS 0 /* always print this */ | ||
36 | #define CTC_DBF_EMERG 0 /* system is unusable */ | ||
37 | #define CTC_DBF_ALERT 1 /* action must be taken immediately */ | ||
38 | #define CTC_DBF_CRIT 2 /* critical conditions */ | ||
39 | #define CTC_DBF_ERROR 3 /* error conditions */ | ||
40 | #define CTC_DBF_WARN 4 /* warning conditions */ | ||
41 | #define CTC_DBF_NOTICE 5 /* normal but significant condition */ | ||
42 | #define CTC_DBF_INFO 5 /* informational */ | ||
43 | #define CTC_DBF_DEBUG 6 /* debug-level messages */ | ||
44 | |||
45 | DECLARE_PER_CPU(char[256], ctcm_dbf_txt_buf); | ||
46 | |||
47 | enum ctcm_dbf_names { | ||
48 | CTCM_DBF_SETUP, | ||
49 | CTCM_DBF_ERROR, | ||
50 | CTCM_DBF_TRACE, | ||
51 | CTCM_DBF_MPC_SETUP, | ||
52 | CTCM_DBF_MPC_ERROR, | ||
53 | CTCM_DBF_MPC_TRACE, | ||
54 | CTCM_DBF_INFOS /* must be last element */ | ||
55 | }; | ||
56 | |||
57 | struct ctcm_dbf_info { | ||
58 | char name[DEBUG_MAX_NAME_LEN]; | ||
59 | int pages; | ||
60 | int areas; | ||
61 | int len; | ||
62 | int level; | ||
63 | debug_info_t *id; | ||
64 | }; | ||
65 | |||
66 | extern struct ctcm_dbf_info ctcm_dbf[CTCM_DBF_INFOS]; | ||
67 | |||
68 | int ctcm_register_dbf_views(void); | ||
69 | void ctcm_unregister_dbf_views(void); | ||
70 | |||
71 | static inline const char *strtail(const char *s, int n) | ||
72 | { | ||
73 | int l = strlen(s); | ||
74 | return (l > n) ? s + (l - n) : s; | ||
75 | } | ||
76 | |||
77 | /* sort out levels early to avoid unnecessary sprintfs */ | ||
78 | static inline int ctcm_dbf_passes(debug_info_t *dbf_grp, int level) | ||
79 | { | ||
80 | return (dbf_grp->level >= level); | ||
81 | } | ||
82 | |||
83 | #define CTCM_FUNTAIL strtail((char *)__func__, 16) | ||
84 | |||
85 | #define CTCM_DBF_TEXT(name, level, text) \ | ||
86 | do { \ | ||
87 | debug_text_event(ctcm_dbf[CTCM_DBF_##name].id, level, text); \ | ||
88 | } while (0) | ||
89 | |||
90 | #define CTCM_DBF_HEX(name, level, addr, len) \ | ||
91 | do { \ | ||
92 | debug_event(ctcm_dbf[CTCM_DBF_##name].id, \ | ||
93 | level, (void *)(addr), len); \ | ||
94 | } while (0) | ||
95 | |||
96 | #define CTCM_DBF_TEXT_(name, level, text...) \ | ||
97 | do { \ | ||
98 | if (ctcm_dbf_passes(ctcm_dbf[CTCM_DBF_##name].id, level)) { \ | ||
99 | char *ctcm_dbf_txt_buf = \ | ||
100 | get_cpu_var(ctcm_dbf_txt_buf); \ | ||
101 | sprintf(ctcm_dbf_txt_buf, text); \ | ||
102 | debug_text_event(ctcm_dbf[CTCM_DBF_##name].id, \ | ||
103 | level, ctcm_dbf_txt_buf); \ | ||
104 | put_cpu_var(ctcm_dbf_txt_buf); \ | ||
105 | } \ | ||
106 | } while (0) | ||
107 | |||
108 | /* | ||
109 | * cat : one of {setup, mpc_setup, trace, mpc_trace, error, mpc_error}. | ||
110 | * dev : netdevice with valid name field. | ||
111 | * text: any text string. | ||
112 | */ | ||
113 | #define CTCM_DBF_DEV_NAME(cat, dev, text) \ | ||
114 | do { \ | ||
115 | CTCM_DBF_TEXT_(cat, CTC_DBF_INFO, "%s(%s) : %s", \ | ||
116 | CTCM_FUNTAIL, dev->name, text); \ | ||
117 | } while (0) | ||
118 | |||
119 | #define MPC_DBF_DEV_NAME(cat, dev, text) \ | ||
120 | do { \ | ||
121 | CTCM_DBF_TEXT_(MPC_##cat, CTC_DBF_INFO, "%s(%s) : %s", \ | ||
122 | CTCM_FUNTAIL, dev->name, text); \ | ||
123 | } while (0) | ||
124 | |||
125 | #define CTCMY_DBF_DEV_NAME(cat, dev, text) \ | ||
126 | do { \ | ||
127 | if (IS_MPCDEV(dev)) \ | ||
128 | MPC_DBF_DEV_NAME(cat, dev, text); \ | ||
129 | else \ | ||
130 | CTCM_DBF_DEV_NAME(cat, dev, text); \ | ||
131 | } while (0) | ||
132 | |||
133 | /* | ||
134 | * cat : one of {setup, mpc_setup, trace, mpc_trace, error, mpc_error}. | ||
135 | * dev : netdevice. | ||
136 | * text: any text string. | ||
137 | */ | ||
138 | #define CTCM_DBF_DEV(cat, dev, text) \ | ||
139 | do { \ | ||
140 | CTCM_DBF_TEXT_(cat, CTC_DBF_INFO, "%s(%p) : %s", \ | ||
141 | CTCM_FUNTAIL, dev, text); \ | ||
142 | } while (0) | ||
143 | |||
144 | #define MPC_DBF_DEV(cat, dev, text) \ | ||
145 | do { \ | ||
146 | CTCM_DBF_TEXT_(MPC_##cat, CTC_DBF_INFO, "%s(%p) : %s", \ | ||
147 | CTCM_FUNTAIL, dev, text); \ | ||
148 | } while (0) | ||
149 | |||
150 | #define CTCMY_DBF_DEV(cat, dev, text) \ | ||
151 | do { \ | ||
152 | if (IS_MPCDEV(dev)) \ | ||
153 | MPC_DBF_DEV(cat, dev, text); \ | ||
154 | else \ | ||
155 | CTCM_DBF_DEV(cat, dev, text); \ | ||
156 | } while (0) | ||
157 | |||
158 | #endif | ||
diff --git a/drivers/s390/net/ctcm_fsms.c b/drivers/s390/net/ctcm_fsms.c new file mode 100644 index 000000000000..2a106f3a076d --- /dev/null +++ b/drivers/s390/net/ctcm_fsms.c | |||
@@ -0,0 +1,2347 @@ | |||
1 | /* | ||
2 | * drivers/s390/net/ctcm_fsms.c | ||
3 | * | ||
4 | * Copyright IBM Corp. 2001, 2007 | ||
5 | * Authors: Fritz Elfert (felfert@millenux.com) | ||
6 | * Peter Tiedemann (ptiedem@de.ibm.com) | ||
7 | * MPC additions : | ||
8 | * Belinda Thompson (belindat@us.ibm.com) | ||
9 | * Andy Richter (richtera@us.ibm.com) | ||
10 | */ | ||
11 | |||
12 | #undef DEBUG | ||
13 | #undef DEBUGDATA | ||
14 | #undef DEBUGCCW | ||
15 | |||
16 | #include <linux/module.h> | ||
17 | #include <linux/init.h> | ||
18 | #include <linux/kernel.h> | ||
19 | #include <linux/slab.h> | ||
20 | #include <linux/errno.h> | ||
21 | #include <linux/types.h> | ||
22 | #include <linux/interrupt.h> | ||
23 | #include <linux/timer.h> | ||
24 | #include <linux/bitops.h> | ||
25 | |||
26 | #include <linux/signal.h> | ||
27 | #include <linux/string.h> | ||
28 | |||
29 | #include <linux/ip.h> | ||
30 | #include <linux/if_arp.h> | ||
31 | #include <linux/tcp.h> | ||
32 | #include <linux/skbuff.h> | ||
33 | #include <linux/ctype.h> | ||
34 | #include <net/dst.h> | ||
35 | |||
36 | #include <linux/io.h> | ||
37 | #include <asm/ccwdev.h> | ||
38 | #include <asm/ccwgroup.h> | ||
39 | #include <linux/uaccess.h> | ||
40 | |||
41 | #include <asm/idals.h> | ||
42 | |||
43 | #include "fsm.h" | ||
44 | #include "cu3088.h" | ||
45 | |||
46 | #include "ctcm_dbug.h" | ||
47 | #include "ctcm_main.h" | ||
48 | #include "ctcm_fsms.h" | ||
49 | |||
50 | const char *dev_state_names[] = { | ||
51 | [DEV_STATE_STOPPED] = "Stopped", | ||
52 | [DEV_STATE_STARTWAIT_RXTX] = "StartWait RXTX", | ||
53 | [DEV_STATE_STARTWAIT_RX] = "StartWait RX", | ||
54 | [DEV_STATE_STARTWAIT_TX] = "StartWait TX", | ||
55 | [DEV_STATE_STOPWAIT_RXTX] = "StopWait RXTX", | ||
56 | [DEV_STATE_STOPWAIT_RX] = "StopWait RX", | ||
57 | [DEV_STATE_STOPWAIT_TX] = "StopWait TX", | ||
58 | [DEV_STATE_RUNNING] = "Running", | ||
59 | }; | ||
60 | |||
61 | const char *dev_event_names[] = { | ||
62 | [DEV_EVENT_START] = "Start", | ||
63 | [DEV_EVENT_STOP] = "Stop", | ||
64 | [DEV_EVENT_RXUP] = "RX up", | ||
65 | [DEV_EVENT_TXUP] = "TX up", | ||
66 | [DEV_EVENT_RXDOWN] = "RX down", | ||
67 | [DEV_EVENT_TXDOWN] = "TX down", | ||
68 | [DEV_EVENT_RESTART] = "Restart", | ||
69 | }; | ||
70 | |||
71 | const char *ctc_ch_event_names[] = { | ||
72 | [CTC_EVENT_IO_SUCCESS] = "ccw_device success", | ||
73 | [CTC_EVENT_IO_EBUSY] = "ccw_device busy", | ||
74 | [CTC_EVENT_IO_ENODEV] = "ccw_device enodev", | ||
75 | [CTC_EVENT_IO_UNKNOWN] = "ccw_device unknown", | ||
76 | [CTC_EVENT_ATTNBUSY] = "Status ATTN & BUSY", | ||
77 | [CTC_EVENT_ATTN] = "Status ATTN", | ||
78 | [CTC_EVENT_BUSY] = "Status BUSY", | ||
79 | [CTC_EVENT_UC_RCRESET] = "Unit check remote reset", | ||
80 | [CTC_EVENT_UC_RSRESET] = "Unit check remote system reset", | ||
81 | [CTC_EVENT_UC_TXTIMEOUT] = "Unit check TX timeout", | ||
82 | [CTC_EVENT_UC_TXPARITY] = "Unit check TX parity", | ||
83 | [CTC_EVENT_UC_HWFAIL] = "Unit check Hardware failure", | ||
84 | [CTC_EVENT_UC_RXPARITY] = "Unit check RX parity", | ||
85 | [CTC_EVENT_UC_ZERO] = "Unit check ZERO", | ||
86 | [CTC_EVENT_UC_UNKNOWN] = "Unit check Unknown", | ||
87 | [CTC_EVENT_SC_UNKNOWN] = "SubChannel check Unknown", | ||
88 | [CTC_EVENT_MC_FAIL] = "Machine check failure", | ||
89 | [CTC_EVENT_MC_GOOD] = "Machine check operational", | ||
90 | [CTC_EVENT_IRQ] = "IRQ normal", | ||
91 | [CTC_EVENT_FINSTAT] = "IRQ final", | ||
92 | [CTC_EVENT_TIMER] = "Timer", | ||
93 | [CTC_EVENT_START] = "Start", | ||
94 | [CTC_EVENT_STOP] = "Stop", | ||
95 | /* | ||
96 | * additional MPC events | ||
97 | */ | ||
98 | [CTC_EVENT_SEND_XID] = "XID Exchange", | ||
99 | [CTC_EVENT_RSWEEP_TIMER] = "MPC Group Sweep Timer", | ||
100 | }; | ||
101 | |||
102 | const char *ctc_ch_state_names[] = { | ||
103 | [CTC_STATE_IDLE] = "Idle", | ||
104 | [CTC_STATE_STOPPED] = "Stopped", | ||
105 | [CTC_STATE_STARTWAIT] = "StartWait", | ||
106 | [CTC_STATE_STARTRETRY] = "StartRetry", | ||
107 | [CTC_STATE_SETUPWAIT] = "SetupWait", | ||
108 | [CTC_STATE_RXINIT] = "RX init", | ||
109 | [CTC_STATE_TXINIT] = "TX init", | ||
110 | [CTC_STATE_RX] = "RX", | ||
111 | [CTC_STATE_TX] = "TX", | ||
112 | [CTC_STATE_RXIDLE] = "RX idle", | ||
113 | [CTC_STATE_TXIDLE] = "TX idle", | ||
114 | [CTC_STATE_RXERR] = "RX error", | ||
115 | [CTC_STATE_TXERR] = "TX error", | ||
116 | [CTC_STATE_TERM] = "Terminating", | ||
117 | [CTC_STATE_DTERM] = "Restarting", | ||
118 | [CTC_STATE_NOTOP] = "Not operational", | ||
119 | /* | ||
120 | * additional MPC states | ||
121 | */ | ||
122 | [CH_XID0_PENDING] = "Pending XID0 Start", | ||
123 | [CH_XID0_INPROGRESS] = "In XID0 Negotiations ", | ||
124 | [CH_XID7_PENDING] = "Pending XID7 P1 Start", | ||
125 | [CH_XID7_PENDING1] = "Active XID7 P1 Exchange ", | ||
126 | [CH_XID7_PENDING2] = "Pending XID7 P2 Start ", | ||
127 | [CH_XID7_PENDING3] = "Active XID7 P2 Exchange ", | ||
128 | [CH_XID7_PENDING4] = "XID7 Complete - Pending READY ", | ||
129 | }; | ||
130 | |||
131 | static void ctcm_action_nop(fsm_instance *fi, int event, void *arg); | ||
132 | |||
133 | /* | ||
134 | * ----- static ctcm actions for channel statemachine ----- | ||
135 | * | ||
136 | */ | ||
137 | static void chx_txdone(fsm_instance *fi, int event, void *arg); | ||
138 | static void chx_rx(fsm_instance *fi, int event, void *arg); | ||
139 | static void chx_rxidle(fsm_instance *fi, int event, void *arg); | ||
140 | static void chx_firstio(fsm_instance *fi, int event, void *arg); | ||
141 | static void ctcm_chx_setmode(fsm_instance *fi, int event, void *arg); | ||
142 | static void ctcm_chx_start(fsm_instance *fi, int event, void *arg); | ||
143 | static void ctcm_chx_haltio(fsm_instance *fi, int event, void *arg); | ||
144 | static void ctcm_chx_stopped(fsm_instance *fi, int event, void *arg); | ||
145 | static void ctcm_chx_stop(fsm_instance *fi, int event, void *arg); | ||
146 | static void ctcm_chx_fail(fsm_instance *fi, int event, void *arg); | ||
147 | static void ctcm_chx_setuperr(fsm_instance *fi, int event, void *arg); | ||
148 | static void ctcm_chx_restart(fsm_instance *fi, int event, void *arg); | ||
149 | static void ctcm_chx_rxiniterr(fsm_instance *fi, int event, void *arg); | ||
150 | static void ctcm_chx_rxinitfail(fsm_instance *fi, int event, void *arg); | ||
151 | static void ctcm_chx_rxdisc(fsm_instance *fi, int event, void *arg); | ||
152 | static void ctcm_chx_txiniterr(fsm_instance *fi, int event, void *arg); | ||
153 | static void ctcm_chx_txretry(fsm_instance *fi, int event, void *arg); | ||
154 | static void ctcm_chx_iofatal(fsm_instance *fi, int event, void *arg); | ||
155 | |||
156 | /* | ||
157 | * ----- static ctcmpc actions for ctcmpc channel statemachine ----- | ||
158 | * | ||
159 | */ | ||
160 | static void ctcmpc_chx_txdone(fsm_instance *fi, int event, void *arg); | ||
161 | static void ctcmpc_chx_rx(fsm_instance *fi, int event, void *arg); | ||
162 | static void ctcmpc_chx_firstio(fsm_instance *fi, int event, void *arg); | ||
163 | /* shared : | ||
164 | static void ctcm_chx_setmode(fsm_instance *fi, int event, void *arg); | ||
165 | static void ctcm_chx_start(fsm_instance *fi, int event, void *arg); | ||
166 | static void ctcm_chx_haltio(fsm_instance *fi, int event, void *arg); | ||
167 | static void ctcm_chx_stopped(fsm_instance *fi, int event, void *arg); | ||
168 | static void ctcm_chx_stop(fsm_instance *fi, int event, void *arg); | ||
169 | static void ctcm_chx_fail(fsm_instance *fi, int event, void *arg); | ||
170 | static void ctcm_chx_setuperr(fsm_instance *fi, int event, void *arg); | ||
171 | static void ctcm_chx_restart(fsm_instance *fi, int event, void *arg); | ||
172 | static void ctcm_chx_rxiniterr(fsm_instance *fi, int event, void *arg); | ||
173 | static void ctcm_chx_rxinitfail(fsm_instance *fi, int event, void *arg); | ||
174 | static void ctcm_chx_rxdisc(fsm_instance *fi, int event, void *arg); | ||
175 | static void ctcm_chx_txiniterr(fsm_instance *fi, int event, void *arg); | ||
176 | static void ctcm_chx_txretry(fsm_instance *fi, int event, void *arg); | ||
177 | static void ctcm_chx_iofatal(fsm_instance *fi, int event, void *arg); | ||
178 | */ | ||
179 | static void ctcmpc_chx_attn(fsm_instance *fsm, int event, void *arg); | ||
180 | static void ctcmpc_chx_attnbusy(fsm_instance *, int, void *); | ||
181 | static void ctcmpc_chx_resend(fsm_instance *, int, void *); | ||
182 | static void ctcmpc_chx_send_sweep(fsm_instance *fsm, int event, void *arg); | ||
183 | |||
184 | /** | ||
185 | * Check return code of a preceeding ccw_device call, halt_IO etc... | ||
186 | * | ||
187 | * ch : The channel, the error belongs to. | ||
188 | * Returns the error code (!= 0) to inspect. | ||
189 | */ | ||
190 | void ctcm_ccw_check_rc(struct channel *ch, int rc, char *msg) | ||
191 | { | ||
192 | CTCM_DBF_TEXT_(ERROR, CTC_DBF_ERROR, | ||
193 | "ccw error %s (%s): %04x\n", ch->id, msg, rc); | ||
194 | switch (rc) { | ||
195 | case -EBUSY: | ||
196 | ctcm_pr_warn("%s (%s): Busy !\n", ch->id, msg); | ||
197 | fsm_event(ch->fsm, CTC_EVENT_IO_EBUSY, ch); | ||
198 | break; | ||
199 | case -ENODEV: | ||
200 | ctcm_pr_emerg("%s (%s): Invalid device called for IO\n", | ||
201 | ch->id, msg); | ||
202 | fsm_event(ch->fsm, CTC_EVENT_IO_ENODEV, ch); | ||
203 | break; | ||
204 | default: | ||
205 | ctcm_pr_emerg("%s (%s): Unknown error in do_IO %04x\n", | ||
206 | ch->id, msg, rc); | ||
207 | fsm_event(ch->fsm, CTC_EVENT_IO_UNKNOWN, ch); | ||
208 | } | ||
209 | } | ||
210 | |||
211 | void ctcm_purge_skb_queue(struct sk_buff_head *q) | ||
212 | { | ||
213 | struct sk_buff *skb; | ||
214 | |||
215 | CTCM_DBF_TEXT(TRACE, 3, __FUNCTION__); | ||
216 | |||
217 | while ((skb = skb_dequeue(q))) { | ||
218 | atomic_dec(&skb->users); | ||
219 | dev_kfree_skb_any(skb); | ||
220 | } | ||
221 | } | ||
222 | |||
223 | /** | ||
224 | * NOP action for statemachines | ||
225 | */ | ||
226 | static void ctcm_action_nop(fsm_instance *fi, int event, void *arg) | ||
227 | { | ||
228 | } | ||
229 | |||
230 | /* | ||
231 | * Actions for channel - statemachines. | ||
232 | */ | ||
233 | |||
234 | /** | ||
235 | * Normal data has been send. Free the corresponding | ||
236 | * skb (it's in io_queue), reset dev->tbusy and | ||
237 | * revert to idle state. | ||
238 | * | ||
239 | * fi An instance of a channel statemachine. | ||
240 | * event The event, just happened. | ||
241 | * arg Generic pointer, casted from channel * upon call. | ||
242 | */ | ||
243 | static void chx_txdone(fsm_instance *fi, int event, void *arg) | ||
244 | { | ||
245 | struct channel *ch = arg; | ||
246 | struct net_device *dev = ch->netdev; | ||
247 | struct ctcm_priv *priv = dev->priv; | ||
248 | struct sk_buff *skb; | ||
249 | int first = 1; | ||
250 | int i; | ||
251 | unsigned long duration; | ||
252 | struct timespec done_stamp = current_kernel_time(); /* xtime */ | ||
253 | |||
254 | duration = | ||
255 | (done_stamp.tv_sec - ch->prof.send_stamp.tv_sec) * 1000000 + | ||
256 | (done_stamp.tv_nsec - ch->prof.send_stamp.tv_nsec) / 1000; | ||
257 | if (duration > ch->prof.tx_time) | ||
258 | ch->prof.tx_time = duration; | ||
259 | |||
260 | if (ch->irb->scsw.count != 0) | ||
261 | ctcm_pr_debug("%s: TX not complete, remaining %d bytes\n", | ||
262 | dev->name, ch->irb->scsw.count); | ||
263 | fsm_deltimer(&ch->timer); | ||
264 | while ((skb = skb_dequeue(&ch->io_queue))) { | ||
265 | priv->stats.tx_packets++; | ||
266 | priv->stats.tx_bytes += skb->len - LL_HEADER_LENGTH; | ||
267 | if (first) { | ||
268 | priv->stats.tx_bytes += 2; | ||
269 | first = 0; | ||
270 | } | ||
271 | atomic_dec(&skb->users); | ||
272 | dev_kfree_skb_irq(skb); | ||
273 | } | ||
274 | spin_lock(&ch->collect_lock); | ||
275 | clear_normalized_cda(&ch->ccw[4]); | ||
276 | if (ch->collect_len > 0) { | ||
277 | int rc; | ||
278 | |||
279 | if (ctcm_checkalloc_buffer(ch)) { | ||
280 | spin_unlock(&ch->collect_lock); | ||
281 | return; | ||
282 | } | ||
283 | ch->trans_skb->data = ch->trans_skb_data; | ||
284 | skb_reset_tail_pointer(ch->trans_skb); | ||
285 | ch->trans_skb->len = 0; | ||
286 | if (ch->prof.maxmulti < (ch->collect_len + 2)) | ||
287 | ch->prof.maxmulti = ch->collect_len + 2; | ||
288 | if (ch->prof.maxcqueue < skb_queue_len(&ch->collect_queue)) | ||
289 | ch->prof.maxcqueue = skb_queue_len(&ch->collect_queue); | ||
290 | *((__u16 *)skb_put(ch->trans_skb, 2)) = ch->collect_len + 2; | ||
291 | i = 0; | ||
292 | while ((skb = skb_dequeue(&ch->collect_queue))) { | ||
293 | skb_copy_from_linear_data(skb, | ||
294 | skb_put(ch->trans_skb, skb->len), skb->len); | ||
295 | priv->stats.tx_packets++; | ||
296 | priv->stats.tx_bytes += skb->len - LL_HEADER_LENGTH; | ||
297 | atomic_dec(&skb->users); | ||
298 | dev_kfree_skb_irq(skb); | ||
299 | i++; | ||
300 | } | ||
301 | ch->collect_len = 0; | ||
302 | spin_unlock(&ch->collect_lock); | ||
303 | ch->ccw[1].count = ch->trans_skb->len; | ||
304 | fsm_addtimer(&ch->timer, CTCM_TIME_5_SEC, CTC_EVENT_TIMER, ch); | ||
305 | ch->prof.send_stamp = current_kernel_time(); /* xtime */ | ||
306 | rc = ccw_device_start(ch->cdev, &ch->ccw[0], | ||
307 | (unsigned long)ch, 0xff, 0); | ||
308 | ch->prof.doios_multi++; | ||
309 | if (rc != 0) { | ||
310 | priv->stats.tx_dropped += i; | ||
311 | priv->stats.tx_errors += i; | ||
312 | fsm_deltimer(&ch->timer); | ||
313 | ctcm_ccw_check_rc(ch, rc, "chained TX"); | ||
314 | } | ||
315 | } else { | ||
316 | spin_unlock(&ch->collect_lock); | ||
317 | fsm_newstate(fi, CTC_STATE_TXIDLE); | ||
318 | } | ||
319 | ctcm_clear_busy_do(dev); | ||
320 | } | ||
321 | |||
322 | /** | ||
323 | * Initial data is sent. | ||
324 | * Notify device statemachine that we are up and | ||
325 | * running. | ||
326 | * | ||
327 | * fi An instance of a channel statemachine. | ||
328 | * event The event, just happened. | ||
329 | * arg Generic pointer, casted from channel * upon call. | ||
330 | */ | ||
331 | void ctcm_chx_txidle(fsm_instance *fi, int event, void *arg) | ||
332 | { | ||
333 | struct channel *ch = arg; | ||
334 | struct net_device *dev = ch->netdev; | ||
335 | struct ctcm_priv *priv = dev->priv; | ||
336 | |||
337 | CTCM_DBF_TEXT(TRACE, 6, __FUNCTION__); | ||
338 | fsm_deltimer(&ch->timer); | ||
339 | fsm_newstate(fi, CTC_STATE_TXIDLE); | ||
340 | fsm_event(priv->fsm, DEV_EVENT_TXUP, ch->netdev); | ||
341 | } | ||
342 | |||
343 | /** | ||
344 | * Got normal data, check for sanity, queue it up, allocate new buffer | ||
345 | * trigger bottom half, and initiate next read. | ||
346 | * | ||
347 | * fi An instance of a channel statemachine. | ||
348 | * event The event, just happened. | ||
349 | * arg Generic pointer, casted from channel * upon call. | ||
350 | */ | ||
351 | static void chx_rx(fsm_instance *fi, int event, void *arg) | ||
352 | { | ||
353 | struct channel *ch = arg; | ||
354 | struct net_device *dev = ch->netdev; | ||
355 | struct ctcm_priv *priv = dev->priv; | ||
356 | int len = ch->max_bufsize - ch->irb->scsw.count; | ||
357 | struct sk_buff *skb = ch->trans_skb; | ||
358 | __u16 block_len = *((__u16 *)skb->data); | ||
359 | int check_len; | ||
360 | int rc; | ||
361 | |||
362 | fsm_deltimer(&ch->timer); | ||
363 | if (len < 8) { | ||
364 | ctcm_pr_debug("%s: got packet with length %d < 8\n", | ||
365 | dev->name, len); | ||
366 | priv->stats.rx_dropped++; | ||
367 | priv->stats.rx_length_errors++; | ||
368 | goto again; | ||
369 | } | ||
370 | if (len > ch->max_bufsize) { | ||
371 | ctcm_pr_debug("%s: got packet with length %d > %d\n", | ||
372 | dev->name, len, ch->max_bufsize); | ||
373 | priv->stats.rx_dropped++; | ||
374 | priv->stats.rx_length_errors++; | ||
375 | goto again; | ||
376 | } | ||
377 | |||
378 | /* | ||
379 | * VM TCP seems to have a bug sending 2 trailing bytes of garbage. | ||
380 | */ | ||
381 | switch (ch->protocol) { | ||
382 | case CTCM_PROTO_S390: | ||
383 | case CTCM_PROTO_OS390: | ||
384 | check_len = block_len + 2; | ||
385 | break; | ||
386 | default: | ||
387 | check_len = block_len; | ||
388 | break; | ||
389 | } | ||
390 | if ((len < block_len) || (len > check_len)) { | ||
391 | ctcm_pr_debug("%s: got block length %d != rx length %d\n", | ||
392 | dev->name, block_len, len); | ||
393 | if (do_debug) | ||
394 | ctcmpc_dump_skb(skb, 0); | ||
395 | |||
396 | *((__u16 *)skb->data) = len; | ||
397 | priv->stats.rx_dropped++; | ||
398 | priv->stats.rx_length_errors++; | ||
399 | goto again; | ||
400 | } | ||
401 | block_len -= 2; | ||
402 | if (block_len > 0) { | ||
403 | *((__u16 *)skb->data) = block_len; | ||
404 | ctcm_unpack_skb(ch, skb); | ||
405 | } | ||
406 | again: | ||
407 | skb->data = ch->trans_skb_data; | ||
408 | skb_reset_tail_pointer(skb); | ||
409 | skb->len = 0; | ||
410 | if (ctcm_checkalloc_buffer(ch)) | ||
411 | return; | ||
412 | ch->ccw[1].count = ch->max_bufsize; | ||
413 | rc = ccw_device_start(ch->cdev, &ch->ccw[0], | ||
414 | (unsigned long)ch, 0xff, 0); | ||
415 | if (rc != 0) | ||
416 | ctcm_ccw_check_rc(ch, rc, "normal RX"); | ||
417 | } | ||
418 | |||
419 | /** | ||
420 | * Initialize connection by sending a __u16 of value 0. | ||
421 | * | ||
422 | * fi An instance of a channel statemachine. | ||
423 | * event The event, just happened. | ||
424 | * arg Generic pointer, casted from channel * upon call. | ||
425 | */ | ||
426 | static void chx_firstio(fsm_instance *fi, int event, void *arg) | ||
427 | { | ||
428 | struct channel *ch = arg; | ||
429 | int rc; | ||
430 | |||
431 | CTCM_DBF_TEXT(TRACE, 6, __FUNCTION__); | ||
432 | |||
433 | if (fsm_getstate(fi) == CTC_STATE_TXIDLE) | ||
434 | ctcm_pr_debug("%s: remote side issued READ?, init.\n", ch->id); | ||
435 | fsm_deltimer(&ch->timer); | ||
436 | if (ctcm_checkalloc_buffer(ch)) | ||
437 | return; | ||
438 | if ((fsm_getstate(fi) == CTC_STATE_SETUPWAIT) && | ||
439 | (ch->protocol == CTCM_PROTO_OS390)) { | ||
440 | /* OS/390 resp. z/OS */ | ||
441 | if (CHANNEL_DIRECTION(ch->flags) == READ) { | ||
442 | *((__u16 *)ch->trans_skb->data) = CTCM_INITIAL_BLOCKLEN; | ||
443 | fsm_addtimer(&ch->timer, CTCM_TIME_5_SEC, | ||
444 | CTC_EVENT_TIMER, ch); | ||
445 | chx_rxidle(fi, event, arg); | ||
446 | } else { | ||
447 | struct net_device *dev = ch->netdev; | ||
448 | struct ctcm_priv *priv = dev->priv; | ||
449 | fsm_newstate(fi, CTC_STATE_TXIDLE); | ||
450 | fsm_event(priv->fsm, DEV_EVENT_TXUP, dev); | ||
451 | } | ||
452 | return; | ||
453 | } | ||
454 | |||
455 | /* | ||
456 | * Don't setup a timer for receiving the initial RX frame | ||
457 | * if in compatibility mode, since VM TCP delays the initial | ||
458 | * frame until it has some data to send. | ||
459 | */ | ||
460 | if ((CHANNEL_DIRECTION(ch->flags) == WRITE) || | ||
461 | (ch->protocol != CTCM_PROTO_S390)) | ||
462 | fsm_addtimer(&ch->timer, CTCM_TIME_5_SEC, CTC_EVENT_TIMER, ch); | ||
463 | |||
464 | *((__u16 *)ch->trans_skb->data) = CTCM_INITIAL_BLOCKLEN; | ||
465 | ch->ccw[1].count = 2; /* Transfer only length */ | ||
466 | |||
467 | fsm_newstate(fi, (CHANNEL_DIRECTION(ch->flags) == READ) | ||
468 | ? CTC_STATE_RXINIT : CTC_STATE_TXINIT); | ||
469 | rc = ccw_device_start(ch->cdev, &ch->ccw[0], | ||
470 | (unsigned long)ch, 0xff, 0); | ||
471 | if (rc != 0) { | ||
472 | fsm_deltimer(&ch->timer); | ||
473 | fsm_newstate(fi, CTC_STATE_SETUPWAIT); | ||
474 | ctcm_ccw_check_rc(ch, rc, "init IO"); | ||
475 | } | ||
476 | /* | ||
477 | * If in compatibility mode since we don't setup a timer, we | ||
478 | * also signal RX channel up immediately. This enables us | ||
479 | * to send packets early which in turn usually triggers some | ||
480 | * reply from VM TCP which brings up the RX channel to it's | ||
481 | * final state. | ||
482 | */ | ||
483 | if ((CHANNEL_DIRECTION(ch->flags) == READ) && | ||
484 | (ch->protocol == CTCM_PROTO_S390)) { | ||
485 | struct net_device *dev = ch->netdev; | ||
486 | struct ctcm_priv *priv = dev->priv; | ||
487 | fsm_event(priv->fsm, DEV_EVENT_RXUP, dev); | ||
488 | } | ||
489 | } | ||
490 | |||
491 | /** | ||
492 | * Got initial data, check it. If OK, | ||
493 | * notify device statemachine that we are up and | ||
494 | * running. | ||
495 | * | ||
496 | * fi An instance of a channel statemachine. | ||
497 | * event The event, just happened. | ||
498 | * arg Generic pointer, casted from channel * upon call. | ||
499 | */ | ||
500 | static void chx_rxidle(fsm_instance *fi, int event, void *arg) | ||
501 | { | ||
502 | struct channel *ch = arg; | ||
503 | struct net_device *dev = ch->netdev; | ||
504 | struct ctcm_priv *priv = dev->priv; | ||
505 | __u16 buflen; | ||
506 | int rc; | ||
507 | |||
508 | CTCM_DBF_TEXT(TRACE, 6, __FUNCTION__); | ||
509 | fsm_deltimer(&ch->timer); | ||
510 | buflen = *((__u16 *)ch->trans_skb->data); | ||
511 | if (do_debug) | ||
512 | ctcm_pr_debug("%s: Initial RX count %d\n", dev->name, buflen); | ||
513 | |||
514 | if (buflen >= CTCM_INITIAL_BLOCKLEN) { | ||
515 | if (ctcm_checkalloc_buffer(ch)) | ||
516 | return; | ||
517 | ch->ccw[1].count = ch->max_bufsize; | ||
518 | fsm_newstate(fi, CTC_STATE_RXIDLE); | ||
519 | rc = ccw_device_start(ch->cdev, &ch->ccw[0], | ||
520 | (unsigned long)ch, 0xff, 0); | ||
521 | if (rc != 0) { | ||
522 | fsm_newstate(fi, CTC_STATE_RXINIT); | ||
523 | ctcm_ccw_check_rc(ch, rc, "initial RX"); | ||
524 | } else | ||
525 | fsm_event(priv->fsm, DEV_EVENT_RXUP, dev); | ||
526 | } else { | ||
527 | if (do_debug) | ||
528 | ctcm_pr_debug("%s: Initial RX count %d not %d\n", | ||
529 | dev->name, buflen, CTCM_INITIAL_BLOCKLEN); | ||
530 | chx_firstio(fi, event, arg); | ||
531 | } | ||
532 | } | ||
533 | |||
534 | /** | ||
535 | * Set channel into extended mode. | ||
536 | * | ||
537 | * fi An instance of a channel statemachine. | ||
538 | * event The event, just happened. | ||
539 | * arg Generic pointer, casted from channel * upon call. | ||
540 | */ | ||
541 | static void ctcm_chx_setmode(fsm_instance *fi, int event, void *arg) | ||
542 | { | ||
543 | struct channel *ch = arg; | ||
544 | int rc; | ||
545 | unsigned long saveflags = 0; | ||
546 | int timeout = CTCM_TIME_5_SEC; | ||
547 | |||
548 | fsm_deltimer(&ch->timer); | ||
549 | if (IS_MPC(ch)) { | ||
550 | timeout = 1500; | ||
551 | if (do_debug) | ||
552 | ctcm_pr_debug("ctcm enter: %s(): cp=%i ch=0x%p id=%s\n", | ||
553 | __FUNCTION__, smp_processor_id(), ch, ch->id); | ||
554 | } | ||
555 | fsm_addtimer(&ch->timer, timeout, CTC_EVENT_TIMER, ch); | ||
556 | fsm_newstate(fi, CTC_STATE_SETUPWAIT); | ||
557 | if (do_debug_ccw && IS_MPC(ch)) | ||
558 | ctcmpc_dumpit((char *)&ch->ccw[6], sizeof(struct ccw1) * 2); | ||
559 | |||
560 | if (event == CTC_EVENT_TIMER) /* only for timer not yet locked */ | ||
561 | spin_lock_irqsave(get_ccwdev_lock(ch->cdev), saveflags); | ||
562 | /* Such conditional locking is undeterministic in | ||
563 | * static view. => ignore sparse warnings here. */ | ||
564 | |||
565 | rc = ccw_device_start(ch->cdev, &ch->ccw[6], | ||
566 | (unsigned long)ch, 0xff, 0); | ||
567 | if (event == CTC_EVENT_TIMER) /* see above comments */ | ||
568 | spin_unlock_irqrestore(get_ccwdev_lock(ch->cdev), saveflags); | ||
569 | if (rc != 0) { | ||
570 | fsm_deltimer(&ch->timer); | ||
571 | fsm_newstate(fi, CTC_STATE_STARTWAIT); | ||
572 | ctcm_ccw_check_rc(ch, rc, "set Mode"); | ||
573 | } else | ||
574 | ch->retry = 0; | ||
575 | } | ||
576 | |||
577 | /** | ||
578 | * Setup channel. | ||
579 | * | ||
580 | * fi An instance of a channel statemachine. | ||
581 | * event The event, just happened. | ||
582 | * arg Generic pointer, casted from channel * upon call. | ||
583 | */ | ||
584 | static void ctcm_chx_start(fsm_instance *fi, int event, void *arg) | ||
585 | { | ||
586 | struct channel *ch = arg; | ||
587 | int rc; | ||
588 | struct net_device *dev; | ||
589 | unsigned long saveflags; | ||
590 | |||
591 | CTCM_DBF_TEXT(TRACE, 5, __FUNCTION__); | ||
592 | if (ch == NULL) { | ||
593 | ctcm_pr_warn("chx_start ch=NULL\n"); | ||
594 | return; | ||
595 | } | ||
596 | if (ch->netdev == NULL) { | ||
597 | ctcm_pr_warn("chx_start dev=NULL, id=%s\n", ch->id); | ||
598 | return; | ||
599 | } | ||
600 | dev = ch->netdev; | ||
601 | |||
602 | if (do_debug) | ||
603 | ctcm_pr_debug("%s: %s channel start\n", dev->name, | ||
604 | (CHANNEL_DIRECTION(ch->flags) == READ) ? "RX" : "TX"); | ||
605 | |||
606 | if (ch->trans_skb != NULL) { | ||
607 | clear_normalized_cda(&ch->ccw[1]); | ||
608 | dev_kfree_skb(ch->trans_skb); | ||
609 | ch->trans_skb = NULL; | ||
610 | } | ||
611 | if (CHANNEL_DIRECTION(ch->flags) == READ) { | ||
612 | ch->ccw[1].cmd_code = CCW_CMD_READ; | ||
613 | ch->ccw[1].flags = CCW_FLAG_SLI; | ||
614 | ch->ccw[1].count = 0; | ||
615 | } else { | ||
616 | ch->ccw[1].cmd_code = CCW_CMD_WRITE; | ||
617 | ch->ccw[1].flags = CCW_FLAG_SLI | CCW_FLAG_CC; | ||
618 | ch->ccw[1].count = 0; | ||
619 | } | ||
620 | if (ctcm_checkalloc_buffer(ch)) { | ||
621 | ctcm_pr_notice("%s: %s trans_skb allocation delayed " | ||
622 | "until first transfer\n", dev->name, | ||
623 | (CHANNEL_DIRECTION(ch->flags) == READ) ? "RX" : "TX"); | ||
624 | } | ||
625 | |||
626 | ch->ccw[0].cmd_code = CCW_CMD_PREPARE; | ||
627 | ch->ccw[0].flags = CCW_FLAG_SLI | CCW_FLAG_CC; | ||
628 | ch->ccw[0].count = 0; | ||
629 | ch->ccw[0].cda = 0; | ||
630 | ch->ccw[2].cmd_code = CCW_CMD_NOOP; /* jointed CE + DE */ | ||
631 | ch->ccw[2].flags = CCW_FLAG_SLI; | ||
632 | ch->ccw[2].count = 0; | ||
633 | ch->ccw[2].cda = 0; | ||
634 | memcpy(&ch->ccw[3], &ch->ccw[0], sizeof(struct ccw1) * 3); | ||
635 | ch->ccw[4].cda = 0; | ||
636 | ch->ccw[4].flags &= ~CCW_FLAG_IDA; | ||
637 | |||
638 | fsm_newstate(fi, CTC_STATE_STARTWAIT); | ||
639 | fsm_addtimer(&ch->timer, 1000, CTC_EVENT_TIMER, ch); | ||
640 | spin_lock_irqsave(get_ccwdev_lock(ch->cdev), saveflags); | ||
641 | rc = ccw_device_halt(ch->cdev, (unsigned long)ch); | ||
642 | spin_unlock_irqrestore(get_ccwdev_lock(ch->cdev), saveflags); | ||
643 | if (rc != 0) { | ||
644 | if (rc != -EBUSY) | ||
645 | fsm_deltimer(&ch->timer); | ||
646 | ctcm_ccw_check_rc(ch, rc, "initial HaltIO"); | ||
647 | } | ||
648 | } | ||
649 | |||
650 | /** | ||
651 | * Shutdown a channel. | ||
652 | * | ||
653 | * fi An instance of a channel statemachine. | ||
654 | * event The event, just happened. | ||
655 | * arg Generic pointer, casted from channel * upon call. | ||
656 | */ | ||
657 | static void ctcm_chx_haltio(fsm_instance *fi, int event, void *arg) | ||
658 | { | ||
659 | struct channel *ch = arg; | ||
660 | unsigned long saveflags = 0; | ||
661 | int rc; | ||
662 | int oldstate; | ||
663 | |||
664 | CTCM_DBF_TEXT(TRACE, 2, __FUNCTION__); | ||
665 | fsm_deltimer(&ch->timer); | ||
666 | if (IS_MPC(ch)) | ||
667 | fsm_deltimer(&ch->sweep_timer); | ||
668 | |||
669 | fsm_addtimer(&ch->timer, CTCM_TIME_5_SEC, CTC_EVENT_TIMER, ch); | ||
670 | |||
671 | if (event == CTC_EVENT_STOP) /* only for STOP not yet locked */ | ||
672 | spin_lock_irqsave(get_ccwdev_lock(ch->cdev), saveflags); | ||
673 | /* Such conditional locking is undeterministic in | ||
674 | * static view. => ignore sparse warnings here. */ | ||
675 | oldstate = fsm_getstate(fi); | ||
676 | fsm_newstate(fi, CTC_STATE_TERM); | ||
677 | rc = ccw_device_halt(ch->cdev, (unsigned long)ch); | ||
678 | |||
679 | if (event == CTC_EVENT_STOP) | ||
680 | spin_unlock_irqrestore(get_ccwdev_lock(ch->cdev), saveflags); | ||
681 | /* see remark above about conditional locking */ | ||
682 | |||
683 | if (rc != 0 && rc != -EBUSY) { | ||
684 | fsm_deltimer(&ch->timer); | ||
685 | if (event != CTC_EVENT_STOP) { | ||
686 | fsm_newstate(fi, oldstate); | ||
687 | ctcm_ccw_check_rc(ch, rc, (char *)__FUNCTION__); | ||
688 | } | ||
689 | } | ||
690 | } | ||
691 | |||
692 | /** | ||
693 | * Cleanup helper for chx_fail and chx_stopped | ||
694 | * cleanup channels queue and notify interface statemachine. | ||
695 | * | ||
696 | * fi An instance of a channel statemachine. | ||
697 | * state The next state (depending on caller). | ||
698 | * ch The channel to operate on. | ||
699 | */ | ||
700 | static void ctcm_chx_cleanup(fsm_instance *fi, int state, | ||
701 | struct channel *ch) | ||
702 | { | ||
703 | struct net_device *dev = ch->netdev; | ||
704 | struct ctcm_priv *priv = dev->priv; | ||
705 | |||
706 | CTCM_DBF_TEXT(TRACE, 3, __FUNCTION__); | ||
707 | |||
708 | fsm_deltimer(&ch->timer); | ||
709 | if (IS_MPC(ch)) | ||
710 | fsm_deltimer(&ch->sweep_timer); | ||
711 | |||
712 | fsm_newstate(fi, state); | ||
713 | if (state == CTC_STATE_STOPPED && ch->trans_skb != NULL) { | ||
714 | clear_normalized_cda(&ch->ccw[1]); | ||
715 | dev_kfree_skb_any(ch->trans_skb); | ||
716 | ch->trans_skb = NULL; | ||
717 | } | ||
718 | |||
719 | ch->th_seg = 0x00; | ||
720 | ch->th_seq_num = 0x00; | ||
721 | if (CHANNEL_DIRECTION(ch->flags) == READ) { | ||
722 | skb_queue_purge(&ch->io_queue); | ||
723 | fsm_event(priv->fsm, DEV_EVENT_RXDOWN, dev); | ||
724 | } else { | ||
725 | ctcm_purge_skb_queue(&ch->io_queue); | ||
726 | if (IS_MPC(ch)) | ||
727 | ctcm_purge_skb_queue(&ch->sweep_queue); | ||
728 | spin_lock(&ch->collect_lock); | ||
729 | ctcm_purge_skb_queue(&ch->collect_queue); | ||
730 | ch->collect_len = 0; | ||
731 | spin_unlock(&ch->collect_lock); | ||
732 | fsm_event(priv->fsm, DEV_EVENT_TXDOWN, dev); | ||
733 | } | ||
734 | } | ||
735 | |||
736 | /** | ||
737 | * A channel has successfully been halted. | ||
738 | * Cleanup it's queue and notify interface statemachine. | ||
739 | * | ||
740 | * fi An instance of a channel statemachine. | ||
741 | * event The event, just happened. | ||
742 | * arg Generic pointer, casted from channel * upon call. | ||
743 | */ | ||
744 | static void ctcm_chx_stopped(fsm_instance *fi, int event, void *arg) | ||
745 | { | ||
746 | CTCM_DBF_TEXT(TRACE, 3, __FUNCTION__); | ||
747 | ctcm_chx_cleanup(fi, CTC_STATE_STOPPED, arg); | ||
748 | } | ||
749 | |||
750 | /** | ||
751 | * A stop command from device statemachine arrived and we are in | ||
752 | * not operational mode. Set state to stopped. | ||
753 | * | ||
754 | * fi An instance of a channel statemachine. | ||
755 | * event The event, just happened. | ||
756 | * arg Generic pointer, casted from channel * upon call. | ||
757 | */ | ||
758 | static void ctcm_chx_stop(fsm_instance *fi, int event, void *arg) | ||
759 | { | ||
760 | fsm_newstate(fi, CTC_STATE_STOPPED); | ||
761 | } | ||
762 | |||
763 | /** | ||
764 | * A machine check for no path, not operational status or gone device has | ||
765 | * happened. | ||
766 | * Cleanup queue and notify interface statemachine. | ||
767 | * | ||
768 | * fi An instance of a channel statemachine. | ||
769 | * event The event, just happened. | ||
770 | * arg Generic pointer, casted from channel * upon call. | ||
771 | */ | ||
772 | static void ctcm_chx_fail(fsm_instance *fi, int event, void *arg) | ||
773 | { | ||
774 | CTCM_DBF_TEXT(TRACE, 3, __FUNCTION__); | ||
775 | ctcm_chx_cleanup(fi, CTC_STATE_NOTOP, arg); | ||
776 | } | ||
777 | |||
778 | /** | ||
779 | * Handle error during setup of channel. | ||
780 | * | ||
781 | * fi An instance of a channel statemachine. | ||
782 | * event The event, just happened. | ||
783 | * arg Generic pointer, casted from channel * upon call. | ||
784 | */ | ||
785 | static void ctcm_chx_setuperr(fsm_instance *fi, int event, void *arg) | ||
786 | { | ||
787 | struct channel *ch = arg; | ||
788 | struct net_device *dev = ch->netdev; | ||
789 | struct ctcm_priv *priv = dev->priv; | ||
790 | |||
791 | /* | ||
792 | * Special case: Got UC_RCRESET on setmode. | ||
793 | * This means that remote side isn't setup. In this case | ||
794 | * simply retry after some 10 secs... | ||
795 | */ | ||
796 | if ((fsm_getstate(fi) == CTC_STATE_SETUPWAIT) && | ||
797 | ((event == CTC_EVENT_UC_RCRESET) || | ||
798 | (event == CTC_EVENT_UC_RSRESET))) { | ||
799 | fsm_newstate(fi, CTC_STATE_STARTRETRY); | ||
800 | fsm_deltimer(&ch->timer); | ||
801 | fsm_addtimer(&ch->timer, CTCM_TIME_5_SEC, CTC_EVENT_TIMER, ch); | ||
802 | if (!IS_MPC(ch) && (CHANNEL_DIRECTION(ch->flags) == READ)) { | ||
803 | int rc = ccw_device_halt(ch->cdev, (unsigned long)ch); | ||
804 | if (rc != 0) | ||
805 | ctcm_ccw_check_rc(ch, rc, | ||
806 | "HaltIO in chx_setuperr"); | ||
807 | } | ||
808 | return; | ||
809 | } | ||
810 | |||
811 | CTCM_DBF_TEXT_(ERROR, CTC_DBF_CRIT, | ||
812 | "%s : %s error during %s channel setup state=%s\n", | ||
813 | dev->name, ctc_ch_event_names[event], | ||
814 | (CHANNEL_DIRECTION(ch->flags) == READ) ? "RX" : "TX", | ||
815 | fsm_getstate_str(fi)); | ||
816 | |||
817 | if (CHANNEL_DIRECTION(ch->flags) == READ) { | ||
818 | fsm_newstate(fi, CTC_STATE_RXERR); | ||
819 | fsm_event(priv->fsm, DEV_EVENT_RXDOWN, dev); | ||
820 | } else { | ||
821 | fsm_newstate(fi, CTC_STATE_TXERR); | ||
822 | fsm_event(priv->fsm, DEV_EVENT_TXDOWN, dev); | ||
823 | } | ||
824 | } | ||
825 | |||
826 | /** | ||
827 | * Restart a channel after an error. | ||
828 | * | ||
829 | * fi An instance of a channel statemachine. | ||
830 | * event The event, just happened. | ||
831 | * arg Generic pointer, casted from channel * upon call. | ||
832 | */ | ||
833 | static void ctcm_chx_restart(fsm_instance *fi, int event, void *arg) | ||
834 | { | ||
835 | struct channel *ch = arg; | ||
836 | struct net_device *dev = ch->netdev; | ||
837 | unsigned long saveflags = 0; | ||
838 | int oldstate; | ||
839 | int rc; | ||
840 | |||
841 | CTCM_DBF_TEXT(TRACE, CTC_DBF_NOTICE, __FUNCTION__); | ||
842 | fsm_deltimer(&ch->timer); | ||
843 | ctcm_pr_debug("%s: %s channel restart\n", dev->name, | ||
844 | (CHANNEL_DIRECTION(ch->flags) == READ) ? "RX" : "TX"); | ||
845 | fsm_addtimer(&ch->timer, CTCM_TIME_5_SEC, CTC_EVENT_TIMER, ch); | ||
846 | oldstate = fsm_getstate(fi); | ||
847 | fsm_newstate(fi, CTC_STATE_STARTWAIT); | ||
848 | if (event == CTC_EVENT_TIMER) /* only for timer not yet locked */ | ||
849 | spin_lock_irqsave(get_ccwdev_lock(ch->cdev), saveflags); | ||
850 | /* Such conditional locking is a known problem for | ||
851 | * sparse because its undeterministic in static view. | ||
852 | * Warnings should be ignored here. */ | ||
853 | rc = ccw_device_halt(ch->cdev, (unsigned long)ch); | ||
854 | if (event == CTC_EVENT_TIMER) | ||
855 | spin_unlock_irqrestore(get_ccwdev_lock(ch->cdev), saveflags); | ||
856 | if (rc != 0) { | ||
857 | if (rc != -EBUSY) { | ||
858 | fsm_deltimer(&ch->timer); | ||
859 | fsm_newstate(fi, oldstate); | ||
860 | } | ||
861 | ctcm_ccw_check_rc(ch, rc, "HaltIO in ctcm_chx_restart"); | ||
862 | } | ||
863 | } | ||
864 | |||
865 | /** | ||
866 | * Handle error during RX initial handshake (exchange of | ||
867 | * 0-length block header) | ||
868 | * | ||
869 | * fi An instance of a channel statemachine. | ||
870 | * event The event, just happened. | ||
871 | * arg Generic pointer, casted from channel * upon call. | ||
872 | */ | ||
873 | static void ctcm_chx_rxiniterr(fsm_instance *fi, int event, void *arg) | ||
874 | { | ||
875 | struct channel *ch = arg; | ||
876 | struct net_device *dev = ch->netdev; | ||
877 | struct ctcm_priv *priv = dev->priv; | ||
878 | |||
879 | CTCM_DBF_TEXT(SETUP, 3, __FUNCTION__); | ||
880 | if (event == CTC_EVENT_TIMER) { | ||
881 | if (!IS_MPCDEV(dev)) | ||
882 | /* TODO : check if MPC deletes timer somewhere */ | ||
883 | fsm_deltimer(&ch->timer); | ||
884 | ctcm_pr_debug("%s: Timeout during RX init handshake\n", | ||
885 | dev->name); | ||
886 | if (ch->retry++ < 3) | ||
887 | ctcm_chx_restart(fi, event, arg); | ||
888 | else { | ||
889 | fsm_newstate(fi, CTC_STATE_RXERR); | ||
890 | fsm_event(priv->fsm, DEV_EVENT_RXDOWN, dev); | ||
891 | } | ||
892 | } else | ||
893 | ctcm_pr_warn("%s: Error during RX init handshake\n", dev->name); | ||
894 | } | ||
895 | |||
896 | /** | ||
897 | * Notify device statemachine if we gave up initialization | ||
898 | * of RX channel. | ||
899 | * | ||
900 | * fi An instance of a channel statemachine. | ||
901 | * event The event, just happened. | ||
902 | * arg Generic pointer, casted from channel * upon call. | ||
903 | */ | ||
904 | static void ctcm_chx_rxinitfail(fsm_instance *fi, int event, void *arg) | ||
905 | { | ||
906 | struct channel *ch = arg; | ||
907 | struct net_device *dev = ch->netdev; | ||
908 | struct ctcm_priv *priv = dev->priv; | ||
909 | |||
910 | CTCM_DBF_TEXT(SETUP, 3, __FUNCTION__); | ||
911 | fsm_newstate(fi, CTC_STATE_RXERR); | ||
912 | ctcm_pr_warn("%s: RX busy. Initialization failed\n", dev->name); | ||
913 | fsm_event(priv->fsm, DEV_EVENT_RXDOWN, dev); | ||
914 | } | ||
915 | |||
916 | /** | ||
917 | * Handle RX Unit check remote reset (remote disconnected) | ||
918 | * | ||
919 | * fi An instance of a channel statemachine. | ||
920 | * event The event, just happened. | ||
921 | * arg Generic pointer, casted from channel * upon call. | ||
922 | */ | ||
923 | static void ctcm_chx_rxdisc(fsm_instance *fi, int event, void *arg) | ||
924 | { | ||
925 | struct channel *ch = arg; | ||
926 | struct channel *ch2; | ||
927 | struct net_device *dev = ch->netdev; | ||
928 | struct ctcm_priv *priv = dev->priv; | ||
929 | |||
930 | CTCM_DBF_DEV_NAME(TRACE, dev, "Got remote disconnect, re-initializing"); | ||
931 | fsm_deltimer(&ch->timer); | ||
932 | if (do_debug) | ||
933 | ctcm_pr_debug("%s: Got remote disconnect, " | ||
934 | "re-initializing ...\n", dev->name); | ||
935 | /* | ||
936 | * Notify device statemachine | ||
937 | */ | ||
938 | fsm_event(priv->fsm, DEV_EVENT_RXDOWN, dev); | ||
939 | fsm_event(priv->fsm, DEV_EVENT_TXDOWN, dev); | ||
940 | |||
941 | fsm_newstate(fi, CTC_STATE_DTERM); | ||
942 | ch2 = priv->channel[WRITE]; | ||
943 | fsm_newstate(ch2->fsm, CTC_STATE_DTERM); | ||
944 | |||
945 | ccw_device_halt(ch->cdev, (unsigned long)ch); | ||
946 | ccw_device_halt(ch2->cdev, (unsigned long)ch2); | ||
947 | } | ||
948 | |||
949 | /** | ||
950 | * Handle error during TX channel initialization. | ||
951 | * | ||
952 | * fi An instance of a channel statemachine. | ||
953 | * event The event, just happened. | ||
954 | * arg Generic pointer, casted from channel * upon call. | ||
955 | */ | ||
956 | static void ctcm_chx_txiniterr(fsm_instance *fi, int event, void *arg) | ||
957 | { | ||
958 | struct channel *ch = arg; | ||
959 | struct net_device *dev = ch->netdev; | ||
960 | struct ctcm_priv *priv = dev->priv; | ||
961 | |||
962 | if (event == CTC_EVENT_TIMER) { | ||
963 | fsm_deltimer(&ch->timer); | ||
964 | CTCM_DBF_DEV_NAME(ERROR, dev, | ||
965 | "Timeout during TX init handshake"); | ||
966 | if (ch->retry++ < 3) | ||
967 | ctcm_chx_restart(fi, event, arg); | ||
968 | else { | ||
969 | fsm_newstate(fi, CTC_STATE_TXERR); | ||
970 | fsm_event(priv->fsm, DEV_EVENT_TXDOWN, dev); | ||
971 | } | ||
972 | } else { | ||
973 | CTCM_DBF_TEXT_(ERROR, CTC_DBF_ERROR, | ||
974 | "%s : %s error during channel setup state=%s", | ||
975 | dev->name, ctc_ch_event_names[event], | ||
976 | fsm_getstate_str(fi)); | ||
977 | |||
978 | ctcm_pr_warn("%s: Error during TX init handshake\n", dev->name); | ||
979 | } | ||
980 | } | ||
981 | |||
982 | /** | ||
983 | * Handle TX timeout by retrying operation. | ||
984 | * | ||
985 | * fi An instance of a channel statemachine. | ||
986 | * event The event, just happened. | ||
987 | * arg Generic pointer, casted from channel * upon call. | ||
988 | */ | ||
989 | static void ctcm_chx_txretry(fsm_instance *fi, int event, void *arg) | ||
990 | { | ||
991 | struct channel *ch = arg; | ||
992 | struct net_device *dev = ch->netdev; | ||
993 | struct ctcm_priv *priv = dev->priv; | ||
994 | struct sk_buff *skb; | ||
995 | |||
996 | if (do_debug) | ||
997 | ctcm_pr_debug("ctcmpc enter: %s(): cp=%i ch=0x%p id=%s\n", | ||
998 | __FUNCTION__, smp_processor_id(), ch, ch->id); | ||
999 | |||
1000 | fsm_deltimer(&ch->timer); | ||
1001 | if (ch->retry++ > 3) { | ||
1002 | struct mpc_group *gptr = priv->mpcg; | ||
1003 | ctcm_pr_debug("%s: TX retry failed, restarting channel\n", | ||
1004 | dev->name); | ||
1005 | fsm_event(priv->fsm, DEV_EVENT_TXDOWN, dev); | ||
1006 | /* call restart if not MPC or if MPC and mpcg fsm is ready. | ||
1007 | use gptr as mpc indicator */ | ||
1008 | if (!(gptr && (fsm_getstate(gptr->fsm) != MPCG_STATE_READY))) | ||
1009 | ctcm_chx_restart(fi, event, arg); | ||
1010 | goto done; | ||
1011 | } | ||
1012 | |||
1013 | ctcm_pr_debug("%s: TX retry %d\n", dev->name, ch->retry); | ||
1014 | skb = skb_peek(&ch->io_queue); | ||
1015 | if (skb) { | ||
1016 | int rc = 0; | ||
1017 | unsigned long saveflags = 0; | ||
1018 | clear_normalized_cda(&ch->ccw[4]); | ||
1019 | ch->ccw[4].count = skb->len; | ||
1020 | if (set_normalized_cda(&ch->ccw[4], skb->data)) { | ||
1021 | ctcm_pr_debug("%s: IDAL alloc failed, chan restart\n", | ||
1022 | dev->name); | ||
1023 | fsm_event(priv->fsm, DEV_EVENT_TXDOWN, dev); | ||
1024 | ctcm_chx_restart(fi, event, arg); | ||
1025 | goto done; | ||
1026 | } | ||
1027 | fsm_addtimer(&ch->timer, 1000, CTC_EVENT_TIMER, ch); | ||
1028 | if (event == CTC_EVENT_TIMER) /* for TIMER not yet locked */ | ||
1029 | spin_lock_irqsave(get_ccwdev_lock(ch->cdev), saveflags); | ||
1030 | /* Such conditional locking is a known problem for | ||
1031 | * sparse because its undeterministic in static view. | ||
1032 | * Warnings should be ignored here. */ | ||
1033 | if (do_debug_ccw) | ||
1034 | ctcmpc_dumpit((char *)&ch->ccw[3], | ||
1035 | sizeof(struct ccw1) * 3); | ||
1036 | |||
1037 | rc = ccw_device_start(ch->cdev, &ch->ccw[3], | ||
1038 | (unsigned long)ch, 0xff, 0); | ||
1039 | if (event == CTC_EVENT_TIMER) | ||
1040 | spin_unlock_irqrestore(get_ccwdev_lock(ch->cdev), | ||
1041 | saveflags); | ||
1042 | if (rc != 0) { | ||
1043 | fsm_deltimer(&ch->timer); | ||
1044 | ctcm_ccw_check_rc(ch, rc, "TX in chx_txretry"); | ||
1045 | ctcm_purge_skb_queue(&ch->io_queue); | ||
1046 | } | ||
1047 | } | ||
1048 | done: | ||
1049 | return; | ||
1050 | } | ||
1051 | |||
1052 | /** | ||
1053 | * Handle fatal errors during an I/O command. | ||
1054 | * | ||
1055 | * fi An instance of a channel statemachine. | ||
1056 | * event The event, just happened. | ||
1057 | * arg Generic pointer, casted from channel * upon call. | ||
1058 | */ | ||
1059 | static void ctcm_chx_iofatal(fsm_instance *fi, int event, void *arg) | ||
1060 | { | ||
1061 | struct channel *ch = arg; | ||
1062 | struct net_device *dev = ch->netdev; | ||
1063 | struct ctcm_priv *priv = dev->priv; | ||
1064 | |||
1065 | CTCM_DBF_TEXT(TRACE, 3, __FUNCTION__); | ||
1066 | fsm_deltimer(&ch->timer); | ||
1067 | ctcm_pr_warn("%s %s : unrecoverable channel error\n", | ||
1068 | CTC_DRIVER_NAME, dev->name); | ||
1069 | if (IS_MPC(ch)) { | ||
1070 | priv->stats.tx_dropped++; | ||
1071 | priv->stats.tx_errors++; | ||
1072 | } | ||
1073 | |||
1074 | if (CHANNEL_DIRECTION(ch->flags) == READ) { | ||
1075 | ctcm_pr_debug("%s: RX I/O error\n", dev->name); | ||
1076 | fsm_newstate(fi, CTC_STATE_RXERR); | ||
1077 | fsm_event(priv->fsm, DEV_EVENT_RXDOWN, dev); | ||
1078 | } else { | ||
1079 | ctcm_pr_debug("%s: TX I/O error\n", dev->name); | ||
1080 | fsm_newstate(fi, CTC_STATE_TXERR); | ||
1081 | fsm_event(priv->fsm, DEV_EVENT_TXDOWN, dev); | ||
1082 | } | ||
1083 | } | ||
1084 | |||
1085 | /* | ||
1086 | * The ctcm statemachine for a channel. | ||
1087 | */ | ||
1088 | const fsm_node ch_fsm[] = { | ||
1089 | { CTC_STATE_STOPPED, CTC_EVENT_STOP, ctcm_action_nop }, | ||
1090 | { CTC_STATE_STOPPED, CTC_EVENT_START, ctcm_chx_start }, | ||
1091 | { CTC_STATE_STOPPED, CTC_EVENT_FINSTAT, ctcm_action_nop }, | ||
1092 | { CTC_STATE_STOPPED, CTC_EVENT_MC_FAIL, ctcm_action_nop }, | ||
1093 | |||
1094 | { CTC_STATE_NOTOP, CTC_EVENT_STOP, ctcm_chx_stop }, | ||
1095 | { CTC_STATE_NOTOP, CTC_EVENT_START, ctcm_action_nop }, | ||
1096 | { CTC_STATE_NOTOP, CTC_EVENT_FINSTAT, ctcm_action_nop }, | ||
1097 | { CTC_STATE_NOTOP, CTC_EVENT_MC_FAIL, ctcm_action_nop }, | ||
1098 | { CTC_STATE_NOTOP, CTC_EVENT_MC_GOOD, ctcm_chx_start }, | ||
1099 | |||
1100 | { CTC_STATE_STARTWAIT, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
1101 | { CTC_STATE_STARTWAIT, CTC_EVENT_START, ctcm_action_nop }, | ||
1102 | { CTC_STATE_STARTWAIT, CTC_EVENT_FINSTAT, ctcm_chx_setmode }, | ||
1103 | { CTC_STATE_STARTWAIT, CTC_EVENT_TIMER, ctcm_chx_setuperr }, | ||
1104 | { CTC_STATE_STARTWAIT, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
1105 | { CTC_STATE_STARTWAIT, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
1106 | |||
1107 | { CTC_STATE_STARTRETRY, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
1108 | { CTC_STATE_STARTRETRY, CTC_EVENT_TIMER, ctcm_chx_setmode }, | ||
1109 | { CTC_STATE_STARTRETRY, CTC_EVENT_FINSTAT, ctcm_action_nop }, | ||
1110 | { CTC_STATE_STARTRETRY, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
1111 | |||
1112 | { CTC_STATE_SETUPWAIT, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
1113 | { CTC_STATE_SETUPWAIT, CTC_EVENT_START, ctcm_action_nop }, | ||
1114 | { CTC_STATE_SETUPWAIT, CTC_EVENT_FINSTAT, chx_firstio }, | ||
1115 | { CTC_STATE_SETUPWAIT, CTC_EVENT_UC_RCRESET, ctcm_chx_setuperr }, | ||
1116 | { CTC_STATE_SETUPWAIT, CTC_EVENT_UC_RSRESET, ctcm_chx_setuperr }, | ||
1117 | { CTC_STATE_SETUPWAIT, CTC_EVENT_TIMER, ctcm_chx_setmode }, | ||
1118 | { CTC_STATE_SETUPWAIT, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
1119 | { CTC_STATE_SETUPWAIT, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
1120 | |||
1121 | { CTC_STATE_RXINIT, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
1122 | { CTC_STATE_RXINIT, CTC_EVENT_START, ctcm_action_nop }, | ||
1123 | { CTC_STATE_RXINIT, CTC_EVENT_FINSTAT, chx_rxidle }, | ||
1124 | { CTC_STATE_RXINIT, CTC_EVENT_UC_RCRESET, ctcm_chx_rxiniterr }, | ||
1125 | { CTC_STATE_RXINIT, CTC_EVENT_UC_RSRESET, ctcm_chx_rxiniterr }, | ||
1126 | { CTC_STATE_RXINIT, CTC_EVENT_TIMER, ctcm_chx_rxiniterr }, | ||
1127 | { CTC_STATE_RXINIT, CTC_EVENT_ATTNBUSY, ctcm_chx_rxinitfail }, | ||
1128 | { CTC_STATE_RXINIT, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
1129 | { CTC_STATE_RXINIT, CTC_EVENT_UC_ZERO, chx_firstio }, | ||
1130 | { CTC_STATE_RXINIT, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
1131 | |||
1132 | { CTC_STATE_RXIDLE, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
1133 | { CTC_STATE_RXIDLE, CTC_EVENT_START, ctcm_action_nop }, | ||
1134 | { CTC_STATE_RXIDLE, CTC_EVENT_FINSTAT, chx_rx }, | ||
1135 | { CTC_STATE_RXIDLE, CTC_EVENT_UC_RCRESET, ctcm_chx_rxdisc }, | ||
1136 | { CTC_STATE_RXIDLE, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
1137 | { CTC_STATE_RXIDLE, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
1138 | { CTC_STATE_RXIDLE, CTC_EVENT_UC_ZERO, chx_rx }, | ||
1139 | |||
1140 | { CTC_STATE_TXINIT, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
1141 | { CTC_STATE_TXINIT, CTC_EVENT_START, ctcm_action_nop }, | ||
1142 | { CTC_STATE_TXINIT, CTC_EVENT_FINSTAT, ctcm_chx_txidle }, | ||
1143 | { CTC_STATE_TXINIT, CTC_EVENT_UC_RCRESET, ctcm_chx_txiniterr }, | ||
1144 | { CTC_STATE_TXINIT, CTC_EVENT_UC_RSRESET, ctcm_chx_txiniterr }, | ||
1145 | { CTC_STATE_TXINIT, CTC_EVENT_TIMER, ctcm_chx_txiniterr }, | ||
1146 | { CTC_STATE_TXINIT, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
1147 | { CTC_STATE_TXINIT, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
1148 | |||
1149 | { CTC_STATE_TXIDLE, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
1150 | { CTC_STATE_TXIDLE, CTC_EVENT_START, ctcm_action_nop }, | ||
1151 | { CTC_STATE_TXIDLE, CTC_EVENT_FINSTAT, chx_firstio }, | ||
1152 | { CTC_STATE_TXIDLE, CTC_EVENT_UC_RCRESET, ctcm_action_nop }, | ||
1153 | { CTC_STATE_TXIDLE, CTC_EVENT_UC_RSRESET, ctcm_action_nop }, | ||
1154 | { CTC_STATE_TXIDLE, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
1155 | { CTC_STATE_TXIDLE, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
1156 | |||
1157 | { CTC_STATE_TERM, CTC_EVENT_STOP, ctcm_action_nop }, | ||
1158 | { CTC_STATE_TERM, CTC_EVENT_START, ctcm_chx_restart }, | ||
1159 | { CTC_STATE_TERM, CTC_EVENT_FINSTAT, ctcm_chx_stopped }, | ||
1160 | { CTC_STATE_TERM, CTC_EVENT_UC_RCRESET, ctcm_action_nop }, | ||
1161 | { CTC_STATE_TERM, CTC_EVENT_UC_RSRESET, ctcm_action_nop }, | ||
1162 | { CTC_STATE_TERM, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
1163 | |||
1164 | { CTC_STATE_DTERM, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
1165 | { CTC_STATE_DTERM, CTC_EVENT_START, ctcm_chx_restart }, | ||
1166 | { CTC_STATE_DTERM, CTC_EVENT_FINSTAT, ctcm_chx_setmode }, | ||
1167 | { CTC_STATE_DTERM, CTC_EVENT_UC_RCRESET, ctcm_action_nop }, | ||
1168 | { CTC_STATE_DTERM, CTC_EVENT_UC_RSRESET, ctcm_action_nop }, | ||
1169 | { CTC_STATE_DTERM, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
1170 | |||
1171 | { CTC_STATE_TX, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
1172 | { CTC_STATE_TX, CTC_EVENT_START, ctcm_action_nop }, | ||
1173 | { CTC_STATE_TX, CTC_EVENT_FINSTAT, chx_txdone }, | ||
1174 | { CTC_STATE_TX, CTC_EVENT_UC_RCRESET, ctcm_chx_txretry }, | ||
1175 | { CTC_STATE_TX, CTC_EVENT_UC_RSRESET, ctcm_chx_txretry }, | ||
1176 | { CTC_STATE_TX, CTC_EVENT_TIMER, ctcm_chx_txretry }, | ||
1177 | { CTC_STATE_TX, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
1178 | { CTC_STATE_TX, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
1179 | |||
1180 | { CTC_STATE_RXERR, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
1181 | { CTC_STATE_TXERR, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
1182 | { CTC_STATE_TXERR, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
1183 | { CTC_STATE_RXERR, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
1184 | }; | ||
1185 | |||
1186 | int ch_fsm_len = ARRAY_SIZE(ch_fsm); | ||
1187 | |||
1188 | /* | ||
1189 | * MPC actions for mpc channel statemachine | ||
1190 | * handling of MPC protocol requires extra | ||
1191 | * statemachine and actions which are prefixed ctcmpc_ . | ||
1192 | * The ctc_ch_states and ctc_ch_state_names, | ||
1193 | * ctc_ch_events and ctc_ch_event_names share the ctcm definitions | ||
1194 | * which are expanded by some elements. | ||
1195 | */ | ||
1196 | |||
1197 | /* | ||
1198 | * Actions for mpc channel statemachine. | ||
1199 | */ | ||
1200 | |||
1201 | /** | ||
1202 | * Normal data has been send. Free the corresponding | ||
1203 | * skb (it's in io_queue), reset dev->tbusy and | ||
1204 | * revert to idle state. | ||
1205 | * | ||
1206 | * fi An instance of a channel statemachine. | ||
1207 | * event The event, just happened. | ||
1208 | * arg Generic pointer, casted from channel * upon call. | ||
1209 | */ | ||
1210 | static void ctcmpc_chx_txdone(fsm_instance *fi, int event, void *arg) | ||
1211 | { | ||
1212 | struct channel *ch = arg; | ||
1213 | struct net_device *dev = ch->netdev; | ||
1214 | struct ctcm_priv *priv = dev->priv; | ||
1215 | struct mpc_group *grp = priv->mpcg; | ||
1216 | struct sk_buff *skb; | ||
1217 | int first = 1; | ||
1218 | int i; | ||
1219 | struct timespec done_stamp; | ||
1220 | __u32 data_space; | ||
1221 | unsigned long duration; | ||
1222 | struct sk_buff *peekskb; | ||
1223 | int rc; | ||
1224 | struct th_header *header; | ||
1225 | struct pdu *p_header; | ||
1226 | |||
1227 | if (do_debug) | ||
1228 | ctcm_pr_debug("%s cp:%i enter: %s()\n", | ||
1229 | dev->name, smp_processor_id(), __FUNCTION__); | ||
1230 | |||
1231 | done_stamp = current_kernel_time(); /* xtime */ | ||
1232 | duration = (done_stamp.tv_sec - ch->prof.send_stamp.tv_sec) * 1000000 | ||
1233 | + (done_stamp.tv_nsec - ch->prof.send_stamp.tv_nsec) / 1000; | ||
1234 | if (duration > ch->prof.tx_time) | ||
1235 | ch->prof.tx_time = duration; | ||
1236 | |||
1237 | if (ch->irb->scsw.count != 0) | ||
1238 | ctcm_pr_debug("%s: TX not complete, remaining %d bytes\n", | ||
1239 | dev->name, ch->irb->scsw.count); | ||
1240 | fsm_deltimer(&ch->timer); | ||
1241 | while ((skb = skb_dequeue(&ch->io_queue))) { | ||
1242 | priv->stats.tx_packets++; | ||
1243 | priv->stats.tx_bytes += skb->len - TH_HEADER_LENGTH; | ||
1244 | if (first) { | ||
1245 | priv->stats.tx_bytes += 2; | ||
1246 | first = 0; | ||
1247 | } | ||
1248 | atomic_dec(&skb->users); | ||
1249 | dev_kfree_skb_irq(skb); | ||
1250 | } | ||
1251 | spin_lock(&ch->collect_lock); | ||
1252 | clear_normalized_cda(&ch->ccw[4]); | ||
1253 | |||
1254 | if ((ch->collect_len <= 0) || (grp->in_sweep != 0)) { | ||
1255 | spin_unlock(&ch->collect_lock); | ||
1256 | fsm_newstate(fi, CTC_STATE_TXIDLE); | ||
1257 | goto done; | ||
1258 | } | ||
1259 | |||
1260 | if (ctcm_checkalloc_buffer(ch)) { | ||
1261 | spin_unlock(&ch->collect_lock); | ||
1262 | goto done; | ||
1263 | } | ||
1264 | ch->trans_skb->data = ch->trans_skb_data; | ||
1265 | skb_reset_tail_pointer(ch->trans_skb); | ||
1266 | ch->trans_skb->len = 0; | ||
1267 | if (ch->prof.maxmulti < (ch->collect_len + TH_HEADER_LENGTH)) | ||
1268 | ch->prof.maxmulti = ch->collect_len + TH_HEADER_LENGTH; | ||
1269 | if (ch->prof.maxcqueue < skb_queue_len(&ch->collect_queue)) | ||
1270 | ch->prof.maxcqueue = skb_queue_len(&ch->collect_queue); | ||
1271 | i = 0; | ||
1272 | |||
1273 | if (do_debug_data) | ||
1274 | ctcm_pr_debug("ctcmpc: %s() building " | ||
1275 | "trans_skb from collect_q \n", __FUNCTION__); | ||
1276 | |||
1277 | data_space = grp->group_max_buflen - TH_HEADER_LENGTH; | ||
1278 | |||
1279 | if (do_debug_data) | ||
1280 | ctcm_pr_debug("ctcmpc: %s() building trans_skb from collect_q" | ||
1281 | " data_space:%04x\n", __FUNCTION__, data_space); | ||
1282 | p_header = NULL; | ||
1283 | while ((skb = skb_dequeue(&ch->collect_queue))) { | ||
1284 | memcpy(skb_put(ch->trans_skb, skb->len), skb->data, skb->len); | ||
1285 | p_header = (struct pdu *) | ||
1286 | (skb_tail_pointer(ch->trans_skb) - skb->len); | ||
1287 | p_header->pdu_flag = 0x00; | ||
1288 | if (skb->protocol == ntohs(ETH_P_SNAP)) | ||
1289 | p_header->pdu_flag |= 0x60; | ||
1290 | else | ||
1291 | p_header->pdu_flag |= 0x20; | ||
1292 | |||
1293 | if (do_debug_data) { | ||
1294 | ctcm_pr_debug("ctcmpc: %s()trans_skb len:%04x \n", | ||
1295 | __FUNCTION__, ch->trans_skb->len); | ||
1296 | ctcm_pr_debug("ctcmpc: %s() pdu header and data" | ||
1297 | " for up to 32 bytes sent to vtam\n", | ||
1298 | __FUNCTION__); | ||
1299 | ctcmpc_dumpit((char *)p_header, | ||
1300 | min_t(int, skb->len, 32)); | ||
1301 | } | ||
1302 | ch->collect_len -= skb->len; | ||
1303 | data_space -= skb->len; | ||
1304 | priv->stats.tx_packets++; | ||
1305 | priv->stats.tx_bytes += skb->len; | ||
1306 | atomic_dec(&skb->users); | ||
1307 | dev_kfree_skb_any(skb); | ||
1308 | peekskb = skb_peek(&ch->collect_queue); | ||
1309 | if (peekskb->len > data_space) | ||
1310 | break; | ||
1311 | i++; | ||
1312 | } | ||
1313 | /* p_header points to the last one we handled */ | ||
1314 | if (p_header) | ||
1315 | p_header->pdu_flag |= PDU_LAST; /*Say it's the last one*/ | ||
1316 | header = kzalloc(TH_HEADER_LENGTH, gfp_type()); | ||
1317 | |||
1318 | if (!header) { | ||
1319 | printk(KERN_WARNING "ctcmpc: OUT OF MEMORY IN %s()" | ||
1320 | ": Data Lost \n", __FUNCTION__); | ||
1321 | spin_unlock(&ch->collect_lock); | ||
1322 | fsm_event(priv->mpcg->fsm, MPCG_EVENT_INOP, dev); | ||
1323 | goto done; | ||
1324 | } | ||
1325 | |||
1326 | header->th_ch_flag = TH_HAS_PDU; /* Normal data */ | ||
1327 | ch->th_seq_num++; | ||
1328 | header->th_seq_num = ch->th_seq_num; | ||
1329 | |||
1330 | if (do_debug_data) | ||
1331 | ctcm_pr_debug("%s: ToVTAM_th_seq= %08x\n" , | ||
1332 | __FUNCTION__, ch->th_seq_num); | ||
1333 | |||
1334 | memcpy(skb_push(ch->trans_skb, TH_HEADER_LENGTH), header, | ||
1335 | TH_HEADER_LENGTH); /* put the TH on the packet */ | ||
1336 | |||
1337 | kfree(header); | ||
1338 | |||
1339 | if (do_debug_data) { | ||
1340 | ctcm_pr_debug("ctcmpc: %s()trans_skb len:%04x \n", | ||
1341 | __FUNCTION__, ch->trans_skb->len); | ||
1342 | |||
1343 | ctcm_pr_debug("ctcmpc: %s() up-to-50 bytes of trans_skb " | ||
1344 | "data to vtam from collect_q\n", __FUNCTION__); | ||
1345 | ctcmpc_dumpit((char *)ch->trans_skb->data, | ||
1346 | min_t(int, ch->trans_skb->len, 50)); | ||
1347 | } | ||
1348 | |||
1349 | spin_unlock(&ch->collect_lock); | ||
1350 | clear_normalized_cda(&ch->ccw[1]); | ||
1351 | if (set_normalized_cda(&ch->ccw[1], ch->trans_skb->data)) { | ||
1352 | dev_kfree_skb_any(ch->trans_skb); | ||
1353 | ch->trans_skb = NULL; | ||
1354 | printk(KERN_WARNING | ||
1355 | "ctcmpc: %s()CCW failure - data lost\n", | ||
1356 | __FUNCTION__); | ||
1357 | fsm_event(priv->mpcg->fsm, MPCG_EVENT_INOP, dev); | ||
1358 | return; | ||
1359 | } | ||
1360 | ch->ccw[1].count = ch->trans_skb->len; | ||
1361 | fsm_addtimer(&ch->timer, CTCM_TIME_5_SEC, CTC_EVENT_TIMER, ch); | ||
1362 | ch->prof.send_stamp = current_kernel_time(); /* xtime */ | ||
1363 | if (do_debug_ccw) | ||
1364 | ctcmpc_dumpit((char *)&ch->ccw[0], sizeof(struct ccw1) * 3); | ||
1365 | rc = ccw_device_start(ch->cdev, &ch->ccw[0], | ||
1366 | (unsigned long)ch, 0xff, 0); | ||
1367 | ch->prof.doios_multi++; | ||
1368 | if (rc != 0) { | ||
1369 | priv->stats.tx_dropped += i; | ||
1370 | priv->stats.tx_errors += i; | ||
1371 | fsm_deltimer(&ch->timer); | ||
1372 | ctcm_ccw_check_rc(ch, rc, "chained TX"); | ||
1373 | } | ||
1374 | done: | ||
1375 | ctcm_clear_busy(dev); | ||
1376 | ctcm_pr_debug("ctcmpc exit: %s %s()\n", dev->name, __FUNCTION__); | ||
1377 | return; | ||
1378 | } | ||
1379 | |||
1380 | /** | ||
1381 | * Got normal data, check for sanity, queue it up, allocate new buffer | ||
1382 | * trigger bottom half, and initiate next read. | ||
1383 | * | ||
1384 | * fi An instance of a channel statemachine. | ||
1385 | * event The event, just happened. | ||
1386 | * arg Generic pointer, casted from channel * upon call. | ||
1387 | */ | ||
1388 | static void ctcmpc_chx_rx(fsm_instance *fi, int event, void *arg) | ||
1389 | { | ||
1390 | struct channel *ch = arg; | ||
1391 | struct net_device *dev = ch->netdev; | ||
1392 | struct ctcm_priv *priv = dev->priv; | ||
1393 | struct mpc_group *grp = priv->mpcg; | ||
1394 | struct sk_buff *skb = ch->trans_skb; | ||
1395 | struct sk_buff *new_skb; | ||
1396 | unsigned long saveflags = 0; /* avoids compiler warning */ | ||
1397 | int len = ch->max_bufsize - ch->irb->scsw.count; | ||
1398 | |||
1399 | if (do_debug_data) { | ||
1400 | CTCM_DBF_TEXT_(TRACE, CTC_DBF_DEBUG, "mpc_ch_rx %s cp:%i %s\n", | ||
1401 | dev->name, smp_processor_id(), ch->id); | ||
1402 | CTCM_DBF_TEXT_(TRACE, CTC_DBF_DEBUG, "mpc_ch_rx: maxbuf: %04x " | ||
1403 | "len: %04x\n", ch->max_bufsize, len); | ||
1404 | } | ||
1405 | fsm_deltimer(&ch->timer); | ||
1406 | |||
1407 | if (skb == NULL) { | ||
1408 | ctcm_pr_debug("ctcmpc exit: %s() TRANS_SKB = NULL \n", | ||
1409 | __FUNCTION__); | ||
1410 | goto again; | ||
1411 | } | ||
1412 | |||
1413 | if (len < TH_HEADER_LENGTH) { | ||
1414 | ctcm_pr_info("%s: got packet with invalid length %d\n", | ||
1415 | dev->name, len); | ||
1416 | priv->stats.rx_dropped++; | ||
1417 | priv->stats.rx_length_errors++; | ||
1418 | } else { | ||
1419 | /* must have valid th header or game over */ | ||
1420 | __u32 block_len = len; | ||
1421 | len = TH_HEADER_LENGTH + XID2_LENGTH + 4; | ||
1422 | new_skb = __dev_alloc_skb(ch->max_bufsize, GFP_ATOMIC); | ||
1423 | |||
1424 | if (new_skb == NULL) { | ||
1425 | printk(KERN_INFO "ctcmpc:%s() NEW_SKB = NULL\n", | ||
1426 | __FUNCTION__); | ||
1427 | printk(KERN_WARNING "ctcmpc: %s() MEMORY ALLOC FAILED" | ||
1428 | " - DATA LOST - MPC FAILED\n", | ||
1429 | __FUNCTION__); | ||
1430 | fsm_event(priv->mpcg->fsm, MPCG_EVENT_INOP, dev); | ||
1431 | goto again; | ||
1432 | } | ||
1433 | switch (fsm_getstate(grp->fsm)) { | ||
1434 | case MPCG_STATE_RESET: | ||
1435 | case MPCG_STATE_INOP: | ||
1436 | dev_kfree_skb_any(new_skb); | ||
1437 | break; | ||
1438 | case MPCG_STATE_FLOWC: | ||
1439 | case MPCG_STATE_READY: | ||
1440 | memcpy(skb_put(new_skb, block_len), | ||
1441 | skb->data, block_len); | ||
1442 | skb_queue_tail(&ch->io_queue, new_skb); | ||
1443 | tasklet_schedule(&ch->ch_tasklet); | ||
1444 | break; | ||
1445 | default: | ||
1446 | memcpy(skb_put(new_skb, len), skb->data, len); | ||
1447 | skb_queue_tail(&ch->io_queue, new_skb); | ||
1448 | tasklet_hi_schedule(&ch->ch_tasklet); | ||
1449 | break; | ||
1450 | } | ||
1451 | } | ||
1452 | |||
1453 | again: | ||
1454 | switch (fsm_getstate(grp->fsm)) { | ||
1455 | int rc, dolock; | ||
1456 | case MPCG_STATE_FLOWC: | ||
1457 | case MPCG_STATE_READY: | ||
1458 | if (ctcm_checkalloc_buffer(ch)) | ||
1459 | break; | ||
1460 | ch->trans_skb->data = ch->trans_skb_data; | ||
1461 | skb_reset_tail_pointer(ch->trans_skb); | ||
1462 | ch->trans_skb->len = 0; | ||
1463 | ch->ccw[1].count = ch->max_bufsize; | ||
1464 | if (do_debug_ccw) | ||
1465 | ctcmpc_dumpit((char *)&ch->ccw[0], | ||
1466 | sizeof(struct ccw1) * 3); | ||
1467 | dolock = !in_irq(); | ||
1468 | if (dolock) | ||
1469 | spin_lock_irqsave( | ||
1470 | get_ccwdev_lock(ch->cdev), saveflags); | ||
1471 | rc = ccw_device_start(ch->cdev, &ch->ccw[0], | ||
1472 | (unsigned long)ch, 0xff, 0); | ||
1473 | if (dolock) /* see remark about conditional locking */ | ||
1474 | spin_unlock_irqrestore( | ||
1475 | get_ccwdev_lock(ch->cdev), saveflags); | ||
1476 | if (rc != 0) | ||
1477 | ctcm_ccw_check_rc(ch, rc, "normal RX"); | ||
1478 | default: | ||
1479 | break; | ||
1480 | } | ||
1481 | |||
1482 | if (do_debug) | ||
1483 | ctcm_pr_debug("ctcmpc exit : %s %s(): ch=0x%p id=%s\n", | ||
1484 | dev->name, __FUNCTION__, ch, ch->id); | ||
1485 | |||
1486 | } | ||
1487 | |||
1488 | /** | ||
1489 | * Initialize connection by sending a __u16 of value 0. | ||
1490 | * | ||
1491 | * fi An instance of a channel statemachine. | ||
1492 | * event The event, just happened. | ||
1493 | * arg Generic pointer, casted from channel * upon call. | ||
1494 | */ | ||
1495 | static void ctcmpc_chx_firstio(fsm_instance *fi, int event, void *arg) | ||
1496 | { | ||
1497 | struct channel *ch = arg; | ||
1498 | struct net_device *dev = ch->netdev; | ||
1499 | struct ctcm_priv *priv = dev->priv; | ||
1500 | |||
1501 | if (do_debug) { | ||
1502 | struct mpc_group *gptr = priv->mpcg; | ||
1503 | ctcm_pr_debug("ctcmpc enter: %s(): ch=0x%p id=%s\n", | ||
1504 | __FUNCTION__, ch, ch->id); | ||
1505 | ctcm_pr_debug("%s() %s chstate:%i grpstate:%i chprotocol:%i\n", | ||
1506 | __FUNCTION__, ch->id, fsm_getstate(fi), | ||
1507 | fsm_getstate(gptr->fsm), ch->protocol); | ||
1508 | } | ||
1509 | if (fsm_getstate(fi) == CTC_STATE_TXIDLE) | ||
1510 | MPC_DBF_DEV_NAME(TRACE, dev, "remote side issued READ? "); | ||
1511 | |||
1512 | fsm_deltimer(&ch->timer); | ||
1513 | if (ctcm_checkalloc_buffer(ch)) | ||
1514 | goto done; | ||
1515 | |||
1516 | switch (fsm_getstate(fi)) { | ||
1517 | case CTC_STATE_STARTRETRY: | ||
1518 | case CTC_STATE_SETUPWAIT: | ||
1519 | if (CHANNEL_DIRECTION(ch->flags) == READ) { | ||
1520 | ctcmpc_chx_rxidle(fi, event, arg); | ||
1521 | } else { | ||
1522 | fsm_newstate(fi, CTC_STATE_TXIDLE); | ||
1523 | fsm_event(priv->fsm, DEV_EVENT_TXUP, dev); | ||
1524 | } | ||
1525 | goto done; | ||
1526 | default: | ||
1527 | break; | ||
1528 | }; | ||
1529 | |||
1530 | fsm_newstate(fi, (CHANNEL_DIRECTION(ch->flags) == READ) | ||
1531 | ? CTC_STATE_RXINIT : CTC_STATE_TXINIT); | ||
1532 | |||
1533 | done: | ||
1534 | if (do_debug) | ||
1535 | ctcm_pr_debug("ctcmpc exit : %s(): ch=0x%p id=%s\n", | ||
1536 | __FUNCTION__, ch, ch->id); | ||
1537 | return; | ||
1538 | } | ||
1539 | |||
1540 | /** | ||
1541 | * Got initial data, check it. If OK, | ||
1542 | * notify device statemachine that we are up and | ||
1543 | * running. | ||
1544 | * | ||
1545 | * fi An instance of a channel statemachine. | ||
1546 | * event The event, just happened. | ||
1547 | * arg Generic pointer, casted from channel * upon call. | ||
1548 | */ | ||
1549 | void ctcmpc_chx_rxidle(fsm_instance *fi, int event, void *arg) | ||
1550 | { | ||
1551 | struct channel *ch = arg; | ||
1552 | struct net_device *dev = ch->netdev; | ||
1553 | struct ctcm_priv *priv = dev->priv; | ||
1554 | struct mpc_group *grp = priv->mpcg; | ||
1555 | int rc; | ||
1556 | unsigned long saveflags = 0; /* avoids compiler warning */ | ||
1557 | |||
1558 | fsm_deltimer(&ch->timer); | ||
1559 | ctcm_pr_debug("%s cp:%i enter: %s()\n", | ||
1560 | dev->name, smp_processor_id(), __FUNCTION__); | ||
1561 | if (do_debug) | ||
1562 | ctcm_pr_debug("%s() %s chstate:%i grpstate:%i\n", | ||
1563 | __FUNCTION__, ch->id, | ||
1564 | fsm_getstate(fi), fsm_getstate(grp->fsm)); | ||
1565 | |||
1566 | fsm_newstate(fi, CTC_STATE_RXIDLE); | ||
1567 | /* XID processing complete */ | ||
1568 | |||
1569 | switch (fsm_getstate(grp->fsm)) { | ||
1570 | case MPCG_STATE_FLOWC: | ||
1571 | case MPCG_STATE_READY: | ||
1572 | if (ctcm_checkalloc_buffer(ch)) | ||
1573 | goto done; | ||
1574 | ch->trans_skb->data = ch->trans_skb_data; | ||
1575 | skb_reset_tail_pointer(ch->trans_skb); | ||
1576 | ch->trans_skb->len = 0; | ||
1577 | ch->ccw[1].count = ch->max_bufsize; | ||
1578 | if (do_debug_ccw) | ||
1579 | ctcmpc_dumpit((char *)&ch->ccw[0], | ||
1580 | sizeof(struct ccw1) * 3); | ||
1581 | if (event == CTC_EVENT_START) | ||
1582 | /* see remark about conditional locking */ | ||
1583 | spin_lock_irqsave(get_ccwdev_lock(ch->cdev), saveflags); | ||
1584 | rc = ccw_device_start(ch->cdev, &ch->ccw[0], | ||
1585 | (unsigned long)ch, 0xff, 0); | ||
1586 | if (event == CTC_EVENT_START) | ||
1587 | spin_unlock_irqrestore( | ||
1588 | get_ccwdev_lock(ch->cdev), saveflags); | ||
1589 | if (rc != 0) { | ||
1590 | fsm_newstate(fi, CTC_STATE_RXINIT); | ||
1591 | ctcm_ccw_check_rc(ch, rc, "initial RX"); | ||
1592 | goto done; | ||
1593 | } | ||
1594 | break; | ||
1595 | default: | ||
1596 | break; | ||
1597 | } | ||
1598 | |||
1599 | fsm_event(priv->fsm, DEV_EVENT_RXUP, dev); | ||
1600 | done: | ||
1601 | if (do_debug) | ||
1602 | ctcm_pr_debug("ctcmpc exit: %s %s()\n", | ||
1603 | dev->name, __FUNCTION__); | ||
1604 | return; | ||
1605 | } | ||
1606 | |||
1607 | /* | ||
1608 | * ctcmpc channel FSM action | ||
1609 | * called from several points in ctcmpc_ch_fsm | ||
1610 | * ctcmpc only | ||
1611 | */ | ||
1612 | static void ctcmpc_chx_attn(fsm_instance *fsm, int event, void *arg) | ||
1613 | { | ||
1614 | struct channel *ch = arg; | ||
1615 | struct net_device *dev = ch->netdev; | ||
1616 | struct ctcm_priv *priv = dev->priv; | ||
1617 | struct mpc_group *grp = priv->mpcg; | ||
1618 | |||
1619 | if (do_debug) { | ||
1620 | ctcm_pr_debug("ctcmpc enter: %s(): cp=%i ch=0x%p id=%s" | ||
1621 | "GrpState:%s ChState:%s\n", | ||
1622 | __FUNCTION__, smp_processor_id(), ch, ch->id, | ||
1623 | fsm_getstate_str(grp->fsm), | ||
1624 | fsm_getstate_str(ch->fsm)); | ||
1625 | } | ||
1626 | |||
1627 | switch (fsm_getstate(grp->fsm)) { | ||
1628 | case MPCG_STATE_XID2INITW: | ||
1629 | /* ok..start yside xid exchanges */ | ||
1630 | if (!ch->in_mpcgroup) | ||
1631 | break; | ||
1632 | if (fsm_getstate(ch->fsm) == CH_XID0_PENDING) { | ||
1633 | fsm_deltimer(&grp->timer); | ||
1634 | fsm_addtimer(&grp->timer, | ||
1635 | MPC_XID_TIMEOUT_VALUE, | ||
1636 | MPCG_EVENT_TIMER, dev); | ||
1637 | fsm_event(grp->fsm, MPCG_EVENT_XID0DO, ch); | ||
1638 | |||
1639 | } else if (fsm_getstate(ch->fsm) < CH_XID7_PENDING1) | ||
1640 | /* attn rcvd before xid0 processed via bh */ | ||
1641 | fsm_newstate(ch->fsm, CH_XID7_PENDING1); | ||
1642 | break; | ||
1643 | case MPCG_STATE_XID2INITX: | ||
1644 | case MPCG_STATE_XID0IOWAIT: | ||
1645 | case MPCG_STATE_XID0IOWAIX: | ||
1646 | /* attn rcvd before xid0 processed on ch | ||
1647 | but mid-xid0 processing for group */ | ||
1648 | if (fsm_getstate(ch->fsm) < CH_XID7_PENDING1) | ||
1649 | fsm_newstate(ch->fsm, CH_XID7_PENDING1); | ||
1650 | break; | ||
1651 | case MPCG_STATE_XID7INITW: | ||
1652 | case MPCG_STATE_XID7INITX: | ||
1653 | case MPCG_STATE_XID7INITI: | ||
1654 | case MPCG_STATE_XID7INITZ: | ||
1655 | switch (fsm_getstate(ch->fsm)) { | ||
1656 | case CH_XID7_PENDING: | ||
1657 | fsm_newstate(ch->fsm, CH_XID7_PENDING1); | ||
1658 | break; | ||
1659 | case CH_XID7_PENDING2: | ||
1660 | fsm_newstate(ch->fsm, CH_XID7_PENDING3); | ||
1661 | break; | ||
1662 | } | ||
1663 | fsm_event(grp->fsm, MPCG_EVENT_XID7DONE, dev); | ||
1664 | break; | ||
1665 | } | ||
1666 | |||
1667 | if (do_debug) | ||
1668 | ctcm_pr_debug("ctcmpc exit : %s(): cp=%i ch=0x%p id=%s\n", | ||
1669 | __FUNCTION__, smp_processor_id(), ch, ch->id); | ||
1670 | return; | ||
1671 | |||
1672 | } | ||
1673 | |||
1674 | /* | ||
1675 | * ctcmpc channel FSM action | ||
1676 | * called from one point in ctcmpc_ch_fsm | ||
1677 | * ctcmpc only | ||
1678 | */ | ||
1679 | static void ctcmpc_chx_attnbusy(fsm_instance *fsm, int event, void *arg) | ||
1680 | { | ||
1681 | struct channel *ch = arg; | ||
1682 | struct net_device *dev = ch->netdev; | ||
1683 | struct ctcm_priv *priv = dev->priv; | ||
1684 | struct mpc_group *grp = priv->mpcg; | ||
1685 | |||
1686 | ctcm_pr_debug("ctcmpc enter: %s %s() %s \nGrpState:%s ChState:%s\n", | ||
1687 | dev->name, | ||
1688 | __FUNCTION__, ch->id, | ||
1689 | fsm_getstate_str(grp->fsm), | ||
1690 | fsm_getstate_str(ch->fsm)); | ||
1691 | |||
1692 | fsm_deltimer(&ch->timer); | ||
1693 | |||
1694 | switch (fsm_getstate(grp->fsm)) { | ||
1695 | case MPCG_STATE_XID0IOWAIT: | ||
1696 | /* vtam wants to be primary.start yside xid exchanges*/ | ||
1697 | /* only receive one attn-busy at a time so must not */ | ||
1698 | /* change state each time */ | ||
1699 | grp->changed_side = 1; | ||
1700 | fsm_newstate(grp->fsm, MPCG_STATE_XID2INITW); | ||
1701 | break; | ||
1702 | case MPCG_STATE_XID2INITW: | ||
1703 | if (grp->changed_side == 1) { | ||
1704 | grp->changed_side = 2; | ||
1705 | break; | ||
1706 | } | ||
1707 | /* process began via call to establish_conn */ | ||
1708 | /* so must report failure instead of reverting */ | ||
1709 | /* back to ready-for-xid passive state */ | ||
1710 | if (grp->estconnfunc) | ||
1711 | goto done; | ||
1712 | /* this attnbusy is NOT the result of xside xid */ | ||
1713 | /* collisions so yside must have been triggered */ | ||
1714 | /* by an ATTN that was not intended to start XID */ | ||
1715 | /* processing. Revert back to ready-for-xid and */ | ||
1716 | /* wait for ATTN interrupt to signal xid start */ | ||
1717 | if (fsm_getstate(ch->fsm) == CH_XID0_INPROGRESS) { | ||
1718 | fsm_newstate(ch->fsm, CH_XID0_PENDING) ; | ||
1719 | fsm_deltimer(&grp->timer); | ||
1720 | goto done; | ||
1721 | } | ||
1722 | fsm_event(grp->fsm, MPCG_EVENT_INOP, dev); | ||
1723 | goto done; | ||
1724 | case MPCG_STATE_XID2INITX: | ||
1725 | /* XID2 was received before ATTN Busy for second | ||
1726 | channel.Send yside xid for second channel. | ||
1727 | */ | ||
1728 | if (grp->changed_side == 1) { | ||
1729 | grp->changed_side = 2; | ||
1730 | break; | ||
1731 | } | ||
1732 | case MPCG_STATE_XID0IOWAIX: | ||
1733 | case MPCG_STATE_XID7INITW: | ||
1734 | case MPCG_STATE_XID7INITX: | ||
1735 | case MPCG_STATE_XID7INITI: | ||
1736 | case MPCG_STATE_XID7INITZ: | ||
1737 | default: | ||
1738 | /* multiple attn-busy indicates too out-of-sync */ | ||
1739 | /* and they are certainly not being received as part */ | ||
1740 | /* of valid mpc group negotiations.. */ | ||
1741 | fsm_event(grp->fsm, MPCG_EVENT_INOP, dev); | ||
1742 | goto done; | ||
1743 | } | ||
1744 | |||
1745 | if (grp->changed_side == 1) { | ||
1746 | fsm_deltimer(&grp->timer); | ||
1747 | fsm_addtimer(&grp->timer, MPC_XID_TIMEOUT_VALUE, | ||
1748 | MPCG_EVENT_TIMER, dev); | ||
1749 | } | ||
1750 | if (ch->in_mpcgroup) | ||
1751 | fsm_event(grp->fsm, MPCG_EVENT_XID0DO, ch); | ||
1752 | else | ||
1753 | printk(KERN_WARNING "ctcmpc: %s() Not all channels have" | ||
1754 | " been added to group\n", __FUNCTION__); | ||
1755 | |||
1756 | done: | ||
1757 | if (do_debug) | ||
1758 | ctcm_pr_debug("ctcmpc exit : %s()%s ch=0x%p id=%s\n", | ||
1759 | __FUNCTION__, dev->name, ch, ch->id); | ||
1760 | |||
1761 | return; | ||
1762 | |||
1763 | } | ||
1764 | |||
1765 | /* | ||
1766 | * ctcmpc channel FSM action | ||
1767 | * called from several points in ctcmpc_ch_fsm | ||
1768 | * ctcmpc only | ||
1769 | */ | ||
1770 | static void ctcmpc_chx_resend(fsm_instance *fsm, int event, void *arg) | ||
1771 | { | ||
1772 | struct channel *ch = arg; | ||
1773 | struct net_device *dev = ch->netdev; | ||
1774 | struct ctcm_priv *priv = dev->priv; | ||
1775 | struct mpc_group *grp = priv->mpcg; | ||
1776 | |||
1777 | ctcm_pr_debug("ctcmpc enter: %s %s() %s \nGrpState:%s ChState:%s\n", | ||
1778 | dev->name, __FUNCTION__, ch->id, | ||
1779 | fsm_getstate_str(grp->fsm), | ||
1780 | fsm_getstate_str(ch->fsm)); | ||
1781 | |||
1782 | fsm_event(grp->fsm, MPCG_EVENT_XID0DO, ch); | ||
1783 | |||
1784 | return; | ||
1785 | } | ||
1786 | |||
1787 | /* | ||
1788 | * ctcmpc channel FSM action | ||
1789 | * called from several points in ctcmpc_ch_fsm | ||
1790 | * ctcmpc only | ||
1791 | */ | ||
1792 | static void ctcmpc_chx_send_sweep(fsm_instance *fsm, int event, void *arg) | ||
1793 | { | ||
1794 | struct channel *ach = arg; | ||
1795 | struct net_device *dev = ach->netdev; | ||
1796 | struct ctcm_priv *priv = dev->priv; | ||
1797 | struct mpc_group *grp = priv->mpcg; | ||
1798 | struct channel *wch = priv->channel[WRITE]; | ||
1799 | struct channel *rch = priv->channel[READ]; | ||
1800 | struct sk_buff *skb; | ||
1801 | struct th_sweep *header; | ||
1802 | int rc = 0; | ||
1803 | unsigned long saveflags = 0; | ||
1804 | |||
1805 | if (do_debug) | ||
1806 | ctcm_pr_debug("ctcmpc enter: %s(): cp=%i ch=0x%p id=%s\n", | ||
1807 | __FUNCTION__, smp_processor_id(), ach, ach->id); | ||
1808 | |||
1809 | if (grp->in_sweep == 0) | ||
1810 | goto done; | ||
1811 | |||
1812 | if (do_debug_data) { | ||
1813 | ctcm_pr_debug("ctcmpc: %s() 1: ToVTAM_th_seq= %08x\n" , | ||
1814 | __FUNCTION__, wch->th_seq_num); | ||
1815 | ctcm_pr_debug("ctcmpc: %s() 1: FromVTAM_th_seq= %08x\n" , | ||
1816 | __FUNCTION__, rch->th_seq_num); | ||
1817 | } | ||
1818 | |||
1819 | if (fsm_getstate(wch->fsm) != CTC_STATE_TXIDLE) { | ||
1820 | /* give the previous IO time to complete */ | ||
1821 | fsm_addtimer(&wch->sweep_timer, | ||
1822 | 200, CTC_EVENT_RSWEEP_TIMER, wch); | ||
1823 | goto done; | ||
1824 | } | ||
1825 | |||
1826 | skb = skb_dequeue(&wch->sweep_queue); | ||
1827 | if (!skb) | ||
1828 | goto done; | ||
1829 | |||
1830 | if (set_normalized_cda(&wch->ccw[4], skb->data)) { | ||
1831 | grp->in_sweep = 0; | ||
1832 | ctcm_clear_busy_do(dev); | ||
1833 | dev_kfree_skb_any(skb); | ||
1834 | fsm_event(grp->fsm, MPCG_EVENT_INOP, dev); | ||
1835 | goto done; | ||
1836 | } else { | ||
1837 | atomic_inc(&skb->users); | ||
1838 | skb_queue_tail(&wch->io_queue, skb); | ||
1839 | } | ||
1840 | |||
1841 | /* send out the sweep */ | ||
1842 | wch->ccw[4].count = skb->len; | ||
1843 | |||
1844 | header = (struct th_sweep *)skb->data; | ||
1845 | switch (header->th.th_ch_flag) { | ||
1846 | case TH_SWEEP_REQ: | ||
1847 | grp->sweep_req_pend_num--; | ||
1848 | break; | ||
1849 | case TH_SWEEP_RESP: | ||
1850 | grp->sweep_rsp_pend_num--; | ||
1851 | break; | ||
1852 | } | ||
1853 | |||
1854 | header->sw.th_last_seq = wch->th_seq_num; | ||
1855 | |||
1856 | if (do_debug_ccw) | ||
1857 | ctcmpc_dumpit((char *)&wch->ccw[3], sizeof(struct ccw1) * 3); | ||
1858 | |||
1859 | ctcm_pr_debug("ctcmpc: %s() sweep packet\n", __FUNCTION__); | ||
1860 | ctcmpc_dumpit((char *)header, TH_SWEEP_LENGTH); | ||
1861 | |||
1862 | fsm_addtimer(&wch->timer, CTCM_TIME_5_SEC, CTC_EVENT_TIMER, wch); | ||
1863 | fsm_newstate(wch->fsm, CTC_STATE_TX); | ||
1864 | |||
1865 | spin_lock_irqsave(get_ccwdev_lock(wch->cdev), saveflags); | ||
1866 | wch->prof.send_stamp = current_kernel_time(); /* xtime */ | ||
1867 | rc = ccw_device_start(wch->cdev, &wch->ccw[3], | ||
1868 | (unsigned long) wch, 0xff, 0); | ||
1869 | spin_unlock_irqrestore(get_ccwdev_lock(wch->cdev), saveflags); | ||
1870 | |||
1871 | if ((grp->sweep_req_pend_num == 0) && | ||
1872 | (grp->sweep_rsp_pend_num == 0)) { | ||
1873 | grp->in_sweep = 0; | ||
1874 | rch->th_seq_num = 0x00; | ||
1875 | wch->th_seq_num = 0x00; | ||
1876 | ctcm_clear_busy_do(dev); | ||
1877 | } | ||
1878 | |||
1879 | if (do_debug_data) { | ||
1880 | ctcm_pr_debug("ctcmpc: %s()2: ToVTAM_th_seq= %08x\n" , | ||
1881 | __FUNCTION__, wch->th_seq_num); | ||
1882 | ctcm_pr_debug("ctcmpc: %s()2: FromVTAM_th_seq= %08x\n" , | ||
1883 | __FUNCTION__, rch->th_seq_num); | ||
1884 | } | ||
1885 | |||
1886 | if (rc != 0) | ||
1887 | ctcm_ccw_check_rc(wch, rc, "send sweep"); | ||
1888 | |||
1889 | done: | ||
1890 | if (do_debug) | ||
1891 | ctcm_pr_debug("ctcmpc exit: %s() %s\n", __FUNCTION__, ach->id); | ||
1892 | return; | ||
1893 | } | ||
1894 | |||
1895 | |||
1896 | /* | ||
1897 | * The ctcmpc statemachine for a channel. | ||
1898 | */ | ||
1899 | |||
1900 | const fsm_node ctcmpc_ch_fsm[] = { | ||
1901 | { CTC_STATE_STOPPED, CTC_EVENT_STOP, ctcm_action_nop }, | ||
1902 | { CTC_STATE_STOPPED, CTC_EVENT_START, ctcm_chx_start }, | ||
1903 | { CTC_STATE_STOPPED, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
1904 | { CTC_STATE_STOPPED, CTC_EVENT_FINSTAT, ctcm_action_nop }, | ||
1905 | { CTC_STATE_STOPPED, CTC_EVENT_MC_FAIL, ctcm_action_nop }, | ||
1906 | |||
1907 | { CTC_STATE_NOTOP, CTC_EVENT_STOP, ctcm_chx_stop }, | ||
1908 | { CTC_STATE_NOTOP, CTC_EVENT_START, ctcm_action_nop }, | ||
1909 | { CTC_STATE_NOTOP, CTC_EVENT_FINSTAT, ctcm_action_nop }, | ||
1910 | { CTC_STATE_NOTOP, CTC_EVENT_MC_FAIL, ctcm_action_nop }, | ||
1911 | { CTC_STATE_NOTOP, CTC_EVENT_MC_GOOD, ctcm_chx_start }, | ||
1912 | { CTC_STATE_NOTOP, CTC_EVENT_UC_RCRESET, ctcm_chx_stop }, | ||
1913 | { CTC_STATE_NOTOP, CTC_EVENT_UC_RSRESET, ctcm_chx_stop }, | ||
1914 | { CTC_STATE_NOTOP, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
1915 | |||
1916 | { CTC_STATE_STARTWAIT, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
1917 | { CTC_STATE_STARTWAIT, CTC_EVENT_START, ctcm_action_nop }, | ||
1918 | { CTC_STATE_STARTWAIT, CTC_EVENT_FINSTAT, ctcm_chx_setmode }, | ||
1919 | { CTC_STATE_STARTWAIT, CTC_EVENT_TIMER, ctcm_chx_setuperr }, | ||
1920 | { CTC_STATE_STARTWAIT, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
1921 | { CTC_STATE_STARTWAIT, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
1922 | |||
1923 | { CTC_STATE_STARTRETRY, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
1924 | { CTC_STATE_STARTRETRY, CTC_EVENT_TIMER, ctcm_chx_setmode }, | ||
1925 | { CTC_STATE_STARTRETRY, CTC_EVENT_FINSTAT, ctcm_chx_setmode }, | ||
1926 | { CTC_STATE_STARTRETRY, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
1927 | { CTC_STATE_STARTRETRY, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
1928 | |||
1929 | { CTC_STATE_SETUPWAIT, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
1930 | { CTC_STATE_SETUPWAIT, CTC_EVENT_START, ctcm_action_nop }, | ||
1931 | { CTC_STATE_SETUPWAIT, CTC_EVENT_FINSTAT, ctcmpc_chx_firstio }, | ||
1932 | { CTC_STATE_SETUPWAIT, CTC_EVENT_UC_RCRESET, ctcm_chx_setuperr }, | ||
1933 | { CTC_STATE_SETUPWAIT, CTC_EVENT_UC_RSRESET, ctcm_chx_setuperr }, | ||
1934 | { CTC_STATE_SETUPWAIT, CTC_EVENT_TIMER, ctcm_chx_setmode }, | ||
1935 | { CTC_STATE_SETUPWAIT, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
1936 | { CTC_STATE_SETUPWAIT, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
1937 | |||
1938 | { CTC_STATE_RXINIT, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
1939 | { CTC_STATE_RXINIT, CTC_EVENT_START, ctcm_action_nop }, | ||
1940 | { CTC_STATE_RXINIT, CTC_EVENT_FINSTAT, ctcmpc_chx_rxidle }, | ||
1941 | { CTC_STATE_RXINIT, CTC_EVENT_UC_RCRESET, ctcm_chx_rxiniterr }, | ||
1942 | { CTC_STATE_RXINIT, CTC_EVENT_UC_RSRESET, ctcm_chx_rxiniterr }, | ||
1943 | { CTC_STATE_RXINIT, CTC_EVENT_TIMER, ctcm_chx_rxiniterr }, | ||
1944 | { CTC_STATE_RXINIT, CTC_EVENT_ATTNBUSY, ctcm_chx_rxinitfail }, | ||
1945 | { CTC_STATE_RXINIT, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
1946 | { CTC_STATE_RXINIT, CTC_EVENT_UC_ZERO, ctcmpc_chx_firstio }, | ||
1947 | { CTC_STATE_RXINIT, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
1948 | |||
1949 | { CH_XID0_PENDING, CTC_EVENT_FINSTAT, ctcm_action_nop }, | ||
1950 | { CH_XID0_PENDING, CTC_EVENT_ATTN, ctcmpc_chx_attn }, | ||
1951 | { CH_XID0_PENDING, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
1952 | { CH_XID0_PENDING, CTC_EVENT_START, ctcm_action_nop }, | ||
1953 | { CH_XID0_PENDING, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
1954 | { CH_XID0_PENDING, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
1955 | { CH_XID0_PENDING, CTC_EVENT_UC_RCRESET, ctcm_chx_setuperr }, | ||
1956 | { CH_XID0_PENDING, CTC_EVENT_UC_RSRESET, ctcm_chx_setuperr }, | ||
1957 | { CH_XID0_PENDING, CTC_EVENT_UC_RSRESET, ctcm_chx_setuperr }, | ||
1958 | { CH_XID0_PENDING, CTC_EVENT_ATTNBUSY, ctcm_chx_iofatal }, | ||
1959 | |||
1960 | { CH_XID0_INPROGRESS, CTC_EVENT_FINSTAT, ctcmpc_chx_rx }, | ||
1961 | { CH_XID0_INPROGRESS, CTC_EVENT_ATTN, ctcmpc_chx_attn }, | ||
1962 | { CH_XID0_INPROGRESS, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
1963 | { CH_XID0_INPROGRESS, CTC_EVENT_START, ctcm_action_nop }, | ||
1964 | { CH_XID0_INPROGRESS, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
1965 | { CH_XID0_INPROGRESS, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
1966 | { CH_XID0_INPROGRESS, CTC_EVENT_UC_ZERO, ctcmpc_chx_rx }, | ||
1967 | { CH_XID0_INPROGRESS, CTC_EVENT_UC_RCRESET, ctcm_chx_setuperr }, | ||
1968 | { CH_XID0_INPROGRESS, CTC_EVENT_ATTNBUSY, ctcmpc_chx_attnbusy }, | ||
1969 | { CH_XID0_INPROGRESS, CTC_EVENT_TIMER, ctcmpc_chx_resend }, | ||
1970 | { CH_XID0_INPROGRESS, CTC_EVENT_IO_EBUSY, ctcm_chx_fail }, | ||
1971 | |||
1972 | { CH_XID7_PENDING, CTC_EVENT_FINSTAT, ctcmpc_chx_rx }, | ||
1973 | { CH_XID7_PENDING, CTC_EVENT_ATTN, ctcmpc_chx_attn }, | ||
1974 | { CH_XID7_PENDING, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
1975 | { CH_XID7_PENDING, CTC_EVENT_START, ctcm_action_nop }, | ||
1976 | { CH_XID7_PENDING, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
1977 | { CH_XID7_PENDING, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
1978 | { CH_XID7_PENDING, CTC_EVENT_UC_ZERO, ctcmpc_chx_rx }, | ||
1979 | { CH_XID7_PENDING, CTC_EVENT_UC_RCRESET, ctcm_chx_setuperr }, | ||
1980 | { CH_XID7_PENDING, CTC_EVENT_UC_RSRESET, ctcm_chx_setuperr }, | ||
1981 | { CH_XID7_PENDING, CTC_EVENT_UC_RSRESET, ctcm_chx_setuperr }, | ||
1982 | { CH_XID7_PENDING, CTC_EVENT_ATTNBUSY, ctcm_chx_iofatal }, | ||
1983 | { CH_XID7_PENDING, CTC_EVENT_TIMER, ctcmpc_chx_resend }, | ||
1984 | { CH_XID7_PENDING, CTC_EVENT_IO_EBUSY, ctcm_chx_fail }, | ||
1985 | |||
1986 | { CH_XID7_PENDING1, CTC_EVENT_FINSTAT, ctcmpc_chx_rx }, | ||
1987 | { CH_XID7_PENDING1, CTC_EVENT_ATTN, ctcmpc_chx_attn }, | ||
1988 | { CH_XID7_PENDING1, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
1989 | { CH_XID7_PENDING1, CTC_EVENT_START, ctcm_action_nop }, | ||
1990 | { CH_XID7_PENDING1, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
1991 | { CH_XID7_PENDING1, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
1992 | { CH_XID7_PENDING1, CTC_EVENT_UC_ZERO, ctcmpc_chx_rx }, | ||
1993 | { CH_XID7_PENDING1, CTC_EVENT_UC_RCRESET, ctcm_chx_setuperr }, | ||
1994 | { CH_XID7_PENDING1, CTC_EVENT_UC_RSRESET, ctcm_chx_setuperr }, | ||
1995 | { CH_XID7_PENDING1, CTC_EVENT_ATTNBUSY, ctcm_chx_iofatal }, | ||
1996 | { CH_XID7_PENDING1, CTC_EVENT_TIMER, ctcmpc_chx_resend }, | ||
1997 | { CH_XID7_PENDING1, CTC_EVENT_IO_EBUSY, ctcm_chx_fail }, | ||
1998 | |||
1999 | { CH_XID7_PENDING2, CTC_EVENT_FINSTAT, ctcmpc_chx_rx }, | ||
2000 | { CH_XID7_PENDING2, CTC_EVENT_ATTN, ctcmpc_chx_attn }, | ||
2001 | { CH_XID7_PENDING2, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
2002 | { CH_XID7_PENDING2, CTC_EVENT_START, ctcm_action_nop }, | ||
2003 | { CH_XID7_PENDING2, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
2004 | { CH_XID7_PENDING2, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
2005 | { CH_XID7_PENDING2, CTC_EVENT_UC_ZERO, ctcmpc_chx_rx }, | ||
2006 | { CH_XID7_PENDING2, CTC_EVENT_UC_RCRESET, ctcm_chx_setuperr }, | ||
2007 | { CH_XID7_PENDING2, CTC_EVENT_UC_RSRESET, ctcm_chx_setuperr }, | ||
2008 | { CH_XID7_PENDING2, CTC_EVENT_ATTNBUSY, ctcm_chx_iofatal }, | ||
2009 | { CH_XID7_PENDING2, CTC_EVENT_TIMER, ctcmpc_chx_resend }, | ||
2010 | { CH_XID7_PENDING2, CTC_EVENT_IO_EBUSY, ctcm_chx_fail }, | ||
2011 | |||
2012 | { CH_XID7_PENDING3, CTC_EVENT_FINSTAT, ctcmpc_chx_rx }, | ||
2013 | { CH_XID7_PENDING3, CTC_EVENT_ATTN, ctcmpc_chx_attn }, | ||
2014 | { CH_XID7_PENDING3, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
2015 | { CH_XID7_PENDING3, CTC_EVENT_START, ctcm_action_nop }, | ||
2016 | { CH_XID7_PENDING3, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
2017 | { CH_XID7_PENDING3, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
2018 | { CH_XID7_PENDING3, CTC_EVENT_UC_ZERO, ctcmpc_chx_rx }, | ||
2019 | { CH_XID7_PENDING3, CTC_EVENT_UC_RCRESET, ctcm_chx_setuperr }, | ||
2020 | { CH_XID7_PENDING3, CTC_EVENT_UC_RSRESET, ctcm_chx_setuperr }, | ||
2021 | { CH_XID7_PENDING3, CTC_EVENT_ATTNBUSY, ctcm_chx_iofatal }, | ||
2022 | { CH_XID7_PENDING3, CTC_EVENT_TIMER, ctcmpc_chx_resend }, | ||
2023 | { CH_XID7_PENDING3, CTC_EVENT_IO_EBUSY, ctcm_chx_fail }, | ||
2024 | |||
2025 | { CH_XID7_PENDING4, CTC_EVENT_FINSTAT, ctcmpc_chx_rx }, | ||
2026 | { CH_XID7_PENDING4, CTC_EVENT_ATTN, ctcmpc_chx_attn }, | ||
2027 | { CH_XID7_PENDING4, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
2028 | { CH_XID7_PENDING4, CTC_EVENT_START, ctcm_action_nop }, | ||
2029 | { CH_XID7_PENDING4, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
2030 | { CH_XID7_PENDING4, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
2031 | { CH_XID7_PENDING4, CTC_EVENT_UC_ZERO, ctcmpc_chx_rx }, | ||
2032 | { CH_XID7_PENDING4, CTC_EVENT_UC_RCRESET, ctcm_chx_setuperr }, | ||
2033 | { CH_XID7_PENDING4, CTC_EVENT_UC_RSRESET, ctcm_chx_setuperr }, | ||
2034 | { CH_XID7_PENDING4, CTC_EVENT_ATTNBUSY, ctcm_chx_iofatal }, | ||
2035 | { CH_XID7_PENDING4, CTC_EVENT_TIMER, ctcmpc_chx_resend }, | ||
2036 | { CH_XID7_PENDING4, CTC_EVENT_IO_EBUSY, ctcm_chx_fail }, | ||
2037 | |||
2038 | { CTC_STATE_RXIDLE, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
2039 | { CTC_STATE_RXIDLE, CTC_EVENT_START, ctcm_action_nop }, | ||
2040 | { CTC_STATE_RXIDLE, CTC_EVENT_FINSTAT, ctcmpc_chx_rx }, | ||
2041 | { CTC_STATE_RXIDLE, CTC_EVENT_UC_RCRESET, ctcm_chx_rxdisc }, | ||
2042 | { CTC_STATE_RXIDLE, CTC_EVENT_UC_RSRESET, ctcm_chx_fail }, | ||
2043 | { CTC_STATE_RXIDLE, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
2044 | { CTC_STATE_RXIDLE, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
2045 | { CTC_STATE_RXIDLE, CTC_EVENT_UC_ZERO, ctcmpc_chx_rx }, | ||
2046 | |||
2047 | { CTC_STATE_TXINIT, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
2048 | { CTC_STATE_TXINIT, CTC_EVENT_START, ctcm_action_nop }, | ||
2049 | { CTC_STATE_TXINIT, CTC_EVENT_FINSTAT, ctcm_chx_txidle }, | ||
2050 | { CTC_STATE_TXINIT, CTC_EVENT_UC_RCRESET, ctcm_chx_txiniterr }, | ||
2051 | { CTC_STATE_TXINIT, CTC_EVENT_UC_RSRESET, ctcm_chx_txiniterr }, | ||
2052 | { CTC_STATE_TXINIT, CTC_EVENT_TIMER, ctcm_chx_txiniterr }, | ||
2053 | { CTC_STATE_TXINIT, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
2054 | { CTC_STATE_TXINIT, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
2055 | { CTC_STATE_TXINIT, CTC_EVENT_RSWEEP_TIMER, ctcmpc_chx_send_sweep }, | ||
2056 | |||
2057 | { CTC_STATE_TXIDLE, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
2058 | { CTC_STATE_TXIDLE, CTC_EVENT_START, ctcm_action_nop }, | ||
2059 | { CTC_STATE_TXIDLE, CTC_EVENT_FINSTAT, ctcmpc_chx_firstio }, | ||
2060 | { CTC_STATE_TXIDLE, CTC_EVENT_UC_RCRESET, ctcm_chx_fail }, | ||
2061 | { CTC_STATE_TXIDLE, CTC_EVENT_UC_RSRESET, ctcm_chx_fail }, | ||
2062 | { CTC_STATE_TXIDLE, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
2063 | { CTC_STATE_TXIDLE, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
2064 | { CTC_STATE_TXIDLE, CTC_EVENT_RSWEEP_TIMER, ctcmpc_chx_send_sweep }, | ||
2065 | |||
2066 | { CTC_STATE_TERM, CTC_EVENT_STOP, ctcm_action_nop }, | ||
2067 | { CTC_STATE_TERM, CTC_EVENT_START, ctcm_chx_restart }, | ||
2068 | { CTC_STATE_TERM, CTC_EVENT_FINSTAT, ctcm_chx_stopped }, | ||
2069 | { CTC_STATE_TERM, CTC_EVENT_UC_RCRESET, ctcm_action_nop }, | ||
2070 | { CTC_STATE_TERM, CTC_EVENT_UC_RSRESET, ctcm_action_nop }, | ||
2071 | { CTC_STATE_TERM, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
2072 | { CTC_STATE_TERM, CTC_EVENT_IO_EBUSY, ctcm_chx_fail }, | ||
2073 | { CTC_STATE_TERM, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
2074 | |||
2075 | { CTC_STATE_DTERM, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
2076 | { CTC_STATE_DTERM, CTC_EVENT_START, ctcm_chx_restart }, | ||
2077 | { CTC_STATE_DTERM, CTC_EVENT_FINSTAT, ctcm_chx_setmode }, | ||
2078 | { CTC_STATE_DTERM, CTC_EVENT_UC_RCRESET, ctcm_action_nop }, | ||
2079 | { CTC_STATE_DTERM, CTC_EVENT_UC_RSRESET, ctcm_action_nop }, | ||
2080 | { CTC_STATE_DTERM, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
2081 | { CTC_STATE_DTERM, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
2082 | |||
2083 | { CTC_STATE_TX, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
2084 | { CTC_STATE_TX, CTC_EVENT_START, ctcm_action_nop }, | ||
2085 | { CTC_STATE_TX, CTC_EVENT_FINSTAT, ctcmpc_chx_txdone }, | ||
2086 | { CTC_STATE_TX, CTC_EVENT_UC_RCRESET, ctcm_chx_fail }, | ||
2087 | { CTC_STATE_TX, CTC_EVENT_UC_RSRESET, ctcm_chx_fail }, | ||
2088 | { CTC_STATE_TX, CTC_EVENT_TIMER, ctcm_chx_txretry }, | ||
2089 | { CTC_STATE_TX, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
2090 | { CTC_STATE_TX, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
2091 | { CTC_STATE_TX, CTC_EVENT_RSWEEP_TIMER, ctcmpc_chx_send_sweep }, | ||
2092 | { CTC_STATE_TX, CTC_EVENT_IO_EBUSY, ctcm_chx_fail }, | ||
2093 | |||
2094 | { CTC_STATE_RXERR, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
2095 | { CTC_STATE_TXERR, CTC_EVENT_STOP, ctcm_chx_haltio }, | ||
2096 | { CTC_STATE_TXERR, CTC_EVENT_IO_ENODEV, ctcm_chx_iofatal }, | ||
2097 | { CTC_STATE_TXERR, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
2098 | { CTC_STATE_RXERR, CTC_EVENT_MC_FAIL, ctcm_chx_fail }, | ||
2099 | }; | ||
2100 | |||
2101 | int mpc_ch_fsm_len = ARRAY_SIZE(ctcmpc_ch_fsm); | ||
2102 | |||
2103 | /* | ||
2104 | * Actions for interface - statemachine. | ||
2105 | */ | ||
2106 | |||
2107 | /** | ||
2108 | * Startup channels by sending CTC_EVENT_START to each channel. | ||
2109 | * | ||
2110 | * fi An instance of an interface statemachine. | ||
2111 | * event The event, just happened. | ||
2112 | * arg Generic pointer, casted from struct net_device * upon call. | ||
2113 | */ | ||
2114 | static void dev_action_start(fsm_instance *fi, int event, void *arg) | ||
2115 | { | ||
2116 | struct net_device *dev = arg; | ||
2117 | struct ctcm_priv *priv = dev->priv; | ||
2118 | int direction; | ||
2119 | |||
2120 | CTCMY_DBF_DEV_NAME(SETUP, dev, ""); | ||
2121 | |||
2122 | fsm_deltimer(&priv->restart_timer); | ||
2123 | fsm_newstate(fi, DEV_STATE_STARTWAIT_RXTX); | ||
2124 | if (IS_MPC(priv)) | ||
2125 | priv->mpcg->channels_terminating = 0; | ||
2126 | for (direction = READ; direction <= WRITE; direction++) { | ||
2127 | struct channel *ch = priv->channel[direction]; | ||
2128 | fsm_event(ch->fsm, CTC_EVENT_START, ch); | ||
2129 | } | ||
2130 | } | ||
2131 | |||
2132 | /** | ||
2133 | * Shutdown channels by sending CTC_EVENT_STOP to each channel. | ||
2134 | * | ||
2135 | * fi An instance of an interface statemachine. | ||
2136 | * event The event, just happened. | ||
2137 | * arg Generic pointer, casted from struct net_device * upon call. | ||
2138 | */ | ||
2139 | static void dev_action_stop(fsm_instance *fi, int event, void *arg) | ||
2140 | { | ||
2141 | int direction; | ||
2142 | struct net_device *dev = arg; | ||
2143 | struct ctcm_priv *priv = dev->priv; | ||
2144 | |||
2145 | CTCMY_DBF_DEV_NAME(SETUP, dev, ""); | ||
2146 | |||
2147 | fsm_newstate(fi, DEV_STATE_STOPWAIT_RXTX); | ||
2148 | for (direction = READ; direction <= WRITE; direction++) { | ||
2149 | struct channel *ch = priv->channel[direction]; | ||
2150 | fsm_event(ch->fsm, CTC_EVENT_STOP, ch); | ||
2151 | ch->th_seq_num = 0x00; | ||
2152 | if (do_debug) | ||
2153 | ctcm_pr_debug("ctcm: %s() CH_th_seq= %08x\n", | ||
2154 | __FUNCTION__, ch->th_seq_num); | ||
2155 | } | ||
2156 | if (IS_MPC(priv)) | ||
2157 | fsm_newstate(priv->mpcg->fsm, MPCG_STATE_RESET); | ||
2158 | } | ||
2159 | |||
2160 | static void dev_action_restart(fsm_instance *fi, int event, void *arg) | ||
2161 | { | ||
2162 | int restart_timer; | ||
2163 | struct net_device *dev = arg; | ||
2164 | struct ctcm_priv *priv = dev->priv; | ||
2165 | |||
2166 | CTCMY_DBF_DEV_NAME(TRACE, dev, ""); | ||
2167 | |||
2168 | if (IS_MPC(priv)) { | ||
2169 | ctcm_pr_info("ctcm: %s Restarting Device and " | ||
2170 | "MPC Group in 5 seconds\n", | ||
2171 | dev->name); | ||
2172 | restart_timer = CTCM_TIME_1_SEC; | ||
2173 | } else { | ||
2174 | ctcm_pr_info("%s: Restarting\n", dev->name); | ||
2175 | restart_timer = CTCM_TIME_5_SEC; | ||
2176 | } | ||
2177 | |||
2178 | dev_action_stop(fi, event, arg); | ||
2179 | fsm_event(priv->fsm, DEV_EVENT_STOP, dev); | ||
2180 | if (IS_MPC(priv)) | ||
2181 | fsm_newstate(priv->mpcg->fsm, MPCG_STATE_RESET); | ||
2182 | |||
2183 | /* going back into start sequence too quickly can */ | ||
2184 | /* result in the other side becoming unreachable due */ | ||
2185 | /* to sense reported when IO is aborted */ | ||
2186 | fsm_addtimer(&priv->restart_timer, restart_timer, | ||
2187 | DEV_EVENT_START, dev); | ||
2188 | } | ||
2189 | |||
2190 | /** | ||
2191 | * Called from channel statemachine | ||
2192 | * when a channel is up and running. | ||
2193 | * | ||
2194 | * fi An instance of an interface statemachine. | ||
2195 | * event The event, just happened. | ||
2196 | * arg Generic pointer, casted from struct net_device * upon call. | ||
2197 | */ | ||
2198 | static void dev_action_chup(fsm_instance *fi, int event, void *arg) | ||
2199 | { | ||
2200 | struct net_device *dev = arg; | ||
2201 | struct ctcm_priv *priv = dev->priv; | ||
2202 | |||
2203 | CTCMY_DBF_DEV_NAME(SETUP, dev, ""); | ||
2204 | |||
2205 | switch (fsm_getstate(fi)) { | ||
2206 | case DEV_STATE_STARTWAIT_RXTX: | ||
2207 | if (event == DEV_EVENT_RXUP) | ||
2208 | fsm_newstate(fi, DEV_STATE_STARTWAIT_TX); | ||
2209 | else | ||
2210 | fsm_newstate(fi, DEV_STATE_STARTWAIT_RX); | ||
2211 | break; | ||
2212 | case DEV_STATE_STARTWAIT_RX: | ||
2213 | if (event == DEV_EVENT_RXUP) { | ||
2214 | fsm_newstate(fi, DEV_STATE_RUNNING); | ||
2215 | ctcm_pr_info("%s: connected with remote side\n", | ||
2216 | dev->name); | ||
2217 | ctcm_clear_busy(dev); | ||
2218 | } | ||
2219 | break; | ||
2220 | case DEV_STATE_STARTWAIT_TX: | ||
2221 | if (event == DEV_EVENT_TXUP) { | ||
2222 | fsm_newstate(fi, DEV_STATE_RUNNING); | ||
2223 | ctcm_pr_info("%s: connected with remote side\n", | ||
2224 | dev->name); | ||
2225 | ctcm_clear_busy(dev); | ||
2226 | } | ||
2227 | break; | ||
2228 | case DEV_STATE_STOPWAIT_TX: | ||
2229 | if (event == DEV_EVENT_RXUP) | ||
2230 | fsm_newstate(fi, DEV_STATE_STOPWAIT_RXTX); | ||
2231 | break; | ||
2232 | case DEV_STATE_STOPWAIT_RX: | ||
2233 | if (event == DEV_EVENT_TXUP) | ||
2234 | fsm_newstate(fi, DEV_STATE_STOPWAIT_RXTX); | ||
2235 | break; | ||
2236 | } | ||
2237 | |||
2238 | if (IS_MPC(priv)) { | ||
2239 | if (event == DEV_EVENT_RXUP) | ||
2240 | mpc_channel_action(priv->channel[READ], | ||
2241 | READ, MPC_CHANNEL_ADD); | ||
2242 | else | ||
2243 | mpc_channel_action(priv->channel[WRITE], | ||
2244 | WRITE, MPC_CHANNEL_ADD); | ||
2245 | } | ||
2246 | } | ||
2247 | |||
2248 | /** | ||
2249 | * Called from device statemachine | ||
2250 | * when a channel has been shutdown. | ||
2251 | * | ||
2252 | * fi An instance of an interface statemachine. | ||
2253 | * event The event, just happened. | ||
2254 | * arg Generic pointer, casted from struct net_device * upon call. | ||
2255 | */ | ||
2256 | static void dev_action_chdown(fsm_instance *fi, int event, void *arg) | ||
2257 | { | ||
2258 | |||
2259 | struct net_device *dev = arg; | ||
2260 | struct ctcm_priv *priv = dev->priv; | ||
2261 | |||
2262 | CTCMY_DBF_DEV_NAME(SETUP, dev, ""); | ||
2263 | |||
2264 | switch (fsm_getstate(fi)) { | ||
2265 | case DEV_STATE_RUNNING: | ||
2266 | if (event == DEV_EVENT_TXDOWN) | ||
2267 | fsm_newstate(fi, DEV_STATE_STARTWAIT_TX); | ||
2268 | else | ||
2269 | fsm_newstate(fi, DEV_STATE_STARTWAIT_RX); | ||
2270 | break; | ||
2271 | case DEV_STATE_STARTWAIT_RX: | ||
2272 | if (event == DEV_EVENT_TXDOWN) | ||
2273 | fsm_newstate(fi, DEV_STATE_STARTWAIT_RXTX); | ||
2274 | break; | ||
2275 | case DEV_STATE_STARTWAIT_TX: | ||
2276 | if (event == DEV_EVENT_RXDOWN) | ||
2277 | fsm_newstate(fi, DEV_STATE_STARTWAIT_RXTX); | ||
2278 | break; | ||
2279 | case DEV_STATE_STOPWAIT_RXTX: | ||
2280 | if (event == DEV_EVENT_TXDOWN) | ||
2281 | fsm_newstate(fi, DEV_STATE_STOPWAIT_RX); | ||
2282 | else | ||
2283 | fsm_newstate(fi, DEV_STATE_STOPWAIT_TX); | ||
2284 | break; | ||
2285 | case DEV_STATE_STOPWAIT_RX: | ||
2286 | if (event == DEV_EVENT_RXDOWN) | ||
2287 | fsm_newstate(fi, DEV_STATE_STOPPED); | ||
2288 | break; | ||
2289 | case DEV_STATE_STOPWAIT_TX: | ||
2290 | if (event == DEV_EVENT_TXDOWN) | ||
2291 | fsm_newstate(fi, DEV_STATE_STOPPED); | ||
2292 | break; | ||
2293 | } | ||
2294 | if (IS_MPC(priv)) { | ||
2295 | if (event == DEV_EVENT_RXDOWN) | ||
2296 | mpc_channel_action(priv->channel[READ], | ||
2297 | READ, MPC_CHANNEL_REMOVE); | ||
2298 | else | ||
2299 | mpc_channel_action(priv->channel[WRITE], | ||
2300 | WRITE, MPC_CHANNEL_REMOVE); | ||
2301 | } | ||
2302 | } | ||
2303 | |||
2304 | const fsm_node dev_fsm[] = { | ||
2305 | { DEV_STATE_STOPPED, DEV_EVENT_START, dev_action_start }, | ||
2306 | { DEV_STATE_STOPWAIT_RXTX, DEV_EVENT_START, dev_action_start }, | ||
2307 | { DEV_STATE_STOPWAIT_RXTX, DEV_EVENT_RXDOWN, dev_action_chdown }, | ||
2308 | { DEV_STATE_STOPWAIT_RXTX, DEV_EVENT_TXDOWN, dev_action_chdown }, | ||
2309 | { DEV_STATE_STOPWAIT_RXTX, DEV_EVENT_RESTART, dev_action_restart }, | ||
2310 | { DEV_STATE_STOPWAIT_RX, DEV_EVENT_START, dev_action_start }, | ||
2311 | { DEV_STATE_STOPWAIT_RX, DEV_EVENT_RXUP, dev_action_chup }, | ||
2312 | { DEV_STATE_STOPWAIT_RX, DEV_EVENT_TXUP, dev_action_chup }, | ||
2313 | { DEV_STATE_STOPWAIT_RX, DEV_EVENT_RXDOWN, dev_action_chdown }, | ||
2314 | { DEV_STATE_STOPWAIT_RX, DEV_EVENT_RESTART, dev_action_restart }, | ||
2315 | { DEV_STATE_STOPWAIT_TX, DEV_EVENT_START, dev_action_start }, | ||
2316 | { DEV_STATE_STOPWAIT_TX, DEV_EVENT_RXUP, dev_action_chup }, | ||
2317 | { DEV_STATE_STOPWAIT_TX, DEV_EVENT_TXUP, dev_action_chup }, | ||
2318 | { DEV_STATE_STOPWAIT_TX, DEV_EVENT_TXDOWN, dev_action_chdown }, | ||
2319 | { DEV_STATE_STOPWAIT_TX, DEV_EVENT_RESTART, dev_action_restart }, | ||
2320 | { DEV_STATE_STARTWAIT_RXTX, DEV_EVENT_STOP, dev_action_stop }, | ||
2321 | { DEV_STATE_STARTWAIT_RXTX, DEV_EVENT_RXUP, dev_action_chup }, | ||
2322 | { DEV_STATE_STARTWAIT_RXTX, DEV_EVENT_TXUP, dev_action_chup }, | ||
2323 | { DEV_STATE_STARTWAIT_RXTX, DEV_EVENT_RXDOWN, dev_action_chdown }, | ||
2324 | { DEV_STATE_STARTWAIT_RXTX, DEV_EVENT_TXDOWN, dev_action_chdown }, | ||
2325 | { DEV_STATE_STARTWAIT_RXTX, DEV_EVENT_RESTART, dev_action_restart }, | ||
2326 | { DEV_STATE_STARTWAIT_TX, DEV_EVENT_STOP, dev_action_stop }, | ||
2327 | { DEV_STATE_STARTWAIT_TX, DEV_EVENT_RXUP, dev_action_chup }, | ||
2328 | { DEV_STATE_STARTWAIT_TX, DEV_EVENT_TXUP, dev_action_chup }, | ||
2329 | { DEV_STATE_STARTWAIT_TX, DEV_EVENT_RXDOWN, dev_action_chdown }, | ||
2330 | { DEV_STATE_STARTWAIT_TX, DEV_EVENT_RESTART, dev_action_restart }, | ||
2331 | { DEV_STATE_STARTWAIT_RX, DEV_EVENT_STOP, dev_action_stop }, | ||
2332 | { DEV_STATE_STARTWAIT_RX, DEV_EVENT_RXUP, dev_action_chup }, | ||
2333 | { DEV_STATE_STARTWAIT_RX, DEV_EVENT_TXUP, dev_action_chup }, | ||
2334 | { DEV_STATE_STARTWAIT_RX, DEV_EVENT_TXDOWN, dev_action_chdown }, | ||
2335 | { DEV_STATE_STARTWAIT_RX, DEV_EVENT_RESTART, dev_action_restart }, | ||
2336 | { DEV_STATE_RUNNING, DEV_EVENT_STOP, dev_action_stop }, | ||
2337 | { DEV_STATE_RUNNING, DEV_EVENT_RXDOWN, dev_action_chdown }, | ||
2338 | { DEV_STATE_RUNNING, DEV_EVENT_TXDOWN, dev_action_chdown }, | ||
2339 | { DEV_STATE_RUNNING, DEV_EVENT_TXUP, ctcm_action_nop }, | ||
2340 | { DEV_STATE_RUNNING, DEV_EVENT_RXUP, ctcm_action_nop }, | ||
2341 | { DEV_STATE_RUNNING, DEV_EVENT_RESTART, dev_action_restart }, | ||
2342 | }; | ||
2343 | |||
2344 | int dev_fsm_len = ARRAY_SIZE(dev_fsm); | ||
2345 | |||
2346 | /* --- This is the END my friend --- */ | ||
2347 | |||
diff --git a/drivers/s390/net/ctcm_fsms.h b/drivers/s390/net/ctcm_fsms.h new file mode 100644 index 000000000000..2326aba9807a --- /dev/null +++ b/drivers/s390/net/ctcm_fsms.h | |||
@@ -0,0 +1,359 @@ | |||
1 | /* | ||
2 | * drivers/s390/net/ctcm_fsms.h | ||
3 | * | ||
4 | * Copyright IBM Corp. 2001, 2007 | ||
5 | * Authors: Fritz Elfert (felfert@millenux.com) | ||
6 | * Peter Tiedemann (ptiedem@de.ibm.com) | ||
7 | * MPC additions : | ||
8 | * Belinda Thompson (belindat@us.ibm.com) | ||
9 | * Andy Richter (richtera@us.ibm.com) | ||
10 | */ | ||
11 | #ifndef _CTCM_FSMS_H_ | ||
12 | #define _CTCM_FSMS_H_ | ||
13 | |||
14 | #include <linux/module.h> | ||
15 | #include <linux/init.h> | ||
16 | #include <linux/kernel.h> | ||
17 | #include <linux/slab.h> | ||
18 | #include <linux/errno.h> | ||
19 | #include <linux/types.h> | ||
20 | #include <linux/interrupt.h> | ||
21 | #include <linux/timer.h> | ||
22 | #include <linux/bitops.h> | ||
23 | |||
24 | #include <linux/signal.h> | ||
25 | #include <linux/string.h> | ||
26 | |||
27 | #include <linux/ip.h> | ||
28 | #include <linux/if_arp.h> | ||
29 | #include <linux/tcp.h> | ||
30 | #include <linux/skbuff.h> | ||
31 | #include <linux/ctype.h> | ||
32 | #include <net/dst.h> | ||
33 | |||
34 | #include <linux/io.h> | ||
35 | #include <asm/ccwdev.h> | ||
36 | #include <asm/ccwgroup.h> | ||
37 | #include <linux/uaccess.h> | ||
38 | |||
39 | #include <asm/idals.h> | ||
40 | |||
41 | #include "fsm.h" | ||
42 | #include "cu3088.h" | ||
43 | #include "ctcm_main.h" | ||
44 | |||
45 | /* | ||
46 | * Definitions for the channel statemachine(s) for ctc and ctcmpc | ||
47 | * | ||
48 | * To allow better kerntyping, prefix-less definitions for channel states | ||
49 | * and channel events have been replaced : | ||
50 | * ch_event... -> ctc_ch_event... | ||
51 | * CH_EVENT... -> CTC_EVENT... | ||
52 | * ch_state... -> ctc_ch_state... | ||
53 | * CH_STATE... -> CTC_STATE... | ||
54 | */ | ||
55 | /* | ||
56 | * Events of the channel statemachine(s) for ctc and ctcmpc | ||
57 | */ | ||
58 | enum ctc_ch_events { | ||
59 | /* | ||
60 | * Events, representing return code of | ||
61 | * I/O operations (ccw_device_start, ccw_device_halt et al.) | ||
62 | */ | ||
63 | CTC_EVENT_IO_SUCCESS, | ||
64 | CTC_EVENT_IO_EBUSY, | ||
65 | CTC_EVENT_IO_ENODEV, | ||
66 | CTC_EVENT_IO_UNKNOWN, | ||
67 | |||
68 | CTC_EVENT_ATTNBUSY, | ||
69 | CTC_EVENT_ATTN, | ||
70 | CTC_EVENT_BUSY, | ||
71 | /* | ||
72 | * Events, representing unit-check | ||
73 | */ | ||
74 | CTC_EVENT_UC_RCRESET, | ||
75 | CTC_EVENT_UC_RSRESET, | ||
76 | CTC_EVENT_UC_TXTIMEOUT, | ||
77 | CTC_EVENT_UC_TXPARITY, | ||
78 | CTC_EVENT_UC_HWFAIL, | ||
79 | CTC_EVENT_UC_RXPARITY, | ||
80 | CTC_EVENT_UC_ZERO, | ||
81 | CTC_EVENT_UC_UNKNOWN, | ||
82 | /* | ||
83 | * Events, representing subchannel-check | ||
84 | */ | ||
85 | CTC_EVENT_SC_UNKNOWN, | ||
86 | /* | ||
87 | * Events, representing machine checks | ||
88 | */ | ||
89 | CTC_EVENT_MC_FAIL, | ||
90 | CTC_EVENT_MC_GOOD, | ||
91 | /* | ||
92 | * Event, representing normal IRQ | ||
93 | */ | ||
94 | CTC_EVENT_IRQ, | ||
95 | CTC_EVENT_FINSTAT, | ||
96 | /* | ||
97 | * Event, representing timer expiry. | ||
98 | */ | ||
99 | CTC_EVENT_TIMER, | ||
100 | /* | ||
101 | * Events, representing commands from upper levels. | ||
102 | */ | ||
103 | CTC_EVENT_START, | ||
104 | CTC_EVENT_STOP, | ||
105 | CTC_NR_EVENTS, | ||
106 | /* | ||
107 | * additional MPC events | ||
108 | */ | ||
109 | CTC_EVENT_SEND_XID = CTC_NR_EVENTS, | ||
110 | CTC_EVENT_RSWEEP_TIMER, | ||
111 | /* | ||
112 | * MUST be always the last element!! | ||
113 | */ | ||
114 | CTC_MPC_NR_EVENTS, | ||
115 | }; | ||
116 | |||
117 | /* | ||
118 | * States of the channel statemachine(s) for ctc and ctcmpc. | ||
119 | */ | ||
120 | enum ctc_ch_states { | ||
121 | /* | ||
122 | * Channel not assigned to any device, | ||
123 | * initial state, direction invalid | ||
124 | */ | ||
125 | CTC_STATE_IDLE, | ||
126 | /* | ||
127 | * Channel assigned but not operating | ||
128 | */ | ||
129 | CTC_STATE_STOPPED, | ||
130 | CTC_STATE_STARTWAIT, | ||
131 | CTC_STATE_STARTRETRY, | ||
132 | CTC_STATE_SETUPWAIT, | ||
133 | CTC_STATE_RXINIT, | ||
134 | CTC_STATE_TXINIT, | ||
135 | CTC_STATE_RX, | ||
136 | CTC_STATE_TX, | ||
137 | CTC_STATE_RXIDLE, | ||
138 | CTC_STATE_TXIDLE, | ||
139 | CTC_STATE_RXERR, | ||
140 | CTC_STATE_TXERR, | ||
141 | CTC_STATE_TERM, | ||
142 | CTC_STATE_DTERM, | ||
143 | CTC_STATE_NOTOP, | ||
144 | CTC_NR_STATES, /* MUST be the last element of non-expanded states */ | ||
145 | /* | ||
146 | * additional MPC states | ||
147 | */ | ||
148 | CH_XID0_PENDING = CTC_NR_STATES, | ||
149 | CH_XID0_INPROGRESS, | ||
150 | CH_XID7_PENDING, | ||
151 | CH_XID7_PENDING1, | ||
152 | CH_XID7_PENDING2, | ||
153 | CH_XID7_PENDING3, | ||
154 | CH_XID7_PENDING4, | ||
155 | CTC_MPC_NR_STATES, /* MUST be the last element of expanded mpc states */ | ||
156 | }; | ||
157 | |||
158 | extern const char *ctc_ch_event_names[]; | ||
159 | |||
160 | extern const char *ctc_ch_state_names[]; | ||
161 | |||
162 | void ctcm_ccw_check_rc(struct channel *ch, int rc, char *msg); | ||
163 | void ctcm_purge_skb_queue(struct sk_buff_head *q); | ||
164 | void fsm_action_nop(fsm_instance *fi, int event, void *arg); | ||
165 | |||
166 | /* | ||
167 | * ----- non-static actions for ctcm channel statemachine ----- | ||
168 | * | ||
169 | */ | ||
170 | void ctcm_chx_txidle(fsm_instance *fi, int event, void *arg); | ||
171 | |||
172 | /* | ||
173 | * ----- FSM (state/event/action) of the ctcm channel statemachine ----- | ||
174 | */ | ||
175 | extern const fsm_node ch_fsm[]; | ||
176 | extern int ch_fsm_len; | ||
177 | |||
178 | |||
179 | /* | ||
180 | * ----- non-static actions for ctcmpc channel statemachine ---- | ||
181 | * | ||
182 | */ | ||
183 | /* shared : | ||
184 | void ctcm_chx_txidle(fsm_instance * fi, int event, void *arg); | ||
185 | */ | ||
186 | void ctcmpc_chx_rxidle(fsm_instance *fi, int event, void *arg); | ||
187 | |||
188 | /* | ||
189 | * ----- FSM (state/event/action) of the ctcmpc channel statemachine ----- | ||
190 | */ | ||
191 | extern const fsm_node ctcmpc_ch_fsm[]; | ||
192 | extern int mpc_ch_fsm_len; | ||
193 | |||
194 | /* | ||
195 | * Definitions for the device interface statemachine for ctc and mpc | ||
196 | */ | ||
197 | |||
198 | /* | ||
199 | * States of the device interface statemachine. | ||
200 | */ | ||
201 | enum dev_states { | ||
202 | DEV_STATE_STOPPED, | ||
203 | DEV_STATE_STARTWAIT_RXTX, | ||
204 | DEV_STATE_STARTWAIT_RX, | ||
205 | DEV_STATE_STARTWAIT_TX, | ||
206 | DEV_STATE_STOPWAIT_RXTX, | ||
207 | DEV_STATE_STOPWAIT_RX, | ||
208 | DEV_STATE_STOPWAIT_TX, | ||
209 | DEV_STATE_RUNNING, | ||
210 | /* | ||
211 | * MUST be always the last element!! | ||
212 | */ | ||
213 | CTCM_NR_DEV_STATES | ||
214 | }; | ||
215 | |||
216 | extern const char *dev_state_names[]; | ||
217 | |||
218 | /* | ||
219 | * Events of the device interface statemachine. | ||
220 | * ctcm and ctcmpc | ||
221 | */ | ||
222 | enum dev_events { | ||
223 | DEV_EVENT_START, | ||
224 | DEV_EVENT_STOP, | ||
225 | DEV_EVENT_RXUP, | ||
226 | DEV_EVENT_TXUP, | ||
227 | DEV_EVENT_RXDOWN, | ||
228 | DEV_EVENT_TXDOWN, | ||
229 | DEV_EVENT_RESTART, | ||
230 | /* | ||
231 | * MUST be always the last element!! | ||
232 | */ | ||
233 | CTCM_NR_DEV_EVENTS | ||
234 | }; | ||
235 | |||
236 | extern const char *dev_event_names[]; | ||
237 | |||
238 | /* | ||
239 | * Actions for the device interface statemachine. | ||
240 | * ctc and ctcmpc | ||
241 | */ | ||
242 | /* | ||
243 | static void dev_action_start(fsm_instance * fi, int event, void *arg); | ||
244 | static void dev_action_stop(fsm_instance * fi, int event, void *arg); | ||
245 | static void dev_action_restart(fsm_instance *fi, int event, void *arg); | ||
246 | static void dev_action_chup(fsm_instance * fi, int event, void *arg); | ||
247 | static void dev_action_chdown(fsm_instance * fi, int event, void *arg); | ||
248 | */ | ||
249 | |||
250 | /* | ||
251 | * The (state/event/action) fsm table of the device interface statemachine. | ||
252 | * ctcm and ctcmpc | ||
253 | */ | ||
254 | extern const fsm_node dev_fsm[]; | ||
255 | extern int dev_fsm_len; | ||
256 | |||
257 | |||
258 | /* | ||
259 | * Definitions for the MPC Group statemachine | ||
260 | */ | ||
261 | |||
262 | /* | ||
263 | * MPC Group Station FSM States | ||
264 | |||
265 | State Name When In This State | ||
266 | ====================== ======================================= | ||
267 | MPCG_STATE_RESET Initial State When Driver Loaded | ||
268 | We receive and send NOTHING | ||
269 | |||
270 | MPCG_STATE_INOP INOP Received. | ||
271 | Group level non-recoverable error | ||
272 | |||
273 | MPCG_STATE_READY XID exchanges for at least 1 write and | ||
274 | 1 read channel have completed. | ||
275 | Group is ready for data transfer. | ||
276 | |||
277 | States from ctc_mpc_alloc_channel | ||
278 | ============================================================== | ||
279 | MPCG_STATE_XID2INITW Awaiting XID2(0) Initiation | ||
280 | ATTN from other side will start | ||
281 | XID negotiations. | ||
282 | Y-side protocol only. | ||
283 | |||
284 | MPCG_STATE_XID2INITX XID2(0) negotiations are in progress. | ||
285 | At least 1, but not all, XID2(0)'s | ||
286 | have been received from partner. | ||
287 | |||
288 | MPCG_STATE_XID7INITW XID2(0) complete | ||
289 | No XID2(7)'s have yet been received. | ||
290 | XID2(7) negotiations pending. | ||
291 | |||
292 | MPCG_STATE_XID7INITX XID2(7) negotiations in progress. | ||
293 | At least 1, but not all, XID2(7)'s | ||
294 | have been received from partner. | ||
295 | |||
296 | MPCG_STATE_XID7INITF XID2(7) negotiations complete. | ||
297 | Transitioning to READY. | ||
298 | |||
299 | MPCG_STATE_READY Ready for Data Transfer. | ||
300 | |||
301 | |||
302 | States from ctc_mpc_establish_connectivity call | ||
303 | ============================================================== | ||
304 | MPCG_STATE_XID0IOWAIT Initiating XID2(0) negotiations. | ||
305 | X-side protocol only. | ||
306 | ATTN-BUSY from other side will convert | ||
307 | this to Y-side protocol and the | ||
308 | ctc_mpc_alloc_channel flow will begin. | ||
309 | |||
310 | MPCG_STATE_XID0IOWAIX XID2(0) negotiations are in progress. | ||
311 | At least 1, but not all, XID2(0)'s | ||
312 | have been received from partner. | ||
313 | |||
314 | MPCG_STATE_XID7INITI XID2(0) complete | ||
315 | No XID2(7)'s have yet been received. | ||
316 | XID2(7) negotiations pending. | ||
317 | |||
318 | MPCG_STATE_XID7INITZ XID2(7) negotiations in progress. | ||
319 | At least 1, but not all, XID2(7)'s | ||
320 | have been received from partner. | ||
321 | |||
322 | MPCG_STATE_XID7INITF XID2(7) negotiations complete. | ||
323 | Transitioning to READY. | ||
324 | |||
325 | MPCG_STATE_READY Ready for Data Transfer. | ||
326 | |||
327 | */ | ||
328 | |||
329 | enum mpcg_events { | ||
330 | MPCG_EVENT_INOP, | ||
331 | MPCG_EVENT_DISCONC, | ||
332 | MPCG_EVENT_XID0DO, | ||
333 | MPCG_EVENT_XID2, | ||
334 | MPCG_EVENT_XID2DONE, | ||
335 | MPCG_EVENT_XID7DONE, | ||
336 | MPCG_EVENT_TIMER, | ||
337 | MPCG_EVENT_DOIO, | ||
338 | MPCG_NR_EVENTS, | ||
339 | }; | ||
340 | |||
341 | enum mpcg_states { | ||
342 | MPCG_STATE_RESET, | ||
343 | MPCG_STATE_INOP, | ||
344 | MPCG_STATE_XID2INITW, | ||
345 | MPCG_STATE_XID2INITX, | ||
346 | MPCG_STATE_XID7INITW, | ||
347 | MPCG_STATE_XID7INITX, | ||
348 | MPCG_STATE_XID0IOWAIT, | ||
349 | MPCG_STATE_XID0IOWAIX, | ||
350 | MPCG_STATE_XID7INITI, | ||
351 | MPCG_STATE_XID7INITZ, | ||
352 | MPCG_STATE_XID7INITF, | ||
353 | MPCG_STATE_FLOWC, | ||
354 | MPCG_STATE_READY, | ||
355 | MPCG_NR_STATES, | ||
356 | }; | ||
357 | |||
358 | #endif | ||
359 | /* --- This is the END my friend --- */ | ||
diff --git a/drivers/s390/net/ctcm_main.c b/drivers/s390/net/ctcm_main.c new file mode 100644 index 000000000000..d52843da4f55 --- /dev/null +++ b/drivers/s390/net/ctcm_main.c | |||
@@ -0,0 +1,1772 @@ | |||
1 | /* | ||
2 | * drivers/s390/net/ctcm_main.c | ||
3 | * | ||
4 | * Copyright IBM Corp. 2001, 2007 | ||
5 | * Author(s): | ||
6 | * Original CTC driver(s): | ||
7 | * Fritz Elfert (felfert@millenux.com) | ||
8 | * Dieter Wellerdiek (wel@de.ibm.com) | ||
9 | * Martin Schwidefsky (schwidefsky@de.ibm.com) | ||
10 | * Denis Joseph Barrow (barrow_dj@yahoo.com) | ||
11 | * Jochen Roehrig (roehrig@de.ibm.com) | ||
12 | * Cornelia Huck <cornelia.huck@de.ibm.com> | ||
13 | * MPC additions: | ||
14 | * Belinda Thompson (belindat@us.ibm.com) | ||
15 | * Andy Richter (richtera@us.ibm.com) | ||
16 | * Revived by: | ||
17 | * Peter Tiedemann (ptiedem@de.ibm.com) | ||
18 | */ | ||
19 | |||
20 | #undef DEBUG | ||
21 | #undef DEBUGDATA | ||
22 | #undef DEBUGCCW | ||
23 | |||
24 | #include <linux/module.h> | ||
25 | #include <linux/init.h> | ||
26 | #include <linux/kernel.h> | ||
27 | #include <linux/slab.h> | ||
28 | #include <linux/errno.h> | ||
29 | #include <linux/types.h> | ||
30 | #include <linux/interrupt.h> | ||
31 | #include <linux/timer.h> | ||
32 | #include <linux/bitops.h> | ||
33 | |||
34 | #include <linux/signal.h> | ||
35 | #include <linux/string.h> | ||
36 | |||
37 | #include <linux/ip.h> | ||
38 | #include <linux/if_arp.h> | ||
39 | #include <linux/tcp.h> | ||
40 | #include <linux/skbuff.h> | ||
41 | #include <linux/ctype.h> | ||
42 | #include <net/dst.h> | ||
43 | |||
44 | #include <linux/io.h> | ||
45 | #include <asm/ccwdev.h> | ||
46 | #include <asm/ccwgroup.h> | ||
47 | #include <linux/uaccess.h> | ||
48 | |||
49 | #include <asm/idals.h> | ||
50 | |||
51 | #include "cu3088.h" | ||
52 | #include "ctcm_fsms.h" | ||
53 | #include "ctcm_main.h" | ||
54 | |||
55 | /* Some common global variables */ | ||
56 | |||
57 | /* | ||
58 | * Linked list of all detected channels. | ||
59 | */ | ||
60 | struct channel *channels; | ||
61 | |||
62 | /** | ||
63 | * Unpack a just received skb and hand it over to | ||
64 | * upper layers. | ||
65 | * | ||
66 | * ch The channel where this skb has been received. | ||
67 | * pskb The received skb. | ||
68 | */ | ||
69 | void ctcm_unpack_skb(struct channel *ch, struct sk_buff *pskb) | ||
70 | { | ||
71 | struct net_device *dev = ch->netdev; | ||
72 | struct ctcm_priv *priv = dev->priv; | ||
73 | __u16 len = *((__u16 *) pskb->data); | ||
74 | |||
75 | skb_put(pskb, 2 + LL_HEADER_LENGTH); | ||
76 | skb_pull(pskb, 2); | ||
77 | pskb->dev = dev; | ||
78 | pskb->ip_summed = CHECKSUM_UNNECESSARY; | ||
79 | while (len > 0) { | ||
80 | struct sk_buff *skb; | ||
81 | int skblen; | ||
82 | struct ll_header *header = (struct ll_header *)pskb->data; | ||
83 | |||
84 | skb_pull(pskb, LL_HEADER_LENGTH); | ||
85 | if ((ch->protocol == CTCM_PROTO_S390) && | ||
86 | (header->type != ETH_P_IP)) { | ||
87 | |||
88 | if (!(ch->logflags & LOG_FLAG_ILLEGALPKT)) { | ||
89 | /* | ||
90 | * Check packet type only if we stick strictly | ||
91 | * to S/390's protocol of OS390. This only | ||
92 | * supports IP. Otherwise allow any packet | ||
93 | * type. | ||
94 | */ | ||
95 | ctcm_pr_warn("%s Illegal packet type 0x%04x " | ||
96 | "received, dropping\n", | ||
97 | dev->name, header->type); | ||
98 | ch->logflags |= LOG_FLAG_ILLEGALPKT; | ||
99 | } | ||
100 | |||
101 | priv->stats.rx_dropped++; | ||
102 | priv->stats.rx_frame_errors++; | ||
103 | return; | ||
104 | } | ||
105 | pskb->protocol = ntohs(header->type); | ||
106 | if (header->length <= LL_HEADER_LENGTH) { | ||
107 | if (!(ch->logflags & LOG_FLAG_ILLEGALSIZE)) { | ||
108 | ctcm_pr_warn( | ||
109 | "%s Illegal packet size %d " | ||
110 | "received (MTU=%d blocklen=%d), " | ||
111 | "dropping\n", dev->name, header->length, | ||
112 | dev->mtu, len); | ||
113 | ch->logflags |= LOG_FLAG_ILLEGALSIZE; | ||
114 | } | ||
115 | |||
116 | priv->stats.rx_dropped++; | ||
117 | priv->stats.rx_length_errors++; | ||
118 | return; | ||
119 | } | ||
120 | header->length -= LL_HEADER_LENGTH; | ||
121 | len -= LL_HEADER_LENGTH; | ||
122 | if ((header->length > skb_tailroom(pskb)) || | ||
123 | (header->length > len)) { | ||
124 | if (!(ch->logflags & LOG_FLAG_OVERRUN)) { | ||
125 | ctcm_pr_warn( | ||
126 | "%s Illegal packet size %d (beyond the" | ||
127 | " end of received data), dropping\n", | ||
128 | dev->name, header->length); | ||
129 | ch->logflags |= LOG_FLAG_OVERRUN; | ||
130 | } | ||
131 | |||
132 | priv->stats.rx_dropped++; | ||
133 | priv->stats.rx_length_errors++; | ||
134 | return; | ||
135 | } | ||
136 | skb_put(pskb, header->length); | ||
137 | skb_reset_mac_header(pskb); | ||
138 | len -= header->length; | ||
139 | skb = dev_alloc_skb(pskb->len); | ||
140 | if (!skb) { | ||
141 | if (!(ch->logflags & LOG_FLAG_NOMEM)) { | ||
142 | ctcm_pr_warn( | ||
143 | "%s Out of memory in ctcm_unpack_skb\n", | ||
144 | dev->name); | ||
145 | ch->logflags |= LOG_FLAG_NOMEM; | ||
146 | } | ||
147 | priv->stats.rx_dropped++; | ||
148 | return; | ||
149 | } | ||
150 | skb_copy_from_linear_data(pskb, skb_put(skb, pskb->len), | ||
151 | pskb->len); | ||
152 | skb_reset_mac_header(skb); | ||
153 | skb->dev = pskb->dev; | ||
154 | skb->protocol = pskb->protocol; | ||
155 | pskb->ip_summed = CHECKSUM_UNNECESSARY; | ||
156 | skblen = skb->len; | ||
157 | /* | ||
158 | * reset logflags | ||
159 | */ | ||
160 | ch->logflags = 0; | ||
161 | priv->stats.rx_packets++; | ||
162 | priv->stats.rx_bytes += skblen; | ||
163 | netif_rx_ni(skb); | ||
164 | dev->last_rx = jiffies; | ||
165 | if (len > 0) { | ||
166 | skb_pull(pskb, header->length); | ||
167 | if (skb_tailroom(pskb) < LL_HEADER_LENGTH) { | ||
168 | if (!(ch->logflags & LOG_FLAG_OVERRUN)) { | ||
169 | CTCM_DBF_DEV_NAME(TRACE, dev, | ||
170 | "Overrun in ctcm_unpack_skb"); | ||
171 | ch->logflags |= LOG_FLAG_OVERRUN; | ||
172 | } | ||
173 | return; | ||
174 | } | ||
175 | skb_put(pskb, LL_HEADER_LENGTH); | ||
176 | } | ||
177 | } | ||
178 | } | ||
179 | |||
180 | /** | ||
181 | * Release a specific channel in the channel list. | ||
182 | * | ||
183 | * ch Pointer to channel struct to be released. | ||
184 | */ | ||
185 | static void channel_free(struct channel *ch) | ||
186 | { | ||
187 | CTCM_DBF_TEXT(TRACE, 2, __FUNCTION__); | ||
188 | ch->flags &= ~CHANNEL_FLAGS_INUSE; | ||
189 | fsm_newstate(ch->fsm, CTC_STATE_IDLE); | ||
190 | } | ||
191 | |||
192 | /** | ||
193 | * Remove a specific channel in the channel list. | ||
194 | * | ||
195 | * ch Pointer to channel struct to be released. | ||
196 | */ | ||
197 | static void channel_remove(struct channel *ch) | ||
198 | { | ||
199 | struct channel **c = &channels; | ||
200 | char chid[CTCM_ID_SIZE+1]; | ||
201 | int ok = 0; | ||
202 | |||
203 | if (ch == NULL) | ||
204 | return; | ||
205 | else | ||
206 | strncpy(chid, ch->id, CTCM_ID_SIZE); | ||
207 | |||
208 | channel_free(ch); | ||
209 | while (*c) { | ||
210 | if (*c == ch) { | ||
211 | *c = ch->next; | ||
212 | fsm_deltimer(&ch->timer); | ||
213 | if (IS_MPC(ch)) | ||
214 | fsm_deltimer(&ch->sweep_timer); | ||
215 | |||
216 | kfree_fsm(ch->fsm); | ||
217 | clear_normalized_cda(&ch->ccw[4]); | ||
218 | if (ch->trans_skb != NULL) { | ||
219 | clear_normalized_cda(&ch->ccw[1]); | ||
220 | dev_kfree_skb_any(ch->trans_skb); | ||
221 | } | ||
222 | if (IS_MPC(ch)) { | ||
223 | tasklet_kill(&ch->ch_tasklet); | ||
224 | tasklet_kill(&ch->ch_disc_tasklet); | ||
225 | kfree(ch->discontact_th); | ||
226 | } | ||
227 | kfree(ch->ccw); | ||
228 | kfree(ch->irb); | ||
229 | kfree(ch); | ||
230 | ok = 1; | ||
231 | break; | ||
232 | } | ||
233 | c = &((*c)->next); | ||
234 | } | ||
235 | |||
236 | CTCM_DBF_TEXT_(SETUP, CTC_DBF_INFO, "%s(%s) %s", CTCM_FUNTAIL, | ||
237 | chid, ok ? "OK" : "failed"); | ||
238 | } | ||
239 | |||
240 | /** | ||
241 | * Get a specific channel from the channel list. | ||
242 | * | ||
243 | * type Type of channel we are interested in. | ||
244 | * id Id of channel we are interested in. | ||
245 | * direction Direction we want to use this channel for. | ||
246 | * | ||
247 | * returns Pointer to a channel or NULL if no matching channel available. | ||
248 | */ | ||
249 | static struct channel *channel_get(enum channel_types type, | ||
250 | char *id, int direction) | ||
251 | { | ||
252 | struct channel *ch = channels; | ||
253 | |||
254 | if (do_debug) { | ||
255 | char buf[64]; | ||
256 | sprintf(buf, "%s(%d, %s, %d)\n", | ||
257 | CTCM_FUNTAIL, type, id, direction); | ||
258 | CTCM_DBF_TEXT(TRACE, CTC_DBF_INFO, buf); | ||
259 | } | ||
260 | while (ch && (strncmp(ch->id, id, CTCM_ID_SIZE) || (ch->type != type))) | ||
261 | ch = ch->next; | ||
262 | if (!ch) { | ||
263 | char buf[64]; | ||
264 | sprintf(buf, "%s(%d, %s, %d) not found in channel list\n", | ||
265 | CTCM_FUNTAIL, type, id, direction); | ||
266 | CTCM_DBF_TEXT(ERROR, CTC_DBF_ERROR, buf); | ||
267 | } else { | ||
268 | if (ch->flags & CHANNEL_FLAGS_INUSE) | ||
269 | ch = NULL; | ||
270 | else { | ||
271 | ch->flags |= CHANNEL_FLAGS_INUSE; | ||
272 | ch->flags &= ~CHANNEL_FLAGS_RWMASK; | ||
273 | ch->flags |= (direction == WRITE) | ||
274 | ? CHANNEL_FLAGS_WRITE : CHANNEL_FLAGS_READ; | ||
275 | fsm_newstate(ch->fsm, CTC_STATE_STOPPED); | ||
276 | } | ||
277 | } | ||
278 | return ch; | ||
279 | } | ||
280 | |||
281 | static long ctcm_check_irb_error(struct ccw_device *cdev, struct irb *irb) | ||
282 | { | ||
283 | if (!IS_ERR(irb)) | ||
284 | return 0; | ||
285 | |||
286 | CTCM_DBF_TEXT_(ERROR, CTC_DBF_WARN, "irb error %ld on device %s\n", | ||
287 | PTR_ERR(irb), cdev->dev.bus_id); | ||
288 | |||
289 | switch (PTR_ERR(irb)) { | ||
290 | case -EIO: | ||
291 | ctcm_pr_warn("i/o-error on device %s\n", cdev->dev.bus_id); | ||
292 | break; | ||
293 | case -ETIMEDOUT: | ||
294 | ctcm_pr_warn("timeout on device %s\n", cdev->dev.bus_id); | ||
295 | break; | ||
296 | default: | ||
297 | ctcm_pr_warn("unknown error %ld on device %s\n", | ||
298 | PTR_ERR(irb), cdev->dev.bus_id); | ||
299 | } | ||
300 | return PTR_ERR(irb); | ||
301 | } | ||
302 | |||
303 | |||
304 | /** | ||
305 | * Check sense of a unit check. | ||
306 | * | ||
307 | * ch The channel, the sense code belongs to. | ||
308 | * sense The sense code to inspect. | ||
309 | */ | ||
310 | static inline void ccw_unit_check(struct channel *ch, unsigned char sense) | ||
311 | { | ||
312 | CTCM_DBF_TEXT(TRACE, 5, __FUNCTION__); | ||
313 | if (sense & SNS0_INTERVENTION_REQ) { | ||
314 | if (sense & 0x01) { | ||
315 | ctcm_pr_debug("%s: Interface disc. or Sel. reset " | ||
316 | "(remote)\n", ch->id); | ||
317 | fsm_event(ch->fsm, CTC_EVENT_UC_RCRESET, ch); | ||
318 | } else { | ||
319 | ctcm_pr_debug("%s: System reset (remote)\n", ch->id); | ||
320 | fsm_event(ch->fsm, CTC_EVENT_UC_RSRESET, ch); | ||
321 | } | ||
322 | } else if (sense & SNS0_EQUIPMENT_CHECK) { | ||
323 | if (sense & SNS0_BUS_OUT_CHECK) { | ||
324 | ctcm_pr_warn("%s: Hardware malfunction (remote)\n", | ||
325 | ch->id); | ||
326 | fsm_event(ch->fsm, CTC_EVENT_UC_HWFAIL, ch); | ||
327 | } else { | ||
328 | ctcm_pr_warn("%s: Read-data parity error (remote)\n", | ||
329 | ch->id); | ||
330 | fsm_event(ch->fsm, CTC_EVENT_UC_RXPARITY, ch); | ||
331 | } | ||
332 | } else if (sense & SNS0_BUS_OUT_CHECK) { | ||
333 | if (sense & 0x04) { | ||
334 | ctcm_pr_warn("%s: Data-streaming timeout)\n", ch->id); | ||
335 | fsm_event(ch->fsm, CTC_EVENT_UC_TXTIMEOUT, ch); | ||
336 | } else { | ||
337 | ctcm_pr_warn("%s: Data-transfer parity error\n", | ||
338 | ch->id); | ||
339 | fsm_event(ch->fsm, CTC_EVENT_UC_TXPARITY, ch); | ||
340 | } | ||
341 | } else if (sense & SNS0_CMD_REJECT) { | ||
342 | ctcm_pr_warn("%s: Command reject\n", ch->id); | ||
343 | } else if (sense == 0) { | ||
344 | ctcm_pr_debug("%s: Unit check ZERO\n", ch->id); | ||
345 | fsm_event(ch->fsm, CTC_EVENT_UC_ZERO, ch); | ||
346 | } else { | ||
347 | ctcm_pr_warn("%s: Unit Check with sense code: %02x\n", | ||
348 | ch->id, sense); | ||
349 | fsm_event(ch->fsm, CTC_EVENT_UC_UNKNOWN, ch); | ||
350 | } | ||
351 | } | ||
352 | |||
353 | int ctcm_ch_alloc_buffer(struct channel *ch) | ||
354 | { | ||
355 | CTCM_DBF_TEXT(TRACE, 5, __FUNCTION__); | ||
356 | |||
357 | clear_normalized_cda(&ch->ccw[1]); | ||
358 | ch->trans_skb = __dev_alloc_skb(ch->max_bufsize, GFP_ATOMIC | GFP_DMA); | ||
359 | if (ch->trans_skb == NULL) { | ||
360 | ctcm_pr_warn("%s: Couldn't alloc %s trans_skb\n", | ||
361 | ch->id, | ||
362 | (CHANNEL_DIRECTION(ch->flags) == READ) ? "RX" : "TX"); | ||
363 | return -ENOMEM; | ||
364 | } | ||
365 | |||
366 | ch->ccw[1].count = ch->max_bufsize; | ||
367 | if (set_normalized_cda(&ch->ccw[1], ch->trans_skb->data)) { | ||
368 | dev_kfree_skb(ch->trans_skb); | ||
369 | ch->trans_skb = NULL; | ||
370 | ctcm_pr_warn("%s: set_normalized_cda for %s " | ||
371 | "trans_skb failed, dropping packets\n", | ||
372 | ch->id, | ||
373 | (CHANNEL_DIRECTION(ch->flags) == READ) ? "RX" : "TX"); | ||
374 | return -ENOMEM; | ||
375 | } | ||
376 | |||
377 | ch->ccw[1].count = 0; | ||
378 | ch->trans_skb_data = ch->trans_skb->data; | ||
379 | ch->flags &= ~CHANNEL_FLAGS_BUFSIZE_CHANGED; | ||
380 | return 0; | ||
381 | } | ||
382 | |||
383 | /* | ||
384 | * Interface API for upper network layers | ||
385 | */ | ||
386 | |||
387 | /** | ||
388 | * Open an interface. | ||
389 | * Called from generic network layer when ifconfig up is run. | ||
390 | * | ||
391 | * dev Pointer to interface struct. | ||
392 | * | ||
393 | * returns 0 on success, -ERRNO on failure. (Never fails.) | ||
394 | */ | ||
395 | int ctcm_open(struct net_device *dev) | ||
396 | { | ||
397 | struct ctcm_priv *priv = dev->priv; | ||
398 | |||
399 | CTCMY_DBF_DEV_NAME(SETUP, dev, ""); | ||
400 | if (!IS_MPC(priv)) | ||
401 | fsm_event(priv->fsm, DEV_EVENT_START, dev); | ||
402 | return 0; | ||
403 | } | ||
404 | |||
405 | /** | ||
406 | * Close an interface. | ||
407 | * Called from generic network layer when ifconfig down is run. | ||
408 | * | ||
409 | * dev Pointer to interface struct. | ||
410 | * | ||
411 | * returns 0 on success, -ERRNO on failure. (Never fails.) | ||
412 | */ | ||
413 | int ctcm_close(struct net_device *dev) | ||
414 | { | ||
415 | struct ctcm_priv *priv = dev->priv; | ||
416 | |||
417 | CTCMY_DBF_DEV_NAME(SETUP, dev, ""); | ||
418 | if (!IS_MPC(priv)) | ||
419 | fsm_event(priv->fsm, DEV_EVENT_STOP, dev); | ||
420 | return 0; | ||
421 | } | ||
422 | |||
423 | |||
424 | /** | ||
425 | * Transmit a packet. | ||
426 | * This is a helper function for ctcm_tx(). | ||
427 | * | ||
428 | * ch Channel to be used for sending. | ||
429 | * skb Pointer to struct sk_buff of packet to send. | ||
430 | * The linklevel header has already been set up | ||
431 | * by ctcm_tx(). | ||
432 | * | ||
433 | * returns 0 on success, -ERRNO on failure. (Never fails.) | ||
434 | */ | ||
435 | static int ctcm_transmit_skb(struct channel *ch, struct sk_buff *skb) | ||
436 | { | ||
437 | unsigned long saveflags; | ||
438 | struct ll_header header; | ||
439 | int rc = 0; | ||
440 | __u16 block_len; | ||
441 | int ccw_idx; | ||
442 | struct sk_buff *nskb; | ||
443 | unsigned long hi; | ||
444 | |||
445 | /* we need to acquire the lock for testing the state | ||
446 | * otherwise we can have an IRQ changing the state to | ||
447 | * TXIDLE after the test but before acquiring the lock. | ||
448 | */ | ||
449 | spin_lock_irqsave(&ch->collect_lock, saveflags); | ||
450 | if (fsm_getstate(ch->fsm) != CTC_STATE_TXIDLE) { | ||
451 | int l = skb->len + LL_HEADER_LENGTH; | ||
452 | |||
453 | if (ch->collect_len + l > ch->max_bufsize - 2) { | ||
454 | spin_unlock_irqrestore(&ch->collect_lock, saveflags); | ||
455 | return -EBUSY; | ||
456 | } else { | ||
457 | atomic_inc(&skb->users); | ||
458 | header.length = l; | ||
459 | header.type = skb->protocol; | ||
460 | header.unused = 0; | ||
461 | memcpy(skb_push(skb, LL_HEADER_LENGTH), &header, | ||
462 | LL_HEADER_LENGTH); | ||
463 | skb_queue_tail(&ch->collect_queue, skb); | ||
464 | ch->collect_len += l; | ||
465 | } | ||
466 | spin_unlock_irqrestore(&ch->collect_lock, saveflags); | ||
467 | goto done; | ||
468 | } | ||
469 | spin_unlock_irqrestore(&ch->collect_lock, saveflags); | ||
470 | /* | ||
471 | * Protect skb against beeing free'd by upper | ||
472 | * layers. | ||
473 | */ | ||
474 | atomic_inc(&skb->users); | ||
475 | ch->prof.txlen += skb->len; | ||
476 | header.length = skb->len + LL_HEADER_LENGTH; | ||
477 | header.type = skb->protocol; | ||
478 | header.unused = 0; | ||
479 | memcpy(skb_push(skb, LL_HEADER_LENGTH), &header, LL_HEADER_LENGTH); | ||
480 | block_len = skb->len + 2; | ||
481 | *((__u16 *)skb_push(skb, 2)) = block_len; | ||
482 | |||
483 | /* | ||
484 | * IDAL support in CTCM is broken, so we have to | ||
485 | * care about skb's above 2G ourselves. | ||
486 | */ | ||
487 | hi = ((unsigned long)skb_tail_pointer(skb) + LL_HEADER_LENGTH) >> 31; | ||
488 | if (hi) { | ||
489 | nskb = alloc_skb(skb->len, GFP_ATOMIC | GFP_DMA); | ||
490 | if (!nskb) { | ||
491 | atomic_dec(&skb->users); | ||
492 | skb_pull(skb, LL_HEADER_LENGTH + 2); | ||
493 | ctcm_clear_busy(ch->netdev); | ||
494 | return -ENOMEM; | ||
495 | } else { | ||
496 | memcpy(skb_put(nskb, skb->len), skb->data, skb->len); | ||
497 | atomic_inc(&nskb->users); | ||
498 | atomic_dec(&skb->users); | ||
499 | dev_kfree_skb_irq(skb); | ||
500 | skb = nskb; | ||
501 | } | ||
502 | } | ||
503 | |||
504 | ch->ccw[4].count = block_len; | ||
505 | if (set_normalized_cda(&ch->ccw[4], skb->data)) { | ||
506 | /* | ||
507 | * idal allocation failed, try via copying to | ||
508 | * trans_skb. trans_skb usually has a pre-allocated | ||
509 | * idal. | ||
510 | */ | ||
511 | if (ctcm_checkalloc_buffer(ch)) { | ||
512 | /* | ||
513 | * Remove our header. It gets added | ||
514 | * again on retransmit. | ||
515 | */ | ||
516 | atomic_dec(&skb->users); | ||
517 | skb_pull(skb, LL_HEADER_LENGTH + 2); | ||
518 | ctcm_clear_busy(ch->netdev); | ||
519 | return -EBUSY; | ||
520 | } | ||
521 | |||
522 | skb_reset_tail_pointer(ch->trans_skb); | ||
523 | ch->trans_skb->len = 0; | ||
524 | ch->ccw[1].count = skb->len; | ||
525 | skb_copy_from_linear_data(skb, | ||
526 | skb_put(ch->trans_skb, skb->len), skb->len); | ||
527 | atomic_dec(&skb->users); | ||
528 | dev_kfree_skb_irq(skb); | ||
529 | ccw_idx = 0; | ||
530 | } else { | ||
531 | skb_queue_tail(&ch->io_queue, skb); | ||
532 | ccw_idx = 3; | ||
533 | } | ||
534 | ch->retry = 0; | ||
535 | fsm_newstate(ch->fsm, CTC_STATE_TX); | ||
536 | fsm_addtimer(&ch->timer, CTCM_TIME_5_SEC, CTC_EVENT_TIMER, ch); | ||
537 | spin_lock_irqsave(get_ccwdev_lock(ch->cdev), saveflags); | ||
538 | ch->prof.send_stamp = current_kernel_time(); /* xtime */ | ||
539 | rc = ccw_device_start(ch->cdev, &ch->ccw[ccw_idx], | ||
540 | (unsigned long)ch, 0xff, 0); | ||
541 | spin_unlock_irqrestore(get_ccwdev_lock(ch->cdev), saveflags); | ||
542 | if (ccw_idx == 3) | ||
543 | ch->prof.doios_single++; | ||
544 | if (rc != 0) { | ||
545 | fsm_deltimer(&ch->timer); | ||
546 | ctcm_ccw_check_rc(ch, rc, "single skb TX"); | ||
547 | if (ccw_idx == 3) | ||
548 | skb_dequeue_tail(&ch->io_queue); | ||
549 | /* | ||
550 | * Remove our header. It gets added | ||
551 | * again on retransmit. | ||
552 | */ | ||
553 | skb_pull(skb, LL_HEADER_LENGTH + 2); | ||
554 | } else if (ccw_idx == 0) { | ||
555 | struct net_device *dev = ch->netdev; | ||
556 | struct ctcm_priv *priv = dev->priv; | ||
557 | priv->stats.tx_packets++; | ||
558 | priv->stats.tx_bytes += skb->len - LL_HEADER_LENGTH; | ||
559 | } | ||
560 | done: | ||
561 | ctcm_clear_busy(ch->netdev); | ||
562 | return rc; | ||
563 | } | ||
564 | |||
565 | static void ctcmpc_send_sweep_req(struct channel *rch) | ||
566 | { | ||
567 | struct net_device *dev = rch->netdev; | ||
568 | struct ctcm_priv *priv; | ||
569 | struct mpc_group *grp; | ||
570 | struct th_sweep *header; | ||
571 | struct sk_buff *sweep_skb; | ||
572 | struct channel *ch; | ||
573 | int rc = 0; | ||
574 | |||
575 | priv = dev->priv; | ||
576 | grp = priv->mpcg; | ||
577 | ch = priv->channel[WRITE]; | ||
578 | |||
579 | if (do_debug) | ||
580 | MPC_DBF_DEV_NAME(TRACE, dev, ch->id); | ||
581 | |||
582 | /* sweep processing is not complete until response and request */ | ||
583 | /* has completed for all read channels in group */ | ||
584 | if (grp->in_sweep == 0) { | ||
585 | grp->in_sweep = 1; | ||
586 | grp->sweep_rsp_pend_num = grp->active_channels[READ]; | ||
587 | grp->sweep_req_pend_num = grp->active_channels[READ]; | ||
588 | } | ||
589 | |||
590 | sweep_skb = __dev_alloc_skb(MPC_BUFSIZE_DEFAULT, GFP_ATOMIC|GFP_DMA); | ||
591 | |||
592 | if (sweep_skb == NULL) { | ||
593 | printk(KERN_INFO "Couldn't alloc sweep_skb\n"); | ||
594 | rc = -ENOMEM; | ||
595 | goto done; | ||
596 | } | ||
597 | |||
598 | header = kmalloc(TH_SWEEP_LENGTH, gfp_type()); | ||
599 | |||
600 | if (!header) { | ||
601 | dev_kfree_skb_any(sweep_skb); | ||
602 | rc = -ENOMEM; | ||
603 | goto done; | ||
604 | } | ||
605 | |||
606 | header->th.th_seg = 0x00 ; | ||
607 | header->th.th_ch_flag = TH_SWEEP_REQ; /* 0x0f */ | ||
608 | header->th.th_blk_flag = 0x00; | ||
609 | header->th.th_is_xid = 0x00; | ||
610 | header->th.th_seq_num = 0x00; | ||
611 | header->sw.th_last_seq = ch->th_seq_num; | ||
612 | |||
613 | memcpy(skb_put(sweep_skb, TH_SWEEP_LENGTH), header, TH_SWEEP_LENGTH); | ||
614 | |||
615 | kfree(header); | ||
616 | |||
617 | dev->trans_start = jiffies; | ||
618 | skb_queue_tail(&ch->sweep_queue, sweep_skb); | ||
619 | |||
620 | fsm_addtimer(&ch->sweep_timer, 100, CTC_EVENT_RSWEEP_TIMER, ch); | ||
621 | |||
622 | return; | ||
623 | |||
624 | done: | ||
625 | if (rc != 0) { | ||
626 | grp->in_sweep = 0; | ||
627 | ctcm_clear_busy(dev); | ||
628 | fsm_event(grp->fsm, MPCG_EVENT_INOP, dev); | ||
629 | } | ||
630 | |||
631 | return; | ||
632 | } | ||
633 | |||
634 | /* | ||
635 | * MPC mode version of transmit_skb | ||
636 | */ | ||
637 | static int ctcmpc_transmit_skb(struct channel *ch, struct sk_buff *skb) | ||
638 | { | ||
639 | struct pdu *p_header; | ||
640 | struct net_device *dev = ch->netdev; | ||
641 | struct ctcm_priv *priv = dev->priv; | ||
642 | struct mpc_group *grp = priv->mpcg; | ||
643 | struct th_header *header; | ||
644 | struct sk_buff *nskb; | ||
645 | int rc = 0; | ||
646 | int ccw_idx; | ||
647 | unsigned long hi; | ||
648 | unsigned long saveflags = 0; /* avoids compiler warning */ | ||
649 | __u16 block_len; | ||
650 | |||
651 | if (do_debug) | ||
652 | ctcm_pr_debug( | ||
653 | "ctcm enter: %s(): %s cp=%i ch=0x%p id=%s state=%s\n", | ||
654 | __FUNCTION__, dev->name, smp_processor_id(), ch, | ||
655 | ch->id, fsm_getstate_str(ch->fsm)); | ||
656 | |||
657 | if ((fsm_getstate(ch->fsm) != CTC_STATE_TXIDLE) || grp->in_sweep) { | ||
658 | spin_lock_irqsave(&ch->collect_lock, saveflags); | ||
659 | atomic_inc(&skb->users); | ||
660 | p_header = kmalloc(PDU_HEADER_LENGTH, gfp_type()); | ||
661 | |||
662 | if (!p_header) { | ||
663 | printk(KERN_WARNING "ctcm: OUT OF MEMORY IN %s():" | ||
664 | " Data Lost \n", __FUNCTION__); | ||
665 | |||
666 | atomic_dec(&skb->users); | ||
667 | dev_kfree_skb_any(skb); | ||
668 | spin_unlock_irqrestore(&ch->collect_lock, saveflags); | ||
669 | fsm_event(priv->mpcg->fsm, MPCG_EVENT_INOP, dev); | ||
670 | goto done; | ||
671 | } | ||
672 | |||
673 | p_header->pdu_offset = skb->len; | ||
674 | p_header->pdu_proto = 0x01; | ||
675 | p_header->pdu_flag = 0x00; | ||
676 | if (skb->protocol == ntohs(ETH_P_SNAP)) { | ||
677 | p_header->pdu_flag |= PDU_FIRST | PDU_CNTL; | ||
678 | } else { | ||
679 | p_header->pdu_flag |= PDU_FIRST; | ||
680 | } | ||
681 | p_header->pdu_seq = 0; | ||
682 | memcpy(skb_push(skb, PDU_HEADER_LENGTH), p_header, | ||
683 | PDU_HEADER_LENGTH); | ||
684 | |||
685 | if (do_debug_data) { | ||
686 | ctcm_pr_debug("ctcm: %s() Putting on collect_q" | ||
687 | " - skb len: %04x \n", __FUNCTION__, skb->len); | ||
688 | ctcm_pr_debug("ctcm: %s() pdu header and data" | ||
689 | " for up to 32 bytes\n", __FUNCTION__); | ||
690 | ctcmpc_dump32((char *)skb->data, skb->len); | ||
691 | } | ||
692 | |||
693 | skb_queue_tail(&ch->collect_queue, skb); | ||
694 | ch->collect_len += skb->len; | ||
695 | kfree(p_header); | ||
696 | |||
697 | spin_unlock_irqrestore(&ch->collect_lock, saveflags); | ||
698 | goto done; | ||
699 | } | ||
700 | |||
701 | /* | ||
702 | * Protect skb against beeing free'd by upper | ||
703 | * layers. | ||
704 | */ | ||
705 | atomic_inc(&skb->users); | ||
706 | |||
707 | block_len = skb->len + TH_HEADER_LENGTH + PDU_HEADER_LENGTH; | ||
708 | /* | ||
709 | * IDAL support in CTCM is broken, so we have to | ||
710 | * care about skb's above 2G ourselves. | ||
711 | */ | ||
712 | hi = ((unsigned long)skb->tail + TH_HEADER_LENGTH) >> 31; | ||
713 | if (hi) { | ||
714 | nskb = __dev_alloc_skb(skb->len, GFP_ATOMIC | GFP_DMA); | ||
715 | if (!nskb) { | ||
716 | printk(KERN_WARNING "ctcm: %s() OUT OF MEMORY" | ||
717 | "- Data Lost \n", __FUNCTION__); | ||
718 | atomic_dec(&skb->users); | ||
719 | dev_kfree_skb_any(skb); | ||
720 | fsm_event(priv->mpcg->fsm, MPCG_EVENT_INOP, dev); | ||
721 | goto done; | ||
722 | } else { | ||
723 | memcpy(skb_put(nskb, skb->len), skb->data, skb->len); | ||
724 | atomic_inc(&nskb->users); | ||
725 | atomic_dec(&skb->users); | ||
726 | dev_kfree_skb_irq(skb); | ||
727 | skb = nskb; | ||
728 | } | ||
729 | } | ||
730 | |||
731 | p_header = kmalloc(PDU_HEADER_LENGTH, gfp_type()); | ||
732 | |||
733 | if (!p_header) { | ||
734 | printk(KERN_WARNING "ctcm: %s() OUT OF MEMORY" | ||
735 | ": Data Lost \n", __FUNCTION__); | ||
736 | |||
737 | atomic_dec(&skb->users); | ||
738 | dev_kfree_skb_any(skb); | ||
739 | fsm_event(priv->mpcg->fsm, MPCG_EVENT_INOP, dev); | ||
740 | goto done; | ||
741 | } | ||
742 | |||
743 | p_header->pdu_offset = skb->len; | ||
744 | p_header->pdu_proto = 0x01; | ||
745 | p_header->pdu_flag = 0x00; | ||
746 | p_header->pdu_seq = 0; | ||
747 | if (skb->protocol == ntohs(ETH_P_SNAP)) { | ||
748 | p_header->pdu_flag |= PDU_FIRST | PDU_CNTL; | ||
749 | } else { | ||
750 | p_header->pdu_flag |= PDU_FIRST; | ||
751 | } | ||
752 | memcpy(skb_push(skb, PDU_HEADER_LENGTH), p_header, PDU_HEADER_LENGTH); | ||
753 | |||
754 | kfree(p_header); | ||
755 | |||
756 | if (ch->collect_len > 0) { | ||
757 | spin_lock_irqsave(&ch->collect_lock, saveflags); | ||
758 | skb_queue_tail(&ch->collect_queue, skb); | ||
759 | ch->collect_len += skb->len; | ||
760 | skb = skb_dequeue(&ch->collect_queue); | ||
761 | ch->collect_len -= skb->len; | ||
762 | spin_unlock_irqrestore(&ch->collect_lock, saveflags); | ||
763 | } | ||
764 | |||
765 | p_header = (struct pdu *)skb->data; | ||
766 | p_header->pdu_flag |= PDU_LAST; | ||
767 | |||
768 | ch->prof.txlen += skb->len - PDU_HEADER_LENGTH; | ||
769 | |||
770 | header = kmalloc(TH_HEADER_LENGTH, gfp_type()); | ||
771 | |||
772 | if (!header) { | ||
773 | printk(KERN_WARNING "ctcm: %s() OUT OF MEMORY: Data Lost \n", | ||
774 | __FUNCTION__); | ||
775 | atomic_dec(&skb->users); | ||
776 | dev_kfree_skb_any(skb); | ||
777 | fsm_event(priv->mpcg->fsm, MPCG_EVENT_INOP, dev); | ||
778 | goto done; | ||
779 | } | ||
780 | |||
781 | header->th_seg = 0x00; | ||
782 | header->th_ch_flag = TH_HAS_PDU; /* Normal data */ | ||
783 | header->th_blk_flag = 0x00; | ||
784 | header->th_is_xid = 0x00; /* Just data here */ | ||
785 | ch->th_seq_num++; | ||
786 | header->th_seq_num = ch->th_seq_num; | ||
787 | |||
788 | if (do_debug_data) | ||
789 | ctcm_pr_debug("ctcm: %s() ToVTAM_th_seq= %08x\n" , | ||
790 | __FUNCTION__, ch->th_seq_num); | ||
791 | |||
792 | /* put the TH on the packet */ | ||
793 | memcpy(skb_push(skb, TH_HEADER_LENGTH), header, TH_HEADER_LENGTH); | ||
794 | |||
795 | kfree(header); | ||
796 | |||
797 | if (do_debug_data) { | ||
798 | ctcm_pr_debug("ctcm: %s(): skb len: %04x \n", | ||
799 | __FUNCTION__, skb->len); | ||
800 | ctcm_pr_debug("ctcm: %s(): pdu header and data for up to 32 " | ||
801 | "bytes sent to vtam\n", __FUNCTION__); | ||
802 | ctcmpc_dump32((char *)skb->data, skb->len); | ||
803 | } | ||
804 | |||
805 | ch->ccw[4].count = skb->len; | ||
806 | if (set_normalized_cda(&ch->ccw[4], skb->data)) { | ||
807 | /* | ||
808 | * idal allocation failed, try via copying to | ||
809 | * trans_skb. trans_skb usually has a pre-allocated | ||
810 | * idal. | ||
811 | */ | ||
812 | if (ctcm_checkalloc_buffer(ch)) { | ||
813 | /* | ||
814 | * Remove our header. It gets added | ||
815 | * again on retransmit. | ||
816 | */ | ||
817 | atomic_dec(&skb->users); | ||
818 | dev_kfree_skb_any(skb); | ||
819 | printk(KERN_WARNING "ctcm: %s()OUT OF MEMORY:" | ||
820 | " Data Lost \n", __FUNCTION__); | ||
821 | fsm_event(priv->mpcg->fsm, MPCG_EVENT_INOP, dev); | ||
822 | goto done; | ||
823 | } | ||
824 | |||
825 | skb_reset_tail_pointer(ch->trans_skb); | ||
826 | ch->trans_skb->len = 0; | ||
827 | ch->ccw[1].count = skb->len; | ||
828 | memcpy(skb_put(ch->trans_skb, skb->len), skb->data, skb->len); | ||
829 | atomic_dec(&skb->users); | ||
830 | dev_kfree_skb_irq(skb); | ||
831 | ccw_idx = 0; | ||
832 | if (do_debug_data) { | ||
833 | ctcm_pr_debug("ctcm: %s() TRANS skb len: %d \n", | ||
834 | __FUNCTION__, ch->trans_skb->len); | ||
835 | ctcm_pr_debug("ctcm: %s up to 32 bytes of data" | ||
836 | " sent to vtam\n", __FUNCTION__); | ||
837 | ctcmpc_dump32((char *)ch->trans_skb->data, | ||
838 | ch->trans_skb->len); | ||
839 | } | ||
840 | } else { | ||
841 | skb_queue_tail(&ch->io_queue, skb); | ||
842 | ccw_idx = 3; | ||
843 | } | ||
844 | ch->retry = 0; | ||
845 | fsm_newstate(ch->fsm, CTC_STATE_TX); | ||
846 | fsm_addtimer(&ch->timer, CTCM_TIME_5_SEC, CTC_EVENT_TIMER, ch); | ||
847 | |||
848 | if (do_debug_ccw) | ||
849 | ctcmpc_dumpit((char *)&ch->ccw[ccw_idx], | ||
850 | sizeof(struct ccw1) * 3); | ||
851 | |||
852 | spin_lock_irqsave(get_ccwdev_lock(ch->cdev), saveflags); | ||
853 | ch->prof.send_stamp = current_kernel_time(); /* xtime */ | ||
854 | rc = ccw_device_start(ch->cdev, &ch->ccw[ccw_idx], | ||
855 | (unsigned long)ch, 0xff, 0); | ||
856 | spin_unlock_irqrestore(get_ccwdev_lock(ch->cdev), saveflags); | ||
857 | if (ccw_idx == 3) | ||
858 | ch->prof.doios_single++; | ||
859 | if (rc != 0) { | ||
860 | fsm_deltimer(&ch->timer); | ||
861 | ctcm_ccw_check_rc(ch, rc, "single skb TX"); | ||
862 | if (ccw_idx == 3) | ||
863 | skb_dequeue_tail(&ch->io_queue); | ||
864 | } else if (ccw_idx == 0) { | ||
865 | priv->stats.tx_packets++; | ||
866 | priv->stats.tx_bytes += skb->len - TH_HEADER_LENGTH; | ||
867 | } | ||
868 | if (ch->th_seq_num > 0xf0000000) /* Chose 4Billion at random. */ | ||
869 | ctcmpc_send_sweep_req(ch); | ||
870 | |||
871 | done: | ||
872 | if (do_debug) | ||
873 | ctcm_pr_debug("ctcm exit: %s %s()\n", dev->name, __FUNCTION__); | ||
874 | return 0; | ||
875 | } | ||
876 | |||
877 | /** | ||
878 | * Start transmission of a packet. | ||
879 | * Called from generic network device layer. | ||
880 | * | ||
881 | * skb Pointer to buffer containing the packet. | ||
882 | * dev Pointer to interface struct. | ||
883 | * | ||
884 | * returns 0 if packet consumed, !0 if packet rejected. | ||
885 | * Note: If we return !0, then the packet is free'd by | ||
886 | * the generic network layer. | ||
887 | */ | ||
888 | /* first merge version - leaving both functions separated */ | ||
889 | static int ctcm_tx(struct sk_buff *skb, struct net_device *dev) | ||
890 | { | ||
891 | int rc = 0; | ||
892 | struct ctcm_priv *priv; | ||
893 | |||
894 | CTCM_DBF_TEXT(TRACE, 5, __FUNCTION__); | ||
895 | priv = dev->priv; | ||
896 | |||
897 | if (skb == NULL) { | ||
898 | ctcm_pr_warn("%s: NULL sk_buff passed\n", dev->name); | ||
899 | priv->stats.tx_dropped++; | ||
900 | return 0; | ||
901 | } | ||
902 | if (skb_headroom(skb) < (LL_HEADER_LENGTH + 2)) { | ||
903 | ctcm_pr_warn("%s: Got sk_buff with head room < %ld bytes\n", | ||
904 | dev->name, LL_HEADER_LENGTH + 2); | ||
905 | dev_kfree_skb(skb); | ||
906 | priv->stats.tx_dropped++; | ||
907 | return 0; | ||
908 | } | ||
909 | |||
910 | /* | ||
911 | * If channels are not running, try to restart them | ||
912 | * and throw away packet. | ||
913 | */ | ||
914 | if (fsm_getstate(priv->fsm) != DEV_STATE_RUNNING) { | ||
915 | fsm_event(priv->fsm, DEV_EVENT_START, dev); | ||
916 | dev_kfree_skb(skb); | ||
917 | priv->stats.tx_dropped++; | ||
918 | priv->stats.tx_errors++; | ||
919 | priv->stats.tx_carrier_errors++; | ||
920 | return 0; | ||
921 | } | ||
922 | |||
923 | if (ctcm_test_and_set_busy(dev)) | ||
924 | return -EBUSY; | ||
925 | |||
926 | dev->trans_start = jiffies; | ||
927 | if (ctcm_transmit_skb(priv->channel[WRITE], skb) != 0) | ||
928 | rc = 1; | ||
929 | return rc; | ||
930 | } | ||
931 | |||
932 | /* unmerged MPC variant of ctcm_tx */ | ||
933 | static int ctcmpc_tx(struct sk_buff *skb, struct net_device *dev) | ||
934 | { | ||
935 | int len = 0; | ||
936 | struct ctcm_priv *priv = NULL; | ||
937 | struct mpc_group *grp = NULL; | ||
938 | struct sk_buff *newskb = NULL; | ||
939 | |||
940 | if (do_debug) | ||
941 | ctcm_pr_debug("ctcmpc enter: %s(): skb:%0lx\n", | ||
942 | __FUNCTION__, (unsigned long)skb); | ||
943 | |||
944 | CTCM_DBF_TEXT_(MPC_TRACE, CTC_DBF_DEBUG, | ||
945 | "ctcmpc enter: %s(): skb:%0lx\n", | ||
946 | __FUNCTION__, (unsigned long)skb); | ||
947 | |||
948 | priv = dev->priv; | ||
949 | grp = priv->mpcg; | ||
950 | /* | ||
951 | * Some sanity checks ... | ||
952 | */ | ||
953 | if (skb == NULL) { | ||
954 | ctcm_pr_warn("ctcmpc: %s: NULL sk_buff passed\n", dev->name); | ||
955 | priv->stats.tx_dropped++; | ||
956 | goto done; | ||
957 | } | ||
958 | if (skb_headroom(skb) < (TH_HEADER_LENGTH + PDU_HEADER_LENGTH)) { | ||
959 | CTCM_DBF_TEXT_(MPC_TRACE, CTC_DBF_WARN, | ||
960 | "%s: Got sk_buff with head room < %ld bytes\n", | ||
961 | dev->name, TH_HEADER_LENGTH + PDU_HEADER_LENGTH); | ||
962 | |||
963 | if (do_debug_data) | ||
964 | ctcmpc_dump32((char *)skb->data, skb->len); | ||
965 | |||
966 | len = skb->len + TH_HEADER_LENGTH + PDU_HEADER_LENGTH; | ||
967 | newskb = __dev_alloc_skb(len, gfp_type() | GFP_DMA); | ||
968 | |||
969 | if (!newskb) { | ||
970 | printk(KERN_WARNING "ctcmpc: %s() OUT OF MEMORY-" | ||
971 | "Data Lost\n", | ||
972 | __FUNCTION__); | ||
973 | |||
974 | dev_kfree_skb_any(skb); | ||
975 | priv->stats.tx_dropped++; | ||
976 | priv->stats.tx_errors++; | ||
977 | priv->stats.tx_carrier_errors++; | ||
978 | fsm_event(grp->fsm, MPCG_EVENT_INOP, dev); | ||
979 | goto done; | ||
980 | } | ||
981 | newskb->protocol = skb->protocol; | ||
982 | skb_reserve(newskb, TH_HEADER_LENGTH + PDU_HEADER_LENGTH); | ||
983 | memcpy(skb_put(newskb, skb->len), skb->data, skb->len); | ||
984 | dev_kfree_skb_any(skb); | ||
985 | skb = newskb; | ||
986 | } | ||
987 | |||
988 | /* | ||
989 | * If channels are not running, | ||
990 | * notify anybody about a link failure and throw | ||
991 | * away packet. | ||
992 | */ | ||
993 | if ((fsm_getstate(priv->fsm) != DEV_STATE_RUNNING) || | ||
994 | (fsm_getstate(grp->fsm) < MPCG_STATE_XID2INITW)) { | ||
995 | dev_kfree_skb_any(skb); | ||
996 | printk(KERN_INFO "ctcmpc: %s() DATA RCVD - MPC GROUP " | ||
997 | "NOT ACTIVE - DROPPED\n", | ||
998 | __FUNCTION__); | ||
999 | priv->stats.tx_dropped++; | ||
1000 | priv->stats.tx_errors++; | ||
1001 | priv->stats.tx_carrier_errors++; | ||
1002 | goto done; | ||
1003 | } | ||
1004 | |||
1005 | if (ctcm_test_and_set_busy(dev)) { | ||
1006 | printk(KERN_WARNING "%s:DEVICE ERR - UNRECOVERABLE DATA LOSS\n", | ||
1007 | __FUNCTION__); | ||
1008 | dev_kfree_skb_any(skb); | ||
1009 | priv->stats.tx_dropped++; | ||
1010 | priv->stats.tx_errors++; | ||
1011 | priv->stats.tx_carrier_errors++; | ||
1012 | fsm_event(grp->fsm, MPCG_EVENT_INOP, dev); | ||
1013 | goto done; | ||
1014 | } | ||
1015 | |||
1016 | dev->trans_start = jiffies; | ||
1017 | if (ctcmpc_transmit_skb(priv->channel[WRITE], skb) != 0) { | ||
1018 | printk(KERN_WARNING "ctcmpc: %s() DEVICE ERROR" | ||
1019 | ": Data Lost \n", | ||
1020 | __FUNCTION__); | ||
1021 | printk(KERN_WARNING "ctcmpc: %s() DEVICE ERROR" | ||
1022 | " - UNRECOVERABLE DATA LOSS\n", | ||
1023 | __FUNCTION__); | ||
1024 | dev_kfree_skb_any(skb); | ||
1025 | priv->stats.tx_dropped++; | ||
1026 | priv->stats.tx_errors++; | ||
1027 | priv->stats.tx_carrier_errors++; | ||
1028 | ctcm_clear_busy(dev); | ||
1029 | fsm_event(grp->fsm, MPCG_EVENT_INOP, dev); | ||
1030 | goto done; | ||
1031 | } | ||
1032 | ctcm_clear_busy(dev); | ||
1033 | done: | ||
1034 | if (do_debug) | ||
1035 | MPC_DBF_DEV_NAME(TRACE, dev, "exit"); | ||
1036 | |||
1037 | return 0; /* handle freeing of skb here */ | ||
1038 | } | ||
1039 | |||
1040 | |||
1041 | /** | ||
1042 | * Sets MTU of an interface. | ||
1043 | * | ||
1044 | * dev Pointer to interface struct. | ||
1045 | * new_mtu The new MTU to use for this interface. | ||
1046 | * | ||
1047 | * returns 0 on success, -EINVAL if MTU is out of valid range. | ||
1048 | * (valid range is 576 .. 65527). If VM is on the | ||
1049 | * remote side, maximum MTU is 32760, however this is | ||
1050 | * not checked here. | ||
1051 | */ | ||
1052 | static int ctcm_change_mtu(struct net_device *dev, int new_mtu) | ||
1053 | { | ||
1054 | struct ctcm_priv *priv; | ||
1055 | int max_bufsize; | ||
1056 | |||
1057 | CTCM_DBF_TEXT(SETUP, CTC_DBF_INFO, __FUNCTION__); | ||
1058 | |||
1059 | if (new_mtu < 576 || new_mtu > 65527) | ||
1060 | return -EINVAL; | ||
1061 | |||
1062 | priv = dev->priv; | ||
1063 | max_bufsize = priv->channel[READ]->max_bufsize; | ||
1064 | |||
1065 | if (IS_MPC(priv)) { | ||
1066 | if (new_mtu > max_bufsize - TH_HEADER_LENGTH) | ||
1067 | return -EINVAL; | ||
1068 | dev->hard_header_len = TH_HEADER_LENGTH + PDU_HEADER_LENGTH; | ||
1069 | } else { | ||
1070 | if (new_mtu > max_bufsize - LL_HEADER_LENGTH - 2) | ||
1071 | return -EINVAL; | ||
1072 | dev->hard_header_len = LL_HEADER_LENGTH + 2; | ||
1073 | } | ||
1074 | dev->mtu = new_mtu; | ||
1075 | return 0; | ||
1076 | } | ||
1077 | |||
1078 | /** | ||
1079 | * Returns interface statistics of a device. | ||
1080 | * | ||
1081 | * dev Pointer to interface struct. | ||
1082 | * | ||
1083 | * returns Pointer to stats struct of this interface. | ||
1084 | */ | ||
1085 | static struct net_device_stats *ctcm_stats(struct net_device *dev) | ||
1086 | { | ||
1087 | return &((struct ctcm_priv *)dev->priv)->stats; | ||
1088 | } | ||
1089 | |||
1090 | |||
1091 | static void ctcm_netdev_unregister(struct net_device *dev) | ||
1092 | { | ||
1093 | CTCM_DBF_TEXT(SETUP, CTC_DBF_INFO, __FUNCTION__); | ||
1094 | if (!dev) | ||
1095 | return; | ||
1096 | unregister_netdev(dev); | ||
1097 | } | ||
1098 | |||
1099 | static int ctcm_netdev_register(struct net_device *dev) | ||
1100 | { | ||
1101 | CTCM_DBF_TEXT(SETUP, CTC_DBF_INFO, __FUNCTION__); | ||
1102 | return register_netdev(dev); | ||
1103 | } | ||
1104 | |||
1105 | static void ctcm_free_netdevice(struct net_device *dev) | ||
1106 | { | ||
1107 | struct ctcm_priv *priv; | ||
1108 | struct mpc_group *grp; | ||
1109 | |||
1110 | CTCM_DBF_TEXT(SETUP, CTC_DBF_INFO, __FUNCTION__); | ||
1111 | |||
1112 | if (!dev) | ||
1113 | return; | ||
1114 | priv = dev->priv; | ||
1115 | if (priv) { | ||
1116 | grp = priv->mpcg; | ||
1117 | if (grp) { | ||
1118 | if (grp->fsm) | ||
1119 | kfree_fsm(grp->fsm); | ||
1120 | if (grp->xid_skb) | ||
1121 | dev_kfree_skb(grp->xid_skb); | ||
1122 | if (grp->rcvd_xid_skb) | ||
1123 | dev_kfree_skb(grp->rcvd_xid_skb); | ||
1124 | tasklet_kill(&grp->mpc_tasklet2); | ||
1125 | kfree(grp); | ||
1126 | priv->mpcg = NULL; | ||
1127 | } | ||
1128 | if (priv->fsm) { | ||
1129 | kfree_fsm(priv->fsm); | ||
1130 | priv->fsm = NULL; | ||
1131 | } | ||
1132 | kfree(priv->xid); | ||
1133 | priv->xid = NULL; | ||
1134 | /* | ||
1135 | * Note: kfree(priv); is done in "opposite" function of | ||
1136 | * allocator function probe_device which is remove_device. | ||
1137 | */ | ||
1138 | } | ||
1139 | #ifdef MODULE | ||
1140 | free_netdev(dev); | ||
1141 | #endif | ||
1142 | } | ||
1143 | |||
1144 | struct mpc_group *ctcmpc_init_mpc_group(struct ctcm_priv *priv); | ||
1145 | |||
1146 | void static ctcm_dev_setup(struct net_device *dev) | ||
1147 | { | ||
1148 | dev->open = ctcm_open; | ||
1149 | dev->stop = ctcm_close; | ||
1150 | dev->get_stats = ctcm_stats; | ||
1151 | dev->change_mtu = ctcm_change_mtu; | ||
1152 | dev->type = ARPHRD_SLIP; | ||
1153 | dev->tx_queue_len = 100; | ||
1154 | dev->flags = IFF_POINTOPOINT | IFF_NOARP; | ||
1155 | } | ||
1156 | |||
1157 | /* | ||
1158 | * Initialize everything of the net device except the name and the | ||
1159 | * channel structs. | ||
1160 | */ | ||
1161 | static struct net_device *ctcm_init_netdevice(struct ctcm_priv *priv) | ||
1162 | { | ||
1163 | struct net_device *dev; | ||
1164 | struct mpc_group *grp; | ||
1165 | if (!priv) | ||
1166 | return NULL; | ||
1167 | |||
1168 | if (IS_MPC(priv)) | ||
1169 | dev = alloc_netdev(0, MPC_DEVICE_GENE, ctcm_dev_setup); | ||
1170 | else | ||
1171 | dev = alloc_netdev(0, CTC_DEVICE_GENE, ctcm_dev_setup); | ||
1172 | |||
1173 | if (!dev) { | ||
1174 | ctcm_pr_err("%s: Out of memory\n", __FUNCTION__); | ||
1175 | return NULL; | ||
1176 | } | ||
1177 | dev->priv = priv; | ||
1178 | priv->fsm = init_fsm("ctcmdev", dev_state_names, dev_event_names, | ||
1179 | CTCM_NR_DEV_STATES, CTCM_NR_DEV_EVENTS, | ||
1180 | dev_fsm, dev_fsm_len, GFP_KERNEL); | ||
1181 | if (priv->fsm == NULL) { | ||
1182 | CTCMY_DBF_DEV(SETUP, dev, "init_fsm error"); | ||
1183 | kfree(dev); | ||
1184 | return NULL; | ||
1185 | } | ||
1186 | fsm_newstate(priv->fsm, DEV_STATE_STOPPED); | ||
1187 | fsm_settimer(priv->fsm, &priv->restart_timer); | ||
1188 | |||
1189 | if (IS_MPC(priv)) { | ||
1190 | /* MPC Group Initializations */ | ||
1191 | grp = ctcmpc_init_mpc_group(priv); | ||
1192 | if (grp == NULL) { | ||
1193 | MPC_DBF_DEV(SETUP, dev, "init_mpc_group error"); | ||
1194 | kfree(dev); | ||
1195 | return NULL; | ||
1196 | } | ||
1197 | tasklet_init(&grp->mpc_tasklet2, | ||
1198 | mpc_group_ready, (unsigned long)dev); | ||
1199 | dev->mtu = MPC_BUFSIZE_DEFAULT - | ||
1200 | TH_HEADER_LENGTH - PDU_HEADER_LENGTH; | ||
1201 | |||
1202 | dev->hard_start_xmit = ctcmpc_tx; | ||
1203 | dev->hard_header_len = TH_HEADER_LENGTH + PDU_HEADER_LENGTH; | ||
1204 | priv->buffer_size = MPC_BUFSIZE_DEFAULT; | ||
1205 | } else { | ||
1206 | dev->mtu = CTCM_BUFSIZE_DEFAULT - LL_HEADER_LENGTH - 2; | ||
1207 | dev->hard_start_xmit = ctcm_tx; | ||
1208 | dev->hard_header_len = LL_HEADER_LENGTH + 2; | ||
1209 | } | ||
1210 | |||
1211 | CTCMY_DBF_DEV(SETUP, dev, "finished"); | ||
1212 | return dev; | ||
1213 | } | ||
1214 | |||
1215 | /** | ||
1216 | * Main IRQ handler. | ||
1217 | * | ||
1218 | * cdev The ccw_device the interrupt is for. | ||
1219 | * intparm interruption parameter. | ||
1220 | * irb interruption response block. | ||
1221 | */ | ||
1222 | static void ctcm_irq_handler(struct ccw_device *cdev, | ||
1223 | unsigned long intparm, struct irb *irb) | ||
1224 | { | ||
1225 | struct channel *ch; | ||
1226 | struct net_device *dev; | ||
1227 | struct ctcm_priv *priv; | ||
1228 | struct ccwgroup_device *cgdev; | ||
1229 | |||
1230 | CTCM_DBF_TEXT(TRACE, CTC_DBF_DEBUG, __FUNCTION__); | ||
1231 | if (ctcm_check_irb_error(cdev, irb)) | ||
1232 | return; | ||
1233 | |||
1234 | cgdev = dev_get_drvdata(&cdev->dev); | ||
1235 | |||
1236 | /* Check for unsolicited interrupts. */ | ||
1237 | if (cgdev == NULL) { | ||
1238 | ctcm_pr_warn("ctcm: Got unsolicited irq: %s c-%02x d-%02x\n", | ||
1239 | cdev->dev.bus_id, irb->scsw.cstat, | ||
1240 | irb->scsw.dstat); | ||
1241 | return; | ||
1242 | } | ||
1243 | |||
1244 | priv = dev_get_drvdata(&cgdev->dev); | ||
1245 | |||
1246 | /* Try to extract channel from driver data. */ | ||
1247 | if (priv->channel[READ]->cdev == cdev) | ||
1248 | ch = priv->channel[READ]; | ||
1249 | else if (priv->channel[WRITE]->cdev == cdev) | ||
1250 | ch = priv->channel[WRITE]; | ||
1251 | else { | ||
1252 | ctcm_pr_err("ctcm: Can't determine channel for interrupt, " | ||
1253 | "device %s\n", cdev->dev.bus_id); | ||
1254 | return; | ||
1255 | } | ||
1256 | |||
1257 | dev = (struct net_device *)(ch->netdev); | ||
1258 | if (dev == NULL) { | ||
1259 | ctcm_pr_crit("ctcm: %s dev=NULL bus_id=%s, ch=0x%p\n", | ||
1260 | __FUNCTION__, cdev->dev.bus_id, ch); | ||
1261 | return; | ||
1262 | } | ||
1263 | |||
1264 | if (do_debug) | ||
1265 | ctcm_pr_debug("%s: interrupt for device: %s " | ||
1266 | "received c-%02x d-%02x\n", | ||
1267 | dev->name, | ||
1268 | ch->id, | ||
1269 | irb->scsw.cstat, | ||
1270 | irb->scsw.dstat); | ||
1271 | |||
1272 | /* Copy interruption response block. */ | ||
1273 | memcpy(ch->irb, irb, sizeof(struct irb)); | ||
1274 | |||
1275 | /* Check for good subchannel return code, otherwise error message */ | ||
1276 | if (irb->scsw.cstat) { | ||
1277 | fsm_event(ch->fsm, CTC_EVENT_SC_UNKNOWN, ch); | ||
1278 | ctcm_pr_warn("%s: subchannel check for dev: %s - %02x %02x\n", | ||
1279 | dev->name, ch->id, irb->scsw.cstat, | ||
1280 | irb->scsw.dstat); | ||
1281 | return; | ||
1282 | } | ||
1283 | |||
1284 | /* Check the reason-code of a unit check */ | ||
1285 | if (irb->scsw.dstat & DEV_STAT_UNIT_CHECK) { | ||
1286 | ccw_unit_check(ch, irb->ecw[0]); | ||
1287 | return; | ||
1288 | } | ||
1289 | if (irb->scsw.dstat & DEV_STAT_BUSY) { | ||
1290 | if (irb->scsw.dstat & DEV_STAT_ATTENTION) | ||
1291 | fsm_event(ch->fsm, CTC_EVENT_ATTNBUSY, ch); | ||
1292 | else | ||
1293 | fsm_event(ch->fsm, CTC_EVENT_BUSY, ch); | ||
1294 | return; | ||
1295 | } | ||
1296 | if (irb->scsw.dstat & DEV_STAT_ATTENTION) { | ||
1297 | fsm_event(ch->fsm, CTC_EVENT_ATTN, ch); | ||
1298 | return; | ||
1299 | } | ||
1300 | if ((irb->scsw.stctl & SCSW_STCTL_SEC_STATUS) || | ||
1301 | (irb->scsw.stctl == SCSW_STCTL_STATUS_PEND) || | ||
1302 | (irb->scsw.stctl == | ||
1303 | (SCSW_STCTL_ALERT_STATUS | SCSW_STCTL_STATUS_PEND))) | ||
1304 | fsm_event(ch->fsm, CTC_EVENT_FINSTAT, ch); | ||
1305 | else | ||
1306 | fsm_event(ch->fsm, CTC_EVENT_IRQ, ch); | ||
1307 | |||
1308 | } | ||
1309 | |||
1310 | /** | ||
1311 | * Add ctcm specific attributes. | ||
1312 | * Add ctcm private data. | ||
1313 | * | ||
1314 | * cgdev pointer to ccwgroup_device just added | ||
1315 | * | ||
1316 | * returns 0 on success, !0 on failure. | ||
1317 | */ | ||
1318 | static int ctcm_probe_device(struct ccwgroup_device *cgdev) | ||
1319 | { | ||
1320 | struct ctcm_priv *priv; | ||
1321 | int rc; | ||
1322 | |||
1323 | CTCM_DBF_TEXT_(SETUP, CTC_DBF_INFO, "%s %p", __FUNCTION__, cgdev); | ||
1324 | |||
1325 | if (!get_device(&cgdev->dev)) | ||
1326 | return -ENODEV; | ||
1327 | |||
1328 | priv = kzalloc(sizeof(struct ctcm_priv), GFP_KERNEL); | ||
1329 | if (!priv) { | ||
1330 | ctcm_pr_err("%s: Out of memory\n", __FUNCTION__); | ||
1331 | put_device(&cgdev->dev); | ||
1332 | return -ENOMEM; | ||
1333 | } | ||
1334 | |||
1335 | rc = ctcm_add_files(&cgdev->dev); | ||
1336 | if (rc) { | ||
1337 | kfree(priv); | ||
1338 | put_device(&cgdev->dev); | ||
1339 | return rc; | ||
1340 | } | ||
1341 | priv->buffer_size = CTCM_BUFSIZE_DEFAULT; | ||
1342 | cgdev->cdev[0]->handler = ctcm_irq_handler; | ||
1343 | cgdev->cdev[1]->handler = ctcm_irq_handler; | ||
1344 | dev_set_drvdata(&cgdev->dev, priv); | ||
1345 | |||
1346 | return 0; | ||
1347 | } | ||
1348 | |||
1349 | /** | ||
1350 | * Add a new channel to the list of channels. | ||
1351 | * Keeps the channel list sorted. | ||
1352 | * | ||
1353 | * cdev The ccw_device to be added. | ||
1354 | * type The type class of the new channel. | ||
1355 | * priv Points to the private data of the ccwgroup_device. | ||
1356 | * | ||
1357 | * returns 0 on success, !0 on error. | ||
1358 | */ | ||
1359 | static int add_channel(struct ccw_device *cdev, enum channel_types type, | ||
1360 | struct ctcm_priv *priv) | ||
1361 | { | ||
1362 | struct channel **c = &channels; | ||
1363 | struct channel *ch; | ||
1364 | int ccw_num; | ||
1365 | int rc = 0; | ||
1366 | |||
1367 | CTCM_DBF_TEXT(TRACE, 2, __FUNCTION__); | ||
1368 | ch = kzalloc(sizeof(struct channel), GFP_KERNEL); | ||
1369 | if (ch == NULL) | ||
1370 | goto nomem_return; | ||
1371 | |||
1372 | ch->protocol = priv->protocol; | ||
1373 | if (IS_MPC(priv)) { | ||
1374 | ch->discontact_th = (struct th_header *) | ||
1375 | kzalloc(TH_HEADER_LENGTH, gfp_type()); | ||
1376 | if (ch->discontact_th == NULL) | ||
1377 | goto nomem_return; | ||
1378 | |||
1379 | ch->discontact_th->th_blk_flag = TH_DISCONTACT; | ||
1380 | tasklet_init(&ch->ch_disc_tasklet, | ||
1381 | mpc_action_send_discontact, (unsigned long)ch); | ||
1382 | |||
1383 | tasklet_init(&ch->ch_tasklet, ctcmpc_bh, (unsigned long)ch); | ||
1384 | ch->max_bufsize = (MPC_BUFSIZE_DEFAULT - 35); | ||
1385 | ccw_num = 17; | ||
1386 | } else | ||
1387 | ccw_num = 8; | ||
1388 | |||
1389 | ch->ccw = (struct ccw1 *) | ||
1390 | kzalloc(ccw_num * sizeof(struct ccw1), GFP_KERNEL | GFP_DMA); | ||
1391 | if (ch->ccw == NULL) | ||
1392 | goto nomem_return; | ||
1393 | |||
1394 | ch->cdev = cdev; | ||
1395 | snprintf(ch->id, CTCM_ID_SIZE, "ch-%s", cdev->dev.bus_id); | ||
1396 | ch->type = type; | ||
1397 | |||
1398 | /** | ||
1399 | * "static" ccws are used in the following way: | ||
1400 | * | ||
1401 | * ccw[0..2] (Channel program for generic I/O): | ||
1402 | * 0: prepare | ||
1403 | * 1: read or write (depending on direction) with fixed | ||
1404 | * buffer (idal allocated once when buffer is allocated) | ||
1405 | * 2: nop | ||
1406 | * ccw[3..5] (Channel program for direct write of packets) | ||
1407 | * 3: prepare | ||
1408 | * 4: write (idal allocated on every write). | ||
1409 | * 5: nop | ||
1410 | * ccw[6..7] (Channel program for initial channel setup): | ||
1411 | * 6: set extended mode | ||
1412 | * 7: nop | ||
1413 | * | ||
1414 | * ch->ccw[0..5] are initialized in ch_action_start because | ||
1415 | * the channel's direction is yet unknown here. | ||
1416 | * | ||
1417 | * ccws used for xid2 negotiations | ||
1418 | * ch-ccw[8-14] need to be used for the XID exchange either | ||
1419 | * X side XID2 Processing | ||
1420 | * 8: write control | ||
1421 | * 9: write th | ||
1422 | * 10: write XID | ||
1423 | * 11: read th from secondary | ||
1424 | * 12: read XID from secondary | ||
1425 | * 13: read 4 byte ID | ||
1426 | * 14: nop | ||
1427 | * Y side XID Processing | ||
1428 | * 8: sense | ||
1429 | * 9: read th | ||
1430 | * 10: read XID | ||
1431 | * 11: write th | ||
1432 | * 12: write XID | ||
1433 | * 13: write 4 byte ID | ||
1434 | * 14: nop | ||
1435 | * | ||
1436 | * ccws used for double noop due to VM timing issues | ||
1437 | * which result in unrecoverable Busy on channel | ||
1438 | * 15: nop | ||
1439 | * 16: nop | ||
1440 | */ | ||
1441 | ch->ccw[6].cmd_code = CCW_CMD_SET_EXTENDED; | ||
1442 | ch->ccw[6].flags = CCW_FLAG_SLI; | ||
1443 | |||
1444 | ch->ccw[7].cmd_code = CCW_CMD_NOOP; | ||
1445 | ch->ccw[7].flags = CCW_FLAG_SLI; | ||
1446 | |||
1447 | if (IS_MPC(priv)) { | ||
1448 | ch->ccw[15].cmd_code = CCW_CMD_WRITE; | ||
1449 | ch->ccw[15].flags = CCW_FLAG_SLI | CCW_FLAG_CC; | ||
1450 | ch->ccw[15].count = TH_HEADER_LENGTH; | ||
1451 | ch->ccw[15].cda = virt_to_phys(ch->discontact_th); | ||
1452 | |||
1453 | ch->ccw[16].cmd_code = CCW_CMD_NOOP; | ||
1454 | ch->ccw[16].flags = CCW_FLAG_SLI; | ||
1455 | |||
1456 | ch->fsm = init_fsm(ch->id, ctc_ch_state_names, | ||
1457 | ctc_ch_event_names, CTC_MPC_NR_STATES, | ||
1458 | CTC_MPC_NR_EVENTS, ctcmpc_ch_fsm, | ||
1459 | mpc_ch_fsm_len, GFP_KERNEL); | ||
1460 | } else { | ||
1461 | ch->fsm = init_fsm(ch->id, ctc_ch_state_names, | ||
1462 | ctc_ch_event_names, CTC_NR_STATES, | ||
1463 | CTC_NR_EVENTS, ch_fsm, | ||
1464 | ch_fsm_len, GFP_KERNEL); | ||
1465 | } | ||
1466 | if (ch->fsm == NULL) | ||
1467 | goto free_return; | ||
1468 | |||
1469 | fsm_newstate(ch->fsm, CTC_STATE_IDLE); | ||
1470 | |||
1471 | ch->irb = kzalloc(sizeof(struct irb), GFP_KERNEL); | ||
1472 | if (ch->irb == NULL) | ||
1473 | goto nomem_return; | ||
1474 | |||
1475 | while (*c && ctcm_less_than((*c)->id, ch->id)) | ||
1476 | c = &(*c)->next; | ||
1477 | |||
1478 | if (*c && (!strncmp((*c)->id, ch->id, CTCM_ID_SIZE))) { | ||
1479 | CTCM_DBF_TEXT_(SETUP, CTC_DBF_INFO, | ||
1480 | "%s (%s) already in list, using old entry", | ||
1481 | __FUNCTION__, (*c)->id); | ||
1482 | |||
1483 | goto free_return; | ||
1484 | } | ||
1485 | |||
1486 | spin_lock_init(&ch->collect_lock); | ||
1487 | |||
1488 | fsm_settimer(ch->fsm, &ch->timer); | ||
1489 | skb_queue_head_init(&ch->io_queue); | ||
1490 | skb_queue_head_init(&ch->collect_queue); | ||
1491 | |||
1492 | if (IS_MPC(priv)) { | ||
1493 | fsm_settimer(ch->fsm, &ch->sweep_timer); | ||
1494 | skb_queue_head_init(&ch->sweep_queue); | ||
1495 | } | ||
1496 | ch->next = *c; | ||
1497 | *c = ch; | ||
1498 | return 0; | ||
1499 | |||
1500 | nomem_return: | ||
1501 | ctcm_pr_warn("ctcm: Out of memory in %s\n", __FUNCTION__); | ||
1502 | rc = -ENOMEM; | ||
1503 | |||
1504 | free_return: /* note that all channel pointers are 0 or valid */ | ||
1505 | kfree(ch->ccw); /* TODO: check that again */ | ||
1506 | kfree(ch->discontact_th); | ||
1507 | kfree_fsm(ch->fsm); | ||
1508 | kfree(ch->irb); | ||
1509 | kfree(ch); | ||
1510 | return rc; | ||
1511 | } | ||
1512 | |||
1513 | /* | ||
1514 | * Return type of a detected device. | ||
1515 | */ | ||
1516 | static enum channel_types get_channel_type(struct ccw_device_id *id) | ||
1517 | { | ||
1518 | enum channel_types type; | ||
1519 | type = (enum channel_types)id->driver_info; | ||
1520 | |||
1521 | if (type == channel_type_ficon) | ||
1522 | type = channel_type_escon; | ||
1523 | |||
1524 | return type; | ||
1525 | } | ||
1526 | |||
1527 | /** | ||
1528 | * | ||
1529 | * Setup an interface. | ||
1530 | * | ||
1531 | * cgdev Device to be setup. | ||
1532 | * | ||
1533 | * returns 0 on success, !0 on failure. | ||
1534 | */ | ||
1535 | static int ctcm_new_device(struct ccwgroup_device *cgdev) | ||
1536 | { | ||
1537 | char read_id[CTCM_ID_SIZE]; | ||
1538 | char write_id[CTCM_ID_SIZE]; | ||
1539 | int direction; | ||
1540 | enum channel_types type; | ||
1541 | struct ctcm_priv *priv; | ||
1542 | struct net_device *dev; | ||
1543 | int ret; | ||
1544 | |||
1545 | CTCM_DBF_TEXT(SETUP, CTC_DBF_INFO, __FUNCTION__); | ||
1546 | |||
1547 | priv = dev_get_drvdata(&cgdev->dev); | ||
1548 | if (!priv) | ||
1549 | return -ENODEV; | ||
1550 | |||
1551 | type = get_channel_type(&cgdev->cdev[0]->id); | ||
1552 | |||
1553 | snprintf(read_id, CTCM_ID_SIZE, "ch-%s", cgdev->cdev[0]->dev.bus_id); | ||
1554 | snprintf(write_id, CTCM_ID_SIZE, "ch-%s", cgdev->cdev[1]->dev.bus_id); | ||
1555 | |||
1556 | ret = add_channel(cgdev->cdev[0], type, priv); | ||
1557 | if (ret) | ||
1558 | return ret; | ||
1559 | ret = add_channel(cgdev->cdev[1], type, priv); | ||
1560 | if (ret) | ||
1561 | return ret; | ||
1562 | |||
1563 | ret = ccw_device_set_online(cgdev->cdev[0]); | ||
1564 | if (ret != 0) { | ||
1565 | CTCM_DBF_TEXT(SETUP, CTC_DBF_WARN, | ||
1566 | "ccw_device_set_online (cdev[0]) failed "); | ||
1567 | ctcm_pr_warn("ccw_device_set_online (cdev[0]) failed " | ||
1568 | "with ret = %d\n", ret); | ||
1569 | } | ||
1570 | |||
1571 | ret = ccw_device_set_online(cgdev->cdev[1]); | ||
1572 | if (ret != 0) { | ||
1573 | CTCM_DBF_TEXT(SETUP, CTC_DBF_WARN, | ||
1574 | "ccw_device_set_online (cdev[1]) failed "); | ||
1575 | ctcm_pr_warn("ccw_device_set_online (cdev[1]) failed " | ||
1576 | "with ret = %d\n", ret); | ||
1577 | } | ||
1578 | |||
1579 | dev = ctcm_init_netdevice(priv); | ||
1580 | |||
1581 | if (dev == NULL) { | ||
1582 | ctcm_pr_warn("ctcm_init_netdevice failed\n"); | ||
1583 | goto out; | ||
1584 | } | ||
1585 | |||
1586 | for (direction = READ; direction <= WRITE; direction++) { | ||
1587 | priv->channel[direction] = | ||
1588 | channel_get(type, direction == READ ? read_id : write_id, | ||
1589 | direction); | ||
1590 | if (priv->channel[direction] == NULL) { | ||
1591 | if (direction == WRITE) | ||
1592 | channel_free(priv->channel[READ]); | ||
1593 | ctcm_free_netdevice(dev); | ||
1594 | goto out; | ||
1595 | } | ||
1596 | priv->channel[direction]->netdev = dev; | ||
1597 | priv->channel[direction]->protocol = priv->protocol; | ||
1598 | priv->channel[direction]->max_bufsize = priv->buffer_size; | ||
1599 | } | ||
1600 | /* sysfs magic */ | ||
1601 | SET_NETDEV_DEV(dev, &cgdev->dev); | ||
1602 | |||
1603 | if (ctcm_netdev_register(dev) != 0) { | ||
1604 | ctcm_free_netdevice(dev); | ||
1605 | goto out; | ||
1606 | } | ||
1607 | |||
1608 | if (ctcm_add_attributes(&cgdev->dev)) { | ||
1609 | ctcm_netdev_unregister(dev); | ||
1610 | /* dev->priv = NULL; why that ???? */ | ||
1611 | ctcm_free_netdevice(dev); | ||
1612 | goto out; | ||
1613 | } | ||
1614 | |||
1615 | strlcpy(priv->fsm->name, dev->name, sizeof(priv->fsm->name)); | ||
1616 | |||
1617 | CTCM_DBF_TEXT_(SETUP, CTC_DBF_INFO, | ||
1618 | "setup(%s) ok : r/w = %s / %s, proto : %d", | ||
1619 | dev->name, priv->channel[READ]->id, | ||
1620 | priv->channel[WRITE]->id, priv->protocol); | ||
1621 | |||
1622 | return 0; | ||
1623 | out: | ||
1624 | ccw_device_set_offline(cgdev->cdev[1]); | ||
1625 | ccw_device_set_offline(cgdev->cdev[0]); | ||
1626 | |||
1627 | return -ENODEV; | ||
1628 | } | ||
1629 | |||
1630 | /** | ||
1631 | * Shutdown an interface. | ||
1632 | * | ||
1633 | * cgdev Device to be shut down. | ||
1634 | * | ||
1635 | * returns 0 on success, !0 on failure. | ||
1636 | */ | ||
1637 | static int ctcm_shutdown_device(struct ccwgroup_device *cgdev) | ||
1638 | { | ||
1639 | struct ctcm_priv *priv; | ||
1640 | struct net_device *dev; | ||
1641 | |||
1642 | priv = dev_get_drvdata(&cgdev->dev); | ||
1643 | if (!priv) | ||
1644 | return -ENODEV; | ||
1645 | |||
1646 | if (priv->channel[READ]) { | ||
1647 | dev = priv->channel[READ]->netdev; | ||
1648 | CTCM_DBF_DEV(SETUP, dev, ""); | ||
1649 | /* Close the device */ | ||
1650 | ctcm_close(dev); | ||
1651 | dev->flags &= ~IFF_RUNNING; | ||
1652 | ctcm_remove_attributes(&cgdev->dev); | ||
1653 | channel_free(priv->channel[READ]); | ||
1654 | } else | ||
1655 | dev = NULL; | ||
1656 | |||
1657 | if (priv->channel[WRITE]) | ||
1658 | channel_free(priv->channel[WRITE]); | ||
1659 | |||
1660 | if (dev) { | ||
1661 | ctcm_netdev_unregister(dev); | ||
1662 | /* dev->priv = NULL; why that ??? */ | ||
1663 | ctcm_free_netdevice(dev); | ||
1664 | } | ||
1665 | |||
1666 | if (priv->fsm) | ||
1667 | kfree_fsm(priv->fsm); | ||
1668 | |||
1669 | ccw_device_set_offline(cgdev->cdev[1]); | ||
1670 | ccw_device_set_offline(cgdev->cdev[0]); | ||
1671 | |||
1672 | if (priv->channel[READ]) | ||
1673 | channel_remove(priv->channel[READ]); | ||
1674 | if (priv->channel[WRITE]) | ||
1675 | channel_remove(priv->channel[WRITE]); | ||
1676 | priv->channel[READ] = priv->channel[WRITE] = NULL; | ||
1677 | |||
1678 | return 0; | ||
1679 | |||
1680 | } | ||
1681 | |||
1682 | |||
1683 | static void ctcm_remove_device(struct ccwgroup_device *cgdev) | ||
1684 | { | ||
1685 | struct ctcm_priv *priv; | ||
1686 | |||
1687 | CTCM_DBF_TEXT(SETUP, CTC_DBF_ERROR, __FUNCTION__); | ||
1688 | |||
1689 | priv = dev_get_drvdata(&cgdev->dev); | ||
1690 | if (!priv) | ||
1691 | return; | ||
1692 | if (cgdev->state == CCWGROUP_ONLINE) | ||
1693 | ctcm_shutdown_device(cgdev); | ||
1694 | ctcm_remove_files(&cgdev->dev); | ||
1695 | dev_set_drvdata(&cgdev->dev, NULL); | ||
1696 | kfree(priv); | ||
1697 | put_device(&cgdev->dev); | ||
1698 | } | ||
1699 | |||
1700 | static struct ccwgroup_driver ctcm_group_driver = { | ||
1701 | .owner = THIS_MODULE, | ||
1702 | .name = CTC_DRIVER_NAME, | ||
1703 | .max_slaves = 2, | ||
1704 | .driver_id = 0xC3E3C3D4, /* CTCM */ | ||
1705 | .probe = ctcm_probe_device, | ||
1706 | .remove = ctcm_remove_device, | ||
1707 | .set_online = ctcm_new_device, | ||
1708 | .set_offline = ctcm_shutdown_device, | ||
1709 | }; | ||
1710 | |||
1711 | |||
1712 | /* | ||
1713 | * Module related routines | ||
1714 | */ | ||
1715 | |||
1716 | /* | ||
1717 | * Prepare to be unloaded. Free IRQ's and release all resources. | ||
1718 | * This is called just before this module is unloaded. It is | ||
1719 | * not called, if the usage count is !0, so we don't need to check | ||
1720 | * for that. | ||
1721 | */ | ||
1722 | static void __exit ctcm_exit(void) | ||
1723 | { | ||
1724 | unregister_cu3088_discipline(&ctcm_group_driver); | ||
1725 | ctcm_unregister_dbf_views(); | ||
1726 | ctcm_pr_info("CTCM driver unloaded\n"); | ||
1727 | } | ||
1728 | |||
1729 | /* | ||
1730 | * Print Banner. | ||
1731 | */ | ||
1732 | static void print_banner(void) | ||
1733 | { | ||
1734 | printk(KERN_INFO "CTCM driver initialized\n"); | ||
1735 | } | ||
1736 | |||
1737 | /** | ||
1738 | * Initialize module. | ||
1739 | * This is called just after the module is loaded. | ||
1740 | * | ||
1741 | * returns 0 on success, !0 on error. | ||
1742 | */ | ||
1743 | static int __init ctcm_init(void) | ||
1744 | { | ||
1745 | int ret; | ||
1746 | |||
1747 | channels = NULL; | ||
1748 | |||
1749 | ret = ctcm_register_dbf_views(); | ||
1750 | if (ret) { | ||
1751 | ctcm_pr_crit("ctcm_init failed with ctcm_register_dbf_views " | ||
1752 | "rc = %d\n", ret); | ||
1753 | return ret; | ||
1754 | } | ||
1755 | ret = register_cu3088_discipline(&ctcm_group_driver); | ||
1756 | if (ret) { | ||
1757 | ctcm_unregister_dbf_views(); | ||
1758 | ctcm_pr_crit("ctcm_init failed with register_cu3088_discipline " | ||
1759 | "(rc = %d)\n", ret); | ||
1760 | return ret; | ||
1761 | } | ||
1762 | print_banner(); | ||
1763 | return ret; | ||
1764 | } | ||
1765 | |||
1766 | module_init(ctcm_init); | ||
1767 | module_exit(ctcm_exit); | ||
1768 | |||
1769 | MODULE_AUTHOR("Peter Tiedemann <ptiedem@de.ibm.com>"); | ||
1770 | MODULE_DESCRIPTION("Network driver for S/390 CTC + CTCMPC (SNA)"); | ||
1771 | MODULE_LICENSE("GPL"); | ||
1772 | |||
diff --git a/drivers/s390/net/ctcm_main.h b/drivers/s390/net/ctcm_main.h new file mode 100644 index 000000000000..95b0c0b6ebc6 --- /dev/null +++ b/drivers/s390/net/ctcm_main.h | |||
@@ -0,0 +1,287 @@ | |||
1 | /* | ||
2 | * drivers/s390/net/ctcm_main.h | ||
3 | * | ||
4 | * Copyright IBM Corp. 2001, 2007 | ||
5 | * Authors: Fritz Elfert (felfert@millenux.com) | ||
6 | * Peter Tiedemann (ptiedem@de.ibm.com) | ||
7 | */ | ||
8 | |||
9 | #ifndef _CTCM_MAIN_H_ | ||
10 | #define _CTCM_MAIN_H_ | ||
11 | |||
12 | #include <asm/ccwdev.h> | ||
13 | #include <asm/ccwgroup.h> | ||
14 | |||
15 | #include <linux/skbuff.h> | ||
16 | #include <linux/netdevice.h> | ||
17 | |||
18 | #include "fsm.h" | ||
19 | #include "cu3088.h" | ||
20 | #include "ctcm_dbug.h" | ||
21 | #include "ctcm_mpc.h" | ||
22 | |||
23 | #define CTC_DRIVER_NAME "ctcm" | ||
24 | #define CTC_DEVICE_NAME "ctc" | ||
25 | #define CTC_DEVICE_GENE "ctc%d" | ||
26 | #define MPC_DEVICE_NAME "mpc" | ||
27 | #define MPC_DEVICE_GENE "mpc%d" | ||
28 | |||
29 | #define CHANNEL_FLAGS_READ 0 | ||
30 | #define CHANNEL_FLAGS_WRITE 1 | ||
31 | #define CHANNEL_FLAGS_INUSE 2 | ||
32 | #define CHANNEL_FLAGS_BUFSIZE_CHANGED 4 | ||
33 | #define CHANNEL_FLAGS_FAILED 8 | ||
34 | #define CHANNEL_FLAGS_WAITIRQ 16 | ||
35 | #define CHANNEL_FLAGS_RWMASK 1 | ||
36 | #define CHANNEL_DIRECTION(f) (f & CHANNEL_FLAGS_RWMASK) | ||
37 | |||
38 | #define LOG_FLAG_ILLEGALPKT 1 | ||
39 | #define LOG_FLAG_ILLEGALSIZE 2 | ||
40 | #define LOG_FLAG_OVERRUN 4 | ||
41 | #define LOG_FLAG_NOMEM 8 | ||
42 | |||
43 | #define ctcm_pr_debug(fmt, arg...) printk(KERN_DEBUG fmt, ##arg) | ||
44 | #define ctcm_pr_info(fmt, arg...) printk(KERN_INFO fmt, ##arg) | ||
45 | #define ctcm_pr_notice(fmt, arg...) printk(KERN_NOTICE fmt, ##arg) | ||
46 | #define ctcm_pr_warn(fmt, arg...) printk(KERN_WARNING fmt, ##arg) | ||
47 | #define ctcm_pr_emerg(fmt, arg...) printk(KERN_EMERG fmt, ##arg) | ||
48 | #define ctcm_pr_err(fmt, arg...) printk(KERN_ERR fmt, ##arg) | ||
49 | #define ctcm_pr_crit(fmt, arg...) printk(KERN_CRIT fmt, ##arg) | ||
50 | |||
51 | /* | ||
52 | * CCW commands, used in this driver. | ||
53 | */ | ||
54 | #define CCW_CMD_WRITE 0x01 | ||
55 | #define CCW_CMD_READ 0x02 | ||
56 | #define CCW_CMD_NOOP 0x03 | ||
57 | #define CCW_CMD_TIC 0x08 | ||
58 | #define CCW_CMD_SENSE_CMD 0x14 | ||
59 | #define CCW_CMD_WRITE_CTL 0x17 | ||
60 | #define CCW_CMD_SET_EXTENDED 0xc3 | ||
61 | #define CCW_CMD_PREPARE 0xe3 | ||
62 | |||
63 | #define CTCM_PROTO_S390 0 | ||
64 | #define CTCM_PROTO_LINUX 1 | ||
65 | #define CTCM_PROTO_LINUX_TTY 2 | ||
66 | #define CTCM_PROTO_OS390 3 | ||
67 | #define CTCM_PROTO_MPC 4 | ||
68 | #define CTCM_PROTO_MAX 4 | ||
69 | |||
70 | #define CTCM_BUFSIZE_LIMIT 65535 | ||
71 | #define CTCM_BUFSIZE_DEFAULT 32768 | ||
72 | #define MPC_BUFSIZE_DEFAULT CTCM_BUFSIZE_LIMIT | ||
73 | |||
74 | #define CTCM_TIME_1_SEC 1000 | ||
75 | #define CTCM_TIME_5_SEC 5000 | ||
76 | #define CTCM_TIME_10_SEC 10000 | ||
77 | |||
78 | #define CTCM_INITIAL_BLOCKLEN 2 | ||
79 | |||
80 | #define READ 0 | ||
81 | #define WRITE 1 | ||
82 | |||
83 | #define CTCM_ID_SIZE BUS_ID_SIZE+3 | ||
84 | |||
85 | struct ctcm_profile { | ||
86 | unsigned long maxmulti; | ||
87 | unsigned long maxcqueue; | ||
88 | unsigned long doios_single; | ||
89 | unsigned long doios_multi; | ||
90 | unsigned long txlen; | ||
91 | unsigned long tx_time; | ||
92 | struct timespec send_stamp; | ||
93 | }; | ||
94 | |||
95 | /* | ||
96 | * Definition of one channel | ||
97 | */ | ||
98 | struct channel { | ||
99 | struct channel *next; | ||
100 | char id[CTCM_ID_SIZE]; | ||
101 | struct ccw_device *cdev; | ||
102 | /* | ||
103 | * Type of this channel. | ||
104 | * CTC/A or Escon for valid channels. | ||
105 | */ | ||
106 | enum channel_types type; | ||
107 | /* | ||
108 | * Misc. flags. See CHANNEL_FLAGS_... below | ||
109 | */ | ||
110 | __u32 flags; | ||
111 | __u16 protocol; /* protocol of this channel (4 = MPC) */ | ||
112 | /* | ||
113 | * I/O and irq related stuff | ||
114 | */ | ||
115 | struct ccw1 *ccw; | ||
116 | struct irb *irb; | ||
117 | /* | ||
118 | * RX/TX buffer size | ||
119 | */ | ||
120 | int max_bufsize; | ||
121 | struct sk_buff *trans_skb; /* transmit/receive buffer */ | ||
122 | struct sk_buff_head io_queue; /* universal I/O queue */ | ||
123 | struct tasklet_struct ch_tasklet; /* MPC ONLY */ | ||
124 | /* | ||
125 | * TX queue for collecting skb's during busy. | ||
126 | */ | ||
127 | struct sk_buff_head collect_queue; | ||
128 | /* | ||
129 | * Amount of data in collect_queue. | ||
130 | */ | ||
131 | int collect_len; | ||
132 | /* | ||
133 | * spinlock for collect_queue and collect_len | ||
134 | */ | ||
135 | spinlock_t collect_lock; | ||
136 | /* | ||
137 | * Timer for detecting unresposive | ||
138 | * I/O operations. | ||
139 | */ | ||
140 | fsm_timer timer; | ||
141 | /* MPC ONLY section begin */ | ||
142 | __u32 th_seq_num; /* SNA TH seq number */ | ||
143 | __u8 th_seg; | ||
144 | __u32 pdu_seq; | ||
145 | struct sk_buff *xid_skb; | ||
146 | char *xid_skb_data; | ||
147 | struct th_header *xid_th; | ||
148 | struct xid2 *xid; | ||
149 | char *xid_id; | ||
150 | struct th_header *rcvd_xid_th; | ||
151 | struct xid2 *rcvd_xid; | ||
152 | char *rcvd_xid_id; | ||
153 | __u8 in_mpcgroup; | ||
154 | fsm_timer sweep_timer; | ||
155 | struct sk_buff_head sweep_queue; | ||
156 | struct th_header *discontact_th; | ||
157 | struct tasklet_struct ch_disc_tasklet; | ||
158 | /* MPC ONLY section end */ | ||
159 | |||
160 | int retry; /* retry counter for misc. operations */ | ||
161 | fsm_instance *fsm; /* finite state machine of this channel */ | ||
162 | struct net_device *netdev; /* corresponding net_device */ | ||
163 | struct ctcm_profile prof; | ||
164 | unsigned char *trans_skb_data; | ||
165 | __u16 logflags; | ||
166 | }; | ||
167 | |||
168 | struct ctcm_priv { | ||
169 | struct net_device_stats stats; | ||
170 | unsigned long tbusy; | ||
171 | |||
172 | /* The MPC group struct of this interface */ | ||
173 | struct mpc_group *mpcg; /* MPC only */ | ||
174 | struct xid2 *xid; /* MPC only */ | ||
175 | |||
176 | /* The finite state machine of this interface */ | ||
177 | fsm_instance *fsm; | ||
178 | |||
179 | /* The protocol of this device */ | ||
180 | __u16 protocol; | ||
181 | |||
182 | /* Timer for restarting after I/O Errors */ | ||
183 | fsm_timer restart_timer; | ||
184 | |||
185 | int buffer_size; /* ctc only */ | ||
186 | |||
187 | struct channel *channel[2]; | ||
188 | }; | ||
189 | |||
190 | int ctcm_open(struct net_device *dev); | ||
191 | int ctcm_close(struct net_device *dev); | ||
192 | |||
193 | /* | ||
194 | * prototypes for non-static sysfs functions | ||
195 | */ | ||
196 | int ctcm_add_attributes(struct device *dev); | ||
197 | void ctcm_remove_attributes(struct device *dev); | ||
198 | int ctcm_add_files(struct device *dev); | ||
199 | void ctcm_remove_files(struct device *dev); | ||
200 | |||
201 | /* | ||
202 | * Compatibility macros for busy handling | ||
203 | * of network devices. | ||
204 | */ | ||
205 | static inline void ctcm_clear_busy_do(struct net_device *dev) | ||
206 | { | ||
207 | clear_bit(0, &(((struct ctcm_priv *)dev->priv)->tbusy)); | ||
208 | netif_wake_queue(dev); | ||
209 | } | ||
210 | |||
211 | static inline void ctcm_clear_busy(struct net_device *dev) | ||
212 | { | ||
213 | struct mpc_group *grp; | ||
214 | grp = ((struct ctcm_priv *)dev->priv)->mpcg; | ||
215 | |||
216 | if (!(grp && grp->in_sweep)) | ||
217 | ctcm_clear_busy_do(dev); | ||
218 | } | ||
219 | |||
220 | |||
221 | static inline int ctcm_test_and_set_busy(struct net_device *dev) | ||
222 | { | ||
223 | netif_stop_queue(dev); | ||
224 | return test_and_set_bit(0, &(((struct ctcm_priv *)dev->priv)->tbusy)); | ||
225 | } | ||
226 | |||
227 | extern int loglevel; | ||
228 | extern struct channel *channels; | ||
229 | |||
230 | void ctcm_unpack_skb(struct channel *ch, struct sk_buff *pskb); | ||
231 | |||
232 | /* | ||
233 | * Functions related to setup and device detection. | ||
234 | */ | ||
235 | |||
236 | static inline int ctcm_less_than(char *id1, char *id2) | ||
237 | { | ||
238 | unsigned long dev1, dev2; | ||
239 | |||
240 | id1 = id1 + 5; | ||
241 | id2 = id2 + 5; | ||
242 | |||
243 | dev1 = simple_strtoul(id1, &id1, 16); | ||
244 | dev2 = simple_strtoul(id2, &id2, 16); | ||
245 | |||
246 | return (dev1 < dev2); | ||
247 | } | ||
248 | |||
249 | int ctcm_ch_alloc_buffer(struct channel *ch); | ||
250 | |||
251 | static inline int ctcm_checkalloc_buffer(struct channel *ch) | ||
252 | { | ||
253 | if (ch->trans_skb == NULL) | ||
254 | return ctcm_ch_alloc_buffer(ch); | ||
255 | if (ch->flags & CHANNEL_FLAGS_BUFSIZE_CHANGED) { | ||
256 | dev_kfree_skb(ch->trans_skb); | ||
257 | return ctcm_ch_alloc_buffer(ch); | ||
258 | } | ||
259 | return 0; | ||
260 | } | ||
261 | |||
262 | struct mpc_group *ctcmpc_init_mpc_group(struct ctcm_priv *priv); | ||
263 | |||
264 | /* test if protocol attribute (of struct ctcm_priv or struct channel) | ||
265 | * has MPC protocol setting. Type is not checked | ||
266 | */ | ||
267 | #define IS_MPC(p) ((p)->protocol == CTCM_PROTO_MPC) | ||
268 | |||
269 | /* test if struct ctcm_priv of struct net_device has MPC protocol setting */ | ||
270 | #define IS_MPCDEV(d) IS_MPC((struct ctcm_priv *)d->priv) | ||
271 | |||
272 | static inline gfp_t gfp_type(void) | ||
273 | { | ||
274 | return in_interrupt() ? GFP_ATOMIC : GFP_KERNEL; | ||
275 | } | ||
276 | |||
277 | /* | ||
278 | * Definition of our link level header. | ||
279 | */ | ||
280 | struct ll_header { | ||
281 | __u16 length; | ||
282 | __u16 type; | ||
283 | __u16 unused; | ||
284 | }; | ||
285 | #define LL_HEADER_LENGTH (sizeof(struct ll_header)) | ||
286 | |||
287 | #endif | ||
diff --git a/drivers/s390/net/ctcm_mpc.c b/drivers/s390/net/ctcm_mpc.c new file mode 100644 index 000000000000..044addee64a2 --- /dev/null +++ b/drivers/s390/net/ctcm_mpc.c | |||
@@ -0,0 +1,2472 @@ | |||
1 | /* | ||
2 | * drivers/s390/net/ctcm_mpc.c | ||
3 | * | ||
4 | * Copyright IBM Corp. 2004, 2007 | ||
5 | * Authors: Belinda Thompson (belindat@us.ibm.com) | ||
6 | * Andy Richter (richtera@us.ibm.com) | ||
7 | * Peter Tiedemann (ptiedem@de.ibm.com) | ||
8 | */ | ||
9 | |||
10 | /* | ||
11 | This module exports functions to be used by CCS: | ||
12 | EXPORT_SYMBOL(ctc_mpc_alloc_channel); | ||
13 | EXPORT_SYMBOL(ctc_mpc_establish_connectivity); | ||
14 | EXPORT_SYMBOL(ctc_mpc_dealloc_ch); | ||
15 | EXPORT_SYMBOL(ctc_mpc_flow_control); | ||
16 | */ | ||
17 | |||
18 | #undef DEBUG | ||
19 | #undef DEBUGDATA | ||
20 | #undef DEBUGCCW | ||
21 | |||
22 | #include <linux/version.h> | ||
23 | #include <linux/module.h> | ||
24 | #include <linux/init.h> | ||
25 | #include <linux/kernel.h> | ||
26 | #include <linux/slab.h> | ||
27 | #include <linux/errno.h> | ||
28 | #include <linux/types.h> | ||
29 | #include <linux/interrupt.h> | ||
30 | #include <linux/timer.h> | ||
31 | #include <linux/sched.h> | ||
32 | |||
33 | #include <linux/signal.h> | ||
34 | #include <linux/string.h> | ||
35 | #include <linux/proc_fs.h> | ||
36 | |||
37 | #include <linux/ip.h> | ||
38 | #include <linux/if_arp.h> | ||
39 | #include <linux/tcp.h> | ||
40 | #include <linux/skbuff.h> | ||
41 | #include <linux/ctype.h> | ||
42 | #include <linux/netdevice.h> | ||
43 | #include <net/dst.h> | ||
44 | |||
45 | #include <linux/io.h> /* instead of <asm/io.h> ok ? */ | ||
46 | #include <asm/ccwdev.h> | ||
47 | #include <asm/ccwgroup.h> | ||
48 | #include <linux/bitops.h> /* instead of <asm/bitops.h> ok ? */ | ||
49 | #include <linux/uaccess.h> /* instead of <asm/uaccess.h> ok ? */ | ||
50 | #include <linux/wait.h> | ||
51 | #include <linux/moduleparam.h> | ||
52 | #include <asm/idals.h> | ||
53 | |||
54 | #include "cu3088.h" | ||
55 | #include "ctcm_mpc.h" | ||
56 | #include "ctcm_main.h" | ||
57 | #include "ctcm_fsms.h" | ||
58 | |||
59 | static const struct xid2 init_xid = { | ||
60 | .xid2_type_id = XID_FM2, | ||
61 | .xid2_len = 0x45, | ||
62 | .xid2_adj_id = 0, | ||
63 | .xid2_rlen = 0x31, | ||
64 | .xid2_resv1 = 0, | ||
65 | .xid2_flag1 = 0, | ||
66 | .xid2_fmtt = 0, | ||
67 | .xid2_flag4 = 0x80, | ||
68 | .xid2_resv2 = 0, | ||
69 | .xid2_tgnum = 0, | ||
70 | .xid2_sender_id = 0, | ||
71 | .xid2_flag2 = 0, | ||
72 | .xid2_option = XID2_0, | ||
73 | .xid2_resv3 = "\x00", | ||
74 | .xid2_resv4 = 0, | ||
75 | .xid2_dlc_type = XID2_READ_SIDE, | ||
76 | .xid2_resv5 = 0, | ||
77 | .xid2_mpc_flag = 0, | ||
78 | .xid2_resv6 = 0, | ||
79 | .xid2_buf_len = (MPC_BUFSIZE_DEFAULT - 35), | ||
80 | }; | ||
81 | |||
82 | static const struct th_header thnorm = { | ||
83 | .th_seg = 0x00, | ||
84 | .th_ch_flag = TH_IS_XID, | ||
85 | .th_blk_flag = TH_DATA_IS_XID, | ||
86 | .th_is_xid = 0x01, | ||
87 | .th_seq_num = 0x00000000, | ||
88 | }; | ||
89 | |||
90 | static const struct th_header thdummy = { | ||
91 | .th_seg = 0x00, | ||
92 | .th_ch_flag = 0x00, | ||
93 | .th_blk_flag = TH_DATA_IS_XID, | ||
94 | .th_is_xid = 0x01, | ||
95 | .th_seq_num = 0x00000000, | ||
96 | }; | ||
97 | |||
98 | /* | ||
99 | * Definition of one MPC group | ||
100 | */ | ||
101 | |||
102 | /* | ||
103 | * Compatibility macros for busy handling | ||
104 | * of network devices. | ||
105 | */ | ||
106 | |||
107 | static void ctcmpc_unpack_skb(struct channel *ch, struct sk_buff *pskb); | ||
108 | |||
109 | /* | ||
110 | * MPC Group state machine actions (static prototypes) | ||
111 | */ | ||
112 | static void mpc_action_nop(fsm_instance *fsm, int event, void *arg); | ||
113 | static void mpc_action_go_ready(fsm_instance *fsm, int event, void *arg); | ||
114 | static void mpc_action_go_inop(fsm_instance *fi, int event, void *arg); | ||
115 | static void mpc_action_timeout(fsm_instance *fi, int event, void *arg); | ||
116 | static int mpc_validate_xid(struct mpcg_info *mpcginfo); | ||
117 | static void mpc_action_yside_xid(fsm_instance *fsm, int event, void *arg); | ||
118 | static void mpc_action_doxid0(fsm_instance *fsm, int event, void *arg); | ||
119 | static void mpc_action_doxid7(fsm_instance *fsm, int event, void *arg); | ||
120 | static void mpc_action_xside_xid(fsm_instance *fsm, int event, void *arg); | ||
121 | static void mpc_action_rcvd_xid0(fsm_instance *fsm, int event, void *arg); | ||
122 | static void mpc_action_rcvd_xid7(fsm_instance *fsm, int event, void *arg); | ||
123 | |||
124 | #ifdef DEBUGDATA | ||
125 | /*-------------------------------------------------------------------* | ||
126 | * Dump buffer format * | ||
127 | * * | ||
128 | *--------------------------------------------------------------------*/ | ||
129 | void ctcmpc_dumpit(char *buf, int len) | ||
130 | { | ||
131 | __u32 ct, sw, rm, dup; | ||
132 | char *ptr, *rptr; | ||
133 | char tbuf[82], tdup[82]; | ||
134 | #if (UTS_MACHINE == s390x) | ||
135 | char addr[22]; | ||
136 | #else | ||
137 | char addr[12]; | ||
138 | #endif | ||
139 | char boff[12]; | ||
140 | char bhex[82], duphex[82]; | ||
141 | char basc[40]; | ||
142 | |||
143 | sw = 0; | ||
144 | rptr = ptr = buf; | ||
145 | rm = 16; | ||
146 | duphex[0] = 0x00; | ||
147 | dup = 0; | ||
148 | |||
149 | for (ct = 0; ct < len; ct++, ptr++, rptr++) { | ||
150 | if (sw == 0) { | ||
151 | #if (UTS_MACHINE == s390x) | ||
152 | sprintf(addr, "%16.16lx", (unsigned long)rptr); | ||
153 | #else | ||
154 | sprintf(addr, "%8.8X", (__u32)rptr); | ||
155 | #endif | ||
156 | |||
157 | sprintf(boff, "%4.4X", (__u32)ct); | ||
158 | bhex[0] = '\0'; | ||
159 | basc[0] = '\0'; | ||
160 | } | ||
161 | if ((sw == 4) || (sw == 12)) | ||
162 | strcat(bhex, " "); | ||
163 | if (sw == 8) | ||
164 | strcat(bhex, " "); | ||
165 | |||
166 | #if (UTS_MACHINE == s390x) | ||
167 | sprintf(tbuf, "%2.2lX", (unsigned long)*ptr); | ||
168 | #else | ||
169 | sprintf(tbuf, "%2.2X", (__u32)*ptr); | ||
170 | #endif | ||
171 | |||
172 | tbuf[2] = '\0'; | ||
173 | strcat(bhex, tbuf); | ||
174 | if ((0 != isprint(*ptr)) && (*ptr >= 0x20)) | ||
175 | basc[sw] = *ptr; | ||
176 | else | ||
177 | basc[sw] = '.'; | ||
178 | |||
179 | basc[sw+1] = '\0'; | ||
180 | sw++; | ||
181 | rm--; | ||
182 | if (sw == 16) { | ||
183 | if ((strcmp(duphex, bhex)) != 0) { | ||
184 | if (dup != 0) { | ||
185 | sprintf(tdup, "Duplicate as above " | ||
186 | "to %s", addr); | ||
187 | printk(KERN_INFO " " | ||
188 | " --- %s ---\n", tdup); | ||
189 | } | ||
190 | printk(KERN_INFO " %s (+%s) : %s [%s]\n", | ||
191 | addr, boff, bhex, basc); | ||
192 | dup = 0; | ||
193 | strcpy(duphex, bhex); | ||
194 | } else | ||
195 | dup++; | ||
196 | |||
197 | sw = 0; | ||
198 | rm = 16; | ||
199 | } | ||
200 | } /* endfor */ | ||
201 | |||
202 | if (sw != 0) { | ||
203 | for ( ; rm > 0; rm--, sw++) { | ||
204 | if ((sw == 4) || (sw == 12)) | ||
205 | strcat(bhex, " "); | ||
206 | if (sw == 8) | ||
207 | strcat(bhex, " "); | ||
208 | strcat(bhex, " "); | ||
209 | strcat(basc, " "); | ||
210 | } | ||
211 | if (dup != 0) { | ||
212 | sprintf(tdup, "Duplicate as above to %s", addr); | ||
213 | printk(KERN_INFO " " | ||
214 | " --- %s ---\n", tdup); | ||
215 | } | ||
216 | printk(KERN_INFO " %s (+%s) : %s [%s]\n", | ||
217 | addr, boff, bhex, basc); | ||
218 | } else { | ||
219 | if (dup >= 1) { | ||
220 | sprintf(tdup, "Duplicate as above to %s", addr); | ||
221 | printk(KERN_INFO " " | ||
222 | " --- %s ---\n", tdup); | ||
223 | } | ||
224 | if (dup != 0) { | ||
225 | printk(KERN_INFO " %s (+%s) : %s [%s]\n", | ||
226 | addr, boff, bhex, basc); | ||
227 | } | ||
228 | } | ||
229 | |||
230 | return; | ||
231 | |||
232 | } /* end of ctcmpc_dumpit */ | ||
233 | #endif | ||
234 | |||
235 | #ifdef DEBUGDATA | ||
236 | /* | ||
237 | * Dump header and first 16 bytes of an sk_buff for debugging purposes. | ||
238 | * | ||
239 | * skb The sk_buff to dump. | ||
240 | * offset Offset relative to skb-data, where to start the dump. | ||
241 | */ | ||
242 | void ctcmpc_dump_skb(struct sk_buff *skb, int offset) | ||
243 | { | ||
244 | unsigned char *p = skb->data; | ||
245 | struct th_header *header; | ||
246 | struct pdu *pheader; | ||
247 | int bl = skb->len; | ||
248 | int i; | ||
249 | |||
250 | if (p == NULL) | ||
251 | return; | ||
252 | |||
253 | p += offset; | ||
254 | header = (struct th_header *)p; | ||
255 | |||
256 | printk(KERN_INFO "dump:\n"); | ||
257 | printk(KERN_INFO "skb len=%d \n", skb->len); | ||
258 | if (skb->len > 2) { | ||
259 | switch (header->th_ch_flag) { | ||
260 | case TH_HAS_PDU: | ||
261 | break; | ||
262 | case 0x00: | ||
263 | case TH_IS_XID: | ||
264 | if ((header->th_blk_flag == TH_DATA_IS_XID) && | ||
265 | (header->th_is_xid == 0x01)) | ||
266 | goto dumpth; | ||
267 | case TH_SWEEP_REQ: | ||
268 | goto dumpth; | ||
269 | case TH_SWEEP_RESP: | ||
270 | goto dumpth; | ||
271 | default: | ||
272 | break; | ||
273 | } | ||
274 | |||
275 | pheader = (struct pdu *)p; | ||
276 | printk(KERN_INFO "pdu->offset: %d hex: %04x\n", | ||
277 | pheader->pdu_offset, pheader->pdu_offset); | ||
278 | printk(KERN_INFO "pdu->flag : %02x\n", pheader->pdu_flag); | ||
279 | printk(KERN_INFO "pdu->proto : %02x\n", pheader->pdu_proto); | ||
280 | printk(KERN_INFO "pdu->seq : %02x\n", pheader->pdu_seq); | ||
281 | goto dumpdata; | ||
282 | |||
283 | dumpth: | ||
284 | printk(KERN_INFO "th->seg : %02x\n", header->th_seg); | ||
285 | printk(KERN_INFO "th->ch : %02x\n", header->th_ch_flag); | ||
286 | printk(KERN_INFO "th->blk_flag: %02x\n", header->th_blk_flag); | ||
287 | printk(KERN_INFO "th->type : %s\n", | ||
288 | (header->th_is_xid) ? "DATA" : "XID"); | ||
289 | printk(KERN_INFO "th->seqnum : %04x\n", header->th_seq_num); | ||
290 | |||
291 | } | ||
292 | dumpdata: | ||
293 | if (bl > 32) | ||
294 | bl = 32; | ||
295 | printk(KERN_INFO "data: "); | ||
296 | for (i = 0; i < bl; i++) | ||
297 | printk(KERN_INFO "%02x%s", *p++, (i % 16) ? " " : "\n<7>"); | ||
298 | printk(KERN_INFO "\n"); | ||
299 | } | ||
300 | #endif | ||
301 | |||
302 | /* | ||
303 | * ctc_mpc_alloc_channel | ||
304 | * (exported interface) | ||
305 | * | ||
306 | * Device Initialization : | ||
307 | * ACTPATH driven IO operations | ||
308 | */ | ||
309 | int ctc_mpc_alloc_channel(int port_num, void (*callback)(int, int)) | ||
310 | { | ||
311 | char device[20]; | ||
312 | struct net_device *dev; | ||
313 | struct mpc_group *grp; | ||
314 | struct ctcm_priv *priv; | ||
315 | |||
316 | ctcm_pr_debug("ctcmpc enter: %s()\n", __FUNCTION__); | ||
317 | |||
318 | sprintf(device, "%s%i", MPC_DEVICE_NAME, port_num); | ||
319 | dev = __dev_get_by_name(&init_net, device); | ||
320 | |||
321 | if (dev == NULL) { | ||
322 | printk(KERN_INFO "ctc_mpc_alloc_channel %s dev=NULL\n", device); | ||
323 | return 1; | ||
324 | } | ||
325 | |||
326 | priv = dev->priv; | ||
327 | grp = priv->mpcg; | ||
328 | if (!grp) | ||
329 | return 1; | ||
330 | |||
331 | grp->allochanfunc = callback; | ||
332 | grp->port_num = port_num; | ||
333 | grp->port_persist = 1; | ||
334 | |||
335 | ctcm_pr_debug("ctcmpc: %s called for device %s state=%s\n", | ||
336 | __FUNCTION__, | ||
337 | dev->name, | ||
338 | fsm_getstate_str(grp->fsm)); | ||
339 | |||
340 | switch (fsm_getstate(grp->fsm)) { | ||
341 | case MPCG_STATE_INOP: | ||
342 | /* Group is in the process of terminating */ | ||
343 | grp->alloc_called = 1; | ||
344 | break; | ||
345 | case MPCG_STATE_RESET: | ||
346 | /* MPC Group will transition to state */ | ||
347 | /* MPCG_STATE_XID2INITW iff the minimum number */ | ||
348 | /* of 1 read and 1 write channel have successfully*/ | ||
349 | /* activated */ | ||
350 | /*fsm_newstate(grp->fsm, MPCG_STATE_XID2INITW);*/ | ||
351 | if (callback) | ||
352 | grp->send_qllc_disc = 1; | ||
353 | case MPCG_STATE_XID0IOWAIT: | ||
354 | fsm_deltimer(&grp->timer); | ||
355 | grp->outstanding_xid2 = 0; | ||
356 | grp->outstanding_xid7 = 0; | ||
357 | grp->outstanding_xid7_p2 = 0; | ||
358 | grp->saved_xid2 = NULL; | ||
359 | if (callback) | ||
360 | ctcm_open(dev); | ||
361 | fsm_event(priv->fsm, DEV_EVENT_START, dev); | ||
362 | break; | ||
363 | case MPCG_STATE_READY: | ||
364 | /* XID exchanges completed after PORT was activated */ | ||
365 | /* Link station already active */ | ||
366 | /* Maybe timing issue...retry callback */ | ||
367 | grp->allocchan_callback_retries++; | ||
368 | if (grp->allocchan_callback_retries < 4) { | ||
369 | if (grp->allochanfunc) | ||
370 | grp->allochanfunc(grp->port_num, | ||
371 | grp->group_max_buflen); | ||
372 | } else { | ||
373 | /* there are problems...bail out */ | ||
374 | /* there may be a state mismatch so restart */ | ||
375 | grp->port_persist = 1; | ||
376 | fsm_event(grp->fsm, MPCG_EVENT_INOP, dev); | ||
377 | grp->allocchan_callback_retries = 0; | ||
378 | } | ||
379 | break; | ||
380 | default: | ||
381 | return 0; | ||
382 | |||
383 | } | ||
384 | |||
385 | ctcm_pr_debug("ctcmpc exit: %s()\n", __FUNCTION__); | ||
386 | return 0; | ||
387 | } | ||
388 | EXPORT_SYMBOL(ctc_mpc_alloc_channel); | ||
389 | |||
390 | /* | ||
391 | * ctc_mpc_establish_connectivity | ||
392 | * (exported interface) | ||
393 | */ | ||
394 | void ctc_mpc_establish_connectivity(int port_num, | ||
395 | void (*callback)(int, int, int)) | ||
396 | { | ||
397 | char device[20]; | ||
398 | struct net_device *dev; | ||
399 | struct mpc_group *grp; | ||
400 | struct ctcm_priv *priv; | ||
401 | struct channel *rch, *wch; | ||
402 | |||
403 | ctcm_pr_debug("ctcmpc enter: %s()\n", __FUNCTION__); | ||
404 | |||
405 | sprintf(device, "%s%i", MPC_DEVICE_NAME, port_num); | ||
406 | dev = __dev_get_by_name(&init_net, device); | ||
407 | |||
408 | if (dev == NULL) { | ||
409 | printk(KERN_INFO "ctc_mpc_establish_connectivity " | ||
410 | "%s dev=NULL\n", device); | ||
411 | return; | ||
412 | } | ||
413 | priv = dev->priv; | ||
414 | rch = priv->channel[READ]; | ||
415 | wch = priv->channel[WRITE]; | ||
416 | |||
417 | grp = priv->mpcg; | ||
418 | |||
419 | ctcm_pr_debug("ctcmpc: %s() called for device %s state=%s\n", | ||
420 | __FUNCTION__, dev->name, | ||
421 | fsm_getstate_str(grp->fsm)); | ||
422 | |||
423 | grp->estconnfunc = callback; | ||
424 | grp->port_num = port_num; | ||
425 | |||
426 | switch (fsm_getstate(grp->fsm)) { | ||
427 | case MPCG_STATE_READY: | ||
428 | /* XID exchanges completed after PORT was activated */ | ||
429 | /* Link station already active */ | ||
430 | /* Maybe timing issue...retry callback */ | ||
431 | fsm_deltimer(&grp->timer); | ||
432 | grp->estconn_callback_retries++; | ||
433 | if (grp->estconn_callback_retries < 4) { | ||
434 | if (grp->estconnfunc) { | ||
435 | grp->estconnfunc(grp->port_num, 0, | ||
436 | grp->group_max_buflen); | ||
437 | grp->estconnfunc = NULL; | ||
438 | } | ||
439 | } else { | ||
440 | /* there are problems...bail out */ | ||
441 | fsm_event(grp->fsm, MPCG_EVENT_INOP, dev); | ||
442 | grp->estconn_callback_retries = 0; | ||
443 | } | ||
444 | break; | ||
445 | case MPCG_STATE_INOP: | ||
446 | case MPCG_STATE_RESET: | ||
447 | /* MPC Group is not ready to start XID - min num of */ | ||
448 | /* 1 read and 1 write channel have not been acquired*/ | ||
449 | printk(KERN_WARNING "ctcmpc: %s() REJECTED ACTIVE XID Req" | ||
450 | "uest - Channel Pair is not Active\n", __FUNCTION__); | ||
451 | if (grp->estconnfunc) { | ||
452 | grp->estconnfunc(grp->port_num, -1, 0); | ||
453 | grp->estconnfunc = NULL; | ||
454 | } | ||
455 | break; | ||
456 | case MPCG_STATE_XID2INITW: | ||
457 | /* alloc channel was called but no XID exchange */ | ||
458 | /* has occurred. initiate xside XID exchange */ | ||
459 | /* make sure yside XID0 processing has not started */ | ||
460 | if ((fsm_getstate(rch->fsm) > CH_XID0_PENDING) || | ||
461 | (fsm_getstate(wch->fsm) > CH_XID0_PENDING)) { | ||
462 | printk(KERN_WARNING "mpc: %s() ABORT ACTIVE XID" | ||
463 | " Request- PASSIVE XID in process\n" | ||
464 | , __FUNCTION__); | ||
465 | break; | ||
466 | } | ||
467 | grp->send_qllc_disc = 1; | ||
468 | fsm_newstate(grp->fsm, MPCG_STATE_XID0IOWAIT); | ||
469 | fsm_deltimer(&grp->timer); | ||
470 | fsm_addtimer(&grp->timer, MPC_XID_TIMEOUT_VALUE, | ||
471 | MPCG_EVENT_TIMER, dev); | ||
472 | grp->outstanding_xid7 = 0; | ||
473 | grp->outstanding_xid7_p2 = 0; | ||
474 | grp->saved_xid2 = NULL; | ||
475 | if ((rch->in_mpcgroup) && | ||
476 | (fsm_getstate(rch->fsm) == CH_XID0_PENDING)) | ||
477 | fsm_event(grp->fsm, MPCG_EVENT_XID0DO, rch); | ||
478 | else { | ||
479 | printk(KERN_WARNING "mpc: %s() Unable to start" | ||
480 | " ACTIVE XID0 on read channel\n", | ||
481 | __FUNCTION__); | ||
482 | if (grp->estconnfunc) { | ||
483 | grp->estconnfunc(grp->port_num, -1, 0); | ||
484 | grp->estconnfunc = NULL; | ||
485 | } | ||
486 | fsm_deltimer(&grp->timer); | ||
487 | goto done; | ||
488 | } | ||
489 | if ((wch->in_mpcgroup) && | ||
490 | (fsm_getstate(wch->fsm) == CH_XID0_PENDING)) | ||
491 | fsm_event(grp->fsm, MPCG_EVENT_XID0DO, wch); | ||
492 | else { | ||
493 | printk(KERN_WARNING "mpc: %s() Unable to start" | ||
494 | " ACTIVE XID0 on write channel\n", | ||
495 | __FUNCTION__); | ||
496 | if (grp->estconnfunc) { | ||
497 | grp->estconnfunc(grp->port_num, -1, 0); | ||
498 | grp->estconnfunc = NULL; | ||
499 | } | ||
500 | fsm_deltimer(&grp->timer); | ||
501 | goto done; | ||
502 | } | ||
503 | break; | ||
504 | case MPCG_STATE_XID0IOWAIT: | ||
505 | /* already in active XID negotiations */ | ||
506 | default: | ||
507 | break; | ||
508 | } | ||
509 | |||
510 | done: | ||
511 | ctcm_pr_debug("ctcmpc exit: %s()\n", __FUNCTION__); | ||
512 | return; | ||
513 | } | ||
514 | EXPORT_SYMBOL(ctc_mpc_establish_connectivity); | ||
515 | |||
516 | /* | ||
517 | * ctc_mpc_dealloc_ch | ||
518 | * (exported interface) | ||
519 | */ | ||
520 | void ctc_mpc_dealloc_ch(int port_num) | ||
521 | { | ||
522 | struct net_device *dev; | ||
523 | char device[20]; | ||
524 | struct ctcm_priv *priv; | ||
525 | struct mpc_group *grp; | ||
526 | |||
527 | ctcm_pr_debug("ctcmpc enter: %s()\n", __FUNCTION__); | ||
528 | sprintf(device, "%s%i", MPC_DEVICE_NAME, port_num); | ||
529 | dev = __dev_get_by_name(&init_net, device); | ||
530 | |||
531 | if (dev == NULL) { | ||
532 | printk(KERN_INFO "%s() %s dev=NULL\n", __FUNCTION__, device); | ||
533 | goto done; | ||
534 | } | ||
535 | |||
536 | ctcm_pr_debug("ctcmpc:%s %s() called for device %s refcount=%d\n", | ||
537 | dev->name, __FUNCTION__, | ||
538 | dev->name, atomic_read(&dev->refcnt)); | ||
539 | |||
540 | priv = dev->priv; | ||
541 | if (priv == NULL) { | ||
542 | printk(KERN_INFO "%s() %s priv=NULL\n", | ||
543 | __FUNCTION__, device); | ||
544 | goto done; | ||
545 | } | ||
546 | fsm_deltimer(&priv->restart_timer); | ||
547 | |||
548 | grp = priv->mpcg; | ||
549 | if (grp == NULL) { | ||
550 | printk(KERN_INFO "%s() %s dev=NULL\n", __FUNCTION__, device); | ||
551 | goto done; | ||
552 | } | ||
553 | grp->channels_terminating = 0; | ||
554 | |||
555 | fsm_deltimer(&grp->timer); | ||
556 | |||
557 | grp->allochanfunc = NULL; | ||
558 | grp->estconnfunc = NULL; | ||
559 | grp->port_persist = 0; | ||
560 | grp->send_qllc_disc = 0; | ||
561 | fsm_event(grp->fsm, MPCG_EVENT_INOP, dev); | ||
562 | |||
563 | ctcm_close(dev); | ||
564 | done: | ||
565 | ctcm_pr_debug("ctcmpc exit: %s()\n", __FUNCTION__); | ||
566 | return; | ||
567 | } | ||
568 | EXPORT_SYMBOL(ctc_mpc_dealloc_ch); | ||
569 | |||
570 | /* | ||
571 | * ctc_mpc_flow_control | ||
572 | * (exported interface) | ||
573 | */ | ||
574 | void ctc_mpc_flow_control(int port_num, int flowc) | ||
575 | { | ||
576 | char device[20]; | ||
577 | struct ctcm_priv *priv; | ||
578 | struct mpc_group *grp; | ||
579 | struct net_device *dev; | ||
580 | struct channel *rch; | ||
581 | int mpcg_state; | ||
582 | |||
583 | ctcm_pr_debug("ctcmpc enter: %s() %i\n", __FUNCTION__, flowc); | ||
584 | |||
585 | sprintf(device, "%s%i", MPC_DEVICE_NAME, port_num); | ||
586 | dev = __dev_get_by_name(&init_net, device); | ||
587 | |||
588 | if (dev == NULL) { | ||
589 | printk(KERN_INFO "ctc_mpc_flow_control %s dev=NULL\n", device); | ||
590 | return; | ||
591 | } | ||
592 | |||
593 | ctcm_pr_debug("ctcmpc: %s %s called \n", dev->name, __FUNCTION__); | ||
594 | |||
595 | priv = dev->priv; | ||
596 | if (priv == NULL) { | ||
597 | printk(KERN_INFO "ctcmpc:%s() %s priv=NULL\n", | ||
598 | __FUNCTION__, device); | ||
599 | return; | ||
600 | } | ||
601 | grp = priv->mpcg; | ||
602 | rch = priv->channel[READ]; | ||
603 | |||
604 | mpcg_state = fsm_getstate(grp->fsm); | ||
605 | switch (flowc) { | ||
606 | case 1: | ||
607 | if (mpcg_state == MPCG_STATE_FLOWC) | ||
608 | break; | ||
609 | if (mpcg_state == MPCG_STATE_READY) { | ||
610 | if (grp->flow_off_called == 1) | ||
611 | grp->flow_off_called = 0; | ||
612 | else | ||
613 | fsm_newstate(grp->fsm, MPCG_STATE_FLOWC); | ||
614 | break; | ||
615 | } | ||
616 | break; | ||
617 | case 0: | ||
618 | if (mpcg_state == MPCG_STATE_FLOWC) { | ||
619 | fsm_newstate(grp->fsm, MPCG_STATE_READY); | ||
620 | /* ensure any data that has accumulated */ | ||
621 | /* on the io_queue will now be sen t */ | ||
622 | tasklet_schedule(&rch->ch_tasklet); | ||
623 | } | ||
624 | /* possible race condition */ | ||
625 | if (mpcg_state == MPCG_STATE_READY) { | ||
626 | grp->flow_off_called = 1; | ||
627 | break; | ||
628 | } | ||
629 | break; | ||
630 | } | ||
631 | |||
632 | ctcm_pr_debug("ctcmpc exit: %s() %i\n", __FUNCTION__, flowc); | ||
633 | } | ||
634 | EXPORT_SYMBOL(ctc_mpc_flow_control); | ||
635 | |||
636 | static int mpc_send_qllc_discontact(struct net_device *); | ||
637 | |||
638 | /* | ||
639 | * helper function of ctcmpc_unpack_skb | ||
640 | */ | ||
641 | static void mpc_rcvd_sweep_resp(struct mpcg_info *mpcginfo) | ||
642 | { | ||
643 | struct channel *rch = mpcginfo->ch; | ||
644 | struct net_device *dev = rch->netdev; | ||
645 | struct ctcm_priv *priv = dev->priv; | ||
646 | struct mpc_group *grp = priv->mpcg; | ||
647 | struct channel *ch = priv->channel[WRITE]; | ||
648 | |||
649 | if (do_debug) | ||
650 | ctcm_pr_debug("ctcmpc enter: %s(): ch=0x%p id=%s\n", | ||
651 | __FUNCTION__, ch, ch->id); | ||
652 | |||
653 | if (do_debug_data) | ||
654 | ctcmpc_dumpit((char *)mpcginfo->sweep, TH_SWEEP_LENGTH); | ||
655 | |||
656 | grp->sweep_rsp_pend_num--; | ||
657 | |||
658 | if ((grp->sweep_req_pend_num == 0) && | ||
659 | (grp->sweep_rsp_pend_num == 0)) { | ||
660 | fsm_deltimer(&ch->sweep_timer); | ||
661 | grp->in_sweep = 0; | ||
662 | rch->th_seq_num = 0x00; | ||
663 | ch->th_seq_num = 0x00; | ||
664 | ctcm_clear_busy_do(dev); | ||
665 | } | ||
666 | |||
667 | kfree(mpcginfo); | ||
668 | |||
669 | return; | ||
670 | |||
671 | } | ||
672 | |||
673 | /* | ||
674 | * helper function of mpc_rcvd_sweep_req | ||
675 | * which is a helper of ctcmpc_unpack_skb | ||
676 | */ | ||
677 | static void ctcmpc_send_sweep_resp(struct channel *rch) | ||
678 | { | ||
679 | struct net_device *dev = rch->netdev; | ||
680 | struct ctcm_priv *priv = dev->priv; | ||
681 | struct mpc_group *grp = priv->mpcg; | ||
682 | int rc = 0; | ||
683 | struct th_sweep *header; | ||
684 | struct sk_buff *sweep_skb; | ||
685 | struct channel *ch = priv->channel[WRITE]; | ||
686 | |||
687 | if (do_debug) | ||
688 | ctcm_pr_debug("ctcmpc exit : %s(): ch=0x%p id=%s\n", | ||
689 | __FUNCTION__, rch, rch->id); | ||
690 | |||
691 | sweep_skb = __dev_alloc_skb(MPC_BUFSIZE_DEFAULT, | ||
692 | GFP_ATOMIC|GFP_DMA); | ||
693 | if (sweep_skb == NULL) { | ||
694 | printk(KERN_INFO "Couldn't alloc sweep_skb\n"); | ||
695 | rc = -ENOMEM; | ||
696 | goto done; | ||
697 | } | ||
698 | |||
699 | header = (struct th_sweep *) | ||
700 | kmalloc(sizeof(struct th_sweep), gfp_type()); | ||
701 | |||
702 | if (!header) { | ||
703 | dev_kfree_skb_any(sweep_skb); | ||
704 | rc = -ENOMEM; | ||
705 | goto done; | ||
706 | } | ||
707 | |||
708 | header->th.th_seg = 0x00 ; | ||
709 | header->th.th_ch_flag = TH_SWEEP_RESP; | ||
710 | header->th.th_blk_flag = 0x00; | ||
711 | header->th.th_is_xid = 0x00; | ||
712 | header->th.th_seq_num = 0x00; | ||
713 | header->sw.th_last_seq = ch->th_seq_num; | ||
714 | |||
715 | memcpy(skb_put(sweep_skb, TH_SWEEP_LENGTH), header, TH_SWEEP_LENGTH); | ||
716 | |||
717 | kfree(header); | ||
718 | |||
719 | dev->trans_start = jiffies; | ||
720 | skb_queue_tail(&ch->sweep_queue, sweep_skb); | ||
721 | |||
722 | fsm_addtimer(&ch->sweep_timer, 100, CTC_EVENT_RSWEEP_TIMER, ch); | ||
723 | |||
724 | return; | ||
725 | |||
726 | done: | ||
727 | if (rc != 0) { | ||
728 | grp->in_sweep = 0; | ||
729 | ctcm_clear_busy_do(dev); | ||
730 | fsm_event(grp->fsm, MPCG_EVENT_INOP, dev); | ||
731 | } | ||
732 | |||
733 | return; | ||
734 | } | ||
735 | |||
736 | /* | ||
737 | * helper function of ctcmpc_unpack_skb | ||
738 | */ | ||
739 | static void mpc_rcvd_sweep_req(struct mpcg_info *mpcginfo) | ||
740 | { | ||
741 | struct channel *rch = mpcginfo->ch; | ||
742 | struct net_device *dev = rch->netdev; | ||
743 | struct ctcm_priv *priv = dev->priv; | ||
744 | struct mpc_group *grp = priv->mpcg; | ||
745 | struct channel *ch = priv->channel[WRITE]; | ||
746 | |||
747 | if (do_debug) | ||
748 | CTCM_DBF_TEXT_(MPC_TRACE, CTC_DBF_DEBUG, | ||
749 | " %s(): ch=0x%p id=%s\n", __FUNCTION__, ch, ch->id); | ||
750 | |||
751 | if (grp->in_sweep == 0) { | ||
752 | grp->in_sweep = 1; | ||
753 | ctcm_test_and_set_busy(dev); | ||
754 | grp->sweep_req_pend_num = grp->active_channels[READ]; | ||
755 | grp->sweep_rsp_pend_num = grp->active_channels[READ]; | ||
756 | } | ||
757 | |||
758 | if (do_debug_data) | ||
759 | ctcmpc_dumpit((char *)mpcginfo->sweep, TH_SWEEP_LENGTH); | ||
760 | |||
761 | grp->sweep_req_pend_num--; | ||
762 | ctcmpc_send_sweep_resp(ch); | ||
763 | kfree(mpcginfo); | ||
764 | return; | ||
765 | } | ||
766 | |||
767 | /* | ||
768 | * MPC Group Station FSM definitions | ||
769 | */ | ||
770 | static const char *mpcg_event_names[] = { | ||
771 | [MPCG_EVENT_INOP] = "INOP Condition", | ||
772 | [MPCG_EVENT_DISCONC] = "Discontact Received", | ||
773 | [MPCG_EVENT_XID0DO] = "Channel Active - Start XID", | ||
774 | [MPCG_EVENT_XID2] = "XID2 Received", | ||
775 | [MPCG_EVENT_XID2DONE] = "XID0 Complete", | ||
776 | [MPCG_EVENT_XID7DONE] = "XID7 Complete", | ||
777 | [MPCG_EVENT_TIMER] = "XID Setup Timer", | ||
778 | [MPCG_EVENT_DOIO] = "XID DoIO", | ||
779 | }; | ||
780 | |||
781 | static const char *mpcg_state_names[] = { | ||
782 | [MPCG_STATE_RESET] = "Reset", | ||
783 | [MPCG_STATE_INOP] = "INOP", | ||
784 | [MPCG_STATE_XID2INITW] = "Passive XID- XID0 Pending Start", | ||
785 | [MPCG_STATE_XID2INITX] = "Passive XID- XID0 Pending Complete", | ||
786 | [MPCG_STATE_XID7INITW] = "Passive XID- XID7 Pending P1 Start", | ||
787 | [MPCG_STATE_XID7INITX] = "Passive XID- XID7 Pending P2 Complete", | ||
788 | [MPCG_STATE_XID0IOWAIT] = "Active XID- XID0 Pending Start", | ||
789 | [MPCG_STATE_XID0IOWAIX] = "Active XID- XID0 Pending Complete", | ||
790 | [MPCG_STATE_XID7INITI] = "Active XID- XID7 Pending Start", | ||
791 | [MPCG_STATE_XID7INITZ] = "Active XID- XID7 Pending Complete ", | ||
792 | [MPCG_STATE_XID7INITF] = "XID - XID7 Complete ", | ||
793 | [MPCG_STATE_FLOWC] = "FLOW CONTROL ON", | ||
794 | [MPCG_STATE_READY] = "READY", | ||
795 | }; | ||
796 | |||
797 | /* | ||
798 | * The MPC Group Station FSM | ||
799 | * 22 events | ||
800 | */ | ||
801 | static const fsm_node mpcg_fsm[] = { | ||
802 | { MPCG_STATE_RESET, MPCG_EVENT_INOP, mpc_action_go_inop }, | ||
803 | { MPCG_STATE_INOP, MPCG_EVENT_INOP, mpc_action_nop }, | ||
804 | { MPCG_STATE_FLOWC, MPCG_EVENT_INOP, mpc_action_go_inop }, | ||
805 | |||
806 | { MPCG_STATE_READY, MPCG_EVENT_DISCONC, mpc_action_discontact }, | ||
807 | { MPCG_STATE_READY, MPCG_EVENT_INOP, mpc_action_go_inop }, | ||
808 | |||
809 | { MPCG_STATE_XID2INITW, MPCG_EVENT_XID0DO, mpc_action_doxid0 }, | ||
810 | { MPCG_STATE_XID2INITW, MPCG_EVENT_XID2, mpc_action_rcvd_xid0 }, | ||
811 | { MPCG_STATE_XID2INITW, MPCG_EVENT_INOP, mpc_action_go_inop }, | ||
812 | { MPCG_STATE_XID2INITW, MPCG_EVENT_TIMER, mpc_action_timeout }, | ||
813 | { MPCG_STATE_XID2INITW, MPCG_EVENT_DOIO, mpc_action_yside_xid }, | ||
814 | |||
815 | { MPCG_STATE_XID2INITX, MPCG_EVENT_XID0DO, mpc_action_doxid0 }, | ||
816 | { MPCG_STATE_XID2INITX, MPCG_EVENT_XID2, mpc_action_rcvd_xid0 }, | ||
817 | { MPCG_STATE_XID2INITX, MPCG_EVENT_INOP, mpc_action_go_inop }, | ||
818 | { MPCG_STATE_XID2INITX, MPCG_EVENT_TIMER, mpc_action_timeout }, | ||
819 | { MPCG_STATE_XID2INITX, MPCG_EVENT_DOIO, mpc_action_yside_xid }, | ||
820 | |||
821 | { MPCG_STATE_XID7INITW, MPCG_EVENT_XID2DONE, mpc_action_doxid7 }, | ||
822 | { MPCG_STATE_XID7INITW, MPCG_EVENT_DISCONC, mpc_action_discontact }, | ||
823 | { MPCG_STATE_XID7INITW, MPCG_EVENT_XID2, mpc_action_rcvd_xid7 }, | ||
824 | { MPCG_STATE_XID7INITW, MPCG_EVENT_INOP, mpc_action_go_inop }, | ||
825 | { MPCG_STATE_XID7INITW, MPCG_EVENT_TIMER, mpc_action_timeout }, | ||
826 | { MPCG_STATE_XID7INITW, MPCG_EVENT_XID7DONE, mpc_action_doxid7 }, | ||
827 | { MPCG_STATE_XID7INITW, MPCG_EVENT_DOIO, mpc_action_yside_xid }, | ||
828 | |||
829 | { MPCG_STATE_XID7INITX, MPCG_EVENT_DISCONC, mpc_action_discontact }, | ||
830 | { MPCG_STATE_XID7INITX, MPCG_EVENT_XID2, mpc_action_rcvd_xid7 }, | ||
831 | { MPCG_STATE_XID7INITX, MPCG_EVENT_INOP, mpc_action_go_inop }, | ||
832 | { MPCG_STATE_XID7INITX, MPCG_EVENT_XID7DONE, mpc_action_doxid7 }, | ||
833 | { MPCG_STATE_XID7INITX, MPCG_EVENT_TIMER, mpc_action_timeout }, | ||
834 | { MPCG_STATE_XID7INITX, MPCG_EVENT_DOIO, mpc_action_yside_xid }, | ||
835 | |||
836 | { MPCG_STATE_XID0IOWAIT, MPCG_EVENT_XID0DO, mpc_action_doxid0 }, | ||
837 | { MPCG_STATE_XID0IOWAIT, MPCG_EVENT_DISCONC, mpc_action_discontact }, | ||
838 | { MPCG_STATE_XID0IOWAIT, MPCG_EVENT_XID2, mpc_action_rcvd_xid0 }, | ||
839 | { MPCG_STATE_XID0IOWAIT, MPCG_EVENT_INOP, mpc_action_go_inop }, | ||
840 | { MPCG_STATE_XID0IOWAIT, MPCG_EVENT_TIMER, mpc_action_timeout }, | ||
841 | { MPCG_STATE_XID0IOWAIT, MPCG_EVENT_DOIO, mpc_action_xside_xid }, | ||
842 | |||
843 | { MPCG_STATE_XID0IOWAIX, MPCG_EVENT_XID0DO, mpc_action_doxid0 }, | ||
844 | { MPCG_STATE_XID0IOWAIX, MPCG_EVENT_DISCONC, mpc_action_discontact }, | ||
845 | { MPCG_STATE_XID0IOWAIX, MPCG_EVENT_XID2, mpc_action_rcvd_xid0 }, | ||
846 | { MPCG_STATE_XID0IOWAIX, MPCG_EVENT_INOP, mpc_action_go_inop }, | ||
847 | { MPCG_STATE_XID0IOWAIX, MPCG_EVENT_TIMER, mpc_action_timeout }, | ||
848 | { MPCG_STATE_XID0IOWAIX, MPCG_EVENT_DOIO, mpc_action_xside_xid }, | ||
849 | |||
850 | { MPCG_STATE_XID7INITI, MPCG_EVENT_XID2DONE, mpc_action_doxid7 }, | ||
851 | { MPCG_STATE_XID7INITI, MPCG_EVENT_XID2, mpc_action_rcvd_xid7 }, | ||
852 | { MPCG_STATE_XID7INITI, MPCG_EVENT_DISCONC, mpc_action_discontact }, | ||
853 | { MPCG_STATE_XID7INITI, MPCG_EVENT_INOP, mpc_action_go_inop }, | ||
854 | { MPCG_STATE_XID7INITI, MPCG_EVENT_TIMER, mpc_action_timeout }, | ||
855 | { MPCG_STATE_XID7INITI, MPCG_EVENT_XID7DONE, mpc_action_doxid7 }, | ||
856 | { MPCG_STATE_XID7INITI, MPCG_EVENT_DOIO, mpc_action_xside_xid }, | ||
857 | |||
858 | { MPCG_STATE_XID7INITZ, MPCG_EVENT_XID2, mpc_action_rcvd_xid7 }, | ||
859 | { MPCG_STATE_XID7INITZ, MPCG_EVENT_XID7DONE, mpc_action_doxid7 }, | ||
860 | { MPCG_STATE_XID7INITZ, MPCG_EVENT_DISCONC, mpc_action_discontact }, | ||
861 | { MPCG_STATE_XID7INITZ, MPCG_EVENT_INOP, mpc_action_go_inop }, | ||
862 | { MPCG_STATE_XID7INITZ, MPCG_EVENT_TIMER, mpc_action_timeout }, | ||
863 | { MPCG_STATE_XID7INITZ, MPCG_EVENT_DOIO, mpc_action_xside_xid }, | ||
864 | |||
865 | { MPCG_STATE_XID7INITF, MPCG_EVENT_INOP, mpc_action_go_inop }, | ||
866 | { MPCG_STATE_XID7INITF, MPCG_EVENT_XID7DONE, mpc_action_go_ready }, | ||
867 | }; | ||
868 | |||
869 | static int mpcg_fsm_len = ARRAY_SIZE(mpcg_fsm); | ||
870 | |||
871 | /* | ||
872 | * MPC Group Station FSM action | ||
873 | * CTCM_PROTO_MPC only | ||
874 | */ | ||
875 | static void mpc_action_go_ready(fsm_instance *fsm, int event, void *arg) | ||
876 | { | ||
877 | struct net_device *dev = arg; | ||
878 | struct ctcm_priv *priv = NULL; | ||
879 | struct mpc_group *grp = NULL; | ||
880 | |||
881 | if (dev == NULL) { | ||
882 | printk(KERN_INFO "%s() dev=NULL\n", __FUNCTION__); | ||
883 | return; | ||
884 | } | ||
885 | |||
886 | ctcm_pr_debug("ctcmpc enter: %s %s()\n", dev->name, __FUNCTION__); | ||
887 | |||
888 | priv = dev->priv; | ||
889 | if (priv == NULL) { | ||
890 | printk(KERN_INFO "%s() priv=NULL\n", __FUNCTION__); | ||
891 | return; | ||
892 | } | ||
893 | |||
894 | grp = priv->mpcg; | ||
895 | if (grp == NULL) { | ||
896 | printk(KERN_INFO "%s() grp=NULL\n", __FUNCTION__); | ||
897 | return; | ||
898 | } | ||
899 | |||
900 | fsm_deltimer(&grp->timer); | ||
901 | |||
902 | if (grp->saved_xid2->xid2_flag2 == 0x40) { | ||
903 | priv->xid->xid2_flag2 = 0x00; | ||
904 | if (grp->estconnfunc) { | ||
905 | grp->estconnfunc(grp->port_num, 1, | ||
906 | grp->group_max_buflen); | ||
907 | grp->estconnfunc = NULL; | ||
908 | } else if (grp->allochanfunc) | ||
909 | grp->send_qllc_disc = 1; | ||
910 | goto done; | ||
911 | } | ||
912 | |||
913 | grp->port_persist = 1; | ||
914 | grp->out_of_sequence = 0; | ||
915 | grp->estconn_called = 0; | ||
916 | |||
917 | tasklet_hi_schedule(&grp->mpc_tasklet2); | ||
918 | |||
919 | ctcm_pr_debug("ctcmpc exit: %s %s()\n", dev->name, __FUNCTION__); | ||
920 | return; | ||
921 | |||
922 | done: | ||
923 | fsm_event(grp->fsm, MPCG_EVENT_INOP, dev); | ||
924 | |||
925 | |||
926 | ctcm_pr_info("ctcmpc: %s()failure occurred\n", __FUNCTION__); | ||
927 | } | ||
928 | |||
929 | /* | ||
930 | * helper of ctcm_init_netdevice | ||
931 | * CTCM_PROTO_MPC only | ||
932 | */ | ||
933 | void mpc_group_ready(unsigned long adev) | ||
934 | { | ||
935 | struct net_device *dev = (struct net_device *)adev; | ||
936 | struct ctcm_priv *priv = NULL; | ||
937 | struct mpc_group *grp = NULL; | ||
938 | struct channel *ch = NULL; | ||
939 | |||
940 | |||
941 | ctcm_pr_debug("ctcmpc enter: %s()\n", __FUNCTION__); | ||
942 | |||
943 | if (dev == NULL) { | ||
944 | printk(KERN_INFO "%s() dev=NULL\n", __FUNCTION__); | ||
945 | return; | ||
946 | } | ||
947 | |||
948 | priv = dev->priv; | ||
949 | if (priv == NULL) { | ||
950 | printk(KERN_INFO "%s() priv=NULL\n", __FUNCTION__); | ||
951 | return; | ||
952 | } | ||
953 | |||
954 | grp = priv->mpcg; | ||
955 | if (grp == NULL) { | ||
956 | printk(KERN_INFO "ctcmpc:%s() grp=NULL\n", __FUNCTION__); | ||
957 | return; | ||
958 | } | ||
959 | |||
960 | printk(KERN_NOTICE "ctcmpc: %s GROUP TRANSITIONED TO READY" | ||
961 | " maxbuf:%d\n", | ||
962 | dev->name, grp->group_max_buflen); | ||
963 | |||
964 | fsm_newstate(grp->fsm, MPCG_STATE_READY); | ||
965 | |||
966 | /* Put up a read on the channel */ | ||
967 | ch = priv->channel[READ]; | ||
968 | ch->pdu_seq = 0; | ||
969 | if (do_debug_data) | ||
970 | ctcm_pr_debug("ctcmpc: %s() ToDCM_pdu_seq= %08x\n" , | ||
971 | __FUNCTION__, ch->pdu_seq); | ||
972 | |||
973 | ctcmpc_chx_rxidle(ch->fsm, CTC_EVENT_START, ch); | ||
974 | /* Put the write channel in idle state */ | ||
975 | ch = priv->channel[WRITE]; | ||
976 | if (ch->collect_len > 0) { | ||
977 | spin_lock(&ch->collect_lock); | ||
978 | ctcm_purge_skb_queue(&ch->collect_queue); | ||
979 | ch->collect_len = 0; | ||
980 | spin_unlock(&ch->collect_lock); | ||
981 | } | ||
982 | ctcm_chx_txidle(ch->fsm, CTC_EVENT_START, ch); | ||
983 | |||
984 | ctcm_clear_busy(dev); | ||
985 | |||
986 | if (grp->estconnfunc) { | ||
987 | grp->estconnfunc(grp->port_num, 0, | ||
988 | grp->group_max_buflen); | ||
989 | grp->estconnfunc = NULL; | ||
990 | } else | ||
991 | if (grp->allochanfunc) | ||
992 | grp->allochanfunc(grp->port_num, | ||
993 | grp->group_max_buflen); | ||
994 | |||
995 | grp->send_qllc_disc = 1; | ||
996 | grp->changed_side = 0; | ||
997 | |||
998 | ctcm_pr_debug("ctcmpc exit: %s()\n", __FUNCTION__); | ||
999 | return; | ||
1000 | |||
1001 | } | ||
1002 | |||
1003 | /* | ||
1004 | * Increment the MPC Group Active Channel Counts | ||
1005 | * helper of dev_action (called from channel fsm) | ||
1006 | */ | ||
1007 | int mpc_channel_action(struct channel *ch, int direction, int action) | ||
1008 | { | ||
1009 | struct net_device *dev = ch->netdev; | ||
1010 | struct ctcm_priv *priv; | ||
1011 | struct mpc_group *grp = NULL; | ||
1012 | int rc = 0; | ||
1013 | |||
1014 | if (do_debug) | ||
1015 | ctcm_pr_debug("ctcmpc enter: %s(): ch=0x%p id=%s\n", | ||
1016 | __FUNCTION__, ch, ch->id); | ||
1017 | |||
1018 | if (dev == NULL) { | ||
1019 | printk(KERN_INFO "ctcmpc_channel_action %i dev=NULL\n", | ||
1020 | action); | ||
1021 | rc = 1; | ||
1022 | goto done; | ||
1023 | } | ||
1024 | |||
1025 | priv = dev->priv; | ||
1026 | if (priv == NULL) { | ||
1027 | printk(KERN_INFO | ||
1028 | "ctcmpc_channel_action%i priv=NULL, dev=%s\n", | ||
1029 | action, dev->name); | ||
1030 | rc = 2; | ||
1031 | goto done; | ||
1032 | } | ||
1033 | |||
1034 | grp = priv->mpcg; | ||
1035 | |||
1036 | if (grp == NULL) { | ||
1037 | printk(KERN_INFO "ctcmpc: %s()%i mpcgroup=NULL, dev=%s\n", | ||
1038 | __FUNCTION__, action, dev->name); | ||
1039 | rc = 3; | ||
1040 | goto done; | ||
1041 | } | ||
1042 | |||
1043 | ctcm_pr_info( | ||
1044 | "ctcmpc: %s() %i(): Grp:%s total_channel_paths=%i " | ||
1045 | "active_channels read=%i, write=%i\n", | ||
1046 | __FUNCTION__, | ||
1047 | action, | ||
1048 | fsm_getstate_str(grp->fsm), | ||
1049 | grp->num_channel_paths, | ||
1050 | grp->active_channels[READ], | ||
1051 | grp->active_channels[WRITE]); | ||
1052 | |||
1053 | if ((action == MPC_CHANNEL_ADD) && (ch->in_mpcgroup == 0)) { | ||
1054 | grp->num_channel_paths++; | ||
1055 | grp->active_channels[direction]++; | ||
1056 | grp->outstanding_xid2++; | ||
1057 | ch->in_mpcgroup = 1; | ||
1058 | |||
1059 | if (ch->xid_skb != NULL) | ||
1060 | dev_kfree_skb_any(ch->xid_skb); | ||
1061 | |||
1062 | ch->xid_skb = __dev_alloc_skb(MPC_BUFSIZE_DEFAULT, | ||
1063 | GFP_ATOMIC | GFP_DMA); | ||
1064 | if (ch->xid_skb == NULL) { | ||
1065 | printk(KERN_INFO "ctcmpc: %s()" | ||
1066 | "Couldn't alloc ch xid_skb\n", __FUNCTION__); | ||
1067 | fsm_event(grp->fsm, MPCG_EVENT_INOP, dev); | ||
1068 | return 1; | ||
1069 | } | ||
1070 | ch->xid_skb_data = ch->xid_skb->data; | ||
1071 | ch->xid_th = (struct th_header *)ch->xid_skb->data; | ||
1072 | skb_put(ch->xid_skb, TH_HEADER_LENGTH); | ||
1073 | ch->xid = (struct xid2 *)skb_tail_pointer(ch->xid_skb); | ||
1074 | skb_put(ch->xid_skb, XID2_LENGTH); | ||
1075 | ch->xid_id = skb_tail_pointer(ch->xid_skb); | ||
1076 | ch->xid_skb->data = ch->xid_skb_data; | ||
1077 | skb_reset_tail_pointer(ch->xid_skb); | ||
1078 | ch->xid_skb->len = 0; | ||
1079 | |||
1080 | memcpy(skb_put(ch->xid_skb, grp->xid_skb->len), | ||
1081 | grp->xid_skb->data, | ||
1082 | grp->xid_skb->len); | ||
1083 | |||
1084 | ch->xid->xid2_dlc_type = ((CHANNEL_DIRECTION(ch->flags) == READ) | ||
1085 | ? XID2_READ_SIDE : XID2_WRITE_SIDE); | ||
1086 | |||
1087 | if (CHANNEL_DIRECTION(ch->flags) == WRITE) | ||
1088 | ch->xid->xid2_buf_len = 0x00; | ||
1089 | |||
1090 | ch->xid_skb->data = ch->xid_skb_data; | ||
1091 | skb_reset_tail_pointer(ch->xid_skb); | ||
1092 | ch->xid_skb->len = 0; | ||
1093 | |||
1094 | fsm_newstate(ch->fsm, CH_XID0_PENDING); | ||
1095 | |||
1096 | if ((grp->active_channels[READ] > 0) && | ||
1097 | (grp->active_channels[WRITE] > 0) && | ||
1098 | (fsm_getstate(grp->fsm) < MPCG_STATE_XID2INITW)) { | ||
1099 | fsm_newstate(grp->fsm, MPCG_STATE_XID2INITW); | ||
1100 | printk(KERN_NOTICE "ctcmpc: %s MPC GROUP " | ||
1101 | "CHANNELS ACTIVE\n", dev->name); | ||
1102 | } | ||
1103 | } else if ((action == MPC_CHANNEL_REMOVE) && | ||
1104 | (ch->in_mpcgroup == 1)) { | ||
1105 | ch->in_mpcgroup = 0; | ||
1106 | grp->num_channel_paths--; | ||
1107 | grp->active_channels[direction]--; | ||
1108 | |||
1109 | if (ch->xid_skb != NULL) | ||
1110 | dev_kfree_skb_any(ch->xid_skb); | ||
1111 | ch->xid_skb = NULL; | ||
1112 | |||
1113 | if (grp->channels_terminating) | ||
1114 | goto done; | ||
1115 | |||
1116 | if (((grp->active_channels[READ] == 0) && | ||
1117 | (grp->active_channels[WRITE] > 0)) | ||
1118 | || ((grp->active_channels[WRITE] == 0) && | ||
1119 | (grp->active_channels[READ] > 0))) | ||
1120 | fsm_event(grp->fsm, MPCG_EVENT_INOP, dev); | ||
1121 | } | ||
1122 | |||
1123 | done: | ||
1124 | |||
1125 | if (do_debug) { | ||
1126 | ctcm_pr_debug( | ||
1127 | "ctcmpc: %s() %i Grp:%s ttl_chan_paths=%i " | ||
1128 | "active_chans read=%i, write=%i\n", | ||
1129 | __FUNCTION__, | ||
1130 | action, | ||
1131 | fsm_getstate_str(grp->fsm), | ||
1132 | grp->num_channel_paths, | ||
1133 | grp->active_channels[READ], | ||
1134 | grp->active_channels[WRITE]); | ||
1135 | |||
1136 | ctcm_pr_debug("ctcmpc exit : %s(): ch=0x%p id=%s\n", | ||
1137 | __FUNCTION__, ch, ch->id); | ||
1138 | } | ||
1139 | return rc; | ||
1140 | |||
1141 | } | ||
1142 | |||
1143 | /** | ||
1144 | * Unpack a just received skb and hand it over to | ||
1145 | * upper layers. | ||
1146 | * special MPC version of unpack_skb. | ||
1147 | * | ||
1148 | * ch The channel where this skb has been received. | ||
1149 | * pskb The received skb. | ||
1150 | */ | ||
1151 | static void ctcmpc_unpack_skb(struct channel *ch, struct sk_buff *pskb) | ||
1152 | { | ||
1153 | struct net_device *dev = ch->netdev; | ||
1154 | struct ctcm_priv *priv = dev->priv; | ||
1155 | struct mpc_group *grp = priv->mpcg; | ||
1156 | struct pdu *curr_pdu; | ||
1157 | struct mpcg_info *mpcginfo; | ||
1158 | struct th_header *header = NULL; | ||
1159 | struct th_sweep *sweep = NULL; | ||
1160 | int pdu_last_seen = 0; | ||
1161 | __u32 new_len; | ||
1162 | struct sk_buff *skb; | ||
1163 | int skblen; | ||
1164 | int sendrc = 0; | ||
1165 | |||
1166 | if (do_debug) | ||
1167 | ctcm_pr_debug("ctcmpc enter: %s() %s cp:%i ch:%s\n", | ||
1168 | __FUNCTION__, dev->name, smp_processor_id(), ch->id); | ||
1169 | |||
1170 | header = (struct th_header *)pskb->data; | ||
1171 | if ((header->th_seg == 0) && | ||
1172 | (header->th_ch_flag == 0) && | ||
1173 | (header->th_blk_flag == 0) && | ||
1174 | (header->th_seq_num == 0)) | ||
1175 | /* nothing for us */ goto done; | ||
1176 | |||
1177 | if (do_debug_data) { | ||
1178 | ctcm_pr_debug("ctcmpc: %s() th_header\n", __FUNCTION__); | ||
1179 | ctcmpc_dumpit((char *)header, TH_HEADER_LENGTH); | ||
1180 | ctcm_pr_debug("ctcmpc: %s() pskb len: %04x \n", | ||
1181 | __FUNCTION__, pskb->len); | ||
1182 | } | ||
1183 | |||
1184 | pskb->dev = dev; | ||
1185 | pskb->ip_summed = CHECKSUM_UNNECESSARY; | ||
1186 | skb_pull(pskb, TH_HEADER_LENGTH); | ||
1187 | |||
1188 | if (likely(header->th_ch_flag == TH_HAS_PDU)) { | ||
1189 | if (do_debug_data) | ||
1190 | ctcm_pr_debug("ctcmpc: %s() came into th_has_pdu\n", | ||
1191 | __FUNCTION__); | ||
1192 | if ((fsm_getstate(grp->fsm) == MPCG_STATE_FLOWC) || | ||
1193 | ((fsm_getstate(grp->fsm) == MPCG_STATE_READY) && | ||
1194 | (header->th_seq_num != ch->th_seq_num + 1) && | ||
1195 | (ch->th_seq_num != 0))) { | ||
1196 | /* This is NOT the next segment * | ||
1197 | * we are not the correct race winner * | ||
1198 | * go away and let someone else win * | ||
1199 | * BUT..this only applies if xid negot * | ||
1200 | * is done * | ||
1201 | */ | ||
1202 | grp->out_of_sequence += 1; | ||
1203 | __skb_push(pskb, TH_HEADER_LENGTH); | ||
1204 | skb_queue_tail(&ch->io_queue, pskb); | ||
1205 | if (do_debug_data) | ||
1206 | ctcm_pr_debug("ctcmpc: %s() th_seq_num " | ||
1207 | "expect:%08x got:%08x\n", __FUNCTION__, | ||
1208 | ch->th_seq_num + 1, header->th_seq_num); | ||
1209 | |||
1210 | return; | ||
1211 | } | ||
1212 | grp->out_of_sequence = 0; | ||
1213 | ch->th_seq_num = header->th_seq_num; | ||
1214 | |||
1215 | if (do_debug_data) | ||
1216 | ctcm_pr_debug("ctcmpc: %s() FromVTAM_th_seq=%08x\n", | ||
1217 | __FUNCTION__, ch->th_seq_num); | ||
1218 | |||
1219 | if (unlikely(fsm_getstate(grp->fsm) != MPCG_STATE_READY)) | ||
1220 | goto done; | ||
1221 | pdu_last_seen = 0; | ||
1222 | while ((pskb->len > 0) && !pdu_last_seen) { | ||
1223 | curr_pdu = (struct pdu *)pskb->data; | ||
1224 | if (do_debug_data) { | ||
1225 | ctcm_pr_debug("ctcm: %s() pdu_header\n", | ||
1226 | __FUNCTION__); | ||
1227 | ctcmpc_dumpit((char *)pskb->data, | ||
1228 | PDU_HEADER_LENGTH); | ||
1229 | ctcm_pr_debug("ctcm: %s() pskb len: %04x \n", | ||
1230 | __FUNCTION__, pskb->len); | ||
1231 | } | ||
1232 | skb_pull(pskb, PDU_HEADER_LENGTH); | ||
1233 | |||
1234 | if (curr_pdu->pdu_flag & PDU_LAST) | ||
1235 | pdu_last_seen = 1; | ||
1236 | if (curr_pdu->pdu_flag & PDU_CNTL) | ||
1237 | pskb->protocol = htons(ETH_P_SNAP); | ||
1238 | else | ||
1239 | pskb->protocol = htons(ETH_P_SNA_DIX); | ||
1240 | |||
1241 | if ((pskb->len <= 0) || (pskb->len > ch->max_bufsize)) { | ||
1242 | printk(KERN_INFO | ||
1243 | "%s Illegal packet size %d " | ||
1244 | "received " | ||
1245 | "dropping\n", dev->name, | ||
1246 | pskb->len); | ||
1247 | priv->stats.rx_dropped++; | ||
1248 | priv->stats.rx_length_errors++; | ||
1249 | goto done; | ||
1250 | } | ||
1251 | skb_reset_mac_header(pskb); | ||
1252 | new_len = curr_pdu->pdu_offset; | ||
1253 | if (do_debug_data) | ||
1254 | ctcm_pr_debug("ctcmpc: %s() new_len: %04x \n", | ||
1255 | __FUNCTION__, new_len); | ||
1256 | if ((new_len == 0) || (new_len > pskb->len)) { | ||
1257 | /* should never happen */ | ||
1258 | /* pskb len must be hosed...bail out */ | ||
1259 | printk(KERN_INFO | ||
1260 | "ctcmpc: %s(): invalid pdu" | ||
1261 | " offset of %04x - data may be" | ||
1262 | "lost\n", __FUNCTION__, new_len); | ||
1263 | goto done; | ||
1264 | } | ||
1265 | skb = __dev_alloc_skb(new_len+4, GFP_ATOMIC); | ||
1266 | |||
1267 | if (!skb) { | ||
1268 | printk(KERN_INFO | ||
1269 | "ctcm: %s Out of memory in " | ||
1270 | "%s()- request-len:%04x \n", | ||
1271 | dev->name, | ||
1272 | __FUNCTION__, | ||
1273 | new_len+4); | ||
1274 | priv->stats.rx_dropped++; | ||
1275 | fsm_event(grp->fsm, | ||
1276 | MPCG_EVENT_INOP, dev); | ||
1277 | goto done; | ||
1278 | } | ||
1279 | |||
1280 | memcpy(skb_put(skb, new_len), | ||
1281 | pskb->data, new_len); | ||
1282 | |||
1283 | skb_reset_mac_header(skb); | ||
1284 | skb->dev = pskb->dev; | ||
1285 | skb->protocol = pskb->protocol; | ||
1286 | skb->ip_summed = CHECKSUM_UNNECESSARY; | ||
1287 | *((__u32 *) skb_push(skb, 4)) = ch->pdu_seq; | ||
1288 | ch->pdu_seq++; | ||
1289 | |||
1290 | if (do_debug_data) | ||
1291 | ctcm_pr_debug("%s: ToDCM_pdu_seq= %08x\n", | ||
1292 | __FUNCTION__, ch->pdu_seq); | ||
1293 | |||
1294 | ctcm_pr_debug("ctcm: %s() skb:%0lx " | ||
1295 | "skb len: %d \n", __FUNCTION__, | ||
1296 | (unsigned long)skb, skb->len); | ||
1297 | if (do_debug_data) { | ||
1298 | ctcm_pr_debug("ctcmpc: %s() up to 32 bytes" | ||
1299 | " of pdu_data sent\n", | ||
1300 | __FUNCTION__); | ||
1301 | ctcmpc_dump32((char *)skb->data, skb->len); | ||
1302 | } | ||
1303 | |||
1304 | skblen = skb->len; | ||
1305 | sendrc = netif_rx(skb); | ||
1306 | priv->stats.rx_packets++; | ||
1307 | priv->stats.rx_bytes += skblen; | ||
1308 | skb_pull(pskb, new_len); /* point to next PDU */ | ||
1309 | } | ||
1310 | } else { | ||
1311 | mpcginfo = (struct mpcg_info *) | ||
1312 | kmalloc(sizeof(struct mpcg_info), gfp_type()); | ||
1313 | if (mpcginfo == NULL) | ||
1314 | goto done; | ||
1315 | |||
1316 | mpcginfo->ch = ch; | ||
1317 | mpcginfo->th = header; | ||
1318 | mpcginfo->skb = pskb; | ||
1319 | ctcm_pr_debug("ctcmpc: %s() Not PDU - may be control pkt\n", | ||
1320 | __FUNCTION__); | ||
1321 | /* it's a sweep? */ | ||
1322 | sweep = (struct th_sweep *)pskb->data; | ||
1323 | mpcginfo->sweep = sweep; | ||
1324 | if (header->th_ch_flag == TH_SWEEP_REQ) | ||
1325 | mpc_rcvd_sweep_req(mpcginfo); | ||
1326 | else if (header->th_ch_flag == TH_SWEEP_RESP) | ||
1327 | mpc_rcvd_sweep_resp(mpcginfo); | ||
1328 | else if (header->th_blk_flag == TH_DATA_IS_XID) { | ||
1329 | struct xid2 *thisxid = (struct xid2 *)pskb->data; | ||
1330 | skb_pull(pskb, XID2_LENGTH); | ||
1331 | mpcginfo->xid = thisxid; | ||
1332 | fsm_event(grp->fsm, MPCG_EVENT_XID2, mpcginfo); | ||
1333 | } else if (header->th_blk_flag == TH_DISCONTACT) | ||
1334 | fsm_event(grp->fsm, MPCG_EVENT_DISCONC, mpcginfo); | ||
1335 | else if (header->th_seq_num != 0) { | ||
1336 | printk(KERN_INFO "%s unexpected packet" | ||
1337 | " expected control pkt\n", dev->name); | ||
1338 | priv->stats.rx_dropped++; | ||
1339 | /* mpcginfo only used for non-data transfers */ | ||
1340 | kfree(mpcginfo); | ||
1341 | if (do_debug_data) | ||
1342 | ctcmpc_dump_skb(pskb, -8); | ||
1343 | } | ||
1344 | } | ||
1345 | done: | ||
1346 | |||
1347 | dev_kfree_skb_any(pskb); | ||
1348 | if (sendrc == NET_RX_DROP) { | ||
1349 | printk(KERN_WARNING "%s %s() NETWORK BACKLOG EXCEEDED" | ||
1350 | " - PACKET DROPPED\n", dev->name, __FUNCTION__); | ||
1351 | fsm_event(grp->fsm, MPCG_EVENT_INOP, dev); | ||
1352 | } | ||
1353 | |||
1354 | if (do_debug) | ||
1355 | ctcm_pr_debug("ctcmpc exit : %s %s(): ch=0x%p id=%s\n", | ||
1356 | dev->name, __FUNCTION__, ch, ch->id); | ||
1357 | } | ||
1358 | |||
1359 | /** | ||
1360 | * tasklet helper for mpc's skb unpacking. | ||
1361 | * | ||
1362 | * ch The channel to work on. | ||
1363 | * Allow flow control back pressure to occur here. | ||
1364 | * Throttling back channel can result in excessive | ||
1365 | * channel inactivity and system deact of channel | ||
1366 | */ | ||
1367 | void ctcmpc_bh(unsigned long thischan) | ||
1368 | { | ||
1369 | struct channel *ch = (struct channel *)thischan; | ||
1370 | struct sk_buff *skb; | ||
1371 | struct net_device *dev = ch->netdev; | ||
1372 | struct ctcm_priv *priv = dev->priv; | ||
1373 | struct mpc_group *grp = priv->mpcg; | ||
1374 | |||
1375 | if (do_debug) | ||
1376 | ctcm_pr_debug("%s cp:%i enter: %s() %s\n", | ||
1377 | dev->name, smp_processor_id(), __FUNCTION__, ch->id); | ||
1378 | /* caller has requested driver to throttle back */ | ||
1379 | while ((fsm_getstate(grp->fsm) != MPCG_STATE_FLOWC) && | ||
1380 | (skb = skb_dequeue(&ch->io_queue))) { | ||
1381 | ctcmpc_unpack_skb(ch, skb); | ||
1382 | if (grp->out_of_sequence > 20) { | ||
1383 | /* assume data loss has occurred if */ | ||
1384 | /* missing seq_num for extended */ | ||
1385 | /* period of time */ | ||
1386 | grp->out_of_sequence = 0; | ||
1387 | fsm_event(grp->fsm, MPCG_EVENT_INOP, dev); | ||
1388 | break; | ||
1389 | } | ||
1390 | if (skb == skb_peek(&ch->io_queue)) | ||
1391 | break; | ||
1392 | } | ||
1393 | if (do_debug) | ||
1394 | ctcm_pr_debug("ctcmpc exit : %s %s(): ch=0x%p id=%s\n", | ||
1395 | dev->name, __FUNCTION__, ch, ch->id); | ||
1396 | return; | ||
1397 | } | ||
1398 | |||
1399 | /* | ||
1400 | * MPC Group Initializations | ||
1401 | */ | ||
1402 | struct mpc_group *ctcmpc_init_mpc_group(struct ctcm_priv *priv) | ||
1403 | { | ||
1404 | struct mpc_group *grp; | ||
1405 | |||
1406 | CTCM_DBF_TEXT(MPC_SETUP, 3, __FUNCTION__); | ||
1407 | |||
1408 | grp = kzalloc(sizeof(struct mpc_group), GFP_KERNEL); | ||
1409 | if (grp == NULL) | ||
1410 | return NULL; | ||
1411 | |||
1412 | grp->fsm = | ||
1413 | init_fsm("mpcg", mpcg_state_names, mpcg_event_names, | ||
1414 | MPCG_NR_STATES, MPCG_NR_EVENTS, mpcg_fsm, | ||
1415 | mpcg_fsm_len, GFP_KERNEL); | ||
1416 | if (grp->fsm == NULL) { | ||
1417 | kfree(grp); | ||
1418 | return NULL; | ||
1419 | } | ||
1420 | |||
1421 | fsm_newstate(grp->fsm, MPCG_STATE_RESET); | ||
1422 | fsm_settimer(grp->fsm, &grp->timer); | ||
1423 | |||
1424 | grp->xid_skb = | ||
1425 | __dev_alloc_skb(MPC_BUFSIZE_DEFAULT, GFP_ATOMIC | GFP_DMA); | ||
1426 | if (grp->xid_skb == NULL) { | ||
1427 | printk(KERN_INFO "Couldn't alloc MPCgroup xid_skb\n"); | ||
1428 | kfree_fsm(grp->fsm); | ||
1429 | kfree(grp); | ||
1430 | return NULL; | ||
1431 | } | ||
1432 | /* base xid for all channels in group */ | ||
1433 | grp->xid_skb_data = grp->xid_skb->data; | ||
1434 | grp->xid_th = (struct th_header *)grp->xid_skb->data; | ||
1435 | memcpy(skb_put(grp->xid_skb, TH_HEADER_LENGTH), | ||
1436 | &thnorm, TH_HEADER_LENGTH); | ||
1437 | |||
1438 | grp->xid = (struct xid2 *) skb_tail_pointer(grp->xid_skb); | ||
1439 | memcpy(skb_put(grp->xid_skb, XID2_LENGTH), &init_xid, XID2_LENGTH); | ||
1440 | grp->xid->xid2_adj_id = jiffies | 0xfff00000; | ||
1441 | grp->xid->xid2_sender_id = jiffies; | ||
1442 | |||
1443 | grp->xid_id = skb_tail_pointer(grp->xid_skb); | ||
1444 | memcpy(skb_put(grp->xid_skb, 4), "VTAM", 4); | ||
1445 | |||
1446 | grp->rcvd_xid_skb = | ||
1447 | __dev_alloc_skb(MPC_BUFSIZE_DEFAULT, GFP_ATOMIC|GFP_DMA); | ||
1448 | if (grp->rcvd_xid_skb == NULL) { | ||
1449 | printk(KERN_INFO "Couldn't alloc MPCgroup rcvd_xid_skb\n"); | ||
1450 | kfree_fsm(grp->fsm); | ||
1451 | dev_kfree_skb(grp->xid_skb); | ||
1452 | kfree(grp); | ||
1453 | return NULL; | ||
1454 | } | ||
1455 | grp->rcvd_xid_data = grp->rcvd_xid_skb->data; | ||
1456 | grp->rcvd_xid_th = (struct th_header *)grp->rcvd_xid_skb->data; | ||
1457 | memcpy(skb_put(grp->rcvd_xid_skb, TH_HEADER_LENGTH), | ||
1458 | &thnorm, TH_HEADER_LENGTH); | ||
1459 | grp->saved_xid2 = NULL; | ||
1460 | priv->xid = grp->xid; | ||
1461 | priv->mpcg = grp; | ||
1462 | return grp; | ||
1463 | } | ||
1464 | |||
1465 | /* | ||
1466 | * The MPC Group Station FSM | ||
1467 | */ | ||
1468 | |||
1469 | /* | ||
1470 | * MPC Group Station FSM actions | ||
1471 | * CTCM_PROTO_MPC only | ||
1472 | */ | ||
1473 | |||
1474 | /** | ||
1475 | * NOP action for statemachines | ||
1476 | */ | ||
1477 | static void mpc_action_nop(fsm_instance *fi, int event, void *arg) | ||
1478 | { | ||
1479 | } | ||
1480 | |||
1481 | /* | ||
1482 | * invoked when the device transitions to dev_stopped | ||
1483 | * MPC will stop each individual channel if a single XID failure | ||
1484 | * occurs, or will intitiate all channels be stopped if a GROUP | ||
1485 | * level failure occurs. | ||
1486 | */ | ||
1487 | static void mpc_action_go_inop(fsm_instance *fi, int event, void *arg) | ||
1488 | { | ||
1489 | struct net_device *dev = arg; | ||
1490 | struct ctcm_priv *priv; | ||
1491 | struct mpc_group *grp; | ||
1492 | int rc = 0; | ||
1493 | struct channel *wch, *rch; | ||
1494 | |||
1495 | if (dev == NULL) { | ||
1496 | printk(KERN_INFO "%s() dev=NULL\n", __FUNCTION__); | ||
1497 | return; | ||
1498 | } | ||
1499 | |||
1500 | ctcm_pr_debug("ctcmpc enter: %s %s()\n", dev->name, __FUNCTION__); | ||
1501 | |||
1502 | priv = dev->priv; | ||
1503 | grp = priv->mpcg; | ||
1504 | grp->flow_off_called = 0; | ||
1505 | |||
1506 | fsm_deltimer(&grp->timer); | ||
1507 | |||
1508 | if (grp->channels_terminating) | ||
1509 | goto done; | ||
1510 | |||
1511 | grp->channels_terminating = 1; | ||
1512 | |||
1513 | grp->saved_state = fsm_getstate(grp->fsm); | ||
1514 | fsm_newstate(grp->fsm, MPCG_STATE_INOP); | ||
1515 | if (grp->saved_state > MPCG_STATE_XID7INITF) | ||
1516 | printk(KERN_NOTICE "%s:MPC GROUP INOPERATIVE\n", dev->name); | ||
1517 | if ((grp->saved_state != MPCG_STATE_RESET) || | ||
1518 | /* dealloc_channel has been called */ | ||
1519 | ((grp->saved_state == MPCG_STATE_RESET) && | ||
1520 | (grp->port_persist == 0))) | ||
1521 | fsm_deltimer(&priv->restart_timer); | ||
1522 | |||
1523 | wch = priv->channel[WRITE]; | ||
1524 | rch = priv->channel[READ]; | ||
1525 | |||
1526 | switch (grp->saved_state) { | ||
1527 | case MPCG_STATE_RESET: | ||
1528 | case MPCG_STATE_INOP: | ||
1529 | case MPCG_STATE_XID2INITW: | ||
1530 | case MPCG_STATE_XID0IOWAIT: | ||
1531 | case MPCG_STATE_XID2INITX: | ||
1532 | case MPCG_STATE_XID7INITW: | ||
1533 | case MPCG_STATE_XID7INITX: | ||
1534 | case MPCG_STATE_XID0IOWAIX: | ||
1535 | case MPCG_STATE_XID7INITI: | ||
1536 | case MPCG_STATE_XID7INITZ: | ||
1537 | case MPCG_STATE_XID7INITF: | ||
1538 | break; | ||
1539 | case MPCG_STATE_FLOWC: | ||
1540 | case MPCG_STATE_READY: | ||
1541 | default: | ||
1542 | tasklet_hi_schedule(&wch->ch_disc_tasklet); | ||
1543 | } | ||
1544 | |||
1545 | grp->xid2_tgnum = 0; | ||
1546 | grp->group_max_buflen = 0; /*min of all received */ | ||
1547 | grp->outstanding_xid2 = 0; | ||
1548 | grp->outstanding_xid7 = 0; | ||
1549 | grp->outstanding_xid7_p2 = 0; | ||
1550 | grp->saved_xid2 = NULL; | ||
1551 | grp->xidnogood = 0; | ||
1552 | grp->changed_side = 0; | ||
1553 | |||
1554 | grp->rcvd_xid_skb->data = grp->rcvd_xid_data; | ||
1555 | skb_reset_tail_pointer(grp->rcvd_xid_skb); | ||
1556 | grp->rcvd_xid_skb->len = 0; | ||
1557 | grp->rcvd_xid_th = (struct th_header *)grp->rcvd_xid_skb->data; | ||
1558 | memcpy(skb_put(grp->rcvd_xid_skb, TH_HEADER_LENGTH), &thnorm, | ||
1559 | TH_HEADER_LENGTH); | ||
1560 | |||
1561 | if (grp->send_qllc_disc == 1) { | ||
1562 | grp->send_qllc_disc = 0; | ||
1563 | rc = mpc_send_qllc_discontact(dev); | ||
1564 | } | ||
1565 | |||
1566 | /* DO NOT issue DEV_EVENT_STOP directly out of this code */ | ||
1567 | /* This can result in INOP of VTAM PU due to halting of */ | ||
1568 | /* outstanding IO which causes a sense to be returned */ | ||
1569 | /* Only about 3 senses are allowed and then IOS/VTAM will*/ | ||
1570 | /* ebcome unreachable without manual intervention */ | ||
1571 | if ((grp->port_persist == 1) || (grp->alloc_called)) { | ||
1572 | grp->alloc_called = 0; | ||
1573 | fsm_deltimer(&priv->restart_timer); | ||
1574 | fsm_addtimer(&priv->restart_timer, | ||
1575 | 500, | ||
1576 | DEV_EVENT_RESTART, | ||
1577 | dev); | ||
1578 | fsm_newstate(grp->fsm, MPCG_STATE_RESET); | ||
1579 | if (grp->saved_state > MPCG_STATE_XID7INITF) | ||
1580 | printk(KERN_NOTICE "%s:MPC GROUP RECOVERY SCHEDULED\n", | ||
1581 | dev->name); | ||
1582 | } else { | ||
1583 | fsm_deltimer(&priv->restart_timer); | ||
1584 | fsm_addtimer(&priv->restart_timer, 500, DEV_EVENT_STOP, dev); | ||
1585 | fsm_newstate(grp->fsm, MPCG_STATE_RESET); | ||
1586 | printk(KERN_NOTICE "%s:MPC GROUP RECOVERY NOT ATTEMPTED\n", | ||
1587 | dev->name); | ||
1588 | } | ||
1589 | |||
1590 | done: | ||
1591 | ctcm_pr_debug("ctcmpc exit:%s %s()\n", dev->name, __FUNCTION__); | ||
1592 | return; | ||
1593 | } | ||
1594 | |||
1595 | /** | ||
1596 | * Handle mpc group action timeout. | ||
1597 | * MPC Group Station FSM action | ||
1598 | * CTCM_PROTO_MPC only | ||
1599 | * | ||
1600 | * fi An instance of an mpc_group fsm. | ||
1601 | * event The event, just happened. | ||
1602 | * arg Generic pointer, casted from net_device * upon call. | ||
1603 | */ | ||
1604 | static void mpc_action_timeout(fsm_instance *fi, int event, void *arg) | ||
1605 | { | ||
1606 | struct net_device *dev = arg; | ||
1607 | struct ctcm_priv *priv; | ||
1608 | struct mpc_group *grp; | ||
1609 | struct channel *wch; | ||
1610 | struct channel *rch; | ||
1611 | |||
1612 | CTCM_DBF_TEXT(MPC_TRACE, 6, __FUNCTION__); | ||
1613 | |||
1614 | if (dev == NULL) { | ||
1615 | CTCM_DBF_TEXT_(MPC_ERROR, 4, "%s: dev=NULL\n", __FUNCTION__); | ||
1616 | return; | ||
1617 | } | ||
1618 | |||
1619 | priv = dev->priv; | ||
1620 | grp = priv->mpcg; | ||
1621 | wch = priv->channel[WRITE]; | ||
1622 | rch = priv->channel[READ]; | ||
1623 | |||
1624 | switch (fsm_getstate(grp->fsm)) { | ||
1625 | case MPCG_STATE_XID2INITW: | ||
1626 | /* Unless there is outstanding IO on the */ | ||
1627 | /* channel just return and wait for ATTN */ | ||
1628 | /* interrupt to begin XID negotiations */ | ||
1629 | if ((fsm_getstate(rch->fsm) == CH_XID0_PENDING) && | ||
1630 | (fsm_getstate(wch->fsm) == CH_XID0_PENDING)) | ||
1631 | break; | ||
1632 | default: | ||
1633 | fsm_event(grp->fsm, MPCG_EVENT_INOP, dev); | ||
1634 | } | ||
1635 | |||
1636 | CTCM_DBF_TEXT_(MPC_TRACE, 6, "%s: dev=%s exit", | ||
1637 | __FUNCTION__, dev->name); | ||
1638 | return; | ||
1639 | } | ||
1640 | |||
1641 | /* | ||
1642 | * MPC Group Station FSM action | ||
1643 | * CTCM_PROTO_MPC only | ||
1644 | */ | ||
1645 | void mpc_action_discontact(fsm_instance *fi, int event, void *arg) | ||
1646 | { | ||
1647 | struct mpcg_info *mpcginfo = arg; | ||
1648 | struct channel *ch = mpcginfo->ch; | ||
1649 | struct net_device *dev = ch->netdev; | ||
1650 | struct ctcm_priv *priv = dev->priv; | ||
1651 | struct mpc_group *grp = priv->mpcg; | ||
1652 | |||
1653 | if (ch == NULL) { | ||
1654 | printk(KERN_INFO "%s() ch=NULL\n", __FUNCTION__); | ||
1655 | return; | ||
1656 | } | ||
1657 | if (ch->netdev == NULL) { | ||
1658 | printk(KERN_INFO "%s() dev=NULL\n", __FUNCTION__); | ||
1659 | return; | ||
1660 | } | ||
1661 | |||
1662 | ctcm_pr_debug("ctcmpc enter: %s %s()\n", dev->name, __FUNCTION__); | ||
1663 | |||
1664 | grp->send_qllc_disc = 1; | ||
1665 | fsm_event(grp->fsm, MPCG_EVENT_INOP, dev); | ||
1666 | |||
1667 | ctcm_pr_debug("ctcmpc exit: %s %s()\n", dev->name, __FUNCTION__); | ||
1668 | return; | ||
1669 | } | ||
1670 | |||
1671 | /* | ||
1672 | * MPC Group Station - not part of FSM | ||
1673 | * CTCM_PROTO_MPC only | ||
1674 | * called from add_channel in ctcm_main.c | ||
1675 | */ | ||
1676 | void mpc_action_send_discontact(unsigned long thischan) | ||
1677 | { | ||
1678 | struct channel *ch; | ||
1679 | struct net_device *dev; | ||
1680 | struct ctcm_priv *priv; | ||
1681 | struct mpc_group *grp; | ||
1682 | int rc = 0; | ||
1683 | unsigned long saveflags; | ||
1684 | |||
1685 | ch = (struct channel *)thischan; | ||
1686 | dev = ch->netdev; | ||
1687 | priv = dev->priv; | ||
1688 | grp = priv->mpcg; | ||
1689 | |||
1690 | ctcm_pr_info("ctcmpc: %s cp:%i enter: %s() GrpState:%s ChState:%s\n", | ||
1691 | dev->name, | ||
1692 | smp_processor_id(), | ||
1693 | __FUNCTION__, | ||
1694 | fsm_getstate_str(grp->fsm), | ||
1695 | fsm_getstate_str(ch->fsm)); | ||
1696 | saveflags = 0; /* avoids compiler warning with | ||
1697 | spin_unlock_irqrestore */ | ||
1698 | |||
1699 | spin_lock_irqsave(get_ccwdev_lock(ch->cdev), saveflags); | ||
1700 | rc = ccw_device_start(ch->cdev, &ch->ccw[15], | ||
1701 | (unsigned long)ch, 0xff, 0); | ||
1702 | spin_unlock_irqrestore(get_ccwdev_lock(ch->cdev), saveflags); | ||
1703 | |||
1704 | if (rc != 0) { | ||
1705 | ctcm_pr_info("ctcmpc: %s() ch:%s IO failed \n", | ||
1706 | __FUNCTION__, | ||
1707 | ch->id); | ||
1708 | ctcm_ccw_check_rc(ch, rc, "send discontact"); | ||
1709 | /* Not checking return code value here */ | ||
1710 | /* Making best effort to notify partner*/ | ||
1711 | /* that MPC Group is going down */ | ||
1712 | } | ||
1713 | |||
1714 | ctcm_pr_debug("ctcmpc exit: %s %s()\n", dev->name, __FUNCTION__); | ||
1715 | return; | ||
1716 | } | ||
1717 | |||
1718 | |||
1719 | /* | ||
1720 | * helper function of mpc FSM | ||
1721 | * CTCM_PROTO_MPC only | ||
1722 | * mpc_action_rcvd_xid7 | ||
1723 | */ | ||
1724 | static int mpc_validate_xid(struct mpcg_info *mpcginfo) | ||
1725 | { | ||
1726 | struct channel *ch = mpcginfo->ch; | ||
1727 | struct net_device *dev = ch->netdev; | ||
1728 | struct ctcm_priv *priv = dev->priv; | ||
1729 | struct mpc_group *grp = priv->mpcg; | ||
1730 | struct xid2 *xid = mpcginfo->xid; | ||
1731 | int failed = 0; | ||
1732 | int rc = 0; | ||
1733 | __u64 our_id, their_id = 0; | ||
1734 | int len; | ||
1735 | |||
1736 | len = TH_HEADER_LENGTH + PDU_HEADER_LENGTH; | ||
1737 | |||
1738 | ctcm_pr_debug("ctcmpc enter: %s()\n", __FUNCTION__); | ||
1739 | |||
1740 | if (mpcginfo->xid == NULL) { | ||
1741 | printk(KERN_INFO "%s() xid=NULL\n", __FUNCTION__); | ||
1742 | rc = 1; | ||
1743 | goto done; | ||
1744 | } | ||
1745 | |||
1746 | ctcm_pr_debug("ctcmpc : %s xid received()\n", __FUNCTION__); | ||
1747 | ctcmpc_dumpit((char *)mpcginfo->xid, XID2_LENGTH); | ||
1748 | |||
1749 | /*the received direction should be the opposite of ours */ | ||
1750 | if (((CHANNEL_DIRECTION(ch->flags) == READ) ? XID2_WRITE_SIDE : | ||
1751 | XID2_READ_SIDE) != xid->xid2_dlc_type) { | ||
1752 | failed = 1; | ||
1753 | printk(KERN_INFO "ctcmpc:%s() XID REJECTED - READ-WRITE CH " | ||
1754 | "Pairing Invalid \n", __FUNCTION__); | ||
1755 | } | ||
1756 | |||
1757 | if (xid->xid2_dlc_type == XID2_READ_SIDE) { | ||
1758 | ctcm_pr_debug("ctcmpc: %s(): grpmaxbuf:%d xid2buflen:%d\n", | ||
1759 | __FUNCTION__, grp->group_max_buflen, | ||
1760 | xid->xid2_buf_len); | ||
1761 | |||
1762 | if (grp->group_max_buflen == 0 || | ||
1763 | grp->group_max_buflen > xid->xid2_buf_len - len) | ||
1764 | grp->group_max_buflen = xid->xid2_buf_len - len; | ||
1765 | } | ||
1766 | |||
1767 | |||
1768 | if (grp->saved_xid2 == NULL) { | ||
1769 | grp->saved_xid2 = | ||
1770 | (struct xid2 *)skb_tail_pointer(grp->rcvd_xid_skb); | ||
1771 | |||
1772 | memcpy(skb_put(grp->rcvd_xid_skb, | ||
1773 | XID2_LENGTH), xid, XID2_LENGTH); | ||
1774 | grp->rcvd_xid_skb->data = grp->rcvd_xid_data; | ||
1775 | |||
1776 | skb_reset_tail_pointer(grp->rcvd_xid_skb); | ||
1777 | grp->rcvd_xid_skb->len = 0; | ||
1778 | |||
1779 | /* convert two 32 bit numbers into 1 64 bit for id compare */ | ||
1780 | our_id = (__u64)priv->xid->xid2_adj_id; | ||
1781 | our_id = our_id << 32; | ||
1782 | our_id = our_id + priv->xid->xid2_sender_id; | ||
1783 | their_id = (__u64)xid->xid2_adj_id; | ||
1784 | their_id = their_id << 32; | ||
1785 | their_id = their_id + xid->xid2_sender_id; | ||
1786 | /* lower id assume the xside role */ | ||
1787 | if (our_id < their_id) { | ||
1788 | grp->roll = XSIDE; | ||
1789 | ctcm_pr_debug("ctcmpc :%s() WE HAVE LOW ID-" | ||
1790 | "TAKE XSIDE\n", __FUNCTION__); | ||
1791 | } else { | ||
1792 | grp->roll = YSIDE; | ||
1793 | ctcm_pr_debug("ctcmpc :%s() WE HAVE HIGH ID-" | ||
1794 | "TAKE YSIDE\n", __FUNCTION__); | ||
1795 | } | ||
1796 | |||
1797 | } else { | ||
1798 | if (xid->xid2_flag4 != grp->saved_xid2->xid2_flag4) { | ||
1799 | failed = 1; | ||
1800 | printk(KERN_INFO "%s XID REJECTED - XID Flag Byte4\n", | ||
1801 | __FUNCTION__); | ||
1802 | } | ||
1803 | if (xid->xid2_flag2 == 0x40) { | ||
1804 | failed = 1; | ||
1805 | printk(KERN_INFO "%s XID REJECTED - XID NOGOOD\n", | ||
1806 | __FUNCTION__); | ||
1807 | } | ||
1808 | if (xid->xid2_adj_id != grp->saved_xid2->xid2_adj_id) { | ||
1809 | failed = 1; | ||
1810 | printk(KERN_INFO "%s XID REJECTED - " | ||
1811 | "Adjacent Station ID Mismatch\n", | ||
1812 | __FUNCTION__); | ||
1813 | } | ||
1814 | if (xid->xid2_sender_id != grp->saved_xid2->xid2_sender_id) { | ||
1815 | failed = 1; | ||
1816 | printk(KERN_INFO "%s XID REJECTED - " | ||
1817 | "Sender Address Mismatch\n", __FUNCTION__); | ||
1818 | |||
1819 | } | ||
1820 | } | ||
1821 | |||
1822 | if (failed) { | ||
1823 | ctcm_pr_info("ctcmpc : %s() failed\n", __FUNCTION__); | ||
1824 | priv->xid->xid2_flag2 = 0x40; | ||
1825 | grp->saved_xid2->xid2_flag2 = 0x40; | ||
1826 | rc = 1; | ||
1827 | } | ||
1828 | |||
1829 | done: | ||
1830 | |||
1831 | ctcm_pr_debug("ctcmpc exit: %s()\n", __FUNCTION__); | ||
1832 | return rc; | ||
1833 | } | ||
1834 | |||
1835 | /* | ||
1836 | * MPC Group Station FSM action | ||
1837 | * CTCM_PROTO_MPC only | ||
1838 | */ | ||
1839 | static void mpc_action_side_xid(fsm_instance *fsm, void *arg, int side) | ||
1840 | { | ||
1841 | struct channel *ch = arg; | ||
1842 | struct ctcm_priv *priv; | ||
1843 | struct mpc_group *grp = NULL; | ||
1844 | struct net_device *dev = NULL; | ||
1845 | int rc = 0; | ||
1846 | int gotlock = 0; | ||
1847 | unsigned long saveflags = 0; /* avoids compiler warning with | ||
1848 | spin_unlock_irqrestore */ | ||
1849 | |||
1850 | if (ch == NULL) { | ||
1851 | printk(KERN_INFO "%s ch=NULL\n", __FUNCTION__); | ||
1852 | goto done; | ||
1853 | } | ||
1854 | |||
1855 | if (do_debug) | ||
1856 | ctcm_pr_debug("ctcmpc enter: %s(): cp=%i ch=0x%p id=%s\n", | ||
1857 | __FUNCTION__, smp_processor_id(), ch, ch->id); | ||
1858 | |||
1859 | dev = ch->netdev; | ||
1860 | if (dev == NULL) { | ||
1861 | printk(KERN_INFO "%s dev=NULL\n", __FUNCTION__); | ||
1862 | goto done; | ||
1863 | } | ||
1864 | |||
1865 | priv = dev->priv; | ||
1866 | if (priv == NULL) { | ||
1867 | printk(KERN_INFO "%s priv=NULL\n", __FUNCTION__); | ||
1868 | goto done; | ||
1869 | } | ||
1870 | |||
1871 | grp = priv->mpcg; | ||
1872 | if (grp == NULL) { | ||
1873 | printk(KERN_INFO "%s grp=NULL\n", __FUNCTION__); | ||
1874 | goto done; | ||
1875 | } | ||
1876 | |||
1877 | if (ctcm_checkalloc_buffer(ch)) | ||
1878 | goto done; | ||
1879 | |||
1880 | /* skb data-buffer referencing: */ | ||
1881 | |||
1882 | ch->trans_skb->data = ch->trans_skb_data; | ||
1883 | skb_reset_tail_pointer(ch->trans_skb); | ||
1884 | ch->trans_skb->len = 0; | ||
1885 | /* result of the previous 3 statements is NOT always | ||
1886 | * already set after ctcm_checkalloc_buffer | ||
1887 | * because of possible reuse of the trans_skb | ||
1888 | */ | ||
1889 | memset(ch->trans_skb->data, 0, 16); | ||
1890 | ch->rcvd_xid_th = (struct th_header *)ch->trans_skb_data; | ||
1891 | /* check is main purpose here: */ | ||
1892 | skb_put(ch->trans_skb, TH_HEADER_LENGTH); | ||
1893 | ch->rcvd_xid = (struct xid2 *)skb_tail_pointer(ch->trans_skb); | ||
1894 | /* check is main purpose here: */ | ||
1895 | skb_put(ch->trans_skb, XID2_LENGTH); | ||
1896 | ch->rcvd_xid_id = skb_tail_pointer(ch->trans_skb); | ||
1897 | /* cleanup back to startpoint */ | ||
1898 | ch->trans_skb->data = ch->trans_skb_data; | ||
1899 | skb_reset_tail_pointer(ch->trans_skb); | ||
1900 | ch->trans_skb->len = 0; | ||
1901 | |||
1902 | /* non-checking rewrite of above skb data-buffer referencing: */ | ||
1903 | /* | ||
1904 | memset(ch->trans_skb->data, 0, 16); | ||
1905 | ch->rcvd_xid_th = (struct th_header *)ch->trans_skb_data; | ||
1906 | ch->rcvd_xid = (struct xid2 *)(ch->trans_skb_data + TH_HEADER_LENGTH); | ||
1907 | ch->rcvd_xid_id = ch->trans_skb_data + TH_HEADER_LENGTH + XID2_LENGTH; | ||
1908 | */ | ||
1909 | |||
1910 | ch->ccw[8].flags = CCW_FLAG_SLI | CCW_FLAG_CC; | ||
1911 | ch->ccw[8].count = 0; | ||
1912 | ch->ccw[8].cda = 0x00; | ||
1913 | |||
1914 | if (side == XSIDE) { | ||
1915 | /* mpc_action_xside_xid */ | ||
1916 | if (ch->xid_th == NULL) { | ||
1917 | printk(KERN_INFO "%s ch->xid_th=NULL\n", __FUNCTION__); | ||
1918 | goto done; | ||
1919 | } | ||
1920 | ch->ccw[9].cmd_code = CCW_CMD_WRITE; | ||
1921 | ch->ccw[9].flags = CCW_FLAG_SLI | CCW_FLAG_CC; | ||
1922 | ch->ccw[9].count = TH_HEADER_LENGTH; | ||
1923 | ch->ccw[9].cda = virt_to_phys(ch->xid_th); | ||
1924 | |||
1925 | if (ch->xid == NULL) { | ||
1926 | printk(KERN_INFO "%s ch->xid=NULL\n", __FUNCTION__); | ||
1927 | goto done; | ||
1928 | } | ||
1929 | |||
1930 | ch->ccw[10].cmd_code = CCW_CMD_WRITE; | ||
1931 | ch->ccw[10].flags = CCW_FLAG_SLI | CCW_FLAG_CC; | ||
1932 | ch->ccw[10].count = XID2_LENGTH; | ||
1933 | ch->ccw[10].cda = virt_to_phys(ch->xid); | ||
1934 | |||
1935 | ch->ccw[11].cmd_code = CCW_CMD_READ; | ||
1936 | ch->ccw[11].flags = CCW_FLAG_SLI | CCW_FLAG_CC; | ||
1937 | ch->ccw[11].count = TH_HEADER_LENGTH; | ||
1938 | ch->ccw[11].cda = virt_to_phys(ch->rcvd_xid_th); | ||
1939 | |||
1940 | ch->ccw[12].cmd_code = CCW_CMD_READ; | ||
1941 | ch->ccw[12].flags = CCW_FLAG_SLI | CCW_FLAG_CC; | ||
1942 | ch->ccw[12].count = XID2_LENGTH; | ||
1943 | ch->ccw[12].cda = virt_to_phys(ch->rcvd_xid); | ||
1944 | |||
1945 | ch->ccw[13].cmd_code = CCW_CMD_READ; | ||
1946 | ch->ccw[13].cda = virt_to_phys(ch->rcvd_xid_id); | ||
1947 | |||
1948 | } else { /* side == YSIDE : mpc_action_yside_xid */ | ||
1949 | ch->ccw[9].cmd_code = CCW_CMD_READ; | ||
1950 | ch->ccw[9].flags = CCW_FLAG_SLI | CCW_FLAG_CC; | ||
1951 | ch->ccw[9].count = TH_HEADER_LENGTH; | ||
1952 | ch->ccw[9].cda = virt_to_phys(ch->rcvd_xid_th); | ||
1953 | |||
1954 | ch->ccw[10].cmd_code = CCW_CMD_READ; | ||
1955 | ch->ccw[10].flags = CCW_FLAG_SLI | CCW_FLAG_CC; | ||
1956 | ch->ccw[10].count = XID2_LENGTH; | ||
1957 | ch->ccw[10].cda = virt_to_phys(ch->rcvd_xid); | ||
1958 | |||
1959 | if (ch->xid_th == NULL) { | ||
1960 | printk(KERN_INFO "%s ch->xid_th=NULL\n", __FUNCTION__); | ||
1961 | goto done; | ||
1962 | } | ||
1963 | ch->ccw[11].cmd_code = CCW_CMD_WRITE; | ||
1964 | ch->ccw[11].flags = CCW_FLAG_SLI | CCW_FLAG_CC; | ||
1965 | ch->ccw[11].count = TH_HEADER_LENGTH; | ||
1966 | ch->ccw[11].cda = virt_to_phys(ch->xid_th); | ||
1967 | |||
1968 | if (ch->xid == NULL) { | ||
1969 | printk(KERN_INFO "%s ch->xid=NULL\n", __FUNCTION__); | ||
1970 | goto done; | ||
1971 | } | ||
1972 | ch->ccw[12].cmd_code = CCW_CMD_WRITE; | ||
1973 | ch->ccw[12].flags = CCW_FLAG_SLI | CCW_FLAG_CC; | ||
1974 | ch->ccw[12].count = XID2_LENGTH; | ||
1975 | ch->ccw[12].cda = virt_to_phys(ch->xid); | ||
1976 | |||
1977 | if (ch->xid_id == NULL) { | ||
1978 | printk(KERN_INFO "%s ch->xid_id=NULL\n", __FUNCTION__); | ||
1979 | goto done; | ||
1980 | } | ||
1981 | ch->ccw[13].cmd_code = CCW_CMD_WRITE; | ||
1982 | ch->ccw[13].cda = virt_to_phys(ch->xid_id); | ||
1983 | |||
1984 | } | ||
1985 | ch->ccw[13].flags = CCW_FLAG_SLI | CCW_FLAG_CC; | ||
1986 | ch->ccw[13].count = 4; | ||
1987 | |||
1988 | ch->ccw[14].cmd_code = CCW_CMD_NOOP; | ||
1989 | ch->ccw[14].flags = CCW_FLAG_SLI; | ||
1990 | ch->ccw[14].count = 0; | ||
1991 | ch->ccw[14].cda = 0; | ||
1992 | |||
1993 | if (do_debug_ccw) | ||
1994 | ctcmpc_dumpit((char *)&ch->ccw[8], sizeof(struct ccw1) * 7); | ||
1995 | |||
1996 | ctcmpc_dumpit((char *)ch->xid_th, TH_HEADER_LENGTH); | ||
1997 | ctcmpc_dumpit((char *)ch->xid, XID2_LENGTH); | ||
1998 | ctcmpc_dumpit((char *)ch->xid_id, 4); | ||
1999 | if (!in_irq()) { | ||
2000 | /* Such conditional locking is a known problem for | ||
2001 | * sparse because its static undeterministic. | ||
2002 | * Warnings should be ignored here. */ | ||
2003 | spin_lock_irqsave(get_ccwdev_lock(ch->cdev), saveflags); | ||
2004 | gotlock = 1; | ||
2005 | } | ||
2006 | |||
2007 | fsm_addtimer(&ch->timer, 5000 , CTC_EVENT_TIMER, ch); | ||
2008 | rc = ccw_device_start(ch->cdev, &ch->ccw[8], | ||
2009 | (unsigned long)ch, 0xff, 0); | ||
2010 | |||
2011 | if (gotlock) /* see remark above about conditional locking */ | ||
2012 | spin_unlock_irqrestore(get_ccwdev_lock(ch->cdev), saveflags); | ||
2013 | |||
2014 | if (rc != 0) { | ||
2015 | ctcm_pr_info("ctcmpc: %s() ch:%s IO failed \n", | ||
2016 | __FUNCTION__, ch->id); | ||
2017 | ctcm_ccw_check_rc(ch, rc, | ||
2018 | (side == XSIDE) ? "x-side XID" : "y-side XID"); | ||
2019 | } | ||
2020 | |||
2021 | done: | ||
2022 | if (do_debug) | ||
2023 | ctcm_pr_debug("ctcmpc exit : %s(): ch=0x%p id=%s\n", | ||
2024 | __FUNCTION__, ch, ch->id); | ||
2025 | return; | ||
2026 | |||
2027 | } | ||
2028 | |||
2029 | /* | ||
2030 | * MPC Group Station FSM action | ||
2031 | * CTCM_PROTO_MPC only | ||
2032 | */ | ||
2033 | static void mpc_action_xside_xid(fsm_instance *fsm, int event, void *arg) | ||
2034 | { | ||
2035 | mpc_action_side_xid(fsm, arg, XSIDE); | ||
2036 | } | ||
2037 | |||
2038 | /* | ||
2039 | * MPC Group Station FSM action | ||
2040 | * CTCM_PROTO_MPC only | ||
2041 | */ | ||
2042 | static void mpc_action_yside_xid(fsm_instance *fsm, int event, void *arg) | ||
2043 | { | ||
2044 | mpc_action_side_xid(fsm, arg, YSIDE); | ||
2045 | } | ||
2046 | |||
2047 | /* | ||
2048 | * MPC Group Station FSM action | ||
2049 | * CTCM_PROTO_MPC only | ||
2050 | */ | ||
2051 | static void mpc_action_doxid0(fsm_instance *fsm, int event, void *arg) | ||
2052 | { | ||
2053 | struct channel *ch = arg; | ||
2054 | struct ctcm_priv *priv; | ||
2055 | struct mpc_group *grp = NULL; | ||
2056 | struct net_device *dev = NULL; | ||
2057 | |||
2058 | if (do_debug) | ||
2059 | ctcm_pr_debug("ctcmpc enter: %s(): cp=%i ch=0x%p id=%s\n", | ||
2060 | __FUNCTION__, smp_processor_id(), ch, ch->id); | ||
2061 | |||
2062 | if (ch == NULL) { | ||
2063 | printk(KERN_WARNING "%s ch=NULL\n", __FUNCTION__); | ||
2064 | goto done; | ||
2065 | } | ||
2066 | |||
2067 | dev = ch->netdev; | ||
2068 | if (dev == NULL) { | ||
2069 | printk(KERN_WARNING "%s dev=NULL\n", __FUNCTION__); | ||
2070 | goto done; | ||
2071 | } | ||
2072 | |||
2073 | priv = dev->priv; | ||
2074 | if (priv == NULL) { | ||
2075 | printk(KERN_WARNING "%s priv=NULL\n", __FUNCTION__); | ||
2076 | goto done; | ||
2077 | } | ||
2078 | |||
2079 | grp = priv->mpcg; | ||
2080 | if (grp == NULL) { | ||
2081 | printk(KERN_WARNING "%s grp=NULL\n", __FUNCTION__); | ||
2082 | goto done; | ||
2083 | } | ||
2084 | |||
2085 | if (ch->xid == NULL) { | ||
2086 | printk(KERN_WARNING "%s ch-xid=NULL\n", __FUNCTION__); | ||
2087 | goto done; | ||
2088 | } | ||
2089 | |||
2090 | fsm_newstate(ch->fsm, CH_XID0_INPROGRESS); | ||
2091 | |||
2092 | ch->xid->xid2_option = XID2_0; | ||
2093 | |||
2094 | switch (fsm_getstate(grp->fsm)) { | ||
2095 | case MPCG_STATE_XID2INITW: | ||
2096 | case MPCG_STATE_XID2INITX: | ||
2097 | ch->ccw[8].cmd_code = CCW_CMD_SENSE_CMD; | ||
2098 | break; | ||
2099 | case MPCG_STATE_XID0IOWAIT: | ||
2100 | case MPCG_STATE_XID0IOWAIX: | ||
2101 | ch->ccw[8].cmd_code = CCW_CMD_WRITE_CTL; | ||
2102 | break; | ||
2103 | } | ||
2104 | |||
2105 | fsm_event(grp->fsm, MPCG_EVENT_DOIO, ch); | ||
2106 | |||
2107 | done: | ||
2108 | if (do_debug) | ||
2109 | ctcm_pr_debug("ctcmpc exit : %s(): ch=0x%p id=%s\n", | ||
2110 | __FUNCTION__, ch, ch->id); | ||
2111 | return; | ||
2112 | |||
2113 | } | ||
2114 | |||
2115 | /* | ||
2116 | * MPC Group Station FSM action | ||
2117 | * CTCM_PROTO_MPC only | ||
2118 | */ | ||
2119 | static void mpc_action_doxid7(fsm_instance *fsm, int event, void *arg) | ||
2120 | { | ||
2121 | struct net_device *dev = arg; | ||
2122 | struct ctcm_priv *priv = NULL; | ||
2123 | struct mpc_group *grp = NULL; | ||
2124 | int direction; | ||
2125 | int rc = 0; | ||
2126 | int send = 0; | ||
2127 | |||
2128 | ctcm_pr_debug("ctcmpc enter: %s() \n", __FUNCTION__); | ||
2129 | |||
2130 | if (dev == NULL) { | ||
2131 | printk(KERN_INFO "%s dev=NULL \n", __FUNCTION__); | ||
2132 | rc = 1; | ||
2133 | goto done; | ||
2134 | } | ||
2135 | |||
2136 | priv = dev->priv; | ||
2137 | if (priv == NULL) { | ||
2138 | printk(KERN_INFO "%s priv=NULL \n", __FUNCTION__); | ||
2139 | rc = 1; | ||
2140 | goto done; | ||
2141 | } | ||
2142 | |||
2143 | grp = priv->mpcg; | ||
2144 | if (grp == NULL) { | ||
2145 | printk(KERN_INFO "%s grp=NULL \n", __FUNCTION__); | ||
2146 | rc = 1; | ||
2147 | goto done; | ||
2148 | } | ||
2149 | |||
2150 | for (direction = READ; direction <= WRITE; direction++) { | ||
2151 | struct channel *ch = priv->channel[direction]; | ||
2152 | struct xid2 *thisxid = ch->xid; | ||
2153 | ch->xid_skb->data = ch->xid_skb_data; | ||
2154 | skb_reset_tail_pointer(ch->xid_skb); | ||
2155 | ch->xid_skb->len = 0; | ||
2156 | thisxid->xid2_option = XID2_7; | ||
2157 | send = 0; | ||
2158 | |||
2159 | /* xid7 phase 1 */ | ||
2160 | if (grp->outstanding_xid7_p2 > 0) { | ||
2161 | if (grp->roll == YSIDE) { | ||
2162 | if (fsm_getstate(ch->fsm) == CH_XID7_PENDING1) { | ||
2163 | fsm_newstate(ch->fsm, CH_XID7_PENDING2); | ||
2164 | ch->ccw[8].cmd_code = CCW_CMD_SENSE_CMD; | ||
2165 | memcpy(skb_put(ch->xid_skb, | ||
2166 | TH_HEADER_LENGTH), | ||
2167 | &thdummy, TH_HEADER_LENGTH); | ||
2168 | send = 1; | ||
2169 | } | ||
2170 | } else if (fsm_getstate(ch->fsm) < CH_XID7_PENDING2) { | ||
2171 | fsm_newstate(ch->fsm, CH_XID7_PENDING2); | ||
2172 | ch->ccw[8].cmd_code = CCW_CMD_WRITE_CTL; | ||
2173 | memcpy(skb_put(ch->xid_skb, | ||
2174 | TH_HEADER_LENGTH), | ||
2175 | &thnorm, TH_HEADER_LENGTH); | ||
2176 | send = 1; | ||
2177 | } | ||
2178 | } else { | ||
2179 | /* xid7 phase 2 */ | ||
2180 | if (grp->roll == YSIDE) { | ||
2181 | if (fsm_getstate(ch->fsm) < CH_XID7_PENDING4) { | ||
2182 | fsm_newstate(ch->fsm, CH_XID7_PENDING4); | ||
2183 | memcpy(skb_put(ch->xid_skb, | ||
2184 | TH_HEADER_LENGTH), | ||
2185 | &thnorm, TH_HEADER_LENGTH); | ||
2186 | ch->ccw[8].cmd_code = CCW_CMD_WRITE_CTL; | ||
2187 | send = 1; | ||
2188 | } | ||
2189 | } else if (fsm_getstate(ch->fsm) == CH_XID7_PENDING3) { | ||
2190 | fsm_newstate(ch->fsm, CH_XID7_PENDING4); | ||
2191 | ch->ccw[8].cmd_code = CCW_CMD_SENSE_CMD; | ||
2192 | memcpy(skb_put(ch->xid_skb, TH_HEADER_LENGTH), | ||
2193 | &thdummy, TH_HEADER_LENGTH); | ||
2194 | send = 1; | ||
2195 | } | ||
2196 | } | ||
2197 | |||
2198 | if (send) | ||
2199 | fsm_event(grp->fsm, MPCG_EVENT_DOIO, ch); | ||
2200 | } | ||
2201 | |||
2202 | done: | ||
2203 | |||
2204 | if (rc != 0) | ||
2205 | fsm_event(grp->fsm, MPCG_EVENT_INOP, dev); | ||
2206 | |||
2207 | return; | ||
2208 | } | ||
2209 | |||
2210 | /* | ||
2211 | * MPC Group Station FSM action | ||
2212 | * CTCM_PROTO_MPC only | ||
2213 | */ | ||
2214 | static void mpc_action_rcvd_xid0(fsm_instance *fsm, int event, void *arg) | ||
2215 | { | ||
2216 | |||
2217 | struct mpcg_info *mpcginfo = arg; | ||
2218 | struct channel *ch = mpcginfo->ch; | ||
2219 | struct net_device *dev = ch->netdev; | ||
2220 | struct ctcm_priv *priv; | ||
2221 | struct mpc_group *grp; | ||
2222 | |||
2223 | if (do_debug) | ||
2224 | ctcm_pr_debug("ctcmpc enter: %s(): cp=%i ch=0x%p id=%s\n", | ||
2225 | __FUNCTION__, smp_processor_id(), ch, ch->id); | ||
2226 | |||
2227 | priv = dev->priv; | ||
2228 | grp = priv->mpcg; | ||
2229 | |||
2230 | ctcm_pr_debug("ctcmpc in:%s() %s xid2:%i xid7:%i xidt_p2:%i \n", | ||
2231 | __FUNCTION__, ch->id, | ||
2232 | grp->outstanding_xid2, | ||
2233 | grp->outstanding_xid7, | ||
2234 | grp->outstanding_xid7_p2); | ||
2235 | |||
2236 | if (fsm_getstate(ch->fsm) < CH_XID7_PENDING) | ||
2237 | fsm_newstate(ch->fsm, CH_XID7_PENDING); | ||
2238 | |||
2239 | grp->outstanding_xid2--; | ||
2240 | grp->outstanding_xid7++; | ||
2241 | grp->outstanding_xid7_p2++; | ||
2242 | |||
2243 | /* must change state before validating xid to */ | ||
2244 | /* properly handle interim interrupts received*/ | ||
2245 | switch (fsm_getstate(grp->fsm)) { | ||
2246 | case MPCG_STATE_XID2INITW: | ||
2247 | fsm_newstate(grp->fsm, MPCG_STATE_XID2INITX); | ||
2248 | mpc_validate_xid(mpcginfo); | ||
2249 | break; | ||
2250 | case MPCG_STATE_XID0IOWAIT: | ||
2251 | fsm_newstate(grp->fsm, MPCG_STATE_XID0IOWAIX); | ||
2252 | mpc_validate_xid(mpcginfo); | ||
2253 | break; | ||
2254 | case MPCG_STATE_XID2INITX: | ||
2255 | if (grp->outstanding_xid2 == 0) { | ||
2256 | fsm_newstate(grp->fsm, MPCG_STATE_XID7INITW); | ||
2257 | mpc_validate_xid(mpcginfo); | ||
2258 | fsm_event(grp->fsm, MPCG_EVENT_XID2DONE, dev); | ||
2259 | } | ||
2260 | break; | ||
2261 | case MPCG_STATE_XID0IOWAIX: | ||
2262 | if (grp->outstanding_xid2 == 0) { | ||
2263 | fsm_newstate(grp->fsm, MPCG_STATE_XID7INITI); | ||
2264 | mpc_validate_xid(mpcginfo); | ||
2265 | fsm_event(grp->fsm, MPCG_EVENT_XID2DONE, dev); | ||
2266 | } | ||
2267 | break; | ||
2268 | } | ||
2269 | kfree(mpcginfo); | ||
2270 | |||
2271 | if (do_debug) { | ||
2272 | ctcm_pr_debug("ctcmpc:%s() %s xid2:%i xid7:%i xidt_p2:%i \n", | ||
2273 | __FUNCTION__, ch->id, | ||
2274 | grp->outstanding_xid2, | ||
2275 | grp->outstanding_xid7, | ||
2276 | grp->outstanding_xid7_p2); | ||
2277 | ctcm_pr_debug("ctcmpc:%s() %s grpstate: %s chanstate: %s \n", | ||
2278 | __FUNCTION__, ch->id, | ||
2279 | fsm_getstate_str(grp->fsm), | ||
2280 | fsm_getstate_str(ch->fsm)); | ||
2281 | } | ||
2282 | return; | ||
2283 | |||
2284 | } | ||
2285 | |||
2286 | |||
2287 | /* | ||
2288 | * MPC Group Station FSM action | ||
2289 | * CTCM_PROTO_MPC only | ||
2290 | */ | ||
2291 | static void mpc_action_rcvd_xid7(fsm_instance *fsm, int event, void *arg) | ||
2292 | { | ||
2293 | struct mpcg_info *mpcginfo = arg; | ||
2294 | struct channel *ch = mpcginfo->ch; | ||
2295 | struct net_device *dev = ch->netdev; | ||
2296 | struct ctcm_priv *priv = dev->priv; | ||
2297 | struct mpc_group *grp = priv->mpcg; | ||
2298 | |||
2299 | if (do_debug) { | ||
2300 | ctcm_pr_debug("ctcmpc enter: %s(): cp=%i ch=0x%p id=%s\n", | ||
2301 | __FUNCTION__, smp_processor_id(), ch, ch->id); | ||
2302 | |||
2303 | ctcm_pr_debug("ctcmpc: outstanding_xid7: %i, " | ||
2304 | " outstanding_xid7_p2: %i\n", | ||
2305 | grp->outstanding_xid7, | ||
2306 | grp->outstanding_xid7_p2); | ||
2307 | } | ||
2308 | |||
2309 | grp->outstanding_xid7--; | ||
2310 | ch->xid_skb->data = ch->xid_skb_data; | ||
2311 | skb_reset_tail_pointer(ch->xid_skb); | ||
2312 | ch->xid_skb->len = 0; | ||
2313 | |||
2314 | switch (fsm_getstate(grp->fsm)) { | ||
2315 | case MPCG_STATE_XID7INITI: | ||
2316 | fsm_newstate(grp->fsm, MPCG_STATE_XID7INITZ); | ||
2317 | mpc_validate_xid(mpcginfo); | ||
2318 | break; | ||
2319 | case MPCG_STATE_XID7INITW: | ||
2320 | fsm_newstate(grp->fsm, MPCG_STATE_XID7INITX); | ||
2321 | mpc_validate_xid(mpcginfo); | ||
2322 | break; | ||
2323 | case MPCG_STATE_XID7INITZ: | ||
2324 | case MPCG_STATE_XID7INITX: | ||
2325 | if (grp->outstanding_xid7 == 0) { | ||
2326 | if (grp->outstanding_xid7_p2 > 0) { | ||
2327 | grp->outstanding_xid7 = | ||
2328 | grp->outstanding_xid7_p2; | ||
2329 | grp->outstanding_xid7_p2 = 0; | ||
2330 | } else | ||
2331 | fsm_newstate(grp->fsm, MPCG_STATE_XID7INITF); | ||
2332 | |||
2333 | mpc_validate_xid(mpcginfo); | ||
2334 | fsm_event(grp->fsm, MPCG_EVENT_XID7DONE, dev); | ||
2335 | break; | ||
2336 | } | ||
2337 | mpc_validate_xid(mpcginfo); | ||
2338 | break; | ||
2339 | } | ||
2340 | |||
2341 | kfree(mpcginfo); | ||
2342 | |||
2343 | if (do_debug) | ||
2344 | ctcm_pr_debug("ctcmpc exit: %s(): cp=%i ch=0x%p id=%s\n", | ||
2345 | __FUNCTION__, smp_processor_id(), ch, ch->id); | ||
2346 | return; | ||
2347 | |||
2348 | } | ||
2349 | |||
2350 | /* | ||
2351 | * mpc_action helper of an MPC Group Station FSM action | ||
2352 | * CTCM_PROTO_MPC only | ||
2353 | */ | ||
2354 | static int mpc_send_qllc_discontact(struct net_device *dev) | ||
2355 | { | ||
2356 | int rc = 0; | ||
2357 | __u32 new_len = 0; | ||
2358 | struct sk_buff *skb; | ||
2359 | struct qllc *qllcptr; | ||
2360 | struct ctcm_priv *priv; | ||
2361 | struct mpc_group *grp; | ||
2362 | |||
2363 | ctcm_pr_debug("ctcmpc enter: %s()\n", __FUNCTION__); | ||
2364 | |||
2365 | if (dev == NULL) { | ||
2366 | printk(KERN_INFO "%s() dev=NULL\n", __FUNCTION__); | ||
2367 | rc = 1; | ||
2368 | goto done; | ||
2369 | } | ||
2370 | |||
2371 | priv = dev->priv; | ||
2372 | if (priv == NULL) { | ||
2373 | printk(KERN_INFO "%s() priv=NULL\n", __FUNCTION__); | ||
2374 | rc = 1; | ||
2375 | goto done; | ||
2376 | } | ||
2377 | |||
2378 | grp = priv->mpcg; | ||
2379 | if (grp == NULL) { | ||
2380 | printk(KERN_INFO "%s() grp=NULL\n", __FUNCTION__); | ||
2381 | rc = 1; | ||
2382 | goto done; | ||
2383 | } | ||
2384 | ctcm_pr_info("ctcmpc: %s() GROUP STATE: %s\n", __FUNCTION__, | ||
2385 | mpcg_state_names[grp->saved_state]); | ||
2386 | |||
2387 | switch (grp->saved_state) { | ||
2388 | /* | ||
2389 | * establish conn callback function is | ||
2390 | * preferred method to report failure | ||
2391 | */ | ||
2392 | case MPCG_STATE_XID0IOWAIT: | ||
2393 | case MPCG_STATE_XID0IOWAIX: | ||
2394 | case MPCG_STATE_XID7INITI: | ||
2395 | case MPCG_STATE_XID7INITZ: | ||
2396 | case MPCG_STATE_XID2INITW: | ||
2397 | case MPCG_STATE_XID2INITX: | ||
2398 | case MPCG_STATE_XID7INITW: | ||
2399 | case MPCG_STATE_XID7INITX: | ||
2400 | if (grp->estconnfunc) { | ||
2401 | grp->estconnfunc(grp->port_num, -1, 0); | ||
2402 | grp->estconnfunc = NULL; | ||
2403 | break; | ||
2404 | } | ||
2405 | case MPCG_STATE_FLOWC: | ||
2406 | case MPCG_STATE_READY: | ||
2407 | grp->send_qllc_disc = 2; | ||
2408 | new_len = sizeof(struct qllc); | ||
2409 | qllcptr = kzalloc(new_len, gfp_type() | GFP_DMA); | ||
2410 | if (qllcptr == NULL) { | ||
2411 | printk(KERN_INFO | ||
2412 | "ctcmpc: Out of memory in %s()\n", | ||
2413 | dev->name); | ||
2414 | rc = 1; | ||
2415 | goto done; | ||
2416 | } | ||
2417 | |||
2418 | qllcptr->qllc_address = 0xcc; | ||
2419 | qllcptr->qllc_commands = 0x03; | ||
2420 | |||
2421 | skb = __dev_alloc_skb(new_len, GFP_ATOMIC); | ||
2422 | |||
2423 | if (skb == NULL) { | ||
2424 | printk(KERN_INFO "%s Out of memory in mpc_send_qllc\n", | ||
2425 | dev->name); | ||
2426 | priv->stats.rx_dropped++; | ||
2427 | rc = 1; | ||
2428 | kfree(qllcptr); | ||
2429 | goto done; | ||
2430 | } | ||
2431 | |||
2432 | memcpy(skb_put(skb, new_len), qllcptr, new_len); | ||
2433 | kfree(qllcptr); | ||
2434 | |||
2435 | if (skb_headroom(skb) < 4) { | ||
2436 | printk(KERN_INFO "ctcmpc: %s() Unable to" | ||
2437 | " build discontact for %s\n", | ||
2438 | __FUNCTION__, dev->name); | ||
2439 | rc = 1; | ||
2440 | dev_kfree_skb_any(skb); | ||
2441 | goto done; | ||
2442 | } | ||
2443 | |||
2444 | *((__u32 *)skb_push(skb, 4)) = priv->channel[READ]->pdu_seq; | ||
2445 | priv->channel[READ]->pdu_seq++; | ||
2446 | if (do_debug_data) | ||
2447 | ctcm_pr_debug("ctcmpc: %s ToDCM_pdu_seq= %08x\n", | ||
2448 | __FUNCTION__, priv->channel[READ]->pdu_seq); | ||
2449 | |||
2450 | /* receipt of CC03 resets anticipated sequence number on | ||
2451 | receiving side */ | ||
2452 | priv->channel[READ]->pdu_seq = 0x00; | ||
2453 | skb_reset_mac_header(skb); | ||
2454 | skb->dev = dev; | ||
2455 | skb->protocol = htons(ETH_P_SNAP); | ||
2456 | skb->ip_summed = CHECKSUM_UNNECESSARY; | ||
2457 | |||
2458 | ctcmpc_dumpit((char *)skb->data, (sizeof(struct qllc) + 4)); | ||
2459 | |||
2460 | netif_rx(skb); | ||
2461 | break; | ||
2462 | default: | ||
2463 | break; | ||
2464 | |||
2465 | } | ||
2466 | |||
2467 | done: | ||
2468 | ctcm_pr_debug("ctcmpc exit: %s()\n", __FUNCTION__); | ||
2469 | return rc; | ||
2470 | } | ||
2471 | /* --- This is the END my friend --- */ | ||
2472 | |||
diff --git a/drivers/s390/net/ctcm_mpc.h b/drivers/s390/net/ctcm_mpc.h new file mode 100644 index 000000000000..f99686069a91 --- /dev/null +++ b/drivers/s390/net/ctcm_mpc.h | |||
@@ -0,0 +1,239 @@ | |||
1 | /* | ||
2 | * drivers/s390/net/ctcm_mpc.h | ||
3 | * | ||
4 | * Copyright IBM Corp. 2007 | ||
5 | * Authors: Peter Tiedemann (ptiedem@de.ibm.com) | ||
6 | * | ||
7 | * MPC additions: | ||
8 | * Belinda Thompson (belindat@us.ibm.com) | ||
9 | * Andy Richter (richtera@us.ibm.com) | ||
10 | */ | ||
11 | |||
12 | #ifndef _CTC_MPC_H_ | ||
13 | #define _CTC_MPC_H_ | ||
14 | |||
15 | #include <linux/skbuff.h> | ||
16 | #include "fsm.h" | ||
17 | |||
18 | /* | ||
19 | * MPC external interface | ||
20 | * Note that ctc_mpc_xyz are called with a lock on ................ | ||
21 | */ | ||
22 | |||
23 | /* port_number is the mpc device 0, 1, 2 etc mpc2 is port_number 2 */ | ||
24 | |||
25 | /* passive open Just wait for XID2 exchange */ | ||
26 | extern int ctc_mpc_alloc_channel(int port, | ||
27 | void (*callback)(int port_num, int max_write_size)); | ||
28 | /* active open Alloc then send XID2 */ | ||
29 | extern void ctc_mpc_establish_connectivity(int port, | ||
30 | void (*callback)(int port_num, int rc, int max_write_size)); | ||
31 | |||
32 | extern void ctc_mpc_dealloc_ch(int port); | ||
33 | extern void ctc_mpc_flow_control(int port, int flowc); | ||
34 | |||
35 | /* | ||
36 | * other MPC Group prototypes and structures | ||
37 | */ | ||
38 | |||
39 | #define ETH_P_SNA_DIX 0x80D5 | ||
40 | |||
41 | /* | ||
42 | * Declaration of an XID2 | ||
43 | * | ||
44 | */ | ||
45 | #define ALLZEROS 0x0000000000000000 | ||
46 | |||
47 | #define XID_FM2 0x20 | ||
48 | #define XID2_0 0x00 | ||
49 | #define XID2_7 0x07 | ||
50 | #define XID2_WRITE_SIDE 0x04 | ||
51 | #define XID2_READ_SIDE 0x05 | ||
52 | |||
53 | struct xid2 { | ||
54 | __u8 xid2_type_id; | ||
55 | __u8 xid2_len; | ||
56 | __u32 xid2_adj_id; | ||
57 | __u8 xid2_rlen; | ||
58 | __u8 xid2_resv1; | ||
59 | __u8 xid2_flag1; | ||
60 | __u8 xid2_fmtt; | ||
61 | __u8 xid2_flag4; | ||
62 | __u16 xid2_resv2; | ||
63 | __u8 xid2_tgnum; | ||
64 | __u32 xid2_sender_id; | ||
65 | __u8 xid2_flag2; | ||
66 | __u8 xid2_option; | ||
67 | char xid2_resv3[8]; | ||
68 | __u16 xid2_resv4; | ||
69 | __u8 xid2_dlc_type; | ||
70 | __u16 xid2_resv5; | ||
71 | __u8 xid2_mpc_flag; | ||
72 | __u8 xid2_resv6; | ||
73 | __u16 xid2_buf_len; | ||
74 | char xid2_buffer[255 - (13 * sizeof(__u8) + | ||
75 | 2 * sizeof(__u32) + | ||
76 | 4 * sizeof(__u16) + | ||
77 | 8 * sizeof(char))]; | ||
78 | } __attribute__ ((packed)); | ||
79 | |||
80 | #define XID2_LENGTH (sizeof(struct xid2)) | ||
81 | |||
82 | struct th_header { | ||
83 | __u8 th_seg; | ||
84 | __u8 th_ch_flag; | ||
85 | #define TH_HAS_PDU 0xf0 | ||
86 | #define TH_IS_XID 0x01 | ||
87 | #define TH_SWEEP_REQ 0xfe | ||
88 | #define TH_SWEEP_RESP 0xff | ||
89 | __u8 th_blk_flag; | ||
90 | #define TH_DATA_IS_XID 0x80 | ||
91 | #define TH_RETRY 0x40 | ||
92 | #define TH_DISCONTACT 0xc0 | ||
93 | #define TH_SEG_BLK 0x20 | ||
94 | #define TH_LAST_SEG 0x10 | ||
95 | #define TH_PDU_PART 0x08 | ||
96 | __u8 th_is_xid; /* is 0x01 if this is XID */ | ||
97 | __u32 th_seq_num; | ||
98 | } __attribute__ ((packed)); | ||
99 | |||
100 | struct th_addon { | ||
101 | __u32 th_last_seq; | ||
102 | __u32 th_resvd; | ||
103 | } __attribute__ ((packed)); | ||
104 | |||
105 | struct th_sweep { | ||
106 | struct th_header th; | ||
107 | struct th_addon sw; | ||
108 | } __attribute__ ((packed)); | ||
109 | |||
110 | #define TH_HEADER_LENGTH (sizeof(struct th_header)) | ||
111 | #define TH_SWEEP_LENGTH (sizeof(struct th_sweep)) | ||
112 | |||
113 | #define PDU_LAST 0x80 | ||
114 | #define PDU_CNTL 0x40 | ||
115 | #define PDU_FIRST 0x20 | ||
116 | |||
117 | struct pdu { | ||
118 | __u32 pdu_offset; | ||
119 | __u8 pdu_flag; | ||
120 | __u8 pdu_proto; /* 0x01 is APPN SNA */ | ||
121 | __u16 pdu_seq; | ||
122 | } __attribute__ ((packed)); | ||
123 | |||
124 | #define PDU_HEADER_LENGTH (sizeof(struct pdu)) | ||
125 | |||
126 | struct qllc { | ||
127 | __u8 qllc_address; | ||
128 | #define QLLC_REQ 0xFF | ||
129 | #define QLLC_RESP 0x00 | ||
130 | __u8 qllc_commands; | ||
131 | #define QLLC_DISCONNECT 0x53 | ||
132 | #define QLLC_UNSEQACK 0x73 | ||
133 | #define QLLC_SETMODE 0x93 | ||
134 | #define QLLC_EXCHID 0xBF | ||
135 | } __attribute__ ((packed)); | ||
136 | |||
137 | |||
138 | /* | ||
139 | * Definition of one MPC group | ||
140 | */ | ||
141 | |||
142 | #define MAX_MPCGCHAN 10 | ||
143 | #define MPC_XID_TIMEOUT_VALUE 10000 | ||
144 | #define MPC_CHANNEL_ADD 0 | ||
145 | #define MPC_CHANNEL_REMOVE 1 | ||
146 | #define MPC_CHANNEL_ATTN 2 | ||
147 | #define XSIDE 1 | ||
148 | #define YSIDE 0 | ||
149 | |||
150 | struct mpcg_info { | ||
151 | struct sk_buff *skb; | ||
152 | struct channel *ch; | ||
153 | struct xid2 *xid; | ||
154 | struct th_sweep *sweep; | ||
155 | struct th_header *th; | ||
156 | }; | ||
157 | |||
158 | struct mpc_group { | ||
159 | struct tasklet_struct mpc_tasklet; | ||
160 | struct tasklet_struct mpc_tasklet2; | ||
161 | int changed_side; | ||
162 | int saved_state; | ||
163 | int channels_terminating; | ||
164 | int out_of_sequence; | ||
165 | int flow_off_called; | ||
166 | int port_num; | ||
167 | int port_persist; | ||
168 | int alloc_called; | ||
169 | __u32 xid2_adj_id; | ||
170 | __u8 xid2_tgnum; | ||
171 | __u32 xid2_sender_id; | ||
172 | int num_channel_paths; | ||
173 | int active_channels[2]; | ||
174 | __u16 group_max_buflen; | ||
175 | int outstanding_xid2; | ||
176 | int outstanding_xid7; | ||
177 | int outstanding_xid7_p2; | ||
178 | int sweep_req_pend_num; | ||
179 | int sweep_rsp_pend_num; | ||
180 | struct sk_buff *xid_skb; | ||
181 | char *xid_skb_data; | ||
182 | struct th_header *xid_th; | ||
183 | struct xid2 *xid; | ||
184 | char *xid_id; | ||
185 | struct th_header *rcvd_xid_th; | ||
186 | struct sk_buff *rcvd_xid_skb; | ||
187 | char *rcvd_xid_data; | ||
188 | __u8 in_sweep; | ||
189 | __u8 roll; | ||
190 | struct xid2 *saved_xid2; | ||
191 | void (*allochanfunc)(int, int); | ||
192 | int allocchan_callback_retries; | ||
193 | void (*estconnfunc)(int, int, int); | ||
194 | int estconn_callback_retries; | ||
195 | int estconn_called; | ||
196 | int xidnogood; | ||
197 | int send_qllc_disc; | ||
198 | fsm_timer timer; | ||
199 | fsm_instance *fsm; /* group xid fsm */ | ||
200 | }; | ||
201 | |||
202 | #ifdef DEBUGDATA | ||
203 | void ctcmpc_dumpit(char *buf, int len); | ||
204 | #else | ||
205 | static inline void ctcmpc_dumpit(char *buf, int len) | ||
206 | { | ||
207 | } | ||
208 | #endif | ||
209 | |||
210 | #ifdef DEBUGDATA | ||
211 | /* | ||
212 | * Dump header and first 16 bytes of an sk_buff for debugging purposes. | ||
213 | * | ||
214 | * skb The struct sk_buff to dump. | ||
215 | * offset Offset relative to skb-data, where to start the dump. | ||
216 | */ | ||
217 | void ctcmpc_dump_skb(struct sk_buff *skb, int offset); | ||
218 | #else | ||
219 | static inline void ctcmpc_dump_skb(struct sk_buff *skb, int offset) | ||
220 | {} | ||
221 | #endif | ||
222 | |||
223 | static inline void ctcmpc_dump32(char *buf, int len) | ||
224 | { | ||
225 | if (len < 32) | ||
226 | ctcmpc_dumpit(buf, len); | ||
227 | else | ||
228 | ctcmpc_dumpit(buf, 32); | ||
229 | } | ||
230 | |||
231 | int ctcmpc_open(struct net_device *); | ||
232 | void ctcm_ccw_check_rc(struct channel *, int, char *); | ||
233 | void mpc_group_ready(unsigned long adev); | ||
234 | int mpc_channel_action(struct channel *ch, int direction, int action); | ||
235 | void mpc_action_send_discontact(unsigned long thischan); | ||
236 | void mpc_action_discontact(fsm_instance *fi, int event, void *arg); | ||
237 | void ctcmpc_bh(unsigned long thischan); | ||
238 | #endif | ||
239 | /* --- This is the END my friend --- */ | ||
diff --git a/drivers/s390/net/ctcm_sysfs.c b/drivers/s390/net/ctcm_sysfs.c new file mode 100644 index 000000000000..bb2d13721d34 --- /dev/null +++ b/drivers/s390/net/ctcm_sysfs.c | |||
@@ -0,0 +1,210 @@ | |||
1 | /* | ||
2 | * drivers/s390/net/ctcm_sysfs.c | ||
3 | * | ||
4 | * Copyright IBM Corp. 2007, 2007 | ||
5 | * Authors: Peter Tiedemann (ptiedem@de.ibm.com) | ||
6 | * | ||
7 | */ | ||
8 | |||
9 | #undef DEBUG | ||
10 | #undef DEBUGDATA | ||
11 | #undef DEBUGCCW | ||
12 | |||
13 | #include <linux/sysfs.h> | ||
14 | #include "ctcm_main.h" | ||
15 | |||
16 | /* | ||
17 | * sysfs attributes | ||
18 | */ | ||
19 | |||
20 | static ssize_t ctcm_buffer_show(struct device *dev, | ||
21 | struct device_attribute *attr, char *buf) | ||
22 | { | ||
23 | struct ctcm_priv *priv = dev_get_drvdata(dev); | ||
24 | |||
25 | if (!priv) | ||
26 | return -ENODEV; | ||
27 | return sprintf(buf, "%d\n", priv->buffer_size); | ||
28 | } | ||
29 | |||
30 | static ssize_t ctcm_buffer_write(struct device *dev, | ||
31 | struct device_attribute *attr, const char *buf, size_t count) | ||
32 | { | ||
33 | struct net_device *ndev; | ||
34 | int bs1; | ||
35 | struct ctcm_priv *priv = dev_get_drvdata(dev); | ||
36 | |||
37 | if (!(priv && priv->channel[READ] && | ||
38 | (ndev = priv->channel[READ]->netdev))) { | ||
39 | CTCM_DBF_TEXT(SETUP, CTC_DBF_ERROR, "bfnondev"); | ||
40 | return -ENODEV; | ||
41 | } | ||
42 | |||
43 | sscanf(buf, "%u", &bs1); | ||
44 | if (bs1 > CTCM_BUFSIZE_LIMIT) | ||
45 | goto einval; | ||
46 | if (bs1 < (576 + LL_HEADER_LENGTH + 2)) | ||
47 | goto einval; | ||
48 | priv->buffer_size = bs1; /* just to overwrite the default */ | ||
49 | |||
50 | if ((ndev->flags & IFF_RUNNING) && | ||
51 | (bs1 < (ndev->mtu + LL_HEADER_LENGTH + 2))) | ||
52 | goto einval; | ||
53 | |||
54 | priv->channel[READ]->max_bufsize = bs1; | ||
55 | priv->channel[WRITE]->max_bufsize = bs1; | ||
56 | if (!(ndev->flags & IFF_RUNNING)) | ||
57 | ndev->mtu = bs1 - LL_HEADER_LENGTH - 2; | ||
58 | priv->channel[READ]->flags |= CHANNEL_FLAGS_BUFSIZE_CHANGED; | ||
59 | priv->channel[WRITE]->flags |= CHANNEL_FLAGS_BUFSIZE_CHANGED; | ||
60 | |||
61 | CTCM_DBF_DEV(SETUP, ndev, buf); | ||
62 | return count; | ||
63 | |||
64 | einval: | ||
65 | CTCM_DBF_DEV(SETUP, ndev, "buff_err"); | ||
66 | return -EINVAL; | ||
67 | } | ||
68 | |||
69 | static void ctcm_print_statistics(struct ctcm_priv *priv) | ||
70 | { | ||
71 | char *sbuf; | ||
72 | char *p; | ||
73 | |||
74 | if (!priv) | ||
75 | return; | ||
76 | sbuf = kmalloc(2048, GFP_KERNEL); | ||
77 | if (sbuf == NULL) | ||
78 | return; | ||
79 | p = sbuf; | ||
80 | |||
81 | p += sprintf(p, " Device FSM state: %s\n", | ||
82 | fsm_getstate_str(priv->fsm)); | ||
83 | p += sprintf(p, " RX channel FSM state: %s\n", | ||
84 | fsm_getstate_str(priv->channel[READ]->fsm)); | ||
85 | p += sprintf(p, " TX channel FSM state: %s\n", | ||
86 | fsm_getstate_str(priv->channel[WRITE]->fsm)); | ||
87 | p += sprintf(p, " Max. TX buffer used: %ld\n", | ||
88 | priv->channel[WRITE]->prof.maxmulti); | ||
89 | p += sprintf(p, " Max. chained SKBs: %ld\n", | ||
90 | priv->channel[WRITE]->prof.maxcqueue); | ||
91 | p += sprintf(p, " TX single write ops: %ld\n", | ||
92 | priv->channel[WRITE]->prof.doios_single); | ||
93 | p += sprintf(p, " TX multi write ops: %ld\n", | ||
94 | priv->channel[WRITE]->prof.doios_multi); | ||
95 | p += sprintf(p, " Netto bytes written: %ld\n", | ||
96 | priv->channel[WRITE]->prof.txlen); | ||
97 | p += sprintf(p, " Max. TX IO-time: %ld\n", | ||
98 | priv->channel[WRITE]->prof.tx_time); | ||
99 | |||
100 | printk(KERN_INFO "Statistics for %s:\n%s", | ||
101 | priv->channel[WRITE]->netdev->name, sbuf); | ||
102 | kfree(sbuf); | ||
103 | return; | ||
104 | } | ||
105 | |||
106 | static ssize_t stats_show(struct device *dev, | ||
107 | struct device_attribute *attr, char *buf) | ||
108 | { | ||
109 | struct ctcm_priv *priv = dev_get_drvdata(dev); | ||
110 | if (!priv) | ||
111 | return -ENODEV; | ||
112 | ctcm_print_statistics(priv); | ||
113 | return sprintf(buf, "0\n"); | ||
114 | } | ||
115 | |||
116 | static ssize_t stats_write(struct device *dev, struct device_attribute *attr, | ||
117 | const char *buf, size_t count) | ||
118 | { | ||
119 | struct ctcm_priv *priv = dev_get_drvdata(dev); | ||
120 | if (!priv) | ||
121 | return -ENODEV; | ||
122 | /* Reset statistics */ | ||
123 | memset(&priv->channel[WRITE]->prof, 0, | ||
124 | sizeof(priv->channel[WRITE]->prof)); | ||
125 | return count; | ||
126 | } | ||
127 | |||
128 | static ssize_t ctcm_proto_show(struct device *dev, | ||
129 | struct device_attribute *attr, char *buf) | ||
130 | { | ||
131 | struct ctcm_priv *priv = dev_get_drvdata(dev); | ||
132 | if (!priv) | ||
133 | return -ENODEV; | ||
134 | |||
135 | return sprintf(buf, "%d\n", priv->protocol); | ||
136 | } | ||
137 | |||
138 | static ssize_t ctcm_proto_store(struct device *dev, | ||
139 | struct device_attribute *attr, const char *buf, size_t count) | ||
140 | { | ||
141 | int value; | ||
142 | struct ctcm_priv *priv = dev_get_drvdata(dev); | ||
143 | |||
144 | if (!priv) | ||
145 | return -ENODEV; | ||
146 | sscanf(buf, "%u", &value); | ||
147 | if (!((value == CTCM_PROTO_S390) || | ||
148 | (value == CTCM_PROTO_LINUX) || | ||
149 | (value == CTCM_PROTO_MPC) || | ||
150 | (value == CTCM_PROTO_OS390))) | ||
151 | return -EINVAL; | ||
152 | priv->protocol = value; | ||
153 | CTCM_DBF_DEV(SETUP, dev, buf); | ||
154 | |||
155 | return count; | ||
156 | } | ||
157 | |||
158 | static ssize_t ctcm_type_show(struct device *dev, | ||
159 | struct device_attribute *attr, char *buf) | ||
160 | { | ||
161 | struct ccwgroup_device *cgdev; | ||
162 | |||
163 | cgdev = to_ccwgroupdev(dev); | ||
164 | if (!cgdev) | ||
165 | return -ENODEV; | ||
166 | |||
167 | return sprintf(buf, "%s\n", | ||
168 | cu3088_type[cgdev->cdev[0]->id.driver_info]); | ||
169 | } | ||
170 | |||
171 | static DEVICE_ATTR(buffer, 0644, ctcm_buffer_show, ctcm_buffer_write); | ||
172 | static DEVICE_ATTR(protocol, 0644, ctcm_proto_show, ctcm_proto_store); | ||
173 | static DEVICE_ATTR(type, 0444, ctcm_type_show, NULL); | ||
174 | static DEVICE_ATTR(stats, 0644, stats_show, stats_write); | ||
175 | |||
176 | static struct attribute *ctcm_attr[] = { | ||
177 | &dev_attr_protocol.attr, | ||
178 | &dev_attr_type.attr, | ||
179 | &dev_attr_buffer.attr, | ||
180 | NULL, | ||
181 | }; | ||
182 | |||
183 | static struct attribute_group ctcm_attr_group = { | ||
184 | .attrs = ctcm_attr, | ||
185 | }; | ||
186 | |||
187 | int ctcm_add_attributes(struct device *dev) | ||
188 | { | ||
189 | int rc; | ||
190 | |||
191 | rc = device_create_file(dev, &dev_attr_stats); | ||
192 | |||
193 | return rc; | ||
194 | } | ||
195 | |||
196 | void ctcm_remove_attributes(struct device *dev) | ||
197 | { | ||
198 | device_remove_file(dev, &dev_attr_stats); | ||
199 | } | ||
200 | |||
201 | int ctcm_add_files(struct device *dev) | ||
202 | { | ||
203 | return sysfs_create_group(&dev->kobj, &ctcm_attr_group); | ||
204 | } | ||
205 | |||
206 | void ctcm_remove_files(struct device *dev) | ||
207 | { | ||
208 | sysfs_remove_group(&dev->kobj, &ctcm_attr_group); | ||
209 | } | ||
210 | |||
diff --git a/drivers/s390/net/ctcmain.c b/drivers/s390/net/ctcmain.c deleted file mode 100644 index 77a503139e32..000000000000 --- a/drivers/s390/net/ctcmain.c +++ /dev/null | |||
@@ -1,3062 +0,0 @@ | |||
1 | /* | ||
2 | * CTC / ESCON network driver | ||
3 | * | ||
4 | * Copyright (C) 2001 IBM Deutschland Entwicklung GmbH, IBM Corporation | ||
5 | * Author(s): Fritz Elfert (elfert@de.ibm.com, felfert@millenux.com) | ||
6 | * Fixes by : Jochen Röhrig (roehrig@de.ibm.com) | ||
7 | * Arnaldo Carvalho de Melo <acme@conectiva.com.br> | ||
8 | Peter Tiedemann (ptiedem@de.ibm.com) | ||
9 | * Driver Model stuff by : Cornelia Huck <cornelia.huck@de.ibm.com> | ||
10 | * | ||
11 | * Documentation used: | ||
12 | * - Principles of Operation (IBM doc#: SA22-7201-06) | ||
13 | * - Common IO/-Device Commands and Self Description (IBM doc#: SA22-7204-02) | ||
14 | * - Common IO/-Device Commands and Self Description (IBM doc#: SN22-5535) | ||
15 | * - ESCON Channel-to-Channel Adapter (IBM doc#: SA22-7203-00) | ||
16 | * - ESCON I/O Interface (IBM doc#: SA22-7202-029 | ||
17 | * | ||
18 | * and the source of the original CTC driver by: | ||
19 | * Dieter Wellerdiek (wel@de.ibm.com) | ||
20 | * Martin Schwidefsky (schwidefsky@de.ibm.com) | ||
21 | * Denis Joseph Barrow (djbarrow@de.ibm.com,barrow_dj@yahoo.com) | ||
22 | * Jochen Röhrig (roehrig@de.ibm.com) | ||
23 | * | ||
24 | * This program is free software; you can redistribute it and/or modify | ||
25 | * it under the terms of the GNU General Public License as published by | ||
26 | * the Free Software Foundation; either version 2, or (at your option) | ||
27 | * any later version. | ||
28 | * | ||
29 | * This program is distributed in the hope that it will be useful, | ||
30 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | ||
31 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | ||
32 | * GNU General Public License for more details. | ||
33 | * | ||
34 | * You should have received a copy of the GNU General Public License | ||
35 | * along with this program; if not, write to the Free Software | ||
36 | * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. | ||
37 | * | ||
38 | */ | ||
39 | #undef DEBUG | ||
40 | #include <linux/module.h> | ||
41 | #include <linux/init.h> | ||
42 | #include <linux/kernel.h> | ||
43 | #include <linux/slab.h> | ||
44 | #include <linux/errno.h> | ||
45 | #include <linux/types.h> | ||
46 | #include <linux/interrupt.h> | ||
47 | #include <linux/timer.h> | ||
48 | #include <linux/bitops.h> | ||
49 | |||
50 | #include <linux/signal.h> | ||
51 | #include <linux/string.h> | ||
52 | |||
53 | #include <linux/ip.h> | ||
54 | #include <linux/if_arp.h> | ||
55 | #include <linux/tcp.h> | ||
56 | #include <linux/skbuff.h> | ||
57 | #include <linux/ctype.h> | ||
58 | #include <net/dst.h> | ||
59 | |||
60 | #include <asm/io.h> | ||
61 | #include <asm/ccwdev.h> | ||
62 | #include <asm/ccwgroup.h> | ||
63 | #include <asm/uaccess.h> | ||
64 | |||
65 | #include <asm/idals.h> | ||
66 | |||
67 | #include "fsm.h" | ||
68 | #include "cu3088.h" | ||
69 | |||
70 | #include "ctcdbug.h" | ||
71 | #include "ctcmain.h" | ||
72 | |||
73 | MODULE_AUTHOR("(C) 2000 IBM Corp. by Fritz Elfert (felfert@millenux.com)"); | ||
74 | MODULE_DESCRIPTION("Linux for S/390 CTC/Escon Driver"); | ||
75 | MODULE_LICENSE("GPL"); | ||
76 | /** | ||
77 | * States of the interface statemachine. | ||
78 | */ | ||
79 | enum dev_states { | ||
80 | DEV_STATE_STOPPED, | ||
81 | DEV_STATE_STARTWAIT_RXTX, | ||
82 | DEV_STATE_STARTWAIT_RX, | ||
83 | DEV_STATE_STARTWAIT_TX, | ||
84 | DEV_STATE_STOPWAIT_RXTX, | ||
85 | DEV_STATE_STOPWAIT_RX, | ||
86 | DEV_STATE_STOPWAIT_TX, | ||
87 | DEV_STATE_RUNNING, | ||
88 | /** | ||
89 | * MUST be always the last element!! | ||
90 | */ | ||
91 | CTC_NR_DEV_STATES | ||
92 | }; | ||
93 | |||
94 | static const char *dev_state_names[] = { | ||
95 | "Stopped", | ||
96 | "StartWait RXTX", | ||
97 | "StartWait RX", | ||
98 | "StartWait TX", | ||
99 | "StopWait RXTX", | ||
100 | "StopWait RX", | ||
101 | "StopWait TX", | ||
102 | "Running", | ||
103 | }; | ||
104 | |||
105 | /** | ||
106 | * Events of the interface statemachine. | ||
107 | */ | ||
108 | enum dev_events { | ||
109 | DEV_EVENT_START, | ||
110 | DEV_EVENT_STOP, | ||
111 | DEV_EVENT_RXUP, | ||
112 | DEV_EVENT_TXUP, | ||
113 | DEV_EVENT_RXDOWN, | ||
114 | DEV_EVENT_TXDOWN, | ||
115 | DEV_EVENT_RESTART, | ||
116 | /** | ||
117 | * MUST be always the last element!! | ||
118 | */ | ||
119 | CTC_NR_DEV_EVENTS | ||
120 | }; | ||
121 | |||
122 | static const char *dev_event_names[] = { | ||
123 | "Start", | ||
124 | "Stop", | ||
125 | "RX up", | ||
126 | "TX up", | ||
127 | "RX down", | ||
128 | "TX down", | ||
129 | "Restart", | ||
130 | }; | ||
131 | |||
132 | /** | ||
133 | * Events of the channel statemachine | ||
134 | */ | ||
135 | enum ch_events { | ||
136 | /** | ||
137 | * Events, representing return code of | ||
138 | * I/O operations (ccw_device_start, ccw_device_halt et al.) | ||
139 | */ | ||
140 | CH_EVENT_IO_SUCCESS, | ||
141 | CH_EVENT_IO_EBUSY, | ||
142 | CH_EVENT_IO_ENODEV, | ||
143 | CH_EVENT_IO_EIO, | ||
144 | CH_EVENT_IO_UNKNOWN, | ||
145 | |||
146 | CH_EVENT_ATTNBUSY, | ||
147 | CH_EVENT_ATTN, | ||
148 | CH_EVENT_BUSY, | ||
149 | |||
150 | /** | ||
151 | * Events, representing unit-check | ||
152 | */ | ||
153 | CH_EVENT_UC_RCRESET, | ||
154 | CH_EVENT_UC_RSRESET, | ||
155 | CH_EVENT_UC_TXTIMEOUT, | ||
156 | CH_EVENT_UC_TXPARITY, | ||
157 | CH_EVENT_UC_HWFAIL, | ||
158 | CH_EVENT_UC_RXPARITY, | ||
159 | CH_EVENT_UC_ZERO, | ||
160 | CH_EVENT_UC_UNKNOWN, | ||
161 | |||
162 | /** | ||
163 | * Events, representing subchannel-check | ||
164 | */ | ||
165 | CH_EVENT_SC_UNKNOWN, | ||
166 | |||
167 | /** | ||
168 | * Events, representing machine checks | ||
169 | */ | ||
170 | CH_EVENT_MC_FAIL, | ||
171 | CH_EVENT_MC_GOOD, | ||
172 | |||
173 | /** | ||
174 | * Event, representing normal IRQ | ||
175 | */ | ||
176 | CH_EVENT_IRQ, | ||
177 | CH_EVENT_FINSTAT, | ||
178 | |||
179 | /** | ||
180 | * Event, representing timer expiry. | ||
181 | */ | ||
182 | CH_EVENT_TIMER, | ||
183 | |||
184 | /** | ||
185 | * Events, representing commands from upper levels. | ||
186 | */ | ||
187 | CH_EVENT_START, | ||
188 | CH_EVENT_STOP, | ||
189 | |||
190 | /** | ||
191 | * MUST be always the last element!! | ||
192 | */ | ||
193 | NR_CH_EVENTS, | ||
194 | }; | ||
195 | |||
196 | /** | ||
197 | * States of the channel statemachine. | ||
198 | */ | ||
199 | enum ch_states { | ||
200 | /** | ||
201 | * Channel not assigned to any device, | ||
202 | * initial state, direction invalid | ||
203 | */ | ||
204 | CH_STATE_IDLE, | ||
205 | |||
206 | /** | ||
207 | * Channel assigned but not operating | ||
208 | */ | ||
209 | CH_STATE_STOPPED, | ||
210 | CH_STATE_STARTWAIT, | ||
211 | CH_STATE_STARTRETRY, | ||
212 | CH_STATE_SETUPWAIT, | ||
213 | CH_STATE_RXINIT, | ||
214 | CH_STATE_TXINIT, | ||
215 | CH_STATE_RX, | ||
216 | CH_STATE_TX, | ||
217 | CH_STATE_RXIDLE, | ||
218 | CH_STATE_TXIDLE, | ||
219 | CH_STATE_RXERR, | ||
220 | CH_STATE_TXERR, | ||
221 | CH_STATE_TERM, | ||
222 | CH_STATE_DTERM, | ||
223 | CH_STATE_NOTOP, | ||
224 | |||
225 | /** | ||
226 | * MUST be always the last element!! | ||
227 | */ | ||
228 | NR_CH_STATES, | ||
229 | }; | ||
230 | |||
231 | static int loglevel = CTC_LOGLEVEL_DEFAULT; | ||
232 | |||
233 | /** | ||
234 | * Linked list of all detected channels. | ||
235 | */ | ||
236 | static struct channel *channels = NULL; | ||
237 | |||
238 | /** | ||
239 | * Print Banner. | ||
240 | */ | ||
241 | static void | ||
242 | print_banner(void) | ||
243 | { | ||
244 | static int printed = 0; | ||
245 | |||
246 | if (printed) | ||
247 | return; | ||
248 | |||
249 | printk(KERN_INFO "CTC driver initialized\n"); | ||
250 | printed = 1; | ||
251 | } | ||
252 | |||
253 | /** | ||
254 | * Return type of a detected device. | ||
255 | */ | ||
256 | static enum channel_types | ||
257 | get_channel_type(struct ccw_device_id *id) | ||
258 | { | ||
259 | enum channel_types type = (enum channel_types) id->driver_info; | ||
260 | |||
261 | if (type == channel_type_ficon) | ||
262 | type = channel_type_escon; | ||
263 | |||
264 | return type; | ||
265 | } | ||
266 | |||
267 | static const char *ch_event_names[] = { | ||
268 | "ccw_device success", | ||
269 | "ccw_device busy", | ||
270 | "ccw_device enodev", | ||
271 | "ccw_device ioerr", | ||
272 | "ccw_device unknown", | ||
273 | |||
274 | "Status ATTN & BUSY", | ||
275 | "Status ATTN", | ||
276 | "Status BUSY", | ||
277 | |||
278 | "Unit check remote reset", | ||
279 | "Unit check remote system reset", | ||
280 | "Unit check TX timeout", | ||
281 | "Unit check TX parity", | ||
282 | "Unit check Hardware failure", | ||
283 | "Unit check RX parity", | ||
284 | "Unit check ZERO", | ||
285 | "Unit check Unknown", | ||
286 | |||
287 | "SubChannel check Unknown", | ||
288 | |||
289 | "Machine check failure", | ||
290 | "Machine check operational", | ||
291 | |||
292 | "IRQ normal", | ||
293 | "IRQ final", | ||
294 | |||
295 | "Timer", | ||
296 | |||
297 | "Start", | ||
298 | "Stop", | ||
299 | }; | ||
300 | |||
301 | static const char *ch_state_names[] = { | ||
302 | "Idle", | ||
303 | "Stopped", | ||
304 | "StartWait", | ||
305 | "StartRetry", | ||
306 | "SetupWait", | ||
307 | "RX init", | ||
308 | "TX init", | ||
309 | "RX", | ||
310 | "TX", | ||
311 | "RX idle", | ||
312 | "TX idle", | ||
313 | "RX error", | ||
314 | "TX error", | ||
315 | "Terminating", | ||
316 | "Restarting", | ||
317 | "Not operational", | ||
318 | }; | ||
319 | |||
320 | #ifdef DEBUG | ||
321 | /** | ||
322 | * Dump header and first 16 bytes of an sk_buff for debugging purposes. | ||
323 | * | ||
324 | * @param skb The sk_buff to dump. | ||
325 | * @param offset Offset relative to skb-data, where to start the dump. | ||
326 | */ | ||
327 | static void | ||
328 | ctc_dump_skb(struct sk_buff *skb, int offset) | ||
329 | { | ||
330 | unsigned char *p = skb->data; | ||
331 | __u16 bl; | ||
332 | struct ll_header *header; | ||
333 | int i; | ||
334 | |||
335 | if (!(loglevel & CTC_LOGLEVEL_DEBUG)) | ||
336 | return; | ||
337 | p += offset; | ||
338 | bl = *((__u16 *) p); | ||
339 | p += 2; | ||
340 | header = (struct ll_header *) p; | ||
341 | p -= 2; | ||
342 | |||
343 | printk(KERN_DEBUG "dump:\n"); | ||
344 | printk(KERN_DEBUG "blocklen=%d %04x\n", bl, bl); | ||
345 | |||
346 | printk(KERN_DEBUG "h->length=%d %04x\n", header->length, | ||
347 | header->length); | ||
348 | printk(KERN_DEBUG "h->type=%04x\n", header->type); | ||
349 | printk(KERN_DEBUG "h->unused=%04x\n", header->unused); | ||
350 | if (bl > 16) | ||
351 | bl = 16; | ||
352 | printk(KERN_DEBUG "data: "); | ||
353 | for (i = 0; i < bl; i++) | ||
354 | printk("%02x%s", *p++, (i % 16) ? " " : "\n<7>"); | ||
355 | printk("\n"); | ||
356 | } | ||
357 | #else | ||
358 | static inline void | ||
359 | ctc_dump_skb(struct sk_buff *skb, int offset) | ||
360 | { | ||
361 | } | ||
362 | #endif | ||
363 | |||
364 | /** | ||
365 | * Unpack a just received skb and hand it over to | ||
366 | * upper layers. | ||
367 | * | ||
368 | * @param ch The channel where this skb has been received. | ||
369 | * @param pskb The received skb. | ||
370 | */ | ||
371 | static void | ||
372 | ctc_unpack_skb(struct channel *ch, struct sk_buff *pskb) | ||
373 | { | ||
374 | struct net_device *dev = ch->netdev; | ||
375 | struct ctc_priv *privptr = (struct ctc_priv *) dev->priv; | ||
376 | __u16 len = *((__u16 *) pskb->data); | ||
377 | |||
378 | DBF_TEXT(trace, 4, __FUNCTION__); | ||
379 | skb_put(pskb, 2 + LL_HEADER_LENGTH); | ||
380 | skb_pull(pskb, 2); | ||
381 | pskb->dev = dev; | ||
382 | pskb->ip_summed = CHECKSUM_UNNECESSARY; | ||
383 | while (len > 0) { | ||
384 | struct sk_buff *skb; | ||
385 | struct ll_header *header = (struct ll_header *) pskb->data; | ||
386 | |||
387 | skb_pull(pskb, LL_HEADER_LENGTH); | ||
388 | if ((ch->protocol == CTC_PROTO_S390) && | ||
389 | (header->type != ETH_P_IP)) { | ||
390 | |||
391 | #ifndef DEBUG | ||
392 | if (!(ch->logflags & LOG_FLAG_ILLEGALPKT)) { | ||
393 | #endif | ||
394 | /** | ||
395 | * Check packet type only if we stick strictly | ||
396 | * to S/390's protocol of OS390. This only | ||
397 | * supports IP. Otherwise allow any packet | ||
398 | * type. | ||
399 | */ | ||
400 | ctc_pr_warn( | ||
401 | "%s Illegal packet type 0x%04x received, dropping\n", | ||
402 | dev->name, header->type); | ||
403 | ch->logflags |= LOG_FLAG_ILLEGALPKT; | ||
404 | #ifndef DEBUG | ||
405 | } | ||
406 | #endif | ||
407 | #ifdef DEBUG | ||
408 | ctc_dump_skb(pskb, -6); | ||
409 | #endif | ||
410 | privptr->stats.rx_dropped++; | ||
411 | privptr->stats.rx_frame_errors++; | ||
412 | return; | ||
413 | } | ||
414 | pskb->protocol = ntohs(header->type); | ||
415 | if (header->length <= LL_HEADER_LENGTH) { | ||
416 | #ifndef DEBUG | ||
417 | if (!(ch->logflags & LOG_FLAG_ILLEGALSIZE)) { | ||
418 | #endif | ||
419 | ctc_pr_warn( | ||
420 | "%s Illegal packet size %d " | ||
421 | "received (MTU=%d blocklen=%d), " | ||
422 | "dropping\n", dev->name, header->length, | ||
423 | dev->mtu, len); | ||
424 | ch->logflags |= LOG_FLAG_ILLEGALSIZE; | ||
425 | #ifndef DEBUG | ||
426 | } | ||
427 | #endif | ||
428 | #ifdef DEBUG | ||
429 | ctc_dump_skb(pskb, -6); | ||
430 | #endif | ||
431 | privptr->stats.rx_dropped++; | ||
432 | privptr->stats.rx_length_errors++; | ||
433 | return; | ||
434 | } | ||
435 | header->length -= LL_HEADER_LENGTH; | ||
436 | len -= LL_HEADER_LENGTH; | ||
437 | if ((header->length > skb_tailroom(pskb)) || | ||
438 | (header->length > len)) { | ||
439 | #ifndef DEBUG | ||
440 | if (!(ch->logflags & LOG_FLAG_OVERRUN)) { | ||
441 | #endif | ||
442 | ctc_pr_warn( | ||
443 | "%s Illegal packet size %d " | ||
444 | "(beyond the end of received data), " | ||
445 | "dropping\n", dev->name, header->length); | ||
446 | ch->logflags |= LOG_FLAG_OVERRUN; | ||
447 | #ifndef DEBUG | ||
448 | } | ||
449 | #endif | ||
450 | #ifdef DEBUG | ||
451 | ctc_dump_skb(pskb, -6); | ||
452 | #endif | ||
453 | privptr->stats.rx_dropped++; | ||
454 | privptr->stats.rx_length_errors++; | ||
455 | return; | ||
456 | } | ||
457 | skb_put(pskb, header->length); | ||
458 | skb_reset_mac_header(pskb); | ||
459 | len -= header->length; | ||
460 | skb = dev_alloc_skb(pskb->len); | ||
461 | if (!skb) { | ||
462 | #ifndef DEBUG | ||
463 | if (!(ch->logflags & LOG_FLAG_NOMEM)) { | ||
464 | #endif | ||
465 | ctc_pr_warn( | ||
466 | "%s Out of memory in ctc_unpack_skb\n", | ||
467 | dev->name); | ||
468 | ch->logflags |= LOG_FLAG_NOMEM; | ||
469 | #ifndef DEBUG | ||
470 | } | ||
471 | #endif | ||
472 | privptr->stats.rx_dropped++; | ||
473 | return; | ||
474 | } | ||
475 | skb_copy_from_linear_data(pskb, skb_put(skb, pskb->len), | ||
476 | pskb->len); | ||
477 | skb_reset_mac_header(skb); | ||
478 | skb->dev = pskb->dev; | ||
479 | skb->protocol = pskb->protocol; | ||
480 | pskb->ip_summed = CHECKSUM_UNNECESSARY; | ||
481 | /** | ||
482 | * reset logflags | ||
483 | */ | ||
484 | ch->logflags = 0; | ||
485 | privptr->stats.rx_packets++; | ||
486 | privptr->stats.rx_bytes += skb->len; | ||
487 | netif_rx_ni(skb); | ||
488 | dev->last_rx = jiffies; | ||
489 | if (len > 0) { | ||
490 | skb_pull(pskb, header->length); | ||
491 | if (skb_tailroom(pskb) < LL_HEADER_LENGTH) { | ||
492 | #ifndef DEBUG | ||
493 | if (!(ch->logflags & LOG_FLAG_OVERRUN)) { | ||
494 | #endif | ||
495 | ctc_pr_warn( | ||
496 | "%s Overrun in ctc_unpack_skb\n", | ||
497 | dev->name); | ||
498 | ch->logflags |= LOG_FLAG_OVERRUN; | ||
499 | #ifndef DEBUG | ||
500 | } | ||
501 | #endif | ||
502 | return; | ||
503 | } | ||
504 | skb_put(pskb, LL_HEADER_LENGTH); | ||
505 | } | ||
506 | } | ||
507 | } | ||
508 | |||
509 | /** | ||
510 | * Check return code of a preceeding ccw_device call, halt_IO etc... | ||
511 | * | ||
512 | * @param ch The channel, the error belongs to. | ||
513 | * @param return_code The error code to inspect. | ||
514 | */ | ||
515 | static void | ||
516 | ccw_check_return_code(struct channel *ch, int return_code, char *msg) | ||
517 | { | ||
518 | DBF_TEXT(trace, 5, __FUNCTION__); | ||
519 | switch (return_code) { | ||
520 | case 0: | ||
521 | fsm_event(ch->fsm, CH_EVENT_IO_SUCCESS, ch); | ||
522 | break; | ||
523 | case -EBUSY: | ||
524 | ctc_pr_warn("%s (%s): Busy !\n", ch->id, msg); | ||
525 | fsm_event(ch->fsm, CH_EVENT_IO_EBUSY, ch); | ||
526 | break; | ||
527 | case -ENODEV: | ||
528 | ctc_pr_emerg("%s (%s): Invalid device called for IO\n", | ||
529 | ch->id, msg); | ||
530 | fsm_event(ch->fsm, CH_EVENT_IO_ENODEV, ch); | ||
531 | break; | ||
532 | case -EIO: | ||
533 | ctc_pr_emerg("%s (%s): Status pending... \n", | ||
534 | ch->id, msg); | ||
535 | fsm_event(ch->fsm, CH_EVENT_IO_EIO, ch); | ||
536 | break; | ||
537 | default: | ||
538 | ctc_pr_emerg("%s (%s): Unknown error in do_IO %04x\n", | ||
539 | ch->id, msg, return_code); | ||
540 | fsm_event(ch->fsm, CH_EVENT_IO_UNKNOWN, ch); | ||
541 | } | ||
542 | } | ||
543 | |||
544 | /** | ||
545 | * Check sense of a unit check. | ||
546 | * | ||
547 | * @param ch The channel, the sense code belongs to. | ||
548 | * @param sense The sense code to inspect. | ||
549 | */ | ||
550 | static void | ||
551 | ccw_unit_check(struct channel *ch, unsigned char sense) | ||
552 | { | ||
553 | DBF_TEXT(trace, 5, __FUNCTION__); | ||
554 | if (sense & SNS0_INTERVENTION_REQ) { | ||
555 | if (sense & 0x01) { | ||
556 | ctc_pr_debug("%s: Interface disc. or Sel. reset " | ||
557 | "(remote)\n", ch->id); | ||
558 | fsm_event(ch->fsm, CH_EVENT_UC_RCRESET, ch); | ||
559 | } else { | ||
560 | ctc_pr_debug("%s: System reset (remote)\n", ch->id); | ||
561 | fsm_event(ch->fsm, CH_EVENT_UC_RSRESET, ch); | ||
562 | } | ||
563 | } else if (sense & SNS0_EQUIPMENT_CHECK) { | ||
564 | if (sense & SNS0_BUS_OUT_CHECK) { | ||
565 | ctc_pr_warn("%s: Hardware malfunction (remote)\n", | ||
566 | ch->id); | ||
567 | fsm_event(ch->fsm, CH_EVENT_UC_HWFAIL, ch); | ||
568 | } else { | ||
569 | ctc_pr_warn("%s: Read-data parity error (remote)\n", | ||
570 | ch->id); | ||
571 | fsm_event(ch->fsm, CH_EVENT_UC_RXPARITY, ch); | ||
572 | } | ||
573 | } else if (sense & SNS0_BUS_OUT_CHECK) { | ||
574 | if (sense & 0x04) { | ||
575 | ctc_pr_warn("%s: Data-streaming timeout)\n", ch->id); | ||
576 | fsm_event(ch->fsm, CH_EVENT_UC_TXTIMEOUT, ch); | ||
577 | } else { | ||
578 | ctc_pr_warn("%s: Data-transfer parity error\n", ch->id); | ||
579 | fsm_event(ch->fsm, CH_EVENT_UC_TXPARITY, ch); | ||
580 | } | ||
581 | } else if (sense & SNS0_CMD_REJECT) { | ||
582 | ctc_pr_warn("%s: Command reject\n", ch->id); | ||
583 | } else if (sense == 0) { | ||
584 | ctc_pr_debug("%s: Unit check ZERO\n", ch->id); | ||
585 | fsm_event(ch->fsm, CH_EVENT_UC_ZERO, ch); | ||
586 | } else { | ||
587 | ctc_pr_warn("%s: Unit Check with sense code: %02x\n", | ||
588 | ch->id, sense); | ||
589 | fsm_event(ch->fsm, CH_EVENT_UC_UNKNOWN, ch); | ||
590 | } | ||
591 | } | ||
592 | |||
593 | static void | ||
594 | ctc_purge_skb_queue(struct sk_buff_head *q) | ||
595 | { | ||
596 | struct sk_buff *skb; | ||
597 | |||
598 | DBF_TEXT(trace, 5, __FUNCTION__); | ||
599 | |||
600 | while ((skb = skb_dequeue(q))) { | ||
601 | atomic_dec(&skb->users); | ||
602 | dev_kfree_skb_irq(skb); | ||
603 | } | ||
604 | } | ||
605 | |||
606 | static int | ||
607 | ctc_checkalloc_buffer(struct channel *ch, int warn) | ||
608 | { | ||
609 | DBF_TEXT(trace, 5, __FUNCTION__); | ||
610 | if ((ch->trans_skb == NULL) || | ||
611 | (ch->flags & CHANNEL_FLAGS_BUFSIZE_CHANGED)) { | ||
612 | if (ch->trans_skb != NULL) | ||
613 | dev_kfree_skb(ch->trans_skb); | ||
614 | clear_normalized_cda(&ch->ccw[1]); | ||
615 | ch->trans_skb = __dev_alloc_skb(ch->max_bufsize, | ||
616 | GFP_ATOMIC | GFP_DMA); | ||
617 | if (ch->trans_skb == NULL) { | ||
618 | if (warn) | ||
619 | ctc_pr_warn( | ||
620 | "%s: Couldn't alloc %s trans_skb\n", | ||
621 | ch->id, | ||
622 | (CHANNEL_DIRECTION(ch->flags) == READ) ? | ||
623 | "RX" : "TX"); | ||
624 | return -ENOMEM; | ||
625 | } | ||
626 | ch->ccw[1].count = ch->max_bufsize; | ||
627 | if (set_normalized_cda(&ch->ccw[1], ch->trans_skb->data)) { | ||
628 | dev_kfree_skb(ch->trans_skb); | ||
629 | ch->trans_skb = NULL; | ||
630 | if (warn) | ||
631 | ctc_pr_warn( | ||
632 | "%s: set_normalized_cda for %s " | ||
633 | "trans_skb failed, dropping packets\n", | ||
634 | ch->id, | ||
635 | (CHANNEL_DIRECTION(ch->flags) == READ) ? | ||
636 | "RX" : "TX"); | ||
637 | return -ENOMEM; | ||
638 | } | ||
639 | ch->ccw[1].count = 0; | ||
640 | ch->trans_skb_data = ch->trans_skb->data; | ||
641 | ch->flags &= ~CHANNEL_FLAGS_BUFSIZE_CHANGED; | ||
642 | } | ||
643 | return 0; | ||
644 | } | ||
645 | |||
646 | /** | ||
647 | * Dummy NOP action for statemachines | ||
648 | */ | ||
649 | static void | ||
650 | fsm_action_nop(fsm_instance * fi, int event, void *arg) | ||
651 | { | ||
652 | } | ||
653 | |||
654 | /** | ||
655 | * Actions for channel - statemachines. | ||
656 | *****************************************************************************/ | ||
657 | |||
658 | /** | ||
659 | * Normal data has been send. Free the corresponding | ||
660 | * skb (it's in io_queue), reset dev->tbusy and | ||
661 | * revert to idle state. | ||
662 | * | ||
663 | * @param fi An instance of a channel statemachine. | ||
664 | * @param event The event, just happened. | ||
665 | * @param arg Generic pointer, casted from channel * upon call. | ||
666 | */ | ||
667 | static void | ||
668 | ch_action_txdone(fsm_instance * fi, int event, void *arg) | ||
669 | { | ||
670 | struct channel *ch = (struct channel *) arg; | ||
671 | struct net_device *dev = ch->netdev; | ||
672 | struct ctc_priv *privptr = dev->priv; | ||
673 | struct sk_buff *skb; | ||
674 | int first = 1; | ||
675 | int i; | ||
676 | unsigned long duration; | ||
677 | struct timespec done_stamp = current_kernel_time(); | ||
678 | |||
679 | DBF_TEXT(trace, 4, __FUNCTION__); | ||
680 | |||
681 | duration = | ||
682 | (done_stamp.tv_sec - ch->prof.send_stamp.tv_sec) * 1000000 + | ||
683 | (done_stamp.tv_nsec - ch->prof.send_stamp.tv_nsec) / 1000; | ||
684 | if (duration > ch->prof.tx_time) | ||
685 | ch->prof.tx_time = duration; | ||
686 | |||
687 | if (ch->irb->scsw.count != 0) | ||
688 | ctc_pr_debug("%s: TX not complete, remaining %d bytes\n", | ||
689 | dev->name, ch->irb->scsw.count); | ||
690 | fsm_deltimer(&ch->timer); | ||
691 | while ((skb = skb_dequeue(&ch->io_queue))) { | ||
692 | privptr->stats.tx_packets++; | ||
693 | privptr->stats.tx_bytes += skb->len - LL_HEADER_LENGTH; | ||
694 | if (first) { | ||
695 | privptr->stats.tx_bytes += 2; | ||
696 | first = 0; | ||
697 | } | ||
698 | atomic_dec(&skb->users); | ||
699 | dev_kfree_skb_irq(skb); | ||
700 | } | ||
701 | spin_lock(&ch->collect_lock); | ||
702 | clear_normalized_cda(&ch->ccw[4]); | ||
703 | if (ch->collect_len > 0) { | ||
704 | int rc; | ||
705 | |||
706 | if (ctc_checkalloc_buffer(ch, 1)) { | ||
707 | spin_unlock(&ch->collect_lock); | ||
708 | return; | ||
709 | } | ||
710 | ch->trans_skb->data = ch->trans_skb_data; | ||
711 | skb_reset_tail_pointer(ch->trans_skb); | ||
712 | ch->trans_skb->len = 0; | ||
713 | if (ch->prof.maxmulti < (ch->collect_len + 2)) | ||
714 | ch->prof.maxmulti = ch->collect_len + 2; | ||
715 | if (ch->prof.maxcqueue < skb_queue_len(&ch->collect_queue)) | ||
716 | ch->prof.maxcqueue = skb_queue_len(&ch->collect_queue); | ||
717 | *((__u16 *) skb_put(ch->trans_skb, 2)) = ch->collect_len + 2; | ||
718 | i = 0; | ||
719 | while ((skb = skb_dequeue(&ch->collect_queue))) { | ||
720 | skb_copy_from_linear_data(skb, skb_put(ch->trans_skb, | ||
721 | skb->len), | ||
722 | skb->len); | ||
723 | privptr->stats.tx_packets++; | ||
724 | privptr->stats.tx_bytes += skb->len - LL_HEADER_LENGTH; | ||
725 | atomic_dec(&skb->users); | ||
726 | dev_kfree_skb_irq(skb); | ||
727 | i++; | ||
728 | } | ||
729 | ch->collect_len = 0; | ||
730 | spin_unlock(&ch->collect_lock); | ||
731 | ch->ccw[1].count = ch->trans_skb->len; | ||
732 | fsm_addtimer(&ch->timer, CTC_TIMEOUT_5SEC, CH_EVENT_TIMER, ch); | ||
733 | ch->prof.send_stamp = current_kernel_time(); | ||
734 | rc = ccw_device_start(ch->cdev, &ch->ccw[0], | ||
735 | (unsigned long) ch, 0xff, 0); | ||
736 | ch->prof.doios_multi++; | ||
737 | if (rc != 0) { | ||
738 | privptr->stats.tx_dropped += i; | ||
739 | privptr->stats.tx_errors += i; | ||
740 | fsm_deltimer(&ch->timer); | ||
741 | ccw_check_return_code(ch, rc, "chained TX"); | ||
742 | } | ||
743 | } else { | ||
744 | spin_unlock(&ch->collect_lock); | ||
745 | fsm_newstate(fi, CH_STATE_TXIDLE); | ||
746 | } | ||
747 | ctc_clear_busy(dev); | ||
748 | } | ||
749 | |||
750 | /** | ||
751 | * Initial data is sent. | ||
752 | * Notify device statemachine that we are up and | ||
753 | * running. | ||
754 | * | ||
755 | * @param fi An instance of a channel statemachine. | ||
756 | * @param event The event, just happened. | ||
757 | * @param arg Generic pointer, casted from channel * upon call. | ||
758 | */ | ||
759 | static void | ||
760 | ch_action_txidle(fsm_instance * fi, int event, void *arg) | ||
761 | { | ||
762 | struct channel *ch = (struct channel *) arg; | ||
763 | |||
764 | DBF_TEXT(trace, 4, __FUNCTION__); | ||
765 | fsm_deltimer(&ch->timer); | ||
766 | fsm_newstate(fi, CH_STATE_TXIDLE); | ||
767 | fsm_event(((struct ctc_priv *) ch->netdev->priv)->fsm, DEV_EVENT_TXUP, | ||
768 | ch->netdev); | ||
769 | } | ||
770 | |||
771 | /** | ||
772 | * Got normal data, check for sanity, queue it up, allocate new buffer | ||
773 | * trigger bottom half, and initiate next read. | ||
774 | * | ||
775 | * @param fi An instance of a channel statemachine. | ||
776 | * @param event The event, just happened. | ||
777 | * @param arg Generic pointer, casted from channel * upon call. | ||
778 | */ | ||
779 | static void | ||
780 | ch_action_rx(fsm_instance * fi, int event, void *arg) | ||
781 | { | ||
782 | struct channel *ch = (struct channel *) arg; | ||
783 | struct net_device *dev = ch->netdev; | ||
784 | struct ctc_priv *privptr = dev->priv; | ||
785 | int len = ch->max_bufsize - ch->irb->scsw.count; | ||
786 | struct sk_buff *skb = ch->trans_skb; | ||
787 | __u16 block_len = *((__u16 *) skb->data); | ||
788 | int check_len; | ||
789 | int rc; | ||
790 | |||
791 | DBF_TEXT(trace, 4, __FUNCTION__); | ||
792 | fsm_deltimer(&ch->timer); | ||
793 | if (len < 8) { | ||
794 | ctc_pr_debug("%s: got packet with length %d < 8\n", | ||
795 | dev->name, len); | ||
796 | privptr->stats.rx_dropped++; | ||
797 | privptr->stats.rx_length_errors++; | ||
798 | goto again; | ||
799 | } | ||
800 | if (len > ch->max_bufsize) { | ||
801 | ctc_pr_debug("%s: got packet with length %d > %d\n", | ||
802 | dev->name, len, ch->max_bufsize); | ||
803 | privptr->stats.rx_dropped++; | ||
804 | privptr->stats.rx_length_errors++; | ||
805 | goto again; | ||
806 | } | ||
807 | |||
808 | /** | ||
809 | * VM TCP seems to have a bug sending 2 trailing bytes of garbage. | ||
810 | */ | ||
811 | switch (ch->protocol) { | ||
812 | case CTC_PROTO_S390: | ||
813 | case CTC_PROTO_OS390: | ||
814 | check_len = block_len + 2; | ||
815 | break; | ||
816 | default: | ||
817 | check_len = block_len; | ||
818 | break; | ||
819 | } | ||
820 | if ((len < block_len) || (len > check_len)) { | ||
821 | ctc_pr_debug("%s: got block length %d != rx length %d\n", | ||
822 | dev->name, block_len, len); | ||
823 | #ifdef DEBUG | ||
824 | ctc_dump_skb(skb, 0); | ||
825 | #endif | ||
826 | *((__u16 *) skb->data) = len; | ||
827 | privptr->stats.rx_dropped++; | ||
828 | privptr->stats.rx_length_errors++; | ||
829 | goto again; | ||
830 | } | ||
831 | block_len -= 2; | ||
832 | if (block_len > 0) { | ||
833 | *((__u16 *) skb->data) = block_len; | ||
834 | ctc_unpack_skb(ch, skb); | ||
835 | } | ||
836 | again: | ||
837 | skb->data = ch->trans_skb_data; | ||
838 | skb_reset_tail_pointer(skb); | ||
839 | skb->len = 0; | ||
840 | if (ctc_checkalloc_buffer(ch, 1)) | ||
841 | return; | ||
842 | ch->ccw[1].count = ch->max_bufsize; | ||
843 | rc = ccw_device_start(ch->cdev, &ch->ccw[0], (unsigned long) ch, 0xff, 0); | ||
844 | if (rc != 0) | ||
845 | ccw_check_return_code(ch, rc, "normal RX"); | ||
846 | } | ||
847 | |||
848 | static void ch_action_rxidle(fsm_instance * fi, int event, void *arg); | ||
849 | |||
850 | /** | ||
851 | * Initialize connection by sending a __u16 of value 0. | ||
852 | * | ||
853 | * @param fi An instance of a channel statemachine. | ||
854 | * @param event The event, just happened. | ||
855 | * @param arg Generic pointer, casted from channel * upon call. | ||
856 | */ | ||
857 | static void | ||
858 | ch_action_firstio(fsm_instance * fi, int event, void *arg) | ||
859 | { | ||
860 | struct channel *ch = (struct channel *) arg; | ||
861 | int rc; | ||
862 | |||
863 | DBF_TEXT(trace, 4, __FUNCTION__); | ||
864 | |||
865 | if (fsm_getstate(fi) == CH_STATE_TXIDLE) | ||
866 | ctc_pr_debug("%s: remote side issued READ?, init ...\n", ch->id); | ||
867 | fsm_deltimer(&ch->timer); | ||
868 | if (ctc_checkalloc_buffer(ch, 1)) | ||
869 | return; | ||
870 | if ((fsm_getstate(fi) == CH_STATE_SETUPWAIT) && | ||
871 | (ch->protocol == CTC_PROTO_OS390)) { | ||
872 | /* OS/390 resp. z/OS */ | ||
873 | if (CHANNEL_DIRECTION(ch->flags) == READ) { | ||
874 | *((__u16 *) ch->trans_skb->data) = CTC_INITIAL_BLOCKLEN; | ||
875 | fsm_addtimer(&ch->timer, CTC_TIMEOUT_5SEC, | ||
876 | CH_EVENT_TIMER, ch); | ||
877 | ch_action_rxidle(fi, event, arg); | ||
878 | } else { | ||
879 | struct net_device *dev = ch->netdev; | ||
880 | fsm_newstate(fi, CH_STATE_TXIDLE); | ||
881 | fsm_event(((struct ctc_priv *) dev->priv)->fsm, | ||
882 | DEV_EVENT_TXUP, dev); | ||
883 | } | ||
884 | return; | ||
885 | } | ||
886 | |||
887 | /** | ||
888 | * Don't setup a timer for receiving the initial RX frame | ||
889 | * if in compatibility mode, since VM TCP delays the initial | ||
890 | * frame until it has some data to send. | ||
891 | */ | ||
892 | if ((CHANNEL_DIRECTION(ch->flags) == WRITE) || | ||
893 | (ch->protocol != CTC_PROTO_S390)) | ||
894 | fsm_addtimer(&ch->timer, CTC_TIMEOUT_5SEC, CH_EVENT_TIMER, ch); | ||
895 | |||
896 | *((__u16 *) ch->trans_skb->data) = CTC_INITIAL_BLOCKLEN; | ||
897 | ch->ccw[1].count = 2; /* Transfer only length */ | ||
898 | |||
899 | fsm_newstate(fi, (CHANNEL_DIRECTION(ch->flags) == READ) | ||
900 | ? CH_STATE_RXINIT : CH_STATE_TXINIT); | ||
901 | rc = ccw_device_start(ch->cdev, &ch->ccw[0], (unsigned long) ch, 0xff, 0); | ||
902 | if (rc != 0) { | ||
903 | fsm_deltimer(&ch->timer); | ||
904 | fsm_newstate(fi, CH_STATE_SETUPWAIT); | ||
905 | ccw_check_return_code(ch, rc, "init IO"); | ||
906 | } | ||
907 | /** | ||
908 | * If in compatibility mode since we don't setup a timer, we | ||
909 | * also signal RX channel up immediately. This enables us | ||
910 | * to send packets early which in turn usually triggers some | ||
911 | * reply from VM TCP which brings up the RX channel to it's | ||
912 | * final state. | ||
913 | */ | ||
914 | if ((CHANNEL_DIRECTION(ch->flags) == READ) && | ||
915 | (ch->protocol == CTC_PROTO_S390)) { | ||
916 | struct net_device *dev = ch->netdev; | ||
917 | fsm_event(((struct ctc_priv *) dev->priv)->fsm, DEV_EVENT_RXUP, | ||
918 | dev); | ||
919 | } | ||
920 | } | ||
921 | |||
922 | /** | ||
923 | * Got initial data, check it. If OK, | ||
924 | * notify device statemachine that we are up and | ||
925 | * running. | ||
926 | * | ||
927 | * @param fi An instance of a channel statemachine. | ||
928 | * @param event The event, just happened. | ||
929 | * @param arg Generic pointer, casted from channel * upon call. | ||
930 | */ | ||
931 | static void | ||
932 | ch_action_rxidle(fsm_instance * fi, int event, void *arg) | ||
933 | { | ||
934 | struct channel *ch = (struct channel *) arg; | ||
935 | struct net_device *dev = ch->netdev; | ||
936 | __u16 buflen; | ||
937 | int rc; | ||
938 | |||
939 | DBF_TEXT(trace, 4, __FUNCTION__); | ||
940 | fsm_deltimer(&ch->timer); | ||
941 | buflen = *((__u16 *) ch->trans_skb->data); | ||
942 | #ifdef DEBUG | ||
943 | ctc_pr_debug("%s: Initial RX count %d\n", dev->name, buflen); | ||
944 | #endif | ||
945 | if (buflen >= CTC_INITIAL_BLOCKLEN) { | ||
946 | if (ctc_checkalloc_buffer(ch, 1)) | ||
947 | return; | ||
948 | ch->ccw[1].count = ch->max_bufsize; | ||
949 | fsm_newstate(fi, CH_STATE_RXIDLE); | ||
950 | rc = ccw_device_start(ch->cdev, &ch->ccw[0], | ||
951 | (unsigned long) ch, 0xff, 0); | ||
952 | if (rc != 0) { | ||
953 | fsm_newstate(fi, CH_STATE_RXINIT); | ||
954 | ccw_check_return_code(ch, rc, "initial RX"); | ||
955 | } else | ||
956 | fsm_event(((struct ctc_priv *) dev->priv)->fsm, | ||
957 | DEV_EVENT_RXUP, dev); | ||
958 | } else { | ||
959 | ctc_pr_debug("%s: Initial RX count %d not %d\n", | ||
960 | dev->name, buflen, CTC_INITIAL_BLOCKLEN); | ||
961 | ch_action_firstio(fi, event, arg); | ||
962 | } | ||
963 | } | ||
964 | |||
965 | /** | ||
966 | * Set channel into extended mode. | ||
967 | * | ||
968 | * @param fi An instance of a channel statemachine. | ||
969 | * @param event The event, just happened. | ||
970 | * @param arg Generic pointer, casted from channel * upon call. | ||
971 | */ | ||
972 | static void | ||
973 | ch_action_setmode(fsm_instance * fi, int event, void *arg) | ||
974 | { | ||
975 | struct channel *ch = (struct channel *) arg; | ||
976 | int rc; | ||
977 | unsigned long saveflags; | ||
978 | |||
979 | DBF_TEXT(trace, 4, __FUNCTION__); | ||
980 | fsm_deltimer(&ch->timer); | ||
981 | fsm_addtimer(&ch->timer, CTC_TIMEOUT_5SEC, CH_EVENT_TIMER, ch); | ||
982 | fsm_newstate(fi, CH_STATE_SETUPWAIT); | ||
983 | saveflags = 0; /* avoids compiler warning with | ||
984 | spin_unlock_irqrestore */ | ||
985 | if (event == CH_EVENT_TIMER) // only for timer not yet locked | ||
986 | spin_lock_irqsave(get_ccwdev_lock(ch->cdev), saveflags); | ||
987 | rc = ccw_device_start(ch->cdev, &ch->ccw[6], (unsigned long) ch, 0xff, 0); | ||
988 | if (event == CH_EVENT_TIMER) | ||
989 | spin_unlock_irqrestore(get_ccwdev_lock(ch->cdev), saveflags); | ||
990 | if (rc != 0) { | ||
991 | fsm_deltimer(&ch->timer); | ||
992 | fsm_newstate(fi, CH_STATE_STARTWAIT); | ||
993 | ccw_check_return_code(ch, rc, "set Mode"); | ||
994 | } else | ||
995 | ch->retry = 0; | ||
996 | } | ||
997 | |||
998 | /** | ||
999 | * Setup channel. | ||
1000 | * | ||
1001 | * @param fi An instance of a channel statemachine. | ||
1002 | * @param event The event, just happened. | ||
1003 | * @param arg Generic pointer, casted from channel * upon call. | ||
1004 | */ | ||
1005 | static void | ||
1006 | ch_action_start(fsm_instance * fi, int event, void *arg) | ||
1007 | { | ||
1008 | struct channel *ch = (struct channel *) arg; | ||
1009 | unsigned long saveflags; | ||
1010 | int rc; | ||
1011 | struct net_device *dev; | ||
1012 | |||
1013 | DBF_TEXT(trace, 4, __FUNCTION__); | ||
1014 | if (ch == NULL) { | ||
1015 | ctc_pr_warn("ch_action_start ch=NULL\n"); | ||
1016 | return; | ||
1017 | } | ||
1018 | if (ch->netdev == NULL) { | ||
1019 | ctc_pr_warn("ch_action_start dev=NULL, id=%s\n", ch->id); | ||
1020 | return; | ||
1021 | } | ||
1022 | dev = ch->netdev; | ||
1023 | |||
1024 | #ifdef DEBUG | ||
1025 | ctc_pr_debug("%s: %s channel start\n", dev->name, | ||
1026 | (CHANNEL_DIRECTION(ch->flags) == READ) ? "RX" : "TX"); | ||
1027 | #endif | ||
1028 | |||
1029 | if (ch->trans_skb != NULL) { | ||
1030 | clear_normalized_cda(&ch->ccw[1]); | ||
1031 | dev_kfree_skb(ch->trans_skb); | ||
1032 | ch->trans_skb = NULL; | ||
1033 | } | ||
1034 | if (CHANNEL_DIRECTION(ch->flags) == READ) { | ||
1035 | ch->ccw[1].cmd_code = CCW_CMD_READ; | ||
1036 | ch->ccw[1].flags = CCW_FLAG_SLI; | ||
1037 | ch->ccw[1].count = 0; | ||
1038 | } else { | ||
1039 | ch->ccw[1].cmd_code = CCW_CMD_WRITE; | ||
1040 | ch->ccw[1].flags = CCW_FLAG_SLI | CCW_FLAG_CC; | ||
1041 | ch->ccw[1].count = 0; | ||
1042 | } | ||
1043 | if (ctc_checkalloc_buffer(ch, 0)) { | ||
1044 | ctc_pr_notice( | ||
1045 | "%s: Could not allocate %s trans_skb, delaying " | ||
1046 | "allocation until first transfer\n", | ||
1047 | dev->name, | ||
1048 | (CHANNEL_DIRECTION(ch->flags) == READ) ? "RX" : "TX"); | ||
1049 | } | ||
1050 | |||
1051 | ch->ccw[0].cmd_code = CCW_CMD_PREPARE; | ||
1052 | ch->ccw[0].flags = CCW_FLAG_SLI | CCW_FLAG_CC; | ||
1053 | ch->ccw[0].count = 0; | ||
1054 | ch->ccw[0].cda = 0; | ||
1055 | ch->ccw[2].cmd_code = CCW_CMD_NOOP; /* jointed CE + DE */ | ||
1056 | ch->ccw[2].flags = CCW_FLAG_SLI; | ||
1057 | ch->ccw[2].count = 0; | ||
1058 | ch->ccw[2].cda = 0; | ||
1059 | memcpy(&ch->ccw[3], &ch->ccw[0], sizeof (struct ccw1) * 3); | ||
1060 | ch->ccw[4].cda = 0; | ||
1061 | ch->ccw[4].flags &= ~CCW_FLAG_IDA; | ||
1062 | |||
1063 | fsm_newstate(fi, CH_STATE_STARTWAIT); | ||
1064 | fsm_addtimer(&ch->timer, 1000, CH_EVENT_TIMER, ch); | ||
1065 | spin_lock_irqsave(get_ccwdev_lock(ch->cdev), saveflags); | ||
1066 | rc = ccw_device_halt(ch->cdev, (unsigned long) ch); | ||
1067 | spin_unlock_irqrestore(get_ccwdev_lock(ch->cdev), saveflags); | ||
1068 | if (rc != 0) { | ||
1069 | if (rc != -EBUSY) | ||
1070 | fsm_deltimer(&ch->timer); | ||
1071 | ccw_check_return_code(ch, rc, "initial HaltIO"); | ||
1072 | } | ||
1073 | #ifdef DEBUG | ||
1074 | ctc_pr_debug("ctc: %s(): leaving\n", __func__); | ||
1075 | #endif | ||
1076 | } | ||
1077 | |||
1078 | /** | ||
1079 | * Shutdown a channel. | ||
1080 | * | ||
1081 | * @param fi An instance of a channel statemachine. | ||
1082 | * @param event The event, just happened. | ||
1083 | * @param arg Generic pointer, casted from channel * upon call. | ||
1084 | */ | ||
1085 | static void | ||
1086 | ch_action_haltio(fsm_instance * fi, int event, void *arg) | ||
1087 | { | ||
1088 | struct channel *ch = (struct channel *) arg; | ||
1089 | unsigned long saveflags; | ||
1090 | int rc; | ||
1091 | int oldstate; | ||
1092 | |||
1093 | DBF_TEXT(trace, 3, __FUNCTION__); | ||
1094 | fsm_deltimer(&ch->timer); | ||
1095 | fsm_addtimer(&ch->timer, CTC_TIMEOUT_5SEC, CH_EVENT_TIMER, ch); | ||
1096 | saveflags = 0; /* avoids comp warning with | ||
1097 | spin_unlock_irqrestore */ | ||
1098 | if (event == CH_EVENT_STOP) // only for STOP not yet locked | ||
1099 | spin_lock_irqsave(get_ccwdev_lock(ch->cdev), saveflags); | ||
1100 | oldstate = fsm_getstate(fi); | ||
1101 | fsm_newstate(fi, CH_STATE_TERM); | ||
1102 | rc = ccw_device_halt(ch->cdev, (unsigned long) ch); | ||
1103 | if (event == CH_EVENT_STOP) | ||
1104 | spin_unlock_irqrestore(get_ccwdev_lock(ch->cdev), saveflags); | ||
1105 | if (rc != 0) { | ||
1106 | if (rc != -EBUSY) { | ||
1107 | fsm_deltimer(&ch->timer); | ||
1108 | fsm_newstate(fi, oldstate); | ||
1109 | } | ||
1110 | ccw_check_return_code(ch, rc, "HaltIO in ch_action_haltio"); | ||
1111 | } | ||
1112 | } | ||
1113 | |||
1114 | /** | ||
1115 | * A channel has successfully been halted. | ||
1116 | * Cleanup it's queue and notify interface statemachine. | ||
1117 | * | ||
1118 | * @param fi An instance of a channel statemachine. | ||
1119 | * @param event The event, just happened. | ||
1120 | * @param arg Generic pointer, casted from channel * upon call. | ||
1121 | */ | ||
1122 | static void | ||
1123 | ch_action_stopped(fsm_instance * fi, int event, void *arg) | ||
1124 | { | ||
1125 | struct channel *ch = (struct channel *) arg; | ||
1126 | struct net_device *dev = ch->netdev; | ||
1127 | |||
1128 | DBF_TEXT(trace, 3, __FUNCTION__); | ||
1129 | fsm_deltimer(&ch->timer); | ||
1130 | fsm_newstate(fi, CH_STATE_STOPPED); | ||
1131 | if (ch->trans_skb != NULL) { | ||
1132 | clear_normalized_cda(&ch->ccw[1]); | ||
1133 | dev_kfree_skb(ch->trans_skb); | ||
1134 | ch->trans_skb = NULL; | ||
1135 | } | ||
1136 | if (CHANNEL_DIRECTION(ch->flags) == READ) { | ||
1137 | skb_queue_purge(&ch->io_queue); | ||
1138 | fsm_event(((struct ctc_priv *) dev->priv)->fsm, | ||
1139 | DEV_EVENT_RXDOWN, dev); | ||
1140 | } else { | ||
1141 | ctc_purge_skb_queue(&ch->io_queue); | ||
1142 | spin_lock(&ch->collect_lock); | ||
1143 | ctc_purge_skb_queue(&ch->collect_queue); | ||
1144 | ch->collect_len = 0; | ||
1145 | spin_unlock(&ch->collect_lock); | ||
1146 | fsm_event(((struct ctc_priv *) dev->priv)->fsm, | ||
1147 | DEV_EVENT_TXDOWN, dev); | ||
1148 | } | ||
1149 | } | ||
1150 | |||
1151 | /** | ||
1152 | * A stop command from device statemachine arrived and we are in | ||
1153 | * not operational mode. Set state to stopped. | ||
1154 | * | ||
1155 | * @param fi An instance of a channel statemachine. | ||
1156 | * @param event The event, just happened. | ||
1157 | * @param arg Generic pointer, casted from channel * upon call. | ||
1158 | */ | ||
1159 | static void | ||
1160 | ch_action_stop(fsm_instance * fi, int event, void *arg) | ||
1161 | { | ||
1162 | fsm_newstate(fi, CH_STATE_STOPPED); | ||
1163 | } | ||
1164 | |||
1165 | /** | ||
1166 | * A machine check for no path, not operational status or gone device has | ||
1167 | * happened. | ||
1168 | * Cleanup queue and notify interface statemachine. | ||
1169 | * | ||
1170 | * @param fi An instance of a channel statemachine. | ||
1171 | * @param event The event, just happened. | ||
1172 | * @param arg Generic pointer, casted from channel * upon call. | ||
1173 | */ | ||
1174 | static void | ||
1175 | ch_action_fail(fsm_instance * fi, int event, void *arg) | ||
1176 | { | ||
1177 | struct channel *ch = (struct channel *) arg; | ||
1178 | struct net_device *dev = ch->netdev; | ||
1179 | |||
1180 | DBF_TEXT(trace, 3, __FUNCTION__); | ||
1181 | fsm_deltimer(&ch->timer); | ||
1182 | fsm_newstate(fi, CH_STATE_NOTOP); | ||
1183 | if (CHANNEL_DIRECTION(ch->flags) == READ) { | ||
1184 | skb_queue_purge(&ch->io_queue); | ||
1185 | fsm_event(((struct ctc_priv *) dev->priv)->fsm, | ||
1186 | DEV_EVENT_RXDOWN, dev); | ||
1187 | } else { | ||
1188 | ctc_purge_skb_queue(&ch->io_queue); | ||
1189 | spin_lock(&ch->collect_lock); | ||
1190 | ctc_purge_skb_queue(&ch->collect_queue); | ||
1191 | ch->collect_len = 0; | ||
1192 | spin_unlock(&ch->collect_lock); | ||
1193 | fsm_event(((struct ctc_priv *) dev->priv)->fsm, | ||
1194 | DEV_EVENT_TXDOWN, dev); | ||
1195 | } | ||
1196 | } | ||
1197 | |||
1198 | /** | ||
1199 | * Handle error during setup of channel. | ||
1200 | * | ||
1201 | * @param fi An instance of a channel statemachine. | ||
1202 | * @param event The event, just happened. | ||
1203 | * @param arg Generic pointer, casted from channel * upon call. | ||
1204 | */ | ||
1205 | static void | ||
1206 | ch_action_setuperr(fsm_instance * fi, int event, void *arg) | ||
1207 | { | ||
1208 | struct channel *ch = (struct channel *) arg; | ||
1209 | struct net_device *dev = ch->netdev; | ||
1210 | |||
1211 | DBF_TEXT(setup, 3, __FUNCTION__); | ||
1212 | /** | ||
1213 | * Special case: Got UC_RCRESET on setmode. | ||
1214 | * This means that remote side isn't setup. In this case | ||
1215 | * simply retry after some 10 secs... | ||
1216 | */ | ||
1217 | if ((fsm_getstate(fi) == CH_STATE_SETUPWAIT) && | ||
1218 | ((event == CH_EVENT_UC_RCRESET) || | ||
1219 | (event == CH_EVENT_UC_RSRESET))) { | ||
1220 | fsm_newstate(fi, CH_STATE_STARTRETRY); | ||
1221 | fsm_deltimer(&ch->timer); | ||
1222 | fsm_addtimer(&ch->timer, CTC_TIMEOUT_5SEC, CH_EVENT_TIMER, ch); | ||
1223 | if (CHANNEL_DIRECTION(ch->flags) == READ) { | ||
1224 | int rc = ccw_device_halt(ch->cdev, (unsigned long) ch); | ||
1225 | if (rc != 0) | ||
1226 | ccw_check_return_code( | ||
1227 | ch, rc, "HaltIO in ch_action_setuperr"); | ||
1228 | } | ||
1229 | return; | ||
1230 | } | ||
1231 | |||
1232 | ctc_pr_debug("%s: Error %s during %s channel setup state=%s\n", | ||
1233 | dev->name, ch_event_names[event], | ||
1234 | (CHANNEL_DIRECTION(ch->flags) == READ) ? "RX" : "TX", | ||
1235 | fsm_getstate_str(fi)); | ||
1236 | if (CHANNEL_DIRECTION(ch->flags) == READ) { | ||
1237 | fsm_newstate(fi, CH_STATE_RXERR); | ||
1238 | fsm_event(((struct ctc_priv *) dev->priv)->fsm, | ||
1239 | DEV_EVENT_RXDOWN, dev); | ||
1240 | } else { | ||
1241 | fsm_newstate(fi, CH_STATE_TXERR); | ||
1242 | fsm_event(((struct ctc_priv *) dev->priv)->fsm, | ||
1243 | DEV_EVENT_TXDOWN, dev); | ||
1244 | } | ||
1245 | } | ||
1246 | |||
1247 | /** | ||
1248 | * Restart a channel after an error. | ||
1249 | * | ||
1250 | * @param fi An instance of a channel statemachine. | ||
1251 | * @param event The event, just happened. | ||
1252 | * @param arg Generic pointer, casted from channel * upon call. | ||
1253 | */ | ||
1254 | static void | ||
1255 | ch_action_restart(fsm_instance * fi, int event, void *arg) | ||
1256 | { | ||
1257 | unsigned long saveflags; | ||
1258 | int oldstate; | ||
1259 | int rc; | ||
1260 | |||
1261 | struct channel *ch = (struct channel *) arg; | ||
1262 | struct net_device *dev = ch->netdev; | ||
1263 | |||
1264 | DBF_TEXT(trace, 3, __FUNCTION__); | ||
1265 | fsm_deltimer(&ch->timer); | ||
1266 | ctc_pr_debug("%s: %s channel restart\n", dev->name, | ||
1267 | (CHANNEL_DIRECTION(ch->flags) == READ) ? "RX" : "TX"); | ||
1268 | fsm_addtimer(&ch->timer, CTC_TIMEOUT_5SEC, CH_EVENT_TIMER, ch); | ||
1269 | oldstate = fsm_getstate(fi); | ||
1270 | fsm_newstate(fi, CH_STATE_STARTWAIT); | ||
1271 | saveflags = 0; /* avoids compiler warning with | ||
1272 | spin_unlock_irqrestore */ | ||
1273 | if (event == CH_EVENT_TIMER) // only for timer not yet locked | ||
1274 | spin_lock_irqsave(get_ccwdev_lock(ch->cdev), saveflags); | ||
1275 | rc = ccw_device_halt(ch->cdev, (unsigned long) ch); | ||
1276 | if (event == CH_EVENT_TIMER) | ||
1277 | spin_unlock_irqrestore(get_ccwdev_lock(ch->cdev), saveflags); | ||
1278 | if (rc != 0) { | ||
1279 | if (rc != -EBUSY) { | ||
1280 | fsm_deltimer(&ch->timer); | ||
1281 | fsm_newstate(fi, oldstate); | ||
1282 | } | ||
1283 | ccw_check_return_code(ch, rc, "HaltIO in ch_action_restart"); | ||
1284 | } | ||
1285 | } | ||
1286 | |||
1287 | /** | ||
1288 | * Handle error during RX initial handshake (exchange of | ||
1289 | * 0-length block header) | ||
1290 | * | ||
1291 | * @param fi An instance of a channel statemachine. | ||
1292 | * @param event The event, just happened. | ||
1293 | * @param arg Generic pointer, casted from channel * upon call. | ||
1294 | */ | ||
1295 | static void | ||
1296 | ch_action_rxiniterr(fsm_instance * fi, int event, void *arg) | ||
1297 | { | ||
1298 | struct channel *ch = (struct channel *) arg; | ||
1299 | struct net_device *dev = ch->netdev; | ||
1300 | |||
1301 | DBF_TEXT(setup, 3, __FUNCTION__); | ||
1302 | if (event == CH_EVENT_TIMER) { | ||
1303 | fsm_deltimer(&ch->timer); | ||
1304 | ctc_pr_debug("%s: Timeout during RX init handshake\n", dev->name); | ||
1305 | if (ch->retry++ < 3) | ||
1306 | ch_action_restart(fi, event, arg); | ||
1307 | else { | ||
1308 | fsm_newstate(fi, CH_STATE_RXERR); | ||
1309 | fsm_event(((struct ctc_priv *) dev->priv)->fsm, | ||
1310 | DEV_EVENT_RXDOWN, dev); | ||
1311 | } | ||
1312 | } else | ||
1313 | ctc_pr_warn("%s: Error during RX init handshake\n", dev->name); | ||
1314 | } | ||
1315 | |||
1316 | /** | ||
1317 | * Notify device statemachine if we gave up initialization | ||
1318 | * of RX channel. | ||
1319 | * | ||
1320 | * @param fi An instance of a channel statemachine. | ||
1321 | * @param event The event, just happened. | ||
1322 | * @param arg Generic pointer, casted from channel * upon call. | ||
1323 | */ | ||
1324 | static void | ||
1325 | ch_action_rxinitfail(fsm_instance * fi, int event, void *arg) | ||
1326 | { | ||
1327 | struct channel *ch = (struct channel *) arg; | ||
1328 | struct net_device *dev = ch->netdev; | ||
1329 | |||
1330 | DBF_TEXT(setup, 3, __FUNCTION__); | ||
1331 | fsm_newstate(fi, CH_STATE_RXERR); | ||
1332 | ctc_pr_warn("%s: RX initialization failed\n", dev->name); | ||
1333 | ctc_pr_warn("%s: RX <-> RX connection detected\n", dev->name); | ||
1334 | fsm_event(((struct ctc_priv *) dev->priv)->fsm, DEV_EVENT_RXDOWN, dev); | ||
1335 | } | ||
1336 | |||
1337 | /** | ||
1338 | * Handle RX Unit check remote reset (remote disconnected) | ||
1339 | * | ||
1340 | * @param fi An instance of a channel statemachine. | ||
1341 | * @param event The event, just happened. | ||
1342 | * @param arg Generic pointer, casted from channel * upon call. | ||
1343 | */ | ||
1344 | static void | ||
1345 | ch_action_rxdisc(fsm_instance * fi, int event, void *arg) | ||
1346 | { | ||
1347 | struct channel *ch = (struct channel *) arg; | ||
1348 | struct channel *ch2; | ||
1349 | struct net_device *dev = ch->netdev; | ||
1350 | |||
1351 | DBF_TEXT(trace, 3, __FUNCTION__); | ||
1352 | fsm_deltimer(&ch->timer); | ||
1353 | ctc_pr_debug("%s: Got remote disconnect, re-initializing ...\n", | ||
1354 | dev->name); | ||
1355 | |||
1356 | /** | ||
1357 | * Notify device statemachine | ||
1358 | */ | ||
1359 | fsm_event(((struct ctc_priv *) dev->priv)->fsm, DEV_EVENT_RXDOWN, dev); | ||
1360 | fsm_event(((struct ctc_priv *) dev->priv)->fsm, DEV_EVENT_TXDOWN, dev); | ||
1361 | |||
1362 | fsm_newstate(fi, CH_STATE_DTERM); | ||
1363 | ch2 = ((struct ctc_priv *) dev->priv)->channel[WRITE]; | ||
1364 | fsm_newstate(ch2->fsm, CH_STATE_DTERM); | ||
1365 | |||
1366 | ccw_device_halt(ch->cdev, (unsigned long) ch); | ||
1367 | ccw_device_halt(ch2->cdev, (unsigned long) ch2); | ||
1368 | } | ||
1369 | |||
1370 | /** | ||
1371 | * Handle error during TX channel initialization. | ||
1372 | * | ||
1373 | * @param fi An instance of a channel statemachine. | ||
1374 | * @param event The event, just happened. | ||
1375 | * @param arg Generic pointer, casted from channel * upon call. | ||
1376 | */ | ||
1377 | static void | ||
1378 | ch_action_txiniterr(fsm_instance * fi, int event, void *arg) | ||
1379 | { | ||
1380 | struct channel *ch = (struct channel *) arg; | ||
1381 | struct net_device *dev = ch->netdev; | ||
1382 | |||
1383 | DBF_TEXT(setup, 2, __FUNCTION__); | ||
1384 | if (event == CH_EVENT_TIMER) { | ||
1385 | fsm_deltimer(&ch->timer); | ||
1386 | ctc_pr_debug("%s: Timeout during TX init handshake\n", dev->name); | ||
1387 | if (ch->retry++ < 3) | ||
1388 | ch_action_restart(fi, event, arg); | ||
1389 | else { | ||
1390 | fsm_newstate(fi, CH_STATE_TXERR); | ||
1391 | fsm_event(((struct ctc_priv *) dev->priv)->fsm, | ||
1392 | DEV_EVENT_TXDOWN, dev); | ||
1393 | } | ||
1394 | } else | ||
1395 | ctc_pr_warn("%s: Error during TX init handshake\n", dev->name); | ||
1396 | } | ||
1397 | |||
1398 | /** | ||
1399 | * Handle TX timeout by retrying operation. | ||
1400 | * | ||
1401 | * @param fi An instance of a channel statemachine. | ||
1402 | * @param event The event, just happened. | ||
1403 | * @param arg Generic pointer, casted from channel * upon call. | ||
1404 | */ | ||
1405 | static void | ||
1406 | ch_action_txretry(fsm_instance * fi, int event, void *arg) | ||
1407 | { | ||
1408 | struct channel *ch = (struct channel *) arg; | ||
1409 | struct net_device *dev = ch->netdev; | ||
1410 | unsigned long saveflags; | ||
1411 | |||
1412 | DBF_TEXT(trace, 4, __FUNCTION__); | ||
1413 | fsm_deltimer(&ch->timer); | ||
1414 | if (ch->retry++ > 3) { | ||
1415 | ctc_pr_debug("%s: TX retry failed, restarting channel\n", | ||
1416 | dev->name); | ||
1417 | fsm_event(((struct ctc_priv *) dev->priv)->fsm, | ||
1418 | DEV_EVENT_TXDOWN, dev); | ||
1419 | ch_action_restart(fi, event, arg); | ||
1420 | } else { | ||
1421 | struct sk_buff *skb; | ||
1422 | |||
1423 | ctc_pr_debug("%s: TX retry %d\n", dev->name, ch->retry); | ||
1424 | if ((skb = skb_peek(&ch->io_queue))) { | ||
1425 | int rc = 0; | ||
1426 | |||
1427 | clear_normalized_cda(&ch->ccw[4]); | ||
1428 | ch->ccw[4].count = skb->len; | ||
1429 | if (set_normalized_cda(&ch->ccw[4], skb->data)) { | ||
1430 | ctc_pr_debug( | ||
1431 | "%s: IDAL alloc failed, chan restart\n", | ||
1432 | dev->name); | ||
1433 | fsm_event(((struct ctc_priv *) dev->priv)->fsm, | ||
1434 | DEV_EVENT_TXDOWN, dev); | ||
1435 | ch_action_restart(fi, event, arg); | ||
1436 | return; | ||
1437 | } | ||
1438 | fsm_addtimer(&ch->timer, 1000, CH_EVENT_TIMER, ch); | ||
1439 | saveflags = 0; /* avoids compiler warning with | ||
1440 | spin_unlock_irqrestore */ | ||
1441 | if (event == CH_EVENT_TIMER) // only for TIMER not yet locked | ||
1442 | spin_lock_irqsave(get_ccwdev_lock(ch->cdev), | ||
1443 | saveflags); | ||
1444 | rc = ccw_device_start(ch->cdev, &ch->ccw[3], | ||
1445 | (unsigned long) ch, 0xff, 0); | ||
1446 | if (event == CH_EVENT_TIMER) | ||
1447 | spin_unlock_irqrestore(get_ccwdev_lock(ch->cdev), | ||
1448 | saveflags); | ||
1449 | if (rc != 0) { | ||
1450 | fsm_deltimer(&ch->timer); | ||
1451 | ccw_check_return_code(ch, rc, "TX in ch_action_txretry"); | ||
1452 | ctc_purge_skb_queue(&ch->io_queue); | ||
1453 | } | ||
1454 | } | ||
1455 | } | ||
1456 | |||
1457 | } | ||
1458 | |||
1459 | /** | ||
1460 | * Handle fatal errors during an I/O command. | ||
1461 | * | ||
1462 | * @param fi An instance of a channel statemachine. | ||
1463 | * @param event The event, just happened. | ||
1464 | * @param arg Generic pointer, casted from channel * upon call. | ||
1465 | */ | ||
1466 | static void | ||
1467 | ch_action_iofatal(fsm_instance * fi, int event, void *arg) | ||
1468 | { | ||
1469 | struct channel *ch = (struct channel *) arg; | ||
1470 | struct net_device *dev = ch->netdev; | ||
1471 | |||
1472 | DBF_TEXT(trace, 3, __FUNCTION__); | ||
1473 | fsm_deltimer(&ch->timer); | ||
1474 | if (CHANNEL_DIRECTION(ch->flags) == READ) { | ||
1475 | ctc_pr_debug("%s: RX I/O error\n", dev->name); | ||
1476 | fsm_newstate(fi, CH_STATE_RXERR); | ||
1477 | fsm_event(((struct ctc_priv *) dev->priv)->fsm, | ||
1478 | DEV_EVENT_RXDOWN, dev); | ||
1479 | } else { | ||
1480 | ctc_pr_debug("%s: TX I/O error\n", dev->name); | ||
1481 | fsm_newstate(fi, CH_STATE_TXERR); | ||
1482 | fsm_event(((struct ctc_priv *) dev->priv)->fsm, | ||
1483 | DEV_EVENT_TXDOWN, dev); | ||
1484 | } | ||
1485 | } | ||
1486 | |||
1487 | static void | ||
1488 | ch_action_reinit(fsm_instance *fi, int event, void *arg) | ||
1489 | { | ||
1490 | struct channel *ch = (struct channel *)arg; | ||
1491 | struct net_device *dev = ch->netdev; | ||
1492 | struct ctc_priv *privptr = dev->priv; | ||
1493 | |||
1494 | DBF_TEXT(trace, 4, __FUNCTION__); | ||
1495 | ch_action_iofatal(fi, event, arg); | ||
1496 | fsm_addtimer(&privptr->restart_timer, 1000, DEV_EVENT_RESTART, dev); | ||
1497 | } | ||
1498 | |||
1499 | /** | ||
1500 | * The statemachine for a channel. | ||
1501 | */ | ||
1502 | static const fsm_node ch_fsm[] = { | ||
1503 | {CH_STATE_STOPPED, CH_EVENT_STOP, fsm_action_nop }, | ||
1504 | {CH_STATE_STOPPED, CH_EVENT_START, ch_action_start }, | ||
1505 | {CH_STATE_STOPPED, CH_EVENT_FINSTAT, fsm_action_nop }, | ||
1506 | {CH_STATE_STOPPED, CH_EVENT_MC_FAIL, fsm_action_nop }, | ||
1507 | |||
1508 | {CH_STATE_NOTOP, CH_EVENT_STOP, ch_action_stop }, | ||
1509 | {CH_STATE_NOTOP, CH_EVENT_START, fsm_action_nop }, | ||
1510 | {CH_STATE_NOTOP, CH_EVENT_FINSTAT, fsm_action_nop }, | ||
1511 | {CH_STATE_NOTOP, CH_EVENT_MC_FAIL, fsm_action_nop }, | ||
1512 | {CH_STATE_NOTOP, CH_EVENT_MC_GOOD, ch_action_start }, | ||
1513 | |||
1514 | {CH_STATE_STARTWAIT, CH_EVENT_STOP, ch_action_haltio }, | ||
1515 | {CH_STATE_STARTWAIT, CH_EVENT_START, fsm_action_nop }, | ||
1516 | {CH_STATE_STARTWAIT, CH_EVENT_FINSTAT, ch_action_setmode }, | ||
1517 | {CH_STATE_STARTWAIT, CH_EVENT_TIMER, ch_action_setuperr }, | ||
1518 | {CH_STATE_STARTWAIT, CH_EVENT_IO_ENODEV, ch_action_iofatal }, | ||
1519 | {CH_STATE_STARTWAIT, CH_EVENT_IO_EIO, ch_action_reinit }, | ||
1520 | {CH_STATE_STARTWAIT, CH_EVENT_MC_FAIL, ch_action_fail }, | ||
1521 | |||
1522 | {CH_STATE_STARTRETRY, CH_EVENT_STOP, ch_action_haltio }, | ||
1523 | {CH_STATE_STARTRETRY, CH_EVENT_TIMER, ch_action_setmode }, | ||
1524 | {CH_STATE_STARTRETRY, CH_EVENT_FINSTAT, fsm_action_nop }, | ||
1525 | {CH_STATE_STARTRETRY, CH_EVENT_MC_FAIL, ch_action_fail }, | ||
1526 | |||
1527 | {CH_STATE_SETUPWAIT, CH_EVENT_STOP, ch_action_haltio }, | ||
1528 | {CH_STATE_SETUPWAIT, CH_EVENT_START, fsm_action_nop }, | ||
1529 | {CH_STATE_SETUPWAIT, CH_EVENT_FINSTAT, ch_action_firstio }, | ||
1530 | {CH_STATE_SETUPWAIT, CH_EVENT_UC_RCRESET, ch_action_setuperr }, | ||
1531 | {CH_STATE_SETUPWAIT, CH_EVENT_UC_RSRESET, ch_action_setuperr }, | ||
1532 | {CH_STATE_SETUPWAIT, CH_EVENT_TIMER, ch_action_setmode }, | ||
1533 | {CH_STATE_SETUPWAIT, CH_EVENT_IO_ENODEV, ch_action_iofatal }, | ||
1534 | {CH_STATE_SETUPWAIT, CH_EVENT_IO_EIO, ch_action_reinit }, | ||
1535 | {CH_STATE_SETUPWAIT, CH_EVENT_MC_FAIL, ch_action_fail }, | ||
1536 | |||
1537 | {CH_STATE_RXINIT, CH_EVENT_STOP, ch_action_haltio }, | ||
1538 | {CH_STATE_RXINIT, CH_EVENT_START, fsm_action_nop }, | ||
1539 | {CH_STATE_RXINIT, CH_EVENT_FINSTAT, ch_action_rxidle }, | ||
1540 | {CH_STATE_RXINIT, CH_EVENT_UC_RCRESET, ch_action_rxiniterr }, | ||
1541 | {CH_STATE_RXINIT, CH_EVENT_UC_RSRESET, ch_action_rxiniterr }, | ||
1542 | {CH_STATE_RXINIT, CH_EVENT_TIMER, ch_action_rxiniterr }, | ||
1543 | {CH_STATE_RXINIT, CH_EVENT_ATTNBUSY, ch_action_rxinitfail }, | ||
1544 | {CH_STATE_RXINIT, CH_EVENT_IO_ENODEV, ch_action_iofatal }, | ||
1545 | {CH_STATE_RXINIT, CH_EVENT_IO_EIO, ch_action_reinit }, | ||
1546 | {CH_STATE_RXINIT, CH_EVENT_UC_ZERO, ch_action_firstio }, | ||
1547 | {CH_STATE_RXINIT, CH_EVENT_MC_FAIL, ch_action_fail }, | ||
1548 | |||
1549 | {CH_STATE_RXIDLE, CH_EVENT_STOP, ch_action_haltio }, | ||
1550 | {CH_STATE_RXIDLE, CH_EVENT_START, fsm_action_nop }, | ||
1551 | {CH_STATE_RXIDLE, CH_EVENT_FINSTAT, ch_action_rx }, | ||
1552 | {CH_STATE_RXIDLE, CH_EVENT_UC_RCRESET, ch_action_rxdisc }, | ||
1553 | // {CH_STATE_RXIDLE, CH_EVENT_UC_RSRESET, ch_action_rxretry }, | ||
1554 | {CH_STATE_RXIDLE, CH_EVENT_IO_ENODEV, ch_action_iofatal }, | ||
1555 | {CH_STATE_RXIDLE, CH_EVENT_IO_EIO, ch_action_reinit }, | ||
1556 | {CH_STATE_RXIDLE, CH_EVENT_MC_FAIL, ch_action_fail }, | ||
1557 | {CH_STATE_RXIDLE, CH_EVENT_UC_ZERO, ch_action_rx }, | ||
1558 | |||
1559 | {CH_STATE_TXINIT, CH_EVENT_STOP, ch_action_haltio }, | ||
1560 | {CH_STATE_TXINIT, CH_EVENT_START, fsm_action_nop }, | ||
1561 | {CH_STATE_TXINIT, CH_EVENT_FINSTAT, ch_action_txidle }, | ||
1562 | {CH_STATE_TXINIT, CH_EVENT_UC_RCRESET, ch_action_txiniterr }, | ||
1563 | {CH_STATE_TXINIT, CH_EVENT_UC_RSRESET, ch_action_txiniterr }, | ||
1564 | {CH_STATE_TXINIT, CH_EVENT_TIMER, ch_action_txiniterr }, | ||
1565 | {CH_STATE_TXINIT, CH_EVENT_IO_ENODEV, ch_action_iofatal }, | ||
1566 | {CH_STATE_TXINIT, CH_EVENT_IO_EIO, ch_action_reinit }, | ||
1567 | {CH_STATE_TXINIT, CH_EVENT_MC_FAIL, ch_action_fail }, | ||
1568 | |||
1569 | {CH_STATE_TXIDLE, CH_EVENT_STOP, ch_action_haltio }, | ||
1570 | {CH_STATE_TXIDLE, CH_EVENT_START, fsm_action_nop }, | ||
1571 | {CH_STATE_TXIDLE, CH_EVENT_FINSTAT, ch_action_firstio }, | ||
1572 | {CH_STATE_TXIDLE, CH_EVENT_UC_RCRESET, fsm_action_nop }, | ||
1573 | {CH_STATE_TXIDLE, CH_EVENT_UC_RSRESET, fsm_action_nop }, | ||
1574 | {CH_STATE_TXIDLE, CH_EVENT_IO_ENODEV, ch_action_iofatal }, | ||
1575 | {CH_STATE_TXIDLE, CH_EVENT_IO_EIO, ch_action_reinit }, | ||
1576 | {CH_STATE_TXIDLE, CH_EVENT_MC_FAIL, ch_action_fail }, | ||
1577 | |||
1578 | {CH_STATE_TERM, CH_EVENT_STOP, fsm_action_nop }, | ||
1579 | {CH_STATE_TERM, CH_EVENT_START, ch_action_restart }, | ||
1580 | {CH_STATE_TERM, CH_EVENT_FINSTAT, ch_action_stopped }, | ||
1581 | {CH_STATE_TERM, CH_EVENT_UC_RCRESET, fsm_action_nop }, | ||
1582 | {CH_STATE_TERM, CH_EVENT_UC_RSRESET, fsm_action_nop }, | ||
1583 | {CH_STATE_TERM, CH_EVENT_MC_FAIL, ch_action_fail }, | ||
1584 | |||
1585 | {CH_STATE_DTERM, CH_EVENT_STOP, ch_action_haltio }, | ||
1586 | {CH_STATE_DTERM, CH_EVENT_START, ch_action_restart }, | ||
1587 | {CH_STATE_DTERM, CH_EVENT_FINSTAT, ch_action_setmode }, | ||
1588 | {CH_STATE_DTERM, CH_EVENT_UC_RCRESET, fsm_action_nop }, | ||
1589 | {CH_STATE_DTERM, CH_EVENT_UC_RSRESET, fsm_action_nop }, | ||
1590 | {CH_STATE_DTERM, CH_EVENT_MC_FAIL, ch_action_fail }, | ||
1591 | |||
1592 | {CH_STATE_TX, CH_EVENT_STOP, ch_action_haltio }, | ||
1593 | {CH_STATE_TX, CH_EVENT_START, fsm_action_nop }, | ||
1594 | {CH_STATE_TX, CH_EVENT_FINSTAT, ch_action_txdone }, | ||
1595 | {CH_STATE_TX, CH_EVENT_UC_RCRESET, ch_action_txretry }, | ||
1596 | {CH_STATE_TX, CH_EVENT_UC_RSRESET, ch_action_txretry }, | ||
1597 | {CH_STATE_TX, CH_EVENT_TIMER, ch_action_txretry }, | ||
1598 | {CH_STATE_TX, CH_EVENT_IO_ENODEV, ch_action_iofatal }, | ||
1599 | {CH_STATE_TX, CH_EVENT_IO_EIO, ch_action_reinit }, | ||
1600 | {CH_STATE_TX, CH_EVENT_MC_FAIL, ch_action_fail }, | ||
1601 | |||
1602 | {CH_STATE_RXERR, CH_EVENT_STOP, ch_action_haltio }, | ||
1603 | {CH_STATE_TXERR, CH_EVENT_STOP, ch_action_haltio }, | ||
1604 | {CH_STATE_TXERR, CH_EVENT_MC_FAIL, ch_action_fail }, | ||
1605 | {CH_STATE_RXERR, CH_EVENT_MC_FAIL, ch_action_fail }, | ||
1606 | }; | ||
1607 | |||
1608 | static const int CH_FSM_LEN = sizeof (ch_fsm) / sizeof (fsm_node); | ||
1609 | |||
1610 | /** | ||
1611 | * Functions related to setup and device detection. | ||
1612 | *****************************************************************************/ | ||
1613 | |||
1614 | static inline int | ||
1615 | less_than(char *id1, char *id2) | ||
1616 | { | ||
1617 | int dev1, dev2, i; | ||
1618 | |||
1619 | for (i = 0; i < 5; i++) { | ||
1620 | id1++; | ||
1621 | id2++; | ||
1622 | } | ||
1623 | dev1 = simple_strtoul(id1, &id1, 16); | ||
1624 | dev2 = simple_strtoul(id2, &id2, 16); | ||
1625 | |||
1626 | return (dev1 < dev2); | ||
1627 | } | ||
1628 | |||
1629 | /** | ||
1630 | * Add a new channel to the list of channels. | ||
1631 | * Keeps the channel list sorted. | ||
1632 | * | ||
1633 | * @param cdev The ccw_device to be added. | ||
1634 | * @param type The type class of the new channel. | ||
1635 | * | ||
1636 | * @return 0 on success, !0 on error. | ||
1637 | */ | ||
1638 | static int | ||
1639 | add_channel(struct ccw_device *cdev, enum channel_types type) | ||
1640 | { | ||
1641 | struct channel **c = &channels; | ||
1642 | struct channel *ch; | ||
1643 | |||
1644 | DBF_TEXT(trace, 2, __FUNCTION__); | ||
1645 | ch = kzalloc(sizeof(struct channel), GFP_KERNEL); | ||
1646 | if (!ch) { | ||
1647 | ctc_pr_warn("ctc: Out of memory in add_channel\n"); | ||
1648 | return -1; | ||
1649 | } | ||
1650 | /* assure all flags and counters are reset */ | ||
1651 | ch->ccw = kzalloc(8 * sizeof(struct ccw1), GFP_KERNEL | GFP_DMA); | ||
1652 | if (!ch->ccw) { | ||
1653 | kfree(ch); | ||
1654 | ctc_pr_warn("ctc: Out of memory in add_channel\n"); | ||
1655 | return -1; | ||
1656 | } | ||
1657 | |||
1658 | |||
1659 | /** | ||
1660 | * "static" ccws are used in the following way: | ||
1661 | * | ||
1662 | * ccw[0..2] (Channel program for generic I/O): | ||
1663 | * 0: prepare | ||
1664 | * 1: read or write (depending on direction) with fixed | ||
1665 | * buffer (idal allocated once when buffer is allocated) | ||
1666 | * 2: nop | ||
1667 | * ccw[3..5] (Channel program for direct write of packets) | ||
1668 | * 3: prepare | ||
1669 | * 4: write (idal allocated on every write). | ||
1670 | * 5: nop | ||
1671 | * ccw[6..7] (Channel program for initial channel setup): | ||
1672 | * 6: set extended mode | ||
1673 | * 7: nop | ||
1674 | * | ||
1675 | * ch->ccw[0..5] are initialized in ch_action_start because | ||
1676 | * the channel's direction is yet unknown here. | ||
1677 | */ | ||
1678 | ch->ccw[6].cmd_code = CCW_CMD_SET_EXTENDED; | ||
1679 | ch->ccw[6].flags = CCW_FLAG_SLI; | ||
1680 | |||
1681 | ch->ccw[7].cmd_code = CCW_CMD_NOOP; | ||
1682 | ch->ccw[7].flags = CCW_FLAG_SLI; | ||
1683 | |||
1684 | ch->cdev = cdev; | ||
1685 | snprintf(ch->id, CTC_ID_SIZE, "ch-%s", cdev->dev.bus_id); | ||
1686 | ch->type = type; | ||
1687 | ch->fsm = init_fsm(ch->id, ch_state_names, | ||
1688 | ch_event_names, NR_CH_STATES, NR_CH_EVENTS, | ||
1689 | ch_fsm, CH_FSM_LEN, GFP_KERNEL); | ||
1690 | if (ch->fsm == NULL) { | ||
1691 | ctc_pr_warn("ctc: Could not create FSM in add_channel\n"); | ||
1692 | kfree(ch->ccw); | ||
1693 | kfree(ch); | ||
1694 | return -1; | ||
1695 | } | ||
1696 | fsm_newstate(ch->fsm, CH_STATE_IDLE); | ||
1697 | ch->irb = kzalloc(sizeof(struct irb), GFP_KERNEL); | ||
1698 | if (!ch->irb) { | ||
1699 | ctc_pr_warn("ctc: Out of memory in add_channel\n"); | ||
1700 | kfree_fsm(ch->fsm); | ||
1701 | kfree(ch->ccw); | ||
1702 | kfree(ch); | ||
1703 | return -1; | ||
1704 | } | ||
1705 | while (*c && less_than((*c)->id, ch->id)) | ||
1706 | c = &(*c)->next; | ||
1707 | if (*c && (!strncmp((*c)->id, ch->id, CTC_ID_SIZE))) { | ||
1708 | ctc_pr_debug( | ||
1709 | "ctc: add_channel: device %s already in list, " | ||
1710 | "using old entry\n", (*c)->id); | ||
1711 | kfree(ch->irb); | ||
1712 | kfree_fsm(ch->fsm); | ||
1713 | kfree(ch->ccw); | ||
1714 | kfree(ch); | ||
1715 | return 0; | ||
1716 | } | ||
1717 | |||
1718 | spin_lock_init(&ch->collect_lock); | ||
1719 | |||
1720 | fsm_settimer(ch->fsm, &ch->timer); | ||
1721 | skb_queue_head_init(&ch->io_queue); | ||
1722 | skb_queue_head_init(&ch->collect_queue); | ||
1723 | ch->next = *c; | ||
1724 | *c = ch; | ||
1725 | return 0; | ||
1726 | } | ||
1727 | |||
1728 | /** | ||
1729 | * Release a specific channel in the channel list. | ||
1730 | * | ||
1731 | * @param ch Pointer to channel struct to be released. | ||
1732 | */ | ||
1733 | static void | ||
1734 | channel_free(struct channel *ch) | ||
1735 | { | ||
1736 | ch->flags &= ~CHANNEL_FLAGS_INUSE; | ||
1737 | fsm_newstate(ch->fsm, CH_STATE_IDLE); | ||
1738 | } | ||
1739 | |||
1740 | /** | ||
1741 | * Remove a specific channel in the channel list. | ||
1742 | * | ||
1743 | * @param ch Pointer to channel struct to be released. | ||
1744 | */ | ||
1745 | static void | ||
1746 | channel_remove(struct channel *ch) | ||
1747 | { | ||
1748 | struct channel **c = &channels; | ||
1749 | |||
1750 | DBF_TEXT(trace, 2, __FUNCTION__); | ||
1751 | if (ch == NULL) | ||
1752 | return; | ||
1753 | |||
1754 | channel_free(ch); | ||
1755 | while (*c) { | ||
1756 | if (*c == ch) { | ||
1757 | *c = ch->next; | ||
1758 | fsm_deltimer(&ch->timer); | ||
1759 | kfree_fsm(ch->fsm); | ||
1760 | clear_normalized_cda(&ch->ccw[4]); | ||
1761 | if (ch->trans_skb != NULL) { | ||
1762 | clear_normalized_cda(&ch->ccw[1]); | ||
1763 | dev_kfree_skb(ch->trans_skb); | ||
1764 | } | ||
1765 | kfree(ch->ccw); | ||
1766 | kfree(ch->irb); | ||
1767 | kfree(ch); | ||
1768 | return; | ||
1769 | } | ||
1770 | c = &((*c)->next); | ||
1771 | } | ||
1772 | } | ||
1773 | |||
1774 | /** | ||
1775 | * Get a specific channel from the channel list. | ||
1776 | * | ||
1777 | * @param type Type of channel we are interested in. | ||
1778 | * @param id Id of channel we are interested in. | ||
1779 | * @param direction Direction we want to use this channel for. | ||
1780 | * | ||
1781 | * @return Pointer to a channel or NULL if no matching channel available. | ||
1782 | */ | ||
1783 | static struct channel | ||
1784 | * | ||
1785 | channel_get(enum channel_types type, char *id, int direction) | ||
1786 | { | ||
1787 | struct channel *ch = channels; | ||
1788 | |||
1789 | DBF_TEXT(trace, 3, __FUNCTION__); | ||
1790 | #ifdef DEBUG | ||
1791 | ctc_pr_debug("ctc: %s(): searching for ch with id %s and type %d\n", | ||
1792 | __func__, id, type); | ||
1793 | #endif | ||
1794 | |||
1795 | while (ch && ((strncmp(ch->id, id, CTC_ID_SIZE)) || (ch->type != type))) { | ||
1796 | #ifdef DEBUG | ||
1797 | ctc_pr_debug("ctc: %s(): ch=0x%p (id=%s, type=%d\n", | ||
1798 | __func__, ch, ch->id, ch->type); | ||
1799 | #endif | ||
1800 | ch = ch->next; | ||
1801 | } | ||
1802 | #ifdef DEBUG | ||
1803 | ctc_pr_debug("ctc: %s(): ch=0x%pq (id=%s, type=%d\n", | ||
1804 | __func__, ch, ch->id, ch->type); | ||
1805 | #endif | ||
1806 | if (!ch) { | ||
1807 | ctc_pr_warn("ctc: %s(): channel with id %s " | ||
1808 | "and type %d not found in channel list\n", | ||
1809 | __func__, id, type); | ||
1810 | } else { | ||
1811 | if (ch->flags & CHANNEL_FLAGS_INUSE) | ||
1812 | ch = NULL; | ||
1813 | else { | ||
1814 | ch->flags |= CHANNEL_FLAGS_INUSE; | ||
1815 | ch->flags &= ~CHANNEL_FLAGS_RWMASK; | ||
1816 | ch->flags |= (direction == WRITE) | ||
1817 | ? CHANNEL_FLAGS_WRITE : CHANNEL_FLAGS_READ; | ||
1818 | fsm_newstate(ch->fsm, CH_STATE_STOPPED); | ||
1819 | } | ||
1820 | } | ||
1821 | return ch; | ||
1822 | } | ||
1823 | |||
1824 | /** | ||
1825 | * Return the channel type by name. | ||
1826 | * | ||
1827 | * @param name Name of network interface. | ||
1828 | * | ||
1829 | * @return Type class of channel to be used for that interface. | ||
1830 | */ | ||
1831 | static enum channel_types inline | ||
1832 | extract_channel_media(char *name) | ||
1833 | { | ||
1834 | enum channel_types ret = channel_type_unknown; | ||
1835 | |||
1836 | if (name != NULL) { | ||
1837 | if (strncmp(name, "ctc", 3) == 0) | ||
1838 | ret = channel_type_parallel; | ||
1839 | if (strncmp(name, "escon", 5) == 0) | ||
1840 | ret = channel_type_escon; | ||
1841 | } | ||
1842 | return ret; | ||
1843 | } | ||
1844 | |||
1845 | static long | ||
1846 | __ctc_check_irb_error(struct ccw_device *cdev, struct irb *irb) | ||
1847 | { | ||
1848 | if (!IS_ERR(irb)) | ||
1849 | return 0; | ||
1850 | |||
1851 | switch (PTR_ERR(irb)) { | ||
1852 | case -EIO: | ||
1853 | ctc_pr_warn("i/o-error on device %s\n", cdev->dev.bus_id); | ||
1854 | // CTC_DBF_TEXT(trace, 2, "ckirberr"); | ||
1855 | // CTC_DBF_TEXT_(trace, 2, " rc%d", -EIO); | ||
1856 | break; | ||
1857 | case -ETIMEDOUT: | ||
1858 | ctc_pr_warn("timeout on device %s\n", cdev->dev.bus_id); | ||
1859 | // CTC_DBF_TEXT(trace, 2, "ckirberr"); | ||
1860 | // CTC_DBF_TEXT_(trace, 2, " rc%d", -ETIMEDOUT); | ||
1861 | break; | ||
1862 | default: | ||
1863 | ctc_pr_warn("unknown error %ld on device %s\n", PTR_ERR(irb), | ||
1864 | cdev->dev.bus_id); | ||
1865 | // CTC_DBF_TEXT(trace, 2, "ckirberr"); | ||
1866 | // CTC_DBF_TEXT(trace, 2, " rc???"); | ||
1867 | } | ||
1868 | return PTR_ERR(irb); | ||
1869 | } | ||
1870 | |||
1871 | /** | ||
1872 | * Main IRQ handler. | ||
1873 | * | ||
1874 | * @param cdev The ccw_device the interrupt is for. | ||
1875 | * @param intparm interruption parameter. | ||
1876 | * @param irb interruption response block. | ||
1877 | */ | ||
1878 | static void | ||
1879 | ctc_irq_handler(struct ccw_device *cdev, unsigned long intparm, struct irb *irb) | ||
1880 | { | ||
1881 | struct channel *ch; | ||
1882 | struct net_device *dev; | ||
1883 | struct ctc_priv *priv; | ||
1884 | |||
1885 | DBF_TEXT(trace, 5, __FUNCTION__); | ||
1886 | if (__ctc_check_irb_error(cdev, irb)) | ||
1887 | return; | ||
1888 | |||
1889 | /* Check for unsolicited interrupts. */ | ||
1890 | if (!cdev->dev.driver_data) { | ||
1891 | ctc_pr_warn("ctc: Got unsolicited irq: %s c-%02x d-%02x\n", | ||
1892 | cdev->dev.bus_id, irb->scsw.cstat, | ||
1893 | irb->scsw.dstat); | ||
1894 | return; | ||
1895 | } | ||
1896 | |||
1897 | priv = ((struct ccwgroup_device *)cdev->dev.driver_data) | ||
1898 | ->dev.driver_data; | ||
1899 | |||
1900 | /* Try to extract channel from driver data. */ | ||
1901 | if (priv->channel[READ]->cdev == cdev) | ||
1902 | ch = priv->channel[READ]; | ||
1903 | else if (priv->channel[WRITE]->cdev == cdev) | ||
1904 | ch = priv->channel[WRITE]; | ||
1905 | else { | ||
1906 | ctc_pr_err("ctc: Can't determine channel for interrupt, " | ||
1907 | "device %s\n", cdev->dev.bus_id); | ||
1908 | return; | ||
1909 | } | ||
1910 | |||
1911 | dev = (struct net_device *) (ch->netdev); | ||
1912 | if (dev == NULL) { | ||
1913 | ctc_pr_crit("ctc: ctc_irq_handler dev=NULL bus_id=%s, ch=0x%p\n", | ||
1914 | cdev->dev.bus_id, ch); | ||
1915 | return; | ||
1916 | } | ||
1917 | |||
1918 | #ifdef DEBUG | ||
1919 | ctc_pr_debug("%s: interrupt for device: %s received c-%02x d-%02x\n", | ||
1920 | dev->name, ch->id, irb->scsw.cstat, irb->scsw.dstat); | ||
1921 | #endif | ||
1922 | |||
1923 | /* Copy interruption response block. */ | ||
1924 | memcpy(ch->irb, irb, sizeof(struct irb)); | ||
1925 | |||
1926 | /* Check for good subchannel return code, otherwise error message */ | ||
1927 | if (ch->irb->scsw.cstat) { | ||
1928 | fsm_event(ch->fsm, CH_EVENT_SC_UNKNOWN, ch); | ||
1929 | ctc_pr_warn("%s: subchannel check for device: %s - %02x %02x\n", | ||
1930 | dev->name, ch->id, ch->irb->scsw.cstat, | ||
1931 | ch->irb->scsw.dstat); | ||
1932 | return; | ||
1933 | } | ||
1934 | |||
1935 | /* Check the reason-code of a unit check */ | ||
1936 | if (ch->irb->scsw.dstat & DEV_STAT_UNIT_CHECK) { | ||
1937 | ccw_unit_check(ch, ch->irb->ecw[0]); | ||
1938 | return; | ||
1939 | } | ||
1940 | if (ch->irb->scsw.dstat & DEV_STAT_BUSY) { | ||
1941 | if (ch->irb->scsw.dstat & DEV_STAT_ATTENTION) | ||
1942 | fsm_event(ch->fsm, CH_EVENT_ATTNBUSY, ch); | ||
1943 | else | ||
1944 | fsm_event(ch->fsm, CH_EVENT_BUSY, ch); | ||
1945 | return; | ||
1946 | } | ||
1947 | if (ch->irb->scsw.dstat & DEV_STAT_ATTENTION) { | ||
1948 | fsm_event(ch->fsm, CH_EVENT_ATTN, ch); | ||
1949 | return; | ||
1950 | } | ||
1951 | if ((ch->irb->scsw.stctl & SCSW_STCTL_SEC_STATUS) || | ||
1952 | (ch->irb->scsw.stctl == SCSW_STCTL_STATUS_PEND) || | ||
1953 | (ch->irb->scsw.stctl == | ||
1954 | (SCSW_STCTL_ALERT_STATUS | SCSW_STCTL_STATUS_PEND))) | ||
1955 | fsm_event(ch->fsm, CH_EVENT_FINSTAT, ch); | ||
1956 | else | ||
1957 | fsm_event(ch->fsm, CH_EVENT_IRQ, ch); | ||
1958 | |||
1959 | } | ||
1960 | |||
1961 | /** | ||
1962 | * Actions for interface - statemachine. | ||
1963 | *****************************************************************************/ | ||
1964 | |||
1965 | /** | ||
1966 | * Startup channels by sending CH_EVENT_START to each channel. | ||
1967 | * | ||
1968 | * @param fi An instance of an interface statemachine. | ||
1969 | * @param event The event, just happened. | ||
1970 | * @param arg Generic pointer, casted from struct net_device * upon call. | ||
1971 | */ | ||
1972 | static void | ||
1973 | dev_action_start(fsm_instance * fi, int event, void *arg) | ||
1974 | { | ||
1975 | struct net_device *dev = (struct net_device *) arg; | ||
1976 | struct ctc_priv *privptr = dev->priv; | ||
1977 | int direction; | ||
1978 | |||
1979 | DBF_TEXT(setup, 3, __FUNCTION__); | ||
1980 | fsm_deltimer(&privptr->restart_timer); | ||
1981 | fsm_newstate(fi, DEV_STATE_STARTWAIT_RXTX); | ||
1982 | for (direction = READ; direction <= WRITE; direction++) { | ||
1983 | struct channel *ch = privptr->channel[direction]; | ||
1984 | fsm_event(ch->fsm, CH_EVENT_START, ch); | ||
1985 | } | ||
1986 | } | ||
1987 | |||
1988 | /** | ||
1989 | * Shutdown channels by sending CH_EVENT_STOP to each channel. | ||
1990 | * | ||
1991 | * @param fi An instance of an interface statemachine. | ||
1992 | * @param event The event, just happened. | ||
1993 | * @param arg Generic pointer, casted from struct net_device * upon call. | ||
1994 | */ | ||
1995 | static void | ||
1996 | dev_action_stop(fsm_instance * fi, int event, void *arg) | ||
1997 | { | ||
1998 | struct net_device *dev = (struct net_device *) arg; | ||
1999 | struct ctc_priv *privptr = dev->priv; | ||
2000 | int direction; | ||
2001 | |||
2002 | DBF_TEXT(trace, 3, __FUNCTION__); | ||
2003 | fsm_newstate(fi, DEV_STATE_STOPWAIT_RXTX); | ||
2004 | for (direction = READ; direction <= WRITE; direction++) { | ||
2005 | struct channel *ch = privptr->channel[direction]; | ||
2006 | fsm_event(ch->fsm, CH_EVENT_STOP, ch); | ||
2007 | } | ||
2008 | } | ||
2009 | static void | ||
2010 | dev_action_restart(fsm_instance *fi, int event, void *arg) | ||
2011 | { | ||
2012 | struct net_device *dev = (struct net_device *)arg; | ||
2013 | struct ctc_priv *privptr = dev->priv; | ||
2014 | |||
2015 | DBF_TEXT(trace, 3, __FUNCTION__); | ||
2016 | ctc_pr_debug("%s: Restarting\n", dev->name); | ||
2017 | dev_action_stop(fi, event, arg); | ||
2018 | fsm_event(privptr->fsm, DEV_EVENT_STOP, dev); | ||
2019 | fsm_addtimer(&privptr->restart_timer, CTC_TIMEOUT_5SEC, | ||
2020 | DEV_EVENT_START, dev); | ||
2021 | } | ||
2022 | |||
2023 | /** | ||
2024 | * Called from channel statemachine | ||
2025 | * when a channel is up and running. | ||
2026 | * | ||
2027 | * @param fi An instance of an interface statemachine. | ||
2028 | * @param event The event, just happened. | ||
2029 | * @param arg Generic pointer, casted from struct net_device * upon call. | ||
2030 | */ | ||
2031 | static void | ||
2032 | dev_action_chup(fsm_instance * fi, int event, void *arg) | ||
2033 | { | ||
2034 | struct net_device *dev = (struct net_device *) arg; | ||
2035 | |||
2036 | DBF_TEXT(trace, 3, __FUNCTION__); | ||
2037 | switch (fsm_getstate(fi)) { | ||
2038 | case DEV_STATE_STARTWAIT_RXTX: | ||
2039 | if (event == DEV_EVENT_RXUP) | ||
2040 | fsm_newstate(fi, DEV_STATE_STARTWAIT_TX); | ||
2041 | else | ||
2042 | fsm_newstate(fi, DEV_STATE_STARTWAIT_RX); | ||
2043 | break; | ||
2044 | case DEV_STATE_STARTWAIT_RX: | ||
2045 | if (event == DEV_EVENT_RXUP) { | ||
2046 | fsm_newstate(fi, DEV_STATE_RUNNING); | ||
2047 | ctc_pr_info("%s: connected with remote side\n", | ||
2048 | dev->name); | ||
2049 | ctc_clear_busy(dev); | ||
2050 | } | ||
2051 | break; | ||
2052 | case DEV_STATE_STARTWAIT_TX: | ||
2053 | if (event == DEV_EVENT_TXUP) { | ||
2054 | fsm_newstate(fi, DEV_STATE_RUNNING); | ||
2055 | ctc_pr_info("%s: connected with remote side\n", | ||
2056 | dev->name); | ||
2057 | ctc_clear_busy(dev); | ||
2058 | } | ||
2059 | break; | ||
2060 | case DEV_STATE_STOPWAIT_TX: | ||
2061 | if (event == DEV_EVENT_RXUP) | ||
2062 | fsm_newstate(fi, DEV_STATE_STOPWAIT_RXTX); | ||
2063 | break; | ||
2064 | case DEV_STATE_STOPWAIT_RX: | ||
2065 | if (event == DEV_EVENT_TXUP) | ||
2066 | fsm_newstate(fi, DEV_STATE_STOPWAIT_RXTX); | ||
2067 | break; | ||
2068 | } | ||
2069 | } | ||
2070 | |||
2071 | /** | ||
2072 | * Called from channel statemachine | ||
2073 | * when a channel has been shutdown. | ||
2074 | * | ||
2075 | * @param fi An instance of an interface statemachine. | ||
2076 | * @param event The event, just happened. | ||
2077 | * @param arg Generic pointer, casted from struct net_device * upon call. | ||
2078 | */ | ||
2079 | static void | ||
2080 | dev_action_chdown(fsm_instance * fi, int event, void *arg) | ||
2081 | { | ||
2082 | |||
2083 | DBF_TEXT(trace, 3, __FUNCTION__); | ||
2084 | switch (fsm_getstate(fi)) { | ||
2085 | case DEV_STATE_RUNNING: | ||
2086 | if (event == DEV_EVENT_TXDOWN) | ||
2087 | fsm_newstate(fi, DEV_STATE_STARTWAIT_TX); | ||
2088 | else | ||
2089 | fsm_newstate(fi, DEV_STATE_STARTWAIT_RX); | ||
2090 | break; | ||
2091 | case DEV_STATE_STARTWAIT_RX: | ||
2092 | if (event == DEV_EVENT_TXDOWN) | ||
2093 | fsm_newstate(fi, DEV_STATE_STARTWAIT_RXTX); | ||
2094 | break; | ||
2095 | case DEV_STATE_STARTWAIT_TX: | ||
2096 | if (event == DEV_EVENT_RXDOWN) | ||
2097 | fsm_newstate(fi, DEV_STATE_STARTWAIT_RXTX); | ||
2098 | break; | ||
2099 | case DEV_STATE_STOPWAIT_RXTX: | ||
2100 | if (event == DEV_EVENT_TXDOWN) | ||
2101 | fsm_newstate(fi, DEV_STATE_STOPWAIT_RX); | ||
2102 | else | ||
2103 | fsm_newstate(fi, DEV_STATE_STOPWAIT_TX); | ||
2104 | break; | ||
2105 | case DEV_STATE_STOPWAIT_RX: | ||
2106 | if (event == DEV_EVENT_RXDOWN) | ||
2107 | fsm_newstate(fi, DEV_STATE_STOPPED); | ||
2108 | break; | ||
2109 | case DEV_STATE_STOPWAIT_TX: | ||
2110 | if (event == DEV_EVENT_TXDOWN) | ||
2111 | fsm_newstate(fi, DEV_STATE_STOPPED); | ||
2112 | break; | ||
2113 | } | ||
2114 | } | ||
2115 | |||
2116 | static const fsm_node dev_fsm[] = { | ||
2117 | {DEV_STATE_STOPPED, DEV_EVENT_START, dev_action_start}, | ||
2118 | |||
2119 | {DEV_STATE_STOPWAIT_RXTX, DEV_EVENT_START, dev_action_start }, | ||
2120 | {DEV_STATE_STOPWAIT_RXTX, DEV_EVENT_RXDOWN, dev_action_chdown }, | ||
2121 | {DEV_STATE_STOPWAIT_RXTX, DEV_EVENT_TXDOWN, dev_action_chdown }, | ||
2122 | {DEV_STATE_STOPWAIT_RXTX, DEV_EVENT_RESTART, dev_action_restart }, | ||
2123 | |||
2124 | {DEV_STATE_STOPWAIT_RX, DEV_EVENT_START, dev_action_start }, | ||
2125 | {DEV_STATE_STOPWAIT_RX, DEV_EVENT_RXUP, dev_action_chup }, | ||
2126 | {DEV_STATE_STOPWAIT_RX, DEV_EVENT_TXUP, dev_action_chup }, | ||
2127 | {DEV_STATE_STOPWAIT_RX, DEV_EVENT_RXDOWN, dev_action_chdown }, | ||
2128 | {DEV_STATE_STOPWAIT_RX, DEV_EVENT_RESTART, dev_action_restart }, | ||
2129 | |||
2130 | {DEV_STATE_STOPWAIT_TX, DEV_EVENT_START, dev_action_start }, | ||
2131 | {DEV_STATE_STOPWAIT_TX, DEV_EVENT_RXUP, dev_action_chup }, | ||
2132 | {DEV_STATE_STOPWAIT_TX, DEV_EVENT_TXUP, dev_action_chup }, | ||
2133 | {DEV_STATE_STOPWAIT_TX, DEV_EVENT_TXDOWN, dev_action_chdown }, | ||
2134 | {DEV_STATE_STOPWAIT_TX, DEV_EVENT_RESTART, dev_action_restart }, | ||
2135 | |||
2136 | {DEV_STATE_STARTWAIT_RXTX, DEV_EVENT_STOP, dev_action_stop }, | ||
2137 | {DEV_STATE_STARTWAIT_RXTX, DEV_EVENT_RXUP, dev_action_chup }, | ||
2138 | {DEV_STATE_STARTWAIT_RXTX, DEV_EVENT_TXUP, dev_action_chup }, | ||
2139 | {DEV_STATE_STARTWAIT_RXTX, DEV_EVENT_RXDOWN, dev_action_chdown }, | ||
2140 | {DEV_STATE_STARTWAIT_RXTX, DEV_EVENT_TXDOWN, dev_action_chdown }, | ||
2141 | {DEV_STATE_STARTWAIT_RXTX, DEV_EVENT_RESTART, dev_action_restart }, | ||
2142 | |||
2143 | {DEV_STATE_STARTWAIT_TX, DEV_EVENT_STOP, dev_action_stop }, | ||
2144 | {DEV_STATE_STARTWAIT_TX, DEV_EVENT_RXUP, dev_action_chup }, | ||
2145 | {DEV_STATE_STARTWAIT_TX, DEV_EVENT_TXUP, dev_action_chup }, | ||
2146 | {DEV_STATE_STARTWAIT_TX, DEV_EVENT_RXDOWN, dev_action_chdown }, | ||
2147 | {DEV_STATE_STARTWAIT_TX, DEV_EVENT_RESTART, dev_action_restart }, | ||
2148 | |||
2149 | {DEV_STATE_STARTWAIT_RX, DEV_EVENT_STOP, dev_action_stop }, | ||
2150 | {DEV_STATE_STARTWAIT_RX, DEV_EVENT_RXUP, dev_action_chup }, | ||
2151 | {DEV_STATE_STARTWAIT_RX, DEV_EVENT_TXUP, dev_action_chup }, | ||
2152 | {DEV_STATE_STARTWAIT_RX, DEV_EVENT_TXDOWN, dev_action_chdown }, | ||
2153 | {DEV_STATE_STARTWAIT_RX, DEV_EVENT_RESTART, dev_action_restart }, | ||
2154 | |||
2155 | {DEV_STATE_RUNNING, DEV_EVENT_STOP, dev_action_stop }, | ||
2156 | {DEV_STATE_RUNNING, DEV_EVENT_RXDOWN, dev_action_chdown }, | ||
2157 | {DEV_STATE_RUNNING, DEV_EVENT_TXDOWN, dev_action_chdown }, | ||
2158 | {DEV_STATE_RUNNING, DEV_EVENT_TXUP, fsm_action_nop }, | ||
2159 | {DEV_STATE_RUNNING, DEV_EVENT_RXUP, fsm_action_nop }, | ||
2160 | {DEV_STATE_RUNNING, DEV_EVENT_RESTART, dev_action_restart }, | ||
2161 | }; | ||
2162 | |||
2163 | static const int DEV_FSM_LEN = sizeof (dev_fsm) / sizeof (fsm_node); | ||
2164 | |||
2165 | /** | ||
2166 | * Transmit a packet. | ||
2167 | * This is a helper function for ctc_tx(). | ||
2168 | * | ||
2169 | * @param ch Channel to be used for sending. | ||
2170 | * @param skb Pointer to struct sk_buff of packet to send. | ||
2171 | * The linklevel header has already been set up | ||
2172 | * by ctc_tx(). | ||
2173 | * | ||
2174 | * @return 0 on success, -ERRNO on failure. (Never fails.) | ||
2175 | */ | ||
2176 | static int | ||
2177 | transmit_skb(struct channel *ch, struct sk_buff *skb) | ||
2178 | { | ||
2179 | unsigned long saveflags; | ||
2180 | struct ll_header header; | ||
2181 | int rc = 0; | ||
2182 | |||
2183 | DBF_TEXT(trace, 5, __FUNCTION__); | ||
2184 | /* we need to acquire the lock for testing the state | ||
2185 | * otherwise we can have an IRQ changing the state to | ||
2186 | * TXIDLE after the test but before acquiring the lock. | ||
2187 | */ | ||
2188 | spin_lock_irqsave(&ch->collect_lock, saveflags); | ||
2189 | if (fsm_getstate(ch->fsm) != CH_STATE_TXIDLE) { | ||
2190 | int l = skb->len + LL_HEADER_LENGTH; | ||
2191 | |||
2192 | if (ch->collect_len + l > ch->max_bufsize - 2) { | ||
2193 | spin_unlock_irqrestore(&ch->collect_lock, saveflags); | ||
2194 | return -EBUSY; | ||
2195 | } else { | ||
2196 | atomic_inc(&skb->users); | ||
2197 | header.length = l; | ||
2198 | header.type = skb->protocol; | ||
2199 | header.unused = 0; | ||
2200 | memcpy(skb_push(skb, LL_HEADER_LENGTH), &header, | ||
2201 | LL_HEADER_LENGTH); | ||
2202 | skb_queue_tail(&ch->collect_queue, skb); | ||
2203 | ch->collect_len += l; | ||
2204 | } | ||
2205 | spin_unlock_irqrestore(&ch->collect_lock, saveflags); | ||
2206 | } else { | ||
2207 | __u16 block_len; | ||
2208 | int ccw_idx; | ||
2209 | struct sk_buff *nskb; | ||
2210 | unsigned long hi; | ||
2211 | spin_unlock_irqrestore(&ch->collect_lock, saveflags); | ||
2212 | /** | ||
2213 | * Protect skb against beeing free'd by upper | ||
2214 | * layers. | ||
2215 | */ | ||
2216 | atomic_inc(&skb->users); | ||
2217 | ch->prof.txlen += skb->len; | ||
2218 | header.length = skb->len + LL_HEADER_LENGTH; | ||
2219 | header.type = skb->protocol; | ||
2220 | header.unused = 0; | ||
2221 | memcpy(skb_push(skb, LL_HEADER_LENGTH), &header, | ||
2222 | LL_HEADER_LENGTH); | ||
2223 | block_len = skb->len + 2; | ||
2224 | *((__u16 *) skb_push(skb, 2)) = block_len; | ||
2225 | |||
2226 | /** | ||
2227 | * IDAL support in CTC is broken, so we have to | ||
2228 | * care about skb's above 2G ourselves. | ||
2229 | */ | ||
2230 | hi = ((unsigned long)skb_tail_pointer(skb) + | ||
2231 | LL_HEADER_LENGTH) >> 31; | ||
2232 | if (hi) { | ||
2233 | nskb = alloc_skb(skb->len, GFP_ATOMIC | GFP_DMA); | ||
2234 | if (!nskb) { | ||
2235 | atomic_dec(&skb->users); | ||
2236 | skb_pull(skb, LL_HEADER_LENGTH + 2); | ||
2237 | ctc_clear_busy(ch->netdev); | ||
2238 | return -ENOMEM; | ||
2239 | } else { | ||
2240 | memcpy(skb_put(nskb, skb->len), | ||
2241 | skb->data, skb->len); | ||
2242 | atomic_inc(&nskb->users); | ||
2243 | atomic_dec(&skb->users); | ||
2244 | dev_kfree_skb_irq(skb); | ||
2245 | skb = nskb; | ||
2246 | } | ||
2247 | } | ||
2248 | |||
2249 | ch->ccw[4].count = block_len; | ||
2250 | if (set_normalized_cda(&ch->ccw[4], skb->data)) { | ||
2251 | /** | ||
2252 | * idal allocation failed, try via copying to | ||
2253 | * trans_skb. trans_skb usually has a pre-allocated | ||
2254 | * idal. | ||
2255 | */ | ||
2256 | if (ctc_checkalloc_buffer(ch, 1)) { | ||
2257 | /** | ||
2258 | * Remove our header. It gets added | ||
2259 | * again on retransmit. | ||
2260 | */ | ||
2261 | atomic_dec(&skb->users); | ||
2262 | skb_pull(skb, LL_HEADER_LENGTH + 2); | ||
2263 | ctc_clear_busy(ch->netdev); | ||
2264 | return -EBUSY; | ||
2265 | } | ||
2266 | |||
2267 | skb_reset_tail_pointer(ch->trans_skb); | ||
2268 | ch->trans_skb->len = 0; | ||
2269 | ch->ccw[1].count = skb->len; | ||
2270 | skb_copy_from_linear_data(skb, skb_put(ch->trans_skb, | ||
2271 | skb->len), | ||
2272 | skb->len); | ||
2273 | atomic_dec(&skb->users); | ||
2274 | dev_kfree_skb_irq(skb); | ||
2275 | ccw_idx = 0; | ||
2276 | } else { | ||
2277 | skb_queue_tail(&ch->io_queue, skb); | ||
2278 | ccw_idx = 3; | ||
2279 | } | ||
2280 | ch->retry = 0; | ||
2281 | fsm_newstate(ch->fsm, CH_STATE_TX); | ||
2282 | fsm_addtimer(&ch->timer, CTC_TIMEOUT_5SEC, CH_EVENT_TIMER, ch); | ||
2283 | spin_lock_irqsave(get_ccwdev_lock(ch->cdev), saveflags); | ||
2284 | ch->prof.send_stamp = current_kernel_time(); | ||
2285 | rc = ccw_device_start(ch->cdev, &ch->ccw[ccw_idx], | ||
2286 | (unsigned long) ch, 0xff, 0); | ||
2287 | spin_unlock_irqrestore(get_ccwdev_lock(ch->cdev), saveflags); | ||
2288 | if (ccw_idx == 3) | ||
2289 | ch->prof.doios_single++; | ||
2290 | if (rc != 0) { | ||
2291 | fsm_deltimer(&ch->timer); | ||
2292 | ccw_check_return_code(ch, rc, "single skb TX"); | ||
2293 | if (ccw_idx == 3) | ||
2294 | skb_dequeue_tail(&ch->io_queue); | ||
2295 | /** | ||
2296 | * Remove our header. It gets added | ||
2297 | * again on retransmit. | ||
2298 | */ | ||
2299 | skb_pull(skb, LL_HEADER_LENGTH + 2); | ||
2300 | } else { | ||
2301 | if (ccw_idx == 0) { | ||
2302 | struct net_device *dev = ch->netdev; | ||
2303 | struct ctc_priv *privptr = dev->priv; | ||
2304 | privptr->stats.tx_packets++; | ||
2305 | privptr->stats.tx_bytes += | ||
2306 | skb->len - LL_HEADER_LENGTH; | ||
2307 | } | ||
2308 | } | ||
2309 | } | ||
2310 | |||
2311 | ctc_clear_busy(ch->netdev); | ||
2312 | return rc; | ||
2313 | } | ||
2314 | |||
2315 | /** | ||
2316 | * Interface API for upper network layers | ||
2317 | *****************************************************************************/ | ||
2318 | |||
2319 | /** | ||
2320 | * Open an interface. | ||
2321 | * Called from generic network layer when ifconfig up is run. | ||
2322 | * | ||
2323 | * @param dev Pointer to interface struct. | ||
2324 | * | ||
2325 | * @return 0 on success, -ERRNO on failure. (Never fails.) | ||
2326 | */ | ||
2327 | static int | ||
2328 | ctc_open(struct net_device * dev) | ||
2329 | { | ||
2330 | DBF_TEXT(trace, 5, __FUNCTION__); | ||
2331 | fsm_event(((struct ctc_priv *) dev->priv)->fsm, DEV_EVENT_START, dev); | ||
2332 | return 0; | ||
2333 | } | ||
2334 | |||
2335 | /** | ||
2336 | * Close an interface. | ||
2337 | * Called from generic network layer when ifconfig down is run. | ||
2338 | * | ||
2339 | * @param dev Pointer to interface struct. | ||
2340 | * | ||
2341 | * @return 0 on success, -ERRNO on failure. (Never fails.) | ||
2342 | */ | ||
2343 | static int | ||
2344 | ctc_close(struct net_device * dev) | ||
2345 | { | ||
2346 | DBF_TEXT(trace, 5, __FUNCTION__); | ||
2347 | fsm_event(((struct ctc_priv *) dev->priv)->fsm, DEV_EVENT_STOP, dev); | ||
2348 | return 0; | ||
2349 | } | ||
2350 | |||
2351 | /** | ||
2352 | * Start transmission of a packet. | ||
2353 | * Called from generic network device layer. | ||
2354 | * | ||
2355 | * @param skb Pointer to buffer containing the packet. | ||
2356 | * @param dev Pointer to interface struct. | ||
2357 | * | ||
2358 | * @return 0 if packet consumed, !0 if packet rejected. | ||
2359 | * Note: If we return !0, then the packet is free'd by | ||
2360 | * the generic network layer. | ||
2361 | */ | ||
2362 | static int | ||
2363 | ctc_tx(struct sk_buff *skb, struct net_device * dev) | ||
2364 | { | ||
2365 | int rc = 0; | ||
2366 | struct ctc_priv *privptr = (struct ctc_priv *) dev->priv; | ||
2367 | |||
2368 | DBF_TEXT(trace, 5, __FUNCTION__); | ||
2369 | /** | ||
2370 | * Some sanity checks ... | ||
2371 | */ | ||
2372 | if (skb == NULL) { | ||
2373 | ctc_pr_warn("%s: NULL sk_buff passed\n", dev->name); | ||
2374 | privptr->stats.tx_dropped++; | ||
2375 | return 0; | ||
2376 | } | ||
2377 | if (skb_headroom(skb) < (LL_HEADER_LENGTH + 2)) { | ||
2378 | ctc_pr_warn("%s: Got sk_buff with head room < %ld bytes\n", | ||
2379 | dev->name, LL_HEADER_LENGTH + 2); | ||
2380 | dev_kfree_skb(skb); | ||
2381 | privptr->stats.tx_dropped++; | ||
2382 | return 0; | ||
2383 | } | ||
2384 | |||
2385 | /** | ||
2386 | * If channels are not running, try to restart them | ||
2387 | * and throw away packet. | ||
2388 | */ | ||
2389 | if (fsm_getstate(privptr->fsm) != DEV_STATE_RUNNING) { | ||
2390 | fsm_event(privptr->fsm, DEV_EVENT_START, dev); | ||
2391 | dev_kfree_skb(skb); | ||
2392 | privptr->stats.tx_dropped++; | ||
2393 | privptr->stats.tx_errors++; | ||
2394 | privptr->stats.tx_carrier_errors++; | ||
2395 | return 0; | ||
2396 | } | ||
2397 | |||
2398 | if (ctc_test_and_set_busy(dev)) | ||
2399 | return -EBUSY; | ||
2400 | |||
2401 | dev->trans_start = jiffies; | ||
2402 | if (transmit_skb(privptr->channel[WRITE], skb) != 0) | ||
2403 | rc = 1; | ||
2404 | return rc; | ||
2405 | } | ||
2406 | |||
2407 | /** | ||
2408 | * Sets MTU of an interface. | ||
2409 | * | ||
2410 | * @param dev Pointer to interface struct. | ||
2411 | * @param new_mtu The new MTU to use for this interface. | ||
2412 | * | ||
2413 | * @return 0 on success, -EINVAL if MTU is out of valid range. | ||
2414 | * (valid range is 576 .. 65527). If VM is on the | ||
2415 | * remote side, maximum MTU is 32760, however this is | ||
2416 | * <em>not</em> checked here. | ||
2417 | */ | ||
2418 | static int | ||
2419 | ctc_change_mtu(struct net_device * dev, int new_mtu) | ||
2420 | { | ||
2421 | struct ctc_priv *privptr = (struct ctc_priv *) dev->priv; | ||
2422 | |||
2423 | DBF_TEXT(trace, 3, __FUNCTION__); | ||
2424 | if ((new_mtu < 576) || (new_mtu > 65527) || | ||
2425 | (new_mtu > (privptr->channel[READ]->max_bufsize - | ||
2426 | LL_HEADER_LENGTH - 2))) | ||
2427 | return -EINVAL; | ||
2428 | dev->mtu = new_mtu; | ||
2429 | dev->hard_header_len = LL_HEADER_LENGTH + 2; | ||
2430 | return 0; | ||
2431 | } | ||
2432 | |||
2433 | /** | ||
2434 | * Returns interface statistics of a device. | ||
2435 | * | ||
2436 | * @param dev Pointer to interface struct. | ||
2437 | * | ||
2438 | * @return Pointer to stats struct of this interface. | ||
2439 | */ | ||
2440 | static struct net_device_stats * | ||
2441 | ctc_stats(struct net_device * dev) | ||
2442 | { | ||
2443 | return &((struct ctc_priv *) dev->priv)->stats; | ||
2444 | } | ||
2445 | |||
2446 | /* | ||
2447 | * sysfs attributes | ||
2448 | */ | ||
2449 | |||
2450 | static ssize_t | ||
2451 | buffer_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
2452 | { | ||
2453 | struct ctc_priv *priv; | ||
2454 | |||
2455 | priv = dev->driver_data; | ||
2456 | if (!priv) | ||
2457 | return -ENODEV; | ||
2458 | return sprintf(buf, "%d\n", | ||
2459 | priv->buffer_size); | ||
2460 | } | ||
2461 | |||
2462 | static ssize_t | ||
2463 | buffer_write(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
2464 | { | ||
2465 | struct ctc_priv *priv; | ||
2466 | struct net_device *ndev; | ||
2467 | int bs1; | ||
2468 | char buffer[16]; | ||
2469 | |||
2470 | DBF_TEXT(trace, 3, __FUNCTION__); | ||
2471 | DBF_TEXT(trace, 3, buf); | ||
2472 | priv = dev->driver_data; | ||
2473 | if (!priv) { | ||
2474 | DBF_TEXT(trace, 3, "bfnopriv"); | ||
2475 | return -ENODEV; | ||
2476 | } | ||
2477 | |||
2478 | sscanf(buf, "%u", &bs1); | ||
2479 | if (bs1 > CTC_BUFSIZE_LIMIT) | ||
2480 | goto einval; | ||
2481 | if (bs1 < (576 + LL_HEADER_LENGTH + 2)) | ||
2482 | goto einval; | ||
2483 | priv->buffer_size = bs1; // just to overwrite the default | ||
2484 | |||
2485 | ndev = priv->channel[READ]->netdev; | ||
2486 | if (!ndev) { | ||
2487 | DBF_TEXT(trace, 3, "bfnondev"); | ||
2488 | return -ENODEV; | ||
2489 | } | ||
2490 | |||
2491 | if ((ndev->flags & IFF_RUNNING) && | ||
2492 | (bs1 < (ndev->mtu + LL_HEADER_LENGTH + 2))) | ||
2493 | goto einval; | ||
2494 | |||
2495 | priv->channel[READ]->max_bufsize = bs1; | ||
2496 | priv->channel[WRITE]->max_bufsize = bs1; | ||
2497 | if (!(ndev->flags & IFF_RUNNING)) | ||
2498 | ndev->mtu = bs1 - LL_HEADER_LENGTH - 2; | ||
2499 | priv->channel[READ]->flags |= CHANNEL_FLAGS_BUFSIZE_CHANGED; | ||
2500 | priv->channel[WRITE]->flags |= CHANNEL_FLAGS_BUFSIZE_CHANGED; | ||
2501 | |||
2502 | sprintf(buffer, "%d",priv->buffer_size); | ||
2503 | DBF_TEXT(trace, 3, buffer); | ||
2504 | return count; | ||
2505 | |||
2506 | einval: | ||
2507 | DBF_TEXT(trace, 3, "buff_err"); | ||
2508 | return -EINVAL; | ||
2509 | } | ||
2510 | |||
2511 | static ssize_t | ||
2512 | loglevel_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
2513 | { | ||
2514 | return sprintf(buf, "%d\n", loglevel); | ||
2515 | } | ||
2516 | |||
2517 | static ssize_t | ||
2518 | loglevel_write(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
2519 | { | ||
2520 | int ll1; | ||
2521 | |||
2522 | DBF_TEXT(trace, 5, __FUNCTION__); | ||
2523 | sscanf(buf, "%i", &ll1); | ||
2524 | |||
2525 | if ((ll1 > CTC_LOGLEVEL_MAX) || (ll1 < 0)) | ||
2526 | return -EINVAL; | ||
2527 | loglevel = ll1; | ||
2528 | return count; | ||
2529 | } | ||
2530 | |||
2531 | static void | ||
2532 | ctc_print_statistics(struct ctc_priv *priv) | ||
2533 | { | ||
2534 | char *sbuf; | ||
2535 | char *p; | ||
2536 | |||
2537 | DBF_TEXT(trace, 4, __FUNCTION__); | ||
2538 | if (!priv) | ||
2539 | return; | ||
2540 | sbuf = kmalloc(2048, GFP_KERNEL); | ||
2541 | if (sbuf == NULL) | ||
2542 | return; | ||
2543 | p = sbuf; | ||
2544 | |||
2545 | p += sprintf(p, " Device FSM state: %s\n", | ||
2546 | fsm_getstate_str(priv->fsm)); | ||
2547 | p += sprintf(p, " RX channel FSM state: %s\n", | ||
2548 | fsm_getstate_str(priv->channel[READ]->fsm)); | ||
2549 | p += sprintf(p, " TX channel FSM state: %s\n", | ||
2550 | fsm_getstate_str(priv->channel[WRITE]->fsm)); | ||
2551 | p += sprintf(p, " Max. TX buffer used: %ld\n", | ||
2552 | priv->channel[WRITE]->prof.maxmulti); | ||
2553 | p += sprintf(p, " Max. chained SKBs: %ld\n", | ||
2554 | priv->channel[WRITE]->prof.maxcqueue); | ||
2555 | p += sprintf(p, " TX single write ops: %ld\n", | ||
2556 | priv->channel[WRITE]->prof.doios_single); | ||
2557 | p += sprintf(p, " TX multi write ops: %ld\n", | ||
2558 | priv->channel[WRITE]->prof.doios_multi); | ||
2559 | p += sprintf(p, " Netto bytes written: %ld\n", | ||
2560 | priv->channel[WRITE]->prof.txlen); | ||
2561 | p += sprintf(p, " Max. TX IO-time: %ld\n", | ||
2562 | priv->channel[WRITE]->prof.tx_time); | ||
2563 | |||
2564 | ctc_pr_debug("Statistics for %s:\n%s", | ||
2565 | priv->channel[WRITE]->netdev->name, sbuf); | ||
2566 | kfree(sbuf); | ||
2567 | return; | ||
2568 | } | ||
2569 | |||
2570 | static ssize_t | ||
2571 | stats_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
2572 | { | ||
2573 | struct ctc_priv *priv = dev->driver_data; | ||
2574 | if (!priv) | ||
2575 | return -ENODEV; | ||
2576 | ctc_print_statistics(priv); | ||
2577 | return sprintf(buf, "0\n"); | ||
2578 | } | ||
2579 | |||
2580 | static ssize_t | ||
2581 | stats_write(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
2582 | { | ||
2583 | struct ctc_priv *priv = dev->driver_data; | ||
2584 | if (!priv) | ||
2585 | return -ENODEV; | ||
2586 | /* Reset statistics */ | ||
2587 | memset(&priv->channel[WRITE]->prof, 0, | ||
2588 | sizeof(priv->channel[WRITE]->prof)); | ||
2589 | return count; | ||
2590 | } | ||
2591 | |||
2592 | static void | ||
2593 | ctc_netdev_unregister(struct net_device * dev) | ||
2594 | { | ||
2595 | struct ctc_priv *privptr; | ||
2596 | |||
2597 | if (!dev) | ||
2598 | return; | ||
2599 | privptr = (struct ctc_priv *) dev->priv; | ||
2600 | unregister_netdev(dev); | ||
2601 | } | ||
2602 | |||
2603 | static int | ||
2604 | ctc_netdev_register(struct net_device * dev) | ||
2605 | { | ||
2606 | return register_netdev(dev); | ||
2607 | } | ||
2608 | |||
2609 | static void | ||
2610 | ctc_free_netdevice(struct net_device * dev, int free_dev) | ||
2611 | { | ||
2612 | struct ctc_priv *privptr; | ||
2613 | if (!dev) | ||
2614 | return; | ||
2615 | privptr = dev->priv; | ||
2616 | if (privptr) { | ||
2617 | if (privptr->fsm) | ||
2618 | kfree_fsm(privptr->fsm); | ||
2619 | kfree(privptr); | ||
2620 | } | ||
2621 | #ifdef MODULE | ||
2622 | if (free_dev) | ||
2623 | free_netdev(dev); | ||
2624 | #endif | ||
2625 | } | ||
2626 | |||
2627 | static ssize_t | ||
2628 | ctc_proto_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
2629 | { | ||
2630 | struct ctc_priv *priv; | ||
2631 | |||
2632 | priv = dev->driver_data; | ||
2633 | if (!priv) | ||
2634 | return -ENODEV; | ||
2635 | |||
2636 | return sprintf(buf, "%d\n", priv->protocol); | ||
2637 | } | ||
2638 | |||
2639 | static ssize_t | ||
2640 | ctc_proto_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
2641 | { | ||
2642 | struct ctc_priv *priv; | ||
2643 | int value; | ||
2644 | |||
2645 | DBF_TEXT(trace, 3, __FUNCTION__); | ||
2646 | pr_debug("%s() called\n", __FUNCTION__); | ||
2647 | |||
2648 | priv = dev->driver_data; | ||
2649 | if (!priv) | ||
2650 | return -ENODEV; | ||
2651 | sscanf(buf, "%u", &value); | ||
2652 | if (!((value == CTC_PROTO_S390) || | ||
2653 | (value == CTC_PROTO_LINUX) || | ||
2654 | (value == CTC_PROTO_OS390))) | ||
2655 | return -EINVAL; | ||
2656 | priv->protocol = value; | ||
2657 | |||
2658 | return count; | ||
2659 | } | ||
2660 | |||
2661 | static ssize_t | ||
2662 | ctc_type_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
2663 | { | ||
2664 | struct ccwgroup_device *cgdev; | ||
2665 | |||
2666 | cgdev = to_ccwgroupdev(dev); | ||
2667 | if (!cgdev) | ||
2668 | return -ENODEV; | ||
2669 | |||
2670 | return sprintf(buf, "%s\n", cu3088_type[cgdev->cdev[0]->id.driver_info]); | ||
2671 | } | ||
2672 | |||
2673 | static DEVICE_ATTR(buffer, 0644, buffer_show, buffer_write); | ||
2674 | static DEVICE_ATTR(protocol, 0644, ctc_proto_show, ctc_proto_store); | ||
2675 | static DEVICE_ATTR(type, 0444, ctc_type_show, NULL); | ||
2676 | |||
2677 | static DEVICE_ATTR(loglevel, 0644, loglevel_show, loglevel_write); | ||
2678 | static DEVICE_ATTR(stats, 0644, stats_show, stats_write); | ||
2679 | |||
2680 | static struct attribute *ctc_attr[] = { | ||
2681 | &dev_attr_protocol.attr, | ||
2682 | &dev_attr_type.attr, | ||
2683 | &dev_attr_buffer.attr, | ||
2684 | NULL, | ||
2685 | }; | ||
2686 | |||
2687 | static struct attribute_group ctc_attr_group = { | ||
2688 | .attrs = ctc_attr, | ||
2689 | }; | ||
2690 | |||
2691 | static int | ||
2692 | ctc_add_attributes(struct device *dev) | ||
2693 | { | ||
2694 | int rc; | ||
2695 | |||
2696 | rc = device_create_file(dev, &dev_attr_loglevel); | ||
2697 | if (rc) | ||
2698 | goto out; | ||
2699 | rc = device_create_file(dev, &dev_attr_stats); | ||
2700 | if (!rc) | ||
2701 | goto out; | ||
2702 | device_remove_file(dev, &dev_attr_loglevel); | ||
2703 | out: | ||
2704 | return rc; | ||
2705 | } | ||
2706 | |||
2707 | static void | ||
2708 | ctc_remove_attributes(struct device *dev) | ||
2709 | { | ||
2710 | device_remove_file(dev, &dev_attr_stats); | ||
2711 | device_remove_file(dev, &dev_attr_loglevel); | ||
2712 | } | ||
2713 | |||
2714 | static int | ||
2715 | ctc_add_files(struct device *dev) | ||
2716 | { | ||
2717 | pr_debug("%s() called\n", __FUNCTION__); | ||
2718 | |||
2719 | return sysfs_create_group(&dev->kobj, &ctc_attr_group); | ||
2720 | } | ||
2721 | |||
2722 | static void | ||
2723 | ctc_remove_files(struct device *dev) | ||
2724 | { | ||
2725 | pr_debug("%s() called\n", __FUNCTION__); | ||
2726 | |||
2727 | sysfs_remove_group(&dev->kobj, &ctc_attr_group); | ||
2728 | } | ||
2729 | |||
2730 | /** | ||
2731 | * Add ctc specific attributes. | ||
2732 | * Add ctc private data. | ||
2733 | * | ||
2734 | * @param cgdev pointer to ccwgroup_device just added | ||
2735 | * | ||
2736 | * @returns 0 on success, !0 on failure. | ||
2737 | */ | ||
2738 | static int | ||
2739 | ctc_probe_device(struct ccwgroup_device *cgdev) | ||
2740 | { | ||
2741 | struct ctc_priv *priv; | ||
2742 | int rc; | ||
2743 | char buffer[16]; | ||
2744 | |||
2745 | pr_debug("%s() called\n", __FUNCTION__); | ||
2746 | DBF_TEXT(setup, 3, __FUNCTION__); | ||
2747 | |||
2748 | if (!get_device(&cgdev->dev)) | ||
2749 | return -ENODEV; | ||
2750 | |||
2751 | priv = kzalloc(sizeof(struct ctc_priv), GFP_KERNEL); | ||
2752 | if (!priv) { | ||
2753 | ctc_pr_err("%s: Out of memory\n", __func__); | ||
2754 | put_device(&cgdev->dev); | ||
2755 | return -ENOMEM; | ||
2756 | } | ||
2757 | |||
2758 | rc = ctc_add_files(&cgdev->dev); | ||
2759 | if (rc) { | ||
2760 | kfree(priv); | ||
2761 | put_device(&cgdev->dev); | ||
2762 | return rc; | ||
2763 | } | ||
2764 | priv->buffer_size = CTC_BUFSIZE_DEFAULT; | ||
2765 | cgdev->cdev[0]->handler = ctc_irq_handler; | ||
2766 | cgdev->cdev[1]->handler = ctc_irq_handler; | ||
2767 | cgdev->dev.driver_data = priv; | ||
2768 | |||
2769 | sprintf(buffer, "%p", priv); | ||
2770 | DBF_TEXT(data, 3, buffer); | ||
2771 | |||
2772 | sprintf(buffer, "%u", (unsigned int)sizeof(struct ctc_priv)); | ||
2773 | DBF_TEXT(data, 3, buffer); | ||
2774 | |||
2775 | sprintf(buffer, "%p", &channels); | ||
2776 | DBF_TEXT(data, 3, buffer); | ||
2777 | |||
2778 | sprintf(buffer, "%u", (unsigned int)sizeof(struct channel)); | ||
2779 | DBF_TEXT(data, 3, buffer); | ||
2780 | |||
2781 | return 0; | ||
2782 | } | ||
2783 | |||
2784 | /** | ||
2785 | * Device setup function called by alloc_netdev(). | ||
2786 | * | ||
2787 | * @param dev Device to be setup. | ||
2788 | */ | ||
2789 | void ctc_init_netdevice(struct net_device * dev) | ||
2790 | { | ||
2791 | DBF_TEXT(setup, 3, __FUNCTION__); | ||
2792 | |||
2793 | if (dev->mtu == 0) | ||
2794 | dev->mtu = CTC_BUFSIZE_DEFAULT - LL_HEADER_LENGTH - 2; | ||
2795 | dev->hard_start_xmit = ctc_tx; | ||
2796 | dev->open = ctc_open; | ||
2797 | dev->stop = ctc_close; | ||
2798 | dev->get_stats = ctc_stats; | ||
2799 | dev->change_mtu = ctc_change_mtu; | ||
2800 | dev->hard_header_len = LL_HEADER_LENGTH + 2; | ||
2801 | dev->addr_len = 0; | ||
2802 | dev->type = ARPHRD_SLIP; | ||
2803 | dev->tx_queue_len = 100; | ||
2804 | dev->flags = IFF_POINTOPOINT | IFF_NOARP; | ||
2805 | } | ||
2806 | |||
2807 | |||
2808 | /** | ||
2809 | * | ||
2810 | * Setup an interface. | ||
2811 | * | ||
2812 | * @param cgdev Device to be setup. | ||
2813 | * | ||
2814 | * @returns 0 on success, !0 on failure. | ||
2815 | */ | ||
2816 | static int | ||
2817 | ctc_new_device(struct ccwgroup_device *cgdev) | ||
2818 | { | ||
2819 | char read_id[CTC_ID_SIZE]; | ||
2820 | char write_id[CTC_ID_SIZE]; | ||
2821 | int direction; | ||
2822 | enum channel_types type; | ||
2823 | struct ctc_priv *privptr; | ||
2824 | struct net_device *dev; | ||
2825 | int ret; | ||
2826 | char buffer[16]; | ||
2827 | |||
2828 | pr_debug("%s() called\n", __FUNCTION__); | ||
2829 | DBF_TEXT(setup, 3, __FUNCTION__); | ||
2830 | |||
2831 | privptr = cgdev->dev.driver_data; | ||
2832 | if (!privptr) | ||
2833 | return -ENODEV; | ||
2834 | |||
2835 | sprintf(buffer, "%d", privptr->buffer_size); | ||
2836 | DBF_TEXT(setup, 3, buffer); | ||
2837 | |||
2838 | type = get_channel_type(&cgdev->cdev[0]->id); | ||
2839 | |||
2840 | snprintf(read_id, CTC_ID_SIZE, "ch-%s", cgdev->cdev[0]->dev.bus_id); | ||
2841 | snprintf(write_id, CTC_ID_SIZE, "ch-%s", cgdev->cdev[1]->dev.bus_id); | ||
2842 | |||
2843 | if (add_channel(cgdev->cdev[0], type)) | ||
2844 | return -ENOMEM; | ||
2845 | if (add_channel(cgdev->cdev[1], type)) | ||
2846 | return -ENOMEM; | ||
2847 | |||
2848 | ret = ccw_device_set_online(cgdev->cdev[0]); | ||
2849 | if (ret != 0) { | ||
2850 | printk(KERN_WARNING | ||
2851 | "ccw_device_set_online (cdev[0]) failed with ret = %d\n", ret); | ||
2852 | } | ||
2853 | |||
2854 | ret = ccw_device_set_online(cgdev->cdev[1]); | ||
2855 | if (ret != 0) { | ||
2856 | printk(KERN_WARNING | ||
2857 | "ccw_device_set_online (cdev[1]) failed with ret = %d\n", ret); | ||
2858 | } | ||
2859 | |||
2860 | dev = alloc_netdev(0, "ctc%d", ctc_init_netdevice); | ||
2861 | if (!dev) { | ||
2862 | ctc_pr_warn("ctc_init_netdevice failed\n"); | ||
2863 | goto out; | ||
2864 | } | ||
2865 | dev->priv = privptr; | ||
2866 | |||
2867 | privptr->fsm = init_fsm("ctcdev", dev_state_names, | ||
2868 | dev_event_names, CTC_NR_DEV_STATES, CTC_NR_DEV_EVENTS, | ||
2869 | dev_fsm, DEV_FSM_LEN, GFP_KERNEL); | ||
2870 | if (privptr->fsm == NULL) { | ||
2871 | free_netdev(dev); | ||
2872 | goto out; | ||
2873 | } | ||
2874 | fsm_newstate(privptr->fsm, DEV_STATE_STOPPED); | ||
2875 | fsm_settimer(privptr->fsm, &privptr->restart_timer); | ||
2876 | |||
2877 | for (direction = READ; direction <= WRITE; direction++) { | ||
2878 | privptr->channel[direction] = | ||
2879 | channel_get(type, direction == READ ? read_id : write_id, | ||
2880 | direction); | ||
2881 | if (privptr->channel[direction] == NULL) { | ||
2882 | if (direction == WRITE) | ||
2883 | channel_free(privptr->channel[READ]); | ||
2884 | |||
2885 | ctc_free_netdevice(dev, 1); | ||
2886 | goto out; | ||
2887 | } | ||
2888 | privptr->channel[direction]->netdev = dev; | ||
2889 | privptr->channel[direction]->protocol = privptr->protocol; | ||
2890 | privptr->channel[direction]->max_bufsize = privptr->buffer_size; | ||
2891 | } | ||
2892 | /* sysfs magic */ | ||
2893 | SET_NETDEV_DEV(dev, &cgdev->dev); | ||
2894 | |||
2895 | if (ctc_netdev_register(dev) != 0) { | ||
2896 | ctc_free_netdevice(dev, 1); | ||
2897 | goto out; | ||
2898 | } | ||
2899 | |||
2900 | if (ctc_add_attributes(&cgdev->dev)) { | ||
2901 | ctc_netdev_unregister(dev); | ||
2902 | dev->priv = NULL; | ||
2903 | ctc_free_netdevice(dev, 1); | ||
2904 | goto out; | ||
2905 | } | ||
2906 | |||
2907 | strlcpy(privptr->fsm->name, dev->name, sizeof (privptr->fsm->name)); | ||
2908 | |||
2909 | print_banner(); | ||
2910 | |||
2911 | ctc_pr_info("%s: read: %s, write: %s, proto: %d\n", | ||
2912 | dev->name, privptr->channel[READ]->id, | ||
2913 | privptr->channel[WRITE]->id, privptr->protocol); | ||
2914 | |||
2915 | return 0; | ||
2916 | out: | ||
2917 | ccw_device_set_offline(cgdev->cdev[1]); | ||
2918 | ccw_device_set_offline(cgdev->cdev[0]); | ||
2919 | |||
2920 | return -ENODEV; | ||
2921 | } | ||
2922 | |||
2923 | /** | ||
2924 | * Shutdown an interface. | ||
2925 | * | ||
2926 | * @param cgdev Device to be shut down. | ||
2927 | * | ||
2928 | * @returns 0 on success, !0 on failure. | ||
2929 | */ | ||
2930 | static int | ||
2931 | ctc_shutdown_device(struct ccwgroup_device *cgdev) | ||
2932 | { | ||
2933 | struct ctc_priv *priv; | ||
2934 | struct net_device *ndev; | ||
2935 | |||
2936 | DBF_TEXT(setup, 3, __FUNCTION__); | ||
2937 | pr_debug("%s() called\n", __FUNCTION__); | ||
2938 | |||
2939 | |||
2940 | priv = cgdev->dev.driver_data; | ||
2941 | ndev = NULL; | ||
2942 | if (!priv) | ||
2943 | return -ENODEV; | ||
2944 | |||
2945 | if (priv->channel[READ]) { | ||
2946 | ndev = priv->channel[READ]->netdev; | ||
2947 | |||
2948 | /* Close the device */ | ||
2949 | ctc_close(ndev); | ||
2950 | ndev->flags &=~IFF_RUNNING; | ||
2951 | |||
2952 | ctc_remove_attributes(&cgdev->dev); | ||
2953 | |||
2954 | channel_free(priv->channel[READ]); | ||
2955 | } | ||
2956 | if (priv->channel[WRITE]) | ||
2957 | channel_free(priv->channel[WRITE]); | ||
2958 | |||
2959 | if (ndev) { | ||
2960 | ctc_netdev_unregister(ndev); | ||
2961 | ndev->priv = NULL; | ||
2962 | ctc_free_netdevice(ndev, 1); | ||
2963 | } | ||
2964 | |||
2965 | if (priv->fsm) | ||
2966 | kfree_fsm(priv->fsm); | ||
2967 | |||
2968 | ccw_device_set_offline(cgdev->cdev[1]); | ||
2969 | ccw_device_set_offline(cgdev->cdev[0]); | ||
2970 | |||
2971 | if (priv->channel[READ]) | ||
2972 | channel_remove(priv->channel[READ]); | ||
2973 | if (priv->channel[WRITE]) | ||
2974 | channel_remove(priv->channel[WRITE]); | ||
2975 | priv->channel[READ] = priv->channel[WRITE] = NULL; | ||
2976 | |||
2977 | return 0; | ||
2978 | |||
2979 | } | ||
2980 | |||
2981 | static void | ||
2982 | ctc_remove_device(struct ccwgroup_device *cgdev) | ||
2983 | { | ||
2984 | struct ctc_priv *priv; | ||
2985 | |||
2986 | pr_debug("%s() called\n", __FUNCTION__); | ||
2987 | DBF_TEXT(setup, 3, __FUNCTION__); | ||
2988 | |||
2989 | priv = cgdev->dev.driver_data; | ||
2990 | if (!priv) | ||
2991 | return; | ||
2992 | if (cgdev->state == CCWGROUP_ONLINE) | ||
2993 | ctc_shutdown_device(cgdev); | ||
2994 | ctc_remove_files(&cgdev->dev); | ||
2995 | cgdev->dev.driver_data = NULL; | ||
2996 | kfree(priv); | ||
2997 | put_device(&cgdev->dev); | ||
2998 | } | ||
2999 | |||
3000 | static struct ccwgroup_driver ctc_group_driver = { | ||
3001 | .owner = THIS_MODULE, | ||
3002 | .name = "ctc", | ||
3003 | .max_slaves = 2, | ||
3004 | .driver_id = 0xC3E3C3, | ||
3005 | .probe = ctc_probe_device, | ||
3006 | .remove = ctc_remove_device, | ||
3007 | .set_online = ctc_new_device, | ||
3008 | .set_offline = ctc_shutdown_device, | ||
3009 | }; | ||
3010 | |||
3011 | /** | ||
3012 | * Module related routines | ||
3013 | *****************************************************************************/ | ||
3014 | |||
3015 | /** | ||
3016 | * Prepare to be unloaded. Free IRQ's and release all resources. | ||
3017 | * This is called just before this module is unloaded. It is | ||
3018 | * <em>not</em> called, if the usage count is !0, so we don't need to check | ||
3019 | * for that. | ||
3020 | */ | ||
3021 | static void __exit | ||
3022 | ctc_exit(void) | ||
3023 | { | ||
3024 | DBF_TEXT(setup, 3, __FUNCTION__); | ||
3025 | unregister_cu3088_discipline(&ctc_group_driver); | ||
3026 | ctc_unregister_dbf_views(); | ||
3027 | ctc_pr_info("CTC driver unloaded\n"); | ||
3028 | } | ||
3029 | |||
3030 | /** | ||
3031 | * Initialize module. | ||
3032 | * This is called just after the module is loaded. | ||
3033 | * | ||
3034 | * @return 0 on success, !0 on error. | ||
3035 | */ | ||
3036 | static int __init | ||
3037 | ctc_init(void) | ||
3038 | { | ||
3039 | int ret = 0; | ||
3040 | |||
3041 | loglevel = CTC_LOGLEVEL_DEFAULT; | ||
3042 | |||
3043 | DBF_TEXT(setup, 3, __FUNCTION__); | ||
3044 | |||
3045 | print_banner(); | ||
3046 | |||
3047 | ret = ctc_register_dbf_views(); | ||
3048 | if (ret){ | ||
3049 | ctc_pr_crit("ctc_init failed with ctc_register_dbf_views rc = %d\n", ret); | ||
3050 | return ret; | ||
3051 | } | ||
3052 | ret = register_cu3088_discipline(&ctc_group_driver); | ||
3053 | if (ret) { | ||
3054 | ctc_unregister_dbf_views(); | ||
3055 | } | ||
3056 | return ret; | ||
3057 | } | ||
3058 | |||
3059 | module_init(ctc_init); | ||
3060 | module_exit(ctc_exit); | ||
3061 | |||
3062 | /* --- This is the END my friend --- */ | ||
diff --git a/drivers/s390/net/ctcmain.h b/drivers/s390/net/ctcmain.h deleted file mode 100644 index 7f305d119f3d..000000000000 --- a/drivers/s390/net/ctcmain.h +++ /dev/null | |||
@@ -1,270 +0,0 @@ | |||
1 | /* | ||
2 | * CTC / ESCON network driver | ||
3 | * | ||
4 | * Copyright (C) 2001 IBM Deutschland Entwicklung GmbH, IBM Corporation | ||
5 | * Author(s): Fritz Elfert (elfert@de.ibm.com, felfert@millenux.com) | ||
6 | Peter Tiedemann (ptiedem@de.ibm.com) | ||
7 | * | ||
8 | * | ||
9 | * Documentation used: | ||
10 | * - Principles of Operation (IBM doc#: SA22-7201-06) | ||
11 | * - Common IO/-Device Commands and Self Description (IBM doc#: SA22-7204-02) | ||
12 | * - Common IO/-Device Commands and Self Description (IBM doc#: SN22-5535) | ||
13 | * - ESCON Channel-to-Channel Adapter (IBM doc#: SA22-7203-00) | ||
14 | * - ESCON I/O Interface (IBM doc#: SA22-7202-029 | ||
15 | * | ||
16 | * This program is free software; you can redistribute it and/or modify | ||
17 | * it under the terms of the GNU General Public License as published by | ||
18 | * the Free Software Foundation; either version 2, or (at your option) | ||
19 | * any later version. | ||
20 | * | ||
21 | * This program is distributed in the hope that it will be useful, | ||
22 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | ||
23 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | ||
24 | * GNU General Public License for more details. | ||
25 | * | ||
26 | * You should have received a copy of the GNU General Public License | ||
27 | * along with this program; if not, write to the Free Software | ||
28 | * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. | ||
29 | * | ||
30 | */ | ||
31 | |||
32 | #ifndef _CTCMAIN_H_ | ||
33 | #define _CTCMAIN_H_ | ||
34 | |||
35 | #include <asm/ccwdev.h> | ||
36 | #include <asm/ccwgroup.h> | ||
37 | |||
38 | #include <linux/skbuff.h> | ||
39 | #include <linux/netdevice.h> | ||
40 | |||
41 | #include "fsm.h" | ||
42 | #include "cu3088.h" | ||
43 | |||
44 | |||
45 | /** | ||
46 | * CCW commands, used in this driver. | ||
47 | */ | ||
48 | #define CCW_CMD_WRITE 0x01 | ||
49 | #define CCW_CMD_READ 0x02 | ||
50 | #define CCW_CMD_SET_EXTENDED 0xc3 | ||
51 | #define CCW_CMD_PREPARE 0xe3 | ||
52 | |||
53 | #define CTC_PROTO_S390 0 | ||
54 | #define CTC_PROTO_LINUX 1 | ||
55 | #define CTC_PROTO_OS390 3 | ||
56 | |||
57 | #define CTC_BUFSIZE_LIMIT 65535 | ||
58 | #define CTC_BUFSIZE_DEFAULT 32768 | ||
59 | |||
60 | #define CTC_TIMEOUT_5SEC 5000 | ||
61 | |||
62 | #define CTC_INITIAL_BLOCKLEN 2 | ||
63 | |||
64 | #define READ 0 | ||
65 | #define WRITE 1 | ||
66 | |||
67 | #define CTC_ID_SIZE BUS_ID_SIZE+3 | ||
68 | |||
69 | |||
70 | struct ctc_profile { | ||
71 | unsigned long maxmulti; | ||
72 | unsigned long maxcqueue; | ||
73 | unsigned long doios_single; | ||
74 | unsigned long doios_multi; | ||
75 | unsigned long txlen; | ||
76 | unsigned long tx_time; | ||
77 | struct timespec send_stamp; | ||
78 | }; | ||
79 | |||
80 | /** | ||
81 | * Definition of one channel | ||
82 | */ | ||
83 | struct channel { | ||
84 | |||
85 | /** | ||
86 | * Pointer to next channel in list. | ||
87 | */ | ||
88 | struct channel *next; | ||
89 | char id[CTC_ID_SIZE]; | ||
90 | struct ccw_device *cdev; | ||
91 | |||
92 | /** | ||
93 | * Type of this channel. | ||
94 | * CTC/A or Escon for valid channels. | ||
95 | */ | ||
96 | enum channel_types type; | ||
97 | |||
98 | /** | ||
99 | * Misc. flags. See CHANNEL_FLAGS_... below | ||
100 | */ | ||
101 | __u32 flags; | ||
102 | |||
103 | /** | ||
104 | * The protocol of this channel | ||
105 | */ | ||
106 | __u16 protocol; | ||
107 | |||
108 | /** | ||
109 | * I/O and irq related stuff | ||
110 | */ | ||
111 | struct ccw1 *ccw; | ||
112 | struct irb *irb; | ||
113 | |||
114 | /** | ||
115 | * RX/TX buffer size | ||
116 | */ | ||
117 | int max_bufsize; | ||
118 | |||
119 | /** | ||
120 | * Transmit/Receive buffer. | ||
121 | */ | ||
122 | struct sk_buff *trans_skb; | ||
123 | |||
124 | /** | ||
125 | * Universal I/O queue. | ||
126 | */ | ||
127 | struct sk_buff_head io_queue; | ||
128 | |||
129 | /** | ||
130 | * TX queue for collecting skb's during busy. | ||
131 | */ | ||
132 | struct sk_buff_head collect_queue; | ||
133 | |||
134 | /** | ||
135 | * Amount of data in collect_queue. | ||
136 | */ | ||
137 | int collect_len; | ||
138 | |||
139 | /** | ||
140 | * spinlock for collect_queue and collect_len | ||
141 | */ | ||
142 | spinlock_t collect_lock; | ||
143 | |||
144 | /** | ||
145 | * Timer for detecting unresposive | ||
146 | * I/O operations. | ||
147 | */ | ||
148 | fsm_timer timer; | ||
149 | |||
150 | /** | ||
151 | * Retry counter for misc. operations. | ||
152 | */ | ||
153 | int retry; | ||
154 | |||
155 | /** | ||
156 | * The finite state machine of this channel | ||
157 | */ | ||
158 | fsm_instance *fsm; | ||
159 | |||
160 | /** | ||
161 | * The corresponding net_device this channel | ||
162 | * belongs to. | ||
163 | */ | ||
164 | struct net_device *netdev; | ||
165 | |||
166 | struct ctc_profile prof; | ||
167 | |||
168 | unsigned char *trans_skb_data; | ||
169 | |||
170 | __u16 logflags; | ||
171 | }; | ||
172 | |||
173 | #define CHANNEL_FLAGS_READ 0 | ||
174 | #define CHANNEL_FLAGS_WRITE 1 | ||
175 | #define CHANNEL_FLAGS_INUSE 2 | ||
176 | #define CHANNEL_FLAGS_BUFSIZE_CHANGED 4 | ||
177 | #define CHANNEL_FLAGS_FAILED 8 | ||
178 | #define CHANNEL_FLAGS_WAITIRQ 16 | ||
179 | #define CHANNEL_FLAGS_RWMASK 1 | ||
180 | #define CHANNEL_DIRECTION(f) (f & CHANNEL_FLAGS_RWMASK) | ||
181 | |||
182 | #define LOG_FLAG_ILLEGALPKT 1 | ||
183 | #define LOG_FLAG_ILLEGALSIZE 2 | ||
184 | #define LOG_FLAG_OVERRUN 4 | ||
185 | #define LOG_FLAG_NOMEM 8 | ||
186 | |||
187 | #define CTC_LOGLEVEL_INFO 1 | ||
188 | #define CTC_LOGLEVEL_NOTICE 2 | ||
189 | #define CTC_LOGLEVEL_WARN 4 | ||
190 | #define CTC_LOGLEVEL_EMERG 8 | ||
191 | #define CTC_LOGLEVEL_ERR 16 | ||
192 | #define CTC_LOGLEVEL_DEBUG 32 | ||
193 | #define CTC_LOGLEVEL_CRIT 64 | ||
194 | |||
195 | #define CTC_LOGLEVEL_DEFAULT \ | ||
196 | (CTC_LOGLEVEL_INFO | CTC_LOGLEVEL_NOTICE | CTC_LOGLEVEL_WARN | CTC_LOGLEVEL_CRIT) | ||
197 | |||
198 | #define CTC_LOGLEVEL_MAX ((CTC_LOGLEVEL_CRIT<<1)-1) | ||
199 | |||
200 | #define ctc_pr_debug(fmt, arg...) \ | ||
201 | do { if (loglevel & CTC_LOGLEVEL_DEBUG) printk(KERN_DEBUG fmt,##arg); } while (0) | ||
202 | |||
203 | #define ctc_pr_info(fmt, arg...) \ | ||
204 | do { if (loglevel & CTC_LOGLEVEL_INFO) printk(KERN_INFO fmt,##arg); } while (0) | ||
205 | |||
206 | #define ctc_pr_notice(fmt, arg...) \ | ||
207 | do { if (loglevel & CTC_LOGLEVEL_NOTICE) printk(KERN_NOTICE fmt,##arg); } while (0) | ||
208 | |||
209 | #define ctc_pr_warn(fmt, arg...) \ | ||
210 | do { if (loglevel & CTC_LOGLEVEL_WARN) printk(KERN_WARNING fmt,##arg); } while (0) | ||
211 | |||
212 | #define ctc_pr_emerg(fmt, arg...) \ | ||
213 | do { if (loglevel & CTC_LOGLEVEL_EMERG) printk(KERN_EMERG fmt,##arg); } while (0) | ||
214 | |||
215 | #define ctc_pr_err(fmt, arg...) \ | ||
216 | do { if (loglevel & CTC_LOGLEVEL_ERR) printk(KERN_ERR fmt,##arg); } while (0) | ||
217 | |||
218 | #define ctc_pr_crit(fmt, arg...) \ | ||
219 | do { if (loglevel & CTC_LOGLEVEL_CRIT) printk(KERN_CRIT fmt,##arg); } while (0) | ||
220 | |||
221 | struct ctc_priv { | ||
222 | struct net_device_stats stats; | ||
223 | unsigned long tbusy; | ||
224 | /** | ||
225 | * The finite state machine of this interface. | ||
226 | */ | ||
227 | fsm_instance *fsm; | ||
228 | /** | ||
229 | * The protocol of this device | ||
230 | */ | ||
231 | __u16 protocol; | ||
232 | /** | ||
233 | * Timer for restarting after I/O Errors | ||
234 | */ | ||
235 | fsm_timer restart_timer; | ||
236 | |||
237 | int buffer_size; | ||
238 | |||
239 | struct channel *channel[2]; | ||
240 | }; | ||
241 | |||
242 | /** | ||
243 | * Definition of our link level header. | ||
244 | */ | ||
245 | struct ll_header { | ||
246 | __u16 length; | ||
247 | __u16 type; | ||
248 | __u16 unused; | ||
249 | }; | ||
250 | #define LL_HEADER_LENGTH (sizeof(struct ll_header)) | ||
251 | |||
252 | /** | ||
253 | * Compatibility macros for busy handling | ||
254 | * of network devices. | ||
255 | */ | ||
256 | static __inline__ void | ||
257 | ctc_clear_busy(struct net_device * dev) | ||
258 | { | ||
259 | clear_bit(0, &(((struct ctc_priv *) dev->priv)->tbusy)); | ||
260 | netif_wake_queue(dev); | ||
261 | } | ||
262 | |||
263 | static __inline__ int | ||
264 | ctc_test_and_set_busy(struct net_device * dev) | ||
265 | { | ||
266 | netif_stop_queue(dev); | ||
267 | return test_and_set_bit(0, &((struct ctc_priv *) dev->priv)->tbusy); | ||
268 | } | ||
269 | |||
270 | #endif | ||
diff --git a/drivers/s390/net/qeth.h b/drivers/s390/net/qeth.h deleted file mode 100644 index 8c6b72d05b1d..000000000000 --- a/drivers/s390/net/qeth.h +++ /dev/null | |||
@@ -1,1253 +0,0 @@ | |||
1 | #ifndef __QETH_H__ | ||
2 | #define __QETH_H__ | ||
3 | |||
4 | #include <linux/if.h> | ||
5 | #include <linux/if_arp.h> | ||
6 | |||
7 | #include <linux/if_tr.h> | ||
8 | #include <linux/trdevice.h> | ||
9 | #include <linux/etherdevice.h> | ||
10 | #include <linux/if_vlan.h> | ||
11 | #include <linux/ctype.h> | ||
12 | |||
13 | #include <net/ipv6.h> | ||
14 | #include <linux/in6.h> | ||
15 | #include <net/if_inet6.h> | ||
16 | #include <net/addrconf.h> | ||
17 | |||
18 | |||
19 | #include <linux/bitops.h> | ||
20 | |||
21 | #include <asm/debug.h> | ||
22 | #include <asm/qdio.h> | ||
23 | #include <asm/ccwdev.h> | ||
24 | #include <asm/ccwgroup.h> | ||
25 | |||
26 | #include "qeth_mpc.h" | ||
27 | |||
28 | #ifdef CONFIG_QETH_IPV6 | ||
29 | #define QETH_VERSION_IPV6 ":IPv6" | ||
30 | #else | ||
31 | #define QETH_VERSION_IPV6 "" | ||
32 | #endif | ||
33 | #ifdef CONFIG_QETH_VLAN | ||
34 | #define QETH_VERSION_VLAN ":VLAN" | ||
35 | #else | ||
36 | #define QETH_VERSION_VLAN "" | ||
37 | #endif | ||
38 | |||
39 | /** | ||
40 | * Debug Facility stuff | ||
41 | */ | ||
42 | #define QETH_DBF_SETUP_NAME "qeth_setup" | ||
43 | #define QETH_DBF_SETUP_LEN 8 | ||
44 | #define QETH_DBF_SETUP_PAGES 8 | ||
45 | #define QETH_DBF_SETUP_NR_AREAS 1 | ||
46 | #define QETH_DBF_SETUP_LEVEL 5 | ||
47 | |||
48 | #define QETH_DBF_MISC_NAME "qeth_misc" | ||
49 | #define QETH_DBF_MISC_LEN 128 | ||
50 | #define QETH_DBF_MISC_PAGES 2 | ||
51 | #define QETH_DBF_MISC_NR_AREAS 1 | ||
52 | #define QETH_DBF_MISC_LEVEL 2 | ||
53 | |||
54 | #define QETH_DBF_DATA_NAME "qeth_data" | ||
55 | #define QETH_DBF_DATA_LEN 96 | ||
56 | #define QETH_DBF_DATA_PAGES 8 | ||
57 | #define QETH_DBF_DATA_NR_AREAS 1 | ||
58 | #define QETH_DBF_DATA_LEVEL 2 | ||
59 | |||
60 | #define QETH_DBF_CONTROL_NAME "qeth_control" | ||
61 | #define QETH_DBF_CONTROL_LEN 256 | ||
62 | #define QETH_DBF_CONTROL_PAGES 8 | ||
63 | #define QETH_DBF_CONTROL_NR_AREAS 2 | ||
64 | #define QETH_DBF_CONTROL_LEVEL 5 | ||
65 | |||
66 | #define QETH_DBF_TRACE_NAME "qeth_trace" | ||
67 | #define QETH_DBF_TRACE_LEN 8 | ||
68 | #define QETH_DBF_TRACE_PAGES 4 | ||
69 | #define QETH_DBF_TRACE_NR_AREAS 2 | ||
70 | #define QETH_DBF_TRACE_LEVEL 3 | ||
71 | extern debug_info_t *qeth_dbf_trace; | ||
72 | |||
73 | #define QETH_DBF_SENSE_NAME "qeth_sense" | ||
74 | #define QETH_DBF_SENSE_LEN 64 | ||
75 | #define QETH_DBF_SENSE_PAGES 2 | ||
76 | #define QETH_DBF_SENSE_NR_AREAS 1 | ||
77 | #define QETH_DBF_SENSE_LEVEL 2 | ||
78 | |||
79 | #define QETH_DBF_QERR_NAME "qeth_qerr" | ||
80 | #define QETH_DBF_QERR_LEN 8 | ||
81 | #define QETH_DBF_QERR_PAGES 2 | ||
82 | #define QETH_DBF_QERR_NR_AREAS 2 | ||
83 | #define QETH_DBF_QERR_LEVEL 2 | ||
84 | |||
85 | #define QETH_DBF_TEXT(name,level,text) \ | ||
86 | do { \ | ||
87 | debug_text_event(qeth_dbf_##name,level,text); \ | ||
88 | } while (0) | ||
89 | |||
90 | #define QETH_DBF_HEX(name,level,addr,len) \ | ||
91 | do { \ | ||
92 | debug_event(qeth_dbf_##name,level,(void*)(addr),len); \ | ||
93 | } while (0) | ||
94 | |||
95 | DECLARE_PER_CPU(char[256], qeth_dbf_txt_buf); | ||
96 | |||
97 | #define QETH_DBF_TEXT_(name,level,text...) \ | ||
98 | do { \ | ||
99 | char* dbf_txt_buf = get_cpu_var(qeth_dbf_txt_buf); \ | ||
100 | sprintf(dbf_txt_buf, text); \ | ||
101 | debug_text_event(qeth_dbf_##name,level,dbf_txt_buf); \ | ||
102 | put_cpu_var(qeth_dbf_txt_buf); \ | ||
103 | } while (0) | ||
104 | |||
105 | #define QETH_DBF_SPRINTF(name,level,text...) \ | ||
106 | do { \ | ||
107 | debug_sprintf_event(qeth_dbf_trace, level, ##text ); \ | ||
108 | debug_sprintf_event(qeth_dbf_trace, level, text ); \ | ||
109 | } while (0) | ||
110 | |||
111 | /** | ||
112 | * some more debug stuff | ||
113 | */ | ||
114 | #define PRINTK_HEADER "qeth: " | ||
115 | |||
116 | #define HEXDUMP16(importance,header,ptr) \ | ||
117 | PRINT_##importance(header "%02x %02x %02x %02x %02x %02x %02x %02x " \ | ||
118 | "%02x %02x %02x %02x %02x %02x %02x %02x\n", \ | ||
119 | *(((char*)ptr)),*(((char*)ptr)+1),*(((char*)ptr)+2), \ | ||
120 | *(((char*)ptr)+3),*(((char*)ptr)+4),*(((char*)ptr)+5), \ | ||
121 | *(((char*)ptr)+6),*(((char*)ptr)+7),*(((char*)ptr)+8), \ | ||
122 | *(((char*)ptr)+9),*(((char*)ptr)+10),*(((char*)ptr)+11), \ | ||
123 | *(((char*)ptr)+12),*(((char*)ptr)+13), \ | ||
124 | *(((char*)ptr)+14),*(((char*)ptr)+15)); \ | ||
125 | PRINT_##importance(header "%02x %02x %02x %02x %02x %02x %02x %02x " \ | ||
126 | "%02x %02x %02x %02x %02x %02x %02x %02x\n", \ | ||
127 | *(((char*)ptr)+16),*(((char*)ptr)+17), \ | ||
128 | *(((char*)ptr)+18),*(((char*)ptr)+19), \ | ||
129 | *(((char*)ptr)+20),*(((char*)ptr)+21), \ | ||
130 | *(((char*)ptr)+22),*(((char*)ptr)+23), \ | ||
131 | *(((char*)ptr)+24),*(((char*)ptr)+25), \ | ||
132 | *(((char*)ptr)+26),*(((char*)ptr)+27), \ | ||
133 | *(((char*)ptr)+28),*(((char*)ptr)+29), \ | ||
134 | *(((char*)ptr)+30),*(((char*)ptr)+31)); | ||
135 | |||
136 | static inline void | ||
137 | qeth_hex_dump(unsigned char *buf, size_t len) | ||
138 | { | ||
139 | size_t i; | ||
140 | |||
141 | for (i = 0; i < len; i++) { | ||
142 | if (i && !(i % 16)) | ||
143 | printk("\n"); | ||
144 | printk("%02x ", *(buf + i)); | ||
145 | } | ||
146 | printk("\n"); | ||
147 | } | ||
148 | |||
149 | #define SENSE_COMMAND_REJECT_BYTE 0 | ||
150 | #define SENSE_COMMAND_REJECT_FLAG 0x80 | ||
151 | #define SENSE_RESETTING_EVENT_BYTE 1 | ||
152 | #define SENSE_RESETTING_EVENT_FLAG 0x80 | ||
153 | |||
154 | /* | ||
155 | * Common IO related definitions | ||
156 | */ | ||
157 | extern struct device *qeth_root_dev; | ||
158 | extern struct ccw_driver qeth_ccw_driver; | ||
159 | extern struct ccwgroup_driver qeth_ccwgroup_driver; | ||
160 | |||
161 | #define CARD_RDEV(card) card->read.ccwdev | ||
162 | #define CARD_WDEV(card) card->write.ccwdev | ||
163 | #define CARD_DDEV(card) card->data.ccwdev | ||
164 | #define CARD_BUS_ID(card) card->gdev->dev.bus_id | ||
165 | #define CARD_RDEV_ID(card) card->read.ccwdev->dev.bus_id | ||
166 | #define CARD_WDEV_ID(card) card->write.ccwdev->dev.bus_id | ||
167 | #define CARD_DDEV_ID(card) card->data.ccwdev->dev.bus_id | ||
168 | #define CHANNEL_ID(channel) channel->ccwdev->dev.bus_id | ||
169 | |||
170 | #define CARD_FROM_CDEV(cdev) (struct qeth_card *) \ | ||
171 | ((struct ccwgroup_device *)cdev->dev.driver_data)\ | ||
172 | ->dev.driver_data; | ||
173 | |||
174 | /** | ||
175 | * card stuff | ||
176 | */ | ||
177 | struct qeth_perf_stats { | ||
178 | unsigned int bufs_rec; | ||
179 | unsigned int bufs_sent; | ||
180 | |||
181 | unsigned int skbs_sent_pack; | ||
182 | unsigned int bufs_sent_pack; | ||
183 | |||
184 | unsigned int sc_dp_p; | ||
185 | unsigned int sc_p_dp; | ||
186 | /* qdio_input_handler: number of times called, time spent in */ | ||
187 | __u64 inbound_start_time; | ||
188 | unsigned int inbound_cnt; | ||
189 | unsigned int inbound_time; | ||
190 | /* qeth_send_packet: number of times called, time spent in */ | ||
191 | __u64 outbound_start_time; | ||
192 | unsigned int outbound_cnt; | ||
193 | unsigned int outbound_time; | ||
194 | /* qdio_output_handler: number of times called, time spent in */ | ||
195 | __u64 outbound_handler_start_time; | ||
196 | unsigned int outbound_handler_cnt; | ||
197 | unsigned int outbound_handler_time; | ||
198 | /* number of calls to and time spent in do_QDIO for inbound queue */ | ||
199 | __u64 inbound_do_qdio_start_time; | ||
200 | unsigned int inbound_do_qdio_cnt; | ||
201 | unsigned int inbound_do_qdio_time; | ||
202 | /* number of calls to and time spent in do_QDIO for outbound queues */ | ||
203 | __u64 outbound_do_qdio_start_time; | ||
204 | unsigned int outbound_do_qdio_cnt; | ||
205 | unsigned int outbound_do_qdio_time; | ||
206 | /* eddp data */ | ||
207 | unsigned int large_send_bytes; | ||
208 | unsigned int large_send_cnt; | ||
209 | unsigned int sg_skbs_sent; | ||
210 | unsigned int sg_frags_sent; | ||
211 | /* initial values when measuring starts */ | ||
212 | unsigned long initial_rx_packets; | ||
213 | unsigned long initial_tx_packets; | ||
214 | /* inbound scatter gather data */ | ||
215 | unsigned int sg_skbs_rx; | ||
216 | unsigned int sg_frags_rx; | ||
217 | unsigned int sg_alloc_page_rx; | ||
218 | }; | ||
219 | |||
220 | /* Routing stuff */ | ||
221 | struct qeth_routing_info { | ||
222 | enum qeth_routing_types type; | ||
223 | }; | ||
224 | |||
225 | /* IPA stuff */ | ||
226 | struct qeth_ipa_info { | ||
227 | __u32 supported_funcs; | ||
228 | __u32 enabled_funcs; | ||
229 | }; | ||
230 | |||
231 | static inline int | ||
232 | qeth_is_ipa_supported(struct qeth_ipa_info *ipa, enum qeth_ipa_funcs func) | ||
233 | { | ||
234 | return (ipa->supported_funcs & func); | ||
235 | } | ||
236 | |||
237 | static inline int | ||
238 | qeth_is_ipa_enabled(struct qeth_ipa_info *ipa, enum qeth_ipa_funcs func) | ||
239 | { | ||
240 | return (ipa->supported_funcs & ipa->enabled_funcs & func); | ||
241 | } | ||
242 | |||
243 | #define qeth_adp_supported(c,f) \ | ||
244 | qeth_is_ipa_supported(&c->options.adp, f) | ||
245 | #define qeth_adp_enabled(c,f) \ | ||
246 | qeth_is_ipa_enabled(&c->options.adp, f) | ||
247 | #define qeth_is_supported(c,f) \ | ||
248 | qeth_is_ipa_supported(&c->options.ipa4, f) | ||
249 | #define qeth_is_enabled(c,f) \ | ||
250 | qeth_is_ipa_enabled(&c->options.ipa4, f) | ||
251 | #ifdef CONFIG_QETH_IPV6 | ||
252 | #define qeth_is_supported6(c,f) \ | ||
253 | qeth_is_ipa_supported(&c->options.ipa6, f) | ||
254 | #define qeth_is_enabled6(c,f) \ | ||
255 | qeth_is_ipa_enabled(&c->options.ipa6, f) | ||
256 | #else /* CONFIG_QETH_IPV6 */ | ||
257 | #define qeth_is_supported6(c,f) 0 | ||
258 | #define qeth_is_enabled6(c,f) 0 | ||
259 | #endif /* CONFIG_QETH_IPV6 */ | ||
260 | #define qeth_is_ipafunc_supported(c,prot,f) \ | ||
261 | (prot==QETH_PROT_IPV6)? qeth_is_supported6(c,f):qeth_is_supported(c,f) | ||
262 | #define qeth_is_ipafunc_enabled(c,prot,f) \ | ||
263 | (prot==QETH_PROT_IPV6)? qeth_is_enabled6(c,f):qeth_is_enabled(c,f) | ||
264 | |||
265 | |||
266 | #define QETH_IDX_FUNC_LEVEL_OSAE_ENA_IPAT 0x0101 | ||
267 | #define QETH_IDX_FUNC_LEVEL_OSAE_DIS_IPAT 0x0101 | ||
268 | #define QETH_IDX_FUNC_LEVEL_IQD_ENA_IPAT 0x4108 | ||
269 | #define QETH_IDX_FUNC_LEVEL_IQD_DIS_IPAT 0x5108 | ||
270 | |||
271 | #define QETH_MODELLIST_ARRAY \ | ||
272 | {{0x1731,0x01,0x1732,0x01,QETH_CARD_TYPE_OSAE,1, \ | ||
273 | QETH_IDX_FUNC_LEVEL_OSAE_ENA_IPAT, \ | ||
274 | QETH_IDX_FUNC_LEVEL_OSAE_DIS_IPAT, \ | ||
275 | QETH_MAX_QUEUES,0}, \ | ||
276 | {0x1731,0x05,0x1732,0x05,QETH_CARD_TYPE_IQD,0, \ | ||
277 | QETH_IDX_FUNC_LEVEL_IQD_ENA_IPAT, \ | ||
278 | QETH_IDX_FUNC_LEVEL_IQD_DIS_IPAT, \ | ||
279 | QETH_MAX_QUEUES,0x103}, \ | ||
280 | {0x1731,0x06,0x1732,0x06,QETH_CARD_TYPE_OSN,0, \ | ||
281 | QETH_IDX_FUNC_LEVEL_OSAE_ENA_IPAT, \ | ||
282 | QETH_IDX_FUNC_LEVEL_OSAE_DIS_IPAT, \ | ||
283 | QETH_MAX_QUEUES,0}, \ | ||
284 | {0,0,0,0,0,0,0,0,0}} | ||
285 | |||
286 | #define QETH_REAL_CARD 1 | ||
287 | #define QETH_VLAN_CARD 2 | ||
288 | #define QETH_BUFSIZE 4096 | ||
289 | |||
290 | /** | ||
291 | * some more defs | ||
292 | */ | ||
293 | #define IF_NAME_LEN 16 | ||
294 | #define QETH_TX_TIMEOUT 100 * HZ | ||
295 | #define QETH_RCD_TIMEOUT 60 * HZ | ||
296 | #define QETH_HEADER_SIZE 32 | ||
297 | #define MAX_PORTNO 15 | ||
298 | #define QETH_FAKE_LL_LEN_ETH ETH_HLEN | ||
299 | #define QETH_FAKE_LL_LEN_TR (sizeof(struct trh_hdr)-TR_MAXRIFLEN+sizeof(struct trllc)) | ||
300 | #define QETH_FAKE_LL_V6_ADDR_POS 24 | ||
301 | |||
302 | /*IPv6 address autoconfiguration stuff*/ | ||
303 | #define UNIQUE_ID_IF_CREATE_ADDR_FAILED 0xfffe | ||
304 | #define UNIQUE_ID_NOT_BY_CARD 0x10000 | ||
305 | |||
306 | /*****************************************************************************/ | ||
307 | /* QDIO queue and buffer handling */ | ||
308 | /*****************************************************************************/ | ||
309 | #define QETH_MAX_QUEUES 4 | ||
310 | #define QETH_IN_BUF_SIZE_DEFAULT 65536 | ||
311 | #define QETH_IN_BUF_COUNT_DEFAULT 16 | ||
312 | #define QETH_IN_BUF_COUNT_MIN 8 | ||
313 | #define QETH_IN_BUF_COUNT_MAX 128 | ||
314 | #define QETH_MAX_BUFFER_ELEMENTS(card) ((card)->qdio.in_buf_size >> 12) | ||
315 | #define QETH_IN_BUF_REQUEUE_THRESHOLD(card) \ | ||
316 | ((card)->qdio.in_buf_pool.buf_count / 2) | ||
317 | |||
318 | /* buffers we have to be behind before we get a PCI */ | ||
319 | #define QETH_PCI_THRESHOLD_A(card) ((card)->qdio.in_buf_pool.buf_count+1) | ||
320 | /*enqueued free buffers left before we get a PCI*/ | ||
321 | #define QETH_PCI_THRESHOLD_B(card) 0 | ||
322 | /*not used unless the microcode gets patched*/ | ||
323 | #define QETH_PCI_TIMER_VALUE(card) 3 | ||
324 | |||
325 | #define QETH_MIN_INPUT_THRESHOLD 1 | ||
326 | #define QETH_MAX_INPUT_THRESHOLD 500 | ||
327 | #define QETH_MIN_OUTPUT_THRESHOLD 1 | ||
328 | #define QETH_MAX_OUTPUT_THRESHOLD 300 | ||
329 | |||
330 | /* priority queing */ | ||
331 | #define QETH_PRIOQ_DEFAULT QETH_NO_PRIO_QUEUEING | ||
332 | #define QETH_DEFAULT_QUEUE 2 | ||
333 | #define QETH_NO_PRIO_QUEUEING 0 | ||
334 | #define QETH_PRIO_Q_ING_PREC 1 | ||
335 | #define QETH_PRIO_Q_ING_TOS 2 | ||
336 | #define IP_TOS_LOWDELAY 0x10 | ||
337 | #define IP_TOS_HIGHTHROUGHPUT 0x08 | ||
338 | #define IP_TOS_HIGHRELIABILITY 0x04 | ||
339 | #define IP_TOS_NOTIMPORTANT 0x02 | ||
340 | |||
341 | /* Packing */ | ||
342 | #define QETH_LOW_WATERMARK_PACK 2 | ||
343 | #define QETH_HIGH_WATERMARK_PACK 5 | ||
344 | #define QETH_WATERMARK_PACK_FUZZ 1 | ||
345 | |||
346 | #define QETH_IP_HEADER_SIZE 40 | ||
347 | |||
348 | /* large receive scatter gather copy break */ | ||
349 | #define QETH_RX_SG_CB (PAGE_SIZE >> 1) | ||
350 | |||
351 | struct qeth_hdr_layer3 { | ||
352 | __u8 id; | ||
353 | __u8 flags; | ||
354 | __u16 inbound_checksum; /*TSO:__u16 seqno */ | ||
355 | __u32 token; /*TSO: __u32 reserved */ | ||
356 | __u16 length; | ||
357 | __u8 vlan_prio; | ||
358 | __u8 ext_flags; | ||
359 | __u16 vlan_id; | ||
360 | __u16 frame_offset; | ||
361 | __u8 dest_addr[16]; | ||
362 | } __attribute__ ((packed)); | ||
363 | |||
364 | struct qeth_hdr_layer2 { | ||
365 | __u8 id; | ||
366 | __u8 flags[3]; | ||
367 | __u8 port_no; | ||
368 | __u8 hdr_length; | ||
369 | __u16 pkt_length; | ||
370 | __u16 seq_no; | ||
371 | __u16 vlan_id; | ||
372 | __u32 reserved; | ||
373 | __u8 reserved2[16]; | ||
374 | } __attribute__ ((packed)); | ||
375 | |||
376 | struct qeth_hdr_osn { | ||
377 | __u8 id; | ||
378 | __u8 reserved; | ||
379 | __u16 seq_no; | ||
380 | __u16 reserved2; | ||
381 | __u16 control_flags; | ||
382 | __u16 pdu_length; | ||
383 | __u8 reserved3[18]; | ||
384 | __u32 ccid; | ||
385 | } __attribute__ ((packed)); | ||
386 | |||
387 | struct qeth_hdr { | ||
388 | union { | ||
389 | struct qeth_hdr_layer2 l2; | ||
390 | struct qeth_hdr_layer3 l3; | ||
391 | struct qeth_hdr_osn osn; | ||
392 | } hdr; | ||
393 | } __attribute__ ((packed)); | ||
394 | |||
395 | /*TCP Segmentation Offload header*/ | ||
396 | struct qeth_hdr_ext_tso { | ||
397 | __u16 hdr_tot_len; | ||
398 | __u8 imb_hdr_no; | ||
399 | __u8 reserved; | ||
400 | __u8 hdr_type; | ||
401 | __u8 hdr_version; | ||
402 | __u16 hdr_len; | ||
403 | __u32 payload_len; | ||
404 | __u16 mss; | ||
405 | __u16 dg_hdr_len; | ||
406 | __u8 padding[16]; | ||
407 | } __attribute__ ((packed)); | ||
408 | |||
409 | struct qeth_hdr_tso { | ||
410 | struct qeth_hdr hdr; /*hdr->hdr.l3.xxx*/ | ||
411 | struct qeth_hdr_ext_tso ext; | ||
412 | } __attribute__ ((packed)); | ||
413 | |||
414 | |||
415 | /* flags for qeth_hdr.flags */ | ||
416 | #define QETH_HDR_PASSTHRU 0x10 | ||
417 | #define QETH_HDR_IPV6 0x80 | ||
418 | #define QETH_HDR_CAST_MASK 0x07 | ||
419 | enum qeth_cast_flags { | ||
420 | QETH_CAST_UNICAST = 0x06, | ||
421 | QETH_CAST_MULTICAST = 0x04, | ||
422 | QETH_CAST_BROADCAST = 0x05, | ||
423 | QETH_CAST_ANYCAST = 0x07, | ||
424 | QETH_CAST_NOCAST = 0x00, | ||
425 | }; | ||
426 | |||
427 | enum qeth_layer2_frame_flags { | ||
428 | QETH_LAYER2_FLAG_MULTICAST = 0x01, | ||
429 | QETH_LAYER2_FLAG_BROADCAST = 0x02, | ||
430 | QETH_LAYER2_FLAG_UNICAST = 0x04, | ||
431 | QETH_LAYER2_FLAG_VLAN = 0x10, | ||
432 | }; | ||
433 | |||
434 | enum qeth_header_ids { | ||
435 | QETH_HEADER_TYPE_LAYER3 = 0x01, | ||
436 | QETH_HEADER_TYPE_LAYER2 = 0x02, | ||
437 | QETH_HEADER_TYPE_TSO = 0x03, | ||
438 | QETH_HEADER_TYPE_OSN = 0x04, | ||
439 | }; | ||
440 | /* flags for qeth_hdr.ext_flags */ | ||
441 | #define QETH_HDR_EXT_VLAN_FRAME 0x01 | ||
442 | #define QETH_HDR_EXT_TOKEN_ID 0x02 | ||
443 | #define QETH_HDR_EXT_INCLUDE_VLAN_TAG 0x04 | ||
444 | #define QETH_HDR_EXT_SRC_MAC_ADDR 0x08 | ||
445 | #define QETH_HDR_EXT_CSUM_HDR_REQ 0x10 | ||
446 | #define QETH_HDR_EXT_CSUM_TRANSP_REQ 0x20 | ||
447 | #define QETH_HDR_EXT_UDP_TSO 0x40 /*bit off for TCP*/ | ||
448 | |||
449 | static inline int | ||
450 | qeth_is_last_sbale(struct qdio_buffer_element *sbale) | ||
451 | { | ||
452 | return (sbale->flags & SBAL_FLAGS_LAST_ENTRY); | ||
453 | } | ||
454 | |||
455 | enum qeth_qdio_buffer_states { | ||
456 | /* | ||
457 | * inbound: read out by driver; owned by hardware in order to be filled | ||
458 | * outbound: owned by driver in order to be filled | ||
459 | */ | ||
460 | QETH_QDIO_BUF_EMPTY, | ||
461 | /* | ||
462 | * inbound: filled by hardware; owned by driver in order to be read out | ||
463 | * outbound: filled by driver; owned by hardware in order to be sent | ||
464 | */ | ||
465 | QETH_QDIO_BUF_PRIMED, | ||
466 | }; | ||
467 | |||
468 | enum qeth_qdio_info_states { | ||
469 | QETH_QDIO_UNINITIALIZED, | ||
470 | QETH_QDIO_ALLOCATED, | ||
471 | QETH_QDIO_ESTABLISHED, | ||
472 | QETH_QDIO_CLEANING | ||
473 | }; | ||
474 | |||
475 | struct qeth_buffer_pool_entry { | ||
476 | struct list_head list; | ||
477 | struct list_head init_list; | ||
478 | void *elements[QDIO_MAX_ELEMENTS_PER_BUFFER]; | ||
479 | }; | ||
480 | |||
481 | struct qeth_qdio_buffer_pool { | ||
482 | struct list_head entry_list; | ||
483 | int buf_count; | ||
484 | }; | ||
485 | |||
486 | struct qeth_qdio_buffer { | ||
487 | struct qdio_buffer *buffer; | ||
488 | volatile enum qeth_qdio_buffer_states state; | ||
489 | /* the buffer pool entry currently associated to this buffer */ | ||
490 | struct qeth_buffer_pool_entry *pool_entry; | ||
491 | }; | ||
492 | |||
493 | struct qeth_qdio_q { | ||
494 | struct qdio_buffer qdio_bufs[QDIO_MAX_BUFFERS_PER_Q]; | ||
495 | struct qeth_qdio_buffer bufs[QDIO_MAX_BUFFERS_PER_Q]; | ||
496 | /* | ||
497 | * buf_to_init means "buffer must be initialized by driver and must | ||
498 | * be made available for hardware" -> state is set to EMPTY | ||
499 | */ | ||
500 | volatile int next_buf_to_init; | ||
501 | } __attribute__ ((aligned(256))); | ||
502 | |||
503 | /* possible types of qeth large_send support */ | ||
504 | enum qeth_large_send_types { | ||
505 | QETH_LARGE_SEND_NO, | ||
506 | QETH_LARGE_SEND_EDDP, | ||
507 | QETH_LARGE_SEND_TSO, | ||
508 | }; | ||
509 | |||
510 | struct qeth_qdio_out_buffer { | ||
511 | struct qdio_buffer *buffer; | ||
512 | atomic_t state; | ||
513 | volatile int next_element_to_fill; | ||
514 | struct sk_buff_head skb_list; | ||
515 | struct list_head ctx_list; | ||
516 | }; | ||
517 | |||
518 | struct qeth_card; | ||
519 | |||
520 | enum qeth_out_q_states { | ||
521 | QETH_OUT_Q_UNLOCKED, | ||
522 | QETH_OUT_Q_LOCKED, | ||
523 | QETH_OUT_Q_LOCKED_FLUSH, | ||
524 | }; | ||
525 | |||
526 | struct qeth_qdio_out_q { | ||
527 | struct qdio_buffer qdio_bufs[QDIO_MAX_BUFFERS_PER_Q]; | ||
528 | struct qeth_qdio_out_buffer bufs[QDIO_MAX_BUFFERS_PER_Q]; | ||
529 | int queue_no; | ||
530 | struct qeth_card *card; | ||
531 | atomic_t state; | ||
532 | volatile int do_pack; | ||
533 | /* | ||
534 | * index of buffer to be filled by driver; state EMPTY or PACKING | ||
535 | */ | ||
536 | volatile int next_buf_to_fill; | ||
537 | /* | ||
538 | * number of buffers that are currently filled (PRIMED) | ||
539 | * -> these buffers are hardware-owned | ||
540 | */ | ||
541 | atomic_t used_buffers; | ||
542 | /* indicates whether PCI flag must be set (or if one is outstanding) */ | ||
543 | atomic_t set_pci_flags_count; | ||
544 | } __attribute__ ((aligned(256))); | ||
545 | |||
546 | struct qeth_qdio_info { | ||
547 | atomic_t state; | ||
548 | /* input */ | ||
549 | struct qeth_qdio_q *in_q; | ||
550 | struct qeth_qdio_buffer_pool in_buf_pool; | ||
551 | struct qeth_qdio_buffer_pool init_pool; | ||
552 | int in_buf_size; | ||
553 | |||
554 | /* output */ | ||
555 | int no_out_queues; | ||
556 | struct qeth_qdio_out_q **out_qs; | ||
557 | |||
558 | /* priority queueing */ | ||
559 | int do_prio_queueing; | ||
560 | int default_out_queue; | ||
561 | }; | ||
562 | |||
563 | enum qeth_send_errors { | ||
564 | QETH_SEND_ERROR_NONE, | ||
565 | QETH_SEND_ERROR_LINK_FAILURE, | ||
566 | QETH_SEND_ERROR_RETRY, | ||
567 | QETH_SEND_ERROR_KICK_IT, | ||
568 | }; | ||
569 | |||
570 | #define QETH_ETH_MAC_V4 0x0100 /* like v4 */ | ||
571 | #define QETH_ETH_MAC_V6 0x3333 /* like v6 */ | ||
572 | /* tr mc mac is longer, but that will be enough to detect mc frames */ | ||
573 | #define QETH_TR_MAC_NC 0xc000 /* non-canonical */ | ||
574 | #define QETH_TR_MAC_C 0x0300 /* canonical */ | ||
575 | |||
576 | #define DEFAULT_ADD_HHLEN 0 | ||
577 | #define MAX_ADD_HHLEN 1024 | ||
578 | |||
579 | /** | ||
580 | * buffer stuff for read channel | ||
581 | */ | ||
582 | #define QETH_CMD_BUFFER_NO 8 | ||
583 | |||
584 | /** | ||
585 | * channel state machine | ||
586 | */ | ||
587 | enum qeth_channel_states { | ||
588 | CH_STATE_UP, | ||
589 | CH_STATE_DOWN, | ||
590 | CH_STATE_ACTIVATING, | ||
591 | CH_STATE_HALTED, | ||
592 | CH_STATE_STOPPED, | ||
593 | CH_STATE_RCD, | ||
594 | CH_STATE_RCD_DONE, | ||
595 | }; | ||
596 | /** | ||
597 | * card state machine | ||
598 | */ | ||
599 | enum qeth_card_states { | ||
600 | CARD_STATE_DOWN, | ||
601 | CARD_STATE_HARDSETUP, | ||
602 | CARD_STATE_SOFTSETUP, | ||
603 | CARD_STATE_UP, | ||
604 | CARD_STATE_RECOVER, | ||
605 | }; | ||
606 | |||
607 | /** | ||
608 | * Protocol versions | ||
609 | */ | ||
610 | enum qeth_prot_versions { | ||
611 | QETH_PROT_IPV4 = 0x0004, | ||
612 | QETH_PROT_IPV6 = 0x0006, | ||
613 | }; | ||
614 | |||
615 | enum qeth_ip_types { | ||
616 | QETH_IP_TYPE_NORMAL, | ||
617 | QETH_IP_TYPE_VIPA, | ||
618 | QETH_IP_TYPE_RXIP, | ||
619 | QETH_IP_TYPE_DEL_ALL_MC, | ||
620 | }; | ||
621 | |||
622 | enum qeth_cmd_buffer_state { | ||
623 | BUF_STATE_FREE, | ||
624 | BUF_STATE_LOCKED, | ||
625 | BUF_STATE_PROCESSED, | ||
626 | }; | ||
627 | /** | ||
628 | * IP address and multicast list | ||
629 | */ | ||
630 | struct qeth_ipaddr { | ||
631 | struct list_head entry; | ||
632 | enum qeth_ip_types type; | ||
633 | enum qeth_ipa_setdelip_flags set_flags; | ||
634 | enum qeth_ipa_setdelip_flags del_flags; | ||
635 | int is_multicast; | ||
636 | volatile int users; | ||
637 | enum qeth_prot_versions proto; | ||
638 | unsigned char mac[OSA_ADDR_LEN]; | ||
639 | union { | ||
640 | struct { | ||
641 | unsigned int addr; | ||
642 | unsigned int mask; | ||
643 | } a4; | ||
644 | struct { | ||
645 | struct in6_addr addr; | ||
646 | unsigned int pfxlen; | ||
647 | } a6; | ||
648 | } u; | ||
649 | }; | ||
650 | |||
651 | struct qeth_ipato_entry { | ||
652 | struct list_head entry; | ||
653 | enum qeth_prot_versions proto; | ||
654 | char addr[16]; | ||
655 | int mask_bits; | ||
656 | }; | ||
657 | |||
658 | struct qeth_ipato { | ||
659 | int enabled; | ||
660 | int invert4; | ||
661 | int invert6; | ||
662 | struct list_head entries; | ||
663 | }; | ||
664 | |||
665 | struct qeth_channel; | ||
666 | |||
667 | struct qeth_cmd_buffer { | ||
668 | enum qeth_cmd_buffer_state state; | ||
669 | struct qeth_channel *channel; | ||
670 | unsigned char *data; | ||
671 | int rc; | ||
672 | void (*callback) (struct qeth_channel *, struct qeth_cmd_buffer *); | ||
673 | }; | ||
674 | |||
675 | |||
676 | /** | ||
677 | * definition of a qeth channel, used for read and write | ||
678 | */ | ||
679 | struct qeth_channel { | ||
680 | enum qeth_channel_states state; | ||
681 | struct ccw1 ccw; | ||
682 | spinlock_t iob_lock; | ||
683 | wait_queue_head_t wait_q; | ||
684 | struct tasklet_struct irq_tasklet; | ||
685 | struct ccw_device *ccwdev; | ||
686 | /*command buffer for control data*/ | ||
687 | struct qeth_cmd_buffer iob[QETH_CMD_BUFFER_NO]; | ||
688 | atomic_t irq_pending; | ||
689 | volatile int io_buf_no; | ||
690 | volatile int buf_no; | ||
691 | }; | ||
692 | |||
693 | /** | ||
694 | * OSA card related definitions | ||
695 | */ | ||
696 | struct qeth_token { | ||
697 | __u32 issuer_rm_w; | ||
698 | __u32 issuer_rm_r; | ||
699 | __u32 cm_filter_w; | ||
700 | __u32 cm_filter_r; | ||
701 | __u32 cm_connection_w; | ||
702 | __u32 cm_connection_r; | ||
703 | __u32 ulp_filter_w; | ||
704 | __u32 ulp_filter_r; | ||
705 | __u32 ulp_connection_w; | ||
706 | __u32 ulp_connection_r; | ||
707 | }; | ||
708 | |||
709 | struct qeth_seqno { | ||
710 | __u32 trans_hdr; | ||
711 | __u32 pdu_hdr; | ||
712 | __u32 pdu_hdr_ack; | ||
713 | __u16 ipa; | ||
714 | __u32 pkt_seqno; | ||
715 | }; | ||
716 | |||
717 | struct qeth_reply { | ||
718 | struct list_head list; | ||
719 | wait_queue_head_t wait_q; | ||
720 | int (*callback)(struct qeth_card *,struct qeth_reply *,unsigned long); | ||
721 | u32 seqno; | ||
722 | unsigned long offset; | ||
723 | atomic_t received; | ||
724 | int rc; | ||
725 | void *param; | ||
726 | struct qeth_card *card; | ||
727 | atomic_t refcnt; | ||
728 | }; | ||
729 | |||
730 | |||
731 | struct qeth_card_blkt { | ||
732 | int time_total; | ||
733 | int inter_packet; | ||
734 | int inter_packet_jumbo; | ||
735 | }; | ||
736 | |||
737 | #define QETH_BROADCAST_WITH_ECHO 0x01 | ||
738 | #define QETH_BROADCAST_WITHOUT_ECHO 0x02 | ||
739 | #define QETH_LAYER2_MAC_READ 0x01 | ||
740 | #define QETH_LAYER2_MAC_REGISTERED 0x02 | ||
741 | struct qeth_card_info { | ||
742 | unsigned short unit_addr2; | ||
743 | unsigned short cula; | ||
744 | unsigned short chpid; | ||
745 | __u16 func_level; | ||
746 | char mcl_level[QETH_MCL_LENGTH + 1]; | ||
747 | int guestlan; | ||
748 | int mac_bits; | ||
749 | int portname_required; | ||
750 | int portno; | ||
751 | char portname[9]; | ||
752 | enum qeth_card_types type; | ||
753 | enum qeth_link_types link_type; | ||
754 | int is_multicast_different; | ||
755 | int initial_mtu; | ||
756 | int max_mtu; | ||
757 | int broadcast_capable; | ||
758 | int unique_id; | ||
759 | struct qeth_card_blkt blkt; | ||
760 | __u32 csum_mask; | ||
761 | enum qeth_ipa_promisc_modes promisc_mode; | ||
762 | }; | ||
763 | |||
764 | struct qeth_card_options { | ||
765 | struct qeth_routing_info route4; | ||
766 | struct qeth_ipa_info ipa4; | ||
767 | struct qeth_ipa_info adp; /*Adapter parameters*/ | ||
768 | #ifdef CONFIG_QETH_IPV6 | ||
769 | struct qeth_routing_info route6; | ||
770 | struct qeth_ipa_info ipa6; | ||
771 | #endif /* QETH_IPV6 */ | ||
772 | enum qeth_checksum_types checksum_type; | ||
773 | int broadcast_mode; | ||
774 | int macaddr_mode; | ||
775 | int fake_broadcast; | ||
776 | int add_hhlen; | ||
777 | int fake_ll; | ||
778 | int layer2; | ||
779 | enum qeth_large_send_types large_send; | ||
780 | int performance_stats; | ||
781 | int rx_sg_cb; | ||
782 | }; | ||
783 | |||
784 | /* | ||
785 | * thread bits for qeth_card thread masks | ||
786 | */ | ||
787 | enum qeth_threads { | ||
788 | QETH_SET_IP_THREAD = 1, | ||
789 | QETH_RECOVER_THREAD = 2, | ||
790 | QETH_SET_PROMISC_MODE_THREAD = 4, | ||
791 | }; | ||
792 | |||
793 | struct qeth_osn_info { | ||
794 | int (*assist_cb)(struct net_device *dev, void *data); | ||
795 | int (*data_cb)(struct sk_buff *skb); | ||
796 | }; | ||
797 | |||
798 | struct qeth_card { | ||
799 | struct list_head list; | ||
800 | enum qeth_card_states state; | ||
801 | int lan_online; | ||
802 | spinlock_t lock; | ||
803 | /*hardware and sysfs stuff*/ | ||
804 | struct ccwgroup_device *gdev; | ||
805 | struct qeth_channel read; | ||
806 | struct qeth_channel write; | ||
807 | struct qeth_channel data; | ||
808 | |||
809 | struct net_device *dev; | ||
810 | struct net_device_stats stats; | ||
811 | |||
812 | struct qeth_card_info info; | ||
813 | struct qeth_token token; | ||
814 | struct qeth_seqno seqno; | ||
815 | struct qeth_card_options options; | ||
816 | |||
817 | wait_queue_head_t wait_q; | ||
818 | #ifdef CONFIG_QETH_VLAN | ||
819 | spinlock_t vlanlock; | ||
820 | struct vlan_group *vlangrp; | ||
821 | #endif | ||
822 | struct work_struct kernel_thread_starter; | ||
823 | spinlock_t thread_mask_lock; | ||
824 | volatile unsigned long thread_start_mask; | ||
825 | volatile unsigned long thread_allowed_mask; | ||
826 | volatile unsigned long thread_running_mask; | ||
827 | spinlock_t ip_lock; | ||
828 | struct list_head ip_list; | ||
829 | struct list_head *ip_tbd_list; | ||
830 | struct qeth_ipato ipato; | ||
831 | struct list_head cmd_waiter_list; | ||
832 | /* QDIO buffer handling */ | ||
833 | struct qeth_qdio_info qdio; | ||
834 | struct qeth_perf_stats perf_stats; | ||
835 | int use_hard_stop; | ||
836 | const struct header_ops *orig_header_ops; | ||
837 | struct qeth_osn_info osn_info; | ||
838 | atomic_t force_alloc_skb; | ||
839 | }; | ||
840 | |||
841 | struct qeth_card_list_struct { | ||
842 | struct list_head list; | ||
843 | rwlock_t rwlock; | ||
844 | }; | ||
845 | |||
846 | extern struct qeth_card_list_struct qeth_card_list; | ||
847 | |||
848 | /*notifier list */ | ||
849 | struct qeth_notify_list_struct { | ||
850 | struct list_head list; | ||
851 | struct task_struct *task; | ||
852 | int signum; | ||
853 | }; | ||
854 | extern spinlock_t qeth_notify_lock; | ||
855 | extern struct list_head qeth_notify_list; | ||
856 | |||
857 | /*some helper functions*/ | ||
858 | |||
859 | #define QETH_CARD_IFNAME(card) (((card)->dev)? (card)->dev->name : "") | ||
860 | |||
861 | static inline __u8 | ||
862 | qeth_get_ipa_adp_type(enum qeth_link_types link_type) | ||
863 | { | ||
864 | switch (link_type) { | ||
865 | case QETH_LINK_TYPE_HSTR: | ||
866 | return 2; | ||
867 | default: | ||
868 | return 1; | ||
869 | } | ||
870 | } | ||
871 | |||
872 | static inline struct sk_buff * | ||
873 | qeth_realloc_headroom(struct qeth_card *card, struct sk_buff *skb, int size) | ||
874 | { | ||
875 | struct sk_buff *new_skb = skb; | ||
876 | |||
877 | if (skb_headroom(skb) >= size) | ||
878 | return skb; | ||
879 | new_skb = skb_realloc_headroom(skb, size); | ||
880 | if (!new_skb) | ||
881 | PRINT_ERR("Could not realloc headroom for qeth_hdr " | ||
882 | "on interface %s", QETH_CARD_IFNAME(card)); | ||
883 | return new_skb; | ||
884 | } | ||
885 | |||
886 | static inline struct sk_buff * | ||
887 | qeth_pskb_unshare(struct sk_buff *skb, gfp_t pri) | ||
888 | { | ||
889 | struct sk_buff *nskb; | ||
890 | if (!skb_cloned(skb)) | ||
891 | return skb; | ||
892 | nskb = skb_copy(skb, pri); | ||
893 | return nskb; | ||
894 | } | ||
895 | |||
896 | static inline void * | ||
897 | qeth_push_skb(struct qeth_card *card, struct sk_buff *skb, int size) | ||
898 | { | ||
899 | void *hdr; | ||
900 | |||
901 | hdr = (void *) skb_push(skb, size); | ||
902 | /* | ||
903 | * sanity check, the Linux memory allocation scheme should | ||
904 | * never present us cases like this one (the qdio header size plus | ||
905 | * the first 40 bytes of the paket cross a 4k boundary) | ||
906 | */ | ||
907 | if ((((unsigned long) hdr) & (~(PAGE_SIZE - 1))) != | ||
908 | (((unsigned long) hdr + size + | ||
909 | QETH_IP_HEADER_SIZE) & (~(PAGE_SIZE - 1)))) { | ||
910 | PRINT_ERR("Misaligned packet on interface %s. Discarded.", | ||
911 | QETH_CARD_IFNAME(card)); | ||
912 | return NULL; | ||
913 | } | ||
914 | return hdr; | ||
915 | } | ||
916 | |||
917 | |||
918 | static inline int | ||
919 | qeth_get_hlen(__u8 link_type) | ||
920 | { | ||
921 | #ifdef CONFIG_QETH_IPV6 | ||
922 | switch (link_type) { | ||
923 | case QETH_LINK_TYPE_HSTR: | ||
924 | case QETH_LINK_TYPE_LANE_TR: | ||
925 | return sizeof(struct qeth_hdr_tso) + TR_HLEN; | ||
926 | default: | ||
927 | #ifdef CONFIG_QETH_VLAN | ||
928 | return sizeof(struct qeth_hdr_tso) + VLAN_ETH_HLEN; | ||
929 | #else | ||
930 | return sizeof(struct qeth_hdr_tso) + ETH_HLEN; | ||
931 | #endif | ||
932 | } | ||
933 | #else /* CONFIG_QETH_IPV6 */ | ||
934 | #ifdef CONFIG_QETH_VLAN | ||
935 | return sizeof(struct qeth_hdr_tso) + VLAN_HLEN; | ||
936 | #else | ||
937 | return sizeof(struct qeth_hdr_tso); | ||
938 | #endif | ||
939 | #endif /* CONFIG_QETH_IPV6 */ | ||
940 | } | ||
941 | |||
942 | static inline unsigned short | ||
943 | qeth_get_netdev_flags(struct qeth_card *card) | ||
944 | { | ||
945 | if (card->options.layer2 && | ||
946 | (card->info.type == QETH_CARD_TYPE_OSAE)) | ||
947 | return 0; | ||
948 | switch (card->info.type) { | ||
949 | case QETH_CARD_TYPE_IQD: | ||
950 | case QETH_CARD_TYPE_OSN: | ||
951 | return IFF_NOARP; | ||
952 | #ifdef CONFIG_QETH_IPV6 | ||
953 | default: | ||
954 | return 0; | ||
955 | #else | ||
956 | default: | ||
957 | return IFF_NOARP; | ||
958 | #endif | ||
959 | } | ||
960 | } | ||
961 | |||
962 | static inline int | ||
963 | qeth_get_initial_mtu_for_card(struct qeth_card * card) | ||
964 | { | ||
965 | switch (card->info.type) { | ||
966 | case QETH_CARD_TYPE_UNKNOWN: | ||
967 | return 1500; | ||
968 | case QETH_CARD_TYPE_IQD: | ||
969 | return card->info.max_mtu; | ||
970 | case QETH_CARD_TYPE_OSAE: | ||
971 | switch (card->info.link_type) { | ||
972 | case QETH_LINK_TYPE_HSTR: | ||
973 | case QETH_LINK_TYPE_LANE_TR: | ||
974 | return 2000; | ||
975 | default: | ||
976 | return 1492; | ||
977 | } | ||
978 | default: | ||
979 | return 1500; | ||
980 | } | ||
981 | } | ||
982 | |||
983 | static inline int | ||
984 | qeth_get_max_mtu_for_card(int cardtype) | ||
985 | { | ||
986 | switch (cardtype) { | ||
987 | |||
988 | case QETH_CARD_TYPE_UNKNOWN: | ||
989 | case QETH_CARD_TYPE_OSAE: | ||
990 | case QETH_CARD_TYPE_OSN: | ||
991 | return 61440; | ||
992 | case QETH_CARD_TYPE_IQD: | ||
993 | return 57344; | ||
994 | default: | ||
995 | return 1500; | ||
996 | } | ||
997 | } | ||
998 | |||
999 | static inline int | ||
1000 | qeth_get_mtu_out_of_mpc(int cardtype) | ||
1001 | { | ||
1002 | switch (cardtype) { | ||
1003 | case QETH_CARD_TYPE_IQD: | ||
1004 | return 1; | ||
1005 | default: | ||
1006 | return 0; | ||
1007 | } | ||
1008 | } | ||
1009 | |||
1010 | static inline int | ||
1011 | qeth_get_mtu_outof_framesize(int framesize) | ||
1012 | { | ||
1013 | switch (framesize) { | ||
1014 | case 0x4000: | ||
1015 | return 8192; | ||
1016 | case 0x6000: | ||
1017 | return 16384; | ||
1018 | case 0xa000: | ||
1019 | return 32768; | ||
1020 | case 0xffff: | ||
1021 | return 57344; | ||
1022 | default: | ||
1023 | return 0; | ||
1024 | } | ||
1025 | } | ||
1026 | |||
1027 | static inline int | ||
1028 | qeth_mtu_is_valid(struct qeth_card * card, int mtu) | ||
1029 | { | ||
1030 | switch (card->info.type) { | ||
1031 | case QETH_CARD_TYPE_OSAE: | ||
1032 | return ((mtu >= 576) && (mtu <= 61440)); | ||
1033 | case QETH_CARD_TYPE_IQD: | ||
1034 | return ((mtu >= 576) && | ||
1035 | (mtu <= card->info.max_mtu + 4096 - 32)); | ||
1036 | case QETH_CARD_TYPE_OSN: | ||
1037 | case QETH_CARD_TYPE_UNKNOWN: | ||
1038 | default: | ||
1039 | return 1; | ||
1040 | } | ||
1041 | } | ||
1042 | |||
1043 | static inline int | ||
1044 | qeth_get_arphdr_type(int cardtype, int linktype) | ||
1045 | { | ||
1046 | switch (cardtype) { | ||
1047 | case QETH_CARD_TYPE_OSAE: | ||
1048 | case QETH_CARD_TYPE_OSN: | ||
1049 | switch (linktype) { | ||
1050 | case QETH_LINK_TYPE_LANE_TR: | ||
1051 | case QETH_LINK_TYPE_HSTR: | ||
1052 | return ARPHRD_IEEE802_TR; | ||
1053 | default: | ||
1054 | return ARPHRD_ETHER; | ||
1055 | } | ||
1056 | case QETH_CARD_TYPE_IQD: | ||
1057 | default: | ||
1058 | return ARPHRD_ETHER; | ||
1059 | } | ||
1060 | } | ||
1061 | |||
1062 | static inline int | ||
1063 | qeth_get_micros(void) | ||
1064 | { | ||
1065 | return (int) (get_clock() >> 12); | ||
1066 | } | ||
1067 | |||
1068 | static inline int | ||
1069 | qeth_get_qdio_q_format(struct qeth_card *card) | ||
1070 | { | ||
1071 | switch (card->info.type) { | ||
1072 | case QETH_CARD_TYPE_IQD: | ||
1073 | return 2; | ||
1074 | default: | ||
1075 | return 0; | ||
1076 | } | ||
1077 | } | ||
1078 | |||
1079 | static inline int | ||
1080 | qeth_isxdigit(char * buf) | ||
1081 | { | ||
1082 | while (*buf) { | ||
1083 | if (!isxdigit(*buf++)) | ||
1084 | return 0; | ||
1085 | } | ||
1086 | return 1; | ||
1087 | } | ||
1088 | |||
1089 | static inline void | ||
1090 | qeth_ipaddr4_to_string(const __u8 *addr, char *buf) | ||
1091 | { | ||
1092 | sprintf(buf, "%i.%i.%i.%i", addr[0], addr[1], addr[2], addr[3]); | ||
1093 | } | ||
1094 | |||
1095 | static inline int | ||
1096 | qeth_string_to_ipaddr4(const char *buf, __u8 *addr) | ||
1097 | { | ||
1098 | int count = 0, rc = 0; | ||
1099 | int in[4]; | ||
1100 | char c; | ||
1101 | |||
1102 | rc = sscanf(buf, "%u.%u.%u.%u%c", | ||
1103 | &in[0], &in[1], &in[2], &in[3], &c); | ||
1104 | if (rc != 4 && (rc != 5 || c != '\n')) | ||
1105 | return -EINVAL; | ||
1106 | for (count = 0; count < 4; count++) { | ||
1107 | if (in[count] > 255) | ||
1108 | return -EINVAL; | ||
1109 | addr[count] = in[count]; | ||
1110 | } | ||
1111 | return 0; | ||
1112 | } | ||
1113 | |||
1114 | static inline void | ||
1115 | qeth_ipaddr6_to_string(const __u8 *addr, char *buf) | ||
1116 | { | ||
1117 | sprintf(buf, "%02x%02x:%02x%02x:%02x%02x:%02x%02x" | ||
1118 | ":%02x%02x:%02x%02x:%02x%02x:%02x%02x", | ||
1119 | addr[0], addr[1], addr[2], addr[3], | ||
1120 | addr[4], addr[5], addr[6], addr[7], | ||
1121 | addr[8], addr[9], addr[10], addr[11], | ||
1122 | addr[12], addr[13], addr[14], addr[15]); | ||
1123 | } | ||
1124 | |||
1125 | static inline int | ||
1126 | qeth_string_to_ipaddr6(const char *buf, __u8 *addr) | ||
1127 | { | ||
1128 | const char *end, *end_tmp, *start; | ||
1129 | __u16 *in; | ||
1130 | char num[5]; | ||
1131 | int num2, cnt, out, found, save_cnt; | ||
1132 | unsigned short in_tmp[8] = {0, }; | ||
1133 | |||
1134 | cnt = out = found = save_cnt = num2 = 0; | ||
1135 | end = start = buf; | ||
1136 | in = (__u16 *) addr; | ||
1137 | memset(in, 0, 16); | ||
1138 | while (*end) { | ||
1139 | end = strchr(start,':'); | ||
1140 | if (end == NULL) { | ||
1141 | end = buf + strlen(buf); | ||
1142 | if ((end_tmp = strchr(start, '\n')) != NULL) | ||
1143 | end = end_tmp; | ||
1144 | out = 1; | ||
1145 | } | ||
1146 | if ((end - start)) { | ||
1147 | memset(num, 0, 5); | ||
1148 | if ((end - start) > 4) | ||
1149 | return -EINVAL; | ||
1150 | memcpy(num, start, end - start); | ||
1151 | if (!qeth_isxdigit(num)) | ||
1152 | return -EINVAL; | ||
1153 | sscanf(start, "%x", &num2); | ||
1154 | if (found) | ||
1155 | in_tmp[save_cnt++] = num2; | ||
1156 | else | ||
1157 | in[cnt++] = num2; | ||
1158 | if (out) | ||
1159 | break; | ||
1160 | } else { | ||
1161 | if (found) | ||
1162 | return -EINVAL; | ||
1163 | found = 1; | ||
1164 | } | ||
1165 | start = ++end; | ||
1166 | } | ||
1167 | if (cnt + save_cnt > 8) | ||
1168 | return -EINVAL; | ||
1169 | cnt = 7; | ||
1170 | while (save_cnt) | ||
1171 | in[cnt--] = in_tmp[--save_cnt]; | ||
1172 | return 0; | ||
1173 | } | ||
1174 | |||
1175 | static inline void | ||
1176 | qeth_ipaddr_to_string(enum qeth_prot_versions proto, const __u8 *addr, | ||
1177 | char *buf) | ||
1178 | { | ||
1179 | if (proto == QETH_PROT_IPV4) | ||
1180 | qeth_ipaddr4_to_string(addr, buf); | ||
1181 | else if (proto == QETH_PROT_IPV6) | ||
1182 | qeth_ipaddr6_to_string(addr, buf); | ||
1183 | } | ||
1184 | |||
1185 | static inline int | ||
1186 | qeth_string_to_ipaddr(const char *buf, enum qeth_prot_versions proto, | ||
1187 | __u8 *addr) | ||
1188 | { | ||
1189 | if (proto == QETH_PROT_IPV4) | ||
1190 | return qeth_string_to_ipaddr4(buf, addr); | ||
1191 | else if (proto == QETH_PROT_IPV6) | ||
1192 | return qeth_string_to_ipaddr6(buf, addr); | ||
1193 | else | ||
1194 | return -EINVAL; | ||
1195 | } | ||
1196 | |||
1197 | extern int | ||
1198 | qeth_setrouting_v4(struct qeth_card *); | ||
1199 | extern int | ||
1200 | qeth_setrouting_v6(struct qeth_card *); | ||
1201 | |||
1202 | extern int | ||
1203 | qeth_add_ipato_entry(struct qeth_card *, struct qeth_ipato_entry *); | ||
1204 | |||
1205 | extern void | ||
1206 | qeth_del_ipato_entry(struct qeth_card *, enum qeth_prot_versions, u8 *, int); | ||
1207 | |||
1208 | extern int | ||
1209 | qeth_add_vipa(struct qeth_card *, enum qeth_prot_versions, const u8 *); | ||
1210 | |||
1211 | extern void | ||
1212 | qeth_del_vipa(struct qeth_card *, enum qeth_prot_versions, const u8 *); | ||
1213 | |||
1214 | extern int | ||
1215 | qeth_add_rxip(struct qeth_card *, enum qeth_prot_versions, const u8 *); | ||
1216 | |||
1217 | extern void | ||
1218 | qeth_del_rxip(struct qeth_card *, enum qeth_prot_versions, const u8 *); | ||
1219 | |||
1220 | extern int | ||
1221 | qeth_notifier_register(struct task_struct *, int ); | ||
1222 | |||
1223 | extern int | ||
1224 | qeth_notifier_unregister(struct task_struct * ); | ||
1225 | |||
1226 | extern void | ||
1227 | qeth_schedule_recovery(struct qeth_card *); | ||
1228 | |||
1229 | extern int | ||
1230 | qeth_realloc_buffer_pool(struct qeth_card *, int); | ||
1231 | |||
1232 | extern int | ||
1233 | qeth_set_large_send(struct qeth_card *, enum qeth_large_send_types); | ||
1234 | |||
1235 | extern void | ||
1236 | qeth_fill_header(struct qeth_card *, struct qeth_hdr *, | ||
1237 | struct sk_buff *, int, int); | ||
1238 | extern void | ||
1239 | qeth_flush_buffers(struct qeth_qdio_out_q *, int, int, int); | ||
1240 | |||
1241 | extern int | ||
1242 | qeth_osn_assist(struct net_device *, void *, int); | ||
1243 | |||
1244 | extern int | ||
1245 | qeth_osn_register(unsigned char *read_dev_no, | ||
1246 | struct net_device **, | ||
1247 | int (*assist_cb)(struct net_device *, void *), | ||
1248 | int (*data_cb)(struct sk_buff *)); | ||
1249 | |||
1250 | extern void | ||
1251 | qeth_osn_deregister(struct net_device *); | ||
1252 | |||
1253 | #endif /* __QETH_H__ */ | ||
diff --git a/drivers/s390/net/qeth_core.h b/drivers/s390/net/qeth_core.h new file mode 100644 index 000000000000..66f4f12503c9 --- /dev/null +++ b/drivers/s390/net/qeth_core.h | |||
@@ -0,0 +1,905 @@ | |||
1 | /* | ||
2 | * drivers/s390/net/qeth_core.h | ||
3 | * | ||
4 | * Copyright IBM Corp. 2007 | ||
5 | * Author(s): Utz Bacher <utz.bacher@de.ibm.com>, | ||
6 | * Frank Pavlic <fpavlic@de.ibm.com>, | ||
7 | * Thomas Spatzier <tspat@de.ibm.com>, | ||
8 | * Frank Blaschka <frank.blaschka@de.ibm.com> | ||
9 | */ | ||
10 | |||
11 | #ifndef __QETH_CORE_H__ | ||
12 | #define __QETH_CORE_H__ | ||
13 | |||
14 | #include <linux/if.h> | ||
15 | #include <linux/if_arp.h> | ||
16 | #include <linux/if_tr.h> | ||
17 | #include <linux/trdevice.h> | ||
18 | #include <linux/etherdevice.h> | ||
19 | #include <linux/if_vlan.h> | ||
20 | #include <linux/ctype.h> | ||
21 | #include <linux/in6.h> | ||
22 | #include <linux/bitops.h> | ||
23 | #include <linux/seq_file.h> | ||
24 | #include <linux/ethtool.h> | ||
25 | |||
26 | #include <net/ipv6.h> | ||
27 | #include <net/if_inet6.h> | ||
28 | #include <net/addrconf.h> | ||
29 | |||
30 | #include <asm/debug.h> | ||
31 | #include <asm/qdio.h> | ||
32 | #include <asm/ccwdev.h> | ||
33 | #include <asm/ccwgroup.h> | ||
34 | |||
35 | #include "qeth_core_mpc.h" | ||
36 | |||
37 | #define KMSG_COMPONENT "qeth" | ||
38 | |||
39 | /** | ||
40 | * Debug Facility stuff | ||
41 | */ | ||
42 | enum qeth_dbf_names { | ||
43 | QETH_DBF_SETUP, | ||
44 | QETH_DBF_QERR, | ||
45 | QETH_DBF_TRACE, | ||
46 | QETH_DBF_MSG, | ||
47 | QETH_DBF_SENSE, | ||
48 | QETH_DBF_MISC, | ||
49 | QETH_DBF_CTRL, | ||
50 | QETH_DBF_INFOS /* must be last element */ | ||
51 | }; | ||
52 | |||
53 | struct qeth_dbf_info { | ||
54 | char name[DEBUG_MAX_NAME_LEN]; | ||
55 | int pages; | ||
56 | int areas; | ||
57 | int len; | ||
58 | int level; | ||
59 | struct debug_view *view; | ||
60 | debug_info_t *id; | ||
61 | }; | ||
62 | |||
63 | #define QETH_DBF_CTRL_LEN 256 | ||
64 | |||
65 | #define QETH_DBF_TEXT(name, level, text) \ | ||
66 | debug_text_event(qeth_dbf[QETH_DBF_##name].id, level, text) | ||
67 | |||
68 | #define QETH_DBF_HEX(name, level, addr, len) \ | ||
69 | debug_event(qeth_dbf[QETH_DBF_##name].id, level, (void *)(addr), len) | ||
70 | |||
71 | #define QETH_DBF_MESSAGE(level, text...) \ | ||
72 | debug_sprintf_event(qeth_dbf[QETH_DBF_MSG].id, level, text) | ||
73 | |||
74 | #define QETH_DBF_TEXT_(name, level, text...) \ | ||
75 | do { \ | ||
76 | if (qeth_dbf_passes(qeth_dbf[QETH_DBF_##name].id, level)) { \ | ||
77 | char *dbf_txt_buf = \ | ||
78 | get_cpu_var(QETH_DBF_TXT_BUF); \ | ||
79 | sprintf(dbf_txt_buf, text); \ | ||
80 | debug_text_event(qeth_dbf[QETH_DBF_##name].id, \ | ||
81 | level, dbf_txt_buf); \ | ||
82 | put_cpu_var(QETH_DBF_TXT_BUF); \ | ||
83 | } \ | ||
84 | } while (0) | ||
85 | |||
86 | /* Allow to sort out low debug levels early to avoid wasted sprints */ | ||
87 | static inline int qeth_dbf_passes(debug_info_t *dbf_grp, int level) | ||
88 | { | ||
89 | return (level <= dbf_grp->level); | ||
90 | } | ||
91 | |||
92 | /** | ||
93 | * some more debug stuff | ||
94 | */ | ||
95 | #define PRINTK_HEADER "qeth: " | ||
96 | |||
97 | #define SENSE_COMMAND_REJECT_BYTE 0 | ||
98 | #define SENSE_COMMAND_REJECT_FLAG 0x80 | ||
99 | #define SENSE_RESETTING_EVENT_BYTE 1 | ||
100 | #define SENSE_RESETTING_EVENT_FLAG 0x80 | ||
101 | |||
102 | /* | ||
103 | * Common IO related definitions | ||
104 | */ | ||
105 | #define CARD_RDEV(card) card->read.ccwdev | ||
106 | #define CARD_WDEV(card) card->write.ccwdev | ||
107 | #define CARD_DDEV(card) card->data.ccwdev | ||
108 | #define CARD_BUS_ID(card) card->gdev->dev.bus_id | ||
109 | #define CARD_RDEV_ID(card) card->read.ccwdev->dev.bus_id | ||
110 | #define CARD_WDEV_ID(card) card->write.ccwdev->dev.bus_id | ||
111 | #define CARD_DDEV_ID(card) card->data.ccwdev->dev.bus_id | ||
112 | #define CHANNEL_ID(channel) channel->ccwdev->dev.bus_id | ||
113 | |||
114 | /** | ||
115 | * card stuff | ||
116 | */ | ||
117 | struct qeth_perf_stats { | ||
118 | unsigned int bufs_rec; | ||
119 | unsigned int bufs_sent; | ||
120 | |||
121 | unsigned int skbs_sent_pack; | ||
122 | unsigned int bufs_sent_pack; | ||
123 | |||
124 | unsigned int sc_dp_p; | ||
125 | unsigned int sc_p_dp; | ||
126 | /* qdio_input_handler: number of times called, time spent in */ | ||
127 | __u64 inbound_start_time; | ||
128 | unsigned int inbound_cnt; | ||
129 | unsigned int inbound_time; | ||
130 | /* qeth_send_packet: number of times called, time spent in */ | ||
131 | __u64 outbound_start_time; | ||
132 | unsigned int outbound_cnt; | ||
133 | unsigned int outbound_time; | ||
134 | /* qdio_output_handler: number of times called, time spent in */ | ||
135 | __u64 outbound_handler_start_time; | ||
136 | unsigned int outbound_handler_cnt; | ||
137 | unsigned int outbound_handler_time; | ||
138 | /* number of calls to and time spent in do_QDIO for inbound queue */ | ||
139 | __u64 inbound_do_qdio_start_time; | ||
140 | unsigned int inbound_do_qdio_cnt; | ||
141 | unsigned int inbound_do_qdio_time; | ||
142 | /* number of calls to and time spent in do_QDIO for outbound queues */ | ||
143 | __u64 outbound_do_qdio_start_time; | ||
144 | unsigned int outbound_do_qdio_cnt; | ||
145 | unsigned int outbound_do_qdio_time; | ||
146 | /* eddp data */ | ||
147 | unsigned int large_send_bytes; | ||
148 | unsigned int large_send_cnt; | ||
149 | unsigned int sg_skbs_sent; | ||
150 | unsigned int sg_frags_sent; | ||
151 | /* initial values when measuring starts */ | ||
152 | unsigned long initial_rx_packets; | ||
153 | unsigned long initial_tx_packets; | ||
154 | /* inbound scatter gather data */ | ||
155 | unsigned int sg_skbs_rx; | ||
156 | unsigned int sg_frags_rx; | ||
157 | unsigned int sg_alloc_page_rx; | ||
158 | }; | ||
159 | |||
160 | /* Routing stuff */ | ||
161 | struct qeth_routing_info { | ||
162 | enum qeth_routing_types type; | ||
163 | }; | ||
164 | |||
165 | /* IPA stuff */ | ||
166 | struct qeth_ipa_info { | ||
167 | __u32 supported_funcs; | ||
168 | __u32 enabled_funcs; | ||
169 | }; | ||
170 | |||
171 | static inline int qeth_is_ipa_supported(struct qeth_ipa_info *ipa, | ||
172 | enum qeth_ipa_funcs func) | ||
173 | { | ||
174 | return (ipa->supported_funcs & func); | ||
175 | } | ||
176 | |||
177 | static inline int qeth_is_ipa_enabled(struct qeth_ipa_info *ipa, | ||
178 | enum qeth_ipa_funcs func) | ||
179 | { | ||
180 | return (ipa->supported_funcs & ipa->enabled_funcs & func); | ||
181 | } | ||
182 | |||
183 | #define qeth_adp_supported(c, f) \ | ||
184 | qeth_is_ipa_supported(&c->options.adp, f) | ||
185 | #define qeth_adp_enabled(c, f) \ | ||
186 | qeth_is_ipa_enabled(&c->options.adp, f) | ||
187 | #define qeth_is_supported(c, f) \ | ||
188 | qeth_is_ipa_supported(&c->options.ipa4, f) | ||
189 | #define qeth_is_enabled(c, f) \ | ||
190 | qeth_is_ipa_enabled(&c->options.ipa4, f) | ||
191 | #define qeth_is_supported6(c, f) \ | ||
192 | qeth_is_ipa_supported(&c->options.ipa6, f) | ||
193 | #define qeth_is_enabled6(c, f) \ | ||
194 | qeth_is_ipa_enabled(&c->options.ipa6, f) | ||
195 | #define qeth_is_ipafunc_supported(c, prot, f) \ | ||
196 | ((prot == QETH_PROT_IPV6) ? \ | ||
197 | qeth_is_supported6(c, f) : qeth_is_supported(c, f)) | ||
198 | #define qeth_is_ipafunc_enabled(c, prot, f) \ | ||
199 | ((prot == QETH_PROT_IPV6) ? \ | ||
200 | qeth_is_enabled6(c, f) : qeth_is_enabled(c, f)) | ||
201 | |||
202 | #define QETH_IDX_FUNC_LEVEL_OSAE_ENA_IPAT 0x0101 | ||
203 | #define QETH_IDX_FUNC_LEVEL_OSAE_DIS_IPAT 0x0101 | ||
204 | #define QETH_IDX_FUNC_LEVEL_IQD_ENA_IPAT 0x4108 | ||
205 | #define QETH_IDX_FUNC_LEVEL_IQD_DIS_IPAT 0x5108 | ||
206 | |||
207 | #define QETH_MODELLIST_ARRAY \ | ||
208 | {{0x1731, 0x01, 0x1732, 0x01, QETH_CARD_TYPE_OSAE, 1, \ | ||
209 | QETH_IDX_FUNC_LEVEL_OSAE_ENA_IPAT, \ | ||
210 | QETH_IDX_FUNC_LEVEL_OSAE_DIS_IPAT, \ | ||
211 | QETH_MAX_QUEUES, 0}, \ | ||
212 | {0x1731, 0x05, 0x1732, 0x05, QETH_CARD_TYPE_IQD, 0, \ | ||
213 | QETH_IDX_FUNC_LEVEL_IQD_ENA_IPAT, \ | ||
214 | QETH_IDX_FUNC_LEVEL_IQD_DIS_IPAT, \ | ||
215 | QETH_MAX_QUEUES, 0x103}, \ | ||
216 | {0x1731, 0x06, 0x1732, 0x06, QETH_CARD_TYPE_OSN, 0, \ | ||
217 | QETH_IDX_FUNC_LEVEL_OSAE_ENA_IPAT, \ | ||
218 | QETH_IDX_FUNC_LEVEL_OSAE_DIS_IPAT, \ | ||
219 | QETH_MAX_QUEUES, 0}, \ | ||
220 | {0, 0, 0, 0, 0, 0, 0, 0, 0} } | ||
221 | |||
222 | #define QETH_REAL_CARD 1 | ||
223 | #define QETH_VLAN_CARD 2 | ||
224 | #define QETH_BUFSIZE 4096 | ||
225 | |||
226 | /** | ||
227 | * some more defs | ||
228 | */ | ||
229 | #define QETH_TX_TIMEOUT 100 * HZ | ||
230 | #define QETH_RCD_TIMEOUT 60 * HZ | ||
231 | #define QETH_HEADER_SIZE 32 | ||
232 | #define QETH_MAX_PORTNO 15 | ||
233 | |||
234 | /*IPv6 address autoconfiguration stuff*/ | ||
235 | #define UNIQUE_ID_IF_CREATE_ADDR_FAILED 0xfffe | ||
236 | #define UNIQUE_ID_NOT_BY_CARD 0x10000 | ||
237 | |||
238 | /*****************************************************************************/ | ||
239 | /* QDIO queue and buffer handling */ | ||
240 | /*****************************************************************************/ | ||
241 | #define QETH_MAX_QUEUES 4 | ||
242 | #define QETH_IN_BUF_SIZE_DEFAULT 65536 | ||
243 | #define QETH_IN_BUF_COUNT_DEFAULT 16 | ||
244 | #define QETH_IN_BUF_COUNT_MIN 8 | ||
245 | #define QETH_IN_BUF_COUNT_MAX 128 | ||
246 | #define QETH_MAX_BUFFER_ELEMENTS(card) ((card)->qdio.in_buf_size >> 12) | ||
247 | #define QETH_IN_BUF_REQUEUE_THRESHOLD(card) \ | ||
248 | ((card)->qdio.in_buf_pool.buf_count / 2) | ||
249 | |||
250 | /* buffers we have to be behind before we get a PCI */ | ||
251 | #define QETH_PCI_THRESHOLD_A(card) ((card)->qdio.in_buf_pool.buf_count+1) | ||
252 | /*enqueued free buffers left before we get a PCI*/ | ||
253 | #define QETH_PCI_THRESHOLD_B(card) 0 | ||
254 | /*not used unless the microcode gets patched*/ | ||
255 | #define QETH_PCI_TIMER_VALUE(card) 3 | ||
256 | |||
257 | #define QETH_MIN_INPUT_THRESHOLD 1 | ||
258 | #define QETH_MAX_INPUT_THRESHOLD 500 | ||
259 | #define QETH_MIN_OUTPUT_THRESHOLD 1 | ||
260 | #define QETH_MAX_OUTPUT_THRESHOLD 300 | ||
261 | |||
262 | /* priority queing */ | ||
263 | #define QETH_PRIOQ_DEFAULT QETH_NO_PRIO_QUEUEING | ||
264 | #define QETH_DEFAULT_QUEUE 2 | ||
265 | #define QETH_NO_PRIO_QUEUEING 0 | ||
266 | #define QETH_PRIO_Q_ING_PREC 1 | ||
267 | #define QETH_PRIO_Q_ING_TOS 2 | ||
268 | #define IP_TOS_LOWDELAY 0x10 | ||
269 | #define IP_TOS_HIGHTHROUGHPUT 0x08 | ||
270 | #define IP_TOS_HIGHRELIABILITY 0x04 | ||
271 | #define IP_TOS_NOTIMPORTANT 0x02 | ||
272 | |||
273 | /* Packing */ | ||
274 | #define QETH_LOW_WATERMARK_PACK 2 | ||
275 | #define QETH_HIGH_WATERMARK_PACK 5 | ||
276 | #define QETH_WATERMARK_PACK_FUZZ 1 | ||
277 | |||
278 | #define QETH_IP_HEADER_SIZE 40 | ||
279 | |||
280 | /* large receive scatter gather copy break */ | ||
281 | #define QETH_RX_SG_CB (PAGE_SIZE >> 1) | ||
282 | |||
283 | struct qeth_hdr_layer3 { | ||
284 | __u8 id; | ||
285 | __u8 flags; | ||
286 | __u16 inbound_checksum; /*TSO:__u16 seqno */ | ||
287 | __u32 token; /*TSO: __u32 reserved */ | ||
288 | __u16 length; | ||
289 | __u8 vlan_prio; | ||
290 | __u8 ext_flags; | ||
291 | __u16 vlan_id; | ||
292 | __u16 frame_offset; | ||
293 | __u8 dest_addr[16]; | ||
294 | } __attribute__ ((packed)); | ||
295 | |||
296 | struct qeth_hdr_layer2 { | ||
297 | __u8 id; | ||
298 | __u8 flags[3]; | ||
299 | __u8 port_no; | ||
300 | __u8 hdr_length; | ||
301 | __u16 pkt_length; | ||
302 | __u16 seq_no; | ||
303 | __u16 vlan_id; | ||
304 | __u32 reserved; | ||
305 | __u8 reserved2[16]; | ||
306 | } __attribute__ ((packed)); | ||
307 | |||
308 | struct qeth_hdr_osn { | ||
309 | __u8 id; | ||
310 | __u8 reserved; | ||
311 | __u16 seq_no; | ||
312 | __u16 reserved2; | ||
313 | __u16 control_flags; | ||
314 | __u16 pdu_length; | ||
315 | __u8 reserved3[18]; | ||
316 | __u32 ccid; | ||
317 | } __attribute__ ((packed)); | ||
318 | |||
319 | struct qeth_hdr { | ||
320 | union { | ||
321 | struct qeth_hdr_layer2 l2; | ||
322 | struct qeth_hdr_layer3 l3; | ||
323 | struct qeth_hdr_osn osn; | ||
324 | } hdr; | ||
325 | } __attribute__ ((packed)); | ||
326 | |||
327 | /*TCP Segmentation Offload header*/ | ||
328 | struct qeth_hdr_ext_tso { | ||
329 | __u16 hdr_tot_len; | ||
330 | __u8 imb_hdr_no; | ||
331 | __u8 reserved; | ||
332 | __u8 hdr_type; | ||
333 | __u8 hdr_version; | ||
334 | __u16 hdr_len; | ||
335 | __u32 payload_len; | ||
336 | __u16 mss; | ||
337 | __u16 dg_hdr_len; | ||
338 | __u8 padding[16]; | ||
339 | } __attribute__ ((packed)); | ||
340 | |||
341 | struct qeth_hdr_tso { | ||
342 | struct qeth_hdr hdr; /*hdr->hdr.l3.xxx*/ | ||
343 | struct qeth_hdr_ext_tso ext; | ||
344 | } __attribute__ ((packed)); | ||
345 | |||
346 | |||
347 | /* flags for qeth_hdr.flags */ | ||
348 | #define QETH_HDR_PASSTHRU 0x10 | ||
349 | #define QETH_HDR_IPV6 0x80 | ||
350 | #define QETH_HDR_CAST_MASK 0x07 | ||
351 | enum qeth_cast_flags { | ||
352 | QETH_CAST_UNICAST = 0x06, | ||
353 | QETH_CAST_MULTICAST = 0x04, | ||
354 | QETH_CAST_BROADCAST = 0x05, | ||
355 | QETH_CAST_ANYCAST = 0x07, | ||
356 | QETH_CAST_NOCAST = 0x00, | ||
357 | }; | ||
358 | |||
359 | enum qeth_layer2_frame_flags { | ||
360 | QETH_LAYER2_FLAG_MULTICAST = 0x01, | ||
361 | QETH_LAYER2_FLAG_BROADCAST = 0x02, | ||
362 | QETH_LAYER2_FLAG_UNICAST = 0x04, | ||
363 | QETH_LAYER2_FLAG_VLAN = 0x10, | ||
364 | }; | ||
365 | |||
366 | enum qeth_header_ids { | ||
367 | QETH_HEADER_TYPE_LAYER3 = 0x01, | ||
368 | QETH_HEADER_TYPE_LAYER2 = 0x02, | ||
369 | QETH_HEADER_TYPE_TSO = 0x03, | ||
370 | QETH_HEADER_TYPE_OSN = 0x04, | ||
371 | }; | ||
372 | /* flags for qeth_hdr.ext_flags */ | ||
373 | #define QETH_HDR_EXT_VLAN_FRAME 0x01 | ||
374 | #define QETH_HDR_EXT_TOKEN_ID 0x02 | ||
375 | #define QETH_HDR_EXT_INCLUDE_VLAN_TAG 0x04 | ||
376 | #define QETH_HDR_EXT_SRC_MAC_ADDR 0x08 | ||
377 | #define QETH_HDR_EXT_CSUM_HDR_REQ 0x10 | ||
378 | #define QETH_HDR_EXT_CSUM_TRANSP_REQ 0x20 | ||
379 | #define QETH_HDR_EXT_UDP_TSO 0x40 /*bit off for TCP*/ | ||
380 | |||
381 | static inline int qeth_is_last_sbale(struct qdio_buffer_element *sbale) | ||
382 | { | ||
383 | return (sbale->flags & SBAL_FLAGS_LAST_ENTRY); | ||
384 | } | ||
385 | |||
386 | enum qeth_qdio_buffer_states { | ||
387 | /* | ||
388 | * inbound: read out by driver; owned by hardware in order to be filled | ||
389 | * outbound: owned by driver in order to be filled | ||
390 | */ | ||
391 | QETH_QDIO_BUF_EMPTY, | ||
392 | /* | ||
393 | * inbound: filled by hardware; owned by driver in order to be read out | ||
394 | * outbound: filled by driver; owned by hardware in order to be sent | ||
395 | */ | ||
396 | QETH_QDIO_BUF_PRIMED, | ||
397 | }; | ||
398 | |||
399 | enum qeth_qdio_info_states { | ||
400 | QETH_QDIO_UNINITIALIZED, | ||
401 | QETH_QDIO_ALLOCATED, | ||
402 | QETH_QDIO_ESTABLISHED, | ||
403 | QETH_QDIO_CLEANING | ||
404 | }; | ||
405 | |||
406 | struct qeth_buffer_pool_entry { | ||
407 | struct list_head list; | ||
408 | struct list_head init_list; | ||
409 | void *elements[QDIO_MAX_ELEMENTS_PER_BUFFER]; | ||
410 | }; | ||
411 | |||
412 | struct qeth_qdio_buffer_pool { | ||
413 | struct list_head entry_list; | ||
414 | int buf_count; | ||
415 | }; | ||
416 | |||
417 | struct qeth_qdio_buffer { | ||
418 | struct qdio_buffer *buffer; | ||
419 | /* the buffer pool entry currently associated to this buffer */ | ||
420 | struct qeth_buffer_pool_entry *pool_entry; | ||
421 | }; | ||
422 | |||
423 | struct qeth_qdio_q { | ||
424 | struct qdio_buffer qdio_bufs[QDIO_MAX_BUFFERS_PER_Q]; | ||
425 | struct qeth_qdio_buffer bufs[QDIO_MAX_BUFFERS_PER_Q]; | ||
426 | int next_buf_to_init; | ||
427 | } __attribute__ ((aligned(256))); | ||
428 | |||
429 | /* possible types of qeth large_send support */ | ||
430 | enum qeth_large_send_types { | ||
431 | QETH_LARGE_SEND_NO, | ||
432 | QETH_LARGE_SEND_EDDP, | ||
433 | QETH_LARGE_SEND_TSO, | ||
434 | }; | ||
435 | |||
436 | struct qeth_qdio_out_buffer { | ||
437 | struct qdio_buffer *buffer; | ||
438 | atomic_t state; | ||
439 | int next_element_to_fill; | ||
440 | struct sk_buff_head skb_list; | ||
441 | struct list_head ctx_list; | ||
442 | }; | ||
443 | |||
444 | struct qeth_card; | ||
445 | |||
446 | enum qeth_out_q_states { | ||
447 | QETH_OUT_Q_UNLOCKED, | ||
448 | QETH_OUT_Q_LOCKED, | ||
449 | QETH_OUT_Q_LOCKED_FLUSH, | ||
450 | }; | ||
451 | |||
452 | struct qeth_qdio_out_q { | ||
453 | struct qdio_buffer qdio_bufs[QDIO_MAX_BUFFERS_PER_Q]; | ||
454 | struct qeth_qdio_out_buffer bufs[QDIO_MAX_BUFFERS_PER_Q]; | ||
455 | int queue_no; | ||
456 | struct qeth_card *card; | ||
457 | atomic_t state; | ||
458 | int do_pack; | ||
459 | /* | ||
460 | * index of buffer to be filled by driver; state EMPTY or PACKING | ||
461 | */ | ||
462 | int next_buf_to_fill; | ||
463 | /* | ||
464 | * number of buffers that are currently filled (PRIMED) | ||
465 | * -> these buffers are hardware-owned | ||
466 | */ | ||
467 | atomic_t used_buffers; | ||
468 | /* indicates whether PCI flag must be set (or if one is outstanding) */ | ||
469 | atomic_t set_pci_flags_count; | ||
470 | } __attribute__ ((aligned(256))); | ||
471 | |||
472 | struct qeth_qdio_info { | ||
473 | atomic_t state; | ||
474 | /* input */ | ||
475 | struct qeth_qdio_q *in_q; | ||
476 | struct qeth_qdio_buffer_pool in_buf_pool; | ||
477 | struct qeth_qdio_buffer_pool init_pool; | ||
478 | int in_buf_size; | ||
479 | |||
480 | /* output */ | ||
481 | int no_out_queues; | ||
482 | struct qeth_qdio_out_q **out_qs; | ||
483 | |||
484 | /* priority queueing */ | ||
485 | int do_prio_queueing; | ||
486 | int default_out_queue; | ||
487 | }; | ||
488 | |||
489 | enum qeth_send_errors { | ||
490 | QETH_SEND_ERROR_NONE, | ||
491 | QETH_SEND_ERROR_LINK_FAILURE, | ||
492 | QETH_SEND_ERROR_RETRY, | ||
493 | QETH_SEND_ERROR_KICK_IT, | ||
494 | }; | ||
495 | |||
496 | #define QETH_ETH_MAC_V4 0x0100 /* like v4 */ | ||
497 | #define QETH_ETH_MAC_V6 0x3333 /* like v6 */ | ||
498 | /* tr mc mac is longer, but that will be enough to detect mc frames */ | ||
499 | #define QETH_TR_MAC_NC 0xc000 /* non-canonical */ | ||
500 | #define QETH_TR_MAC_C 0x0300 /* canonical */ | ||
501 | |||
502 | #define DEFAULT_ADD_HHLEN 0 | ||
503 | #define MAX_ADD_HHLEN 1024 | ||
504 | |||
505 | /** | ||
506 | * buffer stuff for read channel | ||
507 | */ | ||
508 | #define QETH_CMD_BUFFER_NO 8 | ||
509 | |||
510 | /** | ||
511 | * channel state machine | ||
512 | */ | ||
513 | enum qeth_channel_states { | ||
514 | CH_STATE_UP, | ||
515 | CH_STATE_DOWN, | ||
516 | CH_STATE_ACTIVATING, | ||
517 | CH_STATE_HALTED, | ||
518 | CH_STATE_STOPPED, | ||
519 | CH_STATE_RCD, | ||
520 | CH_STATE_RCD_DONE, | ||
521 | }; | ||
522 | /** | ||
523 | * card state machine | ||
524 | */ | ||
525 | enum qeth_card_states { | ||
526 | CARD_STATE_DOWN, | ||
527 | CARD_STATE_HARDSETUP, | ||
528 | CARD_STATE_SOFTSETUP, | ||
529 | CARD_STATE_UP, | ||
530 | CARD_STATE_RECOVER, | ||
531 | }; | ||
532 | |||
533 | /** | ||
534 | * Protocol versions | ||
535 | */ | ||
536 | enum qeth_prot_versions { | ||
537 | QETH_PROT_IPV4 = 0x0004, | ||
538 | QETH_PROT_IPV6 = 0x0006, | ||
539 | }; | ||
540 | |||
541 | enum qeth_ip_types { | ||
542 | QETH_IP_TYPE_NORMAL, | ||
543 | QETH_IP_TYPE_VIPA, | ||
544 | QETH_IP_TYPE_RXIP, | ||
545 | QETH_IP_TYPE_DEL_ALL_MC, | ||
546 | }; | ||
547 | |||
548 | enum qeth_cmd_buffer_state { | ||
549 | BUF_STATE_FREE, | ||
550 | BUF_STATE_LOCKED, | ||
551 | BUF_STATE_PROCESSED, | ||
552 | }; | ||
553 | |||
554 | struct qeth_ipato { | ||
555 | int enabled; | ||
556 | int invert4; | ||
557 | int invert6; | ||
558 | struct list_head entries; | ||
559 | }; | ||
560 | |||
561 | struct qeth_channel; | ||
562 | |||
563 | struct qeth_cmd_buffer { | ||
564 | enum qeth_cmd_buffer_state state; | ||
565 | struct qeth_channel *channel; | ||
566 | unsigned char *data; | ||
567 | int rc; | ||
568 | void (*callback) (struct qeth_channel *, struct qeth_cmd_buffer *); | ||
569 | }; | ||
570 | |||
571 | /** | ||
572 | * definition of a qeth channel, used for read and write | ||
573 | */ | ||
574 | struct qeth_channel { | ||
575 | enum qeth_channel_states state; | ||
576 | struct ccw1 ccw; | ||
577 | spinlock_t iob_lock; | ||
578 | wait_queue_head_t wait_q; | ||
579 | struct tasklet_struct irq_tasklet; | ||
580 | struct ccw_device *ccwdev; | ||
581 | /*command buffer for control data*/ | ||
582 | struct qeth_cmd_buffer iob[QETH_CMD_BUFFER_NO]; | ||
583 | atomic_t irq_pending; | ||
584 | int io_buf_no; | ||
585 | int buf_no; | ||
586 | }; | ||
587 | |||
588 | /** | ||
589 | * OSA card related definitions | ||
590 | */ | ||
591 | struct qeth_token { | ||
592 | __u32 issuer_rm_w; | ||
593 | __u32 issuer_rm_r; | ||
594 | __u32 cm_filter_w; | ||
595 | __u32 cm_filter_r; | ||
596 | __u32 cm_connection_w; | ||
597 | __u32 cm_connection_r; | ||
598 | __u32 ulp_filter_w; | ||
599 | __u32 ulp_filter_r; | ||
600 | __u32 ulp_connection_w; | ||
601 | __u32 ulp_connection_r; | ||
602 | }; | ||
603 | |||
604 | struct qeth_seqno { | ||
605 | __u32 trans_hdr; | ||
606 | __u32 pdu_hdr; | ||
607 | __u32 pdu_hdr_ack; | ||
608 | __u16 ipa; | ||
609 | __u32 pkt_seqno; | ||
610 | }; | ||
611 | |||
612 | struct qeth_reply { | ||
613 | struct list_head list; | ||
614 | wait_queue_head_t wait_q; | ||
615 | int (*callback)(struct qeth_card *, struct qeth_reply *, | ||
616 | unsigned long); | ||
617 | u32 seqno; | ||
618 | unsigned long offset; | ||
619 | atomic_t received; | ||
620 | int rc; | ||
621 | void *param; | ||
622 | struct qeth_card *card; | ||
623 | atomic_t refcnt; | ||
624 | }; | ||
625 | |||
626 | |||
627 | struct qeth_card_blkt { | ||
628 | int time_total; | ||
629 | int inter_packet; | ||
630 | int inter_packet_jumbo; | ||
631 | }; | ||
632 | |||
633 | #define QETH_BROADCAST_WITH_ECHO 0x01 | ||
634 | #define QETH_BROADCAST_WITHOUT_ECHO 0x02 | ||
635 | #define QETH_LAYER2_MAC_READ 0x01 | ||
636 | #define QETH_LAYER2_MAC_REGISTERED 0x02 | ||
637 | struct qeth_card_info { | ||
638 | unsigned short unit_addr2; | ||
639 | unsigned short cula; | ||
640 | unsigned short chpid; | ||
641 | __u16 func_level; | ||
642 | char mcl_level[QETH_MCL_LENGTH + 1]; | ||
643 | int guestlan; | ||
644 | int mac_bits; | ||
645 | int portname_required; | ||
646 | int portno; | ||
647 | char portname[9]; | ||
648 | enum qeth_card_types type; | ||
649 | enum qeth_link_types link_type; | ||
650 | int is_multicast_different; | ||
651 | int initial_mtu; | ||
652 | int max_mtu; | ||
653 | int broadcast_capable; | ||
654 | int unique_id; | ||
655 | struct qeth_card_blkt blkt; | ||
656 | __u32 csum_mask; | ||
657 | enum qeth_ipa_promisc_modes promisc_mode; | ||
658 | }; | ||
659 | |||
660 | struct qeth_card_options { | ||
661 | struct qeth_routing_info route4; | ||
662 | struct qeth_ipa_info ipa4; | ||
663 | struct qeth_ipa_info adp; /*Adapter parameters*/ | ||
664 | struct qeth_routing_info route6; | ||
665 | struct qeth_ipa_info ipa6; | ||
666 | enum qeth_checksum_types checksum_type; | ||
667 | int broadcast_mode; | ||
668 | int macaddr_mode; | ||
669 | int fake_broadcast; | ||
670 | int add_hhlen; | ||
671 | int fake_ll; | ||
672 | int layer2; | ||
673 | enum qeth_large_send_types large_send; | ||
674 | int performance_stats; | ||
675 | int rx_sg_cb; | ||
676 | }; | ||
677 | |||
678 | /* | ||
679 | * thread bits for qeth_card thread masks | ||
680 | */ | ||
681 | enum qeth_threads { | ||
682 | QETH_RECOVER_THREAD = 1, | ||
683 | }; | ||
684 | |||
685 | struct qeth_osn_info { | ||
686 | int (*assist_cb)(struct net_device *dev, void *data); | ||
687 | int (*data_cb)(struct sk_buff *skb); | ||
688 | }; | ||
689 | |||
690 | enum qeth_discipline_id { | ||
691 | QETH_DISCIPLINE_LAYER3 = 0, | ||
692 | QETH_DISCIPLINE_LAYER2 = 1, | ||
693 | }; | ||
694 | |||
695 | struct qeth_discipline { | ||
696 | qdio_handler_t *input_handler; | ||
697 | qdio_handler_t *output_handler; | ||
698 | int (*recover)(void *ptr); | ||
699 | struct ccwgroup_driver *ccwgdriver; | ||
700 | }; | ||
701 | |||
702 | struct qeth_vlan_vid { | ||
703 | struct list_head list; | ||
704 | unsigned short vid; | ||
705 | }; | ||
706 | |||
707 | struct qeth_mc_mac { | ||
708 | struct list_head list; | ||
709 | __u8 mc_addr[MAX_ADDR_LEN]; | ||
710 | unsigned char mc_addrlen; | ||
711 | }; | ||
712 | |||
713 | struct qeth_card { | ||
714 | struct list_head list; | ||
715 | enum qeth_card_states state; | ||
716 | int lan_online; | ||
717 | spinlock_t lock; | ||
718 | struct ccwgroup_device *gdev; | ||
719 | struct qeth_channel read; | ||
720 | struct qeth_channel write; | ||
721 | struct qeth_channel data; | ||
722 | |||
723 | struct net_device *dev; | ||
724 | struct net_device_stats stats; | ||
725 | |||
726 | struct qeth_card_info info; | ||
727 | struct qeth_token token; | ||
728 | struct qeth_seqno seqno; | ||
729 | struct qeth_card_options options; | ||
730 | |||
731 | wait_queue_head_t wait_q; | ||
732 | spinlock_t vlanlock; | ||
733 | spinlock_t mclock; | ||
734 | struct vlan_group *vlangrp; | ||
735 | struct list_head vid_list; | ||
736 | struct list_head mc_list; | ||
737 | struct work_struct kernel_thread_starter; | ||
738 | spinlock_t thread_mask_lock; | ||
739 | unsigned long thread_start_mask; | ||
740 | unsigned long thread_allowed_mask; | ||
741 | unsigned long thread_running_mask; | ||
742 | spinlock_t ip_lock; | ||
743 | struct list_head ip_list; | ||
744 | struct list_head *ip_tbd_list; | ||
745 | struct qeth_ipato ipato; | ||
746 | struct list_head cmd_waiter_list; | ||
747 | /* QDIO buffer handling */ | ||
748 | struct qeth_qdio_info qdio; | ||
749 | struct qeth_perf_stats perf_stats; | ||
750 | int use_hard_stop; | ||
751 | struct qeth_osn_info osn_info; | ||
752 | struct qeth_discipline discipline; | ||
753 | atomic_t force_alloc_skb; | ||
754 | }; | ||
755 | |||
756 | struct qeth_card_list_struct { | ||
757 | struct list_head list; | ||
758 | rwlock_t rwlock; | ||
759 | }; | ||
760 | |||
761 | /*some helper functions*/ | ||
762 | #define QETH_CARD_IFNAME(card) (((card)->dev)? (card)->dev->name : "") | ||
763 | |||
764 | static inline struct qeth_card *CARD_FROM_CDEV(struct ccw_device *cdev) | ||
765 | { | ||
766 | struct qeth_card *card = dev_get_drvdata(&((struct ccwgroup_device *) | ||
767 | dev_get_drvdata(&cdev->dev))->dev); | ||
768 | return card; | ||
769 | } | ||
770 | |||
771 | static inline int qeth_get_micros(void) | ||
772 | { | ||
773 | return (int) (get_clock() >> 12); | ||
774 | } | ||
775 | |||
776 | static inline void *qeth_push_skb(struct qeth_card *card, struct sk_buff *skb, | ||
777 | int size) | ||
778 | { | ||
779 | void *hdr; | ||
780 | |||
781 | hdr = (void *) skb_push(skb, size); | ||
782 | /* | ||
783 | * sanity check, the Linux memory allocation scheme should | ||
784 | * never present us cases like this one (the qdio header size plus | ||
785 | * the first 40 bytes of the paket cross a 4k boundary) | ||
786 | */ | ||
787 | if ((((unsigned long) hdr) & (~(PAGE_SIZE - 1))) != | ||
788 | (((unsigned long) hdr + size + | ||
789 | QETH_IP_HEADER_SIZE) & (~(PAGE_SIZE - 1)))) { | ||
790 | PRINT_ERR("Misaligned packet on interface %s. Discarded.", | ||
791 | QETH_CARD_IFNAME(card)); | ||
792 | return NULL; | ||
793 | } | ||
794 | return hdr; | ||
795 | } | ||
796 | |||
797 | static inline int qeth_get_ip_version(struct sk_buff *skb) | ||
798 | { | ||
799 | switch (skb->protocol) { | ||
800 | case ETH_P_IPV6: | ||
801 | return 6; | ||
802 | case ETH_P_IP: | ||
803 | return 4; | ||
804 | default: | ||
805 | return 0; | ||
806 | } | ||
807 | } | ||
808 | |||
809 | struct qeth_eddp_context; | ||
810 | extern struct ccwgroup_driver qeth_l2_ccwgroup_driver; | ||
811 | extern struct ccwgroup_driver qeth_l3_ccwgroup_driver; | ||
812 | const char *qeth_get_cardname_short(struct qeth_card *); | ||
813 | int qeth_realloc_buffer_pool(struct qeth_card *, int); | ||
814 | int qeth_core_load_discipline(struct qeth_card *, enum qeth_discipline_id); | ||
815 | void qeth_core_free_discipline(struct qeth_card *); | ||
816 | int qeth_core_create_device_attributes(struct device *); | ||
817 | void qeth_core_remove_device_attributes(struct device *); | ||
818 | int qeth_core_create_osn_attributes(struct device *); | ||
819 | void qeth_core_remove_osn_attributes(struct device *); | ||
820 | |||
821 | /* exports for qeth discipline device drivers */ | ||
822 | extern struct qeth_card_list_struct qeth_core_card_list; | ||
823 | |||
824 | extern struct qeth_dbf_info qeth_dbf[QETH_DBF_INFOS]; | ||
825 | |||
826 | void qeth_set_allowed_threads(struct qeth_card *, unsigned long , int); | ||
827 | int qeth_threads_running(struct qeth_card *, unsigned long); | ||
828 | int qeth_wait_for_threads(struct qeth_card *, unsigned long); | ||
829 | int qeth_do_run_thread(struct qeth_card *, unsigned long); | ||
830 | void qeth_clear_thread_start_bit(struct qeth_card *, unsigned long); | ||
831 | void qeth_clear_thread_running_bit(struct qeth_card *, unsigned long); | ||
832 | int qeth_core_hardsetup_card(struct qeth_card *); | ||
833 | void qeth_print_status_message(struct qeth_card *); | ||
834 | int qeth_init_qdio_queues(struct qeth_card *); | ||
835 | int qeth_send_startlan(struct qeth_card *); | ||
836 | int qeth_send_stoplan(struct qeth_card *); | ||
837 | int qeth_send_ipa_cmd(struct qeth_card *, struct qeth_cmd_buffer *, | ||
838 | int (*reply_cb) | ||
839 | (struct qeth_card *, struct qeth_reply *, unsigned long), | ||
840 | void *); | ||
841 | struct qeth_cmd_buffer *qeth_get_ipacmd_buffer(struct qeth_card *, | ||
842 | enum qeth_ipa_cmds, enum qeth_prot_versions); | ||
843 | int qeth_query_setadapterparms(struct qeth_card *); | ||
844 | int qeth_check_qdio_errors(struct qdio_buffer *, unsigned int, | ||
845 | unsigned int, const char *); | ||
846 | void qeth_put_buffer_pool_entry(struct qeth_card *, | ||
847 | struct qeth_buffer_pool_entry *); | ||
848 | void qeth_queue_input_buffer(struct qeth_card *, int); | ||
849 | struct sk_buff *qeth_core_get_next_skb(struct qeth_card *, | ||
850 | struct qdio_buffer *, struct qdio_buffer_element **, int *, | ||
851 | struct qeth_hdr **); | ||
852 | void qeth_schedule_recovery(struct qeth_card *); | ||
853 | void qeth_qdio_output_handler(struct ccw_device *, unsigned int, | ||
854 | unsigned int, unsigned int, | ||
855 | unsigned int, int, int, | ||
856 | unsigned long); | ||
857 | void qeth_clear_ipacmd_list(struct qeth_card *); | ||
858 | int qeth_qdio_clear_card(struct qeth_card *, int); | ||
859 | void qeth_clear_working_pool_list(struct qeth_card *); | ||
860 | void qeth_clear_cmd_buffers(struct qeth_channel *); | ||
861 | void qeth_clear_qdio_buffers(struct qeth_card *); | ||
862 | void qeth_setadp_promisc_mode(struct qeth_card *); | ||
863 | struct net_device_stats *qeth_get_stats(struct net_device *); | ||
864 | int qeth_change_mtu(struct net_device *, int); | ||
865 | int qeth_setadpparms_change_macaddr(struct qeth_card *); | ||
866 | void qeth_tx_timeout(struct net_device *); | ||
867 | void qeth_prepare_control_data(struct qeth_card *, int, | ||
868 | struct qeth_cmd_buffer *); | ||
869 | void qeth_release_buffer(struct qeth_channel *, struct qeth_cmd_buffer *); | ||
870 | void qeth_prepare_ipa_cmd(struct qeth_card *, struct qeth_cmd_buffer *, char); | ||
871 | struct qeth_cmd_buffer *qeth_wait_for_buffer(struct qeth_channel *); | ||
872 | int qeth_mdio_read(struct net_device *, int, int); | ||
873 | int qeth_snmp_command(struct qeth_card *, char __user *); | ||
874 | int qeth_set_large_send(struct qeth_card *, enum qeth_large_send_types); | ||
875 | struct qeth_cmd_buffer *qeth_get_adapter_cmd(struct qeth_card *, __u32, __u32); | ||
876 | int qeth_default_setadapterparms_cb(struct qeth_card *, struct qeth_reply *, | ||
877 | unsigned long); | ||
878 | int qeth_send_control_data(struct qeth_card *, int, struct qeth_cmd_buffer *, | ||
879 | int (*reply_cb)(struct qeth_card *, struct qeth_reply*, unsigned long), | ||
880 | void *reply_param); | ||
881 | int qeth_get_cast_type(struct qeth_card *, struct sk_buff *); | ||
882 | int qeth_get_priority_queue(struct qeth_card *, struct sk_buff *, int, int); | ||
883 | struct sk_buff *qeth_prepare_skb(struct qeth_card *, struct sk_buff *, | ||
884 | struct qeth_hdr **); | ||
885 | int qeth_get_elements_no(struct qeth_card *, void *, struct sk_buff *, int); | ||
886 | int qeth_do_send_packet_fast(struct qeth_card *, struct qeth_qdio_out_q *, | ||
887 | struct sk_buff *, struct qeth_hdr *, int, | ||
888 | struct qeth_eddp_context *); | ||
889 | int qeth_do_send_packet(struct qeth_card *, struct qeth_qdio_out_q *, | ||
890 | struct sk_buff *, struct qeth_hdr *, | ||
891 | int, struct qeth_eddp_context *); | ||
892 | int qeth_core_get_stats_count(struct net_device *); | ||
893 | void qeth_core_get_ethtool_stats(struct net_device *, | ||
894 | struct ethtool_stats *, u64 *); | ||
895 | void qeth_core_get_strings(struct net_device *, u32, u8 *); | ||
896 | void qeth_core_get_drvinfo(struct net_device *, struct ethtool_drvinfo *); | ||
897 | |||
898 | /* exports for OSN */ | ||
899 | int qeth_osn_assist(struct net_device *, void *, int); | ||
900 | int qeth_osn_register(unsigned char *read_dev_no, struct net_device **, | ||
901 | int (*assist_cb)(struct net_device *, void *), | ||
902 | int (*data_cb)(struct sk_buff *)); | ||
903 | void qeth_osn_deregister(struct net_device *); | ||
904 | |||
905 | #endif /* __QETH_CORE_H__ */ | ||
diff --git a/drivers/s390/net/qeth_core_main.c b/drivers/s390/net/qeth_core_main.c new file mode 100644 index 000000000000..055f5c3e7b56 --- /dev/null +++ b/drivers/s390/net/qeth_core_main.c | |||
@@ -0,0 +1,4492 @@ | |||
1 | /* | ||
2 | * drivers/s390/net/qeth_core_main.c | ||
3 | * | ||
4 | * Copyright IBM Corp. 2007 | ||
5 | * Author(s): Utz Bacher <utz.bacher@de.ibm.com>, | ||
6 | * Frank Pavlic <fpavlic@de.ibm.com>, | ||
7 | * Thomas Spatzier <tspat@de.ibm.com>, | ||
8 | * Frank Blaschka <frank.blaschka@de.ibm.com> | ||
9 | */ | ||
10 | |||
11 | #include <linux/module.h> | ||
12 | #include <linux/moduleparam.h> | ||
13 | #include <linux/string.h> | ||
14 | #include <linux/errno.h> | ||
15 | #include <linux/kernel.h> | ||
16 | #include <linux/ip.h> | ||
17 | #include <linux/ipv6.h> | ||
18 | #include <linux/tcp.h> | ||
19 | #include <linux/mii.h> | ||
20 | #include <linux/kthread.h> | ||
21 | |||
22 | #include <asm-s390/ebcdic.h> | ||
23 | #include <asm-s390/io.h> | ||
24 | #include <asm/s390_rdev.h> | ||
25 | |||
26 | #include "qeth_core.h" | ||
27 | #include "qeth_core_offl.h" | ||
28 | |||
29 | static DEFINE_PER_CPU(char[256], qeth_core_dbf_txt_buf); | ||
30 | #define QETH_DBF_TXT_BUF qeth_core_dbf_txt_buf | ||
31 | |||
32 | struct qeth_dbf_info qeth_dbf[QETH_DBF_INFOS] = { | ||
33 | /* define dbf - Name, Pages, Areas, Maxlen, Level, View, Handle */ | ||
34 | /* N P A M L V H */ | ||
35 | [QETH_DBF_SETUP] = {"qeth_setup", | ||
36 | 8, 1, 8, 5, &debug_hex_ascii_view, NULL}, | ||
37 | [QETH_DBF_QERR] = {"qeth_qerr", | ||
38 | 2, 1, 8, 2, &debug_hex_ascii_view, NULL}, | ||
39 | [QETH_DBF_TRACE] = {"qeth_trace", | ||
40 | 4, 1, 8, 3, &debug_hex_ascii_view, NULL}, | ||
41 | [QETH_DBF_MSG] = {"qeth_msg", | ||
42 | 8, 1, 128, 3, &debug_sprintf_view, NULL}, | ||
43 | [QETH_DBF_SENSE] = {"qeth_sense", | ||
44 | 2, 1, 64, 2, &debug_hex_ascii_view, NULL}, | ||
45 | [QETH_DBF_MISC] = {"qeth_misc", | ||
46 | 2, 1, 256, 2, &debug_hex_ascii_view, NULL}, | ||
47 | [QETH_DBF_CTRL] = {"qeth_control", | ||
48 | 8, 1, QETH_DBF_CTRL_LEN, 5, &debug_hex_ascii_view, NULL}, | ||
49 | }; | ||
50 | EXPORT_SYMBOL_GPL(qeth_dbf); | ||
51 | |||
52 | struct qeth_card_list_struct qeth_core_card_list; | ||
53 | EXPORT_SYMBOL_GPL(qeth_core_card_list); | ||
54 | |||
55 | static struct device *qeth_core_root_dev; | ||
56 | static unsigned int known_devices[][10] = QETH_MODELLIST_ARRAY; | ||
57 | static struct lock_class_key qdio_out_skb_queue_key; | ||
58 | |||
59 | static void qeth_send_control_data_cb(struct qeth_channel *, | ||
60 | struct qeth_cmd_buffer *); | ||
61 | static int qeth_issue_next_read(struct qeth_card *); | ||
62 | static struct qeth_cmd_buffer *qeth_get_buffer(struct qeth_channel *); | ||
63 | static void qeth_setup_ccw(struct qeth_channel *, unsigned char *, __u32); | ||
64 | static void qeth_free_buffer_pool(struct qeth_card *); | ||
65 | static int qeth_qdio_establish(struct qeth_card *); | ||
66 | |||
67 | |||
68 | static inline void __qeth_fill_buffer_frag(struct sk_buff *skb, | ||
69 | struct qdio_buffer *buffer, int is_tso, | ||
70 | int *next_element_to_fill) | ||
71 | { | ||
72 | struct skb_frag_struct *frag; | ||
73 | int fragno; | ||
74 | unsigned long addr; | ||
75 | int element, cnt, dlen; | ||
76 | |||
77 | fragno = skb_shinfo(skb)->nr_frags; | ||
78 | element = *next_element_to_fill; | ||
79 | dlen = 0; | ||
80 | |||
81 | if (is_tso) | ||
82 | buffer->element[element].flags = | ||
83 | SBAL_FLAGS_MIDDLE_FRAG; | ||
84 | else | ||
85 | buffer->element[element].flags = | ||
86 | SBAL_FLAGS_FIRST_FRAG; | ||
87 | dlen = skb->len - skb->data_len; | ||
88 | if (dlen) { | ||
89 | buffer->element[element].addr = skb->data; | ||
90 | buffer->element[element].length = dlen; | ||
91 | element++; | ||
92 | } | ||
93 | for (cnt = 0; cnt < fragno; cnt++) { | ||
94 | frag = &skb_shinfo(skb)->frags[cnt]; | ||
95 | addr = (page_to_pfn(frag->page) << PAGE_SHIFT) + | ||
96 | frag->page_offset; | ||
97 | buffer->element[element].addr = (char *)addr; | ||
98 | buffer->element[element].length = frag->size; | ||
99 | if (cnt < (fragno - 1)) | ||
100 | buffer->element[element].flags = | ||
101 | SBAL_FLAGS_MIDDLE_FRAG; | ||
102 | else | ||
103 | buffer->element[element].flags = | ||
104 | SBAL_FLAGS_LAST_FRAG; | ||
105 | element++; | ||
106 | } | ||
107 | *next_element_to_fill = element; | ||
108 | } | ||
109 | |||
110 | static inline const char *qeth_get_cardname(struct qeth_card *card) | ||
111 | { | ||
112 | if (card->info.guestlan) { | ||
113 | switch (card->info.type) { | ||
114 | case QETH_CARD_TYPE_OSAE: | ||
115 | return " Guest LAN QDIO"; | ||
116 | case QETH_CARD_TYPE_IQD: | ||
117 | return " Guest LAN Hiper"; | ||
118 | default: | ||
119 | return " unknown"; | ||
120 | } | ||
121 | } else { | ||
122 | switch (card->info.type) { | ||
123 | case QETH_CARD_TYPE_OSAE: | ||
124 | return " OSD Express"; | ||
125 | case QETH_CARD_TYPE_IQD: | ||
126 | return " HiperSockets"; | ||
127 | case QETH_CARD_TYPE_OSN: | ||
128 | return " OSN QDIO"; | ||
129 | default: | ||
130 | return " unknown"; | ||
131 | } | ||
132 | } | ||
133 | return " n/a"; | ||
134 | } | ||
135 | |||
136 | /* max length to be returned: 14 */ | ||
137 | const char *qeth_get_cardname_short(struct qeth_card *card) | ||
138 | { | ||
139 | if (card->info.guestlan) { | ||
140 | switch (card->info.type) { | ||
141 | case QETH_CARD_TYPE_OSAE: | ||
142 | return "GuestLAN QDIO"; | ||
143 | case QETH_CARD_TYPE_IQD: | ||
144 | return "GuestLAN Hiper"; | ||
145 | default: | ||
146 | return "unknown"; | ||
147 | } | ||
148 | } else { | ||
149 | switch (card->info.type) { | ||
150 | case QETH_CARD_TYPE_OSAE: | ||
151 | switch (card->info.link_type) { | ||
152 | case QETH_LINK_TYPE_FAST_ETH: | ||
153 | return "OSD_100"; | ||
154 | case QETH_LINK_TYPE_HSTR: | ||
155 | return "HSTR"; | ||
156 | case QETH_LINK_TYPE_GBIT_ETH: | ||
157 | return "OSD_1000"; | ||
158 | case QETH_LINK_TYPE_10GBIT_ETH: | ||
159 | return "OSD_10GIG"; | ||
160 | case QETH_LINK_TYPE_LANE_ETH100: | ||
161 | return "OSD_FE_LANE"; | ||
162 | case QETH_LINK_TYPE_LANE_TR: | ||
163 | return "OSD_TR_LANE"; | ||
164 | case QETH_LINK_TYPE_LANE_ETH1000: | ||
165 | return "OSD_GbE_LANE"; | ||
166 | case QETH_LINK_TYPE_LANE: | ||
167 | return "OSD_ATM_LANE"; | ||
168 | default: | ||
169 | return "OSD_Express"; | ||
170 | } | ||
171 | case QETH_CARD_TYPE_IQD: | ||
172 | return "HiperSockets"; | ||
173 | case QETH_CARD_TYPE_OSN: | ||
174 | return "OSN"; | ||
175 | default: | ||
176 | return "unknown"; | ||
177 | } | ||
178 | } | ||
179 | return "n/a"; | ||
180 | } | ||
181 | |||
182 | void qeth_set_allowed_threads(struct qeth_card *card, unsigned long threads, | ||
183 | int clear_start_mask) | ||
184 | { | ||
185 | unsigned long flags; | ||
186 | |||
187 | spin_lock_irqsave(&card->thread_mask_lock, flags); | ||
188 | card->thread_allowed_mask = threads; | ||
189 | if (clear_start_mask) | ||
190 | card->thread_start_mask &= threads; | ||
191 | spin_unlock_irqrestore(&card->thread_mask_lock, flags); | ||
192 | wake_up(&card->wait_q); | ||
193 | } | ||
194 | EXPORT_SYMBOL_GPL(qeth_set_allowed_threads); | ||
195 | |||
196 | int qeth_threads_running(struct qeth_card *card, unsigned long threads) | ||
197 | { | ||
198 | unsigned long flags; | ||
199 | int rc = 0; | ||
200 | |||
201 | spin_lock_irqsave(&card->thread_mask_lock, flags); | ||
202 | rc = (card->thread_running_mask & threads); | ||
203 | spin_unlock_irqrestore(&card->thread_mask_lock, flags); | ||
204 | return rc; | ||
205 | } | ||
206 | EXPORT_SYMBOL_GPL(qeth_threads_running); | ||
207 | |||
208 | int qeth_wait_for_threads(struct qeth_card *card, unsigned long threads) | ||
209 | { | ||
210 | return wait_event_interruptible(card->wait_q, | ||
211 | qeth_threads_running(card, threads) == 0); | ||
212 | } | ||
213 | EXPORT_SYMBOL_GPL(qeth_wait_for_threads); | ||
214 | |||
215 | void qeth_clear_working_pool_list(struct qeth_card *card) | ||
216 | { | ||
217 | struct qeth_buffer_pool_entry *pool_entry, *tmp; | ||
218 | |||
219 | QETH_DBF_TEXT(TRACE, 5, "clwrklst"); | ||
220 | list_for_each_entry_safe(pool_entry, tmp, | ||
221 | &card->qdio.in_buf_pool.entry_list, list){ | ||
222 | list_del(&pool_entry->list); | ||
223 | } | ||
224 | } | ||
225 | EXPORT_SYMBOL_GPL(qeth_clear_working_pool_list); | ||
226 | |||
227 | static int qeth_alloc_buffer_pool(struct qeth_card *card) | ||
228 | { | ||
229 | struct qeth_buffer_pool_entry *pool_entry; | ||
230 | void *ptr; | ||
231 | int i, j; | ||
232 | |||
233 | QETH_DBF_TEXT(TRACE, 5, "alocpool"); | ||
234 | for (i = 0; i < card->qdio.init_pool.buf_count; ++i) { | ||
235 | pool_entry = kmalloc(sizeof(*pool_entry), GFP_KERNEL); | ||
236 | if (!pool_entry) { | ||
237 | qeth_free_buffer_pool(card); | ||
238 | return -ENOMEM; | ||
239 | } | ||
240 | for (j = 0; j < QETH_MAX_BUFFER_ELEMENTS(card); ++j) { | ||
241 | ptr = (void *) __get_free_page(GFP_KERNEL); | ||
242 | if (!ptr) { | ||
243 | while (j > 0) | ||
244 | free_page((unsigned long) | ||
245 | pool_entry->elements[--j]); | ||
246 | kfree(pool_entry); | ||
247 | qeth_free_buffer_pool(card); | ||
248 | return -ENOMEM; | ||
249 | } | ||
250 | pool_entry->elements[j] = ptr; | ||
251 | } | ||
252 | list_add(&pool_entry->init_list, | ||
253 | &card->qdio.init_pool.entry_list); | ||
254 | } | ||
255 | return 0; | ||
256 | } | ||
257 | |||
258 | int qeth_realloc_buffer_pool(struct qeth_card *card, int bufcnt) | ||
259 | { | ||
260 | QETH_DBF_TEXT(TRACE, 2, "realcbp"); | ||
261 | |||
262 | if ((card->state != CARD_STATE_DOWN) && | ||
263 | (card->state != CARD_STATE_RECOVER)) | ||
264 | return -EPERM; | ||
265 | |||
266 | /* TODO: steel/add buffers from/to a running card's buffer pool (?) */ | ||
267 | qeth_clear_working_pool_list(card); | ||
268 | qeth_free_buffer_pool(card); | ||
269 | card->qdio.in_buf_pool.buf_count = bufcnt; | ||
270 | card->qdio.init_pool.buf_count = bufcnt; | ||
271 | return qeth_alloc_buffer_pool(card); | ||
272 | } | ||
273 | |||
274 | int qeth_set_large_send(struct qeth_card *card, | ||
275 | enum qeth_large_send_types type) | ||
276 | { | ||
277 | int rc = 0; | ||
278 | |||
279 | if (card->dev == NULL) { | ||
280 | card->options.large_send = type; | ||
281 | return 0; | ||
282 | } | ||
283 | if (card->state == CARD_STATE_UP) | ||
284 | netif_tx_disable(card->dev); | ||
285 | card->options.large_send = type; | ||
286 | switch (card->options.large_send) { | ||
287 | case QETH_LARGE_SEND_EDDP: | ||
288 | card->dev->features |= NETIF_F_TSO | NETIF_F_SG | | ||
289 | NETIF_F_HW_CSUM; | ||
290 | break; | ||
291 | case QETH_LARGE_SEND_TSO: | ||
292 | if (qeth_is_supported(card, IPA_OUTBOUND_TSO)) { | ||
293 | card->dev->features |= NETIF_F_TSO | NETIF_F_SG | | ||
294 | NETIF_F_HW_CSUM; | ||
295 | } else { | ||
296 | PRINT_WARN("TSO not supported on %s. " | ||
297 | "large_send set to 'no'.\n", | ||
298 | card->dev->name); | ||
299 | card->dev->features &= ~(NETIF_F_TSO | NETIF_F_SG | | ||
300 | NETIF_F_HW_CSUM); | ||
301 | card->options.large_send = QETH_LARGE_SEND_NO; | ||
302 | rc = -EOPNOTSUPP; | ||
303 | } | ||
304 | break; | ||
305 | default: /* includes QETH_LARGE_SEND_NO */ | ||
306 | card->dev->features &= ~(NETIF_F_TSO | NETIF_F_SG | | ||
307 | NETIF_F_HW_CSUM); | ||
308 | break; | ||
309 | } | ||
310 | if (card->state == CARD_STATE_UP) | ||
311 | netif_wake_queue(card->dev); | ||
312 | return rc; | ||
313 | } | ||
314 | EXPORT_SYMBOL_GPL(qeth_set_large_send); | ||
315 | |||
316 | static int qeth_issue_next_read(struct qeth_card *card) | ||
317 | { | ||
318 | int rc; | ||
319 | struct qeth_cmd_buffer *iob; | ||
320 | |||
321 | QETH_DBF_TEXT(TRACE, 5, "issnxrd"); | ||
322 | if (card->read.state != CH_STATE_UP) | ||
323 | return -EIO; | ||
324 | iob = qeth_get_buffer(&card->read); | ||
325 | if (!iob) { | ||
326 | PRINT_WARN("issue_next_read failed: no iob available!\n"); | ||
327 | return -ENOMEM; | ||
328 | } | ||
329 | qeth_setup_ccw(&card->read, iob->data, QETH_BUFSIZE); | ||
330 | QETH_DBF_TEXT(TRACE, 6, "noirqpnd"); | ||
331 | rc = ccw_device_start(card->read.ccwdev, &card->read.ccw, | ||
332 | (addr_t) iob, 0, 0); | ||
333 | if (rc) { | ||
334 | PRINT_ERR("Error in starting next read ccw! rc=%i\n", rc); | ||
335 | atomic_set(&card->read.irq_pending, 0); | ||
336 | qeth_schedule_recovery(card); | ||
337 | wake_up(&card->wait_q); | ||
338 | } | ||
339 | return rc; | ||
340 | } | ||
341 | |||
342 | static struct qeth_reply *qeth_alloc_reply(struct qeth_card *card) | ||
343 | { | ||
344 | struct qeth_reply *reply; | ||
345 | |||
346 | reply = kzalloc(sizeof(struct qeth_reply), GFP_ATOMIC); | ||
347 | if (reply) { | ||
348 | atomic_set(&reply->refcnt, 1); | ||
349 | atomic_set(&reply->received, 0); | ||
350 | reply->card = card; | ||
351 | }; | ||
352 | return reply; | ||
353 | } | ||
354 | |||
355 | static void qeth_get_reply(struct qeth_reply *reply) | ||
356 | { | ||
357 | WARN_ON(atomic_read(&reply->refcnt) <= 0); | ||
358 | atomic_inc(&reply->refcnt); | ||
359 | } | ||
360 | |||
361 | static void qeth_put_reply(struct qeth_reply *reply) | ||
362 | { | ||
363 | WARN_ON(atomic_read(&reply->refcnt) <= 0); | ||
364 | if (atomic_dec_and_test(&reply->refcnt)) | ||
365 | kfree(reply); | ||
366 | } | ||
367 | |||
368 | static void qeth_issue_ipa_msg(struct qeth_ipa_cmd *cmd, int rc, | ||
369 | struct qeth_card *card) | ||
370 | { | ||
371 | char *ipa_name; | ||
372 | int com = cmd->hdr.command; | ||
373 | ipa_name = qeth_get_ipa_cmd_name(com); | ||
374 | if (rc) | ||
375 | QETH_DBF_MESSAGE(2, "IPA: %s(x%X) for %s returned x%X \"%s\"\n", | ||
376 | ipa_name, com, QETH_CARD_IFNAME(card), | ||
377 | rc, qeth_get_ipa_msg(rc)); | ||
378 | else | ||
379 | QETH_DBF_MESSAGE(5, "IPA: %s(x%X) for %s succeeded\n", | ||
380 | ipa_name, com, QETH_CARD_IFNAME(card)); | ||
381 | } | ||
382 | |||
383 | static struct qeth_ipa_cmd *qeth_check_ipa_data(struct qeth_card *card, | ||
384 | struct qeth_cmd_buffer *iob) | ||
385 | { | ||
386 | struct qeth_ipa_cmd *cmd = NULL; | ||
387 | |||
388 | QETH_DBF_TEXT(TRACE, 5, "chkipad"); | ||
389 | if (IS_IPA(iob->data)) { | ||
390 | cmd = (struct qeth_ipa_cmd *) PDU_ENCAPSULATION(iob->data); | ||
391 | if (IS_IPA_REPLY(cmd)) { | ||
392 | if (cmd->hdr.command < IPA_CMD_SETCCID || | ||
393 | cmd->hdr.command > IPA_CMD_MODCCID) | ||
394 | qeth_issue_ipa_msg(cmd, | ||
395 | cmd->hdr.return_code, card); | ||
396 | return cmd; | ||
397 | } else { | ||
398 | switch (cmd->hdr.command) { | ||
399 | case IPA_CMD_STOPLAN: | ||
400 | PRINT_WARN("Link failure on %s (CHPID 0x%X) - " | ||
401 | "there is a network problem or " | ||
402 | "someone pulled the cable or " | ||
403 | "disabled the port.\n", | ||
404 | QETH_CARD_IFNAME(card), | ||
405 | card->info.chpid); | ||
406 | card->lan_online = 0; | ||
407 | if (card->dev && netif_carrier_ok(card->dev)) | ||
408 | netif_carrier_off(card->dev); | ||
409 | return NULL; | ||
410 | case IPA_CMD_STARTLAN: | ||
411 | PRINT_INFO("Link reestablished on %s " | ||
412 | "(CHPID 0x%X). Scheduling " | ||
413 | "IP address reset.\n", | ||
414 | QETH_CARD_IFNAME(card), | ||
415 | card->info.chpid); | ||
416 | netif_carrier_on(card->dev); | ||
417 | card->lan_online = 1; | ||
418 | qeth_schedule_recovery(card); | ||
419 | return NULL; | ||
420 | case IPA_CMD_MODCCID: | ||
421 | return cmd; | ||
422 | case IPA_CMD_REGISTER_LOCAL_ADDR: | ||
423 | QETH_DBF_TEXT(TRACE, 3, "irla"); | ||
424 | break; | ||
425 | case IPA_CMD_UNREGISTER_LOCAL_ADDR: | ||
426 | QETH_DBF_TEXT(TRACE, 3, "urla"); | ||
427 | break; | ||
428 | default: | ||
429 | PRINT_WARN("Received data is IPA " | ||
430 | "but not a reply!\n"); | ||
431 | break; | ||
432 | } | ||
433 | } | ||
434 | } | ||
435 | return cmd; | ||
436 | } | ||
437 | |||
438 | void qeth_clear_ipacmd_list(struct qeth_card *card) | ||
439 | { | ||
440 | struct qeth_reply *reply, *r; | ||
441 | unsigned long flags; | ||
442 | |||
443 | QETH_DBF_TEXT(TRACE, 4, "clipalst"); | ||
444 | |||
445 | spin_lock_irqsave(&card->lock, flags); | ||
446 | list_for_each_entry_safe(reply, r, &card->cmd_waiter_list, list) { | ||
447 | qeth_get_reply(reply); | ||
448 | reply->rc = -EIO; | ||
449 | atomic_inc(&reply->received); | ||
450 | list_del_init(&reply->list); | ||
451 | wake_up(&reply->wait_q); | ||
452 | qeth_put_reply(reply); | ||
453 | } | ||
454 | spin_unlock_irqrestore(&card->lock, flags); | ||
455 | } | ||
456 | EXPORT_SYMBOL_GPL(qeth_clear_ipacmd_list); | ||
457 | |||
458 | static int qeth_check_idx_response(unsigned char *buffer) | ||
459 | { | ||
460 | if (!buffer) | ||
461 | return 0; | ||
462 | |||
463 | QETH_DBF_HEX(CTRL, 2, buffer, QETH_DBF_CTRL_LEN); | ||
464 | if ((buffer[2] & 0xc0) == 0xc0) { | ||
465 | PRINT_WARN("received an IDX TERMINATE " | ||
466 | "with cause code 0x%02x%s\n", | ||
467 | buffer[4], | ||
468 | ((buffer[4] == 0x22) ? | ||
469 | " -- try another portname" : "")); | ||
470 | QETH_DBF_TEXT(TRACE, 2, "ckidxres"); | ||
471 | QETH_DBF_TEXT(TRACE, 2, " idxterm"); | ||
472 | QETH_DBF_TEXT_(TRACE, 2, " rc%d", -EIO); | ||
473 | return -EIO; | ||
474 | } | ||
475 | return 0; | ||
476 | } | ||
477 | |||
478 | static void qeth_setup_ccw(struct qeth_channel *channel, unsigned char *iob, | ||
479 | __u32 len) | ||
480 | { | ||
481 | struct qeth_card *card; | ||
482 | |||
483 | QETH_DBF_TEXT(TRACE, 4, "setupccw"); | ||
484 | card = CARD_FROM_CDEV(channel->ccwdev); | ||
485 | if (channel == &card->read) | ||
486 | memcpy(&channel->ccw, READ_CCW, sizeof(struct ccw1)); | ||
487 | else | ||
488 | memcpy(&channel->ccw, WRITE_CCW, sizeof(struct ccw1)); | ||
489 | channel->ccw.count = len; | ||
490 | channel->ccw.cda = (__u32) __pa(iob); | ||
491 | } | ||
492 | |||
493 | static struct qeth_cmd_buffer *__qeth_get_buffer(struct qeth_channel *channel) | ||
494 | { | ||
495 | __u8 index; | ||
496 | |||
497 | QETH_DBF_TEXT(TRACE, 6, "getbuff"); | ||
498 | index = channel->io_buf_no; | ||
499 | do { | ||
500 | if (channel->iob[index].state == BUF_STATE_FREE) { | ||
501 | channel->iob[index].state = BUF_STATE_LOCKED; | ||
502 | channel->io_buf_no = (channel->io_buf_no + 1) % | ||
503 | QETH_CMD_BUFFER_NO; | ||
504 | memset(channel->iob[index].data, 0, QETH_BUFSIZE); | ||
505 | return channel->iob + index; | ||
506 | } | ||
507 | index = (index + 1) % QETH_CMD_BUFFER_NO; | ||
508 | } while (index != channel->io_buf_no); | ||
509 | |||
510 | return NULL; | ||
511 | } | ||
512 | |||
513 | void qeth_release_buffer(struct qeth_channel *channel, | ||
514 | struct qeth_cmd_buffer *iob) | ||
515 | { | ||
516 | unsigned long flags; | ||
517 | |||
518 | QETH_DBF_TEXT(TRACE, 6, "relbuff"); | ||
519 | spin_lock_irqsave(&channel->iob_lock, flags); | ||
520 | memset(iob->data, 0, QETH_BUFSIZE); | ||
521 | iob->state = BUF_STATE_FREE; | ||
522 | iob->callback = qeth_send_control_data_cb; | ||
523 | iob->rc = 0; | ||
524 | spin_unlock_irqrestore(&channel->iob_lock, flags); | ||
525 | } | ||
526 | EXPORT_SYMBOL_GPL(qeth_release_buffer); | ||
527 | |||
528 | static struct qeth_cmd_buffer *qeth_get_buffer(struct qeth_channel *channel) | ||
529 | { | ||
530 | struct qeth_cmd_buffer *buffer = NULL; | ||
531 | unsigned long flags; | ||
532 | |||
533 | spin_lock_irqsave(&channel->iob_lock, flags); | ||
534 | buffer = __qeth_get_buffer(channel); | ||
535 | spin_unlock_irqrestore(&channel->iob_lock, flags); | ||
536 | return buffer; | ||
537 | } | ||
538 | |||
539 | struct qeth_cmd_buffer *qeth_wait_for_buffer(struct qeth_channel *channel) | ||
540 | { | ||
541 | struct qeth_cmd_buffer *buffer; | ||
542 | wait_event(channel->wait_q, | ||
543 | ((buffer = qeth_get_buffer(channel)) != NULL)); | ||
544 | return buffer; | ||
545 | } | ||
546 | EXPORT_SYMBOL_GPL(qeth_wait_for_buffer); | ||
547 | |||
548 | void qeth_clear_cmd_buffers(struct qeth_channel *channel) | ||
549 | { | ||
550 | int cnt; | ||
551 | |||
552 | for (cnt = 0; cnt < QETH_CMD_BUFFER_NO; cnt++) | ||
553 | qeth_release_buffer(channel, &channel->iob[cnt]); | ||
554 | channel->buf_no = 0; | ||
555 | channel->io_buf_no = 0; | ||
556 | } | ||
557 | EXPORT_SYMBOL_GPL(qeth_clear_cmd_buffers); | ||
558 | |||
559 | static void qeth_send_control_data_cb(struct qeth_channel *channel, | ||
560 | struct qeth_cmd_buffer *iob) | ||
561 | { | ||
562 | struct qeth_card *card; | ||
563 | struct qeth_reply *reply, *r; | ||
564 | struct qeth_ipa_cmd *cmd; | ||
565 | unsigned long flags; | ||
566 | int keep_reply; | ||
567 | |||
568 | QETH_DBF_TEXT(TRACE, 4, "sndctlcb"); | ||
569 | |||
570 | card = CARD_FROM_CDEV(channel->ccwdev); | ||
571 | if (qeth_check_idx_response(iob->data)) { | ||
572 | qeth_clear_ipacmd_list(card); | ||
573 | qeth_schedule_recovery(card); | ||
574 | goto out; | ||
575 | } | ||
576 | |||
577 | cmd = qeth_check_ipa_data(card, iob); | ||
578 | if ((cmd == NULL) && (card->state != CARD_STATE_DOWN)) | ||
579 | goto out; | ||
580 | /*in case of OSN : check if cmd is set */ | ||
581 | if (card->info.type == QETH_CARD_TYPE_OSN && | ||
582 | cmd && | ||
583 | cmd->hdr.command != IPA_CMD_STARTLAN && | ||
584 | card->osn_info.assist_cb != NULL) { | ||
585 | card->osn_info.assist_cb(card->dev, cmd); | ||
586 | goto out; | ||
587 | } | ||
588 | |||
589 | spin_lock_irqsave(&card->lock, flags); | ||
590 | list_for_each_entry_safe(reply, r, &card->cmd_waiter_list, list) { | ||
591 | if ((reply->seqno == QETH_IDX_COMMAND_SEQNO) || | ||
592 | ((cmd) && (reply->seqno == cmd->hdr.seqno))) { | ||
593 | qeth_get_reply(reply); | ||
594 | list_del_init(&reply->list); | ||
595 | spin_unlock_irqrestore(&card->lock, flags); | ||
596 | keep_reply = 0; | ||
597 | if (reply->callback != NULL) { | ||
598 | if (cmd) { | ||
599 | reply->offset = (__u16)((char *)cmd - | ||
600 | (char *)iob->data); | ||
601 | keep_reply = reply->callback(card, | ||
602 | reply, | ||
603 | (unsigned long)cmd); | ||
604 | } else | ||
605 | keep_reply = reply->callback(card, | ||
606 | reply, | ||
607 | (unsigned long)iob); | ||
608 | } | ||
609 | if (cmd) | ||
610 | reply->rc = (u16) cmd->hdr.return_code; | ||
611 | else if (iob->rc) | ||
612 | reply->rc = iob->rc; | ||
613 | if (keep_reply) { | ||
614 | spin_lock_irqsave(&card->lock, flags); | ||
615 | list_add_tail(&reply->list, | ||
616 | &card->cmd_waiter_list); | ||
617 | spin_unlock_irqrestore(&card->lock, flags); | ||
618 | } else { | ||
619 | atomic_inc(&reply->received); | ||
620 | wake_up(&reply->wait_q); | ||
621 | } | ||
622 | qeth_put_reply(reply); | ||
623 | goto out; | ||
624 | } | ||
625 | } | ||
626 | spin_unlock_irqrestore(&card->lock, flags); | ||
627 | out: | ||
628 | memcpy(&card->seqno.pdu_hdr_ack, | ||
629 | QETH_PDU_HEADER_SEQ_NO(iob->data), | ||
630 | QETH_SEQ_NO_LENGTH); | ||
631 | qeth_release_buffer(channel, iob); | ||
632 | } | ||
633 | |||
634 | static int qeth_setup_channel(struct qeth_channel *channel) | ||
635 | { | ||
636 | int cnt; | ||
637 | |||
638 | QETH_DBF_TEXT(SETUP, 2, "setupch"); | ||
639 | for (cnt = 0; cnt < QETH_CMD_BUFFER_NO; cnt++) { | ||
640 | channel->iob[cnt].data = (char *) | ||
641 | kmalloc(QETH_BUFSIZE, GFP_DMA|GFP_KERNEL); | ||
642 | if (channel->iob[cnt].data == NULL) | ||
643 | break; | ||
644 | channel->iob[cnt].state = BUF_STATE_FREE; | ||
645 | channel->iob[cnt].channel = channel; | ||
646 | channel->iob[cnt].callback = qeth_send_control_data_cb; | ||
647 | channel->iob[cnt].rc = 0; | ||
648 | } | ||
649 | if (cnt < QETH_CMD_BUFFER_NO) { | ||
650 | while (cnt-- > 0) | ||
651 | kfree(channel->iob[cnt].data); | ||
652 | return -ENOMEM; | ||
653 | } | ||
654 | channel->buf_no = 0; | ||
655 | channel->io_buf_no = 0; | ||
656 | atomic_set(&channel->irq_pending, 0); | ||
657 | spin_lock_init(&channel->iob_lock); | ||
658 | |||
659 | init_waitqueue_head(&channel->wait_q); | ||
660 | return 0; | ||
661 | } | ||
662 | |||
663 | static int qeth_set_thread_start_bit(struct qeth_card *card, | ||
664 | unsigned long thread) | ||
665 | { | ||
666 | unsigned long flags; | ||
667 | |||
668 | spin_lock_irqsave(&card->thread_mask_lock, flags); | ||
669 | if (!(card->thread_allowed_mask & thread) || | ||
670 | (card->thread_start_mask & thread)) { | ||
671 | spin_unlock_irqrestore(&card->thread_mask_lock, flags); | ||
672 | return -EPERM; | ||
673 | } | ||
674 | card->thread_start_mask |= thread; | ||
675 | spin_unlock_irqrestore(&card->thread_mask_lock, flags); | ||
676 | return 0; | ||
677 | } | ||
678 | |||
679 | void qeth_clear_thread_start_bit(struct qeth_card *card, unsigned long thread) | ||
680 | { | ||
681 | unsigned long flags; | ||
682 | |||
683 | spin_lock_irqsave(&card->thread_mask_lock, flags); | ||
684 | card->thread_start_mask &= ~thread; | ||
685 | spin_unlock_irqrestore(&card->thread_mask_lock, flags); | ||
686 | wake_up(&card->wait_q); | ||
687 | } | ||
688 | EXPORT_SYMBOL_GPL(qeth_clear_thread_start_bit); | ||
689 | |||
690 | void qeth_clear_thread_running_bit(struct qeth_card *card, unsigned long thread) | ||
691 | { | ||
692 | unsigned long flags; | ||
693 | |||
694 | spin_lock_irqsave(&card->thread_mask_lock, flags); | ||
695 | card->thread_running_mask &= ~thread; | ||
696 | spin_unlock_irqrestore(&card->thread_mask_lock, flags); | ||
697 | wake_up(&card->wait_q); | ||
698 | } | ||
699 | EXPORT_SYMBOL_GPL(qeth_clear_thread_running_bit); | ||
700 | |||
701 | static int __qeth_do_run_thread(struct qeth_card *card, unsigned long thread) | ||
702 | { | ||
703 | unsigned long flags; | ||
704 | int rc = 0; | ||
705 | |||
706 | spin_lock_irqsave(&card->thread_mask_lock, flags); | ||
707 | if (card->thread_start_mask & thread) { | ||
708 | if ((card->thread_allowed_mask & thread) && | ||
709 | !(card->thread_running_mask & thread)) { | ||
710 | rc = 1; | ||
711 | card->thread_start_mask &= ~thread; | ||
712 | card->thread_running_mask |= thread; | ||
713 | } else | ||
714 | rc = -EPERM; | ||
715 | } | ||
716 | spin_unlock_irqrestore(&card->thread_mask_lock, flags); | ||
717 | return rc; | ||
718 | } | ||
719 | |||
720 | int qeth_do_run_thread(struct qeth_card *card, unsigned long thread) | ||
721 | { | ||
722 | int rc = 0; | ||
723 | |||
724 | wait_event(card->wait_q, | ||
725 | (rc = __qeth_do_run_thread(card, thread)) >= 0); | ||
726 | return rc; | ||
727 | } | ||
728 | EXPORT_SYMBOL_GPL(qeth_do_run_thread); | ||
729 | |||
730 | void qeth_schedule_recovery(struct qeth_card *card) | ||
731 | { | ||
732 | QETH_DBF_TEXT(TRACE, 2, "startrec"); | ||
733 | if (qeth_set_thread_start_bit(card, QETH_RECOVER_THREAD) == 0) | ||
734 | schedule_work(&card->kernel_thread_starter); | ||
735 | } | ||
736 | EXPORT_SYMBOL_GPL(qeth_schedule_recovery); | ||
737 | |||
738 | static int qeth_get_problem(struct ccw_device *cdev, struct irb *irb) | ||
739 | { | ||
740 | int dstat, cstat; | ||
741 | char *sense; | ||
742 | |||
743 | sense = (char *) irb->ecw; | ||
744 | cstat = irb->scsw.cstat; | ||
745 | dstat = irb->scsw.dstat; | ||
746 | |||
747 | if (cstat & (SCHN_STAT_CHN_CTRL_CHK | SCHN_STAT_INTF_CTRL_CHK | | ||
748 | SCHN_STAT_CHN_DATA_CHK | SCHN_STAT_CHAIN_CHECK | | ||
749 | SCHN_STAT_PROT_CHECK | SCHN_STAT_PROG_CHECK)) { | ||
750 | QETH_DBF_TEXT(TRACE, 2, "CGENCHK"); | ||
751 | PRINT_WARN("check on device %s, dstat=x%x, cstat=x%x ", | ||
752 | cdev->dev.bus_id, dstat, cstat); | ||
753 | print_hex_dump(KERN_WARNING, "qeth: irb ", DUMP_PREFIX_OFFSET, | ||
754 | 16, 1, irb, 64, 1); | ||
755 | return 1; | ||
756 | } | ||
757 | |||
758 | if (dstat & DEV_STAT_UNIT_CHECK) { | ||
759 | if (sense[SENSE_RESETTING_EVENT_BYTE] & | ||
760 | SENSE_RESETTING_EVENT_FLAG) { | ||
761 | QETH_DBF_TEXT(TRACE, 2, "REVIND"); | ||
762 | return 1; | ||
763 | } | ||
764 | if (sense[SENSE_COMMAND_REJECT_BYTE] & | ||
765 | SENSE_COMMAND_REJECT_FLAG) { | ||
766 | QETH_DBF_TEXT(TRACE, 2, "CMDREJi"); | ||
767 | return 0; | ||
768 | } | ||
769 | if ((sense[2] == 0xaf) && (sense[3] == 0xfe)) { | ||
770 | QETH_DBF_TEXT(TRACE, 2, "AFFE"); | ||
771 | return 1; | ||
772 | } | ||
773 | if ((!sense[0]) && (!sense[1]) && (!sense[2]) && (!sense[3])) { | ||
774 | QETH_DBF_TEXT(TRACE, 2, "ZEROSEN"); | ||
775 | return 0; | ||
776 | } | ||
777 | QETH_DBF_TEXT(TRACE, 2, "DGENCHK"); | ||
778 | return 1; | ||
779 | } | ||
780 | return 0; | ||
781 | } | ||
782 | |||
783 | static long __qeth_check_irb_error(struct ccw_device *cdev, | ||
784 | unsigned long intparm, struct irb *irb) | ||
785 | { | ||
786 | if (!IS_ERR(irb)) | ||
787 | return 0; | ||
788 | |||
789 | switch (PTR_ERR(irb)) { | ||
790 | case -EIO: | ||
791 | PRINT_WARN("i/o-error on device %s\n", cdev->dev.bus_id); | ||
792 | QETH_DBF_TEXT(TRACE, 2, "ckirberr"); | ||
793 | QETH_DBF_TEXT_(TRACE, 2, " rc%d", -EIO); | ||
794 | break; | ||
795 | case -ETIMEDOUT: | ||
796 | PRINT_WARN("timeout on device %s\n", cdev->dev.bus_id); | ||
797 | QETH_DBF_TEXT(TRACE, 2, "ckirberr"); | ||
798 | QETH_DBF_TEXT_(TRACE, 2, " rc%d", -ETIMEDOUT); | ||
799 | if (intparm == QETH_RCD_PARM) { | ||
800 | struct qeth_card *card = CARD_FROM_CDEV(cdev); | ||
801 | |||
802 | if (card && (card->data.ccwdev == cdev)) { | ||
803 | card->data.state = CH_STATE_DOWN; | ||
804 | wake_up(&card->wait_q); | ||
805 | } | ||
806 | } | ||
807 | break; | ||
808 | default: | ||
809 | PRINT_WARN("unknown error %ld on device %s\n", PTR_ERR(irb), | ||
810 | cdev->dev.bus_id); | ||
811 | QETH_DBF_TEXT(TRACE, 2, "ckirberr"); | ||
812 | QETH_DBF_TEXT(TRACE, 2, " rc???"); | ||
813 | } | ||
814 | return PTR_ERR(irb); | ||
815 | } | ||
816 | |||
817 | static void qeth_irq(struct ccw_device *cdev, unsigned long intparm, | ||
818 | struct irb *irb) | ||
819 | { | ||
820 | int rc; | ||
821 | int cstat, dstat; | ||
822 | struct qeth_cmd_buffer *buffer; | ||
823 | struct qeth_channel *channel; | ||
824 | struct qeth_card *card; | ||
825 | struct qeth_cmd_buffer *iob; | ||
826 | __u8 index; | ||
827 | |||
828 | QETH_DBF_TEXT(TRACE, 5, "irq"); | ||
829 | |||
830 | if (__qeth_check_irb_error(cdev, intparm, irb)) | ||
831 | return; | ||
832 | cstat = irb->scsw.cstat; | ||
833 | dstat = irb->scsw.dstat; | ||
834 | |||
835 | card = CARD_FROM_CDEV(cdev); | ||
836 | if (!card) | ||
837 | return; | ||
838 | |||
839 | if (card->read.ccwdev == cdev) { | ||
840 | channel = &card->read; | ||
841 | QETH_DBF_TEXT(TRACE, 5, "read"); | ||
842 | } else if (card->write.ccwdev == cdev) { | ||
843 | channel = &card->write; | ||
844 | QETH_DBF_TEXT(TRACE, 5, "write"); | ||
845 | } else { | ||
846 | channel = &card->data; | ||
847 | QETH_DBF_TEXT(TRACE, 5, "data"); | ||
848 | } | ||
849 | atomic_set(&channel->irq_pending, 0); | ||
850 | |||
851 | if (irb->scsw.fctl & (SCSW_FCTL_CLEAR_FUNC)) | ||
852 | channel->state = CH_STATE_STOPPED; | ||
853 | |||
854 | if (irb->scsw.fctl & (SCSW_FCTL_HALT_FUNC)) | ||
855 | channel->state = CH_STATE_HALTED; | ||
856 | |||
857 | /*let's wake up immediately on data channel*/ | ||
858 | if ((channel == &card->data) && (intparm != 0) && | ||
859 | (intparm != QETH_RCD_PARM)) | ||
860 | goto out; | ||
861 | |||
862 | if (intparm == QETH_CLEAR_CHANNEL_PARM) { | ||
863 | QETH_DBF_TEXT(TRACE, 6, "clrchpar"); | ||
864 | /* we don't have to handle this further */ | ||
865 | intparm = 0; | ||
866 | } | ||
867 | if (intparm == QETH_HALT_CHANNEL_PARM) { | ||
868 | QETH_DBF_TEXT(TRACE, 6, "hltchpar"); | ||
869 | /* we don't have to handle this further */ | ||
870 | intparm = 0; | ||
871 | } | ||
872 | if ((dstat & DEV_STAT_UNIT_EXCEP) || | ||
873 | (dstat & DEV_STAT_UNIT_CHECK) || | ||
874 | (cstat)) { | ||
875 | if (irb->esw.esw0.erw.cons) { | ||
876 | /* TODO: we should make this s390dbf */ | ||
877 | PRINT_WARN("sense data available on channel %s.\n", | ||
878 | CHANNEL_ID(channel)); | ||
879 | PRINT_WARN(" cstat 0x%X\n dstat 0x%X\n", cstat, dstat); | ||
880 | print_hex_dump(KERN_WARNING, "qeth: irb ", | ||
881 | DUMP_PREFIX_OFFSET, 16, 1, irb, 32, 1); | ||
882 | print_hex_dump(KERN_WARNING, "qeth: sense data ", | ||
883 | DUMP_PREFIX_OFFSET, 16, 1, irb->ecw, 32, 1); | ||
884 | } | ||
885 | if (intparm == QETH_RCD_PARM) { | ||
886 | channel->state = CH_STATE_DOWN; | ||
887 | goto out; | ||
888 | } | ||
889 | rc = qeth_get_problem(cdev, irb); | ||
890 | if (rc) { | ||
891 | qeth_schedule_recovery(card); | ||
892 | goto out; | ||
893 | } | ||
894 | } | ||
895 | |||
896 | if (intparm == QETH_RCD_PARM) { | ||
897 | channel->state = CH_STATE_RCD_DONE; | ||
898 | goto out; | ||
899 | } | ||
900 | if (intparm) { | ||
901 | buffer = (struct qeth_cmd_buffer *) __va((addr_t)intparm); | ||
902 | buffer->state = BUF_STATE_PROCESSED; | ||
903 | } | ||
904 | if (channel == &card->data) | ||
905 | return; | ||
906 | if (channel == &card->read && | ||
907 | channel->state == CH_STATE_UP) | ||
908 | qeth_issue_next_read(card); | ||
909 | |||
910 | iob = channel->iob; | ||
911 | index = channel->buf_no; | ||
912 | while (iob[index].state == BUF_STATE_PROCESSED) { | ||
913 | if (iob[index].callback != NULL) | ||
914 | iob[index].callback(channel, iob + index); | ||
915 | |||
916 | index = (index + 1) % QETH_CMD_BUFFER_NO; | ||
917 | } | ||
918 | channel->buf_no = index; | ||
919 | out: | ||
920 | wake_up(&card->wait_q); | ||
921 | return; | ||
922 | } | ||
923 | |||
924 | static void qeth_clear_output_buffer(struct qeth_qdio_out_q *queue, | ||
925 | struct qeth_qdio_out_buffer *buf) | ||
926 | { | ||
927 | int i; | ||
928 | struct sk_buff *skb; | ||
929 | |||
930 | /* is PCI flag set on buffer? */ | ||
931 | if (buf->buffer->element[0].flags & 0x40) | ||
932 | atomic_dec(&queue->set_pci_flags_count); | ||
933 | |||
934 | skb = skb_dequeue(&buf->skb_list); | ||
935 | while (skb) { | ||
936 | atomic_dec(&skb->users); | ||
937 | dev_kfree_skb_any(skb); | ||
938 | skb = skb_dequeue(&buf->skb_list); | ||
939 | } | ||
940 | qeth_eddp_buf_release_contexts(buf); | ||
941 | for (i = 0; i < QETH_MAX_BUFFER_ELEMENTS(queue->card); ++i) { | ||
942 | buf->buffer->element[i].length = 0; | ||
943 | buf->buffer->element[i].addr = NULL; | ||
944 | buf->buffer->element[i].flags = 0; | ||
945 | } | ||
946 | buf->next_element_to_fill = 0; | ||
947 | atomic_set(&buf->state, QETH_QDIO_BUF_EMPTY); | ||
948 | } | ||
949 | |||
950 | void qeth_clear_qdio_buffers(struct qeth_card *card) | ||
951 | { | ||
952 | int i, j; | ||
953 | |||
954 | QETH_DBF_TEXT(TRACE, 2, "clearqdbf"); | ||
955 | /* clear outbound buffers to free skbs */ | ||
956 | for (i = 0; i < card->qdio.no_out_queues; ++i) | ||
957 | if (card->qdio.out_qs[i]) { | ||
958 | for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j) | ||
959 | qeth_clear_output_buffer(card->qdio.out_qs[i], | ||
960 | &card->qdio.out_qs[i]->bufs[j]); | ||
961 | } | ||
962 | } | ||
963 | EXPORT_SYMBOL_GPL(qeth_clear_qdio_buffers); | ||
964 | |||
965 | static void qeth_free_buffer_pool(struct qeth_card *card) | ||
966 | { | ||
967 | struct qeth_buffer_pool_entry *pool_entry, *tmp; | ||
968 | int i = 0; | ||
969 | QETH_DBF_TEXT(TRACE, 5, "freepool"); | ||
970 | list_for_each_entry_safe(pool_entry, tmp, | ||
971 | &card->qdio.init_pool.entry_list, init_list){ | ||
972 | for (i = 0; i < QETH_MAX_BUFFER_ELEMENTS(card); ++i) | ||
973 | free_page((unsigned long)pool_entry->elements[i]); | ||
974 | list_del(&pool_entry->init_list); | ||
975 | kfree(pool_entry); | ||
976 | } | ||
977 | } | ||
978 | |||
979 | static void qeth_free_qdio_buffers(struct qeth_card *card) | ||
980 | { | ||
981 | int i, j; | ||
982 | |||
983 | QETH_DBF_TEXT(TRACE, 2, "freeqdbf"); | ||
984 | if (atomic_xchg(&card->qdio.state, QETH_QDIO_UNINITIALIZED) == | ||
985 | QETH_QDIO_UNINITIALIZED) | ||
986 | return; | ||
987 | kfree(card->qdio.in_q); | ||
988 | card->qdio.in_q = NULL; | ||
989 | /* inbound buffer pool */ | ||
990 | qeth_free_buffer_pool(card); | ||
991 | /* free outbound qdio_qs */ | ||
992 | if (card->qdio.out_qs) { | ||
993 | for (i = 0; i < card->qdio.no_out_queues; ++i) { | ||
994 | for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j) | ||
995 | qeth_clear_output_buffer(card->qdio.out_qs[i], | ||
996 | &card->qdio.out_qs[i]->bufs[j]); | ||
997 | kfree(card->qdio.out_qs[i]); | ||
998 | } | ||
999 | kfree(card->qdio.out_qs); | ||
1000 | card->qdio.out_qs = NULL; | ||
1001 | } | ||
1002 | } | ||
1003 | |||
1004 | static void qeth_clean_channel(struct qeth_channel *channel) | ||
1005 | { | ||
1006 | int cnt; | ||
1007 | |||
1008 | QETH_DBF_TEXT(SETUP, 2, "freech"); | ||
1009 | for (cnt = 0; cnt < QETH_CMD_BUFFER_NO; cnt++) | ||
1010 | kfree(channel->iob[cnt].data); | ||
1011 | } | ||
1012 | |||
1013 | static int qeth_is_1920_device(struct qeth_card *card) | ||
1014 | { | ||
1015 | int single_queue = 0; | ||
1016 | struct ccw_device *ccwdev; | ||
1017 | struct channelPath_dsc { | ||
1018 | u8 flags; | ||
1019 | u8 lsn; | ||
1020 | u8 desc; | ||
1021 | u8 chpid; | ||
1022 | u8 swla; | ||
1023 | u8 zeroes; | ||
1024 | u8 chla; | ||
1025 | u8 chpp; | ||
1026 | } *chp_dsc; | ||
1027 | |||
1028 | QETH_DBF_TEXT(SETUP, 2, "chk_1920"); | ||
1029 | |||
1030 | ccwdev = card->data.ccwdev; | ||
1031 | chp_dsc = (struct channelPath_dsc *)ccw_device_get_chp_desc(ccwdev, 0); | ||
1032 | if (chp_dsc != NULL) { | ||
1033 | /* CHPP field bit 6 == 1 -> single queue */ | ||
1034 | single_queue = ((chp_dsc->chpp & 0x02) == 0x02); | ||
1035 | kfree(chp_dsc); | ||
1036 | } | ||
1037 | QETH_DBF_TEXT_(SETUP, 2, "rc:%x", single_queue); | ||
1038 | return single_queue; | ||
1039 | } | ||
1040 | |||
1041 | static void qeth_init_qdio_info(struct qeth_card *card) | ||
1042 | { | ||
1043 | QETH_DBF_TEXT(SETUP, 4, "intqdinf"); | ||
1044 | atomic_set(&card->qdio.state, QETH_QDIO_UNINITIALIZED); | ||
1045 | /* inbound */ | ||
1046 | card->qdio.in_buf_size = QETH_IN_BUF_SIZE_DEFAULT; | ||
1047 | card->qdio.init_pool.buf_count = QETH_IN_BUF_COUNT_DEFAULT; | ||
1048 | card->qdio.in_buf_pool.buf_count = card->qdio.init_pool.buf_count; | ||
1049 | INIT_LIST_HEAD(&card->qdio.in_buf_pool.entry_list); | ||
1050 | INIT_LIST_HEAD(&card->qdio.init_pool.entry_list); | ||
1051 | } | ||
1052 | |||
1053 | static void qeth_set_intial_options(struct qeth_card *card) | ||
1054 | { | ||
1055 | card->options.route4.type = NO_ROUTER; | ||
1056 | card->options.route6.type = NO_ROUTER; | ||
1057 | card->options.checksum_type = QETH_CHECKSUM_DEFAULT; | ||
1058 | card->options.broadcast_mode = QETH_TR_BROADCAST_ALLRINGS; | ||
1059 | card->options.macaddr_mode = QETH_TR_MACADDR_NONCANONICAL; | ||
1060 | card->options.fake_broadcast = 0; | ||
1061 | card->options.add_hhlen = DEFAULT_ADD_HHLEN; | ||
1062 | card->options.fake_ll = 0; | ||
1063 | card->options.performance_stats = 0; | ||
1064 | card->options.rx_sg_cb = QETH_RX_SG_CB; | ||
1065 | } | ||
1066 | |||
1067 | static int qeth_do_start_thread(struct qeth_card *card, unsigned long thread) | ||
1068 | { | ||
1069 | unsigned long flags; | ||
1070 | int rc = 0; | ||
1071 | |||
1072 | spin_lock_irqsave(&card->thread_mask_lock, flags); | ||
1073 | QETH_DBF_TEXT_(TRACE, 4, " %02x%02x%02x", | ||
1074 | (u8) card->thread_start_mask, | ||
1075 | (u8) card->thread_allowed_mask, | ||
1076 | (u8) card->thread_running_mask); | ||
1077 | rc = (card->thread_start_mask & thread); | ||
1078 | spin_unlock_irqrestore(&card->thread_mask_lock, flags); | ||
1079 | return rc; | ||
1080 | } | ||
1081 | |||
1082 | static void qeth_start_kernel_thread(struct work_struct *work) | ||
1083 | { | ||
1084 | struct qeth_card *card = container_of(work, struct qeth_card, | ||
1085 | kernel_thread_starter); | ||
1086 | QETH_DBF_TEXT(TRACE , 2, "strthrd"); | ||
1087 | |||
1088 | if (card->read.state != CH_STATE_UP && | ||
1089 | card->write.state != CH_STATE_UP) | ||
1090 | return; | ||
1091 | if (qeth_do_start_thread(card, QETH_RECOVER_THREAD)) | ||
1092 | kthread_run(card->discipline.recover, (void *) card, | ||
1093 | "qeth_recover"); | ||
1094 | } | ||
1095 | |||
1096 | static int qeth_setup_card(struct qeth_card *card) | ||
1097 | { | ||
1098 | |||
1099 | QETH_DBF_TEXT(SETUP, 2, "setupcrd"); | ||
1100 | QETH_DBF_HEX(SETUP, 2, &card, sizeof(void *)); | ||
1101 | |||
1102 | card->read.state = CH_STATE_DOWN; | ||
1103 | card->write.state = CH_STATE_DOWN; | ||
1104 | card->data.state = CH_STATE_DOWN; | ||
1105 | card->state = CARD_STATE_DOWN; | ||
1106 | card->lan_online = 0; | ||
1107 | card->use_hard_stop = 0; | ||
1108 | card->dev = NULL; | ||
1109 | spin_lock_init(&card->vlanlock); | ||
1110 | spin_lock_init(&card->mclock); | ||
1111 | card->vlangrp = NULL; | ||
1112 | spin_lock_init(&card->lock); | ||
1113 | spin_lock_init(&card->ip_lock); | ||
1114 | spin_lock_init(&card->thread_mask_lock); | ||
1115 | card->thread_start_mask = 0; | ||
1116 | card->thread_allowed_mask = 0; | ||
1117 | card->thread_running_mask = 0; | ||
1118 | INIT_WORK(&card->kernel_thread_starter, qeth_start_kernel_thread); | ||
1119 | INIT_LIST_HEAD(&card->ip_list); | ||
1120 | card->ip_tbd_list = kmalloc(sizeof(struct list_head), GFP_KERNEL); | ||
1121 | if (!card->ip_tbd_list) { | ||
1122 | QETH_DBF_TEXT(SETUP, 0, "iptbdnom"); | ||
1123 | return -ENOMEM; | ||
1124 | } | ||
1125 | INIT_LIST_HEAD(card->ip_tbd_list); | ||
1126 | INIT_LIST_HEAD(&card->cmd_waiter_list); | ||
1127 | init_waitqueue_head(&card->wait_q); | ||
1128 | /* intial options */ | ||
1129 | qeth_set_intial_options(card); | ||
1130 | /* IP address takeover */ | ||
1131 | INIT_LIST_HEAD(&card->ipato.entries); | ||
1132 | card->ipato.enabled = 0; | ||
1133 | card->ipato.invert4 = 0; | ||
1134 | card->ipato.invert6 = 0; | ||
1135 | /* init QDIO stuff */ | ||
1136 | qeth_init_qdio_info(card); | ||
1137 | return 0; | ||
1138 | } | ||
1139 | |||
1140 | static struct qeth_card *qeth_alloc_card(void) | ||
1141 | { | ||
1142 | struct qeth_card *card; | ||
1143 | |||
1144 | QETH_DBF_TEXT(SETUP, 2, "alloccrd"); | ||
1145 | card = kzalloc(sizeof(struct qeth_card), GFP_DMA|GFP_KERNEL); | ||
1146 | if (!card) | ||
1147 | return NULL; | ||
1148 | QETH_DBF_HEX(SETUP, 2, &card, sizeof(void *)); | ||
1149 | if (qeth_setup_channel(&card->read)) { | ||
1150 | kfree(card); | ||
1151 | return NULL; | ||
1152 | } | ||
1153 | if (qeth_setup_channel(&card->write)) { | ||
1154 | qeth_clean_channel(&card->read); | ||
1155 | kfree(card); | ||
1156 | return NULL; | ||
1157 | } | ||
1158 | card->options.layer2 = -1; | ||
1159 | return card; | ||
1160 | } | ||
1161 | |||
1162 | static int qeth_determine_card_type(struct qeth_card *card) | ||
1163 | { | ||
1164 | int i = 0; | ||
1165 | |||
1166 | QETH_DBF_TEXT(SETUP, 2, "detcdtyp"); | ||
1167 | |||
1168 | card->qdio.do_prio_queueing = QETH_PRIOQ_DEFAULT; | ||
1169 | card->qdio.default_out_queue = QETH_DEFAULT_QUEUE; | ||
1170 | while (known_devices[i][4]) { | ||
1171 | if ((CARD_RDEV(card)->id.dev_type == known_devices[i][2]) && | ||
1172 | (CARD_RDEV(card)->id.dev_model == known_devices[i][3])) { | ||
1173 | card->info.type = known_devices[i][4]; | ||
1174 | card->qdio.no_out_queues = known_devices[i][8]; | ||
1175 | card->info.is_multicast_different = known_devices[i][9]; | ||
1176 | if (qeth_is_1920_device(card)) { | ||
1177 | PRINT_INFO("Priority Queueing not able " | ||
1178 | "due to hardware limitations!\n"); | ||
1179 | card->qdio.no_out_queues = 1; | ||
1180 | card->qdio.default_out_queue = 0; | ||
1181 | } | ||
1182 | return 0; | ||
1183 | } | ||
1184 | i++; | ||
1185 | } | ||
1186 | card->info.type = QETH_CARD_TYPE_UNKNOWN; | ||
1187 | PRINT_ERR("unknown card type on device %s\n", CARD_BUS_ID(card)); | ||
1188 | return -ENOENT; | ||
1189 | } | ||
1190 | |||
1191 | static int qeth_clear_channel(struct qeth_channel *channel) | ||
1192 | { | ||
1193 | unsigned long flags; | ||
1194 | struct qeth_card *card; | ||
1195 | int rc; | ||
1196 | |||
1197 | QETH_DBF_TEXT(TRACE, 3, "clearch"); | ||
1198 | card = CARD_FROM_CDEV(channel->ccwdev); | ||
1199 | spin_lock_irqsave(get_ccwdev_lock(channel->ccwdev), flags); | ||
1200 | rc = ccw_device_clear(channel->ccwdev, QETH_CLEAR_CHANNEL_PARM); | ||
1201 | spin_unlock_irqrestore(get_ccwdev_lock(channel->ccwdev), flags); | ||
1202 | |||
1203 | if (rc) | ||
1204 | return rc; | ||
1205 | rc = wait_event_interruptible_timeout(card->wait_q, | ||
1206 | channel->state == CH_STATE_STOPPED, QETH_TIMEOUT); | ||
1207 | if (rc == -ERESTARTSYS) | ||
1208 | return rc; | ||
1209 | if (channel->state != CH_STATE_STOPPED) | ||
1210 | return -ETIME; | ||
1211 | channel->state = CH_STATE_DOWN; | ||
1212 | return 0; | ||
1213 | } | ||
1214 | |||
1215 | static int qeth_halt_channel(struct qeth_channel *channel) | ||
1216 | { | ||
1217 | unsigned long flags; | ||
1218 | struct qeth_card *card; | ||
1219 | int rc; | ||
1220 | |||
1221 | QETH_DBF_TEXT(TRACE, 3, "haltch"); | ||
1222 | card = CARD_FROM_CDEV(channel->ccwdev); | ||
1223 | spin_lock_irqsave(get_ccwdev_lock(channel->ccwdev), flags); | ||
1224 | rc = ccw_device_halt(channel->ccwdev, QETH_HALT_CHANNEL_PARM); | ||
1225 | spin_unlock_irqrestore(get_ccwdev_lock(channel->ccwdev), flags); | ||
1226 | |||
1227 | if (rc) | ||
1228 | return rc; | ||
1229 | rc = wait_event_interruptible_timeout(card->wait_q, | ||
1230 | channel->state == CH_STATE_HALTED, QETH_TIMEOUT); | ||
1231 | if (rc == -ERESTARTSYS) | ||
1232 | return rc; | ||
1233 | if (channel->state != CH_STATE_HALTED) | ||
1234 | return -ETIME; | ||
1235 | return 0; | ||
1236 | } | ||
1237 | |||
1238 | static int qeth_halt_channels(struct qeth_card *card) | ||
1239 | { | ||
1240 | int rc1 = 0, rc2 = 0, rc3 = 0; | ||
1241 | |||
1242 | QETH_DBF_TEXT(TRACE, 3, "haltchs"); | ||
1243 | rc1 = qeth_halt_channel(&card->read); | ||
1244 | rc2 = qeth_halt_channel(&card->write); | ||
1245 | rc3 = qeth_halt_channel(&card->data); | ||
1246 | if (rc1) | ||
1247 | return rc1; | ||
1248 | if (rc2) | ||
1249 | return rc2; | ||
1250 | return rc3; | ||
1251 | } | ||
1252 | |||
1253 | static int qeth_clear_channels(struct qeth_card *card) | ||
1254 | { | ||
1255 | int rc1 = 0, rc2 = 0, rc3 = 0; | ||
1256 | |||
1257 | QETH_DBF_TEXT(TRACE, 3, "clearchs"); | ||
1258 | rc1 = qeth_clear_channel(&card->read); | ||
1259 | rc2 = qeth_clear_channel(&card->write); | ||
1260 | rc3 = qeth_clear_channel(&card->data); | ||
1261 | if (rc1) | ||
1262 | return rc1; | ||
1263 | if (rc2) | ||
1264 | return rc2; | ||
1265 | return rc3; | ||
1266 | } | ||
1267 | |||
1268 | static int qeth_clear_halt_card(struct qeth_card *card, int halt) | ||
1269 | { | ||
1270 | int rc = 0; | ||
1271 | |||
1272 | QETH_DBF_TEXT(TRACE, 3, "clhacrd"); | ||
1273 | QETH_DBF_HEX(TRACE, 3, &card, sizeof(void *)); | ||
1274 | |||
1275 | if (halt) | ||
1276 | rc = qeth_halt_channels(card); | ||
1277 | if (rc) | ||
1278 | return rc; | ||
1279 | return qeth_clear_channels(card); | ||
1280 | } | ||
1281 | |||
1282 | int qeth_qdio_clear_card(struct qeth_card *card, int use_halt) | ||
1283 | { | ||
1284 | int rc = 0; | ||
1285 | |||
1286 | QETH_DBF_TEXT(TRACE, 3, "qdioclr"); | ||
1287 | switch (atomic_cmpxchg(&card->qdio.state, QETH_QDIO_ESTABLISHED, | ||
1288 | QETH_QDIO_CLEANING)) { | ||
1289 | case QETH_QDIO_ESTABLISHED: | ||
1290 | if (card->info.type == QETH_CARD_TYPE_IQD) | ||
1291 | rc = qdio_cleanup(CARD_DDEV(card), | ||
1292 | QDIO_FLAG_CLEANUP_USING_HALT); | ||
1293 | else | ||
1294 | rc = qdio_cleanup(CARD_DDEV(card), | ||
1295 | QDIO_FLAG_CLEANUP_USING_CLEAR); | ||
1296 | if (rc) | ||
1297 | QETH_DBF_TEXT_(TRACE, 3, "1err%d", rc); | ||
1298 | atomic_set(&card->qdio.state, QETH_QDIO_ALLOCATED); | ||
1299 | break; | ||
1300 | case QETH_QDIO_CLEANING: | ||
1301 | return rc; | ||
1302 | default: | ||
1303 | break; | ||
1304 | } | ||
1305 | rc = qeth_clear_halt_card(card, use_halt); | ||
1306 | if (rc) | ||
1307 | QETH_DBF_TEXT_(TRACE, 3, "2err%d", rc); | ||
1308 | card->state = CARD_STATE_DOWN; | ||
1309 | return rc; | ||
1310 | } | ||
1311 | EXPORT_SYMBOL_GPL(qeth_qdio_clear_card); | ||
1312 | |||
1313 | static int qeth_read_conf_data(struct qeth_card *card, void **buffer, | ||
1314 | int *length) | ||
1315 | { | ||
1316 | struct ciw *ciw; | ||
1317 | char *rcd_buf; | ||
1318 | int ret; | ||
1319 | struct qeth_channel *channel = &card->data; | ||
1320 | unsigned long flags; | ||
1321 | |||
1322 | /* | ||
1323 | * scan for RCD command in extended SenseID data | ||
1324 | */ | ||
1325 | ciw = ccw_device_get_ciw(channel->ccwdev, CIW_TYPE_RCD); | ||
1326 | if (!ciw || ciw->cmd == 0) | ||
1327 | return -EOPNOTSUPP; | ||
1328 | rcd_buf = kzalloc(ciw->count, GFP_KERNEL | GFP_DMA); | ||
1329 | if (!rcd_buf) | ||
1330 | return -ENOMEM; | ||
1331 | |||
1332 | channel->ccw.cmd_code = ciw->cmd; | ||
1333 | channel->ccw.cda = (__u32) __pa(rcd_buf); | ||
1334 | channel->ccw.count = ciw->count; | ||
1335 | channel->ccw.flags = CCW_FLAG_SLI; | ||
1336 | channel->state = CH_STATE_RCD; | ||
1337 | spin_lock_irqsave(get_ccwdev_lock(channel->ccwdev), flags); | ||
1338 | ret = ccw_device_start_timeout(channel->ccwdev, &channel->ccw, | ||
1339 | QETH_RCD_PARM, LPM_ANYPATH, 0, | ||
1340 | QETH_RCD_TIMEOUT); | ||
1341 | spin_unlock_irqrestore(get_ccwdev_lock(channel->ccwdev), flags); | ||
1342 | if (!ret) | ||
1343 | wait_event(card->wait_q, | ||
1344 | (channel->state == CH_STATE_RCD_DONE || | ||
1345 | channel->state == CH_STATE_DOWN)); | ||
1346 | if (channel->state == CH_STATE_DOWN) | ||
1347 | ret = -EIO; | ||
1348 | else | ||
1349 | channel->state = CH_STATE_DOWN; | ||
1350 | if (ret) { | ||
1351 | kfree(rcd_buf); | ||
1352 | *buffer = NULL; | ||
1353 | *length = 0; | ||
1354 | } else { | ||
1355 | *length = ciw->count; | ||
1356 | *buffer = rcd_buf; | ||
1357 | } | ||
1358 | return ret; | ||
1359 | } | ||
1360 | |||
1361 | static int qeth_get_unitaddr(struct qeth_card *card) | ||
1362 | { | ||
1363 | int length; | ||
1364 | char *prcd; | ||
1365 | int rc; | ||
1366 | |||
1367 | QETH_DBF_TEXT(SETUP, 2, "getunit"); | ||
1368 | rc = qeth_read_conf_data(card, (void **) &prcd, &length); | ||
1369 | if (rc) { | ||
1370 | PRINT_ERR("qeth_read_conf_data for device %s returned %i\n", | ||
1371 | CARD_DDEV_ID(card), rc); | ||
1372 | return rc; | ||
1373 | } | ||
1374 | card->info.chpid = prcd[30]; | ||
1375 | card->info.unit_addr2 = prcd[31]; | ||
1376 | card->info.cula = prcd[63]; | ||
1377 | card->info.guestlan = ((prcd[0x10] == _ascebc['V']) && | ||
1378 | (prcd[0x11] == _ascebc['M'])); | ||
1379 | kfree(prcd); | ||
1380 | return 0; | ||
1381 | } | ||
1382 | |||
1383 | static void qeth_init_tokens(struct qeth_card *card) | ||
1384 | { | ||
1385 | card->token.issuer_rm_w = 0x00010103UL; | ||
1386 | card->token.cm_filter_w = 0x00010108UL; | ||
1387 | card->token.cm_connection_w = 0x0001010aUL; | ||
1388 | card->token.ulp_filter_w = 0x0001010bUL; | ||
1389 | card->token.ulp_connection_w = 0x0001010dUL; | ||
1390 | } | ||
1391 | |||
1392 | static void qeth_init_func_level(struct qeth_card *card) | ||
1393 | { | ||
1394 | if (card->ipato.enabled) { | ||
1395 | if (card->info.type == QETH_CARD_TYPE_IQD) | ||
1396 | card->info.func_level = | ||
1397 | QETH_IDX_FUNC_LEVEL_IQD_ENA_IPAT; | ||
1398 | else | ||
1399 | card->info.func_level = | ||
1400 | QETH_IDX_FUNC_LEVEL_OSAE_ENA_IPAT; | ||
1401 | } else { | ||
1402 | if (card->info.type == QETH_CARD_TYPE_IQD) | ||
1403 | /*FIXME:why do we have same values for dis and ena for | ||
1404 | osae??? */ | ||
1405 | card->info.func_level = | ||
1406 | QETH_IDX_FUNC_LEVEL_IQD_DIS_IPAT; | ||
1407 | else | ||
1408 | card->info.func_level = | ||
1409 | QETH_IDX_FUNC_LEVEL_OSAE_DIS_IPAT; | ||
1410 | } | ||
1411 | } | ||
1412 | |||
1413 | static inline __u16 qeth_raw_devno_from_bus_id(char *id) | ||
1414 | { | ||
1415 | id += (strlen(id) - 4); | ||
1416 | return (__u16) simple_strtoul(id, &id, 16); | ||
1417 | } | ||
1418 | |||
1419 | static int qeth_idx_activate_get_answer(struct qeth_channel *channel, | ||
1420 | void (*idx_reply_cb)(struct qeth_channel *, | ||
1421 | struct qeth_cmd_buffer *)) | ||
1422 | { | ||
1423 | struct qeth_cmd_buffer *iob; | ||
1424 | unsigned long flags; | ||
1425 | int rc; | ||
1426 | struct qeth_card *card; | ||
1427 | |||
1428 | QETH_DBF_TEXT(SETUP, 2, "idxanswr"); | ||
1429 | card = CARD_FROM_CDEV(channel->ccwdev); | ||
1430 | iob = qeth_get_buffer(channel); | ||
1431 | iob->callback = idx_reply_cb; | ||
1432 | memcpy(&channel->ccw, READ_CCW, sizeof(struct ccw1)); | ||
1433 | channel->ccw.count = QETH_BUFSIZE; | ||
1434 | channel->ccw.cda = (__u32) __pa(iob->data); | ||
1435 | |||
1436 | wait_event(card->wait_q, | ||
1437 | atomic_cmpxchg(&channel->irq_pending, 0, 1) == 0); | ||
1438 | QETH_DBF_TEXT(SETUP, 6, "noirqpnd"); | ||
1439 | spin_lock_irqsave(get_ccwdev_lock(channel->ccwdev), flags); | ||
1440 | rc = ccw_device_start(channel->ccwdev, | ||
1441 | &channel->ccw, (addr_t) iob, 0, 0); | ||
1442 | spin_unlock_irqrestore(get_ccwdev_lock(channel->ccwdev), flags); | ||
1443 | |||
1444 | if (rc) { | ||
1445 | PRINT_ERR("Error2 in activating channel rc=%d\n", rc); | ||
1446 | QETH_DBF_TEXT_(SETUP, 2, "2err%d", rc); | ||
1447 | atomic_set(&channel->irq_pending, 0); | ||
1448 | wake_up(&card->wait_q); | ||
1449 | return rc; | ||
1450 | } | ||
1451 | rc = wait_event_interruptible_timeout(card->wait_q, | ||
1452 | channel->state == CH_STATE_UP, QETH_TIMEOUT); | ||
1453 | if (rc == -ERESTARTSYS) | ||
1454 | return rc; | ||
1455 | if (channel->state != CH_STATE_UP) { | ||
1456 | rc = -ETIME; | ||
1457 | QETH_DBF_TEXT_(SETUP, 2, "3err%d", rc); | ||
1458 | qeth_clear_cmd_buffers(channel); | ||
1459 | } else | ||
1460 | rc = 0; | ||
1461 | return rc; | ||
1462 | } | ||
1463 | |||
1464 | static int qeth_idx_activate_channel(struct qeth_channel *channel, | ||
1465 | void (*idx_reply_cb)(struct qeth_channel *, | ||
1466 | struct qeth_cmd_buffer *)) | ||
1467 | { | ||
1468 | struct qeth_card *card; | ||
1469 | struct qeth_cmd_buffer *iob; | ||
1470 | unsigned long flags; | ||
1471 | __u16 temp; | ||
1472 | __u8 tmp; | ||
1473 | int rc; | ||
1474 | |||
1475 | card = CARD_FROM_CDEV(channel->ccwdev); | ||
1476 | |||
1477 | QETH_DBF_TEXT(SETUP, 2, "idxactch"); | ||
1478 | |||
1479 | iob = qeth_get_buffer(channel); | ||
1480 | iob->callback = idx_reply_cb; | ||
1481 | memcpy(&channel->ccw, WRITE_CCW, sizeof(struct ccw1)); | ||
1482 | channel->ccw.count = IDX_ACTIVATE_SIZE; | ||
1483 | channel->ccw.cda = (__u32) __pa(iob->data); | ||
1484 | if (channel == &card->write) { | ||
1485 | memcpy(iob->data, IDX_ACTIVATE_WRITE, IDX_ACTIVATE_SIZE); | ||
1486 | memcpy(QETH_TRANSPORT_HEADER_SEQ_NO(iob->data), | ||
1487 | &card->seqno.trans_hdr, QETH_SEQ_NO_LENGTH); | ||
1488 | card->seqno.trans_hdr++; | ||
1489 | } else { | ||
1490 | memcpy(iob->data, IDX_ACTIVATE_READ, IDX_ACTIVATE_SIZE); | ||
1491 | memcpy(QETH_TRANSPORT_HEADER_SEQ_NO(iob->data), | ||
1492 | &card->seqno.trans_hdr, QETH_SEQ_NO_LENGTH); | ||
1493 | } | ||
1494 | tmp = ((__u8)card->info.portno) | 0x80; | ||
1495 | memcpy(QETH_IDX_ACT_PNO(iob->data), &tmp, 1); | ||
1496 | memcpy(QETH_IDX_ACT_ISSUER_RM_TOKEN(iob->data), | ||
1497 | &card->token.issuer_rm_w, QETH_MPC_TOKEN_LENGTH); | ||
1498 | memcpy(QETH_IDX_ACT_FUNC_LEVEL(iob->data), | ||
1499 | &card->info.func_level, sizeof(__u16)); | ||
1500 | temp = qeth_raw_devno_from_bus_id(CARD_DDEV_ID(card)); | ||
1501 | memcpy(QETH_IDX_ACT_QDIO_DEV_CUA(iob->data), &temp, 2); | ||
1502 | temp = (card->info.cula << 8) + card->info.unit_addr2; | ||
1503 | memcpy(QETH_IDX_ACT_QDIO_DEV_REALADDR(iob->data), &temp, 2); | ||
1504 | |||
1505 | wait_event(card->wait_q, | ||
1506 | atomic_cmpxchg(&channel->irq_pending, 0, 1) == 0); | ||
1507 | QETH_DBF_TEXT(SETUP, 6, "noirqpnd"); | ||
1508 | spin_lock_irqsave(get_ccwdev_lock(channel->ccwdev), flags); | ||
1509 | rc = ccw_device_start(channel->ccwdev, | ||
1510 | &channel->ccw, (addr_t) iob, 0, 0); | ||
1511 | spin_unlock_irqrestore(get_ccwdev_lock(channel->ccwdev), flags); | ||
1512 | |||
1513 | if (rc) { | ||
1514 | PRINT_ERR("Error1 in activating channel. rc=%d\n", rc); | ||
1515 | QETH_DBF_TEXT_(SETUP, 2, "1err%d", rc); | ||
1516 | atomic_set(&channel->irq_pending, 0); | ||
1517 | wake_up(&card->wait_q); | ||
1518 | return rc; | ||
1519 | } | ||
1520 | rc = wait_event_interruptible_timeout(card->wait_q, | ||
1521 | channel->state == CH_STATE_ACTIVATING, QETH_TIMEOUT); | ||
1522 | if (rc == -ERESTARTSYS) | ||
1523 | return rc; | ||
1524 | if (channel->state != CH_STATE_ACTIVATING) { | ||
1525 | PRINT_WARN("IDX activate timed out!\n"); | ||
1526 | QETH_DBF_TEXT_(SETUP, 2, "2err%d", -ETIME); | ||
1527 | qeth_clear_cmd_buffers(channel); | ||
1528 | return -ETIME; | ||
1529 | } | ||
1530 | return qeth_idx_activate_get_answer(channel, idx_reply_cb); | ||
1531 | } | ||
1532 | |||
1533 | static int qeth_peer_func_level(int level) | ||
1534 | { | ||
1535 | if ((level & 0xff) == 8) | ||
1536 | return (level & 0xff) + 0x400; | ||
1537 | if (((level >> 8) & 3) == 1) | ||
1538 | return (level & 0xff) + 0x200; | ||
1539 | return level; | ||
1540 | } | ||
1541 | |||
1542 | static void qeth_idx_write_cb(struct qeth_channel *channel, | ||
1543 | struct qeth_cmd_buffer *iob) | ||
1544 | { | ||
1545 | struct qeth_card *card; | ||
1546 | __u16 temp; | ||
1547 | |||
1548 | QETH_DBF_TEXT(SETUP , 2, "idxwrcb"); | ||
1549 | |||
1550 | if (channel->state == CH_STATE_DOWN) { | ||
1551 | channel->state = CH_STATE_ACTIVATING; | ||
1552 | goto out; | ||
1553 | } | ||
1554 | card = CARD_FROM_CDEV(channel->ccwdev); | ||
1555 | |||
1556 | if (!(QETH_IS_IDX_ACT_POS_REPLY(iob->data))) { | ||
1557 | if (QETH_IDX_ACT_CAUSE_CODE(iob->data) == 0x19) | ||
1558 | PRINT_ERR("IDX_ACTIVATE on write channel device %s: " | ||
1559 | "adapter exclusively used by another host\n", | ||
1560 | CARD_WDEV_ID(card)); | ||
1561 | else | ||
1562 | PRINT_ERR("IDX_ACTIVATE on write channel device %s: " | ||
1563 | "negative reply\n", CARD_WDEV_ID(card)); | ||
1564 | goto out; | ||
1565 | } | ||
1566 | memcpy(&temp, QETH_IDX_ACT_FUNC_LEVEL(iob->data), 2); | ||
1567 | if ((temp & ~0x0100) != qeth_peer_func_level(card->info.func_level)) { | ||
1568 | PRINT_WARN("IDX_ACTIVATE on write channel device %s: " | ||
1569 | "function level mismatch " | ||
1570 | "(sent: 0x%x, received: 0x%x)\n", | ||
1571 | CARD_WDEV_ID(card), card->info.func_level, temp); | ||
1572 | goto out; | ||
1573 | } | ||
1574 | channel->state = CH_STATE_UP; | ||
1575 | out: | ||
1576 | qeth_release_buffer(channel, iob); | ||
1577 | } | ||
1578 | |||
1579 | static void qeth_idx_read_cb(struct qeth_channel *channel, | ||
1580 | struct qeth_cmd_buffer *iob) | ||
1581 | { | ||
1582 | struct qeth_card *card; | ||
1583 | __u16 temp; | ||
1584 | |||
1585 | QETH_DBF_TEXT(SETUP , 2, "idxrdcb"); | ||
1586 | if (channel->state == CH_STATE_DOWN) { | ||
1587 | channel->state = CH_STATE_ACTIVATING; | ||
1588 | goto out; | ||
1589 | } | ||
1590 | |||
1591 | card = CARD_FROM_CDEV(channel->ccwdev); | ||
1592 | if (qeth_check_idx_response(iob->data)) | ||
1593 | goto out; | ||
1594 | |||
1595 | if (!(QETH_IS_IDX_ACT_POS_REPLY(iob->data))) { | ||
1596 | if (QETH_IDX_ACT_CAUSE_CODE(iob->data) == 0x19) | ||
1597 | PRINT_ERR("IDX_ACTIVATE on read channel device %s: " | ||
1598 | "adapter exclusively used by another host\n", | ||
1599 | CARD_RDEV_ID(card)); | ||
1600 | else | ||
1601 | PRINT_ERR("IDX_ACTIVATE on read channel device %s: " | ||
1602 | "negative reply\n", CARD_RDEV_ID(card)); | ||
1603 | goto out; | ||
1604 | } | ||
1605 | |||
1606 | /** | ||
1607 | * temporary fix for microcode bug | ||
1608 | * to revert it,replace OR by AND | ||
1609 | */ | ||
1610 | if ((!QETH_IDX_NO_PORTNAME_REQUIRED(iob->data)) || | ||
1611 | (card->info.type == QETH_CARD_TYPE_OSAE)) | ||
1612 | card->info.portname_required = 1; | ||
1613 | |||
1614 | memcpy(&temp, QETH_IDX_ACT_FUNC_LEVEL(iob->data), 2); | ||
1615 | if (temp != qeth_peer_func_level(card->info.func_level)) { | ||
1616 | PRINT_WARN("IDX_ACTIVATE on read channel device %s: function " | ||
1617 | "level mismatch (sent: 0x%x, received: 0x%x)\n", | ||
1618 | CARD_RDEV_ID(card), card->info.func_level, temp); | ||
1619 | goto out; | ||
1620 | } | ||
1621 | memcpy(&card->token.issuer_rm_r, | ||
1622 | QETH_IDX_ACT_ISSUER_RM_TOKEN(iob->data), | ||
1623 | QETH_MPC_TOKEN_LENGTH); | ||
1624 | memcpy(&card->info.mcl_level[0], | ||
1625 | QETH_IDX_REPLY_LEVEL(iob->data), QETH_MCL_LENGTH); | ||
1626 | channel->state = CH_STATE_UP; | ||
1627 | out: | ||
1628 | qeth_release_buffer(channel, iob); | ||
1629 | } | ||
1630 | |||
1631 | void qeth_prepare_control_data(struct qeth_card *card, int len, | ||
1632 | struct qeth_cmd_buffer *iob) | ||
1633 | { | ||
1634 | qeth_setup_ccw(&card->write, iob->data, len); | ||
1635 | iob->callback = qeth_release_buffer; | ||
1636 | |||
1637 | memcpy(QETH_TRANSPORT_HEADER_SEQ_NO(iob->data), | ||
1638 | &card->seqno.trans_hdr, QETH_SEQ_NO_LENGTH); | ||
1639 | card->seqno.trans_hdr++; | ||
1640 | memcpy(QETH_PDU_HEADER_SEQ_NO(iob->data), | ||
1641 | &card->seqno.pdu_hdr, QETH_SEQ_NO_LENGTH); | ||
1642 | card->seqno.pdu_hdr++; | ||
1643 | memcpy(QETH_PDU_HEADER_ACK_SEQ_NO(iob->data), | ||
1644 | &card->seqno.pdu_hdr_ack, QETH_SEQ_NO_LENGTH); | ||
1645 | QETH_DBF_HEX(CTRL, 2, iob->data, QETH_DBF_CTRL_LEN); | ||
1646 | } | ||
1647 | EXPORT_SYMBOL_GPL(qeth_prepare_control_data); | ||
1648 | |||
1649 | int qeth_send_control_data(struct qeth_card *card, int len, | ||
1650 | struct qeth_cmd_buffer *iob, | ||
1651 | int (*reply_cb)(struct qeth_card *, struct qeth_reply *, | ||
1652 | unsigned long), | ||
1653 | void *reply_param) | ||
1654 | { | ||
1655 | int rc; | ||
1656 | unsigned long flags; | ||
1657 | struct qeth_reply *reply = NULL; | ||
1658 | unsigned long timeout; | ||
1659 | |||
1660 | QETH_DBF_TEXT(TRACE, 2, "sendctl"); | ||
1661 | |||
1662 | reply = qeth_alloc_reply(card); | ||
1663 | if (!reply) { | ||
1664 | PRINT_WARN("Could not alloc qeth_reply!\n"); | ||
1665 | return -ENOMEM; | ||
1666 | } | ||
1667 | reply->callback = reply_cb; | ||
1668 | reply->param = reply_param; | ||
1669 | if (card->state == CARD_STATE_DOWN) | ||
1670 | reply->seqno = QETH_IDX_COMMAND_SEQNO; | ||
1671 | else | ||
1672 | reply->seqno = card->seqno.ipa++; | ||
1673 | init_waitqueue_head(&reply->wait_q); | ||
1674 | spin_lock_irqsave(&card->lock, flags); | ||
1675 | list_add_tail(&reply->list, &card->cmd_waiter_list); | ||
1676 | spin_unlock_irqrestore(&card->lock, flags); | ||
1677 | QETH_DBF_HEX(CTRL, 2, iob->data, QETH_DBF_CTRL_LEN); | ||
1678 | |||
1679 | while (atomic_cmpxchg(&card->write.irq_pending, 0, 1)) ; | ||
1680 | qeth_prepare_control_data(card, len, iob); | ||
1681 | |||
1682 | if (IS_IPA(iob->data)) | ||
1683 | timeout = jiffies + QETH_IPA_TIMEOUT; | ||
1684 | else | ||
1685 | timeout = jiffies + QETH_TIMEOUT; | ||
1686 | |||
1687 | QETH_DBF_TEXT(TRACE, 6, "noirqpnd"); | ||
1688 | spin_lock_irqsave(get_ccwdev_lock(card->write.ccwdev), flags); | ||
1689 | rc = ccw_device_start(card->write.ccwdev, &card->write.ccw, | ||
1690 | (addr_t) iob, 0, 0); | ||
1691 | spin_unlock_irqrestore(get_ccwdev_lock(card->write.ccwdev), flags); | ||
1692 | if (rc) { | ||
1693 | PRINT_WARN("qeth_send_control_data: " | ||
1694 | "ccw_device_start rc = %i\n", rc); | ||
1695 | QETH_DBF_TEXT_(TRACE, 2, " err%d", rc); | ||
1696 | spin_lock_irqsave(&card->lock, flags); | ||
1697 | list_del_init(&reply->list); | ||
1698 | qeth_put_reply(reply); | ||
1699 | spin_unlock_irqrestore(&card->lock, flags); | ||
1700 | qeth_release_buffer(iob->channel, iob); | ||
1701 | atomic_set(&card->write.irq_pending, 0); | ||
1702 | wake_up(&card->wait_q); | ||
1703 | return rc; | ||
1704 | } | ||
1705 | while (!atomic_read(&reply->received)) { | ||
1706 | if (time_after(jiffies, timeout)) { | ||
1707 | spin_lock_irqsave(&reply->card->lock, flags); | ||
1708 | list_del_init(&reply->list); | ||
1709 | spin_unlock_irqrestore(&reply->card->lock, flags); | ||
1710 | reply->rc = -ETIME; | ||
1711 | atomic_inc(&reply->received); | ||
1712 | wake_up(&reply->wait_q); | ||
1713 | } | ||
1714 | cpu_relax(); | ||
1715 | }; | ||
1716 | rc = reply->rc; | ||
1717 | qeth_put_reply(reply); | ||
1718 | return rc; | ||
1719 | } | ||
1720 | EXPORT_SYMBOL_GPL(qeth_send_control_data); | ||
1721 | |||
1722 | static int qeth_cm_enable_cb(struct qeth_card *card, struct qeth_reply *reply, | ||
1723 | unsigned long data) | ||
1724 | { | ||
1725 | struct qeth_cmd_buffer *iob; | ||
1726 | |||
1727 | QETH_DBF_TEXT(SETUP, 2, "cmenblcb"); | ||
1728 | |||
1729 | iob = (struct qeth_cmd_buffer *) data; | ||
1730 | memcpy(&card->token.cm_filter_r, | ||
1731 | QETH_CM_ENABLE_RESP_FILTER_TOKEN(iob->data), | ||
1732 | QETH_MPC_TOKEN_LENGTH); | ||
1733 | QETH_DBF_TEXT_(SETUP, 2, " rc%d", iob->rc); | ||
1734 | return 0; | ||
1735 | } | ||
1736 | |||
1737 | static int qeth_cm_enable(struct qeth_card *card) | ||
1738 | { | ||
1739 | int rc; | ||
1740 | struct qeth_cmd_buffer *iob; | ||
1741 | |||
1742 | QETH_DBF_TEXT(SETUP, 2, "cmenable"); | ||
1743 | |||
1744 | iob = qeth_wait_for_buffer(&card->write); | ||
1745 | memcpy(iob->data, CM_ENABLE, CM_ENABLE_SIZE); | ||
1746 | memcpy(QETH_CM_ENABLE_ISSUER_RM_TOKEN(iob->data), | ||
1747 | &card->token.issuer_rm_r, QETH_MPC_TOKEN_LENGTH); | ||
1748 | memcpy(QETH_CM_ENABLE_FILTER_TOKEN(iob->data), | ||
1749 | &card->token.cm_filter_w, QETH_MPC_TOKEN_LENGTH); | ||
1750 | |||
1751 | rc = qeth_send_control_data(card, CM_ENABLE_SIZE, iob, | ||
1752 | qeth_cm_enable_cb, NULL); | ||
1753 | return rc; | ||
1754 | } | ||
1755 | |||
1756 | static int qeth_cm_setup_cb(struct qeth_card *card, struct qeth_reply *reply, | ||
1757 | unsigned long data) | ||
1758 | { | ||
1759 | |||
1760 | struct qeth_cmd_buffer *iob; | ||
1761 | |||
1762 | QETH_DBF_TEXT(SETUP, 2, "cmsetpcb"); | ||
1763 | |||
1764 | iob = (struct qeth_cmd_buffer *) data; | ||
1765 | memcpy(&card->token.cm_connection_r, | ||
1766 | QETH_CM_SETUP_RESP_DEST_ADDR(iob->data), | ||
1767 | QETH_MPC_TOKEN_LENGTH); | ||
1768 | QETH_DBF_TEXT_(SETUP, 2, " rc%d", iob->rc); | ||
1769 | return 0; | ||
1770 | } | ||
1771 | |||
1772 | static int qeth_cm_setup(struct qeth_card *card) | ||
1773 | { | ||
1774 | int rc; | ||
1775 | struct qeth_cmd_buffer *iob; | ||
1776 | |||
1777 | QETH_DBF_TEXT(SETUP, 2, "cmsetup"); | ||
1778 | |||
1779 | iob = qeth_wait_for_buffer(&card->write); | ||
1780 | memcpy(iob->data, CM_SETUP, CM_SETUP_SIZE); | ||
1781 | memcpy(QETH_CM_SETUP_DEST_ADDR(iob->data), | ||
1782 | &card->token.issuer_rm_r, QETH_MPC_TOKEN_LENGTH); | ||
1783 | memcpy(QETH_CM_SETUP_CONNECTION_TOKEN(iob->data), | ||
1784 | &card->token.cm_connection_w, QETH_MPC_TOKEN_LENGTH); | ||
1785 | memcpy(QETH_CM_SETUP_FILTER_TOKEN(iob->data), | ||
1786 | &card->token.cm_filter_r, QETH_MPC_TOKEN_LENGTH); | ||
1787 | rc = qeth_send_control_data(card, CM_SETUP_SIZE, iob, | ||
1788 | qeth_cm_setup_cb, NULL); | ||
1789 | return rc; | ||
1790 | |||
1791 | } | ||
1792 | |||
1793 | static inline int qeth_get_initial_mtu_for_card(struct qeth_card *card) | ||
1794 | { | ||
1795 | switch (card->info.type) { | ||
1796 | case QETH_CARD_TYPE_UNKNOWN: | ||
1797 | return 1500; | ||
1798 | case QETH_CARD_TYPE_IQD: | ||
1799 | return card->info.max_mtu; | ||
1800 | case QETH_CARD_TYPE_OSAE: | ||
1801 | switch (card->info.link_type) { | ||
1802 | case QETH_LINK_TYPE_HSTR: | ||
1803 | case QETH_LINK_TYPE_LANE_TR: | ||
1804 | return 2000; | ||
1805 | default: | ||
1806 | return 1492; | ||
1807 | } | ||
1808 | default: | ||
1809 | return 1500; | ||
1810 | } | ||
1811 | } | ||
1812 | |||
1813 | static inline int qeth_get_max_mtu_for_card(int cardtype) | ||
1814 | { | ||
1815 | switch (cardtype) { | ||
1816 | |||
1817 | case QETH_CARD_TYPE_UNKNOWN: | ||
1818 | case QETH_CARD_TYPE_OSAE: | ||
1819 | case QETH_CARD_TYPE_OSN: | ||
1820 | return 61440; | ||
1821 | case QETH_CARD_TYPE_IQD: | ||
1822 | return 57344; | ||
1823 | default: | ||
1824 | return 1500; | ||
1825 | } | ||
1826 | } | ||
1827 | |||
1828 | static inline int qeth_get_mtu_out_of_mpc(int cardtype) | ||
1829 | { | ||
1830 | switch (cardtype) { | ||
1831 | case QETH_CARD_TYPE_IQD: | ||
1832 | return 1; | ||
1833 | default: | ||
1834 | return 0; | ||
1835 | } | ||
1836 | } | ||
1837 | |||
1838 | static inline int qeth_get_mtu_outof_framesize(int framesize) | ||
1839 | { | ||
1840 | switch (framesize) { | ||
1841 | case 0x4000: | ||
1842 | return 8192; | ||
1843 | case 0x6000: | ||
1844 | return 16384; | ||
1845 | case 0xa000: | ||
1846 | return 32768; | ||
1847 | case 0xffff: | ||
1848 | return 57344; | ||
1849 | default: | ||
1850 | return 0; | ||
1851 | } | ||
1852 | } | ||
1853 | |||
1854 | static inline int qeth_mtu_is_valid(struct qeth_card *card, int mtu) | ||
1855 | { | ||
1856 | switch (card->info.type) { | ||
1857 | case QETH_CARD_TYPE_OSAE: | ||
1858 | return ((mtu >= 576) && (mtu <= 61440)); | ||
1859 | case QETH_CARD_TYPE_IQD: | ||
1860 | return ((mtu >= 576) && | ||
1861 | (mtu <= card->info.max_mtu + 4096 - 32)); | ||
1862 | case QETH_CARD_TYPE_OSN: | ||
1863 | case QETH_CARD_TYPE_UNKNOWN: | ||
1864 | default: | ||
1865 | return 1; | ||
1866 | } | ||
1867 | } | ||
1868 | |||
1869 | static int qeth_ulp_enable_cb(struct qeth_card *card, struct qeth_reply *reply, | ||
1870 | unsigned long data) | ||
1871 | { | ||
1872 | |||
1873 | __u16 mtu, framesize; | ||
1874 | __u16 len; | ||
1875 | __u8 link_type; | ||
1876 | struct qeth_cmd_buffer *iob; | ||
1877 | |||
1878 | QETH_DBF_TEXT(SETUP, 2, "ulpenacb"); | ||
1879 | |||
1880 | iob = (struct qeth_cmd_buffer *) data; | ||
1881 | memcpy(&card->token.ulp_filter_r, | ||
1882 | QETH_ULP_ENABLE_RESP_FILTER_TOKEN(iob->data), | ||
1883 | QETH_MPC_TOKEN_LENGTH); | ||
1884 | if (qeth_get_mtu_out_of_mpc(card->info.type)) { | ||
1885 | memcpy(&framesize, QETH_ULP_ENABLE_RESP_MAX_MTU(iob->data), 2); | ||
1886 | mtu = qeth_get_mtu_outof_framesize(framesize); | ||
1887 | if (!mtu) { | ||
1888 | iob->rc = -EINVAL; | ||
1889 | QETH_DBF_TEXT_(SETUP, 2, " rc%d", iob->rc); | ||
1890 | return 0; | ||
1891 | } | ||
1892 | card->info.max_mtu = mtu; | ||
1893 | card->info.initial_mtu = mtu; | ||
1894 | card->qdio.in_buf_size = mtu + 2 * PAGE_SIZE; | ||
1895 | } else { | ||
1896 | card->info.initial_mtu = qeth_get_initial_mtu_for_card(card); | ||
1897 | card->info.max_mtu = qeth_get_max_mtu_for_card(card->info.type); | ||
1898 | card->qdio.in_buf_size = QETH_IN_BUF_SIZE_DEFAULT; | ||
1899 | } | ||
1900 | |||
1901 | memcpy(&len, QETH_ULP_ENABLE_RESP_DIFINFO_LEN(iob->data), 2); | ||
1902 | if (len >= QETH_MPC_DIFINFO_LEN_INDICATES_LINK_TYPE) { | ||
1903 | memcpy(&link_type, | ||
1904 | QETH_ULP_ENABLE_RESP_LINK_TYPE(iob->data), 1); | ||
1905 | card->info.link_type = link_type; | ||
1906 | } else | ||
1907 | card->info.link_type = 0; | ||
1908 | QETH_DBF_TEXT_(SETUP, 2, " rc%d", iob->rc); | ||
1909 | return 0; | ||
1910 | } | ||
1911 | |||
1912 | static int qeth_ulp_enable(struct qeth_card *card) | ||
1913 | { | ||
1914 | int rc; | ||
1915 | char prot_type; | ||
1916 | struct qeth_cmd_buffer *iob; | ||
1917 | |||
1918 | /*FIXME: trace view callbacks*/ | ||
1919 | QETH_DBF_TEXT(SETUP, 2, "ulpenabl"); | ||
1920 | |||
1921 | iob = qeth_wait_for_buffer(&card->write); | ||
1922 | memcpy(iob->data, ULP_ENABLE, ULP_ENABLE_SIZE); | ||
1923 | |||
1924 | *(QETH_ULP_ENABLE_LINKNUM(iob->data)) = | ||
1925 | (__u8) card->info.portno; | ||
1926 | if (card->options.layer2) | ||
1927 | if (card->info.type == QETH_CARD_TYPE_OSN) | ||
1928 | prot_type = QETH_PROT_OSN2; | ||
1929 | else | ||
1930 | prot_type = QETH_PROT_LAYER2; | ||
1931 | else | ||
1932 | prot_type = QETH_PROT_TCPIP; | ||
1933 | |||
1934 | memcpy(QETH_ULP_ENABLE_PROT_TYPE(iob->data), &prot_type, 1); | ||
1935 | memcpy(QETH_ULP_ENABLE_DEST_ADDR(iob->data), | ||
1936 | &card->token.cm_connection_r, QETH_MPC_TOKEN_LENGTH); | ||
1937 | memcpy(QETH_ULP_ENABLE_FILTER_TOKEN(iob->data), | ||
1938 | &card->token.ulp_filter_w, QETH_MPC_TOKEN_LENGTH); | ||
1939 | memcpy(QETH_ULP_ENABLE_PORTNAME_AND_LL(iob->data), | ||
1940 | card->info.portname, 9); | ||
1941 | rc = qeth_send_control_data(card, ULP_ENABLE_SIZE, iob, | ||
1942 | qeth_ulp_enable_cb, NULL); | ||
1943 | return rc; | ||
1944 | |||
1945 | } | ||
1946 | |||
1947 | static int qeth_ulp_setup_cb(struct qeth_card *card, struct qeth_reply *reply, | ||
1948 | unsigned long data) | ||
1949 | { | ||
1950 | struct qeth_cmd_buffer *iob; | ||
1951 | |||
1952 | QETH_DBF_TEXT(SETUP, 2, "ulpstpcb"); | ||
1953 | |||
1954 | iob = (struct qeth_cmd_buffer *) data; | ||
1955 | memcpy(&card->token.ulp_connection_r, | ||
1956 | QETH_ULP_SETUP_RESP_CONNECTION_TOKEN(iob->data), | ||
1957 | QETH_MPC_TOKEN_LENGTH); | ||
1958 | QETH_DBF_TEXT_(SETUP, 2, " rc%d", iob->rc); | ||
1959 | return 0; | ||
1960 | } | ||
1961 | |||
1962 | static int qeth_ulp_setup(struct qeth_card *card) | ||
1963 | { | ||
1964 | int rc; | ||
1965 | __u16 temp; | ||
1966 | struct qeth_cmd_buffer *iob; | ||
1967 | struct ccw_dev_id dev_id; | ||
1968 | |||
1969 | QETH_DBF_TEXT(SETUP, 2, "ulpsetup"); | ||
1970 | |||
1971 | iob = qeth_wait_for_buffer(&card->write); | ||
1972 | memcpy(iob->data, ULP_SETUP, ULP_SETUP_SIZE); | ||
1973 | |||
1974 | memcpy(QETH_ULP_SETUP_DEST_ADDR(iob->data), | ||
1975 | &card->token.cm_connection_r, QETH_MPC_TOKEN_LENGTH); | ||
1976 | memcpy(QETH_ULP_SETUP_CONNECTION_TOKEN(iob->data), | ||
1977 | &card->token.ulp_connection_w, QETH_MPC_TOKEN_LENGTH); | ||
1978 | memcpy(QETH_ULP_SETUP_FILTER_TOKEN(iob->data), | ||
1979 | &card->token.ulp_filter_r, QETH_MPC_TOKEN_LENGTH); | ||
1980 | |||
1981 | ccw_device_get_id(CARD_DDEV(card), &dev_id); | ||
1982 | memcpy(QETH_ULP_SETUP_CUA(iob->data), &dev_id.devno, 2); | ||
1983 | temp = (card->info.cula << 8) + card->info.unit_addr2; | ||
1984 | memcpy(QETH_ULP_SETUP_REAL_DEVADDR(iob->data), &temp, 2); | ||
1985 | rc = qeth_send_control_data(card, ULP_SETUP_SIZE, iob, | ||
1986 | qeth_ulp_setup_cb, NULL); | ||
1987 | return rc; | ||
1988 | } | ||
1989 | |||
1990 | static int qeth_alloc_qdio_buffers(struct qeth_card *card) | ||
1991 | { | ||
1992 | int i, j; | ||
1993 | |||
1994 | QETH_DBF_TEXT(SETUP, 2, "allcqdbf"); | ||
1995 | |||
1996 | if (atomic_cmpxchg(&card->qdio.state, QETH_QDIO_UNINITIALIZED, | ||
1997 | QETH_QDIO_ALLOCATED) != QETH_QDIO_UNINITIALIZED) | ||
1998 | return 0; | ||
1999 | |||
2000 | card->qdio.in_q = kmalloc(sizeof(struct qeth_qdio_q), | ||
2001 | GFP_KERNEL); | ||
2002 | if (!card->qdio.in_q) | ||
2003 | goto out_nomem; | ||
2004 | QETH_DBF_TEXT(SETUP, 2, "inq"); | ||
2005 | QETH_DBF_HEX(SETUP, 2, &card->qdio.in_q, sizeof(void *)); | ||
2006 | memset(card->qdio.in_q, 0, sizeof(struct qeth_qdio_q)); | ||
2007 | /* give inbound qeth_qdio_buffers their qdio_buffers */ | ||
2008 | for (i = 0; i < QDIO_MAX_BUFFERS_PER_Q; ++i) | ||
2009 | card->qdio.in_q->bufs[i].buffer = | ||
2010 | &card->qdio.in_q->qdio_bufs[i]; | ||
2011 | /* inbound buffer pool */ | ||
2012 | if (qeth_alloc_buffer_pool(card)) | ||
2013 | goto out_freeinq; | ||
2014 | /* outbound */ | ||
2015 | card->qdio.out_qs = | ||
2016 | kmalloc(card->qdio.no_out_queues * | ||
2017 | sizeof(struct qeth_qdio_out_q *), GFP_KERNEL); | ||
2018 | if (!card->qdio.out_qs) | ||
2019 | goto out_freepool; | ||
2020 | for (i = 0; i < card->qdio.no_out_queues; ++i) { | ||
2021 | card->qdio.out_qs[i] = kmalloc(sizeof(struct qeth_qdio_out_q), | ||
2022 | GFP_KERNEL); | ||
2023 | if (!card->qdio.out_qs[i]) | ||
2024 | goto out_freeoutq; | ||
2025 | QETH_DBF_TEXT_(SETUP, 2, "outq %i", i); | ||
2026 | QETH_DBF_HEX(SETUP, 2, &card->qdio.out_qs[i], sizeof(void *)); | ||
2027 | memset(card->qdio.out_qs[i], 0, sizeof(struct qeth_qdio_out_q)); | ||
2028 | card->qdio.out_qs[i]->queue_no = i; | ||
2029 | /* give outbound qeth_qdio_buffers their qdio_buffers */ | ||
2030 | for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j) { | ||
2031 | card->qdio.out_qs[i]->bufs[j].buffer = | ||
2032 | &card->qdio.out_qs[i]->qdio_bufs[j]; | ||
2033 | skb_queue_head_init(&card->qdio.out_qs[i]->bufs[j]. | ||
2034 | skb_list); | ||
2035 | lockdep_set_class( | ||
2036 | &card->qdio.out_qs[i]->bufs[j].skb_list.lock, | ||
2037 | &qdio_out_skb_queue_key); | ||
2038 | INIT_LIST_HEAD(&card->qdio.out_qs[i]->bufs[j].ctx_list); | ||
2039 | } | ||
2040 | } | ||
2041 | return 0; | ||
2042 | |||
2043 | out_freeoutq: | ||
2044 | while (i > 0) | ||
2045 | kfree(card->qdio.out_qs[--i]); | ||
2046 | kfree(card->qdio.out_qs); | ||
2047 | card->qdio.out_qs = NULL; | ||
2048 | out_freepool: | ||
2049 | qeth_free_buffer_pool(card); | ||
2050 | out_freeinq: | ||
2051 | kfree(card->qdio.in_q); | ||
2052 | card->qdio.in_q = NULL; | ||
2053 | out_nomem: | ||
2054 | atomic_set(&card->qdio.state, QETH_QDIO_UNINITIALIZED); | ||
2055 | return -ENOMEM; | ||
2056 | } | ||
2057 | |||
2058 | static void qeth_create_qib_param_field(struct qeth_card *card, | ||
2059 | char *param_field) | ||
2060 | { | ||
2061 | |||
2062 | param_field[0] = _ascebc['P']; | ||
2063 | param_field[1] = _ascebc['C']; | ||
2064 | param_field[2] = _ascebc['I']; | ||
2065 | param_field[3] = _ascebc['T']; | ||
2066 | *((unsigned int *) (¶m_field[4])) = QETH_PCI_THRESHOLD_A(card); | ||
2067 | *((unsigned int *) (¶m_field[8])) = QETH_PCI_THRESHOLD_B(card); | ||
2068 | *((unsigned int *) (¶m_field[12])) = QETH_PCI_TIMER_VALUE(card); | ||
2069 | } | ||
2070 | |||
2071 | static void qeth_create_qib_param_field_blkt(struct qeth_card *card, | ||
2072 | char *param_field) | ||
2073 | { | ||
2074 | param_field[16] = _ascebc['B']; | ||
2075 | param_field[17] = _ascebc['L']; | ||
2076 | param_field[18] = _ascebc['K']; | ||
2077 | param_field[19] = _ascebc['T']; | ||
2078 | *((unsigned int *) (¶m_field[20])) = card->info.blkt.time_total; | ||
2079 | *((unsigned int *) (¶m_field[24])) = card->info.blkt.inter_packet; | ||
2080 | *((unsigned int *) (¶m_field[28])) = | ||
2081 | card->info.blkt.inter_packet_jumbo; | ||
2082 | } | ||
2083 | |||
2084 | static int qeth_qdio_activate(struct qeth_card *card) | ||
2085 | { | ||
2086 | QETH_DBF_TEXT(SETUP, 3, "qdioact"); | ||
2087 | return qdio_activate(CARD_DDEV(card), 0); | ||
2088 | } | ||
2089 | |||
2090 | static int qeth_dm_act(struct qeth_card *card) | ||
2091 | { | ||
2092 | int rc; | ||
2093 | struct qeth_cmd_buffer *iob; | ||
2094 | |||
2095 | QETH_DBF_TEXT(SETUP, 2, "dmact"); | ||
2096 | |||
2097 | iob = qeth_wait_for_buffer(&card->write); | ||
2098 | memcpy(iob->data, DM_ACT, DM_ACT_SIZE); | ||
2099 | |||
2100 | memcpy(QETH_DM_ACT_DEST_ADDR(iob->data), | ||
2101 | &card->token.cm_connection_r, QETH_MPC_TOKEN_LENGTH); | ||
2102 | memcpy(QETH_DM_ACT_CONNECTION_TOKEN(iob->data), | ||
2103 | &card->token.ulp_connection_r, QETH_MPC_TOKEN_LENGTH); | ||
2104 | rc = qeth_send_control_data(card, DM_ACT_SIZE, iob, NULL, NULL); | ||
2105 | return rc; | ||
2106 | } | ||
2107 | |||
2108 | static int qeth_mpc_initialize(struct qeth_card *card) | ||
2109 | { | ||
2110 | int rc; | ||
2111 | |||
2112 | QETH_DBF_TEXT(SETUP, 2, "mpcinit"); | ||
2113 | |||
2114 | rc = qeth_issue_next_read(card); | ||
2115 | if (rc) { | ||
2116 | QETH_DBF_TEXT_(SETUP, 2, "1err%d", rc); | ||
2117 | return rc; | ||
2118 | } | ||
2119 | rc = qeth_cm_enable(card); | ||
2120 | if (rc) { | ||
2121 | QETH_DBF_TEXT_(SETUP, 2, "2err%d", rc); | ||
2122 | goto out_qdio; | ||
2123 | } | ||
2124 | rc = qeth_cm_setup(card); | ||
2125 | if (rc) { | ||
2126 | QETH_DBF_TEXT_(SETUP, 2, "3err%d", rc); | ||
2127 | goto out_qdio; | ||
2128 | } | ||
2129 | rc = qeth_ulp_enable(card); | ||
2130 | if (rc) { | ||
2131 | QETH_DBF_TEXT_(SETUP, 2, "4err%d", rc); | ||
2132 | goto out_qdio; | ||
2133 | } | ||
2134 | rc = qeth_ulp_setup(card); | ||
2135 | if (rc) { | ||
2136 | QETH_DBF_TEXT_(SETUP, 2, "5err%d", rc); | ||
2137 | goto out_qdio; | ||
2138 | } | ||
2139 | rc = qeth_alloc_qdio_buffers(card); | ||
2140 | if (rc) { | ||
2141 | QETH_DBF_TEXT_(SETUP, 2, "5err%d", rc); | ||
2142 | goto out_qdio; | ||
2143 | } | ||
2144 | rc = qeth_qdio_establish(card); | ||
2145 | if (rc) { | ||
2146 | QETH_DBF_TEXT_(SETUP, 2, "6err%d", rc); | ||
2147 | qeth_free_qdio_buffers(card); | ||
2148 | goto out_qdio; | ||
2149 | } | ||
2150 | rc = qeth_qdio_activate(card); | ||
2151 | if (rc) { | ||
2152 | QETH_DBF_TEXT_(SETUP, 2, "7err%d", rc); | ||
2153 | goto out_qdio; | ||
2154 | } | ||
2155 | rc = qeth_dm_act(card); | ||
2156 | if (rc) { | ||
2157 | QETH_DBF_TEXT_(SETUP, 2, "8err%d", rc); | ||
2158 | goto out_qdio; | ||
2159 | } | ||
2160 | |||
2161 | return 0; | ||
2162 | out_qdio: | ||
2163 | qeth_qdio_clear_card(card, card->info.type != QETH_CARD_TYPE_IQD); | ||
2164 | return rc; | ||
2165 | } | ||
2166 | |||
2167 | static void qeth_print_status_with_portname(struct qeth_card *card) | ||
2168 | { | ||
2169 | char dbf_text[15]; | ||
2170 | int i; | ||
2171 | |||
2172 | sprintf(dbf_text, "%s", card->info.portname + 1); | ||
2173 | for (i = 0; i < 8; i++) | ||
2174 | dbf_text[i] = | ||
2175 | (char) _ebcasc[(__u8) dbf_text[i]]; | ||
2176 | dbf_text[8] = 0; | ||
2177 | PRINT_INFO("Device %s/%s/%s is a%s card%s%s%s\n" | ||
2178 | "with link type %s (portname: %s)\n", | ||
2179 | CARD_RDEV_ID(card), | ||
2180 | CARD_WDEV_ID(card), | ||
2181 | CARD_DDEV_ID(card), | ||
2182 | qeth_get_cardname(card), | ||
2183 | (card->info.mcl_level[0]) ? " (level: " : "", | ||
2184 | (card->info.mcl_level[0]) ? card->info.mcl_level : "", | ||
2185 | (card->info.mcl_level[0]) ? ")" : "", | ||
2186 | qeth_get_cardname_short(card), | ||
2187 | dbf_text); | ||
2188 | |||
2189 | } | ||
2190 | |||
2191 | static void qeth_print_status_no_portname(struct qeth_card *card) | ||
2192 | { | ||
2193 | if (card->info.portname[0]) | ||
2194 | PRINT_INFO("Device %s/%s/%s is a%s " | ||
2195 | "card%s%s%s\nwith link type %s " | ||
2196 | "(no portname needed by interface).\n", | ||
2197 | CARD_RDEV_ID(card), | ||
2198 | CARD_WDEV_ID(card), | ||
2199 | CARD_DDEV_ID(card), | ||
2200 | qeth_get_cardname(card), | ||
2201 | (card->info.mcl_level[0]) ? " (level: " : "", | ||
2202 | (card->info.mcl_level[0]) ? card->info.mcl_level : "", | ||
2203 | (card->info.mcl_level[0]) ? ")" : "", | ||
2204 | qeth_get_cardname_short(card)); | ||
2205 | else | ||
2206 | PRINT_INFO("Device %s/%s/%s is a%s " | ||
2207 | "card%s%s%s\nwith link type %s.\n", | ||
2208 | CARD_RDEV_ID(card), | ||
2209 | CARD_WDEV_ID(card), | ||
2210 | CARD_DDEV_ID(card), | ||
2211 | qeth_get_cardname(card), | ||
2212 | (card->info.mcl_level[0]) ? " (level: " : "", | ||
2213 | (card->info.mcl_level[0]) ? card->info.mcl_level : "", | ||
2214 | (card->info.mcl_level[0]) ? ")" : "", | ||
2215 | qeth_get_cardname_short(card)); | ||
2216 | } | ||
2217 | |||
2218 | void qeth_print_status_message(struct qeth_card *card) | ||
2219 | { | ||
2220 | switch (card->info.type) { | ||
2221 | case QETH_CARD_TYPE_OSAE: | ||
2222 | /* VM will use a non-zero first character | ||
2223 | * to indicate a HiperSockets like reporting | ||
2224 | * of the level OSA sets the first character to zero | ||
2225 | * */ | ||
2226 | if (!card->info.mcl_level[0]) { | ||
2227 | sprintf(card->info.mcl_level, "%02x%02x", | ||
2228 | card->info.mcl_level[2], | ||
2229 | card->info.mcl_level[3]); | ||
2230 | |||
2231 | card->info.mcl_level[QETH_MCL_LENGTH] = 0; | ||
2232 | break; | ||
2233 | } | ||
2234 | /* fallthrough */ | ||
2235 | case QETH_CARD_TYPE_IQD: | ||
2236 | if (card->info.guestlan) { | ||
2237 | card->info.mcl_level[0] = (char) _ebcasc[(__u8) | ||
2238 | card->info.mcl_level[0]]; | ||
2239 | card->info.mcl_level[1] = (char) _ebcasc[(__u8) | ||
2240 | card->info.mcl_level[1]]; | ||
2241 | card->info.mcl_level[2] = (char) _ebcasc[(__u8) | ||
2242 | card->info.mcl_level[2]]; | ||
2243 | card->info.mcl_level[3] = (char) _ebcasc[(__u8) | ||
2244 | card->info.mcl_level[3]]; | ||
2245 | card->info.mcl_level[QETH_MCL_LENGTH] = 0; | ||
2246 | } | ||
2247 | break; | ||
2248 | default: | ||
2249 | memset(&card->info.mcl_level[0], 0, QETH_MCL_LENGTH + 1); | ||
2250 | } | ||
2251 | if (card->info.portname_required) | ||
2252 | qeth_print_status_with_portname(card); | ||
2253 | else | ||
2254 | qeth_print_status_no_portname(card); | ||
2255 | } | ||
2256 | EXPORT_SYMBOL_GPL(qeth_print_status_message); | ||
2257 | |||
2258 | void qeth_put_buffer_pool_entry(struct qeth_card *card, | ||
2259 | struct qeth_buffer_pool_entry *entry) | ||
2260 | { | ||
2261 | QETH_DBF_TEXT(TRACE, 6, "ptbfplen"); | ||
2262 | list_add_tail(&entry->list, &card->qdio.in_buf_pool.entry_list); | ||
2263 | } | ||
2264 | EXPORT_SYMBOL_GPL(qeth_put_buffer_pool_entry); | ||
2265 | |||
2266 | static void qeth_initialize_working_pool_list(struct qeth_card *card) | ||
2267 | { | ||
2268 | struct qeth_buffer_pool_entry *entry; | ||
2269 | |||
2270 | QETH_DBF_TEXT(TRACE, 5, "inwrklst"); | ||
2271 | |||
2272 | list_for_each_entry(entry, | ||
2273 | &card->qdio.init_pool.entry_list, init_list) { | ||
2274 | qeth_put_buffer_pool_entry(card, entry); | ||
2275 | } | ||
2276 | } | ||
2277 | |||
2278 | static inline struct qeth_buffer_pool_entry *qeth_find_free_buffer_pool_entry( | ||
2279 | struct qeth_card *card) | ||
2280 | { | ||
2281 | struct list_head *plh; | ||
2282 | struct qeth_buffer_pool_entry *entry; | ||
2283 | int i, free; | ||
2284 | struct page *page; | ||
2285 | |||
2286 | if (list_empty(&card->qdio.in_buf_pool.entry_list)) | ||
2287 | return NULL; | ||
2288 | |||
2289 | list_for_each(plh, &card->qdio.in_buf_pool.entry_list) { | ||
2290 | entry = list_entry(plh, struct qeth_buffer_pool_entry, list); | ||
2291 | free = 1; | ||
2292 | for (i = 0; i < QETH_MAX_BUFFER_ELEMENTS(card); ++i) { | ||
2293 | if (page_count(virt_to_page(entry->elements[i])) > 1) { | ||
2294 | free = 0; | ||
2295 | break; | ||
2296 | } | ||
2297 | } | ||
2298 | if (free) { | ||
2299 | list_del_init(&entry->list); | ||
2300 | return entry; | ||
2301 | } | ||
2302 | } | ||
2303 | |||
2304 | /* no free buffer in pool so take first one and swap pages */ | ||
2305 | entry = list_entry(card->qdio.in_buf_pool.entry_list.next, | ||
2306 | struct qeth_buffer_pool_entry, list); | ||
2307 | for (i = 0; i < QETH_MAX_BUFFER_ELEMENTS(card); ++i) { | ||
2308 | if (page_count(virt_to_page(entry->elements[i])) > 1) { | ||
2309 | page = alloc_page(GFP_ATOMIC); | ||
2310 | if (!page) { | ||
2311 | return NULL; | ||
2312 | } else { | ||
2313 | free_page((unsigned long)entry->elements[i]); | ||
2314 | entry->elements[i] = page_address(page); | ||
2315 | if (card->options.performance_stats) | ||
2316 | card->perf_stats.sg_alloc_page_rx++; | ||
2317 | } | ||
2318 | } | ||
2319 | } | ||
2320 | list_del_init(&entry->list); | ||
2321 | return entry; | ||
2322 | } | ||
2323 | |||
2324 | static int qeth_init_input_buffer(struct qeth_card *card, | ||
2325 | struct qeth_qdio_buffer *buf) | ||
2326 | { | ||
2327 | struct qeth_buffer_pool_entry *pool_entry; | ||
2328 | int i; | ||
2329 | |||
2330 | pool_entry = qeth_find_free_buffer_pool_entry(card); | ||
2331 | if (!pool_entry) | ||
2332 | return 1; | ||
2333 | |||
2334 | /* | ||
2335 | * since the buffer is accessed only from the input_tasklet | ||
2336 | * there shouldn't be a need to synchronize; also, since we use | ||
2337 | * the QETH_IN_BUF_REQUEUE_THRESHOLD we should never run out off | ||
2338 | * buffers | ||
2339 | */ | ||
2340 | BUG_ON(!pool_entry); | ||
2341 | |||
2342 | buf->pool_entry = pool_entry; | ||
2343 | for (i = 0; i < QETH_MAX_BUFFER_ELEMENTS(card); ++i) { | ||
2344 | buf->buffer->element[i].length = PAGE_SIZE; | ||
2345 | buf->buffer->element[i].addr = pool_entry->elements[i]; | ||
2346 | if (i == QETH_MAX_BUFFER_ELEMENTS(card) - 1) | ||
2347 | buf->buffer->element[i].flags = SBAL_FLAGS_LAST_ENTRY; | ||
2348 | else | ||
2349 | buf->buffer->element[i].flags = 0; | ||
2350 | } | ||
2351 | return 0; | ||
2352 | } | ||
2353 | |||
2354 | int qeth_init_qdio_queues(struct qeth_card *card) | ||
2355 | { | ||
2356 | int i, j; | ||
2357 | int rc; | ||
2358 | |||
2359 | QETH_DBF_TEXT(SETUP, 2, "initqdqs"); | ||
2360 | |||
2361 | /* inbound queue */ | ||
2362 | memset(card->qdio.in_q->qdio_bufs, 0, | ||
2363 | QDIO_MAX_BUFFERS_PER_Q * sizeof(struct qdio_buffer)); | ||
2364 | qeth_initialize_working_pool_list(card); | ||
2365 | /*give only as many buffers to hardware as we have buffer pool entries*/ | ||
2366 | for (i = 0; i < card->qdio.in_buf_pool.buf_count - 1; ++i) | ||
2367 | qeth_init_input_buffer(card, &card->qdio.in_q->bufs[i]); | ||
2368 | card->qdio.in_q->next_buf_to_init = | ||
2369 | card->qdio.in_buf_pool.buf_count - 1; | ||
2370 | rc = do_QDIO(CARD_DDEV(card), QDIO_FLAG_SYNC_INPUT, 0, 0, | ||
2371 | card->qdio.in_buf_pool.buf_count - 1, NULL); | ||
2372 | if (rc) { | ||
2373 | QETH_DBF_TEXT_(SETUP, 2, "1err%d", rc); | ||
2374 | return rc; | ||
2375 | } | ||
2376 | rc = qdio_synchronize(CARD_DDEV(card), QDIO_FLAG_SYNC_INPUT, 0); | ||
2377 | if (rc) { | ||
2378 | QETH_DBF_TEXT_(SETUP, 2, "2err%d", rc); | ||
2379 | return rc; | ||
2380 | } | ||
2381 | /* outbound queue */ | ||
2382 | for (i = 0; i < card->qdio.no_out_queues; ++i) { | ||
2383 | memset(card->qdio.out_qs[i]->qdio_bufs, 0, | ||
2384 | QDIO_MAX_BUFFERS_PER_Q * sizeof(struct qdio_buffer)); | ||
2385 | for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j) { | ||
2386 | qeth_clear_output_buffer(card->qdio.out_qs[i], | ||
2387 | &card->qdio.out_qs[i]->bufs[j]); | ||
2388 | } | ||
2389 | card->qdio.out_qs[i]->card = card; | ||
2390 | card->qdio.out_qs[i]->next_buf_to_fill = 0; | ||
2391 | card->qdio.out_qs[i]->do_pack = 0; | ||
2392 | atomic_set(&card->qdio.out_qs[i]->used_buffers, 0); | ||
2393 | atomic_set(&card->qdio.out_qs[i]->set_pci_flags_count, 0); | ||
2394 | atomic_set(&card->qdio.out_qs[i]->state, | ||
2395 | QETH_OUT_Q_UNLOCKED); | ||
2396 | } | ||
2397 | return 0; | ||
2398 | } | ||
2399 | EXPORT_SYMBOL_GPL(qeth_init_qdio_queues); | ||
2400 | |||
2401 | static inline __u8 qeth_get_ipa_adp_type(enum qeth_link_types link_type) | ||
2402 | { | ||
2403 | switch (link_type) { | ||
2404 | case QETH_LINK_TYPE_HSTR: | ||
2405 | return 2; | ||
2406 | default: | ||
2407 | return 1; | ||
2408 | } | ||
2409 | } | ||
2410 | |||
2411 | static void qeth_fill_ipacmd_header(struct qeth_card *card, | ||
2412 | struct qeth_ipa_cmd *cmd, __u8 command, | ||
2413 | enum qeth_prot_versions prot) | ||
2414 | { | ||
2415 | memset(cmd, 0, sizeof(struct qeth_ipa_cmd)); | ||
2416 | cmd->hdr.command = command; | ||
2417 | cmd->hdr.initiator = IPA_CMD_INITIATOR_HOST; | ||
2418 | cmd->hdr.seqno = card->seqno.ipa; | ||
2419 | cmd->hdr.adapter_type = qeth_get_ipa_adp_type(card->info.link_type); | ||
2420 | cmd->hdr.rel_adapter_no = (__u8) card->info.portno; | ||
2421 | if (card->options.layer2) | ||
2422 | cmd->hdr.prim_version_no = 2; | ||
2423 | else | ||
2424 | cmd->hdr.prim_version_no = 1; | ||
2425 | cmd->hdr.param_count = 1; | ||
2426 | cmd->hdr.prot_version = prot; | ||
2427 | cmd->hdr.ipa_supported = 0; | ||
2428 | cmd->hdr.ipa_enabled = 0; | ||
2429 | } | ||
2430 | |||
2431 | struct qeth_cmd_buffer *qeth_get_ipacmd_buffer(struct qeth_card *card, | ||
2432 | enum qeth_ipa_cmds ipacmd, enum qeth_prot_versions prot) | ||
2433 | { | ||
2434 | struct qeth_cmd_buffer *iob; | ||
2435 | struct qeth_ipa_cmd *cmd; | ||
2436 | |||
2437 | iob = qeth_wait_for_buffer(&card->write); | ||
2438 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
2439 | qeth_fill_ipacmd_header(card, cmd, ipacmd, prot); | ||
2440 | |||
2441 | return iob; | ||
2442 | } | ||
2443 | EXPORT_SYMBOL_GPL(qeth_get_ipacmd_buffer); | ||
2444 | |||
2445 | void qeth_prepare_ipa_cmd(struct qeth_card *card, struct qeth_cmd_buffer *iob, | ||
2446 | char prot_type) | ||
2447 | { | ||
2448 | memcpy(iob->data, IPA_PDU_HEADER, IPA_PDU_HEADER_SIZE); | ||
2449 | memcpy(QETH_IPA_CMD_PROT_TYPE(iob->data), &prot_type, 1); | ||
2450 | memcpy(QETH_IPA_CMD_DEST_ADDR(iob->data), | ||
2451 | &card->token.ulp_connection_r, QETH_MPC_TOKEN_LENGTH); | ||
2452 | } | ||
2453 | EXPORT_SYMBOL_GPL(qeth_prepare_ipa_cmd); | ||
2454 | |||
2455 | int qeth_send_ipa_cmd(struct qeth_card *card, struct qeth_cmd_buffer *iob, | ||
2456 | int (*reply_cb)(struct qeth_card *, struct qeth_reply*, | ||
2457 | unsigned long), | ||
2458 | void *reply_param) | ||
2459 | { | ||
2460 | int rc; | ||
2461 | char prot_type; | ||
2462 | |||
2463 | QETH_DBF_TEXT(TRACE, 4, "sendipa"); | ||
2464 | |||
2465 | if (card->options.layer2) | ||
2466 | if (card->info.type == QETH_CARD_TYPE_OSN) | ||
2467 | prot_type = QETH_PROT_OSN2; | ||
2468 | else | ||
2469 | prot_type = QETH_PROT_LAYER2; | ||
2470 | else | ||
2471 | prot_type = QETH_PROT_TCPIP; | ||
2472 | qeth_prepare_ipa_cmd(card, iob, prot_type); | ||
2473 | rc = qeth_send_control_data(card, IPA_CMD_LENGTH, | ||
2474 | iob, reply_cb, reply_param); | ||
2475 | return rc; | ||
2476 | } | ||
2477 | EXPORT_SYMBOL_GPL(qeth_send_ipa_cmd); | ||
2478 | |||
2479 | static int qeth_send_startstoplan(struct qeth_card *card, | ||
2480 | enum qeth_ipa_cmds ipacmd, enum qeth_prot_versions prot) | ||
2481 | { | ||
2482 | int rc; | ||
2483 | struct qeth_cmd_buffer *iob; | ||
2484 | |||
2485 | iob = qeth_get_ipacmd_buffer(card, ipacmd, prot); | ||
2486 | rc = qeth_send_ipa_cmd(card, iob, NULL, NULL); | ||
2487 | |||
2488 | return rc; | ||
2489 | } | ||
2490 | |||
2491 | int qeth_send_startlan(struct qeth_card *card) | ||
2492 | { | ||
2493 | int rc; | ||
2494 | |||
2495 | QETH_DBF_TEXT(SETUP, 2, "strtlan"); | ||
2496 | |||
2497 | rc = qeth_send_startstoplan(card, IPA_CMD_STARTLAN, 0); | ||
2498 | return rc; | ||
2499 | } | ||
2500 | EXPORT_SYMBOL_GPL(qeth_send_startlan); | ||
2501 | |||
2502 | int qeth_send_stoplan(struct qeth_card *card) | ||
2503 | { | ||
2504 | int rc = 0; | ||
2505 | |||
2506 | /* | ||
2507 | * TODO: according to the IPA format document page 14, | ||
2508 | * TCP/IP (we!) never issue a STOPLAN | ||
2509 | * is this right ?!? | ||
2510 | */ | ||
2511 | QETH_DBF_TEXT(SETUP, 2, "stoplan"); | ||
2512 | |||
2513 | rc = qeth_send_startstoplan(card, IPA_CMD_STOPLAN, 0); | ||
2514 | return rc; | ||
2515 | } | ||
2516 | EXPORT_SYMBOL_GPL(qeth_send_stoplan); | ||
2517 | |||
2518 | int qeth_default_setadapterparms_cb(struct qeth_card *card, | ||
2519 | struct qeth_reply *reply, unsigned long data) | ||
2520 | { | ||
2521 | struct qeth_ipa_cmd *cmd; | ||
2522 | |||
2523 | QETH_DBF_TEXT(TRACE, 4, "defadpcb"); | ||
2524 | |||
2525 | cmd = (struct qeth_ipa_cmd *) data; | ||
2526 | if (cmd->hdr.return_code == 0) | ||
2527 | cmd->hdr.return_code = | ||
2528 | cmd->data.setadapterparms.hdr.return_code; | ||
2529 | return 0; | ||
2530 | } | ||
2531 | EXPORT_SYMBOL_GPL(qeth_default_setadapterparms_cb); | ||
2532 | |||
2533 | static int qeth_query_setadapterparms_cb(struct qeth_card *card, | ||
2534 | struct qeth_reply *reply, unsigned long data) | ||
2535 | { | ||
2536 | struct qeth_ipa_cmd *cmd; | ||
2537 | |||
2538 | QETH_DBF_TEXT(TRACE, 3, "quyadpcb"); | ||
2539 | |||
2540 | cmd = (struct qeth_ipa_cmd *) data; | ||
2541 | if (cmd->data.setadapterparms.data.query_cmds_supp.lan_type & 0x7f) | ||
2542 | card->info.link_type = | ||
2543 | cmd->data.setadapterparms.data.query_cmds_supp.lan_type; | ||
2544 | card->options.adp.supported_funcs = | ||
2545 | cmd->data.setadapterparms.data.query_cmds_supp.supported_cmds; | ||
2546 | return qeth_default_setadapterparms_cb(card, reply, (unsigned long)cmd); | ||
2547 | } | ||
2548 | |||
2549 | struct qeth_cmd_buffer *qeth_get_adapter_cmd(struct qeth_card *card, | ||
2550 | __u32 command, __u32 cmdlen) | ||
2551 | { | ||
2552 | struct qeth_cmd_buffer *iob; | ||
2553 | struct qeth_ipa_cmd *cmd; | ||
2554 | |||
2555 | iob = qeth_get_ipacmd_buffer(card, IPA_CMD_SETADAPTERPARMS, | ||
2556 | QETH_PROT_IPV4); | ||
2557 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
2558 | cmd->data.setadapterparms.hdr.cmdlength = cmdlen; | ||
2559 | cmd->data.setadapterparms.hdr.command_code = command; | ||
2560 | cmd->data.setadapterparms.hdr.used_total = 1; | ||
2561 | cmd->data.setadapterparms.hdr.seq_no = 1; | ||
2562 | |||
2563 | return iob; | ||
2564 | } | ||
2565 | EXPORT_SYMBOL_GPL(qeth_get_adapter_cmd); | ||
2566 | |||
2567 | int qeth_query_setadapterparms(struct qeth_card *card) | ||
2568 | { | ||
2569 | int rc; | ||
2570 | struct qeth_cmd_buffer *iob; | ||
2571 | |||
2572 | QETH_DBF_TEXT(TRACE, 3, "queryadp"); | ||
2573 | iob = qeth_get_adapter_cmd(card, IPA_SETADP_QUERY_COMMANDS_SUPPORTED, | ||
2574 | sizeof(struct qeth_ipacmd_setadpparms)); | ||
2575 | rc = qeth_send_ipa_cmd(card, iob, qeth_query_setadapterparms_cb, NULL); | ||
2576 | return rc; | ||
2577 | } | ||
2578 | EXPORT_SYMBOL_GPL(qeth_query_setadapterparms); | ||
2579 | |||
2580 | int qeth_check_qdio_errors(struct qdio_buffer *buf, unsigned int qdio_error, | ||
2581 | unsigned int siga_error, const char *dbftext) | ||
2582 | { | ||
2583 | if (qdio_error || siga_error) { | ||
2584 | QETH_DBF_TEXT(TRACE, 2, dbftext); | ||
2585 | QETH_DBF_TEXT(QERR, 2, dbftext); | ||
2586 | QETH_DBF_TEXT_(QERR, 2, " F15=%02X", | ||
2587 | buf->element[15].flags & 0xff); | ||
2588 | QETH_DBF_TEXT_(QERR, 2, " F14=%02X", | ||
2589 | buf->element[14].flags & 0xff); | ||
2590 | QETH_DBF_TEXT_(QERR, 2, " qerr=%X", qdio_error); | ||
2591 | QETH_DBF_TEXT_(QERR, 2, " serr=%X", siga_error); | ||
2592 | return 1; | ||
2593 | } | ||
2594 | return 0; | ||
2595 | } | ||
2596 | EXPORT_SYMBOL_GPL(qeth_check_qdio_errors); | ||
2597 | |||
2598 | void qeth_queue_input_buffer(struct qeth_card *card, int index) | ||
2599 | { | ||
2600 | struct qeth_qdio_q *queue = card->qdio.in_q; | ||
2601 | int count; | ||
2602 | int i; | ||
2603 | int rc; | ||
2604 | int newcount = 0; | ||
2605 | |||
2606 | QETH_DBF_TEXT(TRACE, 6, "queinbuf"); | ||
2607 | count = (index < queue->next_buf_to_init)? | ||
2608 | card->qdio.in_buf_pool.buf_count - | ||
2609 | (queue->next_buf_to_init - index) : | ||
2610 | card->qdio.in_buf_pool.buf_count - | ||
2611 | (queue->next_buf_to_init + QDIO_MAX_BUFFERS_PER_Q - index); | ||
2612 | /* only requeue at a certain threshold to avoid SIGAs */ | ||
2613 | if (count >= QETH_IN_BUF_REQUEUE_THRESHOLD(card)) { | ||
2614 | for (i = queue->next_buf_to_init; | ||
2615 | i < queue->next_buf_to_init + count; ++i) { | ||
2616 | if (qeth_init_input_buffer(card, | ||
2617 | &queue->bufs[i % QDIO_MAX_BUFFERS_PER_Q])) { | ||
2618 | break; | ||
2619 | } else { | ||
2620 | newcount++; | ||
2621 | } | ||
2622 | } | ||
2623 | |||
2624 | if (newcount < count) { | ||
2625 | /* we are in memory shortage so we switch back to | ||
2626 | traditional skb allocation and drop packages */ | ||
2627 | if (!atomic_read(&card->force_alloc_skb) && | ||
2628 | net_ratelimit()) | ||
2629 | PRINT_WARN("Switch to alloc skb\n"); | ||
2630 | atomic_set(&card->force_alloc_skb, 3); | ||
2631 | count = newcount; | ||
2632 | } else { | ||
2633 | if ((atomic_read(&card->force_alloc_skb) == 1) && | ||
2634 | net_ratelimit()) | ||
2635 | PRINT_WARN("Switch to sg\n"); | ||
2636 | atomic_add_unless(&card->force_alloc_skb, -1, 0); | ||
2637 | } | ||
2638 | |||
2639 | /* | ||
2640 | * according to old code it should be avoided to requeue all | ||
2641 | * 128 buffers in order to benefit from PCI avoidance. | ||
2642 | * this function keeps at least one buffer (the buffer at | ||
2643 | * 'index') un-requeued -> this buffer is the first buffer that | ||
2644 | * will be requeued the next time | ||
2645 | */ | ||
2646 | if (card->options.performance_stats) { | ||
2647 | card->perf_stats.inbound_do_qdio_cnt++; | ||
2648 | card->perf_stats.inbound_do_qdio_start_time = | ||
2649 | qeth_get_micros(); | ||
2650 | } | ||
2651 | rc = do_QDIO(CARD_DDEV(card), | ||
2652 | QDIO_FLAG_SYNC_INPUT | QDIO_FLAG_UNDER_INTERRUPT, | ||
2653 | 0, queue->next_buf_to_init, count, NULL); | ||
2654 | if (card->options.performance_stats) | ||
2655 | card->perf_stats.inbound_do_qdio_time += | ||
2656 | qeth_get_micros() - | ||
2657 | card->perf_stats.inbound_do_qdio_start_time; | ||
2658 | if (rc) { | ||
2659 | PRINT_WARN("qeth_queue_input_buffer's do_QDIO " | ||
2660 | "return %i (device %s).\n", | ||
2661 | rc, CARD_DDEV_ID(card)); | ||
2662 | QETH_DBF_TEXT(TRACE, 2, "qinberr"); | ||
2663 | QETH_DBF_TEXT_(TRACE, 2, "%s", CARD_BUS_ID(card)); | ||
2664 | } | ||
2665 | queue->next_buf_to_init = (queue->next_buf_to_init + count) % | ||
2666 | QDIO_MAX_BUFFERS_PER_Q; | ||
2667 | } | ||
2668 | } | ||
2669 | EXPORT_SYMBOL_GPL(qeth_queue_input_buffer); | ||
2670 | |||
2671 | static int qeth_handle_send_error(struct qeth_card *card, | ||
2672 | struct qeth_qdio_out_buffer *buffer, unsigned int qdio_err, | ||
2673 | unsigned int siga_err) | ||
2674 | { | ||
2675 | int sbalf15 = buffer->buffer->element[15].flags & 0xff; | ||
2676 | int cc = siga_err & 3; | ||
2677 | |||
2678 | QETH_DBF_TEXT(TRACE, 6, "hdsnderr"); | ||
2679 | qeth_check_qdio_errors(buffer->buffer, qdio_err, siga_err, "qouterr"); | ||
2680 | switch (cc) { | ||
2681 | case 0: | ||
2682 | if (qdio_err) { | ||
2683 | QETH_DBF_TEXT(TRACE, 1, "lnkfail"); | ||
2684 | QETH_DBF_TEXT_(TRACE, 1, "%s", CARD_BUS_ID(card)); | ||
2685 | QETH_DBF_TEXT_(TRACE, 1, "%04x %02x", | ||
2686 | (u16)qdio_err, (u8)sbalf15); | ||
2687 | return QETH_SEND_ERROR_LINK_FAILURE; | ||
2688 | } | ||
2689 | return QETH_SEND_ERROR_NONE; | ||
2690 | case 2: | ||
2691 | if (siga_err & QDIO_SIGA_ERROR_B_BIT_SET) { | ||
2692 | QETH_DBF_TEXT(TRACE, 1, "SIGAcc2B"); | ||
2693 | QETH_DBF_TEXT_(TRACE, 1, "%s", CARD_BUS_ID(card)); | ||
2694 | return QETH_SEND_ERROR_KICK_IT; | ||
2695 | } | ||
2696 | if ((sbalf15 >= 15) && (sbalf15 <= 31)) | ||
2697 | return QETH_SEND_ERROR_RETRY; | ||
2698 | return QETH_SEND_ERROR_LINK_FAILURE; | ||
2699 | /* look at qdio_error and sbalf 15 */ | ||
2700 | case 1: | ||
2701 | QETH_DBF_TEXT(TRACE, 1, "SIGAcc1"); | ||
2702 | QETH_DBF_TEXT_(TRACE, 1, "%s", CARD_BUS_ID(card)); | ||
2703 | return QETH_SEND_ERROR_LINK_FAILURE; | ||
2704 | case 3: | ||
2705 | default: | ||
2706 | QETH_DBF_TEXT(TRACE, 1, "SIGAcc3"); | ||
2707 | QETH_DBF_TEXT_(TRACE, 1, "%s", CARD_BUS_ID(card)); | ||
2708 | return QETH_SEND_ERROR_KICK_IT; | ||
2709 | } | ||
2710 | } | ||
2711 | |||
2712 | /* | ||
2713 | * Switched to packing state if the number of used buffers on a queue | ||
2714 | * reaches a certain limit. | ||
2715 | */ | ||
2716 | static void qeth_switch_to_packing_if_needed(struct qeth_qdio_out_q *queue) | ||
2717 | { | ||
2718 | if (!queue->do_pack) { | ||
2719 | if (atomic_read(&queue->used_buffers) | ||
2720 | >= QETH_HIGH_WATERMARK_PACK){ | ||
2721 | /* switch non-PACKING -> PACKING */ | ||
2722 | QETH_DBF_TEXT(TRACE, 6, "np->pack"); | ||
2723 | if (queue->card->options.performance_stats) | ||
2724 | queue->card->perf_stats.sc_dp_p++; | ||
2725 | queue->do_pack = 1; | ||
2726 | } | ||
2727 | } | ||
2728 | } | ||
2729 | |||
2730 | /* | ||
2731 | * Switches from packing to non-packing mode. If there is a packing | ||
2732 | * buffer on the queue this buffer will be prepared to be flushed. | ||
2733 | * In that case 1 is returned to inform the caller. If no buffer | ||
2734 | * has to be flushed, zero is returned. | ||
2735 | */ | ||
2736 | static int qeth_switch_to_nonpacking_if_needed(struct qeth_qdio_out_q *queue) | ||
2737 | { | ||
2738 | struct qeth_qdio_out_buffer *buffer; | ||
2739 | int flush_count = 0; | ||
2740 | |||
2741 | if (queue->do_pack) { | ||
2742 | if (atomic_read(&queue->used_buffers) | ||
2743 | <= QETH_LOW_WATERMARK_PACK) { | ||
2744 | /* switch PACKING -> non-PACKING */ | ||
2745 | QETH_DBF_TEXT(TRACE, 6, "pack->np"); | ||
2746 | if (queue->card->options.performance_stats) | ||
2747 | queue->card->perf_stats.sc_p_dp++; | ||
2748 | queue->do_pack = 0; | ||
2749 | /* flush packing buffers */ | ||
2750 | buffer = &queue->bufs[queue->next_buf_to_fill]; | ||
2751 | if ((atomic_read(&buffer->state) == | ||
2752 | QETH_QDIO_BUF_EMPTY) && | ||
2753 | (buffer->next_element_to_fill > 0)) { | ||
2754 | atomic_set(&buffer->state, | ||
2755 | QETH_QDIO_BUF_PRIMED); | ||
2756 | flush_count++; | ||
2757 | queue->next_buf_to_fill = | ||
2758 | (queue->next_buf_to_fill + 1) % | ||
2759 | QDIO_MAX_BUFFERS_PER_Q; | ||
2760 | } | ||
2761 | } | ||
2762 | } | ||
2763 | return flush_count; | ||
2764 | } | ||
2765 | |||
2766 | /* | ||
2767 | * Called to flush a packing buffer if no more pci flags are on the queue. | ||
2768 | * Checks if there is a packing buffer and prepares it to be flushed. | ||
2769 | * In that case returns 1, otherwise zero. | ||
2770 | */ | ||
2771 | static int qeth_flush_buffers_on_no_pci(struct qeth_qdio_out_q *queue) | ||
2772 | { | ||
2773 | struct qeth_qdio_out_buffer *buffer; | ||
2774 | |||
2775 | buffer = &queue->bufs[queue->next_buf_to_fill]; | ||
2776 | if ((atomic_read(&buffer->state) == QETH_QDIO_BUF_EMPTY) && | ||
2777 | (buffer->next_element_to_fill > 0)) { | ||
2778 | /* it's a packing buffer */ | ||
2779 | atomic_set(&buffer->state, QETH_QDIO_BUF_PRIMED); | ||
2780 | queue->next_buf_to_fill = | ||
2781 | (queue->next_buf_to_fill + 1) % QDIO_MAX_BUFFERS_PER_Q; | ||
2782 | return 1; | ||
2783 | } | ||
2784 | return 0; | ||
2785 | } | ||
2786 | |||
2787 | static void qeth_flush_buffers(struct qeth_qdio_out_q *queue, int under_int, | ||
2788 | int index, int count) | ||
2789 | { | ||
2790 | struct qeth_qdio_out_buffer *buf; | ||
2791 | int rc; | ||
2792 | int i; | ||
2793 | unsigned int qdio_flags; | ||
2794 | |||
2795 | QETH_DBF_TEXT(TRACE, 6, "flushbuf"); | ||
2796 | |||
2797 | for (i = index; i < index + count; ++i) { | ||
2798 | buf = &queue->bufs[i % QDIO_MAX_BUFFERS_PER_Q]; | ||
2799 | buf->buffer->element[buf->next_element_to_fill - 1].flags |= | ||
2800 | SBAL_FLAGS_LAST_ENTRY; | ||
2801 | |||
2802 | if (queue->card->info.type == QETH_CARD_TYPE_IQD) | ||
2803 | continue; | ||
2804 | |||
2805 | if (!queue->do_pack) { | ||
2806 | if ((atomic_read(&queue->used_buffers) >= | ||
2807 | (QETH_HIGH_WATERMARK_PACK - | ||
2808 | QETH_WATERMARK_PACK_FUZZ)) && | ||
2809 | !atomic_read(&queue->set_pci_flags_count)) { | ||
2810 | /* it's likely that we'll go to packing | ||
2811 | * mode soon */ | ||
2812 | atomic_inc(&queue->set_pci_flags_count); | ||
2813 | buf->buffer->element[0].flags |= 0x40; | ||
2814 | } | ||
2815 | } else { | ||
2816 | if (!atomic_read(&queue->set_pci_flags_count)) { | ||
2817 | /* | ||
2818 | * there's no outstanding PCI any more, so we | ||
2819 | * have to request a PCI to be sure the the PCI | ||
2820 | * will wake at some time in the future then we | ||
2821 | * can flush packed buffers that might still be | ||
2822 | * hanging around, which can happen if no | ||
2823 | * further send was requested by the stack | ||
2824 | */ | ||
2825 | atomic_inc(&queue->set_pci_flags_count); | ||
2826 | buf->buffer->element[0].flags |= 0x40; | ||
2827 | } | ||
2828 | } | ||
2829 | } | ||
2830 | |||
2831 | queue->card->dev->trans_start = jiffies; | ||
2832 | if (queue->card->options.performance_stats) { | ||
2833 | queue->card->perf_stats.outbound_do_qdio_cnt++; | ||
2834 | queue->card->perf_stats.outbound_do_qdio_start_time = | ||
2835 | qeth_get_micros(); | ||
2836 | } | ||
2837 | qdio_flags = QDIO_FLAG_SYNC_OUTPUT; | ||
2838 | if (under_int) | ||
2839 | qdio_flags |= QDIO_FLAG_UNDER_INTERRUPT; | ||
2840 | if (atomic_read(&queue->set_pci_flags_count)) | ||
2841 | qdio_flags |= QDIO_FLAG_PCI_OUT; | ||
2842 | rc = do_QDIO(CARD_DDEV(queue->card), qdio_flags, | ||
2843 | queue->queue_no, index, count, NULL); | ||
2844 | if (queue->card->options.performance_stats) | ||
2845 | queue->card->perf_stats.outbound_do_qdio_time += | ||
2846 | qeth_get_micros() - | ||
2847 | queue->card->perf_stats.outbound_do_qdio_start_time; | ||
2848 | if (rc) { | ||
2849 | QETH_DBF_TEXT(TRACE, 2, "flushbuf"); | ||
2850 | QETH_DBF_TEXT_(TRACE, 2, " err%d", rc); | ||
2851 | QETH_DBF_TEXT_(TRACE, 2, "%s", CARD_DDEV_ID(queue->card)); | ||
2852 | queue->card->stats.tx_errors += count; | ||
2853 | /* this must not happen under normal circumstances. if it | ||
2854 | * happens something is really wrong -> recover */ | ||
2855 | qeth_schedule_recovery(queue->card); | ||
2856 | return; | ||
2857 | } | ||
2858 | atomic_add(count, &queue->used_buffers); | ||
2859 | if (queue->card->options.performance_stats) | ||
2860 | queue->card->perf_stats.bufs_sent += count; | ||
2861 | } | ||
2862 | |||
2863 | static void qeth_check_outbound_queue(struct qeth_qdio_out_q *queue) | ||
2864 | { | ||
2865 | int index; | ||
2866 | int flush_cnt = 0; | ||
2867 | int q_was_packing = 0; | ||
2868 | |||
2869 | /* | ||
2870 | * check if weed have to switch to non-packing mode or if | ||
2871 | * we have to get a pci flag out on the queue | ||
2872 | */ | ||
2873 | if ((atomic_read(&queue->used_buffers) <= QETH_LOW_WATERMARK_PACK) || | ||
2874 | !atomic_read(&queue->set_pci_flags_count)) { | ||
2875 | if (atomic_xchg(&queue->state, QETH_OUT_Q_LOCKED_FLUSH) == | ||
2876 | QETH_OUT_Q_UNLOCKED) { | ||
2877 | /* | ||
2878 | * If we get in here, there was no action in | ||
2879 | * do_send_packet. So, we check if there is a | ||
2880 | * packing buffer to be flushed here. | ||
2881 | */ | ||
2882 | netif_stop_queue(queue->card->dev); | ||
2883 | index = queue->next_buf_to_fill; | ||
2884 | q_was_packing = queue->do_pack; | ||
2885 | /* queue->do_pack may change */ | ||
2886 | barrier(); | ||
2887 | flush_cnt += qeth_switch_to_nonpacking_if_needed(queue); | ||
2888 | if (!flush_cnt && | ||
2889 | !atomic_read(&queue->set_pci_flags_count)) | ||
2890 | flush_cnt += | ||
2891 | qeth_flush_buffers_on_no_pci(queue); | ||
2892 | if (queue->card->options.performance_stats && | ||
2893 | q_was_packing) | ||
2894 | queue->card->perf_stats.bufs_sent_pack += | ||
2895 | flush_cnt; | ||
2896 | if (flush_cnt) | ||
2897 | qeth_flush_buffers(queue, 1, index, flush_cnt); | ||
2898 | atomic_set(&queue->state, QETH_OUT_Q_UNLOCKED); | ||
2899 | } | ||
2900 | } | ||
2901 | } | ||
2902 | |||
2903 | void qeth_qdio_output_handler(struct ccw_device *ccwdev, unsigned int status, | ||
2904 | unsigned int qdio_error, unsigned int siga_error, | ||
2905 | unsigned int __queue, int first_element, int count, | ||
2906 | unsigned long card_ptr) | ||
2907 | { | ||
2908 | struct qeth_card *card = (struct qeth_card *) card_ptr; | ||
2909 | struct qeth_qdio_out_q *queue = card->qdio.out_qs[__queue]; | ||
2910 | struct qeth_qdio_out_buffer *buffer; | ||
2911 | int i; | ||
2912 | |||
2913 | QETH_DBF_TEXT(TRACE, 6, "qdouhdl"); | ||
2914 | if (status & QDIO_STATUS_LOOK_FOR_ERROR) { | ||
2915 | if (status & QDIO_STATUS_ACTIVATE_CHECK_CONDITION) { | ||
2916 | QETH_DBF_TEXT(TRACE, 2, "achkcond"); | ||
2917 | QETH_DBF_TEXT_(TRACE, 2, "%s", CARD_BUS_ID(card)); | ||
2918 | QETH_DBF_TEXT_(TRACE, 2, "%08x", status); | ||
2919 | netif_stop_queue(card->dev); | ||
2920 | qeth_schedule_recovery(card); | ||
2921 | return; | ||
2922 | } | ||
2923 | } | ||
2924 | if (card->options.performance_stats) { | ||
2925 | card->perf_stats.outbound_handler_cnt++; | ||
2926 | card->perf_stats.outbound_handler_start_time = | ||
2927 | qeth_get_micros(); | ||
2928 | } | ||
2929 | for (i = first_element; i < (first_element + count); ++i) { | ||
2930 | buffer = &queue->bufs[i % QDIO_MAX_BUFFERS_PER_Q]; | ||
2931 | /*we only handle the KICK_IT error by doing a recovery */ | ||
2932 | if (qeth_handle_send_error(card, buffer, | ||
2933 | qdio_error, siga_error) | ||
2934 | == QETH_SEND_ERROR_KICK_IT){ | ||
2935 | netif_stop_queue(card->dev); | ||
2936 | qeth_schedule_recovery(card); | ||
2937 | return; | ||
2938 | } | ||
2939 | qeth_clear_output_buffer(queue, buffer); | ||
2940 | } | ||
2941 | atomic_sub(count, &queue->used_buffers); | ||
2942 | /* check if we need to do something on this outbound queue */ | ||
2943 | if (card->info.type != QETH_CARD_TYPE_IQD) | ||
2944 | qeth_check_outbound_queue(queue); | ||
2945 | |||
2946 | netif_wake_queue(queue->card->dev); | ||
2947 | if (card->options.performance_stats) | ||
2948 | card->perf_stats.outbound_handler_time += qeth_get_micros() - | ||
2949 | card->perf_stats.outbound_handler_start_time; | ||
2950 | } | ||
2951 | EXPORT_SYMBOL_GPL(qeth_qdio_output_handler); | ||
2952 | |||
2953 | int qeth_get_cast_type(struct qeth_card *card, struct sk_buff *skb) | ||
2954 | { | ||
2955 | int cast_type = RTN_UNSPEC; | ||
2956 | |||
2957 | if (card->info.type == QETH_CARD_TYPE_OSN) | ||
2958 | return cast_type; | ||
2959 | |||
2960 | if (skb->dst && skb->dst->neighbour) { | ||
2961 | cast_type = skb->dst->neighbour->type; | ||
2962 | if ((cast_type == RTN_BROADCAST) || | ||
2963 | (cast_type == RTN_MULTICAST) || | ||
2964 | (cast_type == RTN_ANYCAST)) | ||
2965 | return cast_type; | ||
2966 | else | ||
2967 | return RTN_UNSPEC; | ||
2968 | } | ||
2969 | /* try something else */ | ||
2970 | if (skb->protocol == ETH_P_IPV6) | ||
2971 | return (skb_network_header(skb)[24] == 0xff) ? | ||
2972 | RTN_MULTICAST : 0; | ||
2973 | else if (skb->protocol == ETH_P_IP) | ||
2974 | return ((skb_network_header(skb)[16] & 0xf0) == 0xe0) ? | ||
2975 | RTN_MULTICAST : 0; | ||
2976 | /* ... */ | ||
2977 | if (!memcmp(skb->data, skb->dev->broadcast, 6)) | ||
2978 | return RTN_BROADCAST; | ||
2979 | else { | ||
2980 | u16 hdr_mac; | ||
2981 | |||
2982 | hdr_mac = *((u16 *)skb->data); | ||
2983 | /* tr multicast? */ | ||
2984 | switch (card->info.link_type) { | ||
2985 | case QETH_LINK_TYPE_HSTR: | ||
2986 | case QETH_LINK_TYPE_LANE_TR: | ||
2987 | if ((hdr_mac == QETH_TR_MAC_NC) || | ||
2988 | (hdr_mac == QETH_TR_MAC_C)) | ||
2989 | return RTN_MULTICAST; | ||
2990 | break; | ||
2991 | /* eth or so multicast? */ | ||
2992 | default: | ||
2993 | if ((hdr_mac == QETH_ETH_MAC_V4) || | ||
2994 | (hdr_mac == QETH_ETH_MAC_V6)) | ||
2995 | return RTN_MULTICAST; | ||
2996 | } | ||
2997 | } | ||
2998 | return cast_type; | ||
2999 | } | ||
3000 | EXPORT_SYMBOL_GPL(qeth_get_cast_type); | ||
3001 | |||
3002 | int qeth_get_priority_queue(struct qeth_card *card, struct sk_buff *skb, | ||
3003 | int ipv, int cast_type) | ||
3004 | { | ||
3005 | if (!ipv && (card->info.type == QETH_CARD_TYPE_OSAE)) | ||
3006 | return card->qdio.default_out_queue; | ||
3007 | switch (card->qdio.no_out_queues) { | ||
3008 | case 4: | ||
3009 | if (cast_type && card->info.is_multicast_different) | ||
3010 | return card->info.is_multicast_different & | ||
3011 | (card->qdio.no_out_queues - 1); | ||
3012 | if (card->qdio.do_prio_queueing && (ipv == 4)) { | ||
3013 | const u8 tos = ip_hdr(skb)->tos; | ||
3014 | |||
3015 | if (card->qdio.do_prio_queueing == | ||
3016 | QETH_PRIO_Q_ING_TOS) { | ||
3017 | if (tos & IP_TOS_NOTIMPORTANT) | ||
3018 | return 3; | ||
3019 | if (tos & IP_TOS_HIGHRELIABILITY) | ||
3020 | return 2; | ||
3021 | if (tos & IP_TOS_HIGHTHROUGHPUT) | ||
3022 | return 1; | ||
3023 | if (tos & IP_TOS_LOWDELAY) | ||
3024 | return 0; | ||
3025 | } | ||
3026 | if (card->qdio.do_prio_queueing == | ||
3027 | QETH_PRIO_Q_ING_PREC) | ||
3028 | return 3 - (tos >> 6); | ||
3029 | } else if (card->qdio.do_prio_queueing && (ipv == 6)) { | ||
3030 | /* TODO: IPv6!!! */ | ||
3031 | } | ||
3032 | return card->qdio.default_out_queue; | ||
3033 | case 1: /* fallthrough for single-out-queue 1920-device */ | ||
3034 | default: | ||
3035 | return card->qdio.default_out_queue; | ||
3036 | } | ||
3037 | } | ||
3038 | EXPORT_SYMBOL_GPL(qeth_get_priority_queue); | ||
3039 | |||
3040 | static void __qeth_free_new_skb(struct sk_buff *orig_skb, | ||
3041 | struct sk_buff *new_skb) | ||
3042 | { | ||
3043 | if (orig_skb != new_skb) | ||
3044 | dev_kfree_skb_any(new_skb); | ||
3045 | } | ||
3046 | |||
3047 | static inline struct sk_buff *qeth_realloc_headroom(struct qeth_card *card, | ||
3048 | struct sk_buff *skb, int size) | ||
3049 | { | ||
3050 | struct sk_buff *new_skb = skb; | ||
3051 | |||
3052 | if (skb_headroom(skb) >= size) | ||
3053 | return skb; | ||
3054 | new_skb = skb_realloc_headroom(skb, size); | ||
3055 | if (!new_skb) | ||
3056 | PRINT_ERR("Could not realloc headroom for qeth_hdr " | ||
3057 | "on interface %s", QETH_CARD_IFNAME(card)); | ||
3058 | return new_skb; | ||
3059 | } | ||
3060 | |||
3061 | struct sk_buff *qeth_prepare_skb(struct qeth_card *card, struct sk_buff *skb, | ||
3062 | struct qeth_hdr **hdr) | ||
3063 | { | ||
3064 | struct sk_buff *new_skb; | ||
3065 | |||
3066 | QETH_DBF_TEXT(TRACE, 6, "prepskb"); | ||
3067 | |||
3068 | new_skb = qeth_realloc_headroom(card, skb, | ||
3069 | sizeof(struct qeth_hdr)); | ||
3070 | if (!new_skb) | ||
3071 | return NULL; | ||
3072 | |||
3073 | *hdr = ((struct qeth_hdr *)qeth_push_skb(card, new_skb, | ||
3074 | sizeof(struct qeth_hdr))); | ||
3075 | if (*hdr == NULL) { | ||
3076 | __qeth_free_new_skb(skb, new_skb); | ||
3077 | return NULL; | ||
3078 | } | ||
3079 | return new_skb; | ||
3080 | } | ||
3081 | EXPORT_SYMBOL_GPL(qeth_prepare_skb); | ||
3082 | |||
3083 | int qeth_get_elements_no(struct qeth_card *card, void *hdr, | ||
3084 | struct sk_buff *skb, int elems) | ||
3085 | { | ||
3086 | int elements_needed = 0; | ||
3087 | |||
3088 | if (skb_shinfo(skb)->nr_frags > 0) | ||
3089 | elements_needed = (skb_shinfo(skb)->nr_frags + 1); | ||
3090 | if (elements_needed == 0) | ||
3091 | elements_needed = 1 + (((((unsigned long) hdr) % PAGE_SIZE) | ||
3092 | + skb->len) >> PAGE_SHIFT); | ||
3093 | if ((elements_needed + elems) > QETH_MAX_BUFFER_ELEMENTS(card)) { | ||
3094 | PRINT_ERR("Invalid size of IP packet " | ||
3095 | "(Number=%d / Length=%d). Discarded.\n", | ||
3096 | (elements_needed+elems), skb->len); | ||
3097 | return 0; | ||
3098 | } | ||
3099 | return elements_needed; | ||
3100 | } | ||
3101 | EXPORT_SYMBOL_GPL(qeth_get_elements_no); | ||
3102 | |||
3103 | static void __qeth_fill_buffer(struct sk_buff *skb, struct qdio_buffer *buffer, | ||
3104 | int is_tso, int *next_element_to_fill) | ||
3105 | { | ||
3106 | int length = skb->len; | ||
3107 | int length_here; | ||
3108 | int element; | ||
3109 | char *data; | ||
3110 | int first_lap ; | ||
3111 | |||
3112 | element = *next_element_to_fill; | ||
3113 | data = skb->data; | ||
3114 | first_lap = (is_tso == 0 ? 1 : 0); | ||
3115 | |||
3116 | while (length > 0) { | ||
3117 | /* length_here is the remaining amount of data in this page */ | ||
3118 | length_here = PAGE_SIZE - ((unsigned long) data % PAGE_SIZE); | ||
3119 | if (length < length_here) | ||
3120 | length_here = length; | ||
3121 | |||
3122 | buffer->element[element].addr = data; | ||
3123 | buffer->element[element].length = length_here; | ||
3124 | length -= length_here; | ||
3125 | if (!length) { | ||
3126 | if (first_lap) | ||
3127 | buffer->element[element].flags = 0; | ||
3128 | else | ||
3129 | buffer->element[element].flags = | ||
3130 | SBAL_FLAGS_LAST_FRAG; | ||
3131 | } else { | ||
3132 | if (first_lap) | ||
3133 | buffer->element[element].flags = | ||
3134 | SBAL_FLAGS_FIRST_FRAG; | ||
3135 | else | ||
3136 | buffer->element[element].flags = | ||
3137 | SBAL_FLAGS_MIDDLE_FRAG; | ||
3138 | } | ||
3139 | data += length_here; | ||
3140 | element++; | ||
3141 | first_lap = 0; | ||
3142 | } | ||
3143 | *next_element_to_fill = element; | ||
3144 | } | ||
3145 | |||
3146 | static int qeth_fill_buffer(struct qeth_qdio_out_q *queue, | ||
3147 | struct qeth_qdio_out_buffer *buf, struct sk_buff *skb) | ||
3148 | { | ||
3149 | struct qdio_buffer *buffer; | ||
3150 | struct qeth_hdr_tso *hdr; | ||
3151 | int flush_cnt = 0, hdr_len, large_send = 0; | ||
3152 | |||
3153 | QETH_DBF_TEXT(TRACE, 6, "qdfillbf"); | ||
3154 | |||
3155 | buffer = buf->buffer; | ||
3156 | atomic_inc(&skb->users); | ||
3157 | skb_queue_tail(&buf->skb_list, skb); | ||
3158 | |||
3159 | hdr = (struct qeth_hdr_tso *) skb->data; | ||
3160 | /*check first on TSO ....*/ | ||
3161 | if (hdr->hdr.hdr.l3.id == QETH_HEADER_TYPE_TSO) { | ||
3162 | int element = buf->next_element_to_fill; | ||
3163 | |||
3164 | hdr_len = sizeof(struct qeth_hdr_tso) + hdr->ext.dg_hdr_len; | ||
3165 | /*fill first buffer entry only with header information */ | ||
3166 | buffer->element[element].addr = skb->data; | ||
3167 | buffer->element[element].length = hdr_len; | ||
3168 | buffer->element[element].flags = SBAL_FLAGS_FIRST_FRAG; | ||
3169 | buf->next_element_to_fill++; | ||
3170 | skb->data += hdr_len; | ||
3171 | skb->len -= hdr_len; | ||
3172 | large_send = 1; | ||
3173 | } | ||
3174 | if (skb_shinfo(skb)->nr_frags == 0) | ||
3175 | __qeth_fill_buffer(skb, buffer, large_send, | ||
3176 | (int *)&buf->next_element_to_fill); | ||
3177 | else | ||
3178 | __qeth_fill_buffer_frag(skb, buffer, large_send, | ||
3179 | (int *)&buf->next_element_to_fill); | ||
3180 | |||
3181 | if (!queue->do_pack) { | ||
3182 | QETH_DBF_TEXT(TRACE, 6, "fillbfnp"); | ||
3183 | /* set state to PRIMED -> will be flushed */ | ||
3184 | atomic_set(&buf->state, QETH_QDIO_BUF_PRIMED); | ||
3185 | flush_cnt = 1; | ||
3186 | } else { | ||
3187 | QETH_DBF_TEXT(TRACE, 6, "fillbfpa"); | ||
3188 | if (queue->card->options.performance_stats) | ||
3189 | queue->card->perf_stats.skbs_sent_pack++; | ||
3190 | if (buf->next_element_to_fill >= | ||
3191 | QETH_MAX_BUFFER_ELEMENTS(queue->card)) { | ||
3192 | /* | ||
3193 | * packed buffer if full -> set state PRIMED | ||
3194 | * -> will be flushed | ||
3195 | */ | ||
3196 | atomic_set(&buf->state, QETH_QDIO_BUF_PRIMED); | ||
3197 | flush_cnt = 1; | ||
3198 | } | ||
3199 | } | ||
3200 | return flush_cnt; | ||
3201 | } | ||
3202 | |||
3203 | int qeth_do_send_packet_fast(struct qeth_card *card, | ||
3204 | struct qeth_qdio_out_q *queue, struct sk_buff *skb, | ||
3205 | struct qeth_hdr *hdr, int elements_needed, | ||
3206 | struct qeth_eddp_context *ctx) | ||
3207 | { | ||
3208 | struct qeth_qdio_out_buffer *buffer; | ||
3209 | int buffers_needed = 0; | ||
3210 | int flush_cnt = 0; | ||
3211 | int index; | ||
3212 | |||
3213 | QETH_DBF_TEXT(TRACE, 6, "dosndpfa"); | ||
3214 | |||
3215 | /* spin until we get the queue ... */ | ||
3216 | while (atomic_cmpxchg(&queue->state, QETH_OUT_Q_UNLOCKED, | ||
3217 | QETH_OUT_Q_LOCKED) != QETH_OUT_Q_UNLOCKED); | ||
3218 | /* ... now we've got the queue */ | ||
3219 | index = queue->next_buf_to_fill; | ||
3220 | buffer = &queue->bufs[queue->next_buf_to_fill]; | ||
3221 | /* | ||
3222 | * check if buffer is empty to make sure that we do not 'overtake' | ||
3223 | * ourselves and try to fill a buffer that is already primed | ||
3224 | */ | ||
3225 | if (atomic_read(&buffer->state) != QETH_QDIO_BUF_EMPTY) | ||
3226 | goto out; | ||
3227 | if (ctx == NULL) | ||
3228 | queue->next_buf_to_fill = (queue->next_buf_to_fill + 1) % | ||
3229 | QDIO_MAX_BUFFERS_PER_Q; | ||
3230 | else { | ||
3231 | buffers_needed = qeth_eddp_check_buffers_for_context(queue, | ||
3232 | ctx); | ||
3233 | if (buffers_needed < 0) | ||
3234 | goto out; | ||
3235 | queue->next_buf_to_fill = | ||
3236 | (queue->next_buf_to_fill + buffers_needed) % | ||
3237 | QDIO_MAX_BUFFERS_PER_Q; | ||
3238 | } | ||
3239 | atomic_set(&queue->state, QETH_OUT_Q_UNLOCKED); | ||
3240 | if (ctx == NULL) { | ||
3241 | qeth_fill_buffer(queue, buffer, skb); | ||
3242 | qeth_flush_buffers(queue, 0, index, 1); | ||
3243 | } else { | ||
3244 | flush_cnt = qeth_eddp_fill_buffer(queue, ctx, index); | ||
3245 | WARN_ON(buffers_needed != flush_cnt); | ||
3246 | qeth_flush_buffers(queue, 0, index, flush_cnt); | ||
3247 | } | ||
3248 | return 0; | ||
3249 | out: | ||
3250 | atomic_set(&queue->state, QETH_OUT_Q_UNLOCKED); | ||
3251 | return -EBUSY; | ||
3252 | } | ||
3253 | EXPORT_SYMBOL_GPL(qeth_do_send_packet_fast); | ||
3254 | |||
3255 | int qeth_do_send_packet(struct qeth_card *card, struct qeth_qdio_out_q *queue, | ||
3256 | struct sk_buff *skb, struct qeth_hdr *hdr, | ||
3257 | int elements_needed, struct qeth_eddp_context *ctx) | ||
3258 | { | ||
3259 | struct qeth_qdio_out_buffer *buffer; | ||
3260 | int start_index; | ||
3261 | int flush_count = 0; | ||
3262 | int do_pack = 0; | ||
3263 | int tmp; | ||
3264 | int rc = 0; | ||
3265 | |||
3266 | QETH_DBF_TEXT(TRACE, 6, "dosndpkt"); | ||
3267 | |||
3268 | /* spin until we get the queue ... */ | ||
3269 | while (atomic_cmpxchg(&queue->state, QETH_OUT_Q_UNLOCKED, | ||
3270 | QETH_OUT_Q_LOCKED) != QETH_OUT_Q_UNLOCKED); | ||
3271 | start_index = queue->next_buf_to_fill; | ||
3272 | buffer = &queue->bufs[queue->next_buf_to_fill]; | ||
3273 | /* | ||
3274 | * check if buffer is empty to make sure that we do not 'overtake' | ||
3275 | * ourselves and try to fill a buffer that is already primed | ||
3276 | */ | ||
3277 | if (atomic_read(&buffer->state) != QETH_QDIO_BUF_EMPTY) { | ||
3278 | atomic_set(&queue->state, QETH_OUT_Q_UNLOCKED); | ||
3279 | return -EBUSY; | ||
3280 | } | ||
3281 | /* check if we need to switch packing state of this queue */ | ||
3282 | qeth_switch_to_packing_if_needed(queue); | ||
3283 | if (queue->do_pack) { | ||
3284 | do_pack = 1; | ||
3285 | if (ctx == NULL) { | ||
3286 | /* does packet fit in current buffer? */ | ||
3287 | if ((QETH_MAX_BUFFER_ELEMENTS(card) - | ||
3288 | buffer->next_element_to_fill) < elements_needed) { | ||
3289 | /* ... no -> set state PRIMED */ | ||
3290 | atomic_set(&buffer->state, | ||
3291 | QETH_QDIO_BUF_PRIMED); | ||
3292 | flush_count++; | ||
3293 | queue->next_buf_to_fill = | ||
3294 | (queue->next_buf_to_fill + 1) % | ||
3295 | QDIO_MAX_BUFFERS_PER_Q; | ||
3296 | buffer = &queue->bufs[queue->next_buf_to_fill]; | ||
3297 | /* we did a step forward, so check buffer state | ||
3298 | * again */ | ||
3299 | if (atomic_read(&buffer->state) != | ||
3300 | QETH_QDIO_BUF_EMPTY){ | ||
3301 | qeth_flush_buffers(queue, 0, | ||
3302 | start_index, flush_count); | ||
3303 | atomic_set(&queue->state, | ||
3304 | QETH_OUT_Q_UNLOCKED); | ||
3305 | return -EBUSY; | ||
3306 | } | ||
3307 | } | ||
3308 | } else { | ||
3309 | /* check if we have enough elements (including following | ||
3310 | * free buffers) to handle eddp context */ | ||
3311 | if (qeth_eddp_check_buffers_for_context(queue, ctx) | ||
3312 | < 0) { | ||
3313 | if (net_ratelimit()) | ||
3314 | PRINT_WARN("eddp tx_dropped 1\n"); | ||
3315 | rc = -EBUSY; | ||
3316 | goto out; | ||
3317 | } | ||
3318 | } | ||
3319 | } | ||
3320 | if (ctx == NULL) | ||
3321 | tmp = qeth_fill_buffer(queue, buffer, skb); | ||
3322 | else { | ||
3323 | tmp = qeth_eddp_fill_buffer(queue, ctx, | ||
3324 | queue->next_buf_to_fill); | ||
3325 | if (tmp < 0) { | ||
3326 | PRINT_ERR("eddp tx_dropped 2\n"); | ||
3327 | rc = -EBUSY; | ||
3328 | goto out; | ||
3329 | } | ||
3330 | } | ||
3331 | queue->next_buf_to_fill = (queue->next_buf_to_fill + tmp) % | ||
3332 | QDIO_MAX_BUFFERS_PER_Q; | ||
3333 | flush_count += tmp; | ||
3334 | out: | ||
3335 | if (flush_count) | ||
3336 | qeth_flush_buffers(queue, 0, start_index, flush_count); | ||
3337 | else if (!atomic_read(&queue->set_pci_flags_count)) | ||
3338 | atomic_xchg(&queue->state, QETH_OUT_Q_LOCKED_FLUSH); | ||
3339 | /* | ||
3340 | * queue->state will go from LOCKED -> UNLOCKED or from | ||
3341 | * LOCKED_FLUSH -> LOCKED if output_handler wanted to 'notify' us | ||
3342 | * (switch packing state or flush buffer to get another pci flag out). | ||
3343 | * In that case we will enter this loop | ||
3344 | */ | ||
3345 | while (atomic_dec_return(&queue->state)) { | ||
3346 | flush_count = 0; | ||
3347 | start_index = queue->next_buf_to_fill; | ||
3348 | /* check if we can go back to non-packing state */ | ||
3349 | flush_count += qeth_switch_to_nonpacking_if_needed(queue); | ||
3350 | /* | ||
3351 | * check if we need to flush a packing buffer to get a pci | ||
3352 | * flag out on the queue | ||
3353 | */ | ||
3354 | if (!flush_count && !atomic_read(&queue->set_pci_flags_count)) | ||
3355 | flush_count += qeth_flush_buffers_on_no_pci(queue); | ||
3356 | if (flush_count) | ||
3357 | qeth_flush_buffers(queue, 0, start_index, flush_count); | ||
3358 | } | ||
3359 | /* at this point the queue is UNLOCKED again */ | ||
3360 | if (queue->card->options.performance_stats && do_pack) | ||
3361 | queue->card->perf_stats.bufs_sent_pack += flush_count; | ||
3362 | |||
3363 | return rc; | ||
3364 | } | ||
3365 | EXPORT_SYMBOL_GPL(qeth_do_send_packet); | ||
3366 | |||
3367 | static int qeth_setadp_promisc_mode_cb(struct qeth_card *card, | ||
3368 | struct qeth_reply *reply, unsigned long data) | ||
3369 | { | ||
3370 | struct qeth_ipa_cmd *cmd; | ||
3371 | struct qeth_ipacmd_setadpparms *setparms; | ||
3372 | |||
3373 | QETH_DBF_TEXT(TRACE, 4, "prmadpcb"); | ||
3374 | |||
3375 | cmd = (struct qeth_ipa_cmd *) data; | ||
3376 | setparms = &(cmd->data.setadapterparms); | ||
3377 | |||
3378 | qeth_default_setadapterparms_cb(card, reply, (unsigned long)cmd); | ||
3379 | if (cmd->hdr.return_code) { | ||
3380 | QETH_DBF_TEXT_(TRACE, 4, "prmrc%2.2x", cmd->hdr.return_code); | ||
3381 | setparms->data.mode = SET_PROMISC_MODE_OFF; | ||
3382 | } | ||
3383 | card->info.promisc_mode = setparms->data.mode; | ||
3384 | return 0; | ||
3385 | } | ||
3386 | |||
3387 | void qeth_setadp_promisc_mode(struct qeth_card *card) | ||
3388 | { | ||
3389 | enum qeth_ipa_promisc_modes mode; | ||
3390 | struct net_device *dev = card->dev; | ||
3391 | struct qeth_cmd_buffer *iob; | ||
3392 | struct qeth_ipa_cmd *cmd; | ||
3393 | |||
3394 | QETH_DBF_TEXT(TRACE, 4, "setprom"); | ||
3395 | |||
3396 | if (((dev->flags & IFF_PROMISC) && | ||
3397 | (card->info.promisc_mode == SET_PROMISC_MODE_ON)) || | ||
3398 | (!(dev->flags & IFF_PROMISC) && | ||
3399 | (card->info.promisc_mode == SET_PROMISC_MODE_OFF))) | ||
3400 | return; | ||
3401 | mode = SET_PROMISC_MODE_OFF; | ||
3402 | if (dev->flags & IFF_PROMISC) | ||
3403 | mode = SET_PROMISC_MODE_ON; | ||
3404 | QETH_DBF_TEXT_(TRACE, 4, "mode:%x", mode); | ||
3405 | |||
3406 | iob = qeth_get_adapter_cmd(card, IPA_SETADP_SET_PROMISC_MODE, | ||
3407 | sizeof(struct qeth_ipacmd_setadpparms)); | ||
3408 | cmd = (struct qeth_ipa_cmd *)(iob->data + IPA_PDU_HEADER_SIZE); | ||
3409 | cmd->data.setadapterparms.data.mode = mode; | ||
3410 | qeth_send_ipa_cmd(card, iob, qeth_setadp_promisc_mode_cb, NULL); | ||
3411 | } | ||
3412 | EXPORT_SYMBOL_GPL(qeth_setadp_promisc_mode); | ||
3413 | |||
3414 | int qeth_change_mtu(struct net_device *dev, int new_mtu) | ||
3415 | { | ||
3416 | struct qeth_card *card; | ||
3417 | char dbf_text[15]; | ||
3418 | |||
3419 | card = netdev_priv(dev); | ||
3420 | |||
3421 | QETH_DBF_TEXT(TRACE, 4, "chgmtu"); | ||
3422 | sprintf(dbf_text, "%8x", new_mtu); | ||
3423 | QETH_DBF_TEXT(TRACE, 4, dbf_text); | ||
3424 | |||
3425 | if (new_mtu < 64) | ||
3426 | return -EINVAL; | ||
3427 | if (new_mtu > 65535) | ||
3428 | return -EINVAL; | ||
3429 | if ((!qeth_is_supported(card, IPA_IP_FRAGMENTATION)) && | ||
3430 | (!qeth_mtu_is_valid(card, new_mtu))) | ||
3431 | return -EINVAL; | ||
3432 | dev->mtu = new_mtu; | ||
3433 | return 0; | ||
3434 | } | ||
3435 | EXPORT_SYMBOL_GPL(qeth_change_mtu); | ||
3436 | |||
3437 | struct net_device_stats *qeth_get_stats(struct net_device *dev) | ||
3438 | { | ||
3439 | struct qeth_card *card; | ||
3440 | |||
3441 | card = netdev_priv(dev); | ||
3442 | |||
3443 | QETH_DBF_TEXT(TRACE, 5, "getstat"); | ||
3444 | |||
3445 | return &card->stats; | ||
3446 | } | ||
3447 | EXPORT_SYMBOL_GPL(qeth_get_stats); | ||
3448 | |||
3449 | static int qeth_setadpparms_change_macaddr_cb(struct qeth_card *card, | ||
3450 | struct qeth_reply *reply, unsigned long data) | ||
3451 | { | ||
3452 | struct qeth_ipa_cmd *cmd; | ||
3453 | |||
3454 | QETH_DBF_TEXT(TRACE, 4, "chgmaccb"); | ||
3455 | |||
3456 | cmd = (struct qeth_ipa_cmd *) data; | ||
3457 | if (!card->options.layer2 || | ||
3458 | !(card->info.mac_bits & QETH_LAYER2_MAC_READ)) { | ||
3459 | memcpy(card->dev->dev_addr, | ||
3460 | &cmd->data.setadapterparms.data.change_addr.addr, | ||
3461 | OSA_ADDR_LEN); | ||
3462 | card->info.mac_bits |= QETH_LAYER2_MAC_READ; | ||
3463 | } | ||
3464 | qeth_default_setadapterparms_cb(card, reply, (unsigned long) cmd); | ||
3465 | return 0; | ||
3466 | } | ||
3467 | |||
3468 | int qeth_setadpparms_change_macaddr(struct qeth_card *card) | ||
3469 | { | ||
3470 | int rc; | ||
3471 | struct qeth_cmd_buffer *iob; | ||
3472 | struct qeth_ipa_cmd *cmd; | ||
3473 | |||
3474 | QETH_DBF_TEXT(TRACE, 4, "chgmac"); | ||
3475 | |||
3476 | iob = qeth_get_adapter_cmd(card, IPA_SETADP_ALTER_MAC_ADDRESS, | ||
3477 | sizeof(struct qeth_ipacmd_setadpparms)); | ||
3478 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
3479 | cmd->data.setadapterparms.data.change_addr.cmd = CHANGE_ADDR_READ_MAC; | ||
3480 | cmd->data.setadapterparms.data.change_addr.addr_size = OSA_ADDR_LEN; | ||
3481 | memcpy(&cmd->data.setadapterparms.data.change_addr.addr, | ||
3482 | card->dev->dev_addr, OSA_ADDR_LEN); | ||
3483 | rc = qeth_send_ipa_cmd(card, iob, qeth_setadpparms_change_macaddr_cb, | ||
3484 | NULL); | ||
3485 | return rc; | ||
3486 | } | ||
3487 | EXPORT_SYMBOL_GPL(qeth_setadpparms_change_macaddr); | ||
3488 | |||
3489 | void qeth_tx_timeout(struct net_device *dev) | ||
3490 | { | ||
3491 | struct qeth_card *card; | ||
3492 | |||
3493 | card = netdev_priv(dev); | ||
3494 | card->stats.tx_errors++; | ||
3495 | qeth_schedule_recovery(card); | ||
3496 | } | ||
3497 | EXPORT_SYMBOL_GPL(qeth_tx_timeout); | ||
3498 | |||
3499 | int qeth_mdio_read(struct net_device *dev, int phy_id, int regnum) | ||
3500 | { | ||
3501 | struct qeth_card *card = netdev_priv(dev); | ||
3502 | int rc = 0; | ||
3503 | |||
3504 | switch (regnum) { | ||
3505 | case MII_BMCR: /* Basic mode control register */ | ||
3506 | rc = BMCR_FULLDPLX; | ||
3507 | if ((card->info.link_type != QETH_LINK_TYPE_GBIT_ETH) && | ||
3508 | (card->info.link_type != QETH_LINK_TYPE_OSN) && | ||
3509 | (card->info.link_type != QETH_LINK_TYPE_10GBIT_ETH)) | ||
3510 | rc |= BMCR_SPEED100; | ||
3511 | break; | ||
3512 | case MII_BMSR: /* Basic mode status register */ | ||
3513 | rc = BMSR_ERCAP | BMSR_ANEGCOMPLETE | BMSR_LSTATUS | | ||
3514 | BMSR_10HALF | BMSR_10FULL | BMSR_100HALF | BMSR_100FULL | | ||
3515 | BMSR_100BASE4; | ||
3516 | break; | ||
3517 | case MII_PHYSID1: /* PHYS ID 1 */ | ||
3518 | rc = (dev->dev_addr[0] << 16) | (dev->dev_addr[1] << 8) | | ||
3519 | dev->dev_addr[2]; | ||
3520 | rc = (rc >> 5) & 0xFFFF; | ||
3521 | break; | ||
3522 | case MII_PHYSID2: /* PHYS ID 2 */ | ||
3523 | rc = (dev->dev_addr[2] << 10) & 0xFFFF; | ||
3524 | break; | ||
3525 | case MII_ADVERTISE: /* Advertisement control reg */ | ||
3526 | rc = ADVERTISE_ALL; | ||
3527 | break; | ||
3528 | case MII_LPA: /* Link partner ability reg */ | ||
3529 | rc = LPA_10HALF | LPA_10FULL | LPA_100HALF | LPA_100FULL | | ||
3530 | LPA_100BASE4 | LPA_LPACK; | ||
3531 | break; | ||
3532 | case MII_EXPANSION: /* Expansion register */ | ||
3533 | break; | ||
3534 | case MII_DCOUNTER: /* disconnect counter */ | ||
3535 | break; | ||
3536 | case MII_FCSCOUNTER: /* false carrier counter */ | ||
3537 | break; | ||
3538 | case MII_NWAYTEST: /* N-way auto-neg test register */ | ||
3539 | break; | ||
3540 | case MII_RERRCOUNTER: /* rx error counter */ | ||
3541 | rc = card->stats.rx_errors; | ||
3542 | break; | ||
3543 | case MII_SREVISION: /* silicon revision */ | ||
3544 | break; | ||
3545 | case MII_RESV1: /* reserved 1 */ | ||
3546 | break; | ||
3547 | case MII_LBRERROR: /* loopback, rx, bypass error */ | ||
3548 | break; | ||
3549 | case MII_PHYADDR: /* physical address */ | ||
3550 | break; | ||
3551 | case MII_RESV2: /* reserved 2 */ | ||
3552 | break; | ||
3553 | case MII_TPISTATUS: /* TPI status for 10mbps */ | ||
3554 | break; | ||
3555 | case MII_NCONFIG: /* network interface config */ | ||
3556 | break; | ||
3557 | default: | ||
3558 | break; | ||
3559 | } | ||
3560 | return rc; | ||
3561 | } | ||
3562 | EXPORT_SYMBOL_GPL(qeth_mdio_read); | ||
3563 | |||
3564 | static int qeth_send_ipa_snmp_cmd(struct qeth_card *card, | ||
3565 | struct qeth_cmd_buffer *iob, int len, | ||
3566 | int (*reply_cb)(struct qeth_card *, struct qeth_reply *, | ||
3567 | unsigned long), | ||
3568 | void *reply_param) | ||
3569 | { | ||
3570 | u16 s1, s2; | ||
3571 | |||
3572 | QETH_DBF_TEXT(TRACE, 4, "sendsnmp"); | ||
3573 | |||
3574 | memcpy(iob->data, IPA_PDU_HEADER, IPA_PDU_HEADER_SIZE); | ||
3575 | memcpy(QETH_IPA_CMD_DEST_ADDR(iob->data), | ||
3576 | &card->token.ulp_connection_r, QETH_MPC_TOKEN_LENGTH); | ||
3577 | /* adjust PDU length fields in IPA_PDU_HEADER */ | ||
3578 | s1 = (u32) IPA_PDU_HEADER_SIZE + len; | ||
3579 | s2 = (u32) len; | ||
3580 | memcpy(QETH_IPA_PDU_LEN_TOTAL(iob->data), &s1, 2); | ||
3581 | memcpy(QETH_IPA_PDU_LEN_PDU1(iob->data), &s2, 2); | ||
3582 | memcpy(QETH_IPA_PDU_LEN_PDU2(iob->data), &s2, 2); | ||
3583 | memcpy(QETH_IPA_PDU_LEN_PDU3(iob->data), &s2, 2); | ||
3584 | return qeth_send_control_data(card, IPA_PDU_HEADER_SIZE + len, iob, | ||
3585 | reply_cb, reply_param); | ||
3586 | } | ||
3587 | |||
3588 | static int qeth_snmp_command_cb(struct qeth_card *card, | ||
3589 | struct qeth_reply *reply, unsigned long sdata) | ||
3590 | { | ||
3591 | struct qeth_ipa_cmd *cmd; | ||
3592 | struct qeth_arp_query_info *qinfo; | ||
3593 | struct qeth_snmp_cmd *snmp; | ||
3594 | unsigned char *data; | ||
3595 | __u16 data_len; | ||
3596 | |||
3597 | QETH_DBF_TEXT(TRACE, 3, "snpcmdcb"); | ||
3598 | |||
3599 | cmd = (struct qeth_ipa_cmd *) sdata; | ||
3600 | data = (unsigned char *)((char *)cmd - reply->offset); | ||
3601 | qinfo = (struct qeth_arp_query_info *) reply->param; | ||
3602 | snmp = &cmd->data.setadapterparms.data.snmp; | ||
3603 | |||
3604 | if (cmd->hdr.return_code) { | ||
3605 | QETH_DBF_TEXT_(TRACE, 4, "scer1%i", cmd->hdr.return_code); | ||
3606 | return 0; | ||
3607 | } | ||
3608 | if (cmd->data.setadapterparms.hdr.return_code) { | ||
3609 | cmd->hdr.return_code = | ||
3610 | cmd->data.setadapterparms.hdr.return_code; | ||
3611 | QETH_DBF_TEXT_(TRACE, 4, "scer2%i", cmd->hdr.return_code); | ||
3612 | return 0; | ||
3613 | } | ||
3614 | data_len = *((__u16 *)QETH_IPA_PDU_LEN_PDU1(data)); | ||
3615 | if (cmd->data.setadapterparms.hdr.seq_no == 1) | ||
3616 | data_len -= (__u16)((char *)&snmp->data - (char *)cmd); | ||
3617 | else | ||
3618 | data_len -= (__u16)((char *)&snmp->request - (char *)cmd); | ||
3619 | |||
3620 | /* check if there is enough room in userspace */ | ||
3621 | if ((qinfo->udata_len - qinfo->udata_offset) < data_len) { | ||
3622 | QETH_DBF_TEXT_(TRACE, 4, "scer3%i", -ENOMEM); | ||
3623 | cmd->hdr.return_code = -ENOMEM; | ||
3624 | return 0; | ||
3625 | } | ||
3626 | QETH_DBF_TEXT_(TRACE, 4, "snore%i", | ||
3627 | cmd->data.setadapterparms.hdr.used_total); | ||
3628 | QETH_DBF_TEXT_(TRACE, 4, "sseqn%i", | ||
3629 | cmd->data.setadapterparms.hdr.seq_no); | ||
3630 | /*copy entries to user buffer*/ | ||
3631 | if (cmd->data.setadapterparms.hdr.seq_no == 1) { | ||
3632 | memcpy(qinfo->udata + qinfo->udata_offset, | ||
3633 | (char *)snmp, | ||
3634 | data_len + offsetof(struct qeth_snmp_cmd, data)); | ||
3635 | qinfo->udata_offset += offsetof(struct qeth_snmp_cmd, data); | ||
3636 | } else { | ||
3637 | memcpy(qinfo->udata + qinfo->udata_offset, | ||
3638 | (char *)&snmp->request, data_len); | ||
3639 | } | ||
3640 | qinfo->udata_offset += data_len; | ||
3641 | /* check if all replies received ... */ | ||
3642 | QETH_DBF_TEXT_(TRACE, 4, "srtot%i", | ||
3643 | cmd->data.setadapterparms.hdr.used_total); | ||
3644 | QETH_DBF_TEXT_(TRACE, 4, "srseq%i", | ||
3645 | cmd->data.setadapterparms.hdr.seq_no); | ||
3646 | if (cmd->data.setadapterparms.hdr.seq_no < | ||
3647 | cmd->data.setadapterparms.hdr.used_total) | ||
3648 | return 1; | ||
3649 | return 0; | ||
3650 | } | ||
3651 | |||
3652 | int qeth_snmp_command(struct qeth_card *card, char __user *udata) | ||
3653 | { | ||
3654 | struct qeth_cmd_buffer *iob; | ||
3655 | struct qeth_ipa_cmd *cmd; | ||
3656 | struct qeth_snmp_ureq *ureq; | ||
3657 | int req_len; | ||
3658 | struct qeth_arp_query_info qinfo = {0, }; | ||
3659 | int rc = 0; | ||
3660 | |||
3661 | QETH_DBF_TEXT(TRACE, 3, "snmpcmd"); | ||
3662 | |||
3663 | if (card->info.guestlan) | ||
3664 | return -EOPNOTSUPP; | ||
3665 | |||
3666 | if ((!qeth_adp_supported(card, IPA_SETADP_SET_SNMP_CONTROL)) && | ||
3667 | (!card->options.layer2)) { | ||
3668 | PRINT_WARN("SNMP Query MIBS not supported " | ||
3669 | "on %s!\n", QETH_CARD_IFNAME(card)); | ||
3670 | return -EOPNOTSUPP; | ||
3671 | } | ||
3672 | /* skip 4 bytes (data_len struct member) to get req_len */ | ||
3673 | if (copy_from_user(&req_len, udata + sizeof(int), sizeof(int))) | ||
3674 | return -EFAULT; | ||
3675 | ureq = kmalloc(req_len+sizeof(struct qeth_snmp_ureq_hdr), GFP_KERNEL); | ||
3676 | if (!ureq) { | ||
3677 | QETH_DBF_TEXT(TRACE, 2, "snmpnome"); | ||
3678 | return -ENOMEM; | ||
3679 | } | ||
3680 | if (copy_from_user(ureq, udata, | ||
3681 | req_len + sizeof(struct qeth_snmp_ureq_hdr))) { | ||
3682 | kfree(ureq); | ||
3683 | return -EFAULT; | ||
3684 | } | ||
3685 | qinfo.udata_len = ureq->hdr.data_len; | ||
3686 | qinfo.udata = kzalloc(qinfo.udata_len, GFP_KERNEL); | ||
3687 | if (!qinfo.udata) { | ||
3688 | kfree(ureq); | ||
3689 | return -ENOMEM; | ||
3690 | } | ||
3691 | qinfo.udata_offset = sizeof(struct qeth_snmp_ureq_hdr); | ||
3692 | |||
3693 | iob = qeth_get_adapter_cmd(card, IPA_SETADP_SET_SNMP_CONTROL, | ||
3694 | QETH_SNMP_SETADP_CMDLENGTH + req_len); | ||
3695 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
3696 | memcpy(&cmd->data.setadapterparms.data.snmp, &ureq->cmd, req_len); | ||
3697 | rc = qeth_send_ipa_snmp_cmd(card, iob, QETH_SETADP_BASE_LEN + req_len, | ||
3698 | qeth_snmp_command_cb, (void *)&qinfo); | ||
3699 | if (rc) | ||
3700 | PRINT_WARN("SNMP command failed on %s: (0x%x)\n", | ||
3701 | QETH_CARD_IFNAME(card), rc); | ||
3702 | else { | ||
3703 | if (copy_to_user(udata, qinfo.udata, qinfo.udata_len)) | ||
3704 | rc = -EFAULT; | ||
3705 | } | ||
3706 | |||
3707 | kfree(ureq); | ||
3708 | kfree(qinfo.udata); | ||
3709 | return rc; | ||
3710 | } | ||
3711 | EXPORT_SYMBOL_GPL(qeth_snmp_command); | ||
3712 | |||
3713 | static inline int qeth_get_qdio_q_format(struct qeth_card *card) | ||
3714 | { | ||
3715 | switch (card->info.type) { | ||
3716 | case QETH_CARD_TYPE_IQD: | ||
3717 | return 2; | ||
3718 | default: | ||
3719 | return 0; | ||
3720 | } | ||
3721 | } | ||
3722 | |||
3723 | static int qeth_qdio_establish(struct qeth_card *card) | ||
3724 | { | ||
3725 | struct qdio_initialize init_data; | ||
3726 | char *qib_param_field; | ||
3727 | struct qdio_buffer **in_sbal_ptrs; | ||
3728 | struct qdio_buffer **out_sbal_ptrs; | ||
3729 | int i, j, k; | ||
3730 | int rc = 0; | ||
3731 | |||
3732 | QETH_DBF_TEXT(SETUP, 2, "qdioest"); | ||
3733 | |||
3734 | qib_param_field = kzalloc(QDIO_MAX_BUFFERS_PER_Q * sizeof(char), | ||
3735 | GFP_KERNEL); | ||
3736 | if (!qib_param_field) | ||
3737 | return -ENOMEM; | ||
3738 | |||
3739 | qeth_create_qib_param_field(card, qib_param_field); | ||
3740 | qeth_create_qib_param_field_blkt(card, qib_param_field); | ||
3741 | |||
3742 | in_sbal_ptrs = kmalloc(QDIO_MAX_BUFFERS_PER_Q * sizeof(void *), | ||
3743 | GFP_KERNEL); | ||
3744 | if (!in_sbal_ptrs) { | ||
3745 | kfree(qib_param_field); | ||
3746 | return -ENOMEM; | ||
3747 | } | ||
3748 | for (i = 0; i < QDIO_MAX_BUFFERS_PER_Q; ++i) | ||
3749 | in_sbal_ptrs[i] = (struct qdio_buffer *) | ||
3750 | virt_to_phys(card->qdio.in_q->bufs[i].buffer); | ||
3751 | |||
3752 | out_sbal_ptrs = | ||
3753 | kmalloc(card->qdio.no_out_queues * QDIO_MAX_BUFFERS_PER_Q * | ||
3754 | sizeof(void *), GFP_KERNEL); | ||
3755 | if (!out_sbal_ptrs) { | ||
3756 | kfree(in_sbal_ptrs); | ||
3757 | kfree(qib_param_field); | ||
3758 | return -ENOMEM; | ||
3759 | } | ||
3760 | for (i = 0, k = 0; i < card->qdio.no_out_queues; ++i) | ||
3761 | for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j, ++k) { | ||
3762 | out_sbal_ptrs[k] = (struct qdio_buffer *)virt_to_phys( | ||
3763 | card->qdio.out_qs[i]->bufs[j].buffer); | ||
3764 | } | ||
3765 | |||
3766 | memset(&init_data, 0, sizeof(struct qdio_initialize)); | ||
3767 | init_data.cdev = CARD_DDEV(card); | ||
3768 | init_data.q_format = qeth_get_qdio_q_format(card); | ||
3769 | init_data.qib_param_field_format = 0; | ||
3770 | init_data.qib_param_field = qib_param_field; | ||
3771 | init_data.min_input_threshold = QETH_MIN_INPUT_THRESHOLD; | ||
3772 | init_data.max_input_threshold = QETH_MAX_INPUT_THRESHOLD; | ||
3773 | init_data.min_output_threshold = QETH_MIN_OUTPUT_THRESHOLD; | ||
3774 | init_data.max_output_threshold = QETH_MAX_OUTPUT_THRESHOLD; | ||
3775 | init_data.no_input_qs = 1; | ||
3776 | init_data.no_output_qs = card->qdio.no_out_queues; | ||
3777 | init_data.input_handler = card->discipline.input_handler; | ||
3778 | init_data.output_handler = card->discipline.output_handler; | ||
3779 | init_data.int_parm = (unsigned long) card; | ||
3780 | init_data.flags = QDIO_INBOUND_0COPY_SBALS | | ||
3781 | QDIO_OUTBOUND_0COPY_SBALS | | ||
3782 | QDIO_USE_OUTBOUND_PCIS; | ||
3783 | init_data.input_sbal_addr_array = (void **) in_sbal_ptrs; | ||
3784 | init_data.output_sbal_addr_array = (void **) out_sbal_ptrs; | ||
3785 | |||
3786 | if (atomic_cmpxchg(&card->qdio.state, QETH_QDIO_ALLOCATED, | ||
3787 | QETH_QDIO_ESTABLISHED) == QETH_QDIO_ALLOCATED) { | ||
3788 | rc = qdio_initialize(&init_data); | ||
3789 | if (rc) | ||
3790 | atomic_set(&card->qdio.state, QETH_QDIO_ALLOCATED); | ||
3791 | } | ||
3792 | kfree(out_sbal_ptrs); | ||
3793 | kfree(in_sbal_ptrs); | ||
3794 | kfree(qib_param_field); | ||
3795 | return rc; | ||
3796 | } | ||
3797 | |||
3798 | static void qeth_core_free_card(struct qeth_card *card) | ||
3799 | { | ||
3800 | |||
3801 | QETH_DBF_TEXT(SETUP, 2, "freecrd"); | ||
3802 | QETH_DBF_HEX(SETUP, 2, &card, sizeof(void *)); | ||
3803 | qeth_clean_channel(&card->read); | ||
3804 | qeth_clean_channel(&card->write); | ||
3805 | if (card->dev) | ||
3806 | free_netdev(card->dev); | ||
3807 | kfree(card->ip_tbd_list); | ||
3808 | qeth_free_qdio_buffers(card); | ||
3809 | kfree(card); | ||
3810 | } | ||
3811 | |||
3812 | static struct ccw_device_id qeth_ids[] = { | ||
3813 | {CCW_DEVICE(0x1731, 0x01), .driver_info = QETH_CARD_TYPE_OSAE}, | ||
3814 | {CCW_DEVICE(0x1731, 0x05), .driver_info = QETH_CARD_TYPE_IQD}, | ||
3815 | {CCW_DEVICE(0x1731, 0x06), .driver_info = QETH_CARD_TYPE_OSN}, | ||
3816 | {}, | ||
3817 | }; | ||
3818 | MODULE_DEVICE_TABLE(ccw, qeth_ids); | ||
3819 | |||
3820 | static struct ccw_driver qeth_ccw_driver = { | ||
3821 | .name = "qeth", | ||
3822 | .ids = qeth_ids, | ||
3823 | .probe = ccwgroup_probe_ccwdev, | ||
3824 | .remove = ccwgroup_remove_ccwdev, | ||
3825 | }; | ||
3826 | |||
3827 | static int qeth_core_driver_group(const char *buf, struct device *root_dev, | ||
3828 | unsigned long driver_id) | ||
3829 | { | ||
3830 | const char *start, *end; | ||
3831 | char bus_ids[3][BUS_ID_SIZE], *argv[3]; | ||
3832 | int i; | ||
3833 | |||
3834 | start = buf; | ||
3835 | for (i = 0; i < 3; i++) { | ||
3836 | static const char delim[] = { ',', ',', '\n' }; | ||
3837 | int len; | ||
3838 | |||
3839 | end = strchr(start, delim[i]); | ||
3840 | if (!end) | ||
3841 | return -EINVAL; | ||
3842 | len = min_t(ptrdiff_t, BUS_ID_SIZE, end - start); | ||
3843 | strncpy(bus_ids[i], start, len); | ||
3844 | bus_ids[i][len] = '\0'; | ||
3845 | start = end + 1; | ||
3846 | argv[i] = bus_ids[i]; | ||
3847 | } | ||
3848 | |||
3849 | return (ccwgroup_create(root_dev, driver_id, | ||
3850 | &qeth_ccw_driver, 3, argv)); | ||
3851 | } | ||
3852 | |||
3853 | int qeth_core_hardsetup_card(struct qeth_card *card) | ||
3854 | { | ||
3855 | int retries = 3; | ||
3856 | int mpno; | ||
3857 | int rc; | ||
3858 | |||
3859 | QETH_DBF_TEXT(SETUP, 2, "hrdsetup"); | ||
3860 | atomic_set(&card->force_alloc_skb, 0); | ||
3861 | retry: | ||
3862 | if (retries < 3) { | ||
3863 | PRINT_WARN("Retrying to do IDX activates.\n"); | ||
3864 | ccw_device_set_offline(CARD_DDEV(card)); | ||
3865 | ccw_device_set_offline(CARD_WDEV(card)); | ||
3866 | ccw_device_set_offline(CARD_RDEV(card)); | ||
3867 | ccw_device_set_online(CARD_RDEV(card)); | ||
3868 | ccw_device_set_online(CARD_WDEV(card)); | ||
3869 | ccw_device_set_online(CARD_DDEV(card)); | ||
3870 | } | ||
3871 | rc = qeth_qdio_clear_card(card, card->info.type != QETH_CARD_TYPE_IQD); | ||
3872 | if (rc == -ERESTARTSYS) { | ||
3873 | QETH_DBF_TEXT(SETUP, 2, "break1"); | ||
3874 | return rc; | ||
3875 | } else if (rc) { | ||
3876 | QETH_DBF_TEXT_(SETUP, 2, "1err%d", rc); | ||
3877 | if (--retries < 0) | ||
3878 | goto out; | ||
3879 | else | ||
3880 | goto retry; | ||
3881 | } | ||
3882 | |||
3883 | rc = qeth_get_unitaddr(card); | ||
3884 | if (rc) { | ||
3885 | QETH_DBF_TEXT_(SETUP, 2, "2err%d", rc); | ||
3886 | return rc; | ||
3887 | } | ||
3888 | |||
3889 | mpno = QETH_MAX_PORTNO; | ||
3890 | if (card->info.portno > mpno) { | ||
3891 | PRINT_ERR("Device %s does not offer port number %d \n.", | ||
3892 | CARD_BUS_ID(card), card->info.portno); | ||
3893 | rc = -ENODEV; | ||
3894 | goto out; | ||
3895 | } | ||
3896 | qeth_init_tokens(card); | ||
3897 | qeth_init_func_level(card); | ||
3898 | rc = qeth_idx_activate_channel(&card->read, qeth_idx_read_cb); | ||
3899 | if (rc == -ERESTARTSYS) { | ||
3900 | QETH_DBF_TEXT(SETUP, 2, "break2"); | ||
3901 | return rc; | ||
3902 | } else if (rc) { | ||
3903 | QETH_DBF_TEXT_(SETUP, 2, "3err%d", rc); | ||
3904 | if (--retries < 0) | ||
3905 | goto out; | ||
3906 | else | ||
3907 | goto retry; | ||
3908 | } | ||
3909 | rc = qeth_idx_activate_channel(&card->write, qeth_idx_write_cb); | ||
3910 | if (rc == -ERESTARTSYS) { | ||
3911 | QETH_DBF_TEXT(SETUP, 2, "break3"); | ||
3912 | return rc; | ||
3913 | } else if (rc) { | ||
3914 | QETH_DBF_TEXT_(SETUP, 2, "4err%d", rc); | ||
3915 | if (--retries < 0) | ||
3916 | goto out; | ||
3917 | else | ||
3918 | goto retry; | ||
3919 | } | ||
3920 | rc = qeth_mpc_initialize(card); | ||
3921 | if (rc) { | ||
3922 | QETH_DBF_TEXT_(SETUP, 2, "5err%d", rc); | ||
3923 | goto out; | ||
3924 | } | ||
3925 | return 0; | ||
3926 | out: | ||
3927 | PRINT_ERR("Initialization in hardsetup failed! rc=%d\n", rc); | ||
3928 | return rc; | ||
3929 | } | ||
3930 | EXPORT_SYMBOL_GPL(qeth_core_hardsetup_card); | ||
3931 | |||
3932 | static inline int qeth_create_skb_frag(struct qdio_buffer_element *element, | ||
3933 | struct sk_buff **pskb, int offset, int *pfrag, int data_len) | ||
3934 | { | ||
3935 | struct page *page = virt_to_page(element->addr); | ||
3936 | if (*pskb == NULL) { | ||
3937 | /* the upper protocol layers assume that there is data in the | ||
3938 | * skb itself. Copy a small amount (64 bytes) to make them | ||
3939 | * happy. */ | ||
3940 | *pskb = dev_alloc_skb(64 + ETH_HLEN); | ||
3941 | if (!(*pskb)) | ||
3942 | return -ENOMEM; | ||
3943 | skb_reserve(*pskb, ETH_HLEN); | ||
3944 | if (data_len <= 64) { | ||
3945 | memcpy(skb_put(*pskb, data_len), element->addr + offset, | ||
3946 | data_len); | ||
3947 | } else { | ||
3948 | get_page(page); | ||
3949 | memcpy(skb_put(*pskb, 64), element->addr + offset, 64); | ||
3950 | skb_fill_page_desc(*pskb, *pfrag, page, offset + 64, | ||
3951 | data_len - 64); | ||
3952 | (*pskb)->data_len += data_len - 64; | ||
3953 | (*pskb)->len += data_len - 64; | ||
3954 | (*pskb)->truesize += data_len - 64; | ||
3955 | (*pfrag)++; | ||
3956 | } | ||
3957 | } else { | ||
3958 | get_page(page); | ||
3959 | skb_fill_page_desc(*pskb, *pfrag, page, offset, data_len); | ||
3960 | (*pskb)->data_len += data_len; | ||
3961 | (*pskb)->len += data_len; | ||
3962 | (*pskb)->truesize += data_len; | ||
3963 | (*pfrag)++; | ||
3964 | } | ||
3965 | return 0; | ||
3966 | } | ||
3967 | |||
3968 | struct sk_buff *qeth_core_get_next_skb(struct qeth_card *card, | ||
3969 | struct qdio_buffer *buffer, | ||
3970 | struct qdio_buffer_element **__element, int *__offset, | ||
3971 | struct qeth_hdr **hdr) | ||
3972 | { | ||
3973 | struct qdio_buffer_element *element = *__element; | ||
3974 | int offset = *__offset; | ||
3975 | struct sk_buff *skb = NULL; | ||
3976 | int skb_len; | ||
3977 | void *data_ptr; | ||
3978 | int data_len; | ||
3979 | int headroom = 0; | ||
3980 | int use_rx_sg = 0; | ||
3981 | int frag = 0; | ||
3982 | |||
3983 | QETH_DBF_TEXT(TRACE, 6, "nextskb"); | ||
3984 | /* qeth_hdr must not cross element boundaries */ | ||
3985 | if (element->length < offset + sizeof(struct qeth_hdr)) { | ||
3986 | if (qeth_is_last_sbale(element)) | ||
3987 | return NULL; | ||
3988 | element++; | ||
3989 | offset = 0; | ||
3990 | if (element->length < sizeof(struct qeth_hdr)) | ||
3991 | return NULL; | ||
3992 | } | ||
3993 | *hdr = element->addr + offset; | ||
3994 | |||
3995 | offset += sizeof(struct qeth_hdr); | ||
3996 | if (card->options.layer2) { | ||
3997 | if (card->info.type == QETH_CARD_TYPE_OSN) { | ||
3998 | skb_len = (*hdr)->hdr.osn.pdu_length; | ||
3999 | headroom = sizeof(struct qeth_hdr); | ||
4000 | } else { | ||
4001 | skb_len = (*hdr)->hdr.l2.pkt_length; | ||
4002 | } | ||
4003 | } else { | ||
4004 | skb_len = (*hdr)->hdr.l3.length; | ||
4005 | if ((card->info.link_type == QETH_LINK_TYPE_LANE_TR) || | ||
4006 | (card->info.link_type == QETH_LINK_TYPE_HSTR)) | ||
4007 | headroom = TR_HLEN; | ||
4008 | else | ||
4009 | headroom = ETH_HLEN; | ||
4010 | } | ||
4011 | |||
4012 | if (!skb_len) | ||
4013 | return NULL; | ||
4014 | |||
4015 | if ((skb_len >= card->options.rx_sg_cb) && | ||
4016 | (!(card->info.type == QETH_CARD_TYPE_OSN)) && | ||
4017 | (!atomic_read(&card->force_alloc_skb))) { | ||
4018 | use_rx_sg = 1; | ||
4019 | } else { | ||
4020 | skb = dev_alloc_skb(skb_len + headroom); | ||
4021 | if (!skb) | ||
4022 | goto no_mem; | ||
4023 | if (headroom) | ||
4024 | skb_reserve(skb, headroom); | ||
4025 | } | ||
4026 | |||
4027 | data_ptr = element->addr + offset; | ||
4028 | while (skb_len) { | ||
4029 | data_len = min(skb_len, (int)(element->length - offset)); | ||
4030 | if (data_len) { | ||
4031 | if (use_rx_sg) { | ||
4032 | if (qeth_create_skb_frag(element, &skb, offset, | ||
4033 | &frag, data_len)) | ||
4034 | goto no_mem; | ||
4035 | } else { | ||
4036 | memcpy(skb_put(skb, data_len), data_ptr, | ||
4037 | data_len); | ||
4038 | } | ||
4039 | } | ||
4040 | skb_len -= data_len; | ||
4041 | if (skb_len) { | ||
4042 | if (qeth_is_last_sbale(element)) { | ||
4043 | QETH_DBF_TEXT(TRACE, 4, "unexeob"); | ||
4044 | QETH_DBF_TEXT_(TRACE, 4, "%s", | ||
4045 | CARD_BUS_ID(card)); | ||
4046 | QETH_DBF_TEXT(QERR, 2, "unexeob"); | ||
4047 | QETH_DBF_TEXT_(QERR, 2, "%s", | ||
4048 | CARD_BUS_ID(card)); | ||
4049 | QETH_DBF_HEX(MISC, 4, buffer, sizeof(*buffer)); | ||
4050 | dev_kfree_skb_any(skb); | ||
4051 | card->stats.rx_errors++; | ||
4052 | return NULL; | ||
4053 | } | ||
4054 | element++; | ||
4055 | offset = 0; | ||
4056 | data_ptr = element->addr; | ||
4057 | } else { | ||
4058 | offset += data_len; | ||
4059 | } | ||
4060 | } | ||
4061 | *__element = element; | ||
4062 | *__offset = offset; | ||
4063 | if (use_rx_sg && card->options.performance_stats) { | ||
4064 | card->perf_stats.sg_skbs_rx++; | ||
4065 | card->perf_stats.sg_frags_rx += skb_shinfo(skb)->nr_frags; | ||
4066 | } | ||
4067 | return skb; | ||
4068 | no_mem: | ||
4069 | if (net_ratelimit()) { | ||
4070 | PRINT_WARN("No memory for packet received on %s.\n", | ||
4071 | QETH_CARD_IFNAME(card)); | ||
4072 | QETH_DBF_TEXT(TRACE, 2, "noskbmem"); | ||
4073 | QETH_DBF_TEXT_(TRACE, 2, "%s", CARD_BUS_ID(card)); | ||
4074 | } | ||
4075 | card->stats.rx_dropped++; | ||
4076 | return NULL; | ||
4077 | } | ||
4078 | EXPORT_SYMBOL_GPL(qeth_core_get_next_skb); | ||
4079 | |||
4080 | static void qeth_unregister_dbf_views(void) | ||
4081 | { | ||
4082 | int x; | ||
4083 | for (x = 0; x < QETH_DBF_INFOS; x++) { | ||
4084 | debug_unregister(qeth_dbf[x].id); | ||
4085 | qeth_dbf[x].id = NULL; | ||
4086 | } | ||
4087 | } | ||
4088 | |||
4089 | static int qeth_register_dbf_views(void) | ||
4090 | { | ||
4091 | int ret; | ||
4092 | int x; | ||
4093 | |||
4094 | for (x = 0; x < QETH_DBF_INFOS; x++) { | ||
4095 | /* register the areas */ | ||
4096 | qeth_dbf[x].id = debug_register(qeth_dbf[x].name, | ||
4097 | qeth_dbf[x].pages, | ||
4098 | qeth_dbf[x].areas, | ||
4099 | qeth_dbf[x].len); | ||
4100 | if (qeth_dbf[x].id == NULL) { | ||
4101 | qeth_unregister_dbf_views(); | ||
4102 | return -ENOMEM; | ||
4103 | } | ||
4104 | |||
4105 | /* register a view */ | ||
4106 | ret = debug_register_view(qeth_dbf[x].id, qeth_dbf[x].view); | ||
4107 | if (ret) { | ||
4108 | qeth_unregister_dbf_views(); | ||
4109 | return ret; | ||
4110 | } | ||
4111 | |||
4112 | /* set a passing level */ | ||
4113 | debug_set_level(qeth_dbf[x].id, qeth_dbf[x].level); | ||
4114 | } | ||
4115 | |||
4116 | return 0; | ||
4117 | } | ||
4118 | |||
4119 | int qeth_core_load_discipline(struct qeth_card *card, | ||
4120 | enum qeth_discipline_id discipline) | ||
4121 | { | ||
4122 | int rc = 0; | ||
4123 | switch (discipline) { | ||
4124 | case QETH_DISCIPLINE_LAYER3: | ||
4125 | card->discipline.ccwgdriver = try_then_request_module( | ||
4126 | symbol_get(qeth_l3_ccwgroup_driver), | ||
4127 | "qeth_l3"); | ||
4128 | break; | ||
4129 | case QETH_DISCIPLINE_LAYER2: | ||
4130 | card->discipline.ccwgdriver = try_then_request_module( | ||
4131 | symbol_get(qeth_l2_ccwgroup_driver), | ||
4132 | "qeth_l2"); | ||
4133 | break; | ||
4134 | } | ||
4135 | if (!card->discipline.ccwgdriver) { | ||
4136 | PRINT_ERR("Support for discipline %d not present\n", | ||
4137 | discipline); | ||
4138 | rc = -EINVAL; | ||
4139 | } | ||
4140 | return rc; | ||
4141 | } | ||
4142 | |||
4143 | void qeth_core_free_discipline(struct qeth_card *card) | ||
4144 | { | ||
4145 | if (card->options.layer2) | ||
4146 | symbol_put(qeth_l2_ccwgroup_driver); | ||
4147 | else | ||
4148 | symbol_put(qeth_l3_ccwgroup_driver); | ||
4149 | card->discipline.ccwgdriver = NULL; | ||
4150 | } | ||
4151 | |||
4152 | static int qeth_core_probe_device(struct ccwgroup_device *gdev) | ||
4153 | { | ||
4154 | struct qeth_card *card; | ||
4155 | struct device *dev; | ||
4156 | int rc; | ||
4157 | unsigned long flags; | ||
4158 | |||
4159 | QETH_DBF_TEXT(SETUP, 2, "probedev"); | ||
4160 | |||
4161 | dev = &gdev->dev; | ||
4162 | if (!get_device(dev)) | ||
4163 | return -ENODEV; | ||
4164 | |||
4165 | QETH_DBF_TEXT_(SETUP, 2, "%s", gdev->dev.bus_id); | ||
4166 | |||
4167 | card = qeth_alloc_card(); | ||
4168 | if (!card) { | ||
4169 | QETH_DBF_TEXT_(SETUP, 2, "1err%d", -ENOMEM); | ||
4170 | rc = -ENOMEM; | ||
4171 | goto err_dev; | ||
4172 | } | ||
4173 | card->read.ccwdev = gdev->cdev[0]; | ||
4174 | card->write.ccwdev = gdev->cdev[1]; | ||
4175 | card->data.ccwdev = gdev->cdev[2]; | ||
4176 | dev_set_drvdata(&gdev->dev, card); | ||
4177 | card->gdev = gdev; | ||
4178 | gdev->cdev[0]->handler = qeth_irq; | ||
4179 | gdev->cdev[1]->handler = qeth_irq; | ||
4180 | gdev->cdev[2]->handler = qeth_irq; | ||
4181 | |||
4182 | rc = qeth_determine_card_type(card); | ||
4183 | if (rc) { | ||
4184 | PRINT_WARN("%s: not a valid card type\n", __func__); | ||
4185 | QETH_DBF_TEXT_(SETUP, 2, "3err%d", rc); | ||
4186 | goto err_card; | ||
4187 | } | ||
4188 | rc = qeth_setup_card(card); | ||
4189 | if (rc) { | ||
4190 | QETH_DBF_TEXT_(SETUP, 2, "2err%d", rc); | ||
4191 | goto err_card; | ||
4192 | } | ||
4193 | |||
4194 | if (card->info.type == QETH_CARD_TYPE_OSN) { | ||
4195 | rc = qeth_core_create_osn_attributes(dev); | ||
4196 | if (rc) | ||
4197 | goto err_card; | ||
4198 | rc = qeth_core_load_discipline(card, QETH_DISCIPLINE_LAYER2); | ||
4199 | if (rc) { | ||
4200 | qeth_core_remove_osn_attributes(dev); | ||
4201 | goto err_card; | ||
4202 | } | ||
4203 | rc = card->discipline.ccwgdriver->probe(card->gdev); | ||
4204 | if (rc) { | ||
4205 | qeth_core_free_discipline(card); | ||
4206 | qeth_core_remove_osn_attributes(dev); | ||
4207 | goto err_card; | ||
4208 | } | ||
4209 | } else { | ||
4210 | rc = qeth_core_create_device_attributes(dev); | ||
4211 | if (rc) | ||
4212 | goto err_card; | ||
4213 | } | ||
4214 | |||
4215 | write_lock_irqsave(&qeth_core_card_list.rwlock, flags); | ||
4216 | list_add_tail(&card->list, &qeth_core_card_list.list); | ||
4217 | write_unlock_irqrestore(&qeth_core_card_list.rwlock, flags); | ||
4218 | return 0; | ||
4219 | |||
4220 | err_card: | ||
4221 | qeth_core_free_card(card); | ||
4222 | err_dev: | ||
4223 | put_device(dev); | ||
4224 | return rc; | ||
4225 | } | ||
4226 | |||
4227 | static void qeth_core_remove_device(struct ccwgroup_device *gdev) | ||
4228 | { | ||
4229 | unsigned long flags; | ||
4230 | struct qeth_card *card = dev_get_drvdata(&gdev->dev); | ||
4231 | |||
4232 | if (card->discipline.ccwgdriver) { | ||
4233 | card->discipline.ccwgdriver->remove(gdev); | ||
4234 | qeth_core_free_discipline(card); | ||
4235 | } | ||
4236 | |||
4237 | if (card->info.type == QETH_CARD_TYPE_OSN) { | ||
4238 | qeth_core_remove_osn_attributes(&gdev->dev); | ||
4239 | } else { | ||
4240 | qeth_core_remove_device_attributes(&gdev->dev); | ||
4241 | } | ||
4242 | write_lock_irqsave(&qeth_core_card_list.rwlock, flags); | ||
4243 | list_del(&card->list); | ||
4244 | write_unlock_irqrestore(&qeth_core_card_list.rwlock, flags); | ||
4245 | qeth_core_free_card(card); | ||
4246 | dev_set_drvdata(&gdev->dev, NULL); | ||
4247 | put_device(&gdev->dev); | ||
4248 | return; | ||
4249 | } | ||
4250 | |||
4251 | static int qeth_core_set_online(struct ccwgroup_device *gdev) | ||
4252 | { | ||
4253 | struct qeth_card *card = dev_get_drvdata(&gdev->dev); | ||
4254 | int rc = 0; | ||
4255 | int def_discipline; | ||
4256 | |||
4257 | if (!card->discipline.ccwgdriver) { | ||
4258 | if (card->info.type == QETH_CARD_TYPE_IQD) | ||
4259 | def_discipline = QETH_DISCIPLINE_LAYER3; | ||
4260 | else | ||
4261 | def_discipline = QETH_DISCIPLINE_LAYER2; | ||
4262 | rc = qeth_core_load_discipline(card, def_discipline); | ||
4263 | if (rc) | ||
4264 | goto err; | ||
4265 | rc = card->discipline.ccwgdriver->probe(card->gdev); | ||
4266 | if (rc) | ||
4267 | goto err; | ||
4268 | } | ||
4269 | rc = card->discipline.ccwgdriver->set_online(gdev); | ||
4270 | err: | ||
4271 | return rc; | ||
4272 | } | ||
4273 | |||
4274 | static int qeth_core_set_offline(struct ccwgroup_device *gdev) | ||
4275 | { | ||
4276 | struct qeth_card *card = dev_get_drvdata(&gdev->dev); | ||
4277 | return card->discipline.ccwgdriver->set_offline(gdev); | ||
4278 | } | ||
4279 | |||
4280 | static void qeth_core_shutdown(struct ccwgroup_device *gdev) | ||
4281 | { | ||
4282 | struct qeth_card *card = dev_get_drvdata(&gdev->dev); | ||
4283 | if (card->discipline.ccwgdriver && | ||
4284 | card->discipline.ccwgdriver->shutdown) | ||
4285 | card->discipline.ccwgdriver->shutdown(gdev); | ||
4286 | } | ||
4287 | |||
4288 | static struct ccwgroup_driver qeth_core_ccwgroup_driver = { | ||
4289 | .owner = THIS_MODULE, | ||
4290 | .name = "qeth", | ||
4291 | .driver_id = 0xD8C5E3C8, | ||
4292 | .probe = qeth_core_probe_device, | ||
4293 | .remove = qeth_core_remove_device, | ||
4294 | .set_online = qeth_core_set_online, | ||
4295 | .set_offline = qeth_core_set_offline, | ||
4296 | .shutdown = qeth_core_shutdown, | ||
4297 | }; | ||
4298 | |||
4299 | static ssize_t | ||
4300 | qeth_core_driver_group_store(struct device_driver *ddrv, const char *buf, | ||
4301 | size_t count) | ||
4302 | { | ||
4303 | int err; | ||
4304 | err = qeth_core_driver_group(buf, qeth_core_root_dev, | ||
4305 | qeth_core_ccwgroup_driver.driver_id); | ||
4306 | if (err) | ||
4307 | return err; | ||
4308 | else | ||
4309 | return count; | ||
4310 | } | ||
4311 | |||
4312 | static DRIVER_ATTR(group, 0200, NULL, qeth_core_driver_group_store); | ||
4313 | |||
4314 | static struct { | ||
4315 | const char str[ETH_GSTRING_LEN]; | ||
4316 | } qeth_ethtool_stats_keys[] = { | ||
4317 | /* 0 */{"rx skbs"}, | ||
4318 | {"rx buffers"}, | ||
4319 | {"tx skbs"}, | ||
4320 | {"tx buffers"}, | ||
4321 | {"tx skbs no packing"}, | ||
4322 | {"tx buffers no packing"}, | ||
4323 | {"tx skbs packing"}, | ||
4324 | {"tx buffers packing"}, | ||
4325 | {"tx sg skbs"}, | ||
4326 | {"tx sg frags"}, | ||
4327 | /* 10 */{"rx sg skbs"}, | ||
4328 | {"rx sg frags"}, | ||
4329 | {"rx sg page allocs"}, | ||
4330 | {"tx large kbytes"}, | ||
4331 | {"tx large count"}, | ||
4332 | {"tx pk state ch n->p"}, | ||
4333 | {"tx pk state ch p->n"}, | ||
4334 | {"tx pk watermark low"}, | ||
4335 | {"tx pk watermark high"}, | ||
4336 | {"queue 0 buffer usage"}, | ||
4337 | /* 20 */{"queue 1 buffer usage"}, | ||
4338 | {"queue 2 buffer usage"}, | ||
4339 | {"queue 3 buffer usage"}, | ||
4340 | {"rx handler time"}, | ||
4341 | {"rx handler count"}, | ||
4342 | {"rx do_QDIO time"}, | ||
4343 | {"rx do_QDIO count"}, | ||
4344 | {"tx handler time"}, | ||
4345 | {"tx handler count"}, | ||
4346 | {"tx time"}, | ||
4347 | /* 30 */{"tx count"}, | ||
4348 | {"tx do_QDIO time"}, | ||
4349 | {"tx do_QDIO count"}, | ||
4350 | }; | ||
4351 | |||
4352 | int qeth_core_get_stats_count(struct net_device *dev) | ||
4353 | { | ||
4354 | return (sizeof(qeth_ethtool_stats_keys) / ETH_GSTRING_LEN); | ||
4355 | } | ||
4356 | EXPORT_SYMBOL_GPL(qeth_core_get_stats_count); | ||
4357 | |||
4358 | void qeth_core_get_ethtool_stats(struct net_device *dev, | ||
4359 | struct ethtool_stats *stats, u64 *data) | ||
4360 | { | ||
4361 | struct qeth_card *card = netdev_priv(dev); | ||
4362 | data[0] = card->stats.rx_packets - | ||
4363 | card->perf_stats.initial_rx_packets; | ||
4364 | data[1] = card->perf_stats.bufs_rec; | ||
4365 | data[2] = card->stats.tx_packets - | ||
4366 | card->perf_stats.initial_tx_packets; | ||
4367 | data[3] = card->perf_stats.bufs_sent; | ||
4368 | data[4] = card->stats.tx_packets - card->perf_stats.initial_tx_packets | ||
4369 | - card->perf_stats.skbs_sent_pack; | ||
4370 | data[5] = card->perf_stats.bufs_sent - card->perf_stats.bufs_sent_pack; | ||
4371 | data[6] = card->perf_stats.skbs_sent_pack; | ||
4372 | data[7] = card->perf_stats.bufs_sent_pack; | ||
4373 | data[8] = card->perf_stats.sg_skbs_sent; | ||
4374 | data[9] = card->perf_stats.sg_frags_sent; | ||
4375 | data[10] = card->perf_stats.sg_skbs_rx; | ||
4376 | data[11] = card->perf_stats.sg_frags_rx; | ||
4377 | data[12] = card->perf_stats.sg_alloc_page_rx; | ||
4378 | data[13] = (card->perf_stats.large_send_bytes >> 10); | ||
4379 | data[14] = card->perf_stats.large_send_cnt; | ||
4380 | data[15] = card->perf_stats.sc_dp_p; | ||
4381 | data[16] = card->perf_stats.sc_p_dp; | ||
4382 | data[17] = QETH_LOW_WATERMARK_PACK; | ||
4383 | data[18] = QETH_HIGH_WATERMARK_PACK; | ||
4384 | data[19] = atomic_read(&card->qdio.out_qs[0]->used_buffers); | ||
4385 | data[20] = (card->qdio.no_out_queues > 1) ? | ||
4386 | atomic_read(&card->qdio.out_qs[1]->used_buffers) : 0; | ||
4387 | data[21] = (card->qdio.no_out_queues > 2) ? | ||
4388 | atomic_read(&card->qdio.out_qs[2]->used_buffers) : 0; | ||
4389 | data[22] = (card->qdio.no_out_queues > 3) ? | ||
4390 | atomic_read(&card->qdio.out_qs[3]->used_buffers) : 0; | ||
4391 | data[23] = card->perf_stats.inbound_time; | ||
4392 | data[24] = card->perf_stats.inbound_cnt; | ||
4393 | data[25] = card->perf_stats.inbound_do_qdio_time; | ||
4394 | data[26] = card->perf_stats.inbound_do_qdio_cnt; | ||
4395 | data[27] = card->perf_stats.outbound_handler_time; | ||
4396 | data[28] = card->perf_stats.outbound_handler_cnt; | ||
4397 | data[29] = card->perf_stats.outbound_time; | ||
4398 | data[30] = card->perf_stats.outbound_cnt; | ||
4399 | data[31] = card->perf_stats.outbound_do_qdio_time; | ||
4400 | data[32] = card->perf_stats.outbound_do_qdio_cnt; | ||
4401 | } | ||
4402 | EXPORT_SYMBOL_GPL(qeth_core_get_ethtool_stats); | ||
4403 | |||
4404 | void qeth_core_get_strings(struct net_device *dev, u32 stringset, u8 *data) | ||
4405 | { | ||
4406 | switch (stringset) { | ||
4407 | case ETH_SS_STATS: | ||
4408 | memcpy(data, &qeth_ethtool_stats_keys, | ||
4409 | sizeof(qeth_ethtool_stats_keys)); | ||
4410 | break; | ||
4411 | default: | ||
4412 | WARN_ON(1); | ||
4413 | break; | ||
4414 | } | ||
4415 | } | ||
4416 | EXPORT_SYMBOL_GPL(qeth_core_get_strings); | ||
4417 | |||
4418 | void qeth_core_get_drvinfo(struct net_device *dev, | ||
4419 | struct ethtool_drvinfo *info) | ||
4420 | { | ||
4421 | struct qeth_card *card = netdev_priv(dev); | ||
4422 | if (card->options.layer2) | ||
4423 | strcpy(info->driver, "qeth_l2"); | ||
4424 | else | ||
4425 | strcpy(info->driver, "qeth_l3"); | ||
4426 | |||
4427 | strcpy(info->version, "1.0"); | ||
4428 | strcpy(info->fw_version, card->info.mcl_level); | ||
4429 | sprintf(info->bus_info, "%s/%s/%s", | ||
4430 | CARD_RDEV_ID(card), | ||
4431 | CARD_WDEV_ID(card), | ||
4432 | CARD_DDEV_ID(card)); | ||
4433 | } | ||
4434 | EXPORT_SYMBOL_GPL(qeth_core_get_drvinfo); | ||
4435 | |||
4436 | static int __init qeth_core_init(void) | ||
4437 | { | ||
4438 | int rc; | ||
4439 | |||
4440 | PRINT_INFO("loading core functions\n"); | ||
4441 | INIT_LIST_HEAD(&qeth_core_card_list.list); | ||
4442 | rwlock_init(&qeth_core_card_list.rwlock); | ||
4443 | |||
4444 | rc = qeth_register_dbf_views(); | ||
4445 | if (rc) | ||
4446 | goto out_err; | ||
4447 | rc = ccw_driver_register(&qeth_ccw_driver); | ||
4448 | if (rc) | ||
4449 | goto ccw_err; | ||
4450 | rc = ccwgroup_driver_register(&qeth_core_ccwgroup_driver); | ||
4451 | if (rc) | ||
4452 | goto ccwgroup_err; | ||
4453 | rc = driver_create_file(&qeth_core_ccwgroup_driver.driver, | ||
4454 | &driver_attr_group); | ||
4455 | if (rc) | ||
4456 | goto driver_err; | ||
4457 | qeth_core_root_dev = s390_root_dev_register("qeth"); | ||
4458 | rc = IS_ERR(qeth_core_root_dev) ? PTR_ERR(qeth_core_root_dev) : 0; | ||
4459 | if (rc) | ||
4460 | goto register_err; | ||
4461 | return 0; | ||
4462 | |||
4463 | register_err: | ||
4464 | driver_remove_file(&qeth_core_ccwgroup_driver.driver, | ||
4465 | &driver_attr_group); | ||
4466 | driver_err: | ||
4467 | ccwgroup_driver_unregister(&qeth_core_ccwgroup_driver); | ||
4468 | ccwgroup_err: | ||
4469 | ccw_driver_unregister(&qeth_ccw_driver); | ||
4470 | ccw_err: | ||
4471 | qeth_unregister_dbf_views(); | ||
4472 | out_err: | ||
4473 | PRINT_ERR("Initialization failed with code %d\n", rc); | ||
4474 | return rc; | ||
4475 | } | ||
4476 | |||
4477 | static void __exit qeth_core_exit(void) | ||
4478 | { | ||
4479 | s390_root_dev_unregister(qeth_core_root_dev); | ||
4480 | driver_remove_file(&qeth_core_ccwgroup_driver.driver, | ||
4481 | &driver_attr_group); | ||
4482 | ccwgroup_driver_unregister(&qeth_core_ccwgroup_driver); | ||
4483 | ccw_driver_unregister(&qeth_ccw_driver); | ||
4484 | qeth_unregister_dbf_views(); | ||
4485 | PRINT_INFO("core functions removed\n"); | ||
4486 | } | ||
4487 | |||
4488 | module_init(qeth_core_init); | ||
4489 | module_exit(qeth_core_exit); | ||
4490 | MODULE_AUTHOR("Frank Blaschka <frank.blaschka@de.ibm.com>"); | ||
4491 | MODULE_DESCRIPTION("qeth core functions"); | ||
4492 | MODULE_LICENSE("GPL"); | ||
diff --git a/drivers/s390/net/qeth_core_mpc.c b/drivers/s390/net/qeth_core_mpc.c new file mode 100644 index 000000000000..06f4de1f0507 --- /dev/null +++ b/drivers/s390/net/qeth_core_mpc.c | |||
@@ -0,0 +1,266 @@ | |||
1 | /* | ||
2 | * drivers/s390/net/qeth_core_mpc.c | ||
3 | * | ||
4 | * Copyright IBM Corp. 2007 | ||
5 | * Author(s): Frank Pavlic <fpavlic@de.ibm.com>, | ||
6 | * Thomas Spatzier <tspat@de.ibm.com>, | ||
7 | * Frank Blaschka <frank.blaschka@de.ibm.com> | ||
8 | */ | ||
9 | |||
10 | #include <linux/module.h> | ||
11 | #include <asm/cio.h> | ||
12 | #include "qeth_core_mpc.h" | ||
13 | |||
14 | unsigned char IDX_ACTIVATE_READ[] = { | ||
15 | 0x00, 0x00, 0x80, 0x00, 0x00, 0x00, 0x00, 0x00, | ||
16 | 0x19, 0x01, 0x01, 0x80, 0x00, 0x00, 0x00, 0x00, | ||
17 | 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0xc8, 0xc1, | ||
18 | 0xd3, 0xd3, 0xd6, 0xd3, 0xc5, 0x40, 0x00, 0x00, | ||
19 | 0x00, 0x00 | ||
20 | }; | ||
21 | |||
22 | unsigned char IDX_ACTIVATE_WRITE[] = { | ||
23 | 0x00, 0x00, 0x80, 0x00, 0x00, 0x00, 0x00, 0x00, | ||
24 | 0x15, 0x01, 0x01, 0x80, 0x00, 0x00, 0x00, 0x00, | ||
25 | 0xff, 0xff, 0x00, 0x00, 0x00, 0x00, 0xc8, 0xc1, | ||
26 | 0xd3, 0xd3, 0xd6, 0xd3, 0xc5, 0x40, 0x00, 0x00, | ||
27 | 0x00, 0x00 | ||
28 | }; | ||
29 | |||
30 | unsigned char CM_ENABLE[] = { | ||
31 | 0x00, 0xe0, 0x00, 0x00, 0x00, 0x00, 0x00, 0x01, | ||
32 | 0x00, 0x00, 0x00, 0x14, 0x00, 0x00, 0x00, 0x63, | ||
33 | 0x10, 0x00, 0x00, 0x01, | ||
34 | 0x00, 0x00, 0x00, 0x00, | ||
35 | 0x81, 0x7e, 0x00, 0x01, 0x00, 0x00, 0x00, 0x00, | ||
36 | 0x00, 0x00, 0x00, 0x00, 0x00, 0x24, 0x00, 0x23, | ||
37 | 0x00, 0x00, 0x23, 0x05, 0x00, 0x00, 0x00, 0x00, | ||
38 | 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, | ||
39 | 0x01, 0x00, 0x00, 0x23, 0x00, 0x00, 0x00, 0x40, | ||
40 | 0x00, 0x0c, 0x41, 0x02, 0x00, 0x17, 0x00, 0x00, | ||
41 | 0x00, 0x00, 0x00, 0x00, | ||
42 | 0x00, 0x0b, 0x04, 0x01, | ||
43 | 0x7e, 0x04, 0x05, 0x00, 0x01, 0x01, 0x0f, | ||
44 | 0x00, | ||
45 | 0x0c, 0x04, 0x02, 0xff, 0xff, 0xff, 0xff, 0xff, | ||
46 | 0xff, 0xff, 0xff | ||
47 | }; | ||
48 | |||
49 | unsigned char CM_SETUP[] = { | ||
50 | 0x00, 0xe0, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, | ||
51 | 0x00, 0x00, 0x00, 0x14, 0x00, 0x00, 0x00, 0x64, | ||
52 | 0x10, 0x00, 0x00, 0x01, | ||
53 | 0x00, 0x00, 0x00, 0x00, | ||
54 | 0x81, 0x7e, 0x00, 0x01, 0x00, 0x00, 0x00, 0x00, | ||
55 | 0x00, 0x00, 0x00, 0x00, 0x00, 0x24, 0x00, 0x24, | ||
56 | 0x00, 0x00, 0x24, 0x05, 0x00, 0x00, 0x00, 0x00, | ||
57 | 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, | ||
58 | 0x01, 0x00, 0x00, 0x24, 0x00, 0x00, 0x00, 0x40, | ||
59 | 0x00, 0x0c, 0x41, 0x04, 0x00, 0x18, 0x00, 0x00, | ||
60 | 0x00, 0x00, 0x00, 0x00, | ||
61 | 0x00, 0x09, 0x04, 0x04, | ||
62 | 0x05, 0x00, 0x01, 0x01, 0x11, | ||
63 | 0x00, 0x09, 0x04, | ||
64 | 0x05, 0x05, 0x00, 0x00, 0x00, 0x00, | ||
65 | 0x00, 0x06, | ||
66 | 0x04, 0x06, 0xc8, 0x00 | ||
67 | }; | ||
68 | |||
69 | unsigned char ULP_ENABLE[] = { | ||
70 | 0x00, 0xe0, 0x00, 0x00, 0x00, 0x00, 0x00, 0x03, | ||
71 | 0x00, 0x00, 0x00, 0x14, 0x00, 0x00, 0x00, 0x6b, | ||
72 | 0x10, 0x00, 0x00, 0x01, | ||
73 | 0x00, 0x00, 0x00, 0x00, | ||
74 | 0x41, 0x7e, 0x00, 0x01, 0x00, 0x00, 0x00, 0x01, | ||
75 | 0x00, 0x00, 0x00, 0x00, 0x00, 0x24, 0x00, 0x2b, | ||
76 | 0x00, 0x00, 0x2b, 0x05, 0x20, 0x01, 0x00, 0x00, | ||
77 | 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, | ||
78 | 0x01, 0x00, 0x00, 0x2b, 0x00, 0x00, 0x00, 0x40, | ||
79 | 0x00, 0x0c, 0x41, 0x02, 0x00, 0x1f, 0x00, 0x00, | ||
80 | 0x00, 0x00, 0x00, 0x00, | ||
81 | 0x00, 0x0b, 0x04, 0x01, | ||
82 | 0x03, 0x04, 0x05, 0x00, 0x01, 0x01, 0x12, | ||
83 | 0x00, | ||
84 | 0x14, 0x04, 0x0a, 0x00, 0x20, 0x00, 0x00, 0xff, | ||
85 | 0xff, 0x00, 0x08, 0xc8, 0xe8, 0xc4, 0xf1, 0xc7, | ||
86 | 0xf1, 0x00, 0x00 | ||
87 | }; | ||
88 | |||
89 | unsigned char ULP_SETUP[] = { | ||
90 | 0x00, 0xe0, 0x00, 0x00, 0x00, 0x00, 0x00, 0x04, | ||
91 | 0x00, 0x00, 0x00, 0x14, 0x00, 0x00, 0x00, 0x6c, | ||
92 | 0x10, 0x00, 0x00, 0x01, | ||
93 | 0x00, 0x00, 0x00, 0x00, | ||
94 | 0x41, 0x7e, 0x00, 0x01, 0x00, 0x00, 0x00, 0x02, | ||
95 | 0x00, 0x00, 0x00, 0x01, 0x00, 0x24, 0x00, 0x2c, | ||
96 | 0x00, 0x00, 0x2c, 0x05, 0x20, 0x01, 0x00, 0x00, | ||
97 | 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, | ||
98 | 0x01, 0x00, 0x00, 0x2c, 0x00, 0x00, 0x00, 0x40, | ||
99 | 0x00, 0x0c, 0x41, 0x04, 0x00, 0x20, 0x00, 0x00, | ||
100 | 0x00, 0x00, 0x00, 0x00, | ||
101 | 0x00, 0x09, 0x04, 0x04, | ||
102 | 0x05, 0x00, 0x01, 0x01, 0x14, | ||
103 | 0x00, 0x09, 0x04, | ||
104 | 0x05, 0x05, 0x30, 0x01, 0x00, 0x00, | ||
105 | 0x00, 0x06, | ||
106 | 0x04, 0x06, 0x40, 0x00, | ||
107 | 0x00, 0x08, 0x04, 0x0b, | ||
108 | 0x00, 0x00, 0x00, 0x00 | ||
109 | }; | ||
110 | |||
111 | unsigned char DM_ACT[] = { | ||
112 | 0x00, 0xe0, 0x00, 0x00, 0x00, 0x00, 0x00, 0x05, | ||
113 | 0x00, 0x00, 0x00, 0x14, 0x00, 0x00, 0x00, 0x55, | ||
114 | 0x10, 0x00, 0x00, 0x01, | ||
115 | 0x00, 0x00, 0x00, 0x00, | ||
116 | 0x41, 0x7e, 0x00, 0x01, 0x00, 0x00, 0x00, 0x03, | ||
117 | 0x00, 0x00, 0x00, 0x02, 0x00, 0x24, 0x00, 0x15, | ||
118 | 0x00, 0x00, 0x2c, 0x05, 0x20, 0x01, 0x00, 0x00, | ||
119 | 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, | ||
120 | 0x01, 0x00, 0x00, 0x15, 0x00, 0x00, 0x00, 0x40, | ||
121 | 0x00, 0x0c, 0x43, 0x60, 0x00, 0x09, 0x00, 0x00, | ||
122 | 0x00, 0x00, 0x00, 0x00, | ||
123 | 0x00, 0x09, 0x04, 0x04, | ||
124 | 0x05, 0x40, 0x01, 0x01, 0x00 | ||
125 | }; | ||
126 | |||
127 | unsigned char IPA_PDU_HEADER[] = { | ||
128 | 0x00, 0xe0, 0x00, 0x00, 0x77, 0x77, 0x77, 0x77, | ||
129 | 0x00, 0x00, 0x00, 0x14, 0x00, 0x00, | ||
130 | (IPA_PDU_HEADER_SIZE+sizeof(struct qeth_ipa_cmd)) / 256, | ||
131 | (IPA_PDU_HEADER_SIZE+sizeof(struct qeth_ipa_cmd)) % 256, | ||
132 | 0x10, 0x00, 0x00, 0x01, 0x00, 0x00, 0x00, 0x00, | ||
133 | 0xc1, 0x03, 0x00, 0x01, 0x00, 0x00, 0x00, 0x00, | ||
134 | 0x00, 0x00, 0x00, 0x00, 0x00, 0x24, | ||
135 | sizeof(struct qeth_ipa_cmd) / 256, | ||
136 | sizeof(struct qeth_ipa_cmd) % 256, | ||
137 | 0x00, | ||
138 | sizeof(struct qeth_ipa_cmd) / 256, | ||
139 | sizeof(struct qeth_ipa_cmd) % 256, | ||
140 | 0x05, | ||
141 | 0x77, 0x77, 0x77, 0x77, | ||
142 | 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, | ||
143 | 0x01, 0x00, | ||
144 | sizeof(struct qeth_ipa_cmd) / 256, | ||
145 | sizeof(struct qeth_ipa_cmd) % 256, | ||
146 | 0x00, 0x00, 0x00, 0x40, | ||
147 | }; | ||
148 | EXPORT_SYMBOL_GPL(IPA_PDU_HEADER); | ||
149 | |||
150 | unsigned char WRITE_CCW[] = { | ||
151 | 0x01, CCW_FLAG_SLI, 0, 0, | ||
152 | 0, 0, 0, 0 | ||
153 | }; | ||
154 | |||
155 | unsigned char READ_CCW[] = { | ||
156 | 0x02, CCW_FLAG_SLI, 0, 0, | ||
157 | 0, 0, 0, 0 | ||
158 | }; | ||
159 | |||
160 | |||
161 | struct ipa_rc_msg { | ||
162 | enum qeth_ipa_return_codes rc; | ||
163 | char *msg; | ||
164 | }; | ||
165 | |||
166 | static struct ipa_rc_msg qeth_ipa_rc_msg[] = { | ||
167 | {IPA_RC_SUCCESS, "success"}, | ||
168 | {IPA_RC_NOTSUPP, "Command not supported"}, | ||
169 | {IPA_RC_IP_TABLE_FULL, "Add Addr IP Table Full - ipv6"}, | ||
170 | {IPA_RC_UNKNOWN_ERROR, "IPA command failed - reason unknown"}, | ||
171 | {IPA_RC_UNSUPPORTED_COMMAND, "Command not supported"}, | ||
172 | {IPA_RC_DUP_IPV6_REMOTE, "ipv6 address already registered remote"}, | ||
173 | {IPA_RC_DUP_IPV6_HOME, "ipv6 address already registered"}, | ||
174 | {IPA_RC_UNREGISTERED_ADDR, "Address not registered"}, | ||
175 | {IPA_RC_NO_ID_AVAILABLE, "No identifiers available"}, | ||
176 | {IPA_RC_ID_NOT_FOUND, "Identifier not found"}, | ||
177 | {IPA_RC_INVALID_IP_VERSION, "IP version incorrect"}, | ||
178 | {IPA_RC_LAN_FRAME_MISMATCH, "LAN and frame mismatch"}, | ||
179 | {IPA_RC_L2_UNSUPPORTED_CMD, "Unsupported layer 2 command"}, | ||
180 | {IPA_RC_L2_DUP_MAC, "Duplicate MAC address"}, | ||
181 | {IPA_RC_L2_ADDR_TABLE_FULL, "Layer2 address table full"}, | ||
182 | {IPA_RC_L2_DUP_LAYER3_MAC, "Duplicate with layer 3 MAC"}, | ||
183 | {IPA_RC_L2_GMAC_NOT_FOUND, "GMAC not found"}, | ||
184 | {IPA_RC_L2_MAC_NOT_FOUND, "L2 mac address not found"}, | ||
185 | {IPA_RC_L2_INVALID_VLAN_ID, "L2 invalid vlan id"}, | ||
186 | {IPA_RC_L2_DUP_VLAN_ID, "L2 duplicate vlan id"}, | ||
187 | {IPA_RC_L2_VLAN_ID_NOT_FOUND, "L2 vlan id not found"}, | ||
188 | {IPA_RC_DATA_MISMATCH, "Data field mismatch (v4/v6 mixed)"}, | ||
189 | {IPA_RC_INVALID_MTU_SIZE, "Invalid MTU size"}, | ||
190 | {IPA_RC_INVALID_LANTYPE, "Invalid LAN type"}, | ||
191 | {IPA_RC_INVALID_LANNUM, "Invalid LAN num"}, | ||
192 | {IPA_RC_DUPLICATE_IP_ADDRESS, "Address already registered"}, | ||
193 | {IPA_RC_IP_ADDR_TABLE_FULL, "IP address table full"}, | ||
194 | {IPA_RC_LAN_PORT_STATE_ERROR, "LAN port state error"}, | ||
195 | {IPA_RC_SETIP_NO_STARTLAN, "Setip no startlan received"}, | ||
196 | {IPA_RC_SETIP_ALREADY_RECEIVED, "Setip already received"}, | ||
197 | {IPA_RC_IP_ADDR_ALREADY_USED, "IP address already in use on LAN"}, | ||
198 | {IPA_RC_MC_ADDR_NOT_FOUND, "Multicast address not found"}, | ||
199 | {IPA_RC_SETIP_INVALID_VERSION, "SETIP invalid IP version"}, | ||
200 | {IPA_RC_UNSUPPORTED_SUBCMD, "Unsupported assist subcommand"}, | ||
201 | {IPA_RC_ARP_ASSIST_NO_ENABLE, "Only partial success, no enable"}, | ||
202 | {IPA_RC_PRIMARY_ALREADY_DEFINED, "Primary already defined"}, | ||
203 | {IPA_RC_SECOND_ALREADY_DEFINED, "Secondary already defined"}, | ||
204 | {IPA_RC_INVALID_SETRTG_INDICATOR, "Invalid SETRTG indicator"}, | ||
205 | {IPA_RC_MC_ADDR_ALREADY_DEFINED, "Multicast address already defined"}, | ||
206 | {IPA_RC_LAN_OFFLINE, "STRTLAN_LAN_DISABLED - LAN offline"}, | ||
207 | {IPA_RC_INVALID_IP_VERSION2, "Invalid IP version"}, | ||
208 | {IPA_RC_FFFF, "Unknown Error"} | ||
209 | }; | ||
210 | |||
211 | |||
212 | |||
213 | char *qeth_get_ipa_msg(enum qeth_ipa_return_codes rc) | ||
214 | { | ||
215 | int x = 0; | ||
216 | qeth_ipa_rc_msg[sizeof(qeth_ipa_rc_msg) / | ||
217 | sizeof(struct ipa_rc_msg) - 1].rc = rc; | ||
218 | while (qeth_ipa_rc_msg[x].rc != rc) | ||
219 | x++; | ||
220 | return qeth_ipa_rc_msg[x].msg; | ||
221 | } | ||
222 | |||
223 | |||
224 | struct ipa_cmd_names { | ||
225 | enum qeth_ipa_cmds cmd; | ||
226 | char *name; | ||
227 | }; | ||
228 | |||
229 | static struct ipa_cmd_names qeth_ipa_cmd_names[] = { | ||
230 | {IPA_CMD_STARTLAN, "startlan"}, | ||
231 | {IPA_CMD_STOPLAN, "stoplan"}, | ||
232 | {IPA_CMD_SETVMAC, "setvmac"}, | ||
233 | {IPA_CMD_DELVMAC, "delvmac"}, | ||
234 | {IPA_CMD_SETGMAC, "setgmac"}, | ||
235 | {IPA_CMD_DELGMAC, "delgmac"}, | ||
236 | {IPA_CMD_SETVLAN, "setvlan"}, | ||
237 | {IPA_CMD_DELVLAN, "delvlan"}, | ||
238 | {IPA_CMD_SETCCID, "setccid"}, | ||
239 | {IPA_CMD_DELCCID, "delccid"}, | ||
240 | {IPA_CMD_MODCCID, "modccid"}, | ||
241 | {IPA_CMD_SETIP, "setip"}, | ||
242 | {IPA_CMD_QIPASSIST, "qipassist"}, | ||
243 | {IPA_CMD_SETASSPARMS, "setassparms"}, | ||
244 | {IPA_CMD_SETIPM, "setipm"}, | ||
245 | {IPA_CMD_DELIPM, "delipm"}, | ||
246 | {IPA_CMD_SETRTG, "setrtg"}, | ||
247 | {IPA_CMD_DELIP, "delip"}, | ||
248 | {IPA_CMD_SETADAPTERPARMS, "setadapterparms"}, | ||
249 | {IPA_CMD_SET_DIAG_ASS, "set_diag_ass"}, | ||
250 | {IPA_CMD_CREATE_ADDR, "create_addr"}, | ||
251 | {IPA_CMD_DESTROY_ADDR, "destroy_addr"}, | ||
252 | {IPA_CMD_REGISTER_LOCAL_ADDR, "register_local_addr"}, | ||
253 | {IPA_CMD_UNREGISTER_LOCAL_ADDR, "unregister_local_addr"}, | ||
254 | {IPA_CMD_UNKNOWN, "unknown"}, | ||
255 | }; | ||
256 | |||
257 | char *qeth_get_ipa_cmd_name(enum qeth_ipa_cmds cmd) | ||
258 | { | ||
259 | int x = 0; | ||
260 | qeth_ipa_cmd_names[ | ||
261 | sizeof(qeth_ipa_cmd_names) / | ||
262 | sizeof(struct ipa_cmd_names)-1].cmd = cmd; | ||
263 | while (qeth_ipa_cmd_names[x].cmd != cmd) | ||
264 | x++; | ||
265 | return qeth_ipa_cmd_names[x].name; | ||
266 | } | ||
diff --git a/drivers/s390/net/qeth_mpc.h b/drivers/s390/net/qeth_core_mpc.h index 6de2da5ed5fd..18548822e37c 100644 --- a/drivers/s390/net/qeth_mpc.h +++ b/drivers/s390/net/qeth_core_mpc.h | |||
@@ -1,27 +1,25 @@ | |||
1 | /* | 1 | /* |
2 | * linux/drivers/s390/net/qeth_mpc.h | 2 | * drivers/s390/net/qeth_core_mpc.h |
3 | * | ||
4 | * Linux on zSeries OSA Express and HiperSockets support | ||
5 | * | ||
6 | * Copyright 2000,2003 IBM Corporation | ||
7 | * Author(s): Utz Bacher <utz.bacher@de.ibm.com> | ||
8 | * Thomas Spatzier <tspat@de.ibm.com> | ||
9 | * Frank Pavlic <fpavlic@de.ibm.com> | ||
10 | * | 3 | * |
4 | * Copyright IBM Corp. 2007 | ||
5 | * Author(s): Frank Pavlic <fpavlic@de.ibm.com>, | ||
6 | * Thomas Spatzier <tspat@de.ibm.com>, | ||
7 | * Frank Blaschka <frank.blaschka@de.ibm.com> | ||
11 | */ | 8 | */ |
12 | #ifndef __QETH_MPC_H__ | 9 | |
13 | #define __QETH_MPC_H__ | 10 | #ifndef __QETH_CORE_MPC_H__ |
11 | #define __QETH_CORE_MPC_H__ | ||
14 | 12 | ||
15 | #include <asm/qeth.h> | 13 | #include <asm/qeth.h> |
16 | 14 | ||
17 | #define IPA_PDU_HEADER_SIZE 0x40 | 15 | #define IPA_PDU_HEADER_SIZE 0x40 |
18 | #define QETH_IPA_PDU_LEN_TOTAL(buffer) (buffer+0x0e) | 16 | #define QETH_IPA_PDU_LEN_TOTAL(buffer) (buffer + 0x0e) |
19 | #define QETH_IPA_PDU_LEN_PDU1(buffer) (buffer+0x26) | 17 | #define QETH_IPA_PDU_LEN_PDU1(buffer) (buffer + 0x26) |
20 | #define QETH_IPA_PDU_LEN_PDU2(buffer) (buffer+0x29) | 18 | #define QETH_IPA_PDU_LEN_PDU2(buffer) (buffer + 0x29) |
21 | #define QETH_IPA_PDU_LEN_PDU3(buffer) (buffer+0x3a) | 19 | #define QETH_IPA_PDU_LEN_PDU3(buffer) (buffer + 0x3a) |
22 | 20 | ||
23 | extern unsigned char IPA_PDU_HEADER[]; | 21 | extern unsigned char IPA_PDU_HEADER[]; |
24 | #define QETH_IPA_CMD_DEST_ADDR(buffer) (buffer+0x2c) | 22 | #define QETH_IPA_CMD_DEST_ADDR(buffer) (buffer + 0x2c) |
25 | 23 | ||
26 | #define IPA_CMD_LENGTH (IPA_PDU_HEADER_SIZE + sizeof(struct qeth_ipa_cmd)) | 24 | #define IPA_CMD_LENGTH (IPA_PDU_HEADER_SIZE + sizeof(struct qeth_ipa_cmd)) |
27 | 25 | ||
@@ -93,7 +91,8 @@ enum qeth_checksum_types { | |||
93 | */ | 91 | */ |
94 | #define RESET_ROUTING_FLAG 0x10 /* indicate that routing type shall be set */ | 92 | #define RESET_ROUTING_FLAG 0x10 /* indicate that routing type shall be set */ |
95 | enum qeth_routing_types { | 93 | enum qeth_routing_types { |
96 | NO_ROUTER = 0, /* TODO: set to bit flag used in IPA Command */ | 94 | /* TODO: set to bit flag used in IPA Command */ |
95 | NO_ROUTER = 0, | ||
97 | PRIMARY_ROUTER = 1, | 96 | PRIMARY_ROUTER = 1, |
98 | SECONDARY_ROUTER = 2, | 97 | SECONDARY_ROUTER = 2, |
99 | MULTICAST_ROUTER = 3, | 98 | MULTICAST_ROUTER = 3, |
@@ -183,7 +182,7 @@ enum qeth_ipa_return_codes { | |||
183 | IPA_RC_SETIP_NO_STARTLAN = 0xe008, | 182 | IPA_RC_SETIP_NO_STARTLAN = 0xe008, |
184 | IPA_RC_SETIP_ALREADY_RECEIVED = 0xe009, | 183 | IPA_RC_SETIP_ALREADY_RECEIVED = 0xe009, |
185 | IPA_RC_IP_ADDR_ALREADY_USED = 0xe00a, | 184 | IPA_RC_IP_ADDR_ALREADY_USED = 0xe00a, |
186 | IPA_RC_MULTICAST_FULL = 0xe00b, | 185 | IPA_RC_MC_ADDR_NOT_FOUND = 0xe00b, |
187 | IPA_RC_SETIP_INVALID_VERSION = 0xe00d, | 186 | IPA_RC_SETIP_INVALID_VERSION = 0xe00d, |
188 | IPA_RC_UNSUPPORTED_SUBCMD = 0xe00e, | 187 | IPA_RC_UNSUPPORTED_SUBCMD = 0xe00e, |
189 | IPA_RC_ARP_ASSIST_NO_ENABLE = 0xe00f, | 188 | IPA_RC_ARP_ASSIST_NO_ENABLE = 0xe00f, |
@@ -233,14 +232,14 @@ enum qeth_ipa_setdelip_flags { | |||
233 | 232 | ||
234 | /* SETADAPTER IPA Command: ****************************************************/ | 233 | /* SETADAPTER IPA Command: ****************************************************/ |
235 | enum qeth_ipa_setadp_cmd { | 234 | enum qeth_ipa_setadp_cmd { |
236 | IPA_SETADP_QUERY_COMMANDS_SUPPORTED = 0x01, | 235 | IPA_SETADP_QUERY_COMMANDS_SUPPORTED = 0x0001, |
237 | IPA_SETADP_ALTER_MAC_ADDRESS = 0x02, | 236 | IPA_SETADP_ALTER_MAC_ADDRESS = 0x0002, |
238 | IPA_SETADP_ADD_DELETE_GROUP_ADDRESS = 0x04, | 237 | IPA_SETADP_ADD_DELETE_GROUP_ADDRESS = 0x0004, |
239 | IPA_SETADP_ADD_DELETE_FUNCTIONAL_ADDR = 0x08, | 238 | IPA_SETADP_ADD_DELETE_FUNCTIONAL_ADDR = 0x0008, |
240 | IPA_SETADP_SET_ADDRESSING_MODE = 0x10, | 239 | IPA_SETADP_SET_ADDRESSING_MODE = 0x0010, |
241 | IPA_SETADP_SET_CONFIG_PARMS = 0x20, | 240 | IPA_SETADP_SET_CONFIG_PARMS = 0x0020, |
242 | IPA_SETADP_SET_CONFIG_PARMS_EXTENDED = 0x40, | 241 | IPA_SETADP_SET_CONFIG_PARMS_EXTENDED = 0x0040, |
243 | IPA_SETADP_SET_BROADCAST_MODE = 0x80, | 242 | IPA_SETADP_SET_BROADCAST_MODE = 0x0080, |
244 | IPA_SETADP_SEND_OSA_MESSAGE = 0x0100, | 243 | IPA_SETADP_SEND_OSA_MESSAGE = 0x0100, |
245 | IPA_SETADP_SET_SNMP_CONTROL = 0x0200, | 244 | IPA_SETADP_SET_SNMP_CONTROL = 0x0200, |
246 | IPA_SETADP_QUERY_CARD_INFO = 0x0400, | 245 | IPA_SETADP_QUERY_CARD_INFO = 0x0400, |
@@ -397,26 +396,11 @@ struct qeth_ipacmd_setadpparms { | |||
397 | } data; | 396 | } data; |
398 | } __attribute__ ((packed)); | 397 | } __attribute__ ((packed)); |
399 | 398 | ||
400 | /* IPFRAME IPA Command: ***************************************************/ | ||
401 | /* TODO: define in analogy to commands define above */ | ||
402 | |||
403 | /* ADD_ADDR_ENTRY IPA Command: ********************************************/ | ||
404 | /* TODO: define in analogy to commands define above */ | ||
405 | |||
406 | /* DELETE_ADDR_ENTRY IPA Command: *****************************************/ | ||
407 | /* TODO: define in analogy to commands define above */ | ||
408 | |||
409 | /* CREATE_ADDR IPA Command: ***********************************************/ | 399 | /* CREATE_ADDR IPA Command: ***********************************************/ |
410 | struct qeth_create_destroy_address { | 400 | struct qeth_create_destroy_address { |
411 | __u8 unique_id[8]; | 401 | __u8 unique_id[8]; |
412 | } __attribute__ ((packed)); | 402 | } __attribute__ ((packed)); |
413 | 403 | ||
414 | /* REGISTER_LOCAL_ADDR IPA Command: ***************************************/ | ||
415 | /* TODO: define in analogy to commands define above */ | ||
416 | |||
417 | /* UNREGISTER_LOCAL_ADDR IPA Command: *************************************/ | ||
418 | /* TODO: define in analogy to commands define above */ | ||
419 | |||
420 | /* Header for each IPA command */ | 404 | /* Header for each IPA command */ |
421 | struct qeth_ipacmd_hdr { | 405 | struct qeth_ipacmd_hdr { |
422 | __u8 command; | 406 | __u8 command; |
@@ -463,10 +447,8 @@ enum qeth_ipa_arp_return_codes { | |||
463 | }; | 447 | }; |
464 | 448 | ||
465 | 449 | ||
466 | extern char * | 450 | extern char *qeth_get_ipa_msg(enum qeth_ipa_return_codes rc); |
467 | qeth_get_ipa_msg(enum qeth_ipa_return_codes rc); | 451 | extern char *qeth_get_ipa_cmd_name(enum qeth_ipa_cmds cmd); |
468 | extern char * | ||
469 | qeth_get_ipa_cmd_name(enum qeth_ipa_cmds cmd); | ||
470 | 452 | ||
471 | #define QETH_SETASS_BASE_LEN (sizeof(struct qeth_ipacmd_hdr) + \ | 453 | #define QETH_SETASS_BASE_LEN (sizeof(struct qeth_ipacmd_hdr) + \ |
472 | sizeof(struct qeth_ipacmd_setassparms_hdr)) | 454 | sizeof(struct qeth_ipacmd_setassparms_hdr)) |
@@ -492,88 +474,89 @@ extern unsigned char READ_CCW[]; | |||
492 | 474 | ||
493 | extern unsigned char CM_ENABLE[]; | 475 | extern unsigned char CM_ENABLE[]; |
494 | #define CM_ENABLE_SIZE 0x63 | 476 | #define CM_ENABLE_SIZE 0x63 |
495 | #define QETH_CM_ENABLE_ISSUER_RM_TOKEN(buffer) (buffer+0x2c) | 477 | #define QETH_CM_ENABLE_ISSUER_RM_TOKEN(buffer) (buffer + 0x2c) |
496 | #define QETH_CM_ENABLE_FILTER_TOKEN(buffer) (buffer+0x53) | 478 | #define QETH_CM_ENABLE_FILTER_TOKEN(buffer) (buffer + 0x53) |
497 | #define QETH_CM_ENABLE_USER_DATA(buffer) (buffer+0x5b) | 479 | #define QETH_CM_ENABLE_USER_DATA(buffer) (buffer + 0x5b) |
498 | 480 | ||
499 | #define QETH_CM_ENABLE_RESP_FILTER_TOKEN(buffer) \ | 481 | #define QETH_CM_ENABLE_RESP_FILTER_TOKEN(buffer) \ |
500 | (PDU_ENCAPSULATION(buffer)+ 0x13) | 482 | (PDU_ENCAPSULATION(buffer) + 0x13) |
501 | 483 | ||
502 | 484 | ||
503 | extern unsigned char CM_SETUP[]; | 485 | extern unsigned char CM_SETUP[]; |
504 | #define CM_SETUP_SIZE 0x64 | 486 | #define CM_SETUP_SIZE 0x64 |
505 | #define QETH_CM_SETUP_DEST_ADDR(buffer) (buffer+0x2c) | 487 | #define QETH_CM_SETUP_DEST_ADDR(buffer) (buffer + 0x2c) |
506 | #define QETH_CM_SETUP_CONNECTION_TOKEN(buffer) (buffer+0x51) | 488 | #define QETH_CM_SETUP_CONNECTION_TOKEN(buffer) (buffer + 0x51) |
507 | #define QETH_CM_SETUP_FILTER_TOKEN(buffer) (buffer+0x5a) | 489 | #define QETH_CM_SETUP_FILTER_TOKEN(buffer) (buffer + 0x5a) |
508 | 490 | ||
509 | #define QETH_CM_SETUP_RESP_DEST_ADDR(buffer) \ | 491 | #define QETH_CM_SETUP_RESP_DEST_ADDR(buffer) \ |
510 | (PDU_ENCAPSULATION(buffer) + 0x1a) | 492 | (PDU_ENCAPSULATION(buffer) + 0x1a) |
511 | 493 | ||
512 | extern unsigned char ULP_ENABLE[]; | 494 | extern unsigned char ULP_ENABLE[]; |
513 | #define ULP_ENABLE_SIZE 0x6b | 495 | #define ULP_ENABLE_SIZE 0x6b |
514 | #define QETH_ULP_ENABLE_LINKNUM(buffer) (buffer+0x61) | 496 | #define QETH_ULP_ENABLE_LINKNUM(buffer) (buffer + 0x61) |
515 | #define QETH_ULP_ENABLE_DEST_ADDR(buffer) (buffer+0x2c) | 497 | #define QETH_ULP_ENABLE_DEST_ADDR(buffer) (buffer + 0x2c) |
516 | #define QETH_ULP_ENABLE_FILTER_TOKEN(buffer) (buffer+0x53) | 498 | #define QETH_ULP_ENABLE_FILTER_TOKEN(buffer) (buffer + 0x53) |
517 | #define QETH_ULP_ENABLE_PORTNAME_AND_LL(buffer) (buffer+0x62) | 499 | #define QETH_ULP_ENABLE_PORTNAME_AND_LL(buffer) (buffer + 0x62) |
518 | #define QETH_ULP_ENABLE_RESP_FILTER_TOKEN(buffer) \ | 500 | #define QETH_ULP_ENABLE_RESP_FILTER_TOKEN(buffer) \ |
519 | (PDU_ENCAPSULATION(buffer) + 0x13) | 501 | (PDU_ENCAPSULATION(buffer) + 0x13) |
520 | #define QETH_ULP_ENABLE_RESP_MAX_MTU(buffer) \ | 502 | #define QETH_ULP_ENABLE_RESP_MAX_MTU(buffer) \ |
521 | (PDU_ENCAPSULATION(buffer)+ 0x1f) | 503 | (PDU_ENCAPSULATION(buffer) + 0x1f) |
522 | #define QETH_ULP_ENABLE_RESP_DIFINFO_LEN(buffer) \ | 504 | #define QETH_ULP_ENABLE_RESP_DIFINFO_LEN(buffer) \ |
523 | (PDU_ENCAPSULATION(buffer) + 0x17) | 505 | (PDU_ENCAPSULATION(buffer) + 0x17) |
524 | #define QETH_ULP_ENABLE_RESP_LINK_TYPE(buffer) \ | 506 | #define QETH_ULP_ENABLE_RESP_LINK_TYPE(buffer) \ |
525 | (PDU_ENCAPSULATION(buffer)+ 0x2b) | 507 | (PDU_ENCAPSULATION(buffer) + 0x2b) |
526 | /* Layer 2 defintions */ | 508 | /* Layer 2 defintions */ |
527 | #define QETH_PROT_LAYER2 0x08 | 509 | #define QETH_PROT_LAYER2 0x08 |
528 | #define QETH_PROT_TCPIP 0x03 | 510 | #define QETH_PROT_TCPIP 0x03 |
529 | #define QETH_PROT_OSN2 0x0a | 511 | #define QETH_PROT_OSN2 0x0a |
530 | #define QETH_ULP_ENABLE_PROT_TYPE(buffer) (buffer+0x50) | 512 | #define QETH_ULP_ENABLE_PROT_TYPE(buffer) (buffer + 0x50) |
531 | #define QETH_IPA_CMD_PROT_TYPE(buffer) (buffer+0x19) | 513 | #define QETH_IPA_CMD_PROT_TYPE(buffer) (buffer + 0x19) |
532 | 514 | ||
533 | extern unsigned char ULP_SETUP[]; | 515 | extern unsigned char ULP_SETUP[]; |
534 | #define ULP_SETUP_SIZE 0x6c | 516 | #define ULP_SETUP_SIZE 0x6c |
535 | #define QETH_ULP_SETUP_DEST_ADDR(buffer) (buffer+0x2c) | 517 | #define QETH_ULP_SETUP_DEST_ADDR(buffer) (buffer + 0x2c) |
536 | #define QETH_ULP_SETUP_CONNECTION_TOKEN(buffer) (buffer+0x51) | 518 | #define QETH_ULP_SETUP_CONNECTION_TOKEN(buffer) (buffer + 0x51) |
537 | #define QETH_ULP_SETUP_FILTER_TOKEN(buffer) (buffer+0x5a) | 519 | #define QETH_ULP_SETUP_FILTER_TOKEN(buffer) (buffer + 0x5a) |
538 | #define QETH_ULP_SETUP_CUA(buffer) (buffer+0x68) | 520 | #define QETH_ULP_SETUP_CUA(buffer) (buffer + 0x68) |
539 | #define QETH_ULP_SETUP_REAL_DEVADDR(buffer) (buffer+0x6a) | 521 | #define QETH_ULP_SETUP_REAL_DEVADDR(buffer) (buffer + 0x6a) |
540 | 522 | ||
541 | #define QETH_ULP_SETUP_RESP_CONNECTION_TOKEN(buffer) \ | 523 | #define QETH_ULP_SETUP_RESP_CONNECTION_TOKEN(buffer) \ |
542 | (PDU_ENCAPSULATION(buffer)+0x1a) | 524 | (PDU_ENCAPSULATION(buffer) + 0x1a) |
543 | 525 | ||
544 | 526 | ||
545 | extern unsigned char DM_ACT[]; | 527 | extern unsigned char DM_ACT[]; |
546 | #define DM_ACT_SIZE 0x55 | 528 | #define DM_ACT_SIZE 0x55 |
547 | #define QETH_DM_ACT_DEST_ADDR(buffer) (buffer+0x2c) | 529 | #define QETH_DM_ACT_DEST_ADDR(buffer) (buffer + 0x2c) |
548 | #define QETH_DM_ACT_CONNECTION_TOKEN(buffer) (buffer+0x51) | 530 | #define QETH_DM_ACT_CONNECTION_TOKEN(buffer) (buffer + 0x51) |
549 | 531 | ||
550 | 532 | ||
551 | 533 | ||
552 | #define QETH_TRANSPORT_HEADER_SEQ_NO(buffer) (buffer+4) | 534 | #define QETH_TRANSPORT_HEADER_SEQ_NO(buffer) (buffer + 4) |
553 | #define QETH_PDU_HEADER_SEQ_NO(buffer) (buffer+0x1c) | 535 | #define QETH_PDU_HEADER_SEQ_NO(buffer) (buffer + 0x1c) |
554 | #define QETH_PDU_HEADER_ACK_SEQ_NO(buffer) (buffer+0x20) | 536 | #define QETH_PDU_HEADER_ACK_SEQ_NO(buffer) (buffer + 0x20) |
555 | 537 | ||
556 | extern unsigned char IDX_ACTIVATE_READ[]; | 538 | extern unsigned char IDX_ACTIVATE_READ[]; |
557 | extern unsigned char IDX_ACTIVATE_WRITE[]; | 539 | extern unsigned char IDX_ACTIVATE_WRITE[]; |
558 | 540 | ||
559 | #define IDX_ACTIVATE_SIZE 0x22 | 541 | #define IDX_ACTIVATE_SIZE 0x22 |
560 | #define QETH_IDX_ACT_ISSUER_RM_TOKEN(buffer) (buffer+0x0c) | 542 | #define QETH_IDX_ACT_PNO(buffer) (buffer+0x0b) |
561 | #define QETH_IDX_NO_PORTNAME_REQUIRED(buffer) ((buffer)[0x0b]&0x80) | 543 | #define QETH_IDX_ACT_ISSUER_RM_TOKEN(buffer) (buffer + 0x0c) |
562 | #define QETH_IDX_ACT_FUNC_LEVEL(buffer) (buffer+0x10) | 544 | #define QETH_IDX_NO_PORTNAME_REQUIRED(buffer) ((buffer)[0x0b] & 0x80) |
563 | #define QETH_IDX_ACT_DATASET_NAME(buffer) (buffer+0x16) | 545 | #define QETH_IDX_ACT_FUNC_LEVEL(buffer) (buffer + 0x10) |
564 | #define QETH_IDX_ACT_QDIO_DEV_CUA(buffer) (buffer+0x1e) | 546 | #define QETH_IDX_ACT_DATASET_NAME(buffer) (buffer + 0x16) |
565 | #define QETH_IDX_ACT_QDIO_DEV_REALADDR(buffer) (buffer+0x20) | 547 | #define QETH_IDX_ACT_QDIO_DEV_CUA(buffer) (buffer + 0x1e) |
566 | #define QETH_IS_IDX_ACT_POS_REPLY(buffer) (((buffer)[0x08]&3)==2) | 548 | #define QETH_IDX_ACT_QDIO_DEV_REALADDR(buffer) (buffer + 0x20) |
567 | #define QETH_IDX_REPLY_LEVEL(buffer) (buffer+0x12) | 549 | #define QETH_IS_IDX_ACT_POS_REPLY(buffer) (((buffer)[0x08] & 3) == 2) |
550 | #define QETH_IDX_REPLY_LEVEL(buffer) (buffer + 0x12) | ||
568 | #define QETH_IDX_ACT_CAUSE_CODE(buffer) (buffer)[0x09] | 551 | #define QETH_IDX_ACT_CAUSE_CODE(buffer) (buffer)[0x09] |
569 | 552 | ||
570 | #define PDU_ENCAPSULATION(buffer) \ | 553 | #define PDU_ENCAPSULATION(buffer) \ |
571 | (buffer + *(buffer + (*(buffer+0x0b)) + \ | 554 | (buffer + *(buffer + (*(buffer + 0x0b)) + \ |
572 | *(buffer + *(buffer+0x0b)+0x11) +0x07)) | 555 | *(buffer + *(buffer + 0x0b) + 0x11) + 0x07)) |
573 | 556 | ||
574 | #define IS_IPA(buffer) \ | 557 | #define IS_IPA(buffer) \ |
575 | ((buffer) && \ | 558 | ((buffer) && \ |
576 | ( *(buffer + ((*(buffer+0x0b))+4) )==0xc1) ) | 559 | (*(buffer + ((*(buffer + 0x0b)) + 4)) == 0xc1)) |
577 | 560 | ||
578 | #define ADDR_FRAME_TYPE_DIX 1 | 561 | #define ADDR_FRAME_TYPE_DIX 1 |
579 | #define ADDR_FRAME_TYPE_802_3 2 | 562 | #define ADDR_FRAME_TYPE_802_3 2 |
diff --git a/drivers/s390/net/qeth_eddp.c b/drivers/s390/net/qeth_core_offl.c index e3c268cfbffe..822df8362856 100644 --- a/drivers/s390/net/qeth_eddp.c +++ b/drivers/s390/net/qeth_core_offl.c | |||
@@ -1,13 +1,11 @@ | |||
1 | /* | 1 | /* |
2 | * linux/drivers/s390/net/qeth_eddp.c | 2 | * drivers/s390/net/qeth_core_offl.c |
3 | * | ||
4 | * Enhanced Device Driver Packing (EDDP) support for the qeth driver. | ||
5 | * | ||
6 | * Copyright 2004 IBM Corporation | ||
7 | * | ||
8 | * Author(s): Thomas Spatzier <tspat@de.ibm.com> | ||
9 | * | 3 | * |
4 | * Copyright IBM Corp. 2007 | ||
5 | * Author(s): Thomas Spatzier <tspat@de.ibm.com>, | ||
6 | * Frank Blaschka <frank.blaschka@de.ibm.com> | ||
10 | */ | 7 | */ |
8 | |||
11 | #include <linux/errno.h> | 9 | #include <linux/errno.h> |
12 | #include <linux/ip.h> | 10 | #include <linux/ip.h> |
13 | #include <linux/inetdevice.h> | 11 | #include <linux/inetdevice.h> |
@@ -18,14 +16,14 @@ | |||
18 | #include <linux/skbuff.h> | 16 | #include <linux/skbuff.h> |
19 | 17 | ||
20 | #include <net/ip.h> | 18 | #include <net/ip.h> |
19 | #include <net/ip6_checksum.h> | ||
21 | 20 | ||
22 | #include "qeth.h" | 21 | #include "qeth_core.h" |
23 | #include "qeth_mpc.h" | 22 | #include "qeth_core_mpc.h" |
24 | #include "qeth_eddp.h" | 23 | #include "qeth_core_offl.h" |
25 | 24 | ||
26 | int | 25 | int qeth_eddp_check_buffers_for_context(struct qeth_qdio_out_q *queue, |
27 | qeth_eddp_check_buffers_for_context(struct qeth_qdio_out_q *queue, | 26 | struct qeth_eddp_context *ctx) |
28 | struct qeth_eddp_context *ctx) | ||
29 | { | 27 | { |
30 | int index = queue->next_buf_to_fill; | 28 | int index = queue->next_buf_to_fill; |
31 | int elements_needed = ctx->num_elements; | 29 | int elements_needed = ctx->num_elements; |
@@ -33,8 +31,8 @@ qeth_eddp_check_buffers_for_context(struct qeth_qdio_out_q *queue, | |||
33 | int skbs_in_buffer; | 31 | int skbs_in_buffer; |
34 | int buffers_needed = 0; | 32 | int buffers_needed = 0; |
35 | 33 | ||
36 | QETH_DBF_TEXT(trace, 5, "eddpcbfc"); | 34 | QETH_DBF_TEXT(TRACE, 5, "eddpcbfc"); |
37 | while(elements_needed > 0) { | 35 | while (elements_needed > 0) { |
38 | buffers_needed++; | 36 | buffers_needed++; |
39 | if (atomic_read(&queue->bufs[index].state) != | 37 | if (atomic_read(&queue->bufs[index].state) != |
40 | QETH_QDIO_BUF_EMPTY) | 38 | QETH_QDIO_BUF_EMPTY) |
@@ -49,12 +47,11 @@ qeth_eddp_check_buffers_for_context(struct qeth_qdio_out_q *queue, | |||
49 | return buffers_needed; | 47 | return buffers_needed; |
50 | } | 48 | } |
51 | 49 | ||
52 | static void | 50 | static void qeth_eddp_free_context(struct qeth_eddp_context *ctx) |
53 | qeth_eddp_free_context(struct qeth_eddp_context *ctx) | ||
54 | { | 51 | { |
55 | int i; | 52 | int i; |
56 | 53 | ||
57 | QETH_DBF_TEXT(trace, 5, "eddpfctx"); | 54 | QETH_DBF_TEXT(TRACE, 5, "eddpfctx"); |
58 | for (i = 0; i < ctx->num_pages; ++i) | 55 | for (i = 0; i < ctx->num_pages; ++i) |
59 | free_page((unsigned long)ctx->pages[i]); | 56 | free_page((unsigned long)ctx->pages[i]); |
60 | kfree(ctx->pages); | 57 | kfree(ctx->pages); |
@@ -63,26 +60,24 @@ qeth_eddp_free_context(struct qeth_eddp_context *ctx) | |||
63 | } | 60 | } |
64 | 61 | ||
65 | 62 | ||
66 | static inline void | 63 | static void qeth_eddp_get_context(struct qeth_eddp_context *ctx) |
67 | qeth_eddp_get_context(struct qeth_eddp_context *ctx) | ||
68 | { | 64 | { |
69 | atomic_inc(&ctx->refcnt); | 65 | atomic_inc(&ctx->refcnt); |
70 | } | 66 | } |
71 | 67 | ||
72 | void | 68 | void qeth_eddp_put_context(struct qeth_eddp_context *ctx) |
73 | qeth_eddp_put_context(struct qeth_eddp_context *ctx) | ||
74 | { | 69 | { |
75 | if (atomic_dec_return(&ctx->refcnt) == 0) | 70 | if (atomic_dec_return(&ctx->refcnt) == 0) |
76 | qeth_eddp_free_context(ctx); | 71 | qeth_eddp_free_context(ctx); |
77 | } | 72 | } |
73 | EXPORT_SYMBOL_GPL(qeth_eddp_put_context); | ||
78 | 74 | ||
79 | void | 75 | void qeth_eddp_buf_release_contexts(struct qeth_qdio_out_buffer *buf) |
80 | qeth_eddp_buf_release_contexts(struct qeth_qdio_out_buffer *buf) | ||
81 | { | 76 | { |
82 | struct qeth_eddp_context_reference *ref; | 77 | struct qeth_eddp_context_reference *ref; |
83 | 78 | ||
84 | QETH_DBF_TEXT(trace, 6, "eddprctx"); | 79 | QETH_DBF_TEXT(TRACE, 6, "eddprctx"); |
85 | while (!list_empty(&buf->ctx_list)){ | 80 | while (!list_empty(&buf->ctx_list)) { |
86 | ref = list_entry(buf->ctx_list.next, | 81 | ref = list_entry(buf->ctx_list.next, |
87 | struct qeth_eddp_context_reference, list); | 82 | struct qeth_eddp_context_reference, list); |
88 | qeth_eddp_put_context(ref->ctx); | 83 | qeth_eddp_put_context(ref->ctx); |
@@ -91,13 +86,12 @@ qeth_eddp_buf_release_contexts(struct qeth_qdio_out_buffer *buf) | |||
91 | } | 86 | } |
92 | } | 87 | } |
93 | 88 | ||
94 | static int | 89 | static int qeth_eddp_buf_ref_context(struct qeth_qdio_out_buffer *buf, |
95 | qeth_eddp_buf_ref_context(struct qeth_qdio_out_buffer *buf, | 90 | struct qeth_eddp_context *ctx) |
96 | struct qeth_eddp_context *ctx) | ||
97 | { | 91 | { |
98 | struct qeth_eddp_context_reference *ref; | 92 | struct qeth_eddp_context_reference *ref; |
99 | 93 | ||
100 | QETH_DBF_TEXT(trace, 6, "eddprfcx"); | 94 | QETH_DBF_TEXT(TRACE, 6, "eddprfcx"); |
101 | ref = kmalloc(sizeof(struct qeth_eddp_context_reference), GFP_ATOMIC); | 95 | ref = kmalloc(sizeof(struct qeth_eddp_context_reference), GFP_ATOMIC); |
102 | if (ref == NULL) | 96 | if (ref == NULL) |
103 | return -ENOMEM; | 97 | return -ENOMEM; |
@@ -107,10 +101,8 @@ qeth_eddp_buf_ref_context(struct qeth_qdio_out_buffer *buf, | |||
107 | return 0; | 101 | return 0; |
108 | } | 102 | } |
109 | 103 | ||
110 | int | 104 | int qeth_eddp_fill_buffer(struct qeth_qdio_out_q *queue, |
111 | qeth_eddp_fill_buffer(struct qeth_qdio_out_q *queue, | 105 | struct qeth_eddp_context *ctx, int index) |
112 | struct qeth_eddp_context *ctx, | ||
113 | int index) | ||
114 | { | 106 | { |
115 | struct qeth_qdio_out_buffer *buf = NULL; | 107 | struct qeth_qdio_out_buffer *buf = NULL; |
116 | struct qdio_buffer *buffer; | 108 | struct qdio_buffer *buffer; |
@@ -120,10 +112,10 @@ qeth_eddp_fill_buffer(struct qeth_qdio_out_q *queue, | |||
120 | int must_refcnt = 1; | 112 | int must_refcnt = 1; |
121 | int i; | 113 | int i; |
122 | 114 | ||
123 | QETH_DBF_TEXT(trace, 5, "eddpfibu"); | 115 | QETH_DBF_TEXT(TRACE, 5, "eddpfibu"); |
124 | while (elements > 0) { | 116 | while (elements > 0) { |
125 | buf = &queue->bufs[index]; | 117 | buf = &queue->bufs[index]; |
126 | if (atomic_read(&buf->state) != QETH_QDIO_BUF_EMPTY){ | 118 | if (atomic_read(&buf->state) != QETH_QDIO_BUF_EMPTY) { |
127 | /* normally this should not happen since we checked for | 119 | /* normally this should not happen since we checked for |
128 | * available elements in qeth_check_elements_for_context | 120 | * available elements in qeth_check_elements_for_context |
129 | */ | 121 | */ |
@@ -148,9 +140,9 @@ qeth_eddp_fill_buffer(struct qeth_qdio_out_q *queue, | |||
148 | must_refcnt = 1; | 140 | must_refcnt = 1; |
149 | continue; | 141 | continue; |
150 | } | 142 | } |
151 | if (must_refcnt){ | 143 | if (must_refcnt) { |
152 | must_refcnt = 0; | 144 | must_refcnt = 0; |
153 | if (qeth_eddp_buf_ref_context(buf, ctx)){ | 145 | if (qeth_eddp_buf_ref_context(buf, ctx)) { |
154 | PRINT_WARN("no memory to create eddp context " | 146 | PRINT_WARN("no memory to create eddp context " |
155 | "reference\n"); | 147 | "reference\n"); |
156 | goto out_check; | 148 | goto out_check; |
@@ -158,7 +150,7 @@ qeth_eddp_fill_buffer(struct qeth_qdio_out_q *queue, | |||
158 | } | 150 | } |
159 | buffer = buf->buffer; | 151 | buffer = buf->buffer; |
160 | /* fill one skb into buffer */ | 152 | /* fill one skb into buffer */ |
161 | for (i = 0; i < ctx->elements_per_skb; ++i){ | 153 | for (i = 0; i < ctx->elements_per_skb; ++i) { |
162 | if (ctx->elements[element].length != 0) { | 154 | if (ctx->elements[element].length != 0) { |
163 | buffer->element[buf->next_element_to_fill]. | 155 | buffer->element[buf->next_element_to_fill]. |
164 | addr = ctx->elements[element].addr; | 156 | addr = ctx->elements[element].addr; |
@@ -174,16 +166,16 @@ qeth_eddp_fill_buffer(struct qeth_qdio_out_q *queue, | |||
174 | } | 166 | } |
175 | out_check: | 167 | out_check: |
176 | if (!queue->do_pack) { | 168 | if (!queue->do_pack) { |
177 | QETH_DBF_TEXT(trace, 6, "fillbfnp"); | 169 | QETH_DBF_TEXT(TRACE, 6, "fillbfnp"); |
178 | /* set state to PRIMED -> will be flushed */ | 170 | /* set state to PRIMED -> will be flushed */ |
179 | if (buf->next_element_to_fill > 0){ | 171 | if (buf->next_element_to_fill > 0) { |
180 | atomic_set(&buf->state, QETH_QDIO_BUF_PRIMED); | 172 | atomic_set(&buf->state, QETH_QDIO_BUF_PRIMED); |
181 | flush_cnt++; | 173 | flush_cnt++; |
182 | } | 174 | } |
183 | } else { | 175 | } else { |
184 | if (queue->card->options.performance_stats) | 176 | if (queue->card->options.performance_stats) |
185 | queue->card->perf_stats.skbs_sent_pack++; | 177 | queue->card->perf_stats.skbs_sent_pack++; |
186 | QETH_DBF_TEXT(trace, 6, "fillbfpa"); | 178 | QETH_DBF_TEXT(TRACE, 6, "fillbfpa"); |
187 | if (buf->next_element_to_fill >= | 179 | if (buf->next_element_to_fill >= |
188 | QETH_MAX_BUFFER_ELEMENTS(queue->card)) { | 180 | QETH_MAX_BUFFER_ELEMENTS(queue->card)) { |
189 | /* | 181 | /* |
@@ -198,9 +190,8 @@ out: | |||
198 | return flush_cnt; | 190 | return flush_cnt; |
199 | } | 191 | } |
200 | 192 | ||
201 | static void | 193 | static void qeth_eddp_create_segment_hdrs(struct qeth_eddp_context *ctx, |
202 | qeth_eddp_create_segment_hdrs(struct qeth_eddp_context *ctx, | 194 | struct qeth_eddp_data *eddp, int data_len) |
203 | struct qeth_eddp_data *eddp, int data_len) | ||
204 | { | 195 | { |
205 | u8 *page; | 196 | u8 *page; |
206 | int page_remainder; | 197 | int page_remainder; |
@@ -208,7 +199,7 @@ qeth_eddp_create_segment_hdrs(struct qeth_eddp_context *ctx, | |||
208 | int pkt_len; | 199 | int pkt_len; |
209 | struct qeth_eddp_element *element; | 200 | struct qeth_eddp_element *element; |
210 | 201 | ||
211 | QETH_DBF_TEXT(trace, 5, "eddpcrsh"); | 202 | QETH_DBF_TEXT(TRACE, 5, "eddpcrsh"); |
212 | page = ctx->pages[ctx->offset >> PAGE_SHIFT]; | 203 | page = ctx->pages[ctx->offset >> PAGE_SHIFT]; |
213 | page_offset = ctx->offset % PAGE_SIZE; | 204 | page_offset = ctx->offset % PAGE_SIZE; |
214 | element = &ctx->elements[ctx->num_elements]; | 205 | element = &ctx->elements[ctx->num_elements]; |
@@ -220,7 +211,7 @@ qeth_eddp_create_segment_hdrs(struct qeth_eddp_context *ctx, | |||
220 | pkt_len += VLAN_HLEN; | 211 | pkt_len += VLAN_HLEN; |
221 | /* does complete packet fit in current page ? */ | 212 | /* does complete packet fit in current page ? */ |
222 | page_remainder = PAGE_SIZE - page_offset; | 213 | page_remainder = PAGE_SIZE - page_offset; |
223 | if (page_remainder < (sizeof(struct qeth_hdr) + pkt_len)){ | 214 | if (page_remainder < (sizeof(struct qeth_hdr) + pkt_len)) { |
224 | /* no -> go to start of next page */ | 215 | /* no -> go to start of next page */ |
225 | ctx->offset += page_remainder; | 216 | ctx->offset += page_remainder; |
226 | page = ctx->pages[ctx->offset >> PAGE_SHIFT]; | 217 | page = ctx->pages[ctx->offset >> PAGE_SHIFT]; |
@@ -232,14 +223,14 @@ qeth_eddp_create_segment_hdrs(struct qeth_eddp_context *ctx, | |||
232 | ctx->offset += sizeof(struct qeth_hdr); | 223 | ctx->offset += sizeof(struct qeth_hdr); |
233 | page_offset += sizeof(struct qeth_hdr); | 224 | page_offset += sizeof(struct qeth_hdr); |
234 | /* add mac header (?) */ | 225 | /* add mac header (?) */ |
235 | if (eddp->qh.hdr.l2.id == QETH_HEADER_TYPE_LAYER2){ | 226 | if (eddp->qh.hdr.l2.id == QETH_HEADER_TYPE_LAYER2) { |
236 | memcpy(page + page_offset, &eddp->mac, ETH_HLEN); | 227 | memcpy(page + page_offset, &eddp->mac, ETH_HLEN); |
237 | element->length += ETH_HLEN; | 228 | element->length += ETH_HLEN; |
238 | ctx->offset += ETH_HLEN; | 229 | ctx->offset += ETH_HLEN; |
239 | page_offset += ETH_HLEN; | 230 | page_offset += ETH_HLEN; |
240 | } | 231 | } |
241 | /* add VLAN tag */ | 232 | /* add VLAN tag */ |
242 | if (eddp->mac.h_proto == __constant_htons(ETH_P_8021Q)){ | 233 | if (eddp->mac.h_proto == __constant_htons(ETH_P_8021Q)) { |
243 | memcpy(page + page_offset, &eddp->vlan, VLAN_HLEN); | 234 | memcpy(page + page_offset, &eddp->vlan, VLAN_HLEN); |
244 | element->length += VLAN_HLEN; | 235 | element->length += VLAN_HLEN; |
245 | ctx->offset += VLAN_HLEN; | 236 | ctx->offset += VLAN_HLEN; |
@@ -258,16 +249,15 @@ qeth_eddp_create_segment_hdrs(struct qeth_eddp_context *ctx, | |||
258 | ctx->offset += eddp->thl; | 249 | ctx->offset += eddp->thl; |
259 | } | 250 | } |
260 | 251 | ||
261 | static void | 252 | static void qeth_eddp_copy_data_tcp(char *dst, struct qeth_eddp_data *eddp, |
262 | qeth_eddp_copy_data_tcp(char *dst, struct qeth_eddp_data *eddp, int len, | 253 | int len, __wsum *hcsum) |
263 | __wsum *hcsum) | ||
264 | { | 254 | { |
265 | struct skb_frag_struct *frag; | 255 | struct skb_frag_struct *frag; |
266 | int left_in_frag; | 256 | int left_in_frag; |
267 | int copy_len; | 257 | int copy_len; |
268 | u8 *src; | 258 | u8 *src; |
269 | 259 | ||
270 | QETH_DBF_TEXT(trace, 5, "eddpcdtc"); | 260 | QETH_DBF_TEXT(TRACE, 5, "eddpcdtc"); |
271 | if (skb_shinfo(eddp->skb)->nr_frags == 0) { | 261 | if (skb_shinfo(eddp->skb)->nr_frags == 0) { |
272 | skb_copy_from_linear_data_offset(eddp->skb, eddp->skb_offset, | 262 | skb_copy_from_linear_data_offset(eddp->skb, eddp->skb_offset, |
273 | dst, len); | 263 | dst, len); |
@@ -278,16 +268,17 @@ qeth_eddp_copy_data_tcp(char *dst, struct qeth_eddp_data *eddp, int len, | |||
278 | while (len > 0) { | 268 | while (len > 0) { |
279 | if (eddp->frag < 0) { | 269 | if (eddp->frag < 0) { |
280 | /* we're in skb->data */ | 270 | /* we're in skb->data */ |
281 | left_in_frag = (eddp->skb->len - eddp->skb->data_len) | 271 | left_in_frag = (eddp->skb->len - |
272 | eddp->skb->data_len) | ||
282 | - eddp->skb_offset; | 273 | - eddp->skb_offset; |
283 | src = eddp->skb->data + eddp->skb_offset; | 274 | src = eddp->skb->data + eddp->skb_offset; |
284 | } else { | 275 | } else { |
285 | frag = &skb_shinfo(eddp->skb)-> | 276 | frag = &skb_shinfo(eddp->skb)->frags[ |
286 | frags[eddp->frag]; | 277 | eddp->frag]; |
287 | left_in_frag = frag->size - eddp->frag_offset; | 278 | left_in_frag = frag->size - eddp->frag_offset; |
288 | src = (u8 *)( | 279 | src = (u8 *)((page_to_pfn(frag->page) << |
289 | (page_to_pfn(frag->page) << PAGE_SHIFT)+ | 280 | PAGE_SHIFT) + frag->page_offset + |
290 | frag->page_offset + eddp->frag_offset); | 281 | eddp->frag_offset); |
291 | } | 282 | } |
292 | if (left_in_frag <= 0) { | 283 | if (left_in_frag <= 0) { |
293 | eddp->frag++; | 284 | eddp->frag++; |
@@ -305,10 +296,8 @@ qeth_eddp_copy_data_tcp(char *dst, struct qeth_eddp_data *eddp, int len, | |||
305 | } | 296 | } |
306 | } | 297 | } |
307 | 298 | ||
308 | static void | 299 | static void qeth_eddp_create_segment_data_tcp(struct qeth_eddp_context *ctx, |
309 | qeth_eddp_create_segment_data_tcp(struct qeth_eddp_context *ctx, | 300 | struct qeth_eddp_data *eddp, int data_len, __wsum hcsum) |
310 | struct qeth_eddp_data *eddp, int data_len, | ||
311 | __wsum hcsum) | ||
312 | { | 301 | { |
313 | u8 *page; | 302 | u8 *page; |
314 | int page_remainder; | 303 | int page_remainder; |
@@ -316,13 +305,13 @@ qeth_eddp_create_segment_data_tcp(struct qeth_eddp_context *ctx, | |||
316 | struct qeth_eddp_element *element; | 305 | struct qeth_eddp_element *element; |
317 | int first_lap = 1; | 306 | int first_lap = 1; |
318 | 307 | ||
319 | QETH_DBF_TEXT(trace, 5, "eddpcsdt"); | 308 | QETH_DBF_TEXT(TRACE, 5, "eddpcsdt"); |
320 | page = ctx->pages[ctx->offset >> PAGE_SHIFT]; | 309 | page = ctx->pages[ctx->offset >> PAGE_SHIFT]; |
321 | page_offset = ctx->offset % PAGE_SIZE; | 310 | page_offset = ctx->offset % PAGE_SIZE; |
322 | element = &ctx->elements[ctx->num_elements]; | 311 | element = &ctx->elements[ctx->num_elements]; |
323 | while (data_len){ | 312 | while (data_len) { |
324 | page_remainder = PAGE_SIZE - page_offset; | 313 | page_remainder = PAGE_SIZE - page_offset; |
325 | if (page_remainder < data_len){ | 314 | if (page_remainder < data_len) { |
326 | qeth_eddp_copy_data_tcp(page + page_offset, eddp, | 315 | qeth_eddp_copy_data_tcp(page + page_offset, eddp, |
327 | page_remainder, &hcsum); | 316 | page_remainder, &hcsum); |
328 | element->length += page_remainder; | 317 | element->length += page_remainder; |
@@ -352,12 +341,12 @@ qeth_eddp_create_segment_data_tcp(struct qeth_eddp_context *ctx, | |||
352 | ((struct tcphdr *)eddp->th_in_ctx)->check = csum_fold(hcsum); | 341 | ((struct tcphdr *)eddp->th_in_ctx)->check = csum_fold(hcsum); |
353 | } | 342 | } |
354 | 343 | ||
355 | static __wsum | 344 | static __wsum qeth_eddp_check_tcp4_hdr(struct qeth_eddp_data *eddp, |
356 | qeth_eddp_check_tcp4_hdr(struct qeth_eddp_data *eddp, int data_len) | 345 | int data_len) |
357 | { | 346 | { |
358 | __wsum phcsum; /* pseudo header checksum */ | 347 | __wsum phcsum; /* pseudo header checksum */ |
359 | 348 | ||
360 | QETH_DBF_TEXT(trace, 5, "eddpckt4"); | 349 | QETH_DBF_TEXT(TRACE, 5, "eddpckt4"); |
361 | eddp->th.tcp.h.check = 0; | 350 | eddp->th.tcp.h.check = 0; |
362 | /* compute pseudo header checksum */ | 351 | /* compute pseudo header checksum */ |
363 | phcsum = csum_tcpudp_nofold(eddp->nh.ip4.h.saddr, eddp->nh.ip4.h.daddr, | 352 | phcsum = csum_tcpudp_nofold(eddp->nh.ip4.h.saddr, eddp->nh.ip4.h.daddr, |
@@ -366,13 +355,13 @@ qeth_eddp_check_tcp4_hdr(struct qeth_eddp_data *eddp, int data_len) | |||
366 | return csum_partial((u8 *)&eddp->th, eddp->thl, phcsum); | 355 | return csum_partial((u8 *)&eddp->th, eddp->thl, phcsum); |
367 | } | 356 | } |
368 | 357 | ||
369 | static __wsum | 358 | static __wsum qeth_eddp_check_tcp6_hdr(struct qeth_eddp_data *eddp, |
370 | qeth_eddp_check_tcp6_hdr(struct qeth_eddp_data *eddp, int data_len) | 359 | int data_len) |
371 | { | 360 | { |
372 | __be32 proto; | 361 | __be32 proto; |
373 | __wsum phcsum; /* pseudo header checksum */ | 362 | __wsum phcsum; /* pseudo header checksum */ |
374 | 363 | ||
375 | QETH_DBF_TEXT(trace, 5, "eddpckt6"); | 364 | QETH_DBF_TEXT(TRACE, 5, "eddpckt6"); |
376 | eddp->th.tcp.h.check = 0; | 365 | eddp->th.tcp.h.check = 0; |
377 | /* compute pseudo header checksum */ | 366 | /* compute pseudo header checksum */ |
378 | phcsum = csum_partial((u8 *)&eddp->nh.ip6.h.saddr, | 367 | phcsum = csum_partial((u8 *)&eddp->nh.ip6.h.saddr, |
@@ -384,14 +373,14 @@ qeth_eddp_check_tcp6_hdr(struct qeth_eddp_data *eddp, int data_len) | |||
384 | return phcsum; | 373 | return phcsum; |
385 | } | 374 | } |
386 | 375 | ||
387 | static struct qeth_eddp_data * | 376 | static struct qeth_eddp_data *qeth_eddp_create_eddp_data(struct qeth_hdr *qh, |
388 | qeth_eddp_create_eddp_data(struct qeth_hdr *qh, u8 *nh, u8 nhl, u8 *th, u8 thl) | 377 | u8 *nh, u8 nhl, u8 *th, u8 thl) |
389 | { | 378 | { |
390 | struct qeth_eddp_data *eddp; | 379 | struct qeth_eddp_data *eddp; |
391 | 380 | ||
392 | QETH_DBF_TEXT(trace, 5, "eddpcrda"); | 381 | QETH_DBF_TEXT(TRACE, 5, "eddpcrda"); |
393 | eddp = kzalloc(sizeof(struct qeth_eddp_data), GFP_ATOMIC); | 382 | eddp = kzalloc(sizeof(struct qeth_eddp_data), GFP_ATOMIC); |
394 | if (eddp){ | 383 | if (eddp) { |
395 | eddp->nhl = nhl; | 384 | eddp->nhl = nhl; |
396 | eddp->thl = thl; | 385 | eddp->thl = thl; |
397 | memcpy(&eddp->qh, qh, sizeof(struct qeth_hdr)); | 386 | memcpy(&eddp->qh, qh, sizeof(struct qeth_hdr)); |
@@ -402,40 +391,35 @@ qeth_eddp_create_eddp_data(struct qeth_hdr *qh, u8 *nh, u8 nhl, u8 *th, u8 thl) | |||
402 | return eddp; | 391 | return eddp; |
403 | } | 392 | } |
404 | 393 | ||
405 | static void | 394 | static void __qeth_eddp_fill_context_tcp(struct qeth_eddp_context *ctx, |
406 | __qeth_eddp_fill_context_tcp(struct qeth_eddp_context *ctx, | 395 | struct qeth_eddp_data *eddp) |
407 | struct qeth_eddp_data *eddp) | ||
408 | { | 396 | { |
409 | struct tcphdr *tcph; | 397 | struct tcphdr *tcph; |
410 | int data_len; | 398 | int data_len; |
411 | __wsum hcsum; | 399 | __wsum hcsum; |
412 | 400 | ||
413 | QETH_DBF_TEXT(trace, 5, "eddpftcp"); | 401 | QETH_DBF_TEXT(TRACE, 5, "eddpftcp"); |
414 | eddp->skb_offset = sizeof(struct qeth_hdr) + eddp->nhl + eddp->thl; | 402 | eddp->skb_offset = sizeof(struct qeth_hdr) + eddp->nhl + eddp->thl; |
415 | if (eddp->qh.hdr.l2.id == QETH_HEADER_TYPE_LAYER2) { | 403 | if (eddp->qh.hdr.l2.id == QETH_HEADER_TYPE_LAYER2) { |
416 | eddp->skb_offset += sizeof(struct ethhdr); | 404 | eddp->skb_offset += sizeof(struct ethhdr); |
417 | #ifdef CONFIG_QETH_VLAN | 405 | if (eddp->mac.h_proto == __constant_htons(ETH_P_8021Q)) |
418 | if (eddp->mac.h_proto == __constant_htons(ETH_P_8021Q)) | 406 | eddp->skb_offset += VLAN_HLEN; |
419 | eddp->skb_offset += VLAN_HLEN; | 407 | } |
420 | #endif /* CONFIG_QETH_VLAN */ | ||
421 | } | ||
422 | tcph = tcp_hdr(eddp->skb); | 408 | tcph = tcp_hdr(eddp->skb); |
423 | while (eddp->skb_offset < eddp->skb->len) { | 409 | while (eddp->skb_offset < eddp->skb->len) { |
424 | data_len = min((int)skb_shinfo(eddp->skb)->gso_size, | 410 | data_len = min((int)skb_shinfo(eddp->skb)->gso_size, |
425 | (int)(eddp->skb->len - eddp->skb_offset)); | 411 | (int)(eddp->skb->len - eddp->skb_offset)); |
426 | /* prepare qdio hdr */ | 412 | /* prepare qdio hdr */ |
427 | if (eddp->qh.hdr.l2.id == QETH_HEADER_TYPE_LAYER2){ | 413 | if (eddp->qh.hdr.l2.id == QETH_HEADER_TYPE_LAYER2) { |
428 | eddp->qh.hdr.l2.pkt_length = data_len + ETH_HLEN + | 414 | eddp->qh.hdr.l2.pkt_length = data_len + ETH_HLEN + |
429 | eddp->nhl + eddp->thl; | 415 | eddp->nhl + eddp->thl; |
430 | #ifdef CONFIG_QETH_VLAN | ||
431 | if (eddp->mac.h_proto == __constant_htons(ETH_P_8021Q)) | 416 | if (eddp->mac.h_proto == __constant_htons(ETH_P_8021Q)) |
432 | eddp->qh.hdr.l2.pkt_length += VLAN_HLEN; | 417 | eddp->qh.hdr.l2.pkt_length += VLAN_HLEN; |
433 | #endif /* CONFIG_QETH_VLAN */ | ||
434 | } else | 418 | } else |
435 | eddp->qh.hdr.l3.length = data_len + eddp->nhl + | 419 | eddp->qh.hdr.l3.length = data_len + eddp->nhl + |
436 | eddp->thl; | 420 | eddp->thl; |
437 | /* prepare ip hdr */ | 421 | /* prepare ip hdr */ |
438 | if (eddp->skb->protocol == htons(ETH_P_IP)){ | 422 | if (eddp->skb->protocol == htons(ETH_P_IP)) { |
439 | eddp->nh.ip4.h.tot_len = htons(data_len + eddp->nhl + | 423 | eddp->nh.ip4.h.tot_len = htons(data_len + eddp->nhl + |
440 | eddp->thl); | 424 | eddp->thl); |
441 | eddp->nh.ip4.h.check = 0; | 425 | eddp->nh.ip4.h.check = 0; |
@@ -443,9 +427,10 @@ __qeth_eddp_fill_context_tcp(struct qeth_eddp_context *ctx, | |||
443 | ip_fast_csum((u8 *)&eddp->nh.ip4.h, | 427 | ip_fast_csum((u8 *)&eddp->nh.ip4.h, |
444 | eddp->nh.ip4.h.ihl); | 428 | eddp->nh.ip4.h.ihl); |
445 | } else | 429 | } else |
446 | eddp->nh.ip6.h.payload_len = htons(data_len + eddp->thl); | 430 | eddp->nh.ip6.h.payload_len = htons(data_len + |
431 | eddp->thl); | ||
447 | /* prepare tcp hdr */ | 432 | /* prepare tcp hdr */ |
448 | if (data_len == (eddp->skb->len - eddp->skb_offset)){ | 433 | if (data_len == (eddp->skb->len - eddp->skb_offset)) { |
449 | /* last segment -> set FIN and PSH flags */ | 434 | /* last segment -> set FIN and PSH flags */ |
450 | eddp->th.tcp.h.fin = tcph->fin; | 435 | eddp->th.tcp.h.fin = tcph->fin; |
451 | eddp->th.tcp.h.psh = tcph->psh; | 436 | eddp->th.tcp.h.psh = tcph->psh; |
@@ -462,17 +447,17 @@ __qeth_eddp_fill_context_tcp(struct qeth_eddp_context *ctx, | |||
462 | /* prepare headers for next round */ | 447 | /* prepare headers for next round */ |
463 | if (eddp->skb->protocol == htons(ETH_P_IP)) | 448 | if (eddp->skb->protocol == htons(ETH_P_IP)) |
464 | eddp->nh.ip4.h.id = htons(ntohs(eddp->nh.ip4.h.id) + 1); | 449 | eddp->nh.ip4.h.id = htons(ntohs(eddp->nh.ip4.h.id) + 1); |
465 | eddp->th.tcp.h.seq = htonl(ntohl(eddp->th.tcp.h.seq) + data_len); | 450 | eddp->th.tcp.h.seq = htonl(ntohl(eddp->th.tcp.h.seq) + |
451 | data_len); | ||
466 | } | 452 | } |
467 | } | 453 | } |
468 | 454 | ||
469 | static int | 455 | static int qeth_eddp_fill_context_tcp(struct qeth_eddp_context *ctx, |
470 | qeth_eddp_fill_context_tcp(struct qeth_eddp_context *ctx, | 456 | struct sk_buff *skb, struct qeth_hdr *qhdr) |
471 | struct sk_buff *skb, struct qeth_hdr *qhdr) | ||
472 | { | 457 | { |
473 | struct qeth_eddp_data *eddp = NULL; | 458 | struct qeth_eddp_data *eddp = NULL; |
474 | 459 | ||
475 | QETH_DBF_TEXT(trace, 5, "eddpficx"); | 460 | QETH_DBF_TEXT(TRACE, 5, "eddpficx"); |
476 | /* create our segmentation headers and copy original headers */ | 461 | /* create our segmentation headers and copy original headers */ |
477 | if (skb->protocol == htons(ETH_P_IP)) | 462 | if (skb->protocol == htons(ETH_P_IP)) |
478 | eddp = qeth_eddp_create_eddp_data(qhdr, | 463 | eddp = qeth_eddp_create_eddp_data(qhdr, |
@@ -488,18 +473,16 @@ qeth_eddp_fill_context_tcp(struct qeth_eddp_context *ctx, | |||
488 | tcp_hdrlen(skb)); | 473 | tcp_hdrlen(skb)); |
489 | 474 | ||
490 | if (eddp == NULL) { | 475 | if (eddp == NULL) { |
491 | QETH_DBF_TEXT(trace, 2, "eddpfcnm"); | 476 | QETH_DBF_TEXT(TRACE, 2, "eddpfcnm"); |
492 | return -ENOMEM; | 477 | return -ENOMEM; |
493 | } | 478 | } |
494 | if (qhdr->hdr.l2.id == QETH_HEADER_TYPE_LAYER2) { | 479 | if (qhdr->hdr.l2.id == QETH_HEADER_TYPE_LAYER2) { |
495 | skb_set_mac_header(skb, sizeof(struct qeth_hdr)); | 480 | skb_set_mac_header(skb, sizeof(struct qeth_hdr)); |
496 | memcpy(&eddp->mac, eth_hdr(skb), ETH_HLEN); | 481 | memcpy(&eddp->mac, eth_hdr(skb), ETH_HLEN); |
497 | #ifdef CONFIG_QETH_VLAN | ||
498 | if (eddp->mac.h_proto == __constant_htons(ETH_P_8021Q)) { | 482 | if (eddp->mac.h_proto == __constant_htons(ETH_P_8021Q)) { |
499 | eddp->vlan[0] = skb->protocol; | 483 | eddp->vlan[0] = skb->protocol; |
500 | eddp->vlan[1] = htons(vlan_tx_tag_get(skb)); | 484 | eddp->vlan[1] = htons(vlan_tx_tag_get(skb)); |
501 | } | 485 | } |
502 | #endif /* CONFIG_QETH_VLAN */ | ||
503 | } | 486 | } |
504 | /* the next flags will only be set on the last segment */ | 487 | /* the next flags will only be set on the last segment */ |
505 | eddp->th.tcp.h.fin = 0; | 488 | eddp->th.tcp.h.fin = 0; |
@@ -511,16 +494,15 @@ qeth_eddp_fill_context_tcp(struct qeth_eddp_context *ctx, | |||
511 | return 0; | 494 | return 0; |
512 | } | 495 | } |
513 | 496 | ||
514 | static void | 497 | static void qeth_eddp_calc_num_pages(struct qeth_eddp_context *ctx, |
515 | qeth_eddp_calc_num_pages(struct qeth_eddp_context *ctx, struct sk_buff *skb, | 498 | struct sk_buff *skb, int hdr_len) |
516 | int hdr_len) | ||
517 | { | 499 | { |
518 | int skbs_per_page; | 500 | int skbs_per_page; |
519 | 501 | ||
520 | QETH_DBF_TEXT(trace, 5, "eddpcanp"); | 502 | QETH_DBF_TEXT(TRACE, 5, "eddpcanp"); |
521 | /* can we put multiple skbs in one page? */ | 503 | /* can we put multiple skbs in one page? */ |
522 | skbs_per_page = PAGE_SIZE / (skb_shinfo(skb)->gso_size + hdr_len); | 504 | skbs_per_page = PAGE_SIZE / (skb_shinfo(skb)->gso_size + hdr_len); |
523 | if (skbs_per_page > 1){ | 505 | if (skbs_per_page > 1) { |
524 | ctx->num_pages = (skb_shinfo(skb)->gso_segs + 1) / | 506 | ctx->num_pages = (skb_shinfo(skb)->gso_segs + 1) / |
525 | skbs_per_page + 1; | 507 | skbs_per_page + 1; |
526 | ctx->elements_per_skb = 1; | 508 | ctx->elements_per_skb = 1; |
@@ -535,49 +517,47 @@ qeth_eddp_calc_num_pages(struct qeth_eddp_context *ctx, struct sk_buff *skb, | |||
535 | (skb_shinfo(skb)->gso_segs + 1); | 517 | (skb_shinfo(skb)->gso_segs + 1); |
536 | } | 518 | } |
537 | 519 | ||
538 | static struct qeth_eddp_context * | 520 | static struct qeth_eddp_context *qeth_eddp_create_context_generic( |
539 | qeth_eddp_create_context_generic(struct qeth_card *card, struct sk_buff *skb, | 521 | struct qeth_card *card, struct sk_buff *skb, int hdr_len) |
540 | int hdr_len) | ||
541 | { | 522 | { |
542 | struct qeth_eddp_context *ctx = NULL; | 523 | struct qeth_eddp_context *ctx = NULL; |
543 | u8 *addr; | 524 | u8 *addr; |
544 | int i; | 525 | int i; |
545 | 526 | ||
546 | QETH_DBF_TEXT(trace, 5, "creddpcg"); | 527 | QETH_DBF_TEXT(TRACE, 5, "creddpcg"); |
547 | /* create the context and allocate pages */ | 528 | /* create the context and allocate pages */ |
548 | ctx = kzalloc(sizeof(struct qeth_eddp_context), GFP_ATOMIC); | 529 | ctx = kzalloc(sizeof(struct qeth_eddp_context), GFP_ATOMIC); |
549 | if (ctx == NULL){ | 530 | if (ctx == NULL) { |
550 | QETH_DBF_TEXT(trace, 2, "ceddpcn1"); | 531 | QETH_DBF_TEXT(TRACE, 2, "ceddpcn1"); |
551 | return NULL; | 532 | return NULL; |
552 | } | 533 | } |
553 | ctx->type = QETH_LARGE_SEND_EDDP; | 534 | ctx->type = QETH_LARGE_SEND_EDDP; |
554 | qeth_eddp_calc_num_pages(ctx, skb, hdr_len); | 535 | qeth_eddp_calc_num_pages(ctx, skb, hdr_len); |
555 | if (ctx->elements_per_skb > QETH_MAX_BUFFER_ELEMENTS(card)){ | 536 | if (ctx->elements_per_skb > QETH_MAX_BUFFER_ELEMENTS(card)) { |
556 | QETH_DBF_TEXT(trace, 2, "ceddpcis"); | 537 | QETH_DBF_TEXT(TRACE, 2, "ceddpcis"); |
557 | kfree(ctx); | 538 | kfree(ctx); |
558 | return NULL; | 539 | return NULL; |
559 | } | 540 | } |
560 | ctx->pages = kcalloc(ctx->num_pages, sizeof(u8 *), GFP_ATOMIC); | 541 | ctx->pages = kcalloc(ctx->num_pages, sizeof(u8 *), GFP_ATOMIC); |
561 | if (ctx->pages == NULL){ | 542 | if (ctx->pages == NULL) { |
562 | QETH_DBF_TEXT(trace, 2, "ceddpcn2"); | 543 | QETH_DBF_TEXT(TRACE, 2, "ceddpcn2"); |
563 | kfree(ctx); | 544 | kfree(ctx); |
564 | return NULL; | 545 | return NULL; |
565 | } | 546 | } |
566 | for (i = 0; i < ctx->num_pages; ++i){ | 547 | for (i = 0; i < ctx->num_pages; ++i) { |
567 | addr = (u8 *)__get_free_page(GFP_ATOMIC); | 548 | addr = (u8 *)get_zeroed_page(GFP_ATOMIC); |
568 | if (addr == NULL){ | 549 | if (addr == NULL) { |
569 | QETH_DBF_TEXT(trace, 2, "ceddpcn3"); | 550 | QETH_DBF_TEXT(TRACE, 2, "ceddpcn3"); |
570 | ctx->num_pages = i; | 551 | ctx->num_pages = i; |
571 | qeth_eddp_free_context(ctx); | 552 | qeth_eddp_free_context(ctx); |
572 | return NULL; | 553 | return NULL; |
573 | } | 554 | } |
574 | memset(addr, 0, PAGE_SIZE); | ||
575 | ctx->pages[i] = addr; | 555 | ctx->pages[i] = addr; |
576 | } | 556 | } |
577 | ctx->elements = kcalloc(ctx->num_elements, | 557 | ctx->elements = kcalloc(ctx->num_elements, |
578 | sizeof(struct qeth_eddp_element), GFP_ATOMIC); | 558 | sizeof(struct qeth_eddp_element), GFP_ATOMIC); |
579 | if (ctx->elements == NULL){ | 559 | if (ctx->elements == NULL) { |
580 | QETH_DBF_TEXT(trace, 2, "ceddpcn4"); | 560 | QETH_DBF_TEXT(TRACE, 2, "ceddpcn4"); |
581 | qeth_eddp_free_context(ctx); | 561 | qeth_eddp_free_context(ctx); |
582 | return NULL; | 562 | return NULL; |
583 | } | 563 | } |
@@ -587,31 +567,31 @@ qeth_eddp_create_context_generic(struct qeth_card *card, struct sk_buff *skb, | |||
587 | return ctx; | 567 | return ctx; |
588 | } | 568 | } |
589 | 569 | ||
590 | static struct qeth_eddp_context * | 570 | static struct qeth_eddp_context *qeth_eddp_create_context_tcp( |
591 | qeth_eddp_create_context_tcp(struct qeth_card *card, struct sk_buff *skb, | 571 | struct qeth_card *card, struct sk_buff *skb, |
592 | struct qeth_hdr *qhdr) | 572 | struct qeth_hdr *qhdr) |
593 | { | 573 | { |
594 | struct qeth_eddp_context *ctx = NULL; | 574 | struct qeth_eddp_context *ctx = NULL; |
595 | 575 | ||
596 | QETH_DBF_TEXT(trace, 5, "creddpct"); | 576 | QETH_DBF_TEXT(TRACE, 5, "creddpct"); |
597 | if (skb->protocol == htons(ETH_P_IP)) | 577 | if (skb->protocol == htons(ETH_P_IP)) |
598 | ctx = qeth_eddp_create_context_generic(card, skb, | 578 | ctx = qeth_eddp_create_context_generic(card, skb, |
599 | (sizeof(struct qeth_hdr) + | 579 | (sizeof(struct qeth_hdr) + |
600 | ip_hdrlen(skb) + | 580 | ip_hdrlen(skb) + |
601 | tcp_hdrlen(skb))); | 581 | tcp_hdrlen(skb))); |
602 | else if (skb->protocol == htons(ETH_P_IPV6)) | 582 | else if (skb->protocol == htons(ETH_P_IPV6)) |
603 | ctx = qeth_eddp_create_context_generic(card, skb, | 583 | ctx = qeth_eddp_create_context_generic(card, skb, |
604 | sizeof(struct qeth_hdr) + sizeof(struct ipv6hdr) + | 584 | sizeof(struct qeth_hdr) + sizeof(struct ipv6hdr) + |
605 | tcp_hdrlen(skb)); | 585 | tcp_hdrlen(skb)); |
606 | else | 586 | else |
607 | QETH_DBF_TEXT(trace, 2, "cetcpinv"); | 587 | QETH_DBF_TEXT(TRACE, 2, "cetcpinv"); |
608 | 588 | ||
609 | if (ctx == NULL) { | 589 | if (ctx == NULL) { |
610 | QETH_DBF_TEXT(trace, 2, "creddpnl"); | 590 | QETH_DBF_TEXT(TRACE, 2, "creddpnl"); |
611 | return NULL; | 591 | return NULL; |
612 | } | 592 | } |
613 | if (qeth_eddp_fill_context_tcp(ctx, skb, qhdr)){ | 593 | if (qeth_eddp_fill_context_tcp(ctx, skb, qhdr)) { |
614 | QETH_DBF_TEXT(trace, 2, "ceddptfe"); | 594 | QETH_DBF_TEXT(TRACE, 2, "ceddptfe"); |
615 | qeth_eddp_free_context(ctx); | 595 | qeth_eddp_free_context(ctx); |
616 | return NULL; | 596 | return NULL; |
617 | } | 597 | } |
@@ -619,16 +599,103 @@ qeth_eddp_create_context_tcp(struct qeth_card *card, struct sk_buff *skb, | |||
619 | return ctx; | 599 | return ctx; |
620 | } | 600 | } |
621 | 601 | ||
622 | struct qeth_eddp_context * | 602 | struct qeth_eddp_context *qeth_eddp_create_context(struct qeth_card *card, |
623 | qeth_eddp_create_context(struct qeth_card *card, struct sk_buff *skb, | 603 | struct sk_buff *skb, struct qeth_hdr *qhdr, |
624 | struct qeth_hdr *qhdr, unsigned char sk_protocol) | 604 | unsigned char sk_protocol) |
625 | { | 605 | { |
626 | QETH_DBF_TEXT(trace, 5, "creddpc"); | 606 | QETH_DBF_TEXT(TRACE, 5, "creddpc"); |
627 | switch (sk_protocol) { | 607 | switch (sk_protocol) { |
628 | case IPPROTO_TCP: | 608 | case IPPROTO_TCP: |
629 | return qeth_eddp_create_context_tcp(card, skb, qhdr); | 609 | return qeth_eddp_create_context_tcp(card, skb, qhdr); |
630 | default: | 610 | default: |
631 | QETH_DBF_TEXT(trace, 2, "eddpinvp"); | 611 | QETH_DBF_TEXT(TRACE, 2, "eddpinvp"); |
632 | } | 612 | } |
633 | return NULL; | 613 | return NULL; |
634 | } | 614 | } |
615 | EXPORT_SYMBOL_GPL(qeth_eddp_create_context); | ||
616 | |||
617 | void qeth_tso_fill_header(struct qeth_card *card, struct qeth_hdr *qhdr, | ||
618 | struct sk_buff *skb) | ||
619 | { | ||
620 | struct qeth_hdr_tso *hdr = (struct qeth_hdr_tso *)qhdr; | ||
621 | struct tcphdr *tcph = tcp_hdr(skb); | ||
622 | struct iphdr *iph = ip_hdr(skb); | ||
623 | struct ipv6hdr *ip6h = ipv6_hdr(skb); | ||
624 | |||
625 | QETH_DBF_TEXT(TRACE, 5, "tsofhdr"); | ||
626 | |||
627 | /*fix header to TSO values ...*/ | ||
628 | hdr->hdr.hdr.l3.id = QETH_HEADER_TYPE_TSO; | ||
629 | /*set values which are fix for the first approach ...*/ | ||
630 | hdr->ext.hdr_tot_len = (__u16) sizeof(struct qeth_hdr_ext_tso); | ||
631 | hdr->ext.imb_hdr_no = 1; | ||
632 | hdr->ext.hdr_type = 1; | ||
633 | hdr->ext.hdr_version = 1; | ||
634 | hdr->ext.hdr_len = 28; | ||
635 | /*insert non-fix values */ | ||
636 | hdr->ext.mss = skb_shinfo(skb)->gso_size; | ||
637 | hdr->ext.dg_hdr_len = (__u16)(iph->ihl*4 + tcph->doff*4); | ||
638 | hdr->ext.payload_len = (__u16)(skb->len - hdr->ext.dg_hdr_len - | ||
639 | sizeof(struct qeth_hdr_tso)); | ||
640 | tcph->check = 0; | ||
641 | if (skb->protocol == ETH_P_IPV6) { | ||
642 | ip6h->payload_len = 0; | ||
643 | tcph->check = ~csum_ipv6_magic(&ip6h->saddr, &ip6h->daddr, | ||
644 | 0, IPPROTO_TCP, 0); | ||
645 | } else { | ||
646 | /*OSA want us to set these values ...*/ | ||
647 | tcph->check = ~csum_tcpudp_magic(iph->saddr, iph->daddr, | ||
648 | 0, IPPROTO_TCP, 0); | ||
649 | iph->tot_len = 0; | ||
650 | iph->check = 0; | ||
651 | } | ||
652 | } | ||
653 | EXPORT_SYMBOL_GPL(qeth_tso_fill_header); | ||
654 | |||
655 | void qeth_tx_csum(struct sk_buff *skb) | ||
656 | { | ||
657 | int tlen; | ||
658 | if (skb->protocol == htons(ETH_P_IP)) { | ||
659 | tlen = ntohs(ip_hdr(skb)->tot_len) - (ip_hdr(skb)->ihl << 2); | ||
660 | switch (ip_hdr(skb)->protocol) { | ||
661 | case IPPROTO_TCP: | ||
662 | tcp_hdr(skb)->check = 0; | ||
663 | tcp_hdr(skb)->check = csum_tcpudp_magic( | ||
664 | ip_hdr(skb)->saddr, ip_hdr(skb)->daddr, | ||
665 | tlen, ip_hdr(skb)->protocol, | ||
666 | skb_checksum(skb, skb_transport_offset(skb), | ||
667 | tlen, 0)); | ||
668 | break; | ||
669 | case IPPROTO_UDP: | ||
670 | udp_hdr(skb)->check = 0; | ||
671 | udp_hdr(skb)->check = csum_tcpudp_magic( | ||
672 | ip_hdr(skb)->saddr, ip_hdr(skb)->daddr, | ||
673 | tlen, ip_hdr(skb)->protocol, | ||
674 | skb_checksum(skb, skb_transport_offset(skb), | ||
675 | tlen, 0)); | ||
676 | break; | ||
677 | } | ||
678 | } else if (skb->protocol == htons(ETH_P_IPV6)) { | ||
679 | switch (ipv6_hdr(skb)->nexthdr) { | ||
680 | case IPPROTO_TCP: | ||
681 | tcp_hdr(skb)->check = 0; | ||
682 | tcp_hdr(skb)->check = csum_ipv6_magic( | ||
683 | &ipv6_hdr(skb)->saddr, &ipv6_hdr(skb)->daddr, | ||
684 | ipv6_hdr(skb)->payload_len, | ||
685 | ipv6_hdr(skb)->nexthdr, | ||
686 | skb_checksum(skb, skb_transport_offset(skb), | ||
687 | ipv6_hdr(skb)->payload_len, 0)); | ||
688 | break; | ||
689 | case IPPROTO_UDP: | ||
690 | udp_hdr(skb)->check = 0; | ||
691 | udp_hdr(skb)->check = csum_ipv6_magic( | ||
692 | &ipv6_hdr(skb)->saddr, &ipv6_hdr(skb)->daddr, | ||
693 | ipv6_hdr(skb)->payload_len, | ||
694 | ipv6_hdr(skb)->nexthdr, | ||
695 | skb_checksum(skb, skb_transport_offset(skb), | ||
696 | ipv6_hdr(skb)->payload_len, 0)); | ||
697 | break; | ||
698 | } | ||
699 | } | ||
700 | } | ||
701 | EXPORT_SYMBOL_GPL(qeth_tx_csum); | ||
diff --git a/drivers/s390/net/qeth_eddp.h b/drivers/s390/net/qeth_core_offl.h index 52910c9252c0..86bf7df8cf16 100644 --- a/drivers/s390/net/qeth_eddp.h +++ b/drivers/s390/net/qeth_core_offl.h | |||
@@ -1,15 +1,13 @@ | |||
1 | /* | 1 | /* |
2 | * linux/drivers/s390/net/qeth_eddp.h | 2 | * drivers/s390/net/qeth_core_offl.h |
3 | * | ||
4 | * Header file for qeth enhanced device driver packing. | ||
5 | * | ||
6 | * Copyright 2004 IBM Corporation | ||
7 | * | ||
8 | * Author(s): Thomas Spatzier <tspat@de.ibm.com> | ||
9 | * | 3 | * |
4 | * Copyright IBM Corp. 2007 | ||
5 | * Author(s): Thomas Spatzier <tspat@de.ibm.com>, | ||
6 | * Frank Blaschka <frank.blaschka@de.ibm.com> | ||
10 | */ | 7 | */ |
11 | #ifndef __QETH_EDDP_H__ | 8 | |
12 | #define __QETH_EDDP_H__ | 9 | #ifndef __QETH_CORE_OFFL_H__ |
10 | #define __QETH_CORE_OFFL_H__ | ||
13 | 11 | ||
14 | struct qeth_eddp_element { | 12 | struct qeth_eddp_element { |
15 | u32 flags; | 13 | u32 flags; |
@@ -33,25 +31,6 @@ struct qeth_eddp_context_reference { | |||
33 | struct qeth_eddp_context *ctx; | 31 | struct qeth_eddp_context *ctx; |
34 | }; | 32 | }; |
35 | 33 | ||
36 | extern struct qeth_eddp_context * | ||
37 | qeth_eddp_create_context(struct qeth_card *,struct sk_buff *, | ||
38 | struct qeth_hdr *, unsigned char); | ||
39 | |||
40 | extern void | ||
41 | qeth_eddp_put_context(struct qeth_eddp_context *); | ||
42 | |||
43 | extern int | ||
44 | qeth_eddp_fill_buffer(struct qeth_qdio_out_q *,struct qeth_eddp_context *,int); | ||
45 | |||
46 | extern void | ||
47 | qeth_eddp_buf_release_contexts(struct qeth_qdio_out_buffer *); | ||
48 | |||
49 | extern int | ||
50 | qeth_eddp_check_buffers_for_context(struct qeth_qdio_out_q *, | ||
51 | struct qeth_eddp_context *); | ||
52 | /* | ||
53 | * Data used for fragmenting a IP packet. | ||
54 | */ | ||
55 | struct qeth_eddp_data { | 34 | struct qeth_eddp_data { |
56 | struct qeth_hdr qh; | 35 | struct qeth_hdr qh; |
57 | struct ethhdr mac; | 36 | struct ethhdr mac; |
@@ -81,4 +60,17 @@ struct qeth_eddp_data { | |||
81 | int frag_offset; | 60 | int frag_offset; |
82 | } __attribute__ ((packed)); | 61 | } __attribute__ ((packed)); |
83 | 62 | ||
84 | #endif /* __QETH_EDDP_H__ */ | 63 | extern struct qeth_eddp_context *qeth_eddp_create_context(struct qeth_card *, |
64 | struct sk_buff *, struct qeth_hdr *, unsigned char); | ||
65 | extern void qeth_eddp_put_context(struct qeth_eddp_context *); | ||
66 | extern int qeth_eddp_fill_buffer(struct qeth_qdio_out_q *, | ||
67 | struct qeth_eddp_context *, int); | ||
68 | extern void qeth_eddp_buf_release_contexts(struct qeth_qdio_out_buffer *); | ||
69 | extern int qeth_eddp_check_buffers_for_context(struct qeth_qdio_out_q *, | ||
70 | struct qeth_eddp_context *); | ||
71 | |||
72 | void qeth_tso_fill_header(struct qeth_card *, struct qeth_hdr *, | ||
73 | struct sk_buff *); | ||
74 | void qeth_tx_csum(struct sk_buff *skb); | ||
75 | |||
76 | #endif /* __QETH_CORE_EDDP_H__ */ | ||
diff --git a/drivers/s390/net/qeth_core_sys.c b/drivers/s390/net/qeth_core_sys.c new file mode 100644 index 000000000000..08a50f057284 --- /dev/null +++ b/drivers/s390/net/qeth_core_sys.c | |||
@@ -0,0 +1,651 @@ | |||
1 | /* | ||
2 | * drivers/s390/net/qeth_core_sys.c | ||
3 | * | ||
4 | * Copyright IBM Corp. 2007 | ||
5 | * Author(s): Utz Bacher <utz.bacher@de.ibm.com>, | ||
6 | * Frank Pavlic <fpavlic@de.ibm.com>, | ||
7 | * Thomas Spatzier <tspat@de.ibm.com>, | ||
8 | * Frank Blaschka <frank.blaschka@de.ibm.com> | ||
9 | */ | ||
10 | |||
11 | #include <linux/list.h> | ||
12 | #include <linux/rwsem.h> | ||
13 | #include <asm/ebcdic.h> | ||
14 | |||
15 | #include "qeth_core.h" | ||
16 | |||
17 | static ssize_t qeth_dev_state_show(struct device *dev, | ||
18 | struct device_attribute *attr, char *buf) | ||
19 | { | ||
20 | struct qeth_card *card = dev_get_drvdata(dev); | ||
21 | if (!card) | ||
22 | return -EINVAL; | ||
23 | |||
24 | switch (card->state) { | ||
25 | case CARD_STATE_DOWN: | ||
26 | return sprintf(buf, "DOWN\n"); | ||
27 | case CARD_STATE_HARDSETUP: | ||
28 | return sprintf(buf, "HARDSETUP\n"); | ||
29 | case CARD_STATE_SOFTSETUP: | ||
30 | return sprintf(buf, "SOFTSETUP\n"); | ||
31 | case CARD_STATE_UP: | ||
32 | if (card->lan_online) | ||
33 | return sprintf(buf, "UP (LAN ONLINE)\n"); | ||
34 | else | ||
35 | return sprintf(buf, "UP (LAN OFFLINE)\n"); | ||
36 | case CARD_STATE_RECOVER: | ||
37 | return sprintf(buf, "RECOVER\n"); | ||
38 | default: | ||
39 | return sprintf(buf, "UNKNOWN\n"); | ||
40 | } | ||
41 | } | ||
42 | |||
43 | static DEVICE_ATTR(state, 0444, qeth_dev_state_show, NULL); | ||
44 | |||
45 | static ssize_t qeth_dev_chpid_show(struct device *dev, | ||
46 | struct device_attribute *attr, char *buf) | ||
47 | { | ||
48 | struct qeth_card *card = dev_get_drvdata(dev); | ||
49 | if (!card) | ||
50 | return -EINVAL; | ||
51 | |||
52 | return sprintf(buf, "%02X\n", card->info.chpid); | ||
53 | } | ||
54 | |||
55 | static DEVICE_ATTR(chpid, 0444, qeth_dev_chpid_show, NULL); | ||
56 | |||
57 | static ssize_t qeth_dev_if_name_show(struct device *dev, | ||
58 | struct device_attribute *attr, char *buf) | ||
59 | { | ||
60 | struct qeth_card *card = dev_get_drvdata(dev); | ||
61 | if (!card) | ||
62 | return -EINVAL; | ||
63 | return sprintf(buf, "%s\n", QETH_CARD_IFNAME(card)); | ||
64 | } | ||
65 | |||
66 | static DEVICE_ATTR(if_name, 0444, qeth_dev_if_name_show, NULL); | ||
67 | |||
68 | static ssize_t qeth_dev_card_type_show(struct device *dev, | ||
69 | struct device_attribute *attr, char *buf) | ||
70 | { | ||
71 | struct qeth_card *card = dev_get_drvdata(dev); | ||
72 | if (!card) | ||
73 | return -EINVAL; | ||
74 | |||
75 | return sprintf(buf, "%s\n", qeth_get_cardname_short(card)); | ||
76 | } | ||
77 | |||
78 | static DEVICE_ATTR(card_type, 0444, qeth_dev_card_type_show, NULL); | ||
79 | |||
80 | static inline const char *qeth_get_bufsize_str(struct qeth_card *card) | ||
81 | { | ||
82 | if (card->qdio.in_buf_size == 16384) | ||
83 | return "16k"; | ||
84 | else if (card->qdio.in_buf_size == 24576) | ||
85 | return "24k"; | ||
86 | else if (card->qdio.in_buf_size == 32768) | ||
87 | return "32k"; | ||
88 | else if (card->qdio.in_buf_size == 40960) | ||
89 | return "40k"; | ||
90 | else | ||
91 | return "64k"; | ||
92 | } | ||
93 | |||
94 | static ssize_t qeth_dev_inbuf_size_show(struct device *dev, | ||
95 | struct device_attribute *attr, char *buf) | ||
96 | { | ||
97 | struct qeth_card *card = dev_get_drvdata(dev); | ||
98 | if (!card) | ||
99 | return -EINVAL; | ||
100 | |||
101 | return sprintf(buf, "%s\n", qeth_get_bufsize_str(card)); | ||
102 | } | ||
103 | |||
104 | static DEVICE_ATTR(inbuf_size, 0444, qeth_dev_inbuf_size_show, NULL); | ||
105 | |||
106 | static ssize_t qeth_dev_portno_show(struct device *dev, | ||
107 | struct device_attribute *attr, char *buf) | ||
108 | { | ||
109 | struct qeth_card *card = dev_get_drvdata(dev); | ||
110 | if (!card) | ||
111 | return -EINVAL; | ||
112 | |||
113 | return sprintf(buf, "%i\n", card->info.portno); | ||
114 | } | ||
115 | |||
116 | static ssize_t qeth_dev_portno_store(struct device *dev, | ||
117 | struct device_attribute *attr, const char *buf, size_t count) | ||
118 | { | ||
119 | struct qeth_card *card = dev_get_drvdata(dev); | ||
120 | char *tmp; | ||
121 | unsigned int portno; | ||
122 | |||
123 | if (!card) | ||
124 | return -EINVAL; | ||
125 | |||
126 | if ((card->state != CARD_STATE_DOWN) && | ||
127 | (card->state != CARD_STATE_RECOVER)) | ||
128 | return -EPERM; | ||
129 | |||
130 | portno = simple_strtoul(buf, &tmp, 16); | ||
131 | if (portno > QETH_MAX_PORTNO) { | ||
132 | PRINT_WARN("portno 0x%X is out of range\n", portno); | ||
133 | return -EINVAL; | ||
134 | } | ||
135 | |||
136 | card->info.portno = portno; | ||
137 | return count; | ||
138 | } | ||
139 | |||
140 | static DEVICE_ATTR(portno, 0644, qeth_dev_portno_show, qeth_dev_portno_store); | ||
141 | |||
142 | static ssize_t qeth_dev_portname_show(struct device *dev, | ||
143 | struct device_attribute *attr, char *buf) | ||
144 | { | ||
145 | struct qeth_card *card = dev_get_drvdata(dev); | ||
146 | char portname[9] = {0, }; | ||
147 | |||
148 | if (!card) | ||
149 | return -EINVAL; | ||
150 | |||
151 | if (card->info.portname_required) { | ||
152 | memcpy(portname, card->info.portname + 1, 8); | ||
153 | EBCASC(portname, 8); | ||
154 | return sprintf(buf, "%s\n", portname); | ||
155 | } else | ||
156 | return sprintf(buf, "no portname required\n"); | ||
157 | } | ||
158 | |||
159 | static ssize_t qeth_dev_portname_store(struct device *dev, | ||
160 | struct device_attribute *attr, const char *buf, size_t count) | ||
161 | { | ||
162 | struct qeth_card *card = dev_get_drvdata(dev); | ||
163 | char *tmp; | ||
164 | int i; | ||
165 | |||
166 | if (!card) | ||
167 | return -EINVAL; | ||
168 | |||
169 | if ((card->state != CARD_STATE_DOWN) && | ||
170 | (card->state != CARD_STATE_RECOVER)) | ||
171 | return -EPERM; | ||
172 | |||
173 | tmp = strsep((char **) &buf, "\n"); | ||
174 | if ((strlen(tmp) > 8) || (strlen(tmp) == 0)) | ||
175 | return -EINVAL; | ||
176 | |||
177 | card->info.portname[0] = strlen(tmp); | ||
178 | /* for beauty reasons */ | ||
179 | for (i = 1; i < 9; i++) | ||
180 | card->info.portname[i] = ' '; | ||
181 | strcpy(card->info.portname + 1, tmp); | ||
182 | ASCEBC(card->info.portname + 1, 8); | ||
183 | |||
184 | return count; | ||
185 | } | ||
186 | |||
187 | static DEVICE_ATTR(portname, 0644, qeth_dev_portname_show, | ||
188 | qeth_dev_portname_store); | ||
189 | |||
190 | static ssize_t qeth_dev_prioqing_show(struct device *dev, | ||
191 | struct device_attribute *attr, char *buf) | ||
192 | { | ||
193 | struct qeth_card *card = dev_get_drvdata(dev); | ||
194 | |||
195 | if (!card) | ||
196 | return -EINVAL; | ||
197 | |||
198 | switch (card->qdio.do_prio_queueing) { | ||
199 | case QETH_PRIO_Q_ING_PREC: | ||
200 | return sprintf(buf, "%s\n", "by precedence"); | ||
201 | case QETH_PRIO_Q_ING_TOS: | ||
202 | return sprintf(buf, "%s\n", "by type of service"); | ||
203 | default: | ||
204 | return sprintf(buf, "always queue %i\n", | ||
205 | card->qdio.default_out_queue); | ||
206 | } | ||
207 | } | ||
208 | |||
209 | static ssize_t qeth_dev_prioqing_store(struct device *dev, | ||
210 | struct device_attribute *attr, const char *buf, size_t count) | ||
211 | { | ||
212 | struct qeth_card *card = dev_get_drvdata(dev); | ||
213 | char *tmp; | ||
214 | |||
215 | if (!card) | ||
216 | return -EINVAL; | ||
217 | |||
218 | if ((card->state != CARD_STATE_DOWN) && | ||
219 | (card->state != CARD_STATE_RECOVER)) | ||
220 | return -EPERM; | ||
221 | |||
222 | /* check if 1920 devices are supported , | ||
223 | * if though we have to permit priority queueing | ||
224 | */ | ||
225 | if (card->qdio.no_out_queues == 1) { | ||
226 | PRINT_WARN("Priority queueing disabled due " | ||
227 | "to hardware limitations!\n"); | ||
228 | card->qdio.do_prio_queueing = QETH_PRIOQ_DEFAULT; | ||
229 | return -EPERM; | ||
230 | } | ||
231 | |||
232 | tmp = strsep((char **) &buf, "\n"); | ||
233 | if (!strcmp(tmp, "prio_queueing_prec")) | ||
234 | card->qdio.do_prio_queueing = QETH_PRIO_Q_ING_PREC; | ||
235 | else if (!strcmp(tmp, "prio_queueing_tos")) | ||
236 | card->qdio.do_prio_queueing = QETH_PRIO_Q_ING_TOS; | ||
237 | else if (!strcmp(tmp, "no_prio_queueing:0")) { | ||
238 | card->qdio.do_prio_queueing = QETH_NO_PRIO_QUEUEING; | ||
239 | card->qdio.default_out_queue = 0; | ||
240 | } else if (!strcmp(tmp, "no_prio_queueing:1")) { | ||
241 | card->qdio.do_prio_queueing = QETH_NO_PRIO_QUEUEING; | ||
242 | card->qdio.default_out_queue = 1; | ||
243 | } else if (!strcmp(tmp, "no_prio_queueing:2")) { | ||
244 | card->qdio.do_prio_queueing = QETH_NO_PRIO_QUEUEING; | ||
245 | card->qdio.default_out_queue = 2; | ||
246 | } else if (!strcmp(tmp, "no_prio_queueing:3")) { | ||
247 | card->qdio.do_prio_queueing = QETH_NO_PRIO_QUEUEING; | ||
248 | card->qdio.default_out_queue = 3; | ||
249 | } else if (!strcmp(tmp, "no_prio_queueing")) { | ||
250 | card->qdio.do_prio_queueing = QETH_NO_PRIO_QUEUEING; | ||
251 | card->qdio.default_out_queue = QETH_DEFAULT_QUEUE; | ||
252 | } else { | ||
253 | PRINT_WARN("Unknown queueing type '%s'\n", tmp); | ||
254 | return -EINVAL; | ||
255 | } | ||
256 | return count; | ||
257 | } | ||
258 | |||
259 | static DEVICE_ATTR(priority_queueing, 0644, qeth_dev_prioqing_show, | ||
260 | qeth_dev_prioqing_store); | ||
261 | |||
262 | static ssize_t qeth_dev_bufcnt_show(struct device *dev, | ||
263 | struct device_attribute *attr, char *buf) | ||
264 | { | ||
265 | struct qeth_card *card = dev_get_drvdata(dev); | ||
266 | |||
267 | if (!card) | ||
268 | return -EINVAL; | ||
269 | |||
270 | return sprintf(buf, "%i\n", card->qdio.in_buf_pool.buf_count); | ||
271 | } | ||
272 | |||
273 | static ssize_t qeth_dev_bufcnt_store(struct device *dev, | ||
274 | struct device_attribute *attr, const char *buf, size_t count) | ||
275 | { | ||
276 | struct qeth_card *card = dev_get_drvdata(dev); | ||
277 | char *tmp; | ||
278 | int cnt, old_cnt; | ||
279 | int rc; | ||
280 | |||
281 | if (!card) | ||
282 | return -EINVAL; | ||
283 | |||
284 | if ((card->state != CARD_STATE_DOWN) && | ||
285 | (card->state != CARD_STATE_RECOVER)) | ||
286 | return -EPERM; | ||
287 | |||
288 | old_cnt = card->qdio.in_buf_pool.buf_count; | ||
289 | cnt = simple_strtoul(buf, &tmp, 10); | ||
290 | cnt = (cnt < QETH_IN_BUF_COUNT_MIN) ? QETH_IN_BUF_COUNT_MIN : | ||
291 | ((cnt > QETH_IN_BUF_COUNT_MAX) ? QETH_IN_BUF_COUNT_MAX : cnt); | ||
292 | if (old_cnt != cnt) { | ||
293 | rc = qeth_realloc_buffer_pool(card, cnt); | ||
294 | if (rc) | ||
295 | PRINT_WARN("Error (%d) while setting " | ||
296 | "buffer count.\n", rc); | ||
297 | } | ||
298 | return count; | ||
299 | } | ||
300 | |||
301 | static DEVICE_ATTR(buffer_count, 0644, qeth_dev_bufcnt_show, | ||
302 | qeth_dev_bufcnt_store); | ||
303 | |||
304 | static ssize_t qeth_dev_recover_store(struct device *dev, | ||
305 | struct device_attribute *attr, const char *buf, size_t count) | ||
306 | { | ||
307 | struct qeth_card *card = dev_get_drvdata(dev); | ||
308 | char *tmp; | ||
309 | int i; | ||
310 | |||
311 | if (!card) | ||
312 | return -EINVAL; | ||
313 | |||
314 | if (card->state != CARD_STATE_UP) | ||
315 | return -EPERM; | ||
316 | |||
317 | i = simple_strtoul(buf, &tmp, 16); | ||
318 | if (i == 1) | ||
319 | qeth_schedule_recovery(card); | ||
320 | |||
321 | return count; | ||
322 | } | ||
323 | |||
324 | static DEVICE_ATTR(recover, 0200, NULL, qeth_dev_recover_store); | ||
325 | |||
326 | static ssize_t qeth_dev_performance_stats_show(struct device *dev, | ||
327 | struct device_attribute *attr, char *buf) | ||
328 | { | ||
329 | struct qeth_card *card = dev_get_drvdata(dev); | ||
330 | |||
331 | if (!card) | ||
332 | return -EINVAL; | ||
333 | |||
334 | return sprintf(buf, "%i\n", card->options.performance_stats ? 1:0); | ||
335 | } | ||
336 | |||
337 | static ssize_t qeth_dev_performance_stats_store(struct device *dev, | ||
338 | struct device_attribute *attr, const char *buf, size_t count) | ||
339 | { | ||
340 | struct qeth_card *card = dev_get_drvdata(dev); | ||
341 | char *tmp; | ||
342 | int i; | ||
343 | |||
344 | if (!card) | ||
345 | return -EINVAL; | ||
346 | |||
347 | i = simple_strtoul(buf, &tmp, 16); | ||
348 | if ((i == 0) || (i == 1)) { | ||
349 | if (i == card->options.performance_stats) | ||
350 | return count; | ||
351 | card->options.performance_stats = i; | ||
352 | if (i == 0) | ||
353 | memset(&card->perf_stats, 0, | ||
354 | sizeof(struct qeth_perf_stats)); | ||
355 | card->perf_stats.initial_rx_packets = card->stats.rx_packets; | ||
356 | card->perf_stats.initial_tx_packets = card->stats.tx_packets; | ||
357 | } else { | ||
358 | PRINT_WARN("performance_stats: write 0 or 1 to this file!\n"); | ||
359 | return -EINVAL; | ||
360 | } | ||
361 | return count; | ||
362 | } | ||
363 | |||
364 | static DEVICE_ATTR(performance_stats, 0644, qeth_dev_performance_stats_show, | ||
365 | qeth_dev_performance_stats_store); | ||
366 | |||
367 | static ssize_t qeth_dev_layer2_show(struct device *dev, | ||
368 | struct device_attribute *attr, char *buf) | ||
369 | { | ||
370 | struct qeth_card *card = dev_get_drvdata(dev); | ||
371 | |||
372 | if (!card) | ||
373 | return -EINVAL; | ||
374 | |||
375 | return sprintf(buf, "%i\n", card->options.layer2 ? 1:0); | ||
376 | } | ||
377 | |||
378 | static ssize_t qeth_dev_layer2_store(struct device *dev, | ||
379 | struct device_attribute *attr, const char *buf, size_t count) | ||
380 | { | ||
381 | struct qeth_card *card = dev_get_drvdata(dev); | ||
382 | char *tmp; | ||
383 | int i, rc; | ||
384 | enum qeth_discipline_id newdis; | ||
385 | |||
386 | if (!card) | ||
387 | return -EINVAL; | ||
388 | |||
389 | if (((card->state != CARD_STATE_DOWN) && | ||
390 | (card->state != CARD_STATE_RECOVER))) | ||
391 | return -EPERM; | ||
392 | |||
393 | i = simple_strtoul(buf, &tmp, 16); | ||
394 | switch (i) { | ||
395 | case 0: | ||
396 | newdis = QETH_DISCIPLINE_LAYER3; | ||
397 | break; | ||
398 | case 1: | ||
399 | newdis = QETH_DISCIPLINE_LAYER2; | ||
400 | break; | ||
401 | default: | ||
402 | PRINT_WARN("layer2: write 0 or 1 to this file!\n"); | ||
403 | return -EINVAL; | ||
404 | } | ||
405 | |||
406 | if (card->options.layer2 == newdis) { | ||
407 | return count; | ||
408 | } else { | ||
409 | if (card->discipline.ccwgdriver) { | ||
410 | card->discipline.ccwgdriver->remove(card->gdev); | ||
411 | qeth_core_free_discipline(card); | ||
412 | } | ||
413 | } | ||
414 | |||
415 | rc = qeth_core_load_discipline(card, newdis); | ||
416 | if (rc) | ||
417 | return rc; | ||
418 | |||
419 | rc = card->discipline.ccwgdriver->probe(card->gdev); | ||
420 | if (rc) | ||
421 | return rc; | ||
422 | return count; | ||
423 | } | ||
424 | |||
425 | static DEVICE_ATTR(layer2, 0644, qeth_dev_layer2_show, | ||
426 | qeth_dev_layer2_store); | ||
427 | |||
428 | static ssize_t qeth_dev_large_send_show(struct device *dev, | ||
429 | struct device_attribute *attr, char *buf) | ||
430 | { | ||
431 | struct qeth_card *card = dev_get_drvdata(dev); | ||
432 | |||
433 | if (!card) | ||
434 | return -EINVAL; | ||
435 | |||
436 | switch (card->options.large_send) { | ||
437 | case QETH_LARGE_SEND_NO: | ||
438 | return sprintf(buf, "%s\n", "no"); | ||
439 | case QETH_LARGE_SEND_EDDP: | ||
440 | return sprintf(buf, "%s\n", "EDDP"); | ||
441 | case QETH_LARGE_SEND_TSO: | ||
442 | return sprintf(buf, "%s\n", "TSO"); | ||
443 | default: | ||
444 | return sprintf(buf, "%s\n", "N/A"); | ||
445 | } | ||
446 | } | ||
447 | |||
448 | static ssize_t qeth_dev_large_send_store(struct device *dev, | ||
449 | struct device_attribute *attr, const char *buf, size_t count) | ||
450 | { | ||
451 | struct qeth_card *card = dev_get_drvdata(dev); | ||
452 | enum qeth_large_send_types type; | ||
453 | int rc = 0; | ||
454 | char *tmp; | ||
455 | |||
456 | if (!card) | ||
457 | return -EINVAL; | ||
458 | tmp = strsep((char **) &buf, "\n"); | ||
459 | if (!strcmp(tmp, "no")) { | ||
460 | type = QETH_LARGE_SEND_NO; | ||
461 | } else if (!strcmp(tmp, "EDDP")) { | ||
462 | type = QETH_LARGE_SEND_EDDP; | ||
463 | } else if (!strcmp(tmp, "TSO")) { | ||
464 | type = QETH_LARGE_SEND_TSO; | ||
465 | } else { | ||
466 | PRINT_WARN("large_send: invalid mode %s!\n", tmp); | ||
467 | return -EINVAL; | ||
468 | } | ||
469 | if (card->options.large_send == type) | ||
470 | return count; | ||
471 | rc = qeth_set_large_send(card, type); | ||
472 | if (rc) | ||
473 | return rc; | ||
474 | return count; | ||
475 | } | ||
476 | |||
477 | static DEVICE_ATTR(large_send, 0644, qeth_dev_large_send_show, | ||
478 | qeth_dev_large_send_store); | ||
479 | |||
480 | static ssize_t qeth_dev_blkt_show(char *buf, struct qeth_card *card, int value) | ||
481 | { | ||
482 | |||
483 | if (!card) | ||
484 | return -EINVAL; | ||
485 | |||
486 | return sprintf(buf, "%i\n", value); | ||
487 | } | ||
488 | |||
489 | static ssize_t qeth_dev_blkt_store(struct qeth_card *card, | ||
490 | const char *buf, size_t count, int *value, int max_value) | ||
491 | { | ||
492 | char *tmp; | ||
493 | int i; | ||
494 | |||
495 | if (!card) | ||
496 | return -EINVAL; | ||
497 | |||
498 | if ((card->state != CARD_STATE_DOWN) && | ||
499 | (card->state != CARD_STATE_RECOVER)) | ||
500 | return -EPERM; | ||
501 | |||
502 | i = simple_strtoul(buf, &tmp, 10); | ||
503 | if (i <= max_value) { | ||
504 | *value = i; | ||
505 | } else { | ||
506 | PRINT_WARN("blkt total time: write values between" | ||
507 | " 0 and %d to this file!\n", max_value); | ||
508 | return -EINVAL; | ||
509 | } | ||
510 | return count; | ||
511 | } | ||
512 | |||
513 | static ssize_t qeth_dev_blkt_total_show(struct device *dev, | ||
514 | struct device_attribute *attr, char *buf) | ||
515 | { | ||
516 | struct qeth_card *card = dev_get_drvdata(dev); | ||
517 | |||
518 | return qeth_dev_blkt_show(buf, card, card->info.blkt.time_total); | ||
519 | } | ||
520 | |||
521 | static ssize_t qeth_dev_blkt_total_store(struct device *dev, | ||
522 | struct device_attribute *attr, const char *buf, size_t count) | ||
523 | { | ||
524 | struct qeth_card *card = dev_get_drvdata(dev); | ||
525 | |||
526 | return qeth_dev_blkt_store(card, buf, count, | ||
527 | &card->info.blkt.time_total, 1000); | ||
528 | } | ||
529 | |||
530 | |||
531 | |||
532 | static DEVICE_ATTR(total, 0644, qeth_dev_blkt_total_show, | ||
533 | qeth_dev_blkt_total_store); | ||
534 | |||
535 | static ssize_t qeth_dev_blkt_inter_show(struct device *dev, | ||
536 | struct device_attribute *attr, char *buf) | ||
537 | { | ||
538 | struct qeth_card *card = dev_get_drvdata(dev); | ||
539 | |||
540 | return qeth_dev_blkt_show(buf, card, card->info.blkt.inter_packet); | ||
541 | } | ||
542 | |||
543 | static ssize_t qeth_dev_blkt_inter_store(struct device *dev, | ||
544 | struct device_attribute *attr, const char *buf, size_t count) | ||
545 | { | ||
546 | struct qeth_card *card = dev_get_drvdata(dev); | ||
547 | |||
548 | return qeth_dev_blkt_store(card, buf, count, | ||
549 | &card->info.blkt.inter_packet, 100); | ||
550 | } | ||
551 | |||
552 | static DEVICE_ATTR(inter, 0644, qeth_dev_blkt_inter_show, | ||
553 | qeth_dev_blkt_inter_store); | ||
554 | |||
555 | static ssize_t qeth_dev_blkt_inter_jumbo_show(struct device *dev, | ||
556 | struct device_attribute *attr, char *buf) | ||
557 | { | ||
558 | struct qeth_card *card = dev_get_drvdata(dev); | ||
559 | |||
560 | return qeth_dev_blkt_show(buf, card, | ||
561 | card->info.blkt.inter_packet_jumbo); | ||
562 | } | ||
563 | |||
564 | static ssize_t qeth_dev_blkt_inter_jumbo_store(struct device *dev, | ||
565 | struct device_attribute *attr, const char *buf, size_t count) | ||
566 | { | ||
567 | struct qeth_card *card = dev_get_drvdata(dev); | ||
568 | |||
569 | return qeth_dev_blkt_store(card, buf, count, | ||
570 | &card->info.blkt.inter_packet_jumbo, 100); | ||
571 | } | ||
572 | |||
573 | static DEVICE_ATTR(inter_jumbo, 0644, qeth_dev_blkt_inter_jumbo_show, | ||
574 | qeth_dev_blkt_inter_jumbo_store); | ||
575 | |||
576 | static struct attribute *qeth_blkt_device_attrs[] = { | ||
577 | &dev_attr_total.attr, | ||
578 | &dev_attr_inter.attr, | ||
579 | &dev_attr_inter_jumbo.attr, | ||
580 | NULL, | ||
581 | }; | ||
582 | |||
583 | static struct attribute_group qeth_device_blkt_group = { | ||
584 | .name = "blkt", | ||
585 | .attrs = qeth_blkt_device_attrs, | ||
586 | }; | ||
587 | |||
588 | static struct attribute *qeth_device_attrs[] = { | ||
589 | &dev_attr_state.attr, | ||
590 | &dev_attr_chpid.attr, | ||
591 | &dev_attr_if_name.attr, | ||
592 | &dev_attr_card_type.attr, | ||
593 | &dev_attr_inbuf_size.attr, | ||
594 | &dev_attr_portno.attr, | ||
595 | &dev_attr_portname.attr, | ||
596 | &dev_attr_priority_queueing.attr, | ||
597 | &dev_attr_buffer_count.attr, | ||
598 | &dev_attr_recover.attr, | ||
599 | &dev_attr_performance_stats.attr, | ||
600 | &dev_attr_layer2.attr, | ||
601 | &dev_attr_large_send.attr, | ||
602 | NULL, | ||
603 | }; | ||
604 | |||
605 | static struct attribute_group qeth_device_attr_group = { | ||
606 | .attrs = qeth_device_attrs, | ||
607 | }; | ||
608 | |||
609 | static struct attribute *qeth_osn_device_attrs[] = { | ||
610 | &dev_attr_state.attr, | ||
611 | &dev_attr_chpid.attr, | ||
612 | &dev_attr_if_name.attr, | ||
613 | &dev_attr_card_type.attr, | ||
614 | &dev_attr_buffer_count.attr, | ||
615 | &dev_attr_recover.attr, | ||
616 | NULL, | ||
617 | }; | ||
618 | |||
619 | static struct attribute_group qeth_osn_device_attr_group = { | ||
620 | .attrs = qeth_osn_device_attrs, | ||
621 | }; | ||
622 | |||
623 | int qeth_core_create_device_attributes(struct device *dev) | ||
624 | { | ||
625 | int ret; | ||
626 | ret = sysfs_create_group(&dev->kobj, &qeth_device_attr_group); | ||
627 | if (ret) | ||
628 | return ret; | ||
629 | ret = sysfs_create_group(&dev->kobj, &qeth_device_blkt_group); | ||
630 | if (ret) | ||
631 | sysfs_remove_group(&dev->kobj, &qeth_device_attr_group); | ||
632 | |||
633 | return 0; | ||
634 | } | ||
635 | |||
636 | void qeth_core_remove_device_attributes(struct device *dev) | ||
637 | { | ||
638 | sysfs_remove_group(&dev->kobj, &qeth_device_attr_group); | ||
639 | sysfs_remove_group(&dev->kobj, &qeth_device_blkt_group); | ||
640 | } | ||
641 | |||
642 | int qeth_core_create_osn_attributes(struct device *dev) | ||
643 | { | ||
644 | return sysfs_create_group(&dev->kobj, &qeth_osn_device_attr_group); | ||
645 | } | ||
646 | |||
647 | void qeth_core_remove_osn_attributes(struct device *dev) | ||
648 | { | ||
649 | sysfs_remove_group(&dev->kobj, &qeth_osn_device_attr_group); | ||
650 | return; | ||
651 | } | ||
diff --git a/drivers/s390/net/qeth_fs.h b/drivers/s390/net/qeth_fs.h deleted file mode 100644 index 61faf05517d6..000000000000 --- a/drivers/s390/net/qeth_fs.h +++ /dev/null | |||
@@ -1,168 +0,0 @@ | |||
1 | /* | ||
2 | * linux/drivers/s390/net/qeth_fs.h | ||
3 | * | ||
4 | * Linux on zSeries OSA Express and HiperSockets support. | ||
5 | * | ||
6 | * This header file contains definitions related to sysfs and procfs. | ||
7 | * | ||
8 | * Copyright 2000,2003 IBM Corporation | ||
9 | * Author(s): Thomas Spatzier <tspat@de.ibm.com> | ||
10 | * | ||
11 | */ | ||
12 | #ifndef __QETH_FS_H__ | ||
13 | #define __QETH_FS_H__ | ||
14 | |||
15 | #ifdef CONFIG_PROC_FS | ||
16 | extern int | ||
17 | qeth_create_procfs_entries(void); | ||
18 | |||
19 | extern void | ||
20 | qeth_remove_procfs_entries(void); | ||
21 | #else | ||
22 | static inline int | ||
23 | qeth_create_procfs_entries(void) | ||
24 | { | ||
25 | return 0; | ||
26 | } | ||
27 | |||
28 | static inline void | ||
29 | qeth_remove_procfs_entries(void) | ||
30 | { | ||
31 | } | ||
32 | #endif /* CONFIG_PROC_FS */ | ||
33 | |||
34 | extern int | ||
35 | qeth_create_device_attributes(struct device *dev); | ||
36 | |||
37 | extern void | ||
38 | qeth_remove_device_attributes(struct device *dev); | ||
39 | |||
40 | extern int | ||
41 | qeth_create_device_attributes_osn(struct device *dev); | ||
42 | |||
43 | extern void | ||
44 | qeth_remove_device_attributes_osn(struct device *dev); | ||
45 | |||
46 | extern int | ||
47 | qeth_create_driver_attributes(void); | ||
48 | |||
49 | extern void | ||
50 | qeth_remove_driver_attributes(void); | ||
51 | |||
52 | /* | ||
53 | * utility functions used in qeth_proc.c and qeth_sys.c | ||
54 | */ | ||
55 | |||
56 | static inline const char * | ||
57 | qeth_get_checksum_str(struct qeth_card *card) | ||
58 | { | ||
59 | if (card->options.checksum_type == SW_CHECKSUMMING) | ||
60 | return "sw"; | ||
61 | else if (card->options.checksum_type == HW_CHECKSUMMING) | ||
62 | return "hw"; | ||
63 | else | ||
64 | return "no"; | ||
65 | } | ||
66 | |||
67 | static inline const char * | ||
68 | qeth_get_prioq_str(struct qeth_card *card, char *buf) | ||
69 | { | ||
70 | if (card->qdio.do_prio_queueing == QETH_NO_PRIO_QUEUEING) | ||
71 | sprintf(buf, "always_q_%i", card->qdio.default_out_queue); | ||
72 | else | ||
73 | strcpy(buf, (card->qdio.do_prio_queueing == | ||
74 | QETH_PRIO_Q_ING_PREC)? | ||
75 | "by_prec." : "by_ToS"); | ||
76 | return buf; | ||
77 | } | ||
78 | |||
79 | static inline const char * | ||
80 | qeth_get_bufsize_str(struct qeth_card *card) | ||
81 | { | ||
82 | if (card->qdio.in_buf_size == 16384) | ||
83 | return "16k"; | ||
84 | else if (card->qdio.in_buf_size == 24576) | ||
85 | return "24k"; | ||
86 | else if (card->qdio.in_buf_size == 32768) | ||
87 | return "32k"; | ||
88 | else if (card->qdio.in_buf_size == 40960) | ||
89 | return "40k"; | ||
90 | else | ||
91 | return "64k"; | ||
92 | } | ||
93 | |||
94 | static inline const char * | ||
95 | qeth_get_cardname(struct qeth_card *card) | ||
96 | { | ||
97 | if (card->info.guestlan) { | ||
98 | switch (card->info.type) { | ||
99 | case QETH_CARD_TYPE_OSAE: | ||
100 | return " Guest LAN QDIO"; | ||
101 | case QETH_CARD_TYPE_IQD: | ||
102 | return " Guest LAN Hiper"; | ||
103 | default: | ||
104 | return " unknown"; | ||
105 | } | ||
106 | } else { | ||
107 | switch (card->info.type) { | ||
108 | case QETH_CARD_TYPE_OSAE: | ||
109 | return " OSD Express"; | ||
110 | case QETH_CARD_TYPE_IQD: | ||
111 | return " HiperSockets"; | ||
112 | case QETH_CARD_TYPE_OSN: | ||
113 | return " OSN QDIO"; | ||
114 | default: | ||
115 | return " unknown"; | ||
116 | } | ||
117 | } | ||
118 | return " n/a"; | ||
119 | } | ||
120 | |||
121 | /* max length to be returned: 14 */ | ||
122 | static inline const char * | ||
123 | qeth_get_cardname_short(struct qeth_card *card) | ||
124 | { | ||
125 | if (card->info.guestlan){ | ||
126 | switch (card->info.type){ | ||
127 | case QETH_CARD_TYPE_OSAE: | ||
128 | return "GuestLAN QDIO"; | ||
129 | case QETH_CARD_TYPE_IQD: | ||
130 | return "GuestLAN Hiper"; | ||
131 | default: | ||
132 | return "unknown"; | ||
133 | } | ||
134 | } else { | ||
135 | switch (card->info.type) { | ||
136 | case QETH_CARD_TYPE_OSAE: | ||
137 | switch (card->info.link_type) { | ||
138 | case QETH_LINK_TYPE_FAST_ETH: | ||
139 | return "OSD_100"; | ||
140 | case QETH_LINK_TYPE_HSTR: | ||
141 | return "HSTR"; | ||
142 | case QETH_LINK_TYPE_GBIT_ETH: | ||
143 | return "OSD_1000"; | ||
144 | case QETH_LINK_TYPE_10GBIT_ETH: | ||
145 | return "OSD_10GIG"; | ||
146 | case QETH_LINK_TYPE_LANE_ETH100: | ||
147 | return "OSD_FE_LANE"; | ||
148 | case QETH_LINK_TYPE_LANE_TR: | ||
149 | return "OSD_TR_LANE"; | ||
150 | case QETH_LINK_TYPE_LANE_ETH1000: | ||
151 | return "OSD_GbE_LANE"; | ||
152 | case QETH_LINK_TYPE_LANE: | ||
153 | return "OSD_ATM_LANE"; | ||
154 | default: | ||
155 | return "OSD_Express"; | ||
156 | } | ||
157 | case QETH_CARD_TYPE_IQD: | ||
158 | return "HiperSockets"; | ||
159 | case QETH_CARD_TYPE_OSN: | ||
160 | return "OSN"; | ||
161 | default: | ||
162 | return "unknown"; | ||
163 | } | ||
164 | } | ||
165 | return "n/a"; | ||
166 | } | ||
167 | |||
168 | #endif /* __QETH_FS_H__ */ | ||
diff --git a/drivers/s390/net/qeth_l2_main.c b/drivers/s390/net/qeth_l2_main.c new file mode 100644 index 000000000000..3921d1631a78 --- /dev/null +++ b/drivers/s390/net/qeth_l2_main.c | |||
@@ -0,0 +1,1234 @@ | |||
1 | /* | ||
2 | * drivers/s390/net/qeth_l2_main.c | ||
3 | * | ||
4 | * Copyright IBM Corp. 2007 | ||
5 | * Author(s): Utz Bacher <utz.bacher@de.ibm.com>, | ||
6 | * Frank Pavlic <fpavlic@de.ibm.com>, | ||
7 | * Thomas Spatzier <tspat@de.ibm.com>, | ||
8 | * Frank Blaschka <frank.blaschka@de.ibm.com> | ||
9 | */ | ||
10 | |||
11 | #include <linux/module.h> | ||
12 | #include <linux/moduleparam.h> | ||
13 | #include <linux/string.h> | ||
14 | #include <linux/errno.h> | ||
15 | #include <linux/kernel.h> | ||
16 | #include <linux/etherdevice.h> | ||
17 | #include <linux/mii.h> | ||
18 | #include <linux/ip.h> | ||
19 | |||
20 | #include <asm/s390_rdev.h> | ||
21 | |||
22 | #include "qeth_core.h" | ||
23 | #include "qeth_core_offl.h" | ||
24 | |||
25 | #define QETH_DBF_TXT_BUF qeth_l2_dbf_txt_buf | ||
26 | static DEFINE_PER_CPU(char[256], qeth_l2_dbf_txt_buf); | ||
27 | |||
28 | static int qeth_l2_set_offline(struct ccwgroup_device *); | ||
29 | static int qeth_l2_stop(struct net_device *); | ||
30 | static int qeth_l2_send_delmac(struct qeth_card *, __u8 *); | ||
31 | static int qeth_l2_send_setdelmac(struct qeth_card *, __u8 *, | ||
32 | enum qeth_ipa_cmds, | ||
33 | int (*reply_cb) (struct qeth_card *, | ||
34 | struct qeth_reply*, | ||
35 | unsigned long)); | ||
36 | static void qeth_l2_set_multicast_list(struct net_device *); | ||
37 | static int qeth_l2_recover(void *); | ||
38 | |||
39 | static int qeth_l2_do_ioctl(struct net_device *dev, struct ifreq *rq, int cmd) | ||
40 | { | ||
41 | struct qeth_card *card = netdev_priv(dev); | ||
42 | struct mii_ioctl_data *mii_data; | ||
43 | int rc = 0; | ||
44 | |||
45 | if (!card) | ||
46 | return -ENODEV; | ||
47 | |||
48 | if ((card->state != CARD_STATE_UP) && | ||
49 | (card->state != CARD_STATE_SOFTSETUP)) | ||
50 | return -ENODEV; | ||
51 | |||
52 | if (card->info.type == QETH_CARD_TYPE_OSN) | ||
53 | return -EPERM; | ||
54 | |||
55 | switch (cmd) { | ||
56 | case SIOC_QETH_ADP_SET_SNMP_CONTROL: | ||
57 | rc = qeth_snmp_command(card, rq->ifr_ifru.ifru_data); | ||
58 | break; | ||
59 | case SIOC_QETH_GET_CARD_TYPE: | ||
60 | if ((card->info.type == QETH_CARD_TYPE_OSAE) && | ||
61 | !card->info.guestlan) | ||
62 | return 1; | ||
63 | return 0; | ||
64 | break; | ||
65 | case SIOCGMIIPHY: | ||
66 | mii_data = if_mii(rq); | ||
67 | mii_data->phy_id = 0; | ||
68 | break; | ||
69 | case SIOCGMIIREG: | ||
70 | mii_data = if_mii(rq); | ||
71 | if (mii_data->phy_id != 0) | ||
72 | rc = -EINVAL; | ||
73 | else | ||
74 | mii_data->val_out = qeth_mdio_read(dev, | ||
75 | mii_data->phy_id, mii_data->reg_num); | ||
76 | break; | ||
77 | default: | ||
78 | rc = -EOPNOTSUPP; | ||
79 | } | ||
80 | if (rc) | ||
81 | QETH_DBF_TEXT_(TRACE, 2, "ioce%d", rc); | ||
82 | return rc; | ||
83 | } | ||
84 | |||
85 | static int qeth_l2_verify_dev(struct net_device *dev) | ||
86 | { | ||
87 | struct qeth_card *card; | ||
88 | unsigned long flags; | ||
89 | int rc = 0; | ||
90 | |||
91 | read_lock_irqsave(&qeth_core_card_list.rwlock, flags); | ||
92 | list_for_each_entry(card, &qeth_core_card_list.list, list) { | ||
93 | if (card->dev == dev) { | ||
94 | rc = QETH_REAL_CARD; | ||
95 | break; | ||
96 | } | ||
97 | } | ||
98 | read_unlock_irqrestore(&qeth_core_card_list.rwlock, flags); | ||
99 | |||
100 | return rc; | ||
101 | } | ||
102 | |||
103 | static struct net_device *qeth_l2_netdev_by_devno(unsigned char *read_dev_no) | ||
104 | { | ||
105 | struct qeth_card *card; | ||
106 | struct net_device *ndev; | ||
107 | unsigned char *readno; | ||
108 | __u16 temp_dev_no, card_dev_no; | ||
109 | char *endp; | ||
110 | unsigned long flags; | ||
111 | |||
112 | ndev = NULL; | ||
113 | memcpy(&temp_dev_no, read_dev_no, 2); | ||
114 | read_lock_irqsave(&qeth_core_card_list.rwlock, flags); | ||
115 | list_for_each_entry(card, &qeth_core_card_list.list, list) { | ||
116 | readno = CARD_RDEV_ID(card); | ||
117 | readno += (strlen(readno) - 4); | ||
118 | card_dev_no = simple_strtoul(readno, &endp, 16); | ||
119 | if (card_dev_no == temp_dev_no) { | ||
120 | ndev = card->dev; | ||
121 | break; | ||
122 | } | ||
123 | } | ||
124 | read_unlock_irqrestore(&qeth_core_card_list.rwlock, flags); | ||
125 | return ndev; | ||
126 | } | ||
127 | |||
128 | static int qeth_l2_send_setgroupmac_cb(struct qeth_card *card, | ||
129 | struct qeth_reply *reply, | ||
130 | unsigned long data) | ||
131 | { | ||
132 | struct qeth_ipa_cmd *cmd; | ||
133 | __u8 *mac; | ||
134 | |||
135 | QETH_DBF_TEXT(TRACE, 2, "L2Sgmacb"); | ||
136 | cmd = (struct qeth_ipa_cmd *) data; | ||
137 | mac = &cmd->data.setdelmac.mac[0]; | ||
138 | /* MAC already registered, needed in couple/uncouple case */ | ||
139 | if (cmd->hdr.return_code == 0x2005) { | ||
140 | PRINT_WARN("Group MAC %02x:%02x:%02x:%02x:%02x:%02x " \ | ||
141 | "already existing on %s \n", | ||
142 | mac[0], mac[1], mac[2], mac[3], mac[4], mac[5], | ||
143 | QETH_CARD_IFNAME(card)); | ||
144 | cmd->hdr.return_code = 0; | ||
145 | } | ||
146 | if (cmd->hdr.return_code) | ||
147 | PRINT_ERR("Could not set group MAC " \ | ||
148 | "%02x:%02x:%02x:%02x:%02x:%02x on %s: %x\n", | ||
149 | mac[0], mac[1], mac[2], mac[3], mac[4], mac[5], | ||
150 | QETH_CARD_IFNAME(card), cmd->hdr.return_code); | ||
151 | return 0; | ||
152 | } | ||
153 | |||
154 | static int qeth_l2_send_setgroupmac(struct qeth_card *card, __u8 *mac) | ||
155 | { | ||
156 | QETH_DBF_TEXT(TRACE, 2, "L2Sgmac"); | ||
157 | return qeth_l2_send_setdelmac(card, mac, IPA_CMD_SETGMAC, | ||
158 | qeth_l2_send_setgroupmac_cb); | ||
159 | } | ||
160 | |||
161 | static int qeth_l2_send_delgroupmac_cb(struct qeth_card *card, | ||
162 | struct qeth_reply *reply, | ||
163 | unsigned long data) | ||
164 | { | ||
165 | struct qeth_ipa_cmd *cmd; | ||
166 | __u8 *mac; | ||
167 | |||
168 | QETH_DBF_TEXT(TRACE, 2, "L2Dgmacb"); | ||
169 | cmd = (struct qeth_ipa_cmd *) data; | ||
170 | mac = &cmd->data.setdelmac.mac[0]; | ||
171 | if (cmd->hdr.return_code) | ||
172 | PRINT_ERR("Could not delete group MAC " \ | ||
173 | "%02x:%02x:%02x:%02x:%02x:%02x on %s: %x\n", | ||
174 | mac[0], mac[1], mac[2], mac[3], mac[4], mac[5], | ||
175 | QETH_CARD_IFNAME(card), cmd->hdr.return_code); | ||
176 | return 0; | ||
177 | } | ||
178 | |||
179 | static int qeth_l2_send_delgroupmac(struct qeth_card *card, __u8 *mac) | ||
180 | { | ||
181 | QETH_DBF_TEXT(TRACE, 2, "L2Dgmac"); | ||
182 | return qeth_l2_send_setdelmac(card, mac, IPA_CMD_DELGMAC, | ||
183 | qeth_l2_send_delgroupmac_cb); | ||
184 | } | ||
185 | |||
186 | static void qeth_l2_add_mc(struct qeth_card *card, __u8 *mac) | ||
187 | { | ||
188 | struct qeth_mc_mac *mc; | ||
189 | |||
190 | mc = kmalloc(sizeof(struct qeth_mc_mac), GFP_ATOMIC); | ||
191 | |||
192 | if (!mc) { | ||
193 | PRINT_ERR("no mem vor mc mac address\n"); | ||
194 | return; | ||
195 | } | ||
196 | |||
197 | memcpy(mc->mc_addr, mac, OSA_ADDR_LEN); | ||
198 | mc->mc_addrlen = OSA_ADDR_LEN; | ||
199 | |||
200 | if (!qeth_l2_send_setgroupmac(card, mac)) | ||
201 | list_add_tail(&mc->list, &card->mc_list); | ||
202 | else | ||
203 | kfree(mc); | ||
204 | } | ||
205 | |||
206 | static void qeth_l2_del_all_mc(struct qeth_card *card) | ||
207 | { | ||
208 | struct qeth_mc_mac *mc, *tmp; | ||
209 | |||
210 | spin_lock_bh(&card->mclock); | ||
211 | list_for_each_entry_safe(mc, tmp, &card->mc_list, list) { | ||
212 | qeth_l2_send_delgroupmac(card, mc->mc_addr); | ||
213 | list_del(&mc->list); | ||
214 | kfree(mc); | ||
215 | } | ||
216 | spin_unlock_bh(&card->mclock); | ||
217 | } | ||
218 | |||
219 | static void qeth_l2_get_packet_type(struct qeth_card *card, | ||
220 | struct qeth_hdr *hdr, struct sk_buff *skb) | ||
221 | { | ||
222 | __u16 hdr_mac; | ||
223 | |||
224 | if (!memcmp(skb->data + QETH_HEADER_SIZE, | ||
225 | skb->dev->broadcast, 6)) { | ||
226 | /* broadcast? */ | ||
227 | hdr->hdr.l2.flags[2] |= QETH_LAYER2_FLAG_BROADCAST; | ||
228 | return; | ||
229 | } | ||
230 | hdr_mac = *((__u16 *)skb->data); | ||
231 | /* tr multicast? */ | ||
232 | switch (card->info.link_type) { | ||
233 | case QETH_LINK_TYPE_HSTR: | ||
234 | case QETH_LINK_TYPE_LANE_TR: | ||
235 | if ((hdr_mac == QETH_TR_MAC_NC) || | ||
236 | (hdr_mac == QETH_TR_MAC_C)) | ||
237 | hdr->hdr.l2.flags[2] |= QETH_LAYER2_FLAG_MULTICAST; | ||
238 | else | ||
239 | hdr->hdr.l2.flags[2] |= QETH_LAYER2_FLAG_UNICAST; | ||
240 | break; | ||
241 | /* eth or so multicast? */ | ||
242 | default: | ||
243 | if ((hdr_mac == QETH_ETH_MAC_V4) || | ||
244 | (hdr_mac == QETH_ETH_MAC_V6)) | ||
245 | hdr->hdr.l2.flags[2] |= QETH_LAYER2_FLAG_MULTICAST; | ||
246 | else | ||
247 | hdr->hdr.l2.flags[2] |= QETH_LAYER2_FLAG_UNICAST; | ||
248 | } | ||
249 | } | ||
250 | |||
251 | static void qeth_l2_fill_header(struct qeth_card *card, struct qeth_hdr *hdr, | ||
252 | struct sk_buff *skb, int ipv, int cast_type) | ||
253 | { | ||
254 | struct vlan_ethhdr *veth = (struct vlan_ethhdr *)((skb->data) + | ||
255 | QETH_HEADER_SIZE); | ||
256 | |||
257 | memset(hdr, 0, sizeof(struct qeth_hdr)); | ||
258 | hdr->hdr.l2.id = QETH_HEADER_TYPE_LAYER2; | ||
259 | |||
260 | /* set byte byte 3 to casting flags */ | ||
261 | if (cast_type == RTN_MULTICAST) | ||
262 | hdr->hdr.l2.flags[2] |= QETH_LAYER2_FLAG_MULTICAST; | ||
263 | else if (cast_type == RTN_BROADCAST) | ||
264 | hdr->hdr.l2.flags[2] |= QETH_LAYER2_FLAG_BROADCAST; | ||
265 | else | ||
266 | qeth_l2_get_packet_type(card, hdr, skb); | ||
267 | |||
268 | hdr->hdr.l2.pkt_length = skb->len-QETH_HEADER_SIZE; | ||
269 | /* VSWITCH relies on the VLAN | ||
270 | * information to be present in | ||
271 | * the QDIO header */ | ||
272 | if (veth->h_vlan_proto == __constant_htons(ETH_P_8021Q)) { | ||
273 | hdr->hdr.l2.flags[2] |= QETH_LAYER2_FLAG_VLAN; | ||
274 | hdr->hdr.l2.vlan_id = ntohs(veth->h_vlan_TCI); | ||
275 | } | ||
276 | } | ||
277 | |||
278 | static int qeth_l2_send_setdelvlan_cb(struct qeth_card *card, | ||
279 | struct qeth_reply *reply, unsigned long data) | ||
280 | { | ||
281 | struct qeth_ipa_cmd *cmd; | ||
282 | |||
283 | QETH_DBF_TEXT(TRACE, 2, "L2sdvcb"); | ||
284 | cmd = (struct qeth_ipa_cmd *) data; | ||
285 | if (cmd->hdr.return_code) { | ||
286 | PRINT_ERR("Error in processing VLAN %i on %s: 0x%x. " | ||
287 | "Continuing\n", cmd->data.setdelvlan.vlan_id, | ||
288 | QETH_CARD_IFNAME(card), cmd->hdr.return_code); | ||
289 | QETH_DBF_TEXT_(TRACE, 2, "L2VL%4x", cmd->hdr.command); | ||
290 | QETH_DBF_TEXT_(TRACE, 2, "L2%s", CARD_BUS_ID(card)); | ||
291 | QETH_DBF_TEXT_(TRACE, 2, "err%d", cmd->hdr.return_code); | ||
292 | } | ||
293 | return 0; | ||
294 | } | ||
295 | |||
296 | static int qeth_l2_send_setdelvlan(struct qeth_card *card, __u16 i, | ||
297 | enum qeth_ipa_cmds ipacmd) | ||
298 | { | ||
299 | struct qeth_ipa_cmd *cmd; | ||
300 | struct qeth_cmd_buffer *iob; | ||
301 | |||
302 | QETH_DBF_TEXT_(TRACE, 4, "L2sdv%x", ipacmd); | ||
303 | iob = qeth_get_ipacmd_buffer(card, ipacmd, QETH_PROT_IPV4); | ||
304 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
305 | cmd->data.setdelvlan.vlan_id = i; | ||
306 | return qeth_send_ipa_cmd(card, iob, | ||
307 | qeth_l2_send_setdelvlan_cb, NULL); | ||
308 | } | ||
309 | |||
310 | static void qeth_l2_process_vlans(struct qeth_card *card, int clear) | ||
311 | { | ||
312 | struct qeth_vlan_vid *id; | ||
313 | QETH_DBF_TEXT(TRACE, 3, "L2prcvln"); | ||
314 | spin_lock_bh(&card->vlanlock); | ||
315 | list_for_each_entry(id, &card->vid_list, list) { | ||
316 | if (clear) | ||
317 | qeth_l2_send_setdelvlan(card, id->vid, | ||
318 | IPA_CMD_DELVLAN); | ||
319 | else | ||
320 | qeth_l2_send_setdelvlan(card, id->vid, | ||
321 | IPA_CMD_SETVLAN); | ||
322 | } | ||
323 | spin_unlock_bh(&card->vlanlock); | ||
324 | } | ||
325 | |||
326 | static void qeth_l2_vlan_rx_add_vid(struct net_device *dev, unsigned short vid) | ||
327 | { | ||
328 | struct qeth_card *card = netdev_priv(dev); | ||
329 | struct qeth_vlan_vid *id; | ||
330 | |||
331 | QETH_DBF_TEXT_(TRACE, 4, "aid:%d", vid); | ||
332 | id = kmalloc(sizeof(struct qeth_vlan_vid), GFP_ATOMIC); | ||
333 | if (id) { | ||
334 | id->vid = vid; | ||
335 | qeth_l2_send_setdelvlan(card, vid, IPA_CMD_SETVLAN); | ||
336 | spin_lock_bh(&card->vlanlock); | ||
337 | list_add_tail(&id->list, &card->vid_list); | ||
338 | spin_unlock_bh(&card->vlanlock); | ||
339 | } else { | ||
340 | PRINT_ERR("no memory for vid\n"); | ||
341 | } | ||
342 | } | ||
343 | |||
344 | static void qeth_l2_vlan_rx_kill_vid(struct net_device *dev, unsigned short vid) | ||
345 | { | ||
346 | struct qeth_vlan_vid *id, *tmpid = NULL; | ||
347 | struct qeth_card *card = netdev_priv(dev); | ||
348 | |||
349 | QETH_DBF_TEXT_(TRACE, 4, "kid:%d", vid); | ||
350 | spin_lock_bh(&card->vlanlock); | ||
351 | list_for_each_entry(id, &card->vid_list, list) { | ||
352 | if (id->vid == vid) { | ||
353 | list_del(&id->list); | ||
354 | tmpid = id; | ||
355 | break; | ||
356 | } | ||
357 | } | ||
358 | spin_unlock_bh(&card->vlanlock); | ||
359 | if (tmpid) { | ||
360 | qeth_l2_send_setdelvlan(card, vid, IPA_CMD_DELVLAN); | ||
361 | kfree(tmpid); | ||
362 | } | ||
363 | qeth_l2_set_multicast_list(card->dev); | ||
364 | } | ||
365 | |||
366 | static int qeth_l2_stop_card(struct qeth_card *card, int recovery_mode) | ||
367 | { | ||
368 | int rc = 0; | ||
369 | |||
370 | QETH_DBF_TEXT(SETUP , 2, "stopcard"); | ||
371 | QETH_DBF_HEX(SETUP, 2, &card, sizeof(void *)); | ||
372 | |||
373 | qeth_set_allowed_threads(card, 0, 1); | ||
374 | if (qeth_wait_for_threads(card, ~QETH_RECOVER_THREAD)) | ||
375 | return -ERESTARTSYS; | ||
376 | if (card->read.state == CH_STATE_UP && | ||
377 | card->write.state == CH_STATE_UP && | ||
378 | (card->state == CARD_STATE_UP)) { | ||
379 | if (recovery_mode && | ||
380 | card->info.type != QETH_CARD_TYPE_OSN) { | ||
381 | qeth_l2_stop(card->dev); | ||
382 | } else { | ||
383 | rtnl_lock(); | ||
384 | dev_close(card->dev); | ||
385 | rtnl_unlock(); | ||
386 | } | ||
387 | if (!card->use_hard_stop) { | ||
388 | __u8 *mac = &card->dev->dev_addr[0]; | ||
389 | rc = qeth_l2_send_delmac(card, mac); | ||
390 | QETH_DBF_TEXT_(SETUP, 2, "Lerr%d", rc); | ||
391 | } | ||
392 | card->state = CARD_STATE_SOFTSETUP; | ||
393 | } | ||
394 | if (card->state == CARD_STATE_SOFTSETUP) { | ||
395 | qeth_l2_process_vlans(card, 1); | ||
396 | qeth_l2_del_all_mc(card); | ||
397 | qeth_clear_ipacmd_list(card); | ||
398 | card->state = CARD_STATE_HARDSETUP; | ||
399 | } | ||
400 | if (card->state == CARD_STATE_HARDSETUP) { | ||
401 | qeth_qdio_clear_card(card, 0); | ||
402 | qeth_clear_qdio_buffers(card); | ||
403 | qeth_clear_working_pool_list(card); | ||
404 | card->state = CARD_STATE_DOWN; | ||
405 | } | ||
406 | if (card->state == CARD_STATE_DOWN) { | ||
407 | qeth_clear_cmd_buffers(&card->read); | ||
408 | qeth_clear_cmd_buffers(&card->write); | ||
409 | } | ||
410 | card->use_hard_stop = 0; | ||
411 | return rc; | ||
412 | } | ||
413 | |||
414 | static void qeth_l2_process_inbound_buffer(struct qeth_card *card, | ||
415 | struct qeth_qdio_buffer *buf, int index) | ||
416 | { | ||
417 | struct qdio_buffer_element *element; | ||
418 | struct sk_buff *skb; | ||
419 | struct qeth_hdr *hdr; | ||
420 | int offset; | ||
421 | unsigned int len; | ||
422 | |||
423 | /* get first element of current buffer */ | ||
424 | element = (struct qdio_buffer_element *)&buf->buffer->element[0]; | ||
425 | offset = 0; | ||
426 | if (card->options.performance_stats) | ||
427 | card->perf_stats.bufs_rec++; | ||
428 | while ((skb = qeth_core_get_next_skb(card, buf->buffer, &element, | ||
429 | &offset, &hdr))) { | ||
430 | skb->dev = card->dev; | ||
431 | /* is device UP ? */ | ||
432 | if (!(card->dev->flags & IFF_UP)) { | ||
433 | dev_kfree_skb_any(skb); | ||
434 | continue; | ||
435 | } | ||
436 | |||
437 | switch (hdr->hdr.l2.id) { | ||
438 | case QETH_HEADER_TYPE_LAYER2: | ||
439 | skb->pkt_type = PACKET_HOST; | ||
440 | skb->protocol = eth_type_trans(skb, skb->dev); | ||
441 | if (card->options.checksum_type == NO_CHECKSUMMING) | ||
442 | skb->ip_summed = CHECKSUM_UNNECESSARY; | ||
443 | else | ||
444 | skb->ip_summed = CHECKSUM_NONE; | ||
445 | if (skb->protocol == htons(ETH_P_802_2)) | ||
446 | *((__u32 *)skb->cb) = ++card->seqno.pkt_seqno; | ||
447 | len = skb->len; | ||
448 | netif_rx(skb); | ||
449 | break; | ||
450 | case QETH_HEADER_TYPE_OSN: | ||
451 | skb_push(skb, sizeof(struct qeth_hdr)); | ||
452 | skb_copy_to_linear_data(skb, hdr, | ||
453 | sizeof(struct qeth_hdr)); | ||
454 | len = skb->len; | ||
455 | card->osn_info.data_cb(skb); | ||
456 | break; | ||
457 | default: | ||
458 | dev_kfree_skb_any(skb); | ||
459 | QETH_DBF_TEXT(TRACE, 3, "inbunkno"); | ||
460 | QETH_DBF_HEX(CTRL, 3, hdr, QETH_DBF_CTRL_LEN); | ||
461 | continue; | ||
462 | } | ||
463 | card->dev->last_rx = jiffies; | ||
464 | card->stats.rx_packets++; | ||
465 | card->stats.rx_bytes += len; | ||
466 | } | ||
467 | } | ||
468 | |||
469 | static int qeth_l2_send_setdelmac(struct qeth_card *card, __u8 *mac, | ||
470 | enum qeth_ipa_cmds ipacmd, | ||
471 | int (*reply_cb) (struct qeth_card *, | ||
472 | struct qeth_reply*, | ||
473 | unsigned long)) | ||
474 | { | ||
475 | struct qeth_ipa_cmd *cmd; | ||
476 | struct qeth_cmd_buffer *iob; | ||
477 | |||
478 | QETH_DBF_TEXT(TRACE, 2, "L2sdmac"); | ||
479 | iob = qeth_get_ipacmd_buffer(card, ipacmd, QETH_PROT_IPV4); | ||
480 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
481 | cmd->data.setdelmac.mac_length = OSA_ADDR_LEN; | ||
482 | memcpy(&cmd->data.setdelmac.mac, mac, OSA_ADDR_LEN); | ||
483 | return qeth_send_ipa_cmd(card, iob, reply_cb, NULL); | ||
484 | } | ||
485 | |||
486 | static int qeth_l2_send_setmac_cb(struct qeth_card *card, | ||
487 | struct qeth_reply *reply, | ||
488 | unsigned long data) | ||
489 | { | ||
490 | struct qeth_ipa_cmd *cmd; | ||
491 | |||
492 | QETH_DBF_TEXT(TRACE, 2, "L2Smaccb"); | ||
493 | cmd = (struct qeth_ipa_cmd *) data; | ||
494 | if (cmd->hdr.return_code) { | ||
495 | QETH_DBF_TEXT_(TRACE, 2, "L2er%x", cmd->hdr.return_code); | ||
496 | card->info.mac_bits &= ~QETH_LAYER2_MAC_REGISTERED; | ||
497 | cmd->hdr.return_code = -EIO; | ||
498 | } else { | ||
499 | card->info.mac_bits |= QETH_LAYER2_MAC_REGISTERED; | ||
500 | memcpy(card->dev->dev_addr, cmd->data.setdelmac.mac, | ||
501 | OSA_ADDR_LEN); | ||
502 | PRINT_INFO("MAC address %2.2x:%2.2x:%2.2x:%2.2x:%2.2x:%2.2x " | ||
503 | "successfully registered on device %s\n", | ||
504 | card->dev->dev_addr[0], card->dev->dev_addr[1], | ||
505 | card->dev->dev_addr[2], card->dev->dev_addr[3], | ||
506 | card->dev->dev_addr[4], card->dev->dev_addr[5], | ||
507 | card->dev->name); | ||
508 | } | ||
509 | return 0; | ||
510 | } | ||
511 | |||
512 | static int qeth_l2_send_setmac(struct qeth_card *card, __u8 *mac) | ||
513 | { | ||
514 | QETH_DBF_TEXT(TRACE, 2, "L2Setmac"); | ||
515 | return qeth_l2_send_setdelmac(card, mac, IPA_CMD_SETVMAC, | ||
516 | qeth_l2_send_setmac_cb); | ||
517 | } | ||
518 | |||
519 | static int qeth_l2_send_delmac_cb(struct qeth_card *card, | ||
520 | struct qeth_reply *reply, | ||
521 | unsigned long data) | ||
522 | { | ||
523 | struct qeth_ipa_cmd *cmd; | ||
524 | |||
525 | QETH_DBF_TEXT(TRACE, 2, "L2Dmaccb"); | ||
526 | cmd = (struct qeth_ipa_cmd *) data; | ||
527 | if (cmd->hdr.return_code) { | ||
528 | QETH_DBF_TEXT_(TRACE, 2, "err%d", cmd->hdr.return_code); | ||
529 | cmd->hdr.return_code = -EIO; | ||
530 | return 0; | ||
531 | } | ||
532 | card->info.mac_bits &= ~QETH_LAYER2_MAC_REGISTERED; | ||
533 | |||
534 | return 0; | ||
535 | } | ||
536 | |||
537 | static int qeth_l2_send_delmac(struct qeth_card *card, __u8 *mac) | ||
538 | { | ||
539 | QETH_DBF_TEXT(TRACE, 2, "L2Delmac"); | ||
540 | if (!(card->info.mac_bits & QETH_LAYER2_MAC_REGISTERED)) | ||
541 | return 0; | ||
542 | return qeth_l2_send_setdelmac(card, mac, IPA_CMD_DELVMAC, | ||
543 | qeth_l2_send_delmac_cb); | ||
544 | } | ||
545 | |||
546 | static int qeth_l2_request_initial_mac(struct qeth_card *card) | ||
547 | { | ||
548 | int rc = 0; | ||
549 | char vendor_pre[] = {0x02, 0x00, 0x00}; | ||
550 | |||
551 | QETH_DBF_TEXT(SETUP, 2, "doL2init"); | ||
552 | QETH_DBF_TEXT_(SETUP, 2, "doL2%s", CARD_BUS_ID(card)); | ||
553 | |||
554 | rc = qeth_query_setadapterparms(card); | ||
555 | if (rc) { | ||
556 | PRINT_WARN("could not query adapter parameters on device %s: " | ||
557 | "x%x\n", CARD_BUS_ID(card), rc); | ||
558 | } | ||
559 | |||
560 | if (card->info.guestlan) { | ||
561 | rc = qeth_setadpparms_change_macaddr(card); | ||
562 | if (rc) { | ||
563 | PRINT_WARN("couldn't get MAC address on " | ||
564 | "device %s: x%x\n", | ||
565 | CARD_BUS_ID(card), rc); | ||
566 | QETH_DBF_TEXT_(SETUP, 2, "1err%d", rc); | ||
567 | return rc; | ||
568 | } | ||
569 | QETH_DBF_HEX(SETUP, 2, card->dev->dev_addr, OSA_ADDR_LEN); | ||
570 | } else { | ||
571 | random_ether_addr(card->dev->dev_addr); | ||
572 | memcpy(card->dev->dev_addr, vendor_pre, 3); | ||
573 | } | ||
574 | return 0; | ||
575 | } | ||
576 | |||
577 | static int qeth_l2_set_mac_address(struct net_device *dev, void *p) | ||
578 | { | ||
579 | struct sockaddr *addr = p; | ||
580 | struct qeth_card *card = netdev_priv(dev); | ||
581 | int rc = 0; | ||
582 | |||
583 | QETH_DBF_TEXT(TRACE, 3, "setmac"); | ||
584 | |||
585 | if (qeth_l2_verify_dev(dev) != QETH_REAL_CARD) { | ||
586 | QETH_DBF_TEXT(TRACE, 3, "setmcINV"); | ||
587 | return -EOPNOTSUPP; | ||
588 | } | ||
589 | |||
590 | if (card->info.type == QETH_CARD_TYPE_OSN) { | ||
591 | PRINT_WARN("Setting MAC address on %s is not supported.\n", | ||
592 | dev->name); | ||
593 | QETH_DBF_TEXT(TRACE, 3, "setmcOSN"); | ||
594 | return -EOPNOTSUPP; | ||
595 | } | ||
596 | QETH_DBF_TEXT_(TRACE, 3, "%s", CARD_BUS_ID(card)); | ||
597 | QETH_DBF_HEX(TRACE, 3, addr->sa_data, OSA_ADDR_LEN); | ||
598 | rc = qeth_l2_send_delmac(card, &card->dev->dev_addr[0]); | ||
599 | if (!rc) | ||
600 | rc = qeth_l2_send_setmac(card, addr->sa_data); | ||
601 | return rc; | ||
602 | } | ||
603 | |||
604 | static void qeth_l2_set_multicast_list(struct net_device *dev) | ||
605 | { | ||
606 | struct qeth_card *card = netdev_priv(dev); | ||
607 | struct dev_mc_list *dm; | ||
608 | |||
609 | if (card->info.type == QETH_CARD_TYPE_OSN) | ||
610 | return ; | ||
611 | |||
612 | QETH_DBF_TEXT(TRACE, 3, "setmulti"); | ||
613 | qeth_l2_del_all_mc(card); | ||
614 | spin_lock_bh(&card->mclock); | ||
615 | for (dm = dev->mc_list; dm; dm = dm->next) | ||
616 | qeth_l2_add_mc(card, dm->dmi_addr); | ||
617 | spin_unlock_bh(&card->mclock); | ||
618 | if (!qeth_adp_supported(card, IPA_SETADP_SET_PROMISC_MODE)) | ||
619 | return; | ||
620 | qeth_setadp_promisc_mode(card); | ||
621 | } | ||
622 | |||
623 | static int qeth_l2_hard_start_xmit(struct sk_buff *skb, struct net_device *dev) | ||
624 | { | ||
625 | int rc; | ||
626 | struct qeth_hdr *hdr = NULL; | ||
627 | int elements = 0; | ||
628 | struct qeth_card *card = netdev_priv(dev); | ||
629 | struct sk_buff *new_skb = skb; | ||
630 | int ipv = qeth_get_ip_version(skb); | ||
631 | int cast_type = qeth_get_cast_type(card, skb); | ||
632 | struct qeth_qdio_out_q *queue = card->qdio.out_qs | ||
633 | [qeth_get_priority_queue(card, skb, ipv, cast_type)]; | ||
634 | int tx_bytes = skb->len; | ||
635 | enum qeth_large_send_types large_send = QETH_LARGE_SEND_NO; | ||
636 | struct qeth_eddp_context *ctx = NULL; | ||
637 | |||
638 | QETH_DBF_TEXT(TRACE, 6, "l2xmit"); | ||
639 | |||
640 | if ((card->state != CARD_STATE_UP) || !card->lan_online) { | ||
641 | card->stats.tx_carrier_errors++; | ||
642 | goto tx_drop; | ||
643 | } | ||
644 | |||
645 | if ((card->info.type == QETH_CARD_TYPE_OSN) && | ||
646 | (skb->protocol == htons(ETH_P_IPV6))) | ||
647 | goto tx_drop; | ||
648 | |||
649 | if (card->options.performance_stats) { | ||
650 | card->perf_stats.outbound_cnt++; | ||
651 | card->perf_stats.outbound_start_time = qeth_get_micros(); | ||
652 | } | ||
653 | netif_stop_queue(dev); | ||
654 | |||
655 | if (skb_is_gso(skb)) | ||
656 | large_send = QETH_LARGE_SEND_EDDP; | ||
657 | |||
658 | if (card->info.type == QETH_CARD_TYPE_OSN) | ||
659 | hdr = (struct qeth_hdr *)skb->data; | ||
660 | else { | ||
661 | new_skb = qeth_prepare_skb(card, skb, &hdr); | ||
662 | if (!new_skb) | ||
663 | goto tx_drop; | ||
664 | qeth_l2_fill_header(card, hdr, new_skb, ipv, cast_type); | ||
665 | } | ||
666 | |||
667 | if (large_send == QETH_LARGE_SEND_EDDP) { | ||
668 | ctx = qeth_eddp_create_context(card, new_skb, hdr, | ||
669 | skb->sk->sk_protocol); | ||
670 | if (ctx == NULL) { | ||
671 | PRINT_WARN("could not create eddp context\n"); | ||
672 | goto tx_drop; | ||
673 | } | ||
674 | } else { | ||
675 | elements = qeth_get_elements_no(card, (void *)hdr, new_skb, 0); | ||
676 | if (!elements) | ||
677 | goto tx_drop; | ||
678 | } | ||
679 | |||
680 | if ((large_send == QETH_LARGE_SEND_NO) && | ||
681 | (skb->ip_summed == CHECKSUM_PARTIAL)) | ||
682 | qeth_tx_csum(new_skb); | ||
683 | |||
684 | if (card->info.type != QETH_CARD_TYPE_IQD) | ||
685 | rc = qeth_do_send_packet(card, queue, new_skb, hdr, | ||
686 | elements, ctx); | ||
687 | else | ||
688 | rc = qeth_do_send_packet_fast(card, queue, new_skb, hdr, | ||
689 | elements, ctx); | ||
690 | if (!rc) { | ||
691 | card->stats.tx_packets++; | ||
692 | card->stats.tx_bytes += tx_bytes; | ||
693 | if (new_skb != skb) | ||
694 | dev_kfree_skb_any(skb); | ||
695 | if (card->options.performance_stats) { | ||
696 | if (large_send != QETH_LARGE_SEND_NO) { | ||
697 | card->perf_stats.large_send_bytes += tx_bytes; | ||
698 | card->perf_stats.large_send_cnt++; | ||
699 | } | ||
700 | if (skb_shinfo(new_skb)->nr_frags > 0) { | ||
701 | card->perf_stats.sg_skbs_sent++; | ||
702 | /* nr_frags + skb->data */ | ||
703 | card->perf_stats.sg_frags_sent += | ||
704 | skb_shinfo(new_skb)->nr_frags + 1; | ||
705 | } | ||
706 | } | ||
707 | |||
708 | if (ctx != NULL) { | ||
709 | qeth_eddp_put_context(ctx); | ||
710 | dev_kfree_skb_any(new_skb); | ||
711 | } | ||
712 | } else { | ||
713 | if (ctx != NULL) | ||
714 | qeth_eddp_put_context(ctx); | ||
715 | |||
716 | if (rc == -EBUSY) { | ||
717 | if (new_skb != skb) | ||
718 | dev_kfree_skb_any(new_skb); | ||
719 | return NETDEV_TX_BUSY; | ||
720 | } else | ||
721 | goto tx_drop; | ||
722 | } | ||
723 | |||
724 | netif_wake_queue(dev); | ||
725 | if (card->options.performance_stats) | ||
726 | card->perf_stats.outbound_time += qeth_get_micros() - | ||
727 | card->perf_stats.outbound_start_time; | ||
728 | return rc; | ||
729 | |||
730 | tx_drop: | ||
731 | card->stats.tx_dropped++; | ||
732 | card->stats.tx_errors++; | ||
733 | if ((new_skb != skb) && new_skb) | ||
734 | dev_kfree_skb_any(new_skb); | ||
735 | dev_kfree_skb_any(skb); | ||
736 | return NETDEV_TX_OK; | ||
737 | } | ||
738 | |||
739 | static void qeth_l2_qdio_input_handler(struct ccw_device *ccwdev, | ||
740 | unsigned int status, unsigned int qdio_err, | ||
741 | unsigned int siga_err, unsigned int queue, | ||
742 | int first_element, int count, unsigned long card_ptr) | ||
743 | { | ||
744 | struct net_device *net_dev; | ||
745 | struct qeth_card *card; | ||
746 | struct qeth_qdio_buffer *buffer; | ||
747 | int index; | ||
748 | int i; | ||
749 | |||
750 | QETH_DBF_TEXT(TRACE, 6, "qdinput"); | ||
751 | card = (struct qeth_card *) card_ptr; | ||
752 | net_dev = card->dev; | ||
753 | if (card->options.performance_stats) { | ||
754 | card->perf_stats.inbound_cnt++; | ||
755 | card->perf_stats.inbound_start_time = qeth_get_micros(); | ||
756 | } | ||
757 | if (status & QDIO_STATUS_LOOK_FOR_ERROR) { | ||
758 | if (status & QDIO_STATUS_ACTIVATE_CHECK_CONDITION) { | ||
759 | QETH_DBF_TEXT(TRACE, 1, "qdinchk"); | ||
760 | QETH_DBF_TEXT_(TRACE, 1, "%s", CARD_BUS_ID(card)); | ||
761 | QETH_DBF_TEXT_(TRACE, 1, "%04X%04X", first_element, | ||
762 | count); | ||
763 | QETH_DBF_TEXT_(TRACE, 1, "%04X%04X", queue, status); | ||
764 | qeth_schedule_recovery(card); | ||
765 | return; | ||
766 | } | ||
767 | } | ||
768 | for (i = first_element; i < (first_element + count); ++i) { | ||
769 | index = i % QDIO_MAX_BUFFERS_PER_Q; | ||
770 | buffer = &card->qdio.in_q->bufs[index]; | ||
771 | if (!((status & QDIO_STATUS_LOOK_FOR_ERROR) && | ||
772 | qeth_check_qdio_errors(buffer->buffer, | ||
773 | qdio_err, siga_err, "qinerr"))) | ||
774 | qeth_l2_process_inbound_buffer(card, buffer, index); | ||
775 | /* clear buffer and give back to hardware */ | ||
776 | qeth_put_buffer_pool_entry(card, buffer->pool_entry); | ||
777 | qeth_queue_input_buffer(card, index); | ||
778 | } | ||
779 | if (card->options.performance_stats) | ||
780 | card->perf_stats.inbound_time += qeth_get_micros() - | ||
781 | card->perf_stats.inbound_start_time; | ||
782 | } | ||
783 | |||
784 | static int qeth_l2_open(struct net_device *dev) | ||
785 | { | ||
786 | struct qeth_card *card = netdev_priv(dev); | ||
787 | |||
788 | QETH_DBF_TEXT(TRACE, 4, "qethopen"); | ||
789 | if (card->state != CARD_STATE_SOFTSETUP) | ||
790 | return -ENODEV; | ||
791 | |||
792 | if ((card->info.type != QETH_CARD_TYPE_OSN) && | ||
793 | (!(card->info.mac_bits & QETH_LAYER2_MAC_REGISTERED))) { | ||
794 | QETH_DBF_TEXT(TRACE, 4, "nomacadr"); | ||
795 | return -EPERM; | ||
796 | } | ||
797 | card->data.state = CH_STATE_UP; | ||
798 | card->state = CARD_STATE_UP; | ||
799 | card->dev->flags |= IFF_UP; | ||
800 | netif_start_queue(dev); | ||
801 | |||
802 | if (!card->lan_online && netif_carrier_ok(dev)) | ||
803 | netif_carrier_off(dev); | ||
804 | return 0; | ||
805 | } | ||
806 | |||
807 | |||
808 | static int qeth_l2_stop(struct net_device *dev) | ||
809 | { | ||
810 | struct qeth_card *card = netdev_priv(dev); | ||
811 | |||
812 | QETH_DBF_TEXT(TRACE, 4, "qethstop"); | ||
813 | netif_tx_disable(dev); | ||
814 | card->dev->flags &= ~IFF_UP; | ||
815 | if (card->state == CARD_STATE_UP) | ||
816 | card->state = CARD_STATE_SOFTSETUP; | ||
817 | return 0; | ||
818 | } | ||
819 | |||
820 | static int qeth_l2_probe_device(struct ccwgroup_device *gdev) | ||
821 | { | ||
822 | struct qeth_card *card = dev_get_drvdata(&gdev->dev); | ||
823 | |||
824 | INIT_LIST_HEAD(&card->vid_list); | ||
825 | INIT_LIST_HEAD(&card->mc_list); | ||
826 | card->options.layer2 = 1; | ||
827 | card->discipline.input_handler = (qdio_handler_t *) | ||
828 | qeth_l2_qdio_input_handler; | ||
829 | card->discipline.output_handler = (qdio_handler_t *) | ||
830 | qeth_qdio_output_handler; | ||
831 | card->discipline.recover = qeth_l2_recover; | ||
832 | return 0; | ||
833 | } | ||
834 | |||
835 | static void qeth_l2_remove_device(struct ccwgroup_device *cgdev) | ||
836 | { | ||
837 | struct qeth_card *card = dev_get_drvdata(&cgdev->dev); | ||
838 | |||
839 | wait_event(card->wait_q, qeth_threads_running(card, 0xffffffff) == 0); | ||
840 | |||
841 | if (cgdev->state == CCWGROUP_ONLINE) { | ||
842 | card->use_hard_stop = 1; | ||
843 | qeth_l2_set_offline(cgdev); | ||
844 | } | ||
845 | |||
846 | if (card->dev) { | ||
847 | unregister_netdev(card->dev); | ||
848 | card->dev = NULL; | ||
849 | } | ||
850 | |||
851 | qeth_l2_del_all_mc(card); | ||
852 | return; | ||
853 | } | ||
854 | |||
855 | static struct ethtool_ops qeth_l2_ethtool_ops = { | ||
856 | .get_link = ethtool_op_get_link, | ||
857 | .get_tx_csum = ethtool_op_get_tx_csum, | ||
858 | .set_tx_csum = ethtool_op_set_tx_hw_csum, | ||
859 | .get_sg = ethtool_op_get_sg, | ||
860 | .set_sg = ethtool_op_set_sg, | ||
861 | .get_tso = ethtool_op_get_tso, | ||
862 | .set_tso = ethtool_op_set_tso, | ||
863 | .get_strings = qeth_core_get_strings, | ||
864 | .get_ethtool_stats = qeth_core_get_ethtool_stats, | ||
865 | .get_stats_count = qeth_core_get_stats_count, | ||
866 | .get_drvinfo = qeth_core_get_drvinfo, | ||
867 | }; | ||
868 | |||
869 | static struct ethtool_ops qeth_l2_osn_ops = { | ||
870 | .get_strings = qeth_core_get_strings, | ||
871 | .get_ethtool_stats = qeth_core_get_ethtool_stats, | ||
872 | .get_stats_count = qeth_core_get_stats_count, | ||
873 | .get_drvinfo = qeth_core_get_drvinfo, | ||
874 | }; | ||
875 | |||
876 | static int qeth_l2_setup_netdev(struct qeth_card *card) | ||
877 | { | ||
878 | switch (card->info.type) { | ||
879 | case QETH_CARD_TYPE_OSAE: | ||
880 | card->dev = alloc_etherdev(0); | ||
881 | break; | ||
882 | case QETH_CARD_TYPE_IQD: | ||
883 | card->dev = alloc_netdev(0, "hsi%d", ether_setup); | ||
884 | break; | ||
885 | case QETH_CARD_TYPE_OSN: | ||
886 | card->dev = alloc_netdev(0, "osn%d", ether_setup); | ||
887 | card->dev->flags |= IFF_NOARP; | ||
888 | break; | ||
889 | default: | ||
890 | card->dev = alloc_etherdev(0); | ||
891 | } | ||
892 | |||
893 | if (!card->dev) | ||
894 | return -ENODEV; | ||
895 | |||
896 | card->dev->priv = card; | ||
897 | card->dev->tx_timeout = &qeth_tx_timeout; | ||
898 | card->dev->watchdog_timeo = QETH_TX_TIMEOUT; | ||
899 | card->dev->open = qeth_l2_open; | ||
900 | card->dev->stop = qeth_l2_stop; | ||
901 | card->dev->hard_start_xmit = qeth_l2_hard_start_xmit; | ||
902 | card->dev->do_ioctl = qeth_l2_do_ioctl; | ||
903 | card->dev->get_stats = qeth_get_stats; | ||
904 | card->dev->change_mtu = qeth_change_mtu; | ||
905 | card->dev->set_multicast_list = qeth_l2_set_multicast_list; | ||
906 | card->dev->vlan_rx_kill_vid = qeth_l2_vlan_rx_kill_vid; | ||
907 | card->dev->vlan_rx_add_vid = qeth_l2_vlan_rx_add_vid; | ||
908 | card->dev->set_mac_address = qeth_l2_set_mac_address; | ||
909 | card->dev->mtu = card->info.initial_mtu; | ||
910 | if (card->info.type != QETH_CARD_TYPE_OSN) | ||
911 | SET_ETHTOOL_OPS(card->dev, &qeth_l2_ethtool_ops); | ||
912 | else | ||
913 | SET_ETHTOOL_OPS(card->dev, &qeth_l2_osn_ops); | ||
914 | card->dev->features |= NETIF_F_HW_VLAN_FILTER; | ||
915 | card->info.broadcast_capable = 1; | ||
916 | qeth_l2_request_initial_mac(card); | ||
917 | SET_NETDEV_DEV(card->dev, &card->gdev->dev); | ||
918 | return register_netdev(card->dev); | ||
919 | } | ||
920 | |||
921 | static int __qeth_l2_set_online(struct ccwgroup_device *gdev, int recovery_mode) | ||
922 | { | ||
923 | struct qeth_card *card = dev_get_drvdata(&gdev->dev); | ||
924 | int rc = 0; | ||
925 | enum qeth_card_states recover_flag; | ||
926 | |||
927 | BUG_ON(!card); | ||
928 | QETH_DBF_TEXT(SETUP, 2, "setonlin"); | ||
929 | QETH_DBF_HEX(SETUP, 2, &card, sizeof(void *)); | ||
930 | |||
931 | qeth_set_allowed_threads(card, QETH_RECOVER_THREAD, 1); | ||
932 | if (qeth_wait_for_threads(card, ~QETH_RECOVER_THREAD)) { | ||
933 | PRINT_WARN("set_online of card %s interrupted by user!\n", | ||
934 | CARD_BUS_ID(card)); | ||
935 | return -ERESTARTSYS; | ||
936 | } | ||
937 | |||
938 | recover_flag = card->state; | ||
939 | rc = ccw_device_set_online(CARD_RDEV(card)); | ||
940 | if (rc) { | ||
941 | QETH_DBF_TEXT_(SETUP, 2, "1err%d", rc); | ||
942 | return -EIO; | ||
943 | } | ||
944 | rc = ccw_device_set_online(CARD_WDEV(card)); | ||
945 | if (rc) { | ||
946 | QETH_DBF_TEXT_(SETUP, 2, "1err%d", rc); | ||
947 | return -EIO; | ||
948 | } | ||
949 | rc = ccw_device_set_online(CARD_DDEV(card)); | ||
950 | if (rc) { | ||
951 | QETH_DBF_TEXT_(SETUP, 2, "1err%d", rc); | ||
952 | return -EIO; | ||
953 | } | ||
954 | |||
955 | rc = qeth_core_hardsetup_card(card); | ||
956 | if (rc) { | ||
957 | QETH_DBF_TEXT_(SETUP, 2, "2err%d", rc); | ||
958 | goto out_remove; | ||
959 | } | ||
960 | |||
961 | if (!card->dev && qeth_l2_setup_netdev(card)) | ||
962 | goto out_remove; | ||
963 | |||
964 | if (card->info.type != QETH_CARD_TYPE_OSN) | ||
965 | qeth_l2_send_setmac(card, &card->dev->dev_addr[0]); | ||
966 | |||
967 | card->state = CARD_STATE_HARDSETUP; | ||
968 | qeth_print_status_message(card); | ||
969 | |||
970 | /* softsetup */ | ||
971 | QETH_DBF_TEXT(SETUP, 2, "softsetp"); | ||
972 | |||
973 | rc = qeth_send_startlan(card); | ||
974 | if (rc) { | ||
975 | QETH_DBF_TEXT_(SETUP, 2, "1err%d", rc); | ||
976 | if (rc == 0xe080) { | ||
977 | PRINT_WARN("LAN on card %s if offline! " | ||
978 | "Waiting for STARTLAN from card.\n", | ||
979 | CARD_BUS_ID(card)); | ||
980 | card->lan_online = 0; | ||
981 | } | ||
982 | return rc; | ||
983 | } else | ||
984 | card->lan_online = 1; | ||
985 | |||
986 | if (card->info.type != QETH_CARD_TYPE_OSN) { | ||
987 | qeth_set_large_send(card, card->options.large_send); | ||
988 | qeth_l2_process_vlans(card, 0); | ||
989 | } | ||
990 | |||
991 | netif_tx_disable(card->dev); | ||
992 | |||
993 | rc = qeth_init_qdio_queues(card); | ||
994 | if (rc) { | ||
995 | QETH_DBF_TEXT_(SETUP, 2, "6err%d", rc); | ||
996 | goto out_remove; | ||
997 | } | ||
998 | card->state = CARD_STATE_SOFTSETUP; | ||
999 | netif_carrier_on(card->dev); | ||
1000 | |||
1001 | qeth_set_allowed_threads(card, 0xffffffff, 0); | ||
1002 | if (recover_flag == CARD_STATE_RECOVER) { | ||
1003 | if (recovery_mode && | ||
1004 | card->info.type != QETH_CARD_TYPE_OSN) { | ||
1005 | qeth_l2_open(card->dev); | ||
1006 | } else { | ||
1007 | rtnl_lock(); | ||
1008 | dev_open(card->dev); | ||
1009 | rtnl_unlock(); | ||
1010 | } | ||
1011 | /* this also sets saved unicast addresses */ | ||
1012 | qeth_l2_set_multicast_list(card->dev); | ||
1013 | } | ||
1014 | /* let user_space know that device is online */ | ||
1015 | kobject_uevent(&gdev->dev.kobj, KOBJ_CHANGE); | ||
1016 | return 0; | ||
1017 | out_remove: | ||
1018 | card->use_hard_stop = 1; | ||
1019 | qeth_l2_stop_card(card, 0); | ||
1020 | ccw_device_set_offline(CARD_DDEV(card)); | ||
1021 | ccw_device_set_offline(CARD_WDEV(card)); | ||
1022 | ccw_device_set_offline(CARD_RDEV(card)); | ||
1023 | if (recover_flag == CARD_STATE_RECOVER) | ||
1024 | card->state = CARD_STATE_RECOVER; | ||
1025 | else | ||
1026 | card->state = CARD_STATE_DOWN; | ||
1027 | return -ENODEV; | ||
1028 | } | ||
1029 | |||
1030 | static int qeth_l2_set_online(struct ccwgroup_device *gdev) | ||
1031 | { | ||
1032 | return __qeth_l2_set_online(gdev, 0); | ||
1033 | } | ||
1034 | |||
1035 | static int __qeth_l2_set_offline(struct ccwgroup_device *cgdev, | ||
1036 | int recovery_mode) | ||
1037 | { | ||
1038 | struct qeth_card *card = dev_get_drvdata(&cgdev->dev); | ||
1039 | int rc = 0, rc2 = 0, rc3 = 0; | ||
1040 | enum qeth_card_states recover_flag; | ||
1041 | |||
1042 | QETH_DBF_TEXT(SETUP, 3, "setoffl"); | ||
1043 | QETH_DBF_HEX(SETUP, 3, &card, sizeof(void *)); | ||
1044 | |||
1045 | if (card->dev && netif_carrier_ok(card->dev)) | ||
1046 | netif_carrier_off(card->dev); | ||
1047 | recover_flag = card->state; | ||
1048 | if (qeth_l2_stop_card(card, recovery_mode) == -ERESTARTSYS) { | ||
1049 | PRINT_WARN("Stopping card %s interrupted by user!\n", | ||
1050 | CARD_BUS_ID(card)); | ||
1051 | return -ERESTARTSYS; | ||
1052 | } | ||
1053 | rc = ccw_device_set_offline(CARD_DDEV(card)); | ||
1054 | rc2 = ccw_device_set_offline(CARD_WDEV(card)); | ||
1055 | rc3 = ccw_device_set_offline(CARD_RDEV(card)); | ||
1056 | if (!rc) | ||
1057 | rc = (rc2) ? rc2 : rc3; | ||
1058 | if (rc) | ||
1059 | QETH_DBF_TEXT_(SETUP, 2, "1err%d", rc); | ||
1060 | if (recover_flag == CARD_STATE_UP) | ||
1061 | card->state = CARD_STATE_RECOVER; | ||
1062 | /* let user_space know that device is offline */ | ||
1063 | kobject_uevent(&cgdev->dev.kobj, KOBJ_CHANGE); | ||
1064 | return 0; | ||
1065 | } | ||
1066 | |||
1067 | static int qeth_l2_set_offline(struct ccwgroup_device *cgdev) | ||
1068 | { | ||
1069 | return __qeth_l2_set_offline(cgdev, 0); | ||
1070 | } | ||
1071 | |||
1072 | static int qeth_l2_recover(void *ptr) | ||
1073 | { | ||
1074 | struct qeth_card *card; | ||
1075 | int rc = 0; | ||
1076 | |||
1077 | card = (struct qeth_card *) ptr; | ||
1078 | QETH_DBF_TEXT(TRACE, 2, "recover1"); | ||
1079 | QETH_DBF_HEX(TRACE, 2, &card, sizeof(void *)); | ||
1080 | if (!qeth_do_run_thread(card, QETH_RECOVER_THREAD)) | ||
1081 | return 0; | ||
1082 | QETH_DBF_TEXT(TRACE, 2, "recover2"); | ||
1083 | PRINT_WARN("Recovery of device %s started ...\n", | ||
1084 | CARD_BUS_ID(card)); | ||
1085 | card->use_hard_stop = 1; | ||
1086 | __qeth_l2_set_offline(card->gdev, 1); | ||
1087 | rc = __qeth_l2_set_online(card->gdev, 1); | ||
1088 | /* don't run another scheduled recovery */ | ||
1089 | qeth_clear_thread_start_bit(card, QETH_RECOVER_THREAD); | ||
1090 | qeth_clear_thread_running_bit(card, QETH_RECOVER_THREAD); | ||
1091 | if (!rc) | ||
1092 | PRINT_INFO("Device %s successfully recovered!\n", | ||
1093 | CARD_BUS_ID(card)); | ||
1094 | else | ||
1095 | PRINT_INFO("Device %s could not be recovered!\n", | ||
1096 | CARD_BUS_ID(card)); | ||
1097 | return 0; | ||
1098 | } | ||
1099 | |||
1100 | static int __init qeth_l2_init(void) | ||
1101 | { | ||
1102 | PRINT_INFO("register layer 2 discipline\n"); | ||
1103 | return 0; | ||
1104 | } | ||
1105 | |||
1106 | static void __exit qeth_l2_exit(void) | ||
1107 | { | ||
1108 | PRINT_INFO("unregister layer 2 discipline\n"); | ||
1109 | } | ||
1110 | |||
1111 | static void qeth_l2_shutdown(struct ccwgroup_device *gdev) | ||
1112 | { | ||
1113 | struct qeth_card *card = dev_get_drvdata(&gdev->dev); | ||
1114 | qeth_qdio_clear_card(card, 0); | ||
1115 | qeth_clear_qdio_buffers(card); | ||
1116 | } | ||
1117 | |||
1118 | struct ccwgroup_driver qeth_l2_ccwgroup_driver = { | ||
1119 | .probe = qeth_l2_probe_device, | ||
1120 | .remove = qeth_l2_remove_device, | ||
1121 | .set_online = qeth_l2_set_online, | ||
1122 | .set_offline = qeth_l2_set_offline, | ||
1123 | .shutdown = qeth_l2_shutdown, | ||
1124 | }; | ||
1125 | EXPORT_SYMBOL_GPL(qeth_l2_ccwgroup_driver); | ||
1126 | |||
1127 | static int qeth_osn_send_control_data(struct qeth_card *card, int len, | ||
1128 | struct qeth_cmd_buffer *iob) | ||
1129 | { | ||
1130 | unsigned long flags; | ||
1131 | int rc = 0; | ||
1132 | |||
1133 | QETH_DBF_TEXT(TRACE, 5, "osndctrd"); | ||
1134 | |||
1135 | wait_event(card->wait_q, | ||
1136 | atomic_cmpxchg(&card->write.irq_pending, 0, 1) == 0); | ||
1137 | qeth_prepare_control_data(card, len, iob); | ||
1138 | QETH_DBF_TEXT(TRACE, 6, "osnoirqp"); | ||
1139 | spin_lock_irqsave(get_ccwdev_lock(card->write.ccwdev), flags); | ||
1140 | rc = ccw_device_start(card->write.ccwdev, &card->write.ccw, | ||
1141 | (addr_t) iob, 0, 0); | ||
1142 | spin_unlock_irqrestore(get_ccwdev_lock(card->write.ccwdev), flags); | ||
1143 | if (rc) { | ||
1144 | PRINT_WARN("qeth_osn_send_control_data: " | ||
1145 | "ccw_device_start rc = %i\n", rc); | ||
1146 | QETH_DBF_TEXT_(TRACE, 2, " err%d", rc); | ||
1147 | qeth_release_buffer(iob->channel, iob); | ||
1148 | atomic_set(&card->write.irq_pending, 0); | ||
1149 | wake_up(&card->wait_q); | ||
1150 | } | ||
1151 | return rc; | ||
1152 | } | ||
1153 | |||
1154 | static int qeth_osn_send_ipa_cmd(struct qeth_card *card, | ||
1155 | struct qeth_cmd_buffer *iob, int data_len) | ||
1156 | { | ||
1157 | u16 s1, s2; | ||
1158 | |||
1159 | QETH_DBF_TEXT(TRACE, 4, "osndipa"); | ||
1160 | |||
1161 | qeth_prepare_ipa_cmd(card, iob, QETH_PROT_OSN2); | ||
1162 | s1 = (u16)(IPA_PDU_HEADER_SIZE + data_len); | ||
1163 | s2 = (u16)data_len; | ||
1164 | memcpy(QETH_IPA_PDU_LEN_TOTAL(iob->data), &s1, 2); | ||
1165 | memcpy(QETH_IPA_PDU_LEN_PDU1(iob->data), &s2, 2); | ||
1166 | memcpy(QETH_IPA_PDU_LEN_PDU2(iob->data), &s2, 2); | ||
1167 | memcpy(QETH_IPA_PDU_LEN_PDU3(iob->data), &s2, 2); | ||
1168 | return qeth_osn_send_control_data(card, s1, iob); | ||
1169 | } | ||
1170 | |||
1171 | int qeth_osn_assist(struct net_device *dev, void *data, int data_len) | ||
1172 | { | ||
1173 | struct qeth_cmd_buffer *iob; | ||
1174 | struct qeth_card *card; | ||
1175 | int rc; | ||
1176 | |||
1177 | QETH_DBF_TEXT(TRACE, 2, "osnsdmc"); | ||
1178 | if (!dev) | ||
1179 | return -ENODEV; | ||
1180 | card = netdev_priv(dev); | ||
1181 | if (!card) | ||
1182 | return -ENODEV; | ||
1183 | if ((card->state != CARD_STATE_UP) && | ||
1184 | (card->state != CARD_STATE_SOFTSETUP)) | ||
1185 | return -ENODEV; | ||
1186 | iob = qeth_wait_for_buffer(&card->write); | ||
1187 | memcpy(iob->data+IPA_PDU_HEADER_SIZE, data, data_len); | ||
1188 | rc = qeth_osn_send_ipa_cmd(card, iob, data_len); | ||
1189 | return rc; | ||
1190 | } | ||
1191 | EXPORT_SYMBOL(qeth_osn_assist); | ||
1192 | |||
1193 | int qeth_osn_register(unsigned char *read_dev_no, struct net_device **dev, | ||
1194 | int (*assist_cb)(struct net_device *, void *), | ||
1195 | int (*data_cb)(struct sk_buff *)) | ||
1196 | { | ||
1197 | struct qeth_card *card; | ||
1198 | |||
1199 | QETH_DBF_TEXT(TRACE, 2, "osnreg"); | ||
1200 | *dev = qeth_l2_netdev_by_devno(read_dev_no); | ||
1201 | if (*dev == NULL) | ||
1202 | return -ENODEV; | ||
1203 | card = netdev_priv(*dev); | ||
1204 | if (!card) | ||
1205 | return -ENODEV; | ||
1206 | if ((assist_cb == NULL) || (data_cb == NULL)) | ||
1207 | return -EINVAL; | ||
1208 | card->osn_info.assist_cb = assist_cb; | ||
1209 | card->osn_info.data_cb = data_cb; | ||
1210 | return 0; | ||
1211 | } | ||
1212 | EXPORT_SYMBOL(qeth_osn_register); | ||
1213 | |||
1214 | void qeth_osn_deregister(struct net_device *dev) | ||
1215 | { | ||
1216 | struct qeth_card *card; | ||
1217 | |||
1218 | QETH_DBF_TEXT(TRACE, 2, "osndereg"); | ||
1219 | if (!dev) | ||
1220 | return; | ||
1221 | card = netdev_priv(dev); | ||
1222 | if (!card) | ||
1223 | return; | ||
1224 | card->osn_info.assist_cb = NULL; | ||
1225 | card->osn_info.data_cb = NULL; | ||
1226 | return; | ||
1227 | } | ||
1228 | EXPORT_SYMBOL(qeth_osn_deregister); | ||
1229 | |||
1230 | module_init(qeth_l2_init); | ||
1231 | module_exit(qeth_l2_exit); | ||
1232 | MODULE_AUTHOR("Frank Blaschka <frank.blaschka@de.ibm.com>"); | ||
1233 | MODULE_DESCRIPTION("qeth layer 2 discipline"); | ||
1234 | MODULE_LICENSE("GPL"); | ||
diff --git a/drivers/s390/net/qeth_l3.h b/drivers/s390/net/qeth_l3.h new file mode 100644 index 000000000000..1be353593a59 --- /dev/null +++ b/drivers/s390/net/qeth_l3.h | |||
@@ -0,0 +1,67 @@ | |||
1 | /* | ||
2 | * drivers/s390/net/qeth_l3.h | ||
3 | * | ||
4 | * Copyright IBM Corp. 2007 | ||
5 | * Author(s): Utz Bacher <utz.bacher@de.ibm.com>, | ||
6 | * Frank Pavlic <fpavlic@de.ibm.com>, | ||
7 | * Thomas Spatzier <tspat@de.ibm.com>, | ||
8 | * Frank Blaschka <frank.blaschka@de.ibm.com> | ||
9 | */ | ||
10 | |||
11 | #ifndef __QETH_L3_H__ | ||
12 | #define __QETH_L3_H__ | ||
13 | |||
14 | #include "qeth_core.h" | ||
15 | |||
16 | #define QETH_DBF_TXT_BUF qeth_l3_dbf_txt_buf | ||
17 | DECLARE_PER_CPU(char[256], qeth_l3_dbf_txt_buf); | ||
18 | |||
19 | struct qeth_ipaddr { | ||
20 | struct list_head entry; | ||
21 | enum qeth_ip_types type; | ||
22 | enum qeth_ipa_setdelip_flags set_flags; | ||
23 | enum qeth_ipa_setdelip_flags del_flags; | ||
24 | int is_multicast; | ||
25 | int users; | ||
26 | enum qeth_prot_versions proto; | ||
27 | unsigned char mac[OSA_ADDR_LEN]; | ||
28 | union { | ||
29 | struct { | ||
30 | unsigned int addr; | ||
31 | unsigned int mask; | ||
32 | } a4; | ||
33 | struct { | ||
34 | struct in6_addr addr; | ||
35 | unsigned int pfxlen; | ||
36 | } a6; | ||
37 | } u; | ||
38 | }; | ||
39 | |||
40 | struct qeth_ipato_entry { | ||
41 | struct list_head entry; | ||
42 | enum qeth_prot_versions proto; | ||
43 | char addr[16]; | ||
44 | int mask_bits; | ||
45 | }; | ||
46 | |||
47 | |||
48 | void qeth_l3_ipaddr4_to_string(const __u8 *, char *); | ||
49 | int qeth_l3_string_to_ipaddr4(const char *, __u8 *); | ||
50 | void qeth_l3_ipaddr6_to_string(const __u8 *, char *); | ||
51 | int qeth_l3_string_to_ipaddr6(const char *, __u8 *); | ||
52 | void qeth_l3_ipaddr_to_string(enum qeth_prot_versions, const __u8 *, char *); | ||
53 | int qeth_l3_string_to_ipaddr(const char *, enum qeth_prot_versions, __u8 *); | ||
54 | int qeth_l3_create_device_attributes(struct device *); | ||
55 | void qeth_l3_remove_device_attributes(struct device *); | ||
56 | int qeth_l3_setrouting_v4(struct qeth_card *); | ||
57 | int qeth_l3_setrouting_v6(struct qeth_card *); | ||
58 | int qeth_l3_add_ipato_entry(struct qeth_card *, struct qeth_ipato_entry *); | ||
59 | void qeth_l3_del_ipato_entry(struct qeth_card *, enum qeth_prot_versions, | ||
60 | u8 *, int); | ||
61 | int qeth_l3_add_vipa(struct qeth_card *, enum qeth_prot_versions, const u8 *); | ||
62 | void qeth_l3_del_vipa(struct qeth_card *, enum qeth_prot_versions, const u8 *); | ||
63 | int qeth_l3_add_rxip(struct qeth_card *, enum qeth_prot_versions, const u8 *); | ||
64 | void qeth_l3_del_rxip(struct qeth_card *card, enum qeth_prot_versions, | ||
65 | const u8 *); | ||
66 | |||
67 | #endif /* __QETH_L3_H__ */ | ||
diff --git a/drivers/s390/net/qeth_l3_main.c b/drivers/s390/net/qeth_l3_main.c new file mode 100644 index 000000000000..e1bfe56087d6 --- /dev/null +++ b/drivers/s390/net/qeth_l3_main.c | |||
@@ -0,0 +1,3396 @@ | |||
1 | /* | ||
2 | * drivers/s390/net/qeth_l3_main.c | ||
3 | * | ||
4 | * Copyright IBM Corp. 2007 | ||
5 | * Author(s): Utz Bacher <utz.bacher@de.ibm.com>, | ||
6 | * Frank Pavlic <fpavlic@de.ibm.com>, | ||
7 | * Thomas Spatzier <tspat@de.ibm.com>, | ||
8 | * Frank Blaschka <frank.blaschka@de.ibm.com> | ||
9 | */ | ||
10 | |||
11 | #include <linux/module.h> | ||
12 | #include <linux/moduleparam.h> | ||
13 | #include <linux/string.h> | ||
14 | #include <linux/errno.h> | ||
15 | #include <linux/kernel.h> | ||
16 | #include <linux/etherdevice.h> | ||
17 | #include <linux/mii.h> | ||
18 | #include <linux/ip.h> | ||
19 | #include <linux/reboot.h> | ||
20 | #include <linux/inetdevice.h> | ||
21 | #include <linux/igmp.h> | ||
22 | |||
23 | #include <net/ip.h> | ||
24 | #include <net/arp.h> | ||
25 | |||
26 | #include <asm/s390_rdev.h> | ||
27 | |||
28 | #include "qeth_l3.h" | ||
29 | #include "qeth_core_offl.h" | ||
30 | |||
31 | DEFINE_PER_CPU(char[256], qeth_l3_dbf_txt_buf); | ||
32 | |||
33 | static int qeth_l3_set_offline(struct ccwgroup_device *); | ||
34 | static int qeth_l3_recover(void *); | ||
35 | static int qeth_l3_stop(struct net_device *); | ||
36 | static void qeth_l3_set_multicast_list(struct net_device *); | ||
37 | static int qeth_l3_neigh_setup(struct net_device *, struct neigh_parms *); | ||
38 | static int qeth_l3_register_addr_entry(struct qeth_card *, | ||
39 | struct qeth_ipaddr *); | ||
40 | static int qeth_l3_deregister_addr_entry(struct qeth_card *, | ||
41 | struct qeth_ipaddr *); | ||
42 | static int __qeth_l3_set_online(struct ccwgroup_device *, int); | ||
43 | static int __qeth_l3_set_offline(struct ccwgroup_device *, int); | ||
44 | |||
45 | |||
46 | static int qeth_l3_isxdigit(char *buf) | ||
47 | { | ||
48 | while (*buf) { | ||
49 | if (!isxdigit(*buf++)) | ||
50 | return 0; | ||
51 | } | ||
52 | return 1; | ||
53 | } | ||
54 | |||
55 | void qeth_l3_ipaddr4_to_string(const __u8 *addr, char *buf) | ||
56 | { | ||
57 | sprintf(buf, "%i.%i.%i.%i", addr[0], addr[1], addr[2], addr[3]); | ||
58 | } | ||
59 | |||
60 | int qeth_l3_string_to_ipaddr4(const char *buf, __u8 *addr) | ||
61 | { | ||
62 | int count = 0, rc = 0; | ||
63 | int in[4]; | ||
64 | char c; | ||
65 | |||
66 | rc = sscanf(buf, "%u.%u.%u.%u%c", | ||
67 | &in[0], &in[1], &in[2], &in[3], &c); | ||
68 | if (rc != 4 && (rc != 5 || c != '\n')) | ||
69 | return -EINVAL; | ||
70 | for (count = 0; count < 4; count++) { | ||
71 | if (in[count] > 255) | ||
72 | return -EINVAL; | ||
73 | addr[count] = in[count]; | ||
74 | } | ||
75 | return 0; | ||
76 | } | ||
77 | |||
78 | void qeth_l3_ipaddr6_to_string(const __u8 *addr, char *buf) | ||
79 | { | ||
80 | sprintf(buf, "%02x%02x:%02x%02x:%02x%02x:%02x%02x" | ||
81 | ":%02x%02x:%02x%02x:%02x%02x:%02x%02x", | ||
82 | addr[0], addr[1], addr[2], addr[3], | ||
83 | addr[4], addr[5], addr[6], addr[7], | ||
84 | addr[8], addr[9], addr[10], addr[11], | ||
85 | addr[12], addr[13], addr[14], addr[15]); | ||
86 | } | ||
87 | |||
88 | int qeth_l3_string_to_ipaddr6(const char *buf, __u8 *addr) | ||
89 | { | ||
90 | const char *end, *end_tmp, *start; | ||
91 | __u16 *in; | ||
92 | char num[5]; | ||
93 | int num2, cnt, out, found, save_cnt; | ||
94 | unsigned short in_tmp[8] = {0, }; | ||
95 | |||
96 | cnt = out = found = save_cnt = num2 = 0; | ||
97 | end = start = buf; | ||
98 | in = (__u16 *) addr; | ||
99 | memset(in, 0, 16); | ||
100 | while (*end) { | ||
101 | end = strchr(start, ':'); | ||
102 | if (end == NULL) { | ||
103 | end = buf + strlen(buf); | ||
104 | end_tmp = strchr(start, '\n'); | ||
105 | if (end_tmp != NULL) | ||
106 | end = end_tmp; | ||
107 | out = 1; | ||
108 | } | ||
109 | if ((end - start)) { | ||
110 | memset(num, 0, 5); | ||
111 | if ((end - start) > 4) | ||
112 | return -EINVAL; | ||
113 | memcpy(num, start, end - start); | ||
114 | if (!qeth_l3_isxdigit(num)) | ||
115 | return -EINVAL; | ||
116 | sscanf(start, "%x", &num2); | ||
117 | if (found) | ||
118 | in_tmp[save_cnt++] = num2; | ||
119 | else | ||
120 | in[cnt++] = num2; | ||
121 | if (out) | ||
122 | break; | ||
123 | } else { | ||
124 | if (found) | ||
125 | return -EINVAL; | ||
126 | found = 1; | ||
127 | } | ||
128 | start = ++end; | ||
129 | } | ||
130 | if (cnt + save_cnt > 8) | ||
131 | return -EINVAL; | ||
132 | cnt = 7; | ||
133 | while (save_cnt) | ||
134 | in[cnt--] = in_tmp[--save_cnt]; | ||
135 | return 0; | ||
136 | } | ||
137 | |||
138 | void qeth_l3_ipaddr_to_string(enum qeth_prot_versions proto, const __u8 *addr, | ||
139 | char *buf) | ||
140 | { | ||
141 | if (proto == QETH_PROT_IPV4) | ||
142 | qeth_l3_ipaddr4_to_string(addr, buf); | ||
143 | else if (proto == QETH_PROT_IPV6) | ||
144 | qeth_l3_ipaddr6_to_string(addr, buf); | ||
145 | } | ||
146 | |||
147 | int qeth_l3_string_to_ipaddr(const char *buf, enum qeth_prot_versions proto, | ||
148 | __u8 *addr) | ||
149 | { | ||
150 | if (proto == QETH_PROT_IPV4) | ||
151 | return qeth_l3_string_to_ipaddr4(buf, addr); | ||
152 | else if (proto == QETH_PROT_IPV6) | ||
153 | return qeth_l3_string_to_ipaddr6(buf, addr); | ||
154 | else | ||
155 | return -EINVAL; | ||
156 | } | ||
157 | |||
158 | static void qeth_l3_convert_addr_to_bits(u8 *addr, u8 *bits, int len) | ||
159 | { | ||
160 | int i, j; | ||
161 | u8 octet; | ||
162 | |||
163 | for (i = 0; i < len; ++i) { | ||
164 | octet = addr[i]; | ||
165 | for (j = 7; j >= 0; --j) { | ||
166 | bits[i*8 + j] = octet & 1; | ||
167 | octet >>= 1; | ||
168 | } | ||
169 | } | ||
170 | } | ||
171 | |||
172 | static int qeth_l3_is_addr_covered_by_ipato(struct qeth_card *card, | ||
173 | struct qeth_ipaddr *addr) | ||
174 | { | ||
175 | struct qeth_ipato_entry *ipatoe; | ||
176 | u8 addr_bits[128] = {0, }; | ||
177 | u8 ipatoe_bits[128] = {0, }; | ||
178 | int rc = 0; | ||
179 | |||
180 | if (!card->ipato.enabled) | ||
181 | return 0; | ||
182 | |||
183 | qeth_l3_convert_addr_to_bits((u8 *) &addr->u, addr_bits, | ||
184 | (addr->proto == QETH_PROT_IPV4)? 4:16); | ||
185 | list_for_each_entry(ipatoe, &card->ipato.entries, entry) { | ||
186 | if (addr->proto != ipatoe->proto) | ||
187 | continue; | ||
188 | qeth_l3_convert_addr_to_bits(ipatoe->addr, ipatoe_bits, | ||
189 | (ipatoe->proto == QETH_PROT_IPV4) ? | ||
190 | 4 : 16); | ||
191 | if (addr->proto == QETH_PROT_IPV4) | ||
192 | rc = !memcmp(addr_bits, ipatoe_bits, | ||
193 | min(32, ipatoe->mask_bits)); | ||
194 | else | ||
195 | rc = !memcmp(addr_bits, ipatoe_bits, | ||
196 | min(128, ipatoe->mask_bits)); | ||
197 | if (rc) | ||
198 | break; | ||
199 | } | ||
200 | /* invert? */ | ||
201 | if ((addr->proto == QETH_PROT_IPV4) && card->ipato.invert4) | ||
202 | rc = !rc; | ||
203 | else if ((addr->proto == QETH_PROT_IPV6) && card->ipato.invert6) | ||
204 | rc = !rc; | ||
205 | |||
206 | return rc; | ||
207 | } | ||
208 | |||
209 | /* | ||
210 | * Add IP to be added to todo list. If there is already an "add todo" | ||
211 | * in this list we just incremenent the reference count. | ||
212 | * Returns 0 if we just incremented reference count. | ||
213 | */ | ||
214 | static int __qeth_l3_insert_ip_todo(struct qeth_card *card, | ||
215 | struct qeth_ipaddr *addr, int add) | ||
216 | { | ||
217 | struct qeth_ipaddr *tmp, *t; | ||
218 | int found = 0; | ||
219 | |||
220 | list_for_each_entry_safe(tmp, t, card->ip_tbd_list, entry) { | ||
221 | if ((addr->type == QETH_IP_TYPE_DEL_ALL_MC) && | ||
222 | (tmp->type == QETH_IP_TYPE_DEL_ALL_MC)) | ||
223 | return 0; | ||
224 | if ((tmp->proto == QETH_PROT_IPV4) && | ||
225 | (addr->proto == QETH_PROT_IPV4) && | ||
226 | (tmp->type == addr->type) && | ||
227 | (tmp->is_multicast == addr->is_multicast) && | ||
228 | (tmp->u.a4.addr == addr->u.a4.addr) && | ||
229 | (tmp->u.a4.mask == addr->u.a4.mask)) { | ||
230 | found = 1; | ||
231 | break; | ||
232 | } | ||
233 | if ((tmp->proto == QETH_PROT_IPV6) && | ||
234 | (addr->proto == QETH_PROT_IPV6) && | ||
235 | (tmp->type == addr->type) && | ||
236 | (tmp->is_multicast == addr->is_multicast) && | ||
237 | (tmp->u.a6.pfxlen == addr->u.a6.pfxlen) && | ||
238 | (memcmp(&tmp->u.a6.addr, &addr->u.a6.addr, | ||
239 | sizeof(struct in6_addr)) == 0)) { | ||
240 | found = 1; | ||
241 | break; | ||
242 | } | ||
243 | } | ||
244 | if (found) { | ||
245 | if (addr->users != 0) | ||
246 | tmp->users += addr->users; | ||
247 | else | ||
248 | tmp->users += add ? 1 : -1; | ||
249 | if (tmp->users == 0) { | ||
250 | list_del(&tmp->entry); | ||
251 | kfree(tmp); | ||
252 | } | ||
253 | return 0; | ||
254 | } else { | ||
255 | if (addr->type == QETH_IP_TYPE_DEL_ALL_MC) | ||
256 | list_add(&addr->entry, card->ip_tbd_list); | ||
257 | else { | ||
258 | if (addr->users == 0) | ||
259 | addr->users += add ? 1 : -1; | ||
260 | if (add && (addr->type == QETH_IP_TYPE_NORMAL) && | ||
261 | qeth_l3_is_addr_covered_by_ipato(card, addr)) { | ||
262 | QETH_DBF_TEXT(TRACE, 2, "tkovaddr"); | ||
263 | addr->set_flags |= QETH_IPA_SETIP_TAKEOVER_FLAG; | ||
264 | } | ||
265 | list_add_tail(&addr->entry, card->ip_tbd_list); | ||
266 | } | ||
267 | return 1; | ||
268 | } | ||
269 | } | ||
270 | |||
271 | static int qeth_l3_delete_ip(struct qeth_card *card, struct qeth_ipaddr *addr) | ||
272 | { | ||
273 | unsigned long flags; | ||
274 | int rc = 0; | ||
275 | |||
276 | QETH_DBF_TEXT(TRACE, 4, "delip"); | ||
277 | |||
278 | if (addr->proto == QETH_PROT_IPV4) | ||
279 | QETH_DBF_HEX(TRACE, 4, &addr->u.a4.addr, 4); | ||
280 | else { | ||
281 | QETH_DBF_HEX(TRACE, 4, &addr->u.a6.addr, 8); | ||
282 | QETH_DBF_HEX(TRACE, 4, ((char *)&addr->u.a6.addr) + 8, 8); | ||
283 | } | ||
284 | spin_lock_irqsave(&card->ip_lock, flags); | ||
285 | rc = __qeth_l3_insert_ip_todo(card, addr, 0); | ||
286 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
287 | return rc; | ||
288 | } | ||
289 | |||
290 | static int qeth_l3_add_ip(struct qeth_card *card, struct qeth_ipaddr *addr) | ||
291 | { | ||
292 | unsigned long flags; | ||
293 | int rc = 0; | ||
294 | |||
295 | QETH_DBF_TEXT(TRACE, 4, "addip"); | ||
296 | if (addr->proto == QETH_PROT_IPV4) | ||
297 | QETH_DBF_HEX(TRACE, 4, &addr->u.a4.addr, 4); | ||
298 | else { | ||
299 | QETH_DBF_HEX(TRACE, 4, &addr->u.a6.addr, 8); | ||
300 | QETH_DBF_HEX(TRACE, 4, ((char *)&addr->u.a6.addr) + 8, 8); | ||
301 | } | ||
302 | spin_lock_irqsave(&card->ip_lock, flags); | ||
303 | rc = __qeth_l3_insert_ip_todo(card, addr, 1); | ||
304 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
305 | return rc; | ||
306 | } | ||
307 | |||
308 | |||
309 | static struct qeth_ipaddr *qeth_l3_get_addr_buffer( | ||
310 | enum qeth_prot_versions prot) | ||
311 | { | ||
312 | struct qeth_ipaddr *addr; | ||
313 | |||
314 | addr = kzalloc(sizeof(struct qeth_ipaddr), GFP_ATOMIC); | ||
315 | if (addr == NULL) { | ||
316 | PRINT_WARN("Not enough memory to add address\n"); | ||
317 | return NULL; | ||
318 | } | ||
319 | addr->type = QETH_IP_TYPE_NORMAL; | ||
320 | addr->proto = prot; | ||
321 | return addr; | ||
322 | } | ||
323 | |||
324 | static void qeth_l3_delete_mc_addresses(struct qeth_card *card) | ||
325 | { | ||
326 | struct qeth_ipaddr *iptodo; | ||
327 | unsigned long flags; | ||
328 | |||
329 | QETH_DBF_TEXT(TRACE, 4, "delmc"); | ||
330 | iptodo = qeth_l3_get_addr_buffer(QETH_PROT_IPV4); | ||
331 | if (!iptodo) { | ||
332 | QETH_DBF_TEXT(TRACE, 2, "dmcnomem"); | ||
333 | return; | ||
334 | } | ||
335 | iptodo->type = QETH_IP_TYPE_DEL_ALL_MC; | ||
336 | spin_lock_irqsave(&card->ip_lock, flags); | ||
337 | if (!__qeth_l3_insert_ip_todo(card, iptodo, 0)) | ||
338 | kfree(iptodo); | ||
339 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
340 | } | ||
341 | |||
342 | /* | ||
343 | * Add/remove address to/from card's ip list, i.e. try to add or remove | ||
344 | * reference to/from an IP address that is already registered on the card. | ||
345 | * Returns: | ||
346 | * 0 address was on card and its reference count has been adjusted, | ||
347 | * but is still > 0, so nothing has to be done | ||
348 | * also returns 0 if card was not on card and the todo was to delete | ||
349 | * the address -> there is also nothing to be done | ||
350 | * 1 address was not on card and the todo is to add it to the card's ip | ||
351 | * list | ||
352 | * -1 address was on card and its reference count has been decremented | ||
353 | * to <= 0 by the todo -> address must be removed from card | ||
354 | */ | ||
355 | static int __qeth_l3_ref_ip_on_card(struct qeth_card *card, | ||
356 | struct qeth_ipaddr *todo, struct qeth_ipaddr **__addr) | ||
357 | { | ||
358 | struct qeth_ipaddr *addr; | ||
359 | int found = 0; | ||
360 | |||
361 | list_for_each_entry(addr, &card->ip_list, entry) { | ||
362 | if ((addr->proto == QETH_PROT_IPV4) && | ||
363 | (todo->proto == QETH_PROT_IPV4) && | ||
364 | (addr->type == todo->type) && | ||
365 | (addr->u.a4.addr == todo->u.a4.addr) && | ||
366 | (addr->u.a4.mask == todo->u.a4.mask)) { | ||
367 | found = 1; | ||
368 | break; | ||
369 | } | ||
370 | if ((addr->proto == QETH_PROT_IPV6) && | ||
371 | (todo->proto == QETH_PROT_IPV6) && | ||
372 | (addr->type == todo->type) && | ||
373 | (addr->u.a6.pfxlen == todo->u.a6.pfxlen) && | ||
374 | (memcmp(&addr->u.a6.addr, &todo->u.a6.addr, | ||
375 | sizeof(struct in6_addr)) == 0)) { | ||
376 | found = 1; | ||
377 | break; | ||
378 | } | ||
379 | } | ||
380 | if (found) { | ||
381 | addr->users += todo->users; | ||
382 | if (addr->users <= 0) { | ||
383 | *__addr = addr; | ||
384 | return -1; | ||
385 | } else { | ||
386 | /* for VIPA and RXIP limit refcount to 1 */ | ||
387 | if (addr->type != QETH_IP_TYPE_NORMAL) | ||
388 | addr->users = 1; | ||
389 | return 0; | ||
390 | } | ||
391 | } | ||
392 | if (todo->users > 0) { | ||
393 | /* for VIPA and RXIP limit refcount to 1 */ | ||
394 | if (todo->type != QETH_IP_TYPE_NORMAL) | ||
395 | todo->users = 1; | ||
396 | return 1; | ||
397 | } else | ||
398 | return 0; | ||
399 | } | ||
400 | |||
401 | static void __qeth_l3_delete_all_mc(struct qeth_card *card, | ||
402 | unsigned long *flags) | ||
403 | { | ||
404 | struct list_head fail_list; | ||
405 | struct qeth_ipaddr *addr, *tmp; | ||
406 | int rc; | ||
407 | |||
408 | INIT_LIST_HEAD(&fail_list); | ||
409 | again: | ||
410 | list_for_each_entry_safe(addr, tmp, &card->ip_list, entry) { | ||
411 | if (addr->is_multicast) { | ||
412 | list_del(&addr->entry); | ||
413 | spin_unlock_irqrestore(&card->ip_lock, *flags); | ||
414 | rc = qeth_l3_deregister_addr_entry(card, addr); | ||
415 | spin_lock_irqsave(&card->ip_lock, *flags); | ||
416 | if (!rc || (rc == IPA_RC_MC_ADDR_NOT_FOUND)) | ||
417 | kfree(addr); | ||
418 | else | ||
419 | list_add_tail(&addr->entry, &fail_list); | ||
420 | goto again; | ||
421 | } | ||
422 | } | ||
423 | list_splice(&fail_list, &card->ip_list); | ||
424 | } | ||
425 | |||
426 | static void qeth_l3_set_ip_addr_list(struct qeth_card *card) | ||
427 | { | ||
428 | struct list_head *tbd_list; | ||
429 | struct qeth_ipaddr *todo, *addr; | ||
430 | unsigned long flags; | ||
431 | int rc; | ||
432 | |||
433 | QETH_DBF_TEXT(TRACE, 2, "sdiplist"); | ||
434 | QETH_DBF_HEX(TRACE, 2, &card, sizeof(void *)); | ||
435 | |||
436 | spin_lock_irqsave(&card->ip_lock, flags); | ||
437 | tbd_list = card->ip_tbd_list; | ||
438 | card->ip_tbd_list = kmalloc(sizeof(struct list_head), GFP_ATOMIC); | ||
439 | if (!card->ip_tbd_list) { | ||
440 | QETH_DBF_TEXT(TRACE, 0, "silnomem"); | ||
441 | card->ip_tbd_list = tbd_list; | ||
442 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
443 | return; | ||
444 | } else | ||
445 | INIT_LIST_HEAD(card->ip_tbd_list); | ||
446 | |||
447 | while (!list_empty(tbd_list)) { | ||
448 | todo = list_entry(tbd_list->next, struct qeth_ipaddr, entry); | ||
449 | list_del(&todo->entry); | ||
450 | if (todo->type == QETH_IP_TYPE_DEL_ALL_MC) { | ||
451 | __qeth_l3_delete_all_mc(card, &flags); | ||
452 | kfree(todo); | ||
453 | continue; | ||
454 | } | ||
455 | rc = __qeth_l3_ref_ip_on_card(card, todo, &addr); | ||
456 | if (rc == 0) { | ||
457 | /* nothing to be done; only adjusted refcount */ | ||
458 | kfree(todo); | ||
459 | } else if (rc == 1) { | ||
460 | /* new entry to be added to on-card list */ | ||
461 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
462 | rc = qeth_l3_register_addr_entry(card, todo); | ||
463 | spin_lock_irqsave(&card->ip_lock, flags); | ||
464 | if (!rc || (rc == IPA_RC_LAN_OFFLINE)) | ||
465 | list_add_tail(&todo->entry, &card->ip_list); | ||
466 | else | ||
467 | kfree(todo); | ||
468 | } else if (rc == -1) { | ||
469 | /* on-card entry to be removed */ | ||
470 | list_del_init(&addr->entry); | ||
471 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
472 | rc = qeth_l3_deregister_addr_entry(card, addr); | ||
473 | spin_lock_irqsave(&card->ip_lock, flags); | ||
474 | if (!rc || (rc == IPA_RC_PRIMARY_ALREADY_DEFINED)) | ||
475 | kfree(addr); | ||
476 | else | ||
477 | list_add_tail(&addr->entry, &card->ip_list); | ||
478 | kfree(todo); | ||
479 | } | ||
480 | } | ||
481 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
482 | kfree(tbd_list); | ||
483 | } | ||
484 | |||
485 | static void qeth_l3_clear_ip_list(struct qeth_card *card, int clean, | ||
486 | int recover) | ||
487 | { | ||
488 | struct qeth_ipaddr *addr, *tmp; | ||
489 | unsigned long flags; | ||
490 | |||
491 | QETH_DBF_TEXT(TRACE, 4, "clearip"); | ||
492 | spin_lock_irqsave(&card->ip_lock, flags); | ||
493 | /* clear todo list */ | ||
494 | list_for_each_entry_safe(addr, tmp, card->ip_tbd_list, entry) { | ||
495 | list_del(&addr->entry); | ||
496 | kfree(addr); | ||
497 | } | ||
498 | |||
499 | while (!list_empty(&card->ip_list)) { | ||
500 | addr = list_entry(card->ip_list.next, | ||
501 | struct qeth_ipaddr, entry); | ||
502 | list_del_init(&addr->entry); | ||
503 | if (clean) { | ||
504 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
505 | qeth_l3_deregister_addr_entry(card, addr); | ||
506 | spin_lock_irqsave(&card->ip_lock, flags); | ||
507 | } | ||
508 | if (!recover || addr->is_multicast) { | ||
509 | kfree(addr); | ||
510 | continue; | ||
511 | } | ||
512 | list_add_tail(&addr->entry, card->ip_tbd_list); | ||
513 | } | ||
514 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
515 | } | ||
516 | |||
517 | static int qeth_l3_address_exists_in_list(struct list_head *list, | ||
518 | struct qeth_ipaddr *addr, int same_type) | ||
519 | { | ||
520 | struct qeth_ipaddr *tmp; | ||
521 | |||
522 | list_for_each_entry(tmp, list, entry) { | ||
523 | if ((tmp->proto == QETH_PROT_IPV4) && | ||
524 | (addr->proto == QETH_PROT_IPV4) && | ||
525 | ((same_type && (tmp->type == addr->type)) || | ||
526 | (!same_type && (tmp->type != addr->type))) && | ||
527 | (tmp->u.a4.addr == addr->u.a4.addr)) | ||
528 | return 1; | ||
529 | |||
530 | if ((tmp->proto == QETH_PROT_IPV6) && | ||
531 | (addr->proto == QETH_PROT_IPV6) && | ||
532 | ((same_type && (tmp->type == addr->type)) || | ||
533 | (!same_type && (tmp->type != addr->type))) && | ||
534 | (memcmp(&tmp->u.a6.addr, &addr->u.a6.addr, | ||
535 | sizeof(struct in6_addr)) == 0)) | ||
536 | return 1; | ||
537 | |||
538 | } | ||
539 | return 0; | ||
540 | } | ||
541 | |||
542 | static int qeth_l3_send_setdelmc(struct qeth_card *card, | ||
543 | struct qeth_ipaddr *addr, int ipacmd) | ||
544 | { | ||
545 | int rc; | ||
546 | struct qeth_cmd_buffer *iob; | ||
547 | struct qeth_ipa_cmd *cmd; | ||
548 | |||
549 | QETH_DBF_TEXT(TRACE, 4, "setdelmc"); | ||
550 | |||
551 | iob = qeth_get_ipacmd_buffer(card, ipacmd, addr->proto); | ||
552 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
553 | memcpy(&cmd->data.setdelipm.mac, addr->mac, OSA_ADDR_LEN); | ||
554 | if (addr->proto == QETH_PROT_IPV6) | ||
555 | memcpy(cmd->data.setdelipm.ip6, &addr->u.a6.addr, | ||
556 | sizeof(struct in6_addr)); | ||
557 | else | ||
558 | memcpy(&cmd->data.setdelipm.ip4, &addr->u.a4.addr, 4); | ||
559 | |||
560 | rc = qeth_send_ipa_cmd(card, iob, NULL, NULL); | ||
561 | |||
562 | return rc; | ||
563 | } | ||
564 | |||
565 | static void qeth_l3_fill_netmask(u8 *netmask, unsigned int len) | ||
566 | { | ||
567 | int i, j; | ||
568 | for (i = 0; i < 16; i++) { | ||
569 | j = (len) - (i * 8); | ||
570 | if (j >= 8) | ||
571 | netmask[i] = 0xff; | ||
572 | else if (j > 0) | ||
573 | netmask[i] = (u8)(0xFF00 >> j); | ||
574 | else | ||
575 | netmask[i] = 0; | ||
576 | } | ||
577 | } | ||
578 | |||
579 | static int qeth_l3_send_setdelip(struct qeth_card *card, | ||
580 | struct qeth_ipaddr *addr, int ipacmd, unsigned int flags) | ||
581 | { | ||
582 | int rc; | ||
583 | struct qeth_cmd_buffer *iob; | ||
584 | struct qeth_ipa_cmd *cmd; | ||
585 | __u8 netmask[16]; | ||
586 | |||
587 | QETH_DBF_TEXT(TRACE, 4, "setdelip"); | ||
588 | QETH_DBF_TEXT_(TRACE, 4, "flags%02X", flags); | ||
589 | |||
590 | iob = qeth_get_ipacmd_buffer(card, ipacmd, addr->proto); | ||
591 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
592 | if (addr->proto == QETH_PROT_IPV6) { | ||
593 | memcpy(cmd->data.setdelip6.ip_addr, &addr->u.a6.addr, | ||
594 | sizeof(struct in6_addr)); | ||
595 | qeth_l3_fill_netmask(netmask, addr->u.a6.pfxlen); | ||
596 | memcpy(cmd->data.setdelip6.mask, netmask, | ||
597 | sizeof(struct in6_addr)); | ||
598 | cmd->data.setdelip6.flags = flags; | ||
599 | } else { | ||
600 | memcpy(cmd->data.setdelip4.ip_addr, &addr->u.a4.addr, 4); | ||
601 | memcpy(cmd->data.setdelip4.mask, &addr->u.a4.mask, 4); | ||
602 | cmd->data.setdelip4.flags = flags; | ||
603 | } | ||
604 | |||
605 | rc = qeth_send_ipa_cmd(card, iob, NULL, NULL); | ||
606 | |||
607 | return rc; | ||
608 | } | ||
609 | |||
610 | static int qeth_l3_send_setrouting(struct qeth_card *card, | ||
611 | enum qeth_routing_types type, enum qeth_prot_versions prot) | ||
612 | { | ||
613 | int rc; | ||
614 | struct qeth_ipa_cmd *cmd; | ||
615 | struct qeth_cmd_buffer *iob; | ||
616 | |||
617 | QETH_DBF_TEXT(TRACE, 4, "setroutg"); | ||
618 | iob = qeth_get_ipacmd_buffer(card, IPA_CMD_SETRTG, prot); | ||
619 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
620 | cmd->data.setrtg.type = (type); | ||
621 | rc = qeth_send_ipa_cmd(card, iob, NULL, NULL); | ||
622 | |||
623 | return rc; | ||
624 | } | ||
625 | |||
626 | static void qeth_l3_correct_routing_type(struct qeth_card *card, | ||
627 | enum qeth_routing_types *type, enum qeth_prot_versions prot) | ||
628 | { | ||
629 | if (card->info.type == QETH_CARD_TYPE_IQD) { | ||
630 | switch (*type) { | ||
631 | case NO_ROUTER: | ||
632 | case PRIMARY_CONNECTOR: | ||
633 | case SECONDARY_CONNECTOR: | ||
634 | case MULTICAST_ROUTER: | ||
635 | return; | ||
636 | default: | ||
637 | goto out_inval; | ||
638 | } | ||
639 | } else { | ||
640 | switch (*type) { | ||
641 | case NO_ROUTER: | ||
642 | case PRIMARY_ROUTER: | ||
643 | case SECONDARY_ROUTER: | ||
644 | return; | ||
645 | case MULTICAST_ROUTER: | ||
646 | if (qeth_is_ipafunc_supported(card, prot, | ||
647 | IPA_OSA_MC_ROUTER)) | ||
648 | return; | ||
649 | default: | ||
650 | goto out_inval; | ||
651 | } | ||
652 | } | ||
653 | out_inval: | ||
654 | PRINT_WARN("Routing type '%s' not supported for interface %s.\n" | ||
655 | "Router status set to 'no router'.\n", | ||
656 | ((*type == PRIMARY_ROUTER)? "primary router" : | ||
657 | (*type == SECONDARY_ROUTER)? "secondary router" : | ||
658 | (*type == PRIMARY_CONNECTOR)? "primary connector" : | ||
659 | (*type == SECONDARY_CONNECTOR)? "secondary connector" : | ||
660 | (*type == MULTICAST_ROUTER)? "multicast router" : | ||
661 | "unknown"), | ||
662 | card->dev->name); | ||
663 | *type = NO_ROUTER; | ||
664 | } | ||
665 | |||
666 | int qeth_l3_setrouting_v4(struct qeth_card *card) | ||
667 | { | ||
668 | int rc; | ||
669 | |||
670 | QETH_DBF_TEXT(TRACE, 3, "setrtg4"); | ||
671 | |||
672 | qeth_l3_correct_routing_type(card, &card->options.route4.type, | ||
673 | QETH_PROT_IPV4); | ||
674 | |||
675 | rc = qeth_l3_send_setrouting(card, card->options.route4.type, | ||
676 | QETH_PROT_IPV4); | ||
677 | if (rc) { | ||
678 | card->options.route4.type = NO_ROUTER; | ||
679 | PRINT_WARN("Error (0x%04x) while setting routing type on %s. " | ||
680 | "Type set to 'no router'.\n", | ||
681 | rc, QETH_CARD_IFNAME(card)); | ||
682 | } | ||
683 | return rc; | ||
684 | } | ||
685 | |||
686 | int qeth_l3_setrouting_v6(struct qeth_card *card) | ||
687 | { | ||
688 | int rc = 0; | ||
689 | |||
690 | QETH_DBF_TEXT(TRACE, 3, "setrtg6"); | ||
691 | #ifdef CONFIG_QETH_IPV6 | ||
692 | |||
693 | if (!qeth_is_supported(card, IPA_IPV6)) | ||
694 | return 0; | ||
695 | qeth_l3_correct_routing_type(card, &card->options.route6.type, | ||
696 | QETH_PROT_IPV6); | ||
697 | |||
698 | rc = qeth_l3_send_setrouting(card, card->options.route6.type, | ||
699 | QETH_PROT_IPV6); | ||
700 | if (rc) { | ||
701 | card->options.route6.type = NO_ROUTER; | ||
702 | PRINT_WARN("Error (0x%04x) while setting routing type on %s. " | ||
703 | "Type set to 'no router'.\n", | ||
704 | rc, QETH_CARD_IFNAME(card)); | ||
705 | } | ||
706 | #endif | ||
707 | return rc; | ||
708 | } | ||
709 | |||
710 | /* | ||
711 | * IP address takeover related functions | ||
712 | */ | ||
713 | static void qeth_l3_clear_ipato_list(struct qeth_card *card) | ||
714 | { | ||
715 | |||
716 | struct qeth_ipato_entry *ipatoe, *tmp; | ||
717 | unsigned long flags; | ||
718 | |||
719 | spin_lock_irqsave(&card->ip_lock, flags); | ||
720 | list_for_each_entry_safe(ipatoe, tmp, &card->ipato.entries, entry) { | ||
721 | list_del(&ipatoe->entry); | ||
722 | kfree(ipatoe); | ||
723 | } | ||
724 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
725 | } | ||
726 | |||
727 | int qeth_l3_add_ipato_entry(struct qeth_card *card, | ||
728 | struct qeth_ipato_entry *new) | ||
729 | { | ||
730 | struct qeth_ipato_entry *ipatoe; | ||
731 | unsigned long flags; | ||
732 | int rc = 0; | ||
733 | |||
734 | QETH_DBF_TEXT(TRACE, 2, "addipato"); | ||
735 | spin_lock_irqsave(&card->ip_lock, flags); | ||
736 | list_for_each_entry(ipatoe, &card->ipato.entries, entry) { | ||
737 | if (ipatoe->proto != new->proto) | ||
738 | continue; | ||
739 | if (!memcmp(ipatoe->addr, new->addr, | ||
740 | (ipatoe->proto == QETH_PROT_IPV4)? 4:16) && | ||
741 | (ipatoe->mask_bits == new->mask_bits)) { | ||
742 | PRINT_WARN("ipato entry already exists!\n"); | ||
743 | rc = -EEXIST; | ||
744 | break; | ||
745 | } | ||
746 | } | ||
747 | if (!rc) | ||
748 | list_add_tail(&new->entry, &card->ipato.entries); | ||
749 | |||
750 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
751 | return rc; | ||
752 | } | ||
753 | |||
754 | void qeth_l3_del_ipato_entry(struct qeth_card *card, | ||
755 | enum qeth_prot_versions proto, u8 *addr, int mask_bits) | ||
756 | { | ||
757 | struct qeth_ipato_entry *ipatoe, *tmp; | ||
758 | unsigned long flags; | ||
759 | |||
760 | QETH_DBF_TEXT(TRACE, 2, "delipato"); | ||
761 | spin_lock_irqsave(&card->ip_lock, flags); | ||
762 | list_for_each_entry_safe(ipatoe, tmp, &card->ipato.entries, entry) { | ||
763 | if (ipatoe->proto != proto) | ||
764 | continue; | ||
765 | if (!memcmp(ipatoe->addr, addr, | ||
766 | (proto == QETH_PROT_IPV4)? 4:16) && | ||
767 | (ipatoe->mask_bits == mask_bits)) { | ||
768 | list_del(&ipatoe->entry); | ||
769 | kfree(ipatoe); | ||
770 | } | ||
771 | } | ||
772 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
773 | } | ||
774 | |||
775 | /* | ||
776 | * VIPA related functions | ||
777 | */ | ||
778 | int qeth_l3_add_vipa(struct qeth_card *card, enum qeth_prot_versions proto, | ||
779 | const u8 *addr) | ||
780 | { | ||
781 | struct qeth_ipaddr *ipaddr; | ||
782 | unsigned long flags; | ||
783 | int rc = 0; | ||
784 | |||
785 | ipaddr = qeth_l3_get_addr_buffer(proto); | ||
786 | if (ipaddr) { | ||
787 | if (proto == QETH_PROT_IPV4) { | ||
788 | QETH_DBF_TEXT(TRACE, 2, "addvipa4"); | ||
789 | memcpy(&ipaddr->u.a4.addr, addr, 4); | ||
790 | ipaddr->u.a4.mask = 0; | ||
791 | } else if (proto == QETH_PROT_IPV6) { | ||
792 | QETH_DBF_TEXT(TRACE, 2, "addvipa6"); | ||
793 | memcpy(&ipaddr->u.a6.addr, addr, 16); | ||
794 | ipaddr->u.a6.pfxlen = 0; | ||
795 | } | ||
796 | ipaddr->type = QETH_IP_TYPE_VIPA; | ||
797 | ipaddr->set_flags = QETH_IPA_SETIP_VIPA_FLAG; | ||
798 | ipaddr->del_flags = QETH_IPA_DELIP_VIPA_FLAG; | ||
799 | } else | ||
800 | return -ENOMEM; | ||
801 | spin_lock_irqsave(&card->ip_lock, flags); | ||
802 | if (qeth_l3_address_exists_in_list(&card->ip_list, ipaddr, 0) || | ||
803 | qeth_l3_address_exists_in_list(card->ip_tbd_list, ipaddr, 0)) | ||
804 | rc = -EEXIST; | ||
805 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
806 | if (rc) { | ||
807 | PRINT_WARN("Cannot add VIPA. Address already exists!\n"); | ||
808 | return rc; | ||
809 | } | ||
810 | if (!qeth_l3_add_ip(card, ipaddr)) | ||
811 | kfree(ipaddr); | ||
812 | qeth_l3_set_ip_addr_list(card); | ||
813 | return rc; | ||
814 | } | ||
815 | |||
816 | void qeth_l3_del_vipa(struct qeth_card *card, enum qeth_prot_versions proto, | ||
817 | const u8 *addr) | ||
818 | { | ||
819 | struct qeth_ipaddr *ipaddr; | ||
820 | |||
821 | ipaddr = qeth_l3_get_addr_buffer(proto); | ||
822 | if (ipaddr) { | ||
823 | if (proto == QETH_PROT_IPV4) { | ||
824 | QETH_DBF_TEXT(TRACE, 2, "delvipa4"); | ||
825 | memcpy(&ipaddr->u.a4.addr, addr, 4); | ||
826 | ipaddr->u.a4.mask = 0; | ||
827 | } else if (proto == QETH_PROT_IPV6) { | ||
828 | QETH_DBF_TEXT(TRACE, 2, "delvipa6"); | ||
829 | memcpy(&ipaddr->u.a6.addr, addr, 16); | ||
830 | ipaddr->u.a6.pfxlen = 0; | ||
831 | } | ||
832 | ipaddr->type = QETH_IP_TYPE_VIPA; | ||
833 | } else | ||
834 | return; | ||
835 | if (!qeth_l3_delete_ip(card, ipaddr)) | ||
836 | kfree(ipaddr); | ||
837 | qeth_l3_set_ip_addr_list(card); | ||
838 | } | ||
839 | |||
840 | /* | ||
841 | * proxy ARP related functions | ||
842 | */ | ||
843 | int qeth_l3_add_rxip(struct qeth_card *card, enum qeth_prot_versions proto, | ||
844 | const u8 *addr) | ||
845 | { | ||
846 | struct qeth_ipaddr *ipaddr; | ||
847 | unsigned long flags; | ||
848 | int rc = 0; | ||
849 | |||
850 | ipaddr = qeth_l3_get_addr_buffer(proto); | ||
851 | if (ipaddr) { | ||
852 | if (proto == QETH_PROT_IPV4) { | ||
853 | QETH_DBF_TEXT(TRACE, 2, "addrxip4"); | ||
854 | memcpy(&ipaddr->u.a4.addr, addr, 4); | ||
855 | ipaddr->u.a4.mask = 0; | ||
856 | } else if (proto == QETH_PROT_IPV6) { | ||
857 | QETH_DBF_TEXT(TRACE, 2, "addrxip6"); | ||
858 | memcpy(&ipaddr->u.a6.addr, addr, 16); | ||
859 | ipaddr->u.a6.pfxlen = 0; | ||
860 | } | ||
861 | ipaddr->type = QETH_IP_TYPE_RXIP; | ||
862 | ipaddr->set_flags = QETH_IPA_SETIP_TAKEOVER_FLAG; | ||
863 | ipaddr->del_flags = 0; | ||
864 | } else | ||
865 | return -ENOMEM; | ||
866 | spin_lock_irqsave(&card->ip_lock, flags); | ||
867 | if (qeth_l3_address_exists_in_list(&card->ip_list, ipaddr, 0) || | ||
868 | qeth_l3_address_exists_in_list(card->ip_tbd_list, ipaddr, 0)) | ||
869 | rc = -EEXIST; | ||
870 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
871 | if (rc) { | ||
872 | PRINT_WARN("Cannot add RXIP. Address already exists!\n"); | ||
873 | return rc; | ||
874 | } | ||
875 | if (!qeth_l3_add_ip(card, ipaddr)) | ||
876 | kfree(ipaddr); | ||
877 | qeth_l3_set_ip_addr_list(card); | ||
878 | return 0; | ||
879 | } | ||
880 | |||
881 | void qeth_l3_del_rxip(struct qeth_card *card, enum qeth_prot_versions proto, | ||
882 | const u8 *addr) | ||
883 | { | ||
884 | struct qeth_ipaddr *ipaddr; | ||
885 | |||
886 | ipaddr = qeth_l3_get_addr_buffer(proto); | ||
887 | if (ipaddr) { | ||
888 | if (proto == QETH_PROT_IPV4) { | ||
889 | QETH_DBF_TEXT(TRACE, 2, "addrxip4"); | ||
890 | memcpy(&ipaddr->u.a4.addr, addr, 4); | ||
891 | ipaddr->u.a4.mask = 0; | ||
892 | } else if (proto == QETH_PROT_IPV6) { | ||
893 | QETH_DBF_TEXT(TRACE, 2, "addrxip6"); | ||
894 | memcpy(&ipaddr->u.a6.addr, addr, 16); | ||
895 | ipaddr->u.a6.pfxlen = 0; | ||
896 | } | ||
897 | ipaddr->type = QETH_IP_TYPE_RXIP; | ||
898 | } else | ||
899 | return; | ||
900 | if (!qeth_l3_delete_ip(card, ipaddr)) | ||
901 | kfree(ipaddr); | ||
902 | qeth_l3_set_ip_addr_list(card); | ||
903 | } | ||
904 | |||
905 | static int qeth_l3_register_addr_entry(struct qeth_card *card, | ||
906 | struct qeth_ipaddr *addr) | ||
907 | { | ||
908 | char buf[50]; | ||
909 | int rc = 0; | ||
910 | int cnt = 3; | ||
911 | |||
912 | if (addr->proto == QETH_PROT_IPV4) { | ||
913 | QETH_DBF_TEXT(TRACE, 2, "setaddr4"); | ||
914 | QETH_DBF_HEX(TRACE, 3, &addr->u.a4.addr, sizeof(int)); | ||
915 | } else if (addr->proto == QETH_PROT_IPV6) { | ||
916 | QETH_DBF_TEXT(TRACE, 2, "setaddr6"); | ||
917 | QETH_DBF_HEX(TRACE, 3, &addr->u.a6.addr, 8); | ||
918 | QETH_DBF_HEX(TRACE, 3, ((char *)&addr->u.a6.addr) + 8, 8); | ||
919 | } else { | ||
920 | QETH_DBF_TEXT(TRACE, 2, "setaddr?"); | ||
921 | QETH_DBF_HEX(TRACE, 3, addr, sizeof(struct qeth_ipaddr)); | ||
922 | } | ||
923 | do { | ||
924 | if (addr->is_multicast) | ||
925 | rc = qeth_l3_send_setdelmc(card, addr, IPA_CMD_SETIPM); | ||
926 | else | ||
927 | rc = qeth_l3_send_setdelip(card, addr, IPA_CMD_SETIP, | ||
928 | addr->set_flags); | ||
929 | if (rc) | ||
930 | QETH_DBF_TEXT(TRACE, 2, "failed"); | ||
931 | } while ((--cnt > 0) && rc); | ||
932 | if (rc) { | ||
933 | QETH_DBF_TEXT(TRACE, 2, "FAILED"); | ||
934 | qeth_l3_ipaddr_to_string(addr->proto, (u8 *)&addr->u, buf); | ||
935 | PRINT_WARN("Could not register IP address %s (rc=0x%x/%d)\n", | ||
936 | buf, rc, rc); | ||
937 | } | ||
938 | return rc; | ||
939 | } | ||
940 | |||
941 | static int qeth_l3_deregister_addr_entry(struct qeth_card *card, | ||
942 | struct qeth_ipaddr *addr) | ||
943 | { | ||
944 | int rc = 0; | ||
945 | |||
946 | if (addr->proto == QETH_PROT_IPV4) { | ||
947 | QETH_DBF_TEXT(TRACE, 2, "deladdr4"); | ||
948 | QETH_DBF_HEX(TRACE, 3, &addr->u.a4.addr, sizeof(int)); | ||
949 | } else if (addr->proto == QETH_PROT_IPV6) { | ||
950 | QETH_DBF_TEXT(TRACE, 2, "deladdr6"); | ||
951 | QETH_DBF_HEX(TRACE, 3, &addr->u.a6.addr, 8); | ||
952 | QETH_DBF_HEX(TRACE, 3, ((char *)&addr->u.a6.addr) + 8, 8); | ||
953 | } else { | ||
954 | QETH_DBF_TEXT(TRACE, 2, "deladdr?"); | ||
955 | QETH_DBF_HEX(TRACE, 3, addr, sizeof(struct qeth_ipaddr)); | ||
956 | } | ||
957 | if (addr->is_multicast) | ||
958 | rc = qeth_l3_send_setdelmc(card, addr, IPA_CMD_DELIPM); | ||
959 | else | ||
960 | rc = qeth_l3_send_setdelip(card, addr, IPA_CMD_DELIP, | ||
961 | addr->del_flags); | ||
962 | if (rc) { | ||
963 | QETH_DBF_TEXT(TRACE, 2, "failed"); | ||
964 | /* TODO: re-activate this warning as soon as we have a | ||
965 | * clean mirco code | ||
966 | qeth_ipaddr_to_string(addr->proto, (u8 *)&addr->u, buf); | ||
967 | PRINT_WARN("Could not deregister IP address %s (rc=%x)\n", | ||
968 | buf, rc); | ||
969 | */ | ||
970 | } | ||
971 | |||
972 | return rc; | ||
973 | } | ||
974 | |||
975 | static inline u8 qeth_l3_get_qeth_hdr_flags4(int cast_type) | ||
976 | { | ||
977 | if (cast_type == RTN_MULTICAST) | ||
978 | return QETH_CAST_MULTICAST; | ||
979 | if (cast_type == RTN_BROADCAST) | ||
980 | return QETH_CAST_BROADCAST; | ||
981 | return QETH_CAST_UNICAST; | ||
982 | } | ||
983 | |||
984 | static inline u8 qeth_l3_get_qeth_hdr_flags6(int cast_type) | ||
985 | { | ||
986 | u8 ct = QETH_HDR_PASSTHRU | QETH_HDR_IPV6; | ||
987 | if (cast_type == RTN_MULTICAST) | ||
988 | return ct | QETH_CAST_MULTICAST; | ||
989 | if (cast_type == RTN_ANYCAST) | ||
990 | return ct | QETH_CAST_ANYCAST; | ||
991 | if (cast_type == RTN_BROADCAST) | ||
992 | return ct | QETH_CAST_BROADCAST; | ||
993 | return ct | QETH_CAST_UNICAST; | ||
994 | } | ||
995 | |||
996 | static int qeth_l3_send_setadp_mode(struct qeth_card *card, __u32 command, | ||
997 | __u32 mode) | ||
998 | { | ||
999 | int rc; | ||
1000 | struct qeth_cmd_buffer *iob; | ||
1001 | struct qeth_ipa_cmd *cmd; | ||
1002 | |||
1003 | QETH_DBF_TEXT(TRACE, 4, "adpmode"); | ||
1004 | |||
1005 | iob = qeth_get_adapter_cmd(card, command, | ||
1006 | sizeof(struct qeth_ipacmd_setadpparms)); | ||
1007 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
1008 | cmd->data.setadapterparms.data.mode = mode; | ||
1009 | rc = qeth_send_ipa_cmd(card, iob, qeth_default_setadapterparms_cb, | ||
1010 | NULL); | ||
1011 | return rc; | ||
1012 | } | ||
1013 | |||
1014 | static int qeth_l3_setadapter_hstr(struct qeth_card *card) | ||
1015 | { | ||
1016 | int rc; | ||
1017 | |||
1018 | QETH_DBF_TEXT(TRACE, 4, "adphstr"); | ||
1019 | |||
1020 | if (qeth_adp_supported(card, IPA_SETADP_SET_BROADCAST_MODE)) { | ||
1021 | rc = qeth_l3_send_setadp_mode(card, | ||
1022 | IPA_SETADP_SET_BROADCAST_MODE, | ||
1023 | card->options.broadcast_mode); | ||
1024 | if (rc) | ||
1025 | PRINT_WARN("couldn't set broadcast mode on " | ||
1026 | "device %s: x%x\n", | ||
1027 | CARD_BUS_ID(card), rc); | ||
1028 | rc = qeth_l3_send_setadp_mode(card, | ||
1029 | IPA_SETADP_ALTER_MAC_ADDRESS, | ||
1030 | card->options.macaddr_mode); | ||
1031 | if (rc) | ||
1032 | PRINT_WARN("couldn't set macaddr mode on " | ||
1033 | "device %s: x%x\n", CARD_BUS_ID(card), rc); | ||
1034 | return rc; | ||
1035 | } | ||
1036 | if (card->options.broadcast_mode == QETH_TR_BROADCAST_LOCAL) | ||
1037 | PRINT_WARN("set adapter parameters not available " | ||
1038 | "to set broadcast mode, using ALLRINGS " | ||
1039 | "on device %s:\n", CARD_BUS_ID(card)); | ||
1040 | if (card->options.macaddr_mode == QETH_TR_MACADDR_CANONICAL) | ||
1041 | PRINT_WARN("set adapter parameters not available " | ||
1042 | "to set macaddr mode, using NONCANONICAL " | ||
1043 | "on device %s:\n", CARD_BUS_ID(card)); | ||
1044 | return 0; | ||
1045 | } | ||
1046 | |||
1047 | static int qeth_l3_setadapter_parms(struct qeth_card *card) | ||
1048 | { | ||
1049 | int rc; | ||
1050 | |||
1051 | QETH_DBF_TEXT(SETUP, 2, "setadprm"); | ||
1052 | |||
1053 | if (!qeth_is_supported(card, IPA_SETADAPTERPARMS)) { | ||
1054 | PRINT_WARN("set adapter parameters not supported " | ||
1055 | "on device %s.\n", | ||
1056 | CARD_BUS_ID(card)); | ||
1057 | QETH_DBF_TEXT(SETUP, 2, " notsupp"); | ||
1058 | return 0; | ||
1059 | } | ||
1060 | rc = qeth_query_setadapterparms(card); | ||
1061 | if (rc) { | ||
1062 | PRINT_WARN("couldn't set adapter parameters on device %s: " | ||
1063 | "x%x\n", CARD_BUS_ID(card), rc); | ||
1064 | return rc; | ||
1065 | } | ||
1066 | if (qeth_adp_supported(card, IPA_SETADP_ALTER_MAC_ADDRESS)) { | ||
1067 | rc = qeth_setadpparms_change_macaddr(card); | ||
1068 | if (rc) | ||
1069 | PRINT_WARN("couldn't get MAC address on " | ||
1070 | "device %s: x%x\n", | ||
1071 | CARD_BUS_ID(card), rc); | ||
1072 | } | ||
1073 | |||
1074 | if ((card->info.link_type == QETH_LINK_TYPE_HSTR) || | ||
1075 | (card->info.link_type == QETH_LINK_TYPE_LANE_TR)) | ||
1076 | rc = qeth_l3_setadapter_hstr(card); | ||
1077 | |||
1078 | return rc; | ||
1079 | } | ||
1080 | |||
1081 | static int qeth_l3_default_setassparms_cb(struct qeth_card *card, | ||
1082 | struct qeth_reply *reply, unsigned long data) | ||
1083 | { | ||
1084 | struct qeth_ipa_cmd *cmd; | ||
1085 | |||
1086 | QETH_DBF_TEXT(TRACE, 4, "defadpcb"); | ||
1087 | |||
1088 | cmd = (struct qeth_ipa_cmd *) data; | ||
1089 | if (cmd->hdr.return_code == 0) { | ||
1090 | cmd->hdr.return_code = cmd->data.setassparms.hdr.return_code; | ||
1091 | if (cmd->hdr.prot_version == QETH_PROT_IPV4) | ||
1092 | card->options.ipa4.enabled_funcs = cmd->hdr.ipa_enabled; | ||
1093 | if (cmd->hdr.prot_version == QETH_PROT_IPV6) | ||
1094 | card->options.ipa6.enabled_funcs = cmd->hdr.ipa_enabled; | ||
1095 | } | ||
1096 | if (cmd->data.setassparms.hdr.assist_no == IPA_INBOUND_CHECKSUM && | ||
1097 | cmd->data.setassparms.hdr.command_code == IPA_CMD_ASS_START) { | ||
1098 | card->info.csum_mask = cmd->data.setassparms.data.flags_32bit; | ||
1099 | QETH_DBF_TEXT_(TRACE, 3, "csum:%d", card->info.csum_mask); | ||
1100 | } | ||
1101 | return 0; | ||
1102 | } | ||
1103 | |||
1104 | static struct qeth_cmd_buffer *qeth_l3_get_setassparms_cmd( | ||
1105 | struct qeth_card *card, enum qeth_ipa_funcs ipa_func, __u16 cmd_code, | ||
1106 | __u16 len, enum qeth_prot_versions prot) | ||
1107 | { | ||
1108 | struct qeth_cmd_buffer *iob; | ||
1109 | struct qeth_ipa_cmd *cmd; | ||
1110 | |||
1111 | QETH_DBF_TEXT(TRACE, 4, "getasscm"); | ||
1112 | iob = qeth_get_ipacmd_buffer(card, IPA_CMD_SETASSPARMS, prot); | ||
1113 | |||
1114 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
1115 | cmd->data.setassparms.hdr.assist_no = ipa_func; | ||
1116 | cmd->data.setassparms.hdr.length = 8 + len; | ||
1117 | cmd->data.setassparms.hdr.command_code = cmd_code; | ||
1118 | cmd->data.setassparms.hdr.return_code = 0; | ||
1119 | cmd->data.setassparms.hdr.seq_no = 0; | ||
1120 | |||
1121 | return iob; | ||
1122 | } | ||
1123 | |||
1124 | static int qeth_l3_send_setassparms(struct qeth_card *card, | ||
1125 | struct qeth_cmd_buffer *iob, __u16 len, long data, | ||
1126 | int (*reply_cb)(struct qeth_card *, struct qeth_reply *, | ||
1127 | unsigned long), | ||
1128 | void *reply_param) | ||
1129 | { | ||
1130 | int rc; | ||
1131 | struct qeth_ipa_cmd *cmd; | ||
1132 | |||
1133 | QETH_DBF_TEXT(TRACE, 4, "sendassp"); | ||
1134 | |||
1135 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
1136 | if (len <= sizeof(__u32)) | ||
1137 | cmd->data.setassparms.data.flags_32bit = (__u32) data; | ||
1138 | else /* (len > sizeof(__u32)) */ | ||
1139 | memcpy(&cmd->data.setassparms.data, (void *) data, len); | ||
1140 | |||
1141 | rc = qeth_send_ipa_cmd(card, iob, reply_cb, reply_param); | ||
1142 | return rc; | ||
1143 | } | ||
1144 | |||
1145 | #ifdef CONFIG_QETH_IPV6 | ||
1146 | static int qeth_l3_send_simple_setassparms_ipv6(struct qeth_card *card, | ||
1147 | enum qeth_ipa_funcs ipa_func, __u16 cmd_code) | ||
1148 | { | ||
1149 | int rc; | ||
1150 | struct qeth_cmd_buffer *iob; | ||
1151 | |||
1152 | QETH_DBF_TEXT(TRACE, 4, "simassp6"); | ||
1153 | iob = qeth_l3_get_setassparms_cmd(card, ipa_func, cmd_code, | ||
1154 | 0, QETH_PROT_IPV6); | ||
1155 | rc = qeth_l3_send_setassparms(card, iob, 0, 0, | ||
1156 | qeth_l3_default_setassparms_cb, NULL); | ||
1157 | return rc; | ||
1158 | } | ||
1159 | #endif | ||
1160 | |||
1161 | static int qeth_l3_send_simple_setassparms(struct qeth_card *card, | ||
1162 | enum qeth_ipa_funcs ipa_func, __u16 cmd_code, long data) | ||
1163 | { | ||
1164 | int rc; | ||
1165 | int length = 0; | ||
1166 | struct qeth_cmd_buffer *iob; | ||
1167 | |||
1168 | QETH_DBF_TEXT(TRACE, 4, "simassp4"); | ||
1169 | if (data) | ||
1170 | length = sizeof(__u32); | ||
1171 | iob = qeth_l3_get_setassparms_cmd(card, ipa_func, cmd_code, | ||
1172 | length, QETH_PROT_IPV4); | ||
1173 | rc = qeth_l3_send_setassparms(card, iob, length, data, | ||
1174 | qeth_l3_default_setassparms_cb, NULL); | ||
1175 | return rc; | ||
1176 | } | ||
1177 | |||
1178 | static int qeth_l3_start_ipa_arp_processing(struct qeth_card *card) | ||
1179 | { | ||
1180 | int rc; | ||
1181 | |||
1182 | QETH_DBF_TEXT(TRACE, 3, "ipaarp"); | ||
1183 | |||
1184 | if (!qeth_is_supported(card, IPA_ARP_PROCESSING)) { | ||
1185 | PRINT_WARN("ARP processing not supported " | ||
1186 | "on %s!\n", QETH_CARD_IFNAME(card)); | ||
1187 | return 0; | ||
1188 | } | ||
1189 | rc = qeth_l3_send_simple_setassparms(card, IPA_ARP_PROCESSING, | ||
1190 | IPA_CMD_ASS_START, 0); | ||
1191 | if (rc) { | ||
1192 | PRINT_WARN("Could not start ARP processing " | ||
1193 | "assist on %s: 0x%x\n", | ||
1194 | QETH_CARD_IFNAME(card), rc); | ||
1195 | } | ||
1196 | return rc; | ||
1197 | } | ||
1198 | |||
1199 | static int qeth_l3_start_ipa_ip_fragmentation(struct qeth_card *card) | ||
1200 | { | ||
1201 | int rc; | ||
1202 | |||
1203 | QETH_DBF_TEXT(TRACE, 3, "ipaipfrg"); | ||
1204 | |||
1205 | if (!qeth_is_supported(card, IPA_IP_FRAGMENTATION)) { | ||
1206 | PRINT_INFO("Hardware IP fragmentation not supported on %s\n", | ||
1207 | QETH_CARD_IFNAME(card)); | ||
1208 | return -EOPNOTSUPP; | ||
1209 | } | ||
1210 | |||
1211 | rc = qeth_l3_send_simple_setassparms(card, IPA_IP_FRAGMENTATION, | ||
1212 | IPA_CMD_ASS_START, 0); | ||
1213 | if (rc) { | ||
1214 | PRINT_WARN("Could not start Hardware IP fragmentation " | ||
1215 | "assist on %s: 0x%x\n", | ||
1216 | QETH_CARD_IFNAME(card), rc); | ||
1217 | } else | ||
1218 | PRINT_INFO("Hardware IP fragmentation enabled \n"); | ||
1219 | return rc; | ||
1220 | } | ||
1221 | |||
1222 | static int qeth_l3_start_ipa_source_mac(struct qeth_card *card) | ||
1223 | { | ||
1224 | int rc; | ||
1225 | |||
1226 | QETH_DBF_TEXT(TRACE, 3, "stsrcmac"); | ||
1227 | |||
1228 | if (!card->options.fake_ll) | ||
1229 | return -EOPNOTSUPP; | ||
1230 | |||
1231 | if (!qeth_is_supported(card, IPA_SOURCE_MAC)) { | ||
1232 | PRINT_INFO("Inbound source address not " | ||
1233 | "supported on %s\n", QETH_CARD_IFNAME(card)); | ||
1234 | return -EOPNOTSUPP; | ||
1235 | } | ||
1236 | |||
1237 | rc = qeth_l3_send_simple_setassparms(card, IPA_SOURCE_MAC, | ||
1238 | IPA_CMD_ASS_START, 0); | ||
1239 | if (rc) | ||
1240 | PRINT_WARN("Could not start inbound source " | ||
1241 | "assist on %s: 0x%x\n", | ||
1242 | QETH_CARD_IFNAME(card), rc); | ||
1243 | return rc; | ||
1244 | } | ||
1245 | |||
1246 | static int qeth_l3_start_ipa_vlan(struct qeth_card *card) | ||
1247 | { | ||
1248 | int rc = 0; | ||
1249 | |||
1250 | QETH_DBF_TEXT(TRACE, 3, "strtvlan"); | ||
1251 | |||
1252 | if (!qeth_is_supported(card, IPA_FULL_VLAN)) { | ||
1253 | PRINT_WARN("VLAN not supported on %s\n", | ||
1254 | QETH_CARD_IFNAME(card)); | ||
1255 | return -EOPNOTSUPP; | ||
1256 | } | ||
1257 | |||
1258 | rc = qeth_l3_send_simple_setassparms(card, IPA_VLAN_PRIO, | ||
1259 | IPA_CMD_ASS_START, 0); | ||
1260 | if (rc) { | ||
1261 | PRINT_WARN("Could not start vlan " | ||
1262 | "assist on %s: 0x%x\n", | ||
1263 | QETH_CARD_IFNAME(card), rc); | ||
1264 | } else { | ||
1265 | PRINT_INFO("VLAN enabled \n"); | ||
1266 | } | ||
1267 | return rc; | ||
1268 | } | ||
1269 | |||
1270 | static int qeth_l3_start_ipa_multicast(struct qeth_card *card) | ||
1271 | { | ||
1272 | int rc; | ||
1273 | |||
1274 | QETH_DBF_TEXT(TRACE, 3, "stmcast"); | ||
1275 | |||
1276 | if (!qeth_is_supported(card, IPA_MULTICASTING)) { | ||
1277 | PRINT_WARN("Multicast not supported on %s\n", | ||
1278 | QETH_CARD_IFNAME(card)); | ||
1279 | return -EOPNOTSUPP; | ||
1280 | } | ||
1281 | |||
1282 | rc = qeth_l3_send_simple_setassparms(card, IPA_MULTICASTING, | ||
1283 | IPA_CMD_ASS_START, 0); | ||
1284 | if (rc) { | ||
1285 | PRINT_WARN("Could not start multicast " | ||
1286 | "assist on %s: rc=%i\n", | ||
1287 | QETH_CARD_IFNAME(card), rc); | ||
1288 | } else { | ||
1289 | PRINT_INFO("Multicast enabled\n"); | ||
1290 | card->dev->flags |= IFF_MULTICAST; | ||
1291 | } | ||
1292 | return rc; | ||
1293 | } | ||
1294 | |||
1295 | static int qeth_l3_query_ipassists_cb(struct qeth_card *card, | ||
1296 | struct qeth_reply *reply, unsigned long data) | ||
1297 | { | ||
1298 | struct qeth_ipa_cmd *cmd; | ||
1299 | |||
1300 | QETH_DBF_TEXT(SETUP, 2, "qipasscb"); | ||
1301 | |||
1302 | cmd = (struct qeth_ipa_cmd *) data; | ||
1303 | if (cmd->hdr.prot_version == QETH_PROT_IPV4) { | ||
1304 | card->options.ipa4.supported_funcs = cmd->hdr.ipa_supported; | ||
1305 | card->options.ipa4.enabled_funcs = cmd->hdr.ipa_enabled; | ||
1306 | } else { | ||
1307 | card->options.ipa6.supported_funcs = cmd->hdr.ipa_supported; | ||
1308 | card->options.ipa6.enabled_funcs = cmd->hdr.ipa_enabled; | ||
1309 | } | ||
1310 | QETH_DBF_TEXT(SETUP, 2, "suppenbl"); | ||
1311 | QETH_DBF_TEXT_(SETUP, 2, "%x", cmd->hdr.ipa_supported); | ||
1312 | QETH_DBF_TEXT_(SETUP, 2, "%x", cmd->hdr.ipa_enabled); | ||
1313 | return 0; | ||
1314 | } | ||
1315 | |||
1316 | static int qeth_l3_query_ipassists(struct qeth_card *card, | ||
1317 | enum qeth_prot_versions prot) | ||
1318 | { | ||
1319 | int rc; | ||
1320 | struct qeth_cmd_buffer *iob; | ||
1321 | |||
1322 | QETH_DBF_TEXT_(SETUP, 2, "qipassi%i", prot); | ||
1323 | iob = qeth_get_ipacmd_buffer(card, IPA_CMD_QIPASSIST, prot); | ||
1324 | rc = qeth_send_ipa_cmd(card, iob, qeth_l3_query_ipassists_cb, NULL); | ||
1325 | return rc; | ||
1326 | } | ||
1327 | |||
1328 | #ifdef CONFIG_QETH_IPV6 | ||
1329 | static int qeth_l3_softsetup_ipv6(struct qeth_card *card) | ||
1330 | { | ||
1331 | int rc; | ||
1332 | |||
1333 | QETH_DBF_TEXT(TRACE, 3, "softipv6"); | ||
1334 | |||
1335 | if (card->info.type == QETH_CARD_TYPE_IQD) | ||
1336 | goto out; | ||
1337 | |||
1338 | rc = qeth_l3_query_ipassists(card, QETH_PROT_IPV6); | ||
1339 | if (rc) { | ||
1340 | PRINT_ERR("IPv6 query ipassist failed on %s\n", | ||
1341 | QETH_CARD_IFNAME(card)); | ||
1342 | return rc; | ||
1343 | } | ||
1344 | rc = qeth_l3_send_simple_setassparms(card, IPA_IPV6, | ||
1345 | IPA_CMD_ASS_START, 3); | ||
1346 | if (rc) { | ||
1347 | PRINT_WARN("IPv6 start assist (version 4) failed " | ||
1348 | "on %s: 0x%x\n", | ||
1349 | QETH_CARD_IFNAME(card), rc); | ||
1350 | return rc; | ||
1351 | } | ||
1352 | rc = qeth_l3_send_simple_setassparms_ipv6(card, IPA_IPV6, | ||
1353 | IPA_CMD_ASS_START); | ||
1354 | if (rc) { | ||
1355 | PRINT_WARN("IPV6 start assist (version 6) failed " | ||
1356 | "on %s: 0x%x\n", | ||
1357 | QETH_CARD_IFNAME(card), rc); | ||
1358 | return rc; | ||
1359 | } | ||
1360 | rc = qeth_l3_send_simple_setassparms_ipv6(card, IPA_PASSTHRU, | ||
1361 | IPA_CMD_ASS_START); | ||
1362 | if (rc) { | ||
1363 | PRINT_WARN("Could not enable passthrough " | ||
1364 | "on %s: 0x%x\n", | ||
1365 | QETH_CARD_IFNAME(card), rc); | ||
1366 | return rc; | ||
1367 | } | ||
1368 | out: | ||
1369 | PRINT_INFO("IPV6 enabled \n"); | ||
1370 | return 0; | ||
1371 | } | ||
1372 | #endif | ||
1373 | |||
1374 | static int qeth_l3_start_ipa_ipv6(struct qeth_card *card) | ||
1375 | { | ||
1376 | int rc = 0; | ||
1377 | |||
1378 | QETH_DBF_TEXT(TRACE, 3, "strtipv6"); | ||
1379 | |||
1380 | if (!qeth_is_supported(card, IPA_IPV6)) { | ||
1381 | PRINT_WARN("IPv6 not supported on %s\n", | ||
1382 | QETH_CARD_IFNAME(card)); | ||
1383 | return 0; | ||
1384 | } | ||
1385 | #ifdef CONFIG_QETH_IPV6 | ||
1386 | rc = qeth_l3_softsetup_ipv6(card); | ||
1387 | #endif | ||
1388 | return rc ; | ||
1389 | } | ||
1390 | |||
1391 | static int qeth_l3_start_ipa_broadcast(struct qeth_card *card) | ||
1392 | { | ||
1393 | int rc; | ||
1394 | |||
1395 | QETH_DBF_TEXT(TRACE, 3, "stbrdcst"); | ||
1396 | card->info.broadcast_capable = 0; | ||
1397 | if (!qeth_is_supported(card, IPA_FILTERING)) { | ||
1398 | PRINT_WARN("Broadcast not supported on %s\n", | ||
1399 | QETH_CARD_IFNAME(card)); | ||
1400 | rc = -EOPNOTSUPP; | ||
1401 | goto out; | ||
1402 | } | ||
1403 | rc = qeth_l3_send_simple_setassparms(card, IPA_FILTERING, | ||
1404 | IPA_CMD_ASS_START, 0); | ||
1405 | if (rc) { | ||
1406 | PRINT_WARN("Could not enable broadcasting filtering " | ||
1407 | "on %s: 0x%x\n", | ||
1408 | QETH_CARD_IFNAME(card), rc); | ||
1409 | goto out; | ||
1410 | } | ||
1411 | |||
1412 | rc = qeth_l3_send_simple_setassparms(card, IPA_FILTERING, | ||
1413 | IPA_CMD_ASS_CONFIGURE, 1); | ||
1414 | if (rc) { | ||
1415 | PRINT_WARN("Could not set up broadcast filtering on %s: 0x%x\n", | ||
1416 | QETH_CARD_IFNAME(card), rc); | ||
1417 | goto out; | ||
1418 | } | ||
1419 | card->info.broadcast_capable = QETH_BROADCAST_WITH_ECHO; | ||
1420 | PRINT_INFO("Broadcast enabled \n"); | ||
1421 | rc = qeth_l3_send_simple_setassparms(card, IPA_FILTERING, | ||
1422 | IPA_CMD_ASS_ENABLE, 1); | ||
1423 | if (rc) { | ||
1424 | PRINT_WARN("Could not set up broadcast echo filtering on " | ||
1425 | "%s: 0x%x\n", QETH_CARD_IFNAME(card), rc); | ||
1426 | goto out; | ||
1427 | } | ||
1428 | card->info.broadcast_capable = QETH_BROADCAST_WITHOUT_ECHO; | ||
1429 | out: | ||
1430 | if (card->info.broadcast_capable) | ||
1431 | card->dev->flags |= IFF_BROADCAST; | ||
1432 | else | ||
1433 | card->dev->flags &= ~IFF_BROADCAST; | ||
1434 | return rc; | ||
1435 | } | ||
1436 | |||
1437 | static int qeth_l3_send_checksum_command(struct qeth_card *card) | ||
1438 | { | ||
1439 | int rc; | ||
1440 | |||
1441 | rc = qeth_l3_send_simple_setassparms(card, IPA_INBOUND_CHECKSUM, | ||
1442 | IPA_CMD_ASS_START, 0); | ||
1443 | if (rc) { | ||
1444 | PRINT_WARN("Starting Inbound HW Checksumming failed on %s: " | ||
1445 | "0x%x,\ncontinuing using Inbound SW Checksumming\n", | ||
1446 | QETH_CARD_IFNAME(card), rc); | ||
1447 | return rc; | ||
1448 | } | ||
1449 | rc = qeth_l3_send_simple_setassparms(card, IPA_INBOUND_CHECKSUM, | ||
1450 | IPA_CMD_ASS_ENABLE, | ||
1451 | card->info.csum_mask); | ||
1452 | if (rc) { | ||
1453 | PRINT_WARN("Enabling Inbound HW Checksumming failed on %s: " | ||
1454 | "0x%x,\ncontinuing using Inbound SW Checksumming\n", | ||
1455 | QETH_CARD_IFNAME(card), rc); | ||
1456 | return rc; | ||
1457 | } | ||
1458 | return 0; | ||
1459 | } | ||
1460 | |||
1461 | static int qeth_l3_start_ipa_checksum(struct qeth_card *card) | ||
1462 | { | ||
1463 | int rc = 0; | ||
1464 | |||
1465 | QETH_DBF_TEXT(TRACE, 3, "strtcsum"); | ||
1466 | |||
1467 | if (card->options.checksum_type == NO_CHECKSUMMING) { | ||
1468 | PRINT_WARN("Using no checksumming on %s.\n", | ||
1469 | QETH_CARD_IFNAME(card)); | ||
1470 | return 0; | ||
1471 | } | ||
1472 | if (card->options.checksum_type == SW_CHECKSUMMING) { | ||
1473 | PRINT_WARN("Using SW checksumming on %s.\n", | ||
1474 | QETH_CARD_IFNAME(card)); | ||
1475 | return 0; | ||
1476 | } | ||
1477 | if (!qeth_is_supported(card, IPA_INBOUND_CHECKSUM)) { | ||
1478 | PRINT_WARN("Inbound HW Checksumming not " | ||
1479 | "supported on %s,\ncontinuing " | ||
1480 | "using Inbound SW Checksumming\n", | ||
1481 | QETH_CARD_IFNAME(card)); | ||
1482 | card->options.checksum_type = SW_CHECKSUMMING; | ||
1483 | return 0; | ||
1484 | } | ||
1485 | rc = qeth_l3_send_checksum_command(card); | ||
1486 | if (!rc) | ||
1487 | PRINT_INFO("HW Checksumming (inbound) enabled \n"); | ||
1488 | |||
1489 | return rc; | ||
1490 | } | ||
1491 | |||
1492 | static int qeth_l3_start_ipa_tso(struct qeth_card *card) | ||
1493 | { | ||
1494 | int rc; | ||
1495 | |||
1496 | QETH_DBF_TEXT(TRACE, 3, "sttso"); | ||
1497 | |||
1498 | if (!qeth_is_supported(card, IPA_OUTBOUND_TSO)) { | ||
1499 | PRINT_WARN("Outbound TSO not supported on %s\n", | ||
1500 | QETH_CARD_IFNAME(card)); | ||
1501 | rc = -EOPNOTSUPP; | ||
1502 | } else { | ||
1503 | rc = qeth_l3_send_simple_setassparms(card, IPA_OUTBOUND_TSO, | ||
1504 | IPA_CMD_ASS_START, 0); | ||
1505 | if (rc) | ||
1506 | PRINT_WARN("Could not start outbound TSO " | ||
1507 | "assist on %s: rc=%i\n", | ||
1508 | QETH_CARD_IFNAME(card), rc); | ||
1509 | else | ||
1510 | PRINT_INFO("Outbound TSO enabled\n"); | ||
1511 | } | ||
1512 | if (rc && (card->options.large_send == QETH_LARGE_SEND_TSO)) { | ||
1513 | card->options.large_send = QETH_LARGE_SEND_NO; | ||
1514 | card->dev->features &= ~(NETIF_F_TSO | NETIF_F_SG); | ||
1515 | } | ||
1516 | return rc; | ||
1517 | } | ||
1518 | |||
1519 | static int qeth_l3_start_ipassists(struct qeth_card *card) | ||
1520 | { | ||
1521 | QETH_DBF_TEXT(TRACE, 3, "strtipas"); | ||
1522 | qeth_l3_start_ipa_arp_processing(card); /* go on*/ | ||
1523 | qeth_l3_start_ipa_ip_fragmentation(card); /* go on*/ | ||
1524 | qeth_l3_start_ipa_source_mac(card); /* go on*/ | ||
1525 | qeth_l3_start_ipa_vlan(card); /* go on*/ | ||
1526 | qeth_l3_start_ipa_multicast(card); /* go on*/ | ||
1527 | qeth_l3_start_ipa_ipv6(card); /* go on*/ | ||
1528 | qeth_l3_start_ipa_broadcast(card); /* go on*/ | ||
1529 | qeth_l3_start_ipa_checksum(card); /* go on*/ | ||
1530 | qeth_l3_start_ipa_tso(card); /* go on*/ | ||
1531 | return 0; | ||
1532 | } | ||
1533 | |||
1534 | static int qeth_l3_put_unique_id(struct qeth_card *card) | ||
1535 | { | ||
1536 | |||
1537 | int rc = 0; | ||
1538 | struct qeth_cmd_buffer *iob; | ||
1539 | struct qeth_ipa_cmd *cmd; | ||
1540 | |||
1541 | QETH_DBF_TEXT(TRACE, 2, "puniqeid"); | ||
1542 | |||
1543 | if ((card->info.unique_id & UNIQUE_ID_NOT_BY_CARD) == | ||
1544 | UNIQUE_ID_NOT_BY_CARD) | ||
1545 | return -1; | ||
1546 | iob = qeth_get_ipacmd_buffer(card, IPA_CMD_DESTROY_ADDR, | ||
1547 | QETH_PROT_IPV6); | ||
1548 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
1549 | *((__u16 *) &cmd->data.create_destroy_addr.unique_id[6]) = | ||
1550 | card->info.unique_id; | ||
1551 | memcpy(&cmd->data.create_destroy_addr.unique_id[0], | ||
1552 | card->dev->dev_addr, OSA_ADDR_LEN); | ||
1553 | rc = qeth_send_ipa_cmd(card, iob, NULL, NULL); | ||
1554 | return rc; | ||
1555 | } | ||
1556 | |||
1557 | static int qeth_l3_iqd_read_initial_mac_cb(struct qeth_card *card, | ||
1558 | struct qeth_reply *reply, unsigned long data) | ||
1559 | { | ||
1560 | struct qeth_ipa_cmd *cmd; | ||
1561 | |||
1562 | cmd = (struct qeth_ipa_cmd *) data; | ||
1563 | if (cmd->hdr.return_code == 0) | ||
1564 | memcpy(card->dev->dev_addr, | ||
1565 | cmd->data.create_destroy_addr.unique_id, ETH_ALEN); | ||
1566 | else | ||
1567 | random_ether_addr(card->dev->dev_addr); | ||
1568 | |||
1569 | return 0; | ||
1570 | } | ||
1571 | |||
1572 | static int qeth_l3_iqd_read_initial_mac(struct qeth_card *card) | ||
1573 | { | ||
1574 | int rc = 0; | ||
1575 | struct qeth_cmd_buffer *iob; | ||
1576 | struct qeth_ipa_cmd *cmd; | ||
1577 | |||
1578 | QETH_DBF_TEXT(SETUP, 2, "hsrmac"); | ||
1579 | |||
1580 | iob = qeth_get_ipacmd_buffer(card, IPA_CMD_CREATE_ADDR, | ||
1581 | QETH_PROT_IPV6); | ||
1582 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
1583 | *((__u16 *) &cmd->data.create_destroy_addr.unique_id[6]) = | ||
1584 | card->info.unique_id; | ||
1585 | |||
1586 | rc = qeth_send_ipa_cmd(card, iob, qeth_l3_iqd_read_initial_mac_cb, | ||
1587 | NULL); | ||
1588 | return rc; | ||
1589 | } | ||
1590 | |||
1591 | static int qeth_l3_get_unique_id_cb(struct qeth_card *card, | ||
1592 | struct qeth_reply *reply, unsigned long data) | ||
1593 | { | ||
1594 | struct qeth_ipa_cmd *cmd; | ||
1595 | |||
1596 | cmd = (struct qeth_ipa_cmd *) data; | ||
1597 | if (cmd->hdr.return_code == 0) | ||
1598 | card->info.unique_id = *((__u16 *) | ||
1599 | &cmd->data.create_destroy_addr.unique_id[6]); | ||
1600 | else { | ||
1601 | card->info.unique_id = UNIQUE_ID_IF_CREATE_ADDR_FAILED | | ||
1602 | UNIQUE_ID_NOT_BY_CARD; | ||
1603 | PRINT_WARN("couldn't get a unique id from the card on device " | ||
1604 | "%s (result=x%x), using default id. ipv6 " | ||
1605 | "autoconfig on other lpars may lead to duplicate " | ||
1606 | "ip addresses. please use manually " | ||
1607 | "configured ones.\n", | ||
1608 | CARD_BUS_ID(card), cmd->hdr.return_code); | ||
1609 | } | ||
1610 | return 0; | ||
1611 | } | ||
1612 | |||
1613 | static int qeth_l3_get_unique_id(struct qeth_card *card) | ||
1614 | { | ||
1615 | int rc = 0; | ||
1616 | struct qeth_cmd_buffer *iob; | ||
1617 | struct qeth_ipa_cmd *cmd; | ||
1618 | |||
1619 | QETH_DBF_TEXT(SETUP, 2, "guniqeid"); | ||
1620 | |||
1621 | if (!qeth_is_supported(card, IPA_IPV6)) { | ||
1622 | card->info.unique_id = UNIQUE_ID_IF_CREATE_ADDR_FAILED | | ||
1623 | UNIQUE_ID_NOT_BY_CARD; | ||
1624 | return 0; | ||
1625 | } | ||
1626 | |||
1627 | iob = qeth_get_ipacmd_buffer(card, IPA_CMD_CREATE_ADDR, | ||
1628 | QETH_PROT_IPV6); | ||
1629 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
1630 | *((__u16 *) &cmd->data.create_destroy_addr.unique_id[6]) = | ||
1631 | card->info.unique_id; | ||
1632 | |||
1633 | rc = qeth_send_ipa_cmd(card, iob, qeth_l3_get_unique_id_cb, NULL); | ||
1634 | return rc; | ||
1635 | } | ||
1636 | |||
1637 | static void qeth_l3_get_mac_for_ipm(__u32 ipm, char *mac, | ||
1638 | struct net_device *dev) | ||
1639 | { | ||
1640 | if (dev->type == ARPHRD_IEEE802_TR) | ||
1641 | ip_tr_mc_map(ipm, mac); | ||
1642 | else | ||
1643 | ip_eth_mc_map(ipm, mac); | ||
1644 | } | ||
1645 | |||
1646 | static void qeth_l3_add_mc(struct qeth_card *card, struct in_device *in4_dev) | ||
1647 | { | ||
1648 | struct qeth_ipaddr *ipm; | ||
1649 | struct ip_mc_list *im4; | ||
1650 | char buf[MAX_ADDR_LEN]; | ||
1651 | |||
1652 | QETH_DBF_TEXT(TRACE, 4, "addmc"); | ||
1653 | for (im4 = in4_dev->mc_list; im4; im4 = im4->next) { | ||
1654 | qeth_l3_get_mac_for_ipm(im4->multiaddr, buf, in4_dev->dev); | ||
1655 | ipm = qeth_l3_get_addr_buffer(QETH_PROT_IPV4); | ||
1656 | if (!ipm) | ||
1657 | continue; | ||
1658 | ipm->u.a4.addr = im4->multiaddr; | ||
1659 | memcpy(ipm->mac, buf, OSA_ADDR_LEN); | ||
1660 | ipm->is_multicast = 1; | ||
1661 | if (!qeth_l3_add_ip(card, ipm)) | ||
1662 | kfree(ipm); | ||
1663 | } | ||
1664 | } | ||
1665 | |||
1666 | static void qeth_l3_add_vlan_mc(struct qeth_card *card) | ||
1667 | { | ||
1668 | struct in_device *in_dev; | ||
1669 | struct vlan_group *vg; | ||
1670 | int i; | ||
1671 | |||
1672 | QETH_DBF_TEXT(TRACE, 4, "addmcvl"); | ||
1673 | if (!qeth_is_supported(card, IPA_FULL_VLAN) || (card->vlangrp == NULL)) | ||
1674 | return; | ||
1675 | |||
1676 | vg = card->vlangrp; | ||
1677 | for (i = 0; i < VLAN_GROUP_ARRAY_LEN; i++) { | ||
1678 | struct net_device *netdev = vlan_group_get_device(vg, i); | ||
1679 | if (netdev == NULL || | ||
1680 | !(netdev->flags & IFF_UP)) | ||
1681 | continue; | ||
1682 | in_dev = in_dev_get(netdev); | ||
1683 | if (!in_dev) | ||
1684 | continue; | ||
1685 | read_lock(&in_dev->mc_list_lock); | ||
1686 | qeth_l3_add_mc(card, in_dev); | ||
1687 | read_unlock(&in_dev->mc_list_lock); | ||
1688 | in_dev_put(in_dev); | ||
1689 | } | ||
1690 | } | ||
1691 | |||
1692 | static void qeth_l3_add_multicast_ipv4(struct qeth_card *card) | ||
1693 | { | ||
1694 | struct in_device *in4_dev; | ||
1695 | |||
1696 | QETH_DBF_TEXT(TRACE, 4, "chkmcv4"); | ||
1697 | in4_dev = in_dev_get(card->dev); | ||
1698 | if (in4_dev == NULL) | ||
1699 | return; | ||
1700 | read_lock(&in4_dev->mc_list_lock); | ||
1701 | qeth_l3_add_mc(card, in4_dev); | ||
1702 | qeth_l3_add_vlan_mc(card); | ||
1703 | read_unlock(&in4_dev->mc_list_lock); | ||
1704 | in_dev_put(in4_dev); | ||
1705 | } | ||
1706 | |||
1707 | #ifdef CONFIG_QETH_IPV6 | ||
1708 | static void qeth_l3_add_mc6(struct qeth_card *card, struct inet6_dev *in6_dev) | ||
1709 | { | ||
1710 | struct qeth_ipaddr *ipm; | ||
1711 | struct ifmcaddr6 *im6; | ||
1712 | char buf[MAX_ADDR_LEN]; | ||
1713 | |||
1714 | QETH_DBF_TEXT(TRACE, 4, "addmc6"); | ||
1715 | for (im6 = in6_dev->mc_list; im6 != NULL; im6 = im6->next) { | ||
1716 | ndisc_mc_map(&im6->mca_addr, buf, in6_dev->dev, 0); | ||
1717 | ipm = qeth_l3_get_addr_buffer(QETH_PROT_IPV6); | ||
1718 | if (!ipm) | ||
1719 | continue; | ||
1720 | ipm->is_multicast = 1; | ||
1721 | memcpy(ipm->mac, buf, OSA_ADDR_LEN); | ||
1722 | memcpy(&ipm->u.a6.addr, &im6->mca_addr.s6_addr, | ||
1723 | sizeof(struct in6_addr)); | ||
1724 | if (!qeth_l3_add_ip(card, ipm)) | ||
1725 | kfree(ipm); | ||
1726 | } | ||
1727 | } | ||
1728 | |||
1729 | static void qeth_l3_add_vlan_mc6(struct qeth_card *card) | ||
1730 | { | ||
1731 | struct inet6_dev *in_dev; | ||
1732 | struct vlan_group *vg; | ||
1733 | int i; | ||
1734 | |||
1735 | QETH_DBF_TEXT(TRACE, 4, "admc6vl"); | ||
1736 | if (!qeth_is_supported(card, IPA_FULL_VLAN) || (card->vlangrp == NULL)) | ||
1737 | return; | ||
1738 | |||
1739 | vg = card->vlangrp; | ||
1740 | for (i = 0; i < VLAN_GROUP_ARRAY_LEN; i++) { | ||
1741 | struct net_device *netdev = vlan_group_get_device(vg, i); | ||
1742 | if (netdev == NULL || | ||
1743 | !(netdev->flags & IFF_UP)) | ||
1744 | continue; | ||
1745 | in_dev = in6_dev_get(netdev); | ||
1746 | if (!in_dev) | ||
1747 | continue; | ||
1748 | read_lock_bh(&in_dev->lock); | ||
1749 | qeth_l3_add_mc6(card, in_dev); | ||
1750 | read_unlock_bh(&in_dev->lock); | ||
1751 | in6_dev_put(in_dev); | ||
1752 | } | ||
1753 | } | ||
1754 | |||
1755 | static void qeth_l3_add_multicast_ipv6(struct qeth_card *card) | ||
1756 | { | ||
1757 | struct inet6_dev *in6_dev; | ||
1758 | |||
1759 | QETH_DBF_TEXT(TRACE, 4, "chkmcv6"); | ||
1760 | if (!qeth_is_supported(card, IPA_IPV6)) | ||
1761 | return ; | ||
1762 | in6_dev = in6_dev_get(card->dev); | ||
1763 | if (in6_dev == NULL) | ||
1764 | return; | ||
1765 | read_lock_bh(&in6_dev->lock); | ||
1766 | qeth_l3_add_mc6(card, in6_dev); | ||
1767 | qeth_l3_add_vlan_mc6(card); | ||
1768 | read_unlock_bh(&in6_dev->lock); | ||
1769 | in6_dev_put(in6_dev); | ||
1770 | } | ||
1771 | #endif /* CONFIG_QETH_IPV6 */ | ||
1772 | |||
1773 | static void qeth_l3_free_vlan_addresses4(struct qeth_card *card, | ||
1774 | unsigned short vid) | ||
1775 | { | ||
1776 | struct in_device *in_dev; | ||
1777 | struct in_ifaddr *ifa; | ||
1778 | struct qeth_ipaddr *addr; | ||
1779 | |||
1780 | QETH_DBF_TEXT(TRACE, 4, "frvaddr4"); | ||
1781 | |||
1782 | in_dev = in_dev_get(vlan_group_get_device(card->vlangrp, vid)); | ||
1783 | if (!in_dev) | ||
1784 | return; | ||
1785 | for (ifa = in_dev->ifa_list; ifa; ifa = ifa->ifa_next) { | ||
1786 | addr = qeth_l3_get_addr_buffer(QETH_PROT_IPV4); | ||
1787 | if (addr) { | ||
1788 | addr->u.a4.addr = ifa->ifa_address; | ||
1789 | addr->u.a4.mask = ifa->ifa_mask; | ||
1790 | addr->type = QETH_IP_TYPE_NORMAL; | ||
1791 | if (!qeth_l3_delete_ip(card, addr)) | ||
1792 | kfree(addr); | ||
1793 | } | ||
1794 | } | ||
1795 | in_dev_put(in_dev); | ||
1796 | } | ||
1797 | |||
1798 | static void qeth_l3_free_vlan_addresses6(struct qeth_card *card, | ||
1799 | unsigned short vid) | ||
1800 | { | ||
1801 | #ifdef CONFIG_QETH_IPV6 | ||
1802 | struct inet6_dev *in6_dev; | ||
1803 | struct inet6_ifaddr *ifa; | ||
1804 | struct qeth_ipaddr *addr; | ||
1805 | |||
1806 | QETH_DBF_TEXT(TRACE, 4, "frvaddr6"); | ||
1807 | |||
1808 | in6_dev = in6_dev_get(vlan_group_get_device(card->vlangrp, vid)); | ||
1809 | if (!in6_dev) | ||
1810 | return; | ||
1811 | for (ifa = in6_dev->addr_list; ifa; ifa = ifa->lst_next) { | ||
1812 | addr = qeth_l3_get_addr_buffer(QETH_PROT_IPV6); | ||
1813 | if (addr) { | ||
1814 | memcpy(&addr->u.a6.addr, &ifa->addr, | ||
1815 | sizeof(struct in6_addr)); | ||
1816 | addr->u.a6.pfxlen = ifa->prefix_len; | ||
1817 | addr->type = QETH_IP_TYPE_NORMAL; | ||
1818 | if (!qeth_l3_delete_ip(card, addr)) | ||
1819 | kfree(addr); | ||
1820 | } | ||
1821 | } | ||
1822 | in6_dev_put(in6_dev); | ||
1823 | #endif /* CONFIG_QETH_IPV6 */ | ||
1824 | } | ||
1825 | |||
1826 | static void qeth_l3_free_vlan_addresses(struct qeth_card *card, | ||
1827 | unsigned short vid) | ||
1828 | { | ||
1829 | if (!card->vlangrp) | ||
1830 | return; | ||
1831 | qeth_l3_free_vlan_addresses4(card, vid); | ||
1832 | qeth_l3_free_vlan_addresses6(card, vid); | ||
1833 | } | ||
1834 | |||
1835 | static void qeth_l3_vlan_rx_register(struct net_device *dev, | ||
1836 | struct vlan_group *grp) | ||
1837 | { | ||
1838 | struct qeth_card *card = netdev_priv(dev); | ||
1839 | unsigned long flags; | ||
1840 | |||
1841 | QETH_DBF_TEXT(TRACE, 4, "vlanreg"); | ||
1842 | spin_lock_irqsave(&card->vlanlock, flags); | ||
1843 | card->vlangrp = grp; | ||
1844 | spin_unlock_irqrestore(&card->vlanlock, flags); | ||
1845 | } | ||
1846 | |||
1847 | static void qeth_l3_vlan_rx_add_vid(struct net_device *dev, unsigned short vid) | ||
1848 | { | ||
1849 | struct net_device *vlandev; | ||
1850 | struct qeth_card *card = (struct qeth_card *) dev->priv; | ||
1851 | struct in_device *in_dev; | ||
1852 | |||
1853 | if (card->info.type == QETH_CARD_TYPE_IQD) | ||
1854 | return; | ||
1855 | |||
1856 | vlandev = vlan_group_get_device(card->vlangrp, vid); | ||
1857 | vlandev->neigh_setup = qeth_l3_neigh_setup; | ||
1858 | |||
1859 | in_dev = in_dev_get(vlandev); | ||
1860 | #ifdef CONFIG_SYSCTL | ||
1861 | neigh_sysctl_unregister(in_dev->arp_parms); | ||
1862 | #endif | ||
1863 | neigh_parms_release(&arp_tbl, in_dev->arp_parms); | ||
1864 | |||
1865 | in_dev->arp_parms = neigh_parms_alloc(vlandev, &arp_tbl); | ||
1866 | #ifdef CONFIG_SYSCTL | ||
1867 | neigh_sysctl_register(vlandev, in_dev->arp_parms, NET_IPV4, | ||
1868 | NET_IPV4_NEIGH, "ipv4", NULL, NULL); | ||
1869 | #endif | ||
1870 | in_dev_put(in_dev); | ||
1871 | return; | ||
1872 | } | ||
1873 | |||
1874 | static void qeth_l3_vlan_rx_kill_vid(struct net_device *dev, unsigned short vid) | ||
1875 | { | ||
1876 | struct qeth_card *card = netdev_priv(dev); | ||
1877 | unsigned long flags; | ||
1878 | |||
1879 | QETH_DBF_TEXT_(TRACE, 4, "kid:%d", vid); | ||
1880 | spin_lock_irqsave(&card->vlanlock, flags); | ||
1881 | /* unregister IP addresses of vlan device */ | ||
1882 | qeth_l3_free_vlan_addresses(card, vid); | ||
1883 | vlan_group_set_device(card->vlangrp, vid, NULL); | ||
1884 | spin_unlock_irqrestore(&card->vlanlock, flags); | ||
1885 | qeth_l3_set_multicast_list(card->dev); | ||
1886 | } | ||
1887 | |||
1888 | static inline __u16 qeth_l3_rebuild_skb(struct qeth_card *card, | ||
1889 | struct sk_buff *skb, struct qeth_hdr *hdr) | ||
1890 | { | ||
1891 | unsigned short vlan_id = 0; | ||
1892 | __be16 prot; | ||
1893 | struct iphdr *ip_hdr; | ||
1894 | unsigned char tg_addr[MAX_ADDR_LEN]; | ||
1895 | |||
1896 | if (!(hdr->hdr.l3.flags & QETH_HDR_PASSTHRU)) { | ||
1897 | prot = htons((hdr->hdr.l3.flags & QETH_HDR_IPV6)? ETH_P_IPV6 : | ||
1898 | ETH_P_IP); | ||
1899 | switch (hdr->hdr.l3.flags & QETH_HDR_CAST_MASK) { | ||
1900 | case QETH_CAST_MULTICAST: | ||
1901 | switch (prot) { | ||
1902 | #ifdef CONFIG_QETH_IPV6 | ||
1903 | case __constant_htons(ETH_P_IPV6): | ||
1904 | ndisc_mc_map((struct in6_addr *) | ||
1905 | skb->data + 24, | ||
1906 | tg_addr, card->dev, 0); | ||
1907 | break; | ||
1908 | #endif | ||
1909 | case __constant_htons(ETH_P_IP): | ||
1910 | ip_hdr = (struct iphdr *)skb->data; | ||
1911 | (card->dev->type == ARPHRD_IEEE802_TR) ? | ||
1912 | ip_tr_mc_map(ip_hdr->daddr, tg_addr): | ||
1913 | ip_eth_mc_map(ip_hdr->daddr, tg_addr); | ||
1914 | break; | ||
1915 | default: | ||
1916 | memcpy(tg_addr, card->dev->broadcast, | ||
1917 | card->dev->addr_len); | ||
1918 | } | ||
1919 | card->stats.multicast++; | ||
1920 | skb->pkt_type = PACKET_MULTICAST; | ||
1921 | break; | ||
1922 | case QETH_CAST_BROADCAST: | ||
1923 | memcpy(tg_addr, card->dev->broadcast, | ||
1924 | card->dev->addr_len); | ||
1925 | card->stats.multicast++; | ||
1926 | skb->pkt_type = PACKET_BROADCAST; | ||
1927 | break; | ||
1928 | case QETH_CAST_UNICAST: | ||
1929 | case QETH_CAST_ANYCAST: | ||
1930 | case QETH_CAST_NOCAST: | ||
1931 | default: | ||
1932 | skb->pkt_type = PACKET_HOST; | ||
1933 | memcpy(tg_addr, card->dev->dev_addr, | ||
1934 | card->dev->addr_len); | ||
1935 | } | ||
1936 | card->dev->header_ops->create(skb, card->dev, prot, tg_addr, | ||
1937 | "FAKELL", card->dev->addr_len); | ||
1938 | } | ||
1939 | |||
1940 | #ifdef CONFIG_TR | ||
1941 | if (card->dev->type == ARPHRD_IEEE802_TR) | ||
1942 | skb->protocol = tr_type_trans(skb, card->dev); | ||
1943 | else | ||
1944 | #endif | ||
1945 | skb->protocol = eth_type_trans(skb, card->dev); | ||
1946 | |||
1947 | if (hdr->hdr.l3.ext_flags & | ||
1948 | (QETH_HDR_EXT_VLAN_FRAME | QETH_HDR_EXT_INCLUDE_VLAN_TAG)) { | ||
1949 | vlan_id = (hdr->hdr.l3.ext_flags & QETH_HDR_EXT_VLAN_FRAME)? | ||
1950 | hdr->hdr.l3.vlan_id : *((u16 *)&hdr->hdr.l3.dest_addr[12]); | ||
1951 | } | ||
1952 | |||
1953 | skb->ip_summed = card->options.checksum_type; | ||
1954 | if (card->options.checksum_type == HW_CHECKSUMMING) { | ||
1955 | if ((hdr->hdr.l3.ext_flags & | ||
1956 | (QETH_HDR_EXT_CSUM_HDR_REQ | | ||
1957 | QETH_HDR_EXT_CSUM_TRANSP_REQ)) == | ||
1958 | (QETH_HDR_EXT_CSUM_HDR_REQ | | ||
1959 | QETH_HDR_EXT_CSUM_TRANSP_REQ)) | ||
1960 | skb->ip_summed = CHECKSUM_UNNECESSARY; | ||
1961 | else | ||
1962 | skb->ip_summed = SW_CHECKSUMMING; | ||
1963 | } | ||
1964 | |||
1965 | return vlan_id; | ||
1966 | } | ||
1967 | |||
1968 | static void qeth_l3_process_inbound_buffer(struct qeth_card *card, | ||
1969 | struct qeth_qdio_buffer *buf, int index) | ||
1970 | { | ||
1971 | struct qdio_buffer_element *element; | ||
1972 | struct sk_buff *skb; | ||
1973 | struct qeth_hdr *hdr; | ||
1974 | int offset; | ||
1975 | __u16 vlan_tag = 0; | ||
1976 | unsigned int len; | ||
1977 | |||
1978 | /* get first element of current buffer */ | ||
1979 | element = (struct qdio_buffer_element *)&buf->buffer->element[0]; | ||
1980 | offset = 0; | ||
1981 | if (card->options.performance_stats) | ||
1982 | card->perf_stats.bufs_rec++; | ||
1983 | while ((skb = qeth_core_get_next_skb(card, buf->buffer, &element, | ||
1984 | &offset, &hdr))) { | ||
1985 | skb->dev = card->dev; | ||
1986 | /* is device UP ? */ | ||
1987 | if (!(card->dev->flags & IFF_UP)) { | ||
1988 | dev_kfree_skb_any(skb); | ||
1989 | continue; | ||
1990 | } | ||
1991 | |||
1992 | switch (hdr->hdr.l3.id) { | ||
1993 | case QETH_HEADER_TYPE_LAYER3: | ||
1994 | vlan_tag = qeth_l3_rebuild_skb(card, skb, hdr); | ||
1995 | len = skb->len; | ||
1996 | if (vlan_tag) | ||
1997 | if (card->vlangrp) | ||
1998 | vlan_hwaccel_rx(skb, card->vlangrp, | ||
1999 | vlan_tag); | ||
2000 | else { | ||
2001 | dev_kfree_skb_any(skb); | ||
2002 | continue; | ||
2003 | } | ||
2004 | else | ||
2005 | netif_rx(skb); | ||
2006 | break; | ||
2007 | default: | ||
2008 | dev_kfree_skb_any(skb); | ||
2009 | QETH_DBF_TEXT(TRACE, 3, "inbunkno"); | ||
2010 | QETH_DBF_HEX(CTRL, 3, hdr, QETH_DBF_CTRL_LEN); | ||
2011 | continue; | ||
2012 | } | ||
2013 | |||
2014 | card->dev->last_rx = jiffies; | ||
2015 | card->stats.rx_packets++; | ||
2016 | card->stats.rx_bytes += len; | ||
2017 | } | ||
2018 | } | ||
2019 | |||
2020 | static int qeth_l3_verify_vlan_dev(struct net_device *dev, | ||
2021 | struct qeth_card *card) | ||
2022 | { | ||
2023 | int rc = 0; | ||
2024 | struct vlan_group *vg; | ||
2025 | int i; | ||
2026 | |||
2027 | vg = card->vlangrp; | ||
2028 | if (!vg) | ||
2029 | return rc; | ||
2030 | |||
2031 | for (i = 0; i < VLAN_GROUP_ARRAY_LEN; i++) { | ||
2032 | if (vlan_group_get_device(vg, i) == dev) { | ||
2033 | rc = QETH_VLAN_CARD; | ||
2034 | break; | ||
2035 | } | ||
2036 | } | ||
2037 | |||
2038 | if (rc && !(netdev_priv(vlan_dev_info(dev)->real_dev) == (void *)card)) | ||
2039 | return 0; | ||
2040 | |||
2041 | return rc; | ||
2042 | } | ||
2043 | |||
2044 | static int qeth_l3_verify_dev(struct net_device *dev) | ||
2045 | { | ||
2046 | struct qeth_card *card; | ||
2047 | unsigned long flags; | ||
2048 | int rc = 0; | ||
2049 | |||
2050 | read_lock_irqsave(&qeth_core_card_list.rwlock, flags); | ||
2051 | list_for_each_entry(card, &qeth_core_card_list.list, list) { | ||
2052 | if (card->dev == dev) { | ||
2053 | rc = QETH_REAL_CARD; | ||
2054 | break; | ||
2055 | } | ||
2056 | rc = qeth_l3_verify_vlan_dev(dev, card); | ||
2057 | if (rc) | ||
2058 | break; | ||
2059 | } | ||
2060 | read_unlock_irqrestore(&qeth_core_card_list.rwlock, flags); | ||
2061 | |||
2062 | return rc; | ||
2063 | } | ||
2064 | |||
2065 | static struct qeth_card *qeth_l3_get_card_from_dev(struct net_device *dev) | ||
2066 | { | ||
2067 | struct qeth_card *card = NULL; | ||
2068 | int rc; | ||
2069 | |||
2070 | rc = qeth_l3_verify_dev(dev); | ||
2071 | if (rc == QETH_REAL_CARD) | ||
2072 | card = netdev_priv(dev); | ||
2073 | else if (rc == QETH_VLAN_CARD) | ||
2074 | card = netdev_priv(vlan_dev_info(dev)->real_dev); | ||
2075 | if (card->options.layer2) | ||
2076 | card = NULL; | ||
2077 | QETH_DBF_TEXT_(TRACE, 4, "%d", rc); | ||
2078 | return card ; | ||
2079 | } | ||
2080 | |||
2081 | static int qeth_l3_stop_card(struct qeth_card *card, int recovery_mode) | ||
2082 | { | ||
2083 | int rc = 0; | ||
2084 | |||
2085 | QETH_DBF_TEXT(SETUP, 2, "stopcard"); | ||
2086 | QETH_DBF_HEX(SETUP, 2, &card, sizeof(void *)); | ||
2087 | |||
2088 | qeth_set_allowed_threads(card, 0, 1); | ||
2089 | if (qeth_wait_for_threads(card, ~QETH_RECOVER_THREAD)) | ||
2090 | return -ERESTARTSYS; | ||
2091 | if (card->read.state == CH_STATE_UP && | ||
2092 | card->write.state == CH_STATE_UP && | ||
2093 | (card->state == CARD_STATE_UP)) { | ||
2094 | if (recovery_mode) | ||
2095 | qeth_l3_stop(card->dev); | ||
2096 | if (!card->use_hard_stop) { | ||
2097 | rc = qeth_send_stoplan(card); | ||
2098 | if (rc) | ||
2099 | QETH_DBF_TEXT_(SETUP, 2, "1err%d", rc); | ||
2100 | } | ||
2101 | card->state = CARD_STATE_SOFTSETUP; | ||
2102 | } | ||
2103 | if (card->state == CARD_STATE_SOFTSETUP) { | ||
2104 | qeth_l3_clear_ip_list(card, !card->use_hard_stop, 1); | ||
2105 | qeth_clear_ipacmd_list(card); | ||
2106 | card->state = CARD_STATE_HARDSETUP; | ||
2107 | } | ||
2108 | if (card->state == CARD_STATE_HARDSETUP) { | ||
2109 | if (!card->use_hard_stop && | ||
2110 | (card->info.type != QETH_CARD_TYPE_IQD)) { | ||
2111 | rc = qeth_l3_put_unique_id(card); | ||
2112 | if (rc) | ||
2113 | QETH_DBF_TEXT_(SETUP, 2, "2err%d", rc); | ||
2114 | } | ||
2115 | qeth_qdio_clear_card(card, 0); | ||
2116 | qeth_clear_qdio_buffers(card); | ||
2117 | qeth_clear_working_pool_list(card); | ||
2118 | card->state = CARD_STATE_DOWN; | ||
2119 | } | ||
2120 | if (card->state == CARD_STATE_DOWN) { | ||
2121 | qeth_clear_cmd_buffers(&card->read); | ||
2122 | qeth_clear_cmd_buffers(&card->write); | ||
2123 | } | ||
2124 | card->use_hard_stop = 0; | ||
2125 | return rc; | ||
2126 | } | ||
2127 | |||
2128 | static void qeth_l3_set_multicast_list(struct net_device *dev) | ||
2129 | { | ||
2130 | struct qeth_card *card = netdev_priv(dev); | ||
2131 | |||
2132 | QETH_DBF_TEXT(TRACE, 3, "setmulti"); | ||
2133 | qeth_l3_delete_mc_addresses(card); | ||
2134 | qeth_l3_add_multicast_ipv4(card); | ||
2135 | #ifdef CONFIG_QETH_IPV6 | ||
2136 | qeth_l3_add_multicast_ipv6(card); | ||
2137 | #endif | ||
2138 | qeth_l3_set_ip_addr_list(card); | ||
2139 | if (!qeth_adp_supported(card, IPA_SETADP_SET_PROMISC_MODE)) | ||
2140 | return; | ||
2141 | qeth_setadp_promisc_mode(card); | ||
2142 | } | ||
2143 | |||
2144 | static const char *qeth_l3_arp_get_error_cause(int *rc) | ||
2145 | { | ||
2146 | switch (*rc) { | ||
2147 | case QETH_IPA_ARP_RC_FAILED: | ||
2148 | *rc = -EIO; | ||
2149 | return "operation failed"; | ||
2150 | case QETH_IPA_ARP_RC_NOTSUPP: | ||
2151 | *rc = -EOPNOTSUPP; | ||
2152 | return "operation not supported"; | ||
2153 | case QETH_IPA_ARP_RC_OUT_OF_RANGE: | ||
2154 | *rc = -EINVAL; | ||
2155 | return "argument out of range"; | ||
2156 | case QETH_IPA_ARP_RC_Q_NOTSUPP: | ||
2157 | *rc = -EOPNOTSUPP; | ||
2158 | return "query operation not supported"; | ||
2159 | case QETH_IPA_ARP_RC_Q_NO_DATA: | ||
2160 | *rc = -ENOENT; | ||
2161 | return "no query data available"; | ||
2162 | default: | ||
2163 | return "unknown error"; | ||
2164 | } | ||
2165 | } | ||
2166 | |||
2167 | static int qeth_l3_arp_set_no_entries(struct qeth_card *card, int no_entries) | ||
2168 | { | ||
2169 | int tmp; | ||
2170 | int rc; | ||
2171 | |||
2172 | QETH_DBF_TEXT(TRACE, 3, "arpstnoe"); | ||
2173 | |||
2174 | /* | ||
2175 | * currently GuestLAN only supports the ARP assist function | ||
2176 | * IPA_CMD_ASS_ARP_QUERY_INFO, but not IPA_CMD_ASS_ARP_SET_NO_ENTRIES; | ||
2177 | * thus we say EOPNOTSUPP for this ARP function | ||
2178 | */ | ||
2179 | if (card->info.guestlan) | ||
2180 | return -EOPNOTSUPP; | ||
2181 | if (!qeth_is_supported(card, IPA_ARP_PROCESSING)) { | ||
2182 | PRINT_WARN("ARP processing not supported " | ||
2183 | "on %s!\n", QETH_CARD_IFNAME(card)); | ||
2184 | return -EOPNOTSUPP; | ||
2185 | } | ||
2186 | rc = qeth_l3_send_simple_setassparms(card, IPA_ARP_PROCESSING, | ||
2187 | IPA_CMD_ASS_ARP_SET_NO_ENTRIES, | ||
2188 | no_entries); | ||
2189 | if (rc) { | ||
2190 | tmp = rc; | ||
2191 | PRINT_WARN("Could not set number of ARP entries on %s: " | ||
2192 | "%s (0x%x/%d)\n", QETH_CARD_IFNAME(card), | ||
2193 | qeth_l3_arp_get_error_cause(&rc), tmp, tmp); | ||
2194 | } | ||
2195 | return rc; | ||
2196 | } | ||
2197 | |||
2198 | static void qeth_l3_copy_arp_entries_stripped(struct qeth_arp_query_info *qinfo, | ||
2199 | struct qeth_arp_query_data *qdata, int entry_size, | ||
2200 | int uentry_size) | ||
2201 | { | ||
2202 | char *entry_ptr; | ||
2203 | char *uentry_ptr; | ||
2204 | int i; | ||
2205 | |||
2206 | entry_ptr = (char *)&qdata->data; | ||
2207 | uentry_ptr = (char *)(qinfo->udata + qinfo->udata_offset); | ||
2208 | for (i = 0; i < qdata->no_entries; ++i) { | ||
2209 | /* strip off 32 bytes "media specific information" */ | ||
2210 | memcpy(uentry_ptr, (entry_ptr + 32), entry_size - 32); | ||
2211 | entry_ptr += entry_size; | ||
2212 | uentry_ptr += uentry_size; | ||
2213 | } | ||
2214 | } | ||
2215 | |||
2216 | static int qeth_l3_arp_query_cb(struct qeth_card *card, | ||
2217 | struct qeth_reply *reply, unsigned long data) | ||
2218 | { | ||
2219 | struct qeth_ipa_cmd *cmd; | ||
2220 | struct qeth_arp_query_data *qdata; | ||
2221 | struct qeth_arp_query_info *qinfo; | ||
2222 | int entry_size; | ||
2223 | int uentry_size; | ||
2224 | int i; | ||
2225 | |||
2226 | QETH_DBF_TEXT(TRACE, 4, "arpquecb"); | ||
2227 | |||
2228 | qinfo = (struct qeth_arp_query_info *) reply->param; | ||
2229 | cmd = (struct qeth_ipa_cmd *) data; | ||
2230 | if (cmd->hdr.return_code) { | ||
2231 | QETH_DBF_TEXT_(TRACE, 4, "qaer1%i", cmd->hdr.return_code); | ||
2232 | return 0; | ||
2233 | } | ||
2234 | if (cmd->data.setassparms.hdr.return_code) { | ||
2235 | cmd->hdr.return_code = cmd->data.setassparms.hdr.return_code; | ||
2236 | QETH_DBF_TEXT_(TRACE, 4, "qaer2%i", cmd->hdr.return_code); | ||
2237 | return 0; | ||
2238 | } | ||
2239 | qdata = &cmd->data.setassparms.data.query_arp; | ||
2240 | switch (qdata->reply_bits) { | ||
2241 | case 5: | ||
2242 | uentry_size = entry_size = sizeof(struct qeth_arp_qi_entry5); | ||
2243 | if (qinfo->mask_bits & QETH_QARP_STRIP_ENTRIES) | ||
2244 | uentry_size = sizeof(struct qeth_arp_qi_entry5_short); | ||
2245 | break; | ||
2246 | case 7: | ||
2247 | /* fall through to default */ | ||
2248 | default: | ||
2249 | /* tr is the same as eth -> entry7 */ | ||
2250 | uentry_size = entry_size = sizeof(struct qeth_arp_qi_entry7); | ||
2251 | if (qinfo->mask_bits & QETH_QARP_STRIP_ENTRIES) | ||
2252 | uentry_size = sizeof(struct qeth_arp_qi_entry7_short); | ||
2253 | break; | ||
2254 | } | ||
2255 | /* check if there is enough room in userspace */ | ||
2256 | if ((qinfo->udata_len - qinfo->udata_offset) < | ||
2257 | qdata->no_entries * uentry_size){ | ||
2258 | QETH_DBF_TEXT_(TRACE, 4, "qaer3%i", -ENOMEM); | ||
2259 | cmd->hdr.return_code = -ENOMEM; | ||
2260 | PRINT_WARN("query ARP user space buffer is too small for " | ||
2261 | "the returned number of ARP entries. " | ||
2262 | "Aborting query!\n"); | ||
2263 | goto out_error; | ||
2264 | } | ||
2265 | QETH_DBF_TEXT_(TRACE, 4, "anore%i", | ||
2266 | cmd->data.setassparms.hdr.number_of_replies); | ||
2267 | QETH_DBF_TEXT_(TRACE, 4, "aseqn%i", cmd->data.setassparms.hdr.seq_no); | ||
2268 | QETH_DBF_TEXT_(TRACE, 4, "anoen%i", qdata->no_entries); | ||
2269 | |||
2270 | if (qinfo->mask_bits & QETH_QARP_STRIP_ENTRIES) { | ||
2271 | /* strip off "media specific information" */ | ||
2272 | qeth_l3_copy_arp_entries_stripped(qinfo, qdata, entry_size, | ||
2273 | uentry_size); | ||
2274 | } else | ||
2275 | /*copy entries to user buffer*/ | ||
2276 | memcpy(qinfo->udata + qinfo->udata_offset, | ||
2277 | (char *)&qdata->data, qdata->no_entries*uentry_size); | ||
2278 | |||
2279 | qinfo->no_entries += qdata->no_entries; | ||
2280 | qinfo->udata_offset += (qdata->no_entries*uentry_size); | ||
2281 | /* check if all replies received ... */ | ||
2282 | if (cmd->data.setassparms.hdr.seq_no < | ||
2283 | cmd->data.setassparms.hdr.number_of_replies) | ||
2284 | return 1; | ||
2285 | memcpy(qinfo->udata, &qinfo->no_entries, 4); | ||
2286 | /* keep STRIP_ENTRIES flag so the user program can distinguish | ||
2287 | * stripped entries from normal ones */ | ||
2288 | if (qinfo->mask_bits & QETH_QARP_STRIP_ENTRIES) | ||
2289 | qdata->reply_bits |= QETH_QARP_STRIP_ENTRIES; | ||
2290 | memcpy(qinfo->udata + QETH_QARP_MASK_OFFSET, &qdata->reply_bits, 2); | ||
2291 | return 0; | ||
2292 | out_error: | ||
2293 | i = 0; | ||
2294 | memcpy(qinfo->udata, &i, 4); | ||
2295 | return 0; | ||
2296 | } | ||
2297 | |||
2298 | static int qeth_l3_send_ipa_arp_cmd(struct qeth_card *card, | ||
2299 | struct qeth_cmd_buffer *iob, int len, | ||
2300 | int (*reply_cb)(struct qeth_card *, struct qeth_reply *, | ||
2301 | unsigned long), | ||
2302 | void *reply_param) | ||
2303 | { | ||
2304 | QETH_DBF_TEXT(TRACE, 4, "sendarp"); | ||
2305 | |||
2306 | memcpy(iob->data, IPA_PDU_HEADER, IPA_PDU_HEADER_SIZE); | ||
2307 | memcpy(QETH_IPA_CMD_DEST_ADDR(iob->data), | ||
2308 | &card->token.ulp_connection_r, QETH_MPC_TOKEN_LENGTH); | ||
2309 | return qeth_send_control_data(card, IPA_PDU_HEADER_SIZE + len, iob, | ||
2310 | reply_cb, reply_param); | ||
2311 | } | ||
2312 | |||
2313 | static int qeth_l3_arp_query(struct qeth_card *card, char __user *udata) | ||
2314 | { | ||
2315 | struct qeth_cmd_buffer *iob; | ||
2316 | struct qeth_arp_query_info qinfo = {0, }; | ||
2317 | int tmp; | ||
2318 | int rc; | ||
2319 | |||
2320 | QETH_DBF_TEXT(TRACE, 3, "arpquery"); | ||
2321 | |||
2322 | if (!qeth_is_supported(card,/*IPA_QUERY_ARP_ADDR_INFO*/ | ||
2323 | IPA_ARP_PROCESSING)) { | ||
2324 | PRINT_WARN("ARP processing not supported " | ||
2325 | "on %s!\n", QETH_CARD_IFNAME(card)); | ||
2326 | return -EOPNOTSUPP; | ||
2327 | } | ||
2328 | /* get size of userspace buffer and mask_bits -> 6 bytes */ | ||
2329 | if (copy_from_user(&qinfo, udata, 6)) | ||
2330 | return -EFAULT; | ||
2331 | qinfo.udata = kzalloc(qinfo.udata_len, GFP_KERNEL); | ||
2332 | if (!qinfo.udata) | ||
2333 | return -ENOMEM; | ||
2334 | qinfo.udata_offset = QETH_QARP_ENTRIES_OFFSET; | ||
2335 | iob = qeth_l3_get_setassparms_cmd(card, IPA_ARP_PROCESSING, | ||
2336 | IPA_CMD_ASS_ARP_QUERY_INFO, | ||
2337 | sizeof(int), QETH_PROT_IPV4); | ||
2338 | |||
2339 | rc = qeth_l3_send_ipa_arp_cmd(card, iob, | ||
2340 | QETH_SETASS_BASE_LEN+QETH_ARP_CMD_LEN, | ||
2341 | qeth_l3_arp_query_cb, (void *)&qinfo); | ||
2342 | if (rc) { | ||
2343 | tmp = rc; | ||
2344 | PRINT_WARN("Error while querying ARP cache on %s: %s " | ||
2345 | "(0x%x/%d)\n", QETH_CARD_IFNAME(card), | ||
2346 | qeth_l3_arp_get_error_cause(&rc), tmp, tmp); | ||
2347 | if (copy_to_user(udata, qinfo.udata, 4)) | ||
2348 | rc = -EFAULT; | ||
2349 | } else { | ||
2350 | if (copy_to_user(udata, qinfo.udata, qinfo.udata_len)) | ||
2351 | rc = -EFAULT; | ||
2352 | } | ||
2353 | kfree(qinfo.udata); | ||
2354 | return rc; | ||
2355 | } | ||
2356 | |||
2357 | static int qeth_l3_arp_add_entry(struct qeth_card *card, | ||
2358 | struct qeth_arp_cache_entry *entry) | ||
2359 | { | ||
2360 | struct qeth_cmd_buffer *iob; | ||
2361 | char buf[16]; | ||
2362 | int tmp; | ||
2363 | int rc; | ||
2364 | |||
2365 | QETH_DBF_TEXT(TRACE, 3, "arpadent"); | ||
2366 | |||
2367 | /* | ||
2368 | * currently GuestLAN only supports the ARP assist function | ||
2369 | * IPA_CMD_ASS_ARP_QUERY_INFO, but not IPA_CMD_ASS_ARP_ADD_ENTRY; | ||
2370 | * thus we say EOPNOTSUPP for this ARP function | ||
2371 | */ | ||
2372 | if (card->info.guestlan) | ||
2373 | return -EOPNOTSUPP; | ||
2374 | if (!qeth_is_supported(card, IPA_ARP_PROCESSING)) { | ||
2375 | PRINT_WARN("ARP processing not supported " | ||
2376 | "on %s!\n", QETH_CARD_IFNAME(card)); | ||
2377 | return -EOPNOTSUPP; | ||
2378 | } | ||
2379 | |||
2380 | iob = qeth_l3_get_setassparms_cmd(card, IPA_ARP_PROCESSING, | ||
2381 | IPA_CMD_ASS_ARP_ADD_ENTRY, | ||
2382 | sizeof(struct qeth_arp_cache_entry), | ||
2383 | QETH_PROT_IPV4); | ||
2384 | rc = qeth_l3_send_setassparms(card, iob, | ||
2385 | sizeof(struct qeth_arp_cache_entry), | ||
2386 | (unsigned long) entry, | ||
2387 | qeth_l3_default_setassparms_cb, NULL); | ||
2388 | if (rc) { | ||
2389 | tmp = rc; | ||
2390 | qeth_l3_ipaddr4_to_string((u8 *)entry->ipaddr, buf); | ||
2391 | PRINT_WARN("Could not add ARP entry for address %s on %s: " | ||
2392 | "%s (0x%x/%d)\n", | ||
2393 | buf, QETH_CARD_IFNAME(card), | ||
2394 | qeth_l3_arp_get_error_cause(&rc), tmp, tmp); | ||
2395 | } | ||
2396 | return rc; | ||
2397 | } | ||
2398 | |||
2399 | static int qeth_l3_arp_remove_entry(struct qeth_card *card, | ||
2400 | struct qeth_arp_cache_entry *entry) | ||
2401 | { | ||
2402 | struct qeth_cmd_buffer *iob; | ||
2403 | char buf[16] = {0, }; | ||
2404 | int tmp; | ||
2405 | int rc; | ||
2406 | |||
2407 | QETH_DBF_TEXT(TRACE, 3, "arprment"); | ||
2408 | |||
2409 | /* | ||
2410 | * currently GuestLAN only supports the ARP assist function | ||
2411 | * IPA_CMD_ASS_ARP_QUERY_INFO, but not IPA_CMD_ASS_ARP_REMOVE_ENTRY; | ||
2412 | * thus we say EOPNOTSUPP for this ARP function | ||
2413 | */ | ||
2414 | if (card->info.guestlan) | ||
2415 | return -EOPNOTSUPP; | ||
2416 | if (!qeth_is_supported(card, IPA_ARP_PROCESSING)) { | ||
2417 | PRINT_WARN("ARP processing not supported " | ||
2418 | "on %s!\n", QETH_CARD_IFNAME(card)); | ||
2419 | return -EOPNOTSUPP; | ||
2420 | } | ||
2421 | memcpy(buf, entry, 12); | ||
2422 | iob = qeth_l3_get_setassparms_cmd(card, IPA_ARP_PROCESSING, | ||
2423 | IPA_CMD_ASS_ARP_REMOVE_ENTRY, | ||
2424 | 12, | ||
2425 | QETH_PROT_IPV4); | ||
2426 | rc = qeth_l3_send_setassparms(card, iob, | ||
2427 | 12, (unsigned long)buf, | ||
2428 | qeth_l3_default_setassparms_cb, NULL); | ||
2429 | if (rc) { | ||
2430 | tmp = rc; | ||
2431 | memset(buf, 0, 16); | ||
2432 | qeth_l3_ipaddr4_to_string((u8 *)entry->ipaddr, buf); | ||
2433 | PRINT_WARN("Could not delete ARP entry for address %s on %s: " | ||
2434 | "%s (0x%x/%d)\n", | ||
2435 | buf, QETH_CARD_IFNAME(card), | ||
2436 | qeth_l3_arp_get_error_cause(&rc), tmp, tmp); | ||
2437 | } | ||
2438 | return rc; | ||
2439 | } | ||
2440 | |||
2441 | static int qeth_l3_arp_flush_cache(struct qeth_card *card) | ||
2442 | { | ||
2443 | int rc; | ||
2444 | int tmp; | ||
2445 | |||
2446 | QETH_DBF_TEXT(TRACE, 3, "arpflush"); | ||
2447 | |||
2448 | /* | ||
2449 | * currently GuestLAN only supports the ARP assist function | ||
2450 | * IPA_CMD_ASS_ARP_QUERY_INFO, but not IPA_CMD_ASS_ARP_FLUSH_CACHE; | ||
2451 | * thus we say EOPNOTSUPP for this ARP function | ||
2452 | */ | ||
2453 | if (card->info.guestlan || (card->info.type == QETH_CARD_TYPE_IQD)) | ||
2454 | return -EOPNOTSUPP; | ||
2455 | if (!qeth_is_supported(card, IPA_ARP_PROCESSING)) { | ||
2456 | PRINT_WARN("ARP processing not supported " | ||
2457 | "on %s!\n", QETH_CARD_IFNAME(card)); | ||
2458 | return -EOPNOTSUPP; | ||
2459 | } | ||
2460 | rc = qeth_l3_send_simple_setassparms(card, IPA_ARP_PROCESSING, | ||
2461 | IPA_CMD_ASS_ARP_FLUSH_CACHE, 0); | ||
2462 | if (rc) { | ||
2463 | tmp = rc; | ||
2464 | PRINT_WARN("Could not flush ARP cache on %s: %s (0x%x/%d)\n", | ||
2465 | QETH_CARD_IFNAME(card), | ||
2466 | qeth_l3_arp_get_error_cause(&rc), tmp, tmp); | ||
2467 | } | ||
2468 | return rc; | ||
2469 | } | ||
2470 | |||
2471 | static int qeth_l3_do_ioctl(struct net_device *dev, struct ifreq *rq, int cmd) | ||
2472 | { | ||
2473 | struct qeth_card *card = netdev_priv(dev); | ||
2474 | struct qeth_arp_cache_entry arp_entry; | ||
2475 | struct mii_ioctl_data *mii_data; | ||
2476 | int rc = 0; | ||
2477 | |||
2478 | if (!card) | ||
2479 | return -ENODEV; | ||
2480 | |||
2481 | if ((card->state != CARD_STATE_UP) && | ||
2482 | (card->state != CARD_STATE_SOFTSETUP)) | ||
2483 | return -ENODEV; | ||
2484 | |||
2485 | switch (cmd) { | ||
2486 | case SIOC_QETH_ARP_SET_NO_ENTRIES: | ||
2487 | if (!capable(CAP_NET_ADMIN)) { | ||
2488 | rc = -EPERM; | ||
2489 | break; | ||
2490 | } | ||
2491 | rc = qeth_l3_arp_set_no_entries(card, rq->ifr_ifru.ifru_ivalue); | ||
2492 | break; | ||
2493 | case SIOC_QETH_ARP_QUERY_INFO: | ||
2494 | if (!capable(CAP_NET_ADMIN)) { | ||
2495 | rc = -EPERM; | ||
2496 | break; | ||
2497 | } | ||
2498 | rc = qeth_l3_arp_query(card, rq->ifr_ifru.ifru_data); | ||
2499 | break; | ||
2500 | case SIOC_QETH_ARP_ADD_ENTRY: | ||
2501 | if (!capable(CAP_NET_ADMIN)) { | ||
2502 | rc = -EPERM; | ||
2503 | break; | ||
2504 | } | ||
2505 | if (copy_from_user(&arp_entry, rq->ifr_ifru.ifru_data, | ||
2506 | sizeof(struct qeth_arp_cache_entry))) | ||
2507 | rc = -EFAULT; | ||
2508 | else | ||
2509 | rc = qeth_l3_arp_add_entry(card, &arp_entry); | ||
2510 | break; | ||
2511 | case SIOC_QETH_ARP_REMOVE_ENTRY: | ||
2512 | if (!capable(CAP_NET_ADMIN)) { | ||
2513 | rc = -EPERM; | ||
2514 | break; | ||
2515 | } | ||
2516 | if (copy_from_user(&arp_entry, rq->ifr_ifru.ifru_data, | ||
2517 | sizeof(struct qeth_arp_cache_entry))) | ||
2518 | rc = -EFAULT; | ||
2519 | else | ||
2520 | rc = qeth_l3_arp_remove_entry(card, &arp_entry); | ||
2521 | break; | ||
2522 | case SIOC_QETH_ARP_FLUSH_CACHE: | ||
2523 | if (!capable(CAP_NET_ADMIN)) { | ||
2524 | rc = -EPERM; | ||
2525 | break; | ||
2526 | } | ||
2527 | rc = qeth_l3_arp_flush_cache(card); | ||
2528 | break; | ||
2529 | case SIOC_QETH_ADP_SET_SNMP_CONTROL: | ||
2530 | rc = qeth_snmp_command(card, rq->ifr_ifru.ifru_data); | ||
2531 | break; | ||
2532 | case SIOC_QETH_GET_CARD_TYPE: | ||
2533 | if ((card->info.type == QETH_CARD_TYPE_OSAE) && | ||
2534 | !card->info.guestlan) | ||
2535 | return 1; | ||
2536 | return 0; | ||
2537 | break; | ||
2538 | case SIOCGMIIPHY: | ||
2539 | mii_data = if_mii(rq); | ||
2540 | mii_data->phy_id = 0; | ||
2541 | break; | ||
2542 | case SIOCGMIIREG: | ||
2543 | mii_data = if_mii(rq); | ||
2544 | if (mii_data->phy_id != 0) | ||
2545 | rc = -EINVAL; | ||
2546 | else | ||
2547 | mii_data->val_out = qeth_mdio_read(dev, | ||
2548 | mii_data->phy_id, | ||
2549 | mii_data->reg_num); | ||
2550 | break; | ||
2551 | default: | ||
2552 | rc = -EOPNOTSUPP; | ||
2553 | } | ||
2554 | if (rc) | ||
2555 | QETH_DBF_TEXT_(TRACE, 2, "ioce%d", rc); | ||
2556 | return rc; | ||
2557 | } | ||
2558 | |||
2559 | static void qeth_l3_fill_header(struct qeth_card *card, struct qeth_hdr *hdr, | ||
2560 | struct sk_buff *skb, int ipv, int cast_type) | ||
2561 | { | ||
2562 | QETH_DBF_TEXT(TRACE, 6, "fillhdr"); | ||
2563 | |||
2564 | memset(hdr, 0, sizeof(struct qeth_hdr)); | ||
2565 | hdr->hdr.l3.id = QETH_HEADER_TYPE_LAYER3; | ||
2566 | hdr->hdr.l3.ext_flags = 0; | ||
2567 | |||
2568 | /* | ||
2569 | * before we're going to overwrite this location with next hop ip. | ||
2570 | * v6 uses passthrough, v4 sets the tag in the QDIO header. | ||
2571 | */ | ||
2572 | if (card->vlangrp && vlan_tx_tag_present(skb)) { | ||
2573 | hdr->hdr.l3.ext_flags = (ipv == 4) ? | ||
2574 | QETH_HDR_EXT_VLAN_FRAME : | ||
2575 | QETH_HDR_EXT_INCLUDE_VLAN_TAG; | ||
2576 | hdr->hdr.l3.vlan_id = vlan_tx_tag_get(skb); | ||
2577 | } | ||
2578 | |||
2579 | hdr->hdr.l3.length = skb->len - sizeof(struct qeth_hdr); | ||
2580 | if (ipv == 4) { | ||
2581 | /* IPv4 */ | ||
2582 | hdr->hdr.l3.flags = qeth_l3_get_qeth_hdr_flags4(cast_type); | ||
2583 | memset(hdr->hdr.l3.dest_addr, 0, 12); | ||
2584 | if ((skb->dst) && (skb->dst->neighbour)) { | ||
2585 | *((u32 *) (&hdr->hdr.l3.dest_addr[12])) = | ||
2586 | *((u32 *) skb->dst->neighbour->primary_key); | ||
2587 | } else { | ||
2588 | /* fill in destination address used in ip header */ | ||
2589 | *((u32 *) (&hdr->hdr.l3.dest_addr[12])) = | ||
2590 | ip_hdr(skb)->daddr; | ||
2591 | } | ||
2592 | } else if (ipv == 6) { | ||
2593 | /* IPv6 */ | ||
2594 | hdr->hdr.l3.flags = qeth_l3_get_qeth_hdr_flags6(cast_type); | ||
2595 | if (card->info.type == QETH_CARD_TYPE_IQD) | ||
2596 | hdr->hdr.l3.flags &= ~QETH_HDR_PASSTHRU; | ||
2597 | if ((skb->dst) && (skb->dst->neighbour)) { | ||
2598 | memcpy(hdr->hdr.l3.dest_addr, | ||
2599 | skb->dst->neighbour->primary_key, 16); | ||
2600 | } else { | ||
2601 | /* fill in destination address used in ip header */ | ||
2602 | memcpy(hdr->hdr.l3.dest_addr, | ||
2603 | &ipv6_hdr(skb)->daddr, 16); | ||
2604 | } | ||
2605 | } else { | ||
2606 | /* passthrough */ | ||
2607 | if ((skb->dev->type == ARPHRD_IEEE802_TR) && | ||
2608 | !memcmp(skb->data + sizeof(struct qeth_hdr) + | ||
2609 | sizeof(__u16), skb->dev->broadcast, 6)) { | ||
2610 | hdr->hdr.l3.flags = QETH_CAST_BROADCAST | | ||
2611 | QETH_HDR_PASSTHRU; | ||
2612 | } else if (!memcmp(skb->data + sizeof(struct qeth_hdr), | ||
2613 | skb->dev->broadcast, 6)) { | ||
2614 | /* broadcast? */ | ||
2615 | hdr->hdr.l3.flags = QETH_CAST_BROADCAST | | ||
2616 | QETH_HDR_PASSTHRU; | ||
2617 | } else { | ||
2618 | hdr->hdr.l3.flags = (cast_type == RTN_MULTICAST) ? | ||
2619 | QETH_CAST_MULTICAST | QETH_HDR_PASSTHRU : | ||
2620 | QETH_CAST_UNICAST | QETH_HDR_PASSTHRU; | ||
2621 | } | ||
2622 | } | ||
2623 | } | ||
2624 | |||
2625 | static int qeth_l3_hard_start_xmit(struct sk_buff *skb, struct net_device *dev) | ||
2626 | { | ||
2627 | int rc; | ||
2628 | u16 *tag; | ||
2629 | struct qeth_hdr *hdr = NULL; | ||
2630 | int elements_needed = 0; | ||
2631 | struct qeth_card *card = netdev_priv(dev); | ||
2632 | struct sk_buff *new_skb = NULL; | ||
2633 | int ipv = qeth_get_ip_version(skb); | ||
2634 | int cast_type = qeth_get_cast_type(card, skb); | ||
2635 | struct qeth_qdio_out_q *queue = card->qdio.out_qs | ||
2636 | [qeth_get_priority_queue(card, skb, ipv, cast_type)]; | ||
2637 | int tx_bytes = skb->len; | ||
2638 | enum qeth_large_send_types large_send = QETH_LARGE_SEND_NO; | ||
2639 | struct qeth_eddp_context *ctx = NULL; | ||
2640 | |||
2641 | QETH_DBF_TEXT(TRACE, 6, "l3xmit"); | ||
2642 | |||
2643 | if ((card->info.type == QETH_CARD_TYPE_IQD) && | ||
2644 | (skb->protocol != htons(ETH_P_IPV6)) && | ||
2645 | (skb->protocol != htons(ETH_P_IP))) | ||
2646 | goto tx_drop; | ||
2647 | |||
2648 | if ((card->state != CARD_STATE_UP) || !card->lan_online) { | ||
2649 | card->stats.tx_carrier_errors++; | ||
2650 | goto tx_drop; | ||
2651 | } | ||
2652 | |||
2653 | if ((cast_type == RTN_BROADCAST) && | ||
2654 | (card->info.broadcast_capable == 0)) | ||
2655 | goto tx_drop; | ||
2656 | |||
2657 | if (card->options.performance_stats) { | ||
2658 | card->perf_stats.outbound_cnt++; | ||
2659 | card->perf_stats.outbound_start_time = qeth_get_micros(); | ||
2660 | } | ||
2661 | |||
2662 | /* create a clone with writeable headroom */ | ||
2663 | new_skb = skb_realloc_headroom(skb, sizeof(struct qeth_hdr_tso) + | ||
2664 | VLAN_HLEN); | ||
2665 | if (!new_skb) | ||
2666 | goto tx_drop; | ||
2667 | |||
2668 | if (card->info.type == QETH_CARD_TYPE_IQD) { | ||
2669 | skb_pull(new_skb, ETH_HLEN); | ||
2670 | } else { | ||
2671 | if (new_skb->protocol == htons(ETH_P_IP)) { | ||
2672 | if (card->dev->type == ARPHRD_IEEE802_TR) | ||
2673 | skb_pull(new_skb, TR_HLEN); | ||
2674 | else | ||
2675 | skb_pull(new_skb, ETH_HLEN); | ||
2676 | } | ||
2677 | |||
2678 | if (new_skb->protocol == ETH_P_IPV6 && card->vlangrp && | ||
2679 | vlan_tx_tag_present(new_skb)) { | ||
2680 | skb_push(new_skb, VLAN_HLEN); | ||
2681 | skb_copy_to_linear_data(new_skb, new_skb->data + 4, 4); | ||
2682 | skb_copy_to_linear_data_offset(new_skb, 4, | ||
2683 | new_skb->data + 8, 4); | ||
2684 | skb_copy_to_linear_data_offset(new_skb, 8, | ||
2685 | new_skb->data + 12, 4); | ||
2686 | tag = (u16 *)(new_skb->data + 12); | ||
2687 | *tag = __constant_htons(ETH_P_8021Q); | ||
2688 | *(tag + 1) = htons(vlan_tx_tag_get(new_skb)); | ||
2689 | VLAN_TX_SKB_CB(new_skb)->magic = 0; | ||
2690 | } | ||
2691 | } | ||
2692 | |||
2693 | netif_stop_queue(dev); | ||
2694 | |||
2695 | if (skb_is_gso(new_skb)) | ||
2696 | large_send = card->options.large_send; | ||
2697 | |||
2698 | /* fix hardware limitation: as long as we do not have sbal | ||
2699 | * chaining we can not send long frag lists so we temporary | ||
2700 | * switch to EDDP | ||
2701 | */ | ||
2702 | if ((large_send == QETH_LARGE_SEND_TSO) && | ||
2703 | ((skb_shinfo(new_skb)->nr_frags + 2) > 16)) | ||
2704 | large_send = QETH_LARGE_SEND_EDDP; | ||
2705 | |||
2706 | if ((large_send == QETH_LARGE_SEND_TSO) && | ||
2707 | (cast_type == RTN_UNSPEC)) { | ||
2708 | hdr = (struct qeth_hdr *)skb_push(new_skb, | ||
2709 | sizeof(struct qeth_hdr_tso)); | ||
2710 | memset(hdr, 0, sizeof(struct qeth_hdr_tso)); | ||
2711 | qeth_l3_fill_header(card, hdr, new_skb, ipv, cast_type); | ||
2712 | qeth_tso_fill_header(card, hdr, new_skb); | ||
2713 | elements_needed++; | ||
2714 | } else { | ||
2715 | hdr = (struct qeth_hdr *)skb_push(new_skb, | ||
2716 | sizeof(struct qeth_hdr)); | ||
2717 | qeth_l3_fill_header(card, hdr, new_skb, ipv, cast_type); | ||
2718 | } | ||
2719 | |||
2720 | if (large_send == QETH_LARGE_SEND_EDDP) { | ||
2721 | /* new_skb is not owned by a socket so we use skb to get | ||
2722 | * the protocol | ||
2723 | */ | ||
2724 | ctx = qeth_eddp_create_context(card, new_skb, hdr, | ||
2725 | skb->sk->sk_protocol); | ||
2726 | if (ctx == NULL) { | ||
2727 | PRINT_WARN("could not create eddp context\n"); | ||
2728 | goto tx_drop; | ||
2729 | } | ||
2730 | } else { | ||
2731 | int elems = qeth_get_elements_no(card, (void *)hdr, new_skb, | ||
2732 | elements_needed); | ||
2733 | if (!elems) | ||
2734 | goto tx_drop; | ||
2735 | elements_needed += elems; | ||
2736 | } | ||
2737 | |||
2738 | if ((large_send == QETH_LARGE_SEND_NO) && | ||
2739 | (new_skb->ip_summed == CHECKSUM_PARTIAL)) | ||
2740 | qeth_tx_csum(new_skb); | ||
2741 | |||
2742 | if (card->info.type != QETH_CARD_TYPE_IQD) | ||
2743 | rc = qeth_do_send_packet(card, queue, new_skb, hdr, | ||
2744 | elements_needed, ctx); | ||
2745 | else | ||
2746 | rc = qeth_do_send_packet_fast(card, queue, new_skb, hdr, | ||
2747 | elements_needed, ctx); | ||
2748 | |||
2749 | if (!rc) { | ||
2750 | card->stats.tx_packets++; | ||
2751 | card->stats.tx_bytes += tx_bytes; | ||
2752 | if (new_skb != skb) | ||
2753 | dev_kfree_skb_any(skb); | ||
2754 | if (card->options.performance_stats) { | ||
2755 | if (large_send != QETH_LARGE_SEND_NO) { | ||
2756 | card->perf_stats.large_send_bytes += tx_bytes; | ||
2757 | card->perf_stats.large_send_cnt++; | ||
2758 | } | ||
2759 | if (skb_shinfo(new_skb)->nr_frags > 0) { | ||
2760 | card->perf_stats.sg_skbs_sent++; | ||
2761 | /* nr_frags + skb->data */ | ||
2762 | card->perf_stats.sg_frags_sent += | ||
2763 | skb_shinfo(new_skb)->nr_frags + 1; | ||
2764 | } | ||
2765 | } | ||
2766 | |||
2767 | if (ctx != NULL) { | ||
2768 | qeth_eddp_put_context(ctx); | ||
2769 | dev_kfree_skb_any(new_skb); | ||
2770 | } | ||
2771 | } else { | ||
2772 | if (ctx != NULL) | ||
2773 | qeth_eddp_put_context(ctx); | ||
2774 | |||
2775 | if (rc == -EBUSY) { | ||
2776 | if (new_skb != skb) | ||
2777 | dev_kfree_skb_any(new_skb); | ||
2778 | return NETDEV_TX_BUSY; | ||
2779 | } else | ||
2780 | goto tx_drop; | ||
2781 | } | ||
2782 | |||
2783 | netif_wake_queue(dev); | ||
2784 | if (card->options.performance_stats) | ||
2785 | card->perf_stats.outbound_time += qeth_get_micros() - | ||
2786 | card->perf_stats.outbound_start_time; | ||
2787 | return rc; | ||
2788 | |||
2789 | tx_drop: | ||
2790 | card->stats.tx_dropped++; | ||
2791 | card->stats.tx_errors++; | ||
2792 | if ((new_skb != skb) && new_skb) | ||
2793 | dev_kfree_skb_any(new_skb); | ||
2794 | dev_kfree_skb_any(skb); | ||
2795 | return NETDEV_TX_OK; | ||
2796 | } | ||
2797 | |||
2798 | static int qeth_l3_open(struct net_device *dev) | ||
2799 | { | ||
2800 | struct qeth_card *card = netdev_priv(dev); | ||
2801 | |||
2802 | QETH_DBF_TEXT(TRACE, 4, "qethopen"); | ||
2803 | if (card->state != CARD_STATE_SOFTSETUP) | ||
2804 | return -ENODEV; | ||
2805 | card->data.state = CH_STATE_UP; | ||
2806 | card->state = CARD_STATE_UP; | ||
2807 | card->dev->flags |= IFF_UP; | ||
2808 | netif_start_queue(dev); | ||
2809 | |||
2810 | if (!card->lan_online && netif_carrier_ok(dev)) | ||
2811 | netif_carrier_off(dev); | ||
2812 | return 0; | ||
2813 | } | ||
2814 | |||
2815 | static int qeth_l3_stop(struct net_device *dev) | ||
2816 | { | ||
2817 | struct qeth_card *card = netdev_priv(dev); | ||
2818 | |||
2819 | QETH_DBF_TEXT(TRACE, 4, "qethstop"); | ||
2820 | netif_tx_disable(dev); | ||
2821 | card->dev->flags &= ~IFF_UP; | ||
2822 | if (card->state == CARD_STATE_UP) | ||
2823 | card->state = CARD_STATE_SOFTSETUP; | ||
2824 | return 0; | ||
2825 | } | ||
2826 | |||
2827 | static u32 qeth_l3_ethtool_get_rx_csum(struct net_device *dev) | ||
2828 | { | ||
2829 | struct qeth_card *card = netdev_priv(dev); | ||
2830 | |||
2831 | return (card->options.checksum_type == HW_CHECKSUMMING); | ||
2832 | } | ||
2833 | |||
2834 | static int qeth_l3_ethtool_set_rx_csum(struct net_device *dev, u32 data) | ||
2835 | { | ||
2836 | struct qeth_card *card = netdev_priv(dev); | ||
2837 | enum qeth_card_states old_state; | ||
2838 | enum qeth_checksum_types csum_type; | ||
2839 | |||
2840 | if ((card->state != CARD_STATE_UP) && | ||
2841 | (card->state != CARD_STATE_DOWN)) | ||
2842 | return -EPERM; | ||
2843 | |||
2844 | if (data) | ||
2845 | csum_type = HW_CHECKSUMMING; | ||
2846 | else | ||
2847 | csum_type = SW_CHECKSUMMING; | ||
2848 | |||
2849 | if (card->options.checksum_type != csum_type) { | ||
2850 | old_state = card->state; | ||
2851 | if (card->state == CARD_STATE_UP) | ||
2852 | __qeth_l3_set_offline(card->gdev, 1); | ||
2853 | card->options.checksum_type = csum_type; | ||
2854 | if (old_state == CARD_STATE_UP) | ||
2855 | __qeth_l3_set_online(card->gdev, 1); | ||
2856 | } | ||
2857 | return 0; | ||
2858 | } | ||
2859 | |||
2860 | static int qeth_l3_ethtool_set_tso(struct net_device *dev, u32 data) | ||
2861 | { | ||
2862 | struct qeth_card *card = netdev_priv(dev); | ||
2863 | |||
2864 | if (data) { | ||
2865 | if (card->options.large_send == QETH_LARGE_SEND_NO) { | ||
2866 | if (card->info.type == QETH_CARD_TYPE_IQD) | ||
2867 | card->options.large_send = QETH_LARGE_SEND_EDDP; | ||
2868 | else | ||
2869 | card->options.large_send = QETH_LARGE_SEND_TSO; | ||
2870 | dev->features |= NETIF_F_TSO; | ||
2871 | } | ||
2872 | } else { | ||
2873 | dev->features &= ~NETIF_F_TSO; | ||
2874 | card->options.large_send = QETH_LARGE_SEND_NO; | ||
2875 | } | ||
2876 | return 0; | ||
2877 | } | ||
2878 | |||
2879 | static struct ethtool_ops qeth_l3_ethtool_ops = { | ||
2880 | .get_link = ethtool_op_get_link, | ||
2881 | .get_tx_csum = ethtool_op_get_tx_csum, | ||
2882 | .set_tx_csum = ethtool_op_set_tx_hw_csum, | ||
2883 | .get_rx_csum = qeth_l3_ethtool_get_rx_csum, | ||
2884 | .set_rx_csum = qeth_l3_ethtool_set_rx_csum, | ||
2885 | .get_sg = ethtool_op_get_sg, | ||
2886 | .set_sg = ethtool_op_set_sg, | ||
2887 | .get_tso = ethtool_op_get_tso, | ||
2888 | .set_tso = qeth_l3_ethtool_set_tso, | ||
2889 | .get_strings = qeth_core_get_strings, | ||
2890 | .get_ethtool_stats = qeth_core_get_ethtool_stats, | ||
2891 | .get_stats_count = qeth_core_get_stats_count, | ||
2892 | .get_drvinfo = qeth_core_get_drvinfo, | ||
2893 | }; | ||
2894 | |||
2895 | /* | ||
2896 | * we need NOARP for IPv4 but we want neighbor solicitation for IPv6. Setting | ||
2897 | * NOARP on the netdevice is no option because it also turns off neighbor | ||
2898 | * solicitation. For IPv4 we install a neighbor_setup function. We don't want | ||
2899 | * arp resolution but we want the hard header (packet socket will work | ||
2900 | * e.g. tcpdump) | ||
2901 | */ | ||
2902 | static int qeth_l3_neigh_setup_noarp(struct neighbour *n) | ||
2903 | { | ||
2904 | n->nud_state = NUD_NOARP; | ||
2905 | memcpy(n->ha, "FAKELL", 6); | ||
2906 | n->output = n->ops->connected_output; | ||
2907 | return 0; | ||
2908 | } | ||
2909 | |||
2910 | static int | ||
2911 | qeth_l3_neigh_setup(struct net_device *dev, struct neigh_parms *np) | ||
2912 | { | ||
2913 | if (np->tbl->family == AF_INET) | ||
2914 | np->neigh_setup = qeth_l3_neigh_setup_noarp; | ||
2915 | |||
2916 | return 0; | ||
2917 | } | ||
2918 | |||
2919 | static int qeth_l3_setup_netdev(struct qeth_card *card) | ||
2920 | { | ||
2921 | if (card->info.type == QETH_CARD_TYPE_OSAE) { | ||
2922 | if ((card->info.link_type == QETH_LINK_TYPE_LANE_TR) || | ||
2923 | (card->info.link_type == QETH_LINK_TYPE_HSTR)) { | ||
2924 | #ifdef CONFIG_TR | ||
2925 | card->dev = alloc_trdev(0); | ||
2926 | #endif | ||
2927 | if (!card->dev) | ||
2928 | return -ENODEV; | ||
2929 | } else { | ||
2930 | card->dev = alloc_etherdev(0); | ||
2931 | if (!card->dev) | ||
2932 | return -ENODEV; | ||
2933 | card->dev->neigh_setup = qeth_l3_neigh_setup; | ||
2934 | |||
2935 | /*IPv6 address autoconfiguration stuff*/ | ||
2936 | qeth_l3_get_unique_id(card); | ||
2937 | if (!(card->info.unique_id & UNIQUE_ID_NOT_BY_CARD)) | ||
2938 | card->dev->dev_id = card->info.unique_id & | ||
2939 | 0xffff; | ||
2940 | } | ||
2941 | } else if (card->info.type == QETH_CARD_TYPE_IQD) { | ||
2942 | card->dev = alloc_netdev(0, "hsi%d", ether_setup); | ||
2943 | if (!card->dev) | ||
2944 | return -ENODEV; | ||
2945 | card->dev->flags |= IFF_NOARP; | ||
2946 | qeth_l3_iqd_read_initial_mac(card); | ||
2947 | } else | ||
2948 | return -ENODEV; | ||
2949 | |||
2950 | card->dev->hard_start_xmit = qeth_l3_hard_start_xmit; | ||
2951 | card->dev->priv = card; | ||
2952 | card->dev->tx_timeout = &qeth_tx_timeout; | ||
2953 | card->dev->watchdog_timeo = QETH_TX_TIMEOUT; | ||
2954 | card->dev->open = qeth_l3_open; | ||
2955 | card->dev->stop = qeth_l3_stop; | ||
2956 | card->dev->do_ioctl = qeth_l3_do_ioctl; | ||
2957 | card->dev->get_stats = qeth_get_stats; | ||
2958 | card->dev->change_mtu = qeth_change_mtu; | ||
2959 | card->dev->set_multicast_list = qeth_l3_set_multicast_list; | ||
2960 | card->dev->vlan_rx_register = qeth_l3_vlan_rx_register; | ||
2961 | card->dev->vlan_rx_add_vid = qeth_l3_vlan_rx_add_vid; | ||
2962 | card->dev->vlan_rx_kill_vid = qeth_l3_vlan_rx_kill_vid; | ||
2963 | card->dev->mtu = card->info.initial_mtu; | ||
2964 | card->dev->set_mac_address = NULL; | ||
2965 | SET_ETHTOOL_OPS(card->dev, &qeth_l3_ethtool_ops); | ||
2966 | card->dev->features |= NETIF_F_HW_VLAN_TX | | ||
2967 | NETIF_F_HW_VLAN_RX | | ||
2968 | NETIF_F_HW_VLAN_FILTER; | ||
2969 | |||
2970 | SET_NETDEV_DEV(card->dev, &card->gdev->dev); | ||
2971 | return register_netdev(card->dev); | ||
2972 | } | ||
2973 | |||
2974 | static void qeth_l3_qdio_input_handler(struct ccw_device *ccwdev, | ||
2975 | unsigned int status, unsigned int qdio_err, | ||
2976 | unsigned int siga_err, unsigned int queue, int first_element, | ||
2977 | int count, unsigned long card_ptr) | ||
2978 | { | ||
2979 | struct net_device *net_dev; | ||
2980 | struct qeth_card *card; | ||
2981 | struct qeth_qdio_buffer *buffer; | ||
2982 | int index; | ||
2983 | int i; | ||
2984 | |||
2985 | QETH_DBF_TEXT(TRACE, 6, "qdinput"); | ||
2986 | card = (struct qeth_card *) card_ptr; | ||
2987 | net_dev = card->dev; | ||
2988 | if (card->options.performance_stats) { | ||
2989 | card->perf_stats.inbound_cnt++; | ||
2990 | card->perf_stats.inbound_start_time = qeth_get_micros(); | ||
2991 | } | ||
2992 | if (status & QDIO_STATUS_LOOK_FOR_ERROR) { | ||
2993 | if (status & QDIO_STATUS_ACTIVATE_CHECK_CONDITION) { | ||
2994 | QETH_DBF_TEXT(TRACE, 1, "qdinchk"); | ||
2995 | QETH_DBF_TEXT_(TRACE, 1, "%s", CARD_BUS_ID(card)); | ||
2996 | QETH_DBF_TEXT_(TRACE, 1, "%04X%04X", | ||
2997 | first_element, count); | ||
2998 | QETH_DBF_TEXT_(TRACE, 1, "%04X%04X", queue, status); | ||
2999 | qeth_schedule_recovery(card); | ||
3000 | return; | ||
3001 | } | ||
3002 | } | ||
3003 | for (i = first_element; i < (first_element + count); ++i) { | ||
3004 | index = i % QDIO_MAX_BUFFERS_PER_Q; | ||
3005 | buffer = &card->qdio.in_q->bufs[index]; | ||
3006 | if (!((status & QDIO_STATUS_LOOK_FOR_ERROR) && | ||
3007 | qeth_check_qdio_errors(buffer->buffer, | ||
3008 | qdio_err, siga_err, "qinerr"))) | ||
3009 | qeth_l3_process_inbound_buffer(card, buffer, index); | ||
3010 | /* clear buffer and give back to hardware */ | ||
3011 | qeth_put_buffer_pool_entry(card, buffer->pool_entry); | ||
3012 | qeth_queue_input_buffer(card, index); | ||
3013 | } | ||
3014 | if (card->options.performance_stats) | ||
3015 | card->perf_stats.inbound_time += qeth_get_micros() - | ||
3016 | card->perf_stats.inbound_start_time; | ||
3017 | } | ||
3018 | |||
3019 | static int qeth_l3_probe_device(struct ccwgroup_device *gdev) | ||
3020 | { | ||
3021 | struct qeth_card *card = dev_get_drvdata(&gdev->dev); | ||
3022 | |||
3023 | qeth_l3_create_device_attributes(&gdev->dev); | ||
3024 | card->options.layer2 = 0; | ||
3025 | card->discipline.input_handler = (qdio_handler_t *) | ||
3026 | qeth_l3_qdio_input_handler; | ||
3027 | card->discipline.output_handler = (qdio_handler_t *) | ||
3028 | qeth_qdio_output_handler; | ||
3029 | card->discipline.recover = qeth_l3_recover; | ||
3030 | return 0; | ||
3031 | } | ||
3032 | |||
3033 | static void qeth_l3_remove_device(struct ccwgroup_device *cgdev) | ||
3034 | { | ||
3035 | struct qeth_card *card = dev_get_drvdata(&cgdev->dev); | ||
3036 | |||
3037 | wait_event(card->wait_q, qeth_threads_running(card, 0xffffffff) == 0); | ||
3038 | |||
3039 | if (cgdev->state == CCWGROUP_ONLINE) { | ||
3040 | card->use_hard_stop = 1; | ||
3041 | qeth_l3_set_offline(cgdev); | ||
3042 | } | ||
3043 | |||
3044 | if (card->dev) { | ||
3045 | unregister_netdev(card->dev); | ||
3046 | card->dev = NULL; | ||
3047 | } | ||
3048 | |||
3049 | qeth_l3_remove_device_attributes(&cgdev->dev); | ||
3050 | qeth_l3_clear_ip_list(card, 0, 0); | ||
3051 | qeth_l3_clear_ipato_list(card); | ||
3052 | return; | ||
3053 | } | ||
3054 | |||
3055 | static int __qeth_l3_set_online(struct ccwgroup_device *gdev, int recovery_mode) | ||
3056 | { | ||
3057 | struct qeth_card *card = dev_get_drvdata(&gdev->dev); | ||
3058 | int rc = 0; | ||
3059 | enum qeth_card_states recover_flag; | ||
3060 | |||
3061 | BUG_ON(!card); | ||
3062 | QETH_DBF_TEXT(SETUP, 2, "setonlin"); | ||
3063 | QETH_DBF_HEX(SETUP, 2, &card, sizeof(void *)); | ||
3064 | |||
3065 | qeth_set_allowed_threads(card, QETH_RECOVER_THREAD, 1); | ||
3066 | if (qeth_wait_for_threads(card, ~QETH_RECOVER_THREAD)) { | ||
3067 | PRINT_WARN("set_online of card %s interrupted by user!\n", | ||
3068 | CARD_BUS_ID(card)); | ||
3069 | return -ERESTARTSYS; | ||
3070 | } | ||
3071 | |||
3072 | recover_flag = card->state; | ||
3073 | rc = ccw_device_set_online(CARD_RDEV(card)); | ||
3074 | if (rc) { | ||
3075 | QETH_DBF_TEXT_(SETUP, 2, "1err%d", rc); | ||
3076 | return -EIO; | ||
3077 | } | ||
3078 | rc = ccw_device_set_online(CARD_WDEV(card)); | ||
3079 | if (rc) { | ||
3080 | QETH_DBF_TEXT_(SETUP, 2, "1err%d", rc); | ||
3081 | return -EIO; | ||
3082 | } | ||
3083 | rc = ccw_device_set_online(CARD_DDEV(card)); | ||
3084 | if (rc) { | ||
3085 | QETH_DBF_TEXT_(SETUP, 2, "1err%d", rc); | ||
3086 | return -EIO; | ||
3087 | } | ||
3088 | |||
3089 | rc = qeth_core_hardsetup_card(card); | ||
3090 | if (rc) { | ||
3091 | QETH_DBF_TEXT_(SETUP, 2, "2err%d", rc); | ||
3092 | goto out_remove; | ||
3093 | } | ||
3094 | |||
3095 | qeth_l3_query_ipassists(card, QETH_PROT_IPV4); | ||
3096 | |||
3097 | if (!card->dev && qeth_l3_setup_netdev(card)) | ||
3098 | goto out_remove; | ||
3099 | |||
3100 | card->state = CARD_STATE_HARDSETUP; | ||
3101 | qeth_print_status_message(card); | ||
3102 | |||
3103 | /* softsetup */ | ||
3104 | QETH_DBF_TEXT(SETUP, 2, "softsetp"); | ||
3105 | |||
3106 | rc = qeth_send_startlan(card); | ||
3107 | if (rc) { | ||
3108 | QETH_DBF_TEXT_(SETUP, 2, "1err%d", rc); | ||
3109 | if (rc == 0xe080) { | ||
3110 | PRINT_WARN("LAN on card %s if offline! " | ||
3111 | "Waiting for STARTLAN from card.\n", | ||
3112 | CARD_BUS_ID(card)); | ||
3113 | card->lan_online = 0; | ||
3114 | } | ||
3115 | return rc; | ||
3116 | } else | ||
3117 | card->lan_online = 1; | ||
3118 | qeth_set_large_send(card, card->options.large_send); | ||
3119 | |||
3120 | rc = qeth_l3_setadapter_parms(card); | ||
3121 | if (rc) | ||
3122 | QETH_DBF_TEXT_(SETUP, 2, "2err%d", rc); | ||
3123 | rc = qeth_l3_start_ipassists(card); | ||
3124 | if (rc) | ||
3125 | QETH_DBF_TEXT_(SETUP, 2, "3err%d", rc); | ||
3126 | rc = qeth_l3_setrouting_v4(card); | ||
3127 | if (rc) | ||
3128 | QETH_DBF_TEXT_(SETUP, 2, "4err%d", rc); | ||
3129 | rc = qeth_l3_setrouting_v6(card); | ||
3130 | if (rc) | ||
3131 | QETH_DBF_TEXT_(SETUP, 2, "5err%d", rc); | ||
3132 | netif_tx_disable(card->dev); | ||
3133 | |||
3134 | rc = qeth_init_qdio_queues(card); | ||
3135 | if (rc) { | ||
3136 | QETH_DBF_TEXT_(SETUP, 2, "6err%d", rc); | ||
3137 | goto out_remove; | ||
3138 | } | ||
3139 | card->state = CARD_STATE_SOFTSETUP; | ||
3140 | netif_carrier_on(card->dev); | ||
3141 | |||
3142 | qeth_set_allowed_threads(card, 0xffffffff, 0); | ||
3143 | if ((recover_flag == CARD_STATE_RECOVER) && recovery_mode) { | ||
3144 | qeth_l3_open(card->dev); | ||
3145 | qeth_l3_set_multicast_list(card->dev); | ||
3146 | } | ||
3147 | /* let user_space know that device is online */ | ||
3148 | kobject_uevent(&gdev->dev.kobj, KOBJ_CHANGE); | ||
3149 | return 0; | ||
3150 | out_remove: | ||
3151 | card->use_hard_stop = 1; | ||
3152 | qeth_l3_stop_card(card, 0); | ||
3153 | ccw_device_set_offline(CARD_DDEV(card)); | ||
3154 | ccw_device_set_offline(CARD_WDEV(card)); | ||
3155 | ccw_device_set_offline(CARD_RDEV(card)); | ||
3156 | if (recover_flag == CARD_STATE_RECOVER) | ||
3157 | card->state = CARD_STATE_RECOVER; | ||
3158 | else | ||
3159 | card->state = CARD_STATE_DOWN; | ||
3160 | return -ENODEV; | ||
3161 | } | ||
3162 | |||
3163 | static int qeth_l3_set_online(struct ccwgroup_device *gdev) | ||
3164 | { | ||
3165 | return __qeth_l3_set_online(gdev, 0); | ||
3166 | } | ||
3167 | |||
3168 | static int __qeth_l3_set_offline(struct ccwgroup_device *cgdev, | ||
3169 | int recovery_mode) | ||
3170 | { | ||
3171 | struct qeth_card *card = dev_get_drvdata(&cgdev->dev); | ||
3172 | int rc = 0, rc2 = 0, rc3 = 0; | ||
3173 | enum qeth_card_states recover_flag; | ||
3174 | |||
3175 | QETH_DBF_TEXT(SETUP, 3, "setoffl"); | ||
3176 | QETH_DBF_HEX(SETUP, 3, &card, sizeof(void *)); | ||
3177 | |||
3178 | if (card->dev && netif_carrier_ok(card->dev)) | ||
3179 | netif_carrier_off(card->dev); | ||
3180 | recover_flag = card->state; | ||
3181 | if (qeth_l3_stop_card(card, recovery_mode) == -ERESTARTSYS) { | ||
3182 | PRINT_WARN("Stopping card %s interrupted by user!\n", | ||
3183 | CARD_BUS_ID(card)); | ||
3184 | return -ERESTARTSYS; | ||
3185 | } | ||
3186 | rc = ccw_device_set_offline(CARD_DDEV(card)); | ||
3187 | rc2 = ccw_device_set_offline(CARD_WDEV(card)); | ||
3188 | rc3 = ccw_device_set_offline(CARD_RDEV(card)); | ||
3189 | if (!rc) | ||
3190 | rc = (rc2) ? rc2 : rc3; | ||
3191 | if (rc) | ||
3192 | QETH_DBF_TEXT_(SETUP, 2, "1err%d", rc); | ||
3193 | if (recover_flag == CARD_STATE_UP) | ||
3194 | card->state = CARD_STATE_RECOVER; | ||
3195 | /* let user_space know that device is offline */ | ||
3196 | kobject_uevent(&cgdev->dev.kobj, KOBJ_CHANGE); | ||
3197 | return 0; | ||
3198 | } | ||
3199 | |||
3200 | static int qeth_l3_set_offline(struct ccwgroup_device *cgdev) | ||
3201 | { | ||
3202 | return __qeth_l3_set_offline(cgdev, 0); | ||
3203 | } | ||
3204 | |||
3205 | static int qeth_l3_recover(void *ptr) | ||
3206 | { | ||
3207 | struct qeth_card *card; | ||
3208 | int rc = 0; | ||
3209 | |||
3210 | card = (struct qeth_card *) ptr; | ||
3211 | QETH_DBF_TEXT(TRACE, 2, "recover1"); | ||
3212 | QETH_DBF_HEX(TRACE, 2, &card, sizeof(void *)); | ||
3213 | if (!qeth_do_run_thread(card, QETH_RECOVER_THREAD)) | ||
3214 | return 0; | ||
3215 | QETH_DBF_TEXT(TRACE, 2, "recover2"); | ||
3216 | PRINT_WARN("Recovery of device %s started ...\n", | ||
3217 | CARD_BUS_ID(card)); | ||
3218 | card->use_hard_stop = 1; | ||
3219 | __qeth_l3_set_offline(card->gdev, 1); | ||
3220 | rc = __qeth_l3_set_online(card->gdev, 1); | ||
3221 | /* don't run another scheduled recovery */ | ||
3222 | qeth_clear_thread_start_bit(card, QETH_RECOVER_THREAD); | ||
3223 | qeth_clear_thread_running_bit(card, QETH_RECOVER_THREAD); | ||
3224 | if (!rc) | ||
3225 | PRINT_INFO("Device %s successfully recovered!\n", | ||
3226 | CARD_BUS_ID(card)); | ||
3227 | else | ||
3228 | PRINT_INFO("Device %s could not be recovered!\n", | ||
3229 | CARD_BUS_ID(card)); | ||
3230 | return 0; | ||
3231 | } | ||
3232 | |||
3233 | static void qeth_l3_shutdown(struct ccwgroup_device *gdev) | ||
3234 | { | ||
3235 | struct qeth_card *card = dev_get_drvdata(&gdev->dev); | ||
3236 | qeth_l3_clear_ip_list(card, 0, 0); | ||
3237 | qeth_qdio_clear_card(card, 0); | ||
3238 | qeth_clear_qdio_buffers(card); | ||
3239 | } | ||
3240 | |||
3241 | struct ccwgroup_driver qeth_l3_ccwgroup_driver = { | ||
3242 | .probe = qeth_l3_probe_device, | ||
3243 | .remove = qeth_l3_remove_device, | ||
3244 | .set_online = qeth_l3_set_online, | ||
3245 | .set_offline = qeth_l3_set_offline, | ||
3246 | .shutdown = qeth_l3_shutdown, | ||
3247 | }; | ||
3248 | EXPORT_SYMBOL_GPL(qeth_l3_ccwgroup_driver); | ||
3249 | |||
3250 | static int qeth_l3_ip_event(struct notifier_block *this, | ||
3251 | unsigned long event, void *ptr) | ||
3252 | { | ||
3253 | struct in_ifaddr *ifa = (struct in_ifaddr *)ptr; | ||
3254 | struct net_device *dev = (struct net_device *)ifa->ifa_dev->dev; | ||
3255 | struct qeth_ipaddr *addr; | ||
3256 | struct qeth_card *card; | ||
3257 | |||
3258 | if (dev_net(dev) != &init_net) | ||
3259 | return NOTIFY_DONE; | ||
3260 | |||
3261 | QETH_DBF_TEXT(TRACE, 3, "ipevent"); | ||
3262 | card = qeth_l3_get_card_from_dev(dev); | ||
3263 | if (!card) | ||
3264 | return NOTIFY_DONE; | ||
3265 | |||
3266 | addr = qeth_l3_get_addr_buffer(QETH_PROT_IPV4); | ||
3267 | if (addr != NULL) { | ||
3268 | addr->u.a4.addr = ifa->ifa_address; | ||
3269 | addr->u.a4.mask = ifa->ifa_mask; | ||
3270 | addr->type = QETH_IP_TYPE_NORMAL; | ||
3271 | } else | ||
3272 | goto out; | ||
3273 | |||
3274 | switch (event) { | ||
3275 | case NETDEV_UP: | ||
3276 | if (!qeth_l3_add_ip(card, addr)) | ||
3277 | kfree(addr); | ||
3278 | break; | ||
3279 | case NETDEV_DOWN: | ||
3280 | if (!qeth_l3_delete_ip(card, addr)) | ||
3281 | kfree(addr); | ||
3282 | break; | ||
3283 | default: | ||
3284 | break; | ||
3285 | } | ||
3286 | qeth_l3_set_ip_addr_list(card); | ||
3287 | out: | ||
3288 | return NOTIFY_DONE; | ||
3289 | } | ||
3290 | |||
3291 | static struct notifier_block qeth_l3_ip_notifier = { | ||
3292 | qeth_l3_ip_event, | ||
3293 | NULL, | ||
3294 | }; | ||
3295 | |||
3296 | #ifdef CONFIG_QETH_IPV6 | ||
3297 | /** | ||
3298 | * IPv6 event handler | ||
3299 | */ | ||
3300 | static int qeth_l3_ip6_event(struct notifier_block *this, | ||
3301 | unsigned long event, void *ptr) | ||
3302 | { | ||
3303 | struct inet6_ifaddr *ifa = (struct inet6_ifaddr *)ptr; | ||
3304 | struct net_device *dev = (struct net_device *)ifa->idev->dev; | ||
3305 | struct qeth_ipaddr *addr; | ||
3306 | struct qeth_card *card; | ||
3307 | |||
3308 | QETH_DBF_TEXT(TRACE, 3, "ip6event"); | ||
3309 | |||
3310 | card = qeth_l3_get_card_from_dev(dev); | ||
3311 | if (!card) | ||
3312 | return NOTIFY_DONE; | ||
3313 | if (!qeth_is_supported(card, IPA_IPV6)) | ||
3314 | return NOTIFY_DONE; | ||
3315 | |||
3316 | addr = qeth_l3_get_addr_buffer(QETH_PROT_IPV6); | ||
3317 | if (addr != NULL) { | ||
3318 | memcpy(&addr->u.a6.addr, &ifa->addr, sizeof(struct in6_addr)); | ||
3319 | addr->u.a6.pfxlen = ifa->prefix_len; | ||
3320 | addr->type = QETH_IP_TYPE_NORMAL; | ||
3321 | } else | ||
3322 | goto out; | ||
3323 | |||
3324 | switch (event) { | ||
3325 | case NETDEV_UP: | ||
3326 | if (!qeth_l3_add_ip(card, addr)) | ||
3327 | kfree(addr); | ||
3328 | break; | ||
3329 | case NETDEV_DOWN: | ||
3330 | if (!qeth_l3_delete_ip(card, addr)) | ||
3331 | kfree(addr); | ||
3332 | break; | ||
3333 | default: | ||
3334 | break; | ||
3335 | } | ||
3336 | qeth_l3_set_ip_addr_list(card); | ||
3337 | out: | ||
3338 | return NOTIFY_DONE; | ||
3339 | } | ||
3340 | |||
3341 | static struct notifier_block qeth_l3_ip6_notifier = { | ||
3342 | qeth_l3_ip6_event, | ||
3343 | NULL, | ||
3344 | }; | ||
3345 | #endif | ||
3346 | |||
3347 | static int qeth_l3_register_notifiers(void) | ||
3348 | { | ||
3349 | int rc; | ||
3350 | |||
3351 | QETH_DBF_TEXT(TRACE, 5, "regnotif"); | ||
3352 | rc = register_inetaddr_notifier(&qeth_l3_ip_notifier); | ||
3353 | if (rc) | ||
3354 | return rc; | ||
3355 | #ifdef CONFIG_QETH_IPV6 | ||
3356 | rc = register_inet6addr_notifier(&qeth_l3_ip6_notifier); | ||
3357 | if (rc) { | ||
3358 | unregister_inetaddr_notifier(&qeth_l3_ip_notifier); | ||
3359 | return rc; | ||
3360 | } | ||
3361 | #else | ||
3362 | PRINT_WARN("layer 3 discipline no IPv6 support\n"); | ||
3363 | #endif | ||
3364 | return 0; | ||
3365 | } | ||
3366 | |||
3367 | static void qeth_l3_unregister_notifiers(void) | ||
3368 | { | ||
3369 | |||
3370 | QETH_DBF_TEXT(TRACE, 5, "unregnot"); | ||
3371 | BUG_ON(unregister_inetaddr_notifier(&qeth_l3_ip_notifier)); | ||
3372 | #ifdef CONFIG_QETH_IPV6 | ||
3373 | BUG_ON(unregister_inet6addr_notifier(&qeth_l3_ip6_notifier)); | ||
3374 | #endif /* QETH_IPV6 */ | ||
3375 | } | ||
3376 | |||
3377 | static int __init qeth_l3_init(void) | ||
3378 | { | ||
3379 | int rc = 0; | ||
3380 | |||
3381 | PRINT_INFO("register layer 3 discipline\n"); | ||
3382 | rc = qeth_l3_register_notifiers(); | ||
3383 | return rc; | ||
3384 | } | ||
3385 | |||
3386 | static void __exit qeth_l3_exit(void) | ||
3387 | { | ||
3388 | qeth_l3_unregister_notifiers(); | ||
3389 | PRINT_INFO("unregister layer 3 discipline\n"); | ||
3390 | } | ||
3391 | |||
3392 | module_init(qeth_l3_init); | ||
3393 | module_exit(qeth_l3_exit); | ||
3394 | MODULE_AUTHOR("Frank Blaschka <frank.blaschka@de.ibm.com>"); | ||
3395 | MODULE_DESCRIPTION("qeth layer 3 discipline"); | ||
3396 | MODULE_LICENSE("GPL"); | ||
diff --git a/drivers/s390/net/qeth_l3_sys.c b/drivers/s390/net/qeth_l3_sys.c new file mode 100644 index 000000000000..08f51fd902c4 --- /dev/null +++ b/drivers/s390/net/qeth_l3_sys.c | |||
@@ -0,0 +1,1051 @@ | |||
1 | /* | ||
2 | * drivers/s390/net/qeth_l3_sys.c | ||
3 | * | ||
4 | * Copyright IBM Corp. 2007 | ||
5 | * Author(s): Utz Bacher <utz.bacher@de.ibm.com>, | ||
6 | * Frank Pavlic <fpavlic@de.ibm.com>, | ||
7 | * Thomas Spatzier <tspat@de.ibm.com>, | ||
8 | * Frank Blaschka <frank.blaschka@de.ibm.com> | ||
9 | */ | ||
10 | |||
11 | #include "qeth_l3.h" | ||
12 | |||
13 | #define QETH_DEVICE_ATTR(_id, _name, _mode, _show, _store) \ | ||
14 | struct device_attribute dev_attr_##_id = __ATTR(_name, _mode, _show, _store) | ||
15 | |||
16 | static const char *qeth_l3_get_checksum_str(struct qeth_card *card) | ||
17 | { | ||
18 | if (card->options.checksum_type == SW_CHECKSUMMING) | ||
19 | return "sw"; | ||
20 | else if (card->options.checksum_type == HW_CHECKSUMMING) | ||
21 | return "hw"; | ||
22 | else | ||
23 | return "no"; | ||
24 | } | ||
25 | |||
26 | static ssize_t qeth_l3_dev_route_show(struct qeth_card *card, | ||
27 | struct qeth_routing_info *route, char *buf) | ||
28 | { | ||
29 | switch (route->type) { | ||
30 | case PRIMARY_ROUTER: | ||
31 | return sprintf(buf, "%s\n", "primary router"); | ||
32 | case SECONDARY_ROUTER: | ||
33 | return sprintf(buf, "%s\n", "secondary router"); | ||
34 | case MULTICAST_ROUTER: | ||
35 | if (card->info.broadcast_capable == QETH_BROADCAST_WITHOUT_ECHO) | ||
36 | return sprintf(buf, "%s\n", "multicast router+"); | ||
37 | else | ||
38 | return sprintf(buf, "%s\n", "multicast router"); | ||
39 | case PRIMARY_CONNECTOR: | ||
40 | if (card->info.broadcast_capable == QETH_BROADCAST_WITHOUT_ECHO) | ||
41 | return sprintf(buf, "%s\n", "primary connector+"); | ||
42 | else | ||
43 | return sprintf(buf, "%s\n", "primary connector"); | ||
44 | case SECONDARY_CONNECTOR: | ||
45 | if (card->info.broadcast_capable == QETH_BROADCAST_WITHOUT_ECHO) | ||
46 | return sprintf(buf, "%s\n", "secondary connector+"); | ||
47 | else | ||
48 | return sprintf(buf, "%s\n", "secondary connector"); | ||
49 | default: | ||
50 | return sprintf(buf, "%s\n", "no"); | ||
51 | } | ||
52 | } | ||
53 | |||
54 | static ssize_t qeth_l3_dev_route4_show(struct device *dev, | ||
55 | struct device_attribute *attr, char *buf) | ||
56 | { | ||
57 | struct qeth_card *card = dev_get_drvdata(dev); | ||
58 | |||
59 | if (!card) | ||
60 | return -EINVAL; | ||
61 | |||
62 | return qeth_l3_dev_route_show(card, &card->options.route4, buf); | ||
63 | } | ||
64 | |||
65 | static ssize_t qeth_l3_dev_route_store(struct qeth_card *card, | ||
66 | struct qeth_routing_info *route, enum qeth_prot_versions prot, | ||
67 | const char *buf, size_t count) | ||
68 | { | ||
69 | enum qeth_routing_types old_route_type = route->type; | ||
70 | char *tmp; | ||
71 | int rc; | ||
72 | |||
73 | tmp = strsep((char **) &buf, "\n"); | ||
74 | |||
75 | if (!strcmp(tmp, "no_router")) { | ||
76 | route->type = NO_ROUTER; | ||
77 | } else if (!strcmp(tmp, "primary_connector")) { | ||
78 | route->type = PRIMARY_CONNECTOR; | ||
79 | } else if (!strcmp(tmp, "secondary_connector")) { | ||
80 | route->type = SECONDARY_CONNECTOR; | ||
81 | } else if (!strcmp(tmp, "primary_router")) { | ||
82 | route->type = PRIMARY_ROUTER; | ||
83 | } else if (!strcmp(tmp, "secondary_router")) { | ||
84 | route->type = SECONDARY_ROUTER; | ||
85 | } else if (!strcmp(tmp, "multicast_router")) { | ||
86 | route->type = MULTICAST_ROUTER; | ||
87 | } else { | ||
88 | PRINT_WARN("Invalid routing type '%s'.\n", tmp); | ||
89 | return -EINVAL; | ||
90 | } | ||
91 | if (((card->state == CARD_STATE_SOFTSETUP) || | ||
92 | (card->state == CARD_STATE_UP)) && | ||
93 | (old_route_type != route->type)) { | ||
94 | if (prot == QETH_PROT_IPV4) | ||
95 | rc = qeth_l3_setrouting_v4(card); | ||
96 | else if (prot == QETH_PROT_IPV6) | ||
97 | rc = qeth_l3_setrouting_v6(card); | ||
98 | } | ||
99 | return count; | ||
100 | } | ||
101 | |||
102 | static ssize_t qeth_l3_dev_route4_store(struct device *dev, | ||
103 | struct device_attribute *attr, const char *buf, size_t count) | ||
104 | { | ||
105 | struct qeth_card *card = dev_get_drvdata(dev); | ||
106 | |||
107 | if (!card) | ||
108 | return -EINVAL; | ||
109 | |||
110 | return qeth_l3_dev_route_store(card, &card->options.route4, | ||
111 | QETH_PROT_IPV4, buf, count); | ||
112 | } | ||
113 | |||
114 | static DEVICE_ATTR(route4, 0644, qeth_l3_dev_route4_show, | ||
115 | qeth_l3_dev_route4_store); | ||
116 | |||
117 | static ssize_t qeth_l3_dev_route6_show(struct device *dev, | ||
118 | struct device_attribute *attr, char *buf) | ||
119 | { | ||
120 | struct qeth_card *card = dev_get_drvdata(dev); | ||
121 | |||
122 | if (!card) | ||
123 | return -EINVAL; | ||
124 | |||
125 | if (!qeth_is_supported(card, IPA_IPV6)) | ||
126 | return sprintf(buf, "%s\n", "n/a"); | ||
127 | |||
128 | return qeth_l3_dev_route_show(card, &card->options.route6, buf); | ||
129 | } | ||
130 | |||
131 | static ssize_t qeth_l3_dev_route6_store(struct device *dev, | ||
132 | struct device_attribute *attr, const char *buf, size_t count) | ||
133 | { | ||
134 | struct qeth_card *card = dev_get_drvdata(dev); | ||
135 | |||
136 | if (!card) | ||
137 | return -EINVAL; | ||
138 | |||
139 | if (!qeth_is_supported(card, IPA_IPV6)) { | ||
140 | PRINT_WARN("IPv6 not supported for interface %s.\n" | ||
141 | "Routing status no changed.\n", | ||
142 | QETH_CARD_IFNAME(card)); | ||
143 | return -ENOTSUPP; | ||
144 | } | ||
145 | |||
146 | return qeth_l3_dev_route_store(card, &card->options.route6, | ||
147 | QETH_PROT_IPV6, buf, count); | ||
148 | } | ||
149 | |||
150 | static DEVICE_ATTR(route6, 0644, qeth_l3_dev_route6_show, | ||
151 | qeth_l3_dev_route6_store); | ||
152 | |||
153 | static ssize_t qeth_l3_dev_fake_broadcast_show(struct device *dev, | ||
154 | struct device_attribute *attr, char *buf) | ||
155 | { | ||
156 | struct qeth_card *card = dev_get_drvdata(dev); | ||
157 | |||
158 | if (!card) | ||
159 | return -EINVAL; | ||
160 | |||
161 | return sprintf(buf, "%i\n", card->options.fake_broadcast? 1:0); | ||
162 | } | ||
163 | |||
164 | static ssize_t qeth_l3_dev_fake_broadcast_store(struct device *dev, | ||
165 | struct device_attribute *attr, const char *buf, size_t count) | ||
166 | { | ||
167 | struct qeth_card *card = dev_get_drvdata(dev); | ||
168 | char *tmp; | ||
169 | int i; | ||
170 | |||
171 | if (!card) | ||
172 | return -EINVAL; | ||
173 | |||
174 | if ((card->state != CARD_STATE_DOWN) && | ||
175 | (card->state != CARD_STATE_RECOVER)) | ||
176 | return -EPERM; | ||
177 | |||
178 | i = simple_strtoul(buf, &tmp, 16); | ||
179 | if ((i == 0) || (i == 1)) | ||
180 | card->options.fake_broadcast = i; | ||
181 | else { | ||
182 | PRINT_WARN("fake_broadcast: write 0 or 1 to this file!\n"); | ||
183 | return -EINVAL; | ||
184 | } | ||
185 | return count; | ||
186 | } | ||
187 | |||
188 | static DEVICE_ATTR(fake_broadcast, 0644, qeth_l3_dev_fake_broadcast_show, | ||
189 | qeth_l3_dev_fake_broadcast_store); | ||
190 | |||
191 | static ssize_t qeth_l3_dev_broadcast_mode_show(struct device *dev, | ||
192 | struct device_attribute *attr, char *buf) | ||
193 | { | ||
194 | struct qeth_card *card = dev_get_drvdata(dev); | ||
195 | |||
196 | if (!card) | ||
197 | return -EINVAL; | ||
198 | |||
199 | if (!((card->info.link_type == QETH_LINK_TYPE_HSTR) || | ||
200 | (card->info.link_type == QETH_LINK_TYPE_LANE_TR))) | ||
201 | return sprintf(buf, "n/a\n"); | ||
202 | |||
203 | return sprintf(buf, "%s\n", (card->options.broadcast_mode == | ||
204 | QETH_TR_BROADCAST_ALLRINGS)? | ||
205 | "all rings":"local"); | ||
206 | } | ||
207 | |||
208 | static ssize_t qeth_l3_dev_broadcast_mode_store(struct device *dev, | ||
209 | struct device_attribute *attr, const char *buf, size_t count) | ||
210 | { | ||
211 | struct qeth_card *card = dev_get_drvdata(dev); | ||
212 | char *tmp; | ||
213 | |||
214 | if (!card) | ||
215 | return -EINVAL; | ||
216 | |||
217 | if ((card->state != CARD_STATE_DOWN) && | ||
218 | (card->state != CARD_STATE_RECOVER)) | ||
219 | return -EPERM; | ||
220 | |||
221 | if (!((card->info.link_type == QETH_LINK_TYPE_HSTR) || | ||
222 | (card->info.link_type == QETH_LINK_TYPE_LANE_TR))) { | ||
223 | PRINT_WARN("Device is not a tokenring device!\n"); | ||
224 | return -EINVAL; | ||
225 | } | ||
226 | |||
227 | tmp = strsep((char **) &buf, "\n"); | ||
228 | |||
229 | if (!strcmp(tmp, "local")) { | ||
230 | card->options.broadcast_mode = QETH_TR_BROADCAST_LOCAL; | ||
231 | return count; | ||
232 | } else if (!strcmp(tmp, "all_rings")) { | ||
233 | card->options.broadcast_mode = QETH_TR_BROADCAST_ALLRINGS; | ||
234 | return count; | ||
235 | } else { | ||
236 | PRINT_WARN("broadcast_mode: invalid mode %s!\n", | ||
237 | tmp); | ||
238 | return -EINVAL; | ||
239 | } | ||
240 | return count; | ||
241 | } | ||
242 | |||
243 | static DEVICE_ATTR(broadcast_mode, 0644, qeth_l3_dev_broadcast_mode_show, | ||
244 | qeth_l3_dev_broadcast_mode_store); | ||
245 | |||
246 | static ssize_t qeth_l3_dev_canonical_macaddr_show(struct device *dev, | ||
247 | struct device_attribute *attr, char *buf) | ||
248 | { | ||
249 | struct qeth_card *card = dev_get_drvdata(dev); | ||
250 | |||
251 | if (!card) | ||
252 | return -EINVAL; | ||
253 | |||
254 | if (!((card->info.link_type == QETH_LINK_TYPE_HSTR) || | ||
255 | (card->info.link_type == QETH_LINK_TYPE_LANE_TR))) | ||
256 | return sprintf(buf, "n/a\n"); | ||
257 | |||
258 | return sprintf(buf, "%i\n", (card->options.macaddr_mode == | ||
259 | QETH_TR_MACADDR_CANONICAL)? 1:0); | ||
260 | } | ||
261 | |||
262 | static ssize_t qeth_l3_dev_canonical_macaddr_store(struct device *dev, | ||
263 | struct device_attribute *attr, const char *buf, size_t count) | ||
264 | { | ||
265 | struct qeth_card *card = dev_get_drvdata(dev); | ||
266 | char *tmp; | ||
267 | int i; | ||
268 | |||
269 | if (!card) | ||
270 | return -EINVAL; | ||
271 | |||
272 | if ((card->state != CARD_STATE_DOWN) && | ||
273 | (card->state != CARD_STATE_RECOVER)) | ||
274 | return -EPERM; | ||
275 | |||
276 | if (!((card->info.link_type == QETH_LINK_TYPE_HSTR) || | ||
277 | (card->info.link_type == QETH_LINK_TYPE_LANE_TR))) { | ||
278 | PRINT_WARN("Device is not a tokenring device!\n"); | ||
279 | return -EINVAL; | ||
280 | } | ||
281 | |||
282 | i = simple_strtoul(buf, &tmp, 16); | ||
283 | if ((i == 0) || (i == 1)) | ||
284 | card->options.macaddr_mode = i? | ||
285 | QETH_TR_MACADDR_CANONICAL : | ||
286 | QETH_TR_MACADDR_NONCANONICAL; | ||
287 | else { | ||
288 | PRINT_WARN("canonical_macaddr: write 0 or 1 to this file!\n"); | ||
289 | return -EINVAL; | ||
290 | } | ||
291 | return count; | ||
292 | } | ||
293 | |||
294 | static DEVICE_ATTR(canonical_macaddr, 0644, qeth_l3_dev_canonical_macaddr_show, | ||
295 | qeth_l3_dev_canonical_macaddr_store); | ||
296 | |||
297 | static ssize_t qeth_l3_dev_checksum_show(struct device *dev, | ||
298 | struct device_attribute *attr, char *buf) | ||
299 | { | ||
300 | struct qeth_card *card = dev_get_drvdata(dev); | ||
301 | |||
302 | if (!card) | ||
303 | return -EINVAL; | ||
304 | |||
305 | return sprintf(buf, "%s checksumming\n", | ||
306 | qeth_l3_get_checksum_str(card)); | ||
307 | } | ||
308 | |||
309 | static ssize_t qeth_l3_dev_checksum_store(struct device *dev, | ||
310 | struct device_attribute *attr, const char *buf, size_t count) | ||
311 | { | ||
312 | struct qeth_card *card = dev_get_drvdata(dev); | ||
313 | char *tmp; | ||
314 | |||
315 | if (!card) | ||
316 | return -EINVAL; | ||
317 | |||
318 | if ((card->state != CARD_STATE_DOWN) && | ||
319 | (card->state != CARD_STATE_RECOVER)) | ||
320 | return -EPERM; | ||
321 | |||
322 | tmp = strsep((char **) &buf, "\n"); | ||
323 | if (!strcmp(tmp, "sw_checksumming")) | ||
324 | card->options.checksum_type = SW_CHECKSUMMING; | ||
325 | else if (!strcmp(tmp, "hw_checksumming")) | ||
326 | card->options.checksum_type = HW_CHECKSUMMING; | ||
327 | else if (!strcmp(tmp, "no_checksumming")) | ||
328 | card->options.checksum_type = NO_CHECKSUMMING; | ||
329 | else { | ||
330 | PRINT_WARN("Unknown checksumming type '%s'\n", tmp); | ||
331 | return -EINVAL; | ||
332 | } | ||
333 | return count; | ||
334 | } | ||
335 | |||
336 | static DEVICE_ATTR(checksumming, 0644, qeth_l3_dev_checksum_show, | ||
337 | qeth_l3_dev_checksum_store); | ||
338 | |||
339 | static struct attribute *qeth_l3_device_attrs[] = { | ||
340 | &dev_attr_route4.attr, | ||
341 | &dev_attr_route6.attr, | ||
342 | &dev_attr_fake_broadcast.attr, | ||
343 | &dev_attr_broadcast_mode.attr, | ||
344 | &dev_attr_canonical_macaddr.attr, | ||
345 | &dev_attr_checksumming.attr, | ||
346 | NULL, | ||
347 | }; | ||
348 | |||
349 | static struct attribute_group qeth_l3_device_attr_group = { | ||
350 | .attrs = qeth_l3_device_attrs, | ||
351 | }; | ||
352 | |||
353 | static ssize_t qeth_l3_dev_ipato_enable_show(struct device *dev, | ||
354 | struct device_attribute *attr, char *buf) | ||
355 | { | ||
356 | struct qeth_card *card = dev_get_drvdata(dev); | ||
357 | |||
358 | if (!card) | ||
359 | return -EINVAL; | ||
360 | |||
361 | return sprintf(buf, "%i\n", card->ipato.enabled? 1:0); | ||
362 | } | ||
363 | |||
364 | static ssize_t qeth_l3_dev_ipato_enable_store(struct device *dev, | ||
365 | struct device_attribute *attr, const char *buf, size_t count) | ||
366 | { | ||
367 | struct qeth_card *card = dev_get_drvdata(dev); | ||
368 | char *tmp; | ||
369 | |||
370 | if (!card) | ||
371 | return -EINVAL; | ||
372 | |||
373 | if ((card->state != CARD_STATE_DOWN) && | ||
374 | (card->state != CARD_STATE_RECOVER)) | ||
375 | return -EPERM; | ||
376 | |||
377 | tmp = strsep((char **) &buf, "\n"); | ||
378 | if (!strcmp(tmp, "toggle")) { | ||
379 | card->ipato.enabled = (card->ipato.enabled)? 0 : 1; | ||
380 | } else if (!strcmp(tmp, "1")) { | ||
381 | card->ipato.enabled = 1; | ||
382 | } else if (!strcmp(tmp, "0")) { | ||
383 | card->ipato.enabled = 0; | ||
384 | } else { | ||
385 | PRINT_WARN("ipato_enable: write 0, 1 or 'toggle' to " | ||
386 | "this file\n"); | ||
387 | return -EINVAL; | ||
388 | } | ||
389 | return count; | ||
390 | } | ||
391 | |||
392 | static QETH_DEVICE_ATTR(ipato_enable, enable, 0644, | ||
393 | qeth_l3_dev_ipato_enable_show, | ||
394 | qeth_l3_dev_ipato_enable_store); | ||
395 | |||
396 | static ssize_t qeth_l3_dev_ipato_invert4_show(struct device *dev, | ||
397 | struct device_attribute *attr, char *buf) | ||
398 | { | ||
399 | struct qeth_card *card = dev_get_drvdata(dev); | ||
400 | |||
401 | if (!card) | ||
402 | return -EINVAL; | ||
403 | |||
404 | return sprintf(buf, "%i\n", card->ipato.invert4? 1:0); | ||
405 | } | ||
406 | |||
407 | static ssize_t qeth_l3_dev_ipato_invert4_store(struct device *dev, | ||
408 | struct device_attribute *attr, | ||
409 | const char *buf, size_t count) | ||
410 | { | ||
411 | struct qeth_card *card = dev_get_drvdata(dev); | ||
412 | char *tmp; | ||
413 | |||
414 | if (!card) | ||
415 | return -EINVAL; | ||
416 | |||
417 | tmp = strsep((char **) &buf, "\n"); | ||
418 | if (!strcmp(tmp, "toggle")) { | ||
419 | card->ipato.invert4 = (card->ipato.invert4)? 0 : 1; | ||
420 | } else if (!strcmp(tmp, "1")) { | ||
421 | card->ipato.invert4 = 1; | ||
422 | } else if (!strcmp(tmp, "0")) { | ||
423 | card->ipato.invert4 = 0; | ||
424 | } else { | ||
425 | PRINT_WARN("ipato_invert4: write 0, 1 or 'toggle' to " | ||
426 | "this file\n"); | ||
427 | return -EINVAL; | ||
428 | } | ||
429 | return count; | ||
430 | } | ||
431 | |||
432 | static QETH_DEVICE_ATTR(ipato_invert4, invert4, 0644, | ||
433 | qeth_l3_dev_ipato_invert4_show, | ||
434 | qeth_l3_dev_ipato_invert4_store); | ||
435 | |||
436 | static ssize_t qeth_l3_dev_ipato_add_show(char *buf, struct qeth_card *card, | ||
437 | enum qeth_prot_versions proto) | ||
438 | { | ||
439 | struct qeth_ipato_entry *ipatoe; | ||
440 | unsigned long flags; | ||
441 | char addr_str[40]; | ||
442 | int entry_len; /* length of 1 entry string, differs between v4 and v6 */ | ||
443 | int i = 0; | ||
444 | |||
445 | entry_len = (proto == QETH_PROT_IPV4)? 12 : 40; | ||
446 | /* add strlen for "/<mask>\n" */ | ||
447 | entry_len += (proto == QETH_PROT_IPV4)? 5 : 6; | ||
448 | spin_lock_irqsave(&card->ip_lock, flags); | ||
449 | list_for_each_entry(ipatoe, &card->ipato.entries, entry) { | ||
450 | if (ipatoe->proto != proto) | ||
451 | continue; | ||
452 | /* String must not be longer than PAGE_SIZE. So we check if | ||
453 | * string length gets near PAGE_SIZE. Then we can savely display | ||
454 | * the next IPv6 address (worst case, compared to IPv4) */ | ||
455 | if ((PAGE_SIZE - i) <= entry_len) | ||
456 | break; | ||
457 | qeth_l3_ipaddr_to_string(proto, ipatoe->addr, addr_str); | ||
458 | i += snprintf(buf + i, PAGE_SIZE - i, | ||
459 | "%s/%i\n", addr_str, ipatoe->mask_bits); | ||
460 | } | ||
461 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
462 | i += snprintf(buf + i, PAGE_SIZE - i, "\n"); | ||
463 | |||
464 | return i; | ||
465 | } | ||
466 | |||
467 | static ssize_t qeth_l3_dev_ipato_add4_show(struct device *dev, | ||
468 | struct device_attribute *attr, char *buf) | ||
469 | { | ||
470 | struct qeth_card *card = dev_get_drvdata(dev); | ||
471 | |||
472 | if (!card) | ||
473 | return -EINVAL; | ||
474 | |||
475 | return qeth_l3_dev_ipato_add_show(buf, card, QETH_PROT_IPV4); | ||
476 | } | ||
477 | |||
478 | static int qeth_l3_parse_ipatoe(const char *buf, enum qeth_prot_versions proto, | ||
479 | u8 *addr, int *mask_bits) | ||
480 | { | ||
481 | const char *start, *end; | ||
482 | char *tmp; | ||
483 | char buffer[40] = {0, }; | ||
484 | |||
485 | start = buf; | ||
486 | /* get address string */ | ||
487 | end = strchr(start, '/'); | ||
488 | if (!end || (end - start >= 40)) { | ||
489 | PRINT_WARN("Invalid format for ipato_addx/delx. " | ||
490 | "Use <ip addr>/<mask bits>\n"); | ||
491 | return -EINVAL; | ||
492 | } | ||
493 | strncpy(buffer, start, end - start); | ||
494 | if (qeth_l3_string_to_ipaddr(buffer, proto, addr)) { | ||
495 | PRINT_WARN("Invalid IP address format!\n"); | ||
496 | return -EINVAL; | ||
497 | } | ||
498 | start = end + 1; | ||
499 | *mask_bits = simple_strtoul(start, &tmp, 10); | ||
500 | if (!strlen(start) || | ||
501 | (tmp == start) || | ||
502 | (*mask_bits > ((proto == QETH_PROT_IPV4) ? 32 : 128))) { | ||
503 | PRINT_WARN("Invalid mask bits for ipato_addx/delx !\n"); | ||
504 | return -EINVAL; | ||
505 | } | ||
506 | return 0; | ||
507 | } | ||
508 | |||
509 | static ssize_t qeth_l3_dev_ipato_add_store(const char *buf, size_t count, | ||
510 | struct qeth_card *card, enum qeth_prot_versions proto) | ||
511 | { | ||
512 | struct qeth_ipato_entry *ipatoe; | ||
513 | u8 addr[16]; | ||
514 | int mask_bits; | ||
515 | int rc; | ||
516 | |||
517 | rc = qeth_l3_parse_ipatoe(buf, proto, addr, &mask_bits); | ||
518 | if (rc) | ||
519 | return rc; | ||
520 | |||
521 | ipatoe = kzalloc(sizeof(struct qeth_ipato_entry), GFP_KERNEL); | ||
522 | if (!ipatoe) { | ||
523 | PRINT_WARN("No memory to allocate ipato entry\n"); | ||
524 | return -ENOMEM; | ||
525 | } | ||
526 | ipatoe->proto = proto; | ||
527 | memcpy(ipatoe->addr, addr, (proto == QETH_PROT_IPV4)? 4:16); | ||
528 | ipatoe->mask_bits = mask_bits; | ||
529 | |||
530 | rc = qeth_l3_add_ipato_entry(card, ipatoe); | ||
531 | if (rc) { | ||
532 | kfree(ipatoe); | ||
533 | return rc; | ||
534 | } | ||
535 | |||
536 | return count; | ||
537 | } | ||
538 | |||
539 | static ssize_t qeth_l3_dev_ipato_add4_store(struct device *dev, | ||
540 | struct device_attribute *attr, const char *buf, size_t count) | ||
541 | { | ||
542 | struct qeth_card *card = dev_get_drvdata(dev); | ||
543 | |||
544 | if (!card) | ||
545 | return -EINVAL; | ||
546 | |||
547 | return qeth_l3_dev_ipato_add_store(buf, count, card, QETH_PROT_IPV4); | ||
548 | } | ||
549 | |||
550 | static QETH_DEVICE_ATTR(ipato_add4, add4, 0644, | ||
551 | qeth_l3_dev_ipato_add4_show, | ||
552 | qeth_l3_dev_ipato_add4_store); | ||
553 | |||
554 | static ssize_t qeth_l3_dev_ipato_del_store(const char *buf, size_t count, | ||
555 | struct qeth_card *card, enum qeth_prot_versions proto) | ||
556 | { | ||
557 | u8 addr[16]; | ||
558 | int mask_bits; | ||
559 | int rc; | ||
560 | |||
561 | rc = qeth_l3_parse_ipatoe(buf, proto, addr, &mask_bits); | ||
562 | if (rc) | ||
563 | return rc; | ||
564 | |||
565 | qeth_l3_del_ipato_entry(card, proto, addr, mask_bits); | ||
566 | |||
567 | return count; | ||
568 | } | ||
569 | |||
570 | static ssize_t qeth_l3_dev_ipato_del4_store(struct device *dev, | ||
571 | struct device_attribute *attr, const char *buf, size_t count) | ||
572 | { | ||
573 | struct qeth_card *card = dev_get_drvdata(dev); | ||
574 | |||
575 | if (!card) | ||
576 | return -EINVAL; | ||
577 | |||
578 | return qeth_l3_dev_ipato_del_store(buf, count, card, QETH_PROT_IPV4); | ||
579 | } | ||
580 | |||
581 | static QETH_DEVICE_ATTR(ipato_del4, del4, 0200, NULL, | ||
582 | qeth_l3_dev_ipato_del4_store); | ||
583 | |||
584 | static ssize_t qeth_l3_dev_ipato_invert6_show(struct device *dev, | ||
585 | struct device_attribute *attr, char *buf) | ||
586 | { | ||
587 | struct qeth_card *card = dev_get_drvdata(dev); | ||
588 | |||
589 | if (!card) | ||
590 | return -EINVAL; | ||
591 | |||
592 | return sprintf(buf, "%i\n", card->ipato.invert6? 1:0); | ||
593 | } | ||
594 | |||
595 | static ssize_t qeth_l3_dev_ipato_invert6_store(struct device *dev, | ||
596 | struct device_attribute *attr, const char *buf, size_t count) | ||
597 | { | ||
598 | struct qeth_card *card = dev_get_drvdata(dev); | ||
599 | char *tmp; | ||
600 | |||
601 | if (!card) | ||
602 | return -EINVAL; | ||
603 | |||
604 | tmp = strsep((char **) &buf, "\n"); | ||
605 | if (!strcmp(tmp, "toggle")) { | ||
606 | card->ipato.invert6 = (card->ipato.invert6)? 0 : 1; | ||
607 | } else if (!strcmp(tmp, "1")) { | ||
608 | card->ipato.invert6 = 1; | ||
609 | } else if (!strcmp(tmp, "0")) { | ||
610 | card->ipato.invert6 = 0; | ||
611 | } else { | ||
612 | PRINT_WARN("ipato_invert6: write 0, 1 or 'toggle' to " | ||
613 | "this file\n"); | ||
614 | return -EINVAL; | ||
615 | } | ||
616 | return count; | ||
617 | } | ||
618 | |||
619 | static QETH_DEVICE_ATTR(ipato_invert6, invert6, 0644, | ||
620 | qeth_l3_dev_ipato_invert6_show, | ||
621 | qeth_l3_dev_ipato_invert6_store); | ||
622 | |||
623 | |||
624 | static ssize_t qeth_l3_dev_ipato_add6_show(struct device *dev, | ||
625 | struct device_attribute *attr, char *buf) | ||
626 | { | ||
627 | struct qeth_card *card = dev_get_drvdata(dev); | ||
628 | |||
629 | if (!card) | ||
630 | return -EINVAL; | ||
631 | |||
632 | return qeth_l3_dev_ipato_add_show(buf, card, QETH_PROT_IPV6); | ||
633 | } | ||
634 | |||
635 | static ssize_t qeth_l3_dev_ipato_add6_store(struct device *dev, | ||
636 | struct device_attribute *attr, const char *buf, size_t count) | ||
637 | { | ||
638 | struct qeth_card *card = dev_get_drvdata(dev); | ||
639 | |||
640 | if (!card) | ||
641 | return -EINVAL; | ||
642 | |||
643 | return qeth_l3_dev_ipato_add_store(buf, count, card, QETH_PROT_IPV6); | ||
644 | } | ||
645 | |||
646 | static QETH_DEVICE_ATTR(ipato_add6, add6, 0644, | ||
647 | qeth_l3_dev_ipato_add6_show, | ||
648 | qeth_l3_dev_ipato_add6_store); | ||
649 | |||
650 | static ssize_t qeth_l3_dev_ipato_del6_store(struct device *dev, | ||
651 | struct device_attribute *attr, const char *buf, size_t count) | ||
652 | { | ||
653 | struct qeth_card *card = dev_get_drvdata(dev); | ||
654 | |||
655 | if (!card) | ||
656 | return -EINVAL; | ||
657 | |||
658 | return qeth_l3_dev_ipato_del_store(buf, count, card, QETH_PROT_IPV6); | ||
659 | } | ||
660 | |||
661 | static QETH_DEVICE_ATTR(ipato_del6, del6, 0200, NULL, | ||
662 | qeth_l3_dev_ipato_del6_store); | ||
663 | |||
664 | static struct attribute *qeth_ipato_device_attrs[] = { | ||
665 | &dev_attr_ipato_enable.attr, | ||
666 | &dev_attr_ipato_invert4.attr, | ||
667 | &dev_attr_ipato_add4.attr, | ||
668 | &dev_attr_ipato_del4.attr, | ||
669 | &dev_attr_ipato_invert6.attr, | ||
670 | &dev_attr_ipato_add6.attr, | ||
671 | &dev_attr_ipato_del6.attr, | ||
672 | NULL, | ||
673 | }; | ||
674 | |||
675 | static struct attribute_group qeth_device_ipato_group = { | ||
676 | .name = "ipa_takeover", | ||
677 | .attrs = qeth_ipato_device_attrs, | ||
678 | }; | ||
679 | |||
680 | static ssize_t qeth_l3_dev_vipa_add_show(char *buf, struct qeth_card *card, | ||
681 | enum qeth_prot_versions proto) | ||
682 | { | ||
683 | struct qeth_ipaddr *ipaddr; | ||
684 | char addr_str[40]; | ||
685 | int entry_len; /* length of 1 entry string, differs between v4 and v6 */ | ||
686 | unsigned long flags; | ||
687 | int i = 0; | ||
688 | |||
689 | entry_len = (proto == QETH_PROT_IPV4)? 12 : 40; | ||
690 | entry_len += 2; /* \n + terminator */ | ||
691 | spin_lock_irqsave(&card->ip_lock, flags); | ||
692 | list_for_each_entry(ipaddr, &card->ip_list, entry) { | ||
693 | if (ipaddr->proto != proto) | ||
694 | continue; | ||
695 | if (ipaddr->type != QETH_IP_TYPE_VIPA) | ||
696 | continue; | ||
697 | /* String must not be longer than PAGE_SIZE. So we check if | ||
698 | * string length gets near PAGE_SIZE. Then we can savely display | ||
699 | * the next IPv6 address (worst case, compared to IPv4) */ | ||
700 | if ((PAGE_SIZE - i) <= entry_len) | ||
701 | break; | ||
702 | qeth_l3_ipaddr_to_string(proto, (const u8 *)&ipaddr->u, | ||
703 | addr_str); | ||
704 | i += snprintf(buf + i, PAGE_SIZE - i, "%s\n", addr_str); | ||
705 | } | ||
706 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
707 | i += snprintf(buf + i, PAGE_SIZE - i, "\n"); | ||
708 | |||
709 | return i; | ||
710 | } | ||
711 | |||
712 | static ssize_t qeth_l3_dev_vipa_add4_show(struct device *dev, | ||
713 | struct device_attribute *attr, char *buf) | ||
714 | { | ||
715 | struct qeth_card *card = dev_get_drvdata(dev); | ||
716 | |||
717 | if (!card) | ||
718 | return -EINVAL; | ||
719 | |||
720 | return qeth_l3_dev_vipa_add_show(buf, card, QETH_PROT_IPV4); | ||
721 | } | ||
722 | |||
723 | static int qeth_l3_parse_vipae(const char *buf, enum qeth_prot_versions proto, | ||
724 | u8 *addr) | ||
725 | { | ||
726 | if (qeth_l3_string_to_ipaddr(buf, proto, addr)) { | ||
727 | PRINT_WARN("Invalid IP address format!\n"); | ||
728 | return -EINVAL; | ||
729 | } | ||
730 | return 0; | ||
731 | } | ||
732 | |||
733 | static ssize_t qeth_l3_dev_vipa_add_store(const char *buf, size_t count, | ||
734 | struct qeth_card *card, enum qeth_prot_versions proto) | ||
735 | { | ||
736 | u8 addr[16] = {0, }; | ||
737 | int rc; | ||
738 | |||
739 | rc = qeth_l3_parse_vipae(buf, proto, addr); | ||
740 | if (rc) | ||
741 | return rc; | ||
742 | |||
743 | rc = qeth_l3_add_vipa(card, proto, addr); | ||
744 | if (rc) | ||
745 | return rc; | ||
746 | |||
747 | return count; | ||
748 | } | ||
749 | |||
750 | static ssize_t qeth_l3_dev_vipa_add4_store(struct device *dev, | ||
751 | struct device_attribute *attr, const char *buf, size_t count) | ||
752 | { | ||
753 | struct qeth_card *card = dev_get_drvdata(dev); | ||
754 | |||
755 | if (!card) | ||
756 | return -EINVAL; | ||
757 | |||
758 | return qeth_l3_dev_vipa_add_store(buf, count, card, QETH_PROT_IPV4); | ||
759 | } | ||
760 | |||
761 | static QETH_DEVICE_ATTR(vipa_add4, add4, 0644, | ||
762 | qeth_l3_dev_vipa_add4_show, | ||
763 | qeth_l3_dev_vipa_add4_store); | ||
764 | |||
765 | static ssize_t qeth_l3_dev_vipa_del_store(const char *buf, size_t count, | ||
766 | struct qeth_card *card, enum qeth_prot_versions proto) | ||
767 | { | ||
768 | u8 addr[16]; | ||
769 | int rc; | ||
770 | |||
771 | rc = qeth_l3_parse_vipae(buf, proto, addr); | ||
772 | if (rc) | ||
773 | return rc; | ||
774 | |||
775 | qeth_l3_del_vipa(card, proto, addr); | ||
776 | |||
777 | return count; | ||
778 | } | ||
779 | |||
780 | static ssize_t qeth_l3_dev_vipa_del4_store(struct device *dev, | ||
781 | struct device_attribute *attr, const char *buf, size_t count) | ||
782 | { | ||
783 | struct qeth_card *card = dev_get_drvdata(dev); | ||
784 | |||
785 | if (!card) | ||
786 | return -EINVAL; | ||
787 | |||
788 | return qeth_l3_dev_vipa_del_store(buf, count, card, QETH_PROT_IPV4); | ||
789 | } | ||
790 | |||
791 | static QETH_DEVICE_ATTR(vipa_del4, del4, 0200, NULL, | ||
792 | qeth_l3_dev_vipa_del4_store); | ||
793 | |||
794 | static ssize_t qeth_l3_dev_vipa_add6_show(struct device *dev, | ||
795 | struct device_attribute *attr, char *buf) | ||
796 | { | ||
797 | struct qeth_card *card = dev_get_drvdata(dev); | ||
798 | |||
799 | if (!card) | ||
800 | return -EINVAL; | ||
801 | |||
802 | return qeth_l3_dev_vipa_add_show(buf, card, QETH_PROT_IPV6); | ||
803 | } | ||
804 | |||
805 | static ssize_t qeth_l3_dev_vipa_add6_store(struct device *dev, | ||
806 | struct device_attribute *attr, const char *buf, size_t count) | ||
807 | { | ||
808 | struct qeth_card *card = dev_get_drvdata(dev); | ||
809 | |||
810 | if (!card) | ||
811 | return -EINVAL; | ||
812 | |||
813 | return qeth_l3_dev_vipa_add_store(buf, count, card, QETH_PROT_IPV6); | ||
814 | } | ||
815 | |||
816 | static QETH_DEVICE_ATTR(vipa_add6, add6, 0644, | ||
817 | qeth_l3_dev_vipa_add6_show, | ||
818 | qeth_l3_dev_vipa_add6_store); | ||
819 | |||
820 | static ssize_t qeth_l3_dev_vipa_del6_store(struct device *dev, | ||
821 | struct device_attribute *attr, const char *buf, size_t count) | ||
822 | { | ||
823 | struct qeth_card *card = dev_get_drvdata(dev); | ||
824 | |||
825 | if (!card) | ||
826 | return -EINVAL; | ||
827 | |||
828 | return qeth_l3_dev_vipa_del_store(buf, count, card, QETH_PROT_IPV6); | ||
829 | } | ||
830 | |||
831 | static QETH_DEVICE_ATTR(vipa_del6, del6, 0200, NULL, | ||
832 | qeth_l3_dev_vipa_del6_store); | ||
833 | |||
834 | static struct attribute *qeth_vipa_device_attrs[] = { | ||
835 | &dev_attr_vipa_add4.attr, | ||
836 | &dev_attr_vipa_del4.attr, | ||
837 | &dev_attr_vipa_add6.attr, | ||
838 | &dev_attr_vipa_del6.attr, | ||
839 | NULL, | ||
840 | }; | ||
841 | |||
842 | static struct attribute_group qeth_device_vipa_group = { | ||
843 | .name = "vipa", | ||
844 | .attrs = qeth_vipa_device_attrs, | ||
845 | }; | ||
846 | |||
847 | static ssize_t qeth_l3_dev_rxip_add_show(char *buf, struct qeth_card *card, | ||
848 | enum qeth_prot_versions proto) | ||
849 | { | ||
850 | struct qeth_ipaddr *ipaddr; | ||
851 | char addr_str[40]; | ||
852 | int entry_len; /* length of 1 entry string, differs between v4 and v6 */ | ||
853 | unsigned long flags; | ||
854 | int i = 0; | ||
855 | |||
856 | entry_len = (proto == QETH_PROT_IPV4)? 12 : 40; | ||
857 | entry_len += 2; /* \n + terminator */ | ||
858 | spin_lock_irqsave(&card->ip_lock, flags); | ||
859 | list_for_each_entry(ipaddr, &card->ip_list, entry) { | ||
860 | if (ipaddr->proto != proto) | ||
861 | continue; | ||
862 | if (ipaddr->type != QETH_IP_TYPE_RXIP) | ||
863 | continue; | ||
864 | /* String must not be longer than PAGE_SIZE. So we check if | ||
865 | * string length gets near PAGE_SIZE. Then we can savely display | ||
866 | * the next IPv6 address (worst case, compared to IPv4) */ | ||
867 | if ((PAGE_SIZE - i) <= entry_len) | ||
868 | break; | ||
869 | qeth_l3_ipaddr_to_string(proto, (const u8 *)&ipaddr->u, | ||
870 | addr_str); | ||
871 | i += snprintf(buf + i, PAGE_SIZE - i, "%s\n", addr_str); | ||
872 | } | ||
873 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
874 | i += snprintf(buf + i, PAGE_SIZE - i, "\n"); | ||
875 | |||
876 | return i; | ||
877 | } | ||
878 | |||
879 | static ssize_t qeth_l3_dev_rxip_add4_show(struct device *dev, | ||
880 | struct device_attribute *attr, char *buf) | ||
881 | { | ||
882 | struct qeth_card *card = dev_get_drvdata(dev); | ||
883 | |||
884 | if (!card) | ||
885 | return -EINVAL; | ||
886 | |||
887 | return qeth_l3_dev_rxip_add_show(buf, card, QETH_PROT_IPV4); | ||
888 | } | ||
889 | |||
890 | static int qeth_l3_parse_rxipe(const char *buf, enum qeth_prot_versions proto, | ||
891 | u8 *addr) | ||
892 | { | ||
893 | if (qeth_l3_string_to_ipaddr(buf, proto, addr)) { | ||
894 | PRINT_WARN("Invalid IP address format!\n"); | ||
895 | return -EINVAL; | ||
896 | } | ||
897 | return 0; | ||
898 | } | ||
899 | |||
900 | static ssize_t qeth_l3_dev_rxip_add_store(const char *buf, size_t count, | ||
901 | struct qeth_card *card, enum qeth_prot_versions proto) | ||
902 | { | ||
903 | u8 addr[16] = {0, }; | ||
904 | int rc; | ||
905 | |||
906 | rc = qeth_l3_parse_rxipe(buf, proto, addr); | ||
907 | if (rc) | ||
908 | return rc; | ||
909 | |||
910 | rc = qeth_l3_add_rxip(card, proto, addr); | ||
911 | if (rc) | ||
912 | return rc; | ||
913 | |||
914 | return count; | ||
915 | } | ||
916 | |||
917 | static ssize_t qeth_l3_dev_rxip_add4_store(struct device *dev, | ||
918 | struct device_attribute *attr, const char *buf, size_t count) | ||
919 | { | ||
920 | struct qeth_card *card = dev_get_drvdata(dev); | ||
921 | |||
922 | if (!card) | ||
923 | return -EINVAL; | ||
924 | |||
925 | return qeth_l3_dev_rxip_add_store(buf, count, card, QETH_PROT_IPV4); | ||
926 | } | ||
927 | |||
928 | static QETH_DEVICE_ATTR(rxip_add4, add4, 0644, | ||
929 | qeth_l3_dev_rxip_add4_show, | ||
930 | qeth_l3_dev_rxip_add4_store); | ||
931 | |||
932 | static ssize_t qeth_l3_dev_rxip_del_store(const char *buf, size_t count, | ||
933 | struct qeth_card *card, enum qeth_prot_versions proto) | ||
934 | { | ||
935 | u8 addr[16]; | ||
936 | int rc; | ||
937 | |||
938 | rc = qeth_l3_parse_rxipe(buf, proto, addr); | ||
939 | if (rc) | ||
940 | return rc; | ||
941 | |||
942 | qeth_l3_del_rxip(card, proto, addr); | ||
943 | |||
944 | return count; | ||
945 | } | ||
946 | |||
947 | static ssize_t qeth_l3_dev_rxip_del4_store(struct device *dev, | ||
948 | struct device_attribute *attr, const char *buf, size_t count) | ||
949 | { | ||
950 | struct qeth_card *card = dev_get_drvdata(dev); | ||
951 | |||
952 | if (!card) | ||
953 | return -EINVAL; | ||
954 | |||
955 | return qeth_l3_dev_rxip_del_store(buf, count, card, QETH_PROT_IPV4); | ||
956 | } | ||
957 | |||
958 | static QETH_DEVICE_ATTR(rxip_del4, del4, 0200, NULL, | ||
959 | qeth_l3_dev_rxip_del4_store); | ||
960 | |||
961 | static ssize_t qeth_l3_dev_rxip_add6_show(struct device *dev, | ||
962 | struct device_attribute *attr, char *buf) | ||
963 | { | ||
964 | struct qeth_card *card = dev_get_drvdata(dev); | ||
965 | |||
966 | if (!card) | ||
967 | return -EINVAL; | ||
968 | |||
969 | return qeth_l3_dev_rxip_add_show(buf, card, QETH_PROT_IPV6); | ||
970 | } | ||
971 | |||
972 | static ssize_t qeth_l3_dev_rxip_add6_store(struct device *dev, | ||
973 | struct device_attribute *attr, const char *buf, size_t count) | ||
974 | { | ||
975 | struct qeth_card *card = dev_get_drvdata(dev); | ||
976 | |||
977 | if (!card) | ||
978 | return -EINVAL; | ||
979 | |||
980 | return qeth_l3_dev_rxip_add_store(buf, count, card, QETH_PROT_IPV6); | ||
981 | } | ||
982 | |||
983 | static QETH_DEVICE_ATTR(rxip_add6, add6, 0644, | ||
984 | qeth_l3_dev_rxip_add6_show, | ||
985 | qeth_l3_dev_rxip_add6_store); | ||
986 | |||
987 | static ssize_t qeth_l3_dev_rxip_del6_store(struct device *dev, | ||
988 | struct device_attribute *attr, const char *buf, size_t count) | ||
989 | { | ||
990 | struct qeth_card *card = dev_get_drvdata(dev); | ||
991 | |||
992 | if (!card) | ||
993 | return -EINVAL; | ||
994 | |||
995 | return qeth_l3_dev_rxip_del_store(buf, count, card, QETH_PROT_IPV6); | ||
996 | } | ||
997 | |||
998 | static QETH_DEVICE_ATTR(rxip_del6, del6, 0200, NULL, | ||
999 | qeth_l3_dev_rxip_del6_store); | ||
1000 | |||
1001 | static struct attribute *qeth_rxip_device_attrs[] = { | ||
1002 | &dev_attr_rxip_add4.attr, | ||
1003 | &dev_attr_rxip_del4.attr, | ||
1004 | &dev_attr_rxip_add6.attr, | ||
1005 | &dev_attr_rxip_del6.attr, | ||
1006 | NULL, | ||
1007 | }; | ||
1008 | |||
1009 | static struct attribute_group qeth_device_rxip_group = { | ||
1010 | .name = "rxip", | ||
1011 | .attrs = qeth_rxip_device_attrs, | ||
1012 | }; | ||
1013 | |||
1014 | int qeth_l3_create_device_attributes(struct device *dev) | ||
1015 | { | ||
1016 | int ret; | ||
1017 | |||
1018 | ret = sysfs_create_group(&dev->kobj, &qeth_l3_device_attr_group); | ||
1019 | if (ret) | ||
1020 | return ret; | ||
1021 | |||
1022 | ret = sysfs_create_group(&dev->kobj, &qeth_device_ipato_group); | ||
1023 | if (ret) { | ||
1024 | sysfs_remove_group(&dev->kobj, &qeth_l3_device_attr_group); | ||
1025 | return ret; | ||
1026 | } | ||
1027 | |||
1028 | ret = sysfs_create_group(&dev->kobj, &qeth_device_vipa_group); | ||
1029 | if (ret) { | ||
1030 | sysfs_remove_group(&dev->kobj, &qeth_l3_device_attr_group); | ||
1031 | sysfs_remove_group(&dev->kobj, &qeth_device_ipato_group); | ||
1032 | return ret; | ||
1033 | } | ||
1034 | |||
1035 | ret = sysfs_create_group(&dev->kobj, &qeth_device_rxip_group); | ||
1036 | if (ret) { | ||
1037 | sysfs_remove_group(&dev->kobj, &qeth_l3_device_attr_group); | ||
1038 | sysfs_remove_group(&dev->kobj, &qeth_device_ipato_group); | ||
1039 | sysfs_remove_group(&dev->kobj, &qeth_device_vipa_group); | ||
1040 | return ret; | ||
1041 | } | ||
1042 | return 0; | ||
1043 | } | ||
1044 | |||
1045 | void qeth_l3_remove_device_attributes(struct device *dev) | ||
1046 | { | ||
1047 | sysfs_remove_group(&dev->kobj, &qeth_l3_device_attr_group); | ||
1048 | sysfs_remove_group(&dev->kobj, &qeth_device_ipato_group); | ||
1049 | sysfs_remove_group(&dev->kobj, &qeth_device_vipa_group); | ||
1050 | sysfs_remove_group(&dev->kobj, &qeth_device_rxip_group); | ||
1051 | } | ||
diff --git a/drivers/s390/net/qeth_main.c b/drivers/s390/net/qeth_main.c deleted file mode 100644 index 62606ce26e55..000000000000 --- a/drivers/s390/net/qeth_main.c +++ /dev/null | |||
@@ -1,8956 +0,0 @@ | |||
1 | /* | ||
2 | * linux/drivers/s390/net/qeth_main.c | ||
3 | * | ||
4 | * Linux on zSeries OSA Express and HiperSockets support | ||
5 | * | ||
6 | * Copyright 2000,2003 IBM Corporation | ||
7 | * | ||
8 | * Author(s): Original Code written by | ||
9 | * Utz Bacher (utz.bacher@de.ibm.com) | ||
10 | * Rewritten by | ||
11 | * Frank Pavlic (fpavlic@de.ibm.com) and | ||
12 | * Thomas Spatzier <tspat@de.ibm.com> | ||
13 | * | ||
14 | * This program is free software; you can redistribute it and/or modify | ||
15 | * it under the terms of the GNU General Public License as published by | ||
16 | * the Free Software Foundation; either version 2, or (at your option) | ||
17 | * any later version. | ||
18 | * | ||
19 | * This program is distributed in the hope that it will be useful, | ||
20 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | ||
21 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | ||
22 | * GNU General Public License for more details. | ||
23 | * | ||
24 | * You should have received a copy of the GNU General Public License | ||
25 | * along with this program; if not, write to the Free Software | ||
26 | * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. | ||
27 | */ | ||
28 | |||
29 | |||
30 | #include <linux/module.h> | ||
31 | #include <linux/moduleparam.h> | ||
32 | #include <linux/string.h> | ||
33 | #include <linux/errno.h> | ||
34 | #include <linux/mm.h> | ||
35 | #include <linux/ip.h> | ||
36 | #include <linux/inetdevice.h> | ||
37 | #include <linux/netdevice.h> | ||
38 | #include <linux/sched.h> | ||
39 | #include <linux/workqueue.h> | ||
40 | #include <linux/kernel.h> | ||
41 | #include <linux/slab.h> | ||
42 | #include <linux/interrupt.h> | ||
43 | #include <linux/tcp.h> | ||
44 | #include <linux/icmp.h> | ||
45 | #include <linux/skbuff.h> | ||
46 | #include <linux/in.h> | ||
47 | #include <linux/igmp.h> | ||
48 | #include <linux/init.h> | ||
49 | #include <linux/reboot.h> | ||
50 | #include <linux/mii.h> | ||
51 | #include <linux/rcupdate.h> | ||
52 | #include <linux/ethtool.h> | ||
53 | |||
54 | #include <net/arp.h> | ||
55 | #include <net/ip.h> | ||
56 | #include <net/route.h> | ||
57 | |||
58 | #include <asm/ebcdic.h> | ||
59 | #include <asm/io.h> | ||
60 | #include <asm/qeth.h> | ||
61 | #include <asm/timex.h> | ||
62 | #include <asm/semaphore.h> | ||
63 | #include <asm/uaccess.h> | ||
64 | #include <asm/s390_rdev.h> | ||
65 | |||
66 | #include "qeth.h" | ||
67 | #include "qeth_mpc.h" | ||
68 | #include "qeth_fs.h" | ||
69 | #include "qeth_eddp.h" | ||
70 | #include "qeth_tso.h" | ||
71 | |||
72 | static const char *version = "qeth S/390 OSA-Express driver"; | ||
73 | |||
74 | /** | ||
75 | * Debug Facility Stuff | ||
76 | */ | ||
77 | static debug_info_t *qeth_dbf_setup = NULL; | ||
78 | static debug_info_t *qeth_dbf_data = NULL; | ||
79 | static debug_info_t *qeth_dbf_misc = NULL; | ||
80 | static debug_info_t *qeth_dbf_control = NULL; | ||
81 | debug_info_t *qeth_dbf_trace = NULL; | ||
82 | static debug_info_t *qeth_dbf_sense = NULL; | ||
83 | static debug_info_t *qeth_dbf_qerr = NULL; | ||
84 | |||
85 | DEFINE_PER_CPU(char[256], qeth_dbf_txt_buf); | ||
86 | |||
87 | static struct lock_class_key qdio_out_skb_queue_key; | ||
88 | |||
89 | /** | ||
90 | * some more definitions and declarations | ||
91 | */ | ||
92 | static unsigned int known_devices[][10] = QETH_MODELLIST_ARRAY; | ||
93 | |||
94 | /* list of our cards */ | ||
95 | struct qeth_card_list_struct qeth_card_list; | ||
96 | /*process list want to be notified*/ | ||
97 | spinlock_t qeth_notify_lock; | ||
98 | struct list_head qeth_notify_list; | ||
99 | |||
100 | static void qeth_send_control_data_cb(struct qeth_channel *, | ||
101 | struct qeth_cmd_buffer *); | ||
102 | |||
103 | /** | ||
104 | * here we go with function implementation | ||
105 | */ | ||
106 | static void | ||
107 | qeth_init_qdio_info(struct qeth_card *card); | ||
108 | |||
109 | static int | ||
110 | qeth_init_qdio_queues(struct qeth_card *card); | ||
111 | |||
112 | static int | ||
113 | qeth_alloc_qdio_buffers(struct qeth_card *card); | ||
114 | |||
115 | static void | ||
116 | qeth_free_qdio_buffers(struct qeth_card *); | ||
117 | |||
118 | static void | ||
119 | qeth_clear_qdio_buffers(struct qeth_card *); | ||
120 | |||
121 | static void | ||
122 | qeth_clear_ip_list(struct qeth_card *, int, int); | ||
123 | |||
124 | static void | ||
125 | qeth_clear_ipacmd_list(struct qeth_card *); | ||
126 | |||
127 | static int | ||
128 | qeth_qdio_clear_card(struct qeth_card *, int); | ||
129 | |||
130 | static void | ||
131 | qeth_clear_working_pool_list(struct qeth_card *); | ||
132 | |||
133 | static void | ||
134 | qeth_clear_cmd_buffers(struct qeth_channel *); | ||
135 | |||
136 | static int | ||
137 | qeth_stop(struct net_device *); | ||
138 | |||
139 | static void | ||
140 | qeth_clear_ipato_list(struct qeth_card *); | ||
141 | |||
142 | static int | ||
143 | qeth_is_addr_covered_by_ipato(struct qeth_card *, struct qeth_ipaddr *); | ||
144 | |||
145 | static void | ||
146 | qeth_irq_tasklet(unsigned long); | ||
147 | |||
148 | static int | ||
149 | qeth_set_online(struct ccwgroup_device *); | ||
150 | |||
151 | static int | ||
152 | __qeth_set_online(struct ccwgroup_device *gdev, int recovery_mode); | ||
153 | |||
154 | static struct qeth_ipaddr * | ||
155 | qeth_get_addr_buffer(enum qeth_prot_versions); | ||
156 | |||
157 | static void | ||
158 | qeth_set_multicast_list(struct net_device *); | ||
159 | |||
160 | static void | ||
161 | qeth_setadp_promisc_mode(struct qeth_card *); | ||
162 | |||
163 | static int | ||
164 | qeth_hard_header_parse(const struct sk_buff *skb, unsigned char *haddr); | ||
165 | |||
166 | static void | ||
167 | qeth_notify_processes(void) | ||
168 | { | ||
169 | /*notify all registered processes */ | ||
170 | struct qeth_notify_list_struct *n_entry; | ||
171 | |||
172 | QETH_DBF_TEXT(trace,3,"procnoti"); | ||
173 | spin_lock(&qeth_notify_lock); | ||
174 | list_for_each_entry(n_entry, &qeth_notify_list, list) { | ||
175 | send_sig(n_entry->signum, n_entry->task, 1); | ||
176 | } | ||
177 | spin_unlock(&qeth_notify_lock); | ||
178 | |||
179 | } | ||
180 | int | ||
181 | qeth_notifier_unregister(struct task_struct *p) | ||
182 | { | ||
183 | struct qeth_notify_list_struct *n_entry, *tmp; | ||
184 | |||
185 | QETH_DBF_TEXT(trace, 2, "notunreg"); | ||
186 | spin_lock(&qeth_notify_lock); | ||
187 | list_for_each_entry_safe(n_entry, tmp, &qeth_notify_list, list) { | ||
188 | if (n_entry->task == p) { | ||
189 | list_del(&n_entry->list); | ||
190 | kfree(n_entry); | ||
191 | goto out; | ||
192 | } | ||
193 | } | ||
194 | out: | ||
195 | spin_unlock(&qeth_notify_lock); | ||
196 | return 0; | ||
197 | } | ||
198 | int | ||
199 | qeth_notifier_register(struct task_struct *p, int signum) | ||
200 | { | ||
201 | struct qeth_notify_list_struct *n_entry; | ||
202 | |||
203 | /*check first if entry already exists*/ | ||
204 | spin_lock(&qeth_notify_lock); | ||
205 | list_for_each_entry(n_entry, &qeth_notify_list, list) { | ||
206 | if (n_entry->task == p) { | ||
207 | n_entry->signum = signum; | ||
208 | spin_unlock(&qeth_notify_lock); | ||
209 | return 0; | ||
210 | } | ||
211 | } | ||
212 | spin_unlock(&qeth_notify_lock); | ||
213 | |||
214 | n_entry = (struct qeth_notify_list_struct *) | ||
215 | kmalloc(sizeof(struct qeth_notify_list_struct),GFP_KERNEL); | ||
216 | if (!n_entry) | ||
217 | return -ENOMEM; | ||
218 | n_entry->task = p; | ||
219 | n_entry->signum = signum; | ||
220 | spin_lock(&qeth_notify_lock); | ||
221 | list_add(&n_entry->list,&qeth_notify_list); | ||
222 | spin_unlock(&qeth_notify_lock); | ||
223 | return 0; | ||
224 | } | ||
225 | |||
226 | |||
227 | /** | ||
228 | * free channel command buffers | ||
229 | */ | ||
230 | static void | ||
231 | qeth_clean_channel(struct qeth_channel *channel) | ||
232 | { | ||
233 | int cnt; | ||
234 | |||
235 | QETH_DBF_TEXT(setup, 2, "freech"); | ||
236 | for (cnt = 0; cnt < QETH_CMD_BUFFER_NO; cnt++) | ||
237 | kfree(channel->iob[cnt].data); | ||
238 | } | ||
239 | |||
240 | /** | ||
241 | * free card | ||
242 | */ | ||
243 | static void | ||
244 | qeth_free_card(struct qeth_card *card) | ||
245 | { | ||
246 | |||
247 | QETH_DBF_TEXT(setup, 2, "freecrd"); | ||
248 | QETH_DBF_HEX(setup, 2, &card, sizeof(void *)); | ||
249 | qeth_clean_channel(&card->read); | ||
250 | qeth_clean_channel(&card->write); | ||
251 | if (card->dev) | ||
252 | free_netdev(card->dev); | ||
253 | qeth_clear_ip_list(card, 0, 0); | ||
254 | qeth_clear_ipato_list(card); | ||
255 | kfree(card->ip_tbd_list); | ||
256 | qeth_free_qdio_buffers(card); | ||
257 | kfree(card); | ||
258 | } | ||
259 | |||
260 | /** | ||
261 | * alloc memory for command buffer per channel | ||
262 | */ | ||
263 | static int | ||
264 | qeth_setup_channel(struct qeth_channel *channel) | ||
265 | { | ||
266 | int cnt; | ||
267 | |||
268 | QETH_DBF_TEXT(setup, 2, "setupch"); | ||
269 | for (cnt=0; cnt < QETH_CMD_BUFFER_NO; cnt++) { | ||
270 | channel->iob[cnt].data = (char *) | ||
271 | kmalloc(QETH_BUFSIZE, GFP_DMA|GFP_KERNEL); | ||
272 | if (channel->iob[cnt].data == NULL) | ||
273 | break; | ||
274 | channel->iob[cnt].state = BUF_STATE_FREE; | ||
275 | channel->iob[cnt].channel = channel; | ||
276 | channel->iob[cnt].callback = qeth_send_control_data_cb; | ||
277 | channel->iob[cnt].rc = 0; | ||
278 | } | ||
279 | if (cnt < QETH_CMD_BUFFER_NO) { | ||
280 | while (cnt-- > 0) | ||
281 | kfree(channel->iob[cnt].data); | ||
282 | return -ENOMEM; | ||
283 | } | ||
284 | channel->buf_no = 0; | ||
285 | channel->io_buf_no = 0; | ||
286 | atomic_set(&channel->irq_pending, 0); | ||
287 | spin_lock_init(&channel->iob_lock); | ||
288 | |||
289 | init_waitqueue_head(&channel->wait_q); | ||
290 | channel->irq_tasklet.data = (unsigned long) channel; | ||
291 | channel->irq_tasklet.func = qeth_irq_tasklet; | ||
292 | return 0; | ||
293 | } | ||
294 | |||
295 | /** | ||
296 | * alloc memory for card structure | ||
297 | */ | ||
298 | static struct qeth_card * | ||
299 | qeth_alloc_card(void) | ||
300 | { | ||
301 | struct qeth_card *card; | ||
302 | |||
303 | QETH_DBF_TEXT(setup, 2, "alloccrd"); | ||
304 | card = kzalloc(sizeof(struct qeth_card), GFP_DMA|GFP_KERNEL); | ||
305 | if (!card) | ||
306 | return NULL; | ||
307 | QETH_DBF_HEX(setup, 2, &card, sizeof(void *)); | ||
308 | if (qeth_setup_channel(&card->read)) { | ||
309 | kfree(card); | ||
310 | return NULL; | ||
311 | } | ||
312 | if (qeth_setup_channel(&card->write)) { | ||
313 | qeth_clean_channel(&card->read); | ||
314 | kfree(card); | ||
315 | return NULL; | ||
316 | } | ||
317 | return card; | ||
318 | } | ||
319 | |||
320 | static long | ||
321 | __qeth_check_irb_error(struct ccw_device *cdev, unsigned long intparm, | ||
322 | struct irb *irb) | ||
323 | { | ||
324 | if (!IS_ERR(irb)) | ||
325 | return 0; | ||
326 | |||
327 | switch (PTR_ERR(irb)) { | ||
328 | case -EIO: | ||
329 | PRINT_WARN("i/o-error on device %s\n", cdev->dev.bus_id); | ||
330 | QETH_DBF_TEXT(trace, 2, "ckirberr"); | ||
331 | QETH_DBF_TEXT_(trace, 2, " rc%d", -EIO); | ||
332 | break; | ||
333 | case -ETIMEDOUT: | ||
334 | PRINT_WARN("timeout on device %s\n", cdev->dev.bus_id); | ||
335 | QETH_DBF_TEXT(trace, 2, "ckirberr"); | ||
336 | QETH_DBF_TEXT_(trace, 2, " rc%d", -ETIMEDOUT); | ||
337 | if (intparm == QETH_RCD_PARM) { | ||
338 | struct qeth_card *card = CARD_FROM_CDEV(cdev); | ||
339 | |||
340 | if (card && (card->data.ccwdev == cdev)) { | ||
341 | card->data.state = CH_STATE_DOWN; | ||
342 | wake_up(&card->wait_q); | ||
343 | } | ||
344 | } | ||
345 | break; | ||
346 | default: | ||
347 | PRINT_WARN("unknown error %ld on device %s\n", PTR_ERR(irb), | ||
348 | cdev->dev.bus_id); | ||
349 | QETH_DBF_TEXT(trace, 2, "ckirberr"); | ||
350 | QETH_DBF_TEXT(trace, 2, " rc???"); | ||
351 | } | ||
352 | return PTR_ERR(irb); | ||
353 | } | ||
354 | |||
355 | static int | ||
356 | qeth_get_problem(struct ccw_device *cdev, struct irb *irb) | ||
357 | { | ||
358 | int dstat,cstat; | ||
359 | char *sense; | ||
360 | |||
361 | sense = (char *) irb->ecw; | ||
362 | cstat = irb->scsw.cstat; | ||
363 | dstat = irb->scsw.dstat; | ||
364 | |||
365 | if (cstat & (SCHN_STAT_CHN_CTRL_CHK | SCHN_STAT_INTF_CTRL_CHK | | ||
366 | SCHN_STAT_CHN_DATA_CHK | SCHN_STAT_CHAIN_CHECK | | ||
367 | SCHN_STAT_PROT_CHECK | SCHN_STAT_PROG_CHECK)) { | ||
368 | QETH_DBF_TEXT(trace,2, "CGENCHK"); | ||
369 | PRINT_WARN("check on device %s, dstat=x%x, cstat=x%x ", | ||
370 | cdev->dev.bus_id, dstat, cstat); | ||
371 | HEXDUMP16(WARN, "irb: ", irb); | ||
372 | HEXDUMP16(WARN, "irb: ", ((char *) irb) + 32); | ||
373 | return 1; | ||
374 | } | ||
375 | |||
376 | if (dstat & DEV_STAT_UNIT_CHECK) { | ||
377 | if (sense[SENSE_RESETTING_EVENT_BYTE] & | ||
378 | SENSE_RESETTING_EVENT_FLAG) { | ||
379 | QETH_DBF_TEXT(trace,2,"REVIND"); | ||
380 | return 1; | ||
381 | } | ||
382 | if (sense[SENSE_COMMAND_REJECT_BYTE] & | ||
383 | SENSE_COMMAND_REJECT_FLAG) { | ||
384 | QETH_DBF_TEXT(trace,2,"CMDREJi"); | ||
385 | return 0; | ||
386 | } | ||
387 | if ((sense[2] == 0xaf) && (sense[3] == 0xfe)) { | ||
388 | QETH_DBF_TEXT(trace,2,"AFFE"); | ||
389 | return 1; | ||
390 | } | ||
391 | if ((!sense[0]) && (!sense[1]) && (!sense[2]) && (!sense[3])) { | ||
392 | QETH_DBF_TEXT(trace,2,"ZEROSEN"); | ||
393 | return 0; | ||
394 | } | ||
395 | QETH_DBF_TEXT(trace,2,"DGENCHK"); | ||
396 | return 1; | ||
397 | } | ||
398 | return 0; | ||
399 | } | ||
400 | static int qeth_issue_next_read(struct qeth_card *); | ||
401 | |||
402 | /** | ||
403 | * interrupt handler | ||
404 | */ | ||
405 | static void | ||
406 | qeth_irq(struct ccw_device *cdev, unsigned long intparm, struct irb *irb) | ||
407 | { | ||
408 | int rc; | ||
409 | int cstat,dstat; | ||
410 | struct qeth_cmd_buffer *buffer; | ||
411 | struct qeth_channel *channel; | ||
412 | struct qeth_card *card; | ||
413 | |||
414 | QETH_DBF_TEXT(trace,5,"irq"); | ||
415 | |||
416 | if (__qeth_check_irb_error(cdev, intparm, irb)) | ||
417 | return; | ||
418 | cstat = irb->scsw.cstat; | ||
419 | dstat = irb->scsw.dstat; | ||
420 | |||
421 | card = CARD_FROM_CDEV(cdev); | ||
422 | if (!card) | ||
423 | return; | ||
424 | |||
425 | if (card->read.ccwdev == cdev){ | ||
426 | channel = &card->read; | ||
427 | QETH_DBF_TEXT(trace,5,"read"); | ||
428 | } else if (card->write.ccwdev == cdev) { | ||
429 | channel = &card->write; | ||
430 | QETH_DBF_TEXT(trace,5,"write"); | ||
431 | } else { | ||
432 | channel = &card->data; | ||
433 | QETH_DBF_TEXT(trace,5,"data"); | ||
434 | } | ||
435 | atomic_set(&channel->irq_pending, 0); | ||
436 | |||
437 | if (irb->scsw.fctl & (SCSW_FCTL_CLEAR_FUNC)) | ||
438 | channel->state = CH_STATE_STOPPED; | ||
439 | |||
440 | if (irb->scsw.fctl & (SCSW_FCTL_HALT_FUNC)) | ||
441 | channel->state = CH_STATE_HALTED; | ||
442 | |||
443 | /*let's wake up immediately on data channel*/ | ||
444 | if ((channel == &card->data) && (intparm != 0) && | ||
445 | (intparm != QETH_RCD_PARM)) | ||
446 | goto out; | ||
447 | |||
448 | if (intparm == QETH_CLEAR_CHANNEL_PARM) { | ||
449 | QETH_DBF_TEXT(trace, 6, "clrchpar"); | ||
450 | /* we don't have to handle this further */ | ||
451 | intparm = 0; | ||
452 | } | ||
453 | if (intparm == QETH_HALT_CHANNEL_PARM) { | ||
454 | QETH_DBF_TEXT(trace, 6, "hltchpar"); | ||
455 | /* we don't have to handle this further */ | ||
456 | intparm = 0; | ||
457 | } | ||
458 | if ((dstat & DEV_STAT_UNIT_EXCEP) || | ||
459 | (dstat & DEV_STAT_UNIT_CHECK) || | ||
460 | (cstat)) { | ||
461 | if (irb->esw.esw0.erw.cons) { | ||
462 | /* TODO: we should make this s390dbf */ | ||
463 | PRINT_WARN("sense data available on channel %s.\n", | ||
464 | CHANNEL_ID(channel)); | ||
465 | PRINT_WARN(" cstat 0x%X\n dstat 0x%X\n", cstat, dstat); | ||
466 | HEXDUMP16(WARN,"irb: ",irb); | ||
467 | HEXDUMP16(WARN,"sense data: ",irb->ecw); | ||
468 | } | ||
469 | if (intparm == QETH_RCD_PARM) { | ||
470 | channel->state = CH_STATE_DOWN; | ||
471 | goto out; | ||
472 | } | ||
473 | rc = qeth_get_problem(cdev,irb); | ||
474 | if (rc) { | ||
475 | qeth_schedule_recovery(card); | ||
476 | goto out; | ||
477 | } | ||
478 | } | ||
479 | |||
480 | if (intparm == QETH_RCD_PARM) { | ||
481 | channel->state = CH_STATE_RCD_DONE; | ||
482 | goto out; | ||
483 | } | ||
484 | if (intparm) { | ||
485 | buffer = (struct qeth_cmd_buffer *) __va((addr_t)intparm); | ||
486 | buffer->state = BUF_STATE_PROCESSED; | ||
487 | } | ||
488 | if (channel == &card->data) | ||
489 | return; | ||
490 | |||
491 | if (channel == &card->read && | ||
492 | channel->state == CH_STATE_UP) | ||
493 | qeth_issue_next_read(card); | ||
494 | |||
495 | qeth_irq_tasklet((unsigned long)channel); | ||
496 | return; | ||
497 | out: | ||
498 | wake_up(&card->wait_q); | ||
499 | } | ||
500 | |||
501 | /** | ||
502 | * tasklet function scheduled from irq handler | ||
503 | */ | ||
504 | static void | ||
505 | qeth_irq_tasklet(unsigned long data) | ||
506 | { | ||
507 | struct qeth_card *card; | ||
508 | struct qeth_channel *channel; | ||
509 | struct qeth_cmd_buffer *iob; | ||
510 | __u8 index; | ||
511 | |||
512 | QETH_DBF_TEXT(trace,5,"irqtlet"); | ||
513 | channel = (struct qeth_channel *) data; | ||
514 | iob = channel->iob; | ||
515 | index = channel->buf_no; | ||
516 | card = CARD_FROM_CDEV(channel->ccwdev); | ||
517 | while (iob[index].state == BUF_STATE_PROCESSED) { | ||
518 | if (iob[index].callback !=NULL) { | ||
519 | iob[index].callback(channel,iob + index); | ||
520 | } | ||
521 | index = (index + 1) % QETH_CMD_BUFFER_NO; | ||
522 | } | ||
523 | channel->buf_no = index; | ||
524 | wake_up(&card->wait_q); | ||
525 | } | ||
526 | |||
527 | static int qeth_stop_card(struct qeth_card *, int); | ||
528 | |||
529 | static int | ||
530 | __qeth_set_offline(struct ccwgroup_device *cgdev, int recovery_mode) | ||
531 | { | ||
532 | struct qeth_card *card = (struct qeth_card *) cgdev->dev.driver_data; | ||
533 | int rc = 0, rc2 = 0, rc3 = 0; | ||
534 | enum qeth_card_states recover_flag; | ||
535 | |||
536 | QETH_DBF_TEXT(setup, 3, "setoffl"); | ||
537 | QETH_DBF_HEX(setup, 3, &card, sizeof(void *)); | ||
538 | |||
539 | if (card->dev && netif_carrier_ok(card->dev)) | ||
540 | netif_carrier_off(card->dev); | ||
541 | recover_flag = card->state; | ||
542 | if (qeth_stop_card(card, recovery_mode) == -ERESTARTSYS){ | ||
543 | PRINT_WARN("Stopping card %s interrupted by user!\n", | ||
544 | CARD_BUS_ID(card)); | ||
545 | return -ERESTARTSYS; | ||
546 | } | ||
547 | rc = ccw_device_set_offline(CARD_DDEV(card)); | ||
548 | rc2 = ccw_device_set_offline(CARD_WDEV(card)); | ||
549 | rc3 = ccw_device_set_offline(CARD_RDEV(card)); | ||
550 | if (!rc) | ||
551 | rc = (rc2) ? rc2 : rc3; | ||
552 | if (rc) | ||
553 | QETH_DBF_TEXT_(setup, 2, "1err%d", rc); | ||
554 | if (recover_flag == CARD_STATE_UP) | ||
555 | card->state = CARD_STATE_RECOVER; | ||
556 | qeth_notify_processes(); | ||
557 | return 0; | ||
558 | } | ||
559 | |||
560 | static int | ||
561 | qeth_set_offline(struct ccwgroup_device *cgdev) | ||
562 | { | ||
563 | return __qeth_set_offline(cgdev, 0); | ||
564 | } | ||
565 | |||
566 | static int | ||
567 | qeth_threads_running(struct qeth_card *card, unsigned long threads); | ||
568 | |||
569 | |||
570 | static void | ||
571 | qeth_remove_device(struct ccwgroup_device *cgdev) | ||
572 | { | ||
573 | struct qeth_card *card = (struct qeth_card *) cgdev->dev.driver_data; | ||
574 | unsigned long flags; | ||
575 | |||
576 | QETH_DBF_TEXT(setup, 3, "rmdev"); | ||
577 | QETH_DBF_HEX(setup, 3, &card, sizeof(void *)); | ||
578 | |||
579 | if (!card) | ||
580 | return; | ||
581 | |||
582 | wait_event(card->wait_q, qeth_threads_running(card, 0xffffffff) == 0); | ||
583 | |||
584 | if (cgdev->state == CCWGROUP_ONLINE){ | ||
585 | card->use_hard_stop = 1; | ||
586 | qeth_set_offline(cgdev); | ||
587 | } | ||
588 | /* remove form our internal list */ | ||
589 | write_lock_irqsave(&qeth_card_list.rwlock, flags); | ||
590 | list_del(&card->list); | ||
591 | write_unlock_irqrestore(&qeth_card_list.rwlock, flags); | ||
592 | if (card->dev) | ||
593 | unregister_netdev(card->dev); | ||
594 | qeth_remove_device_attributes(&cgdev->dev); | ||
595 | qeth_free_card(card); | ||
596 | cgdev->dev.driver_data = NULL; | ||
597 | put_device(&cgdev->dev); | ||
598 | } | ||
599 | |||
600 | static int | ||
601 | qeth_register_addr_entry(struct qeth_card *, struct qeth_ipaddr *); | ||
602 | static int | ||
603 | qeth_deregister_addr_entry(struct qeth_card *, struct qeth_ipaddr *); | ||
604 | |||
605 | /** | ||
606 | * Add/remove address to/from card's ip list, i.e. try to add or remove | ||
607 | * reference to/from an IP address that is already registered on the card. | ||
608 | * Returns: | ||
609 | * 0 address was on card and its reference count has been adjusted, | ||
610 | * but is still > 0, so nothing has to be done | ||
611 | * also returns 0 if card was not on card and the todo was to delete | ||
612 | * the address -> there is also nothing to be done | ||
613 | * 1 address was not on card and the todo is to add it to the card's ip | ||
614 | * list | ||
615 | * -1 address was on card and its reference count has been decremented | ||
616 | * to <= 0 by the todo -> address must be removed from card | ||
617 | */ | ||
618 | static int | ||
619 | __qeth_ref_ip_on_card(struct qeth_card *card, struct qeth_ipaddr *todo, | ||
620 | struct qeth_ipaddr **__addr) | ||
621 | { | ||
622 | struct qeth_ipaddr *addr; | ||
623 | int found = 0; | ||
624 | |||
625 | list_for_each_entry(addr, &card->ip_list, entry) { | ||
626 | if (card->options.layer2) { | ||
627 | if ((addr->type == todo->type) && | ||
628 | (memcmp(&addr->mac, &todo->mac, | ||
629 | OSA_ADDR_LEN) == 0)) { | ||
630 | found = 1; | ||
631 | break; | ||
632 | } | ||
633 | continue; | ||
634 | } | ||
635 | if ((addr->proto == QETH_PROT_IPV4) && | ||
636 | (todo->proto == QETH_PROT_IPV4) && | ||
637 | (addr->type == todo->type) && | ||
638 | (addr->u.a4.addr == todo->u.a4.addr) && | ||
639 | (addr->u.a4.mask == todo->u.a4.mask)) { | ||
640 | found = 1; | ||
641 | break; | ||
642 | } | ||
643 | if ((addr->proto == QETH_PROT_IPV6) && | ||
644 | (todo->proto == QETH_PROT_IPV6) && | ||
645 | (addr->type == todo->type) && | ||
646 | (addr->u.a6.pfxlen == todo->u.a6.pfxlen) && | ||
647 | (memcmp(&addr->u.a6.addr, &todo->u.a6.addr, | ||
648 | sizeof(struct in6_addr)) == 0)) { | ||
649 | found = 1; | ||
650 | break; | ||
651 | } | ||
652 | } | ||
653 | if (found) { | ||
654 | addr->users += todo->users; | ||
655 | if (addr->users <= 0){ | ||
656 | *__addr = addr; | ||
657 | return -1; | ||
658 | } else { | ||
659 | /* for VIPA and RXIP limit refcount to 1 */ | ||
660 | if (addr->type != QETH_IP_TYPE_NORMAL) | ||
661 | addr->users = 1; | ||
662 | return 0; | ||
663 | } | ||
664 | } | ||
665 | if (todo->users > 0) { | ||
666 | /* for VIPA and RXIP limit refcount to 1 */ | ||
667 | if (todo->type != QETH_IP_TYPE_NORMAL) | ||
668 | todo->users = 1; | ||
669 | return 1; | ||
670 | } else | ||
671 | return 0; | ||
672 | } | ||
673 | |||
674 | static int | ||
675 | __qeth_address_exists_in_list(struct list_head *list, struct qeth_ipaddr *addr, | ||
676 | int same_type) | ||
677 | { | ||
678 | struct qeth_ipaddr *tmp; | ||
679 | |||
680 | list_for_each_entry(tmp, list, entry) { | ||
681 | if ((tmp->proto == QETH_PROT_IPV4) && | ||
682 | (addr->proto == QETH_PROT_IPV4) && | ||
683 | ((same_type && (tmp->type == addr->type)) || | ||
684 | (!same_type && (tmp->type != addr->type)) ) && | ||
685 | (tmp->u.a4.addr == addr->u.a4.addr) ){ | ||
686 | return 1; | ||
687 | } | ||
688 | if ((tmp->proto == QETH_PROT_IPV6) && | ||
689 | (addr->proto == QETH_PROT_IPV6) && | ||
690 | ((same_type && (tmp->type == addr->type)) || | ||
691 | (!same_type && (tmp->type != addr->type)) ) && | ||
692 | (memcmp(&tmp->u.a6.addr, &addr->u.a6.addr, | ||
693 | sizeof(struct in6_addr)) == 0) ) { | ||
694 | return 1; | ||
695 | } | ||
696 | } | ||
697 | return 0; | ||
698 | } | ||
699 | |||
700 | /* | ||
701 | * Add IP to be added to todo list. If there is already an "add todo" | ||
702 | * in this list we just incremenent the reference count. | ||
703 | * Returns 0 if we just incremented reference count. | ||
704 | */ | ||
705 | static int | ||
706 | __qeth_insert_ip_todo(struct qeth_card *card, struct qeth_ipaddr *addr, int add) | ||
707 | { | ||
708 | struct qeth_ipaddr *tmp, *t; | ||
709 | int found = 0; | ||
710 | |||
711 | list_for_each_entry_safe(tmp, t, card->ip_tbd_list, entry) { | ||
712 | if ((addr->type == QETH_IP_TYPE_DEL_ALL_MC) && | ||
713 | (tmp->type == QETH_IP_TYPE_DEL_ALL_MC)) | ||
714 | return 0; | ||
715 | if (card->options.layer2) { | ||
716 | if ((tmp->type == addr->type) && | ||
717 | (tmp->is_multicast == addr->is_multicast) && | ||
718 | (memcmp(&tmp->mac, &addr->mac, | ||
719 | OSA_ADDR_LEN) == 0)) { | ||
720 | found = 1; | ||
721 | break; | ||
722 | } | ||
723 | continue; | ||
724 | } | ||
725 | if ((tmp->proto == QETH_PROT_IPV4) && | ||
726 | (addr->proto == QETH_PROT_IPV4) && | ||
727 | (tmp->type == addr->type) && | ||
728 | (tmp->is_multicast == addr->is_multicast) && | ||
729 | (tmp->u.a4.addr == addr->u.a4.addr) && | ||
730 | (tmp->u.a4.mask == addr->u.a4.mask)) { | ||
731 | found = 1; | ||
732 | break; | ||
733 | } | ||
734 | if ((tmp->proto == QETH_PROT_IPV6) && | ||
735 | (addr->proto == QETH_PROT_IPV6) && | ||
736 | (tmp->type == addr->type) && | ||
737 | (tmp->is_multicast == addr->is_multicast) && | ||
738 | (tmp->u.a6.pfxlen == addr->u.a6.pfxlen) && | ||
739 | (memcmp(&tmp->u.a6.addr, &addr->u.a6.addr, | ||
740 | sizeof(struct in6_addr)) == 0)) { | ||
741 | found = 1; | ||
742 | break; | ||
743 | } | ||
744 | } | ||
745 | if (found){ | ||
746 | if (addr->users != 0) | ||
747 | tmp->users += addr->users; | ||
748 | else | ||
749 | tmp->users += add? 1:-1; | ||
750 | if (tmp->users == 0) { | ||
751 | list_del(&tmp->entry); | ||
752 | kfree(tmp); | ||
753 | } | ||
754 | return 0; | ||
755 | } else { | ||
756 | if (addr->type == QETH_IP_TYPE_DEL_ALL_MC) | ||
757 | list_add(&addr->entry, card->ip_tbd_list); | ||
758 | else { | ||
759 | if (addr->users == 0) | ||
760 | addr->users += add? 1:-1; | ||
761 | if (add && (addr->type == QETH_IP_TYPE_NORMAL) && | ||
762 | qeth_is_addr_covered_by_ipato(card, addr)){ | ||
763 | QETH_DBF_TEXT(trace, 2, "tkovaddr"); | ||
764 | addr->set_flags |= QETH_IPA_SETIP_TAKEOVER_FLAG; | ||
765 | } | ||
766 | list_add_tail(&addr->entry, card->ip_tbd_list); | ||
767 | } | ||
768 | return 1; | ||
769 | } | ||
770 | } | ||
771 | |||
772 | /** | ||
773 | * Remove IP address from list | ||
774 | */ | ||
775 | static int | ||
776 | qeth_delete_ip(struct qeth_card *card, struct qeth_ipaddr *addr) | ||
777 | { | ||
778 | unsigned long flags; | ||
779 | int rc = 0; | ||
780 | |||
781 | QETH_DBF_TEXT(trace, 4, "delip"); | ||
782 | |||
783 | if (card->options.layer2) | ||
784 | QETH_DBF_HEX(trace, 4, &addr->mac, 6); | ||
785 | else if (addr->proto == QETH_PROT_IPV4) | ||
786 | QETH_DBF_HEX(trace, 4, &addr->u.a4.addr, 4); | ||
787 | else { | ||
788 | QETH_DBF_HEX(trace, 4, &addr->u.a6.addr, 8); | ||
789 | QETH_DBF_HEX(trace, 4, ((char *)&addr->u.a6.addr) + 8, 8); | ||
790 | } | ||
791 | spin_lock_irqsave(&card->ip_lock, flags); | ||
792 | rc = __qeth_insert_ip_todo(card, addr, 0); | ||
793 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
794 | return rc; | ||
795 | } | ||
796 | |||
797 | static int | ||
798 | qeth_add_ip(struct qeth_card *card, struct qeth_ipaddr *addr) | ||
799 | { | ||
800 | unsigned long flags; | ||
801 | int rc = 0; | ||
802 | |||
803 | QETH_DBF_TEXT(trace, 4, "addip"); | ||
804 | if (card->options.layer2) | ||
805 | QETH_DBF_HEX(trace, 4, &addr->mac, 6); | ||
806 | else if (addr->proto == QETH_PROT_IPV4) | ||
807 | QETH_DBF_HEX(trace, 4, &addr->u.a4.addr, 4); | ||
808 | else { | ||
809 | QETH_DBF_HEX(trace, 4, &addr->u.a6.addr, 8); | ||
810 | QETH_DBF_HEX(trace, 4, ((char *)&addr->u.a6.addr) + 8, 8); | ||
811 | } | ||
812 | spin_lock_irqsave(&card->ip_lock, flags); | ||
813 | rc = __qeth_insert_ip_todo(card, addr, 1); | ||
814 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
815 | return rc; | ||
816 | } | ||
817 | |||
818 | static void | ||
819 | __qeth_delete_all_mc(struct qeth_card *card, unsigned long *flags) | ||
820 | { | ||
821 | struct qeth_ipaddr *addr, *tmp; | ||
822 | int rc; | ||
823 | again: | ||
824 | list_for_each_entry_safe(addr, tmp, &card->ip_list, entry) { | ||
825 | if (addr->is_multicast) { | ||
826 | list_del(&addr->entry); | ||
827 | spin_unlock_irqrestore(&card->ip_lock, *flags); | ||
828 | rc = qeth_deregister_addr_entry(card, addr); | ||
829 | spin_lock_irqsave(&card->ip_lock, *flags); | ||
830 | if (!rc) { | ||
831 | kfree(addr); | ||
832 | goto again; | ||
833 | } else | ||
834 | list_add(&addr->entry, &card->ip_list); | ||
835 | } | ||
836 | } | ||
837 | } | ||
838 | |||
839 | static void | ||
840 | qeth_set_ip_addr_list(struct qeth_card *card) | ||
841 | { | ||
842 | struct list_head *tbd_list; | ||
843 | struct qeth_ipaddr *todo, *addr; | ||
844 | unsigned long flags; | ||
845 | int rc; | ||
846 | |||
847 | QETH_DBF_TEXT(trace, 2, "sdiplist"); | ||
848 | QETH_DBF_HEX(trace, 2, &card, sizeof(void *)); | ||
849 | |||
850 | spin_lock_irqsave(&card->ip_lock, flags); | ||
851 | tbd_list = card->ip_tbd_list; | ||
852 | card->ip_tbd_list = kmalloc(sizeof(struct list_head), GFP_ATOMIC); | ||
853 | if (!card->ip_tbd_list) { | ||
854 | QETH_DBF_TEXT(trace, 0, "silnomem"); | ||
855 | card->ip_tbd_list = tbd_list; | ||
856 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
857 | return; | ||
858 | } else | ||
859 | INIT_LIST_HEAD(card->ip_tbd_list); | ||
860 | |||
861 | while (!list_empty(tbd_list)){ | ||
862 | todo = list_entry(tbd_list->next, struct qeth_ipaddr, entry); | ||
863 | list_del(&todo->entry); | ||
864 | if (todo->type == QETH_IP_TYPE_DEL_ALL_MC){ | ||
865 | __qeth_delete_all_mc(card, &flags); | ||
866 | kfree(todo); | ||
867 | continue; | ||
868 | } | ||
869 | rc = __qeth_ref_ip_on_card(card, todo, &addr); | ||
870 | if (rc == 0) { | ||
871 | /* nothing to be done; only adjusted refcount */ | ||
872 | kfree(todo); | ||
873 | } else if (rc == 1) { | ||
874 | /* new entry to be added to on-card list */ | ||
875 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
876 | rc = qeth_register_addr_entry(card, todo); | ||
877 | spin_lock_irqsave(&card->ip_lock, flags); | ||
878 | if (!rc) | ||
879 | list_add_tail(&todo->entry, &card->ip_list); | ||
880 | else | ||
881 | kfree(todo); | ||
882 | } else if (rc == -1) { | ||
883 | /* on-card entry to be removed */ | ||
884 | list_del_init(&addr->entry); | ||
885 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
886 | rc = qeth_deregister_addr_entry(card, addr); | ||
887 | spin_lock_irqsave(&card->ip_lock, flags); | ||
888 | if (!rc) | ||
889 | kfree(addr); | ||
890 | else | ||
891 | list_add_tail(&addr->entry, &card->ip_list); | ||
892 | kfree(todo); | ||
893 | } | ||
894 | } | ||
895 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
896 | kfree(tbd_list); | ||
897 | } | ||
898 | |||
899 | static void qeth_delete_mc_addresses(struct qeth_card *); | ||
900 | static void qeth_add_multicast_ipv4(struct qeth_card *); | ||
901 | static void qeth_layer2_add_multicast(struct qeth_card *); | ||
902 | #ifdef CONFIG_QETH_IPV6 | ||
903 | static void qeth_add_multicast_ipv6(struct qeth_card *); | ||
904 | #endif | ||
905 | |||
906 | static int | ||
907 | qeth_set_thread_start_bit(struct qeth_card *card, unsigned long thread) | ||
908 | { | ||
909 | unsigned long flags; | ||
910 | |||
911 | spin_lock_irqsave(&card->thread_mask_lock, flags); | ||
912 | if ( !(card->thread_allowed_mask & thread) || | ||
913 | (card->thread_start_mask & thread) ) { | ||
914 | spin_unlock_irqrestore(&card->thread_mask_lock, flags); | ||
915 | return -EPERM; | ||
916 | } | ||
917 | card->thread_start_mask |= thread; | ||
918 | spin_unlock_irqrestore(&card->thread_mask_lock, flags); | ||
919 | return 0; | ||
920 | } | ||
921 | |||
922 | static void | ||
923 | qeth_clear_thread_start_bit(struct qeth_card *card, unsigned long thread) | ||
924 | { | ||
925 | unsigned long flags; | ||
926 | |||
927 | spin_lock_irqsave(&card->thread_mask_lock, flags); | ||
928 | card->thread_start_mask &= ~thread; | ||
929 | spin_unlock_irqrestore(&card->thread_mask_lock, flags); | ||
930 | wake_up(&card->wait_q); | ||
931 | } | ||
932 | |||
933 | static void | ||
934 | qeth_clear_thread_running_bit(struct qeth_card *card, unsigned long thread) | ||
935 | { | ||
936 | unsigned long flags; | ||
937 | |||
938 | spin_lock_irqsave(&card->thread_mask_lock, flags); | ||
939 | card->thread_running_mask &= ~thread; | ||
940 | spin_unlock_irqrestore(&card->thread_mask_lock, flags); | ||
941 | wake_up(&card->wait_q); | ||
942 | } | ||
943 | |||
944 | static int | ||
945 | __qeth_do_run_thread(struct qeth_card *card, unsigned long thread) | ||
946 | { | ||
947 | unsigned long flags; | ||
948 | int rc = 0; | ||
949 | |||
950 | spin_lock_irqsave(&card->thread_mask_lock, flags); | ||
951 | if (card->thread_start_mask & thread){ | ||
952 | if ((card->thread_allowed_mask & thread) && | ||
953 | !(card->thread_running_mask & thread)){ | ||
954 | rc = 1; | ||
955 | card->thread_start_mask &= ~thread; | ||
956 | card->thread_running_mask |= thread; | ||
957 | } else | ||
958 | rc = -EPERM; | ||
959 | } | ||
960 | spin_unlock_irqrestore(&card->thread_mask_lock, flags); | ||
961 | return rc; | ||
962 | } | ||
963 | |||
964 | static int | ||
965 | qeth_do_run_thread(struct qeth_card *card, unsigned long thread) | ||
966 | { | ||
967 | int rc = 0; | ||
968 | |||
969 | wait_event(card->wait_q, | ||
970 | (rc = __qeth_do_run_thread(card, thread)) >= 0); | ||
971 | return rc; | ||
972 | } | ||
973 | |||
974 | static int | ||
975 | qeth_recover(void *ptr) | ||
976 | { | ||
977 | struct qeth_card *card; | ||
978 | int rc = 0; | ||
979 | |||
980 | card = (struct qeth_card *) ptr; | ||
981 | daemonize("qeth_recover"); | ||
982 | QETH_DBF_TEXT(trace,2,"recover1"); | ||
983 | QETH_DBF_HEX(trace, 2, &card, sizeof(void *)); | ||
984 | if (!qeth_do_run_thread(card, QETH_RECOVER_THREAD)) | ||
985 | return 0; | ||
986 | QETH_DBF_TEXT(trace,2,"recover2"); | ||
987 | PRINT_WARN("Recovery of device %s started ...\n", | ||
988 | CARD_BUS_ID(card)); | ||
989 | card->use_hard_stop = 1; | ||
990 | __qeth_set_offline(card->gdev,1); | ||
991 | rc = __qeth_set_online(card->gdev,1); | ||
992 | /* don't run another scheduled recovery */ | ||
993 | qeth_clear_thread_start_bit(card, QETH_RECOVER_THREAD); | ||
994 | qeth_clear_thread_running_bit(card, QETH_RECOVER_THREAD); | ||
995 | if (!rc) | ||
996 | PRINT_INFO("Device %s successfully recovered!\n", | ||
997 | CARD_BUS_ID(card)); | ||
998 | else | ||
999 | PRINT_INFO("Device %s could not be recovered!\n", | ||
1000 | CARD_BUS_ID(card)); | ||
1001 | return 0; | ||
1002 | } | ||
1003 | |||
1004 | void | ||
1005 | qeth_schedule_recovery(struct qeth_card *card) | ||
1006 | { | ||
1007 | QETH_DBF_TEXT(trace,2,"startrec"); | ||
1008 | if (qeth_set_thread_start_bit(card, QETH_RECOVER_THREAD) == 0) | ||
1009 | schedule_work(&card->kernel_thread_starter); | ||
1010 | } | ||
1011 | |||
1012 | static int | ||
1013 | qeth_do_start_thread(struct qeth_card *card, unsigned long thread) | ||
1014 | { | ||
1015 | unsigned long flags; | ||
1016 | int rc = 0; | ||
1017 | |||
1018 | spin_lock_irqsave(&card->thread_mask_lock, flags); | ||
1019 | QETH_DBF_TEXT_(trace, 4, " %02x%02x%02x", | ||
1020 | (u8) card->thread_start_mask, | ||
1021 | (u8) card->thread_allowed_mask, | ||
1022 | (u8) card->thread_running_mask); | ||
1023 | rc = (card->thread_start_mask & thread); | ||
1024 | spin_unlock_irqrestore(&card->thread_mask_lock, flags); | ||
1025 | return rc; | ||
1026 | } | ||
1027 | |||
1028 | static void | ||
1029 | qeth_start_kernel_thread(struct work_struct *work) | ||
1030 | { | ||
1031 | struct qeth_card *card = container_of(work, struct qeth_card, kernel_thread_starter); | ||
1032 | QETH_DBF_TEXT(trace , 2, "strthrd"); | ||
1033 | |||
1034 | if (card->read.state != CH_STATE_UP && | ||
1035 | card->write.state != CH_STATE_UP) | ||
1036 | return; | ||
1037 | if (qeth_do_start_thread(card, QETH_RECOVER_THREAD)) | ||
1038 | kernel_thread(qeth_recover, (void *) card, SIGCHLD); | ||
1039 | } | ||
1040 | |||
1041 | |||
1042 | static void | ||
1043 | qeth_set_intial_options(struct qeth_card *card) | ||
1044 | { | ||
1045 | card->options.route4.type = NO_ROUTER; | ||
1046 | #ifdef CONFIG_QETH_IPV6 | ||
1047 | card->options.route6.type = NO_ROUTER; | ||
1048 | #endif /* QETH_IPV6 */ | ||
1049 | card->options.checksum_type = QETH_CHECKSUM_DEFAULT; | ||
1050 | card->options.broadcast_mode = QETH_TR_BROADCAST_ALLRINGS; | ||
1051 | card->options.macaddr_mode = QETH_TR_MACADDR_NONCANONICAL; | ||
1052 | card->options.fake_broadcast = 0; | ||
1053 | card->options.add_hhlen = DEFAULT_ADD_HHLEN; | ||
1054 | card->options.fake_ll = 0; | ||
1055 | if (card->info.type == QETH_CARD_TYPE_OSN) | ||
1056 | card->options.layer2 = 1; | ||
1057 | else | ||
1058 | card->options.layer2 = 0; | ||
1059 | card->options.performance_stats = 0; | ||
1060 | card->options.rx_sg_cb = QETH_RX_SG_CB; | ||
1061 | } | ||
1062 | |||
1063 | /** | ||
1064 | * initialize channels ,card and all state machines | ||
1065 | */ | ||
1066 | static int | ||
1067 | qeth_setup_card(struct qeth_card *card) | ||
1068 | { | ||
1069 | |||
1070 | QETH_DBF_TEXT(setup, 2, "setupcrd"); | ||
1071 | QETH_DBF_HEX(setup, 2, &card, sizeof(void *)); | ||
1072 | |||
1073 | card->read.state = CH_STATE_DOWN; | ||
1074 | card->write.state = CH_STATE_DOWN; | ||
1075 | card->data.state = CH_STATE_DOWN; | ||
1076 | card->state = CARD_STATE_DOWN; | ||
1077 | card->lan_online = 0; | ||
1078 | card->use_hard_stop = 0; | ||
1079 | card->dev = NULL; | ||
1080 | #ifdef CONFIG_QETH_VLAN | ||
1081 | spin_lock_init(&card->vlanlock); | ||
1082 | card->vlangrp = NULL; | ||
1083 | #endif | ||
1084 | spin_lock_init(&card->lock); | ||
1085 | spin_lock_init(&card->ip_lock); | ||
1086 | spin_lock_init(&card->thread_mask_lock); | ||
1087 | card->thread_start_mask = 0; | ||
1088 | card->thread_allowed_mask = 0; | ||
1089 | card->thread_running_mask = 0; | ||
1090 | INIT_WORK(&card->kernel_thread_starter, qeth_start_kernel_thread); | ||
1091 | INIT_LIST_HEAD(&card->ip_list); | ||
1092 | card->ip_tbd_list = kmalloc(sizeof(struct list_head), GFP_KERNEL); | ||
1093 | if (!card->ip_tbd_list) { | ||
1094 | QETH_DBF_TEXT(setup, 0, "iptbdnom"); | ||
1095 | return -ENOMEM; | ||
1096 | } | ||
1097 | INIT_LIST_HEAD(card->ip_tbd_list); | ||
1098 | INIT_LIST_HEAD(&card->cmd_waiter_list); | ||
1099 | init_waitqueue_head(&card->wait_q); | ||
1100 | /* intial options */ | ||
1101 | qeth_set_intial_options(card); | ||
1102 | /* IP address takeover */ | ||
1103 | INIT_LIST_HEAD(&card->ipato.entries); | ||
1104 | card->ipato.enabled = 0; | ||
1105 | card->ipato.invert4 = 0; | ||
1106 | card->ipato.invert6 = 0; | ||
1107 | /* init QDIO stuff */ | ||
1108 | qeth_init_qdio_info(card); | ||
1109 | return 0; | ||
1110 | } | ||
1111 | |||
1112 | static int | ||
1113 | is_1920_device (struct qeth_card *card) | ||
1114 | { | ||
1115 | int single_queue = 0; | ||
1116 | struct ccw_device *ccwdev; | ||
1117 | struct channelPath_dsc { | ||
1118 | u8 flags; | ||
1119 | u8 lsn; | ||
1120 | u8 desc; | ||
1121 | u8 chpid; | ||
1122 | u8 swla; | ||
1123 | u8 zeroes; | ||
1124 | u8 chla; | ||
1125 | u8 chpp; | ||
1126 | } *chp_dsc; | ||
1127 | |||
1128 | QETH_DBF_TEXT(setup, 2, "chk_1920"); | ||
1129 | |||
1130 | ccwdev = card->data.ccwdev; | ||
1131 | chp_dsc = (struct channelPath_dsc *)ccw_device_get_chp_desc(ccwdev, 0); | ||
1132 | if (chp_dsc != NULL) { | ||
1133 | /* CHPP field bit 6 == 1 -> single queue */ | ||
1134 | single_queue = ((chp_dsc->chpp & 0x02) == 0x02); | ||
1135 | kfree(chp_dsc); | ||
1136 | } | ||
1137 | QETH_DBF_TEXT_(setup, 2, "rc:%x", single_queue); | ||
1138 | return single_queue; | ||
1139 | } | ||
1140 | |||
1141 | static int | ||
1142 | qeth_determine_card_type(struct qeth_card *card) | ||
1143 | { | ||
1144 | int i = 0; | ||
1145 | |||
1146 | QETH_DBF_TEXT(setup, 2, "detcdtyp"); | ||
1147 | |||
1148 | card->qdio.do_prio_queueing = QETH_PRIOQ_DEFAULT; | ||
1149 | card->qdio.default_out_queue = QETH_DEFAULT_QUEUE; | ||
1150 | while (known_devices[i][4]) { | ||
1151 | if ((CARD_RDEV(card)->id.dev_type == known_devices[i][2]) && | ||
1152 | (CARD_RDEV(card)->id.dev_model == known_devices[i][3])) { | ||
1153 | card->info.type = known_devices[i][4]; | ||
1154 | card->qdio.no_out_queues = known_devices[i][8]; | ||
1155 | card->info.is_multicast_different = known_devices[i][9]; | ||
1156 | if (is_1920_device(card)) { | ||
1157 | PRINT_INFO("Priority Queueing not able " | ||
1158 | "due to hardware limitations!\n"); | ||
1159 | card->qdio.no_out_queues = 1; | ||
1160 | card->qdio.default_out_queue = 0; | ||
1161 | } | ||
1162 | return 0; | ||
1163 | } | ||
1164 | i++; | ||
1165 | } | ||
1166 | card->info.type = QETH_CARD_TYPE_UNKNOWN; | ||
1167 | PRINT_ERR("unknown card type on device %s\n", CARD_BUS_ID(card)); | ||
1168 | return -ENOENT; | ||
1169 | } | ||
1170 | |||
1171 | static int | ||
1172 | qeth_probe_device(struct ccwgroup_device *gdev) | ||
1173 | { | ||
1174 | struct qeth_card *card; | ||
1175 | struct device *dev; | ||
1176 | unsigned long flags; | ||
1177 | int rc; | ||
1178 | |||
1179 | QETH_DBF_TEXT(setup, 2, "probedev"); | ||
1180 | |||
1181 | dev = &gdev->dev; | ||
1182 | if (!get_device(dev)) | ||
1183 | return -ENODEV; | ||
1184 | |||
1185 | QETH_DBF_TEXT_(setup, 2, "%s", gdev->dev.bus_id); | ||
1186 | |||
1187 | card = qeth_alloc_card(); | ||
1188 | if (!card) { | ||
1189 | put_device(dev); | ||
1190 | QETH_DBF_TEXT_(setup, 2, "1err%d", -ENOMEM); | ||
1191 | return -ENOMEM; | ||
1192 | } | ||
1193 | card->read.ccwdev = gdev->cdev[0]; | ||
1194 | card->write.ccwdev = gdev->cdev[1]; | ||
1195 | card->data.ccwdev = gdev->cdev[2]; | ||
1196 | gdev->dev.driver_data = card; | ||
1197 | card->gdev = gdev; | ||
1198 | gdev->cdev[0]->handler = qeth_irq; | ||
1199 | gdev->cdev[1]->handler = qeth_irq; | ||
1200 | gdev->cdev[2]->handler = qeth_irq; | ||
1201 | |||
1202 | if ((rc = qeth_determine_card_type(card))){ | ||
1203 | PRINT_WARN("%s: not a valid card type\n", __func__); | ||
1204 | QETH_DBF_TEXT_(setup, 2, "3err%d", rc); | ||
1205 | put_device(dev); | ||
1206 | qeth_free_card(card); | ||
1207 | return rc; | ||
1208 | } | ||
1209 | if ((rc = qeth_setup_card(card))){ | ||
1210 | QETH_DBF_TEXT_(setup, 2, "2err%d", rc); | ||
1211 | put_device(dev); | ||
1212 | qeth_free_card(card); | ||
1213 | return rc; | ||
1214 | } | ||
1215 | rc = qeth_create_device_attributes(dev); | ||
1216 | if (rc) { | ||
1217 | put_device(dev); | ||
1218 | qeth_free_card(card); | ||
1219 | return rc; | ||
1220 | } | ||
1221 | /* insert into our internal list */ | ||
1222 | write_lock_irqsave(&qeth_card_list.rwlock, flags); | ||
1223 | list_add_tail(&card->list, &qeth_card_list.list); | ||
1224 | write_unlock_irqrestore(&qeth_card_list.rwlock, flags); | ||
1225 | return rc; | ||
1226 | } | ||
1227 | |||
1228 | |||
1229 | static int qeth_read_conf_data(struct qeth_card *card, void **buffer, | ||
1230 | int *length) | ||
1231 | { | ||
1232 | struct ciw *ciw; | ||
1233 | char *rcd_buf; | ||
1234 | int ret; | ||
1235 | struct qeth_channel *channel = &card->data; | ||
1236 | unsigned long flags; | ||
1237 | |||
1238 | /* | ||
1239 | * scan for RCD command in extended SenseID data | ||
1240 | */ | ||
1241 | ciw = ccw_device_get_ciw(channel->ccwdev, CIW_TYPE_RCD); | ||
1242 | if (!ciw || ciw->cmd == 0) | ||
1243 | return -EOPNOTSUPP; | ||
1244 | rcd_buf = kzalloc(ciw->count, GFP_KERNEL | GFP_DMA); | ||
1245 | if (!rcd_buf) | ||
1246 | return -ENOMEM; | ||
1247 | |||
1248 | channel->ccw.cmd_code = ciw->cmd; | ||
1249 | channel->ccw.cda = (__u32) __pa (rcd_buf); | ||
1250 | channel->ccw.count = ciw->count; | ||
1251 | channel->ccw.flags = CCW_FLAG_SLI; | ||
1252 | channel->state = CH_STATE_RCD; | ||
1253 | spin_lock_irqsave(get_ccwdev_lock(channel->ccwdev), flags); | ||
1254 | ret = ccw_device_start_timeout(channel->ccwdev, &channel->ccw, | ||
1255 | QETH_RCD_PARM, LPM_ANYPATH, 0, | ||
1256 | QETH_RCD_TIMEOUT); | ||
1257 | spin_unlock_irqrestore(get_ccwdev_lock(channel->ccwdev), flags); | ||
1258 | if (!ret) | ||
1259 | wait_event(card->wait_q, | ||
1260 | (channel->state == CH_STATE_RCD_DONE || | ||
1261 | channel->state == CH_STATE_DOWN)); | ||
1262 | if (channel->state == CH_STATE_DOWN) | ||
1263 | ret = -EIO; | ||
1264 | else | ||
1265 | channel->state = CH_STATE_DOWN; | ||
1266 | if (ret) { | ||
1267 | kfree(rcd_buf); | ||
1268 | *buffer = NULL; | ||
1269 | *length = 0; | ||
1270 | } else { | ||
1271 | *length = ciw->count; | ||
1272 | *buffer = rcd_buf; | ||
1273 | } | ||
1274 | return ret; | ||
1275 | } | ||
1276 | |||
1277 | static int | ||
1278 | qeth_get_unitaddr(struct qeth_card *card) | ||
1279 | { | ||
1280 | int length; | ||
1281 | char *prcd; | ||
1282 | int rc; | ||
1283 | |||
1284 | QETH_DBF_TEXT(setup, 2, "getunit"); | ||
1285 | rc = qeth_read_conf_data(card, (void **) &prcd, &length); | ||
1286 | if (rc) { | ||
1287 | PRINT_ERR("qeth_read_conf_data for device %s returned %i\n", | ||
1288 | CARD_DDEV_ID(card), rc); | ||
1289 | return rc; | ||
1290 | } | ||
1291 | card->info.chpid = prcd[30]; | ||
1292 | card->info.unit_addr2 = prcd[31]; | ||
1293 | card->info.cula = prcd[63]; | ||
1294 | card->info.guestlan = ((prcd[0x10] == _ascebc['V']) && | ||
1295 | (prcd[0x11] == _ascebc['M'])); | ||
1296 | kfree(prcd); | ||
1297 | return 0; | ||
1298 | } | ||
1299 | |||
1300 | static void | ||
1301 | qeth_init_tokens(struct qeth_card *card) | ||
1302 | { | ||
1303 | card->token.issuer_rm_w = 0x00010103UL; | ||
1304 | card->token.cm_filter_w = 0x00010108UL; | ||
1305 | card->token.cm_connection_w = 0x0001010aUL; | ||
1306 | card->token.ulp_filter_w = 0x0001010bUL; | ||
1307 | card->token.ulp_connection_w = 0x0001010dUL; | ||
1308 | } | ||
1309 | |||
1310 | static inline __u16 | ||
1311 | raw_devno_from_bus_id(char *id) | ||
1312 | { | ||
1313 | id += (strlen(id) - 4); | ||
1314 | return (__u16) simple_strtoul(id, &id, 16); | ||
1315 | } | ||
1316 | /** | ||
1317 | * setup channel | ||
1318 | */ | ||
1319 | static void | ||
1320 | qeth_setup_ccw(struct qeth_channel *channel,unsigned char *iob, __u32 len) | ||
1321 | { | ||
1322 | struct qeth_card *card; | ||
1323 | |||
1324 | QETH_DBF_TEXT(trace, 4, "setupccw"); | ||
1325 | card = CARD_FROM_CDEV(channel->ccwdev); | ||
1326 | if (channel == &card->read) | ||
1327 | memcpy(&channel->ccw, READ_CCW, sizeof(struct ccw1)); | ||
1328 | else | ||
1329 | memcpy(&channel->ccw, WRITE_CCW, sizeof(struct ccw1)); | ||
1330 | channel->ccw.count = len; | ||
1331 | channel->ccw.cda = (__u32) __pa(iob); | ||
1332 | } | ||
1333 | |||
1334 | /** | ||
1335 | * get free buffer for ccws (IDX activation, lancmds,ipassists...) | ||
1336 | */ | ||
1337 | static struct qeth_cmd_buffer * | ||
1338 | __qeth_get_buffer(struct qeth_channel *channel) | ||
1339 | { | ||
1340 | __u8 index; | ||
1341 | |||
1342 | QETH_DBF_TEXT(trace, 6, "getbuff"); | ||
1343 | index = channel->io_buf_no; | ||
1344 | do { | ||
1345 | if (channel->iob[index].state == BUF_STATE_FREE) { | ||
1346 | channel->iob[index].state = BUF_STATE_LOCKED; | ||
1347 | channel->io_buf_no = (channel->io_buf_no + 1) % | ||
1348 | QETH_CMD_BUFFER_NO; | ||
1349 | memset(channel->iob[index].data, 0, QETH_BUFSIZE); | ||
1350 | return channel->iob + index; | ||
1351 | } | ||
1352 | index = (index + 1) % QETH_CMD_BUFFER_NO; | ||
1353 | } while(index != channel->io_buf_no); | ||
1354 | |||
1355 | return NULL; | ||
1356 | } | ||
1357 | |||
1358 | /** | ||
1359 | * release command buffer | ||
1360 | */ | ||
1361 | static void | ||
1362 | qeth_release_buffer(struct qeth_channel *channel, struct qeth_cmd_buffer *iob) | ||
1363 | { | ||
1364 | unsigned long flags; | ||
1365 | |||
1366 | QETH_DBF_TEXT(trace, 6, "relbuff"); | ||
1367 | spin_lock_irqsave(&channel->iob_lock, flags); | ||
1368 | memset(iob->data, 0, QETH_BUFSIZE); | ||
1369 | iob->state = BUF_STATE_FREE; | ||
1370 | iob->callback = qeth_send_control_data_cb; | ||
1371 | iob->rc = 0; | ||
1372 | spin_unlock_irqrestore(&channel->iob_lock, flags); | ||
1373 | } | ||
1374 | |||
1375 | static struct qeth_cmd_buffer * | ||
1376 | qeth_get_buffer(struct qeth_channel *channel) | ||
1377 | { | ||
1378 | struct qeth_cmd_buffer *buffer = NULL; | ||
1379 | unsigned long flags; | ||
1380 | |||
1381 | spin_lock_irqsave(&channel->iob_lock, flags); | ||
1382 | buffer = __qeth_get_buffer(channel); | ||
1383 | spin_unlock_irqrestore(&channel->iob_lock, flags); | ||
1384 | return buffer; | ||
1385 | } | ||
1386 | |||
1387 | static struct qeth_cmd_buffer * | ||
1388 | qeth_wait_for_buffer(struct qeth_channel *channel) | ||
1389 | { | ||
1390 | struct qeth_cmd_buffer *buffer; | ||
1391 | wait_event(channel->wait_q, | ||
1392 | ((buffer = qeth_get_buffer(channel)) != NULL)); | ||
1393 | return buffer; | ||
1394 | } | ||
1395 | |||
1396 | static void | ||
1397 | qeth_clear_cmd_buffers(struct qeth_channel *channel) | ||
1398 | { | ||
1399 | int cnt; | ||
1400 | |||
1401 | for (cnt=0; cnt < QETH_CMD_BUFFER_NO; cnt++) | ||
1402 | qeth_release_buffer(channel,&channel->iob[cnt]); | ||
1403 | channel->buf_no = 0; | ||
1404 | channel->io_buf_no = 0; | ||
1405 | } | ||
1406 | |||
1407 | /** | ||
1408 | * start IDX for read and write channel | ||
1409 | */ | ||
1410 | static int | ||
1411 | qeth_idx_activate_get_answer(struct qeth_channel *channel, | ||
1412 | void (*idx_reply_cb)(struct qeth_channel *, | ||
1413 | struct qeth_cmd_buffer *)) | ||
1414 | { | ||
1415 | struct qeth_cmd_buffer *iob; | ||
1416 | unsigned long flags; | ||
1417 | int rc; | ||
1418 | struct qeth_card *card; | ||
1419 | |||
1420 | QETH_DBF_TEXT(setup, 2, "idxanswr"); | ||
1421 | card = CARD_FROM_CDEV(channel->ccwdev); | ||
1422 | iob = qeth_get_buffer(channel); | ||
1423 | iob->callback = idx_reply_cb; | ||
1424 | memcpy(&channel->ccw, READ_CCW, sizeof(struct ccw1)); | ||
1425 | channel->ccw.count = QETH_BUFSIZE; | ||
1426 | channel->ccw.cda = (__u32) __pa(iob->data); | ||
1427 | |||
1428 | wait_event(card->wait_q, | ||
1429 | atomic_cmpxchg(&channel->irq_pending, 0, 1) == 0); | ||
1430 | QETH_DBF_TEXT(setup, 6, "noirqpnd"); | ||
1431 | spin_lock_irqsave(get_ccwdev_lock(channel->ccwdev), flags); | ||
1432 | rc = ccw_device_start(channel->ccwdev, | ||
1433 | &channel->ccw,(addr_t) iob, 0, 0); | ||
1434 | spin_unlock_irqrestore(get_ccwdev_lock(channel->ccwdev), flags); | ||
1435 | |||
1436 | if (rc) { | ||
1437 | PRINT_ERR("qeth: Error2 in activating channel rc=%d\n",rc); | ||
1438 | QETH_DBF_TEXT_(setup, 2, "2err%d", rc); | ||
1439 | atomic_set(&channel->irq_pending, 0); | ||
1440 | wake_up(&card->wait_q); | ||
1441 | return rc; | ||
1442 | } | ||
1443 | rc = wait_event_interruptible_timeout(card->wait_q, | ||
1444 | channel->state == CH_STATE_UP, QETH_TIMEOUT); | ||
1445 | if (rc == -ERESTARTSYS) | ||
1446 | return rc; | ||
1447 | if (channel->state != CH_STATE_UP){ | ||
1448 | rc = -ETIME; | ||
1449 | QETH_DBF_TEXT_(setup, 2, "3err%d", rc); | ||
1450 | qeth_clear_cmd_buffers(channel); | ||
1451 | } else | ||
1452 | rc = 0; | ||
1453 | return rc; | ||
1454 | } | ||
1455 | |||
1456 | static int | ||
1457 | qeth_idx_activate_channel(struct qeth_channel *channel, | ||
1458 | void (*idx_reply_cb)(struct qeth_channel *, | ||
1459 | struct qeth_cmd_buffer *)) | ||
1460 | { | ||
1461 | struct qeth_card *card; | ||
1462 | struct qeth_cmd_buffer *iob; | ||
1463 | unsigned long flags; | ||
1464 | __u16 temp; | ||
1465 | int rc; | ||
1466 | |||
1467 | card = CARD_FROM_CDEV(channel->ccwdev); | ||
1468 | |||
1469 | QETH_DBF_TEXT(setup, 2, "idxactch"); | ||
1470 | |||
1471 | iob = qeth_get_buffer(channel); | ||
1472 | iob->callback = idx_reply_cb; | ||
1473 | memcpy(&channel->ccw, WRITE_CCW, sizeof(struct ccw1)); | ||
1474 | channel->ccw.count = IDX_ACTIVATE_SIZE; | ||
1475 | channel->ccw.cda = (__u32) __pa(iob->data); | ||
1476 | if (channel == &card->write) { | ||
1477 | memcpy(iob->data, IDX_ACTIVATE_WRITE, IDX_ACTIVATE_SIZE); | ||
1478 | memcpy(QETH_TRANSPORT_HEADER_SEQ_NO(iob->data), | ||
1479 | &card->seqno.trans_hdr, QETH_SEQ_NO_LENGTH); | ||
1480 | card->seqno.trans_hdr++; | ||
1481 | } else { | ||
1482 | memcpy(iob->data, IDX_ACTIVATE_READ, IDX_ACTIVATE_SIZE); | ||
1483 | memcpy(QETH_TRANSPORT_HEADER_SEQ_NO(iob->data), | ||
1484 | &card->seqno.trans_hdr, QETH_SEQ_NO_LENGTH); | ||
1485 | } | ||
1486 | memcpy(QETH_IDX_ACT_ISSUER_RM_TOKEN(iob->data), | ||
1487 | &card->token.issuer_rm_w,QETH_MPC_TOKEN_LENGTH); | ||
1488 | memcpy(QETH_IDX_ACT_FUNC_LEVEL(iob->data), | ||
1489 | &card->info.func_level,sizeof(__u16)); | ||
1490 | temp = raw_devno_from_bus_id(CARD_DDEV_ID(card)); | ||
1491 | memcpy(QETH_IDX_ACT_QDIO_DEV_CUA(iob->data), &temp, 2); | ||
1492 | temp = (card->info.cula << 8) + card->info.unit_addr2; | ||
1493 | memcpy(QETH_IDX_ACT_QDIO_DEV_REALADDR(iob->data), &temp, 2); | ||
1494 | |||
1495 | wait_event(card->wait_q, | ||
1496 | atomic_cmpxchg(&channel->irq_pending, 0, 1) == 0); | ||
1497 | QETH_DBF_TEXT(setup, 6, "noirqpnd"); | ||
1498 | spin_lock_irqsave(get_ccwdev_lock(channel->ccwdev), flags); | ||
1499 | rc = ccw_device_start(channel->ccwdev, | ||
1500 | &channel->ccw,(addr_t) iob, 0, 0); | ||
1501 | spin_unlock_irqrestore(get_ccwdev_lock(channel->ccwdev), flags); | ||
1502 | |||
1503 | if (rc) { | ||
1504 | PRINT_ERR("qeth: Error1 in activating channel. rc=%d\n",rc); | ||
1505 | QETH_DBF_TEXT_(setup, 2, "1err%d", rc); | ||
1506 | atomic_set(&channel->irq_pending, 0); | ||
1507 | wake_up(&card->wait_q); | ||
1508 | return rc; | ||
1509 | } | ||
1510 | rc = wait_event_interruptible_timeout(card->wait_q, | ||
1511 | channel->state == CH_STATE_ACTIVATING, QETH_TIMEOUT); | ||
1512 | if (rc == -ERESTARTSYS) | ||
1513 | return rc; | ||
1514 | if (channel->state != CH_STATE_ACTIVATING) { | ||
1515 | PRINT_WARN("qeth: IDX activate timed out!\n"); | ||
1516 | QETH_DBF_TEXT_(setup, 2, "2err%d", -ETIME); | ||
1517 | qeth_clear_cmd_buffers(channel); | ||
1518 | return -ETIME; | ||
1519 | } | ||
1520 | return qeth_idx_activate_get_answer(channel,idx_reply_cb); | ||
1521 | } | ||
1522 | |||
1523 | static int | ||
1524 | qeth_peer_func_level(int level) | ||
1525 | { | ||
1526 | if ((level & 0xff) == 8) | ||
1527 | return (level & 0xff) + 0x400; | ||
1528 | if (((level >> 8) & 3) == 1) | ||
1529 | return (level & 0xff) + 0x200; | ||
1530 | return level; | ||
1531 | } | ||
1532 | |||
1533 | static void | ||
1534 | qeth_idx_write_cb(struct qeth_channel *channel, struct qeth_cmd_buffer *iob) | ||
1535 | { | ||
1536 | struct qeth_card *card; | ||
1537 | __u16 temp; | ||
1538 | |||
1539 | QETH_DBF_TEXT(setup ,2, "idxwrcb"); | ||
1540 | |||
1541 | if (channel->state == CH_STATE_DOWN) { | ||
1542 | channel->state = CH_STATE_ACTIVATING; | ||
1543 | goto out; | ||
1544 | } | ||
1545 | card = CARD_FROM_CDEV(channel->ccwdev); | ||
1546 | |||
1547 | if (!(QETH_IS_IDX_ACT_POS_REPLY(iob->data))) { | ||
1548 | if (QETH_IDX_ACT_CAUSE_CODE(iob->data) == 0x19) | ||
1549 | PRINT_ERR("IDX_ACTIVATE on write channel device %s: " | ||
1550 | "adapter exclusively used by another host\n", | ||
1551 | CARD_WDEV_ID(card)); | ||
1552 | else | ||
1553 | PRINT_ERR("IDX_ACTIVATE on write channel device %s: " | ||
1554 | "negative reply\n", CARD_WDEV_ID(card)); | ||
1555 | goto out; | ||
1556 | } | ||
1557 | memcpy(&temp, QETH_IDX_ACT_FUNC_LEVEL(iob->data), 2); | ||
1558 | if ((temp & ~0x0100) != qeth_peer_func_level(card->info.func_level)) { | ||
1559 | PRINT_WARN("IDX_ACTIVATE on write channel device %s: " | ||
1560 | "function level mismatch " | ||
1561 | "(sent: 0x%x, received: 0x%x)\n", | ||
1562 | CARD_WDEV_ID(card), card->info.func_level, temp); | ||
1563 | goto out; | ||
1564 | } | ||
1565 | channel->state = CH_STATE_UP; | ||
1566 | out: | ||
1567 | qeth_release_buffer(channel, iob); | ||
1568 | } | ||
1569 | |||
1570 | static int | ||
1571 | qeth_check_idx_response(unsigned char *buffer) | ||
1572 | { | ||
1573 | if (!buffer) | ||
1574 | return 0; | ||
1575 | |||
1576 | QETH_DBF_HEX(control, 2, buffer, QETH_DBF_CONTROL_LEN); | ||
1577 | if ((buffer[2] & 0xc0) == 0xc0) { | ||
1578 | PRINT_WARN("received an IDX TERMINATE " | ||
1579 | "with cause code 0x%02x%s\n", | ||
1580 | buffer[4], | ||
1581 | ((buffer[4] == 0x22) ? | ||
1582 | " -- try another portname" : "")); | ||
1583 | QETH_DBF_TEXT(trace, 2, "ckidxres"); | ||
1584 | QETH_DBF_TEXT(trace, 2, " idxterm"); | ||
1585 | QETH_DBF_TEXT_(trace, 2, " rc%d", -EIO); | ||
1586 | return -EIO; | ||
1587 | } | ||
1588 | return 0; | ||
1589 | } | ||
1590 | |||
1591 | static void | ||
1592 | qeth_idx_read_cb(struct qeth_channel *channel, struct qeth_cmd_buffer *iob) | ||
1593 | { | ||
1594 | struct qeth_card *card; | ||
1595 | __u16 temp; | ||
1596 | |||
1597 | QETH_DBF_TEXT(setup , 2, "idxrdcb"); | ||
1598 | if (channel->state == CH_STATE_DOWN) { | ||
1599 | channel->state = CH_STATE_ACTIVATING; | ||
1600 | goto out; | ||
1601 | } | ||
1602 | |||
1603 | card = CARD_FROM_CDEV(channel->ccwdev); | ||
1604 | if (qeth_check_idx_response(iob->data)) { | ||
1605 | goto out; | ||
1606 | } | ||
1607 | if (!(QETH_IS_IDX_ACT_POS_REPLY(iob->data))) { | ||
1608 | if (QETH_IDX_ACT_CAUSE_CODE(iob->data) == 0x19) | ||
1609 | PRINT_ERR("IDX_ACTIVATE on read channel device %s: " | ||
1610 | "adapter exclusively used by another host\n", | ||
1611 | CARD_RDEV_ID(card)); | ||
1612 | else | ||
1613 | PRINT_ERR("IDX_ACTIVATE on read channel device %s: " | ||
1614 | "negative reply\n", CARD_RDEV_ID(card)); | ||
1615 | goto out; | ||
1616 | } | ||
1617 | |||
1618 | /** | ||
1619 | * temporary fix for microcode bug | ||
1620 | * to revert it,replace OR by AND | ||
1621 | */ | ||
1622 | if ( (!QETH_IDX_NO_PORTNAME_REQUIRED(iob->data)) || | ||
1623 | (card->info.type == QETH_CARD_TYPE_OSAE) ) | ||
1624 | card->info.portname_required = 1; | ||
1625 | |||
1626 | memcpy(&temp, QETH_IDX_ACT_FUNC_LEVEL(iob->data), 2); | ||
1627 | if (temp != qeth_peer_func_level(card->info.func_level)) { | ||
1628 | PRINT_WARN("IDX_ACTIVATE on read channel device %s: function " | ||
1629 | "level mismatch (sent: 0x%x, received: 0x%x)\n", | ||
1630 | CARD_RDEV_ID(card), card->info.func_level, temp); | ||
1631 | goto out; | ||
1632 | } | ||
1633 | memcpy(&card->token.issuer_rm_r, | ||
1634 | QETH_IDX_ACT_ISSUER_RM_TOKEN(iob->data), | ||
1635 | QETH_MPC_TOKEN_LENGTH); | ||
1636 | memcpy(&card->info.mcl_level[0], | ||
1637 | QETH_IDX_REPLY_LEVEL(iob->data), QETH_MCL_LENGTH); | ||
1638 | channel->state = CH_STATE_UP; | ||
1639 | out: | ||
1640 | qeth_release_buffer(channel,iob); | ||
1641 | } | ||
1642 | |||
1643 | static int | ||
1644 | qeth_issue_next_read(struct qeth_card *card) | ||
1645 | { | ||
1646 | int rc; | ||
1647 | struct qeth_cmd_buffer *iob; | ||
1648 | |||
1649 | QETH_DBF_TEXT(trace,5,"issnxrd"); | ||
1650 | if (card->read.state != CH_STATE_UP) | ||
1651 | return -EIO; | ||
1652 | iob = qeth_get_buffer(&card->read); | ||
1653 | if (!iob) { | ||
1654 | PRINT_WARN("issue_next_read failed: no iob available!\n"); | ||
1655 | return -ENOMEM; | ||
1656 | } | ||
1657 | qeth_setup_ccw(&card->read, iob->data, QETH_BUFSIZE); | ||
1658 | QETH_DBF_TEXT(trace, 6, "noirqpnd"); | ||
1659 | rc = ccw_device_start(card->read.ccwdev, &card->read.ccw, | ||
1660 | (addr_t) iob, 0, 0); | ||
1661 | if (rc) { | ||
1662 | PRINT_ERR("Error in starting next read ccw! rc=%i\n", rc); | ||
1663 | atomic_set(&card->read.irq_pending, 0); | ||
1664 | qeth_schedule_recovery(card); | ||
1665 | wake_up(&card->wait_q); | ||
1666 | } | ||
1667 | return rc; | ||
1668 | } | ||
1669 | |||
1670 | static struct qeth_reply * | ||
1671 | qeth_alloc_reply(struct qeth_card *card) | ||
1672 | { | ||
1673 | struct qeth_reply *reply; | ||
1674 | |||
1675 | reply = kzalloc(sizeof(struct qeth_reply), GFP_ATOMIC); | ||
1676 | if (reply){ | ||
1677 | atomic_set(&reply->refcnt, 1); | ||
1678 | atomic_set(&reply->received, 0); | ||
1679 | reply->card = card; | ||
1680 | }; | ||
1681 | return reply; | ||
1682 | } | ||
1683 | |||
1684 | static void | ||
1685 | qeth_get_reply(struct qeth_reply *reply) | ||
1686 | { | ||
1687 | WARN_ON(atomic_read(&reply->refcnt) <= 0); | ||
1688 | atomic_inc(&reply->refcnt); | ||
1689 | } | ||
1690 | |||
1691 | static void | ||
1692 | qeth_put_reply(struct qeth_reply *reply) | ||
1693 | { | ||
1694 | WARN_ON(atomic_read(&reply->refcnt) <= 0); | ||
1695 | if (atomic_dec_and_test(&reply->refcnt)) | ||
1696 | kfree(reply); | ||
1697 | } | ||
1698 | |||
1699 | static void | ||
1700 | qeth_issue_ipa_msg(struct qeth_ipa_cmd *cmd, struct qeth_card *card) | ||
1701 | { | ||
1702 | int rc; | ||
1703 | int com; | ||
1704 | char * ipa_name; | ||
1705 | |||
1706 | com = cmd->hdr.command; | ||
1707 | rc = cmd->hdr.return_code; | ||
1708 | ipa_name = qeth_get_ipa_cmd_name(com); | ||
1709 | |||
1710 | PRINT_ERR("%s(x%X) for %s returned x%X \"%s\"\n", ipa_name, com, | ||
1711 | QETH_CARD_IFNAME(card), rc, qeth_get_ipa_msg(rc)); | ||
1712 | } | ||
1713 | |||
1714 | static struct qeth_ipa_cmd * | ||
1715 | qeth_check_ipa_data(struct qeth_card *card, struct qeth_cmd_buffer *iob) | ||
1716 | { | ||
1717 | struct qeth_ipa_cmd *cmd = NULL; | ||
1718 | |||
1719 | QETH_DBF_TEXT(trace,5,"chkipad"); | ||
1720 | if (IS_IPA(iob->data)){ | ||
1721 | cmd = (struct qeth_ipa_cmd *) PDU_ENCAPSULATION(iob->data); | ||
1722 | if (IS_IPA_REPLY(cmd)) { | ||
1723 | if (cmd->hdr.return_code) | ||
1724 | qeth_issue_ipa_msg(cmd, card); | ||
1725 | return cmd; | ||
1726 | } | ||
1727 | else { | ||
1728 | switch (cmd->hdr.command) { | ||
1729 | case IPA_CMD_STOPLAN: | ||
1730 | PRINT_WARN("Link failure on %s (CHPID 0x%X) - " | ||
1731 | "there is a network problem or " | ||
1732 | "someone pulled the cable or " | ||
1733 | "disabled the port.\n", | ||
1734 | QETH_CARD_IFNAME(card), | ||
1735 | card->info.chpid); | ||
1736 | card->lan_online = 0; | ||
1737 | if (card->dev && netif_carrier_ok(card->dev)) | ||
1738 | netif_carrier_off(card->dev); | ||
1739 | return NULL; | ||
1740 | case IPA_CMD_STARTLAN: | ||
1741 | PRINT_INFO("Link reestablished on %s " | ||
1742 | "(CHPID 0x%X). Scheduling " | ||
1743 | "IP address reset.\n", | ||
1744 | QETH_CARD_IFNAME(card), | ||
1745 | card->info.chpid); | ||
1746 | netif_carrier_on(card->dev); | ||
1747 | qeth_schedule_recovery(card); | ||
1748 | return NULL; | ||
1749 | case IPA_CMD_MODCCID: | ||
1750 | return cmd; | ||
1751 | case IPA_CMD_REGISTER_LOCAL_ADDR: | ||
1752 | QETH_DBF_TEXT(trace,3, "irla"); | ||
1753 | break; | ||
1754 | case IPA_CMD_UNREGISTER_LOCAL_ADDR: | ||
1755 | QETH_DBF_TEXT(trace,3, "urla"); | ||
1756 | break; | ||
1757 | default: | ||
1758 | PRINT_WARN("Received data is IPA " | ||
1759 | "but not a reply!\n"); | ||
1760 | break; | ||
1761 | } | ||
1762 | } | ||
1763 | } | ||
1764 | return cmd; | ||
1765 | } | ||
1766 | |||
1767 | /** | ||
1768 | * wake all waiting ipa commands | ||
1769 | */ | ||
1770 | static void | ||
1771 | qeth_clear_ipacmd_list(struct qeth_card *card) | ||
1772 | { | ||
1773 | struct qeth_reply *reply, *r; | ||
1774 | unsigned long flags; | ||
1775 | |||
1776 | QETH_DBF_TEXT(trace, 4, "clipalst"); | ||
1777 | |||
1778 | spin_lock_irqsave(&card->lock, flags); | ||
1779 | list_for_each_entry_safe(reply, r, &card->cmd_waiter_list, list) { | ||
1780 | qeth_get_reply(reply); | ||
1781 | reply->rc = -EIO; | ||
1782 | atomic_inc(&reply->received); | ||
1783 | list_del_init(&reply->list); | ||
1784 | wake_up(&reply->wait_q); | ||
1785 | qeth_put_reply(reply); | ||
1786 | } | ||
1787 | spin_unlock_irqrestore(&card->lock, flags); | ||
1788 | } | ||
1789 | |||
1790 | static void | ||
1791 | qeth_send_control_data_cb(struct qeth_channel *channel, | ||
1792 | struct qeth_cmd_buffer *iob) | ||
1793 | { | ||
1794 | struct qeth_card *card; | ||
1795 | struct qeth_reply *reply, *r; | ||
1796 | struct qeth_ipa_cmd *cmd; | ||
1797 | unsigned long flags; | ||
1798 | int keep_reply; | ||
1799 | |||
1800 | QETH_DBF_TEXT(trace,4,"sndctlcb"); | ||
1801 | |||
1802 | card = CARD_FROM_CDEV(channel->ccwdev); | ||
1803 | if (qeth_check_idx_response(iob->data)) { | ||
1804 | qeth_clear_ipacmd_list(card); | ||
1805 | qeth_schedule_recovery(card); | ||
1806 | goto out; | ||
1807 | } | ||
1808 | |||
1809 | cmd = qeth_check_ipa_data(card, iob); | ||
1810 | if ((cmd == NULL) && (card->state != CARD_STATE_DOWN)) | ||
1811 | goto out; | ||
1812 | /*in case of OSN : check if cmd is set */ | ||
1813 | if (card->info.type == QETH_CARD_TYPE_OSN && | ||
1814 | cmd && | ||
1815 | cmd->hdr.command != IPA_CMD_STARTLAN && | ||
1816 | card->osn_info.assist_cb != NULL) { | ||
1817 | card->osn_info.assist_cb(card->dev, cmd); | ||
1818 | goto out; | ||
1819 | } | ||
1820 | |||
1821 | spin_lock_irqsave(&card->lock, flags); | ||
1822 | list_for_each_entry_safe(reply, r, &card->cmd_waiter_list, list) { | ||
1823 | if ((reply->seqno == QETH_IDX_COMMAND_SEQNO) || | ||
1824 | ((cmd) && (reply->seqno == cmd->hdr.seqno))) { | ||
1825 | qeth_get_reply(reply); | ||
1826 | list_del_init(&reply->list); | ||
1827 | spin_unlock_irqrestore(&card->lock, flags); | ||
1828 | keep_reply = 0; | ||
1829 | if (reply->callback != NULL) { | ||
1830 | if (cmd) { | ||
1831 | reply->offset = (__u16)((char*)cmd - | ||
1832 | (char *)iob->data); | ||
1833 | keep_reply = reply->callback(card, | ||
1834 | reply, | ||
1835 | (unsigned long)cmd); | ||
1836 | } else | ||
1837 | keep_reply = reply->callback(card, | ||
1838 | reply, | ||
1839 | (unsigned long)iob); | ||
1840 | } | ||
1841 | if (cmd) | ||
1842 | reply->rc = (u16) cmd->hdr.return_code; | ||
1843 | else if (iob->rc) | ||
1844 | reply->rc = iob->rc; | ||
1845 | if (keep_reply) { | ||
1846 | spin_lock_irqsave(&card->lock, flags); | ||
1847 | list_add_tail(&reply->list, | ||
1848 | &card->cmd_waiter_list); | ||
1849 | spin_unlock_irqrestore(&card->lock, flags); | ||
1850 | } else { | ||
1851 | atomic_inc(&reply->received); | ||
1852 | wake_up(&reply->wait_q); | ||
1853 | } | ||
1854 | qeth_put_reply(reply); | ||
1855 | goto out; | ||
1856 | } | ||
1857 | } | ||
1858 | spin_unlock_irqrestore(&card->lock, flags); | ||
1859 | out: | ||
1860 | memcpy(&card->seqno.pdu_hdr_ack, | ||
1861 | QETH_PDU_HEADER_SEQ_NO(iob->data), | ||
1862 | QETH_SEQ_NO_LENGTH); | ||
1863 | qeth_release_buffer(channel,iob); | ||
1864 | } | ||
1865 | |||
1866 | static void | ||
1867 | qeth_prepare_control_data(struct qeth_card *card, int len, | ||
1868 | struct qeth_cmd_buffer *iob) | ||
1869 | { | ||
1870 | qeth_setup_ccw(&card->write,iob->data,len); | ||
1871 | iob->callback = qeth_release_buffer; | ||
1872 | |||
1873 | memcpy(QETH_TRANSPORT_HEADER_SEQ_NO(iob->data), | ||
1874 | &card->seqno.trans_hdr, QETH_SEQ_NO_LENGTH); | ||
1875 | card->seqno.trans_hdr++; | ||
1876 | memcpy(QETH_PDU_HEADER_SEQ_NO(iob->data), | ||
1877 | &card->seqno.pdu_hdr, QETH_SEQ_NO_LENGTH); | ||
1878 | card->seqno.pdu_hdr++; | ||
1879 | memcpy(QETH_PDU_HEADER_ACK_SEQ_NO(iob->data), | ||
1880 | &card->seqno.pdu_hdr_ack, QETH_SEQ_NO_LENGTH); | ||
1881 | QETH_DBF_HEX(control, 2, iob->data, QETH_DBF_CONTROL_LEN); | ||
1882 | } | ||
1883 | |||
1884 | static int | ||
1885 | qeth_send_control_data(struct qeth_card *card, int len, | ||
1886 | struct qeth_cmd_buffer *iob, | ||
1887 | int (*reply_cb) | ||
1888 | (struct qeth_card *, struct qeth_reply*, unsigned long), | ||
1889 | void *reply_param) | ||
1890 | |||
1891 | { | ||
1892 | int rc; | ||
1893 | unsigned long flags; | ||
1894 | struct qeth_reply *reply = NULL; | ||
1895 | unsigned long timeout; | ||
1896 | |||
1897 | QETH_DBF_TEXT(trace, 2, "sendctl"); | ||
1898 | |||
1899 | reply = qeth_alloc_reply(card); | ||
1900 | if (!reply) { | ||
1901 | PRINT_WARN("Could no alloc qeth_reply!\n"); | ||
1902 | return -ENOMEM; | ||
1903 | } | ||
1904 | reply->callback = reply_cb; | ||
1905 | reply->param = reply_param; | ||
1906 | if (card->state == CARD_STATE_DOWN) | ||
1907 | reply->seqno = QETH_IDX_COMMAND_SEQNO; | ||
1908 | else | ||
1909 | reply->seqno = card->seqno.ipa++; | ||
1910 | init_waitqueue_head(&reply->wait_q); | ||
1911 | spin_lock_irqsave(&card->lock, flags); | ||
1912 | list_add_tail(&reply->list, &card->cmd_waiter_list); | ||
1913 | spin_unlock_irqrestore(&card->lock, flags); | ||
1914 | QETH_DBF_HEX(control, 2, iob->data, QETH_DBF_CONTROL_LEN); | ||
1915 | |||
1916 | while (atomic_cmpxchg(&card->write.irq_pending, 0, 1)) ; | ||
1917 | qeth_prepare_control_data(card, len, iob); | ||
1918 | |||
1919 | if (IS_IPA(iob->data)) | ||
1920 | timeout = jiffies + QETH_IPA_TIMEOUT; | ||
1921 | else | ||
1922 | timeout = jiffies + QETH_TIMEOUT; | ||
1923 | |||
1924 | QETH_DBF_TEXT(trace, 6, "noirqpnd"); | ||
1925 | spin_lock_irqsave(get_ccwdev_lock(card->write.ccwdev), flags); | ||
1926 | rc = ccw_device_start(card->write.ccwdev, &card->write.ccw, | ||
1927 | (addr_t) iob, 0, 0); | ||
1928 | spin_unlock_irqrestore(get_ccwdev_lock(card->write.ccwdev), flags); | ||
1929 | if (rc){ | ||
1930 | PRINT_WARN("qeth_send_control_data: " | ||
1931 | "ccw_device_start rc = %i\n", rc); | ||
1932 | QETH_DBF_TEXT_(trace, 2, " err%d", rc); | ||
1933 | spin_lock_irqsave(&card->lock, flags); | ||
1934 | list_del_init(&reply->list); | ||
1935 | qeth_put_reply(reply); | ||
1936 | spin_unlock_irqrestore(&card->lock, flags); | ||
1937 | qeth_release_buffer(iob->channel, iob); | ||
1938 | atomic_set(&card->write.irq_pending, 0); | ||
1939 | wake_up(&card->wait_q); | ||
1940 | return rc; | ||
1941 | } | ||
1942 | while (!atomic_read(&reply->received)) { | ||
1943 | if (time_after(jiffies, timeout)) { | ||
1944 | spin_lock_irqsave(&reply->card->lock, flags); | ||
1945 | list_del_init(&reply->list); | ||
1946 | spin_unlock_irqrestore(&reply->card->lock, flags); | ||
1947 | reply->rc = -ETIME; | ||
1948 | atomic_inc(&reply->received); | ||
1949 | wake_up(&reply->wait_q); | ||
1950 | } | ||
1951 | cpu_relax(); | ||
1952 | }; | ||
1953 | rc = reply->rc; | ||
1954 | qeth_put_reply(reply); | ||
1955 | return rc; | ||
1956 | } | ||
1957 | |||
1958 | static int | ||
1959 | qeth_osn_send_control_data(struct qeth_card *card, int len, | ||
1960 | struct qeth_cmd_buffer *iob) | ||
1961 | { | ||
1962 | unsigned long flags; | ||
1963 | int rc = 0; | ||
1964 | |||
1965 | QETH_DBF_TEXT(trace, 5, "osndctrd"); | ||
1966 | |||
1967 | wait_event(card->wait_q, | ||
1968 | atomic_cmpxchg(&card->write.irq_pending, 0, 1) == 0); | ||
1969 | qeth_prepare_control_data(card, len, iob); | ||
1970 | QETH_DBF_TEXT(trace, 6, "osnoirqp"); | ||
1971 | spin_lock_irqsave(get_ccwdev_lock(card->write.ccwdev), flags); | ||
1972 | rc = ccw_device_start(card->write.ccwdev, &card->write.ccw, | ||
1973 | (addr_t) iob, 0, 0); | ||
1974 | spin_unlock_irqrestore(get_ccwdev_lock(card->write.ccwdev), flags); | ||
1975 | if (rc){ | ||
1976 | PRINT_WARN("qeth_osn_send_control_data: " | ||
1977 | "ccw_device_start rc = %i\n", rc); | ||
1978 | QETH_DBF_TEXT_(trace, 2, " err%d", rc); | ||
1979 | qeth_release_buffer(iob->channel, iob); | ||
1980 | atomic_set(&card->write.irq_pending, 0); | ||
1981 | wake_up(&card->wait_q); | ||
1982 | } | ||
1983 | return rc; | ||
1984 | } | ||
1985 | |||
1986 | static inline void | ||
1987 | qeth_prepare_ipa_cmd(struct qeth_card *card, struct qeth_cmd_buffer *iob, | ||
1988 | char prot_type) | ||
1989 | { | ||
1990 | memcpy(iob->data, IPA_PDU_HEADER, IPA_PDU_HEADER_SIZE); | ||
1991 | memcpy(QETH_IPA_CMD_PROT_TYPE(iob->data),&prot_type,1); | ||
1992 | memcpy(QETH_IPA_CMD_DEST_ADDR(iob->data), | ||
1993 | &card->token.ulp_connection_r, QETH_MPC_TOKEN_LENGTH); | ||
1994 | } | ||
1995 | |||
1996 | static int | ||
1997 | qeth_osn_send_ipa_cmd(struct qeth_card *card, struct qeth_cmd_buffer *iob, | ||
1998 | int data_len) | ||
1999 | { | ||
2000 | u16 s1, s2; | ||
2001 | |||
2002 | QETH_DBF_TEXT(trace,4,"osndipa"); | ||
2003 | |||
2004 | qeth_prepare_ipa_cmd(card, iob, QETH_PROT_OSN2); | ||
2005 | s1 = (u16)(IPA_PDU_HEADER_SIZE + data_len); | ||
2006 | s2 = (u16)data_len; | ||
2007 | memcpy(QETH_IPA_PDU_LEN_TOTAL(iob->data), &s1, 2); | ||
2008 | memcpy(QETH_IPA_PDU_LEN_PDU1(iob->data), &s2, 2); | ||
2009 | memcpy(QETH_IPA_PDU_LEN_PDU2(iob->data), &s2, 2); | ||
2010 | memcpy(QETH_IPA_PDU_LEN_PDU3(iob->data), &s2, 2); | ||
2011 | return qeth_osn_send_control_data(card, s1, iob); | ||
2012 | } | ||
2013 | |||
2014 | static int | ||
2015 | qeth_send_ipa_cmd(struct qeth_card *card, struct qeth_cmd_buffer *iob, | ||
2016 | int (*reply_cb) | ||
2017 | (struct qeth_card *,struct qeth_reply*, unsigned long), | ||
2018 | void *reply_param) | ||
2019 | { | ||
2020 | int rc; | ||
2021 | char prot_type; | ||
2022 | |||
2023 | QETH_DBF_TEXT(trace,4,"sendipa"); | ||
2024 | |||
2025 | if (card->options.layer2) | ||
2026 | if (card->info.type == QETH_CARD_TYPE_OSN) | ||
2027 | prot_type = QETH_PROT_OSN2; | ||
2028 | else | ||
2029 | prot_type = QETH_PROT_LAYER2; | ||
2030 | else | ||
2031 | prot_type = QETH_PROT_TCPIP; | ||
2032 | qeth_prepare_ipa_cmd(card,iob,prot_type); | ||
2033 | rc = qeth_send_control_data(card, IPA_CMD_LENGTH, iob, | ||
2034 | reply_cb, reply_param); | ||
2035 | return rc; | ||
2036 | } | ||
2037 | |||
2038 | |||
2039 | static int | ||
2040 | qeth_cm_enable_cb(struct qeth_card *card, struct qeth_reply *reply, | ||
2041 | unsigned long data) | ||
2042 | { | ||
2043 | struct qeth_cmd_buffer *iob; | ||
2044 | |||
2045 | QETH_DBF_TEXT(setup, 2, "cmenblcb"); | ||
2046 | |||
2047 | iob = (struct qeth_cmd_buffer *) data; | ||
2048 | memcpy(&card->token.cm_filter_r, | ||
2049 | QETH_CM_ENABLE_RESP_FILTER_TOKEN(iob->data), | ||
2050 | QETH_MPC_TOKEN_LENGTH); | ||
2051 | QETH_DBF_TEXT_(setup, 2, " rc%d", iob->rc); | ||
2052 | return 0; | ||
2053 | } | ||
2054 | |||
2055 | static int | ||
2056 | qeth_cm_enable(struct qeth_card *card) | ||
2057 | { | ||
2058 | int rc; | ||
2059 | struct qeth_cmd_buffer *iob; | ||
2060 | |||
2061 | QETH_DBF_TEXT(setup,2,"cmenable"); | ||
2062 | |||
2063 | iob = qeth_wait_for_buffer(&card->write); | ||
2064 | memcpy(iob->data, CM_ENABLE, CM_ENABLE_SIZE); | ||
2065 | memcpy(QETH_CM_ENABLE_ISSUER_RM_TOKEN(iob->data), | ||
2066 | &card->token.issuer_rm_r, QETH_MPC_TOKEN_LENGTH); | ||
2067 | memcpy(QETH_CM_ENABLE_FILTER_TOKEN(iob->data), | ||
2068 | &card->token.cm_filter_w, QETH_MPC_TOKEN_LENGTH); | ||
2069 | |||
2070 | rc = qeth_send_control_data(card, CM_ENABLE_SIZE, iob, | ||
2071 | qeth_cm_enable_cb, NULL); | ||
2072 | return rc; | ||
2073 | } | ||
2074 | |||
2075 | static int | ||
2076 | qeth_cm_setup_cb(struct qeth_card *card, struct qeth_reply *reply, | ||
2077 | unsigned long data) | ||
2078 | { | ||
2079 | |||
2080 | struct qeth_cmd_buffer *iob; | ||
2081 | |||
2082 | QETH_DBF_TEXT(setup, 2, "cmsetpcb"); | ||
2083 | |||
2084 | iob = (struct qeth_cmd_buffer *) data; | ||
2085 | memcpy(&card->token.cm_connection_r, | ||
2086 | QETH_CM_SETUP_RESP_DEST_ADDR(iob->data), | ||
2087 | QETH_MPC_TOKEN_LENGTH); | ||
2088 | QETH_DBF_TEXT_(setup, 2, " rc%d", iob->rc); | ||
2089 | return 0; | ||
2090 | } | ||
2091 | |||
2092 | static int | ||
2093 | qeth_cm_setup(struct qeth_card *card) | ||
2094 | { | ||
2095 | int rc; | ||
2096 | struct qeth_cmd_buffer *iob; | ||
2097 | |||
2098 | QETH_DBF_TEXT(setup,2,"cmsetup"); | ||
2099 | |||
2100 | iob = qeth_wait_for_buffer(&card->write); | ||
2101 | memcpy(iob->data, CM_SETUP, CM_SETUP_SIZE); | ||
2102 | memcpy(QETH_CM_SETUP_DEST_ADDR(iob->data), | ||
2103 | &card->token.issuer_rm_r, QETH_MPC_TOKEN_LENGTH); | ||
2104 | memcpy(QETH_CM_SETUP_CONNECTION_TOKEN(iob->data), | ||
2105 | &card->token.cm_connection_w, QETH_MPC_TOKEN_LENGTH); | ||
2106 | memcpy(QETH_CM_SETUP_FILTER_TOKEN(iob->data), | ||
2107 | &card->token.cm_filter_r, QETH_MPC_TOKEN_LENGTH); | ||
2108 | rc = qeth_send_control_data(card, CM_SETUP_SIZE, iob, | ||
2109 | qeth_cm_setup_cb, NULL); | ||
2110 | return rc; | ||
2111 | |||
2112 | } | ||
2113 | |||
2114 | static int | ||
2115 | qeth_ulp_enable_cb(struct qeth_card *card, struct qeth_reply *reply, | ||
2116 | unsigned long data) | ||
2117 | { | ||
2118 | |||
2119 | __u16 mtu, framesize; | ||
2120 | __u16 len; | ||
2121 | __u8 link_type; | ||
2122 | struct qeth_cmd_buffer *iob; | ||
2123 | |||
2124 | QETH_DBF_TEXT(setup, 2, "ulpenacb"); | ||
2125 | |||
2126 | iob = (struct qeth_cmd_buffer *) data; | ||
2127 | memcpy(&card->token.ulp_filter_r, | ||
2128 | QETH_ULP_ENABLE_RESP_FILTER_TOKEN(iob->data), | ||
2129 | QETH_MPC_TOKEN_LENGTH); | ||
2130 | if (qeth_get_mtu_out_of_mpc(card->info.type)) { | ||
2131 | memcpy(&framesize, QETH_ULP_ENABLE_RESP_MAX_MTU(iob->data), 2); | ||
2132 | mtu = qeth_get_mtu_outof_framesize(framesize); | ||
2133 | if (!mtu) { | ||
2134 | iob->rc = -EINVAL; | ||
2135 | QETH_DBF_TEXT_(setup, 2, " rc%d", iob->rc); | ||
2136 | return 0; | ||
2137 | } | ||
2138 | card->info.max_mtu = mtu; | ||
2139 | card->info.initial_mtu = mtu; | ||
2140 | card->qdio.in_buf_size = mtu + 2 * PAGE_SIZE; | ||
2141 | } else { | ||
2142 | card->info.initial_mtu = qeth_get_initial_mtu_for_card(card); | ||
2143 | card->info.max_mtu = qeth_get_max_mtu_for_card(card->info.type); | ||
2144 | card->qdio.in_buf_size = QETH_IN_BUF_SIZE_DEFAULT; | ||
2145 | } | ||
2146 | |||
2147 | memcpy(&len, QETH_ULP_ENABLE_RESP_DIFINFO_LEN(iob->data), 2); | ||
2148 | if (len >= QETH_MPC_DIFINFO_LEN_INDICATES_LINK_TYPE) { | ||
2149 | memcpy(&link_type, | ||
2150 | QETH_ULP_ENABLE_RESP_LINK_TYPE(iob->data), 1); | ||
2151 | card->info.link_type = link_type; | ||
2152 | } else | ||
2153 | card->info.link_type = 0; | ||
2154 | QETH_DBF_TEXT_(setup, 2, " rc%d", iob->rc); | ||
2155 | return 0; | ||
2156 | } | ||
2157 | |||
2158 | static int | ||
2159 | qeth_ulp_enable(struct qeth_card *card) | ||
2160 | { | ||
2161 | int rc; | ||
2162 | char prot_type; | ||
2163 | struct qeth_cmd_buffer *iob; | ||
2164 | |||
2165 | /*FIXME: trace view callbacks*/ | ||
2166 | QETH_DBF_TEXT(setup,2,"ulpenabl"); | ||
2167 | |||
2168 | iob = qeth_wait_for_buffer(&card->write); | ||
2169 | memcpy(iob->data, ULP_ENABLE, ULP_ENABLE_SIZE); | ||
2170 | |||
2171 | *(QETH_ULP_ENABLE_LINKNUM(iob->data)) = | ||
2172 | (__u8) card->info.portno; | ||
2173 | if (card->options.layer2) | ||
2174 | if (card->info.type == QETH_CARD_TYPE_OSN) | ||
2175 | prot_type = QETH_PROT_OSN2; | ||
2176 | else | ||
2177 | prot_type = QETH_PROT_LAYER2; | ||
2178 | else | ||
2179 | prot_type = QETH_PROT_TCPIP; | ||
2180 | |||
2181 | memcpy(QETH_ULP_ENABLE_PROT_TYPE(iob->data),&prot_type,1); | ||
2182 | memcpy(QETH_ULP_ENABLE_DEST_ADDR(iob->data), | ||
2183 | &card->token.cm_connection_r, QETH_MPC_TOKEN_LENGTH); | ||
2184 | memcpy(QETH_ULP_ENABLE_FILTER_TOKEN(iob->data), | ||
2185 | &card->token.ulp_filter_w, QETH_MPC_TOKEN_LENGTH); | ||
2186 | memcpy(QETH_ULP_ENABLE_PORTNAME_AND_LL(iob->data), | ||
2187 | card->info.portname, 9); | ||
2188 | rc = qeth_send_control_data(card, ULP_ENABLE_SIZE, iob, | ||
2189 | qeth_ulp_enable_cb, NULL); | ||
2190 | return rc; | ||
2191 | |||
2192 | } | ||
2193 | |||
2194 | static int | ||
2195 | qeth_ulp_setup_cb(struct qeth_card *card, struct qeth_reply *reply, | ||
2196 | unsigned long data) | ||
2197 | { | ||
2198 | struct qeth_cmd_buffer *iob; | ||
2199 | |||
2200 | QETH_DBF_TEXT(setup, 2, "ulpstpcb"); | ||
2201 | |||
2202 | iob = (struct qeth_cmd_buffer *) data; | ||
2203 | memcpy(&card->token.ulp_connection_r, | ||
2204 | QETH_ULP_SETUP_RESP_CONNECTION_TOKEN(iob->data), | ||
2205 | QETH_MPC_TOKEN_LENGTH); | ||
2206 | QETH_DBF_TEXT_(setup, 2, " rc%d", iob->rc); | ||
2207 | return 0; | ||
2208 | } | ||
2209 | |||
2210 | static int | ||
2211 | qeth_ulp_setup(struct qeth_card *card) | ||
2212 | { | ||
2213 | int rc; | ||
2214 | __u16 temp; | ||
2215 | struct qeth_cmd_buffer *iob; | ||
2216 | struct ccw_dev_id dev_id; | ||
2217 | |||
2218 | QETH_DBF_TEXT(setup,2,"ulpsetup"); | ||
2219 | |||
2220 | iob = qeth_wait_for_buffer(&card->write); | ||
2221 | memcpy(iob->data, ULP_SETUP, ULP_SETUP_SIZE); | ||
2222 | |||
2223 | memcpy(QETH_ULP_SETUP_DEST_ADDR(iob->data), | ||
2224 | &card->token.cm_connection_r, QETH_MPC_TOKEN_LENGTH); | ||
2225 | memcpy(QETH_ULP_SETUP_CONNECTION_TOKEN(iob->data), | ||
2226 | &card->token.ulp_connection_w, QETH_MPC_TOKEN_LENGTH); | ||
2227 | memcpy(QETH_ULP_SETUP_FILTER_TOKEN(iob->data), | ||
2228 | &card->token.ulp_filter_r, QETH_MPC_TOKEN_LENGTH); | ||
2229 | |||
2230 | ccw_device_get_id(CARD_DDEV(card), &dev_id); | ||
2231 | memcpy(QETH_ULP_SETUP_CUA(iob->data), &dev_id.devno, 2); | ||
2232 | temp = (card->info.cula << 8) + card->info.unit_addr2; | ||
2233 | memcpy(QETH_ULP_SETUP_REAL_DEVADDR(iob->data), &temp, 2); | ||
2234 | rc = qeth_send_control_data(card, ULP_SETUP_SIZE, iob, | ||
2235 | qeth_ulp_setup_cb, NULL); | ||
2236 | return rc; | ||
2237 | } | ||
2238 | |||
2239 | static inline int | ||
2240 | qeth_check_qdio_errors(struct qdio_buffer *buf, unsigned int qdio_error, | ||
2241 | unsigned int siga_error, const char *dbftext) | ||
2242 | { | ||
2243 | if (qdio_error || siga_error) { | ||
2244 | QETH_DBF_TEXT(trace, 2, dbftext); | ||
2245 | QETH_DBF_TEXT(qerr, 2, dbftext); | ||
2246 | QETH_DBF_TEXT_(qerr, 2, " F15=%02X", | ||
2247 | buf->element[15].flags & 0xff); | ||
2248 | QETH_DBF_TEXT_(qerr, 2, " F14=%02X", | ||
2249 | buf->element[14].flags & 0xff); | ||
2250 | QETH_DBF_TEXT_(qerr, 2, " qerr=%X", qdio_error); | ||
2251 | QETH_DBF_TEXT_(qerr, 2, " serr=%X", siga_error); | ||
2252 | return 1; | ||
2253 | } | ||
2254 | return 0; | ||
2255 | } | ||
2256 | |||
2257 | static struct sk_buff * | ||
2258 | qeth_get_skb(unsigned int length, struct qeth_hdr *hdr) | ||
2259 | { | ||
2260 | struct sk_buff* skb; | ||
2261 | int add_len; | ||
2262 | |||
2263 | add_len = 0; | ||
2264 | if (hdr->hdr.osn.id == QETH_HEADER_TYPE_OSN) | ||
2265 | add_len = sizeof(struct qeth_hdr); | ||
2266 | #ifdef CONFIG_QETH_VLAN | ||
2267 | else | ||
2268 | add_len = VLAN_HLEN; | ||
2269 | #endif | ||
2270 | skb = dev_alloc_skb(length + add_len); | ||
2271 | if (skb && add_len) | ||
2272 | skb_reserve(skb, add_len); | ||
2273 | return skb; | ||
2274 | } | ||
2275 | |||
2276 | static inline int | ||
2277 | qeth_create_skb_frag(struct qdio_buffer_element *element, | ||
2278 | struct sk_buff **pskb, | ||
2279 | int offset, int *pfrag, int data_len) | ||
2280 | { | ||
2281 | struct page *page = virt_to_page(element->addr); | ||
2282 | if (*pfrag == 0) { | ||
2283 | /* the upper protocol layers assume that there is data in the | ||
2284 | * skb itself. Copy a small amount (64 bytes) to make them | ||
2285 | * happy. */ | ||
2286 | *pskb = dev_alloc_skb(64 + QETH_FAKE_LL_LEN_ETH); | ||
2287 | if (!(*pskb)) | ||
2288 | return -ENOMEM; | ||
2289 | skb_reserve(*pskb, QETH_FAKE_LL_LEN_ETH); | ||
2290 | if (data_len <= 64) { | ||
2291 | memcpy(skb_put(*pskb, data_len), element->addr + offset, | ||
2292 | data_len); | ||
2293 | } else { | ||
2294 | get_page(page); | ||
2295 | memcpy(skb_put(*pskb, 64), element->addr + offset, 64); | ||
2296 | skb_fill_page_desc(*pskb, *pfrag, page, offset + 64, | ||
2297 | data_len - 64); | ||
2298 | (*pskb)->data_len += data_len - 64; | ||
2299 | (*pskb)->len += data_len - 64; | ||
2300 | (*pskb)->truesize += data_len - 64; | ||
2301 | } | ||
2302 | } else { | ||
2303 | get_page(page); | ||
2304 | skb_fill_page_desc(*pskb, *pfrag, page, offset, data_len); | ||
2305 | (*pskb)->data_len += data_len; | ||
2306 | (*pskb)->len += data_len; | ||
2307 | (*pskb)->truesize += data_len; | ||
2308 | } | ||
2309 | (*pfrag)++; | ||
2310 | return 0; | ||
2311 | } | ||
2312 | |||
2313 | static inline struct qeth_buffer_pool_entry * | ||
2314 | qeth_find_free_buffer_pool_entry(struct qeth_card *card) | ||
2315 | { | ||
2316 | struct list_head *plh; | ||
2317 | struct qeth_buffer_pool_entry *entry; | ||
2318 | int i, free; | ||
2319 | struct page *page; | ||
2320 | |||
2321 | if (list_empty(&card->qdio.in_buf_pool.entry_list)) | ||
2322 | return NULL; | ||
2323 | |||
2324 | list_for_each(plh, &card->qdio.in_buf_pool.entry_list) { | ||
2325 | entry = list_entry(plh, struct qeth_buffer_pool_entry, list); | ||
2326 | free = 1; | ||
2327 | for (i = 0; i < QETH_MAX_BUFFER_ELEMENTS(card); ++i) { | ||
2328 | if (page_count(virt_to_page(entry->elements[i])) > 1) { | ||
2329 | free = 0; | ||
2330 | break; | ||
2331 | } | ||
2332 | } | ||
2333 | if (free) { | ||
2334 | list_del_init(&entry->list); | ||
2335 | return entry; | ||
2336 | } | ||
2337 | } | ||
2338 | |||
2339 | /* no free buffer in pool so take first one and swap pages */ | ||
2340 | entry = list_entry(card->qdio.in_buf_pool.entry_list.next, | ||
2341 | struct qeth_buffer_pool_entry, list); | ||
2342 | for (i = 0; i < QETH_MAX_BUFFER_ELEMENTS(card); ++i) { | ||
2343 | if (page_count(virt_to_page(entry->elements[i])) > 1) { | ||
2344 | page = alloc_page(GFP_ATOMIC|GFP_DMA); | ||
2345 | if (!page) { | ||
2346 | return NULL; | ||
2347 | } else { | ||
2348 | free_page((unsigned long)entry->elements[i]); | ||
2349 | entry->elements[i] = page_address(page); | ||
2350 | if (card->options.performance_stats) | ||
2351 | card->perf_stats.sg_alloc_page_rx++; | ||
2352 | } | ||
2353 | } | ||
2354 | } | ||
2355 | list_del_init(&entry->list); | ||
2356 | return entry; | ||
2357 | } | ||
2358 | |||
2359 | static struct sk_buff * | ||
2360 | qeth_get_next_skb(struct qeth_card *card, struct qdio_buffer *buffer, | ||
2361 | struct qdio_buffer_element **__element, int *__offset, | ||
2362 | struct qeth_hdr **hdr) | ||
2363 | { | ||
2364 | struct qdio_buffer_element *element = *__element; | ||
2365 | int offset = *__offset; | ||
2366 | struct sk_buff *skb = NULL; | ||
2367 | int skb_len; | ||
2368 | void *data_ptr; | ||
2369 | int data_len; | ||
2370 | int use_rx_sg = 0; | ||
2371 | int frag = 0; | ||
2372 | |||
2373 | QETH_DBF_TEXT(trace,6,"nextskb"); | ||
2374 | /* qeth_hdr must not cross element boundaries */ | ||
2375 | if (element->length < offset + sizeof(struct qeth_hdr)){ | ||
2376 | if (qeth_is_last_sbale(element)) | ||
2377 | return NULL; | ||
2378 | element++; | ||
2379 | offset = 0; | ||
2380 | if (element->length < sizeof(struct qeth_hdr)) | ||
2381 | return NULL; | ||
2382 | } | ||
2383 | *hdr = element->addr + offset; | ||
2384 | |||
2385 | offset += sizeof(struct qeth_hdr); | ||
2386 | if (card->options.layer2) | ||
2387 | if (card->info.type == QETH_CARD_TYPE_OSN) | ||
2388 | skb_len = (*hdr)->hdr.osn.pdu_length; | ||
2389 | else | ||
2390 | skb_len = (*hdr)->hdr.l2.pkt_length; | ||
2391 | else | ||
2392 | skb_len = (*hdr)->hdr.l3.length; | ||
2393 | |||
2394 | if (!skb_len) | ||
2395 | return NULL; | ||
2396 | if ((skb_len >= card->options.rx_sg_cb) && | ||
2397 | (!(card->info.type == QETH_CARD_TYPE_OSN)) && | ||
2398 | (!atomic_read(&card->force_alloc_skb))) { | ||
2399 | use_rx_sg = 1; | ||
2400 | } else { | ||
2401 | if (card->options.fake_ll) { | ||
2402 | if (card->dev->type == ARPHRD_IEEE802_TR) { | ||
2403 | if (!(skb = qeth_get_skb(skb_len + | ||
2404 | QETH_FAKE_LL_LEN_TR, *hdr))) | ||
2405 | goto no_mem; | ||
2406 | skb_reserve(skb, QETH_FAKE_LL_LEN_TR); | ||
2407 | } else { | ||
2408 | if (!(skb = qeth_get_skb(skb_len + | ||
2409 | QETH_FAKE_LL_LEN_ETH, *hdr))) | ||
2410 | goto no_mem; | ||
2411 | skb_reserve(skb, QETH_FAKE_LL_LEN_ETH); | ||
2412 | } | ||
2413 | } else { | ||
2414 | skb = qeth_get_skb(skb_len, *hdr); | ||
2415 | if (!skb) | ||
2416 | goto no_mem; | ||
2417 | } | ||
2418 | } | ||
2419 | |||
2420 | data_ptr = element->addr + offset; | ||
2421 | while (skb_len) { | ||
2422 | data_len = min(skb_len, (int)(element->length - offset)); | ||
2423 | if (data_len) { | ||
2424 | if (use_rx_sg) { | ||
2425 | if (qeth_create_skb_frag(element, &skb, offset, | ||
2426 | &frag, data_len)) | ||
2427 | goto no_mem; | ||
2428 | } else { | ||
2429 | memcpy(skb_put(skb, data_len), data_ptr, | ||
2430 | data_len); | ||
2431 | } | ||
2432 | } | ||
2433 | skb_len -= data_len; | ||
2434 | if (skb_len){ | ||
2435 | if (qeth_is_last_sbale(element)){ | ||
2436 | QETH_DBF_TEXT(trace,4,"unexeob"); | ||
2437 | QETH_DBF_TEXT_(trace,4,"%s",CARD_BUS_ID(card)); | ||
2438 | QETH_DBF_TEXT(qerr,2,"unexeob"); | ||
2439 | QETH_DBF_TEXT_(qerr,2,"%s",CARD_BUS_ID(card)); | ||
2440 | QETH_DBF_HEX(misc,4,buffer,sizeof(*buffer)); | ||
2441 | dev_kfree_skb_any(skb); | ||
2442 | card->stats.rx_errors++; | ||
2443 | return NULL; | ||
2444 | } | ||
2445 | element++; | ||
2446 | offset = 0; | ||
2447 | data_ptr = element->addr; | ||
2448 | } else { | ||
2449 | offset += data_len; | ||
2450 | } | ||
2451 | } | ||
2452 | *__element = element; | ||
2453 | *__offset = offset; | ||
2454 | if (use_rx_sg && card->options.performance_stats) { | ||
2455 | card->perf_stats.sg_skbs_rx++; | ||
2456 | card->perf_stats.sg_frags_rx += skb_shinfo(skb)->nr_frags; | ||
2457 | } | ||
2458 | return skb; | ||
2459 | no_mem: | ||
2460 | if (net_ratelimit()){ | ||
2461 | PRINT_WARN("No memory for packet received on %s.\n", | ||
2462 | QETH_CARD_IFNAME(card)); | ||
2463 | QETH_DBF_TEXT(trace,2,"noskbmem"); | ||
2464 | QETH_DBF_TEXT_(trace,2,"%s",CARD_BUS_ID(card)); | ||
2465 | } | ||
2466 | card->stats.rx_dropped++; | ||
2467 | return NULL; | ||
2468 | } | ||
2469 | |||
2470 | static __be16 | ||
2471 | qeth_type_trans(struct sk_buff *skb, struct net_device *dev) | ||
2472 | { | ||
2473 | struct qeth_card *card; | ||
2474 | struct ethhdr *eth; | ||
2475 | |||
2476 | QETH_DBF_TEXT(trace,6,"typtrans"); | ||
2477 | |||
2478 | card = (struct qeth_card *)dev->priv; | ||
2479 | #ifdef CONFIG_TR | ||
2480 | if ((card->info.link_type == QETH_LINK_TYPE_HSTR) || | ||
2481 | (card->info.link_type == QETH_LINK_TYPE_LANE_TR)) | ||
2482 | return tr_type_trans(skb,dev); | ||
2483 | #endif /* CONFIG_TR */ | ||
2484 | skb_reset_mac_header(skb); | ||
2485 | skb_pull(skb, ETH_HLEN ); | ||
2486 | eth = eth_hdr(skb); | ||
2487 | |||
2488 | if (*eth->h_dest & 1) { | ||
2489 | if (memcmp(eth->h_dest, dev->broadcast, ETH_ALEN) == 0) | ||
2490 | skb->pkt_type = PACKET_BROADCAST; | ||
2491 | else | ||
2492 | skb->pkt_type = PACKET_MULTICAST; | ||
2493 | } else if (memcmp(eth->h_dest, dev->dev_addr, ETH_ALEN)) | ||
2494 | skb->pkt_type = PACKET_OTHERHOST; | ||
2495 | |||
2496 | if (ntohs(eth->h_proto) >= 1536) | ||
2497 | return eth->h_proto; | ||
2498 | if (*(unsigned short *) (skb->data) == 0xFFFF) | ||
2499 | return htons(ETH_P_802_3); | ||
2500 | return htons(ETH_P_802_2); | ||
2501 | } | ||
2502 | |||
2503 | static void | ||
2504 | qeth_rebuild_skb_fake_ll_tr(struct qeth_card *card, struct sk_buff *skb, | ||
2505 | struct qeth_hdr *hdr) | ||
2506 | { | ||
2507 | struct trh_hdr *fake_hdr; | ||
2508 | struct trllc *fake_llc; | ||
2509 | struct iphdr *ip_hdr; | ||
2510 | |||
2511 | QETH_DBF_TEXT(trace,5,"skbfktr"); | ||
2512 | skb_set_mac_header(skb, (int)-QETH_FAKE_LL_LEN_TR); | ||
2513 | /* this is a fake ethernet header */ | ||
2514 | fake_hdr = tr_hdr(skb); | ||
2515 | |||
2516 | /* the destination MAC address */ | ||
2517 | switch (skb->pkt_type){ | ||
2518 | case PACKET_MULTICAST: | ||
2519 | switch (skb->protocol){ | ||
2520 | #ifdef CONFIG_QETH_IPV6 | ||
2521 | case __constant_htons(ETH_P_IPV6): | ||
2522 | ndisc_mc_map((struct in6_addr *) | ||
2523 | skb->data + QETH_FAKE_LL_V6_ADDR_POS, | ||
2524 | fake_hdr->daddr, card->dev, 0); | ||
2525 | break; | ||
2526 | #endif /* CONFIG_QETH_IPV6 */ | ||
2527 | case __constant_htons(ETH_P_IP): | ||
2528 | ip_hdr = (struct iphdr *)skb->data; | ||
2529 | ip_tr_mc_map(ip_hdr->daddr, fake_hdr->daddr); | ||
2530 | break; | ||
2531 | default: | ||
2532 | memcpy(fake_hdr->daddr, card->dev->dev_addr, TR_ALEN); | ||
2533 | } | ||
2534 | break; | ||
2535 | case PACKET_BROADCAST: | ||
2536 | memset(fake_hdr->daddr, 0xff, TR_ALEN); | ||
2537 | break; | ||
2538 | default: | ||
2539 | memcpy(fake_hdr->daddr, card->dev->dev_addr, TR_ALEN); | ||
2540 | } | ||
2541 | /* the source MAC address */ | ||
2542 | if (hdr->hdr.l3.ext_flags & QETH_HDR_EXT_SRC_MAC_ADDR) | ||
2543 | memcpy(fake_hdr->saddr, &hdr->hdr.l3.dest_addr[2], TR_ALEN); | ||
2544 | else | ||
2545 | memset(fake_hdr->saddr, 0, TR_ALEN); | ||
2546 | fake_hdr->rcf=0; | ||
2547 | fake_llc = (struct trllc*)&(fake_hdr->rcf); | ||
2548 | fake_llc->dsap = EXTENDED_SAP; | ||
2549 | fake_llc->ssap = EXTENDED_SAP; | ||
2550 | fake_llc->llc = UI_CMD; | ||
2551 | fake_llc->protid[0] = 0; | ||
2552 | fake_llc->protid[1] = 0; | ||
2553 | fake_llc->protid[2] = 0; | ||
2554 | fake_llc->ethertype = ETH_P_IP; | ||
2555 | } | ||
2556 | |||
2557 | static void | ||
2558 | qeth_rebuild_skb_fake_ll_eth(struct qeth_card *card, struct sk_buff *skb, | ||
2559 | struct qeth_hdr *hdr) | ||
2560 | { | ||
2561 | struct ethhdr *fake_hdr; | ||
2562 | struct iphdr *ip_hdr; | ||
2563 | |||
2564 | QETH_DBF_TEXT(trace,5,"skbfketh"); | ||
2565 | skb_set_mac_header(skb, -QETH_FAKE_LL_LEN_ETH); | ||
2566 | /* this is a fake ethernet header */ | ||
2567 | fake_hdr = eth_hdr(skb); | ||
2568 | |||
2569 | /* the destination MAC address */ | ||
2570 | switch (skb->pkt_type){ | ||
2571 | case PACKET_MULTICAST: | ||
2572 | switch (skb->protocol){ | ||
2573 | #ifdef CONFIG_QETH_IPV6 | ||
2574 | case __constant_htons(ETH_P_IPV6): | ||
2575 | ndisc_mc_map((struct in6_addr *) | ||
2576 | skb->data + QETH_FAKE_LL_V6_ADDR_POS, | ||
2577 | fake_hdr->h_dest, card->dev, 0); | ||
2578 | break; | ||
2579 | #endif /* CONFIG_QETH_IPV6 */ | ||
2580 | case __constant_htons(ETH_P_IP): | ||
2581 | ip_hdr = (struct iphdr *)skb->data; | ||
2582 | ip_eth_mc_map(ip_hdr->daddr, fake_hdr->h_dest); | ||
2583 | break; | ||
2584 | default: | ||
2585 | memcpy(fake_hdr->h_dest, card->dev->dev_addr, ETH_ALEN); | ||
2586 | } | ||
2587 | break; | ||
2588 | case PACKET_BROADCAST: | ||
2589 | memset(fake_hdr->h_dest, 0xff, ETH_ALEN); | ||
2590 | break; | ||
2591 | default: | ||
2592 | memcpy(fake_hdr->h_dest, card->dev->dev_addr, ETH_ALEN); | ||
2593 | } | ||
2594 | /* the source MAC address */ | ||
2595 | if (hdr->hdr.l3.ext_flags & QETH_HDR_EXT_SRC_MAC_ADDR) | ||
2596 | memcpy(fake_hdr->h_source, &hdr->hdr.l3.dest_addr[2], ETH_ALEN); | ||
2597 | else | ||
2598 | memset(fake_hdr->h_source, 0, ETH_ALEN); | ||
2599 | /* the protocol */ | ||
2600 | fake_hdr->h_proto = skb->protocol; | ||
2601 | } | ||
2602 | |||
2603 | static inline void | ||
2604 | qeth_rebuild_skb_fake_ll(struct qeth_card *card, struct sk_buff *skb, | ||
2605 | struct qeth_hdr *hdr) | ||
2606 | { | ||
2607 | if (card->dev->type == ARPHRD_IEEE802_TR) | ||
2608 | qeth_rebuild_skb_fake_ll_tr(card, skb, hdr); | ||
2609 | else | ||
2610 | qeth_rebuild_skb_fake_ll_eth(card, skb, hdr); | ||
2611 | } | ||
2612 | |||
2613 | static inline void | ||
2614 | qeth_layer2_rebuild_skb(struct qeth_card *card, struct sk_buff *skb, | ||
2615 | struct qeth_hdr *hdr) | ||
2616 | { | ||
2617 | skb->pkt_type = PACKET_HOST; | ||
2618 | skb->protocol = qeth_type_trans(skb, skb->dev); | ||
2619 | if (card->options.checksum_type == NO_CHECKSUMMING) | ||
2620 | skb->ip_summed = CHECKSUM_UNNECESSARY; | ||
2621 | else | ||
2622 | skb->ip_summed = CHECKSUM_NONE; | ||
2623 | *((__u32 *)skb->cb) = ++card->seqno.pkt_seqno; | ||
2624 | } | ||
2625 | |||
2626 | static __u16 | ||
2627 | qeth_rebuild_skb(struct qeth_card *card, struct sk_buff *skb, | ||
2628 | struct qeth_hdr *hdr) | ||
2629 | { | ||
2630 | unsigned short vlan_id = 0; | ||
2631 | #ifdef CONFIG_QETH_IPV6 | ||
2632 | if (hdr->hdr.l3.flags & QETH_HDR_PASSTHRU) { | ||
2633 | skb->pkt_type = PACKET_HOST; | ||
2634 | skb->protocol = qeth_type_trans(skb, card->dev); | ||
2635 | return 0; | ||
2636 | } | ||
2637 | #endif /* CONFIG_QETH_IPV6 */ | ||
2638 | skb->protocol = htons((hdr->hdr.l3.flags & QETH_HDR_IPV6)? ETH_P_IPV6 : | ||
2639 | ETH_P_IP); | ||
2640 | switch (hdr->hdr.l3.flags & QETH_HDR_CAST_MASK){ | ||
2641 | case QETH_CAST_UNICAST: | ||
2642 | skb->pkt_type = PACKET_HOST; | ||
2643 | break; | ||
2644 | case QETH_CAST_MULTICAST: | ||
2645 | skb->pkt_type = PACKET_MULTICAST; | ||
2646 | card->stats.multicast++; | ||
2647 | break; | ||
2648 | case QETH_CAST_BROADCAST: | ||
2649 | skb->pkt_type = PACKET_BROADCAST; | ||
2650 | card->stats.multicast++; | ||
2651 | break; | ||
2652 | case QETH_CAST_ANYCAST: | ||
2653 | case QETH_CAST_NOCAST: | ||
2654 | default: | ||
2655 | skb->pkt_type = PACKET_HOST; | ||
2656 | } | ||
2657 | |||
2658 | if (hdr->hdr.l3.ext_flags & | ||
2659 | (QETH_HDR_EXT_VLAN_FRAME | QETH_HDR_EXT_INCLUDE_VLAN_TAG)) { | ||
2660 | vlan_id = (hdr->hdr.l3.ext_flags & QETH_HDR_EXT_VLAN_FRAME)? | ||
2661 | hdr->hdr.l3.vlan_id : *((u16 *)&hdr->hdr.l3.dest_addr[12]); | ||
2662 | } | ||
2663 | |||
2664 | if (card->options.fake_ll) | ||
2665 | qeth_rebuild_skb_fake_ll(card, skb, hdr); | ||
2666 | else | ||
2667 | skb_reset_mac_header(skb); | ||
2668 | skb->ip_summed = card->options.checksum_type; | ||
2669 | if (card->options.checksum_type == HW_CHECKSUMMING){ | ||
2670 | if ( (hdr->hdr.l3.ext_flags & | ||
2671 | (QETH_HDR_EXT_CSUM_HDR_REQ | | ||
2672 | QETH_HDR_EXT_CSUM_TRANSP_REQ)) == | ||
2673 | (QETH_HDR_EXT_CSUM_HDR_REQ | | ||
2674 | QETH_HDR_EXT_CSUM_TRANSP_REQ) ) | ||
2675 | skb->ip_summed = CHECKSUM_UNNECESSARY; | ||
2676 | else | ||
2677 | skb->ip_summed = SW_CHECKSUMMING; | ||
2678 | } | ||
2679 | return vlan_id; | ||
2680 | } | ||
2681 | |||
2682 | static void | ||
2683 | qeth_process_inbound_buffer(struct qeth_card *card, | ||
2684 | struct qeth_qdio_buffer *buf, int index) | ||
2685 | { | ||
2686 | struct qdio_buffer_element *element; | ||
2687 | struct sk_buff *skb; | ||
2688 | struct qeth_hdr *hdr; | ||
2689 | int offset; | ||
2690 | int rxrc; | ||
2691 | __u16 vlan_tag = 0; | ||
2692 | |||
2693 | /* get first element of current buffer */ | ||
2694 | element = (struct qdio_buffer_element *)&buf->buffer->element[0]; | ||
2695 | offset = 0; | ||
2696 | if (card->options.performance_stats) | ||
2697 | card->perf_stats.bufs_rec++; | ||
2698 | while((skb = qeth_get_next_skb(card, buf->buffer, &element, | ||
2699 | &offset, &hdr))) { | ||
2700 | skb->dev = card->dev; | ||
2701 | if (hdr->hdr.l2.id == QETH_HEADER_TYPE_LAYER2) | ||
2702 | qeth_layer2_rebuild_skb(card, skb, hdr); | ||
2703 | else if (hdr->hdr.l3.id == QETH_HEADER_TYPE_LAYER3) | ||
2704 | vlan_tag = qeth_rebuild_skb(card, skb, hdr); | ||
2705 | else if (hdr->hdr.osn.id == QETH_HEADER_TYPE_OSN) { | ||
2706 | skb_push(skb, sizeof(struct qeth_hdr)); | ||
2707 | skb_copy_to_linear_data(skb, hdr, | ||
2708 | sizeof(struct qeth_hdr)); | ||
2709 | } else { /* unknown header type */ | ||
2710 | dev_kfree_skb_any(skb); | ||
2711 | QETH_DBF_TEXT(trace, 3, "inbunkno"); | ||
2712 | QETH_DBF_HEX(control, 3, hdr, QETH_DBF_CONTROL_LEN); | ||
2713 | continue; | ||
2714 | } | ||
2715 | /* is device UP ? */ | ||
2716 | if (!(card->dev->flags & IFF_UP)){ | ||
2717 | dev_kfree_skb_any(skb); | ||
2718 | continue; | ||
2719 | } | ||
2720 | if (card->info.type == QETH_CARD_TYPE_OSN) | ||
2721 | rxrc = card->osn_info.data_cb(skb); | ||
2722 | else | ||
2723 | #ifdef CONFIG_QETH_VLAN | ||
2724 | if (vlan_tag) | ||
2725 | if (card->vlangrp) | ||
2726 | vlan_hwaccel_rx(skb, card->vlangrp, vlan_tag); | ||
2727 | else { | ||
2728 | dev_kfree_skb_any(skb); | ||
2729 | continue; | ||
2730 | } | ||
2731 | else | ||
2732 | #endif | ||
2733 | rxrc = netif_rx(skb); | ||
2734 | card->dev->last_rx = jiffies; | ||
2735 | card->stats.rx_packets++; | ||
2736 | card->stats.rx_bytes += skb->len; | ||
2737 | } | ||
2738 | } | ||
2739 | |||
2740 | static int | ||
2741 | qeth_init_input_buffer(struct qeth_card *card, struct qeth_qdio_buffer *buf) | ||
2742 | { | ||
2743 | struct qeth_buffer_pool_entry *pool_entry; | ||
2744 | int i; | ||
2745 | |||
2746 | pool_entry = qeth_find_free_buffer_pool_entry(card); | ||
2747 | if (!pool_entry) | ||
2748 | return 1; | ||
2749 | /* | ||
2750 | * since the buffer is accessed only from the input_tasklet | ||
2751 | * there shouldn't be a need to synchronize; also, since we use | ||
2752 | * the QETH_IN_BUF_REQUEUE_THRESHOLD we should never run out off | ||
2753 | * buffers | ||
2754 | */ | ||
2755 | BUG_ON(!pool_entry); | ||
2756 | |||
2757 | buf->pool_entry = pool_entry; | ||
2758 | for(i = 0; i < QETH_MAX_BUFFER_ELEMENTS(card); ++i){ | ||
2759 | buf->buffer->element[i].length = PAGE_SIZE; | ||
2760 | buf->buffer->element[i].addr = pool_entry->elements[i]; | ||
2761 | if (i == QETH_MAX_BUFFER_ELEMENTS(card) - 1) | ||
2762 | buf->buffer->element[i].flags = SBAL_FLAGS_LAST_ENTRY; | ||
2763 | else | ||
2764 | buf->buffer->element[i].flags = 0; | ||
2765 | } | ||
2766 | buf->state = QETH_QDIO_BUF_EMPTY; | ||
2767 | return 0; | ||
2768 | } | ||
2769 | |||
2770 | static void | ||
2771 | qeth_clear_output_buffer(struct qeth_qdio_out_q *queue, | ||
2772 | struct qeth_qdio_out_buffer *buf) | ||
2773 | { | ||
2774 | int i; | ||
2775 | struct sk_buff *skb; | ||
2776 | |||
2777 | /* is PCI flag set on buffer? */ | ||
2778 | if (buf->buffer->element[0].flags & 0x40) | ||
2779 | atomic_dec(&queue->set_pci_flags_count); | ||
2780 | |||
2781 | while ((skb = skb_dequeue(&buf->skb_list))){ | ||
2782 | atomic_dec(&skb->users); | ||
2783 | dev_kfree_skb_any(skb); | ||
2784 | } | ||
2785 | qeth_eddp_buf_release_contexts(buf); | ||
2786 | for(i = 0; i < QETH_MAX_BUFFER_ELEMENTS(queue->card); ++i){ | ||
2787 | buf->buffer->element[i].length = 0; | ||
2788 | buf->buffer->element[i].addr = NULL; | ||
2789 | buf->buffer->element[i].flags = 0; | ||
2790 | } | ||
2791 | buf->next_element_to_fill = 0; | ||
2792 | atomic_set(&buf->state, QETH_QDIO_BUF_EMPTY); | ||
2793 | } | ||
2794 | |||
2795 | static void | ||
2796 | qeth_queue_input_buffer(struct qeth_card *card, int index) | ||
2797 | { | ||
2798 | struct qeth_qdio_q *queue = card->qdio.in_q; | ||
2799 | int count; | ||
2800 | int i; | ||
2801 | int rc; | ||
2802 | int newcount = 0; | ||
2803 | |||
2804 | QETH_DBF_TEXT(trace,6,"queinbuf"); | ||
2805 | count = (index < queue->next_buf_to_init)? | ||
2806 | card->qdio.in_buf_pool.buf_count - | ||
2807 | (queue->next_buf_to_init - index) : | ||
2808 | card->qdio.in_buf_pool.buf_count - | ||
2809 | (queue->next_buf_to_init + QDIO_MAX_BUFFERS_PER_Q - index); | ||
2810 | /* only requeue at a certain threshold to avoid SIGAs */ | ||
2811 | if (count >= QETH_IN_BUF_REQUEUE_THRESHOLD(card)){ | ||
2812 | for (i = queue->next_buf_to_init; | ||
2813 | i < queue->next_buf_to_init + count; ++i) { | ||
2814 | if (qeth_init_input_buffer(card, | ||
2815 | &queue->bufs[i % QDIO_MAX_BUFFERS_PER_Q])) { | ||
2816 | break; | ||
2817 | } else { | ||
2818 | newcount++; | ||
2819 | } | ||
2820 | } | ||
2821 | |||
2822 | if (newcount < count) { | ||
2823 | /* we are in memory shortage so we switch back to | ||
2824 | traditional skb allocation and drop packages */ | ||
2825 | if (!atomic_read(&card->force_alloc_skb) && | ||
2826 | net_ratelimit()) | ||
2827 | PRINT_WARN("Switch to alloc skb\n"); | ||
2828 | atomic_set(&card->force_alloc_skb, 3); | ||
2829 | count = newcount; | ||
2830 | } else { | ||
2831 | if ((atomic_read(&card->force_alloc_skb) == 1) && | ||
2832 | net_ratelimit()) | ||
2833 | PRINT_WARN("Switch to sg\n"); | ||
2834 | atomic_add_unless(&card->force_alloc_skb, -1, 0); | ||
2835 | } | ||
2836 | |||
2837 | /* | ||
2838 | * according to old code it should be avoided to requeue all | ||
2839 | * 128 buffers in order to benefit from PCI avoidance. | ||
2840 | * this function keeps at least one buffer (the buffer at | ||
2841 | * 'index') un-requeued -> this buffer is the first buffer that | ||
2842 | * will be requeued the next time | ||
2843 | */ | ||
2844 | if (card->options.performance_stats) { | ||
2845 | card->perf_stats.inbound_do_qdio_cnt++; | ||
2846 | card->perf_stats.inbound_do_qdio_start_time = | ||
2847 | qeth_get_micros(); | ||
2848 | } | ||
2849 | rc = do_QDIO(CARD_DDEV(card), | ||
2850 | QDIO_FLAG_SYNC_INPUT | QDIO_FLAG_UNDER_INTERRUPT, | ||
2851 | 0, queue->next_buf_to_init, count, NULL); | ||
2852 | if (card->options.performance_stats) | ||
2853 | card->perf_stats.inbound_do_qdio_time += | ||
2854 | qeth_get_micros() - | ||
2855 | card->perf_stats.inbound_do_qdio_start_time; | ||
2856 | if (rc){ | ||
2857 | PRINT_WARN("qeth_queue_input_buffer's do_QDIO " | ||
2858 | "return %i (device %s).\n", | ||
2859 | rc, CARD_DDEV_ID(card)); | ||
2860 | QETH_DBF_TEXT(trace,2,"qinberr"); | ||
2861 | QETH_DBF_TEXT_(trace,2,"%s",CARD_BUS_ID(card)); | ||
2862 | } | ||
2863 | queue->next_buf_to_init = (queue->next_buf_to_init + count) % | ||
2864 | QDIO_MAX_BUFFERS_PER_Q; | ||
2865 | } | ||
2866 | } | ||
2867 | |||
2868 | static inline void | ||
2869 | qeth_put_buffer_pool_entry(struct qeth_card *card, | ||
2870 | struct qeth_buffer_pool_entry *entry) | ||
2871 | { | ||
2872 | QETH_DBF_TEXT(trace, 6, "ptbfplen"); | ||
2873 | list_add_tail(&entry->list, &card->qdio.in_buf_pool.entry_list); | ||
2874 | } | ||
2875 | |||
2876 | static void | ||
2877 | qeth_qdio_input_handler(struct ccw_device * ccwdev, unsigned int status, | ||
2878 | unsigned int qdio_err, unsigned int siga_err, | ||
2879 | unsigned int queue, int first_element, int count, | ||
2880 | unsigned long card_ptr) | ||
2881 | { | ||
2882 | struct net_device *net_dev; | ||
2883 | struct qeth_card *card; | ||
2884 | struct qeth_qdio_buffer *buffer; | ||
2885 | int index; | ||
2886 | int i; | ||
2887 | |||
2888 | QETH_DBF_TEXT(trace, 6, "qdinput"); | ||
2889 | card = (struct qeth_card *) card_ptr; | ||
2890 | net_dev = card->dev; | ||
2891 | if (card->options.performance_stats) { | ||
2892 | card->perf_stats.inbound_cnt++; | ||
2893 | card->perf_stats.inbound_start_time = qeth_get_micros(); | ||
2894 | } | ||
2895 | if (status & QDIO_STATUS_LOOK_FOR_ERROR) { | ||
2896 | if (status & QDIO_STATUS_ACTIVATE_CHECK_CONDITION){ | ||
2897 | QETH_DBF_TEXT(trace, 1,"qdinchk"); | ||
2898 | QETH_DBF_TEXT_(trace,1,"%s",CARD_BUS_ID(card)); | ||
2899 | QETH_DBF_TEXT_(trace,1,"%04X%04X",first_element,count); | ||
2900 | QETH_DBF_TEXT_(trace,1,"%04X%04X", queue, status); | ||
2901 | qeth_schedule_recovery(card); | ||
2902 | return; | ||
2903 | } | ||
2904 | } | ||
2905 | for (i = first_element; i < (first_element + count); ++i) { | ||
2906 | index = i % QDIO_MAX_BUFFERS_PER_Q; | ||
2907 | buffer = &card->qdio.in_q->bufs[index]; | ||
2908 | if (!((status & QDIO_STATUS_LOOK_FOR_ERROR) && | ||
2909 | qeth_check_qdio_errors(buffer->buffer, | ||
2910 | qdio_err, siga_err,"qinerr"))) | ||
2911 | qeth_process_inbound_buffer(card, buffer, index); | ||
2912 | /* clear buffer and give back to hardware */ | ||
2913 | qeth_put_buffer_pool_entry(card, buffer->pool_entry); | ||
2914 | qeth_queue_input_buffer(card, index); | ||
2915 | } | ||
2916 | if (card->options.performance_stats) | ||
2917 | card->perf_stats.inbound_time += qeth_get_micros() - | ||
2918 | card->perf_stats.inbound_start_time; | ||
2919 | } | ||
2920 | |||
2921 | static int | ||
2922 | qeth_handle_send_error(struct qeth_card *card, | ||
2923 | struct qeth_qdio_out_buffer *buffer, | ||
2924 | unsigned int qdio_err, unsigned int siga_err) | ||
2925 | { | ||
2926 | int sbalf15 = buffer->buffer->element[15].flags & 0xff; | ||
2927 | int cc = siga_err & 3; | ||
2928 | |||
2929 | QETH_DBF_TEXT(trace, 6, "hdsnderr"); | ||
2930 | qeth_check_qdio_errors(buffer->buffer, qdio_err, siga_err, "qouterr"); | ||
2931 | switch (cc) { | ||
2932 | case 0: | ||
2933 | if (qdio_err){ | ||
2934 | QETH_DBF_TEXT(trace, 1,"lnkfail"); | ||
2935 | QETH_DBF_TEXT_(trace,1,"%s",CARD_BUS_ID(card)); | ||
2936 | QETH_DBF_TEXT_(trace,1,"%04x %02x", | ||
2937 | (u16)qdio_err, (u8)sbalf15); | ||
2938 | return QETH_SEND_ERROR_LINK_FAILURE; | ||
2939 | } | ||
2940 | return QETH_SEND_ERROR_NONE; | ||
2941 | case 2: | ||
2942 | if (siga_err & QDIO_SIGA_ERROR_B_BIT_SET) { | ||
2943 | QETH_DBF_TEXT(trace, 1, "SIGAcc2B"); | ||
2944 | QETH_DBF_TEXT_(trace,1,"%s",CARD_BUS_ID(card)); | ||
2945 | return QETH_SEND_ERROR_KICK_IT; | ||
2946 | } | ||
2947 | if ((sbalf15 >= 15) && (sbalf15 <= 31)) | ||
2948 | return QETH_SEND_ERROR_RETRY; | ||
2949 | return QETH_SEND_ERROR_LINK_FAILURE; | ||
2950 | /* look at qdio_error and sbalf 15 */ | ||
2951 | case 1: | ||
2952 | QETH_DBF_TEXT(trace, 1, "SIGAcc1"); | ||
2953 | QETH_DBF_TEXT_(trace,1,"%s",CARD_BUS_ID(card)); | ||
2954 | return QETH_SEND_ERROR_LINK_FAILURE; | ||
2955 | case 3: | ||
2956 | default: | ||
2957 | QETH_DBF_TEXT(trace, 1, "SIGAcc3"); | ||
2958 | QETH_DBF_TEXT_(trace,1,"%s",CARD_BUS_ID(card)); | ||
2959 | return QETH_SEND_ERROR_KICK_IT; | ||
2960 | } | ||
2961 | } | ||
2962 | |||
2963 | void | ||
2964 | qeth_flush_buffers(struct qeth_qdio_out_q *queue, int under_int, | ||
2965 | int index, int count) | ||
2966 | { | ||
2967 | struct qeth_qdio_out_buffer *buf; | ||
2968 | int rc; | ||
2969 | int i; | ||
2970 | unsigned int qdio_flags; | ||
2971 | |||
2972 | QETH_DBF_TEXT(trace, 6, "flushbuf"); | ||
2973 | |||
2974 | for (i = index; i < index + count; ++i) { | ||
2975 | buf = &queue->bufs[i % QDIO_MAX_BUFFERS_PER_Q]; | ||
2976 | buf->buffer->element[buf->next_element_to_fill - 1].flags |= | ||
2977 | SBAL_FLAGS_LAST_ENTRY; | ||
2978 | |||
2979 | if (queue->card->info.type == QETH_CARD_TYPE_IQD) | ||
2980 | continue; | ||
2981 | |||
2982 | if (!queue->do_pack){ | ||
2983 | if ((atomic_read(&queue->used_buffers) >= | ||
2984 | (QETH_HIGH_WATERMARK_PACK - | ||
2985 | QETH_WATERMARK_PACK_FUZZ)) && | ||
2986 | !atomic_read(&queue->set_pci_flags_count)){ | ||
2987 | /* it's likely that we'll go to packing | ||
2988 | * mode soon */ | ||
2989 | atomic_inc(&queue->set_pci_flags_count); | ||
2990 | buf->buffer->element[0].flags |= 0x40; | ||
2991 | } | ||
2992 | } else { | ||
2993 | if (!atomic_read(&queue->set_pci_flags_count)){ | ||
2994 | /* | ||
2995 | * there's no outstanding PCI any more, so we | ||
2996 | * have to request a PCI to be sure that the PCI | ||
2997 | * will wake at some time in the future then we | ||
2998 | * can flush packed buffers that might still be | ||
2999 | * hanging around, which can happen if no | ||
3000 | * further send was requested by the stack | ||
3001 | */ | ||
3002 | atomic_inc(&queue->set_pci_flags_count); | ||
3003 | buf->buffer->element[0].flags |= 0x40; | ||
3004 | } | ||
3005 | } | ||
3006 | } | ||
3007 | |||
3008 | queue->card->dev->trans_start = jiffies; | ||
3009 | if (queue->card->options.performance_stats) { | ||
3010 | queue->card->perf_stats.outbound_do_qdio_cnt++; | ||
3011 | queue->card->perf_stats.outbound_do_qdio_start_time = | ||
3012 | qeth_get_micros(); | ||
3013 | } | ||
3014 | qdio_flags = QDIO_FLAG_SYNC_OUTPUT; | ||
3015 | if (under_int) | ||
3016 | qdio_flags |= QDIO_FLAG_UNDER_INTERRUPT; | ||
3017 | if (atomic_read(&queue->set_pci_flags_count)) | ||
3018 | qdio_flags |= QDIO_FLAG_PCI_OUT; | ||
3019 | rc = do_QDIO(CARD_DDEV(queue->card), qdio_flags, | ||
3020 | queue->queue_no, index, count, NULL); | ||
3021 | if (queue->card->options.performance_stats) | ||
3022 | queue->card->perf_stats.outbound_do_qdio_time += | ||
3023 | qeth_get_micros() - | ||
3024 | queue->card->perf_stats.outbound_do_qdio_start_time; | ||
3025 | if (rc){ | ||
3026 | QETH_DBF_TEXT(trace, 2, "flushbuf"); | ||
3027 | QETH_DBF_TEXT_(trace, 2, " err%d", rc); | ||
3028 | QETH_DBF_TEXT_(trace, 2, "%s", CARD_DDEV_ID(queue->card)); | ||
3029 | queue->card->stats.tx_errors += count; | ||
3030 | /* this must not happen under normal circumstances. if it | ||
3031 | * happens something is really wrong -> recover */ | ||
3032 | qeth_schedule_recovery(queue->card); | ||
3033 | return; | ||
3034 | } | ||
3035 | atomic_add(count, &queue->used_buffers); | ||
3036 | if (queue->card->options.performance_stats) | ||
3037 | queue->card->perf_stats.bufs_sent += count; | ||
3038 | } | ||
3039 | |||
3040 | /* | ||
3041 | * Switched to packing state if the number of used buffers on a queue | ||
3042 | * reaches a certain limit. | ||
3043 | */ | ||
3044 | static void | ||
3045 | qeth_switch_to_packing_if_needed(struct qeth_qdio_out_q *queue) | ||
3046 | { | ||
3047 | if (!queue->do_pack) { | ||
3048 | if (atomic_read(&queue->used_buffers) | ||
3049 | >= QETH_HIGH_WATERMARK_PACK){ | ||
3050 | /* switch non-PACKING -> PACKING */ | ||
3051 | QETH_DBF_TEXT(trace, 6, "np->pack"); | ||
3052 | if (queue->card->options.performance_stats) | ||
3053 | queue->card->perf_stats.sc_dp_p++; | ||
3054 | queue->do_pack = 1; | ||
3055 | } | ||
3056 | } | ||
3057 | } | ||
3058 | |||
3059 | /* | ||
3060 | * Switches from packing to non-packing mode. If there is a packing | ||
3061 | * buffer on the queue this buffer will be prepared to be flushed. | ||
3062 | * In that case 1 is returned to inform the caller. If no buffer | ||
3063 | * has to be flushed, zero is returned. | ||
3064 | */ | ||
3065 | static int | ||
3066 | qeth_switch_to_nonpacking_if_needed(struct qeth_qdio_out_q *queue) | ||
3067 | { | ||
3068 | struct qeth_qdio_out_buffer *buffer; | ||
3069 | int flush_count = 0; | ||
3070 | |||
3071 | if (queue->do_pack) { | ||
3072 | if (atomic_read(&queue->used_buffers) | ||
3073 | <= QETH_LOW_WATERMARK_PACK) { | ||
3074 | /* switch PACKING -> non-PACKING */ | ||
3075 | QETH_DBF_TEXT(trace, 6, "pack->np"); | ||
3076 | if (queue->card->options.performance_stats) | ||
3077 | queue->card->perf_stats.sc_p_dp++; | ||
3078 | queue->do_pack = 0; | ||
3079 | /* flush packing buffers */ | ||
3080 | buffer = &queue->bufs[queue->next_buf_to_fill]; | ||
3081 | if ((atomic_read(&buffer->state) == | ||
3082 | QETH_QDIO_BUF_EMPTY) && | ||
3083 | (buffer->next_element_to_fill > 0)) { | ||
3084 | atomic_set(&buffer->state,QETH_QDIO_BUF_PRIMED); | ||
3085 | flush_count++; | ||
3086 | queue->next_buf_to_fill = | ||
3087 | (queue->next_buf_to_fill + 1) % | ||
3088 | QDIO_MAX_BUFFERS_PER_Q; | ||
3089 | } | ||
3090 | } | ||
3091 | } | ||
3092 | return flush_count; | ||
3093 | } | ||
3094 | |||
3095 | /* | ||
3096 | * Called to flush a packing buffer if no more pci flags are on the queue. | ||
3097 | * Checks if there is a packing buffer and prepares it to be flushed. | ||
3098 | * In that case returns 1, otherwise zero. | ||
3099 | */ | ||
3100 | static int | ||
3101 | qeth_flush_buffers_on_no_pci(struct qeth_qdio_out_q *queue) | ||
3102 | { | ||
3103 | struct qeth_qdio_out_buffer *buffer; | ||
3104 | |||
3105 | buffer = &queue->bufs[queue->next_buf_to_fill]; | ||
3106 | if((atomic_read(&buffer->state) == QETH_QDIO_BUF_EMPTY) && | ||
3107 | (buffer->next_element_to_fill > 0)){ | ||
3108 | /* it's a packing buffer */ | ||
3109 | atomic_set(&buffer->state, QETH_QDIO_BUF_PRIMED); | ||
3110 | queue->next_buf_to_fill = | ||
3111 | (queue->next_buf_to_fill + 1) % QDIO_MAX_BUFFERS_PER_Q; | ||
3112 | return 1; | ||
3113 | } | ||
3114 | return 0; | ||
3115 | } | ||
3116 | |||
3117 | static void | ||
3118 | qeth_check_outbound_queue(struct qeth_qdio_out_q *queue) | ||
3119 | { | ||
3120 | int index; | ||
3121 | int flush_cnt = 0; | ||
3122 | int q_was_packing = 0; | ||
3123 | |||
3124 | /* | ||
3125 | * check if weed have to switch to non-packing mode or if | ||
3126 | * we have to get a pci flag out on the queue | ||
3127 | */ | ||
3128 | if ((atomic_read(&queue->used_buffers) <= QETH_LOW_WATERMARK_PACK) || | ||
3129 | !atomic_read(&queue->set_pci_flags_count)){ | ||
3130 | if (atomic_xchg(&queue->state, QETH_OUT_Q_LOCKED_FLUSH) == | ||
3131 | QETH_OUT_Q_UNLOCKED) { | ||
3132 | /* | ||
3133 | * If we get in here, there was no action in | ||
3134 | * do_send_packet. So, we check if there is a | ||
3135 | * packing buffer to be flushed here. | ||
3136 | */ | ||
3137 | netif_stop_queue(queue->card->dev); | ||
3138 | index = queue->next_buf_to_fill; | ||
3139 | q_was_packing = queue->do_pack; | ||
3140 | flush_cnt += qeth_switch_to_nonpacking_if_needed(queue); | ||
3141 | if (!flush_cnt && | ||
3142 | !atomic_read(&queue->set_pci_flags_count)) | ||
3143 | flush_cnt += | ||
3144 | qeth_flush_buffers_on_no_pci(queue); | ||
3145 | if (queue->card->options.performance_stats && | ||
3146 | q_was_packing) | ||
3147 | queue->card->perf_stats.bufs_sent_pack += | ||
3148 | flush_cnt; | ||
3149 | if (flush_cnt) | ||
3150 | qeth_flush_buffers(queue, 1, index, flush_cnt); | ||
3151 | atomic_set(&queue->state, QETH_OUT_Q_UNLOCKED); | ||
3152 | } | ||
3153 | } | ||
3154 | } | ||
3155 | |||
3156 | static void | ||
3157 | qeth_qdio_output_handler(struct ccw_device * ccwdev, unsigned int status, | ||
3158 | unsigned int qdio_error, unsigned int siga_error, | ||
3159 | unsigned int __queue, int first_element, int count, | ||
3160 | unsigned long card_ptr) | ||
3161 | { | ||
3162 | struct qeth_card *card = (struct qeth_card *) card_ptr; | ||
3163 | struct qeth_qdio_out_q *queue = card->qdio.out_qs[__queue]; | ||
3164 | struct qeth_qdio_out_buffer *buffer; | ||
3165 | int i; | ||
3166 | |||
3167 | QETH_DBF_TEXT(trace, 6, "qdouhdl"); | ||
3168 | if (status & QDIO_STATUS_LOOK_FOR_ERROR) { | ||
3169 | if (status & QDIO_STATUS_ACTIVATE_CHECK_CONDITION){ | ||
3170 | QETH_DBF_TEXT(trace, 2, "achkcond"); | ||
3171 | QETH_DBF_TEXT_(trace, 2, "%s", CARD_BUS_ID(card)); | ||
3172 | QETH_DBF_TEXT_(trace, 2, "%08x", status); | ||
3173 | netif_stop_queue(card->dev); | ||
3174 | qeth_schedule_recovery(card); | ||
3175 | return; | ||
3176 | } | ||
3177 | } | ||
3178 | if (card->options.performance_stats) { | ||
3179 | card->perf_stats.outbound_handler_cnt++; | ||
3180 | card->perf_stats.outbound_handler_start_time = | ||
3181 | qeth_get_micros(); | ||
3182 | } | ||
3183 | for(i = first_element; i < (first_element + count); ++i){ | ||
3184 | buffer = &queue->bufs[i % QDIO_MAX_BUFFERS_PER_Q]; | ||
3185 | /*we only handle the KICK_IT error by doing a recovery */ | ||
3186 | if (qeth_handle_send_error(card, buffer, | ||
3187 | qdio_error, siga_error) | ||
3188 | == QETH_SEND_ERROR_KICK_IT){ | ||
3189 | netif_stop_queue(card->dev); | ||
3190 | qeth_schedule_recovery(card); | ||
3191 | return; | ||
3192 | } | ||
3193 | qeth_clear_output_buffer(queue, buffer); | ||
3194 | } | ||
3195 | atomic_sub(count, &queue->used_buffers); | ||
3196 | /* check if we need to do something on this outbound queue */ | ||
3197 | if (card->info.type != QETH_CARD_TYPE_IQD) | ||
3198 | qeth_check_outbound_queue(queue); | ||
3199 | |||
3200 | netif_wake_queue(queue->card->dev); | ||
3201 | if (card->options.performance_stats) | ||
3202 | card->perf_stats.outbound_handler_time += qeth_get_micros() - | ||
3203 | card->perf_stats.outbound_handler_start_time; | ||
3204 | } | ||
3205 | |||
3206 | static void | ||
3207 | qeth_create_qib_param_field(struct qeth_card *card, char *param_field) | ||
3208 | { | ||
3209 | |||
3210 | param_field[0] = _ascebc['P']; | ||
3211 | param_field[1] = _ascebc['C']; | ||
3212 | param_field[2] = _ascebc['I']; | ||
3213 | param_field[3] = _ascebc['T']; | ||
3214 | *((unsigned int *) (¶m_field[4])) = QETH_PCI_THRESHOLD_A(card); | ||
3215 | *((unsigned int *) (¶m_field[8])) = QETH_PCI_THRESHOLD_B(card); | ||
3216 | *((unsigned int *) (¶m_field[12])) = QETH_PCI_TIMER_VALUE(card); | ||
3217 | } | ||
3218 | |||
3219 | static void | ||
3220 | qeth_create_qib_param_field_blkt(struct qeth_card *card, char *param_field) | ||
3221 | { | ||
3222 | param_field[16] = _ascebc['B']; | ||
3223 | param_field[17] = _ascebc['L']; | ||
3224 | param_field[18] = _ascebc['K']; | ||
3225 | param_field[19] = _ascebc['T']; | ||
3226 | *((unsigned int *) (¶m_field[20])) = card->info.blkt.time_total; | ||
3227 | *((unsigned int *) (¶m_field[24])) = card->info.blkt.inter_packet; | ||
3228 | *((unsigned int *) (¶m_field[28])) = card->info.blkt.inter_packet_jumbo; | ||
3229 | } | ||
3230 | |||
3231 | static void | ||
3232 | qeth_initialize_working_pool_list(struct qeth_card *card) | ||
3233 | { | ||
3234 | struct qeth_buffer_pool_entry *entry; | ||
3235 | |||
3236 | QETH_DBF_TEXT(trace,5,"inwrklst"); | ||
3237 | |||
3238 | list_for_each_entry(entry, | ||
3239 | &card->qdio.init_pool.entry_list, init_list) { | ||
3240 | qeth_put_buffer_pool_entry(card,entry); | ||
3241 | } | ||
3242 | } | ||
3243 | |||
3244 | static void | ||
3245 | qeth_clear_working_pool_list(struct qeth_card *card) | ||
3246 | { | ||
3247 | struct qeth_buffer_pool_entry *pool_entry, *tmp; | ||
3248 | |||
3249 | QETH_DBF_TEXT(trace,5,"clwrklst"); | ||
3250 | list_for_each_entry_safe(pool_entry, tmp, | ||
3251 | &card->qdio.in_buf_pool.entry_list, list){ | ||
3252 | list_del(&pool_entry->list); | ||
3253 | } | ||
3254 | } | ||
3255 | |||
3256 | static void | ||
3257 | qeth_free_buffer_pool(struct qeth_card *card) | ||
3258 | { | ||
3259 | struct qeth_buffer_pool_entry *pool_entry, *tmp; | ||
3260 | int i=0; | ||
3261 | QETH_DBF_TEXT(trace,5,"freepool"); | ||
3262 | list_for_each_entry_safe(pool_entry, tmp, | ||
3263 | &card->qdio.init_pool.entry_list, init_list){ | ||
3264 | for (i = 0; i < QETH_MAX_BUFFER_ELEMENTS(card); ++i) | ||
3265 | free_page((unsigned long)pool_entry->elements[i]); | ||
3266 | list_del(&pool_entry->init_list); | ||
3267 | kfree(pool_entry); | ||
3268 | } | ||
3269 | } | ||
3270 | |||
3271 | static int | ||
3272 | qeth_alloc_buffer_pool(struct qeth_card *card) | ||
3273 | { | ||
3274 | struct qeth_buffer_pool_entry *pool_entry; | ||
3275 | void *ptr; | ||
3276 | int i, j; | ||
3277 | |||
3278 | QETH_DBF_TEXT(trace,5,"alocpool"); | ||
3279 | for (i = 0; i < card->qdio.init_pool.buf_count; ++i){ | ||
3280 | pool_entry = kmalloc(sizeof(*pool_entry), GFP_KERNEL); | ||
3281 | if (!pool_entry){ | ||
3282 | qeth_free_buffer_pool(card); | ||
3283 | return -ENOMEM; | ||
3284 | } | ||
3285 | for(j = 0; j < QETH_MAX_BUFFER_ELEMENTS(card); ++j){ | ||
3286 | ptr = (void *) __get_free_page(GFP_KERNEL|GFP_DMA); | ||
3287 | if (!ptr) { | ||
3288 | while (j > 0) | ||
3289 | free_page((unsigned long) | ||
3290 | pool_entry->elements[--j]); | ||
3291 | kfree(pool_entry); | ||
3292 | qeth_free_buffer_pool(card); | ||
3293 | return -ENOMEM; | ||
3294 | } | ||
3295 | pool_entry->elements[j] = ptr; | ||
3296 | } | ||
3297 | list_add(&pool_entry->init_list, | ||
3298 | &card->qdio.init_pool.entry_list); | ||
3299 | } | ||
3300 | return 0; | ||
3301 | } | ||
3302 | |||
3303 | int | ||
3304 | qeth_realloc_buffer_pool(struct qeth_card *card, int bufcnt) | ||
3305 | { | ||
3306 | QETH_DBF_TEXT(trace, 2, "realcbp"); | ||
3307 | |||
3308 | if ((card->state != CARD_STATE_DOWN) && | ||
3309 | (card->state != CARD_STATE_RECOVER)) | ||
3310 | return -EPERM; | ||
3311 | |||
3312 | /* TODO: steel/add buffers from/to a running card's buffer pool (?) */ | ||
3313 | qeth_clear_working_pool_list(card); | ||
3314 | qeth_free_buffer_pool(card); | ||
3315 | card->qdio.in_buf_pool.buf_count = bufcnt; | ||
3316 | card->qdio.init_pool.buf_count = bufcnt; | ||
3317 | return qeth_alloc_buffer_pool(card); | ||
3318 | } | ||
3319 | |||
3320 | static int | ||
3321 | qeth_alloc_qdio_buffers(struct qeth_card *card) | ||
3322 | { | ||
3323 | int i, j; | ||
3324 | |||
3325 | QETH_DBF_TEXT(setup, 2, "allcqdbf"); | ||
3326 | |||
3327 | if (atomic_cmpxchg(&card->qdio.state, QETH_QDIO_UNINITIALIZED, | ||
3328 | QETH_QDIO_ALLOCATED) != QETH_QDIO_UNINITIALIZED) | ||
3329 | return 0; | ||
3330 | |||
3331 | card->qdio.in_q = kmalloc(sizeof(struct qeth_qdio_q), | ||
3332 | GFP_KERNEL|GFP_DMA); | ||
3333 | if (!card->qdio.in_q) | ||
3334 | goto out_nomem; | ||
3335 | QETH_DBF_TEXT(setup, 2, "inq"); | ||
3336 | QETH_DBF_HEX(setup, 2, &card->qdio.in_q, sizeof(void *)); | ||
3337 | memset(card->qdio.in_q, 0, sizeof(struct qeth_qdio_q)); | ||
3338 | /* give inbound qeth_qdio_buffers their qdio_buffers */ | ||
3339 | for (i = 0; i < QDIO_MAX_BUFFERS_PER_Q; ++i) | ||
3340 | card->qdio.in_q->bufs[i].buffer = | ||
3341 | &card->qdio.in_q->qdio_bufs[i]; | ||
3342 | /* inbound buffer pool */ | ||
3343 | if (qeth_alloc_buffer_pool(card)) | ||
3344 | goto out_freeinq; | ||
3345 | /* outbound */ | ||
3346 | card->qdio.out_qs = | ||
3347 | kmalloc(card->qdio.no_out_queues * | ||
3348 | sizeof(struct qeth_qdio_out_q *), GFP_KERNEL); | ||
3349 | if (!card->qdio.out_qs) | ||
3350 | goto out_freepool; | ||
3351 | for (i = 0; i < card->qdio.no_out_queues; ++i) { | ||
3352 | card->qdio.out_qs[i] = kmalloc(sizeof(struct qeth_qdio_out_q), | ||
3353 | GFP_KERNEL|GFP_DMA); | ||
3354 | if (!card->qdio.out_qs[i]) | ||
3355 | goto out_freeoutq; | ||
3356 | QETH_DBF_TEXT_(setup, 2, "outq %i", i); | ||
3357 | QETH_DBF_HEX(setup, 2, &card->qdio.out_qs[i], sizeof(void *)); | ||
3358 | memset(card->qdio.out_qs[i], 0, sizeof(struct qeth_qdio_out_q)); | ||
3359 | card->qdio.out_qs[i]->queue_no = i; | ||
3360 | /* give outbound qeth_qdio_buffers their qdio_buffers */ | ||
3361 | for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j){ | ||
3362 | card->qdio.out_qs[i]->bufs[j].buffer = | ||
3363 | &card->qdio.out_qs[i]->qdio_bufs[j]; | ||
3364 | skb_queue_head_init(&card->qdio.out_qs[i]->bufs[j]. | ||
3365 | skb_list); | ||
3366 | lockdep_set_class( | ||
3367 | &card->qdio.out_qs[i]->bufs[j].skb_list.lock, | ||
3368 | &qdio_out_skb_queue_key); | ||
3369 | INIT_LIST_HEAD(&card->qdio.out_qs[i]->bufs[j].ctx_list); | ||
3370 | } | ||
3371 | } | ||
3372 | return 0; | ||
3373 | |||
3374 | out_freeoutq: | ||
3375 | while (i > 0) | ||
3376 | kfree(card->qdio.out_qs[--i]); | ||
3377 | kfree(card->qdio.out_qs); | ||
3378 | card->qdio.out_qs = NULL; | ||
3379 | out_freepool: | ||
3380 | qeth_free_buffer_pool(card); | ||
3381 | out_freeinq: | ||
3382 | kfree(card->qdio.in_q); | ||
3383 | card->qdio.in_q = NULL; | ||
3384 | out_nomem: | ||
3385 | atomic_set(&card->qdio.state, QETH_QDIO_UNINITIALIZED); | ||
3386 | return -ENOMEM; | ||
3387 | } | ||
3388 | |||
3389 | static void | ||
3390 | qeth_free_qdio_buffers(struct qeth_card *card) | ||
3391 | { | ||
3392 | int i, j; | ||
3393 | |||
3394 | QETH_DBF_TEXT(trace, 2, "freeqdbf"); | ||
3395 | if (atomic_xchg(&card->qdio.state, QETH_QDIO_UNINITIALIZED) == | ||
3396 | QETH_QDIO_UNINITIALIZED) | ||
3397 | return; | ||
3398 | kfree(card->qdio.in_q); | ||
3399 | card->qdio.in_q = NULL; | ||
3400 | /* inbound buffer pool */ | ||
3401 | qeth_free_buffer_pool(card); | ||
3402 | /* free outbound qdio_qs */ | ||
3403 | if (card->qdio.out_qs) { | ||
3404 | for (i = 0; i < card->qdio.no_out_queues; ++i) { | ||
3405 | for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j) | ||
3406 | qeth_clear_output_buffer(card->qdio.out_qs[i], | ||
3407 | &card->qdio.out_qs[i]->bufs[j]); | ||
3408 | kfree(card->qdio.out_qs[i]); | ||
3409 | } | ||
3410 | kfree(card->qdio.out_qs); | ||
3411 | card->qdio.out_qs = NULL; | ||
3412 | } | ||
3413 | } | ||
3414 | |||
3415 | static void | ||
3416 | qeth_clear_qdio_buffers(struct qeth_card *card) | ||
3417 | { | ||
3418 | int i, j; | ||
3419 | |||
3420 | QETH_DBF_TEXT(trace, 2, "clearqdbf"); | ||
3421 | /* clear outbound buffers to free skbs */ | ||
3422 | for (i = 0; i < card->qdio.no_out_queues; ++i) | ||
3423 | if (card->qdio.out_qs && card->qdio.out_qs[i]) { | ||
3424 | for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j) | ||
3425 | qeth_clear_output_buffer(card->qdio.out_qs[i], | ||
3426 | &card->qdio.out_qs[i]->bufs[j]); | ||
3427 | } | ||
3428 | } | ||
3429 | |||
3430 | static void | ||
3431 | qeth_init_qdio_info(struct qeth_card *card) | ||
3432 | { | ||
3433 | QETH_DBF_TEXT(setup, 4, "intqdinf"); | ||
3434 | atomic_set(&card->qdio.state, QETH_QDIO_UNINITIALIZED); | ||
3435 | /* inbound */ | ||
3436 | card->qdio.in_buf_size = QETH_IN_BUF_SIZE_DEFAULT; | ||
3437 | card->qdio.init_pool.buf_count = QETH_IN_BUF_COUNT_DEFAULT; | ||
3438 | card->qdio.in_buf_pool.buf_count = card->qdio.init_pool.buf_count; | ||
3439 | INIT_LIST_HEAD(&card->qdio.in_buf_pool.entry_list); | ||
3440 | INIT_LIST_HEAD(&card->qdio.init_pool.entry_list); | ||
3441 | } | ||
3442 | |||
3443 | static int | ||
3444 | qeth_init_qdio_queues(struct qeth_card *card) | ||
3445 | { | ||
3446 | int i, j; | ||
3447 | int rc; | ||
3448 | |||
3449 | QETH_DBF_TEXT(setup, 2, "initqdqs"); | ||
3450 | |||
3451 | /* inbound queue */ | ||
3452 | memset(card->qdio.in_q->qdio_bufs, 0, | ||
3453 | QDIO_MAX_BUFFERS_PER_Q * sizeof(struct qdio_buffer)); | ||
3454 | qeth_initialize_working_pool_list(card); | ||
3455 | /*give only as many buffers to hardware as we have buffer pool entries*/ | ||
3456 | for (i = 0; i < card->qdio.in_buf_pool.buf_count - 1; ++i) | ||
3457 | qeth_init_input_buffer(card, &card->qdio.in_q->bufs[i]); | ||
3458 | card->qdio.in_q->next_buf_to_init = card->qdio.in_buf_pool.buf_count - 1; | ||
3459 | rc = do_QDIO(CARD_DDEV(card), QDIO_FLAG_SYNC_INPUT, 0, 0, | ||
3460 | card->qdio.in_buf_pool.buf_count - 1, NULL); | ||
3461 | if (rc) { | ||
3462 | QETH_DBF_TEXT_(setup, 2, "1err%d", rc); | ||
3463 | return rc; | ||
3464 | } | ||
3465 | rc = qdio_synchronize(CARD_DDEV(card), QDIO_FLAG_SYNC_INPUT, 0); | ||
3466 | if (rc) { | ||
3467 | QETH_DBF_TEXT_(setup, 2, "2err%d", rc); | ||
3468 | return rc; | ||
3469 | } | ||
3470 | /* outbound queue */ | ||
3471 | for (i = 0; i < card->qdio.no_out_queues; ++i){ | ||
3472 | memset(card->qdio.out_qs[i]->qdio_bufs, 0, | ||
3473 | QDIO_MAX_BUFFERS_PER_Q * sizeof(struct qdio_buffer)); | ||
3474 | for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j){ | ||
3475 | qeth_clear_output_buffer(card->qdio.out_qs[i], | ||
3476 | &card->qdio.out_qs[i]->bufs[j]); | ||
3477 | } | ||
3478 | card->qdio.out_qs[i]->card = card; | ||
3479 | card->qdio.out_qs[i]->next_buf_to_fill = 0; | ||
3480 | card->qdio.out_qs[i]->do_pack = 0; | ||
3481 | atomic_set(&card->qdio.out_qs[i]->used_buffers,0); | ||
3482 | atomic_set(&card->qdio.out_qs[i]->set_pci_flags_count, 0); | ||
3483 | atomic_set(&card->qdio.out_qs[i]->state, | ||
3484 | QETH_OUT_Q_UNLOCKED); | ||
3485 | } | ||
3486 | return 0; | ||
3487 | } | ||
3488 | |||
3489 | static int | ||
3490 | qeth_qdio_establish(struct qeth_card *card) | ||
3491 | { | ||
3492 | struct qdio_initialize init_data; | ||
3493 | char *qib_param_field; | ||
3494 | struct qdio_buffer **in_sbal_ptrs; | ||
3495 | struct qdio_buffer **out_sbal_ptrs; | ||
3496 | int i, j, k; | ||
3497 | int rc = 0; | ||
3498 | |||
3499 | QETH_DBF_TEXT(setup, 2, "qdioest"); | ||
3500 | |||
3501 | qib_param_field = kzalloc(QDIO_MAX_BUFFERS_PER_Q * sizeof(char), | ||
3502 | GFP_KERNEL); | ||
3503 | if (!qib_param_field) | ||
3504 | return -ENOMEM; | ||
3505 | |||
3506 | qeth_create_qib_param_field(card, qib_param_field); | ||
3507 | qeth_create_qib_param_field_blkt(card, qib_param_field); | ||
3508 | |||
3509 | in_sbal_ptrs = kmalloc(QDIO_MAX_BUFFERS_PER_Q * sizeof(void *), | ||
3510 | GFP_KERNEL); | ||
3511 | if (!in_sbal_ptrs) { | ||
3512 | kfree(qib_param_field); | ||
3513 | return -ENOMEM; | ||
3514 | } | ||
3515 | for(i = 0; i < QDIO_MAX_BUFFERS_PER_Q; ++i) | ||
3516 | in_sbal_ptrs[i] = (struct qdio_buffer *) | ||
3517 | virt_to_phys(card->qdio.in_q->bufs[i].buffer); | ||
3518 | |||
3519 | out_sbal_ptrs = | ||
3520 | kmalloc(card->qdio.no_out_queues * QDIO_MAX_BUFFERS_PER_Q * | ||
3521 | sizeof(void *), GFP_KERNEL); | ||
3522 | if (!out_sbal_ptrs) { | ||
3523 | kfree(in_sbal_ptrs); | ||
3524 | kfree(qib_param_field); | ||
3525 | return -ENOMEM; | ||
3526 | } | ||
3527 | for(i = 0, k = 0; i < card->qdio.no_out_queues; ++i) | ||
3528 | for(j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j, ++k){ | ||
3529 | out_sbal_ptrs[k] = (struct qdio_buffer *) | ||
3530 | virt_to_phys(card->qdio.out_qs[i]-> | ||
3531 | bufs[j].buffer); | ||
3532 | } | ||
3533 | |||
3534 | memset(&init_data, 0, sizeof(struct qdio_initialize)); | ||
3535 | init_data.cdev = CARD_DDEV(card); | ||
3536 | init_data.q_format = qeth_get_qdio_q_format(card); | ||
3537 | init_data.qib_param_field_format = 0; | ||
3538 | init_data.qib_param_field = qib_param_field; | ||
3539 | init_data.min_input_threshold = QETH_MIN_INPUT_THRESHOLD; | ||
3540 | init_data.max_input_threshold = QETH_MAX_INPUT_THRESHOLD; | ||
3541 | init_data.min_output_threshold = QETH_MIN_OUTPUT_THRESHOLD; | ||
3542 | init_data.max_output_threshold = QETH_MAX_OUTPUT_THRESHOLD; | ||
3543 | init_data.no_input_qs = 1; | ||
3544 | init_data.no_output_qs = card->qdio.no_out_queues; | ||
3545 | init_data.input_handler = (qdio_handler_t *) | ||
3546 | qeth_qdio_input_handler; | ||
3547 | init_data.output_handler = (qdio_handler_t *) | ||
3548 | qeth_qdio_output_handler; | ||
3549 | init_data.int_parm = (unsigned long) card; | ||
3550 | init_data.flags = QDIO_INBOUND_0COPY_SBALS | | ||
3551 | QDIO_OUTBOUND_0COPY_SBALS | | ||
3552 | QDIO_USE_OUTBOUND_PCIS; | ||
3553 | init_data.input_sbal_addr_array = (void **) in_sbal_ptrs; | ||
3554 | init_data.output_sbal_addr_array = (void **) out_sbal_ptrs; | ||
3555 | |||
3556 | if (atomic_cmpxchg(&card->qdio.state, QETH_QDIO_ALLOCATED, | ||
3557 | QETH_QDIO_ESTABLISHED) == QETH_QDIO_ALLOCATED) | ||
3558 | if ((rc = qdio_initialize(&init_data))) | ||
3559 | atomic_set(&card->qdio.state, QETH_QDIO_ALLOCATED); | ||
3560 | |||
3561 | kfree(out_sbal_ptrs); | ||
3562 | kfree(in_sbal_ptrs); | ||
3563 | kfree(qib_param_field); | ||
3564 | return rc; | ||
3565 | } | ||
3566 | |||
3567 | static int | ||
3568 | qeth_qdio_activate(struct qeth_card *card) | ||
3569 | { | ||
3570 | QETH_DBF_TEXT(setup,3,"qdioact"); | ||
3571 | return qdio_activate(CARD_DDEV(card), 0); | ||
3572 | } | ||
3573 | |||
3574 | static int | ||
3575 | qeth_clear_channel(struct qeth_channel *channel) | ||
3576 | { | ||
3577 | unsigned long flags; | ||
3578 | struct qeth_card *card; | ||
3579 | int rc; | ||
3580 | |||
3581 | QETH_DBF_TEXT(trace,3,"clearch"); | ||
3582 | card = CARD_FROM_CDEV(channel->ccwdev); | ||
3583 | spin_lock_irqsave(get_ccwdev_lock(channel->ccwdev), flags); | ||
3584 | rc = ccw_device_clear(channel->ccwdev, QETH_CLEAR_CHANNEL_PARM); | ||
3585 | spin_unlock_irqrestore(get_ccwdev_lock(channel->ccwdev), flags); | ||
3586 | |||
3587 | if (rc) | ||
3588 | return rc; | ||
3589 | rc = wait_event_interruptible_timeout(card->wait_q, | ||
3590 | channel->state==CH_STATE_STOPPED, QETH_TIMEOUT); | ||
3591 | if (rc == -ERESTARTSYS) | ||
3592 | return rc; | ||
3593 | if (channel->state != CH_STATE_STOPPED) | ||
3594 | return -ETIME; | ||
3595 | channel->state = CH_STATE_DOWN; | ||
3596 | return 0; | ||
3597 | } | ||
3598 | |||
3599 | static int | ||
3600 | qeth_halt_channel(struct qeth_channel *channel) | ||
3601 | { | ||
3602 | unsigned long flags; | ||
3603 | struct qeth_card *card; | ||
3604 | int rc; | ||
3605 | |||
3606 | QETH_DBF_TEXT(trace,3,"haltch"); | ||
3607 | card = CARD_FROM_CDEV(channel->ccwdev); | ||
3608 | spin_lock_irqsave(get_ccwdev_lock(channel->ccwdev), flags); | ||
3609 | rc = ccw_device_halt(channel->ccwdev, QETH_HALT_CHANNEL_PARM); | ||
3610 | spin_unlock_irqrestore(get_ccwdev_lock(channel->ccwdev), flags); | ||
3611 | |||
3612 | if (rc) | ||
3613 | return rc; | ||
3614 | rc = wait_event_interruptible_timeout(card->wait_q, | ||
3615 | channel->state==CH_STATE_HALTED, QETH_TIMEOUT); | ||
3616 | if (rc == -ERESTARTSYS) | ||
3617 | return rc; | ||
3618 | if (channel->state != CH_STATE_HALTED) | ||
3619 | return -ETIME; | ||
3620 | return 0; | ||
3621 | } | ||
3622 | |||
3623 | static int | ||
3624 | qeth_halt_channels(struct qeth_card *card) | ||
3625 | { | ||
3626 | int rc1 = 0, rc2=0, rc3 = 0; | ||
3627 | |||
3628 | QETH_DBF_TEXT(trace,3,"haltchs"); | ||
3629 | rc1 = qeth_halt_channel(&card->read); | ||
3630 | rc2 = qeth_halt_channel(&card->write); | ||
3631 | rc3 = qeth_halt_channel(&card->data); | ||
3632 | if (rc1) | ||
3633 | return rc1; | ||
3634 | if (rc2) | ||
3635 | return rc2; | ||
3636 | return rc3; | ||
3637 | } | ||
3638 | static int | ||
3639 | qeth_clear_channels(struct qeth_card *card) | ||
3640 | { | ||
3641 | int rc1 = 0, rc2=0, rc3 = 0; | ||
3642 | |||
3643 | QETH_DBF_TEXT(trace,3,"clearchs"); | ||
3644 | rc1 = qeth_clear_channel(&card->read); | ||
3645 | rc2 = qeth_clear_channel(&card->write); | ||
3646 | rc3 = qeth_clear_channel(&card->data); | ||
3647 | if (rc1) | ||
3648 | return rc1; | ||
3649 | if (rc2) | ||
3650 | return rc2; | ||
3651 | return rc3; | ||
3652 | } | ||
3653 | |||
3654 | static int | ||
3655 | qeth_clear_halt_card(struct qeth_card *card, int halt) | ||
3656 | { | ||
3657 | int rc = 0; | ||
3658 | |||
3659 | QETH_DBF_TEXT(trace,3,"clhacrd"); | ||
3660 | QETH_DBF_HEX(trace, 3, &card, sizeof(void *)); | ||
3661 | |||
3662 | if (halt) | ||
3663 | rc = qeth_halt_channels(card); | ||
3664 | if (rc) | ||
3665 | return rc; | ||
3666 | return qeth_clear_channels(card); | ||
3667 | } | ||
3668 | |||
3669 | static int | ||
3670 | qeth_qdio_clear_card(struct qeth_card *card, int use_halt) | ||
3671 | { | ||
3672 | int rc = 0; | ||
3673 | |||
3674 | QETH_DBF_TEXT(trace,3,"qdioclr"); | ||
3675 | switch (atomic_cmpxchg(&card->qdio.state, QETH_QDIO_ESTABLISHED, | ||
3676 | QETH_QDIO_CLEANING)) { | ||
3677 | case QETH_QDIO_ESTABLISHED: | ||
3678 | if ((rc = qdio_cleanup(CARD_DDEV(card), | ||
3679 | (card->info.type == QETH_CARD_TYPE_IQD) ? | ||
3680 | QDIO_FLAG_CLEANUP_USING_HALT : | ||
3681 | QDIO_FLAG_CLEANUP_USING_CLEAR))) | ||
3682 | QETH_DBF_TEXT_(trace, 3, "1err%d", rc); | ||
3683 | atomic_set(&card->qdio.state, QETH_QDIO_ALLOCATED); | ||
3684 | break; | ||
3685 | case QETH_QDIO_CLEANING: | ||
3686 | return rc; | ||
3687 | default: | ||
3688 | break; | ||
3689 | } | ||
3690 | if ((rc = qeth_clear_halt_card(card, use_halt))) | ||
3691 | QETH_DBF_TEXT_(trace, 3, "2err%d", rc); | ||
3692 | card->state = CARD_STATE_DOWN; | ||
3693 | return rc; | ||
3694 | } | ||
3695 | |||
3696 | static int | ||
3697 | qeth_dm_act(struct qeth_card *card) | ||
3698 | { | ||
3699 | int rc; | ||
3700 | struct qeth_cmd_buffer *iob; | ||
3701 | |||
3702 | QETH_DBF_TEXT(setup,2,"dmact"); | ||
3703 | |||
3704 | iob = qeth_wait_for_buffer(&card->write); | ||
3705 | memcpy(iob->data, DM_ACT, DM_ACT_SIZE); | ||
3706 | |||
3707 | memcpy(QETH_DM_ACT_DEST_ADDR(iob->data), | ||
3708 | &card->token.cm_connection_r, QETH_MPC_TOKEN_LENGTH); | ||
3709 | memcpy(QETH_DM_ACT_CONNECTION_TOKEN(iob->data), | ||
3710 | &card->token.ulp_connection_r, QETH_MPC_TOKEN_LENGTH); | ||
3711 | rc = qeth_send_control_data(card, DM_ACT_SIZE, iob, NULL, NULL); | ||
3712 | return rc; | ||
3713 | } | ||
3714 | |||
3715 | static int | ||
3716 | qeth_mpc_initialize(struct qeth_card *card) | ||
3717 | { | ||
3718 | int rc; | ||
3719 | |||
3720 | QETH_DBF_TEXT(setup,2,"mpcinit"); | ||
3721 | |||
3722 | if ((rc = qeth_issue_next_read(card))){ | ||
3723 | QETH_DBF_TEXT_(setup, 2, "1err%d", rc); | ||
3724 | return rc; | ||
3725 | } | ||
3726 | if ((rc = qeth_cm_enable(card))){ | ||
3727 | QETH_DBF_TEXT_(setup, 2, "2err%d", rc); | ||
3728 | goto out_qdio; | ||
3729 | } | ||
3730 | if ((rc = qeth_cm_setup(card))){ | ||
3731 | QETH_DBF_TEXT_(setup, 2, "3err%d", rc); | ||
3732 | goto out_qdio; | ||
3733 | } | ||
3734 | if ((rc = qeth_ulp_enable(card))){ | ||
3735 | QETH_DBF_TEXT_(setup, 2, "4err%d", rc); | ||
3736 | goto out_qdio; | ||
3737 | } | ||
3738 | if ((rc = qeth_ulp_setup(card))){ | ||
3739 | QETH_DBF_TEXT_(setup, 2, "5err%d", rc); | ||
3740 | goto out_qdio; | ||
3741 | } | ||
3742 | if ((rc = qeth_alloc_qdio_buffers(card))){ | ||
3743 | QETH_DBF_TEXT_(setup, 2, "5err%d", rc); | ||
3744 | goto out_qdio; | ||
3745 | } | ||
3746 | if ((rc = qeth_qdio_establish(card))){ | ||
3747 | QETH_DBF_TEXT_(setup, 2, "6err%d", rc); | ||
3748 | qeth_free_qdio_buffers(card); | ||
3749 | goto out_qdio; | ||
3750 | } | ||
3751 | if ((rc = qeth_qdio_activate(card))){ | ||
3752 | QETH_DBF_TEXT_(setup, 2, "7err%d", rc); | ||
3753 | goto out_qdio; | ||
3754 | } | ||
3755 | if ((rc = qeth_dm_act(card))){ | ||
3756 | QETH_DBF_TEXT_(setup, 2, "8err%d", rc); | ||
3757 | goto out_qdio; | ||
3758 | } | ||
3759 | |||
3760 | return 0; | ||
3761 | out_qdio: | ||
3762 | qeth_qdio_clear_card(card, card->info.type!=QETH_CARD_TYPE_IQD); | ||
3763 | return rc; | ||
3764 | } | ||
3765 | |||
3766 | static struct net_device * | ||
3767 | qeth_get_netdevice(enum qeth_card_types type, enum qeth_link_types linktype) | ||
3768 | { | ||
3769 | struct net_device *dev = NULL; | ||
3770 | |||
3771 | switch (type) { | ||
3772 | case QETH_CARD_TYPE_OSAE: | ||
3773 | switch (linktype) { | ||
3774 | case QETH_LINK_TYPE_LANE_TR: | ||
3775 | case QETH_LINK_TYPE_HSTR: | ||
3776 | #ifdef CONFIG_TR | ||
3777 | dev = alloc_trdev(0); | ||
3778 | #endif /* CONFIG_TR */ | ||
3779 | break; | ||
3780 | default: | ||
3781 | dev = alloc_etherdev(0); | ||
3782 | } | ||
3783 | break; | ||
3784 | case QETH_CARD_TYPE_IQD: | ||
3785 | dev = alloc_netdev(0, "hsi%d", ether_setup); | ||
3786 | break; | ||
3787 | case QETH_CARD_TYPE_OSN: | ||
3788 | dev = alloc_netdev(0, "osn%d", ether_setup); | ||
3789 | break; | ||
3790 | default: | ||
3791 | dev = alloc_etherdev(0); | ||
3792 | } | ||
3793 | return dev; | ||
3794 | } | ||
3795 | |||
3796 | /*hard_header fake function; used in case fake_ll is set */ | ||
3797 | static int | ||
3798 | qeth_fake_header(struct sk_buff *skb, struct net_device *dev, | ||
3799 | unsigned short type, const void *daddr, const void *saddr, | ||
3800 | unsigned len) | ||
3801 | { | ||
3802 | if(dev->type == ARPHRD_IEEE802_TR){ | ||
3803 | struct trh_hdr *hdr; | ||
3804 | hdr = (struct trh_hdr *)skb_push(skb, QETH_FAKE_LL_LEN_TR); | ||
3805 | memcpy(hdr->saddr, dev->dev_addr, TR_ALEN); | ||
3806 | memcpy(hdr->daddr, "FAKELL", TR_ALEN); | ||
3807 | return QETH_FAKE_LL_LEN_TR; | ||
3808 | |||
3809 | } else { | ||
3810 | struct ethhdr *hdr; | ||
3811 | hdr = (struct ethhdr *)skb_push(skb, QETH_FAKE_LL_LEN_ETH); | ||
3812 | memcpy(hdr->h_source, dev->dev_addr, ETH_ALEN); | ||
3813 | memcpy(hdr->h_dest, "FAKELL", ETH_ALEN); | ||
3814 | if (type != ETH_P_802_3) | ||
3815 | hdr->h_proto = htons(type); | ||
3816 | else | ||
3817 | hdr->h_proto = htons(len); | ||
3818 | return QETH_FAKE_LL_LEN_ETH; | ||
3819 | |||
3820 | } | ||
3821 | } | ||
3822 | |||
3823 | static const struct header_ops qeth_fake_ops = { | ||
3824 | .create = qeth_fake_header, | ||
3825 | .parse = qeth_hard_header_parse, | ||
3826 | }; | ||
3827 | |||
3828 | static int | ||
3829 | qeth_send_packet(struct qeth_card *, struct sk_buff *); | ||
3830 | |||
3831 | static int | ||
3832 | qeth_hard_start_xmit(struct sk_buff *skb, struct net_device *dev) | ||
3833 | { | ||
3834 | int rc; | ||
3835 | struct qeth_card *card; | ||
3836 | |||
3837 | QETH_DBF_TEXT(trace, 6, "hrdstxmi"); | ||
3838 | card = (struct qeth_card *)dev->priv; | ||
3839 | if (skb==NULL) { | ||
3840 | card->stats.tx_dropped++; | ||
3841 | card->stats.tx_errors++; | ||
3842 | /* return OK; otherwise ksoftirqd goes to 100% */ | ||
3843 | return NETDEV_TX_OK; | ||
3844 | } | ||
3845 | if ((card->state != CARD_STATE_UP) || !card->lan_online) { | ||
3846 | card->stats.tx_dropped++; | ||
3847 | card->stats.tx_errors++; | ||
3848 | card->stats.tx_carrier_errors++; | ||
3849 | dev_kfree_skb_any(skb); | ||
3850 | /* return OK; otherwise ksoftirqd goes to 100% */ | ||
3851 | return NETDEV_TX_OK; | ||
3852 | } | ||
3853 | if (card->options.performance_stats) { | ||
3854 | card->perf_stats.outbound_cnt++; | ||
3855 | card->perf_stats.outbound_start_time = qeth_get_micros(); | ||
3856 | } | ||
3857 | netif_stop_queue(dev); | ||
3858 | if ((rc = qeth_send_packet(card, skb))) { | ||
3859 | if (rc == -EBUSY) { | ||
3860 | return NETDEV_TX_BUSY; | ||
3861 | } else { | ||
3862 | card->stats.tx_errors++; | ||
3863 | card->stats.tx_dropped++; | ||
3864 | dev_kfree_skb_any(skb); | ||
3865 | /*set to OK; otherwise ksoftirqd goes to 100% */ | ||
3866 | rc = NETDEV_TX_OK; | ||
3867 | } | ||
3868 | } | ||
3869 | netif_wake_queue(dev); | ||
3870 | if (card->options.performance_stats) | ||
3871 | card->perf_stats.outbound_time += qeth_get_micros() - | ||
3872 | card->perf_stats.outbound_start_time; | ||
3873 | return rc; | ||
3874 | } | ||
3875 | |||
3876 | static int | ||
3877 | qeth_verify_vlan_dev(struct net_device *dev, struct qeth_card *card) | ||
3878 | { | ||
3879 | int rc = 0; | ||
3880 | #ifdef CONFIG_QETH_VLAN | ||
3881 | struct vlan_group *vg; | ||
3882 | int i; | ||
3883 | |||
3884 | if (!(vg = card->vlangrp)) | ||
3885 | return rc; | ||
3886 | |||
3887 | for (i = 0; i < VLAN_GROUP_ARRAY_LEN; i++){ | ||
3888 | if (vlan_group_get_device(vg, i) == dev){ | ||
3889 | rc = QETH_VLAN_CARD; | ||
3890 | break; | ||
3891 | } | ||
3892 | } | ||
3893 | if (rc && !(vlan_dev_info(dev)->real_dev->priv == (void *)card)) | ||
3894 | return 0; | ||
3895 | |||
3896 | #endif | ||
3897 | return rc; | ||
3898 | } | ||
3899 | |||
3900 | static int | ||
3901 | qeth_verify_dev(struct net_device *dev) | ||
3902 | { | ||
3903 | struct qeth_card *card; | ||
3904 | unsigned long flags; | ||
3905 | int rc = 0; | ||
3906 | |||
3907 | read_lock_irqsave(&qeth_card_list.rwlock, flags); | ||
3908 | list_for_each_entry(card, &qeth_card_list.list, list){ | ||
3909 | if (card->dev == dev){ | ||
3910 | rc = QETH_REAL_CARD; | ||
3911 | break; | ||
3912 | } | ||
3913 | rc = qeth_verify_vlan_dev(dev, card); | ||
3914 | if (rc) | ||
3915 | break; | ||
3916 | } | ||
3917 | read_unlock_irqrestore(&qeth_card_list.rwlock, flags); | ||
3918 | |||
3919 | return rc; | ||
3920 | } | ||
3921 | |||
3922 | static struct qeth_card * | ||
3923 | qeth_get_card_from_dev(struct net_device *dev) | ||
3924 | { | ||
3925 | struct qeth_card *card = NULL; | ||
3926 | int rc; | ||
3927 | |||
3928 | rc = qeth_verify_dev(dev); | ||
3929 | if (rc == QETH_REAL_CARD) | ||
3930 | card = (struct qeth_card *)dev->priv; | ||
3931 | else if (rc == QETH_VLAN_CARD) | ||
3932 | card = (struct qeth_card *) | ||
3933 | vlan_dev_info(dev)->real_dev->priv; | ||
3934 | |||
3935 | QETH_DBF_TEXT_(trace, 4, "%d", rc); | ||
3936 | return card ; | ||
3937 | } | ||
3938 | |||
3939 | static void | ||
3940 | qeth_tx_timeout(struct net_device *dev) | ||
3941 | { | ||
3942 | struct qeth_card *card; | ||
3943 | |||
3944 | card = (struct qeth_card *) dev->priv; | ||
3945 | card->stats.tx_errors++; | ||
3946 | qeth_schedule_recovery(card); | ||
3947 | } | ||
3948 | |||
3949 | static int | ||
3950 | qeth_open(struct net_device *dev) | ||
3951 | { | ||
3952 | struct qeth_card *card; | ||
3953 | |||
3954 | QETH_DBF_TEXT(trace, 4, "qethopen"); | ||
3955 | |||
3956 | card = (struct qeth_card *) dev->priv; | ||
3957 | |||
3958 | if (card->state != CARD_STATE_SOFTSETUP) | ||
3959 | return -ENODEV; | ||
3960 | |||
3961 | if ( (card->info.type != QETH_CARD_TYPE_OSN) && | ||
3962 | (card->options.layer2) && | ||
3963 | (!(card->info.mac_bits & QETH_LAYER2_MAC_REGISTERED))) { | ||
3964 | QETH_DBF_TEXT(trace,4,"nomacadr"); | ||
3965 | return -EPERM; | ||
3966 | } | ||
3967 | card->data.state = CH_STATE_UP; | ||
3968 | card->state = CARD_STATE_UP; | ||
3969 | card->dev->flags |= IFF_UP; | ||
3970 | netif_start_queue(dev); | ||
3971 | |||
3972 | if (!card->lan_online && netif_carrier_ok(dev)) | ||
3973 | netif_carrier_off(dev); | ||
3974 | return 0; | ||
3975 | } | ||
3976 | |||
3977 | static int | ||
3978 | qeth_stop(struct net_device *dev) | ||
3979 | { | ||
3980 | struct qeth_card *card; | ||
3981 | |||
3982 | QETH_DBF_TEXT(trace, 4, "qethstop"); | ||
3983 | |||
3984 | card = (struct qeth_card *) dev->priv; | ||
3985 | |||
3986 | netif_tx_disable(dev); | ||
3987 | card->dev->flags &= ~IFF_UP; | ||
3988 | if (card->state == CARD_STATE_UP) | ||
3989 | card->state = CARD_STATE_SOFTSETUP; | ||
3990 | return 0; | ||
3991 | } | ||
3992 | |||
3993 | static int | ||
3994 | qeth_get_cast_type(struct qeth_card *card, struct sk_buff *skb) | ||
3995 | { | ||
3996 | int cast_type = RTN_UNSPEC; | ||
3997 | |||
3998 | if (card->info.type == QETH_CARD_TYPE_OSN) | ||
3999 | return cast_type; | ||
4000 | |||
4001 | if (skb->dst && skb->dst->neighbour){ | ||
4002 | cast_type = skb->dst->neighbour->type; | ||
4003 | if ((cast_type == RTN_BROADCAST) || | ||
4004 | (cast_type == RTN_MULTICAST) || | ||
4005 | (cast_type == RTN_ANYCAST)) | ||
4006 | return cast_type; | ||
4007 | else | ||
4008 | return RTN_UNSPEC; | ||
4009 | } | ||
4010 | /* try something else */ | ||
4011 | if (skb->protocol == ETH_P_IPV6) | ||
4012 | return (skb_network_header(skb)[24] == 0xff) ? | ||
4013 | RTN_MULTICAST : 0; | ||
4014 | else if (skb->protocol == ETH_P_IP) | ||
4015 | return ((skb_network_header(skb)[16] & 0xf0) == 0xe0) ? | ||
4016 | RTN_MULTICAST : 0; | ||
4017 | /* ... */ | ||
4018 | if (!memcmp(skb->data, skb->dev->broadcast, 6)) | ||
4019 | return RTN_BROADCAST; | ||
4020 | else { | ||
4021 | u16 hdr_mac; | ||
4022 | |||
4023 | hdr_mac = *((u16 *)skb->data); | ||
4024 | /* tr multicast? */ | ||
4025 | switch (card->info.link_type) { | ||
4026 | case QETH_LINK_TYPE_HSTR: | ||
4027 | case QETH_LINK_TYPE_LANE_TR: | ||
4028 | if ((hdr_mac == QETH_TR_MAC_NC) || | ||
4029 | (hdr_mac == QETH_TR_MAC_C)) | ||
4030 | return RTN_MULTICAST; | ||
4031 | break; | ||
4032 | /* eth or so multicast? */ | ||
4033 | default: | ||
4034 | if ((hdr_mac == QETH_ETH_MAC_V4) || | ||
4035 | (hdr_mac == QETH_ETH_MAC_V6)) | ||
4036 | return RTN_MULTICAST; | ||
4037 | } | ||
4038 | } | ||
4039 | return cast_type; | ||
4040 | } | ||
4041 | |||
4042 | static int | ||
4043 | qeth_get_priority_queue(struct qeth_card *card, struct sk_buff *skb, | ||
4044 | int ipv, int cast_type) | ||
4045 | { | ||
4046 | if (!ipv && (card->info.type == QETH_CARD_TYPE_OSAE)) | ||
4047 | return card->qdio.default_out_queue; | ||
4048 | switch (card->qdio.no_out_queues) { | ||
4049 | case 4: | ||
4050 | if (cast_type && card->info.is_multicast_different) | ||
4051 | return card->info.is_multicast_different & | ||
4052 | (card->qdio.no_out_queues - 1); | ||
4053 | if (card->qdio.do_prio_queueing && (ipv == 4)) { | ||
4054 | const u8 tos = ip_hdr(skb)->tos; | ||
4055 | |||
4056 | if (card->qdio.do_prio_queueing==QETH_PRIO_Q_ING_TOS){ | ||
4057 | if (tos & IP_TOS_NOTIMPORTANT) | ||
4058 | return 3; | ||
4059 | if (tos & IP_TOS_HIGHRELIABILITY) | ||
4060 | return 2; | ||
4061 | if (tos & IP_TOS_HIGHTHROUGHPUT) | ||
4062 | return 1; | ||
4063 | if (tos & IP_TOS_LOWDELAY) | ||
4064 | return 0; | ||
4065 | } | ||
4066 | if (card->qdio.do_prio_queueing==QETH_PRIO_Q_ING_PREC) | ||
4067 | return 3 - (tos >> 6); | ||
4068 | } else if (card->qdio.do_prio_queueing && (ipv == 6)) { | ||
4069 | /* TODO: IPv6!!! */ | ||
4070 | } | ||
4071 | return card->qdio.default_out_queue; | ||
4072 | case 1: /* fallthrough for single-out-queue 1920-device */ | ||
4073 | default: | ||
4074 | return card->qdio.default_out_queue; | ||
4075 | } | ||
4076 | } | ||
4077 | |||
4078 | static inline int | ||
4079 | qeth_get_ip_version(struct sk_buff *skb) | ||
4080 | { | ||
4081 | switch (skb->protocol) { | ||
4082 | case ETH_P_IPV6: | ||
4083 | return 6; | ||
4084 | case ETH_P_IP: | ||
4085 | return 4; | ||
4086 | default: | ||
4087 | return 0; | ||
4088 | } | ||
4089 | } | ||
4090 | |||
4091 | static struct qeth_hdr * | ||
4092 | __qeth_prepare_skb(struct qeth_card *card, struct sk_buff *skb, int ipv) | ||
4093 | { | ||
4094 | #ifdef CONFIG_QETH_VLAN | ||
4095 | u16 *tag; | ||
4096 | if (card->vlangrp && vlan_tx_tag_present(skb) && | ||
4097 | ((ipv == 6) || card->options.layer2) ) { | ||
4098 | /* | ||
4099 | * Move the mac addresses (6 bytes src, 6 bytes dest) | ||
4100 | * to the beginning of the new header. We are using three | ||
4101 | * memcpys instead of one memmove to save cycles. | ||
4102 | */ | ||
4103 | skb_push(skb, VLAN_HLEN); | ||
4104 | skb_copy_to_linear_data(skb, skb->data + 4, 4); | ||
4105 | skb_copy_to_linear_data_offset(skb, 4, skb->data + 8, 4); | ||
4106 | skb_copy_to_linear_data_offset(skb, 8, skb->data + 12, 4); | ||
4107 | tag = (u16 *)(skb->data + 12); | ||
4108 | /* | ||
4109 | * first two bytes = ETH_P_8021Q (0x8100) | ||
4110 | * second two bytes = VLANID | ||
4111 | */ | ||
4112 | *tag = __constant_htons(ETH_P_8021Q); | ||
4113 | *(tag + 1) = htons(vlan_tx_tag_get(skb)); | ||
4114 | } | ||
4115 | #endif | ||
4116 | return ((struct qeth_hdr *) | ||
4117 | qeth_push_skb(card, skb, sizeof(struct qeth_hdr))); | ||
4118 | } | ||
4119 | |||
4120 | static void | ||
4121 | __qeth_free_new_skb(struct sk_buff *orig_skb, struct sk_buff *new_skb) | ||
4122 | { | ||
4123 | if (orig_skb != new_skb) | ||
4124 | dev_kfree_skb_any(new_skb); | ||
4125 | } | ||
4126 | |||
4127 | static struct sk_buff * | ||
4128 | qeth_prepare_skb(struct qeth_card *card, struct sk_buff *skb, | ||
4129 | struct qeth_hdr **hdr, int ipv) | ||
4130 | { | ||
4131 | struct sk_buff *new_skb, *new_skb2; | ||
4132 | |||
4133 | QETH_DBF_TEXT(trace, 6, "prepskb"); | ||
4134 | new_skb = skb; | ||
4135 | new_skb = qeth_pskb_unshare(skb, GFP_ATOMIC); | ||
4136 | if (!new_skb) | ||
4137 | return NULL; | ||
4138 | new_skb2 = qeth_realloc_headroom(card, new_skb, | ||
4139 | sizeof(struct qeth_hdr)); | ||
4140 | if (!new_skb2) { | ||
4141 | __qeth_free_new_skb(skb, new_skb); | ||
4142 | return NULL; | ||
4143 | } | ||
4144 | if (new_skb != skb) | ||
4145 | __qeth_free_new_skb(new_skb2, new_skb); | ||
4146 | new_skb = new_skb2; | ||
4147 | *hdr = __qeth_prepare_skb(card, new_skb, ipv); | ||
4148 | if (*hdr == NULL) { | ||
4149 | __qeth_free_new_skb(skb, new_skb); | ||
4150 | return NULL; | ||
4151 | } | ||
4152 | return new_skb; | ||
4153 | } | ||
4154 | |||
4155 | static inline u8 | ||
4156 | qeth_get_qeth_hdr_flags4(int cast_type) | ||
4157 | { | ||
4158 | if (cast_type == RTN_MULTICAST) | ||
4159 | return QETH_CAST_MULTICAST; | ||
4160 | if (cast_type == RTN_BROADCAST) | ||
4161 | return QETH_CAST_BROADCAST; | ||
4162 | return QETH_CAST_UNICAST; | ||
4163 | } | ||
4164 | |||
4165 | static inline u8 | ||
4166 | qeth_get_qeth_hdr_flags6(int cast_type) | ||
4167 | { | ||
4168 | u8 ct = QETH_HDR_PASSTHRU | QETH_HDR_IPV6; | ||
4169 | if (cast_type == RTN_MULTICAST) | ||
4170 | return ct | QETH_CAST_MULTICAST; | ||
4171 | if (cast_type == RTN_ANYCAST) | ||
4172 | return ct | QETH_CAST_ANYCAST; | ||
4173 | if (cast_type == RTN_BROADCAST) | ||
4174 | return ct | QETH_CAST_BROADCAST; | ||
4175 | return ct | QETH_CAST_UNICAST; | ||
4176 | } | ||
4177 | |||
4178 | static void | ||
4179 | qeth_layer2_get_packet_type(struct qeth_card *card, struct qeth_hdr *hdr, | ||
4180 | struct sk_buff *skb) | ||
4181 | { | ||
4182 | __u16 hdr_mac; | ||
4183 | |||
4184 | if (!memcmp(skb->data+QETH_HEADER_SIZE, | ||
4185 | skb->dev->broadcast,6)) { /* broadcast? */ | ||
4186 | *(__u32 *)hdr->hdr.l2.flags |= | ||
4187 | QETH_LAYER2_FLAG_BROADCAST << 8; | ||
4188 | return; | ||
4189 | } | ||
4190 | hdr_mac=*((__u16*)skb->data); | ||
4191 | /* tr multicast? */ | ||
4192 | switch (card->info.link_type) { | ||
4193 | case QETH_LINK_TYPE_HSTR: | ||
4194 | case QETH_LINK_TYPE_LANE_TR: | ||
4195 | if ((hdr_mac == QETH_TR_MAC_NC) || | ||
4196 | (hdr_mac == QETH_TR_MAC_C) ) | ||
4197 | *(__u32 *)hdr->hdr.l2.flags |= | ||
4198 | QETH_LAYER2_FLAG_MULTICAST << 8; | ||
4199 | else | ||
4200 | *(__u32 *)hdr->hdr.l2.flags |= | ||
4201 | QETH_LAYER2_FLAG_UNICAST << 8; | ||
4202 | break; | ||
4203 | /* eth or so multicast? */ | ||
4204 | default: | ||
4205 | if ( (hdr_mac==QETH_ETH_MAC_V4) || | ||
4206 | (hdr_mac==QETH_ETH_MAC_V6) ) | ||
4207 | *(__u32 *)hdr->hdr.l2.flags |= | ||
4208 | QETH_LAYER2_FLAG_MULTICAST << 8; | ||
4209 | else | ||
4210 | *(__u32 *)hdr->hdr.l2.flags |= | ||
4211 | QETH_LAYER2_FLAG_UNICAST << 8; | ||
4212 | } | ||
4213 | } | ||
4214 | |||
4215 | static void | ||
4216 | qeth_layer2_fill_header(struct qeth_card *card, struct qeth_hdr *hdr, | ||
4217 | struct sk_buff *skb, int cast_type) | ||
4218 | { | ||
4219 | memset(hdr, 0, sizeof(struct qeth_hdr)); | ||
4220 | hdr->hdr.l2.id = QETH_HEADER_TYPE_LAYER2; | ||
4221 | |||
4222 | /* set byte 0 to "0x02" and byte 3 to casting flags */ | ||
4223 | if (cast_type==RTN_MULTICAST) | ||
4224 | *(__u32 *)hdr->hdr.l2.flags |= QETH_LAYER2_FLAG_MULTICAST << 8; | ||
4225 | else if (cast_type==RTN_BROADCAST) | ||
4226 | *(__u32 *)hdr->hdr.l2.flags |= QETH_LAYER2_FLAG_BROADCAST << 8; | ||
4227 | else | ||
4228 | qeth_layer2_get_packet_type(card, hdr, skb); | ||
4229 | |||
4230 | hdr->hdr.l2.pkt_length = skb->len-QETH_HEADER_SIZE; | ||
4231 | #ifdef CONFIG_QETH_VLAN | ||
4232 | /* VSWITCH relies on the VLAN | ||
4233 | * information to be present in | ||
4234 | * the QDIO header */ | ||
4235 | if ((card->vlangrp != NULL) && | ||
4236 | vlan_tx_tag_present(skb)) { | ||
4237 | *(__u32 *)hdr->hdr.l2.flags |= QETH_LAYER2_FLAG_VLAN << 8; | ||
4238 | hdr->hdr.l2.vlan_id = vlan_tx_tag_get(skb); | ||
4239 | } | ||
4240 | #endif | ||
4241 | } | ||
4242 | |||
4243 | void | ||
4244 | qeth_fill_header(struct qeth_card *card, struct qeth_hdr *hdr, | ||
4245 | struct sk_buff *skb, int ipv, int cast_type) | ||
4246 | { | ||
4247 | QETH_DBF_TEXT(trace, 6, "fillhdr"); | ||
4248 | |||
4249 | memset(hdr, 0, sizeof(struct qeth_hdr)); | ||
4250 | if (card->options.layer2) { | ||
4251 | qeth_layer2_fill_header(card, hdr, skb, cast_type); | ||
4252 | return; | ||
4253 | } | ||
4254 | hdr->hdr.l3.id = QETH_HEADER_TYPE_LAYER3; | ||
4255 | hdr->hdr.l3.ext_flags = 0; | ||
4256 | #ifdef CONFIG_QETH_VLAN | ||
4257 | /* | ||
4258 | * before we're going to overwrite this location with next hop ip. | ||
4259 | * v6 uses passthrough, v4 sets the tag in the QDIO header. | ||
4260 | */ | ||
4261 | if (card->vlangrp && vlan_tx_tag_present(skb)) { | ||
4262 | hdr->hdr.l3.ext_flags = (ipv == 4) ? | ||
4263 | QETH_HDR_EXT_VLAN_FRAME : | ||
4264 | QETH_HDR_EXT_INCLUDE_VLAN_TAG; | ||
4265 | hdr->hdr.l3.vlan_id = vlan_tx_tag_get(skb); | ||
4266 | } | ||
4267 | #endif /* CONFIG_QETH_VLAN */ | ||
4268 | hdr->hdr.l3.length = skb->len - sizeof(struct qeth_hdr); | ||
4269 | if (ipv == 4) { /* IPv4 */ | ||
4270 | hdr->hdr.l3.flags = qeth_get_qeth_hdr_flags4(cast_type); | ||
4271 | memset(hdr->hdr.l3.dest_addr, 0, 12); | ||
4272 | if ((skb->dst) && (skb->dst->neighbour)) { | ||
4273 | *((u32 *) (&hdr->hdr.l3.dest_addr[12])) = | ||
4274 | *((u32 *) skb->dst->neighbour->primary_key); | ||
4275 | } else { | ||
4276 | /* fill in destination address used in ip header */ | ||
4277 | *((u32 *)(&hdr->hdr.l3.dest_addr[12])) = | ||
4278 | ip_hdr(skb)->daddr; | ||
4279 | } | ||
4280 | } else if (ipv == 6) { /* IPv6 or passthru */ | ||
4281 | hdr->hdr.l3.flags = qeth_get_qeth_hdr_flags6(cast_type); | ||
4282 | if ((skb->dst) && (skb->dst->neighbour)) { | ||
4283 | memcpy(hdr->hdr.l3.dest_addr, | ||
4284 | skb->dst->neighbour->primary_key, 16); | ||
4285 | } else { | ||
4286 | /* fill in destination address used in ip header */ | ||
4287 | memcpy(hdr->hdr.l3.dest_addr, | ||
4288 | &ipv6_hdr(skb)->daddr, 16); | ||
4289 | } | ||
4290 | } else { /* passthrough */ | ||
4291 | if((skb->dev->type == ARPHRD_IEEE802_TR) && | ||
4292 | !memcmp(skb->data + sizeof(struct qeth_hdr) + | ||
4293 | sizeof(__u16), skb->dev->broadcast, 6)) { | ||
4294 | hdr->hdr.l3.flags = QETH_CAST_BROADCAST | | ||
4295 | QETH_HDR_PASSTHRU; | ||
4296 | } else if (!memcmp(skb->data + sizeof(struct qeth_hdr), | ||
4297 | skb->dev->broadcast, 6)) { /* broadcast? */ | ||
4298 | hdr->hdr.l3.flags = QETH_CAST_BROADCAST | | ||
4299 | QETH_HDR_PASSTHRU; | ||
4300 | } else { | ||
4301 | hdr->hdr.l3.flags = (cast_type == RTN_MULTICAST) ? | ||
4302 | QETH_CAST_MULTICAST | QETH_HDR_PASSTHRU : | ||
4303 | QETH_CAST_UNICAST | QETH_HDR_PASSTHRU; | ||
4304 | } | ||
4305 | } | ||
4306 | } | ||
4307 | |||
4308 | static void | ||
4309 | __qeth_fill_buffer(struct sk_buff *skb, struct qdio_buffer *buffer, | ||
4310 | int is_tso, int *next_element_to_fill) | ||
4311 | { | ||
4312 | int length = skb->len; | ||
4313 | int length_here; | ||
4314 | int element; | ||
4315 | char *data; | ||
4316 | int first_lap ; | ||
4317 | |||
4318 | element = *next_element_to_fill; | ||
4319 | data = skb->data; | ||
4320 | first_lap = (is_tso == 0 ? 1 : 0); | ||
4321 | |||
4322 | while (length > 0) { | ||
4323 | /* length_here is the remaining amount of data in this page */ | ||
4324 | length_here = PAGE_SIZE - ((unsigned long) data % PAGE_SIZE); | ||
4325 | if (length < length_here) | ||
4326 | length_here = length; | ||
4327 | |||
4328 | buffer->element[element].addr = data; | ||
4329 | buffer->element[element].length = length_here; | ||
4330 | length -= length_here; | ||
4331 | if (!length) { | ||
4332 | if (first_lap) | ||
4333 | buffer->element[element].flags = 0; | ||
4334 | else | ||
4335 | buffer->element[element].flags = | ||
4336 | SBAL_FLAGS_LAST_FRAG; | ||
4337 | } else { | ||
4338 | if (first_lap) | ||
4339 | buffer->element[element].flags = | ||
4340 | SBAL_FLAGS_FIRST_FRAG; | ||
4341 | else | ||
4342 | buffer->element[element].flags = | ||
4343 | SBAL_FLAGS_MIDDLE_FRAG; | ||
4344 | } | ||
4345 | data += length_here; | ||
4346 | element++; | ||
4347 | first_lap = 0; | ||
4348 | } | ||
4349 | *next_element_to_fill = element; | ||
4350 | } | ||
4351 | |||
4352 | static int | ||
4353 | qeth_fill_buffer(struct qeth_qdio_out_q *queue, | ||
4354 | struct qeth_qdio_out_buffer *buf, | ||
4355 | struct sk_buff *skb) | ||
4356 | { | ||
4357 | struct qdio_buffer *buffer; | ||
4358 | struct qeth_hdr_tso *hdr; | ||
4359 | int flush_cnt = 0, hdr_len, large_send = 0; | ||
4360 | |||
4361 | QETH_DBF_TEXT(trace, 6, "qdfillbf"); | ||
4362 | |||
4363 | buffer = buf->buffer; | ||
4364 | atomic_inc(&skb->users); | ||
4365 | skb_queue_tail(&buf->skb_list, skb); | ||
4366 | |||
4367 | hdr = (struct qeth_hdr_tso *) skb->data; | ||
4368 | /*check first on TSO ....*/ | ||
4369 | if (hdr->hdr.hdr.l3.id == QETH_HEADER_TYPE_TSO) { | ||
4370 | int element = buf->next_element_to_fill; | ||
4371 | |||
4372 | hdr_len = sizeof(struct qeth_hdr_tso) + hdr->ext.dg_hdr_len; | ||
4373 | /*fill first buffer entry only with header information */ | ||
4374 | buffer->element[element].addr = skb->data; | ||
4375 | buffer->element[element].length = hdr_len; | ||
4376 | buffer->element[element].flags = SBAL_FLAGS_FIRST_FRAG; | ||
4377 | buf->next_element_to_fill++; | ||
4378 | skb->data += hdr_len; | ||
4379 | skb->len -= hdr_len; | ||
4380 | large_send = 1; | ||
4381 | } | ||
4382 | if (skb_shinfo(skb)->nr_frags == 0) | ||
4383 | __qeth_fill_buffer(skb, buffer, large_send, | ||
4384 | (int *)&buf->next_element_to_fill); | ||
4385 | else | ||
4386 | __qeth_fill_buffer_frag(skb, buffer, large_send, | ||
4387 | (int *)&buf->next_element_to_fill); | ||
4388 | |||
4389 | if (!queue->do_pack) { | ||
4390 | QETH_DBF_TEXT(trace, 6, "fillbfnp"); | ||
4391 | /* set state to PRIMED -> will be flushed */ | ||
4392 | atomic_set(&buf->state, QETH_QDIO_BUF_PRIMED); | ||
4393 | flush_cnt = 1; | ||
4394 | } else { | ||
4395 | QETH_DBF_TEXT(trace, 6, "fillbfpa"); | ||
4396 | if (queue->card->options.performance_stats) | ||
4397 | queue->card->perf_stats.skbs_sent_pack++; | ||
4398 | if (buf->next_element_to_fill >= | ||
4399 | QETH_MAX_BUFFER_ELEMENTS(queue->card)) { | ||
4400 | /* | ||
4401 | * packed buffer if full -> set state PRIMED | ||
4402 | * -> will be flushed | ||
4403 | */ | ||
4404 | atomic_set(&buf->state, QETH_QDIO_BUF_PRIMED); | ||
4405 | flush_cnt = 1; | ||
4406 | } | ||
4407 | } | ||
4408 | return flush_cnt; | ||
4409 | } | ||
4410 | |||
4411 | static int | ||
4412 | qeth_do_send_packet_fast(struct qeth_card *card, struct qeth_qdio_out_q *queue, | ||
4413 | struct sk_buff *skb, struct qeth_hdr *hdr, | ||
4414 | int elements_needed, | ||
4415 | struct qeth_eddp_context *ctx) | ||
4416 | { | ||
4417 | struct qeth_qdio_out_buffer *buffer; | ||
4418 | int buffers_needed = 0; | ||
4419 | int flush_cnt = 0; | ||
4420 | int index; | ||
4421 | |||
4422 | QETH_DBF_TEXT(trace, 6, "dosndpfa"); | ||
4423 | |||
4424 | /* spin until we get the queue ... */ | ||
4425 | while (atomic_cmpxchg(&queue->state, QETH_OUT_Q_UNLOCKED, | ||
4426 | QETH_OUT_Q_LOCKED) != QETH_OUT_Q_UNLOCKED); | ||
4427 | /* ... now we've got the queue */ | ||
4428 | index = queue->next_buf_to_fill; | ||
4429 | buffer = &queue->bufs[queue->next_buf_to_fill]; | ||
4430 | /* | ||
4431 | * check if buffer is empty to make sure that we do not 'overtake' | ||
4432 | * ourselves and try to fill a buffer that is already primed | ||
4433 | */ | ||
4434 | if (atomic_read(&buffer->state) != QETH_QDIO_BUF_EMPTY) | ||
4435 | goto out; | ||
4436 | if (ctx == NULL) | ||
4437 | queue->next_buf_to_fill = (queue->next_buf_to_fill + 1) % | ||
4438 | QDIO_MAX_BUFFERS_PER_Q; | ||
4439 | else { | ||
4440 | buffers_needed = qeth_eddp_check_buffers_for_context(queue,ctx); | ||
4441 | if (buffers_needed < 0) | ||
4442 | goto out; | ||
4443 | queue->next_buf_to_fill = | ||
4444 | (queue->next_buf_to_fill + buffers_needed) % | ||
4445 | QDIO_MAX_BUFFERS_PER_Q; | ||
4446 | } | ||
4447 | atomic_set(&queue->state, QETH_OUT_Q_UNLOCKED); | ||
4448 | if (ctx == NULL) { | ||
4449 | qeth_fill_buffer(queue, buffer, skb); | ||
4450 | qeth_flush_buffers(queue, 0, index, 1); | ||
4451 | } else { | ||
4452 | flush_cnt = qeth_eddp_fill_buffer(queue, ctx, index); | ||
4453 | WARN_ON(buffers_needed != flush_cnt); | ||
4454 | qeth_flush_buffers(queue, 0, index, flush_cnt); | ||
4455 | } | ||
4456 | return 0; | ||
4457 | out: | ||
4458 | atomic_set(&queue->state, QETH_OUT_Q_UNLOCKED); | ||
4459 | return -EBUSY; | ||
4460 | } | ||
4461 | |||
4462 | static int | ||
4463 | qeth_do_send_packet(struct qeth_card *card, struct qeth_qdio_out_q *queue, | ||
4464 | struct sk_buff *skb, struct qeth_hdr *hdr, | ||
4465 | int elements_needed, struct qeth_eddp_context *ctx) | ||
4466 | { | ||
4467 | struct qeth_qdio_out_buffer *buffer; | ||
4468 | int start_index; | ||
4469 | int flush_count = 0; | ||
4470 | int do_pack = 0; | ||
4471 | int tmp; | ||
4472 | int rc = 0; | ||
4473 | |||
4474 | QETH_DBF_TEXT(trace, 6, "dosndpkt"); | ||
4475 | |||
4476 | /* spin until we get the queue ... */ | ||
4477 | while (atomic_cmpxchg(&queue->state, QETH_OUT_Q_UNLOCKED, | ||
4478 | QETH_OUT_Q_LOCKED) != QETH_OUT_Q_UNLOCKED); | ||
4479 | start_index = queue->next_buf_to_fill; | ||
4480 | buffer = &queue->bufs[queue->next_buf_to_fill]; | ||
4481 | /* | ||
4482 | * check if buffer is empty to make sure that we do not 'overtake' | ||
4483 | * ourselves and try to fill a buffer that is already primed | ||
4484 | */ | ||
4485 | if (atomic_read(&buffer->state) != QETH_QDIO_BUF_EMPTY) { | ||
4486 | atomic_set(&queue->state, QETH_OUT_Q_UNLOCKED); | ||
4487 | return -EBUSY; | ||
4488 | } | ||
4489 | /* check if we need to switch packing state of this queue */ | ||
4490 | qeth_switch_to_packing_if_needed(queue); | ||
4491 | if (queue->do_pack){ | ||
4492 | do_pack = 1; | ||
4493 | if (ctx == NULL) { | ||
4494 | /* does packet fit in current buffer? */ | ||
4495 | if((QETH_MAX_BUFFER_ELEMENTS(card) - | ||
4496 | buffer->next_element_to_fill) < elements_needed){ | ||
4497 | /* ... no -> set state PRIMED */ | ||
4498 | atomic_set(&buffer->state,QETH_QDIO_BUF_PRIMED); | ||
4499 | flush_count++; | ||
4500 | queue->next_buf_to_fill = | ||
4501 | (queue->next_buf_to_fill + 1) % | ||
4502 | QDIO_MAX_BUFFERS_PER_Q; | ||
4503 | buffer = &queue->bufs[queue->next_buf_to_fill]; | ||
4504 | /* we did a step forward, so check buffer state | ||
4505 | * again */ | ||
4506 | if (atomic_read(&buffer->state) != | ||
4507 | QETH_QDIO_BUF_EMPTY){ | ||
4508 | qeth_flush_buffers(queue, 0, start_index, flush_count); | ||
4509 | atomic_set(&queue->state, QETH_OUT_Q_UNLOCKED); | ||
4510 | return -EBUSY; | ||
4511 | } | ||
4512 | } | ||
4513 | } else { | ||
4514 | /* check if we have enough elements (including following | ||
4515 | * free buffers) to handle eddp context */ | ||
4516 | if (qeth_eddp_check_buffers_for_context(queue,ctx) < 0){ | ||
4517 | if (net_ratelimit()) | ||
4518 | PRINT_WARN("eddp tx_dropped 1\n"); | ||
4519 | rc = -EBUSY; | ||
4520 | goto out; | ||
4521 | } | ||
4522 | } | ||
4523 | } | ||
4524 | if (ctx == NULL) | ||
4525 | tmp = qeth_fill_buffer(queue, buffer, skb); | ||
4526 | else { | ||
4527 | tmp = qeth_eddp_fill_buffer(queue,ctx,queue->next_buf_to_fill); | ||
4528 | if (tmp < 0) { | ||
4529 | printk("eddp tx_dropped 2\n"); | ||
4530 | rc = - EBUSY; | ||
4531 | goto out; | ||
4532 | } | ||
4533 | } | ||
4534 | queue->next_buf_to_fill = (queue->next_buf_to_fill + tmp) % | ||
4535 | QDIO_MAX_BUFFERS_PER_Q; | ||
4536 | flush_count += tmp; | ||
4537 | out: | ||
4538 | if (flush_count) | ||
4539 | qeth_flush_buffers(queue, 0, start_index, flush_count); | ||
4540 | else if (!atomic_read(&queue->set_pci_flags_count)) | ||
4541 | atomic_xchg(&queue->state, QETH_OUT_Q_LOCKED_FLUSH); | ||
4542 | /* | ||
4543 | * queue->state will go from LOCKED -> UNLOCKED or from | ||
4544 | * LOCKED_FLUSH -> LOCKED if output_handler wanted to 'notify' us | ||
4545 | * (switch packing state or flush buffer to get another pci flag out). | ||
4546 | * In that case we will enter this loop | ||
4547 | */ | ||
4548 | while (atomic_dec_return(&queue->state)){ | ||
4549 | flush_count = 0; | ||
4550 | start_index = queue->next_buf_to_fill; | ||
4551 | /* check if we can go back to non-packing state */ | ||
4552 | flush_count += qeth_switch_to_nonpacking_if_needed(queue); | ||
4553 | /* | ||
4554 | * check if we need to flush a packing buffer to get a pci | ||
4555 | * flag out on the queue | ||
4556 | */ | ||
4557 | if (!flush_count && !atomic_read(&queue->set_pci_flags_count)) | ||
4558 | flush_count += qeth_flush_buffers_on_no_pci(queue); | ||
4559 | if (flush_count) | ||
4560 | qeth_flush_buffers(queue, 0, start_index, flush_count); | ||
4561 | } | ||
4562 | /* at this point the queue is UNLOCKED again */ | ||
4563 | if (queue->card->options.performance_stats && do_pack) | ||
4564 | queue->card->perf_stats.bufs_sent_pack += flush_count; | ||
4565 | |||
4566 | return rc; | ||
4567 | } | ||
4568 | |||
4569 | static int | ||
4570 | qeth_get_elements_no(struct qeth_card *card, void *hdr, | ||
4571 | struct sk_buff *skb, int elems) | ||
4572 | { | ||
4573 | int elements_needed = 0; | ||
4574 | |||
4575 | if (skb_shinfo(skb)->nr_frags > 0) | ||
4576 | elements_needed = (skb_shinfo(skb)->nr_frags + 1); | ||
4577 | if (elements_needed == 0) | ||
4578 | elements_needed = 1 + (((((unsigned long) hdr) % PAGE_SIZE) | ||
4579 | + skb->len) >> PAGE_SHIFT); | ||
4580 | if ((elements_needed + elems) > QETH_MAX_BUFFER_ELEMENTS(card)){ | ||
4581 | PRINT_ERR("Invalid size of IP packet " | ||
4582 | "(Number=%d / Length=%d). Discarded.\n", | ||
4583 | (elements_needed+elems), skb->len); | ||
4584 | return 0; | ||
4585 | } | ||
4586 | return elements_needed; | ||
4587 | } | ||
4588 | |||
4589 | static void qeth_tx_csum(struct sk_buff *skb) | ||
4590 | { | ||
4591 | int tlen; | ||
4592 | |||
4593 | if (skb->protocol == htons(ETH_P_IP)) { | ||
4594 | tlen = ntohs(ip_hdr(skb)->tot_len) - (ip_hdr(skb)->ihl << 2); | ||
4595 | switch (ip_hdr(skb)->protocol) { | ||
4596 | case IPPROTO_TCP: | ||
4597 | tcp_hdr(skb)->check = 0; | ||
4598 | tcp_hdr(skb)->check = csum_tcpudp_magic( | ||
4599 | ip_hdr(skb)->saddr, ip_hdr(skb)->daddr, | ||
4600 | tlen, ip_hdr(skb)->protocol, | ||
4601 | skb_checksum(skb, skb_transport_offset(skb), | ||
4602 | tlen, 0)); | ||
4603 | break; | ||
4604 | case IPPROTO_UDP: | ||
4605 | udp_hdr(skb)->check = 0; | ||
4606 | udp_hdr(skb)->check = csum_tcpudp_magic( | ||
4607 | ip_hdr(skb)->saddr, ip_hdr(skb)->daddr, | ||
4608 | tlen, ip_hdr(skb)->protocol, | ||
4609 | skb_checksum(skb, skb_transport_offset(skb), | ||
4610 | tlen, 0)); | ||
4611 | break; | ||
4612 | } | ||
4613 | } else if (skb->protocol == htons(ETH_P_IPV6)) { | ||
4614 | switch (ipv6_hdr(skb)->nexthdr) { | ||
4615 | case IPPROTO_TCP: | ||
4616 | tcp_hdr(skb)->check = 0; | ||
4617 | tcp_hdr(skb)->check = csum_ipv6_magic( | ||
4618 | &ipv6_hdr(skb)->saddr, &ipv6_hdr(skb)->daddr, | ||
4619 | ipv6_hdr(skb)->payload_len, | ||
4620 | ipv6_hdr(skb)->nexthdr, | ||
4621 | skb_checksum(skb, skb_transport_offset(skb), | ||
4622 | ipv6_hdr(skb)->payload_len, 0)); | ||
4623 | break; | ||
4624 | case IPPROTO_UDP: | ||
4625 | udp_hdr(skb)->check = 0; | ||
4626 | udp_hdr(skb)->check = csum_ipv6_magic( | ||
4627 | &ipv6_hdr(skb)->saddr, &ipv6_hdr(skb)->daddr, | ||
4628 | ipv6_hdr(skb)->payload_len, | ||
4629 | ipv6_hdr(skb)->nexthdr, | ||
4630 | skb_checksum(skb, skb_transport_offset(skb), | ||
4631 | ipv6_hdr(skb)->payload_len, 0)); | ||
4632 | break; | ||
4633 | } | ||
4634 | } | ||
4635 | } | ||
4636 | |||
4637 | static int | ||
4638 | qeth_send_packet(struct qeth_card *card, struct sk_buff *skb) | ||
4639 | { | ||
4640 | int ipv = 0; | ||
4641 | int cast_type; | ||
4642 | struct qeth_qdio_out_q *queue; | ||
4643 | struct qeth_hdr *hdr = NULL; | ||
4644 | int elements_needed = 0; | ||
4645 | enum qeth_large_send_types large_send = QETH_LARGE_SEND_NO; | ||
4646 | struct qeth_eddp_context *ctx = NULL; | ||
4647 | int tx_bytes = skb->len; | ||
4648 | unsigned short nr_frags = skb_shinfo(skb)->nr_frags; | ||
4649 | unsigned short tso_size = skb_shinfo(skb)->gso_size; | ||
4650 | struct sk_buff *new_skb, *new_skb2; | ||
4651 | int rc; | ||
4652 | |||
4653 | QETH_DBF_TEXT(trace, 6, "sendpkt"); | ||
4654 | |||
4655 | new_skb = skb; | ||
4656 | if ((card->info.type == QETH_CARD_TYPE_OSN) && | ||
4657 | (skb->protocol == htons(ETH_P_IPV6))) | ||
4658 | return -EPERM; | ||
4659 | cast_type = qeth_get_cast_type(card, skb); | ||
4660 | if ((cast_type == RTN_BROADCAST) && | ||
4661 | (card->info.broadcast_capable == 0)) | ||
4662 | return -EPERM; | ||
4663 | queue = card->qdio.out_qs | ||
4664 | [qeth_get_priority_queue(card, skb, ipv, cast_type)]; | ||
4665 | if (!card->options.layer2) { | ||
4666 | ipv = qeth_get_ip_version(skb); | ||
4667 | if ((card->dev->header_ops == &qeth_fake_ops) && ipv) { | ||
4668 | new_skb = qeth_pskb_unshare(skb, GFP_ATOMIC); | ||
4669 | if (!new_skb) | ||
4670 | return -ENOMEM; | ||
4671 | if(card->dev->type == ARPHRD_IEEE802_TR){ | ||
4672 | skb_pull(new_skb, QETH_FAKE_LL_LEN_TR); | ||
4673 | } else { | ||
4674 | skb_pull(new_skb, QETH_FAKE_LL_LEN_ETH); | ||
4675 | } | ||
4676 | } | ||
4677 | } | ||
4678 | if (skb_is_gso(skb)) | ||
4679 | large_send = card->options.large_send; | ||
4680 | /* check on OSN device*/ | ||
4681 | if (card->info.type == QETH_CARD_TYPE_OSN) | ||
4682 | hdr = (struct qeth_hdr *)new_skb->data; | ||
4683 | /*are we able to do TSO ? */ | ||
4684 | if ((large_send == QETH_LARGE_SEND_TSO) && | ||
4685 | (cast_type == RTN_UNSPEC)) { | ||
4686 | rc = qeth_tso_prepare_packet(card, new_skb, ipv, cast_type); | ||
4687 | if (rc) { | ||
4688 | __qeth_free_new_skb(skb, new_skb); | ||
4689 | return rc; | ||
4690 | } | ||
4691 | elements_needed++; | ||
4692 | } else if (card->info.type != QETH_CARD_TYPE_OSN) { | ||
4693 | new_skb2 = qeth_prepare_skb(card, new_skb, &hdr, ipv); | ||
4694 | if (!new_skb2) { | ||
4695 | __qeth_free_new_skb(skb, new_skb); | ||
4696 | return -EINVAL; | ||
4697 | } | ||
4698 | if (new_skb != skb) | ||
4699 | __qeth_free_new_skb(new_skb2, new_skb); | ||
4700 | new_skb = new_skb2; | ||
4701 | qeth_fill_header(card, hdr, new_skb, ipv, cast_type); | ||
4702 | } | ||
4703 | if (large_send == QETH_LARGE_SEND_EDDP) { | ||
4704 | ctx = qeth_eddp_create_context(card, new_skb, hdr, | ||
4705 | skb->sk->sk_protocol); | ||
4706 | if (ctx == NULL) { | ||
4707 | __qeth_free_new_skb(skb, new_skb); | ||
4708 | PRINT_WARN("could not create eddp context\n"); | ||
4709 | return -EINVAL; | ||
4710 | } | ||
4711 | } else { | ||
4712 | int elems = qeth_get_elements_no(card,(void*) hdr, new_skb, | ||
4713 | elements_needed); | ||
4714 | if (!elems) { | ||
4715 | __qeth_free_new_skb(skb, new_skb); | ||
4716 | return -EINVAL; | ||
4717 | } | ||
4718 | elements_needed += elems; | ||
4719 | } | ||
4720 | |||
4721 | if ((large_send == QETH_LARGE_SEND_NO) && | ||
4722 | (skb->ip_summed == CHECKSUM_PARTIAL)) | ||
4723 | qeth_tx_csum(new_skb); | ||
4724 | |||
4725 | if (card->info.type != QETH_CARD_TYPE_IQD) | ||
4726 | rc = qeth_do_send_packet(card, queue, new_skb, hdr, | ||
4727 | elements_needed, ctx); | ||
4728 | else { | ||
4729 | if ((!card->options.layer2) && | ||
4730 | (ipv == 0)) { | ||
4731 | __qeth_free_new_skb(skb, new_skb); | ||
4732 | return -EPERM; | ||
4733 | } | ||
4734 | rc = qeth_do_send_packet_fast(card, queue, new_skb, hdr, | ||
4735 | elements_needed, ctx); | ||
4736 | } | ||
4737 | if (!rc) { | ||
4738 | card->stats.tx_packets++; | ||
4739 | card->stats.tx_bytes += tx_bytes; | ||
4740 | if (new_skb != skb) | ||
4741 | dev_kfree_skb_any(skb); | ||
4742 | if (card->options.performance_stats) { | ||
4743 | if (tso_size && | ||
4744 | !(large_send == QETH_LARGE_SEND_NO)) { | ||
4745 | card->perf_stats.large_send_bytes += tx_bytes; | ||
4746 | card->perf_stats.large_send_cnt++; | ||
4747 | } | ||
4748 | if (nr_frags > 0) { | ||
4749 | card->perf_stats.sg_skbs_sent++; | ||
4750 | /* nr_frags + skb->data */ | ||
4751 | card->perf_stats.sg_frags_sent += | ||
4752 | nr_frags + 1; | ||
4753 | } | ||
4754 | } | ||
4755 | } else { | ||
4756 | card->stats.tx_dropped++; | ||
4757 | __qeth_free_new_skb(skb, new_skb); | ||
4758 | } | ||
4759 | if (ctx != NULL) { | ||
4760 | /* drop creator's reference */ | ||
4761 | qeth_eddp_put_context(ctx); | ||
4762 | /* free skb; it's not referenced by a buffer */ | ||
4763 | if (!rc) | ||
4764 | dev_kfree_skb_any(new_skb); | ||
4765 | } | ||
4766 | return rc; | ||
4767 | } | ||
4768 | |||
4769 | static int | ||
4770 | qeth_mdio_read(struct net_device *dev, int phy_id, int regnum) | ||
4771 | { | ||
4772 | struct qeth_card *card = (struct qeth_card *) dev->priv; | ||
4773 | int rc = 0; | ||
4774 | |||
4775 | switch(regnum){ | ||
4776 | case MII_BMCR: /* Basic mode control register */ | ||
4777 | rc = BMCR_FULLDPLX; | ||
4778 | if ((card->info.link_type != QETH_LINK_TYPE_GBIT_ETH)&& | ||
4779 | (card->info.link_type != QETH_LINK_TYPE_OSN) && | ||
4780 | (card->info.link_type != QETH_LINK_TYPE_10GBIT_ETH)) | ||
4781 | rc |= BMCR_SPEED100; | ||
4782 | break; | ||
4783 | case MII_BMSR: /* Basic mode status register */ | ||
4784 | rc = BMSR_ERCAP | BMSR_ANEGCOMPLETE | BMSR_LSTATUS | | ||
4785 | BMSR_10HALF | BMSR_10FULL | BMSR_100HALF | BMSR_100FULL | | ||
4786 | BMSR_100BASE4; | ||
4787 | break; | ||
4788 | case MII_PHYSID1: /* PHYS ID 1 */ | ||
4789 | rc = (dev->dev_addr[0] << 16) | (dev->dev_addr[1] << 8) | | ||
4790 | dev->dev_addr[2]; | ||
4791 | rc = (rc >> 5) & 0xFFFF; | ||
4792 | break; | ||
4793 | case MII_PHYSID2: /* PHYS ID 2 */ | ||
4794 | rc = (dev->dev_addr[2] << 10) & 0xFFFF; | ||
4795 | break; | ||
4796 | case MII_ADVERTISE: /* Advertisement control reg */ | ||
4797 | rc = ADVERTISE_ALL; | ||
4798 | break; | ||
4799 | case MII_LPA: /* Link partner ability reg */ | ||
4800 | rc = LPA_10HALF | LPA_10FULL | LPA_100HALF | LPA_100FULL | | ||
4801 | LPA_100BASE4 | LPA_LPACK; | ||
4802 | break; | ||
4803 | case MII_EXPANSION: /* Expansion register */ | ||
4804 | break; | ||
4805 | case MII_DCOUNTER: /* disconnect counter */ | ||
4806 | break; | ||
4807 | case MII_FCSCOUNTER: /* false carrier counter */ | ||
4808 | break; | ||
4809 | case MII_NWAYTEST: /* N-way auto-neg test register */ | ||
4810 | break; | ||
4811 | case MII_RERRCOUNTER: /* rx error counter */ | ||
4812 | rc = card->stats.rx_errors; | ||
4813 | break; | ||
4814 | case MII_SREVISION: /* silicon revision */ | ||
4815 | break; | ||
4816 | case MII_RESV1: /* reserved 1 */ | ||
4817 | break; | ||
4818 | case MII_LBRERROR: /* loopback, rx, bypass error */ | ||
4819 | break; | ||
4820 | case MII_PHYADDR: /* physical address */ | ||
4821 | break; | ||
4822 | case MII_RESV2: /* reserved 2 */ | ||
4823 | break; | ||
4824 | case MII_TPISTATUS: /* TPI status for 10mbps */ | ||
4825 | break; | ||
4826 | case MII_NCONFIG: /* network interface config */ | ||
4827 | break; | ||
4828 | default: | ||
4829 | break; | ||
4830 | } | ||
4831 | return rc; | ||
4832 | } | ||
4833 | |||
4834 | |||
4835 | static const char * | ||
4836 | qeth_arp_get_error_cause(int *rc) | ||
4837 | { | ||
4838 | switch (*rc) { | ||
4839 | case QETH_IPA_ARP_RC_FAILED: | ||
4840 | *rc = -EIO; | ||
4841 | return "operation failed"; | ||
4842 | case QETH_IPA_ARP_RC_NOTSUPP: | ||
4843 | *rc = -EOPNOTSUPP; | ||
4844 | return "operation not supported"; | ||
4845 | case QETH_IPA_ARP_RC_OUT_OF_RANGE: | ||
4846 | *rc = -EINVAL; | ||
4847 | return "argument out of range"; | ||
4848 | case QETH_IPA_ARP_RC_Q_NOTSUPP: | ||
4849 | *rc = -EOPNOTSUPP; | ||
4850 | return "query operation not supported"; | ||
4851 | case QETH_IPA_ARP_RC_Q_NO_DATA: | ||
4852 | *rc = -ENOENT; | ||
4853 | return "no query data available"; | ||
4854 | default: | ||
4855 | return "unknown error"; | ||
4856 | } | ||
4857 | } | ||
4858 | |||
4859 | static int | ||
4860 | qeth_send_simple_setassparms(struct qeth_card *, enum qeth_ipa_funcs, | ||
4861 | __u16, long); | ||
4862 | |||
4863 | static int | ||
4864 | qeth_arp_set_no_entries(struct qeth_card *card, int no_entries) | ||
4865 | { | ||
4866 | int tmp; | ||
4867 | int rc; | ||
4868 | |||
4869 | QETH_DBF_TEXT(trace,3,"arpstnoe"); | ||
4870 | |||
4871 | /* | ||
4872 | * currently GuestLAN only supports the ARP assist function | ||
4873 | * IPA_CMD_ASS_ARP_QUERY_INFO, but not IPA_CMD_ASS_ARP_SET_NO_ENTRIES; | ||
4874 | * thus we say EOPNOTSUPP for this ARP function | ||
4875 | */ | ||
4876 | if (card->info.guestlan) | ||
4877 | return -EOPNOTSUPP; | ||
4878 | if (!qeth_is_supported(card,IPA_ARP_PROCESSING)) { | ||
4879 | PRINT_WARN("ARP processing not supported " | ||
4880 | "on %s!\n", QETH_CARD_IFNAME(card)); | ||
4881 | return -EOPNOTSUPP; | ||
4882 | } | ||
4883 | rc = qeth_send_simple_setassparms(card, IPA_ARP_PROCESSING, | ||
4884 | IPA_CMD_ASS_ARP_SET_NO_ENTRIES, | ||
4885 | no_entries); | ||
4886 | if (rc) { | ||
4887 | tmp = rc; | ||
4888 | PRINT_WARN("Could not set number of ARP entries on %s: " | ||
4889 | "%s (0x%x/%d)\n", | ||
4890 | QETH_CARD_IFNAME(card), qeth_arp_get_error_cause(&rc), | ||
4891 | tmp, tmp); | ||
4892 | } | ||
4893 | return rc; | ||
4894 | } | ||
4895 | |||
4896 | static void | ||
4897 | qeth_copy_arp_entries_stripped(struct qeth_arp_query_info *qinfo, | ||
4898 | struct qeth_arp_query_data *qdata, | ||
4899 | int entry_size, int uentry_size) | ||
4900 | { | ||
4901 | char *entry_ptr; | ||
4902 | char *uentry_ptr; | ||
4903 | int i; | ||
4904 | |||
4905 | entry_ptr = (char *)&qdata->data; | ||
4906 | uentry_ptr = (char *)(qinfo->udata + qinfo->udata_offset); | ||
4907 | for (i = 0; i < qdata->no_entries; ++i){ | ||
4908 | /* strip off 32 bytes "media specific information" */ | ||
4909 | memcpy(uentry_ptr, (entry_ptr + 32), entry_size - 32); | ||
4910 | entry_ptr += entry_size; | ||
4911 | uentry_ptr += uentry_size; | ||
4912 | } | ||
4913 | } | ||
4914 | |||
4915 | static int | ||
4916 | qeth_arp_query_cb(struct qeth_card *card, struct qeth_reply *reply, | ||
4917 | unsigned long data) | ||
4918 | { | ||
4919 | struct qeth_ipa_cmd *cmd; | ||
4920 | struct qeth_arp_query_data *qdata; | ||
4921 | struct qeth_arp_query_info *qinfo; | ||
4922 | int entry_size; | ||
4923 | int uentry_size; | ||
4924 | int i; | ||
4925 | |||
4926 | QETH_DBF_TEXT(trace,4,"arpquecb"); | ||
4927 | |||
4928 | qinfo = (struct qeth_arp_query_info *) reply->param; | ||
4929 | cmd = (struct qeth_ipa_cmd *) data; | ||
4930 | if (cmd->hdr.return_code) { | ||
4931 | QETH_DBF_TEXT_(trace,4,"qaer1%i", cmd->hdr.return_code); | ||
4932 | return 0; | ||
4933 | } | ||
4934 | if (cmd->data.setassparms.hdr.return_code) { | ||
4935 | cmd->hdr.return_code = cmd->data.setassparms.hdr.return_code; | ||
4936 | QETH_DBF_TEXT_(trace,4,"qaer2%i", cmd->hdr.return_code); | ||
4937 | return 0; | ||
4938 | } | ||
4939 | qdata = &cmd->data.setassparms.data.query_arp; | ||
4940 | switch(qdata->reply_bits){ | ||
4941 | case 5: | ||
4942 | uentry_size = entry_size = sizeof(struct qeth_arp_qi_entry5); | ||
4943 | if (qinfo->mask_bits & QETH_QARP_STRIP_ENTRIES) | ||
4944 | uentry_size = sizeof(struct qeth_arp_qi_entry5_short); | ||
4945 | break; | ||
4946 | case 7: | ||
4947 | /* fall through to default */ | ||
4948 | default: | ||
4949 | /* tr is the same as eth -> entry7 */ | ||
4950 | uentry_size = entry_size = sizeof(struct qeth_arp_qi_entry7); | ||
4951 | if (qinfo->mask_bits & QETH_QARP_STRIP_ENTRIES) | ||
4952 | uentry_size = sizeof(struct qeth_arp_qi_entry7_short); | ||
4953 | break; | ||
4954 | } | ||
4955 | /* check if there is enough room in userspace */ | ||
4956 | if ((qinfo->udata_len - qinfo->udata_offset) < | ||
4957 | qdata->no_entries * uentry_size){ | ||
4958 | QETH_DBF_TEXT_(trace, 4, "qaer3%i", -ENOMEM); | ||
4959 | cmd->hdr.return_code = -ENOMEM; | ||
4960 | PRINT_WARN("query ARP user space buffer is too small for " | ||
4961 | "the returned number of ARP entries. " | ||
4962 | "Aborting query!\n"); | ||
4963 | goto out_error; | ||
4964 | } | ||
4965 | QETH_DBF_TEXT_(trace, 4, "anore%i", | ||
4966 | cmd->data.setassparms.hdr.number_of_replies); | ||
4967 | QETH_DBF_TEXT_(trace, 4, "aseqn%i", cmd->data.setassparms.hdr.seq_no); | ||
4968 | QETH_DBF_TEXT_(trace, 4, "anoen%i", qdata->no_entries); | ||
4969 | |||
4970 | if (qinfo->mask_bits & QETH_QARP_STRIP_ENTRIES) { | ||
4971 | /* strip off "media specific information" */ | ||
4972 | qeth_copy_arp_entries_stripped(qinfo, qdata, entry_size, | ||
4973 | uentry_size); | ||
4974 | } else | ||
4975 | /*copy entries to user buffer*/ | ||
4976 | memcpy(qinfo->udata + qinfo->udata_offset, | ||
4977 | (char *)&qdata->data, qdata->no_entries*uentry_size); | ||
4978 | |||
4979 | qinfo->no_entries += qdata->no_entries; | ||
4980 | qinfo->udata_offset += (qdata->no_entries*uentry_size); | ||
4981 | /* check if all replies received ... */ | ||
4982 | if (cmd->data.setassparms.hdr.seq_no < | ||
4983 | cmd->data.setassparms.hdr.number_of_replies) | ||
4984 | return 1; | ||
4985 | memcpy(qinfo->udata, &qinfo->no_entries, 4); | ||
4986 | /* keep STRIP_ENTRIES flag so the user program can distinguish | ||
4987 | * stripped entries from normal ones */ | ||
4988 | if (qinfo->mask_bits & QETH_QARP_STRIP_ENTRIES) | ||
4989 | qdata->reply_bits |= QETH_QARP_STRIP_ENTRIES; | ||
4990 | memcpy(qinfo->udata + QETH_QARP_MASK_OFFSET,&qdata->reply_bits,2); | ||
4991 | return 0; | ||
4992 | out_error: | ||
4993 | i = 0; | ||
4994 | memcpy(qinfo->udata, &i, 4); | ||
4995 | return 0; | ||
4996 | } | ||
4997 | |||
4998 | static int | ||
4999 | qeth_send_ipa_arp_cmd(struct qeth_card *card, struct qeth_cmd_buffer *iob, | ||
5000 | int len, int (*reply_cb)(struct qeth_card *, | ||
5001 | struct qeth_reply *, | ||
5002 | unsigned long), | ||
5003 | void *reply_param) | ||
5004 | { | ||
5005 | QETH_DBF_TEXT(trace,4,"sendarp"); | ||
5006 | |||
5007 | memcpy(iob->data, IPA_PDU_HEADER, IPA_PDU_HEADER_SIZE); | ||
5008 | memcpy(QETH_IPA_CMD_DEST_ADDR(iob->data), | ||
5009 | &card->token.ulp_connection_r, QETH_MPC_TOKEN_LENGTH); | ||
5010 | return qeth_send_control_data(card, IPA_PDU_HEADER_SIZE + len, iob, | ||
5011 | reply_cb, reply_param); | ||
5012 | } | ||
5013 | |||
5014 | static int | ||
5015 | qeth_send_ipa_snmp_cmd(struct qeth_card *card, struct qeth_cmd_buffer *iob, | ||
5016 | int len, int (*reply_cb)(struct qeth_card *, | ||
5017 | struct qeth_reply *, | ||
5018 | unsigned long), | ||
5019 | void *reply_param) | ||
5020 | { | ||
5021 | u16 s1, s2; | ||
5022 | |||
5023 | QETH_DBF_TEXT(trace,4,"sendsnmp"); | ||
5024 | |||
5025 | memcpy(iob->data, IPA_PDU_HEADER, IPA_PDU_HEADER_SIZE); | ||
5026 | memcpy(QETH_IPA_CMD_DEST_ADDR(iob->data), | ||
5027 | &card->token.ulp_connection_r, QETH_MPC_TOKEN_LENGTH); | ||
5028 | /* adjust PDU length fields in IPA_PDU_HEADER */ | ||
5029 | s1 = (u32) IPA_PDU_HEADER_SIZE + len; | ||
5030 | s2 = (u32) len; | ||
5031 | memcpy(QETH_IPA_PDU_LEN_TOTAL(iob->data), &s1, 2); | ||
5032 | memcpy(QETH_IPA_PDU_LEN_PDU1(iob->data), &s2, 2); | ||
5033 | memcpy(QETH_IPA_PDU_LEN_PDU2(iob->data), &s2, 2); | ||
5034 | memcpy(QETH_IPA_PDU_LEN_PDU3(iob->data), &s2, 2); | ||
5035 | return qeth_send_control_data(card, IPA_PDU_HEADER_SIZE + len, iob, | ||
5036 | reply_cb, reply_param); | ||
5037 | } | ||
5038 | |||
5039 | static struct qeth_cmd_buffer * | ||
5040 | qeth_get_setassparms_cmd(struct qeth_card *, enum qeth_ipa_funcs, | ||
5041 | __u16, __u16, enum qeth_prot_versions); | ||
5042 | static int | ||
5043 | qeth_arp_query(struct qeth_card *card, char __user *udata) | ||
5044 | { | ||
5045 | struct qeth_cmd_buffer *iob; | ||
5046 | struct qeth_arp_query_info qinfo = {0, }; | ||
5047 | int tmp; | ||
5048 | int rc; | ||
5049 | |||
5050 | QETH_DBF_TEXT(trace,3,"arpquery"); | ||
5051 | |||
5052 | if (!qeth_is_supported(card,/*IPA_QUERY_ARP_ADDR_INFO*/ | ||
5053 | IPA_ARP_PROCESSING)) { | ||
5054 | PRINT_WARN("ARP processing not supported " | ||
5055 | "on %s!\n", QETH_CARD_IFNAME(card)); | ||
5056 | return -EOPNOTSUPP; | ||
5057 | } | ||
5058 | /* get size of userspace buffer and mask_bits -> 6 bytes */ | ||
5059 | if (copy_from_user(&qinfo, udata, 6)) | ||
5060 | return -EFAULT; | ||
5061 | if (!(qinfo.udata = kzalloc(qinfo.udata_len, GFP_KERNEL))) | ||
5062 | return -ENOMEM; | ||
5063 | qinfo.udata_offset = QETH_QARP_ENTRIES_OFFSET; | ||
5064 | iob = qeth_get_setassparms_cmd(card, IPA_ARP_PROCESSING, | ||
5065 | IPA_CMD_ASS_ARP_QUERY_INFO, | ||
5066 | sizeof(int),QETH_PROT_IPV4); | ||
5067 | |||
5068 | rc = qeth_send_ipa_arp_cmd(card, iob, | ||
5069 | QETH_SETASS_BASE_LEN+QETH_ARP_CMD_LEN, | ||
5070 | qeth_arp_query_cb, (void *)&qinfo); | ||
5071 | if (rc) { | ||
5072 | tmp = rc; | ||
5073 | PRINT_WARN("Error while querying ARP cache on %s: %s " | ||
5074 | "(0x%x/%d)\n", | ||
5075 | QETH_CARD_IFNAME(card), qeth_arp_get_error_cause(&rc), | ||
5076 | tmp, tmp); | ||
5077 | if (copy_to_user(udata, qinfo.udata, 4)) | ||
5078 | rc = -EFAULT; | ||
5079 | } else { | ||
5080 | if (copy_to_user(udata, qinfo.udata, qinfo.udata_len)) | ||
5081 | rc = -EFAULT; | ||
5082 | } | ||
5083 | kfree(qinfo.udata); | ||
5084 | return rc; | ||
5085 | } | ||
5086 | |||
5087 | /** | ||
5088 | * SNMP command callback | ||
5089 | */ | ||
5090 | static int | ||
5091 | qeth_snmp_command_cb(struct qeth_card *card, struct qeth_reply *reply, | ||
5092 | unsigned long sdata) | ||
5093 | { | ||
5094 | struct qeth_ipa_cmd *cmd; | ||
5095 | struct qeth_arp_query_info *qinfo; | ||
5096 | struct qeth_snmp_cmd *snmp; | ||
5097 | unsigned char *data; | ||
5098 | __u16 data_len; | ||
5099 | |||
5100 | QETH_DBF_TEXT(trace,3,"snpcmdcb"); | ||
5101 | |||
5102 | cmd = (struct qeth_ipa_cmd *) sdata; | ||
5103 | data = (unsigned char *)((char *)cmd - reply->offset); | ||
5104 | qinfo = (struct qeth_arp_query_info *) reply->param; | ||
5105 | snmp = &cmd->data.setadapterparms.data.snmp; | ||
5106 | |||
5107 | if (cmd->hdr.return_code) { | ||
5108 | QETH_DBF_TEXT_(trace,4,"scer1%i", cmd->hdr.return_code); | ||
5109 | return 0; | ||
5110 | } | ||
5111 | if (cmd->data.setadapterparms.hdr.return_code) { | ||
5112 | cmd->hdr.return_code = cmd->data.setadapterparms.hdr.return_code; | ||
5113 | QETH_DBF_TEXT_(trace,4,"scer2%i", cmd->hdr.return_code); | ||
5114 | return 0; | ||
5115 | } | ||
5116 | data_len = *((__u16*)QETH_IPA_PDU_LEN_PDU1(data)); | ||
5117 | if (cmd->data.setadapterparms.hdr.seq_no == 1) | ||
5118 | data_len -= (__u16)((char *)&snmp->data - (char *)cmd); | ||
5119 | else | ||
5120 | data_len -= (__u16)((char*)&snmp->request - (char *)cmd); | ||
5121 | |||
5122 | /* check if there is enough room in userspace */ | ||
5123 | if ((qinfo->udata_len - qinfo->udata_offset) < data_len) { | ||
5124 | QETH_DBF_TEXT_(trace, 4, "scer3%i", -ENOMEM); | ||
5125 | cmd->hdr.return_code = -ENOMEM; | ||
5126 | return 0; | ||
5127 | } | ||
5128 | QETH_DBF_TEXT_(trace, 4, "snore%i", | ||
5129 | cmd->data.setadapterparms.hdr.used_total); | ||
5130 | QETH_DBF_TEXT_(trace, 4, "sseqn%i", cmd->data.setadapterparms.hdr.seq_no); | ||
5131 | /*copy entries to user buffer*/ | ||
5132 | if (cmd->data.setadapterparms.hdr.seq_no == 1) { | ||
5133 | memcpy(qinfo->udata + qinfo->udata_offset, | ||
5134 | (char *)snmp, | ||
5135 | data_len + offsetof(struct qeth_snmp_cmd,data)); | ||
5136 | qinfo->udata_offset += offsetof(struct qeth_snmp_cmd, data); | ||
5137 | } else { | ||
5138 | memcpy(qinfo->udata + qinfo->udata_offset, | ||
5139 | (char *)&snmp->request, data_len); | ||
5140 | } | ||
5141 | qinfo->udata_offset += data_len; | ||
5142 | /* check if all replies received ... */ | ||
5143 | QETH_DBF_TEXT_(trace, 4, "srtot%i", | ||
5144 | cmd->data.setadapterparms.hdr.used_total); | ||
5145 | QETH_DBF_TEXT_(trace, 4, "srseq%i", | ||
5146 | cmd->data.setadapterparms.hdr.seq_no); | ||
5147 | if (cmd->data.setadapterparms.hdr.seq_no < | ||
5148 | cmd->data.setadapterparms.hdr.used_total) | ||
5149 | return 1; | ||
5150 | return 0; | ||
5151 | } | ||
5152 | |||
5153 | static struct qeth_cmd_buffer * | ||
5154 | qeth_get_ipacmd_buffer(struct qeth_card *, enum qeth_ipa_cmds, | ||
5155 | enum qeth_prot_versions ); | ||
5156 | |||
5157 | static struct qeth_cmd_buffer * | ||
5158 | qeth_get_adapter_cmd(struct qeth_card *card, __u32 command, __u32 cmdlen) | ||
5159 | { | ||
5160 | struct qeth_cmd_buffer *iob; | ||
5161 | struct qeth_ipa_cmd *cmd; | ||
5162 | |||
5163 | iob = qeth_get_ipacmd_buffer(card,IPA_CMD_SETADAPTERPARMS, | ||
5164 | QETH_PROT_IPV4); | ||
5165 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
5166 | cmd->data.setadapterparms.hdr.cmdlength = cmdlen; | ||
5167 | cmd->data.setadapterparms.hdr.command_code = command; | ||
5168 | cmd->data.setadapterparms.hdr.used_total = 1; | ||
5169 | cmd->data.setadapterparms.hdr.seq_no = 1; | ||
5170 | |||
5171 | return iob; | ||
5172 | } | ||
5173 | |||
5174 | /** | ||
5175 | * function to send SNMP commands to OSA-E card | ||
5176 | */ | ||
5177 | static int | ||
5178 | qeth_snmp_command(struct qeth_card *card, char __user *udata) | ||
5179 | { | ||
5180 | struct qeth_cmd_buffer *iob; | ||
5181 | struct qeth_ipa_cmd *cmd; | ||
5182 | struct qeth_snmp_ureq *ureq; | ||
5183 | int req_len; | ||
5184 | struct qeth_arp_query_info qinfo = {0, }; | ||
5185 | int rc = 0; | ||
5186 | |||
5187 | QETH_DBF_TEXT(trace,3,"snmpcmd"); | ||
5188 | |||
5189 | if (card->info.guestlan) | ||
5190 | return -EOPNOTSUPP; | ||
5191 | |||
5192 | if ((!qeth_adp_supported(card,IPA_SETADP_SET_SNMP_CONTROL)) && | ||
5193 | (!card->options.layer2) ) { | ||
5194 | PRINT_WARN("SNMP Query MIBS not supported " | ||
5195 | "on %s!\n", QETH_CARD_IFNAME(card)); | ||
5196 | return -EOPNOTSUPP; | ||
5197 | } | ||
5198 | /* skip 4 bytes (data_len struct member) to get req_len */ | ||
5199 | if (copy_from_user(&req_len, udata + sizeof(int), sizeof(int))) | ||
5200 | return -EFAULT; | ||
5201 | ureq = kmalloc(req_len+sizeof(struct qeth_snmp_ureq_hdr), GFP_KERNEL); | ||
5202 | if (!ureq) { | ||
5203 | QETH_DBF_TEXT(trace, 2, "snmpnome"); | ||
5204 | return -ENOMEM; | ||
5205 | } | ||
5206 | if (copy_from_user(ureq, udata, | ||
5207 | req_len+sizeof(struct qeth_snmp_ureq_hdr))){ | ||
5208 | kfree(ureq); | ||
5209 | return -EFAULT; | ||
5210 | } | ||
5211 | qinfo.udata_len = ureq->hdr.data_len; | ||
5212 | if (!(qinfo.udata = kzalloc(qinfo.udata_len, GFP_KERNEL))){ | ||
5213 | kfree(ureq); | ||
5214 | return -ENOMEM; | ||
5215 | } | ||
5216 | qinfo.udata_offset = sizeof(struct qeth_snmp_ureq_hdr); | ||
5217 | |||
5218 | iob = qeth_get_adapter_cmd(card, IPA_SETADP_SET_SNMP_CONTROL, | ||
5219 | QETH_SNMP_SETADP_CMDLENGTH + req_len); | ||
5220 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
5221 | memcpy(&cmd->data.setadapterparms.data.snmp, &ureq->cmd, req_len); | ||
5222 | rc = qeth_send_ipa_snmp_cmd(card, iob, QETH_SETADP_BASE_LEN + req_len, | ||
5223 | qeth_snmp_command_cb, (void *)&qinfo); | ||
5224 | if (rc) | ||
5225 | PRINT_WARN("SNMP command failed on %s: (0x%x)\n", | ||
5226 | QETH_CARD_IFNAME(card), rc); | ||
5227 | else { | ||
5228 | if (copy_to_user(udata, qinfo.udata, qinfo.udata_len)) | ||
5229 | rc = -EFAULT; | ||
5230 | } | ||
5231 | |||
5232 | kfree(ureq); | ||
5233 | kfree(qinfo.udata); | ||
5234 | return rc; | ||
5235 | } | ||
5236 | |||
5237 | static int | ||
5238 | qeth_default_setassparms_cb(struct qeth_card *, struct qeth_reply *, | ||
5239 | unsigned long); | ||
5240 | |||
5241 | static int | ||
5242 | qeth_default_setadapterparms_cb(struct qeth_card *card, | ||
5243 | struct qeth_reply *reply, | ||
5244 | unsigned long data); | ||
5245 | static int | ||
5246 | qeth_send_setassparms(struct qeth_card *, struct qeth_cmd_buffer *, | ||
5247 | __u16, long, | ||
5248 | int (*reply_cb) | ||
5249 | (struct qeth_card *, struct qeth_reply *, unsigned long), | ||
5250 | void *reply_param); | ||
5251 | |||
5252 | static int | ||
5253 | qeth_arp_add_entry(struct qeth_card *card, struct qeth_arp_cache_entry *entry) | ||
5254 | { | ||
5255 | struct qeth_cmd_buffer *iob; | ||
5256 | char buf[16]; | ||
5257 | int tmp; | ||
5258 | int rc; | ||
5259 | |||
5260 | QETH_DBF_TEXT(trace,3,"arpadent"); | ||
5261 | |||
5262 | /* | ||
5263 | * currently GuestLAN only supports the ARP assist function | ||
5264 | * IPA_CMD_ASS_ARP_QUERY_INFO, but not IPA_CMD_ASS_ARP_ADD_ENTRY; | ||
5265 | * thus we say EOPNOTSUPP for this ARP function | ||
5266 | */ | ||
5267 | if (card->info.guestlan) | ||
5268 | return -EOPNOTSUPP; | ||
5269 | if (!qeth_is_supported(card,IPA_ARP_PROCESSING)) { | ||
5270 | PRINT_WARN("ARP processing not supported " | ||
5271 | "on %s!\n", QETH_CARD_IFNAME(card)); | ||
5272 | return -EOPNOTSUPP; | ||
5273 | } | ||
5274 | |||
5275 | iob = qeth_get_setassparms_cmd(card, IPA_ARP_PROCESSING, | ||
5276 | IPA_CMD_ASS_ARP_ADD_ENTRY, | ||
5277 | sizeof(struct qeth_arp_cache_entry), | ||
5278 | QETH_PROT_IPV4); | ||
5279 | rc = qeth_send_setassparms(card, iob, | ||
5280 | sizeof(struct qeth_arp_cache_entry), | ||
5281 | (unsigned long) entry, | ||
5282 | qeth_default_setassparms_cb, NULL); | ||
5283 | if (rc) { | ||
5284 | tmp = rc; | ||
5285 | qeth_ipaddr4_to_string((u8 *)entry->ipaddr, buf); | ||
5286 | PRINT_WARN("Could not add ARP entry for address %s on %s: " | ||
5287 | "%s (0x%x/%d)\n", | ||
5288 | buf, QETH_CARD_IFNAME(card), | ||
5289 | qeth_arp_get_error_cause(&rc), tmp, tmp); | ||
5290 | } | ||
5291 | return rc; | ||
5292 | } | ||
5293 | |||
5294 | static int | ||
5295 | qeth_arp_remove_entry(struct qeth_card *card, struct qeth_arp_cache_entry *entry) | ||
5296 | { | ||
5297 | struct qeth_cmd_buffer *iob; | ||
5298 | char buf[16] = {0, }; | ||
5299 | int tmp; | ||
5300 | int rc; | ||
5301 | |||
5302 | QETH_DBF_TEXT(trace,3,"arprment"); | ||
5303 | |||
5304 | /* | ||
5305 | * currently GuestLAN only supports the ARP assist function | ||
5306 | * IPA_CMD_ASS_ARP_QUERY_INFO, but not IPA_CMD_ASS_ARP_REMOVE_ENTRY; | ||
5307 | * thus we say EOPNOTSUPP for this ARP function | ||
5308 | */ | ||
5309 | if (card->info.guestlan) | ||
5310 | return -EOPNOTSUPP; | ||
5311 | if (!qeth_is_supported(card,IPA_ARP_PROCESSING)) { | ||
5312 | PRINT_WARN("ARP processing not supported " | ||
5313 | "on %s!\n", QETH_CARD_IFNAME(card)); | ||
5314 | return -EOPNOTSUPP; | ||
5315 | } | ||
5316 | memcpy(buf, entry, 12); | ||
5317 | iob = qeth_get_setassparms_cmd(card, IPA_ARP_PROCESSING, | ||
5318 | IPA_CMD_ASS_ARP_REMOVE_ENTRY, | ||
5319 | 12, | ||
5320 | QETH_PROT_IPV4); | ||
5321 | rc = qeth_send_setassparms(card, iob, | ||
5322 | 12, (unsigned long)buf, | ||
5323 | qeth_default_setassparms_cb, NULL); | ||
5324 | if (rc) { | ||
5325 | tmp = rc; | ||
5326 | memset(buf, 0, 16); | ||
5327 | qeth_ipaddr4_to_string((u8 *)entry->ipaddr, buf); | ||
5328 | PRINT_WARN("Could not delete ARP entry for address %s on %s: " | ||
5329 | "%s (0x%x/%d)\n", | ||
5330 | buf, QETH_CARD_IFNAME(card), | ||
5331 | qeth_arp_get_error_cause(&rc), tmp, tmp); | ||
5332 | } | ||
5333 | return rc; | ||
5334 | } | ||
5335 | |||
5336 | static int | ||
5337 | qeth_arp_flush_cache(struct qeth_card *card) | ||
5338 | { | ||
5339 | int rc; | ||
5340 | int tmp; | ||
5341 | |||
5342 | QETH_DBF_TEXT(trace,3,"arpflush"); | ||
5343 | |||
5344 | /* | ||
5345 | * currently GuestLAN only supports the ARP assist function | ||
5346 | * IPA_CMD_ASS_ARP_QUERY_INFO, but not IPA_CMD_ASS_ARP_FLUSH_CACHE; | ||
5347 | * thus we say EOPNOTSUPP for this ARP function | ||
5348 | */ | ||
5349 | if (card->info.guestlan || (card->info.type == QETH_CARD_TYPE_IQD)) | ||
5350 | return -EOPNOTSUPP; | ||
5351 | if (!qeth_is_supported(card,IPA_ARP_PROCESSING)) { | ||
5352 | PRINT_WARN("ARP processing not supported " | ||
5353 | "on %s!\n", QETH_CARD_IFNAME(card)); | ||
5354 | return -EOPNOTSUPP; | ||
5355 | } | ||
5356 | rc = qeth_send_simple_setassparms(card, IPA_ARP_PROCESSING, | ||
5357 | IPA_CMD_ASS_ARP_FLUSH_CACHE, 0); | ||
5358 | if (rc){ | ||
5359 | tmp = rc; | ||
5360 | PRINT_WARN("Could not flush ARP cache on %s: %s (0x%x/%d)\n", | ||
5361 | QETH_CARD_IFNAME(card), qeth_arp_get_error_cause(&rc), | ||
5362 | tmp, tmp); | ||
5363 | } | ||
5364 | return rc; | ||
5365 | } | ||
5366 | |||
5367 | static int | ||
5368 | qeth_do_ioctl(struct net_device *dev, struct ifreq *rq, int cmd) | ||
5369 | { | ||
5370 | struct qeth_card *card = (struct qeth_card *)dev->priv; | ||
5371 | struct qeth_arp_cache_entry arp_entry; | ||
5372 | struct mii_ioctl_data *mii_data; | ||
5373 | int rc = 0; | ||
5374 | |||
5375 | if (!card) | ||
5376 | return -ENODEV; | ||
5377 | |||
5378 | if ((card->state != CARD_STATE_UP) && | ||
5379 | (card->state != CARD_STATE_SOFTSETUP)) | ||
5380 | return -ENODEV; | ||
5381 | |||
5382 | if (card->info.type == QETH_CARD_TYPE_OSN) | ||
5383 | return -EPERM; | ||
5384 | |||
5385 | switch (cmd){ | ||
5386 | case SIOC_QETH_ARP_SET_NO_ENTRIES: | ||
5387 | if ( !capable(CAP_NET_ADMIN) || | ||
5388 | (card->options.layer2) ) { | ||
5389 | rc = -EPERM; | ||
5390 | break; | ||
5391 | } | ||
5392 | rc = qeth_arp_set_no_entries(card, rq->ifr_ifru.ifru_ivalue); | ||
5393 | break; | ||
5394 | case SIOC_QETH_ARP_QUERY_INFO: | ||
5395 | if ( !capable(CAP_NET_ADMIN) || | ||
5396 | (card->options.layer2) ) { | ||
5397 | rc = -EPERM; | ||
5398 | break; | ||
5399 | } | ||
5400 | rc = qeth_arp_query(card, rq->ifr_ifru.ifru_data); | ||
5401 | break; | ||
5402 | case SIOC_QETH_ARP_ADD_ENTRY: | ||
5403 | if ( !capable(CAP_NET_ADMIN) || | ||
5404 | (card->options.layer2) ) { | ||
5405 | rc = -EPERM; | ||
5406 | break; | ||
5407 | } | ||
5408 | if (copy_from_user(&arp_entry, rq->ifr_ifru.ifru_data, | ||
5409 | sizeof(struct qeth_arp_cache_entry))) | ||
5410 | rc = -EFAULT; | ||
5411 | else | ||
5412 | rc = qeth_arp_add_entry(card, &arp_entry); | ||
5413 | break; | ||
5414 | case SIOC_QETH_ARP_REMOVE_ENTRY: | ||
5415 | if ( !capable(CAP_NET_ADMIN) || | ||
5416 | (card->options.layer2) ) { | ||
5417 | rc = -EPERM; | ||
5418 | break; | ||
5419 | } | ||
5420 | if (copy_from_user(&arp_entry, rq->ifr_ifru.ifru_data, | ||
5421 | sizeof(struct qeth_arp_cache_entry))) | ||
5422 | rc = -EFAULT; | ||
5423 | else | ||
5424 | rc = qeth_arp_remove_entry(card, &arp_entry); | ||
5425 | break; | ||
5426 | case SIOC_QETH_ARP_FLUSH_CACHE: | ||
5427 | if ( !capable(CAP_NET_ADMIN) || | ||
5428 | (card->options.layer2) ) { | ||
5429 | rc = -EPERM; | ||
5430 | break; | ||
5431 | } | ||
5432 | rc = qeth_arp_flush_cache(card); | ||
5433 | break; | ||
5434 | case SIOC_QETH_ADP_SET_SNMP_CONTROL: | ||
5435 | rc = qeth_snmp_command(card, rq->ifr_ifru.ifru_data); | ||
5436 | break; | ||
5437 | case SIOC_QETH_GET_CARD_TYPE: | ||
5438 | if ((card->info.type == QETH_CARD_TYPE_OSAE) && | ||
5439 | !card->info.guestlan) | ||
5440 | return 1; | ||
5441 | return 0; | ||
5442 | break; | ||
5443 | case SIOCGMIIPHY: | ||
5444 | mii_data = if_mii(rq); | ||
5445 | mii_data->phy_id = 0; | ||
5446 | break; | ||
5447 | case SIOCGMIIREG: | ||
5448 | mii_data = if_mii(rq); | ||
5449 | if (mii_data->phy_id != 0) | ||
5450 | rc = -EINVAL; | ||
5451 | else | ||
5452 | mii_data->val_out = qeth_mdio_read(dev,mii_data->phy_id, | ||
5453 | mii_data->reg_num); | ||
5454 | break; | ||
5455 | default: | ||
5456 | rc = -EOPNOTSUPP; | ||
5457 | } | ||
5458 | if (rc) | ||
5459 | QETH_DBF_TEXT_(trace, 2, "ioce%d", rc); | ||
5460 | return rc; | ||
5461 | } | ||
5462 | |||
5463 | static struct net_device_stats * | ||
5464 | qeth_get_stats(struct net_device *dev) | ||
5465 | { | ||
5466 | struct qeth_card *card; | ||
5467 | |||
5468 | card = (struct qeth_card *) (dev->priv); | ||
5469 | |||
5470 | QETH_DBF_TEXT(trace,5,"getstat"); | ||
5471 | |||
5472 | return &card->stats; | ||
5473 | } | ||
5474 | |||
5475 | static int | ||
5476 | qeth_change_mtu(struct net_device *dev, int new_mtu) | ||
5477 | { | ||
5478 | struct qeth_card *card; | ||
5479 | char dbf_text[15]; | ||
5480 | |||
5481 | card = (struct qeth_card *) (dev->priv); | ||
5482 | |||
5483 | QETH_DBF_TEXT(trace,4,"chgmtu"); | ||
5484 | sprintf(dbf_text, "%8x", new_mtu); | ||
5485 | QETH_DBF_TEXT(trace,4,dbf_text); | ||
5486 | |||
5487 | if (new_mtu < 64) | ||
5488 | return -EINVAL; | ||
5489 | if (new_mtu > 65535) | ||
5490 | return -EINVAL; | ||
5491 | if ((!qeth_is_supported(card,IPA_IP_FRAGMENTATION)) && | ||
5492 | (!qeth_mtu_is_valid(card, new_mtu))) | ||
5493 | return -EINVAL; | ||
5494 | dev->mtu = new_mtu; | ||
5495 | return 0; | ||
5496 | } | ||
5497 | |||
5498 | #ifdef CONFIG_QETH_VLAN | ||
5499 | static void | ||
5500 | qeth_vlan_rx_register(struct net_device *dev, struct vlan_group *grp) | ||
5501 | { | ||
5502 | struct qeth_card *card; | ||
5503 | unsigned long flags; | ||
5504 | |||
5505 | QETH_DBF_TEXT(trace,4,"vlanreg"); | ||
5506 | |||
5507 | card = (struct qeth_card *) dev->priv; | ||
5508 | spin_lock_irqsave(&card->vlanlock, flags); | ||
5509 | card->vlangrp = grp; | ||
5510 | spin_unlock_irqrestore(&card->vlanlock, flags); | ||
5511 | } | ||
5512 | |||
5513 | static void | ||
5514 | qeth_free_vlan_buffer(struct qeth_card *card, struct qeth_qdio_out_buffer *buf, | ||
5515 | unsigned short vid) | ||
5516 | { | ||
5517 | int i; | ||
5518 | struct sk_buff *skb; | ||
5519 | struct sk_buff_head tmp_list; | ||
5520 | |||
5521 | skb_queue_head_init(&tmp_list); | ||
5522 | lockdep_set_class(&tmp_list.lock, &qdio_out_skb_queue_key); | ||
5523 | for(i = 0; i < QETH_MAX_BUFFER_ELEMENTS(card); ++i){ | ||
5524 | while ((skb = skb_dequeue(&buf->skb_list))){ | ||
5525 | if (vlan_tx_tag_present(skb) && | ||
5526 | (vlan_tx_tag_get(skb) == vid)) { | ||
5527 | atomic_dec(&skb->users); | ||
5528 | dev_kfree_skb(skb); | ||
5529 | } else | ||
5530 | skb_queue_tail(&tmp_list, skb); | ||
5531 | } | ||
5532 | } | ||
5533 | while ((skb = skb_dequeue(&tmp_list))) | ||
5534 | skb_queue_tail(&buf->skb_list, skb); | ||
5535 | } | ||
5536 | |||
5537 | static void | ||
5538 | qeth_free_vlan_skbs(struct qeth_card *card, unsigned short vid) | ||
5539 | { | ||
5540 | int i, j; | ||
5541 | |||
5542 | QETH_DBF_TEXT(trace, 4, "frvlskbs"); | ||
5543 | for (i = 0; i < card->qdio.no_out_queues; ++i){ | ||
5544 | for (j = 0; j < QDIO_MAX_BUFFERS_PER_Q; ++j) | ||
5545 | qeth_free_vlan_buffer(card, &card->qdio. | ||
5546 | out_qs[i]->bufs[j], vid); | ||
5547 | } | ||
5548 | } | ||
5549 | |||
5550 | static void | ||
5551 | qeth_free_vlan_addresses4(struct qeth_card *card, unsigned short vid) | ||
5552 | { | ||
5553 | struct in_device *in_dev; | ||
5554 | struct in_ifaddr *ifa; | ||
5555 | struct qeth_ipaddr *addr; | ||
5556 | |||
5557 | QETH_DBF_TEXT(trace, 4, "frvaddr4"); | ||
5558 | |||
5559 | rcu_read_lock(); | ||
5560 | in_dev = __in_dev_get_rcu(vlan_group_get_device(card->vlangrp, vid)); | ||
5561 | if (!in_dev) | ||
5562 | goto out; | ||
5563 | for (ifa = in_dev->ifa_list; ifa; ifa = ifa->ifa_next) { | ||
5564 | addr = qeth_get_addr_buffer(QETH_PROT_IPV4); | ||
5565 | if (addr){ | ||
5566 | addr->u.a4.addr = ifa->ifa_address; | ||
5567 | addr->u.a4.mask = ifa->ifa_mask; | ||
5568 | addr->type = QETH_IP_TYPE_NORMAL; | ||
5569 | if (!qeth_delete_ip(card, addr)) | ||
5570 | kfree(addr); | ||
5571 | } | ||
5572 | } | ||
5573 | out: | ||
5574 | rcu_read_unlock(); | ||
5575 | } | ||
5576 | |||
5577 | static void | ||
5578 | qeth_free_vlan_addresses6(struct qeth_card *card, unsigned short vid) | ||
5579 | { | ||
5580 | #ifdef CONFIG_QETH_IPV6 | ||
5581 | struct inet6_dev *in6_dev; | ||
5582 | struct inet6_ifaddr *ifa; | ||
5583 | struct qeth_ipaddr *addr; | ||
5584 | |||
5585 | QETH_DBF_TEXT(trace, 4, "frvaddr6"); | ||
5586 | |||
5587 | in6_dev = in6_dev_get(vlan_group_get_device(card->vlangrp, vid)); | ||
5588 | if (!in6_dev) | ||
5589 | return; | ||
5590 | for (ifa = in6_dev->addr_list; ifa; ifa = ifa->lst_next){ | ||
5591 | addr = qeth_get_addr_buffer(QETH_PROT_IPV6); | ||
5592 | if (addr){ | ||
5593 | memcpy(&addr->u.a6.addr, &ifa->addr, | ||
5594 | sizeof(struct in6_addr)); | ||
5595 | addr->u.a6.pfxlen = ifa->prefix_len; | ||
5596 | addr->type = QETH_IP_TYPE_NORMAL; | ||
5597 | if (!qeth_delete_ip(card, addr)) | ||
5598 | kfree(addr); | ||
5599 | } | ||
5600 | } | ||
5601 | in6_dev_put(in6_dev); | ||
5602 | #endif /* CONFIG_QETH_IPV6 */ | ||
5603 | } | ||
5604 | |||
5605 | static void | ||
5606 | qeth_free_vlan_addresses(struct qeth_card *card, unsigned short vid) | ||
5607 | { | ||
5608 | if (card->options.layer2 || !card->vlangrp) | ||
5609 | return; | ||
5610 | qeth_free_vlan_addresses4(card, vid); | ||
5611 | qeth_free_vlan_addresses6(card, vid); | ||
5612 | } | ||
5613 | |||
5614 | static int | ||
5615 | qeth_layer2_send_setdelvlan_cb(struct qeth_card *card, | ||
5616 | struct qeth_reply *reply, | ||
5617 | unsigned long data) | ||
5618 | { | ||
5619 | struct qeth_ipa_cmd *cmd; | ||
5620 | |||
5621 | QETH_DBF_TEXT(trace, 2, "L2sdvcb"); | ||
5622 | cmd = (struct qeth_ipa_cmd *) data; | ||
5623 | if (cmd->hdr.return_code) { | ||
5624 | PRINT_ERR("Error in processing VLAN %i on %s: 0x%x. " | ||
5625 | "Continuing\n",cmd->data.setdelvlan.vlan_id, | ||
5626 | QETH_CARD_IFNAME(card), cmd->hdr.return_code); | ||
5627 | QETH_DBF_TEXT_(trace, 2, "L2VL%4x", cmd->hdr.command); | ||
5628 | QETH_DBF_TEXT_(trace, 2, "L2%s", CARD_BUS_ID(card)); | ||
5629 | QETH_DBF_TEXT_(trace, 2, "err%d", cmd->hdr.return_code); | ||
5630 | } | ||
5631 | return 0; | ||
5632 | } | ||
5633 | |||
5634 | static int | ||
5635 | qeth_layer2_send_setdelvlan(struct qeth_card *card, __u16 i, | ||
5636 | enum qeth_ipa_cmds ipacmd) | ||
5637 | { | ||
5638 | struct qeth_ipa_cmd *cmd; | ||
5639 | struct qeth_cmd_buffer *iob; | ||
5640 | |||
5641 | QETH_DBF_TEXT_(trace, 4, "L2sdv%x",ipacmd); | ||
5642 | iob = qeth_get_ipacmd_buffer(card, ipacmd, QETH_PROT_IPV4); | ||
5643 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
5644 | cmd->data.setdelvlan.vlan_id = i; | ||
5645 | return qeth_send_ipa_cmd(card, iob, | ||
5646 | qeth_layer2_send_setdelvlan_cb, NULL); | ||
5647 | } | ||
5648 | |||
5649 | static void | ||
5650 | qeth_layer2_process_vlans(struct qeth_card *card, int clear) | ||
5651 | { | ||
5652 | unsigned short i; | ||
5653 | |||
5654 | QETH_DBF_TEXT(trace, 3, "L2prcvln"); | ||
5655 | |||
5656 | if (!card->vlangrp) | ||
5657 | return; | ||
5658 | for (i = 0; i < VLAN_GROUP_ARRAY_LEN; i++) { | ||
5659 | if (vlan_group_get_device(card->vlangrp, i) == NULL) | ||
5660 | continue; | ||
5661 | if (clear) | ||
5662 | qeth_layer2_send_setdelvlan(card, i, IPA_CMD_DELVLAN); | ||
5663 | else | ||
5664 | qeth_layer2_send_setdelvlan(card, i, IPA_CMD_SETVLAN); | ||
5665 | } | ||
5666 | } | ||
5667 | |||
5668 | /*add_vid is layer 2 used only ....*/ | ||
5669 | static void | ||
5670 | qeth_vlan_rx_add_vid(struct net_device *dev, unsigned short vid) | ||
5671 | { | ||
5672 | struct qeth_card *card; | ||
5673 | |||
5674 | QETH_DBF_TEXT_(trace, 4, "aid:%d", vid); | ||
5675 | |||
5676 | card = (struct qeth_card *) dev->priv; | ||
5677 | if (!card->options.layer2) | ||
5678 | return; | ||
5679 | qeth_layer2_send_setdelvlan(card, vid, IPA_CMD_SETVLAN); | ||
5680 | } | ||
5681 | |||
5682 | /*... kill_vid used for both modes*/ | ||
5683 | static void | ||
5684 | qeth_vlan_rx_kill_vid(struct net_device *dev, unsigned short vid) | ||
5685 | { | ||
5686 | struct qeth_card *card; | ||
5687 | unsigned long flags; | ||
5688 | |||
5689 | QETH_DBF_TEXT_(trace, 4, "kid:%d", vid); | ||
5690 | |||
5691 | card = (struct qeth_card *) dev->priv; | ||
5692 | /* free all skbs for the vlan device */ | ||
5693 | qeth_free_vlan_skbs(card, vid); | ||
5694 | spin_lock_irqsave(&card->vlanlock, flags); | ||
5695 | /* unregister IP addresses of vlan device */ | ||
5696 | qeth_free_vlan_addresses(card, vid); | ||
5697 | vlan_group_set_device(card->vlangrp, vid, NULL); | ||
5698 | spin_unlock_irqrestore(&card->vlanlock, flags); | ||
5699 | if (card->options.layer2) | ||
5700 | qeth_layer2_send_setdelvlan(card, vid, IPA_CMD_DELVLAN); | ||
5701 | qeth_set_multicast_list(card->dev); | ||
5702 | } | ||
5703 | #endif | ||
5704 | /** | ||
5705 | * Examine hardware response to SET_PROMISC_MODE | ||
5706 | */ | ||
5707 | static int | ||
5708 | qeth_setadp_promisc_mode_cb(struct qeth_card *card, | ||
5709 | struct qeth_reply *reply, | ||
5710 | unsigned long data) | ||
5711 | { | ||
5712 | struct qeth_ipa_cmd *cmd; | ||
5713 | struct qeth_ipacmd_setadpparms *setparms; | ||
5714 | |||
5715 | QETH_DBF_TEXT(trace,4,"prmadpcb"); | ||
5716 | |||
5717 | cmd = (struct qeth_ipa_cmd *) data; | ||
5718 | setparms = &(cmd->data.setadapterparms); | ||
5719 | |||
5720 | qeth_default_setadapterparms_cb(card, reply, (unsigned long)cmd); | ||
5721 | if (cmd->hdr.return_code) { | ||
5722 | QETH_DBF_TEXT_(trace,4,"prmrc%2.2x",cmd->hdr.return_code); | ||
5723 | setparms->data.mode = SET_PROMISC_MODE_OFF; | ||
5724 | } | ||
5725 | card->info.promisc_mode = setparms->data.mode; | ||
5726 | return 0; | ||
5727 | } | ||
5728 | /* | ||
5729 | * Set promiscuous mode (on or off) (SET_PROMISC_MODE command) | ||
5730 | */ | ||
5731 | static void | ||
5732 | qeth_setadp_promisc_mode(struct qeth_card *card) | ||
5733 | { | ||
5734 | enum qeth_ipa_promisc_modes mode; | ||
5735 | struct net_device *dev = card->dev; | ||
5736 | struct qeth_cmd_buffer *iob; | ||
5737 | struct qeth_ipa_cmd *cmd; | ||
5738 | |||
5739 | QETH_DBF_TEXT(trace, 4, "setprom"); | ||
5740 | |||
5741 | if (((dev->flags & IFF_PROMISC) && | ||
5742 | (card->info.promisc_mode == SET_PROMISC_MODE_ON)) || | ||
5743 | (!(dev->flags & IFF_PROMISC) && | ||
5744 | (card->info.promisc_mode == SET_PROMISC_MODE_OFF))) | ||
5745 | return; | ||
5746 | mode = SET_PROMISC_MODE_OFF; | ||
5747 | if (dev->flags & IFF_PROMISC) | ||
5748 | mode = SET_PROMISC_MODE_ON; | ||
5749 | QETH_DBF_TEXT_(trace, 4, "mode:%x", mode); | ||
5750 | |||
5751 | iob = qeth_get_adapter_cmd(card, IPA_SETADP_SET_PROMISC_MODE, | ||
5752 | sizeof(struct qeth_ipacmd_setadpparms)); | ||
5753 | cmd = (struct qeth_ipa_cmd *)(iob->data + IPA_PDU_HEADER_SIZE); | ||
5754 | cmd->data.setadapterparms.data.mode = mode; | ||
5755 | qeth_send_ipa_cmd(card, iob, qeth_setadp_promisc_mode_cb, NULL); | ||
5756 | } | ||
5757 | |||
5758 | /** | ||
5759 | * set multicast address on card | ||
5760 | */ | ||
5761 | static void | ||
5762 | qeth_set_multicast_list(struct net_device *dev) | ||
5763 | { | ||
5764 | struct qeth_card *card = (struct qeth_card *) dev->priv; | ||
5765 | |||
5766 | if (card->info.type == QETH_CARD_TYPE_OSN) | ||
5767 | return ; | ||
5768 | |||
5769 | QETH_DBF_TEXT(trace, 3, "setmulti"); | ||
5770 | qeth_delete_mc_addresses(card); | ||
5771 | if (card->options.layer2) { | ||
5772 | qeth_layer2_add_multicast(card); | ||
5773 | goto out; | ||
5774 | } | ||
5775 | qeth_add_multicast_ipv4(card); | ||
5776 | #ifdef CONFIG_QETH_IPV6 | ||
5777 | qeth_add_multicast_ipv6(card); | ||
5778 | #endif | ||
5779 | out: | ||
5780 | qeth_set_ip_addr_list(card); | ||
5781 | if (!qeth_adp_supported(card, IPA_SETADP_SET_PROMISC_MODE)) | ||
5782 | return; | ||
5783 | qeth_setadp_promisc_mode(card); | ||
5784 | } | ||
5785 | |||
5786 | static int | ||
5787 | qeth_neigh_setup(struct net_device *dev, struct neigh_parms *np) | ||
5788 | { | ||
5789 | return 0; | ||
5790 | } | ||
5791 | |||
5792 | static void | ||
5793 | qeth_get_mac_for_ipm(__u32 ipm, char *mac, struct net_device *dev) | ||
5794 | { | ||
5795 | if (dev->type == ARPHRD_IEEE802_TR) | ||
5796 | ip_tr_mc_map(ipm, mac); | ||
5797 | else | ||
5798 | ip_eth_mc_map(ipm, mac); | ||
5799 | } | ||
5800 | |||
5801 | static struct qeth_ipaddr * | ||
5802 | qeth_get_addr_buffer(enum qeth_prot_versions prot) | ||
5803 | { | ||
5804 | struct qeth_ipaddr *addr; | ||
5805 | |||
5806 | addr = kzalloc(sizeof(struct qeth_ipaddr), GFP_ATOMIC); | ||
5807 | if (addr == NULL) { | ||
5808 | PRINT_WARN("Not enough memory to add address\n"); | ||
5809 | return NULL; | ||
5810 | } | ||
5811 | addr->type = QETH_IP_TYPE_NORMAL; | ||
5812 | addr->proto = prot; | ||
5813 | return addr; | ||
5814 | } | ||
5815 | |||
5816 | int | ||
5817 | qeth_osn_assist(struct net_device *dev, | ||
5818 | void *data, | ||
5819 | int data_len) | ||
5820 | { | ||
5821 | struct qeth_cmd_buffer *iob; | ||
5822 | struct qeth_card *card; | ||
5823 | int rc; | ||
5824 | |||
5825 | QETH_DBF_TEXT(trace, 2, "osnsdmc"); | ||
5826 | if (!dev) | ||
5827 | return -ENODEV; | ||
5828 | card = (struct qeth_card *)dev->priv; | ||
5829 | if (!card) | ||
5830 | return -ENODEV; | ||
5831 | if ((card->state != CARD_STATE_UP) && | ||
5832 | (card->state != CARD_STATE_SOFTSETUP)) | ||
5833 | return -ENODEV; | ||
5834 | iob = qeth_wait_for_buffer(&card->write); | ||
5835 | memcpy(iob->data+IPA_PDU_HEADER_SIZE, data, data_len); | ||
5836 | rc = qeth_osn_send_ipa_cmd(card, iob, data_len); | ||
5837 | return rc; | ||
5838 | } | ||
5839 | |||
5840 | static struct net_device * | ||
5841 | qeth_netdev_by_devno(unsigned char *read_dev_no) | ||
5842 | { | ||
5843 | struct qeth_card *card; | ||
5844 | struct net_device *ndev; | ||
5845 | unsigned char *readno; | ||
5846 | __u16 temp_dev_no, card_dev_no; | ||
5847 | char *endp; | ||
5848 | unsigned long flags; | ||
5849 | |||
5850 | ndev = NULL; | ||
5851 | memcpy(&temp_dev_no, read_dev_no, 2); | ||
5852 | read_lock_irqsave(&qeth_card_list.rwlock, flags); | ||
5853 | list_for_each_entry(card, &qeth_card_list.list, list) { | ||
5854 | readno = CARD_RDEV_ID(card); | ||
5855 | readno += (strlen(readno) - 4); | ||
5856 | card_dev_no = simple_strtoul(readno, &endp, 16); | ||
5857 | if (card_dev_no == temp_dev_no) { | ||
5858 | ndev = card->dev; | ||
5859 | break; | ||
5860 | } | ||
5861 | } | ||
5862 | read_unlock_irqrestore(&qeth_card_list.rwlock, flags); | ||
5863 | return ndev; | ||
5864 | } | ||
5865 | |||
5866 | int | ||
5867 | qeth_osn_register(unsigned char *read_dev_no, | ||
5868 | struct net_device **dev, | ||
5869 | int (*assist_cb)(struct net_device *, void *), | ||
5870 | int (*data_cb)(struct sk_buff *)) | ||
5871 | { | ||
5872 | struct qeth_card * card; | ||
5873 | |||
5874 | QETH_DBF_TEXT(trace, 2, "osnreg"); | ||
5875 | *dev = qeth_netdev_by_devno(read_dev_no); | ||
5876 | if (*dev == NULL) | ||
5877 | return -ENODEV; | ||
5878 | card = (struct qeth_card *)(*dev)->priv; | ||
5879 | if (!card) | ||
5880 | return -ENODEV; | ||
5881 | if ((assist_cb == NULL) || (data_cb == NULL)) | ||
5882 | return -EINVAL; | ||
5883 | card->osn_info.assist_cb = assist_cb; | ||
5884 | card->osn_info.data_cb = data_cb; | ||
5885 | return 0; | ||
5886 | } | ||
5887 | |||
5888 | void | ||
5889 | qeth_osn_deregister(struct net_device * dev) | ||
5890 | { | ||
5891 | struct qeth_card *card; | ||
5892 | |||
5893 | QETH_DBF_TEXT(trace, 2, "osndereg"); | ||
5894 | if (!dev) | ||
5895 | return; | ||
5896 | card = (struct qeth_card *)dev->priv; | ||
5897 | if (!card) | ||
5898 | return; | ||
5899 | card->osn_info.assist_cb = NULL; | ||
5900 | card->osn_info.data_cb = NULL; | ||
5901 | return; | ||
5902 | } | ||
5903 | |||
5904 | static void | ||
5905 | qeth_delete_mc_addresses(struct qeth_card *card) | ||
5906 | { | ||
5907 | struct qeth_ipaddr *iptodo; | ||
5908 | unsigned long flags; | ||
5909 | |||
5910 | QETH_DBF_TEXT(trace,4,"delmc"); | ||
5911 | iptodo = qeth_get_addr_buffer(QETH_PROT_IPV4); | ||
5912 | if (!iptodo) { | ||
5913 | QETH_DBF_TEXT(trace, 2, "dmcnomem"); | ||
5914 | return; | ||
5915 | } | ||
5916 | iptodo->type = QETH_IP_TYPE_DEL_ALL_MC; | ||
5917 | spin_lock_irqsave(&card->ip_lock, flags); | ||
5918 | if (!__qeth_insert_ip_todo(card, iptodo, 0)) | ||
5919 | kfree(iptodo); | ||
5920 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
5921 | } | ||
5922 | |||
5923 | static void | ||
5924 | qeth_add_mc(struct qeth_card *card, struct in_device *in4_dev) | ||
5925 | { | ||
5926 | struct qeth_ipaddr *ipm; | ||
5927 | struct ip_mc_list *im4; | ||
5928 | char buf[MAX_ADDR_LEN]; | ||
5929 | |||
5930 | QETH_DBF_TEXT(trace,4,"addmc"); | ||
5931 | for (im4 = in4_dev->mc_list; im4; im4 = im4->next) { | ||
5932 | qeth_get_mac_for_ipm(im4->multiaddr, buf, in4_dev->dev); | ||
5933 | ipm = qeth_get_addr_buffer(QETH_PROT_IPV4); | ||
5934 | if (!ipm) | ||
5935 | continue; | ||
5936 | ipm->u.a4.addr = im4->multiaddr; | ||
5937 | memcpy(ipm->mac,buf,OSA_ADDR_LEN); | ||
5938 | ipm->is_multicast = 1; | ||
5939 | if (!qeth_add_ip(card,ipm)) | ||
5940 | kfree(ipm); | ||
5941 | } | ||
5942 | } | ||
5943 | |||
5944 | static inline void | ||
5945 | qeth_add_vlan_mc(struct qeth_card *card) | ||
5946 | { | ||
5947 | #ifdef CONFIG_QETH_VLAN | ||
5948 | struct in_device *in_dev; | ||
5949 | struct vlan_group *vg; | ||
5950 | int i; | ||
5951 | |||
5952 | QETH_DBF_TEXT(trace,4,"addmcvl"); | ||
5953 | if ( ((card->options.layer2 == 0) && | ||
5954 | (!qeth_is_supported(card,IPA_FULL_VLAN))) || | ||
5955 | (card->vlangrp == NULL) ) | ||
5956 | return ; | ||
5957 | |||
5958 | vg = card->vlangrp; | ||
5959 | for (i = 0; i < VLAN_GROUP_ARRAY_LEN; i++) { | ||
5960 | struct net_device *netdev = vlan_group_get_device(vg, i); | ||
5961 | if (netdev == NULL || | ||
5962 | !(netdev->flags & IFF_UP)) | ||
5963 | continue; | ||
5964 | in_dev = in_dev_get(netdev); | ||
5965 | if (!in_dev) | ||
5966 | continue; | ||
5967 | read_lock(&in_dev->mc_list_lock); | ||
5968 | qeth_add_mc(card,in_dev); | ||
5969 | read_unlock(&in_dev->mc_list_lock); | ||
5970 | in_dev_put(in_dev); | ||
5971 | } | ||
5972 | #endif | ||
5973 | } | ||
5974 | |||
5975 | static void | ||
5976 | qeth_add_multicast_ipv4(struct qeth_card *card) | ||
5977 | { | ||
5978 | struct in_device *in4_dev; | ||
5979 | |||
5980 | QETH_DBF_TEXT(trace,4,"chkmcv4"); | ||
5981 | in4_dev = in_dev_get(card->dev); | ||
5982 | if (in4_dev == NULL) | ||
5983 | return; | ||
5984 | read_lock(&in4_dev->mc_list_lock); | ||
5985 | qeth_add_mc(card, in4_dev); | ||
5986 | qeth_add_vlan_mc(card); | ||
5987 | read_unlock(&in4_dev->mc_list_lock); | ||
5988 | in_dev_put(in4_dev); | ||
5989 | } | ||
5990 | |||
5991 | static void | ||
5992 | qeth_layer2_add_multicast(struct qeth_card *card) | ||
5993 | { | ||
5994 | struct qeth_ipaddr *ipm; | ||
5995 | struct dev_mc_list *dm; | ||
5996 | |||
5997 | QETH_DBF_TEXT(trace,4,"L2addmc"); | ||
5998 | for (dm = card->dev->mc_list; dm; dm = dm->next) { | ||
5999 | ipm = qeth_get_addr_buffer(QETH_PROT_IPV4); | ||
6000 | if (!ipm) | ||
6001 | continue; | ||
6002 | memcpy(ipm->mac,dm->dmi_addr,MAX_ADDR_LEN); | ||
6003 | ipm->is_multicast = 1; | ||
6004 | if (!qeth_add_ip(card, ipm)) | ||
6005 | kfree(ipm); | ||
6006 | } | ||
6007 | } | ||
6008 | |||
6009 | #ifdef CONFIG_QETH_IPV6 | ||
6010 | static void | ||
6011 | qeth_add_mc6(struct qeth_card *card, struct inet6_dev *in6_dev) | ||
6012 | { | ||
6013 | struct qeth_ipaddr *ipm; | ||
6014 | struct ifmcaddr6 *im6; | ||
6015 | char buf[MAX_ADDR_LEN]; | ||
6016 | |||
6017 | QETH_DBF_TEXT(trace,4,"addmc6"); | ||
6018 | for (im6 = in6_dev->mc_list; im6 != NULL; im6 = im6->next) { | ||
6019 | ndisc_mc_map(&im6->mca_addr, buf, in6_dev->dev, 0); | ||
6020 | ipm = qeth_get_addr_buffer(QETH_PROT_IPV6); | ||
6021 | if (!ipm) | ||
6022 | continue; | ||
6023 | ipm->is_multicast = 1; | ||
6024 | memcpy(ipm->mac,buf,OSA_ADDR_LEN); | ||
6025 | memcpy(&ipm->u.a6.addr,&im6->mca_addr.s6_addr, | ||
6026 | sizeof(struct in6_addr)); | ||
6027 | if (!qeth_add_ip(card,ipm)) | ||
6028 | kfree(ipm); | ||
6029 | } | ||
6030 | } | ||
6031 | |||
6032 | static inline void | ||
6033 | qeth_add_vlan_mc6(struct qeth_card *card) | ||
6034 | { | ||
6035 | #ifdef CONFIG_QETH_VLAN | ||
6036 | struct inet6_dev *in_dev; | ||
6037 | struct vlan_group *vg; | ||
6038 | int i; | ||
6039 | |||
6040 | QETH_DBF_TEXT(trace,4,"admc6vl"); | ||
6041 | if ( ((card->options.layer2 == 0) && | ||
6042 | (!qeth_is_supported(card,IPA_FULL_VLAN))) || | ||
6043 | (card->vlangrp == NULL)) | ||
6044 | return ; | ||
6045 | |||
6046 | vg = card->vlangrp; | ||
6047 | for (i = 0; i < VLAN_GROUP_ARRAY_LEN; i++) { | ||
6048 | struct net_device *netdev = vlan_group_get_device(vg, i); | ||
6049 | if (netdev == NULL || | ||
6050 | !(netdev->flags & IFF_UP)) | ||
6051 | continue; | ||
6052 | in_dev = in6_dev_get(netdev); | ||
6053 | if (!in_dev) | ||
6054 | continue; | ||
6055 | read_lock_bh(&in_dev->lock); | ||
6056 | qeth_add_mc6(card,in_dev); | ||
6057 | read_unlock_bh(&in_dev->lock); | ||
6058 | in6_dev_put(in_dev); | ||
6059 | } | ||
6060 | #endif /* CONFIG_QETH_VLAN */ | ||
6061 | } | ||
6062 | |||
6063 | static void | ||
6064 | qeth_add_multicast_ipv6(struct qeth_card *card) | ||
6065 | { | ||
6066 | struct inet6_dev *in6_dev; | ||
6067 | |||
6068 | QETH_DBF_TEXT(trace,4,"chkmcv6"); | ||
6069 | if (!qeth_is_supported(card, IPA_IPV6)) | ||
6070 | return ; | ||
6071 | in6_dev = in6_dev_get(card->dev); | ||
6072 | if (in6_dev == NULL) | ||
6073 | return; | ||
6074 | read_lock_bh(&in6_dev->lock); | ||
6075 | qeth_add_mc6(card, in6_dev); | ||
6076 | qeth_add_vlan_mc6(card); | ||
6077 | read_unlock_bh(&in6_dev->lock); | ||
6078 | in6_dev_put(in6_dev); | ||
6079 | } | ||
6080 | #endif /* CONFIG_QETH_IPV6 */ | ||
6081 | |||
6082 | static int | ||
6083 | qeth_layer2_send_setdelmac(struct qeth_card *card, __u8 *mac, | ||
6084 | enum qeth_ipa_cmds ipacmd, | ||
6085 | int (*reply_cb) (struct qeth_card *, | ||
6086 | struct qeth_reply*, | ||
6087 | unsigned long)) | ||
6088 | { | ||
6089 | struct qeth_ipa_cmd *cmd; | ||
6090 | struct qeth_cmd_buffer *iob; | ||
6091 | |||
6092 | QETH_DBF_TEXT(trace, 2, "L2sdmac"); | ||
6093 | iob = qeth_get_ipacmd_buffer(card, ipacmd, QETH_PROT_IPV4); | ||
6094 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
6095 | cmd->data.setdelmac.mac_length = OSA_ADDR_LEN; | ||
6096 | memcpy(&cmd->data.setdelmac.mac, mac, OSA_ADDR_LEN); | ||
6097 | return qeth_send_ipa_cmd(card, iob, reply_cb, NULL); | ||
6098 | } | ||
6099 | |||
6100 | static int | ||
6101 | qeth_layer2_send_setgroupmac_cb(struct qeth_card *card, | ||
6102 | struct qeth_reply *reply, | ||
6103 | unsigned long data) | ||
6104 | { | ||
6105 | struct qeth_ipa_cmd *cmd; | ||
6106 | __u8 *mac; | ||
6107 | |||
6108 | QETH_DBF_TEXT(trace, 2, "L2Sgmacb"); | ||
6109 | cmd = (struct qeth_ipa_cmd *) data; | ||
6110 | mac = &cmd->data.setdelmac.mac[0]; | ||
6111 | /* MAC already registered, needed in couple/uncouple case */ | ||
6112 | if (cmd->hdr.return_code == 0x2005) { | ||
6113 | PRINT_WARN("Group MAC %02x:%02x:%02x:%02x:%02x:%02x " \ | ||
6114 | "already existing on %s \n", | ||
6115 | mac[0], mac[1], mac[2], mac[3], mac[4], mac[5], | ||
6116 | QETH_CARD_IFNAME(card)); | ||
6117 | cmd->hdr.return_code = 0; | ||
6118 | } | ||
6119 | if (cmd->hdr.return_code) | ||
6120 | PRINT_ERR("Could not set group MAC " \ | ||
6121 | "%02x:%02x:%02x:%02x:%02x:%02x on %s: %x\n", | ||
6122 | mac[0], mac[1], mac[2], mac[3], mac[4], mac[5], | ||
6123 | QETH_CARD_IFNAME(card),cmd->hdr.return_code); | ||
6124 | return 0; | ||
6125 | } | ||
6126 | |||
6127 | static int | ||
6128 | qeth_layer2_send_setgroupmac(struct qeth_card *card, __u8 *mac) | ||
6129 | { | ||
6130 | QETH_DBF_TEXT(trace, 2, "L2Sgmac"); | ||
6131 | return qeth_layer2_send_setdelmac(card, mac, IPA_CMD_SETGMAC, | ||
6132 | qeth_layer2_send_setgroupmac_cb); | ||
6133 | } | ||
6134 | |||
6135 | static int | ||
6136 | qeth_layer2_send_delgroupmac_cb(struct qeth_card *card, | ||
6137 | struct qeth_reply *reply, | ||
6138 | unsigned long data) | ||
6139 | { | ||
6140 | struct qeth_ipa_cmd *cmd; | ||
6141 | __u8 *mac; | ||
6142 | |||
6143 | QETH_DBF_TEXT(trace, 2, "L2Dgmacb"); | ||
6144 | cmd = (struct qeth_ipa_cmd *) data; | ||
6145 | mac = &cmd->data.setdelmac.mac[0]; | ||
6146 | if (cmd->hdr.return_code) | ||
6147 | PRINT_ERR("Could not delete group MAC " \ | ||
6148 | "%02x:%02x:%02x:%02x:%02x:%02x on %s: %x\n", | ||
6149 | mac[0], mac[1], mac[2], mac[3], mac[4], mac[5], | ||
6150 | QETH_CARD_IFNAME(card), cmd->hdr.return_code); | ||
6151 | return 0; | ||
6152 | } | ||
6153 | |||
6154 | static int | ||
6155 | qeth_layer2_send_delgroupmac(struct qeth_card *card, __u8 *mac) | ||
6156 | { | ||
6157 | QETH_DBF_TEXT(trace, 2, "L2Dgmac"); | ||
6158 | return qeth_layer2_send_setdelmac(card, mac, IPA_CMD_DELGMAC, | ||
6159 | qeth_layer2_send_delgroupmac_cb); | ||
6160 | } | ||
6161 | |||
6162 | static int | ||
6163 | qeth_layer2_send_setmac_cb(struct qeth_card *card, | ||
6164 | struct qeth_reply *reply, | ||
6165 | unsigned long data) | ||
6166 | { | ||
6167 | struct qeth_ipa_cmd *cmd; | ||
6168 | |||
6169 | QETH_DBF_TEXT(trace, 2, "L2Smaccb"); | ||
6170 | cmd = (struct qeth_ipa_cmd *) data; | ||
6171 | if (cmd->hdr.return_code) { | ||
6172 | QETH_DBF_TEXT_(trace, 2, "L2er%x", cmd->hdr.return_code); | ||
6173 | card->info.mac_bits &= ~QETH_LAYER2_MAC_REGISTERED; | ||
6174 | cmd->hdr.return_code = -EIO; | ||
6175 | } else { | ||
6176 | card->info.mac_bits |= QETH_LAYER2_MAC_REGISTERED; | ||
6177 | memcpy(card->dev->dev_addr,cmd->data.setdelmac.mac, | ||
6178 | OSA_ADDR_LEN); | ||
6179 | PRINT_INFO("MAC address %2.2x:%2.2x:%2.2x:%2.2x:%2.2x:%2.2x " | ||
6180 | "successfully registered on device %s\n", | ||
6181 | card->dev->dev_addr[0], card->dev->dev_addr[1], | ||
6182 | card->dev->dev_addr[2], card->dev->dev_addr[3], | ||
6183 | card->dev->dev_addr[4], card->dev->dev_addr[5], | ||
6184 | card->dev->name); | ||
6185 | } | ||
6186 | return 0; | ||
6187 | } | ||
6188 | |||
6189 | static int | ||
6190 | qeth_layer2_send_setmac(struct qeth_card *card, __u8 *mac) | ||
6191 | { | ||
6192 | QETH_DBF_TEXT(trace, 2, "L2Setmac"); | ||
6193 | return qeth_layer2_send_setdelmac(card, mac, IPA_CMD_SETVMAC, | ||
6194 | qeth_layer2_send_setmac_cb); | ||
6195 | } | ||
6196 | |||
6197 | static int | ||
6198 | qeth_layer2_send_delmac_cb(struct qeth_card *card, | ||
6199 | struct qeth_reply *reply, | ||
6200 | unsigned long data) | ||
6201 | { | ||
6202 | struct qeth_ipa_cmd *cmd; | ||
6203 | |||
6204 | QETH_DBF_TEXT(trace, 2, "L2Dmaccb"); | ||
6205 | cmd = (struct qeth_ipa_cmd *) data; | ||
6206 | if (cmd->hdr.return_code) { | ||
6207 | QETH_DBF_TEXT_(trace, 2, "err%d", cmd->hdr.return_code); | ||
6208 | cmd->hdr.return_code = -EIO; | ||
6209 | return 0; | ||
6210 | } | ||
6211 | card->info.mac_bits &= ~QETH_LAYER2_MAC_REGISTERED; | ||
6212 | |||
6213 | return 0; | ||
6214 | } | ||
6215 | static int | ||
6216 | qeth_layer2_send_delmac(struct qeth_card *card, __u8 *mac) | ||
6217 | { | ||
6218 | QETH_DBF_TEXT(trace, 2, "L2Delmac"); | ||
6219 | if (!(card->info.mac_bits & QETH_LAYER2_MAC_REGISTERED)) | ||
6220 | return 0; | ||
6221 | return qeth_layer2_send_setdelmac(card, mac, IPA_CMD_DELVMAC, | ||
6222 | qeth_layer2_send_delmac_cb); | ||
6223 | } | ||
6224 | |||
6225 | static int | ||
6226 | qeth_layer2_set_mac_address(struct net_device *dev, void *p) | ||
6227 | { | ||
6228 | struct sockaddr *addr = p; | ||
6229 | struct qeth_card *card; | ||
6230 | int rc = 0; | ||
6231 | |||
6232 | QETH_DBF_TEXT(trace, 3, "setmac"); | ||
6233 | |||
6234 | if (qeth_verify_dev(dev) != QETH_REAL_CARD) { | ||
6235 | QETH_DBF_TEXT(trace, 3, "setmcINV"); | ||
6236 | return -EOPNOTSUPP; | ||
6237 | } | ||
6238 | card = (struct qeth_card *) dev->priv; | ||
6239 | |||
6240 | if (!card->options.layer2) { | ||
6241 | PRINT_WARN("Setting MAC address on %s is not supported " | ||
6242 | "in Layer 3 mode.\n", dev->name); | ||
6243 | QETH_DBF_TEXT(trace, 3, "setmcLY3"); | ||
6244 | return -EOPNOTSUPP; | ||
6245 | } | ||
6246 | if (card->info.type == QETH_CARD_TYPE_OSN) { | ||
6247 | PRINT_WARN("Setting MAC address on %s is not supported.\n", | ||
6248 | dev->name); | ||
6249 | QETH_DBF_TEXT(trace, 3, "setmcOSN"); | ||
6250 | return -EOPNOTSUPP; | ||
6251 | } | ||
6252 | QETH_DBF_TEXT_(trace, 3, "%s", CARD_BUS_ID(card)); | ||
6253 | QETH_DBF_HEX(trace, 3, addr->sa_data, OSA_ADDR_LEN); | ||
6254 | rc = qeth_layer2_send_delmac(card, &card->dev->dev_addr[0]); | ||
6255 | if (!rc) | ||
6256 | rc = qeth_layer2_send_setmac(card, addr->sa_data); | ||
6257 | return rc; | ||
6258 | } | ||
6259 | |||
6260 | static void | ||
6261 | qeth_fill_ipacmd_header(struct qeth_card *card, struct qeth_ipa_cmd *cmd, | ||
6262 | __u8 command, enum qeth_prot_versions prot) | ||
6263 | { | ||
6264 | memset(cmd, 0, sizeof (struct qeth_ipa_cmd)); | ||
6265 | cmd->hdr.command = command; | ||
6266 | cmd->hdr.initiator = IPA_CMD_INITIATOR_HOST; | ||
6267 | cmd->hdr.seqno = card->seqno.ipa; | ||
6268 | cmd->hdr.adapter_type = qeth_get_ipa_adp_type(card->info.link_type); | ||
6269 | cmd->hdr.rel_adapter_no = (__u8) card->info.portno; | ||
6270 | if (card->options.layer2) | ||
6271 | cmd->hdr.prim_version_no = 2; | ||
6272 | else | ||
6273 | cmd->hdr.prim_version_no = 1; | ||
6274 | cmd->hdr.param_count = 1; | ||
6275 | cmd->hdr.prot_version = prot; | ||
6276 | cmd->hdr.ipa_supported = 0; | ||
6277 | cmd->hdr.ipa_enabled = 0; | ||
6278 | } | ||
6279 | |||
6280 | static struct qeth_cmd_buffer * | ||
6281 | qeth_get_ipacmd_buffer(struct qeth_card *card, enum qeth_ipa_cmds ipacmd, | ||
6282 | enum qeth_prot_versions prot) | ||
6283 | { | ||
6284 | struct qeth_cmd_buffer *iob; | ||
6285 | struct qeth_ipa_cmd *cmd; | ||
6286 | |||
6287 | iob = qeth_wait_for_buffer(&card->write); | ||
6288 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
6289 | qeth_fill_ipacmd_header(card, cmd, ipacmd, prot); | ||
6290 | |||
6291 | return iob; | ||
6292 | } | ||
6293 | |||
6294 | static int | ||
6295 | qeth_send_setdelmc(struct qeth_card *card, struct qeth_ipaddr *addr, int ipacmd) | ||
6296 | { | ||
6297 | int rc; | ||
6298 | struct qeth_cmd_buffer *iob; | ||
6299 | struct qeth_ipa_cmd *cmd; | ||
6300 | |||
6301 | QETH_DBF_TEXT(trace,4,"setdelmc"); | ||
6302 | |||
6303 | iob = qeth_get_ipacmd_buffer(card, ipacmd, addr->proto); | ||
6304 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
6305 | memcpy(&cmd->data.setdelipm.mac,addr->mac, OSA_ADDR_LEN); | ||
6306 | if (addr->proto == QETH_PROT_IPV6) | ||
6307 | memcpy(cmd->data.setdelipm.ip6, &addr->u.a6.addr, | ||
6308 | sizeof(struct in6_addr)); | ||
6309 | else | ||
6310 | memcpy(&cmd->data.setdelipm.ip4, &addr->u.a4.addr,4); | ||
6311 | |||
6312 | rc = qeth_send_ipa_cmd(card, iob, NULL, NULL); | ||
6313 | |||
6314 | return rc; | ||
6315 | } | ||
6316 | static void | ||
6317 | qeth_fill_netmask(u8 *netmask, unsigned int len) | ||
6318 | { | ||
6319 | int i,j; | ||
6320 | for (i=0;i<16;i++) { | ||
6321 | j=(len)-(i*8); | ||
6322 | if (j >= 8) | ||
6323 | netmask[i] = 0xff; | ||
6324 | else if (j > 0) | ||
6325 | netmask[i] = (u8)(0xFF00>>j); | ||
6326 | else | ||
6327 | netmask[i] = 0; | ||
6328 | } | ||
6329 | } | ||
6330 | |||
6331 | static int | ||
6332 | qeth_send_setdelip(struct qeth_card *card, struct qeth_ipaddr *addr, | ||
6333 | int ipacmd, unsigned int flags) | ||
6334 | { | ||
6335 | int rc; | ||
6336 | struct qeth_cmd_buffer *iob; | ||
6337 | struct qeth_ipa_cmd *cmd; | ||
6338 | __u8 netmask[16]; | ||
6339 | |||
6340 | QETH_DBF_TEXT(trace,4,"setdelip"); | ||
6341 | QETH_DBF_TEXT_(trace,4,"flags%02X", flags); | ||
6342 | |||
6343 | iob = qeth_get_ipacmd_buffer(card, ipacmd, addr->proto); | ||
6344 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
6345 | if (addr->proto == QETH_PROT_IPV6) { | ||
6346 | memcpy(cmd->data.setdelip6.ip_addr, &addr->u.a6.addr, | ||
6347 | sizeof(struct in6_addr)); | ||
6348 | qeth_fill_netmask(netmask,addr->u.a6.pfxlen); | ||
6349 | memcpy(cmd->data.setdelip6.mask, netmask, | ||
6350 | sizeof(struct in6_addr)); | ||
6351 | cmd->data.setdelip6.flags = flags; | ||
6352 | } else { | ||
6353 | memcpy(cmd->data.setdelip4.ip_addr, &addr->u.a4.addr, 4); | ||
6354 | memcpy(cmd->data.setdelip4.mask, &addr->u.a4.mask, 4); | ||
6355 | cmd->data.setdelip4.flags = flags; | ||
6356 | } | ||
6357 | |||
6358 | rc = qeth_send_ipa_cmd(card, iob, NULL, NULL); | ||
6359 | |||
6360 | return rc; | ||
6361 | } | ||
6362 | |||
6363 | static int | ||
6364 | qeth_layer2_register_addr_entry(struct qeth_card *card, | ||
6365 | struct qeth_ipaddr *addr) | ||
6366 | { | ||
6367 | if (!addr->is_multicast) | ||
6368 | return 0; | ||
6369 | QETH_DBF_TEXT(trace, 2, "setgmac"); | ||
6370 | QETH_DBF_HEX(trace,3,&addr->mac[0],OSA_ADDR_LEN); | ||
6371 | return qeth_layer2_send_setgroupmac(card, &addr->mac[0]); | ||
6372 | } | ||
6373 | |||
6374 | static int | ||
6375 | qeth_layer2_deregister_addr_entry(struct qeth_card *card, | ||
6376 | struct qeth_ipaddr *addr) | ||
6377 | { | ||
6378 | if (!addr->is_multicast) | ||
6379 | return 0; | ||
6380 | QETH_DBF_TEXT(trace, 2, "delgmac"); | ||
6381 | QETH_DBF_HEX(trace,3,&addr->mac[0],OSA_ADDR_LEN); | ||
6382 | return qeth_layer2_send_delgroupmac(card, &addr->mac[0]); | ||
6383 | } | ||
6384 | |||
6385 | static int | ||
6386 | qeth_layer3_register_addr_entry(struct qeth_card *card, | ||
6387 | struct qeth_ipaddr *addr) | ||
6388 | { | ||
6389 | char buf[50]; | ||
6390 | int rc; | ||
6391 | int cnt = 3; | ||
6392 | |||
6393 | if (addr->proto == QETH_PROT_IPV4) { | ||
6394 | QETH_DBF_TEXT(trace, 2,"setaddr4"); | ||
6395 | QETH_DBF_HEX(trace, 3, &addr->u.a4.addr, sizeof(int)); | ||
6396 | } else if (addr->proto == QETH_PROT_IPV6) { | ||
6397 | QETH_DBF_TEXT(trace, 2, "setaddr6"); | ||
6398 | QETH_DBF_HEX(trace,3,&addr->u.a6.addr,8); | ||
6399 | QETH_DBF_HEX(trace,3,((char *)&addr->u.a6.addr)+8,8); | ||
6400 | } else { | ||
6401 | QETH_DBF_TEXT(trace, 2, "setaddr?"); | ||
6402 | QETH_DBF_HEX(trace, 3, addr, sizeof(struct qeth_ipaddr)); | ||
6403 | } | ||
6404 | do { | ||
6405 | if (addr->is_multicast) | ||
6406 | rc = qeth_send_setdelmc(card, addr, IPA_CMD_SETIPM); | ||
6407 | else | ||
6408 | rc = qeth_send_setdelip(card, addr, IPA_CMD_SETIP, | ||
6409 | addr->set_flags); | ||
6410 | if (rc) | ||
6411 | QETH_DBF_TEXT(trace, 2, "failed"); | ||
6412 | } while ((--cnt > 0) && rc); | ||
6413 | if (rc){ | ||
6414 | QETH_DBF_TEXT(trace, 2, "FAILED"); | ||
6415 | qeth_ipaddr_to_string(addr->proto, (u8 *)&addr->u, buf); | ||
6416 | PRINT_WARN("Could not register IP address %s (rc=0x%x/%d)\n", | ||
6417 | buf, rc, rc); | ||
6418 | } | ||
6419 | return rc; | ||
6420 | } | ||
6421 | |||
6422 | static int | ||
6423 | qeth_layer3_deregister_addr_entry(struct qeth_card *card, | ||
6424 | struct qeth_ipaddr *addr) | ||
6425 | { | ||
6426 | //char buf[50]; | ||
6427 | int rc; | ||
6428 | |||
6429 | if (addr->proto == QETH_PROT_IPV4) { | ||
6430 | QETH_DBF_TEXT(trace, 2,"deladdr4"); | ||
6431 | QETH_DBF_HEX(trace, 3, &addr->u.a4.addr, sizeof(int)); | ||
6432 | } else if (addr->proto == QETH_PROT_IPV6) { | ||
6433 | QETH_DBF_TEXT(trace, 2, "deladdr6"); | ||
6434 | QETH_DBF_HEX(trace,3,&addr->u.a6.addr,8); | ||
6435 | QETH_DBF_HEX(trace,3,((char *)&addr->u.a6.addr)+8,8); | ||
6436 | } else { | ||
6437 | QETH_DBF_TEXT(trace, 2, "deladdr?"); | ||
6438 | QETH_DBF_HEX(trace, 3, addr, sizeof(struct qeth_ipaddr)); | ||
6439 | } | ||
6440 | if (addr->is_multicast) | ||
6441 | rc = qeth_send_setdelmc(card, addr, IPA_CMD_DELIPM); | ||
6442 | else | ||
6443 | rc = qeth_send_setdelip(card, addr, IPA_CMD_DELIP, | ||
6444 | addr->del_flags); | ||
6445 | if (rc) { | ||
6446 | QETH_DBF_TEXT(trace, 2, "failed"); | ||
6447 | /* TODO: re-activate this warning as soon as we have a | ||
6448 | * clean mirco code | ||
6449 | qeth_ipaddr_to_string(addr->proto, (u8 *)&addr->u, buf); | ||
6450 | PRINT_WARN("Could not deregister IP address %s (rc=%x)\n", | ||
6451 | buf, rc); | ||
6452 | */ | ||
6453 | } | ||
6454 | return rc; | ||
6455 | } | ||
6456 | |||
6457 | static int | ||
6458 | qeth_register_addr_entry(struct qeth_card *card, struct qeth_ipaddr *addr) | ||
6459 | { | ||
6460 | if (card->options.layer2) | ||
6461 | return qeth_layer2_register_addr_entry(card, addr); | ||
6462 | |||
6463 | return qeth_layer3_register_addr_entry(card, addr); | ||
6464 | } | ||
6465 | |||
6466 | static int | ||
6467 | qeth_deregister_addr_entry(struct qeth_card *card, struct qeth_ipaddr *addr) | ||
6468 | { | ||
6469 | if (card->options.layer2) | ||
6470 | return qeth_layer2_deregister_addr_entry(card, addr); | ||
6471 | |||
6472 | return qeth_layer3_deregister_addr_entry(card, addr); | ||
6473 | } | ||
6474 | |||
6475 | static u32 | ||
6476 | qeth_ethtool_get_tx_csum(struct net_device *dev) | ||
6477 | { | ||
6478 | return (dev->features & NETIF_F_HW_CSUM) != 0; | ||
6479 | } | ||
6480 | |||
6481 | static int | ||
6482 | qeth_ethtool_set_tx_csum(struct net_device *dev, u32 data) | ||
6483 | { | ||
6484 | if (data) | ||
6485 | dev->features |= NETIF_F_HW_CSUM; | ||
6486 | else | ||
6487 | dev->features &= ~NETIF_F_HW_CSUM; | ||
6488 | |||
6489 | return 0; | ||
6490 | } | ||
6491 | |||
6492 | static u32 | ||
6493 | qeth_ethtool_get_rx_csum(struct net_device *dev) | ||
6494 | { | ||
6495 | struct qeth_card *card = (struct qeth_card *)dev->priv; | ||
6496 | |||
6497 | return (card->options.checksum_type == HW_CHECKSUMMING); | ||
6498 | } | ||
6499 | |||
6500 | static int | ||
6501 | qeth_ethtool_set_rx_csum(struct net_device *dev, u32 data) | ||
6502 | { | ||
6503 | struct qeth_card *card = (struct qeth_card *)dev->priv; | ||
6504 | |||
6505 | if ((card->state != CARD_STATE_DOWN) && | ||
6506 | (card->state != CARD_STATE_RECOVER)) | ||
6507 | return -EPERM; | ||
6508 | if (data) | ||
6509 | card->options.checksum_type = HW_CHECKSUMMING; | ||
6510 | else | ||
6511 | card->options.checksum_type = SW_CHECKSUMMING; | ||
6512 | return 0; | ||
6513 | } | ||
6514 | |||
6515 | static u32 | ||
6516 | qeth_ethtool_get_sg(struct net_device *dev) | ||
6517 | { | ||
6518 | struct qeth_card *card = (struct qeth_card *)dev->priv; | ||
6519 | |||
6520 | return ((card->options.large_send != QETH_LARGE_SEND_NO) && | ||
6521 | (dev->features & NETIF_F_SG)); | ||
6522 | } | ||
6523 | |||
6524 | static int | ||
6525 | qeth_ethtool_set_sg(struct net_device *dev, u32 data) | ||
6526 | { | ||
6527 | struct qeth_card *card = (struct qeth_card *)dev->priv; | ||
6528 | |||
6529 | if (data) { | ||
6530 | if (card->options.large_send != QETH_LARGE_SEND_NO) | ||
6531 | dev->features |= NETIF_F_SG; | ||
6532 | else { | ||
6533 | dev->features &= ~NETIF_F_SG; | ||
6534 | return -EINVAL; | ||
6535 | } | ||
6536 | } else | ||
6537 | dev->features &= ~NETIF_F_SG; | ||
6538 | return 0; | ||
6539 | } | ||
6540 | |||
6541 | static u32 | ||
6542 | qeth_ethtool_get_tso(struct net_device *dev) | ||
6543 | { | ||
6544 | struct qeth_card *card = (struct qeth_card *)dev->priv; | ||
6545 | |||
6546 | return ((card->options.large_send != QETH_LARGE_SEND_NO) && | ||
6547 | (dev->features & NETIF_F_TSO)); | ||
6548 | } | ||
6549 | |||
6550 | static int | ||
6551 | qeth_ethtool_set_tso(struct net_device *dev, u32 data) | ||
6552 | { | ||
6553 | struct qeth_card *card = (struct qeth_card *)dev->priv; | ||
6554 | |||
6555 | if (data) { | ||
6556 | if (card->options.large_send != QETH_LARGE_SEND_NO) | ||
6557 | dev->features |= NETIF_F_TSO; | ||
6558 | else { | ||
6559 | dev->features &= ~NETIF_F_TSO; | ||
6560 | return -EINVAL; | ||
6561 | } | ||
6562 | } else | ||
6563 | dev->features &= ~NETIF_F_TSO; | ||
6564 | return 0; | ||
6565 | } | ||
6566 | |||
6567 | static struct ethtool_ops qeth_ethtool_ops = { | ||
6568 | .get_tx_csum = qeth_ethtool_get_tx_csum, | ||
6569 | .set_tx_csum = qeth_ethtool_set_tx_csum, | ||
6570 | .get_rx_csum = qeth_ethtool_get_rx_csum, | ||
6571 | .set_rx_csum = qeth_ethtool_set_rx_csum, | ||
6572 | .get_sg = qeth_ethtool_get_sg, | ||
6573 | .set_sg = qeth_ethtool_set_sg, | ||
6574 | .get_tso = qeth_ethtool_get_tso, | ||
6575 | .set_tso = qeth_ethtool_set_tso, | ||
6576 | }; | ||
6577 | |||
6578 | static int | ||
6579 | qeth_hard_header_parse(const struct sk_buff *skb, unsigned char *haddr) | ||
6580 | { | ||
6581 | const struct qeth_card *card; | ||
6582 | const struct ethhdr *eth; | ||
6583 | struct net_device *dev = skb->dev; | ||
6584 | |||
6585 | if (dev->type != ARPHRD_IEEE802_TR) | ||
6586 | return 0; | ||
6587 | |||
6588 | card = qeth_get_card_from_dev(dev); | ||
6589 | if (card->options.layer2) | ||
6590 | goto haveheader; | ||
6591 | #ifdef CONFIG_QETH_IPV6 | ||
6592 | /* cause of the manipulated arp constructor and the ARP | ||
6593 | flag for OSAE devices we have some nasty exceptions */ | ||
6594 | if (card->info.type == QETH_CARD_TYPE_OSAE) { | ||
6595 | if (!card->options.fake_ll) { | ||
6596 | if ((skb->pkt_type==PACKET_OUTGOING) && | ||
6597 | (skb->protocol==ETH_P_IPV6)) | ||
6598 | goto haveheader; | ||
6599 | else | ||
6600 | return 0; | ||
6601 | } else { | ||
6602 | if ((skb->pkt_type==PACKET_OUTGOING) && | ||
6603 | (skb->protocol==ETH_P_IP)) | ||
6604 | return 0; | ||
6605 | else | ||
6606 | goto haveheader; | ||
6607 | } | ||
6608 | } | ||
6609 | #endif | ||
6610 | if (!card->options.fake_ll) | ||
6611 | return 0; | ||
6612 | haveheader: | ||
6613 | eth = eth_hdr(skb); | ||
6614 | memcpy(haddr, eth->h_source, ETH_ALEN); | ||
6615 | return ETH_ALEN; | ||
6616 | } | ||
6617 | |||
6618 | static const struct header_ops qeth_null_ops = { | ||
6619 | .parse = qeth_hard_header_parse, | ||
6620 | }; | ||
6621 | |||
6622 | static int | ||
6623 | qeth_netdev_init(struct net_device *dev) | ||
6624 | { | ||
6625 | struct qeth_card *card; | ||
6626 | |||
6627 | card = (struct qeth_card *) dev->priv; | ||
6628 | |||
6629 | QETH_DBF_TEXT(trace,3,"initdev"); | ||
6630 | |||
6631 | dev->tx_timeout = &qeth_tx_timeout; | ||
6632 | dev->watchdog_timeo = QETH_TX_TIMEOUT; | ||
6633 | dev->open = qeth_open; | ||
6634 | dev->stop = qeth_stop; | ||
6635 | dev->hard_start_xmit = qeth_hard_start_xmit; | ||
6636 | dev->do_ioctl = qeth_do_ioctl; | ||
6637 | dev->get_stats = qeth_get_stats; | ||
6638 | dev->change_mtu = qeth_change_mtu; | ||
6639 | dev->neigh_setup = qeth_neigh_setup; | ||
6640 | dev->set_multicast_list = qeth_set_multicast_list; | ||
6641 | #ifdef CONFIG_QETH_VLAN | ||
6642 | dev->vlan_rx_register = qeth_vlan_rx_register; | ||
6643 | dev->vlan_rx_kill_vid = qeth_vlan_rx_kill_vid; | ||
6644 | dev->vlan_rx_add_vid = qeth_vlan_rx_add_vid; | ||
6645 | #endif | ||
6646 | if (qeth_get_netdev_flags(card) & IFF_NOARP) | ||
6647 | dev->header_ops = &qeth_null_ops; | ||
6648 | |||
6649 | #ifdef CONFIG_QETH_IPV6 | ||
6650 | /*IPv6 address autoconfiguration stuff*/ | ||
6651 | if (!(card->info.unique_id & UNIQUE_ID_NOT_BY_CARD)) | ||
6652 | card->dev->dev_id = card->info.unique_id & 0xffff; | ||
6653 | #endif | ||
6654 | if (card->options.fake_ll && | ||
6655 | (qeth_get_netdev_flags(card) & IFF_NOARP)) | ||
6656 | dev->header_ops = &qeth_fake_ops; | ||
6657 | |||
6658 | dev->set_mac_address = qeth_layer2_set_mac_address; | ||
6659 | dev->flags |= qeth_get_netdev_flags(card); | ||
6660 | if ((card->options.fake_broadcast) || | ||
6661 | (card->info.broadcast_capable)) | ||
6662 | dev->flags |= IFF_BROADCAST; | ||
6663 | dev->hard_header_len = | ||
6664 | qeth_get_hlen(card->info.link_type) + card->options.add_hhlen; | ||
6665 | dev->addr_len = OSA_ADDR_LEN; | ||
6666 | dev->mtu = card->info.initial_mtu; | ||
6667 | if (card->info.type != QETH_CARD_TYPE_OSN) | ||
6668 | SET_ETHTOOL_OPS(dev, &qeth_ethtool_ops); | ||
6669 | return 0; | ||
6670 | } | ||
6671 | |||
6672 | static void | ||
6673 | qeth_init_func_level(struct qeth_card *card) | ||
6674 | { | ||
6675 | if (card->ipato.enabled) { | ||
6676 | if (card->info.type == QETH_CARD_TYPE_IQD) | ||
6677 | card->info.func_level = | ||
6678 | QETH_IDX_FUNC_LEVEL_IQD_ENA_IPAT; | ||
6679 | else | ||
6680 | card->info.func_level = | ||
6681 | QETH_IDX_FUNC_LEVEL_OSAE_ENA_IPAT; | ||
6682 | } else { | ||
6683 | if (card->info.type == QETH_CARD_TYPE_IQD) | ||
6684 | /*FIXME:why do we have same values for dis and ena for osae??? */ | ||
6685 | card->info.func_level = | ||
6686 | QETH_IDX_FUNC_LEVEL_IQD_DIS_IPAT; | ||
6687 | else | ||
6688 | card->info.func_level = | ||
6689 | QETH_IDX_FUNC_LEVEL_OSAE_DIS_IPAT; | ||
6690 | } | ||
6691 | } | ||
6692 | |||
6693 | /** | ||
6694 | * hardsetup card, initialize MPC and QDIO stuff | ||
6695 | */ | ||
6696 | static int | ||
6697 | qeth_hardsetup_card(struct qeth_card *card) | ||
6698 | { | ||
6699 | int retries = 3; | ||
6700 | int rc; | ||
6701 | |||
6702 | QETH_DBF_TEXT(setup, 2, "hrdsetup"); | ||
6703 | |||
6704 | atomic_set(&card->force_alloc_skb, 0); | ||
6705 | retry: | ||
6706 | if (retries < 3){ | ||
6707 | PRINT_WARN("Retrying to do IDX activates.\n"); | ||
6708 | ccw_device_set_offline(CARD_DDEV(card)); | ||
6709 | ccw_device_set_offline(CARD_WDEV(card)); | ||
6710 | ccw_device_set_offline(CARD_RDEV(card)); | ||
6711 | ccw_device_set_online(CARD_RDEV(card)); | ||
6712 | ccw_device_set_online(CARD_WDEV(card)); | ||
6713 | ccw_device_set_online(CARD_DDEV(card)); | ||
6714 | } | ||
6715 | rc = qeth_qdio_clear_card(card,card->info.type!=QETH_CARD_TYPE_IQD); | ||
6716 | if (rc == -ERESTARTSYS) { | ||
6717 | QETH_DBF_TEXT(setup, 2, "break1"); | ||
6718 | return rc; | ||
6719 | } else if (rc) { | ||
6720 | QETH_DBF_TEXT_(setup, 2, "1err%d", rc); | ||
6721 | if (--retries < 0) | ||
6722 | goto out; | ||
6723 | else | ||
6724 | goto retry; | ||
6725 | } | ||
6726 | if ((rc = qeth_get_unitaddr(card))){ | ||
6727 | QETH_DBF_TEXT_(setup, 2, "2err%d", rc); | ||
6728 | return rc; | ||
6729 | } | ||
6730 | qeth_init_tokens(card); | ||
6731 | qeth_init_func_level(card); | ||
6732 | rc = qeth_idx_activate_channel(&card->read, qeth_idx_read_cb); | ||
6733 | if (rc == -ERESTARTSYS) { | ||
6734 | QETH_DBF_TEXT(setup, 2, "break2"); | ||
6735 | return rc; | ||
6736 | } else if (rc) { | ||
6737 | QETH_DBF_TEXT_(setup, 2, "3err%d", rc); | ||
6738 | if (--retries < 0) | ||
6739 | goto out; | ||
6740 | else | ||
6741 | goto retry; | ||
6742 | } | ||
6743 | rc = qeth_idx_activate_channel(&card->write, qeth_idx_write_cb); | ||
6744 | if (rc == -ERESTARTSYS) { | ||
6745 | QETH_DBF_TEXT(setup, 2, "break3"); | ||
6746 | return rc; | ||
6747 | } else if (rc) { | ||
6748 | QETH_DBF_TEXT_(setup, 2, "4err%d", rc); | ||
6749 | if (--retries < 0) | ||
6750 | goto out; | ||
6751 | else | ||
6752 | goto retry; | ||
6753 | } | ||
6754 | if ((rc = qeth_mpc_initialize(card))){ | ||
6755 | QETH_DBF_TEXT_(setup, 2, "5err%d", rc); | ||
6756 | goto out; | ||
6757 | } | ||
6758 | /*network device will be recovered*/ | ||
6759 | if (card->dev) { | ||
6760 | card->dev->header_ops = card->orig_header_ops; | ||
6761 | if (card->options.fake_ll && | ||
6762 | (qeth_get_netdev_flags(card) & IFF_NOARP)) | ||
6763 | card->dev->header_ops = &qeth_fake_ops; | ||
6764 | return 0; | ||
6765 | } | ||
6766 | /* at first set_online allocate netdev */ | ||
6767 | card->dev = qeth_get_netdevice(card->info.type, | ||
6768 | card->info.link_type); | ||
6769 | if (!card->dev){ | ||
6770 | qeth_qdio_clear_card(card, card->info.type != | ||
6771 | QETH_CARD_TYPE_IQD); | ||
6772 | rc = -ENODEV; | ||
6773 | QETH_DBF_TEXT_(setup, 2, "6err%d", rc); | ||
6774 | goto out; | ||
6775 | } | ||
6776 | card->dev->priv = card; | ||
6777 | card->orig_header_ops = card->dev->header_ops; | ||
6778 | card->dev->type = qeth_get_arphdr_type(card->info.type, | ||
6779 | card->info.link_type); | ||
6780 | card->dev->init = qeth_netdev_init; | ||
6781 | return 0; | ||
6782 | out: | ||
6783 | PRINT_ERR("Initialization in hardsetup failed! rc=%d\n", rc); | ||
6784 | return rc; | ||
6785 | } | ||
6786 | |||
6787 | static int | ||
6788 | qeth_default_setassparms_cb(struct qeth_card *card, struct qeth_reply *reply, | ||
6789 | unsigned long data) | ||
6790 | { | ||
6791 | struct qeth_ipa_cmd *cmd; | ||
6792 | |||
6793 | QETH_DBF_TEXT(trace,4,"defadpcb"); | ||
6794 | |||
6795 | cmd = (struct qeth_ipa_cmd *) data; | ||
6796 | if (cmd->hdr.return_code == 0){ | ||
6797 | cmd->hdr.return_code = cmd->data.setassparms.hdr.return_code; | ||
6798 | if (cmd->hdr.prot_version == QETH_PROT_IPV4) | ||
6799 | card->options.ipa4.enabled_funcs = cmd->hdr.ipa_enabled; | ||
6800 | #ifdef CONFIG_QETH_IPV6 | ||
6801 | if (cmd->hdr.prot_version == QETH_PROT_IPV6) | ||
6802 | card->options.ipa6.enabled_funcs = cmd->hdr.ipa_enabled; | ||
6803 | #endif | ||
6804 | } | ||
6805 | if (cmd->data.setassparms.hdr.assist_no == IPA_INBOUND_CHECKSUM && | ||
6806 | cmd->data.setassparms.hdr.command_code == IPA_CMD_ASS_START) { | ||
6807 | card->info.csum_mask = cmd->data.setassparms.data.flags_32bit; | ||
6808 | QETH_DBF_TEXT_(trace, 3, "csum:%d", card->info.csum_mask); | ||
6809 | } | ||
6810 | return 0; | ||
6811 | } | ||
6812 | |||
6813 | static int | ||
6814 | qeth_default_setadapterparms_cb(struct qeth_card *card, | ||
6815 | struct qeth_reply *reply, | ||
6816 | unsigned long data) | ||
6817 | { | ||
6818 | struct qeth_ipa_cmd *cmd; | ||
6819 | |||
6820 | QETH_DBF_TEXT(trace,4,"defadpcb"); | ||
6821 | |||
6822 | cmd = (struct qeth_ipa_cmd *) data; | ||
6823 | if (cmd->hdr.return_code == 0) | ||
6824 | cmd->hdr.return_code = cmd->data.setadapterparms.hdr.return_code; | ||
6825 | return 0; | ||
6826 | } | ||
6827 | |||
6828 | |||
6829 | |||
6830 | static int | ||
6831 | qeth_query_setadapterparms_cb(struct qeth_card *card, struct qeth_reply *reply, | ||
6832 | unsigned long data) | ||
6833 | { | ||
6834 | struct qeth_ipa_cmd *cmd; | ||
6835 | |||
6836 | QETH_DBF_TEXT(trace,3,"quyadpcb"); | ||
6837 | |||
6838 | cmd = (struct qeth_ipa_cmd *) data; | ||
6839 | if (cmd->data.setadapterparms.data.query_cmds_supp.lan_type & 0x7f) | ||
6840 | card->info.link_type = | ||
6841 | cmd->data.setadapterparms.data.query_cmds_supp.lan_type; | ||
6842 | card->options.adp.supported_funcs = | ||
6843 | cmd->data.setadapterparms.data.query_cmds_supp.supported_cmds; | ||
6844 | return qeth_default_setadapterparms_cb(card, reply, (unsigned long)cmd); | ||
6845 | } | ||
6846 | |||
6847 | static int | ||
6848 | qeth_query_setadapterparms(struct qeth_card *card) | ||
6849 | { | ||
6850 | int rc; | ||
6851 | struct qeth_cmd_buffer *iob; | ||
6852 | |||
6853 | QETH_DBF_TEXT(trace,3,"queryadp"); | ||
6854 | iob = qeth_get_adapter_cmd(card, IPA_SETADP_QUERY_COMMANDS_SUPPORTED, | ||
6855 | sizeof(struct qeth_ipacmd_setadpparms)); | ||
6856 | rc = qeth_send_ipa_cmd(card, iob, qeth_query_setadapterparms_cb, NULL); | ||
6857 | return rc; | ||
6858 | } | ||
6859 | |||
6860 | static int | ||
6861 | qeth_setadpparms_change_macaddr_cb(struct qeth_card *card, | ||
6862 | struct qeth_reply *reply, | ||
6863 | unsigned long data) | ||
6864 | { | ||
6865 | struct qeth_ipa_cmd *cmd; | ||
6866 | |||
6867 | QETH_DBF_TEXT(trace,4,"chgmaccb"); | ||
6868 | |||
6869 | cmd = (struct qeth_ipa_cmd *) data; | ||
6870 | if (!card->options.layer2 || | ||
6871 | !(card->info.mac_bits & QETH_LAYER2_MAC_READ)) { | ||
6872 | memcpy(card->dev->dev_addr, | ||
6873 | &cmd->data.setadapterparms.data.change_addr.addr, | ||
6874 | OSA_ADDR_LEN); | ||
6875 | card->info.mac_bits |= QETH_LAYER2_MAC_READ; | ||
6876 | } | ||
6877 | qeth_default_setadapterparms_cb(card, reply, (unsigned long) cmd); | ||
6878 | return 0; | ||
6879 | } | ||
6880 | |||
6881 | static int | ||
6882 | qeth_setadpparms_change_macaddr(struct qeth_card *card) | ||
6883 | { | ||
6884 | int rc; | ||
6885 | struct qeth_cmd_buffer *iob; | ||
6886 | struct qeth_ipa_cmd *cmd; | ||
6887 | |||
6888 | QETH_DBF_TEXT(trace,4,"chgmac"); | ||
6889 | |||
6890 | iob = qeth_get_adapter_cmd(card,IPA_SETADP_ALTER_MAC_ADDRESS, | ||
6891 | sizeof(struct qeth_ipacmd_setadpparms)); | ||
6892 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
6893 | cmd->data.setadapterparms.data.change_addr.cmd = CHANGE_ADDR_READ_MAC; | ||
6894 | cmd->data.setadapterparms.data.change_addr.addr_size = OSA_ADDR_LEN; | ||
6895 | memcpy(&cmd->data.setadapterparms.data.change_addr.addr, | ||
6896 | card->dev->dev_addr, OSA_ADDR_LEN); | ||
6897 | rc = qeth_send_ipa_cmd(card, iob, qeth_setadpparms_change_macaddr_cb, | ||
6898 | NULL); | ||
6899 | return rc; | ||
6900 | } | ||
6901 | |||
6902 | static int | ||
6903 | qeth_send_setadp_mode(struct qeth_card *card, __u32 command, __u32 mode) | ||
6904 | { | ||
6905 | int rc; | ||
6906 | struct qeth_cmd_buffer *iob; | ||
6907 | struct qeth_ipa_cmd *cmd; | ||
6908 | |||
6909 | QETH_DBF_TEXT(trace,4,"adpmode"); | ||
6910 | |||
6911 | iob = qeth_get_adapter_cmd(card, command, | ||
6912 | sizeof(struct qeth_ipacmd_setadpparms)); | ||
6913 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
6914 | cmd->data.setadapterparms.data.mode = mode; | ||
6915 | rc = qeth_send_ipa_cmd(card, iob, qeth_default_setadapterparms_cb, | ||
6916 | NULL); | ||
6917 | return rc; | ||
6918 | } | ||
6919 | |||
6920 | static int | ||
6921 | qeth_setadapter_hstr(struct qeth_card *card) | ||
6922 | { | ||
6923 | int rc; | ||
6924 | |||
6925 | QETH_DBF_TEXT(trace,4,"adphstr"); | ||
6926 | |||
6927 | if (qeth_adp_supported(card,IPA_SETADP_SET_BROADCAST_MODE)) { | ||
6928 | rc = qeth_send_setadp_mode(card, IPA_SETADP_SET_BROADCAST_MODE, | ||
6929 | card->options.broadcast_mode); | ||
6930 | if (rc) | ||
6931 | PRINT_WARN("couldn't set broadcast mode on " | ||
6932 | "device %s: x%x\n", | ||
6933 | CARD_BUS_ID(card), rc); | ||
6934 | rc = qeth_send_setadp_mode(card, IPA_SETADP_ALTER_MAC_ADDRESS, | ||
6935 | card->options.macaddr_mode); | ||
6936 | if (rc) | ||
6937 | PRINT_WARN("couldn't set macaddr mode on " | ||
6938 | "device %s: x%x\n", CARD_BUS_ID(card), rc); | ||
6939 | return rc; | ||
6940 | } | ||
6941 | if (card->options.broadcast_mode == QETH_TR_BROADCAST_LOCAL) | ||
6942 | PRINT_WARN("set adapter parameters not available " | ||
6943 | "to set broadcast mode, using ALLRINGS " | ||
6944 | "on device %s:\n", CARD_BUS_ID(card)); | ||
6945 | if (card->options.macaddr_mode == QETH_TR_MACADDR_CANONICAL) | ||
6946 | PRINT_WARN("set adapter parameters not available " | ||
6947 | "to set macaddr mode, using NONCANONICAL " | ||
6948 | "on device %s:\n", CARD_BUS_ID(card)); | ||
6949 | return 0; | ||
6950 | } | ||
6951 | |||
6952 | static int | ||
6953 | qeth_setadapter_parms(struct qeth_card *card) | ||
6954 | { | ||
6955 | int rc; | ||
6956 | |||
6957 | QETH_DBF_TEXT(setup, 2, "setadprm"); | ||
6958 | |||
6959 | if (!qeth_is_supported(card, IPA_SETADAPTERPARMS)){ | ||
6960 | PRINT_WARN("set adapter parameters not supported " | ||
6961 | "on device %s.\n", | ||
6962 | CARD_BUS_ID(card)); | ||
6963 | QETH_DBF_TEXT(setup, 2, " notsupp"); | ||
6964 | return 0; | ||
6965 | } | ||
6966 | rc = qeth_query_setadapterparms(card); | ||
6967 | if (rc) { | ||
6968 | PRINT_WARN("couldn't set adapter parameters on device %s: " | ||
6969 | "x%x\n", CARD_BUS_ID(card), rc); | ||
6970 | return rc; | ||
6971 | } | ||
6972 | if (qeth_adp_supported(card,IPA_SETADP_ALTER_MAC_ADDRESS)) { | ||
6973 | rc = qeth_setadpparms_change_macaddr(card); | ||
6974 | if (rc) | ||
6975 | PRINT_WARN("couldn't get MAC address on " | ||
6976 | "device %s: x%x\n", | ||
6977 | CARD_BUS_ID(card), rc); | ||
6978 | } | ||
6979 | |||
6980 | if ((card->info.link_type == QETH_LINK_TYPE_HSTR) || | ||
6981 | (card->info.link_type == QETH_LINK_TYPE_LANE_TR)) | ||
6982 | rc = qeth_setadapter_hstr(card); | ||
6983 | |||
6984 | return rc; | ||
6985 | } | ||
6986 | |||
6987 | static int | ||
6988 | qeth_layer2_initialize(struct qeth_card *card) | ||
6989 | { | ||
6990 | int rc = 0; | ||
6991 | |||
6992 | |||
6993 | QETH_DBF_TEXT(setup, 2, "doL2init"); | ||
6994 | QETH_DBF_TEXT_(setup, 2, "doL2%s", CARD_BUS_ID(card)); | ||
6995 | |||
6996 | rc = qeth_query_setadapterparms(card); | ||
6997 | if (rc) { | ||
6998 | PRINT_WARN("could not query adapter parameters on device %s: " | ||
6999 | "x%x\n", CARD_BUS_ID(card), rc); | ||
7000 | } | ||
7001 | |||
7002 | rc = qeth_setadpparms_change_macaddr(card); | ||
7003 | if (rc) { | ||
7004 | PRINT_WARN("couldn't get MAC address on " | ||
7005 | "device %s: x%x\n", | ||
7006 | CARD_BUS_ID(card), rc); | ||
7007 | QETH_DBF_TEXT_(setup, 2,"1err%d",rc); | ||
7008 | return rc; | ||
7009 | } | ||
7010 | QETH_DBF_HEX(setup,2, card->dev->dev_addr, OSA_ADDR_LEN); | ||
7011 | |||
7012 | rc = qeth_layer2_send_setmac(card, &card->dev->dev_addr[0]); | ||
7013 | if (rc) | ||
7014 | QETH_DBF_TEXT_(setup, 2,"2err%d",rc); | ||
7015 | return 0; | ||
7016 | } | ||
7017 | |||
7018 | |||
7019 | static int | ||
7020 | qeth_send_startstoplan(struct qeth_card *card, enum qeth_ipa_cmds ipacmd, | ||
7021 | enum qeth_prot_versions prot) | ||
7022 | { | ||
7023 | int rc; | ||
7024 | struct qeth_cmd_buffer *iob; | ||
7025 | |||
7026 | iob = qeth_get_ipacmd_buffer(card,ipacmd,prot); | ||
7027 | rc = qeth_send_ipa_cmd(card, iob, NULL, NULL); | ||
7028 | |||
7029 | return rc; | ||
7030 | } | ||
7031 | |||
7032 | static int | ||
7033 | qeth_send_startlan(struct qeth_card *card, enum qeth_prot_versions prot) | ||
7034 | { | ||
7035 | int rc; | ||
7036 | |||
7037 | QETH_DBF_TEXT_(setup, 2, "strtlan%i", prot); | ||
7038 | |||
7039 | rc = qeth_send_startstoplan(card, IPA_CMD_STARTLAN, prot); | ||
7040 | return rc; | ||
7041 | } | ||
7042 | |||
7043 | static int | ||
7044 | qeth_send_stoplan(struct qeth_card *card) | ||
7045 | { | ||
7046 | int rc = 0; | ||
7047 | |||
7048 | /* | ||
7049 | * TODO: according to the IPA format document page 14, | ||
7050 | * TCP/IP (we!) never issue a STOPLAN | ||
7051 | * is this right ?!? | ||
7052 | */ | ||
7053 | QETH_DBF_TEXT(trace, 2, "stoplan"); | ||
7054 | |||
7055 | rc = qeth_send_startstoplan(card, IPA_CMD_STOPLAN, QETH_PROT_IPV4); | ||
7056 | return rc; | ||
7057 | } | ||
7058 | |||
7059 | static int | ||
7060 | qeth_query_ipassists_cb(struct qeth_card *card, struct qeth_reply *reply, | ||
7061 | unsigned long data) | ||
7062 | { | ||
7063 | struct qeth_ipa_cmd *cmd; | ||
7064 | |||
7065 | QETH_DBF_TEXT(setup, 2, "qipasscb"); | ||
7066 | |||
7067 | cmd = (struct qeth_ipa_cmd *) data; | ||
7068 | if (cmd->hdr.prot_version == QETH_PROT_IPV4) { | ||
7069 | card->options.ipa4.supported_funcs = cmd->hdr.ipa_supported; | ||
7070 | card->options.ipa4.enabled_funcs = cmd->hdr.ipa_enabled; | ||
7071 | /* Disable IPV6 support hard coded for Hipersockets */ | ||
7072 | if(card->info.type == QETH_CARD_TYPE_IQD) | ||
7073 | card->options.ipa4.supported_funcs &= ~IPA_IPV6; | ||
7074 | } else { | ||
7075 | #ifdef CONFIG_QETH_IPV6 | ||
7076 | card->options.ipa6.supported_funcs = cmd->hdr.ipa_supported; | ||
7077 | card->options.ipa6.enabled_funcs = cmd->hdr.ipa_enabled; | ||
7078 | #endif | ||
7079 | } | ||
7080 | QETH_DBF_TEXT(setup, 2, "suppenbl"); | ||
7081 | QETH_DBF_TEXT_(setup, 2, "%x",cmd->hdr.ipa_supported); | ||
7082 | QETH_DBF_TEXT_(setup, 2, "%x",cmd->hdr.ipa_enabled); | ||
7083 | return 0; | ||
7084 | } | ||
7085 | |||
7086 | static int | ||
7087 | qeth_query_ipassists(struct qeth_card *card, enum qeth_prot_versions prot) | ||
7088 | { | ||
7089 | int rc; | ||
7090 | struct qeth_cmd_buffer *iob; | ||
7091 | |||
7092 | QETH_DBF_TEXT_(setup, 2, "qipassi%i", prot); | ||
7093 | if (card->options.layer2) { | ||
7094 | QETH_DBF_TEXT(setup, 2, "noprmly2"); | ||
7095 | return -EPERM; | ||
7096 | } | ||
7097 | |||
7098 | iob = qeth_get_ipacmd_buffer(card,IPA_CMD_QIPASSIST,prot); | ||
7099 | rc = qeth_send_ipa_cmd(card, iob, qeth_query_ipassists_cb, NULL); | ||
7100 | return rc; | ||
7101 | } | ||
7102 | |||
7103 | static struct qeth_cmd_buffer * | ||
7104 | qeth_get_setassparms_cmd(struct qeth_card *card, enum qeth_ipa_funcs ipa_func, | ||
7105 | __u16 cmd_code, __u16 len, | ||
7106 | enum qeth_prot_versions prot) | ||
7107 | { | ||
7108 | struct qeth_cmd_buffer *iob; | ||
7109 | struct qeth_ipa_cmd *cmd; | ||
7110 | |||
7111 | QETH_DBF_TEXT(trace,4,"getasscm"); | ||
7112 | iob = qeth_get_ipacmd_buffer(card,IPA_CMD_SETASSPARMS,prot); | ||
7113 | |||
7114 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
7115 | cmd->data.setassparms.hdr.assist_no = ipa_func; | ||
7116 | cmd->data.setassparms.hdr.length = 8 + len; | ||
7117 | cmd->data.setassparms.hdr.command_code = cmd_code; | ||
7118 | cmd->data.setassparms.hdr.return_code = 0; | ||
7119 | cmd->data.setassparms.hdr.seq_no = 0; | ||
7120 | |||
7121 | return iob; | ||
7122 | } | ||
7123 | |||
7124 | static int | ||
7125 | qeth_send_setassparms(struct qeth_card *card, struct qeth_cmd_buffer *iob, | ||
7126 | __u16 len, long data, | ||
7127 | int (*reply_cb) | ||
7128 | (struct qeth_card *,struct qeth_reply *,unsigned long), | ||
7129 | void *reply_param) | ||
7130 | { | ||
7131 | int rc; | ||
7132 | struct qeth_ipa_cmd *cmd; | ||
7133 | |||
7134 | QETH_DBF_TEXT(trace,4,"sendassp"); | ||
7135 | |||
7136 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
7137 | if (len <= sizeof(__u32)) | ||
7138 | cmd->data.setassparms.data.flags_32bit = (__u32) data; | ||
7139 | else /* (len > sizeof(__u32)) */ | ||
7140 | memcpy(&cmd->data.setassparms.data, (void *) data, len); | ||
7141 | |||
7142 | rc = qeth_send_ipa_cmd(card, iob, reply_cb, reply_param); | ||
7143 | return rc; | ||
7144 | } | ||
7145 | |||
7146 | #ifdef CONFIG_QETH_IPV6 | ||
7147 | static int | ||
7148 | qeth_send_simple_setassparms_ipv6(struct qeth_card *card, | ||
7149 | enum qeth_ipa_funcs ipa_func, __u16 cmd_code) | ||
7150 | |||
7151 | { | ||
7152 | int rc; | ||
7153 | struct qeth_cmd_buffer *iob; | ||
7154 | |||
7155 | QETH_DBF_TEXT(trace,4,"simassp6"); | ||
7156 | iob = qeth_get_setassparms_cmd(card, ipa_func, cmd_code, | ||
7157 | 0, QETH_PROT_IPV6); | ||
7158 | rc = qeth_send_setassparms(card, iob, 0, 0, | ||
7159 | qeth_default_setassparms_cb, NULL); | ||
7160 | return rc; | ||
7161 | } | ||
7162 | #endif | ||
7163 | |||
7164 | static int | ||
7165 | qeth_send_simple_setassparms(struct qeth_card *card, | ||
7166 | enum qeth_ipa_funcs ipa_func, | ||
7167 | __u16 cmd_code, long data) | ||
7168 | { | ||
7169 | int rc; | ||
7170 | int length = 0; | ||
7171 | struct qeth_cmd_buffer *iob; | ||
7172 | |||
7173 | QETH_DBF_TEXT(trace,4,"simassp4"); | ||
7174 | if (data) | ||
7175 | length = sizeof(__u32); | ||
7176 | iob = qeth_get_setassparms_cmd(card, ipa_func, cmd_code, | ||
7177 | length, QETH_PROT_IPV4); | ||
7178 | rc = qeth_send_setassparms(card, iob, length, data, | ||
7179 | qeth_default_setassparms_cb, NULL); | ||
7180 | return rc; | ||
7181 | } | ||
7182 | |||
7183 | static int | ||
7184 | qeth_start_ipa_arp_processing(struct qeth_card *card) | ||
7185 | { | ||
7186 | int rc; | ||
7187 | |||
7188 | QETH_DBF_TEXT(trace,3,"ipaarp"); | ||
7189 | |||
7190 | if (!qeth_is_supported(card,IPA_ARP_PROCESSING)) { | ||
7191 | PRINT_WARN("ARP processing not supported " | ||
7192 | "on %s!\n", QETH_CARD_IFNAME(card)); | ||
7193 | return 0; | ||
7194 | } | ||
7195 | rc = qeth_send_simple_setassparms(card,IPA_ARP_PROCESSING, | ||
7196 | IPA_CMD_ASS_START, 0); | ||
7197 | if (rc) { | ||
7198 | PRINT_WARN("Could not start ARP processing " | ||
7199 | "assist on %s: 0x%x\n", | ||
7200 | QETH_CARD_IFNAME(card), rc); | ||
7201 | } | ||
7202 | return rc; | ||
7203 | } | ||
7204 | |||
7205 | static int | ||
7206 | qeth_start_ipa_ip_fragmentation(struct qeth_card *card) | ||
7207 | { | ||
7208 | int rc; | ||
7209 | |||
7210 | QETH_DBF_TEXT(trace,3,"ipaipfrg"); | ||
7211 | |||
7212 | if (!qeth_is_supported(card, IPA_IP_FRAGMENTATION)) { | ||
7213 | PRINT_INFO("Hardware IP fragmentation not supported on %s\n", | ||
7214 | QETH_CARD_IFNAME(card)); | ||
7215 | return -EOPNOTSUPP; | ||
7216 | } | ||
7217 | |||
7218 | rc = qeth_send_simple_setassparms(card, IPA_IP_FRAGMENTATION, | ||
7219 | IPA_CMD_ASS_START, 0); | ||
7220 | if (rc) { | ||
7221 | PRINT_WARN("Could not start Hardware IP fragmentation " | ||
7222 | "assist on %s: 0x%x\n", | ||
7223 | QETH_CARD_IFNAME(card), rc); | ||
7224 | } else | ||
7225 | PRINT_INFO("Hardware IP fragmentation enabled \n"); | ||
7226 | return rc; | ||
7227 | } | ||
7228 | |||
7229 | static int | ||
7230 | qeth_start_ipa_source_mac(struct qeth_card *card) | ||
7231 | { | ||
7232 | int rc; | ||
7233 | |||
7234 | QETH_DBF_TEXT(trace,3,"stsrcmac"); | ||
7235 | |||
7236 | if (!card->options.fake_ll) | ||
7237 | return -EOPNOTSUPP; | ||
7238 | |||
7239 | if (!qeth_is_supported(card, IPA_SOURCE_MAC)) { | ||
7240 | PRINT_INFO("Inbound source address not " | ||
7241 | "supported on %s\n", QETH_CARD_IFNAME(card)); | ||
7242 | return -EOPNOTSUPP; | ||
7243 | } | ||
7244 | |||
7245 | rc = qeth_send_simple_setassparms(card, IPA_SOURCE_MAC, | ||
7246 | IPA_CMD_ASS_START, 0); | ||
7247 | if (rc) | ||
7248 | PRINT_WARN("Could not start inbound source " | ||
7249 | "assist on %s: 0x%x\n", | ||
7250 | QETH_CARD_IFNAME(card), rc); | ||
7251 | return rc; | ||
7252 | } | ||
7253 | |||
7254 | static int | ||
7255 | qeth_start_ipa_vlan(struct qeth_card *card) | ||
7256 | { | ||
7257 | int rc = 0; | ||
7258 | |||
7259 | QETH_DBF_TEXT(trace,3,"strtvlan"); | ||
7260 | |||
7261 | #ifdef CONFIG_QETH_VLAN | ||
7262 | if (!qeth_is_supported(card, IPA_FULL_VLAN)) { | ||
7263 | PRINT_WARN("VLAN not supported on %s\n", QETH_CARD_IFNAME(card)); | ||
7264 | return -EOPNOTSUPP; | ||
7265 | } | ||
7266 | |||
7267 | rc = qeth_send_simple_setassparms(card, IPA_VLAN_PRIO, | ||
7268 | IPA_CMD_ASS_START,0); | ||
7269 | if (rc) { | ||
7270 | PRINT_WARN("Could not start vlan " | ||
7271 | "assist on %s: 0x%x\n", | ||
7272 | QETH_CARD_IFNAME(card), rc); | ||
7273 | } else { | ||
7274 | PRINT_INFO("VLAN enabled \n"); | ||
7275 | card->dev->features |= | ||
7276 | NETIF_F_HW_VLAN_FILTER | | ||
7277 | NETIF_F_HW_VLAN_TX | | ||
7278 | NETIF_F_HW_VLAN_RX; | ||
7279 | } | ||
7280 | #endif /* QETH_VLAN */ | ||
7281 | return rc; | ||
7282 | } | ||
7283 | |||
7284 | static int | ||
7285 | qeth_start_ipa_multicast(struct qeth_card *card) | ||
7286 | { | ||
7287 | int rc; | ||
7288 | |||
7289 | QETH_DBF_TEXT(trace,3,"stmcast"); | ||
7290 | |||
7291 | if (!qeth_is_supported(card, IPA_MULTICASTING)) { | ||
7292 | PRINT_WARN("Multicast not supported on %s\n", | ||
7293 | QETH_CARD_IFNAME(card)); | ||
7294 | return -EOPNOTSUPP; | ||
7295 | } | ||
7296 | |||
7297 | rc = qeth_send_simple_setassparms(card, IPA_MULTICASTING, | ||
7298 | IPA_CMD_ASS_START,0); | ||
7299 | if (rc) { | ||
7300 | PRINT_WARN("Could not start multicast " | ||
7301 | "assist on %s: rc=%i\n", | ||
7302 | QETH_CARD_IFNAME(card), rc); | ||
7303 | } else { | ||
7304 | PRINT_INFO("Multicast enabled\n"); | ||
7305 | card->dev->flags |= IFF_MULTICAST; | ||
7306 | } | ||
7307 | return rc; | ||
7308 | } | ||
7309 | |||
7310 | #ifdef CONFIG_QETH_IPV6 | ||
7311 | static int | ||
7312 | qeth_softsetup_ipv6(struct qeth_card *card) | ||
7313 | { | ||
7314 | int rc; | ||
7315 | |||
7316 | QETH_DBF_TEXT(trace,3,"softipv6"); | ||
7317 | |||
7318 | rc = qeth_send_startlan(card, QETH_PROT_IPV6); | ||
7319 | if (rc) { | ||
7320 | PRINT_ERR("IPv6 startlan failed on %s\n", | ||
7321 | QETH_CARD_IFNAME(card)); | ||
7322 | return rc; | ||
7323 | } | ||
7324 | rc = qeth_query_ipassists(card,QETH_PROT_IPV6); | ||
7325 | if (rc) { | ||
7326 | PRINT_ERR("IPv6 query ipassist failed on %s\n", | ||
7327 | QETH_CARD_IFNAME(card)); | ||
7328 | return rc; | ||
7329 | } | ||
7330 | rc = qeth_send_simple_setassparms(card, IPA_IPV6, | ||
7331 | IPA_CMD_ASS_START, 3); | ||
7332 | if (rc) { | ||
7333 | PRINT_WARN("IPv6 start assist (version 4) failed " | ||
7334 | "on %s: 0x%x\n", | ||
7335 | QETH_CARD_IFNAME(card), rc); | ||
7336 | return rc; | ||
7337 | } | ||
7338 | rc = qeth_send_simple_setassparms_ipv6(card, IPA_IPV6, | ||
7339 | IPA_CMD_ASS_START); | ||
7340 | if (rc) { | ||
7341 | PRINT_WARN("IPV6 start assist (version 6) failed " | ||
7342 | "on %s: 0x%x\n", | ||
7343 | QETH_CARD_IFNAME(card), rc); | ||
7344 | return rc; | ||
7345 | } | ||
7346 | rc = qeth_send_simple_setassparms_ipv6(card, IPA_PASSTHRU, | ||
7347 | IPA_CMD_ASS_START); | ||
7348 | if (rc) { | ||
7349 | PRINT_WARN("Could not enable passthrough " | ||
7350 | "on %s: 0x%x\n", | ||
7351 | QETH_CARD_IFNAME(card), rc); | ||
7352 | return rc; | ||
7353 | } | ||
7354 | PRINT_INFO("IPV6 enabled \n"); | ||
7355 | return 0; | ||
7356 | } | ||
7357 | |||
7358 | #endif | ||
7359 | |||
7360 | static int | ||
7361 | qeth_start_ipa_ipv6(struct qeth_card *card) | ||
7362 | { | ||
7363 | int rc = 0; | ||
7364 | #ifdef CONFIG_QETH_IPV6 | ||
7365 | QETH_DBF_TEXT(trace,3,"strtipv6"); | ||
7366 | |||
7367 | if (!qeth_is_supported(card, IPA_IPV6)) { | ||
7368 | PRINT_WARN("IPv6 not supported on %s\n", | ||
7369 | QETH_CARD_IFNAME(card)); | ||
7370 | return 0; | ||
7371 | } | ||
7372 | rc = qeth_softsetup_ipv6(card); | ||
7373 | #endif | ||
7374 | return rc ; | ||
7375 | } | ||
7376 | |||
7377 | static int | ||
7378 | qeth_start_ipa_broadcast(struct qeth_card *card) | ||
7379 | { | ||
7380 | int rc; | ||
7381 | |||
7382 | QETH_DBF_TEXT(trace,3,"stbrdcst"); | ||
7383 | card->info.broadcast_capable = 0; | ||
7384 | if (!qeth_is_supported(card, IPA_FILTERING)) { | ||
7385 | PRINT_WARN("Broadcast not supported on %s\n", | ||
7386 | QETH_CARD_IFNAME(card)); | ||
7387 | rc = -EOPNOTSUPP; | ||
7388 | goto out; | ||
7389 | } | ||
7390 | rc = qeth_send_simple_setassparms(card, IPA_FILTERING, | ||
7391 | IPA_CMD_ASS_START, 0); | ||
7392 | if (rc) { | ||
7393 | PRINT_WARN("Could not enable broadcasting filtering " | ||
7394 | "on %s: 0x%x\n", | ||
7395 | QETH_CARD_IFNAME(card), rc); | ||
7396 | goto out; | ||
7397 | } | ||
7398 | |||
7399 | rc = qeth_send_simple_setassparms(card, IPA_FILTERING, | ||
7400 | IPA_CMD_ASS_CONFIGURE, 1); | ||
7401 | if (rc) { | ||
7402 | PRINT_WARN("Could not set up broadcast filtering on %s: 0x%x\n", | ||
7403 | QETH_CARD_IFNAME(card), rc); | ||
7404 | goto out; | ||
7405 | } | ||
7406 | card->info.broadcast_capable = QETH_BROADCAST_WITH_ECHO; | ||
7407 | PRINT_INFO("Broadcast enabled \n"); | ||
7408 | rc = qeth_send_simple_setassparms(card, IPA_FILTERING, | ||
7409 | IPA_CMD_ASS_ENABLE, 1); | ||
7410 | if (rc) { | ||
7411 | PRINT_WARN("Could not set up broadcast echo filtering on " | ||
7412 | "%s: 0x%x\n", QETH_CARD_IFNAME(card), rc); | ||
7413 | goto out; | ||
7414 | } | ||
7415 | card->info.broadcast_capable = QETH_BROADCAST_WITHOUT_ECHO; | ||
7416 | out: | ||
7417 | if (card->info.broadcast_capable) | ||
7418 | card->dev->flags |= IFF_BROADCAST; | ||
7419 | else | ||
7420 | card->dev->flags &= ~IFF_BROADCAST; | ||
7421 | return rc; | ||
7422 | } | ||
7423 | |||
7424 | static int | ||
7425 | qeth_send_checksum_command(struct qeth_card *card) | ||
7426 | { | ||
7427 | int rc; | ||
7428 | |||
7429 | rc = qeth_send_simple_setassparms(card, IPA_INBOUND_CHECKSUM, | ||
7430 | IPA_CMD_ASS_START, 0); | ||
7431 | if (rc) { | ||
7432 | PRINT_WARN("Starting Inbound HW Checksumming failed on %s: " | ||
7433 | "0x%x,\ncontinuing using Inbound SW Checksumming\n", | ||
7434 | QETH_CARD_IFNAME(card), rc); | ||
7435 | return rc; | ||
7436 | } | ||
7437 | rc = qeth_send_simple_setassparms(card, IPA_INBOUND_CHECKSUM, | ||
7438 | IPA_CMD_ASS_ENABLE, | ||
7439 | card->info.csum_mask); | ||
7440 | if (rc) { | ||
7441 | PRINT_WARN("Enabling Inbound HW Checksumming failed on %s: " | ||
7442 | "0x%x,\ncontinuing using Inbound SW Checksumming\n", | ||
7443 | QETH_CARD_IFNAME(card), rc); | ||
7444 | return rc; | ||
7445 | } | ||
7446 | return 0; | ||
7447 | } | ||
7448 | |||
7449 | static int | ||
7450 | qeth_start_ipa_checksum(struct qeth_card *card) | ||
7451 | { | ||
7452 | int rc = 0; | ||
7453 | |||
7454 | QETH_DBF_TEXT(trace,3,"strtcsum"); | ||
7455 | |||
7456 | if (card->options.checksum_type == NO_CHECKSUMMING) { | ||
7457 | PRINT_WARN("Using no checksumming on %s.\n", | ||
7458 | QETH_CARD_IFNAME(card)); | ||
7459 | return 0; | ||
7460 | } | ||
7461 | if (card->options.checksum_type == SW_CHECKSUMMING) { | ||
7462 | PRINT_WARN("Using SW checksumming on %s.\n", | ||
7463 | QETH_CARD_IFNAME(card)); | ||
7464 | return 0; | ||
7465 | } | ||
7466 | if (!qeth_is_supported(card, IPA_INBOUND_CHECKSUM)) { | ||
7467 | PRINT_WARN("Inbound HW Checksumming not " | ||
7468 | "supported on %s,\ncontinuing " | ||
7469 | "using Inbound SW Checksumming\n", | ||
7470 | QETH_CARD_IFNAME(card)); | ||
7471 | card->options.checksum_type = SW_CHECKSUMMING; | ||
7472 | return 0; | ||
7473 | } | ||
7474 | rc = qeth_send_checksum_command(card); | ||
7475 | if (!rc) { | ||
7476 | PRINT_INFO("HW Checksumming (inbound) enabled \n"); | ||
7477 | } | ||
7478 | return rc; | ||
7479 | } | ||
7480 | |||
7481 | static int | ||
7482 | qeth_start_ipa_tso(struct qeth_card *card) | ||
7483 | { | ||
7484 | int rc; | ||
7485 | |||
7486 | QETH_DBF_TEXT(trace,3,"sttso"); | ||
7487 | |||
7488 | if (!qeth_is_supported(card, IPA_OUTBOUND_TSO)) { | ||
7489 | PRINT_WARN("Outbound TSO not supported on %s\n", | ||
7490 | QETH_CARD_IFNAME(card)); | ||
7491 | rc = -EOPNOTSUPP; | ||
7492 | } else { | ||
7493 | rc = qeth_send_simple_setassparms(card, IPA_OUTBOUND_TSO, | ||
7494 | IPA_CMD_ASS_START,0); | ||
7495 | if (rc) | ||
7496 | PRINT_WARN("Could not start outbound TSO " | ||
7497 | "assist on %s: rc=%i\n", | ||
7498 | QETH_CARD_IFNAME(card), rc); | ||
7499 | else | ||
7500 | PRINT_INFO("Outbound TSO enabled\n"); | ||
7501 | } | ||
7502 | if (rc && (card->options.large_send == QETH_LARGE_SEND_TSO)){ | ||
7503 | card->options.large_send = QETH_LARGE_SEND_NO; | ||
7504 | card->dev->features &= ~(NETIF_F_TSO | NETIF_F_SG | | ||
7505 | NETIF_F_HW_CSUM); | ||
7506 | } | ||
7507 | return rc; | ||
7508 | } | ||
7509 | |||
7510 | static int | ||
7511 | qeth_start_ipassists(struct qeth_card *card) | ||
7512 | { | ||
7513 | QETH_DBF_TEXT(trace,3,"strtipas"); | ||
7514 | qeth_start_ipa_arp_processing(card); /* go on*/ | ||
7515 | qeth_start_ipa_ip_fragmentation(card); /* go on*/ | ||
7516 | qeth_start_ipa_source_mac(card); /* go on*/ | ||
7517 | qeth_start_ipa_vlan(card); /* go on*/ | ||
7518 | qeth_start_ipa_multicast(card); /* go on*/ | ||
7519 | qeth_start_ipa_ipv6(card); /* go on*/ | ||
7520 | qeth_start_ipa_broadcast(card); /* go on*/ | ||
7521 | qeth_start_ipa_checksum(card); /* go on*/ | ||
7522 | qeth_start_ipa_tso(card); /* go on*/ | ||
7523 | return 0; | ||
7524 | } | ||
7525 | |||
7526 | static int | ||
7527 | qeth_send_setrouting(struct qeth_card *card, enum qeth_routing_types type, | ||
7528 | enum qeth_prot_versions prot) | ||
7529 | { | ||
7530 | int rc; | ||
7531 | struct qeth_ipa_cmd *cmd; | ||
7532 | struct qeth_cmd_buffer *iob; | ||
7533 | |||
7534 | QETH_DBF_TEXT(trace,4,"setroutg"); | ||
7535 | iob = qeth_get_ipacmd_buffer(card, IPA_CMD_SETRTG, prot); | ||
7536 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
7537 | cmd->data.setrtg.type = (type); | ||
7538 | rc = qeth_send_ipa_cmd(card, iob, NULL, NULL); | ||
7539 | |||
7540 | return rc; | ||
7541 | |||
7542 | } | ||
7543 | |||
7544 | static void | ||
7545 | qeth_correct_routing_type(struct qeth_card *card, enum qeth_routing_types *type, | ||
7546 | enum qeth_prot_versions prot) | ||
7547 | { | ||
7548 | if (card->info.type == QETH_CARD_TYPE_IQD) { | ||
7549 | switch (*type) { | ||
7550 | case NO_ROUTER: | ||
7551 | case PRIMARY_CONNECTOR: | ||
7552 | case SECONDARY_CONNECTOR: | ||
7553 | case MULTICAST_ROUTER: | ||
7554 | return; | ||
7555 | default: | ||
7556 | goto out_inval; | ||
7557 | } | ||
7558 | } else { | ||
7559 | switch (*type) { | ||
7560 | case NO_ROUTER: | ||
7561 | case PRIMARY_ROUTER: | ||
7562 | case SECONDARY_ROUTER: | ||
7563 | return; | ||
7564 | case MULTICAST_ROUTER: | ||
7565 | if (qeth_is_ipafunc_supported(card, prot, | ||
7566 | IPA_OSA_MC_ROUTER)) | ||
7567 | return; | ||
7568 | default: | ||
7569 | goto out_inval; | ||
7570 | } | ||
7571 | } | ||
7572 | out_inval: | ||
7573 | PRINT_WARN("Routing type '%s' not supported for interface %s.\n" | ||
7574 | "Router status set to 'no router'.\n", | ||
7575 | ((*type == PRIMARY_ROUTER)? "primary router" : | ||
7576 | (*type == SECONDARY_ROUTER)? "secondary router" : | ||
7577 | (*type == PRIMARY_CONNECTOR)? "primary connector" : | ||
7578 | (*type == SECONDARY_CONNECTOR)? "secondary connector" : | ||
7579 | (*type == MULTICAST_ROUTER)? "multicast router" : | ||
7580 | "unknown"), | ||
7581 | card->dev->name); | ||
7582 | *type = NO_ROUTER; | ||
7583 | } | ||
7584 | |||
7585 | int | ||
7586 | qeth_setrouting_v4(struct qeth_card *card) | ||
7587 | { | ||
7588 | int rc; | ||
7589 | |||
7590 | QETH_DBF_TEXT(trace,3,"setrtg4"); | ||
7591 | |||
7592 | qeth_correct_routing_type(card, &card->options.route4.type, | ||
7593 | QETH_PROT_IPV4); | ||
7594 | |||
7595 | rc = qeth_send_setrouting(card, card->options.route4.type, | ||
7596 | QETH_PROT_IPV4); | ||
7597 | if (rc) { | ||
7598 | card->options.route4.type = NO_ROUTER; | ||
7599 | PRINT_WARN("Error (0x%04x) while setting routing type on %s. " | ||
7600 | "Type set to 'no router'.\n", | ||
7601 | rc, QETH_CARD_IFNAME(card)); | ||
7602 | } | ||
7603 | return rc; | ||
7604 | } | ||
7605 | |||
7606 | int | ||
7607 | qeth_setrouting_v6(struct qeth_card *card) | ||
7608 | { | ||
7609 | int rc = 0; | ||
7610 | |||
7611 | QETH_DBF_TEXT(trace,3,"setrtg6"); | ||
7612 | #ifdef CONFIG_QETH_IPV6 | ||
7613 | |||
7614 | if (!qeth_is_supported(card, IPA_IPV6)) | ||
7615 | return 0; | ||
7616 | qeth_correct_routing_type(card, &card->options.route6.type, | ||
7617 | QETH_PROT_IPV6); | ||
7618 | |||
7619 | rc = qeth_send_setrouting(card, card->options.route6.type, | ||
7620 | QETH_PROT_IPV6); | ||
7621 | if (rc) { | ||
7622 | card->options.route6.type = NO_ROUTER; | ||
7623 | PRINT_WARN("Error (0x%04x) while setting routing type on %s. " | ||
7624 | "Type set to 'no router'.\n", | ||
7625 | rc, QETH_CARD_IFNAME(card)); | ||
7626 | } | ||
7627 | #endif | ||
7628 | return rc; | ||
7629 | } | ||
7630 | |||
7631 | int | ||
7632 | qeth_set_large_send(struct qeth_card *card, enum qeth_large_send_types type) | ||
7633 | { | ||
7634 | int rc = 0; | ||
7635 | |||
7636 | if (card->dev == NULL) { | ||
7637 | card->options.large_send = type; | ||
7638 | return 0; | ||
7639 | } | ||
7640 | if (card->state == CARD_STATE_UP) | ||
7641 | netif_tx_disable(card->dev); | ||
7642 | card->options.large_send = type; | ||
7643 | switch (card->options.large_send) { | ||
7644 | case QETH_LARGE_SEND_EDDP: | ||
7645 | card->dev->features |= NETIF_F_TSO | NETIF_F_SG | | ||
7646 | NETIF_F_HW_CSUM; | ||
7647 | break; | ||
7648 | case QETH_LARGE_SEND_TSO: | ||
7649 | if (qeth_is_supported(card, IPA_OUTBOUND_TSO)){ | ||
7650 | card->dev->features |= NETIF_F_TSO | NETIF_F_SG | | ||
7651 | NETIF_F_HW_CSUM; | ||
7652 | } else { | ||
7653 | PRINT_WARN("TSO not supported on %s. " | ||
7654 | "large_send set to 'no'.\n", | ||
7655 | card->dev->name); | ||
7656 | card->dev->features &= ~(NETIF_F_TSO | NETIF_F_SG | | ||
7657 | NETIF_F_HW_CSUM); | ||
7658 | card->options.large_send = QETH_LARGE_SEND_NO; | ||
7659 | rc = -EOPNOTSUPP; | ||
7660 | } | ||
7661 | break; | ||
7662 | default: /* includes QETH_LARGE_SEND_NO */ | ||
7663 | card->dev->features &= ~(NETIF_F_TSO | NETIF_F_SG | | ||
7664 | NETIF_F_HW_CSUM); | ||
7665 | break; | ||
7666 | } | ||
7667 | if (card->state == CARD_STATE_UP) | ||
7668 | netif_wake_queue(card->dev); | ||
7669 | return rc; | ||
7670 | } | ||
7671 | |||
7672 | /* | ||
7673 | * softsetup card: init IPA stuff | ||
7674 | */ | ||
7675 | static int | ||
7676 | qeth_softsetup_card(struct qeth_card *card) | ||
7677 | { | ||
7678 | int rc; | ||
7679 | |||
7680 | QETH_DBF_TEXT(setup, 2, "softsetp"); | ||
7681 | |||
7682 | if ((rc = qeth_send_startlan(card, QETH_PROT_IPV4))){ | ||
7683 | QETH_DBF_TEXT_(setup, 2, "1err%d", rc); | ||
7684 | if (rc == 0xe080){ | ||
7685 | PRINT_WARN("LAN on card %s if offline! " | ||
7686 | "Waiting for STARTLAN from card.\n", | ||
7687 | CARD_BUS_ID(card)); | ||
7688 | card->lan_online = 0; | ||
7689 | } | ||
7690 | return rc; | ||
7691 | } else | ||
7692 | card->lan_online = 1; | ||
7693 | if (card->info.type==QETH_CARD_TYPE_OSN) | ||
7694 | goto out; | ||
7695 | qeth_set_large_send(card, card->options.large_send); | ||
7696 | if (card->options.layer2) { | ||
7697 | card->dev->features |= | ||
7698 | NETIF_F_HW_VLAN_FILTER | | ||
7699 | NETIF_F_HW_VLAN_TX | | ||
7700 | NETIF_F_HW_VLAN_RX; | ||
7701 | card->dev->flags|=IFF_MULTICAST|IFF_BROADCAST; | ||
7702 | card->info.broadcast_capable=1; | ||
7703 | if ((rc = qeth_layer2_initialize(card))) { | ||
7704 | QETH_DBF_TEXT_(setup, 2, "L2err%d", rc); | ||
7705 | return rc; | ||
7706 | } | ||
7707 | #ifdef CONFIG_QETH_VLAN | ||
7708 | qeth_layer2_process_vlans(card, 0); | ||
7709 | #endif | ||
7710 | goto out; | ||
7711 | } | ||
7712 | if ((rc = qeth_setadapter_parms(card))) | ||
7713 | QETH_DBF_TEXT_(setup, 2, "2err%d", rc); | ||
7714 | if ((rc = qeth_start_ipassists(card))) | ||
7715 | QETH_DBF_TEXT_(setup, 2, "3err%d", rc); | ||
7716 | if ((rc = qeth_setrouting_v4(card))) | ||
7717 | QETH_DBF_TEXT_(setup, 2, "4err%d", rc); | ||
7718 | if ((rc = qeth_setrouting_v6(card))) | ||
7719 | QETH_DBF_TEXT_(setup, 2, "5err%d", rc); | ||
7720 | out: | ||
7721 | netif_tx_disable(card->dev); | ||
7722 | return 0; | ||
7723 | } | ||
7724 | |||
7725 | #ifdef CONFIG_QETH_IPV6 | ||
7726 | static int | ||
7727 | qeth_get_unique_id_cb(struct qeth_card *card, struct qeth_reply *reply, | ||
7728 | unsigned long data) | ||
7729 | { | ||
7730 | struct qeth_ipa_cmd *cmd; | ||
7731 | |||
7732 | cmd = (struct qeth_ipa_cmd *) data; | ||
7733 | if (cmd->hdr.return_code == 0) | ||
7734 | card->info.unique_id = *((__u16 *) | ||
7735 | &cmd->data.create_destroy_addr.unique_id[6]); | ||
7736 | else { | ||
7737 | card->info.unique_id = UNIQUE_ID_IF_CREATE_ADDR_FAILED | | ||
7738 | UNIQUE_ID_NOT_BY_CARD; | ||
7739 | PRINT_WARN("couldn't get a unique id from the card on device " | ||
7740 | "%s (result=x%x), using default id. ipv6 " | ||
7741 | "autoconfig on other lpars may lead to duplicate " | ||
7742 | "ip addresses. please use manually " | ||
7743 | "configured ones.\n", | ||
7744 | CARD_BUS_ID(card), cmd->hdr.return_code); | ||
7745 | } | ||
7746 | return 0; | ||
7747 | } | ||
7748 | #endif | ||
7749 | |||
7750 | static int | ||
7751 | qeth_put_unique_id(struct qeth_card *card) | ||
7752 | { | ||
7753 | |||
7754 | int rc = 0; | ||
7755 | #ifdef CONFIG_QETH_IPV6 | ||
7756 | struct qeth_cmd_buffer *iob; | ||
7757 | struct qeth_ipa_cmd *cmd; | ||
7758 | |||
7759 | QETH_DBF_TEXT(trace,2,"puniqeid"); | ||
7760 | |||
7761 | if ((card->info.unique_id & UNIQUE_ID_NOT_BY_CARD) == | ||
7762 | UNIQUE_ID_NOT_BY_CARD) | ||
7763 | return -1; | ||
7764 | iob = qeth_get_ipacmd_buffer(card, IPA_CMD_DESTROY_ADDR, | ||
7765 | QETH_PROT_IPV6); | ||
7766 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
7767 | *((__u16 *) &cmd->data.create_destroy_addr.unique_id[6]) = | ||
7768 | card->info.unique_id; | ||
7769 | memcpy(&cmd->data.create_destroy_addr.unique_id[0], | ||
7770 | card->dev->dev_addr, OSA_ADDR_LEN); | ||
7771 | rc = qeth_send_ipa_cmd(card, iob, NULL, NULL); | ||
7772 | #else | ||
7773 | card->info.unique_id = UNIQUE_ID_IF_CREATE_ADDR_FAILED | | ||
7774 | UNIQUE_ID_NOT_BY_CARD; | ||
7775 | #endif | ||
7776 | return rc; | ||
7777 | } | ||
7778 | |||
7779 | /** | ||
7780 | * Clear IP List | ||
7781 | */ | ||
7782 | static void | ||
7783 | qeth_clear_ip_list(struct qeth_card *card, int clean, int recover) | ||
7784 | { | ||
7785 | struct qeth_ipaddr *addr, *tmp; | ||
7786 | unsigned long flags; | ||
7787 | |||
7788 | QETH_DBF_TEXT(trace,4,"clearip"); | ||
7789 | spin_lock_irqsave(&card->ip_lock, flags); | ||
7790 | /* clear todo list */ | ||
7791 | list_for_each_entry_safe(addr, tmp, card->ip_tbd_list, entry){ | ||
7792 | list_del(&addr->entry); | ||
7793 | kfree(addr); | ||
7794 | } | ||
7795 | |||
7796 | while (!list_empty(&card->ip_list)) { | ||
7797 | addr = list_entry(card->ip_list.next, | ||
7798 | struct qeth_ipaddr, entry); | ||
7799 | list_del_init(&addr->entry); | ||
7800 | if (clean) { | ||
7801 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
7802 | qeth_deregister_addr_entry(card, addr); | ||
7803 | spin_lock_irqsave(&card->ip_lock, flags); | ||
7804 | } | ||
7805 | if (!recover || addr->is_multicast) { | ||
7806 | kfree(addr); | ||
7807 | continue; | ||
7808 | } | ||
7809 | list_add_tail(&addr->entry, card->ip_tbd_list); | ||
7810 | } | ||
7811 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
7812 | } | ||
7813 | |||
7814 | static void | ||
7815 | qeth_set_allowed_threads(struct qeth_card *card, unsigned long threads, | ||
7816 | int clear_start_mask) | ||
7817 | { | ||
7818 | unsigned long flags; | ||
7819 | |||
7820 | spin_lock_irqsave(&card->thread_mask_lock, flags); | ||
7821 | card->thread_allowed_mask = threads; | ||
7822 | if (clear_start_mask) | ||
7823 | card->thread_start_mask &= threads; | ||
7824 | spin_unlock_irqrestore(&card->thread_mask_lock, flags); | ||
7825 | wake_up(&card->wait_q); | ||
7826 | } | ||
7827 | |||
7828 | static int | ||
7829 | qeth_threads_running(struct qeth_card *card, unsigned long threads) | ||
7830 | { | ||
7831 | unsigned long flags; | ||
7832 | int rc = 0; | ||
7833 | |||
7834 | spin_lock_irqsave(&card->thread_mask_lock, flags); | ||
7835 | rc = (card->thread_running_mask & threads); | ||
7836 | spin_unlock_irqrestore(&card->thread_mask_lock, flags); | ||
7837 | return rc; | ||
7838 | } | ||
7839 | |||
7840 | static int | ||
7841 | qeth_wait_for_threads(struct qeth_card *card, unsigned long threads) | ||
7842 | { | ||
7843 | return wait_event_interruptible(card->wait_q, | ||
7844 | qeth_threads_running(card, threads) == 0); | ||
7845 | } | ||
7846 | |||
7847 | static int | ||
7848 | qeth_stop_card(struct qeth_card *card, int recovery_mode) | ||
7849 | { | ||
7850 | int rc = 0; | ||
7851 | |||
7852 | QETH_DBF_TEXT(setup ,2,"stopcard"); | ||
7853 | QETH_DBF_HEX(setup, 2, &card, sizeof(void *)); | ||
7854 | |||
7855 | qeth_set_allowed_threads(card, 0, 1); | ||
7856 | if (qeth_wait_for_threads(card, ~QETH_RECOVER_THREAD)) | ||
7857 | return -ERESTARTSYS; | ||
7858 | if (card->read.state == CH_STATE_UP && | ||
7859 | card->write.state == CH_STATE_UP && | ||
7860 | (card->state == CARD_STATE_UP)) { | ||
7861 | if (recovery_mode && | ||
7862 | card->info.type != QETH_CARD_TYPE_OSN) { | ||
7863 | qeth_stop(card->dev); | ||
7864 | } else { | ||
7865 | rtnl_lock(); | ||
7866 | dev_close(card->dev); | ||
7867 | rtnl_unlock(); | ||
7868 | } | ||
7869 | if (!card->use_hard_stop) { | ||
7870 | __u8 *mac = &card->dev->dev_addr[0]; | ||
7871 | rc = qeth_layer2_send_delmac(card, mac); | ||
7872 | QETH_DBF_TEXT_(setup, 2, "Lerr%d", rc); | ||
7873 | if ((rc = qeth_send_stoplan(card))) | ||
7874 | QETH_DBF_TEXT_(setup, 2, "1err%d", rc); | ||
7875 | } | ||
7876 | card->state = CARD_STATE_SOFTSETUP; | ||
7877 | } | ||
7878 | if (card->state == CARD_STATE_SOFTSETUP) { | ||
7879 | #ifdef CONFIG_QETH_VLAN | ||
7880 | if (card->options.layer2) | ||
7881 | qeth_layer2_process_vlans(card, 1); | ||
7882 | #endif | ||
7883 | qeth_clear_ip_list(card, !card->use_hard_stop, 1); | ||
7884 | qeth_clear_ipacmd_list(card); | ||
7885 | card->state = CARD_STATE_HARDSETUP; | ||
7886 | } | ||
7887 | if (card->state == CARD_STATE_HARDSETUP) { | ||
7888 | if ((!card->use_hard_stop) && | ||
7889 | (!card->options.layer2)) | ||
7890 | if ((rc = qeth_put_unique_id(card))) | ||
7891 | QETH_DBF_TEXT_(setup, 2, "2err%d", rc); | ||
7892 | qeth_qdio_clear_card(card, 0); | ||
7893 | qeth_clear_qdio_buffers(card); | ||
7894 | qeth_clear_working_pool_list(card); | ||
7895 | card->state = CARD_STATE_DOWN; | ||
7896 | } | ||
7897 | if (card->state == CARD_STATE_DOWN) { | ||
7898 | qeth_clear_cmd_buffers(&card->read); | ||
7899 | qeth_clear_cmd_buffers(&card->write); | ||
7900 | } | ||
7901 | card->use_hard_stop = 0; | ||
7902 | return rc; | ||
7903 | } | ||
7904 | |||
7905 | |||
7906 | static int | ||
7907 | qeth_get_unique_id(struct qeth_card *card) | ||
7908 | { | ||
7909 | int rc = 0; | ||
7910 | #ifdef CONFIG_QETH_IPV6 | ||
7911 | struct qeth_cmd_buffer *iob; | ||
7912 | struct qeth_ipa_cmd *cmd; | ||
7913 | |||
7914 | QETH_DBF_TEXT(setup, 2, "guniqeid"); | ||
7915 | |||
7916 | if (!qeth_is_supported(card,IPA_IPV6)) { | ||
7917 | card->info.unique_id = UNIQUE_ID_IF_CREATE_ADDR_FAILED | | ||
7918 | UNIQUE_ID_NOT_BY_CARD; | ||
7919 | return 0; | ||
7920 | } | ||
7921 | |||
7922 | iob = qeth_get_ipacmd_buffer(card, IPA_CMD_CREATE_ADDR, | ||
7923 | QETH_PROT_IPV6); | ||
7924 | cmd = (struct qeth_ipa_cmd *)(iob->data+IPA_PDU_HEADER_SIZE); | ||
7925 | *((__u16 *) &cmd->data.create_destroy_addr.unique_id[6]) = | ||
7926 | card->info.unique_id; | ||
7927 | |||
7928 | rc = qeth_send_ipa_cmd(card, iob, qeth_get_unique_id_cb, NULL); | ||
7929 | #else | ||
7930 | card->info.unique_id = UNIQUE_ID_IF_CREATE_ADDR_FAILED | | ||
7931 | UNIQUE_ID_NOT_BY_CARD; | ||
7932 | #endif | ||
7933 | return rc; | ||
7934 | } | ||
7935 | static void | ||
7936 | qeth_print_status_with_portname(struct qeth_card *card) | ||
7937 | { | ||
7938 | char dbf_text[15]; | ||
7939 | int i; | ||
7940 | |||
7941 | sprintf(dbf_text, "%s", card->info.portname + 1); | ||
7942 | for (i = 0; i < 8; i++) | ||
7943 | dbf_text[i] = | ||
7944 | (char) _ebcasc[(__u8) dbf_text[i]]; | ||
7945 | dbf_text[8] = 0; | ||
7946 | printk("qeth: Device %s/%s/%s is a%s card%s%s%s\n" | ||
7947 | "with link type %s (portname: %s)\n", | ||
7948 | CARD_RDEV_ID(card), | ||
7949 | CARD_WDEV_ID(card), | ||
7950 | CARD_DDEV_ID(card), | ||
7951 | qeth_get_cardname(card), | ||
7952 | (card->info.mcl_level[0]) ? " (level: " : "", | ||
7953 | (card->info.mcl_level[0]) ? card->info.mcl_level : "", | ||
7954 | (card->info.mcl_level[0]) ? ")" : "", | ||
7955 | qeth_get_cardname_short(card), | ||
7956 | dbf_text); | ||
7957 | |||
7958 | } | ||
7959 | |||
7960 | static void | ||
7961 | qeth_print_status_no_portname(struct qeth_card *card) | ||
7962 | { | ||
7963 | if (card->info.portname[0]) | ||
7964 | printk("qeth: Device %s/%s/%s is a%s " | ||
7965 | "card%s%s%s\nwith link type %s " | ||
7966 | "(no portname needed by interface).\n", | ||
7967 | CARD_RDEV_ID(card), | ||
7968 | CARD_WDEV_ID(card), | ||
7969 | CARD_DDEV_ID(card), | ||
7970 | qeth_get_cardname(card), | ||
7971 | (card->info.mcl_level[0]) ? " (level: " : "", | ||
7972 | (card->info.mcl_level[0]) ? card->info.mcl_level : "", | ||
7973 | (card->info.mcl_level[0]) ? ")" : "", | ||
7974 | qeth_get_cardname_short(card)); | ||
7975 | else | ||
7976 | printk("qeth: Device %s/%s/%s is a%s " | ||
7977 | "card%s%s%s\nwith link type %s.\n", | ||
7978 | CARD_RDEV_ID(card), | ||
7979 | CARD_WDEV_ID(card), | ||
7980 | CARD_DDEV_ID(card), | ||
7981 | qeth_get_cardname(card), | ||
7982 | (card->info.mcl_level[0]) ? " (level: " : "", | ||
7983 | (card->info.mcl_level[0]) ? card->info.mcl_level : "", | ||
7984 | (card->info.mcl_level[0]) ? ")" : "", | ||
7985 | qeth_get_cardname_short(card)); | ||
7986 | } | ||
7987 | |||
7988 | static void | ||
7989 | qeth_print_status_message(struct qeth_card *card) | ||
7990 | { | ||
7991 | switch (card->info.type) { | ||
7992 | case QETH_CARD_TYPE_OSAE: | ||
7993 | /* VM will use a non-zero first character | ||
7994 | * to indicate a HiperSockets like reporting | ||
7995 | * of the level OSA sets the first character to zero | ||
7996 | * */ | ||
7997 | if (!card->info.mcl_level[0]) { | ||
7998 | sprintf(card->info.mcl_level,"%02x%02x", | ||
7999 | card->info.mcl_level[2], | ||
8000 | card->info.mcl_level[3]); | ||
8001 | |||
8002 | card->info.mcl_level[QETH_MCL_LENGTH] = 0; | ||
8003 | break; | ||
8004 | } | ||
8005 | /* fallthrough */ | ||
8006 | case QETH_CARD_TYPE_IQD: | ||
8007 | if (card->info.guestlan) { | ||
8008 | card->info.mcl_level[0] = (char) _ebcasc[(__u8) | ||
8009 | card->info.mcl_level[0]]; | ||
8010 | card->info.mcl_level[1] = (char) _ebcasc[(__u8) | ||
8011 | card->info.mcl_level[1]]; | ||
8012 | card->info.mcl_level[2] = (char) _ebcasc[(__u8) | ||
8013 | card->info.mcl_level[2]]; | ||
8014 | card->info.mcl_level[3] = (char) _ebcasc[(__u8) | ||
8015 | card->info.mcl_level[3]]; | ||
8016 | card->info.mcl_level[QETH_MCL_LENGTH] = 0; | ||
8017 | } | ||
8018 | break; | ||
8019 | default: | ||
8020 | memset(&card->info.mcl_level[0], 0, QETH_MCL_LENGTH + 1); | ||
8021 | } | ||
8022 | if (card->info.portname_required) | ||
8023 | qeth_print_status_with_portname(card); | ||
8024 | else | ||
8025 | qeth_print_status_no_portname(card); | ||
8026 | } | ||
8027 | |||
8028 | static int | ||
8029 | qeth_register_netdev(struct qeth_card *card) | ||
8030 | { | ||
8031 | QETH_DBF_TEXT(setup, 3, "regnetd"); | ||
8032 | if (card->dev->reg_state != NETREG_UNINITIALIZED) | ||
8033 | return 0; | ||
8034 | /* sysfs magic */ | ||
8035 | SET_NETDEV_DEV(card->dev, &card->gdev->dev); | ||
8036 | return register_netdev(card->dev); | ||
8037 | } | ||
8038 | |||
8039 | static void | ||
8040 | qeth_start_again(struct qeth_card *card, int recovery_mode) | ||
8041 | { | ||
8042 | QETH_DBF_TEXT(setup ,2, "startag"); | ||
8043 | |||
8044 | if (recovery_mode && | ||
8045 | card->info.type != QETH_CARD_TYPE_OSN) { | ||
8046 | qeth_open(card->dev); | ||
8047 | } else { | ||
8048 | rtnl_lock(); | ||
8049 | dev_open(card->dev); | ||
8050 | rtnl_unlock(); | ||
8051 | } | ||
8052 | /* this also sets saved unicast addresses */ | ||
8053 | qeth_set_multicast_list(card->dev); | ||
8054 | } | ||
8055 | |||
8056 | |||
8057 | /* Layer 2 specific stuff */ | ||
8058 | #define IGNORE_PARAM_EQ(option,value,reset_value,msg) \ | ||
8059 | if (card->options.option == value) { \ | ||
8060 | PRINT_ERR("%s not supported with layer 2 " \ | ||
8061 | "functionality, ignoring option on read" \ | ||
8062 | "channel device %s .\n",msg,CARD_RDEV_ID(card)); \ | ||
8063 | card->options.option = reset_value; \ | ||
8064 | } | ||
8065 | #define IGNORE_PARAM_NEQ(option,value,reset_value,msg) \ | ||
8066 | if (card->options.option != value) { \ | ||
8067 | PRINT_ERR("%s not supported with layer 2 " \ | ||
8068 | "functionality, ignoring option on read" \ | ||
8069 | "channel device %s .\n",msg,CARD_RDEV_ID(card)); \ | ||
8070 | card->options.option = reset_value; \ | ||
8071 | } | ||
8072 | |||
8073 | |||
8074 | static void qeth_make_parameters_consistent(struct qeth_card *card) | ||
8075 | { | ||
8076 | |||
8077 | if (card->options.layer2 == 0) | ||
8078 | return; | ||
8079 | if (card->info.type == QETH_CARD_TYPE_OSN) | ||
8080 | return; | ||
8081 | if (card->info.type == QETH_CARD_TYPE_IQD) { | ||
8082 | PRINT_ERR("Device %s does not support layer 2 functionality." \ | ||
8083 | " Ignoring layer2 option.\n",CARD_BUS_ID(card)); | ||
8084 | card->options.layer2 = 0; | ||
8085 | return; | ||
8086 | } | ||
8087 | IGNORE_PARAM_NEQ(route4.type, NO_ROUTER, NO_ROUTER, | ||
8088 | "Routing options are"); | ||
8089 | #ifdef CONFIG_QETH_IPV6 | ||
8090 | IGNORE_PARAM_NEQ(route6.type, NO_ROUTER, NO_ROUTER, | ||
8091 | "Routing options are"); | ||
8092 | #endif | ||
8093 | IGNORE_PARAM_EQ(checksum_type, HW_CHECKSUMMING, | ||
8094 | QETH_CHECKSUM_DEFAULT, | ||
8095 | "Checksumming options are"); | ||
8096 | IGNORE_PARAM_NEQ(broadcast_mode, QETH_TR_BROADCAST_ALLRINGS, | ||
8097 | QETH_TR_BROADCAST_ALLRINGS, | ||
8098 | "Broadcast mode options are"); | ||
8099 | IGNORE_PARAM_NEQ(macaddr_mode, QETH_TR_MACADDR_NONCANONICAL, | ||
8100 | QETH_TR_MACADDR_NONCANONICAL, | ||
8101 | "Canonical MAC addr options are"); | ||
8102 | IGNORE_PARAM_NEQ(fake_broadcast, 0, 0, | ||
8103 | "Broadcast faking options are"); | ||
8104 | IGNORE_PARAM_NEQ(add_hhlen, DEFAULT_ADD_HHLEN, | ||
8105 | DEFAULT_ADD_HHLEN,"Option add_hhlen is"); | ||
8106 | IGNORE_PARAM_NEQ(fake_ll, 0, 0,"Option fake_ll is"); | ||
8107 | } | ||
8108 | |||
8109 | |||
8110 | static int | ||
8111 | __qeth_set_online(struct ccwgroup_device *gdev, int recovery_mode) | ||
8112 | { | ||
8113 | struct qeth_card *card = gdev->dev.driver_data; | ||
8114 | int rc = 0; | ||
8115 | enum qeth_card_states recover_flag; | ||
8116 | |||
8117 | BUG_ON(!card); | ||
8118 | QETH_DBF_TEXT(setup ,2, "setonlin"); | ||
8119 | QETH_DBF_HEX(setup, 2, &card, sizeof(void *)); | ||
8120 | |||
8121 | qeth_set_allowed_threads(card, QETH_RECOVER_THREAD, 1); | ||
8122 | if (qeth_wait_for_threads(card, ~QETH_RECOVER_THREAD)){ | ||
8123 | PRINT_WARN("set_online of card %s interrupted by user!\n", | ||
8124 | CARD_BUS_ID(card)); | ||
8125 | return -ERESTARTSYS; | ||
8126 | } | ||
8127 | |||
8128 | recover_flag = card->state; | ||
8129 | if ((rc = ccw_device_set_online(CARD_RDEV(card))) || | ||
8130 | (rc = ccw_device_set_online(CARD_WDEV(card))) || | ||
8131 | (rc = ccw_device_set_online(CARD_DDEV(card)))){ | ||
8132 | QETH_DBF_TEXT_(setup, 2, "1err%d", rc); | ||
8133 | return -EIO; | ||
8134 | } | ||
8135 | |||
8136 | qeth_make_parameters_consistent(card); | ||
8137 | |||
8138 | if ((rc = qeth_hardsetup_card(card))){ | ||
8139 | QETH_DBF_TEXT_(setup, 2, "2err%d", rc); | ||
8140 | goto out_remove; | ||
8141 | } | ||
8142 | card->state = CARD_STATE_HARDSETUP; | ||
8143 | |||
8144 | if (!(rc = qeth_query_ipassists(card,QETH_PROT_IPV4))) | ||
8145 | rc = qeth_get_unique_id(card); | ||
8146 | |||
8147 | if (rc && card->options.layer2 == 0) { | ||
8148 | QETH_DBF_TEXT_(setup, 2, "3err%d", rc); | ||
8149 | goto out_remove; | ||
8150 | } | ||
8151 | qeth_print_status_message(card); | ||
8152 | if ((rc = qeth_register_netdev(card))){ | ||
8153 | QETH_DBF_TEXT_(setup, 2, "4err%d", rc); | ||
8154 | goto out_remove; | ||
8155 | } | ||
8156 | if ((rc = qeth_softsetup_card(card))){ | ||
8157 | QETH_DBF_TEXT_(setup, 2, "5err%d", rc); | ||
8158 | goto out_remove; | ||
8159 | } | ||
8160 | |||
8161 | if ((rc = qeth_init_qdio_queues(card))){ | ||
8162 | QETH_DBF_TEXT_(setup, 2, "6err%d", rc); | ||
8163 | goto out_remove; | ||
8164 | } | ||
8165 | card->state = CARD_STATE_SOFTSETUP; | ||
8166 | netif_carrier_on(card->dev); | ||
8167 | |||
8168 | qeth_set_allowed_threads(card, 0xffffffff, 0); | ||
8169 | if (recover_flag == CARD_STATE_RECOVER) | ||
8170 | qeth_start_again(card, recovery_mode); | ||
8171 | qeth_notify_processes(); | ||
8172 | return 0; | ||
8173 | out_remove: | ||
8174 | card->use_hard_stop = 1; | ||
8175 | qeth_stop_card(card, 0); | ||
8176 | ccw_device_set_offline(CARD_DDEV(card)); | ||
8177 | ccw_device_set_offline(CARD_WDEV(card)); | ||
8178 | ccw_device_set_offline(CARD_RDEV(card)); | ||
8179 | if (recover_flag == CARD_STATE_RECOVER) | ||
8180 | card->state = CARD_STATE_RECOVER; | ||
8181 | else | ||
8182 | card->state = CARD_STATE_DOWN; | ||
8183 | return -ENODEV; | ||
8184 | } | ||
8185 | |||
8186 | static int | ||
8187 | qeth_set_online(struct ccwgroup_device *gdev) | ||
8188 | { | ||
8189 | return __qeth_set_online(gdev, 0); | ||
8190 | } | ||
8191 | |||
8192 | static struct ccw_device_id qeth_ids[] = { | ||
8193 | {CCW_DEVICE(0x1731, 0x01), .driver_info = QETH_CARD_TYPE_OSAE}, | ||
8194 | {CCW_DEVICE(0x1731, 0x05), .driver_info = QETH_CARD_TYPE_IQD}, | ||
8195 | {CCW_DEVICE(0x1731, 0x06), .driver_info = QETH_CARD_TYPE_OSN}, | ||
8196 | {}, | ||
8197 | }; | ||
8198 | MODULE_DEVICE_TABLE(ccw, qeth_ids); | ||
8199 | |||
8200 | struct device *qeth_root_dev = NULL; | ||
8201 | |||
8202 | struct ccwgroup_driver qeth_ccwgroup_driver = { | ||
8203 | .owner = THIS_MODULE, | ||
8204 | .name = "qeth", | ||
8205 | .driver_id = 0xD8C5E3C8, | ||
8206 | .probe = qeth_probe_device, | ||
8207 | .remove = qeth_remove_device, | ||
8208 | .set_online = qeth_set_online, | ||
8209 | .set_offline = qeth_set_offline, | ||
8210 | }; | ||
8211 | |||
8212 | struct ccw_driver qeth_ccw_driver = { | ||
8213 | .name = "qeth", | ||
8214 | .ids = qeth_ids, | ||
8215 | .probe = ccwgroup_probe_ccwdev, | ||
8216 | .remove = ccwgroup_remove_ccwdev, | ||
8217 | }; | ||
8218 | |||
8219 | |||
8220 | static void | ||
8221 | qeth_unregister_dbf_views(void) | ||
8222 | { | ||
8223 | if (qeth_dbf_setup) | ||
8224 | debug_unregister(qeth_dbf_setup); | ||
8225 | if (qeth_dbf_qerr) | ||
8226 | debug_unregister(qeth_dbf_qerr); | ||
8227 | if (qeth_dbf_sense) | ||
8228 | debug_unregister(qeth_dbf_sense); | ||
8229 | if (qeth_dbf_misc) | ||
8230 | debug_unregister(qeth_dbf_misc); | ||
8231 | if (qeth_dbf_data) | ||
8232 | debug_unregister(qeth_dbf_data); | ||
8233 | if (qeth_dbf_control) | ||
8234 | debug_unregister(qeth_dbf_control); | ||
8235 | if (qeth_dbf_trace) | ||
8236 | debug_unregister(qeth_dbf_trace); | ||
8237 | } | ||
8238 | static int | ||
8239 | qeth_register_dbf_views(void) | ||
8240 | { | ||
8241 | qeth_dbf_setup = debug_register(QETH_DBF_SETUP_NAME, | ||
8242 | QETH_DBF_SETUP_PAGES, | ||
8243 | QETH_DBF_SETUP_NR_AREAS, | ||
8244 | QETH_DBF_SETUP_LEN); | ||
8245 | qeth_dbf_misc = debug_register(QETH_DBF_MISC_NAME, | ||
8246 | QETH_DBF_MISC_PAGES, | ||
8247 | QETH_DBF_MISC_NR_AREAS, | ||
8248 | QETH_DBF_MISC_LEN); | ||
8249 | qeth_dbf_data = debug_register(QETH_DBF_DATA_NAME, | ||
8250 | QETH_DBF_DATA_PAGES, | ||
8251 | QETH_DBF_DATA_NR_AREAS, | ||
8252 | QETH_DBF_DATA_LEN); | ||
8253 | qeth_dbf_control = debug_register(QETH_DBF_CONTROL_NAME, | ||
8254 | QETH_DBF_CONTROL_PAGES, | ||
8255 | QETH_DBF_CONTROL_NR_AREAS, | ||
8256 | QETH_DBF_CONTROL_LEN); | ||
8257 | qeth_dbf_sense = debug_register(QETH_DBF_SENSE_NAME, | ||
8258 | QETH_DBF_SENSE_PAGES, | ||
8259 | QETH_DBF_SENSE_NR_AREAS, | ||
8260 | QETH_DBF_SENSE_LEN); | ||
8261 | qeth_dbf_qerr = debug_register(QETH_DBF_QERR_NAME, | ||
8262 | QETH_DBF_QERR_PAGES, | ||
8263 | QETH_DBF_QERR_NR_AREAS, | ||
8264 | QETH_DBF_QERR_LEN); | ||
8265 | qeth_dbf_trace = debug_register(QETH_DBF_TRACE_NAME, | ||
8266 | QETH_DBF_TRACE_PAGES, | ||
8267 | QETH_DBF_TRACE_NR_AREAS, | ||
8268 | QETH_DBF_TRACE_LEN); | ||
8269 | |||
8270 | if ((qeth_dbf_setup == NULL) || (qeth_dbf_misc == NULL) || | ||
8271 | (qeth_dbf_data == NULL) || (qeth_dbf_control == NULL) || | ||
8272 | (qeth_dbf_sense == NULL) || (qeth_dbf_qerr == NULL) || | ||
8273 | (qeth_dbf_trace == NULL)) { | ||
8274 | qeth_unregister_dbf_views(); | ||
8275 | return -ENOMEM; | ||
8276 | } | ||
8277 | debug_register_view(qeth_dbf_setup, &debug_hex_ascii_view); | ||
8278 | debug_set_level(qeth_dbf_setup, QETH_DBF_SETUP_LEVEL); | ||
8279 | |||
8280 | debug_register_view(qeth_dbf_misc, &debug_hex_ascii_view); | ||
8281 | debug_set_level(qeth_dbf_misc, QETH_DBF_MISC_LEVEL); | ||
8282 | |||
8283 | debug_register_view(qeth_dbf_data, &debug_hex_ascii_view); | ||
8284 | debug_set_level(qeth_dbf_data, QETH_DBF_DATA_LEVEL); | ||
8285 | |||
8286 | debug_register_view(qeth_dbf_control, &debug_hex_ascii_view); | ||
8287 | debug_set_level(qeth_dbf_control, QETH_DBF_CONTROL_LEVEL); | ||
8288 | |||
8289 | debug_register_view(qeth_dbf_sense, &debug_hex_ascii_view); | ||
8290 | debug_set_level(qeth_dbf_sense, QETH_DBF_SENSE_LEVEL); | ||
8291 | |||
8292 | debug_register_view(qeth_dbf_qerr, &debug_hex_ascii_view); | ||
8293 | debug_set_level(qeth_dbf_qerr, QETH_DBF_QERR_LEVEL); | ||
8294 | |||
8295 | debug_register_view(qeth_dbf_trace, &debug_hex_ascii_view); | ||
8296 | debug_set_level(qeth_dbf_trace, QETH_DBF_TRACE_LEVEL); | ||
8297 | |||
8298 | return 0; | ||
8299 | } | ||
8300 | |||
8301 | #ifdef CONFIG_QETH_IPV6 | ||
8302 | extern struct neigh_table arp_tbl; | ||
8303 | static struct neigh_ops *arp_direct_ops; | ||
8304 | static int (*qeth_old_arp_constructor) (struct neighbour *); | ||
8305 | |||
8306 | static struct neigh_ops arp_direct_ops_template = { | ||
8307 | .family = AF_INET, | ||
8308 | .solicit = NULL, | ||
8309 | .error_report = NULL, | ||
8310 | .output = dev_queue_xmit, | ||
8311 | .connected_output = dev_queue_xmit, | ||
8312 | .hh_output = dev_queue_xmit, | ||
8313 | .queue_xmit = dev_queue_xmit | ||
8314 | }; | ||
8315 | |||
8316 | static int | ||
8317 | qeth_arp_constructor(struct neighbour *neigh) | ||
8318 | { | ||
8319 | struct net_device *dev = neigh->dev; | ||
8320 | struct in_device *in_dev; | ||
8321 | struct neigh_parms *parms; | ||
8322 | struct qeth_card *card; | ||
8323 | |||
8324 | card = qeth_get_card_from_dev(dev); | ||
8325 | if (card == NULL) | ||
8326 | goto out; | ||
8327 | if((card->options.layer2) || | ||
8328 | (card->dev->header_ops == &qeth_fake_ops)) | ||
8329 | goto out; | ||
8330 | |||
8331 | rcu_read_lock(); | ||
8332 | in_dev = __in_dev_get_rcu(dev); | ||
8333 | if (in_dev == NULL) { | ||
8334 | rcu_read_unlock(); | ||
8335 | return -EINVAL; | ||
8336 | } | ||
8337 | |||
8338 | parms = in_dev->arp_parms; | ||
8339 | __neigh_parms_put(neigh->parms); | ||
8340 | neigh->parms = neigh_parms_clone(parms); | ||
8341 | rcu_read_unlock(); | ||
8342 | |||
8343 | neigh->type = inet_addr_type(&init_net, *(__be32 *) neigh->primary_key); | ||
8344 | neigh->nud_state = NUD_NOARP; | ||
8345 | neigh->ops = arp_direct_ops; | ||
8346 | neigh->output = neigh->ops->queue_xmit; | ||
8347 | return 0; | ||
8348 | out: | ||
8349 | return qeth_old_arp_constructor(neigh); | ||
8350 | } | ||
8351 | #endif /*CONFIG_QETH_IPV6*/ | ||
8352 | |||
8353 | /* | ||
8354 | * IP address takeover related functions | ||
8355 | */ | ||
8356 | static void | ||
8357 | qeth_clear_ipato_list(struct qeth_card *card) | ||
8358 | { | ||
8359 | struct qeth_ipato_entry *ipatoe, *tmp; | ||
8360 | unsigned long flags; | ||
8361 | |||
8362 | spin_lock_irqsave(&card->ip_lock, flags); | ||
8363 | list_for_each_entry_safe(ipatoe, tmp, &card->ipato.entries, entry) { | ||
8364 | list_del(&ipatoe->entry); | ||
8365 | kfree(ipatoe); | ||
8366 | } | ||
8367 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
8368 | } | ||
8369 | |||
8370 | int | ||
8371 | qeth_add_ipato_entry(struct qeth_card *card, struct qeth_ipato_entry *new) | ||
8372 | { | ||
8373 | struct qeth_ipato_entry *ipatoe; | ||
8374 | unsigned long flags; | ||
8375 | int rc = 0; | ||
8376 | |||
8377 | QETH_DBF_TEXT(trace, 2, "addipato"); | ||
8378 | spin_lock_irqsave(&card->ip_lock, flags); | ||
8379 | list_for_each_entry(ipatoe, &card->ipato.entries, entry){ | ||
8380 | if (ipatoe->proto != new->proto) | ||
8381 | continue; | ||
8382 | if (!memcmp(ipatoe->addr, new->addr, | ||
8383 | (ipatoe->proto == QETH_PROT_IPV4)? 4:16) && | ||
8384 | (ipatoe->mask_bits == new->mask_bits)){ | ||
8385 | PRINT_WARN("ipato entry already exists!\n"); | ||
8386 | rc = -EEXIST; | ||
8387 | break; | ||
8388 | } | ||
8389 | } | ||
8390 | if (!rc) { | ||
8391 | list_add_tail(&new->entry, &card->ipato.entries); | ||
8392 | } | ||
8393 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
8394 | return rc; | ||
8395 | } | ||
8396 | |||
8397 | void | ||
8398 | qeth_del_ipato_entry(struct qeth_card *card, enum qeth_prot_versions proto, | ||
8399 | u8 *addr, int mask_bits) | ||
8400 | { | ||
8401 | struct qeth_ipato_entry *ipatoe, *tmp; | ||
8402 | unsigned long flags; | ||
8403 | |||
8404 | QETH_DBF_TEXT(trace, 2, "delipato"); | ||
8405 | spin_lock_irqsave(&card->ip_lock, flags); | ||
8406 | list_for_each_entry_safe(ipatoe, tmp, &card->ipato.entries, entry){ | ||
8407 | if (ipatoe->proto != proto) | ||
8408 | continue; | ||
8409 | if (!memcmp(ipatoe->addr, addr, | ||
8410 | (proto == QETH_PROT_IPV4)? 4:16) && | ||
8411 | (ipatoe->mask_bits == mask_bits)){ | ||
8412 | list_del(&ipatoe->entry); | ||
8413 | kfree(ipatoe); | ||
8414 | } | ||
8415 | } | ||
8416 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
8417 | } | ||
8418 | |||
8419 | static void | ||
8420 | qeth_convert_addr_to_bits(u8 *addr, u8 *bits, int len) | ||
8421 | { | ||
8422 | int i, j; | ||
8423 | u8 octet; | ||
8424 | |||
8425 | for (i = 0; i < len; ++i){ | ||
8426 | octet = addr[i]; | ||
8427 | for (j = 7; j >= 0; --j){ | ||
8428 | bits[i*8 + j] = octet & 1; | ||
8429 | octet >>= 1; | ||
8430 | } | ||
8431 | } | ||
8432 | } | ||
8433 | |||
8434 | static int | ||
8435 | qeth_is_addr_covered_by_ipato(struct qeth_card *card, struct qeth_ipaddr *addr) | ||
8436 | { | ||
8437 | struct qeth_ipato_entry *ipatoe; | ||
8438 | u8 addr_bits[128] = {0, }; | ||
8439 | u8 ipatoe_bits[128] = {0, }; | ||
8440 | int rc = 0; | ||
8441 | |||
8442 | if (!card->ipato.enabled) | ||
8443 | return 0; | ||
8444 | |||
8445 | qeth_convert_addr_to_bits((u8 *) &addr->u, addr_bits, | ||
8446 | (addr->proto == QETH_PROT_IPV4)? 4:16); | ||
8447 | list_for_each_entry(ipatoe, &card->ipato.entries, entry){ | ||
8448 | if (addr->proto != ipatoe->proto) | ||
8449 | continue; | ||
8450 | qeth_convert_addr_to_bits(ipatoe->addr, ipatoe_bits, | ||
8451 | (ipatoe->proto==QETH_PROT_IPV4) ? | ||
8452 | 4:16); | ||
8453 | if (addr->proto == QETH_PROT_IPV4) | ||
8454 | rc = !memcmp(addr_bits, ipatoe_bits, | ||
8455 | min(32, ipatoe->mask_bits)); | ||
8456 | else | ||
8457 | rc = !memcmp(addr_bits, ipatoe_bits, | ||
8458 | min(128, ipatoe->mask_bits)); | ||
8459 | if (rc) | ||
8460 | break; | ||
8461 | } | ||
8462 | /* invert? */ | ||
8463 | if ((addr->proto == QETH_PROT_IPV4) && card->ipato.invert4) | ||
8464 | rc = !rc; | ||
8465 | else if ((addr->proto == QETH_PROT_IPV6) && card->ipato.invert6) | ||
8466 | rc = !rc; | ||
8467 | |||
8468 | return rc; | ||
8469 | } | ||
8470 | |||
8471 | /* | ||
8472 | * VIPA related functions | ||
8473 | */ | ||
8474 | int | ||
8475 | qeth_add_vipa(struct qeth_card *card, enum qeth_prot_versions proto, | ||
8476 | const u8 *addr) | ||
8477 | { | ||
8478 | struct qeth_ipaddr *ipaddr; | ||
8479 | unsigned long flags; | ||
8480 | int rc = 0; | ||
8481 | |||
8482 | ipaddr = qeth_get_addr_buffer(proto); | ||
8483 | if (ipaddr){ | ||
8484 | if (proto == QETH_PROT_IPV4){ | ||
8485 | QETH_DBF_TEXT(trace, 2, "addvipa4"); | ||
8486 | memcpy(&ipaddr->u.a4.addr, addr, 4); | ||
8487 | ipaddr->u.a4.mask = 0; | ||
8488 | #ifdef CONFIG_QETH_IPV6 | ||
8489 | } else if (proto == QETH_PROT_IPV6){ | ||
8490 | QETH_DBF_TEXT(trace, 2, "addvipa6"); | ||
8491 | memcpy(&ipaddr->u.a6.addr, addr, 16); | ||
8492 | ipaddr->u.a6.pfxlen = 0; | ||
8493 | #endif | ||
8494 | } | ||
8495 | ipaddr->type = QETH_IP_TYPE_VIPA; | ||
8496 | ipaddr->set_flags = QETH_IPA_SETIP_VIPA_FLAG; | ||
8497 | ipaddr->del_flags = QETH_IPA_DELIP_VIPA_FLAG; | ||
8498 | } else | ||
8499 | return -ENOMEM; | ||
8500 | spin_lock_irqsave(&card->ip_lock, flags); | ||
8501 | if (__qeth_address_exists_in_list(&card->ip_list, ipaddr, 0) || | ||
8502 | __qeth_address_exists_in_list(card->ip_tbd_list, ipaddr, 0)) | ||
8503 | rc = -EEXIST; | ||
8504 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
8505 | if (rc){ | ||
8506 | PRINT_WARN("Cannot add VIPA. Address already exists!\n"); | ||
8507 | return rc; | ||
8508 | } | ||
8509 | if (!qeth_add_ip(card, ipaddr)) | ||
8510 | kfree(ipaddr); | ||
8511 | qeth_set_ip_addr_list(card); | ||
8512 | return rc; | ||
8513 | } | ||
8514 | |||
8515 | void | ||
8516 | qeth_del_vipa(struct qeth_card *card, enum qeth_prot_versions proto, | ||
8517 | const u8 *addr) | ||
8518 | { | ||
8519 | struct qeth_ipaddr *ipaddr; | ||
8520 | |||
8521 | ipaddr = qeth_get_addr_buffer(proto); | ||
8522 | if (ipaddr){ | ||
8523 | if (proto == QETH_PROT_IPV4){ | ||
8524 | QETH_DBF_TEXT(trace, 2, "delvipa4"); | ||
8525 | memcpy(&ipaddr->u.a4.addr, addr, 4); | ||
8526 | ipaddr->u.a4.mask = 0; | ||
8527 | #ifdef CONFIG_QETH_IPV6 | ||
8528 | } else if (proto == QETH_PROT_IPV6){ | ||
8529 | QETH_DBF_TEXT(trace, 2, "delvipa6"); | ||
8530 | memcpy(&ipaddr->u.a6.addr, addr, 16); | ||
8531 | ipaddr->u.a6.pfxlen = 0; | ||
8532 | #endif | ||
8533 | } | ||
8534 | ipaddr->type = QETH_IP_TYPE_VIPA; | ||
8535 | } else | ||
8536 | return; | ||
8537 | if (!qeth_delete_ip(card, ipaddr)) | ||
8538 | kfree(ipaddr); | ||
8539 | qeth_set_ip_addr_list(card); | ||
8540 | } | ||
8541 | |||
8542 | /* | ||
8543 | * proxy ARP related functions | ||
8544 | */ | ||
8545 | int | ||
8546 | qeth_add_rxip(struct qeth_card *card, enum qeth_prot_versions proto, | ||
8547 | const u8 *addr) | ||
8548 | { | ||
8549 | struct qeth_ipaddr *ipaddr; | ||
8550 | unsigned long flags; | ||
8551 | int rc = 0; | ||
8552 | |||
8553 | ipaddr = qeth_get_addr_buffer(proto); | ||
8554 | if (ipaddr){ | ||
8555 | if (proto == QETH_PROT_IPV4){ | ||
8556 | QETH_DBF_TEXT(trace, 2, "addrxip4"); | ||
8557 | memcpy(&ipaddr->u.a4.addr, addr, 4); | ||
8558 | ipaddr->u.a4.mask = 0; | ||
8559 | #ifdef CONFIG_QETH_IPV6 | ||
8560 | } else if (proto == QETH_PROT_IPV6){ | ||
8561 | QETH_DBF_TEXT(trace, 2, "addrxip6"); | ||
8562 | memcpy(&ipaddr->u.a6.addr, addr, 16); | ||
8563 | ipaddr->u.a6.pfxlen = 0; | ||
8564 | #endif | ||
8565 | } | ||
8566 | ipaddr->type = QETH_IP_TYPE_RXIP; | ||
8567 | ipaddr->set_flags = QETH_IPA_SETIP_TAKEOVER_FLAG; | ||
8568 | ipaddr->del_flags = 0; | ||
8569 | } else | ||
8570 | return -ENOMEM; | ||
8571 | spin_lock_irqsave(&card->ip_lock, flags); | ||
8572 | if (__qeth_address_exists_in_list(&card->ip_list, ipaddr, 0) || | ||
8573 | __qeth_address_exists_in_list(card->ip_tbd_list, ipaddr, 0)) | ||
8574 | rc = -EEXIST; | ||
8575 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
8576 | if (rc){ | ||
8577 | PRINT_WARN("Cannot add RXIP. Address already exists!\n"); | ||
8578 | return rc; | ||
8579 | } | ||
8580 | if (!qeth_add_ip(card, ipaddr)) | ||
8581 | kfree(ipaddr); | ||
8582 | qeth_set_ip_addr_list(card); | ||
8583 | return 0; | ||
8584 | } | ||
8585 | |||
8586 | void | ||
8587 | qeth_del_rxip(struct qeth_card *card, enum qeth_prot_versions proto, | ||
8588 | const u8 *addr) | ||
8589 | { | ||
8590 | struct qeth_ipaddr *ipaddr; | ||
8591 | |||
8592 | ipaddr = qeth_get_addr_buffer(proto); | ||
8593 | if (ipaddr){ | ||
8594 | if (proto == QETH_PROT_IPV4){ | ||
8595 | QETH_DBF_TEXT(trace, 2, "addrxip4"); | ||
8596 | memcpy(&ipaddr->u.a4.addr, addr, 4); | ||
8597 | ipaddr->u.a4.mask = 0; | ||
8598 | #ifdef CONFIG_QETH_IPV6 | ||
8599 | } else if (proto == QETH_PROT_IPV6){ | ||
8600 | QETH_DBF_TEXT(trace, 2, "addrxip6"); | ||
8601 | memcpy(&ipaddr->u.a6.addr, addr, 16); | ||
8602 | ipaddr->u.a6.pfxlen = 0; | ||
8603 | #endif | ||
8604 | } | ||
8605 | ipaddr->type = QETH_IP_TYPE_RXIP; | ||
8606 | } else | ||
8607 | return; | ||
8608 | if (!qeth_delete_ip(card, ipaddr)) | ||
8609 | kfree(ipaddr); | ||
8610 | qeth_set_ip_addr_list(card); | ||
8611 | } | ||
8612 | |||
8613 | /** | ||
8614 | * IP event handler | ||
8615 | */ | ||
8616 | static int | ||
8617 | qeth_ip_event(struct notifier_block *this, | ||
8618 | unsigned long event,void *ptr) | ||
8619 | { | ||
8620 | struct in_ifaddr *ifa = (struct in_ifaddr *)ptr; | ||
8621 | struct net_device *dev =(struct net_device *) ifa->ifa_dev->dev; | ||
8622 | struct qeth_ipaddr *addr; | ||
8623 | struct qeth_card *card; | ||
8624 | |||
8625 | QETH_DBF_TEXT(trace,3,"ipevent"); | ||
8626 | card = qeth_get_card_from_dev(dev); | ||
8627 | if (!card) | ||
8628 | return NOTIFY_DONE; | ||
8629 | if (card->options.layer2) | ||
8630 | return NOTIFY_DONE; | ||
8631 | |||
8632 | addr = qeth_get_addr_buffer(QETH_PROT_IPV4); | ||
8633 | if (addr != NULL) { | ||
8634 | addr->u.a4.addr = ifa->ifa_address; | ||
8635 | addr->u.a4.mask = ifa->ifa_mask; | ||
8636 | addr->type = QETH_IP_TYPE_NORMAL; | ||
8637 | } else | ||
8638 | goto out; | ||
8639 | |||
8640 | switch(event) { | ||
8641 | case NETDEV_UP: | ||
8642 | if (!qeth_add_ip(card, addr)) | ||
8643 | kfree(addr); | ||
8644 | break; | ||
8645 | case NETDEV_DOWN: | ||
8646 | if (!qeth_delete_ip(card, addr)) | ||
8647 | kfree(addr); | ||
8648 | break; | ||
8649 | default: | ||
8650 | break; | ||
8651 | } | ||
8652 | qeth_set_ip_addr_list(card); | ||
8653 | out: | ||
8654 | return NOTIFY_DONE; | ||
8655 | } | ||
8656 | |||
8657 | static struct notifier_block qeth_ip_notifier = { | ||
8658 | qeth_ip_event, | ||
8659 | NULL, | ||
8660 | }; | ||
8661 | |||
8662 | #ifdef CONFIG_QETH_IPV6 | ||
8663 | /** | ||
8664 | * IPv6 event handler | ||
8665 | */ | ||
8666 | static int | ||
8667 | qeth_ip6_event(struct notifier_block *this, | ||
8668 | unsigned long event,void *ptr) | ||
8669 | { | ||
8670 | |||
8671 | struct inet6_ifaddr *ifa = (struct inet6_ifaddr *)ptr; | ||
8672 | struct net_device *dev = (struct net_device *)ifa->idev->dev; | ||
8673 | struct qeth_ipaddr *addr; | ||
8674 | struct qeth_card *card; | ||
8675 | |||
8676 | QETH_DBF_TEXT(trace,3,"ip6event"); | ||
8677 | |||
8678 | card = qeth_get_card_from_dev(dev); | ||
8679 | if (!card) | ||
8680 | return NOTIFY_DONE; | ||
8681 | if (!qeth_is_supported(card, IPA_IPV6)) | ||
8682 | return NOTIFY_DONE; | ||
8683 | |||
8684 | addr = qeth_get_addr_buffer(QETH_PROT_IPV6); | ||
8685 | if (addr != NULL) { | ||
8686 | memcpy(&addr->u.a6.addr, &ifa->addr, sizeof(struct in6_addr)); | ||
8687 | addr->u.a6.pfxlen = ifa->prefix_len; | ||
8688 | addr->type = QETH_IP_TYPE_NORMAL; | ||
8689 | } else | ||
8690 | goto out; | ||
8691 | |||
8692 | switch(event) { | ||
8693 | case NETDEV_UP: | ||
8694 | if (!qeth_add_ip(card, addr)) | ||
8695 | kfree(addr); | ||
8696 | break; | ||
8697 | case NETDEV_DOWN: | ||
8698 | if (!qeth_delete_ip(card, addr)) | ||
8699 | kfree(addr); | ||
8700 | break; | ||
8701 | default: | ||
8702 | break; | ||
8703 | } | ||
8704 | qeth_set_ip_addr_list(card); | ||
8705 | out: | ||
8706 | return NOTIFY_DONE; | ||
8707 | } | ||
8708 | |||
8709 | static struct notifier_block qeth_ip6_notifier = { | ||
8710 | qeth_ip6_event, | ||
8711 | NULL, | ||
8712 | }; | ||
8713 | #endif | ||
8714 | |||
8715 | static int | ||
8716 | __qeth_reboot_event_card(struct device *dev, void *data) | ||
8717 | { | ||
8718 | struct qeth_card *card; | ||
8719 | |||
8720 | card = (struct qeth_card *) dev->driver_data; | ||
8721 | qeth_clear_ip_list(card, 0, 0); | ||
8722 | qeth_qdio_clear_card(card, 0); | ||
8723 | qeth_clear_qdio_buffers(card); | ||
8724 | return 0; | ||
8725 | } | ||
8726 | |||
8727 | static int | ||
8728 | qeth_reboot_event(struct notifier_block *this, unsigned long event, void *ptr) | ||
8729 | { | ||
8730 | int ret; | ||
8731 | |||
8732 | ret = driver_for_each_device(&qeth_ccwgroup_driver.driver, NULL, NULL, | ||
8733 | __qeth_reboot_event_card); | ||
8734 | return ret ? NOTIFY_BAD : NOTIFY_DONE; | ||
8735 | } | ||
8736 | |||
8737 | |||
8738 | static struct notifier_block qeth_reboot_notifier = { | ||
8739 | qeth_reboot_event, | ||
8740 | NULL, | ||
8741 | }; | ||
8742 | |||
8743 | static int | ||
8744 | qeth_register_notifiers(void) | ||
8745 | { | ||
8746 | int r; | ||
8747 | |||
8748 | QETH_DBF_TEXT(trace,5,"regnotif"); | ||
8749 | if ((r = register_reboot_notifier(&qeth_reboot_notifier))) | ||
8750 | return r; | ||
8751 | if ((r = register_inetaddr_notifier(&qeth_ip_notifier))) | ||
8752 | goto out_reboot; | ||
8753 | #ifdef CONFIG_QETH_IPV6 | ||
8754 | if ((r = register_inet6addr_notifier(&qeth_ip6_notifier))) | ||
8755 | goto out_ipv4; | ||
8756 | #endif | ||
8757 | return 0; | ||
8758 | |||
8759 | #ifdef CONFIG_QETH_IPV6 | ||
8760 | out_ipv4: | ||
8761 | unregister_inetaddr_notifier(&qeth_ip_notifier); | ||
8762 | #endif | ||
8763 | out_reboot: | ||
8764 | unregister_reboot_notifier(&qeth_reboot_notifier); | ||
8765 | return r; | ||
8766 | } | ||
8767 | |||
8768 | /** | ||
8769 | * unregister all event notifiers | ||
8770 | */ | ||
8771 | static void | ||
8772 | qeth_unregister_notifiers(void) | ||
8773 | { | ||
8774 | |||
8775 | QETH_DBF_TEXT(trace,5,"unregnot"); | ||
8776 | BUG_ON(unregister_reboot_notifier(&qeth_reboot_notifier)); | ||
8777 | BUG_ON(unregister_inetaddr_notifier(&qeth_ip_notifier)); | ||
8778 | #ifdef CONFIG_QETH_IPV6 | ||
8779 | BUG_ON(unregister_inet6addr_notifier(&qeth_ip6_notifier)); | ||
8780 | #endif /* QETH_IPV6 */ | ||
8781 | |||
8782 | } | ||
8783 | |||
8784 | #ifdef CONFIG_QETH_IPV6 | ||
8785 | static int | ||
8786 | qeth_ipv6_init(void) | ||
8787 | { | ||
8788 | qeth_old_arp_constructor = arp_tbl.constructor; | ||
8789 | write_lock_bh(&arp_tbl.lock); | ||
8790 | arp_tbl.constructor = qeth_arp_constructor; | ||
8791 | write_unlock_bh(&arp_tbl.lock); | ||
8792 | |||
8793 | arp_direct_ops = (struct neigh_ops*) | ||
8794 | kmalloc(sizeof(struct neigh_ops), GFP_KERNEL); | ||
8795 | if (!arp_direct_ops) | ||
8796 | return -ENOMEM; | ||
8797 | |||
8798 | memcpy(arp_direct_ops, &arp_direct_ops_template, | ||
8799 | sizeof(struct neigh_ops)); | ||
8800 | |||
8801 | return 0; | ||
8802 | } | ||
8803 | |||
8804 | static void | ||
8805 | qeth_ipv6_uninit(void) | ||
8806 | { | ||
8807 | write_lock_bh(&arp_tbl.lock); | ||
8808 | arp_tbl.constructor = qeth_old_arp_constructor; | ||
8809 | write_unlock_bh(&arp_tbl.lock); | ||
8810 | kfree(arp_direct_ops); | ||
8811 | } | ||
8812 | #endif /* CONFIG_QETH_IPV6 */ | ||
8813 | |||
8814 | static void | ||
8815 | qeth_sysfs_unregister(void) | ||
8816 | { | ||
8817 | s390_root_dev_unregister(qeth_root_dev); | ||
8818 | qeth_remove_driver_attributes(); | ||
8819 | ccw_driver_unregister(&qeth_ccw_driver); | ||
8820 | ccwgroup_driver_unregister(&qeth_ccwgroup_driver); | ||
8821 | } | ||
8822 | |||
8823 | /** | ||
8824 | * register qeth at sysfs | ||
8825 | */ | ||
8826 | static int | ||
8827 | qeth_sysfs_register(void) | ||
8828 | { | ||
8829 | int rc; | ||
8830 | |||
8831 | rc = ccwgroup_driver_register(&qeth_ccwgroup_driver); | ||
8832 | if (rc) | ||
8833 | goto out; | ||
8834 | |||
8835 | rc = ccw_driver_register(&qeth_ccw_driver); | ||
8836 | if (rc) | ||
8837 | goto out_ccw_driver; | ||
8838 | |||
8839 | rc = qeth_create_driver_attributes(); | ||
8840 | if (rc) | ||
8841 | goto out_qeth_attr; | ||
8842 | |||
8843 | qeth_root_dev = s390_root_dev_register("qeth"); | ||
8844 | rc = IS_ERR(qeth_root_dev) ? PTR_ERR(qeth_root_dev) : 0; | ||
8845 | if (!rc) | ||
8846 | goto out; | ||
8847 | |||
8848 | qeth_remove_driver_attributes(); | ||
8849 | out_qeth_attr: | ||
8850 | ccw_driver_unregister(&qeth_ccw_driver); | ||
8851 | out_ccw_driver: | ||
8852 | ccwgroup_driver_unregister(&qeth_ccwgroup_driver); | ||
8853 | out: | ||
8854 | return rc; | ||
8855 | } | ||
8856 | |||
8857 | /*** | ||
8858 | * init function | ||
8859 | */ | ||
8860 | static int __init | ||
8861 | qeth_init(void) | ||
8862 | { | ||
8863 | int rc; | ||
8864 | |||
8865 | PRINT_INFO("loading %s\n", version); | ||
8866 | |||
8867 | INIT_LIST_HEAD(&qeth_card_list.list); | ||
8868 | INIT_LIST_HEAD(&qeth_notify_list); | ||
8869 | spin_lock_init(&qeth_notify_lock); | ||
8870 | rwlock_init(&qeth_card_list.rwlock); | ||
8871 | |||
8872 | rc = qeth_register_dbf_views(); | ||
8873 | if (rc) | ||
8874 | goto out_err; | ||
8875 | |||
8876 | rc = qeth_sysfs_register(); | ||
8877 | if (rc) | ||
8878 | goto out_dbf; | ||
8879 | |||
8880 | #ifdef CONFIG_QETH_IPV6 | ||
8881 | rc = qeth_ipv6_init(); | ||
8882 | if (rc) { | ||
8883 | PRINT_ERR("Out of memory during ipv6 init code = %d\n", rc); | ||
8884 | goto out_sysfs; | ||
8885 | } | ||
8886 | #endif /* QETH_IPV6 */ | ||
8887 | rc = qeth_register_notifiers(); | ||
8888 | if (rc) | ||
8889 | goto out_ipv6; | ||
8890 | rc = qeth_create_procfs_entries(); | ||
8891 | if (rc) | ||
8892 | goto out_notifiers; | ||
8893 | |||
8894 | return rc; | ||
8895 | |||
8896 | out_notifiers: | ||
8897 | qeth_unregister_notifiers(); | ||
8898 | out_ipv6: | ||
8899 | #ifdef CONFIG_QETH_IPV6 | ||
8900 | qeth_ipv6_uninit(); | ||
8901 | out_sysfs: | ||
8902 | #endif /* QETH_IPV6 */ | ||
8903 | qeth_sysfs_unregister(); | ||
8904 | out_dbf: | ||
8905 | qeth_unregister_dbf_views(); | ||
8906 | out_err: | ||
8907 | PRINT_ERR("Initialization failed with code %d\n", rc); | ||
8908 | return rc; | ||
8909 | } | ||
8910 | |||
8911 | static void | ||
8912 | __exit qeth_exit(void) | ||
8913 | { | ||
8914 | struct qeth_card *card, *tmp; | ||
8915 | unsigned long flags; | ||
8916 | |||
8917 | QETH_DBF_TEXT(trace,1, "cleanup."); | ||
8918 | |||
8919 | /* | ||
8920 | * Weed would not need to clean up our devices here, because the | ||
8921 | * common device layer calls qeth_remove_device for each device | ||
8922 | * as soon as we unregister our driver (done in qeth_sysfs_unregister). | ||
8923 | * But we do cleanup here so we can do a "soft" shutdown of our cards. | ||
8924 | * qeth_remove_device called by the common device layer would otherwise | ||
8925 | * do a "hard" shutdown (card->use_hard_stop is set to one in | ||
8926 | * qeth_remove_device). | ||
8927 | */ | ||
8928 | again: | ||
8929 | read_lock_irqsave(&qeth_card_list.rwlock, flags); | ||
8930 | list_for_each_entry_safe(card, tmp, &qeth_card_list.list, list){ | ||
8931 | read_unlock_irqrestore(&qeth_card_list.rwlock, flags); | ||
8932 | qeth_set_offline(card->gdev); | ||
8933 | qeth_remove_device(card->gdev); | ||
8934 | goto again; | ||
8935 | } | ||
8936 | read_unlock_irqrestore(&qeth_card_list.rwlock, flags); | ||
8937 | #ifdef CONFIG_QETH_IPV6 | ||
8938 | qeth_ipv6_uninit(); | ||
8939 | #endif | ||
8940 | qeth_unregister_notifiers(); | ||
8941 | qeth_remove_procfs_entries(); | ||
8942 | qeth_sysfs_unregister(); | ||
8943 | qeth_unregister_dbf_views(); | ||
8944 | printk("qeth: removed\n"); | ||
8945 | } | ||
8946 | |||
8947 | EXPORT_SYMBOL(qeth_osn_register); | ||
8948 | EXPORT_SYMBOL(qeth_osn_deregister); | ||
8949 | EXPORT_SYMBOL(qeth_osn_assist); | ||
8950 | module_init(qeth_init); | ||
8951 | module_exit(qeth_exit); | ||
8952 | MODULE_AUTHOR("Frank Pavlic <fpavlic@de.ibm.com>"); | ||
8953 | MODULE_DESCRIPTION("Linux on zSeries OSA Express and HiperSockets support\n" \ | ||
8954 | "Copyright 2000,2003 IBM Corporation\n"); | ||
8955 | |||
8956 | MODULE_LICENSE("GPL"); | ||
diff --git a/drivers/s390/net/qeth_mpc.c b/drivers/s390/net/qeth_mpc.c deleted file mode 100644 index f29a4bc4f6f2..000000000000 --- a/drivers/s390/net/qeth_mpc.c +++ /dev/null | |||
@@ -1,269 +0,0 @@ | |||
1 | /* | ||
2 | * linux/drivers/s390/net/qeth_mpc.c | ||
3 | * | ||
4 | * Linux on zSeries OSA Express and HiperSockets support | ||
5 | * | ||
6 | * Copyright 2000,2003 IBM Corporation | ||
7 | * Author(s): Frank Pavlic <fpavlic@de.ibm.com> | ||
8 | * Thomas Spatzier <tspat@de.ibm.com> | ||
9 | * | ||
10 | */ | ||
11 | #include <asm/cio.h> | ||
12 | #include "qeth_mpc.h" | ||
13 | |||
14 | unsigned char IDX_ACTIVATE_READ[]={ | ||
15 | 0x00,0x00,0x80,0x00, 0x00,0x00,0x00,0x00, | ||
16 | 0x19,0x01,0x01,0x80, 0x00,0x00,0x00,0x00, | ||
17 | 0x00,0x00,0x00,0x00, 0x00,0x00,0xc8,0xc1, | ||
18 | 0xd3,0xd3,0xd6,0xd3, 0xc5,0x40,0x00,0x00, | ||
19 | 0x00,0x00 | ||
20 | }; | ||
21 | |||
22 | unsigned char IDX_ACTIVATE_WRITE[]={ | ||
23 | 0x00,0x00,0x80,0x00, 0x00,0x00,0x00,0x00, | ||
24 | 0x15,0x01,0x01,0x80, 0x00,0x00,0x00,0x00, | ||
25 | 0xff,0xff,0x00,0x00, 0x00,0x00,0xc8,0xc1, | ||
26 | 0xd3,0xd3,0xd6,0xd3, 0xc5,0x40,0x00,0x00, | ||
27 | 0x00,0x00 | ||
28 | }; | ||
29 | |||
30 | unsigned char CM_ENABLE[]={ | ||
31 | 0x00,0xe0,0x00,0x00, 0x00,0x00,0x00,0x01, | ||
32 | 0x00,0x00,0x00,0x14, 0x00,0x00,0x00,0x63, | ||
33 | 0x10,0x00,0x00,0x01, | ||
34 | 0x00,0x00,0x00,0x00, | ||
35 | 0x81,0x7e,0x00,0x01, 0x00,0x00,0x00,0x00, | ||
36 | 0x00,0x00,0x00,0x00, 0x00,0x24,0x00,0x23, | ||
37 | 0x00,0x00,0x23,0x05, 0x00,0x00,0x00,0x00, | ||
38 | 0x00,0x00,0x00,0x00, 0x00,0x00,0x00,0x00, | ||
39 | 0x01,0x00,0x00,0x23, 0x00,0x00,0x00,0x40, | ||
40 | 0x00,0x0c,0x41,0x02, 0x00,0x17,0x00,0x00, | ||
41 | 0x00,0x00,0x00,0x00, | ||
42 | 0x00,0x0b,0x04,0x01, | ||
43 | 0x7e,0x04,0x05,0x00, 0x01,0x01,0x0f, | ||
44 | 0x00, | ||
45 | 0x0c,0x04,0x02,0xff, 0xff,0xff,0xff,0xff, | ||
46 | 0xff,0xff,0xff | ||
47 | }; | ||
48 | |||
49 | unsigned char CM_SETUP[]={ | ||
50 | 0x00,0xe0,0x00,0x00, 0x00,0x00,0x00,0x02, | ||
51 | 0x00,0x00,0x00,0x14, 0x00,0x00,0x00,0x64, | ||
52 | 0x10,0x00,0x00,0x01, | ||
53 | 0x00,0x00,0x00,0x00, | ||
54 | 0x81,0x7e,0x00,0x01, 0x00,0x00,0x00,0x00, | ||
55 | 0x00,0x00,0x00,0x00, 0x00,0x24,0x00,0x24, | ||
56 | 0x00,0x00,0x24,0x05, 0x00,0x00,0x00,0x00, | ||
57 | 0x00,0x00,0x00,0x00, 0x00,0x00,0x00,0x00, | ||
58 | 0x01,0x00,0x00,0x24, 0x00,0x00,0x00,0x40, | ||
59 | 0x00,0x0c,0x41,0x04, 0x00,0x18,0x00,0x00, | ||
60 | 0x00,0x00,0x00,0x00, | ||
61 | 0x00,0x09,0x04,0x04, | ||
62 | 0x05,0x00,0x01,0x01, 0x11, | ||
63 | 0x00,0x09,0x04, | ||
64 | 0x05,0x05,0x00,0x00, 0x00,0x00, | ||
65 | 0x00,0x06, | ||
66 | 0x04,0x06,0xc8,0x00 | ||
67 | }; | ||
68 | |||
69 | unsigned char ULP_ENABLE[]={ | ||
70 | 0x00,0xe0,0x00,0x00, 0x00,0x00,0x00,0x03, | ||
71 | 0x00,0x00,0x00,0x14, 0x00,0x00,0x00,0x6b, | ||
72 | 0x10,0x00,0x00,0x01, | ||
73 | 0x00,0x00,0x00,0x00, | ||
74 | 0x41,0x7e,0x00,0x01, 0x00,0x00,0x00,0x01, | ||
75 | 0x00,0x00,0x00,0x00, 0x00,0x24,0x00,0x2b, | ||
76 | 0x00,0x00,0x2b,0x05, 0x20,0x01,0x00,0x00, | ||
77 | 0x00,0x00,0x00,0x00, 0x00,0x00,0x00,0x00, | ||
78 | 0x01,0x00,0x00,0x2b, 0x00,0x00,0x00,0x40, | ||
79 | 0x00,0x0c,0x41,0x02, 0x00,0x1f,0x00,0x00, | ||
80 | 0x00,0x00,0x00,0x00, | ||
81 | 0x00,0x0b,0x04,0x01, | ||
82 | 0x03,0x04,0x05,0x00, 0x01,0x01,0x12, | ||
83 | 0x00, | ||
84 | 0x14,0x04,0x0a,0x00, 0x20,0x00,0x00,0xff, | ||
85 | 0xff,0x00,0x08,0xc8, 0xe8,0xc4,0xf1,0xc7, | ||
86 | 0xf1,0x00,0x00 | ||
87 | }; | ||
88 | |||
89 | unsigned char ULP_SETUP[]={ | ||
90 | 0x00,0xe0,0x00,0x00, 0x00,0x00,0x00,0x04, | ||
91 | 0x00,0x00,0x00,0x14, 0x00,0x00,0x00,0x6c, | ||
92 | 0x10,0x00,0x00,0x01, | ||
93 | 0x00,0x00,0x00,0x00, | ||
94 | 0x41,0x7e,0x00,0x01, 0x00,0x00,0x00,0x02, | ||
95 | 0x00,0x00,0x00,0x01, 0x00,0x24,0x00,0x2c, | ||
96 | 0x00,0x00,0x2c,0x05, 0x20,0x01,0x00,0x00, | ||
97 | 0x00,0x00,0x00,0x00, 0x00,0x00,0x00,0x00, | ||
98 | 0x01,0x00,0x00,0x2c, 0x00,0x00,0x00,0x40, | ||
99 | 0x00,0x0c,0x41,0x04, 0x00,0x20,0x00,0x00, | ||
100 | 0x00,0x00,0x00,0x00, | ||
101 | 0x00,0x09,0x04,0x04, | ||
102 | 0x05,0x00,0x01,0x01, 0x14, | ||
103 | 0x00,0x09,0x04, | ||
104 | 0x05,0x05,0x30,0x01, 0x00,0x00, | ||
105 | 0x00,0x06, | ||
106 | 0x04,0x06,0x40,0x00, | ||
107 | 0x00,0x08,0x04,0x0b, | ||
108 | 0x00,0x00,0x00,0x00 | ||
109 | }; | ||
110 | |||
111 | unsigned char DM_ACT[]={ | ||
112 | 0x00,0xe0,0x00,0x00, 0x00,0x00,0x00,0x05, | ||
113 | 0x00,0x00,0x00,0x14, 0x00,0x00,0x00,0x55, | ||
114 | 0x10,0x00,0x00,0x01, | ||
115 | 0x00,0x00,0x00,0x00, | ||
116 | 0x41,0x7e,0x00,0x01, 0x00,0x00,0x00,0x03, | ||
117 | 0x00,0x00,0x00,0x02, 0x00,0x24,0x00,0x15, | ||
118 | 0x00,0x00,0x2c,0x05, 0x20,0x01,0x00,0x00, | ||
119 | 0x00,0x00,0x00,0x00, 0x00,0x00,0x00,0x00, | ||
120 | 0x01,0x00,0x00,0x15, 0x00,0x00,0x00,0x40, | ||
121 | 0x00,0x0c,0x43,0x60, 0x00,0x09,0x00,0x00, | ||
122 | 0x00,0x00,0x00,0x00, | ||
123 | 0x00,0x09,0x04,0x04, | ||
124 | 0x05,0x40,0x01,0x01, 0x00 | ||
125 | }; | ||
126 | |||
127 | unsigned char IPA_PDU_HEADER[]={ | ||
128 | 0x00,0xe0,0x00,0x00, 0x77,0x77,0x77,0x77, | ||
129 | 0x00,0x00,0x00,0x14, 0x00,0x00, | ||
130 | (IPA_PDU_HEADER_SIZE+sizeof(struct qeth_ipa_cmd))/256, | ||
131 | (IPA_PDU_HEADER_SIZE+sizeof(struct qeth_ipa_cmd))%256, | ||
132 | 0x10,0x00,0x00,0x01, 0x00,0x00,0x00,0x00, | ||
133 | 0xc1,0x03,0x00,0x01, 0x00,0x00,0x00,0x00, | ||
134 | 0x00,0x00,0x00,0x00, 0x00,0x24, | ||
135 | sizeof(struct qeth_ipa_cmd)/256, | ||
136 | sizeof(struct qeth_ipa_cmd)%256, | ||
137 | 0x00, | ||
138 | sizeof(struct qeth_ipa_cmd)/256, | ||
139 | sizeof(struct qeth_ipa_cmd)%256, | ||
140 | 0x05, | ||
141 | 0x77,0x77,0x77,0x77, | ||
142 | 0x00,0x00,0x00,0x00, 0x00,0x00,0x00,0x00, | ||
143 | 0x01,0x00, | ||
144 | sizeof(struct qeth_ipa_cmd)/256, | ||
145 | sizeof(struct qeth_ipa_cmd)%256, | ||
146 | 0x00,0x00,0x00,0x40, | ||
147 | }; | ||
148 | |||
149 | unsigned char WRITE_CCW[]={ | ||
150 | 0x01,CCW_FLAG_SLI,0,0, | ||
151 | 0,0,0,0 | ||
152 | }; | ||
153 | |||
154 | unsigned char READ_CCW[]={ | ||
155 | 0x02,CCW_FLAG_SLI,0,0, | ||
156 | 0,0,0,0 | ||
157 | }; | ||
158 | |||
159 | |||
160 | struct ipa_rc_msg { | ||
161 | enum qeth_ipa_return_codes rc; | ||
162 | char *msg; | ||
163 | }; | ||
164 | |||
165 | static struct ipa_rc_msg qeth_ipa_rc_msg[] = { | ||
166 | {IPA_RC_SUCCESS, "success"}, | ||
167 | {IPA_RC_NOTSUPP, "Command not supported"}, | ||
168 | {IPA_RC_IP_TABLE_FULL, "Add Addr IP Table Full - ipv6"}, | ||
169 | {IPA_RC_UNKNOWN_ERROR, "IPA command failed - reason unknown"}, | ||
170 | {IPA_RC_UNSUPPORTED_COMMAND, "Command not supported"}, | ||
171 | {IPA_RC_DUP_IPV6_REMOTE,"ipv6 address already registered remote"}, | ||
172 | {IPA_RC_DUP_IPV6_HOME, "ipv6 address already registered"}, | ||
173 | {IPA_RC_UNREGISTERED_ADDR, "Address not registered"}, | ||
174 | {IPA_RC_NO_ID_AVAILABLE, "No identifiers available"}, | ||
175 | {IPA_RC_ID_NOT_FOUND, "Identifier not found"}, | ||
176 | {IPA_RC_INVALID_IP_VERSION, "IP version incorrect"}, | ||
177 | {IPA_RC_LAN_FRAME_MISMATCH, "LAN and frame mismatch"}, | ||
178 | {IPA_RC_L2_UNSUPPORTED_CMD, "Unsupported layer 2 command"}, | ||
179 | {IPA_RC_L2_DUP_MAC, "Duplicate MAC address"}, | ||
180 | {IPA_RC_L2_ADDR_TABLE_FULL, "Layer2 address table full"}, | ||
181 | {IPA_RC_L2_DUP_LAYER3_MAC, "Duplicate with layer 3 MAC"}, | ||
182 | {IPA_RC_L2_GMAC_NOT_FOUND, "GMAC not found"}, | ||
183 | {IPA_RC_L2_MAC_NOT_FOUND, "L2 mac address not found"}, | ||
184 | {IPA_RC_L2_INVALID_VLAN_ID, "L2 invalid vlan id"}, | ||
185 | {IPA_RC_L2_DUP_VLAN_ID, "L2 duplicate vlan id"}, | ||
186 | {IPA_RC_L2_VLAN_ID_NOT_FOUND, "L2 vlan id not found"}, | ||
187 | {IPA_RC_DATA_MISMATCH, "Data field mismatch (v4/v6 mixed)"}, | ||
188 | {IPA_RC_INVALID_MTU_SIZE, "Invalid MTU size"}, | ||
189 | {IPA_RC_INVALID_LANTYPE, "Invalid LAN type"}, | ||
190 | {IPA_RC_INVALID_LANNUM, "Invalid LAN num"}, | ||
191 | {IPA_RC_DUPLICATE_IP_ADDRESS, "Address already registered"}, | ||
192 | {IPA_RC_IP_ADDR_TABLE_FULL, "IP address table full"}, | ||
193 | {IPA_RC_LAN_PORT_STATE_ERROR, "LAN port state error"}, | ||
194 | {IPA_RC_SETIP_NO_STARTLAN, "Setip no startlan received"}, | ||
195 | {IPA_RC_SETIP_ALREADY_RECEIVED, "Setip already received"}, | ||
196 | {IPA_RC_IP_ADDR_ALREADY_USED, "IP address already in use on LAN"}, | ||
197 | {IPA_RC_MULTICAST_FULL, "No task available, multicast full"}, | ||
198 | {IPA_RC_SETIP_INVALID_VERSION, "SETIP invalid IP version"}, | ||
199 | {IPA_RC_UNSUPPORTED_SUBCMD, "Unsupported assist subcommand"}, | ||
200 | {IPA_RC_ARP_ASSIST_NO_ENABLE, "Only partial success, no enable"}, | ||
201 | {IPA_RC_PRIMARY_ALREADY_DEFINED,"Primary already defined"}, | ||
202 | {IPA_RC_SECOND_ALREADY_DEFINED, "Secondary already defined"}, | ||
203 | {IPA_RC_INVALID_SETRTG_INDICATOR,"Invalid SETRTG indicator"}, | ||
204 | {IPA_RC_MC_ADDR_ALREADY_DEFINED,"Multicast address already defined"}, | ||
205 | {IPA_RC_LAN_OFFLINE, "STRTLAN_LAN_DISABLED - LAN offline"}, | ||
206 | {IPA_RC_INVALID_IP_VERSION2, "Invalid IP version"}, | ||
207 | {IPA_RC_FFFF, "Unknown Error"} | ||
208 | }; | ||
209 | |||
210 | |||
211 | |||
212 | char * | ||
213 | qeth_get_ipa_msg(enum qeth_ipa_return_codes rc) | ||
214 | { | ||
215 | int x = 0; | ||
216 | qeth_ipa_rc_msg[sizeof(qeth_ipa_rc_msg) / | ||
217 | sizeof(struct ipa_rc_msg) - 1].rc = rc; | ||
218 | while(qeth_ipa_rc_msg[x].rc != rc) | ||
219 | x++; | ||
220 | return qeth_ipa_rc_msg[x].msg; | ||
221 | } | ||
222 | |||
223 | |||
224 | struct ipa_cmd_names { | ||
225 | enum qeth_ipa_cmds cmd; | ||
226 | char *name; | ||
227 | }; | ||
228 | |||
229 | static struct ipa_cmd_names qeth_ipa_cmd_names[] = { | ||
230 | {IPA_CMD_STARTLAN, "startlan"}, | ||
231 | {IPA_CMD_STOPLAN, "stoplan"}, | ||
232 | {IPA_CMD_SETVMAC, "setvmac"}, | ||
233 | {IPA_CMD_DELVMAC, "delvmca"}, | ||
234 | {IPA_CMD_SETGMAC, "setgmac"}, | ||
235 | {IPA_CMD_DELGMAC, "delgmac"}, | ||
236 | {IPA_CMD_SETVLAN, "setvlan"}, | ||
237 | {IPA_CMD_DELVLAN, "delvlan"}, | ||
238 | {IPA_CMD_SETCCID, "setccid"}, | ||
239 | {IPA_CMD_DELCCID, "delccid"}, | ||
240 | {IPA_CMD_MODCCID, "setip"}, | ||
241 | {IPA_CMD_SETIP, "setip"}, | ||
242 | {IPA_CMD_QIPASSIST, "qipassist"}, | ||
243 | {IPA_CMD_SETASSPARMS, "setassparms"}, | ||
244 | {IPA_CMD_SETIPM, "setipm"}, | ||
245 | {IPA_CMD_DELIPM, "delipm"}, | ||
246 | {IPA_CMD_SETRTG, "setrtg"}, | ||
247 | {IPA_CMD_DELIP, "delip"}, | ||
248 | {IPA_CMD_SETADAPTERPARMS, "setadapterparms"}, | ||
249 | {IPA_CMD_SET_DIAG_ASS, "set_diag_ass"}, | ||
250 | {IPA_CMD_CREATE_ADDR, "create_addr"}, | ||
251 | {IPA_CMD_DESTROY_ADDR, "destroy_addr"}, | ||
252 | {IPA_CMD_REGISTER_LOCAL_ADDR, "register_local_addr"}, | ||
253 | {IPA_CMD_UNREGISTER_LOCAL_ADDR, "unregister_local_addr"}, | ||
254 | {IPA_CMD_UNKNOWN, "unknown"}, | ||
255 | }; | ||
256 | |||
257 | char * | ||
258 | qeth_get_ipa_cmd_name(enum qeth_ipa_cmds cmd) | ||
259 | { | ||
260 | int x = 0; | ||
261 | qeth_ipa_cmd_names[ | ||
262 | sizeof(qeth_ipa_cmd_names)/ | ||
263 | sizeof(struct ipa_cmd_names)-1].cmd = cmd; | ||
264 | while(qeth_ipa_cmd_names[x].cmd != cmd) | ||
265 | x++; | ||
266 | return qeth_ipa_cmd_names[x].name; | ||
267 | } | ||
268 | |||
269 | |||
diff --git a/drivers/s390/net/qeth_proc.c b/drivers/s390/net/qeth_proc.c deleted file mode 100644 index 46ecd03a597e..000000000000 --- a/drivers/s390/net/qeth_proc.c +++ /dev/null | |||
@@ -1,316 +0,0 @@ | |||
1 | /* | ||
2 | * | ||
3 | * linux/drivers/s390/net/qeth_fs.c | ||
4 | * | ||
5 | * Linux on zSeries OSA Express and HiperSockets support | ||
6 | * This file contains code related to procfs. | ||
7 | * | ||
8 | * Copyright 2000,2003 IBM Corporation | ||
9 | * | ||
10 | * Author(s): Thomas Spatzier <tspat@de.ibm.com> | ||
11 | * | ||
12 | */ | ||
13 | #include <linux/module.h> | ||
14 | #include <linux/init.h> | ||
15 | #include <linux/proc_fs.h> | ||
16 | #include <linux/seq_file.h> | ||
17 | #include <linux/list.h> | ||
18 | #include <linux/rwsem.h> | ||
19 | |||
20 | #include "qeth.h" | ||
21 | #include "qeth_mpc.h" | ||
22 | #include "qeth_fs.h" | ||
23 | |||
24 | /***** /proc/qeth *****/ | ||
25 | #define QETH_PROCFILE_NAME "qeth" | ||
26 | static struct proc_dir_entry *qeth_procfile; | ||
27 | |||
28 | static int | ||
29 | qeth_procfile_seq_match(struct device *dev, void *data) | ||
30 | { | ||
31 | return(dev ? 1 : 0); | ||
32 | } | ||
33 | |||
34 | static void * | ||
35 | qeth_procfile_seq_start(struct seq_file *s, loff_t *offset) | ||
36 | { | ||
37 | struct device *dev = NULL; | ||
38 | loff_t nr = 0; | ||
39 | |||
40 | if (*offset == 0) | ||
41 | return SEQ_START_TOKEN; | ||
42 | while (1) { | ||
43 | dev = driver_find_device(&qeth_ccwgroup_driver.driver, dev, | ||
44 | NULL, qeth_procfile_seq_match); | ||
45 | if (++nr == *offset) | ||
46 | break; | ||
47 | put_device(dev); | ||
48 | } | ||
49 | return dev; | ||
50 | } | ||
51 | |||
52 | static void | ||
53 | qeth_procfile_seq_stop(struct seq_file *s, void* it) | ||
54 | { | ||
55 | } | ||
56 | |||
57 | static void * | ||
58 | qeth_procfile_seq_next(struct seq_file *s, void *it, loff_t *offset) | ||
59 | { | ||
60 | struct device *prev, *next; | ||
61 | |||
62 | if (it == SEQ_START_TOKEN) | ||
63 | prev = NULL; | ||
64 | else | ||
65 | prev = (struct device *) it; | ||
66 | next = driver_find_device(&qeth_ccwgroup_driver.driver, | ||
67 | prev, NULL, qeth_procfile_seq_match); | ||
68 | (*offset)++; | ||
69 | return (void *) next; | ||
70 | } | ||
71 | |||
72 | static inline const char * | ||
73 | qeth_get_router_str(struct qeth_card *card, int ipv) | ||
74 | { | ||
75 | enum qeth_routing_types routing_type = NO_ROUTER; | ||
76 | |||
77 | if (ipv == 4) { | ||
78 | routing_type = card->options.route4.type; | ||
79 | } else { | ||
80 | #ifdef CONFIG_QETH_IPV6 | ||
81 | routing_type = card->options.route6.type; | ||
82 | #else | ||
83 | return "n/a"; | ||
84 | #endif /* CONFIG_QETH_IPV6 */ | ||
85 | } | ||
86 | |||
87 | switch (routing_type){ | ||
88 | case PRIMARY_ROUTER: | ||
89 | return "pri"; | ||
90 | case SECONDARY_ROUTER: | ||
91 | return "sec"; | ||
92 | case MULTICAST_ROUTER: | ||
93 | if (card->info.broadcast_capable == QETH_BROADCAST_WITHOUT_ECHO) | ||
94 | return "mc+"; | ||
95 | return "mc"; | ||
96 | case PRIMARY_CONNECTOR: | ||
97 | if (card->info.broadcast_capable == QETH_BROADCAST_WITHOUT_ECHO) | ||
98 | return "p+c"; | ||
99 | return "p.c"; | ||
100 | case SECONDARY_CONNECTOR: | ||
101 | if (card->info.broadcast_capable == QETH_BROADCAST_WITHOUT_ECHO) | ||
102 | return "s+c"; | ||
103 | return "s.c"; | ||
104 | default: /* NO_ROUTER */ | ||
105 | return "no"; | ||
106 | } | ||
107 | } | ||
108 | |||
109 | static int | ||
110 | qeth_procfile_seq_show(struct seq_file *s, void *it) | ||
111 | { | ||
112 | struct device *device; | ||
113 | struct qeth_card *card; | ||
114 | char tmp[12]; /* for qeth_get_prioq_str */ | ||
115 | |||
116 | if (it == SEQ_START_TOKEN){ | ||
117 | seq_printf(s, "devices CHPID interface " | ||
118 | "cardtype port chksum prio-q'ing rtr4 " | ||
119 | "rtr6 fsz cnt\n"); | ||
120 | seq_printf(s, "-------------------------- ----- ---------- " | ||
121 | "-------------- ---- ------ ---------- ---- " | ||
122 | "---- ----- -----\n"); | ||
123 | } else { | ||
124 | device = (struct device *) it; | ||
125 | card = device->driver_data; | ||
126 | seq_printf(s, "%s/%s/%s x%02X %-10s %-14s %-4i ", | ||
127 | CARD_RDEV_ID(card), | ||
128 | CARD_WDEV_ID(card), | ||
129 | CARD_DDEV_ID(card), | ||
130 | card->info.chpid, | ||
131 | QETH_CARD_IFNAME(card), | ||
132 | qeth_get_cardname_short(card), | ||
133 | card->info.portno); | ||
134 | if (card->lan_online) | ||
135 | seq_printf(s, "%-6s %-10s %-4s %-4s %-5s %-5i\n", | ||
136 | qeth_get_checksum_str(card), | ||
137 | qeth_get_prioq_str(card, tmp), | ||
138 | qeth_get_router_str(card, 4), | ||
139 | qeth_get_router_str(card, 6), | ||
140 | qeth_get_bufsize_str(card), | ||
141 | card->qdio.in_buf_pool.buf_count); | ||
142 | else | ||
143 | seq_printf(s, " +++ LAN OFFLINE +++\n"); | ||
144 | put_device(device); | ||
145 | } | ||
146 | return 0; | ||
147 | } | ||
148 | |||
149 | static const struct seq_operations qeth_procfile_seq_ops = { | ||
150 | .start = qeth_procfile_seq_start, | ||
151 | .stop = qeth_procfile_seq_stop, | ||
152 | .next = qeth_procfile_seq_next, | ||
153 | .show = qeth_procfile_seq_show, | ||
154 | }; | ||
155 | |||
156 | static int | ||
157 | qeth_procfile_open(struct inode *inode, struct file *file) | ||
158 | { | ||
159 | return seq_open(file, &qeth_procfile_seq_ops); | ||
160 | } | ||
161 | |||
162 | static const struct file_operations qeth_procfile_fops = { | ||
163 | .owner = THIS_MODULE, | ||
164 | .open = qeth_procfile_open, | ||
165 | .read = seq_read, | ||
166 | .llseek = seq_lseek, | ||
167 | .release = seq_release, | ||
168 | }; | ||
169 | |||
170 | /***** /proc/qeth_perf *****/ | ||
171 | #define QETH_PERF_PROCFILE_NAME "qeth_perf" | ||
172 | static struct proc_dir_entry *qeth_perf_procfile; | ||
173 | |||
174 | static int | ||
175 | qeth_perf_procfile_seq_show(struct seq_file *s, void *it) | ||
176 | { | ||
177 | struct device *device; | ||
178 | struct qeth_card *card; | ||
179 | |||
180 | |||
181 | if (it == SEQ_START_TOKEN) | ||
182 | return 0; | ||
183 | |||
184 | device = (struct device *) it; | ||
185 | card = device->driver_data; | ||
186 | seq_printf(s, "For card with devnos %s/%s/%s (%s):\n", | ||
187 | CARD_RDEV_ID(card), | ||
188 | CARD_WDEV_ID(card), | ||
189 | CARD_DDEV_ID(card), | ||
190 | QETH_CARD_IFNAME(card) | ||
191 | ); | ||
192 | if (!card->options.performance_stats) | ||
193 | seq_printf(s, "Performance statistics are deactivated.\n"); | ||
194 | seq_printf(s, " Skb's/buffers received : %lu/%u\n" | ||
195 | " Skb's/buffers sent : %lu/%u\n\n", | ||
196 | card->stats.rx_packets - | ||
197 | card->perf_stats.initial_rx_packets, | ||
198 | card->perf_stats.bufs_rec, | ||
199 | card->stats.tx_packets - | ||
200 | card->perf_stats.initial_tx_packets, | ||
201 | card->perf_stats.bufs_sent | ||
202 | ); | ||
203 | seq_printf(s, " Skb's/buffers sent without packing : %lu/%u\n" | ||
204 | " Skb's/buffers sent with packing : %u/%u\n\n", | ||
205 | card->stats.tx_packets - card->perf_stats.initial_tx_packets | ||
206 | - card->perf_stats.skbs_sent_pack, | ||
207 | card->perf_stats.bufs_sent - card->perf_stats.bufs_sent_pack, | ||
208 | card->perf_stats.skbs_sent_pack, | ||
209 | card->perf_stats.bufs_sent_pack | ||
210 | ); | ||
211 | seq_printf(s, " Skbs sent in SG mode : %u\n" | ||
212 | " Skb fragments sent in SG mode : %u\n\n", | ||
213 | card->perf_stats.sg_skbs_sent, | ||
214 | card->perf_stats.sg_frags_sent); | ||
215 | seq_printf(s, " Skbs received in SG mode : %u\n" | ||
216 | " Skb fragments received in SG mode : %u\n" | ||
217 | " Page allocations for rx SG mode : %u\n\n", | ||
218 | card->perf_stats.sg_skbs_rx, | ||
219 | card->perf_stats.sg_frags_rx, | ||
220 | card->perf_stats.sg_alloc_page_rx); | ||
221 | seq_printf(s, " large_send tx (in Kbytes) : %u\n" | ||
222 | " large_send count : %u\n\n", | ||
223 | card->perf_stats.large_send_bytes >> 10, | ||
224 | card->perf_stats.large_send_cnt); | ||
225 | seq_printf(s, " Packing state changes no pkg.->packing : %u/%u\n" | ||
226 | " Watermarks L/H : %i/%i\n" | ||
227 | " Current buffer usage (outbound q's) : " | ||
228 | "%i/%i/%i/%i\n\n", | ||
229 | card->perf_stats.sc_dp_p, card->perf_stats.sc_p_dp, | ||
230 | QETH_LOW_WATERMARK_PACK, QETH_HIGH_WATERMARK_PACK, | ||
231 | atomic_read(&card->qdio.out_qs[0]->used_buffers), | ||
232 | (card->qdio.no_out_queues > 1)? | ||
233 | atomic_read(&card->qdio.out_qs[1]->used_buffers) | ||
234 | : 0, | ||
235 | (card->qdio.no_out_queues > 2)? | ||
236 | atomic_read(&card->qdio.out_qs[2]->used_buffers) | ||
237 | : 0, | ||
238 | (card->qdio.no_out_queues > 3)? | ||
239 | atomic_read(&card->qdio.out_qs[3]->used_buffers) | ||
240 | : 0 | ||
241 | ); | ||
242 | seq_printf(s, " Inbound handler time (in us) : %u\n" | ||
243 | " Inbound handler count : %u\n" | ||
244 | " Inbound do_QDIO time (in us) : %u\n" | ||
245 | " Inbound do_QDIO count : %u\n\n" | ||
246 | " Outbound handler time (in us) : %u\n" | ||
247 | " Outbound handler count : %u\n\n" | ||
248 | " Outbound time (in us, incl QDIO) : %u\n" | ||
249 | " Outbound count : %u\n" | ||
250 | " Outbound do_QDIO time (in us) : %u\n" | ||
251 | " Outbound do_QDIO count : %u\n\n", | ||
252 | card->perf_stats.inbound_time, | ||
253 | card->perf_stats.inbound_cnt, | ||
254 | card->perf_stats.inbound_do_qdio_time, | ||
255 | card->perf_stats.inbound_do_qdio_cnt, | ||
256 | card->perf_stats.outbound_handler_time, | ||
257 | card->perf_stats.outbound_handler_cnt, | ||
258 | card->perf_stats.outbound_time, | ||
259 | card->perf_stats.outbound_cnt, | ||
260 | card->perf_stats.outbound_do_qdio_time, | ||
261 | card->perf_stats.outbound_do_qdio_cnt | ||
262 | ); | ||
263 | put_device(device); | ||
264 | return 0; | ||
265 | } | ||
266 | |||
267 | static const struct seq_operations qeth_perf_procfile_seq_ops = { | ||
268 | .start = qeth_procfile_seq_start, | ||
269 | .stop = qeth_procfile_seq_stop, | ||
270 | .next = qeth_procfile_seq_next, | ||
271 | .show = qeth_perf_procfile_seq_show, | ||
272 | }; | ||
273 | |||
274 | static int | ||
275 | qeth_perf_procfile_open(struct inode *inode, struct file *file) | ||
276 | { | ||
277 | return seq_open(file, &qeth_perf_procfile_seq_ops); | ||
278 | } | ||
279 | |||
280 | static const struct file_operations qeth_perf_procfile_fops = { | ||
281 | .owner = THIS_MODULE, | ||
282 | .open = qeth_perf_procfile_open, | ||
283 | .read = seq_read, | ||
284 | .llseek = seq_lseek, | ||
285 | .release = seq_release, | ||
286 | }; | ||
287 | |||
288 | int __init | ||
289 | qeth_create_procfs_entries(void) | ||
290 | { | ||
291 | qeth_procfile = create_proc_entry(QETH_PROCFILE_NAME, | ||
292 | S_IFREG | 0444, NULL); | ||
293 | if (qeth_procfile) | ||
294 | qeth_procfile->proc_fops = &qeth_procfile_fops; | ||
295 | |||
296 | qeth_perf_procfile = create_proc_entry(QETH_PERF_PROCFILE_NAME, | ||
297 | S_IFREG | 0444, NULL); | ||
298 | if (qeth_perf_procfile) | ||
299 | qeth_perf_procfile->proc_fops = &qeth_perf_procfile_fops; | ||
300 | |||
301 | if (qeth_procfile && | ||
302 | qeth_perf_procfile) | ||
303 | return 0; | ||
304 | else | ||
305 | return -ENOMEM; | ||
306 | } | ||
307 | |||
308 | void __exit | ||
309 | qeth_remove_procfs_entries(void) | ||
310 | { | ||
311 | if (qeth_procfile) | ||
312 | remove_proc_entry(QETH_PROCFILE_NAME, NULL); | ||
313 | if (qeth_perf_procfile) | ||
314 | remove_proc_entry(QETH_PERF_PROCFILE_NAME, NULL); | ||
315 | } | ||
316 | |||
diff --git a/drivers/s390/net/qeth_sys.c b/drivers/s390/net/qeth_sys.c deleted file mode 100644 index 2cc3f3a0e393..000000000000 --- a/drivers/s390/net/qeth_sys.c +++ /dev/null | |||
@@ -1,1858 +0,0 @@ | |||
1 | /* | ||
2 | * | ||
3 | * linux/drivers/s390/net/qeth_sys.c | ||
4 | * | ||
5 | * Linux on zSeries OSA Express and HiperSockets support | ||
6 | * This file contains code related to sysfs. | ||
7 | * | ||
8 | * Copyright 2000,2003 IBM Corporation | ||
9 | * | ||
10 | * Author(s): Thomas Spatzier <tspat@de.ibm.com> | ||
11 | * Frank Pavlic <fpavlic@de.ibm.com> | ||
12 | * | ||
13 | */ | ||
14 | #include <linux/list.h> | ||
15 | #include <linux/rwsem.h> | ||
16 | |||
17 | #include <asm/ebcdic.h> | ||
18 | |||
19 | #include "qeth.h" | ||
20 | #include "qeth_mpc.h" | ||
21 | #include "qeth_fs.h" | ||
22 | |||
23 | /*****************************************************************************/ | ||
24 | /* */ | ||
25 | /* /sys-fs stuff UNDER DEVELOPMENT !!! */ | ||
26 | /* */ | ||
27 | /*****************************************************************************/ | ||
28 | //low/high watermark | ||
29 | |||
30 | static ssize_t | ||
31 | qeth_dev_state_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
32 | { | ||
33 | struct qeth_card *card = dev->driver_data; | ||
34 | if (!card) | ||
35 | return -EINVAL; | ||
36 | |||
37 | switch (card->state) { | ||
38 | case CARD_STATE_DOWN: | ||
39 | return sprintf(buf, "DOWN\n"); | ||
40 | case CARD_STATE_HARDSETUP: | ||
41 | return sprintf(buf, "HARDSETUP\n"); | ||
42 | case CARD_STATE_SOFTSETUP: | ||
43 | return sprintf(buf, "SOFTSETUP\n"); | ||
44 | case CARD_STATE_UP: | ||
45 | if (card->lan_online) | ||
46 | return sprintf(buf, "UP (LAN ONLINE)\n"); | ||
47 | else | ||
48 | return sprintf(buf, "UP (LAN OFFLINE)\n"); | ||
49 | case CARD_STATE_RECOVER: | ||
50 | return sprintf(buf, "RECOVER\n"); | ||
51 | default: | ||
52 | return sprintf(buf, "UNKNOWN\n"); | ||
53 | } | ||
54 | } | ||
55 | |||
56 | static DEVICE_ATTR(state, 0444, qeth_dev_state_show, NULL); | ||
57 | |||
58 | static ssize_t | ||
59 | qeth_dev_chpid_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
60 | { | ||
61 | struct qeth_card *card = dev->driver_data; | ||
62 | if (!card) | ||
63 | return -EINVAL; | ||
64 | |||
65 | return sprintf(buf, "%02X\n", card->info.chpid); | ||
66 | } | ||
67 | |||
68 | static DEVICE_ATTR(chpid, 0444, qeth_dev_chpid_show, NULL); | ||
69 | |||
70 | static ssize_t | ||
71 | qeth_dev_if_name_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
72 | { | ||
73 | struct qeth_card *card = dev->driver_data; | ||
74 | if (!card) | ||
75 | return -EINVAL; | ||
76 | return sprintf(buf, "%s\n", QETH_CARD_IFNAME(card)); | ||
77 | } | ||
78 | |||
79 | static DEVICE_ATTR(if_name, 0444, qeth_dev_if_name_show, NULL); | ||
80 | |||
81 | static ssize_t | ||
82 | qeth_dev_card_type_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
83 | { | ||
84 | struct qeth_card *card = dev->driver_data; | ||
85 | if (!card) | ||
86 | return -EINVAL; | ||
87 | |||
88 | return sprintf(buf, "%s\n", qeth_get_cardname_short(card)); | ||
89 | } | ||
90 | |||
91 | static DEVICE_ATTR(card_type, 0444, qeth_dev_card_type_show, NULL); | ||
92 | |||
93 | static ssize_t | ||
94 | qeth_dev_portno_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
95 | { | ||
96 | struct qeth_card *card = dev->driver_data; | ||
97 | if (!card) | ||
98 | return -EINVAL; | ||
99 | |||
100 | return sprintf(buf, "%i\n", card->info.portno); | ||
101 | } | ||
102 | |||
103 | static ssize_t | ||
104 | qeth_dev_portno_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
105 | { | ||
106 | struct qeth_card *card = dev->driver_data; | ||
107 | char *tmp; | ||
108 | unsigned int portno; | ||
109 | |||
110 | if (!card) | ||
111 | return -EINVAL; | ||
112 | |||
113 | if ((card->state != CARD_STATE_DOWN) && | ||
114 | (card->state != CARD_STATE_RECOVER)) | ||
115 | return -EPERM; | ||
116 | |||
117 | portno = simple_strtoul(buf, &tmp, 16); | ||
118 | if (portno > MAX_PORTNO){ | ||
119 | PRINT_WARN("portno 0x%X is out of range\n", portno); | ||
120 | return -EINVAL; | ||
121 | } | ||
122 | |||
123 | card->info.portno = portno; | ||
124 | return count; | ||
125 | } | ||
126 | |||
127 | static DEVICE_ATTR(portno, 0644, qeth_dev_portno_show, qeth_dev_portno_store); | ||
128 | |||
129 | static ssize_t | ||
130 | qeth_dev_portname_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
131 | { | ||
132 | struct qeth_card *card = dev->driver_data; | ||
133 | char portname[9] = {0, }; | ||
134 | |||
135 | if (!card) | ||
136 | return -EINVAL; | ||
137 | |||
138 | if (card->info.portname_required) { | ||
139 | memcpy(portname, card->info.portname + 1, 8); | ||
140 | EBCASC(portname, 8); | ||
141 | return sprintf(buf, "%s\n", portname); | ||
142 | } else | ||
143 | return sprintf(buf, "no portname required\n"); | ||
144 | } | ||
145 | |||
146 | static ssize_t | ||
147 | qeth_dev_portname_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
148 | { | ||
149 | struct qeth_card *card = dev->driver_data; | ||
150 | char *tmp; | ||
151 | int i; | ||
152 | |||
153 | if (!card) | ||
154 | return -EINVAL; | ||
155 | |||
156 | if ((card->state != CARD_STATE_DOWN) && | ||
157 | (card->state != CARD_STATE_RECOVER)) | ||
158 | return -EPERM; | ||
159 | |||
160 | tmp = strsep((char **) &buf, "\n"); | ||
161 | if ((strlen(tmp) > 8) || (strlen(tmp) == 0)) | ||
162 | return -EINVAL; | ||
163 | |||
164 | card->info.portname[0] = strlen(tmp); | ||
165 | /* for beauty reasons */ | ||
166 | for (i = 1; i < 9; i++) | ||
167 | card->info.portname[i] = ' '; | ||
168 | strcpy(card->info.portname + 1, tmp); | ||
169 | ASCEBC(card->info.portname + 1, 8); | ||
170 | |||
171 | return count; | ||
172 | } | ||
173 | |||
174 | static DEVICE_ATTR(portname, 0644, qeth_dev_portname_show, | ||
175 | qeth_dev_portname_store); | ||
176 | |||
177 | static ssize_t | ||
178 | qeth_dev_checksum_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
179 | { | ||
180 | struct qeth_card *card = dev->driver_data; | ||
181 | |||
182 | if (!card) | ||
183 | return -EINVAL; | ||
184 | |||
185 | return sprintf(buf, "%s checksumming\n", qeth_get_checksum_str(card)); | ||
186 | } | ||
187 | |||
188 | static ssize_t | ||
189 | qeth_dev_checksum_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
190 | { | ||
191 | struct qeth_card *card = dev->driver_data; | ||
192 | char *tmp; | ||
193 | |||
194 | if (!card) | ||
195 | return -EINVAL; | ||
196 | |||
197 | if ((card->state != CARD_STATE_DOWN) && | ||
198 | (card->state != CARD_STATE_RECOVER)) | ||
199 | return -EPERM; | ||
200 | |||
201 | tmp = strsep((char **) &buf, "\n"); | ||
202 | if (!strcmp(tmp, "sw_checksumming")) | ||
203 | card->options.checksum_type = SW_CHECKSUMMING; | ||
204 | else if (!strcmp(tmp, "hw_checksumming")) | ||
205 | card->options.checksum_type = HW_CHECKSUMMING; | ||
206 | else if (!strcmp(tmp, "no_checksumming")) | ||
207 | card->options.checksum_type = NO_CHECKSUMMING; | ||
208 | else { | ||
209 | PRINT_WARN("Unknown checksumming type '%s'\n", tmp); | ||
210 | return -EINVAL; | ||
211 | } | ||
212 | return count; | ||
213 | } | ||
214 | |||
215 | static DEVICE_ATTR(checksumming, 0644, qeth_dev_checksum_show, | ||
216 | qeth_dev_checksum_store); | ||
217 | |||
218 | static ssize_t | ||
219 | qeth_dev_prioqing_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
220 | { | ||
221 | struct qeth_card *card = dev->driver_data; | ||
222 | |||
223 | if (!card) | ||
224 | return -EINVAL; | ||
225 | |||
226 | switch (card->qdio.do_prio_queueing) { | ||
227 | case QETH_PRIO_Q_ING_PREC: | ||
228 | return sprintf(buf, "%s\n", "by precedence"); | ||
229 | case QETH_PRIO_Q_ING_TOS: | ||
230 | return sprintf(buf, "%s\n", "by type of service"); | ||
231 | default: | ||
232 | return sprintf(buf, "always queue %i\n", | ||
233 | card->qdio.default_out_queue); | ||
234 | } | ||
235 | } | ||
236 | |||
237 | static ssize_t | ||
238 | qeth_dev_prioqing_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
239 | { | ||
240 | struct qeth_card *card = dev->driver_data; | ||
241 | char *tmp; | ||
242 | |||
243 | if (!card) | ||
244 | return -EINVAL; | ||
245 | |||
246 | if ((card->state != CARD_STATE_DOWN) && | ||
247 | (card->state != CARD_STATE_RECOVER)) | ||
248 | return -EPERM; | ||
249 | |||
250 | /* check if 1920 devices are supported , | ||
251 | * if though we have to permit priority queueing | ||
252 | */ | ||
253 | if (card->qdio.no_out_queues == 1) { | ||
254 | PRINT_WARN("Priority queueing disabled due " | ||
255 | "to hardware limitations!\n"); | ||
256 | card->qdio.do_prio_queueing = QETH_PRIOQ_DEFAULT; | ||
257 | return -EPERM; | ||
258 | } | ||
259 | |||
260 | tmp = strsep((char **) &buf, "\n"); | ||
261 | if (!strcmp(tmp, "prio_queueing_prec")) | ||
262 | card->qdio.do_prio_queueing = QETH_PRIO_Q_ING_PREC; | ||
263 | else if (!strcmp(tmp, "prio_queueing_tos")) | ||
264 | card->qdio.do_prio_queueing = QETH_PRIO_Q_ING_TOS; | ||
265 | else if (!strcmp(tmp, "no_prio_queueing:0")) { | ||
266 | card->qdio.do_prio_queueing = QETH_NO_PRIO_QUEUEING; | ||
267 | card->qdio.default_out_queue = 0; | ||
268 | } else if (!strcmp(tmp, "no_prio_queueing:1")) { | ||
269 | card->qdio.do_prio_queueing = QETH_NO_PRIO_QUEUEING; | ||
270 | card->qdio.default_out_queue = 1; | ||
271 | } else if (!strcmp(tmp, "no_prio_queueing:2")) { | ||
272 | card->qdio.do_prio_queueing = QETH_NO_PRIO_QUEUEING; | ||
273 | card->qdio.default_out_queue = 2; | ||
274 | } else if (!strcmp(tmp, "no_prio_queueing:3")) { | ||
275 | card->qdio.do_prio_queueing = QETH_NO_PRIO_QUEUEING; | ||
276 | card->qdio.default_out_queue = 3; | ||
277 | } else if (!strcmp(tmp, "no_prio_queueing")) { | ||
278 | card->qdio.do_prio_queueing = QETH_NO_PRIO_QUEUEING; | ||
279 | card->qdio.default_out_queue = QETH_DEFAULT_QUEUE; | ||
280 | } else { | ||
281 | PRINT_WARN("Unknown queueing type '%s'\n", tmp); | ||
282 | return -EINVAL; | ||
283 | } | ||
284 | return count; | ||
285 | } | ||
286 | |||
287 | static DEVICE_ATTR(priority_queueing, 0644, qeth_dev_prioqing_show, | ||
288 | qeth_dev_prioqing_store); | ||
289 | |||
290 | static ssize_t | ||
291 | qeth_dev_bufcnt_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
292 | { | ||
293 | struct qeth_card *card = dev->driver_data; | ||
294 | |||
295 | if (!card) | ||
296 | return -EINVAL; | ||
297 | |||
298 | return sprintf(buf, "%i\n", card->qdio.in_buf_pool.buf_count); | ||
299 | } | ||
300 | |||
301 | static ssize_t | ||
302 | qeth_dev_bufcnt_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
303 | { | ||
304 | struct qeth_card *card = dev->driver_data; | ||
305 | char *tmp; | ||
306 | int cnt, old_cnt; | ||
307 | int rc; | ||
308 | |||
309 | if (!card) | ||
310 | return -EINVAL; | ||
311 | |||
312 | if ((card->state != CARD_STATE_DOWN) && | ||
313 | (card->state != CARD_STATE_RECOVER)) | ||
314 | return -EPERM; | ||
315 | |||
316 | old_cnt = card->qdio.in_buf_pool.buf_count; | ||
317 | cnt = simple_strtoul(buf, &tmp, 10); | ||
318 | cnt = (cnt < QETH_IN_BUF_COUNT_MIN) ? QETH_IN_BUF_COUNT_MIN : | ||
319 | ((cnt > QETH_IN_BUF_COUNT_MAX) ? QETH_IN_BUF_COUNT_MAX : cnt); | ||
320 | if (old_cnt != cnt) { | ||
321 | if ((rc = qeth_realloc_buffer_pool(card, cnt))) | ||
322 | PRINT_WARN("Error (%d) while setting " | ||
323 | "buffer count.\n", rc); | ||
324 | } | ||
325 | return count; | ||
326 | } | ||
327 | |||
328 | static DEVICE_ATTR(buffer_count, 0644, qeth_dev_bufcnt_show, | ||
329 | qeth_dev_bufcnt_store); | ||
330 | |||
331 | static ssize_t | ||
332 | qeth_dev_route_show(struct qeth_card *card, struct qeth_routing_info *route, | ||
333 | char *buf) | ||
334 | { | ||
335 | switch (route->type) { | ||
336 | case PRIMARY_ROUTER: | ||
337 | return sprintf(buf, "%s\n", "primary router"); | ||
338 | case SECONDARY_ROUTER: | ||
339 | return sprintf(buf, "%s\n", "secondary router"); | ||
340 | case MULTICAST_ROUTER: | ||
341 | if (card->info.broadcast_capable == QETH_BROADCAST_WITHOUT_ECHO) | ||
342 | return sprintf(buf, "%s\n", "multicast router+"); | ||
343 | else | ||
344 | return sprintf(buf, "%s\n", "multicast router"); | ||
345 | case PRIMARY_CONNECTOR: | ||
346 | if (card->info.broadcast_capable == QETH_BROADCAST_WITHOUT_ECHO) | ||
347 | return sprintf(buf, "%s\n", "primary connector+"); | ||
348 | else | ||
349 | return sprintf(buf, "%s\n", "primary connector"); | ||
350 | case SECONDARY_CONNECTOR: | ||
351 | if (card->info.broadcast_capable == QETH_BROADCAST_WITHOUT_ECHO) | ||
352 | return sprintf(buf, "%s\n", "secondary connector+"); | ||
353 | else | ||
354 | return sprintf(buf, "%s\n", "secondary connector"); | ||
355 | default: | ||
356 | return sprintf(buf, "%s\n", "no"); | ||
357 | } | ||
358 | } | ||
359 | |||
360 | static ssize_t | ||
361 | qeth_dev_route4_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
362 | { | ||
363 | struct qeth_card *card = dev->driver_data; | ||
364 | |||
365 | if (!card) | ||
366 | return -EINVAL; | ||
367 | |||
368 | return qeth_dev_route_show(card, &card->options.route4, buf); | ||
369 | } | ||
370 | |||
371 | static ssize_t | ||
372 | qeth_dev_route_store(struct qeth_card *card, struct qeth_routing_info *route, | ||
373 | enum qeth_prot_versions prot, const char *buf, size_t count) | ||
374 | { | ||
375 | enum qeth_routing_types old_route_type = route->type; | ||
376 | char *tmp; | ||
377 | int rc; | ||
378 | |||
379 | tmp = strsep((char **) &buf, "\n"); | ||
380 | |||
381 | if (!strcmp(tmp, "no_router")){ | ||
382 | route->type = NO_ROUTER; | ||
383 | } else if (!strcmp(tmp, "primary_connector")) { | ||
384 | route->type = PRIMARY_CONNECTOR; | ||
385 | } else if (!strcmp(tmp, "secondary_connector")) { | ||
386 | route->type = SECONDARY_CONNECTOR; | ||
387 | } else if (!strcmp(tmp, "primary_router")) { | ||
388 | route->type = PRIMARY_ROUTER; | ||
389 | } else if (!strcmp(tmp, "secondary_router")) { | ||
390 | route->type = SECONDARY_ROUTER; | ||
391 | } else if (!strcmp(tmp, "multicast_router")) { | ||
392 | route->type = MULTICAST_ROUTER; | ||
393 | } else { | ||
394 | PRINT_WARN("Invalid routing type '%s'.\n", tmp); | ||
395 | return -EINVAL; | ||
396 | } | ||
397 | if (((card->state == CARD_STATE_SOFTSETUP) || | ||
398 | (card->state == CARD_STATE_UP)) && | ||
399 | (old_route_type != route->type)){ | ||
400 | if (prot == QETH_PROT_IPV4) | ||
401 | rc = qeth_setrouting_v4(card); | ||
402 | else if (prot == QETH_PROT_IPV6) | ||
403 | rc = qeth_setrouting_v6(card); | ||
404 | } | ||
405 | return count; | ||
406 | } | ||
407 | |||
408 | static ssize_t | ||
409 | qeth_dev_route4_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
410 | { | ||
411 | struct qeth_card *card = dev->driver_data; | ||
412 | |||
413 | if (!card) | ||
414 | return -EINVAL; | ||
415 | |||
416 | return qeth_dev_route_store(card, &card->options.route4, | ||
417 | QETH_PROT_IPV4, buf, count); | ||
418 | } | ||
419 | |||
420 | static DEVICE_ATTR(route4, 0644, qeth_dev_route4_show, qeth_dev_route4_store); | ||
421 | |||
422 | #ifdef CONFIG_QETH_IPV6 | ||
423 | static ssize_t | ||
424 | qeth_dev_route6_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
425 | { | ||
426 | struct qeth_card *card = dev->driver_data; | ||
427 | |||
428 | if (!card) | ||
429 | return -EINVAL; | ||
430 | |||
431 | if (!qeth_is_supported(card, IPA_IPV6)) | ||
432 | return sprintf(buf, "%s\n", "n/a"); | ||
433 | |||
434 | return qeth_dev_route_show(card, &card->options.route6, buf); | ||
435 | } | ||
436 | |||
437 | static ssize_t | ||
438 | qeth_dev_route6_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
439 | { | ||
440 | struct qeth_card *card = dev->driver_data; | ||
441 | |||
442 | if (!card) | ||
443 | return -EINVAL; | ||
444 | |||
445 | if (!qeth_is_supported(card, IPA_IPV6)){ | ||
446 | PRINT_WARN("IPv6 not supported for interface %s.\n" | ||
447 | "Routing status no changed.\n", | ||
448 | QETH_CARD_IFNAME(card)); | ||
449 | return -ENOTSUPP; | ||
450 | } | ||
451 | |||
452 | return qeth_dev_route_store(card, &card->options.route6, | ||
453 | QETH_PROT_IPV6, buf, count); | ||
454 | } | ||
455 | |||
456 | static DEVICE_ATTR(route6, 0644, qeth_dev_route6_show, qeth_dev_route6_store); | ||
457 | #endif | ||
458 | |||
459 | static ssize_t | ||
460 | qeth_dev_add_hhlen_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
461 | { | ||
462 | struct qeth_card *card = dev->driver_data; | ||
463 | |||
464 | if (!card) | ||
465 | return -EINVAL; | ||
466 | |||
467 | return sprintf(buf, "%i\n", card->options.add_hhlen); | ||
468 | } | ||
469 | |||
470 | static ssize_t | ||
471 | qeth_dev_add_hhlen_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
472 | { | ||
473 | struct qeth_card *card = dev->driver_data; | ||
474 | char *tmp; | ||
475 | int i; | ||
476 | |||
477 | if (!card) | ||
478 | return -EINVAL; | ||
479 | |||
480 | if ((card->state != CARD_STATE_DOWN) && | ||
481 | (card->state != CARD_STATE_RECOVER)) | ||
482 | return -EPERM; | ||
483 | |||
484 | i = simple_strtoul(buf, &tmp, 10); | ||
485 | if ((i < 0) || (i > MAX_ADD_HHLEN)) { | ||
486 | PRINT_WARN("add_hhlen out of range\n"); | ||
487 | return -EINVAL; | ||
488 | } | ||
489 | card->options.add_hhlen = i; | ||
490 | |||
491 | return count; | ||
492 | } | ||
493 | |||
494 | static DEVICE_ATTR(add_hhlen, 0644, qeth_dev_add_hhlen_show, | ||
495 | qeth_dev_add_hhlen_store); | ||
496 | |||
497 | static ssize_t | ||
498 | qeth_dev_fake_ll_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
499 | { | ||
500 | struct qeth_card *card = dev->driver_data; | ||
501 | |||
502 | if (!card) | ||
503 | return -EINVAL; | ||
504 | |||
505 | return sprintf(buf, "%i\n", card->options.fake_ll? 1:0); | ||
506 | } | ||
507 | |||
508 | static ssize_t | ||
509 | qeth_dev_fake_ll_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
510 | { | ||
511 | struct qeth_card *card = dev->driver_data; | ||
512 | char *tmp; | ||
513 | int i; | ||
514 | |||
515 | if (!card) | ||
516 | return -EINVAL; | ||
517 | |||
518 | if ((card->state != CARD_STATE_DOWN) && | ||
519 | (card->state != CARD_STATE_RECOVER)) | ||
520 | return -EPERM; | ||
521 | |||
522 | i = simple_strtoul(buf, &tmp, 16); | ||
523 | if ((i != 0) && (i != 1)) { | ||
524 | PRINT_WARN("fake_ll: write 0 or 1 to this file!\n"); | ||
525 | return -EINVAL; | ||
526 | } | ||
527 | card->options.fake_ll = i; | ||
528 | return count; | ||
529 | } | ||
530 | |||
531 | static DEVICE_ATTR(fake_ll, 0644, qeth_dev_fake_ll_show, | ||
532 | qeth_dev_fake_ll_store); | ||
533 | |||
534 | static ssize_t | ||
535 | qeth_dev_fake_broadcast_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
536 | { | ||
537 | struct qeth_card *card = dev->driver_data; | ||
538 | |||
539 | if (!card) | ||
540 | return -EINVAL; | ||
541 | |||
542 | return sprintf(buf, "%i\n", card->options.fake_broadcast? 1:0); | ||
543 | } | ||
544 | |||
545 | static ssize_t | ||
546 | qeth_dev_fake_broadcast_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
547 | { | ||
548 | struct qeth_card *card = dev->driver_data; | ||
549 | char *tmp; | ||
550 | int i; | ||
551 | |||
552 | if (!card) | ||
553 | return -EINVAL; | ||
554 | |||
555 | if ((card->state != CARD_STATE_DOWN) && | ||
556 | (card->state != CARD_STATE_RECOVER)) | ||
557 | return -EPERM; | ||
558 | |||
559 | i = simple_strtoul(buf, &tmp, 16); | ||
560 | if ((i == 0) || (i == 1)) | ||
561 | card->options.fake_broadcast = i; | ||
562 | else { | ||
563 | PRINT_WARN("fake_broadcast: write 0 or 1 to this file!\n"); | ||
564 | return -EINVAL; | ||
565 | } | ||
566 | return count; | ||
567 | } | ||
568 | |||
569 | static DEVICE_ATTR(fake_broadcast, 0644, qeth_dev_fake_broadcast_show, | ||
570 | qeth_dev_fake_broadcast_store); | ||
571 | |||
572 | static ssize_t | ||
573 | qeth_dev_recover_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
574 | { | ||
575 | struct qeth_card *card = dev->driver_data; | ||
576 | char *tmp; | ||
577 | int i; | ||
578 | |||
579 | if (!card) | ||
580 | return -EINVAL; | ||
581 | |||
582 | if (card->state != CARD_STATE_UP) | ||
583 | return -EPERM; | ||
584 | |||
585 | i = simple_strtoul(buf, &tmp, 16); | ||
586 | if (i == 1) | ||
587 | qeth_schedule_recovery(card); | ||
588 | |||
589 | return count; | ||
590 | } | ||
591 | |||
592 | static DEVICE_ATTR(recover, 0200, NULL, qeth_dev_recover_store); | ||
593 | |||
594 | static ssize_t | ||
595 | qeth_dev_broadcast_mode_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
596 | { | ||
597 | struct qeth_card *card = dev->driver_data; | ||
598 | |||
599 | if (!card) | ||
600 | return -EINVAL; | ||
601 | |||
602 | if (!((card->info.link_type == QETH_LINK_TYPE_HSTR) || | ||
603 | (card->info.link_type == QETH_LINK_TYPE_LANE_TR))) | ||
604 | return sprintf(buf, "n/a\n"); | ||
605 | |||
606 | return sprintf(buf, "%s\n", (card->options.broadcast_mode == | ||
607 | QETH_TR_BROADCAST_ALLRINGS)? | ||
608 | "all rings":"local"); | ||
609 | } | ||
610 | |||
611 | static ssize_t | ||
612 | qeth_dev_broadcast_mode_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
613 | { | ||
614 | struct qeth_card *card = dev->driver_data; | ||
615 | char *tmp; | ||
616 | |||
617 | if (!card) | ||
618 | return -EINVAL; | ||
619 | |||
620 | if ((card->state != CARD_STATE_DOWN) && | ||
621 | (card->state != CARD_STATE_RECOVER)) | ||
622 | return -EPERM; | ||
623 | |||
624 | if (!((card->info.link_type == QETH_LINK_TYPE_HSTR) || | ||
625 | (card->info.link_type == QETH_LINK_TYPE_LANE_TR))){ | ||
626 | PRINT_WARN("Device is not a tokenring device!\n"); | ||
627 | return -EINVAL; | ||
628 | } | ||
629 | |||
630 | tmp = strsep((char **) &buf, "\n"); | ||
631 | |||
632 | if (!strcmp(tmp, "local")){ | ||
633 | card->options.broadcast_mode = QETH_TR_BROADCAST_LOCAL; | ||
634 | return count; | ||
635 | } else if (!strcmp(tmp, "all_rings")) { | ||
636 | card->options.broadcast_mode = QETH_TR_BROADCAST_ALLRINGS; | ||
637 | return count; | ||
638 | } else { | ||
639 | PRINT_WARN("broadcast_mode: invalid mode %s!\n", | ||
640 | tmp); | ||
641 | return -EINVAL; | ||
642 | } | ||
643 | return count; | ||
644 | } | ||
645 | |||
646 | static DEVICE_ATTR(broadcast_mode, 0644, qeth_dev_broadcast_mode_show, | ||
647 | qeth_dev_broadcast_mode_store); | ||
648 | |||
649 | static ssize_t | ||
650 | qeth_dev_canonical_macaddr_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
651 | { | ||
652 | struct qeth_card *card = dev->driver_data; | ||
653 | |||
654 | if (!card) | ||
655 | return -EINVAL; | ||
656 | |||
657 | if (!((card->info.link_type == QETH_LINK_TYPE_HSTR) || | ||
658 | (card->info.link_type == QETH_LINK_TYPE_LANE_TR))) | ||
659 | return sprintf(buf, "n/a\n"); | ||
660 | |||
661 | return sprintf(buf, "%i\n", (card->options.macaddr_mode == | ||
662 | QETH_TR_MACADDR_CANONICAL)? 1:0); | ||
663 | } | ||
664 | |||
665 | static ssize_t | ||
666 | qeth_dev_canonical_macaddr_store(struct device *dev, struct device_attribute *attr, const char *buf, | ||
667 | size_t count) | ||
668 | { | ||
669 | struct qeth_card *card = dev->driver_data; | ||
670 | char *tmp; | ||
671 | int i; | ||
672 | |||
673 | if (!card) | ||
674 | return -EINVAL; | ||
675 | |||
676 | if ((card->state != CARD_STATE_DOWN) && | ||
677 | (card->state != CARD_STATE_RECOVER)) | ||
678 | return -EPERM; | ||
679 | |||
680 | if (!((card->info.link_type == QETH_LINK_TYPE_HSTR) || | ||
681 | (card->info.link_type == QETH_LINK_TYPE_LANE_TR))){ | ||
682 | PRINT_WARN("Device is not a tokenring device!\n"); | ||
683 | return -EINVAL; | ||
684 | } | ||
685 | |||
686 | i = simple_strtoul(buf, &tmp, 16); | ||
687 | if ((i == 0) || (i == 1)) | ||
688 | card->options.macaddr_mode = i? | ||
689 | QETH_TR_MACADDR_CANONICAL : | ||
690 | QETH_TR_MACADDR_NONCANONICAL; | ||
691 | else { | ||
692 | PRINT_WARN("canonical_macaddr: write 0 or 1 to this file!\n"); | ||
693 | return -EINVAL; | ||
694 | } | ||
695 | return count; | ||
696 | } | ||
697 | |||
698 | static DEVICE_ATTR(canonical_macaddr, 0644, qeth_dev_canonical_macaddr_show, | ||
699 | qeth_dev_canonical_macaddr_store); | ||
700 | |||
701 | static ssize_t | ||
702 | qeth_dev_layer2_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
703 | { | ||
704 | struct qeth_card *card = dev->driver_data; | ||
705 | |||
706 | if (!card) | ||
707 | return -EINVAL; | ||
708 | |||
709 | return sprintf(buf, "%i\n", card->options.layer2 ? 1:0); | ||
710 | } | ||
711 | |||
712 | static ssize_t | ||
713 | qeth_dev_layer2_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
714 | { | ||
715 | struct qeth_card *card = dev->driver_data; | ||
716 | char *tmp; | ||
717 | int i; | ||
718 | |||
719 | if (!card) | ||
720 | return -EINVAL; | ||
721 | if (card->info.type == QETH_CARD_TYPE_IQD) { | ||
722 | PRINT_WARN("Layer2 on Hipersockets is not supported! \n"); | ||
723 | return -EPERM; | ||
724 | } | ||
725 | |||
726 | if (((card->state != CARD_STATE_DOWN) && | ||
727 | (card->state != CARD_STATE_RECOVER))) | ||
728 | return -EPERM; | ||
729 | |||
730 | i = simple_strtoul(buf, &tmp, 16); | ||
731 | if ((i == 0) || (i == 1)) | ||
732 | card->options.layer2 = i; | ||
733 | else { | ||
734 | PRINT_WARN("layer2: write 0 or 1 to this file!\n"); | ||
735 | return -EINVAL; | ||
736 | } | ||
737 | return count; | ||
738 | } | ||
739 | |||
740 | static DEVICE_ATTR(layer2, 0644, qeth_dev_layer2_show, | ||
741 | qeth_dev_layer2_store); | ||
742 | |||
743 | static ssize_t | ||
744 | qeth_dev_performance_stats_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
745 | { | ||
746 | struct qeth_card *card = dev->driver_data; | ||
747 | |||
748 | if (!card) | ||
749 | return -EINVAL; | ||
750 | |||
751 | return sprintf(buf, "%i\n", card->options.performance_stats ? 1:0); | ||
752 | } | ||
753 | |||
754 | static ssize_t | ||
755 | qeth_dev_performance_stats_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
756 | { | ||
757 | struct qeth_card *card = dev->driver_data; | ||
758 | char *tmp; | ||
759 | int i; | ||
760 | |||
761 | if (!card) | ||
762 | return -EINVAL; | ||
763 | |||
764 | i = simple_strtoul(buf, &tmp, 16); | ||
765 | if ((i == 0) || (i == 1)) { | ||
766 | if (i == card->options.performance_stats) | ||
767 | return count; | ||
768 | card->options.performance_stats = i; | ||
769 | if (i == 0) | ||
770 | memset(&card->perf_stats, 0, | ||
771 | sizeof(struct qeth_perf_stats)); | ||
772 | card->perf_stats.initial_rx_packets = card->stats.rx_packets; | ||
773 | card->perf_stats.initial_tx_packets = card->stats.tx_packets; | ||
774 | } else { | ||
775 | PRINT_WARN("performance_stats: write 0 or 1 to this file!\n"); | ||
776 | return -EINVAL; | ||
777 | } | ||
778 | return count; | ||
779 | } | ||
780 | |||
781 | static DEVICE_ATTR(performance_stats, 0644, qeth_dev_performance_stats_show, | ||
782 | qeth_dev_performance_stats_store); | ||
783 | |||
784 | static ssize_t | ||
785 | qeth_dev_large_send_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
786 | { | ||
787 | struct qeth_card *card = dev->driver_data; | ||
788 | |||
789 | if (!card) | ||
790 | return -EINVAL; | ||
791 | |||
792 | switch (card->options.large_send) { | ||
793 | case QETH_LARGE_SEND_NO: | ||
794 | return sprintf(buf, "%s\n", "no"); | ||
795 | case QETH_LARGE_SEND_EDDP: | ||
796 | return sprintf(buf, "%s\n", "EDDP"); | ||
797 | case QETH_LARGE_SEND_TSO: | ||
798 | return sprintf(buf, "%s\n", "TSO"); | ||
799 | default: | ||
800 | return sprintf(buf, "%s\n", "N/A"); | ||
801 | } | ||
802 | } | ||
803 | |||
804 | static ssize_t | ||
805 | qeth_dev_large_send_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
806 | { | ||
807 | struct qeth_card *card = dev->driver_data; | ||
808 | enum qeth_large_send_types type; | ||
809 | int rc = 0; | ||
810 | char *tmp; | ||
811 | |||
812 | if (!card) | ||
813 | return -EINVAL; | ||
814 | tmp = strsep((char **) &buf, "\n"); | ||
815 | if (!strcmp(tmp, "no")){ | ||
816 | type = QETH_LARGE_SEND_NO; | ||
817 | } else if (!strcmp(tmp, "EDDP")) { | ||
818 | type = QETH_LARGE_SEND_EDDP; | ||
819 | } else if (!strcmp(tmp, "TSO")) { | ||
820 | type = QETH_LARGE_SEND_TSO; | ||
821 | } else { | ||
822 | PRINT_WARN("large_send: invalid mode %s!\n", tmp); | ||
823 | return -EINVAL; | ||
824 | } | ||
825 | if (card->options.large_send == type) | ||
826 | return count; | ||
827 | if ((rc = qeth_set_large_send(card, type))) | ||
828 | return rc; | ||
829 | return count; | ||
830 | } | ||
831 | |||
832 | static DEVICE_ATTR(large_send, 0644, qeth_dev_large_send_show, | ||
833 | qeth_dev_large_send_store); | ||
834 | |||
835 | static ssize_t | ||
836 | qeth_dev_blkt_show(char *buf, struct qeth_card *card, int value ) | ||
837 | { | ||
838 | |||
839 | if (!card) | ||
840 | return -EINVAL; | ||
841 | |||
842 | return sprintf(buf, "%i\n", value); | ||
843 | } | ||
844 | |||
845 | static ssize_t | ||
846 | qeth_dev_blkt_store(struct qeth_card *card, const char *buf, size_t count, | ||
847 | int *value, int max_value) | ||
848 | { | ||
849 | char *tmp; | ||
850 | int i; | ||
851 | |||
852 | if (!card) | ||
853 | return -EINVAL; | ||
854 | |||
855 | if ((card->state != CARD_STATE_DOWN) && | ||
856 | (card->state != CARD_STATE_RECOVER)) | ||
857 | return -EPERM; | ||
858 | |||
859 | i = simple_strtoul(buf, &tmp, 10); | ||
860 | if (i <= max_value) { | ||
861 | *value = i; | ||
862 | } else { | ||
863 | PRINT_WARN("blkt total time: write values between" | ||
864 | " 0 and %d to this file!\n", max_value); | ||
865 | return -EINVAL; | ||
866 | } | ||
867 | return count; | ||
868 | } | ||
869 | |||
870 | static ssize_t | ||
871 | qeth_dev_blkt_total_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
872 | { | ||
873 | struct qeth_card *card = dev->driver_data; | ||
874 | |||
875 | return qeth_dev_blkt_show(buf, card, card->info.blkt.time_total); | ||
876 | } | ||
877 | |||
878 | |||
879 | static ssize_t | ||
880 | qeth_dev_blkt_total_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
881 | { | ||
882 | struct qeth_card *card = dev->driver_data; | ||
883 | |||
884 | return qeth_dev_blkt_store(card, buf, count, | ||
885 | &card->info.blkt.time_total,1000); | ||
886 | } | ||
887 | |||
888 | |||
889 | |||
890 | static DEVICE_ATTR(total, 0644, qeth_dev_blkt_total_show, | ||
891 | qeth_dev_blkt_total_store); | ||
892 | |||
893 | static ssize_t | ||
894 | qeth_dev_blkt_inter_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
895 | { | ||
896 | struct qeth_card *card = dev->driver_data; | ||
897 | |||
898 | return qeth_dev_blkt_show(buf, card, card->info.blkt.inter_packet); | ||
899 | } | ||
900 | |||
901 | |||
902 | static ssize_t | ||
903 | qeth_dev_blkt_inter_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
904 | { | ||
905 | struct qeth_card *card = dev->driver_data; | ||
906 | |||
907 | return qeth_dev_blkt_store(card, buf, count, | ||
908 | &card->info.blkt.inter_packet,100); | ||
909 | } | ||
910 | |||
911 | static DEVICE_ATTR(inter, 0644, qeth_dev_blkt_inter_show, | ||
912 | qeth_dev_blkt_inter_store); | ||
913 | |||
914 | static ssize_t | ||
915 | qeth_dev_blkt_inter_jumbo_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
916 | { | ||
917 | struct qeth_card *card = dev->driver_data; | ||
918 | |||
919 | return qeth_dev_blkt_show(buf, card, | ||
920 | card->info.blkt.inter_packet_jumbo); | ||
921 | } | ||
922 | |||
923 | |||
924 | static ssize_t | ||
925 | qeth_dev_blkt_inter_jumbo_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
926 | { | ||
927 | struct qeth_card *card = dev->driver_data; | ||
928 | |||
929 | return qeth_dev_blkt_store(card, buf, count, | ||
930 | &card->info.blkt.inter_packet_jumbo,100); | ||
931 | } | ||
932 | |||
933 | static DEVICE_ATTR(inter_jumbo, 0644, qeth_dev_blkt_inter_jumbo_show, | ||
934 | qeth_dev_blkt_inter_jumbo_store); | ||
935 | |||
936 | static struct device_attribute * qeth_blkt_device_attrs[] = { | ||
937 | &dev_attr_total, | ||
938 | &dev_attr_inter, | ||
939 | &dev_attr_inter_jumbo, | ||
940 | NULL, | ||
941 | }; | ||
942 | |||
943 | static struct attribute_group qeth_device_blkt_group = { | ||
944 | .name = "blkt", | ||
945 | .attrs = (struct attribute **)qeth_blkt_device_attrs, | ||
946 | }; | ||
947 | |||
948 | static struct device_attribute * qeth_device_attrs[] = { | ||
949 | &dev_attr_state, | ||
950 | &dev_attr_chpid, | ||
951 | &dev_attr_if_name, | ||
952 | &dev_attr_card_type, | ||
953 | &dev_attr_portno, | ||
954 | &dev_attr_portname, | ||
955 | &dev_attr_checksumming, | ||
956 | &dev_attr_priority_queueing, | ||
957 | &dev_attr_buffer_count, | ||
958 | &dev_attr_route4, | ||
959 | #ifdef CONFIG_QETH_IPV6 | ||
960 | &dev_attr_route6, | ||
961 | #endif | ||
962 | &dev_attr_add_hhlen, | ||
963 | &dev_attr_fake_ll, | ||
964 | &dev_attr_fake_broadcast, | ||
965 | &dev_attr_recover, | ||
966 | &dev_attr_broadcast_mode, | ||
967 | &dev_attr_canonical_macaddr, | ||
968 | &dev_attr_layer2, | ||
969 | &dev_attr_large_send, | ||
970 | &dev_attr_performance_stats, | ||
971 | NULL, | ||
972 | }; | ||
973 | |||
974 | static struct attribute_group qeth_device_attr_group = { | ||
975 | .attrs = (struct attribute **)qeth_device_attrs, | ||
976 | }; | ||
977 | |||
978 | static struct device_attribute * qeth_osn_device_attrs[] = { | ||
979 | &dev_attr_state, | ||
980 | &dev_attr_chpid, | ||
981 | &dev_attr_if_name, | ||
982 | &dev_attr_card_type, | ||
983 | &dev_attr_buffer_count, | ||
984 | &dev_attr_recover, | ||
985 | NULL, | ||
986 | }; | ||
987 | |||
988 | static struct attribute_group qeth_osn_device_attr_group = { | ||
989 | .attrs = (struct attribute **)qeth_osn_device_attrs, | ||
990 | }; | ||
991 | |||
992 | #define QETH_DEVICE_ATTR(_id,_name,_mode,_show,_store) \ | ||
993 | struct device_attribute dev_attr_##_id = { \ | ||
994 | .attr = {.name=__stringify(_name), .mode=_mode, },\ | ||
995 | .show = _show, \ | ||
996 | .store = _store, \ | ||
997 | }; | ||
998 | |||
999 | static int | ||
1000 | qeth_check_layer2(struct qeth_card *card) | ||
1001 | { | ||
1002 | if (card->options.layer2) | ||
1003 | return -EPERM; | ||
1004 | return 0; | ||
1005 | } | ||
1006 | |||
1007 | |||
1008 | static ssize_t | ||
1009 | qeth_dev_ipato_enable_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
1010 | { | ||
1011 | struct qeth_card *card = dev->driver_data; | ||
1012 | |||
1013 | if (!card) | ||
1014 | return -EINVAL; | ||
1015 | |||
1016 | if (qeth_check_layer2(card)) | ||
1017 | return -EPERM; | ||
1018 | return sprintf(buf, "%i\n", card->ipato.enabled? 1:0); | ||
1019 | } | ||
1020 | |||
1021 | static ssize_t | ||
1022 | qeth_dev_ipato_enable_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
1023 | { | ||
1024 | struct qeth_card *card = dev->driver_data; | ||
1025 | char *tmp; | ||
1026 | |||
1027 | if (!card) | ||
1028 | return -EINVAL; | ||
1029 | |||
1030 | if ((card->state != CARD_STATE_DOWN) && | ||
1031 | (card->state != CARD_STATE_RECOVER)) | ||
1032 | return -EPERM; | ||
1033 | |||
1034 | if (qeth_check_layer2(card)) | ||
1035 | return -EPERM; | ||
1036 | |||
1037 | tmp = strsep((char **) &buf, "\n"); | ||
1038 | if (!strcmp(tmp, "toggle")){ | ||
1039 | card->ipato.enabled = (card->ipato.enabled)? 0 : 1; | ||
1040 | } else if (!strcmp(tmp, "1")){ | ||
1041 | card->ipato.enabled = 1; | ||
1042 | } else if (!strcmp(tmp, "0")){ | ||
1043 | card->ipato.enabled = 0; | ||
1044 | } else { | ||
1045 | PRINT_WARN("ipato_enable: write 0, 1 or 'toggle' to " | ||
1046 | "this file\n"); | ||
1047 | return -EINVAL; | ||
1048 | } | ||
1049 | return count; | ||
1050 | } | ||
1051 | |||
1052 | static QETH_DEVICE_ATTR(ipato_enable, enable, 0644, | ||
1053 | qeth_dev_ipato_enable_show, | ||
1054 | qeth_dev_ipato_enable_store); | ||
1055 | |||
1056 | static ssize_t | ||
1057 | qeth_dev_ipato_invert4_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
1058 | { | ||
1059 | struct qeth_card *card = dev->driver_data; | ||
1060 | |||
1061 | if (!card) | ||
1062 | return -EINVAL; | ||
1063 | |||
1064 | if (qeth_check_layer2(card)) | ||
1065 | return -EPERM; | ||
1066 | |||
1067 | return sprintf(buf, "%i\n", card->ipato.invert4? 1:0); | ||
1068 | } | ||
1069 | |||
1070 | static ssize_t | ||
1071 | qeth_dev_ipato_invert4_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
1072 | { | ||
1073 | struct qeth_card *card = dev->driver_data; | ||
1074 | char *tmp; | ||
1075 | |||
1076 | if (!card) | ||
1077 | return -EINVAL; | ||
1078 | |||
1079 | if (qeth_check_layer2(card)) | ||
1080 | return -EPERM; | ||
1081 | |||
1082 | tmp = strsep((char **) &buf, "\n"); | ||
1083 | if (!strcmp(tmp, "toggle")){ | ||
1084 | card->ipato.invert4 = (card->ipato.invert4)? 0 : 1; | ||
1085 | } else if (!strcmp(tmp, "1")){ | ||
1086 | card->ipato.invert4 = 1; | ||
1087 | } else if (!strcmp(tmp, "0")){ | ||
1088 | card->ipato.invert4 = 0; | ||
1089 | } else { | ||
1090 | PRINT_WARN("ipato_invert4: write 0, 1 or 'toggle' to " | ||
1091 | "this file\n"); | ||
1092 | return -EINVAL; | ||
1093 | } | ||
1094 | return count; | ||
1095 | } | ||
1096 | |||
1097 | static QETH_DEVICE_ATTR(ipato_invert4, invert4, 0644, | ||
1098 | qeth_dev_ipato_invert4_show, | ||
1099 | qeth_dev_ipato_invert4_store); | ||
1100 | |||
1101 | static ssize_t | ||
1102 | qeth_dev_ipato_add_show(char *buf, struct qeth_card *card, | ||
1103 | enum qeth_prot_versions proto) | ||
1104 | { | ||
1105 | struct qeth_ipato_entry *ipatoe; | ||
1106 | unsigned long flags; | ||
1107 | char addr_str[40]; | ||
1108 | int entry_len; /* length of 1 entry string, differs between v4 and v6 */ | ||
1109 | int i = 0; | ||
1110 | |||
1111 | if (qeth_check_layer2(card)) | ||
1112 | return -EPERM; | ||
1113 | |||
1114 | entry_len = (proto == QETH_PROT_IPV4)? 12 : 40; | ||
1115 | /* add strlen for "/<mask>\n" */ | ||
1116 | entry_len += (proto == QETH_PROT_IPV4)? 5 : 6; | ||
1117 | spin_lock_irqsave(&card->ip_lock, flags); | ||
1118 | list_for_each_entry(ipatoe, &card->ipato.entries, entry){ | ||
1119 | if (ipatoe->proto != proto) | ||
1120 | continue; | ||
1121 | /* String must not be longer than PAGE_SIZE. So we check if | ||
1122 | * string length gets near PAGE_SIZE. Then we can savely display | ||
1123 | * the next IPv6 address (worst case, compared to IPv4) */ | ||
1124 | if ((PAGE_SIZE - i) <= entry_len) | ||
1125 | break; | ||
1126 | qeth_ipaddr_to_string(proto, ipatoe->addr, addr_str); | ||
1127 | i += snprintf(buf + i, PAGE_SIZE - i, | ||
1128 | "%s/%i\n", addr_str, ipatoe->mask_bits); | ||
1129 | } | ||
1130 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
1131 | i += snprintf(buf + i, PAGE_SIZE - i, "\n"); | ||
1132 | |||
1133 | return i; | ||
1134 | } | ||
1135 | |||
1136 | static ssize_t | ||
1137 | qeth_dev_ipato_add4_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
1138 | { | ||
1139 | struct qeth_card *card = dev->driver_data; | ||
1140 | |||
1141 | if (!card) | ||
1142 | return -EINVAL; | ||
1143 | |||
1144 | return qeth_dev_ipato_add_show(buf, card, QETH_PROT_IPV4); | ||
1145 | } | ||
1146 | |||
1147 | static int | ||
1148 | qeth_parse_ipatoe(const char* buf, enum qeth_prot_versions proto, | ||
1149 | u8 *addr, int *mask_bits) | ||
1150 | { | ||
1151 | const char *start, *end; | ||
1152 | char *tmp; | ||
1153 | char buffer[40] = {0, }; | ||
1154 | |||
1155 | start = buf; | ||
1156 | /* get address string */ | ||
1157 | end = strchr(start, '/'); | ||
1158 | if (!end || (end - start >= 40)){ | ||
1159 | PRINT_WARN("Invalid format for ipato_addx/delx. " | ||
1160 | "Use <ip addr>/<mask bits>\n"); | ||
1161 | return -EINVAL; | ||
1162 | } | ||
1163 | strncpy(buffer, start, end - start); | ||
1164 | if (qeth_string_to_ipaddr(buffer, proto, addr)){ | ||
1165 | PRINT_WARN("Invalid IP address format!\n"); | ||
1166 | return -EINVAL; | ||
1167 | } | ||
1168 | start = end + 1; | ||
1169 | *mask_bits = simple_strtoul(start, &tmp, 10); | ||
1170 | if (!strlen(start) || | ||
1171 | (tmp == start) || | ||
1172 | (*mask_bits > ((proto == QETH_PROT_IPV4) ? 32 : 128))) { | ||
1173 | PRINT_WARN("Invalid mask bits for ipato_addx/delx !\n"); | ||
1174 | return -EINVAL; | ||
1175 | } | ||
1176 | return 0; | ||
1177 | } | ||
1178 | |||
1179 | static ssize_t | ||
1180 | qeth_dev_ipato_add_store(const char *buf, size_t count, | ||
1181 | struct qeth_card *card, enum qeth_prot_versions proto) | ||
1182 | { | ||
1183 | struct qeth_ipato_entry *ipatoe; | ||
1184 | u8 addr[16]; | ||
1185 | int mask_bits; | ||
1186 | int rc; | ||
1187 | |||
1188 | if (qeth_check_layer2(card)) | ||
1189 | return -EPERM; | ||
1190 | if ((rc = qeth_parse_ipatoe(buf, proto, addr, &mask_bits))) | ||
1191 | return rc; | ||
1192 | |||
1193 | if (!(ipatoe = kzalloc(sizeof(struct qeth_ipato_entry), GFP_KERNEL))){ | ||
1194 | PRINT_WARN("No memory to allocate ipato entry\n"); | ||
1195 | return -ENOMEM; | ||
1196 | } | ||
1197 | ipatoe->proto = proto; | ||
1198 | memcpy(ipatoe->addr, addr, (proto == QETH_PROT_IPV4)? 4:16); | ||
1199 | ipatoe->mask_bits = mask_bits; | ||
1200 | |||
1201 | if ((rc = qeth_add_ipato_entry(card, ipatoe))){ | ||
1202 | kfree(ipatoe); | ||
1203 | return rc; | ||
1204 | } | ||
1205 | |||
1206 | return count; | ||
1207 | } | ||
1208 | |||
1209 | static ssize_t | ||
1210 | qeth_dev_ipato_add4_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
1211 | { | ||
1212 | struct qeth_card *card = dev->driver_data; | ||
1213 | |||
1214 | if (!card) | ||
1215 | return -EINVAL; | ||
1216 | |||
1217 | return qeth_dev_ipato_add_store(buf, count, card, QETH_PROT_IPV4); | ||
1218 | } | ||
1219 | |||
1220 | static QETH_DEVICE_ATTR(ipato_add4, add4, 0644, | ||
1221 | qeth_dev_ipato_add4_show, | ||
1222 | qeth_dev_ipato_add4_store); | ||
1223 | |||
1224 | static ssize_t | ||
1225 | qeth_dev_ipato_del_store(const char *buf, size_t count, | ||
1226 | struct qeth_card *card, enum qeth_prot_versions proto) | ||
1227 | { | ||
1228 | u8 addr[16]; | ||
1229 | int mask_bits; | ||
1230 | int rc; | ||
1231 | |||
1232 | if (qeth_check_layer2(card)) | ||
1233 | return -EPERM; | ||
1234 | if ((rc = qeth_parse_ipatoe(buf, proto, addr, &mask_bits))) | ||
1235 | return rc; | ||
1236 | |||
1237 | qeth_del_ipato_entry(card, proto, addr, mask_bits); | ||
1238 | |||
1239 | return count; | ||
1240 | } | ||
1241 | |||
1242 | static ssize_t | ||
1243 | qeth_dev_ipato_del4_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
1244 | { | ||
1245 | struct qeth_card *card = dev->driver_data; | ||
1246 | |||
1247 | if (!card) | ||
1248 | return -EINVAL; | ||
1249 | |||
1250 | return qeth_dev_ipato_del_store(buf, count, card, QETH_PROT_IPV4); | ||
1251 | } | ||
1252 | |||
1253 | static QETH_DEVICE_ATTR(ipato_del4, del4, 0200, NULL, | ||
1254 | qeth_dev_ipato_del4_store); | ||
1255 | |||
1256 | #ifdef CONFIG_QETH_IPV6 | ||
1257 | static ssize_t | ||
1258 | qeth_dev_ipato_invert6_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
1259 | { | ||
1260 | struct qeth_card *card = dev->driver_data; | ||
1261 | |||
1262 | if (!card) | ||
1263 | return -EINVAL; | ||
1264 | |||
1265 | if (qeth_check_layer2(card)) | ||
1266 | return -EPERM; | ||
1267 | |||
1268 | return sprintf(buf, "%i\n", card->ipato.invert6? 1:0); | ||
1269 | } | ||
1270 | |||
1271 | static ssize_t | ||
1272 | qeth_dev_ipato_invert6_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
1273 | { | ||
1274 | struct qeth_card *card = dev->driver_data; | ||
1275 | char *tmp; | ||
1276 | |||
1277 | if (!card) | ||
1278 | return -EINVAL; | ||
1279 | |||
1280 | if (qeth_check_layer2(card)) | ||
1281 | return -EPERM; | ||
1282 | |||
1283 | tmp = strsep((char **) &buf, "\n"); | ||
1284 | if (!strcmp(tmp, "toggle")){ | ||
1285 | card->ipato.invert6 = (card->ipato.invert6)? 0 : 1; | ||
1286 | } else if (!strcmp(tmp, "1")){ | ||
1287 | card->ipato.invert6 = 1; | ||
1288 | } else if (!strcmp(tmp, "0")){ | ||
1289 | card->ipato.invert6 = 0; | ||
1290 | } else { | ||
1291 | PRINT_WARN("ipato_invert6: write 0, 1 or 'toggle' to " | ||
1292 | "this file\n"); | ||
1293 | return -EINVAL; | ||
1294 | } | ||
1295 | return count; | ||
1296 | } | ||
1297 | |||
1298 | static QETH_DEVICE_ATTR(ipato_invert6, invert6, 0644, | ||
1299 | qeth_dev_ipato_invert6_show, | ||
1300 | qeth_dev_ipato_invert6_store); | ||
1301 | |||
1302 | |||
1303 | static ssize_t | ||
1304 | qeth_dev_ipato_add6_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
1305 | { | ||
1306 | struct qeth_card *card = dev->driver_data; | ||
1307 | |||
1308 | if (!card) | ||
1309 | return -EINVAL; | ||
1310 | |||
1311 | return qeth_dev_ipato_add_show(buf, card, QETH_PROT_IPV6); | ||
1312 | } | ||
1313 | |||
1314 | static ssize_t | ||
1315 | qeth_dev_ipato_add6_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
1316 | { | ||
1317 | struct qeth_card *card = dev->driver_data; | ||
1318 | |||
1319 | if (!card) | ||
1320 | return -EINVAL; | ||
1321 | |||
1322 | return qeth_dev_ipato_add_store(buf, count, card, QETH_PROT_IPV6); | ||
1323 | } | ||
1324 | |||
1325 | static QETH_DEVICE_ATTR(ipato_add6, add6, 0644, | ||
1326 | qeth_dev_ipato_add6_show, | ||
1327 | qeth_dev_ipato_add6_store); | ||
1328 | |||
1329 | static ssize_t | ||
1330 | qeth_dev_ipato_del6_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
1331 | { | ||
1332 | struct qeth_card *card = dev->driver_data; | ||
1333 | |||
1334 | if (!card) | ||
1335 | return -EINVAL; | ||
1336 | |||
1337 | return qeth_dev_ipato_del_store(buf, count, card, QETH_PROT_IPV6); | ||
1338 | } | ||
1339 | |||
1340 | static QETH_DEVICE_ATTR(ipato_del6, del6, 0200, NULL, | ||
1341 | qeth_dev_ipato_del6_store); | ||
1342 | #endif /* CONFIG_QETH_IPV6 */ | ||
1343 | |||
1344 | static struct device_attribute * qeth_ipato_device_attrs[] = { | ||
1345 | &dev_attr_ipato_enable, | ||
1346 | &dev_attr_ipato_invert4, | ||
1347 | &dev_attr_ipato_add4, | ||
1348 | &dev_attr_ipato_del4, | ||
1349 | #ifdef CONFIG_QETH_IPV6 | ||
1350 | &dev_attr_ipato_invert6, | ||
1351 | &dev_attr_ipato_add6, | ||
1352 | &dev_attr_ipato_del6, | ||
1353 | #endif | ||
1354 | NULL, | ||
1355 | }; | ||
1356 | |||
1357 | static struct attribute_group qeth_device_ipato_group = { | ||
1358 | .name = "ipa_takeover", | ||
1359 | .attrs = (struct attribute **)qeth_ipato_device_attrs, | ||
1360 | }; | ||
1361 | |||
1362 | static ssize_t | ||
1363 | qeth_dev_vipa_add_show(char *buf, struct qeth_card *card, | ||
1364 | enum qeth_prot_versions proto) | ||
1365 | { | ||
1366 | struct qeth_ipaddr *ipaddr; | ||
1367 | char addr_str[40]; | ||
1368 | int entry_len; /* length of 1 entry string, differs between v4 and v6 */ | ||
1369 | unsigned long flags; | ||
1370 | int i = 0; | ||
1371 | |||
1372 | if (qeth_check_layer2(card)) | ||
1373 | return -EPERM; | ||
1374 | |||
1375 | entry_len = (proto == QETH_PROT_IPV4)? 12 : 40; | ||
1376 | entry_len += 2; /* \n + terminator */ | ||
1377 | spin_lock_irqsave(&card->ip_lock, flags); | ||
1378 | list_for_each_entry(ipaddr, &card->ip_list, entry){ | ||
1379 | if (ipaddr->proto != proto) | ||
1380 | continue; | ||
1381 | if (ipaddr->type != QETH_IP_TYPE_VIPA) | ||
1382 | continue; | ||
1383 | /* String must not be longer than PAGE_SIZE. So we check if | ||
1384 | * string length gets near PAGE_SIZE. Then we can savely display | ||
1385 | * the next IPv6 address (worst case, compared to IPv4) */ | ||
1386 | if ((PAGE_SIZE - i) <= entry_len) | ||
1387 | break; | ||
1388 | qeth_ipaddr_to_string(proto, (const u8 *)&ipaddr->u, addr_str); | ||
1389 | i += snprintf(buf + i, PAGE_SIZE - i, "%s\n", addr_str); | ||
1390 | } | ||
1391 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
1392 | i += snprintf(buf + i, PAGE_SIZE - i, "\n"); | ||
1393 | |||
1394 | return i; | ||
1395 | } | ||
1396 | |||
1397 | static ssize_t | ||
1398 | qeth_dev_vipa_add4_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
1399 | { | ||
1400 | struct qeth_card *card = dev->driver_data; | ||
1401 | |||
1402 | if (!card) | ||
1403 | return -EINVAL; | ||
1404 | |||
1405 | return qeth_dev_vipa_add_show(buf, card, QETH_PROT_IPV4); | ||
1406 | } | ||
1407 | |||
1408 | static int | ||
1409 | qeth_parse_vipae(const char* buf, enum qeth_prot_versions proto, | ||
1410 | u8 *addr) | ||
1411 | { | ||
1412 | if (qeth_string_to_ipaddr(buf, proto, addr)){ | ||
1413 | PRINT_WARN("Invalid IP address format!\n"); | ||
1414 | return -EINVAL; | ||
1415 | } | ||
1416 | return 0; | ||
1417 | } | ||
1418 | |||
1419 | static ssize_t | ||
1420 | qeth_dev_vipa_add_store(const char *buf, size_t count, | ||
1421 | struct qeth_card *card, enum qeth_prot_versions proto) | ||
1422 | { | ||
1423 | u8 addr[16] = {0, }; | ||
1424 | int rc; | ||
1425 | |||
1426 | if (qeth_check_layer2(card)) | ||
1427 | return -EPERM; | ||
1428 | if ((rc = qeth_parse_vipae(buf, proto, addr))) | ||
1429 | return rc; | ||
1430 | |||
1431 | if ((rc = qeth_add_vipa(card, proto, addr))) | ||
1432 | return rc; | ||
1433 | |||
1434 | return count; | ||
1435 | } | ||
1436 | |||
1437 | static ssize_t | ||
1438 | qeth_dev_vipa_add4_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
1439 | { | ||
1440 | struct qeth_card *card = dev->driver_data; | ||
1441 | |||
1442 | if (!card) | ||
1443 | return -EINVAL; | ||
1444 | |||
1445 | return qeth_dev_vipa_add_store(buf, count, card, QETH_PROT_IPV4); | ||
1446 | } | ||
1447 | |||
1448 | static QETH_DEVICE_ATTR(vipa_add4, add4, 0644, | ||
1449 | qeth_dev_vipa_add4_show, | ||
1450 | qeth_dev_vipa_add4_store); | ||
1451 | |||
1452 | static ssize_t | ||
1453 | qeth_dev_vipa_del_store(const char *buf, size_t count, | ||
1454 | struct qeth_card *card, enum qeth_prot_versions proto) | ||
1455 | { | ||
1456 | u8 addr[16]; | ||
1457 | int rc; | ||
1458 | |||
1459 | if (qeth_check_layer2(card)) | ||
1460 | return -EPERM; | ||
1461 | if ((rc = qeth_parse_vipae(buf, proto, addr))) | ||
1462 | return rc; | ||
1463 | |||
1464 | qeth_del_vipa(card, proto, addr); | ||
1465 | |||
1466 | return count; | ||
1467 | } | ||
1468 | |||
1469 | static ssize_t | ||
1470 | qeth_dev_vipa_del4_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
1471 | { | ||
1472 | struct qeth_card *card = dev->driver_data; | ||
1473 | |||
1474 | if (!card) | ||
1475 | return -EINVAL; | ||
1476 | |||
1477 | return qeth_dev_vipa_del_store(buf, count, card, QETH_PROT_IPV4); | ||
1478 | } | ||
1479 | |||
1480 | static QETH_DEVICE_ATTR(vipa_del4, del4, 0200, NULL, | ||
1481 | qeth_dev_vipa_del4_store); | ||
1482 | |||
1483 | #ifdef CONFIG_QETH_IPV6 | ||
1484 | static ssize_t | ||
1485 | qeth_dev_vipa_add6_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
1486 | { | ||
1487 | struct qeth_card *card = dev->driver_data; | ||
1488 | |||
1489 | if (!card) | ||
1490 | return -EINVAL; | ||
1491 | |||
1492 | return qeth_dev_vipa_add_show(buf, card, QETH_PROT_IPV6); | ||
1493 | } | ||
1494 | |||
1495 | static ssize_t | ||
1496 | qeth_dev_vipa_add6_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
1497 | { | ||
1498 | struct qeth_card *card = dev->driver_data; | ||
1499 | |||
1500 | if (!card) | ||
1501 | return -EINVAL; | ||
1502 | |||
1503 | return qeth_dev_vipa_add_store(buf, count, card, QETH_PROT_IPV6); | ||
1504 | } | ||
1505 | |||
1506 | static QETH_DEVICE_ATTR(vipa_add6, add6, 0644, | ||
1507 | qeth_dev_vipa_add6_show, | ||
1508 | qeth_dev_vipa_add6_store); | ||
1509 | |||
1510 | static ssize_t | ||
1511 | qeth_dev_vipa_del6_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
1512 | { | ||
1513 | struct qeth_card *card = dev->driver_data; | ||
1514 | |||
1515 | if (!card) | ||
1516 | return -EINVAL; | ||
1517 | |||
1518 | if (qeth_check_layer2(card)) | ||
1519 | return -EPERM; | ||
1520 | |||
1521 | return qeth_dev_vipa_del_store(buf, count, card, QETH_PROT_IPV6); | ||
1522 | } | ||
1523 | |||
1524 | static QETH_DEVICE_ATTR(vipa_del6, del6, 0200, NULL, | ||
1525 | qeth_dev_vipa_del6_store); | ||
1526 | #endif /* CONFIG_QETH_IPV6 */ | ||
1527 | |||
1528 | static struct device_attribute * qeth_vipa_device_attrs[] = { | ||
1529 | &dev_attr_vipa_add4, | ||
1530 | &dev_attr_vipa_del4, | ||
1531 | #ifdef CONFIG_QETH_IPV6 | ||
1532 | &dev_attr_vipa_add6, | ||
1533 | &dev_attr_vipa_del6, | ||
1534 | #endif | ||
1535 | NULL, | ||
1536 | }; | ||
1537 | |||
1538 | static struct attribute_group qeth_device_vipa_group = { | ||
1539 | .name = "vipa", | ||
1540 | .attrs = (struct attribute **)qeth_vipa_device_attrs, | ||
1541 | }; | ||
1542 | |||
1543 | static ssize_t | ||
1544 | qeth_dev_rxip_add_show(char *buf, struct qeth_card *card, | ||
1545 | enum qeth_prot_versions proto) | ||
1546 | { | ||
1547 | struct qeth_ipaddr *ipaddr; | ||
1548 | char addr_str[40]; | ||
1549 | int entry_len; /* length of 1 entry string, differs between v4 and v6 */ | ||
1550 | unsigned long flags; | ||
1551 | int i = 0; | ||
1552 | |||
1553 | if (qeth_check_layer2(card)) | ||
1554 | return -EPERM; | ||
1555 | |||
1556 | entry_len = (proto == QETH_PROT_IPV4)? 12 : 40; | ||
1557 | entry_len += 2; /* \n + terminator */ | ||
1558 | spin_lock_irqsave(&card->ip_lock, flags); | ||
1559 | list_for_each_entry(ipaddr, &card->ip_list, entry){ | ||
1560 | if (ipaddr->proto != proto) | ||
1561 | continue; | ||
1562 | if (ipaddr->type != QETH_IP_TYPE_RXIP) | ||
1563 | continue; | ||
1564 | /* String must not be longer than PAGE_SIZE. So we check if | ||
1565 | * string length gets near PAGE_SIZE. Then we can savely display | ||
1566 | * the next IPv6 address (worst case, compared to IPv4) */ | ||
1567 | if ((PAGE_SIZE - i) <= entry_len) | ||
1568 | break; | ||
1569 | qeth_ipaddr_to_string(proto, (const u8 *)&ipaddr->u, addr_str); | ||
1570 | i += snprintf(buf + i, PAGE_SIZE - i, "%s\n", addr_str); | ||
1571 | } | ||
1572 | spin_unlock_irqrestore(&card->ip_lock, flags); | ||
1573 | i += snprintf(buf + i, PAGE_SIZE - i, "\n"); | ||
1574 | |||
1575 | return i; | ||
1576 | } | ||
1577 | |||
1578 | static ssize_t | ||
1579 | qeth_dev_rxip_add4_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
1580 | { | ||
1581 | struct qeth_card *card = dev->driver_data; | ||
1582 | |||
1583 | if (!card) | ||
1584 | return -EINVAL; | ||
1585 | |||
1586 | return qeth_dev_rxip_add_show(buf, card, QETH_PROT_IPV4); | ||
1587 | } | ||
1588 | |||
1589 | static int | ||
1590 | qeth_parse_rxipe(const char* buf, enum qeth_prot_versions proto, | ||
1591 | u8 *addr) | ||
1592 | { | ||
1593 | if (qeth_string_to_ipaddr(buf, proto, addr)){ | ||
1594 | PRINT_WARN("Invalid IP address format!\n"); | ||
1595 | return -EINVAL; | ||
1596 | } | ||
1597 | return 0; | ||
1598 | } | ||
1599 | |||
1600 | static ssize_t | ||
1601 | qeth_dev_rxip_add_store(const char *buf, size_t count, | ||
1602 | struct qeth_card *card, enum qeth_prot_versions proto) | ||
1603 | { | ||
1604 | u8 addr[16] = {0, }; | ||
1605 | int rc; | ||
1606 | |||
1607 | if (qeth_check_layer2(card)) | ||
1608 | return -EPERM; | ||
1609 | if ((rc = qeth_parse_rxipe(buf, proto, addr))) | ||
1610 | return rc; | ||
1611 | |||
1612 | if ((rc = qeth_add_rxip(card, proto, addr))) | ||
1613 | return rc; | ||
1614 | |||
1615 | return count; | ||
1616 | } | ||
1617 | |||
1618 | static ssize_t | ||
1619 | qeth_dev_rxip_add4_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
1620 | { | ||
1621 | struct qeth_card *card = dev->driver_data; | ||
1622 | |||
1623 | if (!card) | ||
1624 | return -EINVAL; | ||
1625 | |||
1626 | return qeth_dev_rxip_add_store(buf, count, card, QETH_PROT_IPV4); | ||
1627 | } | ||
1628 | |||
1629 | static QETH_DEVICE_ATTR(rxip_add4, add4, 0644, | ||
1630 | qeth_dev_rxip_add4_show, | ||
1631 | qeth_dev_rxip_add4_store); | ||
1632 | |||
1633 | static ssize_t | ||
1634 | qeth_dev_rxip_del_store(const char *buf, size_t count, | ||
1635 | struct qeth_card *card, enum qeth_prot_versions proto) | ||
1636 | { | ||
1637 | u8 addr[16]; | ||
1638 | int rc; | ||
1639 | |||
1640 | if (qeth_check_layer2(card)) | ||
1641 | return -EPERM; | ||
1642 | if ((rc = qeth_parse_rxipe(buf, proto, addr))) | ||
1643 | return rc; | ||
1644 | |||
1645 | qeth_del_rxip(card, proto, addr); | ||
1646 | |||
1647 | return count; | ||
1648 | } | ||
1649 | |||
1650 | static ssize_t | ||
1651 | qeth_dev_rxip_del4_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
1652 | { | ||
1653 | struct qeth_card *card = dev->driver_data; | ||
1654 | |||
1655 | if (!card) | ||
1656 | return -EINVAL; | ||
1657 | |||
1658 | return qeth_dev_rxip_del_store(buf, count, card, QETH_PROT_IPV4); | ||
1659 | } | ||
1660 | |||
1661 | static QETH_DEVICE_ATTR(rxip_del4, del4, 0200, NULL, | ||
1662 | qeth_dev_rxip_del4_store); | ||
1663 | |||
1664 | #ifdef CONFIG_QETH_IPV6 | ||
1665 | static ssize_t | ||
1666 | qeth_dev_rxip_add6_show(struct device *dev, struct device_attribute *attr, char *buf) | ||
1667 | { | ||
1668 | struct qeth_card *card = dev->driver_data; | ||
1669 | |||
1670 | if (!card) | ||
1671 | return -EINVAL; | ||
1672 | |||
1673 | return qeth_dev_rxip_add_show(buf, card, QETH_PROT_IPV6); | ||
1674 | } | ||
1675 | |||
1676 | static ssize_t | ||
1677 | qeth_dev_rxip_add6_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
1678 | { | ||
1679 | struct qeth_card *card = dev->driver_data; | ||
1680 | |||
1681 | if (!card) | ||
1682 | return -EINVAL; | ||
1683 | |||
1684 | return qeth_dev_rxip_add_store(buf, count, card, QETH_PROT_IPV6); | ||
1685 | } | ||
1686 | |||
1687 | static QETH_DEVICE_ATTR(rxip_add6, add6, 0644, | ||
1688 | qeth_dev_rxip_add6_show, | ||
1689 | qeth_dev_rxip_add6_store); | ||
1690 | |||
1691 | static ssize_t | ||
1692 | qeth_dev_rxip_del6_store(struct device *dev, struct device_attribute *attr, const char *buf, size_t count) | ||
1693 | { | ||
1694 | struct qeth_card *card = dev->driver_data; | ||
1695 | |||
1696 | if (!card) | ||
1697 | return -EINVAL; | ||
1698 | |||
1699 | return qeth_dev_rxip_del_store(buf, count, card, QETH_PROT_IPV6); | ||
1700 | } | ||
1701 | |||
1702 | static QETH_DEVICE_ATTR(rxip_del6, del6, 0200, NULL, | ||
1703 | qeth_dev_rxip_del6_store); | ||
1704 | #endif /* CONFIG_QETH_IPV6 */ | ||
1705 | |||
1706 | static struct device_attribute * qeth_rxip_device_attrs[] = { | ||
1707 | &dev_attr_rxip_add4, | ||
1708 | &dev_attr_rxip_del4, | ||
1709 | #ifdef CONFIG_QETH_IPV6 | ||
1710 | &dev_attr_rxip_add6, | ||
1711 | &dev_attr_rxip_del6, | ||
1712 | #endif | ||
1713 | NULL, | ||
1714 | }; | ||
1715 | |||
1716 | static struct attribute_group qeth_device_rxip_group = { | ||
1717 | .name = "rxip", | ||
1718 | .attrs = (struct attribute **)qeth_rxip_device_attrs, | ||
1719 | }; | ||
1720 | |||
1721 | int | ||
1722 | qeth_create_device_attributes(struct device *dev) | ||
1723 | { | ||
1724 | int ret; | ||
1725 | struct qeth_card *card = dev->driver_data; | ||
1726 | |||
1727 | if (card->info.type == QETH_CARD_TYPE_OSN) | ||
1728 | return sysfs_create_group(&dev->kobj, | ||
1729 | &qeth_osn_device_attr_group); | ||
1730 | |||
1731 | if ((ret = sysfs_create_group(&dev->kobj, &qeth_device_attr_group))) | ||
1732 | return ret; | ||
1733 | if ((ret = sysfs_create_group(&dev->kobj, &qeth_device_ipato_group))){ | ||
1734 | sysfs_remove_group(&dev->kobj, &qeth_device_attr_group); | ||
1735 | return ret; | ||
1736 | } | ||
1737 | if ((ret = sysfs_create_group(&dev->kobj, &qeth_device_vipa_group))){ | ||
1738 | sysfs_remove_group(&dev->kobj, &qeth_device_attr_group); | ||
1739 | sysfs_remove_group(&dev->kobj, &qeth_device_ipato_group); | ||
1740 | return ret; | ||
1741 | } | ||
1742 | if ((ret = sysfs_create_group(&dev->kobj, &qeth_device_rxip_group))){ | ||
1743 | sysfs_remove_group(&dev->kobj, &qeth_device_attr_group); | ||
1744 | sysfs_remove_group(&dev->kobj, &qeth_device_ipato_group); | ||
1745 | sysfs_remove_group(&dev->kobj, &qeth_device_vipa_group); | ||
1746 | return ret; | ||
1747 | } | ||
1748 | if ((ret = sysfs_create_group(&dev->kobj, &qeth_device_blkt_group))){ | ||
1749 | sysfs_remove_group(&dev->kobj, &qeth_device_attr_group); | ||
1750 | sysfs_remove_group(&dev->kobj, &qeth_device_ipato_group); | ||
1751 | sysfs_remove_group(&dev->kobj, &qeth_device_vipa_group); | ||
1752 | sysfs_remove_group(&dev->kobj, &qeth_device_rxip_group); | ||
1753 | return ret; | ||
1754 | } | ||
1755 | return 0; | ||
1756 | } | ||
1757 | |||
1758 | void | ||
1759 | qeth_remove_device_attributes(struct device *dev) | ||
1760 | { | ||
1761 | struct qeth_card *card = dev->driver_data; | ||
1762 | |||
1763 | if (card->info.type == QETH_CARD_TYPE_OSN) { | ||
1764 | sysfs_remove_group(&dev->kobj, &qeth_osn_device_attr_group); | ||
1765 | return; | ||
1766 | } | ||
1767 | sysfs_remove_group(&dev->kobj, &qeth_device_attr_group); | ||
1768 | sysfs_remove_group(&dev->kobj, &qeth_device_ipato_group); | ||
1769 | sysfs_remove_group(&dev->kobj, &qeth_device_vipa_group); | ||
1770 | sysfs_remove_group(&dev->kobj, &qeth_device_rxip_group); | ||
1771 | sysfs_remove_group(&dev->kobj, &qeth_device_blkt_group); | ||
1772 | } | ||
1773 | |||
1774 | /**********************/ | ||
1775 | /* DRIVER ATTRIBUTES */ | ||
1776 | /**********************/ | ||
1777 | static ssize_t | ||
1778 | qeth_driver_group_store(struct device_driver *ddrv, const char *buf, | ||
1779 | size_t count) | ||
1780 | { | ||
1781 | const char *start, *end; | ||
1782 | char bus_ids[3][BUS_ID_SIZE], *argv[3]; | ||
1783 | int i; | ||
1784 | int err; | ||
1785 | |||
1786 | start = buf; | ||
1787 | for (i = 0; i < 3; i++) { | ||
1788 | static const char delim[] = { ',', ',', '\n' }; | ||
1789 | int len; | ||
1790 | |||
1791 | if (!(end = strchr(start, delim[i]))) | ||
1792 | return -EINVAL; | ||
1793 | len = min_t(ptrdiff_t, BUS_ID_SIZE, end - start); | ||
1794 | strncpy(bus_ids[i], start, len); | ||
1795 | bus_ids[i][len] = '\0'; | ||
1796 | start = end + 1; | ||
1797 | argv[i] = bus_ids[i]; | ||
1798 | } | ||
1799 | err = ccwgroup_create(qeth_root_dev, qeth_ccwgroup_driver.driver_id, | ||
1800 | &qeth_ccw_driver, 3, argv); | ||
1801 | if (err) | ||
1802 | return err; | ||
1803 | else | ||
1804 | return count; | ||
1805 | } | ||
1806 | |||
1807 | |||
1808 | static DRIVER_ATTR(group, 0200, NULL, qeth_driver_group_store); | ||
1809 | |||
1810 | static ssize_t | ||
1811 | qeth_driver_notifier_register_store(struct device_driver *ddrv, const char *buf, | ||
1812 | size_t count) | ||
1813 | { | ||
1814 | int rc; | ||
1815 | int signum; | ||
1816 | char *tmp, *tmp2; | ||
1817 | |||
1818 | tmp = strsep((char **) &buf, "\n"); | ||
1819 | if (!strncmp(tmp, "unregister", 10)){ | ||
1820 | if ((rc = qeth_notifier_unregister(current))) | ||
1821 | return rc; | ||
1822 | return count; | ||
1823 | } | ||
1824 | |||
1825 | signum = simple_strtoul(tmp, &tmp2, 10); | ||
1826 | if ((signum < 0) || (signum > 32)){ | ||
1827 | PRINT_WARN("Signal number %d is out of range\n", signum); | ||
1828 | return -EINVAL; | ||
1829 | } | ||
1830 | if ((rc = qeth_notifier_register(current, signum))) | ||
1831 | return rc; | ||
1832 | |||
1833 | return count; | ||
1834 | } | ||
1835 | |||
1836 | static DRIVER_ATTR(notifier_register, 0200, NULL, | ||
1837 | qeth_driver_notifier_register_store); | ||
1838 | |||
1839 | int | ||
1840 | qeth_create_driver_attributes(void) | ||
1841 | { | ||
1842 | int rc; | ||
1843 | |||
1844 | if ((rc = driver_create_file(&qeth_ccwgroup_driver.driver, | ||
1845 | &driver_attr_group))) | ||
1846 | return rc; | ||
1847 | return driver_create_file(&qeth_ccwgroup_driver.driver, | ||
1848 | &driver_attr_notifier_register); | ||
1849 | } | ||
1850 | |||
1851 | void | ||
1852 | qeth_remove_driver_attributes(void) | ||
1853 | { | ||
1854 | driver_remove_file(&qeth_ccwgroup_driver.driver, | ||
1855 | &driver_attr_group); | ||
1856 | driver_remove_file(&qeth_ccwgroup_driver.driver, | ||
1857 | &driver_attr_notifier_register); | ||
1858 | } | ||
diff --git a/drivers/s390/net/qeth_tso.h b/drivers/s390/net/qeth_tso.h deleted file mode 100644 index c20e923cf9ad..000000000000 --- a/drivers/s390/net/qeth_tso.h +++ /dev/null | |||
@@ -1,148 +0,0 @@ | |||
1 | /* | ||
2 | * linux/drivers/s390/net/qeth_tso.h | ||
3 | * | ||
4 | * Header file for qeth TCP Segmentation Offload support. | ||
5 | * | ||
6 | * Copyright 2004 IBM Corporation | ||
7 | * | ||
8 | * Author(s): Frank Pavlic <fpavlic@de.ibm.com> | ||
9 | * | ||
10 | */ | ||
11 | #ifndef __QETH_TSO_H__ | ||
12 | #define __QETH_TSO_H__ | ||
13 | |||
14 | #include <linux/skbuff.h> | ||
15 | #include <linux/tcp.h> | ||
16 | #include <linux/ip.h> | ||
17 | #include <linux/ipv6.h> | ||
18 | #include <net/ip6_checksum.h> | ||
19 | #include "qeth.h" | ||
20 | #include "qeth_mpc.h" | ||
21 | |||
22 | |||
23 | static inline struct qeth_hdr_tso * | ||
24 | qeth_tso_prepare_skb(struct qeth_card *card, struct sk_buff **skb) | ||
25 | { | ||
26 | QETH_DBF_TEXT(trace, 5, "tsoprsk"); | ||
27 | return qeth_push_skb(card, *skb, sizeof(struct qeth_hdr_tso)); | ||
28 | } | ||
29 | |||
30 | /** | ||
31 | * fill header for a TSO packet | ||
32 | */ | ||
33 | static inline void | ||
34 | qeth_tso_fill_header(struct qeth_card *card, struct sk_buff *skb) | ||
35 | { | ||
36 | struct qeth_hdr_tso *hdr; | ||
37 | struct tcphdr *tcph; | ||
38 | struct iphdr *iph; | ||
39 | |||
40 | QETH_DBF_TEXT(trace, 5, "tsofhdr"); | ||
41 | |||
42 | hdr = (struct qeth_hdr_tso *) skb->data; | ||
43 | iph = ip_hdr(skb); | ||
44 | tcph = tcp_hdr(skb); | ||
45 | /*fix header to TSO values ...*/ | ||
46 | hdr->hdr.hdr.l3.id = QETH_HEADER_TYPE_TSO; | ||
47 | /*set values which are fix for the first approach ...*/ | ||
48 | hdr->ext.hdr_tot_len = (__u16) sizeof(struct qeth_hdr_ext_tso); | ||
49 | hdr->ext.imb_hdr_no = 1; | ||
50 | hdr->ext.hdr_type = 1; | ||
51 | hdr->ext.hdr_version = 1; | ||
52 | hdr->ext.hdr_len = 28; | ||
53 | /*insert non-fix values */ | ||
54 | hdr->ext.mss = skb_shinfo(skb)->gso_size; | ||
55 | hdr->ext.dg_hdr_len = (__u16)(iph->ihl*4 + tcph->doff*4); | ||
56 | hdr->ext.payload_len = (__u16)(skb->len - hdr->ext.dg_hdr_len - | ||
57 | sizeof(struct qeth_hdr_tso)); | ||
58 | } | ||
59 | |||
60 | /** | ||
61 | * change some header values as requested by hardware | ||
62 | */ | ||
63 | static inline void | ||
64 | qeth_tso_set_tcpip_header(struct qeth_card *card, struct sk_buff *skb) | ||
65 | { | ||
66 | struct iphdr *iph = ip_hdr(skb); | ||
67 | struct ipv6hdr *ip6h = ipv6_hdr(skb); | ||
68 | struct tcphdr *tcph = tcp_hdr(skb); | ||
69 | |||
70 | tcph->check = 0; | ||
71 | if (skb->protocol == ETH_P_IPV6) { | ||
72 | ip6h->payload_len = 0; | ||
73 | tcph->check = ~csum_ipv6_magic(&ip6h->saddr, &ip6h->daddr, | ||
74 | 0, IPPROTO_TCP, 0); | ||
75 | return; | ||
76 | } | ||
77 | /*OSA want us to set these values ...*/ | ||
78 | tcph->check = ~csum_tcpudp_magic(iph->saddr, iph->daddr, | ||
79 | 0, IPPROTO_TCP, 0); | ||
80 | iph->tot_len = 0; | ||
81 | iph->check = 0; | ||
82 | } | ||
83 | |||
84 | static inline int | ||
85 | qeth_tso_prepare_packet(struct qeth_card *card, struct sk_buff *skb, | ||
86 | int ipv, int cast_type) | ||
87 | { | ||
88 | struct qeth_hdr_tso *hdr; | ||
89 | |||
90 | QETH_DBF_TEXT(trace, 5, "tsoprep"); | ||
91 | |||
92 | hdr = (struct qeth_hdr_tso *) qeth_tso_prepare_skb(card, &skb); | ||
93 | if (hdr == NULL) { | ||
94 | QETH_DBF_TEXT(trace, 4, "tsoperr"); | ||
95 | return -ENOMEM; | ||
96 | } | ||
97 | memset(hdr, 0, sizeof(struct qeth_hdr_tso)); | ||
98 | /*fill first 32 bytes of qdio header as used | ||
99 | *FIXME: TSO has two struct members | ||
100 | * with different names but same size | ||
101 | * */ | ||
102 | qeth_fill_header(card, &hdr->hdr, skb, ipv, cast_type); | ||
103 | qeth_tso_fill_header(card, skb); | ||
104 | qeth_tso_set_tcpip_header(card, skb); | ||
105 | return 0; | ||
106 | } | ||
107 | |||
108 | static inline void | ||
109 | __qeth_fill_buffer_frag(struct sk_buff *skb, struct qdio_buffer *buffer, | ||
110 | int is_tso, int *next_element_to_fill) | ||
111 | { | ||
112 | struct skb_frag_struct *frag; | ||
113 | int fragno; | ||
114 | unsigned long addr; | ||
115 | int element, cnt, dlen; | ||
116 | |||
117 | fragno = skb_shinfo(skb)->nr_frags; | ||
118 | element = *next_element_to_fill; | ||
119 | dlen = 0; | ||
120 | |||
121 | if (is_tso) | ||
122 | buffer->element[element].flags = | ||
123 | SBAL_FLAGS_MIDDLE_FRAG; | ||
124 | else | ||
125 | buffer->element[element].flags = | ||
126 | SBAL_FLAGS_FIRST_FRAG; | ||
127 | if ( (dlen = (skb->len - skb->data_len)) ) { | ||
128 | buffer->element[element].addr = skb->data; | ||
129 | buffer->element[element].length = dlen; | ||
130 | element++; | ||
131 | } | ||
132 | for (cnt = 0; cnt < fragno; cnt++) { | ||
133 | frag = &skb_shinfo(skb)->frags[cnt]; | ||
134 | addr = (page_to_pfn(frag->page) << PAGE_SHIFT) + | ||
135 | frag->page_offset; | ||
136 | buffer->element[element].addr = (char *)addr; | ||
137 | buffer->element[element].length = frag->size; | ||
138 | if (cnt < (fragno - 1)) | ||
139 | buffer->element[element].flags = | ||
140 | SBAL_FLAGS_MIDDLE_FRAG; | ||
141 | else | ||
142 | buffer->element[element].flags = | ||
143 | SBAL_FLAGS_LAST_FRAG; | ||
144 | element++; | ||
145 | } | ||
146 | *next_element_to_fill = element; | ||
147 | } | ||
148 | #endif /* __QETH_TSO_H__ */ | ||