diff options
author | Johannes Berg <johannes@sipsolutions.net> | 2009-06-02 07:01:37 -0400 |
---|---|---|
committer | John W. Linville <linville@tuxdriver.com> | 2009-06-03 14:06:13 -0400 |
commit | 19d337dff95cbf76edd3ad95c0cee2732c3e1ec5 (patch) | |
tree | 33326eeb09cb9664cc8427a5dc7cd2b08b5a57c3 /net | |
parent | 0f6399c4c525b518644a9b09f8d6fb125a418c4d (diff) |
rfkill: rewrite
This patch completely rewrites the rfkill core to address
the following deficiencies:
* all rfkill drivers need to implement polling where necessary
rather than having one central implementation
* updating the rfkill state cannot be done from arbitrary
contexts, forcing drivers to use schedule_work and requiring
lots of code
* rfkill drivers need to keep track of soft/hard blocked
internally -- the core should do this
* the rfkill API has many unexpected quirks, for example being
asymmetric wrt. alloc/free and register/unregister
* rfkill can call back into a driver from within a function the
driver called -- this is prone to deadlocks and generally
should be avoided
* rfkill-input pointlessly is a separate module
* drivers need to #ifdef rfkill functions (unless they want to
depend on or select RFKILL) -- rfkill should provide inlines
that do nothing if it isn't compiled in
* the rfkill structure is not opaque -- drivers need to initialise
it correctly (lots of sanity checking code required) -- instead
force drivers to pass the right variables to rfkill_alloc()
* the documentation is hard to read because it always assumes the
reader is completely clueless and contains way TOO MANY CAPS
* the rfkill code needlessly uses a lot of locks and atomic
operations in locked sections
* fix LED trigger to actually change the LED when the radio state
changes -- this wasn't done before
Tested-by: Alan Jenkins <alan-jenkins@tuffmail.co.uk>
Signed-off-by: Henrique de Moraes Holschuh <hmh@hmh.eng.br> [thinkpad]
Signed-off-by: Johannes Berg <johannes@sipsolutions.net>
Signed-off-by: John W. Linville <linville@tuxdriver.com>
Diffstat (limited to 'net')
-rw-r--r-- | net/rfkill/Kconfig | 21 | ||||
-rw-r--r-- | net/rfkill/Makefile | 5 | ||||
-rw-r--r-- | net/rfkill/core.c | 896 | ||||
-rw-r--r-- | net/rfkill/input.c | 342 | ||||
-rw-r--r-- | net/rfkill/rfkill-input.c | 390 | ||||
-rw-r--r-- | net/rfkill/rfkill.c | 855 | ||||
-rw-r--r-- | net/rfkill/rfkill.h (renamed from net/rfkill/rfkill-input.h) | 10 | ||||
-rw-r--r-- | net/wimax/Kconfig | 14 | ||||
-rw-r--r-- | net/wimax/op-rfkill.c | 123 |
9 files changed, 1282 insertions, 1374 deletions
diff --git a/net/rfkill/Kconfig b/net/rfkill/Kconfig index 7f807b30cfbb..b47f72fae05d 100644 --- a/net/rfkill/Kconfig +++ b/net/rfkill/Kconfig | |||
@@ -10,22 +10,15 @@ menuconfig RFKILL | |||
10 | To compile this driver as a module, choose M here: the | 10 | To compile this driver as a module, choose M here: the |
11 | module will be called rfkill. | 11 | module will be called rfkill. |
12 | 12 | ||
13 | config RFKILL_INPUT | ||
14 | tristate "Input layer to RF switch connector" | ||
15 | depends on RFKILL && INPUT | ||
16 | help | ||
17 | Say Y here if you want kernel automatically toggle state | ||
18 | of RF switches on and off when user presses appropriate | ||
19 | button or a key on the keyboard. Without this module you | ||
20 | need a some kind of userspace application to control | ||
21 | state of the switches. | ||
22 | |||
23 | To compile this driver as a module, choose M here: the | ||
24 | module will be called rfkill-input. | ||
25 | |||
26 | # LED trigger support | 13 | # LED trigger support |
27 | config RFKILL_LEDS | 14 | config RFKILL_LEDS |
28 | bool | 15 | bool |
29 | depends on RFKILL && LEDS_TRIGGERS | 16 | depends on RFKILL |
17 | depends on LEDS_TRIGGERS = y || RFKILL = LEDS_TRIGGERS | ||
30 | default y | 18 | default y |
31 | 19 | ||
20 | config RFKILL_INPUT | ||
21 | bool | ||
22 | depends on RFKILL | ||
23 | depends on INPUT = y || RFKILL = INPUT | ||
24 | default y | ||
diff --git a/net/rfkill/Makefile b/net/rfkill/Makefile index b38c430be057..662105352691 100644 --- a/net/rfkill/Makefile +++ b/net/rfkill/Makefile | |||
@@ -2,5 +2,6 @@ | |||
2 | # Makefile for the RF switch subsystem. | 2 | # Makefile for the RF switch subsystem. |
3 | # | 3 | # |
4 | 4 | ||
5 | obj-$(CONFIG_RFKILL) += rfkill.o | 5 | rfkill-y += core.o |
6 | obj-$(CONFIG_RFKILL_INPUT) += rfkill-input.o | 6 | rfkill-$(CONFIG_RFKILL_INPUT) += input.o |
7 | obj-$(CONFIG_RFKILL) += rfkill.o | ||
diff --git a/net/rfkill/core.c b/net/rfkill/core.c new file mode 100644 index 000000000000..30a6f8d819b2 --- /dev/null +++ b/net/rfkill/core.c | |||
@@ -0,0 +1,896 @@ | |||
1 | /* | ||
2 | * Copyright (C) 2006 - 2007 Ivo van Doorn | ||
3 | * Copyright (C) 2007 Dmitry Torokhov | ||
4 | * Copyright 2009 Johannes Berg <johannes@sipsolutions.net> | ||
5 | * | ||
6 | * This program is free software; you can redistribute it and/or modify | ||
7 | * it under the terms of the GNU General Public License as published by | ||
8 | * the Free Software Foundation; either version 2 of the License, or | ||
9 | * (at your option) any later version. | ||
10 | * | ||
11 | * This program is distributed in the hope that it will be useful, | ||
12 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | ||
13 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | ||
14 | * GNU General Public License for more details. | ||
15 | * | ||
16 | * You should have received a copy of the GNU General Public License | ||
17 | * along with this program; if not, write to the | ||
18 | * Free Software Foundation, Inc., | ||
19 | * 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. | ||
20 | */ | ||
21 | |||
22 | #include <linux/kernel.h> | ||
23 | #include <linux/module.h> | ||
24 | #include <linux/init.h> | ||
25 | #include <linux/workqueue.h> | ||
26 | #include <linux/capability.h> | ||
27 | #include <linux/list.h> | ||
28 | #include <linux/mutex.h> | ||
29 | #include <linux/rfkill.h> | ||
30 | #include <linux/spinlock.h> | ||
31 | |||
32 | #include "rfkill.h" | ||
33 | |||
34 | #define POLL_INTERVAL (5 * HZ) | ||
35 | |||
36 | #define RFKILL_BLOCK_HW BIT(0) | ||
37 | #define RFKILL_BLOCK_SW BIT(1) | ||
38 | #define RFKILL_BLOCK_SW_PREV BIT(2) | ||
39 | #define RFKILL_BLOCK_ANY (RFKILL_BLOCK_HW |\ | ||
40 | RFKILL_BLOCK_SW |\ | ||
41 | RFKILL_BLOCK_SW_PREV) | ||
42 | #define RFKILL_BLOCK_SW_SETCALL BIT(31) | ||
43 | |||
44 | struct rfkill { | ||
45 | spinlock_t lock; | ||
46 | |||
47 | const char *name; | ||
48 | enum rfkill_type type; | ||
49 | |||
50 | unsigned long state; | ||
51 | |||
52 | bool registered; | ||
53 | bool suspended; | ||
54 | |||
55 | const struct rfkill_ops *ops; | ||
56 | void *data; | ||
57 | |||
58 | #ifdef CONFIG_RFKILL_LEDS | ||
59 | struct led_trigger led_trigger; | ||
60 | const char *ledtrigname; | ||
61 | #endif | ||
62 | |||
63 | struct device dev; | ||
64 | struct list_head node; | ||
65 | |||
66 | struct delayed_work poll_work; | ||
67 | struct work_struct uevent_work; | ||
68 | struct work_struct sync_work; | ||
69 | }; | ||
70 | #define to_rfkill(d) container_of(d, struct rfkill, dev) | ||
71 | |||
72 | |||
73 | |||
74 | MODULE_AUTHOR("Ivo van Doorn <IvDoorn@gmail.com>"); | ||
75 | MODULE_AUTHOR("Johannes Berg <johannes@sipsolutions.net>"); | ||
76 | MODULE_DESCRIPTION("RF switch support"); | ||
77 | MODULE_LICENSE("GPL"); | ||
78 | |||
79 | |||
80 | /* | ||
81 | * The locking here should be made much smarter, we currently have | ||
82 | * a bit of a stupid situation because drivers might want to register | ||
83 | * the rfkill struct under their own lock, and take this lock during | ||
84 | * rfkill method calls -- which will cause an AB-BA deadlock situation. | ||
85 | * | ||
86 | * To fix that, we need to rework this code here to be mostly lock-free | ||
87 | * and only use the mutex for list manipulations, not to protect the | ||
88 | * various other global variables. Then we can avoid holding the mutex | ||
89 | * around driver operations, and all is happy. | ||
90 | */ | ||
91 | static LIST_HEAD(rfkill_list); /* list of registered rf switches */ | ||
92 | static DEFINE_MUTEX(rfkill_global_mutex); | ||
93 | |||
94 | static unsigned int rfkill_default_state = 1; | ||
95 | module_param_named(default_state, rfkill_default_state, uint, 0444); | ||
96 | MODULE_PARM_DESC(default_state, | ||
97 | "Default initial state for all radio types, 0 = radio off"); | ||
98 | |||
99 | static struct { | ||
100 | bool cur, def; | ||
101 | } rfkill_global_states[NUM_RFKILL_TYPES]; | ||
102 | |||
103 | static unsigned long rfkill_states_default_locked; | ||
104 | |||
105 | static bool rfkill_epo_lock_active; | ||
106 | |||
107 | |||
108 | #ifdef CONFIG_RFKILL_LEDS | ||
109 | static void rfkill_led_trigger_event(struct rfkill *rfkill) | ||
110 | { | ||
111 | struct led_trigger *trigger; | ||
112 | |||
113 | if (!rfkill->registered) | ||
114 | return; | ||
115 | |||
116 | trigger = &rfkill->led_trigger; | ||
117 | |||
118 | if (rfkill->state & RFKILL_BLOCK_ANY) | ||
119 | led_trigger_event(trigger, LED_OFF); | ||
120 | else | ||
121 | led_trigger_event(trigger, LED_FULL); | ||
122 | } | ||
123 | |||
124 | static void rfkill_led_trigger_activate(struct led_classdev *led) | ||
125 | { | ||
126 | struct rfkill *rfkill; | ||
127 | |||
128 | rfkill = container_of(led->trigger, struct rfkill, led_trigger); | ||
129 | |||
130 | rfkill_led_trigger_event(rfkill); | ||
131 | } | ||
132 | |||
133 | const char *rfkill_get_led_trigger_name(struct rfkill *rfkill) | ||
134 | { | ||
135 | return rfkill->led_trigger.name; | ||
136 | } | ||
137 | EXPORT_SYMBOL(rfkill_get_led_trigger_name); | ||
138 | |||
139 | void rfkill_set_led_trigger_name(struct rfkill *rfkill, const char *name) | ||
140 | { | ||
141 | BUG_ON(!rfkill); | ||
142 | |||
143 | rfkill->ledtrigname = name; | ||
144 | } | ||
145 | EXPORT_SYMBOL(rfkill_set_led_trigger_name); | ||
146 | |||
147 | static int rfkill_led_trigger_register(struct rfkill *rfkill) | ||
148 | { | ||
149 | rfkill->led_trigger.name = rfkill->ledtrigname | ||
150 | ? : dev_name(&rfkill->dev); | ||
151 | rfkill->led_trigger.activate = rfkill_led_trigger_activate; | ||
152 | return led_trigger_register(&rfkill->led_trigger); | ||
153 | } | ||
154 | |||
155 | static void rfkill_led_trigger_unregister(struct rfkill *rfkill) | ||
156 | { | ||
157 | led_trigger_unregister(&rfkill->led_trigger); | ||
158 | } | ||
159 | #else | ||
160 | static void rfkill_led_trigger_event(struct rfkill *rfkill) | ||
161 | { | ||
162 | } | ||
163 | |||
164 | static inline int rfkill_led_trigger_register(struct rfkill *rfkill) | ||
165 | { | ||
166 | return 0; | ||
167 | } | ||
168 | |||
169 | static inline void rfkill_led_trigger_unregister(struct rfkill *rfkill) | ||
170 | { | ||
171 | } | ||
172 | #endif /* CONFIG_RFKILL_LEDS */ | ||
173 | |||
174 | static void rfkill_uevent(struct rfkill *rfkill) | ||
175 | { | ||
176 | if (!rfkill->registered || rfkill->suspended) | ||
177 | return; | ||
178 | |||
179 | kobject_uevent(&rfkill->dev.kobj, KOBJ_CHANGE); | ||
180 | } | ||
181 | |||
182 | static bool __rfkill_set_hw_state(struct rfkill *rfkill, | ||
183 | bool blocked, bool *change) | ||
184 | { | ||
185 | unsigned long flags; | ||
186 | bool prev, any; | ||
187 | |||
188 | BUG_ON(!rfkill); | ||
189 | |||
190 | spin_lock_irqsave(&rfkill->lock, flags); | ||
191 | prev = !!(rfkill->state & RFKILL_BLOCK_HW); | ||
192 | if (blocked) | ||
193 | rfkill->state |= RFKILL_BLOCK_HW; | ||
194 | else | ||
195 | rfkill->state &= ~RFKILL_BLOCK_HW; | ||
196 | *change = prev != blocked; | ||
197 | any = rfkill->state & RFKILL_BLOCK_ANY; | ||
198 | spin_unlock_irqrestore(&rfkill->lock, flags); | ||
199 | |||
200 | rfkill_led_trigger_event(rfkill); | ||
201 | |||
202 | return any; | ||
203 | } | ||
204 | |||
205 | /** | ||
206 | * rfkill_set_block - wrapper for set_block method | ||
207 | * | ||
208 | * @rfkill: the rfkill struct to use | ||
209 | * @blocked: the new software state | ||
210 | * | ||
211 | * Calls the set_block method (when applicable) and handles notifications | ||
212 | * etc. as well. | ||
213 | */ | ||
214 | static void rfkill_set_block(struct rfkill *rfkill, bool blocked) | ||
215 | { | ||
216 | unsigned long flags; | ||
217 | int err; | ||
218 | |||
219 | /* | ||
220 | * Some platforms (...!) generate input events which affect the | ||
221 | * _hard_ kill state -- whenever something tries to change the | ||
222 | * current software state query the hardware state too. | ||
223 | */ | ||
224 | if (rfkill->ops->query) | ||
225 | rfkill->ops->query(rfkill, rfkill->data); | ||
226 | |||
227 | spin_lock_irqsave(&rfkill->lock, flags); | ||
228 | if (rfkill->state & RFKILL_BLOCK_SW) | ||
229 | rfkill->state |= RFKILL_BLOCK_SW_PREV; | ||
230 | else | ||
231 | rfkill->state &= ~RFKILL_BLOCK_SW_PREV; | ||
232 | |||
233 | if (blocked) | ||
234 | rfkill->state |= RFKILL_BLOCK_SW; | ||
235 | else | ||
236 | rfkill->state &= ~RFKILL_BLOCK_SW; | ||
237 | |||
238 | rfkill->state |= RFKILL_BLOCK_SW_SETCALL; | ||
239 | spin_unlock_irqrestore(&rfkill->lock, flags); | ||
240 | |||
241 | if (unlikely(rfkill->dev.power.power_state.event & PM_EVENT_SLEEP)) | ||
242 | return; | ||
243 | |||
244 | err = rfkill->ops->set_block(rfkill->data, blocked); | ||
245 | |||
246 | spin_lock_irqsave(&rfkill->lock, flags); | ||
247 | if (err) { | ||
248 | /* | ||
249 | * Failed -- reset status to _prev, this may be different | ||
250 | * from what set set _PREV to earlier in this function | ||
251 | * if rfkill_set_sw_state was invoked. | ||
252 | */ | ||
253 | if (rfkill->state & RFKILL_BLOCK_SW_PREV) | ||
254 | rfkill->state |= RFKILL_BLOCK_SW; | ||
255 | else | ||
256 | rfkill->state &= ~RFKILL_BLOCK_SW; | ||
257 | } | ||
258 | rfkill->state &= ~RFKILL_BLOCK_SW_SETCALL; | ||
259 | rfkill->state &= ~RFKILL_BLOCK_SW_PREV; | ||
260 | spin_unlock_irqrestore(&rfkill->lock, flags); | ||
261 | |||
262 | rfkill_led_trigger_event(rfkill); | ||
263 | rfkill_uevent(rfkill); | ||
264 | } | ||
265 | |||
266 | /** | ||
267 | * __rfkill_switch_all - Toggle state of all switches of given type | ||
268 | * @type: type of interfaces to be affected | ||
269 | * @state: the new state | ||
270 | * | ||
271 | * This function sets the state of all switches of given type, | ||
272 | * unless a specific switch is claimed by userspace (in which case, | ||
273 | * that switch is left alone) or suspended. | ||
274 | * | ||
275 | * Caller must have acquired rfkill_global_mutex. | ||
276 | */ | ||
277 | static void __rfkill_switch_all(const enum rfkill_type type, bool blocked) | ||
278 | { | ||
279 | struct rfkill *rfkill; | ||
280 | |||
281 | rfkill_global_states[type].cur = blocked; | ||
282 | list_for_each_entry(rfkill, &rfkill_list, node) { | ||
283 | if (rfkill->type != type) | ||
284 | continue; | ||
285 | |||
286 | rfkill_set_block(rfkill, blocked); | ||
287 | } | ||
288 | } | ||
289 | |||
290 | /** | ||
291 | * rfkill_switch_all - Toggle state of all switches of given type | ||
292 | * @type: type of interfaces to be affected | ||
293 | * @state: the new state | ||
294 | * | ||
295 | * Acquires rfkill_global_mutex and calls __rfkill_switch_all(@type, @state). | ||
296 | * Please refer to __rfkill_switch_all() for details. | ||
297 | * | ||
298 | * Does nothing if the EPO lock is active. | ||
299 | */ | ||
300 | void rfkill_switch_all(enum rfkill_type type, bool blocked) | ||
301 | { | ||
302 | mutex_lock(&rfkill_global_mutex); | ||
303 | |||
304 | if (!rfkill_epo_lock_active) | ||
305 | __rfkill_switch_all(type, blocked); | ||
306 | |||
307 | mutex_unlock(&rfkill_global_mutex); | ||
308 | } | ||
309 | |||
310 | /** | ||
311 | * rfkill_epo - emergency power off all transmitters | ||
312 | * | ||
313 | * This kicks all non-suspended rfkill devices to RFKILL_STATE_SOFT_BLOCKED, | ||
314 | * ignoring everything in its path but rfkill_global_mutex and rfkill->mutex. | ||
315 | * | ||
316 | * The global state before the EPO is saved and can be restored later | ||
317 | * using rfkill_restore_states(). | ||
318 | */ | ||
319 | void rfkill_epo(void) | ||
320 | { | ||
321 | struct rfkill *rfkill; | ||
322 | int i; | ||
323 | |||
324 | mutex_lock(&rfkill_global_mutex); | ||
325 | |||
326 | rfkill_epo_lock_active = true; | ||
327 | list_for_each_entry(rfkill, &rfkill_list, node) | ||
328 | rfkill_set_block(rfkill, true); | ||
329 | |||
330 | for (i = 0; i < NUM_RFKILL_TYPES; i++) { | ||
331 | rfkill_global_states[i].def = rfkill_global_states[i].cur; | ||
332 | rfkill_global_states[i].cur = true; | ||
333 | } | ||
334 | mutex_unlock(&rfkill_global_mutex); | ||
335 | } | ||
336 | |||
337 | /** | ||
338 | * rfkill_restore_states - restore global states | ||
339 | * | ||
340 | * Restore (and sync switches to) the global state from the | ||
341 | * states in rfkill_default_states. This can undo the effects of | ||
342 | * a call to rfkill_epo(). | ||
343 | */ | ||
344 | void rfkill_restore_states(void) | ||
345 | { | ||
346 | int i; | ||
347 | |||
348 | mutex_lock(&rfkill_global_mutex); | ||
349 | |||
350 | rfkill_epo_lock_active = false; | ||
351 | for (i = 0; i < NUM_RFKILL_TYPES; i++) | ||
352 | __rfkill_switch_all(i, rfkill_global_states[i].def); | ||
353 | mutex_unlock(&rfkill_global_mutex); | ||
354 | } | ||
355 | |||
356 | /** | ||
357 | * rfkill_remove_epo_lock - unlock state changes | ||
358 | * | ||
359 | * Used by rfkill-input manually unlock state changes, when | ||
360 | * the EPO switch is deactivated. | ||
361 | */ | ||
362 | void rfkill_remove_epo_lock(void) | ||
363 | { | ||
364 | mutex_lock(&rfkill_global_mutex); | ||
365 | rfkill_epo_lock_active = false; | ||
366 | mutex_unlock(&rfkill_global_mutex); | ||
367 | } | ||
368 | |||
369 | /** | ||
370 | * rfkill_is_epo_lock_active - returns true EPO is active | ||
371 | * | ||
372 | * Returns 0 (false) if there is NOT an active EPO contidion, | ||
373 | * and 1 (true) if there is an active EPO contition, which | ||
374 | * locks all radios in one of the BLOCKED states. | ||
375 | * | ||
376 | * Can be called in atomic context. | ||
377 | */ | ||
378 | bool rfkill_is_epo_lock_active(void) | ||
379 | { | ||
380 | return rfkill_epo_lock_active; | ||
381 | } | ||
382 | |||
383 | /** | ||
384 | * rfkill_get_global_sw_state - returns global state for a type | ||
385 | * @type: the type to get the global state of | ||
386 | * | ||
387 | * Returns the current global state for a given wireless | ||
388 | * device type. | ||
389 | */ | ||
390 | bool rfkill_get_global_sw_state(const enum rfkill_type type) | ||
391 | { | ||
392 | return rfkill_global_states[type].cur; | ||
393 | } | ||
394 | |||
395 | void rfkill_set_global_sw_state(const enum rfkill_type type, bool blocked) | ||
396 | { | ||
397 | mutex_lock(&rfkill_global_mutex); | ||
398 | |||
399 | /* don't allow unblock when epo */ | ||
400 | if (rfkill_epo_lock_active && !blocked) | ||
401 | goto out; | ||
402 | |||
403 | /* too late */ | ||
404 | if (rfkill_states_default_locked & BIT(type)) | ||
405 | goto out; | ||
406 | |||
407 | rfkill_states_default_locked |= BIT(type); | ||
408 | |||
409 | rfkill_global_states[type].cur = blocked; | ||
410 | rfkill_global_states[type].def = blocked; | ||
411 | out: | ||
412 | mutex_unlock(&rfkill_global_mutex); | ||
413 | } | ||
414 | EXPORT_SYMBOL(rfkill_set_global_sw_state); | ||
415 | |||
416 | |||
417 | bool rfkill_set_hw_state(struct rfkill *rfkill, bool blocked) | ||
418 | { | ||
419 | bool ret, change; | ||
420 | |||
421 | ret = __rfkill_set_hw_state(rfkill, blocked, &change); | ||
422 | |||
423 | if (!rfkill->registered) | ||
424 | return ret; | ||
425 | |||
426 | if (change) | ||
427 | schedule_work(&rfkill->uevent_work); | ||
428 | |||
429 | return ret; | ||
430 | } | ||
431 | EXPORT_SYMBOL(rfkill_set_hw_state); | ||
432 | |||
433 | static void __rfkill_set_sw_state(struct rfkill *rfkill, bool blocked) | ||
434 | { | ||
435 | u32 bit = RFKILL_BLOCK_SW; | ||
436 | |||
437 | /* if in a ops->set_block right now, use other bit */ | ||
438 | if (rfkill->state & RFKILL_BLOCK_SW_SETCALL) | ||
439 | bit = RFKILL_BLOCK_SW_PREV; | ||
440 | |||
441 | if (blocked) | ||
442 | rfkill->state |= bit; | ||
443 | else | ||
444 | rfkill->state &= ~bit; | ||
445 | } | ||
446 | |||
447 | bool rfkill_set_sw_state(struct rfkill *rfkill, bool blocked) | ||
448 | { | ||
449 | unsigned long flags; | ||
450 | bool prev, hwblock; | ||
451 | |||
452 | BUG_ON(!rfkill); | ||
453 | |||
454 | spin_lock_irqsave(&rfkill->lock, flags); | ||
455 | prev = !!(rfkill->state & RFKILL_BLOCK_SW); | ||
456 | __rfkill_set_sw_state(rfkill, blocked); | ||
457 | hwblock = !!(rfkill->state & RFKILL_BLOCK_HW); | ||
458 | blocked = blocked || hwblock; | ||
459 | spin_unlock_irqrestore(&rfkill->lock, flags); | ||
460 | |||
461 | if (!rfkill->registered) | ||
462 | return blocked; | ||
463 | |||
464 | if (prev != blocked && !hwblock) | ||
465 | schedule_work(&rfkill->uevent_work); | ||
466 | |||
467 | rfkill_led_trigger_event(rfkill); | ||
468 | |||
469 | return blocked; | ||
470 | } | ||
471 | EXPORT_SYMBOL(rfkill_set_sw_state); | ||
472 | |||
473 | void rfkill_set_states(struct rfkill *rfkill, bool sw, bool hw) | ||
474 | { | ||
475 | unsigned long flags; | ||
476 | bool swprev, hwprev; | ||
477 | |||
478 | BUG_ON(!rfkill); | ||
479 | |||
480 | spin_lock_irqsave(&rfkill->lock, flags); | ||
481 | |||
482 | /* | ||
483 | * No need to care about prev/setblock ... this is for uevent only | ||
484 | * and that will get triggered by rfkill_set_block anyway. | ||
485 | */ | ||
486 | swprev = !!(rfkill->state & RFKILL_BLOCK_SW); | ||
487 | hwprev = !!(rfkill->state & RFKILL_BLOCK_HW); | ||
488 | __rfkill_set_sw_state(rfkill, sw); | ||
489 | |||
490 | spin_unlock_irqrestore(&rfkill->lock, flags); | ||
491 | |||
492 | if (!rfkill->registered) | ||
493 | return; | ||
494 | |||
495 | if (swprev != sw || hwprev != hw) | ||
496 | schedule_work(&rfkill->uevent_work); | ||
497 | |||
498 | rfkill_led_trigger_event(rfkill); | ||
499 | } | ||
500 | EXPORT_SYMBOL(rfkill_set_states); | ||
501 | |||
502 | static ssize_t rfkill_name_show(struct device *dev, | ||
503 | struct device_attribute *attr, | ||
504 | char *buf) | ||
505 | { | ||
506 | struct rfkill *rfkill = to_rfkill(dev); | ||
507 | |||
508 | return sprintf(buf, "%s\n", rfkill->name); | ||
509 | } | ||
510 | |||
511 | static const char *rfkill_get_type_str(enum rfkill_type type) | ||
512 | { | ||
513 | switch (type) { | ||
514 | case RFKILL_TYPE_WLAN: | ||
515 | return "wlan"; | ||
516 | case RFKILL_TYPE_BLUETOOTH: | ||
517 | return "bluetooth"; | ||
518 | case RFKILL_TYPE_UWB: | ||
519 | return "ultrawideband"; | ||
520 | case RFKILL_TYPE_WIMAX: | ||
521 | return "wimax"; | ||
522 | case RFKILL_TYPE_WWAN: | ||
523 | return "wwan"; | ||
524 | default: | ||
525 | BUG(); | ||
526 | } | ||
527 | |||
528 | BUILD_BUG_ON(NUM_RFKILL_TYPES != RFKILL_TYPE_WWAN + 1); | ||
529 | } | ||
530 | |||
531 | static ssize_t rfkill_type_show(struct device *dev, | ||
532 | struct device_attribute *attr, | ||
533 | char *buf) | ||
534 | { | ||
535 | struct rfkill *rfkill = to_rfkill(dev); | ||
536 | |||
537 | return sprintf(buf, "%s\n", rfkill_get_type_str(rfkill->type)); | ||
538 | } | ||
539 | |||
540 | static u8 user_state_from_blocked(unsigned long state) | ||
541 | { | ||
542 | if (state & RFKILL_BLOCK_HW) | ||
543 | return RFKILL_USER_STATE_HARD_BLOCKED; | ||
544 | if (state & RFKILL_BLOCK_SW) | ||
545 | return RFKILL_USER_STATE_SOFT_BLOCKED; | ||
546 | |||
547 | return RFKILL_USER_STATE_UNBLOCKED; | ||
548 | } | ||
549 | |||
550 | static ssize_t rfkill_state_show(struct device *dev, | ||
551 | struct device_attribute *attr, | ||
552 | char *buf) | ||
553 | { | ||
554 | struct rfkill *rfkill = to_rfkill(dev); | ||
555 | unsigned long flags; | ||
556 | u32 state; | ||
557 | |||
558 | spin_lock_irqsave(&rfkill->lock, flags); | ||
559 | state = rfkill->state; | ||
560 | spin_unlock_irqrestore(&rfkill->lock, flags); | ||
561 | |||
562 | return sprintf(buf, "%d\n", user_state_from_blocked(state)); | ||
563 | } | ||
564 | |||
565 | static ssize_t rfkill_state_store(struct device *dev, | ||
566 | struct device_attribute *attr, | ||
567 | const char *buf, size_t count) | ||
568 | { | ||
569 | /* | ||
570 | * The intention was that userspace can only take control over | ||
571 | * a given device when/if rfkill-input doesn't control it due | ||
572 | * to user_claim. Since user_claim is currently unsupported, | ||
573 | * we never support changing the state from userspace -- this | ||
574 | * can be implemented again later. | ||
575 | */ | ||
576 | |||
577 | return -EPERM; | ||
578 | } | ||
579 | |||
580 | static ssize_t rfkill_claim_show(struct device *dev, | ||
581 | struct device_attribute *attr, | ||
582 | char *buf) | ||
583 | { | ||
584 | return sprintf(buf, "%d\n", 0); | ||
585 | } | ||
586 | |||
587 | static ssize_t rfkill_claim_store(struct device *dev, | ||
588 | struct device_attribute *attr, | ||
589 | const char *buf, size_t count) | ||
590 | { | ||
591 | return -EOPNOTSUPP; | ||
592 | } | ||
593 | |||
594 | static struct device_attribute rfkill_dev_attrs[] = { | ||
595 | __ATTR(name, S_IRUGO, rfkill_name_show, NULL), | ||
596 | __ATTR(type, S_IRUGO, rfkill_type_show, NULL), | ||
597 | __ATTR(state, S_IRUGO|S_IWUSR, rfkill_state_show, rfkill_state_store), | ||
598 | __ATTR(claim, S_IRUGO|S_IWUSR, rfkill_claim_show, rfkill_claim_store), | ||
599 | __ATTR_NULL | ||
600 | }; | ||
601 | |||
602 | static void rfkill_release(struct device *dev) | ||
603 | { | ||
604 | struct rfkill *rfkill = to_rfkill(dev); | ||
605 | |||
606 | kfree(rfkill); | ||
607 | } | ||
608 | |||
609 | static int rfkill_dev_uevent(struct device *dev, struct kobj_uevent_env *env) | ||
610 | { | ||
611 | struct rfkill *rfkill = to_rfkill(dev); | ||
612 | unsigned long flags; | ||
613 | u32 state; | ||
614 | int error; | ||
615 | |||
616 | error = add_uevent_var(env, "RFKILL_NAME=%s", rfkill->name); | ||
617 | if (error) | ||
618 | return error; | ||
619 | error = add_uevent_var(env, "RFKILL_TYPE=%s", | ||
620 | rfkill_get_type_str(rfkill->type)); | ||
621 | if (error) | ||
622 | return error; | ||
623 | spin_lock_irqsave(&rfkill->lock, flags); | ||
624 | state = rfkill->state; | ||
625 | spin_unlock_irqrestore(&rfkill->lock, flags); | ||
626 | error = add_uevent_var(env, "RFKILL_STATE=%d", | ||
627 | user_state_from_blocked(state)); | ||
628 | return error; | ||
629 | } | ||
630 | |||
631 | void rfkill_pause_polling(struct rfkill *rfkill) | ||
632 | { | ||
633 | BUG_ON(!rfkill); | ||
634 | |||
635 | if (!rfkill->ops->poll) | ||
636 | return; | ||
637 | |||
638 | cancel_delayed_work_sync(&rfkill->poll_work); | ||
639 | } | ||
640 | EXPORT_SYMBOL(rfkill_pause_polling); | ||
641 | |||
642 | void rfkill_resume_polling(struct rfkill *rfkill) | ||
643 | { | ||
644 | BUG_ON(!rfkill); | ||
645 | |||
646 | if (!rfkill->ops->poll) | ||
647 | return; | ||
648 | |||
649 | schedule_work(&rfkill->poll_work.work); | ||
650 | } | ||
651 | EXPORT_SYMBOL(rfkill_resume_polling); | ||
652 | |||
653 | static int rfkill_suspend(struct device *dev, pm_message_t state) | ||
654 | { | ||
655 | struct rfkill *rfkill = to_rfkill(dev); | ||
656 | |||
657 | rfkill_pause_polling(rfkill); | ||
658 | |||
659 | rfkill->suspended = true; | ||
660 | |||
661 | return 0; | ||
662 | } | ||
663 | |||
664 | static int rfkill_resume(struct device *dev) | ||
665 | { | ||
666 | struct rfkill *rfkill = to_rfkill(dev); | ||
667 | bool cur; | ||
668 | |||
669 | mutex_lock(&rfkill_global_mutex); | ||
670 | cur = rfkill_global_states[rfkill->type].cur; | ||
671 | rfkill_set_block(rfkill, cur); | ||
672 | mutex_unlock(&rfkill_global_mutex); | ||
673 | |||
674 | rfkill->suspended = false; | ||
675 | |||
676 | schedule_work(&rfkill->uevent_work); | ||
677 | |||
678 | rfkill_resume_polling(rfkill); | ||
679 | |||
680 | return 0; | ||
681 | } | ||
682 | |||
683 | static struct class rfkill_class = { | ||
684 | .name = "rfkill", | ||
685 | .dev_release = rfkill_release, | ||
686 | .dev_attrs = rfkill_dev_attrs, | ||
687 | .dev_uevent = rfkill_dev_uevent, | ||
688 | .suspend = rfkill_suspend, | ||
689 | .resume = rfkill_resume, | ||
690 | }; | ||
691 | |||
692 | |||
693 | struct rfkill * __must_check rfkill_alloc(const char *name, | ||
694 | struct device *parent, | ||
695 | const enum rfkill_type type, | ||
696 | const struct rfkill_ops *ops, | ||
697 | void *ops_data) | ||
698 | { | ||
699 | struct rfkill *rfkill; | ||
700 | struct device *dev; | ||
701 | |||
702 | if (WARN_ON(!ops)) | ||
703 | return NULL; | ||
704 | |||
705 | if (WARN_ON(!ops->set_block)) | ||
706 | return NULL; | ||
707 | |||
708 | if (WARN_ON(!name)) | ||
709 | return NULL; | ||
710 | |||
711 | if (WARN_ON(type >= NUM_RFKILL_TYPES)) | ||
712 | return NULL; | ||
713 | |||
714 | rfkill = kzalloc(sizeof(*rfkill), GFP_KERNEL); | ||
715 | if (!rfkill) | ||
716 | return NULL; | ||
717 | |||
718 | spin_lock_init(&rfkill->lock); | ||
719 | INIT_LIST_HEAD(&rfkill->node); | ||
720 | rfkill->type = type; | ||
721 | rfkill->name = name; | ||
722 | rfkill->ops = ops; | ||
723 | rfkill->data = ops_data; | ||
724 | |||
725 | dev = &rfkill->dev; | ||
726 | dev->class = &rfkill_class; | ||
727 | dev->parent = parent; | ||
728 | device_initialize(dev); | ||
729 | |||
730 | return rfkill; | ||
731 | } | ||
732 | EXPORT_SYMBOL(rfkill_alloc); | ||
733 | |||
734 | static void rfkill_poll(struct work_struct *work) | ||
735 | { | ||
736 | struct rfkill *rfkill; | ||
737 | |||
738 | rfkill = container_of(work, struct rfkill, poll_work.work); | ||
739 | |||
740 | /* | ||
741 | * Poll hardware state -- driver will use one of the | ||
742 | * rfkill_set{,_hw,_sw}_state functions and use its | ||
743 | * return value to update the current status. | ||
744 | */ | ||
745 | rfkill->ops->poll(rfkill, rfkill->data); | ||
746 | |||
747 | schedule_delayed_work(&rfkill->poll_work, | ||
748 | round_jiffies_relative(POLL_INTERVAL)); | ||
749 | } | ||
750 | |||
751 | static void rfkill_uevent_work(struct work_struct *work) | ||
752 | { | ||
753 | struct rfkill *rfkill; | ||
754 | |||
755 | rfkill = container_of(work, struct rfkill, uevent_work); | ||
756 | |||
757 | rfkill_uevent(rfkill); | ||
758 | } | ||
759 | |||
760 | static void rfkill_sync_work(struct work_struct *work) | ||
761 | { | ||
762 | struct rfkill *rfkill; | ||
763 | bool cur; | ||
764 | |||
765 | rfkill = container_of(work, struct rfkill, sync_work); | ||
766 | |||
767 | mutex_lock(&rfkill_global_mutex); | ||
768 | cur = rfkill_global_states[rfkill->type].cur; | ||
769 | rfkill_set_block(rfkill, cur); | ||
770 | mutex_unlock(&rfkill_global_mutex); | ||
771 | } | ||
772 | |||
773 | int __must_check rfkill_register(struct rfkill *rfkill) | ||
774 | { | ||
775 | static unsigned long rfkill_no; | ||
776 | struct device *dev = &rfkill->dev; | ||
777 | int error; | ||
778 | |||
779 | BUG_ON(!rfkill); | ||
780 | |||
781 | mutex_lock(&rfkill_global_mutex); | ||
782 | |||
783 | if (rfkill->registered) { | ||
784 | error = -EALREADY; | ||
785 | goto unlock; | ||
786 | } | ||
787 | |||
788 | dev_set_name(dev, "rfkill%lu", rfkill_no); | ||
789 | rfkill_no++; | ||
790 | |||
791 | if (!(rfkill_states_default_locked & BIT(rfkill->type))) { | ||
792 | /* first of its kind */ | ||
793 | BUILD_BUG_ON(NUM_RFKILL_TYPES > | ||
794 | sizeof(rfkill_states_default_locked) * 8); | ||
795 | rfkill_states_default_locked |= BIT(rfkill->type); | ||
796 | rfkill_global_states[rfkill->type].cur = | ||
797 | rfkill_global_states[rfkill->type].def; | ||
798 | } | ||
799 | |||
800 | list_add_tail(&rfkill->node, &rfkill_list); | ||
801 | |||
802 | error = device_add(dev); | ||
803 | if (error) | ||
804 | goto remove; | ||
805 | |||
806 | error = rfkill_led_trigger_register(rfkill); | ||
807 | if (error) | ||
808 | goto devdel; | ||
809 | |||
810 | rfkill->registered = true; | ||
811 | |||
812 | if (rfkill->ops->poll) { | ||
813 | INIT_DELAYED_WORK(&rfkill->poll_work, rfkill_poll); | ||
814 | schedule_delayed_work(&rfkill->poll_work, | ||
815 | round_jiffies_relative(POLL_INTERVAL)); | ||
816 | } | ||
817 | |||
818 | INIT_WORK(&rfkill->uevent_work, rfkill_uevent_work); | ||
819 | |||
820 | INIT_WORK(&rfkill->sync_work, rfkill_sync_work); | ||
821 | schedule_work(&rfkill->sync_work); | ||
822 | |||
823 | mutex_unlock(&rfkill_global_mutex); | ||
824 | return 0; | ||
825 | |||
826 | devdel: | ||
827 | device_del(&rfkill->dev); | ||
828 | remove: | ||
829 | list_del_init(&rfkill->node); | ||
830 | unlock: | ||
831 | mutex_unlock(&rfkill_global_mutex); | ||
832 | return error; | ||
833 | } | ||
834 | EXPORT_SYMBOL(rfkill_register); | ||
835 | |||
836 | void rfkill_unregister(struct rfkill *rfkill) | ||
837 | { | ||
838 | BUG_ON(!rfkill); | ||
839 | |||
840 | if (rfkill->ops->poll) | ||
841 | cancel_delayed_work_sync(&rfkill->poll_work); | ||
842 | |||
843 | cancel_work_sync(&rfkill->uevent_work); | ||
844 | cancel_work_sync(&rfkill->sync_work); | ||
845 | |||
846 | rfkill->registered = false; | ||
847 | |||
848 | device_del(&rfkill->dev); | ||
849 | |||
850 | mutex_lock(&rfkill_global_mutex); | ||
851 | list_del_init(&rfkill->node); | ||
852 | mutex_unlock(&rfkill_global_mutex); | ||
853 | |||
854 | rfkill_led_trigger_unregister(rfkill); | ||
855 | } | ||
856 | EXPORT_SYMBOL(rfkill_unregister); | ||
857 | |||
858 | void rfkill_destroy(struct rfkill *rfkill) | ||
859 | { | ||
860 | if (rfkill) | ||
861 | put_device(&rfkill->dev); | ||
862 | } | ||
863 | EXPORT_SYMBOL(rfkill_destroy); | ||
864 | |||
865 | |||
866 | static int __init rfkill_init(void) | ||
867 | { | ||
868 | int error; | ||
869 | int i; | ||
870 | |||
871 | for (i = 0; i < NUM_RFKILL_TYPES; i++) | ||
872 | rfkill_global_states[i].def = !rfkill_default_state; | ||
873 | |||
874 | error = class_register(&rfkill_class); | ||
875 | if (error) | ||
876 | goto out; | ||
877 | |||
878 | #ifdef CONFIG_RFKILL_INPUT | ||
879 | error = rfkill_handler_init(); | ||
880 | if (error) | ||
881 | class_unregister(&rfkill_class); | ||
882 | #endif | ||
883 | |||
884 | out: | ||
885 | return error; | ||
886 | } | ||
887 | subsys_initcall(rfkill_init); | ||
888 | |||
889 | static void __exit rfkill_exit(void) | ||
890 | { | ||
891 | #ifdef CONFIG_RFKILL_INPUT | ||
892 | rfkill_handler_exit(); | ||
893 | #endif | ||
894 | class_unregister(&rfkill_class); | ||
895 | } | ||
896 | module_exit(rfkill_exit); | ||
diff --git a/net/rfkill/input.c b/net/rfkill/input.c new file mode 100644 index 000000000000..a7295ad5f9cb --- /dev/null +++ b/net/rfkill/input.c | |||
@@ -0,0 +1,342 @@ | |||
1 | /* | ||
2 | * Input layer to RF Kill interface connector | ||
3 | * | ||
4 | * Copyright (c) 2007 Dmitry Torokhov | ||
5 | * Copyright 2009 Johannes Berg <johannes@sipsolutions.net> | ||
6 | * | ||
7 | * This program is free software; you can redistribute it and/or modify it | ||
8 | * under the terms of the GNU General Public License version 2 as published | ||
9 | * by the Free Software Foundation. | ||
10 | * | ||
11 | * If you ever run into a situation in which you have a SW_ type rfkill | ||
12 | * input device, then you can revive code that was removed in the patch | ||
13 | * "rfkill-input: remove unused code". | ||
14 | */ | ||
15 | |||
16 | #include <linux/input.h> | ||
17 | #include <linux/slab.h> | ||
18 | #include <linux/workqueue.h> | ||
19 | #include <linux/init.h> | ||
20 | #include <linux/rfkill.h> | ||
21 | #include <linux/sched.h> | ||
22 | |||
23 | #include "rfkill.h" | ||
24 | |||
25 | enum rfkill_input_master_mode { | ||
26 | RFKILL_INPUT_MASTER_UNLOCK = 0, | ||
27 | RFKILL_INPUT_MASTER_RESTORE = 1, | ||
28 | RFKILL_INPUT_MASTER_UNBLOCKALL = 2, | ||
29 | NUM_RFKILL_INPUT_MASTER_MODES | ||
30 | }; | ||
31 | |||
32 | /* Delay (in ms) between consecutive switch ops */ | ||
33 | #define RFKILL_OPS_DELAY 200 | ||
34 | |||
35 | static enum rfkill_input_master_mode rfkill_master_switch_mode = | ||
36 | RFKILL_INPUT_MASTER_UNBLOCKALL; | ||
37 | module_param_named(master_switch_mode, rfkill_master_switch_mode, uint, 0); | ||
38 | MODULE_PARM_DESC(master_switch_mode, | ||
39 | "SW_RFKILL_ALL ON should: 0=do nothing (only unlock); 1=restore; 2=unblock all"); | ||
40 | |||
41 | static spinlock_t rfkill_op_lock; | ||
42 | static bool rfkill_op_pending; | ||
43 | static unsigned long rfkill_sw_pending[BITS_TO_LONGS(NUM_RFKILL_TYPES)]; | ||
44 | static unsigned long rfkill_sw_state[BITS_TO_LONGS(NUM_RFKILL_TYPES)]; | ||
45 | |||
46 | enum rfkill_sched_op { | ||
47 | RFKILL_GLOBAL_OP_EPO = 0, | ||
48 | RFKILL_GLOBAL_OP_RESTORE, | ||
49 | RFKILL_GLOBAL_OP_UNLOCK, | ||
50 | RFKILL_GLOBAL_OP_UNBLOCK, | ||
51 | }; | ||
52 | |||
53 | static enum rfkill_sched_op rfkill_master_switch_op; | ||
54 | static enum rfkill_sched_op rfkill_op; | ||
55 | |||
56 | static void __rfkill_handle_global_op(enum rfkill_sched_op op) | ||
57 | { | ||
58 | unsigned int i; | ||
59 | |||
60 | switch (op) { | ||
61 | case RFKILL_GLOBAL_OP_EPO: | ||
62 | rfkill_epo(); | ||
63 | break; | ||
64 | case RFKILL_GLOBAL_OP_RESTORE: | ||
65 | rfkill_restore_states(); | ||
66 | break; | ||
67 | case RFKILL_GLOBAL_OP_UNLOCK: | ||
68 | rfkill_remove_epo_lock(); | ||
69 | break; | ||
70 | case RFKILL_GLOBAL_OP_UNBLOCK: | ||
71 | rfkill_remove_epo_lock(); | ||
72 | for (i = 0; i < NUM_RFKILL_TYPES; i++) | ||
73 | rfkill_switch_all(i, false); | ||
74 | break; | ||
75 | default: | ||
76 | /* memory corruption or bug, fail safely */ | ||
77 | rfkill_epo(); | ||
78 | WARN(1, "Unknown requested operation %d! " | ||
79 | "rfkill Emergency Power Off activated\n", | ||
80 | op); | ||
81 | } | ||
82 | } | ||
83 | |||
84 | static void __rfkill_handle_normal_op(const enum rfkill_type type, | ||
85 | const bool complement) | ||
86 | { | ||
87 | bool blocked; | ||
88 | |||
89 | blocked = rfkill_get_global_sw_state(type); | ||
90 | if (complement) | ||
91 | blocked = !blocked; | ||
92 | |||
93 | rfkill_switch_all(type, blocked); | ||
94 | } | ||
95 | |||
96 | static void rfkill_op_handler(struct work_struct *work) | ||
97 | { | ||
98 | unsigned int i; | ||
99 | bool c; | ||
100 | |||
101 | spin_lock_irq(&rfkill_op_lock); | ||
102 | do { | ||
103 | if (rfkill_op_pending) { | ||
104 | enum rfkill_sched_op op = rfkill_op; | ||
105 | rfkill_op_pending = false; | ||
106 | memset(rfkill_sw_pending, 0, | ||
107 | sizeof(rfkill_sw_pending)); | ||
108 | spin_unlock_irq(&rfkill_op_lock); | ||
109 | |||
110 | __rfkill_handle_global_op(op); | ||
111 | |||
112 | spin_lock_irq(&rfkill_op_lock); | ||
113 | |||
114 | /* | ||
115 | * handle global ops first -- during unlocked period | ||
116 | * we might have gotten a new global op. | ||
117 | */ | ||
118 | if (rfkill_op_pending) | ||
119 | continue; | ||
120 | } | ||
121 | |||
122 | if (rfkill_is_epo_lock_active()) | ||
123 | continue; | ||
124 | |||
125 | for (i = 0; i < NUM_RFKILL_TYPES; i++) { | ||
126 | if (__test_and_clear_bit(i, rfkill_sw_pending)) { | ||
127 | c = __test_and_clear_bit(i, rfkill_sw_state); | ||
128 | spin_unlock_irq(&rfkill_op_lock); | ||
129 | |||
130 | __rfkill_handle_normal_op(i, c); | ||
131 | |||
132 | spin_lock_irq(&rfkill_op_lock); | ||
133 | } | ||
134 | } | ||
135 | } while (rfkill_op_pending); | ||
136 | spin_unlock_irq(&rfkill_op_lock); | ||
137 | } | ||
138 | |||
139 | static DECLARE_DELAYED_WORK(rfkill_op_work, rfkill_op_handler); | ||
140 | static unsigned long rfkill_last_scheduled; | ||
141 | |||
142 | static unsigned long rfkill_ratelimit(const unsigned long last) | ||
143 | { | ||
144 | const unsigned long delay = msecs_to_jiffies(RFKILL_OPS_DELAY); | ||
145 | return (time_after(jiffies, last + delay)) ? 0 : delay; | ||
146 | } | ||
147 | |||
148 | static void rfkill_schedule_ratelimited(void) | ||
149 | { | ||
150 | if (delayed_work_pending(&rfkill_op_work)) | ||
151 | return; | ||
152 | schedule_delayed_work(&rfkill_op_work, | ||
153 | rfkill_ratelimit(rfkill_last_scheduled)); | ||
154 | rfkill_last_scheduled = jiffies; | ||
155 | } | ||
156 | |||
157 | static void rfkill_schedule_global_op(enum rfkill_sched_op op) | ||
158 | { | ||
159 | unsigned long flags; | ||
160 | |||
161 | spin_lock_irqsave(&rfkill_op_lock, flags); | ||
162 | rfkill_op = op; | ||
163 | rfkill_op_pending = true; | ||
164 | if (op == RFKILL_GLOBAL_OP_EPO && !rfkill_is_epo_lock_active()) { | ||
165 | /* bypass the limiter for EPO */ | ||
166 | cancel_delayed_work(&rfkill_op_work); | ||
167 | schedule_delayed_work(&rfkill_op_work, 0); | ||
168 | rfkill_last_scheduled = jiffies; | ||
169 | } else | ||
170 | rfkill_schedule_ratelimited(); | ||
171 | spin_unlock_irqrestore(&rfkill_op_lock, flags); | ||
172 | } | ||
173 | |||
174 | static void rfkill_schedule_toggle(enum rfkill_type type) | ||
175 | { | ||
176 | unsigned long flags; | ||
177 | |||
178 | if (rfkill_is_epo_lock_active()) | ||
179 | return; | ||
180 | |||
181 | spin_lock_irqsave(&rfkill_op_lock, flags); | ||
182 | if (!rfkill_op_pending) { | ||
183 | __set_bit(type, rfkill_sw_pending); | ||
184 | __change_bit(type, rfkill_sw_state); | ||
185 | rfkill_schedule_ratelimited(); | ||
186 | } | ||
187 | spin_unlock_irqrestore(&rfkill_op_lock, flags); | ||
188 | } | ||
189 | |||
190 | static void rfkill_schedule_evsw_rfkillall(int state) | ||
191 | { | ||
192 | if (state) | ||
193 | rfkill_schedule_global_op(rfkill_master_switch_op); | ||
194 | else | ||
195 | rfkill_schedule_global_op(RFKILL_GLOBAL_OP_EPO); | ||
196 | } | ||
197 | |||
198 | static void rfkill_event(struct input_handle *handle, unsigned int type, | ||
199 | unsigned int code, int data) | ||
200 | { | ||
201 | if (type == EV_KEY && data == 1) { | ||
202 | switch (code) { | ||
203 | case KEY_WLAN: | ||
204 | rfkill_schedule_toggle(RFKILL_TYPE_WLAN); | ||
205 | break; | ||
206 | case KEY_BLUETOOTH: | ||
207 | rfkill_schedule_toggle(RFKILL_TYPE_BLUETOOTH); | ||
208 | break; | ||
209 | case KEY_UWB: | ||
210 | rfkill_schedule_toggle(RFKILL_TYPE_UWB); | ||
211 | break; | ||
212 | case KEY_WIMAX: | ||
213 | rfkill_schedule_toggle(RFKILL_TYPE_WIMAX); | ||
214 | break; | ||
215 | } | ||
216 | } else if (type == EV_SW && code == SW_RFKILL_ALL) | ||
217 | rfkill_schedule_evsw_rfkillall(data); | ||
218 | } | ||
219 | |||
220 | static int rfkill_connect(struct input_handler *handler, struct input_dev *dev, | ||
221 | const struct input_device_id *id) | ||
222 | { | ||
223 | struct input_handle *handle; | ||
224 | int error; | ||
225 | |||
226 | handle = kzalloc(sizeof(struct input_handle), GFP_KERNEL); | ||
227 | if (!handle) | ||
228 | return -ENOMEM; | ||
229 | |||
230 | handle->dev = dev; | ||
231 | handle->handler = handler; | ||
232 | handle->name = "rfkill"; | ||
233 | |||
234 | /* causes rfkill_start() to be called */ | ||
235 | error = input_register_handle(handle); | ||
236 | if (error) | ||
237 | goto err_free_handle; | ||
238 | |||
239 | error = input_open_device(handle); | ||
240 | if (error) | ||
241 | goto err_unregister_handle; | ||
242 | |||
243 | return 0; | ||
244 | |||
245 | err_unregister_handle: | ||
246 | input_unregister_handle(handle); | ||
247 | err_free_handle: | ||
248 | kfree(handle); | ||
249 | return error; | ||
250 | } | ||
251 | |||
252 | static void rfkill_start(struct input_handle *handle) | ||
253 | { | ||
254 | /* | ||
255 | * Take event_lock to guard against configuration changes, we | ||
256 | * should be able to deal with concurrency with rfkill_event() | ||
257 | * just fine (which event_lock will also avoid). | ||
258 | */ | ||
259 | spin_lock_irq(&handle->dev->event_lock); | ||
260 | |||
261 | if (test_bit(EV_SW, handle->dev->evbit) && | ||
262 | test_bit(SW_RFKILL_ALL, handle->dev->swbit)) | ||
263 | rfkill_schedule_evsw_rfkillall(test_bit(SW_RFKILL_ALL, | ||
264 | handle->dev->sw)); | ||
265 | |||
266 | spin_unlock_irq(&handle->dev->event_lock); | ||
267 | } | ||
268 | |||
269 | static void rfkill_disconnect(struct input_handle *handle) | ||
270 | { | ||
271 | input_close_device(handle); | ||
272 | input_unregister_handle(handle); | ||
273 | kfree(handle); | ||
274 | } | ||
275 | |||
276 | static const struct input_device_id rfkill_ids[] = { | ||
277 | { | ||
278 | .flags = INPUT_DEVICE_ID_MATCH_EVBIT | INPUT_DEVICE_ID_MATCH_KEYBIT, | ||
279 | .evbit = { BIT_MASK(EV_KEY) }, | ||
280 | .keybit = { [BIT_WORD(KEY_WLAN)] = BIT_MASK(KEY_WLAN) }, | ||
281 | }, | ||
282 | { | ||
283 | .flags = INPUT_DEVICE_ID_MATCH_EVBIT | INPUT_DEVICE_ID_MATCH_KEYBIT, | ||
284 | .evbit = { BIT_MASK(EV_KEY) }, | ||
285 | .keybit = { [BIT_WORD(KEY_BLUETOOTH)] = BIT_MASK(KEY_BLUETOOTH) }, | ||
286 | }, | ||
287 | { | ||
288 | .flags = INPUT_DEVICE_ID_MATCH_EVBIT | INPUT_DEVICE_ID_MATCH_KEYBIT, | ||
289 | .evbit = { BIT_MASK(EV_KEY) }, | ||
290 | .keybit = { [BIT_WORD(KEY_UWB)] = BIT_MASK(KEY_UWB) }, | ||
291 | }, | ||
292 | { | ||
293 | .flags = INPUT_DEVICE_ID_MATCH_EVBIT | INPUT_DEVICE_ID_MATCH_KEYBIT, | ||
294 | .evbit = { BIT_MASK(EV_KEY) }, | ||
295 | .keybit = { [BIT_WORD(KEY_WIMAX)] = BIT_MASK(KEY_WIMAX) }, | ||
296 | }, | ||
297 | { | ||
298 | .flags = INPUT_DEVICE_ID_MATCH_EVBIT | INPUT_DEVICE_ID_MATCH_SWBIT, | ||
299 | .evbit = { BIT(EV_SW) }, | ||
300 | .swbit = { [BIT_WORD(SW_RFKILL_ALL)] = BIT_MASK(SW_RFKILL_ALL) }, | ||
301 | }, | ||
302 | { } | ||
303 | }; | ||
304 | |||
305 | static struct input_handler rfkill_handler = { | ||
306 | .name = "rfkill", | ||
307 | .event = rfkill_event, | ||
308 | .connect = rfkill_connect, | ||
309 | .start = rfkill_start, | ||
310 | .disconnect = rfkill_disconnect, | ||
311 | .id_table = rfkill_ids, | ||
312 | }; | ||
313 | |||
314 | int __init rfkill_handler_init(void) | ||
315 | { | ||
316 | switch (rfkill_master_switch_mode) { | ||
317 | case RFKILL_INPUT_MASTER_UNBLOCKALL: | ||
318 | rfkill_master_switch_op = RFKILL_GLOBAL_OP_UNBLOCK; | ||
319 | break; | ||
320 | case RFKILL_INPUT_MASTER_RESTORE: | ||
321 | rfkill_master_switch_op = RFKILL_GLOBAL_OP_RESTORE; | ||
322 | break; | ||
323 | case RFKILL_INPUT_MASTER_UNLOCK: | ||
324 | rfkill_master_switch_op = RFKILL_GLOBAL_OP_UNLOCK; | ||
325 | break; | ||
326 | default: | ||
327 | return -EINVAL; | ||
328 | } | ||
329 | |||
330 | spin_lock_init(&rfkill_op_lock); | ||
331 | |||
332 | /* Avoid delay at first schedule */ | ||
333 | rfkill_last_scheduled = | ||
334 | jiffies - msecs_to_jiffies(RFKILL_OPS_DELAY) - 1; | ||
335 | return input_register_handler(&rfkill_handler); | ||
336 | } | ||
337 | |||
338 | void __exit rfkill_handler_exit(void) | ||
339 | { | ||
340 | input_unregister_handler(&rfkill_handler); | ||
341 | cancel_delayed_work_sync(&rfkill_op_work); | ||
342 | } | ||
diff --git a/net/rfkill/rfkill-input.c b/net/rfkill/rfkill-input.c deleted file mode 100644 index 60a34f3b5f65..000000000000 --- a/net/rfkill/rfkill-input.c +++ /dev/null | |||
@@ -1,390 +0,0 @@ | |||
1 | /* | ||
2 | * Input layer to RF Kill interface connector | ||
3 | * | ||
4 | * Copyright (c) 2007 Dmitry Torokhov | ||
5 | */ | ||
6 | |||
7 | /* | ||
8 | * This program is free software; you can redistribute it and/or modify it | ||
9 | * under the terms of the GNU General Public License version 2 as published | ||
10 | * by the Free Software Foundation. | ||
11 | */ | ||
12 | |||
13 | #include <linux/module.h> | ||
14 | #include <linux/input.h> | ||
15 | #include <linux/slab.h> | ||
16 | #include <linux/workqueue.h> | ||
17 | #include <linux/init.h> | ||
18 | #include <linux/rfkill.h> | ||
19 | #include <linux/sched.h> | ||
20 | |||
21 | #include "rfkill-input.h" | ||
22 | |||
23 | MODULE_AUTHOR("Dmitry Torokhov <dtor@mail.ru>"); | ||
24 | MODULE_DESCRIPTION("Input layer to RF switch connector"); | ||
25 | MODULE_LICENSE("GPL"); | ||
26 | |||
27 | enum rfkill_input_master_mode { | ||
28 | RFKILL_INPUT_MASTER_DONOTHING = 0, | ||
29 | RFKILL_INPUT_MASTER_RESTORE = 1, | ||
30 | RFKILL_INPUT_MASTER_UNBLOCKALL = 2, | ||
31 | RFKILL_INPUT_MASTER_MAX, /* marker */ | ||
32 | }; | ||
33 | |||
34 | /* Delay (in ms) between consecutive switch ops */ | ||
35 | #define RFKILL_OPS_DELAY 200 | ||
36 | |||
37 | static enum rfkill_input_master_mode rfkill_master_switch_mode = | ||
38 | RFKILL_INPUT_MASTER_UNBLOCKALL; | ||
39 | module_param_named(master_switch_mode, rfkill_master_switch_mode, uint, 0); | ||
40 | MODULE_PARM_DESC(master_switch_mode, | ||
41 | "SW_RFKILL_ALL ON should: 0=do nothing; 1=restore; 2=unblock all"); | ||
42 | |||
43 | enum rfkill_global_sched_op { | ||
44 | RFKILL_GLOBAL_OP_EPO = 0, | ||
45 | RFKILL_GLOBAL_OP_RESTORE, | ||
46 | RFKILL_GLOBAL_OP_UNLOCK, | ||
47 | RFKILL_GLOBAL_OP_UNBLOCK, | ||
48 | }; | ||
49 | |||
50 | struct rfkill_task { | ||
51 | struct delayed_work dwork; | ||
52 | |||
53 | /* ensures that task is serialized */ | ||
54 | struct mutex mutex; | ||
55 | |||
56 | /* protects everything below */ | ||
57 | spinlock_t lock; | ||
58 | |||
59 | /* pending regular switch operations (1=pending) */ | ||
60 | unsigned long sw_pending[BITS_TO_LONGS(RFKILL_TYPE_MAX)]; | ||
61 | |||
62 | /* should the state be complemented (1=yes) */ | ||
63 | unsigned long sw_togglestate[BITS_TO_LONGS(RFKILL_TYPE_MAX)]; | ||
64 | |||
65 | bool global_op_pending; | ||
66 | enum rfkill_global_sched_op op; | ||
67 | |||
68 | /* last time it was scheduled */ | ||
69 | unsigned long last_scheduled; | ||
70 | }; | ||
71 | |||
72 | static void __rfkill_handle_global_op(enum rfkill_global_sched_op op) | ||
73 | { | ||
74 | unsigned int i; | ||
75 | |||
76 | switch (op) { | ||
77 | case RFKILL_GLOBAL_OP_EPO: | ||
78 | rfkill_epo(); | ||
79 | break; | ||
80 | case RFKILL_GLOBAL_OP_RESTORE: | ||
81 | rfkill_restore_states(); | ||
82 | break; | ||
83 | case RFKILL_GLOBAL_OP_UNLOCK: | ||
84 | rfkill_remove_epo_lock(); | ||
85 | break; | ||
86 | case RFKILL_GLOBAL_OP_UNBLOCK: | ||
87 | rfkill_remove_epo_lock(); | ||
88 | for (i = 0; i < RFKILL_TYPE_MAX; i++) | ||
89 | rfkill_switch_all(i, RFKILL_STATE_UNBLOCKED); | ||
90 | break; | ||
91 | default: | ||
92 | /* memory corruption or bug, fail safely */ | ||
93 | rfkill_epo(); | ||
94 | WARN(1, "Unknown requested operation %d! " | ||
95 | "rfkill Emergency Power Off activated\n", | ||
96 | op); | ||
97 | } | ||
98 | } | ||
99 | |||
100 | static void __rfkill_handle_normal_op(const enum rfkill_type type, | ||
101 | const bool c) | ||
102 | { | ||
103 | enum rfkill_state state; | ||
104 | |||
105 | state = rfkill_get_global_state(type); | ||
106 | if (c) | ||
107 | state = rfkill_state_complement(state); | ||
108 | |||
109 | rfkill_switch_all(type, state); | ||
110 | } | ||
111 | |||
112 | static void rfkill_task_handler(struct work_struct *work) | ||
113 | { | ||
114 | struct rfkill_task *task = container_of(work, | ||
115 | struct rfkill_task, dwork.work); | ||
116 | bool doit = true; | ||
117 | |||
118 | mutex_lock(&task->mutex); | ||
119 | |||
120 | spin_lock_irq(&task->lock); | ||
121 | while (doit) { | ||
122 | if (task->global_op_pending) { | ||
123 | enum rfkill_global_sched_op op = task->op; | ||
124 | task->global_op_pending = false; | ||
125 | memset(task->sw_pending, 0, sizeof(task->sw_pending)); | ||
126 | spin_unlock_irq(&task->lock); | ||
127 | |||
128 | __rfkill_handle_global_op(op); | ||
129 | |||
130 | /* make sure we do at least one pass with | ||
131 | * !task->global_op_pending */ | ||
132 | spin_lock_irq(&task->lock); | ||
133 | continue; | ||
134 | } else if (!rfkill_is_epo_lock_active()) { | ||
135 | unsigned int i = 0; | ||
136 | |||
137 | while (!task->global_op_pending && | ||
138 | i < RFKILL_TYPE_MAX) { | ||
139 | if (test_and_clear_bit(i, task->sw_pending)) { | ||
140 | bool c; | ||
141 | c = test_and_clear_bit(i, | ||
142 | task->sw_togglestate); | ||
143 | spin_unlock_irq(&task->lock); | ||
144 | |||
145 | __rfkill_handle_normal_op(i, c); | ||
146 | |||
147 | spin_lock_irq(&task->lock); | ||
148 | } | ||
149 | i++; | ||
150 | } | ||
151 | } | ||
152 | doit = task->global_op_pending; | ||
153 | } | ||
154 | spin_unlock_irq(&task->lock); | ||
155 | |||
156 | mutex_unlock(&task->mutex); | ||
157 | } | ||
158 | |||
159 | static struct rfkill_task rfkill_task = { | ||
160 | .dwork = __DELAYED_WORK_INITIALIZER(rfkill_task.dwork, | ||
161 | rfkill_task_handler), | ||
162 | .mutex = __MUTEX_INITIALIZER(rfkill_task.mutex), | ||
163 | .lock = __SPIN_LOCK_UNLOCKED(rfkill_task.lock), | ||
164 | }; | ||
165 | |||
166 | static unsigned long rfkill_ratelimit(const unsigned long last) | ||
167 | { | ||
168 | const unsigned long delay = msecs_to_jiffies(RFKILL_OPS_DELAY); | ||
169 | return (time_after(jiffies, last + delay)) ? 0 : delay; | ||
170 | } | ||
171 | |||
172 | static void rfkill_schedule_ratelimited(void) | ||
173 | { | ||
174 | if (!delayed_work_pending(&rfkill_task.dwork)) { | ||
175 | schedule_delayed_work(&rfkill_task.dwork, | ||
176 | rfkill_ratelimit(rfkill_task.last_scheduled)); | ||
177 | rfkill_task.last_scheduled = jiffies; | ||
178 | } | ||
179 | } | ||
180 | |||
181 | static void rfkill_schedule_global_op(enum rfkill_global_sched_op op) | ||
182 | { | ||
183 | unsigned long flags; | ||
184 | |||
185 | spin_lock_irqsave(&rfkill_task.lock, flags); | ||
186 | rfkill_task.op = op; | ||
187 | rfkill_task.global_op_pending = true; | ||
188 | if (op == RFKILL_GLOBAL_OP_EPO && !rfkill_is_epo_lock_active()) { | ||
189 | /* bypass the limiter for EPO */ | ||
190 | cancel_delayed_work(&rfkill_task.dwork); | ||
191 | schedule_delayed_work(&rfkill_task.dwork, 0); | ||
192 | rfkill_task.last_scheduled = jiffies; | ||
193 | } else | ||
194 | rfkill_schedule_ratelimited(); | ||
195 | spin_unlock_irqrestore(&rfkill_task.lock, flags); | ||
196 | } | ||
197 | |||
198 | static void rfkill_schedule_toggle(enum rfkill_type type) | ||
199 | { | ||
200 | unsigned long flags; | ||
201 | |||
202 | if (rfkill_is_epo_lock_active()) | ||
203 | return; | ||
204 | |||
205 | spin_lock_irqsave(&rfkill_task.lock, flags); | ||
206 | if (!rfkill_task.global_op_pending) { | ||
207 | set_bit(type, rfkill_task.sw_pending); | ||
208 | change_bit(type, rfkill_task.sw_togglestate); | ||
209 | rfkill_schedule_ratelimited(); | ||
210 | } | ||
211 | spin_unlock_irqrestore(&rfkill_task.lock, flags); | ||
212 | } | ||
213 | |||
214 | static void rfkill_schedule_evsw_rfkillall(int state) | ||
215 | { | ||
216 | if (state) { | ||
217 | switch (rfkill_master_switch_mode) { | ||
218 | case RFKILL_INPUT_MASTER_UNBLOCKALL: | ||
219 | rfkill_schedule_global_op(RFKILL_GLOBAL_OP_UNBLOCK); | ||
220 | break; | ||
221 | case RFKILL_INPUT_MASTER_RESTORE: | ||
222 | rfkill_schedule_global_op(RFKILL_GLOBAL_OP_RESTORE); | ||
223 | break; | ||
224 | case RFKILL_INPUT_MASTER_DONOTHING: | ||
225 | rfkill_schedule_global_op(RFKILL_GLOBAL_OP_UNLOCK); | ||
226 | break; | ||
227 | default: | ||
228 | /* memory corruption or driver bug! fail safely */ | ||
229 | rfkill_schedule_global_op(RFKILL_GLOBAL_OP_EPO); | ||
230 | WARN(1, "Unknown rfkill_master_switch_mode (%d), " | ||
231 | "driver bug or memory corruption detected!\n", | ||
232 | rfkill_master_switch_mode); | ||
233 | break; | ||
234 | } | ||
235 | } else | ||
236 | rfkill_schedule_global_op(RFKILL_GLOBAL_OP_EPO); | ||
237 | } | ||
238 | |||
239 | static void rfkill_event(struct input_handle *handle, unsigned int type, | ||
240 | unsigned int code, int data) | ||
241 | { | ||
242 | if (type == EV_KEY && data == 1) { | ||
243 | enum rfkill_type t; | ||
244 | |||
245 | switch (code) { | ||
246 | case KEY_WLAN: | ||
247 | t = RFKILL_TYPE_WLAN; | ||
248 | break; | ||
249 | case KEY_BLUETOOTH: | ||
250 | t = RFKILL_TYPE_BLUETOOTH; | ||
251 | break; | ||
252 | case KEY_UWB: | ||
253 | t = RFKILL_TYPE_UWB; | ||
254 | break; | ||
255 | case KEY_WIMAX: | ||
256 | t = RFKILL_TYPE_WIMAX; | ||
257 | break; | ||
258 | default: | ||
259 | return; | ||
260 | } | ||
261 | rfkill_schedule_toggle(t); | ||
262 | return; | ||
263 | } else if (type == EV_SW) { | ||
264 | switch (code) { | ||
265 | case SW_RFKILL_ALL: | ||
266 | rfkill_schedule_evsw_rfkillall(data); | ||
267 | return; | ||
268 | default: | ||
269 | return; | ||
270 | } | ||
271 | } | ||
272 | } | ||
273 | |||
274 | static int rfkill_connect(struct input_handler *handler, struct input_dev *dev, | ||
275 | const struct input_device_id *id) | ||
276 | { | ||
277 | struct input_handle *handle; | ||
278 | int error; | ||
279 | |||
280 | handle = kzalloc(sizeof(struct input_handle), GFP_KERNEL); | ||
281 | if (!handle) | ||
282 | return -ENOMEM; | ||
283 | |||
284 | handle->dev = dev; | ||
285 | handle->handler = handler; | ||
286 | handle->name = "rfkill"; | ||
287 | |||
288 | /* causes rfkill_start() to be called */ | ||
289 | error = input_register_handle(handle); | ||
290 | if (error) | ||
291 | goto err_free_handle; | ||
292 | |||
293 | error = input_open_device(handle); | ||
294 | if (error) | ||
295 | goto err_unregister_handle; | ||
296 | |||
297 | return 0; | ||
298 | |||
299 | err_unregister_handle: | ||
300 | input_unregister_handle(handle); | ||
301 | err_free_handle: | ||
302 | kfree(handle); | ||
303 | return error; | ||
304 | } | ||
305 | |||
306 | static void rfkill_start(struct input_handle *handle) | ||
307 | { | ||
308 | /* Take event_lock to guard against configuration changes, we | ||
309 | * should be able to deal with concurrency with rfkill_event() | ||
310 | * just fine (which event_lock will also avoid). */ | ||
311 | spin_lock_irq(&handle->dev->event_lock); | ||
312 | |||
313 | if (test_bit(EV_SW, handle->dev->evbit)) { | ||
314 | if (test_bit(SW_RFKILL_ALL, handle->dev->swbit)) | ||
315 | rfkill_schedule_evsw_rfkillall(test_bit(SW_RFKILL_ALL, | ||
316 | handle->dev->sw)); | ||
317 | /* add resync for further EV_SW events here */ | ||
318 | } | ||
319 | |||
320 | spin_unlock_irq(&handle->dev->event_lock); | ||
321 | } | ||
322 | |||
323 | static void rfkill_disconnect(struct input_handle *handle) | ||
324 | { | ||
325 | input_close_device(handle); | ||
326 | input_unregister_handle(handle); | ||
327 | kfree(handle); | ||
328 | } | ||
329 | |||
330 | static const struct input_device_id rfkill_ids[] = { | ||
331 | { | ||
332 | .flags = INPUT_DEVICE_ID_MATCH_EVBIT | INPUT_DEVICE_ID_MATCH_KEYBIT, | ||
333 | .evbit = { BIT_MASK(EV_KEY) }, | ||
334 | .keybit = { [BIT_WORD(KEY_WLAN)] = BIT_MASK(KEY_WLAN) }, | ||
335 | }, | ||
336 | { | ||
337 | .flags = INPUT_DEVICE_ID_MATCH_EVBIT | INPUT_DEVICE_ID_MATCH_KEYBIT, | ||
338 | .evbit = { BIT_MASK(EV_KEY) }, | ||
339 | .keybit = { [BIT_WORD(KEY_BLUETOOTH)] = BIT_MASK(KEY_BLUETOOTH) }, | ||
340 | }, | ||
341 | { | ||
342 | .flags = INPUT_DEVICE_ID_MATCH_EVBIT | INPUT_DEVICE_ID_MATCH_KEYBIT, | ||
343 | .evbit = { BIT_MASK(EV_KEY) }, | ||
344 | .keybit = { [BIT_WORD(KEY_UWB)] = BIT_MASK(KEY_UWB) }, | ||
345 | }, | ||
346 | { | ||
347 | .flags = INPUT_DEVICE_ID_MATCH_EVBIT | INPUT_DEVICE_ID_MATCH_KEYBIT, | ||
348 | .evbit = { BIT_MASK(EV_KEY) }, | ||
349 | .keybit = { [BIT_WORD(KEY_WIMAX)] = BIT_MASK(KEY_WIMAX) }, | ||
350 | }, | ||
351 | { | ||
352 | .flags = INPUT_DEVICE_ID_MATCH_EVBIT | INPUT_DEVICE_ID_MATCH_SWBIT, | ||
353 | .evbit = { BIT(EV_SW) }, | ||
354 | .swbit = { [BIT_WORD(SW_RFKILL_ALL)] = BIT_MASK(SW_RFKILL_ALL) }, | ||
355 | }, | ||
356 | { } | ||
357 | }; | ||
358 | |||
359 | static struct input_handler rfkill_handler = { | ||
360 | .event = rfkill_event, | ||
361 | .connect = rfkill_connect, | ||
362 | .disconnect = rfkill_disconnect, | ||
363 | .start = rfkill_start, | ||
364 | .name = "rfkill", | ||
365 | .id_table = rfkill_ids, | ||
366 | }; | ||
367 | |||
368 | static int __init rfkill_handler_init(void) | ||
369 | { | ||
370 | if (rfkill_master_switch_mode >= RFKILL_INPUT_MASTER_MAX) | ||
371 | return -EINVAL; | ||
372 | |||
373 | /* | ||
374 | * The penalty to not doing this is a possible RFKILL_OPS_DELAY delay | ||
375 | * at the first use. Acceptable, but if we can avoid it, why not? | ||
376 | */ | ||
377 | rfkill_task.last_scheduled = | ||
378 | jiffies - msecs_to_jiffies(RFKILL_OPS_DELAY) - 1; | ||
379 | return input_register_handler(&rfkill_handler); | ||
380 | } | ||
381 | |||
382 | static void __exit rfkill_handler_exit(void) | ||
383 | { | ||
384 | input_unregister_handler(&rfkill_handler); | ||
385 | cancel_delayed_work_sync(&rfkill_task.dwork); | ||
386 | rfkill_remove_epo_lock(); | ||
387 | } | ||
388 | |||
389 | module_init(rfkill_handler_init); | ||
390 | module_exit(rfkill_handler_exit); | ||
diff --git a/net/rfkill/rfkill.c b/net/rfkill/rfkill.c deleted file mode 100644 index 4f5a83183c95..000000000000 --- a/net/rfkill/rfkill.c +++ /dev/null | |||
@@ -1,855 +0,0 @@ | |||
1 | /* | ||
2 | * Copyright (C) 2006 - 2007 Ivo van Doorn | ||
3 | * Copyright (C) 2007 Dmitry Torokhov | ||
4 | * | ||
5 | * This program is free software; you can redistribute it and/or modify | ||
6 | * it under the terms of the GNU General Public License as published by | ||
7 | * the Free Software Foundation; either version 2 of the License, or | ||
8 | * (at your option) any later version. | ||
9 | * | ||
10 | * This program is distributed in the hope that it will be useful, | ||
11 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | ||
12 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | ||
13 | * GNU General Public License for more details. | ||
14 | * | ||
15 | * You should have received a copy of the GNU General Public License | ||
16 | * along with this program; if not, write to the | ||
17 | * Free Software Foundation, Inc., | ||
18 | * 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. | ||
19 | */ | ||
20 | |||
21 | #include <linux/kernel.h> | ||
22 | #include <linux/module.h> | ||
23 | #include <linux/init.h> | ||
24 | #include <linux/workqueue.h> | ||
25 | #include <linux/capability.h> | ||
26 | #include <linux/list.h> | ||
27 | #include <linux/mutex.h> | ||
28 | #include <linux/rfkill.h> | ||
29 | |||
30 | /* Get declaration of rfkill_switch_all() to shut up sparse. */ | ||
31 | #include "rfkill-input.h" | ||
32 | |||
33 | |||
34 | MODULE_AUTHOR("Ivo van Doorn <IvDoorn@gmail.com>"); | ||
35 | MODULE_VERSION("1.0"); | ||
36 | MODULE_DESCRIPTION("RF switch support"); | ||
37 | MODULE_LICENSE("GPL"); | ||
38 | |||
39 | static LIST_HEAD(rfkill_list); /* list of registered rf switches */ | ||
40 | static DEFINE_MUTEX(rfkill_global_mutex); | ||
41 | |||
42 | static unsigned int rfkill_default_state = RFKILL_STATE_UNBLOCKED; | ||
43 | module_param_named(default_state, rfkill_default_state, uint, 0444); | ||
44 | MODULE_PARM_DESC(default_state, | ||
45 | "Default initial state for all radio types, 0 = radio off"); | ||
46 | |||
47 | struct rfkill_gsw_state { | ||
48 | enum rfkill_state current_state; | ||
49 | enum rfkill_state default_state; | ||
50 | }; | ||
51 | |||
52 | static struct rfkill_gsw_state rfkill_global_states[RFKILL_TYPE_MAX]; | ||
53 | static unsigned long rfkill_states_lockdflt[BITS_TO_LONGS(RFKILL_TYPE_MAX)]; | ||
54 | static bool rfkill_epo_lock_active; | ||
55 | |||
56 | |||
57 | #ifdef CONFIG_RFKILL_LEDS | ||
58 | static void rfkill_led_trigger(struct rfkill *rfkill, | ||
59 | enum rfkill_state state) | ||
60 | { | ||
61 | struct led_trigger *led = &rfkill->led_trigger; | ||
62 | |||
63 | if (!led->name) | ||
64 | return; | ||
65 | if (state != RFKILL_STATE_UNBLOCKED) | ||
66 | led_trigger_event(led, LED_OFF); | ||
67 | else | ||
68 | led_trigger_event(led, LED_FULL); | ||
69 | } | ||
70 | |||
71 | static void rfkill_led_trigger_activate(struct led_classdev *led) | ||
72 | { | ||
73 | struct rfkill *rfkill = container_of(led->trigger, | ||
74 | struct rfkill, led_trigger); | ||
75 | |||
76 | rfkill_led_trigger(rfkill, rfkill->state); | ||
77 | } | ||
78 | #else | ||
79 | static inline void rfkill_led_trigger(struct rfkill *rfkill, | ||
80 | enum rfkill_state state) | ||
81 | { | ||
82 | } | ||
83 | #endif /* CONFIG_RFKILL_LEDS */ | ||
84 | |||
85 | static void rfkill_uevent(struct rfkill *rfkill) | ||
86 | { | ||
87 | kobject_uevent(&rfkill->dev.kobj, KOBJ_CHANGE); | ||
88 | } | ||
89 | |||
90 | static void update_rfkill_state(struct rfkill *rfkill) | ||
91 | { | ||
92 | enum rfkill_state newstate, oldstate; | ||
93 | |||
94 | if (rfkill->get_state) { | ||
95 | mutex_lock(&rfkill->mutex); | ||
96 | if (!rfkill->get_state(rfkill->data, &newstate)) { | ||
97 | oldstate = rfkill->state; | ||
98 | rfkill->state = newstate; | ||
99 | if (oldstate != newstate) | ||
100 | rfkill_uevent(rfkill); | ||
101 | } | ||
102 | mutex_unlock(&rfkill->mutex); | ||
103 | } | ||
104 | rfkill_led_trigger(rfkill, rfkill->state); | ||
105 | } | ||
106 | |||
107 | /** | ||
108 | * rfkill_toggle_radio - wrapper for toggle_radio hook | ||
109 | * @rfkill: the rfkill struct to use | ||
110 | * @force: calls toggle_radio even if cache says it is not needed, | ||
111 | * and also makes sure notifications of the state will be | ||
112 | * sent even if it didn't change | ||
113 | * @state: the new state to call toggle_radio() with | ||
114 | * | ||
115 | * Calls rfkill->toggle_radio, enforcing the API for toggle_radio | ||
116 | * calls and handling all the red tape such as issuing notifications | ||
117 | * if the call is successful. | ||
118 | * | ||
119 | * Suspended devices are not touched at all, and -EAGAIN is returned. | ||
120 | * | ||
121 | * Note that the @force parameter cannot override a (possibly cached) | ||
122 | * state of RFKILL_STATE_HARD_BLOCKED. Any device making use of | ||
123 | * RFKILL_STATE_HARD_BLOCKED implements either get_state() or | ||
124 | * rfkill_force_state(), so the cache either is bypassed or valid. | ||
125 | * | ||
126 | * Note that we do call toggle_radio for RFKILL_STATE_SOFT_BLOCKED | ||
127 | * even if the radio is in RFKILL_STATE_HARD_BLOCKED state, so as to | ||
128 | * give the driver a hint that it should double-BLOCK the transmitter. | ||
129 | * | ||
130 | * Caller must have acquired rfkill->mutex. | ||
131 | */ | ||
132 | static int rfkill_toggle_radio(struct rfkill *rfkill, | ||
133 | enum rfkill_state state, | ||
134 | int force) | ||
135 | { | ||
136 | int retval = 0; | ||
137 | enum rfkill_state oldstate, newstate; | ||
138 | |||
139 | if (unlikely(rfkill->dev.power.power_state.event & PM_EVENT_SLEEP)) | ||
140 | return -EBUSY; | ||
141 | |||
142 | oldstate = rfkill->state; | ||
143 | |||
144 | if (rfkill->get_state && !force && | ||
145 | !rfkill->get_state(rfkill->data, &newstate)) { | ||
146 | rfkill->state = newstate; | ||
147 | } | ||
148 | |||
149 | switch (state) { | ||
150 | case RFKILL_STATE_HARD_BLOCKED: | ||
151 | /* typically happens when refreshing hardware state, | ||
152 | * such as on resume */ | ||
153 | state = RFKILL_STATE_SOFT_BLOCKED; | ||
154 | break; | ||
155 | case RFKILL_STATE_UNBLOCKED: | ||
156 | /* force can't override this, only rfkill_force_state() can */ | ||
157 | if (rfkill->state == RFKILL_STATE_HARD_BLOCKED) | ||
158 | return -EPERM; | ||
159 | break; | ||
160 | case RFKILL_STATE_SOFT_BLOCKED: | ||
161 | /* nothing to do, we want to give drivers the hint to double | ||
162 | * BLOCK even a transmitter that is already in state | ||
163 | * RFKILL_STATE_HARD_BLOCKED */ | ||
164 | break; | ||
165 | default: | ||
166 | WARN(1, KERN_WARNING | ||
167 | "rfkill: illegal state %d passed as parameter " | ||
168 | "to rfkill_toggle_radio\n", state); | ||
169 | return -EINVAL; | ||
170 | } | ||
171 | |||
172 | if (force || state != rfkill->state) { | ||
173 | retval = rfkill->toggle_radio(rfkill->data, state); | ||
174 | /* never allow a HARD->SOFT downgrade! */ | ||
175 | if (!retval && rfkill->state != RFKILL_STATE_HARD_BLOCKED) | ||
176 | rfkill->state = state; | ||
177 | } | ||
178 | |||
179 | if (force || rfkill->state != oldstate) | ||
180 | rfkill_uevent(rfkill); | ||
181 | |||
182 | rfkill_led_trigger(rfkill, rfkill->state); | ||
183 | return retval; | ||
184 | } | ||
185 | |||
186 | /** | ||
187 | * __rfkill_switch_all - Toggle state of all switches of given type | ||
188 | * @type: type of interfaces to be affected | ||
189 | * @state: the new state | ||
190 | * | ||
191 | * This function toggles the state of all switches of given type, | ||
192 | * unless a specific switch is claimed by userspace (in which case, | ||
193 | * that switch is left alone) or suspended. | ||
194 | * | ||
195 | * Caller must have acquired rfkill_global_mutex. | ||
196 | */ | ||
197 | static void __rfkill_switch_all(const enum rfkill_type type, | ||
198 | const enum rfkill_state state) | ||
199 | { | ||
200 | struct rfkill *rfkill; | ||
201 | |||
202 | if (WARN((state >= RFKILL_STATE_MAX || type >= RFKILL_TYPE_MAX), | ||
203 | KERN_WARNING | ||
204 | "rfkill: illegal state %d or type %d " | ||
205 | "passed as parameter to __rfkill_switch_all\n", | ||
206 | state, type)) | ||
207 | return; | ||
208 | |||
209 | rfkill_global_states[type].current_state = state; | ||
210 | list_for_each_entry(rfkill, &rfkill_list, node) { | ||
211 | if (rfkill->type == type) { | ||
212 | mutex_lock(&rfkill->mutex); | ||
213 | rfkill_toggle_radio(rfkill, state, 0); | ||
214 | mutex_unlock(&rfkill->mutex); | ||
215 | rfkill_led_trigger(rfkill, rfkill->state); | ||
216 | } | ||
217 | } | ||
218 | } | ||
219 | |||
220 | /** | ||
221 | * rfkill_switch_all - Toggle state of all switches of given type | ||
222 | * @type: type of interfaces to be affected | ||
223 | * @state: the new state | ||
224 | * | ||
225 | * Acquires rfkill_global_mutex and calls __rfkill_switch_all(@type, @state). | ||
226 | * Please refer to __rfkill_switch_all() for details. | ||
227 | * | ||
228 | * Does nothing if the EPO lock is active. | ||
229 | */ | ||
230 | void rfkill_switch_all(enum rfkill_type type, enum rfkill_state state) | ||
231 | { | ||
232 | mutex_lock(&rfkill_global_mutex); | ||
233 | if (!rfkill_epo_lock_active) | ||
234 | __rfkill_switch_all(type, state); | ||
235 | mutex_unlock(&rfkill_global_mutex); | ||
236 | } | ||
237 | EXPORT_SYMBOL(rfkill_switch_all); | ||
238 | |||
239 | /** | ||
240 | * rfkill_epo - emergency power off all transmitters | ||
241 | * | ||
242 | * This kicks all non-suspended rfkill devices to RFKILL_STATE_SOFT_BLOCKED, | ||
243 | * ignoring everything in its path but rfkill_global_mutex and rfkill->mutex. | ||
244 | * | ||
245 | * The global state before the EPO is saved and can be restored later | ||
246 | * using rfkill_restore_states(). | ||
247 | */ | ||
248 | void rfkill_epo(void) | ||
249 | { | ||
250 | struct rfkill *rfkill; | ||
251 | int i; | ||
252 | |||
253 | mutex_lock(&rfkill_global_mutex); | ||
254 | |||
255 | rfkill_epo_lock_active = true; | ||
256 | list_for_each_entry(rfkill, &rfkill_list, node) { | ||
257 | mutex_lock(&rfkill->mutex); | ||
258 | rfkill_toggle_radio(rfkill, RFKILL_STATE_SOFT_BLOCKED, 1); | ||
259 | mutex_unlock(&rfkill->mutex); | ||
260 | } | ||
261 | for (i = 0; i < RFKILL_TYPE_MAX; i++) { | ||
262 | rfkill_global_states[i].default_state = | ||
263 | rfkill_global_states[i].current_state; | ||
264 | rfkill_global_states[i].current_state = | ||
265 | RFKILL_STATE_SOFT_BLOCKED; | ||
266 | } | ||
267 | mutex_unlock(&rfkill_global_mutex); | ||
268 | rfkill_led_trigger(rfkill, rfkill->state); | ||
269 | } | ||
270 | EXPORT_SYMBOL_GPL(rfkill_epo); | ||
271 | |||
272 | /** | ||
273 | * rfkill_restore_states - restore global states | ||
274 | * | ||
275 | * Restore (and sync switches to) the global state from the | ||
276 | * states in rfkill_default_states. This can undo the effects of | ||
277 | * a call to rfkill_epo(). | ||
278 | */ | ||
279 | void rfkill_restore_states(void) | ||
280 | { | ||
281 | int i; | ||
282 | |||
283 | mutex_lock(&rfkill_global_mutex); | ||
284 | |||
285 | rfkill_epo_lock_active = false; | ||
286 | for (i = 0; i < RFKILL_TYPE_MAX; i++) | ||
287 | __rfkill_switch_all(i, rfkill_global_states[i].default_state); | ||
288 | mutex_unlock(&rfkill_global_mutex); | ||
289 | } | ||
290 | EXPORT_SYMBOL_GPL(rfkill_restore_states); | ||
291 | |||
292 | /** | ||
293 | * rfkill_remove_epo_lock - unlock state changes | ||
294 | * | ||
295 | * Used by rfkill-input manually unlock state changes, when | ||
296 | * the EPO switch is deactivated. | ||
297 | */ | ||
298 | void rfkill_remove_epo_lock(void) | ||
299 | { | ||
300 | mutex_lock(&rfkill_global_mutex); | ||
301 | rfkill_epo_lock_active = false; | ||
302 | mutex_unlock(&rfkill_global_mutex); | ||
303 | } | ||
304 | EXPORT_SYMBOL_GPL(rfkill_remove_epo_lock); | ||
305 | |||
306 | /** | ||
307 | * rfkill_is_epo_lock_active - returns true EPO is active | ||
308 | * | ||
309 | * Returns 0 (false) if there is NOT an active EPO contidion, | ||
310 | * and 1 (true) if there is an active EPO contition, which | ||
311 | * locks all radios in one of the BLOCKED states. | ||
312 | * | ||
313 | * Can be called in atomic context. | ||
314 | */ | ||
315 | bool rfkill_is_epo_lock_active(void) | ||
316 | { | ||
317 | return rfkill_epo_lock_active; | ||
318 | } | ||
319 | EXPORT_SYMBOL_GPL(rfkill_is_epo_lock_active); | ||
320 | |||
321 | /** | ||
322 | * rfkill_get_global_state - returns global state for a type | ||
323 | * @type: the type to get the global state of | ||
324 | * | ||
325 | * Returns the current global state for a given wireless | ||
326 | * device type. | ||
327 | */ | ||
328 | enum rfkill_state rfkill_get_global_state(const enum rfkill_type type) | ||
329 | { | ||
330 | return rfkill_global_states[type].current_state; | ||
331 | } | ||
332 | EXPORT_SYMBOL_GPL(rfkill_get_global_state); | ||
333 | |||
334 | /** | ||
335 | * rfkill_force_state - Force the internal rfkill radio state | ||
336 | * @rfkill: pointer to the rfkill class to modify. | ||
337 | * @state: the current radio state the class should be forced to. | ||
338 | * | ||
339 | * This function updates the internal state of the radio cached | ||
340 | * by the rfkill class. It should be used when the driver gets | ||
341 | * a notification by the firmware/hardware of the current *real* | ||
342 | * state of the radio rfkill switch. | ||
343 | * | ||
344 | * Devices which are subject to external changes on their rfkill | ||
345 | * state (such as those caused by a hardware rfkill line) MUST | ||
346 | * have their driver arrange to call rfkill_force_state() as soon | ||
347 | * as possible after such a change. | ||
348 | * | ||
349 | * This function may not be called from an atomic context. | ||
350 | */ | ||
351 | int rfkill_force_state(struct rfkill *rfkill, enum rfkill_state state) | ||
352 | { | ||
353 | enum rfkill_state oldstate; | ||
354 | |||
355 | BUG_ON(!rfkill); | ||
356 | if (WARN((state >= RFKILL_STATE_MAX), | ||
357 | KERN_WARNING | ||
358 | "rfkill: illegal state %d passed as parameter " | ||
359 | "to rfkill_force_state\n", state)) | ||
360 | return -EINVAL; | ||
361 | |||
362 | mutex_lock(&rfkill->mutex); | ||
363 | |||
364 | oldstate = rfkill->state; | ||
365 | rfkill->state = state; | ||
366 | |||
367 | if (state != oldstate) | ||
368 | rfkill_uevent(rfkill); | ||
369 | |||
370 | mutex_unlock(&rfkill->mutex); | ||
371 | rfkill_led_trigger(rfkill, rfkill->state); | ||
372 | |||
373 | return 0; | ||
374 | } | ||
375 | EXPORT_SYMBOL(rfkill_force_state); | ||
376 | |||
377 | static ssize_t rfkill_name_show(struct device *dev, | ||
378 | struct device_attribute *attr, | ||
379 | char *buf) | ||
380 | { | ||
381 | struct rfkill *rfkill = to_rfkill(dev); | ||
382 | |||
383 | return sprintf(buf, "%s\n", rfkill->name); | ||
384 | } | ||
385 | |||
386 | static const char *rfkill_get_type_str(enum rfkill_type type) | ||
387 | { | ||
388 | switch (type) { | ||
389 | case RFKILL_TYPE_WLAN: | ||
390 | return "wlan"; | ||
391 | case RFKILL_TYPE_BLUETOOTH: | ||
392 | return "bluetooth"; | ||
393 | case RFKILL_TYPE_UWB: | ||
394 | return "ultrawideband"; | ||
395 | case RFKILL_TYPE_WIMAX: | ||
396 | return "wimax"; | ||
397 | case RFKILL_TYPE_WWAN: | ||
398 | return "wwan"; | ||
399 | default: | ||
400 | BUG(); | ||
401 | } | ||
402 | } | ||
403 | |||
404 | static ssize_t rfkill_type_show(struct device *dev, | ||
405 | struct device_attribute *attr, | ||
406 | char *buf) | ||
407 | { | ||
408 | struct rfkill *rfkill = to_rfkill(dev); | ||
409 | |||
410 | return sprintf(buf, "%s\n", rfkill_get_type_str(rfkill->type)); | ||
411 | } | ||
412 | |||
413 | static ssize_t rfkill_state_show(struct device *dev, | ||
414 | struct device_attribute *attr, | ||
415 | char *buf) | ||
416 | { | ||
417 | struct rfkill *rfkill = to_rfkill(dev); | ||
418 | |||
419 | update_rfkill_state(rfkill); | ||
420 | return sprintf(buf, "%d\n", rfkill->state); | ||
421 | } | ||
422 | |||
423 | static ssize_t rfkill_state_store(struct device *dev, | ||
424 | struct device_attribute *attr, | ||
425 | const char *buf, size_t count) | ||
426 | { | ||
427 | struct rfkill *rfkill = to_rfkill(dev); | ||
428 | unsigned long state; | ||
429 | int error; | ||
430 | |||
431 | if (!capable(CAP_NET_ADMIN)) | ||
432 | return -EPERM; | ||
433 | |||
434 | error = strict_strtoul(buf, 0, &state); | ||
435 | if (error) | ||
436 | return error; | ||
437 | |||
438 | /* RFKILL_STATE_HARD_BLOCKED is illegal here... */ | ||
439 | if (state != RFKILL_STATE_UNBLOCKED && | ||
440 | state != RFKILL_STATE_SOFT_BLOCKED) | ||
441 | return -EINVAL; | ||
442 | |||
443 | error = mutex_lock_killable(&rfkill->mutex); | ||
444 | if (error) | ||
445 | return error; | ||
446 | |||
447 | if (!rfkill_epo_lock_active) | ||
448 | error = rfkill_toggle_radio(rfkill, state, 0); | ||
449 | else | ||
450 | error = -EPERM; | ||
451 | |||
452 | mutex_unlock(&rfkill->mutex); | ||
453 | |||
454 | return error ? error : count; | ||
455 | } | ||
456 | |||
457 | static ssize_t rfkill_claim_show(struct device *dev, | ||
458 | struct device_attribute *attr, | ||
459 | char *buf) | ||
460 | { | ||
461 | return sprintf(buf, "%d\n", 0); | ||
462 | } | ||
463 | |||
464 | static ssize_t rfkill_claim_store(struct device *dev, | ||
465 | struct device_attribute *attr, | ||
466 | const char *buf, size_t count) | ||
467 | { | ||
468 | return -EOPNOTSUPP; | ||
469 | } | ||
470 | |||
471 | static struct device_attribute rfkill_dev_attrs[] = { | ||
472 | __ATTR(name, S_IRUGO, rfkill_name_show, NULL), | ||
473 | __ATTR(type, S_IRUGO, rfkill_type_show, NULL), | ||
474 | __ATTR(state, S_IRUGO|S_IWUSR, rfkill_state_show, rfkill_state_store), | ||
475 | __ATTR(claim, S_IRUGO|S_IWUSR, rfkill_claim_show, rfkill_claim_store), | ||
476 | __ATTR_NULL | ||
477 | }; | ||
478 | |||
479 | static void rfkill_release(struct device *dev) | ||
480 | { | ||
481 | struct rfkill *rfkill = to_rfkill(dev); | ||
482 | |||
483 | kfree(rfkill); | ||
484 | module_put(THIS_MODULE); | ||
485 | } | ||
486 | |||
487 | #ifdef CONFIG_PM | ||
488 | static int rfkill_suspend(struct device *dev, pm_message_t state) | ||
489 | { | ||
490 | struct rfkill *rfkill = to_rfkill(dev); | ||
491 | |||
492 | /* mark class device as suspended */ | ||
493 | if (dev->power.power_state.event != state.event) | ||
494 | dev->power.power_state = state; | ||
495 | |||
496 | /* store state for the resume handler */ | ||
497 | rfkill->state_for_resume = rfkill->state; | ||
498 | |||
499 | return 0; | ||
500 | } | ||
501 | |||
502 | static int rfkill_resume(struct device *dev) | ||
503 | { | ||
504 | struct rfkill *rfkill = to_rfkill(dev); | ||
505 | enum rfkill_state newstate; | ||
506 | |||
507 | if (dev->power.power_state.event != PM_EVENT_ON) { | ||
508 | mutex_lock(&rfkill->mutex); | ||
509 | |||
510 | dev->power.power_state.event = PM_EVENT_ON; | ||
511 | |||
512 | /* | ||
513 | * rfkill->state could have been modified before we got | ||
514 | * called, and won't be updated by rfkill_toggle_radio() | ||
515 | * in force mode. Sync it FIRST. | ||
516 | */ | ||
517 | if (rfkill->get_state && | ||
518 | !rfkill->get_state(rfkill->data, &newstate)) | ||
519 | rfkill->state = newstate; | ||
520 | |||
521 | /* | ||
522 | * If we are under EPO, kick transmitter offline, | ||
523 | * otherwise restore to pre-suspend state. | ||
524 | * | ||
525 | * Issue a notification in any case | ||
526 | */ | ||
527 | rfkill_toggle_radio(rfkill, | ||
528 | rfkill_epo_lock_active ? | ||
529 | RFKILL_STATE_SOFT_BLOCKED : | ||
530 | rfkill->state_for_resume, | ||
531 | 1); | ||
532 | |||
533 | mutex_unlock(&rfkill->mutex); | ||
534 | rfkill_led_trigger(rfkill, rfkill->state); | ||
535 | } | ||
536 | |||
537 | return 0; | ||
538 | } | ||
539 | #else | ||
540 | #define rfkill_suspend NULL | ||
541 | #define rfkill_resume NULL | ||
542 | #endif | ||
543 | |||
544 | static int rfkill_dev_uevent(struct device *dev, struct kobj_uevent_env *env) | ||
545 | { | ||
546 | struct rfkill *rfkill = to_rfkill(dev); | ||
547 | int error; | ||
548 | |||
549 | error = add_uevent_var(env, "RFKILL_NAME=%s", rfkill->name); | ||
550 | if (error) | ||
551 | return error; | ||
552 | error = add_uevent_var(env, "RFKILL_TYPE=%s", | ||
553 | rfkill_get_type_str(rfkill->type)); | ||
554 | if (error) | ||
555 | return error; | ||
556 | error = add_uevent_var(env, "RFKILL_STATE=%d", rfkill->state); | ||
557 | return error; | ||
558 | } | ||
559 | |||
560 | static struct class rfkill_class = { | ||
561 | .name = "rfkill", | ||
562 | .dev_release = rfkill_release, | ||
563 | .dev_attrs = rfkill_dev_attrs, | ||
564 | .suspend = rfkill_suspend, | ||
565 | .resume = rfkill_resume, | ||
566 | .dev_uevent = rfkill_dev_uevent, | ||
567 | }; | ||
568 | |||
569 | static int rfkill_check_duplicity(const struct rfkill *rfkill) | ||
570 | { | ||
571 | struct rfkill *p; | ||
572 | unsigned long seen[BITS_TO_LONGS(RFKILL_TYPE_MAX)]; | ||
573 | |||
574 | memset(seen, 0, sizeof(seen)); | ||
575 | |||
576 | list_for_each_entry(p, &rfkill_list, node) { | ||
577 | if (WARN((p == rfkill), KERN_WARNING | ||
578 | "rfkill: illegal attempt to register " | ||
579 | "an already registered rfkill struct\n")) | ||
580 | return -EEXIST; | ||
581 | set_bit(p->type, seen); | ||
582 | } | ||
583 | |||
584 | /* 0: first switch of its kind */ | ||
585 | return (test_bit(rfkill->type, seen)) ? 1 : 0; | ||
586 | } | ||
587 | |||
588 | static int rfkill_add_switch(struct rfkill *rfkill) | ||
589 | { | ||
590 | int error; | ||
591 | |||
592 | mutex_lock(&rfkill_global_mutex); | ||
593 | |||
594 | error = rfkill_check_duplicity(rfkill); | ||
595 | if (error < 0) | ||
596 | goto unlock_out; | ||
597 | |||
598 | if (!error) { | ||
599 | /* lock default after first use */ | ||
600 | set_bit(rfkill->type, rfkill_states_lockdflt); | ||
601 | rfkill_global_states[rfkill->type].current_state = | ||
602 | rfkill_global_states[rfkill->type].default_state; | ||
603 | } | ||
604 | |||
605 | rfkill_toggle_radio(rfkill, | ||
606 | rfkill_global_states[rfkill->type].current_state, | ||
607 | 0); | ||
608 | |||
609 | list_add_tail(&rfkill->node, &rfkill_list); | ||
610 | |||
611 | error = 0; | ||
612 | unlock_out: | ||
613 | mutex_unlock(&rfkill_global_mutex); | ||
614 | |||
615 | return error; | ||
616 | } | ||
617 | |||
618 | static void rfkill_remove_switch(struct rfkill *rfkill) | ||
619 | { | ||
620 | mutex_lock(&rfkill_global_mutex); | ||
621 | list_del_init(&rfkill->node); | ||
622 | mutex_unlock(&rfkill_global_mutex); | ||
623 | |||
624 | mutex_lock(&rfkill->mutex); | ||
625 | rfkill_toggle_radio(rfkill, RFKILL_STATE_SOFT_BLOCKED, 1); | ||
626 | mutex_unlock(&rfkill->mutex); | ||
627 | } | ||
628 | |||
629 | /** | ||
630 | * rfkill_allocate - allocate memory for rfkill structure. | ||
631 | * @parent: device that has rf switch on it | ||
632 | * @type: type of the switch (RFKILL_TYPE_*) | ||
633 | * | ||
634 | * This function should be called by the network driver when it needs | ||
635 | * rfkill structure. Once the structure is allocated the driver should | ||
636 | * finish its initialization by setting the name, private data, enable_radio | ||
637 | * and disable_radio methods and then register it with rfkill_register(). | ||
638 | * | ||
639 | * NOTE: If registration fails the structure shoudl be freed by calling | ||
640 | * rfkill_free() otherwise rfkill_unregister() should be used. | ||
641 | */ | ||
642 | struct rfkill * __must_check rfkill_allocate(struct device *parent, | ||
643 | enum rfkill_type type) | ||
644 | { | ||
645 | struct rfkill *rfkill; | ||
646 | struct device *dev; | ||
647 | |||
648 | if (WARN((type >= RFKILL_TYPE_MAX), | ||
649 | KERN_WARNING | ||
650 | "rfkill: illegal type %d passed as parameter " | ||
651 | "to rfkill_allocate\n", type)) | ||
652 | return NULL; | ||
653 | |||
654 | rfkill = kzalloc(sizeof(struct rfkill), GFP_KERNEL); | ||
655 | if (!rfkill) | ||
656 | return NULL; | ||
657 | |||
658 | mutex_init(&rfkill->mutex); | ||
659 | INIT_LIST_HEAD(&rfkill->node); | ||
660 | rfkill->type = type; | ||
661 | |||
662 | dev = &rfkill->dev; | ||
663 | dev->class = &rfkill_class; | ||
664 | dev->parent = parent; | ||
665 | device_initialize(dev); | ||
666 | |||
667 | __module_get(THIS_MODULE); | ||
668 | |||
669 | return rfkill; | ||
670 | } | ||
671 | EXPORT_SYMBOL(rfkill_allocate); | ||
672 | |||
673 | /** | ||
674 | * rfkill_free - Mark rfkill structure for deletion | ||
675 | * @rfkill: rfkill structure to be destroyed | ||
676 | * | ||
677 | * Decrements reference count of the rfkill structure so it is destroyed. | ||
678 | * Note that rfkill_free() should _not_ be called after rfkill_unregister(). | ||
679 | */ | ||
680 | void rfkill_free(struct rfkill *rfkill) | ||
681 | { | ||
682 | if (rfkill) | ||
683 | put_device(&rfkill->dev); | ||
684 | } | ||
685 | EXPORT_SYMBOL(rfkill_free); | ||
686 | |||
687 | static void rfkill_led_trigger_register(struct rfkill *rfkill) | ||
688 | { | ||
689 | #ifdef CONFIG_RFKILL_LEDS | ||
690 | int error; | ||
691 | |||
692 | if (!rfkill->led_trigger.name) | ||
693 | rfkill->led_trigger.name = dev_name(&rfkill->dev); | ||
694 | if (!rfkill->led_trigger.activate) | ||
695 | rfkill->led_trigger.activate = rfkill_led_trigger_activate; | ||
696 | error = led_trigger_register(&rfkill->led_trigger); | ||
697 | if (error) | ||
698 | rfkill->led_trigger.name = NULL; | ||
699 | #endif /* CONFIG_RFKILL_LEDS */ | ||
700 | } | ||
701 | |||
702 | static void rfkill_led_trigger_unregister(struct rfkill *rfkill) | ||
703 | { | ||
704 | #ifdef CONFIG_RFKILL_LEDS | ||
705 | if (rfkill->led_trigger.name) { | ||
706 | led_trigger_unregister(&rfkill->led_trigger); | ||
707 | rfkill->led_trigger.name = NULL; | ||
708 | } | ||
709 | #endif | ||
710 | } | ||
711 | |||
712 | /** | ||
713 | * rfkill_register - Register a rfkill structure. | ||
714 | * @rfkill: rfkill structure to be registered | ||
715 | * | ||
716 | * This function should be called by the network driver when the rfkill | ||
717 | * structure needs to be registered. Immediately from registration the | ||
718 | * switch driver should be able to service calls to toggle_radio. | ||
719 | */ | ||
720 | int __must_check rfkill_register(struct rfkill *rfkill) | ||
721 | { | ||
722 | static atomic_t rfkill_no = ATOMIC_INIT(0); | ||
723 | struct device *dev = &rfkill->dev; | ||
724 | int error; | ||
725 | |||
726 | if (WARN((!rfkill || !rfkill->toggle_radio || | ||
727 | rfkill->type >= RFKILL_TYPE_MAX || | ||
728 | rfkill->state >= RFKILL_STATE_MAX), | ||
729 | KERN_WARNING | ||
730 | "rfkill: attempt to register a " | ||
731 | "badly initialized rfkill struct\n")) | ||
732 | return -EINVAL; | ||
733 | |||
734 | dev_set_name(dev, "rfkill%ld", (long)atomic_inc_return(&rfkill_no) - 1); | ||
735 | |||
736 | rfkill_led_trigger_register(rfkill); | ||
737 | |||
738 | error = rfkill_add_switch(rfkill); | ||
739 | if (error) { | ||
740 | rfkill_led_trigger_unregister(rfkill); | ||
741 | return error; | ||
742 | } | ||
743 | |||
744 | error = device_add(dev); | ||
745 | if (error) { | ||
746 | rfkill_remove_switch(rfkill); | ||
747 | rfkill_led_trigger_unregister(rfkill); | ||
748 | return error; | ||
749 | } | ||
750 | |||
751 | return 0; | ||
752 | } | ||
753 | EXPORT_SYMBOL(rfkill_register); | ||
754 | |||
755 | /** | ||
756 | * rfkill_unregister - Unregister a rfkill structure. | ||
757 | * @rfkill: rfkill structure to be unregistered | ||
758 | * | ||
759 | * This function should be called by the network driver during device | ||
760 | * teardown to destroy rfkill structure. Note that rfkill_free() should | ||
761 | * _not_ be called after rfkill_unregister(). | ||
762 | */ | ||
763 | void rfkill_unregister(struct rfkill *rfkill) | ||
764 | { | ||
765 | BUG_ON(!rfkill); | ||
766 | device_del(&rfkill->dev); | ||
767 | rfkill_remove_switch(rfkill); | ||
768 | rfkill_led_trigger_unregister(rfkill); | ||
769 | put_device(&rfkill->dev); | ||
770 | } | ||
771 | EXPORT_SYMBOL(rfkill_unregister); | ||
772 | |||
773 | /** | ||
774 | * rfkill_set_default - set initial value for a switch type | ||
775 | * @type - the type of switch to set the default state of | ||
776 | * @state - the new default state for that group of switches | ||
777 | * | ||
778 | * Sets the initial state rfkill should use for a given type. | ||
779 | * The following initial states are allowed: RFKILL_STATE_SOFT_BLOCKED | ||
780 | * and RFKILL_STATE_UNBLOCKED. | ||
781 | * | ||
782 | * This function is meant to be used by platform drivers for platforms | ||
783 | * that can save switch state across power down/reboot. | ||
784 | * | ||
785 | * The default state for each switch type can be changed exactly once. | ||
786 | * After a switch of that type is registered, the default state cannot | ||
787 | * be changed anymore. This guards against multiple drivers it the | ||
788 | * same platform trying to set the initial switch default state, which | ||
789 | * is not allowed. | ||
790 | * | ||
791 | * Returns -EPERM if the state has already been set once or is in use, | ||
792 | * so drivers likely want to either ignore or at most printk(KERN_NOTICE) | ||
793 | * if this function returns -EPERM. | ||
794 | * | ||
795 | * Returns 0 if the new default state was set, or an error if it | ||
796 | * could not be set. | ||
797 | */ | ||
798 | int rfkill_set_default(enum rfkill_type type, enum rfkill_state state) | ||
799 | { | ||
800 | int error; | ||
801 | |||
802 | if (WARN((type >= RFKILL_TYPE_MAX || | ||
803 | (state != RFKILL_STATE_SOFT_BLOCKED && | ||
804 | state != RFKILL_STATE_UNBLOCKED)), | ||
805 | KERN_WARNING | ||
806 | "rfkill: illegal state %d or type %d passed as " | ||
807 | "parameter to rfkill_set_default\n", state, type)) | ||
808 | return -EINVAL; | ||
809 | |||
810 | mutex_lock(&rfkill_global_mutex); | ||
811 | |||
812 | if (!test_and_set_bit(type, rfkill_states_lockdflt)) { | ||
813 | rfkill_global_states[type].default_state = state; | ||
814 | rfkill_global_states[type].current_state = state; | ||
815 | error = 0; | ||
816 | } else | ||
817 | error = -EPERM; | ||
818 | |||
819 | mutex_unlock(&rfkill_global_mutex); | ||
820 | return error; | ||
821 | } | ||
822 | EXPORT_SYMBOL_GPL(rfkill_set_default); | ||
823 | |||
824 | /* | ||
825 | * Rfkill module initialization/deinitialization. | ||
826 | */ | ||
827 | static int __init rfkill_init(void) | ||
828 | { | ||
829 | int error; | ||
830 | int i; | ||
831 | |||
832 | /* RFKILL_STATE_HARD_BLOCKED is illegal here... */ | ||
833 | if (rfkill_default_state != RFKILL_STATE_SOFT_BLOCKED && | ||
834 | rfkill_default_state != RFKILL_STATE_UNBLOCKED) | ||
835 | return -EINVAL; | ||
836 | |||
837 | for (i = 0; i < RFKILL_TYPE_MAX; i++) | ||
838 | rfkill_global_states[i].default_state = rfkill_default_state; | ||
839 | |||
840 | error = class_register(&rfkill_class); | ||
841 | if (error) { | ||
842 | printk(KERN_ERR "rfkill: unable to register rfkill class\n"); | ||
843 | return error; | ||
844 | } | ||
845 | |||
846 | return 0; | ||
847 | } | ||
848 | |||
849 | static void __exit rfkill_exit(void) | ||
850 | { | ||
851 | class_unregister(&rfkill_class); | ||
852 | } | ||
853 | |||
854 | subsys_initcall(rfkill_init); | ||
855 | module_exit(rfkill_exit); | ||
diff --git a/net/rfkill/rfkill-input.h b/net/rfkill/rfkill.h index fe8df6b5b935..d1117cb6e4de 100644 --- a/net/rfkill/rfkill-input.h +++ b/net/rfkill/rfkill.h | |||
@@ -1,5 +1,6 @@ | |||
1 | /* | 1 | /* |
2 | * Copyright (C) 2007 Ivo van Doorn | 2 | * Copyright (C) 2007 Ivo van Doorn |
3 | * Copyright 2009 Johannes Berg <johannes@sipsolutions.net> | ||
3 | */ | 4 | */ |
4 | 5 | ||
5 | /* | 6 | /* |
@@ -11,11 +12,16 @@ | |||
11 | #ifndef __RFKILL_INPUT_H | 12 | #ifndef __RFKILL_INPUT_H |
12 | #define __RFKILL_INPUT_H | 13 | #define __RFKILL_INPUT_H |
13 | 14 | ||
14 | void rfkill_switch_all(enum rfkill_type type, enum rfkill_state state); | 15 | /* core code */ |
16 | void rfkill_switch_all(const enum rfkill_type type, bool blocked); | ||
15 | void rfkill_epo(void); | 17 | void rfkill_epo(void); |
16 | void rfkill_restore_states(void); | 18 | void rfkill_restore_states(void); |
17 | void rfkill_remove_epo_lock(void); | 19 | void rfkill_remove_epo_lock(void); |
18 | bool rfkill_is_epo_lock_active(void); | 20 | bool rfkill_is_epo_lock_active(void); |
19 | enum rfkill_state rfkill_get_global_state(const enum rfkill_type type); | 21 | bool rfkill_get_global_sw_state(const enum rfkill_type type); |
22 | |||
23 | /* input handler */ | ||
24 | int rfkill_handler_init(void); | ||
25 | void rfkill_handler_exit(void); | ||
20 | 26 | ||
21 | #endif /* __RFKILL_INPUT_H */ | 27 | #endif /* __RFKILL_INPUT_H */ |
diff --git a/net/wimax/Kconfig b/net/wimax/Kconfig index 1b46747a5f5a..0bdbb6928205 100644 --- a/net/wimax/Kconfig +++ b/net/wimax/Kconfig | |||
@@ -1,23 +1,9 @@ | |||
1 | # | 1 | # |
2 | # WiMAX LAN device configuration | 2 | # WiMAX LAN device configuration |
3 | # | 3 | # |
4 | # Note the ugly 'depends on' on WIMAX: that disallows RFKILL to be a | ||
5 | # module if WIMAX is to be linked in. The WiMAX code is done in such a | ||
6 | # way that it doesn't require and explicit dependency on RFKILL in | ||
7 | # case an embedded system wants to rip it out. | ||
8 | # | ||
9 | # As well, enablement of the RFKILL code means we need the INPUT layer | ||
10 | # support to inject events coming from hw rfkill switches. That | ||
11 | # dependency could be killed if input.h provided appropriate means to | ||
12 | # work when input is disabled. | ||
13 | |||
14 | comment "WiMAX Wireless Broadband support requires CONFIG_INPUT enabled" | ||
15 | depends on INPUT = n && RFKILL != n | ||
16 | 4 | ||
17 | menuconfig WIMAX | 5 | menuconfig WIMAX |
18 | tristate "WiMAX Wireless Broadband support" | 6 | tristate "WiMAX Wireless Broadband support" |
19 | depends on (y && RFKILL != m) || m | ||
20 | depends on (INPUT && RFKILL != n) || RFKILL = n | ||
21 | help | 7 | help |
22 | 8 | ||
23 | Select to configure support for devices that provide | 9 | Select to configure support for devices that provide |
diff --git a/net/wimax/op-rfkill.c b/net/wimax/op-rfkill.c index a3616e2ccb8a..bb102e4aa3e9 100644 --- a/net/wimax/op-rfkill.c +++ b/net/wimax/op-rfkill.c | |||
@@ -29,8 +29,8 @@ | |||
29 | * A non-polled generic rfkill device is embedded into the WiMAX | 29 | * A non-polled generic rfkill device is embedded into the WiMAX |
30 | * subsystem's representation of a device. | 30 | * subsystem's representation of a device. |
31 | * | 31 | * |
32 | * FIXME: Need polled support? use a timer or add the implementation | 32 | * FIXME: Need polled support? Let drivers provide a poll routine |
33 | * to the stack. | 33 | * and hand it to rfkill ops then? |
34 | * | 34 | * |
35 | * All device drivers have to do is after wimax_dev_init(), call | 35 | * All device drivers have to do is after wimax_dev_init(), call |
36 | * wimax_report_rfkill_hw() and wimax_report_rfkill_sw() to update | 36 | * wimax_report_rfkill_hw() and wimax_report_rfkill_sw() to update |
@@ -43,7 +43,7 @@ | |||
43 | * wimax_rfkill() Kernel calling wimax_rfkill() | 43 | * wimax_rfkill() Kernel calling wimax_rfkill() |
44 | * __wimax_rf_toggle_radio() | 44 | * __wimax_rf_toggle_radio() |
45 | * | 45 | * |
46 | * wimax_rfkill_toggle_radio() RF-Kill subsytem calling | 46 | * wimax_rfkill_set_radio_block() RF-Kill subsytem calling |
47 | * __wimax_rf_toggle_radio() | 47 | * __wimax_rf_toggle_radio() |
48 | * | 48 | * |
49 | * __wimax_rf_toggle_radio() | 49 | * __wimax_rf_toggle_radio() |
@@ -65,15 +65,11 @@ | |||
65 | #include <linux/wimax.h> | 65 | #include <linux/wimax.h> |
66 | #include <linux/security.h> | 66 | #include <linux/security.h> |
67 | #include <linux/rfkill.h> | 67 | #include <linux/rfkill.h> |
68 | #include <linux/input.h> | ||
69 | #include "wimax-internal.h" | 68 | #include "wimax-internal.h" |
70 | 69 | ||
71 | #define D_SUBMODULE op_rfkill | 70 | #define D_SUBMODULE op_rfkill |
72 | #include "debug-levels.h" | 71 | #include "debug-levels.h" |
73 | 72 | ||
74 | #if defined(CONFIG_RFKILL) || defined(CONFIG_RFKILL_MODULE) | ||
75 | |||
76 | |||
77 | /** | 73 | /** |
78 | * wimax_report_rfkill_hw - Reports changes in the hardware RF switch | 74 | * wimax_report_rfkill_hw - Reports changes in the hardware RF switch |
79 | * | 75 | * |
@@ -99,7 +95,6 @@ void wimax_report_rfkill_hw(struct wimax_dev *wimax_dev, | |||
99 | int result; | 95 | int result; |
100 | struct device *dev = wimax_dev_to_dev(wimax_dev); | 96 | struct device *dev = wimax_dev_to_dev(wimax_dev); |
101 | enum wimax_st wimax_state; | 97 | enum wimax_st wimax_state; |
102 | enum rfkill_state rfkill_state; | ||
103 | 98 | ||
104 | d_fnstart(3, dev, "(wimax_dev %p state %u)\n", wimax_dev, state); | 99 | d_fnstart(3, dev, "(wimax_dev %p state %u)\n", wimax_dev, state); |
105 | BUG_ON(state == WIMAX_RF_QUERY); | 100 | BUG_ON(state == WIMAX_RF_QUERY); |
@@ -112,16 +107,15 @@ void wimax_report_rfkill_hw(struct wimax_dev *wimax_dev, | |||
112 | 107 | ||
113 | if (state != wimax_dev->rf_hw) { | 108 | if (state != wimax_dev->rf_hw) { |
114 | wimax_dev->rf_hw = state; | 109 | wimax_dev->rf_hw = state; |
115 | rfkill_state = state == WIMAX_RF_ON ? | ||
116 | RFKILL_STATE_UNBLOCKED : RFKILL_STATE_SOFT_BLOCKED; | ||
117 | if (wimax_dev->rf_hw == WIMAX_RF_ON | 110 | if (wimax_dev->rf_hw == WIMAX_RF_ON |
118 | && wimax_dev->rf_sw == WIMAX_RF_ON) | 111 | && wimax_dev->rf_sw == WIMAX_RF_ON) |
119 | wimax_state = WIMAX_ST_READY; | 112 | wimax_state = WIMAX_ST_READY; |
120 | else | 113 | else |
121 | wimax_state = WIMAX_ST_RADIO_OFF; | 114 | wimax_state = WIMAX_ST_RADIO_OFF; |
115 | |||
116 | rfkill_set_hw_state(wimax_dev->rfkill, state == WIMAX_RF_OFF); | ||
117 | |||
122 | __wimax_state_change(wimax_dev, wimax_state); | 118 | __wimax_state_change(wimax_dev, wimax_state); |
123 | input_report_key(wimax_dev->rfkill_input, KEY_WIMAX, | ||
124 | rfkill_state); | ||
125 | } | 119 | } |
126 | error_not_ready: | 120 | error_not_ready: |
127 | mutex_unlock(&wimax_dev->mutex); | 121 | mutex_unlock(&wimax_dev->mutex); |
@@ -174,6 +168,7 @@ void wimax_report_rfkill_sw(struct wimax_dev *wimax_dev, | |||
174 | else | 168 | else |
175 | wimax_state = WIMAX_ST_RADIO_OFF; | 169 | wimax_state = WIMAX_ST_RADIO_OFF; |
176 | __wimax_state_change(wimax_dev, wimax_state); | 170 | __wimax_state_change(wimax_dev, wimax_state); |
171 | rfkill_set_sw_state(wimax_dev->rfkill, state == WIMAX_RF_OFF); | ||
177 | } | 172 | } |
178 | error_not_ready: | 173 | error_not_ready: |
179 | mutex_unlock(&wimax_dev->mutex); | 174 | mutex_unlock(&wimax_dev->mutex); |
@@ -249,36 +244,31 @@ out_no_change: | |||
249 | * | 244 | * |
250 | * NOTE: This call will block until the operation is completed. | 245 | * NOTE: This call will block until the operation is completed. |
251 | */ | 246 | */ |
252 | static | 247 | static int wimax_rfkill_set_radio_block(void *data, bool blocked) |
253 | int wimax_rfkill_toggle_radio(void *data, enum rfkill_state state) | ||
254 | { | 248 | { |
255 | int result; | 249 | int result; |
256 | struct wimax_dev *wimax_dev = data; | 250 | struct wimax_dev *wimax_dev = data; |
257 | struct device *dev = wimax_dev_to_dev(wimax_dev); | 251 | struct device *dev = wimax_dev_to_dev(wimax_dev); |
258 | enum wimax_rf_state rf_state; | 252 | enum wimax_rf_state rf_state; |
259 | 253 | ||
260 | d_fnstart(3, dev, "(wimax_dev %p state %u)\n", wimax_dev, state); | 254 | d_fnstart(3, dev, "(wimax_dev %p blocked %u)\n", wimax_dev, blocked); |
261 | switch (state) { | 255 | rf_state = WIMAX_RF_ON; |
262 | case RFKILL_STATE_SOFT_BLOCKED: | 256 | if (blocked) |
263 | rf_state = WIMAX_RF_OFF; | 257 | rf_state = WIMAX_RF_OFF; |
264 | break; | ||
265 | case RFKILL_STATE_UNBLOCKED: | ||
266 | rf_state = WIMAX_RF_ON; | ||
267 | break; | ||
268 | default: | ||
269 | BUG(); | ||
270 | } | ||
271 | mutex_lock(&wimax_dev->mutex); | 258 | mutex_lock(&wimax_dev->mutex); |
272 | if (wimax_dev->state <= __WIMAX_ST_QUIESCING) | 259 | if (wimax_dev->state <= __WIMAX_ST_QUIESCING) |
273 | result = 0; /* just pretend it didn't happen */ | 260 | result = 0; |
274 | else | 261 | else |
275 | result = __wimax_rf_toggle_radio(wimax_dev, rf_state); | 262 | result = __wimax_rf_toggle_radio(wimax_dev, rf_state); |
276 | mutex_unlock(&wimax_dev->mutex); | 263 | mutex_unlock(&wimax_dev->mutex); |
277 | d_fnend(3, dev, "(wimax_dev %p state %u) = %d\n", | 264 | d_fnend(3, dev, "(wimax_dev %p blocked %u) = %d\n", |
278 | wimax_dev, state, result); | 265 | wimax_dev, blocked, result); |
279 | return result; | 266 | return result; |
280 | } | 267 | } |
281 | 268 | ||
269 | static const struct rfkill_ops wimax_rfkill_ops = { | ||
270 | .set_block = wimax_rfkill_set_radio_block, | ||
271 | }; | ||
282 | 272 | ||
283 | /** | 273 | /** |
284 | * wimax_rfkill - Set the software RF switch state for a WiMAX device | 274 | * wimax_rfkill - Set the software RF switch state for a WiMAX device |
@@ -322,6 +312,7 @@ int wimax_rfkill(struct wimax_dev *wimax_dev, enum wimax_rf_state state) | |||
322 | result = __wimax_rf_toggle_radio(wimax_dev, state); | 312 | result = __wimax_rf_toggle_radio(wimax_dev, state); |
323 | if (result < 0) | 313 | if (result < 0) |
324 | goto error; | 314 | goto error; |
315 | rfkill_set_sw_state(wimax_dev->rfkill, state == WIMAX_RF_OFF); | ||
325 | break; | 316 | break; |
326 | case WIMAX_RF_QUERY: | 317 | case WIMAX_RF_QUERY: |
327 | break; | 318 | break; |
@@ -349,40 +340,20 @@ int wimax_rfkill_add(struct wimax_dev *wimax_dev) | |||
349 | { | 340 | { |
350 | int result; | 341 | int result; |
351 | struct rfkill *rfkill; | 342 | struct rfkill *rfkill; |
352 | struct input_dev *input_dev; | ||
353 | struct device *dev = wimax_dev_to_dev(wimax_dev); | 343 | struct device *dev = wimax_dev_to_dev(wimax_dev); |
354 | 344 | ||
355 | d_fnstart(3, dev, "(wimax_dev %p)\n", wimax_dev); | 345 | d_fnstart(3, dev, "(wimax_dev %p)\n", wimax_dev); |
356 | /* Initialize RF Kill */ | 346 | /* Initialize RF Kill */ |
357 | result = -ENOMEM; | 347 | result = -ENOMEM; |
358 | rfkill = rfkill_allocate(dev, RFKILL_TYPE_WIMAX); | 348 | rfkill = rfkill_alloc(wimax_dev->name, dev, RFKILL_TYPE_WIMAX, |
349 | &wimax_rfkill_ops, wimax_dev); | ||
359 | if (rfkill == NULL) | 350 | if (rfkill == NULL) |
360 | goto error_rfkill_allocate; | 351 | goto error_rfkill_allocate; |
352 | |||
353 | d_printf(1, dev, "rfkill %p\n", rfkill); | ||
354 | |||
361 | wimax_dev->rfkill = rfkill; | 355 | wimax_dev->rfkill = rfkill; |
362 | 356 | ||
363 | rfkill->name = wimax_dev->name; | ||
364 | rfkill->state = RFKILL_STATE_UNBLOCKED; | ||
365 | rfkill->data = wimax_dev; | ||
366 | rfkill->toggle_radio = wimax_rfkill_toggle_radio; | ||
367 | |||
368 | /* Initialize the input device for the hw key */ | ||
369 | input_dev = input_allocate_device(); | ||
370 | if (input_dev == NULL) | ||
371 | goto error_input_allocate; | ||
372 | wimax_dev->rfkill_input = input_dev; | ||
373 | d_printf(1, dev, "rfkill %p input %p\n", rfkill, input_dev); | ||
374 | |||
375 | input_dev->name = wimax_dev->name; | ||
376 | /* FIXME: get a real device bus ID and stuff? do we care? */ | ||
377 | input_dev->id.bustype = BUS_HOST; | ||
378 | input_dev->id.vendor = 0xffff; | ||
379 | input_dev->evbit[0] = BIT(EV_KEY); | ||
380 | set_bit(KEY_WIMAX, input_dev->keybit); | ||
381 | |||
382 | /* Register both */ | ||
383 | result = input_register_device(wimax_dev->rfkill_input); | ||
384 | if (result < 0) | ||
385 | goto error_input_register; | ||
386 | result = rfkill_register(wimax_dev->rfkill); | 357 | result = rfkill_register(wimax_dev->rfkill); |
387 | if (result < 0) | 358 | if (result < 0) |
388 | goto error_rfkill_register; | 359 | goto error_rfkill_register; |
@@ -394,17 +365,8 @@ int wimax_rfkill_add(struct wimax_dev *wimax_dev) | |||
394 | d_fnend(3, dev, "(wimax_dev %p) = 0\n", wimax_dev); | 365 | d_fnend(3, dev, "(wimax_dev %p) = 0\n", wimax_dev); |
395 | return 0; | 366 | return 0; |
396 | 367 | ||
397 | /* if rfkill_register() suceeds, can't use rfkill_free() any | ||
398 | * more, only rfkill_unregister() [it owns the refcount]; with | ||
399 | * the input device we have the same issue--hence the if. */ | ||
400 | error_rfkill_register: | 368 | error_rfkill_register: |
401 | input_unregister_device(wimax_dev->rfkill_input); | 369 | rfkill_destroy(wimax_dev->rfkill); |
402 | wimax_dev->rfkill_input = NULL; | ||
403 | error_input_register: | ||
404 | if (wimax_dev->rfkill_input) | ||
405 | input_free_device(wimax_dev->rfkill_input); | ||
406 | error_input_allocate: | ||
407 | rfkill_free(wimax_dev->rfkill); | ||
408 | error_rfkill_allocate: | 370 | error_rfkill_allocate: |
409 | d_fnend(3, dev, "(wimax_dev %p) = %d\n", wimax_dev, result); | 371 | d_fnend(3, dev, "(wimax_dev %p) = %d\n", wimax_dev, result); |
410 | return result; | 372 | return result; |
@@ -423,45 +385,12 @@ void wimax_rfkill_rm(struct wimax_dev *wimax_dev) | |||
423 | { | 385 | { |
424 | struct device *dev = wimax_dev_to_dev(wimax_dev); | 386 | struct device *dev = wimax_dev_to_dev(wimax_dev); |
425 | d_fnstart(3, dev, "(wimax_dev %p)\n", wimax_dev); | 387 | d_fnstart(3, dev, "(wimax_dev %p)\n", wimax_dev); |
426 | rfkill_unregister(wimax_dev->rfkill); /* frees */ | 388 | rfkill_unregister(wimax_dev->rfkill); |
427 | input_unregister_device(wimax_dev->rfkill_input); | 389 | rfkill_destroy(wimax_dev->rfkill); |
428 | d_fnend(3, dev, "(wimax_dev %p)\n", wimax_dev); | 390 | d_fnend(3, dev, "(wimax_dev %p)\n", wimax_dev); |
429 | } | 391 | } |
430 | 392 | ||
431 | 393 | ||
432 | #else /* #ifdef CONFIG_RFKILL */ | ||
433 | |||
434 | void wimax_report_rfkill_hw(struct wimax_dev *wimax_dev, | ||
435 | enum wimax_rf_state state) | ||
436 | { | ||
437 | } | ||
438 | EXPORT_SYMBOL_GPL(wimax_report_rfkill_hw); | ||
439 | |||
440 | void wimax_report_rfkill_sw(struct wimax_dev *wimax_dev, | ||
441 | enum wimax_rf_state state) | ||
442 | { | ||
443 | } | ||
444 | EXPORT_SYMBOL_GPL(wimax_report_rfkill_sw); | ||
445 | |||
446 | int wimax_rfkill(struct wimax_dev *wimax_dev, | ||
447 | enum wimax_rf_state state) | ||
448 | { | ||
449 | return WIMAX_RF_ON << 1 | WIMAX_RF_ON; | ||
450 | } | ||
451 | EXPORT_SYMBOL_GPL(wimax_rfkill); | ||
452 | |||
453 | int wimax_rfkill_add(struct wimax_dev *wimax_dev) | ||
454 | { | ||
455 | return 0; | ||
456 | } | ||
457 | |||
458 | void wimax_rfkill_rm(struct wimax_dev *wimax_dev) | ||
459 | { | ||
460 | } | ||
461 | |||
462 | #endif /* #ifdef CONFIG_RFKILL */ | ||
463 | |||
464 | |||
465 | /* | 394 | /* |
466 | * Exporting to user space over generic netlink | 395 | * Exporting to user space over generic netlink |
467 | * | 396 | * |