diff options
Diffstat (limited to 'arch/mips/include/asm/bitops.h')
| -rw-r--r-- | arch/mips/include/asm/bitops.h | 672 |
1 files changed, 672 insertions, 0 deletions
diff --git a/arch/mips/include/asm/bitops.h b/arch/mips/include/asm/bitops.h new file mode 100644 index 000000000000..49df8c4c9d25 --- /dev/null +++ b/arch/mips/include/asm/bitops.h | |||
| @@ -0,0 +1,672 @@ | |||
| 1 | /* | ||
| 2 | * This file is subject to the terms and conditions of the GNU General Public | ||
| 3 | * License. See the file "COPYING" in the main directory of this archive | ||
| 4 | * for more details. | ||
| 5 | * | ||
| 6 | * Copyright (c) 1994 - 1997, 99, 2000, 06, 07 Ralf Baechle (ralf@linux-mips.org) | ||
| 7 | * Copyright (c) 1999, 2000 Silicon Graphics, Inc. | ||
| 8 | */ | ||
| 9 | #ifndef _ASM_BITOPS_H | ||
| 10 | #define _ASM_BITOPS_H | ||
| 11 | |||
| 12 | #ifndef _LINUX_BITOPS_H | ||
| 13 | #error only <linux/bitops.h> can be included directly | ||
| 14 | #endif | ||
| 15 | |||
| 16 | #include <linux/compiler.h> | ||
| 17 | #include <linux/irqflags.h> | ||
| 18 | #include <linux/types.h> | ||
| 19 | #include <asm/barrier.h> | ||
| 20 | #include <asm/bug.h> | ||
| 21 | #include <asm/byteorder.h> /* sigh ... */ | ||
| 22 | #include <asm/cpu-features.h> | ||
| 23 | #include <asm/sgidefs.h> | ||
| 24 | #include <asm/war.h> | ||
| 25 | |||
| 26 | #if _MIPS_SZLONG == 32 | ||
| 27 | #define SZLONG_LOG 5 | ||
| 28 | #define SZLONG_MASK 31UL | ||
| 29 | #define __LL "ll " | ||
| 30 | #define __SC "sc " | ||
| 31 | #define __INS "ins " | ||
| 32 | #define __EXT "ext " | ||
| 33 | #elif _MIPS_SZLONG == 64 | ||
| 34 | #define SZLONG_LOG 6 | ||
| 35 | #define SZLONG_MASK 63UL | ||
| 36 | #define __LL "lld " | ||
| 37 | #define __SC "scd " | ||
| 38 | #define __INS "dins " | ||
| 39 | #define __EXT "dext " | ||
| 40 | #endif | ||
| 41 | |||
| 42 | /* | ||
| 43 | * clear_bit() doesn't provide any barrier for the compiler. | ||
| 44 | */ | ||
| 45 | #define smp_mb__before_clear_bit() smp_llsc_mb() | ||
| 46 | #define smp_mb__after_clear_bit() smp_llsc_mb() | ||
| 47 | |||
| 48 | /* | ||
| 49 | * set_bit - Atomically set a bit in memory | ||
| 50 | * @nr: the bit to set | ||
| 51 | * @addr: the address to start counting from | ||
| 52 | * | ||
| 53 | * This function is atomic and may not be reordered. See __set_bit() | ||
| 54 | * if you do not require the atomic guarantees. | ||
| 55 | * Note that @nr may be almost arbitrarily large; this function is not | ||
| 56 | * restricted to acting on a single-word quantity. | ||
| 57 | */ | ||
| 58 | static inline void set_bit(unsigned long nr, volatile unsigned long *addr) | ||
| 59 | { | ||
| 60 | unsigned long *m = ((unsigned long *) addr) + (nr >> SZLONG_LOG); | ||
| 61 | unsigned short bit = nr & SZLONG_MASK; | ||
| 62 | unsigned long temp; | ||
| 63 | |||
| 64 | if (cpu_has_llsc && R10000_LLSC_WAR) { | ||
| 65 | __asm__ __volatile__( | ||
| 66 | " .set mips3 \n" | ||
| 67 | "1: " __LL "%0, %1 # set_bit \n" | ||
| 68 | " or %0, %2 \n" | ||
| 69 | " " __SC "%0, %1 \n" | ||
| 70 | " beqzl %0, 1b \n" | ||
| 71 | " .set mips0 \n" | ||
| 72 | : "=&r" (temp), "=m" (*m) | ||
| 73 | : "ir" (1UL << bit), "m" (*m)); | ||
| 74 | #ifdef CONFIG_CPU_MIPSR2 | ||
| 75 | } else if (__builtin_constant_p(bit)) { | ||
| 76 | __asm__ __volatile__( | ||
| 77 | "1: " __LL "%0, %1 # set_bit \n" | ||
| 78 | " " __INS "%0, %4, %2, 1 \n" | ||
| 79 | " " __SC "%0, %1 \n" | ||
| 80 | " beqz %0, 2f \n" | ||
| 81 | " .subsection 2 \n" | ||
| 82 | "2: b 1b \n" | ||
| 83 | " .previous \n" | ||
| 84 | : "=&r" (temp), "=m" (*m) | ||
| 85 | : "ir" (bit), "m" (*m), "r" (~0)); | ||
| 86 | #endif /* CONFIG_CPU_MIPSR2 */ | ||
| 87 | } else if (cpu_has_llsc) { | ||
| 88 | __asm__ __volatile__( | ||
| 89 | " .set mips3 \n" | ||
| 90 | "1: " __LL "%0, %1 # set_bit \n" | ||
| 91 | " or %0, %2 \n" | ||
| 92 | " " __SC "%0, %1 \n" | ||
| 93 | " beqz %0, 2f \n" | ||
| 94 | " .subsection 2 \n" | ||
| 95 | "2: b 1b \n" | ||
| 96 | " .previous \n" | ||
| 97 | " .set mips0 \n" | ||
| 98 | : "=&r" (temp), "=m" (*m) | ||
| 99 | : "ir" (1UL << bit), "m" (*m)); | ||
| 100 | } else { | ||
| 101 | volatile unsigned long *a = addr; | ||
| 102 | unsigned long mask; | ||
| 103 | unsigned long flags; | ||
| 104 | |||
| 105 | a += nr >> SZLONG_LOG; | ||
| 106 | mask = 1UL << bit; | ||
| 107 | raw_local_irq_save(flags); | ||
| 108 | *a |= mask; | ||
| 109 | raw_local_irq_restore(flags); | ||
| 110 | } | ||
| 111 | } | ||
| 112 | |||
| 113 | /* | ||
| 114 | * clear_bit - Clears a bit in memory | ||
| 115 | * @nr: Bit to clear | ||
| 116 | * @addr: Address to start counting from | ||
| 117 | * | ||
| 118 | * clear_bit() is atomic and may not be reordered. However, it does | ||
| 119 | * not contain a memory barrier, so if it is used for locking purposes, | ||
| 120 | * you should call smp_mb__before_clear_bit() and/or smp_mb__after_clear_bit() | ||
| 121 | * in order to ensure changes are visible on other processors. | ||
| 122 | */ | ||
| 123 | static inline void clear_bit(unsigned long nr, volatile unsigned long *addr) | ||
| 124 | { | ||
| 125 | unsigned long *m = ((unsigned long *) addr) + (nr >> SZLONG_LOG); | ||
| 126 | unsigned short bit = nr & SZLONG_MASK; | ||
| 127 | unsigned long temp; | ||
| 128 | |||
| 129 | if (cpu_has_llsc && R10000_LLSC_WAR) { | ||
| 130 | __asm__ __volatile__( | ||
| 131 | " .set mips3 \n" | ||
| 132 | "1: " __LL "%0, %1 # clear_bit \n" | ||
| 133 | " and %0, %2 \n" | ||
| 134 | " " __SC "%0, %1 \n" | ||
| 135 | " beqzl %0, 1b \n" | ||
| 136 | " .set mips0 \n" | ||
| 137 | : "=&r" (temp), "=m" (*m) | ||
| 138 | : "ir" (~(1UL << bit)), "m" (*m)); | ||
| 139 | #ifdef CONFIG_CPU_MIPSR2 | ||
| 140 | } else if (__builtin_constant_p(bit)) { | ||
| 141 | __asm__ __volatile__( | ||
| 142 | "1: " __LL "%0, %1 # clear_bit \n" | ||
| 143 | " " __INS "%0, $0, %2, 1 \n" | ||
| 144 | " " __SC "%0, %1 \n" | ||
| 145 | " beqz %0, 2f \n" | ||
| 146 | " .subsection 2 \n" | ||
| 147 | "2: b 1b \n" | ||
| 148 | " .previous \n" | ||
| 149 | : "=&r" (temp), "=m" (*m) | ||
| 150 | : "ir" (bit), "m" (*m)); | ||
| 151 | #endif /* CONFIG_CPU_MIPSR2 */ | ||
| 152 | } else if (cpu_has_llsc) { | ||
| 153 | __asm__ __volatile__( | ||
| 154 | " .set mips3 \n" | ||
| 155 | "1: " __LL "%0, %1 # clear_bit \n" | ||
| 156 | " and %0, %2 \n" | ||
| 157 | " " __SC "%0, %1 \n" | ||
| 158 | " beqz %0, 2f \n" | ||
| 159 | " .subsection 2 \n" | ||
| 160 | "2: b 1b \n" | ||
| 161 | " .previous \n" | ||
| 162 | " .set mips0 \n" | ||
| 163 | : "=&r" (temp), "=m" (*m) | ||
| 164 | : "ir" (~(1UL << bit)), "m" (*m)); | ||
| 165 | } else { | ||
| 166 | volatile unsigned long *a = addr; | ||
| 167 | unsigned long mask; | ||
| 168 | unsigned long flags; | ||
| 169 | |||
| 170 | a += nr >> SZLONG_LOG; | ||
| 171 | mask = 1UL << bit; | ||
| 172 | raw_local_irq_save(flags); | ||
| 173 | *a &= ~mask; | ||
| 174 | raw_local_irq_restore(flags); | ||
| 175 | } | ||
| 176 | } | ||
| 177 | |||
| 178 | /* | ||
| 179 | * clear_bit_unlock - Clears a bit in memory | ||
| 180 | * @nr: Bit to clear | ||
| 181 | * @addr: Address to start counting from | ||
| 182 | * | ||
| 183 | * clear_bit() is atomic and implies release semantics before the memory | ||
| 184 | * operation. It can be used for an unlock. | ||
| 185 | */ | ||
| 186 | static inline void clear_bit_unlock(unsigned long nr, volatile unsigned long *addr) | ||
| 187 | { | ||
| 188 | smp_mb__before_clear_bit(); | ||
| 189 | clear_bit(nr, addr); | ||
| 190 | } | ||
| 191 | |||
| 192 | /* | ||
| 193 | * change_bit - Toggle a bit in memory | ||
| 194 | * @nr: Bit to change | ||
| 195 | * @addr: Address to start counting from | ||
| 196 | * | ||
| 197 | * change_bit() is atomic and may not be reordered. | ||
| 198 | * Note that @nr may be almost arbitrarily large; this function is not | ||
| 199 | * restricted to acting on a single-word quantity. | ||
| 200 | */ | ||
| 201 | static inline void change_bit(unsigned long nr, volatile unsigned long *addr) | ||
| 202 | { | ||
| 203 | unsigned short bit = nr & SZLONG_MASK; | ||
| 204 | |||
| 205 | if (cpu_has_llsc && R10000_LLSC_WAR) { | ||
| 206 | unsigned long *m = ((unsigned long *) addr) + (nr >> SZLONG_LOG); | ||
| 207 | unsigned long temp; | ||
| 208 | |||
| 209 | __asm__ __volatile__( | ||
| 210 | " .set mips3 \n" | ||
| 211 | "1: " __LL "%0, %1 # change_bit \n" | ||
| 212 | " xor %0, %2 \n" | ||
| 213 | " " __SC "%0, %1 \n" | ||
| 214 | " beqzl %0, 1b \n" | ||
| 215 | " .set mips0 \n" | ||
| 216 | : "=&r" (temp), "=m" (*m) | ||
| 217 | : "ir" (1UL << bit), "m" (*m)); | ||
| 218 | } else if (cpu_has_llsc) { | ||
| 219 | unsigned long *m = ((unsigned long *) addr) + (nr >> SZLONG_LOG); | ||
| 220 | unsigned long temp; | ||
| 221 | |||
| 222 | __asm__ __volatile__( | ||
| 223 | " .set mips3 \n" | ||
| 224 | "1: " __LL "%0, %1 # change_bit \n" | ||
| 225 | " xor %0, %2 \n" | ||
| 226 | " " __SC "%0, %1 \n" | ||
| 227 | " beqz %0, 2f \n" | ||
| 228 | " .subsection 2 \n" | ||
| 229 | "2: b 1b \n" | ||
| 230 | " .previous \n" | ||
| 231 | " .set mips0 \n" | ||
| 232 | : "=&r" (temp), "=m" (*m) | ||
| 233 | : "ir" (1UL << bit), "m" (*m)); | ||
| 234 | } else { | ||
| 235 | volatile unsigned long *a = addr; | ||
| 236 | unsigned long mask; | ||
| 237 | unsigned long flags; | ||
| 238 | |||
| 239 | a += nr >> SZLONG_LOG; | ||
| 240 | mask = 1UL << bit; | ||
| 241 | raw_local_irq_save(flags); | ||
| 242 | *a ^= mask; | ||
| 243 | raw_local_irq_restore(flags); | ||
| 244 | } | ||
| 245 | } | ||
| 246 | |||
| 247 | /* | ||
| 248 | * test_and_set_bit - Set a bit and return its old value | ||
| 249 | * @nr: Bit to set | ||
| 250 | * @addr: Address to count from | ||
| 251 | * | ||
| 252 | * This operation is atomic and cannot be reordered. | ||
| 253 | * It also implies a memory barrier. | ||
| 254 | */ | ||
| 255 | static inline int test_and_set_bit(unsigned long nr, | ||
| 256 | volatile unsigned long *addr) | ||
| 257 | { | ||
| 258 | unsigned short bit = nr & SZLONG_MASK; | ||
| 259 | unsigned long res; | ||
| 260 | |||
| 261 | smp_llsc_mb(); | ||
| 262 | |||
| 263 | if (cpu_has_llsc && R10000_LLSC_WAR) { | ||
| 264 | unsigned long *m = ((unsigned long *) addr) + (nr >> SZLONG_LOG); | ||
| 265 | unsigned long temp; | ||
| 266 | |||
| 267 | __asm__ __volatile__( | ||
| 268 | " .set mips3 \n" | ||
| 269 | "1: " __LL "%0, %1 # test_and_set_bit \n" | ||
| 270 | " or %2, %0, %3 \n" | ||
| 271 | " " __SC "%2, %1 \n" | ||
| 272 | " beqzl %2, 1b \n" | ||
| 273 | " and %2, %0, %3 \n" | ||
| 274 | " .set mips0 \n" | ||
| 275 | : "=&r" (temp), "=m" (*m), "=&r" (res) | ||
| 276 | : "r" (1UL << bit), "m" (*m) | ||
| 277 | : "memory"); | ||
| 278 | } else if (cpu_has_llsc) { | ||
| 279 | unsigned long *m = ((unsigned long *) addr) + (nr >> SZLONG_LOG); | ||
| 280 | unsigned long temp; | ||
| 281 | |||
| 282 | __asm__ __volatile__( | ||
| 283 | " .set push \n" | ||
| 284 | " .set noreorder \n" | ||
| 285 | " .set mips3 \n" | ||
| 286 | "1: " __LL "%0, %1 # test_and_set_bit \n" | ||
| 287 | " or %2, %0, %3 \n" | ||
| 288 | " " __SC "%2, %1 \n" | ||
| 289 | " beqz %2, 2f \n" | ||
| 290 | " and %2, %0, %3 \n" | ||
| 291 | " .subsection 2 \n" | ||
| 292 | "2: b 1b \n" | ||
| 293 | " nop \n" | ||
| 294 | " .previous \n" | ||
| 295 | " .set pop \n" | ||
| 296 | : "=&r" (temp), "=m" (*m), "=&r" (res) | ||
| 297 | : "r" (1UL << bit), "m" (*m) | ||
| 298 | : "memory"); | ||
| 299 | } else { | ||
| 300 | volatile unsigned long *a = addr; | ||
| 301 | unsigned long mask; | ||
| 302 | unsigned long flags; | ||
| 303 | |||
| 304 | a += nr >> SZLONG_LOG; | ||
| 305 | mask = 1UL << bit; | ||
| 306 | raw_local_irq_save(flags); | ||
| 307 | res = (mask & *a); | ||
| 308 | *a |= mask; | ||
| 309 | raw_local_irq_restore(flags); | ||
| 310 | } | ||
| 311 | |||
| 312 | smp_llsc_mb(); | ||
| 313 | |||
| 314 | return res != 0; | ||
| 315 | } | ||
| 316 | |||
| 317 | /* | ||
| 318 | * test_and_set_bit_lock - Set a bit and return its old value | ||
| 319 | * @nr: Bit to set | ||
| 320 | * @addr: Address to count from | ||
| 321 | * | ||
| 322 | * This operation is atomic and implies acquire ordering semantics | ||
| 323 | * after the memory operation. | ||
| 324 | */ | ||
| 325 | static inline int test_and_set_bit_lock(unsigned long nr, | ||
| 326 | volatile unsigned long *addr) | ||
| 327 | { | ||
| 328 | unsigned short bit = nr & SZLONG_MASK; | ||
| 329 | unsigned long res; | ||
| 330 | |||
| 331 | if (cpu_has_llsc && R10000_LLSC_WAR) { | ||
| 332 | unsigned long *m = ((unsigned long *) addr) + (nr >> SZLONG_LOG); | ||
| 333 | unsigned long temp; | ||
| 334 | |||
| 335 | __asm__ __volatile__( | ||
| 336 | " .set mips3 \n" | ||
| 337 | "1: " __LL "%0, %1 # test_and_set_bit \n" | ||
| 338 | " or %2, %0, %3 \n" | ||
| 339 | " " __SC "%2, %1 \n" | ||
| 340 | " beqzl %2, 1b \n" | ||
| 341 | " and %2, %0, %3 \n" | ||
| 342 | " .set mips0 \n" | ||
| 343 | : "=&r" (temp), "=m" (*m), "=&r" (res) | ||
| 344 | : "r" (1UL << bit), "m" (*m) | ||
| 345 | : "memory"); | ||
| 346 | } else if (cpu_has_llsc) { | ||
| 347 | unsigned long *m = ((unsigned long *) addr) + (nr >> SZLONG_LOG); | ||
| 348 | unsigned long temp; | ||
| 349 | |||
| 350 | __asm__ __volatile__( | ||
| 351 | " .set push \n" | ||
| 352 | " .set noreorder \n" | ||
| 353 | " .set mips3 \n" | ||
| 354 | "1: " __LL "%0, %1 # test_and_set_bit \n" | ||
| 355 | " or %2, %0, %3 \n" | ||
| 356 | " " __SC "%2, %1 \n" | ||
| 357 | " beqz %2, 2f \n" | ||
| 358 | " and %2, %0, %3 \n" | ||
| 359 | " .subsection 2 \n" | ||
| 360 | "2: b 1b \n" | ||
| 361 | " nop \n" | ||
| 362 | " .previous \n" | ||
| 363 | " .set pop \n" | ||
| 364 | : "=&r" (temp), "=m" (*m), "=&r" (res) | ||
| 365 | : "r" (1UL << bit), "m" (*m) | ||
| 366 | : "memory"); | ||
| 367 | } else { | ||
| 368 | volatile unsigned long *a = addr; | ||
| 369 | unsigned long mask; | ||
| 370 | unsigned long flags; | ||
| 371 | |||
| 372 | a += nr >> SZLONG_LOG; | ||
| 373 | mask = 1UL << bit; | ||
| 374 | raw_local_irq_save(flags); | ||
| 375 | res = (mask & *a); | ||
| 376 | *a |= mask; | ||
| 377 | raw_local_irq_restore(flags); | ||
| 378 | } | ||
| 379 | |||
| 380 | smp_llsc_mb(); | ||
| 381 | |||
| 382 | return res != 0; | ||
| 383 | } | ||
| 384 | /* | ||
| 385 | * test_and_clear_bit - Clear a bit and return its old value | ||
| 386 | * @nr: Bit to clear | ||
| 387 | * @addr: Address to count from | ||
| 388 | * | ||
| 389 | * This operation is atomic and cannot be reordered. | ||
| 390 | * It also implies a memory barrier. | ||
| 391 | */ | ||
| 392 | static inline int test_and_clear_bit(unsigned long nr, | ||
| 393 | volatile unsigned long *addr) | ||
| 394 | { | ||
| 395 | unsigned short bit = nr & SZLONG_MASK; | ||
| 396 | unsigned long res; | ||
| 397 | |||
| 398 | smp_llsc_mb(); | ||
| 399 | |||
| 400 | if (cpu_has_llsc && R10000_LLSC_WAR) { | ||
| 401 | unsigned long *m = ((unsigned long *) addr) + (nr >> SZLONG_LOG); | ||
| 402 | unsigned long temp; | ||
| 403 | |||
| 404 | __asm__ __volatile__( | ||
| 405 | " .set mips3 \n" | ||
| 406 | "1: " __LL "%0, %1 # test_and_clear_bit \n" | ||
| 407 | " or %2, %0, %3 \n" | ||
| 408 | " xor %2, %3 \n" | ||
| 409 | " " __SC "%2, %1 \n" | ||
| 410 | " beqzl %2, 1b \n" | ||
| 411 | " and %2, %0, %3 \n" | ||
| 412 | " .set mips0 \n" | ||
| 413 | : "=&r" (temp), "=m" (*m), "=&r" (res) | ||
| 414 | : "r" (1UL << bit), "m" (*m) | ||
| 415 | : "memory"); | ||
| 416 | #ifdef CONFIG_CPU_MIPSR2 | ||
| 417 | } else if (__builtin_constant_p(nr)) { | ||
| 418 | unsigned long *m = ((unsigned long *) addr) + (nr >> SZLONG_LOG); | ||
| 419 | unsigned long temp; | ||
| 420 | |||
| 421 | __asm__ __volatile__( | ||
| 422 | "1: " __LL "%0, %1 # test_and_clear_bit \n" | ||
| 423 | " " __EXT "%2, %0, %3, 1 \n" | ||
| 424 | " " __INS "%0, $0, %3, 1 \n" | ||
| 425 | " " __SC "%0, %1 \n" | ||
| 426 | " beqz %0, 2f \n" | ||
| 427 | " .subsection 2 \n" | ||
| 428 | "2: b 1b \n" | ||
| 429 | " .previous \n" | ||
| 430 | : "=&r" (temp), "=m" (*m), "=&r" (res) | ||
| 431 | : "ir" (bit), "m" (*m) | ||
| 432 | : "memory"); | ||
| 433 | #endif | ||
| 434 | } else if (cpu_has_llsc) { | ||
| 435 | unsigned long *m = ((unsigned long *) addr) + (nr >> SZLONG_LOG); | ||
| 436 | unsigned long temp; | ||
| 437 | |||
| 438 | __asm__ __volatile__( | ||
| 439 | " .set push \n" | ||
| 440 | " .set noreorder \n" | ||
| 441 | " .set mips3 \n" | ||
| 442 | "1: " __LL "%0, %1 # test_and_clear_bit \n" | ||
| 443 | " or %2, %0, %3 \n" | ||
| 444 | " xor %2, %3 \n" | ||
| 445 | " " __SC "%2, %1 \n" | ||
| 446 | " beqz %2, 2f \n" | ||
| 447 | " and %2, %0, %3 \n" | ||
| 448 | " .subsection 2 \n" | ||
| 449 | "2: b 1b \n" | ||
| 450 | " nop \n" | ||
| 451 | " .previous \n" | ||
| 452 | " .set pop \n" | ||
| 453 | : "=&r" (temp), "=m" (*m), "=&r" (res) | ||
| 454 | : "r" (1UL << bit), "m" (*m) | ||
| 455 | : "memory"); | ||
| 456 | } else { | ||
| 457 | volatile unsigned long *a = addr; | ||
| 458 | unsigned long mask; | ||
| 459 | unsigned long flags; | ||
| 460 | |||
| 461 | a += nr >> SZLONG_LOG; | ||
| 462 | mask = 1UL << bit; | ||
| 463 | raw_local_irq_save(flags); | ||
| 464 | res = (mask & *a); | ||
| 465 | *a &= ~mask; | ||
| 466 | raw_local_irq_restore(flags); | ||
| 467 | } | ||
| 468 | |||
| 469 | smp_llsc_mb(); | ||
| 470 | |||
| 471 | return res != 0; | ||
| 472 | } | ||
| 473 | |||
| 474 | /* | ||
| 475 | * test_and_change_bit - Change a bit and return its old value | ||
| 476 | * @nr: Bit to change | ||
| 477 | * @addr: Address to count from | ||
| 478 | * | ||
| 479 | * This operation is atomic and cannot be reordered. | ||
| 480 | * It also implies a memory barrier. | ||
| 481 | */ | ||
| 482 | static inline int test_and_change_bit(unsigned long nr, | ||
| 483 | volatile unsigned long *addr) | ||
| 484 | { | ||
| 485 | unsigned short bit = nr & SZLONG_MASK; | ||
| 486 | unsigned long res; | ||
| 487 | |||
| 488 | smp_llsc_mb(); | ||
| 489 | |||
| 490 | if (cpu_has_llsc && R10000_LLSC_WAR) { | ||
| 491 | unsigned long *m = ((unsigned long *) addr) + (nr >> SZLONG_LOG); | ||
| 492 | unsigned long temp; | ||
| 493 | |||
| 494 | __asm__ __volatile__( | ||
| 495 | " .set mips3 \n" | ||
| 496 | "1: " __LL "%0, %1 # test_and_change_bit \n" | ||
| 497 | " xor %2, %0, %3 \n" | ||
| 498 | " " __SC "%2, %1 \n" | ||
| 499 | " beqzl %2, 1b \n" | ||
| 500 | " and %2, %0, %3 \n" | ||
| 501 | " .set mips0 \n" | ||
| 502 | : "=&r" (temp), "=m" (*m), "=&r" (res) | ||
| 503 | : "r" (1UL << bit), "m" (*m) | ||
| 504 | : "memory"); | ||
| 505 | } else if (cpu_has_llsc) { | ||
| 506 | unsigned long *m = ((unsigned long *) addr) + (nr >> SZLONG_LOG); | ||
| 507 | unsigned long temp; | ||
| 508 | |||
| 509 | __asm__ __volatile__( | ||
| 510 | " .set push \n" | ||
| 511 | " .set noreorder \n" | ||
| 512 | " .set mips3 \n" | ||
| 513 | "1: " __LL "%0, %1 # test_and_change_bit \n" | ||
| 514 | " xor %2, %0, %3 \n" | ||
| 515 | " " __SC "\t%2, %1 \n" | ||
| 516 | " beqz %2, 2f \n" | ||
| 517 | " and %2, %0, %3 \n" | ||
| 518 | " .subsection 2 \n" | ||
| 519 | "2: b 1b \n" | ||
| 520 | " nop \n" | ||
| 521 | " .previous \n" | ||
| 522 | " .set pop \n" | ||
| 523 | : "=&r" (temp), "=m" (*m), "=&r" (res) | ||
| 524 | : "r" (1UL << bit), "m" (*m) | ||
| 525 | : "memory"); | ||
| 526 | } else { | ||
| 527 | volatile unsigned long *a = addr; | ||
| 528 | unsigned long mask; | ||
| 529 | unsigned long flags; | ||
| 530 | |||
| 531 | a += nr >> SZLONG_LOG; | ||
| 532 | mask = 1UL << bit; | ||
| 533 | raw_local_irq_save(flags); | ||
| 534 | res = (mask & *a); | ||
| 535 | *a ^= mask; | ||
| 536 | raw_local_irq_restore(flags); | ||
| 537 | } | ||
| 538 | |||
| 539 | smp_llsc_mb(); | ||
| 540 | |||
| 541 | return res != 0; | ||
| 542 | } | ||
| 543 | |||
| 544 | #include <asm-generic/bitops/non-atomic.h> | ||
| 545 | |||
| 546 | /* | ||
| 547 | * __clear_bit_unlock - Clears a bit in memory | ||
| 548 | * @nr: Bit to clear | ||
| 549 | * @addr: Address to start counting from | ||
| 550 | * | ||
| 551 | * __clear_bit() is non-atomic and implies release semantics before the memory | ||
| 552 | * operation. It can be used for an unlock if no other CPUs can concurrently | ||
| 553 | * modify other bits in the word. | ||
| 554 | */ | ||
| 555 | static inline void __clear_bit_unlock(unsigned long nr, volatile unsigned long *addr) | ||
| 556 | { | ||
| 557 | smp_mb(); | ||
| 558 | __clear_bit(nr, addr); | ||
| 559 | } | ||
| 560 | |||
| 561 | #if defined(CONFIG_CPU_MIPS32) || defined(CONFIG_CPU_MIPS64) | ||
| 562 | |||
| 563 | /* | ||
| 564 | * Return the bit position (0..63) of the most significant 1 bit in a word | ||
| 565 | * Returns -1 if no 1 bit exists | ||
| 566 | */ | ||
| 567 | static inline unsigned long __fls(unsigned long x) | ||
| 568 | { | ||
| 569 | int lz; | ||
| 570 | |||
| 571 | if (sizeof(x) == 4) { | ||
| 572 | __asm__( | ||
| 573 | " .set push \n" | ||
| 574 | " .set mips32 \n" | ||
| 575 | " clz %0, %1 \n" | ||
| 576 | " .set pop \n" | ||
| 577 | : "=r" (lz) | ||
| 578 | : "r" (x)); | ||
| 579 | |||
| 580 | return 31 - lz; | ||
| 581 | } | ||
| 582 | |||
| 583 | BUG_ON(sizeof(x) != 8); | ||
| 584 | |||
| 585 | __asm__( | ||
| 586 | " .set push \n" | ||
| 587 | " .set mips64 \n" | ||
| 588 | " dclz %0, %1 \n" | ||
| 589 | " .set pop \n" | ||
| 590 | : "=r" (lz) | ||
| 591 | : "r" (x)); | ||
| 592 | |||
| 593 | return 63 - lz; | ||
| 594 | } | ||
| 595 | |||
| 596 | /* | ||
| 597 | * __ffs - find first bit in word. | ||
| 598 | * @word: The word to search | ||
| 599 | * | ||
| 600 | * Returns 0..SZLONG-1 | ||
| 601 | * Undefined if no bit exists, so code should check against 0 first. | ||
| 602 | */ | ||
| 603 | static inline unsigned long __ffs(unsigned long word) | ||
| 604 | { | ||
| 605 | return __fls(word & -word); | ||
| 606 | } | ||
| 607 | |||
| 608 | /* | ||
| 609 | * fls - find last bit set. | ||
| 610 | * @word: The word to search | ||
| 611 | * | ||
| 612 | * This is defined the same way as ffs. | ||
| 613 | * Note fls(0) = 0, fls(1) = 1, fls(0x80000000) = 32. | ||
| 614 | */ | ||
| 615 | static inline int fls(int word) | ||
| 616 | { | ||
| 617 | __asm__("clz %0, %1" : "=r" (word) : "r" (word)); | ||
| 618 | |||
| 619 | return 32 - word; | ||
| 620 | } | ||
| 621 | |||
| 622 | #if defined(CONFIG_64BIT) && defined(CONFIG_CPU_MIPS64) | ||
| 623 | static inline int fls64(__u64 word) | ||
| 624 | { | ||
| 625 | __asm__("dclz %0, %1" : "=r" (word) : "r" (word)); | ||
| 626 | |||
| 627 | return 64 - word; | ||
| 628 | } | ||
| 629 | #else | ||
| 630 | #include <asm-generic/bitops/fls64.h> | ||
| 631 | #endif | ||
| 632 | |||
| 633 | /* | ||
| 634 | * ffs - find first bit set. | ||
| 635 | * @word: The word to search | ||
| 636 | * | ||
| 637 | * This is defined the same way as | ||
| 638 | * the libc and compiler builtin ffs routines, therefore | ||
| 639 | * differs in spirit from the above ffz (man ffs). | ||
| 640 | */ | ||
| 641 | static inline int ffs(int word) | ||
| 642 | { | ||
| 643 | if (!word) | ||
| 644 | return 0; | ||
| 645 | |||
| 646 | return fls(word & -word); | ||
| 647 | } | ||
| 648 | |||
| 649 | #else | ||
| 650 | |||
| 651 | #include <asm-generic/bitops/__ffs.h> | ||
| 652 | #include <asm-generic/bitops/__fls.h> | ||
| 653 | #include <asm-generic/bitops/ffs.h> | ||
| 654 | #include <asm-generic/bitops/fls.h> | ||
| 655 | #include <asm-generic/bitops/fls64.h> | ||
| 656 | |||
| 657 | #endif /*defined(CONFIG_CPU_MIPS32) || defined(CONFIG_CPU_MIPS64) */ | ||
| 658 | |||
| 659 | #include <asm-generic/bitops/ffz.h> | ||
| 660 | #include <asm-generic/bitops/find.h> | ||
| 661 | |||
| 662 | #ifdef __KERNEL__ | ||
| 663 | |||
| 664 | #include <asm-generic/bitops/sched.h> | ||
| 665 | #include <asm-generic/bitops/hweight.h> | ||
| 666 | #include <asm-generic/bitops/ext2-non-atomic.h> | ||
| 667 | #include <asm-generic/bitops/ext2-atomic.h> | ||
| 668 | #include <asm-generic/bitops/minix.h> | ||
| 669 | |||
| 670 | #endif /* __KERNEL__ */ | ||
| 671 | |||
| 672 | #endif /* _ASM_BITOPS_H */ | ||
