aboutsummaryrefslogtreecommitdiffstats
path: root/arch/x86/include
diff options
context:
space:
mode:
authorLinus Torvalds <torvalds@linux-foundation.org>2012-02-18 00:48:54 -0500
committerLinus Torvalds <torvalds@linux-foundation.org>2012-02-18 13:19:41 -0500
commitf94edacf998516ac9d849f7bc6949a703977a7f3 (patch)
tree4a5c5487d86f3b3873b8ca1ba1c0b00b832870ad /arch/x86/include
parent4903062b5485f0e2c286a23b44c9b59d9b017d53 (diff)
i387: move TS_USEDFPU flag from thread_info to task_struct
This moves the bit that indicates whether a thread has ownership of the FPU from the TS_USEDFPU bit in thread_info->status to a word of its own (called 'has_fpu') in task_struct->thread.has_fpu. This fixes two independent bugs at the same time: - changing 'thread_info->status' from the scheduler causes nasty problems for the other users of that variable, since it is defined to be thread-synchronous (that's what the "TS_" part of the naming was supposed to indicate). So perfectly valid code could (and did) do ti->status |= TS_RESTORE_SIGMASK; and the compiler was free to do that as separate load, or and store instructions. Which can cause problems with preemption, since a task switch could happen in between, and change the TS_USEDFPU bit. The change to TS_USEDFPU would be overwritten by the final store. In practice, this seldom happened, though, because the 'status' field was seldom used more than once, so gcc would generally tend to generate code that used a read-modify-write instruction and thus happened to avoid this problem - RMW instructions are naturally low fat and preemption-safe. - On x86-32, the current_thread_info() pointer would, during interrupts and softirqs, point to a *copy* of the real thread_info, because x86-32 uses %esp to calculate the thread_info address, and thus the separate irq (and softirq) stacks would cause these kinds of odd thread_info copy aliases. This is normally not a problem, since interrupts aren't supposed to look at thread information anyway (what thread is running at interrupt time really isn't very well-defined), but it confused the heck out of irq_fpu_usable() and the code that tried to squirrel away the FPU state. (It also caused untold confusion for us poor kernel developers). It also turns out that using 'task_struct' is actually much more natural for most of the call sites that care about the FPU state, since they tend to work with the task struct for other reasons anyway (ie scheduling). And the FPU data that we are going to save/restore is found there too. Thanks to Arjan Van De Ven <arjan@linux.intel.com> for pointing us to the %esp issue. Cc: Arjan van de Ven <arjan@linux.intel.com> Reported-and-tested-by: Raphael Prevost <raphael@buro.asia> Acked-and-tested-by: Suresh Siddha <suresh.b.siddha@intel.com> Tested-by: Peter Anvin <hpa@zytor.com> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
Diffstat (limited to 'arch/x86/include')
-rw-r--r--arch/x86/include/asm/i387.h44
-rw-r--r--arch/x86/include/asm/processor.h1
-rw-r--r--arch/x86/include/asm/thread_info.h2
3 files changed, 23 insertions, 24 deletions
diff --git a/arch/x86/include/asm/i387.h b/arch/x86/include/asm/i387.h
index 01b115d8677..f5376676f89 100644
--- a/arch/x86/include/asm/i387.h
+++ b/arch/x86/include/asm/i387.h
@@ -264,21 +264,21 @@ static inline int restore_fpu_checking(struct task_struct *tsk)
264 * be preemption protection *and* they need to be 264 * be preemption protection *and* they need to be
265 * properly paired with the CR0.TS changes! 265 * properly paired with the CR0.TS changes!
266 */ 266 */
267static inline int __thread_has_fpu(struct thread_info *ti) 267static inline int __thread_has_fpu(struct task_struct *tsk)
268{ 268{
269 return ti->status & TS_USEDFPU; 269 return tsk->thread.has_fpu;
270} 270}
271 271
272/* Must be paired with an 'stts' after! */ 272/* Must be paired with an 'stts' after! */
273static inline void __thread_clear_has_fpu(struct thread_info *ti) 273static inline void __thread_clear_has_fpu(struct task_struct *tsk)
274{ 274{
275 ti->status &= ~TS_USEDFPU; 275 tsk->thread.has_fpu = 0;
276} 276}
277 277
278/* Must be paired with a 'clts' before! */ 278/* Must be paired with a 'clts' before! */
279static inline void __thread_set_has_fpu(struct thread_info *ti) 279static inline void __thread_set_has_fpu(struct task_struct *tsk)
280{ 280{
281 ti->status |= TS_USEDFPU; 281 tsk->thread.has_fpu = 1;
282} 282}
283 283
284/* 284/*
@@ -288,16 +288,16 @@ static inline void __thread_set_has_fpu(struct thread_info *ti)
288 * These generally need preemption protection to work, 288 * These generally need preemption protection to work,
289 * do try to avoid using these on their own. 289 * do try to avoid using these on their own.
290 */ 290 */
291static inline void __thread_fpu_end(struct thread_info *ti) 291static inline void __thread_fpu_end(struct task_struct *tsk)
292{ 292{
293 __thread_clear_has_fpu(ti); 293 __thread_clear_has_fpu(tsk);
294 stts(); 294 stts();
295} 295}
296 296
297static inline void __thread_fpu_begin(struct thread_info *ti) 297static inline void __thread_fpu_begin(struct task_struct *tsk)
298{ 298{
299 clts(); 299 clts();
300 __thread_set_has_fpu(ti); 300 __thread_set_has_fpu(tsk);
301} 301}
302 302
303/* 303/*
@@ -308,21 +308,21 @@ extern int restore_i387_xstate(void __user *buf);
308 308
309static inline void __unlazy_fpu(struct task_struct *tsk) 309static inline void __unlazy_fpu(struct task_struct *tsk)
310{ 310{
311 if (__thread_has_fpu(task_thread_info(tsk))) { 311 if (__thread_has_fpu(tsk)) {
312 __save_init_fpu(tsk); 312 __save_init_fpu(tsk);
313 __thread_fpu_end(task_thread_info(tsk)); 313 __thread_fpu_end(tsk);
314 } else 314 } else
315 tsk->fpu_counter = 0; 315 tsk->fpu_counter = 0;
316} 316}
317 317
318static inline void __clear_fpu(struct task_struct *tsk) 318static inline void __clear_fpu(struct task_struct *tsk)
319{ 319{
320 if (__thread_has_fpu(task_thread_info(tsk))) { 320 if (__thread_has_fpu(tsk)) {
321 /* Ignore delayed exceptions from user space */ 321 /* Ignore delayed exceptions from user space */
322 asm volatile("1: fwait\n" 322 asm volatile("1: fwait\n"
323 "2:\n" 323 "2:\n"
324 _ASM_EXTABLE(1b, 2b)); 324 _ASM_EXTABLE(1b, 2b));
325 __thread_fpu_end(task_thread_info(tsk)); 325 __thread_fpu_end(tsk);
326 } 326 }
327} 327}
328 328
@@ -337,7 +337,7 @@ static inline void __clear_fpu(struct task_struct *tsk)
337 */ 337 */
338static inline bool interrupted_kernel_fpu_idle(void) 338static inline bool interrupted_kernel_fpu_idle(void)
339{ 339{
340 return !__thread_has_fpu(current_thread_info()) && 340 return !__thread_has_fpu(current) &&
341 (read_cr0() & X86_CR0_TS); 341 (read_cr0() & X86_CR0_TS);
342} 342}
343 343
@@ -371,12 +371,12 @@ static inline bool irq_fpu_usable(void)
371 371
372static inline void kernel_fpu_begin(void) 372static inline void kernel_fpu_begin(void)
373{ 373{
374 struct thread_info *me = current_thread_info(); 374 struct task_struct *me = current;
375 375
376 WARN_ON_ONCE(!irq_fpu_usable()); 376 WARN_ON_ONCE(!irq_fpu_usable());
377 preempt_disable(); 377 preempt_disable();
378 if (__thread_has_fpu(me)) { 378 if (__thread_has_fpu(me)) {
379 __save_init_fpu(me->task); 379 __save_init_fpu(me);
380 __thread_clear_has_fpu(me); 380 __thread_clear_has_fpu(me);
381 /* We do 'stts()' in kernel_fpu_end() */ 381 /* We do 'stts()' in kernel_fpu_end() */
382 } else 382 } else
@@ -441,13 +441,13 @@ static inline void irq_ts_restore(int TS_state)
441 */ 441 */
442static inline int user_has_fpu(void) 442static inline int user_has_fpu(void)
443{ 443{
444 return __thread_has_fpu(current_thread_info()); 444 return __thread_has_fpu(current);
445} 445}
446 446
447static inline void user_fpu_end(void) 447static inline void user_fpu_end(void)
448{ 448{
449 preempt_disable(); 449 preempt_disable();
450 __thread_fpu_end(current_thread_info()); 450 __thread_fpu_end(current);
451 preempt_enable(); 451 preempt_enable();
452} 452}
453 453
@@ -455,7 +455,7 @@ static inline void user_fpu_begin(void)
455{ 455{
456 preempt_disable(); 456 preempt_disable();
457 if (!user_has_fpu()) 457 if (!user_has_fpu())
458 __thread_fpu_begin(current_thread_info()); 458 __thread_fpu_begin(current);
459 preempt_enable(); 459 preempt_enable();
460} 460}
461 461
@@ -464,10 +464,10 @@ static inline void user_fpu_begin(void)
464 */ 464 */
465static inline void save_init_fpu(struct task_struct *tsk) 465static inline void save_init_fpu(struct task_struct *tsk)
466{ 466{
467 WARN_ON_ONCE(!__thread_has_fpu(task_thread_info(tsk))); 467 WARN_ON_ONCE(!__thread_has_fpu(tsk));
468 preempt_disable(); 468 preempt_disable();
469 __save_init_fpu(tsk); 469 __save_init_fpu(tsk);
470 __thread_fpu_end(task_thread_info(tsk)); 470 __thread_fpu_end(tsk);
471 preempt_enable(); 471 preempt_enable();
472} 472}
473 473
diff --git a/arch/x86/include/asm/processor.h b/arch/x86/include/asm/processor.h
index aa9088c2693..f7c89e231c6 100644
--- a/arch/x86/include/asm/processor.h
+++ b/arch/x86/include/asm/processor.h
@@ -454,6 +454,7 @@ struct thread_struct {
454 unsigned long trap_no; 454 unsigned long trap_no;
455 unsigned long error_code; 455 unsigned long error_code;
456 /* floating point and extended processor state */ 456 /* floating point and extended processor state */
457 unsigned long has_fpu;
457 struct fpu fpu; 458 struct fpu fpu;
458#ifdef CONFIG_X86_32 459#ifdef CONFIG_X86_32
459 /* Virtual 86 mode info */ 460 /* Virtual 86 mode info */
diff --git a/arch/x86/include/asm/thread_info.h b/arch/x86/include/asm/thread_info.h
index bc817cd8b44..cfd8144d552 100644
--- a/arch/x86/include/asm/thread_info.h
+++ b/arch/x86/include/asm/thread_info.h
@@ -247,8 +247,6 @@ static inline struct thread_info *current_thread_info(void)
247 * ever touches our thread-synchronous status, so we don't 247 * ever touches our thread-synchronous status, so we don't
248 * have to worry about atomic accesses. 248 * have to worry about atomic accesses.
249 */ 249 */
250#define TS_USEDFPU 0x0001 /* FPU was used by this task
251 this quantum (SMP) */
252#define TS_COMPAT 0x0002 /* 32bit syscall active (64BIT)*/ 250#define TS_COMPAT 0x0002 /* 32bit syscall active (64BIT)*/
253#define TS_POLLING 0x0004 /* idle task polling need_resched, 251#define TS_POLLING 0x0004 /* idle task polling need_resched,
254 skip sending interrupt */ 252 skip sending interrupt */