diff options
author | Bjoern B. Brandenburg <bbb@cs.unc.edu> | 2011-07-26 22:03:18 -0400 |
---|---|---|
committer | Bjoern B. Brandenburg <bbb@cs.unc.edu> | 2011-11-24 15:21:08 -0500 |
commit | 81b8eb2ae452c241df9b3a1fb2116fa4d5adcb75 (patch) | |
tree | 8ec7804f346b2e573ce384eb221b9e7a81872b75 /litmus/sched_cedf.c | |
parent | 71083a7604e93e44536edde032706348f3a752ca (diff) |
C-EDF: rename lock -> cluster_lock
The macro lock conflicts with locking protocols...
Diffstat (limited to 'litmus/sched_cedf.c')
-rw-r--r-- | litmus/sched_cedf.c | 26 |
1 files changed, 13 insertions, 13 deletions
diff --git a/litmus/sched_cedf.c b/litmus/sched_cedf.c index 8b3f8a7e2609..480c62bc895b 100644 --- a/litmus/sched_cedf.c +++ b/litmus/sched_cedf.c | |||
@@ -99,7 +99,7 @@ typedef struct clusterdomain { | |||
99 | struct bheap_node *heap_node; | 99 | struct bheap_node *heap_node; |
100 | struct bheap cpu_heap; | 100 | struct bheap cpu_heap; |
101 | /* lock for this cluster */ | 101 | /* lock for this cluster */ |
102 | #define lock domain.ready_lock | 102 | #define cluster_lock domain.ready_lock |
103 | } cedf_domain_t; | 103 | } cedf_domain_t; |
104 | 104 | ||
105 | /* a cedf_domain per cluster; allocation is done at init/activation time */ | 105 | /* a cedf_domain per cluster; allocation is done at init/activation time */ |
@@ -331,12 +331,12 @@ static void cedf_release_jobs(rt_domain_t* rt, struct bheap* tasks) | |||
331 | cedf_domain_t* cluster = container_of(rt, cedf_domain_t, domain); | 331 | cedf_domain_t* cluster = container_of(rt, cedf_domain_t, domain); |
332 | unsigned long flags; | 332 | unsigned long flags; |
333 | 333 | ||
334 | raw_spin_lock_irqsave(&cluster->lock, flags); | 334 | raw_spin_lock_irqsave(&cluster->cluster_lock, flags); |
335 | 335 | ||
336 | __merge_ready(&cluster->domain, tasks); | 336 | __merge_ready(&cluster->domain, tasks); |
337 | check_for_preemptions(cluster); | 337 | check_for_preemptions(cluster); |
338 | 338 | ||
339 | raw_spin_unlock_irqrestore(&cluster->lock, flags); | 339 | raw_spin_unlock_irqrestore(&cluster->cluster_lock, flags); |
340 | } | 340 | } |
341 | 341 | ||
342 | /* caller holds cedf_lock */ | 342 | /* caller holds cedf_lock */ |
@@ -427,7 +427,7 @@ static struct task_struct* cedf_schedule(struct task_struct * prev) | |||
427 | } | 427 | } |
428 | #endif | 428 | #endif |
429 | 429 | ||
430 | raw_spin_lock(&cluster->lock); | 430 | raw_spin_lock(&cluster->cluster_lock); |
431 | clear_will_schedule(); | 431 | clear_will_schedule(); |
432 | 432 | ||
433 | /* sanity checking */ | 433 | /* sanity checking */ |
@@ -511,7 +511,7 @@ static struct task_struct* cedf_schedule(struct task_struct * prev) | |||
511 | next = prev; | 511 | next = prev; |
512 | 512 | ||
513 | sched_state_task_picked(); | 513 | sched_state_task_picked(); |
514 | raw_spin_unlock(&cluster->lock); | 514 | raw_spin_unlock(&cluster->cluster_lock); |
515 | 515 | ||
516 | #ifdef WANT_ALL_SCHED_EVENTS | 516 | #ifdef WANT_ALL_SCHED_EVENTS |
517 | TRACE("cedf_lock released, next=0x%p\n", next); | 517 | TRACE("cedf_lock released, next=0x%p\n", next); |
@@ -553,7 +553,7 @@ static void cedf_task_new(struct task_struct * t, int on_rq, int running) | |||
553 | /* the cluster doesn't change even if t is running */ | 553 | /* the cluster doesn't change even if t is running */ |
554 | cluster = task_cpu_cluster(t); | 554 | cluster = task_cpu_cluster(t); |
555 | 555 | ||
556 | raw_spin_lock_irqsave(&cluster->domain.ready_lock, flags); | 556 | raw_spin_lock_irqsave(&cluster->cluster_lock, flags); |
557 | 557 | ||
558 | /* setup job params */ | 558 | /* setup job params */ |
559 | release_at(t, litmus_clock()); | 559 | release_at(t, litmus_clock()); |
@@ -580,7 +580,7 @@ static void cedf_task_new(struct task_struct * t, int on_rq, int running) | |||
580 | t->rt_param.linked_on = NO_CPU; | 580 | t->rt_param.linked_on = NO_CPU; |
581 | 581 | ||
582 | cedf_job_arrival(t); | 582 | cedf_job_arrival(t); |
583 | raw_spin_unlock_irqrestore(&(cluster->domain.ready_lock), flags); | 583 | raw_spin_unlock_irqrestore(&(cluster->cluster_lock), flags); |
584 | } | 584 | } |
585 | 585 | ||
586 | static void cedf_task_wake_up(struct task_struct *task) | 586 | static void cedf_task_wake_up(struct task_struct *task) |
@@ -593,7 +593,7 @@ static void cedf_task_wake_up(struct task_struct *task) | |||
593 | 593 | ||
594 | cluster = task_cpu_cluster(task); | 594 | cluster = task_cpu_cluster(task); |
595 | 595 | ||
596 | raw_spin_lock_irqsave(&cluster->lock, flags); | 596 | raw_spin_lock_irqsave(&cluster->cluster_lock, flags); |
597 | /* We need to take suspensions because of semaphores into | 597 | /* We need to take suspensions because of semaphores into |
598 | * account! If a job resumes after being suspended due to acquiring | 598 | * account! If a job resumes after being suspended due to acquiring |
599 | * a semaphore, it should never be treated as a new job release. | 599 | * a semaphore, it should never be treated as a new job release. |
@@ -616,7 +616,7 @@ static void cedf_task_wake_up(struct task_struct *task) | |||
616 | } | 616 | } |
617 | } | 617 | } |
618 | cedf_job_arrival(task); | 618 | cedf_job_arrival(task); |
619 | raw_spin_unlock_irqrestore(&cluster->lock, flags); | 619 | raw_spin_unlock_irqrestore(&cluster->cluster_lock, flags); |
620 | } | 620 | } |
621 | 621 | ||
622 | static void cedf_task_block(struct task_struct *t) | 622 | static void cedf_task_block(struct task_struct *t) |
@@ -629,9 +629,9 @@ static void cedf_task_block(struct task_struct *t) | |||
629 | cluster = task_cpu_cluster(t); | 629 | cluster = task_cpu_cluster(t); |
630 | 630 | ||
631 | /* unlink if necessary */ | 631 | /* unlink if necessary */ |
632 | raw_spin_lock_irqsave(&cluster->lock, flags); | 632 | raw_spin_lock_irqsave(&cluster->cluster_lock, flags); |
633 | unlink(t); | 633 | unlink(t); |
634 | raw_spin_unlock_irqrestore(&cluster->lock, flags); | 634 | raw_spin_unlock_irqrestore(&cluster->cluster_lock, flags); |
635 | 635 | ||
636 | BUG_ON(!is_realtime(t)); | 636 | BUG_ON(!is_realtime(t)); |
637 | } | 637 | } |
@@ -643,7 +643,7 @@ static void cedf_task_exit(struct task_struct * t) | |||
643 | cedf_domain_t *cluster = task_cpu_cluster(t); | 643 | cedf_domain_t *cluster = task_cpu_cluster(t); |
644 | 644 | ||
645 | /* unlink if necessary */ | 645 | /* unlink if necessary */ |
646 | raw_spin_lock_irqsave(&cluster->lock, flags); | 646 | raw_spin_lock_irqsave(&cluster->cluster_lock, flags); |
647 | unlink(t); | 647 | unlink(t); |
648 | if (tsk_rt(t)->scheduled_on != NO_CPU) { | 648 | if (tsk_rt(t)->scheduled_on != NO_CPU) { |
649 | cpu_entry_t *cpu; | 649 | cpu_entry_t *cpu; |
@@ -651,7 +651,7 @@ static void cedf_task_exit(struct task_struct * t) | |||
651 | cpu->scheduled = NULL; | 651 | cpu->scheduled = NULL; |
652 | tsk_rt(t)->scheduled_on = NO_CPU; | 652 | tsk_rt(t)->scheduled_on = NO_CPU; |
653 | } | 653 | } |
654 | raw_spin_unlock_irqrestore(&cluster->lock, flags); | 654 | raw_spin_unlock_irqrestore(&cluster->cluster_lock, flags); |
655 | 655 | ||
656 | BUG_ON(!is_realtime(t)); | 656 | BUG_ON(!is_realtime(t)); |
657 | TRACE_TASK(t, "RIP\n"); | 657 | TRACE_TASK(t, "RIP\n"); |