/* This file is included from kernel/sched.c */ #include "sched.h" #include #include #include #include #include #include #include #include static void update_time_litmus(struct rq *rq, struct task_struct *p) { u64 delta = rq->clock - p->se.exec_start; if (unlikely((s64)delta < 0)) delta = 0; /* per job counter */ p->rt_param.job_params.exec_time += delta; /* task counter */ p->se.sum_exec_runtime += delta; if (delta) { TRACE_TASK(p, "charged %llu exec time (total:%llu, rem:%llu)\n", delta, p->rt_param.job_params.exec_time, budget_remaining(p)); } /* sched_clock() */ p->se.exec_start = rq->clock; cpuacct_charge(p, delta); } static void double_rq_lock(struct rq *rq1, struct rq *rq2); static void double_rq_unlock(struct rq *rq1, struct rq *rq2); static struct task_struct * litmus_schedule(struct rq *rq, struct task_struct *prev) { struct task_struct *next; #ifdef CONFIG_SMP struct rq* other_rq; long was_running; int from_where; lt_t _maybe_deadlock = 0; #endif /* let the plugin schedule */ next = litmus->schedule(prev); sched_state_plugin_check(); #ifdef CONFIG_SMP /* check if a global plugin pulled a task from a different RQ */ if (next && task_rq(next) != rq) { /* we need to migrate the task */ other_rq = task_rq(next); from_where = other_rq->cpu; TRACE_TASK(next, "migrate from %d\n", from_where); /* while we drop the lock, the prev task could change its * state */ BUG_ON(prev != current); was_running = is_current_running(); /* Don't race with a concurrent switch. This could deadlock in * the case of cross or circular migrations. It's the job of * the plugin to make sure that doesn't happen. */ TRACE_TASK(next, "stack_in_use=%d\n", next->rt_param.stack_in_use); if (next->rt_param.stack_in_use != NO_CPU) { TRACE_TASK(next, "waiting to deschedule\n"); _maybe_deadlock = litmus_clock(); } raw_spin_unlock(&rq->lock); while (next->rt_param.stack_in_use != NO_CPU) { cpu_relax(); mb(); if (next->rt_param.stack_in_use == NO_CPU) TRACE_TASK(next,"descheduled. Proceeding.\n"); if (!litmus->should_wait_for_stack(next)) { /* plugin aborted the wait */ TRACE_TASK(next, "plugin gave up waiting for stack\n"); next = NULL; /* Make sure plugin is given a chance to * reconsider. */ litmus_reschedule_local(); /* give up */ raw_spin_lock(&rq->lock); goto out; } if (from_where != task_rq(next)->cpu) { /* The plugin should not give us something * that other cores are trying to pull, too */ TRACE_TASK(next, "next invalid: task keeps " "shifting around!? " "(%d->%d)\n", from_where, task_rq(next)->cpu); /* bail out */ raw_spin_lock(&rq->lock); litmus->next_became_invalid(next); litmus_reschedule_local(); next = NULL; goto out; } if (lt_before(_maybe_deadlock + 1000000000L, litmus_clock())) { /* We've been spinning for 1s. * Something can't be right! * Let's abandon the task and bail out; at least * we will have debug info instead of a hard * deadlock. */ #ifdef CONFIG_BUG_ON_MIGRATION_DEADLOCK BUG(); #else TRACE_TASK(next,"stack too long in use. " "Deadlock?\n"); next = NULL; /* bail out */ raw_spin_lock(&rq->lock); goto out; #endif } } #ifdef __ARCH_WANT_UNLOCKED_CTXSW if (next->on_cpu) TRACE_TASK(next, "waiting for !oncpu"); while (next->on_cpu) { cpu_relax(); mb(); } #endif double_rq_lock(rq, other_rq); if (other_rq == task_rq(next) && next->rt_param.stack_in_use == NO_CPU) { /* ok, we can grab it */ set_task_cpu(next, rq->cpu); /* release the other CPU's runqueue, but keep ours */ raw_spin_unlock(&other_rq->lock); } else { /* Either it moved or the stack was claimed; both is * bad and forces us to abort the migration. */ TRACE_TASK(next, "next invalid: no longer available\n"); raw_spin_unlock(&other_rq->lock); litmus->next_became_invalid(next); next = NULL; goto out; } if (!litmus->post_migration_validate(next)) { TRACE_TASK(next, "plugin deems task now invalid\n"); litmus_reschedule_local(); next = NULL; } } #endif /* check if the task became invalid while we dropped the lock */ if (next && (!is_realtime(next) || !tsk_rt(next)->present)) { TRACE_TASK(next, "BAD: next (no longer?) valid\n"); litmus->next_became_invalid(next); litmus_reschedule_local(); next = NULL; } if (next) { #ifdef CONFIG_SMP next->rt_param.stack_in_use = rq->cpu; #else next->rt_param.stack_in_use = 0; #endif update_rq_clock(rq); next->se.exec_start = rq->clock; } out: update_enforcement_timer(next); return next; } static void enqueue_task_litmus(struct rq *rq, struct task_struct *p, int flags) { tsk_rt(p)->present = 1; if (flags & ENQUEUE_WAKEUP) { sched_trace_task_resume(p); /* LITMUS^RT plugins need to update the state * _before_ making it available in global structures. * Linux gets away with being lazy about the task state * update. We can't do that, hence we update the task * state already here. * * WARNING: this needs to be re-evaluated when porting * to newer kernel versions. */ p->state = TASK_RUNNING; litmus->task_wake_up(p); rq->litmus.nr_running++; } else { TRACE_TASK(p, "ignoring an enqueue, not a wake up.\n"); p->se.exec_start = rq->clock; } } static void dequeue_task_litmus(struct rq *rq, struct task_struct *p, int flags) { if (flags & DEQUEUE_SLEEP) { #ifdef CONFIG_SCHED_TASK_TRACE tsk_rt(p)->job_params.last_suspension = litmus_clock(); #endif litmus->task_block(p); tsk_rt(p)->present = 0; sched_trace_task_block(p); rq->litmus.nr_running--; } else TRACE_TASK(p, "ignoring a dequeue, not going to sleep.\n"); } static void yield_task_litmus(struct rq *rq) { TS_SYSCALL_IN_START; TS_SYSCALL_IN_END; BUG_ON(rq->curr != current); /* sched_yield() is called to trigger delayed preemptions. * Thus, mark the current task as needing to be rescheduled. * This will cause the scheduler plugin to be invoked, which can * then determine if a preemption is still required. */ clear_exit_np(current); litmus_reschedule_local(); TS_SYSCALL_OUT_START; } /* Plugins are responsible for this. */ static void check_preempt_curr_litmus(struct rq *rq, struct task_struct *p, int flags) { } static void put_prev_task_litmus(struct rq *rq, struct task_struct *p) { } /* pick_next_task_litmus() - litmus_schedule() function * * return the next task to be scheduled */ static struct task_struct *pick_next_task_litmus(struct rq *rq, struct task_struct *prev, struct pin_cookie cookie) { struct task_struct *next; if (is_realtime(prev)) update_time_litmus(rq, prev); lockdep_unpin_lock(&rq->lock, cookie); TS_PLUGIN_SCHED_START; next = litmus_schedule(rq, prev); TS_PLUGIN_SCHED_END; lockdep_repin_lock(&rq->lock, cookie); /* This is a bit backwards: the other classes call put_prev_task() * _after_ they've determined that the class has some queued tasks. * We can't determine this easily because each plugin manages its own * ready queues, and because in the case of globally shared queues, * we really don't know whether we'll have something ready even if * we test here. So we do it in reverse: first ask the plugin to * provide a task, and if we find one, call put_prev_task() on the * previously scheduled task. */ if (next) put_prev_task(rq, prev); return next; } static void task_tick_litmus(struct rq *rq, struct task_struct *p, int queued) { if (is_realtime(p) && !queued) { update_time_litmus(rq, p); /* budget check for QUANTUM_ENFORCEMENT tasks */ if (budget_enforced(p) && budget_exhausted(p)) { litmus_reschedule_local(); } } } static void switched_to_litmus(struct rq *rq, struct task_struct *p) { } static void prio_changed_litmus(struct rq *rq, struct task_struct *p, int oldprio) { } unsigned int get_rr_interval_litmus(struct rq *rq, struct task_struct *p) { /* return infinity */ return 0; } /* This is called when a task became a real-time task, either due to a SCHED_* * class transition or due to PI mutex inheritance. We don't handle Linux PI * mutex inheritance yet (and probably never will). Use LITMUS provided * synchronization primitives instead. */ static void set_curr_task_litmus(struct rq *rq) { rq->curr->se.exec_start = rq->clock; } #ifdef CONFIG_SMP /* execve tries to rebalance task in this scheduling domain. * We don't care about the scheduling domain; can gets called from * exec, fork, wakeup. */ static int select_task_rq_litmus(struct task_struct *p, int cpu, int sd_flag, int flags) { /* preemption is already disabled. * We don't want to change cpu here */ return task_cpu(p); } #endif static void update_curr_litmus(struct rq *rq) { struct task_struct *p = rq->curr; if (!is_realtime(p)) return; update_time_litmus(rq, p); } const struct sched_class litmus_sched_class = { /* From 34f971f6 the stop/migrate worker threads have a class on * their own, which is the highest prio class. We don't support * cpu-hotplug or cpu throttling. Allows Litmus to use up to 1.0 * CPU capacity. */ .next = &fair_sched_class, .enqueue_task = enqueue_task_litmus, .dequeue_task = dequeue_task_litmus, .yield_task = yield_task_litmus, .check_preempt_curr = check_preempt_curr_litmus, .pick_next_task = pick_next_task_litmus, .put_prev_task = put_prev_task_litmus, #ifdef CONFIG_SMP .select_task_rq = select_task_rq_litmus, #endif .set_curr_task = set_curr_task_litmus, .task_tick = task_tick_litmus, .get_rr_interval = get_rr_interval_litmus, .prio_changed = prio_changed_litmus, .switched_to = switched_to_litmus, .update_curr = update_curr_litmus, };