1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
|
/* sched_plugin.c -- core infrastructure for the scheduler plugin system
*
* This file includes the initialization of the plugin system, the no-op Linux
* scheduler plugin, some dummy functions, and some helper functions.
*/
#include <linux/list.h>
#include <linux/spinlock.h>
#include <litmus/litmus.h>
#include <litmus/sched_plugin.h>
#include <litmus/jobs.h>
/*
* Generic function to trigger preemption on either local or remote cpu
* from scheduler plugins. The key feature is that this function is
* non-preemptive section aware and does not invoke the scheduler / send
* IPIs if the to-be-preempted task is actually non-preemptive.
*/
void preempt_if_preemptable(struct task_struct* t, int on_cpu)
{
/* t is the real-time task executing on CPU on_cpu If t is NULL, then
* on_cpu is currently scheduling background work.
*/
int send_ipi;
if (smp_processor_id() == on_cpu) {
/* local CPU case */
if (t) {
/* check if we need to poke userspace */
if (is_user_np(t))
/* yes, poke it */
request_exit_np(t);
else
/* no, see if we are allowed to preempt the
* currently-executing task */
if (!is_kernel_np(t))
set_tsk_need_resched(t);
} else
/* move non-real-time task out of the way */
set_tsk_need_resched(current);
} else {
/* remote CPU case */
if (!t)
/* currently schedules non-real-time work */
send_ipi = 1;
else {
/* currently schedules real-time work */
if (is_user_np(t)) {
/* need to notify user space of delayed
* preemption */
/* to avoid a race, set the flag, then test
* again */
request_exit_np(t);
/* make sure it got written */
mb();
}
/* Only send an ipi if remote task might have raced our
* request, i.e., send an IPI to make sure if it exited
* its critical section.
*/
send_ipi = !is_np(t) && !is_kernel_np(t);
}
if (likely(send_ipi))
smp_send_reschedule(on_cpu);
}
}
/*************************************************************
* Dummy plugin functions *
*************************************************************/
static void litmus_dummy_finish_switch(struct task_struct * prev)
{
}
static struct task_struct* litmus_dummy_schedule(struct task_struct * prev)
{
return NULL;
}
static void litmus_dummy_tick(struct task_struct* tsk)
{
}
static long litmus_dummy_admit_task(struct task_struct* tsk)
{
printk(KERN_CRIT "LITMUS^RT: Linux plugin rejects %s/%d.\n",
tsk->comm, tsk->pid);
return -EINVAL;
}
static void litmus_dummy_task_new(struct task_struct *t, int on_rq, int running)
{
}
static void litmus_dummy_task_wake_up(struct task_struct *task)
{
}
static void litmus_dummy_task_block(struct task_struct *task)
{
}
static void litmus_dummy_task_exit(struct task_struct *task)
{
}
static long litmus_dummy_complete_job(void)
{
return -ENOSYS;
}
static long litmus_dummy_activate_plugin(void)
{
return 0;
}
static long litmus_dummy_deactivate_plugin(void)
{
return 0;
}
#ifdef CONFIG_PI_SEMAPHORES
static long litmus_dummy_inherit_priority(struct pi_semaphore *sem,
struct task_struct *new_owner)
{
return -ENOSYS;
}
static long litmus_dummy_return_priority(struct pi_semaphore *sem)
{
return -ENOSYS;
}
static long litmus_dummy_pi_block(struct pi_semaphore *sem,
struct task_struct *new_waiter)
{
return -ENOSYS;
}
#endif
/* The default scheduler plugin. It doesn't do anything and lets Linux do its
* job.
*/
struct sched_plugin linux_sched_plugin = {
.plugin_name = "Linux",
.tick = litmus_dummy_tick,
.task_new = litmus_dummy_task_new,
.task_exit = litmus_dummy_task_exit,
.task_wake_up = litmus_dummy_task_wake_up,
.task_block = litmus_dummy_task_block,
.complete_job = litmus_dummy_complete_job,
.schedule = litmus_dummy_schedule,
.finish_switch = litmus_dummy_finish_switch,
.activate_plugin = litmus_dummy_activate_plugin,
.deactivate_plugin = litmus_dummy_deactivate_plugin,
#ifdef CONFIG_FMLP
.fmlp_inherit_priority = litmus_dummy_inherit_priority,
.fmlp_return_priority = litmus_dummy_return_priority,
.fmlp_pi_block = litmus_dummy_pi_block,
#endif
.admit_task = litmus_dummy_admit_task
};
/*
* The cluster size is needed in C-EDF: it makes sense only to cluster
* around L2 or L3, so if cluster_cache_index = 2 (default) we cluster
* all the CPUs that shares a L2 cache, while cluster_cache_index = 3
* we cluster all CPs that shares a L3 cache
*/
int cluster_cache_index = 2;
/*
* The reference to current plugin that is used to schedule tasks within
* the system. It stores references to actual function implementations
* Should be initialized by calling "init_***_plugin()"
*/
struct sched_plugin *litmus = &linux_sched_plugin;
/* the list of registered scheduling plugins */
static LIST_HEAD(sched_plugins);
static DEFINE_RAW_SPINLOCK(sched_plugins_lock);
#define CHECK(func) {\
if (!plugin->func) \
plugin->func = litmus_dummy_ ## func;}
#define CHECK_PI(type, func) {\
if (!plugin->type ## _ ## func) \
plugin->type ## _ ## func = litmus_dummy_ ## func; }
/* FIXME: get reference to module */
int register_sched_plugin(struct sched_plugin* plugin)
{
printk(KERN_INFO "Registering LITMUS^RT plugin %s.\n",
plugin->plugin_name);
/* make sure we don't trip over null pointers later */
CHECK(finish_switch);
CHECK(schedule);
CHECK(tick);
CHECK(task_wake_up);
CHECK(task_exit);
CHECK(task_block);
CHECK(task_new);
CHECK(complete_job);
CHECK(activate_plugin);
CHECK(deactivate_plugin);
#ifdef CONFIG_FMLP
CHECK_PI(fmlp, inherit_priority);
CHECK_PI(fmlp, return_priority);
CHECK_PI(fmlp, pi_block);
#endif
CHECK(admit_task);
if (!plugin->release_at)
plugin->release_at = release_at;
raw_spin_lock(&sched_plugins_lock);
list_add(&plugin->list, &sched_plugins);
raw_spin_unlock(&sched_plugins_lock);
return 0;
}
/* FIXME: reference counting, etc. */
struct sched_plugin* find_sched_plugin(const char* name)
{
struct list_head *pos;
struct sched_plugin *plugin;
raw_spin_lock(&sched_plugins_lock);
list_for_each(pos, &sched_plugins) {
plugin = list_entry(pos, struct sched_plugin, list);
if (!strcmp(plugin->plugin_name, name))
goto out_unlock;
}
plugin = NULL;
out_unlock:
raw_spin_unlock(&sched_plugins_lock);
return plugin;
}
int print_sched_plugins(char* buf, int max)
{
int count = 0;
struct list_head *pos;
struct sched_plugin *plugin;
raw_spin_lock(&sched_plugins_lock);
list_for_each(pos, &sched_plugins) {
plugin = list_entry(pos, struct sched_plugin, list);
count += snprintf(buf + count, max - count, "%s\n", plugin->plugin_name);
if (max - count <= 0)
break;
}
raw_spin_unlock(&sched_plugins_lock);
return count;
}
|