1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
|
#include <linux/gfp.h>
#include <linux/cpumask.h>
#include <linux/list.h>
#include <linux/cacheinfo.h>
#include <litmus/debug_trace.h>
#include <litmus/clustered.h>
int get_shared_cpu_map(cpumask_var_t mask, unsigned int cpu, unsigned int index)
{
struct cpu_cacheinfo* info = get_cpu_cacheinfo(cpu);
struct cacheinfo *ci;
if (!info || index >= info->num_leaves) {
TRACE("no shared-cache CPUs: info=%d index=%u\n",
info != NULL, index);
return 1;
}
if (!info->info_list) {
TRACE("no shared-cache CPUs: no info_list (cpu\n");
}
ci = info->info_list + index;
cpumask_copy(mask, &ci->shared_cpu_map);
TRACE("get_shared: P%u@L%u -> %d siblings\n ", cpu, index, cpumask_weight(mask));
return 0;
}
int get_cluster_size(enum cache_level level)
{
cpumask_var_t mask;
int ok;
int num_cpus;
if (level == GLOBAL_CLUSTER)
return num_online_cpus();
else {
if (!zalloc_cpumask_var(&mask, GFP_ATOMIC))
return -ENOMEM;
/* assumes CPU 0 is representative of all CPUs */
ok = get_shared_cpu_map(mask, 0, level);
/* ok == 0 means we got the map; otherwise it's an invalid cache level */
if (ok == 0)
num_cpus = cpumask_weight(mask);
free_cpumask_var(mask);
if (ok == 0)
return num_cpus;
else
return -EINVAL;
}
}
int assign_cpus_to_clusters(enum cache_level level,
struct scheduling_cluster* clusters[],
unsigned int num_clusters,
struct cluster_cpu* cpus[],
unsigned int num_cpus)
{
cpumask_var_t mask;
unsigned int i, free_cluster = 0, low_cpu;
int err = 0;
if (!zalloc_cpumask_var(&mask, GFP_ATOMIC))
return -ENOMEM;
/* clear cluster pointers */
for (i = 0; i < num_cpus; i++) {
cpus[i]->id = i;
cpus[i]->cluster = NULL;
}
/* initialize clusters */
for (i = 0; i < num_clusters; i++) {
clusters[i]->id = i;
INIT_LIST_HEAD(&clusters[i]->cpus);
}
/* Assign each CPU. Two assumtions are made:
* 1) The index of a cpu in cpus corresponds to its processor id (i.e., the index in a cpu mask).
* 2) All cpus that belong to some cluster are online.
*/
for_each_online_cpu(i) {
/* get lowest-id CPU in cluster */
if (level != GLOBAL_CLUSTER) {
err = get_shared_cpu_map(mask, cpus[i]->id, level);
if (err != 0) {
/* ugh... wrong cache level? Either caller screwed up
* or the CPU topology is weird. */
printk(KERN_ERR "Could not set up clusters for L%d sharing (max: L%d).\n",
level, err);
err = -EINVAL;
goto out;
}
low_cpu = cpumask_first(mask);
} else
low_cpu = 0;
if (low_cpu == i) {
/* caller must provide an appropriate number of clusters */
BUG_ON(free_cluster >= num_clusters);
/* create new cluster */
cpus[i]->cluster = clusters[free_cluster++];
} else {
/* low_cpu points to the right cluster
* Assumption: low_cpu is actually online and was processed earlier. */
cpus[i]->cluster = cpus[low_cpu]->cluster;
}
/* enqueue in cpus list */
list_add_tail(&cpus[i]->cluster_list, &cpus[i]->cluster->cpus);
printk(KERN_INFO "Assigning CPU%u to cluster %u\n.", i, cpus[i]->cluster->id);
}
out:
free_cpumask_var(mask);
return err;
}
|