diff options
author | Jonathan Herman <hermanjl@cs.unc.edu> | 2013-04-23 14:01:35 -0400 |
---|---|---|
committer | Jonathan Herman <hermanjl@cs.unc.edu> | 2013-04-23 14:01:35 -0400 |
commit | 7545402506aa76261e18d85af585ff0ac1cf05c1 (patch) | |
tree | 6b5a6d2e819c10311f3b4cdc94174877bdfcfbde | |
parent | 25ccdb0cbc6b959b1f96c89b8bce91963cb67b4c (diff) |
Improved accuracy of sched_trace measurement parsing.
* Measurements from tasks missing > 20% of their scheduling records are
ignored. This is configurable in config/config.py.
* Measurements which only have zero values are ignored.
* If either of these 2 situations are encountered print out a message the
first time using the common.log_once() method. See parse_exps.py for how
this is used with multiple threads.
* Measurements from a task's last job are ignored.
* Miss ratio is calculated only as a fraction of the number of jobs whose
matching release and completion records were found, not just release.
-rw-r--r-- | common.py | 17 | ||||
-rw-r--r-- | config/config.py | 4 | ||||
-rw-r--r-- | parse/point.py | 4 | ||||
-rw-r--r-- | parse/sched.py | 84 | ||||
-rw-r--r-- | parse/tuple_table.py | 4 | ||||
-rwxr-xr-x | parse_exps.py | 159 |
6 files changed, 186 insertions, 86 deletions
@@ -193,3 +193,20 @@ def is_device(dev): | |||
193 | return False | 193 | return False |
194 | mode = os.stat(dev)[stat.ST_MODE] | 194 | mode = os.stat(dev)[stat.ST_MODE] |
195 | return not (not mode & stat.S_IFCHR) | 195 | return not (not mode & stat.S_IFCHR) |
196 | |||
197 | __logged = [] | ||
198 | |||
199 | def set_logged_list(logged): | ||
200 | global __logged | ||
201 | __logged = logged | ||
202 | |||
203 | def log_once(id, msg = None, indent = True): | ||
204 | global __logged | ||
205 | |||
206 | msg = msg if msg else id | ||
207 | |||
208 | if id not in __logged: | ||
209 | __logged += [id] | ||
210 | if indent: | ||
211 | msg = ' ' + msg.strip('\t').replace('\n', '\n\t') | ||
212 | sys.stderr.write('\n' + msg.strip('\n') + '\n') | ||
diff --git a/config/config.py b/config/config.py index b631aa2..5e6f9e3 100644 --- a/config/config.py +++ b/config/config.py | |||
@@ -56,3 +56,7 @@ OVH_ALL_EVENTS = ["%s_%s" % (e, t) for (e,t) in | |||
56 | OVH_ALL_EVENTS += ['RELEASE_LATENCY'] | 56 | OVH_ALL_EVENTS += ['RELEASE_LATENCY'] |
57 | # This event doesn't have a START and END | 57 | # This event doesn't have a START and END |
58 | OVH_BASE_EVENTS += ['RELEASE_LATENCY'] | 58 | OVH_BASE_EVENTS += ['RELEASE_LATENCY'] |
59 | |||
60 | # If a task is missing more than this many records, its measurements | ||
61 | # are not included in sched_trace summaries | ||
62 | MAX_RECORD_LOSS = .2 | ||
diff --git a/parse/point.py b/parse/point.py index ac47c70..b1d9d53 100644 --- a/parse/point.py +++ b/parse/point.py | |||
@@ -133,6 +133,10 @@ class ExpPoint(object): | |||
133 | def get_stats(self): | 133 | def get_stats(self): |
134 | return self.stats.keys() | 134 | return self.stats.keys() |
135 | 135 | ||
136 | def __bool__(self): | ||
137 | return bool(self.stats) | ||
138 | __nonzero__ = __bool__ | ||
139 | |||
136 | 140 | ||
137 | class SummaryPoint(ExpPoint): | 141 | class SummaryPoint(ExpPoint): |
138 | def __init__(self, id="", points=[], typemap = default_typemap): | 142 | def __init__(self, id="", points=[], typemap = default_typemap): |
diff --git a/parse/sched.py b/parse/sched.py index b56324b..4933037 100644 --- a/parse/sched.py +++ b/parse/sched.py | |||
@@ -5,35 +5,55 @@ import struct | |||
5 | import subprocess | 5 | import subprocess |
6 | 6 | ||
7 | from collections import defaultdict,namedtuple | 7 | from collections import defaultdict,namedtuple |
8 | from common import recordtype | 8 | from common import recordtype,log_once |
9 | from point import Measurement | 9 | from point import Measurement |
10 | from ctypes import * | 10 | from ctypes import * |
11 | 11 | ||
12 | class TimeTracker: | 12 | class TimeTracker: |
13 | '''Store stats for durations of time demarcated by sched_trace records.''' | 13 | '''Store stats for durations of time demarcated by sched_trace records.''' |
14 | def __init__(self): | 14 | def __init__(self): |
15 | self.begin = self.avg = self.max = self.num = self.job = 0 | 15 | self.begin = self.avg = self.max = self.num = self.next_job = 0 |
16 | 16 | ||
17 | def store_time(self, record): | 17 | # Count of times the job in start_time matched that in store_time |
18 | self.matches = 0 | ||
19 | # And the times it didn't | ||
20 | self.disjoints = 0 | ||
21 | |||
22 | # Measurements are recorded in store_ time using the previous matching | ||
23 | # record which was passed to store_time. This way, the last record for | ||
24 | # any task is always skipped | ||
25 | self.last_record = None | ||
26 | |||
27 | def store_time(self, next_record): | ||
18 | '''End duration of time.''' | 28 | '''End duration of time.''' |
19 | dur = record.when - self.begin | 29 | dur = (self.last_record.when - self.begin) if self.last_record else -1 |
20 | 30 | ||
21 | if self.job == record.job and dur > 0: | 31 | if self.next_job == next_record.job: |
22 | self.max = max(self.max, dur) | 32 | self.last_record = next_record |
23 | self.avg *= float(self.num / (self.num + 1)) | ||
24 | self.num += 1 | ||
25 | self.avg += dur / float(self.num) | ||
26 | 33 | ||
27 | self.begin = 0 | 34 | if self.last_record: |
28 | self.job = 0 | 35 | self.matches += 1 |
36 | |||
37 | if dur > 0: | ||
38 | self.max = max(self.max, dur) | ||
39 | self.avg *= float(self.num / (self.num + 1)) | ||
40 | self.num += 1 | ||
41 | self.avg += dur / float(self.num) | ||
42 | |||
43 | self.begin = 0 | ||
44 | self.next_job = 0 | ||
45 | else: | ||
46 | self.disjoints += 1 | ||
29 | 47 | ||
30 | def start_time(self, record, time = None): | 48 | def start_time(self, record, time = None): |
31 | '''Start duration of time.''' | 49 | '''Start duration of time.''' |
32 | if not time: | 50 | if self.last_record: |
33 | self.begin = record.when | 51 | if not time: |
34 | else: | 52 | self.begin = self.last_record.when |
35 | self.begin = time | 53 | else: |
36 | self.job = record.job | 54 | self.begin = time |
55 | |||
56 | self.next_job = record.job | ||
37 | 57 | ||
38 | # Data stored for each task | 58 | # Data stored for each task |
39 | TaskParams = namedtuple('TaskParams', ['wcet', 'period', 'cpu']) | 59 | TaskParams = namedtuple('TaskParams', ['wcet', 'period', 'cpu']) |
@@ -203,6 +223,12 @@ def create_task_dict(data_dir, work_dir = None): | |||
203 | 223 | ||
204 | return task_dict | 224 | return task_dict |
205 | 225 | ||
226 | LOSS_MSG = """Found task missing more than %d%% of its scheduling records. | ||
227 | These won't be included in scheduling statistics!"""%(100*conf.MAX_RECORD_LOSS) | ||
228 | SKIP_MSG = """Measurement '%s' has no non-zero values. | ||
229 | Measurements like these are not included in scheduling statistics. | ||
230 | If a measurement is missing, this is why.""" | ||
231 | |||
206 | def extract_sched_data(result, data_dir, work_dir): | 232 | def extract_sched_data(result, data_dir, work_dir): |
207 | task_dict = create_task_dict(data_dir, work_dir) | 233 | task_dict = create_task_dict(data_dir, work_dir) |
208 | stat_data = defaultdict(list) | 234 | stat_data = defaultdict(list) |
@@ -213,19 +239,29 @@ def extract_sched_data(result, data_dir, work_dir): | |||
213 | # Currently unknown where these invalid tasks come from... | 239 | # Currently unknown where these invalid tasks come from... |
214 | continue | 240 | continue |
215 | 241 | ||
216 | miss_ratio = float(tdata.misses.num) / tdata.jobs | 242 | miss = tdata.misses |
217 | stat_data["miss-ratio"].append(float(tdata.misses.num) / tdata.jobs) | 243 | |
244 | record_loss = float(miss.disjoints)/(miss.matches + miss.disjoints) | ||
245 | stat_data["record-loss"].append(record_loss) | ||
246 | |||
247 | if record_loss > conf.MAX_RECORD_LOSS: | ||
248 | log_once(LOSS_MSG) | ||
249 | continue | ||
250 | |||
251 | miss_ratio = float(miss.num) / miss.matches | ||
252 | avg_tard = miss.avg * miss_ratio | ||
253 | |||
254 | stat_data["miss-ratio" ].append(miss_ratio) | ||
218 | 255 | ||
219 | stat_data["max-tard" ].append(tdata.misses.max / tdata.params.wcet) | 256 | stat_data["max-tard"].append(miss.max / tdata.params.period) |
220 | # Scale average down to account for jobs with 0 tardiness | 257 | stat_data["avg-tard"].append(avg_tard / tdata.params.period) |
221 | avg_tard = tdata.misses.avg * miss_ratio | ||
222 | stat_data["avg-tard" ].append(avg_tard / tdata.params.wcet) | ||
223 | 258 | ||
224 | stat_data["avg-block" ].append(tdata.blocks.avg / NSEC_PER_MSEC) | 259 | stat_data["avg-block"].append(tdata.blocks.avg / NSEC_PER_MSEC) |
225 | stat_data["max-block" ].append(tdata.blocks.max / NSEC_PER_MSEC) | 260 | stat_data["max-block"].append(tdata.blocks.max / NSEC_PER_MSEC) |
226 | 261 | ||
227 | # Summarize value groups | 262 | # Summarize value groups |
228 | for name, data in stat_data.iteritems(): | 263 | for name, data in stat_data.iteritems(): |
229 | if not data or not sum(data): | 264 | if not data or not sum(data): |
265 | log_once(SKIP_MSG, SKIP_MSG % name) | ||
230 | continue | 266 | continue |
231 | result[name] = Measurement(str(name)).from_array(data) | 267 | result[name] = Measurement(str(name)).from_array(data) |
diff --git a/parse/tuple_table.py b/parse/tuple_table.py index 47fb6b6..320d9dd 100644 --- a/parse/tuple_table.py +++ b/parse/tuple_table.py | |||
@@ -13,6 +13,10 @@ class TupleTable(object): | |||
13 | def get_col_map(self): | 13 | def get_col_map(self): |
14 | return self.col_map | 14 | return self.col_map |
15 | 15 | ||
16 | def __bool__(self): | ||
17 | return bool(self.table) | ||
18 | __nonzero__ = __bool__ | ||
19 | |||
16 | def __getitem__(self, kv): | 20 | def __getitem__(self, kv): |
17 | key = self.col_map.get_key(kv) | 21 | key = self.col_map.get_key(kv) |
18 | return self.table[key] | 22 | return self.table[key] |
diff --git a/parse_exps.py b/parse_exps.py index c2cbedb..cc4372a 100755 --- a/parse_exps.py +++ b/parse_exps.py | |||
@@ -1,6 +1,8 @@ | |||
1 | #!/usr/bin/env python | 1 | #!/usr/bin/env python |
2 | from __future__ import print_function | 2 | from __future__ import print_function |
3 | 3 | ||
4 | import common as com | ||
5 | import multiprocessing | ||
4 | import os | 6 | import os |
5 | import parse.ft as ft | 7 | import parse.ft as ft |
6 | import parse.sched as st | 8 | import parse.sched as st |
@@ -10,13 +12,12 @@ import sys | |||
10 | import traceback | 12 | import traceback |
11 | 13 | ||
12 | from collections import namedtuple | 14 | from collections import namedtuple |
13 | from common import load_params | ||
14 | from config.config import DEFAULTS,PARAMS | 15 | from config.config import DEFAULTS,PARAMS |
15 | from optparse import OptionParser | 16 | from optparse import OptionParser |
16 | from parse.point import ExpPoint | 17 | from parse.point import ExpPoint |
17 | from parse.tuple_table import TupleTable | 18 | from parse.tuple_table import TupleTable |
18 | from parse.col_map import ColMapBuilder | 19 | from parse.col_map import ColMapBuilder |
19 | from multiprocessing import Pool, cpu_count | 20 | |
20 | 21 | ||
21 | def parse_args(): | 22 | def parse_args(): |
22 | parser = OptionParser("usage: %prog [options] [data_dir]...") | 23 | parser = OptionParser("usage: %prog [options] [data_dir]...") |
@@ -33,18 +34,60 @@ def parse_args(): | |||
33 | parser.add_option('-m', '--write-map', action='store_true', default=False, | 34 | parser.add_option('-m', '--write-map', action='store_true', default=False, |
34 | dest='write_map', | 35 | dest='write_map', |
35 | help='Output map of values instead of csv tree') | 36 | help='Output map of values instead of csv tree') |
36 | parser.add_option('-p', '--processors', default=max(cpu_count() - 1, 1), | 37 | parser.add_option('-p', '--processors', |
38 | default=max(multiprocessing.cpu_count() - 1, 1), | ||
37 | type='int', dest='processors', | 39 | type='int', dest='processors', |
38 | help='number of threads for processing') | 40 | help='number of threads for processing') |
39 | 41 | ||
40 | return parser.parse_args() | 42 | return parser.parse_args() |
41 | 43 | ||
44 | |||
42 | ExpData = namedtuple('ExpData', ['path', 'params', 'work_dir']) | 45 | ExpData = namedtuple('ExpData', ['path', 'params', 'work_dir']) |
43 | 46 | ||
47 | |||
48 | def parse_exp(exp_force): | ||
49 | # Tupled for multiprocessing | ||
50 | exp, force = exp_force | ||
51 | |||
52 | result_file = exp.work_dir + "/exp_point.pkl" | ||
53 | should_load = not force and os.path.exists(result_file) | ||
54 | |||
55 | result = None | ||
56 | if should_load: | ||
57 | with open(result_file, 'rb') as f: | ||
58 | try: | ||
59 | # No need to go through this work twice | ||
60 | result = pickle.load(f) | ||
61 | except: | ||
62 | pass | ||
63 | |||
64 | if not result: | ||
65 | try: | ||
66 | # Create a readable name | ||
67 | name = os.path.relpath(exp.path) | ||
68 | name = name if name != "." else os.path.split(os.getcwd())[1] | ||
69 | |||
70 | result = ExpPoint(name) | ||
71 | |||
72 | # Write overheads into result | ||
73 | cycles = exp.params[PARAMS['cycles']] | ||
74 | ft.extract_ft_data(result, exp.path, exp.work_dir, cycles) | ||
75 | |||
76 | # Write scheduling statistics into result | ||
77 | st.extract_sched_data(result, exp.path, exp.work_dir) | ||
78 | |||
79 | with open(result_file, 'wb') as f: | ||
80 | pickle.dump(result, f) | ||
81 | except: | ||
82 | traceback.print_exc() | ||
83 | |||
84 | return (exp, result) | ||
85 | |||
86 | |||
44 | def get_exp_params(data_dir, cm_builder): | 87 | def get_exp_params(data_dir, cm_builder): |
45 | param_file = "%s/%s" % (data_dir, DEFAULTS['params_file']) | 88 | param_file = "%s/%s" % (data_dir, DEFAULTS['params_file']) |
46 | if os.path.isfile(param_file): | 89 | if os.path.isfile(param_file): |
47 | params = load_params(param_file) | 90 | params = com.load_params(param_file) |
48 | 91 | ||
49 | # Store parameters in cm_builder, which will track which parameters change | 92 | # Store parameters in cm_builder, which will track which parameters change |
50 | # across experiments | 93 | # across experiments |
@@ -83,41 +126,8 @@ def load_exps(exp_dirs, cm_builder, force): | |||
83 | 126 | ||
84 | return exps | 127 | return exps |
85 | 128 | ||
86 | def parse_exp(exp_force): | ||
87 | # Tupled for multiprocessing | ||
88 | exp, force = exp_force | ||
89 | |||
90 | result_file = exp.work_dir + "/exp_point.pkl" | ||
91 | should_load = not force and os.path.exists(result_file) | ||
92 | |||
93 | result = None | ||
94 | if should_load: | ||
95 | with open(result_file, 'rb') as f: | ||
96 | try: | ||
97 | # No need to go through this work twice | ||
98 | result = pickle.load(f) | ||
99 | except: | ||
100 | pass | ||
101 | 129 | ||
102 | if not result: | 130 | def get_dirs(args): |
103 | try: | ||
104 | result = ExpPoint(exp.path) | ||
105 | cycles = exp.params[PARAMS['cycles']] | ||
106 | |||
107 | # Write overheads into result | ||
108 | ft.extract_ft_data(result, exp.path, exp.work_dir, cycles) | ||
109 | |||
110 | # Write scheduling statistics into result | ||
111 | st.extract_sched_data(result, exp.path, exp.work_dir) | ||
112 | |||
113 | with open(result_file, 'wb') as f: | ||
114 | pickle.dump(result, f) | ||
115 | except: | ||
116 | traceback.print_exc() | ||
117 | |||
118 | return (exp, result) | ||
119 | |||
120 | def get_exps(args): | ||
121 | if args: | 131 | if args: |
122 | return args | 132 | return args |
123 | elif os.path.exists(DEFAULTS['out-run']): | 133 | elif os.path.exists(DEFAULTS['out-run']): |
@@ -128,38 +138,32 @@ def get_exps(args): | |||
128 | sys.stderr.write("Reading data from current directory.\n") | 138 | sys.stderr.write("Reading data from current directory.\n") |
129 | return [os.getcwd()] | 139 | return [os.getcwd()] |
130 | 140 | ||
131 | def main(): | ||
132 | opts, args = parse_args() | ||
133 | exp_dirs = get_exps(args) | ||
134 | |||
135 | # Load exp parameters into a ColMap | ||
136 | builder = ColMapBuilder() | ||
137 | exps = load_exps(exp_dirs, builder, opts.force) | ||
138 | 141 | ||
139 | # Don't track changes in ignored parameters | 142 | def fill_table(table, exps, opts): |
140 | if opts.ignore: | 143 | sys.stderr.write("Parsing data...\n") |
141 | for param in opts.ignore.split(","): | ||
142 | builder.try_remove(param) | ||
143 | builder.try_remove(PARAMS['trial']) # Always average multiple trials | ||
144 | builder.try_remove(PARAMS['cycles']) # Only need for feather-trace parsing | ||
145 | 144 | ||
146 | col_map = builder.build() | 145 | procs = min(len(exps), opts.processors) |
147 | result_table = TupleTable(col_map) | 146 | logged = multiprocessing.Manager().list() |
148 | 147 | ||
149 | sys.stderr.write("Parsing data...\n") | 148 | pool = multiprocessing.Pool(processes=procs, |
149 | # Share a list of previously logged messages amongst processes | ||
150 | # This is for the com.log_once method to use | ||
151 | initializer=com.set_logged_list, initargs=(logged,)) | ||
150 | 152 | ||
151 | procs = min(len(exps), opts.processors) | ||
152 | pool = Pool(processes=procs) | ||
153 | pool_args = zip(exps, [opts.force]*len(exps)) | 153 | pool_args = zip(exps, [opts.force]*len(exps)) |
154 | enum = pool.imap_unordered(parse_exp, pool_args, 1) | 154 | enum = pool.imap_unordered(parse_exp, pool_args, 1) |
155 | 155 | ||
156 | try: | 156 | try: |
157 | for i, (exp, result) in enumerate(enum): | 157 | for i, (exp, result) in enumerate(enum): |
158 | if not result: | ||
159 | continue | ||
160 | |||
158 | if opts.verbose: | 161 | if opts.verbose: |
159 | print(result) | 162 | print(result) |
160 | else: | 163 | else: |
161 | sys.stderr.write('\r {0:.2%}'.format(float(i)/len(exps))) | 164 | sys.stderr.write('\r {0:.2%}'.format(float(i)/len(exps))) |
162 | result_table[exp.params] += [result] | 165 | table[exp.params] += [result] |
166 | |||
163 | pool.close() | 167 | pool.close() |
164 | except: | 168 | except: |
165 | pool.terminate() | 169 | pool.terminate() |
@@ -170,16 +174,17 @@ def main(): | |||
170 | 174 | ||
171 | sys.stderr.write('\n') | 175 | sys.stderr.write('\n') |
172 | 176 | ||
173 | if opts.force and os.path.exists(opts.out): | ||
174 | sh.rmtree(opts.out) | ||
175 | 177 | ||
176 | reduced_table = result_table.reduce() | 178 | def write_output(table, opts): |
179 | reduced_table = table.reduce() | ||
177 | 180 | ||
178 | if opts.write_map: | 181 | if opts.write_map: |
179 | sys.stderr.write("Writing python map into %s...\n" % opts.out) | 182 | sys.stderr.write("Writing python map into %s...\n" % opts.out) |
180 | # Write summarized results into map | ||
181 | reduced_table.write_map(opts.out) | 183 | reduced_table.write_map(opts.out) |
182 | else: | 184 | else: |
185 | if opts.force and os.path.exists(opts.out): | ||
186 | sh.rmtree(opts.out) | ||
187 | |||
183 | # Write out csv directories for all variable params | 188 | # Write out csv directories for all variable params |
184 | dir_map = reduced_table.to_dir_map() | 189 | dir_map = reduced_table.to_dir_map() |
185 | 190 | ||
@@ -188,12 +193,42 @@ def main(): | |||
188 | if not opts.verbose: | 193 | if not opts.verbose: |
189 | sys.stderr.write("Too little data to make csv files, " + | 194 | sys.stderr.write("Too little data to make csv files, " + |
190 | "printing results.\n") | 195 | "printing results.\n") |
191 | for key, exp in result_table: | 196 | for key, exp in table: |
192 | for e in exp: | 197 | for e in exp: |
193 | print(e) | 198 | print(e) |
194 | else: | 199 | else: |
195 | sys.stderr.write("Writing csvs into %s...\n" % opts.out) | 200 | sys.stderr.write("Writing csvs into %s...\n" % opts.out) |
196 | dir_map.write(opts.out) | 201 | dir_map.write(opts.out) |
197 | 202 | ||
203 | |||
204 | def main(): | ||
205 | opts, args = parse_args() | ||
206 | exp_dirs = get_dirs(args) | ||
207 | |||
208 | # Load experiment parameters into a ColMap | ||
209 | builder = ColMapBuilder() | ||
210 | exps = load_exps(exp_dirs, builder, opts.force) | ||
211 | |||
212 | # Don't track changes in ignored parameters | ||
213 | if opts.ignore: | ||
214 | for param in opts.ignore.split(","): | ||
215 | builder.try_remove(param) | ||
216 | |||
217 | # Always average multiple trials | ||
218 | builder.try_remove(PARAMS['trial']) | ||
219 | # Only need this for feather-trace parsing | ||
220 | builder.try_remove(PARAMS['cycles']) | ||
221 | |||
222 | col_map = builder.build() | ||
223 | table = TupleTable(col_map) | ||
224 | |||
225 | fill_table(table, exps, opts) | ||
226 | |||
227 | if not table: | ||
228 | sys.stderr.write("Found no data to parse!") | ||
229 | sys.exit(1) | ||
230 | |||
231 | write_output(table, opts) | ||
232 | |||
198 | if __name__ == '__main__': | 233 | if __name__ == '__main__': |
199 | main() | 234 | main() |