diff options
Diffstat (limited to 'lib/debugobjects.c')
-rw-r--r-- | lib/debugobjects.c | 115 |
1 files changed, 91 insertions, 24 deletions
diff --git a/lib/debugobjects.c b/lib/debugobjects.c index 2ac42286cd08..38c23b528f6f 100644 --- a/lib/debugobjects.c +++ b/lib/debugobjects.c | |||
@@ -25,6 +25,7 @@ | |||
25 | 25 | ||
26 | #define ODEBUG_POOL_SIZE 1024 | 26 | #define ODEBUG_POOL_SIZE 1024 |
27 | #define ODEBUG_POOL_MIN_LEVEL 256 | 27 | #define ODEBUG_POOL_MIN_LEVEL 256 |
28 | #define ODEBUG_POOL_PERCPU_SIZE 64 | ||
28 | 29 | ||
29 | #define ODEBUG_CHUNK_SHIFT PAGE_SHIFT | 30 | #define ODEBUG_CHUNK_SHIFT PAGE_SHIFT |
30 | #define ODEBUG_CHUNK_SIZE (1 << ODEBUG_CHUNK_SHIFT) | 31 | #define ODEBUG_CHUNK_SIZE (1 << ODEBUG_CHUNK_SHIFT) |
@@ -35,6 +36,17 @@ struct debug_bucket { | |||
35 | raw_spinlock_t lock; | 36 | raw_spinlock_t lock; |
36 | }; | 37 | }; |
37 | 38 | ||
39 | /* | ||
40 | * Debug object percpu free list | ||
41 | * Access is protected by disabling irq | ||
42 | */ | ||
43 | struct debug_percpu_free { | ||
44 | struct hlist_head free_objs; | ||
45 | int obj_free; | ||
46 | }; | ||
47 | |||
48 | static DEFINE_PER_CPU(struct debug_percpu_free, percpu_obj_pool); | ||
49 | |||
38 | static struct debug_bucket obj_hash[ODEBUG_HASH_SIZE]; | 50 | static struct debug_bucket obj_hash[ODEBUG_HASH_SIZE]; |
39 | 51 | ||
40 | static struct debug_obj obj_static_pool[ODEBUG_POOL_SIZE] __initdata; | 52 | static struct debug_obj obj_static_pool[ODEBUG_POOL_SIZE] __initdata; |
@@ -44,13 +56,19 @@ static DEFINE_RAW_SPINLOCK(pool_lock); | |||
44 | static HLIST_HEAD(obj_pool); | 56 | static HLIST_HEAD(obj_pool); |
45 | static HLIST_HEAD(obj_to_free); | 57 | static HLIST_HEAD(obj_to_free); |
46 | 58 | ||
59 | /* | ||
60 | * Because of the presence of percpu free pools, obj_pool_free will | ||
61 | * under-count those in the percpu free pools. Similarly, obj_pool_used | ||
62 | * will over-count those in the percpu free pools. Adjustments will be | ||
63 | * made at debug_stats_show(). Both obj_pool_min_free and obj_pool_max_used | ||
64 | * can be off. | ||
65 | */ | ||
47 | static int obj_pool_min_free = ODEBUG_POOL_SIZE; | 66 | static int obj_pool_min_free = ODEBUG_POOL_SIZE; |
48 | static int obj_pool_free = ODEBUG_POOL_SIZE; | 67 | static int obj_pool_free = ODEBUG_POOL_SIZE; |
49 | static int obj_pool_used; | 68 | static int obj_pool_used; |
50 | static int obj_pool_max_used; | 69 | static int obj_pool_max_used; |
51 | /* The number of objs on the global free list */ | 70 | /* The number of objs on the global free list */ |
52 | static int obj_nr_tofree; | 71 | static int obj_nr_tofree; |
53 | static struct kmem_cache *obj_cache; | ||
54 | 72 | ||
55 | static int debug_objects_maxchain __read_mostly; | 73 | static int debug_objects_maxchain __read_mostly; |
56 | static int __maybe_unused debug_objects_maxchecked __read_mostly; | 74 | static int __maybe_unused debug_objects_maxchecked __read_mostly; |
@@ -63,6 +81,7 @@ static int debug_objects_pool_size __read_mostly | |||
63 | static int debug_objects_pool_min_level __read_mostly | 81 | static int debug_objects_pool_min_level __read_mostly |
64 | = ODEBUG_POOL_MIN_LEVEL; | 82 | = ODEBUG_POOL_MIN_LEVEL; |
65 | static struct debug_obj_descr *descr_test __read_mostly; | 83 | static struct debug_obj_descr *descr_test __read_mostly; |
84 | static struct kmem_cache *obj_cache __read_mostly; | ||
66 | 85 | ||
67 | /* | 86 | /* |
68 | * Track numbers of kmem_cache_alloc()/free() calls done. | 87 | * Track numbers of kmem_cache_alloc()/free() calls done. |
@@ -163,26 +182,42 @@ static struct debug_obj *lookup_object(void *addr, struct debug_bucket *b) | |||
163 | } | 182 | } |
164 | 183 | ||
165 | /* | 184 | /* |
185 | * Allocate a new object from the hlist | ||
186 | */ | ||
187 | static struct debug_obj *__alloc_object(struct hlist_head *list) | ||
188 | { | ||
189 | struct debug_obj *obj = NULL; | ||
190 | |||
191 | if (list->first) { | ||
192 | obj = hlist_entry(list->first, typeof(*obj), node); | ||
193 | hlist_del(&obj->node); | ||
194 | } | ||
195 | |||
196 | return obj; | ||
197 | } | ||
198 | |||
199 | /* | ||
166 | * Allocate a new object. If the pool is empty, switch off the debugger. | 200 | * Allocate a new object. If the pool is empty, switch off the debugger. |
167 | * Must be called with interrupts disabled. | 201 | * Must be called with interrupts disabled. |
168 | */ | 202 | */ |
169 | static struct debug_obj * | 203 | static struct debug_obj * |
170 | alloc_object(void *addr, struct debug_bucket *b, struct debug_obj_descr *descr) | 204 | alloc_object(void *addr, struct debug_bucket *b, struct debug_obj_descr *descr) |
171 | { | 205 | { |
172 | struct debug_obj *obj = NULL; | 206 | struct debug_percpu_free *percpu_pool; |
173 | 207 | struct debug_obj *obj; | |
174 | raw_spin_lock(&pool_lock); | ||
175 | if (obj_pool.first) { | ||
176 | obj = hlist_entry(obj_pool.first, typeof(*obj), node); | ||
177 | |||
178 | obj->object = addr; | ||
179 | obj->descr = descr; | ||
180 | obj->state = ODEBUG_STATE_NONE; | ||
181 | obj->astate = 0; | ||
182 | hlist_del(&obj->node); | ||
183 | 208 | ||
184 | hlist_add_head(&obj->node, &b->list); | 209 | if (likely(obj_cache)) { |
210 | percpu_pool = this_cpu_ptr(&percpu_obj_pool); | ||
211 | obj = __alloc_object(&percpu_pool->free_objs); | ||
212 | if (obj) { | ||
213 | percpu_pool->obj_free--; | ||
214 | goto init_obj; | ||
215 | } | ||
216 | } | ||
185 | 217 | ||
218 | raw_spin_lock(&pool_lock); | ||
219 | obj = __alloc_object(&obj_pool); | ||
220 | if (obj) { | ||
186 | obj_pool_used++; | 221 | obj_pool_used++; |
187 | if (obj_pool_used > obj_pool_max_used) | 222 | if (obj_pool_used > obj_pool_max_used) |
188 | obj_pool_max_used = obj_pool_used; | 223 | obj_pool_max_used = obj_pool_used; |
@@ -193,6 +228,14 @@ alloc_object(void *addr, struct debug_bucket *b, struct debug_obj_descr *descr) | |||
193 | } | 228 | } |
194 | raw_spin_unlock(&pool_lock); | 229 | raw_spin_unlock(&pool_lock); |
195 | 230 | ||
231 | init_obj: | ||
232 | if (obj) { | ||
233 | obj->object = addr; | ||
234 | obj->descr = descr; | ||
235 | obj->state = ODEBUG_STATE_NONE; | ||
236 | obj->astate = 0; | ||
237 | hlist_add_head(&obj->node, &b->list); | ||
238 | } | ||
196 | return obj; | 239 | return obj; |
197 | } | 240 | } |
198 | 241 | ||
@@ -247,8 +290,21 @@ static bool __free_object(struct debug_obj *obj) | |||
247 | { | 290 | { |
248 | unsigned long flags; | 291 | unsigned long flags; |
249 | bool work; | 292 | bool work; |
293 | struct debug_percpu_free *percpu_pool; | ||
250 | 294 | ||
251 | raw_spin_lock_irqsave(&pool_lock, flags); | 295 | local_irq_save(flags); |
296 | /* | ||
297 | * Try to free it into the percpu pool first. | ||
298 | */ | ||
299 | percpu_pool = this_cpu_ptr(&percpu_obj_pool); | ||
300 | if (obj_cache && percpu_pool->obj_free < ODEBUG_POOL_PERCPU_SIZE) { | ||
301 | hlist_add_head(&obj->node, &percpu_pool->free_objs); | ||
302 | percpu_pool->obj_free++; | ||
303 | local_irq_restore(flags); | ||
304 | return false; | ||
305 | } | ||
306 | |||
307 | raw_spin_lock(&pool_lock); | ||
252 | work = (obj_pool_free > debug_objects_pool_size) && obj_cache; | 308 | work = (obj_pool_free > debug_objects_pool_size) && obj_cache; |
253 | obj_pool_used--; | 309 | obj_pool_used--; |
254 | 310 | ||
@@ -259,7 +315,8 @@ static bool __free_object(struct debug_obj *obj) | |||
259 | obj_pool_free++; | 315 | obj_pool_free++; |
260 | hlist_add_head(&obj->node, &obj_pool); | 316 | hlist_add_head(&obj->node, &obj_pool); |
261 | } | 317 | } |
262 | raw_spin_unlock_irqrestore(&pool_lock, flags); | 318 | raw_spin_unlock(&pool_lock); |
319 | local_irq_restore(flags); | ||
263 | return work; | 320 | return work; |
264 | } | 321 | } |
265 | 322 | ||
@@ -822,13 +879,19 @@ void debug_check_no_obj_freed(const void *address, unsigned long size) | |||
822 | 879 | ||
823 | static int debug_stats_show(struct seq_file *m, void *v) | 880 | static int debug_stats_show(struct seq_file *m, void *v) |
824 | { | 881 | { |
882 | int cpu, obj_percpu_free = 0; | ||
883 | |||
884 | for_each_possible_cpu(cpu) | ||
885 | obj_percpu_free += per_cpu(percpu_obj_pool.obj_free, cpu); | ||
886 | |||
825 | seq_printf(m, "max_chain :%d\n", debug_objects_maxchain); | 887 | seq_printf(m, "max_chain :%d\n", debug_objects_maxchain); |
826 | seq_printf(m, "max_checked :%d\n", debug_objects_maxchecked); | 888 | seq_printf(m, "max_checked :%d\n", debug_objects_maxchecked); |
827 | seq_printf(m, "warnings :%d\n", debug_objects_warnings); | 889 | seq_printf(m, "warnings :%d\n", debug_objects_warnings); |
828 | seq_printf(m, "fixups :%d\n", debug_objects_fixups); | 890 | seq_printf(m, "fixups :%d\n", debug_objects_fixups); |
829 | seq_printf(m, "pool_free :%d\n", obj_pool_free); | 891 | seq_printf(m, "pool_free :%d\n", obj_pool_free + obj_percpu_free); |
892 | seq_printf(m, "pool_pcp_free :%d\n", obj_percpu_free); | ||
830 | seq_printf(m, "pool_min_free :%d\n", obj_pool_min_free); | 893 | seq_printf(m, "pool_min_free :%d\n", obj_pool_min_free); |
831 | seq_printf(m, "pool_used :%d\n", obj_pool_used); | 894 | seq_printf(m, "pool_used :%d\n", obj_pool_used - obj_percpu_free); |
832 | seq_printf(m, "pool_max_used :%d\n", obj_pool_max_used); | 895 | seq_printf(m, "pool_max_used :%d\n", obj_pool_max_used); |
833 | seq_printf(m, "on_free_list :%d\n", obj_nr_tofree); | 896 | seq_printf(m, "on_free_list :%d\n", obj_nr_tofree); |
834 | seq_printf(m, "objs_allocated:%d\n", debug_objects_allocated); | 897 | seq_printf(m, "objs_allocated:%d\n", debug_objects_allocated); |
@@ -1165,9 +1228,20 @@ free: | |||
1165 | */ | 1228 | */ |
1166 | void __init debug_objects_mem_init(void) | 1229 | void __init debug_objects_mem_init(void) |
1167 | { | 1230 | { |
1231 | int cpu; | ||
1232 | |||
1168 | if (!debug_objects_enabled) | 1233 | if (!debug_objects_enabled) |
1169 | return; | 1234 | return; |
1170 | 1235 | ||
1236 | /* | ||
1237 | * Initialize the percpu object pools | ||
1238 | * | ||
1239 | * Initialization is not strictly necessary, but was done for | ||
1240 | * completeness. | ||
1241 | */ | ||
1242 | for_each_possible_cpu(cpu) | ||
1243 | INIT_HLIST_HEAD(&per_cpu(percpu_obj_pool.free_objs, cpu)); | ||
1244 | |||
1171 | obj_cache = kmem_cache_create("debug_objects_cache", | 1245 | obj_cache = kmem_cache_create("debug_objects_cache", |
1172 | sizeof (struct debug_obj), 0, | 1246 | sizeof (struct debug_obj), 0, |
1173 | SLAB_DEBUG_OBJECTS | SLAB_NOLEAKTRACE, | 1247 | SLAB_DEBUG_OBJECTS | SLAB_NOLEAKTRACE, |
@@ -1179,11 +1253,4 @@ void __init debug_objects_mem_init(void) | |||
1179 | pr_warn("out of memory.\n"); | 1253 | pr_warn("out of memory.\n"); |
1180 | } else | 1254 | } else |
1181 | debug_objects_selftest(); | 1255 | debug_objects_selftest(); |
1182 | |||
1183 | /* | ||
1184 | * Increase the thresholds for allocating and freeing objects | ||
1185 | * according to the number of possible CPUs available in the system. | ||
1186 | */ | ||
1187 | debug_objects_pool_size += num_possible_cpus() * 32; | ||
1188 | debug_objects_pool_min_level += num_possible_cpus() * 4; | ||
1189 | } | 1256 | } |