aboutsummaryrefslogtreecommitdiffstats
path: root/kernel/module.c
diff options
context:
space:
mode:
Diffstat (limited to 'kernel/module.c')
-rw-r--r--kernel/module.c159
1 files changed, 105 insertions, 54 deletions
diff --git a/kernel/module.c b/kernel/module.c
index c968d3606dca..e2564580f3f1 100644
--- a/kernel/module.c
+++ b/kernel/module.c
@@ -59,8 +59,6 @@
59#define CREATE_TRACE_POINTS 59#define CREATE_TRACE_POINTS
60#include <trace/events/module.h> 60#include <trace/events/module.h>
61 61
62EXPORT_TRACEPOINT_SYMBOL(module_get);
63
64#if 0 62#if 0
65#define DEBUGP printk 63#define DEBUGP printk
66#else 64#else
@@ -370,27 +368,33 @@ EXPORT_SYMBOL_GPL(find_module);
370 368
371#ifdef CONFIG_SMP 369#ifdef CONFIG_SMP
372 370
373static void *percpu_modalloc(unsigned long size, unsigned long align, 371static inline void __percpu *mod_percpu(struct module *mod)
374 const char *name)
375{ 372{
376 void *ptr; 373 return mod->percpu;
374}
377 375
376static int percpu_modalloc(struct module *mod,
377 unsigned long size, unsigned long align)
378{
378 if (align > PAGE_SIZE) { 379 if (align > PAGE_SIZE) {
379 printk(KERN_WARNING "%s: per-cpu alignment %li > %li\n", 380 printk(KERN_WARNING "%s: per-cpu alignment %li > %li\n",
380 name, align, PAGE_SIZE); 381 mod->name, align, PAGE_SIZE);
381 align = PAGE_SIZE; 382 align = PAGE_SIZE;
382 } 383 }
383 384
384 ptr = __alloc_reserved_percpu(size, align); 385 mod->percpu = __alloc_reserved_percpu(size, align);
385 if (!ptr) 386 if (!mod->percpu) {
386 printk(KERN_WARNING 387 printk(KERN_WARNING
387 "Could not allocate %lu bytes percpu data\n", size); 388 "Could not allocate %lu bytes percpu data\n", size);
388 return ptr; 389 return -ENOMEM;
390 }
391 mod->percpu_size = size;
392 return 0;
389} 393}
390 394
391static void percpu_modfree(void *freeme) 395static void percpu_modfree(struct module *mod)
392{ 396{
393 free_percpu(freeme); 397 free_percpu(mod->percpu);
394} 398}
395 399
396static unsigned int find_pcpusec(Elf_Ehdr *hdr, 400static unsigned int find_pcpusec(Elf_Ehdr *hdr,
@@ -400,24 +404,62 @@ static unsigned int find_pcpusec(Elf_Ehdr *hdr,
400 return find_sec(hdr, sechdrs, secstrings, ".data.percpu"); 404 return find_sec(hdr, sechdrs, secstrings, ".data.percpu");
401} 405}
402 406
403static void percpu_modcopy(void *pcpudest, const void *from, unsigned long size) 407static void percpu_modcopy(struct module *mod,
408 const void *from, unsigned long size)
404{ 409{
405 int cpu; 410 int cpu;
406 411
407 for_each_possible_cpu(cpu) 412 for_each_possible_cpu(cpu)
408 memcpy(pcpudest + per_cpu_offset(cpu), from, size); 413 memcpy(per_cpu_ptr(mod->percpu, cpu), from, size);
414}
415
416/**
417 * is_module_percpu_address - test whether address is from module static percpu
418 * @addr: address to test
419 *
420 * Test whether @addr belongs to module static percpu area.
421 *
422 * RETURNS:
423 * %true if @addr is from module static percpu area
424 */
425bool is_module_percpu_address(unsigned long addr)
426{
427 struct module *mod;
428 unsigned int cpu;
429
430 preempt_disable();
431
432 list_for_each_entry_rcu(mod, &modules, list) {
433 if (!mod->percpu_size)
434 continue;
435 for_each_possible_cpu(cpu) {
436 void *start = per_cpu_ptr(mod->percpu, cpu);
437
438 if ((void *)addr >= start &&
439 (void *)addr < start + mod->percpu_size) {
440 preempt_enable();
441 return true;
442 }
443 }
444 }
445
446 preempt_enable();
447 return false;
409} 448}
410 449
411#else /* ... !CONFIG_SMP */ 450#else /* ... !CONFIG_SMP */
412 451
413static inline void *percpu_modalloc(unsigned long size, unsigned long align, 452static inline void __percpu *mod_percpu(struct module *mod)
414 const char *name)
415{ 453{
416 return NULL; 454 return NULL;
417} 455}
418static inline void percpu_modfree(void *pcpuptr) 456static inline int percpu_modalloc(struct module *mod,
457 unsigned long size, unsigned long align)
458{
459 return -ENOMEM;
460}
461static inline void percpu_modfree(struct module *mod)
419{ 462{
420 BUG();
421} 463}
422static inline unsigned int find_pcpusec(Elf_Ehdr *hdr, 464static inline unsigned int find_pcpusec(Elf_Ehdr *hdr,
423 Elf_Shdr *sechdrs, 465 Elf_Shdr *sechdrs,
@@ -425,12 +467,16 @@ static inline unsigned int find_pcpusec(Elf_Ehdr *hdr,
425{ 467{
426 return 0; 468 return 0;
427} 469}
428static inline void percpu_modcopy(void *pcpudst, const void *src, 470static inline void percpu_modcopy(struct module *mod,
429 unsigned long size) 471 const void *from, unsigned long size)
430{ 472{
431 /* pcpusec should be 0, and size of that section should be 0. */ 473 /* pcpusec should be 0, and size of that section should be 0. */
432 BUG_ON(size != 0); 474 BUG_ON(size != 0);
433} 475}
476bool is_module_percpu_address(unsigned long addr)
477{
478 return false;
479}
434 480
435#endif /* CONFIG_SMP */ 481#endif /* CONFIG_SMP */
436 482
@@ -467,17 +513,22 @@ MODINFO_ATTR(srcversion);
467static char last_unloaded_module[MODULE_NAME_LEN+1]; 513static char last_unloaded_module[MODULE_NAME_LEN+1];
468 514
469#ifdef CONFIG_MODULE_UNLOAD 515#ifdef CONFIG_MODULE_UNLOAD
516
517EXPORT_TRACEPOINT_SYMBOL(module_get);
518
470/* Init the unload section of the module. */ 519/* Init the unload section of the module. */
471static void module_unload_init(struct module *mod) 520static void module_unload_init(struct module *mod)
472{ 521{
473 int cpu; 522 int cpu;
474 523
475 INIT_LIST_HEAD(&mod->modules_which_use_me); 524 INIT_LIST_HEAD(&mod->modules_which_use_me);
476 for_each_possible_cpu(cpu) 525 for_each_possible_cpu(cpu) {
477 per_cpu_ptr(mod->refptr, cpu)->count = 0; 526 per_cpu_ptr(mod->refptr, cpu)->incs = 0;
527 per_cpu_ptr(mod->refptr, cpu)->decs = 0;
528 }
478 529
479 /* Hold reference count during initialization. */ 530 /* Hold reference count during initialization. */
480 __this_cpu_write(mod->refptr->count, 1); 531 __this_cpu_write(mod->refptr->incs, 1);
481 /* Backwards compatibility macros put refcount during init. */ 532 /* Backwards compatibility macros put refcount during init. */
482 mod->waiter = current; 533 mod->waiter = current;
483} 534}
@@ -616,12 +667,28 @@ static int try_stop_module(struct module *mod, int flags, int *forced)
616 667
617unsigned int module_refcount(struct module *mod) 668unsigned int module_refcount(struct module *mod)
618{ 669{
619 unsigned int total = 0; 670 unsigned int incs = 0, decs = 0;
620 int cpu; 671 int cpu;
621 672
622 for_each_possible_cpu(cpu) 673 for_each_possible_cpu(cpu)
623 total += per_cpu_ptr(mod->refptr, cpu)->count; 674 decs += per_cpu_ptr(mod->refptr, cpu)->decs;
624 return total; 675 /*
676 * ensure the incs are added up after the decs.
677 * module_put ensures incs are visible before decs with smp_wmb.
678 *
679 * This 2-count scheme avoids the situation where the refcount
680 * for CPU0 is read, then CPU0 increments the module refcount,
681 * then CPU1 drops that refcount, then the refcount for CPU1 is
682 * read. We would record a decrement but not its corresponding
683 * increment so we would see a low count (disaster).
684 *
685 * Rare situation? But module_refcount can be preempted, and we
686 * might be tallying up 4096+ CPUs. So it is not impossible.
687 */
688 smp_rmb();
689 for_each_possible_cpu(cpu)
690 incs += per_cpu_ptr(mod->refptr, cpu)->incs;
691 return incs - decs;
625} 692}
626EXPORT_SYMBOL(module_refcount); 693EXPORT_SYMBOL(module_refcount);
627 694
@@ -657,16 +724,8 @@ SYSCALL_DEFINE2(delete_module, const char __user *, name_user,
657 return -EFAULT; 724 return -EFAULT;
658 name[MODULE_NAME_LEN-1] = '\0'; 725 name[MODULE_NAME_LEN-1] = '\0';
659 726
660 /* Create stop_machine threads since free_module relies on 727 if (mutex_lock_interruptible(&module_mutex) != 0)
661 * a non-failing stop_machine call. */ 728 return -EINTR;
662 ret = stop_machine_create();
663 if (ret)
664 return ret;
665
666 if (mutex_lock_interruptible(&module_mutex) != 0) {
667 ret = -EINTR;
668 goto out_stop;
669 }
670 729
671 mod = find_module(name); 730 mod = find_module(name);
672 if (!mod) { 731 if (!mod) {
@@ -726,8 +785,6 @@ SYSCALL_DEFINE2(delete_module, const char __user *, name_user,
726 785
727 out: 786 out:
728 mutex_unlock(&module_mutex); 787 mutex_unlock(&module_mutex);
729out_stop:
730 stop_machine_destroy();
731 return ret; 788 return ret;
732} 789}
733 790
@@ -798,10 +855,10 @@ void module_put(struct module *module)
798{ 855{
799 if (module) { 856 if (module) {
800 preempt_disable(); 857 preempt_disable();
801 __this_cpu_dec(module->refptr->count); 858 smp_wmb(); /* see comment in module_refcount */
859 __this_cpu_inc(module->refptr->decs);
802 860
803 trace_module_put(module, _RET_IP_, 861 trace_module_put(module, _RET_IP_);
804 __this_cpu_read(module->refptr->count));
805 /* Maybe they're waiting for us to drop reference? */ 862 /* Maybe they're waiting for us to drop reference? */
806 if (unlikely(!module_is_live(module))) 863 if (unlikely(!module_is_live(module)))
807 wake_up_process(module->waiter); 864 wake_up_process(module->waiter);
@@ -1400,8 +1457,7 @@ static void free_module(struct module *mod)
1400 /* This may be NULL, but that's OK */ 1457 /* This may be NULL, but that's OK */
1401 module_free(mod, mod->module_init); 1458 module_free(mod, mod->module_init);
1402 kfree(mod->args); 1459 kfree(mod->args);
1403 if (mod->percpu) 1460 percpu_modfree(mod);
1404 percpu_modfree(mod->percpu);
1405#if defined(CONFIG_MODULE_UNLOAD) 1461#if defined(CONFIG_MODULE_UNLOAD)
1406 if (mod->refptr) 1462 if (mod->refptr)
1407 free_percpu(mod->refptr); 1463 free_percpu(mod->refptr);
@@ -1520,7 +1576,7 @@ static int simplify_symbols(Elf_Shdr *sechdrs,
1520 default: 1576 default:
1521 /* Divert to percpu allocation if a percpu var. */ 1577 /* Divert to percpu allocation if a percpu var. */
1522 if (sym[i].st_shndx == pcpuindex) 1578 if (sym[i].st_shndx == pcpuindex)
1523 secbase = (unsigned long)mod->percpu; 1579 secbase = (unsigned long)mod_percpu(mod);
1524 else 1580 else
1525 secbase = sechdrs[sym[i].st_shndx].sh_addr; 1581 secbase = sechdrs[sym[i].st_shndx].sh_addr;
1526 sym[i].st_value += secbase; 1582 sym[i].st_value += secbase;
@@ -1954,7 +2010,7 @@ static noinline struct module *load_module(void __user *umod,
1954 unsigned int modindex, versindex, infoindex, pcpuindex; 2010 unsigned int modindex, versindex, infoindex, pcpuindex;
1955 struct module *mod; 2011 struct module *mod;
1956 long err = 0; 2012 long err = 0;
1957 void *percpu = NULL, *ptr = NULL; /* Stops spurious gcc warning */ 2013 void *ptr = NULL; /* Stops spurious gcc warning */
1958 unsigned long symoffs, stroffs, *strmap; 2014 unsigned long symoffs, stroffs, *strmap;
1959 2015
1960 mm_segment_t old_fs; 2016 mm_segment_t old_fs;
@@ -2094,15 +2150,11 @@ static noinline struct module *load_module(void __user *umod,
2094 2150
2095 if (pcpuindex) { 2151 if (pcpuindex) {
2096 /* We have a special allocation for this section. */ 2152 /* We have a special allocation for this section. */
2097 percpu = percpu_modalloc(sechdrs[pcpuindex].sh_size, 2153 err = percpu_modalloc(mod, sechdrs[pcpuindex].sh_size,
2098 sechdrs[pcpuindex].sh_addralign, 2154 sechdrs[pcpuindex].sh_addralign);
2099 mod->name); 2155 if (err)
2100 if (!percpu) {
2101 err = -ENOMEM;
2102 goto free_mod; 2156 goto free_mod;
2103 }
2104 sechdrs[pcpuindex].sh_flags &= ~(unsigned long)SHF_ALLOC; 2157 sechdrs[pcpuindex].sh_flags &= ~(unsigned long)SHF_ALLOC;
2105 mod->percpu = percpu;
2106 } 2158 }
2107 2159
2108 /* Determine total sizes, and put offsets in sh_entsize. For now 2160 /* Determine total sizes, and put offsets in sh_entsize. For now
@@ -2317,7 +2369,7 @@ static noinline struct module *load_module(void __user *umod,
2317 sort_extable(mod->extable, mod->extable + mod->num_exentries); 2369 sort_extable(mod->extable, mod->extable + mod->num_exentries);
2318 2370
2319 /* Finally, copy percpu area over. */ 2371 /* Finally, copy percpu area over. */
2320 percpu_modcopy(mod->percpu, (void *)sechdrs[pcpuindex].sh_addr, 2372 percpu_modcopy(mod, (void *)sechdrs[pcpuindex].sh_addr,
2321 sechdrs[pcpuindex].sh_size); 2373 sechdrs[pcpuindex].sh_size);
2322 2374
2323 add_kallsyms(mod, sechdrs, hdr->e_shnum, symindex, strindex, 2375 add_kallsyms(mod, sechdrs, hdr->e_shnum, symindex, strindex,
@@ -2409,8 +2461,7 @@ static noinline struct module *load_module(void __user *umod,
2409 module_free(mod, mod->module_core); 2461 module_free(mod, mod->module_core);
2410 /* mod will be freed with core. Don't access it beyond this line! */ 2462 /* mod will be freed with core. Don't access it beyond this line! */
2411 free_percpu: 2463 free_percpu:
2412 if (percpu) 2464 percpu_modfree(mod);
2413 percpu_modfree(percpu);
2414 free_mod: 2465 free_mod:
2415 kfree(args); 2466 kfree(args);
2416 kfree(strmap); 2467 kfree(strmap);