diff options
author | Jaroslav Kysela <perex@suse.cz> | 2006-02-01 07:08:56 -0500 |
---|---|---|
committer | Jaroslav Kysela <perex@suse.cz> | 2006-02-01 07:08:56 -0500 |
commit | 847b9d01474f710e7a018186917d05e59e258309 (patch) | |
tree | c0da8777ce350c4b048aa6ed2c41fdd109c42e92 /kernel | |
parent | d1d051b28e9d3c3bed0bd15a2b49df3d04f7768f (diff) | |
parent | a6df590dd8b7644c8e298e3b13442bcd6ceeb739 (diff) |
Merge with rsync://rsync.kernel.org/pub/scm/linux/kernel/git/torvalds/linux-2.6.git
Diffstat (limited to 'kernel')
-rw-r--r-- | kernel/rcutorture.c | 10 | ||||
-rw-r--r-- | kernel/sched.c | 6 | ||||
-rw-r--r-- | kernel/time.c | 2 | ||||
-rw-r--r-- | kernel/user.c | 32 |
4 files changed, 33 insertions, 17 deletions
diff --git a/kernel/rcutorture.c b/kernel/rcutorture.c index 773219907dd8..7712912dbc84 100644 --- a/kernel/rcutorture.c +++ b/kernel/rcutorture.c | |||
@@ -114,16 +114,16 @@ rcu_torture_alloc(void) | |||
114 | { | 114 | { |
115 | struct list_head *p; | 115 | struct list_head *p; |
116 | 116 | ||
117 | spin_lock(&rcu_torture_lock); | 117 | spin_lock_bh(&rcu_torture_lock); |
118 | if (list_empty(&rcu_torture_freelist)) { | 118 | if (list_empty(&rcu_torture_freelist)) { |
119 | atomic_inc(&n_rcu_torture_alloc_fail); | 119 | atomic_inc(&n_rcu_torture_alloc_fail); |
120 | spin_unlock(&rcu_torture_lock); | 120 | spin_unlock_bh(&rcu_torture_lock); |
121 | return NULL; | 121 | return NULL; |
122 | } | 122 | } |
123 | atomic_inc(&n_rcu_torture_alloc); | 123 | atomic_inc(&n_rcu_torture_alloc); |
124 | p = rcu_torture_freelist.next; | 124 | p = rcu_torture_freelist.next; |
125 | list_del_init(p); | 125 | list_del_init(p); |
126 | spin_unlock(&rcu_torture_lock); | 126 | spin_unlock_bh(&rcu_torture_lock); |
127 | return container_of(p, struct rcu_torture, rtort_free); | 127 | return container_of(p, struct rcu_torture, rtort_free); |
128 | } | 128 | } |
129 | 129 | ||
@@ -134,9 +134,9 @@ static void | |||
134 | rcu_torture_free(struct rcu_torture *p) | 134 | rcu_torture_free(struct rcu_torture *p) |
135 | { | 135 | { |
136 | atomic_inc(&n_rcu_torture_free); | 136 | atomic_inc(&n_rcu_torture_free); |
137 | spin_lock(&rcu_torture_lock); | 137 | spin_lock_bh(&rcu_torture_lock); |
138 | list_add_tail(&p->rtort_free, &rcu_torture_freelist); | 138 | list_add_tail(&p->rtort_free, &rcu_torture_freelist); |
139 | spin_unlock(&rcu_torture_lock); | 139 | spin_unlock_bh(&rcu_torture_lock); |
140 | } | 140 | } |
141 | 141 | ||
142 | static void | 142 | static void |
diff --git a/kernel/sched.c b/kernel/sched.c index 3ee2ae45125f..ec7fd9cee306 100644 --- a/kernel/sched.c +++ b/kernel/sched.c | |||
@@ -5141,7 +5141,7 @@ static void init_sched_build_groups(struct sched_group groups[], cpumask_t span, | |||
5141 | #define SEARCH_SCOPE 2 | 5141 | #define SEARCH_SCOPE 2 |
5142 | #define MIN_CACHE_SIZE (64*1024U) | 5142 | #define MIN_CACHE_SIZE (64*1024U) |
5143 | #define DEFAULT_CACHE_SIZE (5*1024*1024U) | 5143 | #define DEFAULT_CACHE_SIZE (5*1024*1024U) |
5144 | #define ITERATIONS 2 | 5144 | #define ITERATIONS 1 |
5145 | #define SIZE_THRESH 130 | 5145 | #define SIZE_THRESH 130 |
5146 | #define COST_THRESH 130 | 5146 | #define COST_THRESH 130 |
5147 | 5147 | ||
@@ -5480,9 +5480,9 @@ static unsigned long long measure_migration_cost(int cpu1, int cpu2) | |||
5480 | break; | 5480 | break; |
5481 | } | 5481 | } |
5482 | /* | 5482 | /* |
5483 | * Increase the cachesize in 5% steps: | 5483 | * Increase the cachesize in 10% steps: |
5484 | */ | 5484 | */ |
5485 | size = size * 20 / 19; | 5485 | size = size * 10 / 9; |
5486 | } | 5486 | } |
5487 | 5487 | ||
5488 | if (migration_debug) | 5488 | if (migration_debug) |
diff --git a/kernel/time.c b/kernel/time.c index 7477b1d2079e..1f23e683d6aa 100644 --- a/kernel/time.c +++ b/kernel/time.c | |||
@@ -155,7 +155,7 @@ int do_sys_settimeofday(struct timespec *tv, struct timezone *tz) | |||
155 | static int firsttime = 1; | 155 | static int firsttime = 1; |
156 | int error = 0; | 156 | int error = 0; |
157 | 157 | ||
158 | if (!timespec_valid(tv)) | 158 | if (tv && !timespec_valid(tv)) |
159 | return -EINVAL; | 159 | return -EINVAL; |
160 | 160 | ||
161 | error = security_settime(tv, tz); | 161 | error = security_settime(tv, tz); |
diff --git a/kernel/user.c b/kernel/user.c index 89e562feb1b1..d9deae43a9ab 100644 --- a/kernel/user.c +++ b/kernel/user.c | |||
@@ -13,6 +13,7 @@ | |||
13 | #include <linux/slab.h> | 13 | #include <linux/slab.h> |
14 | #include <linux/bitops.h> | 14 | #include <linux/bitops.h> |
15 | #include <linux/key.h> | 15 | #include <linux/key.h> |
16 | #include <linux/interrupt.h> | ||
16 | 17 | ||
17 | /* | 18 | /* |
18 | * UID task count cache, to get fast user lookup in "alloc_uid" | 19 | * UID task count cache, to get fast user lookup in "alloc_uid" |
@@ -27,6 +28,16 @@ | |||
27 | 28 | ||
28 | static kmem_cache_t *uid_cachep; | 29 | static kmem_cache_t *uid_cachep; |
29 | static struct list_head uidhash_table[UIDHASH_SZ]; | 30 | static struct list_head uidhash_table[UIDHASH_SZ]; |
31 | |||
32 | /* | ||
33 | * The uidhash_lock is mostly taken from process context, but it is | ||
34 | * occasionally also taken from softirq/tasklet context, when | ||
35 | * task-structs get RCU-freed. Hence all locking must be softirq-safe. | ||
36 | * But free_uid() is also called with local interrupts disabled, and running | ||
37 | * local_bh_enable() with local interrupts disabled is an error - we'll run | ||
38 | * softirq callbacks, and they can unconditionally enable interrupts, and | ||
39 | * the caller of free_uid() didn't expect that.. | ||
40 | */ | ||
30 | static DEFINE_SPINLOCK(uidhash_lock); | 41 | static DEFINE_SPINLOCK(uidhash_lock); |
31 | 42 | ||
32 | struct user_struct root_user = { | 43 | struct user_struct root_user = { |
@@ -82,15 +93,19 @@ static inline struct user_struct *uid_hash_find(uid_t uid, struct list_head *has | |||
82 | struct user_struct *find_user(uid_t uid) | 93 | struct user_struct *find_user(uid_t uid) |
83 | { | 94 | { |
84 | struct user_struct *ret; | 95 | struct user_struct *ret; |
96 | unsigned long flags; | ||
85 | 97 | ||
86 | spin_lock(&uidhash_lock); | 98 | spin_lock_irqsave(&uidhash_lock, flags); |
87 | ret = uid_hash_find(uid, uidhashentry(uid)); | 99 | ret = uid_hash_find(uid, uidhashentry(uid)); |
88 | spin_unlock(&uidhash_lock); | 100 | spin_unlock_irqrestore(&uidhash_lock, flags); |
89 | return ret; | 101 | return ret; |
90 | } | 102 | } |
91 | 103 | ||
92 | void free_uid(struct user_struct *up) | 104 | void free_uid(struct user_struct *up) |
93 | { | 105 | { |
106 | unsigned long flags; | ||
107 | |||
108 | local_irq_save(flags); | ||
94 | if (up && atomic_dec_and_lock(&up->__count, &uidhash_lock)) { | 109 | if (up && atomic_dec_and_lock(&up->__count, &uidhash_lock)) { |
95 | uid_hash_remove(up); | 110 | uid_hash_remove(up); |
96 | key_put(up->uid_keyring); | 111 | key_put(up->uid_keyring); |
@@ -98,6 +113,7 @@ void free_uid(struct user_struct *up) | |||
98 | kmem_cache_free(uid_cachep, up); | 113 | kmem_cache_free(uid_cachep, up); |
99 | spin_unlock(&uidhash_lock); | 114 | spin_unlock(&uidhash_lock); |
100 | } | 115 | } |
116 | local_irq_restore(flags); | ||
101 | } | 117 | } |
102 | 118 | ||
103 | struct user_struct * alloc_uid(uid_t uid) | 119 | struct user_struct * alloc_uid(uid_t uid) |
@@ -105,9 +121,9 @@ struct user_struct * alloc_uid(uid_t uid) | |||
105 | struct list_head *hashent = uidhashentry(uid); | 121 | struct list_head *hashent = uidhashentry(uid); |
106 | struct user_struct *up; | 122 | struct user_struct *up; |
107 | 123 | ||
108 | spin_lock(&uidhash_lock); | 124 | spin_lock_irq(&uidhash_lock); |
109 | up = uid_hash_find(uid, hashent); | 125 | up = uid_hash_find(uid, hashent); |
110 | spin_unlock(&uidhash_lock); | 126 | spin_unlock_irq(&uidhash_lock); |
111 | 127 | ||
112 | if (!up) { | 128 | if (!up) { |
113 | struct user_struct *new; | 129 | struct user_struct *new; |
@@ -137,7 +153,7 @@ struct user_struct * alloc_uid(uid_t uid) | |||
137 | * Before adding this, check whether we raced | 153 | * Before adding this, check whether we raced |
138 | * on adding the same user already.. | 154 | * on adding the same user already.. |
139 | */ | 155 | */ |
140 | spin_lock(&uidhash_lock); | 156 | spin_lock_irq(&uidhash_lock); |
141 | up = uid_hash_find(uid, hashent); | 157 | up = uid_hash_find(uid, hashent); |
142 | if (up) { | 158 | if (up) { |
143 | key_put(new->uid_keyring); | 159 | key_put(new->uid_keyring); |
@@ -147,7 +163,7 @@ struct user_struct * alloc_uid(uid_t uid) | |||
147 | uid_hash_insert(new, hashent); | 163 | uid_hash_insert(new, hashent); |
148 | up = new; | 164 | up = new; |
149 | } | 165 | } |
150 | spin_unlock(&uidhash_lock); | 166 | spin_unlock_irq(&uidhash_lock); |
151 | 167 | ||
152 | } | 168 | } |
153 | return up; | 169 | return up; |
@@ -183,9 +199,9 @@ static int __init uid_cache_init(void) | |||
183 | INIT_LIST_HEAD(uidhash_table + n); | 199 | INIT_LIST_HEAD(uidhash_table + n); |
184 | 200 | ||
185 | /* Insert the root user immediately (init already runs as root) */ | 201 | /* Insert the root user immediately (init already runs as root) */ |
186 | spin_lock(&uidhash_lock); | 202 | spin_lock_irq(&uidhash_lock); |
187 | uid_hash_insert(&root_user, uidhashentry(0)); | 203 | uid_hash_insert(&root_user, uidhashentry(0)); |
188 | spin_unlock(&uidhash_lock); | 204 | spin_unlock_irq(&uidhash_lock); |
189 | 205 | ||
190 | return 0; | 206 | return 0; |
191 | } | 207 | } |