diff options
| author | Sukadev Bhattiprolu <sukadev@us.ibm.com> | 2006-12-08 05:37:58 -0500 |
|---|---|---|
| committer | Linus Torvalds <torvalds@woody.osdl.org> | 2006-12-08 11:28:52 -0500 |
| commit | 61a58c6c238cc81f7742b8cc84212cc55fb57747 (patch) | |
| tree | de0a4338b9c9c42aa5a47293e129282172a6053f /kernel/pid.c | |
| parent | 373beb35cd6b625e0ba4ad98baace12310a26aa8 (diff) | |
[PATCH] rename struct pspace to struct pid_namespace
Rename struct pspace to struct pid_namespace for consistency with other
namespaces (uts_namespace and ipc_namespace). Also rename
include/linux/pspace.h to include/linux/pid_namespace.h and variables from
pspace to pid_ns.
Signed-off-by: Sukadev Bhattiprolu <sukadev@us.ibm.com>
Signed-off-by: Cedric Le Goater <clg@fr.ibm.com>
Cc: Kirill Korotaev <dev@openvz.org>
Cc: Eric W. Biederman <ebiederm@xmission.com>
Cc: Herbert Poetzl <herbert@13thfloor.at>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
Diffstat (limited to 'kernel/pid.c')
| -rw-r--r-- | kernel/pid.c | 49 |
1 files changed, 25 insertions, 24 deletions
diff --git a/kernel/pid.c b/kernel/pid.c index a48879b0b9..25807e1b98 100644 --- a/kernel/pid.c +++ b/kernel/pid.c | |||
| @@ -26,7 +26,7 @@ | |||
| 26 | #include <linux/init.h> | 26 | #include <linux/init.h> |
| 27 | #include <linux/bootmem.h> | 27 | #include <linux/bootmem.h> |
| 28 | #include <linux/hash.h> | 28 | #include <linux/hash.h> |
| 29 | #include <linux/pspace.h> | 29 | #include <linux/pid_namespace.h> |
| 30 | 30 | ||
| 31 | #define pid_hashfn(nr) hash_long((unsigned long)nr, pidhash_shift) | 31 | #define pid_hashfn(nr) hash_long((unsigned long)nr, pidhash_shift) |
| 32 | static struct hlist_head *pid_hash; | 32 | static struct hlist_head *pid_hash; |
| @@ -43,9 +43,10 @@ int pid_max_max = PID_MAX_LIMIT; | |||
| 43 | #define BITS_PER_PAGE (PAGE_SIZE*8) | 43 | #define BITS_PER_PAGE (PAGE_SIZE*8) |
| 44 | #define BITS_PER_PAGE_MASK (BITS_PER_PAGE-1) | 44 | #define BITS_PER_PAGE_MASK (BITS_PER_PAGE-1) |
| 45 | 45 | ||
| 46 | static inline int mk_pid(struct pspace *pspace, struct pidmap *map, int off) | 46 | static inline int mk_pid(struct pid_namespace *pid_ns, |
| 47 | struct pidmap *map, int off) | ||
| 47 | { | 48 | { |
| 48 | return (map - pspace->pidmap)*BITS_PER_PAGE + off; | 49 | return (map - pid_ns->pidmap)*BITS_PER_PAGE + off; |
| 49 | } | 50 | } |
| 50 | 51 | ||
| 51 | #define find_next_offset(map, off) \ | 52 | #define find_next_offset(map, off) \ |
| @@ -57,7 +58,7 @@ static inline int mk_pid(struct pspace *pspace, struct pidmap *map, int off) | |||
| 57 | * value does not cause lots of bitmaps to be allocated, but | 58 | * value does not cause lots of bitmaps to be allocated, but |
| 58 | * the scheme scales to up to 4 million PIDs, runtime. | 59 | * the scheme scales to up to 4 million PIDs, runtime. |
| 59 | */ | 60 | */ |
| 60 | struct pspace init_pspace = { | 61 | struct pid_namespace init_pid_ns = { |
| 61 | .pidmap = { | 62 | .pidmap = { |
| 62 | [ 0 ... PIDMAP_ENTRIES-1] = { ATOMIC_INIT(BITS_PER_PAGE), NULL } | 63 | [ 0 ... PIDMAP_ENTRIES-1] = { ATOMIC_INIT(BITS_PER_PAGE), NULL } |
| 63 | }, | 64 | }, |
| @@ -80,25 +81,25 @@ struct pspace init_pspace = { | |||
| 80 | 81 | ||
| 81 | static __cacheline_aligned_in_smp DEFINE_SPINLOCK(pidmap_lock); | 82 | static __cacheline_aligned_in_smp DEFINE_SPINLOCK(pidmap_lock); |
| 82 | 83 | ||
| 83 | static fastcall void free_pidmap(struct pspace *pspace, int pid) | 84 | static fastcall void free_pidmap(struct pid_namespace *pid_ns, int pid) |
| 84 | { | 85 | { |
| 85 | struct pidmap *map = pspace->pidmap + pid / BITS_PER_PAGE; | 86 | struct pidmap *map = pid_ns->pidmap + pid / BITS_PER_PAGE; |
| 86 | int offset = pid & BITS_PER_PAGE_MASK; | 87 | int offset = pid & BITS_PER_PAGE_MASK; |
| 87 | 88 | ||
| 88 | clear_bit(offset, map->page); | 89 | clear_bit(offset, map->page); |
| 89 | atomic_inc(&map->nr_free); | 90 | atomic_inc(&map->nr_free); |
| 90 | } | 91 | } |
| 91 | 92 | ||
| 92 | static int alloc_pidmap(struct pspace *pspace) | 93 | static int alloc_pidmap(struct pid_namespace *pid_ns) |
| 93 | { | 94 | { |
| 94 | int i, offset, max_scan, pid, last = pspace->last_pid; | 95 | int i, offset, max_scan, pid, last = pid_ns->last_pid; |
| 95 | struct pidmap *map; | 96 | struct pidmap *map; |
| 96 | 97 | ||
| 97 | pid = last + 1; | 98 | pid = last + 1; |
| 98 | if (pid >= pid_max) | 99 | if (pid >= pid_max) |
| 99 | pid = RESERVED_PIDS; | 100 | pid = RESERVED_PIDS; |
| 100 | offset = pid & BITS_PER_PAGE_MASK; | 101 | offset = pid & BITS_PER_PAGE_MASK; |
| 101 | map = &pspace->pidmap[pid/BITS_PER_PAGE]; | 102 | map = &pid_ns->pidmap[pid/BITS_PER_PAGE]; |
| 102 | max_scan = (pid_max + BITS_PER_PAGE - 1)/BITS_PER_PAGE - !offset; | 103 | max_scan = (pid_max + BITS_PER_PAGE - 1)/BITS_PER_PAGE - !offset; |
| 103 | for (i = 0; i <= max_scan; ++i) { | 104 | for (i = 0; i <= max_scan; ++i) { |
| 104 | if (unlikely(!map->page)) { | 105 | if (unlikely(!map->page)) { |
| @@ -120,11 +121,11 @@ static int alloc_pidmap(struct pspace *pspace) | |||
| 120 | do { | 121 | do { |
| 121 | if (!test_and_set_bit(offset, map->page)) { | 122 | if (!test_and_set_bit(offset, map->page)) { |
| 122 | atomic_dec(&map->nr_free); | 123 | atomic_dec(&map->nr_free); |
| 123 | pspace->last_pid = pid; | 124 | pid_ns->last_pid = pid; |
| 124 | return pid; | 125 | return pid; |
| 125 | } | 126 | } |
| 126 | offset = find_next_offset(map, offset); | 127 | offset = find_next_offset(map, offset); |
| 127 | pid = mk_pid(pspace, map, offset); | 128 | pid = mk_pid(pid_ns, map, offset); |
| 128 | /* | 129 | /* |
| 129 | * find_next_offset() found a bit, the pid from it | 130 | * find_next_offset() found a bit, the pid from it |
| 130 | * is in-bounds, and if we fell back to the last | 131 | * is in-bounds, and if we fell back to the last |
| @@ -135,34 +136,34 @@ static int alloc_pidmap(struct pspace *pspace) | |||
| 135 | (i != max_scan || pid < last || | 136 | (i != max_scan || pid < last || |
| 136 | !((last+1) & BITS_PER_PAGE_MASK))); | 137 | !((last+1) & BITS_PER_PAGE_MASK))); |
| 137 | } | 138 | } |
| 138 | if (map < &pspace->pidmap[(pid_max-1)/BITS_PER_PAGE]) { | 139 | if (map < &pid_ns->pidmap[(pid_max-1)/BITS_PER_PAGE]) { |
| 139 | ++map; | 140 | ++map; |
| 140 | offset = 0; | 141 | offset = 0; |
| 141 | } else { | 142 | } else { |
| 142 | map = &pspace->pidmap[0]; | 143 | map = &pid_ns->pidmap[0]; |
| 143 | offset = RESERVED_PIDS; | 144 | offset = RESERVED_PIDS; |
| 144 | if (unlikely(last == offset)) | 145 | if (unlikely(last == offset)) |
| 145 | break; | 146 | break; |
| 146 | } | 147 | } |
| 147 | pid = mk_pid(pspace, map, offset); | 148 | pid = mk_pid(pid_ns, map, offset); |
| 148 | } | 149 | } |
| 149 | return -1; | 150 | return -1; |
| 150 | } | 151 | } |
| 151 | 152 | ||
| 152 | static int next_pidmap(struct pspace *pspace, int last) | 153 | static int next_pidmap(struct pid_namespace *pid_ns, int last) |
| 153 | { | 154 | { |
| 154 | int offset; | 155 | int offset; |
| 155 | struct pidmap *map, *end; | 156 | struct pidmap *map, *end; |
| 156 | 157 | ||
| 157 | offset = (last + 1) & BITS_PER_PAGE_MASK; | 158 | offset = (last + 1) & BITS_PER_PAGE_MASK; |
| 158 | map = &pspace->pidmap[(last + 1)/BITS_PER_PAGE]; | 159 | map = &pid_ns->pidmap[(last + 1)/BITS_PER_PAGE]; |
| 159 | end = &pspace->pidmap[PIDMAP_ENTRIES]; | 160 | end = &pid_ns->pidmap[PIDMAP_ENTRIES]; |
| 160 | for (; map < end; map++, offset = 0) { | 161 | for (; map < end; map++, offset = 0) { |
| 161 | if (unlikely(!map->page)) | 162 | if (unlikely(!map->page)) |
| 162 | continue; | 163 | continue; |
| 163 | offset = find_next_bit((map)->page, BITS_PER_PAGE, offset); | 164 | offset = find_next_bit((map)->page, BITS_PER_PAGE, offset); |
| 164 | if (offset < BITS_PER_PAGE) | 165 | if (offset < BITS_PER_PAGE) |
| 165 | return mk_pid(pspace, map, offset); | 166 | return mk_pid(pid_ns, map, offset); |
| 166 | } | 167 | } |
| 167 | return -1; | 168 | return -1; |
| 168 | } | 169 | } |
| @@ -192,7 +193,7 @@ fastcall void free_pid(struct pid *pid) | |||
| 192 | hlist_del_rcu(&pid->pid_chain); | 193 | hlist_del_rcu(&pid->pid_chain); |
| 193 | spin_unlock_irqrestore(&pidmap_lock, flags); | 194 | spin_unlock_irqrestore(&pidmap_lock, flags); |
| 194 | 195 | ||
| 195 | free_pidmap(&init_pspace, pid->nr); | 196 | free_pidmap(&init_pid_ns, pid->nr); |
| 196 | call_rcu(&pid->rcu, delayed_put_pid); | 197 | call_rcu(&pid->rcu, delayed_put_pid); |
| 197 | } | 198 | } |
| 198 | 199 | ||
| @@ -206,7 +207,7 @@ struct pid *alloc_pid(void) | |||
| 206 | if (!pid) | 207 | if (!pid) |
| 207 | goto out; | 208 | goto out; |
| 208 | 209 | ||
| 209 | nr = alloc_pidmap(&init_pspace); | 210 | nr = alloc_pidmap(&init_pid_ns); |
| 210 | if (nr < 0) | 211 | if (nr < 0) |
| 211 | goto out_free; | 212 | goto out_free; |
| 212 | 213 | ||
| @@ -348,7 +349,7 @@ struct pid *find_ge_pid(int nr) | |||
| 348 | pid = find_pid(nr); | 349 | pid = find_pid(nr); |
| 349 | if (pid) | 350 | if (pid) |
| 350 | break; | 351 | break; |
| 351 | nr = next_pidmap(&init_pspace, nr); | 352 | nr = next_pidmap(&init_pid_ns, nr); |
| 352 | } while (nr > 0); | 353 | } while (nr > 0); |
| 353 | 354 | ||
| 354 | return pid; | 355 | return pid; |
| @@ -382,10 +383,10 @@ void __init pidhash_init(void) | |||
| 382 | 383 | ||
| 383 | void __init pidmap_init(void) | 384 | void __init pidmap_init(void) |
| 384 | { | 385 | { |
| 385 | init_pspace.pidmap[0].page = kzalloc(PAGE_SIZE, GFP_KERNEL); | 386 | init_pid_ns.pidmap[0].page = kzalloc(PAGE_SIZE, GFP_KERNEL); |
| 386 | /* Reserve PID 0. We never call free_pidmap(0) */ | 387 | /* Reserve PID 0. We never call free_pidmap(0) */ |
| 387 | set_bit(0, init_pspace.pidmap[0].page); | 388 | set_bit(0, init_pid_ns.pidmap[0].page); |
| 388 | atomic_dec(&init_pspace.pidmap[0].nr_free); | 389 | atomic_dec(&init_pid_ns.pidmap[0].nr_free); |
| 389 | 390 | ||
| 390 | pid_cachep = kmem_cache_create("pid", sizeof(struct pid), | 391 | pid_cachep = kmem_cache_create("pid", sizeof(struct pid), |
| 391 | __alignof__(struct pid), | 392 | __alignof__(struct pid), |
