diff options
author | Eric W. Biederman <ebiederm@xmission.com> | 2006-10-02 05:17:04 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@g5.osdl.org> | 2006-10-02 10:57:12 -0400 |
commit | 0804ef4b0de7121261f77c565b20a11ac694e877 (patch) | |
tree | ff12e3b999dc2ce66d97fce5d76cd7df073c0d5c /kernel | |
parent | 2bc2d61a9638dab670d8361e928d1a5a291173ef (diff) |
[PATCH] proc: readdir race fix (take 3)
The problem: An opendir, readdir, closedir sequence can fail to report
process ids that are continually in use throughout the sequence of system
calls. For this race to trigger the process that proc_pid_readdir stops at
must exit before readdir is called again.
This can cause ps to fail to report processes, and it is in violation of
posix guarantees and normal application expectations with respect to
readdir.
Currently there is no way to work around this problem in user space short
of providing a gargantuan buffer to user space so the directory read all
happens in on system call.
This patch implements the normal directory semantics for proc, that
guarantee that a directory entry that is neither created nor destroyed
while reading the directory entry will be returned. For directory that are
either created or destroyed during the readdir you may or may not see them.
Furthermore you may seek to a directory offset you have previously seen.
These are the guarantee that ext[23] provides and that posix requires, and
more importantly that user space expects. Plus it is a simple semantic to
implement reliable service. It is just a matter of calling readdir a
second time if you are wondering if something new has show up.
These better semantics are implemented by scanning through the pids in
numerical order and by making the file offset a pid plus a fixed offset.
The pid scan happens on the pid bitmap, which when you look at it is
remarkably efficient for a brute force algorithm. Given that a typical
cache line is 64 bytes and thus covers space for 64*8 == 200 pids. There
are only 40 cache lines for the entire 32K pid space. A typical system
will have 100 pids or more so this is actually fewer cache lines we have to
look at to scan a linked list, and the worst case of having to scan the
entire pid bitmap is pretty reasonable.
If we need something more efficient we can go to a more efficient data
structure for indexing the pids, but for now what we have should be
sufficient.
In addition this takes no additional locks and is actually less code than
what we are doing now.
Also another very subtle bug in this area has been fixed. It is possible
to catch a task in the middle of de_thread where a thread is assuming the
thread of it's thread group leader. This patch carefully handles that case
so if we hit it we don't fail to return the pid, that is undergoing the
de_thread dance.
Thanks to KAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com> for
providing the first fix, pointing this out and working on it.
[oleg@tv-sign.ru: fix it]
Signed-off-by: Eric W. Biederman <ebiederm@xmission.com>
Acked-by: KAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com>
Signed-off-by: Oleg Nesterov <oleg@tv-sign.ru>
Cc: Jean Delvare <jdelvare@suse.de>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
Diffstat (limited to 'kernel')
-rw-r--r-- | kernel/pid.c | 36 |
1 files changed, 36 insertions, 0 deletions
diff --git a/kernel/pid.c b/kernel/pid.c index 8387e8c68193..ed89a732432c 100644 --- a/kernel/pid.c +++ b/kernel/pid.c | |||
@@ -145,6 +145,23 @@ static int alloc_pidmap(void) | |||
145 | return -1; | 145 | return -1; |
146 | } | 146 | } |
147 | 147 | ||
148 | static int next_pidmap(int last) | ||
149 | { | ||
150 | int offset; | ||
151 | pidmap_t *map; | ||
152 | |||
153 | offset = (last + 1) & BITS_PER_PAGE_MASK; | ||
154 | map = &pidmap_array[(last + 1)/BITS_PER_PAGE]; | ||
155 | for (; map < &pidmap_array[PIDMAP_ENTRIES]; map++, offset = 0) { | ||
156 | if (unlikely(!map->page)) | ||
157 | continue; | ||
158 | offset = find_next_bit((map)->page, BITS_PER_PAGE, offset); | ||
159 | if (offset < BITS_PER_PAGE) | ||
160 | return mk_pid(map, offset); | ||
161 | } | ||
162 | return -1; | ||
163 | } | ||
164 | |||
148 | fastcall void put_pid(struct pid *pid) | 165 | fastcall void put_pid(struct pid *pid) |
149 | { | 166 | { |
150 | if (!pid) | 167 | if (!pid) |
@@ -303,6 +320,25 @@ struct pid *find_get_pid(pid_t nr) | |||
303 | } | 320 | } |
304 | 321 | ||
305 | /* | 322 | /* |
323 | * Used by proc to find the first pid that is greater then or equal to nr. | ||
324 | * | ||
325 | * If there is a pid at nr this function is exactly the same as find_pid. | ||
326 | */ | ||
327 | struct pid *find_ge_pid(int nr) | ||
328 | { | ||
329 | struct pid *pid; | ||
330 | |||
331 | do { | ||
332 | pid = find_pid(nr); | ||
333 | if (pid) | ||
334 | break; | ||
335 | nr = next_pidmap(nr); | ||
336 | } while (nr > 0); | ||
337 | |||
338 | return pid; | ||
339 | } | ||
340 | |||
341 | /* | ||
306 | * The pid hash table is scaled according to the amount of memory in the | 342 | * The pid hash table is scaled according to the amount of memory in the |
307 | * machine. From a minimum of 16 slots up to 4096 slots at one gigabyte or | 343 | * machine. From a minimum of 16 slots up to 4096 slots at one gigabyte or |
308 | * more. | 344 | * more. |