diff options
-rw-r--r-- | ipc/sem.c | 123 |
1 files changed, 91 insertions, 32 deletions
@@ -381,7 +381,6 @@ static int try_atomic_semop (struct sem_array * sma, struct sembuf * sops, | |||
381 | sop--; | 381 | sop--; |
382 | } | 382 | } |
383 | 383 | ||
384 | sma->sem_otime = get_seconds(); | ||
385 | return 0; | 384 | return 0; |
386 | 385 | ||
387 | out_of_range: | 386 | out_of_range: |
@@ -404,25 +403,51 @@ undo: | |||
404 | return result; | 403 | return result; |
405 | } | 404 | } |
406 | 405 | ||
407 | /* | 406 | /** wake_up_sem_queue_prepare(q, error): Prepare wake-up |
408 | * Wake up a process waiting on the sem queue with a given error. | 407 | * @q: queue entry that must be signaled |
409 | * The queue is invalid (may not be accessed) after the function returns. | 408 | * @error: Error value for the signal |
409 | * | ||
410 | * Prepare the wake-up of the queue entry q. | ||
410 | */ | 411 | */ |
411 | static void wake_up_sem_queue(struct sem_queue *q, int error) | 412 | static void wake_up_sem_queue_prepare(struct list_head *pt, |
413 | struct sem_queue *q, int error) | ||
412 | { | 414 | { |
413 | /* | 415 | if (list_empty(pt)) { |
414 | * Hold preempt off so that we don't get preempted and have the | 416 | /* |
415 | * wakee busy-wait until we're scheduled back on. We're holding | 417 | * Hold preempt off so that we don't get preempted and have the |
416 | * locks here so it may not strictly be needed, however if the | 418 | * wakee busy-wait until we're scheduled back on. |
417 | * locks become preemptible then this prevents such a problem. | 419 | */ |
418 | */ | 420 | preempt_disable(); |
419 | preempt_disable(); | 421 | } |
420 | q->status = IN_WAKEUP; | 422 | q->status = IN_WAKEUP; |
421 | wake_up_process(q->sleeper); | 423 | q->pid = error; |
422 | /* hands-off: q can disappear immediately after writing q->status. */ | 424 | |
423 | smp_wmb(); | 425 | list_add_tail(&q->simple_list, pt); |
424 | q->status = error; | 426 | } |
425 | preempt_enable(); | 427 | |
428 | /** | ||
429 | * wake_up_sem_queue_do(pt) - do the actual wake-up | ||
430 | * @pt: list of tasks to be woken up | ||
431 | * | ||
432 | * Do the actual wake-up. | ||
433 | * The function is called without any locks held, thus the semaphore array | ||
434 | * could be destroyed already and the tasks can disappear as soon as the | ||
435 | * status is set to the actual return code. | ||
436 | */ | ||
437 | static void wake_up_sem_queue_do(struct list_head *pt) | ||
438 | { | ||
439 | struct sem_queue *q, *t; | ||
440 | int did_something; | ||
441 | |||
442 | did_something = !list_empty(pt); | ||
443 | list_for_each_entry_safe(q, t, pt, simple_list) { | ||
444 | wake_up_process(q->sleeper); | ||
445 | /* q can disappear immediately after writing q->status. */ | ||
446 | smp_wmb(); | ||
447 | q->status = q->pid; | ||
448 | } | ||
449 | if (did_something) | ||
450 | preempt_enable(); | ||
426 | } | 451 | } |
427 | 452 | ||
428 | static void unlink_queue(struct sem_array *sma, struct sem_queue *q) | 453 | static void unlink_queue(struct sem_array *sma, struct sem_queue *q) |
@@ -502,17 +527,22 @@ static int check_restart(struct sem_array *sma, struct sem_queue *q) | |||
502 | * update_queue(sma, semnum): Look for tasks that can be completed. | 527 | * update_queue(sma, semnum): Look for tasks that can be completed. |
503 | * @sma: semaphore array. | 528 | * @sma: semaphore array. |
504 | * @semnum: semaphore that was modified. | 529 | * @semnum: semaphore that was modified. |
530 | * @pt: list head for the tasks that must be woken up. | ||
505 | * | 531 | * |
506 | * update_queue must be called after a semaphore in a semaphore array | 532 | * update_queue must be called after a semaphore in a semaphore array |
507 | * was modified. If multiple semaphore were modified, then @semnum | 533 | * was modified. If multiple semaphore were modified, then @semnum |
508 | * must be set to -1. | 534 | * must be set to -1. |
535 | * The tasks that must be woken up are added to @pt. The return code | ||
536 | * is stored in q->pid. | ||
537 | * The function return 1 if at least one semop was completed successfully. | ||
509 | */ | 538 | */ |
510 | static void update_queue(struct sem_array *sma, int semnum) | 539 | static int update_queue(struct sem_array *sma, int semnum, struct list_head *pt) |
511 | { | 540 | { |
512 | struct sem_queue *q; | 541 | struct sem_queue *q; |
513 | struct list_head *walk; | 542 | struct list_head *walk; |
514 | struct list_head *pending_list; | 543 | struct list_head *pending_list; |
515 | int offset; | 544 | int offset; |
545 | int semop_completed = 0; | ||
516 | 546 | ||
517 | /* if there are complex operations around, then knowing the semaphore | 547 | /* if there are complex operations around, then knowing the semaphore |
518 | * that was modified doesn't help us. Assume that multiple semaphores | 548 | * that was modified doesn't help us. Assume that multiple semaphores |
@@ -557,40 +587,55 @@ again: | |||
557 | 587 | ||
558 | unlink_queue(sma, q); | 588 | unlink_queue(sma, q); |
559 | 589 | ||
560 | if (error) | 590 | if (error) { |
561 | restart = 0; | 591 | restart = 0; |
562 | else | 592 | } else { |
593 | semop_completed = 1; | ||
563 | restart = check_restart(sma, q); | 594 | restart = check_restart(sma, q); |
595 | } | ||
564 | 596 | ||
565 | wake_up_sem_queue(q, error); | 597 | wake_up_sem_queue_prepare(pt, q, error); |
566 | if (restart) | 598 | if (restart) |
567 | goto again; | 599 | goto again; |
568 | } | 600 | } |
601 | return semop_completed; | ||
569 | } | 602 | } |
570 | 603 | ||
571 | /** do_smart_update(sma, sops, nsops): Optimized update_queue | 604 | /** |
605 | * do_smart_update(sma, sops, nsops, otime, pt) - optimized update_queue | ||
572 | * @sma: semaphore array | 606 | * @sma: semaphore array |
573 | * @sops: operations that were performed | 607 | * @sops: operations that were performed |
574 | * @nsops: number of operations | 608 | * @nsops: number of operations |
609 | * @otime: force setting otime | ||
610 | * @pt: list head of the tasks that must be woken up. | ||
575 | * | 611 | * |
576 | * do_smart_update() does the required called to update_queue, based on the | 612 | * do_smart_update() does the required called to update_queue, based on the |
577 | * actual changes that were performed on the semaphore array. | 613 | * actual changes that were performed on the semaphore array. |
614 | * Note that the function does not do the actual wake-up: the caller is | ||
615 | * responsible for calling wake_up_sem_queue_do(@pt). | ||
616 | * It is safe to perform this call after dropping all locks. | ||
578 | */ | 617 | */ |
579 | static void do_smart_update(struct sem_array *sma, struct sembuf *sops, int nsops) | 618 | static void do_smart_update(struct sem_array *sma, struct sembuf *sops, int nsops, |
619 | int otime, struct list_head *pt) | ||
580 | { | 620 | { |
581 | int i; | 621 | int i; |
582 | 622 | ||
583 | if (sma->complex_count || sops == NULL) { | 623 | if (sma->complex_count || sops == NULL) { |
584 | update_queue(sma, -1); | 624 | if (update_queue(sma, -1, pt)) |
585 | return; | 625 | otime = 1; |
626 | goto done; | ||
586 | } | 627 | } |
587 | 628 | ||
588 | for (i = 0; i < nsops; i++) { | 629 | for (i = 0; i < nsops; i++) { |
589 | if (sops[i].sem_op > 0 || | 630 | if (sops[i].sem_op > 0 || |
590 | (sops[i].sem_op < 0 && | 631 | (sops[i].sem_op < 0 && |
591 | sma->sem_base[sops[i].sem_num].semval == 0)) | 632 | sma->sem_base[sops[i].sem_num].semval == 0)) |
592 | update_queue(sma, sops[i].sem_num); | 633 | if (update_queue(sma, sops[i].sem_num, pt)) |
634 | otime = 1; | ||
593 | } | 635 | } |
636 | done: | ||
637 | if (otime) | ||
638 | sma->sem_otime = get_seconds(); | ||
594 | } | 639 | } |
595 | 640 | ||
596 | 641 | ||
@@ -656,6 +701,7 @@ static void freeary(struct ipc_namespace *ns, struct kern_ipc_perm *ipcp) | |||
656 | struct sem_undo *un, *tu; | 701 | struct sem_undo *un, *tu; |
657 | struct sem_queue *q, *tq; | 702 | struct sem_queue *q, *tq; |
658 | struct sem_array *sma = container_of(ipcp, struct sem_array, sem_perm); | 703 | struct sem_array *sma = container_of(ipcp, struct sem_array, sem_perm); |
704 | struct list_head tasks; | ||
659 | 705 | ||
660 | /* Free the existing undo structures for this semaphore set. */ | 706 | /* Free the existing undo structures for this semaphore set. */ |
661 | assert_spin_locked(&sma->sem_perm.lock); | 707 | assert_spin_locked(&sma->sem_perm.lock); |
@@ -669,15 +715,17 @@ static void freeary(struct ipc_namespace *ns, struct kern_ipc_perm *ipcp) | |||
669 | } | 715 | } |
670 | 716 | ||
671 | /* Wake up all pending processes and let them fail with EIDRM. */ | 717 | /* Wake up all pending processes and let them fail with EIDRM. */ |
718 | INIT_LIST_HEAD(&tasks); | ||
672 | list_for_each_entry_safe(q, tq, &sma->sem_pending, list) { | 719 | list_for_each_entry_safe(q, tq, &sma->sem_pending, list) { |
673 | unlink_queue(sma, q); | 720 | unlink_queue(sma, q); |
674 | wake_up_sem_queue(q, -EIDRM); | 721 | wake_up_sem_queue_prepare(&tasks, q, -EIDRM); |
675 | } | 722 | } |
676 | 723 | ||
677 | /* Remove the semaphore set from the IDR */ | 724 | /* Remove the semaphore set from the IDR */ |
678 | sem_rmid(ns, sma); | 725 | sem_rmid(ns, sma); |
679 | sem_unlock(sma); | 726 | sem_unlock(sma); |
680 | 727 | ||
728 | wake_up_sem_queue_do(&tasks); | ||
681 | ns->used_sems -= sma->sem_nsems; | 729 | ns->used_sems -= sma->sem_nsems; |
682 | security_sem_free(sma); | 730 | security_sem_free(sma); |
683 | ipc_rcu_putref(sma); | 731 | ipc_rcu_putref(sma); |
@@ -799,11 +847,13 @@ static int semctl_main(struct ipc_namespace *ns, int semid, int semnum, | |||
799 | ushort fast_sem_io[SEMMSL_FAST]; | 847 | ushort fast_sem_io[SEMMSL_FAST]; |
800 | ushort* sem_io = fast_sem_io; | 848 | ushort* sem_io = fast_sem_io; |
801 | int nsems; | 849 | int nsems; |
850 | struct list_head tasks; | ||
802 | 851 | ||
803 | sma = sem_lock_check(ns, semid); | 852 | sma = sem_lock_check(ns, semid); |
804 | if (IS_ERR(sma)) | 853 | if (IS_ERR(sma)) |
805 | return PTR_ERR(sma); | 854 | return PTR_ERR(sma); |
806 | 855 | ||
856 | INIT_LIST_HEAD(&tasks); | ||
807 | nsems = sma->sem_nsems; | 857 | nsems = sma->sem_nsems; |
808 | 858 | ||
809 | err = -EACCES; | 859 | err = -EACCES; |
@@ -891,7 +941,7 @@ static int semctl_main(struct ipc_namespace *ns, int semid, int semnum, | |||
891 | } | 941 | } |
892 | sma->sem_ctime = get_seconds(); | 942 | sma->sem_ctime = get_seconds(); |
893 | /* maybe some queued-up processes were waiting for this */ | 943 | /* maybe some queued-up processes were waiting for this */ |
894 | update_queue(sma, -1); | 944 | do_smart_update(sma, NULL, 0, 0, &tasks); |
895 | err = 0; | 945 | err = 0; |
896 | goto out_unlock; | 946 | goto out_unlock; |
897 | } | 947 | } |
@@ -933,13 +983,15 @@ static int semctl_main(struct ipc_namespace *ns, int semid, int semnum, | |||
933 | curr->sempid = task_tgid_vnr(current); | 983 | curr->sempid = task_tgid_vnr(current); |
934 | sma->sem_ctime = get_seconds(); | 984 | sma->sem_ctime = get_seconds(); |
935 | /* maybe some queued-up processes were waiting for this */ | 985 | /* maybe some queued-up processes were waiting for this */ |
936 | update_queue(sma, semnum); | 986 | do_smart_update(sma, NULL, 0, 0, &tasks); |
937 | err = 0; | 987 | err = 0; |
938 | goto out_unlock; | 988 | goto out_unlock; |
939 | } | 989 | } |
940 | } | 990 | } |
941 | out_unlock: | 991 | out_unlock: |
942 | sem_unlock(sma); | 992 | sem_unlock(sma); |
993 | wake_up_sem_queue_do(&tasks); | ||
994 | |||
943 | out_free: | 995 | out_free: |
944 | if(sem_io != fast_sem_io) | 996 | if(sem_io != fast_sem_io) |
945 | ipc_free(sem_io, sizeof(ushort)*nsems); | 997 | ipc_free(sem_io, sizeof(ushort)*nsems); |
@@ -1213,6 +1265,7 @@ SYSCALL_DEFINE4(semtimedop, int, semid, struct sembuf __user *, tsops, | |||
1213 | struct sem_queue queue; | 1265 | struct sem_queue queue; |
1214 | unsigned long jiffies_left = 0; | 1266 | unsigned long jiffies_left = 0; |
1215 | struct ipc_namespace *ns; | 1267 | struct ipc_namespace *ns; |
1268 | struct list_head tasks; | ||
1216 | 1269 | ||
1217 | ns = current->nsproxy->ipc_ns; | 1270 | ns = current->nsproxy->ipc_ns; |
1218 | 1271 | ||
@@ -1261,6 +1314,8 @@ SYSCALL_DEFINE4(semtimedop, int, semid, struct sembuf __user *, tsops, | |||
1261 | } else | 1314 | } else |
1262 | un = NULL; | 1315 | un = NULL; |
1263 | 1316 | ||
1317 | INIT_LIST_HEAD(&tasks); | ||
1318 | |||
1264 | sma = sem_lock_check(ns, semid); | 1319 | sma = sem_lock_check(ns, semid); |
1265 | if (IS_ERR(sma)) { | 1320 | if (IS_ERR(sma)) { |
1266 | if (un) | 1321 | if (un) |
@@ -1309,7 +1364,7 @@ SYSCALL_DEFINE4(semtimedop, int, semid, struct sembuf __user *, tsops, | |||
1309 | error = try_atomic_semop (sma, sops, nsops, un, task_tgid_vnr(current)); | 1364 | error = try_atomic_semop (sma, sops, nsops, un, task_tgid_vnr(current)); |
1310 | if (error <= 0) { | 1365 | if (error <= 0) { |
1311 | if (alter && error == 0) | 1366 | if (alter && error == 0) |
1312 | do_smart_update(sma, sops, nsops); | 1367 | do_smart_update(sma, sops, nsops, 1, &tasks); |
1313 | 1368 | ||
1314 | goto out_unlock_free; | 1369 | goto out_unlock_free; |
1315 | } | 1370 | } |
@@ -1386,6 +1441,8 @@ SYSCALL_DEFINE4(semtimedop, int, semid, struct sembuf __user *, tsops, | |||
1386 | 1441 | ||
1387 | out_unlock_free: | 1442 | out_unlock_free: |
1388 | sem_unlock(sma); | 1443 | sem_unlock(sma); |
1444 | |||
1445 | wake_up_sem_queue_do(&tasks); | ||
1389 | out_free: | 1446 | out_free: |
1390 | if(sops != fast_sops) | 1447 | if(sops != fast_sops) |
1391 | kfree(sops); | 1448 | kfree(sops); |
@@ -1446,6 +1503,7 @@ void exit_sem(struct task_struct *tsk) | |||
1446 | for (;;) { | 1503 | for (;;) { |
1447 | struct sem_array *sma; | 1504 | struct sem_array *sma; |
1448 | struct sem_undo *un; | 1505 | struct sem_undo *un; |
1506 | struct list_head tasks; | ||
1449 | int semid; | 1507 | int semid; |
1450 | int i; | 1508 | int i; |
1451 | 1509 | ||
@@ -1509,10 +1567,11 @@ void exit_sem(struct task_struct *tsk) | |||
1509 | semaphore->sempid = task_tgid_vnr(current); | 1567 | semaphore->sempid = task_tgid_vnr(current); |
1510 | } | 1568 | } |
1511 | } | 1569 | } |
1512 | sma->sem_otime = get_seconds(); | ||
1513 | /* maybe some queued-up processes were waiting for this */ | 1570 | /* maybe some queued-up processes were waiting for this */ |
1514 | update_queue(sma, -1); | 1571 | INIT_LIST_HEAD(&tasks); |
1572 | do_smart_update(sma, NULL, 0, 1, &tasks); | ||
1515 | sem_unlock(sma); | 1573 | sem_unlock(sma); |
1574 | wake_up_sem_queue_do(&tasks); | ||
1516 | 1575 | ||
1517 | call_rcu(&un->rcu, free_un); | 1576 | call_rcu(&un->rcu, free_un); |
1518 | } | 1577 | } |