aboutsummaryrefslogtreecommitdiffstats
path: root/arch
diff options
context:
space:
mode:
Diffstat (limited to 'arch')
-rw-r--r--arch/powerpc/platforms/cell/spufs/context.c7
-rw-r--r--arch/powerpc/platforms/cell/spufs/sched.c127
-rw-r--r--arch/powerpc/platforms/cell/spufs/spufs.h1
3 files changed, 127 insertions, 8 deletions
diff --git a/arch/powerpc/platforms/cell/spufs/context.c b/arch/powerpc/platforms/cell/spufs/context.c
index 6ff2a75589f3..f623d963fdc7 100644
--- a/arch/powerpc/platforms/cell/spufs/context.c
+++ b/arch/powerpc/platforms/cell/spufs/context.c
@@ -23,10 +23,14 @@
23#include <linux/fs.h> 23#include <linux/fs.h>
24#include <linux/mm.h> 24#include <linux/mm.h>
25#include <linux/slab.h> 25#include <linux/slab.h>
26#include <asm/atomic.h>
26#include <asm/spu.h> 27#include <asm/spu.h>
27#include <asm/spu_csa.h> 28#include <asm/spu_csa.h>
28#include "spufs.h" 29#include "spufs.h"
29 30
31
32atomic_t nr_spu_contexts = ATOMIC_INIT(0);
33
30struct spu_context *alloc_spu_context(struct spu_gang *gang) 34struct spu_context *alloc_spu_context(struct spu_gang *gang)
31{ 35{
32 struct spu_context *ctx; 36 struct spu_context *ctx;
@@ -55,6 +59,8 @@ struct spu_context *alloc_spu_context(struct spu_gang *gang)
55 spu_gang_add_ctx(gang, ctx); 59 spu_gang_add_ctx(gang, ctx);
56 ctx->cpus_allowed = current->cpus_allowed; 60 ctx->cpus_allowed = current->cpus_allowed;
57 spu_set_timeslice(ctx); 61 spu_set_timeslice(ctx);
62
63 atomic_inc(&nr_spu_contexts);
58 goto out; 64 goto out;
59out_free: 65out_free:
60 kfree(ctx); 66 kfree(ctx);
@@ -74,6 +80,7 @@ void destroy_spu_context(struct kref *kref)
74 if (ctx->gang) 80 if (ctx->gang)
75 spu_gang_remove_ctx(ctx->gang, ctx); 81 spu_gang_remove_ctx(ctx->gang, ctx);
76 BUG_ON(!list_empty(&ctx->rq)); 82 BUG_ON(!list_empty(&ctx->rq));
83 atomic_dec(&nr_spu_contexts);
77 kfree(ctx); 84 kfree(ctx);
78} 85}
79 86
diff --git a/arch/powerpc/platforms/cell/spufs/sched.c b/arch/powerpc/platforms/cell/spufs/sched.c
index 540067550e88..9fc09306c9ae 100644
--- a/arch/powerpc/platforms/cell/spufs/sched.c
+++ b/arch/powerpc/platforms/cell/spufs/sched.c
@@ -36,6 +36,9 @@
36#include <linux/mutex.h> 36#include <linux/mutex.h>
37#include <linux/notifier.h> 37#include <linux/notifier.h>
38#include <linux/kthread.h> 38#include <linux/kthread.h>
39#include <linux/pid_namespace.h>
40#include <linux/proc_fs.h>
41#include <linux/seq_file.h>
39 42
40#include <asm/io.h> 43#include <asm/io.h>
41#include <asm/mmu_context.h> 44#include <asm/mmu_context.h>
@@ -50,8 +53,11 @@ struct spu_prio_array {
50 spinlock_t runq_lock; 53 spinlock_t runq_lock;
51 struct list_head active_list[MAX_NUMNODES]; 54 struct list_head active_list[MAX_NUMNODES];
52 struct mutex active_mutex[MAX_NUMNODES]; 55 struct mutex active_mutex[MAX_NUMNODES];
56 int nr_active[MAX_NUMNODES];
57 int nr_waiting;
53}; 58};
54 59
60static unsigned long spu_avenrun[3];
55static struct spu_prio_array *spu_prio; 61static struct spu_prio_array *spu_prio;
56static struct task_struct *spusched_task; 62static struct task_struct *spusched_task;
57static struct timer_list spusched_timer; 63static struct timer_list spusched_timer;
@@ -169,14 +175,18 @@ static int node_allowed(struct spu_context *ctx, int node)
169 */ 175 */
170static void spu_add_to_active_list(struct spu *spu) 176static void spu_add_to_active_list(struct spu *spu)
171{ 177{
172 mutex_lock(&spu_prio->active_mutex[spu->node]); 178 int node = spu->node;
173 list_add_tail(&spu->list, &spu_prio->active_list[spu->node]); 179
174 mutex_unlock(&spu_prio->active_mutex[spu->node]); 180 mutex_lock(&spu_prio->active_mutex[node]);
181 spu_prio->nr_active[node]++;
182 list_add_tail(&spu->list, &spu_prio->active_list[node]);
183 mutex_unlock(&spu_prio->active_mutex[node]);
175} 184}
176 185
177static void __spu_remove_from_active_list(struct spu *spu) 186static void __spu_remove_from_active_list(struct spu *spu)
178{ 187{
179 list_del_init(&spu->list); 188 list_del_init(&spu->list);
189 spu_prio->nr_active[spu->node]--;
180} 190}
181 191
182/** 192/**
@@ -275,6 +285,7 @@ static void __spu_add_to_rq(struct spu_context *ctx)
275{ 285{
276 int prio = ctx->prio; 286 int prio = ctx->prio;
277 287
288 spu_prio->nr_waiting++;
278 list_add_tail(&ctx->rq, &spu_prio->runq[prio]); 289 list_add_tail(&ctx->rq, &spu_prio->runq[prio]);
279 set_bit(prio, spu_prio->bitmap); 290 set_bit(prio, spu_prio->bitmap);
280} 291}
@@ -283,8 +294,10 @@ static void __spu_del_from_rq(struct spu_context *ctx)
283{ 294{
284 int prio = ctx->prio; 295 int prio = ctx->prio;
285 296
286 if (!list_empty(&ctx->rq)) 297 if (!list_empty(&ctx->rq)) {
287 list_del_init(&ctx->rq); 298 list_del_init(&ctx->rq);
299 spu_prio->nr_waiting--;
300 }
288 if (list_empty(&spu_prio->runq[prio])) 301 if (list_empty(&spu_prio->runq[prio]))
289 clear_bit(prio, spu_prio->bitmap); 302 clear_bit(prio, spu_prio->bitmap);
290} 303}
@@ -567,10 +580,56 @@ static void spusched_tick(struct spu_context *ctx)
567 } 580 }
568} 581}
569 582
583/**
584 * count_active_contexts - count nr of active tasks
585 *
586 * Return the number of tasks currently running or waiting to run.
587 *
588 * Note that we don't take runq_lock / active_mutex here. Reading
589 * a single 32bit value is atomic on powerpc, and we don't care
590 * about memory ordering issues here.
591 */
592static unsigned long count_active_contexts(void)
593{
594 int nr_active = 0, node;
595
596 for (node = 0; node < MAX_NUMNODES; node++)
597 nr_active += spu_prio->nr_active[node];
598 nr_active += spu_prio->nr_waiting;
599
600 return nr_active;
601}
602
603/**
604 * spu_calc_load - given tick count, update the avenrun load estimates.
605 * @tick: tick count
606 *
607 * No locking against reading these values from userspace, as for
608 * the CPU loadavg code.
609 */
610static void spu_calc_load(unsigned long ticks)
611{
612 unsigned long active_tasks; /* fixed-point */
613 static int count = LOAD_FREQ;
614
615 count -= ticks;
616
617 if (unlikely(count < 0)) {
618 active_tasks = count_active_contexts() * FIXED_1;
619 do {
620 CALC_LOAD(spu_avenrun[0], EXP_1, active_tasks);
621 CALC_LOAD(spu_avenrun[1], EXP_5, active_tasks);
622 CALC_LOAD(spu_avenrun[2], EXP_15, active_tasks);
623 count += LOAD_FREQ;
624 } while (count < 0);
625 }
626}
627
570static void spusched_wake(unsigned long data) 628static void spusched_wake(unsigned long data)
571{ 629{
572 mod_timer(&spusched_timer, jiffies + SPUSCHED_TICK); 630 mod_timer(&spusched_timer, jiffies + SPUSCHED_TICK);
573 wake_up_process(spusched_task); 631 wake_up_process(spusched_task);
632 spu_calc_load(SPUSCHED_TICK);
574} 633}
575 634
576static int spusched_thread(void *unused) 635static int spusched_thread(void *unused)
@@ -598,13 +657,52 @@ static int spusched_thread(void *unused)
598 return 0; 657 return 0;
599} 658}
600 659
660#define LOAD_INT(x) ((x) >> FSHIFT)
661#define LOAD_FRAC(x) LOAD_INT(((x) & (FIXED_1-1)) * 100)
662
663static int show_spu_loadavg(struct seq_file *s, void *private)
664{
665 int a, b, c;
666
667 a = spu_avenrun[0] + (FIXED_1/200);
668 b = spu_avenrun[1] + (FIXED_1/200);
669 c = spu_avenrun[2] + (FIXED_1/200);
670
671 /*
672 * Note that last_pid doesn't really make much sense for the
673 * SPU loadavg (it even seems very odd on the CPU side..),
674 * but we include it here to have a 100% compatible interface.
675 */
676 seq_printf(s, "%d.%02d %d.%02d %d.%02d %ld/%d %d\n",
677 LOAD_INT(a), LOAD_FRAC(a),
678 LOAD_INT(b), LOAD_FRAC(b),
679 LOAD_INT(c), LOAD_FRAC(c),
680 count_active_contexts(),
681 atomic_read(&nr_spu_contexts),
682 current->nsproxy->pid_ns->last_pid);
683 return 0;
684}
685
686static int spu_loadavg_open(struct inode *inode, struct file *file)
687{
688 return single_open(file, show_spu_loadavg, NULL);
689}
690
691static const struct file_operations spu_loadavg_fops = {
692 .open = spu_loadavg_open,
693 .read = seq_read,
694 .llseek = seq_lseek,
695 .release = single_release,
696};
697
601int __init spu_sched_init(void) 698int __init spu_sched_init(void)
602{ 699{
603 int i; 700 struct proc_dir_entry *entry;
701 int err = -ENOMEM, i;
604 702
605 spu_prio = kzalloc(sizeof(struct spu_prio_array), GFP_KERNEL); 703 spu_prio = kzalloc(sizeof(struct spu_prio_array), GFP_KERNEL);
606 if (!spu_prio) 704 if (!spu_prio)
607 return -ENOMEM; 705 goto out;
608 706
609 for (i = 0; i < MAX_PRIO; i++) { 707 for (i = 0; i < MAX_PRIO; i++) {
610 INIT_LIST_HEAD(&spu_prio->runq[i]); 708 INIT_LIST_HEAD(&spu_prio->runq[i]);
@@ -619,14 +717,25 @@ int __init spu_sched_init(void)
619 717
620 spusched_task = kthread_run(spusched_thread, NULL, "spusched"); 718 spusched_task = kthread_run(spusched_thread, NULL, "spusched");
621 if (IS_ERR(spusched_task)) { 719 if (IS_ERR(spusched_task)) {
622 kfree(spu_prio); 720 err = PTR_ERR(spusched_task);
623 return PTR_ERR(spusched_task); 721 goto out_free_spu_prio;
624 } 722 }
625 723
724 entry = create_proc_entry("spu_loadavg", 0, NULL);
725 if (!entry)
726 goto out_stop_kthread;
727 entry->proc_fops = &spu_loadavg_fops;
728
626 pr_debug("spusched: tick: %d, min ticks: %d, default ticks: %d\n", 729 pr_debug("spusched: tick: %d, min ticks: %d, default ticks: %d\n",
627 SPUSCHED_TICK, MIN_SPU_TIMESLICE, DEF_SPU_TIMESLICE); 730 SPUSCHED_TICK, MIN_SPU_TIMESLICE, DEF_SPU_TIMESLICE);
628 return 0; 731 return 0;
629 732
733 out_stop_kthread:
734 kthread_stop(spusched_task);
735 out_free_spu_prio:
736 kfree(spu_prio);
737 out:
738 return err;
630} 739}
631 740
632void __exit spu_sched_exit(void) 741void __exit spu_sched_exit(void)
@@ -634,6 +743,8 @@ void __exit spu_sched_exit(void)
634 struct spu *spu, *tmp; 743 struct spu *spu, *tmp;
635 int node; 744 int node;
636 745
746 remove_proc_entry("spu_loadavg", NULL);
747
637 kthread_stop(spusched_task); 748 kthread_stop(spusched_task);
638 749
639 for (node = 0; node < MAX_NUMNODES; node++) { 750 for (node = 0; node < MAX_NUMNODES; node++) {
diff --git a/arch/powerpc/platforms/cell/spufs/spufs.h b/arch/powerpc/platforms/cell/spufs/spufs.h
index 8ff16b4b6bd7..7f5d0b2fdea3 100644
--- a/arch/powerpc/platforms/cell/spufs/spufs.h
+++ b/arch/powerpc/platforms/cell/spufs/spufs.h
@@ -176,6 +176,7 @@ void spu_gang_add_ctx(struct spu_gang *gang, struct spu_context *ctx);
176int spufs_handle_class1(struct spu_context *ctx); 176int spufs_handle_class1(struct spu_context *ctx);
177 177
178/* context management */ 178/* context management */
179extern atomic_t nr_spu_contexts;
179static inline void spu_acquire(struct spu_context *ctx) 180static inline void spu_acquire(struct spu_context *ctx)
180{ 181{
181 mutex_lock(&ctx->state_mutex); 182 mutex_lock(&ctx->state_mutex);