/* CLEANUP: Add comments and make it less messy. * */ #ifndef __UNC_RT_DOMAIN_H__ #define __UNC_RT_DOMAIN_H__ #include #define RELEASE_QUEUE_SLOTS 127 /* prime */ struct _rt_domain; typedef int (*check_resched_needed_t)(struct _rt_domain *rt); typedef void (*release_jobs_t)(struct _rt_domain *rt, struct bheap* tasks); struct release_queue { /* each slot maintains a list of release heaps sorted * by release time */ struct list_head slot[RELEASE_QUEUE_SLOTS]; }; typedef struct _rt_domain { /* runnable rt tasks are in here */ raw_spinlock_t ready_lock; struct bheap ready_queue; /* real-time tasks waiting for release are in here */ raw_spinlock_t release_lock; struct release_queue release_queue; int release_master; /* for moving tasks to the release queue */ raw_spinlock_t tobe_lock; struct list_head tobe_released; /* how do we check if we need to kick another CPU? */ check_resched_needed_t check_resched; /* how do we release jobs? */ release_jobs_t release_jobs; /* how are tasks ordered in the ready queue? */ bheap_prio_t order; } rt_domain_t; struct release_heap { /* list_head for per-time-slot list */ struct list_head list; lt_t release_time; /* all tasks to be released at release_time */ struct bheap heap; /* used to trigger the release */ struct hrtimer timer; /* used to delegate releases */ struct hrtimer_start_on_info info; /* required for the timer callback */ rt_domain_t* dom; }; static inline struct task_struct* __next_ready(rt_domain_t* rt) { struct bheap_node *hn = bheap_peek(rt->order, &rt->ready_queue); if (hn) return bheap2task(hn); else return NULL; } void rt_domain_init(rt_domain_t *rt, bheap_prio_t order, check_resched_needed_t check, release_jobs_t relase); void __add_ready(rt_domain_t* rt, struct task_struct *new); void __merge_ready(rt_domain_t* rt, struct bheap *tasks); void __add_release(rt_domain_t* rt, struct task_struct *task); static inline struct task_struct* __take_ready(rt_domain_t* rt) { struct bheap_node* hn = bheap_take(rt->order, &rt->ready_queue); if (hn) return bheap2task(hn); else return NULL; } static inline struct task_struct* __peek_ready(rt_domain_t* rt) { struct bheap_node* hn = bheap_peek(rt->order, &rt->ready_queue); if (hn) return bheap2task(hn); else return NULL; } static inline int is_queued(struct task_struct *t) { BUG_ON(!tsk_rt(t)->heap_node); return bheap_node_in_heap(tsk_rt(t)->heap_node); } static inline void remove(rt_domain_t* rt, struct task_struct *t) { bheap_delete(rt->order, &rt->ready_queue, tsk_rt(t)->heap_node); } static inline void add_ready(rt_domain_t* rt, struct task_struct *new) { unsigned long flags; /* first we need the write lock for rt_ready_queue */ raw_spin_lock_irqsave(&rt->ready_lock, flags); __add_ready(rt, new); raw_spin_unlock_irqrestore(&rt->ready_lock, flags); } static inline void merge_ready(rt_domain_t* rt, struct bheap* tasks) { unsigned long flags; raw_spin_lock_irqsave(&rt->ready_lock, flags); __merge_ready(rt, tasks); raw_spin_unlock_irqrestore(&rt->ready_lock, flags); } static inline struct task_struct* take_ready(rt_domain_t* rt) { unsigned long flags; struct task_struct* ret; /* first we need the write lock for rt_ready_queue */ raw_spin_lock_irqsave(&rt->ready_lock, flags); ret = __take_ready(rt); raw_spin_unlock_irqrestore(&rt->ready_lock, flags); return ret; } static inline void add_release(rt_domain_t* rt, struct task_struct *task) { unsigned long flags; /* first we need the write lock for rt_ready_queue */ raw_spin_lock_irqsave(&rt->tobe_lock, flags); __add_release(rt, task); raw_spin_unlock_irqrestore(&rt->tobe_lock, flags); } static inline int __jobs_pending(rt_domain_t* rt) { return !bheap_empty(&rt->ready_queue); } static inline int jobs_pending(rt_domain_t* rt) { unsigned long flags; int ret; /* first we need the write lock for rt_ready_queue */ raw_spin_lock_irqsave(&rt->ready_lock, flags); ret = !bheap_empty(&rt->ready_queue); raw_spin_unlock_irqrestore(&rt->ready_lock, flags); return ret; } #endif