forked from Minki/linux
sched/deadline: Add SCHED_DEADLINE SMP-related data structures & logic
Introduces data structures relevant for implementing dynamic migration of -deadline tasks and the logic for checking if runqueues are overloaded with -deadline tasks and for choosing where a task should migrate, when it is the case. Adds also dynamic migrations to SCHED_DEADLINE, so that tasks can be moved among CPUs when necessary. It is also possible to bind a task to a (set of) CPU(s), thus restricting its capability of migrating, or forbidding migrations at all. The very same approach used in sched_rt is utilised: - -deadline tasks are kept into CPU-specific runqueues, - -deadline tasks are migrated among runqueues to achieve the following: * on an M-CPU system the M earliest deadline ready tasks are always running; * affinity/cpusets settings of all the -deadline tasks is always respected. Therefore, this very special form of "load balancing" is done with an active method, i.e., the scheduler pushes or pulls tasks between runqueues when they are woken up and/or (de)scheduled. IOW, every time a preemption occurs, the descheduled task might be sent to some other CPU (depending on its deadline) to continue executing (push). On the other hand, every time a CPU becomes idle, it might pull the second earliest deadline ready task from some other CPU. To enforce this, a pull operation is always attempted before taking any scheduling decision (pre_schedule()), as well as a push one after each scheduling decision (post_schedule()). In addition, when a task arrives or wakes up, the best CPU where to resume it is selected taking into account its affinity mask, the system topology, but also its deadline. E.g., from the scheduling point of view, the best CPU where to wake up (and also where to push) a task is the one which is running the task with the latest deadline among the M executing ones. In order to facilitate these decisions, per-runqueue "caching" of the deadlines of the currently running and of the first ready task is used. Queued but not running tasks are also parked in another rb-tree to speed-up pushes. Signed-off-by: Juri Lelli <juri.lelli@gmail.com> Signed-off-by: Dario Faggioli <raistlin@linux.it> Signed-off-by: Peter Zijlstra <peterz@infradead.org> Link: http://lkml.kernel.org/r/1383831828-15501-5-git-send-email-juri.lelli@gmail.com Signed-off-by: Ingo Molnar <mingo@kernel.org>
This commit is contained in:
parent
aab03e05e8
commit
1baca4ce16
@ -1201,6 +1201,7 @@ struct task_struct {
|
||||
struct list_head tasks;
|
||||
#ifdef CONFIG_SMP
|
||||
struct plist_node pushable_tasks;
|
||||
struct rb_node pushable_dl_tasks;
|
||||
#endif
|
||||
|
||||
struct mm_struct *mm, *active_mm;
|
||||
|
@ -1848,6 +1848,7 @@ int sched_fork(unsigned long clone_flags, struct task_struct *p)
|
||||
init_task_preempt_count(p);
|
||||
#ifdef CONFIG_SMP
|
||||
plist_node_init(&p->pushable_tasks, MAX_PRIO);
|
||||
RB_CLEAR_NODE(&p->pushable_dl_tasks);
|
||||
#endif
|
||||
|
||||
put_cpu();
|
||||
@ -5040,6 +5041,7 @@ static void free_rootdomain(struct rcu_head *rcu)
|
||||
struct root_domain *rd = container_of(rcu, struct root_domain, rcu);
|
||||
|
||||
cpupri_cleanup(&rd->cpupri);
|
||||
free_cpumask_var(rd->dlo_mask);
|
||||
free_cpumask_var(rd->rto_mask);
|
||||
free_cpumask_var(rd->online);
|
||||
free_cpumask_var(rd->span);
|
||||
@ -5091,8 +5093,10 @@ static int init_rootdomain(struct root_domain *rd)
|
||||
goto out;
|
||||
if (!alloc_cpumask_var(&rd->online, GFP_KERNEL))
|
||||
goto free_span;
|
||||
if (!alloc_cpumask_var(&rd->rto_mask, GFP_KERNEL))
|
||||
if (!alloc_cpumask_var(&rd->dlo_mask, GFP_KERNEL))
|
||||
goto free_online;
|
||||
if (!alloc_cpumask_var(&rd->rto_mask, GFP_KERNEL))
|
||||
goto free_dlo_mask;
|
||||
|
||||
if (cpupri_init(&rd->cpupri) != 0)
|
||||
goto free_rto_mask;
|
||||
@ -5100,6 +5104,8 @@ static int init_rootdomain(struct root_domain *rd)
|
||||
|
||||
free_rto_mask:
|
||||
free_cpumask_var(rd->rto_mask);
|
||||
free_dlo_mask:
|
||||
free_cpumask_var(rd->dlo_mask);
|
||||
free_online:
|
||||
free_cpumask_var(rd->online);
|
||||
free_span:
|
||||
@ -6451,6 +6457,7 @@ void __init sched_init_smp(void)
|
||||
free_cpumask_var(non_isolated_cpus);
|
||||
|
||||
init_sched_rt_class();
|
||||
init_sched_dl_class();
|
||||
}
|
||||
#else
|
||||
void __init sched_init_smp(void)
|
||||
|
File diff suppressed because it is too large
Load Diff
@ -1738,7 +1738,7 @@ static void task_woken_rt(struct rq *rq, struct task_struct *p)
|
||||
!test_tsk_need_resched(rq->curr) &&
|
||||
has_pushable_tasks(rq) &&
|
||||
p->nr_cpus_allowed > 1 &&
|
||||
rt_task(rq->curr) &&
|
||||
(dl_task(rq->curr) || rt_task(rq->curr)) &&
|
||||
(rq->curr->nr_cpus_allowed < 2 ||
|
||||
rq->curr->prio <= p->prio))
|
||||
push_rt_tasks(rq);
|
||||
|
@ -385,6 +385,31 @@ struct dl_rq {
|
||||
struct rb_node *rb_leftmost;
|
||||
|
||||
unsigned long dl_nr_running;
|
||||
|
||||
#ifdef CONFIG_SMP
|
||||
/*
|
||||
* Deadline values of the currently executing and the
|
||||
* earliest ready task on this rq. Caching these facilitates
|
||||
* the decision wether or not a ready but not running task
|
||||
* should migrate somewhere else.
|
||||
*/
|
||||
struct {
|
||||
u64 curr;
|
||||
u64 next;
|
||||
} earliest_dl;
|
||||
|
||||
unsigned long dl_nr_migratory;
|
||||
unsigned long dl_nr_total;
|
||||
int overloaded;
|
||||
|
||||
/*
|
||||
* Tasks on this rq that can be pushed away. They are kept in
|
||||
* an rb-tree, ordered by tasks' deadlines, with caching
|
||||
* of the leftmost (earliest deadline) element.
|
||||
*/
|
||||
struct rb_root pushable_dl_tasks_root;
|
||||
struct rb_node *pushable_dl_tasks_leftmost;
|
||||
#endif
|
||||
};
|
||||
|
||||
#ifdef CONFIG_SMP
|
||||
@ -404,6 +429,13 @@ struct root_domain {
|
||||
cpumask_var_t span;
|
||||
cpumask_var_t online;
|
||||
|
||||
/*
|
||||
* The bit corresponding to a CPU gets set here if such CPU has more
|
||||
* than one runnable -deadline task (as it is below for RT tasks).
|
||||
*/
|
||||
cpumask_var_t dlo_mask;
|
||||
atomic_t dlo_count;
|
||||
|
||||
/*
|
||||
* The "RT overload" flag: it gets set if a CPU has more than
|
||||
* one runnable RT task.
|
||||
@ -1095,6 +1127,8 @@ static inline void idle_balance(int cpu, struct rq *rq)
|
||||
extern void sysrq_sched_debug_show(void);
|
||||
extern void sched_init_granularity(void);
|
||||
extern void update_max_interval(void);
|
||||
|
||||
extern void init_sched_dl_class(void);
|
||||
extern void init_sched_rt_class(void);
|
||||
extern void init_sched_fair_class(void);
|
||||
|
||||
|
Loading…
Reference in New Issue
Block a user