BACKPORT: sched: Add task_struct pointer to sched_class::set_curr_task
(cherry picked from commit 03b7fad167efc) In preparation of further separating pick_next_task() and set_curr_task() we have to pass the actual task into it, while there, rename the thing to better pair with put_prev_task(). BUG=b:152605392 TEST=run power_VideoCall test Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Cc: Aaron Lu <aaron.lwe@gmail.com> Cc: Valentin Schneider <valentin.schneider@arm.com> Cc: mingo@kernel.org Cc: Phil Auld <pauld@redhat.com> Cc: Julien Desfossez <jdesfossez@digitalocean.com> Cc: Nishanth Aravamudan <naravamudan@digitalocean.com> Link: https://lkml.kernel.org/r/a96d1bcdd716db4a4c5da2fece647a1456c0ed78.1559129225.git.vpillai@digitalocean.com Signed-off-by: Joel Fernandes <joelaf@google.com> Change-Id: Ic4cdd6e6d0863b6cc5c626cb6ad9c5b81f86aff9 Reviewed-on: https://chromium-review.googlesource.com/c/chromiumos/third_party/kernel/+/2186284 Reviewed-by: Sonny Rao <sonnyrao@chromium.org> Signed-off-by: UtsavBalar1231 <utsavbalar1231@gmail.com>
This commit is contained in:
parent
7560b734a5
commit
3f65aa02df
@ -1651,7 +1651,7 @@ void do_set_cpus_allowed(struct task_struct *p, const struct cpumask *new_mask)
|
||||
if (queued)
|
||||
enqueue_task(rq, p, ENQUEUE_RESTORE | ENQUEUE_NOCLOCK);
|
||||
if (running)
|
||||
set_curr_task(rq, p);
|
||||
set_next_task(rq, p);
|
||||
}
|
||||
|
||||
/*
|
||||
@ -4676,7 +4676,7 @@ void rt_mutex_setprio(struct task_struct *p, struct task_struct *pi_task)
|
||||
if (queued)
|
||||
enqueue_task(rq, p, queue_flag);
|
||||
if (running)
|
||||
set_curr_task(rq, p);
|
||||
set_next_task(rq, p);
|
||||
|
||||
check_class_changed(rq, p, prev_class, oldprio);
|
||||
out_unlock:
|
||||
@ -4743,7 +4743,7 @@ void set_user_nice(struct task_struct *p, long nice)
|
||||
resched_curr(rq);
|
||||
}
|
||||
if (running)
|
||||
set_curr_task(rq, p);
|
||||
set_next_task(rq, p);
|
||||
out_unlock:
|
||||
task_rq_unlock(rq, p, &rf);
|
||||
}
|
||||
@ -5186,7 +5186,7 @@ change:
|
||||
enqueue_task(rq, p, queue_flags);
|
||||
}
|
||||
if (running)
|
||||
set_curr_task(rq, p);
|
||||
set_next_task(rq, p);
|
||||
|
||||
check_class_changed(rq, p, prev_class, oldprio);
|
||||
|
||||
@ -6465,7 +6465,7 @@ void sched_setnuma(struct task_struct *p, int nid)
|
||||
if (queued)
|
||||
enqueue_task(rq, p, ENQUEUE_RESTORE | ENQUEUE_NOCLOCK);
|
||||
if (running)
|
||||
set_curr_task(rq, p);
|
||||
set_next_task(rq, p);
|
||||
task_rq_unlock(rq, p, &rf);
|
||||
}
|
||||
#endif /* CONFIG_NUMA_BALANCING */
|
||||
@ -7651,7 +7651,7 @@ void sched_move_task(struct task_struct *tsk)
|
||||
if (queued)
|
||||
enqueue_task(rq, tsk, queue_flags);
|
||||
if (running)
|
||||
set_curr_task(rq, tsk);
|
||||
set_next_task(rq, tsk);
|
||||
|
||||
task_rq_unlock(rq, tsk, &rf);
|
||||
}
|
||||
|
@ -1873,11 +1873,6 @@ static void task_fork_dl(struct task_struct *p)
|
||||
*/
|
||||
}
|
||||
|
||||
static void set_curr_task_dl(struct rq *rq)
|
||||
{
|
||||
set_next_task_dl(rq, rq->curr);
|
||||
}
|
||||
|
||||
#ifdef CONFIG_SMP
|
||||
|
||||
/* Only try algorithms three times */
|
||||
@ -2469,6 +2464,7 @@ const struct sched_class dl_sched_class = {
|
||||
|
||||
.pick_next_task = pick_next_task_dl,
|
||||
.put_prev_task = put_prev_task_dl,
|
||||
.set_next_task = set_next_task_dl,
|
||||
|
||||
#ifdef CONFIG_SMP
|
||||
.select_task_rq = select_task_rq_dl,
|
||||
@ -2479,7 +2475,6 @@ const struct sched_class dl_sched_class = {
|
||||
.task_woken = task_woken_dl,
|
||||
#endif
|
||||
|
||||
.set_curr_task = set_curr_task_dl,
|
||||
.task_tick = task_tick_dl,
|
||||
.task_fork = task_fork_dl,
|
||||
|
||||
|
@ -12284,9 +12284,19 @@ static void switched_to_fair(struct rq *rq, struct task_struct *p)
|
||||
* This routine is mostly called to set cfs_rq->curr field when a task
|
||||
* migrates between groups/classes.
|
||||
*/
|
||||
static void set_curr_task_fair(struct rq *rq)
|
||||
static void set_next_task_fair(struct rq *rq, struct task_struct *p)
|
||||
{
|
||||
struct sched_entity *se = &rq->curr->se;
|
||||
struct sched_entity *se = &p->se;
|
||||
|
||||
#ifdef CONFIG_SMP
|
||||
if (task_on_rq_queued(p)) {
|
||||
/*
|
||||
* Move the next running task to the front of the list, so our
|
||||
* cfs_tasks list becomes MRU one.
|
||||
*/
|
||||
list_move(&se->group_node, &rq->cfs_tasks);
|
||||
}
|
||||
#endif
|
||||
|
||||
for_each_sched_entity(se) {
|
||||
struct cfs_rq *cfs_rq = cfs_rq_of(se);
|
||||
@ -12557,7 +12567,9 @@ const struct sched_class fair_sched_class = {
|
||||
.check_preempt_curr = check_preempt_wakeup,
|
||||
|
||||
.pick_next_task = pick_next_task_fair,
|
||||
|
||||
.put_prev_task = put_prev_task_fair,
|
||||
.set_next_task = set_next_task_fair,
|
||||
|
||||
#ifdef CONFIG_SMP
|
||||
.select_task_rq = select_task_rq_fair,
|
||||
@ -12570,7 +12582,6 @@ const struct sched_class fair_sched_class = {
|
||||
.set_cpus_allowed = set_cpus_allowed_common,
|
||||
#endif
|
||||
|
||||
.set_curr_task = set_curr_task_fair,
|
||||
.task_tick = task_tick_fair,
|
||||
.task_fork = task_fork_fair,
|
||||
|
||||
|
@ -393,14 +393,25 @@ static void check_preempt_curr_idle(struct rq *rq, struct task_struct *p, int fl
|
||||
resched_curr(rq);
|
||||
}
|
||||
|
||||
static void put_prev_task_idle(struct rq *rq, struct task_struct *prev)
|
||||
{
|
||||
}
|
||||
|
||||
static void set_next_task_idle(struct rq *rq, struct task_struct *next)
|
||||
{
|
||||
update_idle_core(rq);
|
||||
schedstat_inc(rq->sched_goidle);
|
||||
}
|
||||
|
||||
static struct task_struct *
|
||||
pick_next_task_idle(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
|
||||
{
|
||||
put_prev_task(rq, prev);
|
||||
update_idle_core(rq);
|
||||
schedstat_inc(rq->sched_goidle);
|
||||
struct task_struct *next = rq->idle;
|
||||
|
||||
return rq->idle;
|
||||
put_prev_task(rq, prev);
|
||||
set_next_task_idle(rq, next);
|
||||
|
||||
return next;
|
||||
}
|
||||
|
||||
/*
|
||||
@ -416,10 +427,6 @@ dequeue_task_idle(struct rq *rq, struct task_struct *p, int flags)
|
||||
raw_spin_lock_irq(&rq->lock);
|
||||
}
|
||||
|
||||
static void put_prev_task_idle(struct rq *rq, struct task_struct *prev)
|
||||
{
|
||||
}
|
||||
|
||||
/*
|
||||
* scheduler tick hitting a task of our scheduling class.
|
||||
*
|
||||
@ -432,10 +439,6 @@ static void task_tick_idle(struct rq *rq, struct task_struct *curr, int queued)
|
||||
{
|
||||
}
|
||||
|
||||
static void set_curr_task_idle(struct rq *rq)
|
||||
{
|
||||
}
|
||||
|
||||
static void switched_to_idle(struct rq *rq, struct task_struct *p)
|
||||
{
|
||||
BUG();
|
||||
@ -470,13 +473,13 @@ const struct sched_class idle_sched_class = {
|
||||
|
||||
.pick_next_task = pick_next_task_idle,
|
||||
.put_prev_task = put_prev_task_idle,
|
||||
.set_next_task = set_next_task_idle,
|
||||
|
||||
#ifdef CONFIG_SMP
|
||||
.select_task_rq = select_task_rq_idle,
|
||||
.set_cpus_allowed = set_cpus_allowed_common,
|
||||
#endif
|
||||
|
||||
.set_curr_task = set_curr_task_idle,
|
||||
.task_tick = task_tick_idle,
|
||||
|
||||
.get_rr_interval = get_rr_interval_idle,
|
||||
|
@ -2599,11 +2599,6 @@ static void task_tick_rt(struct rq *rq, struct task_struct *p, int queued)
|
||||
}
|
||||
}
|
||||
|
||||
static void set_curr_task_rt(struct rq *rq)
|
||||
{
|
||||
set_next_task_rt(rq, rq->curr);
|
||||
}
|
||||
|
||||
static unsigned int get_rr_interval_rt(struct rq *rq, struct task_struct *task)
|
||||
{
|
||||
/*
|
||||
@ -2625,6 +2620,7 @@ const struct sched_class rt_sched_class = {
|
||||
|
||||
.pick_next_task = pick_next_task_rt,
|
||||
.put_prev_task = put_prev_task_rt,
|
||||
.set_next_task = set_next_task_rt,
|
||||
|
||||
#ifdef CONFIG_SMP
|
||||
.select_task_rq = select_task_rq_rt,
|
||||
@ -2636,7 +2632,6 @@ const struct sched_class rt_sched_class = {
|
||||
.switched_from = switched_from_rt,
|
||||
#endif
|
||||
|
||||
.set_curr_task = set_curr_task_rt,
|
||||
.task_tick = task_tick_rt,
|
||||
|
||||
.get_rr_interval = get_rr_interval_rt,
|
||||
|
@ -1855,6 +1855,7 @@ struct sched_class {
|
||||
struct task_struct *prev,
|
||||
struct rq_flags *rf);
|
||||
void (*put_prev_task)(struct rq *rq, struct task_struct *p);
|
||||
void (*set_next_task)(struct rq *rq, struct task_struct *p);
|
||||
|
||||
#ifdef CONFIG_SMP
|
||||
int (*select_task_rq)(struct task_struct *p, int task_cpu, int sd_flag, int flags,
|
||||
@ -1870,7 +1871,6 @@ struct sched_class {
|
||||
void (*rq_offline)(struct rq *rq);
|
||||
#endif
|
||||
|
||||
void (*set_curr_task)(struct rq *rq);
|
||||
void (*task_tick)(struct rq *rq, struct task_struct *p, int queued);
|
||||
void (*task_fork)(struct task_struct *p);
|
||||
void (*task_dead)(struct task_struct *p);
|
||||
@ -1910,9 +1910,10 @@ static inline void put_prev_task(struct rq *rq, struct task_struct *prev)
|
||||
prev->sched_class->put_prev_task(rq, prev);
|
||||
}
|
||||
|
||||
static inline void set_curr_task(struct rq *rq, struct task_struct *curr)
|
||||
static inline void set_next_task(struct rq *rq, struct task_struct *next)
|
||||
{
|
||||
curr->sched_class->set_curr_task(rq);
|
||||
WARN_ON_ONCE(rq->curr != next);
|
||||
next->sched_class->set_next_task(rq, next);
|
||||
}
|
||||
|
||||
#ifdef CONFIG_SMP
|
||||
|
@ -25,6 +25,11 @@ check_preempt_curr_stop(struct rq *rq, struct task_struct *p, int flags)
|
||||
/* we're never preempted */
|
||||
}
|
||||
|
||||
static void set_next_task_stop(struct rq *rq, struct task_struct *stop)
|
||||
{
|
||||
stop->se.exec_start = rq_clock_task(rq);
|
||||
}
|
||||
|
||||
static struct task_struct *
|
||||
pick_next_task_stop(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
|
||||
{
|
||||
@ -34,8 +39,7 @@ pick_next_task_stop(struct rq *rq, struct task_struct *prev, struct rq_flags *rf
|
||||
return NULL;
|
||||
|
||||
put_prev_task(rq, prev);
|
||||
|
||||
stop->se.exec_start = rq_clock_task(rq);
|
||||
set_next_task_stop(rq, stop);
|
||||
|
||||
return stop;
|
||||
}
|
||||
@ -90,13 +94,6 @@ static void task_tick_stop(struct rq *rq, struct task_struct *curr, int queued)
|
||||
{
|
||||
}
|
||||
|
||||
static void set_curr_task_stop(struct rq *rq)
|
||||
{
|
||||
struct task_struct *stop = rq->stop;
|
||||
|
||||
stop->se.exec_start = rq_clock_task(rq);
|
||||
}
|
||||
|
||||
static void switched_to_stop(struct rq *rq, struct task_struct *p)
|
||||
{
|
||||
BUG(); /* its impossible to change to this class */
|
||||
@ -132,13 +129,13 @@ const struct sched_class stop_sched_class = {
|
||||
|
||||
.pick_next_task = pick_next_task_stop,
|
||||
.put_prev_task = put_prev_task_stop,
|
||||
.set_next_task = set_next_task_stop,
|
||||
|
||||
#ifdef CONFIG_SMP
|
||||
.select_task_rq = select_task_rq_stop,
|
||||
.set_cpus_allowed = set_cpus_allowed_common,
|
||||
#endif
|
||||
|
||||
.set_curr_task = set_curr_task_stop,
|
||||
.task_tick = task_tick_stop,
|
||||
|
||||
.get_rr_interval = get_rr_interval_stop,
|
||||
|
Loading…
Reference in New Issue
Block a user