Loading include/linux/sched_clock.h +1 −1 Original line number Diff line number Diff line Loading @@ -14,7 +14,7 @@ * @sched_clock_mask: Bitmask for two's complement subtraction of non 64bit * clocks. * @read_sched_clock: Current clock source (or dummy source when suspended). * @mult: Multipler for scaled math conversion. * @mult: Multiplier for scaled math conversion. * @shift: Shift value for scaled math conversion. * * Care must be taken when updating this structure; it is read by Loading kernel/sched/core.c +2 −2 Original line number Diff line number Diff line Loading @@ -5506,7 +5506,7 @@ pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) } /* * Try and select tasks for each sibling in decending sched_class * Try and select tasks for each sibling in descending sched_class * order. */ for_each_class(class) { Loading @@ -5520,7 +5520,7 @@ pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) /* * If this sibling doesn't yet have a suitable task to * run; ask for the most elegible task, given the * run; ask for the most eligible task, given the * highest priority task already selected for this * core. */ Loading kernel/sched/fair.c +3 −3 Original line number Diff line number Diff line Loading @@ -10808,11 +10808,11 @@ static inline void task_tick_core(struct rq *rq, struct task_struct *curr) * sched_slice() considers only this active rq and it gets the * whole slice. But during force idle, we have siblings acting * like a single runqueue and hence we need to consider runnable * tasks on this cpu and the forced idle cpu. Ideally, we should * tasks on this CPU and the forced idle CPU. Ideally, we should * go through the forced idle rq, but that would be a perf hit. * We can assume that the forced idle cpu has atleast * We can assume that the forced idle CPU has at least * MIN_NR_TASKS_DURING_FORCEIDLE - 1 tasks and use that to check * if we need to give up the cpu. * if we need to give up the CPU. */ if (rq->core->core_forceidle && rq->cfs.nr_running == 1 && __entity_slice_used(&curr->se, MIN_NR_TASKS_DURING_FORCEIDLE)) Loading Loading
include/linux/sched_clock.h +1 −1 Original line number Diff line number Diff line Loading @@ -14,7 +14,7 @@ * @sched_clock_mask: Bitmask for two's complement subtraction of non 64bit * clocks. * @read_sched_clock: Current clock source (or dummy source when suspended). * @mult: Multipler for scaled math conversion. * @mult: Multiplier for scaled math conversion. * @shift: Shift value for scaled math conversion. * * Care must be taken when updating this structure; it is read by Loading
kernel/sched/core.c +2 −2 Original line number Diff line number Diff line Loading @@ -5506,7 +5506,7 @@ pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) } /* * Try and select tasks for each sibling in decending sched_class * Try and select tasks for each sibling in descending sched_class * order. */ for_each_class(class) { Loading @@ -5520,7 +5520,7 @@ pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) /* * If this sibling doesn't yet have a suitable task to * run; ask for the most elegible task, given the * run; ask for the most eligible task, given the * highest priority task already selected for this * core. */ Loading
kernel/sched/fair.c +3 −3 Original line number Diff line number Diff line Loading @@ -10808,11 +10808,11 @@ static inline void task_tick_core(struct rq *rq, struct task_struct *curr) * sched_slice() considers only this active rq and it gets the * whole slice. But during force idle, we have siblings acting * like a single runqueue and hence we need to consider runnable * tasks on this cpu and the forced idle cpu. Ideally, we should * tasks on this CPU and the forced idle CPU. Ideally, we should * go through the forced idle rq, but that would be a perf hit. * We can assume that the forced idle cpu has atleast * We can assume that the forced idle CPU has at least * MIN_NR_TASKS_DURING_FORCEIDLE - 1 tasks and use that to check * if we need to give up the cpu. * if we need to give up the CPU. */ if (rq->core->core_forceidle && rq->cfs.nr_running == 1 && __entity_slice_used(&curr->se, MIN_NR_TASKS_DURING_FORCEIDLE)) Loading