Loading kernel/sched_fair.c +4 −129 Original line number Diff line number Diff line Loading @@ -408,64 +408,6 @@ static u64 sched_vslice_add(struct cfs_rq *cfs_rq, struct sched_entity *se) return __sched_period(nr_running); } /* * The goal of calc_delta_asym() is to be asymmetrically around NICE_0_LOAD, in * that it favours >=0 over <0. * * -20 | * | * 0 --------+------- * .' * 19 .' * */ static unsigned long calc_delta_asym(unsigned long delta, struct sched_entity *se) { struct load_weight lw = { .weight = NICE_0_LOAD, .inv_weight = 1UL << (WMULT_SHIFT-NICE_0_SHIFT) }; for_each_sched_entity(se) { struct load_weight *se_lw = &se->load; unsigned long rw = cfs_rq_of(se)->load.weight; #ifdef CONFIG_FAIR_SCHED_GROUP struct cfs_rq *cfs_rq = se->my_q; struct task_group *tg = NULL if (cfs_rq) tg = cfs_rq->tg; if (tg && tg->shares < NICE_0_LOAD) { /* * scale shares to what it would have been had * tg->weight been NICE_0_LOAD: * * weight = 1024 * shares / tg->weight */ lw.weight *= se->load.weight; lw.weight /= tg->shares; lw.inv_weight = 0; se_lw = &lw; rw += lw.weight - se->load.weight; } else #endif if (se->load.weight < NICE_0_LOAD) { se_lw = &lw; rw += NICE_0_LOAD - se->load.weight; } delta = calc_delta_mine(delta, rw, se_lw); } return delta; } /* * Update the current task's runtime statistics. Skip current tasks that * are not in our scheduling class. Loading Loading @@ -1281,53 +1223,11 @@ static unsigned long wakeup_gran(struct sched_entity *se) * + nice tasks. */ if (sched_feat(ASYM_GRAN)) gran = calc_delta_asym(sysctl_sched_wakeup_granularity, se); else gran = calc_delta_fair(sysctl_sched_wakeup_granularity, se); gran = calc_delta_mine(gran, NICE_0_LOAD, &se->load); return gran; } /* * Should 'se' preempt 'curr'. * * |s1 * |s2 * |s3 * g * |<--->|c * * w(c, s1) = -1 * w(c, s2) = 0 * w(c, s3) = 1 * */ static int wakeup_preempt_entity(struct sched_entity *curr, struct sched_entity *se) { s64 gran, vdiff = curr->vruntime - se->vruntime; if (vdiff < 0) return -1; gran = wakeup_gran(curr); if (vdiff > gran) return 1; return 0; } /* return depth at which a sched entity is present in the hierarchy */ static inline int depth_se(struct sched_entity *se) { int depth = 0; for_each_sched_entity(se) depth++; return depth; } /* * Preempt the current task with a newly woken task if needed: */ Loading @@ -1336,7 +1236,7 @@ static void check_preempt_wakeup(struct rq *rq, struct task_struct *p, int sync) struct task_struct *curr = rq->curr; struct cfs_rq *cfs_rq = task_cfs_rq(curr); struct sched_entity *se = &curr->se, *pse = &p->se; int se_depth, pse_depth; s64 delta_exec; if (unlikely(rt_prio(p->prio))) { update_rq_clock(rq); Loading Loading @@ -1374,33 +1274,8 @@ static void check_preempt_wakeup(struct rq *rq, struct task_struct *p, int sync) return; } /* * preemption test can be made between sibling entities who are in the * same cfs_rq i.e who have a common parent. Walk up the hierarchy of * both tasks until we find their ancestors who are siblings of common * parent. */ /* First walk up until both entities are at same depth */ se_depth = depth_se(se); pse_depth = depth_se(pse); while (se_depth > pse_depth) { se_depth--; se = parent_entity(se); } while (pse_depth > se_depth) { pse_depth--; pse = parent_entity(pse); } while (!is_same_group(se, pse)) { se = parent_entity(se); pse = parent_entity(pse); } if (wakeup_preempt_entity(se, pse) == 1) delta_exec = se->sum_exec_runtime - se->prev_sum_exec_runtime; if (delta_exec > wakeup_gran(pse)) resched_task(curr); } Loading Loading
kernel/sched_fair.c +4 −129 Original line number Diff line number Diff line Loading @@ -408,64 +408,6 @@ static u64 sched_vslice_add(struct cfs_rq *cfs_rq, struct sched_entity *se) return __sched_period(nr_running); } /* * The goal of calc_delta_asym() is to be asymmetrically around NICE_0_LOAD, in * that it favours >=0 over <0. * * -20 | * | * 0 --------+------- * .' * 19 .' * */ static unsigned long calc_delta_asym(unsigned long delta, struct sched_entity *se) { struct load_weight lw = { .weight = NICE_0_LOAD, .inv_weight = 1UL << (WMULT_SHIFT-NICE_0_SHIFT) }; for_each_sched_entity(se) { struct load_weight *se_lw = &se->load; unsigned long rw = cfs_rq_of(se)->load.weight; #ifdef CONFIG_FAIR_SCHED_GROUP struct cfs_rq *cfs_rq = se->my_q; struct task_group *tg = NULL if (cfs_rq) tg = cfs_rq->tg; if (tg && tg->shares < NICE_0_LOAD) { /* * scale shares to what it would have been had * tg->weight been NICE_0_LOAD: * * weight = 1024 * shares / tg->weight */ lw.weight *= se->load.weight; lw.weight /= tg->shares; lw.inv_weight = 0; se_lw = &lw; rw += lw.weight - se->load.weight; } else #endif if (se->load.weight < NICE_0_LOAD) { se_lw = &lw; rw += NICE_0_LOAD - se->load.weight; } delta = calc_delta_mine(delta, rw, se_lw); } return delta; } /* * Update the current task's runtime statistics. Skip current tasks that * are not in our scheduling class. Loading Loading @@ -1281,53 +1223,11 @@ static unsigned long wakeup_gran(struct sched_entity *se) * + nice tasks. */ if (sched_feat(ASYM_GRAN)) gran = calc_delta_asym(sysctl_sched_wakeup_granularity, se); else gran = calc_delta_fair(sysctl_sched_wakeup_granularity, se); gran = calc_delta_mine(gran, NICE_0_LOAD, &se->load); return gran; } /* * Should 'se' preempt 'curr'. * * |s1 * |s2 * |s3 * g * |<--->|c * * w(c, s1) = -1 * w(c, s2) = 0 * w(c, s3) = 1 * */ static int wakeup_preempt_entity(struct sched_entity *curr, struct sched_entity *se) { s64 gran, vdiff = curr->vruntime - se->vruntime; if (vdiff < 0) return -1; gran = wakeup_gran(curr); if (vdiff > gran) return 1; return 0; } /* return depth at which a sched entity is present in the hierarchy */ static inline int depth_se(struct sched_entity *se) { int depth = 0; for_each_sched_entity(se) depth++; return depth; } /* * Preempt the current task with a newly woken task if needed: */ Loading @@ -1336,7 +1236,7 @@ static void check_preempt_wakeup(struct rq *rq, struct task_struct *p, int sync) struct task_struct *curr = rq->curr; struct cfs_rq *cfs_rq = task_cfs_rq(curr); struct sched_entity *se = &curr->se, *pse = &p->se; int se_depth, pse_depth; s64 delta_exec; if (unlikely(rt_prio(p->prio))) { update_rq_clock(rq); Loading Loading @@ -1374,33 +1274,8 @@ static void check_preempt_wakeup(struct rq *rq, struct task_struct *p, int sync) return; } /* * preemption test can be made between sibling entities who are in the * same cfs_rq i.e who have a common parent. Walk up the hierarchy of * both tasks until we find their ancestors who are siblings of common * parent. */ /* First walk up until both entities are at same depth */ se_depth = depth_se(se); pse_depth = depth_se(pse); while (se_depth > pse_depth) { se_depth--; se = parent_entity(se); } while (pse_depth > se_depth) { pse_depth--; pse = parent_entity(pse); } while (!is_same_group(se, pse)) { se = parent_entity(se); pse = parent_entity(pse); } if (wakeup_preempt_entity(se, pse) == 1) delta_exec = se->sum_exec_runtime - se->prev_sum_exec_runtime; if (delta_exec > wakeup_gran(pse)) resched_task(curr); } Loading