Loading kernel/sched_fair.c +5 −2 Original line number Diff line number Diff line Loading @@ -818,7 +818,7 @@ static void check_preempt_wakeup(struct rq *rq, struct task_struct *p) struct task_struct *curr = rq->curr; struct cfs_rq *cfs_rq = task_cfs_rq(curr); struct sched_entity *se = &curr->se, *pse = &p->se; s64 delta; s64 delta, gran; if (unlikely(rt_prio(p->prio))) { update_rq_clock(rq); Loading @@ -833,8 +833,11 @@ static void check_preempt_wakeup(struct rq *rq, struct task_struct *p) } delta = se->vruntime - pse->vruntime; gran = sysctl_sched_wakeup_granularity; if (unlikely(se->load.weight != NICE_0_LOAD)) gran = calc_delta_fair(gran, &se->load); if (delta > (s64)sysctl_sched_wakeup_granularity) if (delta > gran) resched_task(curr); } Loading Loading
kernel/sched_fair.c +5 −2 Original line number Diff line number Diff line Loading @@ -818,7 +818,7 @@ static void check_preempt_wakeup(struct rq *rq, struct task_struct *p) struct task_struct *curr = rq->curr; struct cfs_rq *cfs_rq = task_cfs_rq(curr); struct sched_entity *se = &curr->se, *pse = &p->se; s64 delta; s64 delta, gran; if (unlikely(rt_prio(p->prio))) { update_rq_clock(rq); Loading @@ -833,8 +833,11 @@ static void check_preempt_wakeup(struct rq *rq, struct task_struct *p) } delta = se->vruntime - pse->vruntime; gran = sysctl_sched_wakeup_granularity; if (unlikely(se->load.weight != NICE_0_LOAD)) gran = calc_delta_fair(gran, &se->load); if (delta > (s64)sysctl_sched_wakeup_granularity) if (delta > gran) resched_task(curr); } Loading