Loading kernel/sched/core.c +0 −7 Original line number Diff line number Diff line Loading @@ -7136,10 +7136,6 @@ void set_curr_task(int cpu, struct task_struct *p) #endif #ifdef CONFIG_RT_GROUP_SCHED #else /* !CONFIG_RT_GROUP_SCHED */ #endif /* CONFIG_RT_GROUP_SCHED */ #ifdef CONFIG_CGROUP_SCHED /* task_group_lock serializes the addition/removal of task groups */ static DEFINE_SPINLOCK(task_group_lock); Loading Loading @@ -7248,9 +7244,6 @@ void sched_move_task(struct task_struct *tsk) } #endif /* CONFIG_CGROUP_SCHED */ #ifdef CONFIG_FAIR_GROUP_SCHED #endif #if defined(CONFIG_RT_GROUP_SCHED) || defined(CONFIG_CFS_BANDWIDTH) static unsigned long to_ratio(u64 period, u64 runtime) { Loading Loading
kernel/sched/core.c +0 −7 Original line number Diff line number Diff line Loading @@ -7136,10 +7136,6 @@ void set_curr_task(int cpu, struct task_struct *p) #endif #ifdef CONFIG_RT_GROUP_SCHED #else /* !CONFIG_RT_GROUP_SCHED */ #endif /* CONFIG_RT_GROUP_SCHED */ #ifdef CONFIG_CGROUP_SCHED /* task_group_lock serializes the addition/removal of task groups */ static DEFINE_SPINLOCK(task_group_lock); Loading Loading @@ -7248,9 +7244,6 @@ void sched_move_task(struct task_struct *tsk) } #endif /* CONFIG_CGROUP_SCHED */ #ifdef CONFIG_FAIR_GROUP_SCHED #endif #if defined(CONFIG_RT_GROUP_SCHED) || defined(CONFIG_CFS_BANDWIDTH) static unsigned long to_ratio(u64 period, u64 runtime) { Loading