mirror of https://github.com/torvalds/linux.git
sched: Clean up and standardize #if/#else/#endif markers in sched/core.c
- Use the standard #ifdef marker format for larger blocks, where appropriate: #if CONFIG_FOO ... #else /* !CONFIG_FOO: */ ... #endif /* !CONFIG_FOO */ - Apply this simplification: -#if defined(CONFIG_FOO) +#ifdef CONFIG_FOO - Fix whitespace noise. - Use vertical alignment to better visualize nested #ifdef blocks, where appropriate: #ifdef CONFIG_FOO # ifdef CONFIG_BAR ... # endif #endif Signed-off-by: Ingo Molnar <mingo@kernel.org> Acked-by: Peter Zijlstra <peterz@infradead.org> Cc: Dietmar Eggemann <dietmar.eggemann@arm.com> Cc: Juri Lelli <juri.lelli@redhat.com> Cc: Linus Torvalds <torvalds@linux-foundation.org> Cc: Mel Gorman <mgorman@suse.de> Cc: Sebastian Andrzej Siewior <bigeasy@linutronix.de> Cc: Shrikanth Hegde <sshegde@linux.ibm.com> Cc: Steven Rostedt <rostedt@goodmis.org> Cc: Valentin Schneider <vschneid@redhat.com> Cc: Vincent Guittot <vincent.guittot@linaro.org> Link: https://lore.kernel.org/r/20250528080924.2273858-4-mingo@kernel.org
This commit is contained in:
parent
bbb1b274e8
commit
b7ebb75856
|
|
@ -481,13 +481,13 @@ void sched_core_put(void)
|
||||||
schedule_work(&_work);
|
schedule_work(&_work);
|
||||||
}
|
}
|
||||||
|
|
||||||
#else /* !CONFIG_SCHED_CORE */
|
#else /* !CONFIG_SCHED_CORE: */
|
||||||
|
|
||||||
static inline void sched_core_enqueue(struct rq *rq, struct task_struct *p) { }
|
static inline void sched_core_enqueue(struct rq *rq, struct task_struct *p) { }
|
||||||
static inline void
|
static inline void
|
||||||
sched_core_dequeue(struct rq *rq, struct task_struct *p, int flags) { }
|
sched_core_dequeue(struct rq *rq, struct task_struct *p, int flags) { }
|
||||||
|
|
||||||
#endif /* CONFIG_SCHED_CORE */
|
#endif /* !CONFIG_SCHED_CORE */
|
||||||
|
|
||||||
/* need a wrapper since we may need to trace from modules */
|
/* need a wrapper since we may need to trace from modules */
|
||||||
EXPORT_TRACEPOINT_SYMBOL(sched_set_state_tp);
|
EXPORT_TRACEPOINT_SYMBOL(sched_set_state_tp);
|
||||||
|
|
@ -667,7 +667,7 @@ void double_rq_lock(struct rq *rq1, struct rq *rq2)
|
||||||
|
|
||||||
double_rq_clock_clear_update(rq1, rq2);
|
double_rq_clock_clear_update(rq1, rq2);
|
||||||
}
|
}
|
||||||
#endif
|
#endif /* CONFIG_SMP */
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* __task_rq_lock - lock the rq @p resides on.
|
* __task_rq_lock - lock the rq @p resides on.
|
||||||
|
|
@ -899,7 +899,7 @@ void hrtick_start(struct rq *rq, u64 delay)
|
||||||
smp_call_function_single_async(cpu_of(rq), &rq->hrtick_csd);
|
smp_call_function_single_async(cpu_of(rq), &rq->hrtick_csd);
|
||||||
}
|
}
|
||||||
|
|
||||||
#else
|
#else /* !CONFIG_SMP: */
|
||||||
/*
|
/*
|
||||||
* Called to set the hrtick timer state.
|
* Called to set the hrtick timer state.
|
||||||
*
|
*
|
||||||
|
|
@ -916,7 +916,7 @@ void hrtick_start(struct rq *rq, u64 delay)
|
||||||
HRTIMER_MODE_REL_PINNED_HARD);
|
HRTIMER_MODE_REL_PINNED_HARD);
|
||||||
}
|
}
|
||||||
|
|
||||||
#endif /* CONFIG_SMP */
|
#endif /* !CONFIG_SMP */
|
||||||
|
|
||||||
static void hrtick_rq_init(struct rq *rq)
|
static void hrtick_rq_init(struct rq *rq)
|
||||||
{
|
{
|
||||||
|
|
@ -925,7 +925,7 @@ static void hrtick_rq_init(struct rq *rq)
|
||||||
#endif
|
#endif
|
||||||
hrtimer_setup(&rq->hrtick_timer, hrtick, CLOCK_MONOTONIC, HRTIMER_MODE_REL_HARD);
|
hrtimer_setup(&rq->hrtick_timer, hrtick, CLOCK_MONOTONIC, HRTIMER_MODE_REL_HARD);
|
||||||
}
|
}
|
||||||
#else /* CONFIG_SCHED_HRTICK */
|
#else /* !CONFIG_SCHED_HRTICK: */
|
||||||
static inline void hrtick_clear(struct rq *rq)
|
static inline void hrtick_clear(struct rq *rq)
|
||||||
{
|
{
|
||||||
}
|
}
|
||||||
|
|
@ -933,7 +933,7 @@ static inline void hrtick_clear(struct rq *rq)
|
||||||
static inline void hrtick_rq_init(struct rq *rq)
|
static inline void hrtick_rq_init(struct rq *rq)
|
||||||
{
|
{
|
||||||
}
|
}
|
||||||
#endif /* CONFIG_SCHED_HRTICK */
|
#endif /* !CONFIG_SCHED_HRTICK */
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* try_cmpxchg based fetch_or() macro so it works for different integer types:
|
* try_cmpxchg based fetch_or() macro so it works for different integer types:
|
||||||
|
|
@ -1971,7 +1971,7 @@ static int sysctl_sched_uclamp_handler(const struct ctl_table *table, int write,
|
||||||
sysctl_sched_uclamp_util_min_rt_default = old_min_rt;
|
sysctl_sched_uclamp_util_min_rt_default = old_min_rt;
|
||||||
return result;
|
return result;
|
||||||
}
|
}
|
||||||
#endif
|
#endif /* CONFIG_SYSCTL */
|
||||||
|
|
||||||
static void uclamp_fork(struct task_struct *p)
|
static void uclamp_fork(struct task_struct *p)
|
||||||
{
|
{
|
||||||
|
|
@ -2037,13 +2037,13 @@ static void __init init_uclamp(void)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#else /* !CONFIG_UCLAMP_TASK */
|
#else /* !CONFIG_UCLAMP_TASK: */
|
||||||
static inline void uclamp_rq_inc(struct rq *rq, struct task_struct *p, int flags) { }
|
static inline void uclamp_rq_inc(struct rq *rq, struct task_struct *p, int flags) { }
|
||||||
static inline void uclamp_rq_dec(struct rq *rq, struct task_struct *p) { }
|
static inline void uclamp_rq_dec(struct rq *rq, struct task_struct *p) { }
|
||||||
static inline void uclamp_fork(struct task_struct *p) { }
|
static inline void uclamp_fork(struct task_struct *p) { }
|
||||||
static inline void uclamp_post_fork(struct task_struct *p) { }
|
static inline void uclamp_post_fork(struct task_struct *p) { }
|
||||||
static inline void init_uclamp(void) { }
|
static inline void init_uclamp(void) { }
|
||||||
#endif /* CONFIG_UCLAMP_TASK */
|
#endif /* !CONFIG_UCLAMP_TASK */
|
||||||
|
|
||||||
bool sched_task_on_rq(struct task_struct *p)
|
bool sched_task_on_rq(struct task_struct *p)
|
||||||
{
|
{
|
||||||
|
|
@ -3661,7 +3661,7 @@ void sched_set_stop_task(int cpu, struct task_struct *stop)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#else /* CONFIG_SMP */
|
#else /* !CONFIG_SMP: */
|
||||||
|
|
||||||
static inline void migrate_disable_switch(struct rq *rq, struct task_struct *p) { }
|
static inline void migrate_disable_switch(struct rq *rq, struct task_struct *p) { }
|
||||||
|
|
||||||
|
|
@ -3770,7 +3770,7 @@ ttwu_do_activate(struct rq *rq, struct task_struct *p, int wake_flags,
|
||||||
|
|
||||||
rq->idle_stamp = 0;
|
rq->idle_stamp = 0;
|
||||||
}
|
}
|
||||||
#endif
|
#endif /* CONFIG_SMP */
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
|
@ -3992,14 +3992,14 @@ static bool ttwu_queue_wakelist(struct task_struct *p, int cpu, int wake_flags)
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
#else /* !CONFIG_SMP */
|
#else /* !CONFIG_SMP: */
|
||||||
|
|
||||||
static inline bool ttwu_queue_wakelist(struct task_struct *p, int cpu, int wake_flags)
|
static inline bool ttwu_queue_wakelist(struct task_struct *p, int cpu, int wake_flags)
|
||||||
{
|
{
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
|
|
||||||
#endif /* CONFIG_SMP */
|
#endif /* !CONFIG_SMP */
|
||||||
|
|
||||||
static void ttwu_queue(struct task_struct *p, int cpu, int wake_flags)
|
static void ttwu_queue(struct task_struct *p, int cpu, int wake_flags)
|
||||||
{
|
{
|
||||||
|
|
@ -4335,9 +4335,9 @@ int try_to_wake_up(struct task_struct *p, unsigned int state, int wake_flags)
|
||||||
psi_ttwu_dequeue(p);
|
psi_ttwu_dequeue(p);
|
||||||
set_task_cpu(p, cpu);
|
set_task_cpu(p, cpu);
|
||||||
}
|
}
|
||||||
#else
|
#else /* !CONFIG_SMP: */
|
||||||
cpu = task_cpu(p);
|
cpu = task_cpu(p);
|
||||||
#endif /* CONFIG_SMP */
|
#endif /* !CONFIG_SMP */
|
||||||
|
|
||||||
ttwu_queue(p, cpu, wake_flags);
|
ttwu_queue(p, cpu, wake_flags);
|
||||||
}
|
}
|
||||||
|
|
@ -4599,8 +4599,8 @@ static int sysctl_numa_balancing(const struct ctl_table *table, int write,
|
||||||
}
|
}
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
#endif
|
#endif /* CONFIG_PROC_SYSCTL */
|
||||||
#endif
|
#endif /* CONFIG_NUMA_BALANCING */
|
||||||
|
|
||||||
#ifdef CONFIG_SCHEDSTATS
|
#ifdef CONFIG_SCHEDSTATS
|
||||||
|
|
||||||
|
|
@ -4787,7 +4787,7 @@ int sched_fork(unsigned long clone_flags, struct task_struct *p)
|
||||||
if (likely(sched_info_on()))
|
if (likely(sched_info_on()))
|
||||||
memset(&p->sched_info, 0, sizeof(p->sched_info));
|
memset(&p->sched_info, 0, sizeof(p->sched_info));
|
||||||
#endif
|
#endif
|
||||||
#if defined(CONFIG_SMP)
|
#ifdef CONFIG_SMP
|
||||||
p->on_cpu = 0;
|
p->on_cpu = 0;
|
||||||
#endif
|
#endif
|
||||||
init_task_preempt_count(p);
|
init_task_preempt_count(p);
|
||||||
|
|
@ -4978,7 +4978,7 @@ fire_sched_out_preempt_notifiers(struct task_struct *curr,
|
||||||
__fire_sched_out_preempt_notifiers(curr, next);
|
__fire_sched_out_preempt_notifiers(curr, next);
|
||||||
}
|
}
|
||||||
|
|
||||||
#else /* !CONFIG_PREEMPT_NOTIFIERS */
|
#else /* !CONFIG_PREEMPT_NOTIFIERS: */
|
||||||
|
|
||||||
static inline void fire_sched_in_preempt_notifiers(struct task_struct *curr)
|
static inline void fire_sched_in_preempt_notifiers(struct task_struct *curr)
|
||||||
{
|
{
|
||||||
|
|
@ -4990,7 +4990,7 @@ fire_sched_out_preempt_notifiers(struct task_struct *curr,
|
||||||
{
|
{
|
||||||
}
|
}
|
||||||
|
|
||||||
#endif /* CONFIG_PREEMPT_NOTIFIERS */
|
#endif /* !CONFIG_PREEMPT_NOTIFIERS */
|
||||||
|
|
||||||
static inline void prepare_task(struct task_struct *next)
|
static inline void prepare_task(struct task_struct *next)
|
||||||
{
|
{
|
||||||
|
|
@ -5107,13 +5107,13 @@ void balance_callbacks(struct rq *rq, struct balance_callback *head)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#else
|
#else /* !CONFIG_SMP: */
|
||||||
|
|
||||||
static inline void __balance_callbacks(struct rq *rq)
|
static inline void __balance_callbacks(struct rq *rq)
|
||||||
{
|
{
|
||||||
}
|
}
|
||||||
|
|
||||||
#endif
|
#endif /* !CONFIG_SMP */
|
||||||
|
|
||||||
static inline void
|
static inline void
|
||||||
prepare_lock_switch(struct rq *rq, struct task_struct *next, struct rq_flags *rf)
|
prepare_lock_switch(struct rq *rq, struct task_struct *next, struct rq_flags *rf)
|
||||||
|
|
@ -5527,7 +5527,7 @@ void sched_exec(void)
|
||||||
stop_one_cpu(task_cpu(p), migration_cpu_stop, &arg);
|
stop_one_cpu(task_cpu(p), migration_cpu_stop, &arg);
|
||||||
}
|
}
|
||||||
|
|
||||||
#endif
|
#endif /* CONFIG_SMP */
|
||||||
|
|
||||||
DEFINE_PER_CPU(struct kernel_stat, kstat);
|
DEFINE_PER_CPU(struct kernel_stat, kstat);
|
||||||
DEFINE_PER_CPU(struct kernel_cpustat, kernel_cpustat);
|
DEFINE_PER_CPU(struct kernel_cpustat, kernel_cpustat);
|
||||||
|
|
@ -5835,10 +5835,10 @@ int __init sched_tick_offload_init(void)
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
#else /* !CONFIG_NO_HZ_FULL */
|
#else /* !CONFIG_NO_HZ_FULL: */
|
||||||
static inline void sched_tick_start(int cpu) { }
|
static inline void sched_tick_start(int cpu) { }
|
||||||
static inline void sched_tick_stop(int cpu) { }
|
static inline void sched_tick_stop(int cpu) { }
|
||||||
#endif
|
#endif /* !CONFIG_NO_HZ_FULL */
|
||||||
|
|
||||||
#if defined(CONFIG_PREEMPTION) && (defined(CONFIG_DEBUG_PREEMPT) || \
|
#if defined(CONFIG_PREEMPTION) && (defined(CONFIG_DEBUG_PREEMPT) || \
|
||||||
defined(CONFIG_TRACE_PREEMPT_TOGGLE))
|
defined(CONFIG_TRACE_PREEMPT_TOGGLE))
|
||||||
|
|
@ -6553,7 +6553,7 @@ static inline void sched_core_cpu_dying(unsigned int cpu)
|
||||||
rq->core = rq;
|
rq->core = rq;
|
||||||
}
|
}
|
||||||
|
|
||||||
#else /* !CONFIG_SCHED_CORE */
|
#else /* !CONFIG_SCHED_CORE: */
|
||||||
|
|
||||||
static inline void sched_core_cpu_starting(unsigned int cpu) {}
|
static inline void sched_core_cpu_starting(unsigned int cpu) {}
|
||||||
static inline void sched_core_cpu_deactivate(unsigned int cpu) {}
|
static inline void sched_core_cpu_deactivate(unsigned int cpu) {}
|
||||||
|
|
@ -6565,7 +6565,7 @@ pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf)
|
||||||
return __pick_next_task(rq, prev, rf);
|
return __pick_next_task(rq, prev, rf);
|
||||||
}
|
}
|
||||||
|
|
||||||
#endif /* CONFIG_SCHED_CORE */
|
#endif /* !CONFIG_SCHED_CORE */
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Constants for the sched_mode argument of __schedule().
|
* Constants for the sched_mode argument of __schedule().
|
||||||
|
|
@ -6992,7 +6992,7 @@ NOKPROBE_SYMBOL(preempt_schedule);
|
||||||
EXPORT_SYMBOL(preempt_schedule);
|
EXPORT_SYMBOL(preempt_schedule);
|
||||||
|
|
||||||
#ifdef CONFIG_PREEMPT_DYNAMIC
|
#ifdef CONFIG_PREEMPT_DYNAMIC
|
||||||
#if defined(CONFIG_HAVE_PREEMPT_DYNAMIC_CALL)
|
# ifdef CONFIG_HAVE_PREEMPT_DYNAMIC_CALL
|
||||||
# ifndef preempt_schedule_dynamic_enabled
|
# ifndef preempt_schedule_dynamic_enabled
|
||||||
# define preempt_schedule_dynamic_enabled preempt_schedule
|
# define preempt_schedule_dynamic_enabled preempt_schedule
|
||||||
# define preempt_schedule_dynamic_disabled NULL
|
# define preempt_schedule_dynamic_disabled NULL
|
||||||
|
|
@ -7010,7 +7010,7 @@ void __sched notrace dynamic_preempt_schedule(void)
|
||||||
NOKPROBE_SYMBOL(dynamic_preempt_schedule);
|
NOKPROBE_SYMBOL(dynamic_preempt_schedule);
|
||||||
EXPORT_SYMBOL(dynamic_preempt_schedule);
|
EXPORT_SYMBOL(dynamic_preempt_schedule);
|
||||||
# endif
|
# endif
|
||||||
#endif
|
#endif /* CONFIG_PREEMPT_DYNAMIC */
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* preempt_schedule_notrace - preempt_schedule called by tracing
|
* preempt_schedule_notrace - preempt_schedule called by tracing
|
||||||
|
|
@ -7301,7 +7301,7 @@ void rt_mutex_setprio(struct task_struct *p, struct task_struct *pi_task)
|
||||||
|
|
||||||
preempt_enable();
|
preempt_enable();
|
||||||
}
|
}
|
||||||
#endif
|
#endif /* CONFIG_RT_MUTEXES */
|
||||||
|
|
||||||
#if !defined(CONFIG_PREEMPTION) || defined(CONFIG_PREEMPT_DYNAMIC)
|
#if !defined(CONFIG_PREEMPTION) || defined(CONFIG_PREEMPT_DYNAMIC)
|
||||||
int __sched __cond_resched(void)
|
int __sched __cond_resched(void)
|
||||||
|
|
@ -7332,7 +7332,7 @@ EXPORT_SYMBOL(__cond_resched);
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
#ifdef CONFIG_PREEMPT_DYNAMIC
|
#ifdef CONFIG_PREEMPT_DYNAMIC
|
||||||
#if defined(CONFIG_HAVE_PREEMPT_DYNAMIC_CALL)
|
# ifdef CONFIG_HAVE_PREEMPT_DYNAMIC_CALL
|
||||||
# define cond_resched_dynamic_enabled __cond_resched
|
# define cond_resched_dynamic_enabled __cond_resched
|
||||||
# define cond_resched_dynamic_disabled ((void *)&__static_call_return0)
|
# define cond_resched_dynamic_disabled ((void *)&__static_call_return0)
|
||||||
DEFINE_STATIC_CALL_RET0(cond_resched, __cond_resched);
|
DEFINE_STATIC_CALL_RET0(cond_resched, __cond_resched);
|
||||||
|
|
@ -7361,7 +7361,7 @@ int __sched dynamic_might_resched(void)
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(dynamic_might_resched);
|
EXPORT_SYMBOL(dynamic_might_resched);
|
||||||
# endif
|
# endif
|
||||||
#endif
|
#endif /* CONFIG_PREEMPT_DYNAMIC */
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* __cond_resched_lock() - if a reschedule is pending, drop the given lock,
|
* __cond_resched_lock() - if a reschedule is pending, drop the given lock,
|
||||||
|
|
@ -8123,7 +8123,7 @@ static void balance_hotplug_wait(void)
|
||||||
TASK_UNINTERRUPTIBLE);
|
TASK_UNINTERRUPTIBLE);
|
||||||
}
|
}
|
||||||
|
|
||||||
#else
|
#else /* !CONFIG_HOTPLUG_CPU: */
|
||||||
|
|
||||||
static inline void balance_push(struct rq *rq)
|
static inline void balance_push(struct rq *rq)
|
||||||
{
|
{
|
||||||
|
|
@ -8137,7 +8137,7 @@ static inline void balance_hotplug_wait(void)
|
||||||
{
|
{
|
||||||
}
|
}
|
||||||
|
|
||||||
#endif /* CONFIG_HOTPLUG_CPU */
|
#endif /* !CONFIG_HOTPLUG_CPU */
|
||||||
|
|
||||||
void set_rq_online(struct rq *rq)
|
void set_rq_online(struct rq *rq)
|
||||||
{
|
{
|
||||||
|
|
@ -8446,7 +8446,7 @@ int sched_cpu_dying(unsigned int cpu)
|
||||||
sched_core_cpu_dying(cpu);
|
sched_core_cpu_dying(cpu);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
#endif
|
#endif /* CONFIG_HOTPLUG_CPU */
|
||||||
|
|
||||||
void __init sched_init_smp(void)
|
void __init sched_init_smp(void)
|
||||||
{
|
{
|
||||||
|
|
@ -8480,12 +8480,12 @@ static int __init migration_init(void)
|
||||||
}
|
}
|
||||||
early_initcall(migration_init);
|
early_initcall(migration_init);
|
||||||
|
|
||||||
#else
|
#else /* !CONFIG_SMP: */
|
||||||
void __init sched_init_smp(void)
|
void __init sched_init_smp(void)
|
||||||
{
|
{
|
||||||
sched_init_granularity();
|
sched_init_granularity();
|
||||||
}
|
}
|
||||||
#endif /* CONFIG_SMP */
|
#endif /* !CONFIG_SMP */
|
||||||
|
|
||||||
int in_sched_functions(unsigned long addr)
|
int in_sched_functions(unsigned long addr)
|
||||||
{
|
{
|
||||||
|
|
@ -8861,8 +8861,8 @@ void __cant_migrate(const char *file, int line)
|
||||||
add_taint(TAINT_WARN, LOCKDEP_STILL_OK);
|
add_taint(TAINT_WARN, LOCKDEP_STILL_OK);
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(__cant_migrate);
|
EXPORT_SYMBOL_GPL(__cant_migrate);
|
||||||
#endif
|
# endif /* CONFIG_SMP */
|
||||||
#endif
|
#endif /* CONFIG_DEBUG_ATOMIC_SLEEP */
|
||||||
|
|
||||||
#ifdef CONFIG_MAGIC_SYSRQ
|
#ifdef CONFIG_MAGIC_SYSRQ
|
||||||
void normalize_rt_tasks(void)
|
void normalize_rt_tasks(void)
|
||||||
|
|
@ -8902,7 +8902,7 @@ void normalize_rt_tasks(void)
|
||||||
|
|
||||||
#endif /* CONFIG_MAGIC_SYSRQ */
|
#endif /* CONFIG_MAGIC_SYSRQ */
|
||||||
|
|
||||||
#if defined(CONFIG_KGDB_KDB)
|
#ifdef CONFIG_KGDB_KDB
|
||||||
/*
|
/*
|
||||||
* These functions are only useful for KDB.
|
* These functions are only useful for KDB.
|
||||||
*
|
*
|
||||||
|
|
@ -8926,7 +8926,7 @@ struct task_struct *curr_task(int cpu)
|
||||||
return cpu_curr(cpu);
|
return cpu_curr(cpu);
|
||||||
}
|
}
|
||||||
|
|
||||||
#endif /* defined(CONFIG_KGDB_KDB) */
|
#endif /* CONFIG_KGDB_KDB */
|
||||||
|
|
||||||
#ifdef CONFIG_CGROUP_SCHED
|
#ifdef CONFIG_CGROUP_SCHED
|
||||||
/* task_group_lock serializes the addition/removal of task groups */
|
/* task_group_lock serializes the addition/removal of task groups */
|
||||||
|
|
@ -9807,7 +9807,7 @@ static int cpu_idle_write_s64(struct cgroup_subsys_state *css,
|
||||||
scx_group_set_idle(css_tg(css), idle);
|
scx_group_set_idle(css_tg(css), idle);
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
#endif
|
#endif /* CONFIG_GROUP_SCHED_WEIGHT */
|
||||||
|
|
||||||
static struct cftype cpu_legacy_files[] = {
|
static struct cftype cpu_legacy_files[] = {
|
||||||
#ifdef CONFIG_GROUP_SCHED_WEIGHT
|
#ifdef CONFIG_GROUP_SCHED_WEIGHT
|
||||||
|
|
@ -9935,7 +9935,7 @@ static int cpu_extra_stat_show(struct seq_file *sf,
|
||||||
cfs_b->nr_periods, cfs_b->nr_throttled,
|
cfs_b->nr_periods, cfs_b->nr_throttled,
|
||||||
throttled_usec, cfs_b->nr_burst, burst_usec);
|
throttled_usec, cfs_b->nr_burst, burst_usec);
|
||||||
}
|
}
|
||||||
#endif
|
#endif /* CONFIG_CFS_BANDWIDTH */
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
@ -10076,7 +10076,7 @@ static ssize_t cpu_max_write(struct kernfs_open_file *of,
|
||||||
ret = tg_set_cfs_bandwidth(tg, period, quota, burst);
|
ret = tg_set_cfs_bandwidth(tg, period, quota, burst);
|
||||||
return ret ?: nbytes;
|
return ret ?: nbytes;
|
||||||
}
|
}
|
||||||
#endif
|
#endif /* CONFIG_CFS_BANDWIDTH */
|
||||||
|
|
||||||
static struct cftype cpu_files[] = {
|
static struct cftype cpu_files[] = {
|
||||||
#ifdef CONFIG_GROUP_SCHED_WEIGHT
|
#ifdef CONFIG_GROUP_SCHED_WEIGHT
|
||||||
|
|
@ -10112,7 +10112,7 @@ static struct cftype cpu_files[] = {
|
||||||
.read_u64 = cpu_cfs_burst_read_u64,
|
.read_u64 = cpu_cfs_burst_read_u64,
|
||||||
.write_u64 = cpu_cfs_burst_write_u64,
|
.write_u64 = cpu_cfs_burst_write_u64,
|
||||||
},
|
},
|
||||||
#endif
|
#endif /* CONFIG_CFS_BANDWIDTH */
|
||||||
#ifdef CONFIG_UCLAMP_TASK_GROUP
|
#ifdef CONFIG_UCLAMP_TASK_GROUP
|
||||||
{
|
{
|
||||||
.name = "uclamp.min",
|
.name = "uclamp.min",
|
||||||
|
|
@ -10126,7 +10126,7 @@ static struct cftype cpu_files[] = {
|
||||||
.seq_show = cpu_uclamp_max_show,
|
.seq_show = cpu_uclamp_max_show,
|
||||||
.write = cpu_uclamp_max_write,
|
.write = cpu_uclamp_max_write,
|
||||||
},
|
},
|
||||||
#endif
|
#endif /* CONFIG_UCLAMP_TASK_GROUP */
|
||||||
{ } /* terminate */
|
{ } /* terminate */
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
@ -10733,7 +10733,7 @@ void sched_mm_cid_fork(struct task_struct *t)
|
||||||
WARN_ON_ONCE(!t->mm || t->mm_cid != -1);
|
WARN_ON_ONCE(!t->mm || t->mm_cid != -1);
|
||||||
t->mm_cid_active = 1;
|
t->mm_cid_active = 1;
|
||||||
}
|
}
|
||||||
#endif
|
#endif /* CONFIG_SCHED_MM_CID */
|
||||||
|
|
||||||
#ifdef CONFIG_SCHED_CLASS_EXT
|
#ifdef CONFIG_SCHED_CLASS_EXT
|
||||||
void sched_deq_and_put_task(struct task_struct *p, int queue_flags,
|
void sched_deq_and_put_task(struct task_struct *p, int queue_flags,
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue