From eb9686932be35a47b6d621cc033a861b51e42741 Mon Sep 17 00:00:00 2001 From: Rick Yiu Date: Mon, 28 Aug 2023 07:08:54 +0000 Subject: [PATCH] ANDROID: sched: Export symbols needed for vendor hooks Bug: 297343949 Change-Id: I0cb65e85b36687bfaae6a185ca373d7fb8de0a77 Signed-off-by: Rick Yiu --- kernel/sched/core.c | 1 + kernel/sched/fair.c | 1 + kernel/sched/pelt.c | 6 ++++-- 3 files changed, 6 insertions(+), 2 deletions(-) diff --git a/kernel/sched/core.c b/kernel/sched/core.c index f534940a660a..8c1911fc5664 100644 --- a/kernel/sched/core.c +++ b/kernel/sched/core.c @@ -2588,6 +2588,7 @@ int push_cpu_stop(void *arg) put_task_struct(p); return 0; } +EXPORT_SYMBOL_GPL(push_cpu_stop); /* * sched_class::set_cpus_allowed must do the below, but is not required to diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c index 06e8f15819dd..d944a6f813c6 100644 --- a/kernel/sched/fair.c +++ b/kernel/sched/fair.c @@ -3350,6 +3350,7 @@ void reweight_task(struct task_struct *p, int prio) reweight_entity(cfs_rq, se, weight); load->inv_weight = sched_prio_to_wmult[prio]; } +EXPORT_SYMBOL_GPL(reweight_task); static inline int throttled_hierarchy(struct cfs_rq *cfs_rq); diff --git a/kernel/sched/pelt.c b/kernel/sched/pelt.c index a005458373be..c1c84f5a3c03 100644 --- a/kernel/sched/pelt.c +++ b/kernel/sched/pelt.c @@ -176,7 +176,7 @@ accumulate_sum(u64 delta, struct sched_avg *sa, * load_avg = u_0` + y*(u_0 + u_1*y + u_2*y^2 + ... ) * = u_0 + u_1*y + u_2*y^2 + ... [re-labeling u_i --> u_{i+1}] */ -static __always_inline int +int ___update_load_sum(u64 now, struct sched_avg *sa, unsigned long load, unsigned long runnable, int running) { @@ -228,6 +228,7 @@ ___update_load_sum(u64 now, struct sched_avg *sa, return 1; } +EXPORT_SYMBOL_GPL(___update_load_sum); /* * When syncing *_avg with *_sum, we must take into account the current @@ -253,7 +254,7 @@ ___update_load_sum(u64 now, struct sched_avg *sa, * the period_contrib of cfs_rq when updating the sched_avg of a sched_entity * if it's more convenient. */ -static __always_inline void +void ___update_load_avg(struct sched_avg *sa, unsigned long load) { u32 divider = get_pelt_divider(sa); @@ -265,6 +266,7 @@ ___update_load_avg(struct sched_avg *sa, unsigned long load) sa->runnable_avg = div_u64(sa->runnable_sum, divider); WRITE_ONCE(sa->util_avg, sa->util_sum / divider); } +EXPORT_SYMBOL_GPL(___update_load_avg); /* * sched_entity: