diff options
| author | Yury Norov (NVIDIA) <yury.norov@gmail.com> | 2025-12-06 22:42:47 -0500 |
|---|---|---|
| committer | Peter Zijlstra <peterz@infradead.org> | 2026-01-08 12:43:56 +0100 |
| commit | 55b39b0cf183b9c682717a55a2fba06da69bba6b (patch) | |
| tree | 12f2feb84eb10d7ef3d13850f7e4e1d70dd64cb7 /kernel/sched | |
| parent | 0ab25ea2a3b3a973fb914d0e47dc9c3c26049e8b (diff) | |
sched/fair: Use cpumask_weight_and() in sched_balance_find_dst_group()
In the group_has_spare case, the function creates a temporary cpumask
to just calculate weight of (p->cpus_ptr & sched_group_span(local)).
We've got a dedicated helper for it.
Signed-off-by: Yury Norov (NVIDIA) <yury.norov@gmail.com>
Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Reviewed-by: Vincent Guittot <vincent.guittot@linaro.org>
Reviewed-by: K Prateek Nayak <kprateek.nayak@amd.com>
Reviewed-by: Fernand Sieber <sieberf@amazon.com>
Link: https://patch.msgid.link/20251207034247.402926-1-yury.norov@gmail.com
Diffstat (limited to 'kernel/sched')
| -rw-r--r-- | kernel/sched/fair.c | 7 |
1 files changed, 3 insertions, 4 deletions
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c index 842a0f20414f..ebee20f75fa0 100644 --- a/kernel/sched/fair.c +++ b/kernel/sched/fair.c @@ -10974,10 +10974,9 @@ sched_balance_find_dst_group(struct sched_domain *sd, struct task_struct *p, int * take care of it. */ if (p->nr_cpus_allowed != NR_CPUS) { - struct cpumask *cpus = this_cpu_cpumask_var_ptr(select_rq_mask); - - cpumask_and(cpus, sched_group_span(local), p->cpus_ptr); - imb_numa_nr = min(cpumask_weight(cpus), sd->imb_numa_nr); + unsigned int w = cpumask_weight_and(p->cpus_ptr, + sched_group_span(local)); + imb_numa_nr = min(w, sd->imb_numa_nr); } imbalance = abs(local_sgs.idle_cpus - idlest_sgs.idle_cpus); |
