Home
last modified time | relevance | path

Searched refs:cpu_of (Results 1 – 8 of 8) sorted by relevance

/openbmc/linux/kernel/sched/
H A Dpelt.h118 delta = cap_scale(delta, arch_scale_cpu_capacity(cpu_of(rq))); in update_rq_clock_pelt()
119 delta = cap_scale(delta, arch_scale_freq_capacity(cpu_of(rq))); in update_rq_clock_pelt()
H A Dpelt.c439 running = cap_scale(running, arch_scale_freq_capacity(cpu_of(rq))); in update_irq_load_avg()
440 running = cap_scale(running, arch_scale_cpu_capacity(cpu_of(rq))); in update_irq_load_avg()
H A Dcore_sched.c242 const struct cpumask *smt_mask = cpu_smt_mask(cpu_of(rq)); in __sched_core_account_forceidle()
H A Dsched.h1190 static inline int cpu_of(struct rq *rq) in cpu_of() function
1282 for_each_cpu(cpu, cpu_smt_mask(cpu_of(rq))) { in sched_core_cookie_match()
2471 int cpu = cpu_of(rq); in sched_update_tick_dependency()
2554 if (!cpu_active(cpu_of(rq))) in hrtick_enabled()
2966 cpu_of(rq))); in cpufreq_update_util()
3124 rq_util = cpu_util_cfs(cpu_of(rq)) + cpu_util_rt(rq); in uclamp_rq_is_capped()
3398 struct mm_cid *pcpu_cid = per_cpu_ptr(mm->pcpu_cid, cpu_of(rq)); in mm_cid_snapshot_time()
H A Dfair.c334 int cpu = cpu_of(rq); in list_add_leaf_cfs_rq()
5089 if (task_fits_cpu(p, cpu_of(rq))) { in update_misfit_status()
5758 se = cfs_rq->tg->se[cpu_of(rq_of(cfs_rq))]; in throttle_cfs_rq()
5826 se = cfs_rq->tg->se[cpu_of(rq)]; in unthrottle_cfs_rq()
6027 if (cpu_of(rq) != this_cpu || in distribute_cfs_runtime()
6514 int cpu = cpu_of(rq); in sched_fair_update_stop_tick()
7254 int core = cpu_of(rq); in __update_idle_core()
9282 int cpu = cpu_of(rq); in __update_blocked_fair()
11381 __cpumask_clear_cpu(cpu_of(busiest), cpus); in load_balance()
11442 stop_one_cpu_nowait(cpu_of(busiest), in load_balance()
[all …]
H A Dcore.c729 steal = paravirt_steal_clock(cpu_of(rq)); in update_rq_clock_task()
764 delta = sched_clock_cpu(cpu_of(rq)) - rq->clock; in update_rq_clock()
791 WARN_ON_ONCE(cpu_of(rq) != smp_processor_id()); in hrtick()
1050 cpu = cpu_of(rq); in resched_curr()
1174 int cpu = cpu_of(rq); in nohz_csd_func()
3062 stop_one_cpu_nowait(cpu_of(rq), migration_cpu_stop, in affine_move_task()
3887 if (WARN_ON_ONCE(task_cpu(p) != cpu_of(rq))) in sched_ttwu_pending()
3888 set_task_cpu(p, cpu_of(rq)); in sched_ttwu_pending()
6082 cpu = cpu_of(rq); in pick_next_task()
6370 int cpu = cpu_of(rq); in sched_core_balance()
[all …]
H A Drt.c566 (rt_rq = iter->rt_rq[cpu_of(rq)]);)
585 int cpu = cpu_of(rq); in sched_rt_rq_enqueue()
603 int cpu = cpu_of(rq_of_rt_rq(rt_rq)); in sched_rt_rq_dequeue()
2565 if (p->prio < rq->curr->prio && cpu_online(cpu_of(rq))) in switched_to_rt()
H A Ddeadline.c1303 int cpu = cpu_of(rq); in update_curr_dl()
2500 src_dl_b = dl_bw_of(cpu_of(rq)); in set_cpus_allowed_dl()