Searched refs:SCHED_CAPACITY_SCALE (Results 1 – 14 of 14) sorted by relevance
/openbmc/linux/arch/x86/kernel/cpu/ |
H A D | aperfmperf.c | 87 static u64 arch_turbo_freq_ratio = SCHED_CAPACITY_SCALE; 88 static u64 arch_max_freq_ratio = SCHED_CAPACITY_SCALE; 92 arch_max_freq_ratio = turbo_disabled ? SCHED_CAPACITY_SCALE : in arch_set_max_freq_ratio() 279 turbo_ratio = div_u64(turbo_freq * SCHED_CAPACITY_SCALE, base_freq); in intel_set_max_freq_ratio() 342 per_cpu(arch_freq_scale, cpu) = SCHED_CAPACITY_SCALE; in disable_freq_invariance_workfn() 348 DEFINE_PER_CPU(unsigned long, arch_freq_scale) = SCHED_CAPACITY_SCALE; 367 if (freq_scale > SCHED_CAPACITY_SCALE) in scale_freq_tick() 368 freq_scale = SCHED_CAPACITY_SCALE; in scale_freq_tick()
|
/openbmc/linux/arch/x86/kernel/acpi/ |
H A D | cppc.c | 90 perf_ratio = div_u64(highest_perf * SCHED_CAPACITY_SCALE, nominal_perf); in amd_set_max_freq_ratio() 92 perf_ratio = (perf_ratio + SCHED_CAPACITY_SCALE) >> 1; in amd_set_max_freq_ratio()
|
/openbmc/linux/include/linux/sched/ |
H A D | topology.h | 259 return SCHED_CAPACITY_SCALE; in arch_scale_cpu_capacity()
|
/openbmc/linux/kernel/sched/ |
H A D | topology.c | 103 if (group->sgc->capacity != SCHED_CAPACITY_SCALE) in sched_domain_debug_one() 987 sg->sgc->capacity = SCHED_CAPACITY_SCALE * cpumask_weight(sg_span); in init_overlap_sched_group() 988 sg->sgc->min_capacity = SCHED_CAPACITY_SCALE; in init_overlap_sched_group() 989 sg->sgc->max_capacity = SCHED_CAPACITY_SCALE; in init_overlap_sched_group() 1214 sg->sgc->capacity = SCHED_CAPACITY_SCALE * cpumask_weight(sched_group_span(sg)); in get_group() 1215 sg->sgc->min_capacity = SCHED_CAPACITY_SCALE; in get_group() 1216 sg->sgc->max_capacity = SCHED_CAPACITY_SCALE; in get_group()
|
H A D | sched.h | 915 unsigned long value : bits_per(SCHED_CAPACITY_SCALE); 916 unsigned long tasks : BITS_PER_LONG - bits_per(SCHED_CAPACITY_SCALE); 2620 return SCHED_CAPACITY_SCALE; in arch_scale_freq_capacity() 3029 return (rq->dl.running_bw * SCHED_CAPACITY_SCALE) >> BW_SHIFT; in cpu_bw_dl() 3132 return max_util != SCHED_CAPACITY_SCALE && rq_util >= max_util; in uclamp_rq_is_capped() 3154 return SCHED_CAPACITY_SCALE; in uclamp_eff_value() 3177 return SCHED_CAPACITY_SCALE; in uclamp_rq_get()
|
H A D | cpufreq_schedutil.c | 9 #define IOWAIT_BOOST_MIN (SCHED_CAPACITY_SCALE / 8) 228 min_t(unsigned int, sg_cpu->iowait_boost << 1, SCHED_CAPACITY_SCALE); in sugov_iowait_boost()
|
H A D | core.c | 1344 static unsigned int __maybe_unused sysctl_sched_uclamp_util_min = SCHED_CAPACITY_SCALE; 1347 static unsigned int __maybe_unused sysctl_sched_uclamp_util_max = SCHED_CAPACITY_SCALE; 1364 static unsigned int sysctl_sched_uclamp_util_min_rt_default = SCHED_CAPACITY_SCALE; 1390 #define UCLAMP_BUCKET_DELTA DIV_ROUND_CLOSEST(SCHED_CAPACITY_SCALE, UCLAMP_BUCKETS) 1404 return SCHED_CAPACITY_SCALE; in uclamp_none() 1842 sysctl_sched_uclamp_util_max > SCHED_CAPACITY_SCALE || in sysctl_sched_uclamp_handler() 1843 sysctl_sched_uclamp_util_min_rt_default > SCHED_CAPACITY_SCALE) { in sysctl_sched_uclamp_handler() 1895 if (util_min + 1 > SCHED_CAPACITY_SCALE + 1) in uclamp_validate() 1902 if (util_max + 1 > SCHED_CAPACITY_SCALE + 1) in uclamp_validate() 10051 rq->cpu_capacity = rq->cpu_capacity_orig = SCHED_CAPACITY_SCALE; in sched_init() [all …]
|
H A D | fair.c | 4883 #define UTIL_EST_MARGIN (SCHED_CAPACITY_SCALE / 100) 5061 uclamp_max_fits = (capacity_orig == SCHED_CAPACITY_SCALE) && (uclamp_max == SCHED_CAPACITY_SCALE); in util_fits_cpu() 9978 sgs->avg_load = (sgs->group_load * SCHED_CAPACITY_SCALE) / in update_sg_lb_stats() 10275 sgs->avg_load = (sgs->group_load * SCHED_CAPACITY_SCALE) / in update_sg_wakeup_stats() 10563 do_div(tmp, 10000 * SCHED_CAPACITY_SCALE); in update_idle_cpu_scan() 10564 tmp = min_t(long, tmp, SCHED_CAPACITY_SCALE); in update_idle_cpu_scan() 10565 y = SCHED_CAPACITY_SCALE - tmp; in update_idle_cpu_scan() 10569 do_div(y, SCHED_CAPACITY_SCALE); in update_idle_cpu_scan() 10783 local->avg_load = (local->group_load * SCHED_CAPACITY_SCALE) / in calculate_imbalance() 10795 sds->avg_load = (sds->total_load * SCHED_CAPACITY_SCALE) / in calculate_imbalance() [all …]
|
H A D | deadline.c | 147 capacity_orig_of(i) == SCHED_CAPACITY_SCALE) { in dl_bw_capacity() 195 return SCHED_CAPACITY_SCALE; in dl_bw_capacity()
|
/openbmc/linux/arch/arm64/kernel/ |
H A D | topology.c | 180 scale = min_t(unsigned long, scale, SCHED_CAPACITY_SCALE); in amu_scale_freq_tick()
|
/openbmc/linux/drivers/base/ |
H A D | arch_topology.c | 127 DEFINE_PER_CPU(unsigned long, arch_freq_scale) = SCHED_CAPACITY_SCALE; 153 DEFINE_PER_CPU(unsigned long, cpu_scale) = SCHED_CAPACITY_SCALE;
|
/openbmc/linux/include/linux/ |
H A D | sched.h | 409 # define SCHED_CAPACITY_SCALE (1L << SCHED_CAPACITY_SHIFT) macro 711 unsigned int value : bits_per(SCHED_CAPACITY_SCALE);
|
/openbmc/linux/drivers/cpufreq/ |
H A D | cppc_cpufreq.c | 388 #define CPPC_EM_COST_GAP (4 * SCHED_CAPACITY_SCALE * CPPC_EM_COST_STEP \
|
/openbmc/linux/init/ |
H A D | Kconfig | 837 will be SCHED_CAPACITY_SCALE/UCLAMP_BUCKETS_COUNT. The higher the
|