Searched refs:lbr_nr (Results 1 – 9 of 9) sorted by relevance
/openbmc/linux/arch/x86/events/intel/ |
H A D | lbr.c | 167 for (i = 0; i < x86_pmu.lbr_nr; i++) in intel_pmu_lbr_reset_32() 175 for (i = 0; i < x86_pmu.lbr_nr; i++) { in intel_pmu_lbr_reset_64() 186 wrmsrl(MSR_ARCH_LBR_DEPTH, x86_pmu.lbr_nr); in intel_pmu_arch_lbr_reset() 193 if (!x86_pmu.lbr_nr) in intel_pmu_lbr_reset() 368 mask = x86_pmu.lbr_nr - 1; in intel_pmu_lbr_restore() 374 for (; i < x86_pmu.lbr_nr; i++) { in intel_pmu_lbr_restore() 395 if (!entries[x86_pmu.lbr_nr - 1].from) in intel_pmu_arch_lbr_restore() 398 for (i = 0; i < x86_pmu.lbr_nr; i++) { in intel_pmu_arch_lbr_restore() 460 mask = x86_pmu.lbr_nr - 1; in intel_pmu_lbr_save() 462 for (i = 0; i < x86_pmu.lbr_nr; i++) { in intel_pmu_lbr_save() [all …]
|
H A D | core.c | 2266 cnt = min_t(unsigned int, cnt, x86_pmu.lbr_nr); in __intel_pmu_snapshot_branch_stack() 2911 if (x86_pmu.lbr_nr) { in intel_pmu_reset() 4724 if (!static_cpu_has(X86_FEATURE_ARCH_LBR) && x86_pmu.lbr_nr) in intel_pmu_cpu_starting() 5149 unsigned long lbr_from_nr = x86_pmu.lbr_from + x86_pmu.lbr_nr; in is_lbr_from() 5490 return snprintf(buf, PAGE_SIZE, "%d\n", x86_pmu.lbr_nr); in branches_show() 5550 return x86_pmu.lbr_nr ? attr->mode : 0; in lbr_is_visible() 6875 x86_pmu.lbr_nr = 0; in intel_pmu_init() 6876 for (i = 0; i < x86_pmu.lbr_nr; i++) { in intel_pmu_init() 6879 x86_pmu.lbr_nr = 0; in intel_pmu_init() 6882 if (x86_pmu.lbr_nr) { in intel_pmu_init() [all …]
|
H A D | ds.c | 1175 sz += x86_pmu.lbr_nr * sizeof(struct lbr_entry); in adaptive_pebs_record_size_update() 1226 ((x86_pmu.lbr_nr-1) << PEBS_DATACFG_LBR_SHIFT); in pebs_update_adaptive_cfg()
|
/openbmc/linux/arch/x86/events/amd/ |
H A D | lbr.c | 170 for (i = 0; i < x86_pmu.lbr_nr; i++) { in amd_pmu_lbr_read() 253 if (!x86_pmu.lbr_nr) in amd_pmu_lbr_setup_filter() 329 if (!x86_pmu.lbr_nr) in amd_pmu_lbr_reset() 333 for (i = 0; i < x86_pmu.lbr_nr; i++) { in amd_pmu_lbr_reset() 348 if (!x86_pmu.lbr_nr) in amd_pmu_lbr_add() 367 if (!x86_pmu.lbr_nr) in amd_pmu_lbr_del() 396 if (!cpuc->lbr_users || !x86_pmu.lbr_nr) in amd_pmu_lbr_enable_all() 419 if (!cpuc->lbr_users || !x86_pmu.lbr_nr) in amd_pmu_lbr_disable_all() 440 x86_pmu.lbr_nr = ebx.split.lbr_v2_stack_sz; in amd_pmu_lbr_init() 442 pr_cont("%d-deep LBR, ", x86_pmu.lbr_nr); in amd_pmu_lbr_init()
|
H A D | brs.c | 62 x86_pmu.lbr_nr = 16; in amd_brs_detect() 86 if (!x86_pmu.lbr_nr) in amd_brs_setup_filter() 148 if (event->attr.sample_period <= x86_pmu.lbr_nr) in amd_brs_hw_config() 170 return (cfg->msroff ? cfg->msroff : x86_pmu.lbr_nr) - 1; in amd_brs_get_tos() 198 pr_cont("%d-deep BRS, ", x86_pmu.lbr_nr); in amd_brs_init() 302 if (WARN_ON_ONCE(cfg.msroff >= x86_pmu.lbr_nr)) in amd_brs_drain()
|
H A D | core.c | 393 if (has_branch_stack(event) && !x86_pmu.lbr_nr) in amd_pmu_hw_config() 542 if (x86_pmu.lbr_nr) in amd_pmu_cpu_reset() 923 if (x86_pmu.lbr_nr) { in amd_pmu_v2_handle_irq() 1258 if (has_branch_stack(event) && *left > x86_pmu.lbr_nr) in amd_pmu_limit_period() 1259 *left -= x86_pmu.lbr_nr; in amd_pmu_limit_period() 1301 return snprintf(buf, PAGE_SIZE, "%d\n", x86_pmu.lbr_nr); in branches_show() 1314 return x86_pmu.lbr_nr ? attr->mode : 0; in amd_branches_is_visible() 1336 return static_cpu_has(X86_FEATURE_BRS) && x86_pmu.lbr_nr ? in amd_brs_is_visible()
|
/openbmc/linux/tools/perf/util/ |
H A D | machine.c | 2557 int lbr_nr = lbr_stack->nr; in lbr_callchain_add_lbr_ip() local 2604 for (i = 0; i < lbr_nr; i++) { in lbr_callchain_add_lbr_ip() 2619 for (i = lbr_nr - 1; i >= 0; i--) { in lbr_callchain_add_lbr_ip() 2631 if (lbr_nr > 0) { in lbr_callchain_add_lbr_ip()
|
/openbmc/linux/arch/x86/events/ |
H A D | perf_event.h | 855 lbr_info, lbr_nr; /* LBR base regs and size */ member
|
H A D | core.c | 554 if (x86_pmu.lbr_nr || x86_pmu.intel_cap.pebs_format >= 2) in x86_pmu_max_precise() 1553 if (x86_pmu.lbr_nr) { in perf_event_print_debug()
|