/openbmc/linux/drivers/gpu/drm/nouveau/nvkm/subdev/pmu/ |
H A D | base.c | 32 struct nvkm_pmu *pmu = device->pmu; in nvkm_pmu_fan_controlled() local 37 if (pmu && pmu->func->code.size) in nvkm_pmu_fan_controlled() 50 if (pmu && pmu->func->pgob) in nvkm_pmu_pgob() 51 pmu->func->pgob(pmu, enable); in nvkm_pmu_pgob() 58 return pmu->func->recv(pmu); in nvkm_pmu_recv() 65 if (!pmu || !pmu->func->send) in nvkm_pmu_send() 76 pmu->func->intr(pmu); in nvkm_pmu_intr() 88 pmu->func->fini(pmu); in nvkm_pmu_fini() 101 return pmu->func->init(pmu); in nvkm_pmu_init() 144 ret = nvkm_falcon_ctor(pmu->func->flcn, &pmu->subdev, pmu->subdev.name, in nvkm_pmu_ctor() [all …]
|
H A D | gt215.c | 37 mutex_lock(&pmu->send.mutex); in gt215_pmu_send() 54 pmu->recv.message = message; in gt215_pmu_send() 55 pmu->recv.process = process; in gt215_pmu_send() 65 pmu->send.base)); in gt215_pmu_send() 77 wait_event(pmu->recv.wait, (pmu->recv.process == 0)); in gt215_pmu_send() 105 pmu->recv.base)); in gt215_pmu_recv() 116 if (pmu->recv.process) { in gt215_pmu_recv() 121 pmu->recv.process = 0; in gt215_pmu_recv() 122 wake_up(&pmu->recv.wait); in gt215_pmu_recv() 207 if (pmu->func->enabled(pmu)) { in gt215_pmu_init() [all …]
|
H A D | gk20a.c | 116 struct gk20a_pmu *pmu = in gk20a_pmu_dvfs_work() local 149 gk20a_pmu_dvfs_target(pmu, &state); in gk20a_pmu_dvfs_work() 153 gk20a_pmu_dvfs_reset_dev_status(pmu); in gk20a_pmu_dvfs_work() 158 gk20a_pmu_fini(struct nvkm_pmu *pmu) in gk20a_pmu_fini() argument 160 struct gk20a_pmu *gpmu = gk20a_pmu(pmu); in gk20a_pmu_fini() 163 nvkm_falcon_put(&pmu->falcon, &pmu->subdev); in gk20a_pmu_fini() 167 gk20a_pmu_init(struct nvkm_pmu *pmu) in gk20a_pmu_init() argument 215 struct gk20a_pmu *pmu; in gk20a_pmu_new() local 218 if (!(pmu = kzalloc(sizeof(*pmu), GFP_KERNEL))) in gk20a_pmu_new() 220 *ppmu = &pmu->base; in gk20a_pmu_new() [all …]
|
H A D | gm20b.c | 42 struct nvkm_pmu *pmu = container_of(falcon, typeof(*pmu), falcon); in gm20b_pmu_acr_bootstrap_falcon() local 129 struct nvkm_pmu *pmu = priv; in gm20b_pmu_acr_init_wpr_callback() local 139 complete_all(&pmu->wpr_ready); in gm20b_pmu_acr_init_wpr_callback() 187 if (!pmu->initmsg_received) { in gm20b_pmu_recv() 188 int ret = pmu->func->initmsg(pmu); in gm20b_pmu_recv() 194 pmu->initmsg_received = true; in gm20b_pmu_recv() 205 flush_work(&pmu->recv.work); in gm20b_pmu_fini() 211 nvkm_falcon_put(&pmu->falcon, &pmu->subdev); in gm20b_pmu_fini() 222 ret = nvkm_falcon_get(&pmu->falcon, &pmu->subdev); in gm20b_pmu_init() 226 pmu->initmsg_received = false; in gm20b_pmu_init() [all …]
|
/openbmc/linux/tools/perf/util/ |
H A D | pmus.c | 113 if (pmu) in perf_pmus__find() 250 bool use_core_pmus = !pmu || pmu->is_core; in perf_pmus__scan() 282 bool use_core_pmus = !pmu || pmu->is_core; in perf_pmus__scan_skip_duplicates() 284 const char *last_pmu_name = (pmu && pmu->name) ? pmu->name : ""; in perf_pmus__scan_skip_duplicates() 376 a_iscpu = as->pmu ? as->pmu->is_core : true; in cmp_sevent() 382 if (as->pmu != bs->pmu) { in cmp_sevent() 419 s->pmu = info->pmu; in perf_pmus__print_pmu_events__callback() 452 while ((pmu = scan_fn(pmu)) != NULL) in perf_pmus__print_pmu_events() 466 while ((pmu = scan_fn(pmu)) != NULL) { in perf_pmus__print_pmu_events() 589 struct perf_pmu *pmu = evsel->pmu; in evsel__find_pmu() local [all …]
|
H A D | pmu.c | 595 .pmu = pmu, in perf_pmu__new_alias() 980 if (pmu_uncore_alias_match(pe->pmu, pmu->name) && in pmu_add_sys_aliases_iter_fn() 1035 pmu = zalloc(sizeof(*pmu)); in perf_pmu__lookup() 1036 if (!pmu) in perf_pmu__lookup() 1076 pmu->max_precise = pmu_max_precise(dirfd, pmu); in perf_pmu__lookup() 1099 struct perf_pmu *pmu = zalloc(sizeof(*pmu)); in perf_pmu__create_placeholder_core_pmu() local 1101 if (!pmu) in perf_pmu__create_placeholder_core_pmu() 1149 return pmu && pmu->auxtrace; in evsel__is_aux_event() 1683 nr += pmu_events_table__num_events(pmu->events_table, pmu) - pmu->cpu_json_aliases; in perf_pmu__num_events() 1732 .pmu = pmu, in perf_pmu__for_each_event() [all …]
|
H A D | pmu.h | 179 const struct perf_pmu *pmu; member 194 void pmu_add_sys_aliases(struct perf_pmu *pmu); 198 int perf_pmu__config_terms(struct perf_pmu *pmu, 214 bool perf_pmu__auto_merge_stats(const struct perf_pmu *pmu); 216 size_t perf_pmu__num_events(struct perf_pmu *pmu); 225 bool perf_pmu__is_software(const struct perf_pmu *pmu); 239 void pmu_add_cpu_aliases_table(struct perf_pmu *pmu, 242 char *perf_pmu__getcpuid(struct perf_pmu *pmu); 249 int perf_pmu__caps_parse(struct perf_pmu *pmu); 254 void perf_pmu__warn_invalid_formats(struct perf_pmu *pmu); [all …]
|
/openbmc/linux/drivers/gpu/drm/i915/ |
H A D | i915_pmu.c | 215 struct i915_pmu *pmu = &i915->pmu; in get_rc6() local 430 struct i915_pmu *pmu = &i915->pmu; in frequency_sample() local 473 struct i915_pmu *pmu = &i915->pmu; in i915_sample() local 591 struct i915_pmu *pmu = &i915->pmu; in i915_pmu_event_init() local 633 struct i915_pmu *pmu = &i915->pmu; in __i915_pmu_event_read() local 692 struct i915_pmu *pmu = &i915->pmu; in i915_pmu_event_read() local 714 struct i915_pmu *pmu = &i915->pmu; in i915_pmu_enable() local 778 struct i915_pmu *pmu = &i915->pmu; in i915_pmu_disable() local 824 struct i915_pmu *pmu = &i915->pmu; in i915_pmu_event_start() local 837 struct i915_pmu *pmu = &i915->pmu; in i915_pmu_event_stop() local [all …]
|
/openbmc/linux/drivers/soc/dove/ |
H A D | pmu.c | 108 pmu->reset.of_node = pmu->of_node; in pmu_reset_init() 146 struct pmu_data *pmu = pmu_dom->pmu; in pmu_domain_power_off() local 180 struct pmu_data *pmu = pmu_dom->pmu; in pmu_domain_power_on() local 315 pmu = kzalloc(sizeof(*pmu), GFP_KERNEL); in dove_init_pmu_legacy() 316 if (!pmu) in dove_init_pmu_legacy() 330 domain->pmu = pmu; in dove_init_pmu_legacy() 390 pmu = kzalloc(sizeof(*pmu), GFP_KERNEL); in dove_init_pmu() 396 pmu->pmc_base = of_iomap(pmu->of_node, 0); in dove_init_pmu() 397 pmu->pmu_base = of_iomap(pmu->of_node, 1); in dove_init_pmu() 398 if (!pmu->pmc_base || !pmu->pmu_base) { in dove_init_pmu() [all …]
|
/openbmc/linux/drivers/perf/ |
H A D | fsl_imx8_ddr_perf.c | 97 struct pmu pmu; member 320 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_is_enhanced_filtered() local 373 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_init() local 393 if (event->group_leader->pmu != event->pmu && in ddr_perf_event_init() 407 if (sibling->pmu != event->pmu && in ddr_perf_event_init() 477 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_update() local 511 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_start() local 528 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_add() local 569 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_stop() local 607 .pmu = (struct pmu) { in ddr_perf_init() [all …]
|
H A D | fsl_imx9_ddr_perf.c | 57 struct pmu pmu; member 401 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_update() local 415 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_init() local 435 if (event->group_leader->pmu != event->pmu && in ddr_perf_event_init() 440 if (sibling->pmu != event->pmu && in ddr_perf_event_init() 453 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_start() local 465 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_add() local 490 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_stop() local 502 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_del() local 529 .pmu = (struct pmu) { in ddr_perf_init() [all …]
|
H A D | marvell_cn10k_ddr_pmu.c | 125 struct pmu pmu; member 325 struct cn10k_ddr_pmu *pmu = to_cn10k_ddr_pmu(event->pmu); in cn10k_ddr_perf_event_init() local 342 if (event->group_leader->pmu != event->pmu && in cn10k_ddr_perf_event_init() 408 struct cn10k_ddr_pmu *pmu = to_cn10k_ddr_pmu(event->pmu); in cn10k_ddr_perf_event_update() local 424 struct cn10k_ddr_pmu *pmu = to_cn10k_ddr_pmu(event->pmu); in cn10k_ddr_perf_event_start() local 512 static void cn10k_ddr_perf_pmu_enable(struct pmu *pmu) in cn10k_ddr_perf_pmu_enable() argument 520 static void cn10k_ddr_perf_pmu_disable(struct pmu *pmu) in cn10k_ddr_perf_pmu_disable() argument 592 cn10k_ddr_perf_pmu_disable(&pmu->pmu); in cn10k_ddr_pmu_overflow_handler() 593 cn10k_ddr_perf_pmu_enable(&pmu->pmu); in cn10k_ddr_pmu_overflow_handler() 627 perf_pmu_migrate_context(&pmu->pmu, cpu, target); in cn10k_ddr_pmu_offline_cpu() [all …]
|
H A D | arm_pmu_platform.c | 37 ret = info->init(pmu); in probe_current_pmu() 112 pmu->pmu.capabilities |= PERF_PMU_CAP_NO_INTERRUPT; in pmu_parse_irqs() 192 struct arm_pmu *pmu; in arm_pmu_device_probe() local 195 pmu = armpmu_alloc(); in arm_pmu_device_probe() 196 if (!pmu) in arm_pmu_device_probe() 199 pmu->plat_device = pdev; in arm_pmu_device_probe() 201 ret = pmu_parse_irqs(pmu); in arm_pmu_device_probe() 216 ret = init_fn(pmu); in arm_pmu_device_probe() 231 ret = armpmu_register(pmu); in arm_pmu_device_probe() 240 armpmu_free_irqs(pmu); in arm_pmu_device_probe() [all …]
|
H A D | riscv_pmu_legacy.c | 116 pmu->ctr_start = pmu_legacy_ctr_start; in pmu_legacy_init() 117 pmu->ctr_stop = NULL; in pmu_legacy_init() 121 pmu->ctr_clear_idx = NULL; in pmu_legacy_init() 122 pmu->ctr_read = pmu_legacy_read_ctr; in pmu_legacy_init() 126 pmu->pmu.capabilities |= PERF_PMU_CAP_NO_INTERRUPT; in pmu_legacy_init() 127 pmu->pmu.capabilities |= PERF_PMU_CAP_NO_EXCLUDE; in pmu_legacy_init() 129 perf_pmu_register(&pmu->pmu, "cpu", PERF_TYPE_RAW); in pmu_legacy_init() 134 struct riscv_pmu *pmu = NULL; in pmu_legacy_device_probe() local 136 pmu = riscv_pmu_alloc(); in pmu_legacy_device_probe() 137 if (!pmu) in pmu_legacy_device_probe() [all …]
|
/openbmc/linux/arch/x86/kvm/vmx/ |
H A D | pmu_intel.c | 77 pmu->fixed_ctr_ctrl = data; in reprogram_fixed_counters() 152 counters = pmu->gp_counters; in intel_rdpmc_ecx_to_pmc() 223 get_fixed_pmc(pmu, msr) || get_fw_gp_pmc(pmu, msr) || in intel_is_valid_msr() 301 pmu->event_count++; in intel_pmu_create_guest_lbr_event() 413 pmu->pebs_enable = data; in intel_pmu_set_msr() 421 pmu->ds_area = data; in intel_pmu_set_msr() 427 pmu->pebs_data_cfg = data; in intel_pmu_set_msr() 514 if (!pmu->version) in intel_pmu_refresh() 527 if (pmu->version == 1) { in intel_pmu_refresh() 550 pmu->global_status_mask = pmu->global_ctrl_mask in intel_pmu_refresh() [all …]
|
/openbmc/linux/arch/x86/kvm/svm/ |
H A D | pmu.c | 114 return pmu->version > 0; in amd_is_valid_msr() 120 return pmu->version > 1; in amd_is_valid_msr() 124 return pmu->version > 1; in amd_is_valid_msr() 170 data &= ~pmu->reserved_bits; in amd_pmu_set_msr() 186 pmu->version = 1; in amd_pmu_refresh() 188 pmu->version = 2; in amd_pmu_refresh() 203 pmu->nr_arch_gp_counters = min_t(unsigned int, pmu->nr_arch_gp_counters, in amd_pmu_refresh() 206 if (pmu->version > 1) { in amd_pmu_refresh() 207 pmu->global_ctrl_mask = ~((1ull << pmu->nr_arch_gp_counters) - 1); in amd_pmu_refresh() 208 pmu->global_status_mask = pmu->global_ctrl_mask; in amd_pmu_refresh() [all …]
|
/openbmc/linux/drivers/pmdomain/starfive/ |
H A D | jh71xx-pmu.c | 71 struct jh71xx_pmu *pmu; member 77 struct jh71xx_pmu *pmu = pmd->pmu; in jh71xx_pmu_get_state() local 89 struct jh71xx_pmu *pmu = pmd->pmu; in jh71xx_pmu_set_state() local 241 pmd->pmu = pmu; in jh71xx_pmu_init_domain() 270 pmu = devm_kzalloc(dev, sizeof(*pmu), GFP_KERNEL); in jh71xx_pmu_probe() 271 if (!pmu) in jh71xx_pmu_probe() 279 if (pmu->irq < 0) in jh71xx_pmu_probe() 280 return pmu->irq; in jh71xx_pmu_probe() 294 if (!pmu->genpd) in jh71xx_pmu_probe() 297 pmu->dev = dev; in jh71xx_pmu_probe() [all …]
|
/openbmc/linux/drivers/perf/amlogic/ |
H A D | meson_ddr_pmu_core.c | 21 struct pmu pmu; member 41 pmu->info.hw_info->enable(&pmu->info); in dmc_pmu_enable() 49 pmu->info.hw_info->disable(&pmu->info); in dmc_pmu_disable() 56 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in meson_ddr_set_axi_filter() local 87 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in meson_ddr_perf_event_update() local 94 pmu->info.hw_info->get_counters(&pmu->info, &dc); in meson_ddr_perf_event_update() 120 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in meson_ddr_perf_event_init() local 145 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in meson_ddr_perf_event_start() local 147 memset(&pmu->counters, 0, sizeof(pmu->counters)); in meson_ddr_perf_event_start() 175 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in meson_ddr_perf_event_stop() local [all …]
|
/openbmc/linux/Documentation/devicetree/bindings/soc/samsung/ |
H A D | exynos-pmu.yaml | 18 - samsung,exynos3250-pmu 19 - samsung,exynos4210-pmu 20 - samsung,exynos4212-pmu 21 - samsung,exynos4412-pmu 22 - samsung,exynos5250-pmu 23 - samsung,exynos5260-pmu 24 - samsung,exynos5410-pmu 25 - samsung,exynos5420-pmu 27 - samsung,exynos7-pmu 28 - samsung,exynos850-pmu [all …]
|
/openbmc/linux/Documentation/devicetree/bindings/arm/ |
H A D | pmu.yaml | 22 - apm,potenza-pmu 24 - apple,blizzard-pmu 26 - apple,icestorm-pmu 28 - arm,arm1136-pmu 29 - arm,arm1176-pmu 31 - arm,cortex-a5-pmu 32 - arm,cortex-a7-pmu 33 - arm,cortex-a8-pmu 34 - arm,cortex-a9-pmu 64 - brcm,vulcan-pmu [all …]
|
/openbmc/u-boot/drivers/power/ |
H A D | ftpmu010.c | 22 oscc = readl(&pmu->OSCC); in ftpmu010_32768osc_enable() 24 writel(oscc, &pmu->OSCC); in ftpmu010_32768osc_enable() 31 oscc = readl(&pmu->OSCC); in ftpmu010_32768osc_enable() 33 writel(oscc, &pmu->OSCC); in ftpmu010_32768osc_enable() 42 mfpsr = readl(&pmu->MFPSR); in ftpmu010_mfpsr_select_dev() 44 writel(mfpsr, &pmu->MFPSR); in ftpmu010_mfpsr_select_dev() 52 mfpsr = readl(&pmu->MFPSR); in ftpmu010_mfpsr_diselect_dev() 54 writel(mfpsr, &pmu->MFPSR); in ftpmu010_mfpsr_diselect_dev() 63 pdllcr0 = readl(&pmu->PDLLCR0); in ftpmu010_dlldis_disable() 65 writel(pdllcr0, &pmu->PDLLCR0); in ftpmu010_dlldis_disable() [all …]
|
/openbmc/linux/Documentation/devicetree/bindings/arm/rockchip/ |
H A D | pmu.yaml | 22 - rockchip,px30-pmu 23 - rockchip,rk3066-pmu 24 - rockchip,rk3128-pmu 25 - rockchip,rk3288-pmu 26 - rockchip,rk3368-pmu 27 - rockchip,rk3399-pmu 28 - rockchip,rk3568-pmu 29 - rockchip,rk3588-pmu 30 - rockchip,rv1126-pmu 39 - rockchip,px30-pmu [all …]
|
/openbmc/linux/arch/x86/events/intel/ |
H A D | uncore.h | 88 struct pmu *pmu; /* for custom pmu ops */ member 122 struct pmu pmu; member 151 struct intel_uncore_pmu *pmu; member 273 box->pmu->type->mmio_offset * box->pmu->pmu_idx; in uncore_mmio_box_ctl() 308 struct intel_uncore_pmu *pmu = box->pmu; in uncore_msr_box_offset() local 310 pmu->type->msr_offsets[pmu->pmu_idx] : in uncore_msr_box_offset() 311 pmu->type->msr_offset * pmu->pmu_idx; in uncore_msr_box_offset() 380 struct intel_uncore_pmu *pmu = box->pmu; in uncore_freerunning_counter() local 385 pmu->type->freerunning[type].box_offsets[pmu->pmu_idx] : in uncore_freerunning_counter() 386 pmu->type->freerunning[type].box_offset * pmu->pmu_idx); in uncore_freerunning_counter() [all …]
|
/openbmc/linux/arch/x86/events/ |
H A D | rapl.c | 110 struct pmu *pmu; member 116 struct pmu pmu; member 212 hrtimer_start(&pmu->hrtimer, pmu->timer_interval, in rapl_start_hrtimer() 257 pmu->n_active++; in __rapl_pmu_event_start() 364 if (!pmu) in rapl_pmu_event_init() 556 pmu->cpu = -1; in rapl_cpu_offline() 564 perf_pmu_migrate_context(pmu->pmu, cpu, target); in rapl_cpu_offline() 574 if (!pmu) { in rapl_cpu_online() 576 if (!pmu) in rapl_cpu_online() 581 pmu->pmu = &rapl_pmus->pmu; in rapl_cpu_online() [all …]
|
/openbmc/linux/Documentation/devicetree/bindings/pinctrl/ |
H A D | marvell,dove-pinctrl.txt | 18 mpp0 0 gpio, pmu, uart2(rts), sdio0(cd), lcd0(pwm), pmu* 19 mpp1 1 gpio, pmu, uart2(cts), sdio0(wp), lcd1(pwm), pmu* 24 mpp4 4 gpio, pmu, uart3(rts), sdio1(cd), spi1(miso), pmu* 25 mpp5 5 gpio, pmu, uart3(cts), sdio1(wp), spi1(cs), pmu* 26 mpp6 6 gpio, pmu, uart3(txd), sdio1(buspwr), spi1(mosi), pmu* 27 mpp7 7 gpio, pmu, uart3(rxd), sdio1(ledctrl), spi1(sck), pmu* 28 mpp8 8 gpio, pmu, watchdog(rstout), pmu* 29 mpp9 9 gpio, pmu, pex1(clkreq), pmu* 30 mpp10 10 gpio, pmu, ssp(sclk), pmu* 37 mpp14 14 gpio, pmu, uart2(txd), sdio1(buspwr), ssp(rxd), pmu* [all …]
|