Home
last modified time | relevance | path

Searched refs:pmu (Results 1 – 25 of 791) sorted by relevance

12345678910>>...32

/openbmc/linux/drivers/gpu/drm/nouveau/nvkm/subdev/pmu/
H A Dbase.c32 struct nvkm_pmu *pmu = device->pmu; in nvkm_pmu_fan_controlled() local
37 if (pmu && pmu->func->code.size) in nvkm_pmu_fan_controlled()
50 if (pmu && pmu->func->pgob) in nvkm_pmu_pgob()
51 pmu->func->pgob(pmu, enable); in nvkm_pmu_pgob()
58 return pmu->func->recv(pmu); in nvkm_pmu_recv()
65 if (!pmu || !pmu->func->send) in nvkm_pmu_send()
76 pmu->func->intr(pmu); in nvkm_pmu_intr()
88 pmu->func->fini(pmu); in nvkm_pmu_fini()
101 return pmu->func->init(pmu); in nvkm_pmu_init()
144 ret = nvkm_falcon_ctor(pmu->func->flcn, &pmu->subdev, pmu->subdev.name, in nvkm_pmu_ctor()
[all …]
H A Dgt215.c37 mutex_lock(&pmu->send.mutex); in gt215_pmu_send()
54 pmu->recv.message = message; in gt215_pmu_send()
55 pmu->recv.process = process; in gt215_pmu_send()
65 pmu->send.base)); in gt215_pmu_send()
77 wait_event(pmu->recv.wait, (pmu->recv.process == 0)); in gt215_pmu_send()
105 pmu->recv.base)); in gt215_pmu_recv()
116 if (pmu->recv.process) { in gt215_pmu_recv()
121 pmu->recv.process = 0; in gt215_pmu_recv()
122 wake_up(&pmu->recv.wait); in gt215_pmu_recv()
207 if (pmu->func->enabled(pmu)) { in gt215_pmu_init()
[all …]
H A Dgk20a.c116 struct gk20a_pmu *pmu = in gk20a_pmu_dvfs_work() local
149 gk20a_pmu_dvfs_target(pmu, &state); in gk20a_pmu_dvfs_work()
153 gk20a_pmu_dvfs_reset_dev_status(pmu); in gk20a_pmu_dvfs_work()
158 gk20a_pmu_fini(struct nvkm_pmu *pmu) in gk20a_pmu_fini() argument
160 struct gk20a_pmu *gpmu = gk20a_pmu(pmu); in gk20a_pmu_fini()
163 nvkm_falcon_put(&pmu->falcon, &pmu->subdev); in gk20a_pmu_fini()
167 gk20a_pmu_init(struct nvkm_pmu *pmu) in gk20a_pmu_init() argument
215 struct gk20a_pmu *pmu; in gk20a_pmu_new() local
218 if (!(pmu = kzalloc(sizeof(*pmu), GFP_KERNEL))) in gk20a_pmu_new()
220 *ppmu = &pmu->base; in gk20a_pmu_new()
[all …]
H A Dgm20b.c42 struct nvkm_pmu *pmu = container_of(falcon, typeof(*pmu), falcon); in gm20b_pmu_acr_bootstrap_falcon() local
129 struct nvkm_pmu *pmu = priv; in gm20b_pmu_acr_init_wpr_callback() local
139 complete_all(&pmu->wpr_ready); in gm20b_pmu_acr_init_wpr_callback()
187 if (!pmu->initmsg_received) { in gm20b_pmu_recv()
188 int ret = pmu->func->initmsg(pmu); in gm20b_pmu_recv()
194 pmu->initmsg_received = true; in gm20b_pmu_recv()
205 flush_work(&pmu->recv.work); in gm20b_pmu_fini()
211 nvkm_falcon_put(&pmu->falcon, &pmu->subdev); in gm20b_pmu_fini()
222 ret = nvkm_falcon_get(&pmu->falcon, &pmu->subdev); in gm20b_pmu_init()
226 pmu->initmsg_received = false; in gm20b_pmu_init()
[all …]
/openbmc/linux/tools/perf/util/
H A Dpmus.c113 if (pmu) in perf_pmus__find()
250 bool use_core_pmus = !pmu || pmu->is_core; in perf_pmus__scan()
282 bool use_core_pmus = !pmu || pmu->is_core; in perf_pmus__scan_skip_duplicates()
284 const char *last_pmu_name = (pmu && pmu->name) ? pmu->name : ""; in perf_pmus__scan_skip_duplicates()
376 a_iscpu = as->pmu ? as->pmu->is_core : true; in cmp_sevent()
382 if (as->pmu != bs->pmu) { in cmp_sevent()
419 s->pmu = info->pmu; in perf_pmus__print_pmu_events__callback()
452 while ((pmu = scan_fn(pmu)) != NULL) in perf_pmus__print_pmu_events()
466 while ((pmu = scan_fn(pmu)) != NULL) { in perf_pmus__print_pmu_events()
589 struct perf_pmu *pmu = evsel->pmu; in evsel__find_pmu() local
[all …]
H A Dpmu.c595 .pmu = pmu, in perf_pmu__new_alias()
980 if (pmu_uncore_alias_match(pe->pmu, pmu->name) && in pmu_add_sys_aliases_iter_fn()
1035 pmu = zalloc(sizeof(*pmu)); in perf_pmu__lookup()
1036 if (!pmu) in perf_pmu__lookup()
1076 pmu->max_precise = pmu_max_precise(dirfd, pmu); in perf_pmu__lookup()
1099 struct perf_pmu *pmu = zalloc(sizeof(*pmu)); in perf_pmu__create_placeholder_core_pmu() local
1101 if (!pmu) in perf_pmu__create_placeholder_core_pmu()
1149 return pmu && pmu->auxtrace; in evsel__is_aux_event()
1683 nr += pmu_events_table__num_events(pmu->events_table, pmu) - pmu->cpu_json_aliases; in perf_pmu__num_events()
1732 .pmu = pmu, in perf_pmu__for_each_event()
[all …]
H A Dpmu.h179 const struct perf_pmu *pmu; member
194 void pmu_add_sys_aliases(struct perf_pmu *pmu);
198 int perf_pmu__config_terms(struct perf_pmu *pmu,
214 bool perf_pmu__auto_merge_stats(const struct perf_pmu *pmu);
216 size_t perf_pmu__num_events(struct perf_pmu *pmu);
225 bool perf_pmu__is_software(const struct perf_pmu *pmu);
239 void pmu_add_cpu_aliases_table(struct perf_pmu *pmu,
242 char *perf_pmu__getcpuid(struct perf_pmu *pmu);
249 int perf_pmu__caps_parse(struct perf_pmu *pmu);
254 void perf_pmu__warn_invalid_formats(struct perf_pmu *pmu);
[all …]
/openbmc/linux/drivers/gpu/drm/i915/
H A Di915_pmu.c215 struct i915_pmu *pmu = &i915->pmu; in get_rc6() local
430 struct i915_pmu *pmu = &i915->pmu; in frequency_sample() local
473 struct i915_pmu *pmu = &i915->pmu; in i915_sample() local
591 struct i915_pmu *pmu = &i915->pmu; in i915_pmu_event_init() local
633 struct i915_pmu *pmu = &i915->pmu; in __i915_pmu_event_read() local
692 struct i915_pmu *pmu = &i915->pmu; in i915_pmu_event_read() local
714 struct i915_pmu *pmu = &i915->pmu; in i915_pmu_enable() local
778 struct i915_pmu *pmu = &i915->pmu; in i915_pmu_disable() local
824 struct i915_pmu *pmu = &i915->pmu; in i915_pmu_event_start() local
837 struct i915_pmu *pmu = &i915->pmu; in i915_pmu_event_stop() local
[all …]
/openbmc/linux/drivers/soc/dove/
H A Dpmu.c108 pmu->reset.of_node = pmu->of_node; in pmu_reset_init()
146 struct pmu_data *pmu = pmu_dom->pmu; in pmu_domain_power_off() local
180 struct pmu_data *pmu = pmu_dom->pmu; in pmu_domain_power_on() local
315 pmu = kzalloc(sizeof(*pmu), GFP_KERNEL); in dove_init_pmu_legacy()
316 if (!pmu) in dove_init_pmu_legacy()
330 domain->pmu = pmu; in dove_init_pmu_legacy()
390 pmu = kzalloc(sizeof(*pmu), GFP_KERNEL); in dove_init_pmu()
396 pmu->pmc_base = of_iomap(pmu->of_node, 0); in dove_init_pmu()
397 pmu->pmu_base = of_iomap(pmu->of_node, 1); in dove_init_pmu()
398 if (!pmu->pmc_base || !pmu->pmu_base) { in dove_init_pmu()
[all …]
/openbmc/linux/drivers/perf/
H A Dfsl_imx8_ddr_perf.c97 struct pmu pmu; member
320 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_is_enhanced_filtered() local
373 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_init() local
393 if (event->group_leader->pmu != event->pmu && in ddr_perf_event_init()
407 if (sibling->pmu != event->pmu && in ddr_perf_event_init()
477 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_update() local
511 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_start() local
528 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_add() local
569 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_stop() local
607 .pmu = (struct pmu) { in ddr_perf_init()
[all …]
H A Dfsl_imx9_ddr_perf.c57 struct pmu pmu; member
401 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_update() local
415 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_init() local
435 if (event->group_leader->pmu != event->pmu && in ddr_perf_event_init()
440 if (sibling->pmu != event->pmu && in ddr_perf_event_init()
453 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_start() local
465 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_add() local
490 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_stop() local
502 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in ddr_perf_event_del() local
529 .pmu = (struct pmu) { in ddr_perf_init()
[all …]
H A Dmarvell_cn10k_ddr_pmu.c125 struct pmu pmu; member
325 struct cn10k_ddr_pmu *pmu = to_cn10k_ddr_pmu(event->pmu); in cn10k_ddr_perf_event_init() local
342 if (event->group_leader->pmu != event->pmu && in cn10k_ddr_perf_event_init()
408 struct cn10k_ddr_pmu *pmu = to_cn10k_ddr_pmu(event->pmu); in cn10k_ddr_perf_event_update() local
424 struct cn10k_ddr_pmu *pmu = to_cn10k_ddr_pmu(event->pmu); in cn10k_ddr_perf_event_start() local
512 static void cn10k_ddr_perf_pmu_enable(struct pmu *pmu) in cn10k_ddr_perf_pmu_enable() argument
520 static void cn10k_ddr_perf_pmu_disable(struct pmu *pmu) in cn10k_ddr_perf_pmu_disable() argument
592 cn10k_ddr_perf_pmu_disable(&pmu->pmu); in cn10k_ddr_pmu_overflow_handler()
593 cn10k_ddr_perf_pmu_enable(&pmu->pmu); in cn10k_ddr_pmu_overflow_handler()
627 perf_pmu_migrate_context(&pmu->pmu, cpu, target); in cn10k_ddr_pmu_offline_cpu()
[all …]
H A Darm_pmu_platform.c37 ret = info->init(pmu); in probe_current_pmu()
112 pmu->pmu.capabilities |= PERF_PMU_CAP_NO_INTERRUPT; in pmu_parse_irqs()
192 struct arm_pmu *pmu; in arm_pmu_device_probe() local
195 pmu = armpmu_alloc(); in arm_pmu_device_probe()
196 if (!pmu) in arm_pmu_device_probe()
199 pmu->plat_device = pdev; in arm_pmu_device_probe()
201 ret = pmu_parse_irqs(pmu); in arm_pmu_device_probe()
216 ret = init_fn(pmu); in arm_pmu_device_probe()
231 ret = armpmu_register(pmu); in arm_pmu_device_probe()
240 armpmu_free_irqs(pmu); in arm_pmu_device_probe()
[all …]
H A Driscv_pmu_legacy.c116 pmu->ctr_start = pmu_legacy_ctr_start; in pmu_legacy_init()
117 pmu->ctr_stop = NULL; in pmu_legacy_init()
121 pmu->ctr_clear_idx = NULL; in pmu_legacy_init()
122 pmu->ctr_read = pmu_legacy_read_ctr; in pmu_legacy_init()
126 pmu->pmu.capabilities |= PERF_PMU_CAP_NO_INTERRUPT; in pmu_legacy_init()
127 pmu->pmu.capabilities |= PERF_PMU_CAP_NO_EXCLUDE; in pmu_legacy_init()
129 perf_pmu_register(&pmu->pmu, "cpu", PERF_TYPE_RAW); in pmu_legacy_init()
134 struct riscv_pmu *pmu = NULL; in pmu_legacy_device_probe() local
136 pmu = riscv_pmu_alloc(); in pmu_legacy_device_probe()
137 if (!pmu) in pmu_legacy_device_probe()
[all …]
/openbmc/linux/arch/x86/kvm/vmx/
H A Dpmu_intel.c77 pmu->fixed_ctr_ctrl = data; in reprogram_fixed_counters()
152 counters = pmu->gp_counters; in intel_rdpmc_ecx_to_pmc()
223 get_fixed_pmc(pmu, msr) || get_fw_gp_pmc(pmu, msr) || in intel_is_valid_msr()
301 pmu->event_count++; in intel_pmu_create_guest_lbr_event()
413 pmu->pebs_enable = data; in intel_pmu_set_msr()
421 pmu->ds_area = data; in intel_pmu_set_msr()
427 pmu->pebs_data_cfg = data; in intel_pmu_set_msr()
514 if (!pmu->version) in intel_pmu_refresh()
527 if (pmu->version == 1) { in intel_pmu_refresh()
550 pmu->global_status_mask = pmu->global_ctrl_mask in intel_pmu_refresh()
[all …]
/openbmc/linux/arch/x86/kvm/svm/
H A Dpmu.c114 return pmu->version > 0; in amd_is_valid_msr()
120 return pmu->version > 1; in amd_is_valid_msr()
124 return pmu->version > 1; in amd_is_valid_msr()
170 data &= ~pmu->reserved_bits; in amd_pmu_set_msr()
186 pmu->version = 1; in amd_pmu_refresh()
188 pmu->version = 2; in amd_pmu_refresh()
203 pmu->nr_arch_gp_counters = min_t(unsigned int, pmu->nr_arch_gp_counters, in amd_pmu_refresh()
206 if (pmu->version > 1) { in amd_pmu_refresh()
207 pmu->global_ctrl_mask = ~((1ull << pmu->nr_arch_gp_counters) - 1); in amd_pmu_refresh()
208 pmu->global_status_mask = pmu->global_ctrl_mask; in amd_pmu_refresh()
[all …]
/openbmc/linux/drivers/pmdomain/starfive/
H A Djh71xx-pmu.c71 struct jh71xx_pmu *pmu; member
77 struct jh71xx_pmu *pmu = pmd->pmu; in jh71xx_pmu_get_state() local
89 struct jh71xx_pmu *pmu = pmd->pmu; in jh71xx_pmu_set_state() local
241 pmd->pmu = pmu; in jh71xx_pmu_init_domain()
270 pmu = devm_kzalloc(dev, sizeof(*pmu), GFP_KERNEL); in jh71xx_pmu_probe()
271 if (!pmu) in jh71xx_pmu_probe()
279 if (pmu->irq < 0) in jh71xx_pmu_probe()
280 return pmu->irq; in jh71xx_pmu_probe()
294 if (!pmu->genpd) in jh71xx_pmu_probe()
297 pmu->dev = dev; in jh71xx_pmu_probe()
[all …]
/openbmc/linux/drivers/perf/amlogic/
H A Dmeson_ddr_pmu_core.c21 struct pmu pmu; member
41 pmu->info.hw_info->enable(&pmu->info); in dmc_pmu_enable()
49 pmu->info.hw_info->disable(&pmu->info); in dmc_pmu_disable()
56 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in meson_ddr_set_axi_filter() local
87 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in meson_ddr_perf_event_update() local
94 pmu->info.hw_info->get_counters(&pmu->info, &dc); in meson_ddr_perf_event_update()
120 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in meson_ddr_perf_event_init() local
145 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in meson_ddr_perf_event_start() local
147 memset(&pmu->counters, 0, sizeof(pmu->counters)); in meson_ddr_perf_event_start()
175 struct ddr_pmu *pmu = to_ddr_pmu(event->pmu); in meson_ddr_perf_event_stop() local
[all …]
/openbmc/linux/Documentation/devicetree/bindings/soc/samsung/
H A Dexynos-pmu.yaml18 - samsung,exynos3250-pmu
19 - samsung,exynos4210-pmu
20 - samsung,exynos4212-pmu
21 - samsung,exynos4412-pmu
22 - samsung,exynos5250-pmu
23 - samsung,exynos5260-pmu
24 - samsung,exynos5410-pmu
25 - samsung,exynos5420-pmu
27 - samsung,exynos7-pmu
28 - samsung,exynos850-pmu
[all …]
/openbmc/linux/Documentation/devicetree/bindings/arm/
H A Dpmu.yaml22 - apm,potenza-pmu
24 - apple,blizzard-pmu
26 - apple,icestorm-pmu
28 - arm,arm1136-pmu
29 - arm,arm1176-pmu
31 - arm,cortex-a5-pmu
32 - arm,cortex-a7-pmu
33 - arm,cortex-a8-pmu
34 - arm,cortex-a9-pmu
64 - brcm,vulcan-pmu
[all …]
/openbmc/u-boot/drivers/power/
H A Dftpmu010.c22 oscc = readl(&pmu->OSCC); in ftpmu010_32768osc_enable()
24 writel(oscc, &pmu->OSCC); in ftpmu010_32768osc_enable()
31 oscc = readl(&pmu->OSCC); in ftpmu010_32768osc_enable()
33 writel(oscc, &pmu->OSCC); in ftpmu010_32768osc_enable()
42 mfpsr = readl(&pmu->MFPSR); in ftpmu010_mfpsr_select_dev()
44 writel(mfpsr, &pmu->MFPSR); in ftpmu010_mfpsr_select_dev()
52 mfpsr = readl(&pmu->MFPSR); in ftpmu010_mfpsr_diselect_dev()
54 writel(mfpsr, &pmu->MFPSR); in ftpmu010_mfpsr_diselect_dev()
63 pdllcr0 = readl(&pmu->PDLLCR0); in ftpmu010_dlldis_disable()
65 writel(pdllcr0, &pmu->PDLLCR0); in ftpmu010_dlldis_disable()
[all …]
/openbmc/linux/Documentation/devicetree/bindings/arm/rockchip/
H A Dpmu.yaml22 - rockchip,px30-pmu
23 - rockchip,rk3066-pmu
24 - rockchip,rk3128-pmu
25 - rockchip,rk3288-pmu
26 - rockchip,rk3368-pmu
27 - rockchip,rk3399-pmu
28 - rockchip,rk3568-pmu
29 - rockchip,rk3588-pmu
30 - rockchip,rv1126-pmu
39 - rockchip,px30-pmu
[all …]
/openbmc/linux/arch/x86/events/intel/
H A Duncore.h88 struct pmu *pmu; /* for custom pmu ops */ member
122 struct pmu pmu; member
151 struct intel_uncore_pmu *pmu; member
273 box->pmu->type->mmio_offset * box->pmu->pmu_idx; in uncore_mmio_box_ctl()
308 struct intel_uncore_pmu *pmu = box->pmu; in uncore_msr_box_offset() local
310 pmu->type->msr_offsets[pmu->pmu_idx] : in uncore_msr_box_offset()
311 pmu->type->msr_offset * pmu->pmu_idx; in uncore_msr_box_offset()
380 struct intel_uncore_pmu *pmu = box->pmu; in uncore_freerunning_counter() local
385 pmu->type->freerunning[type].box_offsets[pmu->pmu_idx] : in uncore_freerunning_counter()
386 pmu->type->freerunning[type].box_offset * pmu->pmu_idx); in uncore_freerunning_counter()
[all …]
/openbmc/linux/arch/x86/events/
H A Drapl.c110 struct pmu *pmu; member
116 struct pmu pmu; member
212 hrtimer_start(&pmu->hrtimer, pmu->timer_interval, in rapl_start_hrtimer()
257 pmu->n_active++; in __rapl_pmu_event_start()
364 if (!pmu) in rapl_pmu_event_init()
556 pmu->cpu = -1; in rapl_cpu_offline()
564 perf_pmu_migrate_context(pmu->pmu, cpu, target); in rapl_cpu_offline()
574 if (!pmu) { in rapl_cpu_online()
576 if (!pmu) in rapl_cpu_online()
581 pmu->pmu = &rapl_pmus->pmu; in rapl_cpu_online()
[all …]
/openbmc/linux/Documentation/devicetree/bindings/pinctrl/
H A Dmarvell,dove-pinctrl.txt18 mpp0 0 gpio, pmu, uart2(rts), sdio0(cd), lcd0(pwm), pmu*
19 mpp1 1 gpio, pmu, uart2(cts), sdio0(wp), lcd1(pwm), pmu*
24 mpp4 4 gpio, pmu, uart3(rts), sdio1(cd), spi1(miso), pmu*
25 mpp5 5 gpio, pmu, uart3(cts), sdio1(wp), spi1(cs), pmu*
26 mpp6 6 gpio, pmu, uart3(txd), sdio1(buspwr), spi1(mosi), pmu*
27 mpp7 7 gpio, pmu, uart3(rxd), sdio1(ledctrl), spi1(sck), pmu*
28 mpp8 8 gpio, pmu, watchdog(rstout), pmu*
29 mpp9 9 gpio, pmu, pex1(clkreq), pmu*
30 mpp10 10 gpio, pmu, ssp(sclk), pmu*
37 mpp14 14 gpio, pmu, uart2(txd), sdio1(buspwr), ssp(rxd), pmu*
[all …]

12345678910>>...32