/openbmc/qemu/target/riscv/ |
H A D | pmu.c | 286 GUINT_TO_POINTER(event_idx)); in riscv_pmu_incr_ctr() 309 uint32_t event_idx; in riscv_pmu_ctr_monitor_instructions() local 322 event_idx = RISCV_PMU_EVENT_HW_INSTRUCTIONS; in riscv_pmu_ctr_monitor_instructions() 324 GUINT_TO_POINTER(event_idx))); in riscv_pmu_ctr_monitor_instructions() 335 uint32_t event_idx; in riscv_pmu_ctr_monitor_cycles() local 348 event_idx = RISCV_PMU_EVENT_HW_CPU_CYCLES; in riscv_pmu_ctr_monitor_cycles() 350 GUINT_TO_POINTER(event_idx))); in riscv_pmu_ctr_monitor_cycles() 382 uint32_t event_idx; in riscv_pmu_update_event_map() local 400 event_idx = value & MHPMEVENT_IDX_MASK; in riscv_pmu_update_event_map() 402 GUINT_TO_POINTER(event_idx))) { in riscv_pmu_update_event_map() [all …]
|
H A D | pmu.h | 33 int riscv_pmu_incr_ctr(RISCVCPU *cpu, enum riscv_pmu_event_idx event_idx);
|
/openbmc/linux/arch/powerpc/perf/ |
H A D | hv-24x7.c | 700 size_t event_idx, in catalog_event_len_validate() argument 711 if (event_idx >= event_entry_count) { in catalog_event_len_validate() 719 event_idx); in catalog_event_len_validate() 727 event_idx, ev_len, event); in catalog_event_len_validate() 732 event_idx, ev_len, ev_end, end, in catalog_event_len_validate() 898 event_idx, nl, name); in create_events_from_catalog() 905 event_idx, nl, name, event->domain); in create_events_from_catalog() 913 event_idx_last = event_idx; in create_events_from_catalog() 940 event = event_data, event_idx = 0; in create_events_from_catalog() 941 event_idx < event_idx_last; in create_events_from_catalog() [all …]
|
H A D | core-book3s.c | 2225 .event_idx = power_pmu_event_idx,
|
/openbmc/qemu/block/ |
H A D | linux-aio.c | 66 int event_idx; member 214 s->event_idx))) { in qemu_laio_process_completions() 215 for (s->event_idx = 0; s->event_idx < s->event_max; ) { in qemu_laio_process_completions() 216 struct iocb *iocb = events[s->event_idx].obj; in qemu_laio_process_completions() 220 laiocb->ret = io_event_ret(&events[s->event_idx]); in qemu_laio_process_completions() 224 s->event_idx++; in qemu_laio_process_completions() 235 s->event_idx = 0; in qemu_laio_process_completions()
|
/openbmc/linux/arch/riscv/kvm/ |
H A D | vcpu_pmu.c | 210 fevent_code = get_event_code(pmc->event_idx); in pmu_ctr_read() 361 fevent_code = get_event_code(pmc->event_idx); in kvm_riscv_vcpu_pmu_ctr_start() 416 fevent_code = get_event_code(pmc->event_idx); in kvm_riscv_vcpu_pmu_ctr_stop() 445 pmc->event_idx = SBI_PMU_EVENT_IDX_INVALID; in kvm_riscv_vcpu_pmu_ctr_stop() 525 pmc->event_idx = eidx; in kvm_riscv_vcpu_pmu_ctr_cfg_match() 587 pmc->event_idx = SBI_PMU_EVENT_IDX_INVALID; in kvm_riscv_vcpu_pmu_init() 624 pmc->event_idx = SBI_PMU_EVENT_IDX_INVALID; in kvm_riscv_vcpu_pmu_deinit()
|
/openbmc/qemu/include/standard-headers/linux/ |
H A D | virtio_ring.h | 220 static inline int vring_need_event(uint16_t event_idx, uint16_t new_idx, uint16_t old) in vring_need_event() argument 227 return (uint16_t)(new_idx - event_idx - 1) < (uint16_t)(new_idx - old); in vring_need_event()
|
/openbmc/linux/include/uapi/linux/ |
H A D | virtio_ring.h | 222 static inline int vring_need_event(__u16 event_idx, __u16 new_idx, __u16 old) in vring_need_event() argument 229 return (__u16)(new_idx - event_idx - 1) < (__u16)(new_idx - old); in vring_need_event()
|
/openbmc/u-boot/include/ |
H A D | virtio_ring.h | 150 static inline int vring_need_event(__u16 event_idx, __u16 new_idx, __u16 old) in vring_need_event() argument 159 return (__u16)(new_idx - event_idx - 1) < (__u16)(new_idx - old); in vring_need_event()
|
/openbmc/linux/drivers/perf/ |
H A D | cxl_pmu.c | 89 int event_idx; /* configurable counters */ member 203 pmu_ev->event_idx = j; in cxl_pmu_parse_caps() 526 int *event_idx) in cxl_pmu_get_event_idx() argument 743 int event_idx = 0; in cxl_pmu_event_add() local 747 rc = cxl_pmu_get_event_idx(event, &idx, &event_idx); in cxl_pmu_event_add() 754 hwc->event_base = event_idx; in cxl_pmu_event_add()
|
H A D | marvell_cn10k_tad_pmu.c | 83 u32 event_idx = event->attr.config; in tad_pmu_event_counter_start() local 99 reg_val = event_idx & 0xFF; in tad_pmu_event_counter_start()
|
H A D | riscv_pmu_sbi.c | 85 uint32_t event_idx; member 429 ret = pmu_cache_event_map[cache_type][cache_op][cache_result].event_idx; in pmu_event_find_cache() 457 ret = pmu_hw_event_map[event->attr.config].event_idx; in pmu_sbi_event_map()
|
H A D | riscv_pmu.c | 416 .event_idx = riscv_pmu_event_idx, in riscv_pmu_alloc()
|
H A D | arm_pmuv3.c | 1215 cpu_pmu->pmu.event_idx = armv8pmu_user_event_idx; in armv8_pmu_init()
|
/openbmc/linux/Documentation/devicetree/bindings/perf/ |
H A D | riscv,pmu.yaml | 48 - description: event_idx, a 20-bit wide encoding of the event type and 64 - description: first event_idx of the range of events 65 - description: last event_idx of the range of events
|
/openbmc/linux/arch/riscv/include/asm/ |
H A D | kvm_vcpu_pmu.h | 39 unsigned long event_idx; member
|
/openbmc/linux/include/net/ |
H A D | amt.h | 369 u8 event_idx; member
|
H A D | cfg80211.h | 7544 int approxlen, int event_idx, gfp_t gfp) in cfg80211_vendor_event_alloc() argument 7548 0, event_idx, approxlen, gfp); in cfg80211_vendor_event_alloc() 7579 int event_idx, gfp_t gfp) in cfg80211_vendor_event_alloc_ucast() argument 7583 portid, event_idx, approxlen, gfp); in cfg80211_vendor_event_alloc_ucast()
|
/openbmc/linux/drivers/net/ |
H A D | amt.c | 904 index = (amt->event_idx + amt->nr_events) % AMT_MAX_EVENTS; in amt_queue_event() 908 amt->event_idx %= AMT_MAX_EVENTS; in amt_queue_event() 2877 event = amt->events[amt->event_idx].event; in amt_event_work() 2878 skb = amt->events[amt->event_idx].skb; in amt_event_work() 2879 amt->events[amt->event_idx].event = AMT_EVENT_NONE; in amt_event_work() 2880 amt->events[amt->event_idx].skb = NULL; in amt_event_work() 2882 amt->event_idx++; in amt_event_work() 2883 amt->event_idx %= AMT_MAX_EVENTS; in amt_event_work() 2990 amt->event_idx = 0; in amt_dev_open()
|
/openbmc/linux/arch/x86/events/intel/ |
H A D | p4.c | 731 static bool p4_event_match_cpu_model(unsigned int event_idx) in p4_event_match_cpu_model() argument 734 if (event_idx == P4_EVENT_INSTR_COMPLETED) { in p4_event_match_cpu_model()
|
/openbmc/linux/drivers/nvme/host/ |
H A D | pci.c | 346 static inline int nvme_dbbuf_need_event(u16 event_idx, u16 new_idx, u16 old) in nvme_dbbuf_need_event() argument 348 return (u16)(new_idx - event_idx - 1) < (u16)(new_idx - old); in nvme_dbbuf_need_event() 356 u16 old_value, event_idx; in nvme_dbbuf_update_and_check_event() local 375 event_idx = le32_to_cpu(*dbbuf_ei); in nvme_dbbuf_update_and_check_event() 376 if (!nvme_dbbuf_need_event(event_idx, value, old_value)) in nvme_dbbuf_update_and_check_event()
|
/openbmc/linux/drivers/virtio/ |
H A D | virtio_ring.c | 1556 u16 new, old, off_wrap, flags, wrap_counter, event_idx; in virtqueue_kick_prepare_packed() local 1592 event_idx = off_wrap & ~(1 << VRING_PACKED_EVENT_F_WRAP_CTR); in virtqueue_kick_prepare_packed() 1594 event_idx -= vq->packed.vring.num; in virtqueue_kick_prepare_packed() 1596 needs_kick = vring_need_event(event_idx, new, old); in virtqueue_kick_prepare_packed()
|
/openbmc/linux/include/linux/ |
H A D | perf_event.h | 450 int (*event_idx) (struct perf_event *event); /*optional */ member
|
/openbmc/linux/drivers/gpu/drm/i915/ |
H A D | i915_pmu.c | 1301 pmu->base.event_idx = i915_pmu_event_event_idx; in i915_pmu_register()
|
/openbmc/qemu/target/arm/ |
H A D | helper.c | 1394 uint16_t event_idx = supported_event_map[event]; in pmevcntr_op_start() local 1395 count = pm_events[event_idx].get_count(env); in pmevcntr_op_start() 1417 uint16_t event_idx = supported_event_map[event]; in pmevcntr_op_finish() local 1424 overflow_in = pm_events[event_idx].ns_per_count(delta); in pmevcntr_op_finish() 1674 uint16_t event_idx = supported_event_map[new_event]; in pmevtyper_write() local 1675 count = pm_events[event_idx].get_count(env); in pmevtyper_write() 1730 uint16_t event_idx = supported_event_map[event]; in pmevtyper_rawwrite() local 1732 pm_events[event_idx].get_count(env); in pmevtyper_rawwrite()
|