/openbmc/linux/drivers/gpu/drm/i915/ |
H A D | i915_tasklet.h | 24 if (!atomic_fetch_inc(&t->count)) in __tasklet_disable_sync_once()
|
H A D | i915_active.c | 498 if (!atomic_fetch_inc(&ref->count)) in __i915_active_activate()
|
/openbmc/linux/drivers/infiniband/hw/hfi1/ |
H A D | netdev_rx.c | 312 if (atomic_fetch_inc(&rx->netdevs)) in hfi1_netdev_rx_init() 399 if (atomic_fetch_inc(&rx->enabled)) in hfi1_netdev_enable_queues()
|
H A D | trace.c | 499 unsigned long packet_count = atomic_fetch_inc(&hist.count); in hfi1_trace_print_rsm_hist() 507 val = atomic_fetch_inc(count); in hfi1_trace_print_rsm_hist()
|
/openbmc/linux/net/mac802154/ |
H A D | util.c | 72 if (!atomic_fetch_inc(&local->phy->hold_txs)) in ieee802154_hold_queue()
|
/openbmc/linux/kernel/ |
H A D | watchdog_perf.c | 151 if (!atomic_fetch_inc(&watchdog_cpus)) in watchdog_hardlockup_enable()
|
/openbmc/linux/drivers/gpu/drm/i915/gt/ |
H A D | intel_timeline.c | 215 if (atomic_fetch_inc(&tl->pin_count)) { in intel_timeline_pin() 262 if (!atomic_fetch_inc(&tl->active_count)) { in intel_timeline_enter()
|
H A D | intel_engine_pm.c | 138 if (!atomic_fetch_inc(&tl->active_count)) in __queue_and_release_pm()
|
H A D | intel_ring.c | 42 if (atomic_fetch_inc(&ring->pin_count)) in intel_ring_pin()
|
H A D | gen8_ppgtt.c | 253 if (atomic_fetch_inc(&pt->used) >> gen8_pd_shift(1) && in __gen8_ppgtt_clear()
|
H A D | intel_rps.c | 1027 if (!atomic_fetch_inc(&slpc->num_waiters)) { in intel_rps_boost() 1037 if (atomic_fetch_inc(&rps->num_waiters)) in intel_rps_boost()
|
/openbmc/linux/drivers/accel/ |
H A D | drm_accel.c | 145 atomic_fetch_inc(&dev->open_count); in accel_open()
|
/openbmc/linux/tools/memory-model/ |
H A D | linux-kernel.def | 84 atomic_fetch_inc(X) __atomic_fetch_op{mb}(X,+,1)
|
/openbmc/linux/drivers/perf/arm_cspmu/ |
H A D | nvidia_cspmu.c | 344 atomic_fetch_inc(&pmu_generic_idx)); in nv_cspmu_format_name()
|
H A D | arm_cspmu.c | 336 atomic_fetch_inc(&pmu_idx[pmu_type])); in arm_cspmu_get_name()
|
/openbmc/linux/io_uring/ |
H A D | poll.c | 83 return !(atomic_fetch_inc(&req->poll_refs) & IO_POLL_REF_MASK); in io_poll_get_ownership_slowpath() 96 return !(atomic_fetch_inc(&req->poll_refs) & IO_POLL_REF_MASK); in io_poll_get_ownership()
|
/openbmc/linux/drivers/gpu/drm/i915/gem/ |
H A D | i915_gem_shrinker.c | 488 if (!atomic_fetch_inc(&obj->mm.shrink_pin) && in i915_gem_object_make_unshrinkable()
|
/openbmc/linux/drivers/net/wwan/t7xx/ |
H A D | t7xx_netdev.c | 96 if (!atomic_fetch_inc(&ccmni_ctl->napi_usr_refcnt)) in t7xx_ccmni_open()
|
/openbmc/linux/net/ipv4/ |
H A D | ip_gre.c | 472 (flags & TUNNEL_SEQ) ? htonl(atomic_fetch_inc(&tunnel->o_seqno)) : 0); in __gre_xmit() 510 (flags & TUNNEL_SEQ) ? htonl(atomic_fetch_inc(&tunnel->o_seqno)) : 0); in gre_fb_xmit() 593 proto, 0, htonl(atomic_fetch_inc(&tunnel->o_seqno))); in erspan_fb_xmit()
|
/openbmc/linux/fs/dlm/ |
H A D | midcomms.c | 976 mh->seq = atomic_fetch_inc(&mh->node->seq_send); in midcomms_new_msg_cb() 1486 h->u.h_seq = cpu_to_le32(atomic_fetch_inc(&rd->node->seq_send)); in midcomms_new_rawmsg_cb()
|
/openbmc/linux/net/ipv6/ |
H A D | ioam6_iptunnel.c | 310 pkt_cnt = atomic_fetch_inc(&ilwt->pkt_cnt); in ioam6_output()
|
H A D | ip6_gre.c | 794 (flags & TUNNEL_SEQ) ? htonl(atomic_fetch_inc(&tunnel->o_seqno)) in __gre6_xmit() 805 (flags & TUNNEL_SEQ) ? htonl(atomic_fetch_inc(&tunnel->o_seqno)) in __gre6_xmit() 1071 gre_build_header(skb, 8, TUNNEL_SEQ, proto, 0, htonl(atomic_fetch_inc(&t->o_seqno))); in ip6erspan_tunnel_xmit()
|
/openbmc/linux/drivers/net/ethernet/netronome/nfp/crypto/ |
H A D | tls.c | 371 if (!atomic_fetch_inc(&nn->ktls_no_space)) in nfp_net_tls_add()
|
/openbmc/linux/drivers/gpu/drm/ |
H A D | drm_file.c | 424 if (!atomic_fetch_inc(&dev->open_count)) in drm_open()
|
/openbmc/linux/include/linux/atomic/ |
H A D | atomic-instrumented.h | 520 atomic_fetch_inc(atomic_t *v) in atomic_fetch_inc() function
|