/openbmc/linux/include/net/ |
H A D | fq_impl.h | 36 idx = flow - fq->flows; in __fq_adjust_removal() 152 flow = &fq->flows[idx]; in fq_flow_classify() 160 tin->flows++; in fq_flow_classify() 173 struct fq_flow *cur = &fq->flows[i]; in fq_find_fattest_flow() 361 fq->flows = kvcalloc(fq->flows_cnt, sizeof(fq->flows[0]), GFP_KERNEL); in fq_init() 362 if (!fq->flows) in fq_init() 367 kvfree(fq->flows); in fq_init() 368 fq->flows = NULL; in fq_init() 373 fq_flow_init(&fq->flows[i]); in fq_init() 384 fq_flow_reset(fq, &fq->flows[i], free_func); in fq_reset() [all …]
|
H A D | fq.h | 57 u32 flows; member 69 struct fq_flow *flows; member
|
/openbmc/linux/drivers/crypto/allwinner/sun8i-ss/ |
H A D | sun8i-ss-core.c | 76 ss->flows[flow].stat_req++; in sun8i_ss_run_task() 132 reinit_completion(&ss->flows[flow].complete); in sun8i_ss_run_task() 133 ss->flows[flow].status = 0; in sun8i_ss_run_task() 138 wait_for_completion_interruptible_timeout(&ss->flows[flow].complete, in sun8i_ss_run_task() 140 if (ss->flows[flow].status == 0) { in sun8i_ss_run_task() 159 ss->flows[flow].status = 1; in ss_irq_handler() 160 complete(&ss->flows[flow].complete); in ss_irq_handler() 483 ss->flows[i].stat_req); in sun8i_ss_debugfs_show() 541 crypto_engine_exit(ss->flows[i].engine); in sun8i_ss_free_flows() 553 ss->flows = devm_kcalloc(ss->dev, MAXFLOW, sizeof(struct sun8i_ss_flow), in allocate_flows() [all …]
|
H A D | sun8i-ss-prng.c | 134 reinit_completion(&ss->flows[flow].complete); in sun8i_ss_prng_generate() 135 ss->flows[flow].status = 0; in sun8i_ss_prng_generate() 141 wait_for_completion_interruptible_timeout(&ss->flows[flow].complete, in sun8i_ss_prng_generate() 143 if (ss->flows[flow].status == 0) { in sun8i_ss_prng_generate()
|
H A D | sun8i-ss-hash.c | 307 ss->flows[flow].stat_req++; in sun8i_ss_run_hash_task() 340 reinit_completion(&ss->flows[flow].complete); in sun8i_ss_run_hash_task() 341 ss->flows[flow].status = 0; in sun8i_ss_run_hash_task() 346 wait_for_completion_interruptible_timeout(&ss->flows[flow].complete, in sun8i_ss_run_hash_task() 348 if (ss->flows[flow].status == 0) { in sun8i_ss_run_hash_task() 424 engine = ss->flows[e].engine; in sun8i_ss_hash_digest() 522 result = ss->flows[rctx->flow].result; in sun8i_ss_hash_run() 523 pad = ss->flows[rctx->flow].pad; in sun8i_ss_hash_run()
|
H A D | sun8i-ss-cipher.c | 130 struct sun8i_ss_flow *sf = &ss->flows[rctx->flow]; in sun8i_ss_setup_ivs() 191 struct sun8i_ss_flow *sf = &ss->flows[rctx->flow]; in sun8i_ss_cipher() 360 engine = op->ss->flows[e].engine; in sun8i_ss_skdecrypt() 381 engine = op->ss->flows[e].engine; in sun8i_ss_skencrypt()
|
/openbmc/linux/samples/bpf/ |
H A D | do_hbm_test.sh | 78 flows=1 150 -f=*|--flows=*) 151 flows="${i#*=}" 278 while [ $flow_cnt -le $flows ] ; do 320 while [ $flow_cnt -le $flows ] ; do 346 iperf3 -c $host -p $port -i 0 -P $flows -f m -t $dur > iperf.$id 366 while [ $flow_cnt -le $flows ] ; do 386 while [ $flow_cnt -le $flows ] ; do
|
/openbmc/linux/net/sched/ |
H A D | sch_fq_codel.c | 53 struct fq_codel_flow *flows; /* Flows table [flows_cnt] */ member 164 flow = &q->flows[idx]; in fq_codel_drop() 204 flow = &q->flows[idx]; in fq_codel_enqueue() 265 q->backlogs[flow - q->flows] -= qdisc_pkt_len(skb); in dequeue_func() 343 struct fq_codel_flow *flow = q->flows + i; in fq_codel_reset() 380 if (q->flows) in fq_codel_change() 456 kvfree(q->flows); in fq_codel_destroy() 488 if (!q->flows) { in fq_codel_init() 489 q->flows = kvcalloc(q->flows_cnt, in fq_codel_init() 492 if (!q->flows) { in fq_codel_init() [all …]
|
H A D | sch_fq_pie.c | 57 struct fq_pie_flow *flows; member 150 sel_flow = &q->flows[idx]; in fq_pie_qdisc_enqueue() 305 if (q->flows) { in fq_pie_change() 392 &q->flows[q->flows_cursor].vars, in fq_pie_timer() 393 q->flows[q->flows_cursor].backlog); in fq_pie_timer() 440 q->flows = kvcalloc(q->flows_cnt, sizeof(struct fq_pie_flow), in fq_pie_init() 442 if (!q->flows) { in fq_pie_init() 447 struct fq_pie_flow *flow = q->flows + idx; in fq_pie_init() 531 struct fq_pie_flow *flow = q->flows + idx; in fq_pie_reset() 549 kvfree(q->flows); in fq_pie_destroy()
|
H A D | sch_fq.c | 120 u32 flows; member 254 q->flows -= fcnt; in fq_gc() 304 if (q->flows >= (2U << q->fq_trees_log) && in fq_classify() 305 q->inactive_flows > q->flows/2) in fq_classify() 359 q->flows++; in fq_classify() 692 q->flows = 0; in fq_reset() 737 q->flows -= fcnt; in fq_rehash() 1024 st.flows = q->flows; in fq_dump_stats()
|
H A D | sch_cake.c | 151 struct cake_flow flows[CAKE_QUEUES]; member 801 q->flows[reduced_hash].set)) { in cake_hash() 819 if (!q->flows[outer_hash + k].set) { in cake_hash() 834 if (!q->flows[outer_hash + k].set) { in cake_hash() 849 if (q->flows[outer_hash + k].set == CAKE_SET_BULK) { in cake_hash() 850 cake_dec_srchost_bulk_flow_count(q, &q->flows[outer_hash + k], flow_mode); in cake_hash() 851 cake_dec_dsthost_bulk_flow_count(q, &q->flows[outer_hash + k], flow_mode); in cake_hash() 876 q->flows[reduced_hash].srchost = srchost_idx; in cake_hash() 878 if (q->flows[reduced_hash].set == CAKE_SET_BULK) in cake_hash() 879 cake_inc_srchost_bulk_flow_count(q, &q->flows[reduced_hash], flow_mode); in cake_hash() [all …]
|
/openbmc/linux/drivers/dma/ti/ |
H A D | k3-udma-glue.c | 84 struct k3_udma_glue_rx_flow *flows; member 645 struct k3_udma_glue_rx_flow *flow = &rx_chn->flows[flow_num]; in k3_udma_glue_release_rx_flow() 665 struct k3_udma_glue_rx_flow *flow = &rx_chn->flows[flow_idx]; in k3_udma_glue_cfg_rx_flow() 962 rx_chn->flows = devm_kcalloc(dev, rx_chn->flow_num, in k3_udma_glue_request_rx_chn_priv() 963 sizeof(*rx_chn->flows), GFP_KERNEL); in k3_udma_glue_request_rx_chn_priv() 964 if (!rx_chn->flows) { in k3_udma_glue_request_rx_chn_priv() 974 rx_chn->flows[i].udma_rflow_id = rx_chn->flow_id_base + i; in k3_udma_glue_request_rx_chn_priv() 1043 rx_chn->flows = devm_kcalloc(dev, rx_chn->flow_num, in k3_udma_glue_request_remote_rx_chn() 1044 sizeof(*rx_chn->flows), GFP_KERNEL); in k3_udma_glue_request_remote_rx_chn() 1045 if (!rx_chn->flows) { in k3_udma_glue_request_remote_rx_chn() [all …]
|
/openbmc/linux/drivers/media/platform/amphion/ |
H A D | vpu_dbg.c | 193 for (i = 0; i < ARRAY_SIZE(inst->flows); i++) { in vpu_dbg_instance() 194 u32 idx = (inst->flow_idx + i) % (ARRAY_SIZE(inst->flows)); in vpu_dbg_instance() 196 if (!inst->flows[idx]) in vpu_dbg_instance() 199 inst->flows[idx] >= VPU_MSG_ID_NOOP ? "M" : "C", in vpu_dbg_instance() 200 vpu_id_name(inst->flows[idx])); in vpu_dbg_instance() 492 inst->flows[inst->flow_idx] = flow; in vpu_inst_record_flow() 493 inst->flow_idx = (inst->flow_idx + 1) % (ARRAY_SIZE(inst->flows)); in vpu_inst_record_flow()
|
/openbmc/linux/drivers/infiniband/hw/hfi1/ |
H A D | tid_rdma.c | 757 u32 generation = rcd->flows[flow_idx].generation; in kern_setup_hw_flow() 775 rcd->flows[flow_idx].generation = in kern_clear_hw_flow() 776 kern_flow_generation_next(rcd->flows[flow_idx].generation); in kern_clear_hw_flow() 804 rcd->flows[fs->index].generation = fs->generation; in hfi1_kern_setup_hw_flow() 853 rcd->flows[i].generation = mask_generation(get_random_u32()); in hfi1_kern_init_ctxt_generations() 1465 struct tid_rdma_flow *flow = &req->flows[req->setup_head]; in hfi1_kern_exp_rcv_setup() 1556 struct tid_rdma_flow *flow = &req->flows[req->clear_tail]; in hfi1_kern_exp_rcv_clear() 1612 kfree(req->flows); in hfi1_kern_exp_rcv_free_flows() 1613 req->flows = NULL; in hfi1_kern_exp_rcv_free_flows() 1634 struct tid_rdma_flow *flows; in hfi1_kern_exp_rcv_alloc_flows() local [all …]
|
/openbmc/linux/Documentation/networking/ |
H A D | nf_flowtable.rst | 33 specifies what flows are placed into the flowtable. Hence, packets follow the 34 classic IP forwarding path unless the user explicitly instruct flows to use this 111 You can identify offloaded flows through the [OFFLOAD] tag when listing your 130 instead the real device is sufficient for the flowtable to track your flows. 198 There is a workqueue that adds the flows to the hardware. Note that a few 202 You can identify hardware offloaded flows through the [HW_OFFLOAD] tag when
|
H A D | scaling.rst | 31 of logical flows. Packets for each flow are steered to a separate receive 188 to the same CPU is CPU load imbalance if flows vary in packet rate. 194 Flow Limit is an optional RPS feature that prioritizes small flows 195 during CPU contention by dropping packets from large flows slightly 196 ahead of those from small flows. It is active only when an RPS or RFS 202 new packet is dropped. Packets from other flows are still only 206 even large flows maintain connectivity. 224 identification of large flows and fewer false positives. The default 261 flows to the CPUs where those flows are being processed. The flow hash 266 same CPU. Indeed, with many flows and few CPUs, it is very likely that [all …]
|
H A D | openvswitch.rst | 16 table" that userspace populates with "flows" that map from keys based 104 A wildcarded flow can represent a group of exact match flows. Each '1' bit 108 by reduce the number of new flows need to be processed by the user space program. 120 two possible approaches: reactively install flows as they miss the kernel 130 The behavior when using overlapping wildcarded flows is undefined. It is the 133 performs best-effort detection of overlapping wildcarded flows and may reject 146 future operations. The kernel is not required to index flows by the original
|
/openbmc/linux/Documentation/admin-guide/pm/ |
H A D | system-wide.rst | 11 suspend-flows
|
/openbmc/linux/Documentation/userspace-api/media/mediactl/ |
H A D | media-controller-model.rst | 26 by an entity flows from the entity's output to one or more entity 31 pads, either on the same entity or on different entities. Data flows
|
/openbmc/linux/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | en_rep.h | 183 struct list_head flows; member 208 struct list_head flows; member
|
H A D | eswitch_offloads.c | 1175 struct mlx5_flow_handle **flows; in esw_add_fdb_peer_miss_rules() local 1194 flows = kvcalloc(nvports, sizeof(*flows), GFP_KERNEL); in esw_add_fdb_peer_miss_rules() 1195 if (!flows) { in esw_add_fdb_peer_miss_rules() 1215 flows[vport->index] = flow; in esw_add_fdb_peer_miss_rules() 1227 flows[vport->index] = flow; in esw_add_fdb_peer_miss_rules() 1241 flows[vport->index] = flow; in esw_add_fdb_peer_miss_rules() 1256 flows[vport->index] = flow; in esw_add_fdb_peer_miss_rules() 1267 esw->fdb_table.offloads.peer_miss_rules[pfindex] = flows; in esw_add_fdb_peer_miss_rules() 1274 if (!flows[vport->index]) in esw_add_fdb_peer_miss_rules() 1276 mlx5_del_flow_rules(flows[vport->index]); in esw_add_fdb_peer_miss_rules() [all …]
|
/openbmc/linux/net/core/ |
H A D | pktgen.c | 415 struct flow_state *flows; member 2310 return !!(pkt_dev->flows[flow].flags & F_INIT); in f_seen() 2318 if (pkt_dev->flows[flow].count >= pkt_dev->lflow) { in f_pick() 2320 pkt_dev->flows[flow].count = 0; in f_pick() 2321 pkt_dev->flows[flow].flags = 0; in f_pick() 2330 if (pkt_dev->flows[flow].count > pkt_dev->lflow) { in f_pick() 2331 pkt_dev->flows[flow].count = 0; in f_pick() 2332 pkt_dev->flows[flow].flags = 0; in f_pick() 2347 struct xfrm_state *x = pkt_dev->flows[flow].x; in get_ipsec_sa() 2366 pkt_dev->flows[flow].x = x; in get_ipsec_sa() [all …]
|
/openbmc/linux/Documentation/admin-guide/blockdev/drbd/ |
H A D | figures.rst | 5 Data flows that Relate some functions, and write packets
|
/openbmc/openbmc/meta-openembedded/meta-networking/recipes-support/sngrep/ |
H A D | sngrep_1.8.2.bb | 2 DESCRIPTION = "Tool for displaying SIP calls message flows from terminal"
|
/openbmc/linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | tc_tun_encap.c | 301 list_for_each_entry(efi, &e->flows, list) { in mlx5e_take_all_encap_flows() 422 list_for_each_entry_safe(efi, tmp, &e->flows, list) { in mlx5e_tc_update_neigh_used_value() 467 WARN_ON(!list_empty(&e->flows)); in mlx5e_encap_dealloc() 484 WARN_ON(!list_empty(&d->flows)); in mlx5e_decap_dealloc() 892 INIT_LIST_HEAD(&e->flows); in mlx5e_attach_encap() 926 list_add(&flow->encaps[out_index].list, &e->flows); in mlx5e_attach_encap() 991 INIT_LIST_HEAD(&d->flows); in mlx5e_attach_decap() 1014 list_add(&flow->l3_to_l2_reformat, &d->flows); in mlx5e_attach_decap()
|