Home
last modified time | relevance | path

Searched refs:tb (Results 101 – 125 of 630) sorted by relevance

12345678910>>...26

/openbmc/linux/net/netfilter/
H A Dnft_log.c147 const struct nlattr * const tb[]) in nft_log_init() argument
155 if (tb[NFTA_LOG_LEVEL] != NULL && in nft_log_init()
156 tb[NFTA_LOG_GROUP] != NULL) in nft_log_init()
158 if (tb[NFTA_LOG_GROUP] != NULL) { in nft_log_init()
160 if (tb[NFTA_LOG_FLAGS] != NULL) in nft_log_init()
164 nla = tb[NFTA_LOG_PREFIX]; in nft_log_init()
176 if (tb[NFTA_LOG_LEVEL] != NULL) { in nft_log_init()
178 ntohl(nla_get_be32(tb[NFTA_LOG_LEVEL])); in nft_log_init()
187 if (tb[NFTA_LOG_FLAGS] != NULL) { in nft_log_init()
198 if (tb[NFTA_LOG_SNAPLEN] != NULL) { in nft_log_init()
[all …]
H A Dnft_range.c51 const struct nlattr * const tb[]) in nft_range_init() argument
65 if (!tb[NFTA_RANGE_SREG] || in nft_range_init()
66 !tb[NFTA_RANGE_OP] || in nft_range_init()
67 !tb[NFTA_RANGE_FROM_DATA] || in nft_range_init()
68 !tb[NFTA_RANGE_TO_DATA]) in nft_range_init()
72 tb[NFTA_RANGE_FROM_DATA]); in nft_range_init()
77 tb[NFTA_RANGE_TO_DATA]); in nft_range_init()
86 err = nft_parse_register_load(tb[NFTA_RANGE_SREG], &priv->sreg, in nft_range_init()
91 err = nft_parse_u32_check(tb[NFTA_RANGE_OP], U8_MAX, &op); in nft_range_init()
H A Dnft_dynset.c162 const struct nlattr * const tb[]) in nft_dynset_init() argument
173 if (tb[NFTA_DYNSET_SET_NAME] == NULL || in nft_dynset_init()
174 tb[NFTA_DYNSET_OP] == NULL || in nft_dynset_init()
175 tb[NFTA_DYNSET_SREG_KEY] == NULL) in nft_dynset_init()
178 if (tb[NFTA_DYNSET_FLAGS]) { in nft_dynset_init()
189 tb[NFTA_DYNSET_SET_NAME], in nft_dynset_init()
190 tb[NFTA_DYNSET_SET_ID], genmask); in nft_dynset_init()
208 if (tb[NFTA_DYNSET_TIMEOUT] != NULL) { in nft_dynset_init()
235 if ((tb[NFTA_DYNSET_EXPR] || tb[NFTA_DYNSET_EXPRESSIONS]) && in nft_dynset_init()
239 if (tb[NFTA_DYNSET_EXPR]) { in nft_dynset_init()
[all …]
H A Dnft_compat.c212 if (!tb[NFTA_RULE_COMPAT_PROTO] || !tb[NFTA_RULE_COMPAT_FLAGS]) in nft_parse_compat()
244 const struct nlattr * const tb[]) in nft_target_init() argument
498 const struct nlattr * const tb[], in __nft_match_init() argument
691 if (tb[NFTA_COMPAT_NAME] == NULL || in nfnl_compat_get_rcu()
692 tb[NFTA_COMPAT_REV] == NULL || in nfnl_compat_get_rcu()
693 tb[NFTA_COMPAT_TYPE] == NULL) in nfnl_compat_get_rcu()
795 if (tb[NFTA_MATCH_NAME] == NULL || in nft_match_select_ops()
796 tb[NFTA_MATCH_REV] == NULL || in nft_match_select_ops()
797 tb[NFTA_MATCH_INFO] == NULL) in nft_match_select_ops()
876 tb[NFTA_TARGET_REV] == NULL || in nft_target_select_ops()
[all …]
H A Dnft_inner.c288 struct nlattr *tb[NFT_EXPR_MAXATTR + 1]; member
293 const struct nlattr * const tb[]) in nft_inner_init() argument
300 if (!tb[NFTA_INNER_FLAGS] || in nft_inner_init()
301 !tb[NFTA_INNER_NUM] || in nft_inner_init()
302 !tb[NFTA_INNER_HDRSIZE] || in nft_inner_init()
303 !tb[NFTA_INNER_TYPE] || in nft_inner_init()
304 !tb[NFTA_INNER_EXPR]) in nft_inner_init()
307 flags = ntohl(nla_get_be32(tb[NFTA_INNER_FLAGS])); in nft_inner_init()
311 num = ntohl(nla_get_be32(tb[NFTA_INNER_NUM])); in nft_inner_init()
316 type = ntohl(nla_get_be32(tb[NFTA_INNER_TYPE])); in nft_inner_init()
[all …]
H A Dnft_payload.c653 if (!tb[NFTA_PAYLOAD_BASE] || !tb[NFTA_PAYLOAD_OFFSET] || in nft_payload_inner_init()
654 !tb[NFTA_PAYLOAD_LEN] || !tb[NFTA_PAYLOAD_DREG]) in nft_payload_inner_init()
938 const struct nlattr * const tb[]) in nft_payload_set_init() argument
948 if (tb[NFTA_PAYLOAD_CSUM_TYPE]) in nft_payload_set_init()
950 if (tb[NFTA_PAYLOAD_CSUM_OFFSET]) { in nft_payload_set_init()
958 if (tb[NFTA_PAYLOAD_CSUM_FLAGS]) { in nft_payload_set_init()
1044 if (tb[NFTA_PAYLOAD_BASE] == NULL || in nft_payload_select_ops()
1046 tb[NFTA_PAYLOAD_LEN] == NULL) in nft_payload_select_ops()
1060 if (tb[NFTA_PAYLOAD_SREG] != NULL) { in nft_payload_select_ops()
1061 if (tb[NFTA_PAYLOAD_DREG] != NULL) in nft_payload_select_ops()
[all …]
/openbmc/linux/net/sched/
H A Dcls_flow.c420 if (tb[TCA_FLOW_KEYS]) { in flow_change()
473 if (tb[TCA_FLOW_MODE]) in flow_change()
480 if (tb[TCA_FLOW_PERTURB]) { in flow_change()
489 if (!tb[TCA_FLOW_KEYS]) in flow_change()
493 if (tb[TCA_FLOW_MODE]) in flow_change()
522 if (tb[TCA_FLOW_KEYS]) { in flow_change()
529 if (tb[TCA_FLOW_MASK]) in flow_change()
531 if (tb[TCA_FLOW_XOR]) in flow_change()
533 if (tb[TCA_FLOW_RSHIFT]) in flow_change()
535 if (tb[TCA_FLOW_ADDEND]) in flow_change()
[all …]
H A Dact_bpf.c195 bpf_num_ops = nla_get_u16(tb[TCA_ACT_BPF_OPS_LEN]); in tcf_bpf_init_from_ops()
200 if (bpf_size != nla_len(tb[TCA_ACT_BPF_OPS])) in tcf_bpf_init_from_ops()
230 bpf_fd = nla_get_u32(tb[TCA_ACT_BPF_FD]); in tcf_bpf_init_from_efd()
236 if (tb[TCA_ACT_BPF_NAME]) { in tcf_bpf_init_from_efd()
286 struct nlattr *tb[TCA_ACT_BPF_MAX + 1]; in tcf_bpf_init() local
303 if (!tb[TCA_ACT_BPF_PARMS]) in tcf_bpf_init()
306 parm = nla_data(tb[TCA_ACT_BPF_PARMS]); in tcf_bpf_init()
335 is_bpf = tb[TCA_ACT_BPF_OPS_LEN] && tb[TCA_ACT_BPF_OPS]; in tcf_bpf_init()
336 is_ebpf = tb[TCA_ACT_BPF_FD]; in tcf_bpf_init()
345 ret = is_bpf ? tcf_bpf_init_from_ops(tb, &cfg) : in tcf_bpf_init()
[all …]
H A Dact_police.c47 struct nlattr *tb[TCA_POLICE_MAX + 1]; in tcf_police_init() local
67 if (tb[TCA_POLICE_TBF] == NULL) in tcf_police_init()
69 size = nla_len(tb[TCA_POLICE_TBF]); in tcf_police_init()
73 parm = nla_data(tb[TCA_POLICE_TBF]); in tcf_police_init()
122 } else if (tb[TCA_POLICE_AVRATE] && in tcf_police_init()
129 if (tb[TCA_POLICE_RESULT]) { in tcf_police_init()
139 if ((tb[TCA_POLICE_PKTRATE64] && !tb[TCA_POLICE_PKTBURST64]) || in tcf_police_init()
140 (!tb[TCA_POLICE_PKTRATE64] && tb[TCA_POLICE_PKTBURST64])) { in tcf_police_init()
170 rate64 = tb[TCA_POLICE_RATE64] ? in tcf_police_init()
192 if (tb[TCA_POLICE_AVRATE]) in tcf_police_init()
[all …]
H A Dsch_gred.c548 if (tb[TCA_GRED_VQ_FLAGS]) in gred_vq_apply()
579 if (!tb[TCA_GRED_VQ_DP]) { in gred_vq_validate()
593 if (tb[TCA_GRED_VQ_FLAGS]) { in gred_vq_validate()
658 if (tb[TCA_GRED_PARMS] == NULL && tb[TCA_GRED_STAB] == NULL) { in gred_change()
659 if (tb[TCA_GRED_LIMIT] != NULL) in gred_change()
665 tb[TCA_GRED_STAB] == NULL || in gred_change()
671 max_P = tb[TCA_GRED_MAX_P] ? nla_get_u32(tb[TCA_GRED_MAX_P]) : 0; in gred_change()
681 if (tb[TCA_GRED_VQ_LIST]) { in gred_change()
711 if (tb[TCA_GRED_VQ_LIST]) in gred_change()
747 if (tb[TCA_GRED_PARMS] || tb[TCA_GRED_STAB]) { in gred_init()
[all …]
H A Dact_skbedit.c125 struct nlattr *tb[TCA_SKBEDIT_MAX + 1]; in tcf_skbedit_init() local
144 if (tb[TCA_SKBEDIT_PARMS] == NULL) in tcf_skbedit_init()
147 if (tb[TCA_SKBEDIT_PRIORITY] != NULL) { in tcf_skbedit_init()
162 if (tb[TCA_SKBEDIT_PTYPE] != NULL) { in tcf_skbedit_init()
169 if (tb[TCA_SKBEDIT_MARK] != NULL) { in tcf_skbedit_init()
171 mark = nla_data(tb[TCA_SKBEDIT_MARK]); in tcf_skbedit_init()
174 if (tb[TCA_SKBEDIT_MASK] != NULL) { in tcf_skbedit_init()
176 mask = nla_data(tb[TCA_SKBEDIT_MASK]); in tcf_skbedit_init()
179 if (tb[TCA_SKBEDIT_FLAGS] != NULL) { in tcf_skbedit_init()
185 if (!tb[TCA_SKBEDIT_QUEUE_MAPPING] || in tcf_skbedit_init()
[all …]
H A Dsch_codel.c137 struct nlattr *tb[TCA_CODEL_MAX + 1]; in codel_change() local
148 if (tb[TCA_CODEL_TARGET]) { in codel_change()
149 u32 target = nla_get_u32(tb[TCA_CODEL_TARGET]); in codel_change()
154 if (tb[TCA_CODEL_CE_THRESHOLD]) { in codel_change()
155 u64 val = nla_get_u32(tb[TCA_CODEL_CE_THRESHOLD]); in codel_change()
160 if (tb[TCA_CODEL_INTERVAL]) { in codel_change()
161 u32 interval = nla_get_u32(tb[TCA_CODEL_INTERVAL]); in codel_change()
166 if (tb[TCA_CODEL_LIMIT]) in codel_change()
167 sch->limit = nla_get_u32(tb[TCA_CODEL_LIMIT]); in codel_change()
169 if (tb[TCA_CODEL_ECN]) in codel_change()
[all …]
H A Dsch_pie.c142 struct nlattr *tb[TCA_PIE_MAX + 1]; in pie_change() local
154 if (tb[TCA_PIE_TARGET]) { in pie_change()
156 u32 target = nla_get_u32(tb[TCA_PIE_TARGET]); in pie_change()
163 if (tb[TCA_PIE_TUPDATE]) in pie_change()
167 if (tb[TCA_PIE_LIMIT]) { in pie_change()
168 u32 limit = nla_get_u32(tb[TCA_PIE_LIMIT]); in pie_change()
174 if (tb[TCA_PIE_ALPHA]) in pie_change()
177 if (tb[TCA_PIE_BETA]) in pie_change()
180 if (tb[TCA_PIE_ECN]) in pie_change()
183 if (tb[TCA_PIE_BYTEMODE]) in pie_change()
[all …]
/openbmc/linux/net/core/
H A Dfib_rules.c529 if (!tb[FRA_SRC] || in fib_nl2rule()
537 if (!tb[FRA_DST] || in fib_nl2rule()
552 if (tb[FRA_PRIORITY]) { in fib_nl2rule()
562 if (tb[FRA_IIFNAME]) { in fib_nl2rule()
572 if (tb[FRA_OIFNAME]) { in fib_nl2rule()
582 if (tb[FRA_FWMARK]) { in fib_nl2rule()
591 if (tb[FRA_FWMASK]) in fib_nl2rule()
594 if (tb[FRA_TUN_ID]) in fib_nl2rule()
598 if (tb[FRA_L3MDEV] && in fib_nl2rule()
615 if (tb[FRA_GOTO]) { in fib_nl2rule()
[all …]
/openbmc/linux/drivers/net/wireless/ath/ath10k/
H A Dtestmode.c385 if (!tb[ATH10K_TM_ATTR_DATA]) { in ath10k_tm_cmd_wmi()
390 if (!tb[ATH10K_TM_ATTR_WMI_CMDID]) { in ath10k_tm_cmd_wmi()
395 buf = nla_data(tb[ATH10K_TM_ATTR_DATA]); in ath10k_tm_cmd_wmi()
396 buf_len = nla_len(tb[ATH10K_TM_ATTR_DATA]); in ath10k_tm_cmd_wmi()
431 struct nlattr *tb[ATH10K_TM_ATTR_MAX + 1]; in ath10k_tm_cmd() local
439 if (!tb[ATH10K_TM_ATTR_CMD]) in ath10k_tm_cmd()
442 switch (nla_get_u32(tb[ATH10K_TM_ATTR_CMD])) { in ath10k_tm_cmd()
444 return ath10k_tm_cmd_get_version(ar, tb); in ath10k_tm_cmd()
446 return ath10k_tm_cmd_utf_start(ar, tb); in ath10k_tm_cmd()
448 return ath10k_tm_cmd_utf_stop(ar, tb); in ath10k_tm_cmd()
[all …]
H A Dwmi-tlv.c147 tb = kcalloc(WMI_TLV_TAG_MAX, sizeof(*tb), gfp); in ath10k_wmi_tlv_parse_alloc()
148 if (!tb) in ath10k_wmi_tlv_parse_alloc()
153 kfree(tb); in ath10k_wmi_tlv_parse_alloc()
157 return tb; in ath10k_wmi_tlv_parse_alloc()
186 kfree(tb); in ath10k_wmi_tlv_event_bcn_tx_status()
211 kfree(tb); in ath10k_wmi_tlv_event_bcn_tx_status()
298 kfree(tb); in ath10k_wmi_tlv_op_pull_peer_stats_info()
369 kfree(tb); in ath10k_wmi_tlv_event_diag_data()
397 kfree(tb); in ath10k_wmi_tlv_event_diag()
431 kfree(tb); in ath10k_wmi_tlv_event_p2p_noa()
[all …]
/openbmc/linux/drivers/thunderbolt/
H A Dnhi.c967 struct tb *tb = pci_get_drvdata(pdev); in __nhi_suspend_noirq() local
992 struct tb *tb = pci_get_drvdata(pdev); in nhi_freeze_noirq() local
1000 struct tb *tb = pci_get_drvdata(pdev); in nhi_thaw_noirq() local
1047 struct tb *tb = pci_get_drvdata(pdev); in nhi_resume_noirq() local
1073 struct tb *tb = pci_get_drvdata(pdev); in nhi_suspend() local
1081 struct tb *tb = pci_get_drvdata(pdev); in nhi_complete() local
1097 struct tb *tb = pci_get_drvdata(pdev); in nhi_runtime_suspend() local
1116 struct tb *tb = pci_get_drvdata(pdev); in nhi_runtime_resume() local
1310 struct tb *tb; in nhi_select_cm() local
1335 struct tb *tb; in nhi_probe() local
[all …]
H A Dswitch.c443 static void tb_dump_port(struct tb *tb, const struct tb_port *port) in tb_dump_port() argument
447 tb_dbg(tb, in tb_dump_port()
1486 tb_dbg(tb, in tb_dump_switch()
2389 sw->tb = tb; in tb_switch_alloc()
2491 sw->tb = tb; in tb_switch_alloc_safe_mode()
2520 struct tb *tb = sw->tb; in tb_switch_configure() local
3446 struct tb *tb; member
3460 if (sw->tb != lookup->tb) in tb_switch_match()
3493 lookup.tb = tb; in tb_switch_find_by_link_depth()
3518 lookup.tb = tb; in tb_switch_find_by_uuid()
[all …]
/openbmc/openbmc/poky/bitbake/lib/bb/
H A Dexceptions.py63 def extract_traceback(tb, context=1): argument
64 frames = inspect.getinnerframes(tb, context)
82 def format_exception(etype, value, tb, context=1, limit=None, formatter=None): argument
85 if hasattr(tb, 'tb_next'):
86 tb = extract_traceback(tb, context)
88 formatted.extend(format_extracted(tb, formatter, limit))
/openbmc/linux/drivers/base/regmap/
H A Dregmap-spi-avmm.c253 tb = br->trans_buf; in br_pkt_phy_tx_prepare()
254 tb_end = tb + br->trans_len; in br_pkt_phy_tx_prepare()
269 *pb = *tb++ ^ 0x20; in br_pkt_phy_tx_prepare()
285 switch (*tb) { in br_pkt_phy_tx_prepare()
299 *pb = *tb++; in br_pkt_phy_tx_prepare()
305 if (tb < tb_end) in br_pkt_phy_tx_prepare()
371 while (tb < tb_limit) { in br_do_rx_and_pkt_phy_parse()
383 if (!tb && pb[i] != PKT_SOP) in br_do_rx_and_pkt_phy_parse()
412 tb = br->trans_buf; in br_do_rx_and_pkt_phy_parse()
446 *tb++ = pb[i] ^ 0x20; in br_do_rx_and_pkt_phy_parse()
[all …]
/openbmc/linux/arch/sparc/kernel/
H A Dirq_64.c1008 tb->cpu_mondo_qmask); in sun4v_register_mondo_queues()
1010 tb->dev_mondo_qmask); in sun4v_register_mondo_queues()
1012 tb->resum_qmask); in sun4v_register_mondo_queues()
1014 tb->nonresum_qmask); in sun4v_register_mondo_queues()
1059 tb->cpu_list_pa = __pa(page); in init_cpu_send_mondo_info()
1071 alloc_one_queue(&tb->cpu_mondo_pa, tb->cpu_mondo_qmask); in sun4v_init_mondo_queues()
1072 alloc_one_queue(&tb->dev_mondo_pa, tb->dev_mondo_qmask); in sun4v_init_mondo_queues()
1073 alloc_one_queue(&tb->resum_mondo_pa, tb->resum_qmask); in sun4v_init_mondo_queues()
1074 alloc_one_queue(&tb->resum_kernel_buf_pa, tb->resum_qmask); in sun4v_init_mondo_queues()
1075 alloc_one_queue(&tb->nonresum_mondo_pa, tb->nonresum_qmask); in sun4v_init_mondo_queues()
[all …]
/openbmc/linux/arch/m68k/coldfire/
H A Dintc-simr.c132 u16 pa, tb; in intc_irq_set_type() local
136 tb = 0x1; in intc_irq_set_type()
139 tb = 0x2; in intc_irq_set_type()
142 tb = 0x3; in intc_irq_set_type()
146 tb = 0; in intc_irq_set_type()
150 if (tb) in intc_irq_set_type()
155 pa = (pa & ~(0x3 << ebit)) | (tb << ebit); in intc_irq_set_type()
/openbmc/linux/include/net/
H A Drtnetlink.h83 struct net_device *(*alloc)(struct nlattr *tb[],
93 int (*validate)(struct nlattr *tb[],
99 struct nlattr *tb[],
103 struct nlattr *tb[],
123 struct nlattr *tb[],
183 struct net *rtnl_link_get_net(struct net *src_net, struct nlattr *tb[]);
187 struct nlattr *tb[],
193 int rtnl_nla_parse_ifinfomsg(struct nlattr **tb, const struct nlattr *nla_peer,
/openbmc/linux/net/dcb/
H A Ddcbnl.c320 if (!tb[DCB_ATTR_CAP]) in dcbnl_getcap()
365 if (!tb[DCB_ATTR_NUMTCS]) in dcbnl_getnumtcs()
411 if (!tb[DCB_ATTR_NUMTCS]) in dcbnl_setnumtcs()
474 if (!tb[DCB_ATTR_APP]) in dcbnl_getapp()
543 if (!tb[DCB_ATTR_APP]) in dcbnl_setapp()
738 if (!tb[DCB_ATTR_STATE]) in dcbnl_setstate()
915 if (!tb[DCB_ATTR_BCN]) in dcbnl_bcn_getcfg()
975 if (!tb[DCB_ATTR_BCN]) in dcbnl_bcn_setcfg()
1571 if (!tb[DCB_ATTR_IEEE]) in dcbnl_ieee_set()
1717 if (!tb[DCB_ATTR_IEEE]) in dcbnl_ieee_del()
[all …]
/openbmc/linux/net/ipv6/
H A Dioam6_iptunnel.c107 struct nlattr *tb[IOAM6_IPTUNNEL_MAX + 1]; in ioam6_build_state() local
124 if ((!tb[IOAM6_IPTUNNEL_FREQ_K] && tb[IOAM6_IPTUNNEL_FREQ_N]) || in ioam6_build_state()
125 (tb[IOAM6_IPTUNNEL_FREQ_K] && !tb[IOAM6_IPTUNNEL_FREQ_N])) { in ioam6_build_state()
128 } else if (!tb[IOAM6_IPTUNNEL_FREQ_K] && !tb[IOAM6_IPTUNNEL_FREQ_N]) { in ioam6_build_state()
132 freq_k = nla_get_u32(tb[IOAM6_IPTUNNEL_FREQ_K]); in ioam6_build_state()
133 freq_n = nla_get_u32(tb[IOAM6_IPTUNNEL_FREQ_N]); in ioam6_build_state()
141 if (!tb[IOAM6_IPTUNNEL_MODE]) in ioam6_build_state()
144 mode = nla_get_u8(tb[IOAM6_IPTUNNEL_MODE]); in ioam6_build_state()
151 if (!tb[IOAM6_IPTUNNEL_TRACE]) { in ioam6_build_state()
156 trace = nla_data(tb[IOAM6_IPTUNNEL_TRACE]); in ioam6_build_state()
[all …]

12345678910>>...26