/openbmc/linux/net/netfilter/ |
H A D | nft_log.c | 147 const struct nlattr * const tb[]) in nft_log_init() argument 155 if (tb[NFTA_LOG_LEVEL] != NULL && in nft_log_init() 156 tb[NFTA_LOG_GROUP] != NULL) in nft_log_init() 158 if (tb[NFTA_LOG_GROUP] != NULL) { in nft_log_init() 160 if (tb[NFTA_LOG_FLAGS] != NULL) in nft_log_init() 164 nla = tb[NFTA_LOG_PREFIX]; in nft_log_init() 176 if (tb[NFTA_LOG_LEVEL] != NULL) { in nft_log_init() 178 ntohl(nla_get_be32(tb[NFTA_LOG_LEVEL])); in nft_log_init() 187 if (tb[NFTA_LOG_FLAGS] != NULL) { in nft_log_init() 198 if (tb[NFTA_LOG_SNAPLEN] != NULL) { in nft_log_init() [all …]
|
H A D | nft_range.c | 51 const struct nlattr * const tb[]) in nft_range_init() argument 65 if (!tb[NFTA_RANGE_SREG] || in nft_range_init() 66 !tb[NFTA_RANGE_OP] || in nft_range_init() 67 !tb[NFTA_RANGE_FROM_DATA] || in nft_range_init() 68 !tb[NFTA_RANGE_TO_DATA]) in nft_range_init() 72 tb[NFTA_RANGE_FROM_DATA]); in nft_range_init() 77 tb[NFTA_RANGE_TO_DATA]); in nft_range_init() 86 err = nft_parse_register_load(tb[NFTA_RANGE_SREG], &priv->sreg, in nft_range_init() 91 err = nft_parse_u32_check(tb[NFTA_RANGE_OP], U8_MAX, &op); in nft_range_init()
|
H A D | nft_dynset.c | 162 const struct nlattr * const tb[]) in nft_dynset_init() argument 173 if (tb[NFTA_DYNSET_SET_NAME] == NULL || in nft_dynset_init() 174 tb[NFTA_DYNSET_OP] == NULL || in nft_dynset_init() 175 tb[NFTA_DYNSET_SREG_KEY] == NULL) in nft_dynset_init() 178 if (tb[NFTA_DYNSET_FLAGS]) { in nft_dynset_init() 189 tb[NFTA_DYNSET_SET_NAME], in nft_dynset_init() 190 tb[NFTA_DYNSET_SET_ID], genmask); in nft_dynset_init() 208 if (tb[NFTA_DYNSET_TIMEOUT] != NULL) { in nft_dynset_init() 235 if ((tb[NFTA_DYNSET_EXPR] || tb[NFTA_DYNSET_EXPRESSIONS]) && in nft_dynset_init() 239 if (tb[NFTA_DYNSET_EXPR]) { in nft_dynset_init() [all …]
|
H A D | nft_compat.c | 212 if (!tb[NFTA_RULE_COMPAT_PROTO] || !tb[NFTA_RULE_COMPAT_FLAGS]) in nft_parse_compat() 244 const struct nlattr * const tb[]) in nft_target_init() argument 498 const struct nlattr * const tb[], in __nft_match_init() argument 691 if (tb[NFTA_COMPAT_NAME] == NULL || in nfnl_compat_get_rcu() 692 tb[NFTA_COMPAT_REV] == NULL || in nfnl_compat_get_rcu() 693 tb[NFTA_COMPAT_TYPE] == NULL) in nfnl_compat_get_rcu() 795 if (tb[NFTA_MATCH_NAME] == NULL || in nft_match_select_ops() 796 tb[NFTA_MATCH_REV] == NULL || in nft_match_select_ops() 797 tb[NFTA_MATCH_INFO] == NULL) in nft_match_select_ops() 876 tb[NFTA_TARGET_REV] == NULL || in nft_target_select_ops() [all …]
|
H A D | nft_inner.c | 288 struct nlattr *tb[NFT_EXPR_MAXATTR + 1]; member 293 const struct nlattr * const tb[]) in nft_inner_init() argument 300 if (!tb[NFTA_INNER_FLAGS] || in nft_inner_init() 301 !tb[NFTA_INNER_NUM] || in nft_inner_init() 302 !tb[NFTA_INNER_HDRSIZE] || in nft_inner_init() 303 !tb[NFTA_INNER_TYPE] || in nft_inner_init() 304 !tb[NFTA_INNER_EXPR]) in nft_inner_init() 307 flags = ntohl(nla_get_be32(tb[NFTA_INNER_FLAGS])); in nft_inner_init() 311 num = ntohl(nla_get_be32(tb[NFTA_INNER_NUM])); in nft_inner_init() 316 type = ntohl(nla_get_be32(tb[NFTA_INNER_TYPE])); in nft_inner_init() [all …]
|
H A D | nft_payload.c | 653 if (!tb[NFTA_PAYLOAD_BASE] || !tb[NFTA_PAYLOAD_OFFSET] || in nft_payload_inner_init() 654 !tb[NFTA_PAYLOAD_LEN] || !tb[NFTA_PAYLOAD_DREG]) in nft_payload_inner_init() 938 const struct nlattr * const tb[]) in nft_payload_set_init() argument 948 if (tb[NFTA_PAYLOAD_CSUM_TYPE]) in nft_payload_set_init() 950 if (tb[NFTA_PAYLOAD_CSUM_OFFSET]) { in nft_payload_set_init() 958 if (tb[NFTA_PAYLOAD_CSUM_FLAGS]) { in nft_payload_set_init() 1044 if (tb[NFTA_PAYLOAD_BASE] == NULL || in nft_payload_select_ops() 1046 tb[NFTA_PAYLOAD_LEN] == NULL) in nft_payload_select_ops() 1060 if (tb[NFTA_PAYLOAD_SREG] != NULL) { in nft_payload_select_ops() 1061 if (tb[NFTA_PAYLOAD_DREG] != NULL) in nft_payload_select_ops() [all …]
|
/openbmc/linux/net/sched/ |
H A D | cls_flow.c | 420 if (tb[TCA_FLOW_KEYS]) { in flow_change() 473 if (tb[TCA_FLOW_MODE]) in flow_change() 480 if (tb[TCA_FLOW_PERTURB]) { in flow_change() 489 if (!tb[TCA_FLOW_KEYS]) in flow_change() 493 if (tb[TCA_FLOW_MODE]) in flow_change() 522 if (tb[TCA_FLOW_KEYS]) { in flow_change() 529 if (tb[TCA_FLOW_MASK]) in flow_change() 531 if (tb[TCA_FLOW_XOR]) in flow_change() 533 if (tb[TCA_FLOW_RSHIFT]) in flow_change() 535 if (tb[TCA_FLOW_ADDEND]) in flow_change() [all …]
|
H A D | act_bpf.c | 195 bpf_num_ops = nla_get_u16(tb[TCA_ACT_BPF_OPS_LEN]); in tcf_bpf_init_from_ops() 200 if (bpf_size != nla_len(tb[TCA_ACT_BPF_OPS])) in tcf_bpf_init_from_ops() 230 bpf_fd = nla_get_u32(tb[TCA_ACT_BPF_FD]); in tcf_bpf_init_from_efd() 236 if (tb[TCA_ACT_BPF_NAME]) { in tcf_bpf_init_from_efd() 286 struct nlattr *tb[TCA_ACT_BPF_MAX + 1]; in tcf_bpf_init() local 303 if (!tb[TCA_ACT_BPF_PARMS]) in tcf_bpf_init() 306 parm = nla_data(tb[TCA_ACT_BPF_PARMS]); in tcf_bpf_init() 335 is_bpf = tb[TCA_ACT_BPF_OPS_LEN] && tb[TCA_ACT_BPF_OPS]; in tcf_bpf_init() 336 is_ebpf = tb[TCA_ACT_BPF_FD]; in tcf_bpf_init() 345 ret = is_bpf ? tcf_bpf_init_from_ops(tb, &cfg) : in tcf_bpf_init() [all …]
|
H A D | act_police.c | 47 struct nlattr *tb[TCA_POLICE_MAX + 1]; in tcf_police_init() local 67 if (tb[TCA_POLICE_TBF] == NULL) in tcf_police_init() 69 size = nla_len(tb[TCA_POLICE_TBF]); in tcf_police_init() 73 parm = nla_data(tb[TCA_POLICE_TBF]); in tcf_police_init() 122 } else if (tb[TCA_POLICE_AVRATE] && in tcf_police_init() 129 if (tb[TCA_POLICE_RESULT]) { in tcf_police_init() 139 if ((tb[TCA_POLICE_PKTRATE64] && !tb[TCA_POLICE_PKTBURST64]) || in tcf_police_init() 140 (!tb[TCA_POLICE_PKTRATE64] && tb[TCA_POLICE_PKTBURST64])) { in tcf_police_init() 170 rate64 = tb[TCA_POLICE_RATE64] ? in tcf_police_init() 192 if (tb[TCA_POLICE_AVRATE]) in tcf_police_init() [all …]
|
H A D | sch_gred.c | 548 if (tb[TCA_GRED_VQ_FLAGS]) in gred_vq_apply() 579 if (!tb[TCA_GRED_VQ_DP]) { in gred_vq_validate() 593 if (tb[TCA_GRED_VQ_FLAGS]) { in gred_vq_validate() 658 if (tb[TCA_GRED_PARMS] == NULL && tb[TCA_GRED_STAB] == NULL) { in gred_change() 659 if (tb[TCA_GRED_LIMIT] != NULL) in gred_change() 665 tb[TCA_GRED_STAB] == NULL || in gred_change() 671 max_P = tb[TCA_GRED_MAX_P] ? nla_get_u32(tb[TCA_GRED_MAX_P]) : 0; in gred_change() 681 if (tb[TCA_GRED_VQ_LIST]) { in gred_change() 711 if (tb[TCA_GRED_VQ_LIST]) in gred_change() 747 if (tb[TCA_GRED_PARMS] || tb[TCA_GRED_STAB]) { in gred_init() [all …]
|
H A D | act_skbedit.c | 125 struct nlattr *tb[TCA_SKBEDIT_MAX + 1]; in tcf_skbedit_init() local 144 if (tb[TCA_SKBEDIT_PARMS] == NULL) in tcf_skbedit_init() 147 if (tb[TCA_SKBEDIT_PRIORITY] != NULL) { in tcf_skbedit_init() 162 if (tb[TCA_SKBEDIT_PTYPE] != NULL) { in tcf_skbedit_init() 169 if (tb[TCA_SKBEDIT_MARK] != NULL) { in tcf_skbedit_init() 171 mark = nla_data(tb[TCA_SKBEDIT_MARK]); in tcf_skbedit_init() 174 if (tb[TCA_SKBEDIT_MASK] != NULL) { in tcf_skbedit_init() 176 mask = nla_data(tb[TCA_SKBEDIT_MASK]); in tcf_skbedit_init() 179 if (tb[TCA_SKBEDIT_FLAGS] != NULL) { in tcf_skbedit_init() 185 if (!tb[TCA_SKBEDIT_QUEUE_MAPPING] || in tcf_skbedit_init() [all …]
|
H A D | sch_codel.c | 137 struct nlattr *tb[TCA_CODEL_MAX + 1]; in codel_change() local 148 if (tb[TCA_CODEL_TARGET]) { in codel_change() 149 u32 target = nla_get_u32(tb[TCA_CODEL_TARGET]); in codel_change() 154 if (tb[TCA_CODEL_CE_THRESHOLD]) { in codel_change() 155 u64 val = nla_get_u32(tb[TCA_CODEL_CE_THRESHOLD]); in codel_change() 160 if (tb[TCA_CODEL_INTERVAL]) { in codel_change() 161 u32 interval = nla_get_u32(tb[TCA_CODEL_INTERVAL]); in codel_change() 166 if (tb[TCA_CODEL_LIMIT]) in codel_change() 167 sch->limit = nla_get_u32(tb[TCA_CODEL_LIMIT]); in codel_change() 169 if (tb[TCA_CODEL_ECN]) in codel_change() [all …]
|
H A D | sch_pie.c | 142 struct nlattr *tb[TCA_PIE_MAX + 1]; in pie_change() local 154 if (tb[TCA_PIE_TARGET]) { in pie_change() 156 u32 target = nla_get_u32(tb[TCA_PIE_TARGET]); in pie_change() 163 if (tb[TCA_PIE_TUPDATE]) in pie_change() 167 if (tb[TCA_PIE_LIMIT]) { in pie_change() 168 u32 limit = nla_get_u32(tb[TCA_PIE_LIMIT]); in pie_change() 174 if (tb[TCA_PIE_ALPHA]) in pie_change() 177 if (tb[TCA_PIE_BETA]) in pie_change() 180 if (tb[TCA_PIE_ECN]) in pie_change() 183 if (tb[TCA_PIE_BYTEMODE]) in pie_change() [all …]
|
/openbmc/linux/net/core/ |
H A D | fib_rules.c | 529 if (!tb[FRA_SRC] || in fib_nl2rule() 537 if (!tb[FRA_DST] || in fib_nl2rule() 552 if (tb[FRA_PRIORITY]) { in fib_nl2rule() 562 if (tb[FRA_IIFNAME]) { in fib_nl2rule() 572 if (tb[FRA_OIFNAME]) { in fib_nl2rule() 582 if (tb[FRA_FWMARK]) { in fib_nl2rule() 591 if (tb[FRA_FWMASK]) in fib_nl2rule() 594 if (tb[FRA_TUN_ID]) in fib_nl2rule() 598 if (tb[FRA_L3MDEV] && in fib_nl2rule() 615 if (tb[FRA_GOTO]) { in fib_nl2rule() [all …]
|
/openbmc/linux/drivers/net/wireless/ath/ath10k/ |
H A D | testmode.c | 385 if (!tb[ATH10K_TM_ATTR_DATA]) { in ath10k_tm_cmd_wmi() 390 if (!tb[ATH10K_TM_ATTR_WMI_CMDID]) { in ath10k_tm_cmd_wmi() 395 buf = nla_data(tb[ATH10K_TM_ATTR_DATA]); in ath10k_tm_cmd_wmi() 396 buf_len = nla_len(tb[ATH10K_TM_ATTR_DATA]); in ath10k_tm_cmd_wmi() 431 struct nlattr *tb[ATH10K_TM_ATTR_MAX + 1]; in ath10k_tm_cmd() local 439 if (!tb[ATH10K_TM_ATTR_CMD]) in ath10k_tm_cmd() 442 switch (nla_get_u32(tb[ATH10K_TM_ATTR_CMD])) { in ath10k_tm_cmd() 444 return ath10k_tm_cmd_get_version(ar, tb); in ath10k_tm_cmd() 446 return ath10k_tm_cmd_utf_start(ar, tb); in ath10k_tm_cmd() 448 return ath10k_tm_cmd_utf_stop(ar, tb); in ath10k_tm_cmd() [all …]
|
H A D | wmi-tlv.c | 147 tb = kcalloc(WMI_TLV_TAG_MAX, sizeof(*tb), gfp); in ath10k_wmi_tlv_parse_alloc() 148 if (!tb) in ath10k_wmi_tlv_parse_alloc() 153 kfree(tb); in ath10k_wmi_tlv_parse_alloc() 157 return tb; in ath10k_wmi_tlv_parse_alloc() 186 kfree(tb); in ath10k_wmi_tlv_event_bcn_tx_status() 211 kfree(tb); in ath10k_wmi_tlv_event_bcn_tx_status() 298 kfree(tb); in ath10k_wmi_tlv_op_pull_peer_stats_info() 369 kfree(tb); in ath10k_wmi_tlv_event_diag_data() 397 kfree(tb); in ath10k_wmi_tlv_event_diag() 431 kfree(tb); in ath10k_wmi_tlv_event_p2p_noa() [all …]
|
/openbmc/linux/drivers/thunderbolt/ |
H A D | nhi.c | 967 struct tb *tb = pci_get_drvdata(pdev); in __nhi_suspend_noirq() local 992 struct tb *tb = pci_get_drvdata(pdev); in nhi_freeze_noirq() local 1000 struct tb *tb = pci_get_drvdata(pdev); in nhi_thaw_noirq() local 1047 struct tb *tb = pci_get_drvdata(pdev); in nhi_resume_noirq() local 1073 struct tb *tb = pci_get_drvdata(pdev); in nhi_suspend() local 1081 struct tb *tb = pci_get_drvdata(pdev); in nhi_complete() local 1097 struct tb *tb = pci_get_drvdata(pdev); in nhi_runtime_suspend() local 1116 struct tb *tb = pci_get_drvdata(pdev); in nhi_runtime_resume() local 1310 struct tb *tb; in nhi_select_cm() local 1335 struct tb *tb; in nhi_probe() local [all …]
|
H A D | switch.c | 443 static void tb_dump_port(struct tb *tb, const struct tb_port *port) in tb_dump_port() argument 447 tb_dbg(tb, in tb_dump_port() 1486 tb_dbg(tb, in tb_dump_switch() 2389 sw->tb = tb; in tb_switch_alloc() 2491 sw->tb = tb; in tb_switch_alloc_safe_mode() 2520 struct tb *tb = sw->tb; in tb_switch_configure() local 3446 struct tb *tb; member 3460 if (sw->tb != lookup->tb) in tb_switch_match() 3493 lookup.tb = tb; in tb_switch_find_by_link_depth() 3518 lookup.tb = tb; in tb_switch_find_by_uuid() [all …]
|
/openbmc/openbmc/poky/bitbake/lib/bb/ |
H A D | exceptions.py | 63 def extract_traceback(tb, context=1): argument 64 frames = inspect.getinnerframes(tb, context) 82 def format_exception(etype, value, tb, context=1, limit=None, formatter=None): argument 85 if hasattr(tb, 'tb_next'): 86 tb = extract_traceback(tb, context) 88 formatted.extend(format_extracted(tb, formatter, limit))
|
/openbmc/linux/drivers/base/regmap/ |
H A D | regmap-spi-avmm.c | 253 tb = br->trans_buf; in br_pkt_phy_tx_prepare() 254 tb_end = tb + br->trans_len; in br_pkt_phy_tx_prepare() 269 *pb = *tb++ ^ 0x20; in br_pkt_phy_tx_prepare() 285 switch (*tb) { in br_pkt_phy_tx_prepare() 299 *pb = *tb++; in br_pkt_phy_tx_prepare() 305 if (tb < tb_end) in br_pkt_phy_tx_prepare() 371 while (tb < tb_limit) { in br_do_rx_and_pkt_phy_parse() 383 if (!tb && pb[i] != PKT_SOP) in br_do_rx_and_pkt_phy_parse() 412 tb = br->trans_buf; in br_do_rx_and_pkt_phy_parse() 446 *tb++ = pb[i] ^ 0x20; in br_do_rx_and_pkt_phy_parse() [all …]
|
/openbmc/linux/arch/sparc/kernel/ |
H A D | irq_64.c | 1008 tb->cpu_mondo_qmask); in sun4v_register_mondo_queues() 1010 tb->dev_mondo_qmask); in sun4v_register_mondo_queues() 1012 tb->resum_qmask); in sun4v_register_mondo_queues() 1014 tb->nonresum_qmask); in sun4v_register_mondo_queues() 1059 tb->cpu_list_pa = __pa(page); in init_cpu_send_mondo_info() 1071 alloc_one_queue(&tb->cpu_mondo_pa, tb->cpu_mondo_qmask); in sun4v_init_mondo_queues() 1072 alloc_one_queue(&tb->dev_mondo_pa, tb->dev_mondo_qmask); in sun4v_init_mondo_queues() 1073 alloc_one_queue(&tb->resum_mondo_pa, tb->resum_qmask); in sun4v_init_mondo_queues() 1074 alloc_one_queue(&tb->resum_kernel_buf_pa, tb->resum_qmask); in sun4v_init_mondo_queues() 1075 alloc_one_queue(&tb->nonresum_mondo_pa, tb->nonresum_qmask); in sun4v_init_mondo_queues() [all …]
|
/openbmc/linux/arch/m68k/coldfire/ |
H A D | intc-simr.c | 132 u16 pa, tb; in intc_irq_set_type() local 136 tb = 0x1; in intc_irq_set_type() 139 tb = 0x2; in intc_irq_set_type() 142 tb = 0x3; in intc_irq_set_type() 146 tb = 0; in intc_irq_set_type() 150 if (tb) in intc_irq_set_type() 155 pa = (pa & ~(0x3 << ebit)) | (tb << ebit); in intc_irq_set_type()
|
/openbmc/linux/include/net/ |
H A D | rtnetlink.h | 83 struct net_device *(*alloc)(struct nlattr *tb[], 93 int (*validate)(struct nlattr *tb[], 99 struct nlattr *tb[], 103 struct nlattr *tb[], 123 struct nlattr *tb[], 183 struct net *rtnl_link_get_net(struct net *src_net, struct nlattr *tb[]); 187 struct nlattr *tb[], 193 int rtnl_nla_parse_ifinfomsg(struct nlattr **tb, const struct nlattr *nla_peer,
|
/openbmc/linux/net/dcb/ |
H A D | dcbnl.c | 320 if (!tb[DCB_ATTR_CAP]) in dcbnl_getcap() 365 if (!tb[DCB_ATTR_NUMTCS]) in dcbnl_getnumtcs() 411 if (!tb[DCB_ATTR_NUMTCS]) in dcbnl_setnumtcs() 474 if (!tb[DCB_ATTR_APP]) in dcbnl_getapp() 543 if (!tb[DCB_ATTR_APP]) in dcbnl_setapp() 738 if (!tb[DCB_ATTR_STATE]) in dcbnl_setstate() 915 if (!tb[DCB_ATTR_BCN]) in dcbnl_bcn_getcfg() 975 if (!tb[DCB_ATTR_BCN]) in dcbnl_bcn_setcfg() 1571 if (!tb[DCB_ATTR_IEEE]) in dcbnl_ieee_set() 1717 if (!tb[DCB_ATTR_IEEE]) in dcbnl_ieee_del() [all …]
|
/openbmc/linux/net/ipv6/ |
H A D | ioam6_iptunnel.c | 107 struct nlattr *tb[IOAM6_IPTUNNEL_MAX + 1]; in ioam6_build_state() local 124 if ((!tb[IOAM6_IPTUNNEL_FREQ_K] && tb[IOAM6_IPTUNNEL_FREQ_N]) || in ioam6_build_state() 125 (tb[IOAM6_IPTUNNEL_FREQ_K] && !tb[IOAM6_IPTUNNEL_FREQ_N])) { in ioam6_build_state() 128 } else if (!tb[IOAM6_IPTUNNEL_FREQ_K] && !tb[IOAM6_IPTUNNEL_FREQ_N]) { in ioam6_build_state() 132 freq_k = nla_get_u32(tb[IOAM6_IPTUNNEL_FREQ_K]); in ioam6_build_state() 133 freq_n = nla_get_u32(tb[IOAM6_IPTUNNEL_FREQ_N]); in ioam6_build_state() 141 if (!tb[IOAM6_IPTUNNEL_MODE]) in ioam6_build_state() 144 mode = nla_get_u8(tb[IOAM6_IPTUNNEL_MODE]); in ioam6_build_state() 151 if (!tb[IOAM6_IPTUNNEL_TRACE]) { in ioam6_build_state() 156 trace = nla_data(tb[IOAM6_IPTUNNEL_TRACE]); in ioam6_build_state() [all …]
|