/openbmc/linux/net/sched/ |
H A D | sch_red.c | 247 if (tb[TCA_RED_PARMS] == NULL || in __red_change() 248 tb[TCA_RED_STAB] == NULL) in __red_change() 251 max_P = tb[TCA_RED_MAX_P] ? nla_get_u32(tb[TCA_RED_MAX_P]) : 0; in __red_change() 253 ctl = nla_data(tb[TCA_RED_PARMS]); in __red_change() 254 stab = nla_data(tb[TCA_RED_STAB]); in __red_change() 339 struct nlattr *tb[TCA_RED_MAX + 1]; in red_init() local 354 err = __red_change(sch, tb, extack); in red_init() 366 tb[TCA_RED_MARK_BLOCK], extack); in red_init() 373 struct nlattr *tb[TCA_RED_MAX + 1]; in red_change() local 387 tb[TCA_RED_MARK_BLOCK], extack); in red_change() [all …]
|
H A D | cls_fw.c | 202 struct fw_filter *f, struct nlattr **tb, in fw_set_parms() argument 215 if (tb[TCA_FW_INDEV]) { in fw_set_parms() 217 ret = tcf_change_indev(net, tb[TCA_FW_INDEV], extack); in fw_set_parms() 224 if (tb[TCA_FW_MASK]) { in fw_set_parms() 225 mask = nla_get_u32(tb[TCA_FW_MASK]); in fw_set_parms() 231 if (tb[TCA_FW_CLASSID]) { in fw_set_parms() 232 f->res.classid = nla_get_u32(tb[TCA_FW_CLASSID]); in fw_set_parms() 247 struct nlattr *tb[TCA_FW_MAX + 1]; in fw_change() local 308 if (tb[TCA_FW_MASK]) in fw_change() 309 mask = nla_get_u32(tb[TCA_FW_MASK]); in fw_change() [all …]
|
H A D | act_skbmod.c | 110 struct nlattr *tb[TCA_SKBMOD_MAX + 1]; in tcf_skbmod_init() local 125 err = nla_parse_nested_deprecated(tb, TCA_SKBMOD_MAX, nla, in tcf_skbmod_init() 130 if (!tb[TCA_SKBMOD_PARMS]) in tcf_skbmod_init() 133 if (tb[TCA_SKBMOD_DMAC]) { in tcf_skbmod_init() 134 daddr = nla_data(tb[TCA_SKBMOD_DMAC]); in tcf_skbmod_init() 138 if (tb[TCA_SKBMOD_SMAC]) { in tcf_skbmod_init() 139 saddr = nla_data(tb[TCA_SKBMOD_SMAC]); in tcf_skbmod_init() 143 if (tb[TCA_SKBMOD_ETYPE]) { in tcf_skbmod_init() 144 eth_type = nla_get_u16(tb[TCA_SKBMOD_ETYPE]); in tcf_skbmod_init() 148 parm = nla_data(tb[TCA_SKBMOD_PARMS]); in tcf_skbmod_init()
|
H A D | cls_route.c | 399 if (tb[TCA_ROUTE4_TO]) { in route4_set_parms() 402 to = nla_get_u32(tb[TCA_ROUTE4_TO]); in route4_set_parms() 408 if (tb[TCA_ROUTE4_FROM]) { in route4_set_parms() 409 if (tb[TCA_ROUTE4_IIF]) in route4_set_parms() 415 } else if (tb[TCA_ROUTE4_IIF]) { in route4_set_parms() 416 id = nla_get_u32(tb[TCA_ROUTE4_IIF]); in route4_set_parms() 452 if (tb[TCA_ROUTE4_TO]) in route4_set_parms() 455 if (tb[TCA_ROUTE4_FROM]) in route4_set_parms() 457 else if (tb[TCA_ROUTE4_IIF]) in route4_set_parms() 464 if (tb[TCA_ROUTE4_CLASSID]) { in route4_set_parms() [all …]
|
H A D | sch_netem.c | 1000 if (tb[TCA_NETEM_LOSS]) { in netem_change() 1031 if (tb[TCA_NETEM_CORR]) in netem_change() 1034 if (tb[TCA_NETEM_REORDER]) in netem_change() 1037 if (tb[TCA_NETEM_CORRUPT]) in netem_change() 1040 if (tb[TCA_NETEM_RATE]) in netem_change() 1043 if (tb[TCA_NETEM_RATE64]) in netem_change() 1047 if (tb[TCA_NETEM_LATENCY64]) in netem_change() 1050 if (tb[TCA_NETEM_JITTER64]) in netem_change() 1053 if (tb[TCA_NETEM_ECN]) in netem_change() 1056 if (tb[TCA_NETEM_SLOT]) in netem_change() [all …]
|
H A D | sch_tbf.c | 355 struct nlattr *tb[TCA_TBF_MAX + 1]; in tbf_change() local 371 if (tb[TCA_TBF_PARMS] == NULL) in tbf_change() 374 qopt = nla_data(tb[TCA_TBF_PARMS]); in tbf_change() 377 tb[TCA_TBF_RTAB], in tbf_change() 382 tb[TCA_TBF_PTAB], in tbf_change() 388 if (tb[TCA_TBF_RATE64]) in tbf_change() 392 if (tb[TCA_TBF_BURST]) { in tbf_change() 400 if (tb[TCA_TBF_PRATE64]) in tbf_change() 410 if (tb[TCA_TBF_PBURST]) { in tbf_change() 454 if (tb[TCA_TBF_PBURST]) in tbf_change() [all …]
|
/openbmc/linux/net/ipv4/netfilter/ |
H A D | nft_dup_ipv4.c | 35 const struct nlattr * const tb[]) in nft_dup_ipv4_init() argument 40 if (tb[NFTA_DUP_SREG_ADDR] == NULL) in nft_dup_ipv4_init() 43 err = nft_parse_register_load(tb[NFTA_DUP_SREG_ADDR], &priv->sreg_addr, in nft_dup_ipv4_init() 48 if (tb[NFTA_DUP_SREG_DEV]) in nft_dup_ipv4_init() 49 err = nft_parse_register_load(tb[NFTA_DUP_SREG_DEV], in nft_dup_ipv4_init()
|
/openbmc/linux/net/ipv6/netfilter/ |
H A D | nft_dup_ipv6.c | 33 const struct nlattr * const tb[]) in nft_dup_ipv6_init() argument 38 if (tb[NFTA_DUP_SREG_ADDR] == NULL) in nft_dup_ipv6_init() 41 err = nft_parse_register_load(tb[NFTA_DUP_SREG_ADDR], &priv->sreg_addr, in nft_dup_ipv6_init() 46 if (tb[NFTA_DUP_SREG_DEV]) in nft_dup_ipv6_init() 47 err = nft_parse_register_load(tb[NFTA_DUP_SREG_DEV], in nft_dup_ipv6_init()
|
/openbmc/linux/arch/powerpc/boot/ |
H A D | 4xx.c | 359 tb = sys_clk; in ibm440gp_fixup_clocks() 362 tb = cpu; in ibm440gp_fixup_clocks() 381 dt_fixup_cpu_clocks(cpu, tb, 0); in ibm440gp_fixup_clocks() 422 u32 ccr1, tb = tmr_clk; in __ibm440eplike_fixup_clocks() local 465 if (tb == 0) { in __ibm440eplike_fixup_clocks() 470 tb = cpu; in __ibm440eplike_fixup_clocks() 472 dt_fixup_cpu_clocks(cpu, tb, 0); in __ibm440eplike_fixup_clocks() 610 tb = cpu; in ibm405gp_fixup_clocks() 628 u32 pllmr0_ccdv, tb, m; in ibm405ep_fixup_clocks() local 651 tb = cpu; in ibm405ep_fixup_clocks() [all …]
|
/openbmc/linux/arch/sparc/mm/ |
H A D | tsb.c | 88 for (i = 0; i < tb->tlb_nr; i++) in __flush_tsb_one() 113 for (i = 0; i < tb->tlb_nr; i++) in __flush_huge_tsb_one() 114 __flush_huge_tsb_one_entry(tsb, tb->vaddrs[i], hash_shift, in __flush_huge_tsb_one() 119 void flush_tsb_user(struct tlb_batch *tb) in flush_tsb_user() argument 121 struct mm_struct *mm = tb->mm; in flush_tsb_user() 126 if (tb->hugepage_shift < REAL_HPAGE_SHIFT) { in flush_tsb_user() 131 if (tb->hugepage_shift == PAGE_SHIFT) in flush_tsb_user() 132 __flush_tsb_one(tb, PAGE_SHIFT, base, nentries); in flush_tsb_user() 135 __flush_huge_tsb_one(tb, PAGE_SHIFT, base, nentries, in flush_tsb_user() 136 tb->hugepage_shift); in flush_tsb_user() [all …]
|
/openbmc/qemu/accel/tcg/ |
H A D | internal-target.h | 57 static inline void tb_unlock_pages(TranslationBlock *tb) { } in tb_unlock_pages() argument 75 static inline vaddr log_pc(CPUState *cpu, const TranslationBlock *tb) in log_pc() argument 77 if (tb_cflags(tb) & CF_PCREL) { in log_pc() 80 return tb->pc; in log_pc()
|
H A D | monitor.c | 108 const TranslationBlock *tb = value; in tb_tree_stats_iter() local 112 tst->host_size += tb->tc.size; in tb_tree_stats_iter() 113 tst->target_size += tb->size; in tb_tree_stats_iter() 114 if (tb->size > tst->max_target_size) { in tb_tree_stats_iter() 115 tst->max_target_size = tb->size; in tb_tree_stats_iter() 117 if (tb->page_addr[1] != -1) { in tb_tree_stats_iter() 120 if (tb->jmp_reset_offset[0] != TB_JMP_OFFSET_INVALID) { in tb_tree_stats_iter() 122 if (tb->jmp_reset_offset[1] != TB_JMP_OFFSET_INVALID) { in tb_tree_stats_iter()
|
/openbmc/qemu/plugins/ |
H A D | api.c | 98 void qemu_plugin_register_vcpu_tb_exec_cb(struct qemu_plugin_tb *tb, in qemu_plugin_register_vcpu_tb_exec_cb() argument 104 plugin_register_dyn_cb__udata(&tb->cbs, cb, flags, udata); in qemu_plugin_register_vcpu_tb_exec_cb() 120 qemu_plugin_register_vcpu_tb_exec_cb(tb, cb, flags, udata); in qemu_plugin_register_vcpu_tb_exec_cond_cb() 123 plugin_register_dyn_cond_cb__udata(&tb->cbs, cb, flags, in qemu_plugin_register_vcpu_tb_exec_cond_cb() 128 struct qemu_plugin_tb *tb, in qemu_plugin_register_vcpu_tb_exec_inline_per_vcpu() argument 238 size_t qemu_plugin_tb_n_insns(const struct qemu_plugin_tb *tb) in qemu_plugin_tb_n_insns() argument 240 return tb->n; in qemu_plugin_tb_n_insns() 243 uint64_t qemu_plugin_tb_vaddr(const struct qemu_plugin_tb *tb) in qemu_plugin_tb_vaddr() argument 250 qemu_plugin_tb_get_insn(const struct qemu_plugin_tb *tb, size_t idx) in qemu_plugin_tb_get_insn() argument 253 if (unlikely(idx >= tb->n)) { in qemu_plugin_tb_get_insn() [all …]
|
/openbmc/linux/net/netfilter/ipset/ |
H A D | ip_set_list_set.c | 360 if (tb[IPSET_ATTR_LINENO]) in list_set_uadt() 361 *lineno = nla_get_u32(tb[IPSET_ATTR_LINENO]); in list_set_uadt() 363 if (unlikely(!tb[IPSET_ATTR_NAME] || in list_set_uadt() 367 ret = ip_set_get_extensions(set, tb, &ext); in list_set_uadt() 379 if (tb[IPSET_ATTR_CADT_FLAGS]) { in list_set_uadt() 385 if (e.before && !tb[IPSET_ATTR_NAMEREF]) { in list_set_uadt() 390 if (tb[IPSET_ATTR_NAMEREF]) { in list_set_uadt() 392 nla_data(tb[IPSET_ATTR_NAMEREF]), in list_set_uadt() 625 if (tb[IPSET_ATTR_SIZE]) in list_set_create() 626 size = ip_set_get_h32(tb[IPSET_ATTR_SIZE]); in list_set_create() [all …]
|
H A D | ip_set_core.c | 334 return nla_data(tb); in ip_set_comment_uget() 431 !!tb[IPSET_ATTR_TIMEOUT]; in add_extension() 475 if (tb[IPSET_ATTR_TIMEOUT]) { in ip_set_get_extensions() 480 if (tb[IPSET_ATTR_BYTES] || tb[IPSET_ATTR_PACKETS]) { in ip_set_get_extensions() 483 if (tb[IPSET_ATTR_BYTES]) in ip_set_get_extensions() 485 tb[IPSET_ATTR_BYTES])); in ip_set_get_extensions() 486 if (tb[IPSET_ATTR_PACKETS]) in ip_set_get_extensions() 490 if (tb[IPSET_ATTR_COMMENT]) { in ip_set_get_extensions() 495 if (tb[IPSET_ATTR_SKBMARK]) { in ip_set_get_extensions() 951 return !tb[IPSET_ATTR_PROTOCOL] || protocol(tb) != IPSET_PROTOCOL; in protocol_failed() [all …]
|
/openbmc/linux/tools/lib/bpf/ |
H A D | netlink.c | 259 if (!tb[CTRL_ATTR_FAMILY_ID]) in parse_genl_family_id() 347 struct nlattr *tb[IFLA_MAX + 1], *attr; in __dump_link_nlmsg() local 357 return dump_link_nlmsg(cookie, ifi, tb); in __dump_link_nlmsg() 370 if (!tb[IFLA_XDP]) in get_xdp_info() 410 struct nlattr *tb[NETDEV_CMD_MAX + 1]; in parse_xdp_features() local 417 if (!tb[NETDEV_A_DEV_IFINDEX] || !tb[NETDEV_A_DEV_XDP_FEATURES]) in parse_xdp_features() 425 if (tb[NETDEV_A_DEV_XDP_ZC_MAX_SEGS]) in parse_xdp_features() 666 if (!tb[TCA_OPTIONS]) in __get_tc_info() 685 struct nlattr *tb[TCA_MAX + 1]; in get_tc_info() local 687 libbpf_nla_parse(tb, TCA_MAX, in get_tc_info() [all …]
|
/openbmc/linux/arch/sparc/kernel/ |
H A D | smp_64.c | 295 struct trap_per_cpu *tb; in ldom_startcpu_cpuid() local 314 tb = &trap_block[cpu]; in ldom_startcpu_cpuid() 670 tb->cpu_list_pa, in hypervisor_xcall_deliver() 774 this_cpu, tot_cpus, tb->cpu_list_pa, tb->cpu_mondo_block_pa); in hypervisor_xcall_deliver() 788 struct trap_per_cpu *tb; in xcall_deliver() local 807 tb = &trap_block[this_cpu]; in xcall_deliver() 826 xcall_deliver_impl(tb, cnt); in xcall_deliver() 1306 tb->cpu_mondo_pa, 0); in cpu_play_dead() 1308 tb->dev_mondo_pa, 0); in cpu_play_dead() 1310 tb->resum_mondo_pa, 0); in cpu_play_dead() [all …]
|
/openbmc/linux/net/netfilter/ |
H A D | nf_conntrack_proto_icmp.c | 274 static int icmp_nlattr_to_tuple(struct nlattr *tb[], in icmp_nlattr_to_tuple() argument 279 if (!tb[CTA_PROTO_ICMP_TYPE]) in icmp_nlattr_to_tuple() 282 tuple->dst.u.icmp.type = nla_get_u8(tb[CTA_PROTO_ICMP_TYPE]); in icmp_nlattr_to_tuple() 289 if (!tb[CTA_PROTO_ICMP_CODE]) in icmp_nlattr_to_tuple() 292 tuple->dst.u.icmp.code = nla_get_u8(tb[CTA_PROTO_ICMP_CODE]); in icmp_nlattr_to_tuple() 296 if (!tb[CTA_PROTO_ICMP_ID]) in icmp_nlattr_to_tuple() 299 tuple->src.u.icmp.id = nla_get_be16(tb[CTA_PROTO_ICMP_ID]); in icmp_nlattr_to_tuple() 321 static int icmp_timeout_nlattr_to_obj(struct nlattr *tb[], in icmp_timeout_nlattr_to_obj() argument 327 if (tb[CTA_TIMEOUT_ICMP_TIMEOUT]) { in icmp_timeout_nlattr_to_obj() 331 ntohl(nla_get_be32(tb[CTA_TIMEOUT_ICMP_TIMEOUT])) * HZ; in icmp_timeout_nlattr_to_obj()
|
H A D | nft_tproxy.c | 193 const struct nlattr * const tb[]) in nft_tproxy_init() argument 199 if (!tb[NFTA_TPROXY_FAMILY] || in nft_tproxy_init() 200 (!tb[NFTA_TPROXY_REG_ADDR] && !tb[NFTA_TPROXY_REG_PORT])) in nft_tproxy_init() 203 priv->family = ntohl(nla_get_be32(tb[NFTA_TPROXY_FAMILY])); in nft_tproxy_init() 223 if (priv->family == NFPROTO_UNSPEC && tb[NFTA_TPROXY_REG_ADDR]) in nft_tproxy_init() 256 if (tb[NFTA_TPROXY_REG_ADDR]) { in nft_tproxy_init() 257 err = nft_parse_register_load(tb[NFTA_TPROXY_REG_ADDR], in nft_tproxy_init() 263 if (tb[NFTA_TPROXY_REG_PORT]) { in nft_tproxy_init() 264 err = nft_parse_register_load(tb[NFTA_TPROXY_REG_PORT], in nft_tproxy_init()
|
H A D | nf_conntrack_proto_icmpv6.c | 249 static int icmpv6_nlattr_to_tuple(struct nlattr *tb[], in icmpv6_nlattr_to_tuple() argument 254 if (!tb[CTA_PROTO_ICMPV6_TYPE]) in icmpv6_nlattr_to_tuple() 257 tuple->dst.u.icmp.type = nla_get_u8(tb[CTA_PROTO_ICMPV6_TYPE]); in icmpv6_nlattr_to_tuple() 265 if (!tb[CTA_PROTO_ICMPV6_CODE]) in icmpv6_nlattr_to_tuple() 268 tuple->dst.u.icmp.code = nla_get_u8(tb[CTA_PROTO_ICMPV6_CODE]); in icmpv6_nlattr_to_tuple() 272 if (!tb[CTA_PROTO_ICMPV6_ID]) in icmpv6_nlattr_to_tuple() 275 tuple->src.u.icmp.id = nla_get_be16(tb[CTA_PROTO_ICMPV6_ID]); in icmpv6_nlattr_to_tuple() 297 static int icmpv6_timeout_nlattr_to_obj(struct nlattr *tb[], in icmpv6_timeout_nlattr_to_obj() argument 305 if (tb[CTA_TIMEOUT_ICMPV6_TIMEOUT]) { in icmpv6_timeout_nlattr_to_obj() 307 ntohl(nla_get_be32(tb[CTA_TIMEOUT_ICMPV6_TIMEOUT])) * HZ; in icmpv6_timeout_nlattr_to_obj()
|
H A D | nft_xfrm.c | 35 const struct nlattr * const tb[]) in nft_xfrm_get_init() argument 42 if (!tb[NFTA_XFRM_KEY] || !tb[NFTA_XFRM_DIR] || !tb[NFTA_XFRM_DREG]) in nft_xfrm_get_init() 54 priv->key = ntohl(nla_get_be32(tb[NFTA_XFRM_KEY])); in nft_xfrm_get_init() 72 dir = nla_get_u8(tb[NFTA_XFRM_DIR]); in nft_xfrm_get_init() 82 if (tb[NFTA_XFRM_SPNUM]) in nft_xfrm_get_init() 83 spnum = ntohl(nla_get_be32(tb[NFTA_XFRM_SPNUM])); in nft_xfrm_get_init() 91 return nft_parse_register_store(ctx, tb[NFTA_XFRM_DREG], &priv->dreg, in nft_xfrm_get_init()
|
H A D | nft_connlimit.c | 60 const struct nlattr * const tb[], in nft_connlimit_do_init() argument 67 if (!tb[NFTA_CONNLIMIT_COUNT]) in nft_connlimit_do_init() 70 limit = ntohl(nla_get_be32(tb[NFTA_CONNLIMIT_COUNT])); in nft_connlimit_do_init() 72 if (tb[NFTA_CONNLIMIT_FLAGS]) { in nft_connlimit_do_init() 73 flags = ntohl(nla_get_be32(tb[NFTA_CONNLIMIT_FLAGS])); in nft_connlimit_do_init() 132 const struct nlattr * const tb[], in nft_connlimit_obj_init() argument 137 return nft_connlimit_do_init(ctx, tb, priv); in nft_connlimit_obj_init() 198 const struct nlattr * const tb[]) in nft_connlimit_init() argument 202 return nft_connlimit_do_init(ctx, tb, priv); in nft_connlimit_init()
|
H A D | nft_counter.c | 59 static int nft_counter_do_init(const struct nlattr * const tb[], in nft_counter_do_init() argument 71 if (tb[NFTA_COUNTER_PACKETS]) { in nft_counter_do_init() 73 be64_to_cpu(nla_get_be64(tb[NFTA_COUNTER_PACKETS])); in nft_counter_do_init() 75 if (tb[NFTA_COUNTER_BYTES]) { in nft_counter_do_init() 77 be64_to_cpu(nla_get_be64(tb[NFTA_COUNTER_BYTES])); in nft_counter_do_init() 85 const struct nlattr * const tb[], in nft_counter_obj_init() argument 90 return nft_counter_do_init(tb, priv); in nft_counter_obj_init() 219 const struct nlattr * const tb[]) in nft_counter_init() argument 223 return nft_counter_do_init(tb, priv); in nft_counter_init()
|
/openbmc/linux/net/phonet/ |
H A D | pn_netlink.c | 55 struct nlattr *tb[IFA_MAX+1]; in addr_doit() local 69 err = nlmsg_parse_deprecated(nlh, sizeof(*ifm), tb, IFA_MAX, in addr_doit() 75 if (tb[IFA_LOCAL] == NULL) in addr_doit() 77 pnaddr = nla_get_u8(tb[IFA_LOCAL]); in addr_doit() 222 struct nlattr *tb[RTA_MAX+1]; in route_doit() local 236 err = nlmsg_parse_deprecated(nlh, sizeof(*rtm), tb, RTA_MAX, in route_doit() 244 if (tb[RTA_DST] == NULL || tb[RTA_OIF] == NULL) in route_doit() 246 dst = nla_get_u8(tb[RTA_DST]); in route_doit() 250 dev = __dev_get_by_index(net, nla_get_u32(tb[RTA_OIF])); in route_doit()
|
/openbmc/linux/net/can/ |
H A D | gw.c | 816 if (tb[CGW_LIM_HOPS]) { in cgw_parse_attr() 846 if (tb[CGW_FDMOD_OR]) { in cgw_parse_attr() 905 if (tb[CGW_MOD_AND]) { in cgw_parse_attr() 921 if (tb[CGW_MOD_OR]) { in cgw_parse_attr() 937 if (tb[CGW_MOD_XOR]) { in cgw_parse_attr() 953 if (tb[CGW_MOD_SET]) { in cgw_parse_attr() 972 if (tb[CGW_CS_CRC8]) { in cgw_parse_attr() 995 if (tb[CGW_CS_XOR]) { in cgw_parse_attr() 1018 if (tb[CGW_MOD_UID]) in cgw_parse_attr() 1029 if (tb[CGW_FILTER]) in cgw_parse_attr() [all …]
|