/openbmc/linux/drivers/bcma/ |
H A D | driver_mips.c | 72 struct bcma_device *mdev = dev->bus->drv_mips.core; in bcma_core_mips_irq() local 81 if (bcma_read32(mdev, BCMA_MIPS_MIPS74K_INTMASK(irq)) & in bcma_core_mips_irq() 92 struct bcma_device *mdev = bus->drv_mips.core; in bcma_core_mips_set_irq() local 102 bcma_write32(mdev, BCMA_MIPS_MIPS74K_INTMASK(0), in bcma_core_mips_set_irq() 103 bcma_read32(mdev, BCMA_MIPS_MIPS74K_INTMASK(0)) & in bcma_core_mips_set_irq() 106 bcma_write32(mdev, BCMA_MIPS_MIPS74K_INTMASK(oldirq), 0); in bcma_core_mips_set_irq() 110 bcma_write32(mdev, BCMA_MIPS_MIPS74K_INTMASK(0), in bcma_core_mips_set_irq() 111 bcma_read32(mdev, BCMA_MIPS_MIPS74K_INTMASK(0)) | in bcma_core_mips_set_irq() 114 u32 irqinitmask = bcma_read32(mdev, in bcma_core_mips_set_irq() 130 bcma_write32(mdev, BCMA_MIPS_MIPS74K_INTMASK(irq), in bcma_core_mips_set_irq()
|
/openbmc/linux/drivers/infiniband/hw/mlx5/ |
H A D | fs.c | 577 MLX5_CAP_FLOWTABLE_NIC_RX(mdev, in is_valid_ethertype() 579 MLX5_CAP_FLOWTABLE_NIC_RX(mdev, in is_valid_ethertype() 750 dev->mdev, log_max_ft_size)); in get_flow_table() 769 if (!MLX5_CAP_FLOWTABLE(dev->mdev, in get_flow_table() 774 dev->mdev, ft_type == MLX5_IB_FT_RX ? in get_flow_table() 833 if (mlx5_core_mp_enabled(dev->mdev) && in set_ecn_ce_spec() 855 if (mlx5_core_mp_enabled(dev->mdev) && in set_cnp_spec() 901 if (!MLX5_CAP_FLOWTABLE(dev->mdev, in mlx5_ib_fs_add_op_fc() 913 if (!MLX5_CAP_FLOWTABLE(dev->mdev, in mlx5_ib_fs_add_op_fc() 2390 obj->mdev = dev->mdev; in UVERBS_HANDLER() [all …]
|
H A D | srq.c | 112 err = mlx5_db_alloc(dev->mdev, &srq->db); in create_srq_kernel() 118 if (mlx5_frag_buf_alloc_node(dev->mdev, buf_size, &srq->buf, in create_srq_kernel() 119 dev->mdev->priv.numa_node)) { in create_srq_kernel() 164 mlx5_frag_buf_free(dev->mdev, &srq->buf); in create_srq_kernel() 167 mlx5_db_free(dev->mdev, &srq->db); in create_srq_kernel() 187 mlx5_frag_buf_free(dev->mdev, &srq->buf); in destroy_srq_kernel() 188 mlx5_db_free(dev->mdev, &srq->db); in destroy_srq_kernel() 202 __u32 max_sge_sz = MLX5_CAP_GEN(dev->mdev, max_wqe_sz_rq) / in mlx5_ib_create_srq() 268 MLX5_CAP_GEN(dev->mdev, log_tag_matching_list_sz)) { in mlx5_ib_create_srq() 412 struct mlx5_core_dev *mdev = dev->mdev; in mlx5_ib_post_srq_recv() local [all …]
|
H A D | qp.c | 478 MLX5_CAP_GEN(dev->mdev, in set_rq_size() 2000 struct mlx5_core_dev *mdev = dev->mdev; in create_xrc_tgt_qp() local 2079 struct mlx5_core_dev *mdev = dev->mdev; in create_dci() local 2240 struct mlx5_core_dev *mdev = dev->mdev; in create_user_qp() local 2441 struct mlx5_core_dev *mdev = dev->mdev; in create_kernel_qp() local 2738 if (mlx5_lag_is_active(dev->mdev) && !MLX5_CAP_GEN(dev->mdev, lag_dct)) in create_dct() 2875 struct mlx5_core_dev *mdev = dev->mdev; in process_vendor_flags() local 2977 struct mlx5_core_dev *mdev = dev->mdev; in process_create_flags() local 5692 struct mlx5_core_dev *mdev = dev->mdev; in handle_drain_completion() local 5752 struct mlx5_core_dev *mdev = dev->mdev; in mlx5_ib_drain_sq() local [all …]
|
H A D | mad.c | 175 struct mlx5_core_dev *mdev; in process_pma_cmd() local 182 if (!mdev) { in process_pma_cmd() 188 mdev = dev->mdev; in process_pma_cmd() 191 if (MLX5_CAP_GEN(dev->mdev, num_ports) == 1 && in process_pma_cmd() 192 !mlx5_core_mp_enabled(mdev)) { in process_pma_cmd() 194 mdev = dev->mdev; in process_pma_cmd() 219 err = mlx5_core_query_vport_counter(mdev, 0, 0, mdev_port_num, in process_pma_cmd() 234 err = query_ib_ppcnt(mdev, mdev_port_num, out_cnt, sz); in process_pma_cmd() 278 if (MLX5_CAP_GEN(dev->mdev, vport_counters) && in mlx5_ib_process_mad() 548 struct mlx5_core_dev *mdev = dev->mdev; in mlx5_query_mad_ifc_port() local [all …]
|
/openbmc/linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | tir.c | 159 struct mlx5_core_dev *mdev, bool reg) in mlx5e_tir_init() argument 163 tir->mdev = mdev; in mlx5e_tir_init() 165 err = mlx5_core_create_tir(tir->mdev, builder->in, &tir->tirn); in mlx5e_tir_init() 170 struct mlx5e_hw_objs *res = &tir->mdev->mlx5e_res.hw_objs; in mlx5e_tir_init() 184 struct mlx5e_hw_objs *res = &tir->mdev->mlx5e_res.hw_objs; in mlx5e_tir_destroy() 197 mlx5_core_destroy_tir(tir->mdev, tir->tirn); in mlx5e_tir_destroy() 202 return mlx5_core_modify_tir(tir->mdev, tir->tirn, builder->in); in mlx5e_tir_modify()
|
H A D | tc_tun_encap.c | 147 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_tc_encap_flows_add() 217 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_tc_encap_flows_del() 411 esw = priv->mdev->priv.eswitch; in mlx5e_tc_update_neigh_used_value() 485 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_encap_put() 498 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_encap_put_locked() 511 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_decap_put() 531 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_detach_encap() 561 struct mlx5_eswitch *esw = priv->mdev->priv.eswitch; in mlx5e_detach_decap() 1038 esw = priv->mdev->priv.eswitch; in mlx5e_tc_tun_encap_dests_set() 1069 esw_attr->dests[out_index].mdev = out_priv->mdev; in mlx5e_tc_tun_encap_dests_set() [all …]
|
H A D | mod_hdr.c | 72 mlx5e_mod_hdr_attach(struct mlx5_core_dev *mdev, in mlx5e_mod_hdr_attach() argument 118 mh->modify_hdr = mlx5_modify_header_alloc(mdev, namespace, in mlx5e_mod_hdr_attach() 135 mlx5e_mod_hdr_detach(mdev, tbl, mh); in mlx5e_mod_hdr_attach() 139 void mlx5e_mod_hdr_detach(struct mlx5_core_dev *mdev, in mlx5e_mod_hdr_detach() argument 149 mlx5_modify_header_dealloc(mdev, mh->modify_hdr); in mlx5e_mod_hdr_detach() 160 mlx5e_mod_hdr_alloc(struct mlx5_core_dev *mdev, int namespace, in mlx5e_mod_hdr_alloc() argument 170 max_hw_actions = mlx5e_mod_hdr_max_actions(mdev, namespace); in mlx5e_mod_hdr_alloc()
|
H A D | health.c | 45 err = mlx5_core_query_cq(cq->mdev, &cq->mcq, out); in mlx5e_health_cq_diag_fmsg() 160 int mlx5e_health_sq_to_ready(struct mlx5_core_dev *mdev, struct net_device *dev, u32 sqn) in mlx5e_health_sq_to_ready() argument 168 err = mlx5e_modify_sq(mdev, sqn, &msp); in mlx5e_health_sq_to_ready() 178 err = mlx5e_modify_sq(mdev, sqn, &msp); in mlx5e_health_sq_to_ready() 260 struct mlx5_core_dev *mdev = priv->mdev; in mlx5e_health_rsc_fmsg_dump() local 267 if (IS_ERR_OR_NULL(mdev->rsc_dump)) in mlx5e_health_rsc_fmsg_dump() 278 cmd = mlx5_rsc_dump_cmd_create(mdev, key); in mlx5e_health_rsc_fmsg_dump() 285 cmd_err = mlx5_rsc_dump_next(mdev, cmd, page, &size); in mlx5e_health_rsc_fmsg_dump()
|
/openbmc/linux/drivers/net/ethernet/mellanox/mlx5/core/en/tc/act/ |
H A D | police.c | 52 return !!mlx5e_get_flow_meters(parse_state->flow->priv->mdev); in tc_act_can_offload_police() 93 if (!(mlx5_fs_get_capabilities(priv->mdev, ns) & in tc_act_parse_police() 132 meter = mlx5e_tc_meter_get(priv->mdev, ¶ms); in tc_act_police_offload() 134 meter = mlx5e_tc_meter_replace(priv->mdev, ¶ms); in tc_act_police_offload() 142 mlx5_core_err(priv->mdev, "Failed to get flow meter %d\n", params.index); in tc_act_police_offload() 157 meter = mlx5e_tc_meter_get(priv->mdev, ¶ms); in tc_act_police_destroy() 160 mlx5_core_err(priv->mdev, "Failed to get flow meter %d\n", params.index); in tc_act_police_destroy() 178 meter = mlx5e_tc_meter_get(priv->mdev, ¶ms); in tc_act_police_stats()
|
/openbmc/linux/drivers/gpio/ |
H A D | gpio-menz127.c | 130 static int men_z127_probe(struct mcb_device *mdev, in men_z127_probe() argument 134 struct device *dev = &mdev->dev; in men_z127_probe() 142 men_z127_gpio->mem = mcb_request_mem(mdev, dev_name(dev)); in men_z127_probe() 155 mcb_set_drvdata(mdev, men_z127_gpio); in men_z127_probe() 157 ret = bgpio_init(&men_z127_gpio->gc, &mdev->dev, 4, in men_z127_probe() 185 static void men_z127_remove(struct mcb_device *mdev) in men_z127_remove() argument 187 struct men_z127_gpio *men_z127_gpio = mcb_get_drvdata(mdev); in men_z127_remove()
|
/openbmc/linux/drivers/net/ethernet/mellanox/mlx5/core/ |
H A D | en_dim.c | 38 struct mlx5_core_dev *mdev, struct mlx5_core_cq *mcq) in mlx5e_complete_dim_work() argument 40 mlx5_core_modify_cq_moderation(mdev, mcq, moder.usec, moder.pkts); in mlx5e_complete_dim_work() 51 mlx5e_complete_dim_work(dim, cur_moder, rq->mdev, &rq->cq.mcq); in mlx5e_rx_dim_work() 61 mlx5e_complete_dim_work(dim, cur_moder, sq->cq.mdev, &sq->cq.mcq); in mlx5e_tx_dim_work()
|
H A D | en_tc.c | 702 mdev = priv->mdev; in mlx5e_hairpin_get_mdev() 718 return mdev; in mlx5e_hairpin_get_mdev() 758 struct mlx5_core_dev *mdev = priv->mdev; in mlx5e_hairpin_create_indirect_rqt() local 907 func_mdev = priv->mdev; in mlx5e_hairpin_create() 3491 return priv->mdev == peer_priv->mdev; in same_port_devs() 3499 fmdev = priv->mdev; in mlx5e_same_hw_devs() 3500 pmdev = peer_priv->mdev; in mlx5e_same_hw_devs() 4487 in_mdev = priv->mdev; in mlx5e_tc_add_fdb_peer_flow() 4673 struct mlx5_core_dev *mdev = priv->mdev; in mlx5e_tc_block_ipsec_offload() local 4678 if (mdev->num_block_tc) in mlx5e_tc_block_ipsec_offload() [all …]
|
/openbmc/linux/drivers/net/ethernet/mellanox/mlx4/ |
H A D | en_tx.c | 54 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_create_tx_ring() local 95 set_dev_node(&mdev->dev->persist->pdev->dev, mdev->dev->numa_node); in mlx4_en_create_tx_ring() 125 ring->bf.uar = &mdev->priv_uar; in mlx4_en_create_tx_ring() 126 ring->bf.uar->map = mdev->uar_map; in mlx4_en_create_tx_ring() 142 priv->mdev->dev->numa_node), in mlx4_en_create_tx_ring() 167 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_destroy_tx_ring() local 172 mlx4_bf_free(mdev->dev, &ring->bf); in mlx4_en_destroy_tx_ring() 174 mlx4_qp_free(mdev->dev, &ring->sp_qp); in mlx4_en_destroy_tx_ring() 189 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_activate_tx_ring() local 223 struct mlx4_en_dev *mdev = priv->mdev; in mlx4_en_deactivate_tx_ring() local [all …]
|
/openbmc/linux/include/linux/mlx5/ |
H A D | driver.h | 406 struct mlx5_core_dev *mdev; member 557 struct mlx5_core_dev *mdev; member 1196 void mlx5_vf_put_core_dev(struct mlx5_core_dev *mdev); 1210 int mlx5_rdma_rn_get_params(struct mlx5_core_dev *mdev, 1336 if (!(MLX5_CAP_GEN_64(mdev, general_obj_types) & in mlx5e_is_macsec_device() 1340 if (!MLX5_CAP_GEN(mdev, log_max_dek)) in mlx5e_is_macsec_device() 1343 if (!MLX5_CAP_MACSEC(mdev, log_max_macsec_offload)) in mlx5e_is_macsec_device() 1346 if (!MLX5_CAP_FLOWTABLE_NIC_RX(mdev, macsec_decrypt) || in mlx5e_is_macsec_device() 1350 if (!MLX5_CAP_FLOWTABLE_NIC_TX(mdev, macsec_encrypt) || in mlx5e_is_macsec_device() 1369 if (((MLX5_CAP_GEN_2(mdev, flow_table_type_2_type) & in mlx5_is_macsec_roce_supported() [all …]
|
/openbmc/linux/drivers/net/ethernet/mellanox/mlx5/core/en_accel/ |
H A D | macsec_stats.c | 28 if (mlx5e_is_macsec_device(priv->mdev)) in MLX5E_DECLARE_STATS_GRP_OP_NUM_STATS() 43 if (!mlx5e_is_macsec_device(priv->mdev)) in MLX5E_DECLARE_STATS_GRP_OP_FILL_STRS() 61 if (!mlx5e_is_macsec_device(priv->mdev)) in MLX5E_DECLARE_STATS_GRP_OP_FILL_STATS() 64 macsec_fs = priv->mdev->macsec_fs; in MLX5E_DECLARE_STATS_GRP_OP_FILL_STATS()
|
/openbmc/linux/drivers/net/wireless/mediatek/mt76/mt7603/ |
H A D | mt7603.h | 235 int mt7603_tx_prepare_skb(struct mt76_dev *mdev, void *txwi_ptr, 240 void mt7603_tx_complete_skb(struct mt76_dev *mdev, struct mt76_queue_entry *e); 242 void mt7603_queue_rx_skb(struct mt76_dev *mdev, enum mt76_rxq_id q, 244 void mt7603_rx_poll_complete(struct mt76_dev *mdev, enum mt76_rxq_id q); 245 void mt7603_sta_ps(struct mt76_dev *mdev, struct ieee80211_sta *sta, bool ps); 246 int mt7603_sta_add(struct mt76_dev *mdev, struct ieee80211_vif *vif, 248 void mt7603_sta_assoc(struct mt76_dev *mdev, struct ieee80211_vif *vif, 250 void mt7603_sta_remove(struct mt76_dev *mdev, struct ieee80211_vif *vif,
|
/openbmc/linux/drivers/net/wireless/mediatek/mt76/ |
H A D | mt76x02_txrx.c | 35 void mt76x02_queue_rx_skb(struct mt76_dev *mdev, enum mt76_rxq_id q, in mt76x02_queue_rx_skb() argument 38 struct mt76x02_dev *dev = container_of(mdev, struct mt76x02_dev, mt76); in mt76x02_queue_rx_skb() 52 mt76_rx(mdev, q, skb); in mt76x02_queue_rx_skb() 123 bool mt76x02_tx_status_data(struct mt76_dev *mdev, u8 *update) in mt76x02_tx_status_data() argument 125 struct mt76x02_dev *dev = container_of(mdev, struct mt76x02_dev, mt76); in mt76x02_tx_status_data() 137 int mt76x02_tx_prepare_skb(struct mt76_dev *mdev, void *txwi_ptr, in mt76x02_tx_prepare_skb() argument 142 struct mt76x02_dev *dev = container_of(mdev, struct mt76x02_dev, mt76); in mt76x02_tx_prepare_skb() 155 pid = mt76_tx_status_skb_add(mdev, wcid, tx_info->skb); in mt76x02_tx_prepare_skb()
|
/openbmc/linux/drivers/media/usb/dvb-usb/ |
H A D | dvb-usb-dvb.c | 102 struct media_device *mdev; in dvb_usb_media_device_init() local 106 mdev = kzalloc(sizeof(*mdev), GFP_KERNEL); in dvb_usb_media_device_init() 107 if (!mdev) in dvb_usb_media_device_init() 110 media_device_usb_init(mdev, udev, d->desc->name); in dvb_usb_media_device_init() 112 dvb_register_media_controller(&adap->dvb_adap, mdev); in dvb_usb_media_device_init() 122 return media_device_register(adap->dvb_adap.mdev); in dvb_usb_media_device_register() 131 if (!adap->dvb_adap.mdev) in dvb_usb_media_device_unregister() 136 media_device_unregister(adap->dvb_adap.mdev); in dvb_usb_media_device_unregister() 137 media_device_cleanup(adap->dvb_adap.mdev); in dvb_usb_media_device_unregister() 138 kfree(adap->dvb_adap.mdev); in dvb_usb_media_device_unregister() [all …]
|
/openbmc/linux/drivers/tty/serial/ |
H A D | men_z135_uart.c | 121 struct mcb_device *mdev; member 264 dev_warn(&uart->mdev->dev, in men_z135_handle_rx() 278 dev_warn(&uart->mdev->dev, in men_z135_handle_rx() 329 dev_err(&uart->mdev->dev, in men_z135_handle_tx() 740 struct mcb_device *mdev = uart->mdev; in men_z135_request_port() local 743 mem = mcb_request_mem(uart->mdev, dev_name(&mdev->dev)); in men_z135_request_port() 814 dev = &mdev->dev; in men_z135_probe() 824 mem = &mdev->mem; in men_z135_probe() 826 mcb_set_drvdata(mdev, uart); in men_z135_probe() 832 uart->port.irq = mcb_get_irq(mdev); in men_z135_probe() [all …]
|
/openbmc/linux/drivers/vdpa/mlx5/net/ |
H A D | mlx5_vnet.c | 402 struct mlx5_core_dev *mdev = ndev->mvdev.mdev; in qp_create() local 538 struct mlx5_core_dev *mdev = ndev->mvdev.mdev; in cq_create() local 615 struct mlx5_core_dev *mdev = ndev->mvdev.mdev; in cq_destroy() local 630 struct mlx5_core_dev *mdev = ndev->mvdev.mdev; in read_umem_params() local 3362 struct mlx5_core_dev *mdev; in mlx5_vdpa_dev_add() local 3371 mdev = mgtdev->madev->mdev; in mlx5_vdpa_dev_add() 3375 dev_warn(mdev->device, in mlx5_vdpa_dev_add() 3386 dev_warn(mdev->device, in mlx5_vdpa_dev_add() 3401 dev_warn(mdev->device, in mlx5_vdpa_dev_add() 3422 mvdev->mdev = mdev; in mlx5_vdpa_dev_add() [all …]
|
/openbmc/linux/drivers/edac/ |
H A D | i10nm_base.c | 582 struct pci_dev *mdev; in get_gnr_mdev() local 594 if (mdev) { in get_gnr_mdev() 597 return mdev; in get_gnr_mdev() 600 pci_dev_put(mdev); in get_gnr_mdev() 632 if (!mdev) in get_ddr_munit() 651 if (!mdev) in get_ddr_munit() 658 return mdev; in get_ddr_munit() 736 if (!mdev) in i10nm_get_ddr_munits() 757 d->imc[lmc].mdev = mdev; in i10nm_get_ddr_munits() 819 if (!mdev) in i10nm_get_hbm_munits() [all …]
|
/openbmc/linux/drivers/net/ethernet/mellanox/mlx5/core/fpga/ |
H A D | sdk.c | 72 if (!fdev->mdev) in mlx5_fpga_mem_read_i2c() 78 err = mlx5_fpga_access_reg(fdev->mdev, actual_size, in mlx5_fpga_mem_read_i2c() 104 if (!fdev->mdev) in mlx5_fpga_mem_write_i2c() 110 err = mlx5_fpga_access_reg(fdev->mdev, actual_size, in mlx5_fpga_mem_write_i2c() 168 return mlx5_fpga_sbu_caps(fdev->mdev, buf, size); in mlx5_fpga_get_sbu_caps()
|
/openbmc/openbmc/poky/meta/recipes-core/initrdscripts/initramfs-framework/ |
H A D | mdev | 6 if [ ! -e /sbin/mdev ]; then 23 echo /sbin/mdev > /proc/sys/kernel/hotplug 24 mdev -s
|
/openbmc/linux/drivers/gpu/drm/arm/display/komeda/ |
H A D | komeda_framebuffer.c | 111 komeda_fb_none_afbc_size_check(struct komeda_dev *mdev, struct komeda_fb *kfb, in komeda_fb_none_afbc_size_check() argument 133 if ((fb->pitches[i] * block_h) % mdev->chip.bus_width) { in komeda_fb_none_afbc_size_check() 135 i, fb->pitches[i], mdev->chip.bus_width); in komeda_fb_none_afbc_size_check() 162 struct komeda_dev *mdev = dev->dev_private; in komeda_fb_create() local 170 kfb->format_caps = komeda_get_format_caps(&mdev->fmt_tbl, in komeda_fb_create() 185 ret = komeda_fb_none_afbc_size_check(mdev, kfb, file, mode_cmd); in komeda_fb_create() 196 kfb->is_va = mdev->iommu ? true : false; in komeda_fb_create() 271 struct komeda_dev *mdev = fb->dev->dev_private; in komeda_fb_is_layer_supported() local 276 supported = komeda_format_mod_supported(&mdev->fmt_tbl, layer_type, in komeda_fb_is_layer_supported()
|