/openbmc/linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
H A D | htb.c | 17 u16 qid; member 81 node->qid = qid; in mlx5e_htb_node_create_leaf() 150 u16 qid; in mlx5e_htb_get_txq_by_classid() local 160 qid = READ_ONCE(node->qid); in mlx5e_htb_get_txq_by_classid() 292 int qid; in mlx5e_htb_leaf_alloc_queue() local 390 qid = node->qid; in mlx5e_htb_leaf_to_inner() 437 if (node->qid == qid) in mlx5e_htb_node_find_by_qid() 460 qid = node->qid; in mlx5e_htb_leaf_del() 518 WRITE_ONCE(node->qid, qid); in mlx5e_htb_leaf_del() 570 qid = node->qid; in mlx5e_htb_leaf_del_last() [all …]
|
H A D | qos.c | 61 ix = qid % params->num_channels; in mlx5e_get_qos_sq() 62 qid /= params->num_channels; in mlx5e_get_qos_sq() 76 int txq_ix, ix, qid, err = 0; in mlx5e_open_qos_sq() local 136 rcu_assign_pointer(qos_sqs[qid], sq); in mlx5e_open_qos_sq() 158 u16 qid; in mlx5e_activate_qos_sq() local 169 priv->txq2sq[qid] = sq; in mlx5e_activate_qos_sq() 187 sq = mlx5e_get_qos_sq(priv, qid); in mlx5e_deactivate_qos_sq() 213 ix = qid % params->num_channels; in mlx5e_close_qos_sq() 214 qid /= params->num_channels; in mlx5e_close_qos_sq() 416 htb_qopt->qid = res; in mlx5e_htb_setup_tc() [all …]
|
H A D | qos.h | 24 void mlx5e_deactivate_qos_sq(struct mlx5e_priv *priv, u16 qid); 25 void mlx5e_close_qos_sq(struct mlx5e_priv *priv, u16 qid); 26 void mlx5e_reactivate_qos_sq(struct mlx5e_priv *priv, u16 qid, struct netdev_queue *txq); 27 void mlx5e_reset_qdisc(struct net_device *dev, u16 qid); 38 u16 mlx5e_qid_from_qos(struct mlx5e_channels *chs, u16 qid);
|
/openbmc/linux/drivers/vdpa/pds/ |
H A D | vdpa_dev.c | 95 iowrite16(qid, pdsv->vqs[qid].notify); in pds_vdpa_kick_vq() 122 free_irq(pdsv->vqs[qid].irq, &pdsv->vqs[qid]); in pds_vdpa_release_irq() 135 __func__, qid, pdsv->vqs[qid].ready, ready); in pds_vdpa_set_vq_ready() 148 err = pds_vdpa_cmd_init_vq(pdsv, qid, invert_idx, &pdsv->vqs[qid]); in pds_vdpa_set_vq_ready() 155 err = pds_vdpa_cmd_reset_vq(pdsv, qid, invert_idx, &pdsv->vqs[qid]); in pds_vdpa_set_vq_ready() 396 for (qid = 0; qid < pdsv->num_vqs; ++qid) { in pds_vdpa_request_irqs() 399 snprintf(pdsv->vqs[qid].irq_name, sizeof(pdsv->vqs[qid].irq_name), in pds_vdpa_request_irqs() 419 while (qid--) in pds_vdpa_request_irqs() 433 int qid; in pds_vdpa_release_irqs() local 438 for (qid = 0; qid < pdsv->num_vqs; qid++) in pds_vdpa_release_irqs() [all …]
|
H A D | cmds.c | 125 int pds_vdpa_cmd_init_vq(struct pds_vdpa_device *pdsv, u16 qid, u16 invert_idx, in pds_vdpa_cmd_init_vq() argument 134 .vdpa_vq_init.qid = cpu_to_le16(qid), in pds_vdpa_cmd_init_vq() 139 .vdpa_vq_init.intr_index = cpu_to_le16(qid), in pds_vdpa_cmd_init_vq() 147 __func__, qid, ilog2(vq_info->q_len), in pds_vdpa_cmd_init_vq() 154 qid, comp.status, ERR_PTR(err)); in pds_vdpa_cmd_init_vq() 159 int pds_vdpa_cmd_reset_vq(struct pds_vdpa_device *pdsv, u16 qid, u16 invert_idx, in pds_vdpa_cmd_reset_vq() argument 168 .vdpa_vq_reset.qid = cpu_to_le16(qid), in pds_vdpa_cmd_reset_vq() 177 qid, comp.status, ERR_PTR(err)); in pds_vdpa_cmd_reset_vq()
|
/openbmc/linux/arch/s390/include/asm/ |
H A D | ap.h | 145 : [qid] "d" (qid) in ap_tapq() 164 qid |= 1UL << 23; /* set T bit*/ in ap_test_queue() 165 return ap_tapq(qid, info); in ap_test_queue() 177 unsigned long reg0 = qid | (1UL << 24); /* fc 1UL is RAPQ */ in ap_rapq() 202 unsigned long reg0 = qid | (2UL << 24); /* fc 2UL is ZAPQ */ in ap_zapq() 295 static inline struct ap_queue_status ap_aqic(ap_qid_t qid, in ap_aqic() argument 375 static inline struct ap_queue_status ap_bapq(ap_qid_t qid) in ap_bapq() argument 377 unsigned long reg0 = qid | (7UL << 24); /* fc 7 is BAPQ */ in ap_bapq() 403 unsigned long reg0 = qid | (8UL << 24); /* fc 8 is AAPQ */ in ap_aapq() 431 static inline struct ap_queue_status ap_nqap(ap_qid_t qid, in ap_nqap() argument [all …]
|
/openbmc/linux/drivers/infiniband/hw/cxgb4/ |
H A D | resource.c | 111 u32 qid; in c4iw_get_cqid() local 119 qid = entry->qid; in c4iw_get_cqid() 143 entry->qid = qid; in c4iw_get_cqid() 157 if (rdev->stats.qid.cur > rdev->stats.qid.max) in c4iw_get_cqid() 158 rdev->stats.qid.max = rdev->stats.qid.cur; in c4iw_get_cqid() 172 entry->qid = qid; in c4iw_put_cqid() 189 qid = entry->qid; in c4iw_get_qpid() 217 entry->qid = qid; in c4iw_get_qpid() 231 if (rdev->stats.qid.cur > rdev->stats.qid.max) in c4iw_get_qpid() 232 rdev->stats.qid.max = rdev->stats.qid.cur; in c4iw_get_qpid() [all …]
|
/openbmc/linux/drivers/net/ethernet/marvell/prestera/ |
H A D | prestera_pci.c | 244 return fw->evt_queue[qid].len; in prestera_fw_evtq_len() 256 u8 qid, u32 idx) in prestera_fw_evtq_rd_set() argument 265 return fw->evt_queue[qid].addr; in prestera_fw_evtq_buf() 298 int qid; in prestera_fw_evtq_pick() local 300 for (qid = 0; qid < fw->evt_qnum; qid++) { in prestera_fw_evtq_pick() 302 return qid; in prestera_fw_evtq_pick() 321 u8 qid; in prestera_fw_evt_work_fn() local 373 return fw->cmd_queue[qid].len; in prestera_fw_cmdq_len() 455 u8 qid; in prestera_fw_init() local 478 for (qid = 0; qid < fw->cmd_qnum; qid++) { in prestera_fw_init() [all …]
|
/openbmc/linux/drivers/nvme/host/ |
H A D | auth.c | 27 int qid; member 82 qid == 0 ? NVME_QID_ANY : qid, in nvme_auth_submit() 245 chap->qid); in nvme_auth_process_dhchap_challenge() 266 chap->qid); in nvme_auth_process_dhchap_challenge() 329 __func__, chap->qid, in nvme_auth_set_dhchap_reply_data() 361 if (chap->qid == 0) in nvme_auth_process_dhchap_success1() 379 chap->qid); in nvme_auth_process_dhchap_success1() 385 if (chap->qid == 0) in nvme_auth_process_dhchap_success1() 448 chap->qid, ret); in nvme_auth_dhchap_setup_host_response() 526 chap->qid, ret); in nvme_auth_dhchap_setup_ctrl_response() [all …]
|
H A D | trace.h | 26 #define parse_nvme_cmd(qid, opcode, fctype, cdw10) \ argument 29 ((qid) ? \ 53 __field(int, qid) 64 __entry->qid = nvme_req_qid(req); 77 __entry->qid, __entry->cid, __entry->nsid, 91 __field(int, qid) 100 __entry->qid = nvme_req_qid(req); 110 __entry->qid, __entry->cid, __entry->result, 147 __field(int, qid) 154 __entry->qid = nvme_req_qid(req); [all …]
|
/openbmc/linux/drivers/vdpa/ifcvf/ |
H A D | ifcvf_base.c | 17 vp_iowrite16(qid, &cfg->queue_select); in ifcvf_set_vq_vector() 88 u16 queue_size, max_size, qid; in ifcvf_get_max_vq_size() local 91 for (qid = 1; qid < hw->nr_vring; qid++) { in ifcvf_get_max_vq_size() 92 queue_size = ifcvf_get_vq_size(hw, qid); in ifcvf_get_max_vq_size() 352 vp_iowrite16(qid, &cfg->queue_select); in ifcvf_set_vq_num() 361 vp_iowrite16(qid, &cfg->queue_select); in ifcvf_set_vq_address() 393 u16 qid; in ifcvf_reset_vring() local 395 for (qid = 0; qid < hw->nr_vring; qid++) { in ifcvf_reset_vring() 396 hw->vring[qid].cb.callback = NULL; in ifcvf_reset_vring() 397 hw->vring[qid].cb.private = NULL; in ifcvf_reset_vring() [all …]
|
H A D | ifcvf_base.h | 107 void ifcvf_notify_queue(struct ifcvf_hw *hw, u16 qid); 119 u16 ifcvf_get_vq_state(struct ifcvf_hw *hw, u16 qid); 120 int ifcvf_set_vq_state(struct ifcvf_hw *hw, u16 qid, u16 num); 124 u16 ifcvf_set_vq_vector(struct ifcvf_hw *hw, u16 qid, int vector); 126 void ifcvf_set_vq_num(struct ifcvf_hw *hw, u16 qid, u32 num); 127 int ifcvf_set_vq_address(struct ifcvf_hw *hw, u16 qid, u64 desc_area, 129 bool ifcvf_get_vq_ready(struct ifcvf_hw *hw, u16 qid); 130 void ifcvf_set_vq_ready(struct ifcvf_hw *hw, u16 qid, bool ready);
|
/openbmc/qemu/hw/display/ |
H A D | trace-events | 61 qxl_destroy_primary(int qid) "%d" 62 qxl_enter_vga_mode(int qid) "%d" 63 qxl_exit_vga_mode(int qid) "%d" 75 qxl_io_read_unexpected(int qid) "%d" 80 qxl_pre_load(int qid) "%d" 81 qxl_pre_save(int qid) "%d" 82 qxl_reset_surfaces(int qid) "%d" 93 qxl_soft_reset(int qid) "%d" 101 qxl_spice_oom(int qid) "%d" 102 qxl_spice_reset_cursor(int qid) "%d" [all …]
|
/openbmc/linux/drivers/scsi/lpfc/ |
H A D | lpfc_debugfs.h | 583 if (phba->sli4_hba.els_wq->queue_id == qid) { in lpfc_debug_dump_wq_by_id() 584 pr_err("ELS WQ[Qid:%d]\n", qid); in lpfc_debug_dump_wq_by_id() 590 pr_err("NVME LS WQ[Qid:%d]\n", qid); in lpfc_debug_dump_wq_by_id() 606 if (phba->sli4_hba.mbx_wq->queue_id == qid) { in lpfc_debug_dump_mq_by_id() 607 printk(KERN_ERR "MBX WQ[Qid:%d]\n", qid); in lpfc_debug_dump_mq_by_id() 623 if (phba->sli4_hba.hdr_rq->queue_id == qid) { in lpfc_debug_dump_rq_by_id() 624 printk(KERN_ERR "HDR RQ[Qid:%d]\n", qid); in lpfc_debug_dump_rq_by_id() 629 printk(KERN_ERR "DAT RQ[Qid:%d]\n", qid); in lpfc_debug_dump_rq_by_id() 658 pr_err("ELS CQ[Qid:%d]\n", qid); in lpfc_debug_dump_cq_by_id() 664 pr_err("NVME LS CQ[Qid:%d]\n", qid); in lpfc_debug_dump_cq_by_id() [all …]
|
/openbmc/linux/drivers/s390/crypto/ |
H A D | ap_queue.c | 158 AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid)); in ap_sm_recv() 210 AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid)); in ap_sm_read() 225 ap_qid_t qid = aq->qid; in ap_sm_write() local 272 AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid)); in ap_sm_write() 313 AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid)); in ap_sm_reset() 356 AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid)); in ap_sm_reset_wait() 397 AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid)); in ap_sm_setirq_wait() 419 AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid)); in ap_sm_assoc_wait() 427 AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid)); in ap_sm_assoc_wait() 677 AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid)); in ap_functions_show() [all …]
|
/openbmc/linux/drivers/nvme/target/ |
H A D | fabrics-cmd-auth.c | 41 __func__, ctrl->cntlid, req->sq->qid, in nvmet_auth_negotiate() 107 __func__, ctrl->cntlid, req->sq->qid, in nvmet_auth_negotiate() 120 __func__, ctrl->cntlid, req->sq->qid, in nvmet_auth_reply() 137 ctrl->cntlid, req->sq->qid); in nvmet_auth_reply() 143 ctrl->cntlid, req->sq->qid); in nvmet_auth_reply() 150 ctrl->cntlid, req->sq->qid); in nvmet_auth_reply() 241 ctrl->cntlid, req->sq->qid); in nvmet_execute_auth_send() 242 if (!req->sq->qid) { in nvmet_execute_auth_send() 329 ctrl->cntlid, req->sq->qid, in nvmet_execute_auth_send() 408 ctrl->cntlid, req->sq->qid); in nvmet_auth_success1() [all …]
|
H A D | trace.h | 28 #define parse_nvme_cmd(qid, opcode, fctype, cdw10) \ argument 31 (qid ? \ 67 __field(int, qid) 80 __entry->qid = req->sq->qid; 94 __entry->qid, __entry->cid, __entry->nsid, 96 show_opcode_name(__entry->qid, __entry->opcode, 98 parse_nvme_cmd(__entry->qid, __entry->opcode, 108 __field(int, qid) 115 __entry->qid = req->cq->qid; 124 __entry->qid, __entry->cid, __entry->result, __entry->status)
|
H A D | fabrics-cmd.c | 140 u16 qid = le16_to_cpu(c->qid); in nvmet_install_queue() local 154 if (ctrl->sqs[qid] != NULL) { in nvmet_install_queue() 155 pr_warn("qid %u has already been created\n", qid); in nvmet_install_queue() 176 nvmet_cq_setup(ctrl, req->cq, qid, sqsize + 1); in nvmet_install_queue() 177 nvmet_sq_setup(ctrl, req->sq, qid, sqsize + 1); in nvmet_install_queue() 188 qid, ctrl->cntlid, ret); in nvmet_install_queue() 189 ctrl->sqs[qid] = NULL; in nvmet_install_queue() 292 u16 qid = le16_to_cpu(c->qid); in nvmet_execute_io_connect() local 327 if (unlikely(qid > ctrl->subsys->max_qid)) { in nvmet_execute_io_connect() 328 pr_warn("invalid queue id (%d)\n", qid); in nvmet_execute_io_connect() [all …]
|
/openbmc/linux/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_process_queue_manager.c | 47 unsigned int qid) in assign_queue_slot_by_qid() argument 61 unsigned int *qid) in find_available_queue_slot() argument 77 *qid = found; in find_available_queue_slot() 225 unsigned int qid) in init_user_queue() argument 272 unsigned int *qid, in pqm_create_queue() argument 317 *qid = q_data->q_id; in pqm_create_queue() 579 unsigned int qid) in pqm_get_kernel_queue() argument 591 unsigned int qid) in pqm_get_user_queue() argument 600 unsigned int qid, in pqm_get_wave_state() argument 610 qid); in pqm_get_wave_state() [all …]
|
/openbmc/linux/drivers/net/ethernet/intel/ixgbe/ |
H A D | ixgbe_xsk.c | 15 int qid = ring->ring_idx; in ixgbe_xsk_pool() local 25 u16 qid) in ixgbe_xsk_pool_enable() argument 31 if (qid >= adapter->num_rx_queues) in ixgbe_xsk_pool_enable() 34 if (qid >= netdev->real_num_rx_queues || in ixgbe_xsk_pool_enable() 35 qid >= netdev->real_num_tx_queues) in ixgbe_xsk_pool_enable() 46 ixgbe_txrx_ring_disable(adapter, qid); in ixgbe_xsk_pool_enable() 48 set_bit(qid, adapter->af_xdp_zc_qps); in ixgbe_xsk_pool_enable() 51 ixgbe_txrx_ring_enable(adapter, qid); in ixgbe_xsk_pool_enable() 84 ixgbe_txrx_ring_enable(adapter, qid); in ixgbe_xsk_pool_disable() 91 u16 qid) in ixgbe_xsk_pool_setup() argument [all …]
|
/openbmc/linux/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | qos.c | 514 WRITE_ONCE(node->qid, qid); in otx2_qos_sw_create_leaf_node() 577 u16 qid; in otx2_get_txq_by_classid() local 585 qid = READ_ONCE(node->qid); in otx2_get_txq_by_classid() 1350 u16 qid; in otx2_qos_leaf_to_inner() local 1398 qid = node->qid; in otx2_qos_leaf_to_inner() 1441 WRITE_ONCE(node->qid, qid); in otx2_qos_leaf_to_inner() 1487 u16 qid; in otx2_qos_leaf_del() local 1499 qid = node->qid; in otx2_qos_leaf_del() 1534 u16 qid; in otx2_qos_leaf_del_last() local 1547 qid = node->qid; in otx2_qos_leaf_del_last() [all …]
|
/openbmc/linux/include/linux/ |
H A D | quota.h | 79 extern qid_t from_kqid(struct user_namespace *to, struct kqid qid); 80 extern qid_t from_kqid_munged(struct user_namespace *to, struct kqid qid); 81 extern bool qid_valid(struct kqid qid); 98 enum quota_type type, qid_t qid) in make_kqid() argument 105 kqid.uid = make_kuid(from, qid); in make_kqid() 108 kqid.gid = make_kgid(from, qid); in make_kqid() 111 kqid.projid = make_kprojid(from, qid); in make_kqid() 189 return from_kqid(ns, qid) != (qid_t) -1; in qid_has_mapping() 339 int (*get_next_id) (struct super_block *sb, struct kqid *qid); 511 extern void quota_send_warning(struct kqid qid, dev_t dev, [all …]
|
/openbmc/linux/fs/xfs/ |
H A D | xfs_quotaops.c | 216 struct kqid qid, in xfs_fs_get_dqblk() argument 225 id = from_kqid(&init_user_ns, qid); in xfs_fs_get_dqblk() 226 return xfs_qm_scall_getquota(mp, id, xfs_quota_type(qid.type), qdq); in xfs_fs_get_dqblk() 233 struct kqid *qid, in xfs_fs_get_nextdqblk() argument 243 id = from_kqid(&init_user_ns, *qid); in xfs_fs_get_nextdqblk() 244 ret = xfs_qm_scall_getquota_next(mp, &id, xfs_quota_type(qid->type), in xfs_fs_get_nextdqblk() 250 *qid = make_kqid(current_user_ns(), qid->type, id); in xfs_fs_get_nextdqblk() 257 struct kqid qid, in xfs_fs_set_dqblk() argument 267 return xfs_qm_scall_setqlim(mp, from_kqid(&init_user_ns, qid), in xfs_fs_set_dqblk() 268 xfs_quota_type(qid.type), qdq); in xfs_fs_set_dqblk()
|
/openbmc/qemu/tests/qtest/libqos/ |
H A D | virtio-9p-client.h | 44 v9fs_qid qid; member 69 v9fs_qid qid; member 145 v9fs_qid *qid; member 227 v9fs_qid *qid; member 309 v9fs_qid *qid; member 343 v9fs_qid *qid; member 376 v9fs_qid *qid; member 468 void v9fs_rattach(P9Req *req, v9fs_qid *qid); 478 void v9fs_rlopen(P9Req *req, v9fs_qid *qid, uint32_t *iounit); 484 void v9fs_rmkdir(P9Req *req, v9fs_qid *qid); [all …]
|
/openbmc/linux/drivers/vdpa/alibaba/ |
H A D | eni_vdpa.c | 273 if (!vp_legacy_get_queue_enable(ldev, qid) in eni_vdpa_set_vq_state() 281 static void eni_vdpa_set_vq_cb(struct vdpa_device *vdpa, u16 qid, in eni_vdpa_set_vq_cb() argument 286 eni_vdpa->vring[qid].cb = *cb; in eni_vdpa_set_vq_cb() 299 vp_legacy_set_queue_address(ldev, qid, 0); in eni_vdpa_set_vq_ready() 306 return vp_legacy_get_queue_enable(ldev, qid); in eni_vdpa_get_vq_ready() 309 static void eni_vdpa_set_vq_num(struct vdpa_device *vdpa, u16 qid, in eni_vdpa_set_vq_num() argument 314 u16 n = vp_legacy_get_queue_size(ldev, qid); in eni_vdpa_set_vq_num() 323 qid, n, num); in eni_vdpa_set_vq_num() 333 vp_legacy_set_queue_address(ldev, qid, pfn); in eni_vdpa_set_vq_address() 338 static void eni_vdpa_kick_vq(struct vdpa_device *vdpa, u16 qid) in eni_vdpa_kick_vq() argument [all …]
|