Home
last modified time | relevance | path

Searched refs:aq (Results 1 – 25 of 55) sorted by relevance

123

/openbmc/linux/drivers/s390/crypto/
H A Dap_queue.c19 static void __ap_flush_queue(struct ap_queue *aq);
25 static inline bool ap_q_supports_bind(struct ap_queue *aq) in ap_q_supports_bind() argument
27 return ap_test_bit(&aq->card->functions, AP_FUNC_EP11) || in ap_q_supports_bind()
28 ap_test_bit(&aq->card->functions, AP_FUNC_ACCEL); in ap_q_supports_bind()
31 static inline bool ap_q_supports_assoc(struct ap_queue *aq) in ap_q_supports_assoc() argument
33 return ap_test_bit(&aq->card->functions, AP_FUNC_EP11); in ap_q_supports_assoc()
45 static int ap_queue_enable_irq(struct ap_queue *aq, void *ind) in ap_queue_enable_irq() argument
52 status = ap_aqic(aq->qid, qirqctrl, virt_to_phys(ind)); in ap_queue_enable_irq()
64 AP_QID_CARD(aq->qid), in ap_queue_enable_irq()
65 AP_QID_QUEUE(aq->qid)); in ap_queue_enable_irq()
[all …]
H A Dap_bus.c448 struct ap_queue *aq = from_timer(aq, t, timeout); in ap_request_timeout() local
450 spin_lock_bh(&aq->lock); in ap_request_timeout()
451 ap_wait(ap_sm_event(aq, AP_SM_EVENT_TIMEOUT)); in ap_request_timeout()
452 spin_unlock_bh(&aq->lock); in ap_request_timeout()
488 struct ap_queue *aq; in ap_tasklet_fn() local
499 hash_for_each(ap_queues, bkt, aq, hnode) { in ap_tasklet_fn()
500 spin_lock_bh(&aq->lock); in ap_tasklet_fn()
501 wait = min(wait, ap_sm_event_loop(aq, AP_SM_EVENT_POLL)); in ap_tasklet_fn()
502 spin_unlock_bh(&aq->lock); in ap_tasklet_fn()
512 struct ap_queue *aq; in ap_pending_requests() local
[all …]
H A Dap_card.c81 struct ap_queue *aq; in request_count_store() local
85 hash_for_each(ap_queues, bkt, aq, hnode) in request_count_store()
86 if (ac == aq->card) in request_count_store()
87 aq->total_request_count = 0; in request_count_store()
100 struct ap_queue *aq; in requestq_count_show() local
106 hash_for_each(ap_queues, bkt, aq, hnode) in requestq_count_show()
107 if (ac == aq->card) in requestq_count_show()
108 reqq_cnt += aq->requestq_count; in requestq_count_show()
119 struct ap_queue *aq; in pendingq_count_show() local
125 hash_for_each(ap_queues, bkt, aq, hnode) in pendingq_count_show()
[all …]
H A Dap_bus.h268 enum ap_sm_wait ap_sm_event(struct ap_queue *aq, enum ap_sm_event event);
269 enum ap_sm_wait ap_sm_event_loop(struct ap_queue *aq, enum ap_sm_event event);
271 int ap_queue_message(struct ap_queue *aq, struct ap_message *ap_msg);
272 void ap_cancel_message(struct ap_queue *aq, struct ap_message *ap_msg);
273 void ap_flush_queue(struct ap_queue *aq);
285 void ap_queue_init_reply(struct ap_queue *aq, struct ap_message *ap_msg);
287 void ap_queue_prepare_remove(struct ap_queue *aq);
288 void ap_queue_remove(struct ap_queue *aq);
289 void ap_queue_init_state(struct ap_queue *aq);
290 void _ap_queue_init_state(struct ap_queue *aq);
H A Dzcrypt_cex4.c649 struct ap_queue *aq = to_ap_queue(&ap_dev->device); in zcrypt_cex4_queue_probe() local
653 if (ap_test_bit(&aq->card->functions, AP_FUNC_ACCEL)) { in zcrypt_cex4_queue_probe()
654 zq = zcrypt_queue_alloc(aq->card->maxmsgsize); in zcrypt_cex4_queue_probe()
659 } else if (ap_test_bit(&aq->card->functions, AP_FUNC_COPRO)) { in zcrypt_cex4_queue_probe()
660 zq = zcrypt_queue_alloc(aq->card->maxmsgsize); in zcrypt_cex4_queue_probe()
665 } else if (ap_test_bit(&aq->card->functions, AP_FUNC_EP11)) { in zcrypt_cex4_queue_probe()
666 zq = zcrypt_queue_alloc(aq->card->maxmsgsize); in zcrypt_cex4_queue_probe()
675 zq->queue = aq; in zcrypt_cex4_queue_probe()
678 ap_queue_init_state(aq); in zcrypt_cex4_queue_probe()
679 ap_queue_init_reply(aq, &zq->reply); in zcrypt_cex4_queue_probe()
[all …]
H A Dzcrypt_queue.c44 struct ap_queue *aq = to_ap_queue(dev); in online_show() local
45 int online = aq->config && zq->online ? 1 : 0; in online_show()
55 struct ap_queue *aq = to_ap_queue(dev); in online_store() local
62 if (online && (!aq->config || !aq->card->config)) in online_store()
72 ap_send_online_uevent(&aq->ap_dev, online); in online_store()
/openbmc/linux/drivers/net/ethernet/intel/i40e/
H A Di40e_adminq.c21 hw->aq.asq.tail = I40E_VF_ATQT1; in i40e_adminq_init_regs()
22 hw->aq.asq.head = I40E_VF_ATQH1; in i40e_adminq_init_regs()
23 hw->aq.asq.len = I40E_VF_ATQLEN1; in i40e_adminq_init_regs()
24 hw->aq.asq.bal = I40E_VF_ATQBAL1; in i40e_adminq_init_regs()
25 hw->aq.asq.bah = I40E_VF_ATQBAH1; in i40e_adminq_init_regs()
26 hw->aq.arq.tail = I40E_VF_ARQT1; in i40e_adminq_init_regs()
27 hw->aq.arq.head = I40E_VF_ARQH1; in i40e_adminq_init_regs()
28 hw->aq.arq.len = I40E_VF_ARQLEN1; in i40e_adminq_init_regs()
29 hw->aq.arq.bal = I40E_VF_ARQBAL1; in i40e_adminq_init_regs()
30 hw->aq.arq.bah = I40E_VF_ARQBAH1; in i40e_adminq_init_regs()
[all …]
H A Di40e_nvm.c79 access, time_left, ret_code, hw->aq.asq_last_status); in i40e_acquire_nvm()
101 time_left, ret_code, hw->aq.asq_last_status); in i40e_acquire_nvm()
129 (total_delay < hw->aq.asq_cmd_timeout)) { in i40e_release_nvm()
877 mutex_lock(&hw->aq.arq_mutex); in i40e_nvmupd_command()
915 mutex_unlock(&hw->aq.arq_mutex); in i40e_nvmupd_command()
943 hw->aq.asq_last_status); in i40e_nvmupd_state_init()
954 hw->aq.asq_last_status); in i40e_nvmupd_state_init()
968 hw->aq.asq_last_status); in i40e_nvmupd_state_init()
985 hw->aq.asq_last_status); in i40e_nvmupd_state_init()
1002 hw->aq.asq_last_status); in i40e_nvmupd_state_init()
[all …]
H A Di40e_main.c1850 i40e_aq_str(hw, hw->aq.asq_last_status)); in i40e_set_mac()
1882 i40e_aq_str(hw, hw->aq.asq_last_status)); in i40e_config_rss_aq()
1894 i40e_aq_str(hw, hw->aq.asq_last_status)); in i40e_config_rss_aq()
2475 i40e_aq_str(hw, hw->aq.asq_last_status), in i40e_aqc_broadcast_filter()
2517 i40e_aq_str(hw, hw->aq.asq_last_status)); in i40e_set_promiscuous()
2529 i40e_aq_str(hw, hw->aq.asq_last_status)); in i40e_set_promiscuous()
2539 i40e_aq_str(hw, hw->aq.asq_last_status)); in i40e_set_promiscuous()
2659 filter_list_len = hw->aq.asq_buf_size / in i40e_sync_vsi_filters()
2720 filter_list_len = hw->aq.asq_buf_size / in i40e_sync_vsi_filters()
2842 hw->aq.asq_last_status); in i40e_sync_vsi_filters()
[all …]
H A Di40e_client.c356 cdev->lan_info.fw_maj_ver = pf->hw.aq.fw_maj_ver; in i40e_client_add_instance()
357 cdev->lan_info.fw_min_ver = pf->hw.aq.fw_min_ver; in i40e_client_add_instance()
358 cdev->lan_info.fw_build = pf->hw.aq.fw_build; in i40e_client_add_instance()
549 err, hw->aq.asq_last_status); in i40e_client_virtchnl_send()
691 pf->hw.aq.asq_last_status)); in i40e_client_update_vsi_ctxt()
719 pf->hw.aq.asq_last_status)); in i40e_client_update_vsi_ctxt()
H A Di40e_debugfs.c508 ring = &(hw->aq.asq); in i40e_dbg_dump_aq_desc()
521 ring = &(hw->aq.arq); in i40e_dbg_dump_aq_desc()
1064 pf->hw.aq.asq_last_status); in i40e_dbg_command_write()
1177 ret, pf->hw.aq.asq_last_status); in i40e_dbg_command_write()
1317 desc->opcode, pf->hw.aq.asq_last_status); in i40e_dbg_command_write()
1378 desc->opcode, pf->hw.aq.asq_last_status); in i40e_dbg_command_write()
1410 pf->hw.aq.asq_last_status); in i40e_dbg_command_write()
1421 __func__, pf->hw.aq.asq_last_status); in i40e_dbg_command_write()
1439 __func__, pf->hw.aq.asq_last_status); in i40e_dbg_command_write()
1447 pf->hw.aq.asq_last_status); in i40e_dbg_command_write()
[all …]
H A Di40e_common.c199 if (hw->aq.asq.len) in i40e_check_asq_alive()
200 return !!(rd32(hw, hw->aq.asq.len) & in i40e_check_asq_alive()
1346 switch (hw->aq.asq_last_status) { in i40e_aq_get_phy_capabilities()
1360 } while ((hw->aq.asq_last_status == I40E_AQ_RC_EAGAIN) && in i40e_aq_get_phy_capabilities()
1368 hw->aq.api_maj_ver == I40E_FW_API_VERSION_MAJOR && in i40e_aq_get_phy_capabilities()
1369 hw->aq.api_min_ver >= I40E_MINOR_VER_GET_LINK_INFO_XL710) { in i40e_aq_get_phy_capabilities()
1640 (hw->aq.fw_maj_ver < 4 || (hw->aq.fw_maj_ver == 4 && in i40e_aq_get_link_info()
1641 hw->aq.fw_min_ver < 40)) && hw_link_info->phy_type == 0xE) in i40e_aq_get_link_info()
1751 static bool i40e_is_aq_api_ver_ge(struct i40e_adminq_info *aq, u16 maj, in i40e_is_aq_api_ver_ge() argument
1754 return (aq->api_maj_ver > maj || in i40e_is_aq_api_ver_ge()
[all …]
/openbmc/linux/drivers/net/ethernet/intel/iavf/
H A Diavf_adminq.c19 hw->aq.asq.tail = IAVF_VF_ATQT1; in iavf_adminq_init_regs()
20 hw->aq.asq.head = IAVF_VF_ATQH1; in iavf_adminq_init_regs()
21 hw->aq.asq.len = IAVF_VF_ATQLEN1; in iavf_adminq_init_regs()
22 hw->aq.asq.bal = IAVF_VF_ATQBAL1; in iavf_adminq_init_regs()
23 hw->aq.asq.bah = IAVF_VF_ATQBAH1; in iavf_adminq_init_regs()
24 hw->aq.arq.tail = IAVF_VF_ARQT1; in iavf_adminq_init_regs()
25 hw->aq.arq.head = IAVF_VF_ARQH1; in iavf_adminq_init_regs()
26 hw->aq.arq.len = IAVF_VF_ARQLEN1; in iavf_adminq_init_regs()
27 hw->aq.arq.bal = IAVF_VF_ARQBAL1; in iavf_adminq_init_regs()
28 hw->aq.arq.bah = IAVF_VF_ARQBAH1; in iavf_adminq_init_regs()
[all …]
/openbmc/linux/drivers/infiniband/hw/efa/
H A Defa_com.c126 struct efa_com_admin_queue *aq = &edev->aq; in efa_com_admin_init_sq() local
127 struct efa_com_admin_sq *sq = &aq->sq; in efa_com_admin_init_sq()
128 u16 size = aq->depth * sizeof(*sq->entries); in efa_com_admin_init_sq()
134 dma_alloc_coherent(aq->dmadev, size, &sq->dma_addr, GFP_KERNEL); in efa_com_admin_init_sq()
152 EFA_SET(&aq_caps, EFA_REGS_AQ_CAPS_AQ_DEPTH, aq->depth); in efa_com_admin_init_sq()
163 struct efa_com_admin_queue *aq = &edev->aq; in efa_com_admin_init_cq() local
164 struct efa_com_admin_cq *cq = &aq->cq; in efa_com_admin_init_cq()
165 u16 size = aq->depth * sizeof(*cq->entries); in efa_com_admin_init_cq()
171 dma_alloc_coherent(aq->dmadev, size, &cq->dma_addr, GFP_KERNEL); in efa_com_admin_init_cq()
186 EFA_SET(&acq_caps, EFA_REGS_ACQ_CAPS_ACQ_DEPTH, aq->depth); in efa_com_admin_init_cq()
[all …]
H A Defa_com_cmd.c15 struct efa_com_admin_queue *aq = &edev->aq; in efa_com_create_qp() local
35 err = efa_com_cmd_exec(aq, in efa_com_create_qp()
60 struct efa_com_admin_queue *aq = &edev->aq; in efa_com_modify_qp() local
75 err = efa_com_cmd_exec(aq, in efa_com_modify_qp()
95 struct efa_com_admin_queue *aq = &edev->aq; in efa_com_query_qp() local
103 err = efa_com_cmd_exec(aq, in efa_com_query_qp()
129 struct efa_com_admin_queue *aq = &edev->aq; in efa_com_destroy_qp() local
135 err = efa_com_cmd_exec(aq, in efa_com_destroy_qp()
156 struct efa_com_admin_queue *aq = &edev->aq; in efa_com_create_cq() local
179 err = efa_com_cmd_exec(aq, in efa_com_create_cq()
[all …]
H A Defa_com.h107 struct efa_com_admin_queue aq; member
171 int efa_com_cmd_exec(struct efa_com_admin_queue *aq,
/openbmc/linux/drivers/spi/
H A Datmel-quadspi.c167 int (*set_cfg)(struct atmel_qspi *aq, const struct spi_mem_op *op,
239 static u32 atmel_qspi_read(struct atmel_qspi *aq, u32 offset) in atmel_qspi_read() argument
241 u32 value = readl_relaxed(aq->regs + offset); in atmel_qspi_read()
246 dev_vdbg(&aq->pdev->dev, "read 0x%08x from %s\n", value, in atmel_qspi_read()
253 static void atmel_qspi_write(u32 value, struct atmel_qspi *aq, u32 offset) in atmel_qspi_write() argument
258 dev_vdbg(&aq->pdev->dev, "write 0x%08x into %s\n", value, in atmel_qspi_write()
262 writel_relaxed(value, aq->regs + offset); in atmel_qspi_write()
308 static int atmel_qspi_set_cfg(struct atmel_qspi *aq, in atmel_qspi_set_cfg() argument
391 if (!(aq->mr & QSPI_MR_SMM)) { in atmel_qspi_set_cfg()
392 aq->mr |= QSPI_MR_SMM; in atmel_qspi_set_cfg()
[all …]
/openbmc/linux/drivers/net/ethernet/marvell/octeontx2/nic/
H A Dcn10k.c77 struct nix_cn10k_aq_enq_req *aq; in cn10k_sq_aq_init() local
81 aq = otx2_mbox_alloc_msg_nix_cn10k_aq_enq(&pfvf->mbox); in cn10k_sq_aq_init()
82 if (!aq) in cn10k_sq_aq_init()
85 aq->sq.cq = pfvf->hw.rx_queues + qidx; in cn10k_sq_aq_init()
86 aq->sq.max_sqe_size = NIX_MAXSQESZ_W16; /* 128 byte */ in cn10k_sq_aq_init()
87 aq->sq.cq_ena = 1; in cn10k_sq_aq_init()
88 aq->sq.ena = 1; in cn10k_sq_aq_init()
89 aq->sq.smq = otx2_get_smq_idx(pfvf, qidx); in cn10k_sq_aq_init()
90 aq->sq.smq_rr_weight = mtu_to_dwrr_weight(pfvf, pfvf->tx_max_pktlen); in cn10k_sq_aq_init()
91 aq->sq.default_chan = pfvf->hw.tx_chan_base; in cn10k_sq_aq_init()
[all …]
H A Dotx2_common.c316 struct nix_aq_enq_req *aq; in otx2_set_rss_table() local
323 aq = otx2_mbox_alloc_msg_nix_aq_enq(mbox); in otx2_set_rss_table()
324 if (!aq) { in otx2_set_rss_table()
333 aq = otx2_mbox_alloc_msg_nix_aq_enq(mbox); in otx2_set_rss_table()
334 if (!aq) { in otx2_set_rss_table()
340 aq->rss.rq = rss_ctx->ind_tbl[idx]; in otx2_set_rss_table()
343 aq->qidx = index + idx; in otx2_set_rss_table()
344 aq->ctype = NIX_AQ_CTYPE_RSS; in otx2_set_rss_table()
345 aq->op = NIX_AQ_INSTOP_INIT; in otx2_set_rss_table()
856 struct nix_aq_enq_req *aq; in otx2_rq_init() local
[all …]
H A Dotx2_dcbnl.c340 struct nix_aq_enq_req *aq; in otx2_update_bpid_in_rqctx() local
357 aq = otx2_mbox_alloc_msg_nix_aq_enq(&pfvf->mbox); in otx2_update_bpid_in_rqctx()
358 if (!aq) { in otx2_update_bpid_in_rqctx()
363 aq->cq.bpid = pfvf->bpid[vlan_prio]; in otx2_update_bpid_in_rqctx()
364 aq->cq_mask.bpid = GENMASK(8, 0); in otx2_update_bpid_in_rqctx()
367 aq->qidx = qidx; in otx2_update_bpid_in_rqctx()
368 aq->ctype = NIX_AQ_CTYPE_CQ; in otx2_update_bpid_in_rqctx()
369 aq->op = NIX_AQ_INSTOP_WRITE; in otx2_update_bpid_in_rqctx()
/openbmc/linux/drivers/net/ethernet/marvell/octeontx2/af/
H A Drvu_npa.c18 struct admin_queue *aq = block->aq; in npa_aq_enqueue_wait() local
23 result = (struct npa_aq_res_s *)aq->res->base; in npa_aq_enqueue_wait()
29 memcpy((void *)(aq->inst->base + (head * aq->inst->entry_sz)), in npa_aq_enqueue_wait()
30 (void *)inst, aq->inst->entry_sz); in npa_aq_enqueue_wait()
69 struct admin_queue *aq; in rvu_npa_aq_enq_inst() local
83 aq = block->aq; in rvu_npa_aq_enq_inst()
84 if (!aq) { in rvu_npa_aq_enq_inst()
101 inst.res_addr = (u64)aq->res->iova; in rvu_npa_aq_enq_inst()
106 spin_lock(&aq->lock); in rvu_npa_aq_enq_inst()
109 memset(aq->res->base, 0, aq->res->entry_sz); in rvu_npa_aq_enq_inst()
[all …]
/openbmc/linux/lib/
H A Dkasprintf.c19 va_list aq; in kvasprintf() local
21 va_copy(aq, ap); in kvasprintf()
22 first = vsnprintf(NULL, 0, fmt, aq); in kvasprintf()
23 va_end(aq); in kvasprintf()
/openbmc/linux/arch/riscv/net/
H A Dbpf_jit.h274 static inline u32 rv_amo_insn(u8 funct5, u8 aq, u8 rl, u8 rs2, u8 rs1, in rv_amo_insn() argument
277 u8 funct7 = (funct5 << 2) | (aq << 1) | rl; in rv_amo_insn()
557 static inline u32 rv_amoadd_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl) in rv_amoadd_w() argument
559 return rv_amo_insn(0, aq, rl, rs2, rs1, 2, rd, 0x2f); in rv_amoadd_w()
562 static inline u32 rv_amoand_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl) in rv_amoand_w() argument
564 return rv_amo_insn(0xc, aq, rl, rs2, rs1, 2, rd, 0x2f); in rv_amoand_w()
567 static inline u32 rv_amoor_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl) in rv_amoor_w() argument
569 return rv_amo_insn(0x8, aq, rl, rs2, rs1, 2, rd, 0x2f); in rv_amoor_w()
572 static inline u32 rv_amoxor_w(u8 rd, u8 rs2, u8 rs1, u8 aq, u8 rl) in rv_amoxor_w() argument
574 return rv_amo_insn(0x4, aq, rl, rs2, rs1, 2, rd, 0x2f); in rv_amoxor_w()
[all …]
/openbmc/linux/drivers/crypto/hisilicon/sec2/
H A Dsec_crypto.c1218 struct aead_request *aq = req->aead_req.aead_req; in GEN_SEC_AEAD_SETKEY_FUNC() local
1220 return sec_cipher_map(ctx, req, aq->src, aq->dst); in GEN_SEC_AEAD_SETKEY_FUNC()
1225 struct aead_request *aq = req->aead_req.aead_req; in sec_aead_sgl_unmap() local
1227 sec_cipher_unmap(ctx, req, aq->src, aq->dst); in sec_aead_sgl_unmap()
1528 struct aead_request *aq = a_req->aead_req; in sec_auth_bd_fill_xcm() local
1529 struct crypto_aead *tfm = crypto_aead_reqtfm(aq); in sec_auth_bd_fill_xcm()
1545 sec_sqe->type2.alen_ivllen = cpu_to_le32(aq->assoclen); in sec_auth_bd_fill_xcm()
1547 sec_sqe->type2.cipher_src_offset = cpu_to_le16((u16)aq->assoclen); in sec_auth_bd_fill_xcm()
1556 struct aead_request *aq = a_req->aead_req; in sec_auth_bd_fill_xcm_v3() local
1557 struct crypto_aead *tfm = crypto_aead_reqtfm(aq); in sec_auth_bd_fill_xcm_v3()
[all …]
/openbmc/linux/arch/riscv/include/asm/
H A Dasm.h26 #define REG_AMOSWAP_AQ __REG_SEL(amoswap.d.aq, amoswap.w.aq)

123