/openbmc/linux/drivers/gpu/drm/nouveau/nvkm/falcon/ |
H A D | msgq.c | 29 msgq->position = nvkm_falcon_rd32(msgq->qmgr->falcon, msgq->tail_reg); in nvkm_falcon_msgq_open() 38 nvkm_falcon_wr32(falcon, msgq->tail_reg, msgq->position); in nvkm_falcon_msgq_close() 46 u32 head = nvkm_falcon_rd32(msgq->qmgr->falcon, msgq->head_reg); in nvkm_falcon_msgq_empty() 47 u32 tail = nvkm_falcon_rd32(msgq->qmgr->falcon, msgq->tail_reg); in nvkm_falcon_msgq_empty() 60 msgq->position = msgq->offset; in nvkm_falcon_msgq_pop() 161 msgq->head_reg = falcon->func->msgq.head; in nvkm_falcon_msgq_recv_initmsg() 162 msgq->tail_reg = falcon->func->msgq.tail; in nvkm_falcon_msgq_recv_initmsg() 182 msgq->head_reg = func->msgq.head + index * func->msgq.stride; in nvkm_falcon_msgq_init() 183 msgq->tail_reg = func->msgq.tail + index * func->msgq.stride; in nvkm_falcon_msgq_init() 194 if (msgq) { in nvkm_falcon_msgq_del() [all …]
|
H A D | Kbuild | 5 nvkm-y += nvkm/falcon/msgq.o
|
/openbmc/linux/drivers/scsi/arm/ |
H A D | msgqueue.c | 26 if ((mq = msgq->free) != NULL) in mqe_alloc() 27 msgq->free = mq->next; in mqe_alloc() 41 mq->next = msgq->free; in mqe_free() 42 msgq->free = mq; in mqe_free() 55 msgq->qe = NULL; in msgqueue_initialise() 56 msgq->free = &msgq->entries[0]; in msgqueue_initialise() 59 msgq->entries[i].next = &msgq->entries[i + 1]; in msgqueue_initialise() 70 void msgqueue_free(MsgQueue_t *msgq) in msgqueue_free() argument 133 mqp = &msgq->qe; in msgqueue_addmsg() 154 mqe_free(msgq, mq); in msgqueue_flush() [all …]
|
H A D | msgqueue.h | 36 extern void msgqueue_initialise(MsgQueue_t *msgq); 43 extern void msgqueue_free(MsgQueue_t *msgq); 51 extern int msgqueue_msglength(MsgQueue_t *msgq); 60 extern struct message *msgqueue_getmsg(MsgQueue_t *msgq, int msgno); 70 extern int msgqueue_addmsg(MsgQueue_t *msgq, int length, ...); 77 extern void msgqueue_flush(MsgQueue_t *msgq);
|
/openbmc/linux/drivers/net/ethernet/brocade/bna/ |
H A D | bfa_msgq.c | 305 cmdq->msgq = msgq; in bfa_msgq_cmdq_attach() 492 rspq->msgq = msgq; in bfa_msgq_rspq_attach() 520 bfa_nw_ioc_mbox_queue(msgq->ioc, &msgq->init_mb, NULL, NULL); in bfa_msgq_init() 557 bfa_wc_init(&msgq->init_wc, bfa_msgq_init, msgq); in bfa_msgq_notify() 597 msgq->rspq.addr.pa = pa; in bfa_msgq_memclaim() 603 msgq->ioc = ioc; in bfa_msgq_attach() 605 bfa_msgq_cmdq_attach(&msgq->cmdq, msgq); in bfa_msgq_attach() 606 bfa_msgq_rspq_attach(&msgq->rspq, msgq); in bfa_msgq_attach() 608 bfa_nw_ioc_mbox_regisr(msgq->ioc, BFI_MC_MSGQ, bfa_msgq_isr, msgq); in bfa_msgq_attach() 609 bfa_ioc_notify_init(&msgq->ioc_notify, bfa_msgq_notify, msgq); in bfa_msgq_attach() [all …]
|
H A D | bfa_msgq.h | 77 struct bfa_msgq *msgq; member 104 struct bfa_msgq *msgq; member 119 void bfa_msgq_memclaim(struct bfa_msgq *msgq, u8 *kva, u64 pa); 120 void bfa_msgq_attach(struct bfa_msgq *msgq, struct bfa_ioc *ioc); 121 void bfa_msgq_regisr(struct bfa_msgq *msgq, enum bfi_mclass mc, 123 void bfa_msgq_cmd_post(struct bfa_msgq *msgq, 125 void bfa_msgq_rsp_copy(struct bfa_msgq *msgq, u8 *buf, size_t buf_len);
|
H A D | bna_enet.c | 418 bfa_msgq_cmd_post(ðport->bna->msgq, ðport->msgq_cmd); in bna_bfi_ethport_admin_up() 435 bfa_msgq_cmd_post(ðport->bna->msgq, ðport->msgq_cmd); in bna_bfi_ethport_admin_down() 456 bfa_msgq_cmd_post(ðport->bna->msgq, ðport->msgq_cmd); in bna_bfi_ethport_lpbk_up() 473 bfa_msgq_cmd_post(ðport->bna->msgq, ðport->msgq_cmd); in bna_bfi_ethport_lpbk_down() 1171 bfa_msgq_cmd_post(&enet->bna->msgq, &enet->msgq_cmd); in bna_bfi_pause_set() 1630 bfa_msgq_cmd_post(&ioceth->bna->msgq, &ioceth->msgq_cmd); in bna_bfi_attr_get() 1729 bfa_msgq_attach(&bna->msgq, &ioceth->ioc); in bna_ioceth_init() 1730 bfa_msgq_memclaim(&bna->msgq, kva, dma); in bna_ioceth_init() 1731 bfa_msgq_regisr(&bna->msgq, BFI_MC_ENET, bna_msgq_rsp_handler, bna); in bna_ioceth_init() 1858 bfa_msgq_cmd_post(&bna->msgq, &bna->stats_mod.stats_get_cmd); in bna_bfi_stats_get()
|
H A D | bna_tx_rx.c | 190 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_ucast_req() 206 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_mcast_add_req() 222 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_mcast_del_req() 237 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_mcast_filter_req() 252 bfa_msgq_cmd_post(&rxf->rx->bna->msgq, &rxf->msgq_cmd); in bna_bfi_rx_promisc_req() 1711 bfa_msgq_cmd_post(&rx->bna->msgq, &rx->msgq_cmd); in bna_bfi_rx_enet_start() 1725 bfa_msgq_cmd_post(&rx->bna->msgq, &rx->msgq_cmd); in bna_bfi_rx_enet_stop() 2093 bfa_msgq_rsp_copy(&rx->bna->msgq, (u8 *)cfg_rsp, in bna_bfi_rx_enet_start_rsp() 3131 bfa_msgq_cmd_post(&tx->bna->msgq, &tx->msgq_cmd); in bna_bfi_tx_enet_start() 3145 bfa_msgq_cmd_post(&tx->bna->msgq, &tx->msgq_cmd); in bna_bfi_tx_enet_stop() [all …]
|
H A D | bna_types.h | 928 struct bfa_msgq msgq; member
|
/openbmc/linux/drivers/misc/bcm-vk/ |
H A D | bcm_vk_msg.c | 468 chan->msgq[j] = msgq; in bcm_vk_sync_msgq() 484 readw_relaxed(&msgq->num), in bcm_vk_sync_msgq() 501 msgq++; in bcm_vk_sync_msgq() 555 msgq = chan->msgq[q_num]; in bcm_vk_append_ib_sgl() 618 msgq = chan->msgq[q_num]; in bcm_to_v_msg_enqueue() 667 readl_relaxed(&msgq->num), in bcm_to_v_msg_enqueue() 671 msgq_occupied(msgq, qinfo), in bcm_to_v_msg_enqueue() 673 readl_relaxed(&msgq->size)); in bcm_to_v_msg_enqueue() 803 msgq = chan->msgq[q_num]; in bcm_to_h_msg_dequeue() 865 readl_relaxed(&msgq->num), in bcm_to_h_msg_dequeue() [all …]
|
H A D | bcm_vk_msg.h | 136 struct bcm_vk_msgq __iomem *msgq[VK_MSGQ_MAX_NR]; member
|
/openbmc/linux/drivers/gpu/drm/nouveau/nvkm/engine/sec2/ |
H A D | gp102.c | 127 ret = nvkm_falcon_msgq_recv_initmsg(sec2->msgq, &msg, sizeof(msg)); in gp102_sec2_initmsg() 137 nvkm_falcon_msgq_init(sec2->msgq, in gp102_sec2_initmsg() 172 if (!nvkm_falcon_msgq_empty(sec2->msgq)) in gp102_sec2_intr() 173 nvkm_falcon_msgq_recv(sec2->msgq); in gp102_sec2_intr() 215 .msgq = { 0xa30, 0xa34, 8 },
|
H A D | ga102.c | 35 ret = nvkm_falcon_msgq_recv_initmsg(sec2->msgq, &msg, sizeof(msg)); in ga102_sec2_initmsg() 45 nvkm_falcon_msgq_init(sec2->msgq, msg.queue_info[i].index, in ga102_sec2_initmsg() 137 .msgq = { 0xc80, 0xc84, 8 },
|
H A D | base.c | 118 nvkm_falcon_msgq_del(&sec2->msgq); in nvkm_sec2_dtor() 160 (ret = nvkm_falcon_msgq_new(sec2->qmgr, "msgq", &sec2->msgq))) in nvkm_sec2_new_()
|
H A D | tu102.c | 44 .msgq = { 0xc80, 0xc84, 8 },
|
/openbmc/linux/drivers/gpu/drm/nouveau/nvkm/subdev/pmu/ |
H A D | gm20b.c | 164 ret = nvkm_falcon_msgq_recv_initmsg(pmu->msgq, &msg, sizeof(msg)); in gm20b_pmu_initmsg() 178 nvkm_falcon_msgq_init(pmu->msgq, msg.queue_info[4].index, in gm20b_pmu_initmsg() 197 nvkm_falcon_msgq_recv(pmu->msgq); in gm20b_pmu_recv()
|
H A D | base.c | 108 nvkm_falcon_msgq_del(&pmu->msgq); in nvkm_pmu_dtor() 152 (ret = nvkm_falcon_msgq_new(pmu->qmgr, "msgq", &pmu->msgq))) in nvkm_pmu_ctor()
|
H A D | gp102.c | 39 .msgq = { 0x4c8, 0x4cc, 0 },
|
H A D | gm200.c | 58 .msgq = { 0x4c8, 0x4cc, 0 },
|
/openbmc/linux/drivers/gpu/drm/nouveau/include/nvkm/engine/ |
H A D | sec2.h | 18 struct nvkm_falcon_msgq *msgq; member
|
H A D | falcon.h | 88 } cmdq, msgq; member
|
/openbmc/linux/drivers/isdn/mISDN/ |
H A D | stack.c | 28 skb_queue_tail(&st->msgq, skb); in _queue_message() 221 skb = skb_dequeue(&st->msgq); in mISDNStackd() 226 skb = skb_dequeue(&st->msgq); in mISDNStackd() 266 if (!skb_queue_empty(&st->msgq)) in mISDNStackd() 315 skb_queue_purge(&st->msgq); in mISDNStackd() 379 skb_queue_head_init(&newst->msgq); in create_stack()
|
/openbmc/linux/drivers/gpu/drm/nouveau/include/nvkm/subdev/ |
H A D | pmu.h | 15 struct nvkm_falcon_msgq *msgq; member
|
/openbmc/linux/include/linux/ |
H A D | mISDNif.h | 511 struct sk_buff_head msgq; member
|