/openbmc/linux/drivers/media/usb/uvc/ |
H A D | uvc_queue.c | 220 queue->queue.type = type; in uvc_queue_init() 222 queue->queue.drv_priv = queue; in uvc_queue_init() 227 queue->queue.lock = &queue->mutex; in uvc_queue_init() 234 queue->queue.io_modes |= VB2_DMABUF; in uvc_queue_init() 235 queue->queue.ops = &uvc_queue_qops; in uvc_queue_init() 239 ret = vb2_queue_init(&queue->queue); in uvc_queue_init() 254 vb2_queue_release(&queue->queue); in uvc_queue_release() 268 ret = vb2_reqbufs(&queue->queue, rb); in uvc_request_buffers() 279 ret = vb2_querybuf(&queue->queue, buf); in uvc_query_buffer() 315 ret = vb2_expbuf(&queue->queue, exp); in uvc_export_buffer() [all …]
|
/openbmc/linux/drivers/usb/gadget/function/ |
H A D | uvc_queue.c | 140 queue->queue.type = type; in uvcg_queue_init() 142 queue->queue.drv_priv = queue; in uvcg_queue_init() 144 queue->queue.ops = &uvc_queue_qops; in uvcg_queue_init() 145 queue->queue.lock = lock; in uvcg_queue_init() 155 queue->queue.dev = dev; in uvcg_queue_init() 157 ret = vb2_queue_init(&queue->queue); in uvcg_queue_init() 173 vb2_queue_release(&queue->queue); in uvcg_free_buffers() 184 ret = vb2_reqbufs(&queue->queue, rb); in uvcg_alloc_buffers() 223 return vb2_mmap(&queue->queue, vma); in uvcg_queue_mmap() 301 ret = vb2_streamon(&queue->queue, queue->queue.type); in uvcg_queue_enable() [all …]
|
/openbmc/linux/drivers/net/wireless/st/cw1200/ |
H A D | queue.c | 97 list_for_each_entry_safe(iter, tmp, &queue->queue, head) { in __cw1200_queue_gc() 119 if (queue->num_queued <= (queue->capacity >> 1)) { in __cw1200_queue_gc() 172 memset(queue, 0, sizeof(*queue)); in cw1200_queue_init() 177 INIT_LIST_HEAD(&queue->queue); in cw1200_queue_init() 211 list_splice_tail_init(&queue->queue, &queue->pending); in cw1200_queue_clear() 268 ret = queue->num_queued - queue->num_pending; in cw1200_queue_get_num_queued() 296 list_move_tail(&item->head, &queue->queue); in cw1200_queue_put() 343 list_for_each_entry(item, &queue->queue, head) { in cw1200_queue_get() 408 list_move(&item->head, &queue->queue); in cw1200_queue_requeue() 434 list_move(&item->head, &queue->queue); in cw1200_queue_requeue_all() [all …]
|
/openbmc/linux/drivers/net/wireless/broadcom/b43legacy/ |
H A D | pio.c | 205 struct b43legacy_pioqueue *queue = packet->queue; in free_txpacket() local 219 struct b43legacy_pioqueue *queue = packet->queue; in pio_tx_packet() local 241 if (queue->tx_devq_used + octets > queue->tx_devq_size) in pio_tx_packet() 308 packet->queue = queue; in setup_txqueues() 323 queue = kzalloc(sizeof(*queue), GFP_KERNEL); in b43legacy_setup_pioqueue() 324 if (!queue) in b43legacy_setup_pioqueue() 381 if (!queue) in b43legacy_destroy_pioqueue() 413 if (!queue) in b43legacy_pio_init() 418 if (!queue) in b43legacy_pio_init() 423 if (!queue) in b43legacy_pio_init() [all …]
|
/openbmc/linux/drivers/net/xen-netback/ |
H A D | rx.c | 92 if (queue->rx_queue_len >= queue->rx_queue_max) { in xenvif_rx_queue_tail() 119 xenvif_update_needed_slots(queue, skb_peek(&queue->rx_queue)); in xenvif_rx_dequeue() 122 if (queue->rx_queue_len < queue->rx_queue_max) { in xenvif_rx_dequeue() 125 txq = netdev_get_tx_queue(queue->vif->dev, queue->id); in xenvif_rx_dequeue() 164 gnttab_batch_copy(queue->rx_copy.op, queue->rx_copy.num); in xenvif_rx_copy_flush() 204 op = &queue->rx_copy.op[queue->rx_copy.num]; in xenvif_rx_copy_add() 226 queue->rx_copy.idx[queue->rx_copy.num] = queue->rx.req_cons; in xenvif_rx_copy_add() 332 queue->rx.rsp_prod_pvt = queue->rx.req_cons; in xenvif_rx_complete() 473 req = RING_GET_REQUEST(&queue->rx, queue->rx.req_cons); in xenvif_rx_skb() 474 rsp = RING_GET_RESPONSE(&queue->rx, queue->rx.req_cons); in xenvif_rx_skb() [all …]
|
H A D | netback.c | 162 wake_up(&queue->wq); in xenvif_kick_thread() 190 max_credit = queue->remaining_credit + queue->credit_bytes; in tx_add_credit() 200 struct xenvif_queue *queue = from_timer(queue, t, credit_timeout); in xenvif_tx_credit_callback() local 201 tx_add_credit(queue); in xenvif_tx_credit_callback() 345 queue->pages_to_map[mop-queue->tx_map_ops] = queue->mmap_pages[pending_idx]; in xenvif_tx_create_map_op() 932 if (queue->tx.sring->req_prod - queue->tx.req_cons > in xenvif_tx_build_gops() 937 queue->tx.sring->req_prod, queue->tx.req_cons, in xenvif_tx_build_gops() 1296 BUG_ON(queue->dealloc_prod - queue->dealloc_cons >= in xenvif_zerocopy_callback() 1341 queue->pages_to_unmap[gop - queue->tx_unmap_ops] = in xenvif_tx_dealloc_action() 1521 return queue->dealloc_cons != queue->dealloc_prod; in tx_dealloc_work_todo() [all …]
|
H A D | interface.c | 311 if (queue->tx_irq != queue->rx_irq) in xenvif_up() 326 if (queue->tx_irq != queue->rx_irq) in xenvif_down() 562 queue->credit_bytes = queue->remaining_credit = ~0UL; in xenvif_init_queue() 686 unbind_from_irqhandler(queue->tx_irq, queue); in xenvif_disconnect_queue() 687 if (queue->tx_irq == queue->rx_irq) in xenvif_disconnect_queue() 693 unbind_from_irqhandler(queue->rx_irq, queue); in xenvif_disconnect_queue() 748 queue->name, queue); in xenvif_connect_data() 751 queue->tx_irq = queue->rx_irq = err; in xenvif_connect_data() 755 snprintf(queue->tx_irq_name, sizeof(queue->tx_irq_name), in xenvif_connect_data() 759 queue->tx_irq_name, queue); in xenvif_connect_data() [all …]
|
/openbmc/linux/drivers/iio/buffer/ |
H A D | industrialio-buffer-dma.c | 183 block->queue = queue; in iio_dma_buffer_alloc_block() 194 struct iio_dma_buffer_queue *queue = block->queue; in _iio_dma_buffer_block_done() local 215 struct iio_dma_buffer_queue *queue = block->queue; in iio_dma_buffer_block_done() local 369 if (!queue->ops) in iio_dma_buffer_submit_block() 374 ret = queue->ops->submit(queue, block); in iio_dma_buffer_submit_block() 434 if (queue->ops && queue->ops->abort) in iio_dma_buffer_disable() 435 queue->ops->abort(queue); in iio_dma_buffer_disable() 615 queue->buffer.watermark = queue->buffer.length / 2; in iio_dma_buffer_init() 616 queue->dev = dev; in iio_dma_buffer_init() 617 queue->ops = ops; in iio_dma_buffer_init() [all …]
|
/openbmc/linux/drivers/nvme/target/ |
H A D | tcp.c | 432 struct nvmet_tcp_queue *queue = cmd->queue; in nvmet_setup_c2h_data_pdu() local 465 struct nvmet_tcp_queue *queue = cmd->queue; in nvmet_setup_r2t_pdu() local 490 struct nvmet_tcp_queue *queue = cmd->queue; in nvmet_setup_response_pdu() local 549 struct nvmet_tcp_queue *queue = cmd->queue; in nvmet_tcp_queue_response() local 609 struct nvmet_tcp_queue *queue = cmd->queue; in nvmet_try_send_data() local 721 struct nvmet_tcp_queue *queue = cmd->queue; in nvmet_try_send_ddgst() local 1023 queue->idx, queue->state); in nvmet_tcp_done_recv_pdu() 1035 queue->cmd = nvmet_tcp_get_cmd(queue); in nvmet_tcp_done_recv_pdu() 1039 queue->idx, queue->nr_cmds, queue->send_list_len, in nvmet_tcp_done_recv_pdu() 1081 queue->cmd->req.execute(&queue->cmd->req); in nvmet_tcp_done_recv_pdu() [all …]
|
H A D | rdma.c | 662 struct nvmet_rdma_queue *queue = rsp->queue; in nvmet_rdma_release_rsp() local 946 struct nvmet_rdma_queue *queue = rsp->queue; in nvmet_rdma_execute_command() local 1023 cmd->queue = queue; in nvmet_rdma_recv_done() 1034 rsp->queue = queue; in nvmet_rdma_recv_done() 1293 if (queue->port->pi_enable && queue->host_qid) in nvmet_rdma_create_queue_ib() 1301 queue->qp = queue->cm_id->qp; in nvmet_rdma_create_queue_ib() 1311 queue->cmds[i].queue = queue; in nvmet_rdma_create_queue_ib() 1346 nvmet_rdma_free_cmds(queue->dev, queue->cmds, in nvmet_rdma_free_queue() 1420 queue = kzalloc(sizeof(*queue), GFP_KERNEL); in nvmet_rdma_alloc_queue() 1464 queue->comp_vector = !queue->host_qid ? 0 : in nvmet_rdma_alloc_queue() [all …]
|
/openbmc/linux/drivers/misc/genwqe/ |
H A D | card_ddcb.c | 84 return queue->ddcb_next == queue->ddcb_act; in queue_empty() 89 if (queue->ddcb_next >= queue->ddcb_act) in queue_enqueued_ddcbs() 90 return queue->ddcb_next - queue->ddcb_act; in queue_enqueued_ddcbs() 92 return queue->ddcb_max - (queue->ddcb_act - queue->ddcb_next); in queue_enqueued_ddcbs() 332 struct ddcb_queue *queue = req->queue; in copy_ddcb_results() local 451 queue->ddcb_act = (queue->ddcb_act + 1) % queue->ddcb_max; in genwqe_check_ddcb_queue() 487 queue = req->queue; in __genwqe_wait_ddcb() 586 queue->ddcb_next = (queue->ddcb_next + 1) % queue->ddcb_max; in get_next_ddcb() 625 struct ddcb_queue *queue = req->queue; in __genwqe_purge_ddcb() local 772 queue = req->queue = &cd->queue; in __genwqe_enqueue_ddcb() [all …]
|
/openbmc/linux/drivers/crypto/cavium/zip/ |
H A D | zip_device.c | 59 return ((zip_dev->iq[queue].sw_head - zip_dev->iq[queue].sw_tail) * in zip_cmd_queue_consumed() 81 u32 queue = 0; in zip_load_instr() local 91 queue = 0; in zip_load_instr() 93 queue = 1; in zip_load_instr() 128 zip_dev->iq[queue].sw_head = zip_dev->iq[queue].sw_tail; in zip_load_instr() 151 zip_dev->iq[queue].sw_head, zip_dev->iq[queue].sw_tail, in zip_load_instr() 152 zip_dev->iq[queue].hw_tail); in zip_load_instr() 166 return queue; in zip_load_instr() 187 zip_dev->iq[queue].hw_tail = zip_dev->iq[queue].sw_head; in zip_update_cmd_bufs() 197 zip_dev->iq[queue].sw_head, zip_dev->iq[queue].sw_tail, in zip_update_cmd_bufs() [all …]
|
/openbmc/qemu/net/ |
H A D | queue.c | 65 NetQueue *queue; in qemu_new_net_queue() local 71 queue->nq_count = 0; in qemu_new_net_queue() 76 queue->delivering = 0; in qemu_new_net_queue() 78 return queue; in qemu_new_net_queue() 90 g_free(queue); in qemu_del_net_queue() 102 if (queue->nq_count >= queue->nq_maxlen && !sent_cb) { in qemu_net_queue_append() 112 queue->nq_count++; in qemu_net_queue_append() 127 if (queue->nq_count >= queue->nq_maxlen && !sent_cb) { in qemu_net_queue_append_iov() 147 queue->nq_count++; in qemu_net_queue_append_iov() 164 ret = queue->deliver(sender, flags, &iov, 1, queue->opaque); in qemu_net_queue_deliver() [all …]
|
/openbmc/linux/drivers/net/wireguard/ |
H A D | queueing.c | 30 memset(queue, 0, sizeof(*queue)); in wg_packet_queue_init() 31 queue->last_cpu = -1; in wg_packet_queue_init() 35 queue->worker = wg_packet_percpu_multicore_worker_alloc(function, queue); in wg_packet_queue_init() 36 if (!queue->worker) { in wg_packet_queue_init() 51 #define STUB(queue) ((struct sk_buff *)&queue->empty) argument 56 queue->head = queue->tail = STUB(queue); in wg_prev_queue_init() 57 queue->peeked = NULL; in wg_prev_queue_init() 87 queue->tail = next; in wg_prev_queue_dequeue() 92 queue->tail = next; in wg_prev_queue_dequeue() 98 __wg_prev_queue_enqueue(queue, STUB(queue)); in wg_prev_queue_dequeue() [all …]
|
/openbmc/linux/drivers/nvme/host/ |
H A D | tcp.c | 189 return queue - queue->ctrl->queues; in nvme_tcp_queue_id() 340 struct nvme_tcp_queue *queue = req->queue; in nvme_tcp_queue_request() local 492 req->queue = queue; in nvme_tcp_init_request() 630 struct nvme_tcp_queue *queue = req->queue; in nvme_tcp_setup_h2c_data_pdu() local 858 if (queue->recv_ddgst != queue->exp_ddgst) { in nvme_tcp_recv_ddgst() 928 if (likely(queue && queue->rd_enabled) && in nvme_tcp_data_ready() 995 struct nvme_tcp_queue *queue = req->queue; in nvme_tcp_try_send_data() local 1058 struct nvme_tcp_queue *queue = req->queue; in nvme_tcp_try_send_cmd_pdu() local 1099 struct nvme_tcp_queue *queue = req->queue; in nvme_tcp_try_send_data_pdu() local 1133 struct nvme_tcp_queue *queue = req->queue; in nvme_tcp_try_send_ddgst() local [all …]
|
H A D | rdma.c | 161 return queue - queue->ctrl->queues; in nvme_rdma_queue_idx() 281 queue->qp = queue->cm_id->qp; in nvme_rdma_create_qp() 313 req->queue = queue; in nvme_rdma_init_request() 418 ib_cq_pool_put(queue->ib_cq, queue->cq_size); in nvme_rdma_free_cq() 475 queue->ib_cq = ib_alloc_cq(ibdev, queue, queue->cq_size, in nvme_rdma_create_cq() 505 queue->cq_size = cq_factor * queue->queue_size + 1; in nvme_rdma_create_queue_ib() 535 queue->queue_size, nvme_rdma_queue_idx(queue)); in nvme_rdma_create_queue_ib() 546 queue->queue_size, nvme_rdma_queue_idx(queue)); in nvme_rdma_create_queue_ib() 1937 struct nvme_rdma_queue *queue = req->queue; in nvme_rdma_complete_timed_out() local 1946 struct nvme_rdma_queue *queue = req->queue; in nvme_rdma_timeout() local [all …]
|
/openbmc/linux/drivers/net/ |
H A D | xen-netfront.c | 482 struct netfront_queue *queue = info->queue; in xennet_tx_setup_grant() local 619 .queue = queue, in xennet_xdp_xmit_one() 784 info.queue = queue; in xennet_start_xmit() 1821 if (queue->tx_irq && (queue->tx_irq == queue->rx_irq)) in xennet_disconnect_backend() 1823 if (queue->tx_irq && (queue->tx_irq != queue->rx_irq)) { in xennet_disconnect_backend() 1828 queue->tx_irq = queue->rx_irq = 0; in xennet_disconnect_backend() 1913 queue->rx_evtchn = queue->tx_evtchn; in setup_netfront_single() 1914 queue->rx_irq = queue->tx_irq = err; in setup_netfront_single() 1940 queue->tx_irq_name, queue); in setup_netfront_split() 1949 queue->rx_irq_name, queue); in setup_netfront_split() [all …]
|
/openbmc/linux/drivers/net/wireless/ralink/rt2x00/ |
H A D | rt2x00queue.c | 25 struct data_queue *queue = entry->queue; in rt2x00queue_alloc_rxskb() local 37 frame_size = queue->data_size + queue->desc_size + queue->winfo_size; in rt2x00queue_alloc_rxskb() 532 struct data_queue *queue = entry->queue; in rt2x00queue_write_tx_descriptor() local 557 queue->rt2x00dev->ops->lib->kick_queue(queue); in rt2x00queue_kick_tx_queue() 856 entry = &queue->entries[queue->index[index]]; in rt2x00queue_get_entry() 866 struct data_queue *queue = entry->queue; in rt2x00queue_index_inc() local 878 if (queue->index[index] >= queue->limit) in rt2x00queue_index_inc() 944 queue->rt2x00dev->ops->lib->kick_queue(queue); in rt2x00queue_unpause_queue() 964 queue->rt2x00dev->ops->lib->start_queue(queue); in rt2x00queue_start_queue() 983 queue->rt2x00dev->ops->lib->stop_queue(queue); in rt2x00queue_stop_queue() [all …]
|
/openbmc/linux/drivers/scsi/arm/ |
H A D | queue.c | 65 INIT_LIST_HEAD(&queue->head); in queue_initialise() 66 INIT_LIST_HEAD(&queue->free); in queue_initialise() 83 return queue->alloc != NULL; in queue_initialise() 95 kfree(queue->alloc); in queue_free() 115 if (list_empty(&queue->free)) in __queue_add() 118 l = queue->free.next; in __queue_add() 128 list_add(l, &queue->head); in __queue_add() 150 list_add(ent, &queue->free); in __queue_remove() 195 SCpnt = __queue_remove(queue, queue->head.next); in queue_remove() 247 __queue_remove(queue, l); in queue_remove_all_target() [all …]
|
/openbmc/linux/drivers/net/ethernet/ibm/ehea/ |
H A D | ehea_qmr.h | 208 return hw_qeit_calc(queue, queue->current_q_offset); in hw_qeit_get() 213 queue->current_q_offset += queue->qe_size; in hw_qeit_inc() 214 if (queue->current_q_offset >= queue->queue_length) { in hw_qeit_inc() 217 queue->toggle_state = (~queue->toggle_state) & 1; in hw_qeit_inc() 224 hw_qeit_inc(queue); in hw_qeit_get_inc() 236 hw_qeit_inc(queue); in hw_qeit_get_inc_valid() 237 pref = hw_qeit_calc(queue, queue->current_q_offset); in hw_qeit_get_inc_valid() 251 pref = hw_qeit_calc(queue, queue->current_q_offset); in hw_qeit_get_valid() 269 u64 last_entry_in_q = queue->queue_length - queue->qe_size; in hw_qeit_eq_get_inc() 273 queue->current_q_offset += queue->qe_size; in hw_qeit_eq_get_inc() [all …]
|
/openbmc/linux/drivers/crypto/hisilicon/sec/ |
H A D | sec_drv.c | 703 while (test_bit(queue->expected, queue->unprocessed)) { in sec_isr_handle() 704 clear_bit(queue->expected, queue->unprocessed); in sec_isr_handle() 708 queue->shadow[queue->expected]); in sec_isr_handle() 709 queue->shadow[queue->expected] = NULL; in sec_isr_handle() 710 queue->expected = (queue->expected + 1) % in sec_isr_handle() 735 IRQF_TRIGGER_RISING, queue->name, queue); in sec_queue_irq_init() 747 free_irq(queue->task_irq, queue); in sec_queue_irq_uninit() 907 sec_queue_cmdbase_addr(queue, queue->ring_cmd.paddr); in sec_queue_hw_init() 909 sec_queue_outorder_addr(queue, queue->ring_cq.paddr); in sec_queue_hw_init() 911 sec_queue_errbase_addr(queue, queue->ring_db.paddr); in sec_queue_hw_init() [all …]
|
/openbmc/linux/drivers/soc/ixp4xx/ |
H A D | ixp4xx-qmgr.c | 35 qmgr_queue_descs[queue], queue, val); in qmgr_put_entry() 48 qmgr_queue_descs[queue], queue, val); in qmgr_get_entry() 86 if (queue >= HALF_QUEUES) in qmgr_stat_below_low_watermark() 100 if (queue >= HALF_QUEUES) in qmgr_stat_full() 136 irq_pdevs[queue] = pdev; in qmgr_set_irq() 208 int half = queue / 32; in qmgr_enable_irq() 220 int half = queue / 32; in qmgr_disable_irq() 252 BUG_ON(queue >= QUEUES); in qmgr_request_queue() 318 qmgr_queue_descs[queue], queue, addr); in qmgr_request_queue() 355 qmgr_queue_descs[queue], queue); in qmgr_release_queue() [all …]
|
/openbmc/linux/include/drm/ |
H A D | spsc_queue.h | 50 queue->head = NULL; in spsc_queue_init() 51 atomic_long_set(&queue->tail, (long)&queue->head); in spsc_queue_init() 52 atomic_set(&queue->job_count, 0); in spsc_queue_init() 57 return queue->head; in spsc_queue_peek() 62 return atomic_read(&queue->job_count); in spsc_queue_count() 75 atomic_inc(&queue->job_count); in spsc_queue_push() 85 return tail == &queue->head; in spsc_queue_push() 96 node = READ_ONCE(queue->head); in spsc_queue_pop() 102 WRITE_ONCE(queue->head, next); in spsc_queue_pop() 107 if (atomic_long_cmpxchg(&queue->tail, in spsc_queue_pop() [all …]
|
/openbmc/linux/net/sunrpc/ |
H A D | sched.c | 140 rpc_set_waitqueue_priority(queue, queue->maxpriority); in rpc_reset_waitqueue_priority() 215 queue->qlen++; in __rpc_add_wait_queue() 240 queue->qlen--; in __rpc_remove_wait_queue() 252 queue->qlen = 0; in __rpc_init_priority_wait_queue() 622 q = &queue->tasks[queue->priority]; in __rpc_find_next_queued_priority() 624 queue->nr--; in __rpc_find_next_queued_priority() 634 q = &queue->tasks[queue->maxpriority]; in __rpc_find_next_queued_priority() 641 } while (q != &queue->tasks[queue->priority]); in __rpc_find_next_queued_priority() 647 rpc_set_waitqueue_priority(queue, (unsigned int)(q - &queue->tasks[0])); in __rpc_find_next_queued_priority() 870 if (queue) in rpc_signal_task() [all …]
|
/openbmc/linux/drivers/net/wireless/ath/ath5k/ |
H A D | qcu.c | 161 qi = &ah->ah_txq[queue]; in ath5k_hw_set_tx_queueprops() 206 unsigned int queue; in ath5k_hw_setup_tx_queue() local 231 queue = AR5K_TX_QUEUE_ID_UAPSD; in ath5k_hw_setup_tx_queue() 237 queue = AR5K_TX_QUEUE_ID_CAB; in ath5k_hw_setup_tx_queue() 264 return queue; in ath5k_hw_setup_tx_queue() 282 unsigned int queue) in ath5k_hw_set_tx_retry_limits() argument 288 if (queue > 0) in ath5k_hw_set_tx_retry_limits() 371 AR5K_QUEUE_CBRCFG(queue)); in ath5k_hw_reset_tx_queue() 402 AR5K_QUEUE_DFS_MISC(queue)); in ath5k_hw_reset_tx_queue() 407 AR5K_QUEUE_DFS_MISC(queue)); in ath5k_hw_reset_tx_queue() [all …]
|