Home
last modified time | relevance | path

Searched refs:qp (Results 1 – 25 of 381) sorted by relevance

12345678910>>...16

/openbmc/linux/drivers/infiniband/sw/rxe/
H A Drxe_qp.c156 qp->mtu = ib_mtu_enum_to_int(qp->attr.path_mtu); in rxe_qp_init_misc()
211 qp->sq.queue->buf, qp->sq.queue->buf_size, in rxe_init_sq()
247 qp->sk->sk->sk_user_data = qp; in rxe_qp_init_req()
262 qp->req.wqe_index = queue_get_producer(qp->sq.queue, in rxe_qp_init_req()
268 rxe_init_task(&qp->req.task, qp, rxe_requester); in rxe_qp_init_req()
269 rxe_init_task(&qp->comp.task, qp, rxe_completer); in rxe_qp_init_req()
302 qp->rq.queue->buf, qp->rq.queue->buf_size, in rxe_init_rq()
340 rxe_init_task(&qp->resp.task, qp, rxe_responder); in rxe_qp_init_resp()
732 qp->resp.psn = qp->attr.rq_psn; in rxe_qp_from_attr()
744 qp->req.psn = qp->attr.sq_psn; in rxe_qp_from_attr()
[all …]
H A Drxe_comp.c117 struct rxe_qp *qp = from_timer(qp, t, retrans_timer); in retransmit_timer() local
176 qp->comp.retry_cnt = qp->attr.retry_cnt; in reset_retry_counters()
177 qp->comp.rnr_retry = qp->attr.rnr_retry; in reset_retry_counters()
413 wc->qp = &qp->ibqp; in make_send_cqe()
489 if (qp->attr.sq_draining && qp->comp.psn == qp->req.psn) { in comp_check_sq_drain_done()
497 ev.element.qp = &qp->ibqp; in comp_check_sq_drain_done()
580 wc->qp = &qp->ibqp; in flush_send_wqe()
617 struct rxe_qp *qp = pkt->qp; in free_pkt() local
621 rxe_put(qp); in free_pkt()
640 psn_compare(qp->req.psn, qp->comp.psn) > 0) in reset_retry_timer()
[all …]
H A Drxe_resp.c299 qp->resp.wqe = &qp->resp.srq_wqe.wqe; in get_srq_wqe()
596 err = rxe_mr_copy(qp->resp.mr, qp->resp.va + qp->resp.offset, in write_data_in()
692 qp->resp.ack_psn = qp->resp.psn; in process_flush()
726 qp->resp.ack_psn = qp->resp.psn; in atomic_reply()
754 iova = qp->resp.va + qp->resp.offset; in atomic_write_reply()
765 qp->resp.ack_psn = qp->resp.psn; in atomic_write_reply()
796 ack->qp = qp; in prepare_ack_packet()
1051 qp->resp.ack_psn = qp->resp.psn; in execute()
1085 wc->qp = &qp->ibqp; in do_complete()
1453 wc->qp = &qp->ibqp; in flush_recv_wqe()
[all …]
H A Drxe_req.c26 qp->req.opcode = next_opcode(qp, wqe, in retry_first_write_send()
53 qp->req.psn = qp->comp.psn; in req_retry()
90 qp->mtu; in req_retry()
101 struct rxe_qp *qp = from_timer(qp, t, rnr_nak_timer); in rnr_nak_timer() local
151 ev.element.qp = &qp->ibqp; in req_check_sq_drain_done()
574 qp->req.psn = (qp->req.psn + 1) & BTH_PSN_MASK; in update_wqe_psn()
657 qp->req.wqe_index = queue_next_index(qp->sq.queue, qp->req.wqe_index); in rxe_do_local_ops()
744 psn_compare(qp->req.psn, (qp->comp.psn + in rxe_requester()
784 qp->req.psn = (qp->req.psn + 1) & BTH_PSN_MASK; in rxe_requester()
798 pkt.qp = qp; in rxe_requester()
[all …]
/openbmc/linux/drivers/infiniband/hw/qib/
H A Dqib_rc.c253 rvt_send_complete(qp, wqe, qp->s_last != qp->s_acked ? in qib_make_rc_req()
474 qp->s_len = restart_sge(&qp->s_sge, wqe, qp->s_psn, pmtu); in qib_make_rc_req()
499 if (qp->s_cur >= qp->s_size) in qib_make_rc_req()
513 qp->s_len = restart_sge(&qp->s_sge, wqe, qp->s_psn, pmtu); in qib_make_rc_req()
538 if (qp->s_cur >= qp->s_size) in qib_make_rc_req()
980 qp->s_acked = qp->s_cur; in do_rc_completion()
1881 wc.qp = &qp->ibqp; in qib_rc_rcv()
2084 qp->r_ack_psn = qp->r_psn; in qib_rc_rcv()
2096 qp->r_ack_psn = qp->r_psn; in qib_rc_rcv()
2110 qp->r_ack_psn = qp->r_psn; in qib_rc_rcv()
[all …]
H A Dqib_uc.c71 wqe = rvt_get_swqe_ptr(qp, qp->s_last); in qib_make_uc_req()
85 wqe = rvt_get_swqe_ptr(qp, qp->s_cur); in qib_make_uc_req()
125 if (++qp->s_cur >= qp->s_size) in qib_make_uc_req()
154 if (++qp->s_cur >= qp->s_size) in qib_make_uc_req()
183 if (++qp->s_cur >= qp->s_size) in qib_make_uc_req()
208 if (++qp->s_cur >= qp->s_size) in qib_make_uc_req()
214 qp->s_cur_sge = &qp->s_sge; in qib_make_uc_req()
337 qp->r_sge = qp->s_rdma_read_sge; in qib_uc_rcv()
348 qp->s_rdma_read_sge = qp->r_sge; in qib_uc_rcv()
394 wc.qp = &qp->ibqp; in qib_uc_rcv()
[all …]
H A Dqib_ruc.c49 qp->remote_ah_attr = qp->alt_ah_attr; in qib_migrate_qp()
50 qp->port_num = rdma_ah_get_port_num(&qp->alt_ah_attr); in qib_migrate_qp()
51 qp->s_pkey_index = qp->s_alt_pkey_index; in qib_migrate_qp()
54 ev.element.qp = &qp->ibqp; in qib_migrate_qp()
56 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in qib_migrate_qp()
210 struct qib_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in qib_make_ruc_header()
220 qp->s_hdrwords += in qib_make_ruc_header()
252 qib_do_send(qp); in _qib_do_send()
266 struct qib_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in qib_do_send()
304 if (qib_verbs_send(qp, priv->s_hdr, qp->s_hdrwords, in qib_do_send()
[all …]
H A Dqib_ud.c68 if (!qp) { in qib_ud_loopback()
166 rvt_copy_sge(qp, &qp->r_sge, &grh, in qib_ud_loopback()
204 wc.qp = &qp->ibqp; in qib_ud_loopback()
251 if (qp->s_last == READ_ONCE(qp->s_head)) in qib_make_ud_req()
258 wqe = rvt_get_swqe_ptr(qp, qp->s_last); in qib_make_ud_req()
264 if (qp->s_cur == READ_ONCE(qp->s_head)) in qib_make_ud_req()
267 wqe = rvt_get_swqe_ptr(qp, qp->s_cur); in qib_make_ud_req()
273 ibp = to_iport(qp->ibqp.device, qp->port_num); in qib_make_ud_req()
315 qp->s_cur_sge = &qp->s_sge; in qib_make_ud_req()
550 rvt_copy_sge(qp, &qp->r_sge, &hdr->u.l.grh, in qib_ud_rcv()
[all …]
/openbmc/linux/drivers/infiniband/hw/hfi1/
H A Drc.c906 qp->s_len = restart_sge(&qp->s_sge, wqe, qp->s_psn, pmtu); in hfi1_make_rc_req()
937 if (qp->s_cur >= qp->s_size) in hfi1_make_rc_req()
951 qp->s_len = restart_sge(&qp->s_sge, wqe, qp->s_psn, pmtu); in hfi1_make_rc_req()
977 if (qp->s_cur >= qp->s_size) in hfi1_make_rc_req()
1885 qp->s_acked = qp->s_cur; in do_rc_completion()
1894 if (qp->state == IB_QPS_SQD && qp->s_acked == qp->s_cur) in do_rc_completion()
2945 wc.qp = &qp->ibqp; in hfi1_rc_rcv()
3176 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv()
3184 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv()
3194 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv()
[all …]
H A Dqp.c449 qp = tx->qp; in iowait_sleep()
587 qp->s_last == qp->s_acked && in qp_idle()
588 qp->s_acked == qp->s_cur && in qp_idle()
589 qp->s_cur == qp->s_tail && in qp_idle()
590 qp->s_tail == qp->s_head; in qp_idle()
601 struct rvt_qp *qp = iter->qp; in qp_iter_print() local
631 qp->s_psn, qp->s_next_psn, in qp_iter_print()
634 qp->s_last, qp->s_acked, qp->s_cur, in qp_iter_print()
635 qp->s_tail, qp->s_head, qp->s_size, in qp_iter_print()
789 ev.element.qp = &qp->ibqp; in hfi1_migrate_qp()
[all …]
H A Duc.c49 wqe = rvt_get_swqe_ptr(qp, qp->s_last); in hfi1_make_uc_req()
72 wqe = rvt_get_swqe_ptr(qp, qp->s_cur); in hfi1_make_uc_req()
93 if (qp->s_last != qp->s_cur) in hfi1_make_uc_req()
95 if (++qp->s_cur == qp->s_size) in hfi1_make_uc_req()
138 if (++qp->s_cur >= qp->s_size) in hfi1_make_uc_req()
167 if (++qp->s_cur >= qp->s_size) in hfi1_make_uc_req()
197 if (++qp->s_cur >= qp->s_size) in hfi1_make_uc_req()
223 if (++qp->s_cur >= qp->s_size) in hfi1_make_uc_req()
264 struct rvt_qp *qp = packet->qp; in hfi1_uc_rcv() local
412 wc.qp = &qp->ibqp; in hfi1_uc_rcv()
[all …]
H A Druc.c31 struct rvt_qp *qp = packet->qp; in hfi1_ruc_check_hdr() local
74 hfi1_migrate_qp(qp); in hfi1_ruc_check_hdr()
168 clear_ahg(qp); in build_ahg()
171 if (qp->s_ahgidx < 0) in build_ahg()
174 qp->s_ahgpsn = npsn; in build_ahg()
282 build_ahg(qp, bth2); in hfi1_make_ruc_header_16B()
360 build_ahg(qp, bth2); in hfi1_make_ruc_header_9B()
481 hfi1_do_send(qp, true); in _hfi1_do_send()
500 ps.ibp = to_iport(qp->ibqp.device, qp->port_num); in hfi1_do_send()
512 rvt_ruc_loopback(qp); in hfi1_do_send()
[all …]
H A Dtid_rdma.c1669 req->qp = qp; in hfi1_init_trdma_req()
1709 struct rvt_qp *qp = req->qp; in hfi1_build_tid_rdma_read_packet() local
2006 qp->r_ack_psn = qp->r_psn; in tid_rdma_rcv_error()
2333 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv_tid_rdma_read_req()
2341 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv_tid_rdma_read_req()
2892 packet->qp = qp; in hfi1_handle_kdeth_eflags()
3217 s_prev = (qp->s_cur == 0 ? qp->s_size : qp->s_cur) - 1; in hfi1_tid_rdma_wqe_interlock()
3603 qp->r_ack_psn = qp->r_psn; in hfi1_tid_write_alloc_resources()
3824 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv_tid_rdma_write_req()
3832 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv_tid_rdma_write_req()
[all …]
H A Dtrace_tid.h196 TP_ARGS(qp),
226 TP_ARGS(qp)
232 TP_ARGS(qp)
238 TP_ARGS(qp)
244 TP_ARGS(qp)
250 TP_ARGS(qp)
360 __entry->qpn = qp ? qp->ibqp.qp_num : 0;
931 TP_ARGS(qp)
937 TP_ARGS(qp)
943 TP_ARGS(qp)
[all …]
H A Dqp.h51 struct hfi1_qp_priv *priv = qp->priv; in hfi1_send_ok()
55 (qp->s_flags & RVT_S_RESP_PENDING) || in hfi1_send_ok()
56 !(qp->s_flags & RVT_S_ANY_WAIT_SEND)); in hfi1_send_ok()
64 struct hfi1_qp_priv *priv = qp->priv; in clear_ahg()
68 if (priv->s_sde && qp->s_ahgidx >= 0) in clear_ahg()
70 qp->s_ahgidx = -1; in clear_ahg()
88 void hfi1_migrate_qp(struct rvt_qp *qp);
96 void notify_qp_reset(struct rvt_qp *qp);
100 void notify_error_qp(struct rvt_qp *qp);
101 void stop_send_queue(struct rvt_qp *qp);
[all …]
/openbmc/linux/drivers/infiniband/sw/rdmavt/
H A Dqp.c715 ev.element.qp = &qp->ibqp; in rvt_qp_mr_clean()
1316 wc.qp = &qp->ibqp; in rvt_error_qp()
1381 rcu_assign_pointer(rvp->qp[qp->ibqp.qp_num], qp); in rvt_insert_qp()
1542 qp->s_draining = qp->s_last != qp->s_cur; in rvt_modify_qp()
1648 ev.element.qp = &qp->ibqp; in rvt_modify_qp()
1819 wc.qp = &qp->ibqp; in rvt_post_recv()
1948 qp->ibqp.qp_num, qp->s_size, qp->s_avail, in rvt_qp_is_avail()
1949 qp->s_head, qp->s_tail, qp->s_cur, in rvt_qp_is_avail()
2308 wc.qp = &qp->ibqp; in init_sge()
2723 iter->qp = qp; in rvt_qp_iter_next()
[all …]
/openbmc/linux/drivers/ntb/
H A Dntb_transport.c472 if (!qp || !qp->link_is_up) in debugfs_read()
661 entry->qp = qp; in ntb_transport_setup_qp_mw()
953 qp->event_handler(qp->cb_data, qp->link_is_up); in ntb_qp_link_cleanup()
1154 qp->event_handler(qp->cb_data, qp->link_is_up); in ntb_qp_link_work()
1224 qp->debugfs_dir, qp, in ntb_transport_init_queue()
1470 qp->rx_handler(qp, qp->cb_data, cb_data, len); in ntb_complete_rxc()
1627 offset = qp->rx_buff + qp->rx_max_frame * qp->rx_index; in ntb_process_rxc()
1781 qp->tx_handler(qp, qp->cb_data, entry->cb_data, in ntb_tx_copy_callback()
1911 qp->tx_handler(qp, qp->cb_data, NULL, -EIO); in ntb_process_tx()
2063 entry->qp = qp; in ntb_transport_create_queue()
[all …]
/openbmc/linux/drivers/infiniband/sw/siw/
H A Dsiw_qp.c171 if (qp->cep) { in siw_qp_llp_close()
173 qp->cep = NULL; in siw_qp_llp_close()
371 memset(&qp->term_info, 0, sizeof(qp->term_info)); in siw_init_terminate()
688 qp->attrs.orq_size, qp->attrs.irq_size); in siw_qp_nextstate_from_idle()
929 if (qp->attrs.orq_size && !siw_orq_empty(qp)) { in siw_activate_tx_from_sq()
965 qp->sq_get++; in siw_activate_tx_from_sq()
988 irqe = &qp->irq[qp->irq_get % qp->attrs.irq_size]; in siw_activate_tx()
1201 sqe = &qp->orq[qp->orq_get % qp->attrs.orq_size]; in siw_sq_flush()
1237 sqe = &qp->sendq[qp->sq_get % qp->attrs.sq_size]; in siw_sq_flush()
1301 &qp->recvq[qp->rq_get % qp->attrs.rq_size]; in siw_rq_flush()
[all …]
/openbmc/linux/drivers/infiniband/hw/mthca/
H A Dmthca_qp.c257 qp->port = qp->alt_port; in mthca_qp_event()
261 event.element.qp = &qp->ibqp; in mthca_qp_event()
841 qp->sq.last = get_send_wqe(qp, qp->sq.max - 1); in __mthca_modify_qp()
844 qp->rq.last = get_recv_wqe(qp, qp->rq.max - 1); in __mthca_modify_qp()
1052 qp->send_wqe_offset = ALIGN(qp->rq.max << qp->rq.wqe_shift, in mthca_alloc_wqe_buf()
1072 &qp->queue, &qp->is_direct, pd, 0, &qp->mr); in mthca_alloc_wqe_buf()
1088 &qp->queue, qp->is_direct, &qp->mr); in mthca_free_wqe_buf()
1138 qp->qpn, &qp->rq.db); in mthca_alloc_memfree()
1143 qp->qpn, &qp->sq.db); in mthca_alloc_memfree()
1246 qp->sq.last = get_send_wqe(qp, qp->sq.max - 1); in mthca_alloc_qp_common()
[all …]
/openbmc/linux/drivers/net/ethernet/qlogic/qed/
H A Dqed_roce.c210 if (qp->vlan_id) { in qed_roce_get_qp_tc()
217 qp->icid, tc, qp->vlan_id ? "enabled" : "disabled"); in qed_roce_get_qp_tc()
244 if (!qp->irq) { in qed_roce_sp_create_responder()
343 qp->cq_prod = 0; in qed_roce_sp_create_responder()
355 qp->irq, qp->irq_phys_addr); in qed_roce_sp_create_responder()
382 if (!qp->orq) { in qed_roce_sp_create_requester()
486 qp->orq, qp->orq_phys_addr); in qed_roce_sp_create_requester()
736 qp->irq, qp->irq_phys_addr); in qed_roce_sp_destroy_qp_responder()
798 qp->orq, qp->orq_phys_addr); in qed_roce_sp_destroy_qp_requester()
828 if ((!(qp->resp_offloaded)) && (!(qp->req_offloaded))) { in qed_roce_query_qp()
[all …]
/openbmc/linux/drivers/infiniband/hw/mlx4/
H A Dqp.c181 return get_wqe(qp, qp->rq.offset + (n << qp->rq.wqe_shift)); in get_recv_wqe()
186 return get_wqe(qp, qp->sq.offset + (n << qp->sq.wqe_shift)); in get_send_wqe()
274 qpe_work->qp = qp; in mlx4_ib_qp_event()
425 qp->buf_size = (qp->rq.wqe_cnt << qp->rq.wqe_shift) + in set_kernel_sq_size()
429 qp->sq.offset = qp->rq.wqe_cnt << qp->rq.wqe_shift; in set_kernel_sq_size()
431 qp->rq.offset = qp->sq.wqe_cnt << qp->sq.wqe_shift; in set_kernel_sq_size()
464 qp->buf_size = (qp->rq.wqe_cnt << qp->rq.wqe_shift) + in set_user_sq_size()
774 qp->ibqp.qp_num = qp->mqp.qpn; in _mlx4_ib_create_qp_rss()
918 qp->buf_size = (qp->rq.wqe_cnt << qp->rq.wqe_shift) + in create_rq()
1412 if (qp->pri.smac || (!qp->pri.smac && qp->pri.smac_port)) { in destroy_qp_common()
[all …]
/openbmc/linux/drivers/infiniband/hw/erdma/
H A Derdma_qp.c33 if (qp->cep) { in erdma_qp_llp_close()
35 qp->cep = NULL; in erdma_qp_llp_close()
45 if (qp) in erdma_get_ibqp()
46 return &qp->ibqp; in erdma_get_ibqp()
140 if (qp->cep) { in erdma_modify_qp_internal()
182 mod_delayed_work(qp->dev->reflush_wq, &qp->reflush_dwork, in erdma_modify_qp_internal()
515 if ((u16)(sq_pi - qp->kern_qp.sq_ci) >= qp->attrs.sq_size) { in erdma_post_send()
534 mod_delayed_work(qp->dev->reflush_wq, &qp->reflush_dwork, in erdma_post_send()
544 get_queue_entry(qp->kern_qp.rq_buf, qp->kern_qp.rq_pi, in erdma_post_recv_one()
563 qp->kern_qp.rwr_tbl[qp->kern_qp.rq_pi & (qp->attrs.rq_size - 1)] = in erdma_post_recv_one()
[all …]
/openbmc/linux/drivers/infiniband/hw/vmw_pvrdma/
H A Dpvrdma_qp.c150 qp->npages_recv = (qp->rq.wqe_cnt * qp->rq.wqe_size + PAGE_SIZE - 1) / in pvrdma_set_rq_size()
300 qp->npages = qp->npages_send + qp->npages_recv; in pvrdma_create_qp()
312 qp->npages = qp->npages_send + qp->npages_recv; in pvrdma_create_qp()
318 qp->rq.offset = qp->npages_send * PAGE_SIZE; in pvrdma_create_qp()
344 qp->sq.ring = qp->pdir.pages[0]; in pvrdma_create_qp()
345 qp->rq.ring = is_srq ? NULL : &qp->sq.ring[1]; in pvrdma_create_qp()
542 qp->state; in pvrdma_modify_qp()
637 qp->sq.offset + n * qp->sq.wqe_size); in get_sq_wqe()
643 qp->rq.offset + n * qp->rq.wqe_size); in get_rq_wqe()
696 qp->sq.ring, qp->sq.wqe_cnt, &tail))) { in pvrdma_post_send()
[all …]
/openbmc/linux/drivers/infiniband/hw/irdma/
H A Duk.c134 writel(qp->qp_id, qp->wqe_alloc_db); in irdma_uk_qp_post_wr()
138 writel(qp->qp_id, qp->wqe_alloc_db); in irdma_uk_qp_post_wr()
143 qp->initial_ring.head = qp->sq_ring.head; in irdma_uk_qp_post_wr()
184 qp->swqe_polarity = !qp->swqe_polarity; in irdma_qp_get_next_send_wqe()
191 wqe_0 = qp->sq_base[IRDMA_RING_CURRENT_HEAD(qp->sq_ring)].elem; in irdma_qp_get_next_send_wqe()
219 qp->rwqe_polarity = !qp->rwqe_polarity; in irdma_qp_get_next_recv_wqe()
221 wqe = qp->rq_base[*wqe_idx * qp->rq_wqe_size_multiplier].elem; in irdma_qp_get_next_recv_wqe()
1080 if (!qp || qp->destroy_pending) { in irdma_uk_cq_poll_cmpl()
1100 info->wr_id = qp->rq_wrid_array[qp->rq_ring.tail]; in irdma_uk_cq_poll_cmpl()
1128 if (wqe_idx < qp->conn_wqes && qp->sq_ring.head == qp->sq_ring.tail) { in irdma_uk_cq_poll_cmpl()
[all …]
/openbmc/linux/net/ipv4/
H A Dip_fragment.c90 qp->ecn = 0; in ip4_frag_init()
98 struct ipq *qp; in ip4_frag_free() local
101 if (qp->peer) in ip4_frag_free()
157 ipq_kill(qp); in ip_expire()
199 ipq_put(qp); in ip_expire()
268 qp->iif = 0; in ip_frag_reinit()
269 qp->ecn = 0; in ip_frag_reinit()
375 qp->q.meat == qp->q.len) { in ip_frag_queue()
421 ipq_kill(qp); in ip_frag_reasm()
457 if (qp->max_df_size == qp->q.max_size) { in ip_frag_reasm()
[all …]

12345678910>>...16