/openbmc/linux/drivers/infiniband/sw/rxe/ |
H A D | rxe_qp.c | 156 qp->mtu = ib_mtu_enum_to_int(qp->attr.path_mtu); in rxe_qp_init_misc() 211 qp->sq.queue->buf, qp->sq.queue->buf_size, in rxe_init_sq() 247 qp->sk->sk->sk_user_data = qp; in rxe_qp_init_req() 262 qp->req.wqe_index = queue_get_producer(qp->sq.queue, in rxe_qp_init_req() 268 rxe_init_task(&qp->req.task, qp, rxe_requester); in rxe_qp_init_req() 269 rxe_init_task(&qp->comp.task, qp, rxe_completer); in rxe_qp_init_req() 302 qp->rq.queue->buf, qp->rq.queue->buf_size, in rxe_init_rq() 340 rxe_init_task(&qp->resp.task, qp, rxe_responder); in rxe_qp_init_resp() 732 qp->resp.psn = qp->attr.rq_psn; in rxe_qp_from_attr() 744 qp->req.psn = qp->attr.sq_psn; in rxe_qp_from_attr() [all …]
|
H A D | rxe_comp.c | 117 struct rxe_qp *qp = from_timer(qp, t, retrans_timer); in retransmit_timer() local 176 qp->comp.retry_cnt = qp->attr.retry_cnt; in reset_retry_counters() 177 qp->comp.rnr_retry = qp->attr.rnr_retry; in reset_retry_counters() 413 wc->qp = &qp->ibqp; in make_send_cqe() 489 if (qp->attr.sq_draining && qp->comp.psn == qp->req.psn) { in comp_check_sq_drain_done() 497 ev.element.qp = &qp->ibqp; in comp_check_sq_drain_done() 580 wc->qp = &qp->ibqp; in flush_send_wqe() 617 struct rxe_qp *qp = pkt->qp; in free_pkt() local 621 rxe_put(qp); in free_pkt() 640 psn_compare(qp->req.psn, qp->comp.psn) > 0) in reset_retry_timer() [all …]
|
H A D | rxe_resp.c | 299 qp->resp.wqe = &qp->resp.srq_wqe.wqe; in get_srq_wqe() 596 err = rxe_mr_copy(qp->resp.mr, qp->resp.va + qp->resp.offset, in write_data_in() 692 qp->resp.ack_psn = qp->resp.psn; in process_flush() 726 qp->resp.ack_psn = qp->resp.psn; in atomic_reply() 754 iova = qp->resp.va + qp->resp.offset; in atomic_write_reply() 765 qp->resp.ack_psn = qp->resp.psn; in atomic_write_reply() 796 ack->qp = qp; in prepare_ack_packet() 1051 qp->resp.ack_psn = qp->resp.psn; in execute() 1085 wc->qp = &qp->ibqp; in do_complete() 1453 wc->qp = &qp->ibqp; in flush_recv_wqe() [all …]
|
H A D | rxe_req.c | 26 qp->req.opcode = next_opcode(qp, wqe, in retry_first_write_send() 53 qp->req.psn = qp->comp.psn; in req_retry() 90 qp->mtu; in req_retry() 101 struct rxe_qp *qp = from_timer(qp, t, rnr_nak_timer); in rnr_nak_timer() local 151 ev.element.qp = &qp->ibqp; in req_check_sq_drain_done() 574 qp->req.psn = (qp->req.psn + 1) & BTH_PSN_MASK; in update_wqe_psn() 657 qp->req.wqe_index = queue_next_index(qp->sq.queue, qp->req.wqe_index); in rxe_do_local_ops() 744 psn_compare(qp->req.psn, (qp->comp.psn + in rxe_requester() 784 qp->req.psn = (qp->req.psn + 1) & BTH_PSN_MASK; in rxe_requester() 798 pkt.qp = qp; in rxe_requester() [all …]
|
/openbmc/linux/drivers/infiniband/hw/qib/ |
H A D | qib_rc.c | 253 rvt_send_complete(qp, wqe, qp->s_last != qp->s_acked ? in qib_make_rc_req() 474 qp->s_len = restart_sge(&qp->s_sge, wqe, qp->s_psn, pmtu); in qib_make_rc_req() 499 if (qp->s_cur >= qp->s_size) in qib_make_rc_req() 513 qp->s_len = restart_sge(&qp->s_sge, wqe, qp->s_psn, pmtu); in qib_make_rc_req() 538 if (qp->s_cur >= qp->s_size) in qib_make_rc_req() 980 qp->s_acked = qp->s_cur; in do_rc_completion() 1881 wc.qp = &qp->ibqp; in qib_rc_rcv() 2084 qp->r_ack_psn = qp->r_psn; in qib_rc_rcv() 2096 qp->r_ack_psn = qp->r_psn; in qib_rc_rcv() 2110 qp->r_ack_psn = qp->r_psn; in qib_rc_rcv() [all …]
|
H A D | qib_uc.c | 71 wqe = rvt_get_swqe_ptr(qp, qp->s_last); in qib_make_uc_req() 85 wqe = rvt_get_swqe_ptr(qp, qp->s_cur); in qib_make_uc_req() 125 if (++qp->s_cur >= qp->s_size) in qib_make_uc_req() 154 if (++qp->s_cur >= qp->s_size) in qib_make_uc_req() 183 if (++qp->s_cur >= qp->s_size) in qib_make_uc_req() 208 if (++qp->s_cur >= qp->s_size) in qib_make_uc_req() 214 qp->s_cur_sge = &qp->s_sge; in qib_make_uc_req() 337 qp->r_sge = qp->s_rdma_read_sge; in qib_uc_rcv() 348 qp->s_rdma_read_sge = qp->r_sge; in qib_uc_rcv() 394 wc.qp = &qp->ibqp; in qib_uc_rcv() [all …]
|
H A D | qib_ruc.c | 49 qp->remote_ah_attr = qp->alt_ah_attr; in qib_migrate_qp() 50 qp->port_num = rdma_ah_get_port_num(&qp->alt_ah_attr); in qib_migrate_qp() 51 qp->s_pkey_index = qp->s_alt_pkey_index; in qib_migrate_qp() 54 ev.element.qp = &qp->ibqp; in qib_migrate_qp() 56 qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); in qib_migrate_qp() 210 struct qib_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in qib_make_ruc_header() 220 qp->s_hdrwords += in qib_make_ruc_header() 252 qib_do_send(qp); in _qib_do_send() 266 struct qib_ibport *ibp = to_iport(qp->ibqp.device, qp->port_num); in qib_do_send() 304 if (qib_verbs_send(qp, priv->s_hdr, qp->s_hdrwords, in qib_do_send() [all …]
|
H A D | qib_ud.c | 68 if (!qp) { in qib_ud_loopback() 166 rvt_copy_sge(qp, &qp->r_sge, &grh, in qib_ud_loopback() 204 wc.qp = &qp->ibqp; in qib_ud_loopback() 251 if (qp->s_last == READ_ONCE(qp->s_head)) in qib_make_ud_req() 258 wqe = rvt_get_swqe_ptr(qp, qp->s_last); in qib_make_ud_req() 264 if (qp->s_cur == READ_ONCE(qp->s_head)) in qib_make_ud_req() 267 wqe = rvt_get_swqe_ptr(qp, qp->s_cur); in qib_make_ud_req() 273 ibp = to_iport(qp->ibqp.device, qp->port_num); in qib_make_ud_req() 315 qp->s_cur_sge = &qp->s_sge; in qib_make_ud_req() 550 rvt_copy_sge(qp, &qp->r_sge, &hdr->u.l.grh, in qib_ud_rcv() [all …]
|
/openbmc/linux/drivers/infiniband/hw/hfi1/ |
H A D | rc.c | 906 qp->s_len = restart_sge(&qp->s_sge, wqe, qp->s_psn, pmtu); in hfi1_make_rc_req() 937 if (qp->s_cur >= qp->s_size) in hfi1_make_rc_req() 951 qp->s_len = restart_sge(&qp->s_sge, wqe, qp->s_psn, pmtu); in hfi1_make_rc_req() 977 if (qp->s_cur >= qp->s_size) in hfi1_make_rc_req() 1885 qp->s_acked = qp->s_cur; in do_rc_completion() 1894 if (qp->state == IB_QPS_SQD && qp->s_acked == qp->s_cur) in do_rc_completion() 2945 wc.qp = &qp->ibqp; in hfi1_rc_rcv() 3176 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv() 3184 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv() 3194 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv() [all …]
|
H A D | qp.c | 449 qp = tx->qp; in iowait_sleep() 587 qp->s_last == qp->s_acked && in qp_idle() 588 qp->s_acked == qp->s_cur && in qp_idle() 589 qp->s_cur == qp->s_tail && in qp_idle() 590 qp->s_tail == qp->s_head; in qp_idle() 601 struct rvt_qp *qp = iter->qp; in qp_iter_print() local 631 qp->s_psn, qp->s_next_psn, in qp_iter_print() 634 qp->s_last, qp->s_acked, qp->s_cur, in qp_iter_print() 635 qp->s_tail, qp->s_head, qp->s_size, in qp_iter_print() 789 ev.element.qp = &qp->ibqp; in hfi1_migrate_qp() [all …]
|
H A D | uc.c | 49 wqe = rvt_get_swqe_ptr(qp, qp->s_last); in hfi1_make_uc_req() 72 wqe = rvt_get_swqe_ptr(qp, qp->s_cur); in hfi1_make_uc_req() 93 if (qp->s_last != qp->s_cur) in hfi1_make_uc_req() 95 if (++qp->s_cur == qp->s_size) in hfi1_make_uc_req() 138 if (++qp->s_cur >= qp->s_size) in hfi1_make_uc_req() 167 if (++qp->s_cur >= qp->s_size) in hfi1_make_uc_req() 197 if (++qp->s_cur >= qp->s_size) in hfi1_make_uc_req() 223 if (++qp->s_cur >= qp->s_size) in hfi1_make_uc_req() 264 struct rvt_qp *qp = packet->qp; in hfi1_uc_rcv() local 412 wc.qp = &qp->ibqp; in hfi1_uc_rcv() [all …]
|
H A D | ruc.c | 31 struct rvt_qp *qp = packet->qp; in hfi1_ruc_check_hdr() local 74 hfi1_migrate_qp(qp); in hfi1_ruc_check_hdr() 168 clear_ahg(qp); in build_ahg() 171 if (qp->s_ahgidx < 0) in build_ahg() 174 qp->s_ahgpsn = npsn; in build_ahg() 282 build_ahg(qp, bth2); in hfi1_make_ruc_header_16B() 360 build_ahg(qp, bth2); in hfi1_make_ruc_header_9B() 481 hfi1_do_send(qp, true); in _hfi1_do_send() 500 ps.ibp = to_iport(qp->ibqp.device, qp->port_num); in hfi1_do_send() 512 rvt_ruc_loopback(qp); in hfi1_do_send() [all …]
|
H A D | tid_rdma.c | 1669 req->qp = qp; in hfi1_init_trdma_req() 1709 struct rvt_qp *qp = req->qp; in hfi1_build_tid_rdma_read_packet() local 2006 qp->r_ack_psn = qp->r_psn; in tid_rdma_rcv_error() 2333 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv_tid_rdma_read_req() 2341 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv_tid_rdma_read_req() 2892 packet->qp = qp; in hfi1_handle_kdeth_eflags() 3217 s_prev = (qp->s_cur == 0 ? qp->s_size : qp->s_cur) - 1; in hfi1_tid_rdma_wqe_interlock() 3603 qp->r_ack_psn = qp->r_psn; in hfi1_tid_write_alloc_resources() 3824 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv_tid_rdma_write_req() 3832 qp->r_ack_psn = qp->r_psn; in hfi1_rc_rcv_tid_rdma_write_req() [all …]
|
H A D | trace_tid.h | 196 TP_ARGS(qp), 226 TP_ARGS(qp) 232 TP_ARGS(qp) 238 TP_ARGS(qp) 244 TP_ARGS(qp) 250 TP_ARGS(qp) 360 __entry->qpn = qp ? qp->ibqp.qp_num : 0; 931 TP_ARGS(qp) 937 TP_ARGS(qp) 943 TP_ARGS(qp) [all …]
|
H A D | qp.h | 51 struct hfi1_qp_priv *priv = qp->priv; in hfi1_send_ok() 55 (qp->s_flags & RVT_S_RESP_PENDING) || in hfi1_send_ok() 56 !(qp->s_flags & RVT_S_ANY_WAIT_SEND)); in hfi1_send_ok() 64 struct hfi1_qp_priv *priv = qp->priv; in clear_ahg() 68 if (priv->s_sde && qp->s_ahgidx >= 0) in clear_ahg() 70 qp->s_ahgidx = -1; in clear_ahg() 88 void hfi1_migrate_qp(struct rvt_qp *qp); 96 void notify_qp_reset(struct rvt_qp *qp); 100 void notify_error_qp(struct rvt_qp *qp); 101 void stop_send_queue(struct rvt_qp *qp); [all …]
|
/openbmc/linux/drivers/infiniband/sw/rdmavt/ |
H A D | qp.c | 715 ev.element.qp = &qp->ibqp; in rvt_qp_mr_clean() 1316 wc.qp = &qp->ibqp; in rvt_error_qp() 1381 rcu_assign_pointer(rvp->qp[qp->ibqp.qp_num], qp); in rvt_insert_qp() 1542 qp->s_draining = qp->s_last != qp->s_cur; in rvt_modify_qp() 1648 ev.element.qp = &qp->ibqp; in rvt_modify_qp() 1819 wc.qp = &qp->ibqp; in rvt_post_recv() 1948 qp->ibqp.qp_num, qp->s_size, qp->s_avail, in rvt_qp_is_avail() 1949 qp->s_head, qp->s_tail, qp->s_cur, in rvt_qp_is_avail() 2308 wc.qp = &qp->ibqp; in init_sge() 2723 iter->qp = qp; in rvt_qp_iter_next() [all …]
|
/openbmc/linux/drivers/ntb/ |
H A D | ntb_transport.c | 472 if (!qp || !qp->link_is_up) in debugfs_read() 661 entry->qp = qp; in ntb_transport_setup_qp_mw() 953 qp->event_handler(qp->cb_data, qp->link_is_up); in ntb_qp_link_cleanup() 1154 qp->event_handler(qp->cb_data, qp->link_is_up); in ntb_qp_link_work() 1224 qp->debugfs_dir, qp, in ntb_transport_init_queue() 1470 qp->rx_handler(qp, qp->cb_data, cb_data, len); in ntb_complete_rxc() 1627 offset = qp->rx_buff + qp->rx_max_frame * qp->rx_index; in ntb_process_rxc() 1781 qp->tx_handler(qp, qp->cb_data, entry->cb_data, in ntb_tx_copy_callback() 1911 qp->tx_handler(qp, qp->cb_data, NULL, -EIO); in ntb_process_tx() 2063 entry->qp = qp; in ntb_transport_create_queue() [all …]
|
/openbmc/linux/drivers/infiniband/sw/siw/ |
H A D | siw_qp.c | 171 if (qp->cep) { in siw_qp_llp_close() 173 qp->cep = NULL; in siw_qp_llp_close() 371 memset(&qp->term_info, 0, sizeof(qp->term_info)); in siw_init_terminate() 688 qp->attrs.orq_size, qp->attrs.irq_size); in siw_qp_nextstate_from_idle() 929 if (qp->attrs.orq_size && !siw_orq_empty(qp)) { in siw_activate_tx_from_sq() 965 qp->sq_get++; in siw_activate_tx_from_sq() 988 irqe = &qp->irq[qp->irq_get % qp->attrs.irq_size]; in siw_activate_tx() 1201 sqe = &qp->orq[qp->orq_get % qp->attrs.orq_size]; in siw_sq_flush() 1237 sqe = &qp->sendq[qp->sq_get % qp->attrs.sq_size]; in siw_sq_flush() 1301 &qp->recvq[qp->rq_get % qp->attrs.rq_size]; in siw_rq_flush() [all …]
|
/openbmc/linux/drivers/infiniband/hw/mthca/ |
H A D | mthca_qp.c | 257 qp->port = qp->alt_port; in mthca_qp_event() 261 event.element.qp = &qp->ibqp; in mthca_qp_event() 841 qp->sq.last = get_send_wqe(qp, qp->sq.max - 1); in __mthca_modify_qp() 844 qp->rq.last = get_recv_wqe(qp, qp->rq.max - 1); in __mthca_modify_qp() 1052 qp->send_wqe_offset = ALIGN(qp->rq.max << qp->rq.wqe_shift, in mthca_alloc_wqe_buf() 1072 &qp->queue, &qp->is_direct, pd, 0, &qp->mr); in mthca_alloc_wqe_buf() 1088 &qp->queue, qp->is_direct, &qp->mr); in mthca_free_wqe_buf() 1138 qp->qpn, &qp->rq.db); in mthca_alloc_memfree() 1143 qp->qpn, &qp->sq.db); in mthca_alloc_memfree() 1246 qp->sq.last = get_send_wqe(qp, qp->sq.max - 1); in mthca_alloc_qp_common() [all …]
|
/openbmc/linux/drivers/net/ethernet/qlogic/qed/ |
H A D | qed_roce.c | 210 if (qp->vlan_id) { in qed_roce_get_qp_tc() 217 qp->icid, tc, qp->vlan_id ? "enabled" : "disabled"); in qed_roce_get_qp_tc() 244 if (!qp->irq) { in qed_roce_sp_create_responder() 343 qp->cq_prod = 0; in qed_roce_sp_create_responder() 355 qp->irq, qp->irq_phys_addr); in qed_roce_sp_create_responder() 382 if (!qp->orq) { in qed_roce_sp_create_requester() 486 qp->orq, qp->orq_phys_addr); in qed_roce_sp_create_requester() 736 qp->irq, qp->irq_phys_addr); in qed_roce_sp_destroy_qp_responder() 798 qp->orq, qp->orq_phys_addr); in qed_roce_sp_destroy_qp_requester() 828 if ((!(qp->resp_offloaded)) && (!(qp->req_offloaded))) { in qed_roce_query_qp() [all …]
|
/openbmc/linux/drivers/infiniband/hw/mlx4/ |
H A D | qp.c | 181 return get_wqe(qp, qp->rq.offset + (n << qp->rq.wqe_shift)); in get_recv_wqe() 186 return get_wqe(qp, qp->sq.offset + (n << qp->sq.wqe_shift)); in get_send_wqe() 274 qpe_work->qp = qp; in mlx4_ib_qp_event() 425 qp->buf_size = (qp->rq.wqe_cnt << qp->rq.wqe_shift) + in set_kernel_sq_size() 429 qp->sq.offset = qp->rq.wqe_cnt << qp->rq.wqe_shift; in set_kernel_sq_size() 431 qp->rq.offset = qp->sq.wqe_cnt << qp->sq.wqe_shift; in set_kernel_sq_size() 464 qp->buf_size = (qp->rq.wqe_cnt << qp->rq.wqe_shift) + in set_user_sq_size() 774 qp->ibqp.qp_num = qp->mqp.qpn; in _mlx4_ib_create_qp_rss() 918 qp->buf_size = (qp->rq.wqe_cnt << qp->rq.wqe_shift) + in create_rq() 1412 if (qp->pri.smac || (!qp->pri.smac && qp->pri.smac_port)) { in destroy_qp_common() [all …]
|
/openbmc/linux/drivers/infiniband/hw/erdma/ |
H A D | erdma_qp.c | 33 if (qp->cep) { in erdma_qp_llp_close() 35 qp->cep = NULL; in erdma_qp_llp_close() 45 if (qp) in erdma_get_ibqp() 46 return &qp->ibqp; in erdma_get_ibqp() 140 if (qp->cep) { in erdma_modify_qp_internal() 182 mod_delayed_work(qp->dev->reflush_wq, &qp->reflush_dwork, in erdma_modify_qp_internal() 515 if ((u16)(sq_pi - qp->kern_qp.sq_ci) >= qp->attrs.sq_size) { in erdma_post_send() 534 mod_delayed_work(qp->dev->reflush_wq, &qp->reflush_dwork, in erdma_post_send() 544 get_queue_entry(qp->kern_qp.rq_buf, qp->kern_qp.rq_pi, in erdma_post_recv_one() 563 qp->kern_qp.rwr_tbl[qp->kern_qp.rq_pi & (qp->attrs.rq_size - 1)] = in erdma_post_recv_one() [all …]
|
/openbmc/linux/drivers/infiniband/hw/vmw_pvrdma/ |
H A D | pvrdma_qp.c | 150 qp->npages_recv = (qp->rq.wqe_cnt * qp->rq.wqe_size + PAGE_SIZE - 1) / in pvrdma_set_rq_size() 300 qp->npages = qp->npages_send + qp->npages_recv; in pvrdma_create_qp() 312 qp->npages = qp->npages_send + qp->npages_recv; in pvrdma_create_qp() 318 qp->rq.offset = qp->npages_send * PAGE_SIZE; in pvrdma_create_qp() 344 qp->sq.ring = qp->pdir.pages[0]; in pvrdma_create_qp() 345 qp->rq.ring = is_srq ? NULL : &qp->sq.ring[1]; in pvrdma_create_qp() 542 qp->state; in pvrdma_modify_qp() 637 qp->sq.offset + n * qp->sq.wqe_size); in get_sq_wqe() 643 qp->rq.offset + n * qp->rq.wqe_size); in get_rq_wqe() 696 qp->sq.ring, qp->sq.wqe_cnt, &tail))) { in pvrdma_post_send() [all …]
|
/openbmc/linux/drivers/infiniband/hw/irdma/ |
H A D | uk.c | 134 writel(qp->qp_id, qp->wqe_alloc_db); in irdma_uk_qp_post_wr() 138 writel(qp->qp_id, qp->wqe_alloc_db); in irdma_uk_qp_post_wr() 143 qp->initial_ring.head = qp->sq_ring.head; in irdma_uk_qp_post_wr() 184 qp->swqe_polarity = !qp->swqe_polarity; in irdma_qp_get_next_send_wqe() 191 wqe_0 = qp->sq_base[IRDMA_RING_CURRENT_HEAD(qp->sq_ring)].elem; in irdma_qp_get_next_send_wqe() 219 qp->rwqe_polarity = !qp->rwqe_polarity; in irdma_qp_get_next_recv_wqe() 221 wqe = qp->rq_base[*wqe_idx * qp->rq_wqe_size_multiplier].elem; in irdma_qp_get_next_recv_wqe() 1080 if (!qp || qp->destroy_pending) { in irdma_uk_cq_poll_cmpl() 1100 info->wr_id = qp->rq_wrid_array[qp->rq_ring.tail]; in irdma_uk_cq_poll_cmpl() 1128 if (wqe_idx < qp->conn_wqes && qp->sq_ring.head == qp->sq_ring.tail) { in irdma_uk_cq_poll_cmpl() [all …]
|
/openbmc/linux/net/ipv4/ |
H A D | ip_fragment.c | 90 qp->ecn = 0; in ip4_frag_init() 98 struct ipq *qp; in ip4_frag_free() local 101 if (qp->peer) in ip4_frag_free() 157 ipq_kill(qp); in ip_expire() 199 ipq_put(qp); in ip_expire() 268 qp->iif = 0; in ip_frag_reinit() 269 qp->ecn = 0; in ip_frag_reinit() 375 qp->q.meat == qp->q.len) { in ip_frag_queue() 421 ipq_kill(qp); in ip_frag_reasm() 457 if (qp->max_df_size == qp->q.max_size) { in ip_frag_reasm() [all …]
|