xref: /openbmc/linux/drivers/infiniband/sw/rxe/rxe.h (revision 2a954832)
1 /* SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB */
2 /*
3  * Copyright (c) 2016 Mellanox Technologies Ltd. All rights reserved.
4  * Copyright (c) 2015 System Fabric Works, Inc. All rights reserved.
5  */
6 
7 #ifndef RXE_H
8 #define RXE_H
9 
10 #ifdef pr_fmt
11 #undef pr_fmt
12 #endif
13 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
14 
15 #include <linux/skbuff.h>
16 
17 #include <rdma/ib_verbs.h>
18 #include <rdma/ib_user_verbs.h>
19 #include <rdma/ib_pack.h>
20 #include <rdma/ib_smi.h>
21 #include <rdma/ib_umem.h>
22 #include <rdma/ib_cache.h>
23 #include <rdma/ib_addr.h>
24 #include <crypto/hash.h>
25 
26 #include "rxe_net.h"
27 #include "rxe_opcode.h"
28 #include "rxe_hdr.h"
29 #include "rxe_param.h"
30 #include "rxe_verbs.h"
31 #include "rxe_loc.h"
32 
33 /*
34  * Version 1 and Version 2 are identical on 64 bit machines, but on 32 bit
35  * machines Version 2 has a different struct layout.
36  */
37 #define RXE_UVERBS_ABI_VERSION		2
38 
39 #define RXE_ROCE_V2_SPORT		(0xc000)
40 
41 #define rxe_dbg(fmt, ...) pr_debug("%s: " fmt "\n", __func__, ##__VA_ARGS__)
42 #define rxe_dbg_dev(rxe, fmt, ...) ibdev_dbg(&(rxe)->ib_dev,		\
43 		"%s: " fmt, __func__, ##__VA_ARGS__)
44 #define rxe_dbg_uc(uc, fmt, ...) ibdev_dbg((uc)->ibuc.device,		\
45 		"uc#%d %s: " fmt, (uc)->elem.index, __func__, ##__VA_ARGS__)
46 #define rxe_dbg_pd(pd, fmt, ...) ibdev_dbg((pd)->ibpd.device,		\
47 		"pd#%d %s: " fmt, (pd)->elem.index, __func__, ##__VA_ARGS__)
48 #define rxe_dbg_ah(ah, fmt, ...) ibdev_dbg((ah)->ibah.device,		\
49 		"ah#%d %s: " fmt, (ah)->elem.index, __func__, ##__VA_ARGS__)
50 #define rxe_dbg_srq(srq, fmt, ...) ibdev_dbg((srq)->ibsrq.device,	\
51 		"srq#%d %s: " fmt, (srq)->elem.index, __func__, ##__VA_ARGS__)
52 #define rxe_dbg_qp(qp, fmt, ...) ibdev_dbg((qp)->ibqp.device,		\
53 		"qp#%d %s: " fmt, (qp)->elem.index, __func__, ##__VA_ARGS__)
54 #define rxe_dbg_cq(cq, fmt, ...) ibdev_dbg((cq)->ibcq.device,		\
55 		"cq#%d %s: " fmt, (cq)->elem.index, __func__, ##__VA_ARGS__)
56 #define rxe_dbg_mr(mr, fmt, ...) ibdev_dbg((mr)->ibmr.device,		\
57 		"mr#%d %s:  " fmt, (mr)->elem.index, __func__, ##__VA_ARGS__)
58 #define rxe_dbg_mw(mw, fmt, ...) ibdev_dbg((mw)->ibmw.device,		\
59 		"mw#%d %s:  " fmt, (mw)->elem.index, __func__, ##__VA_ARGS__)
60 
61 #define rxe_err(fmt, ...) pr_err_ratelimited("%s: " fmt "\n", __func__, \
62 					##__VA_ARGS__)
63 #define rxe_err_dev(rxe, fmt, ...) ibdev_err_ratelimited(&(rxe)->ib_dev, \
64 		"%s: " fmt, __func__, ##__VA_ARGS__)
65 #define rxe_err_uc(uc, fmt, ...) ibdev_err_ratelimited((uc)->ibuc.device, \
66 		"uc#%d %s: " fmt, (uc)->elem.index, __func__, ##__VA_ARGS__)
67 #define rxe_err_pd(pd, fmt, ...) ibdev_err_ratelimited((pd)->ibpd.device, \
68 		"pd#%d %s: " fmt, (pd)->elem.index, __func__, ##__VA_ARGS__)
69 #define rxe_err_ah(ah, fmt, ...) ibdev_err_ratelimited((ah)->ibah.device, \
70 		"ah#%d %s: " fmt, (ah)->elem.index, __func__, ##__VA_ARGS__)
71 #define rxe_err_srq(srq, fmt, ...) ibdev_err_ratelimited((srq)->ibsrq.device, \
72 		"srq#%d %s: " fmt, (srq)->elem.index, __func__, ##__VA_ARGS__)
73 #define rxe_err_qp(qp, fmt, ...) ibdev_err_ratelimited((qp)->ibqp.device, \
74 		"qp#%d %s: " fmt, (qp)->elem.index, __func__, ##__VA_ARGS__)
75 #define rxe_err_cq(cq, fmt, ...) ibdev_err_ratelimited((cq)->ibcq.device, \
76 		"cq#%d %s: " fmt, (cq)->elem.index, __func__, ##__VA_ARGS__)
77 #define rxe_err_mr(mr, fmt, ...) ibdev_err_ratelimited((mr)->ibmr.device, \
78 		"mr#%d %s:  " fmt, (mr)->elem.index, __func__, ##__VA_ARGS__)
79 #define rxe_err_mw(mw, fmt, ...) ibdev_err_ratelimited((mw)->ibmw.device, \
80 		"mw#%d %s:  " fmt, (mw)->elem.index, __func__, ##__VA_ARGS__)
81 
82 #define rxe_info(fmt, ...) pr_info_ratelimited("%s: " fmt "\n", __func__, \
83 					##__VA_ARGS__)
84 #define rxe_info_dev(rxe, fmt, ...) ibdev_info_ratelimited(&(rxe)->ib_dev, \
85 		"%s: " fmt, __func__, ##__VA_ARGS__)
86 #define rxe_info_uc(uc, fmt, ...) ibdev_info_ratelimited((uc)->ibuc.device, \
87 		"uc#%d %s: " fmt, (uc)->elem.index, __func__, ##__VA_ARGS__)
88 #define rxe_info_pd(pd, fmt, ...) ibdev_info_ratelimited((pd)->ibpd.device, \
89 		"pd#%d %s: " fmt, (pd)->elem.index, __func__, ##__VA_ARGS__)
90 #define rxe_info_ah(ah, fmt, ...) ibdev_info_ratelimited((ah)->ibah.device, \
91 		"ah#%d %s: " fmt, (ah)->elem.index, __func__, ##__VA_ARGS__)
92 #define rxe_info_srq(srq, fmt, ...) ibdev_info_ratelimited((srq)->ibsrq.device, \
93 		"srq#%d %s: " fmt, (srq)->elem.index, __func__, ##__VA_ARGS__)
94 #define rxe_info_qp(qp, fmt, ...) ibdev_info_ratelimited((qp)->ibqp.device, \
95 		"qp#%d %s: " fmt, (qp)->elem.index, __func__, ##__VA_ARGS__)
96 #define rxe_info_cq(cq, fmt, ...) ibdev_info_ratelimited((cq)->ibcq.device, \
97 		"cq#%d %s: " fmt, (cq)->elem.index, __func__, ##__VA_ARGS__)
98 #define rxe_info_mr(mr, fmt, ...) ibdev_info_ratelimited((mr)->ibmr.device, \
99 		"mr#%d %s:  " fmt, (mr)->elem.index, __func__, ##__VA_ARGS__)
100 #define rxe_info_mw(mw, fmt, ...) ibdev_info_ratelimited((mw)->ibmw.device, \
101 		"mw#%d %s:  " fmt, (mw)->elem.index, __func__, ##__VA_ARGS__)
102 
103 /* responder states */
104 enum resp_states {
105 	RESPST_NONE,
106 	RESPST_GET_REQ,
107 	RESPST_CHK_PSN,
108 	RESPST_CHK_OP_SEQ,
109 	RESPST_CHK_OP_VALID,
110 	RESPST_CHK_RESOURCE,
111 	RESPST_CHK_LENGTH,
112 	RESPST_CHK_RKEY,
113 	RESPST_EXECUTE,
114 	RESPST_READ_REPLY,
115 	RESPST_ATOMIC_REPLY,
116 	RESPST_ATOMIC_WRITE_REPLY,
117 	RESPST_PROCESS_FLUSH,
118 	RESPST_COMPLETE,
119 	RESPST_ACKNOWLEDGE,
120 	RESPST_CLEANUP,
121 	RESPST_DUPLICATE_REQUEST,
122 	RESPST_ERR_MALFORMED_WQE,
123 	RESPST_ERR_UNSUPPORTED_OPCODE,
124 	RESPST_ERR_MISALIGNED_ATOMIC,
125 	RESPST_ERR_PSN_OUT_OF_SEQ,
126 	RESPST_ERR_MISSING_OPCODE_FIRST,
127 	RESPST_ERR_MISSING_OPCODE_LAST_C,
128 	RESPST_ERR_MISSING_OPCODE_LAST_D1E,
129 	RESPST_ERR_TOO_MANY_RDMA_ATM_REQ,
130 	RESPST_ERR_RNR,
131 	RESPST_ERR_RKEY_VIOLATION,
132 	RESPST_ERR_INVALIDATE_RKEY,
133 	RESPST_ERR_LENGTH,
134 	RESPST_ERR_CQ_OVERFLOW,
135 	RESPST_ERROR,
136 	RESPST_DONE,
137 	RESPST_EXIT,
138 };
139 
140 void rxe_set_mtu(struct rxe_dev *rxe, unsigned int dev_mtu);
141 
142 int rxe_add(struct rxe_dev *rxe, unsigned int mtu, const char *ibdev_name);
143 
144 void rxe_rcv(struct sk_buff *skb);
145 
146 /* The caller must do a matching ib_device_put(&dev->ib_dev) */
147 static inline struct rxe_dev *rxe_get_dev_from_net(struct net_device *ndev)
148 {
149 	struct ib_device *ibdev =
150 		ib_device_get_by_netdev(ndev, RDMA_DRIVER_RXE);
151 
152 	if (!ibdev)
153 		return NULL;
154 	return container_of(ibdev, struct rxe_dev, ib_dev);
155 }
156 
157 void rxe_port_up(struct rxe_dev *rxe);
158 void rxe_port_down(struct rxe_dev *rxe);
159 void rxe_set_port_state(struct rxe_dev *rxe);
160 
161 #endif /* RXE_H */
162