1 // SPDX-License-Identifier: GPL-2.0 or BSD-3-Clause
2 /*
3 * Copyright(c) 2015-2018 Intel Corporation.
4 */
5
6 #include <linux/net.h>
7 #include <rdma/opa_addr.h>
8 #define OPA_NUM_PKEY_BLOCKS_PER_SMP (OPA_SMP_DR_DATA_SIZE \
9 / (OPA_PARTITION_TABLE_BLK_SIZE * sizeof(u16)))
10
11 #include "hfi.h"
12 #include "mad.h"
13 #include "trace.h"
14 #include "qp.h"
15 #include "vnic.h"
16
17 /* the reset value from the FM is supposed to be 0xffff, handle both */
18 #define OPA_LINK_WIDTH_RESET_OLD 0x0fff
19 #define OPA_LINK_WIDTH_RESET 0xffff
20
21 struct trap_node {
22 struct list_head list;
23 struct opa_mad_notice_attr data;
24 __be64 tid;
25 int len;
26 u32 retry;
27 u8 in_use;
28 u8 repress;
29 };
30
smp_length_check(u32 data_size,u32 request_len)31 static int smp_length_check(u32 data_size, u32 request_len)
32 {
33 if (unlikely(request_len < data_size))
34 return -EINVAL;
35
36 return 0;
37 }
38
reply(struct ib_mad_hdr * smp)39 static int reply(struct ib_mad_hdr *smp)
40 {
41 /*
42 * The verbs framework will handle the directed/LID route
43 * packet changes.
44 */
45 smp->method = IB_MGMT_METHOD_GET_RESP;
46 if (smp->mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE)
47 smp->status |= IB_SMP_DIRECTION;
48 return IB_MAD_RESULT_SUCCESS | IB_MAD_RESULT_REPLY;
49 }
50
clear_opa_smp_data(struct opa_smp * smp)51 static inline void clear_opa_smp_data(struct opa_smp *smp)
52 {
53 void *data = opa_get_smp_data(smp);
54 size_t size = opa_get_smp_data_size(smp);
55
56 memset(data, 0, size);
57 }
58
hfi1_lookup_pkey_value(struct hfi1_ibport * ibp,int pkey_idx)59 static u16 hfi1_lookup_pkey_value(struct hfi1_ibport *ibp, int pkey_idx)
60 {
61 struct hfi1_pportdata *ppd = ppd_from_ibp(ibp);
62
63 if (pkey_idx < ARRAY_SIZE(ppd->pkeys))
64 return ppd->pkeys[pkey_idx];
65
66 return 0;
67 }
68
hfi1_event_pkey_change(struct hfi1_devdata * dd,u32 port)69 void hfi1_event_pkey_change(struct hfi1_devdata *dd, u32 port)
70 {
71 struct ib_event event;
72
73 event.event = IB_EVENT_PKEY_CHANGE;
74 event.device = &dd->verbs_dev.rdi.ibdev;
75 event.element.port_num = port;
76 ib_dispatch_event(&event);
77 }
78
79 /*
80 * If the port is down, clean up all pending traps. We need to be careful
81 * with the given trap, because it may be queued.
82 */
cleanup_traps(struct hfi1_ibport * ibp,struct trap_node * trap)83 static void cleanup_traps(struct hfi1_ibport *ibp, struct trap_node *trap)
84 {
85 struct trap_node *node, *q;
86 unsigned long flags;
87 struct list_head trap_list;
88 int i;
89
90 for (i = 0; i < RVT_MAX_TRAP_LISTS; i++) {
91 spin_lock_irqsave(&ibp->rvp.lock, flags);
92 list_replace_init(&ibp->rvp.trap_lists[i].list, &trap_list);
93 ibp->rvp.trap_lists[i].list_len = 0;
94 spin_unlock_irqrestore(&ibp->rvp.lock, flags);
95
96 /*
97 * Remove all items from the list, freeing all the non-given
98 * traps.
99 */
100 list_for_each_entry_safe(node, q, &trap_list, list) {
101 list_del(&node->list);
102 if (node != trap)
103 kfree(node);
104 }
105 }
106
107 /*
108 * If this wasn't on one of the lists it would not be freed. If it
109 * was on the list, it is now safe to free.
110 */
111 kfree(trap);
112 }
113
check_and_add_trap(struct hfi1_ibport * ibp,struct trap_node * trap)114 static struct trap_node *check_and_add_trap(struct hfi1_ibport *ibp,
115 struct trap_node *trap)
116 {
117 struct trap_node *node;
118 struct trap_list *trap_list;
119 unsigned long flags;
120 unsigned long timeout;
121 int found = 0;
122 unsigned int queue_id;
123 static int trap_count;
124
125 queue_id = trap->data.generic_type & 0x0F;
126 if (queue_id >= RVT_MAX_TRAP_LISTS) {
127 trap_count++;
128 pr_err_ratelimited("hfi1: Invalid trap 0x%0x dropped. Total dropped: %d\n",
129 trap->data.generic_type, trap_count);
130 kfree(trap);
131 return NULL;
132 }
133
134 /*
135 * Since the retry (handle timeout) does not remove a trap request
136 * from the list, all we have to do is compare the node.
137 */
138 spin_lock_irqsave(&ibp->rvp.lock, flags);
139 trap_list = &ibp->rvp.trap_lists[queue_id];
140
141 list_for_each_entry(node, &trap_list->list, list) {
142 if (node == trap) {
143 node->retry++;
144 found = 1;
145 break;
146 }
147 }
148
149 /* If it is not on the list, add it, limited to RVT-MAX_TRAP_LEN. */
150 if (!found) {
151 if (trap_list->list_len < RVT_MAX_TRAP_LEN) {
152 trap_list->list_len++;
153 list_add_tail(&trap->list, &trap_list->list);
154 } else {
155 pr_warn_ratelimited("hfi1: Maximum trap limit reached for 0x%0x traps\n",
156 trap->data.generic_type);
157 kfree(trap);
158 }
159 }
160
161 /*
162 * Next check to see if there is a timer pending. If not, set it up
163 * and get the first trap from the list.
164 */
165 node = NULL;
166 if (!timer_pending(&ibp->rvp.trap_timer)) {
167 /*
168 * o14-2
169 * If the time out is set we have to wait until it expires
170 * before the trap can be sent.
171 * This should be > RVT_TRAP_TIMEOUT
172 */
173 timeout = (RVT_TRAP_TIMEOUT *
174 (1UL << ibp->rvp.subnet_timeout)) / 1000;
175 mod_timer(&ibp->rvp.trap_timer,
176 jiffies + usecs_to_jiffies(timeout));
177 node = list_first_entry(&trap_list->list, struct trap_node,
178 list);
179 node->in_use = 1;
180 }
181 spin_unlock_irqrestore(&ibp->rvp.lock, flags);
182
183 return node;
184 }
185
subn_handle_opa_trap_repress(struct hfi1_ibport * ibp,struct opa_smp * smp)186 static void subn_handle_opa_trap_repress(struct hfi1_ibport *ibp,
187 struct opa_smp *smp)
188 {
189 struct trap_list *trap_list;
190 struct trap_node *trap;
191 unsigned long flags;
192 int i;
193
194 if (smp->attr_id != IB_SMP_ATTR_NOTICE)
195 return;
196
197 spin_lock_irqsave(&ibp->rvp.lock, flags);
198 for (i = 0; i < RVT_MAX_TRAP_LISTS; i++) {
199 trap_list = &ibp->rvp.trap_lists[i];
200 trap = list_first_entry_or_null(&trap_list->list,
201 struct trap_node, list);
202 if (trap && trap->tid == smp->tid) {
203 if (trap->in_use) {
204 trap->repress = 1;
205 } else {
206 trap_list->list_len--;
207 list_del(&trap->list);
208 kfree(trap);
209 }
210 break;
211 }
212 }
213 spin_unlock_irqrestore(&ibp->rvp.lock, flags);
214 }
215
hfi1_update_sm_ah_attr(struct hfi1_ibport * ibp,struct rdma_ah_attr * attr,u32 dlid)216 static void hfi1_update_sm_ah_attr(struct hfi1_ibport *ibp,
217 struct rdma_ah_attr *attr, u32 dlid)
218 {
219 rdma_ah_set_dlid(attr, dlid);
220 rdma_ah_set_port_num(attr, ppd_from_ibp(ibp)->port);
221 if (dlid >= be16_to_cpu(IB_MULTICAST_LID_BASE)) {
222 struct ib_global_route *grh = rdma_ah_retrieve_grh(attr);
223
224 rdma_ah_set_ah_flags(attr, IB_AH_GRH);
225 grh->sgid_index = 0;
226 grh->hop_limit = 1;
227 grh->dgid.global.subnet_prefix =
228 ibp->rvp.gid_prefix;
229 grh->dgid.global.interface_id = OPA_MAKE_ID(dlid);
230 }
231 }
232
hfi1_modify_qp0_ah(struct hfi1_ibport * ibp,struct rvt_ah * ah,u32 dlid)233 static int hfi1_modify_qp0_ah(struct hfi1_ibport *ibp,
234 struct rvt_ah *ah, u32 dlid)
235 {
236 struct rdma_ah_attr attr;
237 struct rvt_qp *qp0;
238 int ret = -EINVAL;
239
240 memset(&attr, 0, sizeof(attr));
241 attr.type = ah->ibah.type;
242 hfi1_update_sm_ah_attr(ibp, &attr, dlid);
243 rcu_read_lock();
244 qp0 = rcu_dereference(ibp->rvp.qp[0]);
245 if (qp0)
246 ret = rdma_modify_ah(&ah->ibah, &attr);
247 rcu_read_unlock();
248 return ret;
249 }
250
hfi1_create_qp0_ah(struct hfi1_ibport * ibp,u32 dlid)251 static struct ib_ah *hfi1_create_qp0_ah(struct hfi1_ibport *ibp, u32 dlid)
252 {
253 struct rdma_ah_attr attr;
254 struct ib_ah *ah = ERR_PTR(-EINVAL);
255 struct rvt_qp *qp0;
256 struct hfi1_pportdata *ppd = ppd_from_ibp(ibp);
257 struct hfi1_devdata *dd = dd_from_ppd(ppd);
258 u32 port_num = ppd->port;
259
260 memset(&attr, 0, sizeof(attr));
261 attr.type = rdma_ah_find_type(&dd->verbs_dev.rdi.ibdev, port_num);
262 hfi1_update_sm_ah_attr(ibp, &attr, dlid);
263 rcu_read_lock();
264 qp0 = rcu_dereference(ibp->rvp.qp[0]);
265 if (qp0)
266 ah = rdma_create_ah(qp0->ibqp.pd, &attr, 0);
267 rcu_read_unlock();
268 return ah;
269 }
270
send_trap(struct hfi1_ibport * ibp,struct trap_node * trap)271 static void send_trap(struct hfi1_ibport *ibp, struct trap_node *trap)
272 {
273 struct ib_mad_send_buf *send_buf;
274 struct ib_mad_agent *agent;
275 struct opa_smp *smp;
276 unsigned long flags;
277 int pkey_idx;
278 u32 qpn = ppd_from_ibp(ibp)->sm_trap_qp;
279
280 agent = ibp->rvp.send_agent;
281 if (!agent) {
282 cleanup_traps(ibp, trap);
283 return;
284 }
285
286 /* o14-3.2.1 */
287 if (driver_lstate(ppd_from_ibp(ibp)) != IB_PORT_ACTIVE) {
288 cleanup_traps(ibp, trap);
289 return;
290 }
291
292 /* Add the trap to the list if necessary and see if we can send it */
293 trap = check_and_add_trap(ibp, trap);
294 if (!trap)
295 return;
296
297 pkey_idx = hfi1_lookup_pkey_idx(ibp, LIM_MGMT_P_KEY);
298 if (pkey_idx < 0) {
299 pr_warn("%s: failed to find limited mgmt pkey, defaulting 0x%x\n",
300 __func__, hfi1_get_pkey(ibp, 1));
301 pkey_idx = 1;
302 }
303
304 send_buf = ib_create_send_mad(agent, qpn, pkey_idx, 0,
305 IB_MGMT_MAD_HDR, IB_MGMT_MAD_DATA,
306 GFP_ATOMIC, IB_MGMT_BASE_VERSION);
307 if (IS_ERR(send_buf))
308 return;
309
310 smp = send_buf->mad;
311 smp->base_version = OPA_MGMT_BASE_VERSION;
312 smp->mgmt_class = IB_MGMT_CLASS_SUBN_LID_ROUTED;
313 smp->class_version = OPA_SM_CLASS_VERSION;
314 smp->method = IB_MGMT_METHOD_TRAP;
315
316 /* Only update the transaction ID for new traps (o13-5). */
317 if (trap->tid == 0) {
318 ibp->rvp.tid++;
319 /* make sure that tid != 0 */
320 if (ibp->rvp.tid == 0)
321 ibp->rvp.tid++;
322 trap->tid = cpu_to_be64(ibp->rvp.tid);
323 }
324 smp->tid = trap->tid;
325
326 smp->attr_id = IB_SMP_ATTR_NOTICE;
327 /* o14-1: smp->mkey = 0; */
328
329 memcpy(smp->route.lid.data, &trap->data, trap->len);
330
331 spin_lock_irqsave(&ibp->rvp.lock, flags);
332 if (!ibp->rvp.sm_ah) {
333 if (ibp->rvp.sm_lid != be16_to_cpu(IB_LID_PERMISSIVE)) {
334 struct ib_ah *ah;
335
336 ah = hfi1_create_qp0_ah(ibp, ibp->rvp.sm_lid);
337 if (IS_ERR(ah)) {
338 spin_unlock_irqrestore(&ibp->rvp.lock, flags);
339 return;
340 }
341 send_buf->ah = ah;
342 ibp->rvp.sm_ah = ibah_to_rvtah(ah);
343 } else {
344 spin_unlock_irqrestore(&ibp->rvp.lock, flags);
345 return;
346 }
347 } else {
348 send_buf->ah = &ibp->rvp.sm_ah->ibah;
349 }
350
351 /*
352 * If the trap was repressed while things were getting set up, don't
353 * bother sending it. This could happen for a retry.
354 */
355 if (trap->repress) {
356 list_del(&trap->list);
357 spin_unlock_irqrestore(&ibp->rvp.lock, flags);
358 kfree(trap);
359 ib_free_send_mad(send_buf);
360 return;
361 }
362
363 trap->in_use = 0;
364 spin_unlock_irqrestore(&ibp->rvp.lock, flags);
365
366 if (ib_post_send_mad(send_buf, NULL))
367 ib_free_send_mad(send_buf);
368 }
369
hfi1_handle_trap_timer(struct timer_list * t)370 void hfi1_handle_trap_timer(struct timer_list *t)
371 {
372 struct hfi1_ibport *ibp = from_timer(ibp, t, rvp.trap_timer);
373 struct trap_node *trap = NULL;
374 unsigned long flags;
375 int i;
376
377 /* Find the trap with the highest priority */
378 spin_lock_irqsave(&ibp->rvp.lock, flags);
379 for (i = 0; !trap && i < RVT_MAX_TRAP_LISTS; i++) {
380 trap = list_first_entry_or_null(&ibp->rvp.trap_lists[i].list,
381 struct trap_node, list);
382 }
383 spin_unlock_irqrestore(&ibp->rvp.lock, flags);
384
385 if (trap)
386 send_trap(ibp, trap);
387 }
388
create_trap_node(u8 type,__be16 trap_num,u32 lid)389 static struct trap_node *create_trap_node(u8 type, __be16 trap_num, u32 lid)
390 {
391 struct trap_node *trap;
392
393 trap = kzalloc(sizeof(*trap), GFP_ATOMIC);
394 if (!trap)
395 return NULL;
396
397 INIT_LIST_HEAD(&trap->list);
398 trap->data.generic_type = type;
399 trap->data.prod_type_lsb = IB_NOTICE_PROD_CA;
400 trap->data.trap_num = trap_num;
401 trap->data.issuer_lid = cpu_to_be32(lid);
402
403 return trap;
404 }
405
406 /*
407 * Send a bad P_Key trap (ch. 14.3.8).
408 */
hfi1_bad_pkey(struct hfi1_ibport * ibp,u32 key,u32 sl,u32 qp1,u32 qp2,u32 lid1,u32 lid2)409 void hfi1_bad_pkey(struct hfi1_ibport *ibp, u32 key, u32 sl,
410 u32 qp1, u32 qp2, u32 lid1, u32 lid2)
411 {
412 struct trap_node *trap;
413 u32 lid = ppd_from_ibp(ibp)->lid;
414
415 ibp->rvp.n_pkt_drops++;
416 ibp->rvp.pkey_violations++;
417
418 trap = create_trap_node(IB_NOTICE_TYPE_SECURITY, OPA_TRAP_BAD_P_KEY,
419 lid);
420 if (!trap)
421 return;
422
423 /* Send violation trap */
424 trap->data.ntc_257_258.lid1 = cpu_to_be32(lid1);
425 trap->data.ntc_257_258.lid2 = cpu_to_be32(lid2);
426 trap->data.ntc_257_258.key = cpu_to_be32(key);
427 trap->data.ntc_257_258.sl = sl << 3;
428 trap->data.ntc_257_258.qp1 = cpu_to_be32(qp1);
429 trap->data.ntc_257_258.qp2 = cpu_to_be32(qp2);
430
431 trap->len = sizeof(trap->data);
432 send_trap(ibp, trap);
433 }
434
435 /*
436 * Send a bad M_Key trap (ch. 14.3.9).
437 */
bad_mkey(struct hfi1_ibport * ibp,struct ib_mad_hdr * mad,__be64 mkey,__be32 dr_slid,u8 return_path[],u8 hop_cnt)438 static void bad_mkey(struct hfi1_ibport *ibp, struct ib_mad_hdr *mad,
439 __be64 mkey, __be32 dr_slid, u8 return_path[], u8 hop_cnt)
440 {
441 struct trap_node *trap;
442 u32 lid = ppd_from_ibp(ibp)->lid;
443
444 trap = create_trap_node(IB_NOTICE_TYPE_SECURITY, OPA_TRAP_BAD_M_KEY,
445 lid);
446 if (!trap)
447 return;
448
449 /* Send violation trap */
450 trap->data.ntc_256.lid = trap->data.issuer_lid;
451 trap->data.ntc_256.method = mad->method;
452 trap->data.ntc_256.attr_id = mad->attr_id;
453 trap->data.ntc_256.attr_mod = mad->attr_mod;
454 trap->data.ntc_256.mkey = mkey;
455 if (mad->mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) {
456 trap->data.ntc_256.dr_slid = dr_slid;
457 trap->data.ntc_256.dr_trunc_hop = IB_NOTICE_TRAP_DR_NOTICE;
458 if (hop_cnt > ARRAY_SIZE(trap->data.ntc_256.dr_rtn_path)) {
459 trap->data.ntc_256.dr_trunc_hop |=
460 IB_NOTICE_TRAP_DR_TRUNC;
461 hop_cnt = ARRAY_SIZE(trap->data.ntc_256.dr_rtn_path);
462 }
463 trap->data.ntc_256.dr_trunc_hop |= hop_cnt;
464 memcpy(trap->data.ntc_256.dr_rtn_path, return_path,
465 hop_cnt);
466 }
467
468 trap->len = sizeof(trap->data);
469
470 send_trap(ibp, trap);
471 }
472
473 /*
474 * Send a Port Capability Mask Changed trap (ch. 14.3.11).
475 */
hfi1_cap_mask_chg(struct rvt_dev_info * rdi,u32 port_num)476 void hfi1_cap_mask_chg(struct rvt_dev_info *rdi, u32 port_num)
477 {
478 struct trap_node *trap;
479 struct hfi1_ibdev *verbs_dev = dev_from_rdi(rdi);
480 struct hfi1_devdata *dd = dd_from_dev(verbs_dev);
481 struct hfi1_ibport *ibp = &dd->pport[port_num - 1].ibport_data;
482 u32 lid = ppd_from_ibp(ibp)->lid;
483
484 trap = create_trap_node(IB_NOTICE_TYPE_INFO,
485 OPA_TRAP_CHANGE_CAPABILITY,
486 lid);
487 if (!trap)
488 return;
489
490 trap->data.ntc_144.lid = trap->data.issuer_lid;
491 trap->data.ntc_144.new_cap_mask = cpu_to_be32(ibp->rvp.port_cap_flags);
492 trap->data.ntc_144.cap_mask3 = cpu_to_be16(ibp->rvp.port_cap3_flags);
493
494 trap->len = sizeof(trap->data);
495 send_trap(ibp, trap);
496 }
497
498 /*
499 * Send a System Image GUID Changed trap (ch. 14.3.12).
500 */
hfi1_sys_guid_chg(struct hfi1_ibport * ibp)501 void hfi1_sys_guid_chg(struct hfi1_ibport *ibp)
502 {
503 struct trap_node *trap;
504 u32 lid = ppd_from_ibp(ibp)->lid;
505
506 trap = create_trap_node(IB_NOTICE_TYPE_INFO, OPA_TRAP_CHANGE_SYSGUID,
507 lid);
508 if (!trap)
509 return;
510
511 trap->data.ntc_145.new_sys_guid = ib_hfi1_sys_image_guid;
512 trap->data.ntc_145.lid = trap->data.issuer_lid;
513
514 trap->len = sizeof(trap->data);
515 send_trap(ibp, trap);
516 }
517
518 /*
519 * Send a Node Description Changed trap (ch. 14.3.13).
520 */
hfi1_node_desc_chg(struct hfi1_ibport * ibp)521 void hfi1_node_desc_chg(struct hfi1_ibport *ibp)
522 {
523 struct trap_node *trap;
524 u32 lid = ppd_from_ibp(ibp)->lid;
525
526 trap = create_trap_node(IB_NOTICE_TYPE_INFO,
527 OPA_TRAP_CHANGE_CAPABILITY,
528 lid);
529 if (!trap)
530 return;
531
532 trap->data.ntc_144.lid = trap->data.issuer_lid;
533 trap->data.ntc_144.change_flags =
534 cpu_to_be16(OPA_NOTICE_TRAP_NODE_DESC_CHG);
535
536 trap->len = sizeof(trap->data);
537 send_trap(ibp, trap);
538 }
539
__subn_get_opa_nodedesc(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)540 static int __subn_get_opa_nodedesc(struct opa_smp *smp, u32 am,
541 u8 *data, struct ib_device *ibdev,
542 u32 port, u32 *resp_len, u32 max_len)
543 {
544 struct opa_node_description *nd;
545
546 if (am || smp_length_check(sizeof(*nd), max_len)) {
547 smp->status |= IB_SMP_INVALID_FIELD;
548 return reply((struct ib_mad_hdr *)smp);
549 }
550
551 nd = (struct opa_node_description *)data;
552
553 memcpy(nd->data, ibdev->node_desc, sizeof(nd->data));
554
555 if (resp_len)
556 *resp_len += sizeof(*nd);
557
558 return reply((struct ib_mad_hdr *)smp);
559 }
560
__subn_get_opa_nodeinfo(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)561 static int __subn_get_opa_nodeinfo(struct opa_smp *smp, u32 am, u8 *data,
562 struct ib_device *ibdev, u32 port,
563 u32 *resp_len, u32 max_len)
564 {
565 struct opa_node_info *ni;
566 struct hfi1_devdata *dd = dd_from_ibdev(ibdev);
567 u32 pidx = port - 1; /* IB number port from 1, hw from 0 */
568
569 ni = (struct opa_node_info *)data;
570
571 /* GUID 0 is illegal */
572 if (am || pidx >= dd->num_pports || ibdev->node_guid == 0 ||
573 smp_length_check(sizeof(*ni), max_len) ||
574 get_sguid(to_iport(ibdev, port), HFI1_PORT_GUID_INDEX) == 0) {
575 smp->status |= IB_SMP_INVALID_FIELD;
576 return reply((struct ib_mad_hdr *)smp);
577 }
578
579 ni->port_guid = get_sguid(to_iport(ibdev, port), HFI1_PORT_GUID_INDEX);
580 ni->base_version = OPA_MGMT_BASE_VERSION;
581 ni->class_version = OPA_SM_CLASS_VERSION;
582 ni->node_type = 1; /* channel adapter */
583 ni->num_ports = ibdev->phys_port_cnt;
584 /* This is already in network order */
585 ni->system_image_guid = ib_hfi1_sys_image_guid;
586 ni->node_guid = ibdev->node_guid;
587 ni->partition_cap = cpu_to_be16(hfi1_get_npkeys(dd));
588 ni->device_id = cpu_to_be16(dd->pcidev->device);
589 ni->revision = cpu_to_be32(dd->minrev);
590 ni->local_port_num = port;
591 ni->vendor_id[0] = dd->oui1;
592 ni->vendor_id[1] = dd->oui2;
593 ni->vendor_id[2] = dd->oui3;
594
595 if (resp_len)
596 *resp_len += sizeof(*ni);
597
598 return reply((struct ib_mad_hdr *)smp);
599 }
600
subn_get_nodeinfo(struct ib_smp * smp,struct ib_device * ibdev,u32 port)601 static int subn_get_nodeinfo(struct ib_smp *smp, struct ib_device *ibdev,
602 u32 port)
603 {
604 struct ib_node_info *nip = (struct ib_node_info *)&smp->data;
605 struct hfi1_devdata *dd = dd_from_ibdev(ibdev);
606 u32 pidx = port - 1; /* IB number port from 1, hw from 0 */
607
608 /* GUID 0 is illegal */
609 if (smp->attr_mod || pidx >= dd->num_pports ||
610 ibdev->node_guid == 0 ||
611 get_sguid(to_iport(ibdev, port), HFI1_PORT_GUID_INDEX) == 0) {
612 smp->status |= IB_SMP_INVALID_FIELD;
613 return reply((struct ib_mad_hdr *)smp);
614 }
615
616 nip->port_guid = get_sguid(to_iport(ibdev, port), HFI1_PORT_GUID_INDEX);
617 nip->base_version = OPA_MGMT_BASE_VERSION;
618 nip->class_version = OPA_SM_CLASS_VERSION;
619 nip->node_type = 1; /* channel adapter */
620 nip->num_ports = ibdev->phys_port_cnt;
621 /* This is already in network order */
622 nip->sys_guid = ib_hfi1_sys_image_guid;
623 nip->node_guid = ibdev->node_guid;
624 nip->partition_cap = cpu_to_be16(hfi1_get_npkeys(dd));
625 nip->device_id = cpu_to_be16(dd->pcidev->device);
626 nip->revision = cpu_to_be32(dd->minrev);
627 nip->local_port_num = port;
628 nip->vendor_id[0] = dd->oui1;
629 nip->vendor_id[1] = dd->oui2;
630 nip->vendor_id[2] = dd->oui3;
631
632 return reply((struct ib_mad_hdr *)smp);
633 }
634
set_link_width_enabled(struct hfi1_pportdata * ppd,u32 w)635 static void set_link_width_enabled(struct hfi1_pportdata *ppd, u32 w)
636 {
637 (void)hfi1_set_ib_cfg(ppd, HFI1_IB_CFG_LWID_ENB, w);
638 }
639
set_link_width_downgrade_enabled(struct hfi1_pportdata * ppd,u32 w)640 static void set_link_width_downgrade_enabled(struct hfi1_pportdata *ppd, u32 w)
641 {
642 (void)hfi1_set_ib_cfg(ppd, HFI1_IB_CFG_LWID_DG_ENB, w);
643 }
644
set_link_speed_enabled(struct hfi1_pportdata * ppd,u32 s)645 static void set_link_speed_enabled(struct hfi1_pportdata *ppd, u32 s)
646 {
647 (void)hfi1_set_ib_cfg(ppd, HFI1_IB_CFG_SPD_ENB, s);
648 }
649
check_mkey(struct hfi1_ibport * ibp,struct ib_mad_hdr * mad,int mad_flags,__be64 mkey,__be32 dr_slid,u8 return_path[],u8 hop_cnt)650 static int check_mkey(struct hfi1_ibport *ibp, struct ib_mad_hdr *mad,
651 int mad_flags, __be64 mkey, __be32 dr_slid,
652 u8 return_path[], u8 hop_cnt)
653 {
654 int valid_mkey = 0;
655 int ret = 0;
656
657 /* Is the mkey in the process of expiring? */
658 if (ibp->rvp.mkey_lease_timeout &&
659 time_after_eq(jiffies, ibp->rvp.mkey_lease_timeout)) {
660 /* Clear timeout and mkey protection field. */
661 ibp->rvp.mkey_lease_timeout = 0;
662 ibp->rvp.mkeyprot = 0;
663 }
664
665 if ((mad_flags & IB_MAD_IGNORE_MKEY) || ibp->rvp.mkey == 0 ||
666 ibp->rvp.mkey == mkey)
667 valid_mkey = 1;
668
669 /* Unset lease timeout on any valid Get/Set/TrapRepress */
670 if (valid_mkey && ibp->rvp.mkey_lease_timeout &&
671 (mad->method == IB_MGMT_METHOD_GET ||
672 mad->method == IB_MGMT_METHOD_SET ||
673 mad->method == IB_MGMT_METHOD_TRAP_REPRESS))
674 ibp->rvp.mkey_lease_timeout = 0;
675
676 if (!valid_mkey) {
677 switch (mad->method) {
678 case IB_MGMT_METHOD_GET:
679 /* Bad mkey not a violation below level 2 */
680 if (ibp->rvp.mkeyprot < 2)
681 break;
682 fallthrough;
683 case IB_MGMT_METHOD_SET:
684 case IB_MGMT_METHOD_TRAP_REPRESS:
685 if (ibp->rvp.mkey_violations != 0xFFFF)
686 ++ibp->rvp.mkey_violations;
687 if (!ibp->rvp.mkey_lease_timeout &&
688 ibp->rvp.mkey_lease_period)
689 ibp->rvp.mkey_lease_timeout = jiffies +
690 ibp->rvp.mkey_lease_period * HZ;
691 /* Generate a trap notice. */
692 bad_mkey(ibp, mad, mkey, dr_slid, return_path,
693 hop_cnt);
694 ret = 1;
695 }
696 }
697
698 return ret;
699 }
700
701 /*
702 * The SMA caches reads from LCB registers in case the LCB is unavailable.
703 * (The LCB is unavailable in certain link states, for example.)
704 */
705 struct lcb_datum {
706 u32 off;
707 u64 val;
708 };
709
710 static struct lcb_datum lcb_cache[] = {
711 { DC_LCB_STS_ROUND_TRIP_LTP_CNT, 0 },
712 };
713
write_lcb_cache(u32 off,u64 val)714 static int write_lcb_cache(u32 off, u64 val)
715 {
716 int i;
717
718 for (i = 0; i < ARRAY_SIZE(lcb_cache); i++) {
719 if (lcb_cache[i].off == off) {
720 lcb_cache[i].val = val;
721 return 0;
722 }
723 }
724
725 pr_warn("%s bad offset 0x%x\n", __func__, off);
726 return -1;
727 }
728
read_lcb_cache(u32 off,u64 * val)729 static int read_lcb_cache(u32 off, u64 *val)
730 {
731 int i;
732
733 for (i = 0; i < ARRAY_SIZE(lcb_cache); i++) {
734 if (lcb_cache[i].off == off) {
735 *val = lcb_cache[i].val;
736 return 0;
737 }
738 }
739
740 pr_warn("%s bad offset 0x%x\n", __func__, off);
741 return -1;
742 }
743
read_ltp_rtt(struct hfi1_devdata * dd)744 void read_ltp_rtt(struct hfi1_devdata *dd)
745 {
746 u64 reg;
747
748 if (read_lcb_csr(dd, DC_LCB_STS_ROUND_TRIP_LTP_CNT, ®))
749 dd_dev_err(dd, "%s: unable to read LTP RTT\n", __func__);
750 else
751 write_lcb_cache(DC_LCB_STS_ROUND_TRIP_LTP_CNT, reg);
752 }
753
__subn_get_opa_portinfo(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)754 static int __subn_get_opa_portinfo(struct opa_smp *smp, u32 am, u8 *data,
755 struct ib_device *ibdev, u32 port,
756 u32 *resp_len, u32 max_len)
757 {
758 int i;
759 struct hfi1_devdata *dd;
760 struct hfi1_pportdata *ppd;
761 struct hfi1_ibport *ibp;
762 struct opa_port_info *pi = (struct opa_port_info *)data;
763 u8 mtu;
764 u8 credit_rate;
765 u8 is_beaconing_active;
766 u32 state;
767 u32 num_ports = OPA_AM_NPORT(am);
768 u32 start_of_sm_config = OPA_AM_START_SM_CFG(am);
769 u32 buffer_units;
770 u64 tmp = 0;
771
772 if (num_ports != 1 || smp_length_check(sizeof(*pi), max_len)) {
773 smp->status |= IB_SMP_INVALID_FIELD;
774 return reply((struct ib_mad_hdr *)smp);
775 }
776
777 dd = dd_from_ibdev(ibdev);
778 /* IB numbers ports from 1, hw from 0 */
779 ppd = dd->pport + (port - 1);
780 ibp = &ppd->ibport_data;
781
782 if (ppd->vls_supported / 2 > ARRAY_SIZE(pi->neigh_mtu.pvlx_to_mtu) ||
783 ppd->vls_supported > ARRAY_SIZE(dd->vld)) {
784 smp->status |= IB_SMP_INVALID_FIELD;
785 return reply((struct ib_mad_hdr *)smp);
786 }
787
788 pi->lid = cpu_to_be32(ppd->lid);
789
790 /* Only return the mkey if the protection field allows it. */
791 if (!(smp->method == IB_MGMT_METHOD_GET &&
792 ibp->rvp.mkey != smp->mkey &&
793 ibp->rvp.mkeyprot == 1))
794 pi->mkey = ibp->rvp.mkey;
795
796 pi->subnet_prefix = ibp->rvp.gid_prefix;
797 pi->sm_lid = cpu_to_be32(ibp->rvp.sm_lid);
798 pi->ib_cap_mask = cpu_to_be32(ibp->rvp.port_cap_flags);
799 pi->mkey_lease_period = cpu_to_be16(ibp->rvp.mkey_lease_period);
800 pi->sm_trap_qp = cpu_to_be32(ppd->sm_trap_qp);
801 pi->sa_qp = cpu_to_be32(ppd->sa_qp);
802
803 pi->link_width.enabled = cpu_to_be16(ppd->link_width_enabled);
804 pi->link_width.supported = cpu_to_be16(ppd->link_width_supported);
805 pi->link_width.active = cpu_to_be16(ppd->link_width_active);
806
807 pi->link_width_downgrade.supported =
808 cpu_to_be16(ppd->link_width_downgrade_supported);
809 pi->link_width_downgrade.enabled =
810 cpu_to_be16(ppd->link_width_downgrade_enabled);
811 pi->link_width_downgrade.tx_active =
812 cpu_to_be16(ppd->link_width_downgrade_tx_active);
813 pi->link_width_downgrade.rx_active =
814 cpu_to_be16(ppd->link_width_downgrade_rx_active);
815
816 pi->link_speed.supported = cpu_to_be16(ppd->link_speed_supported);
817 pi->link_speed.active = cpu_to_be16(ppd->link_speed_active);
818 pi->link_speed.enabled = cpu_to_be16(ppd->link_speed_enabled);
819
820 state = driver_lstate(ppd);
821
822 if (start_of_sm_config && (state == IB_PORT_INIT))
823 ppd->is_sm_config_started = 1;
824
825 pi->port_phys_conf = (ppd->port_type & 0xf);
826
827 pi->port_states.ledenable_offlinereason = ppd->neighbor_normal << 4;
828 pi->port_states.ledenable_offlinereason |=
829 ppd->is_sm_config_started << 5;
830 /*
831 * This pairs with the memory barrier in hfi1_start_led_override to
832 * ensure that we read the correct state of LED beaconing represented
833 * by led_override_timer_active
834 */
835 smp_rmb();
836 is_beaconing_active = !!atomic_read(&ppd->led_override_timer_active);
837 pi->port_states.ledenable_offlinereason |= is_beaconing_active << 6;
838 pi->port_states.ledenable_offlinereason |=
839 ppd->offline_disabled_reason;
840
841 pi->port_states.portphysstate_portstate =
842 (driver_pstate(ppd) << 4) | state;
843
844 pi->mkeyprotect_lmc = (ibp->rvp.mkeyprot << 6) | ppd->lmc;
845
846 memset(pi->neigh_mtu.pvlx_to_mtu, 0, sizeof(pi->neigh_mtu.pvlx_to_mtu));
847 for (i = 0; i < ppd->vls_supported; i++) {
848 mtu = mtu_to_enum(dd->vld[i].mtu, HFI1_DEFAULT_ACTIVE_MTU);
849 if ((i % 2) == 0)
850 pi->neigh_mtu.pvlx_to_mtu[i / 2] |= (mtu << 4);
851 else
852 pi->neigh_mtu.pvlx_to_mtu[i / 2] |= mtu;
853 }
854 /* don't forget VL 15 */
855 mtu = mtu_to_enum(dd->vld[15].mtu, 2048);
856 pi->neigh_mtu.pvlx_to_mtu[15 / 2] |= mtu;
857 pi->smsl = ibp->rvp.sm_sl & OPA_PI_MASK_SMSL;
858 pi->operational_vls = hfi1_get_ib_cfg(ppd, HFI1_IB_CFG_OP_VLS);
859 pi->partenforce_filterraw |=
860 (ppd->linkinit_reason & OPA_PI_MASK_LINKINIT_REASON);
861 if (ppd->part_enforce & HFI1_PART_ENFORCE_IN)
862 pi->partenforce_filterraw |= OPA_PI_MASK_PARTITION_ENFORCE_IN;
863 if (ppd->part_enforce & HFI1_PART_ENFORCE_OUT)
864 pi->partenforce_filterraw |= OPA_PI_MASK_PARTITION_ENFORCE_OUT;
865 pi->mkey_violations = cpu_to_be16(ibp->rvp.mkey_violations);
866 /* P_KeyViolations are counted by hardware. */
867 pi->pkey_violations = cpu_to_be16(ibp->rvp.pkey_violations);
868 pi->qkey_violations = cpu_to_be16(ibp->rvp.qkey_violations);
869
870 pi->vl.cap = ppd->vls_supported;
871 pi->vl.high_limit = cpu_to_be16(ibp->rvp.vl_high_limit);
872 pi->vl.arb_high_cap = (u8)hfi1_get_ib_cfg(ppd, HFI1_IB_CFG_VL_HIGH_CAP);
873 pi->vl.arb_low_cap = (u8)hfi1_get_ib_cfg(ppd, HFI1_IB_CFG_VL_LOW_CAP);
874
875 pi->clientrereg_subnettimeout = ibp->rvp.subnet_timeout;
876
877 pi->port_link_mode = cpu_to_be16(OPA_PORT_LINK_MODE_OPA << 10 |
878 OPA_PORT_LINK_MODE_OPA << 5 |
879 OPA_PORT_LINK_MODE_OPA);
880
881 pi->port_ltp_crc_mode = cpu_to_be16(ppd->port_ltp_crc_mode);
882
883 pi->port_mode = cpu_to_be16(
884 ppd->is_active_optimize_enabled ?
885 OPA_PI_MASK_PORT_ACTIVE_OPTOMIZE : 0);
886
887 pi->port_packet_format.supported =
888 cpu_to_be16(OPA_PORT_PACKET_FORMAT_9B |
889 OPA_PORT_PACKET_FORMAT_16B);
890 pi->port_packet_format.enabled =
891 cpu_to_be16(OPA_PORT_PACKET_FORMAT_9B |
892 OPA_PORT_PACKET_FORMAT_16B);
893
894 /* flit_control.interleave is (OPA V1, version .76):
895 * bits use
896 * ---- ---
897 * 2 res
898 * 2 DistanceSupported
899 * 2 DistanceEnabled
900 * 5 MaxNextLevelTxEnabled
901 * 5 MaxNestLevelRxSupported
902 *
903 * HFI supports only "distance mode 1" (see OPA V1, version .76,
904 * section 9.6.2), so set DistanceSupported, DistanceEnabled
905 * to 0x1.
906 */
907 pi->flit_control.interleave = cpu_to_be16(0x1400);
908
909 pi->link_down_reason = ppd->local_link_down_reason.sma;
910 pi->neigh_link_down_reason = ppd->neigh_link_down_reason.sma;
911 pi->port_error_action = cpu_to_be32(ppd->port_error_action);
912 pi->mtucap = mtu_to_enum(hfi1_max_mtu, IB_MTU_4096);
913
914 /* 32.768 usec. response time (guessing) */
915 pi->resptimevalue = 3;
916
917 pi->local_port_num = port;
918
919 /* buffer info for FM */
920 pi->overall_buffer_space = cpu_to_be16(dd->link_credits);
921
922 pi->neigh_node_guid = cpu_to_be64(ppd->neighbor_guid);
923 pi->neigh_port_num = ppd->neighbor_port_number;
924 pi->port_neigh_mode =
925 (ppd->neighbor_type & OPA_PI_MASK_NEIGH_NODE_TYPE) |
926 (ppd->mgmt_allowed ? OPA_PI_MASK_NEIGH_MGMT_ALLOWED : 0) |
927 (ppd->neighbor_fm_security ?
928 OPA_PI_MASK_NEIGH_FW_AUTH_BYPASS : 0);
929
930 /* HFIs shall always return VL15 credits to their
931 * neighbor in a timely manner, without any credit return pacing.
932 */
933 credit_rate = 0;
934 buffer_units = (dd->vau) & OPA_PI_MASK_BUF_UNIT_BUF_ALLOC;
935 buffer_units |= (dd->vcu << 3) & OPA_PI_MASK_BUF_UNIT_CREDIT_ACK;
936 buffer_units |= (credit_rate << 6) &
937 OPA_PI_MASK_BUF_UNIT_VL15_CREDIT_RATE;
938 buffer_units |= (dd->vl15_init << 11) & OPA_PI_MASK_BUF_UNIT_VL15_INIT;
939 pi->buffer_units = cpu_to_be32(buffer_units);
940
941 pi->opa_cap_mask = cpu_to_be16(ibp->rvp.port_cap3_flags);
942 pi->collectivemask_multicastmask = ((OPA_COLLECTIVE_NR & 0x7)
943 << 3 | (OPA_MCAST_NR & 0x7));
944
945 /* HFI supports a replay buffer 128 LTPs in size */
946 pi->replay_depth.buffer = 0x80;
947 /* read the cached value of DC_LCB_STS_ROUND_TRIP_LTP_CNT */
948 read_lcb_cache(DC_LCB_STS_ROUND_TRIP_LTP_CNT, &tmp);
949
950 /*
951 * this counter is 16 bits wide, but the replay_depth.wire
952 * variable is only 8 bits
953 */
954 if (tmp > 0xff)
955 tmp = 0xff;
956 pi->replay_depth.wire = tmp;
957
958 if (resp_len)
959 *resp_len += sizeof(struct opa_port_info);
960
961 return reply((struct ib_mad_hdr *)smp);
962 }
963
964 /**
965 * get_pkeys - return the PKEY table
966 * @dd: the hfi1_ib device
967 * @port: the IB port number
968 * @pkeys: the pkey table is placed here
969 */
get_pkeys(struct hfi1_devdata * dd,u32 port,u16 * pkeys)970 static int get_pkeys(struct hfi1_devdata *dd, u32 port, u16 *pkeys)
971 {
972 struct hfi1_pportdata *ppd = dd->pport + port - 1;
973
974 memcpy(pkeys, ppd->pkeys, sizeof(ppd->pkeys));
975
976 return 0;
977 }
978
__subn_get_opa_pkeytable(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)979 static int __subn_get_opa_pkeytable(struct opa_smp *smp, u32 am, u8 *data,
980 struct ib_device *ibdev, u32 port,
981 u32 *resp_len, u32 max_len)
982 {
983 struct hfi1_devdata *dd = dd_from_ibdev(ibdev);
984 u32 n_blocks_req = OPA_AM_NBLK(am);
985 u32 start_block = am & 0x7ff;
986 __be16 *p;
987 u16 *q;
988 int i;
989 u16 n_blocks_avail;
990 unsigned npkeys = hfi1_get_npkeys(dd);
991 size_t size;
992
993 if (n_blocks_req == 0) {
994 pr_warn("OPA Get PKey AM Invalid : P = %d; B = 0x%x; N = 0x%x\n",
995 port, start_block, n_blocks_req);
996 smp->status |= IB_SMP_INVALID_FIELD;
997 return reply((struct ib_mad_hdr *)smp);
998 }
999
1000 n_blocks_avail = (u16)(npkeys / OPA_PARTITION_TABLE_BLK_SIZE) + 1;
1001
1002 size = (n_blocks_req * OPA_PARTITION_TABLE_BLK_SIZE) * sizeof(u16);
1003
1004 if (smp_length_check(size, max_len)) {
1005 smp->status |= IB_SMP_INVALID_FIELD;
1006 return reply((struct ib_mad_hdr *)smp);
1007 }
1008
1009 if (start_block + n_blocks_req > n_blocks_avail ||
1010 n_blocks_req > OPA_NUM_PKEY_BLOCKS_PER_SMP) {
1011 pr_warn("OPA Get PKey AM Invalid : s 0x%x; req 0x%x; "
1012 "avail 0x%x; blk/smp 0x%lx\n",
1013 start_block, n_blocks_req, n_blocks_avail,
1014 OPA_NUM_PKEY_BLOCKS_PER_SMP);
1015 smp->status |= IB_SMP_INVALID_FIELD;
1016 return reply((struct ib_mad_hdr *)smp);
1017 }
1018
1019 p = (__be16 *)data;
1020 q = (u16 *)data;
1021 /* get the real pkeys if we are requesting the first block */
1022 if (start_block == 0) {
1023 get_pkeys(dd, port, q);
1024 for (i = 0; i < npkeys; i++)
1025 p[i] = cpu_to_be16(q[i]);
1026 if (resp_len)
1027 *resp_len += size;
1028 } else {
1029 smp->status |= IB_SMP_INVALID_FIELD;
1030 }
1031 return reply((struct ib_mad_hdr *)smp);
1032 }
1033
1034 enum {
1035 HFI_TRANSITION_DISALLOWED,
1036 HFI_TRANSITION_IGNORED,
1037 HFI_TRANSITION_ALLOWED,
1038 HFI_TRANSITION_UNDEFINED,
1039 };
1040
1041 /*
1042 * Use shortened names to improve readability of
1043 * {logical,physical}_state_transitions
1044 */
1045 enum {
1046 __D = HFI_TRANSITION_DISALLOWED,
1047 __I = HFI_TRANSITION_IGNORED,
1048 __A = HFI_TRANSITION_ALLOWED,
1049 __U = HFI_TRANSITION_UNDEFINED,
1050 };
1051
1052 /*
1053 * IB_PORTPHYSSTATE_POLLING (2) through OPA_PORTPHYSSTATE_MAX (11) are
1054 * represented in physical_state_transitions.
1055 */
1056 #define __N_PHYSTATES (OPA_PORTPHYSSTATE_MAX - IB_PORTPHYSSTATE_POLLING + 1)
1057
1058 /*
1059 * Within physical_state_transitions, rows represent "old" states,
1060 * columns "new" states, and physical_state_transitions.allowed[old][new]
1061 * indicates if the transition from old state to new state is legal (see
1062 * OPAg1v1, Table 6-4).
1063 */
1064 static const struct {
1065 u8 allowed[__N_PHYSTATES][__N_PHYSTATES];
1066 } physical_state_transitions = {
1067 {
1068 /* 2 3 4 5 6 7 8 9 10 11 */
1069 /* 2 */ { __A, __A, __D, __D, __D, __D, __D, __D, __D, __D },
1070 /* 3 */ { __A, __I, __D, __D, __D, __D, __D, __D, __D, __A },
1071 /* 4 */ { __U, __U, __U, __U, __U, __U, __U, __U, __U, __U },
1072 /* 5 */ { __A, __A, __D, __I, __D, __D, __D, __D, __D, __D },
1073 /* 6 */ { __U, __U, __U, __U, __U, __U, __U, __U, __U, __U },
1074 /* 7 */ { __D, __A, __D, __D, __D, __I, __D, __D, __D, __D },
1075 /* 8 */ { __U, __U, __U, __U, __U, __U, __U, __U, __U, __U },
1076 /* 9 */ { __I, __A, __D, __D, __D, __D, __D, __I, __D, __D },
1077 /*10 */ { __U, __U, __U, __U, __U, __U, __U, __U, __U, __U },
1078 /*11 */ { __D, __A, __D, __D, __D, __D, __D, __D, __D, __I },
1079 }
1080 };
1081
1082 /*
1083 * IB_PORT_DOWN (1) through IB_PORT_ACTIVE_DEFER (5) are represented
1084 * logical_state_transitions
1085 */
1086
1087 #define __N_LOGICAL_STATES (IB_PORT_ACTIVE_DEFER - IB_PORT_DOWN + 1)
1088
1089 /*
1090 * Within logical_state_transitions rows represent "old" states,
1091 * columns "new" states, and logical_state_transitions.allowed[old][new]
1092 * indicates if the transition from old state to new state is legal (see
1093 * OPAg1v1, Table 9-12).
1094 */
1095 static const struct {
1096 u8 allowed[__N_LOGICAL_STATES][__N_LOGICAL_STATES];
1097 } logical_state_transitions = {
1098 {
1099 /* 1 2 3 4 5 */
1100 /* 1 */ { __I, __D, __D, __D, __U},
1101 /* 2 */ { __D, __I, __A, __D, __U},
1102 /* 3 */ { __D, __D, __I, __A, __U},
1103 /* 4 */ { __D, __D, __I, __I, __U},
1104 /* 5 */ { __U, __U, __U, __U, __U},
1105 }
1106 };
1107
logical_transition_allowed(int old,int new)1108 static int logical_transition_allowed(int old, int new)
1109 {
1110 if (old < IB_PORT_NOP || old > IB_PORT_ACTIVE_DEFER ||
1111 new < IB_PORT_NOP || new > IB_PORT_ACTIVE_DEFER) {
1112 pr_warn("invalid logical state(s) (old %d new %d)\n",
1113 old, new);
1114 return HFI_TRANSITION_UNDEFINED;
1115 }
1116
1117 if (new == IB_PORT_NOP)
1118 return HFI_TRANSITION_ALLOWED; /* always allowed */
1119
1120 /* adjust states for indexing into logical_state_transitions */
1121 old -= IB_PORT_DOWN;
1122 new -= IB_PORT_DOWN;
1123
1124 if (old < 0 || new < 0)
1125 return HFI_TRANSITION_UNDEFINED;
1126 return logical_state_transitions.allowed[old][new];
1127 }
1128
physical_transition_allowed(int old,int new)1129 static int physical_transition_allowed(int old, int new)
1130 {
1131 if (old < IB_PORTPHYSSTATE_NOP || old > OPA_PORTPHYSSTATE_MAX ||
1132 new < IB_PORTPHYSSTATE_NOP || new > OPA_PORTPHYSSTATE_MAX) {
1133 pr_warn("invalid physical state(s) (old %d new %d)\n",
1134 old, new);
1135 return HFI_TRANSITION_UNDEFINED;
1136 }
1137
1138 if (new == IB_PORTPHYSSTATE_NOP)
1139 return HFI_TRANSITION_ALLOWED; /* always allowed */
1140
1141 /* adjust states for indexing into physical_state_transitions */
1142 old -= IB_PORTPHYSSTATE_POLLING;
1143 new -= IB_PORTPHYSSTATE_POLLING;
1144
1145 if (old < 0 || new < 0)
1146 return HFI_TRANSITION_UNDEFINED;
1147 return physical_state_transitions.allowed[old][new];
1148 }
1149
port_states_transition_allowed(struct hfi1_pportdata * ppd,u32 logical_new,u32 physical_new)1150 static int port_states_transition_allowed(struct hfi1_pportdata *ppd,
1151 u32 logical_new, u32 physical_new)
1152 {
1153 u32 physical_old = driver_pstate(ppd);
1154 u32 logical_old = driver_lstate(ppd);
1155 int ret, logical_allowed, physical_allowed;
1156
1157 ret = logical_transition_allowed(logical_old, logical_new);
1158 logical_allowed = ret;
1159
1160 if (ret == HFI_TRANSITION_DISALLOWED ||
1161 ret == HFI_TRANSITION_UNDEFINED) {
1162 pr_warn("invalid logical state transition %s -> %s\n",
1163 opa_lstate_name(logical_old),
1164 opa_lstate_name(logical_new));
1165 return ret;
1166 }
1167
1168 ret = physical_transition_allowed(physical_old, physical_new);
1169 physical_allowed = ret;
1170
1171 if (ret == HFI_TRANSITION_DISALLOWED ||
1172 ret == HFI_TRANSITION_UNDEFINED) {
1173 pr_warn("invalid physical state transition %s -> %s\n",
1174 opa_pstate_name(physical_old),
1175 opa_pstate_name(physical_new));
1176 return ret;
1177 }
1178
1179 if (logical_allowed == HFI_TRANSITION_IGNORED &&
1180 physical_allowed == HFI_TRANSITION_IGNORED)
1181 return HFI_TRANSITION_IGNORED;
1182
1183 /*
1184 * A change request of Physical Port State from
1185 * 'Offline' to 'Polling' should be ignored.
1186 */
1187 if ((physical_old == OPA_PORTPHYSSTATE_OFFLINE) &&
1188 (physical_new == IB_PORTPHYSSTATE_POLLING))
1189 return HFI_TRANSITION_IGNORED;
1190
1191 /*
1192 * Either physical_allowed or logical_allowed is
1193 * HFI_TRANSITION_ALLOWED.
1194 */
1195 return HFI_TRANSITION_ALLOWED;
1196 }
1197
set_port_states(struct hfi1_pportdata * ppd,struct opa_smp * smp,u32 logical_state,u32 phys_state,int local_mad)1198 static int set_port_states(struct hfi1_pportdata *ppd, struct opa_smp *smp,
1199 u32 logical_state, u32 phys_state, int local_mad)
1200 {
1201 struct hfi1_devdata *dd = ppd->dd;
1202 u32 link_state;
1203 int ret;
1204
1205 ret = port_states_transition_allowed(ppd, logical_state, phys_state);
1206 if (ret == HFI_TRANSITION_DISALLOWED ||
1207 ret == HFI_TRANSITION_UNDEFINED) {
1208 /* error message emitted above */
1209 smp->status |= IB_SMP_INVALID_FIELD;
1210 return 0;
1211 }
1212
1213 if (ret == HFI_TRANSITION_IGNORED)
1214 return 0;
1215
1216 if ((phys_state != IB_PORTPHYSSTATE_NOP) &&
1217 !(logical_state == IB_PORT_DOWN ||
1218 logical_state == IB_PORT_NOP)){
1219 pr_warn("SubnSet(OPA_PortInfo) port state invalid: logical_state 0x%x physical_state 0x%x\n",
1220 logical_state, phys_state);
1221 smp->status |= IB_SMP_INVALID_FIELD;
1222 }
1223
1224 /*
1225 * Logical state changes are summarized in OPAv1g1 spec.,
1226 * Table 9-12; physical state changes are summarized in
1227 * OPAv1g1 spec., Table 6.4.
1228 */
1229 switch (logical_state) {
1230 case IB_PORT_NOP:
1231 if (phys_state == IB_PORTPHYSSTATE_NOP)
1232 break;
1233 fallthrough;
1234 case IB_PORT_DOWN:
1235 if (phys_state == IB_PORTPHYSSTATE_NOP) {
1236 link_state = HLS_DN_DOWNDEF;
1237 } else if (phys_state == IB_PORTPHYSSTATE_POLLING) {
1238 link_state = HLS_DN_POLL;
1239 set_link_down_reason(ppd, OPA_LINKDOWN_REASON_FM_BOUNCE,
1240 0, OPA_LINKDOWN_REASON_FM_BOUNCE);
1241 } else if (phys_state == IB_PORTPHYSSTATE_DISABLED) {
1242 link_state = HLS_DN_DISABLE;
1243 } else {
1244 pr_warn("SubnSet(OPA_PortInfo) invalid physical state 0x%x\n",
1245 phys_state);
1246 smp->status |= IB_SMP_INVALID_FIELD;
1247 break;
1248 }
1249
1250 if ((link_state == HLS_DN_POLL ||
1251 link_state == HLS_DN_DOWNDEF)) {
1252 /*
1253 * Going to poll. No matter what the current state,
1254 * always move offline first, then tune and start the
1255 * link. This correctly handles a FM link bounce and
1256 * a link enable. Going offline is a no-op if already
1257 * offline.
1258 */
1259 set_link_state(ppd, HLS_DN_OFFLINE);
1260 start_link(ppd);
1261 } else {
1262 set_link_state(ppd, link_state);
1263 }
1264 if (link_state == HLS_DN_DISABLE &&
1265 (ppd->offline_disabled_reason >
1266 HFI1_ODR_MASK(OPA_LINKDOWN_REASON_SMA_DISABLED) ||
1267 ppd->offline_disabled_reason ==
1268 HFI1_ODR_MASK(OPA_LINKDOWN_REASON_NONE)))
1269 ppd->offline_disabled_reason =
1270 HFI1_ODR_MASK(OPA_LINKDOWN_REASON_SMA_DISABLED);
1271 /*
1272 * Don't send a reply if the response would be sent
1273 * through the disabled port.
1274 */
1275 if (link_state == HLS_DN_DISABLE && !local_mad)
1276 return IB_MAD_RESULT_SUCCESS | IB_MAD_RESULT_CONSUMED;
1277 break;
1278 case IB_PORT_ARMED:
1279 ret = set_link_state(ppd, HLS_UP_ARMED);
1280 if (!ret)
1281 send_idle_sma(dd, SMA_IDLE_ARM);
1282 break;
1283 case IB_PORT_ACTIVE:
1284 if (ppd->neighbor_normal) {
1285 ret = set_link_state(ppd, HLS_UP_ACTIVE);
1286 if (ret == 0)
1287 send_idle_sma(dd, SMA_IDLE_ACTIVE);
1288 } else {
1289 pr_warn("SubnSet(OPA_PortInfo) Cannot move to Active with NeighborNormal 0\n");
1290 smp->status |= IB_SMP_INVALID_FIELD;
1291 }
1292 break;
1293 default:
1294 pr_warn("SubnSet(OPA_PortInfo) invalid logical state 0x%x\n",
1295 logical_state);
1296 smp->status |= IB_SMP_INVALID_FIELD;
1297 }
1298
1299 return 0;
1300 }
1301
1302 /*
1303 * subn_set_opa_portinfo - set port information
1304 * @smp: the incoming SM packet
1305 * @ibdev: the infiniband device
1306 * @port: the port on the device
1307 *
1308 */
__subn_set_opa_portinfo(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len,int local_mad)1309 static int __subn_set_opa_portinfo(struct opa_smp *smp, u32 am, u8 *data,
1310 struct ib_device *ibdev, u32 port,
1311 u32 *resp_len, u32 max_len, int local_mad)
1312 {
1313 struct opa_port_info *pi = (struct opa_port_info *)data;
1314 struct ib_event event;
1315 struct hfi1_devdata *dd;
1316 struct hfi1_pportdata *ppd;
1317 struct hfi1_ibport *ibp;
1318 u8 clientrereg;
1319 unsigned long flags;
1320 u32 smlid;
1321 u32 lid;
1322 u8 ls_old, ls_new, ps_new;
1323 u8 vls;
1324 u8 msl;
1325 u8 crc_enabled;
1326 u16 lse, lwe, mtu;
1327 u32 num_ports = OPA_AM_NPORT(am);
1328 u32 start_of_sm_config = OPA_AM_START_SM_CFG(am);
1329 int ret, i, invalid = 0, call_set_mtu = 0;
1330 int call_link_downgrade_policy = 0;
1331
1332 if (num_ports != 1 ||
1333 smp_length_check(sizeof(*pi), max_len)) {
1334 smp->status |= IB_SMP_INVALID_FIELD;
1335 return reply((struct ib_mad_hdr *)smp);
1336 }
1337
1338 lid = be32_to_cpu(pi->lid);
1339 if (lid & 0xFF000000) {
1340 pr_warn("OPA_PortInfo lid out of range: %X\n", lid);
1341 smp->status |= IB_SMP_INVALID_FIELD;
1342 goto get_only;
1343 }
1344
1345
1346 smlid = be32_to_cpu(pi->sm_lid);
1347 if (smlid & 0xFF000000) {
1348 pr_warn("OPA_PortInfo SM lid out of range: %X\n", smlid);
1349 smp->status |= IB_SMP_INVALID_FIELD;
1350 goto get_only;
1351 }
1352
1353 clientrereg = (pi->clientrereg_subnettimeout &
1354 OPA_PI_MASK_CLIENT_REREGISTER);
1355
1356 dd = dd_from_ibdev(ibdev);
1357 /* IB numbers ports from 1, hw from 0 */
1358 ppd = dd->pport + (port - 1);
1359 ibp = &ppd->ibport_data;
1360 event.device = ibdev;
1361 event.element.port_num = port;
1362
1363 ls_old = driver_lstate(ppd);
1364
1365 ibp->rvp.mkey = pi->mkey;
1366 if (ibp->rvp.gid_prefix != pi->subnet_prefix) {
1367 ibp->rvp.gid_prefix = pi->subnet_prefix;
1368 event.event = IB_EVENT_GID_CHANGE;
1369 ib_dispatch_event(&event);
1370 }
1371 ibp->rvp.mkey_lease_period = be16_to_cpu(pi->mkey_lease_period);
1372
1373 /* Must be a valid unicast LID address. */
1374 if ((lid == 0 && ls_old > IB_PORT_INIT) ||
1375 (hfi1_is_16B_mcast(lid))) {
1376 smp->status |= IB_SMP_INVALID_FIELD;
1377 pr_warn("SubnSet(OPA_PortInfo) lid invalid 0x%x\n",
1378 lid);
1379 } else if (ppd->lid != lid ||
1380 ppd->lmc != (pi->mkeyprotect_lmc & OPA_PI_MASK_LMC)) {
1381 if (ppd->lid != lid)
1382 hfi1_set_uevent_bits(ppd, _HFI1_EVENT_LID_CHANGE_BIT);
1383 if (ppd->lmc != (pi->mkeyprotect_lmc & OPA_PI_MASK_LMC))
1384 hfi1_set_uevent_bits(ppd, _HFI1_EVENT_LMC_CHANGE_BIT);
1385 hfi1_set_lid(ppd, lid, pi->mkeyprotect_lmc & OPA_PI_MASK_LMC);
1386 event.event = IB_EVENT_LID_CHANGE;
1387 ib_dispatch_event(&event);
1388
1389 if (HFI1_PORT_GUID_INDEX + 1 < HFI1_GUIDS_PER_PORT) {
1390 /* Manufacture GID from LID to support extended
1391 * addresses
1392 */
1393 ppd->guids[HFI1_PORT_GUID_INDEX + 1] =
1394 be64_to_cpu(OPA_MAKE_ID(lid));
1395 event.event = IB_EVENT_GID_CHANGE;
1396 ib_dispatch_event(&event);
1397 }
1398 }
1399
1400 msl = pi->smsl & OPA_PI_MASK_SMSL;
1401 if (pi->partenforce_filterraw & OPA_PI_MASK_LINKINIT_REASON)
1402 ppd->linkinit_reason =
1403 (pi->partenforce_filterraw &
1404 OPA_PI_MASK_LINKINIT_REASON);
1405
1406 /* Must be a valid unicast LID address. */
1407 if ((smlid == 0 && ls_old > IB_PORT_INIT) ||
1408 (hfi1_is_16B_mcast(smlid))) {
1409 smp->status |= IB_SMP_INVALID_FIELD;
1410 pr_warn("SubnSet(OPA_PortInfo) smlid invalid 0x%x\n", smlid);
1411 } else if (smlid != ibp->rvp.sm_lid || msl != ibp->rvp.sm_sl) {
1412 pr_warn("SubnSet(OPA_PortInfo) smlid 0x%x\n", smlid);
1413 spin_lock_irqsave(&ibp->rvp.lock, flags);
1414 if (ibp->rvp.sm_ah) {
1415 if (smlid != ibp->rvp.sm_lid)
1416 hfi1_modify_qp0_ah(ibp, ibp->rvp.sm_ah, smlid);
1417 if (msl != ibp->rvp.sm_sl)
1418 rdma_ah_set_sl(&ibp->rvp.sm_ah->attr, msl);
1419 }
1420 spin_unlock_irqrestore(&ibp->rvp.lock, flags);
1421 if (smlid != ibp->rvp.sm_lid)
1422 ibp->rvp.sm_lid = smlid;
1423 if (msl != ibp->rvp.sm_sl)
1424 ibp->rvp.sm_sl = msl;
1425 event.event = IB_EVENT_SM_CHANGE;
1426 ib_dispatch_event(&event);
1427 }
1428
1429 if (pi->link_down_reason == 0) {
1430 ppd->local_link_down_reason.sma = 0;
1431 ppd->local_link_down_reason.latest = 0;
1432 }
1433
1434 if (pi->neigh_link_down_reason == 0) {
1435 ppd->neigh_link_down_reason.sma = 0;
1436 ppd->neigh_link_down_reason.latest = 0;
1437 }
1438
1439 ppd->sm_trap_qp = be32_to_cpu(pi->sm_trap_qp);
1440 ppd->sa_qp = be32_to_cpu(pi->sa_qp);
1441
1442 ppd->port_error_action = be32_to_cpu(pi->port_error_action);
1443 lwe = be16_to_cpu(pi->link_width.enabled);
1444 if (lwe) {
1445 if (lwe == OPA_LINK_WIDTH_RESET ||
1446 lwe == OPA_LINK_WIDTH_RESET_OLD)
1447 set_link_width_enabled(ppd, ppd->link_width_supported);
1448 else if ((lwe & ~ppd->link_width_supported) == 0)
1449 set_link_width_enabled(ppd, lwe);
1450 else
1451 smp->status |= IB_SMP_INVALID_FIELD;
1452 }
1453 lwe = be16_to_cpu(pi->link_width_downgrade.enabled);
1454 /* LWD.E is always applied - 0 means "disabled" */
1455 if (lwe == OPA_LINK_WIDTH_RESET ||
1456 lwe == OPA_LINK_WIDTH_RESET_OLD) {
1457 set_link_width_downgrade_enabled(ppd,
1458 ppd->
1459 link_width_downgrade_supported
1460 );
1461 } else if ((lwe & ~ppd->link_width_downgrade_supported) == 0) {
1462 /* only set and apply if something changed */
1463 if (lwe != ppd->link_width_downgrade_enabled) {
1464 set_link_width_downgrade_enabled(ppd, lwe);
1465 call_link_downgrade_policy = 1;
1466 }
1467 } else {
1468 smp->status |= IB_SMP_INVALID_FIELD;
1469 }
1470 lse = be16_to_cpu(pi->link_speed.enabled);
1471 if (lse) {
1472 if (lse & be16_to_cpu(pi->link_speed.supported))
1473 set_link_speed_enabled(ppd, lse);
1474 else
1475 smp->status |= IB_SMP_INVALID_FIELD;
1476 }
1477
1478 ibp->rvp.mkeyprot =
1479 (pi->mkeyprotect_lmc & OPA_PI_MASK_MKEY_PROT_BIT) >> 6;
1480 ibp->rvp.vl_high_limit = be16_to_cpu(pi->vl.high_limit) & 0xFF;
1481 (void)hfi1_set_ib_cfg(ppd, HFI1_IB_CFG_VL_HIGH_LIMIT,
1482 ibp->rvp.vl_high_limit);
1483
1484 if (ppd->vls_supported / 2 > ARRAY_SIZE(pi->neigh_mtu.pvlx_to_mtu) ||
1485 ppd->vls_supported > ARRAY_SIZE(dd->vld)) {
1486 smp->status |= IB_SMP_INVALID_FIELD;
1487 return reply((struct ib_mad_hdr *)smp);
1488 }
1489 for (i = 0; i < ppd->vls_supported; i++) {
1490 if ((i % 2) == 0)
1491 mtu = enum_to_mtu((pi->neigh_mtu.pvlx_to_mtu[i / 2] >>
1492 4) & 0xF);
1493 else
1494 mtu = enum_to_mtu(pi->neigh_mtu.pvlx_to_mtu[i / 2] &
1495 0xF);
1496 if (mtu == 0xffff) {
1497 pr_warn("SubnSet(OPA_PortInfo) mtu invalid %d (0x%x)\n",
1498 mtu,
1499 (pi->neigh_mtu.pvlx_to_mtu[0] >> 4) & 0xF);
1500 smp->status |= IB_SMP_INVALID_FIELD;
1501 mtu = hfi1_max_mtu; /* use a valid MTU */
1502 }
1503 if (dd->vld[i].mtu != mtu) {
1504 dd_dev_info(dd,
1505 "MTU change on vl %d from %d to %d\n",
1506 i, dd->vld[i].mtu, mtu);
1507 dd->vld[i].mtu = mtu;
1508 call_set_mtu++;
1509 }
1510 }
1511 /* As per OPAV1 spec: VL15 must support and be configured
1512 * for operation with a 2048 or larger MTU.
1513 */
1514 mtu = enum_to_mtu(pi->neigh_mtu.pvlx_to_mtu[15 / 2] & 0xF);
1515 if (mtu < 2048 || mtu == 0xffff)
1516 mtu = 2048;
1517 if (dd->vld[15].mtu != mtu) {
1518 dd_dev_info(dd,
1519 "MTU change on vl 15 from %d to %d\n",
1520 dd->vld[15].mtu, mtu);
1521 dd->vld[15].mtu = mtu;
1522 call_set_mtu++;
1523 }
1524 if (call_set_mtu)
1525 set_mtu(ppd);
1526
1527 /* Set operational VLs */
1528 vls = pi->operational_vls & OPA_PI_MASK_OPERATIONAL_VL;
1529 if (vls) {
1530 if (vls > ppd->vls_supported) {
1531 pr_warn("SubnSet(OPA_PortInfo) VL's supported invalid %d\n",
1532 pi->operational_vls);
1533 smp->status |= IB_SMP_INVALID_FIELD;
1534 } else {
1535 if (hfi1_set_ib_cfg(ppd, HFI1_IB_CFG_OP_VLS,
1536 vls) == -EINVAL)
1537 smp->status |= IB_SMP_INVALID_FIELD;
1538 }
1539 }
1540
1541 if (pi->mkey_violations == 0)
1542 ibp->rvp.mkey_violations = 0;
1543
1544 if (pi->pkey_violations == 0)
1545 ibp->rvp.pkey_violations = 0;
1546
1547 if (pi->qkey_violations == 0)
1548 ibp->rvp.qkey_violations = 0;
1549
1550 ibp->rvp.subnet_timeout =
1551 pi->clientrereg_subnettimeout & OPA_PI_MASK_SUBNET_TIMEOUT;
1552
1553 crc_enabled = be16_to_cpu(pi->port_ltp_crc_mode);
1554 crc_enabled >>= 4;
1555 crc_enabled &= 0xf;
1556
1557 if (crc_enabled != 0)
1558 ppd->port_crc_mode_enabled = port_ltp_to_cap(crc_enabled);
1559
1560 ppd->is_active_optimize_enabled =
1561 !!(be16_to_cpu(pi->port_mode)
1562 & OPA_PI_MASK_PORT_ACTIVE_OPTOMIZE);
1563
1564 ls_new = pi->port_states.portphysstate_portstate &
1565 OPA_PI_MASK_PORT_STATE;
1566 ps_new = (pi->port_states.portphysstate_portstate &
1567 OPA_PI_MASK_PORT_PHYSICAL_STATE) >> 4;
1568
1569 if (ls_old == IB_PORT_INIT) {
1570 if (start_of_sm_config) {
1571 if (ls_new == ls_old || (ls_new == IB_PORT_ARMED))
1572 ppd->is_sm_config_started = 1;
1573 } else if (ls_new == IB_PORT_ARMED) {
1574 if (ppd->is_sm_config_started == 0) {
1575 invalid = 1;
1576 smp->status |= IB_SMP_INVALID_FIELD;
1577 }
1578 }
1579 }
1580
1581 /* Handle CLIENT_REREGISTER event b/c SM asked us for it */
1582 if (clientrereg) {
1583 event.event = IB_EVENT_CLIENT_REREGISTER;
1584 ib_dispatch_event(&event);
1585 }
1586
1587 /*
1588 * Do the port state change now that the other link parameters
1589 * have been set.
1590 * Changing the port physical state only makes sense if the link
1591 * is down or is being set to down.
1592 */
1593
1594 if (!invalid) {
1595 ret = set_port_states(ppd, smp, ls_new, ps_new, local_mad);
1596 if (ret)
1597 return ret;
1598 }
1599
1600 ret = __subn_get_opa_portinfo(smp, am, data, ibdev, port, resp_len,
1601 max_len);
1602
1603 /* restore re-reg bit per o14-12.2.1 */
1604 pi->clientrereg_subnettimeout |= clientrereg;
1605
1606 /*
1607 * Apply the new link downgrade policy. This may result in a link
1608 * bounce. Do this after everything else so things are settled.
1609 * Possible problem: if setting the port state above fails, then
1610 * the policy change is not applied.
1611 */
1612 if (call_link_downgrade_policy)
1613 apply_link_downgrade_policy(ppd, 0);
1614
1615 return ret;
1616
1617 get_only:
1618 return __subn_get_opa_portinfo(smp, am, data, ibdev, port, resp_len,
1619 max_len);
1620 }
1621
1622 /**
1623 * set_pkeys - set the PKEY table for ctxt 0
1624 * @dd: the hfi1_ib device
1625 * @port: the IB port number
1626 * @pkeys: the PKEY table
1627 */
set_pkeys(struct hfi1_devdata * dd,u32 port,u16 * pkeys)1628 static int set_pkeys(struct hfi1_devdata *dd, u32 port, u16 *pkeys)
1629 {
1630 struct hfi1_pportdata *ppd;
1631 int i;
1632 int changed = 0;
1633 int update_includes_mgmt_partition = 0;
1634
1635 /*
1636 * IB port one/two always maps to context zero/one,
1637 * always a kernel context, no locking needed
1638 * If we get here with ppd setup, no need to check
1639 * that rcd is valid.
1640 */
1641 ppd = dd->pport + (port - 1);
1642 /*
1643 * If the update does not include the management pkey, don't do it.
1644 */
1645 for (i = 0; i < ARRAY_SIZE(ppd->pkeys); i++) {
1646 if (pkeys[i] == LIM_MGMT_P_KEY) {
1647 update_includes_mgmt_partition = 1;
1648 break;
1649 }
1650 }
1651
1652 if (!update_includes_mgmt_partition)
1653 return 1;
1654
1655 for (i = 0; i < ARRAY_SIZE(ppd->pkeys); i++) {
1656 u16 key = pkeys[i];
1657 u16 okey = ppd->pkeys[i];
1658
1659 if (key == okey)
1660 continue;
1661 /*
1662 * The SM gives us the complete PKey table. We have
1663 * to ensure that we put the PKeys in the matching
1664 * slots.
1665 */
1666 ppd->pkeys[i] = key;
1667 changed = 1;
1668 }
1669
1670 if (changed) {
1671 (void)hfi1_set_ib_cfg(ppd, HFI1_IB_CFG_PKEYS, 0);
1672 hfi1_event_pkey_change(dd, port);
1673 }
1674
1675 return 0;
1676 }
1677
__subn_set_opa_pkeytable(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)1678 static int __subn_set_opa_pkeytable(struct opa_smp *smp, u32 am, u8 *data,
1679 struct ib_device *ibdev, u32 port,
1680 u32 *resp_len, u32 max_len)
1681 {
1682 struct hfi1_devdata *dd = dd_from_ibdev(ibdev);
1683 u32 n_blocks_sent = OPA_AM_NBLK(am);
1684 u32 start_block = am & 0x7ff;
1685 u16 *p = (u16 *)data;
1686 __be16 *q = (__be16 *)data;
1687 int i;
1688 u16 n_blocks_avail;
1689 unsigned npkeys = hfi1_get_npkeys(dd);
1690 u32 size = 0;
1691
1692 if (n_blocks_sent == 0) {
1693 pr_warn("OPA Get PKey AM Invalid : P = %u; B = 0x%x; N = 0x%x\n",
1694 port, start_block, n_blocks_sent);
1695 smp->status |= IB_SMP_INVALID_FIELD;
1696 return reply((struct ib_mad_hdr *)smp);
1697 }
1698
1699 n_blocks_avail = (u16)(npkeys / OPA_PARTITION_TABLE_BLK_SIZE) + 1;
1700
1701 size = sizeof(u16) * (n_blocks_sent * OPA_PARTITION_TABLE_BLK_SIZE);
1702
1703 if (smp_length_check(size, max_len)) {
1704 smp->status |= IB_SMP_INVALID_FIELD;
1705 return reply((struct ib_mad_hdr *)smp);
1706 }
1707
1708 if (start_block + n_blocks_sent > n_blocks_avail ||
1709 n_blocks_sent > OPA_NUM_PKEY_BLOCKS_PER_SMP) {
1710 pr_warn("OPA Set PKey AM Invalid : s 0x%x; req 0x%x; avail 0x%x; blk/smp 0x%lx\n",
1711 start_block, n_blocks_sent, n_blocks_avail,
1712 OPA_NUM_PKEY_BLOCKS_PER_SMP);
1713 smp->status |= IB_SMP_INVALID_FIELD;
1714 return reply((struct ib_mad_hdr *)smp);
1715 }
1716
1717 for (i = 0; i < n_blocks_sent * OPA_PARTITION_TABLE_BLK_SIZE; i++)
1718 p[i] = be16_to_cpu(q[i]);
1719
1720 if (start_block == 0 && set_pkeys(dd, port, p) != 0) {
1721 smp->status |= IB_SMP_INVALID_FIELD;
1722 return reply((struct ib_mad_hdr *)smp);
1723 }
1724
1725 return __subn_get_opa_pkeytable(smp, am, data, ibdev, port, resp_len,
1726 max_len);
1727 }
1728
1729 #define ILLEGAL_VL 12
1730 /*
1731 * filter_sc2vlt changes mappings to VL15 to ILLEGAL_VL (except
1732 * for SC15, which must map to VL15). If we don't remap things this
1733 * way it is possible for VL15 counters to increment when we try to
1734 * send on a SC which is mapped to an invalid VL.
1735 * When getting the table convert ILLEGAL_VL back to VL15.
1736 */
filter_sc2vlt(void * data,bool set)1737 static void filter_sc2vlt(void *data, bool set)
1738 {
1739 int i;
1740 u8 *pd = data;
1741
1742 for (i = 0; i < OPA_MAX_SCS; i++) {
1743 if (i == 15)
1744 continue;
1745
1746 if (set) {
1747 if ((pd[i] & 0x1f) == 0xf)
1748 pd[i] = ILLEGAL_VL;
1749 } else {
1750 if ((pd[i] & 0x1f) == ILLEGAL_VL)
1751 pd[i] = 0xf;
1752 }
1753 }
1754 }
1755
set_sc2vlt_tables(struct hfi1_devdata * dd,void * data)1756 static int set_sc2vlt_tables(struct hfi1_devdata *dd, void *data)
1757 {
1758 u64 *val = data;
1759
1760 filter_sc2vlt(data, true);
1761
1762 write_csr(dd, SEND_SC2VLT0, *val++);
1763 write_csr(dd, SEND_SC2VLT1, *val++);
1764 write_csr(dd, SEND_SC2VLT2, *val++);
1765 write_csr(dd, SEND_SC2VLT3, *val++);
1766 write_seqlock_irq(&dd->sc2vl_lock);
1767 memcpy(dd->sc2vl, data, sizeof(dd->sc2vl));
1768 write_sequnlock_irq(&dd->sc2vl_lock);
1769 return 0;
1770 }
1771
get_sc2vlt_tables(struct hfi1_devdata * dd,void * data)1772 static int get_sc2vlt_tables(struct hfi1_devdata *dd, void *data)
1773 {
1774 u64 *val = (u64 *)data;
1775
1776 *val++ = read_csr(dd, SEND_SC2VLT0);
1777 *val++ = read_csr(dd, SEND_SC2VLT1);
1778 *val++ = read_csr(dd, SEND_SC2VLT2);
1779 *val++ = read_csr(dd, SEND_SC2VLT3);
1780
1781 filter_sc2vlt((u64 *)data, false);
1782 return 0;
1783 }
1784
__subn_get_opa_sl_to_sc(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)1785 static int __subn_get_opa_sl_to_sc(struct opa_smp *smp, u32 am, u8 *data,
1786 struct ib_device *ibdev, u32 port,
1787 u32 *resp_len, u32 max_len)
1788 {
1789 struct hfi1_ibport *ibp = to_iport(ibdev, port);
1790 u8 *p = data;
1791 size_t size = ARRAY_SIZE(ibp->sl_to_sc); /* == 32 */
1792 unsigned i;
1793
1794 if (am || smp_length_check(size, max_len)) {
1795 smp->status |= IB_SMP_INVALID_FIELD;
1796 return reply((struct ib_mad_hdr *)smp);
1797 }
1798
1799 for (i = 0; i < ARRAY_SIZE(ibp->sl_to_sc); i++)
1800 *p++ = ibp->sl_to_sc[i];
1801
1802 if (resp_len)
1803 *resp_len += size;
1804
1805 return reply((struct ib_mad_hdr *)smp);
1806 }
1807
__subn_set_opa_sl_to_sc(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)1808 static int __subn_set_opa_sl_to_sc(struct opa_smp *smp, u32 am, u8 *data,
1809 struct ib_device *ibdev, u32 port,
1810 u32 *resp_len, u32 max_len)
1811 {
1812 struct hfi1_ibport *ibp = to_iport(ibdev, port);
1813 u8 *p = data;
1814 size_t size = ARRAY_SIZE(ibp->sl_to_sc);
1815 int i;
1816 u8 sc;
1817
1818 if (am || smp_length_check(size, max_len)) {
1819 smp->status |= IB_SMP_INVALID_FIELD;
1820 return reply((struct ib_mad_hdr *)smp);
1821 }
1822
1823 for (i = 0; i < ARRAY_SIZE(ibp->sl_to_sc); i++) {
1824 sc = *p++;
1825 if (ibp->sl_to_sc[i] != sc) {
1826 ibp->sl_to_sc[i] = sc;
1827
1828 /* Put all stale qps into error state */
1829 hfi1_error_port_qps(ibp, i);
1830 }
1831 }
1832
1833 return __subn_get_opa_sl_to_sc(smp, am, data, ibdev, port, resp_len,
1834 max_len);
1835 }
1836
__subn_get_opa_sc_to_sl(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)1837 static int __subn_get_opa_sc_to_sl(struct opa_smp *smp, u32 am, u8 *data,
1838 struct ib_device *ibdev, u32 port,
1839 u32 *resp_len, u32 max_len)
1840 {
1841 struct hfi1_ibport *ibp = to_iport(ibdev, port);
1842 u8 *p = data;
1843 size_t size = ARRAY_SIZE(ibp->sc_to_sl); /* == 32 */
1844 unsigned i;
1845
1846 if (am || smp_length_check(size, max_len)) {
1847 smp->status |= IB_SMP_INVALID_FIELD;
1848 return reply((struct ib_mad_hdr *)smp);
1849 }
1850
1851 for (i = 0; i < ARRAY_SIZE(ibp->sc_to_sl); i++)
1852 *p++ = ibp->sc_to_sl[i];
1853
1854 if (resp_len)
1855 *resp_len += size;
1856
1857 return reply((struct ib_mad_hdr *)smp);
1858 }
1859
__subn_set_opa_sc_to_sl(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)1860 static int __subn_set_opa_sc_to_sl(struct opa_smp *smp, u32 am, u8 *data,
1861 struct ib_device *ibdev, u32 port,
1862 u32 *resp_len, u32 max_len)
1863 {
1864 struct hfi1_ibport *ibp = to_iport(ibdev, port);
1865 size_t size = ARRAY_SIZE(ibp->sc_to_sl);
1866 u8 *p = data;
1867 int i;
1868
1869 if (am || smp_length_check(size, max_len)) {
1870 smp->status |= IB_SMP_INVALID_FIELD;
1871 return reply((struct ib_mad_hdr *)smp);
1872 }
1873
1874 for (i = 0; i < ARRAY_SIZE(ibp->sc_to_sl); i++)
1875 ibp->sc_to_sl[i] = *p++;
1876
1877 return __subn_get_opa_sc_to_sl(smp, am, data, ibdev, port, resp_len,
1878 max_len);
1879 }
1880
__subn_get_opa_sc_to_vlt(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)1881 static int __subn_get_opa_sc_to_vlt(struct opa_smp *smp, u32 am, u8 *data,
1882 struct ib_device *ibdev, u32 port,
1883 u32 *resp_len, u32 max_len)
1884 {
1885 u32 n_blocks = OPA_AM_NBLK(am);
1886 struct hfi1_devdata *dd = dd_from_ibdev(ibdev);
1887 void *vp = (void *)data;
1888 size_t size = 4 * sizeof(u64);
1889
1890 if (n_blocks != 1 || smp_length_check(size, max_len)) {
1891 smp->status |= IB_SMP_INVALID_FIELD;
1892 return reply((struct ib_mad_hdr *)smp);
1893 }
1894
1895 get_sc2vlt_tables(dd, vp);
1896
1897 if (resp_len)
1898 *resp_len += size;
1899
1900 return reply((struct ib_mad_hdr *)smp);
1901 }
1902
__subn_set_opa_sc_to_vlt(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)1903 static int __subn_set_opa_sc_to_vlt(struct opa_smp *smp, u32 am, u8 *data,
1904 struct ib_device *ibdev, u32 port,
1905 u32 *resp_len, u32 max_len)
1906 {
1907 u32 n_blocks = OPA_AM_NBLK(am);
1908 int async_update = OPA_AM_ASYNC(am);
1909 struct hfi1_devdata *dd = dd_from_ibdev(ibdev);
1910 void *vp = (void *)data;
1911 struct hfi1_pportdata *ppd;
1912 int lstate;
1913 /*
1914 * set_sc2vlt_tables writes the information contained in *data
1915 * to four 64-bit registers SendSC2VLt[0-3]. We need to make
1916 * sure *max_len is not greater than the total size of the four
1917 * SendSC2VLt[0-3] registers.
1918 */
1919 size_t size = 4 * sizeof(u64);
1920
1921 if (n_blocks != 1 || async_update || smp_length_check(size, max_len)) {
1922 smp->status |= IB_SMP_INVALID_FIELD;
1923 return reply((struct ib_mad_hdr *)smp);
1924 }
1925
1926 /* IB numbers ports from 1, hw from 0 */
1927 ppd = dd->pport + (port - 1);
1928 lstate = driver_lstate(ppd);
1929 /*
1930 * it's known that async_update is 0 by this point, but include
1931 * the explicit check for clarity
1932 */
1933 if (!async_update &&
1934 (lstate == IB_PORT_ARMED || lstate == IB_PORT_ACTIVE)) {
1935 smp->status |= IB_SMP_INVALID_FIELD;
1936 return reply((struct ib_mad_hdr *)smp);
1937 }
1938
1939 set_sc2vlt_tables(dd, vp);
1940
1941 return __subn_get_opa_sc_to_vlt(smp, am, data, ibdev, port, resp_len,
1942 max_len);
1943 }
1944
__subn_get_opa_sc_to_vlnt(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)1945 static int __subn_get_opa_sc_to_vlnt(struct opa_smp *smp, u32 am, u8 *data,
1946 struct ib_device *ibdev, u32 port,
1947 u32 *resp_len, u32 max_len)
1948 {
1949 u32 n_blocks = OPA_AM_NPORT(am);
1950 struct hfi1_devdata *dd = dd_from_ibdev(ibdev);
1951 struct hfi1_pportdata *ppd;
1952 void *vp = (void *)data;
1953 int size = sizeof(struct sc2vlnt);
1954
1955 if (n_blocks != 1 || smp_length_check(size, max_len)) {
1956 smp->status |= IB_SMP_INVALID_FIELD;
1957 return reply((struct ib_mad_hdr *)smp);
1958 }
1959
1960 ppd = dd->pport + (port - 1);
1961
1962 fm_get_table(ppd, FM_TBL_SC2VLNT, vp);
1963
1964 if (resp_len)
1965 *resp_len += size;
1966
1967 return reply((struct ib_mad_hdr *)smp);
1968 }
1969
__subn_set_opa_sc_to_vlnt(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)1970 static int __subn_set_opa_sc_to_vlnt(struct opa_smp *smp, u32 am, u8 *data,
1971 struct ib_device *ibdev, u32 port,
1972 u32 *resp_len, u32 max_len)
1973 {
1974 u32 n_blocks = OPA_AM_NPORT(am);
1975 struct hfi1_devdata *dd = dd_from_ibdev(ibdev);
1976 struct hfi1_pportdata *ppd;
1977 void *vp = (void *)data;
1978 int lstate;
1979 int size = sizeof(struct sc2vlnt);
1980
1981 if (n_blocks != 1 || smp_length_check(size, max_len)) {
1982 smp->status |= IB_SMP_INVALID_FIELD;
1983 return reply((struct ib_mad_hdr *)smp);
1984 }
1985
1986 /* IB numbers ports from 1, hw from 0 */
1987 ppd = dd->pport + (port - 1);
1988 lstate = driver_lstate(ppd);
1989 if (lstate == IB_PORT_ARMED || lstate == IB_PORT_ACTIVE) {
1990 smp->status |= IB_SMP_INVALID_FIELD;
1991 return reply((struct ib_mad_hdr *)smp);
1992 }
1993
1994 ppd = dd->pport + (port - 1);
1995
1996 fm_set_table(ppd, FM_TBL_SC2VLNT, vp);
1997
1998 return __subn_get_opa_sc_to_vlnt(smp, am, data, ibdev, port,
1999 resp_len, max_len);
2000 }
2001
__subn_get_opa_psi(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)2002 static int __subn_get_opa_psi(struct opa_smp *smp, u32 am, u8 *data,
2003 struct ib_device *ibdev, u32 port,
2004 u32 *resp_len, u32 max_len)
2005 {
2006 u32 nports = OPA_AM_NPORT(am);
2007 u32 start_of_sm_config = OPA_AM_START_SM_CFG(am);
2008 u32 lstate;
2009 struct hfi1_ibport *ibp;
2010 struct hfi1_pportdata *ppd;
2011 struct opa_port_state_info *psi = (struct opa_port_state_info *)data;
2012
2013 if (nports != 1 || smp_length_check(sizeof(*psi), max_len)) {
2014 smp->status |= IB_SMP_INVALID_FIELD;
2015 return reply((struct ib_mad_hdr *)smp);
2016 }
2017
2018 ibp = to_iport(ibdev, port);
2019 ppd = ppd_from_ibp(ibp);
2020
2021 lstate = driver_lstate(ppd);
2022
2023 if (start_of_sm_config && (lstate == IB_PORT_INIT))
2024 ppd->is_sm_config_started = 1;
2025
2026 psi->port_states.ledenable_offlinereason = ppd->neighbor_normal << 4;
2027 psi->port_states.ledenable_offlinereason |=
2028 ppd->is_sm_config_started << 5;
2029 psi->port_states.ledenable_offlinereason |=
2030 ppd->offline_disabled_reason;
2031
2032 psi->port_states.portphysstate_portstate =
2033 (driver_pstate(ppd) << 4) | (lstate & 0xf);
2034 psi->link_width_downgrade_tx_active =
2035 cpu_to_be16(ppd->link_width_downgrade_tx_active);
2036 psi->link_width_downgrade_rx_active =
2037 cpu_to_be16(ppd->link_width_downgrade_rx_active);
2038 if (resp_len)
2039 *resp_len += sizeof(struct opa_port_state_info);
2040
2041 return reply((struct ib_mad_hdr *)smp);
2042 }
2043
__subn_set_opa_psi(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len,int local_mad)2044 static int __subn_set_opa_psi(struct opa_smp *smp, u32 am, u8 *data,
2045 struct ib_device *ibdev, u32 port,
2046 u32 *resp_len, u32 max_len, int local_mad)
2047 {
2048 u32 nports = OPA_AM_NPORT(am);
2049 u32 start_of_sm_config = OPA_AM_START_SM_CFG(am);
2050 u32 ls_old;
2051 u8 ls_new, ps_new;
2052 struct hfi1_ibport *ibp;
2053 struct hfi1_pportdata *ppd;
2054 struct opa_port_state_info *psi = (struct opa_port_state_info *)data;
2055 int ret, invalid = 0;
2056
2057 if (nports != 1 || smp_length_check(sizeof(*psi), max_len)) {
2058 smp->status |= IB_SMP_INVALID_FIELD;
2059 return reply((struct ib_mad_hdr *)smp);
2060 }
2061
2062 ibp = to_iport(ibdev, port);
2063 ppd = ppd_from_ibp(ibp);
2064
2065 ls_old = driver_lstate(ppd);
2066
2067 ls_new = port_states_to_logical_state(&psi->port_states);
2068 ps_new = port_states_to_phys_state(&psi->port_states);
2069
2070 if (ls_old == IB_PORT_INIT) {
2071 if (start_of_sm_config) {
2072 if (ls_new == ls_old || (ls_new == IB_PORT_ARMED))
2073 ppd->is_sm_config_started = 1;
2074 } else if (ls_new == IB_PORT_ARMED) {
2075 if (ppd->is_sm_config_started == 0) {
2076 invalid = 1;
2077 smp->status |= IB_SMP_INVALID_FIELD;
2078 }
2079 }
2080 }
2081
2082 if (!invalid) {
2083 ret = set_port_states(ppd, smp, ls_new, ps_new, local_mad);
2084 if (ret)
2085 return ret;
2086 }
2087
2088 return __subn_get_opa_psi(smp, am, data, ibdev, port, resp_len,
2089 max_len);
2090 }
2091
__subn_get_opa_cable_info(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)2092 static int __subn_get_opa_cable_info(struct opa_smp *smp, u32 am, u8 *data,
2093 struct ib_device *ibdev, u32 port,
2094 u32 *resp_len, u32 max_len)
2095 {
2096 struct hfi1_devdata *dd = dd_from_ibdev(ibdev);
2097 u32 addr = OPA_AM_CI_ADDR(am);
2098 u32 len = OPA_AM_CI_LEN(am) + 1;
2099 int ret;
2100
2101 if (dd->pport->port_type != PORT_TYPE_QSFP ||
2102 smp_length_check(len, max_len)) {
2103 smp->status |= IB_SMP_INVALID_FIELD;
2104 return reply((struct ib_mad_hdr *)smp);
2105 }
2106
2107 #define __CI_PAGE_SIZE BIT(7) /* 128 bytes */
2108 #define __CI_PAGE_MASK ~(__CI_PAGE_SIZE - 1)
2109 #define __CI_PAGE_NUM(a) ((a) & __CI_PAGE_MASK)
2110
2111 /*
2112 * check that addr is within spec, and
2113 * addr and (addr + len - 1) are on the same "page"
2114 */
2115 if (addr >= 4096 ||
2116 (__CI_PAGE_NUM(addr) != __CI_PAGE_NUM(addr + len - 1))) {
2117 smp->status |= IB_SMP_INVALID_FIELD;
2118 return reply((struct ib_mad_hdr *)smp);
2119 }
2120
2121 ret = get_cable_info(dd, port, addr, len, data);
2122
2123 if (ret == -ENODEV) {
2124 smp->status |= IB_SMP_UNSUP_METH_ATTR;
2125 return reply((struct ib_mad_hdr *)smp);
2126 }
2127
2128 /* The address range for the CableInfo SMA query is wider than the
2129 * memory available on the QSFP cable. We want to return a valid
2130 * response, albeit zeroed out, for address ranges beyond available
2131 * memory but that are within the CableInfo query spec
2132 */
2133 if (ret < 0 && ret != -ERANGE) {
2134 smp->status |= IB_SMP_INVALID_FIELD;
2135 return reply((struct ib_mad_hdr *)smp);
2136 }
2137
2138 if (resp_len)
2139 *resp_len += len;
2140
2141 return reply((struct ib_mad_hdr *)smp);
2142 }
2143
__subn_get_opa_bct(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)2144 static int __subn_get_opa_bct(struct opa_smp *smp, u32 am, u8 *data,
2145 struct ib_device *ibdev, u32 port, u32 *resp_len,
2146 u32 max_len)
2147 {
2148 u32 num_ports = OPA_AM_NPORT(am);
2149 struct hfi1_devdata *dd = dd_from_ibdev(ibdev);
2150 struct hfi1_pportdata *ppd;
2151 struct buffer_control *p = (struct buffer_control *)data;
2152 int size = sizeof(struct buffer_control);
2153
2154 if (num_ports != 1 || smp_length_check(size, max_len)) {
2155 smp->status |= IB_SMP_INVALID_FIELD;
2156 return reply((struct ib_mad_hdr *)smp);
2157 }
2158
2159 ppd = dd->pport + (port - 1);
2160 fm_get_table(ppd, FM_TBL_BUFFER_CONTROL, p);
2161 trace_bct_get(dd, p);
2162 if (resp_len)
2163 *resp_len += size;
2164
2165 return reply((struct ib_mad_hdr *)smp);
2166 }
2167
__subn_set_opa_bct(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)2168 static int __subn_set_opa_bct(struct opa_smp *smp, u32 am, u8 *data,
2169 struct ib_device *ibdev, u32 port, u32 *resp_len,
2170 u32 max_len)
2171 {
2172 u32 num_ports = OPA_AM_NPORT(am);
2173 struct hfi1_devdata *dd = dd_from_ibdev(ibdev);
2174 struct hfi1_pportdata *ppd;
2175 struct buffer_control *p = (struct buffer_control *)data;
2176
2177 if (num_ports != 1 || smp_length_check(sizeof(*p), max_len)) {
2178 smp->status |= IB_SMP_INVALID_FIELD;
2179 return reply((struct ib_mad_hdr *)smp);
2180 }
2181 ppd = dd->pport + (port - 1);
2182 trace_bct_set(dd, p);
2183 if (fm_set_table(ppd, FM_TBL_BUFFER_CONTROL, p) < 0) {
2184 smp->status |= IB_SMP_INVALID_FIELD;
2185 return reply((struct ib_mad_hdr *)smp);
2186 }
2187
2188 return __subn_get_opa_bct(smp, am, data, ibdev, port, resp_len,
2189 max_len);
2190 }
2191
__subn_get_opa_vl_arb(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)2192 static int __subn_get_opa_vl_arb(struct opa_smp *smp, u32 am, u8 *data,
2193 struct ib_device *ibdev, u32 port,
2194 u32 *resp_len, u32 max_len)
2195 {
2196 struct hfi1_pportdata *ppd = ppd_from_ibp(to_iport(ibdev, port));
2197 u32 num_ports = OPA_AM_NPORT(am);
2198 u8 section = (am & 0x00ff0000) >> 16;
2199 u8 *p = data;
2200 int size = 256;
2201
2202 if (num_ports != 1 || smp_length_check(size, max_len)) {
2203 smp->status |= IB_SMP_INVALID_FIELD;
2204 return reply((struct ib_mad_hdr *)smp);
2205 }
2206
2207 switch (section) {
2208 case OPA_VLARB_LOW_ELEMENTS:
2209 fm_get_table(ppd, FM_TBL_VL_LOW_ARB, p);
2210 break;
2211 case OPA_VLARB_HIGH_ELEMENTS:
2212 fm_get_table(ppd, FM_TBL_VL_HIGH_ARB, p);
2213 break;
2214 case OPA_VLARB_PREEMPT_ELEMENTS:
2215 fm_get_table(ppd, FM_TBL_VL_PREEMPT_ELEMS, p);
2216 break;
2217 case OPA_VLARB_PREEMPT_MATRIX:
2218 fm_get_table(ppd, FM_TBL_VL_PREEMPT_MATRIX, p);
2219 break;
2220 default:
2221 pr_warn("OPA SubnGet(VL Arb) AM Invalid : 0x%x\n",
2222 be32_to_cpu(smp->attr_mod));
2223 smp->status |= IB_SMP_INVALID_FIELD;
2224 size = 0;
2225 break;
2226 }
2227
2228 if (size > 0 && resp_len)
2229 *resp_len += size;
2230
2231 return reply((struct ib_mad_hdr *)smp);
2232 }
2233
__subn_set_opa_vl_arb(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)2234 static int __subn_set_opa_vl_arb(struct opa_smp *smp, u32 am, u8 *data,
2235 struct ib_device *ibdev, u32 port,
2236 u32 *resp_len, u32 max_len)
2237 {
2238 struct hfi1_pportdata *ppd = ppd_from_ibp(to_iport(ibdev, port));
2239 u32 num_ports = OPA_AM_NPORT(am);
2240 u8 section = (am & 0x00ff0000) >> 16;
2241 u8 *p = data;
2242 int size = 256;
2243
2244 if (num_ports != 1 || smp_length_check(size, max_len)) {
2245 smp->status |= IB_SMP_INVALID_FIELD;
2246 return reply((struct ib_mad_hdr *)smp);
2247 }
2248
2249 switch (section) {
2250 case OPA_VLARB_LOW_ELEMENTS:
2251 (void)fm_set_table(ppd, FM_TBL_VL_LOW_ARB, p);
2252 break;
2253 case OPA_VLARB_HIGH_ELEMENTS:
2254 (void)fm_set_table(ppd, FM_TBL_VL_HIGH_ARB, p);
2255 break;
2256 /*
2257 * neither OPA_VLARB_PREEMPT_ELEMENTS, or OPA_VLARB_PREEMPT_MATRIX
2258 * can be changed from the default values
2259 */
2260 case OPA_VLARB_PREEMPT_ELEMENTS:
2261 case OPA_VLARB_PREEMPT_MATRIX:
2262 smp->status |= IB_SMP_UNSUP_METH_ATTR;
2263 break;
2264 default:
2265 pr_warn("OPA SubnSet(VL Arb) AM Invalid : 0x%x\n",
2266 be32_to_cpu(smp->attr_mod));
2267 smp->status |= IB_SMP_INVALID_FIELD;
2268 break;
2269 }
2270
2271 return __subn_get_opa_vl_arb(smp, am, data, ibdev, port, resp_len,
2272 max_len);
2273 }
2274
2275 struct opa_pma_mad {
2276 struct ib_mad_hdr mad_hdr;
2277 u8 data[2024];
2278 } __packed;
2279
2280 struct opa_port_status_req {
2281 __u8 port_num;
2282 __u8 reserved[3];
2283 __be32 vl_select_mask;
2284 };
2285
2286 #define VL_MASK_ALL 0x00000000000080ffUL
2287
2288 struct opa_port_status_rsp {
2289 __u8 port_num;
2290 __u8 reserved[3];
2291 __be32 vl_select_mask;
2292
2293 /* Data counters */
2294 __be64 port_xmit_data;
2295 __be64 port_rcv_data;
2296 __be64 port_xmit_pkts;
2297 __be64 port_rcv_pkts;
2298 __be64 port_multicast_xmit_pkts;
2299 __be64 port_multicast_rcv_pkts;
2300 __be64 port_xmit_wait;
2301 __be64 sw_port_congestion;
2302 __be64 port_rcv_fecn;
2303 __be64 port_rcv_becn;
2304 __be64 port_xmit_time_cong;
2305 __be64 port_xmit_wasted_bw;
2306 __be64 port_xmit_wait_data;
2307 __be64 port_rcv_bubble;
2308 __be64 port_mark_fecn;
2309 /* Error counters */
2310 __be64 port_rcv_constraint_errors;
2311 __be64 port_rcv_switch_relay_errors;
2312 __be64 port_xmit_discards;
2313 __be64 port_xmit_constraint_errors;
2314 __be64 port_rcv_remote_physical_errors;
2315 __be64 local_link_integrity_errors;
2316 __be64 port_rcv_errors;
2317 __be64 excessive_buffer_overruns;
2318 __be64 fm_config_errors;
2319 __be32 link_error_recovery;
2320 __be32 link_downed;
2321 u8 uncorrectable_errors;
2322
2323 u8 link_quality_indicator; /* 5res, 3bit */
2324 u8 res2[6];
2325 struct _vls_pctrs {
2326 /* per-VL Data counters */
2327 __be64 port_vl_xmit_data;
2328 __be64 port_vl_rcv_data;
2329 __be64 port_vl_xmit_pkts;
2330 __be64 port_vl_rcv_pkts;
2331 __be64 port_vl_xmit_wait;
2332 __be64 sw_port_vl_congestion;
2333 __be64 port_vl_rcv_fecn;
2334 __be64 port_vl_rcv_becn;
2335 __be64 port_xmit_time_cong;
2336 __be64 port_vl_xmit_wasted_bw;
2337 __be64 port_vl_xmit_wait_data;
2338 __be64 port_vl_rcv_bubble;
2339 __be64 port_vl_mark_fecn;
2340 __be64 port_vl_xmit_discards;
2341 } vls[]; /* real array size defined by # bits set in vl_select_mask */
2342 };
2343
2344 enum counter_selects {
2345 CS_PORT_XMIT_DATA = (1 << 31),
2346 CS_PORT_RCV_DATA = (1 << 30),
2347 CS_PORT_XMIT_PKTS = (1 << 29),
2348 CS_PORT_RCV_PKTS = (1 << 28),
2349 CS_PORT_MCAST_XMIT_PKTS = (1 << 27),
2350 CS_PORT_MCAST_RCV_PKTS = (1 << 26),
2351 CS_PORT_XMIT_WAIT = (1 << 25),
2352 CS_SW_PORT_CONGESTION = (1 << 24),
2353 CS_PORT_RCV_FECN = (1 << 23),
2354 CS_PORT_RCV_BECN = (1 << 22),
2355 CS_PORT_XMIT_TIME_CONG = (1 << 21),
2356 CS_PORT_XMIT_WASTED_BW = (1 << 20),
2357 CS_PORT_XMIT_WAIT_DATA = (1 << 19),
2358 CS_PORT_RCV_BUBBLE = (1 << 18),
2359 CS_PORT_MARK_FECN = (1 << 17),
2360 CS_PORT_RCV_CONSTRAINT_ERRORS = (1 << 16),
2361 CS_PORT_RCV_SWITCH_RELAY_ERRORS = (1 << 15),
2362 CS_PORT_XMIT_DISCARDS = (1 << 14),
2363 CS_PORT_XMIT_CONSTRAINT_ERRORS = (1 << 13),
2364 CS_PORT_RCV_REMOTE_PHYSICAL_ERRORS = (1 << 12),
2365 CS_LOCAL_LINK_INTEGRITY_ERRORS = (1 << 11),
2366 CS_PORT_RCV_ERRORS = (1 << 10),
2367 CS_EXCESSIVE_BUFFER_OVERRUNS = (1 << 9),
2368 CS_FM_CONFIG_ERRORS = (1 << 8),
2369 CS_LINK_ERROR_RECOVERY = (1 << 7),
2370 CS_LINK_DOWNED = (1 << 6),
2371 CS_UNCORRECTABLE_ERRORS = (1 << 5),
2372 };
2373
2374 struct opa_clear_port_status {
2375 __be64 port_select_mask[4];
2376 __be32 counter_select_mask;
2377 };
2378
2379 struct opa_aggregate {
2380 __be16 attr_id;
2381 __be16 err_reqlength; /* 1 bit, 8 res, 7 bit */
2382 __be32 attr_mod;
2383 u8 data[];
2384 };
2385
2386 #define MSK_LLI 0x000000f0
2387 #define MSK_LLI_SFT 4
2388 #define MSK_LER 0x0000000f
2389 #define MSK_LER_SFT 0
2390 #define ADD_LLI 8
2391 #define ADD_LER 2
2392
2393 /* Request contains first three fields, response contains those plus the rest */
2394 struct opa_port_data_counters_msg {
2395 __be64 port_select_mask[4];
2396 __be32 vl_select_mask;
2397 __be32 resolution;
2398
2399 /* Response fields follow */
2400 struct _port_dctrs {
2401 u8 port_number;
2402 u8 reserved2[3];
2403 __be32 link_quality_indicator; /* 29res, 3bit */
2404
2405 /* Data counters */
2406 __be64 port_xmit_data;
2407 __be64 port_rcv_data;
2408 __be64 port_xmit_pkts;
2409 __be64 port_rcv_pkts;
2410 __be64 port_multicast_xmit_pkts;
2411 __be64 port_multicast_rcv_pkts;
2412 __be64 port_xmit_wait;
2413 __be64 sw_port_congestion;
2414 __be64 port_rcv_fecn;
2415 __be64 port_rcv_becn;
2416 __be64 port_xmit_time_cong;
2417 __be64 port_xmit_wasted_bw;
2418 __be64 port_xmit_wait_data;
2419 __be64 port_rcv_bubble;
2420 __be64 port_mark_fecn;
2421
2422 __be64 port_error_counter_summary;
2423 /* Sum of error counts/port */
2424
2425 struct _vls_dctrs {
2426 /* per-VL Data counters */
2427 __be64 port_vl_xmit_data;
2428 __be64 port_vl_rcv_data;
2429 __be64 port_vl_xmit_pkts;
2430 __be64 port_vl_rcv_pkts;
2431 __be64 port_vl_xmit_wait;
2432 __be64 sw_port_vl_congestion;
2433 __be64 port_vl_rcv_fecn;
2434 __be64 port_vl_rcv_becn;
2435 __be64 port_xmit_time_cong;
2436 __be64 port_vl_xmit_wasted_bw;
2437 __be64 port_vl_xmit_wait_data;
2438 __be64 port_vl_rcv_bubble;
2439 __be64 port_vl_mark_fecn;
2440 } vls[];
2441 /* array size defined by #bits set in vl_select_mask*/
2442 } port;
2443 };
2444
2445 struct opa_port_error_counters64_msg {
2446 /*
2447 * Request contains first two fields, response contains the
2448 * whole magilla
2449 */
2450 __be64 port_select_mask[4];
2451 __be32 vl_select_mask;
2452
2453 /* Response-only fields follow */
2454 __be32 reserved1;
2455 struct _port_ectrs {
2456 u8 port_number;
2457 u8 reserved2[7];
2458 __be64 port_rcv_constraint_errors;
2459 __be64 port_rcv_switch_relay_errors;
2460 __be64 port_xmit_discards;
2461 __be64 port_xmit_constraint_errors;
2462 __be64 port_rcv_remote_physical_errors;
2463 __be64 local_link_integrity_errors;
2464 __be64 port_rcv_errors;
2465 __be64 excessive_buffer_overruns;
2466 __be64 fm_config_errors;
2467 __be32 link_error_recovery;
2468 __be32 link_downed;
2469 u8 uncorrectable_errors;
2470 u8 reserved3[7];
2471 struct _vls_ectrs {
2472 __be64 port_vl_xmit_discards;
2473 } vls[];
2474 /* array size defined by #bits set in vl_select_mask */
2475 } port;
2476 };
2477
2478 struct opa_port_error_info_msg {
2479 __be64 port_select_mask[4];
2480 __be32 error_info_select_mask;
2481 __be32 reserved1;
2482 struct _port_ei {
2483 u8 port_number;
2484 u8 reserved2[7];
2485
2486 /* PortRcvErrorInfo */
2487 struct {
2488 u8 status_and_code;
2489 union {
2490 u8 raw[17];
2491 struct {
2492 /* EI1to12 format */
2493 u8 packet_flit1[8];
2494 u8 packet_flit2[8];
2495 u8 remaining_flit_bits12;
2496 } ei1to12;
2497 struct {
2498 u8 packet_bytes[8];
2499 u8 remaining_flit_bits;
2500 } ei13;
2501 } ei;
2502 u8 reserved3[6];
2503 } __packed port_rcv_ei;
2504
2505 /* ExcessiveBufferOverrunInfo */
2506 struct {
2507 u8 status_and_sc;
2508 u8 reserved4[7];
2509 } __packed excessive_buffer_overrun_ei;
2510
2511 /* PortXmitConstraintErrorInfo */
2512 struct {
2513 u8 status;
2514 u8 reserved5;
2515 __be16 pkey;
2516 __be32 slid;
2517 } __packed port_xmit_constraint_ei;
2518
2519 /* PortRcvConstraintErrorInfo */
2520 struct {
2521 u8 status;
2522 u8 reserved6;
2523 __be16 pkey;
2524 __be32 slid;
2525 } __packed port_rcv_constraint_ei;
2526
2527 /* PortRcvSwitchRelayErrorInfo */
2528 struct {
2529 u8 status_and_code;
2530 u8 reserved7[3];
2531 __u32 error_info;
2532 } __packed port_rcv_switch_relay_ei;
2533
2534 /* UncorrectableErrorInfo */
2535 struct {
2536 u8 status_and_code;
2537 u8 reserved8;
2538 } __packed uncorrectable_ei;
2539
2540 /* FMConfigErrorInfo */
2541 struct {
2542 u8 status_and_code;
2543 u8 error_info;
2544 } __packed fm_config_ei;
2545 __u32 reserved9;
2546 } port;
2547 };
2548
2549 /* opa_port_error_info_msg error_info_select_mask bit definitions */
2550 enum error_info_selects {
2551 ES_PORT_RCV_ERROR_INFO = (1 << 31),
2552 ES_EXCESSIVE_BUFFER_OVERRUN_INFO = (1 << 30),
2553 ES_PORT_XMIT_CONSTRAINT_ERROR_INFO = (1 << 29),
2554 ES_PORT_RCV_CONSTRAINT_ERROR_INFO = (1 << 28),
2555 ES_PORT_RCV_SWITCH_RELAY_ERROR_INFO = (1 << 27),
2556 ES_UNCORRECTABLE_ERROR_INFO = (1 << 26),
2557 ES_FM_CONFIG_ERROR_INFO = (1 << 25)
2558 };
2559
pma_get_opa_classportinfo(struct opa_pma_mad * pmp,struct ib_device * ibdev,u32 * resp_len)2560 static int pma_get_opa_classportinfo(struct opa_pma_mad *pmp,
2561 struct ib_device *ibdev, u32 *resp_len)
2562 {
2563 struct opa_class_port_info *p =
2564 (struct opa_class_port_info *)pmp->data;
2565
2566 memset(pmp->data, 0, sizeof(pmp->data));
2567
2568 if (pmp->mad_hdr.attr_mod != 0)
2569 pmp->mad_hdr.status |= IB_SMP_INVALID_FIELD;
2570
2571 p->base_version = OPA_MGMT_BASE_VERSION;
2572 p->class_version = OPA_SM_CLASS_VERSION;
2573 /*
2574 * Expected response time is 4.096 usec. * 2^18 == 1.073741824 sec.
2575 */
2576 p->cap_mask2_resp_time = cpu_to_be32(18);
2577
2578 if (resp_len)
2579 *resp_len += sizeof(*p);
2580
2581 return reply((struct ib_mad_hdr *)pmp);
2582 }
2583
a0_portstatus(struct hfi1_pportdata * ppd,struct opa_port_status_rsp * rsp)2584 static void a0_portstatus(struct hfi1_pportdata *ppd,
2585 struct opa_port_status_rsp *rsp)
2586 {
2587 if (!is_bx(ppd->dd)) {
2588 unsigned long vl;
2589 u64 sum_vl_xmit_wait = 0;
2590 unsigned long vl_all_mask = VL_MASK_ALL;
2591
2592 for_each_set_bit(vl, &vl_all_mask, BITS_PER_LONG) {
2593 u64 tmp = sum_vl_xmit_wait +
2594 read_port_cntr(ppd, C_TX_WAIT_VL,
2595 idx_from_vl(vl));
2596 if (tmp < sum_vl_xmit_wait) {
2597 /* we wrapped */
2598 sum_vl_xmit_wait = (u64)~0;
2599 break;
2600 }
2601 sum_vl_xmit_wait = tmp;
2602 }
2603 if (be64_to_cpu(rsp->port_xmit_wait) > sum_vl_xmit_wait)
2604 rsp->port_xmit_wait = cpu_to_be64(sum_vl_xmit_wait);
2605 }
2606 }
2607
2608 /**
2609 * tx_link_width - convert link width bitmask to integer
2610 * value representing actual link width.
2611 * @link_width: width of active link
2612 * @return: return index of the bit set in link_width var
2613 *
2614 * The function convert and return the index of bit set
2615 * that indicate the current link width.
2616 */
tx_link_width(u16 link_width)2617 u16 tx_link_width(u16 link_width)
2618 {
2619 int n = LINK_WIDTH_DEFAULT;
2620 u16 tx_width = n;
2621
2622 while (link_width && n) {
2623 if (link_width & (1 << (n - 1))) {
2624 tx_width = n;
2625 break;
2626 }
2627 n--;
2628 }
2629
2630 return tx_width;
2631 }
2632
2633 /**
2634 * get_xmit_wait_counters - Convert HFI 's SendWaitCnt/SendWaitVlCnt
2635 * counter in unit of TXE cycle times to flit times.
2636 * @ppd: info of physical Hfi port
2637 * @link_width: width of active link
2638 * @link_speed: speed of active link
2639 * @vl: represent VL0-VL7, VL15 for PortVLXmitWait counters request
2640 * and if vl value is C_VL_COUNT, it represent SendWaitCnt
2641 * counter request
2642 * @return: return SendWaitCnt/SendWaitVlCnt counter value per vl.
2643 *
2644 * Convert SendWaitCnt/SendWaitVlCnt counter from TXE cycle times to
2645 * flit times. Call this function to samples these counters. This
2646 * function will calculate for previous state transition and update
2647 * current state at end of function using ppd->prev_link_width and
2648 * ppd->port_vl_xmit_wait_last to port_vl_xmit_wait_curr and link_width.
2649 */
get_xmit_wait_counters(struct hfi1_pportdata * ppd,u16 link_width,u16 link_speed,int vl)2650 u64 get_xmit_wait_counters(struct hfi1_pportdata *ppd,
2651 u16 link_width, u16 link_speed, int vl)
2652 {
2653 u64 port_vl_xmit_wait_curr;
2654 u64 delta_vl_xmit_wait;
2655 u64 xmit_wait_val;
2656
2657 if (vl > C_VL_COUNT)
2658 return 0;
2659 if (vl < C_VL_COUNT)
2660 port_vl_xmit_wait_curr =
2661 read_port_cntr(ppd, C_TX_WAIT_VL, vl);
2662 else
2663 port_vl_xmit_wait_curr =
2664 read_port_cntr(ppd, C_TX_WAIT, CNTR_INVALID_VL);
2665
2666 xmit_wait_val =
2667 port_vl_xmit_wait_curr -
2668 ppd->port_vl_xmit_wait_last[vl];
2669 delta_vl_xmit_wait =
2670 convert_xmit_counter(xmit_wait_val,
2671 ppd->prev_link_width,
2672 link_speed);
2673
2674 ppd->vl_xmit_flit_cnt[vl] += delta_vl_xmit_wait;
2675 ppd->port_vl_xmit_wait_last[vl] = port_vl_xmit_wait_curr;
2676 ppd->prev_link_width = link_width;
2677
2678 return ppd->vl_xmit_flit_cnt[vl];
2679 }
2680
pma_get_opa_portstatus(struct opa_pma_mad * pmp,struct ib_device * ibdev,u32 port,u32 * resp_len)2681 static int pma_get_opa_portstatus(struct opa_pma_mad *pmp,
2682 struct ib_device *ibdev,
2683 u32 port, u32 *resp_len)
2684 {
2685 struct opa_port_status_req *req =
2686 (struct opa_port_status_req *)pmp->data;
2687 struct hfi1_devdata *dd = dd_from_ibdev(ibdev);
2688 struct opa_port_status_rsp *rsp;
2689 unsigned long vl_select_mask = be32_to_cpu(req->vl_select_mask);
2690 unsigned long vl;
2691 size_t response_data_size;
2692 u32 nports = be32_to_cpu(pmp->mad_hdr.attr_mod) >> 24;
2693 u32 port_num = req->port_num;
2694 u8 num_vls = hweight64(vl_select_mask);
2695 struct _vls_pctrs *vlinfo;
2696 struct hfi1_ibport *ibp = to_iport(ibdev, port);
2697 struct hfi1_pportdata *ppd = ppd_from_ibp(ibp);
2698 int vfi;
2699 u64 tmp, tmp2;
2700 u16 link_width;
2701 u16 link_speed;
2702
2703 response_data_size = struct_size(rsp, vls, num_vls);
2704 if (response_data_size > sizeof(pmp->data)) {
2705 pmp->mad_hdr.status |= OPA_PM_STATUS_REQUEST_TOO_LARGE;
2706 return reply((struct ib_mad_hdr *)pmp);
2707 }
2708
2709 if (nports != 1 || (port_num && port_num != port) ||
2710 num_vls > OPA_MAX_VLS || (vl_select_mask & ~VL_MASK_ALL)) {
2711 pmp->mad_hdr.status |= IB_SMP_INVALID_FIELD;
2712 return reply((struct ib_mad_hdr *)pmp);
2713 }
2714
2715 memset(pmp->data, 0, sizeof(pmp->data));
2716
2717 rsp = (struct opa_port_status_rsp *)pmp->data;
2718 if (port_num)
2719 rsp->port_num = port_num;
2720 else
2721 rsp->port_num = port;
2722
2723 rsp->port_rcv_constraint_errors =
2724 cpu_to_be64(read_port_cntr(ppd, C_SW_RCV_CSTR_ERR,
2725 CNTR_INVALID_VL));
2726
2727 hfi1_read_link_quality(dd, &rsp->link_quality_indicator);
2728
2729 rsp->vl_select_mask = cpu_to_be32((u32)vl_select_mask);
2730 rsp->port_xmit_data = cpu_to_be64(read_dev_cntr(dd, C_DC_XMIT_FLITS,
2731 CNTR_INVALID_VL));
2732 rsp->port_rcv_data = cpu_to_be64(read_dev_cntr(dd, C_DC_RCV_FLITS,
2733 CNTR_INVALID_VL));
2734 rsp->port_xmit_pkts = cpu_to_be64(read_dev_cntr(dd, C_DC_XMIT_PKTS,
2735 CNTR_INVALID_VL));
2736 rsp->port_rcv_pkts = cpu_to_be64(read_dev_cntr(dd, C_DC_RCV_PKTS,
2737 CNTR_INVALID_VL));
2738 rsp->port_multicast_xmit_pkts =
2739 cpu_to_be64(read_dev_cntr(dd, C_DC_MC_XMIT_PKTS,
2740 CNTR_INVALID_VL));
2741 rsp->port_multicast_rcv_pkts =
2742 cpu_to_be64(read_dev_cntr(dd, C_DC_MC_RCV_PKTS,
2743 CNTR_INVALID_VL));
2744 /*
2745 * Convert PortXmitWait counter from TXE cycle times
2746 * to flit times.
2747 */
2748 link_width =
2749 tx_link_width(ppd->link_width_downgrade_tx_active);
2750 link_speed = get_link_speed(ppd->link_speed_active);
2751 rsp->port_xmit_wait =
2752 cpu_to_be64(get_xmit_wait_counters(ppd, link_width,
2753 link_speed, C_VL_COUNT));
2754 rsp->port_rcv_fecn =
2755 cpu_to_be64(read_dev_cntr(dd, C_DC_RCV_FCN, CNTR_INVALID_VL));
2756 rsp->port_rcv_becn =
2757 cpu_to_be64(read_dev_cntr(dd, C_DC_RCV_BCN, CNTR_INVALID_VL));
2758 rsp->port_xmit_discards =
2759 cpu_to_be64(read_port_cntr(ppd, C_SW_XMIT_DSCD,
2760 CNTR_INVALID_VL));
2761 rsp->port_xmit_constraint_errors =
2762 cpu_to_be64(read_port_cntr(ppd, C_SW_XMIT_CSTR_ERR,
2763 CNTR_INVALID_VL));
2764 rsp->port_rcv_remote_physical_errors =
2765 cpu_to_be64(read_dev_cntr(dd, C_DC_RMT_PHY_ERR,
2766 CNTR_INVALID_VL));
2767 rsp->local_link_integrity_errors =
2768 cpu_to_be64(read_dev_cntr(dd, C_DC_RX_REPLAY,
2769 CNTR_INVALID_VL));
2770 tmp = read_dev_cntr(dd, C_DC_SEQ_CRC_CNT, CNTR_INVALID_VL);
2771 tmp2 = tmp + read_dev_cntr(dd, C_DC_REINIT_FROM_PEER_CNT,
2772 CNTR_INVALID_VL);
2773 if (tmp2 > (u32)UINT_MAX || tmp2 < tmp) {
2774 /* overflow/wrapped */
2775 rsp->link_error_recovery = cpu_to_be32(~0);
2776 } else {
2777 rsp->link_error_recovery = cpu_to_be32(tmp2);
2778 }
2779 rsp->port_rcv_errors =
2780 cpu_to_be64(read_dev_cntr(dd, C_DC_RCV_ERR, CNTR_INVALID_VL));
2781 rsp->excessive_buffer_overruns =
2782 cpu_to_be64(read_dev_cntr(dd, C_RCV_OVF, CNTR_INVALID_VL));
2783 rsp->fm_config_errors =
2784 cpu_to_be64(read_dev_cntr(dd, C_DC_FM_CFG_ERR,
2785 CNTR_INVALID_VL));
2786 rsp->link_downed = cpu_to_be32(read_port_cntr(ppd, C_SW_LINK_DOWN,
2787 CNTR_INVALID_VL));
2788
2789 /* rsp->uncorrectable_errors is 8 bits wide, and it pegs at 0xff */
2790 tmp = read_dev_cntr(dd, C_DC_UNC_ERR, CNTR_INVALID_VL);
2791 rsp->uncorrectable_errors = tmp < 0x100 ? (tmp & 0xff) : 0xff;
2792
2793 vlinfo = &rsp->vls[0];
2794 vfi = 0;
2795 /* The vl_select_mask has been checked above, and we know
2796 * that it contains only entries which represent valid VLs.
2797 * So in the for_each_set_bit() loop below, we don't need
2798 * any additional checks for vl.
2799 */
2800 for_each_set_bit(vl, &vl_select_mask, BITS_PER_LONG) {
2801 memset(vlinfo, 0, sizeof(*vlinfo));
2802
2803 tmp = read_dev_cntr(dd, C_DC_RX_FLIT_VL, idx_from_vl(vl));
2804 rsp->vls[vfi].port_vl_rcv_data = cpu_to_be64(tmp);
2805
2806 rsp->vls[vfi].port_vl_rcv_pkts =
2807 cpu_to_be64(read_dev_cntr(dd, C_DC_RX_PKT_VL,
2808 idx_from_vl(vl)));
2809
2810 rsp->vls[vfi].port_vl_xmit_data =
2811 cpu_to_be64(read_port_cntr(ppd, C_TX_FLIT_VL,
2812 idx_from_vl(vl)));
2813
2814 rsp->vls[vfi].port_vl_xmit_pkts =
2815 cpu_to_be64(read_port_cntr(ppd, C_TX_PKT_VL,
2816 idx_from_vl(vl)));
2817 /*
2818 * Convert PortVlXmitWait counter from TXE cycle
2819 * times to flit times.
2820 */
2821 rsp->vls[vfi].port_vl_xmit_wait =
2822 cpu_to_be64(get_xmit_wait_counters(ppd, link_width,
2823 link_speed,
2824 idx_from_vl(vl)));
2825
2826 rsp->vls[vfi].port_vl_rcv_fecn =
2827 cpu_to_be64(read_dev_cntr(dd, C_DC_RCV_FCN_VL,
2828 idx_from_vl(vl)));
2829
2830 rsp->vls[vfi].port_vl_rcv_becn =
2831 cpu_to_be64(read_dev_cntr(dd, C_DC_RCV_BCN_VL,
2832 idx_from_vl(vl)));
2833
2834 rsp->vls[vfi].port_vl_xmit_discards =
2835 cpu_to_be64(read_port_cntr(ppd, C_SW_XMIT_DSCD_VL,
2836 idx_from_vl(vl)));
2837 vlinfo++;
2838 vfi++;
2839 }
2840
2841 a0_portstatus(ppd, rsp);
2842
2843 if (resp_len)
2844 *resp_len += response_data_size;
2845
2846 return reply((struct ib_mad_hdr *)pmp);
2847 }
2848
get_error_counter_summary(struct ib_device * ibdev,u32 port,u8 res_lli,u8 res_ler)2849 static u64 get_error_counter_summary(struct ib_device *ibdev, u32 port,
2850 u8 res_lli, u8 res_ler)
2851 {
2852 struct hfi1_devdata *dd = dd_from_ibdev(ibdev);
2853 struct hfi1_ibport *ibp = to_iport(ibdev, port);
2854 struct hfi1_pportdata *ppd = ppd_from_ibp(ibp);
2855 u64 error_counter_summary = 0, tmp;
2856
2857 error_counter_summary += read_port_cntr(ppd, C_SW_RCV_CSTR_ERR,
2858 CNTR_INVALID_VL);
2859 /* port_rcv_switch_relay_errors is 0 for HFIs */
2860 error_counter_summary += read_port_cntr(ppd, C_SW_XMIT_DSCD,
2861 CNTR_INVALID_VL);
2862 error_counter_summary += read_port_cntr(ppd, C_SW_XMIT_CSTR_ERR,
2863 CNTR_INVALID_VL);
2864 error_counter_summary += read_dev_cntr(dd, C_DC_RMT_PHY_ERR,
2865 CNTR_INVALID_VL);
2866 /* local link integrity must be right-shifted by the lli resolution */
2867 error_counter_summary += (read_dev_cntr(dd, C_DC_RX_REPLAY,
2868 CNTR_INVALID_VL) >> res_lli);
2869 /* link error recovery must b right-shifted by the ler resolution */
2870 tmp = read_dev_cntr(dd, C_DC_SEQ_CRC_CNT, CNTR_INVALID_VL);
2871 tmp += read_dev_cntr(dd, C_DC_REINIT_FROM_PEER_CNT, CNTR_INVALID_VL);
2872 error_counter_summary += (tmp >> res_ler);
2873 error_counter_summary += read_dev_cntr(dd, C_DC_RCV_ERR,
2874 CNTR_INVALID_VL);
2875 error_counter_summary += read_dev_cntr(dd, C_RCV_OVF, CNTR_INVALID_VL);
2876 error_counter_summary += read_dev_cntr(dd, C_DC_FM_CFG_ERR,
2877 CNTR_INVALID_VL);
2878 /* ppd->link_downed is a 32-bit value */
2879 error_counter_summary += read_port_cntr(ppd, C_SW_LINK_DOWN,
2880 CNTR_INVALID_VL);
2881 tmp = read_dev_cntr(dd, C_DC_UNC_ERR, CNTR_INVALID_VL);
2882 /* this is an 8-bit quantity */
2883 error_counter_summary += tmp < 0x100 ? (tmp & 0xff) : 0xff;
2884
2885 return error_counter_summary;
2886 }
2887
a0_datacounters(struct hfi1_pportdata * ppd,struct _port_dctrs * rsp)2888 static void a0_datacounters(struct hfi1_pportdata *ppd, struct _port_dctrs *rsp)
2889 {
2890 if (!is_bx(ppd->dd)) {
2891 unsigned long vl;
2892 u64 sum_vl_xmit_wait = 0;
2893 unsigned long vl_all_mask = VL_MASK_ALL;
2894
2895 for_each_set_bit(vl, &vl_all_mask, BITS_PER_LONG) {
2896 u64 tmp = sum_vl_xmit_wait +
2897 read_port_cntr(ppd, C_TX_WAIT_VL,
2898 idx_from_vl(vl));
2899 if (tmp < sum_vl_xmit_wait) {
2900 /* we wrapped */
2901 sum_vl_xmit_wait = (u64)~0;
2902 break;
2903 }
2904 sum_vl_xmit_wait = tmp;
2905 }
2906 if (be64_to_cpu(rsp->port_xmit_wait) > sum_vl_xmit_wait)
2907 rsp->port_xmit_wait = cpu_to_be64(sum_vl_xmit_wait);
2908 }
2909 }
2910
pma_get_opa_port_dctrs(struct ib_device * ibdev,struct _port_dctrs * rsp)2911 static void pma_get_opa_port_dctrs(struct ib_device *ibdev,
2912 struct _port_dctrs *rsp)
2913 {
2914 struct hfi1_devdata *dd = dd_from_ibdev(ibdev);
2915
2916 rsp->port_xmit_data = cpu_to_be64(read_dev_cntr(dd, C_DC_XMIT_FLITS,
2917 CNTR_INVALID_VL));
2918 rsp->port_rcv_data = cpu_to_be64(read_dev_cntr(dd, C_DC_RCV_FLITS,
2919 CNTR_INVALID_VL));
2920 rsp->port_xmit_pkts = cpu_to_be64(read_dev_cntr(dd, C_DC_XMIT_PKTS,
2921 CNTR_INVALID_VL));
2922 rsp->port_rcv_pkts = cpu_to_be64(read_dev_cntr(dd, C_DC_RCV_PKTS,
2923 CNTR_INVALID_VL));
2924 rsp->port_multicast_xmit_pkts =
2925 cpu_to_be64(read_dev_cntr(dd, C_DC_MC_XMIT_PKTS,
2926 CNTR_INVALID_VL));
2927 rsp->port_multicast_rcv_pkts =
2928 cpu_to_be64(read_dev_cntr(dd, C_DC_MC_RCV_PKTS,
2929 CNTR_INVALID_VL));
2930 }
2931
pma_get_opa_datacounters(struct opa_pma_mad * pmp,struct ib_device * ibdev,u32 port,u32 * resp_len)2932 static int pma_get_opa_datacounters(struct opa_pma_mad *pmp,
2933 struct ib_device *ibdev,
2934 u32 port, u32 *resp_len)
2935 {
2936 struct opa_port_data_counters_msg *req =
2937 (struct opa_port_data_counters_msg *)pmp->data;
2938 struct hfi1_devdata *dd = dd_from_ibdev(ibdev);
2939 struct hfi1_ibport *ibp = to_iport(ibdev, port);
2940 struct hfi1_pportdata *ppd = ppd_from_ibp(ibp);
2941 struct _port_dctrs *rsp;
2942 struct _vls_dctrs *vlinfo;
2943 size_t response_data_size;
2944 u32 num_ports;
2945 u8 lq, num_vls;
2946 u8 res_lli, res_ler;
2947 u64 port_mask;
2948 u32 port_num;
2949 unsigned long vl;
2950 unsigned long vl_select_mask;
2951 int vfi;
2952 u16 link_width;
2953 u16 link_speed;
2954
2955 num_ports = be32_to_cpu(pmp->mad_hdr.attr_mod) >> 24;
2956 num_vls = hweight32(be32_to_cpu(req->vl_select_mask));
2957 vl_select_mask = be32_to_cpu(req->vl_select_mask);
2958 res_lli = (u8)(be32_to_cpu(req->resolution) & MSK_LLI) >> MSK_LLI_SFT;
2959 res_lli = res_lli ? res_lli + ADD_LLI : 0;
2960 res_ler = (u8)(be32_to_cpu(req->resolution) & MSK_LER) >> MSK_LER_SFT;
2961 res_ler = res_ler ? res_ler + ADD_LER : 0;
2962
2963 if (num_ports != 1 || (vl_select_mask & ~VL_MASK_ALL)) {
2964 pmp->mad_hdr.status |= IB_SMP_INVALID_FIELD;
2965 return reply((struct ib_mad_hdr *)pmp);
2966 }
2967
2968 /* Sanity check */
2969 response_data_size = struct_size(req, port.vls, num_vls);
2970
2971 if (response_data_size > sizeof(pmp->data)) {
2972 pmp->mad_hdr.status |= IB_SMP_INVALID_FIELD;
2973 return reply((struct ib_mad_hdr *)pmp);
2974 }
2975
2976 /*
2977 * The bit set in the mask needs to be consistent with the
2978 * port the request came in on.
2979 */
2980 port_mask = be64_to_cpu(req->port_select_mask[3]);
2981 port_num = find_first_bit((unsigned long *)&port_mask,
2982 sizeof(port_mask) * 8);
2983
2984 if (port_num != port) {
2985 pmp->mad_hdr.status |= IB_SMP_INVALID_FIELD;
2986 return reply((struct ib_mad_hdr *)pmp);
2987 }
2988
2989 rsp = &req->port;
2990 memset(rsp, 0, sizeof(*rsp));
2991
2992 rsp->port_number = port;
2993 /*
2994 * Note that link_quality_indicator is a 32 bit quantity in
2995 * 'datacounters' queries (as opposed to 'portinfo' queries,
2996 * where it's a byte).
2997 */
2998 hfi1_read_link_quality(dd, &lq);
2999 rsp->link_quality_indicator = cpu_to_be32((u32)lq);
3000 pma_get_opa_port_dctrs(ibdev, rsp);
3001
3002 /*
3003 * Convert PortXmitWait counter from TXE
3004 * cycle times to flit times.
3005 */
3006 link_width =
3007 tx_link_width(ppd->link_width_downgrade_tx_active);
3008 link_speed = get_link_speed(ppd->link_speed_active);
3009 rsp->port_xmit_wait =
3010 cpu_to_be64(get_xmit_wait_counters(ppd, link_width,
3011 link_speed, C_VL_COUNT));
3012 rsp->port_rcv_fecn =
3013 cpu_to_be64(read_dev_cntr(dd, C_DC_RCV_FCN, CNTR_INVALID_VL));
3014 rsp->port_rcv_becn =
3015 cpu_to_be64(read_dev_cntr(dd, C_DC_RCV_BCN, CNTR_INVALID_VL));
3016 rsp->port_error_counter_summary =
3017 cpu_to_be64(get_error_counter_summary(ibdev, port,
3018 res_lli, res_ler));
3019
3020 vlinfo = &rsp->vls[0];
3021 vfi = 0;
3022 /* The vl_select_mask has been checked above, and we know
3023 * that it contains only entries which represent valid VLs.
3024 * So in the for_each_set_bit() loop below, we don't need
3025 * any additional checks for vl.
3026 */
3027 for_each_set_bit(vl, &vl_select_mask, BITS_PER_LONG) {
3028 memset(vlinfo, 0, sizeof(*vlinfo));
3029
3030 rsp->vls[vfi].port_vl_xmit_data =
3031 cpu_to_be64(read_port_cntr(ppd, C_TX_FLIT_VL,
3032 idx_from_vl(vl)));
3033
3034 rsp->vls[vfi].port_vl_rcv_data =
3035 cpu_to_be64(read_dev_cntr(dd, C_DC_RX_FLIT_VL,
3036 idx_from_vl(vl)));
3037
3038 rsp->vls[vfi].port_vl_xmit_pkts =
3039 cpu_to_be64(read_port_cntr(ppd, C_TX_PKT_VL,
3040 idx_from_vl(vl)));
3041
3042 rsp->vls[vfi].port_vl_rcv_pkts =
3043 cpu_to_be64(read_dev_cntr(dd, C_DC_RX_PKT_VL,
3044 idx_from_vl(vl)));
3045
3046 /*
3047 * Convert PortVlXmitWait counter from TXE
3048 * cycle times to flit times.
3049 */
3050 rsp->vls[vfi].port_vl_xmit_wait =
3051 cpu_to_be64(get_xmit_wait_counters(ppd, link_width,
3052 link_speed,
3053 idx_from_vl(vl)));
3054
3055 rsp->vls[vfi].port_vl_rcv_fecn =
3056 cpu_to_be64(read_dev_cntr(dd, C_DC_RCV_FCN_VL,
3057 idx_from_vl(vl)));
3058 rsp->vls[vfi].port_vl_rcv_becn =
3059 cpu_to_be64(read_dev_cntr(dd, C_DC_RCV_BCN_VL,
3060 idx_from_vl(vl)));
3061
3062 /* rsp->port_vl_xmit_time_cong is 0 for HFIs */
3063 /* rsp->port_vl_xmit_wasted_bw ??? */
3064 /* port_vl_xmit_wait_data - TXE (table 13-9 HFI spec) ???
3065 * does this differ from rsp->vls[vfi].port_vl_xmit_wait
3066 */
3067 /*rsp->vls[vfi].port_vl_mark_fecn =
3068 * cpu_to_be64(read_csr(dd, DCC_PRF_PORT_VL_MARK_FECN_CNT
3069 * + offset));
3070 */
3071 vlinfo++;
3072 vfi++;
3073 }
3074
3075 a0_datacounters(ppd, rsp);
3076
3077 if (resp_len)
3078 *resp_len += response_data_size;
3079
3080 return reply((struct ib_mad_hdr *)pmp);
3081 }
3082
pma_get_ib_portcounters_ext(struct ib_pma_mad * pmp,struct ib_device * ibdev,u32 port)3083 static int pma_get_ib_portcounters_ext(struct ib_pma_mad *pmp,
3084 struct ib_device *ibdev, u32 port)
3085 {
3086 struct ib_pma_portcounters_ext *p = (struct ib_pma_portcounters_ext *)
3087 pmp->data;
3088 struct _port_dctrs rsp;
3089
3090 if (pmp->mad_hdr.attr_mod != 0 || p->port_select != port) {
3091 pmp->mad_hdr.status |= IB_SMP_INVALID_FIELD;
3092 goto bail;
3093 }
3094
3095 memset(&rsp, 0, sizeof(rsp));
3096 pma_get_opa_port_dctrs(ibdev, &rsp);
3097
3098 p->port_xmit_data = rsp.port_xmit_data;
3099 p->port_rcv_data = rsp.port_rcv_data;
3100 p->port_xmit_packets = rsp.port_xmit_pkts;
3101 p->port_rcv_packets = rsp.port_rcv_pkts;
3102 p->port_unicast_xmit_packets = 0;
3103 p->port_unicast_rcv_packets = 0;
3104 p->port_multicast_xmit_packets = rsp.port_multicast_xmit_pkts;
3105 p->port_multicast_rcv_packets = rsp.port_multicast_rcv_pkts;
3106
3107 bail:
3108 return reply((struct ib_mad_hdr *)pmp);
3109 }
3110
pma_get_opa_port_ectrs(struct ib_device * ibdev,struct _port_ectrs * rsp,u32 port)3111 static void pma_get_opa_port_ectrs(struct ib_device *ibdev,
3112 struct _port_ectrs *rsp, u32 port)
3113 {
3114 u64 tmp, tmp2;
3115 struct hfi1_devdata *dd = dd_from_ibdev(ibdev);
3116 struct hfi1_ibport *ibp = to_iport(ibdev, port);
3117 struct hfi1_pportdata *ppd = ppd_from_ibp(ibp);
3118
3119 tmp = read_dev_cntr(dd, C_DC_SEQ_CRC_CNT, CNTR_INVALID_VL);
3120 tmp2 = tmp + read_dev_cntr(dd, C_DC_REINIT_FROM_PEER_CNT,
3121 CNTR_INVALID_VL);
3122 if (tmp2 > (u32)UINT_MAX || tmp2 < tmp) {
3123 /* overflow/wrapped */
3124 rsp->link_error_recovery = cpu_to_be32(~0);
3125 } else {
3126 rsp->link_error_recovery = cpu_to_be32(tmp2);
3127 }
3128
3129 rsp->link_downed = cpu_to_be32(read_port_cntr(ppd, C_SW_LINK_DOWN,
3130 CNTR_INVALID_VL));
3131 rsp->port_rcv_errors =
3132 cpu_to_be64(read_dev_cntr(dd, C_DC_RCV_ERR, CNTR_INVALID_VL));
3133 rsp->port_rcv_remote_physical_errors =
3134 cpu_to_be64(read_dev_cntr(dd, C_DC_RMT_PHY_ERR,
3135 CNTR_INVALID_VL));
3136 rsp->port_rcv_switch_relay_errors = 0;
3137 rsp->port_xmit_discards =
3138 cpu_to_be64(read_port_cntr(ppd, C_SW_XMIT_DSCD,
3139 CNTR_INVALID_VL));
3140 rsp->port_xmit_constraint_errors =
3141 cpu_to_be64(read_port_cntr(ppd, C_SW_XMIT_CSTR_ERR,
3142 CNTR_INVALID_VL));
3143 rsp->port_rcv_constraint_errors =
3144 cpu_to_be64(read_port_cntr(ppd, C_SW_RCV_CSTR_ERR,
3145 CNTR_INVALID_VL));
3146 rsp->local_link_integrity_errors =
3147 cpu_to_be64(read_dev_cntr(dd, C_DC_RX_REPLAY,
3148 CNTR_INVALID_VL));
3149 rsp->excessive_buffer_overruns =
3150 cpu_to_be64(read_dev_cntr(dd, C_RCV_OVF, CNTR_INVALID_VL));
3151 }
3152
pma_get_opa_porterrors(struct opa_pma_mad * pmp,struct ib_device * ibdev,u32 port,u32 * resp_len)3153 static int pma_get_opa_porterrors(struct opa_pma_mad *pmp,
3154 struct ib_device *ibdev,
3155 u32 port, u32 *resp_len)
3156 {
3157 size_t response_data_size;
3158 struct _port_ectrs *rsp;
3159 u32 port_num;
3160 struct opa_port_error_counters64_msg *req;
3161 struct hfi1_devdata *dd = dd_from_ibdev(ibdev);
3162 u32 num_ports;
3163 u8 num_pslm;
3164 u8 num_vls;
3165 struct hfi1_ibport *ibp;
3166 struct hfi1_pportdata *ppd;
3167 struct _vls_ectrs *vlinfo;
3168 unsigned long vl;
3169 u64 port_mask, tmp;
3170 unsigned long vl_select_mask;
3171 int vfi;
3172
3173 req = (struct opa_port_error_counters64_msg *)pmp->data;
3174
3175 num_ports = be32_to_cpu(pmp->mad_hdr.attr_mod) >> 24;
3176
3177 num_pslm = hweight64(be64_to_cpu(req->port_select_mask[3]));
3178 num_vls = hweight32(be32_to_cpu(req->vl_select_mask));
3179
3180 if (num_ports != 1 || num_ports != num_pslm) {
3181 pmp->mad_hdr.status |= IB_SMP_INVALID_FIELD;
3182 return reply((struct ib_mad_hdr *)pmp);
3183 }
3184
3185 response_data_size = struct_size(req, port.vls, num_vls);
3186
3187 if (response_data_size > sizeof(pmp->data)) {
3188 pmp->mad_hdr.status |= IB_SMP_INVALID_FIELD;
3189 return reply((struct ib_mad_hdr *)pmp);
3190 }
3191 /*
3192 * The bit set in the mask needs to be consistent with the
3193 * port the request came in on.
3194 */
3195 port_mask = be64_to_cpu(req->port_select_mask[3]);
3196 port_num = find_first_bit((unsigned long *)&port_mask,
3197 sizeof(port_mask) * 8);
3198
3199 if (port_num != port) {
3200 pmp->mad_hdr.status |= IB_SMP_INVALID_FIELD;
3201 return reply((struct ib_mad_hdr *)pmp);
3202 }
3203
3204 rsp = &req->port;
3205
3206 ibp = to_iport(ibdev, port_num);
3207 ppd = ppd_from_ibp(ibp);
3208
3209 memset(rsp, 0, sizeof(*rsp));
3210 rsp->port_number = port_num;
3211
3212 pma_get_opa_port_ectrs(ibdev, rsp, port_num);
3213
3214 rsp->port_rcv_remote_physical_errors =
3215 cpu_to_be64(read_dev_cntr(dd, C_DC_RMT_PHY_ERR,
3216 CNTR_INVALID_VL));
3217 rsp->fm_config_errors =
3218 cpu_to_be64(read_dev_cntr(dd, C_DC_FM_CFG_ERR,
3219 CNTR_INVALID_VL));
3220 tmp = read_dev_cntr(dd, C_DC_UNC_ERR, CNTR_INVALID_VL);
3221
3222 rsp->uncorrectable_errors = tmp < 0x100 ? (tmp & 0xff) : 0xff;
3223 rsp->port_rcv_errors =
3224 cpu_to_be64(read_dev_cntr(dd, C_DC_RCV_ERR, CNTR_INVALID_VL));
3225 vlinfo = &rsp->vls[0];
3226 vfi = 0;
3227 vl_select_mask = be32_to_cpu(req->vl_select_mask);
3228 for_each_set_bit(vl, &vl_select_mask, BITS_PER_LONG) {
3229 memset(vlinfo, 0, sizeof(*vlinfo));
3230 rsp->vls[vfi].port_vl_xmit_discards =
3231 cpu_to_be64(read_port_cntr(ppd, C_SW_XMIT_DSCD_VL,
3232 idx_from_vl(vl)));
3233 vlinfo += 1;
3234 vfi++;
3235 }
3236
3237 if (resp_len)
3238 *resp_len += response_data_size;
3239
3240 return reply((struct ib_mad_hdr *)pmp);
3241 }
3242
pma_get_ib_portcounters(struct ib_pma_mad * pmp,struct ib_device * ibdev,u32 port)3243 static int pma_get_ib_portcounters(struct ib_pma_mad *pmp,
3244 struct ib_device *ibdev, u32 port)
3245 {
3246 struct ib_pma_portcounters *p = (struct ib_pma_portcounters *)
3247 pmp->data;
3248 struct _port_ectrs rsp;
3249 u64 temp_link_overrun_errors;
3250 u64 temp_64;
3251 u32 temp_32;
3252
3253 memset(&rsp, 0, sizeof(rsp));
3254 pma_get_opa_port_ectrs(ibdev, &rsp, port);
3255
3256 if (pmp->mad_hdr.attr_mod != 0 || p->port_select != port) {
3257 pmp->mad_hdr.status |= IB_SMP_INVALID_FIELD;
3258 goto bail;
3259 }
3260
3261 p->symbol_error_counter = 0; /* N/A for OPA */
3262
3263 temp_32 = be32_to_cpu(rsp.link_error_recovery);
3264 if (temp_32 > 0xFFUL)
3265 p->link_error_recovery_counter = 0xFF;
3266 else
3267 p->link_error_recovery_counter = (u8)temp_32;
3268
3269 temp_32 = be32_to_cpu(rsp.link_downed);
3270 if (temp_32 > 0xFFUL)
3271 p->link_downed_counter = 0xFF;
3272 else
3273 p->link_downed_counter = (u8)temp_32;
3274
3275 temp_64 = be64_to_cpu(rsp.port_rcv_errors);
3276 if (temp_64 > 0xFFFFUL)
3277 p->port_rcv_errors = cpu_to_be16(0xFFFF);
3278 else
3279 p->port_rcv_errors = cpu_to_be16((u16)temp_64);
3280
3281 temp_64 = be64_to_cpu(rsp.port_rcv_remote_physical_errors);
3282 if (temp_64 > 0xFFFFUL)
3283 p->port_rcv_remphys_errors = cpu_to_be16(0xFFFF);
3284 else
3285 p->port_rcv_remphys_errors = cpu_to_be16((u16)temp_64);
3286
3287 temp_64 = be64_to_cpu(rsp.port_rcv_switch_relay_errors);
3288 p->port_rcv_switch_relay_errors = cpu_to_be16((u16)temp_64);
3289
3290 temp_64 = be64_to_cpu(rsp.port_xmit_discards);
3291 if (temp_64 > 0xFFFFUL)
3292 p->port_xmit_discards = cpu_to_be16(0xFFFF);
3293 else
3294 p->port_xmit_discards = cpu_to_be16((u16)temp_64);
3295
3296 temp_64 = be64_to_cpu(rsp.port_xmit_constraint_errors);
3297 if (temp_64 > 0xFFUL)
3298 p->port_xmit_constraint_errors = 0xFF;
3299 else
3300 p->port_xmit_constraint_errors = (u8)temp_64;
3301
3302 temp_64 = be64_to_cpu(rsp.port_rcv_constraint_errors);
3303 if (temp_64 > 0xFFUL)
3304 p->port_rcv_constraint_errors = 0xFFUL;
3305 else
3306 p->port_rcv_constraint_errors = (u8)temp_64;
3307
3308 /* LocalLink: 7:4, BufferOverrun: 3:0 */
3309 temp_64 = be64_to_cpu(rsp.local_link_integrity_errors);
3310 if (temp_64 > 0xFUL)
3311 temp_64 = 0xFUL;
3312
3313 temp_link_overrun_errors = temp_64 << 4;
3314
3315 temp_64 = be64_to_cpu(rsp.excessive_buffer_overruns);
3316 if (temp_64 > 0xFUL)
3317 temp_64 = 0xFUL;
3318 temp_link_overrun_errors |= temp_64;
3319
3320 p->link_overrun_errors = (u8)temp_link_overrun_errors;
3321
3322 p->vl15_dropped = 0; /* N/A for OPA */
3323
3324 bail:
3325 return reply((struct ib_mad_hdr *)pmp);
3326 }
3327
pma_get_opa_errorinfo(struct opa_pma_mad * pmp,struct ib_device * ibdev,u32 port,u32 * resp_len)3328 static int pma_get_opa_errorinfo(struct opa_pma_mad *pmp,
3329 struct ib_device *ibdev,
3330 u32 port, u32 *resp_len)
3331 {
3332 size_t response_data_size;
3333 struct _port_ei *rsp;
3334 struct opa_port_error_info_msg *req;
3335 struct hfi1_devdata *dd = dd_from_ibdev(ibdev);
3336 u64 port_mask;
3337 u32 num_ports;
3338 u32 port_num;
3339 u8 num_pslm;
3340 u64 reg;
3341
3342 req = (struct opa_port_error_info_msg *)pmp->data;
3343 rsp = &req->port;
3344
3345 num_ports = OPA_AM_NPORT(be32_to_cpu(pmp->mad_hdr.attr_mod));
3346 num_pslm = hweight64(be64_to_cpu(req->port_select_mask[3]));
3347
3348 memset(rsp, 0, sizeof(*rsp));
3349
3350 if (num_ports != 1 || num_ports != num_pslm) {
3351 pmp->mad_hdr.status |= IB_SMP_INVALID_FIELD;
3352 return reply((struct ib_mad_hdr *)pmp);
3353 }
3354
3355 /* Sanity check */
3356 response_data_size = sizeof(struct opa_port_error_info_msg);
3357
3358 if (response_data_size > sizeof(pmp->data)) {
3359 pmp->mad_hdr.status |= IB_SMP_INVALID_FIELD;
3360 return reply((struct ib_mad_hdr *)pmp);
3361 }
3362
3363 /*
3364 * The bit set in the mask needs to be consistent with the port
3365 * the request came in on.
3366 */
3367 port_mask = be64_to_cpu(req->port_select_mask[3]);
3368 port_num = find_first_bit((unsigned long *)&port_mask,
3369 sizeof(port_mask) * 8);
3370
3371 if (port_num != port) {
3372 pmp->mad_hdr.status |= IB_SMP_INVALID_FIELD;
3373 return reply((struct ib_mad_hdr *)pmp);
3374 }
3375 rsp->port_number = port;
3376
3377 /* PortRcvErrorInfo */
3378 rsp->port_rcv_ei.status_and_code =
3379 dd->err_info_rcvport.status_and_code;
3380 memcpy(&rsp->port_rcv_ei.ei.ei1to12.packet_flit1,
3381 &dd->err_info_rcvport.packet_flit1, sizeof(u64));
3382 memcpy(&rsp->port_rcv_ei.ei.ei1to12.packet_flit2,
3383 &dd->err_info_rcvport.packet_flit2, sizeof(u64));
3384
3385 /* ExcessiverBufferOverrunInfo */
3386 reg = read_csr(dd, RCV_ERR_INFO);
3387 if (reg & RCV_ERR_INFO_RCV_EXCESS_BUFFER_OVERRUN_SMASK) {
3388 /*
3389 * if the RcvExcessBufferOverrun bit is set, save SC of
3390 * first pkt that encountered an excess buffer overrun
3391 */
3392 u8 tmp = (u8)reg;
3393
3394 tmp &= RCV_ERR_INFO_RCV_EXCESS_BUFFER_OVERRUN_SC_SMASK;
3395 tmp <<= 2;
3396 rsp->excessive_buffer_overrun_ei.status_and_sc = tmp;
3397 /* set the status bit */
3398 rsp->excessive_buffer_overrun_ei.status_and_sc |= 0x80;
3399 }
3400
3401 rsp->port_xmit_constraint_ei.status =
3402 dd->err_info_xmit_constraint.status;
3403 rsp->port_xmit_constraint_ei.pkey =
3404 cpu_to_be16(dd->err_info_xmit_constraint.pkey);
3405 rsp->port_xmit_constraint_ei.slid =
3406 cpu_to_be32(dd->err_info_xmit_constraint.slid);
3407
3408 rsp->port_rcv_constraint_ei.status =
3409 dd->err_info_rcv_constraint.status;
3410 rsp->port_rcv_constraint_ei.pkey =
3411 cpu_to_be16(dd->err_info_rcv_constraint.pkey);
3412 rsp->port_rcv_constraint_ei.slid =
3413 cpu_to_be32(dd->err_info_rcv_constraint.slid);
3414
3415 /* UncorrectableErrorInfo */
3416 rsp->uncorrectable_ei.status_and_code = dd->err_info_uncorrectable;
3417
3418 /* FMConfigErrorInfo */
3419 rsp->fm_config_ei.status_and_code = dd->err_info_fmconfig;
3420
3421 if (resp_len)
3422 *resp_len += response_data_size;
3423
3424 return reply((struct ib_mad_hdr *)pmp);
3425 }
3426
pma_set_opa_portstatus(struct opa_pma_mad * pmp,struct ib_device * ibdev,u32 port,u32 * resp_len)3427 static int pma_set_opa_portstatus(struct opa_pma_mad *pmp,
3428 struct ib_device *ibdev,
3429 u32 port, u32 *resp_len)
3430 {
3431 struct opa_clear_port_status *req =
3432 (struct opa_clear_port_status *)pmp->data;
3433 struct hfi1_devdata *dd = dd_from_ibdev(ibdev);
3434 struct hfi1_ibport *ibp = to_iport(ibdev, port);
3435 struct hfi1_pportdata *ppd = ppd_from_ibp(ibp);
3436 u32 nports = be32_to_cpu(pmp->mad_hdr.attr_mod) >> 24;
3437 u64 portn = be64_to_cpu(req->port_select_mask[3]);
3438 u32 counter_select = be32_to_cpu(req->counter_select_mask);
3439 unsigned long vl_select_mask = VL_MASK_ALL; /* clear all per-vl cnts */
3440 unsigned long vl;
3441
3442 if ((nports != 1) || (portn != 1 << port)) {
3443 pmp->mad_hdr.status |= IB_SMP_INVALID_FIELD;
3444 return reply((struct ib_mad_hdr *)pmp);
3445 }
3446 /*
3447 * only counters returned by pma_get_opa_portstatus() are
3448 * handled, so when pma_get_opa_portstatus() gets a fix,
3449 * the corresponding change should be made here as well.
3450 */
3451
3452 if (counter_select & CS_PORT_XMIT_DATA)
3453 write_dev_cntr(dd, C_DC_XMIT_FLITS, CNTR_INVALID_VL, 0);
3454
3455 if (counter_select & CS_PORT_RCV_DATA)
3456 write_dev_cntr(dd, C_DC_RCV_FLITS, CNTR_INVALID_VL, 0);
3457
3458 if (counter_select & CS_PORT_XMIT_PKTS)
3459 write_dev_cntr(dd, C_DC_XMIT_PKTS, CNTR_INVALID_VL, 0);
3460
3461 if (counter_select & CS_PORT_RCV_PKTS)
3462 write_dev_cntr(dd, C_DC_RCV_PKTS, CNTR_INVALID_VL, 0);
3463
3464 if (counter_select & CS_PORT_MCAST_XMIT_PKTS)
3465 write_dev_cntr(dd, C_DC_MC_XMIT_PKTS, CNTR_INVALID_VL, 0);
3466
3467 if (counter_select & CS_PORT_MCAST_RCV_PKTS)
3468 write_dev_cntr(dd, C_DC_MC_RCV_PKTS, CNTR_INVALID_VL, 0);
3469
3470 if (counter_select & CS_PORT_XMIT_WAIT) {
3471 write_port_cntr(ppd, C_TX_WAIT, CNTR_INVALID_VL, 0);
3472 ppd->port_vl_xmit_wait_last[C_VL_COUNT] = 0;
3473 ppd->vl_xmit_flit_cnt[C_VL_COUNT] = 0;
3474 }
3475 /* ignore cs_sw_portCongestion for HFIs */
3476
3477 if (counter_select & CS_PORT_RCV_FECN)
3478 write_dev_cntr(dd, C_DC_RCV_FCN, CNTR_INVALID_VL, 0);
3479
3480 if (counter_select & CS_PORT_RCV_BECN)
3481 write_dev_cntr(dd, C_DC_RCV_BCN, CNTR_INVALID_VL, 0);
3482
3483 /* ignore cs_port_xmit_time_cong for HFIs */
3484 /* ignore cs_port_xmit_wasted_bw for now */
3485 /* ignore cs_port_xmit_wait_data for now */
3486 if (counter_select & CS_PORT_RCV_BUBBLE)
3487 write_dev_cntr(dd, C_DC_RCV_BBL, CNTR_INVALID_VL, 0);
3488
3489 /* Only applicable for switch */
3490 /* if (counter_select & CS_PORT_MARK_FECN)
3491 * write_csr(dd, DCC_PRF_PORT_MARK_FECN_CNT, 0);
3492 */
3493
3494 if (counter_select & CS_PORT_RCV_CONSTRAINT_ERRORS)
3495 write_port_cntr(ppd, C_SW_RCV_CSTR_ERR, CNTR_INVALID_VL, 0);
3496
3497 /* ignore cs_port_rcv_switch_relay_errors for HFIs */
3498 if (counter_select & CS_PORT_XMIT_DISCARDS)
3499 write_port_cntr(ppd, C_SW_XMIT_DSCD, CNTR_INVALID_VL, 0);
3500
3501 if (counter_select & CS_PORT_XMIT_CONSTRAINT_ERRORS)
3502 write_port_cntr(ppd, C_SW_XMIT_CSTR_ERR, CNTR_INVALID_VL, 0);
3503
3504 if (counter_select & CS_PORT_RCV_REMOTE_PHYSICAL_ERRORS)
3505 write_dev_cntr(dd, C_DC_RMT_PHY_ERR, CNTR_INVALID_VL, 0);
3506
3507 if (counter_select & CS_LOCAL_LINK_INTEGRITY_ERRORS)
3508 write_dev_cntr(dd, C_DC_RX_REPLAY, CNTR_INVALID_VL, 0);
3509
3510 if (counter_select & CS_LINK_ERROR_RECOVERY) {
3511 write_dev_cntr(dd, C_DC_SEQ_CRC_CNT, CNTR_INVALID_VL, 0);
3512 write_dev_cntr(dd, C_DC_REINIT_FROM_PEER_CNT,
3513 CNTR_INVALID_VL, 0);
3514 }
3515
3516 if (counter_select & CS_PORT_RCV_ERRORS)
3517 write_dev_cntr(dd, C_DC_RCV_ERR, CNTR_INVALID_VL, 0);
3518
3519 if (counter_select & CS_EXCESSIVE_BUFFER_OVERRUNS) {
3520 write_dev_cntr(dd, C_RCV_OVF, CNTR_INVALID_VL, 0);
3521 dd->rcv_ovfl_cnt = 0;
3522 }
3523
3524 if (counter_select & CS_FM_CONFIG_ERRORS)
3525 write_dev_cntr(dd, C_DC_FM_CFG_ERR, CNTR_INVALID_VL, 0);
3526
3527 if (counter_select & CS_LINK_DOWNED)
3528 write_port_cntr(ppd, C_SW_LINK_DOWN, CNTR_INVALID_VL, 0);
3529
3530 if (counter_select & CS_UNCORRECTABLE_ERRORS)
3531 write_dev_cntr(dd, C_DC_UNC_ERR, CNTR_INVALID_VL, 0);
3532
3533 for_each_set_bit(vl, &vl_select_mask, BITS_PER_LONG) {
3534 if (counter_select & CS_PORT_XMIT_DATA)
3535 write_port_cntr(ppd, C_TX_FLIT_VL, idx_from_vl(vl), 0);
3536
3537 if (counter_select & CS_PORT_RCV_DATA)
3538 write_dev_cntr(dd, C_DC_RX_FLIT_VL, idx_from_vl(vl), 0);
3539
3540 if (counter_select & CS_PORT_XMIT_PKTS)
3541 write_port_cntr(ppd, C_TX_PKT_VL, idx_from_vl(vl), 0);
3542
3543 if (counter_select & CS_PORT_RCV_PKTS)
3544 write_dev_cntr(dd, C_DC_RX_PKT_VL, idx_from_vl(vl), 0);
3545
3546 if (counter_select & CS_PORT_XMIT_WAIT) {
3547 write_port_cntr(ppd, C_TX_WAIT_VL, idx_from_vl(vl), 0);
3548 ppd->port_vl_xmit_wait_last[idx_from_vl(vl)] = 0;
3549 ppd->vl_xmit_flit_cnt[idx_from_vl(vl)] = 0;
3550 }
3551
3552 /* sw_port_vl_congestion is 0 for HFIs */
3553 if (counter_select & CS_PORT_RCV_FECN)
3554 write_dev_cntr(dd, C_DC_RCV_FCN_VL, idx_from_vl(vl), 0);
3555
3556 if (counter_select & CS_PORT_RCV_BECN)
3557 write_dev_cntr(dd, C_DC_RCV_BCN_VL, idx_from_vl(vl), 0);
3558
3559 /* port_vl_xmit_time_cong is 0 for HFIs */
3560 /* port_vl_xmit_wasted_bw ??? */
3561 /* port_vl_xmit_wait_data - TXE (table 13-9 HFI spec) ??? */
3562 if (counter_select & CS_PORT_RCV_BUBBLE)
3563 write_dev_cntr(dd, C_DC_RCV_BBL_VL, idx_from_vl(vl), 0);
3564
3565 /* if (counter_select & CS_PORT_MARK_FECN)
3566 * write_csr(dd, DCC_PRF_PORT_VL_MARK_FECN_CNT + offset, 0);
3567 */
3568 if (counter_select & C_SW_XMIT_DSCD_VL)
3569 write_port_cntr(ppd, C_SW_XMIT_DSCD_VL,
3570 idx_from_vl(vl), 0);
3571 }
3572
3573 if (resp_len)
3574 *resp_len += sizeof(*req);
3575
3576 return reply((struct ib_mad_hdr *)pmp);
3577 }
3578
pma_set_opa_errorinfo(struct opa_pma_mad * pmp,struct ib_device * ibdev,u32 port,u32 * resp_len)3579 static int pma_set_opa_errorinfo(struct opa_pma_mad *pmp,
3580 struct ib_device *ibdev,
3581 u32 port, u32 *resp_len)
3582 {
3583 struct _port_ei *rsp;
3584 struct opa_port_error_info_msg *req;
3585 struct hfi1_devdata *dd = dd_from_ibdev(ibdev);
3586 u64 port_mask;
3587 u32 num_ports;
3588 u32 port_num;
3589 u8 num_pslm;
3590 u32 error_info_select;
3591
3592 req = (struct opa_port_error_info_msg *)pmp->data;
3593 rsp = &req->port;
3594
3595 num_ports = OPA_AM_NPORT(be32_to_cpu(pmp->mad_hdr.attr_mod));
3596 num_pslm = hweight64(be64_to_cpu(req->port_select_mask[3]));
3597
3598 memset(rsp, 0, sizeof(*rsp));
3599
3600 if (num_ports != 1 || num_ports != num_pslm) {
3601 pmp->mad_hdr.status |= IB_SMP_INVALID_FIELD;
3602 return reply((struct ib_mad_hdr *)pmp);
3603 }
3604
3605 /*
3606 * The bit set in the mask needs to be consistent with the port
3607 * the request came in on.
3608 */
3609 port_mask = be64_to_cpu(req->port_select_mask[3]);
3610 port_num = find_first_bit((unsigned long *)&port_mask,
3611 sizeof(port_mask) * 8);
3612
3613 if (port_num != port) {
3614 pmp->mad_hdr.status |= IB_SMP_INVALID_FIELD;
3615 return reply((struct ib_mad_hdr *)pmp);
3616 }
3617
3618 error_info_select = be32_to_cpu(req->error_info_select_mask);
3619
3620 /* PortRcvErrorInfo */
3621 if (error_info_select & ES_PORT_RCV_ERROR_INFO)
3622 /* turn off status bit */
3623 dd->err_info_rcvport.status_and_code &= ~OPA_EI_STATUS_SMASK;
3624
3625 /* ExcessiverBufferOverrunInfo */
3626 if (error_info_select & ES_EXCESSIVE_BUFFER_OVERRUN_INFO)
3627 /*
3628 * status bit is essentially kept in the h/w - bit 5 of
3629 * RCV_ERR_INFO
3630 */
3631 write_csr(dd, RCV_ERR_INFO,
3632 RCV_ERR_INFO_RCV_EXCESS_BUFFER_OVERRUN_SMASK);
3633
3634 if (error_info_select & ES_PORT_XMIT_CONSTRAINT_ERROR_INFO)
3635 dd->err_info_xmit_constraint.status &= ~OPA_EI_STATUS_SMASK;
3636
3637 if (error_info_select & ES_PORT_RCV_CONSTRAINT_ERROR_INFO)
3638 dd->err_info_rcv_constraint.status &= ~OPA_EI_STATUS_SMASK;
3639
3640 /* UncorrectableErrorInfo */
3641 if (error_info_select & ES_UNCORRECTABLE_ERROR_INFO)
3642 /* turn off status bit */
3643 dd->err_info_uncorrectable &= ~OPA_EI_STATUS_SMASK;
3644
3645 /* FMConfigErrorInfo */
3646 if (error_info_select & ES_FM_CONFIG_ERROR_INFO)
3647 /* turn off status bit */
3648 dd->err_info_fmconfig &= ~OPA_EI_STATUS_SMASK;
3649
3650 if (resp_len)
3651 *resp_len += sizeof(*req);
3652
3653 return reply((struct ib_mad_hdr *)pmp);
3654 }
3655
3656 struct opa_congestion_info_attr {
3657 __be16 congestion_info;
3658 u8 control_table_cap; /* Multiple of 64 entry unit CCTs */
3659 u8 congestion_log_length;
3660 } __packed;
3661
__subn_get_opa_cong_info(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)3662 static int __subn_get_opa_cong_info(struct opa_smp *smp, u32 am, u8 *data,
3663 struct ib_device *ibdev, u32 port,
3664 u32 *resp_len, u32 max_len)
3665 {
3666 struct opa_congestion_info_attr *p =
3667 (struct opa_congestion_info_attr *)data;
3668 struct hfi1_ibport *ibp = to_iport(ibdev, port);
3669 struct hfi1_pportdata *ppd = ppd_from_ibp(ibp);
3670
3671 if (smp_length_check(sizeof(*p), max_len)) {
3672 smp->status |= IB_SMP_INVALID_FIELD;
3673 return reply((struct ib_mad_hdr *)smp);
3674 }
3675
3676 p->congestion_info = 0;
3677 p->control_table_cap = ppd->cc_max_table_entries;
3678 p->congestion_log_length = OPA_CONG_LOG_ELEMS;
3679
3680 if (resp_len)
3681 *resp_len += sizeof(*p);
3682
3683 return reply((struct ib_mad_hdr *)smp);
3684 }
3685
__subn_get_opa_cong_setting(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)3686 static int __subn_get_opa_cong_setting(struct opa_smp *smp, u32 am,
3687 u8 *data, struct ib_device *ibdev,
3688 u32 port, u32 *resp_len, u32 max_len)
3689 {
3690 int i;
3691 struct opa_congestion_setting_attr *p =
3692 (struct opa_congestion_setting_attr *)data;
3693 struct hfi1_ibport *ibp = to_iport(ibdev, port);
3694 struct hfi1_pportdata *ppd = ppd_from_ibp(ibp);
3695 struct opa_congestion_setting_entry_shadow *entries;
3696 struct cc_state *cc_state;
3697
3698 if (smp_length_check(sizeof(*p), max_len)) {
3699 smp->status |= IB_SMP_INVALID_FIELD;
3700 return reply((struct ib_mad_hdr *)smp);
3701 }
3702
3703 rcu_read_lock();
3704
3705 cc_state = get_cc_state(ppd);
3706
3707 if (!cc_state) {
3708 rcu_read_unlock();
3709 return reply((struct ib_mad_hdr *)smp);
3710 }
3711
3712 entries = cc_state->cong_setting.entries;
3713 p->port_control = cpu_to_be16(cc_state->cong_setting.port_control);
3714 p->control_map = cpu_to_be32(cc_state->cong_setting.control_map);
3715 for (i = 0; i < OPA_MAX_SLS; i++) {
3716 p->entries[i].ccti_increase = entries[i].ccti_increase;
3717 p->entries[i].ccti_timer = cpu_to_be16(entries[i].ccti_timer);
3718 p->entries[i].trigger_threshold =
3719 entries[i].trigger_threshold;
3720 p->entries[i].ccti_min = entries[i].ccti_min;
3721 }
3722
3723 rcu_read_unlock();
3724
3725 if (resp_len)
3726 *resp_len += sizeof(*p);
3727
3728 return reply((struct ib_mad_hdr *)smp);
3729 }
3730
3731 /*
3732 * Apply congestion control information stored in the ppd to the
3733 * active structure.
3734 */
apply_cc_state(struct hfi1_pportdata * ppd)3735 static void apply_cc_state(struct hfi1_pportdata *ppd)
3736 {
3737 struct cc_state *old_cc_state, *new_cc_state;
3738
3739 new_cc_state = kzalloc(sizeof(*new_cc_state), GFP_KERNEL);
3740 if (!new_cc_state)
3741 return;
3742
3743 /*
3744 * Hold the lock for updating *and* to prevent ppd information
3745 * from changing during the update.
3746 */
3747 spin_lock(&ppd->cc_state_lock);
3748
3749 old_cc_state = get_cc_state_protected(ppd);
3750 if (!old_cc_state) {
3751 /* never active, or shutting down */
3752 spin_unlock(&ppd->cc_state_lock);
3753 kfree(new_cc_state);
3754 return;
3755 }
3756
3757 *new_cc_state = *old_cc_state;
3758
3759 if (ppd->total_cct_entry)
3760 new_cc_state->cct.ccti_limit = ppd->total_cct_entry - 1;
3761 else
3762 new_cc_state->cct.ccti_limit = 0;
3763
3764 memcpy(new_cc_state->cct.entries, ppd->ccti_entries,
3765 ppd->total_cct_entry * sizeof(struct ib_cc_table_entry));
3766
3767 new_cc_state->cong_setting.port_control = IB_CC_CCS_PC_SL_BASED;
3768 new_cc_state->cong_setting.control_map = ppd->cc_sl_control_map;
3769 memcpy(new_cc_state->cong_setting.entries, ppd->congestion_entries,
3770 OPA_MAX_SLS * sizeof(struct opa_congestion_setting_entry));
3771
3772 rcu_assign_pointer(ppd->cc_state, new_cc_state);
3773
3774 spin_unlock(&ppd->cc_state_lock);
3775
3776 kfree_rcu(old_cc_state, rcu);
3777 }
3778
__subn_set_opa_cong_setting(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)3779 static int __subn_set_opa_cong_setting(struct opa_smp *smp, u32 am, u8 *data,
3780 struct ib_device *ibdev, u32 port,
3781 u32 *resp_len, u32 max_len)
3782 {
3783 struct opa_congestion_setting_attr *p =
3784 (struct opa_congestion_setting_attr *)data;
3785 struct hfi1_ibport *ibp = to_iport(ibdev, port);
3786 struct hfi1_pportdata *ppd = ppd_from_ibp(ibp);
3787 struct opa_congestion_setting_entry_shadow *entries;
3788 int i;
3789
3790 if (smp_length_check(sizeof(*p), max_len)) {
3791 smp->status |= IB_SMP_INVALID_FIELD;
3792 return reply((struct ib_mad_hdr *)smp);
3793 }
3794
3795 /*
3796 * Save details from packet into the ppd. Hold the cc_state_lock so
3797 * our information is consistent with anyone trying to apply the state.
3798 */
3799 spin_lock(&ppd->cc_state_lock);
3800 ppd->cc_sl_control_map = be32_to_cpu(p->control_map);
3801
3802 entries = ppd->congestion_entries;
3803 for (i = 0; i < OPA_MAX_SLS; i++) {
3804 entries[i].ccti_increase = p->entries[i].ccti_increase;
3805 entries[i].ccti_timer = be16_to_cpu(p->entries[i].ccti_timer);
3806 entries[i].trigger_threshold =
3807 p->entries[i].trigger_threshold;
3808 entries[i].ccti_min = p->entries[i].ccti_min;
3809 }
3810 spin_unlock(&ppd->cc_state_lock);
3811
3812 /* now apply the information */
3813 apply_cc_state(ppd);
3814
3815 return __subn_get_opa_cong_setting(smp, am, data, ibdev, port,
3816 resp_len, max_len);
3817 }
3818
__subn_get_opa_hfi1_cong_log(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)3819 static int __subn_get_opa_hfi1_cong_log(struct opa_smp *smp, u32 am,
3820 u8 *data, struct ib_device *ibdev,
3821 u32 port, u32 *resp_len, u32 max_len)
3822 {
3823 struct hfi1_ibport *ibp = to_iport(ibdev, port);
3824 struct hfi1_pportdata *ppd = ppd_from_ibp(ibp);
3825 struct opa_hfi1_cong_log *cong_log = (struct opa_hfi1_cong_log *)data;
3826 u64 ts;
3827 int i;
3828
3829 if (am || smp_length_check(sizeof(*cong_log), max_len)) {
3830 smp->status |= IB_SMP_INVALID_FIELD;
3831 return reply((struct ib_mad_hdr *)smp);
3832 }
3833
3834 spin_lock_irq(&ppd->cc_log_lock);
3835
3836 cong_log->log_type = OPA_CC_LOG_TYPE_HFI;
3837 cong_log->congestion_flags = 0;
3838 cong_log->threshold_event_counter =
3839 cpu_to_be16(ppd->threshold_event_counter);
3840 memcpy(cong_log->threshold_cong_event_map,
3841 ppd->threshold_cong_event_map,
3842 sizeof(cong_log->threshold_cong_event_map));
3843 /* keep timestamp in units of 1.024 usec */
3844 ts = ktime_get_ns() / 1024;
3845 cong_log->current_time_stamp = cpu_to_be32(ts);
3846 for (i = 0; i < OPA_CONG_LOG_ELEMS; i++) {
3847 struct opa_hfi1_cong_log_event_internal *cce =
3848 &ppd->cc_events[ppd->cc_mad_idx++];
3849 if (ppd->cc_mad_idx == OPA_CONG_LOG_ELEMS)
3850 ppd->cc_mad_idx = 0;
3851 /*
3852 * Entries which are older than twice the time
3853 * required to wrap the counter are supposed to
3854 * be zeroed (CA10-49 IBTA, release 1.2.1, V1).
3855 */
3856 if ((ts - cce->timestamp) / 2 > U32_MAX)
3857 continue;
3858 memcpy(cong_log->events[i].local_qp_cn_entry, &cce->lqpn, 3);
3859 memcpy(cong_log->events[i].remote_qp_number_cn_entry,
3860 &cce->rqpn, 3);
3861 cong_log->events[i].sl_svc_type_cn_entry =
3862 ((cce->sl & 0x1f) << 3) | (cce->svc_type & 0x7);
3863 cong_log->events[i].remote_lid_cn_entry =
3864 cpu_to_be32(cce->rlid);
3865 cong_log->events[i].timestamp_cn_entry =
3866 cpu_to_be32(cce->timestamp);
3867 }
3868
3869 /*
3870 * Reset threshold_cong_event_map, and threshold_event_counter
3871 * to 0 when log is read.
3872 */
3873 memset(ppd->threshold_cong_event_map, 0x0,
3874 sizeof(ppd->threshold_cong_event_map));
3875 ppd->threshold_event_counter = 0;
3876
3877 spin_unlock_irq(&ppd->cc_log_lock);
3878
3879 if (resp_len)
3880 *resp_len += sizeof(struct opa_hfi1_cong_log);
3881
3882 return reply((struct ib_mad_hdr *)smp);
3883 }
3884
__subn_get_opa_cc_table(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)3885 static int __subn_get_opa_cc_table(struct opa_smp *smp, u32 am, u8 *data,
3886 struct ib_device *ibdev, u32 port,
3887 u32 *resp_len, u32 max_len)
3888 {
3889 struct ib_cc_table_attr *cc_table_attr =
3890 (struct ib_cc_table_attr *)data;
3891 struct hfi1_ibport *ibp = to_iport(ibdev, port);
3892 struct hfi1_pportdata *ppd = ppd_from_ibp(ibp);
3893 u32 start_block = OPA_AM_START_BLK(am);
3894 u32 n_blocks = OPA_AM_NBLK(am);
3895 struct ib_cc_table_entry_shadow *entries;
3896 int i, j;
3897 u32 sentry, eentry;
3898 struct cc_state *cc_state;
3899 u32 size = sizeof(u16) * (IB_CCT_ENTRIES * n_blocks + 1);
3900
3901 /* sanity check n_blocks, start_block */
3902 if (n_blocks == 0 || smp_length_check(size, max_len) ||
3903 start_block + n_blocks > ppd->cc_max_table_entries) {
3904 smp->status |= IB_SMP_INVALID_FIELD;
3905 return reply((struct ib_mad_hdr *)smp);
3906 }
3907
3908 rcu_read_lock();
3909
3910 cc_state = get_cc_state(ppd);
3911
3912 if (!cc_state) {
3913 rcu_read_unlock();
3914 return reply((struct ib_mad_hdr *)smp);
3915 }
3916
3917 sentry = start_block * IB_CCT_ENTRIES;
3918 eentry = sentry + (IB_CCT_ENTRIES * n_blocks);
3919
3920 cc_table_attr->ccti_limit = cpu_to_be16(cc_state->cct.ccti_limit);
3921
3922 entries = cc_state->cct.entries;
3923
3924 /* return n_blocks, though the last block may not be full */
3925 for (j = 0, i = sentry; i < eentry; j++, i++)
3926 cc_table_attr->ccti_entries[j].entry =
3927 cpu_to_be16(entries[i].entry);
3928
3929 rcu_read_unlock();
3930
3931 if (resp_len)
3932 *resp_len += size;
3933
3934 return reply((struct ib_mad_hdr *)smp);
3935 }
3936
__subn_set_opa_cc_table(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)3937 static int __subn_set_opa_cc_table(struct opa_smp *smp, u32 am, u8 *data,
3938 struct ib_device *ibdev, u32 port,
3939 u32 *resp_len, u32 max_len)
3940 {
3941 struct ib_cc_table_attr *p = (struct ib_cc_table_attr *)data;
3942 struct hfi1_ibport *ibp = to_iport(ibdev, port);
3943 struct hfi1_pportdata *ppd = ppd_from_ibp(ibp);
3944 u32 start_block = OPA_AM_START_BLK(am);
3945 u32 n_blocks = OPA_AM_NBLK(am);
3946 struct ib_cc_table_entry_shadow *entries;
3947 int i, j;
3948 u32 sentry, eentry;
3949 u16 ccti_limit;
3950 u32 size = sizeof(u16) * (IB_CCT_ENTRIES * n_blocks + 1);
3951
3952 /* sanity check n_blocks, start_block */
3953 if (n_blocks == 0 || smp_length_check(size, max_len) ||
3954 start_block + n_blocks > ppd->cc_max_table_entries) {
3955 smp->status |= IB_SMP_INVALID_FIELD;
3956 return reply((struct ib_mad_hdr *)smp);
3957 }
3958
3959 sentry = start_block * IB_CCT_ENTRIES;
3960 eentry = sentry + ((n_blocks - 1) * IB_CCT_ENTRIES) +
3961 (be16_to_cpu(p->ccti_limit)) % IB_CCT_ENTRIES + 1;
3962
3963 /* sanity check ccti_limit */
3964 ccti_limit = be16_to_cpu(p->ccti_limit);
3965 if (ccti_limit + 1 > eentry) {
3966 smp->status |= IB_SMP_INVALID_FIELD;
3967 return reply((struct ib_mad_hdr *)smp);
3968 }
3969
3970 /*
3971 * Save details from packet into the ppd. Hold the cc_state_lock so
3972 * our information is consistent with anyone trying to apply the state.
3973 */
3974 spin_lock(&ppd->cc_state_lock);
3975 ppd->total_cct_entry = ccti_limit + 1;
3976 entries = ppd->ccti_entries;
3977 for (j = 0, i = sentry; i < eentry; j++, i++)
3978 entries[i].entry = be16_to_cpu(p->ccti_entries[j].entry);
3979 spin_unlock(&ppd->cc_state_lock);
3980
3981 /* now apply the information */
3982 apply_cc_state(ppd);
3983
3984 return __subn_get_opa_cc_table(smp, am, data, ibdev, port, resp_len,
3985 max_len);
3986 }
3987
3988 struct opa_led_info {
3989 __be32 rsvd_led_mask;
3990 __be32 rsvd;
3991 };
3992
3993 #define OPA_LED_SHIFT 31
3994 #define OPA_LED_MASK BIT(OPA_LED_SHIFT)
3995
__subn_get_opa_led_info(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)3996 static int __subn_get_opa_led_info(struct opa_smp *smp, u32 am, u8 *data,
3997 struct ib_device *ibdev, u32 port,
3998 u32 *resp_len, u32 max_len)
3999 {
4000 struct hfi1_devdata *dd = dd_from_ibdev(ibdev);
4001 struct hfi1_pportdata *ppd = dd->pport;
4002 struct opa_led_info *p = (struct opa_led_info *)data;
4003 u32 nport = OPA_AM_NPORT(am);
4004 u32 is_beaconing_active;
4005
4006 if (nport != 1 || smp_length_check(sizeof(*p), max_len)) {
4007 smp->status |= IB_SMP_INVALID_FIELD;
4008 return reply((struct ib_mad_hdr *)smp);
4009 }
4010
4011 /*
4012 * This pairs with the memory barrier in hfi1_start_led_override to
4013 * ensure that we read the correct state of LED beaconing represented
4014 * by led_override_timer_active
4015 */
4016 smp_rmb();
4017 is_beaconing_active = !!atomic_read(&ppd->led_override_timer_active);
4018 p->rsvd_led_mask = cpu_to_be32(is_beaconing_active << OPA_LED_SHIFT);
4019
4020 if (resp_len)
4021 *resp_len += sizeof(struct opa_led_info);
4022
4023 return reply((struct ib_mad_hdr *)smp);
4024 }
4025
__subn_set_opa_led_info(struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)4026 static int __subn_set_opa_led_info(struct opa_smp *smp, u32 am, u8 *data,
4027 struct ib_device *ibdev, u32 port,
4028 u32 *resp_len, u32 max_len)
4029 {
4030 struct hfi1_devdata *dd = dd_from_ibdev(ibdev);
4031 struct opa_led_info *p = (struct opa_led_info *)data;
4032 u32 nport = OPA_AM_NPORT(am);
4033 int on = !!(be32_to_cpu(p->rsvd_led_mask) & OPA_LED_MASK);
4034
4035 if (nport != 1 || smp_length_check(sizeof(*p), max_len)) {
4036 smp->status |= IB_SMP_INVALID_FIELD;
4037 return reply((struct ib_mad_hdr *)smp);
4038 }
4039
4040 if (on)
4041 hfi1_start_led_override(dd->pport, 2000, 1500);
4042 else
4043 shutdown_led_override(dd->pport);
4044
4045 return __subn_get_opa_led_info(smp, am, data, ibdev, port, resp_len,
4046 max_len);
4047 }
4048
subn_get_opa_sma(__be16 attr_id,struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len)4049 static int subn_get_opa_sma(__be16 attr_id, struct opa_smp *smp, u32 am,
4050 u8 *data, struct ib_device *ibdev, u32 port,
4051 u32 *resp_len, u32 max_len)
4052 {
4053 int ret;
4054 struct hfi1_ibport *ibp = to_iport(ibdev, port);
4055
4056 switch (attr_id) {
4057 case IB_SMP_ATTR_NODE_DESC:
4058 ret = __subn_get_opa_nodedesc(smp, am, data, ibdev, port,
4059 resp_len, max_len);
4060 break;
4061 case IB_SMP_ATTR_NODE_INFO:
4062 ret = __subn_get_opa_nodeinfo(smp, am, data, ibdev, port,
4063 resp_len, max_len);
4064 break;
4065 case IB_SMP_ATTR_PORT_INFO:
4066 ret = __subn_get_opa_portinfo(smp, am, data, ibdev, port,
4067 resp_len, max_len);
4068 break;
4069 case IB_SMP_ATTR_PKEY_TABLE:
4070 ret = __subn_get_opa_pkeytable(smp, am, data, ibdev, port,
4071 resp_len, max_len);
4072 break;
4073 case OPA_ATTRIB_ID_SL_TO_SC_MAP:
4074 ret = __subn_get_opa_sl_to_sc(smp, am, data, ibdev, port,
4075 resp_len, max_len);
4076 break;
4077 case OPA_ATTRIB_ID_SC_TO_SL_MAP:
4078 ret = __subn_get_opa_sc_to_sl(smp, am, data, ibdev, port,
4079 resp_len, max_len);
4080 break;
4081 case OPA_ATTRIB_ID_SC_TO_VLT_MAP:
4082 ret = __subn_get_opa_sc_to_vlt(smp, am, data, ibdev, port,
4083 resp_len, max_len);
4084 break;
4085 case OPA_ATTRIB_ID_SC_TO_VLNT_MAP:
4086 ret = __subn_get_opa_sc_to_vlnt(smp, am, data, ibdev, port,
4087 resp_len, max_len);
4088 break;
4089 case OPA_ATTRIB_ID_PORT_STATE_INFO:
4090 ret = __subn_get_opa_psi(smp, am, data, ibdev, port,
4091 resp_len, max_len);
4092 break;
4093 case OPA_ATTRIB_ID_BUFFER_CONTROL_TABLE:
4094 ret = __subn_get_opa_bct(smp, am, data, ibdev, port,
4095 resp_len, max_len);
4096 break;
4097 case OPA_ATTRIB_ID_CABLE_INFO:
4098 ret = __subn_get_opa_cable_info(smp, am, data, ibdev, port,
4099 resp_len, max_len);
4100 break;
4101 case IB_SMP_ATTR_VL_ARB_TABLE:
4102 ret = __subn_get_opa_vl_arb(smp, am, data, ibdev, port,
4103 resp_len, max_len);
4104 break;
4105 case OPA_ATTRIB_ID_CONGESTION_INFO:
4106 ret = __subn_get_opa_cong_info(smp, am, data, ibdev, port,
4107 resp_len, max_len);
4108 break;
4109 case OPA_ATTRIB_ID_HFI_CONGESTION_SETTING:
4110 ret = __subn_get_opa_cong_setting(smp, am, data, ibdev,
4111 port, resp_len, max_len);
4112 break;
4113 case OPA_ATTRIB_ID_HFI_CONGESTION_LOG:
4114 ret = __subn_get_opa_hfi1_cong_log(smp, am, data, ibdev,
4115 port, resp_len, max_len);
4116 break;
4117 case OPA_ATTRIB_ID_CONGESTION_CONTROL_TABLE:
4118 ret = __subn_get_opa_cc_table(smp, am, data, ibdev, port,
4119 resp_len, max_len);
4120 break;
4121 case IB_SMP_ATTR_LED_INFO:
4122 ret = __subn_get_opa_led_info(smp, am, data, ibdev, port,
4123 resp_len, max_len);
4124 break;
4125 case IB_SMP_ATTR_SM_INFO:
4126 if (ibp->rvp.port_cap_flags & IB_PORT_SM_DISABLED)
4127 return IB_MAD_RESULT_SUCCESS | IB_MAD_RESULT_CONSUMED;
4128 if (ibp->rvp.port_cap_flags & IB_PORT_SM)
4129 return IB_MAD_RESULT_SUCCESS;
4130 fallthrough;
4131 default:
4132 smp->status |= IB_SMP_UNSUP_METH_ATTR;
4133 ret = reply((struct ib_mad_hdr *)smp);
4134 break;
4135 }
4136 return ret;
4137 }
4138
subn_set_opa_sma(__be16 attr_id,struct opa_smp * smp,u32 am,u8 * data,struct ib_device * ibdev,u32 port,u32 * resp_len,u32 max_len,int local_mad)4139 static int subn_set_opa_sma(__be16 attr_id, struct opa_smp *smp, u32 am,
4140 u8 *data, struct ib_device *ibdev, u32 port,
4141 u32 *resp_len, u32 max_len, int local_mad)
4142 {
4143 int ret;
4144 struct hfi1_ibport *ibp = to_iport(ibdev, port);
4145
4146 switch (attr_id) {
4147 case IB_SMP_ATTR_PORT_INFO:
4148 ret = __subn_set_opa_portinfo(smp, am, data, ibdev, port,
4149 resp_len, max_len, local_mad);
4150 break;
4151 case IB_SMP_ATTR_PKEY_TABLE:
4152 ret = __subn_set_opa_pkeytable(smp, am, data, ibdev, port,
4153 resp_len, max_len);
4154 break;
4155 case OPA_ATTRIB_ID_SL_TO_SC_MAP:
4156 ret = __subn_set_opa_sl_to_sc(smp, am, data, ibdev, port,
4157 resp_len, max_len);
4158 break;
4159 case OPA_ATTRIB_ID_SC_TO_SL_MAP:
4160 ret = __subn_set_opa_sc_to_sl(smp, am, data, ibdev, port,
4161 resp_len, max_len);
4162 break;
4163 case OPA_ATTRIB_ID_SC_TO_VLT_MAP:
4164 ret = __subn_set_opa_sc_to_vlt(smp, am, data, ibdev, port,
4165 resp_len, max_len);
4166 break;
4167 case OPA_ATTRIB_ID_SC_TO_VLNT_MAP:
4168 ret = __subn_set_opa_sc_to_vlnt(smp, am, data, ibdev, port,
4169 resp_len, max_len);
4170 break;
4171 case OPA_ATTRIB_ID_PORT_STATE_INFO:
4172 ret = __subn_set_opa_psi(smp, am, data, ibdev, port,
4173 resp_len, max_len, local_mad);
4174 break;
4175 case OPA_ATTRIB_ID_BUFFER_CONTROL_TABLE:
4176 ret = __subn_set_opa_bct(smp, am, data, ibdev, port,
4177 resp_len, max_len);
4178 break;
4179 case IB_SMP_ATTR_VL_ARB_TABLE:
4180 ret = __subn_set_opa_vl_arb(smp, am, data, ibdev, port,
4181 resp_len, max_len);
4182 break;
4183 case OPA_ATTRIB_ID_HFI_CONGESTION_SETTING:
4184 ret = __subn_set_opa_cong_setting(smp, am, data, ibdev,
4185 port, resp_len, max_len);
4186 break;
4187 case OPA_ATTRIB_ID_CONGESTION_CONTROL_TABLE:
4188 ret = __subn_set_opa_cc_table(smp, am, data, ibdev, port,
4189 resp_len, max_len);
4190 break;
4191 case IB_SMP_ATTR_LED_INFO:
4192 ret = __subn_set_opa_led_info(smp, am, data, ibdev, port,
4193 resp_len, max_len);
4194 break;
4195 case IB_SMP_ATTR_SM_INFO:
4196 if (ibp->rvp.port_cap_flags & IB_PORT_SM_DISABLED)
4197 return IB_MAD_RESULT_SUCCESS | IB_MAD_RESULT_CONSUMED;
4198 if (ibp->rvp.port_cap_flags & IB_PORT_SM)
4199 return IB_MAD_RESULT_SUCCESS;
4200 fallthrough;
4201 default:
4202 smp->status |= IB_SMP_UNSUP_METH_ATTR;
4203 ret = reply((struct ib_mad_hdr *)smp);
4204 break;
4205 }
4206 return ret;
4207 }
4208
set_aggr_error(struct opa_aggregate * ag)4209 static inline void set_aggr_error(struct opa_aggregate *ag)
4210 {
4211 ag->err_reqlength |= cpu_to_be16(0x8000);
4212 }
4213
subn_get_opa_aggregate(struct opa_smp * smp,struct ib_device * ibdev,u32 port,u32 * resp_len)4214 static int subn_get_opa_aggregate(struct opa_smp *smp,
4215 struct ib_device *ibdev, u32 port,
4216 u32 *resp_len)
4217 {
4218 int i;
4219 u32 num_attr = be32_to_cpu(smp->attr_mod) & 0x000000ff;
4220 u8 *next_smp = opa_get_smp_data(smp);
4221
4222 if (num_attr < 1 || num_attr > 117) {
4223 smp->status |= IB_SMP_INVALID_FIELD;
4224 return reply((struct ib_mad_hdr *)smp);
4225 }
4226
4227 for (i = 0; i < num_attr; i++) {
4228 struct opa_aggregate *agg;
4229 size_t agg_data_len;
4230 size_t agg_size;
4231 u32 am;
4232
4233 agg = (struct opa_aggregate *)next_smp;
4234 agg_data_len = (be16_to_cpu(agg->err_reqlength) & 0x007f) * 8;
4235 agg_size = sizeof(*agg) + agg_data_len;
4236 am = be32_to_cpu(agg->attr_mod);
4237
4238 *resp_len += agg_size;
4239
4240 if (next_smp + agg_size > ((u8 *)smp) + sizeof(*smp)) {
4241 smp->status |= IB_SMP_INVALID_FIELD;
4242 return reply((struct ib_mad_hdr *)smp);
4243 }
4244
4245 /* zero the payload for this segment */
4246 memset(next_smp + sizeof(*agg), 0, agg_data_len);
4247
4248 (void)subn_get_opa_sma(agg->attr_id, smp, am, agg->data,
4249 ibdev, port, NULL, (u32)agg_data_len);
4250
4251 if (smp->status & IB_SMP_INVALID_FIELD)
4252 break;
4253 if (smp->status & ~IB_SMP_DIRECTION) {
4254 set_aggr_error(agg);
4255 return reply((struct ib_mad_hdr *)smp);
4256 }
4257 next_smp += agg_size;
4258 }
4259
4260 return reply((struct ib_mad_hdr *)smp);
4261 }
4262
subn_set_opa_aggregate(struct opa_smp * smp,struct ib_device * ibdev,u32 port,u32 * resp_len,int local_mad)4263 static int subn_set_opa_aggregate(struct opa_smp *smp,
4264 struct ib_device *ibdev, u32 port,
4265 u32 *resp_len, int local_mad)
4266 {
4267 int i;
4268 u32 num_attr = be32_to_cpu(smp->attr_mod) & 0x000000ff;
4269 u8 *next_smp = opa_get_smp_data(smp);
4270
4271 if (num_attr < 1 || num_attr > 117) {
4272 smp->status |= IB_SMP_INVALID_FIELD;
4273 return reply((struct ib_mad_hdr *)smp);
4274 }
4275
4276 for (i = 0; i < num_attr; i++) {
4277 struct opa_aggregate *agg;
4278 size_t agg_data_len;
4279 size_t agg_size;
4280 u32 am;
4281
4282 agg = (struct opa_aggregate *)next_smp;
4283 agg_data_len = (be16_to_cpu(agg->err_reqlength) & 0x007f) * 8;
4284 agg_size = sizeof(*agg) + agg_data_len;
4285 am = be32_to_cpu(agg->attr_mod);
4286
4287 *resp_len += agg_size;
4288
4289 if (next_smp + agg_size > ((u8 *)smp) + sizeof(*smp)) {
4290 smp->status |= IB_SMP_INVALID_FIELD;
4291 return reply((struct ib_mad_hdr *)smp);
4292 }
4293
4294 (void)subn_set_opa_sma(agg->attr_id, smp, am, agg->data,
4295 ibdev, port, NULL, (u32)agg_data_len,
4296 local_mad);
4297
4298 if (smp->status & IB_SMP_INVALID_FIELD)
4299 break;
4300 if (smp->status & ~IB_SMP_DIRECTION) {
4301 set_aggr_error(agg);
4302 return reply((struct ib_mad_hdr *)smp);
4303 }
4304 next_smp += agg_size;
4305 }
4306
4307 return reply((struct ib_mad_hdr *)smp);
4308 }
4309
4310 /*
4311 * OPAv1 specifies that, on the transition to link up, these counters
4312 * are cleared:
4313 * PortRcvErrors [*]
4314 * LinkErrorRecovery
4315 * LocalLinkIntegrityErrors
4316 * ExcessiveBufferOverruns [*]
4317 *
4318 * [*] Error info associated with these counters is retained, but the
4319 * error info status is reset to 0.
4320 */
clear_linkup_counters(struct hfi1_devdata * dd)4321 void clear_linkup_counters(struct hfi1_devdata *dd)
4322 {
4323 /* PortRcvErrors */
4324 write_dev_cntr(dd, C_DC_RCV_ERR, CNTR_INVALID_VL, 0);
4325 dd->err_info_rcvport.status_and_code &= ~OPA_EI_STATUS_SMASK;
4326 /* LinkErrorRecovery */
4327 write_dev_cntr(dd, C_DC_SEQ_CRC_CNT, CNTR_INVALID_VL, 0);
4328 write_dev_cntr(dd, C_DC_REINIT_FROM_PEER_CNT, CNTR_INVALID_VL, 0);
4329 /* LocalLinkIntegrityErrors */
4330 write_dev_cntr(dd, C_DC_RX_REPLAY, CNTR_INVALID_VL, 0);
4331 /* ExcessiveBufferOverruns */
4332 write_dev_cntr(dd, C_RCV_OVF, CNTR_INVALID_VL, 0);
4333 dd->rcv_ovfl_cnt = 0;
4334 dd->err_info_xmit_constraint.status &= ~OPA_EI_STATUS_SMASK;
4335 }
4336
is_full_mgmt_pkey_in_table(struct hfi1_ibport * ibp)4337 static int is_full_mgmt_pkey_in_table(struct hfi1_ibport *ibp)
4338 {
4339 unsigned int i;
4340 struct hfi1_pportdata *ppd = ppd_from_ibp(ibp);
4341
4342 for (i = 0; i < ARRAY_SIZE(ppd->pkeys); ++i)
4343 if (ppd->pkeys[i] == FULL_MGMT_P_KEY)
4344 return 1;
4345
4346 return 0;
4347 }
4348
4349 /*
4350 * is_local_mad() returns 1 if 'mad' is sent from, and destined to the
4351 * local node, 0 otherwise.
4352 */
is_local_mad(struct hfi1_ibport * ibp,const struct opa_mad * mad,const struct ib_wc * in_wc)4353 static int is_local_mad(struct hfi1_ibport *ibp, const struct opa_mad *mad,
4354 const struct ib_wc *in_wc)
4355 {
4356 struct hfi1_pportdata *ppd = ppd_from_ibp(ibp);
4357 const struct opa_smp *smp = (const struct opa_smp *)mad;
4358
4359 if (smp->mgmt_class == IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE) {
4360 return (smp->hop_cnt == 0 &&
4361 smp->route.dr.dr_slid == OPA_LID_PERMISSIVE &&
4362 smp->route.dr.dr_dlid == OPA_LID_PERMISSIVE);
4363 }
4364
4365 return (in_wc->slid == ppd->lid);
4366 }
4367
4368 /*
4369 * opa_local_smp_check() should only be called on MADs for which
4370 * is_local_mad() returns true. It applies the SMP checks that are
4371 * specific to SMPs which are sent from, and destined to this node.
4372 * opa_local_smp_check() returns 0 if the SMP passes its checks, 1
4373 * otherwise.
4374 *
4375 * SMPs which arrive from other nodes are instead checked by
4376 * opa_smp_check().
4377 */
opa_local_smp_check(struct hfi1_ibport * ibp,const struct ib_wc * in_wc)4378 static int opa_local_smp_check(struct hfi1_ibport *ibp,
4379 const struct ib_wc *in_wc)
4380 {
4381 struct hfi1_pportdata *ppd = ppd_from_ibp(ibp);
4382 u16 pkey;
4383
4384 if (in_wc->pkey_index >= ARRAY_SIZE(ppd->pkeys))
4385 return 1;
4386
4387 pkey = ppd->pkeys[in_wc->pkey_index];
4388 /*
4389 * We need to do the "node-local" checks specified in OPAv1,
4390 * rev 0.90, section 9.10.26, which are:
4391 * - pkey is 0x7fff, or 0xffff
4392 * - Source QPN == 0 || Destination QPN == 0
4393 * - the MAD header's management class is either
4394 * IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE or
4395 * IB_MGMT_CLASS_SUBN_LID_ROUTED
4396 * - SLID != 0
4397 *
4398 * However, we know (and so don't need to check again) that,
4399 * for local SMPs, the MAD stack passes MADs with:
4400 * - Source QPN of 0
4401 * - MAD mgmt_class is IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE
4402 * - SLID is either: OPA_LID_PERMISSIVE (0xFFFFFFFF), or
4403 * our own port's lid
4404 *
4405 */
4406 if (pkey == LIM_MGMT_P_KEY || pkey == FULL_MGMT_P_KEY)
4407 return 0;
4408 ingress_pkey_table_fail(ppd, pkey, in_wc->slid);
4409 return 1;
4410 }
4411
4412 /**
4413 * hfi1_pkey_validation_pma - It validates PKEYs for incoming PMA MAD packets.
4414 * @ibp: IB port data
4415 * @in_mad: MAD packet with header and data
4416 * @in_wc: Work completion data such as source LID, port number, etc.
4417 *
4418 * These are all the possible logic rules for validating a pkey:
4419 *
4420 * a) If pkey neither FULL_MGMT_P_KEY nor LIM_MGMT_P_KEY,
4421 * and NOT self-originated packet:
4422 * Drop MAD packet as it should always be part of the
4423 * management partition unless it's a self-originated packet.
4424 *
4425 * b) If pkey_index -> FULL_MGMT_P_KEY, and LIM_MGMT_P_KEY in pkey table:
4426 * The packet is coming from a management node and the receiving node
4427 * is also a management node, so it is safe for the packet to go through.
4428 *
4429 * c) If pkey_index -> FULL_MGMT_P_KEY, and LIM_MGMT_P_KEY is NOT in pkey table:
4430 * Drop the packet as LIM_MGMT_P_KEY should always be in the pkey table.
4431 * It could be an FM misconfiguration.
4432 *
4433 * d) If pkey_index -> LIM_MGMT_P_KEY and FULL_MGMT_P_KEY is NOT in pkey table:
4434 * It is safe for the packet to go through since a non-management node is
4435 * talking to another non-management node.
4436 *
4437 * e) If pkey_index -> LIM_MGMT_P_KEY and FULL_MGMT_P_KEY in pkey table:
4438 * Drop the packet because a non-management node is talking to a
4439 * management node, and it could be an attack.
4440 *
4441 * For the implementation, these rules can be simplied to only checking
4442 * for (a) and (e). There's no need to check for rule (b) as
4443 * the packet doesn't need to be dropped. Rule (c) is not possible in
4444 * the driver as LIM_MGMT_P_KEY is always in the pkey table.
4445 *
4446 * Return:
4447 * 0 - pkey is okay, -EINVAL it's a bad pkey
4448 */
hfi1_pkey_validation_pma(struct hfi1_ibport * ibp,const struct opa_mad * in_mad,const struct ib_wc * in_wc)4449 static int hfi1_pkey_validation_pma(struct hfi1_ibport *ibp,
4450 const struct opa_mad *in_mad,
4451 const struct ib_wc *in_wc)
4452 {
4453 u16 pkey_value = hfi1_lookup_pkey_value(ibp, in_wc->pkey_index);
4454
4455 /* Rule (a) from above */
4456 if (!is_local_mad(ibp, in_mad, in_wc) &&
4457 pkey_value != LIM_MGMT_P_KEY &&
4458 pkey_value != FULL_MGMT_P_KEY)
4459 return -EINVAL;
4460
4461 /* Rule (e) from above */
4462 if (pkey_value == LIM_MGMT_P_KEY &&
4463 is_full_mgmt_pkey_in_table(ibp))
4464 return -EINVAL;
4465
4466 return 0;
4467 }
4468
process_subn_opa(struct ib_device * ibdev,int mad_flags,u32 port,const struct opa_mad * in_mad,struct opa_mad * out_mad,u32 * resp_len,int local_mad)4469 static int process_subn_opa(struct ib_device *ibdev, int mad_flags,
4470 u32 port, const struct opa_mad *in_mad,
4471 struct opa_mad *out_mad,
4472 u32 *resp_len, int local_mad)
4473 {
4474 struct opa_smp *smp = (struct opa_smp *)out_mad;
4475 struct hfi1_ibport *ibp = to_iport(ibdev, port);
4476 u8 *data;
4477 u32 am, data_size;
4478 __be16 attr_id;
4479 int ret;
4480
4481 *out_mad = *in_mad;
4482 data = opa_get_smp_data(smp);
4483 data_size = (u32)opa_get_smp_data_size(smp);
4484
4485 am = be32_to_cpu(smp->attr_mod);
4486 attr_id = smp->attr_id;
4487 if (smp->class_version != OPA_SM_CLASS_VERSION) {
4488 smp->status |= IB_SMP_UNSUP_VERSION;
4489 ret = reply((struct ib_mad_hdr *)smp);
4490 return ret;
4491 }
4492 ret = check_mkey(ibp, (struct ib_mad_hdr *)smp, mad_flags, smp->mkey,
4493 smp->route.dr.dr_slid, smp->route.dr.return_path,
4494 smp->hop_cnt);
4495 if (ret) {
4496 u32 port_num = be32_to_cpu(smp->attr_mod);
4497
4498 /*
4499 * If this is a get/set portinfo, we already check the
4500 * M_Key if the MAD is for another port and the M_Key
4501 * is OK on the receiving port. This check is needed
4502 * to increment the error counters when the M_Key
4503 * fails to match on *both* ports.
4504 */
4505 if (attr_id == IB_SMP_ATTR_PORT_INFO &&
4506 (smp->method == IB_MGMT_METHOD_GET ||
4507 smp->method == IB_MGMT_METHOD_SET) &&
4508 port_num && port_num <= ibdev->phys_port_cnt &&
4509 port != port_num)
4510 (void)check_mkey(to_iport(ibdev, port_num),
4511 (struct ib_mad_hdr *)smp, 0,
4512 smp->mkey, smp->route.dr.dr_slid,
4513 smp->route.dr.return_path,
4514 smp->hop_cnt);
4515 ret = IB_MAD_RESULT_FAILURE;
4516 return ret;
4517 }
4518
4519 *resp_len = opa_get_smp_header_size(smp);
4520
4521 switch (smp->method) {
4522 case IB_MGMT_METHOD_GET:
4523 switch (attr_id) {
4524 default:
4525 clear_opa_smp_data(smp);
4526 ret = subn_get_opa_sma(attr_id, smp, am, data,
4527 ibdev, port, resp_len,
4528 data_size);
4529 break;
4530 case OPA_ATTRIB_ID_AGGREGATE:
4531 ret = subn_get_opa_aggregate(smp, ibdev, port,
4532 resp_len);
4533 break;
4534 }
4535 break;
4536 case IB_MGMT_METHOD_SET:
4537 switch (attr_id) {
4538 default:
4539 ret = subn_set_opa_sma(attr_id, smp, am, data,
4540 ibdev, port, resp_len,
4541 data_size, local_mad);
4542 break;
4543 case OPA_ATTRIB_ID_AGGREGATE:
4544 ret = subn_set_opa_aggregate(smp, ibdev, port,
4545 resp_len, local_mad);
4546 break;
4547 }
4548 break;
4549 case IB_MGMT_METHOD_TRAP:
4550 case IB_MGMT_METHOD_REPORT:
4551 case IB_MGMT_METHOD_REPORT_RESP:
4552 case IB_MGMT_METHOD_GET_RESP:
4553 /*
4554 * The ib_mad module will call us to process responses
4555 * before checking for other consumers.
4556 * Just tell the caller to process it normally.
4557 */
4558 ret = IB_MAD_RESULT_SUCCESS;
4559 break;
4560 case IB_MGMT_METHOD_TRAP_REPRESS:
4561 subn_handle_opa_trap_repress(ibp, smp);
4562 /* Always successful */
4563 ret = IB_MAD_RESULT_SUCCESS;
4564 break;
4565 default:
4566 smp->status |= IB_SMP_UNSUP_METHOD;
4567 ret = reply((struct ib_mad_hdr *)smp);
4568 break;
4569 }
4570
4571 return ret;
4572 }
4573
process_subn(struct ib_device * ibdev,int mad_flags,u32 port,const struct ib_mad * in_mad,struct ib_mad * out_mad)4574 static int process_subn(struct ib_device *ibdev, int mad_flags,
4575 u32 port, const struct ib_mad *in_mad,
4576 struct ib_mad *out_mad)
4577 {
4578 struct ib_smp *smp = (struct ib_smp *)out_mad;
4579 struct hfi1_ibport *ibp = to_iport(ibdev, port);
4580 int ret;
4581
4582 *out_mad = *in_mad;
4583 if (smp->class_version != 1) {
4584 smp->status |= IB_SMP_UNSUP_VERSION;
4585 ret = reply((struct ib_mad_hdr *)smp);
4586 return ret;
4587 }
4588
4589 ret = check_mkey(ibp, (struct ib_mad_hdr *)smp, mad_flags,
4590 smp->mkey, (__force __be32)smp->dr_slid,
4591 smp->return_path, smp->hop_cnt);
4592 if (ret) {
4593 u32 port_num = be32_to_cpu(smp->attr_mod);
4594
4595 /*
4596 * If this is a get/set portinfo, we already check the
4597 * M_Key if the MAD is for another port and the M_Key
4598 * is OK on the receiving port. This check is needed
4599 * to increment the error counters when the M_Key
4600 * fails to match on *both* ports.
4601 */
4602 if (in_mad->mad_hdr.attr_id == IB_SMP_ATTR_PORT_INFO &&
4603 (smp->method == IB_MGMT_METHOD_GET ||
4604 smp->method == IB_MGMT_METHOD_SET) &&
4605 port_num && port_num <= ibdev->phys_port_cnt &&
4606 port != port_num)
4607 (void)check_mkey(to_iport(ibdev, port_num),
4608 (struct ib_mad_hdr *)smp, 0,
4609 smp->mkey,
4610 (__force __be32)smp->dr_slid,
4611 smp->return_path, smp->hop_cnt);
4612 ret = IB_MAD_RESULT_FAILURE;
4613 return ret;
4614 }
4615
4616 switch (smp->method) {
4617 case IB_MGMT_METHOD_GET:
4618 switch (smp->attr_id) {
4619 case IB_SMP_ATTR_NODE_INFO:
4620 ret = subn_get_nodeinfo(smp, ibdev, port);
4621 break;
4622 default:
4623 smp->status |= IB_SMP_UNSUP_METH_ATTR;
4624 ret = reply((struct ib_mad_hdr *)smp);
4625 break;
4626 }
4627 break;
4628 }
4629
4630 return ret;
4631 }
4632
process_perf(struct ib_device * ibdev,u32 port,const struct ib_mad * in_mad,struct ib_mad * out_mad)4633 static int process_perf(struct ib_device *ibdev, u32 port,
4634 const struct ib_mad *in_mad,
4635 struct ib_mad *out_mad)
4636 {
4637 struct ib_pma_mad *pmp = (struct ib_pma_mad *)out_mad;
4638 struct ib_class_port_info *cpi = (struct ib_class_port_info *)
4639 &pmp->data;
4640 int ret = IB_MAD_RESULT_FAILURE;
4641
4642 *out_mad = *in_mad;
4643 if (pmp->mad_hdr.class_version != 1) {
4644 pmp->mad_hdr.status |= IB_SMP_UNSUP_VERSION;
4645 ret = reply((struct ib_mad_hdr *)pmp);
4646 return ret;
4647 }
4648
4649 switch (pmp->mad_hdr.method) {
4650 case IB_MGMT_METHOD_GET:
4651 switch (pmp->mad_hdr.attr_id) {
4652 case IB_PMA_PORT_COUNTERS:
4653 ret = pma_get_ib_portcounters(pmp, ibdev, port);
4654 break;
4655 case IB_PMA_PORT_COUNTERS_EXT:
4656 ret = pma_get_ib_portcounters_ext(pmp, ibdev, port);
4657 break;
4658 case IB_PMA_CLASS_PORT_INFO:
4659 cpi->capability_mask = IB_PMA_CLASS_CAP_EXT_WIDTH;
4660 ret = reply((struct ib_mad_hdr *)pmp);
4661 break;
4662 default:
4663 pmp->mad_hdr.status |= IB_SMP_UNSUP_METH_ATTR;
4664 ret = reply((struct ib_mad_hdr *)pmp);
4665 break;
4666 }
4667 break;
4668
4669 case IB_MGMT_METHOD_SET:
4670 if (pmp->mad_hdr.attr_id) {
4671 pmp->mad_hdr.status |= IB_SMP_UNSUP_METH_ATTR;
4672 ret = reply((struct ib_mad_hdr *)pmp);
4673 }
4674 break;
4675
4676 case IB_MGMT_METHOD_TRAP:
4677 case IB_MGMT_METHOD_GET_RESP:
4678 /*
4679 * The ib_mad module will call us to process responses
4680 * before checking for other consumers.
4681 * Just tell the caller to process it normally.
4682 */
4683 ret = IB_MAD_RESULT_SUCCESS;
4684 break;
4685
4686 default:
4687 pmp->mad_hdr.status |= IB_SMP_UNSUP_METHOD;
4688 ret = reply((struct ib_mad_hdr *)pmp);
4689 break;
4690 }
4691
4692 return ret;
4693 }
4694
process_perf_opa(struct ib_device * ibdev,u32 port,const struct opa_mad * in_mad,struct opa_mad * out_mad,u32 * resp_len)4695 static int process_perf_opa(struct ib_device *ibdev, u32 port,
4696 const struct opa_mad *in_mad,
4697 struct opa_mad *out_mad, u32 *resp_len)
4698 {
4699 struct opa_pma_mad *pmp = (struct opa_pma_mad *)out_mad;
4700 int ret;
4701
4702 *out_mad = *in_mad;
4703
4704 if (pmp->mad_hdr.class_version != OPA_SM_CLASS_VERSION) {
4705 pmp->mad_hdr.status |= IB_SMP_UNSUP_VERSION;
4706 return reply((struct ib_mad_hdr *)pmp);
4707 }
4708
4709 *resp_len = sizeof(pmp->mad_hdr);
4710
4711 switch (pmp->mad_hdr.method) {
4712 case IB_MGMT_METHOD_GET:
4713 switch (pmp->mad_hdr.attr_id) {
4714 case IB_PMA_CLASS_PORT_INFO:
4715 ret = pma_get_opa_classportinfo(pmp, ibdev, resp_len);
4716 break;
4717 case OPA_PM_ATTRIB_ID_PORT_STATUS:
4718 ret = pma_get_opa_portstatus(pmp, ibdev, port,
4719 resp_len);
4720 break;
4721 case OPA_PM_ATTRIB_ID_DATA_PORT_COUNTERS:
4722 ret = pma_get_opa_datacounters(pmp, ibdev, port,
4723 resp_len);
4724 break;
4725 case OPA_PM_ATTRIB_ID_ERROR_PORT_COUNTERS:
4726 ret = pma_get_opa_porterrors(pmp, ibdev, port,
4727 resp_len);
4728 break;
4729 case OPA_PM_ATTRIB_ID_ERROR_INFO:
4730 ret = pma_get_opa_errorinfo(pmp, ibdev, port,
4731 resp_len);
4732 break;
4733 default:
4734 pmp->mad_hdr.status |= IB_SMP_UNSUP_METH_ATTR;
4735 ret = reply((struct ib_mad_hdr *)pmp);
4736 break;
4737 }
4738 break;
4739
4740 case IB_MGMT_METHOD_SET:
4741 switch (pmp->mad_hdr.attr_id) {
4742 case OPA_PM_ATTRIB_ID_CLEAR_PORT_STATUS:
4743 ret = pma_set_opa_portstatus(pmp, ibdev, port,
4744 resp_len);
4745 break;
4746 case OPA_PM_ATTRIB_ID_ERROR_INFO:
4747 ret = pma_set_opa_errorinfo(pmp, ibdev, port,
4748 resp_len);
4749 break;
4750 default:
4751 pmp->mad_hdr.status |= IB_SMP_UNSUP_METH_ATTR;
4752 ret = reply((struct ib_mad_hdr *)pmp);
4753 break;
4754 }
4755 break;
4756
4757 case IB_MGMT_METHOD_TRAP:
4758 case IB_MGMT_METHOD_GET_RESP:
4759 /*
4760 * The ib_mad module will call us to process responses
4761 * before checking for other consumers.
4762 * Just tell the caller to process it normally.
4763 */
4764 ret = IB_MAD_RESULT_SUCCESS;
4765 break;
4766
4767 default:
4768 pmp->mad_hdr.status |= IB_SMP_UNSUP_METHOD;
4769 ret = reply((struct ib_mad_hdr *)pmp);
4770 break;
4771 }
4772
4773 return ret;
4774 }
4775
hfi1_process_opa_mad(struct ib_device * ibdev,int mad_flags,u32 port,const struct ib_wc * in_wc,const struct ib_grh * in_grh,const struct opa_mad * in_mad,struct opa_mad * out_mad,size_t * out_mad_size,u16 * out_mad_pkey_index)4776 static int hfi1_process_opa_mad(struct ib_device *ibdev, int mad_flags,
4777 u32 port, const struct ib_wc *in_wc,
4778 const struct ib_grh *in_grh,
4779 const struct opa_mad *in_mad,
4780 struct opa_mad *out_mad, size_t *out_mad_size,
4781 u16 *out_mad_pkey_index)
4782 {
4783 int ret;
4784 int pkey_idx;
4785 int local_mad = 0;
4786 u32 resp_len = in_wc->byte_len - sizeof(*in_grh);
4787 struct hfi1_ibport *ibp = to_iport(ibdev, port);
4788
4789 pkey_idx = hfi1_lookup_pkey_idx(ibp, LIM_MGMT_P_KEY);
4790 if (pkey_idx < 0) {
4791 pr_warn("failed to find limited mgmt pkey, defaulting 0x%x\n",
4792 hfi1_get_pkey(ibp, 1));
4793 pkey_idx = 1;
4794 }
4795 *out_mad_pkey_index = (u16)pkey_idx;
4796
4797 switch (in_mad->mad_hdr.mgmt_class) {
4798 case IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE:
4799 case IB_MGMT_CLASS_SUBN_LID_ROUTED:
4800 local_mad = is_local_mad(ibp, in_mad, in_wc);
4801 if (local_mad) {
4802 ret = opa_local_smp_check(ibp, in_wc);
4803 if (ret)
4804 return IB_MAD_RESULT_FAILURE;
4805 }
4806 ret = process_subn_opa(ibdev, mad_flags, port, in_mad,
4807 out_mad, &resp_len, local_mad);
4808 goto bail;
4809 case IB_MGMT_CLASS_PERF_MGMT:
4810 ret = hfi1_pkey_validation_pma(ibp, in_mad, in_wc);
4811 if (ret)
4812 return IB_MAD_RESULT_FAILURE;
4813
4814 ret = process_perf_opa(ibdev, port, in_mad, out_mad, &resp_len);
4815 goto bail;
4816
4817 default:
4818 ret = IB_MAD_RESULT_SUCCESS;
4819 }
4820
4821 bail:
4822 if (ret & IB_MAD_RESULT_REPLY)
4823 *out_mad_size = round_up(resp_len, 8);
4824 else if (ret & IB_MAD_RESULT_SUCCESS)
4825 *out_mad_size = in_wc->byte_len - sizeof(struct ib_grh);
4826
4827 return ret;
4828 }
4829
hfi1_process_ib_mad(struct ib_device * ibdev,int mad_flags,u32 port,const struct ib_wc * in_wc,const struct ib_grh * in_grh,const struct ib_mad * in_mad,struct ib_mad * out_mad)4830 static int hfi1_process_ib_mad(struct ib_device *ibdev, int mad_flags, u32 port,
4831 const struct ib_wc *in_wc,
4832 const struct ib_grh *in_grh,
4833 const struct ib_mad *in_mad,
4834 struct ib_mad *out_mad)
4835 {
4836 int ret;
4837
4838 switch (in_mad->mad_hdr.mgmt_class) {
4839 case IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE:
4840 case IB_MGMT_CLASS_SUBN_LID_ROUTED:
4841 ret = process_subn(ibdev, mad_flags, port, in_mad, out_mad);
4842 break;
4843 case IB_MGMT_CLASS_PERF_MGMT:
4844 ret = process_perf(ibdev, port, in_mad, out_mad);
4845 break;
4846 default:
4847 ret = IB_MAD_RESULT_SUCCESS;
4848 break;
4849 }
4850
4851 return ret;
4852 }
4853
4854 /**
4855 * hfi1_process_mad - process an incoming MAD packet
4856 * @ibdev: the infiniband device this packet came in on
4857 * @mad_flags: MAD flags
4858 * @port: the port number this packet came in on
4859 * @in_wc: the work completion entry for this packet
4860 * @in_grh: the global route header for this packet
4861 * @in_mad: the incoming MAD
4862 * @out_mad: any outgoing MAD reply
4863 * @out_mad_size: size of the outgoing MAD reply
4864 * @out_mad_pkey_index: used to apss back the packet key index
4865 *
4866 * Returns IB_MAD_RESULT_SUCCESS if this is a MAD that we are not
4867 * interested in processing.
4868 *
4869 * Note that the verbs framework has already done the MAD sanity checks,
4870 * and hop count/pointer updating for IB_MGMT_CLASS_SUBN_DIRECTED_ROUTE
4871 * MADs.
4872 *
4873 * This is called by the ib_mad module.
4874 */
hfi1_process_mad(struct ib_device * ibdev,int mad_flags,u32 port,const struct ib_wc * in_wc,const struct ib_grh * in_grh,const struct ib_mad * in_mad,struct ib_mad * out_mad,size_t * out_mad_size,u16 * out_mad_pkey_index)4875 int hfi1_process_mad(struct ib_device *ibdev, int mad_flags, u32 port,
4876 const struct ib_wc *in_wc, const struct ib_grh *in_grh,
4877 const struct ib_mad *in_mad, struct ib_mad *out_mad,
4878 size_t *out_mad_size, u16 *out_mad_pkey_index)
4879 {
4880 switch (in_mad->mad_hdr.base_version) {
4881 case OPA_MGMT_BASE_VERSION:
4882 return hfi1_process_opa_mad(ibdev, mad_flags, port,
4883 in_wc, in_grh,
4884 (struct opa_mad *)in_mad,
4885 (struct opa_mad *)out_mad,
4886 out_mad_size,
4887 out_mad_pkey_index);
4888 case IB_MGMT_BASE_VERSION:
4889 return hfi1_process_ib_mad(ibdev, mad_flags, port, in_wc,
4890 in_grh, in_mad, out_mad);
4891 default:
4892 break;
4893 }
4894
4895 return IB_MAD_RESULT_FAILURE;
4896 }
4897