1 // SPDX-License-Identifier: (GPL-2.0 OR MIT) 2 /* Copyright 2020-2021 NXP 3 */ 4 #include <net/devlink.h> 5 #include "ocelot.h" 6 7 /* The queue system tracks four resource consumptions: 8 * Resource 0: Memory tracked per source port 9 * Resource 1: Frame references tracked per source port 10 * Resource 2: Memory tracked per destination port 11 * Resource 3: Frame references tracked per destination port 12 */ 13 #define OCELOT_RESOURCE_SZ 256 14 #define OCELOT_NUM_RESOURCES 4 15 16 #define BUF_xxxx_I (0 * OCELOT_RESOURCE_SZ) 17 #define REF_xxxx_I (1 * OCELOT_RESOURCE_SZ) 18 #define BUF_xxxx_E (2 * OCELOT_RESOURCE_SZ) 19 #define REF_xxxx_E (3 * OCELOT_RESOURCE_SZ) 20 21 /* For each resource type there are 4 types of watermarks: 22 * Q_RSRV: reservation per QoS class per port 23 * PRIO_SHR: sharing watermark per QoS class across all ports 24 * P_RSRV: reservation per port 25 * COL_SHR: sharing watermark per color (drop precedence) across all ports 26 */ 27 #define xxx_Q_RSRV_x 0 28 #define xxx_PRIO_SHR_x 216 29 #define xxx_P_RSRV_x 224 30 #define xxx_COL_SHR_x 254 31 32 /* Reservation Watermarks 33 * ---------------------- 34 * 35 * For setting up the reserved areas, egress watermarks exist per port and per 36 * QoS class for both ingress and egress. 37 */ 38 39 /* Amount of packet buffer 40 * | per QoS class 41 * | | reserved 42 * | | | per egress port 43 * | | | | 44 * V V v v 45 * BUF_Q_RSRV_E 46 */ 47 #define BUF_Q_RSRV_E(port, prio) \ 48 (BUF_xxxx_E + xxx_Q_RSRV_x + OCELOT_NUM_TC * (port) + (prio)) 49 50 /* Amount of packet buffer 51 * | for all port's traffic classes 52 * | | reserved 53 * | | | per egress port 54 * | | | | 55 * V V v v 56 * BUF_P_RSRV_E 57 */ 58 #define BUF_P_RSRV_E(port) \ 59 (BUF_xxxx_E + xxx_P_RSRV_x + (port)) 60 61 /* Amount of packet buffer 62 * | per QoS class 63 * | | reserved 64 * | | | per ingress port 65 * | | | | 66 * V V v v 67 * BUF_Q_RSRV_I 68 */ 69 #define BUF_Q_RSRV_I(port, prio) \ 70 (BUF_xxxx_I + xxx_Q_RSRV_x + OCELOT_NUM_TC * (port) + (prio)) 71 72 /* Amount of packet buffer 73 * | for all port's traffic classes 74 * | | reserved 75 * | | | per ingress port 76 * | | | | 77 * V V v v 78 * BUF_P_RSRV_I 79 */ 80 #define BUF_P_RSRV_I(port) \ 81 (BUF_xxxx_I + xxx_P_RSRV_x + (port)) 82 83 /* Amount of frame references 84 * | per QoS class 85 * | | reserved 86 * | | | per egress port 87 * | | | | 88 * V V v v 89 * REF_Q_RSRV_E 90 */ 91 #define REF_Q_RSRV_E(port, prio) \ 92 (REF_xxxx_E + xxx_Q_RSRV_x + OCELOT_NUM_TC * (port) + (prio)) 93 94 /* Amount of frame references 95 * | for all port's traffic classes 96 * | | reserved 97 * | | | per egress port 98 * | | | | 99 * V V v v 100 * REF_P_RSRV_E 101 */ 102 #define REF_P_RSRV_E(port) \ 103 (REF_xxxx_E + xxx_P_RSRV_x + (port)) 104 105 /* Amount of frame references 106 * | per QoS class 107 * | | reserved 108 * | | | per ingress port 109 * | | | | 110 * V V v v 111 * REF_Q_RSRV_I 112 */ 113 #define REF_Q_RSRV_I(port, prio) \ 114 (REF_xxxx_I + xxx_Q_RSRV_x + OCELOT_NUM_TC * (port) + (prio)) 115 116 /* Amount of frame references 117 * | for all port's traffic classes 118 * | | reserved 119 * | | | per ingress port 120 * | | | | 121 * V V v v 122 * REF_P_RSRV_I 123 */ 124 #define REF_P_RSRV_I(port) \ 125 (REF_xxxx_I + xxx_P_RSRV_x + (port)) 126 127 /* Sharing Watermarks 128 * ------------------ 129 * 130 * The shared memory area is shared between all ports. 131 */ 132 133 /* Amount of buffer 134 * | per QoS class 135 * | | from the shared memory area 136 * | | | for egress traffic 137 * | | | | 138 * V V v v 139 * BUF_PRIO_SHR_E 140 */ 141 #define BUF_PRIO_SHR_E(prio) \ 142 (BUF_xxxx_E + xxx_PRIO_SHR_x + (prio)) 143 144 /* Amount of buffer 145 * | per color (drop precedence level) 146 * | | from the shared memory area 147 * | | | for egress traffic 148 * | | | | 149 * V V v v 150 * BUF_COL_SHR_E 151 */ 152 #define BUF_COL_SHR_E(dp) \ 153 (BUF_xxxx_E + xxx_COL_SHR_x + (1 - (dp))) 154 155 /* Amount of buffer 156 * | per QoS class 157 * | | from the shared memory area 158 * | | | for ingress traffic 159 * | | | | 160 * V V v v 161 * BUF_PRIO_SHR_I 162 */ 163 #define BUF_PRIO_SHR_I(prio) \ 164 (BUF_xxxx_I + xxx_PRIO_SHR_x + (prio)) 165 166 /* Amount of buffer 167 * | per color (drop precedence level) 168 * | | from the shared memory area 169 * | | | for ingress traffic 170 * | | | | 171 * V V v v 172 * BUF_COL_SHR_I 173 */ 174 #define BUF_COL_SHR_I(dp) \ 175 (BUF_xxxx_I + xxx_COL_SHR_x + (1 - (dp))) 176 177 /* Amount of frame references 178 * | per QoS class 179 * | | from the shared area 180 * | | | for egress traffic 181 * | | | | 182 * V V v v 183 * REF_PRIO_SHR_E 184 */ 185 #define REF_PRIO_SHR_E(prio) \ 186 (REF_xxxx_E + xxx_PRIO_SHR_x + (prio)) 187 188 /* Amount of frame references 189 * | per color (drop precedence level) 190 * | | from the shared area 191 * | | | for egress traffic 192 * | | | | 193 * V V v v 194 * REF_COL_SHR_E 195 */ 196 #define REF_COL_SHR_E(dp) \ 197 (REF_xxxx_E + xxx_COL_SHR_x + (1 - (dp))) 198 199 /* Amount of frame references 200 * | per QoS class 201 * | | from the shared area 202 * | | | for ingress traffic 203 * | | | | 204 * V V v v 205 * REF_PRIO_SHR_I 206 */ 207 #define REF_PRIO_SHR_I(prio) \ 208 (REF_xxxx_I + xxx_PRIO_SHR_x + (prio)) 209 210 /* Amount of frame references 211 * | per color (drop precedence level) 212 * | | from the shared area 213 * | | | for ingress traffic 214 * | | | | 215 * V V v v 216 * REF_COL_SHR_I 217 */ 218 #define REF_COL_SHR_I(dp) \ 219 (REF_xxxx_I + xxx_COL_SHR_x + (1 - (dp))) 220 221 static u32 ocelot_wm_read(struct ocelot *ocelot, int index) 222 { 223 int wm = ocelot_read_gix(ocelot, QSYS_RES_CFG, index); 224 225 return ocelot->ops->wm_dec(wm); 226 } 227 228 static void ocelot_wm_write(struct ocelot *ocelot, int index, u32 val) 229 { 230 u32 wm = ocelot->ops->wm_enc(val); 231 232 ocelot_write_gix(ocelot, wm, QSYS_RES_CFG, index); 233 } 234 235 static void ocelot_wm_status(struct ocelot *ocelot, int index, u32 *inuse, 236 u32 *maxuse) 237 { 238 int res_stat = ocelot_read_gix(ocelot, QSYS_RES_STAT, index); 239 240 return ocelot->ops->wm_stat(res_stat, inuse, maxuse); 241 } 242 243 /* The hardware comes out of reset with strange defaults: the sum of all 244 * reservations for frame memory is larger than the total buffer size. 245 * One has to wonder how can the reservation watermarks still guarantee 246 * anything under congestion. 247 * Bring some sense into the hardware by changing the defaults to disable all 248 * reservations and rely only on the sharing watermark for frames with drop 249 * precedence 0. The user can still explicitly request reservations per port 250 * and per port-tc through devlink-sb. 251 */ 252 static void ocelot_disable_reservation_watermarks(struct ocelot *ocelot, 253 int port) 254 { 255 int prio; 256 257 for (prio = 0; prio < OCELOT_NUM_TC; prio++) { 258 ocelot_wm_write(ocelot, BUF_Q_RSRV_I(port, prio), 0); 259 ocelot_wm_write(ocelot, BUF_Q_RSRV_E(port, prio), 0); 260 ocelot_wm_write(ocelot, REF_Q_RSRV_I(port, prio), 0); 261 ocelot_wm_write(ocelot, REF_Q_RSRV_E(port, prio), 0); 262 } 263 264 ocelot_wm_write(ocelot, BUF_P_RSRV_I(port), 0); 265 ocelot_wm_write(ocelot, BUF_P_RSRV_E(port), 0); 266 ocelot_wm_write(ocelot, REF_P_RSRV_I(port), 0); 267 ocelot_wm_write(ocelot, REF_P_RSRV_E(port), 0); 268 } 269 270 /* We want the sharing watermarks to consume all nonreserved resources, for 271 * efficient resource utilization (a single traffic flow should be able to use 272 * up the entire buffer space and frame resources as long as there's no 273 * interference). 274 * The switch has 10 sharing watermarks per lookup: 8 per traffic class and 2 275 * per color (drop precedence). 276 * The trouble with configuring these sharing watermarks is that: 277 * (1) There's a risk that we overcommit the resources if we configure 278 * (a) all 8 per-TC sharing watermarks to the max 279 * (b) all 2 per-color sharing watermarks to the max 280 * (2) There's a risk that we undercommit the resources if we configure 281 * (a) all 8 per-TC sharing watermarks to "max / 8" 282 * (b) all 2 per-color sharing watermarks to "max / 2" 283 * So for Linux, let's just disable the sharing watermarks per traffic class 284 * (setting them to 0 will make them always exceeded), and rely only on the 285 * sharing watermark for drop priority 0. So frames with drop priority set to 1 286 * by QoS classification or policing will still be allowed, but only as long as 287 * the port and port-TC reservations are not exceeded. 288 */ 289 static void ocelot_disable_tc_sharing_watermarks(struct ocelot *ocelot) 290 { 291 int prio; 292 293 for (prio = 0; prio < OCELOT_NUM_TC; prio++) { 294 ocelot_wm_write(ocelot, BUF_PRIO_SHR_I(prio), 0); 295 ocelot_wm_write(ocelot, BUF_PRIO_SHR_E(prio), 0); 296 ocelot_wm_write(ocelot, REF_PRIO_SHR_I(prio), 0); 297 ocelot_wm_write(ocelot, REF_PRIO_SHR_E(prio), 0); 298 } 299 } 300 301 static void ocelot_get_buf_rsrv(struct ocelot *ocelot, u32 *buf_rsrv_i, 302 u32 *buf_rsrv_e) 303 { 304 int port, prio; 305 306 *buf_rsrv_i = 0; 307 *buf_rsrv_e = 0; 308 309 for (port = 0; port <= ocelot->num_phys_ports; port++) { 310 for (prio = 0; prio < OCELOT_NUM_TC; prio++) { 311 *buf_rsrv_i += ocelot_wm_read(ocelot, 312 BUF_Q_RSRV_I(port, prio)); 313 *buf_rsrv_e += ocelot_wm_read(ocelot, 314 BUF_Q_RSRV_E(port, prio)); 315 } 316 317 *buf_rsrv_i += ocelot_wm_read(ocelot, BUF_P_RSRV_I(port)); 318 *buf_rsrv_e += ocelot_wm_read(ocelot, BUF_P_RSRV_E(port)); 319 } 320 321 *buf_rsrv_i *= OCELOT_BUFFER_CELL_SZ; 322 *buf_rsrv_e *= OCELOT_BUFFER_CELL_SZ; 323 } 324 325 static void ocelot_get_ref_rsrv(struct ocelot *ocelot, u32 *ref_rsrv_i, 326 u32 *ref_rsrv_e) 327 { 328 int port, prio; 329 330 *ref_rsrv_i = 0; 331 *ref_rsrv_e = 0; 332 333 for (port = 0; port <= ocelot->num_phys_ports; port++) { 334 for (prio = 0; prio < OCELOT_NUM_TC; prio++) { 335 *ref_rsrv_i += ocelot_wm_read(ocelot, 336 REF_Q_RSRV_I(port, prio)); 337 *ref_rsrv_e += ocelot_wm_read(ocelot, 338 REF_Q_RSRV_E(port, prio)); 339 } 340 341 *ref_rsrv_i += ocelot_wm_read(ocelot, REF_P_RSRV_I(port)); 342 *ref_rsrv_e += ocelot_wm_read(ocelot, REF_P_RSRV_E(port)); 343 } 344 } 345 346 /* Calculate all reservations, then set up the sharing watermark for DP=0 to 347 * consume the remaining resources up to the pool's configured size. 348 */ 349 static void ocelot_setup_sharing_watermarks(struct ocelot *ocelot) 350 { 351 u32 buf_rsrv_i, buf_rsrv_e; 352 u32 ref_rsrv_i, ref_rsrv_e; 353 u32 buf_shr_i, buf_shr_e; 354 u32 ref_shr_i, ref_shr_e; 355 356 ocelot_get_buf_rsrv(ocelot, &buf_rsrv_i, &buf_rsrv_e); 357 ocelot_get_ref_rsrv(ocelot, &ref_rsrv_i, &ref_rsrv_e); 358 359 buf_shr_i = ocelot->pool_size[OCELOT_SB_BUF][OCELOT_SB_POOL_ING] - 360 buf_rsrv_i; 361 buf_shr_e = ocelot->pool_size[OCELOT_SB_BUF][OCELOT_SB_POOL_EGR] - 362 buf_rsrv_e; 363 ref_shr_i = ocelot->pool_size[OCELOT_SB_REF][OCELOT_SB_POOL_ING] - 364 ref_rsrv_i; 365 ref_shr_e = ocelot->pool_size[OCELOT_SB_REF][OCELOT_SB_POOL_EGR] - 366 ref_rsrv_e; 367 368 buf_shr_i /= OCELOT_BUFFER_CELL_SZ; 369 buf_shr_e /= OCELOT_BUFFER_CELL_SZ; 370 371 ocelot_wm_write(ocelot, BUF_COL_SHR_I(0), buf_shr_i); 372 ocelot_wm_write(ocelot, BUF_COL_SHR_E(0), buf_shr_e); 373 ocelot_wm_write(ocelot, REF_COL_SHR_E(0), ref_shr_e); 374 ocelot_wm_write(ocelot, REF_COL_SHR_I(0), ref_shr_i); 375 ocelot_wm_write(ocelot, BUF_COL_SHR_I(1), 0); 376 ocelot_wm_write(ocelot, BUF_COL_SHR_E(1), 0); 377 ocelot_wm_write(ocelot, REF_COL_SHR_E(1), 0); 378 ocelot_wm_write(ocelot, REF_COL_SHR_I(1), 0); 379 } 380 381 /* Ensure that all reservations can be enforced */ 382 static int ocelot_watermark_validate(struct ocelot *ocelot, 383 struct netlink_ext_ack *extack) 384 { 385 u32 buf_rsrv_i, buf_rsrv_e; 386 u32 ref_rsrv_i, ref_rsrv_e; 387 388 ocelot_get_buf_rsrv(ocelot, &buf_rsrv_i, &buf_rsrv_e); 389 ocelot_get_ref_rsrv(ocelot, &ref_rsrv_i, &ref_rsrv_e); 390 391 if (buf_rsrv_i > ocelot->pool_size[OCELOT_SB_BUF][OCELOT_SB_POOL_ING]) { 392 NL_SET_ERR_MSG_MOD(extack, 393 "Ingress frame reservations exceed pool size"); 394 return -ERANGE; 395 } 396 if (buf_rsrv_e > ocelot->pool_size[OCELOT_SB_BUF][OCELOT_SB_POOL_EGR]) { 397 NL_SET_ERR_MSG_MOD(extack, 398 "Egress frame reservations exceed pool size"); 399 return -ERANGE; 400 } 401 if (ref_rsrv_i > ocelot->pool_size[OCELOT_SB_REF][OCELOT_SB_POOL_ING]) { 402 NL_SET_ERR_MSG_MOD(extack, 403 "Ingress reference reservations exceed pool size"); 404 return -ERANGE; 405 } 406 if (ref_rsrv_e > ocelot->pool_size[OCELOT_SB_REF][OCELOT_SB_POOL_EGR]) { 407 NL_SET_ERR_MSG_MOD(extack, 408 "Egress reference reservations exceed pool size"); 409 return -ERANGE; 410 } 411 412 return 0; 413 } 414 415 /* The hardware works like this: 416 * 417 * Frame forwarding decision taken 418 * | 419 * v 420 * +--------------------+--------------------+--------------------+ 421 * | | | | 422 * v v v v 423 * Ingress memory Egress memory Ingress frame Egress frame 424 * check check reference check reference check 425 * | | | | 426 * v v v v 427 * BUF_Q_RSRV_I ok BUF_Q_RSRV_E ok REF_Q_RSRV_I ok REF_Q_RSRV_E ok 428 *(src port, prio) -+ (dst port, prio) -+ (src port, prio) -+ (dst port, prio) -+ 429 * | | | | | | | | 430 * |exceeded | |exceeded | |exceeded | |exceeded | 431 * v | v | v | v | 432 * BUF_P_RSRV_I ok| BUF_P_RSRV_E ok| REF_P_RSRV_I ok| REF_P_RSRV_E ok| 433 * (src port) ----+ (dst port) ----+ (src port) ----+ (dst port) -----+ 434 * | | | | | | | | 435 * |exceeded | |exceeded | |exceeded | |exceeded | 436 * v | v | v | v | 437 * BUF_PRIO_SHR_I ok| BUF_PRIO_SHR_E ok| REF_PRIO_SHR_I ok| REF_PRIO_SHR_E ok| 438 * (prio) ------+ (prio) ------+ (prio) ------+ (prio) -------+ 439 * | | | | | | | | 440 * |exceeded | |exceeded | |exceeded | |exceeded | 441 * v | v | v | v | 442 * BUF_COL_SHR_I ok| BUF_COL_SHR_E ok| REF_COL_SHR_I ok| REF_COL_SHR_E ok| 443 * (dp) -------+ (dp) -------+ (dp) -------+ (dp) --------+ 444 * | | | | | | | | 445 * |exceeded | |exceeded | |exceeded | |exceeded | 446 * v v v v v v v v 447 * fail success fail success fail success fail success 448 * | | | | | | | | 449 * v v v v v v v v 450 * +-----+----+ +-----+----+ +-----+----+ +-----+-----+ 451 * | | | | 452 * +-------> OR <-------+ +-------> OR <-------+ 453 * | | 454 * v v 455 * +----------------> AND <-----------------+ 456 * | 457 * v 458 * FIFO drop / accept 459 * 460 * We are modeling each of the 4 parallel lookups as a devlink-sb pool. 461 * At least one (ingress or egress) memory pool and one (ingress or egress) 462 * frame reference pool need to have resources for frame acceptance to succeed. 463 * 464 * The following watermarks are controlled explicitly through devlink-sb: 465 * BUF_Q_RSRV_I, BUF_Q_RSRV_E, REF_Q_RSRV_I, REF_Q_RSRV_E 466 * BUF_P_RSRV_I, BUF_P_RSRV_E, REF_P_RSRV_I, REF_P_RSRV_E 467 * The following watermarks are controlled implicitly through devlink-sb: 468 * BUF_COL_SHR_I, BUF_COL_SHR_E, REF_COL_SHR_I, REF_COL_SHR_E 469 * The following watermarks are unused and disabled: 470 * BUF_PRIO_SHR_I, BUF_PRIO_SHR_E, REF_PRIO_SHR_I, REF_PRIO_SHR_E 471 * 472 * This function overrides the hardware defaults with more sane ones (no 473 * reservations by default, let sharing use all resources) and disables the 474 * unused watermarks. 475 */ 476 static void ocelot_watermark_init(struct ocelot *ocelot) 477 { 478 int all_tcs = GENMASK(OCELOT_NUM_TC - 1, 0); 479 int port; 480 481 ocelot_write(ocelot, all_tcs, QSYS_RES_QOS_MODE); 482 483 for (port = 0; port <= ocelot->num_phys_ports; port++) 484 ocelot_disable_reservation_watermarks(ocelot, port); 485 486 ocelot_disable_tc_sharing_watermarks(ocelot); 487 ocelot_setup_sharing_watermarks(ocelot); 488 } 489 490 /* Watermark encode 491 * Bit 8: Unit; 0:1, 1:16 492 * Bit 7-0: Value to be multiplied with unit 493 */ 494 u16 ocelot_wm_enc(u16 value) 495 { 496 WARN_ON(value >= 16 * BIT(8)); 497 498 if (value >= BIT(8)) 499 return BIT(8) | (value / 16); 500 501 return value; 502 } 503 EXPORT_SYMBOL(ocelot_wm_enc); 504 505 u16 ocelot_wm_dec(u16 wm) 506 { 507 if (wm & BIT(8)) 508 return (wm & GENMASK(7, 0)) * 16; 509 510 return wm; 511 } 512 EXPORT_SYMBOL(ocelot_wm_dec); 513 514 void ocelot_wm_stat(u32 val, u32 *inuse, u32 *maxuse) 515 { 516 *inuse = (val & GENMASK(23, 12)) >> 12; 517 *maxuse = val & GENMASK(11, 0); 518 } 519 EXPORT_SYMBOL(ocelot_wm_stat); 520 521 /* Pool size and type are fixed up at runtime. Keeping this structure to 522 * look up the cell size multipliers. 523 */ 524 static const struct devlink_sb_pool_info ocelot_sb_pool[] = { 525 [OCELOT_SB_BUF] = { 526 .cell_size = OCELOT_BUFFER_CELL_SZ, 527 .threshold_type = DEVLINK_SB_THRESHOLD_TYPE_STATIC, 528 }, 529 [OCELOT_SB_REF] = { 530 .cell_size = 1, 531 .threshold_type = DEVLINK_SB_THRESHOLD_TYPE_STATIC, 532 }, 533 }; 534 535 /* Returns the pool size configured through ocelot_sb_pool_set */ 536 int ocelot_sb_pool_get(struct ocelot *ocelot, unsigned int sb_index, 537 u16 pool_index, 538 struct devlink_sb_pool_info *pool_info) 539 { 540 if (sb_index >= OCELOT_SB_NUM) 541 return -ENODEV; 542 if (pool_index >= OCELOT_SB_POOL_NUM) 543 return -ENODEV; 544 545 *pool_info = ocelot_sb_pool[sb_index]; 546 pool_info->size = ocelot->pool_size[sb_index][pool_index]; 547 if (pool_index) 548 pool_info->pool_type = DEVLINK_SB_POOL_TYPE_INGRESS; 549 else 550 pool_info->pool_type = DEVLINK_SB_POOL_TYPE_EGRESS; 551 552 return 0; 553 } 554 EXPORT_SYMBOL(ocelot_sb_pool_get); 555 556 /* The pool size received here configures the total amount of resources used on 557 * ingress (or on egress, depending upon the pool index). The pool size, minus 558 * the values for the port and port-tc reservations, is written into the 559 * COL_SHR(dp=0) sharing watermark. 560 */ 561 int ocelot_sb_pool_set(struct ocelot *ocelot, unsigned int sb_index, 562 u16 pool_index, u32 size, 563 enum devlink_sb_threshold_type threshold_type, 564 struct netlink_ext_ack *extack) 565 { 566 u32 old_pool_size; 567 int err; 568 569 if (sb_index >= OCELOT_SB_NUM) { 570 NL_SET_ERR_MSG_MOD(extack, 571 "Invalid sb, use 0 for buffers and 1 for frame references"); 572 return -ENODEV; 573 } 574 if (pool_index >= OCELOT_SB_POOL_NUM) { 575 NL_SET_ERR_MSG_MOD(extack, 576 "Invalid pool, use 0 for ingress and 1 for egress"); 577 return -ENODEV; 578 } 579 if (threshold_type != DEVLINK_SB_THRESHOLD_TYPE_STATIC) { 580 NL_SET_ERR_MSG_MOD(extack, 581 "Only static threshold supported"); 582 return -EOPNOTSUPP; 583 } 584 585 old_pool_size = ocelot->pool_size[sb_index][pool_index]; 586 ocelot->pool_size[sb_index][pool_index] = size; 587 588 err = ocelot_watermark_validate(ocelot, extack); 589 if (err) { 590 ocelot->pool_size[sb_index][pool_index] = old_pool_size; 591 return err; 592 } 593 594 ocelot_setup_sharing_watermarks(ocelot); 595 596 return 0; 597 } 598 EXPORT_SYMBOL(ocelot_sb_pool_set); 599 600 /* This retrieves the configuration made with ocelot_sb_port_pool_set */ 601 int ocelot_sb_port_pool_get(struct ocelot *ocelot, int port, 602 unsigned int sb_index, u16 pool_index, 603 u32 *p_threshold) 604 { 605 int wm_index; 606 607 switch (sb_index) { 608 case OCELOT_SB_BUF: 609 if (pool_index == OCELOT_SB_POOL_ING) 610 wm_index = BUF_P_RSRV_I(port); 611 else 612 wm_index = BUF_P_RSRV_E(port); 613 break; 614 case OCELOT_SB_REF: 615 if (pool_index == OCELOT_SB_POOL_ING) 616 wm_index = REF_P_RSRV_I(port); 617 else 618 wm_index = REF_P_RSRV_E(port); 619 break; 620 default: 621 return -ENODEV; 622 } 623 624 *p_threshold = ocelot_wm_read(ocelot, wm_index); 625 *p_threshold *= ocelot_sb_pool[sb_index].cell_size; 626 627 return 0; 628 } 629 EXPORT_SYMBOL(ocelot_sb_port_pool_get); 630 631 /* This configures the P_RSRV per-port reserved resource watermark */ 632 int ocelot_sb_port_pool_set(struct ocelot *ocelot, int port, 633 unsigned int sb_index, u16 pool_index, 634 u32 threshold, struct netlink_ext_ack *extack) 635 { 636 int wm_index, err; 637 u32 old_thr; 638 639 switch (sb_index) { 640 case OCELOT_SB_BUF: 641 if (pool_index == OCELOT_SB_POOL_ING) 642 wm_index = BUF_P_RSRV_I(port); 643 else 644 wm_index = BUF_P_RSRV_E(port); 645 break; 646 case OCELOT_SB_REF: 647 if (pool_index == OCELOT_SB_POOL_ING) 648 wm_index = REF_P_RSRV_I(port); 649 else 650 wm_index = REF_P_RSRV_E(port); 651 break; 652 default: 653 NL_SET_ERR_MSG_MOD(extack, "Invalid shared buffer"); 654 return -ENODEV; 655 } 656 657 threshold /= ocelot_sb_pool[sb_index].cell_size; 658 659 old_thr = ocelot_wm_read(ocelot, wm_index); 660 ocelot_wm_write(ocelot, wm_index, threshold); 661 662 err = ocelot_watermark_validate(ocelot, extack); 663 if (err) { 664 ocelot_wm_write(ocelot, wm_index, old_thr); 665 return err; 666 } 667 668 ocelot_setup_sharing_watermarks(ocelot); 669 670 return 0; 671 } 672 EXPORT_SYMBOL(ocelot_sb_port_pool_set); 673 674 /* This retrieves the configuration done by ocelot_sb_tc_pool_bind_set */ 675 int ocelot_sb_tc_pool_bind_get(struct ocelot *ocelot, int port, 676 unsigned int sb_index, u16 tc_index, 677 enum devlink_sb_pool_type pool_type, 678 u16 *p_pool_index, u32 *p_threshold) 679 { 680 int wm_index; 681 682 switch (sb_index) { 683 case OCELOT_SB_BUF: 684 if (pool_type == DEVLINK_SB_POOL_TYPE_INGRESS) 685 wm_index = BUF_Q_RSRV_I(port, tc_index); 686 else 687 wm_index = BUF_Q_RSRV_E(port, tc_index); 688 break; 689 case OCELOT_SB_REF: 690 if (pool_type == DEVLINK_SB_POOL_TYPE_INGRESS) 691 wm_index = REF_Q_RSRV_I(port, tc_index); 692 else 693 wm_index = REF_Q_RSRV_E(port, tc_index); 694 break; 695 default: 696 return -ENODEV; 697 } 698 699 *p_threshold = ocelot_wm_read(ocelot, wm_index); 700 *p_threshold *= ocelot_sb_pool[sb_index].cell_size; 701 702 if (pool_type == DEVLINK_SB_POOL_TYPE_INGRESS) 703 *p_pool_index = 0; 704 else 705 *p_pool_index = 1; 706 707 return 0; 708 } 709 EXPORT_SYMBOL(ocelot_sb_tc_pool_bind_get); 710 711 /* This configures the Q_RSRV per-port-tc reserved resource watermark */ 712 int ocelot_sb_tc_pool_bind_set(struct ocelot *ocelot, int port, 713 unsigned int sb_index, u16 tc_index, 714 enum devlink_sb_pool_type pool_type, 715 u16 pool_index, u32 threshold, 716 struct netlink_ext_ack *extack) 717 { 718 int wm_index, err; 719 u32 old_thr; 720 721 /* Paranoid check? */ 722 if (pool_index == OCELOT_SB_POOL_ING && 723 pool_type != DEVLINK_SB_POOL_TYPE_INGRESS) 724 return -EINVAL; 725 if (pool_index == OCELOT_SB_POOL_EGR && 726 pool_type != DEVLINK_SB_POOL_TYPE_EGRESS) 727 return -EINVAL; 728 729 switch (sb_index) { 730 case OCELOT_SB_BUF: 731 if (pool_type == DEVLINK_SB_POOL_TYPE_INGRESS) 732 wm_index = BUF_Q_RSRV_I(port, tc_index); 733 else 734 wm_index = BUF_Q_RSRV_E(port, tc_index); 735 break; 736 case OCELOT_SB_REF: 737 if (pool_type == DEVLINK_SB_POOL_TYPE_INGRESS) 738 wm_index = REF_Q_RSRV_I(port, tc_index); 739 else 740 wm_index = REF_Q_RSRV_E(port, tc_index); 741 break; 742 default: 743 NL_SET_ERR_MSG_MOD(extack, "Invalid shared buffer"); 744 return -ENODEV; 745 } 746 747 threshold /= ocelot_sb_pool[sb_index].cell_size; 748 749 old_thr = ocelot_wm_read(ocelot, wm_index); 750 ocelot_wm_write(ocelot, wm_index, threshold); 751 err = ocelot_watermark_validate(ocelot, extack); 752 if (err) { 753 ocelot_wm_write(ocelot, wm_index, old_thr); 754 return err; 755 } 756 757 ocelot_setup_sharing_watermarks(ocelot); 758 759 return 0; 760 } 761 EXPORT_SYMBOL(ocelot_sb_tc_pool_bind_set); 762 763 /* The hardware does not support atomic snapshots, we'll read out the 764 * occupancy registers individually and have this as just a stub. 765 */ 766 int ocelot_sb_occ_snapshot(struct ocelot *ocelot, unsigned int sb_index) 767 { 768 return 0; 769 } 770 EXPORT_SYMBOL(ocelot_sb_occ_snapshot); 771 772 /* The watermark occupancy registers are cleared upon read, 773 * so let's read them. 774 */ 775 int ocelot_sb_occ_max_clear(struct ocelot *ocelot, unsigned int sb_index) 776 { 777 u32 inuse, maxuse; 778 int port, prio; 779 780 switch (sb_index) { 781 case OCELOT_SB_BUF: 782 for (port = 0; port <= ocelot->num_phys_ports; port++) { 783 for (prio = 0; prio < OCELOT_NUM_TC; prio++) { 784 ocelot_wm_status(ocelot, BUF_Q_RSRV_I(port, prio), 785 &inuse, &maxuse); 786 ocelot_wm_status(ocelot, BUF_Q_RSRV_E(port, prio), 787 &inuse, &maxuse); 788 } 789 ocelot_wm_status(ocelot, BUF_P_RSRV_I(port), 790 &inuse, &maxuse); 791 ocelot_wm_status(ocelot, BUF_P_RSRV_E(port), 792 &inuse, &maxuse); 793 } 794 break; 795 case OCELOT_SB_REF: 796 for (port = 0; port <= ocelot->num_phys_ports; port++) { 797 for (prio = 0; prio < OCELOT_NUM_TC; prio++) { 798 ocelot_wm_status(ocelot, REF_Q_RSRV_I(port, prio), 799 &inuse, &maxuse); 800 ocelot_wm_status(ocelot, REF_Q_RSRV_E(port, prio), 801 &inuse, &maxuse); 802 } 803 ocelot_wm_status(ocelot, REF_P_RSRV_I(port), 804 &inuse, &maxuse); 805 ocelot_wm_status(ocelot, REF_P_RSRV_E(port), 806 &inuse, &maxuse); 807 } 808 break; 809 default: 810 return -ENODEV; 811 } 812 813 return 0; 814 } 815 EXPORT_SYMBOL(ocelot_sb_occ_max_clear); 816 817 /* This retrieves the watermark occupancy for per-port P_RSRV watermarks */ 818 int ocelot_sb_occ_port_pool_get(struct ocelot *ocelot, int port, 819 unsigned int sb_index, u16 pool_index, 820 u32 *p_cur, u32 *p_max) 821 { 822 int wm_index; 823 824 switch (sb_index) { 825 case OCELOT_SB_BUF: 826 if (pool_index == OCELOT_SB_POOL_ING) 827 wm_index = BUF_P_RSRV_I(port); 828 else 829 wm_index = BUF_P_RSRV_E(port); 830 break; 831 case OCELOT_SB_REF: 832 if (pool_index == OCELOT_SB_POOL_ING) 833 wm_index = REF_P_RSRV_I(port); 834 else 835 wm_index = REF_P_RSRV_E(port); 836 break; 837 default: 838 return -ENODEV; 839 } 840 841 ocelot_wm_status(ocelot, wm_index, p_cur, p_max); 842 *p_cur *= ocelot_sb_pool[sb_index].cell_size; 843 *p_max *= ocelot_sb_pool[sb_index].cell_size; 844 845 return 0; 846 } 847 EXPORT_SYMBOL(ocelot_sb_occ_port_pool_get); 848 849 /* This retrieves the watermark occupancy for per-port-tc Q_RSRV watermarks */ 850 int ocelot_sb_occ_tc_port_bind_get(struct ocelot *ocelot, int port, 851 unsigned int sb_index, u16 tc_index, 852 enum devlink_sb_pool_type pool_type, 853 u32 *p_cur, u32 *p_max) 854 { 855 int wm_index; 856 857 switch (sb_index) { 858 case OCELOT_SB_BUF: 859 if (pool_type == DEVLINK_SB_POOL_TYPE_INGRESS) 860 wm_index = BUF_Q_RSRV_I(port, tc_index); 861 else 862 wm_index = BUF_Q_RSRV_E(port, tc_index); 863 break; 864 case OCELOT_SB_REF: 865 if (pool_type == DEVLINK_SB_POOL_TYPE_INGRESS) 866 wm_index = REF_Q_RSRV_I(port, tc_index); 867 else 868 wm_index = REF_Q_RSRV_E(port, tc_index); 869 break; 870 default: 871 return -ENODEV; 872 } 873 874 ocelot_wm_status(ocelot, wm_index, p_cur, p_max); 875 *p_cur *= ocelot_sb_pool[sb_index].cell_size; 876 *p_max *= ocelot_sb_pool[sb_index].cell_size; 877 878 return 0; 879 } 880 EXPORT_SYMBOL(ocelot_sb_occ_tc_port_bind_get); 881 882 int ocelot_devlink_sb_register(struct ocelot *ocelot) 883 { 884 int err; 885 886 err = devlink_sb_register(ocelot->devlink, OCELOT_SB_BUF, 887 ocelot->packet_buffer_size, 1, 1, 888 OCELOT_NUM_TC, OCELOT_NUM_TC); 889 if (err) 890 return err; 891 892 err = devlink_sb_register(ocelot->devlink, OCELOT_SB_REF, 893 ocelot->num_frame_refs, 1, 1, 894 OCELOT_NUM_TC, OCELOT_NUM_TC); 895 if (err) { 896 devlink_sb_unregister(ocelot->devlink, OCELOT_SB_BUF); 897 return err; 898 } 899 900 ocelot->pool_size[OCELOT_SB_BUF][OCELOT_SB_POOL_ING] = ocelot->packet_buffer_size; 901 ocelot->pool_size[OCELOT_SB_BUF][OCELOT_SB_POOL_EGR] = ocelot->packet_buffer_size; 902 ocelot->pool_size[OCELOT_SB_REF][OCELOT_SB_POOL_ING] = ocelot->num_frame_refs; 903 ocelot->pool_size[OCELOT_SB_REF][OCELOT_SB_POOL_EGR] = ocelot->num_frame_refs; 904 905 ocelot_watermark_init(ocelot); 906 907 return 0; 908 } 909 EXPORT_SYMBOL(ocelot_devlink_sb_register); 910 911 void ocelot_devlink_sb_unregister(struct ocelot *ocelot) 912 { 913 devlink_sb_unregister(ocelot->devlink, OCELOT_SB_BUF); 914 devlink_sb_unregister(ocelot->devlink, OCELOT_SB_REF); 915 } 916 EXPORT_SYMBOL(ocelot_devlink_sb_unregister); 917