1 // SPDX-License-Identifier: GPL-2.0-only 2 /* 3 * net/sched/sch_ets.c Enhanced Transmission Selection scheduler 4 * 5 * Description 6 * ----------- 7 * 8 * The Enhanced Transmission Selection scheduler is a classful queuing 9 * discipline that merges functionality of PRIO and DRR qdiscs in one scheduler. 10 * ETS makes it easy to configure a set of strict and bandwidth-sharing bands to 11 * implement the transmission selection described in 802.1Qaz. 12 * 13 * Although ETS is technically classful, it's not possible to add and remove 14 * classes at will. Instead one specifies number of classes, how many are 15 * PRIO-like and how many DRR-like, and quanta for the latter. 16 * 17 * Algorithm 18 * --------- 19 * 20 * The strict classes, if any, are tried for traffic first: first band 0, if it 21 * has no traffic then band 1, etc. 22 * 23 * When there is no traffic in any of the strict queues, the bandwidth-sharing 24 * ones are tried next. Each band is assigned a deficit counter, initialized to 25 * "quantum" of that band. ETS maintains a list of active bandwidth-sharing 26 * bands whose qdiscs are non-empty. A packet is dequeued from the band at the 27 * head of the list if the packet size is smaller or equal to the deficit 28 * counter. If the counter is too small, it is increased by "quantum" and the 29 * scheduler moves on to the next band in the active list. 30 */ 31 32 #include <linux/module.h> 33 #include <net/gen_stats.h> 34 #include <net/netlink.h> 35 #include <net/pkt_cls.h> 36 #include <net/pkt_sched.h> 37 #include <net/sch_generic.h> 38 39 struct ets_class { 40 struct list_head alist; /* In struct ets_sched.active. */ 41 struct Qdisc *qdisc; 42 u32 quantum; 43 u32 deficit; 44 struct gnet_stats_basic_sync bstats; 45 struct gnet_stats_queue qstats; 46 }; 47 48 struct ets_sched { 49 struct list_head active; 50 struct tcf_proto __rcu *filter_list; 51 struct tcf_block *block; 52 unsigned int nbands; 53 unsigned int nstrict; 54 u8 prio2band[TC_PRIO_MAX + 1]; 55 struct ets_class classes[TCQ_ETS_MAX_BANDS]; 56 }; 57 58 static const struct nla_policy ets_policy[TCA_ETS_MAX + 1] = { 59 [TCA_ETS_NBANDS] = { .type = NLA_U8 }, 60 [TCA_ETS_NSTRICT] = { .type = NLA_U8 }, 61 [TCA_ETS_QUANTA] = { .type = NLA_NESTED }, 62 [TCA_ETS_PRIOMAP] = { .type = NLA_NESTED }, 63 }; 64 65 static const struct nla_policy ets_priomap_policy[TCA_ETS_MAX + 1] = { 66 [TCA_ETS_PRIOMAP_BAND] = { .type = NLA_U8 }, 67 }; 68 69 static const struct nla_policy ets_quanta_policy[TCA_ETS_MAX + 1] = { 70 [TCA_ETS_QUANTA_BAND] = { .type = NLA_U32 }, 71 }; 72 73 static const struct nla_policy ets_class_policy[TCA_ETS_MAX + 1] = { 74 [TCA_ETS_QUANTA_BAND] = { .type = NLA_U32 }, 75 }; 76 77 static int ets_quantum_parse(struct Qdisc *sch, const struct nlattr *attr, 78 unsigned int *quantum, 79 struct netlink_ext_ack *extack) 80 { 81 *quantum = nla_get_u32(attr); 82 if (!*quantum) { 83 NL_SET_ERR_MSG(extack, "ETS quantum cannot be zero"); 84 return -EINVAL; 85 } 86 return 0; 87 } 88 89 static struct ets_class * 90 ets_class_from_arg(struct Qdisc *sch, unsigned long arg) 91 { 92 struct ets_sched *q = qdisc_priv(sch); 93 94 return &q->classes[arg - 1]; 95 } 96 97 static u32 ets_class_id(struct Qdisc *sch, const struct ets_class *cl) 98 { 99 struct ets_sched *q = qdisc_priv(sch); 100 int band = cl - q->classes; 101 102 return TC_H_MAKE(sch->handle, band + 1); 103 } 104 105 static void ets_offload_change(struct Qdisc *sch) 106 { 107 struct net_device *dev = qdisc_dev(sch); 108 struct ets_sched *q = qdisc_priv(sch); 109 struct tc_ets_qopt_offload qopt; 110 unsigned int w_psum_prev = 0; 111 unsigned int q_psum = 0; 112 unsigned int q_sum = 0; 113 unsigned int quantum; 114 unsigned int w_psum; 115 unsigned int weight; 116 unsigned int i; 117 118 if (!tc_can_offload(dev) || !dev->netdev_ops->ndo_setup_tc) 119 return; 120 121 qopt.command = TC_ETS_REPLACE; 122 qopt.handle = sch->handle; 123 qopt.parent = sch->parent; 124 qopt.replace_params.bands = q->nbands; 125 qopt.replace_params.qstats = &sch->qstats; 126 memcpy(&qopt.replace_params.priomap, 127 q->prio2band, sizeof(q->prio2band)); 128 129 for (i = 0; i < q->nbands; i++) 130 q_sum += q->classes[i].quantum; 131 132 for (i = 0; i < q->nbands; i++) { 133 quantum = q->classes[i].quantum; 134 q_psum += quantum; 135 w_psum = quantum ? q_psum * 100 / q_sum : 0; 136 weight = w_psum - w_psum_prev; 137 w_psum_prev = w_psum; 138 139 qopt.replace_params.quanta[i] = quantum; 140 qopt.replace_params.weights[i] = weight; 141 } 142 143 dev->netdev_ops->ndo_setup_tc(dev, TC_SETUP_QDISC_ETS, &qopt); 144 } 145 146 static void ets_offload_destroy(struct Qdisc *sch) 147 { 148 struct net_device *dev = qdisc_dev(sch); 149 struct tc_ets_qopt_offload qopt; 150 151 if (!tc_can_offload(dev) || !dev->netdev_ops->ndo_setup_tc) 152 return; 153 154 qopt.command = TC_ETS_DESTROY; 155 qopt.handle = sch->handle; 156 qopt.parent = sch->parent; 157 dev->netdev_ops->ndo_setup_tc(dev, TC_SETUP_QDISC_ETS, &qopt); 158 } 159 160 static void ets_offload_graft(struct Qdisc *sch, struct Qdisc *new, 161 struct Qdisc *old, unsigned long arg, 162 struct netlink_ext_ack *extack) 163 { 164 struct net_device *dev = qdisc_dev(sch); 165 struct tc_ets_qopt_offload qopt; 166 167 qopt.command = TC_ETS_GRAFT; 168 qopt.handle = sch->handle; 169 qopt.parent = sch->parent; 170 qopt.graft_params.band = arg - 1; 171 qopt.graft_params.child_handle = new->handle; 172 173 qdisc_offload_graft_helper(dev, sch, new, old, TC_SETUP_QDISC_ETS, 174 &qopt, extack); 175 } 176 177 static int ets_offload_dump(struct Qdisc *sch) 178 { 179 struct tc_ets_qopt_offload qopt; 180 181 qopt.command = TC_ETS_STATS; 182 qopt.handle = sch->handle; 183 qopt.parent = sch->parent; 184 qopt.stats.bstats = &sch->bstats; 185 qopt.stats.qstats = &sch->qstats; 186 187 return qdisc_offload_dump_helper(sch, TC_SETUP_QDISC_ETS, &qopt); 188 } 189 190 static bool ets_class_is_strict(struct ets_sched *q, const struct ets_class *cl) 191 { 192 unsigned int band = cl - q->classes; 193 194 return band < q->nstrict; 195 } 196 197 static int ets_class_change(struct Qdisc *sch, u32 classid, u32 parentid, 198 struct nlattr **tca, unsigned long *arg, 199 struct netlink_ext_ack *extack) 200 { 201 struct ets_class *cl = ets_class_from_arg(sch, *arg); 202 struct ets_sched *q = qdisc_priv(sch); 203 struct nlattr *opt = tca[TCA_OPTIONS]; 204 struct nlattr *tb[TCA_ETS_MAX + 1]; 205 unsigned int quantum; 206 int err; 207 208 /* Classes can be added and removed only through Qdisc_ops.change 209 * interface. 210 */ 211 if (!cl) { 212 NL_SET_ERR_MSG(extack, "Fine-grained class addition and removal is not supported"); 213 return -EOPNOTSUPP; 214 } 215 216 if (!opt) { 217 NL_SET_ERR_MSG(extack, "ETS options are required for this operation"); 218 return -EINVAL; 219 } 220 221 err = nla_parse_nested(tb, TCA_ETS_MAX, opt, ets_class_policy, extack); 222 if (err < 0) 223 return err; 224 225 if (!tb[TCA_ETS_QUANTA_BAND]) 226 /* Nothing to configure. */ 227 return 0; 228 229 if (ets_class_is_strict(q, cl)) { 230 NL_SET_ERR_MSG(extack, "Strict bands do not have a configurable quantum"); 231 return -EINVAL; 232 } 233 234 err = ets_quantum_parse(sch, tb[TCA_ETS_QUANTA_BAND], &quantum, 235 extack); 236 if (err) 237 return err; 238 239 sch_tree_lock(sch); 240 cl->quantum = quantum; 241 sch_tree_unlock(sch); 242 243 ets_offload_change(sch); 244 return 0; 245 } 246 247 static int ets_class_graft(struct Qdisc *sch, unsigned long arg, 248 struct Qdisc *new, struct Qdisc **old, 249 struct netlink_ext_ack *extack) 250 { 251 struct ets_class *cl = ets_class_from_arg(sch, arg); 252 253 if (!new) { 254 new = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, 255 ets_class_id(sch, cl), NULL); 256 if (!new) 257 new = &noop_qdisc; 258 else 259 qdisc_hash_add(new, true); 260 } 261 262 *old = qdisc_replace(sch, new, &cl->qdisc); 263 ets_offload_graft(sch, new, *old, arg, extack); 264 return 0; 265 } 266 267 static struct Qdisc *ets_class_leaf(struct Qdisc *sch, unsigned long arg) 268 { 269 struct ets_class *cl = ets_class_from_arg(sch, arg); 270 271 return cl->qdisc; 272 } 273 274 static unsigned long ets_class_find(struct Qdisc *sch, u32 classid) 275 { 276 unsigned long band = TC_H_MIN(classid); 277 struct ets_sched *q = qdisc_priv(sch); 278 279 if (band - 1 >= q->nbands) 280 return 0; 281 return band; 282 } 283 284 static void ets_class_qlen_notify(struct Qdisc *sch, unsigned long arg) 285 { 286 struct ets_class *cl = ets_class_from_arg(sch, arg); 287 struct ets_sched *q = qdisc_priv(sch); 288 289 /* We get notified about zero-length child Qdiscs as well if they are 290 * offloaded. Those aren't on the active list though, so don't attempt 291 * to remove them. 292 */ 293 if (!ets_class_is_strict(q, cl) && sch->q.qlen) 294 list_del(&cl->alist); 295 } 296 297 static int ets_class_dump(struct Qdisc *sch, unsigned long arg, 298 struct sk_buff *skb, struct tcmsg *tcm) 299 { 300 struct ets_class *cl = ets_class_from_arg(sch, arg); 301 struct ets_sched *q = qdisc_priv(sch); 302 struct nlattr *nest; 303 304 tcm->tcm_parent = TC_H_ROOT; 305 tcm->tcm_handle = ets_class_id(sch, cl); 306 tcm->tcm_info = cl->qdisc->handle; 307 308 nest = nla_nest_start_noflag(skb, TCA_OPTIONS); 309 if (!nest) 310 goto nla_put_failure; 311 if (!ets_class_is_strict(q, cl)) { 312 if (nla_put_u32(skb, TCA_ETS_QUANTA_BAND, cl->quantum)) 313 goto nla_put_failure; 314 } 315 return nla_nest_end(skb, nest); 316 317 nla_put_failure: 318 nla_nest_cancel(skb, nest); 319 return -EMSGSIZE; 320 } 321 322 static int ets_class_dump_stats(struct Qdisc *sch, unsigned long arg, 323 struct gnet_dump *d) 324 { 325 struct ets_class *cl = ets_class_from_arg(sch, arg); 326 struct Qdisc *cl_q = cl->qdisc; 327 328 if (gnet_stats_copy_basic(d, NULL, &cl_q->bstats, true) < 0 || 329 qdisc_qstats_copy(d, cl_q) < 0) 330 return -1; 331 332 return 0; 333 } 334 335 static void ets_qdisc_walk(struct Qdisc *sch, struct qdisc_walker *arg) 336 { 337 struct ets_sched *q = qdisc_priv(sch); 338 int i; 339 340 if (arg->stop) 341 return; 342 343 for (i = 0; i < q->nbands; i++) { 344 if (arg->count < arg->skip) { 345 arg->count++; 346 continue; 347 } 348 if (arg->fn(sch, i + 1, arg) < 0) { 349 arg->stop = 1; 350 break; 351 } 352 arg->count++; 353 } 354 } 355 356 static struct tcf_block * 357 ets_qdisc_tcf_block(struct Qdisc *sch, unsigned long cl, 358 struct netlink_ext_ack *extack) 359 { 360 struct ets_sched *q = qdisc_priv(sch); 361 362 if (cl) { 363 NL_SET_ERR_MSG(extack, "ETS classid must be zero"); 364 return NULL; 365 } 366 367 return q->block; 368 } 369 370 static unsigned long ets_qdisc_bind_tcf(struct Qdisc *sch, unsigned long parent, 371 u32 classid) 372 { 373 return ets_class_find(sch, classid); 374 } 375 376 static void ets_qdisc_unbind_tcf(struct Qdisc *sch, unsigned long arg) 377 { 378 } 379 380 static struct ets_class *ets_classify(struct sk_buff *skb, struct Qdisc *sch, 381 int *qerr) 382 { 383 struct ets_sched *q = qdisc_priv(sch); 384 u32 band = skb->priority; 385 struct tcf_result res; 386 struct tcf_proto *fl; 387 int err; 388 389 *qerr = NET_XMIT_SUCCESS | __NET_XMIT_BYPASS; 390 if (TC_H_MAJ(skb->priority) != sch->handle) { 391 fl = rcu_dereference_bh(q->filter_list); 392 err = tcf_classify(skb, NULL, fl, &res, false); 393 #ifdef CONFIG_NET_CLS_ACT 394 switch (err) { 395 case TC_ACT_STOLEN: 396 case TC_ACT_QUEUED: 397 case TC_ACT_TRAP: 398 *qerr = NET_XMIT_SUCCESS | __NET_XMIT_STOLEN; 399 fallthrough; 400 case TC_ACT_SHOT: 401 return NULL; 402 } 403 #endif 404 if (!fl || err < 0) { 405 if (TC_H_MAJ(band)) 406 band = 0; 407 return &q->classes[q->prio2band[band & TC_PRIO_MAX]]; 408 } 409 band = res.classid; 410 } 411 band = TC_H_MIN(band) - 1; 412 if (band >= q->nbands) 413 return &q->classes[q->prio2band[0]]; 414 return &q->classes[band]; 415 } 416 417 static int ets_qdisc_enqueue(struct sk_buff *skb, struct Qdisc *sch, 418 struct sk_buff **to_free) 419 { 420 unsigned int len = qdisc_pkt_len(skb); 421 struct ets_sched *q = qdisc_priv(sch); 422 struct ets_class *cl; 423 int err = 0; 424 bool first; 425 426 cl = ets_classify(skb, sch, &err); 427 if (!cl) { 428 if (err & __NET_XMIT_BYPASS) 429 qdisc_qstats_drop(sch); 430 __qdisc_drop(skb, to_free); 431 return err; 432 } 433 434 first = !cl->qdisc->q.qlen; 435 err = qdisc_enqueue(skb, cl->qdisc, to_free); 436 if (unlikely(err != NET_XMIT_SUCCESS)) { 437 if (net_xmit_drop_count(err)) { 438 cl->qstats.drops++; 439 qdisc_qstats_drop(sch); 440 } 441 return err; 442 } 443 444 if (first && !ets_class_is_strict(q, cl)) { 445 list_add_tail(&cl->alist, &q->active); 446 cl->deficit = cl->quantum; 447 } 448 449 sch->qstats.backlog += len; 450 sch->q.qlen++; 451 return err; 452 } 453 454 static struct sk_buff * 455 ets_qdisc_dequeue_skb(struct Qdisc *sch, struct sk_buff *skb) 456 { 457 qdisc_bstats_update(sch, skb); 458 qdisc_qstats_backlog_dec(sch, skb); 459 sch->q.qlen--; 460 return skb; 461 } 462 463 static struct sk_buff *ets_qdisc_dequeue(struct Qdisc *sch) 464 { 465 struct ets_sched *q = qdisc_priv(sch); 466 struct ets_class *cl; 467 struct sk_buff *skb; 468 unsigned int band; 469 unsigned int len; 470 471 while (1) { 472 for (band = 0; band < q->nstrict; band++) { 473 cl = &q->classes[band]; 474 skb = qdisc_dequeue_peeked(cl->qdisc); 475 if (skb) 476 return ets_qdisc_dequeue_skb(sch, skb); 477 } 478 479 if (list_empty(&q->active)) 480 goto out; 481 482 cl = list_first_entry(&q->active, struct ets_class, alist); 483 skb = cl->qdisc->ops->peek(cl->qdisc); 484 if (!skb) { 485 qdisc_warn_nonwc(__func__, cl->qdisc); 486 goto out; 487 } 488 489 len = qdisc_pkt_len(skb); 490 if (len <= cl->deficit) { 491 cl->deficit -= len; 492 skb = qdisc_dequeue_peeked(cl->qdisc); 493 if (unlikely(!skb)) 494 goto out; 495 if (cl->qdisc->q.qlen == 0) 496 list_del(&cl->alist); 497 return ets_qdisc_dequeue_skb(sch, skb); 498 } 499 500 cl->deficit += cl->quantum; 501 list_move_tail(&cl->alist, &q->active); 502 } 503 out: 504 return NULL; 505 } 506 507 static int ets_qdisc_priomap_parse(struct nlattr *priomap_attr, 508 unsigned int nbands, u8 *priomap, 509 struct netlink_ext_ack *extack) 510 { 511 const struct nlattr *attr; 512 int prio = 0; 513 u8 band; 514 int rem; 515 int err; 516 517 err = __nla_validate_nested(priomap_attr, TCA_ETS_MAX, 518 ets_priomap_policy, NL_VALIDATE_STRICT, 519 extack); 520 if (err) 521 return err; 522 523 nla_for_each_nested(attr, priomap_attr, rem) { 524 switch (nla_type(attr)) { 525 case TCA_ETS_PRIOMAP_BAND: 526 if (prio > TC_PRIO_MAX) { 527 NL_SET_ERR_MSG_MOD(extack, "Too many priorities in ETS priomap"); 528 return -EINVAL; 529 } 530 band = nla_get_u8(attr); 531 if (band >= nbands) { 532 NL_SET_ERR_MSG_MOD(extack, "Invalid band number in ETS priomap"); 533 return -EINVAL; 534 } 535 priomap[prio++] = band; 536 break; 537 default: 538 WARN_ON_ONCE(1); /* Validate should have caught this. */ 539 return -EINVAL; 540 } 541 } 542 543 return 0; 544 } 545 546 static int ets_qdisc_quanta_parse(struct Qdisc *sch, struct nlattr *quanta_attr, 547 unsigned int nbands, unsigned int nstrict, 548 unsigned int *quanta, 549 struct netlink_ext_ack *extack) 550 { 551 const struct nlattr *attr; 552 int band = nstrict; 553 int rem; 554 int err; 555 556 err = __nla_validate_nested(quanta_attr, TCA_ETS_MAX, 557 ets_quanta_policy, NL_VALIDATE_STRICT, 558 extack); 559 if (err < 0) 560 return err; 561 562 nla_for_each_nested(attr, quanta_attr, rem) { 563 switch (nla_type(attr)) { 564 case TCA_ETS_QUANTA_BAND: 565 if (band >= nbands) { 566 NL_SET_ERR_MSG_MOD(extack, "ETS quanta has more values than bands"); 567 return -EINVAL; 568 } 569 err = ets_quantum_parse(sch, attr, &quanta[band++], 570 extack); 571 if (err) 572 return err; 573 break; 574 default: 575 WARN_ON_ONCE(1); /* Validate should have caught this. */ 576 return -EINVAL; 577 } 578 } 579 580 return 0; 581 } 582 583 static int ets_qdisc_change(struct Qdisc *sch, struct nlattr *opt, 584 struct netlink_ext_ack *extack) 585 { 586 unsigned int quanta[TCQ_ETS_MAX_BANDS] = {0}; 587 struct Qdisc *queues[TCQ_ETS_MAX_BANDS]; 588 struct ets_sched *q = qdisc_priv(sch); 589 struct nlattr *tb[TCA_ETS_MAX + 1]; 590 unsigned int oldbands = q->nbands; 591 u8 priomap[TC_PRIO_MAX + 1]; 592 unsigned int nstrict = 0; 593 unsigned int nbands; 594 unsigned int i; 595 int err; 596 597 if (!opt) { 598 NL_SET_ERR_MSG(extack, "ETS options are required for this operation"); 599 return -EINVAL; 600 } 601 602 err = nla_parse_nested(tb, TCA_ETS_MAX, opt, ets_policy, extack); 603 if (err < 0) 604 return err; 605 606 if (!tb[TCA_ETS_NBANDS]) { 607 NL_SET_ERR_MSG_MOD(extack, "Number of bands is a required argument"); 608 return -EINVAL; 609 } 610 nbands = nla_get_u8(tb[TCA_ETS_NBANDS]); 611 if (nbands < 1 || nbands > TCQ_ETS_MAX_BANDS) { 612 NL_SET_ERR_MSG_MOD(extack, "Invalid number of bands"); 613 return -EINVAL; 614 } 615 /* Unless overridden, traffic goes to the last band. */ 616 memset(priomap, nbands - 1, sizeof(priomap)); 617 618 if (tb[TCA_ETS_NSTRICT]) { 619 nstrict = nla_get_u8(tb[TCA_ETS_NSTRICT]); 620 if (nstrict > nbands) { 621 NL_SET_ERR_MSG_MOD(extack, "Invalid number of strict bands"); 622 return -EINVAL; 623 } 624 } 625 626 if (tb[TCA_ETS_PRIOMAP]) { 627 err = ets_qdisc_priomap_parse(tb[TCA_ETS_PRIOMAP], 628 nbands, priomap, extack); 629 if (err) 630 return err; 631 } 632 633 if (tb[TCA_ETS_QUANTA]) { 634 err = ets_qdisc_quanta_parse(sch, tb[TCA_ETS_QUANTA], 635 nbands, nstrict, quanta, extack); 636 if (err) 637 return err; 638 } 639 /* If there are more bands than strict + quanta provided, the remaining 640 * ones are ETS with quantum of MTU. Initialize the missing values here. 641 */ 642 for (i = nstrict; i < nbands; i++) { 643 if (!quanta[i]) 644 quanta[i] = psched_mtu(qdisc_dev(sch)); 645 } 646 647 /* Before commit, make sure we can allocate all new qdiscs */ 648 for (i = oldbands; i < nbands; i++) { 649 queues[i] = qdisc_create_dflt(sch->dev_queue, &pfifo_qdisc_ops, 650 ets_class_id(sch, &q->classes[i]), 651 extack); 652 if (!queues[i]) { 653 while (i > oldbands) 654 qdisc_put(queues[--i]); 655 return -ENOMEM; 656 } 657 } 658 659 sch_tree_lock(sch); 660 661 q->nbands = nbands; 662 for (i = nstrict; i < q->nstrict; i++) { 663 if (q->classes[i].qdisc->q.qlen) { 664 list_add_tail(&q->classes[i].alist, &q->active); 665 q->classes[i].deficit = quanta[i]; 666 } 667 } 668 for (i = q->nbands; i < oldbands; i++) { 669 qdisc_tree_flush_backlog(q->classes[i].qdisc); 670 if (i >= q->nstrict) 671 list_del(&q->classes[i].alist); 672 } 673 q->nstrict = nstrict; 674 memcpy(q->prio2band, priomap, sizeof(priomap)); 675 676 for (i = 0; i < q->nbands; i++) 677 q->classes[i].quantum = quanta[i]; 678 679 for (i = oldbands; i < q->nbands; i++) { 680 q->classes[i].qdisc = queues[i]; 681 if (q->classes[i].qdisc != &noop_qdisc) 682 qdisc_hash_add(q->classes[i].qdisc, true); 683 } 684 685 sch_tree_unlock(sch); 686 687 ets_offload_change(sch); 688 for (i = q->nbands; i < oldbands; i++) { 689 qdisc_put(q->classes[i].qdisc); 690 q->classes[i].qdisc = NULL; 691 q->classes[i].quantum = 0; 692 q->classes[i].deficit = 0; 693 gnet_stats_basic_sync_init(&q->classes[i].bstats); 694 memset(&q->classes[i].qstats, 0, sizeof(q->classes[i].qstats)); 695 } 696 return 0; 697 } 698 699 static int ets_qdisc_init(struct Qdisc *sch, struct nlattr *opt, 700 struct netlink_ext_ack *extack) 701 { 702 struct ets_sched *q = qdisc_priv(sch); 703 int err, i; 704 705 if (!opt) 706 return -EINVAL; 707 708 err = tcf_block_get(&q->block, &q->filter_list, sch, extack); 709 if (err) 710 return err; 711 712 INIT_LIST_HEAD(&q->active); 713 for (i = 0; i < TCQ_ETS_MAX_BANDS; i++) 714 INIT_LIST_HEAD(&q->classes[i].alist); 715 716 return ets_qdisc_change(sch, opt, extack); 717 } 718 719 static void ets_qdisc_reset(struct Qdisc *sch) 720 { 721 struct ets_sched *q = qdisc_priv(sch); 722 int band; 723 724 for (band = q->nstrict; band < q->nbands; band++) { 725 if (q->classes[band].qdisc->q.qlen) 726 list_del(&q->classes[band].alist); 727 } 728 for (band = 0; band < q->nbands; band++) 729 qdisc_reset(q->classes[band].qdisc); 730 sch->qstats.backlog = 0; 731 sch->q.qlen = 0; 732 } 733 734 static void ets_qdisc_destroy(struct Qdisc *sch) 735 { 736 struct ets_sched *q = qdisc_priv(sch); 737 int band; 738 739 ets_offload_destroy(sch); 740 tcf_block_put(q->block); 741 for (band = 0; band < q->nbands; band++) 742 qdisc_put(q->classes[band].qdisc); 743 } 744 745 static int ets_qdisc_dump(struct Qdisc *sch, struct sk_buff *skb) 746 { 747 struct ets_sched *q = qdisc_priv(sch); 748 struct nlattr *opts; 749 struct nlattr *nest; 750 int band; 751 int prio; 752 int err; 753 754 err = ets_offload_dump(sch); 755 if (err) 756 return err; 757 758 opts = nla_nest_start_noflag(skb, TCA_OPTIONS); 759 if (!opts) 760 goto nla_err; 761 762 if (nla_put_u8(skb, TCA_ETS_NBANDS, q->nbands)) 763 goto nla_err; 764 765 if (q->nstrict && 766 nla_put_u8(skb, TCA_ETS_NSTRICT, q->nstrict)) 767 goto nla_err; 768 769 if (q->nbands > q->nstrict) { 770 nest = nla_nest_start(skb, TCA_ETS_QUANTA); 771 if (!nest) 772 goto nla_err; 773 774 for (band = q->nstrict; band < q->nbands; band++) { 775 if (nla_put_u32(skb, TCA_ETS_QUANTA_BAND, 776 q->classes[band].quantum)) 777 goto nla_err; 778 } 779 780 nla_nest_end(skb, nest); 781 } 782 783 nest = nla_nest_start(skb, TCA_ETS_PRIOMAP); 784 if (!nest) 785 goto nla_err; 786 787 for (prio = 0; prio <= TC_PRIO_MAX; prio++) { 788 if (nla_put_u8(skb, TCA_ETS_PRIOMAP_BAND, q->prio2band[prio])) 789 goto nla_err; 790 } 791 792 nla_nest_end(skb, nest); 793 794 return nla_nest_end(skb, opts); 795 796 nla_err: 797 nla_nest_cancel(skb, opts); 798 return -EMSGSIZE; 799 } 800 801 static const struct Qdisc_class_ops ets_class_ops = { 802 .change = ets_class_change, 803 .graft = ets_class_graft, 804 .leaf = ets_class_leaf, 805 .find = ets_class_find, 806 .qlen_notify = ets_class_qlen_notify, 807 .dump = ets_class_dump, 808 .dump_stats = ets_class_dump_stats, 809 .walk = ets_qdisc_walk, 810 .tcf_block = ets_qdisc_tcf_block, 811 .bind_tcf = ets_qdisc_bind_tcf, 812 .unbind_tcf = ets_qdisc_unbind_tcf, 813 }; 814 815 static struct Qdisc_ops ets_qdisc_ops __read_mostly = { 816 .cl_ops = &ets_class_ops, 817 .id = "ets", 818 .priv_size = sizeof(struct ets_sched), 819 .enqueue = ets_qdisc_enqueue, 820 .dequeue = ets_qdisc_dequeue, 821 .peek = qdisc_peek_dequeued, 822 .change = ets_qdisc_change, 823 .init = ets_qdisc_init, 824 .reset = ets_qdisc_reset, 825 .destroy = ets_qdisc_destroy, 826 .dump = ets_qdisc_dump, 827 .owner = THIS_MODULE, 828 }; 829 830 static int __init ets_init(void) 831 { 832 return register_qdisc(&ets_qdisc_ops); 833 } 834 835 static void __exit ets_exit(void) 836 { 837 unregister_qdisc(&ets_qdisc_ops); 838 } 839 840 module_init(ets_init); 841 module_exit(ets_exit); 842 MODULE_LICENSE("GPL"); 843