1// SPDX-License-Identifier: GPL-2.0-only
2/*
3 * net/sched/sch_ets.c Enhanced Transmission Selection scheduler
4 *
5 * Description
6 * -----------
7 *
8 * The Enhanced Transmission Selection scheduler is a classful queuing
9 * discipline that merges functionality of PRIO and DRR qdiscs in one scheduler.
10 * ETS makes it easy to configure a set of strict and bandwidth-sharing bands to
11 * implement the transmission selection described in 802.1Qaz.
12 *
13 * Although ETS is technically classful, it's not possible to add and remove
14 * classes at will. Instead one specifies number of classes, how many are
15 * PRIO-like and how many DRR-like, and quanta for the latter.
16 *
17 * Algorithm
18 * ---------
19 *
20 * The strict classes, if any, are tried for traffic first: first band 0, if it
21 * has no traffic then band 1, etc.
22 *
23 * When there is no traffic in any of the strict queues, the bandwidth-sharing
24 * ones are tried next. Each band is assigned a deficit counter, initialized to
25 * "quantum" of that band. ETS maintains a list of active bandwidth-sharing
26 * bands whose qdiscs are non-empty. A packet is dequeued from the band at the
27 * head of the list if the packet size is smaller or equal to the deficit
28 * counter. If the counter is too small, it is increased by "quantum" and the
29 * scheduler moves on to the next band in the active list.
30 */
31
32#include <linux/module.h>
33#include <net/gen_stats.h>
34#include <net/netlink.h>
35#include <net/pkt_cls.h>
36#include <net/pkt_sched.h>
37#include <net/sch_generic.h>
38
39struct ets_class {
40 struct list_head alist; /* In struct ets_sched.active. */
41 struct Qdisc *qdisc;
42 u32 quantum;
43 u32 deficit;
44 struct gnet_stats_basic_sync bstats;
45 struct gnet_stats_queue qstats;
46};
47
48struct ets_sched {
49 struct list_head active;
50 struct tcf_proto __rcu *filter_list;
51 struct tcf_block *block;
52 unsigned int nbands;
53 unsigned int nstrict;
54 u8 prio2band[TC_PRIO_MAX + 1];
55 struct ets_class classes[TCQ_ETS_MAX_BANDS];
56};
57
58static const struct nla_policy ets_policy[TCA_ETS_MAX + 1] = {
59 [TCA_ETS_NBANDS] = { .type = NLA_U8 },
60 [TCA_ETS_NSTRICT] = { .type = NLA_U8 },
61 [TCA_ETS_QUANTA] = { .type = NLA_NESTED },
62 [TCA_ETS_PRIOMAP] = { .type = NLA_NESTED },
63};
64
65static const struct nla_policy ets_priomap_policy[TCA_ETS_MAX + 1] = {
66 [TCA_ETS_PRIOMAP_BAND] = { .type = NLA_U8 },
67};
68
69static const struct nla_policy ets_quanta_policy[TCA_ETS_MAX + 1] = {
70 [TCA_ETS_QUANTA_BAND] = { .type = NLA_U32 },
71};
72
73static const struct nla_policy ets_class_policy[TCA_ETS_MAX + 1] = {
74 [TCA_ETS_QUANTA_BAND] = { .type = NLA_U32 },
75};
76
77static int ets_quantum_parse(struct Qdisc *sch, const struct nlattr *attr,
78 unsigned int *quantum,
79 struct netlink_ext_ack *extack)
80{
81 *quantum = nla_get_u32(nla: attr);
82 if (!*quantum) {
83 NL_SET_ERR_MSG(extack, "ETS quantum cannot be zero");
84 return -EINVAL;
85 }
86 return 0;
87}
88
89static struct ets_class *
90ets_class_from_arg(struct Qdisc *sch, unsigned long arg)
91{
92 struct ets_sched *q = qdisc_priv(sch);
93
94 return &q->classes[arg - 1];
95}
96
97static u32 ets_class_id(struct Qdisc *sch, const struct ets_class *cl)
98{
99 struct ets_sched *q = qdisc_priv(sch);
100 int band = cl - q->classes;
101
102 return TC_H_MAKE(sch->handle, band + 1);
103}
104
105static void ets_offload_change(struct Qdisc *sch)
106{
107 struct net_device *dev = qdisc_dev(qdisc: sch);
108 struct ets_sched *q = qdisc_priv(sch);
109 struct tc_ets_qopt_offload qopt;
110 unsigned int w_psum_prev = 0;
111 unsigned int q_psum = 0;
112 unsigned int q_sum = 0;
113 unsigned int quantum;
114 unsigned int w_psum;
115 unsigned int weight;
116 unsigned int i;
117
118 if (!tc_can_offload(dev) || !dev->netdev_ops->ndo_setup_tc)
119 return;
120
121 qopt.command = TC_ETS_REPLACE;
122 qopt.handle = sch->handle;
123 qopt.parent = sch->parent;
124 qopt.replace_params.bands = q->nbands;
125 qopt.replace_params.qstats = &sch->qstats;
126 memcpy(&qopt.replace_params.priomap,
127 q->prio2band, sizeof(q->prio2band));
128
129 for (i = 0; i < q->nbands; i++)
130 q_sum += q->classes[i].quantum;
131
132 for (i = 0; i < q->nbands; i++) {
133 quantum = q->classes[i].quantum;
134 q_psum += quantum;
135 w_psum = quantum ? q_psum * 100 / q_sum : 0;
136 weight = w_psum - w_psum_prev;
137 w_psum_prev = w_psum;
138
139 qopt.replace_params.quanta[i] = quantum;
140 qopt.replace_params.weights[i] = weight;
141 }
142
143 dev->netdev_ops->ndo_setup_tc(dev, TC_SETUP_QDISC_ETS, &qopt);
144}
145
146static void ets_offload_destroy(struct Qdisc *sch)
147{
148 struct net_device *dev = qdisc_dev(qdisc: sch);
149 struct tc_ets_qopt_offload qopt;
150
151 if (!tc_can_offload(dev) || !dev->netdev_ops->ndo_setup_tc)
152 return;
153
154 qopt.command = TC_ETS_DESTROY;
155 qopt.handle = sch->handle;
156 qopt.parent = sch->parent;
157 dev->netdev_ops->ndo_setup_tc(dev, TC_SETUP_QDISC_ETS, &qopt);
158}
159
160static void ets_offload_graft(struct Qdisc *sch, struct Qdisc *new,
161 struct Qdisc *old, unsigned long arg,
162 struct netlink_ext_ack *extack)
163{
164 struct net_device *dev = qdisc_dev(qdisc: sch);
165 struct tc_ets_qopt_offload qopt;
166
167 qopt.command = TC_ETS_GRAFT;
168 qopt.handle = sch->handle;
169 qopt.parent = sch->parent;
170 qopt.graft_params.band = arg - 1;
171 qopt.graft_params.child_handle = new->handle;
172
173 qdisc_offload_graft_helper(dev, sch, new, old, type: TC_SETUP_QDISC_ETS,
174 type_data: &qopt, extack);
175}
176
177static int ets_offload_dump(struct Qdisc *sch)
178{
179 struct tc_ets_qopt_offload qopt;
180
181 qopt.command = TC_ETS_STATS;
182 qopt.handle = sch->handle;
183 qopt.parent = sch->parent;
184 qopt.stats.bstats = &sch->bstats;
185 qopt.stats.qstats = &sch->qstats;
186
187 return qdisc_offload_dump_helper(q: sch, type: TC_SETUP_QDISC_ETS, type_data: &qopt);
188}
189
190static bool ets_class_is_strict(struct ets_sched *q, const struct ets_class *cl)
191{
192 unsigned int band = cl - q->classes;
193
194 return band < q->nstrict;
195}
196
197static int ets_class_change(struct Qdisc *sch, u32 classid, u32 parentid,
198 struct nlattr **tca, unsigned long *arg,
199 struct netlink_ext_ack *extack)
200{
201 struct ets_class *cl = ets_class_from_arg(sch, arg: *arg);
202 struct ets_sched *q = qdisc_priv(sch);
203 struct nlattr *opt = tca[TCA_OPTIONS];
204 struct nlattr *tb[TCA_ETS_MAX + 1];
205 unsigned int quantum;
206 int err;
207
208 /* Classes can be added and removed only through Qdisc_ops.change
209 * interface.
210 */
211 if (!cl) {
212 NL_SET_ERR_MSG(extack, "Fine-grained class addition and removal is not supported");
213 return -EOPNOTSUPP;
214 }
215
216 if (!opt) {
217 NL_SET_ERR_MSG(extack, "ETS options are required for this operation");
218 return -EINVAL;
219 }
220
221 err = nla_parse_nested(tb, TCA_ETS_MAX, nla: opt, policy: ets_class_policy, extack);
222 if (err < 0)
223 return err;
224
225 if (!tb[TCA_ETS_QUANTA_BAND])
226 /* Nothing to configure. */
227 return 0;
228
229 if (ets_class_is_strict(q, cl)) {
230 NL_SET_ERR_MSG(extack, "Strict bands do not have a configurable quantum");
231 return -EINVAL;
232 }
233
234 err = ets_quantum_parse(sch, attr: tb[TCA_ETS_QUANTA_BAND], quantum: &quantum,
235 extack);
236 if (err)
237 return err;
238
239 sch_tree_lock(q: sch);
240 cl->quantum = quantum;
241 sch_tree_unlock(q: sch);
242
243 ets_offload_change(sch);
244 return 0;
245}
246
247static int ets_class_graft(struct Qdisc *sch, unsigned long arg,
248 struct Qdisc *new, struct Qdisc **old,
249 struct netlink_ext_ack *extack)
250{
251 struct ets_class *cl = ets_class_from_arg(sch, arg);
252
253 if (!new) {
254 new = qdisc_create_dflt(dev_queue: sch->dev_queue, ops: &pfifo_qdisc_ops,
255 parentid: ets_class_id(sch, cl), NULL);
256 if (!new)
257 new = &noop_qdisc;
258 else
259 qdisc_hash_add(q: new, invisible: true);
260 }
261
262 *old = qdisc_replace(sch, new, pold: &cl->qdisc);
263 ets_offload_graft(sch, new, old: *old, arg, extack);
264 return 0;
265}
266
267static struct Qdisc *ets_class_leaf(struct Qdisc *sch, unsigned long arg)
268{
269 struct ets_class *cl = ets_class_from_arg(sch, arg);
270
271 return cl->qdisc;
272}
273
274static unsigned long ets_class_find(struct Qdisc *sch, u32 classid)
275{
276 unsigned long band = TC_H_MIN(classid);
277 struct ets_sched *q = qdisc_priv(sch);
278
279 if (band - 1 >= q->nbands)
280 return 0;
281 return band;
282}
283
284static void ets_class_qlen_notify(struct Qdisc *sch, unsigned long arg)
285{
286 struct ets_class *cl = ets_class_from_arg(sch, arg);
287 struct ets_sched *q = qdisc_priv(sch);
288
289 /* We get notified about zero-length child Qdiscs as well if they are
290 * offloaded. Those aren't on the active list though, so don't attempt
291 * to remove them.
292 */
293 if (!ets_class_is_strict(q, cl) && sch->q.qlen)
294 list_del(entry: &cl->alist);
295}
296
297static int ets_class_dump(struct Qdisc *sch, unsigned long arg,
298 struct sk_buff *skb, struct tcmsg *tcm)
299{
300 struct ets_class *cl = ets_class_from_arg(sch, arg);
301 struct ets_sched *q = qdisc_priv(sch);
302 struct nlattr *nest;
303
304 tcm->tcm_parent = TC_H_ROOT;
305 tcm->tcm_handle = ets_class_id(sch, cl);
306 tcm->tcm_info = cl->qdisc->handle;
307
308 nest = nla_nest_start_noflag(skb, attrtype: TCA_OPTIONS);
309 if (!nest)
310 goto nla_put_failure;
311 if (!ets_class_is_strict(q, cl)) {
312 if (nla_put_u32(skb, attrtype: TCA_ETS_QUANTA_BAND, value: cl->quantum))
313 goto nla_put_failure;
314 }
315 return nla_nest_end(skb, start: nest);
316
317nla_put_failure:
318 nla_nest_cancel(skb, start: nest);
319 return -EMSGSIZE;
320}
321
322static int ets_class_dump_stats(struct Qdisc *sch, unsigned long arg,
323 struct gnet_dump *d)
324{
325 struct ets_class *cl = ets_class_from_arg(sch, arg);
326 struct Qdisc *cl_q = cl->qdisc;
327
328 if (gnet_stats_copy_basic(d, NULL, b: &cl_q->bstats, running: true) < 0 ||
329 qdisc_qstats_copy(d, sch: cl_q) < 0)
330 return -1;
331
332 return 0;
333}
334
335static void ets_qdisc_walk(struct Qdisc *sch, struct qdisc_walker *arg)
336{
337 struct ets_sched *q = qdisc_priv(sch);
338 int i;
339
340 if (arg->stop)
341 return;
342
343 for (i = 0; i < q->nbands; i++) {
344 if (!tc_qdisc_stats_dump(sch, cl: i + 1, arg))
345 break;
346 }
347}
348
349static struct tcf_block *
350ets_qdisc_tcf_block(struct Qdisc *sch, unsigned long cl,
351 struct netlink_ext_ack *extack)
352{
353 struct ets_sched *q = qdisc_priv(sch);
354
355 if (cl) {
356 NL_SET_ERR_MSG(extack, "ETS classid must be zero");
357 return NULL;
358 }
359
360 return q->block;
361}
362
363static unsigned long ets_qdisc_bind_tcf(struct Qdisc *sch, unsigned long parent,
364 u32 classid)
365{
366 return ets_class_find(sch, classid);
367}
368
369static void ets_qdisc_unbind_tcf(struct Qdisc *sch, unsigned long arg)
370{
371}
372
373static struct ets_class *ets_classify(struct sk_buff *skb, struct Qdisc *sch,
374 int *qerr)
375{
376 struct ets_sched *q = qdisc_priv(sch);
377 u32 band = skb->priority;
378 struct tcf_result res;
379 struct tcf_proto *fl;
380 int err;
381
382 *qerr = NET_XMIT_SUCCESS | __NET_XMIT_BYPASS;
383 if (TC_H_MAJ(skb->priority) != sch->handle) {
384 fl = rcu_dereference_bh(q->filter_list);
385 err = tcf_classify(skb, NULL, tp: fl, res: &res, compat_mode: false);
386#ifdef CONFIG_NET_CLS_ACT
387 switch (err) {
388 case TC_ACT_STOLEN:
389 case TC_ACT_QUEUED:
390 case TC_ACT_TRAP:
391 *qerr = NET_XMIT_SUCCESS | __NET_XMIT_STOLEN;
392 fallthrough;
393 case TC_ACT_SHOT:
394 return NULL;
395 }
396#endif
397 if (!fl || err < 0) {
398 if (TC_H_MAJ(band))
399 band = 0;
400 return &q->classes[q->prio2band[band & TC_PRIO_MAX]];
401 }
402 band = res.classid;
403 }
404 band = TC_H_MIN(band) - 1;
405 if (band >= q->nbands)
406 return &q->classes[q->prio2band[0]];
407 return &q->classes[band];
408}
409
410static int ets_qdisc_enqueue(struct sk_buff *skb, struct Qdisc *sch,
411 struct sk_buff **to_free)
412{
413 unsigned int len = qdisc_pkt_len(skb);
414 struct ets_sched *q = qdisc_priv(sch);
415 struct ets_class *cl;
416 int err = 0;
417 bool first;
418
419 cl = ets_classify(skb, sch, qerr: &err);
420 if (!cl) {
421 if (err & __NET_XMIT_BYPASS)
422 qdisc_qstats_drop(sch);
423 __qdisc_drop(skb, to_free);
424 return err;
425 }
426
427 first = !cl->qdisc->q.qlen;
428 err = qdisc_enqueue(skb, sch: cl->qdisc, to_free);
429 if (unlikely(err != NET_XMIT_SUCCESS)) {
430 if (net_xmit_drop_count(err)) {
431 cl->qstats.drops++;
432 qdisc_qstats_drop(sch);
433 }
434 return err;
435 }
436
437 if (first && !ets_class_is_strict(q, cl)) {
438 list_add_tail(new: &cl->alist, head: &q->active);
439 cl->deficit = cl->quantum;
440 }
441
442 sch->qstats.backlog += len;
443 sch->q.qlen++;
444 return err;
445}
446
447static struct sk_buff *
448ets_qdisc_dequeue_skb(struct Qdisc *sch, struct sk_buff *skb)
449{
450 qdisc_bstats_update(sch, skb);
451 qdisc_qstats_backlog_dec(sch, skb);
452 sch->q.qlen--;
453 return skb;
454}
455
456static struct sk_buff *ets_qdisc_dequeue(struct Qdisc *sch)
457{
458 struct ets_sched *q = qdisc_priv(sch);
459 struct ets_class *cl;
460 struct sk_buff *skb;
461 unsigned int band;
462 unsigned int len;
463
464 while (1) {
465 for (band = 0; band < q->nstrict; band++) {
466 cl = &q->classes[band];
467 skb = qdisc_dequeue_peeked(sch: cl->qdisc);
468 if (skb)
469 return ets_qdisc_dequeue_skb(sch, skb);
470 }
471
472 if (list_empty(head: &q->active))
473 goto out;
474
475 cl = list_first_entry(&q->active, struct ets_class, alist);
476 skb = cl->qdisc->ops->peek(cl->qdisc);
477 if (!skb) {
478 qdisc_warn_nonwc(txt: __func__, qdisc: cl->qdisc);
479 goto out;
480 }
481
482 len = qdisc_pkt_len(skb);
483 if (len <= cl->deficit) {
484 cl->deficit -= len;
485 skb = qdisc_dequeue_peeked(sch: cl->qdisc);
486 if (unlikely(!skb))
487 goto out;
488 if (cl->qdisc->q.qlen == 0)
489 list_del(entry: &cl->alist);
490 return ets_qdisc_dequeue_skb(sch, skb);
491 }
492
493 cl->deficit += cl->quantum;
494 list_move_tail(list: &cl->alist, head: &q->active);
495 }
496out:
497 return NULL;
498}
499
500static int ets_qdisc_priomap_parse(struct nlattr *priomap_attr,
501 unsigned int nbands, u8 *priomap,
502 struct netlink_ext_ack *extack)
503{
504 const struct nlattr *attr;
505 int prio = 0;
506 u8 band;
507 int rem;
508 int err;
509
510 err = __nla_validate_nested(start: priomap_attr, TCA_ETS_MAX,
511 policy: ets_priomap_policy, NL_VALIDATE_STRICT,
512 extack);
513 if (err)
514 return err;
515
516 nla_for_each_nested(attr, priomap_attr, rem) {
517 switch (nla_type(nla: attr)) {
518 case TCA_ETS_PRIOMAP_BAND:
519 if (prio > TC_PRIO_MAX) {
520 NL_SET_ERR_MSG_MOD(extack, "Too many priorities in ETS priomap");
521 return -EINVAL;
522 }
523 band = nla_get_u8(nla: attr);
524 if (band >= nbands) {
525 NL_SET_ERR_MSG_MOD(extack, "Invalid band number in ETS priomap");
526 return -EINVAL;
527 }
528 priomap[prio++] = band;
529 break;
530 default:
531 WARN_ON_ONCE(1); /* Validate should have caught this. */
532 return -EINVAL;
533 }
534 }
535
536 return 0;
537}
538
539static int ets_qdisc_quanta_parse(struct Qdisc *sch, struct nlattr *quanta_attr,
540 unsigned int nbands, unsigned int nstrict,
541 unsigned int *quanta,
542 struct netlink_ext_ack *extack)
543{
544 const struct nlattr *attr;
545 int band = nstrict;
546 int rem;
547 int err;
548
549 err = __nla_validate_nested(start: quanta_attr, TCA_ETS_MAX,
550 policy: ets_quanta_policy, NL_VALIDATE_STRICT,
551 extack);
552 if (err < 0)
553 return err;
554
555 nla_for_each_nested(attr, quanta_attr, rem) {
556 switch (nla_type(nla: attr)) {
557 case TCA_ETS_QUANTA_BAND:
558 if (band >= nbands) {
559 NL_SET_ERR_MSG_MOD(extack, "ETS quanta has more values than bands");
560 return -EINVAL;
561 }
562 err = ets_quantum_parse(sch, attr, quantum: &quanta[band++],
563 extack);
564 if (err)
565 return err;
566 break;
567 default:
568 WARN_ON_ONCE(1); /* Validate should have caught this. */
569 return -EINVAL;
570 }
571 }
572
573 return 0;
574}
575
576static int ets_qdisc_change(struct Qdisc *sch, struct nlattr *opt,
577 struct netlink_ext_ack *extack)
578{
579 unsigned int quanta[TCQ_ETS_MAX_BANDS] = {0};
580 struct Qdisc *queues[TCQ_ETS_MAX_BANDS];
581 struct ets_sched *q = qdisc_priv(sch);
582 struct nlattr *tb[TCA_ETS_MAX + 1];
583 unsigned int oldbands = q->nbands;
584 u8 priomap[TC_PRIO_MAX + 1];
585 unsigned int nstrict = 0;
586 unsigned int nbands;
587 unsigned int i;
588 int err;
589
590 err = nla_parse_nested(tb, TCA_ETS_MAX, nla: opt, policy: ets_policy, extack);
591 if (err < 0)
592 return err;
593
594 if (!tb[TCA_ETS_NBANDS]) {
595 NL_SET_ERR_MSG_MOD(extack, "Number of bands is a required argument");
596 return -EINVAL;
597 }
598 nbands = nla_get_u8(nla: tb[TCA_ETS_NBANDS]);
599 if (nbands < 1 || nbands > TCQ_ETS_MAX_BANDS) {
600 NL_SET_ERR_MSG_MOD(extack, "Invalid number of bands");
601 return -EINVAL;
602 }
603 /* Unless overridden, traffic goes to the last band. */
604 memset(priomap, nbands - 1, sizeof(priomap));
605
606 if (tb[TCA_ETS_NSTRICT]) {
607 nstrict = nla_get_u8(nla: tb[TCA_ETS_NSTRICT]);
608 if (nstrict > nbands) {
609 NL_SET_ERR_MSG_MOD(extack, "Invalid number of strict bands");
610 return -EINVAL;
611 }
612 }
613
614 if (tb[TCA_ETS_PRIOMAP]) {
615 err = ets_qdisc_priomap_parse(priomap_attr: tb[TCA_ETS_PRIOMAP],
616 nbands, priomap, extack);
617 if (err)
618 return err;
619 }
620
621 if (tb[TCA_ETS_QUANTA]) {
622 err = ets_qdisc_quanta_parse(sch, quanta_attr: tb[TCA_ETS_QUANTA],
623 nbands, nstrict, quanta, extack);
624 if (err)
625 return err;
626 }
627 /* If there are more bands than strict + quanta provided, the remaining
628 * ones are ETS with quantum of MTU. Initialize the missing values here.
629 */
630 for (i = nstrict; i < nbands; i++) {
631 if (!quanta[i])
632 quanta[i] = psched_mtu(dev: qdisc_dev(qdisc: sch));
633 }
634
635 /* Before commit, make sure we can allocate all new qdiscs */
636 for (i = oldbands; i < nbands; i++) {
637 queues[i] = qdisc_create_dflt(dev_queue: sch->dev_queue, ops: &pfifo_qdisc_ops,
638 parentid: ets_class_id(sch, cl: &q->classes[i]),
639 extack);
640 if (!queues[i]) {
641 while (i > oldbands)
642 qdisc_put(qdisc: queues[--i]);
643 return -ENOMEM;
644 }
645 }
646
647 sch_tree_lock(q: sch);
648
649 q->nbands = nbands;
650 for (i = nstrict; i < q->nstrict; i++) {
651 if (q->classes[i].qdisc->q.qlen) {
652 list_add_tail(new: &q->classes[i].alist, head: &q->active);
653 q->classes[i].deficit = quanta[i];
654 }
655 }
656 for (i = q->nbands; i < oldbands; i++) {
657 if (i >= q->nstrict && q->classes[i].qdisc->q.qlen)
658 list_del(entry: &q->classes[i].alist);
659 qdisc_tree_flush_backlog(sch: q->classes[i].qdisc);
660 }
661 q->nstrict = nstrict;
662 memcpy(q->prio2band, priomap, sizeof(priomap));
663
664 for (i = 0; i < q->nbands; i++)
665 q->classes[i].quantum = quanta[i];
666
667 for (i = oldbands; i < q->nbands; i++) {
668 q->classes[i].qdisc = queues[i];
669 if (q->classes[i].qdisc != &noop_qdisc)
670 qdisc_hash_add(q: q->classes[i].qdisc, invisible: true);
671 }
672
673 sch_tree_unlock(q: sch);
674
675 ets_offload_change(sch);
676 for (i = q->nbands; i < oldbands; i++) {
677 qdisc_put(qdisc: q->classes[i].qdisc);
678 q->classes[i].qdisc = NULL;
679 q->classes[i].quantum = 0;
680 q->classes[i].deficit = 0;
681 gnet_stats_basic_sync_init(b: &q->classes[i].bstats);
682 memset(&q->classes[i].qstats, 0, sizeof(q->classes[i].qstats));
683 }
684 return 0;
685}
686
687static int ets_qdisc_init(struct Qdisc *sch, struct nlattr *opt,
688 struct netlink_ext_ack *extack)
689{
690 struct ets_sched *q = qdisc_priv(sch);
691 int err, i;
692
693 if (!opt)
694 return -EINVAL;
695
696 err = tcf_block_get(p_block: &q->block, p_filter_chain: &q->filter_list, q: sch, extack);
697 if (err)
698 return err;
699
700 INIT_LIST_HEAD(list: &q->active);
701 for (i = 0; i < TCQ_ETS_MAX_BANDS; i++)
702 INIT_LIST_HEAD(list: &q->classes[i].alist);
703
704 return ets_qdisc_change(sch, opt, extack);
705}
706
707static void ets_qdisc_reset(struct Qdisc *sch)
708{
709 struct ets_sched *q = qdisc_priv(sch);
710 int band;
711
712 for (band = q->nstrict; band < q->nbands; band++) {
713 if (q->classes[band].qdisc->q.qlen)
714 list_del(entry: &q->classes[band].alist);
715 }
716 for (band = 0; band < q->nbands; band++)
717 qdisc_reset(qdisc: q->classes[band].qdisc);
718}
719
720static void ets_qdisc_destroy(struct Qdisc *sch)
721{
722 struct ets_sched *q = qdisc_priv(sch);
723 int band;
724
725 ets_offload_destroy(sch);
726 tcf_block_put(block: q->block);
727 for (band = 0; band < q->nbands; band++)
728 qdisc_put(qdisc: q->classes[band].qdisc);
729}
730
731static int ets_qdisc_dump(struct Qdisc *sch, struct sk_buff *skb)
732{
733 struct ets_sched *q = qdisc_priv(sch);
734 struct nlattr *opts;
735 struct nlattr *nest;
736 int band;
737 int prio;
738 int err;
739
740 err = ets_offload_dump(sch);
741 if (err)
742 return err;
743
744 opts = nla_nest_start_noflag(skb, attrtype: TCA_OPTIONS);
745 if (!opts)
746 goto nla_err;
747
748 if (nla_put_u8(skb, attrtype: TCA_ETS_NBANDS, value: q->nbands))
749 goto nla_err;
750
751 if (q->nstrict &&
752 nla_put_u8(skb, attrtype: TCA_ETS_NSTRICT, value: q->nstrict))
753 goto nla_err;
754
755 if (q->nbands > q->nstrict) {
756 nest = nla_nest_start(skb, attrtype: TCA_ETS_QUANTA);
757 if (!nest)
758 goto nla_err;
759
760 for (band = q->nstrict; band < q->nbands; band++) {
761 if (nla_put_u32(skb, attrtype: TCA_ETS_QUANTA_BAND,
762 value: q->classes[band].quantum))
763 goto nla_err;
764 }
765
766 nla_nest_end(skb, start: nest);
767 }
768
769 nest = nla_nest_start(skb, attrtype: TCA_ETS_PRIOMAP);
770 if (!nest)
771 goto nla_err;
772
773 for (prio = 0; prio <= TC_PRIO_MAX; prio++) {
774 if (nla_put_u8(skb, attrtype: TCA_ETS_PRIOMAP_BAND, value: q->prio2band[prio]))
775 goto nla_err;
776 }
777
778 nla_nest_end(skb, start: nest);
779
780 return nla_nest_end(skb, start: opts);
781
782nla_err:
783 nla_nest_cancel(skb, start: opts);
784 return -EMSGSIZE;
785}
786
787static const struct Qdisc_class_ops ets_class_ops = {
788 .change = ets_class_change,
789 .graft = ets_class_graft,
790 .leaf = ets_class_leaf,
791 .find = ets_class_find,
792 .qlen_notify = ets_class_qlen_notify,
793 .dump = ets_class_dump,
794 .dump_stats = ets_class_dump_stats,
795 .walk = ets_qdisc_walk,
796 .tcf_block = ets_qdisc_tcf_block,
797 .bind_tcf = ets_qdisc_bind_tcf,
798 .unbind_tcf = ets_qdisc_unbind_tcf,
799};
800
801static struct Qdisc_ops ets_qdisc_ops __read_mostly = {
802 .cl_ops = &ets_class_ops,
803 .id = "ets",
804 .priv_size = sizeof(struct ets_sched),
805 .enqueue = ets_qdisc_enqueue,
806 .dequeue = ets_qdisc_dequeue,
807 .peek = qdisc_peek_dequeued,
808 .change = ets_qdisc_change,
809 .init = ets_qdisc_init,
810 .reset = ets_qdisc_reset,
811 .destroy = ets_qdisc_destroy,
812 .dump = ets_qdisc_dump,
813 .owner = THIS_MODULE,
814};
815MODULE_ALIAS_NET_SCH("ets");
816
817static int __init ets_init(void)
818{
819 return register_qdisc(qops: &ets_qdisc_ops);
820}
821
822static void __exit ets_exit(void)
823{
824 unregister_qdisc(qops: &ets_qdisc_ops);
825}
826
827module_init(ets_init);
828module_exit(ets_exit);
829MODULE_LICENSE("GPL");
830MODULE_DESCRIPTION("Enhanced Transmission Selection(ETS) scheduler");
831

source code of linux/net/sched/sch_ets.c