sch_skbprio.c 7.4 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309
  1. // SPDX-License-Identifier: GPL-2.0-or-later
  2. /*
  3. * net/sched/sch_skbprio.c SKB Priority Queue.
  4. *
  5. * Authors: Nishanth Devarajan, <[email protected]>
  6. * Cody Doucette, <[email protected]>
  7. * original idea by Michel Machado, Cody Doucette, and Qiaobin Fu
  8. */
  9. #include <linux/string.h>
  10. #include <linux/module.h>
  11. #include <linux/slab.h>
  12. #include <linux/types.h>
  13. #include <linux/kernel.h>
  14. #include <linux/errno.h>
  15. #include <linux/skbuff.h>
  16. #include <net/pkt_sched.h>
  17. #include <net/sch_generic.h>
  18. #include <net/inet_ecn.h>
  19. /* SKB Priority Queue
  20. * =================================
  21. *
  22. * Skbprio (SKB Priority Queue) is a queueing discipline that prioritizes
  23. * packets according to their skb->priority field. Under congestion,
  24. * Skbprio drops already-enqueued lower priority packets to make space
  25. * available for higher priority packets; it was conceived as a solution
  26. * for denial-of-service defenses that need to route packets with different
  27. * priorities as a mean to overcome DoS attacks.
  28. */
  29. struct skbprio_sched_data {
  30. /* Queue state. */
  31. struct sk_buff_head qdiscs[SKBPRIO_MAX_PRIORITY];
  32. struct gnet_stats_queue qstats[SKBPRIO_MAX_PRIORITY];
  33. u16 highest_prio;
  34. u16 lowest_prio;
  35. };
  36. static u16 calc_new_high_prio(const struct skbprio_sched_data *q)
  37. {
  38. int prio;
  39. for (prio = q->highest_prio - 1; prio >= q->lowest_prio; prio--) {
  40. if (!skb_queue_empty(&q->qdiscs[prio]))
  41. return prio;
  42. }
  43. /* SKB queue is empty, return 0 (default highest priority setting). */
  44. return 0;
  45. }
  46. static u16 calc_new_low_prio(const struct skbprio_sched_data *q)
  47. {
  48. int prio;
  49. for (prio = q->lowest_prio + 1; prio <= q->highest_prio; prio++) {
  50. if (!skb_queue_empty(&q->qdiscs[prio]))
  51. return prio;
  52. }
  53. /* SKB queue is empty, return SKBPRIO_MAX_PRIORITY - 1
  54. * (default lowest priority setting).
  55. */
  56. return SKBPRIO_MAX_PRIORITY - 1;
  57. }
  58. static int skbprio_enqueue(struct sk_buff *skb, struct Qdisc *sch,
  59. struct sk_buff **to_free)
  60. {
  61. const unsigned int max_priority = SKBPRIO_MAX_PRIORITY - 1;
  62. struct skbprio_sched_data *q = qdisc_priv(sch);
  63. struct sk_buff_head *qdisc;
  64. struct sk_buff_head *lp_qdisc;
  65. struct sk_buff *to_drop;
  66. u16 prio, lp;
  67. /* Obtain the priority of @skb. */
  68. prio = min(skb->priority, max_priority);
  69. qdisc = &q->qdiscs[prio];
  70. if (sch->q.qlen < sch->limit) {
  71. __skb_queue_tail(qdisc, skb);
  72. qdisc_qstats_backlog_inc(sch, skb);
  73. q->qstats[prio].backlog += qdisc_pkt_len(skb);
  74. /* Check to update highest and lowest priorities. */
  75. if (prio > q->highest_prio)
  76. q->highest_prio = prio;
  77. if (prio < q->lowest_prio)
  78. q->lowest_prio = prio;
  79. sch->q.qlen++;
  80. return NET_XMIT_SUCCESS;
  81. }
  82. /* If this packet has the lowest priority, drop it. */
  83. lp = q->lowest_prio;
  84. if (prio <= lp) {
  85. q->qstats[prio].drops++;
  86. q->qstats[prio].overlimits++;
  87. return qdisc_drop(skb, sch, to_free);
  88. }
  89. __skb_queue_tail(qdisc, skb);
  90. qdisc_qstats_backlog_inc(sch, skb);
  91. q->qstats[prio].backlog += qdisc_pkt_len(skb);
  92. /* Drop the packet at the tail of the lowest priority qdisc. */
  93. lp_qdisc = &q->qdiscs[lp];
  94. to_drop = __skb_dequeue_tail(lp_qdisc);
  95. BUG_ON(!to_drop);
  96. qdisc_qstats_backlog_dec(sch, to_drop);
  97. qdisc_drop(to_drop, sch, to_free);
  98. q->qstats[lp].backlog -= qdisc_pkt_len(to_drop);
  99. q->qstats[lp].drops++;
  100. q->qstats[lp].overlimits++;
  101. /* Check to update highest and lowest priorities. */
  102. if (skb_queue_empty(lp_qdisc)) {
  103. if (q->lowest_prio == q->highest_prio) {
  104. /* The incoming packet is the only packet in queue. */
  105. BUG_ON(sch->q.qlen != 1);
  106. q->lowest_prio = prio;
  107. q->highest_prio = prio;
  108. } else {
  109. q->lowest_prio = calc_new_low_prio(q);
  110. }
  111. }
  112. if (prio > q->highest_prio)
  113. q->highest_prio = prio;
  114. return NET_XMIT_CN;
  115. }
  116. static struct sk_buff *skbprio_dequeue(struct Qdisc *sch)
  117. {
  118. struct skbprio_sched_data *q = qdisc_priv(sch);
  119. struct sk_buff_head *hpq = &q->qdiscs[q->highest_prio];
  120. struct sk_buff *skb = __skb_dequeue(hpq);
  121. if (unlikely(!skb))
  122. return NULL;
  123. sch->q.qlen--;
  124. qdisc_qstats_backlog_dec(sch, skb);
  125. qdisc_bstats_update(sch, skb);
  126. q->qstats[q->highest_prio].backlog -= qdisc_pkt_len(skb);
  127. /* Update highest priority field. */
  128. if (skb_queue_empty(hpq)) {
  129. if (q->lowest_prio == q->highest_prio) {
  130. BUG_ON(sch->q.qlen);
  131. q->highest_prio = 0;
  132. q->lowest_prio = SKBPRIO_MAX_PRIORITY - 1;
  133. } else {
  134. q->highest_prio = calc_new_high_prio(q);
  135. }
  136. }
  137. return skb;
  138. }
  139. static int skbprio_change(struct Qdisc *sch, struct nlattr *opt,
  140. struct netlink_ext_ack *extack)
  141. {
  142. struct tc_skbprio_qopt *ctl = nla_data(opt);
  143. if (opt->nla_len != nla_attr_size(sizeof(*ctl)))
  144. return -EINVAL;
  145. sch->limit = ctl->limit;
  146. return 0;
  147. }
  148. static int skbprio_init(struct Qdisc *sch, struct nlattr *opt,
  149. struct netlink_ext_ack *extack)
  150. {
  151. struct skbprio_sched_data *q = qdisc_priv(sch);
  152. int prio;
  153. /* Initialise all queues, one for each possible priority. */
  154. for (prio = 0; prio < SKBPRIO_MAX_PRIORITY; prio++)
  155. __skb_queue_head_init(&q->qdiscs[prio]);
  156. memset(&q->qstats, 0, sizeof(q->qstats));
  157. q->highest_prio = 0;
  158. q->lowest_prio = SKBPRIO_MAX_PRIORITY - 1;
  159. sch->limit = 64;
  160. if (!opt)
  161. return 0;
  162. return skbprio_change(sch, opt, extack);
  163. }
  164. static int skbprio_dump(struct Qdisc *sch, struct sk_buff *skb)
  165. {
  166. struct tc_skbprio_qopt opt;
  167. opt.limit = sch->limit;
  168. if (nla_put(skb, TCA_OPTIONS, sizeof(opt), &opt))
  169. return -1;
  170. return skb->len;
  171. }
  172. static void skbprio_reset(struct Qdisc *sch)
  173. {
  174. struct skbprio_sched_data *q = qdisc_priv(sch);
  175. int prio;
  176. for (prio = 0; prio < SKBPRIO_MAX_PRIORITY; prio++)
  177. __skb_queue_purge(&q->qdiscs[prio]);
  178. memset(&q->qstats, 0, sizeof(q->qstats));
  179. q->highest_prio = 0;
  180. q->lowest_prio = SKBPRIO_MAX_PRIORITY - 1;
  181. }
  182. static void skbprio_destroy(struct Qdisc *sch)
  183. {
  184. struct skbprio_sched_data *q = qdisc_priv(sch);
  185. int prio;
  186. for (prio = 0; prio < SKBPRIO_MAX_PRIORITY; prio++)
  187. __skb_queue_purge(&q->qdiscs[prio]);
  188. }
  189. static struct Qdisc *skbprio_leaf(struct Qdisc *sch, unsigned long arg)
  190. {
  191. return NULL;
  192. }
  193. static unsigned long skbprio_find(struct Qdisc *sch, u32 classid)
  194. {
  195. return 0;
  196. }
  197. static int skbprio_dump_class(struct Qdisc *sch, unsigned long cl,
  198. struct sk_buff *skb, struct tcmsg *tcm)
  199. {
  200. tcm->tcm_handle |= TC_H_MIN(cl);
  201. return 0;
  202. }
  203. static int skbprio_dump_class_stats(struct Qdisc *sch, unsigned long cl,
  204. struct gnet_dump *d)
  205. {
  206. struct skbprio_sched_data *q = qdisc_priv(sch);
  207. if (gnet_stats_copy_queue(d, NULL, &q->qstats[cl - 1],
  208. q->qstats[cl - 1].qlen) < 0)
  209. return -1;
  210. return 0;
  211. }
  212. static void skbprio_walk(struct Qdisc *sch, struct qdisc_walker *arg)
  213. {
  214. unsigned int i;
  215. if (arg->stop)
  216. return;
  217. for (i = 0; i < SKBPRIO_MAX_PRIORITY; i++) {
  218. if (!tc_qdisc_stats_dump(sch, i + 1, arg))
  219. break;
  220. }
  221. }
  222. static const struct Qdisc_class_ops skbprio_class_ops = {
  223. .leaf = skbprio_leaf,
  224. .find = skbprio_find,
  225. .dump = skbprio_dump_class,
  226. .dump_stats = skbprio_dump_class_stats,
  227. .walk = skbprio_walk,
  228. };
  229. static struct Qdisc_ops skbprio_qdisc_ops __read_mostly = {
  230. .cl_ops = &skbprio_class_ops,
  231. .id = "skbprio",
  232. .priv_size = sizeof(struct skbprio_sched_data),
  233. .enqueue = skbprio_enqueue,
  234. .dequeue = skbprio_dequeue,
  235. .peek = qdisc_peek_dequeued,
  236. .init = skbprio_init,
  237. .reset = skbprio_reset,
  238. .change = skbprio_change,
  239. .dump = skbprio_dump,
  240. .destroy = skbprio_destroy,
  241. .owner = THIS_MODULE,
  242. };
  243. static int __init skbprio_module_init(void)
  244. {
  245. return register_qdisc(&skbprio_qdisc_ops);
  246. }
  247. static void __exit skbprio_module_exit(void)
  248. {
  249. unregister_qdisc(&skbprio_qdisc_ops);
  250. }
  251. module_init(skbprio_module_init)
  252. module_exit(skbprio_module_exit)
  253. MODULE_LICENSE("GPL");