Warning: This file is not a C or C++ file. It does not have highlighting.
1 | /* SPDX-License-Identifier: GPL-2.0 WITH Linux-syscall-note */ |
---|---|
2 | #ifndef __LINUX_PKT_SCHED_H |
3 | #define __LINUX_PKT_SCHED_H |
4 | |
5 | #include <linux/types.h> |
6 | |
7 | /* Logical priority bands not depending on specific packet scheduler. |
8 | Every scheduler will map them to real traffic classes, if it has |
9 | no more precise mechanism to classify packets. |
10 | |
11 | These numbers have no special meaning, though their coincidence |
12 | with obsolete IPv6 values is not occasional :-). New IPv6 drafts |
13 | preferred full anarchy inspired by diffserv group. |
14 | |
15 | Note: TC_PRIO_BESTEFFORT does not mean that it is the most unhappy |
16 | class, actually, as rule it will be handled with more care than |
17 | filler or even bulk. |
18 | */ |
19 | |
20 | #define TC_PRIO_BESTEFFORT 0 |
21 | #define TC_PRIO_FILLER 1 |
22 | #define TC_PRIO_BULK 2 |
23 | #define TC_PRIO_INTERACTIVE_BULK 4 |
24 | #define TC_PRIO_INTERACTIVE 6 |
25 | #define TC_PRIO_CONTROL 7 |
26 | |
27 | #define TC_PRIO_MAX 15 |
28 | |
29 | /* Generic queue statistics, available for all the elements. |
30 | Particular schedulers may have also their private records. |
31 | */ |
32 | |
33 | struct tc_stats { |
34 | __u64 bytes; /* Number of enqueued bytes */ |
35 | __u32 packets; /* Number of enqueued packets */ |
36 | __u32 drops; /* Packets dropped because of lack of resources */ |
37 | __u32 overlimits; /* Number of throttle events when this |
38 | * flow goes out of allocated bandwidth */ |
39 | __u32 bps; /* Current flow byte rate */ |
40 | __u32 pps; /* Current flow packet rate */ |
41 | __u32 qlen; |
42 | __u32 backlog; |
43 | }; |
44 | |
45 | struct tc_estimator { |
46 | signed char interval; |
47 | unsigned char ewma_log; |
48 | }; |
49 | |
50 | /* "Handles" |
51 | --------- |
52 | |
53 | All the traffic control objects have 32bit identifiers, or "handles". |
54 | |
55 | They can be considered as opaque numbers from user API viewpoint, |
56 | but actually they always consist of two fields: major and |
57 | minor numbers, which are interpreted by kernel specially, |
58 | that may be used by applications, though not recommended. |
59 | |
60 | F.e. qdisc handles always have minor number equal to zero, |
61 | classes (or flows) have major equal to parent qdisc major, and |
62 | minor uniquely identifying class inside qdisc. |
63 | |
64 | Macros to manipulate handles: |
65 | */ |
66 | |
67 | #define TC_H_MAJ_MASK (0xFFFF0000U) |
68 | #define TC_H_MIN_MASK (0x0000FFFFU) |
69 | #define TC_H_MAJ(h) ((h)&TC_H_MAJ_MASK) |
70 | #define TC_H_MIN(h) ((h)&TC_H_MIN_MASK) |
71 | #define TC_H_MAKE(maj,min) (((maj)&TC_H_MAJ_MASK)|((min)&TC_H_MIN_MASK)) |
72 | |
73 | #define TC_H_UNSPEC (0U) |
74 | #define TC_H_ROOT (0xFFFFFFFFU) |
75 | #define TC_H_INGRESS (0xFFFFFFF1U) |
76 | #define TC_H_CLSACT TC_H_INGRESS |
77 | |
78 | #define TC_H_MIN_PRIORITY 0xFFE0U |
79 | #define TC_H_MIN_INGRESS 0xFFF2U |
80 | #define TC_H_MIN_EGRESS 0xFFF3U |
81 | |
82 | /* Need to corrospond to iproute2 tc/tc_core.h "enum link_layer" */ |
83 | enum tc_link_layer { |
84 | TC_LINKLAYER_UNAWARE, /* Indicate unaware old iproute2 util */ |
85 | TC_LINKLAYER_ETHERNET, |
86 | TC_LINKLAYER_ATM, |
87 | }; |
88 | #define TC_LINKLAYER_MASK 0x0F /* limit use to lower 4 bits */ |
89 | |
90 | struct tc_ratespec { |
91 | unsigned char cell_log; |
92 | __u8 linklayer; /* lower 4 bits */ |
93 | unsigned short overhead; |
94 | short cell_align; |
95 | unsigned short mpu; |
96 | __u32 rate; |
97 | }; |
98 | |
99 | #define TC_RTAB_SIZE 1024 |
100 | |
101 | struct tc_sizespec { |
102 | unsigned char cell_log; |
103 | unsigned char size_log; |
104 | short cell_align; |
105 | int overhead; |
106 | unsigned int linklayer; |
107 | unsigned int mpu; |
108 | unsigned int mtu; |
109 | unsigned int tsize; |
110 | }; |
111 | |
112 | enum { |
113 | TCA_STAB_UNSPEC, |
114 | TCA_STAB_BASE, |
115 | TCA_STAB_DATA, |
116 | __TCA_STAB_MAX |
117 | }; |
118 | |
119 | #define TCA_STAB_MAX (__TCA_STAB_MAX - 1) |
120 | |
121 | /* FIFO section */ |
122 | |
123 | struct tc_fifo_qopt { |
124 | __u32 limit; /* Queue length: bytes for bfifo, packets for pfifo */ |
125 | }; |
126 | |
127 | /* SKBPRIO section */ |
128 | |
129 | /* |
130 | * Priorities go from zero to (SKBPRIO_MAX_PRIORITY - 1). |
131 | * SKBPRIO_MAX_PRIORITY should be at least 64 in order for skbprio to be able |
132 | * to map one to one the DS field of IPV4 and IPV6 headers. |
133 | * Memory allocation grows linearly with SKBPRIO_MAX_PRIORITY. |
134 | */ |
135 | |
136 | #define SKBPRIO_MAX_PRIORITY 64 |
137 | |
138 | struct tc_skbprio_qopt { |
139 | __u32 limit; /* Queue length in packets. */ |
140 | }; |
141 | |
142 | /* PRIO section */ |
143 | |
144 | #define TCQ_PRIO_BANDS 16 |
145 | #define TCQ_MIN_PRIO_BANDS 2 |
146 | |
147 | struct tc_prio_qopt { |
148 | int bands; /* Number of bands */ |
149 | __u8 priomap[TC_PRIO_MAX+1]; /* Map: logical priority -> PRIO band */ |
150 | }; |
151 | |
152 | /* MULTIQ section */ |
153 | |
154 | struct tc_multiq_qopt { |
155 | __u16 bands; /* Number of bands */ |
156 | __u16 max_bands; /* Maximum number of queues */ |
157 | }; |
158 | |
159 | /* PLUG section */ |
160 | |
161 | #define TCQ_PLUG_BUFFER 0 |
162 | #define TCQ_PLUG_RELEASE_ONE 1 |
163 | #define TCQ_PLUG_RELEASE_INDEFINITE 2 |
164 | #define TCQ_PLUG_LIMIT 3 |
165 | |
166 | struct tc_plug_qopt { |
167 | /* TCQ_PLUG_BUFFER: Inset a plug into the queue and |
168 | * buffer any incoming packets |
169 | * TCQ_PLUG_RELEASE_ONE: Dequeue packets from queue head |
170 | * to beginning of the next plug. |
171 | * TCQ_PLUG_RELEASE_INDEFINITE: Dequeue all packets from queue. |
172 | * Stop buffering packets until the next TCQ_PLUG_BUFFER |
173 | * command is received (just act as a pass-thru queue). |
174 | * TCQ_PLUG_LIMIT: Increase/decrease queue size |
175 | */ |
176 | int action; |
177 | __u32 limit; |
178 | }; |
179 | |
180 | /* TBF section */ |
181 | |
182 | struct tc_tbf_qopt { |
183 | struct tc_ratespec rate; |
184 | struct tc_ratespec peakrate; |
185 | __u32 limit; |
186 | __u32 buffer; |
187 | __u32 mtu; |
188 | }; |
189 | |
190 | enum { |
191 | TCA_TBF_UNSPEC, |
192 | TCA_TBF_PARMS, |
193 | TCA_TBF_RTAB, |
194 | TCA_TBF_PTAB, |
195 | TCA_TBF_RATE64, |
196 | TCA_TBF_PRATE64, |
197 | TCA_TBF_BURST, |
198 | TCA_TBF_PBURST, |
199 | TCA_TBF_PAD, |
200 | __TCA_TBF_MAX, |
201 | }; |
202 | |
203 | #define TCA_TBF_MAX (__TCA_TBF_MAX - 1) |
204 | |
205 | |
206 | /* TEQL section */ |
207 | |
208 | /* TEQL does not require any parameters */ |
209 | |
210 | /* SFQ section */ |
211 | |
212 | struct tc_sfq_qopt { |
213 | unsigned quantum; /* Bytes per round allocated to flow */ |
214 | int perturb_period; /* Period of hash perturbation */ |
215 | __u32 limit; /* Maximal packets in queue */ |
216 | unsigned divisor; /* Hash divisor */ |
217 | unsigned flows; /* Maximal number of flows */ |
218 | }; |
219 | |
220 | struct tc_sfqred_stats { |
221 | __u32 prob_drop; /* Early drops, below max threshold */ |
222 | __u32 forced_drop; /* Early drops, after max threshold */ |
223 | __u32 prob_mark; /* Marked packets, below max threshold */ |
224 | __u32 forced_mark; /* Marked packets, after max threshold */ |
225 | __u32 prob_mark_head; /* Marked packets, below max threshold */ |
226 | __u32 forced_mark_head;/* Marked packets, after max threshold */ |
227 | }; |
228 | |
229 | struct tc_sfq_qopt_v1 { |
230 | struct tc_sfq_qopt v0; |
231 | unsigned int depth; /* max number of packets per flow */ |
232 | unsigned int headdrop; |
233 | /* SFQRED parameters */ |
234 | __u32 limit; /* HARD maximal flow queue length (bytes) */ |
235 | __u32 qth_min; /* Min average length threshold (bytes) */ |
236 | __u32 qth_max; /* Max average length threshold (bytes) */ |
237 | unsigned char Wlog; /* log(W) */ |
238 | unsigned char Plog; /* log(P_max/(qth_max-qth_min)) */ |
239 | unsigned char Scell_log; /* cell size for idle damping */ |
240 | unsigned char flags; |
241 | __u32 max_P; /* probability, high resolution */ |
242 | /* SFQRED stats */ |
243 | struct tc_sfqred_stats stats; |
244 | }; |
245 | |
246 | |
247 | struct tc_sfq_xstats { |
248 | __s32 allot; |
249 | }; |
250 | |
251 | /* RED section */ |
252 | |
253 | enum { |
254 | TCA_RED_UNSPEC, |
255 | TCA_RED_PARMS, |
256 | TCA_RED_STAB, |
257 | TCA_RED_MAX_P, |
258 | __TCA_RED_MAX, |
259 | }; |
260 | |
261 | #define TCA_RED_MAX (__TCA_RED_MAX - 1) |
262 | |
263 | struct tc_red_qopt { |
264 | __u32 limit; /* HARD maximal queue length (bytes) */ |
265 | __u32 qth_min; /* Min average length threshold (bytes) */ |
266 | __u32 qth_max; /* Max average length threshold (bytes) */ |
267 | unsigned char Wlog; /* log(W) */ |
268 | unsigned char Plog; /* log(P_max/(qth_max-qth_min)) */ |
269 | unsigned char Scell_log; /* cell size for idle damping */ |
270 | unsigned char flags; |
271 | #define TC_RED_ECN 1 |
272 | #define TC_RED_HARDDROP 2 |
273 | #define TC_RED_ADAPTATIVE 4 |
274 | }; |
275 | |
276 | struct tc_red_xstats { |
277 | __u32 early; /* Early drops */ |
278 | __u32 pdrop; /* Drops due to queue limits */ |
279 | __u32 other; /* Drops due to drop() calls */ |
280 | __u32 marked; /* Marked packets */ |
281 | }; |
282 | |
283 | /* GRED section */ |
284 | |
285 | #define MAX_DPs 16 |
286 | |
287 | enum { |
288 | TCA_GRED_UNSPEC, |
289 | TCA_GRED_PARMS, |
290 | TCA_GRED_STAB, |
291 | TCA_GRED_DPS, |
292 | TCA_GRED_MAX_P, |
293 | TCA_GRED_LIMIT, |
294 | TCA_GRED_VQ_LIST, /* nested TCA_GRED_VQ_ENTRY */ |
295 | __TCA_GRED_MAX, |
296 | }; |
297 | |
298 | #define TCA_GRED_MAX (__TCA_GRED_MAX - 1) |
299 | |
300 | enum { |
301 | TCA_GRED_VQ_ENTRY_UNSPEC, |
302 | TCA_GRED_VQ_ENTRY, /* nested TCA_GRED_VQ_* */ |
303 | __TCA_GRED_VQ_ENTRY_MAX, |
304 | }; |
305 | #define TCA_GRED_VQ_ENTRY_MAX (__TCA_GRED_VQ_ENTRY_MAX - 1) |
306 | |
307 | enum { |
308 | TCA_GRED_VQ_UNSPEC, |
309 | TCA_GRED_VQ_PAD, |
310 | TCA_GRED_VQ_DP, /* u32 */ |
311 | TCA_GRED_VQ_STAT_BYTES, /* u64 */ |
312 | TCA_GRED_VQ_STAT_PACKETS, /* u32 */ |
313 | TCA_GRED_VQ_STAT_BACKLOG, /* u32 */ |
314 | TCA_GRED_VQ_STAT_PROB_DROP, /* u32 */ |
315 | TCA_GRED_VQ_STAT_PROB_MARK, /* u32 */ |
316 | TCA_GRED_VQ_STAT_FORCED_DROP, /* u32 */ |
317 | TCA_GRED_VQ_STAT_FORCED_MARK, /* u32 */ |
318 | TCA_GRED_VQ_STAT_PDROP, /* u32 */ |
319 | TCA_GRED_VQ_STAT_OTHER, /* u32 */ |
320 | TCA_GRED_VQ_FLAGS, /* u32 */ |
321 | __TCA_GRED_VQ_MAX |
322 | }; |
323 | |
324 | #define TCA_GRED_VQ_MAX (__TCA_GRED_VQ_MAX - 1) |
325 | |
326 | struct tc_gred_qopt { |
327 | __u32 limit; /* HARD maximal queue length (bytes) */ |
328 | __u32 qth_min; /* Min average length threshold (bytes) */ |
329 | __u32 qth_max; /* Max average length threshold (bytes) */ |
330 | __u32 DP; /* up to 2^32 DPs */ |
331 | __u32 backlog; |
332 | __u32 qave; |
333 | __u32 forced; |
334 | __u32 early; |
335 | __u32 other; |
336 | __u32 pdrop; |
337 | __u8 Wlog; /* log(W) */ |
338 | __u8 Plog; /* log(P_max/(qth_max-qth_min)) */ |
339 | __u8 Scell_log; /* cell size for idle damping */ |
340 | __u8 prio; /* prio of this VQ */ |
341 | __u32 packets; |
342 | __u32 bytesin; |
343 | }; |
344 | |
345 | /* gred setup */ |
346 | struct tc_gred_sopt { |
347 | __u32 DPs; |
348 | __u32 def_DP; |
349 | __u8 grio; |
350 | __u8 flags; |
351 | __u16 pad1; |
352 | }; |
353 | |
354 | /* CHOKe section */ |
355 | |
356 | enum { |
357 | TCA_CHOKE_UNSPEC, |
358 | TCA_CHOKE_PARMS, |
359 | TCA_CHOKE_STAB, |
360 | TCA_CHOKE_MAX_P, |
361 | __TCA_CHOKE_MAX, |
362 | }; |
363 | |
364 | #define TCA_CHOKE_MAX (__TCA_CHOKE_MAX - 1) |
365 | |
366 | struct tc_choke_qopt { |
367 | __u32 limit; /* Hard queue length (packets) */ |
368 | __u32 qth_min; /* Min average threshold (packets) */ |
369 | __u32 qth_max; /* Max average threshold (packets) */ |
370 | unsigned char Wlog; /* log(W) */ |
371 | unsigned char Plog; /* log(P_max/(qth_max-qth_min)) */ |
372 | unsigned char Scell_log; /* cell size for idle damping */ |
373 | unsigned char flags; /* see RED flags */ |
374 | }; |
375 | |
376 | struct tc_choke_xstats { |
377 | __u32 early; /* Early drops */ |
378 | __u32 pdrop; /* Drops due to queue limits */ |
379 | __u32 other; /* Drops due to drop() calls */ |
380 | __u32 marked; /* Marked packets */ |
381 | __u32 matched; /* Drops due to flow match */ |
382 | }; |
383 | |
384 | /* HTB section */ |
385 | #define TC_HTB_NUMPRIO 8 |
386 | #define TC_HTB_MAXDEPTH 8 |
387 | #define TC_HTB_PROTOVER 3 /* the same as HTB and TC's major */ |
388 | |
389 | struct tc_htb_opt { |
390 | struct tc_ratespec rate; |
391 | struct tc_ratespec ceil; |
392 | __u32 buffer; |
393 | __u32 cbuffer; |
394 | __u32 quantum; |
395 | __u32 level; /* out only */ |
396 | __u32 prio; |
397 | }; |
398 | struct tc_htb_glob { |
399 | __u32 version; /* to match HTB/TC */ |
400 | __u32 rate2quantum; /* bps->quantum divisor */ |
401 | __u32 defcls; /* default class number */ |
402 | __u32 debug; /* debug flags */ |
403 | |
404 | /* stats */ |
405 | __u32 direct_pkts; /* count of non shaped packets */ |
406 | }; |
407 | enum { |
408 | TCA_HTB_UNSPEC, |
409 | TCA_HTB_PARMS, |
410 | TCA_HTB_INIT, |
411 | TCA_HTB_CTAB, |
412 | TCA_HTB_RTAB, |
413 | TCA_HTB_DIRECT_QLEN, |
414 | TCA_HTB_RATE64, |
415 | TCA_HTB_CEIL64, |
416 | TCA_HTB_PAD, |
417 | TCA_HTB_OFFLOAD, |
418 | __TCA_HTB_MAX, |
419 | }; |
420 | |
421 | #define TCA_HTB_MAX (__TCA_HTB_MAX - 1) |
422 | |
423 | struct tc_htb_xstats { |
424 | __u32 lends; |
425 | __u32 borrows; |
426 | __u32 giants; /* unused since 'Make HTB scheduler work with TSO.' */ |
427 | __s32 tokens; |
428 | __s32 ctokens; |
429 | }; |
430 | |
431 | /* HFSC section */ |
432 | |
433 | struct tc_hfsc_qopt { |
434 | __u16 defcls; /* default class */ |
435 | }; |
436 | |
437 | struct tc_service_curve { |
438 | __u32 m1; /* slope of the first segment in bps */ |
439 | __u32 d; /* x-projection of the first segment in us */ |
440 | __u32 m2; /* slope of the second segment in bps */ |
441 | }; |
442 | |
443 | struct tc_hfsc_stats { |
444 | __u64 work; /* total work done */ |
445 | __u64 rtwork; /* work done by real-time criteria */ |
446 | __u32 period; /* current period */ |
447 | __u32 level; /* class level in hierarchy */ |
448 | }; |
449 | |
450 | enum { |
451 | TCA_HFSC_UNSPEC, |
452 | TCA_HFSC_RSC, |
453 | TCA_HFSC_FSC, |
454 | TCA_HFSC_USC, |
455 | __TCA_HFSC_MAX, |
456 | }; |
457 | |
458 | #define TCA_HFSC_MAX (__TCA_HFSC_MAX - 1) |
459 | |
460 | /* Network emulator */ |
461 | |
462 | enum { |
463 | TCA_NETEM_UNSPEC, |
464 | TCA_NETEM_CORR, |
465 | TCA_NETEM_DELAY_DIST, |
466 | TCA_NETEM_REORDER, |
467 | TCA_NETEM_CORRUPT, |
468 | TCA_NETEM_LOSS, |
469 | TCA_NETEM_RATE, |
470 | TCA_NETEM_ECN, |
471 | TCA_NETEM_RATE64, |
472 | TCA_NETEM_PAD, |
473 | TCA_NETEM_LATENCY64, |
474 | TCA_NETEM_JITTER64, |
475 | TCA_NETEM_SLOT, |
476 | TCA_NETEM_SLOT_DIST, |
477 | __TCA_NETEM_MAX, |
478 | }; |
479 | |
480 | #define TCA_NETEM_MAX (__TCA_NETEM_MAX - 1) |
481 | |
482 | struct tc_netem_qopt { |
483 | __u32 latency; /* added delay (us) */ |
484 | __u32 limit; /* fifo limit (packets) */ |
485 | __u32 loss; /* random packet loss (0=none ~0=100%) */ |
486 | __u32 gap; /* re-ordering gap (0 for none) */ |
487 | __u32 duplicate; /* random packet dup (0=none ~0=100%) */ |
488 | __u32 jitter; /* random jitter in latency (us) */ |
489 | }; |
490 | |
491 | struct tc_netem_corr { |
492 | __u32 delay_corr; /* delay correlation */ |
493 | __u32 loss_corr; /* packet loss correlation */ |
494 | __u32 dup_corr; /* duplicate correlation */ |
495 | }; |
496 | |
497 | struct tc_netem_reorder { |
498 | __u32 probability; |
499 | __u32 correlation; |
500 | }; |
501 | |
502 | struct tc_netem_corrupt { |
503 | __u32 probability; |
504 | __u32 correlation; |
505 | }; |
506 | |
507 | struct tc_netem_rate { |
508 | __u32 rate; /* byte/s */ |
509 | __s32 packet_overhead; |
510 | __u32 cell_size; |
511 | __s32 cell_overhead; |
512 | }; |
513 | |
514 | struct tc_netem_slot { |
515 | __s64 min_delay; /* nsec */ |
516 | __s64 max_delay; |
517 | __s32 max_packets; |
518 | __s32 max_bytes; |
519 | __s64 dist_delay; /* nsec */ |
520 | __s64 dist_jitter; /* nsec */ |
521 | }; |
522 | |
523 | enum { |
524 | NETEM_LOSS_UNSPEC, |
525 | NETEM_LOSS_GI, /* General Intuitive - 4 state model */ |
526 | NETEM_LOSS_GE, /* Gilbert Elliot models */ |
527 | __NETEM_LOSS_MAX |
528 | }; |
529 | #define NETEM_LOSS_MAX (__NETEM_LOSS_MAX - 1) |
530 | |
531 | /* State transition probabilities for 4 state model */ |
532 | struct tc_netem_gimodel { |
533 | __u32 p13; |
534 | __u32 p31; |
535 | __u32 p32; |
536 | __u32 p14; |
537 | __u32 p23; |
538 | }; |
539 | |
540 | /* Gilbert-Elliot models */ |
541 | struct tc_netem_gemodel { |
542 | __u32 p; |
543 | __u32 r; |
544 | __u32 h; |
545 | __u32 k1; |
546 | }; |
547 | |
548 | #define NETEM_DIST_SCALE 8192 |
549 | #define NETEM_DIST_MAX 16384 |
550 | |
551 | /* DRR */ |
552 | |
553 | enum { |
554 | TCA_DRR_UNSPEC, |
555 | TCA_DRR_QUANTUM, |
556 | __TCA_DRR_MAX |
557 | }; |
558 | |
559 | #define TCA_DRR_MAX (__TCA_DRR_MAX - 1) |
560 | |
561 | struct tc_drr_stats { |
562 | __u32 deficit; |
563 | }; |
564 | |
565 | /* MQPRIO */ |
566 | #define TC_QOPT_BITMASK 15 |
567 | #define TC_QOPT_MAX_QUEUE 16 |
568 | |
569 | enum { |
570 | TC_MQPRIO_HW_OFFLOAD_NONE, /* no offload requested */ |
571 | TC_MQPRIO_HW_OFFLOAD_TCS, /* offload TCs, no queue counts */ |
572 | __TC_MQPRIO_HW_OFFLOAD_MAX |
573 | }; |
574 | |
575 | #define TC_MQPRIO_HW_OFFLOAD_MAX (__TC_MQPRIO_HW_OFFLOAD_MAX - 1) |
576 | |
577 | enum { |
578 | TC_MQPRIO_MODE_DCB, |
579 | TC_MQPRIO_MODE_CHANNEL, |
580 | __TC_MQPRIO_MODE_MAX |
581 | }; |
582 | |
583 | #define __TC_MQPRIO_MODE_MAX (__TC_MQPRIO_MODE_MAX - 1) |
584 | |
585 | enum { |
586 | TC_MQPRIO_SHAPER_DCB, |
587 | TC_MQPRIO_SHAPER_BW_RATE, /* Add new shapers below */ |
588 | __TC_MQPRIO_SHAPER_MAX |
589 | }; |
590 | |
591 | #define __TC_MQPRIO_SHAPER_MAX (__TC_MQPRIO_SHAPER_MAX - 1) |
592 | |
593 | struct tc_mqprio_qopt { |
594 | __u8 num_tc; |
595 | __u8 prio_tc_map[TC_QOPT_BITMASK + 1]; |
596 | __u8 hw; |
597 | __u16 count[TC_QOPT_MAX_QUEUE]; |
598 | __u16 offset[TC_QOPT_MAX_QUEUE]; |
599 | }; |
600 | |
601 | #define TC_MQPRIO_F_MODE 0x1 |
602 | #define TC_MQPRIO_F_SHAPER 0x2 |
603 | #define TC_MQPRIO_F_MIN_RATE 0x4 |
604 | #define TC_MQPRIO_F_MAX_RATE 0x8 |
605 | |
606 | enum { |
607 | TCA_MQPRIO_UNSPEC, |
608 | TCA_MQPRIO_MODE, |
609 | TCA_MQPRIO_SHAPER, |
610 | TCA_MQPRIO_MIN_RATE64, |
611 | TCA_MQPRIO_MAX_RATE64, |
612 | __TCA_MQPRIO_MAX, |
613 | }; |
614 | |
615 | #define TCA_MQPRIO_MAX (__TCA_MQPRIO_MAX - 1) |
616 | |
617 | /* SFB */ |
618 | |
619 | enum { |
620 | TCA_SFB_UNSPEC, |
621 | TCA_SFB_PARMS, |
622 | __TCA_SFB_MAX, |
623 | }; |
624 | |
625 | #define TCA_SFB_MAX (__TCA_SFB_MAX - 1) |
626 | |
627 | /* |
628 | * Note: increment, decrement are Q0.16 fixed-point values. |
629 | */ |
630 | struct tc_sfb_qopt { |
631 | __u32 rehash_interval; /* delay between hash move, in ms */ |
632 | __u32 warmup_time; /* double buffering warmup time in ms (warmup_time < rehash_interval) */ |
633 | __u32 max; /* max len of qlen_min */ |
634 | __u32 bin_size; /* maximum queue length per bin */ |
635 | __u32 increment; /* probability increment, (d1 in Blue) */ |
636 | __u32 decrement; /* probability decrement, (d2 in Blue) */ |
637 | __u32 limit; /* max SFB queue length */ |
638 | __u32 penalty_rate; /* inelastic flows are rate limited to 'rate' pps */ |
639 | __u32 penalty_burst; |
640 | }; |
641 | |
642 | struct tc_sfb_xstats { |
643 | __u32 earlydrop; |
644 | __u32 penaltydrop; |
645 | __u32 bucketdrop; |
646 | __u32 queuedrop; |
647 | __u32 childdrop; /* drops in child qdisc */ |
648 | __u32 marked; |
649 | __u32 maxqlen; |
650 | __u32 maxprob; |
651 | __u32 avgprob; |
652 | }; |
653 | |
654 | #define SFB_MAX_PROB 0xFFFF |
655 | |
656 | /* QFQ */ |
657 | enum { |
658 | TCA_QFQ_UNSPEC, |
659 | TCA_QFQ_WEIGHT, |
660 | TCA_QFQ_LMAX, |
661 | __TCA_QFQ_MAX |
662 | }; |
663 | |
664 | #define TCA_QFQ_MAX (__TCA_QFQ_MAX - 1) |
665 | |
666 | struct tc_qfq_stats { |
667 | __u32 weight; |
668 | __u32 lmax; |
669 | }; |
670 | |
671 | /* CODEL */ |
672 | |
673 | enum { |
674 | TCA_CODEL_UNSPEC, |
675 | TCA_CODEL_TARGET, |
676 | TCA_CODEL_LIMIT, |
677 | TCA_CODEL_INTERVAL, |
678 | TCA_CODEL_ECN, |
679 | TCA_CODEL_CE_THRESHOLD, |
680 | __TCA_CODEL_MAX |
681 | }; |
682 | |
683 | #define TCA_CODEL_MAX (__TCA_CODEL_MAX - 1) |
684 | |
685 | struct tc_codel_xstats { |
686 | __u32 maxpacket; /* largest packet we've seen so far */ |
687 | __u32 count; /* how many drops we've done since the last time we |
688 | * entered dropping state |
689 | */ |
690 | __u32 lastcount; /* count at entry to dropping state */ |
691 | __u32 ldelay; /* in-queue delay seen by most recently dequeued packet */ |
692 | __s32 drop_next; /* time to drop next packet */ |
693 | __u32 drop_overlimit; /* number of time max qdisc packet limit was hit */ |
694 | __u32 ecn_mark; /* number of packets we ECN marked instead of dropped */ |
695 | __u32 dropping; /* are we in dropping state ? */ |
696 | __u32 ce_mark; /* number of CE marked packets because of ce_threshold */ |
697 | }; |
698 | |
699 | /* FQ_CODEL */ |
700 | |
701 | enum { |
702 | TCA_FQ_CODEL_UNSPEC, |
703 | TCA_FQ_CODEL_TARGET, |
704 | TCA_FQ_CODEL_LIMIT, |
705 | TCA_FQ_CODEL_INTERVAL, |
706 | TCA_FQ_CODEL_ECN, |
707 | TCA_FQ_CODEL_FLOWS, |
708 | TCA_FQ_CODEL_QUANTUM, |
709 | TCA_FQ_CODEL_CE_THRESHOLD, |
710 | TCA_FQ_CODEL_DROP_BATCH_SIZE, |
711 | TCA_FQ_CODEL_MEMORY_LIMIT, |
712 | __TCA_FQ_CODEL_MAX |
713 | }; |
714 | |
715 | #define TCA_FQ_CODEL_MAX (__TCA_FQ_CODEL_MAX - 1) |
716 | |
717 | enum { |
718 | TCA_FQ_CODEL_XSTATS_QDISC, |
719 | TCA_FQ_CODEL_XSTATS_CLASS, |
720 | }; |
721 | |
722 | struct tc_fq_codel_qd_stats { |
723 | __u32 maxpacket; /* largest packet we've seen so far */ |
724 | __u32 drop_overlimit; /* number of time max qdisc |
725 | * packet limit was hit |
726 | */ |
727 | __u32 ecn_mark; /* number of packets we ECN marked |
728 | * instead of being dropped |
729 | */ |
730 | __u32 new_flow_count; /* number of time packets |
731 | * created a 'new flow' |
732 | */ |
733 | __u32 new_flows_len; /* count of flows in new list */ |
734 | __u32 old_flows_len; /* count of flows in old list */ |
735 | __u32 ce_mark; /* packets above ce_threshold */ |
736 | __u32 memory_usage; /* in bytes */ |
737 | __u32 drop_overmemory; |
738 | }; |
739 | |
740 | struct tc_fq_codel_cl_stats { |
741 | __s32 deficit; |
742 | __u32 ldelay; /* in-queue delay seen by most recently |
743 | * dequeued packet |
744 | */ |
745 | __u32 count; |
746 | __u32 lastcount; |
747 | __u32 dropping; |
748 | __s32 drop_next; |
749 | }; |
750 | |
751 | struct tc_fq_codel_xstats { |
752 | __u32 type; |
753 | union { |
754 | struct tc_fq_codel_qd_stats qdisc_stats; |
755 | struct tc_fq_codel_cl_stats class_stats; |
756 | }; |
757 | }; |
758 | |
759 | /* FQ */ |
760 | |
761 | enum { |
762 | TCA_FQ_UNSPEC, |
763 | |
764 | TCA_FQ_PLIMIT, /* limit of total number of packets in queue */ |
765 | |
766 | TCA_FQ_FLOW_PLIMIT, /* limit of packets per flow */ |
767 | |
768 | TCA_FQ_QUANTUM, /* RR quantum */ |
769 | |
770 | TCA_FQ_INITIAL_QUANTUM, /* RR quantum for new flow */ |
771 | |
772 | TCA_FQ_RATE_ENABLE, /* enable/disable rate limiting */ |
773 | |
774 | TCA_FQ_FLOW_DEFAULT_RATE,/* obsolete, do not use */ |
775 | |
776 | TCA_FQ_FLOW_MAX_RATE, /* per flow max rate */ |
777 | |
778 | TCA_FQ_BUCKETS_LOG, /* log2(number of buckets) */ |
779 | |
780 | TCA_FQ_FLOW_REFILL_DELAY, /* flow credit refill delay in usec */ |
781 | |
782 | TCA_FQ_ORPHAN_MASK, /* mask applied to orphaned skb hashes */ |
783 | |
784 | TCA_FQ_LOW_RATE_THRESHOLD, /* per packet delay under this rate */ |
785 | |
786 | TCA_FQ_CE_THRESHOLD, /* DCTCP-like CE-marking threshold */ |
787 | |
788 | __TCA_FQ_MAX |
789 | }; |
790 | |
791 | #define TCA_FQ_MAX (__TCA_FQ_MAX - 1) |
792 | |
793 | struct tc_fq_qd_stats { |
794 | __u64 gc_flows; |
795 | __u64 highprio_packets; |
796 | __u64 tcp_retrans; |
797 | __u64 throttled; |
798 | __u64 flows_plimit; |
799 | __u64 pkts_too_long; |
800 | __u64 allocation_errors; |
801 | __s64 time_next_delayed_flow; |
802 | __u32 flows; |
803 | __u32 inactive_flows; |
804 | __u32 throttled_flows; |
805 | __u32 unthrottle_latency_ns; |
806 | __u64 ce_mark; /* packets above ce_threshold */ |
807 | }; |
808 | |
809 | /* Heavy-Hitter Filter */ |
810 | |
811 | enum { |
812 | TCA_HHF_UNSPEC, |
813 | TCA_HHF_BACKLOG_LIMIT, |
814 | TCA_HHF_QUANTUM, |
815 | TCA_HHF_HH_FLOWS_LIMIT, |
816 | TCA_HHF_RESET_TIMEOUT, |
817 | TCA_HHF_ADMIT_BYTES, |
818 | TCA_HHF_EVICT_TIMEOUT, |
819 | TCA_HHF_NON_HH_WEIGHT, |
820 | __TCA_HHF_MAX |
821 | }; |
822 | |
823 | #define TCA_HHF_MAX (__TCA_HHF_MAX - 1) |
824 | |
825 | struct tc_hhf_xstats { |
826 | __u32 drop_overlimit; /* number of times max qdisc packet limit |
827 | * was hit |
828 | */ |
829 | __u32 hh_overlimit; /* number of times max heavy-hitters was hit */ |
830 | __u32 hh_tot_count; /* number of captured heavy-hitters so far */ |
831 | __u32 hh_cur_count; /* number of current heavy-hitters */ |
832 | }; |
833 | |
834 | /* PIE */ |
835 | enum { |
836 | TCA_PIE_UNSPEC, |
837 | TCA_PIE_TARGET, |
838 | TCA_PIE_LIMIT, |
839 | TCA_PIE_TUPDATE, |
840 | TCA_PIE_ALPHA, |
841 | TCA_PIE_BETA, |
842 | TCA_PIE_ECN, |
843 | TCA_PIE_BYTEMODE, |
844 | __TCA_PIE_MAX |
845 | }; |
846 | #define TCA_PIE_MAX (__TCA_PIE_MAX - 1) |
847 | |
848 | struct tc_pie_xstats { |
849 | __u32 prob; /* current probability */ |
850 | __u32 delay; /* current delay in ms */ |
851 | __u32 avg_dq_rate; /* current average dq_rate in bits/pie_time */ |
852 | __u32 packets_in; /* total number of packets enqueued */ |
853 | __u32 dropped; /* packets dropped due to pie_action */ |
854 | __u32 overlimit; /* dropped due to lack of space in queue */ |
855 | __u32 maxq; /* maximum queue size */ |
856 | __u32 ecn_mark; /* packets marked with ecn*/ |
857 | }; |
858 | |
859 | /* CBS */ |
860 | struct tc_cbs_qopt { |
861 | __u8 offload; |
862 | __u8 _pad[3]; |
863 | __s32 hicredit; |
864 | __s32 locredit; |
865 | __s32 idleslope; |
866 | __s32 sendslope; |
867 | }; |
868 | |
869 | enum { |
870 | TCA_CBS_UNSPEC, |
871 | TCA_CBS_PARMS, |
872 | __TCA_CBS_MAX, |
873 | }; |
874 | |
875 | #define TCA_CBS_MAX (__TCA_CBS_MAX - 1) |
876 | |
877 | |
878 | /* ETF */ |
879 | struct tc_etf_qopt { |
880 | __s32 delta; |
881 | __s32 clockid; |
882 | __u32 flags; |
883 | #define TC_ETF_DEADLINE_MODE_ON BIT(0) |
884 | #define TC_ETF_OFFLOAD_ON BIT(1) |
885 | }; |
886 | |
887 | enum { |
888 | TCA_ETF_UNSPEC, |
889 | TCA_ETF_PARMS, |
890 | __TCA_ETF_MAX, |
891 | }; |
892 | |
893 | #define TCA_ETF_MAX (__TCA_ETF_MAX - 1) |
894 | |
895 | |
896 | /* CAKE */ |
897 | enum { |
898 | TCA_CAKE_UNSPEC, |
899 | TCA_CAKE_PAD, |
900 | TCA_CAKE_BASE_RATE64, |
901 | TCA_CAKE_DIFFSERV_MODE, |
902 | TCA_CAKE_ATM, |
903 | TCA_CAKE_FLOW_MODE, |
904 | TCA_CAKE_OVERHEAD, |
905 | TCA_CAKE_RTT, |
906 | TCA_CAKE_TARGET, |
907 | TCA_CAKE_AUTORATE, |
908 | TCA_CAKE_MEMORY, |
909 | TCA_CAKE_NAT, |
910 | TCA_CAKE_RAW, |
911 | TCA_CAKE_WASH, |
912 | TCA_CAKE_MPU, |
913 | TCA_CAKE_INGRESS, |
914 | TCA_CAKE_ACK_FILTER, |
915 | TCA_CAKE_SPLIT_GSO, |
916 | __TCA_CAKE_MAX |
917 | }; |
918 | #define TCA_CAKE_MAX (__TCA_CAKE_MAX - 1) |
919 | |
920 | enum { |
921 | __TCA_CAKE_STATS_INVALID, |
922 | TCA_CAKE_STATS_PAD, |
923 | TCA_CAKE_STATS_CAPACITY_ESTIMATE64, |
924 | TCA_CAKE_STATS_MEMORY_LIMIT, |
925 | TCA_CAKE_STATS_MEMORY_USED, |
926 | TCA_CAKE_STATS_AVG_NETOFF, |
927 | TCA_CAKE_STATS_MIN_NETLEN, |
928 | TCA_CAKE_STATS_MAX_NETLEN, |
929 | TCA_CAKE_STATS_MIN_ADJLEN, |
930 | TCA_CAKE_STATS_MAX_ADJLEN, |
931 | TCA_CAKE_STATS_TIN_STATS, |
932 | TCA_CAKE_STATS_DEFICIT, |
933 | TCA_CAKE_STATS_COBALT_COUNT, |
934 | TCA_CAKE_STATS_DROPPING, |
935 | TCA_CAKE_STATS_DROP_NEXT_US, |
936 | TCA_CAKE_STATS_P_DROP, |
937 | TCA_CAKE_STATS_BLUE_TIMER_US, |
938 | __TCA_CAKE_STATS_MAX |
939 | }; |
940 | #define TCA_CAKE_STATS_MAX (__TCA_CAKE_STATS_MAX - 1) |
941 | |
942 | enum { |
943 | __TCA_CAKE_TIN_STATS_INVALID, |
944 | TCA_CAKE_TIN_STATS_PAD, |
945 | TCA_CAKE_TIN_STATS_SENT_PACKETS, |
946 | TCA_CAKE_TIN_STATS_SENT_BYTES64, |
947 | TCA_CAKE_TIN_STATS_DROPPED_PACKETS, |
948 | TCA_CAKE_TIN_STATS_DROPPED_BYTES64, |
949 | TCA_CAKE_TIN_STATS_ACKS_DROPPED_PACKETS, |
950 | TCA_CAKE_TIN_STATS_ACKS_DROPPED_BYTES64, |
951 | TCA_CAKE_TIN_STATS_ECN_MARKED_PACKETS, |
952 | TCA_CAKE_TIN_STATS_ECN_MARKED_BYTES64, |
953 | TCA_CAKE_TIN_STATS_BACKLOG_PACKETS, |
954 | TCA_CAKE_TIN_STATS_BACKLOG_BYTES, |
955 | TCA_CAKE_TIN_STATS_THRESHOLD_RATE64, |
956 | TCA_CAKE_TIN_STATS_TARGET_US, |
957 | TCA_CAKE_TIN_STATS_INTERVAL_US, |
958 | TCA_CAKE_TIN_STATS_WAY_INDIRECT_HITS, |
959 | TCA_CAKE_TIN_STATS_WAY_MISSES, |
960 | TCA_CAKE_TIN_STATS_WAY_COLLISIONS, |
961 | TCA_CAKE_TIN_STATS_PEAK_DELAY_US, |
962 | TCA_CAKE_TIN_STATS_AVG_DELAY_US, |
963 | TCA_CAKE_TIN_STATS_BASE_DELAY_US, |
964 | TCA_CAKE_TIN_STATS_SPARSE_FLOWS, |
965 | TCA_CAKE_TIN_STATS_BULK_FLOWS, |
966 | TCA_CAKE_TIN_STATS_UNRESPONSIVE_FLOWS, |
967 | TCA_CAKE_TIN_STATS_MAX_SKBLEN, |
968 | TCA_CAKE_TIN_STATS_FLOW_QUANTUM, |
969 | __TCA_CAKE_TIN_STATS_MAX |
970 | }; |
971 | #define TCA_CAKE_TIN_STATS_MAX (__TCA_CAKE_TIN_STATS_MAX - 1) |
972 | #define TC_CAKE_MAX_TINS (8) |
973 | |
974 | enum { |
975 | CAKE_FLOW_NONE = 0, |
976 | CAKE_FLOW_SRC_IP, |
977 | CAKE_FLOW_DST_IP, |
978 | CAKE_FLOW_HOSTS, /* = CAKE_FLOW_SRC_IP | CAKE_FLOW_DST_IP */ |
979 | CAKE_FLOW_FLOWS, |
980 | CAKE_FLOW_DUAL_SRC, /* = CAKE_FLOW_SRC_IP | CAKE_FLOW_FLOWS */ |
981 | CAKE_FLOW_DUAL_DST, /* = CAKE_FLOW_DST_IP | CAKE_FLOW_FLOWS */ |
982 | CAKE_FLOW_TRIPLE, /* = CAKE_FLOW_HOSTS | CAKE_FLOW_FLOWS */ |
983 | CAKE_FLOW_MAX, |
984 | }; |
985 | |
986 | enum { |
987 | CAKE_DIFFSERV_DIFFSERV3 = 0, |
988 | CAKE_DIFFSERV_DIFFSERV4, |
989 | CAKE_DIFFSERV_DIFFSERV8, |
990 | CAKE_DIFFSERV_BESTEFFORT, |
991 | CAKE_DIFFSERV_PRECEDENCE, |
992 | CAKE_DIFFSERV_MAX |
993 | }; |
994 | |
995 | enum { |
996 | CAKE_ACK_NONE = 0, |
997 | CAKE_ACK_FILTER, |
998 | CAKE_ACK_AGGRESSIVE, |
999 | CAKE_ACK_MAX |
1000 | }; |
1001 | |
1002 | enum { |
1003 | CAKE_ATM_NONE = 0, |
1004 | CAKE_ATM_ATM, |
1005 | CAKE_ATM_PTM, |
1006 | CAKE_ATM_MAX |
1007 | }; |
1008 | |
1009 | |
1010 | /* TAPRIO */ |
1011 | enum { |
1012 | TC_TAPRIO_CMD_SET_GATES = 0x00, |
1013 | TC_TAPRIO_CMD_SET_AND_HOLD = 0x01, |
1014 | TC_TAPRIO_CMD_SET_AND_RELEASE = 0x02, |
1015 | }; |
1016 | |
1017 | enum { |
1018 | TCA_TAPRIO_SCHED_ENTRY_UNSPEC, |
1019 | TCA_TAPRIO_SCHED_ENTRY_INDEX, /* u32 */ |
1020 | TCA_TAPRIO_SCHED_ENTRY_CMD, /* u8 */ |
1021 | TCA_TAPRIO_SCHED_ENTRY_GATE_MASK, /* u32 */ |
1022 | TCA_TAPRIO_SCHED_ENTRY_INTERVAL, /* u32 */ |
1023 | __TCA_TAPRIO_SCHED_ENTRY_MAX, |
1024 | }; |
1025 | #define TCA_TAPRIO_SCHED_ENTRY_MAX (__TCA_TAPRIO_SCHED_ENTRY_MAX - 1) |
1026 | |
1027 | /* The format for schedule entry list is: |
1028 | * [TCA_TAPRIO_SCHED_ENTRY_LIST] |
1029 | * [TCA_TAPRIO_SCHED_ENTRY] |
1030 | * [TCA_TAPRIO_SCHED_ENTRY_CMD] |
1031 | * [TCA_TAPRIO_SCHED_ENTRY_GATES] |
1032 | * [TCA_TAPRIO_SCHED_ENTRY_INTERVAL] |
1033 | */ |
1034 | enum { |
1035 | TCA_TAPRIO_SCHED_UNSPEC, |
1036 | TCA_TAPRIO_SCHED_ENTRY, |
1037 | __TCA_TAPRIO_SCHED_MAX, |
1038 | }; |
1039 | |
1040 | #define TCA_TAPRIO_SCHED_MAX (__TCA_TAPRIO_SCHED_MAX - 1) |
1041 | |
1042 | enum { |
1043 | TCA_TAPRIO_ATTR_UNSPEC, |
1044 | TCA_TAPRIO_ATTR_PRIOMAP, /* struct tc_mqprio_qopt */ |
1045 | TCA_TAPRIO_ATTR_SCHED_ENTRY_LIST, /* nested of entry */ |
1046 | TCA_TAPRIO_ATTR_SCHED_BASE_TIME, /* s64 */ |
1047 | TCA_TAPRIO_ATTR_SCHED_SINGLE_ENTRY, /* single entry */ |
1048 | TCA_TAPRIO_ATTR_SCHED_CLOCKID, /* s32 */ |
1049 | TCA_TAPRIO_PAD, |
1050 | __TCA_TAPRIO_ATTR_MAX, |
1051 | }; |
1052 | |
1053 | #define TCA_TAPRIO_ATTR_MAX (__TCA_TAPRIO_ATTR_MAX - 1) |
1054 | |
1055 | #endif |
1056 |
Warning: This file is not a C or C++ file. It does not have highlighting.