1#ifndef __LINUX_PKT_SCHED_H
2#define __LINUX_PKT_SCHED_H
3
4#include <linux/types.h>
5
6/* Logical priority bands not depending on specific packet scheduler.
7   Every scheduler will map them to real traffic classes, if it has
8   no more precise mechanism to classify packets.
9
10   These numbers have no special meaning, though their coincidence
11   with obsolete IPv6 values is not occasional :-). New IPv6 drafts
12   preferred full anarchy inspired by diffserv group.
13
14   Note: TC_PRIO_BESTEFFORT does not mean that it is the most unhappy
15   class, actually, as rule it will be handled with more care than
16   filler or even bulk.
17 */
18
19#define TC_PRIO_BESTEFFORT		0
20#define TC_PRIO_FILLER			1
21#define TC_PRIO_BULK			2
22#define TC_PRIO_INTERACTIVE_BULK	4
23#define TC_PRIO_INTERACTIVE		6
24#define TC_PRIO_CONTROL			7
25
26#define TC_PRIO_MAX			15
27
28/* Generic queue statistics, available for all the elements.
29   Particular schedulers may have also their private records.
30 */
31
32struct tc_stats {
33	__u64	bytes;			/* Number of enqueued bytes */
34	__u32	packets;		/* Number of enqueued packets	*/
35	__u32	drops;			/* Packets dropped because of lack of resources */
36	__u32	overlimits;		/* Number of throttle events when this
37					 * flow goes out of allocated bandwidth */
38	__u32	bps;			/* Current flow byte rate */
39	__u32	pps;			/* Current flow packet rate */
40	__u32	qlen;
41	__u32	backlog;
42};
43
44struct tc_estimator {
45	signed char	interval;
46	unsigned char	ewma_log;
47};
48
49/* "Handles"
50   ---------
51
52    All the traffic control objects have 32bit identifiers, or "handles".
53
54    They can be considered as opaque numbers from user API viewpoint,
55    but actually they always consist of two fields: major and
56    minor numbers, which are interpreted by kernel specially,
57    that may be used by applications, though not recommended.
58
59    F.e. qdisc handles always have minor number equal to zero,
60    classes (or flows) have major equal to parent qdisc major, and
61    minor uniquely identifying class inside qdisc.
62
63    Macros to manipulate handles:
64 */
65
66#define TC_H_MAJ_MASK (0xFFFF0000U)
67#define TC_H_MIN_MASK (0x0000FFFFU)
68#define TC_H_MAJ(h) ((h)&TC_H_MAJ_MASK)
69#define TC_H_MIN(h) ((h)&TC_H_MIN_MASK)
70#define TC_H_MAKE(maj,min) (((maj)&TC_H_MAJ_MASK)|((min)&TC_H_MIN_MASK))
71
72#define TC_H_UNSPEC	(0U)
73#define TC_H_ROOT	(0xFFFFFFFFU)
74#define TC_H_INGRESS    (0xFFFFFFF1U)
75
76/* Need to corrospond to iproute2 tc/tc_core.h "enum link_layer" */
77enum tc_link_layer {
78	TC_LINKLAYER_UNAWARE, /* Indicate unaware old iproute2 util */
79	TC_LINKLAYER_ETHERNET,
80	TC_LINKLAYER_ATM,
81};
82#define TC_LINKLAYER_MASK 0x0F /* limit use to lower 4 bits */
83
84struct tc_ratespec {
85	unsigned char	cell_log;
86	__u8		linklayer; /* lower 4 bits */
87	unsigned short	overhead;
88	short		cell_align;
89	unsigned short	mpu;
90	__u32		rate;
91};
92
93#define TC_RTAB_SIZE	1024
94
95struct tc_sizespec {
96	unsigned char	cell_log;
97	unsigned char	size_log;
98	short		cell_align;
99	int		overhead;
100	unsigned int	linklayer;
101	unsigned int	mpu;
102	unsigned int	mtu;
103	unsigned int	tsize;
104};
105
106enum {
107	TCA_STAB_UNSPEC,
108	TCA_STAB_BASE,
109	TCA_STAB_DATA,
110	__TCA_STAB_MAX
111};
112
113#define TCA_STAB_MAX (__TCA_STAB_MAX - 1)
114
115/* FIFO section */
116
117struct tc_fifo_qopt {
118	__u32	limit;	/* Queue length: bytes for bfifo, packets for pfifo */
119};
120
121/* PRIO section */
122
123#define TCQ_PRIO_BANDS	16
124#define TCQ_MIN_PRIO_BANDS 2
125
126struct tc_prio_qopt {
127	int	bands;			/* Number of bands */
128	__u8	priomap[TC_PRIO_MAX+1];	/* Map: logical priority -> PRIO band */
129};
130
131/* MULTIQ section */
132
133struct tc_multiq_qopt {
134	__u16	bands;			/* Number of bands */
135	__u16	max_bands;		/* Maximum number of queues */
136};
137
138/* PLUG section */
139
140#define TCQ_PLUG_BUFFER                0
141#define TCQ_PLUG_RELEASE_ONE           1
142#define TCQ_PLUG_RELEASE_INDEFINITE    2
143#define TCQ_PLUG_LIMIT                 3
144
145struct tc_plug_qopt {
146	/* TCQ_PLUG_BUFFER: Inset a plug into the queue and
147	 *  buffer any incoming packets
148	 * TCQ_PLUG_RELEASE_ONE: Dequeue packets from queue head
149	 *   to beginning of the next plug.
150	 * TCQ_PLUG_RELEASE_INDEFINITE: Dequeue all packets from queue.
151	 *   Stop buffering packets until the next TCQ_PLUG_BUFFER
152	 *   command is received (just act as a pass-thru queue).
153	 * TCQ_PLUG_LIMIT: Increase/decrease queue size
154	 */
155	int             action;
156	__u32           limit;
157};
158
159/* TBF section */
160
161struct tc_tbf_qopt {
162	struct tc_ratespec rate;
163	struct tc_ratespec peakrate;
164	__u32		limit;
165	__u32		buffer;
166	__u32		mtu;
167};
168
169enum {
170	TCA_TBF_UNSPEC,
171	TCA_TBF_PARMS,
172	TCA_TBF_RTAB,
173	TCA_TBF_PTAB,
174	TCA_TBF_RATE64,
175	TCA_TBF_PRATE64,
176	TCA_TBF_BURST,
177	TCA_TBF_PBURST,
178	__TCA_TBF_MAX,
179};
180
181#define TCA_TBF_MAX (__TCA_TBF_MAX - 1)
182
183
184/* TEQL section */
185
186/* TEQL does not require any parameters */
187
188/* SFQ section */
189
190struct tc_sfq_qopt {
191	unsigned	quantum;	/* Bytes per round allocated to flow */
192	int		perturb_period;	/* Period of hash perturbation */
193	__u32		limit;		/* Maximal packets in queue */
194	unsigned	divisor;	/* Hash divisor  */
195	unsigned	flows;		/* Maximal number of flows  */
196};
197
198struct tc_sfqred_stats {
199	__u32           prob_drop;      /* Early drops, below max threshold */
200	__u32           forced_drop;	/* Early drops, after max threshold */
201	__u32           prob_mark;      /* Marked packets, below max threshold */
202	__u32           forced_mark;    /* Marked packets, after max threshold */
203	__u32           prob_mark_head; /* Marked packets, below max threshold */
204	__u32           forced_mark_head;/* Marked packets, after max threshold */
205};
206
207struct tc_sfq_qopt_v1 {
208	struct tc_sfq_qopt v0;
209	unsigned int	depth;		/* max number of packets per flow */
210	unsigned int	headdrop;
211/* SFQRED parameters */
212	__u32		limit;		/* HARD maximal flow queue length (bytes) */
213	__u32		qth_min;	/* Min average length threshold (bytes) */
214	__u32		qth_max;	/* Max average length threshold (bytes) */
215	unsigned char   Wlog;		/* log(W)		*/
216	unsigned char   Plog;		/* log(P_max/(qth_max-qth_min))	*/
217	unsigned char   Scell_log;	/* cell size for idle damping */
218	unsigned char	flags;
219	__u32		max_P;		/* probability, high resolution */
220/* SFQRED stats */
221	struct tc_sfqred_stats stats;
222};
223
224
225struct tc_sfq_xstats {
226	__s32		allot;
227};
228
229/* RED section */
230
231enum {
232	TCA_RED_UNSPEC,
233	TCA_RED_PARMS,
234	TCA_RED_STAB,
235	TCA_RED_MAX_P,
236	__TCA_RED_MAX,
237};
238
239#define TCA_RED_MAX (__TCA_RED_MAX - 1)
240
241struct tc_red_qopt {
242	__u32		limit;		/* HARD maximal queue length (bytes)	*/
243	__u32		qth_min;	/* Min average length threshold (bytes) */
244	__u32		qth_max;	/* Max average length threshold (bytes) */
245	unsigned char   Wlog;		/* log(W)		*/
246	unsigned char   Plog;		/* log(P_max/(qth_max-qth_min))	*/
247	unsigned char   Scell_log;	/* cell size for idle damping */
248	unsigned char	flags;
249#define TC_RED_ECN		1
250#define TC_RED_HARDDROP		2
251#define TC_RED_ADAPTATIVE	4
252};
253
254struct tc_red_xstats {
255	__u32           early;          /* Early drops */
256	__u32           pdrop;          /* Drops due to queue limits */
257	__u32           other;          /* Drops due to drop() calls */
258	__u32           marked;         /* Marked packets */
259};
260
261/* GRED section */
262
263#define MAX_DPs 16
264
265enum {
266       TCA_GRED_UNSPEC,
267       TCA_GRED_PARMS,
268       TCA_GRED_STAB,
269       TCA_GRED_DPS,
270       TCA_GRED_MAX_P,
271	   __TCA_GRED_MAX,
272};
273
274#define TCA_GRED_MAX (__TCA_GRED_MAX - 1)
275
276struct tc_gred_qopt {
277	__u32		limit;        /* HARD maximal queue length (bytes)    */
278	__u32		qth_min;      /* Min average length threshold (bytes) */
279	__u32		qth_max;      /* Max average length threshold (bytes) */
280	__u32		DP;           /* up to 2^32 DPs */
281	__u32		backlog;
282	__u32		qave;
283	__u32		forced;
284	__u32		early;
285	__u32		other;
286	__u32		pdrop;
287	__u8		Wlog;         /* log(W)               */
288	__u8		Plog;         /* log(P_max/(qth_max-qth_min)) */
289	__u8		Scell_log;    /* cell size for idle damping */
290	__u8		prio;         /* prio of this VQ */
291	__u32		packets;
292	__u32		bytesin;
293};
294
295/* gred setup */
296struct tc_gred_sopt {
297	__u32		DPs;
298	__u32		def_DP;
299	__u8		grio;
300	__u8		flags;
301	__u16		pad1;
302};
303
304/* CHOKe section */
305
306enum {
307	TCA_CHOKE_UNSPEC,
308	TCA_CHOKE_PARMS,
309	TCA_CHOKE_STAB,
310	TCA_CHOKE_MAX_P,
311	__TCA_CHOKE_MAX,
312};
313
314#define TCA_CHOKE_MAX (__TCA_CHOKE_MAX - 1)
315
316struct tc_choke_qopt {
317	__u32		limit;		/* Hard queue length (packets)	*/
318	__u32		qth_min;	/* Min average threshold (packets) */
319	__u32		qth_max;	/* Max average threshold (packets) */
320	unsigned char   Wlog;		/* log(W)		*/
321	unsigned char   Plog;		/* log(P_max/(qth_max-qth_min))	*/
322	unsigned char   Scell_log;	/* cell size for idle damping */
323	unsigned char	flags;		/* see RED flags */
324};
325
326struct tc_choke_xstats {
327	__u32		early;          /* Early drops */
328	__u32		pdrop;          /* Drops due to queue limits */
329	__u32		other;          /* Drops due to drop() calls */
330	__u32		marked;         /* Marked packets */
331	__u32		matched;	/* Drops due to flow match */
332};
333
334/* HTB section */
335#define TC_HTB_NUMPRIO		8
336#define TC_HTB_MAXDEPTH		8
337#define TC_HTB_PROTOVER		3 /* the same as HTB and TC's major */
338
339struct tc_htb_opt {
340	struct tc_ratespec 	rate;
341	struct tc_ratespec 	ceil;
342	__u32	buffer;
343	__u32	cbuffer;
344	__u32	quantum;
345	__u32	level;		/* out only */
346	__u32	prio;
347};
348struct tc_htb_glob {
349	__u32 version;		/* to match HTB/TC */
350    	__u32 rate2quantum;	/* bps->quantum divisor */
351    	__u32 defcls;		/* default class number */
352	__u32 debug;		/* debug flags */
353
354	/* stats */
355	__u32 direct_pkts; /* count of non shaped packets */
356};
357enum {
358	TCA_HTB_UNSPEC,
359	TCA_HTB_PARMS,
360	TCA_HTB_INIT,
361	TCA_HTB_CTAB,
362	TCA_HTB_RTAB,
363	TCA_HTB_DIRECT_QLEN,
364	TCA_HTB_RATE64,
365	TCA_HTB_CEIL64,
366	__TCA_HTB_MAX,
367};
368
369#define TCA_HTB_MAX (__TCA_HTB_MAX - 1)
370
371struct tc_htb_xstats {
372	__u32 lends;
373	__u32 borrows;
374	__u32 giants;	/* too big packets (rate will not be accurate) */
375	__u32 tokens;
376	__u32 ctokens;
377};
378
379/* HFSC section */
380
381struct tc_hfsc_qopt {
382	__u16	defcls;		/* default class */
383};
384
385struct tc_service_curve {
386	__u32	m1;		/* slope of the first segment in bps */
387	__u32	d;		/* x-projection of the first segment in us */
388	__u32	m2;		/* slope of the second segment in bps */
389};
390
391struct tc_hfsc_stats {
392	__u64	work;		/* total work done */
393	__u64	rtwork;		/* work done by real-time criteria */
394	__u32	period;		/* current period */
395	__u32	level;		/* class level in hierarchy */
396};
397
398enum {
399	TCA_HFSC_UNSPEC,
400	TCA_HFSC_RSC,
401	TCA_HFSC_FSC,
402	TCA_HFSC_USC,
403	__TCA_HFSC_MAX,
404};
405
406#define TCA_HFSC_MAX (__TCA_HFSC_MAX - 1)
407
408
409/* CBQ section */
410
411#define TC_CBQ_MAXPRIO		8
412#define TC_CBQ_MAXLEVEL		8
413#define TC_CBQ_DEF_EWMA		5
414
415struct tc_cbq_lssopt {
416	unsigned char	change;
417	unsigned char	flags;
418#define TCF_CBQ_LSS_BOUNDED	1
419#define TCF_CBQ_LSS_ISOLATED	2
420	unsigned char  	ewma_log;
421	unsigned char  	level;
422#define TCF_CBQ_LSS_FLAGS	1
423#define TCF_CBQ_LSS_EWMA	2
424#define TCF_CBQ_LSS_MAXIDLE	4
425#define TCF_CBQ_LSS_MINIDLE	8
426#define TCF_CBQ_LSS_OFFTIME	0x10
427#define TCF_CBQ_LSS_AVPKT	0x20
428	__u32		maxidle;
429	__u32		minidle;
430	__u32		offtime;
431	__u32		avpkt;
432};
433
434struct tc_cbq_wrropt {
435	unsigned char	flags;
436	unsigned char	priority;
437	unsigned char	cpriority;
438	unsigned char	__reserved;
439	__u32		allot;
440	__u32		weight;
441};
442
443struct tc_cbq_ovl {
444	unsigned char	strategy;
445#define	TC_CBQ_OVL_CLASSIC	0
446#define	TC_CBQ_OVL_DELAY	1
447#define	TC_CBQ_OVL_LOWPRIO	2
448#define	TC_CBQ_OVL_DROP		3
449#define	TC_CBQ_OVL_RCLASSIC	4
450	unsigned char	priority2;
451	__u16		pad;
452	__u32		penalty;
453};
454
455struct tc_cbq_police {
456	unsigned char	police;
457	unsigned char	__res1;
458	unsigned short	__res2;
459};
460
461struct tc_cbq_fopt {
462	__u32		split;
463	__u32		defmap;
464	__u32		defchange;
465};
466
467struct tc_cbq_xstats {
468	__u32		borrows;
469	__u32		overactions;
470	__s32		avgidle;
471	__s32		undertime;
472};
473
474enum {
475	TCA_CBQ_UNSPEC,
476	TCA_CBQ_LSSOPT,
477	TCA_CBQ_WRROPT,
478	TCA_CBQ_FOPT,
479	TCA_CBQ_OVL_STRATEGY,
480	TCA_CBQ_RATE,
481	TCA_CBQ_RTAB,
482	TCA_CBQ_POLICE,
483	__TCA_CBQ_MAX,
484};
485
486#define TCA_CBQ_MAX	(__TCA_CBQ_MAX - 1)
487
488/* dsmark section */
489
490enum {
491	TCA_DSMARK_UNSPEC,
492	TCA_DSMARK_INDICES,
493	TCA_DSMARK_DEFAULT_INDEX,
494	TCA_DSMARK_SET_TC_INDEX,
495	TCA_DSMARK_MASK,
496	TCA_DSMARK_VALUE,
497	__TCA_DSMARK_MAX,
498};
499
500#define TCA_DSMARK_MAX (__TCA_DSMARK_MAX - 1)
501
502/* ATM  section */
503
504enum {
505	TCA_ATM_UNSPEC,
506	TCA_ATM_FD,		/* file/socket descriptor */
507	TCA_ATM_PTR,		/* pointer to descriptor - later */
508	TCA_ATM_HDR,		/* LL header */
509	TCA_ATM_EXCESS,		/* excess traffic class (0 for CLP)  */
510	TCA_ATM_ADDR,		/* PVC address (for output only) */
511	TCA_ATM_STATE,		/* VC state (ATM_VS_*; for output only) */
512	__TCA_ATM_MAX,
513};
514
515#define TCA_ATM_MAX	(__TCA_ATM_MAX - 1)
516
517/* Network emulator */
518
519enum {
520	TCA_NETEM_UNSPEC,
521	TCA_NETEM_CORR,
522	TCA_NETEM_DELAY_DIST,
523	TCA_NETEM_REORDER,
524	TCA_NETEM_CORRUPT,
525	TCA_NETEM_LOSS,
526	TCA_NETEM_RATE,
527	TCA_NETEM_ECN,
528	TCA_NETEM_RATE64,
529	__TCA_NETEM_MAX,
530};
531
532#define TCA_NETEM_MAX (__TCA_NETEM_MAX - 1)
533
534struct tc_netem_qopt {
535	__u32	latency;	/* added delay (us) */
536	__u32   limit;		/* fifo limit (packets) */
537	__u32	loss;		/* random packet loss (0=none ~0=100%) */
538	__u32	gap;		/* re-ordering gap (0 for none) */
539	__u32   duplicate;	/* random packet dup  (0=none ~0=100%) */
540	__u32	jitter;		/* random jitter in latency (us) */
541};
542
543struct tc_netem_corr {
544	__u32	delay_corr;	/* delay correlation */
545	__u32	loss_corr;	/* packet loss correlation */
546	__u32	dup_corr;	/* duplicate correlation  */
547};
548
549struct tc_netem_reorder {
550	__u32	probability;
551	__u32	correlation;
552};
553
554struct tc_netem_corrupt {
555	__u32	probability;
556	__u32	correlation;
557};
558
559struct tc_netem_rate {
560	__u32	rate;	/* byte/s */
561	__s32	packet_overhead;
562	__u32	cell_size;
563	__s32	cell_overhead;
564};
565
566enum {
567	NETEM_LOSS_UNSPEC,
568	NETEM_LOSS_GI,		/* General Intuitive - 4 state model */
569	NETEM_LOSS_GE,		/* Gilbert Elliot models */
570	__NETEM_LOSS_MAX
571};
572#define NETEM_LOSS_MAX (__NETEM_LOSS_MAX - 1)
573
574/* State transition probabilities for 4 state model */
575struct tc_netem_gimodel {
576	__u32	p13;
577	__u32	p31;
578	__u32	p32;
579	__u32	p14;
580	__u32	p23;
581};
582
583/* Gilbert-Elliot models */
584struct tc_netem_gemodel {
585	__u32 p;
586	__u32 r;
587	__u32 h;
588	__u32 k1;
589};
590
591#define NETEM_DIST_SCALE	8192
592#define NETEM_DIST_MAX		16384
593
594/* DRR */
595
596enum {
597	TCA_DRR_UNSPEC,
598	TCA_DRR_QUANTUM,
599	__TCA_DRR_MAX
600};
601
602#define TCA_DRR_MAX	(__TCA_DRR_MAX - 1)
603
604struct tc_drr_stats {
605	__u32	deficit;
606};
607
608/* MQPRIO */
609#define TC_QOPT_BITMASK 15
610#define TC_QOPT_MAX_QUEUE 16
611
612struct tc_mqprio_qopt {
613	__u8	num_tc;
614	__u8	prio_tc_map[TC_QOPT_BITMASK + 1];
615	__u8	hw;
616	__u16	count[TC_QOPT_MAX_QUEUE];
617	__u16	offset[TC_QOPT_MAX_QUEUE];
618};
619
620/* SFB */
621
622enum {
623	TCA_SFB_UNSPEC,
624	TCA_SFB_PARMS,
625	__TCA_SFB_MAX,
626};
627
628#define TCA_SFB_MAX (__TCA_SFB_MAX - 1)
629
630/*
631 * Note: increment, decrement are Q0.16 fixed-point values.
632 */
633struct tc_sfb_qopt {
634	__u32 rehash_interval;	/* delay between hash move, in ms */
635	__u32 warmup_time;	/* double buffering warmup time in ms (warmup_time < rehash_interval) */
636	__u32 max;		/* max len of qlen_min */
637	__u32 bin_size;		/* maximum queue length per bin */
638	__u32 increment;	/* probability increment, (d1 in Blue) */
639	__u32 decrement;	/* probability decrement, (d2 in Blue) */
640	__u32 limit;		/* max SFB queue length */
641	__u32 penalty_rate;	/* inelastic flows are rate limited to 'rate' pps */
642	__u32 penalty_burst;
643};
644
645struct tc_sfb_xstats {
646	__u32 earlydrop;
647	__u32 penaltydrop;
648	__u32 bucketdrop;
649	__u32 queuedrop;
650	__u32 childdrop; /* drops in child qdisc */
651	__u32 marked;
652	__u32 maxqlen;
653	__u32 maxprob;
654	__u32 avgprob;
655};
656
657#define SFB_MAX_PROB 0xFFFF
658
659/* QFQ */
660enum {
661	TCA_QFQ_UNSPEC,
662	TCA_QFQ_WEIGHT,
663	TCA_QFQ_LMAX,
664	__TCA_QFQ_MAX
665};
666
667#define TCA_QFQ_MAX	(__TCA_QFQ_MAX - 1)
668
669struct tc_qfq_stats {
670	__u32 weight;
671	__u32 lmax;
672};
673
674/* CODEL */
675
676enum {
677	TCA_CODEL_UNSPEC,
678	TCA_CODEL_TARGET,
679	TCA_CODEL_LIMIT,
680	TCA_CODEL_INTERVAL,
681	TCA_CODEL_ECN,
682	__TCA_CODEL_MAX
683};
684
685#define TCA_CODEL_MAX	(__TCA_CODEL_MAX - 1)
686
687struct tc_codel_xstats {
688	__u32	maxpacket; /* largest packet we've seen so far */
689	__u32	count;	   /* how many drops we've done since the last time we
690			    * entered dropping state
691			    */
692	__u32	lastcount; /* count at entry to dropping state */
693	__u32	ldelay;    /* in-queue delay seen by most recently dequeued packet */
694	__s32	drop_next; /* time to drop next packet */
695	__u32	drop_overlimit; /* number of time max qdisc packet limit was hit */
696	__u32	ecn_mark;  /* number of packets we ECN marked instead of dropped */
697	__u32	dropping;  /* are we in dropping state ? */
698};
699
700/* FQ_CODEL */
701
702enum {
703	TCA_FQ_CODEL_UNSPEC,
704	TCA_FQ_CODEL_TARGET,
705	TCA_FQ_CODEL_LIMIT,
706	TCA_FQ_CODEL_INTERVAL,
707	TCA_FQ_CODEL_ECN,
708	TCA_FQ_CODEL_FLOWS,
709	TCA_FQ_CODEL_QUANTUM,
710	__TCA_FQ_CODEL_MAX
711};
712
713#define TCA_FQ_CODEL_MAX	(__TCA_FQ_CODEL_MAX - 1)
714
715enum {
716	TCA_FQ_CODEL_XSTATS_QDISC,
717	TCA_FQ_CODEL_XSTATS_CLASS,
718};
719
720struct tc_fq_codel_qd_stats {
721	__u32	maxpacket;	/* largest packet we've seen so far */
722	__u32	drop_overlimit; /* number of time max qdisc
723				 * packet limit was hit
724				 */
725	__u32	ecn_mark;	/* number of packets we ECN marked
726				 * instead of being dropped
727				 */
728	__u32	new_flow_count; /* number of time packets
729				 * created a 'new flow'
730				 */
731	__u32	new_flows_len;	/* count of flows in new list */
732	__u32	old_flows_len;	/* count of flows in old list */
733};
734
735struct tc_fq_codel_cl_stats {
736	__s32	deficit;
737	__u32	ldelay;		/* in-queue delay seen by most recently
738				 * dequeued packet
739				 */
740	__u32	count;
741	__u32	lastcount;
742	__u32	dropping;
743	__s32	drop_next;
744};
745
746struct tc_fq_codel_xstats {
747	__u32	type;
748	union {
749		struct tc_fq_codel_qd_stats qdisc_stats;
750		struct tc_fq_codel_cl_stats class_stats;
751	};
752};
753
754/* FQ */
755
756enum {
757	TCA_FQ_UNSPEC,
758
759	TCA_FQ_PLIMIT,		/* limit of total number of packets in queue */
760
761	TCA_FQ_FLOW_PLIMIT,	/* limit of packets per flow */
762
763	TCA_FQ_QUANTUM,		/* RR quantum */
764
765	TCA_FQ_INITIAL_QUANTUM,		/* RR quantum for new flow */
766
767	TCA_FQ_RATE_ENABLE,	/* enable/disable rate limiting */
768
769	TCA_FQ_FLOW_DEFAULT_RATE,/* obsolete, do not use */
770
771	TCA_FQ_FLOW_MAX_RATE,	/* per flow max rate */
772
773	TCA_FQ_BUCKETS_LOG,	/* log2(number of buckets) */
774
775	TCA_FQ_FLOW_REFILL_DELAY,	/* flow credit refill delay in usec */
776
777	TCA_FQ_ORPHAN_MASK,	/* mask applied to orphaned skb hashes */
778
779	__TCA_FQ_MAX
780};
781
782#define TCA_FQ_MAX	(__TCA_FQ_MAX - 1)
783
784struct tc_fq_qd_stats {
785	__u64	gc_flows;
786	__u64	highprio_packets;
787	__u64	tcp_retrans;
788	__u64	throttled;
789	__u64	flows_plimit;
790	__u64	pkts_too_long;
791	__u64	allocation_errors;
792	__s64	time_next_delayed_flow;
793	__u32	flows;
794	__u32	inactive_flows;
795	__u32	throttled_flows;
796	__u32	pad;
797};
798
799/* Heavy-Hitter Filter */
800
801enum {
802	TCA_HHF_UNSPEC,
803	TCA_HHF_BACKLOG_LIMIT,
804	TCA_HHF_QUANTUM,
805	TCA_HHF_HH_FLOWS_LIMIT,
806	TCA_HHF_RESET_TIMEOUT,
807	TCA_HHF_ADMIT_BYTES,
808	TCA_HHF_EVICT_TIMEOUT,
809	TCA_HHF_NON_HH_WEIGHT,
810	__TCA_HHF_MAX
811};
812
813#define TCA_HHF_MAX	(__TCA_HHF_MAX - 1)
814
815struct tc_hhf_xstats {
816	__u32	drop_overlimit; /* number of times max qdisc packet limit
817				 * was hit
818				 */
819	__u32	hh_overlimit;   /* number of times max heavy-hitters was hit */
820	__u32	hh_tot_count;   /* number of captured heavy-hitters so far */
821	__u32	hh_cur_count;   /* number of current heavy-hitters */
822};
823
824/* PIE */
825enum {
826	TCA_PIE_UNSPEC,
827	TCA_PIE_TARGET,
828	TCA_PIE_LIMIT,
829	TCA_PIE_TUPDATE,
830	TCA_PIE_ALPHA,
831	TCA_PIE_BETA,
832	TCA_PIE_ECN,
833	TCA_PIE_BYTEMODE,
834	__TCA_PIE_MAX
835};
836#define TCA_PIE_MAX   (__TCA_PIE_MAX - 1)
837
838struct tc_pie_xstats {
839	__u32 prob;             /* current probability */
840	__u32 delay;            /* current delay in ms */
841	__u32 avg_dq_rate;      /* current average dq_rate in bits/pie_time */
842	__u32 packets_in;       /* total number of packets enqueued */
843	__u32 dropped;          /* packets dropped due to pie_action */
844	__u32 overlimit;        /* dropped due to lack of space in queue */
845	__u32 maxq;             /* maximum queue size */
846	__u32 ecn_mark;         /* packets marked with ecn*/
847};
848#endif
849