Commit | Line | Data |
---|---|---|
b2441318 | 1 | /* SPDX-License-Identifier: GPL-2.0 */ |
1da177e4 LT |
2 | #ifndef __NET_PKT_SCHED_H |
3 | #define __NET_PKT_SCHED_H | |
4 | ||
538e43a4 | 5 | #include <linux/jiffies.h> |
641b9e0e | 6 | #include <linux/ktime.h> |
d8b9605d | 7 | #include <linux/if_vlan.h> |
855319be | 8 | #include <linux/netdevice.h> |
1da177e4 | 9 | #include <net/sch_generic.h> |
855319be | 10 | #include <net/net_namespace.h> |
861932ec | 11 | #include <uapi/linux/pkt_sched.h> |
1da177e4 | 12 | |
d0a81f67 | 13 | #define DEFAULT_TX_QUEUE_LEN 1000 |
b193e15a | 14 | #define STAB_SIZE_LOG_MAX 30 |
d0a81f67 | 15 | |
fd2c3ef7 | 16 | struct qdisc_walker { |
1da177e4 LT |
17 | int stop; |
18 | int skip; | |
19 | int count; | |
20 | int (*fn)(struct Qdisc *, unsigned long cl, struct qdisc_walker *); | |
21 | }; | |
22 | ||
1da177e4 LT |
23 | static inline void *qdisc_priv(struct Qdisc *q) |
24 | { | |
846e463a | 25 | return &q->privdata; |
1da177e4 LT |
26 | } |
27 | ||
6e1978a9 AP |
28 | static inline struct Qdisc *qdisc_from_priv(void *priv) |
29 | { | |
30 | return container_of(priv, struct Qdisc, privdata); | |
31 | } | |
32 | ||
1da177e4 LT |
33 | /* |
34 | Timer resolution MUST BE < 10% of min_schedulable_packet_size/bandwidth | |
35 | ||
36 | Normal IP packet size ~ 512byte, hence: | |
37 | ||
38 | 0.5Kbyte/1Mbyte/sec = 0.5msec, so that we need 50usec timer for | |
39 | 10Mbit ethernet. | |
40 | ||
41 | 10msec resolution -> <50Kbit/sec. | |
42 | ||
43 | The result: [34]86 is not good choice for QoS router :-( | |
44 | ||
25985edc | 45 | The things are not so bad, because we may use artificial |
1da177e4 LT |
46 | clock evaluated by integration of network data flow |
47 | in the most critical places. | |
1da177e4 LT |
48 | */ |
49 | ||
1da177e4 LT |
50 | typedef u64 psched_time_t; |
51 | typedef long psched_tdiff_t; | |
52 | ||
a4a710c4 JP |
53 | /* Avoid doing 64 bit divide */ |
54 | #define PSCHED_SHIFT 6 | |
ca44d6e6 JP |
55 | #define PSCHED_TICKS2NS(x) ((s64)(x) << PSCHED_SHIFT) |
56 | #define PSCHED_NS2TICKS(x) ((x) >> PSCHED_SHIFT) | |
1da177e4 | 57 | |
ca44d6e6 | 58 | #define PSCHED_TICKS_PER_SEC PSCHED_NS2TICKS(NSEC_PER_SEC) |
a084980d | 59 | #define PSCHED_PASTPERFECT 0 |
1da177e4 | 60 | |
3bebcda2 PM |
61 | static inline psched_time_t psched_get_time(void) |
62 | { | |
d2de875c | 63 | return PSCHED_NS2TICKS(ktime_get_ns()); |
3bebcda2 PM |
64 | } |
65 | ||
4179477f PM |
66 | struct qdisc_watchdog { |
67 | struct hrtimer timer; | |
68 | struct Qdisc *qdisc; | |
69 | }; | |
70 | ||
860b642b VCG |
71 | void qdisc_watchdog_init_clockid(struct qdisc_watchdog *wd, struct Qdisc *qdisc, |
72 | clockid_t clockid); | |
5c15257f | 73 | void qdisc_watchdog_init(struct qdisc_watchdog *wd, struct Qdisc *qdisc); |
efe074c2 ED |
74 | |
75 | void qdisc_watchdog_schedule_range_ns(struct qdisc_watchdog *wd, u64 expires, | |
76 | u64 delta_ns); | |
77 | ||
78 | static inline void qdisc_watchdog_schedule_ns(struct qdisc_watchdog *wd, | |
79 | u64 expires) | |
80 | { | |
81 | return qdisc_watchdog_schedule_range_ns(wd, expires, 0ULL); | |
82 | } | |
34c5d292 JP |
83 | |
84 | static inline void qdisc_watchdog_schedule(struct qdisc_watchdog *wd, | |
85 | psched_time_t expires) | |
86 | { | |
45f50bed | 87 | qdisc_watchdog_schedule_ns(wd, PSCHED_TICKS2NS(expires)); |
34c5d292 JP |
88 | } |
89 | ||
5c15257f | 90 | void qdisc_watchdog_cancel(struct qdisc_watchdog *wd); |
4179477f | 91 | |
1da177e4 LT |
92 | extern struct Qdisc_ops pfifo_qdisc_ops; |
93 | extern struct Qdisc_ops bfifo_qdisc_ops; | |
57dbb2d8 | 94 | extern struct Qdisc_ops pfifo_head_drop_qdisc_ops; |
1da177e4 | 95 | |
5c15257f JP |
96 | int fifo_set_limit(struct Qdisc *q, unsigned int limit); |
97 | struct Qdisc *fifo_create_dflt(struct Qdisc *sch, struct Qdisc_ops *ops, | |
a38a9882 AA |
98 | unsigned int limit, |
99 | struct netlink_ext_ack *extack); | |
5c15257f JP |
100 | |
101 | int register_qdisc(struct Qdisc_ops *qops); | |
52327d2e | 102 | void unregister_qdisc(struct Qdisc_ops *qops); |
6da7c8fc | 103 | void qdisc_get_default(char *id, size_t len); |
104 | int qdisc_set_default(const char *id); | |
105 | ||
49b49971 | 106 | void qdisc_hash_add(struct Qdisc *q, bool invisible); |
59cc1f61 | 107 | void qdisc_hash_del(struct Qdisc *q); |
5c15257f | 108 | struct Qdisc *qdisc_lookup(struct net_device *dev, u32 handle); |
3a7d0d07 | 109 | struct Qdisc *qdisc_lookup_rcu(struct net_device *dev, u32 handle); |
5c15257f | 110 | struct qdisc_rate_table *qdisc_get_rtab(struct tc_ratespec *r, |
e9bc3fa2 AA |
111 | struct nlattr *tab, |
112 | struct netlink_ext_ack *extack); | |
5c15257f JP |
113 | void qdisc_put_rtab(struct qdisc_rate_table *tab); |
114 | void qdisc_put_stab(struct qdisc_size_table *tab); | |
6e765a00 | 115 | void qdisc_warn_nonwc(const char *txt, struct Qdisc *qdisc); |
29b86cda JF |
116 | bool sch_direct_xmit(struct sk_buff *skb, struct Qdisc *q, |
117 | struct net_device *dev, struct netdev_queue *txq, | |
118 | spinlock_t *root_lock, bool validate); | |
5c15257f JP |
119 | |
120 | void __qdisc_run(struct Qdisc *q); | |
1da177e4 | 121 | |
37437bb2 | 122 | static inline void qdisc_run(struct Qdisc *q) |
1da177e4 | 123 | { |
6c148184 | 124 | if (qdisc_run_begin(q)) { |
102b55ee | 125 | __qdisc_run(q); |
6c148184 JF |
126 | qdisc_run_end(q); |
127 | } | |
d8b9605d JP |
128 | } |
129 | ||
1da177e4 LT |
130 | /* Calculate maximal size of packet seen by hard_start_xmit |
131 | routine of this device. | |
132 | */ | |
95c96174 | 133 | static inline unsigned int psched_mtu(const struct net_device *dev) |
1da177e4 | 134 | { |
3b04ddde | 135 | return dev->mtu + dev->hard_header_len; |
1da177e4 LT |
136 | } |
137 | ||
855319be JP |
138 | static inline struct net *qdisc_net(struct Qdisc *q) |
139 | { | |
140 | return dev_net(q->dev_queue->dev); | |
141 | } | |
142 | ||
aac4daa8 VO |
143 | struct tc_query_caps_base { |
144 | enum tc_setup_type type; | |
145 | void *caps; | |
146 | }; | |
147 | ||
3d0bd028 VCG |
148 | struct tc_cbs_qopt_offload { |
149 | u8 enable; | |
150 | s32 queue; | |
151 | s32 hicredit; | |
152 | s32 locredit; | |
153 | s32 idleslope; | |
154 | s32 sendslope; | |
155 | }; | |
156 | ||
88cab771 JSP |
157 | struct tc_etf_qopt_offload { |
158 | u8 enable; | |
159 | s32 queue; | |
160 | }; | |
161 | ||
19278d76 VO |
162 | struct tc_mqprio_caps { |
163 | bool validate_queue_counts:1; | |
164 | }; | |
165 | ||
9adafe2b VO |
166 | struct tc_mqprio_qopt_offload { |
167 | /* struct tc_mqprio_qopt must always be the first element */ | |
168 | struct tc_mqprio_qopt qopt; | |
c54876cd | 169 | struct netlink_ext_ack *extack; |
9adafe2b VO |
170 | u16 mode; |
171 | u16 shaper; | |
172 | u32 flags; | |
173 | u64 min_rate[TC_QOPT_MAX_QUEUE]; | |
174 | u64 max_rate[TC_QOPT_MAX_QUEUE]; | |
f62af20b | 175 | unsigned long preemptible_tcs; |
9adafe2b VO |
176 | }; |
177 | ||
a54fc09e VO |
178 | struct tc_taprio_caps { |
179 | bool supports_queue_max_sdu:1; | |
522d15ea | 180 | bool gate_mask_per_txq:1; |
2f530df7 VO |
181 | /* Device expects lower TXQ numbers to have higher priority over higher |
182 | * TXQs, regardless of their TC mapping. DO NOT USE FOR NEW DRIVERS, | |
183 | * INSTEAD ENFORCE A PROPER TC:TXQ MAPPING COMING FROM USER SPACE. | |
184 | */ | |
185 | bool broken_mqprio:1; | |
a54fc09e VO |
186 | }; |
187 | ||
9c66d156 VCG |
188 | struct tc_taprio_sched_entry { |
189 | u8 command; /* TC_TAPRIO_CMD_* */ | |
190 | ||
191 | /* The gate_mask in the offloading side refers to traffic classes */ | |
192 | u32 gate_mask; | |
193 | u32 interval; | |
194 | }; | |
195 | ||
196 | struct tc_taprio_qopt_offload { | |
09c794c0 | 197 | struct tc_mqprio_qopt_offload mqprio; |
c54876cd | 198 | struct netlink_ext_ack *extack; |
9c66d156 VCG |
199 | u8 enable; |
200 | ktime_t base_time; | |
201 | u64 cycle_time; | |
202 | u64 cycle_time_extension; | |
a54fc09e | 203 | u32 max_sdu[TC_MAX_QUEUE]; |
9c66d156 VCG |
204 | |
205 | size_t num_entries; | |
b90feaff | 206 | struct tc_taprio_sched_entry entries[]; |
9c66d156 VCG |
207 | }; |
208 | ||
d7be266a VO |
209 | #if IS_ENABLED(CONFIG_NET_SCH_TAPRIO) |
210 | ||
9c66d156 VCG |
211 | /* Reference counting */ |
212 | struct tc_taprio_qopt_offload *taprio_offload_get(struct tc_taprio_qopt_offload | |
213 | *offload); | |
214 | void taprio_offload_free(struct tc_taprio_qopt_offload *offload); | |
215 | ||
d7be266a VO |
216 | #else |
217 | ||
218 | /* Reference counting */ | |
219 | static inline struct tc_taprio_qopt_offload * | |
220 | taprio_offload_get(struct tc_taprio_qopt_offload *offload) | |
221 | { | |
222 | return NULL; | |
223 | } | |
224 | ||
225 | static inline void taprio_offload_free(struct tc_taprio_qopt_offload *offload) | |
226 | { | |
227 | } | |
228 | ||
229 | #endif | |
230 | ||
847cbfc0 VO |
231 | /* Ensure skb_mstamp_ns, which might have been populated with the txtime, is |
232 | * not mistaken for a software timestamp, because this will otherwise prevent | |
233 | * the dispatch of hardware timestamps to the socket. | |
234 | */ | |
235 | static inline void skb_txtime_consumed(struct sk_buff *skb) | |
236 | { | |
237 | skb->tstamp = ktime_set(0, 0); | |
238 | } | |
239 | ||
ec624fe7 PB |
240 | struct tc_skb_cb { |
241 | struct qdisc_skb_cb qdisc_cb; | |
242 | ||
243 | u16 mru; | |
6f022c2d PB |
244 | u8 post_ct:1; |
245 | u8 post_ct_snat:1; | |
246 | u8 post_ct_dnat:1; | |
38495958 | 247 | u16 zone; /* Only valid if post_ct = true */ |
ec624fe7 PB |
248 | }; |
249 | ||
250 | static inline struct tc_skb_cb *tc_skb_cb(const struct sk_buff *skb) | |
251 | { | |
252 | struct tc_skb_cb *cb = (struct tc_skb_cb *)skb->cb; | |
253 | ||
254 | BUILD_BUG_ON(sizeof(*cb) > sizeof_field(struct sk_buff, cb)); | |
255 | return cb; | |
256 | } | |
257 | ||
d7a68e56 ZS |
258 | static inline bool tc_qdisc_stats_dump(struct Qdisc *sch, |
259 | unsigned long cl, | |
260 | struct qdisc_walker *arg) | |
261 | { | |
262 | if (arg->count >= arg->skip && arg->fn(sch, cl, arg) < 0) { | |
263 | arg->stop = 1; | |
264 | return false; | |
265 | } | |
266 | ||
267 | arg->count++; | |
268 | return true; | |
269 | } | |
270 | ||
1da177e4 | 271 | #endif |