Commit | Line | Data |
---|---|---|
2874c5fd | 1 | /* SPDX-License-Identifier: GPL-2.0-or-later */ |
1da177e4 LT |
2 | /* |
3 | * VLAN An implementation of 802.1Q VLAN tagging. | |
4 | * | |
5 | * Authors: Ben Greear <greearb@candelatech.com> | |
1da177e4 | 6 | */ |
1da177e4 LT |
7 | #ifndef _LINUX_IF_VLAN_H_ |
8 | #define _LINUX_IF_VLAN_H_ | |
9 | ||
1da177e4 | 10 | #include <linux/netdevice.h> |
0610d11b | 11 | #include <linux/etherdevice.h> |
65ac6a5f | 12 | #include <linux/rtnetlink.h> |
187f1882 | 13 | #include <linux/bug.h> |
607ca46e | 14 | #include <uapi/linux/if_vlan.h> |
1da177e4 | 15 | |
a3f671b3 JP |
16 | #define VLAN_HLEN 4 /* The additional bytes required by VLAN |
17 | * (in addition to the Ethernet header) | |
1da177e4 | 18 | */ |
1da177e4 LT |
19 | #define VLAN_ETH_HLEN 18 /* Total octets in header. */ |
20 | #define VLAN_ETH_ZLEN 64 /* Min. octets in frame sans FCS */ | |
21 | ||
22 | /* | |
23 | * According to 802.3ac, the packet can be 4 bytes longer. --Klika Jan | |
24 | */ | |
25 | #define VLAN_ETH_DATA_LEN 1500 /* Max. octets in payload */ | |
26 | #define VLAN_ETH_FRAME_LEN 1518 /* Max. octets in frame sans FCS */ | |
27 | ||
469acedd THJ |
28 | #define VLAN_MAX_DEPTH 8 /* Max. number of nested VLAN tags parsed */ |
29 | ||
740c15d0 PM |
30 | /* |
31 | * struct vlan_hdr - vlan header | |
32 | * @h_vlan_TCI: priority and VLAN ID | |
33 | * @h_vlan_encapsulated_proto: packet type ID or len | |
34 | */ | |
35 | struct vlan_hdr { | |
36 | __be16 h_vlan_TCI; | |
37 | __be16 h_vlan_encapsulated_proto; | |
38 | }; | |
39 | ||
40 | /** | |
41 | * struct vlan_ethhdr - vlan ethernet header (ethhdr + vlan_hdr) | |
42 | * @h_dest: destination ethernet address | |
43 | * @h_source: source ethernet address | |
a7bf5804 | 44 | * @h_vlan_proto: ethernet protocol |
740c15d0 PM |
45 | * @h_vlan_TCI: priority and VLAN ID |
46 | * @h_vlan_encapsulated_proto: packet type ID or len | |
47 | */ | |
1da177e4 | 48 | struct vlan_ethhdr { |
6d5c900e KC |
49 | struct_group(addrs, |
50 | unsigned char h_dest[ETH_ALEN]; | |
51 | unsigned char h_source[ETH_ALEN]; | |
52 | ); | |
740c15d0 PM |
53 | __be16 h_vlan_proto; |
54 | __be16 h_vlan_TCI; | |
55 | __be16 h_vlan_encapsulated_proto; | |
1da177e4 LT |
56 | }; |
57 | ||
58 | #include <linux/skbuff.h> | |
59 | ||
60 | static inline struct vlan_ethhdr *vlan_eth_hdr(const struct sk_buff *skb) | |
61 | { | |
98e399f8 | 62 | return (struct vlan_ethhdr *)skb_mac_header(skb); |
1da177e4 LT |
63 | } |
64 | ||
1f5020ac VO |
65 | /* Prefer this version in TX path, instead of |
66 | * skb_reset_mac_header() + vlan_eth_hdr() | |
67 | */ | |
68 | static inline struct vlan_ethhdr *skb_vlan_eth_hdr(const struct sk_buff *skb) | |
69 | { | |
70 | return (struct vlan_ethhdr *)skb->data; | |
71 | } | |
72 | ||
05423b24 ED |
73 | #define VLAN_PRIO_MASK 0xe000 /* Priority Code Point */ |
74 | #define VLAN_PRIO_SHIFT 13 | |
a2e768b8 | 75 | #define VLAN_CFI_MASK 0x1000 /* Canonical Format Indicator / Drop Eligible Indicator */ |
05423b24 | 76 | #define VLAN_VID_MASK 0x0fff /* VLAN Identifier */ |
b738127d | 77 | #define VLAN_N_VID 4096 |
1da177e4 LT |
78 | |
79 | /* found in socket.c */ | |
881d966b | 80 | extern void vlan_ioctl_set(int (*hook)(struct net *, void __user *)); |
1da177e4 | 81 | |
b618aaa9 | 82 | static inline bool is_vlan_dev(const struct net_device *dev) |
6dcbbe25 NH |
83 | { |
84 | return dev->priv_flags & IFF_802_1Q_VLAN; | |
85 | } | |
86 | ||
354259fa | 87 | #define skb_vlan_tag_present(__skb) (!!(__skb)->vlan_all) |
0c4b2d37 | 88 | #define skb_vlan_tag_get(__skb) ((__skb)->vlan_tci) |
df8a39de | 89 | #define skb_vlan_tag_get_id(__skb) ((__skb)->vlan_tci & VLAN_VID_MASK) |
a2e768b8 | 90 | #define skb_vlan_tag_get_cfi(__skb) (!!((__skb)->vlan_tci & VLAN_CFI_MASK)) |
9b319148 | 91 | #define skb_vlan_tag_get_prio(__skb) (((__skb)->vlan_tci & VLAN_PRIO_MASK) >> VLAN_PRIO_SHIFT) |
1da177e4 | 92 | |
9daae9bd GP |
93 | static inline int vlan_get_rx_ctag_filter_info(struct net_device *dev) |
94 | { | |
95 | ASSERT_RTNL(); | |
96 | return notifier_to_errno(call_netdevice_notifiers(NETDEV_CVLAN_FILTER_PUSH_INFO, dev)); | |
97 | } | |
98 | ||
99 | static inline void vlan_drop_rx_ctag_filter_info(struct net_device *dev) | |
100 | { | |
101 | ASSERT_RTNL(); | |
102 | call_netdevice_notifiers(NETDEV_CVLAN_FILTER_DROP_INFO, dev); | |
103 | } | |
104 | ||
105 | static inline int vlan_get_rx_stag_filter_info(struct net_device *dev) | |
106 | { | |
107 | ASSERT_RTNL(); | |
108 | return notifier_to_errno(call_netdevice_notifiers(NETDEV_SVLAN_FILTER_PUSH_INFO, dev)); | |
109 | } | |
110 | ||
111 | static inline void vlan_drop_rx_stag_filter_info(struct net_device *dev) | |
112 | { | |
113 | ASSERT_RTNL(); | |
114 | call_netdevice_notifiers(NETDEV_SVLAN_FILTER_DROP_INFO, dev); | |
115 | } | |
116 | ||
e267cb96 DM |
117 | /** |
118 | * struct vlan_pcpu_stats - VLAN percpu rx/tx stats | |
119 | * @rx_packets: number of received packets | |
120 | * @rx_bytes: number of received bytes | |
121 | * @rx_multicast: number of received multicast packets | |
122 | * @tx_packets: number of transmitted packets | |
123 | * @tx_bytes: number of transmitted bytes | |
124 | * @syncp: synchronization point for 64bit counters | |
125 | * @rx_errors: number of rx errors | |
126 | * @tx_dropped: number of tx drops | |
127 | */ | |
128 | struct vlan_pcpu_stats { | |
09cca53c ED |
129 | u64_stats_t rx_packets; |
130 | u64_stats_t rx_bytes; | |
131 | u64_stats_t rx_multicast; | |
132 | u64_stats_t tx_packets; | |
133 | u64_stats_t tx_bytes; | |
e267cb96 DM |
134 | struct u64_stats_sync syncp; |
135 | u32 rx_errors; | |
136 | u32 tx_dropped; | |
137 | }; | |
138 | ||
1e85c9b6 HFS |
139 | #if defined(CONFIG_VLAN_8021Q) || defined(CONFIG_VLAN_8021Q_MODULE) |
140 | ||
f06c7f9f | 141 | extern struct net_device *__vlan_find_dev_deep_rcu(struct net_device *real_dev, |
1e85c9b6 | 142 | __be16 vlan_proto, u16 vlan_id); |
960abf68 IK |
143 | extern int vlan_for_each(struct net_device *dev, |
144 | int (*action)(struct net_device *dev, int vid, | |
145 | void *arg), void *arg); | |
1e85c9b6 HFS |
146 | extern struct net_device *vlan_dev_real_dev(const struct net_device *dev); |
147 | extern u16 vlan_dev_vlan_id(const struct net_device *dev); | |
71e415e4 | 148 | extern __be16 vlan_dev_vlan_proto(const struct net_device *dev); |
1e85c9b6 HFS |
149 | |
150 | /** | |
151 | * struct vlan_priority_tci_mapping - vlan egress priority mappings | |
152 | * @priority: skb priority | |
153 | * @vlan_qos: vlan priority: (skb->priority << 13) & 0xE000 | |
154 | * @next: pointer to next struct | |
155 | */ | |
156 | struct vlan_priority_tci_mapping { | |
157 | u32 priority; | |
158 | u16 vlan_qos; | |
159 | struct vlan_priority_tci_mapping *next; | |
160 | }; | |
161 | ||
e267cb96 DM |
162 | struct proc_dir_entry; |
163 | struct netpoll; | |
164 | ||
165 | /** | |
166 | * struct vlan_dev_priv - VLAN private device data | |
167 | * @nr_ingress_mappings: number of ingress priority mappings | |
168 | * @ingress_priority_map: ingress priority mappings | |
169 | * @nr_egress_mappings: number of egress priority mappings | |
170 | * @egress_priority_map: hash of egress priority mappings | |
171 | * @vlan_proto: VLAN encapsulation protocol | |
172 | * @vlan_id: VLAN identifier | |
173 | * @flags: device flags | |
174 | * @real_dev: underlying netdevice | |
19c9ebf6 | 175 | * @dev_tracker: refcount tracker for @real_dev reference |
e267cb96 DM |
176 | * @real_dev_addr: address of underlying netdevice |
177 | * @dent: proc dir entry | |
178 | * @vlan_pcpu_stats: ptr to percpu rx stats | |
179 | */ | |
180 | struct vlan_dev_priv { | |
181 | unsigned int nr_ingress_mappings; | |
182 | u32 ingress_priority_map[8]; | |
183 | unsigned int nr_egress_mappings; | |
184 | struct vlan_priority_tci_mapping *egress_priority_map[16]; | |
185 | ||
186 | __be16 vlan_proto; | |
187 | u16 vlan_id; | |
188 | u16 flags; | |
189 | ||
190 | struct net_device *real_dev; | |
19c9ebf6 ED |
191 | netdevice_tracker dev_tracker; |
192 | ||
e267cb96 DM |
193 | unsigned char real_dev_addr[ETH_ALEN]; |
194 | ||
195 | struct proc_dir_entry *dent; | |
196 | struct vlan_pcpu_stats __percpu *vlan_pcpu_stats; | |
197 | #ifdef CONFIG_NET_POLL_CONTROLLER | |
198 | struct netpoll *netpoll; | |
199 | #endif | |
200 | }; | |
201 | ||
202 | static inline struct vlan_dev_priv *vlan_dev_priv(const struct net_device *dev) | |
203 | { | |
204 | return netdev_priv(dev); | |
205 | } | |
206 | ||
207 | static inline u16 | |
208 | vlan_dev_get_egress_qos_mask(struct net_device *dev, u32 skprio) | |
209 | { | |
210 | struct vlan_priority_tci_mapping *mp; | |
211 | ||
212 | smp_rmb(); /* coupled with smp_wmb() in vlan_dev_set_egress_priority() */ | |
213 | ||
214 | mp = vlan_dev_priv(dev)->egress_priority_map[(skprio & 0xF)]; | |
215 | while (mp) { | |
216 | if (mp->priority == skprio) { | |
217 | return mp->vlan_qos; /* This should already be shifted | |
218 | * to mask correctly with the | |
219 | * VLAN's TCI */ | |
220 | } | |
221 | mp = mp->next; | |
222 | } | |
223 | return 0; | |
224 | } | |
225 | ||
48cc32d3 | 226 | extern bool vlan_do_receive(struct sk_buff **skb); |
9b22ea56 | 227 | |
80d5c368 PM |
228 | extern int vlan_vid_add(struct net_device *dev, __be16 proto, u16 vid); |
229 | extern void vlan_vid_del(struct net_device *dev, __be16 proto, u16 vid); | |
87002b03 | 230 | |
348a1443 JP |
231 | extern int vlan_vids_add_by_dev(struct net_device *dev, |
232 | const struct net_device *by_dev); | |
233 | extern void vlan_vids_del_by_dev(struct net_device *dev, | |
234 | const struct net_device *by_dev); | |
9b361c13 JP |
235 | |
236 | extern bool vlan_uses_dev(const struct net_device *dev); | |
e1618d46 | 237 | |
7750f403 | 238 | #else |
cec9c133 | 239 | static inline struct net_device * |
f06c7f9f | 240 | __vlan_find_dev_deep_rcu(struct net_device *real_dev, |
9fae27b3 | 241 | __be16 vlan_proto, u16 vlan_id) |
cec9c133 JP |
242 | { |
243 | return NULL; | |
244 | } | |
245 | ||
960abf68 IK |
246 | static inline int |
247 | vlan_for_each(struct net_device *dev, | |
248 | int (*action)(struct net_device *dev, int vid, void *arg), | |
249 | void *arg) | |
250 | { | |
251 | return 0; | |
252 | } | |
253 | ||
22d1ba74 PM |
254 | static inline struct net_device *vlan_dev_real_dev(const struct net_device *dev) |
255 | { | |
256 | BUG(); | |
257 | return NULL; | |
258 | } | |
259 | ||
260 | static inline u16 vlan_dev_vlan_id(const struct net_device *dev) | |
261 | { | |
262 | BUG(); | |
263 | return 0; | |
264 | } | |
265 | ||
71e415e4 | 266 | static inline __be16 vlan_dev_vlan_proto(const struct net_device *dev) |
267 | { | |
268 | BUG(); | |
269 | return 0; | |
270 | } | |
271 | ||
d3243539 EP |
272 | static inline u16 vlan_dev_get_egress_qos_mask(struct net_device *dev, |
273 | u32 skprio) | |
274 | { | |
275 | return 0; | |
276 | } | |
277 | ||
48cc32d3 | 278 | static inline bool vlan_do_receive(struct sk_buff **skb) |
9b22ea56 | 279 | { |
3701e513 | 280 | return false; |
9b22ea56 | 281 | } |
e1c096e2 | 282 | |
9fae27b3 | 283 | static inline int vlan_vid_add(struct net_device *dev, __be16 proto, u16 vid) |
87002b03 JP |
284 | { |
285 | return 0; | |
286 | } | |
287 | ||
9fae27b3 | 288 | static inline void vlan_vid_del(struct net_device *dev, __be16 proto, u16 vid) |
87002b03 JP |
289 | { |
290 | } | |
348a1443 JP |
291 | |
292 | static inline int vlan_vids_add_by_dev(struct net_device *dev, | |
293 | const struct net_device *by_dev) | |
294 | { | |
295 | return 0; | |
296 | } | |
297 | ||
298 | static inline void vlan_vids_del_by_dev(struct net_device *dev, | |
299 | const struct net_device *by_dev) | |
300 | { | |
301 | } | |
9b361c13 JP |
302 | |
303 | static inline bool vlan_uses_dev(const struct net_device *dev) | |
304 | { | |
305 | return false; | |
306 | } | |
7750f403 | 307 | #endif |
1da177e4 | 308 | |
fe19c4f9 EG |
309 | /** |
310 | * eth_type_vlan - check for valid vlan ether type. | |
311 | * @ethertype: ether type to check | |
312 | * | |
313 | * Returns true if the ether type is a vlan ether type. | |
314 | */ | |
315 | static inline bool eth_type_vlan(__be16 ethertype) | |
316 | { | |
317 | switch (ethertype) { | |
318 | case htons(ETH_P_8021Q): | |
319 | case htons(ETH_P_8021AD): | |
320 | return true; | |
321 | default: | |
322 | return false; | |
323 | } | |
324 | } | |
325 | ||
86a9bad3 PM |
326 | static inline bool vlan_hw_offload_capable(netdev_features_t features, |
327 | __be16 proto) | |
328 | { | |
329 | if (proto == htons(ETH_P_8021Q) && features & NETIF_F_HW_VLAN_CTAG_TX) | |
330 | return true; | |
8ad227ff PM |
331 | if (proto == htons(ETH_P_8021AD) && features & NETIF_F_HW_VLAN_STAG_TX) |
332 | return true; | |
86a9bad3 PM |
333 | return false; |
334 | } | |
335 | ||
1da177e4 | 336 | /** |
cbe7128c | 337 | * __vlan_insert_inner_tag - inner VLAN tag inserting |
1da177e4 | 338 | * @skb: skbuff to tag |
86a9bad3 | 339 | * @vlan_proto: VLAN encapsulation protocol |
9bb8582e | 340 | * @vlan_tci: VLAN TCI to insert |
cbe7128c | 341 | * @mac_len: MAC header length including outer vlan headers |
1da177e4 | 342 | * |
cbe7128c | 343 | * Inserts the VLAN tag into @skb as part of the payload at offset mac_len |
8051ac76 | 344 | * Returns error if skb_cow_head fails. |
0b5c9db1 JP |
345 | * |
346 | * Does not change skb->protocol so this function can be used during receive. | |
1da177e4 | 347 | */ |
cbe7128c TM |
348 | static inline int __vlan_insert_inner_tag(struct sk_buff *skb, |
349 | __be16 vlan_proto, u16 vlan_tci, | |
350 | unsigned int mac_len) | |
1da177e4 LT |
351 | { |
352 | struct vlan_ethhdr *veth; | |
353 | ||
15255a43 JP |
354 | if (skb_cow_head(skb, VLAN_HLEN) < 0) |
355 | return -ENOMEM; | |
356 | ||
cbe7128c | 357 | skb_push(skb, VLAN_HLEN); |
1da177e4 | 358 | |
cbe7128c | 359 | /* Move the mac header sans proto to the beginning of the new header. */ |
c769accd TM |
360 | if (likely(mac_len > ETH_TLEN)) |
361 | memmove(skb->data, skb->data + VLAN_HLEN, mac_len - ETH_TLEN); | |
f90615ad VO |
362 | if (skb_mac_header_was_set(skb)) |
363 | skb->mac_header -= VLAN_HLEN; | |
1da177e4 | 364 | |
cbe7128c TM |
365 | veth = (struct vlan_ethhdr *)(skb->data + mac_len - ETH_HLEN); |
366 | ||
1da177e4 | 367 | /* first, the ethernet type */ |
c769accd TM |
368 | if (likely(mac_len >= ETH_TLEN)) { |
369 | /* h_vlan_encapsulated_proto should already be populated, and | |
370 | * skb->data has space for h_vlan_proto | |
371 | */ | |
372 | veth->h_vlan_proto = vlan_proto; | |
373 | } else { | |
374 | /* h_vlan_encapsulated_proto should not be populated, and | |
375 | * skb->data has no space for h_vlan_proto | |
376 | */ | |
377 | veth->h_vlan_encapsulated_proto = skb->protocol; | |
378 | } | |
1da177e4 | 379 | |
9bb8582e PM |
380 | /* now, the TCI */ |
381 | veth->h_vlan_TCI = htons(vlan_tci); | |
1da177e4 | 382 | |
15255a43 JP |
383 | return 0; |
384 | } | |
385 | ||
386 | /** | |
cbe7128c | 387 | * __vlan_insert_tag - regular VLAN tag inserting |
15255a43 JP |
388 | * @skb: skbuff to tag |
389 | * @vlan_proto: VLAN encapsulation protocol | |
390 | * @vlan_tci: VLAN TCI to insert | |
391 | * | |
392 | * Inserts the VLAN tag into @skb as part of the payload | |
8051ac76 | 393 | * Returns error if skb_cow_head fails. |
cbe7128c TM |
394 | * |
395 | * Does not change skb->protocol so this function can be used during receive. | |
396 | */ | |
397 | static inline int __vlan_insert_tag(struct sk_buff *skb, | |
398 | __be16 vlan_proto, u16 vlan_tci) | |
399 | { | |
400 | return __vlan_insert_inner_tag(skb, vlan_proto, vlan_tci, ETH_HLEN); | |
401 | } | |
402 | ||
403 | /** | |
404 | * vlan_insert_inner_tag - inner VLAN tag inserting | |
405 | * @skb: skbuff to tag | |
406 | * @vlan_proto: VLAN encapsulation protocol | |
407 | * @vlan_tci: VLAN TCI to insert | |
408 | * @mac_len: MAC header length including outer vlan headers | |
409 | * | |
410 | * Inserts the VLAN tag into @skb as part of the payload at offset mac_len | |
7740bb88 | 411 | * Returns a VLAN tagged skb. This might change skb->head. |
15255a43 JP |
412 | * |
413 | * Following the skb_unshare() example, in case of error, the calling function | |
414 | * doesn't have to worry about freeing the original skb. | |
415 | * | |
416 | * Does not change skb->protocol so this function can be used during receive. | |
417 | */ | |
cbe7128c TM |
418 | static inline struct sk_buff *vlan_insert_inner_tag(struct sk_buff *skb, |
419 | __be16 vlan_proto, | |
420 | u16 vlan_tci, | |
421 | unsigned int mac_len) | |
15255a43 JP |
422 | { |
423 | int err; | |
424 | ||
cbe7128c | 425 | err = __vlan_insert_inner_tag(skb, vlan_proto, vlan_tci, mac_len); |
15255a43 JP |
426 | if (err) { |
427 | dev_kfree_skb_any(skb); | |
428 | return NULL; | |
429 | } | |
0b5c9db1 JP |
430 | return skb; |
431 | } | |
1da177e4 | 432 | |
cbe7128c TM |
433 | /** |
434 | * vlan_insert_tag - regular VLAN tag inserting | |
435 | * @skb: skbuff to tag | |
436 | * @vlan_proto: VLAN encapsulation protocol | |
437 | * @vlan_tci: VLAN TCI to insert | |
438 | * | |
439 | * Inserts the VLAN tag into @skb as part of the payload | |
7740bb88 | 440 | * Returns a VLAN tagged skb. This might change skb->head. |
cbe7128c TM |
441 | * |
442 | * Following the skb_unshare() example, in case of error, the calling function | |
443 | * doesn't have to worry about freeing the original skb. | |
444 | * | |
445 | * Does not change skb->protocol so this function can be used during receive. | |
446 | */ | |
447 | static inline struct sk_buff *vlan_insert_tag(struct sk_buff *skb, | |
448 | __be16 vlan_proto, u16 vlan_tci) | |
449 | { | |
450 | return vlan_insert_inner_tag(skb, vlan_proto, vlan_tci, ETH_HLEN); | |
451 | } | |
452 | ||
0b5c9db1 | 453 | /** |
62749e2c | 454 | * vlan_insert_tag_set_proto - regular VLAN tag inserting |
0b5c9db1 | 455 | * @skb: skbuff to tag |
62749e2c | 456 | * @vlan_proto: VLAN encapsulation protocol |
0b5c9db1 JP |
457 | * @vlan_tci: VLAN TCI to insert |
458 | * | |
459 | * Inserts the VLAN tag into @skb as part of the payload | |
7740bb88 | 460 | * Returns a VLAN tagged skb. This might change skb->head. |
0b5c9db1 JP |
461 | * |
462 | * Following the skb_unshare() example, in case of error, the calling function | |
463 | * doesn't have to worry about freeing the original skb. | |
464 | */ | |
62749e2c JP |
465 | static inline struct sk_buff *vlan_insert_tag_set_proto(struct sk_buff *skb, |
466 | __be16 vlan_proto, | |
467 | u16 vlan_tci) | |
0b5c9db1 | 468 | { |
86a9bad3 | 469 | skb = vlan_insert_tag(skb, vlan_proto, vlan_tci); |
0b5c9db1 | 470 | if (skb) |
86a9bad3 | 471 | skb->protocol = vlan_proto; |
1da177e4 LT |
472 | return skb; |
473 | } | |
474 | ||
c8accd5a MM |
475 | /** |
476 | * __vlan_hwaccel_clear_tag - clear hardware accelerated VLAN info | |
477 | * @skb: skbuff to clear | |
478 | * | |
479 | * Clears the VLAN information from @skb | |
480 | */ | |
481 | static inline void __vlan_hwaccel_clear_tag(struct sk_buff *skb) | |
482 | { | |
354259fa | 483 | skb->vlan_all = 0; |
c8accd5a MM |
484 | } |
485 | ||
e0a6b809 MM |
486 | /** |
487 | * __vlan_hwaccel_copy_tag - copy hardware accelerated VLAN info from another skb | |
488 | * @dst: skbuff to copy to | |
489 | * @src: skbuff to copy from | |
490 | * | |
491 | * Copies VLAN information from @src to @dst (for branchless code) | |
492 | */ | |
493 | static inline void __vlan_hwaccel_copy_tag(struct sk_buff *dst, const struct sk_buff *src) | |
494 | { | |
354259fa | 495 | dst->vlan_all = src->vlan_all; |
e0a6b809 MM |
496 | } |
497 | ||
5968250c JP |
498 | /* |
499 | * __vlan_hwaccel_push_inside - pushes vlan tag to the payload | |
500 | * @skb: skbuff to tag | |
501 | * | |
502 | * Pushes the VLAN tag from @skb->vlan_tci inside to the payload. | |
503 | * | |
504 | * Following the skb_unshare() example, in case of error, the calling function | |
505 | * doesn't have to worry about freeing the original skb. | |
506 | */ | |
507 | static inline struct sk_buff *__vlan_hwaccel_push_inside(struct sk_buff *skb) | |
508 | { | |
509 | skb = vlan_insert_tag_set_proto(skb, skb->vlan_proto, | |
df8a39de | 510 | skb_vlan_tag_get(skb)); |
5968250c | 511 | if (likely(skb)) |
c8accd5a | 512 | __vlan_hwaccel_clear_tag(skb); |
5968250c JP |
513 | return skb; |
514 | } | |
5968250c | 515 | |
1da177e4 LT |
516 | /** |
517 | * __vlan_hwaccel_put_tag - hardware accelerated VLAN inserting | |
518 | * @skb: skbuff to tag | |
86a9bad3 | 519 | * @vlan_proto: VLAN encapsulation protocol |
9bb8582e | 520 | * @vlan_tci: VLAN TCI to insert |
1da177e4 | 521 | * |
6aa895b0 | 522 | * Puts the VLAN TCI in @skb->vlan_tci and lets the device do the rest |
1da177e4 | 523 | */ |
b960a0ac JP |
524 | static inline void __vlan_hwaccel_put_tag(struct sk_buff *skb, |
525 | __be16 vlan_proto, u16 vlan_tci) | |
1da177e4 | 526 | { |
86a9bad3 | 527 | skb->vlan_proto = vlan_proto; |
0c4b2d37 | 528 | skb->vlan_tci = vlan_tci; |
1da177e4 LT |
529 | } |
530 | ||
1da177e4 LT |
531 | /** |
532 | * __vlan_get_tag - get the VLAN ID that is part of the payload | |
533 | * @skb: skbuff to query | |
f4fb874c | 534 | * @vlan_tci: buffer to store value |
9bb8582e | 535 | * |
1da177e4 LT |
536 | * Returns error if the skb is not of VLAN type |
537 | */ | |
9bb8582e | 538 | static inline int __vlan_get_tag(const struct sk_buff *skb, u16 *vlan_tci) |
1da177e4 | 539 | { |
1f5020ac | 540 | struct vlan_ethhdr *veth = skb_vlan_eth_hdr(skb); |
1da177e4 | 541 | |
fe19c4f9 | 542 | if (!eth_type_vlan(veth->h_vlan_proto)) |
537fec07 | 543 | return -ENODATA; |
1da177e4 | 544 | |
9bb8582e | 545 | *vlan_tci = ntohs(veth->h_vlan_TCI); |
1da177e4 LT |
546 | return 0; |
547 | } | |
548 | ||
549 | /** | |
550 | * __vlan_hwaccel_get_tag - get the VLAN ID that is in @skb->cb[] | |
551 | * @skb: skbuff to query | |
f4fb874c | 552 | * @vlan_tci: buffer to store value |
9bb8582e | 553 | * |
6aa895b0 | 554 | * Returns error if @skb->vlan_tci is not set correctly |
1da177e4 | 555 | */ |
18149935 | 556 | static inline int __vlan_hwaccel_get_tag(const struct sk_buff *skb, |
9bb8582e | 557 | u16 *vlan_tci) |
1da177e4 | 558 | { |
df8a39de JP |
559 | if (skb_vlan_tag_present(skb)) { |
560 | *vlan_tci = skb_vlan_tag_get(skb); | |
1da177e4 LT |
561 | return 0; |
562 | } else { | |
9bb8582e | 563 | *vlan_tci = 0; |
537fec07 | 564 | return -ENODATA; |
1da177e4 LT |
565 | } |
566 | } | |
567 | ||
1da177e4 LT |
568 | /** |
569 | * vlan_get_tag - get the VLAN ID from the skb | |
570 | * @skb: skbuff to query | |
f4fb874c | 571 | * @vlan_tci: buffer to store value |
9bb8582e | 572 | * |
1da177e4 LT |
573 | * Returns error if the skb is not VLAN tagged |
574 | */ | |
9bb8582e | 575 | static inline int vlan_get_tag(const struct sk_buff *skb, u16 *vlan_tci) |
1da177e4 | 576 | { |
f646968f | 577 | if (skb->dev->features & NETIF_F_HW_VLAN_CTAG_TX) { |
9bb8582e | 578 | return __vlan_hwaccel_get_tag(skb, vlan_tci); |
1da177e4 | 579 | } else { |
9bb8582e | 580 | return __vlan_get_tag(skb, vlan_tci); |
1da177e4 LT |
581 | } |
582 | } | |
583 | ||
0a85df00 HZ |
584 | /** |
585 | * vlan_get_protocol - get protocol EtherType. | |
586 | * @skb: skbuff to query | |
d4bcef3f TM |
587 | * @type: first vlan protocol |
588 | * @depth: buffer to store length of eth and vlan tags in bytes | |
0a85df00 HZ |
589 | * |
590 | * Returns the EtherType of the packet, regardless of whether it is | |
591 | * vlan encapsulated (normal or hardware accelerated) or not. | |
592 | */ | |
469acedd | 593 | static inline __be16 __vlan_get_protocol(const struct sk_buff *skb, __be16 type, |
d4bcef3f | 594 | int *depth) |
0a85df00 | 595 | { |
469acedd | 596 | unsigned int vlan_depth = skb->mac_len, parse_depth = VLAN_MAX_DEPTH; |
d4bcef3f TM |
597 | |
598 | /* if type is 802.1Q/AD then the header should already be | |
599 | * present at mac_len - VLAN_HLEN (if mac_len > 0), or at | |
600 | * ETH_HLEN otherwise | |
601 | */ | |
fe19c4f9 | 602 | if (eth_type_vlan(type)) { |
d4bcef3f TM |
603 | if (vlan_depth) { |
604 | if (WARN_ON(vlan_depth < VLAN_HLEN)) | |
605 | return 0; | |
606 | vlan_depth -= VLAN_HLEN; | |
607 | } else { | |
608 | vlan_depth = ETH_HLEN; | |
609 | } | |
610 | do { | |
469acedd | 611 | struct vlan_hdr vhdr, *vh; |
d4bcef3f | 612 | |
469acedd THJ |
613 | vh = skb_header_pointer(skb, vlan_depth, sizeof(vhdr), &vhdr); |
614 | if (unlikely(!vh || !--parse_depth)) | |
d4bcef3f TM |
615 | return 0; |
616 | ||
d4bcef3f TM |
617 | type = vh->h_vlan_encapsulated_proto; |
618 | vlan_depth += VLAN_HLEN; | |
fe19c4f9 | 619 | } while (eth_type_vlan(type)); |
0a85df00 HZ |
620 | } |
621 | ||
d4bcef3f TM |
622 | if (depth) |
623 | *depth = vlan_depth; | |
624 | ||
625 | return type; | |
626 | } | |
627 | ||
628 | /** | |
629 | * vlan_get_protocol - get protocol EtherType. | |
630 | * @skb: skbuff to query | |
631 | * | |
632 | * Returns the EtherType of the packet, regardless of whether it is | |
633 | * vlan encapsulated (normal or hardware accelerated) or not. | |
634 | */ | |
469acedd | 635 | static inline __be16 vlan_get_protocol(const struct sk_buff *skb) |
d4bcef3f TM |
636 | { |
637 | return __vlan_get_protocol(skb, skb->protocol, NULL); | |
0a85df00 | 638 | } |
396cf943 | 639 | |
4063384e ED |
640 | /* This version of __vlan_get_protocol() also pulls mac header in skb->head */ |
641 | static inline __be16 vlan_get_protocol_and_depth(struct sk_buff *skb, | |
642 | __be16 type, int *depth) | |
643 | { | |
644 | int maclen; | |
645 | ||
646 | type = __vlan_get_protocol(skb, type, &maclen); | |
647 | ||
648 | if (type) { | |
649 | if (!pskb_may_pull(skb, maclen)) | |
650 | type = 0; | |
651 | else if (depth) | |
652 | *depth = maclen; | |
653 | } | |
654 | return type; | |
655 | } | |
656 | ||
469acedd THJ |
657 | /* A getter for the SKB protocol field which will handle VLAN tags consistently |
658 | * whether VLAN acceleration is enabled or not. | |
659 | */ | |
660 | static inline __be16 skb_protocol(const struct sk_buff *skb, bool skip_vlan) | |
661 | { | |
662 | if (!skip_vlan) | |
663 | /* VLAN acceleration strips the VLAN header from the skb and | |
664 | * moves it to skb->vlan_proto | |
665 | */ | |
666 | return skb_vlan_tag_present(skb) ? skb->vlan_proto : skb->protocol; | |
667 | ||
668 | return vlan_get_protocol(skb); | |
669 | } | |
670 | ||
396cf943 PS |
671 | static inline void vlan_set_encap_proto(struct sk_buff *skb, |
672 | struct vlan_hdr *vhdr) | |
673 | { | |
674 | __be16 proto; | |
da8c8724 | 675 | unsigned short *rawp; |
396cf943 PS |
676 | |
677 | /* | |
678 | * Was a VLAN packet, grab the encapsulated protocol, which the layer | |
679 | * three protocols care about. | |
680 | */ | |
681 | ||
682 | proto = vhdr->h_vlan_encapsulated_proto; | |
9545b22d | 683 | if (eth_proto_is_802_3(proto)) { |
396cf943 PS |
684 | skb->protocol = proto; |
685 | return; | |
686 | } | |
687 | ||
da8c8724 CW |
688 | rawp = (unsigned short *)(vhdr + 1); |
689 | if (*rawp == 0xFFFF) | |
396cf943 PS |
690 | /* |
691 | * This is a magic hack to spot IPX packets. Older Novell | |
692 | * breaks the protocol design and runs IPX over 802.3 without | |
693 | * an 802.2 LLC layer. We look for FFFF which isn't a used | |
694 | * 802.2 SSAP/DSAP. This won't work for fault tolerant netware | |
695 | * but does for the rest. | |
696 | */ | |
697 | skb->protocol = htons(ETH_P_802_3); | |
698 | else | |
699 | /* | |
700 | * Real 802.2 LLC | |
701 | */ | |
702 | skb->protocol = htons(ETH_P_802_2); | |
703 | } | |
44a40855 | 704 | |
0bcf2e4a VO |
705 | /** |
706 | * vlan_remove_tag - remove outer VLAN tag from payload | |
707 | * @skb: skbuff to remove tag from | |
708 | * @vlan_tci: buffer to store value | |
709 | * | |
710 | * Expects the skb to contain a VLAN tag in the payload, and to have skb->data | |
711 | * pointing at the MAC header. | |
712 | * | |
713 | * Returns a new pointer to skb->data, or NULL on failure to pull. | |
714 | */ | |
715 | static inline void *vlan_remove_tag(struct sk_buff *skb, u16 *vlan_tci) | |
716 | { | |
717 | struct vlan_hdr *vhdr = (struct vlan_hdr *)(skb->data + ETH_HLEN); | |
718 | ||
719 | *vlan_tci = ntohs(vhdr->h_vlan_TCI); | |
720 | ||
721 | memmove(skb->data + VLAN_HLEN, skb->data, 2 * ETH_ALEN); | |
722 | vlan_set_encap_proto(skb, vhdr); | |
723 | return __skb_pull(skb, VLAN_HLEN); | |
724 | } | |
725 | ||
f5a7fb88 TM |
726 | /** |
727 | * skb_vlan_tagged - check if skb is vlan tagged. | |
728 | * @skb: skbuff to query | |
729 | * | |
730 | * Returns true if the skb is tagged, regardless of whether it is hardware | |
731 | * accelerated or not. | |
732 | */ | |
733 | static inline bool skb_vlan_tagged(const struct sk_buff *skb) | |
734 | { | |
735 | if (!skb_vlan_tag_present(skb) && | |
fe19c4f9 | 736 | likely(!eth_type_vlan(skb->protocol))) |
f5a7fb88 TM |
737 | return false; |
738 | ||
739 | return true; | |
740 | } | |
741 | ||
742 | /** | |
743 | * skb_vlan_tagged_multi - check if skb is vlan tagged with multiple headers. | |
744 | * @skb: skbuff to query | |
745 | * | |
746 | * Returns true if the skb is tagged with multiple vlan headers, regardless | |
747 | * of whether it is hardware accelerated or not. | |
748 | */ | |
7ce23672 | 749 | static inline bool skb_vlan_tagged_multi(struct sk_buff *skb) |
f5a7fb88 TM |
750 | { |
751 | __be16 protocol = skb->protocol; | |
752 | ||
753 | if (!skb_vlan_tag_present(skb)) { | |
754 | struct vlan_ethhdr *veh; | |
755 | ||
fe19c4f9 | 756 | if (likely(!eth_type_vlan(protocol))) |
f5a7fb88 TM |
757 | return false; |
758 | ||
7ce23672 TM |
759 | if (unlikely(!pskb_may_pull(skb, VLAN_ETH_HLEN))) |
760 | return false; | |
761 | ||
1f5020ac | 762 | veh = skb_vlan_eth_hdr(skb); |
f5a7fb88 TM |
763 | protocol = veh->h_vlan_encapsulated_proto; |
764 | } | |
765 | ||
fe19c4f9 | 766 | if (!eth_type_vlan(protocol)) |
f5a7fb88 TM |
767 | return false; |
768 | ||
769 | return true; | |
770 | } | |
771 | ||
8cb65d00 TM |
772 | /** |
773 | * vlan_features_check - drop unsafe features for skb with multiple tags. | |
774 | * @skb: skbuff to query | |
775 | * @features: features to be checked | |
776 | * | |
777 | * Returns features without unsafe ones if the skb has multiple tags. | |
778 | */ | |
7ce23672 | 779 | static inline netdev_features_t vlan_features_check(struct sk_buff *skb, |
8cb65d00 TM |
780 | netdev_features_t features) |
781 | { | |
35d2f80b VY |
782 | if (skb_vlan_tagged_multi(skb)) { |
783 | /* In the case of multi-tagged packets, use a direct mask | |
784 | * instead of using netdev_interesect_features(), to make | |
785 | * sure that only devices supporting NETIF_F_HW_CSUM will | |
786 | * have checksum offloading support. | |
787 | */ | |
788 | features &= NETIF_F_SG | NETIF_F_HIGHDMA | NETIF_F_HW_CSUM | | |
789 | NETIF_F_FRAGLIST | NETIF_F_HW_VLAN_CTAG_TX | | |
790 | NETIF_F_HW_VLAN_STAG_TX; | |
791 | } | |
8cb65d00 TM |
792 | |
793 | return features; | |
794 | } | |
795 | ||
66e5133f TM |
796 | /** |
797 | * compare_vlan_header - Compare two vlan headers | |
798 | * @h1: Pointer to vlan header | |
799 | * @h2: Pointer to vlan header | |
800 | * | |
801 | * Compare two vlan headers, returns 0 if equal. | |
802 | * | |
803 | * Please note that alignment of h1 & h2 are only guaranteed to be 16 bits. | |
804 | */ | |
805 | static inline unsigned long compare_vlan_header(const struct vlan_hdr *h1, | |
806 | const struct vlan_hdr *h2) | |
807 | { | |
808 | #if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) | |
809 | return *(u32 *)h1 ^ *(u32 *)h2; | |
810 | #else | |
811 | return ((__force u32)h1->h_vlan_TCI ^ (__force u32)h2->h_vlan_TCI) | | |
812 | ((__force u32)h1->h_vlan_encapsulated_proto ^ | |
813 | (__force u32)h2->h_vlan_encapsulated_proto); | |
814 | #endif | |
815 | } | |
1da177e4 | 816 | #endif /* !(_LINUX_IF_VLAN_H_) */ |