1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22#ifndef _LINUX_ETHERDEVICE_H
23#define _LINUX_ETHERDEVICE_H
24
25#include <linux/if_ether.h>
26#include <linux/netdevice.h>
27#include <linux/random.h>
28#include <asm/unaligned.h>
29#include <asm/bitsperlong.h>
30
31#ifdef __KERNEL__
32struct device;
33int eth_platform_get_mac_address(struct device *dev, u8 *mac_addr);
34unsigned char *arch_get_platform_mac_address(void);
35u32 eth_get_headlen(void *data, unsigned int max_len);
36__be16 eth_type_trans(struct sk_buff *skb, struct net_device *dev);
37extern const struct header_ops eth_header_ops;
38
39int eth_header(struct sk_buff *skb, struct net_device *dev, unsigned short type,
40 const void *daddr, const void *saddr, unsigned len);
41int eth_header_parse(const struct sk_buff *skb, unsigned char *haddr);
42int eth_header_cache(const struct neighbour *neigh, struct hh_cache *hh,
43 __be16 type);
44void eth_header_cache_update(struct hh_cache *hh, const struct net_device *dev,
45 const unsigned char *haddr);
46int eth_prepare_mac_addr_change(struct net_device *dev, void *p);
47void eth_commit_mac_addr_change(struct net_device *dev, void *p);
48int eth_mac_addr(struct net_device *dev, void *p);
49int eth_change_mtu(struct net_device *dev, int new_mtu);
50int eth_validate_addr(struct net_device *dev);
51
52struct net_device *alloc_etherdev_mqs(int sizeof_priv, unsigned int txqs,
53 unsigned int rxqs);
54#define alloc_etherdev(sizeof_priv) alloc_etherdev_mq(sizeof_priv, 1)
55#define alloc_etherdev_mq(sizeof_priv, count) alloc_etherdev_mqs(sizeof_priv, count, count)
56
57struct net_device *devm_alloc_etherdev_mqs(struct device *dev, int sizeof_priv,
58 unsigned int txqs,
59 unsigned int rxqs);
60#define devm_alloc_etherdev(dev, sizeof_priv) devm_alloc_etherdev_mqs(dev, sizeof_priv, 1, 1)
61
62struct sk_buff **eth_gro_receive(struct sk_buff **head,
63 struct sk_buff *skb);
64int eth_gro_complete(struct sk_buff *skb, int nhoff);
65
66
67static const u8 eth_reserved_addr_base[ETH_ALEN] __aligned(2) =
68{ 0x01, 0x80, 0xc2, 0x00, 0x00, 0x00 };
69#define eth_stp_addr eth_reserved_addr_base
70
71
72
73
74
75
76
77
78
79
80static inline bool is_link_local_ether_addr(const u8 *addr)
81{
82 __be16 *a = (__be16 *)addr;
83 static const __be16 *b = (const __be16 *)eth_reserved_addr_base;
84 static const __be16 m = cpu_to_be16(0xfff0);
85
86#if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS)
87 return (((*(const u32 *)addr) ^ (*(const u32 *)b)) |
88 (__force int)((a[2] ^ b[2]) & m)) == 0;
89#else
90 return ((a[0] ^ b[0]) | (a[1] ^ b[1]) | ((a[2] ^ b[2]) & m)) == 0;
91#endif
92}
93
94
95
96
97
98
99
100
101
102static inline bool is_zero_ether_addr(const u8 *addr)
103{
104#if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS)
105 return ((*(const u32 *)addr) | (*(const u16 *)(addr + 4))) == 0;
106#else
107 return (*(const u16 *)(addr + 0) |
108 *(const u16 *)(addr + 2) |
109 *(const u16 *)(addr + 4)) == 0;
110#endif
111}
112
113
114
115
116
117
118
119
120static inline bool is_multicast_ether_addr(const u8 *addr)
121{
122#if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS)
123 u32 a = *(const u32 *)addr;
124#else
125 u16 a = *(const u16 *)addr;
126#endif
127#ifdef __BIG_ENDIAN
128 return 0x01 & (a >> ((sizeof(a) * 8) - 8));
129#else
130 return 0x01 & a;
131#endif
132}
133
134static inline bool is_multicast_ether_addr_64bits(const u8 addr[6+2])
135{
136#if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64
137#ifdef __BIG_ENDIAN
138 return 0x01 & ((*(const u64 *)addr) >> 56);
139#else
140 return 0x01 & (*(const u64 *)addr);
141#endif
142#else
143 return is_multicast_ether_addr(addr);
144#endif
145}
146
147
148
149
150
151
152
153static inline bool is_local_ether_addr(const u8 *addr)
154{
155 return 0x02 & addr[0];
156}
157
158
159
160
161
162
163
164
165
166static inline bool is_broadcast_ether_addr(const u8 *addr)
167{
168 return (*(const u16 *)(addr + 0) &
169 *(const u16 *)(addr + 2) &
170 *(const u16 *)(addr + 4)) == 0xffff;
171}
172
173
174
175
176
177
178
179static inline bool is_unicast_ether_addr(const u8 *addr)
180{
181 return !is_multicast_ether_addr(addr);
182}
183
184
185
186
187
188
189
190
191
192
193
194
195static inline bool is_valid_ether_addr(const u8 *addr)
196{
197
198
199 return !is_multicast_ether_addr(addr) && !is_zero_ether_addr(addr);
200}
201
202
203
204
205
206
207
208
209
210static inline bool eth_proto_is_802_3(__be16 proto)
211{
212#ifndef __BIG_ENDIAN
213
214 proto &= htons(0xFF00);
215#endif
216
217 return (__force u16)proto >= (__force u16)htons(ETH_P_802_3_MIN);
218}
219
220
221
222
223
224
225
226
227static inline void eth_random_addr(u8 *addr)
228{
229 get_random_bytes(addr, ETH_ALEN);
230 addr[0] &= 0xfe;
231 addr[0] |= 0x02;
232}
233
234#define random_ether_addr(addr) eth_random_addr(addr)
235
236
237
238
239
240
241
242static inline void eth_broadcast_addr(u8 *addr)
243{
244 memset(addr, 0xff, ETH_ALEN);
245}
246
247
248
249
250
251
252
253static inline void eth_zero_addr(u8 *addr)
254{
255 memset(addr, 0x00, ETH_ALEN);
256}
257
258
259
260
261
262
263
264
265
266
267static inline void eth_hw_addr_random(struct net_device *dev)
268{
269 dev->addr_assign_type = NET_ADDR_RANDOM;
270 eth_random_addr(dev->dev_addr);
271}
272
273
274
275
276
277
278
279
280static inline void ether_addr_copy(u8 *dst, const u8 *src)
281{
282#if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS)
283 *(u32 *)dst = *(const u32 *)src;
284 *(u16 *)(dst + 4) = *(const u16 *)(src + 4);
285#else
286 u16 *a = (u16 *)dst;
287 const u16 *b = (const u16 *)src;
288
289 a[0] = b[0];
290 a[1] = b[1];
291 a[2] = b[2];
292#endif
293}
294
295
296
297
298
299
300
301
302
303static inline void eth_hw_addr_inherit(struct net_device *dst,
304 struct net_device *src)
305{
306 dst->addr_assign_type = src->addr_assign_type;
307 ether_addr_copy(dst->dev_addr, src->dev_addr);
308}
309
310
311
312
313
314
315
316
317
318
319static inline bool ether_addr_equal(const u8 *addr1, const u8 *addr2)
320{
321#if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS)
322 u32 fold = ((*(const u32 *)addr1) ^ (*(const u32 *)addr2)) |
323 ((*(const u16 *)(addr1 + 4)) ^ (*(const u16 *)(addr2 + 4)));
324
325 return fold == 0;
326#else
327 const u16 *a = (const u16 *)addr1;
328 const u16 *b = (const u16 *)addr2;
329
330 return ((a[0] ^ b[0]) | (a[1] ^ b[1]) | (a[2] ^ b[2])) == 0;
331#endif
332}
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348static inline bool ether_addr_equal_64bits(const u8 addr1[6+2],
349 const u8 addr2[6+2])
350{
351#if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64
352 u64 fold = (*(const u64 *)addr1) ^ (*(const u64 *)addr2);
353
354#ifdef __BIG_ENDIAN
355 return (fold >> 16) == 0;
356#else
357 return (fold << 16) == 0;
358#endif
359#else
360 return ether_addr_equal(addr1, addr2);
361#endif
362}
363
364
365
366
367
368
369
370
371
372
373static inline bool ether_addr_equal_unaligned(const u8 *addr1, const u8 *addr2)
374{
375#if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS)
376 return ether_addr_equal(addr1, addr2);
377#else
378 return memcmp(addr1, addr2, ETH_ALEN) == 0;
379#endif
380}
381
382
383
384
385
386
387
388
389
390
391
392static inline bool ether_addr_equal_masked(const u8 *addr1, const u8 *addr2,
393 const u8 *mask)
394{
395 int i;
396
397 for (i = 0; i < ETH_ALEN; i++) {
398 if ((addr1[i] ^ addr2[i]) & mask[i])
399 return false;
400 }
401
402 return true;
403}
404
405
406
407
408
409
410
411static inline u64 ether_addr_to_u64(const u8 *addr)
412{
413 u64 u = 0;
414 int i;
415
416 for (i = 0; i < ETH_ALEN; i++)
417 u = u << 8 | addr[i];
418
419 return u;
420}
421
422
423
424
425
426
427static inline void u64_to_ether_addr(u64 u, u8 *addr)
428{
429 int i;
430
431 for (i = ETH_ALEN - 1; i >= 0; i--) {
432 addr[i] = u & 0xff;
433 u = u >> 8;
434 }
435}
436
437
438
439
440
441
442static inline void eth_addr_dec(u8 *addr)
443{
444 u64 u = ether_addr_to_u64(addr);
445
446 u--;
447 u64_to_ether_addr(u, addr);
448}
449
450
451
452
453
454
455
456
457
458
459
460
461static inline bool is_etherdev_addr(const struct net_device *dev,
462 const u8 addr[6 + 2])
463{
464 struct netdev_hw_addr *ha;
465 bool res = false;
466
467 rcu_read_lock();
468 for_each_dev_addr(dev, ha) {
469 res = ether_addr_equal_64bits(addr, ha->addr);
470 if (res)
471 break;
472 }
473 rcu_read_unlock();
474 return res;
475}
476#endif
477
478
479
480
481
482
483
484
485
486
487
488
489
490static inline unsigned long compare_ether_header(const void *a, const void *b)
491{
492#if defined(CONFIG_HAVE_EFFICIENT_UNALIGNED_ACCESS) && BITS_PER_LONG == 64
493 unsigned long fold;
494
495
496
497
498
499
500
501
502
503 fold = *(unsigned long *)a ^ *(unsigned long *)b;
504 fold |= *(unsigned long *)(a + 6) ^ *(unsigned long *)(b + 6);
505 return fold;
506#else
507 u32 *a32 = (u32 *)((u8 *)a + 2);
508 u32 *b32 = (u32 *)((u8 *)b + 2);
509
510 return (*(u16 *)a ^ *(u16 *)b) | (a32[0] ^ b32[0]) |
511 (a32[1] ^ b32[1]) | (a32[2] ^ b32[2]);
512#endif
513}
514
515
516
517
518
519
520
521
522static inline int eth_skb_pad(struct sk_buff *skb)
523{
524 return skb_put_padto(skb, ETH_ZLEN);
525}
526
527#endif
528