1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19#include "distributed-arp-table.h"
20#include "main.h"
21
22#include <linux/atomic.h>
23#include <linux/bitops.h>
24#include <linux/byteorder/generic.h>
25#include <linux/errno.h>
26#include <linux/etherdevice.h>
27#include <linux/gfp.h>
28#include <linux/if_arp.h>
29#include <linux/if_ether.h>
30#include <linux/if_vlan.h>
31#include <linux/in.h>
32#include <linux/jiffies.h>
33#include <linux/kernel.h>
34#include <linux/kref.h>
35#include <linux/list.h>
36#include <linux/rculist.h>
37#include <linux/rcupdate.h>
38#include <linux/seq_file.h>
39#include <linux/skbuff.h>
40#include <linux/slab.h>
41#include <linux/spinlock.h>
42#include <linux/stddef.h>
43#include <linux/string.h>
44#include <linux/workqueue.h>
45#include <net/arp.h>
46
47#include "bridge_loop_avoidance.h"
48#include "hard-interface.h"
49#include "hash.h"
50#include "log.h"
51#include "originator.h"
52#include "send.h"
53#include "translation-table.h"
54#include "tvlv.h"
55
56static void batadv_dat_purge(struct work_struct *work);
57
58
59
60
61
62static void batadv_dat_start_timer(struct batadv_priv *bat_priv)
63{
64 INIT_DELAYED_WORK(&bat_priv->dat.work, batadv_dat_purge);
65 queue_delayed_work(batadv_event_workqueue, &bat_priv->dat.work,
66 msecs_to_jiffies(10000));
67}
68
69
70
71
72
73
74static void batadv_dat_entry_release(struct kref *ref)
75{
76 struct batadv_dat_entry *dat_entry;
77
78 dat_entry = container_of(ref, struct batadv_dat_entry, refcount);
79
80 kfree_rcu(dat_entry, rcu);
81}
82
83
84
85
86
87
88static void batadv_dat_entry_put(struct batadv_dat_entry *dat_entry)
89{
90 kref_put(&dat_entry->refcount, batadv_dat_entry_release);
91}
92
93
94
95
96
97
98
99static bool batadv_dat_to_purge(struct batadv_dat_entry *dat_entry)
100{
101 return batadv_has_timed_out(dat_entry->last_update,
102 BATADV_DAT_ENTRY_TIMEOUT);
103}
104
105
106
107
108
109
110
111
112
113
114
115
116static void __batadv_dat_purge(struct batadv_priv *bat_priv,
117 bool (*to_purge)(struct batadv_dat_entry *))
118{
119 spinlock_t *list_lock;
120 struct batadv_dat_entry *dat_entry;
121 struct hlist_node *node_tmp;
122 struct hlist_head *head;
123 u32 i;
124
125 if (!bat_priv->dat.hash)
126 return;
127
128 for (i = 0; i < bat_priv->dat.hash->size; i++) {
129 head = &bat_priv->dat.hash->table[i];
130 list_lock = &bat_priv->dat.hash->list_locks[i];
131
132 spin_lock_bh(list_lock);
133 hlist_for_each_entry_safe(dat_entry, node_tmp, head,
134 hash_entry) {
135
136
137
138 if (to_purge && !to_purge(dat_entry))
139 continue;
140
141 hlist_del_rcu(&dat_entry->hash_entry);
142 batadv_dat_entry_put(dat_entry);
143 }
144 spin_unlock_bh(list_lock);
145 }
146}
147
148
149
150
151
152
153static void batadv_dat_purge(struct work_struct *work)
154{
155 struct delayed_work *delayed_work;
156 struct batadv_priv_dat *priv_dat;
157 struct batadv_priv *bat_priv;
158
159 delayed_work = to_delayed_work(work);
160 priv_dat = container_of(delayed_work, struct batadv_priv_dat, work);
161 bat_priv = container_of(priv_dat, struct batadv_priv, dat);
162
163 __batadv_dat_purge(bat_priv, batadv_dat_to_purge);
164 batadv_dat_start_timer(bat_priv);
165}
166
167
168
169
170
171
172
173
174static bool batadv_compare_dat(const struct hlist_node *node, const void *data2)
175{
176 const void *data1 = container_of(node, struct batadv_dat_entry,
177 hash_entry);
178
179 return memcmp(data1, data2, sizeof(__be32)) == 0;
180}
181
182
183
184
185
186
187
188
189static u8 *batadv_arp_hw_src(struct sk_buff *skb, int hdr_size)
190{
191 u8 *addr;
192
193 addr = (u8 *)(skb->data + hdr_size);
194 addr += ETH_HLEN + sizeof(struct arphdr);
195
196 return addr;
197}
198
199
200
201
202
203
204
205
206static __be32 batadv_arp_ip_src(struct sk_buff *skb, int hdr_size)
207{
208 return *(__be32 *)(batadv_arp_hw_src(skb, hdr_size) + ETH_ALEN);
209}
210
211
212
213
214
215
216
217
218static u8 *batadv_arp_hw_dst(struct sk_buff *skb, int hdr_size)
219{
220 return batadv_arp_hw_src(skb, hdr_size) + ETH_ALEN + 4;
221}
222
223
224
225
226
227
228
229
230static __be32 batadv_arp_ip_dst(struct sk_buff *skb, int hdr_size)
231{
232 return *(__be32 *)(batadv_arp_hw_src(skb, hdr_size) + ETH_ALEN * 2 + 4);
233}
234
235
236
237
238
239
240
241
242static u32 batadv_hash_dat(const void *data, u32 size)
243{
244 u32 hash = 0;
245 const struct batadv_dat_entry *dat = data;
246 const unsigned char *key;
247 u32 i;
248
249 key = (const unsigned char *)&dat->ip;
250 for (i = 0; i < sizeof(dat->ip); i++) {
251 hash += key[i];
252 hash += (hash << 10);
253 hash ^= (hash >> 6);
254 }
255
256 key = (const unsigned char *)&dat->vid;
257 for (i = 0; i < sizeof(dat->vid); i++) {
258 hash += key[i];
259 hash += (hash << 10);
260 hash ^= (hash >> 6);
261 }
262
263 hash += (hash << 3);
264 hash ^= (hash >> 11);
265 hash += (hash << 15);
266
267 return hash % size;
268}
269
270
271
272
273
274
275
276
277
278
279static struct batadv_dat_entry *
280batadv_dat_entry_hash_find(struct batadv_priv *bat_priv, __be32 ip,
281 unsigned short vid)
282{
283 struct hlist_head *head;
284 struct batadv_dat_entry to_find, *dat_entry, *dat_entry_tmp = NULL;
285 struct batadv_hashtable *hash = bat_priv->dat.hash;
286 u32 index;
287
288 if (!hash)
289 return NULL;
290
291 to_find.ip = ip;
292 to_find.vid = vid;
293
294 index = batadv_hash_dat(&to_find, hash->size);
295 head = &hash->table[index];
296
297 rcu_read_lock();
298 hlist_for_each_entry_rcu(dat_entry, head, hash_entry) {
299 if (dat_entry->ip != ip)
300 continue;
301
302 if (!kref_get_unless_zero(&dat_entry->refcount))
303 continue;
304
305 dat_entry_tmp = dat_entry;
306 break;
307 }
308 rcu_read_unlock();
309
310 return dat_entry_tmp;
311}
312
313
314
315
316
317
318
319
320static void batadv_dat_entry_add(struct batadv_priv *bat_priv, __be32 ip,
321 u8 *mac_addr, unsigned short vid)
322{
323 struct batadv_dat_entry *dat_entry;
324 int hash_added;
325
326 dat_entry = batadv_dat_entry_hash_find(bat_priv, ip, vid);
327
328 if (dat_entry) {
329 if (!batadv_compare_eth(dat_entry->mac_addr, mac_addr))
330 ether_addr_copy(dat_entry->mac_addr, mac_addr);
331 dat_entry->last_update = jiffies;
332 batadv_dbg(BATADV_DBG_DAT, bat_priv,
333 "Entry updated: %pI4 %pM (vid: %d)\n",
334 &dat_entry->ip, dat_entry->mac_addr,
335 batadv_print_vid(vid));
336 goto out;
337 }
338
339 dat_entry = kmalloc(sizeof(*dat_entry), GFP_ATOMIC);
340 if (!dat_entry)
341 goto out;
342
343 dat_entry->ip = ip;
344 dat_entry->vid = vid;
345 ether_addr_copy(dat_entry->mac_addr, mac_addr);
346 dat_entry->last_update = jiffies;
347 kref_init(&dat_entry->refcount);
348
349 kref_get(&dat_entry->refcount);
350 hash_added = batadv_hash_add(bat_priv->dat.hash, batadv_compare_dat,
351 batadv_hash_dat, dat_entry,
352 &dat_entry->hash_entry);
353
354 if (unlikely(hash_added != 0)) {
355
356 batadv_dat_entry_put(dat_entry);
357 goto out;
358 }
359
360 batadv_dbg(BATADV_DBG_DAT, bat_priv, "New entry added: %pI4 %pM (vid: %d)\n",
361 &dat_entry->ip, dat_entry->mac_addr, batadv_print_vid(vid));
362
363out:
364 if (dat_entry)
365 batadv_dat_entry_put(dat_entry);
366}
367
368#ifdef CONFIG_BATMAN_ADV_DEBUG
369
370
371
372
373
374
375
376
377
378static void batadv_dbg_arp(struct batadv_priv *bat_priv, struct sk_buff *skb,
379 int hdr_size, char *msg)
380{
381 struct batadv_unicast_4addr_packet *unicast_4addr_packet;
382 struct batadv_bcast_packet *bcast_pkt;
383 u8 *orig_addr;
384 __be32 ip_src, ip_dst;
385
386 if (msg)
387 batadv_dbg(BATADV_DBG_DAT, bat_priv, "%s\n", msg);
388
389 ip_src = batadv_arp_ip_src(skb, hdr_size);
390 ip_dst = batadv_arp_ip_dst(skb, hdr_size);
391 batadv_dbg(BATADV_DBG_DAT, bat_priv,
392 "ARP MSG = [src: %pM-%pI4 dst: %pM-%pI4]\n",
393 batadv_arp_hw_src(skb, hdr_size), &ip_src,
394 batadv_arp_hw_dst(skb, hdr_size), &ip_dst);
395
396 if (hdr_size < sizeof(struct batadv_unicast_packet))
397 return;
398
399 unicast_4addr_packet = (struct batadv_unicast_4addr_packet *)skb->data;
400
401 switch (unicast_4addr_packet->u.packet_type) {
402 case BATADV_UNICAST:
403 batadv_dbg(BATADV_DBG_DAT, bat_priv,
404 "* encapsulated within a UNICAST packet\n");
405 break;
406 case BATADV_UNICAST_4ADDR:
407 batadv_dbg(BATADV_DBG_DAT, bat_priv,
408 "* encapsulated within a UNICAST_4ADDR packet (src: %pM)\n",
409 unicast_4addr_packet->src);
410 switch (unicast_4addr_packet->subtype) {
411 case BATADV_P_DAT_DHT_PUT:
412 batadv_dbg(BATADV_DBG_DAT, bat_priv, "* type: DAT_DHT_PUT\n");
413 break;
414 case BATADV_P_DAT_DHT_GET:
415 batadv_dbg(BATADV_DBG_DAT, bat_priv, "* type: DAT_DHT_GET\n");
416 break;
417 case BATADV_P_DAT_CACHE_REPLY:
418 batadv_dbg(BATADV_DBG_DAT, bat_priv,
419 "* type: DAT_CACHE_REPLY\n");
420 break;
421 case BATADV_P_DATA:
422 batadv_dbg(BATADV_DBG_DAT, bat_priv, "* type: DATA\n");
423 break;
424 default:
425 batadv_dbg(BATADV_DBG_DAT, bat_priv, "* type: Unknown (%u)!\n",
426 unicast_4addr_packet->u.packet_type);
427 }
428 break;
429 case BATADV_BCAST:
430 bcast_pkt = (struct batadv_bcast_packet *)unicast_4addr_packet;
431 orig_addr = bcast_pkt->orig;
432 batadv_dbg(BATADV_DBG_DAT, bat_priv,
433 "* encapsulated within a BCAST packet (src: %pM)\n",
434 orig_addr);
435 break;
436 default:
437 batadv_dbg(BATADV_DBG_DAT, bat_priv,
438 "* encapsulated within an unknown packet type (0x%x)\n",
439 unicast_4addr_packet->u.packet_type);
440 }
441}
442
443#else
444
445static void batadv_dbg_arp(struct batadv_priv *bat_priv, struct sk_buff *skb,
446 int hdr_size, char *msg)
447{
448}
449
450#endif
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465static bool batadv_is_orig_node_eligible(struct batadv_dat_candidate *res,
466 int select, batadv_dat_addr_t tmp_max,
467 batadv_dat_addr_t max,
468 batadv_dat_addr_t last_max,
469 struct batadv_orig_node *candidate,
470 struct batadv_orig_node *max_orig_node)
471{
472 bool ret = false;
473 int j;
474
475
476 if (!test_bit(BATADV_ORIG_CAPA_HAS_DAT, &candidate->capabilities))
477 goto out;
478
479
480 for (j = 0; j < select; j++)
481 if (res[j].orig_node == candidate)
482 break;
483
484 if (j < select)
485 goto out;
486
487 if (tmp_max > last_max)
488 goto out;
489
490
491
492 if (tmp_max < max)
493 goto out;
494
495
496
497 if (tmp_max == max && max_orig_node &&
498 batadv_compare_eth(candidate->orig, max_orig_node->orig) > 0)
499 goto out;
500
501 ret = true;
502out:
503 return ret;
504}
505
506
507
508
509
510
511
512
513
514static void batadv_choose_next_candidate(struct batadv_priv *bat_priv,
515 struct batadv_dat_candidate *cands,
516 int select, batadv_dat_addr_t ip_key,
517 batadv_dat_addr_t *last_max)
518{
519 batadv_dat_addr_t max = 0;
520 batadv_dat_addr_t tmp_max = 0;
521 struct batadv_orig_node *orig_node, *max_orig_node = NULL;
522 struct batadv_hashtable *hash = bat_priv->orig_hash;
523 struct hlist_head *head;
524 int i;
525
526
527
528
529 cands[select].type = BATADV_DAT_CANDIDATE_NOT_FOUND;
530
531
532
533
534 for (i = 0; i < hash->size; i++) {
535 head = &hash->table[i];
536
537 rcu_read_lock();
538 hlist_for_each_entry_rcu(orig_node, head, hash_entry) {
539
540 tmp_max = BATADV_DAT_ADDR_MAX - orig_node->dat_addr +
541 ip_key;
542
543 if (!batadv_is_orig_node_eligible(cands, select,
544 tmp_max, max,
545 *last_max, orig_node,
546 max_orig_node))
547 continue;
548
549 if (!kref_get_unless_zero(&orig_node->refcount))
550 continue;
551
552 max = tmp_max;
553 if (max_orig_node)
554 batadv_orig_node_put(max_orig_node);
555 max_orig_node = orig_node;
556 }
557 rcu_read_unlock();
558 }
559 if (max_orig_node) {
560 cands[select].type = BATADV_DAT_CANDIDATE_ORIG;
561 cands[select].orig_node = max_orig_node;
562 batadv_dbg(BATADV_DBG_DAT, bat_priv,
563 "dat_select_candidates() %d: selected %pM addr=%u dist=%u\n",
564 select, max_orig_node->orig, max_orig_node->dat_addr,
565 max);
566 }
567 *last_max = max;
568}
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583static struct batadv_dat_candidate *
584batadv_dat_select_candidates(struct batadv_priv *bat_priv, __be32 ip_dst,
585 unsigned short vid)
586{
587 int select;
588 batadv_dat_addr_t last_max = BATADV_DAT_ADDR_MAX, ip_key;
589 struct batadv_dat_candidate *res;
590 struct batadv_dat_entry dat;
591
592 if (!bat_priv->orig_hash)
593 return NULL;
594
595 res = kmalloc_array(BATADV_DAT_CANDIDATES_NUM, sizeof(*res),
596 GFP_ATOMIC);
597 if (!res)
598 return NULL;
599
600 dat.ip = ip_dst;
601 dat.vid = vid;
602 ip_key = (batadv_dat_addr_t)batadv_hash_dat(&dat,
603 BATADV_DAT_ADDR_MAX);
604
605 batadv_dbg(BATADV_DBG_DAT, bat_priv,
606 "%s(): IP=%pI4 hash(IP)=%u\n", __func__, &ip_dst,
607 ip_key);
608
609 for (select = 0; select < BATADV_DAT_CANDIDATES_NUM; select++)
610 batadv_choose_next_candidate(bat_priv, res, select, ip_key,
611 &last_max);
612
613 return res;
614}
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630static bool batadv_dat_send_data(struct batadv_priv *bat_priv,
631 struct sk_buff *skb, __be32 ip,
632 unsigned short vid, int packet_subtype)
633{
634 int i;
635 bool ret = false;
636 int send_status;
637 struct batadv_neigh_node *neigh_node = NULL;
638 struct sk_buff *tmp_skb;
639 struct batadv_dat_candidate *cand;
640
641 cand = batadv_dat_select_candidates(bat_priv, ip, vid);
642 if (!cand)
643 goto out;
644
645 batadv_dbg(BATADV_DBG_DAT, bat_priv, "DHT_SEND for %pI4\n", &ip);
646
647 for (i = 0; i < BATADV_DAT_CANDIDATES_NUM; i++) {
648 if (cand[i].type == BATADV_DAT_CANDIDATE_NOT_FOUND)
649 continue;
650
651 neigh_node = batadv_orig_router_get(cand[i].orig_node,
652 BATADV_IF_DEFAULT);
653 if (!neigh_node)
654 goto free_orig;
655
656 tmp_skb = pskb_copy_for_clone(skb, GFP_ATOMIC);
657 if (!batadv_send_skb_prepare_unicast_4addr(bat_priv, tmp_skb,
658 cand[i].orig_node,
659 packet_subtype)) {
660 kfree_skb(tmp_skb);
661 goto free_neigh;
662 }
663
664 send_status = batadv_send_unicast_skb(tmp_skb, neigh_node);
665 if (send_status == NET_XMIT_SUCCESS) {
666
667 switch (packet_subtype) {
668 case BATADV_P_DAT_DHT_GET:
669 batadv_inc_counter(bat_priv,
670 BATADV_CNT_DAT_GET_TX);
671 break;
672 case BATADV_P_DAT_DHT_PUT:
673 batadv_inc_counter(bat_priv,
674 BATADV_CNT_DAT_PUT_TX);
675 break;
676 }
677
678
679 ret = true;
680 }
681free_neigh:
682 batadv_neigh_node_put(neigh_node);
683free_orig:
684 batadv_orig_node_put(cand[i].orig_node);
685 }
686
687out:
688 kfree(cand);
689 return ret;
690}
691
692
693
694
695
696
697static void batadv_dat_tvlv_container_update(struct batadv_priv *bat_priv)
698{
699 char dat_mode;
700
701 dat_mode = atomic_read(&bat_priv->distributed_arp_table);
702
703 switch (dat_mode) {
704 case 0:
705 batadv_tvlv_container_unregister(bat_priv, BATADV_TVLV_DAT, 1);
706 break;
707 case 1:
708 batadv_tvlv_container_register(bat_priv, BATADV_TVLV_DAT, 1,
709 NULL, 0);
710 break;
711 }
712}
713
714
715
716
717
718
719void batadv_dat_status_update(struct net_device *net_dev)
720{
721 struct batadv_priv *bat_priv = netdev_priv(net_dev);
722
723 batadv_dat_tvlv_container_update(bat_priv);
724}
725
726
727
728
729
730
731
732
733
734static void batadv_dat_tvlv_ogm_handler_v1(struct batadv_priv *bat_priv,
735 struct batadv_orig_node *orig,
736 u8 flags,
737 void *tvlv_value, u16 tvlv_value_len)
738{
739 if (flags & BATADV_TVLV_HANDLER_OGM_CIFNOTFND)
740 clear_bit(BATADV_ORIG_CAPA_HAS_DAT, &orig->capabilities);
741 else
742 set_bit(BATADV_ORIG_CAPA_HAS_DAT, &orig->capabilities);
743}
744
745
746
747
748
749static void batadv_dat_hash_free(struct batadv_priv *bat_priv)
750{
751 if (!bat_priv->dat.hash)
752 return;
753
754 __batadv_dat_purge(bat_priv, NULL);
755
756 batadv_hash_destroy(bat_priv->dat.hash);
757
758 bat_priv->dat.hash = NULL;
759}
760
761
762
763
764
765
766
767int batadv_dat_init(struct batadv_priv *bat_priv)
768{
769 if (bat_priv->dat.hash)
770 return 0;
771
772 bat_priv->dat.hash = batadv_hash_new(1024);
773
774 if (!bat_priv->dat.hash)
775 return -ENOMEM;
776
777 batadv_dat_start_timer(bat_priv);
778
779 batadv_tvlv_handler_register(bat_priv, batadv_dat_tvlv_ogm_handler_v1,
780 NULL, BATADV_TVLV_DAT, 1,
781 BATADV_TVLV_HANDLER_OGM_CIFNOTFND);
782 batadv_dat_tvlv_container_update(bat_priv);
783 return 0;
784}
785
786
787
788
789
790void batadv_dat_free(struct batadv_priv *bat_priv)
791{
792 batadv_tvlv_container_unregister(bat_priv, BATADV_TVLV_DAT, 1);
793 batadv_tvlv_handler_unregister(bat_priv, BATADV_TVLV_DAT, 1);
794
795 cancel_delayed_work_sync(&bat_priv->dat.work);
796
797 batadv_dat_hash_free(bat_priv);
798}
799
800#ifdef CONFIG_BATMAN_ADV_DEBUGFS
801
802
803
804
805
806
807
808int batadv_dat_cache_seq_print_text(struct seq_file *seq, void *offset)
809{
810 struct net_device *net_dev = (struct net_device *)seq->private;
811 struct batadv_priv *bat_priv = netdev_priv(net_dev);
812 struct batadv_hashtable *hash = bat_priv->dat.hash;
813 struct batadv_dat_entry *dat_entry;
814 struct batadv_hard_iface *primary_if;
815 struct hlist_head *head;
816 unsigned long last_seen_jiffies;
817 int last_seen_msecs, last_seen_secs, last_seen_mins;
818 u32 i;
819
820 primary_if = batadv_seq_print_text_primary_if_get(seq);
821 if (!primary_if)
822 goto out;
823
824 seq_printf(seq, "Distributed ARP Table (%s):\n", net_dev->name);
825 seq_puts(seq,
826 " IPv4 MAC VID last-seen\n");
827
828 for (i = 0; i < hash->size; i++) {
829 head = &hash->table[i];
830
831 rcu_read_lock();
832 hlist_for_each_entry_rcu(dat_entry, head, hash_entry) {
833 last_seen_jiffies = jiffies - dat_entry->last_update;
834 last_seen_msecs = jiffies_to_msecs(last_seen_jiffies);
835 last_seen_mins = last_seen_msecs / 60000;
836 last_seen_msecs = last_seen_msecs % 60000;
837 last_seen_secs = last_seen_msecs / 1000;
838
839 seq_printf(seq, " * %15pI4 %pM %4i %6i:%02i\n",
840 &dat_entry->ip, dat_entry->mac_addr,
841 batadv_print_vid(dat_entry->vid),
842 last_seen_mins, last_seen_secs);
843 }
844 rcu_read_unlock();
845 }
846
847out:
848 if (primary_if)
849 batadv_hardif_put(primary_if);
850 return 0;
851}
852#endif
853
854
855
856
857
858
859
860
861
862static u16 batadv_arp_get_type(struct batadv_priv *bat_priv,
863 struct sk_buff *skb, int hdr_size)
864{
865 struct arphdr *arphdr;
866 struct ethhdr *ethhdr;
867 __be32 ip_src, ip_dst;
868 u8 *hw_src, *hw_dst;
869 u16 type = 0;
870
871
872 if (unlikely(!pskb_may_pull(skb, hdr_size + ETH_HLEN)))
873 goto out;
874
875 ethhdr = (struct ethhdr *)(skb->data + hdr_size);
876
877 if (ethhdr->h_proto != htons(ETH_P_ARP))
878 goto out;
879
880
881 if (unlikely(!pskb_may_pull(skb, hdr_size + ETH_HLEN +
882 arp_hdr_len(skb->dev))))
883 goto out;
884
885 arphdr = (struct arphdr *)(skb->data + hdr_size + ETH_HLEN);
886
887
888 if (arphdr->ar_hrd != htons(ARPHRD_ETHER))
889 goto out;
890
891 if (arphdr->ar_pro != htons(ETH_P_IP))
892 goto out;
893
894 if (arphdr->ar_hln != ETH_ALEN)
895 goto out;
896
897 if (arphdr->ar_pln != 4)
898 goto out;
899
900
901
902
903 ip_src = batadv_arp_ip_src(skb, hdr_size);
904 ip_dst = batadv_arp_ip_dst(skb, hdr_size);
905 if (ipv4_is_loopback(ip_src) || ipv4_is_multicast(ip_src) ||
906 ipv4_is_loopback(ip_dst) || ipv4_is_multicast(ip_dst) ||
907 ipv4_is_zeronet(ip_src) || ipv4_is_lbcast(ip_src) ||
908 ipv4_is_zeronet(ip_dst) || ipv4_is_lbcast(ip_dst))
909 goto out;
910
911 hw_src = batadv_arp_hw_src(skb, hdr_size);
912 if (is_zero_ether_addr(hw_src) || is_multicast_ether_addr(hw_src))
913 goto out;
914
915
916 if (arphdr->ar_op != htons(ARPOP_REQUEST)) {
917 hw_dst = batadv_arp_hw_dst(skb, hdr_size);
918 if (is_zero_ether_addr(hw_dst) ||
919 is_multicast_ether_addr(hw_dst))
920 goto out;
921 }
922
923 type = ntohs(arphdr->ar_op);
924out:
925 return type;
926}
927
928
929
930
931
932
933
934
935
936
937static unsigned short batadv_dat_get_vid(struct sk_buff *skb, int *hdr_size)
938{
939 unsigned short vid;
940
941 vid = batadv_get_vid(skb, *hdr_size);
942
943
944
945
946
947
948 if (vid & BATADV_VLAN_HAS_TAG)
949 *hdr_size += VLAN_HLEN;
950
951 return vid;
952}
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968static struct sk_buff *
969batadv_dat_arp_create_reply(struct batadv_priv *bat_priv, __be32 ip_src,
970 __be32 ip_dst, u8 *hw_src, u8 *hw_dst,
971 unsigned short vid)
972{
973 struct sk_buff *skb;
974
975 skb = arp_create(ARPOP_REPLY, ETH_P_ARP, ip_dst, bat_priv->soft_iface,
976 ip_src, hw_dst, hw_src, hw_dst);
977 if (!skb)
978 return NULL;
979
980 skb_reset_mac_header(skb);
981
982 if (vid & BATADV_VLAN_HAS_TAG)
983 skb = vlan_insert_tag(skb, htons(ETH_P_8021Q),
984 vid & VLAN_VID_MASK);
985
986 return skb;
987}
988
989
990
991
992
993
994
995
996
997
998
999bool batadv_dat_snoop_outgoing_arp_request(struct batadv_priv *bat_priv,
1000 struct sk_buff *skb)
1001{
1002 u16 type = 0;
1003 __be32 ip_dst, ip_src;
1004 u8 *hw_src;
1005 bool ret = false;
1006 struct batadv_dat_entry *dat_entry = NULL;
1007 struct sk_buff *skb_new;
1008 struct net_device *soft_iface = bat_priv->soft_iface;
1009 int hdr_size = 0;
1010 unsigned short vid;
1011
1012 if (!atomic_read(&bat_priv->distributed_arp_table))
1013 goto out;
1014
1015 vid = batadv_dat_get_vid(skb, &hdr_size);
1016
1017 type = batadv_arp_get_type(bat_priv, skb, hdr_size);
1018
1019
1020
1021 if (type != ARPOP_REQUEST)
1022 goto out;
1023
1024 batadv_dbg_arp(bat_priv, skb, hdr_size, "Parsing outgoing ARP REQUEST");
1025
1026 ip_src = batadv_arp_ip_src(skb, hdr_size);
1027 hw_src = batadv_arp_hw_src(skb, hdr_size);
1028 ip_dst = batadv_arp_ip_dst(skb, hdr_size);
1029
1030 batadv_dat_entry_add(bat_priv, ip_src, hw_src, vid);
1031
1032 dat_entry = batadv_dat_entry_hash_find(bat_priv, ip_dst, vid);
1033 if (dat_entry) {
1034
1035
1036
1037
1038
1039
1040
1041
1042 if (batadv_is_my_client(bat_priv, dat_entry->mac_addr, vid)) {
1043 ret = true;
1044 goto out;
1045 }
1046
1047
1048
1049
1050
1051
1052 if (!batadv_bla_check_claim(bat_priv,
1053 dat_entry->mac_addr, vid)) {
1054 batadv_dbg(BATADV_DBG_DAT, bat_priv,
1055 "Device %pM claimed by another backbone gw. Don't send ARP reply!",
1056 dat_entry->mac_addr);
1057 ret = true;
1058 goto out;
1059 }
1060
1061 skb_new = batadv_dat_arp_create_reply(bat_priv, ip_dst, ip_src,
1062 dat_entry->mac_addr,
1063 hw_src, vid);
1064 if (!skb_new)
1065 goto out;
1066
1067 skb_new->protocol = eth_type_trans(skb_new, soft_iface);
1068
1069 batadv_inc_counter(bat_priv, BATADV_CNT_RX);
1070 batadv_add_counter(bat_priv, BATADV_CNT_RX_BYTES,
1071 skb->len + ETH_HLEN + hdr_size);
1072
1073 netif_rx(skb_new);
1074 batadv_dbg(BATADV_DBG_DAT, bat_priv, "ARP request replied locally\n");
1075 ret = true;
1076 } else {
1077
1078 ret = batadv_dat_send_data(bat_priv, skb, ip_dst, vid,
1079 BATADV_P_DAT_DHT_GET);
1080 }
1081out:
1082 if (dat_entry)
1083 batadv_dat_entry_put(dat_entry);
1084 return ret;
1085}
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096bool batadv_dat_snoop_incoming_arp_request(struct batadv_priv *bat_priv,
1097 struct sk_buff *skb, int hdr_size)
1098{
1099 u16 type;
1100 __be32 ip_src, ip_dst;
1101 u8 *hw_src;
1102 struct sk_buff *skb_new;
1103 struct batadv_dat_entry *dat_entry = NULL;
1104 bool ret = false;
1105 unsigned short vid;
1106 int err;
1107
1108 if (!atomic_read(&bat_priv->distributed_arp_table))
1109 goto out;
1110
1111 vid = batadv_dat_get_vid(skb, &hdr_size);
1112
1113 type = batadv_arp_get_type(bat_priv, skb, hdr_size);
1114 if (type != ARPOP_REQUEST)
1115 goto out;
1116
1117 hw_src = batadv_arp_hw_src(skb, hdr_size);
1118 ip_src = batadv_arp_ip_src(skb, hdr_size);
1119 ip_dst = batadv_arp_ip_dst(skb, hdr_size);
1120
1121 batadv_dbg_arp(bat_priv, skb, hdr_size, "Parsing incoming ARP REQUEST");
1122
1123 batadv_dat_entry_add(bat_priv, ip_src, hw_src, vid);
1124
1125 dat_entry = batadv_dat_entry_hash_find(bat_priv, ip_dst, vid);
1126 if (!dat_entry)
1127 goto out;
1128
1129 skb_new = batadv_dat_arp_create_reply(bat_priv, ip_dst, ip_src,
1130 dat_entry->mac_addr, hw_src, vid);
1131 if (!skb_new)
1132 goto out;
1133
1134
1135
1136
1137
1138 if (hdr_size == sizeof(struct batadv_unicast_4addr_packet))
1139 err = batadv_send_skb_via_tt_4addr(bat_priv, skb_new,
1140 BATADV_P_DAT_CACHE_REPLY,
1141 NULL, vid);
1142 else
1143 err = batadv_send_skb_via_tt(bat_priv, skb_new, NULL, vid);
1144
1145 if (err != NET_XMIT_DROP) {
1146 batadv_inc_counter(bat_priv, BATADV_CNT_DAT_CACHED_REPLY_TX);
1147 ret = true;
1148 }
1149out:
1150 if (dat_entry)
1151 batadv_dat_entry_put(dat_entry);
1152 if (ret)
1153 kfree_skb(skb);
1154 return ret;
1155}
1156
1157
1158
1159
1160
1161
1162void batadv_dat_snoop_outgoing_arp_reply(struct batadv_priv *bat_priv,
1163 struct sk_buff *skb)
1164{
1165 u16 type;
1166 __be32 ip_src, ip_dst;
1167 u8 *hw_src, *hw_dst;
1168 int hdr_size = 0;
1169 unsigned short vid;
1170
1171 if (!atomic_read(&bat_priv->distributed_arp_table))
1172 return;
1173
1174 vid = batadv_dat_get_vid(skb, &hdr_size);
1175
1176 type = batadv_arp_get_type(bat_priv, skb, hdr_size);
1177 if (type != ARPOP_REPLY)
1178 return;
1179
1180 batadv_dbg_arp(bat_priv, skb, hdr_size, "Parsing outgoing ARP REPLY");
1181
1182 hw_src = batadv_arp_hw_src(skb, hdr_size);
1183 ip_src = batadv_arp_ip_src(skb, hdr_size);
1184 hw_dst = batadv_arp_hw_dst(skb, hdr_size);
1185 ip_dst = batadv_arp_ip_dst(skb, hdr_size);
1186
1187 batadv_dat_entry_add(bat_priv, ip_src, hw_src, vid);
1188 batadv_dat_entry_add(bat_priv, ip_dst, hw_dst, vid);
1189
1190
1191
1192
1193 batadv_dat_send_data(bat_priv, skb, ip_src, vid, BATADV_P_DAT_DHT_PUT);
1194 batadv_dat_send_data(bat_priv, skb, ip_dst, vid, BATADV_P_DAT_DHT_PUT);
1195}
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207bool batadv_dat_snoop_incoming_arp_reply(struct batadv_priv *bat_priv,
1208 struct sk_buff *skb, int hdr_size)
1209{
1210 struct batadv_dat_entry *dat_entry = NULL;
1211 u16 type;
1212 __be32 ip_src, ip_dst;
1213 u8 *hw_src, *hw_dst;
1214 bool dropped = false;
1215 unsigned short vid;
1216
1217 if (!atomic_read(&bat_priv->distributed_arp_table))
1218 goto out;
1219
1220 vid = batadv_dat_get_vid(skb, &hdr_size);
1221
1222 type = batadv_arp_get_type(bat_priv, skb, hdr_size);
1223 if (type != ARPOP_REPLY)
1224 goto out;
1225
1226 batadv_dbg_arp(bat_priv, skb, hdr_size, "Parsing incoming ARP REPLY");
1227
1228 hw_src = batadv_arp_hw_src(skb, hdr_size);
1229 ip_src = batadv_arp_ip_src(skb, hdr_size);
1230 hw_dst = batadv_arp_hw_dst(skb, hdr_size);
1231 ip_dst = batadv_arp_ip_dst(skb, hdr_size);
1232
1233
1234
1235
1236
1237
1238
1239 dat_entry = batadv_dat_entry_hash_find(bat_priv, ip_src, vid);
1240 if (dat_entry && batadv_compare_eth(hw_src, dat_entry->mac_addr)) {
1241 batadv_dbg(BATADV_DBG_DAT, bat_priv, "Doubled ARP reply removed: ARP MSG = [src: %pM-%pI4 dst: %pM-%pI4]; dat_entry: %pM-%pI4\n",
1242 hw_src, &ip_src, hw_dst, &ip_dst,
1243 dat_entry->mac_addr, &dat_entry->ip);
1244 dropped = true;
1245 goto out;
1246 }
1247
1248
1249
1250
1251 batadv_dat_entry_add(bat_priv, ip_src, hw_src, vid);
1252 batadv_dat_entry_add(bat_priv, ip_dst, hw_dst, vid);
1253
1254
1255
1256
1257
1258
1259
1260 if (!batadv_bla_check_claim(bat_priv, hw_src, vid)) {
1261 batadv_dbg(BATADV_DBG_DAT, bat_priv,
1262 "Device %pM claimed by another backbone gw. Drop ARP reply.\n",
1263 hw_src);
1264 dropped = true;
1265 goto out;
1266 }
1267
1268
1269
1270
1271 dropped = !batadv_is_my_client(bat_priv, hw_dst, vid);
1272
1273
1274
1275
1276 dropped |= batadv_is_my_client(bat_priv, hw_src, vid);
1277out:
1278 if (dropped)
1279 kfree_skb(skb);
1280 if (dat_entry)
1281 batadv_dat_entry_put(dat_entry);
1282
1283 return dropped;
1284}
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294bool batadv_dat_drop_broadcast_packet(struct batadv_priv *bat_priv,
1295 struct batadv_forw_packet *forw_packet)
1296{
1297 u16 type;
1298 __be32 ip_dst;
1299 struct batadv_dat_entry *dat_entry = NULL;
1300 bool ret = false;
1301 int hdr_size = sizeof(struct batadv_bcast_packet);
1302 unsigned short vid;
1303
1304 if (!atomic_read(&bat_priv->distributed_arp_table))
1305 goto out;
1306
1307
1308
1309
1310 if (batadv_forw_packet_is_rebroadcast(forw_packet))
1311 goto out;
1312
1313 vid = batadv_dat_get_vid(forw_packet->skb, &hdr_size);
1314
1315 type = batadv_arp_get_type(bat_priv, forw_packet->skb, hdr_size);
1316 if (type != ARPOP_REQUEST)
1317 goto out;
1318
1319 ip_dst = batadv_arp_ip_dst(forw_packet->skb, hdr_size);
1320 dat_entry = batadv_dat_entry_hash_find(bat_priv, ip_dst, vid);
1321
1322 if (!dat_entry) {
1323 batadv_dbg(BATADV_DBG_DAT, bat_priv,
1324 "ARP Request for %pI4: fallback\n", &ip_dst);
1325 goto out;
1326 }
1327
1328 batadv_dbg(BATADV_DBG_DAT, bat_priv,
1329 "ARP Request for %pI4: fallback prevented\n", &ip_dst);
1330 ret = true;
1331
1332out:
1333 if (dat_entry)
1334 batadv_dat_entry_put(dat_entry);
1335 return ret;
1336}
1337