blob: ee9cead7654502cf5f9b3db1de7870c88babd46d [file] [log] [blame]
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001// SPDX-License-Identifier: GPL-2.0
Olivier Deprez157378f2022-04-04 15:47:50 +02002/* Copyright (C) 2011-2020 B.A.T.M.A.N. contributors:
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00003 *
4 * Simon Wunderlich
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00005 */
6
7#include "bridge_loop_avoidance.h"
8#include "main.h"
9
10#include <linux/atomic.h>
11#include <linux/byteorder/generic.h>
12#include <linux/compiler.h>
13#include <linux/crc16.h>
14#include <linux/errno.h>
15#include <linux/etherdevice.h>
16#include <linux/gfp.h>
17#include <linux/if_arp.h>
18#include <linux/if_ether.h>
19#include <linux/if_vlan.h>
20#include <linux/jhash.h>
21#include <linux/jiffies.h>
22#include <linux/kernel.h>
23#include <linux/kref.h>
24#include <linux/list.h>
25#include <linux/lockdep.h>
26#include <linux/netdevice.h>
27#include <linux/netlink.h>
Olivier Deprez0e641232021-09-23 10:07:05 +020028#include <linux/preempt.h>
Andrew Scullb4b6d4a2019-01-02 15:54:55 +000029#include <linux/rculist.h>
30#include <linux/rcupdate.h>
31#include <linux/seq_file.h>
32#include <linux/skbuff.h>
33#include <linux/slab.h>
34#include <linux/spinlock.h>
35#include <linux/stddef.h>
36#include <linux/string.h>
37#include <linux/workqueue.h>
38#include <net/arp.h>
39#include <net/genetlink.h>
40#include <net/netlink.h>
41#include <net/sock.h>
42#include <uapi/linux/batadv_packet.h>
43#include <uapi/linux/batman_adv.h>
44
45#include "hard-interface.h"
46#include "hash.h"
47#include "log.h"
48#include "netlink.h"
49#include "originator.h"
50#include "soft-interface.h"
Andrew Scullb4b6d4a2019-01-02 15:54:55 +000051#include "translation-table.h"
52
53static const u8 batadv_announce_mac[4] = {0x43, 0x05, 0x43, 0x05};
54
55static void batadv_bla_periodic_work(struct work_struct *work);
56static void
57batadv_bla_send_announce(struct batadv_priv *bat_priv,
58 struct batadv_bla_backbone_gw *backbone_gw);
59
60/**
61 * batadv_choose_claim() - choose the right bucket for a claim.
62 * @data: data to hash
63 * @size: size of the hash table
64 *
65 * Return: the hash index of the claim
66 */
67static inline u32 batadv_choose_claim(const void *data, u32 size)
68{
69 struct batadv_bla_claim *claim = (struct batadv_bla_claim *)data;
70 u32 hash = 0;
71
72 hash = jhash(&claim->addr, sizeof(claim->addr), hash);
73 hash = jhash(&claim->vid, sizeof(claim->vid), hash);
74
75 return hash % size;
76}
77
78/**
79 * batadv_choose_backbone_gw() - choose the right bucket for a backbone gateway.
80 * @data: data to hash
81 * @size: size of the hash table
82 *
83 * Return: the hash index of the backbone gateway
84 */
85static inline u32 batadv_choose_backbone_gw(const void *data, u32 size)
86{
Olivier Deprez0e641232021-09-23 10:07:05 +020087 const struct batadv_bla_backbone_gw *gw;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +000088 u32 hash = 0;
89
Olivier Deprez0e641232021-09-23 10:07:05 +020090 gw = (struct batadv_bla_backbone_gw *)data;
91 hash = jhash(&gw->orig, sizeof(gw->orig), hash);
92 hash = jhash(&gw->vid, sizeof(gw->vid), hash);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +000093
94 return hash % size;
95}
96
97/**
98 * batadv_compare_backbone_gw() - compare address and vid of two backbone gws
99 * @node: list node of the first entry to compare
100 * @data2: pointer to the second backbone gateway
101 *
102 * Return: true if the backbones have the same data, false otherwise
103 */
104static bool batadv_compare_backbone_gw(const struct hlist_node *node,
105 const void *data2)
106{
107 const void *data1 = container_of(node, struct batadv_bla_backbone_gw,
108 hash_entry);
109 const struct batadv_bla_backbone_gw *gw1 = data1;
110 const struct batadv_bla_backbone_gw *gw2 = data2;
111
112 if (!batadv_compare_eth(gw1->orig, gw2->orig))
113 return false;
114
115 if (gw1->vid != gw2->vid)
116 return false;
117
118 return true;
119}
120
121/**
122 * batadv_compare_claim() - compare address and vid of two claims
123 * @node: list node of the first entry to compare
124 * @data2: pointer to the second claims
125 *
126 * Return: true if the claim have the same data, 0 otherwise
127 */
128static bool batadv_compare_claim(const struct hlist_node *node,
129 const void *data2)
130{
131 const void *data1 = container_of(node, struct batadv_bla_claim,
132 hash_entry);
133 const struct batadv_bla_claim *cl1 = data1;
134 const struct batadv_bla_claim *cl2 = data2;
135
136 if (!batadv_compare_eth(cl1->addr, cl2->addr))
137 return false;
138
139 if (cl1->vid != cl2->vid)
140 return false;
141
142 return true;
143}
144
145/**
146 * batadv_backbone_gw_release() - release backbone gw from lists and queue for
147 * free after rcu grace period
148 * @ref: kref pointer of the backbone gw
149 */
150static void batadv_backbone_gw_release(struct kref *ref)
151{
152 struct batadv_bla_backbone_gw *backbone_gw;
153
154 backbone_gw = container_of(ref, struct batadv_bla_backbone_gw,
155 refcount);
156
157 kfree_rcu(backbone_gw, rcu);
158}
159
160/**
161 * batadv_backbone_gw_put() - decrement the backbone gw refcounter and possibly
162 * release it
163 * @backbone_gw: backbone gateway to be free'd
164 */
165static void batadv_backbone_gw_put(struct batadv_bla_backbone_gw *backbone_gw)
166{
167 kref_put(&backbone_gw->refcount, batadv_backbone_gw_release);
168}
169
170/**
171 * batadv_claim_release() - release claim from lists and queue for free after
172 * rcu grace period
173 * @ref: kref pointer of the claim
174 */
175static void batadv_claim_release(struct kref *ref)
176{
177 struct batadv_bla_claim *claim;
178 struct batadv_bla_backbone_gw *old_backbone_gw;
179
180 claim = container_of(ref, struct batadv_bla_claim, refcount);
181
182 spin_lock_bh(&claim->backbone_lock);
183 old_backbone_gw = claim->backbone_gw;
184 claim->backbone_gw = NULL;
185 spin_unlock_bh(&claim->backbone_lock);
186
187 spin_lock_bh(&old_backbone_gw->crc_lock);
188 old_backbone_gw->crc ^= crc16(0, claim->addr, ETH_ALEN);
189 spin_unlock_bh(&old_backbone_gw->crc_lock);
190
191 batadv_backbone_gw_put(old_backbone_gw);
192
193 kfree_rcu(claim, rcu);
194}
195
196/**
197 * batadv_claim_put() - decrement the claim refcounter and possibly release it
198 * @claim: claim to be free'd
199 */
200static void batadv_claim_put(struct batadv_bla_claim *claim)
201{
202 kref_put(&claim->refcount, batadv_claim_release);
203}
204
205/**
206 * batadv_claim_hash_find() - looks for a claim in the claim hash
207 * @bat_priv: the bat priv with all the soft interface information
208 * @data: search data (may be local/static data)
209 *
210 * Return: claim if found or NULL otherwise.
211 */
212static struct batadv_bla_claim *
213batadv_claim_hash_find(struct batadv_priv *bat_priv,
214 struct batadv_bla_claim *data)
215{
216 struct batadv_hashtable *hash = bat_priv->bla.claim_hash;
217 struct hlist_head *head;
218 struct batadv_bla_claim *claim;
219 struct batadv_bla_claim *claim_tmp = NULL;
220 int index;
221
222 if (!hash)
223 return NULL;
224
225 index = batadv_choose_claim(data, hash->size);
226 head = &hash->table[index];
227
228 rcu_read_lock();
229 hlist_for_each_entry_rcu(claim, head, hash_entry) {
230 if (!batadv_compare_claim(&claim->hash_entry, data))
231 continue;
232
233 if (!kref_get_unless_zero(&claim->refcount))
234 continue;
235
236 claim_tmp = claim;
237 break;
238 }
239 rcu_read_unlock();
240
241 return claim_tmp;
242}
243
244/**
245 * batadv_backbone_hash_find() - looks for a backbone gateway in the hash
246 * @bat_priv: the bat priv with all the soft interface information
247 * @addr: the address of the originator
248 * @vid: the VLAN ID
249 *
250 * Return: backbone gateway if found or NULL otherwise
251 */
252static struct batadv_bla_backbone_gw *
253batadv_backbone_hash_find(struct batadv_priv *bat_priv, u8 *addr,
254 unsigned short vid)
255{
256 struct batadv_hashtable *hash = bat_priv->bla.backbone_hash;
257 struct hlist_head *head;
258 struct batadv_bla_backbone_gw search_entry, *backbone_gw;
259 struct batadv_bla_backbone_gw *backbone_gw_tmp = NULL;
260 int index;
261
262 if (!hash)
263 return NULL;
264
265 ether_addr_copy(search_entry.orig, addr);
266 search_entry.vid = vid;
267
268 index = batadv_choose_backbone_gw(&search_entry, hash->size);
269 head = &hash->table[index];
270
271 rcu_read_lock();
272 hlist_for_each_entry_rcu(backbone_gw, head, hash_entry) {
273 if (!batadv_compare_backbone_gw(&backbone_gw->hash_entry,
274 &search_entry))
275 continue;
276
277 if (!kref_get_unless_zero(&backbone_gw->refcount))
278 continue;
279
280 backbone_gw_tmp = backbone_gw;
281 break;
282 }
283 rcu_read_unlock();
284
285 return backbone_gw_tmp;
286}
287
288/**
289 * batadv_bla_del_backbone_claims() - delete all claims for a backbone
290 * @backbone_gw: backbone gateway where the claims should be removed
291 */
292static void
293batadv_bla_del_backbone_claims(struct batadv_bla_backbone_gw *backbone_gw)
294{
295 struct batadv_hashtable *hash;
296 struct hlist_node *node_tmp;
297 struct hlist_head *head;
298 struct batadv_bla_claim *claim;
299 int i;
300 spinlock_t *list_lock; /* protects write access to the hash lists */
301
302 hash = backbone_gw->bat_priv->bla.claim_hash;
303 if (!hash)
304 return;
305
306 for (i = 0; i < hash->size; i++) {
307 head = &hash->table[i];
308 list_lock = &hash->list_locks[i];
309
310 spin_lock_bh(list_lock);
311 hlist_for_each_entry_safe(claim, node_tmp,
312 head, hash_entry) {
313 if (claim->backbone_gw != backbone_gw)
314 continue;
315
316 batadv_claim_put(claim);
317 hlist_del_rcu(&claim->hash_entry);
318 }
319 spin_unlock_bh(list_lock);
320 }
321
322 /* all claims gone, initialize CRC */
323 spin_lock_bh(&backbone_gw->crc_lock);
324 backbone_gw->crc = BATADV_BLA_CRC_INIT;
325 spin_unlock_bh(&backbone_gw->crc_lock);
326}
327
328/**
329 * batadv_bla_send_claim() - sends a claim frame according to the provided info
330 * @bat_priv: the bat priv with all the soft interface information
331 * @mac: the mac address to be announced within the claim
332 * @vid: the VLAN ID
333 * @claimtype: the type of the claim (CLAIM, UNCLAIM, ANNOUNCE, ...)
334 */
335static void batadv_bla_send_claim(struct batadv_priv *bat_priv, u8 *mac,
336 unsigned short vid, int claimtype)
337{
338 struct sk_buff *skb;
339 struct ethhdr *ethhdr;
340 struct batadv_hard_iface *primary_if;
341 struct net_device *soft_iface;
342 u8 *hw_src;
343 struct batadv_bla_claim_dst local_claim_dest;
344 __be32 zeroip = 0;
345
346 primary_if = batadv_primary_if_get_selected(bat_priv);
347 if (!primary_if)
348 return;
349
350 memcpy(&local_claim_dest, &bat_priv->bla.claim_dest,
351 sizeof(local_claim_dest));
352 local_claim_dest.type = claimtype;
353
354 soft_iface = primary_if->soft_iface;
355
356 skb = arp_create(ARPOP_REPLY, ETH_P_ARP,
357 /* IP DST: 0.0.0.0 */
358 zeroip,
359 primary_if->soft_iface,
360 /* IP SRC: 0.0.0.0 */
361 zeroip,
362 /* Ethernet DST: Broadcast */
363 NULL,
364 /* Ethernet SRC/HW SRC: originator mac */
365 primary_if->net_dev->dev_addr,
366 /* HW DST: FF:43:05:XX:YY:YY
367 * with XX = claim type
368 * and YY:YY = group id
369 */
370 (u8 *)&local_claim_dest);
371
372 if (!skb)
373 goto out;
374
375 ethhdr = (struct ethhdr *)skb->data;
376 hw_src = (u8 *)ethhdr + ETH_HLEN + sizeof(struct arphdr);
377
378 /* now we pretend that the client would have sent this ... */
379 switch (claimtype) {
380 case BATADV_CLAIM_TYPE_CLAIM:
381 /* normal claim frame
382 * set Ethernet SRC to the clients mac
383 */
384 ether_addr_copy(ethhdr->h_source, mac);
385 batadv_dbg(BATADV_DBG_BLA, bat_priv,
386 "%s(): CLAIM %pM on vid %d\n", __func__, mac,
387 batadv_print_vid(vid));
388 break;
389 case BATADV_CLAIM_TYPE_UNCLAIM:
390 /* unclaim frame
391 * set HW SRC to the clients mac
392 */
393 ether_addr_copy(hw_src, mac);
394 batadv_dbg(BATADV_DBG_BLA, bat_priv,
395 "%s(): UNCLAIM %pM on vid %d\n", __func__, mac,
396 batadv_print_vid(vid));
397 break;
398 case BATADV_CLAIM_TYPE_ANNOUNCE:
399 /* announcement frame
400 * set HW SRC to the special mac containg the crc
401 */
402 ether_addr_copy(hw_src, mac);
403 batadv_dbg(BATADV_DBG_BLA, bat_priv,
404 "%s(): ANNOUNCE of %pM on vid %d\n", __func__,
405 ethhdr->h_source, batadv_print_vid(vid));
406 break;
407 case BATADV_CLAIM_TYPE_REQUEST:
408 /* request frame
409 * set HW SRC and header destination to the receiving backbone
410 * gws mac
411 */
412 ether_addr_copy(hw_src, mac);
413 ether_addr_copy(ethhdr->h_dest, mac);
414 batadv_dbg(BATADV_DBG_BLA, bat_priv,
415 "%s(): REQUEST of %pM to %pM on vid %d\n", __func__,
416 ethhdr->h_source, ethhdr->h_dest,
417 batadv_print_vid(vid));
418 break;
419 case BATADV_CLAIM_TYPE_LOOPDETECT:
420 ether_addr_copy(ethhdr->h_source, mac);
421 batadv_dbg(BATADV_DBG_BLA, bat_priv,
422 "%s(): LOOPDETECT of %pM to %pM on vid %d\n",
423 __func__, ethhdr->h_source, ethhdr->h_dest,
424 batadv_print_vid(vid));
425
426 break;
427 }
428
429 if (vid & BATADV_VLAN_HAS_TAG) {
430 skb = vlan_insert_tag(skb, htons(ETH_P_8021Q),
431 vid & VLAN_VID_MASK);
432 if (!skb)
433 goto out;
434 }
435
436 skb_reset_mac_header(skb);
437 skb->protocol = eth_type_trans(skb, soft_iface);
438 batadv_inc_counter(bat_priv, BATADV_CNT_RX);
439 batadv_add_counter(bat_priv, BATADV_CNT_RX_BYTES,
440 skb->len + ETH_HLEN);
441
Olivier Deprez0e641232021-09-23 10:07:05 +0200442 if (in_interrupt())
443 netif_rx(skb);
444 else
445 netif_rx_ni(skb);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000446out:
447 if (primary_if)
448 batadv_hardif_put(primary_if);
449}
450
451/**
452 * batadv_bla_loopdetect_report() - worker for reporting the loop
453 * @work: work queue item
454 *
455 * Throws an uevent, as the loopdetect check function can't do that itself
456 * since the kernel may sleep while throwing uevents.
457 */
458static void batadv_bla_loopdetect_report(struct work_struct *work)
459{
460 struct batadv_bla_backbone_gw *backbone_gw;
461 struct batadv_priv *bat_priv;
462 char vid_str[6] = { '\0' };
463
464 backbone_gw = container_of(work, struct batadv_bla_backbone_gw,
465 report_work);
466 bat_priv = backbone_gw->bat_priv;
467
468 batadv_info(bat_priv->soft_iface,
469 "Possible loop on VLAN %d detected which can't be handled by BLA - please check your network setup!\n",
470 batadv_print_vid(backbone_gw->vid));
471 snprintf(vid_str, sizeof(vid_str), "%d",
472 batadv_print_vid(backbone_gw->vid));
473 vid_str[sizeof(vid_str) - 1] = 0;
474
475 batadv_throw_uevent(bat_priv, BATADV_UEV_BLA, BATADV_UEV_LOOPDETECT,
476 vid_str);
477
478 batadv_backbone_gw_put(backbone_gw);
479}
480
481/**
482 * batadv_bla_get_backbone_gw() - finds or creates a backbone gateway
483 * @bat_priv: the bat priv with all the soft interface information
484 * @orig: the mac address of the originator
485 * @vid: the VLAN ID
486 * @own_backbone: set if the requested backbone is local
487 *
488 * Return: the (possibly created) backbone gateway or NULL on error
489 */
490static struct batadv_bla_backbone_gw *
491batadv_bla_get_backbone_gw(struct batadv_priv *bat_priv, u8 *orig,
492 unsigned short vid, bool own_backbone)
493{
494 struct batadv_bla_backbone_gw *entry;
495 struct batadv_orig_node *orig_node;
496 int hash_added;
497
498 entry = batadv_backbone_hash_find(bat_priv, orig, vid);
499
500 if (entry)
501 return entry;
502
503 batadv_dbg(BATADV_DBG_BLA, bat_priv,
504 "%s(): not found (%pM, %d), creating new entry\n", __func__,
505 orig, batadv_print_vid(vid));
506
507 entry = kzalloc(sizeof(*entry), GFP_ATOMIC);
508 if (!entry)
509 return NULL;
510
511 entry->vid = vid;
512 entry->lasttime = jiffies;
513 entry->crc = BATADV_BLA_CRC_INIT;
514 entry->bat_priv = bat_priv;
515 spin_lock_init(&entry->crc_lock);
516 atomic_set(&entry->request_sent, 0);
517 atomic_set(&entry->wait_periods, 0);
518 ether_addr_copy(entry->orig, orig);
519 INIT_WORK(&entry->report_work, batadv_bla_loopdetect_report);
520 kref_init(&entry->refcount);
521
522 kref_get(&entry->refcount);
523 hash_added = batadv_hash_add(bat_priv->bla.backbone_hash,
524 batadv_compare_backbone_gw,
525 batadv_choose_backbone_gw, entry,
526 &entry->hash_entry);
527
528 if (unlikely(hash_added != 0)) {
529 /* hash failed, free the structure */
530 kfree(entry);
531 return NULL;
532 }
533
534 /* this is a gateway now, remove any TT entry on this VLAN */
535 orig_node = batadv_orig_hash_find(bat_priv, orig);
536 if (orig_node) {
537 batadv_tt_global_del_orig(bat_priv, orig_node, vid,
538 "became a backbone gateway");
539 batadv_orig_node_put(orig_node);
540 }
541
542 if (own_backbone) {
543 batadv_bla_send_announce(bat_priv, entry);
544
545 /* this will be decreased in the worker thread */
546 atomic_inc(&entry->request_sent);
547 atomic_set(&entry->wait_periods, BATADV_BLA_WAIT_PERIODS);
548 atomic_inc(&bat_priv->bla.num_requests);
549 }
550
551 return entry;
552}
553
554/**
555 * batadv_bla_update_own_backbone_gw() - updates the own backbone gw for a VLAN
556 * @bat_priv: the bat priv with all the soft interface information
557 * @primary_if: the selected primary interface
558 * @vid: VLAN identifier
559 *
560 * update or add the own backbone gw to make sure we announce
561 * where we receive other backbone gws
562 */
563static void
564batadv_bla_update_own_backbone_gw(struct batadv_priv *bat_priv,
565 struct batadv_hard_iface *primary_if,
566 unsigned short vid)
567{
568 struct batadv_bla_backbone_gw *backbone_gw;
569
570 backbone_gw = batadv_bla_get_backbone_gw(bat_priv,
571 primary_if->net_dev->dev_addr,
572 vid, true);
573 if (unlikely(!backbone_gw))
574 return;
575
576 backbone_gw->lasttime = jiffies;
577 batadv_backbone_gw_put(backbone_gw);
578}
579
580/**
581 * batadv_bla_answer_request() - answer a bla request by sending own claims
582 * @bat_priv: the bat priv with all the soft interface information
583 * @primary_if: interface where the request came on
584 * @vid: the vid where the request came on
585 *
586 * Repeat all of our own claims, and finally send an ANNOUNCE frame
587 * to allow the requester another check if the CRC is correct now.
588 */
589static void batadv_bla_answer_request(struct batadv_priv *bat_priv,
590 struct batadv_hard_iface *primary_if,
591 unsigned short vid)
592{
593 struct hlist_head *head;
594 struct batadv_hashtable *hash;
595 struct batadv_bla_claim *claim;
596 struct batadv_bla_backbone_gw *backbone_gw;
597 int i;
598
599 batadv_dbg(BATADV_DBG_BLA, bat_priv,
600 "%s(): received a claim request, send all of our own claims again\n",
601 __func__);
602
603 backbone_gw = batadv_backbone_hash_find(bat_priv,
604 primary_if->net_dev->dev_addr,
605 vid);
606 if (!backbone_gw)
607 return;
608
609 hash = bat_priv->bla.claim_hash;
610 for (i = 0; i < hash->size; i++) {
611 head = &hash->table[i];
612
613 rcu_read_lock();
614 hlist_for_each_entry_rcu(claim, head, hash_entry) {
615 /* only own claims are interesting */
616 if (claim->backbone_gw != backbone_gw)
617 continue;
618
619 batadv_bla_send_claim(bat_priv, claim->addr, claim->vid,
620 BATADV_CLAIM_TYPE_CLAIM);
621 }
622 rcu_read_unlock();
623 }
624
625 /* finally, send an announcement frame */
626 batadv_bla_send_announce(bat_priv, backbone_gw);
627 batadv_backbone_gw_put(backbone_gw);
628}
629
630/**
631 * batadv_bla_send_request() - send a request to repeat claims
632 * @backbone_gw: the backbone gateway from whom we are out of sync
633 *
634 * When the crc is wrong, ask the backbone gateway for a full table update.
635 * After the request, it will repeat all of his own claims and finally
636 * send an announcement claim with which we can check again.
637 */
638static void batadv_bla_send_request(struct batadv_bla_backbone_gw *backbone_gw)
639{
640 /* first, remove all old entries */
641 batadv_bla_del_backbone_claims(backbone_gw);
642
643 batadv_dbg(BATADV_DBG_BLA, backbone_gw->bat_priv,
644 "Sending REQUEST to %pM\n", backbone_gw->orig);
645
646 /* send request */
647 batadv_bla_send_claim(backbone_gw->bat_priv, backbone_gw->orig,
648 backbone_gw->vid, BATADV_CLAIM_TYPE_REQUEST);
649
650 /* no local broadcasts should be sent or received, for now. */
651 if (!atomic_read(&backbone_gw->request_sent)) {
652 atomic_inc(&backbone_gw->bat_priv->bla.num_requests);
653 atomic_set(&backbone_gw->request_sent, 1);
654 }
655}
656
657/**
658 * batadv_bla_send_announce() - Send an announcement frame
659 * @bat_priv: the bat priv with all the soft interface information
660 * @backbone_gw: our backbone gateway which should be announced
661 */
662static void batadv_bla_send_announce(struct batadv_priv *bat_priv,
663 struct batadv_bla_backbone_gw *backbone_gw)
664{
665 u8 mac[ETH_ALEN];
666 __be16 crc;
667
668 memcpy(mac, batadv_announce_mac, 4);
669 spin_lock_bh(&backbone_gw->crc_lock);
670 crc = htons(backbone_gw->crc);
671 spin_unlock_bh(&backbone_gw->crc_lock);
672 memcpy(&mac[4], &crc, 2);
673
674 batadv_bla_send_claim(bat_priv, mac, backbone_gw->vid,
675 BATADV_CLAIM_TYPE_ANNOUNCE);
676}
677
678/**
679 * batadv_bla_add_claim() - Adds a claim in the claim hash
680 * @bat_priv: the bat priv with all the soft interface information
681 * @mac: the mac address of the claim
682 * @vid: the VLAN ID of the frame
683 * @backbone_gw: the backbone gateway which claims it
684 */
685static void batadv_bla_add_claim(struct batadv_priv *bat_priv,
686 const u8 *mac, const unsigned short vid,
687 struct batadv_bla_backbone_gw *backbone_gw)
688{
689 struct batadv_bla_backbone_gw *old_backbone_gw;
690 struct batadv_bla_claim *claim;
691 struct batadv_bla_claim search_claim;
692 bool remove_crc = false;
693 int hash_added;
694
695 ether_addr_copy(search_claim.addr, mac);
696 search_claim.vid = vid;
697 claim = batadv_claim_hash_find(bat_priv, &search_claim);
698
699 /* create a new claim entry if it does not exist yet. */
700 if (!claim) {
701 claim = kzalloc(sizeof(*claim), GFP_ATOMIC);
702 if (!claim)
703 return;
704
705 ether_addr_copy(claim->addr, mac);
706 spin_lock_init(&claim->backbone_lock);
707 claim->vid = vid;
708 claim->lasttime = jiffies;
709 kref_get(&backbone_gw->refcount);
710 claim->backbone_gw = backbone_gw;
711 kref_init(&claim->refcount);
712
713 batadv_dbg(BATADV_DBG_BLA, bat_priv,
714 "%s(): adding new entry %pM, vid %d to hash ...\n",
715 __func__, mac, batadv_print_vid(vid));
716
717 kref_get(&claim->refcount);
718 hash_added = batadv_hash_add(bat_priv->bla.claim_hash,
719 batadv_compare_claim,
720 batadv_choose_claim, claim,
721 &claim->hash_entry);
722
723 if (unlikely(hash_added != 0)) {
724 /* only local changes happened. */
725 kfree(claim);
726 return;
727 }
728 } else {
729 claim->lasttime = jiffies;
730 if (claim->backbone_gw == backbone_gw)
731 /* no need to register a new backbone */
732 goto claim_free_ref;
733
734 batadv_dbg(BATADV_DBG_BLA, bat_priv,
735 "%s(): changing ownership for %pM, vid %d to gw %pM\n",
736 __func__, mac, batadv_print_vid(vid),
737 backbone_gw->orig);
738
739 remove_crc = true;
740 }
741
742 /* replace backbone_gw atomically and adjust reference counters */
743 spin_lock_bh(&claim->backbone_lock);
744 old_backbone_gw = claim->backbone_gw;
745 kref_get(&backbone_gw->refcount);
746 claim->backbone_gw = backbone_gw;
747 spin_unlock_bh(&claim->backbone_lock);
748
749 if (remove_crc) {
750 /* remove claim address from old backbone_gw */
751 spin_lock_bh(&old_backbone_gw->crc_lock);
752 old_backbone_gw->crc ^= crc16(0, claim->addr, ETH_ALEN);
753 spin_unlock_bh(&old_backbone_gw->crc_lock);
754 }
755
756 batadv_backbone_gw_put(old_backbone_gw);
757
758 /* add claim address to new backbone_gw */
759 spin_lock_bh(&backbone_gw->crc_lock);
760 backbone_gw->crc ^= crc16(0, claim->addr, ETH_ALEN);
761 spin_unlock_bh(&backbone_gw->crc_lock);
762 backbone_gw->lasttime = jiffies;
763
764claim_free_ref:
765 batadv_claim_put(claim);
766}
767
768/**
769 * batadv_bla_claim_get_backbone_gw() - Get valid reference for backbone_gw of
770 * claim
771 * @claim: claim whose backbone_gw should be returned
772 *
773 * Return: valid reference to claim::backbone_gw
774 */
775static struct batadv_bla_backbone_gw *
776batadv_bla_claim_get_backbone_gw(struct batadv_bla_claim *claim)
777{
778 struct batadv_bla_backbone_gw *backbone_gw;
779
780 spin_lock_bh(&claim->backbone_lock);
781 backbone_gw = claim->backbone_gw;
782 kref_get(&backbone_gw->refcount);
783 spin_unlock_bh(&claim->backbone_lock);
784
785 return backbone_gw;
786}
787
788/**
789 * batadv_bla_del_claim() - delete a claim from the claim hash
790 * @bat_priv: the bat priv with all the soft interface information
791 * @mac: mac address of the claim to be removed
792 * @vid: VLAN id for the claim to be removed
793 */
794static void batadv_bla_del_claim(struct batadv_priv *bat_priv,
795 const u8 *mac, const unsigned short vid)
796{
797 struct batadv_bla_claim search_claim, *claim;
David Brazdil0f672f62019-12-10 10:32:29 +0000798 struct batadv_bla_claim *claim_removed_entry;
799 struct hlist_node *claim_removed_node;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000800
801 ether_addr_copy(search_claim.addr, mac);
802 search_claim.vid = vid;
803 claim = batadv_claim_hash_find(bat_priv, &search_claim);
804 if (!claim)
805 return;
806
807 batadv_dbg(BATADV_DBG_BLA, bat_priv, "%s(): %pM, vid %d\n", __func__,
808 mac, batadv_print_vid(vid));
809
David Brazdil0f672f62019-12-10 10:32:29 +0000810 claim_removed_node = batadv_hash_remove(bat_priv->bla.claim_hash,
811 batadv_compare_claim,
812 batadv_choose_claim, claim);
813 if (!claim_removed_node)
814 goto free_claim;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000815
David Brazdil0f672f62019-12-10 10:32:29 +0000816 /* reference from the hash is gone */
817 claim_removed_entry = hlist_entry(claim_removed_node,
818 struct batadv_bla_claim, hash_entry);
819 batadv_claim_put(claim_removed_entry);
820
821free_claim:
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000822 /* don't need the reference from hash_find() anymore */
823 batadv_claim_put(claim);
824}
825
826/**
827 * batadv_handle_announce() - check for ANNOUNCE frame
828 * @bat_priv: the bat priv with all the soft interface information
829 * @an_addr: announcement mac address (ARP Sender HW address)
830 * @backbone_addr: originator address of the sender (Ethernet source MAC)
831 * @vid: the VLAN ID of the frame
832 *
833 * Return: true if handled
834 */
835static bool batadv_handle_announce(struct batadv_priv *bat_priv, u8 *an_addr,
836 u8 *backbone_addr, unsigned short vid)
837{
838 struct batadv_bla_backbone_gw *backbone_gw;
839 u16 backbone_crc, crc;
840
841 if (memcmp(an_addr, batadv_announce_mac, 4) != 0)
842 return false;
843
844 backbone_gw = batadv_bla_get_backbone_gw(bat_priv, backbone_addr, vid,
845 false);
846
847 if (unlikely(!backbone_gw))
848 return true;
849
850 /* handle as ANNOUNCE frame */
851 backbone_gw->lasttime = jiffies;
Olivier Deprez157378f2022-04-04 15:47:50 +0200852 crc = ntohs(*((__force __be16 *)(&an_addr[4])));
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000853
854 batadv_dbg(BATADV_DBG_BLA, bat_priv,
855 "%s(): ANNOUNCE vid %d (sent by %pM)... CRC = %#.4x\n",
856 __func__, batadv_print_vid(vid), backbone_gw->orig, crc);
857
858 spin_lock_bh(&backbone_gw->crc_lock);
859 backbone_crc = backbone_gw->crc;
860 spin_unlock_bh(&backbone_gw->crc_lock);
861
862 if (backbone_crc != crc) {
863 batadv_dbg(BATADV_DBG_BLA, backbone_gw->bat_priv,
864 "%s(): CRC FAILED for %pM/%d (my = %#.4x, sent = %#.4x)\n",
865 __func__, backbone_gw->orig,
866 batadv_print_vid(backbone_gw->vid),
867 backbone_crc, crc);
868
869 batadv_bla_send_request(backbone_gw);
870 } else {
871 /* if we have sent a request and the crc was OK,
872 * we can allow traffic again.
873 */
874 if (atomic_read(&backbone_gw->request_sent)) {
875 atomic_dec(&backbone_gw->bat_priv->bla.num_requests);
876 atomic_set(&backbone_gw->request_sent, 0);
877 }
878 }
879
880 batadv_backbone_gw_put(backbone_gw);
881 return true;
882}
883
884/**
885 * batadv_handle_request() - check for REQUEST frame
886 * @bat_priv: the bat priv with all the soft interface information
887 * @primary_if: the primary hard interface of this batman soft interface
888 * @backbone_addr: backbone address to be requested (ARP sender HW MAC)
889 * @ethhdr: ethernet header of a packet
890 * @vid: the VLAN ID of the frame
891 *
892 * Return: true if handled
893 */
894static bool batadv_handle_request(struct batadv_priv *bat_priv,
895 struct batadv_hard_iface *primary_if,
896 u8 *backbone_addr, struct ethhdr *ethhdr,
897 unsigned short vid)
898{
899 /* check for REQUEST frame */
900 if (!batadv_compare_eth(backbone_addr, ethhdr->h_dest))
901 return false;
902
903 /* sanity check, this should not happen on a normal switch,
904 * we ignore it in this case.
905 */
906 if (!batadv_compare_eth(ethhdr->h_dest, primary_if->net_dev->dev_addr))
907 return true;
908
909 batadv_dbg(BATADV_DBG_BLA, bat_priv,
910 "%s(): REQUEST vid %d (sent by %pM)...\n",
911 __func__, batadv_print_vid(vid), ethhdr->h_source);
912
913 batadv_bla_answer_request(bat_priv, primary_if, vid);
914 return true;
915}
916
917/**
918 * batadv_handle_unclaim() - check for UNCLAIM frame
919 * @bat_priv: the bat priv with all the soft interface information
920 * @primary_if: the primary hard interface of this batman soft interface
921 * @backbone_addr: originator address of the backbone (Ethernet source)
922 * @claim_addr: Client to be unclaimed (ARP sender HW MAC)
923 * @vid: the VLAN ID of the frame
924 *
925 * Return: true if handled
926 */
927static bool batadv_handle_unclaim(struct batadv_priv *bat_priv,
928 struct batadv_hard_iface *primary_if,
929 u8 *backbone_addr, u8 *claim_addr,
930 unsigned short vid)
931{
932 struct batadv_bla_backbone_gw *backbone_gw;
933
934 /* unclaim in any case if it is our own */
935 if (primary_if && batadv_compare_eth(backbone_addr,
936 primary_if->net_dev->dev_addr))
937 batadv_bla_send_claim(bat_priv, claim_addr, vid,
938 BATADV_CLAIM_TYPE_UNCLAIM);
939
940 backbone_gw = batadv_backbone_hash_find(bat_priv, backbone_addr, vid);
941
942 if (!backbone_gw)
943 return true;
944
945 /* this must be an UNCLAIM frame */
946 batadv_dbg(BATADV_DBG_BLA, bat_priv,
947 "%s(): UNCLAIM %pM on vid %d (sent by %pM)...\n", __func__,
948 claim_addr, batadv_print_vid(vid), backbone_gw->orig);
949
950 batadv_bla_del_claim(bat_priv, claim_addr, vid);
951 batadv_backbone_gw_put(backbone_gw);
952 return true;
953}
954
955/**
956 * batadv_handle_claim() - check for CLAIM frame
957 * @bat_priv: the bat priv with all the soft interface information
958 * @primary_if: the primary hard interface of this batman soft interface
959 * @backbone_addr: originator address of the backbone (Ethernet Source)
960 * @claim_addr: client mac address to be claimed (ARP sender HW MAC)
961 * @vid: the VLAN ID of the frame
962 *
963 * Return: true if handled
964 */
965static bool batadv_handle_claim(struct batadv_priv *bat_priv,
966 struct batadv_hard_iface *primary_if,
967 u8 *backbone_addr, u8 *claim_addr,
968 unsigned short vid)
969{
970 struct batadv_bla_backbone_gw *backbone_gw;
971
972 /* register the gateway if not yet available, and add the claim. */
973
974 backbone_gw = batadv_bla_get_backbone_gw(bat_priv, backbone_addr, vid,
975 false);
976
977 if (unlikely(!backbone_gw))
978 return true;
979
980 /* this must be a CLAIM frame */
981 batadv_bla_add_claim(bat_priv, claim_addr, vid, backbone_gw);
982 if (batadv_compare_eth(backbone_addr, primary_if->net_dev->dev_addr))
983 batadv_bla_send_claim(bat_priv, claim_addr, vid,
984 BATADV_CLAIM_TYPE_CLAIM);
985
986 /* TODO: we could call something like tt_local_del() here. */
987
988 batadv_backbone_gw_put(backbone_gw);
989 return true;
990}
991
992/**
993 * batadv_check_claim_group() - check for claim group membership
994 * @bat_priv: the bat priv with all the soft interface information
995 * @primary_if: the primary interface of this batman interface
996 * @hw_src: the Hardware source in the ARP Header
997 * @hw_dst: the Hardware destination in the ARP Header
998 * @ethhdr: pointer to the Ethernet header of the claim frame
999 *
Olivier Deprez157378f2022-04-04 15:47:50 +02001000 * checks if it is a claim packet and if it's on the same group.
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001001 * This function also applies the group ID of the sender
1002 * if it is in the same mesh.
1003 *
1004 * Return:
1005 * 2 - if it is a claim packet and on the same group
1006 * 1 - if is a claim packet from another group
1007 * 0 - if it is not a claim packet
1008 */
1009static int batadv_check_claim_group(struct batadv_priv *bat_priv,
1010 struct batadv_hard_iface *primary_if,
1011 u8 *hw_src, u8 *hw_dst,
1012 struct ethhdr *ethhdr)
1013{
1014 u8 *backbone_addr;
1015 struct batadv_orig_node *orig_node;
1016 struct batadv_bla_claim_dst *bla_dst, *bla_dst_own;
1017
1018 bla_dst = (struct batadv_bla_claim_dst *)hw_dst;
1019 bla_dst_own = &bat_priv->bla.claim_dest;
1020
1021 /* if announcement packet, use the source,
1022 * otherwise assume it is in the hw_src
1023 */
1024 switch (bla_dst->type) {
1025 case BATADV_CLAIM_TYPE_CLAIM:
1026 backbone_addr = hw_src;
1027 break;
1028 case BATADV_CLAIM_TYPE_REQUEST:
1029 case BATADV_CLAIM_TYPE_ANNOUNCE:
1030 case BATADV_CLAIM_TYPE_UNCLAIM:
1031 backbone_addr = ethhdr->h_source;
1032 break;
1033 default:
1034 return 0;
1035 }
1036
1037 /* don't accept claim frames from ourselves */
1038 if (batadv_compare_eth(backbone_addr, primary_if->net_dev->dev_addr))
1039 return 0;
1040
1041 /* if its already the same group, it is fine. */
1042 if (bla_dst->group == bla_dst_own->group)
1043 return 2;
1044
1045 /* lets see if this originator is in our mesh */
1046 orig_node = batadv_orig_hash_find(bat_priv, backbone_addr);
1047
1048 /* dont accept claims from gateways which are not in
1049 * the same mesh or group.
1050 */
1051 if (!orig_node)
1052 return 1;
1053
1054 /* if our mesh friends mac is bigger, use it for ourselves. */
1055 if (ntohs(bla_dst->group) > ntohs(bla_dst_own->group)) {
1056 batadv_dbg(BATADV_DBG_BLA, bat_priv,
1057 "taking other backbones claim group: %#.4x\n",
1058 ntohs(bla_dst->group));
1059 bla_dst_own->group = bla_dst->group;
1060 }
1061
1062 batadv_orig_node_put(orig_node);
1063
1064 return 2;
1065}
1066
1067/**
1068 * batadv_bla_process_claim() - Check if this is a claim frame, and process it
1069 * @bat_priv: the bat priv with all the soft interface information
1070 * @primary_if: the primary hard interface of this batman soft interface
1071 * @skb: the frame to be checked
1072 *
1073 * Return: true if it was a claim frame, otherwise return false to
1074 * tell the callee that it can use the frame on its own.
1075 */
1076static bool batadv_bla_process_claim(struct batadv_priv *bat_priv,
1077 struct batadv_hard_iface *primary_if,
1078 struct sk_buff *skb)
1079{
1080 struct batadv_bla_claim_dst *bla_dst, *bla_dst_own;
1081 u8 *hw_src, *hw_dst;
1082 struct vlan_hdr *vhdr, vhdr_buf;
1083 struct ethhdr *ethhdr;
1084 struct arphdr *arphdr;
1085 unsigned short vid;
1086 int vlan_depth = 0;
1087 __be16 proto;
1088 int headlen;
1089 int ret;
1090
1091 vid = batadv_get_vid(skb, 0);
1092 ethhdr = eth_hdr(skb);
1093
1094 proto = ethhdr->h_proto;
1095 headlen = ETH_HLEN;
1096 if (vid & BATADV_VLAN_HAS_TAG) {
1097 /* Traverse the VLAN/Ethertypes.
1098 *
1099 * At this point it is known that the first protocol is a VLAN
1100 * header, so start checking at the encapsulated protocol.
1101 *
1102 * The depth of the VLAN headers is recorded to drop BLA claim
1103 * frames encapsulated into multiple VLAN headers (QinQ).
1104 */
1105 do {
1106 vhdr = skb_header_pointer(skb, headlen, VLAN_HLEN,
1107 &vhdr_buf);
1108 if (!vhdr)
1109 return false;
1110
1111 proto = vhdr->h_vlan_encapsulated_proto;
1112 headlen += VLAN_HLEN;
1113 vlan_depth++;
1114 } while (proto == htons(ETH_P_8021Q));
1115 }
1116
1117 if (proto != htons(ETH_P_ARP))
1118 return false; /* not a claim frame */
1119
1120 /* this must be a ARP frame. check if it is a claim. */
1121
1122 if (unlikely(!pskb_may_pull(skb, headlen + arp_hdr_len(skb->dev))))
1123 return false;
1124
1125 /* pskb_may_pull() may have modified the pointers, get ethhdr again */
1126 ethhdr = eth_hdr(skb);
1127 arphdr = (struct arphdr *)((u8 *)ethhdr + headlen);
1128
1129 /* Check whether the ARP frame carries a valid
1130 * IP information
1131 */
1132 if (arphdr->ar_hrd != htons(ARPHRD_ETHER))
1133 return false;
1134 if (arphdr->ar_pro != htons(ETH_P_IP))
1135 return false;
1136 if (arphdr->ar_hln != ETH_ALEN)
1137 return false;
1138 if (arphdr->ar_pln != 4)
1139 return false;
1140
1141 hw_src = (u8 *)arphdr + sizeof(struct arphdr);
1142 hw_dst = hw_src + ETH_ALEN + 4;
1143 bla_dst = (struct batadv_bla_claim_dst *)hw_dst;
1144 bla_dst_own = &bat_priv->bla.claim_dest;
1145
1146 /* check if it is a claim frame in general */
1147 if (memcmp(bla_dst->magic, bla_dst_own->magic,
1148 sizeof(bla_dst->magic)) != 0)
1149 return false;
1150
1151 /* check if there is a claim frame encapsulated deeper in (QinQ) and
1152 * drop that, as this is not supported by BLA but should also not be
1153 * sent via the mesh.
1154 */
1155 if (vlan_depth > 1)
1156 return true;
1157
1158 /* Let the loopdetect frames on the mesh in any case. */
1159 if (bla_dst->type == BATADV_CLAIM_TYPE_LOOPDETECT)
1160 return false;
1161
1162 /* check if it is a claim frame. */
1163 ret = batadv_check_claim_group(bat_priv, primary_if, hw_src, hw_dst,
1164 ethhdr);
1165 if (ret == 1)
1166 batadv_dbg(BATADV_DBG_BLA, bat_priv,
1167 "%s(): received a claim frame from another group. From: %pM on vid %d ...(hw_src %pM, hw_dst %pM)\n",
1168 __func__, ethhdr->h_source, batadv_print_vid(vid),
1169 hw_src, hw_dst);
1170
1171 if (ret < 2)
1172 return !!ret;
1173
1174 /* become a backbone gw ourselves on this vlan if not happened yet */
1175 batadv_bla_update_own_backbone_gw(bat_priv, primary_if, vid);
1176
1177 /* check for the different types of claim frames ... */
1178 switch (bla_dst->type) {
1179 case BATADV_CLAIM_TYPE_CLAIM:
1180 if (batadv_handle_claim(bat_priv, primary_if, hw_src,
1181 ethhdr->h_source, vid))
1182 return true;
1183 break;
1184 case BATADV_CLAIM_TYPE_UNCLAIM:
1185 if (batadv_handle_unclaim(bat_priv, primary_if,
1186 ethhdr->h_source, hw_src, vid))
1187 return true;
1188 break;
1189
1190 case BATADV_CLAIM_TYPE_ANNOUNCE:
1191 if (batadv_handle_announce(bat_priv, hw_src, ethhdr->h_source,
1192 vid))
1193 return true;
1194 break;
1195 case BATADV_CLAIM_TYPE_REQUEST:
1196 if (batadv_handle_request(bat_priv, primary_if, hw_src, ethhdr,
1197 vid))
1198 return true;
1199 break;
1200 }
1201
1202 batadv_dbg(BATADV_DBG_BLA, bat_priv,
1203 "%s(): ERROR - this looks like a claim frame, but is useless. eth src %pM on vid %d ...(hw_src %pM, hw_dst %pM)\n",
1204 __func__, ethhdr->h_source, batadv_print_vid(vid), hw_src,
1205 hw_dst);
1206 return true;
1207}
1208
1209/**
1210 * batadv_bla_purge_backbone_gw() - Remove backbone gateways after a timeout or
1211 * immediately
1212 * @bat_priv: the bat priv with all the soft interface information
1213 * @now: whether the whole hash shall be wiped now
1214 *
1215 * Check when we last heard from other nodes, and remove them in case of
1216 * a time out, or clean all backbone gws if now is set.
1217 */
1218static void batadv_bla_purge_backbone_gw(struct batadv_priv *bat_priv, int now)
1219{
1220 struct batadv_bla_backbone_gw *backbone_gw;
1221 struct hlist_node *node_tmp;
1222 struct hlist_head *head;
1223 struct batadv_hashtable *hash;
1224 spinlock_t *list_lock; /* protects write access to the hash lists */
1225 int i;
1226
1227 hash = bat_priv->bla.backbone_hash;
1228 if (!hash)
1229 return;
1230
1231 for (i = 0; i < hash->size; i++) {
1232 head = &hash->table[i];
1233 list_lock = &hash->list_locks[i];
1234
1235 spin_lock_bh(list_lock);
1236 hlist_for_each_entry_safe(backbone_gw, node_tmp,
1237 head, hash_entry) {
1238 if (now)
1239 goto purge_now;
1240 if (!batadv_has_timed_out(backbone_gw->lasttime,
1241 BATADV_BLA_BACKBONE_TIMEOUT))
1242 continue;
1243
1244 batadv_dbg(BATADV_DBG_BLA, backbone_gw->bat_priv,
1245 "%s(): backbone gw %pM timed out\n",
1246 __func__, backbone_gw->orig);
1247
1248purge_now:
1249 /* don't wait for the pending request anymore */
1250 if (atomic_read(&backbone_gw->request_sent))
1251 atomic_dec(&bat_priv->bla.num_requests);
1252
1253 batadv_bla_del_backbone_claims(backbone_gw);
1254
1255 hlist_del_rcu(&backbone_gw->hash_entry);
1256 batadv_backbone_gw_put(backbone_gw);
1257 }
1258 spin_unlock_bh(list_lock);
1259 }
1260}
1261
1262/**
1263 * batadv_bla_purge_claims() - Remove claims after a timeout or immediately
1264 * @bat_priv: the bat priv with all the soft interface information
1265 * @primary_if: the selected primary interface, may be NULL if now is set
1266 * @now: whether the whole hash shall be wiped now
1267 *
1268 * Check when we heard last time from our own claims, and remove them in case of
1269 * a time out, or clean all claims if now is set
1270 */
1271static void batadv_bla_purge_claims(struct batadv_priv *bat_priv,
1272 struct batadv_hard_iface *primary_if,
1273 int now)
1274{
1275 struct batadv_bla_backbone_gw *backbone_gw;
1276 struct batadv_bla_claim *claim;
1277 struct hlist_head *head;
1278 struct batadv_hashtable *hash;
1279 int i;
1280
1281 hash = bat_priv->bla.claim_hash;
1282 if (!hash)
1283 return;
1284
1285 for (i = 0; i < hash->size; i++) {
1286 head = &hash->table[i];
1287
1288 rcu_read_lock();
1289 hlist_for_each_entry_rcu(claim, head, hash_entry) {
1290 backbone_gw = batadv_bla_claim_get_backbone_gw(claim);
1291 if (now)
1292 goto purge_now;
1293
1294 if (!batadv_compare_eth(backbone_gw->orig,
1295 primary_if->net_dev->dev_addr))
1296 goto skip;
1297
1298 if (!batadv_has_timed_out(claim->lasttime,
1299 BATADV_BLA_CLAIM_TIMEOUT))
1300 goto skip;
1301
1302 batadv_dbg(BATADV_DBG_BLA, bat_priv,
1303 "%s(): timed out.\n", __func__);
1304
1305purge_now:
1306 batadv_dbg(BATADV_DBG_BLA, bat_priv,
1307 "%s(): %pM, vid %d\n", __func__,
1308 claim->addr, claim->vid);
1309
1310 batadv_handle_unclaim(bat_priv, primary_if,
1311 backbone_gw->orig,
1312 claim->addr, claim->vid);
1313skip:
1314 batadv_backbone_gw_put(backbone_gw);
1315 }
1316 rcu_read_unlock();
1317 }
1318}
1319
1320/**
1321 * batadv_bla_update_orig_address() - Update the backbone gateways when the own
1322 * originator address changes
1323 * @bat_priv: the bat priv with all the soft interface information
1324 * @primary_if: the new selected primary_if
1325 * @oldif: the old primary interface, may be NULL
1326 */
1327void batadv_bla_update_orig_address(struct batadv_priv *bat_priv,
1328 struct batadv_hard_iface *primary_if,
1329 struct batadv_hard_iface *oldif)
1330{
1331 struct batadv_bla_backbone_gw *backbone_gw;
1332 struct hlist_head *head;
1333 struct batadv_hashtable *hash;
1334 __be16 group;
1335 int i;
1336
1337 /* reset bridge loop avoidance group id */
1338 group = htons(crc16(0, primary_if->net_dev->dev_addr, ETH_ALEN));
1339 bat_priv->bla.claim_dest.group = group;
1340
1341 /* purge everything when bridge loop avoidance is turned off */
1342 if (!atomic_read(&bat_priv->bridge_loop_avoidance))
1343 oldif = NULL;
1344
1345 if (!oldif) {
1346 batadv_bla_purge_claims(bat_priv, NULL, 1);
1347 batadv_bla_purge_backbone_gw(bat_priv, 1);
1348 return;
1349 }
1350
1351 hash = bat_priv->bla.backbone_hash;
1352 if (!hash)
1353 return;
1354
1355 for (i = 0; i < hash->size; i++) {
1356 head = &hash->table[i];
1357
1358 rcu_read_lock();
1359 hlist_for_each_entry_rcu(backbone_gw, head, hash_entry) {
1360 /* own orig still holds the old value. */
1361 if (!batadv_compare_eth(backbone_gw->orig,
1362 oldif->net_dev->dev_addr))
1363 continue;
1364
1365 ether_addr_copy(backbone_gw->orig,
1366 primary_if->net_dev->dev_addr);
1367 /* send an announce frame so others will ask for our
1368 * claims and update their tables.
1369 */
1370 batadv_bla_send_announce(bat_priv, backbone_gw);
1371 }
1372 rcu_read_unlock();
1373 }
1374}
1375
1376/**
1377 * batadv_bla_send_loopdetect() - send a loopdetect frame
1378 * @bat_priv: the bat priv with all the soft interface information
1379 * @backbone_gw: the backbone gateway for which a loop should be detected
1380 *
1381 * To detect loops that the bridge loop avoidance can't handle, send a loop
1382 * detection packet on the backbone. Unlike other BLA frames, this frame will
1383 * be allowed on the mesh by other nodes. If it is received on the mesh, this
1384 * indicates that there is a loop.
1385 */
1386static void
1387batadv_bla_send_loopdetect(struct batadv_priv *bat_priv,
1388 struct batadv_bla_backbone_gw *backbone_gw)
1389{
1390 batadv_dbg(BATADV_DBG_BLA, bat_priv, "Send loopdetect frame for vid %d\n",
1391 backbone_gw->vid);
1392 batadv_bla_send_claim(bat_priv, bat_priv->bla.loopdetect_addr,
1393 backbone_gw->vid, BATADV_CLAIM_TYPE_LOOPDETECT);
1394}
1395
1396/**
1397 * batadv_bla_status_update() - purge bla interfaces if necessary
1398 * @net_dev: the soft interface net device
1399 */
1400void batadv_bla_status_update(struct net_device *net_dev)
1401{
1402 struct batadv_priv *bat_priv = netdev_priv(net_dev);
1403 struct batadv_hard_iface *primary_if;
1404
1405 primary_if = batadv_primary_if_get_selected(bat_priv);
1406 if (!primary_if)
1407 return;
1408
1409 /* this function already purges everything when bla is disabled,
1410 * so just call that one.
1411 */
1412 batadv_bla_update_orig_address(bat_priv, primary_if, primary_if);
1413 batadv_hardif_put(primary_if);
1414}
1415
1416/**
1417 * batadv_bla_periodic_work() - performs periodic bla work
1418 * @work: kernel work struct
1419 *
1420 * periodic work to do:
1421 * * purge structures when they are too old
1422 * * send announcements
1423 */
1424static void batadv_bla_periodic_work(struct work_struct *work)
1425{
1426 struct delayed_work *delayed_work;
1427 struct batadv_priv *bat_priv;
1428 struct batadv_priv_bla *priv_bla;
1429 struct hlist_head *head;
1430 struct batadv_bla_backbone_gw *backbone_gw;
1431 struct batadv_hashtable *hash;
1432 struct batadv_hard_iface *primary_if;
1433 bool send_loopdetect = false;
1434 int i;
1435
1436 delayed_work = to_delayed_work(work);
1437 priv_bla = container_of(delayed_work, struct batadv_priv_bla, work);
1438 bat_priv = container_of(priv_bla, struct batadv_priv, bla);
1439 primary_if = batadv_primary_if_get_selected(bat_priv);
1440 if (!primary_if)
1441 goto out;
1442
1443 batadv_bla_purge_claims(bat_priv, primary_if, 0);
1444 batadv_bla_purge_backbone_gw(bat_priv, 0);
1445
1446 if (!atomic_read(&bat_priv->bridge_loop_avoidance))
1447 goto out;
1448
1449 if (atomic_dec_and_test(&bat_priv->bla.loopdetect_next)) {
1450 /* set a new random mac address for the next bridge loop
1451 * detection frames. Set the locally administered bit to avoid
1452 * collisions with users mac addresses.
1453 */
1454 eth_random_addr(bat_priv->bla.loopdetect_addr);
1455 bat_priv->bla.loopdetect_addr[0] = 0xba;
1456 bat_priv->bla.loopdetect_addr[1] = 0xbe;
1457 bat_priv->bla.loopdetect_lasttime = jiffies;
1458 atomic_set(&bat_priv->bla.loopdetect_next,
1459 BATADV_BLA_LOOPDETECT_PERIODS);
1460
1461 /* mark for sending loop detect on all VLANs */
1462 send_loopdetect = true;
1463 }
1464
1465 hash = bat_priv->bla.backbone_hash;
1466 if (!hash)
1467 goto out;
1468
1469 for (i = 0; i < hash->size; i++) {
1470 head = &hash->table[i];
1471
1472 rcu_read_lock();
1473 hlist_for_each_entry_rcu(backbone_gw, head, hash_entry) {
1474 if (!batadv_compare_eth(backbone_gw->orig,
1475 primary_if->net_dev->dev_addr))
1476 continue;
1477
1478 backbone_gw->lasttime = jiffies;
1479
1480 batadv_bla_send_announce(bat_priv, backbone_gw);
1481 if (send_loopdetect)
1482 batadv_bla_send_loopdetect(bat_priv,
1483 backbone_gw);
1484
1485 /* request_sent is only set after creation to avoid
1486 * problems when we are not yet known as backbone gw
1487 * in the backbone.
1488 *
1489 * We can reset this now after we waited some periods
1490 * to give bridge forward delays and bla group forming
1491 * some grace time.
1492 */
1493
1494 if (atomic_read(&backbone_gw->request_sent) == 0)
1495 continue;
1496
1497 if (!atomic_dec_and_test(&backbone_gw->wait_periods))
1498 continue;
1499
1500 atomic_dec(&backbone_gw->bat_priv->bla.num_requests);
1501 atomic_set(&backbone_gw->request_sent, 0);
1502 }
1503 rcu_read_unlock();
1504 }
1505out:
1506 if (primary_if)
1507 batadv_hardif_put(primary_if);
1508
1509 queue_delayed_work(batadv_event_workqueue, &bat_priv->bla.work,
1510 msecs_to_jiffies(BATADV_BLA_PERIOD_LENGTH));
1511}
1512
1513/* The hash for claim and backbone hash receive the same key because they
1514 * are getting initialized by hash_new with the same key. Reinitializing
1515 * them with to different keys to allow nested locking without generating
1516 * lockdep warnings
1517 */
1518static struct lock_class_key batadv_claim_hash_lock_class_key;
1519static struct lock_class_key batadv_backbone_hash_lock_class_key;
1520
1521/**
1522 * batadv_bla_init() - initialize all bla structures
1523 * @bat_priv: the bat priv with all the soft interface information
1524 *
1525 * Return: 0 on success, < 0 on error.
1526 */
1527int batadv_bla_init(struct batadv_priv *bat_priv)
1528{
1529 int i;
1530 u8 claim_dest[ETH_ALEN] = {0xff, 0x43, 0x05, 0x00, 0x00, 0x00};
1531 struct batadv_hard_iface *primary_if;
1532 u16 crc;
1533 unsigned long entrytime;
1534
1535 spin_lock_init(&bat_priv->bla.bcast_duplist_lock);
1536
1537 batadv_dbg(BATADV_DBG_BLA, bat_priv, "bla hash registering\n");
1538
1539 /* setting claim destination address */
1540 memcpy(&bat_priv->bla.claim_dest.magic, claim_dest, 3);
1541 bat_priv->bla.claim_dest.type = 0;
1542 primary_if = batadv_primary_if_get_selected(bat_priv);
1543 if (primary_if) {
1544 crc = crc16(0, primary_if->net_dev->dev_addr, ETH_ALEN);
1545 bat_priv->bla.claim_dest.group = htons(crc);
1546 batadv_hardif_put(primary_if);
1547 } else {
1548 bat_priv->bla.claim_dest.group = 0; /* will be set later */
1549 }
1550
1551 /* initialize the duplicate list */
1552 entrytime = jiffies - msecs_to_jiffies(BATADV_DUPLIST_TIMEOUT);
1553 for (i = 0; i < BATADV_DUPLIST_SIZE; i++)
1554 bat_priv->bla.bcast_duplist[i].entrytime = entrytime;
1555 bat_priv->bla.bcast_duplist_curr = 0;
1556
1557 atomic_set(&bat_priv->bla.loopdetect_next,
1558 BATADV_BLA_LOOPDETECT_PERIODS);
1559
1560 if (bat_priv->bla.claim_hash)
1561 return 0;
1562
1563 bat_priv->bla.claim_hash = batadv_hash_new(128);
Olivier Deprez157378f2022-04-04 15:47:50 +02001564 if (!bat_priv->bla.claim_hash)
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001565 return -ENOMEM;
1566
Olivier Deprez157378f2022-04-04 15:47:50 +02001567 bat_priv->bla.backbone_hash = batadv_hash_new(32);
1568 if (!bat_priv->bla.backbone_hash) {
1569 batadv_hash_destroy(bat_priv->bla.claim_hash);
1570 return -ENOMEM;
1571 }
1572
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001573 batadv_hash_set_lock_class(bat_priv->bla.claim_hash,
1574 &batadv_claim_hash_lock_class_key);
1575 batadv_hash_set_lock_class(bat_priv->bla.backbone_hash,
1576 &batadv_backbone_hash_lock_class_key);
1577
1578 batadv_dbg(BATADV_DBG_BLA, bat_priv, "bla hashes initialized\n");
1579
1580 INIT_DELAYED_WORK(&bat_priv->bla.work, batadv_bla_periodic_work);
1581
1582 queue_delayed_work(batadv_event_workqueue, &bat_priv->bla.work,
1583 msecs_to_jiffies(BATADV_BLA_PERIOD_LENGTH));
1584 return 0;
1585}
1586
1587/**
Olivier Deprez0e641232021-09-23 10:07:05 +02001588 * batadv_bla_check_duplist() - Check if a frame is in the broadcast dup.
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001589 * @bat_priv: the bat priv with all the soft interface information
Olivier Deprez0e641232021-09-23 10:07:05 +02001590 * @skb: contains the multicast packet to be checked
1591 * @payload_ptr: pointer to position inside the head buffer of the skb
1592 * marking the start of the data to be CRC'ed
1593 * @orig: originator mac address, NULL if unknown
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001594 *
Olivier Deprez0e641232021-09-23 10:07:05 +02001595 * Check if it is on our broadcast list. Another gateway might have sent the
1596 * same packet because it is connected to the same backbone, so we have to
1597 * remove this duplicate.
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001598 *
1599 * This is performed by checking the CRC, which will tell us
1600 * with a good chance that it is the same packet. If it is furthermore
1601 * sent by another host, drop it. We allow equal packets from
1602 * the same host however as this might be intended.
1603 *
1604 * Return: true if a packet is in the duplicate list, false otherwise.
1605 */
Olivier Deprez0e641232021-09-23 10:07:05 +02001606static bool batadv_bla_check_duplist(struct batadv_priv *bat_priv,
1607 struct sk_buff *skb, u8 *payload_ptr,
1608 const u8 *orig)
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001609{
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001610 struct batadv_bcast_duplist_entry *entry;
1611 bool ret = false;
Olivier Deprez0e641232021-09-23 10:07:05 +02001612 int i, curr;
1613 __be32 crc;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001614
1615 /* calculate the crc ... */
Olivier Deprez0e641232021-09-23 10:07:05 +02001616 crc = batadv_skb_crc32(skb, payload_ptr);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001617
1618 spin_lock_bh(&bat_priv->bla.bcast_duplist_lock);
1619
1620 for (i = 0; i < BATADV_DUPLIST_SIZE; i++) {
1621 curr = (bat_priv->bla.bcast_duplist_curr + i);
1622 curr %= BATADV_DUPLIST_SIZE;
1623 entry = &bat_priv->bla.bcast_duplist[curr];
1624
1625 /* we can stop searching if the entry is too old ;
1626 * later entries will be even older
1627 */
1628 if (batadv_has_timed_out(entry->entrytime,
1629 BATADV_DUPLIST_TIMEOUT))
1630 break;
1631
1632 if (entry->crc != crc)
1633 continue;
1634
Olivier Deprez0e641232021-09-23 10:07:05 +02001635 /* are the originators both known and not anonymous? */
1636 if (orig && !is_zero_ether_addr(orig) &&
1637 !is_zero_ether_addr(entry->orig)) {
1638 /* If known, check if the new frame came from
1639 * the same originator:
1640 * We are safe to take identical frames from the
1641 * same orig, if known, as multiplications in
1642 * the mesh are detected via the (orig, seqno) pair.
1643 * So we can be a bit more liberal here and allow
1644 * identical frames from the same orig which the source
1645 * host might have sent multiple times on purpose.
1646 */
1647 if (batadv_compare_eth(entry->orig, orig))
1648 continue;
1649 }
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001650
1651 /* this entry seems to match: same crc, not too old,
1652 * and from another gw. therefore return true to forbid it.
1653 */
1654 ret = true;
1655 goto out;
1656 }
1657 /* not found, add a new entry (overwrite the oldest entry)
1658 * and allow it, its the first occurrence.
1659 */
1660 curr = (bat_priv->bla.bcast_duplist_curr + BATADV_DUPLIST_SIZE - 1);
1661 curr %= BATADV_DUPLIST_SIZE;
1662 entry = &bat_priv->bla.bcast_duplist[curr];
1663 entry->crc = crc;
1664 entry->entrytime = jiffies;
Olivier Deprez0e641232021-09-23 10:07:05 +02001665
1666 /* known originator */
1667 if (orig)
1668 ether_addr_copy(entry->orig, orig);
1669 /* anonymous originator */
1670 else
1671 eth_zero_addr(entry->orig);
1672
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001673 bat_priv->bla.bcast_duplist_curr = curr;
1674
1675out:
1676 spin_unlock_bh(&bat_priv->bla.bcast_duplist_lock);
1677
1678 return ret;
1679}
1680
1681/**
Olivier Deprez0e641232021-09-23 10:07:05 +02001682 * batadv_bla_check_ucast_duplist() - Check if a frame is in the broadcast dup.
1683 * @bat_priv: the bat priv with all the soft interface information
1684 * @skb: contains the multicast packet to be checked, decapsulated from a
1685 * unicast_packet
1686 *
1687 * Check if it is on our broadcast list. Another gateway might have sent the
1688 * same packet because it is connected to the same backbone, so we have to
1689 * remove this duplicate.
1690 *
1691 * Return: true if a packet is in the duplicate list, false otherwise.
1692 */
1693static bool batadv_bla_check_ucast_duplist(struct batadv_priv *bat_priv,
1694 struct sk_buff *skb)
1695{
1696 return batadv_bla_check_duplist(bat_priv, skb, (u8 *)skb->data, NULL);
1697}
1698
1699/**
1700 * batadv_bla_check_bcast_duplist() - Check if a frame is in the broadcast dup.
1701 * @bat_priv: the bat priv with all the soft interface information
1702 * @skb: contains the bcast_packet to be checked
1703 *
1704 * Check if it is on our broadcast list. Another gateway might have sent the
1705 * same packet because it is connected to the same backbone, so we have to
1706 * remove this duplicate.
1707 *
1708 * Return: true if a packet is in the duplicate list, false otherwise.
1709 */
1710bool batadv_bla_check_bcast_duplist(struct batadv_priv *bat_priv,
1711 struct sk_buff *skb)
1712{
1713 struct batadv_bcast_packet *bcast_packet;
1714 u8 *payload_ptr;
1715
1716 bcast_packet = (struct batadv_bcast_packet *)skb->data;
1717 payload_ptr = (u8 *)(bcast_packet + 1);
1718
1719 return batadv_bla_check_duplist(bat_priv, skb, payload_ptr,
1720 bcast_packet->orig);
1721}
1722
1723/**
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001724 * batadv_bla_is_backbone_gw_orig() - Check if the originator is a gateway for
1725 * the VLAN identified by vid.
1726 * @bat_priv: the bat priv with all the soft interface information
1727 * @orig: originator mac address
1728 * @vid: VLAN identifier
1729 *
1730 * Return: true if orig is a backbone for this vid, false otherwise.
1731 */
1732bool batadv_bla_is_backbone_gw_orig(struct batadv_priv *bat_priv, u8 *orig,
1733 unsigned short vid)
1734{
1735 struct batadv_hashtable *hash = bat_priv->bla.backbone_hash;
1736 struct hlist_head *head;
1737 struct batadv_bla_backbone_gw *backbone_gw;
1738 int i;
1739
1740 if (!atomic_read(&bat_priv->bridge_loop_avoidance))
1741 return false;
1742
1743 if (!hash)
1744 return false;
1745
1746 for (i = 0; i < hash->size; i++) {
1747 head = &hash->table[i];
1748
1749 rcu_read_lock();
1750 hlist_for_each_entry_rcu(backbone_gw, head, hash_entry) {
1751 if (batadv_compare_eth(backbone_gw->orig, orig) &&
1752 backbone_gw->vid == vid) {
1753 rcu_read_unlock();
1754 return true;
1755 }
1756 }
1757 rcu_read_unlock();
1758 }
1759
1760 return false;
1761}
1762
1763/**
1764 * batadv_bla_is_backbone_gw() - check if originator is a backbone gw for a VLAN
1765 * @skb: the frame to be checked
1766 * @orig_node: the orig_node of the frame
1767 * @hdr_size: maximum length of the frame
1768 *
1769 * Return: true if the orig_node is also a gateway on the soft interface,
1770 * otherwise it returns false.
1771 */
1772bool batadv_bla_is_backbone_gw(struct sk_buff *skb,
1773 struct batadv_orig_node *orig_node, int hdr_size)
1774{
1775 struct batadv_bla_backbone_gw *backbone_gw;
1776 unsigned short vid;
1777
1778 if (!atomic_read(&orig_node->bat_priv->bridge_loop_avoidance))
1779 return false;
1780
1781 /* first, find out the vid. */
1782 if (!pskb_may_pull(skb, hdr_size + ETH_HLEN))
1783 return false;
1784
1785 vid = batadv_get_vid(skb, hdr_size);
1786
1787 /* see if this originator is a backbone gw for this VLAN */
1788 backbone_gw = batadv_backbone_hash_find(orig_node->bat_priv,
1789 orig_node->orig, vid);
1790 if (!backbone_gw)
1791 return false;
1792
1793 batadv_backbone_gw_put(backbone_gw);
1794 return true;
1795}
1796
1797/**
1798 * batadv_bla_free() - free all bla structures
1799 * @bat_priv: the bat priv with all the soft interface information
1800 *
1801 * for softinterface free or module unload
1802 */
1803void batadv_bla_free(struct batadv_priv *bat_priv)
1804{
1805 struct batadv_hard_iface *primary_if;
1806
1807 cancel_delayed_work_sync(&bat_priv->bla.work);
1808 primary_if = batadv_primary_if_get_selected(bat_priv);
1809
1810 if (bat_priv->bla.claim_hash) {
1811 batadv_bla_purge_claims(bat_priv, primary_if, 1);
1812 batadv_hash_destroy(bat_priv->bla.claim_hash);
1813 bat_priv->bla.claim_hash = NULL;
1814 }
1815 if (bat_priv->bla.backbone_hash) {
1816 batadv_bla_purge_backbone_gw(bat_priv, 1);
1817 batadv_hash_destroy(bat_priv->bla.backbone_hash);
1818 bat_priv->bla.backbone_hash = NULL;
1819 }
1820 if (primary_if)
1821 batadv_hardif_put(primary_if);
1822}
1823
1824/**
1825 * batadv_bla_loopdetect_check() - check and handle a detected loop
1826 * @bat_priv: the bat priv with all the soft interface information
1827 * @skb: the packet to check
1828 * @primary_if: interface where the request came on
1829 * @vid: the VLAN ID of the frame
1830 *
1831 * Checks if this packet is a loop detect frame which has been sent by us,
Olivier Deprez157378f2022-04-04 15:47:50 +02001832 * throws an uevent and logs the event if that is the case.
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001833 *
1834 * Return: true if it is a loop detect frame which is to be dropped, false
1835 * otherwise.
1836 */
1837static bool
1838batadv_bla_loopdetect_check(struct batadv_priv *bat_priv, struct sk_buff *skb,
1839 struct batadv_hard_iface *primary_if,
1840 unsigned short vid)
1841{
1842 struct batadv_bla_backbone_gw *backbone_gw;
1843 struct ethhdr *ethhdr;
1844 bool ret;
1845
1846 ethhdr = eth_hdr(skb);
1847
1848 /* Only check for the MAC address and skip more checks here for
1849 * performance reasons - this function is on the hotpath, after all.
1850 */
1851 if (!batadv_compare_eth(ethhdr->h_source,
1852 bat_priv->bla.loopdetect_addr))
1853 return false;
1854
1855 /* If the packet came too late, don't forward it on the mesh
1856 * but don't consider that as loop. It might be a coincidence.
1857 */
1858 if (batadv_has_timed_out(bat_priv->bla.loopdetect_lasttime,
1859 BATADV_BLA_LOOPDETECT_TIMEOUT))
1860 return true;
1861
1862 backbone_gw = batadv_bla_get_backbone_gw(bat_priv,
1863 primary_if->net_dev->dev_addr,
1864 vid, true);
1865 if (unlikely(!backbone_gw))
1866 return true;
1867
1868 ret = queue_work(batadv_event_workqueue, &backbone_gw->report_work);
1869
Olivier Deprez157378f2022-04-04 15:47:50 +02001870 /* backbone_gw is unreferenced in the report work function
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001871 * if queue_work() call was successful
1872 */
1873 if (!ret)
1874 batadv_backbone_gw_put(backbone_gw);
1875
1876 return true;
1877}
1878
1879/**
1880 * batadv_bla_rx() - check packets coming from the mesh.
1881 * @bat_priv: the bat priv with all the soft interface information
1882 * @skb: the frame to be checked
1883 * @vid: the VLAN ID of the frame
Olivier Deprez0e641232021-09-23 10:07:05 +02001884 * @packet_type: the batman packet type this frame came in
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001885 *
1886 * batadv_bla_rx avoidance checks if:
1887 * * we have to race for a claim
1888 * * if the frame is allowed on the LAN
1889 *
Olivier Deprez157378f2022-04-04 15:47:50 +02001890 * In these cases, the skb is further handled by this function
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001891 *
1892 * Return: true if handled, otherwise it returns false and the caller shall
1893 * further process the skb.
1894 */
1895bool batadv_bla_rx(struct batadv_priv *bat_priv, struct sk_buff *skb,
Olivier Deprez0e641232021-09-23 10:07:05 +02001896 unsigned short vid, int packet_type)
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001897{
1898 struct batadv_bla_backbone_gw *backbone_gw;
1899 struct ethhdr *ethhdr;
1900 struct batadv_bla_claim search_claim, *claim = NULL;
1901 struct batadv_hard_iface *primary_if;
1902 bool own_claim;
1903 bool ret;
1904
1905 ethhdr = eth_hdr(skb);
1906
1907 primary_if = batadv_primary_if_get_selected(bat_priv);
1908 if (!primary_if)
1909 goto handled;
1910
1911 if (!atomic_read(&bat_priv->bridge_loop_avoidance))
1912 goto allow;
1913
1914 if (batadv_bla_loopdetect_check(bat_priv, skb, primary_if, vid))
1915 goto handled;
1916
1917 if (unlikely(atomic_read(&bat_priv->bla.num_requests)))
Olivier Deprez0e641232021-09-23 10:07:05 +02001918 /* don't allow multicast packets while requests are in flight */
1919 if (is_multicast_ether_addr(ethhdr->h_dest))
1920 /* Both broadcast flooding or multicast-via-unicasts
1921 * delivery might send to multiple backbone gateways
1922 * sharing the same LAN and therefore need to coordinate
1923 * which backbone gateway forwards into the LAN,
1924 * by claiming the payload source address.
1925 *
1926 * Broadcast flooding and multicast-via-unicasts
1927 * delivery use the following two batman packet types.
1928 * Note: explicitly exclude BATADV_UNICAST_4ADDR,
1929 * as the DHCP gateway feature will send explicitly
1930 * to only one BLA gateway, so the claiming process
1931 * should be avoided there.
1932 */
1933 if (packet_type == BATADV_BCAST ||
1934 packet_type == BATADV_UNICAST)
1935 goto handled;
1936
1937 /* potential duplicates from foreign BLA backbone gateways via
1938 * multicast-in-unicast packets
1939 */
1940 if (is_multicast_ether_addr(ethhdr->h_dest) &&
1941 packet_type == BATADV_UNICAST &&
1942 batadv_bla_check_ucast_duplist(bat_priv, skb))
1943 goto handled;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001944
1945 ether_addr_copy(search_claim.addr, ethhdr->h_source);
1946 search_claim.vid = vid;
1947 claim = batadv_claim_hash_find(bat_priv, &search_claim);
1948
1949 if (!claim) {
1950 /* possible optimization: race for a claim */
1951 /* No claim exists yet, claim it for us!
1952 */
1953
1954 batadv_dbg(BATADV_DBG_BLA, bat_priv,
1955 "%s(): Unclaimed MAC %pM found. Claim it. Local: %s\n",
1956 __func__, ethhdr->h_source,
1957 batadv_is_my_client(bat_priv,
1958 ethhdr->h_source, vid) ?
1959 "yes" : "no");
1960 batadv_handle_claim(bat_priv, primary_if,
1961 primary_if->net_dev->dev_addr,
1962 ethhdr->h_source, vid);
1963 goto allow;
1964 }
1965
1966 /* if it is our own claim ... */
1967 backbone_gw = batadv_bla_claim_get_backbone_gw(claim);
1968 own_claim = batadv_compare_eth(backbone_gw->orig,
1969 primary_if->net_dev->dev_addr);
1970 batadv_backbone_gw_put(backbone_gw);
1971
1972 if (own_claim) {
1973 /* ... allow it in any case */
1974 claim->lasttime = jiffies;
1975 goto allow;
1976 }
1977
Olivier Deprez0e641232021-09-23 10:07:05 +02001978 /* if it is a multicast ... */
1979 if (is_multicast_ether_addr(ethhdr->h_dest) &&
1980 (packet_type == BATADV_BCAST || packet_type == BATADV_UNICAST)) {
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001981 /* ... drop it. the responsible gateway is in charge.
1982 *
Olivier Deprez0e641232021-09-23 10:07:05 +02001983 * We need to check packet type because with the gateway
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001984 * feature, broadcasts (like DHCP requests) may be sent
Olivier Deprez0e641232021-09-23 10:07:05 +02001985 * using a unicast 4 address packet type. See comment above.
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001986 */
1987 goto handled;
1988 } else {
1989 /* seems the client considers us as its best gateway.
1990 * send a claim and update the claim table
1991 * immediately.
1992 */
1993 batadv_handle_claim(bat_priv, primary_if,
1994 primary_if->net_dev->dev_addr,
1995 ethhdr->h_source, vid);
1996 goto allow;
1997 }
1998allow:
1999 batadv_bla_update_own_backbone_gw(bat_priv, primary_if, vid);
2000 ret = false;
2001 goto out;
2002
2003handled:
2004 kfree_skb(skb);
2005 ret = true;
2006
2007out:
2008 if (primary_if)
2009 batadv_hardif_put(primary_if);
2010 if (claim)
2011 batadv_claim_put(claim);
2012 return ret;
2013}
2014
2015/**
2016 * batadv_bla_tx() - check packets going into the mesh
2017 * @bat_priv: the bat priv with all the soft interface information
2018 * @skb: the frame to be checked
2019 * @vid: the VLAN ID of the frame
2020 *
2021 * batadv_bla_tx checks if:
2022 * * a claim was received which has to be processed
2023 * * the frame is allowed on the mesh
2024 *
2025 * in these cases, the skb is further handled by this function.
2026 *
2027 * This call might reallocate skb data.
2028 *
2029 * Return: true if handled, otherwise it returns false and the caller shall
2030 * further process the skb.
2031 */
2032bool batadv_bla_tx(struct batadv_priv *bat_priv, struct sk_buff *skb,
2033 unsigned short vid)
2034{
2035 struct ethhdr *ethhdr;
2036 struct batadv_bla_claim search_claim, *claim = NULL;
2037 struct batadv_bla_backbone_gw *backbone_gw;
2038 struct batadv_hard_iface *primary_if;
2039 bool client_roamed;
2040 bool ret = false;
2041
2042 primary_if = batadv_primary_if_get_selected(bat_priv);
2043 if (!primary_if)
2044 goto out;
2045
2046 if (!atomic_read(&bat_priv->bridge_loop_avoidance))
2047 goto allow;
2048
2049 if (batadv_bla_process_claim(bat_priv, primary_if, skb))
2050 goto handled;
2051
2052 ethhdr = eth_hdr(skb);
2053
2054 if (unlikely(atomic_read(&bat_priv->bla.num_requests)))
2055 /* don't allow broadcasts while requests are in flight */
2056 if (is_multicast_ether_addr(ethhdr->h_dest))
2057 goto handled;
2058
2059 ether_addr_copy(search_claim.addr, ethhdr->h_source);
2060 search_claim.vid = vid;
2061
2062 claim = batadv_claim_hash_find(bat_priv, &search_claim);
2063
2064 /* if no claim exists, allow it. */
2065 if (!claim)
2066 goto allow;
2067
2068 /* check if we are responsible. */
2069 backbone_gw = batadv_bla_claim_get_backbone_gw(claim);
2070 client_roamed = batadv_compare_eth(backbone_gw->orig,
2071 primary_if->net_dev->dev_addr);
2072 batadv_backbone_gw_put(backbone_gw);
2073
2074 if (client_roamed) {
2075 /* if yes, the client has roamed and we have
2076 * to unclaim it.
2077 */
2078 if (batadv_has_timed_out(claim->lasttime, 100)) {
2079 /* only unclaim if the last claim entry is
2080 * older than 100 ms to make sure we really
2081 * have a roaming client here.
2082 */
2083 batadv_dbg(BATADV_DBG_BLA, bat_priv, "%s(): Roaming client %pM detected. Unclaim it.\n",
2084 __func__, ethhdr->h_source);
2085 batadv_handle_unclaim(bat_priv, primary_if,
2086 primary_if->net_dev->dev_addr,
2087 ethhdr->h_source, vid);
2088 goto allow;
2089 } else {
2090 batadv_dbg(BATADV_DBG_BLA, bat_priv, "%s(): Race for claim %pM detected. Drop packet.\n",
2091 __func__, ethhdr->h_source);
2092 goto handled;
2093 }
2094 }
2095
2096 /* check if it is a multicast/broadcast frame */
2097 if (is_multicast_ether_addr(ethhdr->h_dest)) {
2098 /* drop it. the responsible gateway has forwarded it into
2099 * the backbone network.
2100 */
2101 goto handled;
2102 } else {
2103 /* we must allow it. at least if we are
2104 * responsible for the DESTINATION.
2105 */
2106 goto allow;
2107 }
2108allow:
2109 batadv_bla_update_own_backbone_gw(bat_priv, primary_if, vid);
2110 ret = false;
2111 goto out;
2112handled:
2113 ret = true;
2114out:
2115 if (primary_if)
2116 batadv_hardif_put(primary_if);
2117 if (claim)
2118 batadv_claim_put(claim);
2119 return ret;
2120}
2121
2122#ifdef CONFIG_BATMAN_ADV_DEBUGFS
2123/**
2124 * batadv_bla_claim_table_seq_print_text() - print the claim table in a seq file
2125 * @seq: seq file to print on
2126 * @offset: not used
2127 *
2128 * Return: always 0
2129 */
2130int batadv_bla_claim_table_seq_print_text(struct seq_file *seq, void *offset)
2131{
2132 struct net_device *net_dev = (struct net_device *)seq->private;
2133 struct batadv_priv *bat_priv = netdev_priv(net_dev);
2134 struct batadv_hashtable *hash = bat_priv->bla.claim_hash;
2135 struct batadv_bla_backbone_gw *backbone_gw;
2136 struct batadv_bla_claim *claim;
2137 struct batadv_hard_iface *primary_if;
2138 struct hlist_head *head;
2139 u16 backbone_crc;
2140 u32 i;
2141 bool is_own;
2142 u8 *primary_addr;
2143
2144 primary_if = batadv_seq_print_text_primary_if_get(seq);
2145 if (!primary_if)
2146 goto out;
2147
2148 primary_addr = primary_if->net_dev->dev_addr;
2149 seq_printf(seq,
2150 "Claims announced for the mesh %s (orig %pM, group id %#.4x)\n",
2151 net_dev->name, primary_addr,
2152 ntohs(bat_priv->bla.claim_dest.group));
2153 seq_puts(seq,
2154 " Client VID Originator [o] (CRC )\n");
2155 for (i = 0; i < hash->size; i++) {
2156 head = &hash->table[i];
2157
2158 rcu_read_lock();
2159 hlist_for_each_entry_rcu(claim, head, hash_entry) {
2160 backbone_gw = batadv_bla_claim_get_backbone_gw(claim);
2161
2162 is_own = batadv_compare_eth(backbone_gw->orig,
2163 primary_addr);
2164
2165 spin_lock_bh(&backbone_gw->crc_lock);
2166 backbone_crc = backbone_gw->crc;
2167 spin_unlock_bh(&backbone_gw->crc_lock);
2168 seq_printf(seq, " * %pM on %5d by %pM [%c] (%#.4x)\n",
2169 claim->addr, batadv_print_vid(claim->vid),
2170 backbone_gw->orig,
2171 (is_own ? 'x' : ' '),
2172 backbone_crc);
2173
2174 batadv_backbone_gw_put(backbone_gw);
2175 }
2176 rcu_read_unlock();
2177 }
2178out:
2179 if (primary_if)
2180 batadv_hardif_put(primary_if);
2181 return 0;
2182}
2183#endif
2184
2185/**
2186 * batadv_bla_claim_dump_entry() - dump one entry of the claim table
2187 * to a netlink socket
2188 * @msg: buffer for the message
2189 * @portid: netlink port
David Brazdil0f672f62019-12-10 10:32:29 +00002190 * @cb: Control block containing additional options
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002191 * @primary_if: primary interface
2192 * @claim: entry to dump
2193 *
2194 * Return: 0 or error code.
2195 */
2196static int
David Brazdil0f672f62019-12-10 10:32:29 +00002197batadv_bla_claim_dump_entry(struct sk_buff *msg, u32 portid,
2198 struct netlink_callback *cb,
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002199 struct batadv_hard_iface *primary_if,
2200 struct batadv_bla_claim *claim)
2201{
2202 u8 *primary_addr = primary_if->net_dev->dev_addr;
2203 u16 backbone_crc;
2204 bool is_own;
2205 void *hdr;
2206 int ret = -EINVAL;
2207
David Brazdil0f672f62019-12-10 10:32:29 +00002208 hdr = genlmsg_put(msg, portid, cb->nlh->nlmsg_seq,
2209 &batadv_netlink_family, NLM_F_MULTI,
2210 BATADV_CMD_GET_BLA_CLAIM);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002211 if (!hdr) {
2212 ret = -ENOBUFS;
2213 goto out;
2214 }
2215
David Brazdil0f672f62019-12-10 10:32:29 +00002216 genl_dump_check_consistent(cb, hdr);
2217
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002218 is_own = batadv_compare_eth(claim->backbone_gw->orig,
2219 primary_addr);
2220
2221 spin_lock_bh(&claim->backbone_gw->crc_lock);
2222 backbone_crc = claim->backbone_gw->crc;
2223 spin_unlock_bh(&claim->backbone_gw->crc_lock);
2224
2225 if (is_own)
2226 if (nla_put_flag(msg, BATADV_ATTR_BLA_OWN)) {
2227 genlmsg_cancel(msg, hdr);
2228 goto out;
2229 }
2230
2231 if (nla_put(msg, BATADV_ATTR_BLA_ADDRESS, ETH_ALEN, claim->addr) ||
2232 nla_put_u16(msg, BATADV_ATTR_BLA_VID, claim->vid) ||
2233 nla_put(msg, BATADV_ATTR_BLA_BACKBONE, ETH_ALEN,
2234 claim->backbone_gw->orig) ||
2235 nla_put_u16(msg, BATADV_ATTR_BLA_CRC,
2236 backbone_crc)) {
2237 genlmsg_cancel(msg, hdr);
2238 goto out;
2239 }
2240
2241 genlmsg_end(msg, hdr);
2242 ret = 0;
2243
2244out:
2245 return ret;
2246}
2247
2248/**
2249 * batadv_bla_claim_dump_bucket() - dump one bucket of the claim table
2250 * to a netlink socket
2251 * @msg: buffer for the message
2252 * @portid: netlink port
David Brazdil0f672f62019-12-10 10:32:29 +00002253 * @cb: Control block containing additional options
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002254 * @primary_if: primary interface
David Brazdil0f672f62019-12-10 10:32:29 +00002255 * @hash: hash to dump
2256 * @bucket: bucket index to dump
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002257 * @idx_skip: How many entries to skip
2258 *
2259 * Return: always 0.
2260 */
2261static int
David Brazdil0f672f62019-12-10 10:32:29 +00002262batadv_bla_claim_dump_bucket(struct sk_buff *msg, u32 portid,
2263 struct netlink_callback *cb,
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002264 struct batadv_hard_iface *primary_if,
David Brazdil0f672f62019-12-10 10:32:29 +00002265 struct batadv_hashtable *hash, unsigned int bucket,
2266 int *idx_skip)
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002267{
2268 struct batadv_bla_claim *claim;
2269 int idx = 0;
2270 int ret = 0;
2271
David Brazdil0f672f62019-12-10 10:32:29 +00002272 spin_lock_bh(&hash->list_locks[bucket]);
2273 cb->seq = atomic_read(&hash->generation) << 1 | 1;
2274
2275 hlist_for_each_entry(claim, &hash->table[bucket], hash_entry) {
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002276 if (idx++ < *idx_skip)
2277 continue;
2278
David Brazdil0f672f62019-12-10 10:32:29 +00002279 ret = batadv_bla_claim_dump_entry(msg, portid, cb,
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002280 primary_if, claim);
2281 if (ret) {
2282 *idx_skip = idx - 1;
2283 goto unlock;
2284 }
2285 }
2286
2287 *idx_skip = 0;
2288unlock:
David Brazdil0f672f62019-12-10 10:32:29 +00002289 spin_unlock_bh(&hash->list_locks[bucket]);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002290 return ret;
2291}
2292
2293/**
2294 * batadv_bla_claim_dump() - dump claim table to a netlink socket
2295 * @msg: buffer for the message
2296 * @cb: callback structure containing arguments
2297 *
2298 * Return: message length.
2299 */
2300int batadv_bla_claim_dump(struct sk_buff *msg, struct netlink_callback *cb)
2301{
2302 struct batadv_hard_iface *primary_if = NULL;
2303 int portid = NETLINK_CB(cb->skb).portid;
2304 struct net *net = sock_net(cb->skb->sk);
2305 struct net_device *soft_iface;
2306 struct batadv_hashtable *hash;
2307 struct batadv_priv *bat_priv;
2308 int bucket = cb->args[0];
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002309 int idx = cb->args[1];
2310 int ifindex;
2311 int ret = 0;
2312
2313 ifindex = batadv_netlink_get_ifindex(cb->nlh,
2314 BATADV_ATTR_MESH_IFINDEX);
2315 if (!ifindex)
2316 return -EINVAL;
2317
2318 soft_iface = dev_get_by_index(net, ifindex);
2319 if (!soft_iface || !batadv_softif_is_valid(soft_iface)) {
2320 ret = -ENODEV;
2321 goto out;
2322 }
2323
2324 bat_priv = netdev_priv(soft_iface);
2325 hash = bat_priv->bla.claim_hash;
2326
2327 primary_if = batadv_primary_if_get_selected(bat_priv);
2328 if (!primary_if || primary_if->if_status != BATADV_IF_ACTIVE) {
2329 ret = -ENOENT;
2330 goto out;
2331 }
2332
2333 while (bucket < hash->size) {
David Brazdil0f672f62019-12-10 10:32:29 +00002334 if (batadv_bla_claim_dump_bucket(msg, portid, cb, primary_if,
2335 hash, bucket, &idx))
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002336 break;
2337 bucket++;
2338 }
2339
2340 cb->args[0] = bucket;
2341 cb->args[1] = idx;
2342
2343 ret = msg->len;
2344
2345out:
2346 if (primary_if)
2347 batadv_hardif_put(primary_if);
2348
2349 if (soft_iface)
2350 dev_put(soft_iface);
2351
2352 return ret;
2353}
2354
2355#ifdef CONFIG_BATMAN_ADV_DEBUGFS
2356/**
2357 * batadv_bla_backbone_table_seq_print_text() - print the backbone table in a
2358 * seq file
2359 * @seq: seq file to print on
2360 * @offset: not used
2361 *
2362 * Return: always 0
2363 */
2364int batadv_bla_backbone_table_seq_print_text(struct seq_file *seq, void *offset)
2365{
2366 struct net_device *net_dev = (struct net_device *)seq->private;
2367 struct batadv_priv *bat_priv = netdev_priv(net_dev);
2368 struct batadv_hashtable *hash = bat_priv->bla.backbone_hash;
2369 struct batadv_bla_backbone_gw *backbone_gw;
2370 struct batadv_hard_iface *primary_if;
2371 struct hlist_head *head;
2372 int secs, msecs;
2373 u16 backbone_crc;
2374 u32 i;
2375 bool is_own;
2376 u8 *primary_addr;
2377
2378 primary_if = batadv_seq_print_text_primary_if_get(seq);
2379 if (!primary_if)
2380 goto out;
2381
2382 primary_addr = primary_if->net_dev->dev_addr;
2383 seq_printf(seq,
2384 "Backbones announced for the mesh %s (orig %pM, group id %#.4x)\n",
2385 net_dev->name, primary_addr,
2386 ntohs(bat_priv->bla.claim_dest.group));
2387 seq_puts(seq, " Originator VID last seen (CRC )\n");
2388 for (i = 0; i < hash->size; i++) {
2389 head = &hash->table[i];
2390
2391 rcu_read_lock();
2392 hlist_for_each_entry_rcu(backbone_gw, head, hash_entry) {
2393 msecs = jiffies_to_msecs(jiffies -
2394 backbone_gw->lasttime);
2395 secs = msecs / 1000;
2396 msecs = msecs % 1000;
2397
2398 is_own = batadv_compare_eth(backbone_gw->orig,
2399 primary_addr);
2400 if (is_own)
2401 continue;
2402
2403 spin_lock_bh(&backbone_gw->crc_lock);
2404 backbone_crc = backbone_gw->crc;
2405 spin_unlock_bh(&backbone_gw->crc_lock);
2406
2407 seq_printf(seq, " * %pM on %5d %4i.%03is (%#.4x)\n",
2408 backbone_gw->orig,
2409 batadv_print_vid(backbone_gw->vid), secs,
2410 msecs, backbone_crc);
2411 }
2412 rcu_read_unlock();
2413 }
2414out:
2415 if (primary_if)
2416 batadv_hardif_put(primary_if);
2417 return 0;
2418}
2419#endif
2420
2421/**
2422 * batadv_bla_backbone_dump_entry() - dump one entry of the backbone table to a
2423 * netlink socket
2424 * @msg: buffer for the message
2425 * @portid: netlink port
David Brazdil0f672f62019-12-10 10:32:29 +00002426 * @cb: Control block containing additional options
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002427 * @primary_if: primary interface
2428 * @backbone_gw: entry to dump
2429 *
2430 * Return: 0 or error code.
2431 */
2432static int
David Brazdil0f672f62019-12-10 10:32:29 +00002433batadv_bla_backbone_dump_entry(struct sk_buff *msg, u32 portid,
2434 struct netlink_callback *cb,
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002435 struct batadv_hard_iface *primary_if,
2436 struct batadv_bla_backbone_gw *backbone_gw)
2437{
2438 u8 *primary_addr = primary_if->net_dev->dev_addr;
2439 u16 backbone_crc;
2440 bool is_own;
2441 int msecs;
2442 void *hdr;
2443 int ret = -EINVAL;
2444
David Brazdil0f672f62019-12-10 10:32:29 +00002445 hdr = genlmsg_put(msg, portid, cb->nlh->nlmsg_seq,
2446 &batadv_netlink_family, NLM_F_MULTI,
2447 BATADV_CMD_GET_BLA_BACKBONE);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002448 if (!hdr) {
2449 ret = -ENOBUFS;
2450 goto out;
2451 }
2452
David Brazdil0f672f62019-12-10 10:32:29 +00002453 genl_dump_check_consistent(cb, hdr);
2454
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002455 is_own = batadv_compare_eth(backbone_gw->orig, primary_addr);
2456
2457 spin_lock_bh(&backbone_gw->crc_lock);
2458 backbone_crc = backbone_gw->crc;
2459 spin_unlock_bh(&backbone_gw->crc_lock);
2460
2461 msecs = jiffies_to_msecs(jiffies - backbone_gw->lasttime);
2462
2463 if (is_own)
2464 if (nla_put_flag(msg, BATADV_ATTR_BLA_OWN)) {
2465 genlmsg_cancel(msg, hdr);
2466 goto out;
2467 }
2468
2469 if (nla_put(msg, BATADV_ATTR_BLA_BACKBONE, ETH_ALEN,
2470 backbone_gw->orig) ||
2471 nla_put_u16(msg, BATADV_ATTR_BLA_VID, backbone_gw->vid) ||
2472 nla_put_u16(msg, BATADV_ATTR_BLA_CRC,
2473 backbone_crc) ||
2474 nla_put_u32(msg, BATADV_ATTR_LAST_SEEN_MSECS, msecs)) {
2475 genlmsg_cancel(msg, hdr);
2476 goto out;
2477 }
2478
2479 genlmsg_end(msg, hdr);
2480 ret = 0;
2481
2482out:
2483 return ret;
2484}
2485
2486/**
2487 * batadv_bla_backbone_dump_bucket() - dump one bucket of the backbone table to
2488 * a netlink socket
2489 * @msg: buffer for the message
2490 * @portid: netlink port
David Brazdil0f672f62019-12-10 10:32:29 +00002491 * @cb: Control block containing additional options
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002492 * @primary_if: primary interface
David Brazdil0f672f62019-12-10 10:32:29 +00002493 * @hash: hash to dump
2494 * @bucket: bucket index to dump
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002495 * @idx_skip: How many entries to skip
2496 *
2497 * Return: always 0.
2498 */
2499static int
David Brazdil0f672f62019-12-10 10:32:29 +00002500batadv_bla_backbone_dump_bucket(struct sk_buff *msg, u32 portid,
2501 struct netlink_callback *cb,
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002502 struct batadv_hard_iface *primary_if,
David Brazdil0f672f62019-12-10 10:32:29 +00002503 struct batadv_hashtable *hash,
2504 unsigned int bucket, int *idx_skip)
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002505{
2506 struct batadv_bla_backbone_gw *backbone_gw;
2507 int idx = 0;
2508 int ret = 0;
2509
David Brazdil0f672f62019-12-10 10:32:29 +00002510 spin_lock_bh(&hash->list_locks[bucket]);
2511 cb->seq = atomic_read(&hash->generation) << 1 | 1;
2512
2513 hlist_for_each_entry(backbone_gw, &hash->table[bucket], hash_entry) {
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002514 if (idx++ < *idx_skip)
2515 continue;
2516
David Brazdil0f672f62019-12-10 10:32:29 +00002517 ret = batadv_bla_backbone_dump_entry(msg, portid, cb,
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002518 primary_if, backbone_gw);
2519 if (ret) {
2520 *idx_skip = idx - 1;
2521 goto unlock;
2522 }
2523 }
2524
2525 *idx_skip = 0;
2526unlock:
David Brazdil0f672f62019-12-10 10:32:29 +00002527 spin_unlock_bh(&hash->list_locks[bucket]);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002528 return ret;
2529}
2530
2531/**
2532 * batadv_bla_backbone_dump() - dump backbone table to a netlink socket
2533 * @msg: buffer for the message
2534 * @cb: callback structure containing arguments
2535 *
2536 * Return: message length.
2537 */
2538int batadv_bla_backbone_dump(struct sk_buff *msg, struct netlink_callback *cb)
2539{
2540 struct batadv_hard_iface *primary_if = NULL;
2541 int portid = NETLINK_CB(cb->skb).portid;
2542 struct net *net = sock_net(cb->skb->sk);
2543 struct net_device *soft_iface;
2544 struct batadv_hashtable *hash;
2545 struct batadv_priv *bat_priv;
2546 int bucket = cb->args[0];
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002547 int idx = cb->args[1];
2548 int ifindex;
2549 int ret = 0;
2550
2551 ifindex = batadv_netlink_get_ifindex(cb->nlh,
2552 BATADV_ATTR_MESH_IFINDEX);
2553 if (!ifindex)
2554 return -EINVAL;
2555
2556 soft_iface = dev_get_by_index(net, ifindex);
2557 if (!soft_iface || !batadv_softif_is_valid(soft_iface)) {
2558 ret = -ENODEV;
2559 goto out;
2560 }
2561
2562 bat_priv = netdev_priv(soft_iface);
2563 hash = bat_priv->bla.backbone_hash;
2564
2565 primary_if = batadv_primary_if_get_selected(bat_priv);
2566 if (!primary_if || primary_if->if_status != BATADV_IF_ACTIVE) {
2567 ret = -ENOENT;
2568 goto out;
2569 }
2570
2571 while (bucket < hash->size) {
David Brazdil0f672f62019-12-10 10:32:29 +00002572 if (batadv_bla_backbone_dump_bucket(msg, portid, cb, primary_if,
2573 hash, bucket, &idx))
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002574 break;
2575 bucket++;
2576 }
2577
2578 cb->args[0] = bucket;
2579 cb->args[1] = idx;
2580
2581 ret = msg->len;
2582
2583out:
2584 if (primary_if)
2585 batadv_hardif_put(primary_if);
2586
2587 if (soft_iface)
2588 dev_put(soft_iface);
2589
2590 return ret;
2591}
2592
2593#ifdef CONFIG_BATMAN_ADV_DAT
2594/**
2595 * batadv_bla_check_claim() - check if address is claimed
2596 *
2597 * @bat_priv: the bat priv with all the soft interface information
2598 * @addr: mac address of which the claim status is checked
2599 * @vid: the VLAN ID
2600 *
2601 * addr is checked if this address is claimed by the local device itself.
2602 *
2603 * Return: true if bla is disabled or the mac is claimed by the device,
2604 * false if the device addr is already claimed by another gateway
2605 */
2606bool batadv_bla_check_claim(struct batadv_priv *bat_priv,
2607 u8 *addr, unsigned short vid)
2608{
2609 struct batadv_bla_claim search_claim;
2610 struct batadv_bla_claim *claim = NULL;
2611 struct batadv_hard_iface *primary_if = NULL;
2612 bool ret = true;
2613
2614 if (!atomic_read(&bat_priv->bridge_loop_avoidance))
2615 return ret;
2616
2617 primary_if = batadv_primary_if_get_selected(bat_priv);
2618 if (!primary_if)
2619 return ret;
2620
2621 /* First look if the mac address is claimed */
2622 ether_addr_copy(search_claim.addr, addr);
2623 search_claim.vid = vid;
2624
2625 claim = batadv_claim_hash_find(bat_priv, &search_claim);
2626
2627 /* If there is a claim and we are not owner of the claim,
2628 * return false.
2629 */
2630 if (claim) {
2631 if (!batadv_compare_eth(claim->backbone_gw->orig,
2632 primary_if->net_dev->dev_addr))
2633 ret = false;
2634 batadv_claim_put(claim);
2635 }
2636
2637 batadv_hardif_put(primary_if);
2638 return ret;
2639}
2640#endif