You've already forked linux-apfs
mirror of
https://github.com/linux-apfs/linux-apfs.git
synced 2026-05-01 15:00:59 -07:00
batman-adv: Split batadv_priv in sub-structures for features
The structure batadv_priv grows everytime a new feature is introduced. It gets hard to find the parts of the struct that belongs to a specific feature. This becomes even harder by the fact that not every feature uses a prefix in the member name. The variables for bridge loop avoidence, gateway handling, translation table and visualization server are moved into separate structs that are included in the bat_priv main struct. Signed-off-by: Sven Eckelmann <sven@narfation.org> Signed-off-by: Antonio Quartulli <ordex@autistici.org>
This commit is contained in:
committed by
Antonio Quartulli
parent
624463079e
commit
807736f6e0
@@ -603,8 +603,8 @@ static void batadv_iv_ogm_schedule(struct batadv_hard_iface *hard_iface)
|
||||
htonl((uint32_t)atomic_read(&hard_iface->seqno));
|
||||
atomic_inc(&hard_iface->seqno);
|
||||
|
||||
batadv_ogm_packet->ttvn = atomic_read(&bat_priv->ttvn);
|
||||
batadv_ogm_packet->tt_crc = htons(bat_priv->tt_crc);
|
||||
batadv_ogm_packet->ttvn = atomic_read(&bat_priv->tt.vn);
|
||||
batadv_ogm_packet->tt_crc = htons(bat_priv->tt.local_crc);
|
||||
if (tt_num_changes >= 0)
|
||||
batadv_ogm_packet->tt_num_changes = tt_num_changes;
|
||||
|
||||
|
||||
@@ -133,7 +133,7 @@ static void batadv_claim_free_ref(struct batadv_claim *claim)
|
||||
static struct batadv_claim *batadv_claim_hash_find(struct batadv_priv *bat_priv,
|
||||
struct batadv_claim *data)
|
||||
{
|
||||
struct batadv_hashtable *hash = bat_priv->claim_hash;
|
||||
struct batadv_hashtable *hash = bat_priv->bla.claim_hash;
|
||||
struct hlist_head *head;
|
||||
struct hlist_node *node;
|
||||
struct batadv_claim *claim;
|
||||
@@ -174,7 +174,7 @@ static struct batadv_backbone_gw *
|
||||
batadv_backbone_hash_find(struct batadv_priv *bat_priv,
|
||||
uint8_t *addr, short vid)
|
||||
{
|
||||
struct batadv_hashtable *hash = bat_priv->backbone_hash;
|
||||
struct batadv_hashtable *hash = bat_priv->bla.backbone_hash;
|
||||
struct hlist_head *head;
|
||||
struct hlist_node *node;
|
||||
struct batadv_backbone_gw search_entry, *backbone_gw;
|
||||
@@ -218,7 +218,7 @@ batadv_bla_del_backbone_claims(struct batadv_backbone_gw *backbone_gw)
|
||||
int i;
|
||||
spinlock_t *list_lock; /* protects write access to the hash lists */
|
||||
|
||||
hash = backbone_gw->bat_priv->claim_hash;
|
||||
hash = backbone_gw->bat_priv->bla.claim_hash;
|
||||
if (!hash)
|
||||
return;
|
||||
|
||||
@@ -265,7 +265,7 @@ static void batadv_bla_send_claim(struct batadv_priv *bat_priv, uint8_t *mac,
|
||||
if (!primary_if)
|
||||
return;
|
||||
|
||||
memcpy(&local_claim_dest, &bat_priv->claim_dest,
|
||||
memcpy(&local_claim_dest, &bat_priv->bla.claim_dest,
|
||||
sizeof(local_claim_dest));
|
||||
local_claim_dest.type = claimtype;
|
||||
|
||||
@@ -391,7 +391,7 @@ batadv_bla_get_backbone_gw(struct batadv_priv *bat_priv, uint8_t *orig,
|
||||
/* one for the hash, one for returning */
|
||||
atomic_set(&entry->refcount, 2);
|
||||
|
||||
hash_added = batadv_hash_add(bat_priv->backbone_hash,
|
||||
hash_added = batadv_hash_add(bat_priv->bla.backbone_hash,
|
||||
batadv_compare_backbone_gw,
|
||||
batadv_choose_backbone_gw, entry,
|
||||
&entry->hash_entry);
|
||||
@@ -458,7 +458,7 @@ static void batadv_bla_answer_request(struct batadv_priv *bat_priv,
|
||||
if (!backbone_gw)
|
||||
return;
|
||||
|
||||
hash = bat_priv->claim_hash;
|
||||
hash = bat_priv->bla.claim_hash;
|
||||
for (i = 0; i < hash->size; i++) {
|
||||
head = &hash->table[i];
|
||||
|
||||
@@ -499,7 +499,7 @@ static void batadv_bla_send_request(struct batadv_backbone_gw *backbone_gw)
|
||||
|
||||
/* no local broadcasts should be sent or received, for now. */
|
||||
if (!atomic_read(&backbone_gw->request_sent)) {
|
||||
atomic_inc(&backbone_gw->bat_priv->bla_num_requests);
|
||||
atomic_inc(&backbone_gw->bat_priv->bla.num_requests);
|
||||
atomic_set(&backbone_gw->request_sent, 1);
|
||||
}
|
||||
}
|
||||
@@ -559,7 +559,7 @@ static void batadv_bla_add_claim(struct batadv_priv *bat_priv,
|
||||
batadv_dbg(BATADV_DBG_BLA, bat_priv,
|
||||
"bla_add_claim(): adding new entry %pM, vid %d to hash ...\n",
|
||||
mac, vid);
|
||||
hash_added = batadv_hash_add(bat_priv->claim_hash,
|
||||
hash_added = batadv_hash_add(bat_priv->bla.claim_hash,
|
||||
batadv_compare_claim,
|
||||
batadv_choose_claim, claim,
|
||||
&claim->hash_entry);
|
||||
@@ -612,7 +612,7 @@ static void batadv_bla_del_claim(struct batadv_priv *bat_priv,
|
||||
batadv_dbg(BATADV_DBG_BLA, bat_priv, "bla_del_claim(): %pM, vid %d\n",
|
||||
mac, vid);
|
||||
|
||||
batadv_hash_remove(bat_priv->claim_hash, batadv_compare_claim,
|
||||
batadv_hash_remove(bat_priv->bla.claim_hash, batadv_compare_claim,
|
||||
batadv_choose_claim, claim);
|
||||
batadv_claim_free_ref(claim); /* reference from the hash is gone */
|
||||
|
||||
@@ -659,7 +659,7 @@ static int batadv_handle_announce(struct batadv_priv *bat_priv,
|
||||
* we can allow traffic again.
|
||||
*/
|
||||
if (atomic_read(&backbone_gw->request_sent)) {
|
||||
atomic_dec(&backbone_gw->bat_priv->bla_num_requests);
|
||||
atomic_dec(&backbone_gw->bat_priv->bla.num_requests);
|
||||
atomic_set(&backbone_gw->request_sent, 0);
|
||||
}
|
||||
}
|
||||
@@ -774,7 +774,7 @@ static int batadv_check_claim_group(struct batadv_priv *bat_priv,
|
||||
struct batadv_bla_claim_dst *bla_dst, *bla_dst_own;
|
||||
|
||||
bla_dst = (struct batadv_bla_claim_dst *)hw_dst;
|
||||
bla_dst_own = &bat_priv->claim_dest;
|
||||
bla_dst_own = &bat_priv->bla.claim_dest;
|
||||
|
||||
/* check if it is a claim packet in general */
|
||||
if (memcmp(bla_dst->magic, bla_dst_own->magic,
|
||||
@@ -947,7 +947,7 @@ static void batadv_bla_purge_backbone_gw(struct batadv_priv *bat_priv, int now)
|
||||
spinlock_t *list_lock; /* protects write access to the hash lists */
|
||||
int i;
|
||||
|
||||
hash = bat_priv->backbone_hash;
|
||||
hash = bat_priv->bla.backbone_hash;
|
||||
if (!hash)
|
||||
return;
|
||||
|
||||
@@ -971,7 +971,7 @@ static void batadv_bla_purge_backbone_gw(struct batadv_priv *bat_priv, int now)
|
||||
purge_now:
|
||||
/* don't wait for the pending request anymore */
|
||||
if (atomic_read(&backbone_gw->request_sent))
|
||||
atomic_dec(&bat_priv->bla_num_requests);
|
||||
atomic_dec(&bat_priv->bla.num_requests);
|
||||
|
||||
batadv_bla_del_backbone_claims(backbone_gw);
|
||||
|
||||
@@ -1001,7 +1001,7 @@ static void batadv_bla_purge_claims(struct batadv_priv *bat_priv,
|
||||
struct batadv_hashtable *hash;
|
||||
int i;
|
||||
|
||||
hash = bat_priv->claim_hash;
|
||||
hash = bat_priv->bla.claim_hash;
|
||||
if (!hash)
|
||||
return;
|
||||
|
||||
@@ -1048,11 +1048,12 @@ void batadv_bla_update_orig_address(struct batadv_priv *bat_priv,
|
||||
struct hlist_node *node;
|
||||
struct hlist_head *head;
|
||||
struct batadv_hashtable *hash;
|
||||
__be16 group;
|
||||
int i;
|
||||
|
||||
/* reset bridge loop avoidance group id */
|
||||
bat_priv->claim_dest.group =
|
||||
htons(crc16(0, primary_if->net_dev->dev_addr, ETH_ALEN));
|
||||
group = htons(crc16(0, primary_if->net_dev->dev_addr, ETH_ALEN));
|
||||
bat_priv->bla.claim_dest.group = group;
|
||||
|
||||
if (!oldif) {
|
||||
batadv_bla_purge_claims(bat_priv, NULL, 1);
|
||||
@@ -1060,7 +1061,7 @@ void batadv_bla_update_orig_address(struct batadv_priv *bat_priv,
|
||||
return;
|
||||
}
|
||||
|
||||
hash = bat_priv->backbone_hash;
|
||||
hash = bat_priv->bla.backbone_hash;
|
||||
if (!hash)
|
||||
return;
|
||||
|
||||
@@ -1090,8 +1091,8 @@ void batadv_bla_update_orig_address(struct batadv_priv *bat_priv,
|
||||
/* (re)start the timer */
|
||||
static void batadv_bla_start_timer(struct batadv_priv *bat_priv)
|
||||
{
|
||||
INIT_DELAYED_WORK(&bat_priv->bla_work, batadv_bla_periodic_work);
|
||||
queue_delayed_work(batadv_event_workqueue, &bat_priv->bla_work,
|
||||
INIT_DELAYED_WORK(&bat_priv->bla.work, batadv_bla_periodic_work);
|
||||
queue_delayed_work(batadv_event_workqueue, &bat_priv->bla.work,
|
||||
msecs_to_jiffies(BATADV_BLA_PERIOD_LENGTH));
|
||||
}
|
||||
|
||||
@@ -1104,6 +1105,7 @@ static void batadv_bla_periodic_work(struct work_struct *work)
|
||||
struct delayed_work *delayed_work =
|
||||
container_of(work, struct delayed_work, work);
|
||||
struct batadv_priv *bat_priv;
|
||||
struct batadv_priv_bla *priv_bla;
|
||||
struct hlist_node *node;
|
||||
struct hlist_head *head;
|
||||
struct batadv_backbone_gw *backbone_gw;
|
||||
@@ -1111,7 +1113,8 @@ static void batadv_bla_periodic_work(struct work_struct *work)
|
||||
struct batadv_hard_iface *primary_if;
|
||||
int i;
|
||||
|
||||
bat_priv = container_of(delayed_work, struct batadv_priv, bla_work);
|
||||
priv_bla = container_of(delayed_work, struct batadv_priv_bla, work);
|
||||
bat_priv = container_of(priv_bla, struct batadv_priv, bla);
|
||||
primary_if = batadv_primary_if_get_selected(bat_priv);
|
||||
if (!primary_if)
|
||||
goto out;
|
||||
@@ -1122,7 +1125,7 @@ static void batadv_bla_periodic_work(struct work_struct *work)
|
||||
if (!atomic_read(&bat_priv->bridge_loop_avoidance))
|
||||
goto out;
|
||||
|
||||
hash = bat_priv->backbone_hash;
|
||||
hash = bat_priv->bla.backbone_hash;
|
||||
if (!hash)
|
||||
goto out;
|
||||
|
||||
@@ -1162,40 +1165,41 @@ int batadv_bla_init(struct batadv_priv *bat_priv)
|
||||
int i;
|
||||
uint8_t claim_dest[ETH_ALEN] = {0xff, 0x43, 0x05, 0x00, 0x00, 0x00};
|
||||
struct batadv_hard_iface *primary_if;
|
||||
uint16_t crc;
|
||||
unsigned long entrytime;
|
||||
|
||||
batadv_dbg(BATADV_DBG_BLA, bat_priv, "bla hash registering\n");
|
||||
|
||||
/* setting claim destination address */
|
||||
memcpy(&bat_priv->claim_dest.magic, claim_dest, 3);
|
||||
bat_priv->claim_dest.type = 0;
|
||||
memcpy(&bat_priv->bla.claim_dest.magic, claim_dest, 3);
|
||||
bat_priv->bla.claim_dest.type = 0;
|
||||
primary_if = batadv_primary_if_get_selected(bat_priv);
|
||||
if (primary_if) {
|
||||
bat_priv->claim_dest.group =
|
||||
htons(crc16(0, primary_if->net_dev->dev_addr,
|
||||
ETH_ALEN));
|
||||
crc = crc16(0, primary_if->net_dev->dev_addr, ETH_ALEN);
|
||||
bat_priv->bla.claim_dest.group = htons(crc);
|
||||
batadv_hardif_free_ref(primary_if);
|
||||
} else {
|
||||
bat_priv->claim_dest.group = 0; /* will be set later */
|
||||
bat_priv->bla.claim_dest.group = 0; /* will be set later */
|
||||
}
|
||||
|
||||
/* initialize the duplicate list */
|
||||
entrytime = jiffies - msecs_to_jiffies(BATADV_DUPLIST_TIMEOUT);
|
||||
for (i = 0; i < BATADV_DUPLIST_SIZE; i++)
|
||||
bat_priv->bcast_duplist[i].entrytime =
|
||||
jiffies - msecs_to_jiffies(BATADV_DUPLIST_TIMEOUT);
|
||||
bat_priv->bcast_duplist_curr = 0;
|
||||
bat_priv->bla.bcast_duplist[i].entrytime = entrytime;
|
||||
bat_priv->bla.bcast_duplist_curr = 0;
|
||||
|
||||
if (bat_priv->claim_hash)
|
||||
if (bat_priv->bla.claim_hash)
|
||||
return 0;
|
||||
|
||||
bat_priv->claim_hash = batadv_hash_new(128);
|
||||
bat_priv->backbone_hash = batadv_hash_new(32);
|
||||
bat_priv->bla.claim_hash = batadv_hash_new(128);
|
||||
bat_priv->bla.backbone_hash = batadv_hash_new(32);
|
||||
|
||||
if (!bat_priv->claim_hash || !bat_priv->backbone_hash)
|
||||
if (!bat_priv->bla.claim_hash || !bat_priv->bla.backbone_hash)
|
||||
return -ENOMEM;
|
||||
|
||||
batadv_hash_set_lock_class(bat_priv->claim_hash,
|
||||
batadv_hash_set_lock_class(bat_priv->bla.claim_hash,
|
||||
&batadv_claim_hash_lock_class_key);
|
||||
batadv_hash_set_lock_class(bat_priv->backbone_hash,
|
||||
batadv_hash_set_lock_class(bat_priv->bla.backbone_hash,
|
||||
&batadv_backbone_hash_lock_class_key);
|
||||
|
||||
batadv_dbg(BATADV_DBG_BLA, bat_priv, "bla hashes initialized\n");
|
||||
@@ -1236,8 +1240,9 @@ int batadv_bla_check_bcast_duplist(struct batadv_priv *bat_priv,
|
||||
crc = crc16(0, content, length);
|
||||
|
||||
for (i = 0; i < BATADV_DUPLIST_SIZE; i++) {
|
||||
curr = (bat_priv->bcast_duplist_curr + i) % BATADV_DUPLIST_SIZE;
|
||||
entry = &bat_priv->bcast_duplist[curr];
|
||||
curr = (bat_priv->bla.bcast_duplist_curr + i);
|
||||
curr %= BATADV_DUPLIST_SIZE;
|
||||
entry = &bat_priv->bla.bcast_duplist[curr];
|
||||
|
||||
/* we can stop searching if the entry is too old ;
|
||||
* later entries will be even older
|
||||
@@ -1258,13 +1263,13 @@ int batadv_bla_check_bcast_duplist(struct batadv_priv *bat_priv,
|
||||
return 1;
|
||||
}
|
||||
/* not found, add a new entry (overwrite the oldest entry) */
|
||||
curr = (bat_priv->bcast_duplist_curr + BATADV_DUPLIST_SIZE - 1);
|
||||
curr = (bat_priv->bla.bcast_duplist_curr + BATADV_DUPLIST_SIZE - 1);
|
||||
curr %= BATADV_DUPLIST_SIZE;
|
||||
entry = &bat_priv->bcast_duplist[curr];
|
||||
entry = &bat_priv->bla.bcast_duplist[curr];
|
||||
entry->crc = crc;
|
||||
entry->entrytime = jiffies;
|
||||
memcpy(entry->orig, bcast_packet->orig, ETH_ALEN);
|
||||
bat_priv->bcast_duplist_curr = curr;
|
||||
bat_priv->bla.bcast_duplist_curr = curr;
|
||||
|
||||
/* allow it, its the first occurence. */
|
||||
return 0;
|
||||
@@ -1281,7 +1286,7 @@ int batadv_bla_check_bcast_duplist(struct batadv_priv *bat_priv,
|
||||
*/
|
||||
int batadv_bla_is_backbone_gw_orig(struct batadv_priv *bat_priv, uint8_t *orig)
|
||||
{
|
||||
struct batadv_hashtable *hash = bat_priv->backbone_hash;
|
||||
struct batadv_hashtable *hash = bat_priv->bla.backbone_hash;
|
||||
struct hlist_head *head;
|
||||
struct hlist_node *node;
|
||||
struct batadv_backbone_gw *backbone_gw;
|
||||
@@ -1361,18 +1366,18 @@ void batadv_bla_free(struct batadv_priv *bat_priv)
|
||||
{
|
||||
struct batadv_hard_iface *primary_if;
|
||||
|
||||
cancel_delayed_work_sync(&bat_priv->bla_work);
|
||||
cancel_delayed_work_sync(&bat_priv->bla.work);
|
||||
primary_if = batadv_primary_if_get_selected(bat_priv);
|
||||
|
||||
if (bat_priv->claim_hash) {
|
||||
if (bat_priv->bla.claim_hash) {
|
||||
batadv_bla_purge_claims(bat_priv, primary_if, 1);
|
||||
batadv_hash_destroy(bat_priv->claim_hash);
|
||||
bat_priv->claim_hash = NULL;
|
||||
batadv_hash_destroy(bat_priv->bla.claim_hash);
|
||||
bat_priv->bla.claim_hash = NULL;
|
||||
}
|
||||
if (bat_priv->backbone_hash) {
|
||||
if (bat_priv->bla.backbone_hash) {
|
||||
batadv_bla_purge_backbone_gw(bat_priv, 1);
|
||||
batadv_hash_destroy(bat_priv->backbone_hash);
|
||||
bat_priv->backbone_hash = NULL;
|
||||
batadv_hash_destroy(bat_priv->bla.backbone_hash);
|
||||
bat_priv->bla.backbone_hash = NULL;
|
||||
}
|
||||
if (primary_if)
|
||||
batadv_hardif_free_ref(primary_if);
|
||||
@@ -1411,7 +1416,7 @@ int batadv_bla_rx(struct batadv_priv *bat_priv, struct sk_buff *skb, short vid,
|
||||
goto allow;
|
||||
|
||||
|
||||
if (unlikely(atomic_read(&bat_priv->bla_num_requests)))
|
||||
if (unlikely(atomic_read(&bat_priv->bla.num_requests)))
|
||||
/* don't allow broadcasts while requests are in flight */
|
||||
if (is_multicast_ether_addr(ethhdr->h_dest) && is_bcast)
|
||||
goto handled;
|
||||
@@ -1510,7 +1515,7 @@ int batadv_bla_tx(struct batadv_priv *bat_priv, struct sk_buff *skb, short vid)
|
||||
|
||||
ethhdr = (struct ethhdr *)skb_mac_header(skb);
|
||||
|
||||
if (unlikely(atomic_read(&bat_priv->bla_num_requests)))
|
||||
if (unlikely(atomic_read(&bat_priv->bla.num_requests)))
|
||||
/* don't allow broadcasts while requests are in flight */
|
||||
if (is_multicast_ether_addr(ethhdr->h_dest))
|
||||
goto handled;
|
||||
@@ -1566,7 +1571,7 @@ int batadv_bla_claim_table_seq_print_text(struct seq_file *seq, void *offset)
|
||||
{
|
||||
struct net_device *net_dev = (struct net_device *)seq->private;
|
||||
struct batadv_priv *bat_priv = netdev_priv(net_dev);
|
||||
struct batadv_hashtable *hash = bat_priv->claim_hash;
|
||||
struct batadv_hashtable *hash = bat_priv->bla.claim_hash;
|
||||
struct batadv_claim *claim;
|
||||
struct batadv_hard_iface *primary_if;
|
||||
struct hlist_node *node;
|
||||
@@ -1595,7 +1600,7 @@ int batadv_bla_claim_table_seq_print_text(struct seq_file *seq, void *offset)
|
||||
seq_printf(seq,
|
||||
"Claims announced for the mesh %s (orig %pM, group id %04x)\n",
|
||||
net_dev->name, primary_addr,
|
||||
ntohs(bat_priv->claim_dest.group));
|
||||
ntohs(bat_priv->bla.claim_dest.group));
|
||||
seq_printf(seq, " %-17s %-5s %-17s [o] (%-4s)\n",
|
||||
"Client", "VID", "Originator", "CRC");
|
||||
for (i = 0; i < hash->size; i++) {
|
||||
@@ -1623,7 +1628,7 @@ int batadv_bla_backbone_table_seq_print_text(struct seq_file *seq, void *offset)
|
||||
{
|
||||
struct net_device *net_dev = (struct net_device *)seq->private;
|
||||
struct batadv_priv *bat_priv = netdev_priv(net_dev);
|
||||
struct batadv_hashtable *hash = bat_priv->backbone_hash;
|
||||
struct batadv_hashtable *hash = bat_priv->bla.backbone_hash;
|
||||
struct batadv_backbone_gw *backbone_gw;
|
||||
struct batadv_hard_iface *primary_if;
|
||||
struct hlist_node *node;
|
||||
@@ -1653,7 +1658,7 @@ int batadv_bla_backbone_table_seq_print_text(struct seq_file *seq, void *offset)
|
||||
seq_printf(seq,
|
||||
"Backbones announced for the mesh %s (orig %pM, group id %04x)\n",
|
||||
net_dev->name, primary_addr,
|
||||
ntohs(bat_priv->claim_dest.group));
|
||||
ntohs(bat_priv->bla.claim_dest.group));
|
||||
seq_printf(seq, " %-17s %-5s %-9s (%-4s)\n",
|
||||
"Originator", "VID", "last seen", "CRC");
|
||||
for (i = 0; i < hash->size; i++) {
|
||||
|
||||
@@ -48,7 +48,7 @@ batadv_gw_get_selected_gw_node(struct batadv_priv *bat_priv)
|
||||
struct batadv_gw_node *gw_node;
|
||||
|
||||
rcu_read_lock();
|
||||
gw_node = rcu_dereference(bat_priv->curr_gw);
|
||||
gw_node = rcu_dereference(bat_priv->gw.curr_gw);
|
||||
if (!gw_node)
|
||||
goto out;
|
||||
|
||||
@@ -91,23 +91,23 @@ static void batadv_gw_select(struct batadv_priv *bat_priv,
|
||||
{
|
||||
struct batadv_gw_node *curr_gw_node;
|
||||
|
||||
spin_lock_bh(&bat_priv->gw_list_lock);
|
||||
spin_lock_bh(&bat_priv->gw.list_lock);
|
||||
|
||||
if (new_gw_node && !atomic_inc_not_zero(&new_gw_node->refcount))
|
||||
new_gw_node = NULL;
|
||||
|
||||
curr_gw_node = rcu_dereference_protected(bat_priv->curr_gw, 1);
|
||||
rcu_assign_pointer(bat_priv->curr_gw, new_gw_node);
|
||||
curr_gw_node = rcu_dereference_protected(bat_priv->gw.curr_gw, 1);
|
||||
rcu_assign_pointer(bat_priv->gw.curr_gw, new_gw_node);
|
||||
|
||||
if (curr_gw_node)
|
||||
batadv_gw_node_free_ref(curr_gw_node);
|
||||
|
||||
spin_unlock_bh(&bat_priv->gw_list_lock);
|
||||
spin_unlock_bh(&bat_priv->gw.list_lock);
|
||||
}
|
||||
|
||||
void batadv_gw_deselect(struct batadv_priv *bat_priv)
|
||||
{
|
||||
atomic_set(&bat_priv->gw_reselect, 1);
|
||||
atomic_set(&bat_priv->gw.reselect, 1);
|
||||
}
|
||||
|
||||
static struct batadv_gw_node *
|
||||
@@ -122,7 +122,7 @@ batadv_gw_get_best_gw_node(struct batadv_priv *bat_priv)
|
||||
struct batadv_orig_node *orig_node;
|
||||
|
||||
rcu_read_lock();
|
||||
hlist_for_each_entry_rcu(gw_node, node, &bat_priv->gw_list, list) {
|
||||
hlist_for_each_entry_rcu(gw_node, node, &bat_priv->gw.list, list) {
|
||||
if (gw_node->deleted)
|
||||
continue;
|
||||
|
||||
@@ -202,7 +202,7 @@ void batadv_gw_election(struct batadv_priv *bat_priv)
|
||||
|
||||
curr_gw = batadv_gw_get_selected_gw_node(bat_priv);
|
||||
|
||||
if (!batadv_atomic_dec_not_zero(&bat_priv->gw_reselect) && curr_gw)
|
||||
if (!batadv_atomic_dec_not_zero(&bat_priv->gw.reselect) && curr_gw)
|
||||
goto out;
|
||||
|
||||
next_gw = batadv_gw_get_best_gw_node(bat_priv);
|
||||
@@ -321,9 +321,9 @@ static void batadv_gw_node_add(struct batadv_priv *bat_priv,
|
||||
gw_node->orig_node = orig_node;
|
||||
atomic_set(&gw_node->refcount, 1);
|
||||
|
||||
spin_lock_bh(&bat_priv->gw_list_lock);
|
||||
hlist_add_head_rcu(&gw_node->list, &bat_priv->gw_list);
|
||||
spin_unlock_bh(&bat_priv->gw_list_lock);
|
||||
spin_lock_bh(&bat_priv->gw.list_lock);
|
||||
hlist_add_head_rcu(&gw_node->list, &bat_priv->gw.list);
|
||||
spin_unlock_bh(&bat_priv->gw.list_lock);
|
||||
|
||||
batadv_gw_bandwidth_to_kbit(new_gwflags, &down, &up);
|
||||
batadv_dbg(BATADV_DBG_BATMAN, bat_priv,
|
||||
@@ -350,7 +350,7 @@ void batadv_gw_node_update(struct batadv_priv *bat_priv,
|
||||
curr_gw = batadv_gw_get_selected_gw_node(bat_priv);
|
||||
|
||||
rcu_read_lock();
|
||||
hlist_for_each_entry_rcu(gw_node, node, &bat_priv->gw_list, list) {
|
||||
hlist_for_each_entry_rcu(gw_node, node, &bat_priv->gw.list, list) {
|
||||
if (gw_node->orig_node != orig_node)
|
||||
continue;
|
||||
|
||||
@@ -404,10 +404,10 @@ void batadv_gw_node_purge(struct batadv_priv *bat_priv)
|
||||
|
||||
curr_gw = batadv_gw_get_selected_gw_node(bat_priv);
|
||||
|
||||
spin_lock_bh(&bat_priv->gw_list_lock);
|
||||
spin_lock_bh(&bat_priv->gw.list_lock);
|
||||
|
||||
hlist_for_each_entry_safe(gw_node, node, node_tmp,
|
||||
&bat_priv->gw_list, list) {
|
||||
&bat_priv->gw.list, list) {
|
||||
if (((!gw_node->deleted) ||
|
||||
(time_before(jiffies, gw_node->deleted + timeout))) &&
|
||||
atomic_read(&bat_priv->mesh_state) == BATADV_MESH_ACTIVE)
|
||||
@@ -420,7 +420,7 @@ void batadv_gw_node_purge(struct batadv_priv *bat_priv)
|
||||
batadv_gw_node_free_ref(gw_node);
|
||||
}
|
||||
|
||||
spin_unlock_bh(&bat_priv->gw_list_lock);
|
||||
spin_unlock_bh(&bat_priv->gw.list_lock);
|
||||
|
||||
/* gw_deselect() needs to acquire the gw_list_lock */
|
||||
if (do_deselect)
|
||||
@@ -496,7 +496,7 @@ int batadv_gw_client_seq_print_text(struct seq_file *seq, void *offset)
|
||||
primary_if->net_dev->dev_addr, net_dev->name);
|
||||
|
||||
rcu_read_lock();
|
||||
hlist_for_each_entry_rcu(gw_node, node, &bat_priv->gw_list, list) {
|
||||
hlist_for_each_entry_rcu(gw_node, node, &bat_priv->gw.list, list) {
|
||||
if (gw_node->deleted)
|
||||
continue;
|
||||
|
||||
|
||||
@@ -103,13 +103,14 @@ static void batadv_primary_if_update_addr(struct batadv_priv *bat_priv,
|
||||
{
|
||||
struct batadv_vis_packet *vis_packet;
|
||||
struct batadv_hard_iface *primary_if;
|
||||
struct sk_buff *skb;
|
||||
|
||||
primary_if = batadv_primary_if_get_selected(bat_priv);
|
||||
if (!primary_if)
|
||||
goto out;
|
||||
|
||||
vis_packet = (struct batadv_vis_packet *)
|
||||
bat_priv->my_vis_info->skb_packet->data;
|
||||
skb = bat_priv->vis.my_info->skb_packet;
|
||||
vis_packet = (struct batadv_vis_packet *)skb->data;
|
||||
memcpy(vis_packet->vis_orig, primary_if->net_dev->dev_addr, ETH_ALEN);
|
||||
memcpy(vis_packet->sender_orig,
|
||||
primary_if->net_dev->dev_addr, ETH_ALEN);
|
||||
|
||||
+12
-12
@@ -94,20 +94,20 @@ int batadv_mesh_init(struct net_device *soft_iface)
|
||||
|
||||
spin_lock_init(&bat_priv->forw_bat_list_lock);
|
||||
spin_lock_init(&bat_priv->forw_bcast_list_lock);
|
||||
spin_lock_init(&bat_priv->tt_changes_list_lock);
|
||||
spin_lock_init(&bat_priv->tt_req_list_lock);
|
||||
spin_lock_init(&bat_priv->tt_roam_list_lock);
|
||||
spin_lock_init(&bat_priv->tt_buff_lock);
|
||||
spin_lock_init(&bat_priv->gw_list_lock);
|
||||
spin_lock_init(&bat_priv->vis_hash_lock);
|
||||
spin_lock_init(&bat_priv->vis_list_lock);
|
||||
spin_lock_init(&bat_priv->tt.changes_list_lock);
|
||||
spin_lock_init(&bat_priv->tt.req_list_lock);
|
||||
spin_lock_init(&bat_priv->tt.roam_list_lock);
|
||||
spin_lock_init(&bat_priv->tt.last_changeset_lock);
|
||||
spin_lock_init(&bat_priv->gw.list_lock);
|
||||
spin_lock_init(&bat_priv->vis.hash_lock);
|
||||
spin_lock_init(&bat_priv->vis.list_lock);
|
||||
|
||||
INIT_HLIST_HEAD(&bat_priv->forw_bat_list);
|
||||
INIT_HLIST_HEAD(&bat_priv->forw_bcast_list);
|
||||
INIT_HLIST_HEAD(&bat_priv->gw_list);
|
||||
INIT_LIST_HEAD(&bat_priv->tt_changes_list);
|
||||
INIT_LIST_HEAD(&bat_priv->tt_req_list);
|
||||
INIT_LIST_HEAD(&bat_priv->tt_roam_list);
|
||||
INIT_HLIST_HEAD(&bat_priv->gw.list);
|
||||
INIT_LIST_HEAD(&bat_priv->tt.changes_list);
|
||||
INIT_LIST_HEAD(&bat_priv->tt.req_list);
|
||||
INIT_LIST_HEAD(&bat_priv->tt.roam_list);
|
||||
|
||||
ret = batadv_originator_init(bat_priv);
|
||||
if (ret < 0)
|
||||
@@ -128,7 +128,7 @@ int batadv_mesh_init(struct net_device *soft_iface)
|
||||
if (ret < 0)
|
||||
goto err;
|
||||
|
||||
atomic_set(&bat_priv->gw_reselect, 0);
|
||||
atomic_set(&bat_priv->gw.reselect, 0);
|
||||
atomic_set(&bat_priv->mesh_state, BATADV_MESH_ACTIVE);
|
||||
|
||||
return 0;
|
||||
|
||||
@@ -721,7 +721,7 @@ int batadv_recv_roam_adv(struct sk_buff *skb, struct batadv_hard_iface *recv_if)
|
||||
* been incremented yet. This flag will make me check all the incoming
|
||||
* packets for the correct destination.
|
||||
*/
|
||||
bat_priv->tt_poss_change = true;
|
||||
bat_priv->tt.poss_change = true;
|
||||
|
||||
batadv_orig_node_free_ref(orig_node);
|
||||
out:
|
||||
@@ -947,8 +947,8 @@ static int batadv_check_unicast_ttvn(struct batadv_priv *bat_priv,
|
||||
unicast_packet = (struct batadv_unicast_packet *)skb->data;
|
||||
|
||||
if (batadv_is_my_mac(unicast_packet->dest)) {
|
||||
tt_poss_change = bat_priv->tt_poss_change;
|
||||
curr_ttvn = (uint8_t)atomic_read(&bat_priv->ttvn);
|
||||
tt_poss_change = bat_priv->tt.poss_change;
|
||||
curr_ttvn = (uint8_t)atomic_read(&bat_priv->tt.vn);
|
||||
} else {
|
||||
orig_node = batadv_orig_hash_find(bat_priv,
|
||||
unicast_packet->dest);
|
||||
|
||||
@@ -420,14 +420,15 @@ struct net_device *batadv_softif_create(const char *name)
|
||||
|
||||
atomic_set(&bat_priv->mesh_state, BATADV_MESH_INACTIVE);
|
||||
atomic_set(&bat_priv->bcast_seqno, 1);
|
||||
atomic_set(&bat_priv->ttvn, 0);
|
||||
atomic_set(&bat_priv->tt_local_changes, 0);
|
||||
atomic_set(&bat_priv->tt_ogm_append_cnt, 0);
|
||||
atomic_set(&bat_priv->bla_num_requests, 0);
|
||||
|
||||
bat_priv->tt_buff = NULL;
|
||||
bat_priv->tt_buff_len = 0;
|
||||
bat_priv->tt_poss_change = false;
|
||||
atomic_set(&bat_priv->tt.vn, 0);
|
||||
atomic_set(&bat_priv->tt.local_changes, 0);
|
||||
atomic_set(&bat_priv->tt.ogm_append_cnt, 0);
|
||||
#ifdef CONFIG_BATMAN_ADV_BLA
|
||||
atomic_set(&bat_priv->bla.num_requests, 0);
|
||||
#endif
|
||||
bat_priv->tt.last_changeset = NULL;
|
||||
bat_priv->tt.last_changeset_len = 0;
|
||||
bat_priv->tt.poss_change = false;
|
||||
|
||||
bat_priv->primary_if = NULL;
|
||||
bat_priv->num_ifaces = 0;
|
||||
|
||||
+105
-103
File diff suppressed because it is too large
Load Diff
+67
-46
@@ -165,6 +165,67 @@ enum batadv_counters {
|
||||
BATADV_CNT_NUM,
|
||||
};
|
||||
|
||||
/**
|
||||
* struct batadv_priv_tt - per mesh interface translation table data
|
||||
* @vn: translation table version number
|
||||
* @local_changes: changes registered in an originator interval
|
||||
* @poss_change: Detect an ongoing roaming phase. If true, then this node
|
||||
* received a roaming_adv and has to inspect every packet directed to it to
|
||||
* check whether it still is the true destination or not. This flag will be
|
||||
* reset to false as soon as the this node's ttvn is increased
|
||||
* @changes_list: tracks tt local changes within an originator interval
|
||||
* @req_list: list of pending tt_requests
|
||||
* @local_crc: Checksum of the local table, recomputed before sending a new OGM
|
||||
*/
|
||||
struct batadv_priv_tt {
|
||||
atomic_t vn;
|
||||
atomic_t ogm_append_cnt;
|
||||
atomic_t local_changes;
|
||||
bool poss_change;
|
||||
struct list_head changes_list;
|
||||
struct batadv_hashtable *local_hash;
|
||||
struct batadv_hashtable *global_hash;
|
||||
struct list_head req_list;
|
||||
struct list_head roam_list;
|
||||
spinlock_t changes_list_lock; /* protects changes */
|
||||
spinlock_t req_list_lock; /* protects req_list */
|
||||
spinlock_t roam_list_lock; /* protects roam_list */
|
||||
atomic_t local_entry_num;
|
||||
uint16_t local_crc;
|
||||
unsigned char *last_changeset;
|
||||
int16_t last_changeset_len;
|
||||
spinlock_t last_changeset_lock; /* protects last_changeset */
|
||||
struct delayed_work work;
|
||||
};
|
||||
|
||||
#ifdef CONFIG_BATMAN_ADV_BLA
|
||||
struct batadv_priv_bla {
|
||||
atomic_t num_requests; /* number of bla requests in flight */
|
||||
struct batadv_hashtable *claim_hash;
|
||||
struct batadv_hashtable *backbone_hash;
|
||||
struct batadv_bcast_duplist_entry bcast_duplist[BATADV_DUPLIST_SIZE];
|
||||
int bcast_duplist_curr;
|
||||
struct batadv_bla_claim_dst claim_dest;
|
||||
struct delayed_work work;
|
||||
};
|
||||
#endif
|
||||
|
||||
struct batadv_priv_gw {
|
||||
struct hlist_head list;
|
||||
spinlock_t list_lock; /* protects gw_list and curr_gw */
|
||||
struct batadv_gw_node __rcu *curr_gw; /* rcu protected pointer */
|
||||
atomic_t reselect;
|
||||
};
|
||||
|
||||
struct batadv_priv_vis {
|
||||
struct list_head send_list;
|
||||
struct batadv_hashtable *hash;
|
||||
spinlock_t hash_lock; /* protects hash */
|
||||
spinlock_t list_lock; /* protects info::recv_list */
|
||||
struct delayed_work work;
|
||||
struct batadv_vis_info *my_info;
|
||||
};
|
||||
|
||||
struct batadv_priv {
|
||||
atomic_t mesh_state;
|
||||
struct net_device_stats stats;
|
||||
@@ -184,64 +245,24 @@ struct batadv_priv {
|
||||
atomic_t bcast_seqno;
|
||||
atomic_t bcast_queue_left;
|
||||
atomic_t batman_queue_left;
|
||||
atomic_t ttvn; /* translation table version number */
|
||||
atomic_t tt_ogm_append_cnt;
|
||||
atomic_t tt_local_changes; /* changes registered in a OGM interval */
|
||||
atomic_t bla_num_requests; /* number of bla requests in flight */
|
||||
/* The tt_poss_change flag is used to detect an ongoing roaming phase.
|
||||
* If true, then I received a Roaming_adv and I have to inspect every
|
||||
* packet directed to me to check whether I am still the true
|
||||
* destination or not. This flag will be reset to false as soon as I
|
||||
* increase my TTVN
|
||||
*/
|
||||
bool tt_poss_change;
|
||||
char num_ifaces;
|
||||
struct batadv_debug_log *debug_log;
|
||||
struct kobject *mesh_obj;
|
||||
struct dentry *debug_dir;
|
||||
struct hlist_head forw_bat_list;
|
||||
struct hlist_head forw_bcast_list;
|
||||
struct hlist_head gw_list;
|
||||
struct list_head tt_changes_list; /* tracks changes in a OGM int */
|
||||
struct list_head vis_send_list;
|
||||
struct batadv_hashtable *orig_hash;
|
||||
struct batadv_hashtable *tt_local_hash;
|
||||
struct batadv_hashtable *tt_global_hash;
|
||||
#ifdef CONFIG_BATMAN_ADV_BLA
|
||||
struct batadv_hashtable *claim_hash;
|
||||
struct batadv_hashtable *backbone_hash;
|
||||
#endif
|
||||
struct list_head tt_req_list; /* list of pending tt_requests */
|
||||
struct list_head tt_roam_list;
|
||||
struct batadv_hashtable *vis_hash;
|
||||
#ifdef CONFIG_BATMAN_ADV_BLA
|
||||
struct batadv_bcast_duplist_entry bcast_duplist[BATADV_DUPLIST_SIZE];
|
||||
int bcast_duplist_curr;
|
||||
struct batadv_bla_claim_dst claim_dest;
|
||||
#endif
|
||||
spinlock_t forw_bat_list_lock; /* protects forw_bat_list */
|
||||
spinlock_t forw_bcast_list_lock; /* protects */
|
||||
spinlock_t tt_changes_list_lock; /* protects tt_changes */
|
||||
spinlock_t tt_req_list_lock; /* protects tt_req_list */
|
||||
spinlock_t tt_roam_list_lock; /* protects tt_roam_list */
|
||||
spinlock_t gw_list_lock; /* protects gw_list and curr_gw */
|
||||
spinlock_t vis_hash_lock; /* protects vis_hash */
|
||||
spinlock_t vis_list_lock; /* protects vis_info::recv_list */
|
||||
atomic_t num_local_tt;
|
||||
/* Checksum of the local table, recomputed before sending a new OGM */
|
||||
uint16_t tt_crc;
|
||||
unsigned char *tt_buff;
|
||||
int16_t tt_buff_len;
|
||||
spinlock_t tt_buff_lock; /* protects tt_buff */
|
||||
struct delayed_work tt_work;
|
||||
struct delayed_work orig_work;
|
||||
struct delayed_work vis_work;
|
||||
struct delayed_work bla_work;
|
||||
struct batadv_gw_node __rcu *curr_gw; /* rcu protected pointer */
|
||||
atomic_t gw_reselect;
|
||||
struct batadv_hard_iface __rcu *primary_if; /* rcu protected pointer */
|
||||
struct batadv_vis_info *my_vis_info;
|
||||
struct batadv_algo_ops *bat_algo_ops;
|
||||
#ifdef CONFIG_BATMAN_ADV_BLA
|
||||
struct batadv_priv_bla bla;
|
||||
#endif
|
||||
struct batadv_priv_gw gw;
|
||||
struct batadv_priv_tt tt;
|
||||
struct batadv_priv_vis vis;
|
||||
};
|
||||
|
||||
struct batadv_socket_client {
|
||||
|
||||
+66
-64
@@ -41,13 +41,13 @@ static void batadv_free_info(struct kref *ref)
|
||||
bat_priv = info->bat_priv;
|
||||
|
||||
list_del_init(&info->send_list);
|
||||
spin_lock_bh(&bat_priv->vis_list_lock);
|
||||
spin_lock_bh(&bat_priv->vis.list_lock);
|
||||
list_for_each_entry_safe(entry, tmp, &info->recv_list, list) {
|
||||
list_del(&entry->list);
|
||||
kfree(entry);
|
||||
}
|
||||
|
||||
spin_unlock_bh(&bat_priv->vis_list_lock);
|
||||
spin_unlock_bh(&bat_priv->vis.list_lock);
|
||||
kfree_skb(info->skb_packet);
|
||||
kfree(info);
|
||||
}
|
||||
@@ -94,7 +94,7 @@ static uint32_t batadv_vis_info_choose(const void *data, uint32_t size)
|
||||
static struct batadv_vis_info *
|
||||
batadv_vis_hash_find(struct batadv_priv *bat_priv, const void *data)
|
||||
{
|
||||
struct batadv_hashtable *hash = bat_priv->vis_hash;
|
||||
struct batadv_hashtable *hash = bat_priv->vis.hash;
|
||||
struct hlist_head *head;
|
||||
struct hlist_node *node;
|
||||
struct batadv_vis_info *vis_info, *vis_info_tmp = NULL;
|
||||
@@ -252,7 +252,7 @@ int batadv_vis_seq_print_text(struct seq_file *seq, void *offset)
|
||||
struct hlist_head *head;
|
||||
struct net_device *net_dev = (struct net_device *)seq->private;
|
||||
struct batadv_priv *bat_priv = netdev_priv(net_dev);
|
||||
struct batadv_hashtable *hash = bat_priv->vis_hash;
|
||||
struct batadv_hashtable *hash = bat_priv->vis.hash;
|
||||
uint32_t i;
|
||||
int ret = 0;
|
||||
int vis_server = atomic_read(&bat_priv->vis_mode);
|
||||
@@ -264,12 +264,12 @@ int batadv_vis_seq_print_text(struct seq_file *seq, void *offset)
|
||||
if (vis_server == BATADV_VIS_TYPE_CLIENT_UPDATE)
|
||||
goto out;
|
||||
|
||||
spin_lock_bh(&bat_priv->vis_hash_lock);
|
||||
spin_lock_bh(&bat_priv->vis.hash_lock);
|
||||
for (i = 0; i < hash->size; i++) {
|
||||
head = &hash->table[i];
|
||||
batadv_vis_seq_print_text_bucket(seq, head);
|
||||
}
|
||||
spin_unlock_bh(&bat_priv->vis_hash_lock);
|
||||
spin_unlock_bh(&bat_priv->vis.hash_lock);
|
||||
|
||||
out:
|
||||
if (primary_if)
|
||||
@@ -285,7 +285,7 @@ static void batadv_send_list_add(struct batadv_priv *bat_priv,
|
||||
{
|
||||
if (list_empty(&info->send_list)) {
|
||||
kref_get(&info->refcount);
|
||||
list_add_tail(&info->send_list, &bat_priv->vis_send_list);
|
||||
list_add_tail(&info->send_list, &bat_priv->vis.send_list);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -311,9 +311,9 @@ static void batadv_recv_list_add(struct batadv_priv *bat_priv,
|
||||
return;
|
||||
|
||||
memcpy(entry->mac, mac, ETH_ALEN);
|
||||
spin_lock_bh(&bat_priv->vis_list_lock);
|
||||
spin_lock_bh(&bat_priv->vis.list_lock);
|
||||
list_add_tail(&entry->list, recv_list);
|
||||
spin_unlock_bh(&bat_priv->vis_list_lock);
|
||||
spin_unlock_bh(&bat_priv->vis.list_lock);
|
||||
}
|
||||
|
||||
/* returns 1 if this mac is in the recv_list */
|
||||
@@ -323,14 +323,14 @@ static int batadv_recv_list_is_in(struct batadv_priv *bat_priv,
|
||||
{
|
||||
const struct batadv_recvlist_node *entry;
|
||||
|
||||
spin_lock_bh(&bat_priv->vis_list_lock);
|
||||
spin_lock_bh(&bat_priv->vis.list_lock);
|
||||
list_for_each_entry(entry, recv_list, list) {
|
||||
if (batadv_compare_eth(entry->mac, mac)) {
|
||||
spin_unlock_bh(&bat_priv->vis_list_lock);
|
||||
spin_unlock_bh(&bat_priv->vis.list_lock);
|
||||
return 1;
|
||||
}
|
||||
}
|
||||
spin_unlock_bh(&bat_priv->vis_list_lock);
|
||||
spin_unlock_bh(&bat_priv->vis.list_lock);
|
||||
return 0;
|
||||
}
|
||||
|
||||
@@ -354,7 +354,7 @@ batadv_add_packet(struct batadv_priv *bat_priv,
|
||||
|
||||
*is_new = 0;
|
||||
/* sanity check */
|
||||
if (!bat_priv->vis_hash)
|
||||
if (!bat_priv->vis.hash)
|
||||
return NULL;
|
||||
|
||||
/* see if the packet is already in vis_hash */
|
||||
@@ -385,7 +385,7 @@ batadv_add_packet(struct batadv_priv *bat_priv,
|
||||
}
|
||||
}
|
||||
/* remove old entry */
|
||||
batadv_hash_remove(bat_priv->vis_hash, batadv_vis_info_cmp,
|
||||
batadv_hash_remove(bat_priv->vis.hash, batadv_vis_info_cmp,
|
||||
batadv_vis_info_choose, old_info);
|
||||
batadv_send_list_del(old_info);
|
||||
kref_put(&old_info->refcount, batadv_free_info);
|
||||
@@ -426,7 +426,7 @@ batadv_add_packet(struct batadv_priv *bat_priv,
|
||||
batadv_recv_list_add(bat_priv, &info->recv_list, packet->sender_orig);
|
||||
|
||||
/* try to add it */
|
||||
hash_added = batadv_hash_add(bat_priv->vis_hash, batadv_vis_info_cmp,
|
||||
hash_added = batadv_hash_add(bat_priv->vis.hash, batadv_vis_info_cmp,
|
||||
batadv_vis_info_choose, info,
|
||||
&info->hash_entry);
|
||||
if (hash_added != 0) {
|
||||
@@ -449,7 +449,7 @@ void batadv_receive_server_sync_packet(struct batadv_priv *bat_priv,
|
||||
|
||||
make_broadcast = (vis_server == BATADV_VIS_TYPE_SERVER_SYNC);
|
||||
|
||||
spin_lock_bh(&bat_priv->vis_hash_lock);
|
||||
spin_lock_bh(&bat_priv->vis.hash_lock);
|
||||
info = batadv_add_packet(bat_priv, vis_packet, vis_info_len,
|
||||
&is_new, make_broadcast);
|
||||
if (!info)
|
||||
@@ -461,7 +461,7 @@ void batadv_receive_server_sync_packet(struct batadv_priv *bat_priv,
|
||||
if (vis_server == BATADV_VIS_TYPE_SERVER_SYNC && is_new)
|
||||
batadv_send_list_add(bat_priv, info);
|
||||
end:
|
||||
spin_unlock_bh(&bat_priv->vis_hash_lock);
|
||||
spin_unlock_bh(&bat_priv->vis.hash_lock);
|
||||
}
|
||||
|
||||
/* handle an incoming client update packet and schedule forward if needed. */
|
||||
@@ -484,7 +484,7 @@ void batadv_receive_client_update_packet(struct batadv_priv *bat_priv,
|
||||
batadv_is_my_mac(vis_packet->target_orig))
|
||||
are_target = 1;
|
||||
|
||||
spin_lock_bh(&bat_priv->vis_hash_lock);
|
||||
spin_lock_bh(&bat_priv->vis.hash_lock);
|
||||
info = batadv_add_packet(bat_priv, vis_packet, vis_info_len,
|
||||
&is_new, are_target);
|
||||
|
||||
@@ -505,7 +505,7 @@ void batadv_receive_client_update_packet(struct batadv_priv *bat_priv,
|
||||
}
|
||||
|
||||
end:
|
||||
spin_unlock_bh(&bat_priv->vis_hash_lock);
|
||||
spin_unlock_bh(&bat_priv->vis.hash_lock);
|
||||
}
|
||||
|
||||
/* Walk the originators and find the VIS server with the best tq. Set the packet
|
||||
@@ -574,7 +574,7 @@ static int batadv_generate_vis_packet(struct batadv_priv *bat_priv)
|
||||
struct hlist_head *head;
|
||||
struct batadv_orig_node *orig_node;
|
||||
struct batadv_neigh_node *router;
|
||||
struct batadv_vis_info *info = bat_priv->my_vis_info;
|
||||
struct batadv_vis_info *info = bat_priv->vis.my_info;
|
||||
struct batadv_vis_packet *packet;
|
||||
struct batadv_vis_info_entry *entry;
|
||||
struct batadv_tt_common_entry *tt_common_entry;
|
||||
@@ -636,7 +636,7 @@ next:
|
||||
rcu_read_unlock();
|
||||
}
|
||||
|
||||
hash = bat_priv->tt_local_hash;
|
||||
hash = bat_priv->tt.local_hash;
|
||||
|
||||
for (i = 0; i < hash->size; i++) {
|
||||
head = &hash->table[i];
|
||||
@@ -671,7 +671,7 @@ unlock:
|
||||
static void batadv_purge_vis_packets(struct batadv_priv *bat_priv)
|
||||
{
|
||||
uint32_t i;
|
||||
struct batadv_hashtable *hash = bat_priv->vis_hash;
|
||||
struct batadv_hashtable *hash = bat_priv->vis.hash;
|
||||
struct hlist_node *node, *node_tmp;
|
||||
struct hlist_head *head;
|
||||
struct batadv_vis_info *info;
|
||||
@@ -682,7 +682,7 @@ static void batadv_purge_vis_packets(struct batadv_priv *bat_priv)
|
||||
hlist_for_each_entry_safe(info, node, node_tmp,
|
||||
head, hash_entry) {
|
||||
/* never purge own data. */
|
||||
if (info == bat_priv->my_vis_info)
|
||||
if (info == bat_priv->vis.my_info)
|
||||
continue;
|
||||
|
||||
if (batadv_has_timed_out(info->first_seen,
|
||||
@@ -817,31 +817,33 @@ static void batadv_send_vis_packets(struct work_struct *work)
|
||||
struct delayed_work *delayed_work =
|
||||
container_of(work, struct delayed_work, work);
|
||||
struct batadv_priv *bat_priv;
|
||||
struct batadv_priv_vis *priv_vis;
|
||||
struct batadv_vis_info *info;
|
||||
|
||||
bat_priv = container_of(delayed_work, struct batadv_priv, vis_work);
|
||||
spin_lock_bh(&bat_priv->vis_hash_lock);
|
||||
priv_vis = container_of(delayed_work, struct batadv_priv_vis, work);
|
||||
bat_priv = container_of(priv_vis, struct batadv_priv, vis);
|
||||
spin_lock_bh(&bat_priv->vis.hash_lock);
|
||||
batadv_purge_vis_packets(bat_priv);
|
||||
|
||||
if (batadv_generate_vis_packet(bat_priv) == 0) {
|
||||
/* schedule if generation was successful */
|
||||
batadv_send_list_add(bat_priv, bat_priv->my_vis_info);
|
||||
batadv_send_list_add(bat_priv, bat_priv->vis.my_info);
|
||||
}
|
||||
|
||||
while (!list_empty(&bat_priv->vis_send_list)) {
|
||||
info = list_first_entry(&bat_priv->vis_send_list,
|
||||
while (!list_empty(&bat_priv->vis.send_list)) {
|
||||
info = list_first_entry(&bat_priv->vis.send_list,
|
||||
typeof(*info), send_list);
|
||||
|
||||
kref_get(&info->refcount);
|
||||
spin_unlock_bh(&bat_priv->vis_hash_lock);
|
||||
spin_unlock_bh(&bat_priv->vis.hash_lock);
|
||||
|
||||
batadv_send_vis_packet(bat_priv, info);
|
||||
|
||||
spin_lock_bh(&bat_priv->vis_hash_lock);
|
||||
spin_lock_bh(&bat_priv->vis.hash_lock);
|
||||
batadv_send_list_del(info);
|
||||
kref_put(&info->refcount, batadv_free_info);
|
||||
}
|
||||
spin_unlock_bh(&bat_priv->vis_hash_lock);
|
||||
spin_unlock_bh(&bat_priv->vis.hash_lock);
|
||||
batadv_start_vis_timer(bat_priv);
|
||||
}
|
||||
|
||||
@@ -856,37 +858,37 @@ int batadv_vis_init(struct batadv_priv *bat_priv)
|
||||
unsigned long first_seen;
|
||||
struct sk_buff *tmp_skb;
|
||||
|
||||
if (bat_priv->vis_hash)
|
||||
if (bat_priv->vis.hash)
|
||||
return 0;
|
||||
|
||||
spin_lock_bh(&bat_priv->vis_hash_lock);
|
||||
spin_lock_bh(&bat_priv->vis.hash_lock);
|
||||
|
||||
bat_priv->vis_hash = batadv_hash_new(256);
|
||||
if (!bat_priv->vis_hash) {
|
||||
bat_priv->vis.hash = batadv_hash_new(256);
|
||||
if (!bat_priv->vis.hash) {
|
||||
pr_err("Can't initialize vis_hash\n");
|
||||
goto err;
|
||||
}
|
||||
|
||||
bat_priv->my_vis_info = kmalloc(BATADV_MAX_VIS_PACKET_SIZE, GFP_ATOMIC);
|
||||
if (!bat_priv->my_vis_info)
|
||||
bat_priv->vis.my_info = kmalloc(BATADV_MAX_VIS_PACKET_SIZE, GFP_ATOMIC);
|
||||
if (!bat_priv->vis.my_info)
|
||||
goto err;
|
||||
|
||||
len = sizeof(*packet) + BATADV_MAX_VIS_PACKET_SIZE + ETH_HLEN;
|
||||
bat_priv->my_vis_info->skb_packet = dev_alloc_skb(len);
|
||||
if (!bat_priv->my_vis_info->skb_packet)
|
||||
bat_priv->vis.my_info->skb_packet = dev_alloc_skb(len);
|
||||
if (!bat_priv->vis.my_info->skb_packet)
|
||||
goto free_info;
|
||||
|
||||
skb_reserve(bat_priv->my_vis_info->skb_packet, ETH_HLEN);
|
||||
tmp_skb = bat_priv->my_vis_info->skb_packet;
|
||||
skb_reserve(bat_priv->vis.my_info->skb_packet, ETH_HLEN);
|
||||
tmp_skb = bat_priv->vis.my_info->skb_packet;
|
||||
packet = (struct batadv_vis_packet *)skb_put(tmp_skb, sizeof(*packet));
|
||||
|
||||
/* prefill the vis info */
|
||||
first_seen = jiffies - msecs_to_jiffies(BATADV_VIS_INTERVAL);
|
||||
bat_priv->my_vis_info->first_seen = first_seen;
|
||||
INIT_LIST_HEAD(&bat_priv->my_vis_info->recv_list);
|
||||
INIT_LIST_HEAD(&bat_priv->my_vis_info->send_list);
|
||||
kref_init(&bat_priv->my_vis_info->refcount);
|
||||
bat_priv->my_vis_info->bat_priv = bat_priv;
|
||||
bat_priv->vis.my_info->first_seen = first_seen;
|
||||
INIT_LIST_HEAD(&bat_priv->vis.my_info->recv_list);
|
||||
INIT_LIST_HEAD(&bat_priv->vis.my_info->send_list);
|
||||
kref_init(&bat_priv->vis.my_info->refcount);
|
||||
bat_priv->vis.my_info->bat_priv = bat_priv;
|
||||
packet->header.version = BATADV_COMPAT_VERSION;
|
||||
packet->header.packet_type = BATADV_VIS;
|
||||
packet->header.ttl = BATADV_TTL;
|
||||
@@ -894,28 +896,28 @@ int batadv_vis_init(struct batadv_priv *bat_priv)
|
||||
packet->reserved = 0;
|
||||
packet->entries = 0;
|
||||
|
||||
INIT_LIST_HEAD(&bat_priv->vis_send_list);
|
||||
INIT_LIST_HEAD(&bat_priv->vis.send_list);
|
||||
|
||||
hash_added = batadv_hash_add(bat_priv->vis_hash, batadv_vis_info_cmp,
|
||||
hash_added = batadv_hash_add(bat_priv->vis.hash, batadv_vis_info_cmp,
|
||||
batadv_vis_info_choose,
|
||||
bat_priv->my_vis_info,
|
||||
&bat_priv->my_vis_info->hash_entry);
|
||||
bat_priv->vis.my_info,
|
||||
&bat_priv->vis.my_info->hash_entry);
|
||||
if (hash_added != 0) {
|
||||
pr_err("Can't add own vis packet into hash\n");
|
||||
/* not in hash, need to remove it manually. */
|
||||
kref_put(&bat_priv->my_vis_info->refcount, batadv_free_info);
|
||||
kref_put(&bat_priv->vis.my_info->refcount, batadv_free_info);
|
||||
goto err;
|
||||
}
|
||||
|
||||
spin_unlock_bh(&bat_priv->vis_hash_lock);
|
||||
spin_unlock_bh(&bat_priv->vis.hash_lock);
|
||||
batadv_start_vis_timer(bat_priv);
|
||||
return 0;
|
||||
|
||||
free_info:
|
||||
kfree(bat_priv->my_vis_info);
|
||||
bat_priv->my_vis_info = NULL;
|
||||
kfree(bat_priv->vis.my_info);
|
||||
bat_priv->vis.my_info = NULL;
|
||||
err:
|
||||
spin_unlock_bh(&bat_priv->vis_hash_lock);
|
||||
spin_unlock_bh(&bat_priv->vis.hash_lock);
|
||||
batadv_vis_quit(bat_priv);
|
||||
return -ENOMEM;
|
||||
}
|
||||
@@ -933,23 +935,23 @@ static void batadv_free_info_ref(struct hlist_node *node, void *arg)
|
||||
/* shutdown vis-server */
|
||||
void batadv_vis_quit(struct batadv_priv *bat_priv)
|
||||
{
|
||||
if (!bat_priv->vis_hash)
|
||||
if (!bat_priv->vis.hash)
|
||||
return;
|
||||
|
||||
cancel_delayed_work_sync(&bat_priv->vis_work);
|
||||
cancel_delayed_work_sync(&bat_priv->vis.work);
|
||||
|
||||
spin_lock_bh(&bat_priv->vis_hash_lock);
|
||||
spin_lock_bh(&bat_priv->vis.hash_lock);
|
||||
/* properly remove, kill timers ... */
|
||||
batadv_hash_delete(bat_priv->vis_hash, batadv_free_info_ref, NULL);
|
||||
bat_priv->vis_hash = NULL;
|
||||
bat_priv->my_vis_info = NULL;
|
||||
spin_unlock_bh(&bat_priv->vis_hash_lock);
|
||||
batadv_hash_delete(bat_priv->vis.hash, batadv_free_info_ref, NULL);
|
||||
bat_priv->vis.hash = NULL;
|
||||
bat_priv->vis.my_info = NULL;
|
||||
spin_unlock_bh(&bat_priv->vis.hash_lock);
|
||||
}
|
||||
|
||||
/* schedule packets for (re)transmission */
|
||||
static void batadv_start_vis_timer(struct batadv_priv *bat_priv)
|
||||
{
|
||||
INIT_DELAYED_WORK(&bat_priv->vis_work, batadv_send_vis_packets);
|
||||
queue_delayed_work(batadv_event_workqueue, &bat_priv->vis_work,
|
||||
INIT_DELAYED_WORK(&bat_priv->vis.work, batadv_send_vis_packets);
|
||||
queue_delayed_work(batadv_event_workqueue, &bat_priv->vis.work,
|
||||
msecs_to_jiffies(BATADV_VIS_INTERVAL));
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user