|
@@ -26,40 +26,40 @@
|
|
#include "hash.h"
|
|
#include "hash.h"
|
|
#include "originator.h"
|
|
#include "originator.h"
|
|
|
|
|
|
-static void hna_local_purge(struct work_struct *work);
|
|
|
|
-static void _hna_global_del_orig(struct bat_priv *bat_priv,
|
|
|
|
- struct hna_global_entry *hna_global_entry,
|
|
|
|
|
|
+static void tt_local_purge(struct work_struct *work);
|
|
|
|
+static void _tt_global_del_orig(struct bat_priv *bat_priv,
|
|
|
|
+ struct tt_global_entry *tt_global_entry,
|
|
char *message);
|
|
char *message);
|
|
|
|
|
|
/* returns 1 if they are the same mac addr */
|
|
/* returns 1 if they are the same mac addr */
|
|
-static int compare_lhna(struct hlist_node *node, void *data2)
|
|
|
|
|
|
+static int compare_ltt(struct hlist_node *node, void *data2)
|
|
{
|
|
{
|
|
- void *data1 = container_of(node, struct hna_local_entry, hash_entry);
|
|
|
|
|
|
+ void *data1 = container_of(node, struct tt_local_entry, hash_entry);
|
|
|
|
|
|
return (memcmp(data1, data2, ETH_ALEN) == 0 ? 1 : 0);
|
|
return (memcmp(data1, data2, ETH_ALEN) == 0 ? 1 : 0);
|
|
}
|
|
}
|
|
|
|
|
|
/* returns 1 if they are the same mac addr */
|
|
/* returns 1 if they are the same mac addr */
|
|
-static int compare_ghna(struct hlist_node *node, void *data2)
|
|
|
|
|
|
+static int compare_gtt(struct hlist_node *node, void *data2)
|
|
{
|
|
{
|
|
- void *data1 = container_of(node, struct hna_global_entry, hash_entry);
|
|
|
|
|
|
+ void *data1 = container_of(node, struct tt_global_entry, hash_entry);
|
|
|
|
|
|
return (memcmp(data1, data2, ETH_ALEN) == 0 ? 1 : 0);
|
|
return (memcmp(data1, data2, ETH_ALEN) == 0 ? 1 : 0);
|
|
}
|
|
}
|
|
|
|
|
|
-static void hna_local_start_timer(struct bat_priv *bat_priv)
|
|
|
|
|
|
+static void tt_local_start_timer(struct bat_priv *bat_priv)
|
|
{
|
|
{
|
|
- INIT_DELAYED_WORK(&bat_priv->hna_work, hna_local_purge);
|
|
|
|
- queue_delayed_work(bat_event_workqueue, &bat_priv->hna_work, 10 * HZ);
|
|
|
|
|
|
+ INIT_DELAYED_WORK(&bat_priv->tt_work, tt_local_purge);
|
|
|
|
+ queue_delayed_work(bat_event_workqueue, &bat_priv->tt_work, 10 * HZ);
|
|
}
|
|
}
|
|
|
|
|
|
-static struct hna_local_entry *hna_local_hash_find(struct bat_priv *bat_priv,
|
|
|
|
|
|
+static struct tt_local_entry *tt_local_hash_find(struct bat_priv *bat_priv,
|
|
void *data)
|
|
void *data)
|
|
{
|
|
{
|
|
- struct hashtable_t *hash = bat_priv->hna_local_hash;
|
|
|
|
|
|
+ struct hashtable_t *hash = bat_priv->tt_local_hash;
|
|
struct hlist_head *head;
|
|
struct hlist_head *head;
|
|
struct hlist_node *node;
|
|
struct hlist_node *node;
|
|
- struct hna_local_entry *hna_local_entry, *hna_local_entry_tmp = NULL;
|
|
|
|
|
|
+ struct tt_local_entry *tt_local_entry, *tt_local_entry_tmp = NULL;
|
|
int index;
|
|
int index;
|
|
|
|
|
|
if (!hash)
|
|
if (!hash)
|
|
@@ -69,26 +69,26 @@ static struct hna_local_entry *hna_local_hash_find(struct bat_priv *bat_priv,
|
|
head = &hash->table[index];
|
|
head = &hash->table[index];
|
|
|
|
|
|
rcu_read_lock();
|
|
rcu_read_lock();
|
|
- hlist_for_each_entry_rcu(hna_local_entry, node, head, hash_entry) {
|
|
|
|
- if (!compare_eth(hna_local_entry, data))
|
|
|
|
|
|
+ hlist_for_each_entry_rcu(tt_local_entry, node, head, hash_entry) {
|
|
|
|
+ if (!compare_eth(tt_local_entry, data))
|
|
continue;
|
|
continue;
|
|
|
|
|
|
- hna_local_entry_tmp = hna_local_entry;
|
|
|
|
|
|
+ tt_local_entry_tmp = tt_local_entry;
|
|
break;
|
|
break;
|
|
}
|
|
}
|
|
rcu_read_unlock();
|
|
rcu_read_unlock();
|
|
|
|
|
|
- return hna_local_entry_tmp;
|
|
|
|
|
|
+ return tt_local_entry_tmp;
|
|
}
|
|
}
|
|
|
|
|
|
-static struct hna_global_entry *hna_global_hash_find(struct bat_priv *bat_priv,
|
|
|
|
|
|
+static struct tt_global_entry *tt_global_hash_find(struct bat_priv *bat_priv,
|
|
void *data)
|
|
void *data)
|
|
{
|
|
{
|
|
- struct hashtable_t *hash = bat_priv->hna_global_hash;
|
|
|
|
|
|
+ struct hashtable_t *hash = bat_priv->tt_global_hash;
|
|
struct hlist_head *head;
|
|
struct hlist_head *head;
|
|
struct hlist_node *node;
|
|
struct hlist_node *node;
|
|
- struct hna_global_entry *hna_global_entry;
|
|
|
|
- struct hna_global_entry *hna_global_entry_tmp = NULL;
|
|
|
|
|
|
+ struct tt_global_entry *tt_global_entry;
|
|
|
|
+ struct tt_global_entry *tt_global_entry_tmp = NULL;
|
|
int index;
|
|
int index;
|
|
|
|
|
|
if (!hash)
|
|
if (!hash)
|
|
@@ -98,125 +98,125 @@ static struct hna_global_entry *hna_global_hash_find(struct bat_priv *bat_priv,
|
|
head = &hash->table[index];
|
|
head = &hash->table[index];
|
|
|
|
|
|
rcu_read_lock();
|
|
rcu_read_lock();
|
|
- hlist_for_each_entry_rcu(hna_global_entry, node, head, hash_entry) {
|
|
|
|
- if (!compare_eth(hna_global_entry, data))
|
|
|
|
|
|
+ hlist_for_each_entry_rcu(tt_global_entry, node, head, hash_entry) {
|
|
|
|
+ if (!compare_eth(tt_global_entry, data))
|
|
continue;
|
|
continue;
|
|
|
|
|
|
- hna_global_entry_tmp = hna_global_entry;
|
|
|
|
|
|
+ tt_global_entry_tmp = tt_global_entry;
|
|
break;
|
|
break;
|
|
}
|
|
}
|
|
rcu_read_unlock();
|
|
rcu_read_unlock();
|
|
|
|
|
|
- return hna_global_entry_tmp;
|
|
|
|
|
|
+ return tt_global_entry_tmp;
|
|
}
|
|
}
|
|
|
|
|
|
-int hna_local_init(struct bat_priv *bat_priv)
|
|
|
|
|
|
+int tt_local_init(struct bat_priv *bat_priv)
|
|
{
|
|
{
|
|
- if (bat_priv->hna_local_hash)
|
|
|
|
|
|
+ if (bat_priv->tt_local_hash)
|
|
return 1;
|
|
return 1;
|
|
|
|
|
|
- bat_priv->hna_local_hash = hash_new(1024);
|
|
|
|
|
|
+ bat_priv->tt_local_hash = hash_new(1024);
|
|
|
|
|
|
- if (!bat_priv->hna_local_hash)
|
|
|
|
|
|
+ if (!bat_priv->tt_local_hash)
|
|
return 0;
|
|
return 0;
|
|
|
|
|
|
- atomic_set(&bat_priv->hna_local_changed, 0);
|
|
|
|
- hna_local_start_timer(bat_priv);
|
|
|
|
|
|
+ atomic_set(&bat_priv->tt_local_changed, 0);
|
|
|
|
+ tt_local_start_timer(bat_priv);
|
|
|
|
|
|
return 1;
|
|
return 1;
|
|
}
|
|
}
|
|
|
|
|
|
-void hna_local_add(struct net_device *soft_iface, uint8_t *addr)
|
|
|
|
|
|
+void tt_local_add(struct net_device *soft_iface, uint8_t *addr)
|
|
{
|
|
{
|
|
struct bat_priv *bat_priv = netdev_priv(soft_iface);
|
|
struct bat_priv *bat_priv = netdev_priv(soft_iface);
|
|
- struct hna_local_entry *hna_local_entry;
|
|
|
|
- struct hna_global_entry *hna_global_entry;
|
|
|
|
|
|
+ struct tt_local_entry *tt_local_entry;
|
|
|
|
+ struct tt_global_entry *tt_global_entry;
|
|
int required_bytes;
|
|
int required_bytes;
|
|
|
|
|
|
- spin_lock_bh(&bat_priv->hna_lhash_lock);
|
|
|
|
- hna_local_entry = hna_local_hash_find(bat_priv, addr);
|
|
|
|
- spin_unlock_bh(&bat_priv->hna_lhash_lock);
|
|
|
|
|
|
+ spin_lock_bh(&bat_priv->tt_lhash_lock);
|
|
|
|
+ tt_local_entry = tt_local_hash_find(bat_priv, addr);
|
|
|
|
+ spin_unlock_bh(&bat_priv->tt_lhash_lock);
|
|
|
|
|
|
- if (hna_local_entry) {
|
|
|
|
- hna_local_entry->last_seen = jiffies;
|
|
|
|
|
|
+ if (tt_local_entry) {
|
|
|
|
+ tt_local_entry->last_seen = jiffies;
|
|
return;
|
|
return;
|
|
}
|
|
}
|
|
|
|
|
|
/* only announce as many hosts as possible in the batman-packet and
|
|
/* only announce as many hosts as possible in the batman-packet and
|
|
- space in batman_packet->num_hna That also should give a limit to
|
|
|
|
|
|
+ space in batman_packet->num_tt That also should give a limit to
|
|
MAC-flooding. */
|
|
MAC-flooding. */
|
|
- required_bytes = (bat_priv->num_local_hna + 1) * ETH_ALEN;
|
|
|
|
|
|
+ required_bytes = (bat_priv->num_local_tt + 1) * ETH_ALEN;
|
|
required_bytes += BAT_PACKET_LEN;
|
|
required_bytes += BAT_PACKET_LEN;
|
|
|
|
|
|
if ((required_bytes > ETH_DATA_LEN) ||
|
|
if ((required_bytes > ETH_DATA_LEN) ||
|
|
(atomic_read(&bat_priv->aggregated_ogms) &&
|
|
(atomic_read(&bat_priv->aggregated_ogms) &&
|
|
required_bytes > MAX_AGGREGATION_BYTES) ||
|
|
required_bytes > MAX_AGGREGATION_BYTES) ||
|
|
- (bat_priv->num_local_hna + 1 > 255)) {
|
|
|
|
|
|
+ (bat_priv->num_local_tt + 1 > 255)) {
|
|
bat_dbg(DBG_ROUTES, bat_priv,
|
|
bat_dbg(DBG_ROUTES, bat_priv,
|
|
- "Can't add new local hna entry (%pM): "
|
|
|
|
- "number of local hna entries exceeds packet size\n",
|
|
|
|
|
|
+ "Can't add new local tt entry (%pM): "
|
|
|
|
+ "number of local tt entries exceeds packet size\n",
|
|
addr);
|
|
addr);
|
|
return;
|
|
return;
|
|
}
|
|
}
|
|
|
|
|
|
bat_dbg(DBG_ROUTES, bat_priv,
|
|
bat_dbg(DBG_ROUTES, bat_priv,
|
|
- "Creating new local hna entry: %pM\n", addr);
|
|
|
|
|
|
+ "Creating new local tt entry: %pM\n", addr);
|
|
|
|
|
|
- hna_local_entry = kmalloc(sizeof(struct hna_local_entry), GFP_ATOMIC);
|
|
|
|
- if (!hna_local_entry)
|
|
|
|
|
|
+ tt_local_entry = kmalloc(sizeof(struct tt_local_entry), GFP_ATOMIC);
|
|
|
|
+ if (!tt_local_entry)
|
|
return;
|
|
return;
|
|
|
|
|
|
- memcpy(hna_local_entry->addr, addr, ETH_ALEN);
|
|
|
|
- hna_local_entry->last_seen = jiffies;
|
|
|
|
|
|
+ memcpy(tt_local_entry->addr, addr, ETH_ALEN);
|
|
|
|
+ tt_local_entry->last_seen = jiffies;
|
|
|
|
|
|
/* the batman interface mac address should never be purged */
|
|
/* the batman interface mac address should never be purged */
|
|
if (compare_eth(addr, soft_iface->dev_addr))
|
|
if (compare_eth(addr, soft_iface->dev_addr))
|
|
- hna_local_entry->never_purge = 1;
|
|
|
|
|
|
+ tt_local_entry->never_purge = 1;
|
|
else
|
|
else
|
|
- hna_local_entry->never_purge = 0;
|
|
|
|
|
|
+ tt_local_entry->never_purge = 0;
|
|
|
|
|
|
- spin_lock_bh(&bat_priv->hna_lhash_lock);
|
|
|
|
|
|
+ spin_lock_bh(&bat_priv->tt_lhash_lock);
|
|
|
|
|
|
- hash_add(bat_priv->hna_local_hash, compare_lhna, choose_orig,
|
|
|
|
- hna_local_entry, &hna_local_entry->hash_entry);
|
|
|
|
- bat_priv->num_local_hna++;
|
|
|
|
- atomic_set(&bat_priv->hna_local_changed, 1);
|
|
|
|
|
|
+ hash_add(bat_priv->tt_local_hash, compare_ltt, choose_orig,
|
|
|
|
+ tt_local_entry, &tt_local_entry->hash_entry);
|
|
|
|
+ bat_priv->num_local_tt++;
|
|
|
|
+ atomic_set(&bat_priv->tt_local_changed, 1);
|
|
|
|
|
|
- spin_unlock_bh(&bat_priv->hna_lhash_lock);
|
|
|
|
|
|
+ spin_unlock_bh(&bat_priv->tt_lhash_lock);
|
|
|
|
|
|
/* remove address from global hash if present */
|
|
/* remove address from global hash if present */
|
|
- spin_lock_bh(&bat_priv->hna_ghash_lock);
|
|
|
|
|
|
+ spin_lock_bh(&bat_priv->tt_ghash_lock);
|
|
|
|
|
|
- hna_global_entry = hna_global_hash_find(bat_priv, addr);
|
|
|
|
|
|
+ tt_global_entry = tt_global_hash_find(bat_priv, addr);
|
|
|
|
|
|
- if (hna_global_entry)
|
|
|
|
- _hna_global_del_orig(bat_priv, hna_global_entry,
|
|
|
|
- "local hna received");
|
|
|
|
|
|
+ if (tt_global_entry)
|
|
|
|
+ _tt_global_del_orig(bat_priv, tt_global_entry,
|
|
|
|
+ "local tt received");
|
|
|
|
|
|
- spin_unlock_bh(&bat_priv->hna_ghash_lock);
|
|
|
|
|
|
+ spin_unlock_bh(&bat_priv->tt_ghash_lock);
|
|
}
|
|
}
|
|
|
|
|
|
-int hna_local_fill_buffer(struct bat_priv *bat_priv,
|
|
|
|
|
|
+int tt_local_fill_buffer(struct bat_priv *bat_priv,
|
|
unsigned char *buff, int buff_len)
|
|
unsigned char *buff, int buff_len)
|
|
{
|
|
{
|
|
- struct hashtable_t *hash = bat_priv->hna_local_hash;
|
|
|
|
- struct hna_local_entry *hna_local_entry;
|
|
|
|
|
|
+ struct hashtable_t *hash = bat_priv->tt_local_hash;
|
|
|
|
+ struct tt_local_entry *tt_local_entry;
|
|
struct hlist_node *node;
|
|
struct hlist_node *node;
|
|
struct hlist_head *head;
|
|
struct hlist_head *head;
|
|
int i, count = 0;
|
|
int i, count = 0;
|
|
|
|
|
|
- spin_lock_bh(&bat_priv->hna_lhash_lock);
|
|
|
|
|
|
+ spin_lock_bh(&bat_priv->tt_lhash_lock);
|
|
|
|
|
|
for (i = 0; i < hash->size; i++) {
|
|
for (i = 0; i < hash->size; i++) {
|
|
head = &hash->table[i];
|
|
head = &hash->table[i];
|
|
|
|
|
|
rcu_read_lock();
|
|
rcu_read_lock();
|
|
- hlist_for_each_entry_rcu(hna_local_entry, node,
|
|
|
|
|
|
+ hlist_for_each_entry_rcu(tt_local_entry, node,
|
|
head, hash_entry) {
|
|
head, hash_entry) {
|
|
if (buff_len < (count + 1) * ETH_ALEN)
|
|
if (buff_len < (count + 1) * ETH_ALEN)
|
|
break;
|
|
break;
|
|
|
|
|
|
- memcpy(buff + (count * ETH_ALEN), hna_local_entry->addr,
|
|
|
|
|
|
+ memcpy(buff + (count * ETH_ALEN), tt_local_entry->addr,
|
|
ETH_ALEN);
|
|
ETH_ALEN);
|
|
|
|
|
|
count++;
|
|
count++;
|
|
@@ -224,20 +224,20 @@ int hna_local_fill_buffer(struct bat_priv *bat_priv,
|
|
rcu_read_unlock();
|
|
rcu_read_unlock();
|
|
}
|
|
}
|
|
|
|
|
|
- /* if we did not get all new local hnas see you next time ;-) */
|
|
|
|
- if (count == bat_priv->num_local_hna)
|
|
|
|
- atomic_set(&bat_priv->hna_local_changed, 0);
|
|
|
|
|
|
+ /* if we did not get all new local tts see you next time ;-) */
|
|
|
|
+ if (count == bat_priv->num_local_tt)
|
|
|
|
+ atomic_set(&bat_priv->tt_local_changed, 0);
|
|
|
|
|
|
- spin_unlock_bh(&bat_priv->hna_lhash_lock);
|
|
|
|
|
|
+ spin_unlock_bh(&bat_priv->tt_lhash_lock);
|
|
return count;
|
|
return count;
|
|
}
|
|
}
|
|
|
|
|
|
-int hna_local_seq_print_text(struct seq_file *seq, void *offset)
|
|
|
|
|
|
+int tt_local_seq_print_text(struct seq_file *seq, void *offset)
|
|
{
|
|
{
|
|
struct net_device *net_dev = (struct net_device *)seq->private;
|
|
struct net_device *net_dev = (struct net_device *)seq->private;
|
|
struct bat_priv *bat_priv = netdev_priv(net_dev);
|
|
struct bat_priv *bat_priv = netdev_priv(net_dev);
|
|
- struct hashtable_t *hash = bat_priv->hna_local_hash;
|
|
|
|
- struct hna_local_entry *hna_local_entry;
|
|
|
|
|
|
+ struct hashtable_t *hash = bat_priv->tt_local_hash;
|
|
|
|
+ struct tt_local_entry *tt_local_entry;
|
|
struct hard_iface *primary_if;
|
|
struct hard_iface *primary_if;
|
|
struct hlist_node *node;
|
|
struct hlist_node *node;
|
|
struct hlist_head *head;
|
|
struct hlist_head *head;
|
|
@@ -261,10 +261,10 @@ int hna_local_seq_print_text(struct seq_file *seq, void *offset)
|
|
}
|
|
}
|
|
|
|
|
|
seq_printf(seq, "Locally retrieved addresses (from %s) "
|
|
seq_printf(seq, "Locally retrieved addresses (from %s) "
|
|
- "announced via HNA:\n",
|
|
|
|
|
|
+ "announced via TT:\n",
|
|
net_dev->name);
|
|
net_dev->name);
|
|
|
|
|
|
- spin_lock_bh(&bat_priv->hna_lhash_lock);
|
|
|
|
|
|
+ spin_lock_bh(&bat_priv->tt_lhash_lock);
|
|
|
|
|
|
buf_size = 1;
|
|
buf_size = 1;
|
|
/* Estimate length for: " * xx:xx:xx:xx:xx:xx\n" */
|
|
/* Estimate length for: " * xx:xx:xx:xx:xx:xx\n" */
|
|
@@ -279,7 +279,7 @@ int hna_local_seq_print_text(struct seq_file *seq, void *offset)
|
|
|
|
|
|
buff = kmalloc(buf_size, GFP_ATOMIC);
|
|
buff = kmalloc(buf_size, GFP_ATOMIC);
|
|
if (!buff) {
|
|
if (!buff) {
|
|
- spin_unlock_bh(&bat_priv->hna_lhash_lock);
|
|
|
|
|
|
+ spin_unlock_bh(&bat_priv->tt_lhash_lock);
|
|
ret = -ENOMEM;
|
|
ret = -ENOMEM;
|
|
goto out;
|
|
goto out;
|
|
}
|
|
}
|
|
@@ -291,15 +291,15 @@ int hna_local_seq_print_text(struct seq_file *seq, void *offset)
|
|
head = &hash->table[i];
|
|
head = &hash->table[i];
|
|
|
|
|
|
rcu_read_lock();
|
|
rcu_read_lock();
|
|
- hlist_for_each_entry_rcu(hna_local_entry, node,
|
|
|
|
|
|
+ hlist_for_each_entry_rcu(tt_local_entry, node,
|
|
head, hash_entry) {
|
|
head, hash_entry) {
|
|
pos += snprintf(buff + pos, 22, " * %pM\n",
|
|
pos += snprintf(buff + pos, 22, " * %pM\n",
|
|
- hna_local_entry->addr);
|
|
|
|
|
|
+ tt_local_entry->addr);
|
|
}
|
|
}
|
|
rcu_read_unlock();
|
|
rcu_read_unlock();
|
|
}
|
|
}
|
|
|
|
|
|
- spin_unlock_bh(&bat_priv->hna_lhash_lock);
|
|
|
|
|
|
+ spin_unlock_bh(&bat_priv->tt_lhash_lock);
|
|
|
|
|
|
seq_printf(seq, "%s", buff);
|
|
seq_printf(seq, "%s", buff);
|
|
kfree(buff);
|
|
kfree(buff);
|
|
@@ -309,180 +309,180 @@ out:
|
|
return ret;
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
|
|
-static void _hna_local_del(struct hlist_node *node, void *arg)
|
|
|
|
|
|
+static void _tt_local_del(struct hlist_node *node, void *arg)
|
|
{
|
|
{
|
|
struct bat_priv *bat_priv = (struct bat_priv *)arg;
|
|
struct bat_priv *bat_priv = (struct bat_priv *)arg;
|
|
- void *data = container_of(node, struct hna_local_entry, hash_entry);
|
|
|
|
|
|
+ void *data = container_of(node, struct tt_local_entry, hash_entry);
|
|
|
|
|
|
kfree(data);
|
|
kfree(data);
|
|
- bat_priv->num_local_hna--;
|
|
|
|
- atomic_set(&bat_priv->hna_local_changed, 1);
|
|
|
|
|
|
+ bat_priv->num_local_tt--;
|
|
|
|
+ atomic_set(&bat_priv->tt_local_changed, 1);
|
|
}
|
|
}
|
|
|
|
|
|
-static void hna_local_del(struct bat_priv *bat_priv,
|
|
|
|
- struct hna_local_entry *hna_local_entry,
|
|
|
|
|
|
+static void tt_local_del(struct bat_priv *bat_priv,
|
|
|
|
+ struct tt_local_entry *tt_local_entry,
|
|
char *message)
|
|
char *message)
|
|
{
|
|
{
|
|
- bat_dbg(DBG_ROUTES, bat_priv, "Deleting local hna entry (%pM): %s\n",
|
|
|
|
- hna_local_entry->addr, message);
|
|
|
|
|
|
+ bat_dbg(DBG_ROUTES, bat_priv, "Deleting local tt entry (%pM): %s\n",
|
|
|
|
+ tt_local_entry->addr, message);
|
|
|
|
|
|
- hash_remove(bat_priv->hna_local_hash, compare_lhna, choose_orig,
|
|
|
|
- hna_local_entry->addr);
|
|
|
|
- _hna_local_del(&hna_local_entry->hash_entry, bat_priv);
|
|
|
|
|
|
+ hash_remove(bat_priv->tt_local_hash, compare_ltt, choose_orig,
|
|
|
|
+ tt_local_entry->addr);
|
|
|
|
+ _tt_local_del(&tt_local_entry->hash_entry, bat_priv);
|
|
}
|
|
}
|
|
|
|
|
|
-void hna_local_remove(struct bat_priv *bat_priv,
|
|
|
|
|
|
+void tt_local_remove(struct bat_priv *bat_priv,
|
|
uint8_t *addr, char *message)
|
|
uint8_t *addr, char *message)
|
|
{
|
|
{
|
|
- struct hna_local_entry *hna_local_entry;
|
|
|
|
|
|
+ struct tt_local_entry *tt_local_entry;
|
|
|
|
|
|
- spin_lock_bh(&bat_priv->hna_lhash_lock);
|
|
|
|
|
|
+ spin_lock_bh(&bat_priv->tt_lhash_lock);
|
|
|
|
|
|
- hna_local_entry = hna_local_hash_find(bat_priv, addr);
|
|
|
|
|
|
+ tt_local_entry = tt_local_hash_find(bat_priv, addr);
|
|
|
|
|
|
- if (hna_local_entry)
|
|
|
|
- hna_local_del(bat_priv, hna_local_entry, message);
|
|
|
|
|
|
+ if (tt_local_entry)
|
|
|
|
+ tt_local_del(bat_priv, tt_local_entry, message);
|
|
|
|
|
|
- spin_unlock_bh(&bat_priv->hna_lhash_lock);
|
|
|
|
|
|
+ spin_unlock_bh(&bat_priv->tt_lhash_lock);
|
|
}
|
|
}
|
|
|
|
|
|
-static void hna_local_purge(struct work_struct *work)
|
|
|
|
|
|
+static void tt_local_purge(struct work_struct *work)
|
|
{
|
|
{
|
|
struct delayed_work *delayed_work =
|
|
struct delayed_work *delayed_work =
|
|
container_of(work, struct delayed_work, work);
|
|
container_of(work, struct delayed_work, work);
|
|
struct bat_priv *bat_priv =
|
|
struct bat_priv *bat_priv =
|
|
- container_of(delayed_work, struct bat_priv, hna_work);
|
|
|
|
- struct hashtable_t *hash = bat_priv->hna_local_hash;
|
|
|
|
- struct hna_local_entry *hna_local_entry;
|
|
|
|
|
|
+ container_of(delayed_work, struct bat_priv, tt_work);
|
|
|
|
+ struct hashtable_t *hash = bat_priv->tt_local_hash;
|
|
|
|
+ struct tt_local_entry *tt_local_entry;
|
|
struct hlist_node *node, *node_tmp;
|
|
struct hlist_node *node, *node_tmp;
|
|
struct hlist_head *head;
|
|
struct hlist_head *head;
|
|
unsigned long timeout;
|
|
unsigned long timeout;
|
|
int i;
|
|
int i;
|
|
|
|
|
|
- spin_lock_bh(&bat_priv->hna_lhash_lock);
|
|
|
|
|
|
+ spin_lock_bh(&bat_priv->tt_lhash_lock);
|
|
|
|
|
|
for (i = 0; i < hash->size; i++) {
|
|
for (i = 0; i < hash->size; i++) {
|
|
head = &hash->table[i];
|
|
head = &hash->table[i];
|
|
|
|
|
|
- hlist_for_each_entry_safe(hna_local_entry, node, node_tmp,
|
|
|
|
|
|
+ hlist_for_each_entry_safe(tt_local_entry, node, node_tmp,
|
|
head, hash_entry) {
|
|
head, hash_entry) {
|
|
- if (hna_local_entry->never_purge)
|
|
|
|
|
|
+ if (tt_local_entry->never_purge)
|
|
continue;
|
|
continue;
|
|
|
|
|
|
- timeout = hna_local_entry->last_seen;
|
|
|
|
- timeout += LOCAL_HNA_TIMEOUT * HZ;
|
|
|
|
|
|
+ timeout = tt_local_entry->last_seen;
|
|
|
|
+ timeout += TT_LOCAL_TIMEOUT * HZ;
|
|
|
|
|
|
if (time_before(jiffies, timeout))
|
|
if (time_before(jiffies, timeout))
|
|
continue;
|
|
continue;
|
|
|
|
|
|
- hna_local_del(bat_priv, hna_local_entry,
|
|
|
|
|
|
+ tt_local_del(bat_priv, tt_local_entry,
|
|
"address timed out");
|
|
"address timed out");
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
- spin_unlock_bh(&bat_priv->hna_lhash_lock);
|
|
|
|
- hna_local_start_timer(bat_priv);
|
|
|
|
|
|
+ spin_unlock_bh(&bat_priv->tt_lhash_lock);
|
|
|
|
+ tt_local_start_timer(bat_priv);
|
|
}
|
|
}
|
|
|
|
|
|
-void hna_local_free(struct bat_priv *bat_priv)
|
|
|
|
|
|
+void tt_local_free(struct bat_priv *bat_priv)
|
|
{
|
|
{
|
|
- if (!bat_priv->hna_local_hash)
|
|
|
|
|
|
+ if (!bat_priv->tt_local_hash)
|
|
return;
|
|
return;
|
|
|
|
|
|
- cancel_delayed_work_sync(&bat_priv->hna_work);
|
|
|
|
- hash_delete(bat_priv->hna_local_hash, _hna_local_del, bat_priv);
|
|
|
|
- bat_priv->hna_local_hash = NULL;
|
|
|
|
|
|
+ cancel_delayed_work_sync(&bat_priv->tt_work);
|
|
|
|
+ hash_delete(bat_priv->tt_local_hash, _tt_local_del, bat_priv);
|
|
|
|
+ bat_priv->tt_local_hash = NULL;
|
|
}
|
|
}
|
|
|
|
|
|
-int hna_global_init(struct bat_priv *bat_priv)
|
|
|
|
|
|
+int tt_global_init(struct bat_priv *bat_priv)
|
|
{
|
|
{
|
|
- if (bat_priv->hna_global_hash)
|
|
|
|
|
|
+ if (bat_priv->tt_global_hash)
|
|
return 1;
|
|
return 1;
|
|
|
|
|
|
- bat_priv->hna_global_hash = hash_new(1024);
|
|
|
|
|
|
+ bat_priv->tt_global_hash = hash_new(1024);
|
|
|
|
|
|
- if (!bat_priv->hna_global_hash)
|
|
|
|
|
|
+ if (!bat_priv->tt_global_hash)
|
|
return 0;
|
|
return 0;
|
|
|
|
|
|
return 1;
|
|
return 1;
|
|
}
|
|
}
|
|
|
|
|
|
-void hna_global_add_orig(struct bat_priv *bat_priv,
|
|
|
|
|
|
+void tt_global_add_orig(struct bat_priv *bat_priv,
|
|
struct orig_node *orig_node,
|
|
struct orig_node *orig_node,
|
|
- unsigned char *hna_buff, int hna_buff_len)
|
|
|
|
|
|
+ unsigned char *tt_buff, int tt_buff_len)
|
|
{
|
|
{
|
|
- struct hna_global_entry *hna_global_entry;
|
|
|
|
- struct hna_local_entry *hna_local_entry;
|
|
|
|
- int hna_buff_count = 0;
|
|
|
|
- unsigned char *hna_ptr;
|
|
|
|
|
|
+ struct tt_global_entry *tt_global_entry;
|
|
|
|
+ struct tt_local_entry *tt_local_entry;
|
|
|
|
+ int tt_buff_count = 0;
|
|
|
|
+ unsigned char *tt_ptr;
|
|
|
|
|
|
- while ((hna_buff_count + 1) * ETH_ALEN <= hna_buff_len) {
|
|
|
|
- spin_lock_bh(&bat_priv->hna_ghash_lock);
|
|
|
|
|
|
+ while ((tt_buff_count + 1) * ETH_ALEN <= tt_buff_len) {
|
|
|
|
+ spin_lock_bh(&bat_priv->tt_ghash_lock);
|
|
|
|
|
|
- hna_ptr = hna_buff + (hna_buff_count * ETH_ALEN);
|
|
|
|
- hna_global_entry = hna_global_hash_find(bat_priv, hna_ptr);
|
|
|
|
|
|
+ tt_ptr = tt_buff + (tt_buff_count * ETH_ALEN);
|
|
|
|
+ tt_global_entry = tt_global_hash_find(bat_priv, tt_ptr);
|
|
|
|
|
|
- if (!hna_global_entry) {
|
|
|
|
- spin_unlock_bh(&bat_priv->hna_ghash_lock);
|
|
|
|
|
|
+ if (!tt_global_entry) {
|
|
|
|
+ spin_unlock_bh(&bat_priv->tt_ghash_lock);
|
|
|
|
|
|
- hna_global_entry =
|
|
|
|
- kmalloc(sizeof(struct hna_global_entry),
|
|
|
|
|
|
+ tt_global_entry =
|
|
|
|
+ kmalloc(sizeof(struct tt_global_entry),
|
|
GFP_ATOMIC);
|
|
GFP_ATOMIC);
|
|
|
|
|
|
- if (!hna_global_entry)
|
|
|
|
|
|
+ if (!tt_global_entry)
|
|
break;
|
|
break;
|
|
|
|
|
|
- memcpy(hna_global_entry->addr, hna_ptr, ETH_ALEN);
|
|
|
|
|
|
+ memcpy(tt_global_entry->addr, tt_ptr, ETH_ALEN);
|
|
|
|
|
|
bat_dbg(DBG_ROUTES, bat_priv,
|
|
bat_dbg(DBG_ROUTES, bat_priv,
|
|
- "Creating new global hna entry: "
|
|
|
|
|
|
+ "Creating new global tt entry: "
|
|
"%pM (via %pM)\n",
|
|
"%pM (via %pM)\n",
|
|
- hna_global_entry->addr, orig_node->orig);
|
|
|
|
|
|
+ tt_global_entry->addr, orig_node->orig);
|
|
|
|
|
|
- spin_lock_bh(&bat_priv->hna_ghash_lock);
|
|
|
|
- hash_add(bat_priv->hna_global_hash, compare_ghna,
|
|
|
|
- choose_orig, hna_global_entry,
|
|
|
|
- &hna_global_entry->hash_entry);
|
|
|
|
|
|
+ spin_lock_bh(&bat_priv->tt_ghash_lock);
|
|
|
|
+ hash_add(bat_priv->tt_global_hash, compare_gtt,
|
|
|
|
+ choose_orig, tt_global_entry,
|
|
|
|
+ &tt_global_entry->hash_entry);
|
|
|
|
|
|
}
|
|
}
|
|
|
|
|
|
- hna_global_entry->orig_node = orig_node;
|
|
|
|
- spin_unlock_bh(&bat_priv->hna_ghash_lock);
|
|
|
|
|
|
+ tt_global_entry->orig_node = orig_node;
|
|
|
|
+ spin_unlock_bh(&bat_priv->tt_ghash_lock);
|
|
|
|
|
|
/* remove address from local hash if present */
|
|
/* remove address from local hash if present */
|
|
- spin_lock_bh(&bat_priv->hna_lhash_lock);
|
|
|
|
|
|
+ spin_lock_bh(&bat_priv->tt_lhash_lock);
|
|
|
|
|
|
- hna_ptr = hna_buff + (hna_buff_count * ETH_ALEN);
|
|
|
|
- hna_local_entry = hna_local_hash_find(bat_priv, hna_ptr);
|
|
|
|
|
|
+ tt_ptr = tt_buff + (tt_buff_count * ETH_ALEN);
|
|
|
|
+ tt_local_entry = tt_local_hash_find(bat_priv, tt_ptr);
|
|
|
|
|
|
- if (hna_local_entry)
|
|
|
|
- hna_local_del(bat_priv, hna_local_entry,
|
|
|
|
- "global hna received");
|
|
|
|
|
|
+ if (tt_local_entry)
|
|
|
|
+ tt_local_del(bat_priv, tt_local_entry,
|
|
|
|
+ "global tt received");
|
|
|
|
|
|
- spin_unlock_bh(&bat_priv->hna_lhash_lock);
|
|
|
|
|
|
+ spin_unlock_bh(&bat_priv->tt_lhash_lock);
|
|
|
|
|
|
- hna_buff_count++;
|
|
|
|
|
|
+ tt_buff_count++;
|
|
}
|
|
}
|
|
|
|
|
|
/* initialize, and overwrite if malloc succeeds */
|
|
/* initialize, and overwrite if malloc succeeds */
|
|
- orig_node->hna_buff = NULL;
|
|
|
|
- orig_node->hna_buff_len = 0;
|
|
|
|
-
|
|
|
|
- if (hna_buff_len > 0) {
|
|
|
|
- orig_node->hna_buff = kmalloc(hna_buff_len, GFP_ATOMIC);
|
|
|
|
- if (orig_node->hna_buff) {
|
|
|
|
- memcpy(orig_node->hna_buff, hna_buff, hna_buff_len);
|
|
|
|
- orig_node->hna_buff_len = hna_buff_len;
|
|
|
|
|
|
+ orig_node->tt_buff = NULL;
|
|
|
|
+ orig_node->tt_buff_len = 0;
|
|
|
|
+
|
|
|
|
+ if (tt_buff_len > 0) {
|
|
|
|
+ orig_node->tt_buff = kmalloc(tt_buff_len, GFP_ATOMIC);
|
|
|
|
+ if (orig_node->tt_buff) {
|
|
|
|
+ memcpy(orig_node->tt_buff, tt_buff, tt_buff_len);
|
|
|
|
+ orig_node->tt_buff_len = tt_buff_len;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
|
|
-int hna_global_seq_print_text(struct seq_file *seq, void *offset)
|
|
|
|
|
|
+int tt_global_seq_print_text(struct seq_file *seq, void *offset)
|
|
{
|
|
{
|
|
struct net_device *net_dev = (struct net_device *)seq->private;
|
|
struct net_device *net_dev = (struct net_device *)seq->private;
|
|
struct bat_priv *bat_priv = netdev_priv(net_dev);
|
|
struct bat_priv *bat_priv = netdev_priv(net_dev);
|
|
- struct hashtable_t *hash = bat_priv->hna_global_hash;
|
|
|
|
- struct hna_global_entry *hna_global_entry;
|
|
|
|
|
|
+ struct hashtable_t *hash = bat_priv->tt_global_hash;
|
|
|
|
+ struct tt_global_entry *tt_global_entry;
|
|
struct hard_iface *primary_if;
|
|
struct hard_iface *primary_if;
|
|
struct hlist_node *node;
|
|
struct hlist_node *node;
|
|
struct hlist_head *head;
|
|
struct hlist_head *head;
|
|
@@ -505,10 +505,11 @@ int hna_global_seq_print_text(struct seq_file *seq, void *offset)
|
|
goto out;
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
|
|
- seq_printf(seq, "Globally announced HNAs received via the mesh %s\n",
|
|
|
|
|
|
+ seq_printf(seq,
|
|
|
|
+ "Globally announced TT entries received via the mesh %s\n",
|
|
net_dev->name);
|
|
net_dev->name);
|
|
|
|
|
|
- spin_lock_bh(&bat_priv->hna_ghash_lock);
|
|
|
|
|
|
+ spin_lock_bh(&bat_priv->tt_ghash_lock);
|
|
|
|
|
|
buf_size = 1;
|
|
buf_size = 1;
|
|
/* Estimate length for: " * xx:xx:xx:xx:xx:xx via xx:xx:xx:xx:xx:xx\n"*/
|
|
/* Estimate length for: " * xx:xx:xx:xx:xx:xx via xx:xx:xx:xx:xx:xx\n"*/
|
|
@@ -523,7 +524,7 @@ int hna_global_seq_print_text(struct seq_file *seq, void *offset)
|
|
|
|
|
|
buff = kmalloc(buf_size, GFP_ATOMIC);
|
|
buff = kmalloc(buf_size, GFP_ATOMIC);
|
|
if (!buff) {
|
|
if (!buff) {
|
|
- spin_unlock_bh(&bat_priv->hna_ghash_lock);
|
|
|
|
|
|
+ spin_unlock_bh(&bat_priv->tt_ghash_lock);
|
|
ret = -ENOMEM;
|
|
ret = -ENOMEM;
|
|
goto out;
|
|
goto out;
|
|
}
|
|
}
|
|
@@ -534,17 +535,17 @@ int hna_global_seq_print_text(struct seq_file *seq, void *offset)
|
|
head = &hash->table[i];
|
|
head = &hash->table[i];
|
|
|
|
|
|
rcu_read_lock();
|
|
rcu_read_lock();
|
|
- hlist_for_each_entry_rcu(hna_global_entry, node,
|
|
|
|
|
|
+ hlist_for_each_entry_rcu(tt_global_entry, node,
|
|
head, hash_entry) {
|
|
head, hash_entry) {
|
|
pos += snprintf(buff + pos, 44,
|
|
pos += snprintf(buff + pos, 44,
|
|
" * %pM via %pM\n",
|
|
" * %pM via %pM\n",
|
|
- hna_global_entry->addr,
|
|
|
|
- hna_global_entry->orig_node->orig);
|
|
|
|
|
|
+ tt_global_entry->addr,
|
|
|
|
+ tt_global_entry->orig_node->orig);
|
|
}
|
|
}
|
|
rcu_read_unlock();
|
|
rcu_read_unlock();
|
|
}
|
|
}
|
|
|
|
|
|
- spin_unlock_bh(&bat_priv->hna_ghash_lock);
|
|
|
|
|
|
+ spin_unlock_bh(&bat_priv->tt_ghash_lock);
|
|
|
|
|
|
seq_printf(seq, "%s", buff);
|
|
seq_printf(seq, "%s", buff);
|
|
kfree(buff);
|
|
kfree(buff);
|
|
@@ -554,84 +555,84 @@ out:
|
|
return ret;
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
|
|
-static void _hna_global_del_orig(struct bat_priv *bat_priv,
|
|
|
|
- struct hna_global_entry *hna_global_entry,
|
|
|
|
|
|
+static void _tt_global_del_orig(struct bat_priv *bat_priv,
|
|
|
|
+ struct tt_global_entry *tt_global_entry,
|
|
char *message)
|
|
char *message)
|
|
{
|
|
{
|
|
bat_dbg(DBG_ROUTES, bat_priv,
|
|
bat_dbg(DBG_ROUTES, bat_priv,
|
|
- "Deleting global hna entry %pM (via %pM): %s\n",
|
|
|
|
- hna_global_entry->addr, hna_global_entry->orig_node->orig,
|
|
|
|
|
|
+ "Deleting global tt entry %pM (via %pM): %s\n",
|
|
|
|
+ tt_global_entry->addr, tt_global_entry->orig_node->orig,
|
|
message);
|
|
message);
|
|
|
|
|
|
- hash_remove(bat_priv->hna_global_hash, compare_ghna, choose_orig,
|
|
|
|
- hna_global_entry->addr);
|
|
|
|
- kfree(hna_global_entry);
|
|
|
|
|
|
+ hash_remove(bat_priv->tt_global_hash, compare_gtt, choose_orig,
|
|
|
|
+ tt_global_entry->addr);
|
|
|
|
+ kfree(tt_global_entry);
|
|
}
|
|
}
|
|
|
|
|
|
-void hna_global_del_orig(struct bat_priv *bat_priv,
|
|
|
|
|
|
+void tt_global_del_orig(struct bat_priv *bat_priv,
|
|
struct orig_node *orig_node, char *message)
|
|
struct orig_node *orig_node, char *message)
|
|
{
|
|
{
|
|
- struct hna_global_entry *hna_global_entry;
|
|
|
|
- int hna_buff_count = 0;
|
|
|
|
- unsigned char *hna_ptr;
|
|
|
|
|
|
+ struct tt_global_entry *tt_global_entry;
|
|
|
|
+ int tt_buff_count = 0;
|
|
|
|
+ unsigned char *tt_ptr;
|
|
|
|
|
|
- if (orig_node->hna_buff_len == 0)
|
|
|
|
|
|
+ if (orig_node->tt_buff_len == 0)
|
|
return;
|
|
return;
|
|
|
|
|
|
- spin_lock_bh(&bat_priv->hna_ghash_lock);
|
|
|
|
|
|
+ spin_lock_bh(&bat_priv->tt_ghash_lock);
|
|
|
|
|
|
- while ((hna_buff_count + 1) * ETH_ALEN <= orig_node->hna_buff_len) {
|
|
|
|
- hna_ptr = orig_node->hna_buff + (hna_buff_count * ETH_ALEN);
|
|
|
|
- hna_global_entry = hna_global_hash_find(bat_priv, hna_ptr);
|
|
|
|
|
|
+ while ((tt_buff_count + 1) * ETH_ALEN <= orig_node->tt_buff_len) {
|
|
|
|
+ tt_ptr = orig_node->tt_buff + (tt_buff_count * ETH_ALEN);
|
|
|
|
+ tt_global_entry = tt_global_hash_find(bat_priv, tt_ptr);
|
|
|
|
|
|
- if ((hna_global_entry) &&
|
|
|
|
- (hna_global_entry->orig_node == orig_node))
|
|
|
|
- _hna_global_del_orig(bat_priv, hna_global_entry,
|
|
|
|
|
|
+ if ((tt_global_entry) &&
|
|
|
|
+ (tt_global_entry->orig_node == orig_node))
|
|
|
|
+ _tt_global_del_orig(bat_priv, tt_global_entry,
|
|
message);
|
|
message);
|
|
|
|
|
|
- hna_buff_count++;
|
|
|
|
|
|
+ tt_buff_count++;
|
|
}
|
|
}
|
|
|
|
|
|
- spin_unlock_bh(&bat_priv->hna_ghash_lock);
|
|
|
|
|
|
+ spin_unlock_bh(&bat_priv->tt_ghash_lock);
|
|
|
|
|
|
- orig_node->hna_buff_len = 0;
|
|
|
|
- kfree(orig_node->hna_buff);
|
|
|
|
- orig_node->hna_buff = NULL;
|
|
|
|
|
|
+ orig_node->tt_buff_len = 0;
|
|
|
|
+ kfree(orig_node->tt_buff);
|
|
|
|
+ orig_node->tt_buff = NULL;
|
|
}
|
|
}
|
|
|
|
|
|
-static void hna_global_del(struct hlist_node *node, void *arg)
|
|
|
|
|
|
+static void tt_global_del(struct hlist_node *node, void *arg)
|
|
{
|
|
{
|
|
- void *data = container_of(node, struct hna_global_entry, hash_entry);
|
|
|
|
|
|
+ void *data = container_of(node, struct tt_global_entry, hash_entry);
|
|
|
|
|
|
kfree(data);
|
|
kfree(data);
|
|
}
|
|
}
|
|
|
|
|
|
-void hna_global_free(struct bat_priv *bat_priv)
|
|
|
|
|
|
+void tt_global_free(struct bat_priv *bat_priv)
|
|
{
|
|
{
|
|
- if (!bat_priv->hna_global_hash)
|
|
|
|
|
|
+ if (!bat_priv->tt_global_hash)
|
|
return;
|
|
return;
|
|
|
|
|
|
- hash_delete(bat_priv->hna_global_hash, hna_global_del, NULL);
|
|
|
|
- bat_priv->hna_global_hash = NULL;
|
|
|
|
|
|
+ hash_delete(bat_priv->tt_global_hash, tt_global_del, NULL);
|
|
|
|
+ bat_priv->tt_global_hash = NULL;
|
|
}
|
|
}
|
|
|
|
|
|
struct orig_node *transtable_search(struct bat_priv *bat_priv, uint8_t *addr)
|
|
struct orig_node *transtable_search(struct bat_priv *bat_priv, uint8_t *addr)
|
|
{
|
|
{
|
|
- struct hna_global_entry *hna_global_entry;
|
|
|
|
|
|
+ struct tt_global_entry *tt_global_entry;
|
|
struct orig_node *orig_node = NULL;
|
|
struct orig_node *orig_node = NULL;
|
|
|
|
|
|
- spin_lock_bh(&bat_priv->hna_ghash_lock);
|
|
|
|
- hna_global_entry = hna_global_hash_find(bat_priv, addr);
|
|
|
|
|
|
+ spin_lock_bh(&bat_priv->tt_ghash_lock);
|
|
|
|
+ tt_global_entry = tt_global_hash_find(bat_priv, addr);
|
|
|
|
|
|
- if (!hna_global_entry)
|
|
|
|
|
|
+ if (!tt_global_entry)
|
|
goto out;
|
|
goto out;
|
|
|
|
|
|
- if (!atomic_inc_not_zero(&hna_global_entry->orig_node->refcount))
|
|
|
|
|
|
+ if (!atomic_inc_not_zero(&tt_global_entry->orig_node->refcount))
|
|
goto out;
|
|
goto out;
|
|
|
|
|
|
- orig_node = hna_global_entry->orig_node;
|
|
|
|
|
|
+ orig_node = tt_global_entry->orig_node;
|
|
|
|
|
|
out:
|
|
out:
|
|
- spin_unlock_bh(&bat_priv->hna_ghash_lock);
|
|
|
|
|
|
+ spin_unlock_bh(&bat_priv->tt_ghash_lock);
|
|
return orig_node;
|
|
return orig_node;
|
|
}
|
|
}
|