summaryrefslogtreecommitdiff
path: root/net/mac80211
diff options
context:
space:
mode:
authorBob Copeland <me@bobcopeland.com>2016-03-19 05:11:30 +0300
committerJohannes Berg <johannes.berg@intel.com>2016-04-05 22:34:51 +0300
commit18b27ff7d2e232b0f07f2f51aa8052ff2a617908 (patch)
tree78c8d16a5ef5430a1eaf062b6ec296fbca033972 /net/mac80211
parent47a0489ce1e518f4936c7fedb93b3d2abd7ccd2e (diff)
downloadlinux-18b27ff7d2e232b0f07f2f51aa8052ff2a617908.tar.xz
mac80211: mesh: embed gates hlist head directly
Since we have converted the mesh path tables to rhashtable, we are no longer swapping out the entire mesh_pathtbl pointer with RCU. As a result, we no longer need indirection to the hlist head for the gates list and can simply embed it, saving a pair of pointer-sized allocations. Signed-off-by: Bob Copeland <me@bobcopeland.com> Signed-off-by: Johannes Berg <johannes.berg@intel.com>
Diffstat (limited to 'net/mac80211')
-rw-r--r--net/mac80211/mesh.h2
-rw-r--r--net/mac80211/mesh_pathtbl.c18
2 files changed, 5 insertions, 15 deletions
diff --git a/net/mac80211/mesh.h b/net/mac80211/mesh.h
index bc3f9a32b5a4..46b540a25d9d 100644
--- a/net/mac80211/mesh.h
+++ b/net/mac80211/mesh.h
@@ -134,7 +134,7 @@ struct mesh_path {
*/
struct mesh_table {
atomic_t entries; /* Up to MAX_MESH_NEIGHBOURS */
- struct hlist_head *known_gates;
+ struct hlist_head known_gates;
spinlock_t gates_lock;
struct rhashtable rhead;
diff --git a/net/mac80211/mesh_pathtbl.c b/net/mac80211/mesh_pathtbl.c
index 1c9412a29ca3..6db2ddfa0695 100644
--- a/net/mac80211/mesh_pathtbl.c
+++ b/net/mac80211/mesh_pathtbl.c
@@ -58,12 +58,7 @@ static struct mesh_table *mesh_table_alloc(void)
if (!newtbl)
return NULL;
- newtbl->known_gates = kzalloc(sizeof(struct hlist_head), GFP_ATOMIC);
- if (!newtbl->known_gates) {
- kfree(newtbl);
- return NULL;
- }
- INIT_HLIST_HEAD(newtbl->known_gates);
+ INIT_HLIST_HEAD(&newtbl->known_gates);
atomic_set(&newtbl->entries, 0);
spin_lock_init(&newtbl->gates_lock);
@@ -341,7 +336,7 @@ int mesh_path_add_gate(struct mesh_path *mpath)
mpath->sdata->u.mesh.num_gates++;
spin_lock(&tbl->gates_lock);
- hlist_add_head_rcu(&mpath->gate_list, tbl->known_gates);
+ hlist_add_head_rcu(&mpath->gate_list, &tbl->known_gates);
spin_unlock(&tbl->gates_lock);
spin_unlock_bh(&mpath->state_lock);
@@ -759,16 +754,11 @@ int mesh_path_send_to_gates(struct mesh_path *mpath)
struct mesh_path *from_mpath = mpath;
struct mesh_path *gate;
bool copy = false;
- struct hlist_head *known_gates;
tbl = sdata->u.mesh.mesh_paths;
- known_gates = tbl->known_gates;
-
- if (!known_gates)
- return -EHOSTUNREACH;
rcu_read_lock();
- hlist_for_each_entry_rcu(gate, known_gates, gate_list) {
+ hlist_for_each_entry_rcu(gate, &tbl->known_gates, gate_list) {
if (gate->flags & MESH_PATH_ACTIVE) {
mpath_dbg(sdata, "Forwarding to %pM\n", gate->dst);
mesh_path_move_to_queue(gate, from_mpath, copy);
@@ -781,7 +771,7 @@ int mesh_path_send_to_gates(struct mesh_path *mpath)
}
}
- hlist_for_each_entry_rcu(gate, known_gates, gate_list) {
+ hlist_for_each_entry_rcu(gate, &tbl->known_gates, gate_list) {
mpath_dbg(sdata, "Sending to %pM\n", gate->dst);
mesh_path_tx_pending(gate);
}