Add the rt linux 4.1.3-rt3 as base
[kvmfornfv.git] / kernel / drivers / net / wireless / brcm80211 / brcmfmac / flowring.c
diff --git a/kernel/drivers/net/wireless/brcm80211/brcmfmac/flowring.c b/kernel/drivers/net/wireless/brcm80211/brcmfmac/flowring.c
new file mode 100644 (file)
index 0000000..eb13253
--- /dev/null
@@ -0,0 +1,499 @@
+/* Copyright (c) 2014 Broadcom Corporation
+ *
+ * Permission to use, copy, modify, and/or distribute this software for any
+ * purpose with or without fee is hereby granted, provided that the above
+ * copyright notice and this permission notice appear in all copies.
+ *
+ * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
+ * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
+ * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY
+ * SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
+ * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION
+ * OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN
+ * CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
+ */
+
+
+#include <linux/types.h>
+#include <linux/netdevice.h>
+#include <linux/etherdevice.h>
+#include <brcmu_utils.h>
+
+#include "core.h"
+#include "debug.h"
+#include "bus.h"
+#include "proto.h"
+#include "flowring.h"
+#include "msgbuf.h"
+#include "common.h"
+
+
+#define BRCMF_FLOWRING_HIGH            1024
+#define BRCMF_FLOWRING_LOW             (BRCMF_FLOWRING_HIGH - 256)
+#define BRCMF_FLOWRING_INVALID_IFIDX   0xff
+
+#define BRCMF_FLOWRING_HASH_AP(da, fifo, ifidx) (da[5] + fifo + ifidx * 16)
+#define BRCMF_FLOWRING_HASH_STA(fifo, ifidx) (fifo + ifidx * 16)
+
+static const u8 brcmf_flowring_prio2fifo[] = {
+       1,
+       0,
+       0,
+       1,
+       2,
+       2,
+       3,
+       3
+};
+
+
+static bool
+brcmf_flowring_is_tdls_mac(struct brcmf_flowring *flow, u8 mac[ETH_ALEN])
+{
+       struct brcmf_flowring_tdls_entry *search;
+
+       search = flow->tdls_entry;
+
+       while (search) {
+               if (memcmp(search->mac, mac, ETH_ALEN) == 0)
+                       return true;
+               search = search->next;
+       }
+
+       return false;
+}
+
+
+u32 brcmf_flowring_lookup(struct brcmf_flowring *flow, u8 da[ETH_ALEN],
+                         u8 prio, u8 ifidx)
+{
+       struct brcmf_flowring_hash *hash;
+       u8 hash_idx;
+       u32 i;
+       bool found;
+       bool sta;
+       u8 fifo;
+       u8 *mac;
+
+       fifo = brcmf_flowring_prio2fifo[prio];
+       sta = (flow->addr_mode[ifidx] == ADDR_INDIRECT);
+       mac = da;
+       if ((!sta) && (is_multicast_ether_addr(da))) {
+               mac = (u8 *)ALLFFMAC;
+               fifo = 0;
+       }
+       if ((sta) && (flow->tdls_active) &&
+           (brcmf_flowring_is_tdls_mac(flow, da))) {
+               sta = false;
+       }
+       hash_idx =  sta ? BRCMF_FLOWRING_HASH_STA(fifo, ifidx) :
+                         BRCMF_FLOWRING_HASH_AP(mac, fifo, ifidx);
+       found = false;
+       hash = flow->hash;
+       for (i = 0; i < BRCMF_FLOWRING_HASHSIZE; i++) {
+               if ((sta || (memcmp(hash[hash_idx].mac, mac, ETH_ALEN) == 0)) &&
+                   (hash[hash_idx].fifo == fifo) &&
+                   (hash[hash_idx].ifidx == ifidx)) {
+                       found = true;
+                       break;
+               }
+               hash_idx++;
+       }
+       if (found)
+               return hash[hash_idx].flowid;
+
+       return BRCMF_FLOWRING_INVALID_ID;
+}
+
+
+u32 brcmf_flowring_create(struct brcmf_flowring *flow, u8 da[ETH_ALEN],
+                         u8 prio, u8 ifidx)
+{
+       struct brcmf_flowring_ring *ring;
+       struct brcmf_flowring_hash *hash;
+       u8 hash_idx;
+       u32 i;
+       bool found;
+       u8 fifo;
+       bool sta;
+       u8 *mac;
+
+       fifo = brcmf_flowring_prio2fifo[prio];
+       sta = (flow->addr_mode[ifidx] == ADDR_INDIRECT);
+       mac = da;
+       if ((!sta) && (is_multicast_ether_addr(da))) {
+               mac = (u8 *)ALLFFMAC;
+               fifo = 0;
+       }
+       if ((sta) && (flow->tdls_active) &&
+           (brcmf_flowring_is_tdls_mac(flow, da))) {
+               sta = false;
+       }
+       hash_idx =  sta ? BRCMF_FLOWRING_HASH_STA(fifo, ifidx) :
+                         BRCMF_FLOWRING_HASH_AP(mac, fifo, ifidx);
+       found = false;
+       hash = flow->hash;
+       for (i = 0; i < BRCMF_FLOWRING_HASHSIZE; i++) {
+               if ((hash[hash_idx].ifidx == BRCMF_FLOWRING_INVALID_IFIDX) &&
+                   (is_zero_ether_addr(hash[hash_idx].mac))) {
+                       found = true;
+                       break;
+               }
+               hash_idx++;
+       }
+       if (found) {
+               for (i = 0; i < flow->nrofrings; i++) {
+                       if (flow->rings[i] == NULL)
+                               break;
+               }
+               if (i == flow->nrofrings)
+                       return -ENOMEM;
+
+               ring = kzalloc(sizeof(*ring), GFP_ATOMIC);
+               if (!ring)
+                       return -ENOMEM;
+
+               memcpy(hash[hash_idx].mac, mac, ETH_ALEN);
+               hash[hash_idx].fifo = fifo;
+               hash[hash_idx].ifidx = ifidx;
+               hash[hash_idx].flowid = i;
+
+               ring->hash_id = hash_idx;
+               ring->status = RING_CLOSED;
+               skb_queue_head_init(&ring->skblist);
+               flow->rings[i] = ring;
+
+               return i;
+       }
+       return BRCMF_FLOWRING_INVALID_ID;
+}
+
+
+u8 brcmf_flowring_tid(struct brcmf_flowring *flow, u8 flowid)
+{
+       struct brcmf_flowring_ring *ring;
+
+       ring = flow->rings[flowid];
+
+       return flow->hash[ring->hash_id].fifo;
+}
+
+
+static void brcmf_flowring_block(struct brcmf_flowring *flow, u8 flowid,
+                                bool blocked)
+{
+       struct brcmf_flowring_ring *ring;
+       struct brcmf_bus *bus_if;
+       struct brcmf_pub *drvr;
+       struct brcmf_if *ifp;
+       bool currently_blocked;
+       int i;
+       u8 ifidx;
+       unsigned long flags;
+
+       spin_lock_irqsave(&flow->block_lock, flags);
+
+       ring = flow->rings[flowid];
+       ifidx = brcmf_flowring_ifidx_get(flow, flowid);
+
+       currently_blocked = false;
+       for (i = 0; i < flow->nrofrings; i++) {
+               if (flow->rings[i]) {
+                       ring = flow->rings[i];
+                       if ((ring->status == RING_OPEN) &&
+                           (brcmf_flowring_ifidx_get(flow, i) == ifidx)) {
+                               if (ring->blocked) {
+                                       currently_blocked = true;
+                                       break;
+                               }
+                       }
+               }
+       }
+       ring->blocked = blocked;
+       if (currently_blocked == blocked) {
+               spin_unlock_irqrestore(&flow->block_lock, flags);
+               return;
+       }
+
+       bus_if = dev_get_drvdata(flow->dev);
+       drvr = bus_if->drvr;
+       ifp = drvr->iflist[ifidx];
+       brcmf_txflowblock_if(ifp, BRCMF_NETIF_STOP_REASON_FLOW, blocked);
+
+       spin_unlock_irqrestore(&flow->block_lock, flags);
+}
+
+
+void brcmf_flowring_delete(struct brcmf_flowring *flow, u8 flowid)
+{
+       struct brcmf_flowring_ring *ring;
+       u8 hash_idx;
+       struct sk_buff *skb;
+
+       ring = flow->rings[flowid];
+       if (!ring)
+               return;
+       brcmf_flowring_block(flow, flowid, false);
+       hash_idx = ring->hash_id;
+       flow->hash[hash_idx].ifidx = BRCMF_FLOWRING_INVALID_IFIDX;
+       eth_zero_addr(flow->hash[hash_idx].mac);
+       flow->rings[flowid] = NULL;
+
+       skb = skb_dequeue(&ring->skblist);
+       while (skb) {
+               brcmu_pkt_buf_free_skb(skb);
+               skb = skb_dequeue(&ring->skblist);
+       }
+
+       kfree(ring);
+}
+
+
+void brcmf_flowring_enqueue(struct brcmf_flowring *flow, u8 flowid,
+                           struct sk_buff *skb)
+{
+       struct brcmf_flowring_ring *ring;
+
+       ring = flow->rings[flowid];
+
+       skb_queue_tail(&ring->skblist, skb);
+
+       if (!ring->blocked &&
+           (skb_queue_len(&ring->skblist) > BRCMF_FLOWRING_HIGH)) {
+               brcmf_flowring_block(flow, flowid, true);
+               brcmf_dbg(MSGBUF, "Flowcontrol: BLOCK for ring %d\n", flowid);
+               /* To prevent (work around) possible race condition, check
+                * queue len again. It is also possible to use locking to
+                * protect, but that is undesirable for every enqueue and
+                * dequeue. This simple check will solve a possible race
+                * condition if it occurs.
+                */
+               if (skb_queue_len(&ring->skblist) < BRCMF_FLOWRING_LOW)
+                       brcmf_flowring_block(flow, flowid, false);
+       }
+}
+
+
+struct sk_buff *brcmf_flowring_dequeue(struct brcmf_flowring *flow, u8 flowid)
+{
+       struct brcmf_flowring_ring *ring;
+       struct sk_buff *skb;
+
+       ring = flow->rings[flowid];
+       if (ring->status != RING_OPEN)
+               return NULL;
+
+       skb = skb_dequeue(&ring->skblist);
+
+       if (ring->blocked &&
+           (skb_queue_len(&ring->skblist) < BRCMF_FLOWRING_LOW)) {
+               brcmf_flowring_block(flow, flowid, false);
+               brcmf_dbg(MSGBUF, "Flowcontrol: OPEN for ring %d\n", flowid);
+       }
+
+       return skb;
+}
+
+
+void brcmf_flowring_reinsert(struct brcmf_flowring *flow, u8 flowid,
+                            struct sk_buff *skb)
+{
+       struct brcmf_flowring_ring *ring;
+
+       ring = flow->rings[flowid];
+
+       skb_queue_head(&ring->skblist, skb);
+}
+
+
+u32 brcmf_flowring_qlen(struct brcmf_flowring *flow, u8 flowid)
+{
+       struct brcmf_flowring_ring *ring;
+
+       ring = flow->rings[flowid];
+       if (!ring)
+               return 0;
+
+       if (ring->status != RING_OPEN)
+               return 0;
+
+       return skb_queue_len(&ring->skblist);
+}
+
+
+void brcmf_flowring_open(struct brcmf_flowring *flow, u8 flowid)
+{
+       struct brcmf_flowring_ring *ring;
+
+       ring = flow->rings[flowid];
+       if (!ring) {
+               brcmf_err("Ring NULL, for flowid %d\n", flowid);
+               return;
+       }
+
+       ring->status = RING_OPEN;
+}
+
+
+u8 brcmf_flowring_ifidx_get(struct brcmf_flowring *flow, u8 flowid)
+{
+       struct brcmf_flowring_ring *ring;
+       u8 hash_idx;
+
+       ring = flow->rings[flowid];
+       hash_idx = ring->hash_id;
+
+       return flow->hash[hash_idx].ifidx;
+}
+
+
+struct brcmf_flowring *brcmf_flowring_attach(struct device *dev, u16 nrofrings)
+{
+       struct brcmf_flowring *flow;
+       u32 i;
+
+       flow = kzalloc(sizeof(*flow), GFP_KERNEL);
+       if (flow) {
+               flow->dev = dev;
+               flow->nrofrings = nrofrings;
+               spin_lock_init(&flow->block_lock);
+               for (i = 0; i < ARRAY_SIZE(flow->addr_mode); i++)
+                       flow->addr_mode[i] = ADDR_INDIRECT;
+               for (i = 0; i < ARRAY_SIZE(flow->hash); i++)
+                       flow->hash[i].ifidx = BRCMF_FLOWRING_INVALID_IFIDX;
+               flow->rings = kcalloc(nrofrings, sizeof(*flow->rings),
+                                     GFP_KERNEL);
+               if (!flow->rings) {
+                       kfree(flow);
+                       flow = NULL;
+               }
+       }
+
+       return flow;
+}
+
+
+void brcmf_flowring_detach(struct brcmf_flowring *flow)
+{
+       struct brcmf_bus *bus_if = dev_get_drvdata(flow->dev);
+       struct brcmf_pub *drvr = bus_if->drvr;
+       struct brcmf_flowring_tdls_entry *search;
+       struct brcmf_flowring_tdls_entry *remove;
+       u8 flowid;
+
+       for (flowid = 0; flowid < flow->nrofrings; flowid++) {
+               if (flow->rings[flowid])
+                       brcmf_msgbuf_delete_flowring(drvr, flowid);
+       }
+
+       search = flow->tdls_entry;
+       while (search) {
+               remove = search;
+               search = search->next;
+               kfree(remove);
+       }
+       kfree(flow->rings);
+       kfree(flow);
+}
+
+
+void brcmf_flowring_configure_addr_mode(struct brcmf_flowring *flow, int ifidx,
+                                       enum proto_addr_mode addr_mode)
+{
+       struct brcmf_bus *bus_if = dev_get_drvdata(flow->dev);
+       struct brcmf_pub *drvr = bus_if->drvr;
+       u32 i;
+       u8 flowid;
+
+       if (flow->addr_mode[ifidx] != addr_mode) {
+               for (i = 0; i < ARRAY_SIZE(flow->hash); i++) {
+                       if (flow->hash[i].ifidx == ifidx) {
+                               flowid = flow->hash[i].flowid;
+                               if (flow->rings[flowid]->status != RING_OPEN)
+                                       continue;
+                               flow->rings[flowid]->status = RING_CLOSING;
+                               brcmf_msgbuf_delete_flowring(drvr, flowid);
+                       }
+               }
+               flow->addr_mode[ifidx] = addr_mode;
+       }
+}
+
+
+void brcmf_flowring_delete_peer(struct brcmf_flowring *flow, int ifidx,
+                               u8 peer[ETH_ALEN])
+{
+       struct brcmf_bus *bus_if = dev_get_drvdata(flow->dev);
+       struct brcmf_pub *drvr = bus_if->drvr;
+       struct brcmf_flowring_hash *hash;
+       struct brcmf_flowring_tdls_entry *prev;
+       struct brcmf_flowring_tdls_entry *search;
+       u32 i;
+       u8 flowid;
+       bool sta;
+
+       sta = (flow->addr_mode[ifidx] == ADDR_INDIRECT);
+
+       search = flow->tdls_entry;
+       prev = NULL;
+       while (search) {
+               if (memcmp(search->mac, peer, ETH_ALEN) == 0) {
+                       sta = false;
+                       break;
+               }
+               prev = search;
+               search = search->next;
+       }
+
+       hash = flow->hash;
+       for (i = 0; i < BRCMF_FLOWRING_HASHSIZE; i++) {
+               if ((sta || (memcmp(hash[i].mac, peer, ETH_ALEN) == 0)) &&
+                   (hash[i].ifidx == ifidx)) {
+                       flowid = flow->hash[i].flowid;
+                       if (flow->rings[flowid]->status == RING_OPEN) {
+                               flow->rings[flowid]->status = RING_CLOSING;
+                               brcmf_msgbuf_delete_flowring(drvr, flowid);
+                       }
+               }
+       }
+
+       if (search) {
+               if (prev)
+                       prev->next = search->next;
+               else
+                       flow->tdls_entry = search->next;
+               kfree(search);
+               if (flow->tdls_entry == NULL)
+                       flow->tdls_active = false;
+       }
+}
+
+
+void brcmf_flowring_add_tdls_peer(struct brcmf_flowring *flow, int ifidx,
+                                 u8 peer[ETH_ALEN])
+{
+       struct brcmf_flowring_tdls_entry *tdls_entry;
+       struct brcmf_flowring_tdls_entry *search;
+
+       tdls_entry = kzalloc(sizeof(*tdls_entry), GFP_ATOMIC);
+       if (tdls_entry == NULL)
+               return;
+
+       memcpy(tdls_entry->mac, peer, ETH_ALEN);
+       tdls_entry->next = NULL;
+       if (flow->tdls_entry == NULL) {
+               flow->tdls_entry = tdls_entry;
+       } else {
+               search = flow->tdls_entry;
+               if (memcmp(search->mac, peer, ETH_ALEN) == 0)
+                       return;
+               while (search->next) {
+                       search = search->next;
+                       if (memcmp(search->mac, peer, ETH_ALEN) == 0)
+                               return;
+               }
+               search->next = tdls_entry;
+       }
+
+       flow->tdls_active = true;
+}