diff options
Diffstat (limited to 'package/madwifi/patches/300-napi_polling.patch')
-rw-r--r-- | package/madwifi/patches/300-napi_polling.patch | 536 |
1 files changed, 0 insertions, 536 deletions
diff --git a/package/madwifi/patches/300-napi_polling.patch b/package/madwifi/patches/300-napi_polling.patch deleted file mode 100644 index bde768494..000000000 --- a/package/madwifi/patches/300-napi_polling.patch +++ /dev/null @@ -1,536 +0,0 @@ ---- a/ath/if_ath.c -+++ b/ath/if_ath.c -@@ -184,7 +184,11 @@ static void ath_recv_mgmt(struct ieee802 - struct sk_buff *, int, int, u_int64_t); - static void ath_setdefantenna(struct ath_softc *, u_int); - static struct ath_txq *ath_txq_setup(struct ath_softc *, int, int); --static void ath_rx_tasklet(TQUEUE_ARG); -+#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) -+static int ath_rx_poll(struct napi_struct *napi, int budget); -+#else -+static int ath_rx_poll(struct net_device *dev, int *budget); -+#endif - static int ath_hardstart(struct sk_buff *, struct net_device *); - static int ath_mgtstart(struct ieee80211com *, struct sk_buff *); - #ifdef ATH_SUPERG_COMP -@@ -376,6 +380,9 @@ static u_int32_t ath_set_clamped_maxtxpo - u_int32_t new_clamped_maxtxpower); - static u_int32_t ath_get_real_maxtxpower(struct ath_softc *sc); - -+static void ath_poll_disable(struct net_device *dev); -+static void ath_poll_enable(struct net_device *dev); -+ - /* calibrate every 30 secs in steady state but check every second at first. */ - static int ath_calinterval = ATH_SHORT_CALINTERVAL; - static int ath_countrycode = CTRY_DEFAULT; /* country code */ -@@ -547,7 +554,6 @@ ath_attach(u_int16_t devid, struct net_d - - atomic_set(&sc->sc_txbuf_counter, 0); - -- ATH_INIT_TQUEUE(&sc->sc_rxtq, ath_rx_tasklet, dev); - ATH_INIT_TQUEUE(&sc->sc_txtq, ath_tx_tasklet, dev); - ATH_INIT_TQUEUE(&sc->sc_bmisstq, ath_bmiss_tasklet, dev); - ATH_INIT_TQUEUE(&sc->sc_bstucktq, ath_bstuck_tasklet, dev); -@@ -821,6 +827,12 @@ ath_attach(u_int16_t devid, struct net_d - dev->set_mac_address = ath_set_mac_address; - dev->change_mtu = ath_change_mtu; - dev->tx_queue_len = ATH_TXBUF - ATH_TXBUF_MGT_RESERVED; -+#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) -+ netif_napi_add(dev, &sc->sc_napi, ath_rx_poll, 64); -+#else -+ dev->poll = ath_rx_poll; -+ dev->weight = 64; -+#endif - #ifdef USE_HEADERLEN_RESV - dev->hard_header_len += sizeof(struct ieee80211_qosframe) + - sizeof(struct llc) + -@@ -2220,6 +2232,7 @@ ath_intr(int irq, void *dev_id, struct p - (status & HAL_INT_GLOBAL) ? " HAL_INT_GLOBAL" : "" - ); - -+ sc->sc_isr = status; - status &= sc->sc_imask; /* discard unasked for bits */ - /* As soon as we know we have a real interrupt we intend to service, - * we will check to see if we need an initial hardware TSF reading. -@@ -2277,7 +2290,21 @@ ath_intr(int irq, void *dev_id, struct p - } - if (status & (HAL_INT_RX | HAL_INT_RXPHY)) { - ath_uapsd_processtriggers(sc, hw_tsf); -- ATH_SCHEDULE_TQUEUE(&sc->sc_rxtq, &needmark); -+ sc->sc_isr &= ~HAL_INT_RX; -+#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) -+ if (netif_rx_schedule_prep(dev, &sc->sc_napi)) -+#else -+ if (netif_rx_schedule_prep(dev)) -+#endif -+ { -+ sc->sc_imask &= ~HAL_INT_RX; -+ ath_hal_intrset(ah, sc->sc_imask); -+#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) -+ __netif_rx_schedule(dev, &sc->sc_napi); -+#else -+ __netif_rx_schedule(dev); -+#endif -+ } - } - if (status & HAL_INT_TX) { - #ifdef ATH_SUPERG_DYNTURBO -@@ -2303,6 +2330,11 @@ ath_intr(int irq, void *dev_id, struct p - } - } - #endif -+ /* disable transmit interrupt */ -+ sc->sc_isr &= ~HAL_INT_TX; -+ ath_hal_intrset(ah, sc->sc_imask & ~HAL_INT_TX); -+ sc->sc_imask &= ~HAL_INT_TX; -+ - ATH_SCHEDULE_TQUEUE(&sc->sc_txtq, &needmark); - } - if (status & HAL_INT_BMISS) { -@@ -2515,6 +2547,7 @@ ath_init(struct net_device *dev) - if (sc->sc_tx99 != NULL) - sc->sc_tx99->start(sc->sc_tx99); - #endif -+ ath_poll_enable(dev); - - done: - ATH_UNLOCK(sc); -@@ -2555,6 +2588,9 @@ ath_stop_locked(struct net_device *dev) - if (sc->sc_tx99 != NULL) - sc->sc_tx99->stop(sc->sc_tx99); - #endif -+#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) -+ ath_poll_disable(dev); -+#endif - netif_stop_queue(dev); /* XXX re-enabled by ath_newstate */ - dev->flags &= ~IFF_RUNNING; /* NB: avoid recursion */ - ieee80211_stop_running(ic); /* stop all VAPs */ -@@ -4013,12 +4049,47 @@ ath_key_set(struct ieee80211vap *vap, co - return ath_keyset(sc, k, mac, vap->iv_bss); - } - -+static void ath_poll_disable(struct net_device *dev) -+{ -+ struct ath_softc *sc = dev->priv; -+ -+ /* -+ * XXX Using in_softirq is not right since we might -+ * be called from other soft irq contexts than -+ * ath_rx_poll -+ */ -+ if (!in_softirq()) { -+#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) -+ napi_disable(&sc->sc_napi); -+#else -+ netif_poll_disable(dev); -+#endif -+ } -+} -+ -+static void ath_poll_enable(struct net_device *dev) -+{ -+ struct ath_softc *sc = dev->priv; -+ -+ /* NB: see above */ -+ if (!in_softirq()) { -+#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) -+ napi_enable(&sc->sc_napi); -+#else -+ netif_poll_enable(dev); -+#endif -+ } -+} -+ -+ - /* - * Block/unblock tx+rx processing while a key change is done. - * We assume the caller serializes key management operations - * so we only need to worry about synchronization with other - * uses that originate in the driver. - */ -+#define IS_UP(_dev) \ -+ (((_dev)->flags & (IFF_RUNNING|IFF_UP)) == (IFF_RUNNING|IFF_UP)) - static void - ath_key_update_begin(struct ieee80211vap *vap) - { -@@ -4032,14 +4103,9 @@ ath_key_update_begin(struct ieee80211vap - * When called from the rx tasklet we cannot use - * tasklet_disable because it will block waiting - * for us to complete execution. -- * -- * XXX Using in_softirq is not right since we might -- * be called from other soft irq contexts than -- * ath_rx_tasklet. - */ -- if (!in_softirq()) -- tasklet_disable(&sc->sc_rxtq); -- netif_stop_queue(dev); -+ if (IS_UP(vap->iv_dev)) -+ netif_stop_queue(dev); - } - - static void -@@ -4051,9 +4117,9 @@ ath_key_update_end(struct ieee80211vap * - #endif - - DPRINTF(sc, ATH_DEBUG_KEYCACHE, "End\n"); -- netif_wake_queue(dev); -- if (!in_softirq()) /* NB: see above */ -- tasklet_enable(&sc->sc_rxtq); -+ -+ if (IS_UP(vap->iv_dev)) -+ netif_wake_queue(dev); - } - - /* -@@ -6360,15 +6426,25 @@ ath_setdefantenna(struct ath_softc *sc, - sc->sc_rxotherant = 0; - } - --static void --ath_rx_tasklet(TQUEUE_ARG data) -+static int -+#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) -+ath_rx_poll(struct napi_struct *napi, int budget) -+#else -+ath_rx_poll(struct net_device *dev, int *budget) -+#endif - { - #define PA2DESC(_sc, _pa) \ - ((struct ath_desc *)((caddr_t)(_sc)->sc_rxdma.dd_desc + \ - ((_pa) - (_sc)->sc_rxdma.dd_desc_paddr))) -- struct net_device *dev = (struct net_device *)data; -- struct ath_buf *bf; -+#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) -+ struct ath_softc *sc = container_of(napi, struct ath_softc, sc_napi); -+ struct net_device *dev = sc->sc_dev; -+ u_int rx_limit = budget; -+#else - struct ath_softc *sc = dev->priv; -+ u_int rx_limit = min(dev->quota, *budget); -+#endif -+ struct ath_buf *bf; - struct ieee80211com *ic = &sc->sc_ic; - struct ath_hal *ah = sc ? sc->sc_ah : NULL; - struct ath_desc *ds; -@@ -6378,8 +6454,10 @@ ath_rx_tasklet(TQUEUE_ARG data) - unsigned int len; - int type; - u_int phyerr; -+ u_int processed = 0, early_stop = 0; - - DPRINTF(sc, ATH_DEBUG_RX_PROC, "invoked\n"); -+process_rx_again: - do { - bf = STAILQ_FIRST(&sc->sc_rxbuf); - if (bf == NULL) { /* XXX ??? can this happen */ -@@ -6403,6 +6481,15 @@ ath_rx_tasklet(TQUEUE_ARG data) - /* NB: never process the self-linked entry at the end */ - break; - } -+ -+ if (rx_limit-- < 2) { -+ early_stop = 1; -+ break; -+ } -+#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) -+ processed++; -+#endif -+ - skb = bf->bf_skb; - if (skb == NULL) { - EPRINTF(sc, "Dropping; buffer contains NULL skbuff.\n"); -@@ -6450,6 +6537,7 @@ ath_rx_tasklet(TQUEUE_ARG data) - sc->sc_stats.ast_rx_phyerr++; - phyerr = rs->rs_phyerr & 0x1f; - sc->sc_stats.ast_rx_phy[phyerr]++; -+ goto rx_next; - } - if (rs->rs_status & HAL_RXERR_DECRYPT) { - /* -@@ -6645,9 +6733,39 @@ rx_next: - STAILQ_INSERT_TAIL(&sc->sc_rxbuf, bf, bf_list); - ATH_RXBUF_UNLOCK_IRQ(sc); - } while (ath_rxbuf_init(sc, bf) == 0); -+ if (!early_stop) { -+ unsigned long flags; -+ /* Check if more data is received while we were -+ * processing the descriptor chain. -+ */ -+ local_irq_save(flags); -+ if (sc->sc_isr & HAL_INT_RX) { -+ u_int64_t hw_tsf = ath_hal_gettsf64(ah); -+ sc->sc_isr &= ~HAL_INT_RX; -+ local_irq_restore(flags); -+ ath_uapsd_processtriggers(sc, hw_tsf); -+ goto process_rx_again; -+ } -+ local_irq_restore(flags); -+ } -+ -+#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) -+ netif_rx_complete(dev, napi); -+#else -+ netif_rx_complete(dev); -+ *budget -= processed; -+ dev->quota -= processed; -+#endif -+ sc->sc_imask |= HAL_INT_RX; -+ ath_hal_intrset(ah, sc->sc_imask); - - /* rx signal state monitoring */ - ath_hal_rxmonitor(ah, &sc->sc_halstats, &sc->sc_curchan); -+#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) -+ return processed; -+#else -+ return early_stop; -+#endif - #undef PA2DESC - } - -@@ -8298,12 +8416,24 @@ ath_tx_tasklet_q0(TQUEUE_ARG data) - { - struct net_device *dev = (struct net_device *)data; - struct ath_softc *sc = dev->priv; -+ unsigned long flags; - -+process_tx_again: - if (txqactive(sc->sc_ah, 0)) - ath_tx_processq(sc, &sc->sc_txq[0]); - if (txqactive(sc->sc_ah, sc->sc_cabq->axq_qnum)) - ath_tx_processq(sc, sc->sc_cabq); - -+ local_irq_save(flags); -+ if (sc->sc_isr & HAL_INT_TX) { -+ sc->sc_isr &= ~HAL_INT_TX; -+ local_irq_restore(flags); -+ goto process_tx_again; -+ } -+ sc->sc_imask |= HAL_INT_TX; -+ ath_hal_intrset(sc->sc_ah, sc->sc_imask); -+ local_irq_restore(flags); -+ - netif_wake_queue(dev); - - if (sc->sc_softled) -@@ -8319,7 +8449,9 @@ ath_tx_tasklet_q0123(TQUEUE_ARG data) - { - struct net_device *dev = (struct net_device *)data; - struct ath_softc *sc = dev->priv; -+ unsigned long flags; - -+process_tx_again: - /* - * Process each active queue. - */ -@@ -8340,6 +8472,16 @@ ath_tx_tasklet_q0123(TQUEUE_ARG data) - if (sc->sc_uapsdq && txqactive(sc->sc_ah, sc->sc_uapsdq->axq_qnum)) - ath_tx_processq(sc, sc->sc_uapsdq); - -+ local_irq_save(flags); -+ if (sc->sc_isr & HAL_INT_TX) { -+ sc->sc_isr &= ~HAL_INT_TX; -+ local_irq_restore(flags); -+ goto process_tx_again; -+ } -+ sc->sc_imask |= HAL_INT_TX; -+ ath_hal_intrset(sc->sc_ah, sc->sc_imask); -+ local_irq_restore(flags); -+ - netif_wake_queue(dev); - - if (sc->sc_softled) -@@ -8355,13 +8497,25 @@ ath_tx_tasklet(TQUEUE_ARG data) - struct net_device *dev = (struct net_device *)data; - struct ath_softc *sc = dev->priv; - unsigned int i; -+ unsigned long flags; - - /* Process each active queue. This includes sc_cabq, sc_xrtq and - * sc_uapsdq */ -+process_tx_again: - for (i = 0; i < HAL_NUM_TX_QUEUES; i++) - if (ATH_TXQ_SETUP(sc, i) && txqactive(sc->sc_ah, i)) - ath_tx_processq(sc, &sc->sc_txq[i]); - -+ local_irq_save(flags); -+ if (sc->sc_isr & HAL_INT_TX) { -+ sc->sc_isr &= ~HAL_INT_TX; -+ local_irq_restore(flags); -+ goto process_tx_again; -+ } -+ sc->sc_imask |= HAL_INT_TX; -+ ath_hal_intrset(sc->sc_ah, sc->sc_imask); -+ local_irq_restore(flags); -+ - netif_wake_queue(dev); - - if (sc->sc_softled) -@@ -10296,9 +10450,9 @@ ath_change_mtu(struct net_device *dev, i - dev->mtu = mtu; - if ((dev->flags & IFF_RUNNING) && !sc->sc_invalid) { - /* NB: the rx buffers may need to be reallocated */ -- tasklet_disable(&sc->sc_rxtq); -+ ath_poll_disable(dev); - error = ath_reset(dev); -- tasklet_enable(&sc->sc_rxtq); -+ ath_poll_enable(dev); - } - ATH_UNLOCK(sc); - ---- a/ath/if_athvar.h -+++ b/ath/if_athvar.h -@@ -53,6 +53,10 @@ - # include <asm/bitops.h> - #endif - -+#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,0) -+#define irqs_disabled() 0 -+#endif -+ - /* - * Deduce if tasklets are available. If not then - * fall back to using the immediate work queue. -@@ -616,6 +620,9 @@ struct ath_rp { - struct ath_softc { - struct ieee80211com sc_ic; /* NB: must be first */ - struct net_device *sc_dev; -+#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) -+ struct napi_struct sc_napi; -+#endif - void __iomem *sc_iobase; /* address of the device */ - struct semaphore sc_lock; /* dev-level lock */ - struct net_device_stats sc_devstats; /* device statistics */ -@@ -730,7 +737,6 @@ struct ath_softc { - struct ath_buf *sc_rxbufcur; /* current rx buffer */ - u_int32_t *sc_rxlink; /* link ptr in last RX desc */ - spinlock_t sc_rxbuflock; -- struct ATH_TQ_STRUCT sc_rxtq; /* rx intr tasklet */ - struct ATH_TQ_STRUCT sc_rxorntq; /* rxorn intr tasklet */ - u_int8_t sc_defant; /* current default antenna */ - u_int8_t sc_rxotherant; /* RXs on non-default antenna */ -@@ -745,6 +751,7 @@ struct ath_softc { - u_int sc_txintrperiod; /* tx interrupt batching */ - struct ath_txq sc_txq[HAL_NUM_TX_QUEUES]; - struct ath_txq *sc_ac2q[WME_NUM_AC]; /* WME AC -> h/w qnum */ -+ HAL_INT sc_isr; /* unmasked ISR state */ - struct ATH_TQ_STRUCT sc_txtq; /* tx intr tasklet */ - u_int8_t sc_grppoll_str[GRPPOLL_RATE_STR_LEN]; - struct ath_descdma sc_bdma; /* beacon descriptors */ -@@ -858,6 +865,8 @@ typedef void (*ath_callback) (struct ath - #define ATH_TXBUF_LOCK_CHECK(_sc) - #endif - -+#define ATH_DISABLE_INTR local_irq_disable -+#define ATH_ENABLE_INTR local_irq_enable - - #define ATH_RXBUF_LOCK_INIT(_sc) spin_lock_init(&(_sc)->sc_rxbuflock) - #define ATH_RXBUF_LOCK_DESTROY(_sc) ---- a/net80211/ieee80211_input.c -+++ b/net80211/ieee80211_input.c -@@ -1198,7 +1198,7 @@ ieee80211_deliver_data(struct ieee80211_ - /* attach vlan tag */ - struct ieee80211_node *ni_tmp = SKB_CB(skb)->ni; - if (vlan_hwaccel_receive_skb(skb, vap->iv_vlgrp, ni->ni_vlan) == NET_RX_DROP) { -- /* If netif_rx dropped the packet because -+ /* If netif_receive_skb dropped the packet because - * device was too busy */ - if (ni_tmp != NULL) { - /* node reference was leaked */ -@@ -1209,8 +1209,8 @@ ieee80211_deliver_data(struct ieee80211_ - skb = NULL; /* SKB is no longer ours */ - } else { - struct ieee80211_node *ni_tmp = SKB_CB(skb)->ni; -- if (netif_rx(skb) == NET_RX_DROP) { -- /* If netif_rx dropped the packet because -+ if (netif_receive_skb(skb) == NET_RX_DROP) { -+ /* If netif_receive_skb dropped the packet because - * device was too busy */ - if (ni_tmp != NULL) { - /* node reference was leaked */ -@@ -2322,8 +2322,8 @@ forward_mgmt_to_app(struct ieee80211vap - skb1->protocol = __constant_htons(0x0019); /* ETH_P_80211_RAW */ - - ni_tmp = SKB_CB(skb1)->ni; -- if (netif_rx(skb1) == NET_RX_DROP) { -- /* If netif_rx dropped the packet because -+ if (netif_receive_skb(skb1) == NET_RX_DROP) { -+ /* If netif_receive_skb dropped the packet because - * device was too busy */ - if (ni_tmp != NULL) { - /* node reference was leaked */ ---- a/net80211/ieee80211_monitor.c -+++ b/net80211/ieee80211_monitor.c -@@ -584,8 +584,8 @@ ieee80211_input_monitor(struct ieee80211 - skb1->protocol = - __constant_htons(0x0019); /* ETH_P_80211_RAW */ - -- if (netif_rx(skb1) == NET_RX_DROP) { -- /* If netif_rx dropped the packet because -+ if (netif_receive_skb(skb1) == NET_RX_DROP) { -+ /* If netif_receive_skb dropped the packet because - * device was too busy, reclaim the ref. in - * the skb. */ - if (SKB_CB(skb1)->ni != NULL) ---- a/net80211/ieee80211_skb.c -+++ b/net80211/ieee80211_skb.c -@@ -73,7 +73,7 @@ - #undef dev_queue_xmit - #undef kfree_skb - #undef kfree_skb_fast --#undef netif_rx -+#undef netif_receive_skb - #undef pskb_copy - #undef skb_clone - #undef skb_copy -@@ -638,8 +638,8 @@ int vlan_hwaccel_receive_skb_debug(stru - grp, vlan_tag); - } - --int netif_rx_debug(struct sk_buff *skb, const char* func, int line) { -- return netif_rx(untrack_skb(skb, 0, func, line, __func__, __LINE__)); -+int netif_receive_skb_debug(struct sk_buff *skb, const char* func, int line) { -+ return netif_receive_skb(untrack_skb(skb, 0, func, line, __func__, __LINE__)); - } - - struct sk_buff * alloc_skb_debug(unsigned int length, gfp_t gfp_mask, -@@ -760,7 +760,7 @@ struct sk_buff * skb_copy_expand_debug(c - } - - EXPORT_SYMBOL(vlan_hwaccel_receive_skb_debug); --EXPORT_SYMBOL(netif_rx_debug); -+EXPORT_SYMBOL(netif_receive_skb_debug); - EXPORT_SYMBOL(alloc_skb_debug); - EXPORT_SYMBOL(dev_alloc_skb_debug); - EXPORT_SYMBOL(skb_clone_debug); ---- a/net80211/ieee80211_skb.h -+++ b/net80211/ieee80211_skb.h -@@ -116,7 +116,7 @@ int ieee80211_skb_references(void); - int vlan_hwaccel_receive_skb_debug(struct sk_buff *skb, - struct vlan_group *grp, unsigned short vlan_tag, - const char* func, int line); --int netif_rx_debug(struct sk_buff *skb, const char* func, int line); -+int netif_receive_skb_debug(struct sk_buff *skb, const char* func, int line); - struct sk_buff * alloc_skb_debug(unsigned int length, gfp_t gfp_mask, - const char *func, int line); - struct sk_buff * dev_alloc_skb_debug(unsigned int length, -@@ -151,7 +151,7 @@ struct sk_buff * skb_copy_expand_debug(c - #undef dev_queue_xmit - #undef kfree_skb - #undef kfree_skb_fast --#undef netif_rx -+#undef netif_receive_skb - #undef pskb_copy - #undef skb_clone - #undef skb_copy -@@ -168,8 +168,8 @@ struct sk_buff * skb_copy_expand_debug(c - skb_copy_expand_debug(_skb, _newheadroom, _newtailroom, _gfp_mask, __func__, __LINE__) - #define vlan_hwaccel_receive_skb(_skb, _grp, _tag) \ - vlan_hwaccel_receive_skb_debug(_skb, _grp, _tag, __func__, __LINE__) --#define netif_rx(_skb) \ -- netif_rx_debug(_skb, __func__, __LINE__) -+#define netif_receive_skb(_skb) \ -+ netif_receive_skb_debug(_skb, __func__, __LINE__) - #define alloc_skb(_length, _gfp_mask) \ - alloc_skb_debug(_length, _gfp_mask, __func__, __LINE__) - #define dev_alloc_skb(_length) \ |