Index: madwifi-trunk-r3314/ath/if_ath.c =================================================================== --- madwifi-trunk-r3314.orig/ath/if_ath.c 2008-03-06 23:49:24.344018025 +0100 +++ madwifi-trunk-r3314/ath/if_ath.c 2008-03-07 00:38:01.010996975 +0100 @@ -184,7 +184,11 @@ struct sk_buff *, int, int, u_int64_t); static void ath_setdefantenna(struct ath_softc *, u_int); static struct ath_txq *ath_txq_setup(struct ath_softc *, int, int); -static void ath_rx_tasklet(TQUEUE_ARG); +#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) +static int ath_rx_poll(struct napi_struct *napi, int budget); +#else +static int ath_rx_poll(struct net_device *dev, int *budget); +#endif static int ath_hardstart(struct sk_buff *, struct net_device *); static int ath_mgtstart(struct ieee80211com *, struct sk_buff *); #ifdef ATH_SUPERG_COMP @@ -374,6 +378,9 @@ u_int32_t new_clamped_maxtxpower); static u_int32_t ath_get_real_maxtxpower(struct ath_softc *sc); +static void ath_poll_disable(struct net_device *dev); +static void ath_poll_enable(struct net_device *dev); + /* calibrate every 30 secs in steady state but check every second at first. */ static int ath_calinterval = ATH_SHORT_CALINTERVAL; static int ath_countrycode = CTRY_DEFAULT; /* country code */ @@ -545,7 +552,6 @@ atomic_set(&sc->sc_txbuf_counter, 0); - ATH_INIT_TQUEUE(&sc->sc_rxtq, ath_rx_tasklet, dev); ATH_INIT_TQUEUE(&sc->sc_txtq, ath_tx_tasklet, dev); ATH_INIT_TQUEUE(&sc->sc_bmisstq, ath_bmiss_tasklet, dev); ATH_INIT_TQUEUE(&sc->sc_bstucktq, ath_bstuck_tasklet, dev); @@ -819,6 +825,12 @@ dev->set_mac_address = ath_set_mac_address; dev->change_mtu = ath_change_mtu; dev->tx_queue_len = ATH_TXBUF - ATH_TXBUF_MGT_RESERVED; +#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) + netif_napi_add(dev, &sc->sc_napi, ath_rx_poll, 64); +#else + dev->poll = ath_rx_poll; + dev->weight = 64; +#endif #ifdef USE_HEADERLEN_RESV dev->hard_header_len += sizeof(struct ieee80211_qosframe) + sizeof(struct llc) + @@ -2213,6 +2225,7 @@ (status & HAL_INT_GLOBAL) ? " HAL_INT_GLOBAL" : "" ); + sc->sc_isr = status; status &= sc->sc_imask; /* discard unasked for bits */ /* As soon as we know we have a real interrupt we intend to service, * we will check to see if we need an initial hardware TSF reading. @@ -2270,7 +2283,23 @@ } if (status & (HAL_INT_RX | HAL_INT_RXPHY)) { ath_uapsd_processtriggers(sc, hw_tsf); - ATH_SCHEDULE_TQUEUE(&sc->sc_rxtq, &needmark); + sc->sc_isr &= ~HAL_INT_RX; +#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) + if (netif_rx_schedule_prep(dev, &sc->sc_napi)) +#else + if (netif_rx_schedule_prep(dev)) +#endif + { +#ifndef ATH_PRECISE_TSF + sc->sc_imask &= ~HAL_INT_RX; + ath_hal_intrset(ah, sc->sc_imask); +#endif +#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) + __netif_rx_schedule(dev, &sc->sc_napi); +#else + __netif_rx_schedule(dev); +#endif + } } if (status & HAL_INT_TX) { #ifdef ATH_SUPERG_DYNTURBO @@ -2296,6 +2325,11 @@ } } #endif + /* disable transmit interrupt */ + sc->sc_isr &= ~HAL_INT_TX; + ath_hal_intrset(ah, sc->sc_imask & ~HAL_INT_TX); + sc->sc_imask &= ~HAL_INT_TX; + ATH_SCHEDULE_TQUEUE(&sc->sc_txtq, &needmark); } if (status & HAL_INT_BMISS) { @@ -2508,6 +2542,7 @@ if (sc->sc_tx99 != NULL) sc->sc_tx99->start(sc->sc_tx99); #endif + ath_poll_enable(dev); done: ATH_UNLOCK(sc); @@ -2548,6 +2583,9 @@ if (sc->sc_tx99 != NULL) sc->sc_tx99->stop(sc->sc_tx99); #endif +#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) + ath_poll_disable(dev); +#endif netif_stop_queue(dev); /* XXX re-enabled by ath_newstate */ dev->flags &= ~IFF_RUNNING; /* NB: avoid recursion */ ieee80211_stop_running(ic); /* stop all VAPs */ @@ -4006,6 +4044,39 @@ return ath_keyset(sc, k, mac, vap->iv_bss); } +static void ath_poll_disable(struct net_device *dev) +{ + struct ath_softc *sc = dev->priv; + + /* + * XXX Using in_softirq is not right since we might + * be called from other soft irq contexts than + * ath_rx_poll + */ + if (!in_softirq()) { +#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) + napi_disable(&sc->sc_napi); +#else + netif_poll_disable(dev); +#endif + } +} + +static void ath_poll_enable(struct net_device *dev) +{ + struct ath_softc *sc = dev->priv; + + /* NB: see above */ + if (!in_softirq()) { +#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) + napi_enable(&sc->sc_napi); +#else + netif_poll_enable(dev); +#endif + } +} + + /* * Block/unblock tx+rx processing while a key change is done. * We assume the caller serializes key management operations @@ -4023,13 +4094,7 @@ * When called from the rx tasklet we cannot use * tasklet_disable because it will block waiting * for us to complete execution. - * - * XXX Using in_softirq is not right since we might - * be called from other soft irq contexts than - * ath_rx_tasklet. */ - if (!in_softirq()) - tasklet_disable(&sc->sc_rxtq); netif_stop_queue(dev); } @@ -4040,9 +4105,9 @@ struct ath_softc *sc = dev->priv; DPRINTF(sc, ATH_DEBUG_KEYCACHE, "End\n"); - netif_wake_queue(dev); - if (!in_softirq()) /* NB: see above */ - tasklet_enable(&sc->sc_rxtq); + + if (dev->flags&IFF_RUNNING) + netif_wake_queue(dev); } /* @@ -6347,15 +6412,25 @@ sc->sc_rxotherant = 0; } -static void -ath_rx_tasklet(TQUEUE_ARG data) +static int +#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) +ath_rx_poll(struct napi_struct *napi, int budget) +#else +ath_rx_poll(struct net_device *dev, int *budget) +#endif { #define PA2DESC(_sc, _pa) \ ((struct ath_desc *)((caddr_t)(_sc)->sc_rxdma.dd_desc + \ ((_pa) - (_sc)->sc_rxdma.dd_desc_paddr))) - struct net_device *dev = (struct net_device *)data; - struct ath_buf *bf; +#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) + struct ath_softc *sc = container_of(napi, struct ath_softc, sc_napi); + struct net_device *dev = sc->sc_dev; + u_int rx_limit = budget; +#else struct ath_softc *sc = dev->priv; + u_int rx_limit = min(dev->quota, *budget); +#endif + struct ath_buf *bf; struct ieee80211com *ic = &sc->sc_ic; struct ath_hal *ah = sc ? sc->sc_ah : NULL; struct ath_desc *ds; @@ -6365,8 +6440,10 @@ unsigned int len; int type; u_int phyerr; + u_int processed = 0, early_stop = 0; DPRINTF(sc, ATH_DEBUG_RX_PROC, "invoked\n"); +process_rx_again: do { bf = STAILQ_FIRST(&sc->sc_rxbuf); if (bf == NULL) { /* XXX ??? can this happen */ @@ -6390,6 +6467,15 @@ /* NB: never process the self-linked entry at the end */ break; } + +#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) + processed++; +#endif + if (rx_limit-- < 0) { + early_stop = 1; + break; + } + skb = bf->bf_skb; if (skb == NULL) { EPRINTF(sc, "Dropping; buffer contains NULL skbuff.\n"); @@ -6437,6 +6523,7 @@ sc->sc_stats.ast_rx_phyerr++; phyerr = rs->rs_phyerr & 0x1f; sc->sc_stats.ast_rx_phy[phyerr]++; + goto rx_next; } if (rs->rs_status & HAL_RXERR_DECRYPT) { /* @@ -6632,9 +6719,43 @@ STAILQ_INSERT_TAIL(&sc->sc_rxbuf, bf, bf_list); ATH_RXBUF_UNLOCK_IRQ(sc); } while (ath_rxbuf_init(sc, bf) == 0); + if (!early_stop) { + unsigned long flags; + /* Check if more data is received while we were + * processing the descriptor chain. + */ +#ifndef ATH_PRECISE_TSF + local_irq_save(flags); + if (sc->sc_isr & HAL_INT_RX) { + u_int64_t hw_tsf = ath_hal_gettsf64(ah); + sc->sc_isr &= ~HAL_INT_RX; + local_irq_restore(flags); + ath_uapsd_processtriggers(sc, hw_tsf); + goto process_rx_again; + } +#endif +#ifndef ATH_PRECISE_TSF + sc->sc_imask |= HAL_INT_RX; + ath_hal_intrset(ah, sc->sc_imask); + local_irq_restore(flags); +#endif + } + +#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) + netif_rx_complete(dev, napi); +#else + netif_rx_complete(dev); + *budget -= processed; + dev->quota -= processed; +#endif /* rx signal state monitoring */ ath_hal_rxmonitor(ah, &sc->sc_halstats, &sc->sc_curchan); +#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) + return processed; +#else + return early_stop; +#endif #undef PA2DESC } @@ -8285,12 +8406,24 @@ { struct net_device *dev = (struct net_device *)data; struct ath_softc *sc = dev->priv; + unsigned long flags; +process_tx_again: if (txqactive(sc->sc_ah, 0)) ath_tx_processq(sc, &sc->sc_txq[0]); if (txqactive(sc->sc_ah, sc->sc_cabq->axq_qnum)) ath_tx_processq(sc, sc->sc_cabq); + local_irq_save(flags); + if (sc->sc_isr & HAL_INT_TX) { + sc->sc_isr &= ~HAL_INT_TX; + local_irq_restore(flags); + goto process_tx_again; + } + sc->sc_imask |= HAL_INT_TX; + ath_hal_intrset(sc->sc_ah, sc->sc_imask); + local_irq_restore(flags); + netif_wake_queue(dev); if (sc->sc_softled) @@ -8306,7 +8439,9 @@ { struct net_device *dev = (struct net_device *)data; struct ath_softc *sc = dev->priv; + unsigned long flags; +process_tx_again: /* * Process each active queue. */ @@ -8327,6 +8462,16 @@ if (sc->sc_uapsdq && txqactive(sc->sc_ah, sc->sc_uapsdq->axq_qnum)) ath_tx_processq(sc, sc->sc_uapsdq); + local_irq_save(flags); + if (sc->sc_isr & HAL_INT_TX) { + sc->sc_isr &= ~HAL_INT_TX; + local_irq_restore(flags); + goto process_tx_again; + } + sc->sc_imask |= HAL_INT_TX; + ath_hal_intrset(sc->sc_ah, sc->sc_imask); + local_irq_restore(flags); + netif_wake_queue(dev); if (sc->sc_softled) @@ -8342,13 +8487,25 @@ struct net_device *dev = (struct net_device *)data; struct ath_softc *sc = dev->priv; unsigned int i; + unsigned long flags; /* Process each active queue. This includes sc_cabq, sc_xrtq and * sc_uapsdq */ +process_tx_again: for (i = 0; i < HAL_NUM_TX_QUEUES; i++) if (ATH_TXQ_SETUP(sc, i) && txqactive(sc->sc_ah, i)) ath_tx_processq(sc, &sc->sc_txq[i]); + local_irq_save(flags); + if (sc->sc_isr & HAL_INT_TX) { + sc->sc_isr &= ~HAL_INT_TX; + local_irq_restore(flags); + goto process_tx_again; + } + sc->sc_imask |= HAL_INT_TX; + ath_hal_intrset(sc->sc_ah, sc->sc_imask); + local_irq_restore(flags); + netif_wake_queue(dev); if (sc->sc_softled) @@ -8423,6 +8580,7 @@ ath_draintxq(struct ath_softc *sc) { struct ath_hal *ah = sc->sc_ah; + int npend = 0; unsigned int i; /* XXX return value */ @@ -10281,9 +10439,9 @@ dev->mtu = mtu; if ((dev->flags & IFF_RUNNING) && !sc->sc_invalid) { /* NB: the rx buffers may need to be reallocated */ - tasklet_disable(&sc->sc_rxtq); + ath_poll_disable(dev); error = ath_reset(dev); - tasklet_enable(&sc->sc_rxtq); + ath_poll_enable(dev); } ATH_UNLOCK(sc); Index: madwifi-trunk-r3314/ath/if_athvar.h =================================================================== --- madwifi-trunk-r3314.orig/ath/if_athvar.h 2008-03-06 23:49:24.296015306 +0100 +++ madwifi-trunk-r3314/ath/if_athvar.h 2008-03-07 00:34:32.792958008 +0100 @@ -53,6 +53,10 @@ # include #endif +#if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,0) +#define irqs_disabled() 0 +#endif + /* * Deduce if tasklets are available. If not then * fall back to using the immediate work queue. @@ -616,6 +620,9 @@ struct ath_softc { struct ieee80211com sc_ic; /* NB: must be first */ struct net_device *sc_dev; +#if LINUX_VERSION_CODE >= KERNEL_VERSION(2,6,24) + struct napi_struct sc_napi; +#endif void __iomem *sc_iobase; /* address of the device */ struct semaphore sc_lock; /* dev-level lock */ struct net_device_stats sc_devstats; /* device statistics */ @@ -730,7 +737,6 @@ struct ath_buf *sc_rxbufcur; /* current rx buffer */ u_int32_t *sc_rxlink; /* link ptr in last RX desc */ spinlock_t sc_rxbuflock; - struct ATH_TQ_STRUCT sc_rxtq; /* rx intr tasklet */ struct ATH_TQ_STRUCT sc_rxorntq; /* rxorn intr tasklet */ u_int8_t sc_defant; /* current default antenna */ u_int8_t sc_rxotherant; /* RXs on non-default antenna */ @@ -745,6 +751,7 @@ u_int sc_txintrperiod; /* tx interrupt batching */ struct ath_txq sc_txq[HAL_NUM_TX_QUEUES]; struct ath_txq *sc_ac2q[WME_NUM_AC]; /* WME AC -> h/w qnum */ + HAL_INT sc_isr; /* unmasked ISR state */ struct ATH_TQ_STRUCT sc_txtq; /* tx intr tasklet */ u_int8_t sc_grppoll_str[GRPPOLL_RATE_STR_LEN]; struct ath_descdma sc_bdma; /* beacon descriptors */ @@ -858,6 +865,8 @@ #define ATH_TXBUF_LOCK_CHECK(_sc) #endif +#define ATH_DISABLE_INTR local_irq_disable +#define ATH_ENABLE_INTR local_irq_enable #define ATH_RXBUF_LOCK_INIT(_sc) spin_lock_init(&(_sc)->sc_rxbuflock) #define ATH_RXBUF_LOCK_DESTROY(_sc) Index: madwifi-trunk-r3314/net80211/ieee80211_input.c =================================================================== --- madwifi-trunk-r3314.orig/net80211/ieee80211_input.c 2008-03-06 23:49:11.027259154 +0100 +++ madwifi-trunk-r3314/net80211/ieee80211_input.c 2008-03-07 00:34:41.589459294 +0100 @@ -1198,7 +1198,7 @@ /* attach vlan tag */ struct ieee80211_node *ni_tmp = SKB_CB(skb)->ni; if (vlan_hwaccel_receive_skb(skb, vap->iv_vlgrp, ni->ni_vlan) == NET_RX_DROP) { - /* If netif_rx dropped the packet because + /* If netif_receive_skb dropped the packet because * device was too busy */ if (ni_tmp != NULL) { /* node reference was leaked */ @@ -1209,8 +1209,8 @@ skb = NULL; /* SKB is no longer ours */ } else { struct ieee80211_node *ni_tmp = SKB_CB(skb)->ni; - if (netif_rx(skb) == NET_RX_DROP) { - /* If netif_rx dropped the packet because + if (netif_receive_skb(skb) == NET_RX_DROP) { + /* If netif_receive_skb dropped the packet because * device was too busy */ if (ni_tmp != NULL) { /* node reference was leaked */ @@ -2322,8 +2322,8 @@ skb1->protocol = __constant_htons(0x0019); /* ETH_P_80211_RAW */ ni_tmp = SKB_CB(skb1)->ni; - if (netif_rx(skb1) == NET_RX_DROP) { - /* If netif_rx dropped the packet because + if (netif_receive_skb(skb1) == NET_RX_DROP) { + /* If netif_receive_skb dropped the packet because * device was too busy */ if (ni_tmp != NULL) { /* node reference was leaked */ Index: madwifi-trunk-r3314/net80211/ieee80211_monitor.c =================================================================== --- madwifi-trunk-r3314.orig/net80211/ieee80211_monitor.c 2008-03-06 23:49:11.031259383 +0100 +++ madwifi-trunk-r3314/net80211/ieee80211_monitor.c 2008-03-07 00:33:59.235045655 +0100 @@ -584,8 +584,8 @@ skb1->protocol = __constant_htons(0x0019); /* ETH_P_80211_RAW */ - if (netif_rx(skb1) == NET_RX_DROP) { - /* If netif_rx dropped the packet because + if (netif_receive_skb(skb1) == NET_RX_DROP) { + /* If netif_receive_skb dropped the packet because * device was too busy, reclaim the ref. in * the skb. */ if (SKB_CB(skb1)->ni != NULL) Index: madwifi-trunk-r3314/net80211/ieee80211_skb.c =================================================================== --- madwifi-trunk-r3314.orig/net80211/ieee80211_skb.c 2008-03-06 23:49:11.039259840 +0100 +++ madwifi-trunk-r3314/net80211/ieee80211_skb.c 2008-03-06 23:49:24.384020316 +0100 @@ -73,7 +73,7 @@ #undef dev_queue_xmit #undef kfree_skb #undef kfree_skb_fast -#undef netif_rx +#undef netif_receive_skb #undef pskb_copy #undef skb_clone #undef skb_copy @@ -638,8 +638,8 @@ grp, vlan_tag); } -int netif_rx_debug(struct sk_buff *skb, const char* func, int line) { - return netif_rx(untrack_skb(skb, 0, func, line, __func__, __LINE__)); +int netif_receive_skb_debug(struct sk_buff *skb, const char* func, int line) { + return netif_receive_skb(untrack_skb(skb, 0, func, line, __func__, __LINE__)); } struct sk_buff * alloc_skb_debug(unsigned int length, gfp_t gfp_mask, @@ -760,7 +760,7 @@ } EXPORT_SYMBOL(vlan_hwaccel_receive_skb_debug); -EXPORT_SYMBOL(netif_rx_debug); +EXPORT_SYMBOL(netif_receive_skb_debug); EXPORT_SYMBOL(alloc_skb_debug); EXPORT_SYMBOL(dev_alloc_skb_debug); EXPORT_SYMBOL(skb_clone_debug); Index: madwifi-trunk-r3314/net80211/ieee80211_skb.h =================================================================== --- madwifi-trunk-r3314.orig/net80211/ieee80211_skb.h 2008-03-06 23:49:11.047260294 +0100 +++ madwifi-trunk-r3314/net80211/ieee80211_skb.h 2008-03-06 23:49:24.384020316 +0100 @@ -116,7 +116,7 @@ int vlan_hwaccel_receive_skb_debug(struct sk_buff *skb, struct vlan_group *grp, unsigned short vlan_tag, const char* func, int line); -int netif_rx_debug(struct sk_buff *skb, const char* func, int line); +int netif_receive_skb_debug(struct sk_buff *skb, const char* func, int line); struct sk_buff * alloc_skb_debug(unsigned int length, gfp_t gfp_mask, const char *func, int line); struct sk_buff * dev_alloc_skb_debug(unsigned int length, @@ -151,7 +151,7 @@ #undef dev_queue_xmit #undef kfree_skb #undef kfree_skb_fast -#undef netif_rx +#undef netif_receive_skb #undef pskb_copy #undef skb_clone #undef skb_copy @@ -168,8 +168,8 @@ skb_copy_expand_debug(_skb, _newheadroom, _newtailroom, _gfp_mask, __func__, __LINE__) #define vlan_hwaccel_receive_skb(_skb, _grp, _tag) \ vlan_hwaccel_receive_skb_debug(_skb, _grp, _tag, __func__, __LINE__) -#define netif_rx(_skb) \ - netif_rx_debug(_skb, __func__, __LINE__) +#define netif_receive_skb(_skb) \ + netif_receive_skb_debug(_skb, __func__, __LINE__) #define alloc_skb(_length, _gfp_mask) \ alloc_skb_debug(_length, _gfp_mask, __func__, __LINE__) #define dev_alloc_skb(_length) \