1203c4805SLuis R. Rodriguez /* 25b68138eSSujith Manoharan * Copyright (c) 2008-2011 Atheros Communications Inc. 3203c4805SLuis R. Rodriguez * 4203c4805SLuis R. Rodriguez * Permission to use, copy, modify, and/or distribute this software for any 5203c4805SLuis R. Rodriguez * purpose with or without fee is hereby granted, provided that the above 6203c4805SLuis R. Rodriguez * copyright notice and this permission notice appear in all copies. 7203c4805SLuis R. Rodriguez * 8203c4805SLuis R. Rodriguez * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES 9203c4805SLuis R. Rodriguez * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF 10203c4805SLuis R. Rodriguez * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR 11203c4805SLuis R. Rodriguez * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES 12203c4805SLuis R. Rodriguez * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN 13203c4805SLuis R. Rodriguez * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF 14203c4805SLuis R. Rodriguez * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. 15203c4805SLuis R. Rodriguez */ 16203c4805SLuis R. Rodriguez 17b7f080cfSAlexey Dobriyan #include <linux/dma-mapping.h> 18203c4805SLuis R. Rodriguez #include "ath9k.h" 19b622a720SLuis R. Rodriguez #include "ar9003_mac.h" 20203c4805SLuis R. Rodriguez 21b5c80475SFelix Fietkau #define SKB_CB_ATHBUF(__skb) (*((struct ath_buf **)__skb->cb)) 22b5c80475SFelix Fietkau 23ededf1f8SVasanthakumar Thiagarajan static inline bool ath9k_check_auto_sleep(struct ath_softc *sc) 24ededf1f8SVasanthakumar Thiagarajan { 25ededf1f8SVasanthakumar Thiagarajan return sc->ps_enabled && 26ededf1f8SVasanthakumar Thiagarajan (sc->sc_ah->caps.hw_caps & ATH9K_HW_CAP_AUTOSLEEP); 27ededf1f8SVasanthakumar Thiagarajan } 28ededf1f8SVasanthakumar Thiagarajan 29203c4805SLuis R. Rodriguez /* 30203c4805SLuis R. Rodriguez * Setup and link descriptors. 31203c4805SLuis R. Rodriguez * 32203c4805SLuis R. Rodriguez * 11N: we can no longer afford to self link the last descriptor. 33203c4805SLuis R. Rodriguez * MAC acknowledges BA status as long as it copies frames to host 34203c4805SLuis R. Rodriguez * buffer (or rx fifo). This can incorrectly acknowledge packets 35203c4805SLuis R. Rodriguez * to a sender if last desc is self-linked. 36203c4805SLuis R. Rodriguez */ 37203c4805SLuis R. Rodriguez static void ath_rx_buf_link(struct ath_softc *sc, struct ath_buf *bf) 38203c4805SLuis R. Rodriguez { 39203c4805SLuis R. Rodriguez struct ath_hw *ah = sc->sc_ah; 40cc861f74SLuis R. Rodriguez struct ath_common *common = ath9k_hw_common(ah); 41203c4805SLuis R. Rodriguez struct ath_desc *ds; 42203c4805SLuis R. Rodriguez struct sk_buff *skb; 43203c4805SLuis R. Rodriguez 44203c4805SLuis R. Rodriguez ATH_RXBUF_RESET(bf); 45203c4805SLuis R. Rodriguez 46203c4805SLuis R. Rodriguez ds = bf->bf_desc; 47203c4805SLuis R. Rodriguez ds->ds_link = 0; /* link to null */ 48203c4805SLuis R. Rodriguez ds->ds_data = bf->bf_buf_addr; 49203c4805SLuis R. Rodriguez 50203c4805SLuis R. Rodriguez /* virtual addr of the beginning of the buffer. */ 51203c4805SLuis R. Rodriguez skb = bf->bf_mpdu; 529680e8a3SLuis R. Rodriguez BUG_ON(skb == NULL); 53203c4805SLuis R. Rodriguez ds->ds_vdata = skb->data; 54203c4805SLuis R. Rodriguez 55cc861f74SLuis R. Rodriguez /* 56cc861f74SLuis R. Rodriguez * setup rx descriptors. The rx_bufsize here tells the hardware 57203c4805SLuis R. Rodriguez * how much data it can DMA to us and that we are prepared 58cc861f74SLuis R. Rodriguez * to process 59cc861f74SLuis R. Rodriguez */ 60203c4805SLuis R. Rodriguez ath9k_hw_setuprxdesc(ah, ds, 61cc861f74SLuis R. Rodriguez common->rx_bufsize, 62203c4805SLuis R. Rodriguez 0); 63203c4805SLuis R. Rodriguez 64203c4805SLuis R. Rodriguez if (sc->rx.rxlink == NULL) 65203c4805SLuis R. Rodriguez ath9k_hw_putrxbuf(ah, bf->bf_daddr); 66203c4805SLuis R. Rodriguez else 67203c4805SLuis R. Rodriguez *sc->rx.rxlink = bf->bf_daddr; 68203c4805SLuis R. Rodriguez 69203c4805SLuis R. Rodriguez sc->rx.rxlink = &ds->ds_link; 70203c4805SLuis R. Rodriguez } 71203c4805SLuis R. Rodriguez 72203c4805SLuis R. Rodriguez static void ath_setdefantenna(struct ath_softc *sc, u32 antenna) 73203c4805SLuis R. Rodriguez { 74203c4805SLuis R. Rodriguez /* XXX block beacon interrupts */ 75203c4805SLuis R. Rodriguez ath9k_hw_setantenna(sc->sc_ah, antenna); 76203c4805SLuis R. Rodriguez sc->rx.defant = antenna; 77203c4805SLuis R. Rodriguez sc->rx.rxotherant = 0; 78203c4805SLuis R. Rodriguez } 79203c4805SLuis R. Rodriguez 80203c4805SLuis R. Rodriguez static void ath_opmode_init(struct ath_softc *sc) 81203c4805SLuis R. Rodriguez { 82203c4805SLuis R. Rodriguez struct ath_hw *ah = sc->sc_ah; 831510718dSLuis R. Rodriguez struct ath_common *common = ath9k_hw_common(ah); 841510718dSLuis R. Rodriguez 85203c4805SLuis R. Rodriguez u32 rfilt, mfilt[2]; 86203c4805SLuis R. Rodriguez 87203c4805SLuis R. Rodriguez /* configure rx filter */ 88203c4805SLuis R. Rodriguez rfilt = ath_calcrxfilter(sc); 89203c4805SLuis R. Rodriguez ath9k_hw_setrxfilter(ah, rfilt); 90203c4805SLuis R. Rodriguez 91203c4805SLuis R. Rodriguez /* configure bssid mask */ 9213b81559SLuis R. Rodriguez ath_hw_setbssidmask(common); 93203c4805SLuis R. Rodriguez 94203c4805SLuis R. Rodriguez /* configure operational mode */ 95203c4805SLuis R. Rodriguez ath9k_hw_setopmode(ah); 96203c4805SLuis R. Rodriguez 97203c4805SLuis R. Rodriguez /* calculate and install multicast filter */ 98203c4805SLuis R. Rodriguez mfilt[0] = mfilt[1] = ~0; 99203c4805SLuis R. Rodriguez ath9k_hw_setmcastfilter(ah, mfilt[0], mfilt[1]); 100203c4805SLuis R. Rodriguez } 101203c4805SLuis R. Rodriguez 102b5c80475SFelix Fietkau static bool ath_rx_edma_buf_link(struct ath_softc *sc, 103b5c80475SFelix Fietkau enum ath9k_rx_qtype qtype) 104b5c80475SFelix Fietkau { 105b5c80475SFelix Fietkau struct ath_hw *ah = sc->sc_ah; 106b5c80475SFelix Fietkau struct ath_rx_edma *rx_edma; 107b5c80475SFelix Fietkau struct sk_buff *skb; 108b5c80475SFelix Fietkau struct ath_buf *bf; 109b5c80475SFelix Fietkau 110b5c80475SFelix Fietkau rx_edma = &sc->rx.rx_edma[qtype]; 111b5c80475SFelix Fietkau if (skb_queue_len(&rx_edma->rx_fifo) >= rx_edma->rx_fifo_hwsize) 112b5c80475SFelix Fietkau return false; 113b5c80475SFelix Fietkau 114b5c80475SFelix Fietkau bf = list_first_entry(&sc->rx.rxbuf, struct ath_buf, list); 115b5c80475SFelix Fietkau list_del_init(&bf->list); 116b5c80475SFelix Fietkau 117b5c80475SFelix Fietkau skb = bf->bf_mpdu; 118b5c80475SFelix Fietkau 119b5c80475SFelix Fietkau ATH_RXBUF_RESET(bf); 120b5c80475SFelix Fietkau memset(skb->data, 0, ah->caps.rx_status_len); 121b5c80475SFelix Fietkau dma_sync_single_for_device(sc->dev, bf->bf_buf_addr, 122b5c80475SFelix Fietkau ah->caps.rx_status_len, DMA_TO_DEVICE); 123b5c80475SFelix Fietkau 124b5c80475SFelix Fietkau SKB_CB_ATHBUF(skb) = bf; 125b5c80475SFelix Fietkau ath9k_hw_addrxbuf_edma(ah, bf->bf_buf_addr, qtype); 126b5c80475SFelix Fietkau skb_queue_tail(&rx_edma->rx_fifo, skb); 127b5c80475SFelix Fietkau 128b5c80475SFelix Fietkau return true; 129b5c80475SFelix Fietkau } 130b5c80475SFelix Fietkau 131b5c80475SFelix Fietkau static void ath_rx_addbuffer_edma(struct ath_softc *sc, 132b5c80475SFelix Fietkau enum ath9k_rx_qtype qtype, int size) 133b5c80475SFelix Fietkau { 134b5c80475SFelix Fietkau struct ath_common *common = ath9k_hw_common(sc->sc_ah); 1356a01f0c0SMohammed Shafi Shajakhan struct ath_buf *bf, *tbf; 136b5c80475SFelix Fietkau 137b5c80475SFelix Fietkau if (list_empty(&sc->rx.rxbuf)) { 138d2182b69SJoe Perches ath_dbg(common, QUEUE, "No free rx buf available\n"); 139b5c80475SFelix Fietkau return; 140b5c80475SFelix Fietkau } 141b5c80475SFelix Fietkau 1426a01f0c0SMohammed Shafi Shajakhan list_for_each_entry_safe(bf, tbf, &sc->rx.rxbuf, list) 143b5c80475SFelix Fietkau if (!ath_rx_edma_buf_link(sc, qtype)) 144b5c80475SFelix Fietkau break; 145b5c80475SFelix Fietkau 146b5c80475SFelix Fietkau } 147b5c80475SFelix Fietkau 148b5c80475SFelix Fietkau static void ath_rx_remove_buffer(struct ath_softc *sc, 149b5c80475SFelix Fietkau enum ath9k_rx_qtype qtype) 150b5c80475SFelix Fietkau { 151b5c80475SFelix Fietkau struct ath_buf *bf; 152b5c80475SFelix Fietkau struct ath_rx_edma *rx_edma; 153b5c80475SFelix Fietkau struct sk_buff *skb; 154b5c80475SFelix Fietkau 155b5c80475SFelix Fietkau rx_edma = &sc->rx.rx_edma[qtype]; 156b5c80475SFelix Fietkau 157b5c80475SFelix Fietkau while ((skb = skb_dequeue(&rx_edma->rx_fifo)) != NULL) { 158b5c80475SFelix Fietkau bf = SKB_CB_ATHBUF(skb); 159b5c80475SFelix Fietkau BUG_ON(!bf); 160b5c80475SFelix Fietkau list_add_tail(&bf->list, &sc->rx.rxbuf); 161b5c80475SFelix Fietkau } 162b5c80475SFelix Fietkau } 163b5c80475SFelix Fietkau 164b5c80475SFelix Fietkau static void ath_rx_edma_cleanup(struct ath_softc *sc) 165b5c80475SFelix Fietkau { 166ba542385SMohammed Shafi Shajakhan struct ath_hw *ah = sc->sc_ah; 167ba542385SMohammed Shafi Shajakhan struct ath_common *common = ath9k_hw_common(ah); 168b5c80475SFelix Fietkau struct ath_buf *bf; 169b5c80475SFelix Fietkau 170b5c80475SFelix Fietkau ath_rx_remove_buffer(sc, ATH9K_RX_QUEUE_LP); 171b5c80475SFelix Fietkau ath_rx_remove_buffer(sc, ATH9K_RX_QUEUE_HP); 172b5c80475SFelix Fietkau 173b5c80475SFelix Fietkau list_for_each_entry(bf, &sc->rx.rxbuf, list) { 174ba542385SMohammed Shafi Shajakhan if (bf->bf_mpdu) { 175ba542385SMohammed Shafi Shajakhan dma_unmap_single(sc->dev, bf->bf_buf_addr, 176ba542385SMohammed Shafi Shajakhan common->rx_bufsize, 177ba542385SMohammed Shafi Shajakhan DMA_BIDIRECTIONAL); 178b5c80475SFelix Fietkau dev_kfree_skb_any(bf->bf_mpdu); 179ba542385SMohammed Shafi Shajakhan bf->bf_buf_addr = 0; 180ba542385SMohammed Shafi Shajakhan bf->bf_mpdu = NULL; 181ba542385SMohammed Shafi Shajakhan } 182b5c80475SFelix Fietkau } 183b5c80475SFelix Fietkau 184b5c80475SFelix Fietkau INIT_LIST_HEAD(&sc->rx.rxbuf); 185b5c80475SFelix Fietkau 186b5c80475SFelix Fietkau kfree(sc->rx.rx_bufptr); 187b5c80475SFelix Fietkau sc->rx.rx_bufptr = NULL; 188b5c80475SFelix Fietkau } 189b5c80475SFelix Fietkau 190b5c80475SFelix Fietkau static void ath_rx_edma_init_queue(struct ath_rx_edma *rx_edma, int size) 191b5c80475SFelix Fietkau { 192b5c80475SFelix Fietkau skb_queue_head_init(&rx_edma->rx_fifo); 193b5c80475SFelix Fietkau rx_edma->rx_fifo_hwsize = size; 194b5c80475SFelix Fietkau } 195b5c80475SFelix Fietkau 196b5c80475SFelix Fietkau static int ath_rx_edma_init(struct ath_softc *sc, int nbufs) 197b5c80475SFelix Fietkau { 198b5c80475SFelix Fietkau struct ath_common *common = ath9k_hw_common(sc->sc_ah); 199b5c80475SFelix Fietkau struct ath_hw *ah = sc->sc_ah; 200b5c80475SFelix Fietkau struct sk_buff *skb; 201b5c80475SFelix Fietkau struct ath_buf *bf; 202b5c80475SFelix Fietkau int error = 0, i; 203b5c80475SFelix Fietkau u32 size; 204b5c80475SFelix Fietkau 205b5c80475SFelix Fietkau ath9k_hw_set_rx_bufsize(ah, common->rx_bufsize - 206b5c80475SFelix Fietkau ah->caps.rx_status_len); 207b5c80475SFelix Fietkau 208b5c80475SFelix Fietkau ath_rx_edma_init_queue(&sc->rx.rx_edma[ATH9K_RX_QUEUE_LP], 209b5c80475SFelix Fietkau ah->caps.rx_lp_qdepth); 210b5c80475SFelix Fietkau ath_rx_edma_init_queue(&sc->rx.rx_edma[ATH9K_RX_QUEUE_HP], 211b5c80475SFelix Fietkau ah->caps.rx_hp_qdepth); 212b5c80475SFelix Fietkau 213b5c80475SFelix Fietkau size = sizeof(struct ath_buf) * nbufs; 214b5c80475SFelix Fietkau bf = kzalloc(size, GFP_KERNEL); 215b5c80475SFelix Fietkau if (!bf) 216b5c80475SFelix Fietkau return -ENOMEM; 217b5c80475SFelix Fietkau 218b5c80475SFelix Fietkau INIT_LIST_HEAD(&sc->rx.rxbuf); 219b5c80475SFelix Fietkau sc->rx.rx_bufptr = bf; 220b5c80475SFelix Fietkau 221b5c80475SFelix Fietkau for (i = 0; i < nbufs; i++, bf++) { 222b5c80475SFelix Fietkau skb = ath_rxbuf_alloc(common, common->rx_bufsize, GFP_KERNEL); 223b5c80475SFelix Fietkau if (!skb) { 224b5c80475SFelix Fietkau error = -ENOMEM; 225b5c80475SFelix Fietkau goto rx_init_fail; 226b5c80475SFelix Fietkau } 227b5c80475SFelix Fietkau 228b5c80475SFelix Fietkau memset(skb->data, 0, common->rx_bufsize); 229b5c80475SFelix Fietkau bf->bf_mpdu = skb; 230b5c80475SFelix Fietkau 231b5c80475SFelix Fietkau bf->bf_buf_addr = dma_map_single(sc->dev, skb->data, 232b5c80475SFelix Fietkau common->rx_bufsize, 233b5c80475SFelix Fietkau DMA_BIDIRECTIONAL); 234b5c80475SFelix Fietkau if (unlikely(dma_mapping_error(sc->dev, 235b5c80475SFelix Fietkau bf->bf_buf_addr))) { 236b5c80475SFelix Fietkau dev_kfree_skb_any(skb); 237b5c80475SFelix Fietkau bf->bf_mpdu = NULL; 2386cf9e995SBen Greear bf->bf_buf_addr = 0; 2393800276aSJoe Perches ath_err(common, 240b5c80475SFelix Fietkau "dma_mapping_error() on RX init\n"); 241b5c80475SFelix Fietkau error = -ENOMEM; 242b5c80475SFelix Fietkau goto rx_init_fail; 243b5c80475SFelix Fietkau } 244b5c80475SFelix Fietkau 245b5c80475SFelix Fietkau list_add_tail(&bf->list, &sc->rx.rxbuf); 246b5c80475SFelix Fietkau } 247b5c80475SFelix Fietkau 248b5c80475SFelix Fietkau return 0; 249b5c80475SFelix Fietkau 250b5c80475SFelix Fietkau rx_init_fail: 251b5c80475SFelix Fietkau ath_rx_edma_cleanup(sc); 252b5c80475SFelix Fietkau return error; 253b5c80475SFelix Fietkau } 254b5c80475SFelix Fietkau 255b5c80475SFelix Fietkau static void ath_edma_start_recv(struct ath_softc *sc) 256b5c80475SFelix Fietkau { 257b5c80475SFelix Fietkau spin_lock_bh(&sc->rx.rxbuflock); 258b5c80475SFelix Fietkau 259b5c80475SFelix Fietkau ath9k_hw_rxena(sc->sc_ah); 260b5c80475SFelix Fietkau 261b5c80475SFelix Fietkau ath_rx_addbuffer_edma(sc, ATH9K_RX_QUEUE_HP, 262b5c80475SFelix Fietkau sc->rx.rx_edma[ATH9K_RX_QUEUE_HP].rx_fifo_hwsize); 263b5c80475SFelix Fietkau 264b5c80475SFelix Fietkau ath_rx_addbuffer_edma(sc, ATH9K_RX_QUEUE_LP, 265b5c80475SFelix Fietkau sc->rx.rx_edma[ATH9K_RX_QUEUE_LP].rx_fifo_hwsize); 266b5c80475SFelix Fietkau 267b5c80475SFelix Fietkau ath_opmode_init(sc); 268b5c80475SFelix Fietkau 2694cb54fa3SSujith Manoharan ath9k_hw_startpcureceive(sc->sc_ah, !!(sc->hw->conf.flags & IEEE80211_CONF_OFFCHANNEL)); 2707583c550SLuis R. Rodriguez 2717583c550SLuis R. Rodriguez spin_unlock_bh(&sc->rx.rxbuflock); 272b5c80475SFelix Fietkau } 273b5c80475SFelix Fietkau 274b5c80475SFelix Fietkau static void ath_edma_stop_recv(struct ath_softc *sc) 275b5c80475SFelix Fietkau { 276b5c80475SFelix Fietkau ath_rx_remove_buffer(sc, ATH9K_RX_QUEUE_HP); 277b5c80475SFelix Fietkau ath_rx_remove_buffer(sc, ATH9K_RX_QUEUE_LP); 278b5c80475SFelix Fietkau } 279b5c80475SFelix Fietkau 280203c4805SLuis R. Rodriguez int ath_rx_init(struct ath_softc *sc, int nbufs) 281203c4805SLuis R. Rodriguez { 28227c51f1aSLuis R. Rodriguez struct ath_common *common = ath9k_hw_common(sc->sc_ah); 283203c4805SLuis R. Rodriguez struct sk_buff *skb; 284203c4805SLuis R. Rodriguez struct ath_buf *bf; 285203c4805SLuis R. Rodriguez int error = 0; 286203c4805SLuis R. Rodriguez 2874bdd1e97SLuis R. Rodriguez spin_lock_init(&sc->sc_pcu_lock); 288203c4805SLuis R. Rodriguez spin_lock_init(&sc->rx.rxbuflock); 289781b14a3SSujith Manoharan clear_bit(SC_OP_RXFLUSH, &sc->sc_flags); 290203c4805SLuis R. Rodriguez 2910d95521eSFelix Fietkau common->rx_bufsize = IEEE80211_MAX_MPDU_LEN / 2 + 2920d95521eSFelix Fietkau sc->sc_ah->caps.rx_status_len; 2930d95521eSFelix Fietkau 294b5c80475SFelix Fietkau if (sc->sc_ah->caps.hw_caps & ATH9K_HW_CAP_EDMA) { 295b5c80475SFelix Fietkau return ath_rx_edma_init(sc, nbufs); 296b5c80475SFelix Fietkau } else { 297d2182b69SJoe Perches ath_dbg(common, CONFIG, "cachelsz %u rxbufsize %u\n", 298cc861f74SLuis R. Rodriguez common->cachelsz, common->rx_bufsize); 299203c4805SLuis R. Rodriguez 300203c4805SLuis R. Rodriguez /* Initialize rx descriptors */ 301203c4805SLuis R. Rodriguez 302203c4805SLuis R. Rodriguez error = ath_descdma_setup(sc, &sc->rx.rxdma, &sc->rx.rxbuf, 3034adfcdedSVasanthakumar Thiagarajan "rx", nbufs, 1, 0); 304203c4805SLuis R. Rodriguez if (error != 0) { 3053800276aSJoe Perches ath_err(common, 306b5c80475SFelix Fietkau "failed to allocate rx descriptors: %d\n", 307b5c80475SFelix Fietkau error); 308203c4805SLuis R. Rodriguez goto err; 309203c4805SLuis R. Rodriguez } 310203c4805SLuis R. Rodriguez 311203c4805SLuis R. Rodriguez list_for_each_entry(bf, &sc->rx.rxbuf, list) { 312b5c80475SFelix Fietkau skb = ath_rxbuf_alloc(common, common->rx_bufsize, 313b5c80475SFelix Fietkau GFP_KERNEL); 314203c4805SLuis R. Rodriguez if (skb == NULL) { 315203c4805SLuis R. Rodriguez error = -ENOMEM; 316203c4805SLuis R. Rodriguez goto err; 317203c4805SLuis R. Rodriguez } 318203c4805SLuis R. Rodriguez 319203c4805SLuis R. Rodriguez bf->bf_mpdu = skb; 320203c4805SLuis R. Rodriguez bf->bf_buf_addr = dma_map_single(sc->dev, skb->data, 321cc861f74SLuis R. Rodriguez common->rx_bufsize, 322203c4805SLuis R. Rodriguez DMA_FROM_DEVICE); 323203c4805SLuis R. Rodriguez if (unlikely(dma_mapping_error(sc->dev, 324203c4805SLuis R. Rodriguez bf->bf_buf_addr))) { 325203c4805SLuis R. Rodriguez dev_kfree_skb_any(skb); 326203c4805SLuis R. Rodriguez bf->bf_mpdu = NULL; 3276cf9e995SBen Greear bf->bf_buf_addr = 0; 3283800276aSJoe Perches ath_err(common, 329203c4805SLuis R. Rodriguez "dma_mapping_error() on RX init\n"); 330203c4805SLuis R. Rodriguez error = -ENOMEM; 331203c4805SLuis R. Rodriguez goto err; 332203c4805SLuis R. Rodriguez } 333203c4805SLuis R. Rodriguez } 334203c4805SLuis R. Rodriguez sc->rx.rxlink = NULL; 335b5c80475SFelix Fietkau } 336203c4805SLuis R. Rodriguez 337203c4805SLuis R. Rodriguez err: 338203c4805SLuis R. Rodriguez if (error) 339203c4805SLuis R. Rodriguez ath_rx_cleanup(sc); 340203c4805SLuis R. Rodriguez 341203c4805SLuis R. Rodriguez return error; 342203c4805SLuis R. Rodriguez } 343203c4805SLuis R. Rodriguez 344203c4805SLuis R. Rodriguez void ath_rx_cleanup(struct ath_softc *sc) 345203c4805SLuis R. Rodriguez { 346cc861f74SLuis R. Rodriguez struct ath_hw *ah = sc->sc_ah; 347cc861f74SLuis R. Rodriguez struct ath_common *common = ath9k_hw_common(ah); 348203c4805SLuis R. Rodriguez struct sk_buff *skb; 349203c4805SLuis R. Rodriguez struct ath_buf *bf; 350203c4805SLuis R. Rodriguez 351b5c80475SFelix Fietkau if (sc->sc_ah->caps.hw_caps & ATH9K_HW_CAP_EDMA) { 352b5c80475SFelix Fietkau ath_rx_edma_cleanup(sc); 353b5c80475SFelix Fietkau return; 354b5c80475SFelix Fietkau } else { 355203c4805SLuis R. Rodriguez list_for_each_entry(bf, &sc->rx.rxbuf, list) { 356203c4805SLuis R. Rodriguez skb = bf->bf_mpdu; 357203c4805SLuis R. Rodriguez if (skb) { 358203c4805SLuis R. Rodriguez dma_unmap_single(sc->dev, bf->bf_buf_addr, 359b5c80475SFelix Fietkau common->rx_bufsize, 360b5c80475SFelix Fietkau DMA_FROM_DEVICE); 361203c4805SLuis R. Rodriguez dev_kfree_skb(skb); 3626cf9e995SBen Greear bf->bf_buf_addr = 0; 3636cf9e995SBen Greear bf->bf_mpdu = NULL; 364203c4805SLuis R. Rodriguez } 365203c4805SLuis R. Rodriguez } 366203c4805SLuis R. Rodriguez 367203c4805SLuis R. Rodriguez if (sc->rx.rxdma.dd_desc_len != 0) 368203c4805SLuis R. Rodriguez ath_descdma_cleanup(sc, &sc->rx.rxdma, &sc->rx.rxbuf); 369203c4805SLuis R. Rodriguez } 370b5c80475SFelix Fietkau } 371203c4805SLuis R. Rodriguez 372203c4805SLuis R. Rodriguez /* 373203c4805SLuis R. Rodriguez * Calculate the receive filter according to the 374203c4805SLuis R. Rodriguez * operating mode and state: 375203c4805SLuis R. Rodriguez * 376203c4805SLuis R. Rodriguez * o always accept unicast, broadcast, and multicast traffic 377203c4805SLuis R. Rodriguez * o maintain current state of phy error reception (the hal 378203c4805SLuis R. Rodriguez * may enable phy error frames for noise immunity work) 379203c4805SLuis R. Rodriguez * o probe request frames are accepted only when operating in 380203c4805SLuis R. Rodriguez * hostap, adhoc, or monitor modes 381203c4805SLuis R. Rodriguez * o enable promiscuous mode according to the interface state 382203c4805SLuis R. Rodriguez * o accept beacons: 383203c4805SLuis R. Rodriguez * - when operating in adhoc mode so the 802.11 layer creates 384203c4805SLuis R. Rodriguez * node table entries for peers, 385203c4805SLuis R. Rodriguez * - when operating in station mode for collecting rssi data when 386203c4805SLuis R. Rodriguez * the station is otherwise quiet, or 387203c4805SLuis R. Rodriguez * - when operating as a repeater so we see repeater-sta beacons 388203c4805SLuis R. Rodriguez * - when scanning 389203c4805SLuis R. Rodriguez */ 390203c4805SLuis R. Rodriguez 391203c4805SLuis R. Rodriguez u32 ath_calcrxfilter(struct ath_softc *sc) 392203c4805SLuis R. Rodriguez { 393203c4805SLuis R. Rodriguez u32 rfilt; 394203c4805SLuis R. Rodriguez 395ac06697cSFelix Fietkau rfilt = ATH9K_RX_FILTER_UCAST | ATH9K_RX_FILTER_BCAST 396203c4805SLuis R. Rodriguez | ATH9K_RX_FILTER_MCAST; 397203c4805SLuis R. Rodriguez 3989c1d8e4aSJouni Malinen if (sc->rx.rxfilter & FIF_PROBE_REQ) 399203c4805SLuis R. Rodriguez rfilt |= ATH9K_RX_FILTER_PROBEREQ; 400203c4805SLuis R. Rodriguez 401203c4805SLuis R. Rodriguez /* 402203c4805SLuis R. Rodriguez * Set promiscuous mode when FIF_PROMISC_IN_BSS is enabled for station 403203c4805SLuis R. Rodriguez * mode interface or when in monitor mode. AP mode does not need this 404203c4805SLuis R. Rodriguez * since it receives all in-BSS frames anyway. 405203c4805SLuis R. Rodriguez */ 4062e286947SFelix Fietkau if (sc->sc_ah->is_monitoring) 407203c4805SLuis R. Rodriguez rfilt |= ATH9K_RX_FILTER_PROM; 408203c4805SLuis R. Rodriguez 409203c4805SLuis R. Rodriguez if (sc->rx.rxfilter & FIF_CONTROL) 410203c4805SLuis R. Rodriguez rfilt |= ATH9K_RX_FILTER_CONTROL; 411203c4805SLuis R. Rodriguez 412203c4805SLuis R. Rodriguez if ((sc->sc_ah->opmode == NL80211_IFTYPE_STATION) && 413cfda6695SBen Greear (sc->nvifs <= 1) && 414203c4805SLuis R. Rodriguez !(sc->rx.rxfilter & FIF_BCN_PRBRESP_PROMISC)) 415203c4805SLuis R. Rodriguez rfilt |= ATH9K_RX_FILTER_MYBEACON; 416203c4805SLuis R. Rodriguez else 417203c4805SLuis R. Rodriguez rfilt |= ATH9K_RX_FILTER_BEACON; 418203c4805SLuis R. Rodriguez 419264bbec8SFelix Fietkau if ((sc->sc_ah->opmode == NL80211_IFTYPE_AP) || 42066afad01SSenthil Balasubramanian (sc->rx.rxfilter & FIF_PSPOLL)) 421203c4805SLuis R. Rodriguez rfilt |= ATH9K_RX_FILTER_PSPOLL; 422203c4805SLuis R. Rodriguez 4237ea310beSSujith if (conf_is_ht(&sc->hw->conf)) 4247ea310beSSujith rfilt |= ATH9K_RX_FILTER_COMP_BAR; 4257ea310beSSujith 4267545daf4SFelix Fietkau if (sc->nvifs > 1 || (sc->rx.rxfilter & FIF_OTHER_BSS)) { 4275eb6ba83SJavier Cardona /* The following may also be needed for other older chips */ 4285eb6ba83SJavier Cardona if (sc->sc_ah->hw_version.macVersion == AR_SREV_VERSION_9160) 4295eb6ba83SJavier Cardona rfilt |= ATH9K_RX_FILTER_PROM; 430203c4805SLuis R. Rodriguez rfilt |= ATH9K_RX_FILTER_MCAST_BCAST_ALL; 431203c4805SLuis R. Rodriguez } 432203c4805SLuis R. Rodriguez 433203c4805SLuis R. Rodriguez return rfilt; 434203c4805SLuis R. Rodriguez 435203c4805SLuis R. Rodriguez } 436203c4805SLuis R. Rodriguez 437203c4805SLuis R. Rodriguez int ath_startrecv(struct ath_softc *sc) 438203c4805SLuis R. Rodriguez { 439203c4805SLuis R. Rodriguez struct ath_hw *ah = sc->sc_ah; 440203c4805SLuis R. Rodriguez struct ath_buf *bf, *tbf; 441203c4805SLuis R. Rodriguez 442b5c80475SFelix Fietkau if (ah->caps.hw_caps & ATH9K_HW_CAP_EDMA) { 443b5c80475SFelix Fietkau ath_edma_start_recv(sc); 444b5c80475SFelix Fietkau return 0; 445b5c80475SFelix Fietkau } 446b5c80475SFelix Fietkau 447203c4805SLuis R. Rodriguez spin_lock_bh(&sc->rx.rxbuflock); 448203c4805SLuis R. Rodriguez if (list_empty(&sc->rx.rxbuf)) 449203c4805SLuis R. Rodriguez goto start_recv; 450203c4805SLuis R. Rodriguez 451203c4805SLuis R. Rodriguez sc->rx.rxlink = NULL; 452203c4805SLuis R. Rodriguez list_for_each_entry_safe(bf, tbf, &sc->rx.rxbuf, list) { 453203c4805SLuis R. Rodriguez ath_rx_buf_link(sc, bf); 454203c4805SLuis R. Rodriguez } 455203c4805SLuis R. Rodriguez 456203c4805SLuis R. Rodriguez /* We could have deleted elements so the list may be empty now */ 457203c4805SLuis R. Rodriguez if (list_empty(&sc->rx.rxbuf)) 458203c4805SLuis R. Rodriguez goto start_recv; 459203c4805SLuis R. Rodriguez 460203c4805SLuis R. Rodriguez bf = list_first_entry(&sc->rx.rxbuf, struct ath_buf, list); 461203c4805SLuis R. Rodriguez ath9k_hw_putrxbuf(ah, bf->bf_daddr); 462203c4805SLuis R. Rodriguez ath9k_hw_rxena(ah); 463203c4805SLuis R. Rodriguez 464203c4805SLuis R. Rodriguez start_recv: 465203c4805SLuis R. Rodriguez ath_opmode_init(sc); 4664cb54fa3SSujith Manoharan ath9k_hw_startpcureceive(ah, !!(sc->hw->conf.flags & IEEE80211_CONF_OFFCHANNEL)); 467203c4805SLuis R. Rodriguez 4687583c550SLuis R. Rodriguez spin_unlock_bh(&sc->rx.rxbuflock); 4697583c550SLuis R. Rodriguez 470203c4805SLuis R. Rodriguez return 0; 471203c4805SLuis R. Rodriguez } 472203c4805SLuis R. Rodriguez 473203c4805SLuis R. Rodriguez bool ath_stoprecv(struct ath_softc *sc) 474203c4805SLuis R. Rodriguez { 475203c4805SLuis R. Rodriguez struct ath_hw *ah = sc->sc_ah; 4765882da02SFelix Fietkau bool stopped, reset = false; 477203c4805SLuis R. Rodriguez 4781e450285SLuis R. Rodriguez spin_lock_bh(&sc->rx.rxbuflock); 479d47844a0SFelix Fietkau ath9k_hw_abortpcurecv(ah); 480203c4805SLuis R. Rodriguez ath9k_hw_setrxfilter(ah, 0); 4815882da02SFelix Fietkau stopped = ath9k_hw_stopdmarecv(ah, &reset); 482b5c80475SFelix Fietkau 483b5c80475SFelix Fietkau if (sc->sc_ah->caps.hw_caps & ATH9K_HW_CAP_EDMA) 484b5c80475SFelix Fietkau ath_edma_stop_recv(sc); 485b5c80475SFelix Fietkau else 486203c4805SLuis R. Rodriguez sc->rx.rxlink = NULL; 4871e450285SLuis R. Rodriguez spin_unlock_bh(&sc->rx.rxbuflock); 488203c4805SLuis R. Rodriguez 489d584747bSRajkumar Manoharan if (!(ah->ah_flags & AH_UNPLUGGED) && 490d584747bSRajkumar Manoharan unlikely(!stopped)) { 491d7fd1b50SBen Greear ath_err(ath9k_hw_common(sc->sc_ah), 492d7fd1b50SBen Greear "Could not stop RX, we could be " 49378a7685eSLuis R. Rodriguez "confusing the DMA engine when we start RX up\n"); 494d7fd1b50SBen Greear ATH_DBG_WARN_ON_ONCE(!stopped); 495d7fd1b50SBen Greear } 4962232d31bSFelix Fietkau return stopped && !reset; 497203c4805SLuis R. Rodriguez } 498203c4805SLuis R. Rodriguez 499203c4805SLuis R. Rodriguez void ath_flushrecv(struct ath_softc *sc) 500203c4805SLuis R. Rodriguez { 501781b14a3SSujith Manoharan set_bit(SC_OP_RXFLUSH, &sc->sc_flags); 502b5c80475SFelix Fietkau if (sc->sc_ah->caps.hw_caps & ATH9K_HW_CAP_EDMA) 503b5c80475SFelix Fietkau ath_rx_tasklet(sc, 1, true); 504b5c80475SFelix Fietkau ath_rx_tasklet(sc, 1, false); 505781b14a3SSujith Manoharan clear_bit(SC_OP_RXFLUSH, &sc->sc_flags); 506203c4805SLuis R. Rodriguez } 507203c4805SLuis R. Rodriguez 508cc65965cSJouni Malinen static bool ath_beacon_dtim_pending_cab(struct sk_buff *skb) 509cc65965cSJouni Malinen { 510cc65965cSJouni Malinen /* Check whether the Beacon frame has DTIM indicating buffered bc/mc */ 511cc65965cSJouni Malinen struct ieee80211_mgmt *mgmt; 512cc65965cSJouni Malinen u8 *pos, *end, id, elen; 513cc65965cSJouni Malinen struct ieee80211_tim_ie *tim; 514cc65965cSJouni Malinen 515cc65965cSJouni Malinen mgmt = (struct ieee80211_mgmt *)skb->data; 516cc65965cSJouni Malinen pos = mgmt->u.beacon.variable; 517cc65965cSJouni Malinen end = skb->data + skb->len; 518cc65965cSJouni Malinen 519cc65965cSJouni Malinen while (pos + 2 < end) { 520cc65965cSJouni Malinen id = *pos++; 521cc65965cSJouni Malinen elen = *pos++; 522cc65965cSJouni Malinen if (pos + elen > end) 523cc65965cSJouni Malinen break; 524cc65965cSJouni Malinen 525cc65965cSJouni Malinen if (id == WLAN_EID_TIM) { 526cc65965cSJouni Malinen if (elen < sizeof(*tim)) 527cc65965cSJouni Malinen break; 528cc65965cSJouni Malinen tim = (struct ieee80211_tim_ie *) pos; 529cc65965cSJouni Malinen if (tim->dtim_count != 0) 530cc65965cSJouni Malinen break; 531cc65965cSJouni Malinen return tim->bitmap_ctrl & 0x01; 532cc65965cSJouni Malinen } 533cc65965cSJouni Malinen 534cc65965cSJouni Malinen pos += elen; 535cc65965cSJouni Malinen } 536cc65965cSJouni Malinen 537cc65965cSJouni Malinen return false; 538cc65965cSJouni Malinen } 539cc65965cSJouni Malinen 540cc65965cSJouni Malinen static void ath_rx_ps_beacon(struct ath_softc *sc, struct sk_buff *skb) 541cc65965cSJouni Malinen { 5421510718dSLuis R. Rodriguez struct ath_common *common = ath9k_hw_common(sc->sc_ah); 543cc65965cSJouni Malinen 544cc65965cSJouni Malinen if (skb->len < 24 + 8 + 2 + 2) 545cc65965cSJouni Malinen return; 546cc65965cSJouni Malinen 5471b04b930SSujith sc->ps_flags &= ~PS_WAIT_FOR_BEACON; 548293dc5dfSGabor Juhos 5491b04b930SSujith if (sc->ps_flags & PS_BEACON_SYNC) { 5501b04b930SSujith sc->ps_flags &= ~PS_BEACON_SYNC; 551d2182b69SJoe Perches ath_dbg(common, PS, 552226afe68SJoe Perches "Reconfigure Beacon timers based on timestamp from the AP\n"); 55399e4d43aSRajkumar Manoharan ath_set_beacon(sc); 554ccdfeab6SJouni Malinen } 555ccdfeab6SJouni Malinen 556cc65965cSJouni Malinen if (ath_beacon_dtim_pending_cab(skb)) { 557cc65965cSJouni Malinen /* 558cc65965cSJouni Malinen * Remain awake waiting for buffered broadcast/multicast 55958f5fffdSGabor Juhos * frames. If the last broadcast/multicast frame is not 56058f5fffdSGabor Juhos * received properly, the next beacon frame will work as 56158f5fffdSGabor Juhos * a backup trigger for returning into NETWORK SLEEP state, 56258f5fffdSGabor Juhos * so we are waiting for it as well. 563cc65965cSJouni Malinen */ 564d2182b69SJoe Perches ath_dbg(common, PS, 565226afe68SJoe Perches "Received DTIM beacon indicating buffered broadcast/multicast frame(s)\n"); 5661b04b930SSujith sc->ps_flags |= PS_WAIT_FOR_CAB | PS_WAIT_FOR_BEACON; 567cc65965cSJouni Malinen return; 568cc65965cSJouni Malinen } 569cc65965cSJouni Malinen 5701b04b930SSujith if (sc->ps_flags & PS_WAIT_FOR_CAB) { 571cc65965cSJouni Malinen /* 572cc65965cSJouni Malinen * This can happen if a broadcast frame is dropped or the AP 573cc65965cSJouni Malinen * fails to send a frame indicating that all CAB frames have 574cc65965cSJouni Malinen * been delivered. 575cc65965cSJouni Malinen */ 5761b04b930SSujith sc->ps_flags &= ~PS_WAIT_FOR_CAB; 577d2182b69SJoe Perches ath_dbg(common, PS, "PS wait for CAB frames timed out\n"); 578cc65965cSJouni Malinen } 579cc65965cSJouni Malinen } 580cc65965cSJouni Malinen 581f73c604cSRajkumar Manoharan static void ath_rx_ps(struct ath_softc *sc, struct sk_buff *skb, bool mybeacon) 582cc65965cSJouni Malinen { 583cc65965cSJouni Malinen struct ieee80211_hdr *hdr; 584c46917bbSLuis R. Rodriguez struct ath_common *common = ath9k_hw_common(sc->sc_ah); 585cc65965cSJouni Malinen 586cc65965cSJouni Malinen hdr = (struct ieee80211_hdr *)skb->data; 587cc65965cSJouni Malinen 588cc65965cSJouni Malinen /* Process Beacon and CAB receive in PS state */ 589ededf1f8SVasanthakumar Thiagarajan if (((sc->ps_flags & PS_WAIT_FOR_BEACON) || ath9k_check_auto_sleep(sc)) 590*07c15a3fSSujith Manoharan && mybeacon) { 591cc65965cSJouni Malinen ath_rx_ps_beacon(sc, skb); 592*07c15a3fSSujith Manoharan } else if ((sc->ps_flags & PS_WAIT_FOR_CAB) && 593cc65965cSJouni Malinen (ieee80211_is_data(hdr->frame_control) || 594cc65965cSJouni Malinen ieee80211_is_action(hdr->frame_control)) && 595cc65965cSJouni Malinen is_multicast_ether_addr(hdr->addr1) && 596cc65965cSJouni Malinen !ieee80211_has_moredata(hdr->frame_control)) { 597cc65965cSJouni Malinen /* 598cc65965cSJouni Malinen * No more broadcast/multicast frames to be received at this 599cc65965cSJouni Malinen * point. 600cc65965cSJouni Malinen */ 6013fac6dfdSSenthil Balasubramanian sc->ps_flags &= ~(PS_WAIT_FOR_CAB | PS_WAIT_FOR_BEACON); 602d2182b69SJoe Perches ath_dbg(common, PS, 603c46917bbSLuis R. Rodriguez "All PS CAB frames received, back to sleep\n"); 6041b04b930SSujith } else if ((sc->ps_flags & PS_WAIT_FOR_PSPOLL_DATA) && 6059a23f9caSJouni Malinen !is_multicast_ether_addr(hdr->addr1) && 6069a23f9caSJouni Malinen !ieee80211_has_morefrags(hdr->frame_control)) { 6071b04b930SSujith sc->ps_flags &= ~PS_WAIT_FOR_PSPOLL_DATA; 608d2182b69SJoe Perches ath_dbg(common, PS, 609226afe68SJoe Perches "Going back to sleep after having received PS-Poll data (0x%lx)\n", 6101b04b930SSujith sc->ps_flags & (PS_WAIT_FOR_BEACON | 6111b04b930SSujith PS_WAIT_FOR_CAB | 6121b04b930SSujith PS_WAIT_FOR_PSPOLL_DATA | 6131b04b930SSujith PS_WAIT_FOR_TX_ACK)); 614cc65965cSJouni Malinen } 615cc65965cSJouni Malinen } 616cc65965cSJouni Malinen 617b5c80475SFelix Fietkau static bool ath_edma_get_buffers(struct ath_softc *sc, 6183a2923e8SFelix Fietkau enum ath9k_rx_qtype qtype, 6193a2923e8SFelix Fietkau struct ath_rx_status *rs, 6203a2923e8SFelix Fietkau struct ath_buf **dest) 621203c4805SLuis R. Rodriguez { 622b5c80475SFelix Fietkau struct ath_rx_edma *rx_edma = &sc->rx.rx_edma[qtype]; 623203c4805SLuis R. Rodriguez struct ath_hw *ah = sc->sc_ah; 62427c51f1aSLuis R. Rodriguez struct ath_common *common = ath9k_hw_common(ah); 625b5c80475SFelix Fietkau struct sk_buff *skb; 626b5c80475SFelix Fietkau struct ath_buf *bf; 627b5c80475SFelix Fietkau int ret; 628203c4805SLuis R. Rodriguez 629b5c80475SFelix Fietkau skb = skb_peek(&rx_edma->rx_fifo); 630b5c80475SFelix Fietkau if (!skb) 631b5c80475SFelix Fietkau return false; 632203c4805SLuis R. Rodriguez 633b5c80475SFelix Fietkau bf = SKB_CB_ATHBUF(skb); 634b5c80475SFelix Fietkau BUG_ON(!bf); 635b5c80475SFelix Fietkau 636ce9426d1SMing Lei dma_sync_single_for_cpu(sc->dev, bf->bf_buf_addr, 637b5c80475SFelix Fietkau common->rx_bufsize, DMA_FROM_DEVICE); 638b5c80475SFelix Fietkau 6393a2923e8SFelix Fietkau ret = ath9k_hw_process_rxdesc_edma(ah, rs, skb->data); 640ce9426d1SMing Lei if (ret == -EINPROGRESS) { 641ce9426d1SMing Lei /*let device gain the buffer again*/ 642ce9426d1SMing Lei dma_sync_single_for_device(sc->dev, bf->bf_buf_addr, 643ce9426d1SMing Lei common->rx_bufsize, DMA_FROM_DEVICE); 644b5c80475SFelix Fietkau return false; 645ce9426d1SMing Lei } 646b5c80475SFelix Fietkau 647b5c80475SFelix Fietkau __skb_unlink(skb, &rx_edma->rx_fifo); 648b5c80475SFelix Fietkau if (ret == -EINVAL) { 649b5c80475SFelix Fietkau /* corrupt descriptor, skip this one and the following one */ 650b5c80475SFelix Fietkau list_add_tail(&bf->list, &sc->rx.rxbuf); 651b5c80475SFelix Fietkau ath_rx_edma_buf_link(sc, qtype); 652b5c80475SFelix Fietkau 6533a2923e8SFelix Fietkau skb = skb_peek(&rx_edma->rx_fifo); 6543a2923e8SFelix Fietkau if (skb) { 655b5c80475SFelix Fietkau bf = SKB_CB_ATHBUF(skb); 656b5c80475SFelix Fietkau BUG_ON(!bf); 657b5c80475SFelix Fietkau 658b5c80475SFelix Fietkau __skb_unlink(skb, &rx_edma->rx_fifo); 659b5c80475SFelix Fietkau list_add_tail(&bf->list, &sc->rx.rxbuf); 660b5c80475SFelix Fietkau ath_rx_edma_buf_link(sc, qtype); 6613a2923e8SFelix Fietkau } else { 6623a2923e8SFelix Fietkau bf = NULL; 663b5c80475SFelix Fietkau } 6643a2923e8SFelix Fietkau } 665b5c80475SFelix Fietkau 6663a2923e8SFelix Fietkau *dest = bf; 667b5c80475SFelix Fietkau return true; 668b5c80475SFelix Fietkau } 669b5c80475SFelix Fietkau 670b5c80475SFelix Fietkau static struct ath_buf *ath_edma_get_next_rx_buf(struct ath_softc *sc, 671b5c80475SFelix Fietkau struct ath_rx_status *rs, 672b5c80475SFelix Fietkau enum ath9k_rx_qtype qtype) 673b5c80475SFelix Fietkau { 6743a2923e8SFelix Fietkau struct ath_buf *bf = NULL; 675b5c80475SFelix Fietkau 6763a2923e8SFelix Fietkau while (ath_edma_get_buffers(sc, qtype, rs, &bf)) { 6773a2923e8SFelix Fietkau if (!bf) 6783a2923e8SFelix Fietkau continue; 679b5c80475SFelix Fietkau 680b5c80475SFelix Fietkau return bf; 681b5c80475SFelix Fietkau } 6823a2923e8SFelix Fietkau return NULL; 6833a2923e8SFelix Fietkau } 684b5c80475SFelix Fietkau 685b5c80475SFelix Fietkau static struct ath_buf *ath_get_next_rx_buf(struct ath_softc *sc, 686b5c80475SFelix Fietkau struct ath_rx_status *rs) 687b5c80475SFelix Fietkau { 688b5c80475SFelix Fietkau struct ath_hw *ah = sc->sc_ah; 689b5c80475SFelix Fietkau struct ath_common *common = ath9k_hw_common(ah); 690b5c80475SFelix Fietkau struct ath_desc *ds; 691b5c80475SFelix Fietkau struct ath_buf *bf; 692b5c80475SFelix Fietkau int ret; 693203c4805SLuis R. Rodriguez 694203c4805SLuis R. Rodriguez if (list_empty(&sc->rx.rxbuf)) { 695203c4805SLuis R. Rodriguez sc->rx.rxlink = NULL; 696b5c80475SFelix Fietkau return NULL; 697203c4805SLuis R. Rodriguez } 698203c4805SLuis R. Rodriguez 699203c4805SLuis R. Rodriguez bf = list_first_entry(&sc->rx.rxbuf, struct ath_buf, list); 700203c4805SLuis R. Rodriguez ds = bf->bf_desc; 701203c4805SLuis R. Rodriguez 702203c4805SLuis R. Rodriguez /* 703203c4805SLuis R. Rodriguez * Must provide the virtual address of the current 704203c4805SLuis R. Rodriguez * descriptor, the physical address, and the virtual 705203c4805SLuis R. Rodriguez * address of the next descriptor in the h/w chain. 706203c4805SLuis R. Rodriguez * This allows the HAL to look ahead to see if the 707203c4805SLuis R. Rodriguez * hardware is done with a descriptor by checking the 708203c4805SLuis R. Rodriguez * done bit in the following descriptor and the address 709203c4805SLuis R. Rodriguez * of the current descriptor the DMA engine is working 710203c4805SLuis R. Rodriguez * on. All this is necessary because of our use of 711203c4805SLuis R. Rodriguez * a self-linked list to avoid rx overruns. 712203c4805SLuis R. Rodriguez */ 7133de21116SRajkumar Manoharan ret = ath9k_hw_rxprocdesc(ah, ds, rs); 714b5c80475SFelix Fietkau if (ret == -EINPROGRESS) { 71529bffa96SFelix Fietkau struct ath_rx_status trs; 716203c4805SLuis R. Rodriguez struct ath_buf *tbf; 717203c4805SLuis R. Rodriguez struct ath_desc *tds; 718203c4805SLuis R. Rodriguez 71929bffa96SFelix Fietkau memset(&trs, 0, sizeof(trs)); 720203c4805SLuis R. Rodriguez if (list_is_last(&bf->list, &sc->rx.rxbuf)) { 721203c4805SLuis R. Rodriguez sc->rx.rxlink = NULL; 722b5c80475SFelix Fietkau return NULL; 723203c4805SLuis R. Rodriguez } 724203c4805SLuis R. Rodriguez 725203c4805SLuis R. Rodriguez tbf = list_entry(bf->list.next, struct ath_buf, list); 726203c4805SLuis R. Rodriguez 727203c4805SLuis R. Rodriguez /* 728203c4805SLuis R. Rodriguez * On some hardware the descriptor status words could 729203c4805SLuis R. Rodriguez * get corrupted, including the done bit. Because of 730203c4805SLuis R. Rodriguez * this, check if the next descriptor's done bit is 731203c4805SLuis R. Rodriguez * set or not. 732203c4805SLuis R. Rodriguez * 733203c4805SLuis R. Rodriguez * If the next descriptor's done bit is set, the current 734203c4805SLuis R. Rodriguez * descriptor has been corrupted. Force s/w to discard 735203c4805SLuis R. Rodriguez * this descriptor and continue... 736203c4805SLuis R. Rodriguez */ 737203c4805SLuis R. Rodriguez 738203c4805SLuis R. Rodriguez tds = tbf->bf_desc; 7393de21116SRajkumar Manoharan ret = ath9k_hw_rxprocdesc(ah, tds, &trs); 740b5c80475SFelix Fietkau if (ret == -EINPROGRESS) 741b5c80475SFelix Fietkau return NULL; 742203c4805SLuis R. Rodriguez } 743203c4805SLuis R. Rodriguez 744b5c80475SFelix Fietkau if (!bf->bf_mpdu) 745b5c80475SFelix Fietkau return bf; 746203c4805SLuis R. Rodriguez 747203c4805SLuis R. Rodriguez /* 748203c4805SLuis R. Rodriguez * Synchronize the DMA transfer with CPU before 749203c4805SLuis R. Rodriguez * 1. accessing the frame 750203c4805SLuis R. Rodriguez * 2. requeueing the same buffer to h/w 751203c4805SLuis R. Rodriguez */ 752ce9426d1SMing Lei dma_sync_single_for_cpu(sc->dev, bf->bf_buf_addr, 753cc861f74SLuis R. Rodriguez common->rx_bufsize, 754203c4805SLuis R. Rodriguez DMA_FROM_DEVICE); 755203c4805SLuis R. Rodriguez 756b5c80475SFelix Fietkau return bf; 757b5c80475SFelix Fietkau } 758b5c80475SFelix Fietkau 759d435700fSSujith /* Assumes you've already done the endian to CPU conversion */ 760d435700fSSujith static bool ath9k_rx_accept(struct ath_common *common, 7619f167f64SVasanthakumar Thiagarajan struct ieee80211_hdr *hdr, 762d435700fSSujith struct ieee80211_rx_status *rxs, 763d435700fSSujith struct ath_rx_status *rx_stats, 764d435700fSSujith bool *decrypt_error) 765d435700fSSujith { 766ec205999SFelix Fietkau struct ath_softc *sc = (struct ath_softc *) common->priv; 76766760eacSFelix Fietkau bool is_mc, is_valid_tkip, strip_mic, mic_error; 768d435700fSSujith struct ath_hw *ah = common->ah; 769d435700fSSujith __le16 fc; 770b7b1b512SVasanthakumar Thiagarajan u8 rx_status_len = ah->caps.rx_status_len; 771d435700fSSujith 772d435700fSSujith fc = hdr->frame_control; 773d435700fSSujith 77466760eacSFelix Fietkau is_mc = !!is_multicast_ether_addr(hdr->addr1); 77566760eacSFelix Fietkau is_valid_tkip = rx_stats->rs_keyix != ATH9K_RXKEYIX_INVALID && 77666760eacSFelix Fietkau test_bit(rx_stats->rs_keyix, common->tkip_keymap); 777152e585dSBill Jordan strip_mic = is_valid_tkip && ieee80211_is_data(fc) && 7782a5783b8SMichael Liang ieee80211_has_protected(fc) && 779152e585dSBill Jordan !(rx_stats->rs_status & 780846d9363SFelix Fietkau (ATH9K_RXERR_DECRYPT | ATH9K_RXERR_CRC | ATH9K_RXERR_MIC | 781846d9363SFelix Fietkau ATH9K_RXERR_KEYMISS)); 78266760eacSFelix Fietkau 783f88373faSFelix Fietkau /* 784f88373faSFelix Fietkau * Key miss events are only relevant for pairwise keys where the 785f88373faSFelix Fietkau * descriptor does contain a valid key index. This has been observed 786f88373faSFelix Fietkau * mostly with CCMP encryption. 787f88373faSFelix Fietkau */ 788f88373faSFelix Fietkau if (rx_stats->rs_keyix == ATH9K_RXKEYIX_INVALID) 789f88373faSFelix Fietkau rx_stats->rs_status &= ~ATH9K_RXERR_KEYMISS; 790f88373faSFelix Fietkau 79115072189SBen Greear if (!rx_stats->rs_datalen) { 79215072189SBen Greear RX_STAT_INC(rx_len_err); 793d435700fSSujith return false; 79415072189SBen Greear } 79515072189SBen Greear 796d435700fSSujith /* 797d435700fSSujith * rs_status follows rs_datalen so if rs_datalen is too large 798d435700fSSujith * we can take a hint that hardware corrupted it, so ignore 799d435700fSSujith * those frames. 800d435700fSSujith */ 80115072189SBen Greear if (rx_stats->rs_datalen > (common->rx_bufsize - rx_status_len)) { 80215072189SBen Greear RX_STAT_INC(rx_len_err); 803d435700fSSujith return false; 80415072189SBen Greear } 805d435700fSSujith 8060d95521eSFelix Fietkau /* Only use error bits from the last fragment */ 807d435700fSSujith if (rx_stats->rs_more) 8080d95521eSFelix Fietkau return true; 809d435700fSSujith 81066760eacSFelix Fietkau mic_error = is_valid_tkip && !ieee80211_is_ctl(fc) && 81166760eacSFelix Fietkau !ieee80211_has_morefrags(fc) && 81266760eacSFelix Fietkau !(le16_to_cpu(hdr->seq_ctrl) & IEEE80211_SCTL_FRAG) && 81366760eacSFelix Fietkau (rx_stats->rs_status & ATH9K_RXERR_MIC); 81466760eacSFelix Fietkau 815d435700fSSujith /* 816d435700fSSujith * The rx_stats->rs_status will not be set until the end of the 817d435700fSSujith * chained descriptors so it can be ignored if rs_more is set. The 818d435700fSSujith * rs_more will be false at the last element of the chained 819d435700fSSujith * descriptors. 820d435700fSSujith */ 821d435700fSSujith if (rx_stats->rs_status != 0) { 822846d9363SFelix Fietkau u8 status_mask; 823846d9363SFelix Fietkau 82466760eacSFelix Fietkau if (rx_stats->rs_status & ATH9K_RXERR_CRC) { 825d435700fSSujith rxs->flag |= RX_FLAG_FAILED_FCS_CRC; 82666760eacSFelix Fietkau mic_error = false; 82766760eacSFelix Fietkau } 828d435700fSSujith if (rx_stats->rs_status & ATH9K_RXERR_PHY) 829d435700fSSujith return false; 830d435700fSSujith 831846d9363SFelix Fietkau if ((rx_stats->rs_status & ATH9K_RXERR_DECRYPT) || 832846d9363SFelix Fietkau (!is_mc && (rx_stats->rs_status & ATH9K_RXERR_KEYMISS))) { 833d435700fSSujith *decrypt_error = true; 83466760eacSFelix Fietkau mic_error = false; 835d435700fSSujith } 83666760eacSFelix Fietkau 837d435700fSSujith /* 838d435700fSSujith * Reject error frames with the exception of 839d435700fSSujith * decryption and MIC failures. For monitor mode, 840d435700fSSujith * we also ignore the CRC error. 841d435700fSSujith */ 842846d9363SFelix Fietkau status_mask = ATH9K_RXERR_DECRYPT | ATH9K_RXERR_MIC | 843846d9363SFelix Fietkau ATH9K_RXERR_KEYMISS; 844846d9363SFelix Fietkau 845ec205999SFelix Fietkau if (ah->is_monitoring && (sc->rx.rxfilter & FIF_FCSFAIL)) 846846d9363SFelix Fietkau status_mask |= ATH9K_RXERR_CRC; 847846d9363SFelix Fietkau 848846d9363SFelix Fietkau if (rx_stats->rs_status & ~status_mask) 849d435700fSSujith return false; 850d435700fSSujith } 85166760eacSFelix Fietkau 85266760eacSFelix Fietkau /* 85366760eacSFelix Fietkau * For unicast frames the MIC error bit can have false positives, 85466760eacSFelix Fietkau * so all MIC error reports need to be validated in software. 85566760eacSFelix Fietkau * False negatives are not common, so skip software verification 85666760eacSFelix Fietkau * if the hardware considers the MIC valid. 85766760eacSFelix Fietkau */ 85866760eacSFelix Fietkau if (strip_mic) 85966760eacSFelix Fietkau rxs->flag |= RX_FLAG_MMIC_STRIPPED; 86066760eacSFelix Fietkau else if (is_mc && mic_error) 86166760eacSFelix Fietkau rxs->flag |= RX_FLAG_MMIC_ERROR; 86266760eacSFelix Fietkau 863d435700fSSujith return true; 864d435700fSSujith } 865d435700fSSujith 866d435700fSSujith static int ath9k_process_rate(struct ath_common *common, 867d435700fSSujith struct ieee80211_hw *hw, 868d435700fSSujith struct ath_rx_status *rx_stats, 8699f167f64SVasanthakumar Thiagarajan struct ieee80211_rx_status *rxs) 870d435700fSSujith { 871d435700fSSujith struct ieee80211_supported_band *sband; 872d435700fSSujith enum ieee80211_band band; 873d435700fSSujith unsigned int i = 0; 874990e08a0SBen Greear struct ath_softc __maybe_unused *sc = common->priv; 875d435700fSSujith 876d435700fSSujith band = hw->conf.channel->band; 877d435700fSSujith sband = hw->wiphy->bands[band]; 878d435700fSSujith 879d435700fSSujith if (rx_stats->rs_rate & 0x80) { 880d435700fSSujith /* HT rate */ 881d435700fSSujith rxs->flag |= RX_FLAG_HT; 882d435700fSSujith if (rx_stats->rs_flags & ATH9K_RX_2040) 883d435700fSSujith rxs->flag |= RX_FLAG_40MHZ; 884d435700fSSujith if (rx_stats->rs_flags & ATH9K_RX_GI) 885d435700fSSujith rxs->flag |= RX_FLAG_SHORT_GI; 886d435700fSSujith rxs->rate_idx = rx_stats->rs_rate & 0x7f; 887d435700fSSujith return 0; 888d435700fSSujith } 889d435700fSSujith 890d435700fSSujith for (i = 0; i < sband->n_bitrates; i++) { 891d435700fSSujith if (sband->bitrates[i].hw_value == rx_stats->rs_rate) { 892d435700fSSujith rxs->rate_idx = i; 893d435700fSSujith return 0; 894d435700fSSujith } 895d435700fSSujith if (sband->bitrates[i].hw_value_short == rx_stats->rs_rate) { 896d435700fSSujith rxs->flag |= RX_FLAG_SHORTPRE; 897d435700fSSujith rxs->rate_idx = i; 898d435700fSSujith return 0; 899d435700fSSujith } 900d435700fSSujith } 901d435700fSSujith 902d435700fSSujith /* 903d435700fSSujith * No valid hardware bitrate found -- we should not get here 904d435700fSSujith * because hardware has already validated this frame as OK. 905d435700fSSujith */ 906d2182b69SJoe Perches ath_dbg(common, ANY, 907226afe68SJoe Perches "unsupported hw bitrate detected 0x%02x using 1 Mbit\n", 908226afe68SJoe Perches rx_stats->rs_rate); 90915072189SBen Greear RX_STAT_INC(rx_rate_err); 910d435700fSSujith return -EINVAL; 911d435700fSSujith } 912d435700fSSujith 913d435700fSSujith static void ath9k_process_rssi(struct ath_common *common, 914d435700fSSujith struct ieee80211_hw *hw, 9159f167f64SVasanthakumar Thiagarajan struct ieee80211_hdr *hdr, 916d435700fSSujith struct ath_rx_status *rx_stats) 917d435700fSSujith { 9189ac58615SFelix Fietkau struct ath_softc *sc = hw->priv; 919d435700fSSujith struct ath_hw *ah = common->ah; 9209fa23e17SFelix Fietkau int last_rssi; 9212ef16755SFelix Fietkau int rssi = rx_stats->rs_rssi; 922d435700fSSujith 923cf3af748SRajkumar Manoharan if (!rx_stats->is_mybeacon || 924cf3af748SRajkumar Manoharan ((ah->opmode != NL80211_IFTYPE_STATION) && 925cf3af748SRajkumar Manoharan (ah->opmode != NL80211_IFTYPE_ADHOC))) 9269fa23e17SFelix Fietkau return; 9279fa23e17SFelix Fietkau 9289fa23e17SFelix Fietkau if (rx_stats->rs_rssi != ATH9K_RSSI_BAD && !rx_stats->rs_moreaggr) 9299ac58615SFelix Fietkau ATH_RSSI_LPF(sc->last_rssi, rx_stats->rs_rssi); 930686b9cb9SBen Greear 9319ac58615SFelix Fietkau last_rssi = sc->last_rssi; 932d435700fSSujith if (likely(last_rssi != ATH_RSSI_DUMMY_MARKER)) 9332ef16755SFelix Fietkau rssi = ATH_EP_RND(last_rssi, ATH_RSSI_EP_MULTIPLIER); 9342ef16755SFelix Fietkau if (rssi < 0) 9352ef16755SFelix Fietkau rssi = 0; 936d435700fSSujith 937d435700fSSujith /* Update Beacon RSSI, this is used by ANI. */ 9382ef16755SFelix Fietkau ah->stats.avgbrssi = rssi; 939d435700fSSujith } 940d435700fSSujith 941d435700fSSujith /* 942d435700fSSujith * For Decrypt or Demic errors, we only mark packet status here and always push 943d435700fSSujith * up the frame up to let mac80211 handle the actual error case, be it no 944d435700fSSujith * decryption key or real decryption error. This let us keep statistics there. 945d435700fSSujith */ 946d435700fSSujith static int ath9k_rx_skb_preprocess(struct ath_common *common, 947d435700fSSujith struct ieee80211_hw *hw, 9489f167f64SVasanthakumar Thiagarajan struct ieee80211_hdr *hdr, 949d435700fSSujith struct ath_rx_status *rx_stats, 950d435700fSSujith struct ieee80211_rx_status *rx_status, 951d435700fSSujith bool *decrypt_error) 952d435700fSSujith { 953f749b946SFelix Fietkau struct ath_hw *ah = common->ah; 954f749b946SFelix Fietkau 955d435700fSSujith /* 956d435700fSSujith * everything but the rate is checked here, the rate check is done 957d435700fSSujith * separately to avoid doing two lookups for a rate for each frame. 958d435700fSSujith */ 9599f167f64SVasanthakumar Thiagarajan if (!ath9k_rx_accept(common, hdr, rx_status, rx_stats, decrypt_error)) 960d435700fSSujith return -EINVAL; 961d435700fSSujith 9620d95521eSFelix Fietkau /* Only use status info from the last fragment */ 9630d95521eSFelix Fietkau if (rx_stats->rs_more) 9640d95521eSFelix Fietkau return 0; 9650d95521eSFelix Fietkau 9669f167f64SVasanthakumar Thiagarajan ath9k_process_rssi(common, hw, hdr, rx_stats); 967d435700fSSujith 9689f167f64SVasanthakumar Thiagarajan if (ath9k_process_rate(common, hw, rx_stats, rx_status)) 969d435700fSSujith return -EINVAL; 970d435700fSSujith 971d435700fSSujith rx_status->band = hw->conf.channel->band; 972d435700fSSujith rx_status->freq = hw->conf.channel->center_freq; 973f749b946SFelix Fietkau rx_status->signal = ah->noise + rx_stats->rs_rssi; 974d435700fSSujith rx_status->antenna = rx_stats->rs_antenna; 9756ebacbb7SJohannes Berg rx_status->flag |= RX_FLAG_MACTIME_MPDU; 9762ef16755SFelix Fietkau if (rx_stats->rs_moreaggr) 9772ef16755SFelix Fietkau rx_status->flag |= RX_FLAG_NO_SIGNAL_VAL; 978d435700fSSujith 979d435700fSSujith return 0; 980d435700fSSujith } 981d435700fSSujith 982d435700fSSujith static void ath9k_rx_skb_postprocess(struct ath_common *common, 983d435700fSSujith struct sk_buff *skb, 984d435700fSSujith struct ath_rx_status *rx_stats, 985d435700fSSujith struct ieee80211_rx_status *rxs, 986d435700fSSujith bool decrypt_error) 987d435700fSSujith { 988d435700fSSujith struct ath_hw *ah = common->ah; 989d435700fSSujith struct ieee80211_hdr *hdr; 990d435700fSSujith int hdrlen, padpos, padsize; 991d435700fSSujith u8 keyix; 992d435700fSSujith __le16 fc; 993d435700fSSujith 994d435700fSSujith /* see if any padding is done by the hw and remove it */ 995d435700fSSujith hdr = (struct ieee80211_hdr *) skb->data; 996d435700fSSujith hdrlen = ieee80211_get_hdrlen_from_skb(skb); 997d435700fSSujith fc = hdr->frame_control; 998d435700fSSujith padpos = ath9k_cmn_padpos(hdr->frame_control); 999d435700fSSujith 1000d435700fSSujith /* The MAC header is padded to have 32-bit boundary if the 1001d435700fSSujith * packet payload is non-zero. The general calculation for 1002d435700fSSujith * padsize would take into account odd header lengths: 1003d435700fSSujith * padsize = (4 - padpos % 4) % 4; However, since only 1004d435700fSSujith * even-length headers are used, padding can only be 0 or 2 1005d435700fSSujith * bytes and we can optimize this a bit. In addition, we must 1006d435700fSSujith * not try to remove padding from short control frames that do 1007d435700fSSujith * not have payload. */ 1008d435700fSSujith padsize = padpos & 3; 1009d435700fSSujith if (padsize && skb->len>=padpos+padsize+FCS_LEN) { 1010d435700fSSujith memmove(skb->data + padsize, skb->data, padpos); 1011d435700fSSujith skb_pull(skb, padsize); 1012d435700fSSujith } 1013d435700fSSujith 1014d435700fSSujith keyix = rx_stats->rs_keyix; 1015d435700fSSujith 1016d435700fSSujith if (!(keyix == ATH9K_RXKEYIX_INVALID) && !decrypt_error && 1017d435700fSSujith ieee80211_has_protected(fc)) { 1018d435700fSSujith rxs->flag |= RX_FLAG_DECRYPTED; 1019d435700fSSujith } else if (ieee80211_has_protected(fc) 1020d435700fSSujith && !decrypt_error && skb->len >= hdrlen + 4) { 1021d435700fSSujith keyix = skb->data[hdrlen + 3] >> 6; 1022d435700fSSujith 1023d435700fSSujith if (test_bit(keyix, common->keymap)) 1024d435700fSSujith rxs->flag |= RX_FLAG_DECRYPTED; 1025d435700fSSujith } 1026d435700fSSujith if (ah->sw_mgmt_crypto && 1027d435700fSSujith (rxs->flag & RX_FLAG_DECRYPTED) && 1028d435700fSSujith ieee80211_is_mgmt(fc)) 1029d435700fSSujith /* Use software decrypt for management frames. */ 1030d435700fSSujith rxs->flag &= ~RX_FLAG_DECRYPTED; 1031d435700fSSujith } 1032b5c80475SFelix Fietkau 1033b5c80475SFelix Fietkau int ath_rx_tasklet(struct ath_softc *sc, int flush, bool hp) 1034b5c80475SFelix Fietkau { 1035b5c80475SFelix Fietkau struct ath_buf *bf; 10360d95521eSFelix Fietkau struct sk_buff *skb = NULL, *requeue_skb, *hdr_skb; 1037b5c80475SFelix Fietkau struct ieee80211_rx_status *rxs; 1038b5c80475SFelix Fietkau struct ath_hw *ah = sc->sc_ah; 1039b5c80475SFelix Fietkau struct ath_common *common = ath9k_hw_common(ah); 10407545daf4SFelix Fietkau struct ieee80211_hw *hw = sc->hw; 1041b5c80475SFelix Fietkau struct ieee80211_hdr *hdr; 1042b5c80475SFelix Fietkau int retval; 1043b5c80475SFelix Fietkau bool decrypt_error = false; 1044b5c80475SFelix Fietkau struct ath_rx_status rs; 1045b5c80475SFelix Fietkau enum ath9k_rx_qtype qtype; 1046b5c80475SFelix Fietkau bool edma = !!(ah->caps.hw_caps & ATH9K_HW_CAP_EDMA); 1047b5c80475SFelix Fietkau int dma_type; 10485c6dd921SVasanthakumar Thiagarajan u8 rx_status_len = ah->caps.rx_status_len; 1049a6d2055bSFelix Fietkau u64 tsf = 0; 1050a6d2055bSFelix Fietkau u32 tsf_lower = 0; 10518ab2cd09SLuis R. Rodriguez unsigned long flags; 1052b5c80475SFelix Fietkau 1053b5c80475SFelix Fietkau if (edma) 1054b5c80475SFelix Fietkau dma_type = DMA_BIDIRECTIONAL; 105556824223SMing Lei else 105656824223SMing Lei dma_type = DMA_FROM_DEVICE; 1057b5c80475SFelix Fietkau 1058b5c80475SFelix Fietkau qtype = hp ? ATH9K_RX_QUEUE_HP : ATH9K_RX_QUEUE_LP; 1059b5c80475SFelix Fietkau spin_lock_bh(&sc->rx.rxbuflock); 1060b5c80475SFelix Fietkau 1061a6d2055bSFelix Fietkau tsf = ath9k_hw_gettsf64(ah); 1062a6d2055bSFelix Fietkau tsf_lower = tsf & 0xffffffff; 1063a6d2055bSFelix Fietkau 1064b5c80475SFelix Fietkau do { 1065b5c80475SFelix Fietkau /* If handling rx interrupt and flush is in progress => exit */ 1066781b14a3SSujith Manoharan if (test_bit(SC_OP_RXFLUSH, &sc->sc_flags) && (flush == 0)) 1067b5c80475SFelix Fietkau break; 1068b5c80475SFelix Fietkau 1069b5c80475SFelix Fietkau memset(&rs, 0, sizeof(rs)); 1070b5c80475SFelix Fietkau if (edma) 1071b5c80475SFelix Fietkau bf = ath_edma_get_next_rx_buf(sc, &rs, qtype); 1072b5c80475SFelix Fietkau else 1073b5c80475SFelix Fietkau bf = ath_get_next_rx_buf(sc, &rs); 1074b5c80475SFelix Fietkau 1075b5c80475SFelix Fietkau if (!bf) 1076b5c80475SFelix Fietkau break; 1077b5c80475SFelix Fietkau 1078b5c80475SFelix Fietkau skb = bf->bf_mpdu; 1079b5c80475SFelix Fietkau if (!skb) 1080b5c80475SFelix Fietkau continue; 1081b5c80475SFelix Fietkau 10820d95521eSFelix Fietkau /* 10830d95521eSFelix Fietkau * Take frame header from the first fragment and RX status from 10840d95521eSFelix Fietkau * the last one. 10850d95521eSFelix Fietkau */ 10860d95521eSFelix Fietkau if (sc->rx.frag) 10870d95521eSFelix Fietkau hdr_skb = sc->rx.frag; 10880d95521eSFelix Fietkau else 10890d95521eSFelix Fietkau hdr_skb = skb; 10900d95521eSFelix Fietkau 10910d95521eSFelix Fietkau hdr = (struct ieee80211_hdr *) (hdr_skb->data + rx_status_len); 10920d95521eSFelix Fietkau rxs = IEEE80211_SKB_RXCB(hdr_skb); 109315072189SBen Greear if (ieee80211_is_beacon(hdr->frame_control)) { 109415072189SBen Greear RX_STAT_INC(rx_beacons); 109515072189SBen Greear if (!is_zero_ether_addr(common->curbssid) && 10962e42e474SJoe Perches ether_addr_equal(hdr->addr3, common->curbssid)) 1097cf3af748SRajkumar Manoharan rs.is_mybeacon = true; 1098cf3af748SRajkumar Manoharan else 1099cf3af748SRajkumar Manoharan rs.is_mybeacon = false; 110015072189SBen Greear } 110115072189SBen Greear else 110215072189SBen Greear rs.is_mybeacon = false; 11035ca42627SLuis R. Rodriguez 11046995fb80SRajkumar Manoharan sc->rx.num_pkts++; 110529bffa96SFelix Fietkau ath_debug_stat_rx(sc, &rs); 11061395d3f0SSujith 1107203c4805SLuis R. Rodriguez /* 1108203c4805SLuis R. Rodriguez * If we're asked to flush receive queue, directly 1109203c4805SLuis R. Rodriguez * chain it back at the queue without processing it. 1110203c4805SLuis R. Rodriguez */ 1111781b14a3SSujith Manoharan if (test_bit(SC_OP_RXFLUSH, &sc->sc_flags)) { 111215072189SBen Greear RX_STAT_INC(rx_drop_rxflush); 11130d95521eSFelix Fietkau goto requeue_drop_frag; 111415072189SBen Greear } 1115203c4805SLuis R. Rodriguez 1116ffb1c56aSAshok Nagarajan memset(rxs, 0, sizeof(struct ieee80211_rx_status)); 1117ffb1c56aSAshok Nagarajan 1118a6d2055bSFelix Fietkau rxs->mactime = (tsf & ~0xffffffffULL) | rs.rs_tstamp; 1119a6d2055bSFelix Fietkau if (rs.rs_tstamp > tsf_lower && 1120a6d2055bSFelix Fietkau unlikely(rs.rs_tstamp - tsf_lower > 0x10000000)) 1121a6d2055bSFelix Fietkau rxs->mactime -= 0x100000000ULL; 1122a6d2055bSFelix Fietkau 1123a6d2055bSFelix Fietkau if (rs.rs_tstamp < tsf_lower && 1124a6d2055bSFelix Fietkau unlikely(tsf_lower - rs.rs_tstamp > 0x10000000)) 1125a6d2055bSFelix Fietkau rxs->mactime += 0x100000000ULL; 1126a6d2055bSFelix Fietkau 112783c76570SZefir Kurtisi retval = ath9k_rx_skb_preprocess(common, hw, hdr, &rs, 112883c76570SZefir Kurtisi rxs, &decrypt_error); 112983c76570SZefir Kurtisi if (retval) 113083c76570SZefir Kurtisi goto requeue_drop_frag; 113183c76570SZefir Kurtisi 113201e18918SRajkumar Manoharan if (rs.is_mybeacon) { 113301e18918SRajkumar Manoharan sc->hw_busy_count = 0; 113401e18918SRajkumar Manoharan ath_start_rx_poll(sc, 3); 113501e18918SRajkumar Manoharan } 1136203c4805SLuis R. Rodriguez /* Ensure we always have an skb to requeue once we are done 1137203c4805SLuis R. Rodriguez * processing the current buffer's skb */ 1138cc861f74SLuis R. Rodriguez requeue_skb = ath_rxbuf_alloc(common, common->rx_bufsize, GFP_ATOMIC); 1139203c4805SLuis R. Rodriguez 1140203c4805SLuis R. Rodriguez /* If there is no memory we ignore the current RX'd frame, 1141203c4805SLuis R. Rodriguez * tell hardware it can give us a new frame using the old 1142203c4805SLuis R. Rodriguez * skb and put it at the tail of the sc->rx.rxbuf list for 1143203c4805SLuis R. Rodriguez * processing. */ 114415072189SBen Greear if (!requeue_skb) { 114515072189SBen Greear RX_STAT_INC(rx_oom_err); 11460d95521eSFelix Fietkau goto requeue_drop_frag; 114715072189SBen Greear } 1148203c4805SLuis R. Rodriguez 1149203c4805SLuis R. Rodriguez /* Unmap the frame */ 1150203c4805SLuis R. Rodriguez dma_unmap_single(sc->dev, bf->bf_buf_addr, 1151cc861f74SLuis R. Rodriguez common->rx_bufsize, 1152b5c80475SFelix Fietkau dma_type); 1153203c4805SLuis R. Rodriguez 1154b5c80475SFelix Fietkau skb_put(skb, rs.rs_datalen + ah->caps.rx_status_len); 1155b5c80475SFelix Fietkau if (ah->caps.rx_status_len) 1156b5c80475SFelix Fietkau skb_pull(skb, ah->caps.rx_status_len); 1157203c4805SLuis R. Rodriguez 11580d95521eSFelix Fietkau if (!rs.rs_more) 11590d95521eSFelix Fietkau ath9k_rx_skb_postprocess(common, hdr_skb, &rs, 1160c9b14170SLuis R. Rodriguez rxs, decrypt_error); 1161203c4805SLuis R. Rodriguez 1162203c4805SLuis R. Rodriguez /* We will now give hardware our shiny new allocated skb */ 1163203c4805SLuis R. Rodriguez bf->bf_mpdu = requeue_skb; 1164203c4805SLuis R. Rodriguez bf->bf_buf_addr = dma_map_single(sc->dev, requeue_skb->data, 1165cc861f74SLuis R. Rodriguez common->rx_bufsize, 1166b5c80475SFelix Fietkau dma_type); 1167203c4805SLuis R. Rodriguez if (unlikely(dma_mapping_error(sc->dev, 1168203c4805SLuis R. Rodriguez bf->bf_buf_addr))) { 1169203c4805SLuis R. Rodriguez dev_kfree_skb_any(requeue_skb); 1170203c4805SLuis R. Rodriguez bf->bf_mpdu = NULL; 11716cf9e995SBen Greear bf->bf_buf_addr = 0; 11723800276aSJoe Perches ath_err(common, "dma_mapping_error() on RX\n"); 11737545daf4SFelix Fietkau ieee80211_rx(hw, skb); 1174203c4805SLuis R. Rodriguez break; 1175203c4805SLuis R. Rodriguez } 1176203c4805SLuis R. Rodriguez 11770d95521eSFelix Fietkau if (rs.rs_more) { 117815072189SBen Greear RX_STAT_INC(rx_frags); 11790d95521eSFelix Fietkau /* 11800d95521eSFelix Fietkau * rs_more indicates chained descriptors which can be 11810d95521eSFelix Fietkau * used to link buffers together for a sort of 11820d95521eSFelix Fietkau * scatter-gather operation. 11830d95521eSFelix Fietkau */ 11840d95521eSFelix Fietkau if (sc->rx.frag) { 11850d95521eSFelix Fietkau /* too many fragments - cannot handle frame */ 11860d95521eSFelix Fietkau dev_kfree_skb_any(sc->rx.frag); 11870d95521eSFelix Fietkau dev_kfree_skb_any(skb); 118815072189SBen Greear RX_STAT_INC(rx_too_many_frags_err); 11890d95521eSFelix Fietkau skb = NULL; 11900d95521eSFelix Fietkau } 11910d95521eSFelix Fietkau sc->rx.frag = skb; 11920d95521eSFelix Fietkau goto requeue; 11930d95521eSFelix Fietkau } 11940d95521eSFelix Fietkau 11950d95521eSFelix Fietkau if (sc->rx.frag) { 11960d95521eSFelix Fietkau int space = skb->len - skb_tailroom(hdr_skb); 11970d95521eSFelix Fietkau 11980d95521eSFelix Fietkau if (pskb_expand_head(hdr_skb, 0, space, GFP_ATOMIC) < 0) { 11990d95521eSFelix Fietkau dev_kfree_skb(skb); 120015072189SBen Greear RX_STAT_INC(rx_oom_err); 12010d95521eSFelix Fietkau goto requeue_drop_frag; 12020d95521eSFelix Fietkau } 12030d95521eSFelix Fietkau 1204b5447ff9SEric Dumazet sc->rx.frag = NULL; 1205b5447ff9SEric Dumazet 12060d95521eSFelix Fietkau skb_copy_from_linear_data(skb, skb_put(hdr_skb, skb->len), 12070d95521eSFelix Fietkau skb->len); 12080d95521eSFelix Fietkau dev_kfree_skb_any(skb); 12090d95521eSFelix Fietkau skb = hdr_skb; 12100d95521eSFelix Fietkau } 12110d95521eSFelix Fietkau 1212eb840a80SMohammed Shafi Shajakhan 1213eb840a80SMohammed Shafi Shajakhan if (ah->caps.hw_caps & ATH9K_HW_CAP_ANT_DIV_COMB) { 1214eb840a80SMohammed Shafi Shajakhan 1215203c4805SLuis R. Rodriguez /* 1216eb840a80SMohammed Shafi Shajakhan * change the default rx antenna if rx diversity 1217eb840a80SMohammed Shafi Shajakhan * chooses the other antenna 3 times in a row. 1218203c4805SLuis R. Rodriguez */ 121929bffa96SFelix Fietkau if (sc->rx.defant != rs.rs_antenna) { 1220203c4805SLuis R. Rodriguez if (++sc->rx.rxotherant >= 3) 122129bffa96SFelix Fietkau ath_setdefantenna(sc, rs.rs_antenna); 1222203c4805SLuis R. Rodriguez } else { 1223203c4805SLuis R. Rodriguez sc->rx.rxotherant = 0; 1224203c4805SLuis R. Rodriguez } 1225203c4805SLuis R. Rodriguez 1226eb840a80SMohammed Shafi Shajakhan } 1227eb840a80SMohammed Shafi Shajakhan 122866760eacSFelix Fietkau if (rxs->flag & RX_FLAG_MMIC_STRIPPED) 122966760eacSFelix Fietkau skb_trim(skb, skb->len - 8); 123066760eacSFelix Fietkau 12318ab2cd09SLuis R. Rodriguez spin_lock_irqsave(&sc->sc_pm_lock, flags); 1232aaef24b4SMohammed Shafi Shajakhan if ((sc->ps_flags & (PS_WAIT_FOR_BEACON | 12331b04b930SSujith PS_WAIT_FOR_CAB | 1234aaef24b4SMohammed Shafi Shajakhan PS_WAIT_FOR_PSPOLL_DATA)) || 1235cedc7e3dSMohammed Shafi Shajakhan ath9k_check_auto_sleep(sc)) 1236f73c604cSRajkumar Manoharan ath_rx_ps(sc, skb, rs.is_mybeacon); 12378ab2cd09SLuis R. Rodriguez spin_unlock_irqrestore(&sc->sc_pm_lock, flags); 1238cc65965cSJouni Malinen 123943c35284SFelix Fietkau if ((ah->caps.hw_caps & ATH9K_HW_CAP_ANT_DIV_COMB) && sc->ant_rx == 3) 1240102885a5SVasanthakumar Thiagarajan ath_ant_comb_scan(sc, &rs); 1241102885a5SVasanthakumar Thiagarajan 12427545daf4SFelix Fietkau ieee80211_rx(hw, skb); 1243cc65965cSJouni Malinen 12440d95521eSFelix Fietkau requeue_drop_frag: 12450d95521eSFelix Fietkau if (sc->rx.frag) { 12460d95521eSFelix Fietkau dev_kfree_skb_any(sc->rx.frag); 12470d95521eSFelix Fietkau sc->rx.frag = NULL; 12480d95521eSFelix Fietkau } 1249203c4805SLuis R. Rodriguez requeue: 1250b5c80475SFelix Fietkau if (edma) { 1251b5c80475SFelix Fietkau list_add_tail(&bf->list, &sc->rx.rxbuf); 1252b5c80475SFelix Fietkau ath_rx_edma_buf_link(sc, qtype); 1253b5c80475SFelix Fietkau } else { 1254203c4805SLuis R. Rodriguez list_move_tail(&bf->list, &sc->rx.rxbuf); 1255203c4805SLuis R. Rodriguez ath_rx_buf_link(sc, bf); 12563483288cSFelix Fietkau if (!flush) 125795294973SFelix Fietkau ath9k_hw_rxena(ah); 1258b5c80475SFelix Fietkau } 1259203c4805SLuis R. Rodriguez } while (1); 1260203c4805SLuis R. Rodriguez 1261203c4805SLuis R. Rodriguez spin_unlock_bh(&sc->rx.rxbuflock); 1262203c4805SLuis R. Rodriguez 126329ab0b36SRajkumar Manoharan if (!(ah->imask & ATH9K_INT_RXEOL)) { 126429ab0b36SRajkumar Manoharan ah->imask |= (ATH9K_INT_RXEOL | ATH9K_INT_RXORN); 126572d874c6SFelix Fietkau ath9k_hw_set_interrupts(ah); 126629ab0b36SRajkumar Manoharan } 126729ab0b36SRajkumar Manoharan 1268203c4805SLuis R. Rodriguez return 0; 1269203c4805SLuis R. Rodriguez } 1270