2 * Copyright (c) 2010-2011 Atheros Communications Inc.
4 * Permission to use, copy, modify, and/or distribute this software for any
5 * purpose with or without fee is hereby granted, provided that the above
6 * copyright notice and this permission notice appear in all copies.
8 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
9 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
10 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
11 * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
12 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
13 * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
14 * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
16 #include <linux/export.h>
18 #include "ar9003_mac.h"
19 #include "ar9003_mci.h"
21 static void ar9003_hw_rx_enable(struct ath_hw *hw)
23 REG_WRITE(hw, AR_CR, 0);
27 ar9003_set_txdesc(struct ath_hw *ah, void *ds, struct ath_tx_info *i)
29 struct ar9003_txc *ads = ds;
31 u32 val, ctl12, ctl17;
34 desc_len = ((AR_SREV_9462(ah) || AR_SREV_9565(ah)) ? 0x18 : 0x17);
36 val = (ATHEROS_VENDOR_ID << AR_DescId_S) |
37 (1 << AR_TxRxDesc_S) |
38 (1 << AR_CtrlStat_S) |
39 (i->qcu << AR_TxQcuNum_S) | desc_len;
42 WRITE_ONCE(ads->info, val);
45 WRITE_ONCE(ads->link, i->link);
47 checksum += i->buf_addr[0];
48 WRITE_ONCE(ads->data0, i->buf_addr[0]);
49 checksum += i->buf_addr[1];
50 WRITE_ONCE(ads->data1, i->buf_addr[1]);
51 checksum += i->buf_addr[2];
52 WRITE_ONCE(ads->data2, i->buf_addr[2]);
53 checksum += i->buf_addr[3];
54 WRITE_ONCE(ads->data3, i->buf_addr[3]);
56 checksum += (val = (i->buf_len[0] << AR_BufLen_S) & AR_BufLen);
57 WRITE_ONCE(ads->ctl3, val);
58 checksum += (val = (i->buf_len[1] << AR_BufLen_S) & AR_BufLen);
59 WRITE_ONCE(ads->ctl5, val);
60 checksum += (val = (i->buf_len[2] << AR_BufLen_S) & AR_BufLen);
61 WRITE_ONCE(ads->ctl7, val);
62 checksum += (val = (i->buf_len[3] << AR_BufLen_S) & AR_BufLen);
63 WRITE_ONCE(ads->ctl9, val);
65 checksum = (u16) (((checksum & 0xffff) + (checksum >> 16)) & 0xffff);
66 WRITE_ONCE(ads->ctl10, checksum);
68 if (i->is_first || i->is_last) {
69 WRITE_ONCE(ads->ctl13, set11nTries(i->rates, 0)
70 | set11nTries(i->rates, 1)
71 | set11nTries(i->rates, 2)
72 | set11nTries(i->rates, 3)
73 | (i->dur_update ? AR_DurUpdateEna : 0)
74 | SM(0, AR_BurstDur));
76 WRITE_ONCE(ads->ctl14, set11nRate(i->rates, 0)
77 | set11nRate(i->rates, 1)
78 | set11nRate(i->rates, 2)
79 | set11nRate(i->rates, 3));
81 WRITE_ONCE(ads->ctl13, 0);
82 WRITE_ONCE(ads->ctl14, 0);
90 ctl17 = SM(i->keytype, AR_EncrType);
92 WRITE_ONCE(ads->ctl11, 0);
93 WRITE_ONCE(ads->ctl12, i->is_last ? 0 : AR_TxMore);
94 WRITE_ONCE(ads->ctl15, 0);
95 WRITE_ONCE(ads->ctl16, 0);
96 WRITE_ONCE(ads->ctl17, ctl17);
97 WRITE_ONCE(ads->ctl18, 0);
98 WRITE_ONCE(ads->ctl19, 0);
102 WRITE_ONCE(ads->ctl11, (i->pkt_len & AR_FrameLen)
103 | (i->flags & ATH9K_TXDESC_VMF ? AR_VirtMoreFrag : 0)
104 | SM(i->txpower[0], AR_XmitPower0)
105 | (i->flags & ATH9K_TXDESC_VEOL ? AR_VEOL : 0)
106 | (i->keyix != ATH9K_TXKEYIX_INVALID ? AR_DestIdxValid : 0)
107 | (i->flags & ATH9K_TXDESC_LOWRXCHAIN ? AR_LowRxChain : 0)
108 | (i->flags & ATH9K_TXDESC_CLRDMASK ? AR_ClrDestMask : 0)
109 | (i->flags & ATH9K_TXDESC_RTSENA ? AR_RTSEnable :
110 (i->flags & ATH9K_TXDESC_CTSENA ? AR_CTSEnable : 0)));
112 ctl12 = (i->keyix != ATH9K_TXKEYIX_INVALID ?
113 SM(i->keyix, AR_DestIdx) : 0)
114 | SM(i->type, AR_FrameType)
115 | (i->flags & ATH9K_TXDESC_NOACK ? AR_NoAck : 0)
116 | (i->flags & ATH9K_TXDESC_EXT_ONLY ? AR_ExtOnly : 0)
117 | (i->flags & ATH9K_TXDESC_EXT_AND_CTL ? AR_ExtAndCtl : 0);
119 ctl17 |= (i->flags & ATH9K_TXDESC_LDPC ? AR_LDPC : 0);
122 ctl17 |= SM(i->aggr_len, AR_AggrLen);
124 case AGGR_BUF_MIDDLE:
125 ctl12 |= AR_IsAggr | AR_MoreAggr;
126 ctl17 |= SM(i->ndelim, AR_PadDelim);
135 val = (i->flags & ATH9K_TXDESC_PAPRD) >> ATH9K_TXDESC_PAPRD_S;
136 ctl12 |= SM(val, AR_PAPRDChainMask);
138 WRITE_ONCE(ads->ctl12, ctl12);
139 WRITE_ONCE(ads->ctl17, ctl17);
141 WRITE_ONCE(ads->ctl15, set11nPktDurRTSCTS(i->rates, 0)
142 | set11nPktDurRTSCTS(i->rates, 1));
144 WRITE_ONCE(ads->ctl16, set11nPktDurRTSCTS(i->rates, 2)
145 | set11nPktDurRTSCTS(i->rates, 3));
147 WRITE_ONCE(ads->ctl18, set11nRateFlags(i->rates, 0)
148 | set11nRateFlags(i->rates, 1)
149 | set11nRateFlags(i->rates, 2)
150 | set11nRateFlags(i->rates, 3)
151 | SM(i->rtscts_rate, AR_RTSCTSRate));
153 WRITE_ONCE(ads->ctl19, AR_Not_Sounding);
155 WRITE_ONCE(ads->ctl20, SM(i->txpower[1], AR_XmitPower1));
156 WRITE_ONCE(ads->ctl21, SM(i->txpower[2], AR_XmitPower2));
157 WRITE_ONCE(ads->ctl22, SM(i->txpower[3], AR_XmitPower3));
160 static u16 ar9003_calc_ptr_chksum(struct ar9003_txc *ads)
164 checksum = ads->info + ads->link
165 + ads->data0 + ads->ctl3
166 + ads->data1 + ads->ctl5
167 + ads->data2 + ads->ctl7
168 + ads->data3 + ads->ctl9;
170 return ((checksum & 0xffff) + (checksum >> 16)) & AR_TxPtrChkSum;
173 static void ar9003_hw_set_desc_link(void *ds, u32 ds_link)
175 struct ar9003_txc *ads = ds;
178 ads->ctl10 &= ~AR_TxPtrChkSum;
179 ads->ctl10 |= ar9003_calc_ptr_chksum(ads);
182 static bool ar9003_hw_get_isr(struct ath_hw *ah, enum ath9k_int *masked,
187 struct ath9k_hw_capabilities *pCap = &ah->caps;
188 struct ath_common *common = ath9k_hw_common(ah);
189 u32 sync_cause = 0, async_cause, async_mask = AR_INTR_MAC_IRQ;
192 if (ath9k_hw_mci_is_enabled(ah))
193 async_mask |= AR_INTR_ASYNC_MASK_MCI;
195 async_cause = REG_READ(ah, AR_INTR_ASYNC_CAUSE);
197 if (async_cause & async_mask) {
198 if ((REG_READ(ah, AR_RTC_STATUS) & AR_RTC_STATUS_M)
200 isr = REG_READ(ah, AR_ISR);
204 sync_cause = REG_READ(ah, AR_INTR_SYNC_CAUSE) & AR_INTR_SYNC_DEFAULT;
208 if (!isr && !sync_cause && !async_cause)
212 if (isr & AR_ISR_BCNMISC) {
214 isr2 = REG_READ(ah, AR_ISR_S2);
216 mask2 |= ((isr2 & AR_ISR_S2_TIM) >>
218 mask2 |= ((isr2 & AR_ISR_S2_DTIM) >>
220 mask2 |= ((isr2 & AR_ISR_S2_DTIMSYNC) >>
221 MAP_ISR_S2_DTIMSYNC);
222 mask2 |= ((isr2 & AR_ISR_S2_CABEND) >>
224 mask2 |= ((isr2 & AR_ISR_S2_GTT) <<
226 mask2 |= ((isr2 & AR_ISR_S2_CST) <<
228 mask2 |= ((isr2 & AR_ISR_S2_TSFOOR) >>
230 mask2 |= ((isr2 & AR_ISR_S2_BB_WATCHDOG) >>
231 MAP_ISR_S2_BB_WATCHDOG);
233 if (!(pCap->hw_caps & ATH9K_HW_CAP_RAC_SUPPORTED)) {
234 REG_WRITE(ah, AR_ISR_S2, isr2);
235 isr &= ~AR_ISR_BCNMISC;
239 if ((pCap->hw_caps & ATH9K_HW_CAP_RAC_SUPPORTED))
240 isr = REG_READ(ah, AR_ISR_RAC);
242 if (isr == 0xffffffff) {
247 *masked = isr & ATH9K_INT_COMMON;
249 if (ah->config.rx_intr_mitigation)
250 if (isr & (AR_ISR_RXMINTR | AR_ISR_RXINTM))
251 *masked |= ATH9K_INT_RXLP;
253 if (ah->config.tx_intr_mitigation)
254 if (isr & (AR_ISR_TXMINTR | AR_ISR_TXINTM))
255 *masked |= ATH9K_INT_TX;
257 if (isr & (AR_ISR_LP_RXOK | AR_ISR_RXERR))
258 *masked |= ATH9K_INT_RXLP;
260 if (isr & AR_ISR_HP_RXOK)
261 *masked |= ATH9K_INT_RXHP;
263 if (isr & (AR_ISR_TXOK | AR_ISR_TXERR | AR_ISR_TXEOL)) {
264 *masked |= ATH9K_INT_TX;
266 if (!(pCap->hw_caps & ATH9K_HW_CAP_RAC_SUPPORTED)) {
268 s0 = REG_READ(ah, AR_ISR_S0);
269 REG_WRITE(ah, AR_ISR_S0, s0);
270 s1 = REG_READ(ah, AR_ISR_S1);
271 REG_WRITE(ah, AR_ISR_S1, s1);
273 isr &= ~(AR_ISR_TXOK | AR_ISR_TXERR |
278 if (isr & AR_ISR_GENTMR) {
281 if (pCap->hw_caps & ATH9K_HW_CAP_RAC_SUPPORTED)
282 s5 = REG_READ(ah, AR_ISR_S5_S);
284 s5 = REG_READ(ah, AR_ISR_S5);
286 ah->intr_gen_timer_trigger =
287 MS(s5, AR_ISR_S5_GENTIMER_TRIG);
289 ah->intr_gen_timer_thresh =
290 MS(s5, AR_ISR_S5_GENTIMER_THRESH);
292 if (ah->intr_gen_timer_trigger)
293 *masked |= ATH9K_INT_GENTIMER;
295 if (!(pCap->hw_caps & ATH9K_HW_CAP_RAC_SUPPORTED)) {
296 REG_WRITE(ah, AR_ISR_S5, s5);
297 isr &= ~AR_ISR_GENTMR;
304 if (!(pCap->hw_caps & ATH9K_HW_CAP_RAC_SUPPORTED)) {
305 REG_WRITE(ah, AR_ISR, isr);
307 (void) REG_READ(ah, AR_ISR);
310 if (*masked & ATH9K_INT_BB_WATCHDOG)
311 ar9003_hw_bb_watchdog_read(ah);
314 if (async_cause & AR_INTR_ASYNC_MASK_MCI)
315 ar9003_mci_get_isr(ah, masked);
319 *sync_cause_p = sync_cause;
322 (AR_INTR_SYNC_HOST1_FATAL | AR_INTR_SYNC_HOST1_PERR))
326 if (sync_cause & AR_INTR_SYNC_HOST1_FATAL) {
328 "received PCI FATAL interrupt\n");
330 if (sync_cause & AR_INTR_SYNC_HOST1_PERR) {
332 "received PCI PERR interrupt\n");
334 *masked |= ATH9K_INT_FATAL;
337 if (sync_cause & AR_INTR_SYNC_RADM_CPL_TIMEOUT) {
338 REG_WRITE(ah, AR_RC, AR_RC_HOSTIF);
339 REG_WRITE(ah, AR_RC, 0);
340 *masked |= ATH9K_INT_FATAL;
343 if (sync_cause & AR_INTR_SYNC_LOCAL_TIMEOUT)
344 ath_dbg(common, INTERRUPT,
345 "AR_INTR_SYNC_LOCAL_TIMEOUT\n");
347 REG_WRITE(ah, AR_INTR_SYNC_CAUSE_CLR, sync_cause);
348 (void) REG_READ(ah, AR_INTR_SYNC_CAUSE_CLR);
354 static int ar9003_hw_proc_txdesc(struct ath_hw *ah, void *ds,
355 struct ath_tx_status *ts)
357 struct ar9003_txs *ads;
360 ads = &ah->ts_ring[ah->ts_tail];
362 status = READ_ONCE(ads->status8);
363 if ((status & AR_TxDone) == 0)
366 ah->ts_tail = (ah->ts_tail + 1) % ah->ts_size;
368 if ((MS(ads->ds_info, AR_DescId) != ATHEROS_VENDOR_ID) ||
369 (MS(ads->ds_info, AR_TxRxDesc) != 1)) {
370 ath_dbg(ath9k_hw_common(ah), XMIT,
371 "Tx Descriptor error %x\n", ads->ds_info);
372 memset(ads, 0, sizeof(*ads));
376 ts->ts_rateindex = MS(status, AR_FinalTxIdx);
377 ts->ts_seqnum = MS(status, AR_SeqNum);
378 ts->tid = MS(status, AR_TxTid);
380 ts->qid = MS(ads->ds_info, AR_TxQcuNum);
381 ts->desc_id = MS(ads->status1, AR_TxDescId);
382 ts->ts_tstamp = ads->status4;
386 if (status & AR_TxOpExceeded)
387 ts->ts_status |= ATH9K_TXERR_XTXOP;
388 status = READ_ONCE(ads->status2);
389 ts->ts_rssi_ctl0 = MS(status, AR_TxRSSIAnt00);
390 ts->ts_rssi_ctl1 = MS(status, AR_TxRSSIAnt01);
391 ts->ts_rssi_ctl2 = MS(status, AR_TxRSSIAnt02);
392 if (status & AR_TxBaStatus) {
393 ts->ts_flags |= ATH9K_TX_BA;
394 ts->ba_low = ads->status5;
395 ts->ba_high = ads->status6;
398 status = READ_ONCE(ads->status3);
399 if (status & AR_ExcessiveRetries)
400 ts->ts_status |= ATH9K_TXERR_XRETRY;
401 if (status & AR_Filtered)
402 ts->ts_status |= ATH9K_TXERR_FILT;
403 if (status & AR_FIFOUnderrun) {
404 ts->ts_status |= ATH9K_TXERR_FIFO;
405 ath9k_hw_updatetxtriglevel(ah, true);
407 if (status & AR_TxTimerExpired)
408 ts->ts_status |= ATH9K_TXERR_TIMER_EXPIRED;
409 if (status & AR_DescCfgErr)
410 ts->ts_flags |= ATH9K_TX_DESC_CFG_ERR;
411 if (status & AR_TxDataUnderrun) {
412 ts->ts_flags |= ATH9K_TX_DATA_UNDERRUN;
413 ath9k_hw_updatetxtriglevel(ah, true);
415 if (status & AR_TxDelimUnderrun) {
416 ts->ts_flags |= ATH9K_TX_DELIM_UNDERRUN;
417 ath9k_hw_updatetxtriglevel(ah, true);
419 ts->ts_shortretry = MS(status, AR_RTSFailCnt);
420 ts->ts_longretry = MS(status, AR_DataFailCnt);
421 ts->ts_virtcol = MS(status, AR_VirtRetryCnt);
423 status = READ_ONCE(ads->status7);
424 ts->ts_rssi = MS(status, AR_TxRSSICombined);
425 ts->ts_rssi_ext0 = MS(status, AR_TxRSSIAnt10);
426 ts->ts_rssi_ext1 = MS(status, AR_TxRSSIAnt11);
427 ts->ts_rssi_ext2 = MS(status, AR_TxRSSIAnt12);
429 memset(ads, 0, sizeof(*ads));
434 static int ar9003_hw_get_duration(struct ath_hw *ah, const void *ds, int index)
436 const struct ar9003_txc *adc = ds;
440 return MS(READ_ONCE(adc->ctl15), AR_PacketDur0);
442 return MS(READ_ONCE(adc->ctl15), AR_PacketDur1);
444 return MS(READ_ONCE(adc->ctl16), AR_PacketDur2);
446 return MS(READ_ONCE(adc->ctl16), AR_PacketDur3);
452 void ar9003_hw_attach_mac_ops(struct ath_hw *hw)
454 struct ath_hw_ops *ops = ath9k_hw_ops(hw);
456 ops->rx_enable = ar9003_hw_rx_enable;
457 ops->set_desc_link = ar9003_hw_set_desc_link;
458 ops->get_isr = ar9003_hw_get_isr;
459 ops->set_txdesc = ar9003_set_txdesc;
460 ops->proc_txdesc = ar9003_hw_proc_txdesc;
461 ops->get_duration = ar9003_hw_get_duration;
464 void ath9k_hw_set_rx_bufsize(struct ath_hw *ah, u16 buf_size)
466 REG_WRITE(ah, AR_DATABUF_SIZE, buf_size & AR_DATABUF_SIZE_MASK);
468 EXPORT_SYMBOL(ath9k_hw_set_rx_bufsize);
470 void ath9k_hw_addrxbuf_edma(struct ath_hw *ah, u32 rxdp,
471 enum ath9k_rx_qtype qtype)
473 if (qtype == ATH9K_RX_QUEUE_HP)
474 REG_WRITE(ah, AR_HP_RXDP, rxdp);
476 REG_WRITE(ah, AR_LP_RXDP, rxdp);
478 EXPORT_SYMBOL(ath9k_hw_addrxbuf_edma);
480 int ath9k_hw_process_rxdesc_edma(struct ath_hw *ah, struct ath_rx_status *rxs,
483 struct ar9003_rxs *rxsp = buf_addr;
486 if ((rxsp->status11 & AR_RxDone) == 0)
489 if (MS(rxsp->ds_info, AR_DescId) != 0x168c)
492 if ((rxsp->ds_info & (AR_TxRxDesc | AR_CtrlStat)) != 0)
498 rxs->bw = RATE_INFO_BW_20;
500 rxs->rs_datalen = rxsp->status2 & AR_DataLen;
501 rxs->rs_tstamp = rxsp->status3;
504 rxs->rs_rssi = MS(rxsp->status5, AR_RxRSSICombined);
505 rxs->rs_rssi_ctl[0] = MS(rxsp->status1, AR_RxRSSIAnt00);
506 rxs->rs_rssi_ctl[1] = MS(rxsp->status1, AR_RxRSSIAnt01);
507 rxs->rs_rssi_ctl[2] = MS(rxsp->status1, AR_RxRSSIAnt02);
508 rxs->rs_rssi_ext[0] = MS(rxsp->status5, AR_RxRSSIAnt10);
509 rxs->rs_rssi_ext[1] = MS(rxsp->status5, AR_RxRSSIAnt11);
510 rxs->rs_rssi_ext[2] = MS(rxsp->status5, AR_RxRSSIAnt12);
512 if (rxsp->status11 & AR_RxKeyIdxValid)
513 rxs->rs_keyix = MS(rxsp->status11, AR_KeyIdx);
515 rxs->rs_keyix = ATH9K_RXKEYIX_INVALID;
517 rxs->rs_rate = MS(rxsp->status1, AR_RxRate);
518 rxs->rs_more = (rxsp->status2 & AR_RxMore) ? 1 : 0;
520 rxs->rs_firstaggr = (rxsp->status11 & AR_RxFirstAggr) ? 1 : 0;
521 rxs->rs_isaggr = (rxsp->status11 & AR_RxAggr) ? 1 : 0;
522 rxs->rs_moreaggr = (rxsp->status11 & AR_RxMoreAggr) ? 1 : 0;
523 rxs->rs_antenna = (MS(rxsp->status4, AR_RxAntenna) & 0x7);
524 rxs->enc_flags |= (rxsp->status4 & AR_GI) ? RX_ENC_FLAG_SHORT_GI : 0;
525 rxs->bw = (rxsp->status4 & AR_2040) ? RATE_INFO_BW_40 : RATE_INFO_BW_20;
527 rxs->evm0 = rxsp->status6;
528 rxs->evm1 = rxsp->status7;
529 rxs->evm2 = rxsp->status8;
530 rxs->evm3 = rxsp->status9;
531 rxs->evm4 = (rxsp->status10 & 0xffff);
533 if (rxsp->status11 & AR_PreDelimCRCErr)
534 rxs->rs_flags |= ATH9K_RX_DELIM_CRC_PRE;
536 if (rxsp->status11 & AR_PostDelimCRCErr)
537 rxs->rs_flags |= ATH9K_RX_DELIM_CRC_POST;
539 if (rxsp->status11 & AR_DecryptBusyErr)
540 rxs->rs_flags |= ATH9K_RX_DECRYPT_BUSY;
542 if ((rxsp->status11 & AR_RxFrameOK) == 0) {
544 * AR_CRCErr will bet set to true if we're on the last
545 * subframe and the AR_PostDelimCRCErr is caught.
546 * In a way this also gives us a guarantee that when
547 * (!(AR_CRCErr) && (AR_PostDelimCRCErr)) we cannot
548 * possibly be reviewing the last subframe. AR_CRCErr
549 * is the CRC of the actual data.
551 if (rxsp->status11 & AR_CRCErr)
552 rxs->rs_status |= ATH9K_RXERR_CRC;
553 else if (rxsp->status11 & AR_DecryptCRCErr)
554 rxs->rs_status |= ATH9K_RXERR_DECRYPT;
555 else if (rxsp->status11 & AR_MichaelErr)
556 rxs->rs_status |= ATH9K_RXERR_MIC;
557 if (rxsp->status11 & AR_PHYErr) {
558 phyerr = MS(rxsp->status11, AR_PHYErrCode);
560 * If we reach a point here where AR_PostDelimCRCErr is
561 * true it implies we're *not* on the last subframe. In
562 * in that case that we know already that the CRC of
563 * the frame was OK, and MAC would send an ACK for that
564 * subframe, even if we did get a phy error of type
565 * ATH9K_PHYERR_OFDM_RESTART. This is only applicable
566 * to frame that are prior to the last subframe.
567 * The AR_PostDelimCRCErr is the CRC for the MPDU
568 * delimiter, which contains the 4 reserved bits,
569 * the MPDU length (12 bits), and follows the MPDU
570 * delimiter for an A-MPDU subframe (0x4E = 'N' ASCII).
572 if ((phyerr == ATH9K_PHYERR_OFDM_RESTART) &&
573 (rxsp->status11 & AR_PostDelimCRCErr)) {
576 rxs->rs_status |= ATH9K_RXERR_PHY;
577 rxs->rs_phyerr = phyerr;
582 if (rxsp->status11 & AR_KeyMiss)
583 rxs->rs_status |= ATH9K_RXERR_KEYMISS;
587 EXPORT_SYMBOL(ath9k_hw_process_rxdesc_edma);
589 void ath9k_hw_reset_txstatus_ring(struct ath_hw *ah)
593 memset((void *) ah->ts_ring, 0,
594 ah->ts_size * sizeof(struct ar9003_txs));
596 ath_dbg(ath9k_hw_common(ah), XMIT,
597 "TS Start 0x%x End 0x%x Virt %p, Size %d\n",
598 ah->ts_paddr_start, ah->ts_paddr_end,
599 ah->ts_ring, ah->ts_size);
601 REG_WRITE(ah, AR_Q_STATUS_RING_START, ah->ts_paddr_start);
602 REG_WRITE(ah, AR_Q_STATUS_RING_END, ah->ts_paddr_end);
605 void ath9k_hw_setup_statusring(struct ath_hw *ah, void *ts_start,
610 ah->ts_paddr_start = ts_paddr_start;
611 ah->ts_paddr_end = ts_paddr_start + (size * sizeof(struct ar9003_txs));
613 ah->ts_ring = ts_start;
615 ath9k_hw_reset_txstatus_ring(ah);
617 EXPORT_SYMBOL(ath9k_hw_setup_statusring);