Hi Stefan, for me the patch works in mode 11n:
before (OpenBSD 6.7-current (GENERIC.MP) #304: Fri Jun 26 02:08:50 MDT 2020) bandwidth min/avg/max/std-dev = 2.354/12.319/15.391/3.850 Mbps with patch (OpenBSD 6.7-current (GENERIC.MP) #0: Mon Jun 29 09:35:24 GMT 2020) bandwidth min/avg/max/std-dev = 12.174/31.411/57.746/15.154 Mbps iwm0 at pci2 dev 0 function 0 "Intel Dual Band Wireless-AC 8265" rev 0x78, msi iwm0: hw rev 0x230, fw ver 34.0.1, address 60:f6:77:bc:3a:04 (mode 11g: bandwidth min/avg/max/std-dev = 0.620/0.844/1.101/0.153 Mbps) mbk Uwe On 26 Jun 14:45, Stefan Sperling wrote: > This patch adds support for 11n Tx aggregation to iwm(4). > > Please help with testing if you can by running the patch and using wifi > as usual. Nothing should change, except that Tx speed may potentially > improve. If you have time to run before/after performance measurements with > tcpbench or such, that would be nice. But it's not required for testing. > > If Tx aggregation is active then netstat will show a non-zero output block ack > agreement counter: > > $ netstat -W iwm0 | grep 'output block' > 3 new output block ack agreements > 0 output block ack agreements timed out > > It would be great to get at least one test for all the chipsets the driver > supports: 7260, 7265, 3160, 3165, 3168, 8260, 8265, 9260, 9560 > The behaviour of the access point also matters a great deal. It won't > hurt to test the same chipset against several different access points. > > I have tested this version on 8265 only so far. I've run older revisions > of this patch on 7265 so I'm confident that this chip will work, too. > So far, the APs I have tested against are athn(4) in 11a mode and in 11n > mode with the 'nomimo' nwflag, and a Sagemcom 11ac AP. All on 5Ghz channels. > > diff refs/heads/master refs/heads/txagg > blob - 3a75d07a60a7eb4c66540474e47aeffd7a85250a > blob + 853bdd1290ad509f5fce7b5bf20550f458a2b460 > --- sys/dev/pci/if_iwm.c > +++ sys/dev/pci/if_iwm.c > @@ -144,6 +144,8 @@ > #include <net80211/ieee80211_amrr.h> > #include <net80211/ieee80211_mira.h> > #include <net80211/ieee80211_radiotap.h> > +#include <net80211/ieee80211_priv.h> /* for SEQ_LT */ > +#undef DPRINTF /* defined in ieee80211_priv.h */ > > #define DEVNAME(_s) ((_s)->sc_dev.dv_xname) > > @@ -299,7 +301,8 @@ int iwm_nic_rx_mq_init(struct iwm_softc *); > int iwm_nic_tx_init(struct iwm_softc *); > int iwm_nic_init(struct iwm_softc *); > int iwm_enable_ac_txq(struct iwm_softc *, int, int); > -int iwm_enable_txq(struct iwm_softc *, int, int, int); > +int iwm_enable_txq(struct iwm_softc *, int, int, int, int, uint8_t, > + uint16_t); > int iwm_post_alive(struct iwm_softc *); > struct iwm_phy_db_entry *iwm_phy_db_get_section(struct iwm_softc *, uint16_t, > uint16_t); > @@ -334,12 +337,12 @@ void iwm_ampdu_rx_stop(struct ieee80211com *, struct > i > uint8_t); > void iwm_sta_rx_agg(struct iwm_softc *, struct ieee80211_node *, uint8_t, > uint16_t, uint16_t, int); > -#ifdef notyet > +void iwm_sta_tx_agg(struct iwm_softc *, struct ieee80211_node *, uint8_t, > + uint16_t, uint16_t, int); > int iwm_ampdu_tx_start(struct ieee80211com *, struct ieee80211_node *, > uint8_t); > void iwm_ampdu_tx_stop(struct ieee80211com *, struct ieee80211_node *, > uint8_t); > -#endif > void iwm_ba_task(void *); > > int iwm_parse_nvm_data(struct iwm_softc *, const uint16_t *, > @@ -372,14 +375,25 @@ int iwm_rxmq_get_signal_strength(struct iwm_softc > *, s > void iwm_rx_rx_phy_cmd(struct iwm_softc *, struct iwm_rx_packet *, > struct iwm_rx_data *); > int iwm_get_noise(const struct iwm_statistics_rx_non_phy *); > +void iwm_txq_advance(struct iwm_softc *, struct iwm_tx_ring *, int); > +void iwm_ampdu_tx_done(struct iwm_softc *, struct iwm_cmd_header *, > + struct iwm_node *, struct iwm_tx_ring *, uint32_t, uint8_t, > + uint8_t, uint16_t, int, struct iwm_agg_tx_status *); > int iwm_ccmp_decap(struct iwm_softc *, struct mbuf *, > struct ieee80211_node *); > void iwm_rx_frame(struct iwm_softc *, struct mbuf *, int, uint32_t, int, int, > uint32_t, struct ieee80211_rxinfo *, struct mbuf_list *); > -void iwm_rx_tx_cmd_single(struct iwm_softc *, struct iwm_rx_packet *, > - struct iwm_node *, int, int); > +void iwm_rx_tx_cmd_single(struct iwm_softc *, struct iwm_tx_resp *, > + struct iwm_node *, int, int, int); > +void iwm_txd_done(struct iwm_softc *, struct iwm_tx_data *); > void iwm_rx_tx_cmd(struct iwm_softc *, struct iwm_rx_packet *, > struct iwm_rx_data *); > +void iwm_clear_oactive(struct iwm_softc *, struct iwm_tx_ring *); > +void iwm_mira_choose(struct iwm_softc *, struct ieee80211_node *); > +void iwm_ampdu_rate_control(struct iwm_softc *, struct ieee80211_node *, > + struct iwm_tx_ring *, int, uint16_t, uint16_t); > +void iwm_rx_ba(struct iwm_softc *, struct iwm_rx_packet *, > + struct iwm_rx_data *); > void iwm_rx_bmiss(struct iwm_softc *, struct iwm_rx_packet *, > struct iwm_rx_data *); > int iwm_binding_cmd(struct iwm_softc *, struct iwm_node *, uint32_t); > @@ -399,6 +413,7 @@ int iwm_send_cmd_pdu_status(struct iwm_softc *, > uint32 > void iwm_free_resp(struct iwm_softc *, struct iwm_host_cmd *); > void iwm_cmd_done(struct iwm_softc *, int, int, int); > void iwm_update_sched(struct iwm_softc *, int, int, uint8_t, uint16_t); > +void iwm_reset_sched(struct iwm_softc *, int, int, uint8_t); > const struct iwm_rate *iwm_tx_fill_cmd(struct iwm_softc *, struct iwm_node *, > struct ieee80211_frame *, struct iwm_tx_cmd *); > int iwm_tx(struct iwm_softc *, struct mbuf *, struct ieee80211_node *, int); > @@ -1306,17 +1321,17 @@ iwm_alloc_tx_ring(struct iwm_softc *sc, struct iwm_tx_ > * The command is queue 0 (sc->txq[0]), and 4 mgmt/data frame queues > * are sc->tqx[IWM_DQA_MIN_MGMT_QUEUE + ac], i.e. sc->txq[5:8], > * in order to provide one queue per EDCA category. > + * Tx aggregation requires additional queues, one queue per TID for > + * which aggregation is enabled. We map TID 0-7 to sc->txq[10:17]. > * > - * In non-DQA mode, we use rings 0 through 9 (0-3 are EDCA, 9 is cmd). > + * In non-DQA mode, we use rings 0 through 9 (0-3 are EDCA, 9 is cmd), > + * and Tx aggregation is not supported. > * > - * Tx aggregation will require additional queues (one queue per TID > - * for which aggregation is enabled) but we do not implement this yet. > - * > * Unfortunately, we cannot tell if DQA will be used until the > * firmware gets loaded later, so just allocate sufficient rings > * in order to satisfy both cases. > */ > - if (qid > IWM_CMD_QUEUE) > + if (qid > IWM_LAST_AGG_TX_QUEUE) > return 0; > > size = IWM_TX_RING_COUNT * sizeof(struct iwm_device_cmd); > @@ -1380,6 +1395,7 @@ iwm_reset_tx_ring(struct iwm_softc *sc, struct iwm_tx_ > bus_dmamap_sync(sc->sc_dmat, ring->desc_dma.map, 0, > ring->desc_dma.size, BUS_DMASYNC_PREWRITE); > sc->qfullmsk &= ~(1 << ring->qid); > + sc->qenablemsk &= ~(1 << ring->qid); > /* 7000 family NICs are locked while commands are in progress. */ > if (ring->qid == sc->cmdqid && ring->queued > 0) { > if (sc->sc_device_family == IWM_DEVICE_FAMILY_7000) > @@ -2208,6 +2224,18 @@ const uint8_t iwm_ac_to_tx_fifo[] = { > IWM_TX_FIFO_VO, > }; > > +/* Map a TID to an ieee80211_edca_ac category. */ > +const uint8_t iwm_tid_to_ac[IWM_MAX_TID_COUNT] = { > + EDCA_AC_BE, > + EDCA_AC_BK, > + EDCA_AC_BK, > + EDCA_AC_BE, > + EDCA_AC_VI, > + EDCA_AC_VI, > + EDCA_AC_VO, > + EDCA_AC_VO, > +}; > + > int > iwm_enable_ac_txq(struct iwm_softc *sc, int qid, int fifo) > { > @@ -2250,28 +2278,48 @@ iwm_enable_ac_txq(struct iwm_softc *sc, int qid, int f > } > > int > -iwm_enable_txq(struct iwm_softc *sc, int sta_id, int qid, int fifo) > +iwm_enable_txq(struct iwm_softc *sc, int sta_id, int qid, int fifo, > + int aggregate, uint8_t tid, uint16_t ssn) > { > + struct iwm_tx_ring *ring = &sc->txq[qid]; > struct iwm_scd_txq_cfg_cmd cmd; > - int err; > + int err, idx, scd_bug; > > iwm_nic_assert_locked(sc); > > - IWM_WRITE(sc, IWM_HBUS_TARG_WRPTR, qid << 8 | 0); > + /* > + * If we need to move the SCD write pointer by steps of > + * 0x40, 0x80 or 0xc0, it gets stuck. > + * This is really ugly, but this is the easiest way out for > + * this sad hardware issue. > + * This bug has been fixed on devices 9000 and up. > + */ > + scd_bug = !sc->sc_mqrx_supported && > + !((ssn - ring->cur) & 0x3f) && > + (ssn != ring->cur); > + if (scd_bug) > + ssn = (ssn + 1) & 0xfff; > > + idx = IWM_AGG_SSN_TO_TXQ_IDX(ssn); > + IWM_WRITE(sc, IWM_HBUS_TARG_WRPTR, qid << 8 | idx); > + ring->cur = idx; > + ring->tail = idx; > + > memset(&cmd, 0, sizeof(cmd)); > + cmd.tid = tid; > cmd.scd_queue = qid; > cmd.enable = 1; > cmd.sta_id = sta_id; > cmd.tx_fifo = fifo; > - cmd.aggregate = 0; > + cmd.aggregate = aggregate; > + cmd.ssn = htole16(ssn); > cmd.window = IWM_FRAME_LIMIT; > > - err = iwm_send_cmd_pdu(sc, IWM_SCD_QUEUE_CFG, 0, > - sizeof(cmd), &cmd); > + err = iwm_send_cmd_pdu(sc, IWM_SCD_QUEUE_CFG, 0, sizeof(cmd), &cmd); > if (err) > return err; > > + sc->qenablemsk |= (1 << qid); > return 0; > } > > @@ -2950,8 +2998,12 @@ iwm_sta_rx_agg(struct iwm_softc *sc, struct ieee80211_ > if (start) { > sc->sc_rx_ba_sessions++; > ieee80211_addba_req_accept(ic, ni, tid); > - } else if (sc->sc_rx_ba_sessions > 0) > - sc->sc_rx_ba_sessions--; > + in->ampdu_rx_tid_mask |= (1 << tid); > + } else { > + in->ampdu_rx_tid_mask &= ~(1 << tid); > + if (sc->sc_rx_ba_sessions > 0) > + sc->sc_rx_ba_sessions--; > + } > } else if (start) > ieee80211_addba_req_refuse(ic, ni, tid); > > @@ -2959,6 +3011,75 @@ iwm_sta_rx_agg(struct iwm_softc *sc, struct ieee80211_ > } > > void > +iwm_sta_tx_agg(struct iwm_softc *sc, struct ieee80211_node *ni, uint8_t tid, > + uint16_t ssn, uint16_t winsize, int start) > +{ > + struct ieee80211com *ic = &sc->sc_ic; > + struct iwm_node *in = (void *)ni; > + int qid = IWM_FIRST_AGG_TX_QUEUE + tid; > + enum ieee80211_edca_ac ac = iwm_tid_to_ac[tid]; > + int fifo = iwm_ac_to_tx_fifo[ac]; > + int err; > + > + if (qid > IWM_LAST_AGG_TX_QUEUE || !iwm_nic_lock(sc)) { > + ieee80211_addba_resp_refuse(ic, ni, tid, > + IEEE80211_STATUS_UNSPECIFIED); > + return; > + } > + > + if (start) { > + if ((sc->qenablemsk & (1 << qid)) == 0) { > + struct iwm_tx_ring *ring = &sc->txq[qid]; > + > + err = iwm_enable_txq(sc, IWM_STATION_ID, qid, fifo, > + 1, tid, ssn); > + if (err) > + goto done; > + /* > + * If iwm_enable_txq() employed the SCD hardware bug > + * workaround we must skip the frame with seqnum SSN. > + */ > + if (IWM_AGG_SSN_TO_TXQ_IDX(ring->cur) != > + IWM_AGG_SSN_TO_TXQ_IDX(ssn)) { > + ssn = (ssn + 1) & 0xfff; > + KASSERT(IWM_AGG_SSN_TO_TXQ_IDX(ring->cur) == > + IWM_AGG_SSN_TO_TXQ_IDX(ssn)); > + ieee80211_output_ba_move_window(ic, ni, > + tid, ssn); > + ni->ni_qos_txseqs[tid] = ssn; > + } > + } > + in->ampdu_tx_tid_mask |= (1 << tid); > + err = iwm_add_sta_cmd(sc, in, 1); > + if (err) { > + printf("%s: could not update sta (error %d)\n", > + DEVNAME(sc), err); > + in->ampdu_tx_tid_mask &= ~(1 << tid); > + goto done; > + } > + } else { > + in->ampdu_tx_tid_mask &= ~(1 << tid); > + err = iwm_add_sta_cmd(sc, in, 1); > + if (err) { > + printf("%s: could not update sta (error %d)\n", > + DEVNAME(sc), err); > + in->ampdu_tx_tid_mask |= (1 << tid); > + goto done; > + } > + } > + > +done: > + iwm_nic_unlock(sc); > + if (start) { > + if (err) > + ieee80211_addba_resp_refuse(ic, ni, tid, > + IEEE80211_STATUS_UNSPECIFIED); > + else > + ieee80211_addba_resp_accept(ic, ni, tid); > + } > +} > + > +void > iwm_htprot_task(void *arg) > { > struct iwm_softc *sc = arg; > @@ -3002,19 +3123,53 @@ iwm_ba_task(void *arg) > struct ieee80211com *ic = &sc->sc_ic; > struct ieee80211_node *ni = ic->ic_bss; > int s = splnet(); > + int tid; > > - if (sc->sc_flags & IWM_FLAG_SHUTDOWN) { > + if ((sc->sc_flags & IWM_FLAG_SHUTDOWN) || > + ic->ic_state != IEEE80211_S_RUN) { > refcnt_rele_wake(&sc->task_refs); > splx(s); > return; > } > > - if (sc->ba_start) > - iwm_sta_rx_agg(sc, ni, sc->ba_tid, sc->ba_ssn, > - sc->ba_winsize, 1); > - else > - iwm_sta_rx_agg(sc, ni, sc->ba_tid, 0, 0, 0); > + if (sc->ba_flags & IWM_RX_BA_START) { > + for (tid = 0; tid < IWM_MAX_TID_COUNT; tid++) { > + if ((sc->rx_ba_start.tid_mask & (1 << tid)) == 0) > + continue; > + iwm_sta_rx_agg(sc, ni, tid, sc->rx_ba_start.ssn[tid], > + sc->rx_ba_start.winsize[tid], 1); > + sc->rx_ba_start.tid_mask &= ~(1 << tid); > + } > + } > + > + if (sc->ba_flags & IWM_RX_BA_STOP) { > + for (tid = 0; tid < IWM_MAX_TID_COUNT; tid++) { > + if ((sc->rx_ba_stop.tid_mask & (1 << tid)) == 0) > + continue; > + iwm_sta_rx_agg(sc, ni, tid, 0, 0, 0); > + sc->rx_ba_stop.tid_mask &= ~(1 << tid); > + } > + } > > + if (sc->ba_flags & IWM_TX_BA_START) { > + for (tid = 0; tid < IWM_MAX_TID_COUNT; tid++) { > + if ((sc->tx_ba_start.tid_mask & (1 << tid)) == 0) > + continue; > + iwm_sta_tx_agg(sc, ni, tid, sc->tx_ba_start.ssn[tid], > + sc->tx_ba_start.winsize[tid], 1); > + sc->tx_ba_start.tid_mask &= ~(1 << tid); > + } > + } > + > + if (sc->ba_flags & IWM_TX_BA_STOP) { > + for (tid = 0; tid < IWM_MAX_TID_COUNT; tid++) { > + if ((sc->tx_ba_stop.tid_mask & (1 << tid)) == 0) > + continue; > + iwm_sta_tx_agg(sc, ni, tid, 0, 0, 0); > + sc->tx_ba_stop.tid_mask &= ~(1 << tid); > + } > + } > + > refcnt_rele_wake(&sc->task_refs); > splx(s); > } > @@ -3029,14 +3184,23 @@ iwm_ampdu_rx_start(struct ieee80211com *ic, struct iee > { > struct ieee80211_rx_ba *ba = &ni->ni_rx_ba[tid]; > struct iwm_softc *sc = IC2IFP(ic)->if_softc; > + struct iwm_node *in = (void *)ni; > > - if (sc->sc_rx_ba_sessions >= IWM_MAX_RX_BA_SESSIONS) > + if (tid >= IWM_MAX_TID_COUNT) > + return EINVAL; > + > + if (sc->sc_rx_ba_sessions >= IWM_MAX_RX_BA_SESSIONS || > + (in->ampdu_rx_tid_mask & (1 << tid)) != 0) > return ENOSPC; > > - sc->ba_start = 1; > - sc->ba_tid = tid; > - sc->ba_ssn = htole16(ba->ba_winstart); > - sc->ba_winsize = htole16(ba->ba_winsize); > + if (sc->rx_ba_start.tid_mask & (1 << tid) || > + sc->rx_ba_stop.tid_mask & (1 << tid)) > + return EAGAIN; > + > + sc->ba_flags |= IWM_RX_BA_START; > + sc->rx_ba_start.tid_mask |= (1 << tid); > + sc->rx_ba_start.ssn[tid] = htole16(ba->ba_winstart); > + sc->rx_ba_start.winsize[tid] = htole16(ba->ba_winsize); > iwm_add_task(sc, systq, &sc->ba_task); > > return EBUSY; > @@ -3051,13 +3215,69 @@ iwm_ampdu_rx_stop(struct ieee80211com *ic, struct ieee > uint8_t tid) > { > struct iwm_softc *sc = IC2IFP(ic)->if_softc; > + struct iwm_node *in = (void *)ni; > > - sc->ba_start = 0; > - sc->ba_tid = tid; > + if (in->ampdu_rx_tid_mask & (1 << tid)) { > + sc->ba_flags |= IWM_RX_BA_STOP; > + sc->rx_ba_stop.tid_mask |= (1 << tid); > + iwm_add_task(sc, systq, &sc->ba_task); > + } > +} > + > +int > +iwm_ampdu_tx_start(struct ieee80211com *ic, struct ieee80211_node *ni, > + uint8_t tid) > +{ > + struct iwm_softc *sc = IC2IFP(ic)->if_softc; > + struct iwm_node *in = (void *)ni; > + struct ieee80211_tx_ba *ba = &ni->ni_tx_ba[tid]; > + > + /* We only implement Tx aggregation with DQA-capable firmware. */ > + if (!isset(sc->sc_enabled_capa, IWM_UCODE_TLV_CAPA_DQA_SUPPORT)) > + return ENOTSUP; > + > + /* Ensure we can map this TID to an aggregation queue. */ > + if (tid >= IWM_MAX_TID_COUNT) > + return EINVAL; > + > + /* We only support a fixed Tx aggregation window size, for now. */ > + if (ba->ba_winsize != IWM_FRAME_LIMIT) > + return ENOTSUP; > + > + if ((in->ampdu_tx_tid_mask & (1 << tid)) != 0) > + return ENOSPC; > + > + if (sc->tx_ba_start.tid_mask & (1 << tid) || > + sc->tx_ba_stop.tid_mask & (1 << tid)) > + return EAGAIN; > + > + sc->ba_flags |= IWM_TX_BA_START; > + sc->tx_ba_start.tid_mask |= (1 << tid); > + sc->tx_ba_start.ssn[tid] = htole16(ba->ba_winstart); > + sc->tx_ba_start.winsize[tid] = htole16(ba->ba_winsize); > iwm_add_task(sc, systq, &sc->ba_task); > + > + return EBUSY; > } > > void > +iwm_ampdu_tx_stop(struct ieee80211com *ic, struct ieee80211_node *ni, > + uint8_t tid) > +{ > + struct iwm_softc *sc = IC2IFP(ic)->if_softc; > + struct iwm_node *in = (void *)ni; > + > + if (tid >= IWM_MAX_TID_COUNT) > + return; > + > + if (in->ampdu_tx_tid_mask & (1 << tid)) { > + sc->ba_flags |= IWM_TX_BA_STOP; > + sc->tx_ba_stop.tid_mask |= (1 << tid); > + iwm_add_task(sc, systq, &sc->ba_task); > + } > +} > + > +void > iwm_set_hw_address_8000(struct iwm_softc *sc, struct iwm_nvm_data *data, > const uint16_t *mac_override, const uint16_t *nvm_hw) > { > @@ -4238,13 +4458,178 @@ iwm_rx_mpdu_mq(struct iwm_softc *sc, struct mbuf *m, > v > } > > void > -iwm_rx_tx_cmd_single(struct iwm_softc *sc, struct iwm_rx_packet *pkt, > - struct iwm_node *in, int txmcs, int txrate) > +iwm_txq_advance(struct iwm_softc *sc, struct iwm_tx_ring *ring, int idx) > { > + struct iwm_tx_data *txd; > + > + while (ring->tail != idx) { > + txd = &ring->data[ring->tail]; > + if (txd->m != NULL) { > + if (ring->qid < IWM_FIRST_AGG_TX_QUEUE) > + DPRINTF(("%s: missed Tx completion: tail=%d " > + "idx=%d\n", __func__, ring->tail, idx)); > + iwm_reset_sched(sc, ring->qid, ring->tail, > IWM_STATION_ID); > + iwm_txd_done(sc, txd); > + ring->queued--; > + } > + ring->tail = (ring->tail + 1) % IWM_TX_RING_COUNT; > + } > +} > + > +void > +iwm_ampdu_tx_done(struct iwm_softc *sc, struct iwm_cmd_header *cmd_hdr, > + struct iwm_node *in, struct iwm_tx_ring *txq, uint32_t initial_rate, > + uint8_t nframes, uint8_t failure_frame, uint16_t ssn, int status, > + struct iwm_agg_tx_status *agg_status) > +{ > struct ieee80211com *ic = &sc->sc_ic; > + int tid = cmd_hdr->qid - IWM_FIRST_AGG_TX_QUEUE; > + struct iwm_tx_data *txdata = &txq->data[cmd_hdr->idx]; > struct ieee80211_node *ni = &in->in_ni; > + int txfail = (status != IWM_TX_STATUS_SUCCESS && > + status != IWM_TX_STATUS_DIRECT_DONE); > + struct ieee80211_tx_ba *ba; > + > + sc->sc_tx_timer = 0; > + > + if (ic->ic_state != IEEE80211_S_RUN) > + return; > + > + if (nframes > 1) { > + int ampdu_id, have_ampdu_id = 0, ampdu_size = 0; > + int i; > + > + /* Compute the size of this A-MPDU. */ > + for (i = 0; i < nframes; i++) { > + uint8_t qid = agg_status[i].qid; > + uint8_t idx = agg_status[i].idx; > + > + if (qid != cmd_hdr->qid) > + continue; > + > + txdata = &txq->data[idx]; > + if (txdata->m == NULL) > + continue; > + > + ampdu_size += txdata->totlen + IEEE80211_CRC_LEN; > + } > + > + /* > + * For each subframe collect Tx status, retries, and Tx rate. > + * (The Tx rate is the same for all subframes in this batch.) > + */ > + for (i = 0; i < nframes; i++) { > + uint8_t qid = agg_status[i].qid; > + uint8_t idx = agg_status[i].idx; > + uint16_t txstatus = (le16toh(agg_status[i].status) & > + IWM_AGG_TX_STATE_STATUS_MSK); > + uint16_t trycnt = (le16toh(agg_status[i].status) & > + IWM_AGG_TX_STATE_TRY_CNT_MSK) >> > + IWM_AGG_TX_STATE_TRY_CNT_POS; > + > + if (qid != cmd_hdr->qid) > + continue; > + > + txdata = &txq->data[idx]; > + if (txdata->m == NULL) > + continue; > + > + if (initial_rate & IWM_RATE_MCS_HT_MSK) > + txdata->ampdu_txmcs = (initial_rate & > + (IWM_RATE_HT_MCS_RATE_CODE_MSK | > + IWM_RATE_HT_MCS_NSS_MSK)); > + if (txstatus != IWM_AGG_TX_STATE_TRANSMITTED) > + txdata->txfail++; > + if (trycnt > 1) > + txdata->retries++; > + > + /* > + * Assign a common ID to all subframes of this A-MPDU. > + * This ID will be used during Tx rate control to > + * infer the ACK status of individual subframes. > + */ > + if (!have_ampdu_id) { > + ampdu_id = txdata->in->next_ampdu_id++; > + have_ampdu_id = 1; > + } > + txdata->ampdu_id = ampdu_id; > + > + /* > + * We will also need to know the total number of > + * subframes and the size of this A-MPDU. We store > + * this redundantly on each subframe because firmware > + * only reports acknowledged subframes via compressed > + * block-ack notification. This way we will know what > + * the total number of subframes and size were even if > + * just one of these subframes gets acknowledged. > + */ > + txdata->ampdu_nframes = nframes; > + txdata->ampdu_size = ampdu_size; > + } > + return; > + } > + > + if (ni == NULL) > + return; > + > + ba = &ni->ni_tx_ba[tid]; > + if (ba->ba_state != IEEE80211_BA_AGREED) > + return; > + > + /* This is a final single-frame Tx attempt. */ > + DPRINTFN(3, ("%s: final tx status=0x%x qid=%d queued=%d idx=%d ssn=%u " > + "bitmap=0x%llx\n", __func__, status, desc->qid, txq->queued, > + desc->idx, ssn, ba->ba_bitmap)); > + > + /* > + * Skip rate control if our Tx rate is fixed. > + * Don't report frames to MiRA which were sent at a different > + * Tx rate than ni->ni_txmcs. > + */ > + if (ic->ic_fixed_mcs == -1 && txdata->txmcs == ni->ni_txmcs) { > + in->in_mn.frames++; > + in->in_mn.agglen = 1; > + in->in_mn.ampdu_size = txdata->totlen + IEEE80211_CRC_LEN; > + if (failure_frame > 0) > + in->in_mn.retries++; > + if (txfail) > + in->in_mn.txfail++; > + iwm_mira_choose(sc, ni); > + } > + > + if (txfail) > + ieee80211_tx_compressed_bar(ic, ni, tid, ssn); > + else if (!SEQ_LT(ssn, ba->ba_winstart)) { > + /* > + * Move window forward if SSN lies beyond end of window, > + * otherwise we can't record the ACK for this frame. > + * Non-acked frames which left holes in the bitmap near > + * the beginning of the window must be discarded. > + */ > + uint16_t s = ssn; > + while (SEQ_LT(ba->ba_winend, s)) { > + ieee80211_output_ba_move_window(ic, ni, tid, s); > + iwm_txq_advance(sc, txq, IWM_AGG_SSN_TO_TXQ_IDX(s)); > + s = (s + 1) % 0xfff; > + } > + /* SSN should now be within window; set corresponding bit. */ > + ieee80211_output_ba_record_ack(ic, ni, tid, ssn); > + } > + > + /* Move window forward up to the first hole in the bitmap. */ > + ieee80211_output_ba_move_window_to_first_unacked(ic, ni, tid, ssn); > + iwm_txq_advance(sc, txq, IWM_AGG_SSN_TO_TXQ_IDX(ba->ba_winstart)); > + > + iwm_clear_oactive(sc, txq); > +} > + > +void > +iwm_rx_tx_cmd_single(struct iwm_softc *sc, struct iwm_tx_resp *tx_resp, > + struct iwm_node *in, int txmcs, int txrate, int qid) > +{ > + struct ieee80211com *ic = &sc->sc_ic; > + struct ieee80211_node *ni = &in->in_ni; > struct ifnet *ifp = IC2IFP(ic); > - struct iwm_tx_resp *tx_resp = (void *)pkt->data; > int status = le16toh(tx_resp->status.status) & IWM_TX_STATUS_MSK; > int txfail; > > @@ -4277,22 +4662,8 @@ iwm_rx_tx_cmd_single(struct iwm_softc *sc, struct iwm_ > in->in_mn.retries += tx_resp->failure_frame; > if (txfail) > in->in_mn.txfail += tx_resp->frame_count; > - if (ic->ic_state == IEEE80211_S_RUN) { > - int best_mcs; > - > - ieee80211_mira_choose(&in->in_mn, ic, &in->in_ni); > - /* > - * If MiRA has chosen a new TX rate we must update > - * the firwmare's LQ rate table from process context. > - * ni_txmcs may change again before the task runs so > - * cache the chosen rate in the iwm_node structure. > - */ > - best_mcs = ieee80211_mira_get_best_mcs(&in->in_mn); > - if (best_mcs != in->chosen_txmcs) { > - in->chosen_txmcs = best_mcs; > - iwm_setrates(in, 1); > - } > - } > + if (ic->ic_state == IEEE80211_S_RUN) > + iwm_mira_choose(sc, ni); > } > > if (txfail) > @@ -4313,49 +4684,91 @@ iwm_txd_done(struct iwm_softc *sc, struct iwm_tx_data > KASSERT(txd->in); > ieee80211_release_node(ic, &txd->in->in_ni); > txd->in = NULL; > + > + txd->retries = 0; > + txd->txfail = 0; > + txd->txmcs = 0; > + txd->ampdu_txmcs = 0; > + txd->txrate = 0; > } > > void > iwm_rx_tx_cmd(struct iwm_softc *sc, struct iwm_rx_packet *pkt, > struct iwm_rx_data *data) > { > - struct ieee80211com *ic = &sc->sc_ic; > - struct ifnet *ifp = IC2IFP(ic); > struct iwm_cmd_header *cmd_hdr = &pkt->hdr; > int idx = cmd_hdr->idx; > int qid = cmd_hdr->qid; > struct iwm_tx_ring *ring = &sc->txq[qid]; > struct iwm_tx_data *txd; > + struct iwm_tx_resp *tx_resp = (void *)pkt->data; > + uint32_t ssn; > + uint32_t len = iwm_rx_packet_len(pkt); > > bus_dmamap_sync(sc->sc_dmat, data->map, 0, IWM_RBUF_SIZE, > BUS_DMASYNC_POSTREAD); > > sc->sc_tx_timer = 0; > > + /* Sanity checks. */ > + if (sizeof(*tx_resp) > len) > + return; > + if (qid < IWM_FIRST_AGG_TX_QUEUE && tx_resp->frame_count > 1) > + return; > + if (qid >= IWM_FIRST_AGG_TX_QUEUE && sizeof(*tx_resp) + sizeof(ssn) + > + tx_resp->frame_count * sizeof(tx_resp->status) > len) > + return; > + > + /* > + * In the multi-frame case the firmware has just transmitted a bunch > + * of frames in an A-MPDU. The final Tx status of those frames won't > + * be known until the peer ACKs subframes with a block ack or firmware > + * gives up on a particular subframe. > + * Subframes for which the firmware never sees an ACK will be retried > + * and will eventually arrive here as a single-frame Tx failure. > + * So there is nothing to do, for now. > + */ > + if (tx_resp->frame_count != 1) > + return; > + > txd = &ring->data[idx]; > if (txd->m == NULL) > return; > > - iwm_rx_tx_cmd_single(sc, pkt, txd->in, txd->txmcs, txd->txrate); > - iwm_txd_done(sc, txd); > + if (qid >= IWM_FIRST_AGG_TX_QUEUE) { > + int status; > > - /* > - * XXX Sometimes we miss Tx completion interrupts. > - * We cannot check Tx success/failure for affected frames; just free > - * the associated mbuf and release the associated node reference. > - */ > - while (ring->tail != idx) { > - txd = &ring->data[ring->tail]; > - if (txd->m != NULL) { > - DPRINTF(("%s: missed Tx completion: tail=%d idx=%d\n", > - __func__, ring->tail, idx)); > - iwm_txd_done(sc, txd); > - ring->queued--; > - } > - ring->tail = (ring->tail + 1) % IWM_TX_RING_COUNT; > + memcpy(&ssn, &tx_resp->status + tx_resp->frame_count, > sizeof(ssn)); > + ssn = le32toh(ssn) & 0xfff; > + status = le16toh(tx_resp->status.status) & IWM_TX_STATUS_MSK; > + iwm_ampdu_tx_done(sc, cmd_hdr, txd->in, ring, > + le32toh(tx_resp->initial_rate), tx_resp->frame_count, > + tx_resp->failure_frame, ssn, status, &tx_resp->status); > + } else { > + iwm_rx_tx_cmd_single(sc, tx_resp, txd->in, txd->txmcs, > + txd->txrate, qid); > + iwm_reset_sched(sc, qid, idx, IWM_STATION_ID); > + iwm_txd_done(sc, txd); > + ring->queued--; > + > + /* > + * XXX Sometimes we miss Tx completion interrupts. > + * We cannot check Tx success/failure for affected frames; > + * just free the associated mbuf and release the associated > + * node reference. > + */ > + iwm_txq_advance(sc, ring, idx); > + iwm_clear_oactive(sc, ring); > } > +} > > - if (--ring->queued < IWM_TX_RING_LOMARK) { > +void > +iwm_clear_oactive(struct iwm_softc *sc, struct iwm_tx_ring *ring) > +{ > + struct ieee80211com *ic = &sc->sc_ic; > + struct ifnet *ifp = IC2IFP(ic); > + > + if (ring->queued < IWM_TX_RING_LOMARK) { > sc->qfullmsk &= ~(1 << ring->qid); > if (sc->qfullmsk == 0 && ifq_is_oactive(&ifp->if_snd)) { > ifq_clr_oactive(&ifp->if_snd); > @@ -4370,6 +4783,183 @@ iwm_rx_tx_cmd(struct iwm_softc *sc, struct iwm_rx_pack > } > > void > +iwm_mira_choose(struct iwm_softc *sc, struct ieee80211_node *ni) > +{ > + struct ieee80211com *ic = &sc->sc_ic; > + struct iwm_node *in = (void *)ni; > + int best_mcs = ieee80211_mira_get_best_mcs(&in->in_mn); > + > + ieee80211_mira_choose(&in->in_mn, ic, ni); > + > + /* > + * Update firmware's LQ retry table if MiRA has chosen a new MCS. > + * > + * We only need to do this if the best MCS has changed because > + * we ask firmware to use a fixed MCS while MiRA is probing a > + * candidate MCS. > + * While not probing we ask firmware to retry at lower rates in case > + * Tx at the newly chosen best MCS ends up failing, and then report > + * any resulting Tx retries to MiRA in order to trigger probing. > + */ > + if (best_mcs != ieee80211_mira_get_best_mcs(&in->in_mn)) { > + in->chosen_txmcs = best_mcs; > + iwm_setrates(in, 1); > + } > +} > + > +void > +iwm_ampdu_rate_control(struct iwm_softc *sc, struct ieee80211_node *ni, > + struct iwm_tx_ring *txq, int tid, uint16_t seq, uint16_t ssn) > +{ > + struct ieee80211com *ic = &sc->sc_ic; > + struct iwm_node *in = (void *)ni; > + struct ieee80211_tx_ba *ba = &ni->ni_tx_ba[tid]; > + int min_ampdu_id, max_ampdu_id, id; > + int idx, end_idx; > + > + /* Determine the min/max IDs we assigned to AMPDUs in this range. */ > + idx = IWM_AGG_SSN_TO_TXQ_IDX(seq); > + end_idx = IWM_AGG_SSN_TO_TXQ_IDX(ssn); > + min_ampdu_id = txq->data[idx].ampdu_id; > + max_ampdu_id = min_ampdu_id; > + while (idx != end_idx) { > + struct iwm_tx_data *txdata = &txq->data[idx]; > + > + if (txdata->m != NULL) { > + if (min_ampdu_id > txdata->ampdu_id) > + min_ampdu_id = txdata->ampdu_id; > + if (max_ampdu_id < txdata->ampdu_id) > + max_ampdu_id = txdata->ampdu_id; > + } > + > + idx = (idx + 1) % IWM_TX_RING_COUNT; > + } > + > + /* > + * Update Tx rate statistics for A-MPDUs before firmware's BA window. > + */ > + for (id = min_ampdu_id; id <= max_ampdu_id; id++) { > + int have_ack = 0, bit = 0; > + idx = IWM_AGG_SSN_TO_TXQ_IDX(seq); > + end_idx = IWM_AGG_SSN_TO_TXQ_IDX(ssn); > + in->in_mn.agglen = 0; > + in->in_mn.ampdu_size = 0; > + while (idx != end_idx) { > + struct iwm_tx_data *txdata = &txq->data[idx]; > + uint16_t s = (seq + bit) & 0xfff; > + /* > + * We can assume that this subframe has been ACKed > + * because ACK failures come as single frames and > + * before failing an A-MPDU subframe the firmware > + * sends it as a single frame at least once. > + * > + * However, when this A-MPDU was transmitted we > + * learned how many subframes it contained. > + * So if firmware isn't reporting all subframes now > + * we can deduce an ACK failure for missing frames. > + */ > + if (txdata->m != NULL && txdata->ampdu_id == id && > + txdata->ampdu_txmcs == ni->ni_txmcs && > + txdata->ampdu_nframes > 0 && > + (SEQ_LT(ba->ba_winend, s) || > + (ba->ba_bitmap & (1 << bit)) == 0)) { > + have_ack++; > + in->in_mn.frames = txdata->ampdu_nframes; > + in->in_mn.agglen = txdata->ampdu_nframes; > + in->in_mn.ampdu_size = txdata->ampdu_size; > + if (txdata->retries > 1) > + in->in_mn.retries++; > + if (!SEQ_LT(ba->ba_winend, s)) > + ieee80211_output_ba_record_ack(ic, ni, > + tid, s); > + } > + > + idx = (idx + 1) % IWM_TX_RING_COUNT; > + bit++; > + } > + > + if (have_ack > 0) { > + in->in_mn.txfail = in->in_mn.frames - have_ack; > + iwm_mira_choose(sc, ni); > + } > + } > +} > + > +void > +iwm_rx_ba(struct iwm_softc *sc, struct iwm_rx_packet *pkt, > + struct iwm_rx_data *data) > +{ > + struct iwm_ba_notif *ban = (void *)pkt->data; > + struct ieee80211com *ic = &sc->sc_ic; > + struct ieee80211_node *ni; > + struct ieee80211_tx_ba *ba; > + struct iwm_node *in; > + struct iwm_tx_ring *ring; > + uint16_t seq, ssn, idx; > + int qid; > + > + if (ic->ic_state != IEEE80211_S_RUN) > + return; > + > + if (iwm_rx_packet_payload_len(pkt) < sizeof(*ban)) > + return; > + > + if (ban->sta_id != IWM_STATION_ID || > + !IEEE80211_ADDR_EQ(ic->ic_bss->ni_macaddr, ban->sta_addr)) > + return; > + > + ni = ic->ic_bss; > + in = (void *)ni; > + > + qid = le16toh(ban->scd_flow); > + if (qid < IWM_FIRST_AGG_TX_QUEUE || qid > IWM_LAST_AGG_TX_QUEUE) > + return; > + > + /* Protect against a firmware bug where the queue/TID are off. */ > + if (qid != IWM_FIRST_AGG_TX_QUEUE + ban->tid) > + return; > + > + ba = &ni->ni_tx_ba[ban->tid]; > + if (ba->ba_state != IEEE80211_BA_AGREED) > + return; > + > + ring = &sc->txq[qid]; > + ssn = le16toh(ban->scd_ssn); /* BA window starting sequence number */ > + idx = IWM_AGG_SSN_TO_TXQ_IDX(ssn); > + > + /* > + * The first bit in ban->bitmap corresponds to the sequence number > + * stored in the sequence control field ban->seq_ctl. > + * Multiple BA notifications in a row may be using this number, with > + * additional bits being set in cba->bitmap. It is unclear how the > + * firmware decides to shift this window forward. > + */ > + seq = le16toh(ban->seq_ctl) >> IEEE80211_SEQ_SEQ_SHIFT; > + > + /* > + * The firmware's new BA window starting sequence number > + * corresponds to the first hole in ban->scd_ssn, implying > + * that all frames between 'seq' and 'ssn' have been acked. > + */ > + ssn = le16toh(ban->scd_ssn); > + > + /* Skip rate control if our Tx rate is fixed. */ > + if (ic->ic_fixed_mcs != -1) > + iwm_ampdu_rate_control(sc, ni, ring, ban->tid, seq, ssn); > + > + /* > + * SSN corresponds to the first (perhaps not yet transmitted) frame > + * in firmware's BA window. Firmware is not going to retransmit any > + * frames before its BA window so mark them all as done. > + */ > + if (SEQ_LT(ba->ba_winstart, ssn)) { > + ieee80211_output_ba_move_window(ic, ni, ban->tid, ssn); > + iwm_txq_advance(sc, ring, IWM_AGG_SSN_TO_TXQ_IDX(ssn)); > + iwm_clear_oactive(sc, ring); > + } > +} > + > +void > iwm_rx_bmiss(struct iwm_softc *sc, struct iwm_rx_packet *pkt, > struct iwm_rx_data *data) > { > @@ -4638,9 +5228,6 @@ iwm_send_cmd(struct iwm_softc *sc, struct iwm_host_cmd > } > } > > -#if 0 > - iwm_update_sched(sc, ring->qid, ring->cur, 0, 0); > -#endif > /* Kick command ring. */ > ring->queued++; > ring->cur = (ring->cur + 1) % IWM_TX_RING_COUNT; > @@ -4771,7 +5358,6 @@ iwm_cmd_done(struct iwm_softc *sc, int qid, int idx, i > } > } > > -#if 0 > /* > * necessary only for block ack mode > */ > @@ -4780,32 +5366,49 @@ iwm_update_sched(struct iwm_softc *sc, int qid, int id > uint16_t len) > { > struct iwm_agn_scd_bc_tbl *scd_bc_tbl; > - uint16_t w_val; > + uint16_t val; > > scd_bc_tbl = sc->sched_dma.vaddr; > > - len += 8; /* magic numbers came naturally from paris */ > + len += IWM_TX_CRC_SIZE + IWM_TX_DELIMITER_SIZE; > if (sc->sc_capaflags & IWM_UCODE_TLV_FLAGS_DW_BC_TABLE) > len = roundup(len, 4) / 4; > > - w_val = htole16(sta_id << 12 | len); > + val = htole16(sta_id << 12 | len); > > + bus_dmamap_sync(sc->sc_dmat, sc->sched_dma.map, > + 0, sc->sched_dma.size, BUS_DMASYNC_PREWRITE); > + > /* Update TX scheduler. */ > - scd_bc_tbl[qid].tfd_offset[idx] = w_val; > + scd_bc_tbl[qid].tfd_offset[idx] = val; > + if (idx < IWM_TFD_QUEUE_SIZE_BC_DUP) > + scd_bc_tbl[qid].tfd_offset[IWM_TFD_QUEUE_SIZE_MAX + idx] = val; > + > bus_dmamap_sync(sc->sc_dmat, sc->sched_dma.map, > - (char *)(void *)w - (char *)(void *)sc->sched_dma.vaddr, > - sizeof(uint16_t), BUS_DMASYNC_PREWRITE); > + 0, sc->sched_dma.size, BUS_DMASYNC_POSTWRITE); > +} > > - /* I really wonder what this is ?!? */ > - if (idx < IWM_TFD_QUEUE_SIZE_BC_DUP) { > - scd_bc_tbl[qid].tfd_offset[IWM_TFD_QUEUE_SIZE_MAX + idx] = > w_val; > - bus_dmamap_sync(sc->sc_dmat, sc->sched_dma.map, > - (char *)(void *)(w + IWM_TFD_QUEUE_SIZE_MAX) - > - (char *)(void *)sc->sched_dma.vaddr, > - sizeof (uint16_t), BUS_DMASYNC_PREWRITE); > - } > +void > +iwm_reset_sched(struct iwm_softc *sc, int qid, int idx, uint8_t sta_id) > +{ > + struct iwm_agn_scd_bc_tbl *scd_bc_tbl; > + uint16_t val; > + > + scd_bc_tbl = sc->sched_dma.vaddr; > + > + val = htole16(1 | (sta_id << 12)); > + > + bus_dmamap_sync(sc->sc_dmat, sc->sched_dma.map, > + 0, sc->sched_dma.size, BUS_DMASYNC_PREWRITE); > + > + /* Update TX scheduler. */ > + scd_bc_tbl[qid].tfd_offset[idx] = val; > + if (idx < IWM_TFD_QUEUE_SIZE_BC_DUP) > + scd_bc_tbl[qid].tfd_offset[IWM_TFD_QUEUE_SIZE_MAX + idx] = val; > + > + bus_dmamap_sync(sc->sc_dmat, sc->sched_dma.map, > + 0, sc->sched_dma.size, BUS_DMASYNC_POSTWRITE); > } > -#endif > > /* > * Fill in various bit for management frames, and leave them > @@ -4897,19 +5500,24 @@ iwm_tx(struct iwm_softc *sc, struct mbuf *m, struct ie > uint32_t flags; > u_int hdrlen; > bus_dma_segment_t *seg; > - uint8_t tid, type; > + uint8_t tid, type, subtype; > int i, totlen, err, pad; > - int hdrlen2, rtsthres = ic->ic_rtsthreshold; > + int qid, hasqos, rtsthres = ic->ic_rtsthreshold; > > wh = mtod(m, struct ieee80211_frame *); > - hdrlen = ieee80211_get_hdrlen(wh); > type = wh->i_fc[0] & IEEE80211_FC0_TYPE_MASK; > + subtype = wh->i_fc[0] & IEEE80211_FC0_SUBTYPE_MASK; > + if (type == IEEE80211_FC0_TYPE_CTL) > + hdrlen = sizeof(struct ieee80211_frame_min); > + else > + hdrlen = ieee80211_get_hdrlen(wh); > > - hdrlen2 = (ieee80211_has_qos(wh)) ? > - sizeof (struct ieee80211_qosframe) : > - sizeof (struct ieee80211_frame); > + hasqos = ieee80211_has_qos(wh); > > - tid = 0; > + if (type == IEEE80211_FC0_TYPE_DATA) > + tid = IWM_TID_NON_QOS; > + else > + tid = IWM_MAX_TID_COUNT; > > /* > * Map EDCA categories to Tx data queues. > @@ -4918,14 +5526,31 @@ iwm_tx(struct iwm_softc *sc, struct mbuf *m, struct ie > * need to share Tx queues between stations because we only implement > * client mode; the firmware's station table contains only one entry > * which represents our access point. > - * > - * Tx aggregation will require additional queues (one queue per TID > - * for which aggregation is enabled) but we do not implement this yet. > */ > if (isset(sc->sc_enabled_capa, IWM_UCODE_TLV_CAPA_DQA_SUPPORT)) > - ring = &sc->txq[IWM_DQA_MIN_MGMT_QUEUE + ac]; > + qid = IWM_DQA_MIN_MGMT_QUEUE + ac; > else > - ring = &sc->txq[ac]; > + qid = ac; > + > + /* If possible, put this frame on an aggregation queue. */ > + if (hasqos) { > + struct ieee80211_tx_ba *ba; > + uint16_t qos = ieee80211_get_qos(wh); > + int qostid = qos & IEEE80211_QOS_TID; > + int qosac = ieee80211_up_to_ac(ic, qostid); > + > + ba = &ni->ni_tx_ba[qostid]; > + if (!IEEE80211_IS_MULTICAST(wh->i_addr1) && > + type == IEEE80211_FC0_TYPE_DATA && > + (in->ampdu_tx_tid_mask & (1 << qostid)) && > + ba->ba_state == IEEE80211_BA_AGREED) { > + qid = IWM_FIRST_AGG_TX_QUEUE + qostid; > + tid = qostid; > + ac = qosac; > + } > + } > + > + ring = &sc->txq[qid]; > desc = &ring->desc[ring->cur]; > memset(desc, 0, sizeof(*desc)); > data = &ring->data[ring->cur]; > @@ -5004,14 +5629,28 @@ iwm_tx(struct iwm_softc *sc, struct mbuf *m, struct ie > tx->sta_id = IWM_STATION_ID; > > if (type == IEEE80211_FC0_TYPE_MGT) { > - uint8_t subtype = wh->i_fc[0] & IEEE80211_FC0_SUBTYPE_MASK; > - > if (subtype == IEEE80211_FC0_SUBTYPE_ASSOC_REQ || > subtype == IEEE80211_FC0_SUBTYPE_REASSOC_REQ) > tx->pm_frame_timeout = htole16(3); > else > tx->pm_frame_timeout = htole16(2); > } else { > + if (type == IEEE80211_FC0_TYPE_CTL && > + subtype == IEEE80211_FC0_SUBTYPE_BAR) { > + struct ieee80211_frame_min *wbar; > + uint8_t *frm; > + uint16_t ctl; > + > + flags |= IWM_TX_CMD_FLG_ACK | IWM_TX_CMD_FLG_BAR; > + tx->data_retry_limit = IWM_BAR_DFAULT_RETRY_LIMIT; > + > + wbar = mtod(m, struct ieee80211_frame_min *); > + frm = (uint8_t *)&wbar[1]; > + memcpy(&ctl, frm, sizeof(ctl)); > + tid = (le16toh(ctl) & IEEE80211_BA_TID_INFO_MASK) >> > + IEEE80211_BA_TID_INFO_SHIFT; > + } > + > tx->pm_frame_timeout = htole16(0); > } > > @@ -5058,7 +5697,9 @@ iwm_tx(struct iwm_softc *sc, struct mbuf *m, struct ie > tx->sec_ctl = 0; > } > > - flags |= IWM_TX_CMD_FLG_BT_DIS | IWM_TX_CMD_FLG_SEQ_CTL; > + flags |= IWM_TX_CMD_FLG_BT_DIS; > + if (!hasqos) > + flags |= IWM_TX_CMD_FLG_SEQ_CTL; > > tx->tx_flags |= htole32(flags); > > @@ -5085,9 +5726,11 @@ iwm_tx(struct iwm_softc *sc, struct mbuf *m, struct ie > } > } > data->m = m; > + data->totlen = totlen; > data->in = in; > data->txmcs = ni->ni_txmcs; > data->txrate = ni->ni_txrate; > + data->ampdu_txmcs = ni->ni_txmcs; /* updated upon Tx interrupt */ > > /* Fill TX descriptor. */ > desc->num_tbs = 2 + data->map->dm_nsegs; > @@ -5118,9 +5761,7 @@ iwm_tx(struct iwm_softc *sc, struct mbuf *m, struct ie > (char *)(void *)desc - (char *)(void *)ring->desc_dma.vaddr, > sizeof (*desc), BUS_DMASYNC_PREWRITE); > > -#if 0 > iwm_update_sched(sc, ring->qid, ring->cur, tx->sta_id, > le16toh(tx->len)); > -#endif > > /* Kick TX ring. */ > ring->cur = (ring->cur + 1) % IWM_TX_RING_COUNT; > @@ -5336,6 +5977,7 @@ iwm_add_sta_cmd(struct iwm_softc *sc, struct iwm_node > uint32_t status; > size_t cmdsize; > struct ieee80211com *ic = &sc->sc_ic; > + uint16_t tid_disable_tx = 0xffff; > > if (!update && (sc->sc_flags & IWM_FLAG_STA_ACTIVE)) > panic("STA already added"); > @@ -5362,7 +6004,7 @@ iwm_add_sta_cmd(struct iwm_softc *sc, struct iwm_node > else > qid = IWM_AUX_QUEUE; > add_sta_cmd.tfd_queue_msk |= htole32(1 << qid); > - } else if (!update) { > + } else { > int ac; > for (ac = 0; ac < EDCA_NUM_AC; ac++) { > int qid = ac; > @@ -5371,15 +6013,33 @@ iwm_add_sta_cmd(struct iwm_softc *sc, struct iwm_node > qid += IWM_DQA_MIN_MGMT_QUEUE; > add_sta_cmd.tfd_queue_msk |= htole32(1 << qid); > } > - IEEE80211_ADDR_COPY(&add_sta_cmd.addr, in->in_ni.ni_bssid); > } > + if (!update) { > + if (ic->ic_opmode == IEEE80211_M_MONITOR) > + IEEE80211_ADDR_COPY(&add_sta_cmd.addr, > + etherbroadcastaddr); > + else > + IEEE80211_ADDR_COPY(&add_sta_cmd.addr, > + in->in_ni.ni_bssid); > + } > add_sta_cmd.add_modify = update ? 1 : 0; > add_sta_cmd.station_flags_msk > |= htole32(IWM_STA_FLG_FAT_EN_MSK | IWM_STA_FLG_MIMO_EN_MSK); > - add_sta_cmd.tid_disable_tx = htole16(0xffff); > - if (update) > - add_sta_cmd.modify_mask |= (IWM_STA_MODIFY_TID_DISABLE_TX); > + if (update) { > + int tid, qid; > + for (tid = 0; tid < IWM_MAX_TID_COUNT; tid++) { > + if ((in->ampdu_tx_tid_mask & (1 << tid)) == 0) > + continue; > > + qid = IWM_FIRST_AGG_TX_QUEUE + tid; > + add_sta_cmd.tfd_queue_msk |= htole32(1 << qid); > + tid_disable_tx &= ~(1 << tid); > + add_sta_cmd.modify_mask |= (IWM_STA_MODIFY_QUEUES | > + IWM_STA_MODIFY_TID_DISABLE_TX); > + } > + } > + add_sta_cmd.tid_disable_tx = htole16(tid_disable_tx); > + > if (in->in_ni.ni_flags & IEEE80211_NODE_HT) { > add_sta_cmd.station_flags_msk > |= htole32(IWM_STA_FLG_MAX_AGG_SIZE_MSK | > @@ -5444,7 +6104,7 @@ iwm_add_aux_sta(struct iwm_softc *sc) > if (isset(sc->sc_enabled_capa, IWM_UCODE_TLV_CAPA_DQA_SUPPORT)) { > qid = IWM_DQA_AUX_QUEUE; > err = iwm_enable_txq(sc, IWM_AUX_STA_ID, qid, > - IWM_TX_FIFO_MCAST); > + IWM_TX_FIFO_MCAST, 0, IWM_MAX_TID_COUNT, 0); > } else { > qid = IWM_AUX_QUEUE; > err = iwm_enable_ac_txq(sc, qid, IWM_TX_FIFO_MCAST); > @@ -6582,6 +7242,9 @@ iwm_auth(struct iwm_softc *sc) > > splassert(IPL_NET); > > + in->ampdu_rx_tid_mask = 0; > + in->ampdu_tx_tid_mask = 0; > + > if (ic->ic_opmode == IEEE80211_M_MONITOR) > sc->sc_phyctxt[0].channel = ic->ic_ibss_chan; > else > @@ -7156,11 +7819,7 @@ iwm_setrates(struct iwm_node *in, int async) > > lqcmd.agg_time_limit = htole16(4000); /* 4ms */ > lqcmd.agg_disable_start_th = 3; > -#ifdef notyet > lqcmd.agg_frame_cnt_limit = 0x3f; > -#else > - lqcmd.agg_frame_cnt_limit = 1; /* tx agg disabled */ > -#endif > > cmd.data[0] = &lqcmd; > iwm_send_cmd(sc, &cmd); > @@ -7892,7 +8551,7 @@ iwm_init_hw(struct iwm_softc *sc) > else > qid = IWM_AUX_QUEUE; > err = iwm_enable_txq(sc, IWM_MONITOR_STA_ID, qid, > - iwm_ac_to_tx_fifo[EDCA_AC_BE]); > + iwm_ac_to_tx_fifo[EDCA_AC_BE], 0, IWM_MAX_TID_COUNT, 0); > if (err) { > printf("%s: could not enable monitor inject Tx queue " > "(error %d)\n", DEVNAME(sc), err); > @@ -7906,7 +8565,7 @@ iwm_init_hw(struct iwm_softc *sc) > else > qid = ac; > err = iwm_enable_txq(sc, IWM_STATION_ID, qid, > - iwm_ac_to_tx_fifo[ac]); > + iwm_ac_to_tx_fifo[ac], 0, IWM_TID_NON_QOS, 0); > if (err) { > printf("%s: could not enable Tx queue %d " > "(error %d)\n", DEVNAME(sc), ac, err); > @@ -8578,6 +9237,10 @@ iwm_rx_pkt(struct iwm_softc *sc, struct iwm_rx_data *d > iwm_rx_tx_cmd(sc, pkt, data); > break; > > + case IWM_BA_NOTIF: > + iwm_rx_ba(sc, pkt, data); > + break; > + > case IWM_MISSED_BEACONS_NOTIFICATION: > iwm_rx_bmiss(sc, pkt, data); > break; > @@ -8943,9 +9606,9 @@ iwm_intr(void *arg) > DPRINTF(("driver status:\n")); > for (i = 0; i < IWM_MAX_QUEUES; i++) { > struct iwm_tx_ring *ring = &sc->txq[i]; > - DPRINTF((" tx ring %2d: qid=%-2d cur=%-3d " > + DPRINTF((" tx ring %2d: qid=%-2d tail=%-3d cur=%-3d " > "queued=%-3d\n", > - i, ring->qid, ring->cur, ring->queued)); > + i, ring->qid, ring->tail, ring->cur, ring->queued)); > } > DPRINTF((" rx ring: cur=%d\n", sc->rxq.cur)); > DPRINTF((" 802.11 state %s\n", > @@ -9053,9 +9716,9 @@ iwm_intr_msix(void *arg) > DPRINTF(("driver status:\n")); > for (i = 0; i < IWM_MAX_QUEUES; i++) { > struct iwm_tx_ring *ring = &sc->txq[i]; > - DPRINTF((" tx ring %2d: qid=%-2d cur=%-3d " > + DPRINTF((" tx ring %2d: qid=%-2d tail=%-3d cur=%-3d " > "queued=%-3d\n", > - i, ring->qid, ring->cur, ring->queued)); > + i, ring->qid, ring->tail, ring->cur, ring->queued)); > } > DPRINTF((" rx ring: cur=%d\n", sc->rxq.cur)); > DPRINTF((" 802.11 state %s\n", > @@ -9465,6 +10128,7 @@ iwm_attach(struct device *parent, struct device *self, > > /* Set device capabilities. */ > ic->ic_caps = > + IEEE80211_C_QOS | IEEE80211_C_TX_AMPDU | /* A-MPDU */ > IEEE80211_C_WEP | /* WEP */ > IEEE80211_C_RSN | /* WPA/RSN */ > IEEE80211_C_SCANALL | /* device scans all channels at once */ > @@ -9529,10 +10193,8 @@ iwm_attach(struct device *parent, struct device *self, > ic->ic_update_htprot = iwm_update_htprot; > ic->ic_ampdu_rx_start = iwm_ampdu_rx_start; > ic->ic_ampdu_rx_stop = iwm_ampdu_rx_stop; > -#ifdef notyet > ic->ic_ampdu_tx_start = iwm_ampdu_tx_start; > ic->ic_ampdu_tx_stop = iwm_ampdu_tx_stop; > -#endif > /* > * We cannot read the MAC address without loading the > * firmware from disk. Postpone until mountroot is done. > blob - 201ce69014b9422335a6d698cd4a3cc3f314b2b5 > blob + 1e2e4e01e2a98f60221b72fc6e82a1246f7b9cef > --- sys/dev/pci/if_iwmreg.h > +++ sys/dev/pci/if_iwmreg.h > @@ -1837,6 +1837,9 @@ struct iwm_agn_scd_bc_tbl { > uint16_t tfd_offset[IWM_TFD_QUEUE_BC_SIZE]; > } __packed; > > +#define IWM_TX_CRC_SIZE 4 > +#define IWM_TX_DELIMITER_SIZE 4 > + > /* Maximum number of Tx queues. */ > #define IWM_MAX_QUEUES 31 > > @@ -1875,6 +1878,11 @@ struct iwm_agn_scd_bc_tbl { > #define IWM_DQA_MIN_DATA_QUEUE 10 > #define IWM_DQA_MAX_DATA_QUEUE 31 > > +/* Reserve 8 DQA Tx queues, from 10 up to 17, for A-MPDU aggregation. */ > +#define IWM_MAX_TID_COUNT 8 > +#define IWM_FIRST_AGG_TX_QUEUE IWM_DQA_MIN_DATA_QUEUE > +#define IWM_LAST_AGG_TX_QUEUE (IWM_FIRST_AGG_TX_QUEUE + > IWM_MAX_TID_COUNT - 1) > + > /* legacy non-DQA queues; the legacy command queue uses a different number! > */ > #define IWM_OFFCHANNEL_QUEUE 8 > #define IWM_CMD_QUEUE 9 > @@ -4627,7 +4635,8 @@ struct iwm_lq_cmd { > /* > * TID for non QoS frames - to be written in tid_tspec > */ > -#define IWM_TID_NON_QOS IWM_MAX_TID_COUNT > +#define IWM_TID_NON_QOS 0 > +#define IWM_TID_MGMT 15 > > /* > * Limits on the retransmissions - to be written in {data,rts}_retry_limit > @@ -4898,21 +4907,23 @@ struct iwm_tx_resp { > /** > * struct iwm_ba_notif - notifies about reception of BA > * ( IWM_BA_NOTIF = 0xc5 ) > - * @sta_addr_lo32: lower 32 bits of the MAC address > - * @sta_addr_hi16: upper 16 bits of the MAC address > + * @sta_addr: MAC address > * @sta_id: Index of recipient (BA-sending) station in fw's station table > * @tid: tid of the session > - * @seq_ctl: sequence control field from IEEE80211 frame header (it is > unclear > - * which frame this relates to; info or reverse engineering welcome) > + * @seq_ctl: sequence control field from IEEE80211 frame header (the first > + * bit in @bitmap corresponds to the sequence number stored here) > * @bitmap: the bitmap of the BA notification as seen in the air > * @scd_flow: the tx queue this BA relates to > * @scd_ssn: the index of the last contiguously sent packet > * @txed: number of Txed frames in this batch > * @txed_2_done: number of Acked frames in this batch > + * @reduced_txp: power reduced according to TPC. This is the actual value and > + * not a copy from the LQ command. Thus, if not the first rate was used > + * for Tx-ing then this value will be set to 0 by FW. > + * @reserved1: reserved > */ > struct iwm_ba_notif { > - uint32_t sta_addr_lo32; > - uint16_t sta_addr_hi16; > + uint8_t sta_addr[ETHER_ADDR_LEN]; > uint16_t reserved; > > uint8_t sta_id; > @@ -4923,6 +4934,7 @@ struct iwm_ba_notif { > uint16_t scd_ssn; > uint8_t txed; > uint8_t txed_2_done; > + uint8_t reduced_txp; > uint16_t reserved1; > } __packed; > > blob - 89abe2c1dbdf5ac3ccbf710994380502530ef2a8 > blob + 7d9e26bffe0f1658c771bf85768797c23e94e147 > --- sys/dev/pci/if_iwmvar.h > +++ sys/dev/pci/if_iwmvar.h > @@ -252,14 +252,26 @@ struct iwm_fw_paging { > #define IWM_TX_RING_LOMARK 192 > #define IWM_TX_RING_HIMARK 224 > > +/* For aggregation queues, index must be aligned to frame sequence number. */ > +#define IWM_AGG_SSN_TO_TXQ_IDX(x) ((x) & (IWM_TX_RING_COUNT - 1)) > + > struct iwm_tx_data { > bus_dmamap_t map; > bus_addr_t cmd_paddr; > bus_addr_t scratch_paddr; > struct mbuf *m; > struct iwm_node *in; > + int totlen; > + int retries; > + int txfail; > int txmcs; > int txrate; > + > + /* A-MPDU subframes */ > + int ampdu_id; > + int ampdu_txmcs; > + int ampdu_nframes; > + int ampdu_size; > }; > > struct iwm_tx_ring { > @@ -363,6 +375,12 @@ struct iwm_bf_data { > int last_cqm_event; > }; > > +struct iwm_ba_param { > + uint16_t tid_mask; > + uint16_t ssn[IWM_MAX_TID_COUNT]; > + uint16_t winsize[IWM_MAX_TID_COUNT]; > +}; > + > struct iwm_softc { > struct device sc_dev; > struct ieee80211com sc_ic; > @@ -381,10 +399,15 @@ struct iwm_softc { > > /* Task for firmware BlockAck setup/teardown and its arguments. */ > struct task ba_task; > - int ba_start; > - int ba_tid; > - uint16_t ba_ssn; > - uint16_t ba_winsize; > + int ba_flags; > +#define IWM_RX_BA_START 0x01 > +#define IWM_TX_BA_START 0x02 > +#define IWM_RX_BA_STOP 0x04 > +#define IWM_TX_BA_STOP 0x08 > + struct iwm_ba_param rx_ba_start; > + struct iwm_ba_param rx_ba_stop; > + struct iwm_ba_param tx_ba_start; > + struct iwm_ba_param tx_ba_stop; > > /* Task for HT protection updates. */ > struct task htprot_task; > @@ -407,6 +430,7 @@ struct iwm_softc { > struct iwm_rx_ring rxq; > int qfullmsk; > int cmdqid; > + int qenablemsk; > > int sc_sf_state; > > @@ -551,6 +575,12 @@ struct iwm_node { > int chosen_txrate; > struct ieee80211_mira_node in_mn; > int chosen_txmcs; > + > + uint32_t next_ampdu_id; > + > + /* Currently active Rx/Tx block ack sessions; tracked per TID. */ > + uint8_t ampdu_rx_tid_mask; > + uint8_t ampdu_tx_tid_mask; > }; > #define IWM_STATION_ID 0 > #define IWM_AUX_STA_ID 1 > --