Hi Stefan,
for me the patch works in mode 11n:
before (OpenBSD 6.7-current (GENERIC.MP) #304: Fri Jun 26 02:08:50 MDT 2020)
bandwidth min/avg/max/std-dev = 2.354/12.319/15.391/3.850 Mbps
with patch (OpenBSD 6.7-current (GENERIC.MP) #0: Mon Jun 29 09:35:24 GMT 2020)
bandwidth min/avg/max/std-dev = 12.174/31.411/57.746/15.154 Mbps
iwm0 at pci2 dev 0 function 0 "Intel Dual Band Wireless-AC 8265" rev 0x78, msi
iwm0: hw rev 0x230, fw ver 34.0.1, address 60:f6:77:bc:3a:04
(mode 11g: bandwidth min/avg/max/std-dev = 0.620/0.844/1.101/0.153 Mbps)
mbk Uwe
On 26 Jun 14:45, Stefan Sperling wrote:
> This patch adds support for 11n Tx aggregation to iwm(4).
>
> Please help with testing if you can by running the patch and using wifi
> as usual. Nothing should change, except that Tx speed may potentially
> improve. If you have time to run before/after performance measurements with
> tcpbench or such, that would be nice. But it's not required for testing.
>
> If Tx aggregation is active then netstat will show a non-zero output block ack
> agreement counter:
>
> $ netstat -W iwm0 | grep 'output block'
> 3 new output block ack agreements
> 0 output block ack agreements timed out
>
> It would be great to get at least one test for all the chipsets the driver
> supports: 7260, 7265, 3160, 3165, 3168, 8260, 8265, 9260, 9560
> The behaviour of the access point also matters a great deal. It won't
> hurt to test the same chipset against several different access points.
>
> I have tested this version on 8265 only so far. I've run older revisions
> of this patch on 7265 so I'm confident that this chip will work, too.
> So far, the APs I have tested against are athn(4) in 11a mode and in 11n
> mode with the 'nomimo' nwflag, and a Sagemcom 11ac AP. All on 5Ghz channels.
>
> diff refs/heads/master refs/heads/txagg
> blob - 3a75d07a60a7eb4c66540474e47aeffd7a85250a
> blob + 853bdd1290ad509f5fce7b5bf20550f458a2b460
> --- sys/dev/pci/if_iwm.c
> +++ sys/dev/pci/if_iwm.c
> @@ -144,6 +144,8 @@
> #include <net80211/ieee80211_amrr.h>
> #include <net80211/ieee80211_mira.h>
> #include <net80211/ieee80211_radiotap.h>
> +#include <net80211/ieee80211_priv.h> /* for SEQ_LT */
> +#undef DPRINTF /* defined in ieee80211_priv.h */
>
> #define DEVNAME(_s) ((_s)->sc_dev.dv_xname)
>
> @@ -299,7 +301,8 @@ int iwm_nic_rx_mq_init(struct iwm_softc *);
> int iwm_nic_tx_init(struct iwm_softc *);
> int iwm_nic_init(struct iwm_softc *);
> int iwm_enable_ac_txq(struct iwm_softc *, int, int);
> -int iwm_enable_txq(struct iwm_softc *, int, int, int);
> +int iwm_enable_txq(struct iwm_softc *, int, int, int, int, uint8_t,
> + uint16_t);
> int iwm_post_alive(struct iwm_softc *);
> struct iwm_phy_db_entry *iwm_phy_db_get_section(struct iwm_softc *, uint16_t,
> uint16_t);
> @@ -334,12 +337,12 @@ void iwm_ampdu_rx_stop(struct ieee80211com *, struct
> i
> uint8_t);
> void iwm_sta_rx_agg(struct iwm_softc *, struct ieee80211_node *, uint8_t,
> uint16_t, uint16_t, int);
> -#ifdef notyet
> +void iwm_sta_tx_agg(struct iwm_softc *, struct ieee80211_node *, uint8_t,
> + uint16_t, uint16_t, int);
> int iwm_ampdu_tx_start(struct ieee80211com *, struct ieee80211_node *,
> uint8_t);
> void iwm_ampdu_tx_stop(struct ieee80211com *, struct ieee80211_node *,
> uint8_t);
> -#endif
> void iwm_ba_task(void *);
>
> int iwm_parse_nvm_data(struct iwm_softc *, const uint16_t *,
> @@ -372,14 +375,25 @@ int iwm_rxmq_get_signal_strength(struct iwm_softc
> *, s
> void iwm_rx_rx_phy_cmd(struct iwm_softc *, struct iwm_rx_packet *,
> struct iwm_rx_data *);
> int iwm_get_noise(const struct iwm_statistics_rx_non_phy *);
> +void iwm_txq_advance(struct iwm_softc *, struct iwm_tx_ring *, int);
> +void iwm_ampdu_tx_done(struct iwm_softc *, struct iwm_cmd_header *,
> + struct iwm_node *, struct iwm_tx_ring *, uint32_t, uint8_t,
> + uint8_t, uint16_t, int, struct iwm_agg_tx_status *);
> int iwm_ccmp_decap(struct iwm_softc *, struct mbuf *,
> struct ieee80211_node *);
> void iwm_rx_frame(struct iwm_softc *, struct mbuf *, int, uint32_t, int, int,
> uint32_t, struct ieee80211_rxinfo *, struct mbuf_list *);
> -void iwm_rx_tx_cmd_single(struct iwm_softc *, struct iwm_rx_packet *,
> - struct iwm_node *, int, int);
> +void iwm_rx_tx_cmd_single(struct iwm_softc *, struct iwm_tx_resp *,
> + struct iwm_node *, int, int, int);
> +void iwm_txd_done(struct iwm_softc *, struct iwm_tx_data *);
> void iwm_rx_tx_cmd(struct iwm_softc *, struct iwm_rx_packet *,
> struct iwm_rx_data *);
> +void iwm_clear_oactive(struct iwm_softc *, struct iwm_tx_ring *);
> +void iwm_mira_choose(struct iwm_softc *, struct ieee80211_node *);
> +void iwm_ampdu_rate_control(struct iwm_softc *, struct ieee80211_node *,
> + struct iwm_tx_ring *, int, uint16_t, uint16_t);
> +void iwm_rx_ba(struct iwm_softc *, struct iwm_rx_packet *,
> + struct iwm_rx_data *);
> void iwm_rx_bmiss(struct iwm_softc *, struct iwm_rx_packet *,
> struct iwm_rx_data *);
> int iwm_binding_cmd(struct iwm_softc *, struct iwm_node *, uint32_t);
> @@ -399,6 +413,7 @@ int iwm_send_cmd_pdu_status(struct iwm_softc *,
> uint32
> void iwm_free_resp(struct iwm_softc *, struct iwm_host_cmd *);
> void iwm_cmd_done(struct iwm_softc *, int, int, int);
> void iwm_update_sched(struct iwm_softc *, int, int, uint8_t, uint16_t);
> +void iwm_reset_sched(struct iwm_softc *, int, int, uint8_t);
> const struct iwm_rate *iwm_tx_fill_cmd(struct iwm_softc *, struct iwm_node *,
> struct ieee80211_frame *, struct iwm_tx_cmd *);
> int iwm_tx(struct iwm_softc *, struct mbuf *, struct ieee80211_node *, int);
> @@ -1306,17 +1321,17 @@ iwm_alloc_tx_ring(struct iwm_softc *sc, struct iwm_tx_
> * The command is queue 0 (sc->txq[0]), and 4 mgmt/data frame queues
> * are sc->tqx[IWM_DQA_MIN_MGMT_QUEUE + ac], i.e. sc->txq[5:8],
> * in order to provide one queue per EDCA category.
> + * Tx aggregation requires additional queues, one queue per TID for
> + * which aggregation is enabled. We map TID 0-7 to sc->txq[10:17].
> *
> - * In non-DQA mode, we use rings 0 through 9 (0-3 are EDCA, 9 is cmd).
> + * In non-DQA mode, we use rings 0 through 9 (0-3 are EDCA, 9 is cmd),
> + * and Tx aggregation is not supported.
> *
> - * Tx aggregation will require additional queues (one queue per TID
> - * for which aggregation is enabled) but we do not implement this yet.
> - *
> * Unfortunately, we cannot tell if DQA will be used until the
> * firmware gets loaded later, so just allocate sufficient rings
> * in order to satisfy both cases.
> */
> - if (qid > IWM_CMD_QUEUE)
> + if (qid > IWM_LAST_AGG_TX_QUEUE)
> return 0;
>
> size = IWM_TX_RING_COUNT * sizeof(struct iwm_device_cmd);
> @@ -1380,6 +1395,7 @@ iwm_reset_tx_ring(struct iwm_softc *sc, struct iwm_tx_
> bus_dmamap_sync(sc->sc_dmat, ring->desc_dma.map, 0,
> ring->desc_dma.size, BUS_DMASYNC_PREWRITE);
> sc->qfullmsk &= ~(1 << ring->qid);
> + sc->qenablemsk &= ~(1 << ring->qid);
> /* 7000 family NICs are locked while commands are in progress. */
> if (ring->qid == sc->cmdqid && ring->queued > 0) {
> if (sc->sc_device_family == IWM_DEVICE_FAMILY_7000)
> @@ -2208,6 +2224,18 @@ const uint8_t iwm_ac_to_tx_fifo[] = {
> IWM_TX_FIFO_VO,
> };
>
> +/* Map a TID to an ieee80211_edca_ac category. */
> +const uint8_t iwm_tid_to_ac[IWM_MAX_TID_COUNT] = {
> + EDCA_AC_BE,
> + EDCA_AC_BK,
> + EDCA_AC_BK,
> + EDCA_AC_BE,
> + EDCA_AC_VI,
> + EDCA_AC_VI,
> + EDCA_AC_VO,
> + EDCA_AC_VO,
> +};
> +
> int
> iwm_enable_ac_txq(struct iwm_softc *sc, int qid, int fifo)
> {
> @@ -2250,28 +2278,48 @@ iwm_enable_ac_txq(struct iwm_softc *sc, int qid, int f
> }
>
> int
> -iwm_enable_txq(struct iwm_softc *sc, int sta_id, int qid, int fifo)
> +iwm_enable_txq(struct iwm_softc *sc, int sta_id, int qid, int fifo,
> + int aggregate, uint8_t tid, uint16_t ssn)
> {
> + struct iwm_tx_ring *ring = &sc->txq[qid];
> struct iwm_scd_txq_cfg_cmd cmd;
> - int err;
> + int err, idx, scd_bug;
>
> iwm_nic_assert_locked(sc);
>
> - IWM_WRITE(sc, IWM_HBUS_TARG_WRPTR, qid << 8 | 0);
> + /*
> + * If we need to move the SCD write pointer by steps of
> + * 0x40, 0x80 or 0xc0, it gets stuck.
> + * This is really ugly, but this is the easiest way out for
> + * this sad hardware issue.
> + * This bug has been fixed on devices 9000 and up.
> + */
> + scd_bug = !sc->sc_mqrx_supported &&
> + !((ssn - ring->cur) & 0x3f) &&
> + (ssn != ring->cur);
> + if (scd_bug)
> + ssn = (ssn + 1) & 0xfff;
>
> + idx = IWM_AGG_SSN_TO_TXQ_IDX(ssn);
> + IWM_WRITE(sc, IWM_HBUS_TARG_WRPTR, qid << 8 | idx);
> + ring->cur = idx;
> + ring->tail = idx;
> +
> memset(&cmd, 0, sizeof(cmd));
> + cmd.tid = tid;
> cmd.scd_queue = qid;
> cmd.enable = 1;
> cmd.sta_id = sta_id;
> cmd.tx_fifo = fifo;
> - cmd.aggregate = 0;
> + cmd.aggregate = aggregate;
> + cmd.ssn = htole16(ssn);
> cmd.window = IWM_FRAME_LIMIT;
>
> - err = iwm_send_cmd_pdu(sc, IWM_SCD_QUEUE_CFG, 0,
> - sizeof(cmd), &cmd);
> + err = iwm_send_cmd_pdu(sc, IWM_SCD_QUEUE_CFG, 0, sizeof(cmd), &cmd);
> if (err)
> return err;
>
> + sc->qenablemsk |= (1 << qid);
> return 0;
> }
>
> @@ -2950,8 +2998,12 @@ iwm_sta_rx_agg(struct iwm_softc *sc, struct ieee80211_
> if (start) {
> sc->sc_rx_ba_sessions++;
> ieee80211_addba_req_accept(ic, ni, tid);
> - } else if (sc->sc_rx_ba_sessions > 0)
> - sc->sc_rx_ba_sessions--;
> + in->ampdu_rx_tid_mask |= (1 << tid);
> + } else {
> + in->ampdu_rx_tid_mask &= ~(1 << tid);
> + if (sc->sc_rx_ba_sessions > 0)
> + sc->sc_rx_ba_sessions--;
> + }
> } else if (start)
> ieee80211_addba_req_refuse(ic, ni, tid);
>
> @@ -2959,6 +3011,75 @@ iwm_sta_rx_agg(struct iwm_softc *sc, struct ieee80211_
> }
>
> void
> +iwm_sta_tx_agg(struct iwm_softc *sc, struct ieee80211_node *ni, uint8_t tid,
> + uint16_t ssn, uint16_t winsize, int start)
> +{
> + struct ieee80211com *ic = &sc->sc_ic;
> + struct iwm_node *in = (void *)ni;
> + int qid = IWM_FIRST_AGG_TX_QUEUE + tid;
> + enum ieee80211_edca_ac ac = iwm_tid_to_ac[tid];
> + int fifo = iwm_ac_to_tx_fifo[ac];
> + int err;
> +
> + if (qid > IWM_LAST_AGG_TX_QUEUE || !iwm_nic_lock(sc)) {
> + ieee80211_addba_resp_refuse(ic, ni, tid,
> + IEEE80211_STATUS_UNSPECIFIED);
> + return;
> + }
> +
> + if (start) {
> + if ((sc->qenablemsk & (1 << qid)) == 0) {
> + struct iwm_tx_ring *ring = &sc->txq[qid];
> +
> + err = iwm_enable_txq(sc, IWM_STATION_ID, qid, fifo,
> + 1, tid, ssn);
> + if (err)
> + goto done;
> + /*
> + * If iwm_enable_txq() employed the SCD hardware bug
> + * workaround we must skip the frame with seqnum SSN.
> + */
> + if (IWM_AGG_SSN_TO_TXQ_IDX(ring->cur) !=
> + IWM_AGG_SSN_TO_TXQ_IDX(ssn)) {
> + ssn = (ssn + 1) & 0xfff;
> + KASSERT(IWM_AGG_SSN_TO_TXQ_IDX(ring->cur) ==
> + IWM_AGG_SSN_TO_TXQ_IDX(ssn));
> + ieee80211_output_ba_move_window(ic, ni,
> + tid, ssn);
> + ni->ni_qos_txseqs[tid] = ssn;
> + }
> + }
> + in->ampdu_tx_tid_mask |= (1 << tid);
> + err = iwm_add_sta_cmd(sc, in, 1);
> + if (err) {
> + printf("%s: could not update sta (error %d)\n",
> + DEVNAME(sc), err);
> + in->ampdu_tx_tid_mask &= ~(1 << tid);
> + goto done;
> + }
> + } else {
> + in->ampdu_tx_tid_mask &= ~(1 << tid);
> + err = iwm_add_sta_cmd(sc, in, 1);
> + if (err) {
> + printf("%s: could not update sta (error %d)\n",
> + DEVNAME(sc), err);
> + in->ampdu_tx_tid_mask |= (1 << tid);
> + goto done;
> + }
> + }
> +
> +done:
> + iwm_nic_unlock(sc);
> + if (start) {
> + if (err)
> + ieee80211_addba_resp_refuse(ic, ni, tid,
> + IEEE80211_STATUS_UNSPECIFIED);
> + else
> + ieee80211_addba_resp_accept(ic, ni, tid);
> + }
> +}
> +
> +void
> iwm_htprot_task(void *arg)
> {
> struct iwm_softc *sc = arg;
> @@ -3002,19 +3123,53 @@ iwm_ba_task(void *arg)
> struct ieee80211com *ic = &sc->sc_ic;
> struct ieee80211_node *ni = ic->ic_bss;
> int s = splnet();
> + int tid;
>
> - if (sc->sc_flags & IWM_FLAG_SHUTDOWN) {
> + if ((sc->sc_flags & IWM_FLAG_SHUTDOWN) ||
> + ic->ic_state != IEEE80211_S_RUN) {
> refcnt_rele_wake(&sc->task_refs);
> splx(s);
> return;
> }
>
> - if (sc->ba_start)
> - iwm_sta_rx_agg(sc, ni, sc->ba_tid, sc->ba_ssn,
> - sc->ba_winsize, 1);
> - else
> - iwm_sta_rx_agg(sc, ni, sc->ba_tid, 0, 0, 0);
> + if (sc->ba_flags & IWM_RX_BA_START) {
> + for (tid = 0; tid < IWM_MAX_TID_COUNT; tid++) {
> + if ((sc->rx_ba_start.tid_mask & (1 << tid)) == 0)
> + continue;
> + iwm_sta_rx_agg(sc, ni, tid, sc->rx_ba_start.ssn[tid],
> + sc->rx_ba_start.winsize[tid], 1);
> + sc->rx_ba_start.tid_mask &= ~(1 << tid);
> + }
> + }
> +
> + if (sc->ba_flags & IWM_RX_BA_STOP) {
> + for (tid = 0; tid < IWM_MAX_TID_COUNT; tid++) {
> + if ((sc->rx_ba_stop.tid_mask & (1 << tid)) == 0)
> + continue;
> + iwm_sta_rx_agg(sc, ni, tid, 0, 0, 0);
> + sc->rx_ba_stop.tid_mask &= ~(1 << tid);
> + }
> + }
>
> + if (sc->ba_flags & IWM_TX_BA_START) {
> + for (tid = 0; tid < IWM_MAX_TID_COUNT; tid++) {
> + if ((sc->tx_ba_start.tid_mask & (1 << tid)) == 0)
> + continue;
> + iwm_sta_tx_agg(sc, ni, tid, sc->tx_ba_start.ssn[tid],
> + sc->tx_ba_start.winsize[tid], 1);
> + sc->tx_ba_start.tid_mask &= ~(1 << tid);
> + }
> + }
> +
> + if (sc->ba_flags & IWM_TX_BA_STOP) {
> + for (tid = 0; tid < IWM_MAX_TID_COUNT; tid++) {
> + if ((sc->tx_ba_stop.tid_mask & (1 << tid)) == 0)
> + continue;
> + iwm_sta_tx_agg(sc, ni, tid, 0, 0, 0);
> + sc->tx_ba_stop.tid_mask &= ~(1 << tid);
> + }
> + }
> +
> refcnt_rele_wake(&sc->task_refs);
> splx(s);
> }
> @@ -3029,14 +3184,23 @@ iwm_ampdu_rx_start(struct ieee80211com *ic, struct iee
> {
> struct ieee80211_rx_ba *ba = &ni->ni_rx_ba[tid];
> struct iwm_softc *sc = IC2IFP(ic)->if_softc;
> + struct iwm_node *in = (void *)ni;
>
> - if (sc->sc_rx_ba_sessions >= IWM_MAX_RX_BA_SESSIONS)
> + if (tid >= IWM_MAX_TID_COUNT)
> + return EINVAL;
> +
> + if (sc->sc_rx_ba_sessions >= IWM_MAX_RX_BA_SESSIONS ||
> + (in->ampdu_rx_tid_mask & (1 << tid)) != 0)
> return ENOSPC;
>
> - sc->ba_start = 1;
> - sc->ba_tid = tid;
> - sc->ba_ssn = htole16(ba->ba_winstart);
> - sc->ba_winsize = htole16(ba->ba_winsize);
> + if (sc->rx_ba_start.tid_mask & (1 << tid) ||
> + sc->rx_ba_stop.tid_mask & (1 << tid))
> + return EAGAIN;
> +
> + sc->ba_flags |= IWM_RX_BA_START;
> + sc->rx_ba_start.tid_mask |= (1 << tid);
> + sc->rx_ba_start.ssn[tid] = htole16(ba->ba_winstart);
> + sc->rx_ba_start.winsize[tid] = htole16(ba->ba_winsize);
> iwm_add_task(sc, systq, &sc->ba_task);
>
> return EBUSY;
> @@ -3051,13 +3215,69 @@ iwm_ampdu_rx_stop(struct ieee80211com *ic, struct ieee
> uint8_t tid)
> {
> struct iwm_softc *sc = IC2IFP(ic)->if_softc;
> + struct iwm_node *in = (void *)ni;
>
> - sc->ba_start = 0;
> - sc->ba_tid = tid;
> + if (in->ampdu_rx_tid_mask & (1 << tid)) {
> + sc->ba_flags |= IWM_RX_BA_STOP;
> + sc->rx_ba_stop.tid_mask |= (1 << tid);
> + iwm_add_task(sc, systq, &sc->ba_task);
> + }
> +}
> +
> +int
> +iwm_ampdu_tx_start(struct ieee80211com *ic, struct ieee80211_node *ni,
> + uint8_t tid)
> +{
> + struct iwm_softc *sc = IC2IFP(ic)->if_softc;
> + struct iwm_node *in = (void *)ni;
> + struct ieee80211_tx_ba *ba = &ni->ni_tx_ba[tid];
> +
> + /* We only implement Tx aggregation with DQA-capable firmware. */
> + if (!isset(sc->sc_enabled_capa, IWM_UCODE_TLV_CAPA_DQA_SUPPORT))
> + return ENOTSUP;
> +
> + /* Ensure we can map this TID to an aggregation queue. */
> + if (tid >= IWM_MAX_TID_COUNT)
> + return EINVAL;
> +
> + /* We only support a fixed Tx aggregation window size, for now. */
> + if (ba->ba_winsize != IWM_FRAME_LIMIT)
> + return ENOTSUP;
> +
> + if ((in->ampdu_tx_tid_mask & (1 << tid)) != 0)
> + return ENOSPC;
> +
> + if (sc->tx_ba_start.tid_mask & (1 << tid) ||
> + sc->tx_ba_stop.tid_mask & (1 << tid))
> + return EAGAIN;
> +
> + sc->ba_flags |= IWM_TX_BA_START;
> + sc->tx_ba_start.tid_mask |= (1 << tid);
> + sc->tx_ba_start.ssn[tid] = htole16(ba->ba_winstart);
> + sc->tx_ba_start.winsize[tid] = htole16(ba->ba_winsize);
> iwm_add_task(sc, systq, &sc->ba_task);
> +
> + return EBUSY;
> }
>
> void
> +iwm_ampdu_tx_stop(struct ieee80211com *ic, struct ieee80211_node *ni,
> + uint8_t tid)
> +{
> + struct iwm_softc *sc = IC2IFP(ic)->if_softc;
> + struct iwm_node *in = (void *)ni;
> +
> + if (tid >= IWM_MAX_TID_COUNT)
> + return;
> +
> + if (in->ampdu_tx_tid_mask & (1 << tid)) {
> + sc->ba_flags |= IWM_TX_BA_STOP;
> + sc->tx_ba_stop.tid_mask |= (1 << tid);
> + iwm_add_task(sc, systq, &sc->ba_task);
> + }
> +}
> +
> +void
> iwm_set_hw_address_8000(struct iwm_softc *sc, struct iwm_nvm_data *data,
> const uint16_t *mac_override, const uint16_t *nvm_hw)
> {
> @@ -4238,13 +4458,178 @@ iwm_rx_mpdu_mq(struct iwm_softc *sc, struct mbuf *m,
> v
> }
>
> void
> -iwm_rx_tx_cmd_single(struct iwm_softc *sc, struct iwm_rx_packet *pkt,
> - struct iwm_node *in, int txmcs, int txrate)
> +iwm_txq_advance(struct iwm_softc *sc, struct iwm_tx_ring *ring, int idx)
> {
> + struct iwm_tx_data *txd;
> +
> + while (ring->tail != idx) {
> + txd = &ring->data[ring->tail];
> + if (txd->m != NULL) {
> + if (ring->qid < IWM_FIRST_AGG_TX_QUEUE)
> + DPRINTF(("%s: missed Tx completion: tail=%d "
> + "idx=%d\n", __func__, ring->tail, idx));
> + iwm_reset_sched(sc, ring->qid, ring->tail,
> IWM_STATION_ID);
> + iwm_txd_done(sc, txd);
> + ring->queued--;
> + }
> + ring->tail = (ring->tail + 1) % IWM_TX_RING_COUNT;
> + }
> +}
> +
> +void
> +iwm_ampdu_tx_done(struct iwm_softc *sc, struct iwm_cmd_header *cmd_hdr,
> + struct iwm_node *in, struct iwm_tx_ring *txq, uint32_t initial_rate,
> + uint8_t nframes, uint8_t failure_frame, uint16_t ssn, int status,
> + struct iwm_agg_tx_status *agg_status)
> +{
> struct ieee80211com *ic = &sc->sc_ic;
> + int tid = cmd_hdr->qid - IWM_FIRST_AGG_TX_QUEUE;
> + struct iwm_tx_data *txdata = &txq->data[cmd_hdr->idx];
> struct ieee80211_node *ni = &in->in_ni;
> + int txfail = (status != IWM_TX_STATUS_SUCCESS &&
> + status != IWM_TX_STATUS_DIRECT_DONE);
> + struct ieee80211_tx_ba *ba;
> +
> + sc->sc_tx_timer = 0;
> +
> + if (ic->ic_state != IEEE80211_S_RUN)
> + return;
> +
> + if (nframes > 1) {
> + int ampdu_id, have_ampdu_id = 0, ampdu_size = 0;
> + int i;
> +
> + /* Compute the size of this A-MPDU. */
> + for (i = 0; i < nframes; i++) {
> + uint8_t qid = agg_status[i].qid;
> + uint8_t idx = agg_status[i].idx;
> +
> + if (qid != cmd_hdr->qid)
> + continue;
> +
> + txdata = &txq->data[idx];
> + if (txdata->m == NULL)
> + continue;
> +
> + ampdu_size += txdata->totlen + IEEE80211_CRC_LEN;
> + }
> +
> + /*
> + * For each subframe collect Tx status, retries, and Tx rate.
> + * (The Tx rate is the same for all subframes in this batch.)
> + */
> + for (i = 0; i < nframes; i++) {
> + uint8_t qid = agg_status[i].qid;
> + uint8_t idx = agg_status[i].idx;
> + uint16_t txstatus = (le16toh(agg_status[i].status) &
> + IWM_AGG_TX_STATE_STATUS_MSK);
> + uint16_t trycnt = (le16toh(agg_status[i].status) &
> + IWM_AGG_TX_STATE_TRY_CNT_MSK) >>
> + IWM_AGG_TX_STATE_TRY_CNT_POS;
> +
> + if (qid != cmd_hdr->qid)
> + continue;
> +
> + txdata = &txq->data[idx];
> + if (txdata->m == NULL)
> + continue;
> +
> + if (initial_rate & IWM_RATE_MCS_HT_MSK)
> + txdata->ampdu_txmcs = (initial_rate &
> + (IWM_RATE_HT_MCS_RATE_CODE_MSK |
> + IWM_RATE_HT_MCS_NSS_MSK));
> + if (txstatus != IWM_AGG_TX_STATE_TRANSMITTED)
> + txdata->txfail++;
> + if (trycnt > 1)
> + txdata->retries++;
> +
> + /*
> + * Assign a common ID to all subframes of this A-MPDU.
> + * This ID will be used during Tx rate control to
> + * infer the ACK status of individual subframes.
> + */
> + if (!have_ampdu_id) {
> + ampdu_id = txdata->in->next_ampdu_id++;
> + have_ampdu_id = 1;
> + }
> + txdata->ampdu_id = ampdu_id;
> +
> + /*
> + * We will also need to know the total number of
> + * subframes and the size of this A-MPDU. We store
> + * this redundantly on each subframe because firmware
> + * only reports acknowledged subframes via compressed
> + * block-ack notification. This way we will know what
> + * the total number of subframes and size were even if
> + * just one of these subframes gets acknowledged.
> + */
> + txdata->ampdu_nframes = nframes;
> + txdata->ampdu_size = ampdu_size;
> + }
> + return;
> + }
> +
> + if (ni == NULL)
> + return;
> +
> + ba = &ni->ni_tx_ba[tid];
> + if (ba->ba_state != IEEE80211_BA_AGREED)
> + return;
> +
> + /* This is a final single-frame Tx attempt. */
> + DPRINTFN(3, ("%s: final tx status=0x%x qid=%d queued=%d idx=%d ssn=%u "
> + "bitmap=0x%llx\n", __func__, status, desc->qid, txq->queued,
> + desc->idx, ssn, ba->ba_bitmap));
> +
> + /*
> + * Skip rate control if our Tx rate is fixed.
> + * Don't report frames to MiRA which were sent at a different
> + * Tx rate than ni->ni_txmcs.
> + */
> + if (ic->ic_fixed_mcs == -1 && txdata->txmcs == ni->ni_txmcs) {
> + in->in_mn.frames++;
> + in->in_mn.agglen = 1;
> + in->in_mn.ampdu_size = txdata->totlen + IEEE80211_CRC_LEN;
> + if (failure_frame > 0)
> + in->in_mn.retries++;
> + if (txfail)
> + in->in_mn.txfail++;
> + iwm_mira_choose(sc, ni);
> + }
> +
> + if (txfail)
> + ieee80211_tx_compressed_bar(ic, ni, tid, ssn);
> + else if (!SEQ_LT(ssn, ba->ba_winstart)) {
> + /*
> + * Move window forward if SSN lies beyond end of window,
> + * otherwise we can't record the ACK for this frame.
> + * Non-acked frames which left holes in the bitmap near
> + * the beginning of the window must be discarded.
> + */
> + uint16_t s = ssn;
> + while (SEQ_LT(ba->ba_winend, s)) {
> + ieee80211_output_ba_move_window(ic, ni, tid, s);
> + iwm_txq_advance(sc, txq, IWM_AGG_SSN_TO_TXQ_IDX(s));
> + s = (s + 1) % 0xfff;
> + }
> + /* SSN should now be within window; set corresponding bit. */
> + ieee80211_output_ba_record_ack(ic, ni, tid, ssn);
> + }
> +
> + /* Move window forward up to the first hole in the bitmap. */
> + ieee80211_output_ba_move_window_to_first_unacked(ic, ni, tid, ssn);
> + iwm_txq_advance(sc, txq, IWM_AGG_SSN_TO_TXQ_IDX(ba->ba_winstart));
> +
> + iwm_clear_oactive(sc, txq);
> +}
> +
> +void
> +iwm_rx_tx_cmd_single(struct iwm_softc *sc, struct iwm_tx_resp *tx_resp,
> + struct iwm_node *in, int txmcs, int txrate, int qid)
> +{
> + struct ieee80211com *ic = &sc->sc_ic;
> + struct ieee80211_node *ni = &in->in_ni;
> struct ifnet *ifp = IC2IFP(ic);
> - struct iwm_tx_resp *tx_resp = (void *)pkt->data;
> int status = le16toh(tx_resp->status.status) & IWM_TX_STATUS_MSK;
> int txfail;
>
> @@ -4277,22 +4662,8 @@ iwm_rx_tx_cmd_single(struct iwm_softc *sc, struct iwm_
> in->in_mn.retries += tx_resp->failure_frame;
> if (txfail)
> in->in_mn.txfail += tx_resp->frame_count;
> - if (ic->ic_state == IEEE80211_S_RUN) {
> - int best_mcs;
> -
> - ieee80211_mira_choose(&in->in_mn, ic, &in->in_ni);
> - /*
> - * If MiRA has chosen a new TX rate we must update
> - * the firwmare's LQ rate table from process context.
> - * ni_txmcs may change again before the task runs so
> - * cache the chosen rate in the iwm_node structure.
> - */
> - best_mcs = ieee80211_mira_get_best_mcs(&in->in_mn);
> - if (best_mcs != in->chosen_txmcs) {
> - in->chosen_txmcs = best_mcs;
> - iwm_setrates(in, 1);
> - }
> - }
> + if (ic->ic_state == IEEE80211_S_RUN)
> + iwm_mira_choose(sc, ni);
> }
>
> if (txfail)
> @@ -4313,49 +4684,91 @@ iwm_txd_done(struct iwm_softc *sc, struct iwm_tx_data
> KASSERT(txd->in);
> ieee80211_release_node(ic, &txd->in->in_ni);
> txd->in = NULL;
> +
> + txd->retries = 0;
> + txd->txfail = 0;
> + txd->txmcs = 0;
> + txd->ampdu_txmcs = 0;
> + txd->txrate = 0;
> }
>
> void
> iwm_rx_tx_cmd(struct iwm_softc *sc, struct iwm_rx_packet *pkt,
> struct iwm_rx_data *data)
> {
> - struct ieee80211com *ic = &sc->sc_ic;
> - struct ifnet *ifp = IC2IFP(ic);
> struct iwm_cmd_header *cmd_hdr = &pkt->hdr;
> int idx = cmd_hdr->idx;
> int qid = cmd_hdr->qid;
> struct iwm_tx_ring *ring = &sc->txq[qid];
> struct iwm_tx_data *txd;
> + struct iwm_tx_resp *tx_resp = (void *)pkt->data;
> + uint32_t ssn;
> + uint32_t len = iwm_rx_packet_len(pkt);
>
> bus_dmamap_sync(sc->sc_dmat, data->map, 0, IWM_RBUF_SIZE,
> BUS_DMASYNC_POSTREAD);
>
> sc->sc_tx_timer = 0;
>
> + /* Sanity checks. */
> + if (sizeof(*tx_resp) > len)
> + return;
> + if (qid < IWM_FIRST_AGG_TX_QUEUE && tx_resp->frame_count > 1)
> + return;
> + if (qid >= IWM_FIRST_AGG_TX_QUEUE && sizeof(*tx_resp) + sizeof(ssn) +
> + tx_resp->frame_count * sizeof(tx_resp->status) > len)
> + return;
> +
> + /*
> + * In the multi-frame case the firmware has just transmitted a bunch
> + * of frames in an A-MPDU. The final Tx status of those frames won't
> + * be known until the peer ACKs subframes with a block ack or firmware
> + * gives up on a particular subframe.
> + * Subframes for which the firmware never sees an ACK will be retried
> + * and will eventually arrive here as a single-frame Tx failure.
> + * So there is nothing to do, for now.
> + */
> + if (tx_resp->frame_count != 1)
> + return;
> +
> txd = &ring->data[idx];
> if (txd->m == NULL)
> return;
>
> - iwm_rx_tx_cmd_single(sc, pkt, txd->in, txd->txmcs, txd->txrate);
> - iwm_txd_done(sc, txd);
> + if (qid >= IWM_FIRST_AGG_TX_QUEUE) {
> + int status;
>
> - /*
> - * XXX Sometimes we miss Tx completion interrupts.
> - * We cannot check Tx success/failure for affected frames; just free
> - * the associated mbuf and release the associated node reference.
> - */
> - while (ring->tail != idx) {
> - txd = &ring->data[ring->tail];
> - if (txd->m != NULL) {
> - DPRINTF(("%s: missed Tx completion: tail=%d idx=%d\n",
> - __func__, ring->tail, idx));
> - iwm_txd_done(sc, txd);
> - ring->queued--;
> - }
> - ring->tail = (ring->tail + 1) % IWM_TX_RING_COUNT;
> + memcpy(&ssn, &tx_resp->status + tx_resp->frame_count,
> sizeof(ssn));
> + ssn = le32toh(ssn) & 0xfff;
> + status = le16toh(tx_resp->status.status) & IWM_TX_STATUS_MSK;
> + iwm_ampdu_tx_done(sc, cmd_hdr, txd->in, ring,
> + le32toh(tx_resp->initial_rate), tx_resp->frame_count,
> + tx_resp->failure_frame, ssn, status, &tx_resp->status);
> + } else {
> + iwm_rx_tx_cmd_single(sc, tx_resp, txd->in, txd->txmcs,
> + txd->txrate, qid);
> + iwm_reset_sched(sc, qid, idx, IWM_STATION_ID);
> + iwm_txd_done(sc, txd);
> + ring->queued--;
> +
> + /*
> + * XXX Sometimes we miss Tx completion interrupts.
> + * We cannot check Tx success/failure for affected frames;
> + * just free the associated mbuf and release the associated
> + * node reference.
> + */
> + iwm_txq_advance(sc, ring, idx);
> + iwm_clear_oactive(sc, ring);
> }
> +}
>
> - if (--ring->queued < IWM_TX_RING_LOMARK) {
> +void
> +iwm_clear_oactive(struct iwm_softc *sc, struct iwm_tx_ring *ring)
> +{
> + struct ieee80211com *ic = &sc->sc_ic;
> + struct ifnet *ifp = IC2IFP(ic);
> +
> + if (ring->queued < IWM_TX_RING_LOMARK) {
> sc->qfullmsk &= ~(1 << ring->qid);
> if (sc->qfullmsk == 0 && ifq_is_oactive(&ifp->if_snd)) {
> ifq_clr_oactive(&ifp->if_snd);
> @@ -4370,6 +4783,183 @@ iwm_rx_tx_cmd(struct iwm_softc *sc, struct iwm_rx_pack
> }
>
> void
> +iwm_mira_choose(struct iwm_softc *sc, struct ieee80211_node *ni)
> +{
> + struct ieee80211com *ic = &sc->sc_ic;
> + struct iwm_node *in = (void *)ni;
> + int best_mcs = ieee80211_mira_get_best_mcs(&in->in_mn);
> +
> + ieee80211_mira_choose(&in->in_mn, ic, ni);
> +
> + /*
> + * Update firmware's LQ retry table if MiRA has chosen a new MCS.
> + *
> + * We only need to do this if the best MCS has changed because
> + * we ask firmware to use a fixed MCS while MiRA is probing a
> + * candidate MCS.
> + * While not probing we ask firmware to retry at lower rates in case
> + * Tx at the newly chosen best MCS ends up failing, and then report
> + * any resulting Tx retries to MiRA in order to trigger probing.
> + */
> + if (best_mcs != ieee80211_mira_get_best_mcs(&in->in_mn)) {
> + in->chosen_txmcs = best_mcs;
> + iwm_setrates(in, 1);
> + }
> +}
> +
> +void
> +iwm_ampdu_rate_control(struct iwm_softc *sc, struct ieee80211_node *ni,
> + struct iwm_tx_ring *txq, int tid, uint16_t seq, uint16_t ssn)
> +{
> + struct ieee80211com *ic = &sc->sc_ic;
> + struct iwm_node *in = (void *)ni;
> + struct ieee80211_tx_ba *ba = &ni->ni_tx_ba[tid];
> + int min_ampdu_id, max_ampdu_id, id;
> + int idx, end_idx;
> +
> + /* Determine the min/max IDs we assigned to AMPDUs in this range. */
> + idx = IWM_AGG_SSN_TO_TXQ_IDX(seq);
> + end_idx = IWM_AGG_SSN_TO_TXQ_IDX(ssn);
> + min_ampdu_id = txq->data[idx].ampdu_id;
> + max_ampdu_id = min_ampdu_id;
> + while (idx != end_idx) {
> + struct iwm_tx_data *txdata = &txq->data[idx];
> +
> + if (txdata->m != NULL) {
> + if (min_ampdu_id > txdata->ampdu_id)
> + min_ampdu_id = txdata->ampdu_id;
> + if (max_ampdu_id < txdata->ampdu_id)
> + max_ampdu_id = txdata->ampdu_id;
> + }
> +
> + idx = (idx + 1) % IWM_TX_RING_COUNT;
> + }
> +
> + /*
> + * Update Tx rate statistics for A-MPDUs before firmware's BA window.
> + */
> + for (id = min_ampdu_id; id <= max_ampdu_id; id++) {
> + int have_ack = 0, bit = 0;
> + idx = IWM_AGG_SSN_TO_TXQ_IDX(seq);
> + end_idx = IWM_AGG_SSN_TO_TXQ_IDX(ssn);
> + in->in_mn.agglen = 0;
> + in->in_mn.ampdu_size = 0;
> + while (idx != end_idx) {
> + struct iwm_tx_data *txdata = &txq->data[idx];
> + uint16_t s = (seq + bit) & 0xfff;
> + /*
> + * We can assume that this subframe has been ACKed
> + * because ACK failures come as single frames and
> + * before failing an A-MPDU subframe the firmware
> + * sends it as a single frame at least once.
> + *
> + * However, when this A-MPDU was transmitted we
> + * learned how many subframes it contained.
> + * So if firmware isn't reporting all subframes now
> + * we can deduce an ACK failure for missing frames.
> + */
> + if (txdata->m != NULL && txdata->ampdu_id == id &&
> + txdata->ampdu_txmcs == ni->ni_txmcs &&
> + txdata->ampdu_nframes > 0 &&
> + (SEQ_LT(ba->ba_winend, s) ||
> + (ba->ba_bitmap & (1 << bit)) == 0)) {
> + have_ack++;
> + in->in_mn.frames = txdata->ampdu_nframes;
> + in->in_mn.agglen = txdata->ampdu_nframes;
> + in->in_mn.ampdu_size = txdata->ampdu_size;
> + if (txdata->retries > 1)
> + in->in_mn.retries++;
> + if (!SEQ_LT(ba->ba_winend, s))
> + ieee80211_output_ba_record_ack(ic, ni,
> + tid, s);
> + }
> +
> + idx = (idx + 1) % IWM_TX_RING_COUNT;
> + bit++;
> + }
> +
> + if (have_ack > 0) {
> + in->in_mn.txfail = in->in_mn.frames - have_ack;
> + iwm_mira_choose(sc, ni);
> + }
> + }
> +}
> +
> +void
> +iwm_rx_ba(struct iwm_softc *sc, struct iwm_rx_packet *pkt,
> + struct iwm_rx_data *data)
> +{
> + struct iwm_ba_notif *ban = (void *)pkt->data;
> + struct ieee80211com *ic = &sc->sc_ic;
> + struct ieee80211_node *ni;
> + struct ieee80211_tx_ba *ba;
> + struct iwm_node *in;
> + struct iwm_tx_ring *ring;
> + uint16_t seq, ssn, idx;
> + int qid;
> +
> + if (ic->ic_state != IEEE80211_S_RUN)
> + return;
> +
> + if (iwm_rx_packet_payload_len(pkt) < sizeof(*ban))
> + return;
> +
> + if (ban->sta_id != IWM_STATION_ID ||
> + !IEEE80211_ADDR_EQ(ic->ic_bss->ni_macaddr, ban->sta_addr))
> + return;
> +
> + ni = ic->ic_bss;
> + in = (void *)ni;
> +
> + qid = le16toh(ban->scd_flow);
> + if (qid < IWM_FIRST_AGG_TX_QUEUE || qid > IWM_LAST_AGG_TX_QUEUE)
> + return;
> +
> + /* Protect against a firmware bug where the queue/TID are off. */
> + if (qid != IWM_FIRST_AGG_TX_QUEUE + ban->tid)
> + return;
> +
> + ba = &ni->ni_tx_ba[ban->tid];
> + if (ba->ba_state != IEEE80211_BA_AGREED)
> + return;
> +
> + ring = &sc->txq[qid];
> + ssn = le16toh(ban->scd_ssn); /* BA window starting sequence number */
> + idx = IWM_AGG_SSN_TO_TXQ_IDX(ssn);
> +
> + /*
> + * The first bit in ban->bitmap corresponds to the sequence number
> + * stored in the sequence control field ban->seq_ctl.
> + * Multiple BA notifications in a row may be using this number, with
> + * additional bits being set in cba->bitmap. It is unclear how the
> + * firmware decides to shift this window forward.
> + */
> + seq = le16toh(ban->seq_ctl) >> IEEE80211_SEQ_SEQ_SHIFT;
> +
> + /*
> + * The firmware's new BA window starting sequence number
> + * corresponds to the first hole in ban->scd_ssn, implying
> + * that all frames between 'seq' and 'ssn' have been acked.
> + */
> + ssn = le16toh(ban->scd_ssn);
> +
> + /* Skip rate control if our Tx rate is fixed. */
> + if (ic->ic_fixed_mcs != -1)
> + iwm_ampdu_rate_control(sc, ni, ring, ban->tid, seq, ssn);
> +
> + /*
> + * SSN corresponds to the first (perhaps not yet transmitted) frame
> + * in firmware's BA window. Firmware is not going to retransmit any
> + * frames before its BA window so mark them all as done.
> + */
> + if (SEQ_LT(ba->ba_winstart, ssn)) {
> + ieee80211_output_ba_move_window(ic, ni, ban->tid, ssn);
> + iwm_txq_advance(sc, ring, IWM_AGG_SSN_TO_TXQ_IDX(ssn));
> + iwm_clear_oactive(sc, ring);
> + }
> +}
> +
> +void
> iwm_rx_bmiss(struct iwm_softc *sc, struct iwm_rx_packet *pkt,
> struct iwm_rx_data *data)
> {
> @@ -4638,9 +5228,6 @@ iwm_send_cmd(struct iwm_softc *sc, struct iwm_host_cmd
> }
> }
>
> -#if 0
> - iwm_update_sched(sc, ring->qid, ring->cur, 0, 0);
> -#endif
> /* Kick command ring. */
> ring->queued++;
> ring->cur = (ring->cur + 1) % IWM_TX_RING_COUNT;
> @@ -4771,7 +5358,6 @@ iwm_cmd_done(struct iwm_softc *sc, int qid, int idx, i
> }
> }
>
> -#if 0
> /*
> * necessary only for block ack mode
> */
> @@ -4780,32 +5366,49 @@ iwm_update_sched(struct iwm_softc *sc, int qid, int id
> uint16_t len)
> {
> struct iwm_agn_scd_bc_tbl *scd_bc_tbl;
> - uint16_t w_val;
> + uint16_t val;
>
> scd_bc_tbl = sc->sched_dma.vaddr;
>
> - len += 8; /* magic numbers came naturally from paris */
> + len += IWM_TX_CRC_SIZE + IWM_TX_DELIMITER_SIZE;
> if (sc->sc_capaflags & IWM_UCODE_TLV_FLAGS_DW_BC_TABLE)
> len = roundup(len, 4) / 4;
>
> - w_val = htole16(sta_id << 12 | len);
> + val = htole16(sta_id << 12 | len);
>
> + bus_dmamap_sync(sc->sc_dmat, sc->sched_dma.map,
> + 0, sc->sched_dma.size, BUS_DMASYNC_PREWRITE);
> +
> /* Update TX scheduler. */
> - scd_bc_tbl[qid].tfd_offset[idx] = w_val;
> + scd_bc_tbl[qid].tfd_offset[idx] = val;
> + if (idx < IWM_TFD_QUEUE_SIZE_BC_DUP)
> + scd_bc_tbl[qid].tfd_offset[IWM_TFD_QUEUE_SIZE_MAX + idx] = val;
> +
> bus_dmamap_sync(sc->sc_dmat, sc->sched_dma.map,
> - (char *)(void *)w - (char *)(void *)sc->sched_dma.vaddr,
> - sizeof(uint16_t), BUS_DMASYNC_PREWRITE);
> + 0, sc->sched_dma.size, BUS_DMASYNC_POSTWRITE);
> +}
>
> - /* I really wonder what this is ?!? */
> - if (idx < IWM_TFD_QUEUE_SIZE_BC_DUP) {
> - scd_bc_tbl[qid].tfd_offset[IWM_TFD_QUEUE_SIZE_MAX + idx] =
> w_val;
> - bus_dmamap_sync(sc->sc_dmat, sc->sched_dma.map,
> - (char *)(void *)(w + IWM_TFD_QUEUE_SIZE_MAX) -
> - (char *)(void *)sc->sched_dma.vaddr,
> - sizeof (uint16_t), BUS_DMASYNC_PREWRITE);
> - }
> +void
> +iwm_reset_sched(struct iwm_softc *sc, int qid, int idx, uint8_t sta_id)
> +{
> + struct iwm_agn_scd_bc_tbl *scd_bc_tbl;
> + uint16_t val;
> +
> + scd_bc_tbl = sc->sched_dma.vaddr;
> +
> + val = htole16(1 | (sta_id << 12));
> +
> + bus_dmamap_sync(sc->sc_dmat, sc->sched_dma.map,
> + 0, sc->sched_dma.size, BUS_DMASYNC_PREWRITE);
> +
> + /* Update TX scheduler. */
> + scd_bc_tbl[qid].tfd_offset[idx] = val;
> + if (idx < IWM_TFD_QUEUE_SIZE_BC_DUP)
> + scd_bc_tbl[qid].tfd_offset[IWM_TFD_QUEUE_SIZE_MAX + idx] = val;
> +
> + bus_dmamap_sync(sc->sc_dmat, sc->sched_dma.map,
> + 0, sc->sched_dma.size, BUS_DMASYNC_POSTWRITE);
> }
> -#endif
>
> /*
> * Fill in various bit for management frames, and leave them
> @@ -4897,19 +5500,24 @@ iwm_tx(struct iwm_softc *sc, struct mbuf *m, struct ie
> uint32_t flags;
> u_int hdrlen;
> bus_dma_segment_t *seg;
> - uint8_t tid, type;
> + uint8_t tid, type, subtype;
> int i, totlen, err, pad;
> - int hdrlen2, rtsthres = ic->ic_rtsthreshold;
> + int qid, hasqos, rtsthres = ic->ic_rtsthreshold;
>
> wh = mtod(m, struct ieee80211_frame *);
> - hdrlen = ieee80211_get_hdrlen(wh);
> type = wh->i_fc[0] & IEEE80211_FC0_TYPE_MASK;
> + subtype = wh->i_fc[0] & IEEE80211_FC0_SUBTYPE_MASK;
> + if (type == IEEE80211_FC0_TYPE_CTL)
> + hdrlen = sizeof(struct ieee80211_frame_min);
> + else
> + hdrlen = ieee80211_get_hdrlen(wh);
>
> - hdrlen2 = (ieee80211_has_qos(wh)) ?
> - sizeof (struct ieee80211_qosframe) :
> - sizeof (struct ieee80211_frame);
> + hasqos = ieee80211_has_qos(wh);
>
> - tid = 0;
> + if (type == IEEE80211_FC0_TYPE_DATA)
> + tid = IWM_TID_NON_QOS;
> + else
> + tid = IWM_MAX_TID_COUNT;
>
> /*
> * Map EDCA categories to Tx data queues.
> @@ -4918,14 +5526,31 @@ iwm_tx(struct iwm_softc *sc, struct mbuf *m, struct ie
> * need to share Tx queues between stations because we only implement
> * client mode; the firmware's station table contains only one entry
> * which represents our access point.
> - *
> - * Tx aggregation will require additional queues (one queue per TID
> - * for which aggregation is enabled) but we do not implement this yet.
> */
> if (isset(sc->sc_enabled_capa, IWM_UCODE_TLV_CAPA_DQA_SUPPORT))
> - ring = &sc->txq[IWM_DQA_MIN_MGMT_QUEUE + ac];
> + qid = IWM_DQA_MIN_MGMT_QUEUE + ac;
> else
> - ring = &sc->txq[ac];
> + qid = ac;
> +
> + /* If possible, put this frame on an aggregation queue. */
> + if (hasqos) {
> + struct ieee80211_tx_ba *ba;
> + uint16_t qos = ieee80211_get_qos(wh);
> + int qostid = qos & IEEE80211_QOS_TID;
> + int qosac = ieee80211_up_to_ac(ic, qostid);
> +
> + ba = &ni->ni_tx_ba[qostid];
> + if (!IEEE80211_IS_MULTICAST(wh->i_addr1) &&
> + type == IEEE80211_FC0_TYPE_DATA &&
> + (in->ampdu_tx_tid_mask & (1 << qostid)) &&
> + ba->ba_state == IEEE80211_BA_AGREED) {
> + qid = IWM_FIRST_AGG_TX_QUEUE + qostid;
> + tid = qostid;
> + ac = qosac;
> + }
> + }
> +
> + ring = &sc->txq[qid];
> desc = &ring->desc[ring->cur];
> memset(desc, 0, sizeof(*desc));
> data = &ring->data[ring->cur];
> @@ -5004,14 +5629,28 @@ iwm_tx(struct iwm_softc *sc, struct mbuf *m, struct ie
> tx->sta_id = IWM_STATION_ID;
>
> if (type == IEEE80211_FC0_TYPE_MGT) {
> - uint8_t subtype = wh->i_fc[0] & IEEE80211_FC0_SUBTYPE_MASK;
> -
> if (subtype == IEEE80211_FC0_SUBTYPE_ASSOC_REQ ||
> subtype == IEEE80211_FC0_SUBTYPE_REASSOC_REQ)
> tx->pm_frame_timeout = htole16(3);
> else
> tx->pm_frame_timeout = htole16(2);
> } else {
> + if (type == IEEE80211_FC0_TYPE_CTL &&
> + subtype == IEEE80211_FC0_SUBTYPE_BAR) {
> + struct ieee80211_frame_min *wbar;
> + uint8_t *frm;
> + uint16_t ctl;
> +
> + flags |= IWM_TX_CMD_FLG_ACK | IWM_TX_CMD_FLG_BAR;
> + tx->data_retry_limit = IWM_BAR_DFAULT_RETRY_LIMIT;
> +
> + wbar = mtod(m, struct ieee80211_frame_min *);
> + frm = (uint8_t *)&wbar[1];
> + memcpy(&ctl, frm, sizeof(ctl));
> + tid = (le16toh(ctl) & IEEE80211_BA_TID_INFO_MASK) >>
> + IEEE80211_BA_TID_INFO_SHIFT;
> + }
> +
> tx->pm_frame_timeout = htole16(0);
> }
>
> @@ -5058,7 +5697,9 @@ iwm_tx(struct iwm_softc *sc, struct mbuf *m, struct ie
> tx->sec_ctl = 0;
> }
>
> - flags |= IWM_TX_CMD_FLG_BT_DIS | IWM_TX_CMD_FLG_SEQ_CTL;
> + flags |= IWM_TX_CMD_FLG_BT_DIS;
> + if (!hasqos)
> + flags |= IWM_TX_CMD_FLG_SEQ_CTL;
>
> tx->tx_flags |= htole32(flags);
>
> @@ -5085,9 +5726,11 @@ iwm_tx(struct iwm_softc *sc, struct mbuf *m, struct ie
> }
> }
> data->m = m;
> + data->totlen = totlen;
> data->in = in;
> data->txmcs = ni->ni_txmcs;
> data->txrate = ni->ni_txrate;
> + data->ampdu_txmcs = ni->ni_txmcs; /* updated upon Tx interrupt */
>
> /* Fill TX descriptor. */
> desc->num_tbs = 2 + data->map->dm_nsegs;
> @@ -5118,9 +5761,7 @@ iwm_tx(struct iwm_softc *sc, struct mbuf *m, struct ie
> (char *)(void *)desc - (char *)(void *)ring->desc_dma.vaddr,
> sizeof (*desc), BUS_DMASYNC_PREWRITE);
>
> -#if 0
> iwm_update_sched(sc, ring->qid, ring->cur, tx->sta_id,
> le16toh(tx->len));
> -#endif
>
> /* Kick TX ring. */
> ring->cur = (ring->cur + 1) % IWM_TX_RING_COUNT;
> @@ -5336,6 +5977,7 @@ iwm_add_sta_cmd(struct iwm_softc *sc, struct iwm_node
> uint32_t status;
> size_t cmdsize;
> struct ieee80211com *ic = &sc->sc_ic;
> + uint16_t tid_disable_tx = 0xffff;
>
> if (!update && (sc->sc_flags & IWM_FLAG_STA_ACTIVE))
> panic("STA already added");
> @@ -5362,7 +6004,7 @@ iwm_add_sta_cmd(struct iwm_softc *sc, struct iwm_node
> else
> qid = IWM_AUX_QUEUE;
> add_sta_cmd.tfd_queue_msk |= htole32(1 << qid);
> - } else if (!update) {
> + } else {
> int ac;
> for (ac = 0; ac < EDCA_NUM_AC; ac++) {
> int qid = ac;
> @@ -5371,15 +6013,33 @@ iwm_add_sta_cmd(struct iwm_softc *sc, struct iwm_node
> qid += IWM_DQA_MIN_MGMT_QUEUE;
> add_sta_cmd.tfd_queue_msk |= htole32(1 << qid);
> }
> - IEEE80211_ADDR_COPY(&add_sta_cmd.addr, in->in_ni.ni_bssid);
> }
> + if (!update) {
> + if (ic->ic_opmode == IEEE80211_M_MONITOR)
> + IEEE80211_ADDR_COPY(&add_sta_cmd.addr,
> + etherbroadcastaddr);
> + else
> + IEEE80211_ADDR_COPY(&add_sta_cmd.addr,
> + in->in_ni.ni_bssid);
> + }
> add_sta_cmd.add_modify = update ? 1 : 0;
> add_sta_cmd.station_flags_msk
> |= htole32(IWM_STA_FLG_FAT_EN_MSK | IWM_STA_FLG_MIMO_EN_MSK);
> - add_sta_cmd.tid_disable_tx = htole16(0xffff);
> - if (update)
> - add_sta_cmd.modify_mask |= (IWM_STA_MODIFY_TID_DISABLE_TX);
> + if (update) {
> + int tid, qid;
> + for (tid = 0; tid < IWM_MAX_TID_COUNT; tid++) {
> + if ((in->ampdu_tx_tid_mask & (1 << tid)) == 0)
> + continue;
>
> + qid = IWM_FIRST_AGG_TX_QUEUE + tid;
> + add_sta_cmd.tfd_queue_msk |= htole32(1 << qid);
> + tid_disable_tx &= ~(1 << tid);
> + add_sta_cmd.modify_mask |= (IWM_STA_MODIFY_QUEUES |
> + IWM_STA_MODIFY_TID_DISABLE_TX);
> + }
> + }
> + add_sta_cmd.tid_disable_tx = htole16(tid_disable_tx);
> +
> if (in->in_ni.ni_flags & IEEE80211_NODE_HT) {
> add_sta_cmd.station_flags_msk
> |= htole32(IWM_STA_FLG_MAX_AGG_SIZE_MSK |
> @@ -5444,7 +6104,7 @@ iwm_add_aux_sta(struct iwm_softc *sc)
> if (isset(sc->sc_enabled_capa, IWM_UCODE_TLV_CAPA_DQA_SUPPORT)) {
> qid = IWM_DQA_AUX_QUEUE;
> err = iwm_enable_txq(sc, IWM_AUX_STA_ID, qid,
> - IWM_TX_FIFO_MCAST);
> + IWM_TX_FIFO_MCAST, 0, IWM_MAX_TID_COUNT, 0);
> } else {
> qid = IWM_AUX_QUEUE;
> err = iwm_enable_ac_txq(sc, qid, IWM_TX_FIFO_MCAST);
> @@ -6582,6 +7242,9 @@ iwm_auth(struct iwm_softc *sc)
>
> splassert(IPL_NET);
>
> + in->ampdu_rx_tid_mask = 0;
> + in->ampdu_tx_tid_mask = 0;
> +
> if (ic->ic_opmode == IEEE80211_M_MONITOR)
> sc->sc_phyctxt[0].channel = ic->ic_ibss_chan;
> else
> @@ -7156,11 +7819,7 @@ iwm_setrates(struct iwm_node *in, int async)
>
> lqcmd.agg_time_limit = htole16(4000); /* 4ms */
> lqcmd.agg_disable_start_th = 3;
> -#ifdef notyet
> lqcmd.agg_frame_cnt_limit = 0x3f;
> -#else
> - lqcmd.agg_frame_cnt_limit = 1; /* tx agg disabled */
> -#endif
>
> cmd.data[0] = &lqcmd;
> iwm_send_cmd(sc, &cmd);
> @@ -7892,7 +8551,7 @@ iwm_init_hw(struct iwm_softc *sc)
> else
> qid = IWM_AUX_QUEUE;
> err = iwm_enable_txq(sc, IWM_MONITOR_STA_ID, qid,
> - iwm_ac_to_tx_fifo[EDCA_AC_BE]);
> + iwm_ac_to_tx_fifo[EDCA_AC_BE], 0, IWM_MAX_TID_COUNT, 0);
> if (err) {
> printf("%s: could not enable monitor inject Tx queue "
> "(error %d)\n", DEVNAME(sc), err);
> @@ -7906,7 +8565,7 @@ iwm_init_hw(struct iwm_softc *sc)
> else
> qid = ac;
> err = iwm_enable_txq(sc, IWM_STATION_ID, qid,
> - iwm_ac_to_tx_fifo[ac]);
> + iwm_ac_to_tx_fifo[ac], 0, IWM_TID_NON_QOS, 0);
> if (err) {
> printf("%s: could not enable Tx queue %d "
> "(error %d)\n", DEVNAME(sc), ac, err);
> @@ -8578,6 +9237,10 @@ iwm_rx_pkt(struct iwm_softc *sc, struct iwm_rx_data *d
> iwm_rx_tx_cmd(sc, pkt, data);
> break;
>
> + case IWM_BA_NOTIF:
> + iwm_rx_ba(sc, pkt, data);
> + break;
> +
> case IWM_MISSED_BEACONS_NOTIFICATION:
> iwm_rx_bmiss(sc, pkt, data);
> break;
> @@ -8943,9 +9606,9 @@ iwm_intr(void *arg)
> DPRINTF(("driver status:\n"));
> for (i = 0; i < IWM_MAX_QUEUES; i++) {
> struct iwm_tx_ring *ring = &sc->txq[i];
> - DPRINTF((" tx ring %2d: qid=%-2d cur=%-3d "
> + DPRINTF((" tx ring %2d: qid=%-2d tail=%-3d cur=%-3d "
> "queued=%-3d\n",
> - i, ring->qid, ring->cur, ring->queued));
> + i, ring->qid, ring->tail, ring->cur, ring->queued));
> }
> DPRINTF((" rx ring: cur=%d\n", sc->rxq.cur));
> DPRINTF((" 802.11 state %s\n",
> @@ -9053,9 +9716,9 @@ iwm_intr_msix(void *arg)
> DPRINTF(("driver status:\n"));
> for (i = 0; i < IWM_MAX_QUEUES; i++) {
> struct iwm_tx_ring *ring = &sc->txq[i];
> - DPRINTF((" tx ring %2d: qid=%-2d cur=%-3d "
> + DPRINTF((" tx ring %2d: qid=%-2d tail=%-3d cur=%-3d "
> "queued=%-3d\n",
> - i, ring->qid, ring->cur, ring->queued));
> + i, ring->qid, ring->tail, ring->cur, ring->queued));
> }
> DPRINTF((" rx ring: cur=%d\n", sc->rxq.cur));
> DPRINTF((" 802.11 state %s\n",
> @@ -9465,6 +10128,7 @@ iwm_attach(struct device *parent, struct device *self,
>
> /* Set device capabilities. */
> ic->ic_caps =
> + IEEE80211_C_QOS | IEEE80211_C_TX_AMPDU | /* A-MPDU */
> IEEE80211_C_WEP | /* WEP */
> IEEE80211_C_RSN | /* WPA/RSN */
> IEEE80211_C_SCANALL | /* device scans all channels at once */
> @@ -9529,10 +10193,8 @@ iwm_attach(struct device *parent, struct device *self,
> ic->ic_update_htprot = iwm_update_htprot;
> ic->ic_ampdu_rx_start = iwm_ampdu_rx_start;
> ic->ic_ampdu_rx_stop = iwm_ampdu_rx_stop;
> -#ifdef notyet
> ic->ic_ampdu_tx_start = iwm_ampdu_tx_start;
> ic->ic_ampdu_tx_stop = iwm_ampdu_tx_stop;
> -#endif
> /*
> * We cannot read the MAC address without loading the
> * firmware from disk. Postpone until mountroot is done.
> blob - 201ce69014b9422335a6d698cd4a3cc3f314b2b5
> blob + 1e2e4e01e2a98f60221b72fc6e82a1246f7b9cef
> --- sys/dev/pci/if_iwmreg.h
> +++ sys/dev/pci/if_iwmreg.h
> @@ -1837,6 +1837,9 @@ struct iwm_agn_scd_bc_tbl {
> uint16_t tfd_offset[IWM_TFD_QUEUE_BC_SIZE];
> } __packed;
>
> +#define IWM_TX_CRC_SIZE 4
> +#define IWM_TX_DELIMITER_SIZE 4
> +
> /* Maximum number of Tx queues. */
> #define IWM_MAX_QUEUES 31
>
> @@ -1875,6 +1878,11 @@ struct iwm_agn_scd_bc_tbl {
> #define IWM_DQA_MIN_DATA_QUEUE 10
> #define IWM_DQA_MAX_DATA_QUEUE 31
>
> +/* Reserve 8 DQA Tx queues, from 10 up to 17, for A-MPDU aggregation. */
> +#define IWM_MAX_TID_COUNT 8
> +#define IWM_FIRST_AGG_TX_QUEUE IWM_DQA_MIN_DATA_QUEUE
> +#define IWM_LAST_AGG_TX_QUEUE (IWM_FIRST_AGG_TX_QUEUE +
> IWM_MAX_TID_COUNT - 1)
> +
> /* legacy non-DQA queues; the legacy command queue uses a different number!
> */
> #define IWM_OFFCHANNEL_QUEUE 8
> #define IWM_CMD_QUEUE 9
> @@ -4627,7 +4635,8 @@ struct iwm_lq_cmd {
> /*
> * TID for non QoS frames - to be written in tid_tspec
> */
> -#define IWM_TID_NON_QOS IWM_MAX_TID_COUNT
> +#define IWM_TID_NON_QOS 0
> +#define IWM_TID_MGMT 15
>
> /*
> * Limits on the retransmissions - to be written in {data,rts}_retry_limit
> @@ -4898,21 +4907,23 @@ struct iwm_tx_resp {
> /**
> * struct iwm_ba_notif - notifies about reception of BA
> * ( IWM_BA_NOTIF = 0xc5 )
> - * @sta_addr_lo32: lower 32 bits of the MAC address
> - * @sta_addr_hi16: upper 16 bits of the MAC address
> + * @sta_addr: MAC address
> * @sta_id: Index of recipient (BA-sending) station in fw's station table
> * @tid: tid of the session
> - * @seq_ctl: sequence control field from IEEE80211 frame header (it is
> unclear
> - * which frame this relates to; info or reverse engineering welcome)
> + * @seq_ctl: sequence control field from IEEE80211 frame header (the first
> + * bit in @bitmap corresponds to the sequence number stored here)
> * @bitmap: the bitmap of the BA notification as seen in the air
> * @scd_flow: the tx queue this BA relates to
> * @scd_ssn: the index of the last contiguously sent packet
> * @txed: number of Txed frames in this batch
> * @txed_2_done: number of Acked frames in this batch
> + * @reduced_txp: power reduced according to TPC. This is the actual value and
> + * not a copy from the LQ command. Thus, if not the first rate was used
> + * for Tx-ing then this value will be set to 0 by FW.
> + * @reserved1: reserved
> */
> struct iwm_ba_notif {
> - uint32_t sta_addr_lo32;
> - uint16_t sta_addr_hi16;
> + uint8_t sta_addr[ETHER_ADDR_LEN];
> uint16_t reserved;
>
> uint8_t sta_id;
> @@ -4923,6 +4934,7 @@ struct iwm_ba_notif {
> uint16_t scd_ssn;
> uint8_t txed;
> uint8_t txed_2_done;
> + uint8_t reduced_txp;
> uint16_t reserved1;
> } __packed;
>
> blob - 89abe2c1dbdf5ac3ccbf710994380502530ef2a8
> blob + 7d9e26bffe0f1658c771bf85768797c23e94e147
> --- sys/dev/pci/if_iwmvar.h
> +++ sys/dev/pci/if_iwmvar.h
> @@ -252,14 +252,26 @@ struct iwm_fw_paging {
> #define IWM_TX_RING_LOMARK 192
> #define IWM_TX_RING_HIMARK 224
>
> +/* For aggregation queues, index must be aligned to frame sequence number. */
> +#define IWM_AGG_SSN_TO_TXQ_IDX(x) ((x) & (IWM_TX_RING_COUNT - 1))
> +
> struct iwm_tx_data {
> bus_dmamap_t map;
> bus_addr_t cmd_paddr;
> bus_addr_t scratch_paddr;
> struct mbuf *m;
> struct iwm_node *in;
> + int totlen;
> + int retries;
> + int txfail;
> int txmcs;
> int txrate;
> +
> + /* A-MPDU subframes */
> + int ampdu_id;
> + int ampdu_txmcs;
> + int ampdu_nframes;
> + int ampdu_size;
> };
>
> struct iwm_tx_ring {
> @@ -363,6 +375,12 @@ struct iwm_bf_data {
> int last_cqm_event;
> };
>
> +struct iwm_ba_param {
> + uint16_t tid_mask;
> + uint16_t ssn[IWM_MAX_TID_COUNT];
> + uint16_t winsize[IWM_MAX_TID_COUNT];
> +};
> +
> struct iwm_softc {
> struct device sc_dev;
> struct ieee80211com sc_ic;
> @@ -381,10 +399,15 @@ struct iwm_softc {
>
> /* Task for firmware BlockAck setup/teardown and its arguments. */
> struct task ba_task;
> - int ba_start;
> - int ba_tid;
> - uint16_t ba_ssn;
> - uint16_t ba_winsize;
> + int ba_flags;
> +#define IWM_RX_BA_START 0x01
> +#define IWM_TX_BA_START 0x02
> +#define IWM_RX_BA_STOP 0x04
> +#define IWM_TX_BA_STOP 0x08
> + struct iwm_ba_param rx_ba_start;
> + struct iwm_ba_param rx_ba_stop;
> + struct iwm_ba_param tx_ba_start;
> + struct iwm_ba_param tx_ba_stop;
>
> /* Task for HT protection updates. */
> struct task htprot_task;
> @@ -407,6 +430,7 @@ struct iwm_softc {
> struct iwm_rx_ring rxq;
> int qfullmsk;
> int cmdqid;
> + int qenablemsk;
>
> int sc_sf_state;
>
> @@ -551,6 +575,12 @@ struct iwm_node {
> int chosen_txrate;
> struct ieee80211_mira_node in_mn;
> int chosen_txmcs;
> +
> + uint32_t next_ampdu_id;
> +
> + /* Currently active Rx/Tx block ack sessions; tracked per TID. */
> + uint8_t ampdu_rx_tid_mask;
> + uint8_t ampdu_tx_tid_mask;
> };
> #define IWM_STATION_ID 0
> #define IWM_AUX_STA_ID 1
>
--