GL.iNet GL-MT6000 - AQL and WiFi Latency

I am not having the best of the results after a while. On my MT6000 after hours I get 'persistent Wireless's slowdowns.

Dumping the stats shows this:

cat /sys/kernel/debug/ieee80211/phy*/aql*
AC     AQL pending
VO     0 us
VI     0 us
BE     4960 us.     <---- this has become stuck, does not change after multiple dumps
BK     4 us
BC/MC  0 us
total  4964 us
AC      AQL limit low   AQL limit high
VO      5000            5000
VI      5000            5000
BE      5000            5000
BK      5000            5000
BC/MC   50000

I have a 802.11s BATMAN + AP setup on SNAPSHOT, r26740.

I believe the tight AQL limits in my environment is exposing something.

I will revert to AQL defaults to monitor behaviour.


Only the 2.4GHz radio is affected right?

A few questions for you...

Are you running @pesa1234's build?

What firmware are you running? (dmesg | grep Firmware)

1 Like

Clean Snapshot and it's old MTK firmware, not pesa's build in my case.

[   12.112457] mt798x-wmac 18000000.wifi: WM Firmware Ve
rsion: ____000000, Build Time: 20221012174725
[   12.196474] mt798x-wmac 18000000.wifi: WA Firmware Ve
rsion: DEV_000000, Build Time: 20221012174937

Had to keep the MT6000 with aql_low:2500 & aql_high:12000 for now.

How come the TUF 6000 is performing significantly better? It’s not a secret that I’m a TUF 6000 fan-boy, but now I’m actually curious? Compare the equivalent graph for the MY-6000 and it’s a BIG difference in performance or am I missing something?

I’m trying to follow here, but I’m a bit lost at the moment. Are you referring to this unit?

Secondly, which graphs are you comparing specifically to where you’re seeing the significant performance difference?

1 Like

Gungernuts graph a few posts above

TUF Latency

I think this needs to be revisited. I'm not at all trying to put down the TUF AX6000 or TUF AX4000, but the graph @Gingernut posted doesn't seem to me to indicate a significant performance increase over the MT6000.

If you look at some of the more recent graphs I posted in another MT6000 thread for @pesa1234's build, we have been working hard to tune some AQL settings and achieved some pretty respectable results:

The graph that @Gingernut posted here seems to indicate some significant latency struggles especially in the combined test (third test on the graph). Those latency spikes have largely been smoothed out with tuning in @pesa1234's build.


Thanks for the explanation!

For comparison - my Crusader results using TP-Link EAP615 (ramips/mt7621) with OpenWrt snapshot from Jun23 - 5Ghz@80Mhz

AQL set to 5000/12000

AQL set to 1500/5000

Looks like using 1500/5000 results in latency improvement of >25%

edit: these results are from a test with 1 associated station - later Today I will retest with >3 associated stations

edit2: Retest with 5 associated stations

CH48 80MHz AQL 1500/5000 11:42

CH100 80MHz AQL 1500/5000 11:46

CH100 80MHz AQL 5000/12000 11:49

CH48 80MHz AQL 5000/12000 11:51


@dtaht Any thoughts on the testing that @ed8 did here?

Thought #1 - I should read the OpenWrt forums more often.
Thought #2 - what a difference a clean channel (100?) makes!!, and also how hardware retries muck with things
Thought #3 - I wish crusaders graphs were as directly comparable as flent's. In particular I care that the AQM get tested and this test does not do that. (well, the slope of the crusader staggered start test sort of does). A packet capture can help too, with the RTT plot from Wireshark (tcptrace -G, and an also)

But to me the lowered AQL values seem to indicate goodness across the board.

Regrettably I don't have the mental energy to understand all else that under test? If you care (as I do) that a gaming wifi station do better with multiple stations present, flent's rtt_fair test is good, and anything consistently less than 20ms is good.

1 Like

Using 2500/5000 again now but on updated SNAPSHOT, r26912 for the MT6000. Looks better & stable now and nothing seems to be getting stuck after 16 hours.

1 Like

That's awesome news! Out of curiosity, have you tried running @pesa1234's build?


Yea I've been keeping low the same and messing with larger high settings. Currently doing AQL of 2500/8000 for a little more peak throughput with nice results. Running the latest snapshot on 6.6.38 with some various tweaks.

1 Like