2022-12-10 22:43:44 -06:00
* Overview
Setting up 4 simulation environments is in order:
| Plan Size | Characteristics |
|-----------+-----------------|
| Small | 1024 users |
| Medium | 10000 users |
| Large | 40000 users |
| Humongous | 100k users |
** IP address ranges
2001:0002::/48 and 198.18.0.0/15 are the experimental address ranges.
For testing outside of the network, perhaps a BGP AS for this would help
** DNS service
nsupdate for reverse dns? Hosts file?
* Subscriber Plans
2022-12-11 10:09:00 -06:00
| Plan Type | Down | Up | MinDown | MinUP | Cap | | | |
|------------+---------+---------+---------+---------+-----+---+---+---|
| OldDSL | 4Mbit | 384kbit | 1Mbit | | | | | |
| OldDSL6 | 6Mbit | 1Mbit | 3Mbit | | | | | |
| Cable20 | 20Mbit | 4Mbit | 12mbit | | | | | |
| LifeLine | 25Mbit | 3Mbit | | | | | | |
| FCCmin | 25Mbit | 3Mbit | | | | | | |
| NTIAserved | 100MBit | 20Mbit | 20Mbit | 5Mbit | | | | |
| Biz50 | 50MBit | 20Mbit | 50Mbit | 20Mbit | | | | |
| Biz100 | 100MBit | 100Mbit | 100Mbit | 100Mbit | | | | |
| Biz200 | 200Mbit | 20Mbit | 200Mbit | 20Mbit | | | | |
| Biz500 | 500Mbit | 50Mbit | | | | | | |
| Sym1000 | 1Gbit | 1Gbit | | | | | | |
| Sym100 | 100Mbit | 100Mbit | | | | | | |
| Sym10 | 10Mbit | 10Mbit | | | | | | |
2022-12-10 22:43:44 -06:00
* Topologies
A variety of topologies need to be emulated.
** Distance
** Hops
** Different technologies
* Routing Protocols
Both ISIS and OSPF are in use
* Underlying framing
PPPoe, MPLS are in use
CGNATs might become an issue
IPv6 deploying?
2022-12-11 10:09:00 -06:00
* Tries
Some customers (particularly [[https://seclists.org/nanog/2022/Nov/102 ][business ones ]]) have multiple IP addresses.
Everyone, using IPv6, has multiple IP addresses.
Worse, these are often dynamic in origin.
2022-12-10 22:43:44 -06:00
* Transport Technologies
| Plan | Down | Up | | | | | | |
|----------------------+------+-----+---+---+---+---+---+---|
| WiFi | | | | | | | | |
| Cable | 2ms | 6ms | | | | | | |
| Docsis LL | 1ms | | | | | | | |
| Wireless half duplex | | | | | | | | |
* Test tools
** Flent
2022-12-11 11:05:09 -06:00
** TRex
2022-12-10 22:43:44 -06:00
** goresponsiveness
** crusader
** speedtest.net
2022-12-11 11:05:09 -06:00
** Web PLT
Alexa top 10000
2022-12-10 22:43:44 -06:00
** samknows
** Netflix video quality
** VOIP MOS
** Videoconferencing Quality
* Congestion Controls
2022-12-11 10:14:18 -06:00
| Control | Why | ecn | | |
|----------+------------+-----+---+---|
| cubic | Linux | S | | |
| reno | BSD | S | | |
| ledbat | Torrent | N | | |
| ledbat++ | Microsoft | Y | | |
| curved | Microsoft | S | | |
| bbr | Google | N | | |
| bbrv2 | Google exp | L | | |
| prague | IETF exp | L | | |
2022-12-10 22:43:44 -06:00
* Diffserv Markings
EF,LE,CS1, CS5,NQB,CS6
* Failure Modes
2022-12-10 23:04:25 -06:00
** AQM
fq_codel was not designed to be a middlebox. The FQ component works well,
the AQM component, doesn't work as well as it could.
[[./img/middlebox_compared.png ]]
2022-12-10 23:37:23 -06:00
Putting cake on the CPE also results in a metric ton less jitter.
[[./img/tonlessjitter.png ]]
2022-12-10 23:04:25 -06:00
Similarly, many devices are themselves the bottleneck, still, so they
accumulate a ton of RTT themselves, and monitoring the RTT and doing
something about it would possibly help.
2022-12-10 23:19:55 -06:00
Validating the the RTT metrics reported by pping line up with the
actual measurements from actual flows is important. Also, what are
the effects of ack-filtering on pping?
2022-12-10 23:04:25 -06:00
** FQ
2022-12-10 23:19:55 -06:00
The FQ methods we use are really good for most traffic types, and
could be even better if more applications did single packet pacing and
were more sensitive to delay and jitter.
2022-12-10 23:04:25 -06:00
** Encapsulations
2022-12-10 23:19:55 -06:00
We have no insight into QUIC or VPN traffic. This is going to get
worse over time. The only thing we have for quic is the: [[https://www.ietfjournal.org/enabling-internet-measurement-with-the-quic-spin-bit/ ][spin bit ]] -
which is probably [[https://http3-explained.haxx.se/en/quic/quic-spinbit ][not widely implemented ]]. The best insight we actually
have is queue accomulation and packet drop/mark behaviors.
2022-12-10 23:04:25 -06:00
** HTB is bursty
2022-12-10 23:19:55 -06:00
** DSCP
What DSCPs are in common use today?
** DROP_MONITOR
2022-12-10 23:37:23 -06:00
We have 2600 (not kidding, 2600) places where packets can be dropped
in the kernel, not to mention other parts of the network. A nice
feature would be to be able to track retransmits relative to drops.