APNIC Pty Ltd.

31/07/2024 | Press release | Distributed by Public on 31/07/2024 01:51

Bytes from IETF 120 — BBR 1, 2, 3

During the recent IETF 120 meeting it was suggested that we made a mistake by calling the end-to-end transport flow control algorithms 'congestion control', as this term has negative connotations about the network and the quality of the user experience.

If we had called this function something like 'performance optimization,' the focus might have been different. Instead of concentrating on avoiding network congestion and packet loss (which are intrinsic to most packet networks), we would have looked at how to interpret these signals to optimize the algorithm's adaptation to the network path.

On this topic of performance optimization, I'd like to discuss one particular presentation from the ACM/IRTF Applied Networking Research Workshop held at IETF 120, 'BBRv3 in the public Internet: a boon or a bane? '

Some time back, well over a couple of decades ago, there was one dominant TCP control algorithm used within the public Internet, characterized by the generic term Additive Increase, Multiplicative Decrease (AIMD). Each individual TCP flow would gently increase its sending rate in a linear manner (increasing the sending rate by one segment per Round-Trip Time (RTT)), and this rate inflation would continue until either the sender ran out of local buffer space, or one or more of the queues within the network had filled to the point of overflow and packets were dropped.

In response to such packet loss, the TCP flow rate would drop as soon as it was informed of the loss, halving the sending rate, and it would then resume this gentle rate increase. As long as all concurrent TCP sessions behaved in much the same manner, then a shared network path element would be more or less equally used by all these TCP sessions.

With the increasing diversity of network element characteristics in terms of bandwidth, stability, latency, and reliability, this approximate uniformity of TCP behaviour has been challenged. In the myriad of alternate TCP flow control algorithms, two other general approaches have emerged into the mainstream of adoption.

The first enlists the assistance of the network to inform the TCP session of the onset of network load before packet loss, preventing the loss of TCP's control signal. This Explicit Congestion Notification (ECN) is a marker placed on an IP packet when the network element is experiencing internal queue formation. The TCP packet receiver mirrors this congestion signal back to the sender in its ACK stream and the sender can then respond with a flow modification comparable to detection of packet loss. It allows a sender to perform adaptive rate control without deliberately pushing the network into overload and avoiding the extended hiatus that can be caused by receiver timeouts or inefficiencies due to packet loss. This approach of enlisting the network's cooperation in early congestion signalling lies behind the work on Low Latency, Low Loss, Scalable (L4S) throughput, which seeks to improve the user experience through more responsive TCP flow behaviours.

Another approach that has gathered momentum through widespread adoption is Google's Bottleneck Bandwidth and Round-trip propagation time (BBR) flow control algorithm, which replaces packet loss with round-trip delay inflation as the signal of incipient network congestion and decreases the sensitivity of the session response from a continuous control function to a behaviour modification performed only at periodic intervals.

Various measurement exercises have shown very different TCP responses to different path characteristics. BBR is not that sensitive to packet loss and can perform far better in environments where elevated intermittent bit error rates can cause non-congestion packet loss, often seen in mobile data scenarios, and more recently in Starlink services.

ECN-aware controls are better at avoiding protracted control signal loss and can provide superior overall throughput results when compared to loss-based TCP sessions driven by congestion control protocols such as Reno and CUBIC. These various approaches not only have different behaviours in response to particular network path characteristics but respond differently to each other when sharing a common network path.

The initial version of BBR (v1) was able to exert greater flow pressure on concurrent loss-based TCP sessions (Figure 1).

Given this performance edge, BBR was quickly adopted by many content delivery platforms. By 2019, 22% of Alexa's top 20,000 websites used BBR. At the time, BBR accounted for an estimated 40% or more of the overall traffic volume on the public Internet.

BBR v2 attempted to redress some of this unfairness by reacting to both packet loss and ECN signals by moderating its sending rate in addition to BBR's probing profile. This revision to BBR also adjusted its probe response to be less aggressive in pre-empting network capacity. The result was a BBR behaviour that appeared to coexist on more or less equal terms with CUBIC (Figure 2).

Work has continued on BBR and BBRv3 was released in 2023. This release is a tuning release that was intended to improve BBR's coexistence with CUBIC and Reno. The work presented at this workshop centred on a benchtop test using a bottleneck 100Mbps link with a 10ms delay and passing concurrent TCP sessions across the link. The basic result is that under such idealized conditions, BBRv3 exerts a flow pressure equivalent to 16 concurrent CUBIC sessions, which is much the same as the result for BBRv1. When a single BBRv3 flow shares a link with a single CUBIC session, the BBR session quickly occupies much of the available link capacity (Figure 3).

If ECN signals are added to the network, then the difference in the responses of the two protocols is even greater (Figure 4).

It appears to me that our efforts in trying to optimize TCP flow behaviours by interpreting the extremes of network behaviour (packet loss) and managing the flow parameters based on that interpretation, together with some assumptions about network queuing behaviours, have some fundamental limitations. The coarseness of the TCP response in loss-based algorithms assumes (and even requires) the presence of network buffers to perform a form of behaviour adaptation between the host and the network.

The BBR approach is different in that it periodically provokes the network to react and interprets the network's response in relation to this provocation. This periodic probing has resulted in a control protocol that is evidently more capable of exerting continuous flow pressure on concurrent flows. It points to an observation that network queues can be driven to become a source of delay and inefficiency in terms of data transfer.

The BBR model places more control into the control algorithms that operate on the end systems and tries to optimize their behaviour in a manner that avoids using the network's queuing responses. A reasonable objective of any adaptive performance control protocol is to fill the network's circuits, but at the same time not to fill the network's internal buffers!

The widespread adoption of BBR in the Internet indicates an ongoing decline in loss-based congestion control protocols. I personally suspect that the critical service metrics for these newer adaptive rate control protocols are the extent to which they outperform CUBIC and other loss-based congestion protocols, and how well they can fairly share network resources across concurrent sessions.

The views expressed by the authors of this blog are their own and do not necessarily reflect the views of APNIC. Please note a Code of Conduct applies to this blog.