Rate Pace is a TCP Express™ feature that you should be using. Most TCP profile options have difficult tradeoffs. When we introduced Rate Pace in F5® TMOS® 11.5.0, it was unquestionably better for throughput but had difficult implications for CPU efficiency in some deployments. We've whittled away at those efficiency issues, and today Rate Pace is an essentially free improvement in application throughput.
TCP's various windows sometimes allow senders to transmit a burst of packets all at once. This burst traditionally happens at the line rate of the local interface. If this line rate exceeds the throughput of the bottleneck router, packets will accumulate in the router buffer and even overflow, causing drops. Packet losses then cause TCP to throttle back its bandwidth estimate of the path, with bad implications for performance.
The solution is obvious. TCP should space packet transmissions out in accordance with the bottleneck bandwidth. This allows the bottleneck router to deal with each packet before we send another one.
And indeed, a quick test shows some benefit. The chart below shows throughput of a 50 MB download over a 54Mbps, 10ms Round Trip Time (RTT) WAN with a 64 KB router buffer, for two congestion controls. Rate Pace benefits Highspeed congestion control significantly. Woodside, which is less prone to fill queues to exhaustion, only benefits a bit.
The hard part is figuring out what that bottleneck bandwidth should be. How exactly TMOS does this depends on whether the profile has a setting for rate-pace-max-rate, which we introduced in TMOS 12.0.
If there is no configured maximum rate, rate pacing is not in effect at all until there is a packet loss, which indicates TCP has reached the maximum bandwidth of the path. At that point, TCP limits the overall rate to the congestion window (cwnd) divided by the RTT. Assuming there is data available at the BIG-IP, TCP will send out a full-size (Maximum Segment Size, or MSS) packet every (RTT * MSS / cwnd) milliseconds.
If the profile configures rate-pace-max-rate, we expect that the user has set it to reflect knowledge of bottleneck bandwidth in the path. That rate limit is in effect from the beginning of the connection, even before any packet loss. After a packet loss, it uses the cwnd/rtt calculation above, but the value cannot exceed the maximum rate.
Rate Pacing, in general, is not on by default in our built-in TCP profiles. It's one of several reasons that mptcp-mobile-optimized, which enables it, is generally the highest-performing of our built-in profiles. It's also a reason we're planning a refresh of our built-in TCP profiles. Until then, boost your performance by simply turning this on in the profiles you use.