aboutgitcodebugslistschat
diff options
context:
space:
mode:
authorJon Maloy <jmaloy@redhat.com>2026-04-25 15:58:18 -0400
committerStefano Brivio <sbrivio@redhat.com>2026-05-07 08:03:14 +0200
commita458719f01838f0d5867e817d84a497637999ee1 (patch)
treec4a406dca11a06e0c5ccc1587817af3ec16a45af
parentec96f0124282338cd2b2e65ff1aa3def8882ae23 (diff)
downloadpasst-a458719f01838f0d5867e817d84a497637999ee1.tar
passt-a458719f01838f0d5867e817d84a497637999ee1.tar.gz
passt-a458719f01838f0d5867e817d84a497637999ee1.tar.bz2
passt-a458719f01838f0d5867e817d84a497637999ee1.tar.lz
passt-a458719f01838f0d5867e817d84a497637999ee1.tar.xz
passt-a458719f01838f0d5867e817d84a497637999ee1.tar.zst
passt-a458719f01838f0d5867e817d84a497637999ee1.zip
tcp: Use SO_MEMINFO for accurate send buffer overhead accounting
The TCP window advertised to the guest/container must balance two competing needs: large enough to trigger kernel socket buffer auto-tuning, but not so large that sendmsg() partially fails, causing retransmissions. The current approach uses the difference (SNDBUF_GET() - SIOCOUTQ), but SNDBUF_GET() returns a scaled value that only roughly accounts for per-skb overhead. The clamped_scale approximation doesn't accurately track the actual per-segment overhead, which can lead to both excessive retransmissions and reduced throughput. We now introduce the use of SO_MEMINFO to obtain SK_MEMINFO_SNDBUF and SK_MEMINFO_WMEM_QUEUED from the kernel. The latter is presented in the kernel's own accounting units, i.e. including the sk_buff overhead, and matches exactly what the kernel's own sk_stream_memory_free() function is using. When data is queued and the overhead ratio is observable, we calculate the per-segment overhead as (wmem_queued - sendq) / num_segments, then determine how many additional segments should fit in the remaining buffer space, considering the calculated per-mss overhead. This approach treats segments as discrete quantities, and produces a more accurate estimate of available buffer space than a linear scaling factor does. When the ratio cannot be observed, e.g. because the queue is empty or we are in a transient state, we fall back to the existing clamped_scale calculation (scaling between 100% and 75% of buffer capacity). When SO_MEMINFO succeeds, we also use SK_MEMINFO_SNDBUF directly to set SNDBUF, avoiding a separate SO_SNDBUF getsockopt() call. If SO_MEMINFO is unavailable, we fall back to the pre-existing SNDBUF_GET() - SIOCOUTQ calculation. Link: https://bugs.passt.top/show_bug.cgi?id=138 Link: https://github.com/containers/podman/issues/28219 Analysed-by: Yumei Huang <yuhuang@redhat.com> Signed-off-by: Jon Maloy <jmaloy@redhat.com> Signed-off-by: Stefano Brivio <sbrivio@redhat.com>
-rw-r--r--tcp.c137
-rw-r--r--tcp_conn.h2
2 files changed, 89 insertions, 50 deletions
diff --git a/tcp.c b/tcp.c
index a481bd3..92d9797 100644
--- a/tcp.c
+++ b/tcp.c
@@ -295,6 +295,7 @@
#include <arpa/inet.h>
#include <linux/sockios.h>
+#include <linux/sock_diag.h>
#include "checksum.h"
#include "util.h"
@@ -1018,6 +1019,90 @@ size_t tcp_fill_headers(const struct ctx *c, struct tcp_tap_conn *conn,
}
/**
+ * tcp_wnd_from_sndbuf() - Calculate window from available send buffer space
+ * @s: Socket file descriptor
+ * @conn: Connection pointer
+ * @tinfo: tcp_info from kernel
+ *
+ * Return: window value to advertise, not scaled
+ */
+static uint32_t tcp_wnd_from_sndbuf(int s, struct tcp_tap_conn *conn,
+ const struct tcp_info_linux *tinfo)
+{
+ uint32_t rtt_ms_ceiling = DIV_ROUND_UP(tinfo->tcpi_rtt, 1000);
+ uint32_t mem[SK_MEMINFO_VARS];
+ socklen_t mem_sl = sizeof(mem);
+ int mss = MSS_GET(conn);
+ uint32_t limit, sendq;
+
+ if (ioctl(s, SIOCOUTQ, &sendq)) {
+ debug_perror("SIOCOUTQ on socket %i, assuming 0", s);
+ sendq = 0;
+ }
+
+ if (getsockopt(s, SOL_SOCKET, SO_MEMINFO, &mem, &mem_sl)) {
+ tcp_get_sndbuf(conn);
+
+ if (sendq > SNDBUF_GET(conn)) /* Due to memory pressure? */
+ limit = 0;
+ else
+ limit = SNDBUF_GET(conn) - sendq;
+ } else {
+ uint32_t sndbuf = mem[SK_MEMINFO_SNDBUF];
+ uint32_t wmemq = mem[SK_MEMINFO_WMEM_QUEUED];
+ uint32_t scaled = clamped_scale(sndbuf, sndbuf, SNDBUF_SMALL,
+ SNDBUF_BIG, 75);
+
+ SNDBUF_SET(conn, MIN(INT_MAX, scaled));
+
+ if (wmemq > sndbuf) {
+ limit = 0;
+ } else if (!sendq || !mss || wmemq <= sendq) {
+ limit = SNDBUF_GET(conn) - wmemq;
+ } else {
+ uint32_t used_segs = MAX(sendq / mss, 1);
+ uint32_t overhead = (wmemq - sendq) / used_segs;
+ uint32_t remaining = sndbuf - wmemq;
+ uint32_t avail_segs = remaining / (mss + overhead);
+
+ limit = avail_segs * mss;
+ }
+ }
+
+ /* If the sender uses mechanisms to prevent Silly Window
+ * Syndrome (SWS, described in RFC 813 Section 3) it's critical
+ * that, should the window ever become less than the MSS, we
+ * advertise a new value once it increases again to be above it.
+ *
+ * The mechanism to avoid SWS in the kernel is, implicitly,
+ * implemented by Nagle's algorithm (which was proposed after
+ * RFC 813).
+ *
+ * To this end, for simplicity, approximate a window value below
+ * the MSS to zero, as we already have mechanisms in place to
+ * force updates after the window becomes zero. This matches the
+ * suggestion from RFC 813, Section 4.
+ *
+ * But don't do this if, either:
+ *
+ * - there's nothing in the outbound queue: the size of the
+ * sending buffer is limiting us, and it won't increase if we
+ * don't send data, so there's no point in waiting, or
+ *
+ * - we haven't sent data in a while (somewhat arbitrarily, ten
+ * times the RTT), as that might indicate that the receiver
+ * will only process data in batches that are large enough,
+ * but we won't send enough to fill one because we're stuck
+ * with pending data in the outbound queue
+ */
+ if (limit < (uint32_t)MSS_GET(conn) && sendq &&
+ tinfo->tcpi_last_data_sent < rtt_ms_ceiling * 10)
+ limit = 0;
+
+ return MIN(tinfo->tcpi_snd_wnd, limit);
+}
+
+/**
* tcp_update_seqack_wnd() - Update ACK sequence and window to guest/tap
* @c: Execution context
* @conn: Connection pointer
@@ -1124,56 +1209,10 @@ int tcp_update_seqack_wnd(const struct ctx *c, struct tcp_tap_conn *conn,
}
}
- if ((conn->flags & LOCAL) || tcp_rtt_dst_low(conn)) {
+ if ((conn->flags & LOCAL) || tcp_rtt_dst_low(conn))
new_wnd_to_tap = tinfo->tcpi_snd_wnd;
- } else {
- unsigned rtt_ms_ceiling = DIV_ROUND_UP(tinfo->tcpi_rtt, 1000);
- uint32_t sendq;
- int limit;
-
- if (ioctl(s, SIOCOUTQ, &sendq)) {
- debug_perror("SIOCOUTQ on socket %i, assuming 0", s);
- sendq = 0;
- }
- tcp_get_sndbuf(conn);
-
- if ((int)sendq > SNDBUF_GET(conn)) /* Due to memory pressure? */
- limit = 0;
- else
- limit = SNDBUF_GET(conn) - (int)sendq;
-
- /* If the sender uses mechanisms to prevent Silly Window
- * Syndrome (SWS, described in RFC 813 Section 3) it's critical
- * that, should the window ever become less than the MSS, we
- * advertise a new value once it increases again to be above it.
- *
- * The mechanism to avoid SWS in the kernel is, implicitly,
- * implemented by Nagle's algorithm (which was proposed after
- * RFC 813).
- *
- * To this end, for simplicity, approximate a window value below
- * the MSS to zero, as we already have mechanisms in place to
- * force updates after the window becomes zero. This matches the
- * suggestion from RFC 813, Section 4.
- *
- * But don't do this if, either:
- *
- * - there's nothing in the outbound queue: the size of the
- * sending buffer is limiting us, and it won't increase if we
- * don't send data, so there's no point in waiting, or
- *
- * - we haven't sent data in a while (somewhat arbitrarily, ten
- * times the RTT), as that might indicate that the receiver
- * will only process data in batches that are large enough,
- * but we won't send enough to fill one because we're stuck
- * with pending data in the outbound queue
- */
- if (limit < MSS_GET(conn) && sendq &&
- tinfo->tcpi_last_data_sent < rtt_ms_ceiling * 10)
- limit = 0;
-
- new_wnd_to_tap = MIN((int)tinfo->tcpi_snd_wnd, limit);
- }
+ else
+ new_wnd_to_tap = tcp_wnd_from_sndbuf(s, conn, tinfo);
new_wnd_to_tap = MIN(new_wnd_to_tap, MAX_WINDOW);
if (!(conn->events & ESTABLISHED))
diff --git a/tcp_conn.h b/tcp_conn.h
index 6985426..9f5bee0 100644
--- a/tcp_conn.h
+++ b/tcp_conn.h
@@ -98,7 +98,7 @@ struct tcp_tap_conn {
#define SNDBUF_BITS 24
unsigned int sndbuf :SNDBUF_BITS;
#define SNDBUF_SET(conn, bytes) (conn->sndbuf = ((bytes) >> (32 - SNDBUF_BITS)))
-#define SNDBUF_GET(conn) (conn->sndbuf << (32 - SNDBUF_BITS))
+#define SNDBUF_GET(conn) ((uint32_t)(conn->sndbuf << (32 - SNDBUF_BITS)))
uint8_t seq_dup_ack_approx;