diff options
author | Eric Dumazet <edumazet@google.com> | 2024-04-23 15:56:20 +0300 |
---|---|---|
committer | Jakub Kicinski <kuba@kernel.org> | 2024-04-25 22:15:02 +0300 |
commit | ec00ed472bdb7d0af840da68c8c11bff9f4d9caa (patch) | |
tree | 7b9bb64d8735b8f6f94c6b211959765da79f2f46 /include/linux/pci_ids.h | |
parent | e6be197f23c5cf15fb4a5acac213b77f80e8cf96 (diff) | |
download | linux-ec00ed472bdb7d0af840da68c8c11bff9f4d9caa.tar.xz |
tcp: avoid premature drops in tcp_add_backlog()
While testing TCP performance with latest trees,
I saw suspect SOCKET_BACKLOG drops.
tcp_add_backlog() computes its limit with :
limit = (u32)READ_ONCE(sk->sk_rcvbuf) +
(u32)(READ_ONCE(sk->sk_sndbuf) >> 1);
limit += 64 * 1024;
This does not take into account that sk->sk_backlog.len
is reset only at the very end of __release_sock().
Both sk->sk_backlog.len and sk->sk_rmem_alloc could reach
sk_rcvbuf in normal conditions.
We should double sk->sk_rcvbuf contribution in the formula
to absorb bubbles in the backlog, which happen more often
for very fast flows.
This change maintains decent protection against abuses.
Fixes: c377411f2494 ("net: sk_add_backlog() take rmem_alloc into account")
Signed-off-by: Eric Dumazet <edumazet@google.com>
Link: https://lore.kernel.org/r/20240423125620.3309458-1-edumazet@google.com
Signed-off-by: Jakub Kicinski <kuba@kernel.org>
Diffstat (limited to 'include/linux/pci_ids.h')
0 files changed, 0 insertions, 0 deletions