On 10/29, Mina Almasry wrote:
Check we're going to free a reasonable number of frags in token_count before starting the loop, to prevent looping too long.
Also minor code cleanups:
- Flip checks to reduce indentation.
 - Use sizeof(*tokens) everywhere for consistentcy.
 Cc: Yi Lai yi1.lai@linux.intel.com
Signed-off-by: Mina Almasry almasrymina@google.com
net/core/sock.c | 46 ++++++++++++++++++++++++++++------------------ 1 file changed, 28 insertions(+), 18 deletions(-)
diff --git a/net/core/sock.c b/net/core/sock.c index 7f398bd07fb7..8603b8d87f2e 100644 --- a/net/core/sock.c +++ b/net/core/sock.c @@ -1047,11 +1047,12 @@ static int sock_reserve_memory(struct sock *sk, int bytes)
#ifdef CONFIG_PAGE_POOL
-/* This is the number of tokens that the user can SO_DEVMEM_DONTNEED in +/* This is the number of frags that the user can SO_DEVMEM_DONTNEED in
- 1 syscall. The limit exists to limit the amount of memory the kernel
 
- allocates to copy these tokens.
 
- allocates to copy these tokens, and to prevent looping over the frags for
 */
- too long.
 -#define MAX_DONTNEED_TOKENS 128 +#define MAX_DONTNEED_FRAGS 1024
static noinline_for_stack int sock_devmem_dontneed(struct sock *sk, sockptr_t optval, unsigned int optlen) @@ -1059,43 +1060,52 @@ sock_devmem_dontneed(struct sock *sk, sockptr_t optval, unsigned int optlen) unsigned int num_tokens, i, j, k, netmem_num = 0; struct dmabuf_token *tokens; netmem_ref netmems[16];
u64 num_frags = 0; int ret = 0;
if (!sk_is_tcp(sk)) return -EBADF;
- if (optlen % sizeof(struct dmabuf_token) ||
 optlen > sizeof(*tokens) * MAX_DONTNEED_TOKENS)
- if (optlen % sizeof(*tokens) ||
  return -EINVAL;optlen > sizeof(*tokens) * MAX_DONTNEED_FRAGS)
- tokens = kvmalloc_array(optlen, sizeof(*tokens), GFP_KERNEL);
 
- num_tokens = optlen / sizeof(*tokens);
 - tokens = kvmalloc_array(num_tokens, sizeof(*tokens), GFP_KERNEL); if (!tokens) return -ENOMEM;
 
- num_tokens = optlen / sizeof(struct dmabuf_token); if (copy_from_sockptr(tokens, optval, optlen)) { kvfree(tokens); return -EFAULT; }
 
- for (i = 0; i < num_tokens; i++) {
 num_frags += tokens[i].token_count;if (num_frags > MAX_DONTNEED_FRAGS) {kvfree(tokens);return -E2BIG;}- }
 - xa_lock_bh(&sk->sk_user_frags); for (i = 0; i < num_tokens; i++) { for (j = 0; j < tokens[i].token_count; j++) { netmem_ref netmem = (__force netmem_ref)__xa_erase( &sk->sk_user_frags, tokens[i].token_start + j);
 
if (netmem &&!WARN_ON_ONCE(!netmem_is_net_iov(netmem))) {netmems[netmem_num++] = netmem;if (netmem_num == ARRAY_SIZE(netmems)) {xa_unlock_bh(&sk->sk_user_frags);for (k = 0; k < netmem_num; k++)WARN_ON_ONCE(!napi_pp_put_page(netmems[k]));netmem_num = 0;xa_lock_bh(&sk->sk_user_frags);}ret++;
[..]
if (!netmem || WARN_ON_ONCE(!netmem_is_net_iov(netmem)))continue;
Any reason we are not returning explicit error to the callers here? That probably needs some mechanism to signal which particular one failed so the users can restart?