|
@@ -604,7 +604,7 @@ static inline void __sk_add_backlog(struct sock *sk, struct sk_buff *skb)
|
|
|
}
|
|
|
|
|
|
/* The per-socket spinlock must be held here. */
|
|
|
-static inline int sk_add_backlog(struct sock *sk, struct sk_buff *skb)
|
|
|
+static inline __must_check int sk_add_backlog(struct sock *sk, struct sk_buff *skb)
|
|
|
{
|
|
|
if (sk->sk_backlog.len >= max(sk->sk_backlog.limit, sk->sk_rcvbuf << 1))
|
|
|
return -ENOBUFS;
|