|
@@ -125,7 +125,7 @@ static void handle_tx(struct vhost_net *net)
|
|
mutex_lock(&vq->mutex);
|
|
mutex_lock(&vq->mutex);
|
|
vhost_disable_notify(vq);
|
|
vhost_disable_notify(vq);
|
|
|
|
|
|
- if (wmem < sock->sk->sk_sndbuf * 2)
|
|
|
|
|
|
+ if (wmem < sock->sk->sk_sndbuf / 2)
|
|
tx_poll_stop(net);
|
|
tx_poll_stop(net);
|
|
hdr_size = vq->hdr_size;
|
|
hdr_size = vq->hdr_size;
|
|
|
|
|
|
@@ -508,12 +508,12 @@ static long vhost_net_set_backend(struct vhost_net *n, unsigned index, int fd)
|
|
/* Verify that ring has been setup correctly. */
|
|
/* Verify that ring has been setup correctly. */
|
|
if (!vhost_vq_access_ok(vq)) {
|
|
if (!vhost_vq_access_ok(vq)) {
|
|
r = -EFAULT;
|
|
r = -EFAULT;
|
|
- goto err;
|
|
|
|
|
|
+ goto err_vq;
|
|
}
|
|
}
|
|
sock = get_socket(fd);
|
|
sock = get_socket(fd);
|
|
if (IS_ERR(sock)) {
|
|
if (IS_ERR(sock)) {
|
|
r = PTR_ERR(sock);
|
|
r = PTR_ERR(sock);
|
|
- goto err;
|
|
|
|
|
|
+ goto err_vq;
|
|
}
|
|
}
|
|
|
|
|
|
/* start polling new socket */
|
|
/* start polling new socket */
|
|
@@ -524,12 +524,14 @@ static long vhost_net_set_backend(struct vhost_net *n, unsigned index, int fd)
|
|
vhost_net_disable_vq(n, vq);
|
|
vhost_net_disable_vq(n, vq);
|
|
rcu_assign_pointer(vq->private_data, sock);
|
|
rcu_assign_pointer(vq->private_data, sock);
|
|
vhost_net_enable_vq(n, vq);
|
|
vhost_net_enable_vq(n, vq);
|
|
- mutex_unlock(&vq->mutex);
|
|
|
|
done:
|
|
done:
|
|
if (oldsock) {
|
|
if (oldsock) {
|
|
vhost_net_flush_vq(n, index);
|
|
vhost_net_flush_vq(n, index);
|
|
fput(oldsock->file);
|
|
fput(oldsock->file);
|
|
}
|
|
}
|
|
|
|
+
|
|
|
|
+err_vq:
|
|
|
|
+ mutex_unlock(&vq->mutex);
|
|
err:
|
|
err:
|
|
mutex_unlock(&n->dev.mutex);
|
|
mutex_unlock(&n->dev.mutex);
|
|
return r;
|
|
return r;
|