|
@@ -311,7 +311,7 @@ ip_vs_bypass_xmit_v6(struct sk_buff *skb, struct ip_vs_conn *cp,
|
|
|
mtu = dst_mtu(&rt->u.dst);
|
|
|
if (skb->len > mtu) {
|
|
|
dst_release(&rt->u.dst);
|
|
|
- icmpv6_send(skb, ICMPV6_PKT_TOOBIG, 0, mtu, skb->dev);
|
|
|
+ icmpv6_send(skb, ICMPV6_PKT_TOOBIG, 0, mtu);
|
|
|
IP_VS_DBG_RL("%s(): frag needed\n", __func__);
|
|
|
goto tx_error;
|
|
|
}
|
|
@@ -454,7 +454,7 @@ ip_vs_nat_xmit_v6(struct sk_buff *skb, struct ip_vs_conn *cp,
|
|
|
mtu = dst_mtu(&rt->u.dst);
|
|
|
if (skb->len > mtu) {
|
|
|
dst_release(&rt->u.dst);
|
|
|
- icmpv6_send(skb, ICMPV6_PKT_TOOBIG, 0, mtu, skb->dev);
|
|
|
+ icmpv6_send(skb, ICMPV6_PKT_TOOBIG, 0, mtu);
|
|
|
IP_VS_DBG_RL_PKT(0, pp, skb, 0,
|
|
|
"ip_vs_nat_xmit_v6(): frag needed for");
|
|
|
goto tx_error;
|
|
@@ -672,7 +672,7 @@ ip_vs_tunnel_xmit_v6(struct sk_buff *skb, struct ip_vs_conn *cp,
|
|
|
skb_dst(skb)->ops->update_pmtu(skb_dst(skb), mtu);
|
|
|
|
|
|
if (mtu < ntohs(old_iph->payload_len) + sizeof(struct ipv6hdr)) {
|
|
|
- icmpv6_send(skb, ICMPV6_PKT_TOOBIG, 0, mtu, skb->dev);
|
|
|
+ icmpv6_send(skb, ICMPV6_PKT_TOOBIG, 0, mtu);
|
|
|
dst_release(&rt->u.dst);
|
|
|
IP_VS_DBG_RL("%s(): frag needed\n", __func__);
|
|
|
goto tx_error;
|
|
@@ -814,7 +814,7 @@ ip_vs_dr_xmit_v6(struct sk_buff *skb, struct ip_vs_conn *cp,
|
|
|
/* MTU checking */
|
|
|
mtu = dst_mtu(&rt->u.dst);
|
|
|
if (skb->len > mtu) {
|
|
|
- icmpv6_send(skb, ICMPV6_PKT_TOOBIG, 0, mtu, skb->dev);
|
|
|
+ icmpv6_send(skb, ICMPV6_PKT_TOOBIG, 0, mtu);
|
|
|
dst_release(&rt->u.dst);
|
|
|
IP_VS_DBG_RL("%s(): frag needed\n", __func__);
|
|
|
goto tx_error;
|
|
@@ -965,7 +965,7 @@ ip_vs_icmp_xmit_v6(struct sk_buff *skb, struct ip_vs_conn *cp,
|
|
|
mtu = dst_mtu(&rt->u.dst);
|
|
|
if (skb->len > mtu) {
|
|
|
dst_release(&rt->u.dst);
|
|
|
- icmpv6_send(skb, ICMPV6_PKT_TOOBIG, 0, mtu, skb->dev);
|
|
|
+ icmpv6_send(skb, ICMPV6_PKT_TOOBIG, 0, mtu);
|
|
|
IP_VS_DBG_RL("%s(): frag needed\n", __func__);
|
|
|
goto tx_error;
|
|
|
}
|