|
@@ -62,8 +62,6 @@ static int rq2_entries = EHEA_DEF_ENTRIES_RQ2;
|
|
static int rq3_entries = EHEA_DEF_ENTRIES_RQ3;
|
|
static int rq3_entries = EHEA_DEF_ENTRIES_RQ3;
|
|
static int sq_entries = EHEA_DEF_ENTRIES_SQ;
|
|
static int sq_entries = EHEA_DEF_ENTRIES_SQ;
|
|
static int use_mcs = 1;
|
|
static int use_mcs = 1;
|
|
-static int use_lro;
|
|
|
|
-static int lro_max_aggr = EHEA_LRO_MAX_AGGR;
|
|
|
|
static int prop_carrier_state;
|
|
static int prop_carrier_state;
|
|
|
|
|
|
module_param(msg_level, int, 0);
|
|
module_param(msg_level, int, 0);
|
|
@@ -73,8 +71,6 @@ module_param(rq3_entries, int, 0);
|
|
module_param(sq_entries, int, 0);
|
|
module_param(sq_entries, int, 0);
|
|
module_param(prop_carrier_state, int, 0);
|
|
module_param(prop_carrier_state, int, 0);
|
|
module_param(use_mcs, int, 0);
|
|
module_param(use_mcs, int, 0);
|
|
-module_param(use_lro, int, 0);
|
|
|
|
-module_param(lro_max_aggr, int, 0);
|
|
|
|
|
|
|
|
MODULE_PARM_DESC(msg_level, "msg_level");
|
|
MODULE_PARM_DESC(msg_level, "msg_level");
|
|
MODULE_PARM_DESC(prop_carrier_state, "Propagate carrier state of physical "
|
|
MODULE_PARM_DESC(prop_carrier_state, "Propagate carrier state of physical "
|
|
@@ -94,11 +90,6 @@ MODULE_PARM_DESC(sq_entries, " Number of entries for the Send Queue "
|
|
MODULE_PARM_DESC(use_mcs, " Multiple receive queues, 1: enable, 0: disable, "
|
|
MODULE_PARM_DESC(use_mcs, " Multiple receive queues, 1: enable, 0: disable, "
|
|
"Default = 1");
|
|
"Default = 1");
|
|
|
|
|
|
-MODULE_PARM_DESC(lro_max_aggr, " LRO: Max packets to be aggregated. Default = "
|
|
|
|
- __MODULE_STRING(EHEA_LRO_MAX_AGGR));
|
|
|
|
-MODULE_PARM_DESC(use_lro, " Large Receive Offload, 1: enable, 0: disable, "
|
|
|
|
- "Default = 0");
|
|
|
|
-
|
|
|
|
static int port_name_cnt;
|
|
static int port_name_cnt;
|
|
static LIST_HEAD(adapter_list);
|
|
static LIST_HEAD(adapter_list);
|
|
static unsigned long ehea_driver_flags;
|
|
static unsigned long ehea_driver_flags;
|
|
@@ -656,47 +647,13 @@ static int ehea_treat_poll_error(struct ehea_port_res *pr, int rq,
|
|
return 0;
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
|
|
-static int get_skb_hdr(struct sk_buff *skb, void **iphdr,
|
|
|
|
- void **tcph, u64 *hdr_flags, void *priv)
|
|
|
|
-{
|
|
|
|
- struct ehea_cqe *cqe = priv;
|
|
|
|
- unsigned int ip_len;
|
|
|
|
- struct iphdr *iph;
|
|
|
|
-
|
|
|
|
- /* non tcp/udp packets */
|
|
|
|
- if (!cqe->header_length)
|
|
|
|
- return -1;
|
|
|
|
-
|
|
|
|
- /* non tcp packet */
|
|
|
|
- skb_reset_network_header(skb);
|
|
|
|
- iph = ip_hdr(skb);
|
|
|
|
- if (iph->protocol != IPPROTO_TCP)
|
|
|
|
- return -1;
|
|
|
|
-
|
|
|
|
- ip_len = ip_hdrlen(skb);
|
|
|
|
- skb_set_transport_header(skb, ip_len);
|
|
|
|
- *tcph = tcp_hdr(skb);
|
|
|
|
-
|
|
|
|
- /* check if ip header and tcp header are complete */
|
|
|
|
- if (ntohs(iph->tot_len) < ip_len + tcp_hdrlen(skb))
|
|
|
|
- return -1;
|
|
|
|
-
|
|
|
|
- *hdr_flags = LRO_IPV4 | LRO_TCP;
|
|
|
|
- *iphdr = iph;
|
|
|
|
-
|
|
|
|
- return 0;
|
|
|
|
-}
|
|
|
|
-
|
|
|
|
static void ehea_proc_skb(struct ehea_port_res *pr, struct ehea_cqe *cqe,
|
|
static void ehea_proc_skb(struct ehea_port_res *pr, struct ehea_cqe *cqe,
|
|
struct sk_buff *skb)
|
|
struct sk_buff *skb)
|
|
{
|
|
{
|
|
if (cqe->status & EHEA_CQE_VLAN_TAG_XTRACT)
|
|
if (cqe->status & EHEA_CQE_VLAN_TAG_XTRACT)
|
|
__vlan_hwaccel_put_tag(skb, cqe->vlan_tag);
|
|
__vlan_hwaccel_put_tag(skb, cqe->vlan_tag);
|
|
|
|
|
|
- if (skb->dev->features & NETIF_F_LRO)
|
|
|
|
- lro_receive_skb(&pr->lro_mgr, skb, cqe);
|
|
|
|
- else
|
|
|
|
- netif_receive_skb(skb);
|
|
|
|
|
|
+ netif_receive_skb(skb);
|
|
}
|
|
}
|
|
|
|
|
|
static int ehea_proc_rwqes(struct net_device *dev,
|
|
static int ehea_proc_rwqes(struct net_device *dev,
|
|
@@ -786,8 +743,6 @@ static int ehea_proc_rwqes(struct net_device *dev,
|
|
}
|
|
}
|
|
cqe = ehea_poll_rq1(qp, &wqe_index);
|
|
cqe = ehea_poll_rq1(qp, &wqe_index);
|
|
}
|
|
}
|
|
- if (dev->features & NETIF_F_LRO)
|
|
|
|
- lro_flush_all(&pr->lro_mgr);
|
|
|
|
|
|
|
|
pr->rx_packets += processed;
|
|
pr->rx_packets += processed;
|
|
pr->rx_bytes += processed_bytes;
|
|
pr->rx_bytes += processed_bytes;
|
|
@@ -1611,15 +1566,6 @@ static int ehea_init_port_res(struct ehea_port *port, struct ehea_port_res *pr,
|
|
|
|
|
|
netif_napi_add(pr->port->netdev, &pr->napi, ehea_poll, 64);
|
|
netif_napi_add(pr->port->netdev, &pr->napi, ehea_poll, 64);
|
|
|
|
|
|
- pr->lro_mgr.max_aggr = pr->port->lro_max_aggr;
|
|
|
|
- pr->lro_mgr.max_desc = MAX_LRO_DESCRIPTORS;
|
|
|
|
- pr->lro_mgr.lro_arr = pr->lro_desc;
|
|
|
|
- pr->lro_mgr.get_skb_header = get_skb_hdr;
|
|
|
|
- pr->lro_mgr.features = LRO_F_NAPI | LRO_F_EXTRACT_VLAN_ID;
|
|
|
|
- pr->lro_mgr.dev = port->netdev;
|
|
|
|
- pr->lro_mgr.ip_summed = CHECKSUM_UNNECESSARY;
|
|
|
|
- pr->lro_mgr.ip_summed_aggr = CHECKSUM_UNNECESSARY;
|
|
|
|
-
|
|
|
|
ret = 0;
|
|
ret = 0;
|
|
goto out;
|
|
goto out;
|
|
|
|
|
|
@@ -3082,9 +3028,6 @@ struct ehea_port *ehea_setup_single_port(struct ehea_adapter *adapter,
|
|
NETIF_F_IP_CSUM;
|
|
NETIF_F_IP_CSUM;
|
|
dev->watchdog_timeo = EHEA_WATCH_DOG_TIMEOUT;
|
|
dev->watchdog_timeo = EHEA_WATCH_DOG_TIMEOUT;
|
|
|
|
|
|
- if (use_lro)
|
|
|
|
- dev->features |= NETIF_F_LRO;
|
|
|
|
-
|
|
|
|
INIT_WORK(&port->reset_task, ehea_reset_port);
|
|
INIT_WORK(&port->reset_task, ehea_reset_port);
|
|
INIT_DELAYED_WORK(&port->stats_work, ehea_update_stats);
|
|
INIT_DELAYED_WORK(&port->stats_work, ehea_update_stats);
|
|
|
|
|
|
@@ -3098,8 +3041,6 @@ struct ehea_port *ehea_setup_single_port(struct ehea_adapter *adapter,
|
|
goto out_unreg_port;
|
|
goto out_unreg_port;
|
|
}
|
|
}
|
|
|
|
|
|
- port->lro_max_aggr = lro_max_aggr;
|
|
|
|
-
|
|
|
|
ret = ehea_get_jumboframe_status(port, &jumbo);
|
|
ret = ehea_get_jumboframe_status(port, &jumbo);
|
|
if (ret)
|
|
if (ret)
|
|
netdev_err(dev, "failed determining jumbo frame status\n");
|
|
netdev_err(dev, "failed determining jumbo frame status\n");
|