|
@@ -829,7 +829,7 @@ static int skge_rx_fill(struct skge_port *skge)
|
|
do {
|
|
do {
|
|
struct sk_buff *skb;
|
|
struct sk_buff *skb;
|
|
|
|
|
|
- skb = dev_alloc_skb(skge->rx_buf_size + NET_IP_ALIGN);
|
|
|
|
|
|
+ skb = alloc_skb(skge->rx_buf_size + NET_IP_ALIGN, GFP_KERNEL);
|
|
if (!skb)
|
|
if (!skb)
|
|
return -ENOMEM;
|
|
return -ENOMEM;
|
|
|
|
|
|
@@ -2592,7 +2592,7 @@ static inline struct sk_buff *skge_rx_get(struct skge_port *skge,
|
|
goto error;
|
|
goto error;
|
|
|
|
|
|
if (len < RX_COPY_THRESHOLD) {
|
|
if (len < RX_COPY_THRESHOLD) {
|
|
- skb = dev_alloc_skb(len + 2);
|
|
|
|
|
|
+ skb = alloc_skb(len + 2, GFP_ATOMIC);
|
|
if (!skb)
|
|
if (!skb)
|
|
goto resubmit;
|
|
goto resubmit;
|
|
|
|
|
|
@@ -2607,7 +2607,7 @@ static inline struct sk_buff *skge_rx_get(struct skge_port *skge,
|
|
skge_rx_reuse(e, skge->rx_buf_size);
|
|
skge_rx_reuse(e, skge->rx_buf_size);
|
|
} else {
|
|
} else {
|
|
struct sk_buff *nskb;
|
|
struct sk_buff *nskb;
|
|
- nskb = dev_alloc_skb(skge->rx_buf_size + NET_IP_ALIGN);
|
|
|
|
|
|
+ nskb = alloc_skb(skge->rx_buf_size + NET_IP_ALIGN, GFP_ATOMIC);
|
|
if (!nskb)
|
|
if (!nskb)
|
|
goto resubmit;
|
|
goto resubmit;
|
|
|
|
|