Cleanup transmit buffers using NAPI. This allows the transmit routine
to leave interrupts enabled, and that improves performance.
Signed-off-by: Stephen Hemminger [EMAIL PROTECTED]
--- skge-2.6.orig/drivers/net/skge.c
+++ skge-2.6/drivers/net/skge.c
@@ -2307,16 +2307,13 @@ static int skge_xmit_frame(struct sk_buf
int i;
u32 control, len;
u64 map;
- unsigned long flags;
skb = skb_padto(skb, ETH_ZLEN);
if (!skb)
return NETDEV_TX_OK;
- local_irq_save(flags);
if (!spin_trylock(skge-tx_lock)) {
/* Collision - tell upper layer to requeue */
- local_irq_restore(flags);
return NETDEV_TX_LOCKED;
}
@@ -2327,7 +2324,7 @@ static int skge_xmit_frame(struct sk_buf
printk(KERN_WARNING PFX %s: ring full when queue
awake!\n,
dev-name);
}
- spin_unlock_irqrestore(skge-tx_lock, flags);
+ spin_unlock(skge-tx_lock);
return NETDEV_TX_BUSY;
}
@@ -2403,7 +2400,7 @@ static int skge_xmit_frame(struct sk_buf
}
dev-trans_start = jiffies;
- spin_unlock_irqrestore(skge-tx_lock, flags);
+ spin_unlock(skge-tx_lock);
return NETDEV_TX_OK;
}
@@ -2416,7 +2413,7 @@ static inline void skge_tx_free(struct s
pci_unmap_addr(e, mapaddr),
pci_unmap_len(e, maplen),
PCI_DMA_TODEVICE);
- dev_kfree_skb_any(e-skb);
+ dev_kfree_skb(e-skb);
e-skb = NULL;
} else {
pci_unmap_page(hw-pdev,
@@ -2430,15 +2427,14 @@ static void skge_tx_clean(struct skge_po
{
struct skge_ring *ring = skge-tx_ring;
struct skge_element *e;
- unsigned long flags;
- spin_lock_irqsave(skge-tx_lock, flags);
+ spin_lock_bh(skge-tx_lock);
for (e = ring-to_clean; e != ring-to_use; e = e-next) {
++skge-tx_avail;
skge_tx_free(skge-hw, e);
}
ring-to_clean = e;
- spin_unlock_irqrestore(skge-tx_lock, flags);
+ spin_unlock_bh(skge-tx_lock);
}
static void skge_tx_timeout(struct net_device *dev)
@@ -2663,6 +2659,37 @@ resubmit:
return NULL;
}
+static void skge_tx_done(struct skge_port *skge)
+{
+ struct skge_ring *ring = skge-tx_ring;
+ struct skge_element *e;
+
+ spin_lock(skge-tx_lock);
+ for (e = ring-to_clean; prefetch(e-next), e != ring-to_use; e =
e-next) {
+ struct skge_tx_desc *td = e-desc;
+ u32 control;
+
+ rmb();
+ control = td-control;
+ if (control BMU_OWN)
+ break;
+
+ if (unlikely(netif_msg_tx_done(skge)))
+ printk(KERN_DEBUG PFX %s: tx done slot %td status
0x%x\n,
+ skge-netdev-name, e - ring-start, td-status);
+
+ skge_tx_free(skge-hw, e);
+ e-skb = NULL;
+ ++skge-tx_avail;
+ }
+ ring-to_clean = e;
+ skge_write8(skge-hw, Q_ADDR(txqaddr[skge-port], Q_CSR), CSR_IRQ_CL_F);
+
+ if (skge-tx_avail MAX_SKB_FRAGS + 1)
+ netif_wake_queue(skge-netdev);
+
+ spin_unlock(skge-tx_lock);
+}
static int skge_poll(struct net_device *dev, int *budget)
{
@@ -2670,8 +2697,10 @@ static int skge_poll(struct net_device *
struct skge_hw *hw = skge-hw;
struct skge_ring *ring = skge-rx_ring;
struct skge_element *e;
- unsigned int to_do = min(dev-quota, *budget);
- unsigned int work_done = 0;
+ int to_do = min(dev-quota, *budget);
+ int work_done = 0;
+
+ skge_tx_done(skge);
for (e = ring-to_clean; prefetch(e-next), work_done to_do; e =
e-next) {
struct skge_rx_desc *rd = e-desc;
@@ -2714,40 +2743,6 @@ static int skge_poll(struct net_device *
return 0;
}
-static inline void skge_tx_intr(struct net_device *dev)
-{
- struct skge_port *skge = netdev_priv(dev);
- struct skge_hw *hw = skge-hw;
- struct skge_ring *ring = skge-tx_ring;
- struct skge_element *e;
-
- spin_lock(skge-tx_lock);
- for (e = ring-to_clean; prefetch(e-next), e != ring-to_use; e =
e-next) {
- struct skge_tx_desc *td = e-desc;
- u32 control;
-
- rmb();
- control = td-control;
- if (control BMU_OWN)
- break;
-
- if (unlikely(netif_msg_tx_done(skge)))
- printk(KERN_DEBUG PFX %s: tx done slot %td status
0x%x\n,
- dev-name, e - ring-start, td-status);
-
- skge_tx_free(hw, e);
- e-skb = NULL;
- ++skge-tx_avail;
- }
- ring-to_clean = e;
- skge_write8(hw,