On 12/1/2016 11:48 PM, Pavel Machek wrote:

@@ -2771,12 +2771,8 @@ static netdev_features_t stmmac_fix_features(struct 
net_device *dev,
                features &= ~NETIF_F_CSUM_MASK;

        /* Disable tso if asked by ethtool */
-       if ((priv->plat->tso_en) && (priv->dma_cap.tsoen)) {
-               if (features & NETIF_F_TSO)
-                       priv->tso = true;
-               else
-                       priv->tso = false;
-       }
+       if ((priv->plat->tso_en) && (priv->dma_cap.tsoen))
+               priv->tso = !!(features & NETIF_F_TSO);


Pavel, this really seems arbitrary.

Whilst I really appreciate you're looking into this driver a bit because
of some issues you are trying to resolve, I'd like to ask that you not
start bombarding me with nit-pick cleanups here and there and instead
concentrate on the real bug or issue.

Well, fixing clean code is easier than fixing strange code... Plus I
was hoping to make the mainainers to talk. The driver is listed as
supported after all.

Absolutely, I am available to support you, better I can.
So no problem to clarify strange or complex parts of the driver
and find/try new solutions to enhance it.

Anyway... since you asked. I belive I have way to disable NAPI / tx
coalescing in the driver. Unfortunately, locking is missing on the rx
path, and needs to be extended to _irqsave variant on tx path.

I have just replied to a previous thread about that...

To be honest, I have in the box just a patch to fix lock on lpi
as I had discussed in this mailing list some week ago.
I will provide it asap.


So patch currently looks like this (hand edited, can't be
applied, got it working few hours ago). Does it look acceptable?

I'd prefer this to go after the patch that pulls common code to single
place, so that single place needs to be patched. Plus I guess I should
add ifdefs, so that more advanced NAPI / tx coalescing code can be
reactivated when it is fixed. Trivial fixes can go on top. Does that
sound like a plan?

Hmm, what I find strange is that, just this code is running since a
long time on several platforms and Chip versions. No raise condition
have been found or lock protection problems (also proving look
mechanisms).

I'd like to avoid to break old compatibilities and having the same
performances but if there are some bugs I can support to review
and test. Indeed, this year we have added the 4.x but some parts
of the code (for TSO) should be self-contained. So I cannot image
regressions on common part of the code... I let Alex to do a double
check.

Pavel, I ask you sorry if I missed some problems so, if you can
(as D. Miller asked) to send us a cover letter + all patches
I will try to reply soon. I can do also some tests if you ask
me that! I could run on 3.x and 4.x but I cannot promise you
benchmarks.

Which tree do you want patches against?

https://git.kernel.org/cgit/linux/kernel/git/davem/net-next.git/ ?

I think that bug fixing should be on top of net.git but I let Miller
to decide.

Best Regards
Peppe


Best regards,
                                                                Pavel


diff --git a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c 
b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
index 0b706a7..c0016c8 100644
--- a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
+++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c
@@ -1395,9 +1397,10 @@ static void __stmmac_tx_clean(struct stmmac_priv *priv)

 static void stmmac_tx_clean(struct stmmac_priv *priv)
 {
-       spin_lock(&priv->tx_lock);
+       unsigned long flags;
+       spin_lock_irqsave(&priv->tx_lock, flags);
        __stmmac_tx_clean(priv);
-       spin_unlock(&priv->tx_lock);     
+       spin_unlock_irqrestore(&priv->tx_lock, flags);   
 }

 static inline void stmmac_enable_dma_irq(struct stmmac_priv *priv)
@@ -1441,6 +1444,8 @@ static void stmmac_tx_err(struct stmmac_priv *priv)
        netif_wake_queue(priv->dev);
 }

+static int stmmac_rx(struct stmmac_priv *priv, int limit);
+
 /**
  * stmmac_dma_interrupt - DMA ISR
  * @priv: driver private structure
@@ -1452,10 +1457,17 @@ static void stmmac_dma_interrupt(struct stmmac_priv 
*priv)
 {
        int status;
        int rxfifosz = priv->plat->rx_fifo_size;
+       unsigned long flags;

        status = priv->hw->dma->dma_interrupt(priv->ioaddr, &priv->xstats);
        if (likely((status & handle_rx)) || (status & handle_tx)) {
+               int r;
+               spin_lock_irqsave(&priv->tx_lock, flags);
+               r = stmmac_rx(priv, 999);
+               spin_unlock_irqrestore(&priv->tx_lock, flags);           
+#if 0
                if (likely(napi_schedule_prep(&priv->napi))) {
                        //pr_err("napi: schedule\n");
                        stmmac_disable_dma_irq(priv);
@@ -1463,7 +1475,8 @@ static void stmmac_dma_interrupt(struct stmmac_priv *priv)
                } else
                        pr_err("napi: schedule failed\n");
 #endif
+               stmmac_tx_clean(priv);
        }
        if (unlikely(status & tx_hard_error_bump_tc)) {
                /* Try to bump up the dma threshold on this failure */
@@ -1638,7 +1651,7 @@ static void stmmac_tx_timer(unsigned long data)
 {
        struct stmmac_priv *priv = (struct stmmac_priv *)data;

-       stmmac_tx_clean(priv);
+       //stmmac_tx_clean(priv);
 }

 /**
@@ -1990,6 +2003,8 @@ static void stmmac_xmit_common(struct sk_buff *skb, 
struct net_device *dev, int
        if (likely(priv->tx_coal_frames > priv->tx_count_frames)) {
                mod_timer(&priv->txtimer,
                          STMMAC_COAL_TIMER(priv->tx_coal_timer));
+               priv->hw->desc->set_tx_ic(desc);
+               priv->xstats.tx_set_ic_bit++;
        } else {
                priv->tx_count_frames = 0;
                priv->hw->desc->set_tx_ic(desc);
@@ -2038,8 +2053,9 @@ static netdev_tx_t stmmac_tso_xmit(struct sk_buff *skb, 
struct net_device *dev)
        struct dma_desc *desc, *first, *mss_desc = NULL;
        u8 proto_hdr_len;
        int i;
+       unsigned long flags;

-       spin_lock(&priv->tx_lock);
+       spin_lock_irqsave(&priv->tx_lock, flags);

        /* Compute header lengths */
        proto_hdr_len = skb_transport_offset(skb) + tcp_hdrlen(skb);
@@ -2052,7 +2068,7 @@ static netdev_tx_t stmmac_tso_xmit(struct sk_buff *skb, 
struct net_device *dev)
                        /* This is a hard error, log it. */
                        pr_err("%s: Tx Ring full when queue awake\n", __func__);
                }
-               spin_unlock(&priv->tx_lock);
+               spin_unlock_irqrestore(&priv->tx_lock, flags);
                return NETDEV_TX_BUSY;
        }

@@ -2168,11 +2184,11 @@ static netdev_tx_t stmmac_tso_xmit(struct sk_buff *skb, 
struct net_device *dev)
        priv->hw->dma->set_tx_tail_ptr(priv->ioaddr, priv->tx_tail_addr,
                                       STMMAC_CHAN0);

-       spin_unlock(&priv->tx_lock);
+       spin_unlock_irqrestore(&priv->tx_lock, flags);
        return NETDEV_TX_OK;

 dma_map_err:
-       spin_unlock(&priv->tx_lock);
+       spin_unlock_irqrestore(&priv->tx_lock, flags);
        dev_err(priv->device, "Tx dma map failed\n");
        dev_kfree_skb(skb);
        priv->dev->stats.tx_dropped++;
@@ -2197,6 +2213,7 @@ static netdev_tx_t stmmac_xmit(struct sk_buff *skb, 
struct net_device *dev)
        struct dma_desc *desc, *first;
        unsigned int enh_desc;
        unsigned int des;
+       unsigned int flags;

        /* Manage oversized TCP frames for GMAC4 device */
        if (skb_is_gso(skb) && priv->tso) {
@@ -2204,7 +2221,7 @@ static netdev_tx_t stmmac_xmit(struct sk_buff *skb, 
struct net_device *dev)
                        return stmmac_tso_xmit(skb, dev);
        }

-       spin_lock(&priv->tx_lock);
+       spin_lock_irqsave(&priv->tx_lock, flags);

        if (unlikely(stmmac_tx_avail(priv) < nfrags + 1)) {
                if (!netif_queue_stopped(dev)) {
@@ -2212,7 +2229,7 @@ static netdev_tx_t stmmac_xmit(struct sk_buff *skb, 
struct net_device *dev)
                        /* This is a hard error, log it. */
                        pr_err("%s: Tx Ring full when queue awake\n", __func__);
                }
-               spin_unlock(&priv->tx_lock);
+               spin_unlock_irqrestore(&priv->tx_lock, flags);
                return NETDEV_TX_BUSY;
        }

@@ -2347,11 +2364,11 @@ static netdev_tx_t stmmac_xmit(struct sk_buff *skb, 
struct net_device *dev)
                priv->hw->dma->set_tx_tail_ptr(priv->ioaddr, priv->tx_tail_addr,
                                               STMMAC_CHAN0);

-       spin_unlock(&priv->tx_lock);
+       spin_unlock_irqrestore(&priv->tx_lock, flags);
        return NETDEV_TX_OK;

 dma_map_err:
-       spin_unlock(&priv->tx_lock);
+       spin_unlock_irqrestore(&priv->tx_lock, flags);
        dev_err(priv->device, "Tx dma map failed\n");
        dev_kfree_skb(skb);
        priv->dev->stats.tx_dropped++;
@@ -2634,7 +2651,8 @@ static int stmmac_rx(struct stmmac_priv *priv, int limit)
                        else
                                skb->ip_summed = CHECKSUM_UNNECESSARY;

-                       napi_gro_receive(&priv->napi, skb);
+                       //napi_gro_receive(&priv->napi, skb);
+                       netif_rx(skb);

                        priv->dev->stats.rx_packets++;
                        priv->dev->stats.rx_bytes += frame_len;
@@ -2662,6 +2680,7 @@ static int stmmac_poll(struct napi_struct *napi, int 
budget)
        struct stmmac_priv *priv = container_of(napi, struct stmmac_priv, napi);
        int work_done;

+       BUG();
        priv->xstats.napi_poll++;
        stmmac_tx_clean(priv);





Reply via email to