cns3xxx: fix RX softIRQ loop
[openwrt/openwrt.git] / target / linux / cns3xxx / files / drivers / net / ethernet / cavium / cns3xxx_eth.c
index 74a3ca09053dcd6ca5471d24bca6b194155c2462..51b01876efe6a61a4a734890eaeea5bbe231fd90 100644 (file)
@@ -306,8 +306,6 @@ static struct switch_regs __iomem *mdio_regs; /* mdio command and status only */
 struct mii_bus *mdio_bus;
 static int ports_open;
 static struct port *switch_port_tab[4];
-static struct dma_pool *rx_dma_pool;
-static struct dma_pool *tx_dma_pool;
 struct net_device *napi_dev;
 
 static int cns3xxx_mdio_cmd(struct mii_bus *bus, int phy_id, int location,
@@ -520,14 +518,14 @@ static void cns3xxx_alloc_rx_buf(struct sw *sw, int received)
        unsigned int phys;
 
        for (received += rx_ring->alloc_count; received > 0; received--) {
-               buf = kmalloc(RX_SEGMENT_ALLOC_SIZE, GFP_ATOMIC);
+               buf = napi_alloc_frag(RX_SEGMENT_ALLOC_SIZE);
                if (!buf)
                        break;
 
                phys = dma_map_single(sw->dev, buf + SKB_HEAD_ALIGN,
                                      RX_SEGMENT_MRU, DMA_FROM_DEVICE);
                if (dma_mapping_error(sw->dev, phys)) {
-                       kfree(buf);
+                       skb_free_frag(buf);
                        break;
                }
 
@@ -638,7 +636,7 @@ static int eth_poll(struct napi_struct *napi, int budget)
                dma_unmap_single(sw->dev, rx_ring->phys_tab[i],
                                 RX_SEGMENT_MRU, DMA_FROM_DEVICE);
 
-               skb = build_skb(rx_ring->buff_tab[i], 0);
+               skb = build_skb(rx_ring->buff_tab[i], RX_SEGMENT_ALLOC_SIZE);
                if (!skb)
                        break;
 
@@ -719,8 +717,8 @@ static int eth_poll(struct napi_struct *napi, int budget)
                enable_irq(sw->rx_irq);
                budget = 0;
 
-               /* if rx descriptors are full schedule another poll */
-               if (rx_ring->desc[(i-1) & (RX_DESCS-1)].cown)
+               /* If 1 or more frames came in during IRQ enable, re-schedule */
+               if (rx_ring->desc[i].cown)
                        eth_schedule_poll(sw);
        }
 
@@ -898,21 +896,18 @@ static int init_rings(struct sw *sw)
 
        __raw_writel(QUEUE_THRESHOLD, &sw->regs->dma_ring_ctrl);
 
-       if (!(rx_dma_pool = dma_pool_create(DRV_NAME, sw->dev,
-                                           RX_POOL_ALLOC_SIZE, 32, 0)))
+       rx_ring->desc = dmam_alloc_coherent(sw->dev, RX_POOL_ALLOC_SIZE,
+                                           &rx_ring->phys_addr, GFP_KERNEL);
+       if (!rx_ring->desc)
                return -ENOMEM;
 
-       if (!(rx_ring->desc = dma_pool_alloc(rx_dma_pool, GFP_KERNEL,
-                                             &rx_ring->phys_addr)))
-               return -ENOMEM;
-       memset(rx_ring->desc, 0, RX_POOL_ALLOC_SIZE);
-
        /* Setup RX buffers */
+       memset(rx_ring->desc, 0, RX_POOL_ALLOC_SIZE);
        for (i = 0; i < RX_DESCS; i++) {
                struct rx_desc *desc = &(rx_ring)->desc[i];
                void *buf;
 
-               buf = kzalloc(RX_SEGMENT_ALLOC_SIZE, GFP_KERNEL);
+               buf = netdev_alloc_frag(RX_SEGMENT_ALLOC_SIZE);
                if (!buf)
                        return -ENOMEM;
 
@@ -934,16 +929,13 @@ static int init_rings(struct sw *sw)
        __raw_writel(rx_ring->phys_addr, &sw->regs->fs_desc_ptr0);
        __raw_writel(rx_ring->phys_addr, &sw->regs->fs_desc_base_addr0);
 
-       if (!(tx_dma_pool = dma_pool_create(DRV_NAME, sw->dev,
-                                           TX_POOL_ALLOC_SIZE, 32, 0)))
+       tx_ring->desc = dmam_alloc_coherent(sw->dev, TX_POOL_ALLOC_SIZE,
+                                           &tx_ring->phys_addr, GFP_KERNEL);
+       if (!tx_ring->desc)
                return -ENOMEM;
 
-       if (!(tx_ring->desc = dma_pool_alloc(tx_dma_pool, GFP_KERNEL,
-                                             &tx_ring->phys_addr)))
-               return -ENOMEM;
-       memset(tx_ring->desc, 0, TX_POOL_ALLOC_SIZE);
-
        /* Setup TX buffers */
+       memset(tx_ring->desc, 0, TX_POOL_ALLOC_SIZE);
        for (i = 0; i < TX_DESCS; i++) {
                struct tx_desc *desc = &(tx_ring)->desc[i];
                tx_ring->buff_tab[i] = 0;
@@ -961,39 +953,30 @@ static int init_rings(struct sw *sw)
 static void destroy_rings(struct sw *sw)
 {
        int i;
-       if (sw->rx_ring.desc) {
-               for (i = 0; i < RX_DESCS; i++) {
-                       struct _rx_ring *rx_ring = &sw->rx_ring;
-                       struct rx_desc *desc = &(rx_ring)->desc[i];
-                       struct sk_buff *skb = sw->rx_ring.buff_tab[i];
-
-                       if (!skb)
-                               continue;
-
-                       dma_unmap_single(sw->dev, desc->sdp, RX_SEGMENT_MRU,
-                                        DMA_FROM_DEVICE);
-                       dev_kfree_skb(skb);
-               }
-               dma_pool_free(rx_dma_pool, sw->rx_ring.desc, sw->rx_ring.phys_addr);
-               dma_pool_destroy(rx_dma_pool);
-               rx_dma_pool = 0;
-               sw->rx_ring.desc = 0;
+
+       for (i = 0; i < RX_DESCS; i++) {
+               struct _rx_ring *rx_ring = &sw->rx_ring;
+               struct rx_desc *desc = &(rx_ring)->desc[i];
+               void *buf = sw->rx_ring.buff_tab[i];
+
+               if (!buf)
+                       continue;
+
+               dma_unmap_single(sw->dev, desc->sdp, RX_SEGMENT_MRU,
+                                DMA_FROM_DEVICE);
+               skb_free_frag(buf);
        }
-       if (sw->tx_ring.desc) {
-               for (i = 0; i < TX_DESCS; i++) {
-                       struct _tx_ring *tx_ring = &sw->tx_ring;
-                       struct tx_desc *desc = &(tx_ring)->desc[i];
-                       struct sk_buff *skb = sw->tx_ring.buff_tab[i];
-                       if (skb) {
-                               dma_unmap_single(sw->dev, desc->sdp,
-                                       skb->len, DMA_TO_DEVICE);
-                               dev_kfree_skb(skb);
-                       }
-               }
-               dma_pool_free(tx_dma_pool, sw->tx_ring.desc, sw->tx_ring.phys_addr);
-               dma_pool_destroy(tx_dma_pool);
-               tx_dma_pool = 0;
-               sw->tx_ring.desc = 0;
+
+       for (i = 0; i < TX_DESCS; i++) {
+               struct _tx_ring *tx_ring = &sw->tx_ring;
+               struct tx_desc *desc = &(tx_ring)->desc[i];
+               struct sk_buff *skb = sw->tx_ring.buff_tab[i];
+
+               if (!skb)
+                       continue;
+
+               dma_unmap_single(sw->dev, desc->sdp, skb->len, DMA_TO_DEVICE);
+               dev_kfree_skb(skb);
        }
 }
 
@@ -1227,7 +1210,6 @@ static int eth_init_one(struct platform_device *pdev)
                     CRC_STRIPPING, &sw->regs->mac_glob_cfg);
 
        if ((err = init_rings(sw)) != 0) {
-               destroy_rings(sw);
                err = -ENOMEM;
                goto err_free;
        }
@@ -1314,8 +1296,8 @@ static int eth_remove_one(struct platform_device *pdev)
        struct net_device *dev = platform_get_drvdata(pdev);
        struct sw *sw = netdev_priv(dev);
        int i;
-       destroy_rings(sw);
 
+       destroy_rings(sw);
        for (i = 3; i >= 0; i--) {
                if (switch_port_tab[i]) {
                        struct port *port = switch_port_tab[i];