Message ID | dfc9cb24be56c18b27d9ee5f81612faedb8e5a8c.1491467980.git.jpinto@synopsys.com |
---|---|
State | Accepted, archived |
Delegated to: | David Miller |
Headers | show |
Survived 10/10 reboot + ping test Tested-by: Niklas Cassel <niklas.cassel@axis.com> On 04/06/2017 10:49 AM, Joao Pinto wrote: > This patch breaks several functions into RX and TX scopes, which > will be useful when adding multiple buffers mechanism. > > Signed-off-by: Joao Pinto <jpinto@synopsys.com> > --- > changes v2->v3: > - just to keep up with patch-set version > changes v1->v2: > - RX and TX inconsistency > - stmmac_free_rx_buffers renamed to stmmac_free_rx_buffer > - stmmac_free_tx_buffers renamed to stmmac_free_tx_buffer > - some useless comments were removed > > drivers/net/ethernet/stmicro/stmmac/stmmac_main.c | 352 ++++++++++++++++------ > 1 file changed, 266 insertions(+), 86 deletions(-) > > diff --git a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c > index 7cbda41..ff839e1 100644 > --- a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c > +++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c > @@ -889,24 +889,41 @@ static int stmmac_init_phy(struct net_device *dev) > return 0; > } > > -static void stmmac_display_rings(struct stmmac_priv *priv) > +static void stmmac_display_rx_rings(struct stmmac_priv *priv) > { > - void *head_rx, *head_tx; > + void *head_rx; > > - if (priv->extend_desc) { > + if (priv->extend_desc) > head_rx = (void *)priv->dma_erx; > - head_tx = (void *)priv->dma_etx; > - } else { > + else > head_rx = (void *)priv->dma_rx; > - head_tx = (void *)priv->dma_tx; > - } > > - /* Display Rx ring */ > + /* Display RX ring */ > priv->hw->desc->display_ring(head_rx, DMA_RX_SIZE, true); > - /* Display Tx ring */ > +} > + > +static void stmmac_display_tx_rings(struct stmmac_priv *priv) > +{ > + void *head_tx; > + > + if (priv->extend_desc) > + head_tx = (void *)priv->dma_etx; > + else > + head_tx = (void *)priv->dma_tx; > + > + /* Display TX ring */ > priv->hw->desc->display_ring(head_tx, DMA_TX_SIZE, false); > } > > +static void stmmac_display_rings(struct stmmac_priv *priv) > +{ > + /* Display RX ring */ > + stmmac_display_rx_rings(priv); > + > + /* Display TX ring */ > + stmmac_display_tx_rings(priv); > +} > + > static int stmmac_set_bfsize(int mtu, int bufsize) > { > int ret = bufsize; > @@ -924,16 +941,16 @@ static int stmmac_set_bfsize(int mtu, int bufsize) > } > > /** > - * stmmac_clear_descriptors - clear descriptors > + * stmmac_clear_rx_descriptors - clear RX descriptors > * @priv: driver private structure > - * Description: this function is called to clear the tx and rx descriptors > + * Description: this function is called to clear the RX descriptors > * in case of both basic and extended descriptors are used. > */ > -static void stmmac_clear_descriptors(struct stmmac_priv *priv) > +static void stmmac_clear_rx_descriptors(struct stmmac_priv *priv) > { > int i; > > - /* Clear the Rx/Tx descriptors */ > + /* Clear the RX descriptors */ > for (i = 0; i < DMA_RX_SIZE; i++) > if (priv->extend_desc) > priv->hw->desc->init_rx_desc(&priv->dma_erx[i].basic, > @@ -943,6 +960,19 @@ static void stmmac_clear_descriptors(struct stmmac_priv *priv) > priv->hw->desc->init_rx_desc(&priv->dma_rx[i], > priv->use_riwt, priv->mode, > (i == DMA_RX_SIZE - 1)); > +} > + > +/** > + * stmmac_clear_tx_descriptors - clear tx descriptors > + * @priv: driver private structure > + * Description: this function is called to clear the TX descriptors > + * in case of both basic and extended descriptors are used. > + */ > +static void stmmac_clear_tx_descriptors(struct stmmac_priv *priv) > +{ > + int i; > + > + /* Clear the TX descriptors */ > for (i = 0; i < DMA_TX_SIZE; i++) > if (priv->extend_desc) > priv->hw->desc->init_tx_desc(&priv->dma_etx[i].basic, > @@ -955,6 +985,21 @@ static void stmmac_clear_descriptors(struct stmmac_priv *priv) > } > > /** > + * stmmac_clear_descriptors - clear descriptors > + * @priv: driver private structure > + * Description: this function is called to clear the TX and RX descriptors > + * in case of both basic and extended descriptors are used. > + */ > +static void stmmac_clear_descriptors(struct stmmac_priv *priv) > +{ > + /* Clear the RX descriptors */ > + stmmac_clear_rx_descriptors(priv); > + > + /* Clear the TX descriptors */ > + stmmac_clear_tx_descriptors(priv); > +} > + > +/** > * stmmac_init_rx_buffers - init the RX descriptor buffer. > * @priv: driver private structure > * @p: descriptor pointer > @@ -996,7 +1041,12 @@ static int stmmac_init_rx_buffers(struct stmmac_priv *priv, struct dma_desc *p, > return 0; > } > > -static void stmmac_free_rx_buffers(struct stmmac_priv *priv, int i) > +/** > + * stmmac_free_rx_buffer - free RX dma buffers > + * @priv: private structure > + * @i: buffer index. > + */ > +static void stmmac_free_rx_buffer(struct stmmac_priv *priv, int i) > { > if (priv->rx_skbuff[i]) { > dma_unmap_single(priv->device, priv->rx_skbuff_dma[i], > @@ -1007,14 +1057,42 @@ static void stmmac_free_rx_buffers(struct stmmac_priv *priv, int i) > } > > /** > - * init_dma_desc_rings - init the RX/TX descriptor rings > + * stmmac_free_tx_buffer - free RX dma buffers > + * @priv: private structure > + * @i: buffer index. > + */ > +static void stmmac_free_tx_buffer(struct stmmac_priv *priv, int i) > +{ > + if (priv->tx_skbuff_dma[i].buf) { > + if (priv->tx_skbuff_dma[i].map_as_page) > + dma_unmap_page(priv->device, > + priv->tx_skbuff_dma[i].buf, > + priv->tx_skbuff_dma[i].len, > + DMA_TO_DEVICE); > + else > + dma_unmap_single(priv->device, > + priv->tx_skbuff_dma[i].buf, > + priv->tx_skbuff_dma[i].len, > + DMA_TO_DEVICE); > + } > + > + if (priv->tx_skbuff[i]) { > + dev_kfree_skb_any(priv->tx_skbuff[i]); > + priv->tx_skbuff[i] = NULL; > + priv->tx_skbuff_dma[i].buf = 0; > + priv->tx_skbuff_dma[i].map_as_page = false; > + } > +} > + > +/** > + * init_dma_rx_desc_rings - init the RX descriptor rings > * @dev: net device structure > * @flags: gfp flag. > - * Description: this function initializes the DMA RX/TX descriptors > + * Description: this function initializes the DMA RX descriptors > * and allocates the socket buffers. It supports the chained and ring > * modes. > */ > -static int init_dma_desc_rings(struct net_device *dev, gfp_t flags) > +static int init_dma_rx_desc_rings(struct net_device *dev, gfp_t flags) > { > int i; > struct stmmac_priv *priv = netdev_priv(dev); > @@ -1030,10 +1108,8 @@ static int init_dma_desc_rings(struct net_device *dev, gfp_t flags) > priv->dma_buf_sz = bfsize; > > netif_dbg(priv, probe, priv->dev, > - "(%s) dma_rx_phy=0x%08x dma_tx_phy=0x%08x\n", > - __func__, (u32)priv->dma_rx_phy, (u32)priv->dma_tx_phy); > + "(%s) dma_rx_phy=0x%08x\n", __func__, (u32)priv->dma_rx_phy); > > - /* RX INITIALIZATION */ > netif_dbg(priv, probe, priv->dev, > "SKB addresses:\nskb\t\tskb data\tdma data\n"); > > @@ -1058,20 +1134,46 @@ static int init_dma_desc_rings(struct net_device *dev, gfp_t flags) > > /* Setup the chained descriptor addresses */ > if (priv->mode == STMMAC_CHAIN_MODE) { > - if (priv->extend_desc) { > + if (priv->extend_desc) > priv->hw->mode->init(priv->dma_erx, priv->dma_rx_phy, > DMA_RX_SIZE, 1); > - priv->hw->mode->init(priv->dma_etx, priv->dma_tx_phy, > - DMA_TX_SIZE, 1); > - } else { > + else > priv->hw->mode->init(priv->dma_rx, priv->dma_rx_phy, > DMA_RX_SIZE, 0); > + } > + > + return 0; > +err_init_rx_buffers: > + while (--i >= 0) > + stmmac_free_rx_buffer(priv, i); > + return ret; > +} > + > +/** > + * init_dma_tx_desc_rings - init the TX descriptor rings > + * @dev: net device structure. > + * Description: this function initializes the DMA TX descriptors > + * and allocates the socket buffers. It supports the chained and ring > + * modes. > + */ > +static int init_dma_tx_desc_rings(struct net_device *dev) > +{ > + struct stmmac_priv *priv = netdev_priv(dev); > + int i; > + > + netif_dbg(priv, probe, priv->dev, > + "(%s) dma_tx_phy=0x%08x\n", __func__, (u32)priv->dma_tx_phy); > + > + /* Setup the chained descriptor addresses */ > + if (priv->mode == STMMAC_CHAIN_MODE) { > + if (priv->extend_desc) > + priv->hw->mode->init(priv->dma_etx, priv->dma_tx_phy, > + DMA_TX_SIZE, 1); > + else > priv->hw->mode->init(priv->dma_tx, priv->dma_tx_phy, > DMA_TX_SIZE, 0); > - } > } > > - /* TX INITIALIZATION */ > for (i = 0; i < DMA_TX_SIZE; i++) { > struct dma_desc *p; > if (priv->extend_desc) > @@ -1099,62 +1201,69 @@ static int init_dma_desc_rings(struct net_device *dev, gfp_t flags) > priv->cur_tx = 0; > netdev_reset_queue(priv->dev); > > + return 0; > +} > + > +/** > + * init_dma_desc_rings - init the RX/TX descriptor rings > + * @dev: net device structure > + * @flags: gfp flag. > + * Description: this function initializes the DMA RX/TX descriptors > + * and allocates the socket buffers. It supports the chained and ring > + * modes. > + */ > +static int init_dma_desc_rings(struct net_device *dev, gfp_t flags) > +{ > + struct stmmac_priv *priv = netdev_priv(dev); > + int ret; > + > + ret = init_dma_rx_desc_rings(dev, flags); > + if (ret) > + return ret; > + > + ret = init_dma_tx_desc_rings(dev); > + > stmmac_clear_descriptors(priv); > > if (netif_msg_hw(priv)) > stmmac_display_rings(priv); > > - return 0; > -err_init_rx_buffers: > - while (--i >= 0) > - stmmac_free_rx_buffers(priv, i); > return ret; > } > > +/** > + * dma_free_rx_skbufs - free RX dma buffers > + * @priv: private structure > + */ > static void dma_free_rx_skbufs(struct stmmac_priv *priv) > { > int i; > > for (i = 0; i < DMA_RX_SIZE; i++) > - stmmac_free_rx_buffers(priv, i); > + stmmac_free_rx_buffer(priv, i); > } > > +/** > + * dma_free_tx_skbufs - free TX dma buffers > + * @priv: private structure > + */ > static void dma_free_tx_skbufs(struct stmmac_priv *priv) > { > int i; > > - for (i = 0; i < DMA_TX_SIZE; i++) { > - if (priv->tx_skbuff_dma[i].buf) { > - if (priv->tx_skbuff_dma[i].map_as_page) > - dma_unmap_page(priv->device, > - priv->tx_skbuff_dma[i].buf, > - priv->tx_skbuff_dma[i].len, > - DMA_TO_DEVICE); > - else > - dma_unmap_single(priv->device, > - priv->tx_skbuff_dma[i].buf, > - priv->tx_skbuff_dma[i].len, > - DMA_TO_DEVICE); > - } > - > - if (priv->tx_skbuff[i]) { > - dev_kfree_skb_any(priv->tx_skbuff[i]); > - priv->tx_skbuff[i] = NULL; > - priv->tx_skbuff_dma[i].buf = 0; > - priv->tx_skbuff_dma[i].map_as_page = false; > - } > - } > + for (i = 0; i < DMA_TX_SIZE; i++) > + stmmac_free_tx_buffer(priv, i); > } > > /** > - * alloc_dma_desc_resources - alloc TX/RX resources. > + * alloc_dma_rx_desc_resources - alloc RX resources. > * @priv: private structure > * Description: according to which descriptor can be used (extend or basic) > * this function allocates the resources for TX and RX paths. In case of > * reception, for example, it pre-allocated the RX socket buffer in order to > * allow zero-copy mechanism. > */ > -static int alloc_dma_desc_resources(struct stmmac_priv *priv) > +static int alloc_dma_rx_desc_resources(struct stmmac_priv *priv) > { > int ret = -ENOMEM; > > @@ -1168,11 +1277,50 @@ static int alloc_dma_desc_resources(struct stmmac_priv *priv) > if (!priv->rx_skbuff) > goto err_rx_skbuff; > > + if (priv->extend_desc) { > + priv->dma_erx = dma_zalloc_coherent(priv->device, DMA_RX_SIZE * > + sizeof(struct > + dma_extended_desc), > + &priv->dma_rx_phy, > + GFP_KERNEL); > + if (!priv->dma_erx) > + goto err_dma; > + > + } else { > + priv->dma_rx = dma_zalloc_coherent(priv->device, DMA_RX_SIZE * > + sizeof(struct dma_desc), > + &priv->dma_rx_phy, > + GFP_KERNEL); > + if (!priv->dma_rx) > + goto err_dma; > + } > + > + return 0; > + > +err_dma: > + kfree(priv->rx_skbuff); > +err_rx_skbuff: > + kfree(priv->rx_skbuff_dma); > + return ret; > +} > + > +/** > + * alloc_dma_tx_desc_resources - alloc TX resources. > + * @priv: private structure > + * Description: according to which descriptor can be used (extend or basic) > + * this function allocates the resources for TX and RX paths. In case of > + * reception, for example, it pre-allocated the RX socket buffer in order to > + * allow zero-copy mechanism. > + */ > +static int alloc_dma_tx_desc_resources(struct stmmac_priv *priv) > +{ > + int ret = -ENOMEM; > + > priv->tx_skbuff_dma = kmalloc_array(DMA_TX_SIZE, > sizeof(*priv->tx_skbuff_dma), > GFP_KERNEL); > if (!priv->tx_skbuff_dma) > - goto err_tx_skbuff_dma; > + return -ENOMEM; > > priv->tx_skbuff = kmalloc_array(DMA_TX_SIZE, sizeof(struct sk_buff *), > GFP_KERNEL); > @@ -1180,14 +1328,6 @@ static int alloc_dma_desc_resources(struct stmmac_priv *priv) > goto err_tx_skbuff; > > if (priv->extend_desc) { > - priv->dma_erx = dma_zalloc_coherent(priv->device, DMA_RX_SIZE * > - sizeof(struct > - dma_extended_desc), > - &priv->dma_rx_phy, > - GFP_KERNEL); > - if (!priv->dma_erx) > - goto err_dma; > - > priv->dma_etx = dma_zalloc_coherent(priv->device, DMA_TX_SIZE * > sizeof(struct > dma_extended_desc), > @@ -1200,13 +1340,6 @@ static int alloc_dma_desc_resources(struct stmmac_priv *priv) > goto err_dma; > } > } else { > - priv->dma_rx = dma_zalloc_coherent(priv->device, DMA_RX_SIZE * > - sizeof(struct dma_desc), > - &priv->dma_rx_phy, > - GFP_KERNEL); > - if (!priv->dma_rx) > - goto err_dma; > - > priv->dma_tx = dma_zalloc_coherent(priv->device, DMA_TX_SIZE * > sizeof(struct dma_desc), > &priv->dma_tx_phy, > @@ -1225,42 +1358,89 @@ static int alloc_dma_desc_resources(struct stmmac_priv *priv) > kfree(priv->tx_skbuff); > err_tx_skbuff: > kfree(priv->tx_skbuff_dma); > -err_tx_skbuff_dma: > - kfree(priv->rx_skbuff); > -err_rx_skbuff: > - kfree(priv->rx_skbuff_dma); > return ret; > } > > -static void free_dma_desc_resources(struct stmmac_priv *priv) > +/** > + * alloc_dma_desc_resources - alloc TX/RX resources. > + * @priv: private structure > + * Description: according to which descriptor can be used (extend or basic) > + * this function allocates the resources for TX and RX paths. In case of > + * reception, for example, it pre-allocated the RX socket buffer in order to > + * allow zero-copy mechanism. > + */ > +static int alloc_dma_desc_resources(struct stmmac_priv *priv) > { > - /* Release the DMA TX/RX socket buffers */ > + int ret = alloc_dma_rx_desc_resources(priv); > + > + if (ret) > + return ret; > + > + ret = alloc_dma_tx_desc_resources(priv); > + > + return ret; > +} > + > +/** > + * free_dma_rx_desc_resources - free RX dma desc resources > + * @priv: private structure > + */ > +static void free_dma_rx_desc_resources(struct stmmac_priv *priv) > +{ > + /* Release the DMA RX socket buffers */ > dma_free_rx_skbufs(priv); > - dma_free_tx_skbufs(priv); > > /* Free DMA regions of consistent memory previously allocated */ > - if (!priv->extend_desc) { > - dma_free_coherent(priv->device, > - DMA_TX_SIZE * sizeof(struct dma_desc), > - priv->dma_tx, priv->dma_tx_phy); > + if (!priv->extend_desc) > dma_free_coherent(priv->device, > DMA_RX_SIZE * sizeof(struct dma_desc), > priv->dma_rx, priv->dma_rx_phy); > - } else { > - dma_free_coherent(priv->device, DMA_TX_SIZE * > - sizeof(struct dma_extended_desc), > - priv->dma_etx, priv->dma_tx_phy); > + else > dma_free_coherent(priv->device, DMA_RX_SIZE * > sizeof(struct dma_extended_desc), > priv->dma_erx, priv->dma_rx_phy); > - } > + > kfree(priv->rx_skbuff_dma); > kfree(priv->rx_skbuff); > +} > + > +/** > + * free_dma_tx_desc_resources - free TX dma desc resources > + * @priv: private structure > + */ > +static void free_dma_tx_desc_resources(struct stmmac_priv *priv) > +{ > + /* Release the DMA TX socket buffers */ > + dma_free_tx_skbufs(priv); > + > + /* Free DMA regions of consistent memory previously allocated */ > + if (!priv->extend_desc) > + dma_free_coherent(priv->device, > + DMA_TX_SIZE * sizeof(struct dma_desc), > + priv->dma_tx, priv->dma_tx_phy); > + else > + dma_free_coherent(priv->device, DMA_TX_SIZE * > + sizeof(struct dma_extended_desc), > + priv->dma_etx, priv->dma_tx_phy); > + > kfree(priv->tx_skbuff_dma); > kfree(priv->tx_skbuff); > } > > /** > + * free_dma_desc_resources - free dma desc resources > + * @priv: private structure > + */ > +static void free_dma_desc_resources(struct stmmac_priv *priv) > +{ > + /* Release the DMA RX socket buffers */ > + free_dma_rx_desc_resources(priv); > + > + /* Release the DMA TX socket buffers */ > + free_dma_tx_desc_resources(priv); > +} > + > +/** > * stmmac_mac_enable_rx_queues - Enable MAC rx queues > * @priv: driver private structure > * Description: It is used for enabling the rx queues in the MAC >
diff --git a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c index 7cbda41..ff839e1 100644 --- a/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c +++ b/drivers/net/ethernet/stmicro/stmmac/stmmac_main.c @@ -889,24 +889,41 @@ static int stmmac_init_phy(struct net_device *dev) return 0; } -static void stmmac_display_rings(struct stmmac_priv *priv) +static void stmmac_display_rx_rings(struct stmmac_priv *priv) { - void *head_rx, *head_tx; + void *head_rx; - if (priv->extend_desc) { + if (priv->extend_desc) head_rx = (void *)priv->dma_erx; - head_tx = (void *)priv->dma_etx; - } else { + else head_rx = (void *)priv->dma_rx; - head_tx = (void *)priv->dma_tx; - } - /* Display Rx ring */ + /* Display RX ring */ priv->hw->desc->display_ring(head_rx, DMA_RX_SIZE, true); - /* Display Tx ring */ +} + +static void stmmac_display_tx_rings(struct stmmac_priv *priv) +{ + void *head_tx; + + if (priv->extend_desc) + head_tx = (void *)priv->dma_etx; + else + head_tx = (void *)priv->dma_tx; + + /* Display TX ring */ priv->hw->desc->display_ring(head_tx, DMA_TX_SIZE, false); } +static void stmmac_display_rings(struct stmmac_priv *priv) +{ + /* Display RX ring */ + stmmac_display_rx_rings(priv); + + /* Display TX ring */ + stmmac_display_tx_rings(priv); +} + static int stmmac_set_bfsize(int mtu, int bufsize) { int ret = bufsize; @@ -924,16 +941,16 @@ static int stmmac_set_bfsize(int mtu, int bufsize) } /** - * stmmac_clear_descriptors - clear descriptors + * stmmac_clear_rx_descriptors - clear RX descriptors * @priv: driver private structure - * Description: this function is called to clear the tx and rx descriptors + * Description: this function is called to clear the RX descriptors * in case of both basic and extended descriptors are used. */ -static void stmmac_clear_descriptors(struct stmmac_priv *priv) +static void stmmac_clear_rx_descriptors(struct stmmac_priv *priv) { int i; - /* Clear the Rx/Tx descriptors */ + /* Clear the RX descriptors */ for (i = 0; i < DMA_RX_SIZE; i++) if (priv->extend_desc) priv->hw->desc->init_rx_desc(&priv->dma_erx[i].basic, @@ -943,6 +960,19 @@ static void stmmac_clear_descriptors(struct stmmac_priv *priv) priv->hw->desc->init_rx_desc(&priv->dma_rx[i], priv->use_riwt, priv->mode, (i == DMA_RX_SIZE - 1)); +} + +/** + * stmmac_clear_tx_descriptors - clear tx descriptors + * @priv: driver private structure + * Description: this function is called to clear the TX descriptors + * in case of both basic and extended descriptors are used. + */ +static void stmmac_clear_tx_descriptors(struct stmmac_priv *priv) +{ + int i; + + /* Clear the TX descriptors */ for (i = 0; i < DMA_TX_SIZE; i++) if (priv->extend_desc) priv->hw->desc->init_tx_desc(&priv->dma_etx[i].basic, @@ -955,6 +985,21 @@ static void stmmac_clear_descriptors(struct stmmac_priv *priv) } /** + * stmmac_clear_descriptors - clear descriptors + * @priv: driver private structure + * Description: this function is called to clear the TX and RX descriptors + * in case of both basic and extended descriptors are used. + */ +static void stmmac_clear_descriptors(struct stmmac_priv *priv) +{ + /* Clear the RX descriptors */ + stmmac_clear_rx_descriptors(priv); + + /* Clear the TX descriptors */ + stmmac_clear_tx_descriptors(priv); +} + +/** * stmmac_init_rx_buffers - init the RX descriptor buffer. * @priv: driver private structure * @p: descriptor pointer @@ -996,7 +1041,12 @@ static int stmmac_init_rx_buffers(struct stmmac_priv *priv, struct dma_desc *p, return 0; } -static void stmmac_free_rx_buffers(struct stmmac_priv *priv, int i) +/** + * stmmac_free_rx_buffer - free RX dma buffers + * @priv: private structure + * @i: buffer index. + */ +static void stmmac_free_rx_buffer(struct stmmac_priv *priv, int i) { if (priv->rx_skbuff[i]) { dma_unmap_single(priv->device, priv->rx_skbuff_dma[i], @@ -1007,14 +1057,42 @@ static void stmmac_free_rx_buffers(struct stmmac_priv *priv, int i) } /** - * init_dma_desc_rings - init the RX/TX descriptor rings + * stmmac_free_tx_buffer - free RX dma buffers + * @priv: private structure + * @i: buffer index. + */ +static void stmmac_free_tx_buffer(struct stmmac_priv *priv, int i) +{ + if (priv->tx_skbuff_dma[i].buf) { + if (priv->tx_skbuff_dma[i].map_as_page) + dma_unmap_page(priv->device, + priv->tx_skbuff_dma[i].buf, + priv->tx_skbuff_dma[i].len, + DMA_TO_DEVICE); + else + dma_unmap_single(priv->device, + priv->tx_skbuff_dma[i].buf, + priv->tx_skbuff_dma[i].len, + DMA_TO_DEVICE); + } + + if (priv->tx_skbuff[i]) { + dev_kfree_skb_any(priv->tx_skbuff[i]); + priv->tx_skbuff[i] = NULL; + priv->tx_skbuff_dma[i].buf = 0; + priv->tx_skbuff_dma[i].map_as_page = false; + } +} + +/** + * init_dma_rx_desc_rings - init the RX descriptor rings * @dev: net device structure * @flags: gfp flag. - * Description: this function initializes the DMA RX/TX descriptors + * Description: this function initializes the DMA RX descriptors * and allocates the socket buffers. It supports the chained and ring * modes. */ -static int init_dma_desc_rings(struct net_device *dev, gfp_t flags) +static int init_dma_rx_desc_rings(struct net_device *dev, gfp_t flags) { int i; struct stmmac_priv *priv = netdev_priv(dev); @@ -1030,10 +1108,8 @@ static int init_dma_desc_rings(struct net_device *dev, gfp_t flags) priv->dma_buf_sz = bfsize; netif_dbg(priv, probe, priv->dev, - "(%s) dma_rx_phy=0x%08x dma_tx_phy=0x%08x\n", - __func__, (u32)priv->dma_rx_phy, (u32)priv->dma_tx_phy); + "(%s) dma_rx_phy=0x%08x\n", __func__, (u32)priv->dma_rx_phy); - /* RX INITIALIZATION */ netif_dbg(priv, probe, priv->dev, "SKB addresses:\nskb\t\tskb data\tdma data\n"); @@ -1058,20 +1134,46 @@ static int init_dma_desc_rings(struct net_device *dev, gfp_t flags) /* Setup the chained descriptor addresses */ if (priv->mode == STMMAC_CHAIN_MODE) { - if (priv->extend_desc) { + if (priv->extend_desc) priv->hw->mode->init(priv->dma_erx, priv->dma_rx_phy, DMA_RX_SIZE, 1); - priv->hw->mode->init(priv->dma_etx, priv->dma_tx_phy, - DMA_TX_SIZE, 1); - } else { + else priv->hw->mode->init(priv->dma_rx, priv->dma_rx_phy, DMA_RX_SIZE, 0); + } + + return 0; +err_init_rx_buffers: + while (--i >= 0) + stmmac_free_rx_buffer(priv, i); + return ret; +} + +/** + * init_dma_tx_desc_rings - init the TX descriptor rings + * @dev: net device structure. + * Description: this function initializes the DMA TX descriptors + * and allocates the socket buffers. It supports the chained and ring + * modes. + */ +static int init_dma_tx_desc_rings(struct net_device *dev) +{ + struct stmmac_priv *priv = netdev_priv(dev); + int i; + + netif_dbg(priv, probe, priv->dev, + "(%s) dma_tx_phy=0x%08x\n", __func__, (u32)priv->dma_tx_phy); + + /* Setup the chained descriptor addresses */ + if (priv->mode == STMMAC_CHAIN_MODE) { + if (priv->extend_desc) + priv->hw->mode->init(priv->dma_etx, priv->dma_tx_phy, + DMA_TX_SIZE, 1); + else priv->hw->mode->init(priv->dma_tx, priv->dma_tx_phy, DMA_TX_SIZE, 0); - } } - /* TX INITIALIZATION */ for (i = 0; i < DMA_TX_SIZE; i++) { struct dma_desc *p; if (priv->extend_desc) @@ -1099,62 +1201,69 @@ static int init_dma_desc_rings(struct net_device *dev, gfp_t flags) priv->cur_tx = 0; netdev_reset_queue(priv->dev); + return 0; +} + +/** + * init_dma_desc_rings - init the RX/TX descriptor rings + * @dev: net device structure + * @flags: gfp flag. + * Description: this function initializes the DMA RX/TX descriptors + * and allocates the socket buffers. It supports the chained and ring + * modes. + */ +static int init_dma_desc_rings(struct net_device *dev, gfp_t flags) +{ + struct stmmac_priv *priv = netdev_priv(dev); + int ret; + + ret = init_dma_rx_desc_rings(dev, flags); + if (ret) + return ret; + + ret = init_dma_tx_desc_rings(dev); + stmmac_clear_descriptors(priv); if (netif_msg_hw(priv)) stmmac_display_rings(priv); - return 0; -err_init_rx_buffers: - while (--i >= 0) - stmmac_free_rx_buffers(priv, i); return ret; } +/** + * dma_free_rx_skbufs - free RX dma buffers + * @priv: private structure + */ static void dma_free_rx_skbufs(struct stmmac_priv *priv) { int i; for (i = 0; i < DMA_RX_SIZE; i++) - stmmac_free_rx_buffers(priv, i); + stmmac_free_rx_buffer(priv, i); } +/** + * dma_free_tx_skbufs - free TX dma buffers + * @priv: private structure + */ static void dma_free_tx_skbufs(struct stmmac_priv *priv) { int i; - for (i = 0; i < DMA_TX_SIZE; i++) { - if (priv->tx_skbuff_dma[i].buf) { - if (priv->tx_skbuff_dma[i].map_as_page) - dma_unmap_page(priv->device, - priv->tx_skbuff_dma[i].buf, - priv->tx_skbuff_dma[i].len, - DMA_TO_DEVICE); - else - dma_unmap_single(priv->device, - priv->tx_skbuff_dma[i].buf, - priv->tx_skbuff_dma[i].len, - DMA_TO_DEVICE); - } - - if (priv->tx_skbuff[i]) { - dev_kfree_skb_any(priv->tx_skbuff[i]); - priv->tx_skbuff[i] = NULL; - priv->tx_skbuff_dma[i].buf = 0; - priv->tx_skbuff_dma[i].map_as_page = false; - } - } + for (i = 0; i < DMA_TX_SIZE; i++) + stmmac_free_tx_buffer(priv, i); } /** - * alloc_dma_desc_resources - alloc TX/RX resources. + * alloc_dma_rx_desc_resources - alloc RX resources. * @priv: private structure * Description: according to which descriptor can be used (extend or basic) * this function allocates the resources for TX and RX paths. In case of * reception, for example, it pre-allocated the RX socket buffer in order to * allow zero-copy mechanism. */ -static int alloc_dma_desc_resources(struct stmmac_priv *priv) +static int alloc_dma_rx_desc_resources(struct stmmac_priv *priv) { int ret = -ENOMEM; @@ -1168,11 +1277,50 @@ static int alloc_dma_desc_resources(struct stmmac_priv *priv) if (!priv->rx_skbuff) goto err_rx_skbuff; + if (priv->extend_desc) { + priv->dma_erx = dma_zalloc_coherent(priv->device, DMA_RX_SIZE * + sizeof(struct + dma_extended_desc), + &priv->dma_rx_phy, + GFP_KERNEL); + if (!priv->dma_erx) + goto err_dma; + + } else { + priv->dma_rx = dma_zalloc_coherent(priv->device, DMA_RX_SIZE * + sizeof(struct dma_desc), + &priv->dma_rx_phy, + GFP_KERNEL); + if (!priv->dma_rx) + goto err_dma; + } + + return 0; + +err_dma: + kfree(priv->rx_skbuff); +err_rx_skbuff: + kfree(priv->rx_skbuff_dma); + return ret; +} + +/** + * alloc_dma_tx_desc_resources - alloc TX resources. + * @priv: private structure + * Description: according to which descriptor can be used (extend or basic) + * this function allocates the resources for TX and RX paths. In case of + * reception, for example, it pre-allocated the RX socket buffer in order to + * allow zero-copy mechanism. + */ +static int alloc_dma_tx_desc_resources(struct stmmac_priv *priv) +{ + int ret = -ENOMEM; + priv->tx_skbuff_dma = kmalloc_array(DMA_TX_SIZE, sizeof(*priv->tx_skbuff_dma), GFP_KERNEL); if (!priv->tx_skbuff_dma) - goto err_tx_skbuff_dma; + return -ENOMEM; priv->tx_skbuff = kmalloc_array(DMA_TX_SIZE, sizeof(struct sk_buff *), GFP_KERNEL); @@ -1180,14 +1328,6 @@ static int alloc_dma_desc_resources(struct stmmac_priv *priv) goto err_tx_skbuff; if (priv->extend_desc) { - priv->dma_erx = dma_zalloc_coherent(priv->device, DMA_RX_SIZE * - sizeof(struct - dma_extended_desc), - &priv->dma_rx_phy, - GFP_KERNEL); - if (!priv->dma_erx) - goto err_dma; - priv->dma_etx = dma_zalloc_coherent(priv->device, DMA_TX_SIZE * sizeof(struct dma_extended_desc), @@ -1200,13 +1340,6 @@ static int alloc_dma_desc_resources(struct stmmac_priv *priv) goto err_dma; } } else { - priv->dma_rx = dma_zalloc_coherent(priv->device, DMA_RX_SIZE * - sizeof(struct dma_desc), - &priv->dma_rx_phy, - GFP_KERNEL); - if (!priv->dma_rx) - goto err_dma; - priv->dma_tx = dma_zalloc_coherent(priv->device, DMA_TX_SIZE * sizeof(struct dma_desc), &priv->dma_tx_phy, @@ -1225,42 +1358,89 @@ static int alloc_dma_desc_resources(struct stmmac_priv *priv) kfree(priv->tx_skbuff); err_tx_skbuff: kfree(priv->tx_skbuff_dma); -err_tx_skbuff_dma: - kfree(priv->rx_skbuff); -err_rx_skbuff: - kfree(priv->rx_skbuff_dma); return ret; } -static void free_dma_desc_resources(struct stmmac_priv *priv) +/** + * alloc_dma_desc_resources - alloc TX/RX resources. + * @priv: private structure + * Description: according to which descriptor can be used (extend or basic) + * this function allocates the resources for TX and RX paths. In case of + * reception, for example, it pre-allocated the RX socket buffer in order to + * allow zero-copy mechanism. + */ +static int alloc_dma_desc_resources(struct stmmac_priv *priv) { - /* Release the DMA TX/RX socket buffers */ + int ret = alloc_dma_rx_desc_resources(priv); + + if (ret) + return ret; + + ret = alloc_dma_tx_desc_resources(priv); + + return ret; +} + +/** + * free_dma_rx_desc_resources - free RX dma desc resources + * @priv: private structure + */ +static void free_dma_rx_desc_resources(struct stmmac_priv *priv) +{ + /* Release the DMA RX socket buffers */ dma_free_rx_skbufs(priv); - dma_free_tx_skbufs(priv); /* Free DMA regions of consistent memory previously allocated */ - if (!priv->extend_desc) { - dma_free_coherent(priv->device, - DMA_TX_SIZE * sizeof(struct dma_desc), - priv->dma_tx, priv->dma_tx_phy); + if (!priv->extend_desc) dma_free_coherent(priv->device, DMA_RX_SIZE * sizeof(struct dma_desc), priv->dma_rx, priv->dma_rx_phy); - } else { - dma_free_coherent(priv->device, DMA_TX_SIZE * - sizeof(struct dma_extended_desc), - priv->dma_etx, priv->dma_tx_phy); + else dma_free_coherent(priv->device, DMA_RX_SIZE * sizeof(struct dma_extended_desc), priv->dma_erx, priv->dma_rx_phy); - } + kfree(priv->rx_skbuff_dma); kfree(priv->rx_skbuff); +} + +/** + * free_dma_tx_desc_resources - free TX dma desc resources + * @priv: private structure + */ +static void free_dma_tx_desc_resources(struct stmmac_priv *priv) +{ + /* Release the DMA TX socket buffers */ + dma_free_tx_skbufs(priv); + + /* Free DMA regions of consistent memory previously allocated */ + if (!priv->extend_desc) + dma_free_coherent(priv->device, + DMA_TX_SIZE * sizeof(struct dma_desc), + priv->dma_tx, priv->dma_tx_phy); + else + dma_free_coherent(priv->device, DMA_TX_SIZE * + sizeof(struct dma_extended_desc), + priv->dma_etx, priv->dma_tx_phy); + kfree(priv->tx_skbuff_dma); kfree(priv->tx_skbuff); } /** + * free_dma_desc_resources - free dma desc resources + * @priv: private structure + */ +static void free_dma_desc_resources(struct stmmac_priv *priv) +{ + /* Release the DMA RX socket buffers */ + free_dma_rx_desc_resources(priv); + + /* Release the DMA TX socket buffers */ + free_dma_tx_desc_resources(priv); +} + +/** * stmmac_mac_enable_rx_queues - Enable MAC rx queues * @priv: driver private structure * Description: It is used for enabling the rx queues in the MAC
This patch breaks several functions into RX and TX scopes, which will be useful when adding multiple buffers mechanism. Signed-off-by: Joao Pinto <jpinto@synopsys.com> --- changes v2->v3: - just to keep up with patch-set version changes v1->v2: - RX and TX inconsistency - stmmac_free_rx_buffers renamed to stmmac_free_rx_buffer - stmmac_free_tx_buffers renamed to stmmac_free_tx_buffer - some useless comments were removed drivers/net/ethernet/stmicro/stmmac/stmmac_main.c | 352 ++++++++++++++++------ 1 file changed, 266 insertions(+), 86 deletions(-)