mlx4: reduce rx ring page_cache size
authorEric Dumazet <edumazet@google.com>
Wed, 8 Mar 2017 16:17:10 +0000 (08:17 -0800)
committerDavid S. Miller <davem@davemloft.net>
Thu, 9 Mar 2017 17:54:46 +0000 (09:54 -0800)
We only need to store the page and dma address.

Signed-off-by: Eric Dumazet <edumazet@google.com>
Acked-by: Tariq Toukan <tariqt@mellanox.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
drivers/net/ethernet/mellanox/mlx4/en_rx.c
drivers/net/ethernet/mellanox/mlx4/en_tx.c
drivers/net/ethernet/mellanox/mlx4/mlx4_en.h

index bb33032a280f00ee62cc39d4261e72543ed0434e..453313d404e3698b0d41a8220005b3834c9d68a1 100644 (file)
@@ -250,7 +250,10 @@ static int mlx4_en_prepare_rx_desc(struct mlx4_en_priv *priv,
                                        (index << priv->log_rx_info);
 
        if (ring->page_cache.index > 0) {
-               frags[0] = ring->page_cache.buf[--ring->page_cache.index];
+               ring->page_cache.index--;
+               frags[0].page = ring->page_cache.buf[ring->page_cache.index].page;
+               frags[0].dma  = ring->page_cache.buf[ring->page_cache.index].dma;
+               frags[0].page_offset = XDP_PACKET_HEADROOM;
                rx_desc->data[0].addr = cpu_to_be64(frags[0].dma +
                                                    frags[0].page_offset);
                return 0;
@@ -537,7 +540,9 @@ bool mlx4_en_rx_recycle(struct mlx4_en_rx_ring *ring,
        if (cache->index >= MLX4_EN_CACHE_SIZE)
                return false;
 
-       cache->buf[cache->index++] = *frame;
+       cache->buf[cache->index].page = frame->page;
+       cache->buf[cache->index].dma = frame->dma;
+       cache->index++;
        return true;
 }
 
@@ -567,11 +572,9 @@ void mlx4_en_deactivate_rx_ring(struct mlx4_en_priv *priv,
        int i;
 
        for (i = 0; i < ring->page_cache.index; i++) {
-               struct mlx4_en_rx_alloc *frame = &ring->page_cache.buf[i];
-
-               dma_unmap_page(priv->ddev, frame->dma, frame->page_size,
-                              priv->dma_dir);
-               put_page(frame->page);
+               dma_unmap_page(priv->ddev, ring->page_cache.buf[i].dma,
+                              PAGE_SIZE, priv->dma_dir);
+               put_page(ring->page_cache.buf[i].page);
        }
        ring->page_cache.index = 0;
        mlx4_en_free_rx_buf(priv, ring);
index 98bc67a7249b14f8857fe1fd6baa40ae3ec5a880..e0c5ffb3e3a6607456e1f191b0b8c8becfc71219 100644 (file)
@@ -354,8 +354,6 @@ u32 mlx4_en_recycle_tx_desc(struct mlx4_en_priv *priv,
        struct mlx4_en_rx_alloc frame = {
                .page = tx_info->page,
                .dma = tx_info->map0_dma,
-               .page_offset = XDP_PACKET_HEADROOM,
-               .page_size = PAGE_SIZE,
        };
 
        if (!mlx4_en_rx_recycle(ring->recycle_ring, &frame)) {
index fc7b4da5d8c376455c642d9512a8565d88db42a3..c98940c6acd02c1bd8c4f82107d82979ead051d1 100644 (file)
@@ -268,9 +268,13 @@ struct mlx4_en_rx_alloc {
 };
 
 #define MLX4_EN_CACHE_SIZE (2 * NAPI_POLL_WEIGHT)
+
 struct mlx4_en_page_cache {
        u32 index;
-       struct mlx4_en_rx_alloc buf[MLX4_EN_CACHE_SIZE];
+       struct {
+               struct page     *page;
+               dma_addr_t      dma;
+       } buf[MLX4_EN_CACHE_SIZE];
 };
 
 struct mlx4_en_priv;