net/mlx5e: RX, Remove unnecessary recycle parameter and page_cache stats
authorDragos Tatulea <dtatulea@nvidia.com>
Wed, 22 Feb 2023 09:27:38 +0000 (11:27 +0200)
committerSaeed Mahameed <saeedm@nvidia.com>
Tue, 28 Mar 2023 20:43:59 +0000 (13:43 -0700)
The recycle parameter used during page release is no longer
necessary: the page pool can detect when the page cannot be
recycled to the cache or ring without any outside hint.

The page pool will also take care of cleaning up after itself
once all the inflight pages have been released. So no need to
explicitly release pages to the system.

Remove the internal page_cache stats as the mlx5e_page_cache
struct no longer exists.

Delete the documentation entries along with the stats.

Signed-off-by: Dragos Tatulea <dtatulea@nvidia.com>
Reviewed-by: Tariq Toukan <tariqt@nvidia.com>
Signed-off-by: Saeed Mahameed <saeedm@nvidia.com>
Documentation/networking/device_drivers/ethernet/mellanox/mlx5/counters.rst
drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c
drivers/net/ethernet/mellanox/mlx5/core/en_rx.c
drivers/net/ethernet/mellanox/mlx5/core/en_stats.c
drivers/net/ethernet/mellanox/mlx5/core/en_stats.h

index 4cd8e869762bf7505d066b8f4885112e95ddedbe..6b2d1fe74ecf24634a05afeb884c6335ce08540f 100644 (file)
@@ -346,32 +346,6 @@ the software port.
      - The number of receive packets with CQE compression on ring i [#accel]_.
      - Acceleration
 
-   * - `rx[i]_cache_reuse`
-     - The number of events of successful reuse of a page from a driver's
-       internal page cache.
-     - Acceleration
-
-   * - `rx[i]_cache_full`
-     - The number of events of full internal page cache where driver can't put a
-       page back to the cache for recycling (page will be freed).
-     - Acceleration
-
-   * - `rx[i]_cache_empty`
-     - The number of events where cache was empty - no page to give. Driver
-       shall allocate new page.
-     - Acceleration
-
-   * - `rx[i]_cache_busy`
-     - The number of events where cache head was busy and cannot be recycled.
-       Driver allocated new page.
-     - Acceleration
-
-   * - `rx[i]_cache_waive`
-     - The number of cache evacuation. This can occur due to page move to
-       another NUMA node or page was pfmemalloc-ed and should be freed as soon
-       as possible.
-     - Acceleration
-
    * - `rx[i]_arfs_err`
      - Number of flow rules that failed to be added to the flow table.
      - Error
index ca6ac9772d22bbc5d43ff1149cef447d7439f078..15d15d0e5ef9f07b97945ea69492f4e6ccbc4114 100644 (file)
@@ -505,7 +505,6 @@ mlx5e_xmit_xdp_frame(struct mlx5e_xdpsq *sq, struct mlx5e_xmit_data *xdptxd,
 static void mlx5e_free_xdpsq_desc(struct mlx5e_xdpsq *sq,
                                  struct mlx5e_xdp_wqe_info *wi,
                                  u32 *xsk_frames,
-                                 bool recycle,
                                  struct xdp_frame_bulk *bq)
 {
        struct mlx5e_xdp_info_fifo *xdpi_fifo = &sq->db.xdpi_fifo;
@@ -524,7 +523,7 @@ static void mlx5e_free_xdpsq_desc(struct mlx5e_xdpsq *sq,
                case MLX5E_XDP_XMIT_MODE_PAGE:
                        /* XDP_TX from the regular RQ */
                        page_pool_put_defragged_page(xdpi.page.rq->page_pool,
-                                                    xdpi.page.page, -1, recycle);
+                                                    xdpi.page.page, -1, true);
                        break;
                case MLX5E_XDP_XMIT_MODE_XSK:
                        /* AF_XDP send */
@@ -578,7 +577,7 @@ bool mlx5e_poll_xdpsq_cq(struct mlx5e_cq *cq)
 
                        sqcc += wi->num_wqebbs;
 
-                       mlx5e_free_xdpsq_desc(sq, wi, &xsk_frames, true, &bq);
+                       mlx5e_free_xdpsq_desc(sq, wi, &xsk_frames, &bq);
                } while (!last_wqe);
 
                if (unlikely(get_cqe_opcode(cqe) != MLX5_CQE_REQ)) {
@@ -625,7 +624,7 @@ void mlx5e_free_xdpsq_descs(struct mlx5e_xdpsq *sq)
 
                sq->cc += wi->num_wqebbs;
 
-               mlx5e_free_xdpsq_desc(sq, wi, &xsk_frames, false, &bq);
+               mlx5e_free_xdpsq_desc(sq, wi, &xsk_frames, &bq);
        }
 
        xdp_flush_frame_bulk(&bq);
index df5dbef9e5ec00f195879ea64b22f56933fe9c28..1049805571c6ec51f3d3f2f0372e1b78b7392f04 100644 (file)
@@ -293,14 +293,13 @@ static int mlx5e_page_alloc_fragmented(struct mlx5e_rq *rq,
 }
 
 static void mlx5e_page_release_fragmented(struct mlx5e_rq *rq,
-                                         struct mlx5e_frag_page *frag_page,
-                                         bool recycle)
+                                         struct mlx5e_frag_page *frag_page)
 {
        u16 drain_count = MLX5E_PAGECNT_BIAS_MAX - frag_page->frags;
        struct page *page = frag_page->page;
 
        if (page_pool_defrag_page(page, drain_count) == 0)
-               page_pool_put_defragged_page(rq->page_pool, page, -1, recycle);
+               page_pool_put_defragged_page(rq->page_pool, page, -1, true);
 }
 
 static inline int mlx5e_get_rx_frag(struct mlx5e_rq *rq,
@@ -330,11 +329,10 @@ static bool mlx5e_frag_can_release(struct mlx5e_wqe_frag_info *frag)
 }
 
 static inline void mlx5e_put_rx_frag(struct mlx5e_rq *rq,
-                                    struct mlx5e_wqe_frag_info *frag,
-                                    bool recycle)
+                                    struct mlx5e_wqe_frag_info *frag)
 {
        if (mlx5e_frag_can_release(frag))
-               mlx5e_page_release_fragmented(rq, frag->frag_page, recycle);
+               mlx5e_page_release_fragmented(rq, frag->frag_page);
 }
 
 static inline struct mlx5e_wqe_frag_info *get_frag(struct mlx5e_rq *rq, u16 ix)
@@ -368,19 +366,18 @@ static int mlx5e_alloc_rx_wqe(struct mlx5e_rq *rq, struct mlx5e_rx_wqe_cyc *wqe,
 
 free_frags:
        while (--i >= 0)
-               mlx5e_put_rx_frag(rq, --frag, true);
+               mlx5e_put_rx_frag(rq, --frag);
 
        return err;
 }
 
 static inline void mlx5e_free_rx_wqe(struct mlx5e_rq *rq,
-                                    struct mlx5e_wqe_frag_info *wi,
-                                    bool recycle)
+                                    struct mlx5e_wqe_frag_info *wi)
 {
        int i;
 
        for (i = 0; i < rq->wqe.info.num_frags; i++, wi++)
-               mlx5e_put_rx_frag(rq, wi, recycle);
+               mlx5e_put_rx_frag(rq, wi);
 }
 
 static void mlx5e_xsk_free_rx_wqe(struct mlx5e_wqe_frag_info *wi)
@@ -396,7 +393,7 @@ static void mlx5e_dealloc_rx_wqe(struct mlx5e_rq *rq, u16 ix)
        if (rq->xsk_pool)
                mlx5e_xsk_free_rx_wqe(wi);
        else
-               mlx5e_free_rx_wqe(rq, wi, false);
+               mlx5e_free_rx_wqe(rq, wi);
 }
 
 static void mlx5e_xsk_free_rx_wqes(struct mlx5e_rq *rq, u16 ix, int wqe_bulk)
@@ -427,7 +424,7 @@ static void mlx5e_free_rx_wqes(struct mlx5e_rq *rq, u16 ix, int wqe_bulk)
                struct mlx5e_wqe_frag_info *wi;
 
                wi = get_frag(rq, j);
-               mlx5e_free_rx_wqe(rq, wi, true);
+               mlx5e_free_rx_wqe(rq, wi);
        }
 }
 
@@ -502,7 +499,7 @@ mlx5e_copy_skb_header(struct mlx5e_rq *rq, struct sk_buff *skb,
 }
 
 static void
-mlx5e_free_rx_mpwqe(struct mlx5e_rq *rq, struct mlx5e_mpw_info *wi, bool recycle)
+mlx5e_free_rx_mpwqe(struct mlx5e_rq *rq, struct mlx5e_mpw_info *wi)
 {
        bool no_xdp_xmit;
        int i;
@@ -516,9 +513,9 @@ mlx5e_free_rx_mpwqe(struct mlx5e_rq *rq, struct mlx5e_mpw_info *wi, bool recycle
        if (rq->xsk_pool) {
                struct xdp_buff **xsk_buffs = wi->alloc_units.xsk_buffs;
 
-               /* The `recycle` parameter is ignored, and the page is always
-                * put into the Reuse Ring, because there is no way to return
-                * the page to the userspace when the interface goes down.
+               /* The page is always put into the Reuse Ring, because there
+                * is no way to return the page to userspace when the interface
+                * goes down.
                 */
                for (i = 0; i < rq->mpwqe.pages_per_wqe; i++)
                        if (no_xdp_xmit || !test_bit(i, wi->skip_release_bitmap))
@@ -529,7 +526,7 @@ mlx5e_free_rx_mpwqe(struct mlx5e_rq *rq, struct mlx5e_mpw_info *wi, bool recycle
                                struct mlx5e_frag_page *frag_page;
 
                                frag_page = &wi->alloc_units.frag_pages[i];
-                               mlx5e_page_release_fragmented(rq, frag_page, recycle);
+                               mlx5e_page_release_fragmented(rq, frag_page);
                        }
                }
        }
@@ -663,7 +660,7 @@ err_unmap:
                dma_info = &shampo->info[--index];
                if (!(i & (MLX5E_SHAMPO_WQ_HEADER_PER_PAGE - 1))) {
                        dma_info->addr = ALIGN_DOWN(dma_info->addr, PAGE_SIZE);
-                       mlx5e_page_release_fragmented(rq, dma_info->frag_page, true);
+                       mlx5e_page_release_fragmented(rq, dma_info->frag_page);
                }
        }
        rq->stats->buff_alloc_err++;
@@ -781,7 +778,7 @@ static int mlx5e_alloc_rx_mpwqe(struct mlx5e_rq *rq, u16 ix)
 err_unmap:
        while (--i >= 0) {
                frag_page--;
-               mlx5e_page_release_fragmented(rq, frag_page, true);
+               mlx5e_page_release_fragmented(rq, frag_page);
        }
 
 err:
@@ -815,7 +812,7 @@ void mlx5e_shampo_dealloc_hd(struct mlx5e_rq *rq, u16 len, u16 start, bool close
                hd_info->addr = ALIGN_DOWN(hd_info->addr, PAGE_SIZE);
                if (hd_info->frag_page && hd_info->frag_page != deleted_page) {
                        deleted_page = hd_info->frag_page;
-                       mlx5e_page_release_fragmented(rq, hd_info->frag_page, false);
+                       mlx5e_page_release_fragmented(rq, hd_info->frag_page);
                }
 
                hd_info->frag_page = NULL;
@@ -833,8 +830,8 @@ void mlx5e_shampo_dealloc_hd(struct mlx5e_rq *rq, u16 len, u16 start, bool close
 static void mlx5e_dealloc_rx_mpwqe(struct mlx5e_rq *rq, u16 ix)
 {
        struct mlx5e_mpw_info *wi = mlx5e_get_mpw_info(rq, ix);
-       /* Don't recycle, this function is called on rq/netdev close */
-       mlx5e_free_rx_mpwqe(rq, wi, false);
+       /* This function is called on rq/netdev close. */
+       mlx5e_free_rx_mpwqe(rq, wi);
 }
 
 INDIRECT_CALLABLE_SCOPE bool mlx5e_post_rx_wqes(struct mlx5e_rq *rq)
@@ -1058,7 +1055,7 @@ INDIRECT_CALLABLE_SCOPE bool mlx5e_post_rx_mpwqes(struct mlx5e_rq *rq)
                struct mlx5e_mpw_info *wi = mlx5e_get_mpw_info(rq, head);
 
                /* Deferred free for better page pool cache usage. */
-               mlx5e_free_rx_mpwqe(rq, wi, true);
+               mlx5e_free_rx_mpwqe(rq, wi);
 
                alloc_err = rq->xsk_pool ? mlx5e_xsk_alloc_rx_mpwqe(rq, head) :
                                           mlx5e_alloc_rx_mpwqe(rq, head);
@@ -1739,7 +1736,7 @@ mlx5e_skb_from_cqe_nonlinear(struct mlx5e_rq *rq, struct mlx5e_wqe_frag_info *wi
                        int i;
 
                        for (i = wi - head_wi; i < rq->wqe.info.num_frags; i++)
-                               mlx5e_put_rx_frag(rq, &head_wi[i], true);
+                               mlx5e_put_rx_frag(rq, &head_wi[i]);
                }
                return NULL; /* page/packet was consumed by XDP */
        }
@@ -2158,7 +2155,7 @@ mlx5e_free_rx_shampo_hd_entry(struct mlx5e_rq *rq, u16 header_index)
                struct mlx5e_dma_info *dma_info = &shampo->info[header_index];
 
                dma_info->addr = ALIGN_DOWN(addr, PAGE_SIZE);
-               mlx5e_page_release_fragmented(rq, dma_info->frag_page, true);
+               mlx5e_page_release_fragmented(rq, dma_info->frag_page);
        }
        bitmap_clear(shampo->bitmap, header_index, 1);
 }
index 4478223c172094570d11be942c313eaba0761ae0..f1d9596905c6c8dad60be9e8c454f47c0f24dbd4 100644 (file)
@@ -179,11 +179,6 @@ static const struct counter_desc sw_stats_desc[] = {
        { MLX5E_DECLARE_STAT(struct mlx5e_sw_stats, rx_buff_alloc_err) },
        { MLX5E_DECLARE_STAT(struct mlx5e_sw_stats, rx_cqe_compress_blks) },
        { MLX5E_DECLARE_STAT(struct mlx5e_sw_stats, rx_cqe_compress_pkts) },
-       { MLX5E_DECLARE_STAT(struct mlx5e_sw_stats, rx_cache_reuse) },
-       { MLX5E_DECLARE_STAT(struct mlx5e_sw_stats, rx_cache_full) },
-       { MLX5E_DECLARE_STAT(struct mlx5e_sw_stats, rx_cache_empty) },
-       { MLX5E_DECLARE_STAT(struct mlx5e_sw_stats, rx_cache_busy) },
-       { MLX5E_DECLARE_STAT(struct mlx5e_sw_stats, rx_cache_waive) },
        { MLX5E_DECLARE_STAT(struct mlx5e_sw_stats, rx_congst_umr) },
        { MLX5E_DECLARE_STAT(struct mlx5e_sw_stats, rx_arfs_err) },
        { MLX5E_DECLARE_STAT(struct mlx5e_sw_stats, rx_recover) },
@@ -358,11 +353,6 @@ static void mlx5e_stats_grp_sw_update_stats_rq_stats(struct mlx5e_sw_stats *s,
        s->rx_buff_alloc_err          += rq_stats->buff_alloc_err;
        s->rx_cqe_compress_blks       += rq_stats->cqe_compress_blks;
        s->rx_cqe_compress_pkts       += rq_stats->cqe_compress_pkts;
-       s->rx_cache_reuse             += rq_stats->cache_reuse;
-       s->rx_cache_full              += rq_stats->cache_full;
-       s->rx_cache_empty             += rq_stats->cache_empty;
-       s->rx_cache_busy              += rq_stats->cache_busy;
-       s->rx_cache_waive             += rq_stats->cache_waive;
        s->rx_congst_umr              += rq_stats->congst_umr;
        s->rx_arfs_err                += rq_stats->arfs_err;
        s->rx_recover                 += rq_stats->recover;
@@ -1978,11 +1968,6 @@ static const struct counter_desc rq_stats_desc[] = {
        { MLX5E_DECLARE_RX_STAT(struct mlx5e_rq_stats, buff_alloc_err) },
        { MLX5E_DECLARE_RX_STAT(struct mlx5e_rq_stats, cqe_compress_blks) },
        { MLX5E_DECLARE_RX_STAT(struct mlx5e_rq_stats, cqe_compress_pkts) },
-       { MLX5E_DECLARE_RX_STAT(struct mlx5e_rq_stats, cache_reuse) },
-       { MLX5E_DECLARE_RX_STAT(struct mlx5e_rq_stats, cache_full) },
-       { MLX5E_DECLARE_RX_STAT(struct mlx5e_rq_stats, cache_empty) },
-       { MLX5E_DECLARE_RX_STAT(struct mlx5e_rq_stats, cache_busy) },
-       { MLX5E_DECLARE_RX_STAT(struct mlx5e_rq_stats, cache_waive) },
        { MLX5E_DECLARE_RX_STAT(struct mlx5e_rq_stats, congst_umr) },
        { MLX5E_DECLARE_RX_STAT(struct mlx5e_rq_stats, arfs_err) },
        { MLX5E_DECLARE_RX_STAT(struct mlx5e_rq_stats, recover) },
@@ -2163,11 +2148,6 @@ static const struct counter_desc ptp_rq_stats_desc[] = {
        { MLX5E_DECLARE_PTP_RQ_STAT(struct mlx5e_rq_stats, buff_alloc_err) },
        { MLX5E_DECLARE_PTP_RQ_STAT(struct mlx5e_rq_stats, cqe_compress_blks) },
        { MLX5E_DECLARE_PTP_RQ_STAT(struct mlx5e_rq_stats, cqe_compress_pkts) },
-       { MLX5E_DECLARE_PTP_RQ_STAT(struct mlx5e_rq_stats, cache_reuse) },
-       { MLX5E_DECLARE_PTP_RQ_STAT(struct mlx5e_rq_stats, cache_full) },
-       { MLX5E_DECLARE_PTP_RQ_STAT(struct mlx5e_rq_stats, cache_empty) },
-       { MLX5E_DECLARE_PTP_RQ_STAT(struct mlx5e_rq_stats, cache_busy) },
-       { MLX5E_DECLARE_PTP_RQ_STAT(struct mlx5e_rq_stats, cache_waive) },
        { MLX5E_DECLARE_PTP_RQ_STAT(struct mlx5e_rq_stats, congst_umr) },
        { MLX5E_DECLARE_PTP_RQ_STAT(struct mlx5e_rq_stats, arfs_err) },
        { MLX5E_DECLARE_PTP_RQ_STAT(struct mlx5e_rq_stats, recover) },
index b77100b60b5053db92117738433ffa1989fc6f4f..1ff8a06027dcfc4987a01b24dfd4e2740b15ca23 100644 (file)
@@ -193,11 +193,6 @@ struct mlx5e_sw_stats {
        u64 rx_buff_alloc_err;
        u64 rx_cqe_compress_blks;
        u64 rx_cqe_compress_pkts;
-       u64 rx_cache_reuse;
-       u64 rx_cache_full;
-       u64 rx_cache_empty;
-       u64 rx_cache_busy;
-       u64 rx_cache_waive;
        u64 rx_congst_umr;
        u64 rx_arfs_err;
        u64 rx_recover;
@@ -362,11 +357,6 @@ struct mlx5e_rq_stats {
        u64 buff_alloc_err;
        u64 cqe_compress_blks;
        u64 cqe_compress_pkts;
-       u64 cache_reuse;
-       u64 cache_full;
-       u64 cache_empty;
-       u64 cache_busy;
-       u64 cache_waive;
        u64 congst_umr;
        u64 arfs_err;
        u64 recover;