ice: use prefetch methods
authorJesse Brandeburg <jesse.brandeburg@intel.com>
Wed, 27 Oct 2021 19:38:36 +0000 (12:38 -0700)
committerTony Nguyen <anthony.l.nguyen@intel.com>
Wed, 15 Dec 2021 16:46:28 +0000 (08:46 -0800)
The kernel provides some prefetch mechanisms to speed up commonly
cold cache line accesses during receive processing. Since these are
software structures it helps to have these strategically placed
prefetches.

Be careful to call BQL prefetch complete only for non XDP queues.

Co-developed-by: Piotr Raczynski <piotr.raczynski@intel.com>
Signed-off-by: Piotr Raczynski <piotr.raczynski@intel.com>
Signed-off-by: Jesse Brandeburg <jesse.brandeburg@intel.com>
Tested-by: Gurucharan G <gurucharanx.g@intel.com>
Signed-off-by: Tony Nguyen <anthony.l.nguyen@intel.com>
drivers/net/ethernet/intel/ice/ice_txrx.c

index 12a2edd13877a1a40b4c564c018260ffe75e4a52..de9247d45c392e58d0abee1e4bd19e7d14c1a244 100644 (file)
@@ -3,8 +3,9 @@
 
 /* The driver transmit and receive code */
 
-#include <linux/prefetch.h>
 #include <linux/mm.h>
+#include <linux/netdevice.h>
+#include <linux/prefetch.h>
 #include <linux/bpf_trace.h>
 #include <net/dsfield.h>
 #include <net/xdp.h>
@@ -219,6 +220,10 @@ static bool ice_clean_tx_irq(struct ice_tx_ring *tx_ring, int napi_budget)
        struct ice_tx_desc *tx_desc;
        struct ice_tx_buf *tx_buf;
 
+       /* get the bql data ready */
+       if (!ice_ring_is_xdp(tx_ring))
+               netdev_txq_bql_complete_prefetchw(txring_txq(tx_ring));
+
        tx_buf = &tx_ring->tx_buf[i];
        tx_desc = ICE_TX_DESC(tx_ring, i);
        i -= tx_ring->count;
@@ -232,6 +237,9 @@ static bool ice_clean_tx_irq(struct ice_tx_ring *tx_ring, int napi_budget)
                if (!eop_desc)
                        break;
 
+               /* follow the guidelines of other drivers */
+               prefetchw(&tx_buf->skb->users);
+
                smp_rmb();      /* prevent any other reads prior to eop_desc */
 
                ice_trace(clean_tx_irq, tx_ring, tx_desc, tx_buf);
@@ -2265,6 +2273,9 @@ ice_xmit_frame_ring(struct sk_buff *skb, struct ice_tx_ring *tx_ring)
                return NETDEV_TX_BUSY;
        }
 
+       /* prefetch for bql data which is infrequently used */
+       netdev_txq_bql_enqueue_prefetchw(txring_txq(tx_ring));
+
        offload.tx_ring = tx_ring;
 
        /* record the location of the first descriptor for this packet */