return page;
 }
 
-static inline u8 *__bnxt_alloc_rx_data(struct bnxt *bp, dma_addr_t *mapping,
+static inline u8 *__bnxt_alloc_rx_frag(struct bnxt *bp, dma_addr_t *mapping,
                                       gfp_t gfp)
 {
        u8 *data;
        struct pci_dev *pdev = bp->pdev;
 
-       data = kmalloc(bp->rx_buf_size, gfp);
+       if (gfp == GFP_ATOMIC)
+               data = napi_alloc_frag(bp->rx_buf_size);
+       else
+               data = netdev_alloc_frag(bp->rx_buf_size);
        if (!data)
                return NULL;
 
                                        DMA_ATTR_WEAK_ORDERING);
 
        if (dma_mapping_error(&pdev->dev, *mapping)) {
-               kfree(data);
+               skb_free_frag(data);
                data = NULL;
        }
        return data;
                rx_buf->data = page;
                rx_buf->data_ptr = page_address(page) + bp->rx_offset;
        } else {
-               u8 *data = __bnxt_alloc_rx_data(bp, &mapping, gfp);
+               u8 *data = __bnxt_alloc_rx_frag(bp, &mapping, gfp);
 
                if (!data)
                        return -ENOMEM;
                return NULL;
        }
 
-       skb = build_skb(data, 0);
+       skb = build_skb(data, bp->rx_buf_size);
        dma_unmap_single_attrs(&bp->pdev->dev, dma_addr, bp->rx_buf_use_size,
                               bp->rx_dir, DMA_ATTR_WEAK_ORDERING);
        if (!skb) {
-               kfree(data);
+               skb_free_frag(data);
                return NULL;
        }
 
                u8 *new_data;
                dma_addr_t new_mapping;
 
-               new_data = __bnxt_alloc_rx_data(bp, &new_mapping, GFP_ATOMIC);
+               new_data = __bnxt_alloc_rx_frag(bp, &new_mapping, GFP_ATOMIC);
                if (!new_data) {
                        bnxt_abort_tpa(cpr, idx, agg_bufs);
                        cpr->sw_stats.rx.rx_oom_discards += 1;
                tpa_info->data_ptr = new_data + bp->rx_offset;
                tpa_info->mapping = new_mapping;
 
-               skb = build_skb(data, 0);
+               skb = build_skb(data, bp->rx_buf_size);
                dma_unmap_single_attrs(&bp->pdev->dev, mapping,
                                       bp->rx_buf_use_size, bp->rx_dir,
                                       DMA_ATTR_WEAK_ORDERING);
 
                if (!skb) {
-                       kfree(data);
+                       skb_free_frag(data);
                        bnxt_abort_tpa(cpr, idx, agg_bufs);
                        cpr->sw_stats.rx.rx_oom_discards += 1;
                        return NULL;
 
                tpa_info->data = NULL;
 
-               kfree(data);
+               skb_free_frag(data);
        }
 
 skip_rx_tpa_free:
                        dma_unmap_single_attrs(&pdev->dev, mapping,
                                               bp->rx_buf_use_size, bp->rx_dir,
                                               DMA_ATTR_WEAK_ORDERING);
-                       kfree(data);
+                       skb_free_frag(data);
                }
        }
 
                u8 *data;
 
                for (i = 0; i < bp->max_tpa; i++) {
-                       data = __bnxt_alloc_rx_data(bp, &mapping, GFP_KERNEL);
+                       data = __bnxt_alloc_rx_frag(bp, &mapping, GFP_KERNEL);
                        if (!data)
                                return -ENOMEM;