Using the NAPI or netdev frag cache along with other drivers can lead to
32 KiB pages being held for a long time, despite only being used for
very few page fragment.
This can happen if the ethernet driver grabs one or two fragments for rx
ring refill, while other drivers use (and free up) the remaining
fragments. The 32 KiB higher-order page can only be freed once all users
have freed their fragments, which only happens after the rings of all
drivers holding the fragments have wrapped around.
Depending on the traffic patterns, this can waste a lot of memory and
look a lot like a memory leak
Signed-off-by: Felix Fietkau <nbd@nbd.name>
static void fe_clean_rx(struct fe_priv *priv)
{
static void fe_clean_rx(struct fe_priv *priv)
{
struct fe_rx_ring *ring = &priv->rx_ring;
struct fe_rx_ring *ring = &priv->rx_ring;
+ struct page *page;
+ int i;
if (ring->rx_data) {
for (i = 0; i < ring->rx_ring_size; i++)
if (ring->rx_data) {
for (i = 0; i < ring->rx_ring_size; i++)
ring->rx_phys);
ring->rx_dma = NULL;
}
ring->rx_phys);
ring->rx_dma = NULL;
}
+
+ if (!ring->frag_cache.va)
+ return;
+
+ page = virt_to_page(ring->frag_cache.va);
+ __page_frag_cache_drain(page, ring->frag_cache.pagecnt_bias);
+ memset(&ring->frag_cache, 0, sizeof(ring->frag_cache));
}
static int fe_alloc_rx(struct fe_priv *priv)
}
static int fe_alloc_rx(struct fe_priv *priv)
goto no_rx_mem;
for (i = 0; i < ring->rx_ring_size; i++) {
goto no_rx_mem;
for (i = 0; i < ring->rx_ring_size; i++) {
- ring->rx_data[i] = netdev_alloc_frag(ring->frag_size);
+ ring->rx_data[i] = page_frag_alloc(&ring->frag_cache,
+ ring->frag_size,
+ GFP_KERNEL);
if (!ring->rx_data[i])
goto no_rx_mem;
}
if (!ring->rx_data[i])
goto no_rx_mem;
}
break;
/* alloc new buffer */
break;
/* alloc new buffer */
- new_data = netdev_alloc_frag(ring->frag_size);
+ new_data = page_frag_alloc(&ring->frag_cache, ring->frag_size,
+ GFP_ATOMIC);
if (unlikely(!new_data)) {
stats->rx_dropped++;
goto release_desc;
if (unlikely(!new_data)) {
stats->rx_dropped++;
goto release_desc;
+ struct page_frag_cache frag_cache;
struct fe_rx_dma *rx_dma;
u8 **rx_data;
dma_addr_t rx_phys;
struct fe_rx_dma *rx_dma;
u8 **rx_data;
dma_addr_t rx_phys;