netmem_set_pp(netmem, NULL);
 }
 
-static __always_inline void __page_pool_release_page_dma(struct page_pool *pool,
-                                                        netmem_ref netmem)
+static __always_inline void __page_pool_release_netmem_dma(struct page_pool *pool,
+                                                          netmem_ref netmem)
 {
        struct page *old, *page = netmem_to_page(netmem);
        unsigned long id;
        if (static_branch_unlikely(&page_pool_mem_providers) && pool->mp_ops)
                put = pool->mp_ops->release_netmem(pool, netmem);
        else
-               __page_pool_release_page_dma(pool, netmem);
+               __page_pool_release_netmem_dma(pool, netmem);
 
        /* This may be the last page returned, releasing the pool, so
         * it is not safe to reference pool afterwards.
                }
 
                xa_for_each(&pool->dma_mapped, id, ptr)
-                       __page_pool_release_page_dma(pool, page_to_netmem(ptr));
+                       __page_pool_release_netmem_dma(pool, page_to_netmem((struct page *)ptr));
        }
 
        /* No more consumers should exist, but producers could still