/* Slow/Control-path information follows */
        struct page_pool_params_slow slow;
+       /* User-facing fields, protected by page_pools_lock */
+       struct {
+               u32 id;
+       } user;
 };
 
 struct page *page_pool_alloc_pages(struct page_pool *pool, gfp_t gfp);
 
 obj-$(CONFIG_NETDEV_ADDR_LIST_TEST) += dev_addr_lists_test.o
 
 obj-y += net-sysfs.o
-obj-$(CONFIG_PAGE_POOL) += page_pool.o
+obj-$(CONFIG_PAGE_POOL) += page_pool.o page_pool_user.o
 obj-$(CONFIG_PROC_FS) += net-procfs.o
 obj-$(CONFIG_NET_PKTGEN) += pktgen.o
 obj-$(CONFIG_NETPOLL) += netpoll.o
 
 
 #include <trace/events/page_pool.h>
 
+#include "page_pool_priv.h"
+
 #define DEFER_TIME (msecs_to_jiffies(1000))
 #define DEFER_WARN_INTERVAL (60 * HZ)
 
                return ERR_PTR(-ENOMEM);
 
        err = page_pool_init(pool, params);
-       if (err < 0) {
-               pr_warn("%s() gave up with errno %d\n", __func__, err);
-               kfree(pool);
-               return ERR_PTR(err);
-       }
+       if (err < 0)
+               goto err_free;
+
+       err = page_pool_list(pool);
+       if (err)
+               goto err_uninit;
 
        return pool;
+
+err_uninit:
+       page_pool_uninit(pool);
+err_free:
+       pr_warn("%s() gave up with errno %d\n", __func__, err);
+       kfree(pool);
+       return ERR_PTR(err);
 }
 EXPORT_SYMBOL(page_pool_create);
 
        if (pool->disconnect)
                pool->disconnect(pool);
 
+       page_pool_unlist(pool);
        page_pool_uninit(pool);
        kfree(pool);
 }
 
--- /dev/null
+/* SPDX-License-Identifier: GPL-2.0 */
+
+#ifndef __PAGE_POOL_PRIV_H
+#define __PAGE_POOL_PRIV_H
+
+int page_pool_list(struct page_pool *pool);
+void page_pool_unlist(struct page_pool *pool);
+
+#endif
 
--- /dev/null
+// SPDX-License-Identifier: GPL-2.0
+
+#include <linux/mutex.h>
+#include <linux/xarray.h>
+#include <net/page_pool/types.h>
+
+#include "page_pool_priv.h"
+
+static DEFINE_XARRAY_FLAGS(page_pools, XA_FLAGS_ALLOC1);
+static DEFINE_MUTEX(page_pools_lock);
+
+int page_pool_list(struct page_pool *pool)
+{
+       static u32 id_alloc_next;
+       int err;
+
+       mutex_lock(&page_pools_lock);
+       err = xa_alloc_cyclic(&page_pools, &pool->user.id, pool, xa_limit_32b,
+                             &id_alloc_next, GFP_KERNEL);
+       if (err < 0)
+               goto err_unlock;
+
+       mutex_unlock(&page_pools_lock);
+       return 0;
+
+err_unlock:
+       mutex_unlock(&page_pools_lock);
+       return err;
+}
+
+void page_pool_unlist(struct page_pool *pool)
+{
+       mutex_lock(&page_pools_lock);
+       xa_erase(&page_pools, pool->user.id);
+       mutex_unlock(&page_pools_lock);
+}