]> www.infradead.org Git - users/dwmw2/linux.git/commitdiff
nvmet: make nvmet_wq visible in sysfs
authorGuixin Liu <kanie@linux.alibaba.com>
Thu, 31 Oct 2024 02:27:20 +0000 (10:27 +0800)
committerKeith Busch <kbusch@kernel.org>
Tue, 5 Nov 2024 16:36:18 +0000 (08:36 -0800)
In some complex scenarios, we deploy multiple tasks on a single machine
(hybrid deployment), such as Docker containers for function computation
(background processing), real-time tasks, monitoring, event handling,
and management, along with an NVMe target server.

Each of these components is restricted to its own CPU cores to prevent
mutual interference and ensure strict isolation. To achieve this level
of isolation for nvmet_wq we need to  use sysfs tunables such as
cpumask that are currently not accessible.

Add WQ_SYSFS flag to alloc_workqueue() when creating nvmet_wq so
workqueue tunables are exported in the userspace via sysfs.

with this patch :-

  nvme (nvme-6.13) # ls /sys/devices/virtual/workqueue/nvmet-wq/
  affinity_scope  affinity_strict  cpumask  max_active  nice per_cpu
  power  subsystem  uevent

Signed-off-by: Guixin Liu <kanie@linux.alibaba.com>
Reviewed-by: Chaitanya Kulkarni <kch@nvidia.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Keith Busch <kbusch@kernel.org>
drivers/nvme/target/core.c

index ed2424f8a396e015d6debd1027dcb75795b673d5..15b25f464e7761001ccf333acedc5ab93fafb953 100644 (file)
@@ -1717,7 +1717,7 @@ static int __init nvmet_init(void)
                goto out_free_zbd_work_queue;
 
        nvmet_wq = alloc_workqueue("nvmet-wq",
-                       WQ_MEM_RECLAIM | WQ_UNBOUND, 0);
+                       WQ_MEM_RECLAIM | WQ_UNBOUND | WQ_SYSFS, 0);
        if (!nvmet_wq)
                goto out_free_buffered_work_queue;