}
 }
 
-static int blkdev_zone_reset_all_emulated(struct block_device *bdev,
-                                         gfp_t gfp_mask)
+static int blkdev_zone_reset_all_emulated(struct block_device *bdev)
 {
        struct gendisk *disk = bdev->bd_disk;
        sector_t capacity = bdev_nr_sectors(bdev);
                }
 
                bio = blk_next_bio(bio, bdev, 0, REQ_OP_ZONE_RESET | REQ_SYNC,
-                                  gfp_mask);
+                                  GFP_KERNEL);
                bio->bi_iter.bi_sector = sector;
                sector += zone_sectors;
 
        return ret;
 }
 
-static int blkdev_zone_reset_all(struct block_device *bdev, gfp_t gfp_mask)
+static int blkdev_zone_reset_all(struct block_device *bdev)
 {
        struct bio bio;
 
  * @sector:    Start sector of the first zone to operate on
  * @nr_sectors:        Number of sectors, should be at least the length of one zone and
  *             must be zone size aligned.
- * @gfp_mask:  Memory allocation flags (for bio_alloc)
  *
  * Description:
  *    Perform the specified operation on the range of zones specified by
  *    or finish request.
  */
 int blkdev_zone_mgmt(struct block_device *bdev, enum req_op op,
-                    sector_t sector, sector_t nr_sectors, gfp_t gfp_mask)
+                    sector_t sector, sector_t nr_sectors)
 {
        struct request_queue *q = bdev_get_queue(bdev);
        sector_t zone_sectors = bdev_zone_sectors(bdev);
         */
        if (op == REQ_OP_ZONE_RESET && sector == 0 && nr_sectors == capacity) {
                if (!blk_queue_zone_resetall(q))
-                       return blkdev_zone_reset_all_emulated(bdev, gfp_mask);
-               return blkdev_zone_reset_all(bdev, gfp_mask);
+                       return blkdev_zone_reset_all_emulated(bdev);
+               return blkdev_zone_reset_all(bdev);
        }
 
        while (sector < end_sector) {
-               bio = blk_next_bio(bio, bdev, 0, op | REQ_SYNC, gfp_mask);
+               bio = blk_next_bio(bio, bdev, 0, op | REQ_SYNC, GFP_KERNEL);
                bio->bi_iter.bi_sector = sector;
                sector += zone_sectors;
 
                return -ENOTTY;
        }
 
-       ret = blkdev_zone_mgmt(bdev, op, zrange.sector, zrange.nr_sectors,
-                              GFP_KERNEL);
+       ret = blkdev_zone_mgmt(bdev, op, zrange.sector, zrange.nr_sectors);
 
 fail:
        if (cmd == BLKRESETZONE)
 
                noio_flag = memalloc_noio_save();
                ret = blkdev_zone_mgmt(dev->bdev, REQ_OP_ZONE_RESET,
                                       dmz_start_sect(zmd, zone),
-                                      zmd->zone_nr_sectors, GFP_KERNEL);
+                                      zmd->zone_nr_sectors);
                memalloc_noio_restore(noio_flag);
                if (ret) {
                        dmz_dev_err(dev, "Reset zone %u failed %d",
 
        switch (zsa_req_op(req->cmd->zms.zsa)) {
        case REQ_OP_ZONE_RESET:
                ret = blkdev_zone_mgmt(req->ns->bdev, REQ_OP_ZONE_RESET, 0,
-                                      get_capacity(req->ns->bdev->bd_disk),
-                                      GFP_KERNEL);
+                                      get_capacity(req->ns->bdev->bd_disk));
                if (ret < 0)
                        return blkdev_zone_mgmt_errno_to_nvme_status(ret);
                break;
                goto out;
        }
 
-       ret = blkdev_zone_mgmt(bdev, op, sect, zone_sectors, GFP_KERNEL);
+       ret = blkdev_zone_mgmt(bdev, op, sect, zone_sectors);
        if (ret < 0)
                status = blkdev_zone_mgmt_errno_to_nvme_status(ret);
 
 
 
                        nofs_flags = memalloc_nofs_save();
                        ret = blkdev_zone_mgmt(bdev, REQ_OP_ZONE_RESET,
-                                              reset->start, reset->len,
-                                              GFP_KERNEL);
+                                              reset->start, reset->len);
                        memalloc_nofs_restore(nofs_flags);
                        if (ret)
                                return ret;
                                nofs_flags = memalloc_nofs_save();
                                ret = blkdev_zone_mgmt(device->bdev,
                                                REQ_OP_ZONE_FINISH, zone->start,
-                                               zone->len, GFP_KERNEL);
+                                               zone->len);
                                memalloc_nofs_restore(nofs_flags);
                                if (ret)
                                        return ret;
        nofs_flags = memalloc_nofs_save();
        ret = blkdev_zone_mgmt(bdev, REQ_OP_ZONE_RESET,
                               zone_start_sector(sb_zone, bdev),
-                              zone_sectors * BTRFS_NR_SB_LOG_ZONES,
-                              GFP_KERNEL);
+                              zone_sectors * BTRFS_NR_SB_LOG_ZONES);
        memalloc_nofs_restore(nofs_flags);
        return ret;
 }
        *bytes = 0;
        nofs_flags = memalloc_nofs_save();
        ret = blkdev_zone_mgmt(device->bdev, REQ_OP_ZONE_RESET,
-                              physical >> SECTOR_SHIFT, length >> SECTOR_SHIFT,
-                              GFP_KERNEL);
+                              physical >> SECTOR_SHIFT, length >> SECTOR_SHIFT);
        memalloc_nofs_restore(nofs_flags);
        if (ret)
                return ret;
                nofs_flags = memalloc_nofs_save();
                ret = blkdev_zone_mgmt(device->bdev, REQ_OP_ZONE_FINISH,
                                       physical >> SECTOR_SHIFT,
-                                      zinfo->zone_size >> SECTOR_SHIFT,
-                                      GFP_KERNEL);
+                                      zinfo->zone_size >> SECTOR_SHIFT);
                memalloc_nofs_restore(nofs_flags);
 
                if (ret)
 
                        trace_f2fs_issue_reset_zone(bdev, blkstart);
                        nofs_flags = memalloc_nofs_save();
                        ret = blkdev_zone_mgmt(bdev, REQ_OP_ZONE_RESET,
-                                               sector, nr_sects, GFP_KERNEL);
+                                               sector, nr_sects);
                        memalloc_nofs_restore(nofs_flags);
                        return ret;
                }
 
        nofs_flags = memalloc_nofs_save();
        ret = blkdev_zone_mgmt(fdev->bdev, REQ_OP_ZONE_FINISH,
-                               zone->start, zone->len, GFP_KERNEL);
+                               zone->start, zone->len);
        memalloc_nofs_restore(nofs_flags);
        if (ret == -EOPNOTSUPP) {
                ret = blkdev_issue_zeroout(fdev->bdev, zone->wp,
 
 
        trace_zonefs_zone_mgmt(sb, z, op);
        ret = blkdev_zone_mgmt(sb->s_bdev, op, z->z_sector,
-                              z->z_size >> SECTOR_SHIFT, GFP_KERNEL);
+                              z->z_size >> SECTOR_SHIFT);
        if (ret) {
                zonefs_err(sb,
                           "Zone management operation %s at %llu failed %d\n",
 
 int blkdev_report_zones(struct block_device *bdev, sector_t sector,
                unsigned int nr_zones, report_zones_cb cb, void *data);
 int blkdev_zone_mgmt(struct block_device *bdev, enum req_op op,
-               sector_t sectors, sector_t nr_sectors, gfp_t gfp_mask);
+               sector_t sectors, sector_t nr_sectors);
 int blk_revalidate_disk_zones(struct gendisk *disk,
                void (*update_driver_data)(struct gendisk *disk));