}
 }
 
-void generic_start_io_acct(int rw, unsigned long sectors,
-                          struct hd_struct *part)
+void generic_start_io_acct(struct request_queue *q, int rw,
+                          unsigned long sectors, struct hd_struct *part)
 {
        int cpu = part_stat_lock();
 
-       part_round_stats(cpu, part);
+       part_round_stats(q, cpu, part);
        part_stat_inc(cpu, part, ios[rw]);
        part_stat_add(cpu, part, sectors[rw], sectors);
-       part_inc_in_flight(part, rw);
+       part_inc_in_flight(q, part, rw);
 
        part_stat_unlock();
 }
 EXPORT_SYMBOL(generic_start_io_acct);
 
-void generic_end_io_acct(int rw, struct hd_struct *part,
-                        unsigned long start_time)
+void generic_end_io_acct(struct request_queue *q, int rw,
+                        struct hd_struct *part, unsigned long start_time)
 {
        unsigned long duration = jiffies - start_time;
        int cpu = part_stat_lock();
 
        part_stat_add(cpu, part, ticks[rw], duration);
-       part_round_stats(cpu, part);
-       part_dec_in_flight(part, rw);
+       part_round_stats(q, cpu, part);
+       part_dec_in_flight(q, part, rw);
 
        part_stat_unlock();
 }
 
        __elv_add_request(q, rq, where);
 }
 
-static void part_round_stats_single(int cpu, struct hd_struct *part,
-                                   unsigned long now)
+static void part_round_stats_single(struct request_queue *q, int cpu,
+                                   struct hd_struct *part, unsigned long now)
 {
        int inflight;
 
        if (now == part->stamp)
                return;
 
-       inflight = part_in_flight(part);
+       inflight = part_in_flight(q, part);
        if (inflight) {
                __part_stat_add(cpu, part, time_in_queue,
                                inflight * (now - part->stamp));
 
 /**
  * part_round_stats() - Round off the performance stats on a struct disk_stats.
+ * @q: target block queue
  * @cpu: cpu number for stats access
  * @part: target partition
  *
  * /proc/diskstats.  This accounts immediately for all queue usage up to
  * the current jiffies and restarts the counters again.
  */
-void part_round_stats(int cpu, struct hd_struct *part)
+void part_round_stats(struct request_queue *q, int cpu, struct hd_struct *part)
 {
        unsigned long now = jiffies;
 
        if (part->partno)
-               part_round_stats_single(cpu, &part_to_disk(part)->part0, now);
-       part_round_stats_single(cpu, part, now);
+               part_round_stats_single(q, cpu, &part_to_disk(part)->part0,
+                                               now);
+       part_round_stats_single(q, cpu, part, now);
 }
 EXPORT_SYMBOL_GPL(part_round_stats);
 
 
                part_stat_inc(cpu, part, ios[rw]);
                part_stat_add(cpu, part, ticks[rw], duration);
-               part_round_stats(cpu, part);
-               part_dec_in_flight(part, rw);
+               part_round_stats(req->q, cpu, part);
+               part_dec_in_flight(req->q, part, rw);
 
                hd_struct_put(part);
                part_stat_unlock();
                        part = &rq->rq_disk->part0;
                        hd_struct_get(part);
                }
-               part_round_stats(cpu, part);
-               part_inc_in_flight(part, rw);
+               part_round_stats(rq->q, cpu, part);
+               part_inc_in_flight(rq->q, part, rw);
                rq->part = part;
        }
 
 
                cpu = part_stat_lock();
                part = req->part;
 
-               part_round_stats(cpu, part);
-               part_dec_in_flight(part, rq_data_dir(req));
+               part_round_stats(req->q, cpu, part);
+               part_dec_in_flight(req->q, part, rq_data_dir(req));
 
                hd_struct_put(part);
                part_stat_unlock();
 
        disk_part_iter_init(&piter, gp, DISK_PITER_INCL_EMPTY_PART0);
        while ((hd = disk_part_iter_next(&piter))) {
                cpu = part_stat_lock();
-               part_round_stats(cpu, hd);
+               part_round_stats(gp->queue, cpu, hd);
                part_stat_unlock();
                seq_printf(seqf, "%4d %7d %s %lu %lu %lu "
                           "%u %lu %lu %lu %u %u %u %u\n",
                           part_stat_read(hd, merges[WRITE]),
                           part_stat_read(hd, sectors[WRITE]),
                           jiffies_to_msecs(part_stat_read(hd, ticks[WRITE])),
-                          part_in_flight(hd),
+                          part_in_flight(gp->queue, hd),
                           jiffies_to_msecs(part_stat_read(hd, io_ticks)),
                           jiffies_to_msecs(part_stat_read(hd, time_in_queue))
                        );
 
                       struct device_attribute *attr, char *buf)
 {
        struct hd_struct *p = dev_to_part(dev);
+       struct request_queue *q = dev_to_disk(dev)->queue;
        int cpu;
 
        cpu = part_stat_lock();
-       part_round_stats(cpu, p);
+       part_round_stats(q, cpu, p);
        part_stat_unlock();
        return sprintf(buf,
                "%8lu %8lu %8llu %8u "
                part_stat_read(p, merges[WRITE]),
                (unsigned long long)part_stat_read(p, sectors[WRITE]),
                jiffies_to_msecs(part_stat_read(p, ticks[WRITE])),
-               part_in_flight(p),
+               part_in_flight(q, p),
                jiffies_to_msecs(part_stat_read(p, io_ticks)),
                jiffies_to_msecs(part_stat_read(p, time_in_queue)));
 }
 
 /* Update disk stats at start of I/O request */
 static void _drbd_start_io_acct(struct drbd_device *device, struct drbd_request *req)
 {
-       generic_start_io_acct(bio_data_dir(req->master_bio), req->i.size >> 9,
-                             &device->vdisk->part0);
+       struct request_queue *q = device->rq_queue;
+
+       generic_start_io_acct(q, bio_data_dir(req->master_bio),
+                               req->i.size >> 9, &device->vdisk->part0);
 }
 
 /* Update disk stats when completing request upwards */
 static void _drbd_end_io_acct(struct drbd_device *device, struct drbd_request *req)
 {
-       generic_end_io_acct(bio_data_dir(req->master_bio),
+       struct request_queue *q = device->rq_queue;
+
+       generic_end_io_acct(q, bio_data_dir(req->master_bio),
                            &device->vdisk->part0, req->start_jif);
 }
 
 
 
 static void disk_stats_start(struct rsxx_cardinfo *card, struct bio *bio)
 {
-       generic_start_io_acct(bio_data_dir(bio), bio_sectors(bio),
+       generic_start_io_acct(card->queue, bio_data_dir(bio), bio_sectors(bio),
                             &card->gendisk->part0);
 }
 
                                struct bio *bio,
                                unsigned long start_time)
 {
-       generic_end_io_acct(bio_data_dir(bio), &card->gendisk->part0,
-                          start_time);
+       generic_end_io_acct(card->queue, bio_data_dir(bio),
+                               &card->gendisk->part0, start_time);
 }
 
 static void bio_dma_done_cb(struct rsxx_cardinfo *card,
 
 {
        unsigned long start_time = jiffies;
        int rw_acct = is_write ? REQ_OP_WRITE : REQ_OP_READ;
+       struct request_queue *q = zram->disk->queue;
        int ret;
 
-       generic_start_io_acct(rw_acct, bvec->bv_len >> SECTOR_SHIFT,
+       generic_start_io_acct(q, rw_acct, bvec->bv_len >> SECTOR_SHIFT,
                        &zram->disk->part0);
 
        if (!is_write) {
                ret = zram_bvec_write(zram, bvec, index, offset);
        }
 
-       generic_end_io_acct(rw_acct, &zram->disk->part0, start_time);
+       generic_end_io_acct(q, rw_acct, &zram->disk->part0, start_time);
 
        if (unlikely(ret)) {
                if (!is_write)
 
 static void bio_complete(struct search *s)
 {
        if (s->orig_bio) {
-               generic_end_io_acct(bio_data_dir(s->orig_bio),
+               struct request_queue *q = bdev_get_queue(s->orig_bio->bi_bdev);
+               generic_end_io_acct(q, bio_data_dir(s->orig_bio),
                                    &s->d->disk->part0, s->start_time);
 
                trace_bcache_request_end(s->d, s->orig_bio);
        struct cached_dev *dc = container_of(d, struct cached_dev, disk);
        int rw = bio_data_dir(bio);
 
-       generic_start_io_acct(rw, bio_sectors(bio), &d->disk->part0);
+       generic_start_io_acct(q, rw, bio_sectors(bio), &d->disk->part0);
 
        bio->bi_bdev = dc->bdev;
        bio->bi_iter.bi_sector += dc->sb.data_offset;
        struct bcache_device *d = bio->bi_bdev->bd_disk->private_data;
        int rw = bio_data_dir(bio);
 
-       generic_start_io_acct(rw, bio_sectors(bio), &d->disk->part0);
+       generic_start_io_acct(q, rw, bio_sectors(bio), &d->disk->part0);
 
        s = search_alloc(bio, d);
        cl = &s->cl;
 
        io->start_time = jiffies;
 
        cpu = part_stat_lock();
-       part_round_stats(cpu, &dm_disk(md)->part0);
+       part_round_stats(md->queue, cpu, &dm_disk(md)->part0);
        part_stat_unlock();
        atomic_set(&dm_disk(md)->part0.in_flight[rw],
                atomic_inc_return(&md->pending[rw]));
        int pending;
        int rw = bio_data_dir(bio);
 
-       generic_end_io_acct(rw, &dm_disk(md)->part0, io->start_time);
+       generic_end_io_acct(md->queue, rw, &dm_disk(md)->part0, io->start_time);
 
        if (unlikely(dm_stats_used(&md->stats)))
                dm_stats_account_io(&md->stats, bio_data_dir(bio),
 
        map = dm_get_live_table(md, &srcu_idx);
 
-       generic_start_io_acct(rw, bio_sectors(bio), &dm_disk(md)->part0);
+       generic_start_io_acct(q, rw, bio_sectors(bio), &dm_disk(md)->part0);
 
        /* if we're suspended, we have to queue this io for later */
        if (unlikely(test_bit(DMF_BLOCK_IO_FOR_SUSPEND, &md->flags))) {
 
                return false;
 
        *start = jiffies;
-       generic_start_io_acct(bio_data_dir(bio),
+       generic_start_io_acct(disk->queue, bio_data_dir(bio),
                              bio_sectors(bio), &disk->part0);
        return true;
 }
 {
        struct gendisk *disk = bio->bi_bdev->bd_disk;
 
-       generic_end_io_acct(bio_data_dir(bio), &disk->part0, start);
+       generic_end_io_acct(disk->queue, bio_data_dir(bio), &disk->part0,
+                               start);
 }
 static inline bool is_bad_pmem(struct badblocks *bb, sector_t sector,
                unsigned int len)
 
 extern void bio_set_pages_dirty(struct bio *bio);
 extern void bio_check_pages_dirty(struct bio *bio);
 
-void generic_start_io_acct(int rw, unsigned long sectors,
-                          struct hd_struct *part);
-void generic_end_io_acct(int rw, struct hd_struct *part,
-                        unsigned long start_time);
+void generic_start_io_acct(struct request_queue *q, int rw,
+                               unsigned long sectors, struct hd_struct *part);
+void generic_end_io_acct(struct request_queue *q, int rw,
+                               struct hd_struct *part,
+                               unsigned long start_time);
 
 #ifndef ARCH_IMPLEMENTS_FLUSH_DCACHE_PAGE
 # error        "You should define ARCH_IMPLEMENTS_FLUSH_DCACHE_PAGE for your platform"
 
 #define part_stat_sub(cpu, gendiskp, field, subnd)                     \
        part_stat_add(cpu, gendiskp, field, -subnd)
 
-static inline void part_inc_in_flight(struct hd_struct *part, int rw)
+static inline void part_inc_in_flight(struct request_queue *q,
+                                     struct hd_struct *part, int rw)
 {
        atomic_inc(&part->in_flight[rw]);
        if (part->partno)
                atomic_inc(&part_to_disk(part)->part0.in_flight[rw]);
 }
 
-static inline void part_dec_in_flight(struct hd_struct *part, int rw)
+static inline void part_dec_in_flight(struct request_queue *q,
+                                     struct hd_struct *part, int rw)
 {
        atomic_dec(&part->in_flight[rw]);
        if (part->partno)
                atomic_dec(&part_to_disk(part)->part0.in_flight[rw]);
 }
 
-static inline int part_in_flight(struct hd_struct *part)
+static inline int part_in_flight(struct request_queue *q,
+                                struct hd_struct *part)
 {
-       return atomic_read(&part->in_flight[0]) + atomic_read(&part->in_flight[1]);
+       return atomic_read(&part->in_flight[0]) +
+                       atomic_read(&part->in_flight[1]);
 }
 
 static inline struct partition_meta_info *alloc_part_info(struct gendisk *disk)
 }
 
 /* block/blk-core.c */
-extern void part_round_stats(int cpu, struct hd_struct *part);
+extern void part_round_stats(struct request_queue *q, int cpu, struct hd_struct *part);
 
 /* block/genhd.c */
 extern void device_add_disk(struct device *parent, struct gendisk *disk);