u32 *cs;
        int i;
 
-       if (!IS_GEN(rq->i915, 7) || rq->engine->id != RCS0) {
-               drm_dbg(&rq->i915->drm, "sol reset is gen7/rcs only\n");
+       if (!IS_GEN(rq->engine->i915, 7) || rq->engine->id != RCS0) {
+               drm_dbg(&rq->engine->i915->drm, "sol reset is gen7/rcs only\n");
                return -EINVAL;
        }
 
 
        cmd = MI_FLUSH;
        if (mode & EMIT_INVALIDATE) {
                cmd |= MI_EXE_FLUSH;
-               if (IS_G4X(rq->i915) || IS_GEN(rq->i915, 5))
+               if (IS_G4X(rq->engine->i915) || IS_GEN(rq->engine->i915, 5))
                        cmd |= MI_INVALIDATE_ISP;
        }
 
 
        *cs++ = MI_STORE_DWORD_IMM_GEN4 | MI_USE_GGTT;
        *cs++ = lower_32_bits(offset);
        *cs++ = upper_32_bits(offset);
-       *cs++ = intel_sseu_make_rpcs(rq->i915, &sseu);
+       *cs++ = intel_sseu_make_rpcs(rq->engine->i915, &sseu);
 
        intel_ring_advance(rq, cs);
 
 
        if (!frame)
                return -ENOMEM;
 
-       frame->rq.i915 = engine->i915;
        frame->rq.engine = engine;
        frame->rq.context = ce;
        rcu_assign_pointer(frame->rq.timeline, ce->timeline);
        }
 }
 
-static int print_sched_attr(struct drm_i915_private *i915,
-                           const struct i915_sched_attr *attr,
+static int print_sched_attr(const struct i915_sched_attr *attr,
                            char *buf, int x, int len)
 {
        if (attr->priority == I915_PRIORITY_INVALID)
        char buf[80] = "";
        int x = 0;
 
-       x = print_sched_attr(rq->i915, &rq->sched.attr, buf, x, sizeof(buf));
+       x = print_sched_attr(&rq->sched.attr, buf, x, sizeof(buf));
 
        drm_printf(m, "%s %llx:%llx%s%s %s @ %dms: %s\n",
                   prefix,
 
        int err, i;
        u32 *cs;
 
-       GEM_BUG_ON(intel_vgpu_active(rq->i915));
+       GEM_BUG_ON(intel_vgpu_active(rq->engine->i915));
 
        /*
         * Beware ye of the dragons, this sequence is magic!
                 * On GEN9: before VF_CACHE_INVALIDATE we need to emit a NULL
                 * pipe control.
                 */
-               if (IS_GEN(request->i915, 9))
+               if (IS_GEN(request->engine->i915, 9))
                        vf_flush_wa = true;
 
                /* WaForGAMHang:kbl */
-               if (IS_KBL_REVID(request->i915, 0, KBL_REVID_B0))
+               if (IS_KBL_REVID(request->engine->i915, 0, KBL_REVID_B0))
                        dc_flush_wa = true;
        }
 
 
                                 struct intel_context *ce,
                                 u32 flags)
 {
-       struct drm_i915_private *i915 = rq->i915;
        struct intel_engine_cs *engine = rq->engine;
+       struct drm_i915_private *i915 = engine->i915;
        enum intel_engine_id id;
        const int num_engines =
                IS_HASWELL(i915) ? RUNTIME_INFO(i915)->num_engines - 1 : 0;
 
 static int remap_l3_slice(struct i915_request *rq, int slice)
 {
-       u32 *cs, *remap_info = rq->i915->l3_parity.remap_info[slice];
+       u32 *cs, *remap_info = rq->engine->i915->l3_parity.remap_info[slice];
        int i;
 
        if (!remap_info)
        void **residuals = NULL;
        int ret;
 
-       GEM_BUG_ON(HAS_EXECLISTS(rq->i915));
+       GEM_BUG_ON(HAS_EXECLISTS(engine->i915));
 
        if (engine->wa_ctx.vma && ce != engine->kernel_context) {
                if (engine->wa_ctx.vma->private != ce) {
 
            const struct i915_wa_list *wal,
            struct i915_vma *vma)
 {
-       struct drm_i915_private *i915 = rq->i915;
+       struct drm_i915_private *i915 = rq->engine->i915;
        unsigned int i, count = 0;
        const struct i915_wa *wa;
        u32 srm, *cs;
 
        err = 0;
        for (i = 0, wa = wal->list; i < wal->count; i++, wa++) {
-               if (mcr_range(rq->i915, i915_mmio_reg_offset(wa->reg)))
+               if (mcr_range(rq->engine->i915, i915_mmio_reg_offset(wa->reg)))
                        continue;
 
                if (!wa_verify(wa, results[i], wal->name, from))
 
                return PTR_ERR(cs);
 
        cmd = MI_STORE_REGISTER_MEM | MI_USE_GGTT;
-       if (INTEL_GEN(rq->i915) >= 8)
+       if (INTEL_GEN(rq->engine->i915) >= 8)
                cmd++;
        *cs++ = cmd;
        *cs++ = i915_mmio_reg_offset(RING_TIMESTAMP(rq->engine->mmio_base));
 
 {
        u32 addr;
 
-       if (HAS_GLOBAL_MOCS_REGISTERS(rq->i915))
+       if (HAS_GLOBAL_MOCS_REGISTERS(rq->engine->i915))
                addr = global_mocs_offset();
        else
                addr = mocs_offset(rq->engine);
 
        }
 
        cmd = MI_STORE_REGISTER_MEM | MI_USE_GGTT;
-       if (INTEL_GEN(rq->i915) >= 8)
+       if (INTEL_GEN(rq->engine->i915) >= 8)
                cmd++;
 
        *cs++ = cmd;
                int pass;
 
                for (pass = 0; pass < 2; pass++) {
+                       struct i915_gpu_error *error = >->i915->gpu_error;
                        struct intel_context *ce;
                        unsigned int resets =
-                               i915_reset_engine_count(>->i915->gpu_error,
-                                                       engine);
+                               i915_reset_engine_count(error, engine);
                        const u32 *res;
 
                        /* Use a sacrifical context */
                                 engine->name, READ_ONCE(*res));
 
                        if (resets !=
-                           i915_reset_engine_count(>->i915->gpu_error,
-                                                   engine)) {
+                           i915_reset_engine_count(error, engine)) {
                                pr_err("%s: GPU reset required\n",
                                       engine->name);
                                add_taint_for_CI(TAINT_WARN);
 
        if (IS_ERR(cs))
                return PTR_ERR(cs);
 
-       if (INTEL_GEN(rq->i915) >= 8) {
+       if (INTEL_GEN(rq->engine->i915) >= 8) {
                *cs++ = MI_STORE_DWORD_IMM_GEN4 | MI_USE_GGTT;
                *cs++ = addr;
                *cs++ = 0;
                *cs++ = value;
-       } else if (INTEL_GEN(rq->i915) >= 4) {
+       } else if (INTEL_GEN(rq->engine->i915) >= 4) {
                *cs++ = MI_STORE_DWORD_IMM_GEN4 | MI_USE_GGTT;
                *cs++ = 0;
                *cs++ = addr;
 
        u32 *cs;
        int err;
 
-       if (IS_GEN(req->i915, 9) && is_inhibit_context(req->context))
+       if (IS_GEN(req->engine->i915, 9) && is_inhibit_context(req->context))
                intel_vgpu_restore_inhibit_context(vgpu, req);
 
        /*
        context_page_num = rq->engine->context_size;
        context_page_num = context_page_num >> PAGE_SHIFT;
 
-       if (IS_BROADWELL(rq->i915) && rq->engine->id == RCS0)
+       if (IS_BROADWELL(rq->engine->i915) && rq->engine->id == RCS0)
                context_page_num = 19;
 
        context_base = (void *) ctx->lrc_reg_state -
 
 
 static const char *i915_fence_get_driver_name(struct dma_fence *fence)
 {
-       return dev_name(to_request(fence)->i915->drm.dev);
+       return dev_name(to_request(fence)->engine->i915->drm.dev);
 }
 
 static const char *i915_fence_get_timeline_name(struct dma_fence *fence)
                }
        }
 
-       rq->i915 = ce->engine->i915;
        rq->context = ce;
        rq->engine = ce->engine;
        rq->ring = ce->ring;
                      struct i915_request *from,
                      u32 seqno)
 {
-       const int has_token = INTEL_GEN(to->i915) >= 12;
+       const int has_token = INTEL_GEN(to->engine->i915) >= 12;
        u32 hwsp_offset;
        int len, err;
        u32 *cs;
 
-       GEM_BUG_ON(INTEL_GEN(to->i915) < 8);
+       GEM_BUG_ON(INTEL_GEN(to->engine->i915) < 8);
        GEM_BUG_ON(i915_request_has_initial_breadcrumb(to));
 
        /* We need to pin the signaler's HWSP until we are finished reading. */
 {
        mark_external(rq);
        return i915_sw_fence_await_dma_fence(&rq->submit, fence,
-                                            i915_fence_context_timeout(rq->i915,
+                                            i915_fence_context_timeout(rq->engine->i915,
                                                                        fence->context),
                                             I915_FENCE_GFP);
 }
         * (bad for battery).
         */
        if (flags & I915_WAIT_PRIORITY) {
-               if (!i915_request_started(rq) && INTEL_GEN(rq->i915) >= 6)
+               if (!i915_request_started(rq) &&
+                   INTEL_GEN(rq->engine->i915) >= 6)
                        intel_rps_boost(rq);
        }
 
 
        struct dma_fence fence;
        spinlock_t lock;
 
-       /** On Which ring this request was generated */
-       struct drm_i915_private *i915;
-
        /**
         * Context and ring buffer related to this request
         * Contexts are refcounted, so when this request is associated with a
 
                             ),
 
            TP_fast_assign(
-                          __entry->dev = rq->i915->drm.primary->index;
+                          __entry->dev = rq->engine->i915->drm.primary->index;
                           __entry->class = rq->engine->uabi_class;
                           __entry->instance = rq->engine->uabi_instance;
                           __entry->ctx = rq->fence.context;
                             ),
 
            TP_fast_assign(
-                          __entry->dev = rq->i915->drm.primary->index;
+                          __entry->dev = rq->engine->i915->drm.primary->index;
                           __entry->class = rq->engine->uabi_class;
                           __entry->instance = rq->engine->uabi_instance;
                           __entry->ctx = rq->fence.context;
                            ),
 
            TP_fast_assign(
-                          __entry->dev = rq->i915->drm.primary->index;
+                          __entry->dev = rq->engine->i915->drm.primary->index;
                           __entry->class = rq->engine->uabi_class;
                           __entry->instance = rq->engine->uabi_instance;
                           __entry->ctx = rq->fence.context;
                            ),
 
            TP_fast_assign(
-                          __entry->dev = rq->i915->drm.primary->index;
+                          __entry->dev = rq->engine->i915->drm.primary->index;
                           __entry->class = rq->engine->uabi_class;
                           __entry->instance = rq->engine->uabi_instance;
                           __entry->ctx = rq->fence.context;
             * less desirable.
             */
            TP_fast_assign(
-                          __entry->dev = rq->i915->drm.primary->index;
+                          __entry->dev = rq->engine->i915->drm.primary->index;
                           __entry->class = rq->engine->uabi_class;
                           __entry->instance = rq->engine->uabi_instance;
                           __entry->ctx = rq->fence.context;
 
                return PTR_ERR(cs);
 
        len = 5;
-       if (INTEL_GEN(rq->i915) >= 8)
+       if (INTEL_GEN(rq->engine->i915) >= 8)
                len++;
 
        *cs++ = GFX_OP_PIPE_CONTROL(len);
 
 
        batch = spin->batch;
 
-       if (INTEL_GEN(rq->i915) >= 8) {
+       if (INTEL_GEN(rq->engine->i915) >= 8) {
                *batch++ = MI_STORE_DWORD_IMM_GEN4;
                *batch++ = lower_32_bits(hws_address(hws, rq));
                *batch++ = upper_32_bits(hws_address(hws, rq));
-       } else if (INTEL_GEN(rq->i915) >= 6) {
+       } else if (INTEL_GEN(rq->engine->i915) >= 6) {
                *batch++ = MI_STORE_DWORD_IMM_GEN4;
                *batch++ = 0;
                *batch++ = hws_address(hws, rq);
-       } else if (INTEL_GEN(rq->i915) >= 4) {
+       } else if (INTEL_GEN(rq->engine->i915) >= 4) {
                *batch++ = MI_STORE_DWORD_IMM_GEN4 | MI_USE_GGTT;
                *batch++ = 0;
                *batch++ = hws_address(hws, rq);
 
        *batch++ = arbitration_command;
 
-       if (INTEL_GEN(rq->i915) >= 8)
+       if (INTEL_GEN(rq->engine->i915) >= 8)
                *batch++ = MI_BATCH_BUFFER_START | BIT(8) | 1;
-       else if (IS_HASWELL(rq->i915))
+       else if (IS_HASWELL(rq->engine->i915))
                *batch++ = MI_BATCH_BUFFER_START | MI_BATCH_PPGTT_HSW;
-       else if (INTEL_GEN(rq->i915) >= 6)
+       else if (INTEL_GEN(rq->engine->i915) >= 6)
                *batch++ = MI_BATCH_BUFFER_START;
        else
                *batch++ = MI_BATCH_BUFFER_START | MI_BATCH_GTT;
        }
 
        flags = 0;
-       if (INTEL_GEN(rq->i915) <= 5)
+       if (INTEL_GEN(rq->engine->i915) <= 5)
                flags |= I915_DISPATCH_SECURE;
        err = engine->emit_bb_start(rq, vma->node.start, PAGE_SIZE, flags);