static void emulate_exception(struct x86_emulate_ctxt *ctxt, int vec,
                                      u32 error, bool valid)
 {
-       ctxt->exception = vec;
-       ctxt->error_code = error;
-       ctxt->error_code_valid = valid;
+       ctxt->exception.vector = vec;
+       ctxt->exception.error_code = error;
+       ctxt->exception.error_code_valid = valid;
 }
 
 static void emulate_gp(struct x86_emulate_ctxt *ctxt, int err)
 
        if (ctxt->mode == X86EMUL_MODE_PROT64 && (c->d & No64)) {
                emulate_ud(ctxt);
+               rc = X86EMUL_PROPAGATE_FAULT;
                goto done;
        }
 
        /* LOCK prefix is allowed only with some instructions */
        if (c->lock_prefix && (!(c->d & Lock) || c->dst.type != OP_MEM)) {
                emulate_ud(ctxt);
+               rc = X86EMUL_PROPAGATE_FAULT;
                goto done;
        }
 
        if ((c->d & SrcMask) == SrcMemFAddr && c->src.type != OP_MEM) {
                emulate_ud(ctxt);
+               rc = X86EMUL_PROPAGATE_FAULT;
                goto done;
        }
 
        /* Privileged instruction can be executed only in CPL=0 */
        if ((c->d & Priv) && ops->cpl(ctxt->vcpu)) {
                emulate_gp(ctxt, 0);
+               rc = X86EMUL_PROPAGATE_FAULT;
                goto done;
        }
 
        case 0x8c:  /* mov r/m, sreg */
                if (c->modrm_reg > VCPU_SREG_GS) {
                        emulate_ud(ctxt);
+                       rc = X86EMUL_PROPAGATE_FAULT;
                        goto done;
                }
                c->dst.val = ops->get_segment_selector(c->modrm_reg, ctxt->vcpu);
                if (c->modrm_reg == VCPU_SREG_CS ||
                    c->modrm_reg > VCPU_SREG_GS) {
                        emulate_ud(ctxt);
+                       rc = X86EMUL_PROPAGATE_FAULT;
                        goto done;
                }
 
                c->dst.bytes = min(c->dst.bytes, 4u);
                if (!emulator_io_permited(ctxt, ops, c->src.val, c->dst.bytes)) {
                        emulate_gp(ctxt, 0);
+                       rc = X86EMUL_PROPAGATE_FAULT;
                        goto done;
                }
                if (!pio_in_emulated(ctxt, ops, c->dst.bytes, c->src.val,
                if (!emulator_io_permited(ctxt, ops, c->dst.val,
                                          c->src.bytes)) {
                        emulate_gp(ctxt, 0);
+                       rc = X86EMUL_PROPAGATE_FAULT;
                        goto done;
                }
                ops->pio_out_emulated(c->src.bytes, c->dst.val,
        case 0xfa: /* cli */
                if (emulator_bad_iopl(ctxt, ops)) {
                        emulate_gp(ctxt, 0);
+                       rc = X86EMUL_PROPAGATE_FAULT;
                        goto done;
                } else
                        ctxt->eflags &= ~X86_EFLAGS_IF;
        case 0xfb: /* sti */
                if (emulator_bad_iopl(ctxt, ops)) {
                        emulate_gp(ctxt, 0);
+                       rc = X86EMUL_PROPAGATE_FAULT;
                        goto done;
                } else {
                        ctxt->interruptibility = KVM_X86_SHADOW_INT_STI;
        ctxt->eip = c->eip;
 
 done:
+       if (rc == X86EMUL_PROPAGATE_FAULT)
+               ctxt->have_exception = true;
        return (rc == X86EMUL_UNHANDLEABLE) ? EMULATION_FAILED : EMULATION_OK;
 
 twobyte_insn:
                        break;
                case 5: /* not defined */
                        emulate_ud(ctxt);
+                       rc = X86EMUL_PROPAGATE_FAULT;
                        goto done;
                case 7: /* invlpg*/
                        emulate_invlpg(ctxt->vcpu,
                case 5 ... 7:
                case 9 ... 15:
                        emulate_ud(ctxt);
+                       rc = X86EMUL_PROPAGATE_FAULT;
                        goto done;
                }
                c->dst.val = ops->get_cr(c->modrm_reg, ctxt->vcpu);
                if ((ops->get_cr(4, ctxt->vcpu) & X86_CR4_DE) &&
                    (c->modrm_reg == 4 || c->modrm_reg == 5)) {
                        emulate_ud(ctxt);
+                       rc = X86EMUL_PROPAGATE_FAULT;
                        goto done;
                }
                ops->get_dr(c->modrm_reg, &c->dst.val, ctxt->vcpu);
        case 0x22: /* mov reg, cr */
                if (ops->set_cr(c->modrm_reg, c->src.val, ctxt->vcpu)) {
                        emulate_gp(ctxt, 0);
+                       rc = X86EMUL_PROPAGATE_FAULT;
                        goto done;
                }
                c->dst.type = OP_NONE;
                if ((ops->get_cr(4, ctxt->vcpu) & X86_CR4_DE) &&
                    (c->modrm_reg == 4 || c->modrm_reg == 5)) {
                        emulate_ud(ctxt);
+                       rc = X86EMUL_PROPAGATE_FAULT;
                        goto done;
                }
 
                                 ~0ULL : ~0U), ctxt->vcpu) < 0) {
                        /* #UD condition is already handled by the code above */
                        emulate_gp(ctxt, 0);
+                       rc = X86EMUL_PROPAGATE_FAULT;
                        goto done;
                }
 
                        | ((u64)c->regs[VCPU_REGS_RDX] << 32);
                if (ops->set_msr(ctxt->vcpu, c->regs[VCPU_REGS_RCX], msr_data)) {
                        emulate_gp(ctxt, 0);
+                       rc = X86EMUL_PROPAGATE_FAULT;
                        goto done;
                }
                rc = X86EMUL_CONTINUE;
                /* rdmsr */
                if (ops->get_msr(ctxt->vcpu, c->regs[VCPU_REGS_RCX], &msr_data)) {
                        emulate_gp(ctxt, 0);
+                       rc = X86EMUL_PROPAGATE_FAULT;
                        goto done;
                } else {
                        c->regs[VCPU_REGS_RAX] = (u32)msr_data;
 
 static void inject_emulated_exception(struct kvm_vcpu *vcpu)
 {
        struct x86_emulate_ctxt *ctxt = &vcpu->arch.emulate_ctxt;
-       if (ctxt->exception == PF_VECTOR)
+       if (ctxt->exception.vector == PF_VECTOR)
                kvm_propagate_fault(vcpu);
-       else if (ctxt->error_code_valid)
-               kvm_queue_exception_e(vcpu, ctxt->exception, ctxt->error_code);
+       else if (ctxt->exception.error_code_valid)
+               kvm_queue_exception_e(vcpu, ctxt->exception.vector,
+                                     ctxt->exception.error_code);
        else
-               kvm_queue_exception(vcpu, ctxt->exception);
+               kvm_queue_exception(vcpu, ctxt->exception.vector);
 }
 
 static void init_emulate_ctxt(struct kvm_vcpu *vcpu)
        if (!(emulation_type & EMULTYPE_NO_DECODE)) {
                init_emulate_ctxt(vcpu);
                vcpu->arch.emulate_ctxt.interruptibility = 0;
-               vcpu->arch.emulate_ctxt.exception = -1;
+               vcpu->arch.emulate_ctxt.have_exception = false;
                vcpu->arch.emulate_ctxt.perm_ok = false;
 
                r = x86_decode_insn(&vcpu->arch.emulate_ctxt);
        }
 
 done:
-       if (vcpu->arch.emulate_ctxt.exception >= 0) {
+       if (vcpu->arch.emulate_ctxt.have_exception) {
                inject_emulated_exception(vcpu);
                r = EMULATE_DONE;
        } else if (vcpu->arch.pio.count) {