mmu_page_add_parent_pte(vcpu, sp, parent_pte);
                if (sp->unsync_children) {
-                       set_bit(KVM_REQ_MMU_SYNC, &vcpu->requests);
+                       kvm_make_request(KVM_REQ_MMU_SYNC, vcpu);
                        kvm_mmu_mark_parents_unsync(sp);
                } else if (sp->unsync)
                        kvm_mmu_mark_parents_unsync(sp);
        int ret = 0;
 
        if (!kvm_is_visible_gfn(vcpu->kvm, root_gfn)) {
-               set_bit(KVM_REQ_TRIPLE_FAULT, &vcpu->requests);
+               kvm_make_request(KVM_REQ_TRIPLE_FAULT, vcpu);
                ret = 1;
        }
 
 void kvm_mmu_flush_tlb(struct kvm_vcpu *vcpu)
 {
        ++vcpu->stat.tlb_flush;
-       set_bit(KVM_REQ_TLB_FLUSH, &vcpu->requests);
+       kvm_make_request(KVM_REQ_TLB_FLUSH, vcpu);
 }
 
 static void paging_new_cr3(struct kvm_vcpu *vcpu)
 
        prev_nr = vcpu->arch.exception.nr;
        if (prev_nr == DF_VECTOR) {
                /* triple fault -> shutdown */
-               set_bit(KVM_REQ_TRIPLE_FAULT, &vcpu->requests);
+               kvm_make_request(KVM_REQ_TRIPLE_FAULT, vcpu);
                return;
        }
        class1 = exception_class(prev_nr);
 
        if (!vcpu->time_page)
                return 0;
-       set_bit(KVM_REQ_KVMCLOCK_UPDATE, &v->requests);
+       kvm_make_request(KVM_REQ_KVMCLOCK_UPDATE, v);
        return 1;
 }
 
                        printk(KERN_DEBUG "kvm: set_mce: "
                               "injects mce exception while "
                               "previous one is in progress!\n");
-                       set_bit(KVM_REQ_TRIPLE_FAULT, &vcpu->requests);
+                       kvm_make_request(KVM_REQ_TRIPLE_FAULT, vcpu);
                        return 0;
                }
                if (banks[1] & MCI_STATUS_VAL)
                vcpu->run->request_interrupt_window;
 
        if (vcpu->requests)
-               if (test_and_clear_bit(KVM_REQ_MMU_RELOAD, &vcpu->requests))
+               if (kvm_check_request(KVM_REQ_MMU_RELOAD, vcpu))
                        kvm_mmu_unload(vcpu);
 
        r = kvm_mmu_reload(vcpu);
                goto out;
 
        if (vcpu->requests) {
-               if (test_and_clear_bit(KVM_REQ_MIGRATE_TIMER, &vcpu->requests))
+               if (kvm_check_request(KVM_REQ_MIGRATE_TIMER, vcpu))
                        __kvm_migrate_timers(vcpu);
-               if (test_and_clear_bit(KVM_REQ_KVMCLOCK_UPDATE, &vcpu->requests))
+               if (kvm_check_request(KVM_REQ_KVMCLOCK_UPDATE, vcpu))
                        kvm_write_guest_time(vcpu);
-               if (test_and_clear_bit(KVM_REQ_MMU_SYNC, &vcpu->requests))
+               if (kvm_check_request(KVM_REQ_MMU_SYNC, vcpu))
                        kvm_mmu_sync_roots(vcpu);
-               if (test_and_clear_bit(KVM_REQ_TLB_FLUSH, &vcpu->requests))
+               if (kvm_check_request(KVM_REQ_TLB_FLUSH, vcpu))
                        kvm_x86_ops->tlb_flush(vcpu);
-               if (test_and_clear_bit(KVM_REQ_REPORT_TPR_ACCESS,
-                                      &vcpu->requests)) {
+               if (kvm_check_request(KVM_REQ_REPORT_TPR_ACCESS, vcpu)) {
                        vcpu->run->exit_reason = KVM_EXIT_TPR_ACCESS;
                        r = 0;
                        goto out;
                }
-               if (test_and_clear_bit(KVM_REQ_TRIPLE_FAULT, &vcpu->requests)) {
+               if (kvm_check_request(KVM_REQ_TRIPLE_FAULT, vcpu)) {
                        vcpu->run->exit_reason = KVM_EXIT_SHUTDOWN;
                        r = 0;
                        goto out;
                }
-               if (test_and_clear_bit(KVM_REQ_DEACTIVATE_FPU, &vcpu->requests)) {
+               if (kvm_check_request(KVM_REQ_DEACTIVATE_FPU, vcpu)) {
                        vcpu->fpu_active = 0;
                        kvm_x86_ops->fpu_deactivate(vcpu);
                }
                        srcu_read_unlock(&kvm->srcu, vcpu->srcu_idx);
                        kvm_vcpu_block(vcpu);
                        vcpu->srcu_idx = srcu_read_lock(&kvm->srcu);
-                       if (test_and_clear_bit(KVM_REQ_UNHALT, &vcpu->requests))
+                       if (kvm_check_request(KVM_REQ_UNHALT, vcpu))
                        {
                                switch(vcpu->arch.mp_state) {
                                case KVM_MP_STATE_HALTED:
        vcpu->guest_fpu_loaded = 0;
        fpu_save_init(&vcpu->arch.guest_fpu);
        ++vcpu->stat.fpu_reload;
-       set_bit(KVM_REQ_DEACTIVATE_FPU, &vcpu->requests);
+       kvm_make_request(KVM_REQ_DEACTIVATE_FPU, vcpu);
        trace_kvm_fpu(0);
 }
 
 
        raw_spin_lock(&kvm->requests_lock);
        me = smp_processor_id();
        kvm_for_each_vcpu(i, vcpu, kvm) {
-               if (test_and_set_bit(req, &vcpu->requests))
+               if (kvm_make_check_request(req, vcpu))
                        continue;
                cpu = vcpu->cpu;
                if (cpus != NULL && cpu != -1 && cpu != me)
                prepare_to_wait(&vcpu->wq, &wait, TASK_INTERRUPTIBLE);
 
                if (kvm_arch_vcpu_runnable(vcpu)) {
-                       set_bit(KVM_REQ_UNHALT, &vcpu->requests);
+                       kvm_make_request(KVM_REQ_UNHALT, vcpu);
                        break;
                }
                if (kvm_cpu_has_pending_timer(vcpu))