struct esca_block *sca = vcpu->kvm->arch.sca;
 
                clear_bit_inv(vcpu->vcpu_id, (unsigned long *) sca->mcn);
-               if (sca->cpu[vcpu->vcpu_id].sda == (__u64) vcpu->arch.sie_block)
-                       sca->cpu[vcpu->vcpu_id].sda = 0;
+               sca->cpu[vcpu->vcpu_id].sda = 0;
        } else {
                struct bsca_block *sca = vcpu->kvm->arch.sca;
 
                clear_bit_inv(vcpu->vcpu_id, (unsigned long *) &sca->mcn);
-               if (sca->cpu[vcpu->vcpu_id].sda == (__u64) vcpu->arch.sie_block)
-                       sca->cpu[vcpu->vcpu_id].sda = 0;
+               sca->cpu[vcpu->vcpu_id].sda = 0;
        }
        read_unlock(&vcpu->kvm->arch.sca_lock);
 }
        if (kvm->arch.use_esca) {
                struct esca_block *sca = kvm->arch.sca;
 
-               if (!sca->cpu[id].sda)
-                       sca->cpu[id].sda = (__u64) vcpu->arch.sie_block;
+               sca->cpu[id].sda = (__u64) vcpu->arch.sie_block;
                vcpu->arch.sie_block->scaoh = (__u32)(((__u64)sca) >> 32);
                vcpu->arch.sie_block->scaol = (__u32)(__u64)sca & ~0x3fU;
                vcpu->arch.sie_block->ecb2 |= 0x04U;
        } else {
                struct bsca_block *sca = kvm->arch.sca;
 
-               if (!sca->cpu[id].sda)
-                       sca->cpu[id].sda = (__u64) vcpu->arch.sie_block;
+               sca->cpu[id].sda = (__u64) vcpu->arch.sie_block;
                vcpu->arch.sie_block->scaoh = (__u32)(((__u64)sca) >> 32);
                vcpu->arch.sie_block->scaol = (__u32)(__u64)sca;
                set_bit_inv(id, (unsigned long *) &sca->mcn);