return 0;
 }
 
+#define GIC_LPI_MAX_INTID      ((1 << INTERRUPT_ID_BITS_ITS) - 1)
+
 /*
  * Create a snapshot of the current LPIs targeting @vcpu, so that we can
  * enumerate those LPIs without holding any lock.
 int vgic_copy_lpi_list(struct kvm *kvm, struct kvm_vcpu *vcpu, u32 **intid_ptr)
 {
        struct vgic_dist *dist = &kvm->arch.vgic;
+       XA_STATE(xas, &dist->lpi_xa, GIC_LPI_OFFSET);
        struct vgic_irq *irq;
        unsigned long flags;
        u32 *intids;
                return -ENOMEM;
 
        raw_spin_lock_irqsave(&dist->lpi_list_lock, flags);
-       list_for_each_entry(irq, &dist->lpi_list_head, lpi_list) {
+       rcu_read_lock();
+
+       xas_for_each(&xas, irq, GIC_LPI_MAX_INTID) {
                if (i == irq_count)
                        break;
                /* We don't need to "get" the IRQ, as we hold the list lock. */
                        continue;
                intids[i++] = irq->intid;
        }
+
+       rcu_read_unlock();
        raw_spin_unlock_irqrestore(&dist->lpi_list_lock, flags);
 
        *intid_ptr = intids;