return (vcpu_cp15(vcpu, c1_SCTLR) & 0b101) == 0b101;
 }
 
-static inline void __clean_dcache_guest_page(struct kvm_vcpu *vcpu,
-                                            kvm_pfn_t pfn,
-                                            unsigned long size)
+static inline void __clean_dcache_guest_page(kvm_pfn_t pfn, unsigned long size)
 {
        /*
         * Clean the dcache to the Point of Coherency.
        }
 }
 
-static inline void __invalidate_icache_guest_page(struct kvm_vcpu *vcpu,
-                                                 kvm_pfn_t pfn,
+static inline void __invalidate_icache_guest_page(kvm_pfn_t pfn,
                                                  unsigned long size)
 {
        u32 iclsz;
 
        return (vcpu_sys_reg(vcpu, SCTLR_EL1) & 0b101) == 0b101;
 }
 
-static inline void __clean_dcache_guest_page(struct kvm_vcpu *vcpu,
-                                            kvm_pfn_t pfn,
-                                            unsigned long size)
+static inline void __clean_dcache_guest_page(kvm_pfn_t pfn, unsigned long size)
 {
        void *va = page_address(pfn_to_page(pfn));
 
        kvm_flush_dcache_to_poc(va, size);
 }
 
-static inline void __invalidate_icache_guest_page(struct kvm_vcpu *vcpu,
-                                                 kvm_pfn_t pfn,
+static inline void __invalidate_icache_guest_page(kvm_pfn_t pfn,
                                                  unsigned long size)
 {
        if (icache_is_aliasing()) {
 
        kvm_mmu_write_protect_pt_masked(kvm, slot, gfn_offset, mask);
 }
 
-static void clean_dcache_guest_page(struct kvm_vcpu *vcpu, kvm_pfn_t pfn,
-                                   unsigned long size)
+static void clean_dcache_guest_page(kvm_pfn_t pfn, unsigned long size)
 {
-       __clean_dcache_guest_page(vcpu, pfn, size);
+       __clean_dcache_guest_page(pfn, size);
 }
 
-static void invalidate_icache_guest_page(struct kvm_vcpu *vcpu, kvm_pfn_t pfn,
-                                        unsigned long size)
+static void invalidate_icache_guest_page(kvm_pfn_t pfn, unsigned long size)
 {
-       __invalidate_icache_guest_page(vcpu, pfn, size);
+       __invalidate_icache_guest_page(pfn, size);
 }
 
 static void kvm_send_hwpoison_signal(unsigned long address,
                }
 
                if (fault_status != FSC_PERM)
-                       clean_dcache_guest_page(vcpu, pfn, PMD_SIZE);
+                       clean_dcache_guest_page(pfn, PMD_SIZE);
 
                if (exec_fault) {
                        new_pmd = kvm_s2pmd_mkexec(new_pmd);
-                       invalidate_icache_guest_page(vcpu, pfn, PMD_SIZE);
+                       invalidate_icache_guest_page(pfn, PMD_SIZE);
                } else if (fault_status == FSC_PERM) {
                        /* Preserve execute if XN was already cleared */
                        if (stage2_is_exec(kvm, fault_ipa))
                }
 
                if (fault_status != FSC_PERM)
-                       clean_dcache_guest_page(vcpu, pfn, PAGE_SIZE);
+                       clean_dcache_guest_page(pfn, PAGE_SIZE);
 
                if (exec_fault) {
                        new_pte = kvm_s2pte_mkexec(new_pte);
-                       invalidate_icache_guest_page(vcpu, pfn, PAGE_SIZE);
+                       invalidate_icache_guest_page(pfn, PAGE_SIZE);
                } else if (fault_status == FSC_PERM) {
                        /* Preserve execute if XN was already cleared */
                        if (stage2_is_exec(kvm, fault_ipa))