#ifdef CONFIG_DYNAMIC_FTRACE_WITH_REGS
 #define ARCH_SUPPORTS_FTRACE_OPS 1
 #else
-#define MCOUNT_ADDR            ((unsigned long)function_nocfi(_mcount))
+#define MCOUNT_ADDR            ((unsigned long)_mcount)
 #endif
 
 /* The BL at the callsite's adjusted rec->ip */
 
                ttbr1 |= TTBR_CNP_BIT;
        }
 
-       replace_phys = (void *)__pa_symbol(function_nocfi(idmap_cpu_replace_ttbr1));
+       replace_phys = (void *)__pa_symbol(idmap_cpu_replace_ttbr1);
 
        __cpu_install_idmap(idmap);
        replace_phys(ttbr1);
 
         * that read this address need to convert this address to the
         * Boot-Loader's endianness before jumping.
         */
-       writeq_relaxed(__pa_symbol(function_nocfi(secondary_entry)),
+       writeq_relaxed(__pa_symbol(secondary_entry),
                       &mailbox->entry_point);
        writel_relaxed(cpu_entry->gic_cpu_id, &mailbox->cpu_id);
 
 
        if (arm64_use_ng_mappings)
                return;
 
-       remap_fn = (void *)__pa_symbol(function_nocfi(idmap_kpti_install_ng_mappings));
+       remap_fn = (void *)__pa_symbol(idmap_kpti_install_ng_mappings);
 
        if (!cpu) {
                alloc = __get_free_pages(GFP_ATOMIC | __GFP_ZERO, order);
 
        unsigned long pc;
        u32 new;
 
-       pc = (unsigned long)function_nocfi(ftrace_call);
+       pc = (unsigned long)ftrace_call;
        new = aarch64_insn_gen_branch_imm(pc, (unsigned long)func,
                                          AARCH64_INSN_BRANCH_LINK);
 
 
                typeof(cpu_soft_restart) *restart;
 
                cpu_install_idmap();
-               restart = (void *)__pa_symbol(function_nocfi(cpu_soft_restart));
+               restart = (void *)__pa_symbol(cpu_soft_restart);
                restart(is_hyp_nvhe(), kimage->start, kimage->arch.dtb_mem,
                        0, 0);
        } else {
 
 
 static int cpu_psci_cpu_boot(unsigned int cpu)
 {
-       phys_addr_t pa_secondary_entry = __pa_symbol(function_nocfi(secondary_entry));
+       phys_addr_t pa_secondary_entry = __pa_symbol(secondary_entry);
        int err = psci_ops.cpu_on(cpu_logical_map(cpu), pa_secondary_entry);
        if (err)
                pr_err("failed to boot CPU%d (%d)\n", cpu, err);
 
 static int smp_spin_table_cpu_prepare(unsigned int cpu)
 {
        __le64 __iomem *release_addr;
-       phys_addr_t pa_holding_pen = __pa_symbol(function_nocfi(secondary_holding_pen));
+       phys_addr_t pa_holding_pen = __pa_symbol(secondary_holding_pen);
 
        if (!cpu_release_addr[cpu])
                return -ENODEV;
 
 static int psci_suspend_finisher(unsigned long state)
 {
        u32 power_state = state;
-       phys_addr_t pa_cpu_resume = __pa_symbol(function_nocfi(cpu_resume));
+       phys_addr_t pa_cpu_resume = __pa_symbol(cpu_resume);
 
        return psci_ops.cpu_suspend(power_state, pa_cpu_resume);
 }
 
 static int psci_system_suspend(unsigned long unused)
 {
-       phys_addr_t pa_cpu_resume = __pa_symbol(function_nocfi(cpu_resume));
+       phys_addr_t pa_cpu_resume = __pa_symbol(cpu_resume);
 
        return invoke_psci_fn(PSCI_FN_NATIVE(1_0, SYSTEM_SUSPEND),
                              pa_cpu_resume, 0, 0);
 
 
        pr_info("attempting bad copy_to_user from kernel text: %px\n",
                vm_mmap);
-       if (copy_to_user((void __user *)user_addr, function_nocfi(vm_mmap),
+       if (copy_to_user((void __user *)user_addr, vm_mmap,
                         unconst + PAGE_SIZE)) {
                pr_warn("copy_to_user failed, but lacked Oops\n");
                goto free_user;
 
        __v;                                                            \
 })
 
-/*
- * With CONFIG_CFI_CLANG, the compiler replaces function addresses in
- * instrumented C code with jump table addresses. Architectures that
- * support CFI can define this macro to return the actual function address
- * when needed.
- */
-#ifndef function_nocfi
-#define function_nocfi(x) (x)
-#endif
-
 #endif /* __KERNEL__ */
 
 /*