#else  /* !__ASSEMBLY__ */
 
 #ifdef CONFIG_ARC_HAS_ICCM
-#define __arcfp_code __section(.text.arcfp)
+#define __arcfp_code __section(".text.arcfp")
 #else
-#define __arcfp_code __section(.text)
+#define __arcfp_code __section(".text")
 #endif
 
 #ifdef CONFIG_ARC_HAS_DCCM
-#define __arcfp_data __section(.data.arcfp)
+#define __arcfp_data __section(".data.arcfp")
 #else
-#define __arcfp_data __section(.data)
+#define __arcfp_data __section(".data")
 #endif
 
 #endif /* __ASSEMBLY__ */
 
  */
 #define MACHINE_START(_type, _name)                    \
 static const struct machine_desc __mach_desc_##_type   \
-__used __section(.arch.info.init) = {                  \
+__used __section(".arch.info.init") = {                        \
        .name           = _name,
 
 #define MACHINE_END                            \
 
 #include <asm/io.h>
 #include <asm/mach_desc.h>
 
-int arc_hsdk_axi_dmac_coherent __section(.data) = 0;
+int arc_hsdk_axi_dmac_coherent __section(".data") = 0;
 
 #define ARC_CCM_UNUSED_ADDR    0x60000000
 
 
 #define ARCH_SLAB_MINALIGN 8
 #endif
 
-#define __read_mostly __attribute__((__section__(".data..read_mostly")))
+#define __read_mostly __section(".data..read_mostly")
 
 #endif
 
 
 #define CPUIDLE_METHOD_OF_DECLARE(name, _method, _ops)                 \
        static const struct of_cpuidle_method __cpuidle_method_of_table_##name \
-       __used __section(__cpuidle_method_of_table)                     \
+       __used __section("__cpuidle_method_of_table")                   \
        = { .method = _method, .ops = _ops }
 
 extern int arm_cpuidle_suspend(int index);
 
 #include <linux/pgtable.h>
 
 /* Tag a function as requiring to be executed via an identity mapping. */
-#define __idmap __section(.idmap.text) noinline notrace
+#define __idmap __section(".idmap.text") noinline notrace
 
 extern pgd_t *idmap_pgd;
 
 
 #define MACHINE_START(_type,_name)                     \
 static const struct machine_desc __mach_desc_##_type   \
  __used                                                        \
- __attribute__((__section__(".arch.info.init"))) = {   \
+ __section(".arch.info.init") = {                      \
        .nr             = MACH_TYPE_##_type,            \
        .name           = _name,
 
 #define DT_MACHINE_START(_name, _namestr)              \
 static const struct machine_desc __mach_desc_##_name   \
  __used                                                        \
- __attribute__((__section__(".arch.info.init"))) = {   \
+ __section(".arch.info.init") = {                      \
        .nr             = ~0,                           \
        .name           = _namestr,
 
 
 #include <uapi/asm/setup.h>
 
 
-#define __tag __used __attribute__((__section__(".taglist.init")))
+#define __tag __used __section(".taglist.init")
 #define __tagtable(tag, fn) \
 static const struct tagtable __tagtable_##fn __tag = { tag, fn }
 
 
 
 #define CPU_METHOD_OF_DECLARE(name, _method, _ops)                     \
        static const struct of_cpu_method __cpu_method_of_table_##name  \
-               __used __section(__cpu_method_of_table)                 \
+               __used __section("__cpu_method_of_table")               \
                = { .method = _method, .ops = _ops }
 /*
  * set platform specific SMP operations
 
 #include <linux/compiler.h>
 
 /* Tag variables with this */
-#define __tcmdata __section(.tcm.data)
+#define __tcmdata __section(".tcm.data")
 /* Tag constants with this */
-#define __tcmconst __section(.tcm.rodata)
+#define __tcmconst __section(".tcm.rodata")
 /* Tag functions inside TCM called from outside TCM with this */
-#define __tcmfunc __attribute__((long_call)) __section(.tcm.text) noinline
+#define __tcmfunc __attribute__((long_call)) __section(".tcm.text") noinline
 /* Tag function inside TCM called from inside TCM  with this */
-#define __tcmlocalfunc __section(.tcm.text)
+#define __tcmlocalfunc __section(".tcm.text")
 
 void *tcm_alloc(size_t len);
 void tcm_free(void *addr, size_t len);
 
 extern struct of_cpuidle_method __cpuidle_method_of_table[];
 
 static const struct of_cpuidle_method __cpuidle_method_of_table_sentinel
-       __used __section(__cpuidle_method_of_table_end);
+       __used __section("__cpuidle_method_of_table_end");
 
 static struct cpuidle_ops cpuidle_ops[NR_CPUS] __ro_after_init;
 
 
 extern struct of_cpu_method __cpu_method_of_table[];
 
 static const struct of_cpu_method __cpu_method_of_table_sentinel
-       __used __section(__cpu_method_of_table_end);
+       __used __section("__cpu_method_of_table_end");
 
 
 static int __init set_smp_ops_by_method(struct device_node *node)
 
        return (read_cpuid_cachetype() >> CTR_CWG_SHIFT) & CTR_CWG_MASK;
 }
 
-#define __read_mostly __section(.data..read_mostly)
+#define __read_mostly __section(".data..read_mostly")
 
 static inline int cache_line_size_of_cpu(void)
 {
 
 }
 
 /* we will fill this structure from the stub, so don't put it in .bss */
-struct screen_info screen_info __section(.data);
+struct screen_info screen_info __section(".data");
 
 int __init efi_create_mapping(struct mm_struct *mm, efi_memory_desc_t *md)
 {
 
 #include <asm/smp_plat.h>
 
 extern void secondary_holding_pen(void);
-volatile unsigned long __section(.mmuoff.data.read)
+volatile unsigned long __section(".mmuoff.data.read")
 secondary_holding_pen_release = INVALID_HWID;
 
 static phys_addr_t cpu_release_addr[NR_CPUS];
 
 u64 idmap_t0sz = TCR_T0SZ(VA_BITS);
 u64 idmap_ptrs_per_pgd = PTRS_PER_PGD;
 
-u64 __section(.mmuoff.data.write) vabits_actual;
+u64 __section(".mmuoff.data.write") vabits_actual;
 EXPORT_SYMBOL(vabits_actual);
 
 u64 kimage_voffset __ro_after_init;
 
 #include <linux/compiler.h>
 
 /* Tag variables with this */
-#define __tcmdata __section(.tcm.data)
+#define __tcmdata __section(".tcm.data")
 /* Tag constants with this */
-#define __tcmconst __section(.tcm.rodata)
+#define __tcmconst __section(".tcm.rodata")
 /* Tag functions inside TCM called from outside TCM with this */
-#define __tcmfunc __section(.tcm.text) noinline
+#define __tcmfunc __section(".tcm.text") noinline
 /* Tag function inside TCM called from inside TCM  with this */
-#define __tcmlocalfunc __section(.tcm.text)
+#define __tcmlocalfunc __section(".tcm.text")
 
 void *tcm_alloc(size_t len);
 void tcm_free(void *addr, size_t len);
 
 # define SMP_CACHE_BYTES       (1 << 3)
 #endif
 
-#define __read_mostly __attribute__((__section__(".data..read_mostly")))
+#define __read_mostly __section(".data..read_mostly")
 
 #endif /* _ASM_IA64_CACHE_H */
 
  * ASM code. Default position is BSS section which is cleared
  * in machine_early_init().
  */
-char cmd_line[COMMAND_LINE_SIZE] __attribute__ ((section(".data")));
+char cmd_line[COMMAND_LINE_SIZE] __section(".data");
 
 void __init setup_arch(char **cmdline_p)
 {
 
 #define L1_CACHE_SHIFT         CONFIG_MIPS_L1_CACHE_SHIFT
 #define L1_CACHE_BYTES         (1 << L1_CACHE_SHIFT)
 
-#define __read_mostly __attribute__((__section__(".data..read_mostly")))
+#define __read_mostly __section(".data..read_mostly")
 
 #endif /* _ASM_CACHE_H */
 
 
 #define MIPS_MACHINE(name)                                             \
        static const struct mips_machine __mips_mach_##name             \
-               __used __section(.mips.machines.init)
+               __used __section(".mips.machines.init")
 
 #define for_each_mips_machine(mach)                                    \
        for ((mach) = (struct mips_machine *)&__mips_machines_start;    \
 
 #include <asm/prom.h>
 
 #ifdef CONFIG_MIPS_ELF_APPENDED_DTB
-const char __section(.appended_dtb) __appended_dtb[0x100000];
+const char __section(".appended_dtb") __appended_dtb[0x100000];
 #endif /* CONFIG_MIPS_ELF_APPENDED_DTB */
 
 struct cpuinfo_mips cpu_data[NR_CPUS] __read_mostly;
 
  * size, and waste space.  So we place it in its own section and align
  * it in the linker script.
  */
-pgd_t swapper_pg_dir[PTRS_PER_PGD] __section(.bss..swapper_pg_dir);
+pgd_t swapper_pg_dir[PTRS_PER_PGD] __section(".bss..swapper_pg_dir");
 #ifndef __PAGETABLE_PUD_FOLDED
 pud_t invalid_pud_table[PTRS_PER_PUD] __page_aligned_bss;
 #endif
 
 
 #define ARCH_DMA_MINALIGN      L1_CACHE_BYTES
 
-#define __read_mostly __section(.data..read_mostly)
+#define __read_mostly __section(".data..read_mostly")
 
 void parisc_cache_init(void);  /* initializes cache-flushing */
 void disable_sr_hashing_asm(int); /* low level support for above */
 
 })
 
 #ifdef CONFIG_SMP
-# define __lock_aligned __section(.data..lock_aligned)
+# define __lock_aligned __section(".data..lock_aligned")
 #endif
 
 #endif /* __PARISC_LDCW_H */
 
 #include <asm/ftrace.h>
 #include <asm/patch.h>
 
-#define __hot __attribute__ ((__section__ (".text.hot")))
+#define __hot __section(".text.hot")
 
 #ifdef CONFIG_FUNCTION_GRAPH_TRACER
 /*
 
  * guarantee that global objects will be laid out in memory in the same order
  * as the order of declaration, so put these in different sections and use
  * the linker script to order them. */
-pmd_t pmd0[PTRS_PER_PMD] __attribute__ ((__section__ (".data..vm0.pmd"), aligned(PAGE_SIZE)));
+pmd_t pmd0[PTRS_PER_PMD] __section(".data..vm0.pmd") __attribute__ ((aligned(PAGE_SIZE)));
 #endif
 
-pgd_t swapper_pg_dir[PTRS_PER_PGD] __attribute__ ((__section__ (".data..vm0.pgd"), aligned(PAGE_SIZE)));
-pte_t pg0[PT_INITIAL * PTRS_PER_PTE] __attribute__ ((__section__ (".data..vm0.pte"), aligned(PAGE_SIZE)));
+pgd_t swapper_pg_dir[PTRS_PER_PGD] __section(".data..vm0.pgd") __attribute__ ((aligned(PAGE_SIZE)));
+pte_t pg0[PT_INITIAL * PTRS_PER_PTE] __section(".data..vm0.pte") __attribute__ ((aligned(PAGE_SIZE)));
 
 static struct resource data_resource = {
        .name   = "Kernel data",
 
 
 #endif
 
-#define __read_mostly __section(.data..read_mostly)
+#define __read_mostly __section(".data..read_mostly")
 
 #ifdef CONFIG_PPC_BOOK3S_32
 extern long _get_L2CR(void);
 
 extern struct machdep_calls ppc_md;
 extern struct machdep_calls *machine_id;
 
-#define __machine_desc __attribute__ ((__section__ (".machine.desc")))
+#define __machine_desc __section(".machine.desc")
 
 #define define_machine(name)                                   \
        extern struct machdep_calls mach_##name;                \
 
 static void scrollscreen(void);
 #endif
 
-#define __force_data __section(.data)
+#define __force_data __section(".data")
 
 static int g_loc_X __force_data;
 static int g_loc_Y __force_data;
 
 #include <linux/linux_logo.h>
 
 /* All of prom_init bss lives here */
-#define __prombss __section(.bss.prominit)
+#define __prombss __section(".bss.prominit")
 
 /*
  * Eventually bump that one up
 
 #ifdef CONFIG_BUG
 
 #define WARN_ON_ONCE_RM(condition)     ({                      \
-       static bool __section(.data.unlikely) __warned;         \
+       static bool __section(".data.unlikely") __warned;       \
        int __ret_warn_once = !!(condition);                    \
                                                                \
        if (unlikely(__ret_warn_once && !__warned)) {           \
 
 
 #define SOC_EARLY_INIT_DECLARE(name, compat, fn)                       \
        static const struct of_device_id __soc_early_init__##name       \
-               __used __section(__soc_early_init_table)                \
+               __used __section("__soc_early_init_table")              \
                 = { .compatible = compat, .data = fn  }
 
 void soc_early_init(void);
        }                                                               \
                                                                        \
        static const struct soc_builtin_dtb __soc_builtin_dtb__##name   \
-               __used __section(__soc_builtin_dtb_table) =             \
+               __used __section("__soc_builtin_dtb_table") =           \
        {                                                               \
                .vendor_id = vendor,                                    \
                .arch_id   = arch,                                      \
 
 
 const struct cpu_operations *cpu_ops[NR_CPUS] __ro_after_init;
 
-void *__cpu_up_stack_pointer[NR_CPUS] __section(.data);
-void *__cpu_up_task_pointer[NR_CPUS] __section(.data);
+void *__cpu_up_stack_pointer[NR_CPUS] __section(".data");
+void *__cpu_up_task_pointer[NR_CPUS] __section(".data");
 
 extern const struct cpu_operations cpu_ops_sbi;
 extern const struct cpu_operations cpu_ops_spinwait;
 
 #include "head.h"
 
 #if defined(CONFIG_DUMMY_CONSOLE) || defined(CONFIG_EFI)
-struct screen_info screen_info __section(.data) = {
+struct screen_info screen_info __section(".data") = {
        .orig_video_lines       = 30,
        .orig_video_cols        = 80,
        .orig_video_mode        = 0,
  * This is used before the kernel initializes the BSS so it can't be in the
  * BSS.
  */
-atomic_t hart_lottery __section(.sdata);
+atomic_t hart_lottery __section(".sdata");
 unsigned long boot_cpu_hartid;
 static DEFINE_PER_CPU(struct cpu, cpu_devices);
 
 
        .diag0c = _diag0c_dma,
        .diag308_reset = _diag308_reset_dma
 };
-static struct diag210 _diag210_tmp_dma __section(.dma.data);
+static struct diag210 _diag210_tmp_dma __section(".dma.data");
 struct diag210 *__bootdata_preserved(__diag210_tmp_dma) = &_diag210_tmp_dma;
 
 void error(char *x)
 
 #define L1_CACHE_SHIFT     8
 #define NET_SKB_PAD       32
 
-#define __read_mostly __section(.data..read_mostly)
+#define __read_mostly __section(".data..read_mostly")
 
 #endif
 
  * final .boot.data section, which should be identical in the decompressor and
  * the decompressed kernel (that is checked during the build).
  */
-#define __bootdata(var) __section(.boot.data.var) var
+#define __bootdata(var) __section(".boot.data.var") var
 
 /*
  * .boot.preserved.data is similar to .boot.data, but it is not part of the
  * .init section and thus will be preserved for later use in the decompressed
  * kernel.
  */
-#define __bootdata_preserved(var) __section(.boot.preserved.data.var) var
+#define __bootdata_preserved(var) __section(".boot.preserved.data.var") var
 
 extern unsigned long __sdma, __edma;
 extern unsigned long __stext_dma, __etext_dma;
 
 #include <asm/uv.h>
 #include <linux/virtio_config.h>
 
-pgd_t swapper_pg_dir[PTRS_PER_PGD] __section(.bss..swapper_pg_dir);
+pgd_t swapper_pg_dir[PTRS_PER_PGD] __section(".bss..swapper_pg_dir");
 
 unsigned long empty_zero_page, zero_page_mask;
 EXPORT_SYMBOL(empty_zero_page);
 
 
 extern const struct of_cpu_method __cpu_method_of_table[];
 const struct of_cpu_method __cpu_method_of_table_sentinel
-       __section(__cpu_method_of_table_end);
+       __section("__cpu_method_of_table_end");
 
 static void sh_of_smp_probe(void)
 {
 
 
 #define L1_CACHE_BYTES         (1 << L1_CACHE_SHIFT)
 
-#define __read_mostly __attribute__((__section__(".data..read_mostly")))
+#define __read_mostly __section(".data..read_mostly")
 
 #ifndef __ASSEMBLY__
 struct cache_info {
 
 #define get_system_type()      sh_mv.mv_name
 
 #define __initmv \
-       __used __section(.machvec.init)
+       __used __section(".machvec.init")
 
 #endif /* _ASM_SH_MACHVEC_H */
 
 
 #define CPU_METHOD_OF_DECLARE(name, _method, _ops)                     \
        static const struct of_cpu_method __cpu_method_of_table_##name  \
-               __used __section(__cpu_method_of_table)                 \
+               __used __section("__cpu_method_of_table")               \
                = { .method = _method, .ops = _ops }
 
 #else
 
 
 #define SMP_CACHE_BYTES (1 << SMP_CACHE_BYTES_SHIFT)
 
-#define __read_mostly __attribute__((__section__(".data..read_mostly")))
+#define __read_mostly __section(".data..read_mostly")
 
 #endif /* !(_SPARC_CACHE_H) */
 
 static void draw_byte_16(unsigned char *bits, unsigned int *base, int rb);
 static void draw_byte_8(unsigned char *bits, unsigned int *base, int rb);
 
-#define __force_data __attribute__((__section__(".data")))
+#define __force_data __section(".data")
 
 static int g_loc_X __force_data;
 static int g_loc_Y __force_data;
 
 
 /* These are for everybody (although not all archs will actually
    discard it in modules) */
-#define __init         __section(.init.text)
-#define __initdata     __section(.init.data)
-#define __exitdata     __section(.exit.data)
-#define __exit_call    __used __section(.exitcall.exit)
+#define __init         __section(".init.text")
+#define __initdata     __section(".init.data")
+#define __exitdata     __section(".exit.data")
+#define __exit_call    __used __section(".exitcall.exit")
 
 #ifdef MODULE
-#define __exit         __section(.exit.text)
+#define __exit         __section(".exit.text")
 #else
-#define __exit         __used __section(.exit.text)
+#define __exit         __used __section(".exit.text")
 #endif
 
 #endif
  * Mark functions and data as being only used at initialization
  * or exit time.
  */
-#define __uml_init_setup       __used __section(.uml.setup.init)
-#define __uml_setup_help       __used __section(.uml.help.init)
-#define __uml_postsetup_call   __used __section(.uml.postsetup.init)
-#define __uml_exit_call                __used __section(.uml.exitcall.exit)
+#define __uml_init_setup       __used __section(".uml.setup.init")
+#define __uml_setup_help       __used __section(".uml.help.init")
+#define __uml_postsetup_call   __used __section(".uml.postsetup.init")
+#define __uml_exit_call                __used __section(".uml.exitcall.exit")
 
 #ifdef __UM_HOST__
 
 
 #define __exitcall(fn) static exitcall_t __exitcall_##fn __exit_call = fn
 
-#define __init_call    __used __section(.initcall.init)
+#define __init_call    __used __section(".initcall.init")
 
 #endif
 
 
  * on some systems.
  */
 
-void __attribute__ ((__section__ (".__syscall_stub")))
+void __section(".__syscall_stub")
 stub_clone_handler(void)
 {
        struct stub_data *data = (struct stub_data *) STUB_DATA;
 
 };
 
 union thread_union cpu0_irqstack
-       __attribute__((__section__(".data..init_irqstack"))) =
+       __section(".data..init_irqstack") =
                { .thread_info = INIT_THREAD_INFO(init_task) };
 
 /* Changed in setup_arch, which is called in early boot */
 
 
 #ifdef CONFIG_X86_5LEVEL
 /* __pgtable_l5_enabled needs to be in .data to avoid being cleared along with .bss */
-unsigned int __section(.data) __pgtable_l5_enabled;
-unsigned int __section(.data) pgdir_shift = 39;
-unsigned int __section(.data) ptrs_per_p4d = 1;
+unsigned int __section(".data") __pgtable_l5_enabled;
+unsigned int __section(".data") pgdir_shift = 39;
+unsigned int __section(".data") ptrs_per_p4d = 1;
 #endif
 
 struct paging_config {
  * Avoid putting the pointer into .bss as it will be cleared between
  * paging_prepare() and extract_kernel().
  */
-unsigned long *trampoline_32bit __section(.data);
+unsigned long *trampoline_32bit __section(".data");
 
 extern struct boot_params *boot_params;
 int cmdline_find_option_bool(const char *option);
 
  * error during initialization.
  */
 
-static void __attribute__((section(".inittext"))) serial_putchar(int ch)
+static void __section(".inittext") serial_putchar(int ch)
 {
        unsigned timeout = 0xffff;
 
        outb(ch, early_serial_base + TXR);
 }
 
-static void __attribute__((section(".inittext"))) bios_putchar(int ch)
+static void __section(".inittext") bios_putchar(int ch)
 {
        struct biosregs ireg;
 
        intcall(0x10, &ireg, NULL);
 }
 
-void __attribute__((section(".inittext"))) putchar(int ch)
+void __section(".inittext") putchar(int ch)
 {
        if (ch == '\n')
                putchar('\r');  /* \n -> \r\n */
                serial_putchar(ch);
 }
 
-void __attribute__((section(".inittext"))) puts(const char *str)
+void __section(".inittext") puts(const char *str)
 {
        while (*str)
                putchar(*str++);
 
        u16 xmode_n;            /* Size of unprobed mode range */
 };
 
-#define __videocard struct card_info __attribute__((used,section(".videocards")))
+#define __videocard struct card_info __section(".videocards") __attribute__((used))
 extern struct card_info video_cards[], video_cards_end[];
 
 int mode_defined(u16 mode);    /* video.c */
 
 #define apic_driver(sym)                                       \
        static const struct apic *__apicdrivers_##sym __used            \
        __aligned(sizeof(struct apic *))                        \
-       __section(.apicdrivers) = { &sym }
+       __section(".apicdrivers") = { &sym }
 
 #define apic_drivers(sym1, sym2)                                       \
        static struct apic *__apicdrivers_##sym1##sym2[2] __used        \
        __aligned(sizeof(struct apic *))                                \
-       __section(.apicdrivers) = { &sym1, &sym2 }
+       __section(".apicdrivers") = { &sym1, &sym2 }
 
 extern struct apic *__apicdrivers[], *__apicdrivers_end[];
 
 
 #define L1_CACHE_SHIFT (CONFIG_X86_L1_CACHE_SHIFT)
 #define L1_CACHE_BYTES (1 << L1_CACHE_SHIFT)
 
-#define __read_mostly __attribute__((__section__(".data..read_mostly")))
+#define __read_mostly __section(".data..read_mostly")
 
 #define INTERNODE_CACHE_SHIFT CONFIG_X86_INTERNODE_CACHE_SHIFT
 #define INTERNODE_CACHE_BYTES (1 << INTERNODE_CACHE_SHIFT)
 
 
 #define sfi_device(i)                                                          \
        static const struct devs_id *const __intel_mid_sfi_##i##_dev __used     \
-       __attribute__((__section__(".x86_intel_mid_dev.init"))) = &i
+       __section(".x86_intel_mid_dev.init") = &i
 
 /**
 * struct mid_sd_board_info - template for SD device creation
 
 #include <asm/nospec-branch.h>
 
 /* Provide __cpuidle; we can't safely include <linux/cpu.h> */
-#define __cpuidle __attribute__((__section__(".cpuidle.text")))
+#define __cpuidle __section(".cpuidle.text")
 
 /*
  * Interrupt control:
 
 bool sev_active(void);
 bool sev_es_active(void);
 
-#define __bss_decrypted __attribute__((__section__(".bss..decrypted")))
+#define __bss_decrypted __section(".bss..decrypted")
 
 #else  /* !CONFIG_AMD_MEM_ENCRYPT */
 
 
  * executable.)
  */
 #define RESERVE_BRK(name,sz)                                           \
-       static void __section(.discard.text) __used notrace             \
+       static void __section(".discard.text") __used notrace           \
        __brk_reservation_fn_##name##__(void) {                         \
                asm volatile (                                          \
                        ".pushsection .brk_reservation,\"aw\",@nobits;" \
 
 
 #define cpu_dev_register(cpu_devX) \
        static const struct cpu_dev *const __cpu_dev_##cpu_devX __used \
-       __attribute__((__section__(".x86_cpu_dev.init"))) = \
+       __section(".x86_cpu_dev.init") = \
        &cpu_devX;
 
 extern const struct cpu_dev *const __x86_cpu_dev_start[],
 
        .address = 0,
 };
 
-#define __head __section(.head.text)
+#define __head __section(".head.text")
 
 static void __head *fixup_pointer(void *ptr, unsigned long physaddr)
 {
 
  * reside in the .data section so as not to be zeroed out when the .bss
  * section is later cleared.
  */
-u64 sme_me_mask __section(.data) = 0;
-u64 sev_status __section(.data) = 0;
+u64 sme_me_mask __section(".data") = 0;
+u64 sev_status __section(".data") = 0;
 EXPORT_SYMBOL(sme_me_mask);
 DEFINE_STATIC_KEY_FALSE(sev_enable_key);
 EXPORT_SYMBOL_GPL(sev_enable_key);
 
-bool sev_enabled __section(.data);
+bool sev_enabled __section(".data");
 
 /* Buffer used for early in-place encryption by BSP, no locking needed */
 static char sme_early_buffer[PAGE_SIZE] __initdata __aligned(PAGE_SIZE);
 
  * section is 2MB aligned to allow for simple pagetable setup using only
  * PMD entries (see vmlinux.lds.S).
  */
-static char sme_workarea[2 * PMD_PAGE_SIZE] __section(.init.scratch);
+static char sme_workarea[2 * PMD_PAGE_SIZE] __section(".init.scratch");
 
 static char sme_cmdline_arg[] __initdata = "mem_encrypt";
 static char sme_cmdline_on[]  __initdata = "on";
 
  * pvh_bootparams and pvh_start_info need to live in the data segment since
  * they are used after startup_{32|64}, which clear .bss, are invoked.
  */
-struct boot_params pvh_bootparams __attribute__((section(".data")));
-struct hvm_start_info pvh_start_info __attribute__((section(".data")));
+struct boot_params pvh_bootparams __section(".data");
+struct hvm_start_info pvh_start_info __section(".data");
 
 unsigned int pvh_start_info_sz = sizeof(pvh_start_info);
 
 
 
 #include "../boot/string.h"
 
-u8 purgatory_sha256_digest[SHA256_DIGEST_SIZE] __section(.kexec-purgatory);
+u8 purgatory_sha256_digest[SHA256_DIGEST_SIZE] __section(".kexec-purgatory");
 
-struct kexec_sha_region purgatory_sha_regions[KEXEC_SEGMENT_MAX] __section(.kexec-purgatory);
+struct kexec_sha_region purgatory_sha_regions[KEXEC_SEGMENT_MAX] __section(".kexec-purgatory");
 
 static int verify_sha256_digest(void)
 {
 
 #include <sysdep/mcontext.h>
 #include <sys/ucontext.h>
 
-void __attribute__ ((__section__ (".__syscall_stub")))
+void __section(".__syscall_stub")
 stub_segv_handler(int sig, siginfo_t *info, void *p)
 {
        ucontext_t *uc = p;
 
  * NB: needs to live in .data because it's used by xen_prepare_pvh which runs
  * before clearing the bss.
  */
-uint32_t xen_start_flags __attribute__((section(".data"))) = 0;
+uint32_t xen_start_flags __section(".data") = 0;
 EXPORT_SYMBOL(xen_start_flags);
 
 /*
 
  * The variable xen_pvh needs to live in the data segment since it is used
  * after startup_{32|64} is invoked, which will clear the .bss segment.
  */
-bool xen_pvh __attribute__((section(".data"))) = 0;
+bool xen_pvh __section(".data") = 0;
 
 void __init xen_pvh_init(struct boot_params *boot_params)
 {
 
 } tagtable_t;
 
 #define __tagtable(tag, fn) static tagtable_t __tagtable_##fn          \
-       __attribute__((used, section(".taglist"))) = { tag, fn }
+       __section(".taglist") __attribute__((used)) = { tag, fn }
 
 /* parse current tag */
 
 
 
 extern struct of_device_id __clk_of_table;
 static const struct of_device_id __clk_of_table_sentinel
-       __used __section(__clk_of_table_end);
+       __used __section("__clk_of_table_end");
 
 static LIST_HEAD(of_clk_providers);
 static DEFINE_MUTEX(of_clk_mutex);
 
 extern struct of_device_id __timer_of_table[];
 
 static const struct of_device_id __timer_of_table_sentinel
-       __used __section(__timer_of_table_end);
+       __used __section("__timer_of_table_end");
 
 void __init timer_probe(void)
 {
 
  * special section.
  */
 static const struct of_device_id
-irqchip_of_match_end __used __section(__irqchip_of_table_end);
+irqchip_of_match_end __used __section("__irqchip_of_table_end");
 
 extern struct of_device_id __irqchip_of_table[];
 
 
 }
 
 static const struct of_device_id __rmem_of_table_sentinel
-       __used __section(__reservedmem_of_table_end);
+       __used __section("__reservedmem_of_table_end");
 
 /**
  * __reserved_mem_init_node() - call region specific reserved memory init code
 
 
 #define THERMAL_TABLE_ENTRY(table, name)                       \
        static typeof(name) *__thermal_table_entry_##name       \
-       __used __section(__##table##_thermal_table) = &name
+       __used __section("__" #table "_thermal_table") = &name
 
 #define THERMAL_GOVERNOR_DECLARE(name) THERMAL_TABLE_ENTRY(governor, name)
 
 
 
 #define xfs_printk_once(func, dev, fmt, ...)                   \
 ({                                                             \
-       static bool __section(.data.once) __print_once;         \
+       static bool __section(".data.once") __print_once;       \
        bool __ret_print_once = !__print_once;                  \
                                                                \
        if (!__print_once) {                                    \
 
 
 #ifndef WARN_ON_ONCE
 #define WARN_ON_ONCE(condition)        ({                              \
-       static bool __section(.data.once) __warned;             \
+       static bool __section(".data.once") __warned;           \
        int __ret_warn_once = !!(condition);                    \
                                                                \
        if (unlikely(__ret_warn_once && !__warned)) {           \
 #endif
 
 #define WARN_ONCE(condition, format...)        ({                      \
-       static bool __section(.data.once) __warned;             \
+       static bool __section(".data.once") __warned;           \
        int __ret_warn_once = !!(condition);                    \
                                                                \
        if (unlikely(__ret_warn_once && !__warned)) {           \
 })
 
 #define WARN_TAINT_ONCE(condition, taint, format...)   ({      \
-       static bool __section(.data.once) __warned;             \
+       static bool __section(".data.once") __warned;           \
        int __ret_warn_once = !!(condition);                    \
                                                                \
        if (unlikely(__ret_warn_once && !__warned)) {           \
 
  */
 #define ALLOW_ERROR_INJECTION(fname, _etype)                           \
 static struct error_injection_entry __used                             \
-       __attribute__((__section__("_error_injection_whitelist")))      \
+       __section("_error_injection_whitelist")                         \
        _eil_addr_##fname = {                                           \
                .addr = (unsigned long)fname,                           \
                .etype = EI_ETYPE_##_etype,                             \
 
  */
 # define __NOKPROBE_SYMBOL(fname)                              \
 static unsigned long __used                                    \
-       __attribute__((__section__("_kprobe_blacklist")))       \
+       __section("_kprobe_blacklist")                          \
        _kbl_addr_##fname = (unsigned long)fname;
 # define NOKPROBE_SYMBOL(fname)        __NOKPROBE_SYMBOL(fname)
 /* Use this to forbid a kprobes attach on very low level functions */
-# define __kprobes     __attribute__((__section__(".kprobes.text")))
+# define __kprobes     __section(".kprobes.text")
 # define nokprobe_inline       __always_inline
 #else
 # define NOKPROBE_SYMBOL(fname)
 
        static struct kunit_suite *unique_array[] = { __VA_ARGS__, NULL };     \
        kunit_test_suites_for_module(unique_array);                            \
        static struct kunit_suite **unique_suites                              \
-       __used __section(.kunit_test_suites) = unique_array
+       __used __section(".kunit_test_suites") = unique_array
 
 /**
  * kunit_test_suites() - used to register one or more &struct kunit_suite
 
 #define ACPI_DECLARE_PROBE_ENTRY(table, name, table_id, subtable,      \
                                 valid, data, fn)                       \
        static const struct acpi_probe_entry __acpi_probe_##name        \
-               __used __section(__##table##_acpi_probe_table) = {      \
+               __used __section("__" #table "_acpi_probe_table") = {   \
                        .id = table_id,                                 \
                        .type = subtable,                               \
                        .subtable_valid = valid,                        \
 #define ACPI_DECLARE_SUBTABLE_PROBE_ENTRY(table, name, table_id,       \
                                          subtable, valid, data, fn)    \
        static const struct acpi_probe_entry __acpi_probe_##name        \
-               __used __section(__##table##_acpi_probe_table) = {      \
+               __used __section("__" #table "_acpi_probe_table") = {   \
                        .id = table_id,                                 \
                        .type = subtable,                               \
                        .subtable_valid = valid,                        \
 
  * but may get written to during init, so can't live in .rodata (via "const").
  */
 #ifndef __ro_after_init
-#define __ro_after_init __attribute__((__section__(".data..ro_after_init")))
+#define __ro_after_init __section(".data..ro_after_init")
 #endif
 
 #ifndef ____cacheline_aligned
 
                        long ______r;                                   \
                        static struct ftrace_likely_data                \
                                __aligned(4)                            \
-                               __section(_ftrace_annotated_branch)     \
+                               __section("_ftrace_annotated_branch")   \
                                ______f = {                             \
                                .data.func = __func__,                  \
                                .data.file = __FILE__,                  \
 #define __trace_if_value(cond) ({                      \
        static struct ftrace_branch_data                \
                __aligned(4)                            \
-               __section(_ftrace_branch)               \
+               __section("_ftrace_branch")             \
                __if_trace = {                          \
                        .func = __func__,               \
                        .file = __FILE__,               \
        ".popsection\n\t"
 
 /* Annotate a C jump table to allow objtool to follow the code flow */
-#define __annotate_jump_table __section(.rodata..c_jump_table)
+#define __annotate_jump_table __section(".rodata..c_jump_table")
 
 #else
 #define annotate_reachable()
  * visible to the compiler.
  */
 #define __ADDRESSABLE(sym) \
-       static void * __section(.discard.addressable) __used \
+       static void * __section(".discard.addressable") __used \
                __UNIQUE_ID(__PASTE(__addressable_,sym)) = (void *)&sym;
 
 /**
 
  *   gcc: https://gcc.gnu.org/onlinedocs/gcc/Common-Variable-Attributes.html#index-section-variable-attribute
  * clang: https://clang.llvm.org/docs/AttributeReference.html#section-declspec-allocate
  */
-#define __section(S)                    __attribute__((__section__(#S)))
+#define __section(section)              __attribute__((__section__(section)))
 
 /*
  *   gcc: https://gcc.gnu.org/onlinedocs/gcc/Common-Function-Attributes.html#index-unused-function-attribute
 
 void cpu_idle_poll_ctrl(bool enable);
 
 /* Attach to any functions which should be considered cpuidle. */
-#define __cpuidle      __attribute__((__section__(".cpuidle.text")))
+#define __cpuidle      __section(".cpuidle.text")
 
 bool cpu_in_idle(unsigned long pc);
 
 
 
 #define DEFINE_DYNAMIC_DEBUG_METADATA(name, fmt)               \
        static struct _ddebug  __aligned(8)                     \
-       __section(__dyndbg) name = {                            \
+       __section("__dyndbg") name = {                          \
                .modname = KBUILD_MODNAME,                      \
                .function = __func__,                           \
                .filename = __FILE__,                           \
 
  * discarded in the final link stage.
  */
 #define __ksym_marker(sym)     \
-       static int __ksym_marker_##sym[0] __section(.discard.ksym) __used
+       static int __ksym_marker_##sym[0] __section(".discard.ksym") __used
 
 #define __EXPORT_SYMBOL(sym, sec, ns)                                  \
        __ksym_marker(sym);                                             \
 
 
 #define DECLARE_BUILTIN_FIRMWARE_SIZE(name, blob, size)                             \
        static const struct builtin_fw __fw_concat(__builtin_fw,__COUNTER__) \
-       __used __section(.builtin_fw) = { name, blob, size }
+       __used __section(".builtin_fw") = { name, blob, size }
 
 #if defined(CONFIG_FW_LOADER) || (defined(CONFIG_FW_LOADER_MODULE) && defined(MODULE))
 int request_firmware(const struct firmware **fw, const char *name,
 
 
 /* These are for everybody (although not all archs will actually
    discard it in modules) */
-#define __init         __section(.init.text) __cold  __latent_entropy __noinitretpoline
-#define __initdata     __section(.init.data)
-#define __initconst    __section(.init.rodata)
-#define __exitdata     __section(.exit.data)
-#define __exit_call    __used __section(.exitcall.exit)
+#define __init         __section(".init.text") __cold  __latent_entropy __noinitretpoline
+#define __initdata     __section(".init.data")
+#define __initconst    __section(".init.rodata")
+#define __exitdata     __section(".exit.data")
+#define __exit_call    __used __section(".exitcall.exit")
 
 /*
  * modpost check for section mismatches during the kernel build.
  *
  * The markers follow same syntax rules as __init / __initdata.
  */
-#define __ref            __section(.ref.text) noinline
-#define __refdata        __section(.ref.data)
-#define __refconst       __section(.ref.rodata)
+#define __ref            __section(".ref.text") noinline
+#define __refdata        __section(".ref.data")
+#define __refconst       __section(".ref.rodata")
 
 #ifdef MODULE
 #define __exitused
 #define __exitused  __used
 #endif
 
-#define __exit          __section(.exit.text) __exitused __cold notrace
+#define __exit          __section(".exit.text") __exitused __cold notrace
 
 /* Used for MEMORY_HOTPLUG */
-#define __meminit        __section(.meminit.text) __cold notrace \
+#define __meminit        __section(".meminit.text") __cold notrace \
                                                  __latent_entropy
-#define __meminitdata    __section(.meminit.data)
-#define __meminitconst   __section(.meminit.rodata)
-#define __memexit        __section(.memexit.text) __exitused __cold notrace
-#define __memexitdata    __section(.memexit.data)
-#define __memexitconst   __section(.memexit.rodata)
+#define __meminitdata    __section(".meminit.data")
+#define __meminitconst   __section(".meminit.rodata")
+#define __memexit        __section(".memexit.text") __exitused __cold notrace
+#define __memexitdata    __section(".memexit.data")
+#define __memexitconst   __section(".memexit.rodata")
 
 /* For assembly routines */
 #define __HEAD         .section        ".head.text","ax"
        static const char __setup_str_##unique_id[] __initconst         \
                __aligned(1) = str;                                     \
        static struct obs_kernel_param __setup_##unique_id              \
-               __used __section(.init.setup)                           \
+               __used __section(".init.setup")                         \
                __attribute__((aligned((sizeof(long)))))                \
                = { __setup_str_##unique_id, fn, early }
 
 #endif
 
 /* Data marked not to be saved by software suspend */
-#define __nosavedata __section(.data..nosave)
+#define __nosavedata __section(".data..nosave")
 
 #ifdef MODULE
 #define __exit_p(x) x
 
 
 /* Attach to the init_task data structure for proper alignment */
 #ifdef CONFIG_ARCH_TASK_STRUCT_ON_STACK
-#define __init_task_data __attribute__((__section__(".data..init_task")))
+#define __init_task_data __section(".data..init_task")
 #else
 #define __init_task_data /**/
 #endif
 
 /* Attach to the thread_info data structure for proper alignment */
-#define __init_thread_info __attribute__((__section__(".data..init_thread_info")))
+#define __init_thread_info __section(".data..init_thread_info")
 
 #endif
 
  * We want to know which function is an entrypoint of a hardirq or a softirq.
  */
 #ifndef __irq_entry
-# define __irq_entry    __attribute__((__section__(".irqentry.text")))
+# define __irq_entry    __section(".irqentry.text")
 #endif
 
-#define __softirq_entry  __attribute__((__section__(".softirqentry.text")))
+#define __softirq_entry  __section(".softirqentry.text")
 
 #endif
 
 #define do_trace_printk(fmt, args...)                                  \
 do {                                                                   \
        static const char *trace_printk_fmt __used                      \
-               __attribute__((section("__trace_printk_fmt"))) =        \
+               __section("__trace_printk_fmt") =                       \
                __builtin_constant_p(fmt) ? fmt : NULL;                 \
                                                                        \
        __trace_printk_check_format(fmt, ##args);                       \
 
 #define trace_puts(str) ({                                             \
        static const char *trace_printk_fmt __used                      \
-               __attribute__((section("__trace_printk_fmt"))) =        \
+               __section("__trace_printk_fmt") =                       \
                __builtin_constant_p(str) ? str : NULL;                 \
                                                                        \
        if (__builtin_constant_p(str))                                  \
 do {                                                                   \
        if (__builtin_constant_p(fmt)) {                                \
                static const char *trace_printk_fmt __used              \
-                 __attribute__((section("__trace_printk_fmt"))) =      \
+                 __section("__trace_printk_fmt") =                     \
                        __builtin_constant_p(fmt) ? fmt : NULL;         \
                                                                        \
                __ftrace_vbprintk(_THIS_IP_, trace_printk_fmt, vargs);  \
 
                  __stringify(name))
 #endif
 
-#define __page_aligned_data    __section(.data..page_aligned) __aligned(PAGE_SIZE)
-#define __page_aligned_bss     __section(.bss..page_aligned) __aligned(PAGE_SIZE)
+#define __page_aligned_data    __section(".data..page_aligned") __aligned(PAGE_SIZE)
+#define __page_aligned_bss     __section(".bss..page_aligned") __aligned(PAGE_SIZE)
 
 /*
  * For assembly routines.
 
 
 #define DEFINE_LSM(lsm)                                                        \
        static struct lsm_info __lsm_##lsm                              \
-               __used __section(.lsm_info.init)                        \
+               __used __section(".lsm_info.init")                      \
                __aligned(sizeof(unsigned long))
 
 #define DEFINE_EARLY_LSM(lsm)                                          \
        static struct lsm_info __early_lsm_##lsm                        \
-               __used __section(.early_lsm_info.init)                  \
+               __used __section(".early_lsm_info.init")                \
                __aligned(sizeof(unsigned long))
 
 #ifdef CONFIG_SECURITY_SELINUX_DISABLE
 
                .version        = _version,                             \
        };                                                              \
        static const struct module_version_attribute                    \
-       __used __attribute__ ((__section__ ("__modver")))               \
+       __used __section("__modver")                                    \
        * __moduleparam_const __modver_attr = &___modver_attr
 #endif
 
 
 
 #define __MODULE_INFO(tag, name, info)                                   \
 static const char __UNIQUE_ID(name)[]                                    \
-  __used __attribute__((section(".modinfo"), unused, aligned(1)))        \
+  __used __section(".modinfo") __attribute__((unused, aligned(1)))       \
   = __MODULE_INFO_PREFIX __stringify(tag) "=" info
 
 #define __MODULE_PARM_TYPE(name, _type)                                          \
        static const char __param_str_##name[] = prefix #name;          \
        static struct kernel_param __moduleparam_const __param_##name   \
        __used                                                          \
-    __attribute__ ((unused,__section__ ("__param"),aligned(sizeof(void *)))) \
+    __section("__param") __attribute__ ((unused, aligned(sizeof(void *)))) \
        = { __param_str_##name, THIS_MODULE, ops,                       \
            VERIFY_OCTAL_PERMISSIONS(perm), level, flags, { arg } }
 
 
  * those functions so they get relocated to ram.
  */
 #ifdef CONFIG_XIP_KERNEL
-#define __xipram noinline __attribute__ ((__section__ (".xiptext")))
+#define __xipram noinline __section(".xiptext")
 #endif
 
 /*
 
  * For more information, see tools/objtool/Documentation/stack-validation.txt.
  */
 #define STACK_FRAME_NON_STANDARD(func) \
-       static void __used __section(.discard.func_stack_frame_non_standard) \
+       static void __used __section(".discard.func_stack_frame_non_standard") \
                *__func_stack_frame_non_standard_##func = func
 
 #else /* __ASSEMBLY__ */
 
 #if defined(CONFIG_OF) && !defined(MODULE)
 #define _OF_DECLARE(table, name, compat, fn, fn_type)                  \
        static const struct of_device_id __of_table_##name              \
-               __used __section(__##table##_of_table)                  \
+               __used __section("__" #table "_of_table")               \
                 = { .compatible = compat,                              \
                     .data = (fn == (fn_type)NULL) ? fn : fn  }
 #else
 
        PER_CPU_ATTRIBUTES
 
 #define __PCPU_DUMMY_ATTRS                                             \
-       __attribute__((section(".discard"), unused))
+       __section(".discard") __attribute__((unused))
 
 /*
  * s390 and alpha modules require percpu variables to be defined as
 
 #ifdef CONFIG_PRINTK
 #define printk_once(fmt, ...)                                  \
 ({                                                             \
-       static bool __section(.data.once) __print_once;         \
+       static bool __section(".data.once") __print_once;       \
        bool __ret_print_once = !__print_once;                  \
                                                                \
        if (!__print_once) {                                    \
 })
 #define printk_deferred_once(fmt, ...)                         \
 ({                                                             \
-       static bool __section(.data.once) __print_once;         \
+       static bool __section(".data.once") __print_once;       \
        bool __ret_print_once = !__print_once;                  \
                                                                \
        if (!__print_once) {                                    \
 
  */
 #define RCU_LOCKDEP_WARN(c, s)                                         \
        do {                                                            \
-               static bool __section(.data.unlikely) __warned;         \
+               static bool __section(".data.unlikely") __warned;       \
                if (debug_lockdep_rcu_enabled() && !__warned && (c)) {  \
                        __warned = true;                                \
                        lockdep_rcu_suspicious(__FILE__, __LINE__, s);  \
 
 #endif
 
 /* Attach to any functions which should be ignored in wchan output. */
-#define __sched                __attribute__((__section__(".sched.text")))
+#define __sched                __section(".sched.text")
 
 /* Linker adds these: start and end of __sched functions */
 extern char __sched_text_start[], __sched_text_end[];
 
                    .compatible = compat,                               \
                    .setup = fn  };                                     \
        static const struct earlycon_id EARLYCON_USED_OR_UNUSED         \
-               __section(__earlycon_table)                             \
+               __section("__earlycon_table")                           \
                * const __PASTE(__p, unique_id) = &unique_id
 
 #define OF_EARLYCON_DECLARE(_name, compat, fn)                         \
 
 #define LOCK_SECTION_END                        \
         ".previous\n\t"
 
-#define __lockfunc __attribute__((section(".spinlock.text")))
+#define __lockfunc __section(".spinlock.text")
 
 /*
  * Pull the arch_spinlock_t and arch_rwlock_t definitions:
 
                .flags                  = TRACE_EVENT_FL_CAP_ANY,       \
        };                                                              \
        static struct trace_event_call __used                           \
-         __attribute__((section("_ftrace_events")))                    \
+         __section("_ftrace_events")                                   \
         *__event_enter_##sname = &event_enter_##sname;
 
 #define SYSCALL_TRACE_EXIT_EVENT(sname)                                        \
                .flags                  = TRACE_EVENT_FL_CAP_ANY,       \
        };                                                              \
        static struct trace_event_call __used                           \
-         __attribute__((section("_ftrace_events")))                    \
+         __section("_ftrace_events")                                   \
        *__event_exit_##sname = &event_exit_##sname;
 
 #define SYSCALL_METADATA(sname, nb, ...)                       \
                .enter_fields   = LIST_HEAD_INIT(__syscall_meta_##sname.enter_fields), \
        };                                                      \
        static struct syscall_metadata __used                   \
-         __attribute__((section("__syscalls_metadata")))       \
+         __section("__syscalls_metadata")                      \
         *__p_syscall_meta_##sname = &__syscall_meta_##sname;
 
 static inline int is_syscall_trace_event(struct trace_event_call *tp_event)
 
        tracing_record_cmdline(current);                                \
        if (__builtin_constant_p(fmt)) {                                \
                static const char *trace_printk_fmt                     \
-                 __attribute__((section("__trace_printk_fmt"))) =      \
+                 __section("__trace_printk_fmt") =                     \
                        __builtin_constant_p(fmt) ? fmt : NULL;         \
                                                                        \
                __trace_bprintk(ip, trace_printk_fmt, ##args);          \
 
 
 #define __TRACEPOINT_ENTRY(name)                                        \
        static tracepoint_ptr_t __tracepoint_ptr_##name __used           \
-       __section(__tracepoints_ptrs) = &__tracepoint_##name
+       __section("__tracepoints_ptrs") = &__tracepoint_##name
 #endif
 
 #endif /* _LINUX_TRACEPOINT_H */
  */
 #define DEFINE_TRACE_FN(_name, _reg, _unreg, proto, args)              \
        static const char __tpstrtab_##_name[]                          \
-       __section(__tracepoints_strings) = #_name;                      \
+       __section("__tracepoints_strings") = #_name;                    \
        extern struct static_call_key STATIC_CALL_KEY(tp_func_##_name); \
        int __traceiter_##_name(void *__data, proto);                   \
        struct tracepoint __tracepoint_##_name  __used                  \
-       __section(__tracepoints) = {                                    \
+       __section("__tracepoints") = {                                  \
                .name = __tpstrtab_##_name,                             \
                .key = STATIC_KEY_INIT_FALSE,                           \
                .static_call_key = &STATIC_CALL_KEY(tp_func_##_name),   \
                static const char *___tp_str __tracepoint_string = str; \
                ___tp_str;                                              \
        })
-#define __tracepoint_string    __used __section(__tracepoint_str)
+#define __tracepoint_string    __used __section("__tracepoint_str")
 #else
 /*
  * tracepoint_string() is used to save the string address for userspace
 
        struct bpf_raw_event_map event;                                 \
        btf_trace_##call handler;                                       \
 } __bpf_trace_tp_map_##call __used                                     \
-__attribute__((section("__bpf_raw_tp_map"))) = {                       \
+__section("__bpf_raw_tp_map") = {                                      \
        .event = {                                                      \
                .tp             = &__tracepoint_##call,                 \
                .bpf_func       = __bpf_trace_##template,               \
 
                .eval_value = a                         \
        };                                              \
        static struct trace_eval_map __used             \
-       __attribute__((section("_ftrace_eval_map")))    \
+       __section("_ftrace_eval_map")                   \
        *TRACE_SYSTEM##_##a = &__##TRACE_SYSTEM##_##a
 
 #undef TRACE_DEFINE_SIZEOF
                .eval_value = sizeof(a)                 \
        };                                              \
        static struct trace_eval_map __used             \
-       __attribute__((section("_ftrace_eval_map")))    \
+       __section("_ftrace_eval_map")                   \
        *TRACE_SYSTEM##_##a = &__##TRACE_SYSTEM##_##a
 
 /*
  * // its only safe to use pointers when doing linker tricks to
  * // create an array.
  * static struct trace_event_call __used
- * __attribute__((section("_ftrace_events"))) *__event_<call> = &event_<call>;
+ * __section("_ftrace_events") *__event_<call> = &event_<call>;
  *
  */
 
        .flags                  = TRACE_EVENT_FL_TRACEPOINT,            \
 };                                                                     \
 static struct trace_event_call __used                                  \
-__attribute__((section("_ftrace_events"))) *__event_##call = &event_##call
+__section("_ftrace_events") *__event_##call = &event_##call
 
 #undef DEFINE_EVENT_PRINT
 #define DEFINE_EVENT_PRINT(template, call, proto, args, print)         \
        .flags                  = TRACE_EVENT_FL_TRACEPOINT,            \
 };                                                                     \
 static struct trace_event_call __used                                  \
-__attribute__((section("_ftrace_events"))) *__event_##call = &event_##call
+__section("_ftrace_events") *__event_##call = &event_##call
 
 #include TRACE_INCLUDE(TRACE_INCLUDE_FILE)
 
  * has one (eg: FRV).
  */
 extern const unsigned int kallsyms_num_syms
-__attribute__((weak, section(".rodata")));
+__section(".rodata") __attribute__((weak));
 
 extern const unsigned long kallsyms_relative_base
-__attribute__((weak, section(".rodata")));
+__section(".rodata") __attribute__((weak));
 
 extern const char kallsyms_token_table[] __weak;
 extern const u16 kallsyms_token_index[] __weak;
 
 }
 
 const struct sched_class dl_sched_class
-       __attribute__((section("__dl_sched_class"))) = {
+       __section("__dl_sched_class") = {
        .enqueue_task           = enqueue_task_dl,
        .dequeue_task           = dequeue_task_dl,
        .yield_task             = yield_task_dl,
 
  * All the scheduling class methods:
  */
 const struct sched_class fair_sched_class
-       __attribute__((section("__fair_sched_class"))) = {
+       __section("__fair_sched_class") = {
        .enqueue_task           = enqueue_task_fair,
        .dequeue_task           = dequeue_task_fair,
        .yield_task             = yield_task_fair,
 
  * Simple, special scheduling class for the per-CPU idle tasks:
  */
 const struct sched_class idle_sched_class
-       __attribute__((section("__idle_sched_class"))) = {
+       __section("__idle_sched_class") = {
        /* no enqueue/yield_task for idle tasks */
 
        /* dequeue is not valid, we print a debug message there: */
 
 }
 
 const struct sched_class rt_sched_class
-       __attribute__((section("__rt_sched_class"))) = {
+       __section("__rt_sched_class") = {
        .enqueue_task           = enqueue_task_rt,
        .dequeue_task           = dequeue_task_rt,
        .yield_task             = yield_task_rt,
 
  * Simple, special scheduling class for the per-CPU stop tasks:
  */
 const struct sched_class stop_sched_class
-       __attribute__((section("__stop_sched_class"))) = {
+       __section("__stop_sched_class") = {
 
        .enqueue_task           = enqueue_task_stop,
        .dequeue_task           = dequeue_task_stop,
 
 
 /* Use this for memory failure errors */
 #define MEM_FAIL(condition, fmt, ...) ({                       \
-       static bool __section(.data.once) __warned;             \
+       static bool __section(".data.once") __warned;           \
        int __ret_warn_once = !!(condition);                    \
                                                                \
        if (unlikely(__ret_warn_once && !__warned)) {           \
 
        .flags                  = TRACE_EVENT_FL_IGNORE_ENABLE,         \
 };                                                                     \
 static struct trace_event_call __used                                          \
-__attribute__((section("_ftrace_events"))) *__event_##call = &event_##call;
+__section("_ftrace_events") *__event_##call = &event_##call;
 
 #undef FTRACE_ENTRY
 #define FTRACE_ENTRY(call, struct_name, etype, tstruct, print)         \
 
        buf_printf(b, "MODULE_INFO(name, KBUILD_MODNAME);\n");
        buf_printf(b, "\n");
        buf_printf(b, "__visible struct module __this_module\n");
-       buf_printf(b, "__section(.gnu.linkonce.this_module) = {\n");
+       buf_printf(b, "__section(\".gnu.linkonce.this_module\") = {\n");
        buf_printf(b, "\t.name = KBUILD_MODNAME,\n");
        if (mod->has_init)
                buf_printf(b, "\t.init = init_module,\n");
 
        buf_printf(b, "\n");
        buf_printf(b, "static const struct modversion_info ____versions[]\n");
-       buf_printf(b, "__used __section(__versions) = {\n");
+       buf_printf(b, "__used __section(\"__versions\") = {\n");
 
        for (s = mod->unres; s; s = s->next) {
                if (!s->module)
 
  * For more information, see tools/objtool/Documentation/stack-validation.txt.
  */
 #define STACK_FRAME_NON_STANDARD(func) \
-       static void __used __section(.discard.func_stack_frame_non_standard) \
+       static void __used __section(".discard.func_stack_frame_non_standard") \
                *__func_stack_frame_non_standard_##func = func
 
 #else /* __ASSEMBLY__ */