* Freed pages are either returned to buddy allocator or held back to be used
  * as metadata pages.
  */
-bool __init kmsan_memblock_free_pages(struct page *page, unsigned int order);
+bool __init __must_check kmsan_memblock_free_pages(struct page *page,
+                                                  unsigned int order);
 
 /**
  * kmsan_alloc_page() - Notify KMSAN about an alloc_pages() call.
  * vmalloc metadata address range. Returns 0 on success, callers must check
  * for non-zero return value.
  */
-int kmsan_vmap_pages_range_noflush(unsigned long start, unsigned long end,
-                                  pgprot_t prot, struct page **pages,
-                                  unsigned int page_shift);
+int __must_check kmsan_vmap_pages_range_noflush(unsigned long start,
+                                               unsigned long end,
+                                               pgprot_t prot,
+                                               struct page **pages,
+                                               unsigned int page_shift);
 
 /**
  * kmsan_vunmap_kernel_range_noflush() - Notify KMSAN about a vunmap.
  * virtual memory. Returns 0 on success, callers must check for non-zero return
  * value.
  */
-int kmsan_ioremap_page_range(unsigned long addr, unsigned long end,
-                            phys_addr_t phys_addr, pgprot_t prot,
-                            unsigned int page_shift);
+int __must_check kmsan_ioremap_page_range(unsigned long addr, unsigned long end,
+                                         phys_addr_t phys_addr, pgprot_t prot,
+                                         unsigned int page_shift);
 
 /**
  * kmsan_iounmap_page_range() - Notify KMSAN about a iounmap_page_range() call.
 {
 }
 
-static inline bool kmsan_memblock_free_pages(struct page *page,
-                                            unsigned int order)
+static inline bool __must_check kmsan_memblock_free_pages(struct page *page,
+                                                         unsigned int order)
 {
        return true;
 }
 {
 }
 
-static inline int kmsan_alloc_page(struct page *page, unsigned int order,
-                                  gfp_t flags)
+static inline void kmsan_alloc_page(struct page *page, unsigned int order,
+                                   gfp_t flags)
 {
-       return 0;
 }
 
 static inline void kmsan_free_page(struct page *page, unsigned int order)
 {
 }
 
-static inline int kmsan_vmap_pages_range_noflush(unsigned long start,
-                                                unsigned long end,
-                                                pgprot_t prot,
-                                                struct page **pages,
-                                                unsigned int page_shift)
+static inline int __must_check kmsan_vmap_pages_range_noflush(
+       unsigned long start, unsigned long end, pgprot_t prot,
+       struct page **pages, unsigned int page_shift)
 {
        return 0;
 }
 {
 }
 
-static inline int kmsan_ioremap_page_range(unsigned long start,
-                                          unsigned long end,
-                                          phys_addr_t phys_addr, pgprot_t prot,
-                                          unsigned int page_shift)
+static inline int __must_check kmsan_ioremap_page_range(unsigned long start,
+                                                       unsigned long end,
+                                                       phys_addr_t phys_addr,
+                                                       pgprot_t prot,
+                                                       unsigned int page_shift)
 {
        return 0;
 }