}
 
 /* Called with interrupts disabled.  */
-static inline void user_enter_irqoff(void)
+static __always_inline void user_enter_irqoff(void)
 {
        if (context_tracking_enabled())
                __context_tracking_enter(CONTEXT_USER);
 
 }
-static inline void user_exit_irqoff(void)
+static __always_inline void user_exit_irqoff(void)
 {
        if (context_tracking_enabled())
                __context_tracking_exit(CONTEXT_USER);
  * is enabled.  If context tracking is disabled, returns
  * CONTEXT_DISABLED.  This should be used primarily for debugging.
  */
-static inline enum ctx_state ct_state(void)
+static __always_inline enum ctx_state ct_state(void)
 {
        return context_tracking_enabled() ?
                this_cpu_read(context_tracking.state) : CONTEXT_DISABLED;
 
 extern struct static_key_false context_tracking_key;
 DECLARE_PER_CPU(struct context_tracking, context_tracking);
 
-static inline bool context_tracking_enabled(void)
+static __always_inline bool context_tracking_enabled(void)
 {
        return static_branch_unlikely(&context_tracking_key);
 }
 
-static inline bool context_tracking_enabled_cpu(int cpu)
+static __always_inline bool context_tracking_enabled_cpu(int cpu)
 {
        return context_tracking_enabled() && per_cpu(context_tracking.active, cpu);
 }
        return context_tracking_enabled() && __this_cpu_read(context_tracking.active);
 }
 
-static inline bool context_tracking_in_user(void)
+static __always_inline bool context_tracking_in_user(void)
 {
        return __this_cpu_read(context_tracking.state) == CONTEXT_USER;
 }
 
 DEFINE_PER_CPU(struct context_tracking, context_tracking);
 EXPORT_SYMBOL_GPL(context_tracking);
 
-static bool context_tracking_recursion_enter(void)
+static noinstr bool context_tracking_recursion_enter(void)
 {
        int recursion;
 
        return false;
 }
 
-static void context_tracking_recursion_exit(void)
+static __always_inline void context_tracking_recursion_exit(void)
 {
        __this_cpu_dec(context_tracking.recursion);
 }
  * instructions to execute won't use any RCU read side critical section
  * because this function sets RCU in extended quiescent state.
  */
-void __context_tracking_enter(enum ctx_state state)
+void noinstr __context_tracking_enter(enum ctx_state state)
 {
        /* Kernel threads aren't supposed to go to userspace */
        WARN_ON_ONCE(!current->mm);
                         * on the tick.
                         */
                        if (state == CONTEXT_USER) {
+                               instrumentation_begin();
                                trace_user_enter(0);
                                vtime_user_enter(current);
+                               instrumentation_end();
                        }
                        rcu_user_enter();
                }
        }
        context_tracking_recursion_exit();
 }
-NOKPROBE_SYMBOL(__context_tracking_enter);
 EXPORT_SYMBOL_GPL(__context_tracking_enter);
 
 void context_tracking_enter(enum ctx_state state)
  * This call supports re-entrancy. This way it can be called from any exception
  * handler without needing to know if we came from userspace or not.
  */
-void __context_tracking_exit(enum ctx_state state)
+void noinstr __context_tracking_exit(enum ctx_state state)
 {
        if (!context_tracking_recursion_enter())
                return;
                         */
                        rcu_user_exit();
                        if (state == CONTEXT_USER) {
+                               instrumentation_begin();
                                vtime_user_exit(current);
                                trace_user_exit(0);
+                               instrumentation_end();
                        }
                }
                __this_cpu_write(context_tracking.state, CONTEXT_KERNEL);
        }
        context_tracking_recursion_exit();
 }
-NOKPROBE_SYMBOL(__context_tracking_exit);
 EXPORT_SYMBOL_GPL(__context_tracking_exit);
 
 void context_tracking_exit(enum ctx_state state)