__field_packed( unsigned long,  ret,            retval  )
                __field_packed( int,            ret,            depth   )
                __field_packed( unsigned int,   ret,            overrun )
-               __field_packed( unsigned long long, ret,        calltime)
-               __field_packed( unsigned long long, ret,        rettime )
+               __field(unsigned long long,     calltime                )
+               __field(unsigned long long,     rettime                 )
        ),
 
        F_printk("<-- %ps (%d) (start: %llx  end: %llx) over: %d retval: %lx",
                __field_packed( unsigned long,  ret,            func    )
                __field_packed( int,            ret,            depth   )
                __field_packed( unsigned int,   ret,            overrun )
-               __field_packed( unsigned long long, ret,        calltime)
-               __field_packed( unsigned long long, ret,        rettime )
+               __field(unsigned long long,     calltime                )
+               __field(unsigned long long,     rettime                 )
        ),
 
        F_printk("<-- %ps (%d) (start: %llx  end: %llx) over: %d",
 
        struct ftrace_graph_ret ret = {
                .func     = ip,
                .depth    = 0,
-               .calltime = time,
-               .rettime  = time,
        };
 
        __trace_graph_entry(tr, &ent, trace_ctx);
-       __trace_graph_return(tr, &ret, trace_ctx);
+       __trace_graph_return(tr, &ret, trace_ctx, time, time);
 }
 
 void
 }
 
 void __trace_graph_return(struct trace_array *tr,
-                               struct ftrace_graph_ret *trace,
-                               unsigned int trace_ctx)
+                         struct ftrace_graph_ret *trace,
+                         unsigned int trace_ctx,
+                         u64 calltime, u64 rettime)
 {
        struct ring_buffer_event *event;
        struct trace_buffer *buffer = tr->array_buffer.buffer;
                return;
        entry   = ring_buffer_event_data(event);
        entry->ret                              = *trace;
+       entry->calltime                         = calltime;
+       entry->rettime                          = rettime;
        trace_buffer_unlock_commit_nostack(buffer, event);
 }
 
        struct trace_array_cpu *data;
        struct fgraph_times *ftimes;
        unsigned int trace_ctx;
+       u64 calltime, rettime;
        long disabled;
        int size;
        int cpu;
 
+       rettime = trace_clock_local();
+
        ftrace_graph_addr_finish(gops, trace);
 
        if (*task_var & TRACE_GRAPH_NOTRACE) {
 
        handle_nosleeptime(trace, ftimes, size);
 
-       trace->calltime = ftimes->calltime;
+       calltime = ftimes->calltime;
 
        preempt_disable_notrace();
        cpu = raw_smp_processor_id();
        disabled = atomic_read(&data->disabled);
        if (likely(!disabled)) {
                trace_ctx = tracing_gen_ctx();
-               __trace_graph_return(tr, trace, trace_ctx);
+               __trace_graph_return(tr, trace, trace_ctx, calltime, rettime);
        }
        preempt_enable_notrace();
 }
 
        handle_nosleeptime(trace, ftimes, size);
 
-       trace->calltime = ftimes->calltime;
-
        if (tracing_thresh &&
-           (trace->rettime - ftimes->calltime < tracing_thresh))
+           (trace_clock_local() - ftimes->calltime < tracing_thresh))
                return;
        else
                trace_graph_return(trace, gops, fregs);
 
        graph_ret = &ret_entry->ret;
        call = &entry->graph_ent;
-       duration = graph_ret->rettime - graph_ret->calltime;
+       duration = ret_entry->rettime - ret_entry->calltime;
 
        func = call->func + iter->tr->text_delta;
 
 }
 
 static enum print_line_t
-print_graph_return(struct ftrace_graph_ret *trace, struct trace_seq *s,
+print_graph_return(struct ftrace_graph_ret_entry *retentry, struct trace_seq *s,
                   struct trace_entry *ent, struct trace_iterator *iter,
                   u32 flags)
 {
-       unsigned long long duration = trace->rettime - trace->calltime;
+       struct ftrace_graph_ret *trace = &retentry->ret;
+       u64 calltime = retentry->calltime;
+       u64 rettime = retentry->rettime;
+       unsigned long long duration = rettime - calltime;
        struct fgraph_data *data = iter->private;
        struct trace_array *tr = iter->tr;
        unsigned long func;
        case TRACE_GRAPH_RET: {
                struct ftrace_graph_ret_entry *field;
                trace_assign_type(field, entry);
-               return print_graph_return(&field->ret, s, entry, iter, flags);
+               return print_graph_return(field, s, entry, iter, flags);
        }
        case TRACE_STACK:
        case TRACE_FN: