read-only, and can be pinned. */
 static void __xen_pgd_pin(struct mm_struct *mm, pgd_t *pgd)
 {
+       trace_xen_mmu_pgd_pin(mm, pgd);
+
        xen_mc_batch();
 
        if (__xen_pgd_walk(mm, pgd, xen_pin_page, USER_LIMIT)) {
 /* Release a pagetables pages back as normal RW */
 static void __xen_pgd_unpin(struct mm_struct *mm, pgd_t *pgd)
 {
+       trace_xen_mmu_pgd_unpin(mm, pgd);
+
        xen_mc_batch();
 
        xen_do_pin(MMUEXT_UNPIN_TABLE, PFN_DOWN(__pa(pgd)));
 
                      __entry->pinned ? "" : "un")
        );
 
+TRACE_EVENT(xen_mmu_pgd_pin,
+           TP_PROTO(struct mm_struct *mm, pgd_t *pgd),
+           TP_ARGS(mm, pgd),
+           TP_STRUCT__entry(
+                   __field(struct mm_struct *, mm)
+                   __field(pgd_t *, pgd)
+                   ),
+           TP_fast_assign(__entry->mm = mm;
+                          __entry->pgd = pgd),
+           TP_printk("mm %p pgd %p", __entry->mm, __entry->pgd)
+       );
+
+TRACE_EVENT(xen_mmu_pgd_unpin,
+           TP_PROTO(struct mm_struct *mm, pgd_t *pgd),
+           TP_ARGS(mm, pgd),
+           TP_STRUCT__entry(
+                   __field(struct mm_struct *, mm)
+                   __field(pgd_t *, pgd)
+                   ),
+           TP_fast_assign(__entry->mm = mm;
+                          __entry->pgd = pgd),
+           TP_printk("mm %p pgd %p", __entry->mm, __entry->pgd)
+       );
+
 #endif /*  _TRACE_XEN_H */
 
 /* This part must be outside protection */