unsigned int uses_cmm:1;
        /* The gmaps associated with this context are allowed to use huge pages. */
        unsigned int allow_gmap_hpage_1m:1;
-       /* The mmu context is for compat task */
-       unsigned int compat_mm:1;
 } mm_context_t;
 
 #define INIT_MM_CONTEXT(name)                                             \
 
        atomic_set(&mm->context.flush_count, 0);
        mm->context.gmap_asce = 0;
        mm->context.flush_mm = 0;
-       mm->context.compat_mm = test_thread_flag(TIF_31BIT);
 #ifdef CONFIG_PGSTE
        mm->context.alloc_pgste = page_table_allocate_pgste ||
                test_thread_flag(TIF_PGSTE) ||
                mm->context.asce = __pa(mm->pgd) | _ASCE_TABLE_LENGTH |
                                   _ASCE_USER_BITS | _ASCE_TYPE_REGION2;
                break;
-       case _REGION3_SIZE:
-               /* forked 2-level compat task, set new asce with new mm->pgd */
-               mm->context.asce = __pa(mm->pgd) | _ASCE_TABLE_LENGTH |
-                                  _ASCE_USER_BITS | _ASCE_TYPE_SEGMENT;
        }
        crst_table_init((unsigned long *) mm->pgd, pgd_entry_type(mm));
        return 0;
 
 }
 
 int crst_table_upgrade(struct mm_struct *mm, unsigned long limit);
-void crst_table_downgrade(struct mm_struct *);
 
 static inline unsigned long check_asce_limit(struct mm_struct *mm, unsigned long addr,
                                             unsigned long len)
 
 static inline pgd_t *pgd_alloc(struct mm_struct *mm)
 {
-       unsigned long *table = crst_table_alloc(mm);
-
-       if (!table)
-               return NULL;
-       if (mm->context.asce_limit == _REGION3_SIZE) {
-               /* Forking a compat process with 2 page table levels */
-               if (!pgtable_pmd_page_ctor(virt_to_page(table))) {
-                       crst_table_free(mm, table);
-                       return NULL;
-               }
-       }
-       return (pgd_t *) table;
+       return (pgd_t *) crst_table_alloc(mm);
 }
 
 static inline void pgd_free(struct mm_struct *mm, pgd_t *pgd)
 {
-       if (mm->context.asce_limit == _REGION3_SIZE)
-               pgtable_pmd_page_dtor(virt_to_page(pgd));
        crst_table_free(mm, (unsigned long *) pgd);
 }
 
 
        regs->psw.mask  = PSW_USER_BITS | PSW_MASK_BA;                  \
        regs->psw.addr  = new_psw;                                      \
        regs->gprs[15]  = new_stackp;                                   \
-       crst_table_downgrade(current->mm);                              \
        execve_tail();                                                  \
 } while (0)
 
 
        return -ENOMEM;
 }
 
-void crst_table_downgrade(struct mm_struct *mm)
-{
-       pgd_t *pgd;
-
-       /* downgrade should only happen from 3 to 2 levels (compat only) */
-       VM_BUG_ON(mm->context.asce_limit != _REGION2_SIZE);
-
-       if (current->active_mm == mm) {
-               clear_user_asce();
-               __tlb_flush_mm(mm);
-       }
-
-       pgd = mm->pgd;
-       mm_dec_nr_pmds(mm);
-       mm->pgd = (pgd_t *) (pgd_val(*pgd) & _REGION_ENTRY_ORIGIN);
-       mm->context.asce_limit = _REGION3_SIZE;
-       mm->context.asce = __pa(mm->pgd) | _ASCE_TABLE_LENGTH |
-                          _ASCE_USER_BITS | _ASCE_TYPE_SEGMENT;
-       crst_table_free(mm, (unsigned long *) pgd);
-
-       if (current->active_mm == mm)
-               set_user_asce(mm);
-}
-
 static inline unsigned int atomic_xor_bits(atomic_t *v, unsigned int bits)
 {
        unsigned int old, new;