}
 
 static void a6xx_set_pagetable(struct a6xx_gpu *a6xx_gpu,
-               struct msm_ringbuffer *ring, struct msm_file_private *ctx)
+               struct msm_ringbuffer *ring, struct msm_gem_submit *submit)
 {
        bool sysprof = refcount_read(&a6xx_gpu->base.base.sysprof_active) > 1;
+       struct msm_file_private *ctx = submit->queue->ctx;
        struct adreno_gpu *adreno_gpu = &a6xx_gpu->base;
        phys_addr_t ttbr;
        u32 asid;
        if (msm_iommu_pagetable_params(ctx->aspace->mmu, &ttbr, &asid))
                return;
 
+       if (adreno_gpu->info->family >= ADRENO_7XX_GEN1) {
+               /* Wait for previous submit to complete before continuing: */
+               OUT_PKT7(ring, CP_WAIT_TIMESTAMP, 4);
+               OUT_RING(ring, 0);
+               OUT_RING(ring, lower_32_bits(rbmemptr(ring, fence)));
+               OUT_RING(ring, upper_32_bits(rbmemptr(ring, fence)));
+               OUT_RING(ring, submit->seqno - 1);
+       }
+
        if (!sysprof) {
                if (!adreno_is_a7xx(adreno_gpu)) {
                        /* Turn off protected mode to write to special registers */
        struct msm_ringbuffer *ring = submit->ring;
        unsigned int i, ibs = 0;
 
-       a6xx_set_pagetable(a6xx_gpu, ring, submit->queue->ctx);
+       a6xx_set_pagetable(a6xx_gpu, ring, submit);
 
        get_stats_counter(ring, REG_A6XX_RBBM_PERFCTR_CP(0),
                rbmemptr_stats(ring, index, cpcycles_start));
        OUT_PKT7(ring, CP_THREAD_CONTROL, 1);
        OUT_RING(ring, CP_THREAD_CONTROL_0_SYNC_THREADS | CP_SET_THREAD_BR);
 
-       a6xx_set_pagetable(a6xx_gpu, ring, submit->queue->ctx);
+       a6xx_set_pagetable(a6xx_gpu, ring, submit);
 
        get_stats_counter(ring, REG_A7XX_RBBM_PERFCTR_CP(0),
                rbmemptr_stats(ring, index, cpcycles_start));