static ssize_t debug_stat_show(struct device *dev,
                struct device_attribute *attr, char *buf)
 {
-       int version = 1;
+       int version = 2;
        struct zram *zram = dev_to_zram(dev);
        ssize_t ret;
 
        down_read(&zram->init_lock);
        ret = scnprintf(buf, PAGE_SIZE,
-                       "version: %d\n%8llu %8llu\n",
+                       "version: %d\n%8llu\n",
                        version,
-                       (u64)atomic64_read(&zram->stats.writestall),
                        (u64)atomic64_read(&zram->stats.miss_free));
        up_read(&zram->init_lock);
 
        }
        kunmap_atomic(mem);
 
-compress_again:
        zstrm = zcomp_stream_get(zram->comp);
        src = kmap_atomic(page);
        ret = zcomp_compress(zstrm, src, &comp_len);
        if (unlikely(ret)) {
                zcomp_stream_put(zram->comp);
                pr_err("Compression failed! err=%d\n", ret);
-               zs_free(zram->mem_pool, handle);
                return ret;
        }
 
        if (comp_len >= huge_class_size)
                comp_len = PAGE_SIZE;
-       /*
-        * handle allocation has 2 paths:
-        * a) fast path is executed with preemption disabled (for
-        *  per-cpu streams) and has __GFP_DIRECT_RECLAIM bit clear,
-        *  since we can't sleep;
-        * b) slow path enables preemption and attempts to allocate
-        *  the page with __GFP_DIRECT_RECLAIM bit set. we have to
-        *  put per-cpu compression stream and, thus, to re-do
-        *  the compression once handle is allocated.
-        *
-        * if we have a 'non-null' handle here then we are coming
-        * from the slow path and handle has already been allocated.
-        */
-       if (!handle)
-               handle = zs_malloc(zram->mem_pool, comp_len,
-                               __GFP_KSWAPD_RECLAIM |
-                               __GFP_NOWARN |
-                               __GFP_HIGHMEM |
-                               __GFP_MOVABLE);
-       if (!handle) {
+
+       handle = zs_malloc(zram->mem_pool, comp_len,
+                       __GFP_KSWAPD_RECLAIM |
+                       __GFP_NOWARN |
+                       __GFP_HIGHMEM |
+                       __GFP_MOVABLE);
+
+       if (unlikely(!handle)) {
                zcomp_stream_put(zram->comp);
-               atomic64_inc(&zram->stats.writestall);
-               handle = zs_malloc(zram->mem_pool, comp_len,
-                               GFP_NOIO | __GFP_HIGHMEM |
-                               __GFP_MOVABLE);
-               if (handle)
-                       goto compress_again;
                return -ENOMEM;
        }
 
        if (ZRAM_LOGICAL_BLOCK_SIZE == PAGE_SIZE)
                blk_queue_max_write_zeroes_sectors(zram->disk->queue, UINT_MAX);
 
-       blk_queue_flag_set(QUEUE_FLAG_STABLE_WRITES, zram->disk->queue);
        ret = device_add_disk(NULL, zram->disk, zram_disk_groups);
        if (ret)
                goto out_cleanup_disk;