#include <linux/genhd.h>
 #include <linux/highmem.h>
 #include <linux/slab.h>
-#include <linux/lzo.h>
 #include <linux/string.h>
 #include <linux/vmalloc.h>
 
 /* Globals */
 static int zram_major;
 static struct zram *zram_devices;
+static const char *default_compressor = "lzo";
 
 /* Module params (documentation at end) */
 static unsigned int num_devices = 1;
 static void zram_meta_free(struct zram_meta *meta)
 {
        zs_destroy_pool(meta->mem_pool);
-       kfree(meta->compress_workmem);
-       free_pages((unsigned long)meta->compress_buffer, 1);
        vfree(meta->table);
        kfree(meta);
 }
        if (!meta)
                goto out;
 
-       meta->compress_workmem = kzalloc(LZO1X_MEM_COMPRESS, GFP_KERNEL);
-       if (!meta->compress_workmem)
-               goto free_meta;
-
-       meta->compress_buffer =
-               (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, 1);
-       if (!meta->compress_buffer) {
-               pr_err("Error allocating compressor buffer space\n");
-               goto free_workmem;
-       }
-
        num_pages = disksize >> PAGE_SHIFT;
        meta->table = vzalloc(num_pages * sizeof(*meta->table));
        if (!meta->table) {
                pr_err("Error allocating zram address table\n");
-               goto free_buffer;
+               goto free_meta;
        }
 
        meta->mem_pool = zs_create_pool(GFP_NOIO | __GFP_HIGHMEM);
        }
 
        rwlock_init(&meta->tb_lock);
-       mutex_init(&meta->buffer_lock);
        return meta;
 
 free_table:
        vfree(meta->table);
-free_buffer:
-       free_pages((unsigned long)meta->compress_buffer, 1);
-free_workmem:
-       kfree(meta->compress_workmem);
 free_meta:
        kfree(meta);
        meta = NULL;
 
 static int zram_decompress_page(struct zram *zram, char *mem, u32 index)
 {
-       int ret = LZO_E_OK;
-       size_t clen = PAGE_SIZE;
+       int ret = 0;
        unsigned char *cmem;
        struct zram_meta *meta = zram->meta;
        unsigned long handle;
        if (size == PAGE_SIZE)
                copy_page(mem, cmem);
        else
-               ret = lzo1x_decompress_safe(cmem, size, mem, &clen);
+               ret = zcomp_decompress(zram->comp, cmem, size, mem);
        zs_unmap_object(meta->mem_pool, handle);
        read_unlock(&meta->tb_lock);
 
        /* Should NEVER happen. Return bio error if it does. */
-       if (unlikely(ret != LZO_E_OK)) {
+       if (unlikely(ret)) {
                pr_err("Decompression failed! err=%d, page=%u\n", ret, index);
                atomic64_inc(&zram->stats.failed_reads);
                return ret;
 
        ret = zram_decompress_page(zram, uncmem, index);
        /* Should NEVER happen. Return bio error if it does. */
-       if (unlikely(ret != LZO_E_OK))
+       if (unlikely(ret))
                goto out_cleanup;
 
        if (is_partial_io(bvec))
        struct page *page;
        unsigned char *user_mem, *cmem, *src, *uncmem = NULL;
        struct zram_meta *meta = zram->meta;
+       struct zcomp_strm *zstrm;
        bool locked = false;
 
        page = bvec->bv_page;
-       src = meta->compress_buffer;
-
        if (is_partial_io(bvec)) {
                /*
                 * This is a partial IO. We need to read the full page
                        goto out;
        }
 
-       mutex_lock(&meta->buffer_lock);
+       zstrm = zcomp_strm_find(zram->comp);
        locked = true;
        user_mem = kmap_atomic(page);
 
                goto out;
        }
 
-       ret = lzo1x_1_compress(uncmem, PAGE_SIZE, src, &clen,
-                              meta->compress_workmem);
+       ret = zcomp_compress(zram->comp, zstrm, uncmem, &clen);
        if (!is_partial_io(bvec)) {
                kunmap_atomic(user_mem);
                user_mem = NULL;
                uncmem = NULL;
        }
 
-       if (unlikely(ret != LZO_E_OK)) {
+       if (unlikely(ret)) {
                pr_err("Compression failed! err=%d\n", ret);
                goto out;
        }
-
+       src = zstrm->buffer;
        if (unlikely(clen > max_zpage_size)) {
                clen = PAGE_SIZE;
-               src = NULL;
                if (is_partial_io(bvec))
                        src = uncmem;
        }
                memcpy(cmem, src, clen);
        }
 
+       zcomp_strm_release(zram->comp, zstrm);
+       locked = false;
        zs_unmap_object(meta->mem_pool, handle);
 
        /*
        atomic64_inc(&zram->stats.pages_stored);
 out:
        if (locked)
-               mutex_unlock(&meta->buffer_lock);
+               zcomp_strm_release(zram->comp, zstrm);
        if (is_partial_io(bvec))
                kfree(uncmem);
-
        if (ret)
                atomic64_inc(&zram->stats.failed_writes);
        return ret;
                zs_free(meta->mem_pool, handle);
        }
 
+       zcomp_destroy(zram->comp);
        zram_meta_free(zram->meta);
        zram->meta = NULL;
        /* Reset stats */
        u64 disksize;
        struct zram_meta *meta;
        struct zram *zram = dev_to_zram(dev);
+       int err;
 
        disksize = memparse(buf, NULL);
        if (!disksize)
 
        down_write(&zram->init_lock);
        if (init_done(zram)) {
-               zram_meta_free(meta);
-               up_write(&zram->init_lock);
                pr_info("Cannot change disksize for initialized device\n");
-               return -EBUSY;
+               err = -EBUSY;
+               goto out_free_meta;
+       }
+
+       zram->comp = zcomp_create(default_compressor);
+       if (!zram->comp) {
+               pr_info("Cannot initialise %s compressing backend\n",
+                               default_compressor);
+               err = -EINVAL;
+               goto out_free_meta;
        }
 
        zram->meta = meta;
        up_write(&zram->init_lock);
 
        return len;
+
+out_free_meta:
+       up_write(&zram->init_lock);
+       zram_meta_free(meta);
+       return err;
 }
 
 static ssize_t reset_store(struct device *dev,