idx = (virt - bo->start) >> PAGE_SHIFT;
                offset = (virt - bo->start) - (idx << PAGE_SHIFT);
 
-               src = (char *)kmap(bo->page_obj[idx].page) + offset;
+               src = (char *)kmap(bo->pages[idx]) + offset;
 
                if ((bytes + offset) >= PAGE_SIZE) {
                        len = PAGE_SIZE - offset;
 
                clflush_cache_range(src, len);
 
-               kunmap(bo->page_obj[idx].page);
+               kunmap(bo->pages[idx]);
        }
 
        return 0;
                offset = (virt - bo->start) - (idx << PAGE_SHIFT);
 
                if (in_atomic())
-                       des = (char *)kmap_atomic(bo->page_obj[idx].page);
+                       des = (char *)kmap_atomic(bo->pages[idx]);
                else
-                       des = (char *)kmap(bo->page_obj[idx].page);
+                       des = (char *)kmap(bo->pages[idx]);
 
                if (!des) {
                        dev_err(atomisp_dev,
                         */
                        kunmap_atomic(des - offset);
                else
-                       kunmap(bo->page_obj[idx].page);
+                       kunmap(bo->pages[idx]);
        }
 
        return 0;
                idx = (virt - bo->start) >> PAGE_SHIFT;
                offset = (virt - bo->start) - (idx << PAGE_SHIFT);
 
-               des = (char *)kmap(bo->page_obj[idx].page) + offset;
+               des = (char *)kmap(bo->pages[idx]) + offset;
 
                if ((bytes + offset) >= PAGE_SIZE) {
                        len = PAGE_SIZE - offset;
 
                clflush_cache_range(des, len);
 
-               kunmap(bo->page_obj[idx].page);
+               kunmap(bo->pages[idx]);
        }
 
        return 0;
        idx = (virt - bo->start) >> PAGE_SHIFT;
        offset = (virt - bo->start) - (idx << PAGE_SHIFT);
 
-       return page_to_phys(bo->page_obj[idx].page) + offset;
+       return page_to_phys(bo->pages[idx]) + offset;
 }
 
 int hmm_mmap(struct vm_area_struct *vma, ia_css_ptr virt)
 
        int i, ret;
 
        for (i = 0; i < free_pgnr; i++) {
-               ret = set_pages_wb(bo->page_obj[i].page, 1);
+               ret = set_pages_wb(bo->pages[i], 1);
                if (ret)
                        dev_err(atomisp_dev,
                                "set page to WB err ...ret = %d\n",
                address be valid,it maybe memory corruption by lowmemory
                */
                if (!ret) {
-                       __free_pages(bo->page_obj[i].page, 0);
+                       __free_pages(bo->pages[i], 0);
                }
        }
 }
 
        pgnr = bo->pgnr;
 
-       bo->page_obj = kmalloc_array(pgnr, sizeof(struct hmm_page_object),
-                                    GFP_KERNEL);
-       if (unlikely(!bo->page_obj))
-               return -ENOMEM;
-
        i = 0;
        alloc_pgnr = 0;
 
                        }
 
                        for (j = 0; j < blk_pgnr; j++, i++) {
-                               bo->page_obj[i].page = pages + j;
+                               bo->pages[i] = pages + j;
                        }
 
                        pgnr -= blk_pgnr;
 cleanup:
        alloc_pgnr = i;
        free_private_bo_pages(bo, alloc_pgnr);
-
-       kfree(bo->page_obj);
-
        return -ENOMEM;
 }
 
-static void free_private_pages(struct hmm_buffer_object *bo)
-{
-       free_private_bo_pages(bo, bo->pgnr);
-       kfree(bo->page_obj);
-}
-
 static void free_user_pages(struct hmm_buffer_object *bo,
                            unsigned int page_nr)
 {
                for (i = 0; i < page_nr; i++)
                        put_page(bo->pages[i]);
        }
-       kfree(bo->pages);
-       kfree(bo->page_obj);
 }
 
 /*
                            const void __user *userptr)
 {
        int page_nr;
-       int i;
        struct vm_area_struct *vma;
-       struct page **pages;
-
-       pages = kmalloc_array(bo->pgnr, sizeof(struct page *), GFP_KERNEL);
-       if (unlikely(!pages))
-               return -ENOMEM;
-
-       bo->page_obj = kmalloc_array(bo->pgnr, sizeof(struct hmm_page_object),
-                                    GFP_KERNEL);
-       if (unlikely(!bo->page_obj)) {
-               kfree(pages);
-               return -ENOMEM;
-       }
 
        mutex_unlock(&bo->mutex);
        mmap_read_lock(current->mm);
        mmap_read_unlock(current->mm);
        if (!vma) {
                dev_err(atomisp_dev, "find_vma failed\n");
-               kfree(bo->page_obj);
-               kfree(pages);
                mutex_lock(&bo->mutex);
                return -EFAULT;
        }
 
        userptr = untagged_addr(userptr);
 
-       bo->pages = pages;
-
        if (vma->vm_flags & (VM_IO | VM_PFNMAP)) {
                page_nr = pin_user_pages((unsigned long)userptr, bo->pgnr,
                                         FOLL_LONGTERM | FOLL_WRITE,
-                                        pages, NULL);
+                                        bo->pages, NULL);
                bo->mem_type = HMM_BO_MEM_TYPE_PFN;
        } else {
                /*Handle frame buffer allocated in user space*/
                mutex_unlock(&bo->mutex);
                page_nr = get_user_pages_fast((unsigned long)userptr,
-                                             (int)(bo->pgnr), 1, pages);
+                                             (int)(bo->pgnr), 1, bo->pages);
                mutex_lock(&bo->mutex);
                bo->mem_type = HMM_BO_MEM_TYPE_USER;
        }
                goto out_of_mem;
        }
 
-       for (i = 0; i < bo->pgnr; i++) {
-               bo->page_obj[i].page = pages[i];
-       }
-
        return 0;
 
 out_of_mem:
        mutex_lock(&bo->mutex);
        check_bo_status_no_goto(bo, HMM_BO_PAGE_ALLOCED, status_err);
 
+       bo->pages = kmalloc_array(bo->pgnr, sizeof(struct page *), GFP_KERNEL);
+       if (unlikely(!bo->pages)) {
+               ret = -ENOMEM;
+               goto alloc_err;
+       }
+
        /*
         * TO DO:
         * add HMM_BO_USER type
        return 0;
 
 alloc_err:
+       kfree(bo->pages);
        mutex_unlock(&bo->mutex);
        dev_err(atomisp_dev, "alloc pages err...\n");
        return ret;
        bo->status &= (~HMM_BO_PAGE_ALLOCED);
 
        if (bo->type == HMM_BO_PRIVATE)
-               free_private_pages(bo);
+               free_private_bo_pages(bo, bo->pgnr);
        else if (bo->type == HMM_BO_USER)
                free_user_pages(bo, bo->pgnr);
        else
                dev_err(atomisp_dev, "invalid buffer type.\n");
+
+       kfree(bo->pages);
        mutex_unlock(&bo->mutex);
 
        return;
        for (i = 0; i < bo->pgnr; i++) {
                ret =
                    isp_mmu_map(&bdev->mmu, virt,
-                               page_to_phys(bo->page_obj[i].page), 1);
+                               page_to_phys(bo->pages[i]), 1);
                if (ret)
                        goto map_err;
                virt += (1 << PAGE_SHIFT);
 
 void *hmm_bo_vmap(struct hmm_buffer_object *bo, bool cached)
 {
-       struct page **pages;
-       int i;
-
        check_bo_null_return(bo, NULL);
 
        mutex_lock(&bo->mutex);
                bo->status &= ~(HMM_BO_VMAPED | HMM_BO_VMAPED_CACHED);
        }
 
-       pages = kmalloc_array(bo->pgnr, sizeof(*pages), GFP_KERNEL);
-       if (unlikely(!pages)) {
-               mutex_unlock(&bo->mutex);
-               return NULL;
-       }
-
-       for (i = 0; i < bo->pgnr; i++)
-               pages[i] = bo->page_obj[i].page;
-
-       bo->vmap_addr = vmap(pages, bo->pgnr, VM_MAP,
+       bo->vmap_addr = vmap(bo->pages, bo->pgnr, VM_MAP,
                             cached ? PAGE_KERNEL : PAGE_KERNEL_NOCACHE);
        if (unlikely(!bo->vmap_addr)) {
-               kfree(pages);
                mutex_unlock(&bo->mutex);
                dev_err(atomisp_dev, "vmap failed...\n");
                return NULL;
        }
        bo->status |= (cached ? HMM_BO_VMAPED_CACHED : HMM_BO_VMAPED);
 
-       kfree(pages);
-
        mutex_unlock(&bo->mutex);
        return bo->vmap_addr;
 }
 
        virt = vma->vm_start;
        for (i = 0; i < pgnr; i++) {
-               pfn = page_to_pfn(bo->page_obj[i].page);
+               pfn = page_to_pfn(bo->pages[i]);
                if (remap_pfn_range(vma, virt, pfn, PAGE_SIZE, PAGE_SHARED)) {
                        dev_warn(atomisp_dev,
                                 "remap_pfn_range failed: virt = 0x%x, pfn = 0x%x, mapped_pgnr = %d\n",