}
 
 static void ath6kl_sdio_setup_scat_data(struct hif_scatter_req *scat_req,
-                                       struct hif_scatter_req_priv *s_req_priv,
                                        struct mmc_data *data)
 {
        struct scatterlist *sg;
                                                    MMC_DATA_READ;
 
        /* fill SG entries */
-       sg = s_req_priv->sgentries;
+       sg = scat_req->sgentries;
        sg_init_table(sg, scat_req->scat_entries);
 
        /* assemble SG list */
        }
 
        /* set scatter-gather table for request */
-       data->sg = s_req_priv->sgentries;
+       data->sg = scat_req->sgentries;
        data->sg_len = scat_req->scat_entries;
 }
 
        memset(&cmd, 0, sizeof(struct mmc_command));
        memset(&data, 0, sizeof(struct mmc_data));
 
-       ath6kl_sdio_setup_scat_data(scat_req, scat_req->req_priv, &data);
+       ath6kl_sdio_setup_scat_data(scat_req, &data);
 
        opcode = (scat_req->req & HIF_FIXED_ADDRESS) ?
                  CMD53_ARG_FIXED_ADDRESS : CMD53_ARG_INCR_ADDRESS;
                list_del(&s_req->list);
                spin_unlock_irqrestore(&ar_sdio->scat_lock, flag);
 
-               if (s_req->req_priv && s_req->req_priv->busrequest)
-                       ath6kl_sdio_free_bus_req(ar_sdio,
-                                                s_req->req_priv->busrequest);
+               if (s_req->busrequest)
+                       ath6kl_sdio_free_bus_req(ar_sdio, s_req->busrequest);
                kfree(s_req->virt_dma_buf);
-               kfree(s_req->req_priv);
+               kfree(s_req->sgentries);
                kfree(s_req);
 
                spin_lock_irqsave(&ar_sdio->scat_lock, flag);
 {
        struct hif_scatter_req *s_req;
        struct bus_request *bus_req;
-       int i, scat_req_sz, scat_list_sz;
+       int i, scat_req_sz, scat_list_sz, sg_sz;
 
        /* check if host supports scatter and it meets our requirements */
        if (ar_sdio->func->card->host->max_segs < MAX_SCATTER_ENTRIES_PER_REQ) {
                       sizeof(struct hif_scatter_item);
        scat_req_sz = sizeof(*s_req) + scat_list_sz;
 
+       sg_sz = sizeof(struct scatterlist) * MAX_SCATTER_ENTRIES_PER_REQ;
+
        for (i = 0; i < MAX_SCATTER_REQUESTS; i++) {
                /* allocate the scatter request */
                s_req = kzalloc(scat_req_sz, GFP_KERNEL);
                if (!s_req)
                        goto fail_setup_scat;
 
-               /* allocate the private request blob */
-               s_req->req_priv = kzalloc(sizeof(*s_req->req_priv), GFP_KERNEL);
+               /* allocate sglist */
+               s_req->sgentries = kzalloc(sg_sz, GFP_KERNEL);
 
-               if (!s_req->req_priv) {
+               if (!s_req->sgentries) {
                        kfree(s_req);
                        goto fail_setup_scat;
                }
                /* allocate a bus request for this scatter request */
                bus_req = ath6kl_sdio_alloc_busreq(ar_sdio);
                if (!bus_req) {
-                       kfree(s_req->req_priv);
+                       kfree(s_req->sgentries);
                        kfree(s_req);
                        goto fail_setup_scat;
                }
 
                /* assign the scatter request to this bus request */
                bus_req->scat_req = s_req;
-               s_req->req_priv->busrequest = bus_req;
+               s_req->busrequest = bus_req;
                /* add it to the scatter pool */
                hif_scatter_req_add(ar_sdio->ar, s_req);
        }
                                        struct hif_scatter_req *scat_req)
 {
        struct ath6kl_sdio *ar_sdio = ath6kl_sdio_priv(ar);
-       struct hif_scatter_req_priv *req_priv = scat_req->req_priv;
        u32 request = scat_req->req;
        int status = 0;
        unsigned long flags;
 
        if (request & HIF_SYNCHRONOUS) {
                sdio_claim_host(ar_sdio->func);
-               status = ath6kl_sdio_scat_rw(ar_sdio, req_priv->busrequest);
+               status = ath6kl_sdio_scat_rw(ar_sdio, scat_req->busrequest);
                sdio_release_host(ar_sdio->func);
        } else {
                spin_lock_irqsave(&ar_sdio->wr_async_lock, flags);
-               list_add_tail(&req_priv->busrequest->list, &ar_sdio->wr_asyncq);
+               list_add_tail(&scat_req->busrequest->list, &ar_sdio->wr_asyncq);
                spin_unlock_irqrestore(&ar_sdio->wr_async_lock, flags);
                queue_work(ar->ath6kl_wq, &ar_sdio->wr_async_work);
        }