/* assign pages to this device */
for (x = 0; x < (1UL << order); x++)
- p[x].mapping = adev->mman.bdev.dev_mapping;
+ page_folio(&p[x])->mapping = adev->mman.bdev.dev_mapping;
/* If the hardware does not support UTCL2 snooping of the CPU caches
* then set_memory_wc() could be used as a workaround to mark the pages
kfree(sg);
p = virt_to_page(adev->gart.ptr);
for (x = 0; x < (1UL << order); x++)
- p[x].mapping = NULL;
+ page_folio(&p[x])->mapping = NULL;
__free_pages(p, order);
adev->gart.ptr = NULL;
return ret;
for (i = 0; i < ttm->num_pages; ++i)
- ttm->pages[i]->mapping = bdev->dev_mapping;
+ page_folio(ttm->pages[i])->mapping = bdev->dev_mapping;
return 0;
}
return;
for (i = 0; i < ttm->num_pages; ++i)
- ttm->pages[i]->mapping = NULL;
+ page_folio(ttm->pages[i])->mapping = NULL;
adev = amdgpu_ttm_adev(bdev);
return -EPERM;
p = pfn_to_page(pfn);
- if (p->mapping != adev->mman.bdev.dev_mapping)
+ if (page_folio(p)->mapping != adev->mman.bdev.dev_mapping)
return -EPERM;
ptr = kmap_local_page(p);
return -EPERM;
p = pfn_to_page(pfn);
- if (p->mapping != adev->mman.bdev.dev_mapping)
+ if (page_folio(p)->mapping != adev->mman.bdev.dev_mapping)
return -EPERM;
ptr = kmap_local_page(p);