static void
hugetlb_vmdelete_list(struct rb_root_cached *root, pgoff_t start, pgoff_t end,
- unsigned long zap_flags)
+ zap_flags_t zap_flags)
{
struct vm_area_struct *vma;
int *);
void unmap_hugepage_range(struct vm_area_struct *,
unsigned long, unsigned long, struct page *,
- unsigned long);
+ zap_flags_t);
void __unmap_hugepage_range_final(struct mmu_gather *tlb,
struct vm_area_struct *vma,
unsigned long start, unsigned long end,
- struct page *ref_page, unsigned long zap_flags);
+ struct page *ref_page, zap_flags_t zap_flags);
void hugetlb_report_meminfo(struct seq_file *);
int hugetlb_report_node_meminfo(char *buf, int len, int nid);
void hugetlb_show_meminfo(void);
static inline void __unmap_hugepage_range_final(struct mmu_gather *tlb,
struct vm_area_struct *vma, unsigned long start,
unsigned long end, struct page *ref_page,
- unsigned long zap_flags)
+ zap_flags_t zap_flags)
{
BUG();
}
static void __unmap_hugepage_range(struct mmu_gather *tlb, struct vm_area_struct *vma,
unsigned long start, unsigned long end,
- struct page *ref_page, unsigned long zap_flags)
+ struct page *ref_page, zap_flags_t zap_flags)
{
struct mm_struct *mm = vma->vm_mm;
unsigned long address;
void __unmap_hugepage_range_final(struct mmu_gather *tlb,
struct vm_area_struct *vma, unsigned long start,
unsigned long end, struct page *ref_page,
- unsigned long zap_flags)
+ zap_flags_t zap_flags)
{
__unmap_hugepage_range(tlb, vma, start, end, ref_page, zap_flags);
void unmap_hugepage_range(struct vm_area_struct *vma, unsigned long start,
unsigned long end, struct page *ref_page,
- unsigned long zap_flags)
+ zap_flags_t zap_flags)
{
struct mmu_gather tlb;
* safe to do nothing in this case.
*/
if (vma->vm_file) {
- unsigned long zap_flags = details ?
+ zap_flags_t zap_flags = details ?
details->zap_flags : 0;
i_mmap_lock_write(vma->vm_file->f_mapping);
__unmap_hugepage_range_final(tlb, vma, start, end,