pte_unmap(pte);
                goto out;
        }
-       inc_mm_counter(mm, rss);
+       inc_mm_counter(mm, anon_rss);
        lru_cache_add_active(page);
        set_pte_at(mm, address, pte, pte_mkdirty(pte_mkwrite(mk_pte(
                                        page, vma->vm_page_prot))));
 
                jiffies_to_clock_t(it_real_value),
                start_time,
                vsize,
-               mm ? get_mm_counter(mm, rss) : 0, /* you might want to shift this left 3 */
+               mm ? get_mm_rss(mm) : 0,
                rsslim,
                mm ? mm->start_code : 0,
                mm ? mm->end_code : 0,
 
                "VmPTE:\t%8lu kB\n",
                (mm->total_vm - mm->reserved_vm) << (PAGE_SHIFT-10),
                mm->locked_vm << (PAGE_SHIFT-10),
-               get_mm_counter(mm, rss) << (PAGE_SHIFT-10),
+               get_mm_rss(mm) << (PAGE_SHIFT-10),
                data << (PAGE_SHIFT-10),
                mm->stack_vm << (PAGE_SHIFT-10), text, lib,
                (PTRS_PER_PTE*sizeof(pte_t)*mm->nr_ptes) >> 10);
 int task_statm(struct mm_struct *mm, int *shared, int *text,
               int *data, int *resident)
 {
-       int rss = get_mm_counter(mm, rss);
-
-       *shared = rss - get_mm_counter(mm, anon_rss);
+       *shared = get_mm_counter(mm, file_rss);
        *text = (PAGE_ALIGN(mm->end_code) - (mm->start_code & PAGE_MASK))
                                                                >> PAGE_SHIFT;
        *data = mm->total_vm - mm->shared_vm;
-       *resident = rss;
+       *resident = *shared + get_mm_counter(mm, anon_rss);
        return mm->total_vm;
 }
 
 
 #define add_mm_counter(mm, member, value) (mm)->_##member += (value)
 #define inc_mm_counter(mm, member) (mm)->_##member++
 #define dec_mm_counter(mm, member) (mm)->_##member--
+#define get_mm_rss(mm) ((mm)->_file_rss + (mm)->_anon_rss)
+
 typedef unsigned long mm_counter_t;
 
 struct mm_struct {
        unsigned long exec_vm, stack_vm, reserved_vm, def_flags, nr_ptes;
 
        /* Special counters protected by the page_table_lock */
-       mm_counter_t _rss;
+       mm_counter_t _file_rss;
        mm_counter_t _anon_rss;
 
        unsigned long saved_auxv[AT_VECTOR_SIZE]; /* for /proc/PID/auxv */
 
                if (delta == 0)
                        return;
                tsk->acct_stimexpd = tsk->stime;
-               tsk->acct_rss_mem1 += delta * get_mm_counter(tsk->mm, rss);
+               tsk->acct_rss_mem1 += delta * get_mm_rss(tsk->mm);
                tsk->acct_vm_mem1 += delta * tsk->mm->total_vm;
        }
 }
 
        INIT_LIST_HEAD(&mm->mmlist);
        mm->core_waiters = 0;
        mm->nr_ptes = 0;
-       set_mm_counter(mm, rss, 0);
+       set_mm_counter(mm, file_rss, 0);
        set_mm_counter(mm, anon_rss, 0);
        spin_lock_init(&mm->page_table_lock);
        rwlock_init(&mm->ioctx_list_lock);
        if (retval)
                goto free_pt;
 
-       mm->hiwater_rss = get_mm_counter(mm,rss);
+       mm->hiwater_rss = get_mm_rss(mm);
        mm->hiwater_vm = mm->total_vm;
 
 good_mm:
 
                                        set_page_dirty(page);
                                page_remove_rmap(page);
                                page_cache_release(page);
-                               dec_mm_counter(mm, rss);
+                               dec_mm_counter(mm, file_rss);
                        }
                }
        } else {
 
        zap_pte(mm, vma, addr, pte);
 
-       inc_mm_counter(mm,rss);
+       inc_mm_counter(mm, file_rss);
        flush_icache_page(vma, page);
        set_pte_at(mm, addr, pte, mk_pte(page, prot));
        page_add_file_rmap(page);
 
                        entry = *src_pte;
                        ptepage = pte_page(entry);
                        get_page(ptepage);
-                       add_mm_counter(dst, rss, HPAGE_SIZE / PAGE_SIZE);
+                       add_mm_counter(dst, file_rss, HPAGE_SIZE / PAGE_SIZE);
                        set_huge_pte_at(dst, addr, dst_pte, entry);
                }
                spin_unlock(&src->page_table_lock);
 
                page = pte_page(pte);
                put_page(page);
-               add_mm_counter(mm, rss,  - (HPAGE_SIZE / PAGE_SIZE));
+               add_mm_counter(mm, file_rss, (int) -(HPAGE_SIZE / PAGE_SIZE));
        }
        flush_tlb_range(vma, start, end);
 }
                                goto out;
                        }
                }
-               add_mm_counter(mm, rss, HPAGE_SIZE / PAGE_SIZE);
+               add_mm_counter(mm, file_rss, HPAGE_SIZE / PAGE_SIZE);
                set_huge_pte_at(mm, addr, pte, make_huge_pte(vma, page));
        }
 out:
 
                pte = pte_mkclean(pte);
        pte = pte_mkold(pte);
        get_page(page);
-       inc_mm_counter(dst_mm, rss);
        if (PageAnon(page))
                inc_mm_counter(dst_mm, anon_rss);
+       else
+               inc_mm_counter(dst_mm, file_rss);
        set_pte_at(dst_mm, addr, dst_pte, pte);
        page_dup_rmap(page);
 }
                                        set_page_dirty(page);
                                if (pte_young(ptent))
                                        mark_page_accessed(page);
+                               dec_mm_counter(tlb->mm, file_rss);
                        }
-                       dec_mm_counter(tlb->mm, rss);
                        page_remove_rmap(page);
                        tlb_remove_page(tlb, page);
                        continue;
        spin_lock(&mm->page_table_lock);
        page_table = pte_offset_map(pmd, address);
        if (likely(pte_same(*page_table, orig_pte))) {
-               if (PageAnon(old_page))
-                       dec_mm_counter(mm, anon_rss);
                if (PageReserved(old_page))
-                       inc_mm_counter(mm, rss);
-               else
+                       inc_mm_counter(mm, anon_rss);
+               else {
                        page_remove_rmap(old_page);
-
+                       if (!PageAnon(old_page)) {
+                               inc_mm_counter(mm, anon_rss);
+                               dec_mm_counter(mm, file_rss);
+                       }
+               }
                flush_cache_page(vma, address, pfn);
                entry = mk_pte(new_page, vma->vm_page_prot);
                entry = maybe_mkwrite(pte_mkdirty(entry), vma);
 
        /* The page isn't present yet, go ahead with the fault. */
 
-       inc_mm_counter(mm, rss);
+       inc_mm_counter(mm, anon_rss);
        pte = mk_pte(page, vma->vm_page_prot);
        if (write_access && can_share_swap_page(page)) {
                pte = maybe_mkwrite(pte_mkdirty(pte), vma);
                        page_cache_release(page);
                        goto unlock;
                }
-               inc_mm_counter(mm, rss);
+               inc_mm_counter(mm, anon_rss);
                entry = mk_pte(page, vma->vm_page_prot);
                entry = maybe_mkwrite(pte_mkdirty(entry), vma);
                lru_cache_add_active(page);
         */
        /* Only go through if we didn't race with anybody else... */
        if (pte_none(*page_table)) {
-               if (!PageReserved(new_page))
-                       inc_mm_counter(mm, rss);
-
                flush_icache_page(vma, new_page);
                entry = mk_pte(new_page, vma->vm_page_prot);
                if (write_access)
                        entry = maybe_mkwrite(pte_mkdirty(entry), vma);
                set_pte_at(mm, address, page_table, entry);
                if (anon) {
+                       inc_mm_counter(mm, anon_rss);
                        lru_cache_add_active(new_page);
                        page_add_anon_rmap(new_page, vma, address);
-               } else
+               } else if (!PageReserved(new_page)) {
+                       inc_mm_counter(mm, file_rss);
                        page_add_file_rmap(new_page);
+               }
        } else {
                /* One of our sibling threads was faster, back out. */
                page_cache_release(new_page);
 void update_mem_hiwater(struct task_struct *tsk)
 {
        if (tsk->mm) {
-               unsigned long rss = get_mm_counter(tsk->mm, rss);
+               unsigned long rss = get_mm_rss(tsk->mm);
 
                if (tsk->mm->hiwater_rss < rss)
                        tsk->mm->hiwater_rss = rss;
 
        unsigned long rss;
 
        if (likely(tsk->mm)) {
-               rss = get_mm_counter(tsk->mm, rss);
+               rss = get_mm_rss(tsk->mm);
                if (tsk->mm->hiwater_rss < rss)
                        tsk->mm->hiwater_rss = rss;
                if (tsk->mm->hiwater_vm < tsk->mm->total_vm)
 
 {
        BUG_ON(PageReserved(page));
 
-       inc_mm_counter(vma->vm_mm, anon_rss);
-
        if (atomic_inc_and_test(&page->_mapcount)) {
                struct anon_vma *anon_vma = vma->anon_vma;
 
                set_pte_at(mm, address, pte, swp_entry_to_pte(entry));
                BUG_ON(pte_file(*pte));
                dec_mm_counter(mm, anon_rss);
-       }
+       } else
+               dec_mm_counter(mm, file_rss);
 
-       dec_mm_counter(mm, rss);
        page_remove_rmap(page);
        page_cache_release(page);
 
 
                page_remove_rmap(page);
                page_cache_release(page);
-               dec_mm_counter(mm, rss);
+               dec_mm_counter(mm, file_rss);
                (*mapcount)--;
        }
 
 
 static void unuse_pte(struct vm_area_struct *vma, pte_t *pte,
                unsigned long addr, swp_entry_t entry, struct page *page)
 {
-       inc_mm_counter(vma->vm_mm, rss);
+       inc_mm_counter(vma->vm_mm, anon_rss);
        get_page(page);
        set_pte_at(vma->vm_mm, addr, pte,
                   pte_mkold(mk_pte(page, vma->vm_page_prot)));