Lines Matching refs:pte

488 static void smaps_pte_entry(pte_t *pte, unsigned long addr,  in smaps_pte_entry()  argument
495 if (pte_present(*pte)) { in smaps_pte_entry()
496 page = vm_normal_page(vma, addr, *pte); in smaps_pte_entry()
497 } else if (is_swap_pte(*pte)) { in smaps_pte_entry()
498 swp_entry_t swpent = pte_to_swp_entry(*pte); in smaps_pte_entry()
519 smaps_account(mss, page, PAGE_SIZE, pte_young(*pte), pte_dirty(*pte)); in smaps_pte_entry()
549 pte_t *pte; in smaps_pte_range() local
565 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in smaps_pte_range()
566 for (; addr != end; pte++, addr += PAGE_SIZE) in smaps_pte_range()
567 smaps_pte_entry(pte, addr, walk); in smaps_pte_range()
568 pte_unmap_unlock(pte - 1, ptl); in smaps_pte_range()
632 static int smaps_hugetlb_range(pte_t *pte, unsigned long hmask, in smaps_hugetlb_range() argument
640 if (pte_present(*pte)) { in smaps_hugetlb_range()
641 page = vm_normal_page(vma, addr, *pte); in smaps_hugetlb_range()
642 } else if (is_swap_pte(*pte)) { in smaps_hugetlb_range()
643 swp_entry_t swpent = pte_to_swp_entry(*pte); in smaps_hugetlb_range()
784 unsigned long addr, pte_t *pte) in clear_soft_dirty() argument
792 pte_t ptent = *pte; in clear_soft_dirty()
795 ptent = ptep_modify_prot_start(vma->vm_mm, addr, pte); in clear_soft_dirty()
798 ptep_modify_prot_commit(vma->vm_mm, addr, pte, ptent); in clear_soft_dirty()
801 set_pte_at(vma->vm_mm, addr, pte, ptent); in clear_soft_dirty()
806 unsigned long addr, pte_t *pte) in clear_soft_dirty() argument
837 pte_t *pte, ptent; in clear_refs_pte_range() local
861 pte = pte_offset_map_lock(vma->vm_mm, pmd, addr, &ptl); in clear_refs_pte_range()
862 for (; addr != end; pte++, addr += PAGE_SIZE) { in clear_refs_pte_range()
863 ptent = *pte; in clear_refs_pte_range()
866 clear_soft_dirty(vma, addr, pte); in clear_refs_pte_range()
878 ptep_test_and_clear_young(vma, addr, pte); in clear_refs_pte_range()
882 pte_unmap_unlock(pte - 1, ptl); in clear_refs_pte_range()
1071 struct vm_area_struct *vma, unsigned long addr, pte_t pte) in pte_to_pagemap_entry() argument
1076 if (pte_present(pte)) { in pte_to_pagemap_entry()
1078 frame = pte_pfn(pte); in pte_to_pagemap_entry()
1080 page = vm_normal_page(vma, addr, pte); in pte_to_pagemap_entry()
1081 if (pte_soft_dirty(pte)) in pte_to_pagemap_entry()
1083 } else if (is_swap_pte(pte)) { in pte_to_pagemap_entry()
1085 if (pte_swp_soft_dirty(pte)) in pte_to_pagemap_entry()
1087 entry = pte_to_swp_entry(pte); in pte_to_pagemap_entry()
1111 pte_t *pte, *orig_pte; in pagemap_pmd_range() local
1161 orig_pte = pte = pte_offset_map_lock(walk->mm, pmdp, addr, &ptl); in pagemap_pmd_range()
1162 for (; addr < end; pte++, addr += PAGE_SIZE) { in pagemap_pmd_range()
1165 pme = pte_to_pagemap_entry(pm, vma, addr, *pte); in pagemap_pmd_range()
1187 pte_t pte; in pagemap_hugetlb_range() local
1192 pte = huge_ptep_get(ptep); in pagemap_hugetlb_range()
1193 if (pte_present(pte)) { in pagemap_hugetlb_range()
1194 struct page *page = pte_page(pte); in pagemap_hugetlb_range()
1204 frame = pte_pfn(pte) + in pagemap_hugetlb_range()
1415 static struct page *can_gather_numa_stats(pte_t pte, struct vm_area_struct *vma, in can_gather_numa_stats() argument
1421 if (!pte_present(pte)) in can_gather_numa_stats()
1424 page = vm_normal_page(vma, addr, pte); in can_gather_numa_stats()
1471 pte_t *pte; in gather_pte_stats() local
1488 orig_pte = pte = pte_offset_map_lock(walk->mm, pmd, addr, &ptl); in gather_pte_stats()
1490 struct page *page = can_gather_numa_stats(*pte, vma, addr); in gather_pte_stats()
1493 gather_stats(page, md, pte_dirty(*pte), 1); in gather_pte_stats()
1495 } while (pte++, addr += PAGE_SIZE, addr != end); in gather_pte_stats()
1500 static int gather_hugetlb_stats(pte_t *pte, unsigned long hmask, in gather_hugetlb_stats() argument
1503 pte_t huge_pte = huge_ptep_get(pte); in gather_hugetlb_stats()
1520 static int gather_hugetlb_stats(pte_t *pte, unsigned long hmask, in gather_hugetlb_stats() argument