Lines Matching refs:mm
1144 struct mm_struct *mm; member
1180 extern int handle_mm_fault(struct mm_struct *mm, struct vm_area_struct *vma,
1182 extern int fixup_user_fault(struct task_struct *tsk, struct mm_struct *mm,
1185 static inline int handle_mm_fault(struct mm_struct *mm, in handle_mm_fault() argument
1194 struct mm_struct *mm, unsigned long address, in fixup_user_fault() argument
1204 extern int access_remote_vm(struct mm_struct *mm, unsigned long addr,
1207 long __get_user_pages(struct task_struct *tsk, struct mm_struct *mm,
1211 long get_user_pages(struct task_struct *tsk, struct mm_struct *mm,
1215 long get_user_pages_locked(struct task_struct *tsk, struct mm_struct *mm,
1219 long __get_user_pages_unlocked(struct task_struct *tsk, struct mm_struct *mm,
1223 long get_user_pages_unlocked(struct task_struct *tsk, struct mm_struct *mm,
1300 static inline unsigned long get_mm_counter(struct mm_struct *mm, int member) in get_mm_counter() argument
1302 long val = atomic_long_read(&mm->rss_stat.count[member]); in get_mm_counter()
1315 static inline void add_mm_counter(struct mm_struct *mm, int member, long value) in add_mm_counter() argument
1317 atomic_long_add(value, &mm->rss_stat.count[member]); in add_mm_counter()
1320 static inline void inc_mm_counter(struct mm_struct *mm, int member) in inc_mm_counter() argument
1322 atomic_long_inc(&mm->rss_stat.count[member]); in inc_mm_counter()
1325 static inline void dec_mm_counter(struct mm_struct *mm, int member) in dec_mm_counter() argument
1327 atomic_long_dec(&mm->rss_stat.count[member]); in dec_mm_counter()
1330 static inline unsigned long get_mm_rss(struct mm_struct *mm) in get_mm_rss() argument
1332 return get_mm_counter(mm, MM_FILEPAGES) + in get_mm_rss()
1333 get_mm_counter(mm, MM_ANONPAGES); in get_mm_rss()
1336 static inline unsigned long get_mm_hiwater_rss(struct mm_struct *mm) in get_mm_hiwater_rss() argument
1338 return max(mm->hiwater_rss, get_mm_rss(mm)); in get_mm_hiwater_rss()
1341 static inline unsigned long get_mm_hiwater_vm(struct mm_struct *mm) in get_mm_hiwater_vm() argument
1343 return max(mm->hiwater_vm, mm->total_vm); in get_mm_hiwater_vm()
1346 static inline void update_hiwater_rss(struct mm_struct *mm) in update_hiwater_rss() argument
1348 unsigned long _rss = get_mm_rss(mm); in update_hiwater_rss()
1350 if ((mm)->hiwater_rss < _rss) in update_hiwater_rss()
1351 (mm)->hiwater_rss = _rss; in update_hiwater_rss()
1354 static inline void update_hiwater_vm(struct mm_struct *mm) in update_hiwater_vm() argument
1356 if (mm->hiwater_vm < mm->total_vm) in update_hiwater_vm()
1357 mm->hiwater_vm = mm->total_vm; in update_hiwater_vm()
1360 static inline void reset_mm_hiwater_rss(struct mm_struct *mm) in reset_mm_hiwater_rss() argument
1362 mm->hiwater_rss = get_mm_rss(mm); in reset_mm_hiwater_rss()
1366 struct mm_struct *mm) in setmax_mm_hiwater_rss() argument
1368 unsigned long hiwater_rss = get_mm_hiwater_rss(mm); in setmax_mm_hiwater_rss()
1375 void sync_mm_rss(struct mm_struct *mm);
1377 static inline void sync_mm_rss(struct mm_struct *mm) in sync_mm_rss() argument
1384 extern pte_t *__get_locked_pte(struct mm_struct *mm, unsigned long addr,
1386 static inline pte_t *get_locked_pte(struct mm_struct *mm, unsigned long addr, in get_locked_pte() argument
1390 __cond_lock(*ptl, ptep = __get_locked_pte(mm, addr, ptl)); in get_locked_pte()
1395 static inline int __pud_alloc(struct mm_struct *mm, pgd_t *pgd, in __pud_alloc() argument
1401 int __pud_alloc(struct mm_struct *mm, pgd_t *pgd, unsigned long address);
1405 static inline int __pmd_alloc(struct mm_struct *mm, pud_t *pud, in __pmd_alloc() argument
1411 static inline void mm_nr_pmds_init(struct mm_struct *mm) {} in mm_nr_pmds_init() argument
1413 static inline unsigned long mm_nr_pmds(struct mm_struct *mm) in mm_nr_pmds() argument
1418 static inline void mm_inc_nr_pmds(struct mm_struct *mm) {} in mm_inc_nr_pmds() argument
1419 static inline void mm_dec_nr_pmds(struct mm_struct *mm) {} in mm_dec_nr_pmds() argument
1422 int __pmd_alloc(struct mm_struct *mm, pud_t *pud, unsigned long address);
1424 static inline void mm_nr_pmds_init(struct mm_struct *mm) in mm_nr_pmds_init() argument
1426 atomic_long_set(&mm->nr_pmds, 0); in mm_nr_pmds_init()
1429 static inline unsigned long mm_nr_pmds(struct mm_struct *mm) in mm_nr_pmds() argument
1431 return atomic_long_read(&mm->nr_pmds); in mm_nr_pmds()
1434 static inline void mm_inc_nr_pmds(struct mm_struct *mm) in mm_inc_nr_pmds() argument
1436 atomic_long_inc(&mm->nr_pmds); in mm_inc_nr_pmds()
1439 static inline void mm_dec_nr_pmds(struct mm_struct *mm) in mm_dec_nr_pmds() argument
1441 atomic_long_dec(&mm->nr_pmds); in mm_dec_nr_pmds()
1445 int __pte_alloc(struct mm_struct *mm, struct vm_area_struct *vma,
1454 static inline pud_t *pud_alloc(struct mm_struct *mm, pgd_t *pgd, unsigned long address) in pud_alloc() argument
1456 return (unlikely(pgd_none(*pgd)) && __pud_alloc(mm, pgd, address))? in pud_alloc()
1460 static inline pmd_t *pmd_alloc(struct mm_struct *mm, pud_t *pud, unsigned long address) in pmd_alloc() argument
1462 return (unlikely(pud_none(*pud)) && __pmd_alloc(mm, pud, address))? in pmd_alloc()
1497 static inline spinlock_t *pte_lockptr(struct mm_struct *mm, pmd_t *pmd) in pte_lockptr() argument
1530 static inline spinlock_t *pte_lockptr(struct mm_struct *mm, pmd_t *pmd) in pte_lockptr() argument
1532 return &mm->page_table_lock; in pte_lockptr()
1557 #define pte_offset_map_lock(mm, pmd, address, ptlp) \ argument
1559 spinlock_t *__ptl = pte_lockptr(mm, pmd); \
1571 #define pte_alloc_map(mm, vma, pmd, address) \ argument
1572 ((unlikely(pmd_none(*(pmd))) && __pte_alloc(mm, vma, \
1576 #define pte_alloc_map_lock(mm, pmd, address, ptlp) \ argument
1577 ((unlikely(pmd_none(*(pmd))) && __pte_alloc(mm, NULL, \
1579 NULL: pte_offset_map_lock(mm, pmd, address, ptlp))
1593 static inline spinlock_t *pmd_lockptr(struct mm_struct *mm, pmd_t *pmd) in pmd_lockptr() argument
1614 #define pmd_huge_pte(mm, pmd) (pmd_to_page(pmd)->pmd_huge_pte) argument
1618 static inline spinlock_t *pmd_lockptr(struct mm_struct *mm, pmd_t *pmd) in pmd_lockptr() argument
1620 return &mm->page_table_lock; in pmd_lockptr()
1626 #define pmd_huge_pte(mm, pmd) ((mm)->pmd_huge_pte) argument
1630 static inline spinlock_t *pmd_lock(struct mm_struct *mm, pmd_t *pmd) in pmd_lock() argument
1632 spinlock_t *ptl = pmd_lockptr(mm, pmd); in pmd_lock()
1823 extern int __vm_enough_memory(struct mm_struct *mm, long pages, int cap_sys_admin);
1856 extern int mm_take_all_locks(struct mm_struct *mm);
1857 extern void mm_drop_all_locks(struct mm_struct *mm);
1859 extern void set_mm_exe_file(struct mm_struct *mm, struct file *new_exe_file);
1860 extern struct file *get_mm_exe_file(struct mm_struct *mm);
1862 extern int may_expand_vm(struct mm_struct *mm, unsigned long npages);
1863 extern struct vm_area_struct *_install_special_mapping(struct mm_struct *mm,
1868 extern int install_special_mapping(struct mm_struct *mm,
1981 extern struct vm_area_struct * find_vma(struct mm_struct * mm, unsigned long addr);
1982 extern struct vm_area_struct * find_vma_prev(struct mm_struct * mm, unsigned long addr,
1987 static inline struct vm_area_struct * find_vma_intersection(struct mm_struct * mm, unsigned long st… in find_vma_intersection() argument
1989 struct vm_area_struct * vma = find_vma(mm,start_addr); in find_vma_intersection()
2002 static inline struct vm_area_struct *find_exact_vma(struct mm_struct *mm, in find_exact_vma() argument
2005 struct vm_area_struct *vma = find_vma(mm, vm_start); in find_exact_vma()
2070 extern int apply_to_page_range(struct mm_struct *mm, unsigned long address,
2076 static inline void vm_stat_account(struct mm_struct *mm, in vm_stat_account() argument
2079 mm->total_vm += pages; in vm_stat_account()
2112 extern struct vm_area_struct *get_gate_vma(struct mm_struct *mm);
2114 extern int in_gate_area(struct mm_struct *mm, unsigned long addr);
2116 static inline struct vm_area_struct *get_gate_vma(struct mm_struct *mm) in get_gate_vma() argument
2121 static inline int in_gate_area(struct mm_struct *mm, unsigned long addr) in in_gate_area() argument