Lines Matching refs:numa_group

871 struct numa_group {  struct
901 return p->numa_group ? p->numa_group->gid : 0; in task_numa_group_id() argument
926 if (!p->numa_group) in group_faults()
929 return p->numa_group->faults[task_faults_idx(NUMA_MEM, nid, 0)] + in group_faults()
930 p->numa_group->faults[task_faults_idx(NUMA_MEM, nid, 1)]; in group_faults()
933 static inline unsigned long group_faults_cpu(struct numa_group *group, int nid) in group_faults_cpu()
1034 if (!p->numa_group) in group_weight()
1037 total_faults = p->numa_group->total_faults; in group_weight()
1051 struct numa_group *ng = p->numa_group; in should_numa_migrate_memory()
1267 long imp = env->p->numa_group ? groupimp : taskimp; in task_numa_compare()
1309 if (cur->numa_group == env->p->numa_group) { in task_numa_compare()
1316 if (cur->numa_group) in task_numa_compare()
1324 if (cur->numa_group) in task_numa_compare()
1475 if (env.best_cpu == -1 || (p->numa_group && in task_numa_migrate()
1476 nodes_weight(p->numa_group->active_nodes) > 1)) { in task_numa_migrate()
1509 if (p->numa_group) { in task_numa_migrate()
1515 if (node_isset(nid, p->numa_group->active_nodes)) in task_numa_migrate()
1575 static void update_numa_active_node_mask(struct numa_group *numa_group) in update_numa_active_node_mask() argument
1581 faults = group_faults_cpu(numa_group, nid); in update_numa_active_node_mask()
1587 faults = group_faults_cpu(numa_group, nid); in update_numa_active_node_mask()
1588 if (!node_isset(nid, numa_group->active_nodes)) { in update_numa_active_node_mask()
1590 node_set(nid, numa_group->active_nodes); in update_numa_active_node_mask()
1592 node_clear(nid, numa_group->active_nodes); in update_numa_active_node_mask()
1808 if (p->numa_group) { in task_numa_placement()
1809 group_lock = &p->numa_group->lock; in task_numa_placement()
1850 if (p->numa_group) { in task_numa_placement()
1858 p->numa_group->faults[mem_idx] += diff; in task_numa_placement()
1859 p->numa_group->faults_cpu[mem_idx] += f_diff; in task_numa_placement()
1860 p->numa_group->total_faults += diff; in task_numa_placement()
1861 group_faults += p->numa_group->faults[mem_idx]; in task_numa_placement()
1878 if (p->numa_group) { in task_numa_placement()
1879 update_numa_active_node_mask(p->numa_group); in task_numa_placement()
1894 static inline int get_numa_group(struct numa_group *grp) in get_numa_group()
1899 static inline void put_numa_group(struct numa_group *grp) in put_numa_group()
1908 struct numa_group *grp, *my_grp; in task_numa_group()
1914 if (unlikely(!p->numa_group)) { in task_numa_group()
1915 unsigned int size = sizeof(struct numa_group) + in task_numa_group()
1937 rcu_assign_pointer(p->numa_group, grp); in task_numa_group()
1946 grp = rcu_dereference(tsk->numa_group); in task_numa_group()
1950 my_grp = p->numa_group; in task_numa_group()
2002 rcu_assign_pointer(p->numa_group, grp); in task_numa_group()
2014 struct numa_group *grp = p->numa_group; in task_numa_free()
2027 RCU_INIT_POINTER(p->numa_group, NULL); in task_numa_free()
2084 if (!priv && !local && p->numa_group && in task_numa_fault()
2085 node_isset(cpu_node, p->numa_group->active_nodes) && in task_numa_fault()
2086 node_isset(mem_node, p->numa_group->active_nodes)) in task_numa_fault()
5476 struct numa_group *numa_group = rcu_dereference(p->numa_group); in migrate_improves_locality() local
5490 if (numa_group) { in migrate_improves_locality()
5492 if (node_isset(src_nid, numa_group->active_nodes)) in migrate_improves_locality()
5496 if (node_isset(dst_nid, numa_group->active_nodes)) in migrate_improves_locality()
5512 struct numa_group *numa_group = rcu_dereference(p->numa_group); in migrate_degrades_locality() local
5527 if (numa_group) { in migrate_degrades_locality()
5529 if (node_isset(dst_nid, numa_group->active_nodes)) in migrate_degrades_locality()
5533 if (node_isset(src_nid, numa_group->active_nodes)) in migrate_degrades_locality()