Lines Matching refs:sgc
5461 if (group->sgc->capacity != SCHED_CAPACITY_SCALE) { in sched_domain_debug_one()
5463 group->sgc->capacity); in sched_domain_debug_one()
5688 if (free_sgc && atomic_dec_and_test(&sg->sgc->ref)) in free_sched_groups()
5689 kfree(sg->sgc); in free_sched_groups()
5707 kfree(sd->groups->sgc); in free_sched_domain()
5914 sg->sgc = *per_cpu_ptr(sdd->sgc, i); in build_overlap_sched_groups()
5915 if (atomic_inc_return(&sg->sgc->ref) == 1) in build_overlap_sched_groups()
5923 sg->sgc->capacity = SCHED_CAPACITY_SCALE * cpumask_weight(sg_span); in build_overlap_sched_groups()
5961 (*sg)->sgc = *per_cpu_ptr(sdd->sgc, cpu); in get_group()
5962 atomic_set(&(*sg)->sgc->ref, 1); /* for claim_allocations */ in get_group()
6049 atomic_set(&sg->sgc->nr_busy_cpus, sg->group_weight); in init_sched_groups_capacity()
6140 if (atomic_read(&(*per_cpu_ptr(sdd->sgc, cpu))->ref)) in claim_allocations()
6141 *per_cpu_ptr(sdd->sgc, cpu) = NULL; in claim_allocations()
6603 sdd->sgc = alloc_percpu(struct sched_group_capacity *); in __sdt_alloc()
6604 if (!sdd->sgc) in __sdt_alloc()
6610 struct sched_group_capacity *sgc; in __sdt_alloc() local
6628 sgc = kzalloc_node(sizeof(struct sched_group_capacity) + cpumask_size(), in __sdt_alloc()
6630 if (!sgc) in __sdt_alloc()
6633 *per_cpu_ptr(sdd->sgc, j) = sgc; in __sdt_alloc()
6660 if (sdd->sgc) in __sdt_free()
6661 kfree(*per_cpu_ptr(sdd->sgc, j)); in __sdt_free()
6667 free_percpu(sdd->sgc); in __sdt_free()
6668 sdd->sgc = NULL; in __sdt_free()