alloc_state       117 arch/powerpc/include/asm/spu.h 	enum { SPU_FREE, SPU_USED } alloc_state;
alloc_state       573 arch/powerpc/platforms/cell/spu_base.c 	spu->alloc_state = SPU_FREE;
alloc_state       193 arch/powerpc/platforms/cell/spufs/sched.c 			if (spu->alloc_state != SPU_FREE) {
alloc_state       576 arch/powerpc/platforms/cell/spufs/sched.c 			if (spu && spu->alloc_state == SPU_FREE)
alloc_state       593 arch/powerpc/platforms/cell/spufs/sched.c 			if (spu->alloc_state == SPU_FREE)
alloc_state       604 arch/powerpc/platforms/cell/spufs/sched.c 	spu->alloc_state = SPU_USED;
alloc_state       716 arch/powerpc/platforms/cell/spufs/sched.c 		spu->alloc_state = SPU_USED;
alloc_state       758 arch/powerpc/platforms/cell/spufs/sched.c 		spu->alloc_state = SPU_FREE;
alloc_state      1140 arch/powerpc/platforms/cell/spufs/sched.c 			if (spu->alloc_state != SPU_FREE)
alloc_state      1141 arch/powerpc/platforms/cell/spufs/sched.c 				spu->alloc_state = SPU_FREE;
alloc_state        84 drivers/crypto/cavium/zip/zip_device.h 	u8 alloc_state;
alloc_state      5586 fs/btrfs/extent-tree.c 		find_first_clear_extent_bit(&device->alloc_state, start,
alloc_state      5612 fs/btrfs/extent-tree.c 			set_extent_bits(&device->alloc_state, start,
alloc_state       362 fs/btrfs/extent_map.c 		set_extent_bits_nowait(&device->alloc_state, stripe->physical,
alloc_state       377 fs/btrfs/extent_map.c 		__clear_extent_bit(&device->alloc_state, stripe->physical,
alloc_state       343 fs/btrfs/volumes.c 	extent_io_tree_release(&device->alloc_state);
alloc_state       407 fs/btrfs/volumes.c 	extent_io_tree_init(NULL, &dev->alloc_state, 0, NULL);
alloc_state      1555 fs/btrfs/volumes.c 	if (!find_first_extent_bit(&device->alloc_state, *start,
alloc_state       151 fs/btrfs/volumes.h 	struct extent_io_tree alloc_state;
alloc_state       674 fs/gfs2/bmap.c 	enum alloc_state state;
alloc_state      1987 kernel/sched/topology.c 	enum s_alloc alloc_state = sa_none;
alloc_state      1998 kernel/sched/topology.c 	alloc_state = __visit_domain_allocation_hell(&d, cpu_map);
alloc_state      1999 kernel/sched/topology.c 	if (alloc_state != sa_rootdomain)
alloc_state      2080 kernel/sched/topology.c 	__free_domain_allocs(&d, alloc_state, cpu_map);
alloc_state       406 net/ipv4/netfilter/arp_tables.c 		 struct xt_percpu_counter_alloc_state *alloc_state)
alloc_state       412 net/ipv4/netfilter/arp_tables.c 	if (!xt_percpu_counter_alloc(alloc_state, &e->counters))
alloc_state       523 net/ipv4/netfilter/arp_tables.c 	struct xt_percpu_counter_alloc_state alloc_state = { 0 };
alloc_state       578 net/ipv4/netfilter/arp_tables.c 				       &alloc_state);
alloc_state       518 net/ipv4/netfilter/ip_tables.c 		 struct xt_percpu_counter_alloc_state *alloc_state)
alloc_state       527 net/ipv4/netfilter/ip_tables.c 	if (!xt_percpu_counter_alloc(alloc_state, &e->counters))
alloc_state       664 net/ipv4/netfilter/ip_tables.c 	struct xt_percpu_counter_alloc_state alloc_state = { 0 };
alloc_state       718 net/ipv4/netfilter/ip_tables.c 				       &alloc_state);
alloc_state       537 net/ipv6/netfilter/ip6_tables.c 		 struct xt_percpu_counter_alloc_state *alloc_state)
alloc_state       546 net/ipv6/netfilter/ip6_tables.c 	if (!xt_percpu_counter_alloc(alloc_state, &e->counters))
alloc_state       681 net/ipv6/netfilter/ip6_tables.c 	struct xt_percpu_counter_alloc_state alloc_state = { 0 };
alloc_state       735 net/ipv6/netfilter/ip6_tables.c 				       &alloc_state);