alloc_state 117 arch/powerpc/include/asm/spu.h enum { SPU_FREE, SPU_USED } alloc_state; alloc_state 573 arch/powerpc/platforms/cell/spu_base.c spu->alloc_state = SPU_FREE; alloc_state 193 arch/powerpc/platforms/cell/spufs/sched.c if (spu->alloc_state != SPU_FREE) { alloc_state 576 arch/powerpc/platforms/cell/spufs/sched.c if (spu && spu->alloc_state == SPU_FREE) alloc_state 593 arch/powerpc/platforms/cell/spufs/sched.c if (spu->alloc_state == SPU_FREE) alloc_state 604 arch/powerpc/platforms/cell/spufs/sched.c spu->alloc_state = SPU_USED; alloc_state 716 arch/powerpc/platforms/cell/spufs/sched.c spu->alloc_state = SPU_USED; alloc_state 758 arch/powerpc/platforms/cell/spufs/sched.c spu->alloc_state = SPU_FREE; alloc_state 1140 arch/powerpc/platforms/cell/spufs/sched.c if (spu->alloc_state != SPU_FREE) alloc_state 1141 arch/powerpc/platforms/cell/spufs/sched.c spu->alloc_state = SPU_FREE; alloc_state 84 drivers/crypto/cavium/zip/zip_device.h u8 alloc_state; alloc_state 5586 fs/btrfs/extent-tree.c find_first_clear_extent_bit(&device->alloc_state, start, alloc_state 5612 fs/btrfs/extent-tree.c set_extent_bits(&device->alloc_state, start, alloc_state 362 fs/btrfs/extent_map.c set_extent_bits_nowait(&device->alloc_state, stripe->physical, alloc_state 377 fs/btrfs/extent_map.c __clear_extent_bit(&device->alloc_state, stripe->physical, alloc_state 343 fs/btrfs/volumes.c extent_io_tree_release(&device->alloc_state); alloc_state 407 fs/btrfs/volumes.c extent_io_tree_init(NULL, &dev->alloc_state, 0, NULL); alloc_state 1555 fs/btrfs/volumes.c if (!find_first_extent_bit(&device->alloc_state, *start, alloc_state 151 fs/btrfs/volumes.h struct extent_io_tree alloc_state; alloc_state 674 fs/gfs2/bmap.c enum alloc_state state; alloc_state 1987 kernel/sched/topology.c enum s_alloc alloc_state = sa_none; alloc_state 1998 kernel/sched/topology.c alloc_state = __visit_domain_allocation_hell(&d, cpu_map); alloc_state 1999 kernel/sched/topology.c if (alloc_state != sa_rootdomain) alloc_state 2080 kernel/sched/topology.c __free_domain_allocs(&d, alloc_state, cpu_map); alloc_state 406 net/ipv4/netfilter/arp_tables.c struct xt_percpu_counter_alloc_state *alloc_state) alloc_state 412 net/ipv4/netfilter/arp_tables.c if (!xt_percpu_counter_alloc(alloc_state, &e->counters)) alloc_state 523 net/ipv4/netfilter/arp_tables.c struct xt_percpu_counter_alloc_state alloc_state = { 0 }; alloc_state 578 net/ipv4/netfilter/arp_tables.c &alloc_state); alloc_state 518 net/ipv4/netfilter/ip_tables.c struct xt_percpu_counter_alloc_state *alloc_state) alloc_state 527 net/ipv4/netfilter/ip_tables.c if (!xt_percpu_counter_alloc(alloc_state, &e->counters)) alloc_state 664 net/ipv4/netfilter/ip_tables.c struct xt_percpu_counter_alloc_state alloc_state = { 0 }; alloc_state 718 net/ipv4/netfilter/ip_tables.c &alloc_state); alloc_state 537 net/ipv6/netfilter/ip6_tables.c struct xt_percpu_counter_alloc_state *alloc_state) alloc_state 546 net/ipv6/netfilter/ip6_tables.c if (!xt_percpu_counter_alloc(alloc_state, &e->counters)) alloc_state 681 net/ipv6/netfilter/ip6_tables.c struct xt_percpu_counter_alloc_state alloc_state = { 0 }; alloc_state 735 net/ipv6/netfilter/ip6_tables.c &alloc_state);