numa 190 arch/s390/numa/mode_emu.c static struct toptree *node_for_core(struct toptree *numa, struct toptree *core, numa 199 arch/s390/numa/mode_emu.c toptree_for_each(node, numa, NODE) { numa 220 arch/s390/numa/mode_emu.c static void toptree_to_numa_single(struct toptree *numa, struct toptree *phys, numa 226 arch/s390/numa/mode_emu.c node = node_for_core(numa, core, extra); numa 260 arch/s390/numa/mode_emu.c static void move_level_to_numa(struct toptree *numa, struct toptree *phys, numa 265 arch/s390/numa/mode_emu.c toptree_for_each(node, numa, NODE) numa 272 arch/s390/numa/mode_emu.c static void toptree_to_numa_first(struct toptree *numa, struct toptree *phys) numa 277 arch/s390/numa/mode_emu.c move_level_to_numa(numa, phys, DRAWER, true); numa 278 arch/s390/numa/mode_emu.c move_level_to_numa(numa, phys, DRAWER, false); numa 279 arch/s390/numa/mode_emu.c move_level_to_numa(numa, phys, BOOK, true); numa 280 arch/s390/numa/mode_emu.c move_level_to_numa(numa, phys, BOOK, false); numa 281 arch/s390/numa/mode_emu.c move_level_to_numa(numa, phys, MC, true); numa 282 arch/s390/numa/mode_emu.c move_level_to_numa(numa, phys, MC, false); numa 284 arch/s390/numa/mode_emu.c toptree_for_each(core, numa, CORE) numa 330 arch/s390/numa/mode_emu.c struct toptree *numa; numa 335 arch/s390/numa/mode_emu.c numa = toptree_new(TOPTREE_ID_NUMA, emu_nodes); numa 337 arch/s390/numa/mode_emu.c toptree_to_numa_first(numa, phys); numa 340 arch/s390/numa/mode_emu.c toptree_to_numa_single(numa, phys, 0); numa 341 arch/s390/numa/mode_emu.c toptree_to_numa_single(numa, phys, 1); numa 342 arch/s390/numa/mode_emu.c toptree_unify_tree(numa); numa 345 arch/s390/numa/mode_emu.c return numa; numa 396 arch/s390/numa/mode_emu.c static void toptree_to_topology(struct toptree *numa) numa 406 arch/s390/numa/mode_emu.c toptree_for_each(core, numa, CORE) numa 459 arch/s390/numa/mode_emu.c struct toptree *phys, *numa; numa 464 arch/s390/numa/mode_emu.c numa = toptree_to_numa(phys); numa 466 arch/s390/numa/mode_emu.c toptree_to_topology(numa); numa 467 arch/s390/numa/mode_emu.c toptree_free(numa); numa 948 drivers/infiniband/hw/hfi1/file_ops.c int ret, numa; numa 970 drivers/infiniband/hw/hfi1/file_ops.c numa = cpu_to_node(fd->rec_cpu_num); numa 972 drivers/infiniband/hw/hfi1/file_ops.c numa = numa_node_id(); numa 973 drivers/infiniband/hw/hfi1/file_ops.c ret = hfi1_create_ctxtdata(dd->pport, numa, &uctxt); numa 1485 drivers/infiniband/hw/hfi1/hfi.h int hfi1_create_ctxtdata(struct hfi1_pportdata *ppd, int numa, numa 344 drivers/infiniband/hw/hfi1/init.c int hfi1_create_ctxtdata(struct hfi1_pportdata *ppd, int numa, numa 356 drivers/infiniband/hw/hfi1/init.c rcd = kzalloc_node(sizeof(*rcd), GFP_KERNEL, numa); numa 373 drivers/infiniband/hw/hfi1/init.c rcd->numa_id = numa; numa 454 drivers/infiniband/hw/hfi1/init.c GFP_KERNEL, numa); numa 460 drivers/infiniband/hw/hfi1/init.c GFP_KERNEL, numa); numa 480 drivers/infiniband/hw/hfi1/init.c GFP_KERNEL, numa); numa 697 drivers/infiniband/hw/hfi1/pio.c uint hdrqentsize, int numa) numa 714 drivers/infiniband/hw/hfi1/pio.c sc = kzalloc_node(sizeof(*sc), GFP_KERNEL, numa); numa 740 drivers/infiniband/hw/hfi1/pio.c sc->node = numa; numa 857 drivers/infiniband/hw/hfi1/pio.c GFP_KERNEL, numa); numa 285 drivers/infiniband/hw/hfi1/pio.h uint hdrqentsize, int numa); numa 951 drivers/net/ethernet/mellanox/mlx5/core/en_main.c static int mlx5e_alloc_xdpsq_fifo(struct mlx5e_xdpsq *sq, int numa) numa 958 drivers/net/ethernet/mellanox/mlx5/core/en_main.c GFP_KERNEL, numa); numa 969 drivers/net/ethernet/mellanox/mlx5/core/en_main.c static int mlx5e_alloc_xdpsq_db(struct mlx5e_xdpsq *sq, int numa) numa 975 drivers/net/ethernet/mellanox/mlx5/core/en_main.c GFP_KERNEL, numa); numa 979 drivers/net/ethernet/mellanox/mlx5/core/en_main.c err = mlx5e_alloc_xdpsq_fifo(sq, numa); numa 1043 drivers/net/ethernet/mellanox/mlx5/core/en_main.c static int mlx5e_alloc_icosq_db(struct mlx5e_icosq *sq, int numa) numa 1049 drivers/net/ethernet/mellanox/mlx5/core/en_main.c GFP_KERNEL, numa); numa 1108 drivers/net/ethernet/mellanox/mlx5/core/en_main.c static int mlx5e_alloc_txqsq_db(struct mlx5e_txqsq *sq, int numa) numa 1115 drivers/net/ethernet/mellanox/mlx5/core/en_main.c GFP_KERNEL, numa); numa 1118 drivers/net/ethernet/mellanox/mlx5/core/en_main.c GFP_KERNEL, numa); numa 340 kernel/bpf/cpumap.c int numa, err, i; numa 343 kernel/bpf/cpumap.c numa = cpu_to_node(cpu); numa 345 kernel/bpf/cpumap.c rcpu = kzalloc_node(sizeof(*rcpu), gfp, numa); numa 361 kernel/bpf/cpumap.c rcpu->queue = kzalloc_node(sizeof(*rcpu->queue), gfp, numa); numa 374 kernel/bpf/cpumap.c rcpu->kthread = kthread_create_on_node(cpu_map_kthread_run, rcpu, numa, numa 5543 kernel/workqueue.c __ATTR(numa, 0644, wq_numa_show, wq_numa_store), numa 47 tools/vm/slabinfo.c int numa[MAX_NODES]; numa 274 tools/vm/slabinfo.c static void decode_numa_list(int *numa, char *t) numa 279 tools/vm/slabinfo.c memset(numa, 0, MAX_NODES * sizeof(int)); numa 290 tools/vm/slabinfo.c numa[node] = nr; numa 391 tools/vm/slabinfo.c store_size(b, s->numa[node]); numa 1236 tools/vm/slabinfo.c decode_numa_list(slab->numa, t);