refs 46 arch/x86/include/asm/refcount.h : [var] "+m" (r->refs.counter) refs 55 arch/x86/include/asm/refcount.h : [var] "+m" (r->refs.counter) refs 63 arch/x86/include/asm/refcount.h : [var] "+m" (r->refs.counter) refs 72 arch/x86/include/asm/refcount.h r->refs.counter, e, "er", i, "cx"); refs 86 arch/x86/include/asm/refcount.h r->refs.counter, e, "cx"); refs 101 arch/x86/include/asm/refcount.h c = atomic_read(&(r->refs)); refs 111 arch/x86/include/asm/refcount.h : : [var] "m" (r->refs.counter) refs 116 arch/x86/include/asm/refcount.h } while (!atomic_try_cmpxchg(&(r->refs), &c, result)); refs 315 drivers/android/binder.c struct hlist_head refs; refs 1266 drivers/android/binder.c !hlist_empty(&node->refs)) refs 1276 drivers/android/binder.c if (hlist_empty(&node->refs) && !node->local_strong_refs && refs 1491 drivers/android/binder.c hlist_add_head(&new_ref->node_entry, &node->refs); refs 4305 drivers/android/binder.c weak = !hlist_empty(&node->refs) || refs 5343 drivers/android/binder.c static int binder_node_release(struct binder_node *node, int refs) refs 5358 drivers/android/binder.c if (hlist_empty(&node->refs) && node->tmp_refs == 1) { refs 5363 drivers/android/binder.c return refs; refs 5375 drivers/android/binder.c hlist_for_each_entry(ref, &node->refs, node_entry) { refs 5376 drivers/android/binder.c refs++; refs 5401 drivers/android/binder.c node->debug_id, refs, death); refs 5405 drivers/android/binder.c return refs; refs 5673 drivers/android/binder.c hlist_for_each_entry(ref, &node->refs, node_entry) refs 5683 drivers/android/binder.c hlist_for_each_entry(ref, &node->refs, node_entry) refs 588 drivers/base/swnode.c refnode = software_node_fwnode(ref->refs[index].node); refs 607 drivers/base/swnode.c args->args[i] = ref->refs[index].args[i]; refs 111 drivers/block/nbd.c refcount_t refs; refs 248 drivers/block/nbd.c if (refcount_dec_and_mutex_lock(&nbd->refs, refs 1460 drivers/block/nbd.c if (!refcount_inc_not_zero(&nbd->refs)) { refs 1479 drivers/block/nbd.c refcount_inc(&nbd->refs); refs 1724 drivers/block/nbd.c refcount_set(&nbd->refs, 1); refs 1874 drivers/block/nbd.c if (!refcount_inc_not_zero(&nbd->refs)) { refs 2021 drivers/block/nbd.c if (!refcount_inc_not_zero(&nbd->refs)) { refs 2062 drivers/block/nbd.c if (!refcount_inc_not_zero(&nbd->refs)) { refs 2109 drivers/block/nbd.c refcount_inc(&nbd->refs); refs 2421 drivers/block/nbd.c if (refcount_read(&nbd->refs) != 1) refs 61 drivers/block/virtio_blk.c refcount_t refs; refs 409 drivers/block/virtio_blk.c refcount_inc(&vblk->refs); refs 414 drivers/block/virtio_blk.c if (refcount_dec_and_test(&vblk->refs)) { refs 836 drivers/block/virtio_blk.c refcount_set(&vblk->refs, 1); refs 522 drivers/gpu/drm/i915/display/intel_display_power.c int refs = hweight64(power_well->desc->domains & refs 525 drivers/gpu/drm/i915/display/intel_display_power.c WARN_ON(refs > power_well->count); refs 527 drivers/gpu/drm/i915/display/intel_display_power.c return refs; refs 17 drivers/gpu/drm/nouveau/include/nvkm/core/event.h int *refs; refs 155 drivers/gpu/drm/nouveau/nouveau_gem.c if (--vma->refs == 0) { refs 65 drivers/gpu/drm/nouveau/nouveau_vmm.c if (vma && --vma->refs <= 0) { refs 86 drivers/gpu/drm/nouveau/nouveau_vmm.c vma->refs++; refs 93 drivers/gpu/drm/nouveau/nouveau_vmm.c vma->refs = 1; refs 9 drivers/gpu/drm/nouveau/nouveau_vmm.h int refs; refs 31 drivers/gpu/drm/nouveau/nvkm/core/event.c if (--event->refs[index * event->types_nr + type] == 0) { refs 44 drivers/gpu/drm/nouveau/nvkm/core/event.c if (++event->refs[index * event->types_nr + type] == 1) { refs 58 drivers/gpu/drm/nouveau/nvkm/core/event.c if (!event->refs || WARN_ON(index >= event->index_nr)) refs 77 drivers/gpu/drm/nouveau/nvkm/core/event.c if (event->refs) { refs 78 drivers/gpu/drm/nouveau/nvkm/core/event.c kfree(event->refs); refs 79 drivers/gpu/drm/nouveau/nvkm/core/event.c event->refs = NULL; refs 87 drivers/gpu/drm/nouveau/nvkm/core/event.c event->refs = kzalloc(array3_size(index_nr, types_nr, refs 88 drivers/gpu/drm/nouveau/nvkm/core/event.c sizeof(*event->refs)), refs 90 drivers/gpu/drm/nouveau/nvkm/core/event.c if (!event->refs) refs 139 drivers/gpu/drm/nouveau/nvkm/core/notify.c if ((notify->event = event), event->refs) { refs 116 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c u32 refs; refs 133 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c ptc->refs = 0; refs 155 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c if (pt->ptc->refs < 8 /* Heuristic. */ && !force) { refs 157 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c pt->ptc->refs++; refs 195 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/base.c ptc->refs--; refs 150 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (--pgd->refs[0]) { refs 155 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (!pgt->refs[!type]) { refs 191 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (!pgt->refs[!type]) refs 216 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (!pgt->refs[0]) refs 277 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c pgt->refs[type] -= ptes; refs 280 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (desc->type == SPT && (pgt->refs[0] || pgt->refs[1])) refs 284 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (!pgt->refs[type]) { refs 315 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (!pgt->refs[0]) refs 371 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c pgt->refs[type] += ptes; refs 426 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c pgd->refs[0]++; refs 440 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (desc->type == LPT && pgt->refs[1]) { refs 491 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (!pgd->refs[0]) refs 556 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c if (ref && !pgt->refs[desc[it.lvl - 1].type == SPT]) { refs 1075 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.c vmm->pd->refs[0] = 1; refs 15 drivers/gpu/drm/nouveau/nvkm/subdev/mmu/vmm.h u32 refs[2]; refs 337 drivers/gpu/drm/tegra/drm.c struct drm_gem_object **refs; refs 371 drivers/gpu/drm/tegra/drm.c refs = kmalloc_array(num_refs, sizeof(*refs), GFP_KERNEL); refs 372 drivers/gpu/drm/tegra/drm.c if (!refs) { refs 408 drivers/gpu/drm/tegra/drm.c refs[num_refs++] = &obj->gem; refs 438 drivers/gpu/drm/tegra/drm.c refs[num_refs++] = &obj->gem; refs 452 drivers/gpu/drm/tegra/drm.c refs[num_refs++] = &obj->gem; refs 495 drivers/gpu/drm/tegra/drm.c drm_gem_object_put_unlocked(refs[num_refs]); refs 497 drivers/gpu/drm/tegra/drm.c kfree(refs); refs 683 drivers/misc/sgi-xp/xpc.h s32 refs = atomic_dec_return(&ch->references); refs 685 drivers/misc/sgi-xp/xpc.h DBUG_ON(refs < 0); refs 686 drivers/misc/sgi-xp/xpc.h if (refs == 0) refs 700 drivers/misc/sgi-xp/xpc.h s32 refs = atomic_dec_return(&part->references); refs 702 drivers/misc/sgi-xp/xpc.h DBUG_ON(refs < 0); refs 703 drivers/misc/sgi-xp/xpc.h if (refs == 0 && part->setup_state == XPC_P_SS_WTEARDOWN) refs 488 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c refcount_inc(&dpaa_bp_array[bpid]->refs); refs 499 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c refcount_set(&dpaa_bp->refs, 1); refs 587 drivers/net/ethernet/freescale/dpaa/dpaa_eth.c if (!refcount_dec_and_test(&bp->refs)) refs 103 drivers/net/ethernet/freescale/dpaa/dpaa_eth.h refcount_t refs; refs 746 drivers/net/ethernet/mellanox/mlx4/mlx4.h int refs[MLX4_MAX_MAC_NUM]; refs 769 drivers/net/ethernet/mellanox/mlx4/mlx4.h int refs[MLX4_MAX_VLAN_NUM]; refs 69 drivers/net/ethernet/mellanox/mlx4/port.c table->refs[i] = 0; refs 83 drivers/net/ethernet/mellanox/mlx4/port.c table->refs[i] = 0; refs 118 drivers/net/ethernet/mellanox/mlx4/port.c if (table->refs[i] && refs 157 drivers/net/ethernet/mellanox/mlx4/port.c if (!table->refs[i]) refs 232 drivers/net/ethernet/mellanox/mlx4/port.c dup_table->refs[index_at_port]) { refs 242 drivers/net/ethernet/mellanox/mlx4/port.c if (!table->refs[index_at_dup_port] || refs 251 drivers/net/ethernet/mellanox/mlx4/port.c if (!table->refs[i]) { refs 255 drivers/net/ethernet/mellanox/mlx4/port.c if (!dup_table->refs[i]) refs 265 drivers/net/ethernet/mellanox/mlx4/port.c ++table->refs[i]; refs 308 drivers/net/ethernet/mellanox/mlx4/port.c table->refs[free] = 1; refs 312 drivers/net/ethernet/mellanox/mlx4/port.c dup_table->refs[free] = 0; refs 412 drivers/net/ethernet/mellanox/mlx4/port.c if (--table->refs[index] || table->is_dup[index]) { refs 415 drivers/net/ethernet/mellanox/mlx4/port.c if (!table->refs[index]) refs 427 drivers/net/ethernet/mellanox/mlx4/port.c if (dup_table->refs[index]) refs 564 drivers/net/ethernet/mellanox/mlx4/port.c if (table->refs[i] && refs 632 drivers/net/ethernet/mellanox/mlx4/port.c dup_table->refs[index_at_port]) { refs 642 drivers/net/ethernet/mellanox/mlx4/port.c if (!table->refs[index_at_dup_port] || refs 651 drivers/net/ethernet/mellanox/mlx4/port.c if (!table->refs[i]) { refs 655 drivers/net/ethernet/mellanox/mlx4/port.c if (!dup_table->refs[i]) refs 660 drivers/net/ethernet/mellanox/mlx4/port.c if ((table->refs[i] || table->is_dup[i]) && refs 666 drivers/net/ethernet/mellanox/mlx4/port.c ++table->refs[i]; refs 697 drivers/net/ethernet/mellanox/mlx4/port.c table->refs[free] = 1; refs 704 drivers/net/ethernet/mellanox/mlx4/port.c table->refs[free] = 0; refs 710 drivers/net/ethernet/mellanox/mlx4/port.c dup_table->refs[free] = 0; refs 792 drivers/net/ethernet/mellanox/mlx4/port.c if (--table->refs[index] || table->is_dup[index]) { refs 794 drivers/net/ethernet/mellanox/mlx4/port.c table->refs[index], index); refs 795 drivers/net/ethernet/mellanox/mlx4/port.c if (!table->refs[index]) refs 805 drivers/net/ethernet/mellanox/mlx4/port.c if (dup_table->refs[index]) refs 920 drivers/net/ethernet/mellanox/mlx4/port.c if (!t1->refs[i]) { refs 925 drivers/net/ethernet/mellanox/mlx4/port.c if (!t2->refs[i]) { refs 1027 drivers/net/ethernet/mellanox/mlx4/port.c if (!t1->refs[i]) { refs 1032 drivers/net/ethernet/mellanox/mlx4/port.c if (!t2->refs[i]) { refs 527 drivers/pci/controller/pci-hyperv.c refcount_t refs; refs 580 drivers/pci/controller/pci-hyperv.c refcount_inc(&hpdev->refs); refs 585 drivers/pci/controller/pci-hyperv.c if (refcount_dec_and_test(&hpdev->refs)) refs 1904 drivers/pci/controller/pci-hyperv.c refcount_set(&hpdev->refs, 1); refs 43 drivers/staging/most/core.c int refs; refs 935 drivers/staging/most/core.c if (c->pipe0.refs && c->pipe0.comp->tx_completion) refs 938 drivers/staging/most/core.c if (c->pipe1.refs && c->pipe1.comp->tx_completion) refs 1048 drivers/staging/most/core.c if (c->pipe0.refs && c->pipe1.refs && refs 1081 drivers/staging/most/core.c if (c->pipe0.refs && c->pipe1.refs && refs 1153 drivers/staging/most/core.c if (c->pipe0.refs && c->pipe0.comp->rx_completion && refs 1157 drivers/staging/most/core.c if (c->pipe1.refs && c->pipe1.comp->rx_completion && refs 1186 drivers/staging/most/core.c if (c->pipe0.refs + c->pipe1.refs > 0) refs 1226 drivers/staging/most/core.c c->pipe0.refs++; refs 1228 drivers/staging/most/core.c c->pipe1.refs++; refs 1259 drivers/staging/most/core.c if (c->pipe0.refs + c->pipe1.refs >= 2) refs 1292 drivers/staging/most/core.c c->pipe0.refs--; refs 1294 drivers/staging/most/core.c c->pipe1.refs--; refs 572 drivers/tty/tty_io.c int refs; refs 619 drivers/tty/tty_io.c refs = tty_signal_session_leader(tty, exit_session); refs 621 drivers/tty/tty_io.c while (refs--) refs 194 drivers/tty/tty_jobctrl.c int refs = 0; refs 205 drivers/tty/tty_jobctrl.c refs++; refs 230 drivers/tty/tty_jobctrl.c return refs; refs 49 drivers/xen/gntdev-dmabuf.c grant_ref_t *refs; refs 471 drivers/xen/gntdev-dmabuf.c int count, u32 domid, u32 *refs, u32 *fd) refs 483 drivers/xen/gntdev-dmabuf.c map->grants[i].ref = refs[i]; refs 522 drivers/xen/gntdev-dmabuf.c dmabuf_imp_grant_foreign_access(struct page **pages, u32 *refs, refs 546 drivers/xen/gntdev-dmabuf.c refs[i] = cur_ref; refs 556 drivers/xen/gntdev-dmabuf.c static void dmabuf_imp_end_foreign_access(u32 *refs, int count) refs 561 drivers/xen/gntdev-dmabuf.c if (refs[i] != GRANT_INVALID_REF) refs 562 drivers/xen/gntdev-dmabuf.c gnttab_end_foreign_access(refs[i], 0, 0UL); refs 568 drivers/xen/gntdev-dmabuf.c kfree(gntdev_dmabuf->u.imp.refs); refs 581 drivers/xen/gntdev-dmabuf.c gntdev_dmabuf->u.imp.refs = kcalloc(count, refs 582 drivers/xen/gntdev-dmabuf.c sizeof(gntdev_dmabuf->u.imp.refs[0]), refs 584 drivers/xen/gntdev-dmabuf.c if (!gntdev_dmabuf->u.imp.refs) refs 596 drivers/xen/gntdev-dmabuf.c gntdev_dmabuf->u.imp.refs[i] = GRANT_INVALID_REF; refs 672 drivers/xen/gntdev-dmabuf.c gntdev_dmabuf->u.imp.refs, refs 686 drivers/xen/gntdev-dmabuf.c dmabuf_imp_end_foreign_access(gntdev_dmabuf->u.imp.refs, count); refs 732 drivers/xen/gntdev-dmabuf.c dmabuf_imp_end_foreign_access(gntdev_dmabuf->u.imp.refs, refs 762 drivers/xen/gntdev-dmabuf.c u32 *refs; refs 777 drivers/xen/gntdev-dmabuf.c refs = kcalloc(op.count, sizeof(*refs), GFP_KERNEL); refs 778 drivers/xen/gntdev-dmabuf.c if (!refs) refs 781 drivers/xen/gntdev-dmabuf.c if (copy_from_user(refs, u->refs, sizeof(*refs) * op.count) != 0) { refs 787 drivers/xen/gntdev-dmabuf.c op.domid, refs, &op.fd); refs 795 drivers/xen/gntdev-dmabuf.c kfree(refs); refs 830 drivers/xen/gntdev-dmabuf.c if (copy_to_user(u->refs, gntdev_dmabuf->u.imp.refs, refs 831 drivers/xen/gntdev-dmabuf.c sizeof(*u->refs) * op.count) != 0) { refs 685 drivers/xen/gntdev.c if (copy_from_user(map->grants, &u->refs, refs 1115 fs/btrfs/backref.c u64 time_seq, struct ulist *refs, refs 1182 fs/btrfs/backref.c refcount_inc(&head->refs); refs 1307 fs/btrfs/backref.c ret = ulist_add_merge_ptr(refs, ref->parent, refs 1902 fs/btrfs/backref.c struct ulist *refs = NULL; refs 1929 fs/btrfs/backref.c tree_mod_seq_elem.seq, &refs, refs 1935 fs/btrfs/backref.c while (!ret && (ref_node = ulist_next(refs, &ref_uiter))) { refs 1957 fs/btrfs/backref.c free_leaf_list(refs); refs 140 fs/btrfs/ctree.c if (atomic_inc_not_zero(&eb->refs)) { refs 891 fs/btrfs/ctree.c u64 refs; refs 917 fs/btrfs/ctree.c &refs, &flags); refs 920 fs/btrfs/ctree.c if (refs == 0) { refs 926 fs/btrfs/ctree.c refs = 1; refs 938 fs/btrfs/ctree.c if (refs > 1) { refs 1037 fs/btrfs/ctree.h refcount_t refs; refs 1600 fs/btrfs/ctree.h BTRFS_SETGET_FUNCS(extent_refs, struct btrfs_extent_item, refs, 64); refs 1605 fs/btrfs/ctree.h BTRFS_SETGET_FUNCS(extent_refs_v0, struct btrfs_extent_item_v0, refs, 32); refs 1942 fs/btrfs/ctree.h BTRFS_SETGET_FUNCS(disk_root_refs, struct btrfs_root_item, refs, 32); refs 1951 fs/btrfs/ctree.h BTRFS_SETGET_STACK_FUNCS(root_refs, struct btrfs_root_item, refs, 32); refs 2407 fs/btrfs/ctree.h u64 offset, int metadata, u64 *refs, u64 *flags); refs 46 fs/btrfs/delayed-inode.c refcount_set(&delayed_node->refs, 0); refs 75 fs/btrfs/delayed-inode.c refcount_inc(&node->refs); refs 84 fs/btrfs/delayed-inode.c refcount_inc(&node->refs); /* can be accessed */ refs 106 fs/btrfs/delayed-inode.c if (refcount_inc_not_zero(&node->refs)) { refs 107 fs/btrfs/delayed-inode.c refcount_inc(&node->refs); refs 141 fs/btrfs/delayed-inode.c refcount_set(&node->refs, 2); refs 182 fs/btrfs/delayed-inode.c refcount_inc(&node->refs); /* inserted into list */ refs 196 fs/btrfs/delayed-inode.c refcount_dec(&node->refs); /* not in the list */ refs 217 fs/btrfs/delayed-inode.c refcount_inc(&node->refs); refs 244 fs/btrfs/delayed-inode.c refcount_inc(&next->refs); refs 269 fs/btrfs/delayed-inode.c if (refcount_dec_and_test(&delayed_node->refs)) { refs 277 fs/btrfs/delayed-inode.c ASSERT(refcount_read(&delayed_node->refs) == 0); refs 303 fs/btrfs/delayed-inode.c refcount_inc(&node->refs); refs 325 fs/btrfs/delayed-inode.c refcount_set(&item->refs, 1); refs 503 fs/btrfs/delayed-inode.c if (refcount_dec_and_test(&item->refs)) refs 1611 fs/btrfs/delayed-inode.c refcount_inc(&item->refs); refs 1618 fs/btrfs/delayed-inode.c refcount_inc(&item->refs); refs 1632 fs/btrfs/delayed-inode.c refcount_dec(&delayed_node->refs); refs 1645 fs/btrfs/delayed-inode.c if (refcount_dec_and_test(&curr->refs)) refs 1651 fs/btrfs/delayed-inode.c if (refcount_dec_and_test(&curr->refs)) refs 1706 fs/btrfs/delayed-inode.c if (refcount_dec_and_test(&curr->refs)) refs 1723 fs/btrfs/delayed-inode.c if (refcount_dec_and_test(&curr->refs)) refs 1970 fs/btrfs/delayed-inode.c if (!refcount_inc_not_zero(&delayed_nodes[i]->refs)) refs 57 fs/btrfs/delayed-inode.h refcount_t refs; refs 70 fs/btrfs/delayed-inode.h refcount_t refs; refs 401 fs/btrfs/delayed-ref.c refcount_inc(&head->refs); refs 767 fs/btrfs/delayed-ref.c refcount_set(&head_ref->refs, 1); refs 893 fs/btrfs/delayed-ref.c refcount_set(&ref->refs, 1); refs 36 fs/btrfs/delayed-ref.h refcount_t refs; refs 74 fs/btrfs/delayed-ref.h refcount_t refs; refs 311 fs/btrfs/delayed-ref.h WARN_ON(refcount_read(&ref->refs) == 0); refs 312 fs/btrfs/delayed-ref.h if (refcount_dec_and_test(&ref->refs)) { refs 331 fs/btrfs/delayed-ref.h if (refcount_dec_and_test(&head->refs)) refs 1006 fs/btrfs/disk-io.c BUG_ON(!atomic_read(&eb->refs)); refs 1167 fs/btrfs/disk-io.c refcount_set(&root->refs, 1); refs 100 fs/btrfs/disk-io.h if (refcount_inc_not_zero(&root->refs)) refs 107 fs/btrfs/disk-io.h if (refcount_dec_and_test(&root->refs)) refs 152 fs/btrfs/extent-tree.c u64 offset, int metadata, u64 *refs, u64 *flags) refs 241 fs/btrfs/extent-tree.c refcount_inc(&head->refs); refs 268 fs/btrfs/extent-tree.c if (refs) refs 269 fs/btrfs/extent-tree.c *refs = num_refs; refs 1023 fs/btrfs/extent-tree.c u64 refs; refs 1037 fs/btrfs/extent-tree.c refs = btrfs_extent_refs(leaf, ei); refs 1038 fs/btrfs/extent-tree.c refs += refs_to_add; refs 1039 fs/btrfs/extent-tree.c btrfs_set_extent_refs(leaf, ei, refs); refs 1117 fs/btrfs/extent-tree.c u64 refs; refs 1120 fs/btrfs/extent-tree.c refs = btrfs_extent_refs(leaf, ei); refs 1121 fs/btrfs/extent-tree.c WARN_ON(refs_to_mod < 0 && refs + refs_to_mod <= 0); refs 1122 fs/btrfs/extent-tree.c refs += refs_to_mod; refs 1123 fs/btrfs/extent-tree.c btrfs_set_extent_refs(leaf, ei, refs); refs 1136 fs/btrfs/extent-tree.c refs = btrfs_extent_data_ref_count(leaf, dref); refs 1139 fs/btrfs/extent-tree.c refs = btrfs_shared_data_ref_count(leaf, sref); refs 1141 fs/btrfs/extent-tree.c refs = 1; refs 1145 fs/btrfs/extent-tree.c BUG_ON(refs_to_mod < 0 && refs < -refs_to_mod); refs 1146 fs/btrfs/extent-tree.c refs += refs_to_mod; refs 1148 fs/btrfs/extent-tree.c if (refs > 0) { refs 1150 fs/btrfs/extent-tree.c btrfs_set_extent_data_ref_count(leaf, dref, refs); refs 1152 fs/btrfs/extent-tree.c btrfs_set_shared_data_ref_count(leaf, sref, refs); refs 1443 fs/btrfs/extent-tree.c u64 refs; refs 1467 fs/btrfs/extent-tree.c refs = btrfs_extent_refs(leaf, item); refs 1468 fs/btrfs/extent-tree.c btrfs_set_extent_refs(leaf, item, refs + refs_to_add); refs 2203 fs/btrfs/extent-tree.c refcount_inc(&head->refs); refs 2271 fs/btrfs/extent-tree.c refcount_inc(&head->refs); refs 2967 fs/btrfs/extent-tree.c u64 refs; refs 3102 fs/btrfs/extent-tree.c refs = btrfs_extent_refs(leaf, ei); refs 3103 fs/btrfs/extent-tree.c if (refs < refs_to_drop) { refs 3106 fs/btrfs/extent-tree.c refs_to_drop, refs, bytenr); refs 3111 fs/btrfs/extent-tree.c refs -= refs_to_drop; refs 3113 fs/btrfs/extent-tree.c if (refs > 0) { refs 3123 fs/btrfs/extent-tree.c btrfs_set_extent_refs(leaf, ei, refs); refs 4577 fs/btrfs/extent-tree.c u64 refs[BTRFS_MAX_LEVEL]; refs 4603 fs/btrfs/extent-tree.c u64 refs; refs 4641 fs/btrfs/extent-tree.c wc->level - 1, 1, &refs, refs 4646 fs/btrfs/extent-tree.c BUG_ON(refs == 0); refs 4649 fs/btrfs/extent-tree.c if (refs == 1) refs 4703 fs/btrfs/extent-tree.c ((wc->stage == DROP_REFERENCE && wc->refs[level] != 1) || refs 4708 fs/btrfs/extent-tree.c &wc->refs[level], refs 4713 fs/btrfs/extent-tree.c BUG_ON(wc->refs[level] == 0); refs 4717 fs/btrfs/extent-tree.c if (wc->refs[level] > 1) refs 4841 fs/btrfs/extent-tree.c &wc->refs[level - 1], refs 4846 fs/btrfs/extent-tree.c if (unlikely(wc->refs[level - 1] == 0)) { refs 4854 fs/btrfs/extent-tree.c if (wc->refs[level - 1] > 1) { refs 4916 fs/btrfs/extent-tree.c wc->refs[level - 1] = 0; refs 5043 fs/btrfs/extent-tree.c &wc->refs[level], refs 5050 fs/btrfs/extent-tree.c BUG_ON(wc->refs[level] == 0); refs 5051 fs/btrfs/extent-tree.c if (wc->refs[level] == 1) { refs 5060 fs/btrfs/extent-tree.c BUG_ON(wc->refs[level] > 1 && !path->locks[level]); refs 5062 fs/btrfs/extent-tree.c if (wc->refs[level] == 1) { refs 5101 fs/btrfs/extent-tree.c btrfs_free_tree_block(trans, root, eb, parent, wc->refs[level] == 1); refs 5103 fs/btrfs/extent-tree.c wc->refs[level] = 0; refs 5284 fs/btrfs/extent-tree.c level, 1, &wc->refs[level], refs 5290 fs/btrfs/extent-tree.c BUG_ON(wc->refs[level] == 0); refs 5297 fs/btrfs/extent-tree.c WARN_ON(wc->refs[level] != 1); refs 5468 fs/btrfs/extent-tree.c wc->refs[parent_level] = 1; refs 73 fs/btrfs/extent_io.c refcount_read(&state->refs)); refs 81 fs/btrfs/extent_io.c eb->start, eb->len, atomic_read(&eb->refs), eb->bflags); refs 308 fs/btrfs/extent_io.c refcount_set(&state->refs, 1); refs 318 fs/btrfs/extent_io.c if (refcount_dec_and_test(&state->refs)) { refs 716 fs/btrfs/extent_io.c refcount_dec(&cached->refs); refs 868 fs/btrfs/extent_io.c refcount_inc(&state->refs); refs 913 fs/btrfs/extent_io.c refcount_inc(&state->refs); refs 1719 fs/btrfs/extent_io.c refcount_inc(&state->refs); refs 3032 fs/btrfs/extent_io.c refcount_inc(&em->refs); refs 3043 fs/btrfs/extent_io.c refcount_inc(&em->refs); refs 3997 fs/btrfs/extent_io.c ret = atomic_inc_not_zero(&eb->refs); refs 4940 fs/btrfs/extent_io.c atomic_set(&eb->refs, 1); refs 5027 fs/btrfs/extent_io.c int refs; refs 5048 fs/btrfs/extent_io.c refs = atomic_read(&eb->refs); refs 5049 fs/btrfs/extent_io.c if (refs >= 2 && test_bit(EXTENT_BUFFER_TREE_REF, &eb->bflags)) refs 5054 fs/btrfs/extent_io.c atomic_inc(&eb->refs); refs 5082 fs/btrfs/extent_io.c if (eb && atomic_inc_not_zero(&eb->refs)) { refs 5198 fs/btrfs/extent_io.c if (atomic_inc_not_zero(&exists->refs)) { refs 5265 fs/btrfs/extent_io.c WARN_ON(!atomic_dec_and_test(&eb->refs)); refs 5287 fs/btrfs/extent_io.c WARN_ON(atomic_read(&eb->refs) == 0); refs 5288 fs/btrfs/extent_io.c if (atomic_dec_and_test(&eb->refs)) { refs 5320 fs/btrfs/extent_io.c int refs; refs 5326 fs/btrfs/extent_io.c refs = atomic_read(&eb->refs); refs 5327 fs/btrfs/extent_io.c if ((!test_bit(EXTENT_BUFFER_UNMAPPED, &eb->bflags) && refs <= 3) refs 5329 fs/btrfs/extent_io.c refs == 1)) refs 5331 fs/btrfs/extent_io.c old = atomic_cmpxchg(&eb->refs, refs, refs - 1); refs 5332 fs/btrfs/extent_io.c if (old == refs) refs 5337 fs/btrfs/extent_io.c if (atomic_read(&eb->refs) == 2 && refs 5341 fs/btrfs/extent_io.c atomic_dec(&eb->refs); refs 5358 fs/btrfs/extent_io.c if (atomic_read(&eb->refs) == 2 && !extent_buffer_under_io(eb) && refs 5360 fs/btrfs/extent_io.c atomic_dec(&eb->refs); refs 5389 fs/btrfs/extent_io.c WARN_ON(atomic_read(&eb->refs) == 0); refs 5403 fs/btrfs/extent_io.c WARN_ON(atomic_read(&eb->refs) == 0); refs 6111 fs/btrfs/extent_io.c if (atomic_read(&eb->refs) != 1 || extent_buffer_under_io(eb)) { refs 146 fs/btrfs/extent_io.h refcount_t refs; refs 164 fs/btrfs/extent_io.h atomic_t refs; refs 447 fs/btrfs/extent_io.h atomic_inc(&eb->refs); refs 60 fs/btrfs/extent_map.c refcount_set(&em->refs, 1); refs 76 fs/btrfs/extent_map.c WARN_ON(refcount_read(&em->refs) == 0); refs 77 fs/btrfs/extent_map.c if (refcount_dec_and_test(&em->refs)) { refs 244 fs/btrfs/extent_map.c if (refcount_read(&em->refs) > 2) refs 342 fs/btrfs/extent_map.c refcount_inc(&em->refs); refs 438 fs/btrfs/extent_map.c refcount_inc(&em->refs); refs 54 fs/btrfs/extent_map.h refcount_t refs; refs 207 fs/btrfs/ordered-data.c refcount_set(&entry->refs, 1); refs 358 fs/btrfs/ordered-data.c refcount_inc(&entry->refs); refs 422 fs/btrfs/ordered-data.c refcount_inc(&entry->refs); refs 439 fs/btrfs/ordered-data.c if (refcount_dec_and_test(&entry->refs)) { refs 546 fs/btrfs/ordered-data.c refcount_inc(&ordered->refs); refs 729 fs/btrfs/ordered-data.c refcount_inc(&entry->refs); refs 770 fs/btrfs/ordered-data.c refcount_inc(&entry->refs); refs 793 fs/btrfs/ordered-data.c refcount_inc(&entry->refs); refs 1010 fs/btrfs/ordered-data.c refcount_dec(&cache->refs); refs 102 fs/btrfs/ordered-data.h refcount_t refs; refs 156 fs/btrfs/print-tree.c atomic_read(&eb->refs), eb->write_locks, refs 143 fs/btrfs/raid56.c refcount_t refs; refs 389 fs/btrfs/raid56.c refcount_dec(&rbio->refs); refs 480 fs/btrfs/raid56.c refcount_inc(&rbio->refs); refs 694 fs/btrfs/raid56.c refcount_dec(&cur->refs); refs 743 fs/btrfs/raid56.c refcount_inc(&rbio->refs); refs 789 fs/btrfs/raid56.c refcount_dec(&rbio->refs); refs 806 fs/btrfs/raid56.c refcount_inc(&next->refs); refs 839 fs/btrfs/raid56.c if (!refcount_dec_and_test(&rbio->refs)) refs 1010 fs/btrfs/raid56.c refcount_set(&rbio->refs, 1); refs 70 fs/btrfs/ref-verify.c struct rb_root refs; refs 245 fs/btrfs/ref-verify.c while((n = rb_first(&be->refs))) { refs 247 fs/btrfs/ref-verify.c rb_erase(&ref->node, &be->refs); refs 298 fs/btrfs/ref-verify.c be->refs = RB_ROOT; refs 342 fs/btrfs/ref-verify.c exist = insert_ref_entry(&be->refs, ref); refs 371 fs/btrfs/ref-verify.c if (insert_ref_entry(&be->refs, ref)) { refs 409 fs/btrfs/ref-verify.c if (insert_ref_entry(&be->refs, ref)) { refs 643 fs/btrfs/ref-verify.c for (n = rb_first(&be->refs); n; n = rb_next(n)) { refs 817 fs/btrfs/ref-verify.c exist = insert_ref_entry(&be->refs, ref); refs 831 fs/btrfs/ref-verify.c rb_erase(&exist->node, &be->refs); refs 57 fs/btrfs/scrub.c refcount_t refs; refs 72 fs/btrfs/scrub.c atomic_t refs; refs 105 fs/btrfs/scrub.c refcount_t refs; /* free mem on transition to zero */ refs 135 fs/btrfs/scrub.c refcount_t refs; refs 190 fs/btrfs/scrub.c refcount_t refs; refs 205 fs/btrfs/scrub.c u64 refs; refs 266 fs/btrfs/scrub.c refcount_inc(&sctx->refs); refs 339 fs/btrfs/scrub.c entry->refs++; refs 351 fs/btrfs/scrub.c ret->refs = 1; refs 505 fs/btrfs/scrub.c if (fstripe_lock->refs == 0) { refs 510 fs/btrfs/scrub.c fstripe_lock->refs--; refs 513 fs/btrfs/scrub.c if (fstripe_lock->refs == 0) { refs 571 fs/btrfs/scrub.c if (refcount_dec_and_test(&sctx->refs)) refs 584 fs/btrfs/scrub.c refcount_set(&sctx->refs, 1); refs 801 fs/btrfs/scrub.c refcount_inc(&recover->refs); refs 807 fs/btrfs/scrub.c if (refcount_dec_and_test(&recover->refs)) { refs 1318 fs/btrfs/scrub.c refcount_set(&recover->refs, 1); refs 1989 fs/btrfs/scrub.c refcount_inc(&sblock->refs); refs 1994 fs/btrfs/scrub.c if (refcount_dec_and_test(&sblock->refs)) { refs 2008 fs/btrfs/scrub.c atomic_inc(&spage->refs); refs 2013 fs/btrfs/scrub.c if (atomic_dec_and_test(&spage->refs)) { refs 2239 fs/btrfs/scrub.c refcount_set(&sblock->refs, 1); refs 2544 fs/btrfs/scrub.c refcount_set(&sblock->refs, 1); refs 2807 fs/btrfs/scrub.c refcount_inc(&sparity->refs); refs 2812 fs/btrfs/scrub.c if (!refcount_dec_and_test(&sparity->refs)) refs 2864 fs/btrfs/scrub.c refcount_set(&sparity->refs, 1); refs 4209 fs/btrfs/send.c void *ctx, struct list_head *refs) refs 4232 fs/btrfs/send.c ret = __record_ref(refs, dir, gen, p); refs 6030 fs/btrfs/send.c struct list_head *refs; refs 6040 fs/btrfs/send.c ppctx->refs); refs 6066 fs/btrfs/send.c ctx.refs = &deleted_refs; refs 21 fs/btrfs/tests/extent-map-tests.c if (refcount_read(&em->refs) != 1) { refs 25 fs/btrfs/tests/extent-map-tests.c em->block_len, refcount_read(&em->refs)); refs 27 fs/btrfs/tests/extent-map-tests.c refcount_set(&em->refs, 1); refs 74 fs/btrfs/tests/qgroup-tests.c u64 refs; refs 99 fs/btrfs/tests/qgroup-tests.c refs = btrfs_extent_refs(path->nodes[0], item); refs 100 fs/btrfs/tests/qgroup-tests.c btrfs_set_extent_refs(path->nodes[0], item, refs + 1); refs 158 fs/btrfs/tests/qgroup-tests.c u64 refs; refs 183 fs/btrfs/tests/qgroup-tests.c refs = btrfs_extent_refs(path->nodes[0], item); refs 184 fs/btrfs/tests/qgroup-tests.c btrfs_set_extent_refs(path->nodes[0], item, refs - 1); refs 4427 fs/btrfs/tree-log.c refcount_inc(&em->refs); refs 5610 fs/btrfs/volumes.c refcount_set(&bbio->refs, 1); refs 5617 fs/btrfs/volumes.c WARN_ON(!refcount_read(&bbio->refs)); refs 5618 fs/btrfs/volumes.c refcount_inc(&bbio->refs); refs 5625 fs/btrfs/volumes.c if (refcount_dec_and_test(&bbio->refs)) refs 327 fs/btrfs/volumes.h refcount_t refs; refs 252 fs/cifs/dfs_cache.c static inline void dump_refs(const struct dfs_info3_param *refs, int numrefs) refs 258 fs/cifs/dfs_cache.c const struct dfs_info3_param *ref = &refs[i]; refs 369 fs/cifs/dfs_cache.c static int copy_ref_data(const struct dfs_info3_param *refs, int numrefs, refs 374 fs/cifs/dfs_cache.c ce->ce_ttl = refs[0].ttl; refs 376 fs/cifs/dfs_cache.c ce->ce_srvtype = refs[0].server_type; refs 377 fs/cifs/dfs_cache.c ce->ce_flags = refs[0].ref_flag; refs 378 fs/cifs/dfs_cache.c ce->ce_path_consumed = refs[0].path_consumed; refs 383 fs/cifs/dfs_cache.c t = alloc_tgt(refs[i].node_name); refs 405 fs/cifs/dfs_cache.c alloc_cache_entry(const char *path, const struct dfs_info3_param *refs, refs 423 fs/cifs/dfs_cache.c rc = copy_ref_data(refs, numrefs, ce, NULL); refs 458 fs/cifs/dfs_cache.c const struct dfs_info3_param *refs, int numrefs) refs 462 fs/cifs/dfs_cache.c ce = alloc_cache_entry(path, refs, numrefs); refs 564 fs/cifs/dfs_cache.c __update_cache_entry(const char *path, const struct dfs_info3_param *refs, refs 586 fs/cifs/dfs_cache.c rc = copy_ref_data(refs, numrefs, ce, th); refs 602 fs/cifs/dfs_cache.c struct dfs_info3_param *refs = NULL; refs 617 fs/cifs/dfs_cache.c rc = ses->server->ops->get_dfs_refer(xid, ses, path, &refs, &numrefs, refs 622 fs/cifs/dfs_cache.c ce = __update_cache_entry(path, refs, numrefs); refs 624 fs/cifs/dfs_cache.c dump_refs(refs, numrefs); refs 625 fs/cifs/dfs_cache.c free_dfs_info_array(refs, numrefs); refs 1384 fs/cifs/dfs_cache.c struct dfs_info3_param *refs = NULL; refs 1423 fs/cifs/dfs_cache.c rc = ses->server->ops->get_dfs_refer(xid, ses, path, &refs, refs 1428 fs/cifs/dfs_cache.c ce = __update_cache_entry(npath, refs, numrefs); refs 1430 fs/cifs/dfs_cache.c dump_refs(refs, numrefs); refs 1431 fs/cifs/dfs_cache.c free_dfs_info_array(refs, numrefs); refs 709 fs/file.c static struct file *__fget(unsigned int fd, fmode_t mask, unsigned int refs) refs 724 fs/file.c else if (!get_file_rcu_many(file, refs)) refs 732 fs/file.c struct file *fget_many(unsigned int fd, unsigned int refs) refs 734 fs/file.c return __fget(fd, FMODE_PATH, refs); refs 334 fs/file_table.c void fput_many(struct file *file, unsigned int refs) refs 336 fs/file_table.c if (atomic_long_sub_and_test(refs, &file->f_count)) { refs 510 fs/hugetlbfs/inode.c resv_map_release(&resv_map->refs); refs 809 fs/hugetlbfs/inode.c kref_put(&resv_map->refs, resv_map_release); refs 178 fs/io_uring.c struct percpu_ref refs; refs 319 fs/io_uring.c refcount_t refs; refs 391 fs/io_uring.c struct io_ring_ctx *ctx = container_of(ref, struct io_ring_ctx, refs); refs 405 fs/io_uring.c if (percpu_ref_init(&ctx->refs, io_ring_ctx_ref_free, refs 620 fs/io_uring.c if (!percpu_ref_tryget(&ctx->refs)) refs 657 fs/io_uring.c refcount_set(&req->refs, 2); refs 662 fs/io_uring.c percpu_ref_put(&ctx->refs); refs 670 fs/io_uring.c percpu_ref_put_many(&ctx->refs, *nr); refs 679 fs/io_uring.c percpu_ref_put(&req->ctx->refs); refs 743 fs/io_uring.c if (refcount_dec_and_test(&req->refs)) refs 780 fs/io_uring.c if (refcount_dec_and_test(&req->refs)) { refs 2717 fs/io_uring.c refcount_dec(&shadow_req->refs); refs 2911 fs/io_uring.c refcount_dec(&shadow_req->refs); refs 3641 fs/io_uring.c percpu_ref_exit(&ctx->refs); refs 3681 fs/io_uring.c percpu_ref_kill(&ctx->refs); refs 3751 fs/io_uring.c if (!percpu_ref_tryget(&ctx->refs)) refs 3787 fs/io_uring.c percpu_ref_put(&ctx->refs); refs 4025 fs/io_uring.c if (percpu_ref_is_dying(&ctx->refs)) refs 4028 fs/io_uring.c percpu_ref_kill(&ctx->refs); refs 4079 fs/io_uring.c percpu_ref_reinit(&ctx->refs); refs 283 fs/ocfs2/dlm/dlmcommon.h struct kref refs; refs 868 fs/ocfs2/dlm/dlmcommon.h kref_get(&res->refs); refs 94 fs/ocfs2/dlm/dlmdebug.c res->last_used, kref_read(&res->refs), refs 509 fs/ocfs2/dlm/dlmdebug.c kref_read(&res->refs)); refs 471 fs/ocfs2/dlm/dlmmaster.c res = container_of(kref, struct dlm_lock_resource, refs); refs 522 fs/ocfs2/dlm/dlmmaster.c kref_put(&res->refs, dlm_lockres_release); refs 558 fs/ocfs2/dlm/dlmmaster.c kref_init(&res->refs); refs 187 include/linux/bpf_verifier.h struct bpf_reference_state *refs; refs 48 include/linux/file.h extern struct file *fget_many(unsigned int fd, unsigned int refs); refs 43 include/linux/hugetlb.h struct kref refs; refs 403 include/linux/property.h const struct software_node_ref_args *refs; refs 20 include/linux/refcount.h atomic_t refs; refs 23 include/linux/refcount.h #define REFCOUNT_INIT(n) { .refs = ATOMIC_INIT(n), } refs 32 include/linux/refcount.h atomic_set(&r->refs, n); refs 43 include/linux/refcount.h return atomic_read(&r->refs); refs 76 include/linux/refcount.h return atomic_add_unless(&r->refs, i, 0); refs 81 include/linux/refcount.h atomic_add(i, &r->refs); refs 86 include/linux/refcount.h return atomic_add_unless(&r->refs, 1, 0); refs 91 include/linux/refcount.h atomic_inc(&r->refs); refs 96 include/linux/refcount.h return atomic_sub_and_test(i, &r->refs); refs 101 include/linux/refcount.h return atomic_dec_and_test(&r->refs); refs 106 include/linux/refcount.h atomic_dec(&r->refs); refs 292 include/linux/rmap.h #define try_to_unmap(page, refs) false refs 433 include/sound/pcm.h refcount_t refs; refs 261 include/trace/events/btrfs.h __field( int, refs ) refs 274 include/trace/events/btrfs.h __entry->refs = refcount_read(&map->refs); refs 290 include/trace/events/btrfs.h __entry->refs, __entry->compress_type) refs 491 include/trace/events/btrfs.h __field( int, refs ) refs 505 include/trace/events/btrfs.h __entry->refs = refcount_read(&ordered->refs); refs 525 include/trace/events/btrfs.h __entry->compress_type, __entry->refs) refs 994 include/trace/events/btrfs.h __field( int, refs ) refs 1003 include/trace/events/btrfs.h __entry->refs = atomic_read(&buf->refs); refs 1012 include/trace/events/btrfs.h __entry->refs, refs 474 include/uapi/linux/btrfs_tree.h __le64 refs; refs 480 include/uapi/linux/btrfs_tree.h __le32 refs; refs 617 include/uapi/linux/btrfs_tree.h __le32 refs; refs 69 include/uapi/linux/xdp_diag.h __u32 refs; refs 64 include/uapi/xen/gntdev.h struct ioctl_gntdev_grant_ref refs[1]; refs 250 include/uapi/xen/gntdev.h __u32 refs[1]; refs 292 include/uapi/xen/gntdev.h __u32 refs[1]; refs 31 kernel/audit_tree.c atomic_long_t refs; refs 141 kernel/audit_tree.c if (atomic_long_dec_and_test(&chunk->refs)) refs 202 kernel/audit_tree.c atomic_long_set(&chunk->refs, 1); refs 256 kernel/audit_tree.c atomic_long_inc(&p->refs); refs 528 kernel/bpf/verifier.c if (state->acquired_refs && state->refs[0].id) { refs 529 kernel/bpf/verifier.c verbose(env, " refs=%d", state->refs[0].id); refs 531 kernel/bpf/verifier.c if (state->refs[i].id) refs 532 kernel/bpf/verifier.c verbose(env, ",%d", state->refs[i].id); refs 553 kernel/bpf/verifier.c COPY_STATE_FN(reference, acquired_refs, refs, 1) refs 593 kernel/bpf/verifier.c REALLOC_STATE_FN(reference, acquired_refs, refs, 1) refs 629 kernel/bpf/verifier.c state->refs[new_ofs].id = id; refs 630 kernel/bpf/verifier.c state->refs[new_ofs].insn_idx = insn_idx; refs 642 kernel/bpf/verifier.c if (state->refs[i].id == ptr_id) { refs 644 kernel/bpf/verifier.c memcpy(&state->refs[i], &state->refs[last_idx], refs 645 kernel/bpf/verifier.c sizeof(*state->refs)); refs 646 kernel/bpf/verifier.c memset(&state->refs[last_idx], 0, sizeof(*state->refs)); refs 670 kernel/bpf/verifier.c kfree(state->refs); refs 3967 kernel/bpf/verifier.c state->refs[i].id, state->refs[i].insn_idx); refs 7080 kernel/bpf/verifier.c return !memcmp(old->refs, cur->refs, refs 7081 kernel/bpf/verifier.c sizeof(*old->refs) * old->acquired_refs); refs 66 lib/refcount.c unsigned int new, val = atomic_read(&r->refs); refs 79 lib/refcount.c } while (!atomic_try_cmpxchg_relaxed(&r->refs, &val, new)); refs 123 lib/refcount.c unsigned int new, val = atomic_read(&r->refs); refs 134 lib/refcount.c } while (!atomic_try_cmpxchg_relaxed(&r->refs, &val, new)); refs 182 lib/refcount.c unsigned int new, val = atomic_read(&r->refs); refs 194 lib/refcount.c } while (!atomic_try_cmpxchg_release(&r->refs, &val, new)); refs 260 lib/refcount.c return atomic_try_cmpxchg_release(&r->refs, &val, 0); refs 277 lib/refcount.c unsigned int new, val = atomic_read(&r->refs); refs 292 lib/refcount.c } while (!atomic_try_cmpxchg_release(&r->refs, &val, new)); refs 1815 mm/gup.c static inline struct page *try_get_compound_head(struct page *page, int refs) refs 1820 mm/gup.c if (unlikely(!page_cache_add_speculative(head, refs))) refs 1997 mm/gup.c int refs; refs 2011 mm/gup.c refs = 0; refs 2020 mm/gup.c refs++; refs 2023 mm/gup.c head = try_get_compound_head(head, refs); refs 2025 mm/gup.c *nr -= refs; refs 2031 mm/gup.c *nr -= refs; refs 2032 mm/gup.c while (refs--) refs 2072 mm/gup.c int refs; refs 2083 mm/gup.c refs = 0; refs 2089 mm/gup.c refs++; refs 2092 mm/gup.c head = try_get_compound_head(pmd_page(orig), refs); refs 2094 mm/gup.c *nr -= refs; refs 2099 mm/gup.c *nr -= refs; refs 2100 mm/gup.c while (refs--) refs 2113 mm/gup.c int refs; refs 2124 mm/gup.c refs = 0; refs 2130 mm/gup.c refs++; refs 2133 mm/gup.c head = try_get_compound_head(pud_page(orig), refs); refs 2135 mm/gup.c *nr -= refs; refs 2140 mm/gup.c *nr -= refs; refs 2141 mm/gup.c while (refs--) refs 2154 mm/gup.c int refs; refs 2161 mm/gup.c refs = 0; refs 2167 mm/gup.c refs++; refs 2170 mm/gup.c head = try_get_compound_head(pgd_page(orig), refs); refs 2172 mm/gup.c *nr -= refs; refs 2177 mm/gup.c *nr -= refs; refs 2178 mm/gup.c while (refs--) refs 710 mm/hugetlb.c kref_init(&resv_map->refs); refs 725 mm/hugetlb.c struct resv_map *resv_map = container_of(ref, struct resv_map, refs); refs 3292 mm/hugetlb.c kref_get(&resv->refs); refs 3311 mm/hugetlb.c kref_put(&resv->refs, resv_map_release); refs 4760 mm/hugetlb.c kref_put(&resv_map->refs, resv_map_release); refs 42 net/netfilter/nf_conntrack_ecache.c struct nf_conn *refs[16]; refs 67 net/netfilter/nf_conntrack_ecache.c refs[evicted] = ct; refs 69 net/netfilter/nf_conntrack_ecache.c if (++evicted >= ARRAY_SIZE(refs)) { refs 79 net/netfilter/nf_conntrack_ecache.c nf_ct_put(refs[--evicted]); refs 66 net/xdp/xsk_diag.c du.refs = refcount_read(&umem->users); refs 80 sound/core/pcm_native.c refcount_set(&group->refs, 1); refs 1113 sound/core/pcm_native.c do_free = refcount_dec_and_test(&group->refs); refs 1135 sound/core/pcm_native.c refcount_inc(&group->refs); refs 2043 sound/core/pcm_native.c refcount_inc(&target_group->refs); refs 2080 sound/core/pcm_native.c refcount_dec(&group->refs); refs 2087 sound/core/pcm_native.c do_free = refcount_dec_and_test(&group->refs); refs 53 tools/include/linux/refcount.h atomic_t refs; refs 56 tools/include/linux/refcount.h #define REFCOUNT_INIT(n) { .refs = ATOMIC_INIT(n), } refs 60 tools/include/linux/refcount.h atomic_set(&r->refs, n); refs 65 tools/include/linux/refcount.h return atomic_read(&r->refs); refs 78 tools/include/linux/refcount.h unsigned int old, new, val = atomic_read(&r->refs); refs 89 tools/include/linux/refcount.h old = atomic_cmpxchg_relaxed(&r->refs, val, new); refs 123 tools/include/linux/refcount.h unsigned int old, new, val = atomic_read(&r->refs); refs 135 tools/include/linux/refcount.h old = atomic_cmpxchg_release(&r->refs, val, new); refs 278 tools/perf/arch/x86/util/intel-bts.c struct intel_bts_snapshot_ref *refs; refs 286 tools/perf/arch/x86/util/intel-bts.c refs = calloc(new_cnt, sz); refs 287 tools/perf/arch/x86/util/intel-bts.c if (!refs) refs 290 tools/perf/arch/x86/util/intel-bts.c memcpy(refs, btsr->snapshot_refs, cnt * sz); refs 292 tools/perf/arch/x86/util/intel-bts.c btsr->snapshot_refs = refs; refs 839 tools/perf/arch/x86/util/intel-pt.c struct intel_pt_snapshot_ref *refs; refs 847 tools/perf/arch/x86/util/intel-pt.c refs = calloc(new_cnt, sz); refs 848 tools/perf/arch/x86/util/intel-pt.c if (!refs) refs 851 tools/perf/arch/x86/util/intel-pt.c memcpy(refs, ptr->snapshot_refs, cnt * sz); refs 853 tools/perf/arch/x86/util/intel-pt.c ptr->snapshot_refs = refs; refs 31 tools/vm/slabinfo.c int refs; refs 1108 tools/vm/slabinfo.c s->refs++; refs 1126 tools/vm/slabinfo.c if (!show_single_ref && a->slab->refs == 1) refs 1156 tools/vm/slabinfo.c if (s->refs > 1 && !show_first_alias) refs 1213 tools/vm/slabinfo.c slab->refs = 0;