atomic 144 arch/arc/include/asm/arcregs.h unsigned int div_rem:4, pad2:4, ldd:1, unalign:1, atomic:1, be:1, atomic 147 arch/arc/include/asm/arcregs.h unsigned int ver:8, pad1:12, be:1, atomic:1, unalign:1, atomic 252 arch/arc/kernel/setup.c cpu->isa.atomic = IS_ENABLED(CONFIG_ARC_HAS_LLSC); atomic 256 arch/arc/kernel/setup.c cpu->isa.atomic = bcr.info & 1; atomic 309 arch/arc/kernel/setup.c IS_AVAIL2(cpu->isa.atomic, "atomic ", CONFIG_ARC_HAS_LLSC), atomic 88 arch/arm/lib/uaccess_with_memcpy.c int atomic; atomic 96 arch/arm/lib/uaccess_with_memcpy.c atomic = faulthandler_disabled(); atomic 98 arch/arm/lib/uaccess_with_memcpy.c if (!atomic) atomic 106 arch/arm/lib/uaccess_with_memcpy.c if (!atomic) atomic 110 arch/arm/lib/uaccess_with_memcpy.c if (!atomic) atomic 130 arch/arm/lib/uaccess_with_memcpy.c if (!atomic) atomic 210 arch/csky/include/asm/atomic.h #include <asm-generic/atomic.h> atomic 362 arch/powerpc/platforms/powernv/opal.c static int __opal_put_chars(uint32_t vtermno, const char *data, int total_len, bool atomic) atomic 372 arch/powerpc/platforms/powernv/opal.c if (atomic) atomic 402 arch/powerpc/platforms/powernv/opal.c if (atomic) { atomic 412 arch/powerpc/platforms/powernv/opal.c if (atomic) atomic 58 arch/riscv/include/asm/atomic.h void atomic##prefix##_##op(c_type i, atomic##prefix##_t *v) \ atomic 92 arch/riscv/include/asm/atomic.h c_type atomic##prefix##_fetch_##op##_relaxed(c_type i, \ atomic 93 arch/riscv/include/asm/atomic.h atomic##prefix##_t *v) \ atomic 104 arch/riscv/include/asm/atomic.h c_type atomic##prefix##_fetch_##op(c_type i, atomic##prefix##_t *v) \ atomic 117 arch/riscv/include/asm/atomic.h c_type atomic##prefix##_##op##_return_relaxed(c_type i, \ atomic 118 arch/riscv/include/asm/atomic.h atomic##prefix##_t *v) \ atomic 120 arch/riscv/include/asm/atomic.h return atomic##prefix##_fetch_##op##_relaxed(i, v) c_op I; \ atomic 123 arch/riscv/include/asm/atomic.h c_type atomic##prefix##_##op##_return(c_type i, atomic##prefix##_t *v) \ atomic 125 arch/riscv/include/asm/atomic.h return atomic##prefix##_fetch_##op(i, v) c_op I; \ atomic 249 arch/riscv/include/asm/atomic.h c_t atomic##prefix##_xchg_relaxed(atomic##prefix##_t *v, c_t n) \ atomic 254 arch/riscv/include/asm/atomic.h c_t atomic##prefix##_xchg_acquire(atomic##prefix##_t *v, c_t n) \ atomic 259 arch/riscv/include/asm/atomic.h c_t atomic##prefix##_xchg_release(atomic##prefix##_t *v, c_t n) \ atomic 264 arch/riscv/include/asm/atomic.h c_t atomic##prefix##_xchg(atomic##prefix##_t *v, c_t n) \ atomic 269 arch/riscv/include/asm/atomic.h c_t atomic##prefix##_cmpxchg_relaxed(atomic##prefix##_t *v, \ atomic 275 arch/riscv/include/asm/atomic.h c_t atomic##prefix##_cmpxchg_acquire(atomic##prefix##_t *v, \ atomic 281 arch/riscv/include/asm/atomic.h c_t atomic##prefix##_cmpxchg_release(atomic##prefix##_t *v, \ atomic 287 arch/riscv/include/asm/atomic.h c_t atomic##prefix##_cmpxchg(atomic##prefix##_t *v, c_t o, c_t n) \ atomic 22 arch/um/include/shared/kern_util.h extern unsigned long alloc_stack(int order, int atomic); atomic 66 arch/um/kernel/process.c unsigned long alloc_stack(int order, int atomic) atomic 71 arch/um/kernel/process.c if (atomic) atomic 61 arch/x86/crypto/aegis128-aesni-glue.c struct aead_request *req, bool atomic); atomic 40 crypto/aegis128-core.c struct aead_request *req, bool atomic); atomic 477 crypto/skcipher.c struct skcipher_request *req, bool atomic) atomic 487 crypto/skcipher.c walk->flags &= atomic ? ~SKCIPHER_WALK_SLEEP : ~0; atomic 511 crypto/skcipher.c struct aead_request *req, bool atomic) atomic 546 crypto/skcipher.c if (atomic) atomic 553 crypto/skcipher.c bool atomic) atomic 557 crypto/skcipher.c return skcipher_walk_aead_common(walk, req, atomic); atomic 562 crypto/skcipher.c struct aead_request *req, bool atomic) atomic 566 crypto/skcipher.c return skcipher_walk_aead_common(walk, req, atomic); atomic 571 crypto/skcipher.c struct aead_request *req, bool atomic) atomic 577 crypto/skcipher.c return skcipher_walk_aead_common(walk, req, atomic); atomic 205 drivers/dma/mediatek/mtk-cqdma.c static int mtk_cqdma_poll_engine_done(struct mtk_cqdma_pchan *pc, bool atomic) atomic 209 drivers/dma/mediatek/mtk-cqdma.c if (!atomic) atomic 1831 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c int x, int y, int atomic) atomic 1849 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c if (!atomic && !crtc->primary->fb) { atomic 1854 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c if (atomic) atomic 1868 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c if (!atomic) { atomic 2061 drivers/gpu/drm/amd/amdgpu/dce_v10_0.c if (!atomic && fb && fb != crtc->primary->fb) { atomic 1873 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c int x, int y, int atomic) atomic 1891 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c if (!atomic && !crtc->primary->fb) { atomic 1896 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c if (atomic) atomic 1910 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c if (!atomic) { atomic 2103 drivers/gpu/drm/amd/amdgpu/dce_v11_0.c if (!atomic && fb && fb != crtc->primary->fb) { atomic 1789 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c int x, int y, int atomic) atomic 1806 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c if (!atomic && !crtc->primary->fb) { atomic 1811 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c if (atomic) atomic 1825 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c if (!atomic) { atomic 1996 drivers/gpu/drm/amd/amdgpu/dce_v6_0.c if (!atomic && fb && fb != crtc->primary->fb) { atomic 1760 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c int x, int y, int atomic) atomic 1778 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c if (!atomic && !crtc->primary->fb) { atomic 1783 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c if (atomic) atomic 1797 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c if (!atomic) { atomic 1970 drivers/gpu/drm/amd/amdgpu/dce_v8_0.c if (!atomic && fb && fb != crtc->primary->fb) { atomic 530 drivers/gpu/drm/ast/ast_mode.c int x, int y, int atomic) atomic 536 drivers/gpu/drm/ast/ast_mode.c if (!atomic && fb) { atomic 2595 drivers/gpu/drm/drm_atomic_helper.c bool atomic) atomic 2602 drivers/gpu/drm/drm_atomic_helper.c if (atomic && crtc_funcs && crtc_funcs->atomic_begin) atomic 2617 drivers/gpu/drm/drm_atomic_helper.c if (atomic && crtc_funcs && crtc_funcs->atomic_flush) atomic 1299 drivers/gpu/drm/drm_atomic_uapi.c if (!file_priv->atomic) atomic 2221 drivers/gpu/drm/drm_connector.c ret = drm_mode_object_get_properties(&connector->base, file_priv->atomic, atomic 151 drivers/gpu/drm/drm_crtc_internal.h int drm_mode_object_get_properties(struct drm_mode_object *obj, bool atomic, atomic 346 drivers/gpu/drm/drm_ioctl.c file_priv->atomic = req->value; atomic 359 drivers/gpu/drm/drm_ioctl.c if (!file_priv->atomic) atomic 338 drivers/gpu/drm/drm_mode_object.c int drm_mode_object_get_properties(struct drm_mode_object *obj, bool atomic, atomic 349 drivers/gpu/drm/drm_mode_object.c if ((prop->flags & DRM_MODE_PROP_ATOMIC) && !atomic) atomic 408 drivers/gpu/drm/drm_mode_object.c ret = drm_mode_object_get_properties(obj, file_priv->atomic, atomic 861 drivers/gpu/drm/mgag200/mgag200_mode.c int x, int y, int atomic) atomic 867 drivers/gpu/drm/mgag200/mgag200_mode.c if (!atomic && fb) { atomic 819 drivers/gpu/drm/nouveau/dispnv04/crtc.c int x, int y, bool atomic) atomic 832 drivers/gpu/drm/nouveau/dispnv04/crtc.c if (!atomic && !crtc->primary->fb) { atomic 840 drivers/gpu/drm/nouveau/dispnv04/crtc.c if (atomic) { atomic 87 drivers/gpu/drm/nouveau/nouveau_drm.c MODULE_PARM_DESC(atomic, "Expose atomic ioctl (default: disabled)"); atomic 89 drivers/gpu/drm/nouveau/nouveau_drm.c module_param_named(atomic, nouveau_atomic, int, 0400); atomic 1146 drivers/gpu/drm/radeon/atombios_crtc.c int x, int y, int atomic) atomic 1164 drivers/gpu/drm/radeon/atombios_crtc.c if (!atomic && !crtc->primary->fb) { atomic 1169 drivers/gpu/drm/radeon/atombios_crtc.c if (atomic) atomic 1183 drivers/gpu/drm/radeon/atombios_crtc.c if (atomic) atomic 1451 drivers/gpu/drm/radeon/atombios_crtc.c if (!atomic && fb && fb != crtc->primary->fb) { atomic 1468 drivers/gpu/drm/radeon/atombios_crtc.c int x, int y, int atomic) atomic 1485 drivers/gpu/drm/radeon/atombios_crtc.c if (!atomic && !crtc->primary->fb) { atomic 1490 drivers/gpu/drm/radeon/atombios_crtc.c if (atomic) atomic 1504 drivers/gpu/drm/radeon/atombios_crtc.c if (atomic) atomic 1660 drivers/gpu/drm/radeon/atombios_crtc.c if (!atomic && fb && fb != crtc->primary->fb) { atomic 375 drivers/gpu/drm/radeon/radeon_legacy_crtc.c int x, int y, int atomic) atomic 393 drivers/gpu/drm/radeon/radeon_legacy_crtc.c if (!atomic && !crtc->primary->fb) { atomic 398 drivers/gpu/drm/radeon/radeon_legacy_crtc.c if (atomic) atomic 448 drivers/gpu/drm/radeon/radeon_legacy_crtc.c if (!atomic && fb && fb != crtc->primary->fb) { atomic 558 drivers/gpu/drm/radeon/radeon_legacy_crtc.c if (!atomic && fb && fb != crtc->primary->fb) { atomic 867 drivers/gpu/drm/radeon/radeon_mode.h int x, int y, int atomic); atomic 1028 drivers/i2c/busses/i2c-img-scb.c bool atomic = false; atomic 1057 drivers/i2c/busses/i2c-img-scb.c atomic = true; atomic 1096 drivers/i2c/busses/i2c-img-scb.c if (atomic) { atomic 200 drivers/i2c/busses/i2c-tegra-bpmp.c bool atomic) atomic 215 drivers/i2c/busses/i2c-tegra-bpmp.c if (atomic) atomic 225 drivers/i2c/busses/i2c-tegra-bpmp.c bool atomic) atomic 247 drivers/i2c/busses/i2c-tegra-bpmp.c err = tegra_bpmp_i2c_msg_xfer(i2c, &request, &response, atomic); atomic 442 drivers/iio/adc/sun4i-gpadc-iio.c unsigned int *irq, atomic_t *atomic) atomic 460 drivers/iio/adc/sun4i-gpadc-iio.c atomic_set(atomic, 1); atomic 482 drivers/iio/adc/sun4i-gpadc-iio.c atomic_set(atomic, 0); atomic 2099 drivers/infiniband/core/uverbs_cmd.c struct ib_atomic_wr *atomic; atomic 2101 drivers/infiniband/core/uverbs_cmd.c next_size = sizeof(*atomic); atomic 2102 drivers/infiniband/core/uverbs_cmd.c atomic = alloc_wr(next_size, user_wr->num_sge); atomic 2103 drivers/infiniband/core/uverbs_cmd.c if (!atomic) { atomic 2108 drivers/infiniband/core/uverbs_cmd.c atomic->remote_addr = user_wr->wr.atomic.remote_addr; atomic 2109 drivers/infiniband/core/uverbs_cmd.c atomic->compare_add = user_wr->wr.atomic.compare_add; atomic 2110 drivers/infiniband/core/uverbs_cmd.c atomic->swap = user_wr->wr.atomic.swap; atomic 2111 drivers/infiniband/core/uverbs_cmd.c atomic->rkey = user_wr->wr.atomic.rkey; atomic 2113 drivers/infiniband/core/uverbs_cmd.c next = &atomic->wr; atomic 2132 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->atomic.cmp_data = atomic_wr(wr)->compare_add; atomic 2133 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->atomic.swap_data = atomic_wr(wr)->swap; atomic 2137 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->atomic.cmp_data = atomic_wr(wr)->compare_add; atomic 2142 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->atomic.remote_va = atomic_wr(wr)->remote_addr; atomic 2143 drivers/infiniband/hw/bnxt_re/ib_verbs.c wqe->atomic.r_key = atomic_wr(wr)->rkey; atomic 1678 drivers/infiniband/hw/bnxt_re/qplib_fp.c sqe->remote_key = cpu_to_le32(wqe->atomic.r_key); atomic 1679 drivers/infiniband/hw/bnxt_re/qplib_fp.c sqe->remote_va = cpu_to_le64(wqe->atomic.remote_va); atomic 1680 drivers/infiniband/hw/bnxt_re/qplib_fp.c sqe->swap_data = cpu_to_le64(wqe->atomic.swap_data); atomic 1681 drivers/infiniband/hw/bnxt_re/qplib_fp.c sqe->cmp_data = cpu_to_le64(wqe->atomic.cmp_data); atomic 179 drivers/infiniband/hw/bnxt_re/qplib_fp.h } atomic; atomic 1489 drivers/infiniband/hw/mlx5/mlx5_ib.h MLX5_CAP_GEN(dev->mdev, atomic) && atomic 364 drivers/infiniband/hw/mlx5/odp.c if (MLX5_CAP_ODP(dev->mdev, rc_odp_caps.atomic)) atomic 382 drivers/infiniband/hw/mlx5/odp.c if (MLX5_CAP_ODP(dev->mdev, xrc_odp_caps.atomic)) atomic 1916 drivers/infiniband/hw/mlx5/qp.c u8 atomic = MLX5_CAP_GEN(dev->mdev, atomic); atomic 1920 drivers/infiniband/hw/mlx5/qp.c if (!atomic) atomic 4164 drivers/infiniband/hw/mlx5/qp.c static __be64 frwr_mkey_mask(bool atomic) atomic 4180 drivers/infiniband/hw/mlx5/qp.c if (atomic) atomic 4208 drivers/infiniband/hw/mlx5/qp.c struct mlx5_ib_mr *mr, u8 flags, bool atomic) atomic 4216 drivers/infiniband/hw/mlx5/qp.c umr->mkey_mask = frwr_mkey_mask(atomic); atomic 4256 drivers/infiniband/hw/mlx5/qp.c static __be64 get_umr_update_access_mask(int atomic) atomic 4265 drivers/infiniband/hw/mlx5/qp.c if (atomic) atomic 4292 drivers/infiniband/hw/mlx5/qp.c const struct ib_send_wr *wr, int atomic) atomic 4318 drivers/infiniband/hw/mlx5/qp.c umr->mkey_mask |= get_umr_update_access_mask(atomic); atomic 4818 drivers/infiniband/hw/mlx5/qp.c bool atomic = wr->access & IB_ACCESS_REMOTE_ATOMIC; atomic 4821 drivers/infiniband/hw/mlx5/qp.c if (!mlx5_ib_can_use_umr(dev, atomic)) { atomic 4842 drivers/infiniband/hw/mlx5/qp.c set_reg_umr_seg(*seg, mr, flags, atomic); atomic 5271 drivers/infiniband/hw/mlx5/qp.c err = set_reg_umr_segment(dev, seg, wr, !!(MLX5_CAP_GEN(mdev, atomic))); atomic 759 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->wr.atomic.remote_addr = atomic 761 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->wr.atomic.rkey = atomic_wr(wr)->rkey; atomic 762 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->wr.atomic.compare_add = atomic 765 drivers/infiniband/hw/vmw_pvrdma/pvrdma_qp.c wqe_hdr->wr.atomic.swap = atomic 156 drivers/infiniband/sw/rxe/rxe_qp.c kfree_skb(res->atomic.skb); atomic 93 drivers/infiniband/sw/rxe/rxe_req.c wqe->wr.wr.atomic.remote_addr : atomic 454 drivers/infiniband/sw/rxe/rxe_req.c atmeth_set_swap_add(pkt, ibwr->wr.atomic.swap); atomic 455 drivers/infiniband/sw/rxe/rxe_req.c atmeth_set_comp(pkt, ibwr->wr.atomic.compare_add); atomic 457 drivers/infiniband/sw/rxe/rxe_req.c atmeth_set_swap_add(pkt, ibwr->wr.atomic.compare_add); atomic 459 drivers/infiniband/sw/rxe/rxe_req.c atmeth_set_rkey(pkt, ibwr->wr.atomic.rkey); atomic 1008 drivers/infiniband/sw/rxe/rxe_resp.c res->atomic.skb = skb; atomic 1141 drivers/infiniband/sw/rxe/rxe_resp.c skb_get(res->atomic.skb); atomic 1143 drivers/infiniband/sw/rxe/rxe_resp.c rc = rxe_xmit_packet(qp, pkt, res->atomic.skb); atomic 565 drivers/infiniband/sw/rxe/rxe_verbs.c wr->wr.atomic.remote_addr = atomic 567 drivers/infiniband/sw/rxe/rxe_verbs.c wr->wr.atomic.compare_add = atomic 569 drivers/infiniband/sw/rxe/rxe_verbs.c wr->wr.atomic.swap = atomic_wr(ibwr)->swap; atomic 570 drivers/infiniband/sw/rxe/rxe_verbs.c wr->wr.atomic.rkey = atomic_wr(ibwr)->rkey; atomic 184 drivers/infiniband/sw/rxe/rxe_verbs.h } atomic; atomic 499 drivers/iommu/arm-smmu-v3.c } atomic; atomic 1377 drivers/iommu/arm-smmu-v3.c &cmdq->q.llq.atomic.prod); atomic 4905 drivers/media/dvb-frontends/drx39xyj/drxj.c set_agc_rf(struct drx_demod_instance *demod, struct drxj_cfg_agc *agc_settings, bool atomic) atomic 4919 drivers/media/dvb-frontends/drx39xyj/drxj.c if (atomic) { atomic 5119 drivers/media/dvb-frontends/drx39xyj/drxj.c set_agc_if(struct drx_demod_instance *demod, struct drxj_cfg_agc *agc_settings, bool atomic) atomic 5133 drivers/media/dvb-frontends/drx39xyj/drxj.c if (atomic) { atomic 253 drivers/misc/sgi-gru/grufault.c int write, int atomic, unsigned long *gpa, int *pageshift) atomic 271 drivers/misc/sgi-gru/grufault.c if (atomic) atomic 310 drivers/misc/sgi-gru/grufault.c struct gru_thread_state *gts, int atomic, atomic 332 drivers/misc/sgi-gru/grufault.c ret = gru_vtop(gts, vaddr, write, atomic, &gpa, &pageshift); atomic 338 drivers/misc/sgi-gru/grufault.c atomic ? "atomic" : "non-atomic", gru->gs_gid, gts, tfh, atomic 362 drivers/misc/sgi-gru/grufault.c int pageshift = 0, asid, write, ret, atomic = !cbk, indexway; atomic 413 drivers/misc/sgi-gru/grufault.c ret = gru_vtop(gts, vaddr, write, atomic, &gpa, &pageshift); atomic 421 drivers/misc/sgi-gru/grufault.c if (atomic || !gru_update_cch(gts)) { atomic 428 drivers/misc/sgi-gru/grufault.c gru_preload_tlb(gru, gts, atomic, vaddr, asid, write, tlb_preload_count, tfh, cbe); atomic 439 drivers/misc/sgi-gru/grufault.c atomic ? "atomic" : "non-atomic", gru->gs_gid, gts, tfh, vaddr, asid, atomic 70 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h #define INIT_ULPTX_WR(w, wrlen, atomic, tid) do { \ atomic 72 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h FW_WR_ATOMIC_V(atomic)); \ atomic 174 drivers/net/ethernet/mellanox/mlx5/core/fw.c if (MLX5_CAP_GEN(dev, atomic)) { atomic 426 drivers/net/ethernet/mellanox/mlx5/core/main.c if (MLX5_CAP_GEN(dev, atomic)) { atomic 498 drivers/net/ethernet/mellanox/mlx5/core/main.c ODP_CAP_SET_MAX(dev, xrc_odp_caps.atomic); atomic 504 drivers/net/ethernet/mellanox/mlx5/core/main.c ODP_CAP_SET_MAX(dev, dc_odp_caps.atomic); atomic 378 drivers/net/ethernet/myricom/myri10ge/myri10ge.c struct myri10ge_cmd *data, int atomic) atomic 408 drivers/net/ethernet/myricom/myri10ge/myri10ge.c if (atomic) { atomic 830 drivers/net/ethernet/myricom/myri10ge/myri10ge.c myri10ge_change_promisc(struct myri10ge_priv *mgp, int promisc, int atomic) atomic 836 drivers/net/ethernet/myricom/myri10ge/myri10ge.c status = myri10ge_send_cmd(mgp, ctl, &cmd, atomic); atomic 41 drivers/net/ethernet/sfc/efx.h void efx_fast_push_rx_descriptors(struct efx_rx_queue *rx_queue, bool atomic); atomic 44 drivers/net/ethernet/sfc/falcon/efx.h void ef4_fast_push_rx_descriptors(struct ef4_rx_queue *rx_queue, bool atomic); atomic 149 drivers/net/ethernet/sfc/falcon/rx.c static int ef4_init_rx_buffers(struct ef4_rx_queue *rx_queue, bool atomic) atomic 164 drivers/net/ethernet/sfc/falcon/rx.c (atomic ? GFP_ATOMIC : GFP_KERNEL), atomic 324 drivers/net/ethernet/sfc/falcon/rx.c void ef4_fast_push_rx_descriptors(struct ef4_rx_queue *rx_queue, bool atomic) atomic 357 drivers/net/ethernet/sfc/falcon/rx.c rc = ef4_init_rx_buffers(rx_queue, atomic); atomic 149 drivers/net/ethernet/sfc/rx.c static int efx_init_rx_buffers(struct efx_rx_queue *rx_queue, bool atomic) atomic 164 drivers/net/ethernet/sfc/rx.c (atomic ? GFP_ATOMIC : GFP_KERNEL), atomic 324 drivers/net/ethernet/sfc/rx.c void efx_fast_push_rx_descriptors(struct efx_rx_queue *rx_queue, bool atomic) atomic 357 drivers/net/ethernet/sfc/rx.c rc = efx_init_rx_buffers(rx_queue, atomic); atomic 303 fs/btrfs/disk-io.c int atomic) atomic 312 fs/btrfs/disk-io.c if (atomic) atomic 4108 fs/btrfs/disk-io.c int atomic) atomic 4118 fs/btrfs/disk-io.c parent_transid, atomic); atomic 113 fs/btrfs/disk-io.h int atomic); atomic 19 fs/erofs/xattr.c static inline void xattr_iter_end(struct xattr_iter *it, bool atomic) atomic 22 fs/erofs/xattr.c if (!atomic) atomic 38 fs/erofs/zpvec.h bool atomic) atomic 43 fs/erofs/zpvec.h if (atomic) atomic 72 fs/erofs/zpvec.h bool atomic) atomic 76 fs/erofs/zpvec.h z_erofs_pagevec_ctor_exit(ctor, atomic); atomic 80 fs/erofs/zpvec.h ctor->pages = atomic ? atomic 3064 fs/f2fs/f2fs.h struct writeback_control *wbc, bool atomic, atomic 208 fs/f2fs/file.c int datasync, bool atomic) atomic 287 fs/f2fs/file.c ret = f2fs_fsync_node_pages(sbi, inode, &wbc, atomic, &seq_id); atomic 312 fs/f2fs/file.c if (!atomic) { atomic 322 fs/f2fs/file.c if (!atomic && F2FS_OPTION(sbi).fsync_mode != FSYNC_MODE_NOBARRIER) atomic 1498 fs/f2fs/node.c static int __write_node_page(struct page *page, bool atomic, bool *submitted, atomic 1562 fs/f2fs/node.c if (atomic && !test_opt(sbi, NOBARRIER)) atomic 1650 fs/f2fs/node.c struct writeback_control *wbc, bool atomic, atomic 1662 fs/f2fs/node.c if (atomic) { atomic 1711 fs/f2fs/node.c if (!atomic || page == last_page) { atomic 1728 fs/f2fs/node.c ret = __write_node_page(page, atomic && atomic 1752 fs/f2fs/node.c if (!ret && atomic && !marked) { atomic 1157 fs/nfs/nfs4proc.c if (cinfo->atomic && cinfo->before == inode_peek_iversion_raw(dir)) { atomic 4231 fs/nfs/nfs4xdr.c cinfo->atomic = be32_to_cpup(p++); atomic 2041 fs/nfsd/nfs4xdr.c *p++ = cpu_to_be32(c->atomic); atomic 78 fs/nfsd/xdr4.h u32 atomic; atomic 718 fs/nfsd/xdr4.h cinfo->atomic = (u32)fhp->fh_post_saved; atomic 305 fs/ubifs/orphan.c static int do_write_orph_node(struct ubifs_info *c, int len, int atomic) atomic 309 fs/ubifs/orphan.c if (atomic) { atomic 336 fs/ubifs/orphan.c static int write_orph_node(struct ubifs_info *c, int atomic) atomic 385 fs/ubifs/orphan.c err = do_write_orph_node(c, len, atomic); atomic 399 fs/ubifs/orphan.c static int write_orph_nodes(struct ubifs_info *c, int atomic) atomic 404 fs/ubifs/orphan.c err = write_orph_node(c, atomic); atomic 408 fs/ubifs/orphan.c if (atomic) { atomic 479 fs/ubifs/orphan.c int avail, atomic = 0, err; atomic 488 fs/ubifs/orphan.c atomic = 1; atomic 490 fs/ubifs/orphan.c err = write_orph_nodes(c, atomic); atomic 139 include/crypto/internal/skcipher.h bool atomic); atomic 144 include/crypto/internal/skcipher.h bool atomic); atomic 146 include/crypto/internal/skcipher.h struct aead_request *req, bool atomic); atomic 148 include/crypto/internal/skcipher.h struct aead_request *req, bool atomic); atomic 96 include/drm/drm_atomic_helper.h bool atomic); atomic 186 include/drm/drm_file.h bool atomic; atomic 722 include/linux/kvm_host.h bool atomic, bool *async, bool write_fault, atomic 763 include/linux/kvm_host.h struct gfn_to_pfn_cache *cache, bool atomic); atomic 767 include/linux/kvm_host.h struct gfn_to_pfn_cache *cache, bool dirty, bool atomic); atomic 437 include/linux/mlx5/mlx5_ifc.h u8 atomic[0x1]; atomic 1291 include/linux/mlx5/mlx5_ifc.h u8 atomic[0x1]; atomic 179 include/linux/nfs_xdr.h u32 atomic; atomic 454 include/linux/spinlock.h extern int _atomic_dec_and_lock(atomic_t *atomic, spinlock_t *lock); atomic 455 include/linux/spinlock.h #define atomic_dec_and_lock(atomic, lock) \ atomic 456 include/linux/spinlock.h __cond_lock(lock, _atomic_dec_and_lock(atomic, lock)) atomic 458 include/linux/spinlock.h extern int _atomic_dec_and_lock_irqsave(atomic_t *atomic, spinlock_t *lock, atomic 460 include/linux/spinlock.h #define atomic_dec_and_lock_irqsave(atomic, lock, flags) \ atomic 461 include/linux/spinlock.h __cond_lock(lock, _atomic_dec_and_lock_irqsave(atomic, lock, &(flags))) atomic 554 include/sound/gus.h int atomic, atomic 51 include/sound/seq_kernel.h int (*event_input)(struct snd_seq_event *ev, int direct, void *private_data, int atomic, int hop); atomic 63 include/sound/seq_kernel.h int snd_seq_kernel_client_dispatch(int client, struct snd_seq_event *ev, int atomic, int hop); atomic 814 include/uapi/rdma/ib_user_verbs.h } atomic; atomic 90 include/uapi/rdma/rdma_user_rxe.h } atomic; atomic 248 include/uapi/rdma/vmw_pvrdma-abi.h } atomic; atomic 22 lib/atomic64_test.c atomic##bit##_set(&v, v0); \ atomic 24 lib/atomic64_test.c atomic##bit##_##op(val, &v); \ atomic 26 lib/atomic64_test.c WARN(atomic##bit##_read(&v) != r, "%Lx != %Lx\n", \ atomic 27 lib/atomic64_test.c (unsigned long long)atomic##bit##_read(&v), \ atomic 46 lib/atomic64_test.c atomic##bit##_set(&v, v0); \ atomic 49 lib/atomic64_test.c BUG_ON(atomic##bit##_##op(val, &v) != r); \ atomic 50 lib/atomic64_test.c BUG_ON(atomic##bit##_read(&v) != r); \ atomic 55 lib/atomic64_test.c atomic##bit##_set(&v, v0); \ atomic 58 lib/atomic64_test.c BUG_ON(atomic##bit##_##op(val, &v) != v0); \ atomic 59 lib/atomic64_test.c BUG_ON(atomic##bit##_read(&v) != r); \ atomic 74 lib/atomic64_test.c atomic##bit##_set(&v, init); \ atomic 75 lib/atomic64_test.c BUG_ON(atomic##bit##_##op(&v, ##args) != ret); \ atomic 76 lib/atomic64_test.c BUG_ON(atomic##bit##_read(&v) != expect); \ atomic 21 lib/dec_and_lock.c int _atomic_dec_and_lock(atomic_t *atomic, spinlock_t *lock) atomic 24 lib/dec_and_lock.c if (atomic_add_unless(atomic, -1, 1)) atomic 29 lib/dec_and_lock.c if (atomic_dec_and_test(atomic)) atomic 37 lib/dec_and_lock.c int _atomic_dec_and_lock_irqsave(atomic_t *atomic, spinlock_t *lock, atomic 41 lib/dec_and_lock.c if (atomic_add_unless(atomic, -1, 1)) atomic 46 lib/dec_and_lock.c if (atomic_dec_and_test(atomic)) atomic 133 net/rds/ib_send.c rds_ib_send_complete(container_of(op, struct rds_message, atomic), atomic 173 net/rds/ib_send.c rm = container_of(send->s_op, struct rds_message, atomic); atomic 794 net/rds/ib_send.c rds_message_addref(container_of(send->s_op, struct rds_message, atomic)); atomic 167 net/rds/message.c if (rm->atomic.op_active) atomic 168 net/rds/message.c rds_atomic_free_op(&rm->atomic); atomic 169 net/rds/message.c if (rm->atomic.op_rdma_mr) atomic 170 net/rds/message.c rds_mr_put(rm->atomic.op_rdma_mr); atomic 794 net/rds/rdma.c || rm->atomic.op_active) atomic 802 net/rds/rdma.c rm->atomic.op_type = RDS_ATOMIC_TYPE_FADD; atomic 803 net/rds/rdma.c rm->atomic.op_m_fadd.add = args->fadd.add; atomic 804 net/rds/rdma.c rm->atomic.op_m_fadd.nocarry_mask = 0; atomic 807 net/rds/rdma.c rm->atomic.op_type = RDS_ATOMIC_TYPE_FADD; atomic 808 net/rds/rdma.c rm->atomic.op_m_fadd.add = args->m_fadd.add; atomic 809 net/rds/rdma.c rm->atomic.op_m_fadd.nocarry_mask = args->m_fadd.nocarry_mask; atomic 812 net/rds/rdma.c rm->atomic.op_type = RDS_ATOMIC_TYPE_CSWP; atomic 813 net/rds/rdma.c rm->atomic.op_m_cswp.compare = args->cswp.compare; atomic 814 net/rds/rdma.c rm->atomic.op_m_cswp.swap = args->cswp.swap; atomic 815 net/rds/rdma.c rm->atomic.op_m_cswp.compare_mask = ~0; atomic 816 net/rds/rdma.c rm->atomic.op_m_cswp.swap_mask = ~0; atomic 819 net/rds/rdma.c rm->atomic.op_type = RDS_ATOMIC_TYPE_CSWP; atomic 820 net/rds/rdma.c rm->atomic.op_m_cswp.compare = args->m_cswp.compare; atomic 821 net/rds/rdma.c rm->atomic.op_m_cswp.swap = args->m_cswp.swap; atomic 822 net/rds/rdma.c rm->atomic.op_m_cswp.compare_mask = args->m_cswp.compare_mask; atomic 823 net/rds/rdma.c rm->atomic.op_m_cswp.swap_mask = args->m_cswp.swap_mask; atomic 829 net/rds/rdma.c rm->atomic.op_notify = !!(args->flags & RDS_RDMA_NOTIFY_ME); atomic 830 net/rds/rdma.c rm->atomic.op_silent = !!(args->flags & RDS_RDMA_SILENT); atomic 831 net/rds/rdma.c rm->atomic.op_active = 1; atomic 832 net/rds/rdma.c rm->atomic.op_recverr = rs->rs_recverr; atomic 833 net/rds/rdma.c rm->atomic.op_sg = rds_message_alloc_sgs(rm, 1); atomic 834 net/rds/rdma.c if (IS_ERR(rm->atomic.op_sg)) { atomic 835 net/rds/rdma.c ret = PTR_ERR(rm->atomic.op_sg); atomic 850 net/rds/rdma.c sg_set_page(rm->atomic.op_sg, page, 8, offset_in_page(args->local_addr)); atomic 852 net/rds/rdma.c if (rm->atomic.op_notify || rm->atomic.op_recverr) { atomic 858 net/rds/rdma.c rm->atomic.op_notifier = kmalloc(sizeof(*rm->atomic.op_notifier), GFP_KERNEL); atomic 859 net/rds/rdma.c if (!rm->atomic.op_notifier) { atomic 864 net/rds/rdma.c rm->atomic.op_notifier->n_user_token = args->user_token; atomic 865 net/rds/rdma.c rm->atomic.op_notifier->n_status = RDS_RDMA_SUCCESS; atomic 868 net/rds/rdma.c rm->atomic.op_rkey = rds_rdma_cookie_key(args->cookie); atomic 869 net/rds/rdma.c rm->atomic.op_remote_addr = args->remote_addr + rds_rdma_cookie_offset(args->cookie); atomic 875 net/rds/rdma.c rm->atomic.op_active = 0; atomic 876 net/rds/rdma.c kfree(rm->atomic.op_notifier); atomic 463 net/rds/rds.h } atomic; atomic 326 net/rds/send.c if (rm->atomic.op_active && !cp->cp_xmit_atomic_sent) { atomic 327 net/rds/send.c rm->m_final_op = &rm->atomic; atomic 332 net/rds/send.c ret = conn->c_trans->xmit_atomic(conn, &rm->atomic); atomic 353 net/rds/send.c ops_present = (rm->atomic.op_active || rm->rdma.op_active); atomic 354 net/rds/send.c if (rm->atomic.op_active && !rm->atomic.op_silent) atomic 545 net/rds/send.c ao = &rm->atomic; atomic 587 net/rds/send.c ao = &rm->atomic; atomic 1331 net/rds/send.c if (rm->atomic.op_active && !conn->c_trans->xmit_atomic) { atomic 1333 net/rds/send.c &rm->atomic, conn->c_trans->xmit_atomic); atomic 297 net/sched/act_ife.c int len, bool atomic, bool exists) atomic 302 net/sched/act_ife.c mi = kzalloc(sizeof(*mi), atomic ? GFP_ATOMIC : GFP_KERNEL); atomic 309 net/sched/act_ife.c ret = ops->alloc(mi, metaval, atomic ? GFP_ATOMIC : GFP_KERNEL); atomic 136 sound/core/seq/oss/seq_oss_device.h snd_seq_oss_dispatch(struct seq_oss_devinfo *dp, struct snd_seq_event *ev, int atomic, int hop) atomic 138 sound/core/seq/oss/seq_oss_device.h return snd_seq_kernel_client_dispatch(dp->cseq, ev, atomic, hop); atomic 423 sound/core/seq/oss/seq_oss_event.c int atomic, int hop) atomic 96 sound/core/seq/oss/seq_oss_event.h int snd_seq_oss_event_input(struct snd_seq_event *ev, int direct, void *private_data, int atomic, int hop); atomic 40 sound/core/seq/oss/seq_oss_init.c static int receive_announce(struct snd_seq_event *ev, int direct, void *private, int atomic, int hop); atomic 123 sound/core/seq/oss/seq_oss_init.c receive_announce(struct snd_seq_event *ev, int direct, void *private, int atomic, int hop) atomic 127 sound/core/seq/oss/seq_oss_init.c if (atomic) atomic 68 sound/core/seq/seq_clientmgr.c int err, int atomic, int hop); atomic 71 sound/core/seq/seq_clientmgr.c int filter, int atomic, int hop); atomic 535 sound/core/seq/seq_clientmgr.c int err, int atomic, int hop) atomic 557 sound/core/seq/seq_clientmgr.c result = snd_seq_deliver_single_event(NULL, &bounce_ev, 0, atomic, hop + 1); atomic 603 sound/core/seq/seq_clientmgr.c int filter, int atomic, int hop) atomic 640 sound/core/seq/seq_clientmgr.c atomic, hop); atomic 653 sound/core/seq/seq_clientmgr.c result = bounce_error_event(client, event, result, atomic, hop); atomic 664 sound/core/seq/seq_clientmgr.c int atomic, int hop) atomic 680 sound/core/seq/seq_clientmgr.c if (atomic) atomic 694 sound/core/seq/seq_clientmgr.c 0, atomic, hop); atomic 705 sound/core/seq/seq_clientmgr.c if (atomic) atomic 721 sound/core/seq/seq_clientmgr.c int atomic, int hop) atomic 737 sound/core/seq/seq_clientmgr.c atomic, hop); atomic 757 sound/core/seq/seq_clientmgr.c struct snd_seq_event *event, int atomic, int hop) atomic 772 sound/core/seq/seq_clientmgr.c err = port_broadcast_event(client, event, atomic, hop); atomic 777 sound/core/seq/seq_clientmgr.c atomic, hop); atomic 793 sound/core/seq/seq_clientmgr.c int atomic, int hop) atomic 810 sound/core/seq/seq_clientmgr.c int atomic, int hop) atomic 823 sound/core/seq/seq_clientmgr.c snd_BUG_ON(atomic && (event->data.ext.len & SNDRV_SEQ_EXT_USRPTR))) atomic 828 sound/core/seq/seq_clientmgr.c result = deliver_to_subscribers(client, event, atomic, hop); atomic 832 sound/core/seq/seq_clientmgr.c result = broadcast_event(client, event, atomic, hop); atomic 834 sound/core/seq/seq_clientmgr.c result = multicast_event(client, event, atomic, hop); atomic 836 sound/core/seq/seq_clientmgr.c result = port_broadcast_event(client, event, atomic, hop); atomic 839 sound/core/seq/seq_clientmgr.c result = snd_seq_deliver_single_event(client, event, 0, atomic, hop); atomic 854 sound/core/seq/seq_clientmgr.c int snd_seq_dispatch_event(struct snd_seq_event_cell *cell, int atomic, int hop) atomic 878 sound/core/seq/seq_clientmgr.c result = snd_seq_deliver_event(client, &tmpev, atomic, hop); atomic 905 sound/core/seq/seq_clientmgr.c if (snd_seq_enqueue_event(cell, atomic, hop) < 0) atomic 913 sound/core/seq/seq_clientmgr.c result = snd_seq_deliver_event(client, &cell->event, atomic, hop); atomic 929 sound/core/seq/seq_clientmgr.c int atomic, int hop, atomic 958 sound/core/seq/seq_clientmgr.c return snd_seq_deliver_event(client, event, atomic, hop); atomic 968 sound/core/seq/seq_clientmgr.c err = snd_seq_event_dup(client->pool, event, &cell, !blocking || atomic, atomic 974 sound/core/seq/seq_clientmgr.c if ((err = snd_seq_enqueue_event(cell, atomic, hop)) < 0) { atomic 2301 sound/core/seq/seq_clientmgr.c int atomic, int hop) atomic 2323 sound/core/seq/seq_clientmgr.c result = snd_seq_deliver_event(cptr, ev, atomic, hop); atomic 79 sound/core/seq/seq_clientmgr.h int snd_seq_dispatch_event(struct snd_seq_event_cell *cell, int atomic, int hop); atomic 75 sound/core/seq/seq_dummy.c int atomic, int hop) atomic 90 sound/core/seq/seq_dummy.c return snd_seq_kernel_client_dispatch(p->client, &tmpev, atomic, hop); atomic 115 sound/core/seq/seq_midi.c void *private_data, int atomic, int hop) atomic 58 sound/core/seq/seq_ports.h int atomic, int hop); atomic 237 sound/core/seq/seq_queue.c void snd_seq_check_queue(struct snd_seq_queue *q, int atomic, int hop) atomic 264 sound/core/seq/seq_queue.c snd_seq_dispatch_event(cell, atomic, hop); atomic 273 sound/core/seq/seq_queue.c snd_seq_dispatch_event(cell, atomic, hop); atomic 289 sound/core/seq/seq_queue.c int snd_seq_enqueue_event(struct snd_seq_event_cell *cell, int atomic, int hop) atomic 333 sound/core/seq/seq_queue.c snd_seq_check_queue(q, atomic, hop); atomic 641 sound/core/seq/seq_queue.c int atomic, int hop) atomic 656 sound/core/seq/seq_queue.c snd_seq_kernel_client_dispatch(SNDRV_SEQ_CLIENT_SYSTEM, &sev, atomic, hop); atomic 665 sound/core/seq/seq_queue.c int atomic, int hop) atomic 672 sound/core/seq/seq_queue.c queue_broadcast_event(q, ev, atomic, hop); atomic 677 sound/core/seq/seq_queue.c queue_broadcast_event(q, ev, atomic, hop); atomic 682 sound/core/seq/seq_queue.c queue_broadcast_event(q, ev, atomic, hop); atomic 687 sound/core/seq/seq_queue.c queue_broadcast_event(q, ev, atomic, hop); atomic 692 sound/core/seq/seq_queue.c queue_broadcast_event(q, ev, atomic, hop); atomic 698 sound/core/seq/seq_queue.c queue_broadcast_event(q, ev, atomic, hop); atomic 705 sound/core/seq/seq_queue.c queue_broadcast_event(q, ev, atomic, hop); atomic 716 sound/core/seq/seq_queue.c int snd_seq_control_queue(struct snd_seq_event *ev, int atomic, int hop) atomic 732 sound/core/seq/seq_queue.c snd_seq_queue_process_event(q, ev, atomic, hop); atomic 69 sound/core/seq/seq_queue.h int snd_seq_enqueue_event(struct snd_seq_event_cell *cell, int atomic, int hop); atomic 84 sound/core/seq/seq_queue.h void snd_seq_check_queue(struct snd_seq_queue *q, int atomic, int hop); atomic 96 sound/core/seq/seq_queue.h int snd_seq_control_queue(struct snd_seq_event *ev, int atomic, int hop); atomic 101 sound/core/seq/seq_system.c static int event_input_timer(struct snd_seq_event * ev, int direct, void *private_data, int atomic, int hop) atomic 103 sound/core/seq/seq_system.c return snd_seq_control_queue(ev, atomic, hop); atomic 67 sound/core/seq/seq_virmidi.c bool atomic) atomic 73 sound/core/seq/seq_virmidi.c if (atomic) atomic 91 sound/core/seq/seq_virmidi.c if (atomic) atomic 103 sound/core/seq/seq_virmidi.c void *private_data, int atomic, int hop) atomic 110 sound/core/seq/seq_virmidi.c return snd_virmidi_dev_receive_event(rdev, ev, atomic); atomic 29 sound/drivers/opl3/opl3_oss.c void *private_data, int atomic, int hop) atomic 142 sound/drivers/opl3/opl3_seq.c void *private_data, int atomic, int hop) atomic 112 sound/drivers/opl4/opl4_seq.c void *private_data, int atomic, int hop) atomic 182 sound/isa/gus/gus_dma.c int atomic, atomic 188 sound/isa/gus/gus_dma.c block = kmalloc(sizeof(*block), atomic ? GFP_ATOMIC : GFP_KERNEL); atomic 167 sound/pci/pcxhr/pcxhr_core.c unsigned int itdsp, int atomic) atomic 192 sound/pci/pcxhr/pcxhr_core.c if (atomic) atomic 198 sound/pci/pcxhr/pcxhr_core.c if (atomic) atomic 27 sound/synth/emux/emux_oss.c void *private, int atomic, int hop); atomic 30 sound/synth/emux/emux_oss.c int cmd, unsigned char *event, int atomic, int hop); atomic 32 sound/synth/emux/emux_oss.c int cmd, unsigned char *event, int atomic, int hop); atomic 34 sound/synth/emux/emux_oss.c int ch, int param, int val, int atomic, int hop); atomic 288 sound/synth/emux/emux_oss.c int atomic, int hop) atomic 301 sound/synth/emux/emux_oss.c return snd_emux_event_input(ev, direct, private_data, atomic, hop); atomic 309 sound/synth/emux/emux_oss.c emuspec_control(emu, p, cmd, data, atomic, hop); atomic 311 sound/synth/emux/emux_oss.c gusspec_control(emu, p, cmd, data, atomic, hop); atomic 321 sound/synth/emux/emux_oss.c unsigned char *event, int atomic, int hop) atomic 361 sound/synth/emux/emux_oss.c fake_event(emu, port, voice, MIDI_CTL_ALL_NOTES_OFF, 0, atomic, hop); atomic 364 sound/synth/emux/emux_oss.c fake_event(emu, port, voice, MIDI_CTL_ALL_SOUNDS_OFF, 0, atomic, hop); atomic 416 sound/synth/emux/emux_oss.c unsigned char *event, int atomic, int hop) atomic 481 sound/synth/emux/emux_oss.c fake_event(struct snd_emux *emu, struct snd_emux_port *port, int ch, int param, int val, int atomic, int hop) atomic 489 sound/synth/emux/emux_oss.c snd_emux_event_input(&ev, 0, port, atomic, hop); atomic 240 sound/synth/emux/emux_seq.c int atomic, int hop) atomic 25 sound/synth/emux/emux_voice.h int atomic, int hop); atomic 1657 virt/kvm/kvm_main.c static kvm_pfn_t hva_to_pfn(unsigned long addr, bool atomic, bool *async, atomic 1665 virt/kvm/kvm_main.c BUG_ON(atomic && async); atomic 1670 virt/kvm/kvm_main.c if (atomic) atomic 1706 virt/kvm/kvm_main.c bool atomic, bool *async, bool write_fault, atomic 1729 virt/kvm/kvm_main.c return hva_to_pfn(addr, atomic, async, write_fault, atomic 1846 virt/kvm/kvm_main.c bool atomic) atomic 1860 virt/kvm/kvm_main.c if (atomic) atomic 1866 virt/kvm/kvm_main.c if (atomic) atomic 1875 virt/kvm/kvm_main.c if (atomic) atomic 1880 virt/kvm/kvm_main.c } else if (!atomic) { atomic 1899 virt/kvm/kvm_main.c struct gfn_to_pfn_cache *cache, bool atomic) atomic 1902 virt/kvm/kvm_main.c cache, atomic); atomic 1916 virt/kvm/kvm_main.c bool dirty, bool atomic) atomic 1925 virt/kvm/kvm_main.c if (atomic) atomic 1931 virt/kvm/kvm_main.c else if (!atomic) atomic 1950 virt/kvm/kvm_main.c struct gfn_to_pfn_cache *cache, bool dirty, bool atomic) atomic 1953 virt/kvm/kvm_main.c cache, dirty, atomic);