rw                 79 arch/arc/include/asm/spinlock.h static inline void arch_read_lock(arch_rwlock_t *rw)
rw                101 arch/arc/include/asm/spinlock.h 	: [rwlock]	"r"	(&(rw->counter)),
rw                109 arch/arc/include/asm/spinlock.h static inline int arch_read_trylock(arch_rwlock_t *rw)
rw                125 arch/arc/include/asm/spinlock.h 	: [rwlock]	"r"	(&(rw->counter)),
rw                134 arch/arc/include/asm/spinlock.h static inline void arch_write_lock(arch_rwlock_t *rw)
rw                158 arch/arc/include/asm/spinlock.h 	: [rwlock]	"r"	(&(rw->counter)),
rw                167 arch/arc/include/asm/spinlock.h static inline int arch_write_trylock(arch_rwlock_t *rw)
rw                183 arch/arc/include/asm/spinlock.h 	: [rwlock]	"r"	(&(rw->counter)),
rw                193 arch/arc/include/asm/spinlock.h static inline void arch_read_unlock(arch_rwlock_t *rw)
rw                209 arch/arc/include/asm/spinlock.h 	: [rwlock]	"r"	(&(rw->counter))
rw                213 arch/arc/include/asm/spinlock.h static inline void arch_write_unlock(arch_rwlock_t *rw)
rw                217 arch/arc/include/asm/spinlock.h 	WRITE_ONCE(rw->counter, __ARCH_RW_LOCK_UNLOCKED__);
rw                304 arch/arc/include/asm/spinlock.h static inline int arch_read_trylock(arch_rwlock_t *rw)
rw                310 arch/arc/include/asm/spinlock.h 	arch_spin_lock(&(rw->lock_mutex));
rw                316 arch/arc/include/asm/spinlock.h 	if (rw->counter > 0) {
rw                317 arch/arc/include/asm/spinlock.h 		rw->counter--;
rw                321 arch/arc/include/asm/spinlock.h 	arch_spin_unlock(&(rw->lock_mutex));
rw                328 arch/arc/include/asm/spinlock.h static inline int arch_write_trylock(arch_rwlock_t *rw)
rw                334 arch/arc/include/asm/spinlock.h 	arch_spin_lock(&(rw->lock_mutex));
rw                342 arch/arc/include/asm/spinlock.h 	if (rw->counter == __ARCH_RW_LOCK_UNLOCKED__) {
rw                343 arch/arc/include/asm/spinlock.h 		rw->counter = 0;
rw                346 arch/arc/include/asm/spinlock.h 	arch_spin_unlock(&(rw->lock_mutex));
rw                352 arch/arc/include/asm/spinlock.h static inline void arch_read_lock(arch_rwlock_t *rw)
rw                354 arch/arc/include/asm/spinlock.h 	while (!arch_read_trylock(rw))
rw                358 arch/arc/include/asm/spinlock.h static inline void arch_write_lock(arch_rwlock_t *rw)
rw                360 arch/arc/include/asm/spinlock.h 	while (!arch_write_trylock(rw))
rw                364 arch/arc/include/asm/spinlock.h static inline void arch_read_unlock(arch_rwlock_t *rw)
rw                369 arch/arc/include/asm/spinlock.h 	arch_spin_lock(&(rw->lock_mutex));
rw                370 arch/arc/include/asm/spinlock.h 	rw->counter++;
rw                371 arch/arc/include/asm/spinlock.h 	arch_spin_unlock(&(rw->lock_mutex));
rw                375 arch/arc/include/asm/spinlock.h static inline void arch_write_unlock(arch_rwlock_t *rw)
rw                380 arch/arc/include/asm/spinlock.h 	arch_spin_lock(&(rw->lock_mutex));
rw                381 arch/arc/include/asm/spinlock.h 	rw->counter = __ARCH_RW_LOCK_UNLOCKED__;
rw                382 arch/arc/include/asm/spinlock.h 	arch_spin_unlock(&(rw->lock_mutex));
rw                139 arch/arm/include/asm/spinlock.h static inline void arch_write_lock(arch_rwlock_t *rw)
rw                143 arch/arm/include/asm/spinlock.h 	prefetchw(&rw->lock);
rw                152 arch/arm/include/asm/spinlock.h 	: "r" (&rw->lock), "r" (0x80000000)
rw                158 arch/arm/include/asm/spinlock.h static inline int arch_write_trylock(arch_rwlock_t *rw)
rw                162 arch/arm/include/asm/spinlock.h 	prefetchw(&rw->lock);
rw                170 arch/arm/include/asm/spinlock.h 		: "r" (&rw->lock), "r" (0x80000000)
rw                182 arch/arm/include/asm/spinlock.h static inline void arch_write_unlock(arch_rwlock_t *rw)
rw                189 arch/arm/include/asm/spinlock.h 	: "r" (&rw->lock), "r" (0)
rw                207 arch/arm/include/asm/spinlock.h static inline void arch_read_lock(arch_rwlock_t *rw)
rw                211 arch/arm/include/asm/spinlock.h 	prefetchw(&rw->lock);
rw                221 arch/arm/include/asm/spinlock.h 	: "r" (&rw->lock)
rw                227 arch/arm/include/asm/spinlock.h static inline void arch_read_unlock(arch_rwlock_t *rw)
rw                233 arch/arm/include/asm/spinlock.h 	prefetchw(&rw->lock);
rw                241 arch/arm/include/asm/spinlock.h 	: "r" (&rw->lock)
rw                248 arch/arm/include/asm/spinlock.h static inline int arch_read_trylock(arch_rwlock_t *rw)
rw                252 arch/arm/include/asm/spinlock.h 	prefetchw(&rw->lock);
rw                260 arch/arm/include/asm/spinlock.h 		: "r" (&rw->lock)
rw                165 arch/ia64/include/asm/spinlock.h #define arch_read_lock_flags(rw, flags) arch_read_lock(rw)
rw                167 arch/ia64/include/asm/spinlock.h #define arch_read_lock(rw)								\
rw                169 arch/ia64/include/asm/spinlock.h 	arch_rwlock_t *__read_lock_ptr = (rw);						\
rw                180 arch/ia64/include/asm/spinlock.h #define arch_read_unlock(rw)					\
rw                182 arch/ia64/include/asm/spinlock.h 	arch_rwlock_t *__read_lock_ptr = (rw);			\
rw                214 arch/ia64/include/asm/spinlock.h #define arch_write_lock(rw) arch_write_lock_flags(rw, 0)
rw                216 arch/ia64/include/asm/spinlock.h #define arch_write_trylock(rw)							\
rw                224 arch/ia64/include/asm/spinlock.h 		: "=r"(result) : "r"(rw) : "ar.ccv", "r29", "memory");		\
rw                248 arch/ia64/include/asm/spinlock.h #define arch_write_trylock(rw)						\
rw                252 arch/ia64/include/asm/spinlock.h 	ia64_val = ia64_cmpxchg4_acq((__u32 *)(rw), ia64_set_val, 0);	\
rw                 59 arch/ia64/include/asm/syscall.h 	struct pt_regs *regs, unsigned long *args, int rw);
rw               2143 arch/ia64/kernel/ptrace.c 	int rw;
rw               2165 arch/ia64/kernel/ptrace.c 		if (args->rw)
rw               2173 arch/ia64/kernel/ptrace.c 	if (!args->rw) {
rw               2182 arch/ia64/kernel/ptrace.c 	struct pt_regs *regs, unsigned long *args, int rw)
rw               2189 arch/ia64/kernel/ptrace.c 		.rw = rw,
rw               1060 arch/microblaze/pci/pci-common.c #define NULL_PCI_OP(rw, size, type)					\
rw               1062 arch/microblaze/pci/pci-common.c null_##rw##_config_##size(struct pci_dev *dev, int offset, type val)	\
rw               1104 arch/microblaze/pci/pci-common.c #define EARLY_PCI_OP(rw, size, type)					\
rw               1105 arch/microblaze/pci/pci-common.c int early_##rw##_config_##size(struct pci_controller *hose, int bus,	\
rw               1108 arch/microblaze/pci/pci-common.c 	return pci_bus_##rw##_config_##size(fake_pci_bus(hose, bus),	\
rw                322 arch/mips/include/asm/octeon/cvmx-mio-defs.h 		uint64_t rw:1;
rw                340 arch/mips/include/asm/octeon/cvmx-mio-defs.h 		uint64_t rw:1;
rw                883 arch/mips/include/asm/octeon/cvmx-mio-defs.h 		uint64_t rw:1;
rw                891 arch/mips/include/asm/octeon/cvmx-mio-defs.h 		uint64_t rw:1;
rw               2501 arch/mips/include/asm/octeon/cvmx-mio-defs.h 		uint64_t rw:1;
rw               2519 arch/mips/include/asm/octeon/cvmx-mio-defs.h 		uint64_t rw:1;
rw                 66 arch/mips/lasat/picvue.c 	data |= picvue->rw;
rw                 95 arch/mips/lasat/picvue.c 	data &= ~picvue->rw;
rw                 13 arch/mips/lasat/picvue.h 	u32 rw;
rw                 73 arch/parisc/include/asm/spinlock.h static  __inline__ void arch_read_lock(arch_rwlock_t *rw)
rw                 77 arch/parisc/include/asm/spinlock.h 	arch_spin_lock_flags(&rw->lock, flags);
rw                 78 arch/parisc/include/asm/spinlock.h 	rw->counter++;
rw                 79 arch/parisc/include/asm/spinlock.h 	arch_spin_unlock(&rw->lock);
rw                 85 arch/parisc/include/asm/spinlock.h static  __inline__ void arch_read_unlock(arch_rwlock_t *rw)
rw                 89 arch/parisc/include/asm/spinlock.h 	arch_spin_lock_flags(&rw->lock, flags);
rw                 90 arch/parisc/include/asm/spinlock.h 	rw->counter--;
rw                 91 arch/parisc/include/asm/spinlock.h 	arch_spin_unlock(&rw->lock);
rw                 97 arch/parisc/include/asm/spinlock.h static __inline__ int arch_read_trylock(arch_rwlock_t *rw)
rw                102 arch/parisc/include/asm/spinlock.h 	if (arch_spin_trylock(&rw->lock)) {
rw                103 arch/parisc/include/asm/spinlock.h 		rw->counter++;
rw                104 arch/parisc/include/asm/spinlock.h 		arch_spin_unlock(&rw->lock);
rw                111 arch/parisc/include/asm/spinlock.h 	if (rw->counter < 0)
rw                115 arch/parisc/include/asm/spinlock.h 	while (arch_spin_is_locked(&rw->lock) && rw->counter >= 0)
rw                123 arch/parisc/include/asm/spinlock.h static __inline__ void arch_write_lock(arch_rwlock_t *rw)
rw                128 arch/parisc/include/asm/spinlock.h 	arch_spin_lock_flags(&rw->lock, flags);
rw                130 arch/parisc/include/asm/spinlock.h 	if (rw->counter != 0) {
rw                131 arch/parisc/include/asm/spinlock.h 		arch_spin_unlock(&rw->lock);
rw                134 arch/parisc/include/asm/spinlock.h 		while (rw->counter != 0)
rw                140 arch/parisc/include/asm/spinlock.h 	rw->counter = -1; /* mark as write-locked */
rw                145 arch/parisc/include/asm/spinlock.h static __inline__ void arch_write_unlock(arch_rwlock_t *rw)
rw                147 arch/parisc/include/asm/spinlock.h 	rw->counter = 0;
rw                148 arch/parisc/include/asm/spinlock.h 	arch_spin_unlock(&rw->lock);
rw                153 arch/parisc/include/asm/spinlock.h static __inline__ int arch_write_trylock(arch_rwlock_t *rw)
rw                159 arch/parisc/include/asm/spinlock.h 	if (arch_spin_trylock(&rw->lock)) {
rw                160 arch/parisc/include/asm/spinlock.h 		if (rw->counter == 0) {
rw                161 arch/parisc/include/asm/spinlock.h 			rw->counter = -1;
rw                165 arch/parisc/include/asm/spinlock.h 			arch_spin_unlock(&rw->lock);
rw                209 arch/powerpc/include/asm/spinlock.h static inline long __arch_read_trylock(arch_rwlock_t *rw)
rw                223 arch/powerpc/include/asm/spinlock.h 	: "r" (&rw->lock)
rw                233 arch/powerpc/include/asm/spinlock.h static inline long __arch_write_trylock(arch_rwlock_t *rw)
rw                247 arch/powerpc/include/asm/spinlock.h 	: "r" (token), "r" (&rw->lock)
rw                253 arch/powerpc/include/asm/spinlock.h static inline void arch_read_lock(arch_rwlock_t *rw)
rw                256 arch/powerpc/include/asm/spinlock.h 		if (likely(__arch_read_trylock(rw) > 0))
rw                261 arch/powerpc/include/asm/spinlock.h 				splpar_rw_yield(rw);
rw                262 arch/powerpc/include/asm/spinlock.h 		} while (unlikely(rw->lock < 0));
rw                267 arch/powerpc/include/asm/spinlock.h static inline void arch_write_lock(arch_rwlock_t *rw)
rw                270 arch/powerpc/include/asm/spinlock.h 		if (likely(__arch_write_trylock(rw) == 0))
rw                275 arch/powerpc/include/asm/spinlock.h 				splpar_rw_yield(rw);
rw                276 arch/powerpc/include/asm/spinlock.h 		} while (unlikely(rw->lock != 0));
rw                281 arch/powerpc/include/asm/spinlock.h static inline int arch_read_trylock(arch_rwlock_t *rw)
rw                283 arch/powerpc/include/asm/spinlock.h 	return __arch_read_trylock(rw) > 0;
rw                286 arch/powerpc/include/asm/spinlock.h static inline int arch_write_trylock(arch_rwlock_t *rw)
rw                288 arch/powerpc/include/asm/spinlock.h 	return __arch_write_trylock(rw) == 0;
rw                291 arch/powerpc/include/asm/spinlock.h static inline void arch_read_unlock(arch_rwlock_t *rw)
rw                304 arch/powerpc/include/asm/spinlock.h 	: "r"(&rw->lock)
rw                308 arch/powerpc/include/asm/spinlock.h static inline void arch_write_unlock(arch_rwlock_t *rw)
rw                312 arch/powerpc/include/asm/spinlock.h 	rw->lock = 0;
rw               1515 arch/powerpc/kernel/pci-common.c #define NULL_PCI_OP(rw, size, type)					\
rw               1517 arch/powerpc/kernel/pci-common.c null_##rw##_config_##size(struct pci_dev *dev, int offset, type val)	\
rw               1560 arch/powerpc/kernel/pci-common.c #define EARLY_PCI_OP(rw, size, type)					\
rw               1561 arch/powerpc/kernel/pci-common.c int early_##rw##_config_##size(struct pci_controller *hose, int bus,	\
rw               1564 arch/powerpc/kernel/pci-common.c 	return pci_bus_##rw##_config_##size(fake_pci_bus(hose, bus),	\
rw                 46 arch/powerpc/lib/locks.c void splpar_rw_yield(arch_rwlock_t *rw)
rw                 51 arch/powerpc/lib/locks.c 	lock_value = rw->lock;
rw                 60 arch/powerpc/lib/locks.c 	if (rw->lock != lock_value)
rw                115 arch/powerpc/platforms/powermac/low_i2c.c 	int			rw;
rw                284 arch/powerpc/platforms/powermac/low_i2c.c 			else if (host->rw) {
rw                444 arch/powerpc/platforms/powermac/low_i2c.c 	host->rw = (addrdir & 1);
rw                 26 arch/riscv/include/asm/barrier.h #define __smp_mb()	RISCV_FENCE(rw,rw)
rw                 33 arch/riscv/include/asm/barrier.h 	RISCV_FENCE(rw,w);						\
rw                 41 arch/riscv/include/asm/barrier.h 	RISCV_FENCE(r,rw);						\
rw                 62 arch/riscv/include/asm/barrier.h #define smp_mb__after_spinlock()	RISCV_FENCE(rw,rw)
rw                106 arch/s390/include/asm/spinlock.h #define arch_read_relax(rw) barrier()
rw                107 arch/s390/include/asm/spinlock.h #define arch_write_relax(rw) barrier()
rw                112 arch/s390/include/asm/spinlock.h static inline void arch_read_lock(arch_rwlock_t *rw)
rw                116 arch/s390/include/asm/spinlock.h 	old = __atomic_add(1, &rw->cnts);
rw                118 arch/s390/include/asm/spinlock.h 		arch_read_lock_wait(rw);
rw                121 arch/s390/include/asm/spinlock.h static inline void arch_read_unlock(arch_rwlock_t *rw)
rw                123 arch/s390/include/asm/spinlock.h 	__atomic_add_const_barrier(-1, &rw->cnts);
rw                126 arch/s390/include/asm/spinlock.h static inline void arch_write_lock(arch_rwlock_t *rw)
rw                128 arch/s390/include/asm/spinlock.h 	if (!__atomic_cmpxchg_bool(&rw->cnts, 0, 0x30000))
rw                129 arch/s390/include/asm/spinlock.h 		arch_write_lock_wait(rw);
rw                132 arch/s390/include/asm/spinlock.h static inline void arch_write_unlock(arch_rwlock_t *rw)
rw                134 arch/s390/include/asm/spinlock.h 	__atomic_add_barrier(-0x30000, &rw->cnts);
rw                138 arch/s390/include/asm/spinlock.h static inline int arch_read_trylock(arch_rwlock_t *rw)
rw                142 arch/s390/include/asm/spinlock.h 	old = READ_ONCE(rw->cnts);
rw                144 arch/s390/include/asm/spinlock.h 		__atomic_cmpxchg_bool(&rw->cnts, old, old + 1));
rw                147 arch/s390/include/asm/spinlock.h static inline int arch_write_trylock(arch_rwlock_t *rw)
rw                151 arch/s390/include/asm/spinlock.h 	old = READ_ONCE(rw->cnts);
rw                152 arch/s390/include/asm/spinlock.h 	return !old && __atomic_cmpxchg_bool(&rw->cnts, 0, 0x30000);
rw                270 arch/s390/lib/spinlock.c void arch_read_lock_wait(arch_rwlock_t *rw)
rw                273 arch/s390/lib/spinlock.c 		while (READ_ONCE(rw->cnts) & 0x10000)
rw                279 arch/s390/lib/spinlock.c 	__atomic_add_const(-1, &rw->cnts);
rw                281 arch/s390/lib/spinlock.c 	arch_spin_lock(&rw->wait);
rw                283 arch/s390/lib/spinlock.c 	__atomic_add_const(1, &rw->cnts);
rw                285 arch/s390/lib/spinlock.c 	while (READ_ONCE(rw->cnts) & 0x10000)
rw                287 arch/s390/lib/spinlock.c 	arch_spin_unlock(&rw->wait);
rw                291 arch/s390/lib/spinlock.c void arch_write_lock_wait(arch_rwlock_t *rw)
rw                296 arch/s390/lib/spinlock.c 	__atomic_add(0x20000, &rw->cnts);
rw                299 arch/s390/lib/spinlock.c 	arch_spin_lock(&rw->wait);
rw                302 arch/s390/lib/spinlock.c 		old = READ_ONCE(rw->cnts);
rw                304 arch/s390/lib/spinlock.c 		    __atomic_cmpxchg_bool(&rw->cnts, old, old | 0x10000))
rw                310 arch/s390/lib/spinlock.c 	arch_spin_unlock(&rw->wait);
rw                 27 arch/s390/pci/pci_event.c 	u32 rw		:  1;		/* read/write */
rw                 33 arch/sh/drivers/pci/common.c #define EARLY_PCI_OP(rw, size, type)					\
rw                 34 arch/sh/drivers/pci/common.c int __init early_##rw##_config_##size(struct pci_channel *hose,		\
rw                 37 arch/sh/drivers/pci/common.c 	return pci_##rw##_config_##size(				\
rw                 52 arch/sh/include/asm/spinlock-cas.h static inline void arch_read_lock(arch_rwlock_t *rw)
rw                 55 arch/sh/include/asm/spinlock-cas.h 	do old = rw->lock;
rw                 56 arch/sh/include/asm/spinlock-cas.h 	while (!old || __sl_cas(&rw->lock, old, old-1) != old);
rw                 59 arch/sh/include/asm/spinlock-cas.h static inline void arch_read_unlock(arch_rwlock_t *rw)
rw                 62 arch/sh/include/asm/spinlock-cas.h 	do old = rw->lock;
rw                 63 arch/sh/include/asm/spinlock-cas.h 	while (__sl_cas(&rw->lock, old, old+1) != old);
rw                 66 arch/sh/include/asm/spinlock-cas.h static inline void arch_write_lock(arch_rwlock_t *rw)
rw                 68 arch/sh/include/asm/spinlock-cas.h 	while (__sl_cas(&rw->lock, RW_LOCK_BIAS, 0) != RW_LOCK_BIAS);
rw                 71 arch/sh/include/asm/spinlock-cas.h static inline void arch_write_unlock(arch_rwlock_t *rw)
rw                 73 arch/sh/include/asm/spinlock-cas.h 	__sl_cas(&rw->lock, 0, RW_LOCK_BIAS);
rw                 76 arch/sh/include/asm/spinlock-cas.h static inline int arch_read_trylock(arch_rwlock_t *rw)
rw                 79 arch/sh/include/asm/spinlock-cas.h 	do old = rw->lock;
rw                 80 arch/sh/include/asm/spinlock-cas.h 	while (old && __sl_cas(&rw->lock, old, old-1) != old);
rw                 84 arch/sh/include/asm/spinlock-cas.h static inline int arch_write_trylock(arch_rwlock_t *rw)
rw                 86 arch/sh/include/asm/spinlock-cas.h 	return __sl_cas(&rw->lock, RW_LOCK_BIAS, 0) == RW_LOCK_BIAS;
rw                 90 arch/sh/include/asm/spinlock-llsc.h static inline void arch_read_lock(arch_rwlock_t *rw)
rw                103 arch/sh/include/asm/spinlock-llsc.h 		: "r" (&rw->lock)
rw                108 arch/sh/include/asm/spinlock-llsc.h static inline void arch_read_unlock(arch_rwlock_t *rw)
rw                119 arch/sh/include/asm/spinlock-llsc.h 		: "r" (&rw->lock)
rw                124 arch/sh/include/asm/spinlock-llsc.h static inline void arch_write_lock(arch_rwlock_t *rw)
rw                137 arch/sh/include/asm/spinlock-llsc.h 		: "r" (&rw->lock), "r" (RW_LOCK_BIAS)
rw                142 arch/sh/include/asm/spinlock-llsc.h static inline void arch_write_unlock(arch_rwlock_t *rw)
rw                147 arch/sh/include/asm/spinlock-llsc.h 		: "r" (&rw->lock), "r" (RW_LOCK_BIAS)
rw                152 arch/sh/include/asm/spinlock-llsc.h static inline int arch_read_trylock(arch_rwlock_t *rw)
rw                168 arch/sh/include/asm/spinlock-llsc.h 		: "r" (&rw->lock)
rw                175 arch/sh/include/asm/spinlock-llsc.h static inline int arch_write_trylock(arch_rwlock_t *rw)
rw                191 arch/sh/include/asm/spinlock-llsc.h 		: "r" (&rw->lock), "r" (RW_LOCK_BIAS)
rw                 79 arch/sparc/include/asm/spinlock_32.h static inline void __arch_read_lock(arch_rwlock_t *rw)
rw                 82 arch/sparc/include/asm/spinlock_32.h 	lp = rw;
rw                 99 arch/sparc/include/asm/spinlock_32.h static inline void __arch_read_unlock(arch_rwlock_t *rw)
rw                102 arch/sparc/include/asm/spinlock_32.h 	lp = rw;
rw                119 arch/sparc/include/asm/spinlock_32.h static inline void arch_write_lock(arch_rwlock_t *rw)
rw                122 arch/sparc/include/asm/spinlock_32.h 	lp = rw;
rw                142 arch/sparc/include/asm/spinlock_32.h static inline int arch_write_trylock(arch_rwlock_t *rw)
rw                148 arch/sparc/include/asm/spinlock_32.h 			     : "r" (&rw->lock)
rw                152 arch/sparc/include/asm/spinlock_32.h 		val = rw->lock & ~0xff;
rw                154 arch/sparc/include/asm/spinlock_32.h 			((volatile u8*)&rw->lock)[3] = 0;
rw                156 arch/sparc/include/asm/spinlock_32.h 			*(volatile u32*)&rw->lock = ~0U;
rw                162 arch/sparc/include/asm/spinlock_32.h static inline int __arch_read_trylock(arch_rwlock_t *rw)
rw                166 arch/sparc/include/asm/spinlock_32.h 	lp = rw;
rw                124 arch/sparc/kernel/process_32.c 	struct reg_window32 *rw = (struct reg_window32 *) r->u_regs[14];
rw                140 arch/sparc/kernel/process_32.c 	       rw->locals[0], rw->locals[1], rw->locals[2], rw->locals[3],
rw                141 arch/sparc/kernel/process_32.c 	       rw->locals[4], rw->locals[5], rw->locals[6], rw->locals[7]);
rw                143 arch/sparc/kernel/process_32.c 	       rw->ins[0], rw->ins[1], rw->ins[2], rw->ins[3],
rw                144 arch/sparc/kernel/process_32.c 	       rw->ins[4], rw->ins[5], rw->ins[6], rw->ins[7]);
rw                155 arch/sparc/kernel/process_32.c 	struct reg_window32 *rw;
rw                171 arch/sparc/kernel/process_32.c 		rw = (struct reg_window32 *) fp;
rw                172 arch/sparc/kernel/process_32.c 		pc = rw->ins[7];
rw                175 arch/sparc/kernel/process_32.c 		fp = rw->ins[6];
rw                466 arch/sparc/kernel/process_32.c 	struct reg_window32 *rw;
rw                479 arch/sparc/kernel/process_32.c 		rw = (struct reg_window32 *) fp;
rw                480 arch/sparc/kernel/process_32.c 		pc = rw->ins[7];
rw                485 arch/sparc/kernel/process_32.c 		fp = rw->ins[6] + bias;
rw                111 arch/sparc/kernel/process_64.c 	struct reg_window32 __user *rw;
rw                116 arch/sparc/kernel/process_64.c 	rw = compat_ptr((unsigned int)regs->u_regs[14]);
rw                119 arch/sparc/kernel/process_64.c 	if (copy_from_user (&r_w, rw, sizeof(r_w))) {
rw                140 arch/sparc/kernel/process_64.c 	struct reg_window __user *rw;
rw                147 arch/sparc/kernel/process_64.c 		rw = (struct reg_window __user *)
rw                154 arch/sparc/kernel/process_64.c 			if (copy_from_user (&r_w, rw, sizeof(r_w))) {
rw                219 arch/sparc/kernel/process_64.c 		struct reg_window *rw;
rw                221 arch/sparc/kernel/process_64.c 		rw = (struct reg_window *)
rw                223 arch/sparc/kernel/process_64.c 		if (kstack_valid(tp, (unsigned long) rw)) {
rw                224 arch/sparc/kernel/process_64.c 			rp->i7 = rw->ins[7];
rw                225 arch/sparc/kernel/process_64.c 			rw = (struct reg_window *)
rw                226 arch/sparc/kernel/process_64.c 				(rw->ins[6] + STACK_BIAS);
rw                227 arch/sparc/kernel/process_64.c 			if (kstack_valid(tp, (unsigned long) rw))
rw                228 arch/sparc/kernel/process_64.c 				rp->rpc = rw->ins[7];
rw                774 arch/sparc/kernel/process_64.c 	struct reg_window *rw;
rw                789 arch/sparc/kernel/process_64.c 		rw = (struct reg_window *) fp;
rw                790 arch/sparc/kernel/process_64.c 		pc = rw->ins[7];
rw                795 arch/sparc/kernel/process_64.c 		fp = rw->ins[6] + bias;
rw                 72 arch/sparc/kernel/traps_32.c 		struct reg_window32 *rw = (struct reg_window32 *)regs->u_regs[UREG_FP];
rw                 78 arch/sparc/kernel/traps_32.c 		while(rw					&&
rw                 80 arch/sparc/kernel/traps_32.c                       (((unsigned long) rw) >= PAGE_OFFSET)	&&
rw                 81 arch/sparc/kernel/traps_32.c 		      !(((unsigned long) rw) & 0x7)) {
rw                 82 arch/sparc/kernel/traps_32.c 			printk("Caller[%08lx]: %pS\n", rw->ins[7],
rw                 83 arch/sparc/kernel/traps_32.c 			       (void *) rw->ins[7]);
rw                 84 arch/sparc/kernel/traps_32.c 			rw = (struct reg_window32 *)rw->ins[6];
rw               2515 arch/sparc/kernel/traps_64.c static inline struct reg_window *kernel_stack_up(struct reg_window *rw)
rw               2517 arch/sparc/kernel/traps_64.c 	unsigned long fp = rw->ins[6];
rw               2544 arch/sparc/kernel/traps_64.c 		struct reg_window *rw = (struct reg_window *)
rw               2550 arch/sparc/kernel/traps_64.c 		while (rw &&
rw               2552 arch/sparc/kernel/traps_64.c 		       kstack_valid(tp, (unsigned long) rw)) {
rw               2553 arch/sparc/kernel/traps_64.c 			printk("Caller[%016lx]: %pS\n", rw->ins[7],
rw               2554 arch/sparc/kernel/traps_64.c 			       (void *) rw->ins[7]);
rw               2556 arch/sparc/kernel/traps_64.c 			rw = kernel_stack_up(rw);
rw                101 arch/x86/kvm/trace.h 	TP_PROTO(unsigned int rw, unsigned int port, unsigned int size,
rw                103 arch/x86/kvm/trace.h 	TP_ARGS(rw, port, size, count, data),
rw                106 arch/x86/kvm/trace.h 		__field(	unsigned int, 	rw		)
rw                114 arch/x86/kvm/trace.h 		__entry->rw		= rw;
rw                127 arch/x86/kvm/trace.h 		  __entry->rw ? "write" : "read",
rw                195 arch/x86/kvm/trace.h 	TP_PROTO(unsigned int rw, unsigned int reg, unsigned int val),
rw                196 arch/x86/kvm/trace.h 	TP_ARGS(rw, reg, val),
rw                199 arch/x86/kvm/trace.h 		__field(	unsigned int,	rw		)
rw                205 arch/x86/kvm/trace.h 		__entry->rw		= rw;
rw                211 arch/x86/kvm/trace.h 		  __entry->rw ? "write" : "read",
rw                362 arch/x86/kvm/trace.h 	TP_PROTO(unsigned int rw, unsigned int cr, unsigned long val),
rw                363 arch/x86/kvm/trace.h 	TP_ARGS(rw, cr, val),
rw                366 arch/x86/kvm/trace.h 		__field(	unsigned int,	rw		)
rw                372 arch/x86/kvm/trace.h 		__entry->rw		= rw;
rw                378 arch/x86/kvm/trace.h 		  __entry->rw ? "write" : "read",
rw               1323 arch/x86/kvm/trace.h 	    TP_PROTO(u32 vcpu, u32 offset, bool ft, bool rw, u32 vec),
rw               1324 arch/x86/kvm/trace.h 	    TP_ARGS(vcpu, offset, ft, rw, vec),
rw               1330 arch/x86/kvm/trace.h 		__field(bool, rw)
rw               1338 arch/x86/kvm/trace.h 		__entry->rw = rw;
rw               1347 arch/x86/kvm/trace.h 		  __entry->rw ? "write" : "read",
rw               1350 block/blk-core.c 	int rw = rq_data_dir(rq);
rw               1359 block/blk-core.c 		part_stat_inc(part, merges[rw]);
rw               1374 block/blk-core.c 		part_inc_in_flight(rq->q, part, rw);
rw               1282 block/blk-iocost.c 	int cpu, rw;
rw               1288 block/blk-iocost.c 		for (rw = READ; rw <= WRITE; rw++) {
rw               1289 block/blk-iocost.c 			u32 this_met = READ_ONCE(stat->missed[rw].nr_met);
rw               1290 block/blk-iocost.c 			u32 this_missed = READ_ONCE(stat->missed[rw].nr_missed);
rw               1292 block/blk-iocost.c 			nr_met[rw] += this_met - stat->missed[rw].last_met;
rw               1293 block/blk-iocost.c 			nr_missed[rw] += this_missed - stat->missed[rw].last_missed;
rw               1294 block/blk-iocost.c 			stat->missed[rw].last_met = this_met;
rw               1295 block/blk-iocost.c 			stat->missed[rw].last_missed = this_missed;
rw               1303 block/blk-iocost.c 	for (rw = READ; rw <= WRITE; rw++) {
rw               1304 block/blk-iocost.c 		if (nr_met[rw] + nr_missed[rw])
rw               1305 block/blk-iocost.c 			missed_ppm_ar[rw] =
rw               1306 block/blk-iocost.c 				DIV64_U64_ROUND_UP((u64)nr_missed[rw] * MILLION,
rw               1307 block/blk-iocost.c 						   nr_met[rw] + nr_missed[rw]);
rw               1309 block/blk-iocost.c 			missed_ppm_ar[rw] = 0;
rw               1875 block/blk-iocost.c 	int pidx, rw;
rw               1883 block/blk-iocost.c 		rw = READ;
rw               1887 block/blk-iocost.c 		rw = WRITE;
rw               1897 block/blk-iocost.c 		this_cpu_inc(ioc->pcpu_stat->missed[rw].nr_met);
rw               1899 block/blk-iocost.c 		this_cpu_inc(ioc->pcpu_stat->missed[rw].nr_missed);
rw                292 block/blk-throttle.c static uint64_t tg_bps_limit(struct throtl_grp *tg, int rw)
rw                302 block/blk-throttle.c 	ret = tg->bps[rw][td->limit_index];
rw                306 block/blk-throttle.c 		    tg->iops[rw][td->limit_index])
rw                312 block/blk-throttle.c 	if (td->limit_index == LIMIT_MAX && tg->bps[rw][LIMIT_LOW] &&
rw                313 block/blk-throttle.c 	    tg->bps[rw][LIMIT_LOW] != tg->bps[rw][LIMIT_MAX]) {
rw                316 block/blk-throttle.c 		adjusted = throtl_adjusted_limit(tg->bps[rw][LIMIT_LOW], td);
rw                317 block/blk-throttle.c 		ret = min(tg->bps[rw][LIMIT_MAX], adjusted);
rw                322 block/blk-throttle.c static unsigned int tg_iops_limit(struct throtl_grp *tg, int rw)
rw                332 block/blk-throttle.c 	ret = tg->iops[rw][td->limit_index];
rw                336 block/blk-throttle.c 		    tg->bps[rw][td->limit_index])
rw                342 block/blk-throttle.c 	if (td->limit_index == LIMIT_MAX && tg->iops[rw][LIMIT_LOW] &&
rw                343 block/blk-throttle.c 	    tg->iops[rw][LIMIT_LOW] != tg->iops[rw][LIMIT_MAX]) {
rw                346 block/blk-throttle.c 		adjusted = throtl_adjusted_limit(tg->iops[rw][LIMIT_LOW], td);
rw                349 block/blk-throttle.c 		ret = min_t(unsigned int, tg->iops[rw][LIMIT_MAX], adjusted);
rw                486 block/blk-throttle.c 	int rw;
rw                494 block/blk-throttle.c 	for (rw = READ; rw <= WRITE; rw++) {
rw                495 block/blk-throttle.c 		throtl_qnode_init(&tg->qnode_on_self[rw], tg);
rw                496 block/blk-throttle.c 		throtl_qnode_init(&tg->qnode_on_parent[rw], tg);
rw                553 block/blk-throttle.c 	int rw;
rw                555 block/blk-throttle.c 	for (rw = READ; rw <= WRITE; rw++)
rw                556 block/blk-throttle.c 		tg->has_rules[rw] = (parent_tg && parent_tg->has_rules[rw]) ||
rw                558 block/blk-throttle.c 			 (tg_bps_limit(tg, rw) != U64_MAX ||
rw                559 block/blk-throttle.c 			  tg_iops_limit(tg, rw) != UINT_MAX));
rw                760 block/blk-throttle.c 		bool rw, unsigned long start)
rw                762 block/blk-throttle.c 	tg->bytes_disp[rw] = 0;
rw                763 block/blk-throttle.c 	tg->io_disp[rw] = 0;
rw                771 block/blk-throttle.c 	if (time_after_eq(start, tg->slice_start[rw]))
rw                772 block/blk-throttle.c 		tg->slice_start[rw] = start;
rw                774 block/blk-throttle.c 	tg->slice_end[rw] = jiffies + tg->td->throtl_slice;
rw                777 block/blk-throttle.c 		   rw == READ ? 'R' : 'W', tg->slice_start[rw],
rw                778 block/blk-throttle.c 		   tg->slice_end[rw], jiffies);
rw                781 block/blk-throttle.c static inline void throtl_start_new_slice(struct throtl_grp *tg, bool rw)
rw                783 block/blk-throttle.c 	tg->bytes_disp[rw] = 0;
rw                784 block/blk-throttle.c 	tg->io_disp[rw] = 0;
rw                785 block/blk-throttle.c 	tg->slice_start[rw] = jiffies;
rw                786 block/blk-throttle.c 	tg->slice_end[rw] = jiffies + tg->td->throtl_slice;
rw                789 block/blk-throttle.c 		   rw == READ ? 'R' : 'W', tg->slice_start[rw],
rw                790 block/blk-throttle.c 		   tg->slice_end[rw], jiffies);
rw                793 block/blk-throttle.c static inline void throtl_set_slice_end(struct throtl_grp *tg, bool rw,
rw                796 block/blk-throttle.c 	tg->slice_end[rw] = roundup(jiffy_end, tg->td->throtl_slice);
rw                799 block/blk-throttle.c static inline void throtl_extend_slice(struct throtl_grp *tg, bool rw,
rw                802 block/blk-throttle.c 	tg->slice_end[rw] = roundup(jiffy_end, tg->td->throtl_slice);
rw                805 block/blk-throttle.c 		   rw == READ ? 'R' : 'W', tg->slice_start[rw],
rw                806 block/blk-throttle.c 		   tg->slice_end[rw], jiffies);
rw                810 block/blk-throttle.c static bool throtl_slice_used(struct throtl_grp *tg, bool rw)
rw                812 block/blk-throttle.c 	if (time_in_range(jiffies, tg->slice_start[rw], tg->slice_end[rw]))
rw                819 block/blk-throttle.c static inline void throtl_trim_slice(struct throtl_grp *tg, bool rw)
rw                824 block/blk-throttle.c 	BUG_ON(time_before(tg->slice_end[rw], tg->slice_start[rw]));
rw                831 block/blk-throttle.c 	if (throtl_slice_used(tg, rw))
rw                842 block/blk-throttle.c 	throtl_set_slice_end(tg, rw, jiffies + tg->td->throtl_slice);
rw                844 block/blk-throttle.c 	time_elapsed = jiffies - tg->slice_start[rw];
rw                850 block/blk-throttle.c 	tmp = tg_bps_limit(tg, rw) * tg->td->throtl_slice * nr_slices;
rw                854 block/blk-throttle.c 	io_trim = (tg_iops_limit(tg, rw) * tg->td->throtl_slice * nr_slices) /
rw                860 block/blk-throttle.c 	if (tg->bytes_disp[rw] >= bytes_trim)
rw                861 block/blk-throttle.c 		tg->bytes_disp[rw] -= bytes_trim;
rw                863 block/blk-throttle.c 		tg->bytes_disp[rw] = 0;
rw                865 block/blk-throttle.c 	if (tg->io_disp[rw] >= io_trim)
rw                866 block/blk-throttle.c 		tg->io_disp[rw] -= io_trim;
rw                868 block/blk-throttle.c 		tg->io_disp[rw] = 0;
rw                870 block/blk-throttle.c 	tg->slice_start[rw] += nr_slices * tg->td->throtl_slice;
rw                874 block/blk-throttle.c 		   rw == READ ? 'R' : 'W', nr_slices, bytes_trim, io_trim,
rw                875 block/blk-throttle.c 		   tg->slice_start[rw], tg->slice_end[rw], jiffies);
rw                881 block/blk-throttle.c 	bool rw = bio_data_dir(bio);
rw                886 block/blk-throttle.c 	jiffy_elapsed = jiffies - tg->slice_start[rw];
rw                898 block/blk-throttle.c 	tmp = (u64)tg_iops_limit(tg, rw) * jiffy_elapsed_rnd;
rw                906 block/blk-throttle.c 	if (tg->io_disp[rw] + 1 <= io_allowed) {
rw                923 block/blk-throttle.c 	bool rw = bio_data_dir(bio);
rw                928 block/blk-throttle.c 	jiffy_elapsed = jiffy_elapsed_rnd = jiffies - tg->slice_start[rw];
rw                936 block/blk-throttle.c 	tmp = tg_bps_limit(tg, rw) * jiffy_elapsed_rnd;
rw                940 block/blk-throttle.c 	if (tg->bytes_disp[rw] + bio_size <= bytes_allowed) {
rw                947 block/blk-throttle.c 	extra_bytes = tg->bytes_disp[rw] + bio_size - bytes_allowed;
rw                948 block/blk-throttle.c 	jiffy_wait = div64_u64(extra_bytes * HZ, tg_bps_limit(tg, rw));
rw                970 block/blk-throttle.c 	bool rw = bio_data_dir(bio);
rw                979 block/blk-throttle.c 	BUG_ON(tg->service_queue.nr_queued[rw] &&
rw                980 block/blk-throttle.c 	       bio != throtl_peek_queued(&tg->service_queue.queued[rw]));
rw                983 block/blk-throttle.c 	if (tg_bps_limit(tg, rw) == U64_MAX &&
rw                984 block/blk-throttle.c 	    tg_iops_limit(tg, rw) == UINT_MAX) {
rw                997 block/blk-throttle.c 	if (throtl_slice_used(tg, rw) && !(tg->service_queue.nr_queued[rw]))
rw                998 block/blk-throttle.c 		throtl_start_new_slice(tg, rw);
rw               1000 block/blk-throttle.c 		if (time_before(tg->slice_end[rw],
rw               1002 block/blk-throttle.c 			throtl_extend_slice(tg, rw,
rw               1018 block/blk-throttle.c 	if (time_before(tg->slice_end[rw], jiffies + max_wait))
rw               1019 block/blk-throttle.c 		throtl_extend_slice(tg, rw, jiffies + max_wait);
rw               1026 block/blk-throttle.c 	bool rw = bio_data_dir(bio);
rw               1030 block/blk-throttle.c 	tg->bytes_disp[rw] += bio_size;
rw               1031 block/blk-throttle.c 	tg->io_disp[rw]++;
rw               1032 block/blk-throttle.c 	tg->last_bytes_disp[rw] += bio_size;
rw               1033 block/blk-throttle.c 	tg->last_io_disp[rw]++;
rw               1058 block/blk-throttle.c 	bool rw = bio_data_dir(bio);
rw               1061 block/blk-throttle.c 		qn = &tg->qnode_on_self[rw];
rw               1069 block/blk-throttle.c 	if (!sq->nr_queued[rw])
rw               1072 block/blk-throttle.c 	throtl_qnode_add_bio(bio, qn, &sq->queued[rw]);
rw               1074 block/blk-throttle.c 	sq->nr_queued[rw]++;
rw               1105 block/blk-throttle.c 					struct throtl_grp *parent_tg, bool rw)
rw               1107 block/blk-throttle.c 	if (throtl_slice_used(parent_tg, rw)) {
rw               1108 block/blk-throttle.c 		throtl_start_new_slice_with_credit(parent_tg, rw,
rw               1109 block/blk-throttle.c 				child_tg->slice_start[rw]);
rw               1114 block/blk-throttle.c static void tg_dispatch_one_bio(struct throtl_grp *tg, bool rw)
rw               1128 block/blk-throttle.c 	bio = throtl_pop_queued(&sq->queued[rw], &tg_to_put);
rw               1129 block/blk-throttle.c 	sq->nr_queued[rw]--;
rw               1141 block/blk-throttle.c 		throtl_add_bio_tg(bio, &tg->qnode_on_parent[rw], parent_tg);
rw               1142 block/blk-throttle.c 		start_parent_slice_with_credit(tg, parent_tg, rw);
rw               1144 block/blk-throttle.c 		throtl_qnode_add_bio(bio, &tg->qnode_on_parent[rw],
rw               1145 block/blk-throttle.c 				     &parent_sq->queued[rw]);
rw               1146 block/blk-throttle.c 		BUG_ON(tg->td->nr_queued[rw] <= 0);
rw               1147 block/blk-throttle.c 		tg->td->nr_queued[rw]--;
rw               1150 block/blk-throttle.c 	throtl_trim_slice(tg, rw);
rw               1312 block/blk-throttle.c 	int rw;
rw               1317 block/blk-throttle.c 	for (rw = READ; rw <= WRITE; rw++)
rw               1318 block/blk-throttle.c 		while ((bio = throtl_pop_queued(&td_sq->queued[rw], NULL)))
rw               2038 block/blk-throttle.c 	int i, cpu, rw;
rw               2049 block/blk-throttle.c 	for (rw = READ; rw <= WRITE; rw++) {
rw               2051 block/blk-throttle.c 			struct latency_bucket *tmp = &td->tmp_buckets[rw][i];
rw               2057 block/blk-throttle.c 				bucket = per_cpu_ptr(td->latency_buckets[rw],
rw               2068 block/blk-throttle.c 				latency[rw] = tmp->total_latency;
rw               2072 block/blk-throttle.c 				latency[rw] /= samples;
rw               2073 block/blk-throttle.c 				if (latency[rw] == 0)
rw               2075 block/blk-throttle.c 				avg_latency[rw][i].latency = latency[rw];
rw               2080 block/blk-throttle.c 	for (rw = READ; rw <= WRITE; rw++) {
rw               2082 block/blk-throttle.c 			if (!avg_latency[rw][i].latency) {
rw               2083 block/blk-throttle.c 				if (td->avg_buckets[rw][i].latency < last_latency[rw])
rw               2084 block/blk-throttle.c 					td->avg_buckets[rw][i].latency =
rw               2085 block/blk-throttle.c 						last_latency[rw];
rw               2089 block/blk-throttle.c 			if (!td->avg_buckets[rw][i].valid)
rw               2090 block/blk-throttle.c 				latency[rw] = avg_latency[rw][i].latency;
rw               2092 block/blk-throttle.c 				latency[rw] = (td->avg_buckets[rw][i].latency * 7 +
rw               2093 block/blk-throttle.c 					avg_latency[rw][i].latency) >> 3;
rw               2095 block/blk-throttle.c 			td->avg_buckets[rw][i].latency = max(latency[rw],
rw               2096 block/blk-throttle.c 				last_latency[rw]);
rw               2097 block/blk-throttle.c 			td->avg_buckets[rw][i].valid = true;
rw               2098 block/blk-throttle.c 			last_latency[rw] = td->avg_buckets[rw][i].latency;
rw               2123 block/blk-throttle.c 	bool rw = bio_data_dir(bio);
rw               2130 block/blk-throttle.c 	if (bio_flagged(bio, BIO_THROTTLED) || !tg->has_rules[rw])
rw               2143 block/blk-throttle.c 		if (tg->last_low_overflow_time[rw] == 0)
rw               2144 block/blk-throttle.c 			tg->last_low_overflow_time[rw] = jiffies;
rw               2148 block/blk-throttle.c 		if (sq->nr_queued[rw])
rw               2153 block/blk-throttle.c 			tg->last_low_overflow_time[rw] = jiffies;
rw               2175 block/blk-throttle.c 		throtl_trim_slice(tg, rw);
rw               2182 block/blk-throttle.c 		qn = &tg->qnode_on_parent[rw];
rw               2191 block/blk-throttle.c 		   rw == READ ? 'R' : 'W',
rw               2192 block/blk-throttle.c 		   tg->bytes_disp[rw], bio->bi_iter.bi_size,
rw               2193 block/blk-throttle.c 		   tg_bps_limit(tg, rw),
rw               2194 block/blk-throttle.c 		   tg->io_disp[rw], tg_iops_limit(tg, rw),
rw               2197 block/blk-throttle.c 	tg->last_low_overflow_time[rw] = jiffies;
rw               2199 block/blk-throttle.c 	td->nr_queued[rw]++;
rw               2263 block/blk-throttle.c 	int rw = bio_data_dir(bio);
rw               2289 block/blk-throttle.c 		threshold = tg->td->avg_buckets[rw][bucket].latency +
rw               2343 block/blk-throttle.c 	int rw;
rw               2363 block/blk-throttle.c 	for (rw = READ; rw <= WRITE; rw++)
rw               2364 block/blk-throttle.c 		while ((bio = throtl_pop_queued(&td->service_queue.queued[rw],
rw                458 block/blk-wbt.c static inline unsigned int get_limit(struct rq_wb *rwb, unsigned long rw)
rw                469 block/blk-wbt.c 	if ((rw & REQ_OP_MASK) == REQ_OP_DISCARD)
rw                480 block/blk-wbt.c 	if ((rw & REQ_HIPRIO) || wb_recent_wait(rwb) || current_is_kswapd())
rw                482 block/blk-wbt.c 	else if ((rw & REQ_BACKGROUND) || close_io(rwb)) {
rw                497 block/blk-wbt.c 	unsigned long rw;
rw                503 block/blk-wbt.c 	return rq_wait_inc_below(rqw, get_limit(data->rwb, data->rw));
rw                517 block/blk-wbt.c 		       unsigned long rw)
rw                523 block/blk-wbt.c 		.rw = rw,
rw                290 block/bounce.c 	int rw = bio_data_dir(*bio_orig);
rw                330 block/bounce.c 		if (rw == WRITE) {
rw                348 block/bounce.c 		if (rw == READ)
rw                352 block/bounce.c 		if (rw == READ)
rw                 49 block/genhd.c  void part_inc_in_flight(struct request_queue *q, struct hd_struct *part, int rw)
rw                 54 block/genhd.c  	part_stat_local_inc(part, in_flight[rw]);
rw                 56 block/genhd.c  		part_stat_local_inc(&part_to_disk(part)->part0, in_flight[rw]);
rw                 59 block/genhd.c  void part_dec_in_flight(struct request_queue *q, struct hd_struct *part, int rw)
rw                 64 block/genhd.c  	part_stat_local_dec(part, in_flight[rw]);
rw                 66 block/genhd.c  		part_stat_local_dec(&part_to_disk(part)->part0, in_flight[rw]);
rw               2437 drivers/acpi/nfit/core.c 		resource_size_t dpa, void *iobuf, size_t len, int rw,
rw               2447 drivers/acpi/nfit/core.c 	write_blk_ctl(nfit_blk, lane, dpa, len, rw);
rw               2464 drivers/acpi/nfit/core.c 		if (rw)
rw               2478 drivers/acpi/nfit/core.c 	if (rw)
rw               2486 drivers/acpi/nfit/core.c 		resource_size_t dpa, void *iobuf, u64 len, int rw)
rw               2499 drivers/acpi/nfit/core.c 				iobuf + copied, c, rw, lane);
rw                245 drivers/acpi/nfit/nfit.h 			void *iobuf, u64 len, int rw);
rw                543 drivers/ata/libata-sff.c 			       unsigned int buflen, int rw)
rw                550 drivers/ata/libata-sff.c 	if (rw == READ)
rw                566 drivers/ata/libata-sff.c 		if (rw == READ) {
rw                598 drivers/ata/libata-sff.c 			       unsigned int buflen, int rw)
rw                607 drivers/ata/libata-sff.c 		return ata_sff_data_xfer(qc, buf, buflen, rw);
rw                610 drivers/ata/libata-sff.c 	if (rw == READ)
rw                626 drivers/ata/libata-sff.c 		if (rw == READ) {
rw                777 drivers/ata/libata-sff.c 	int rw = (qc->tf.flags & ATA_TFLAG_WRITE) ? WRITE : READ;
rw                812 drivers/ata/libata-sff.c 	consumed = ap->ops->sff_data_xfer(qc, buf + offset, count, rw);
rw               2947 drivers/ata/libata-sff.c 	unsigned int rw = (qc->tf.flags & ATA_TFLAG_WRITE);
rw               2957 drivers/ata/libata-sff.c 	if (!rw)
rw                381 drivers/ata/pata_arasan_cf.c 		u32 rw = acdev->qc->tf.flags & ATA_TFLAG_WRITE;
rw                383 drivers/ata/pata_arasan_cf.c 		dev_err(acdev->host->dev, "%s TimeOut", rw ? "write" : "read");
rw                 67 drivers/ata/pata_buddha.c 					 unsigned int buflen, int rw)
rw                 75 drivers/ata/pata_buddha.c 	if (rw == READ)
rw                 87 drivers/ata/pata_buddha.c 		if (rw == READ) {
rw                480 drivers/ata/pata_ep93xx.c 					  unsigned int buflen, int rw)
rw                489 drivers/ata/pata_ep93xx.c 		if (rw == READ)
rw                503 drivers/ata/pata_ep93xx.c 		if (rw == READ) {
rw                 45 drivers/ata/pata_falcon.c 					  unsigned int buflen, int rw)
rw                 59 drivers/ata/pata_falcon.c 	if (rw == READ) {
rw                 78 drivers/ata/pata_falcon.c 		if (rw == READ) {
rw                 45 drivers/ata/pata_gayle.c 					 unsigned int buflen, int rw)
rw                 53 drivers/ata/pata_gayle.c 	if (rw == READ)
rw                 65 drivers/ata/pata_gayle.c 		if (rw == READ) {
rw                 40 drivers/ata/pata_ixp4xx_cf.c 				unsigned char *buf, unsigned int buflen, int rw)
rw                 56 drivers/ata/pata_ixp4xx_cf.c 	if (rw == READ)
rw                 68 drivers/ata/pata_ixp4xx_cf.c 		if (rw == READ) {
rw                293 drivers/ata/pata_legacy.c 			unsigned char *buf, unsigned int buflen, int rw)
rw                312 drivers/ata/pata_legacy.c 		if (rw == READ)
rw                319 drivers/ata/pata_legacy.c 			if (rw == READ) {
rw                330 drivers/ata/pata_legacy.c 		buflen = ata_sff_data_xfer32(qc, buf, buflen, rw);
rw                694 drivers/ata/pata_legacy.c 				    unsigned int buflen, int rw)
rw                702 drivers/ata/pata_legacy.c 		if (rw == WRITE)
rw                709 drivers/ata/pata_legacy.c 			if (rw == WRITE) {
rw                719 drivers/ata/pata_legacy.c 		return ata_sff_data_xfer(qc, buf, buflen, rw);
rw                124 drivers/ata/pata_ns87415.c 	unsigned int rw = (qc->tf.flags & ATA_TFLAG_WRITE);
rw                137 drivers/ata/pata_ns87415.c 	if (!rw)
rw                296 drivers/ata/pata_octeon_cf.c 					 int rw)
rw                304 drivers/ata/pata_octeon_cf.c 	if (rw) {
rw                335 drivers/ata/pata_octeon_cf.c 					  int rw)
rw                343 drivers/ata/pata_octeon_cf.c 	if (rw) {
rw                367 drivers/ata/pata_octeon_cf.c 		if (rw == READ) {
rw                588 drivers/ata/pata_octeon_cf.c 	mio_boot_dma_cfg.s.rw = ((qc->tf.flags & ATA_TFLAG_WRITE) != 0);
rw                609 drivers/ata/pata_octeon_cf.c 		(mio_boot_dma_cfg.s.rw) ? "write" : "read", sg->length,
rw                 92 drivers/ata/pata_pcmcia.c 				unsigned char *buf, unsigned int buflen, int rw)
rw                 96 drivers/ata/pata_pcmcia.c 	if (rw == READ)
rw                265 drivers/ata/pata_samsung_cf.c 				unsigned char *buf, unsigned int buflen, int rw)
rw                274 drivers/ata/pata_samsung_cf.c 	if (rw == READ)
rw               1913 drivers/ata/sata_mv.c 	unsigned int rw = (qc->tf.flags & ATA_TFLAG_WRITE);
rw               1914 drivers/ata/sata_mv.c 	u32 cmd = (rw ? 0 : ATA_DMA_WR) | ATA_DMA_START;
rw               2154 drivers/ata/sata_nv.c 	unsigned int rw;
rw               2168 drivers/ata/sata_nv.c 	rw = qc->tf.flags & ATA_TFLAG_WRITE;
rw               2177 drivers/ata/sata_nv.c 	if (!rw)
rw                451 drivers/ata/sata_rcar.c 					      unsigned int buflen, int rw)
rw                458 drivers/ata/sata_rcar.c 	if (rw == READ)
rw                474 drivers/ata/sata_rcar.c 		if (rw == READ) {
rw                564 drivers/ata/sata_rcar.c 	unsigned int rw = qc->tf.flags & ATA_TFLAG_WRITE;
rw                580 drivers/ata/sata_rcar.c 	if (!rw)
rw                278 drivers/ata/sata_sil.c 	unsigned int rw = (qc->tf.flags & ATA_TFLAG_WRITE);
rw                287 drivers/ata/sata_sil.c 	if (!rw)
rw                233 drivers/ata/sata_svw.c 	unsigned int rw = (qc->tf.flags & ATA_TFLAG_WRITE);
rw                244 drivers/ata/sata_svw.c 	if (!rw)
rw                627 drivers/ata/sata_sx4.c 	unsigned int rw = (qc->tf.flags & ATA_TFLAG_WRITE);
rw                641 drivers/ata/sata_sx4.c 	if (rw && qc->tf.protocol == ATA_PROT_DMA) {
rw                212 drivers/auxdisplay/panel.c 		int rw;
rw                946 drivers/auxdisplay/panel.c 		lcd.pins.rw = PIN_INITP;
rw                992 drivers/auxdisplay/panel.c 		lcd.pins.rw = lcd_rw_pin;
rw               1025 drivers/auxdisplay/panel.c 		if (lcd.pins.rw == PIN_NOT_SET)
rw               1026 drivers/auxdisplay/panel.c 			lcd.pins.rw = DEFAULT_LCD_PIN_RW;
rw               1038 drivers/auxdisplay/panel.c 	if (lcd.pins.rw == PIN_NOT_SET)
rw               1039 drivers/auxdisplay/panel.c 		lcd.pins.rw = PIN_NONE;
rw               1059 drivers/auxdisplay/panel.c 	pin_to_bits(lcd.pins.rw, lcd_bits[LCD_PORT_D][LCD_BIT_RW],
rw               1741 drivers/auxdisplay/panel.c 		lcd.pins.rw = lcd_rw_pin;
rw               1211 drivers/block/drbd/drbd_req.c 	const int rw = bio_data_dir(bio);
rw               1240 drivers/block/drbd/drbd_req.c 	if (rw == WRITE && req->private_bio && req->i.size
rw               1325 drivers/block/drbd/drbd_req.c 	const int rw = bio_data_dir(req->master_bio);
rw               1331 drivers/block/drbd/drbd_req.c 	if (rw == WRITE) {
rw               1358 drivers/block/drbd/drbd_req.c 	if (rw != WRITE) {
rw               1369 drivers/block/drbd/drbd_req.c 		if (rw == WRITE)
rw               1375 drivers/block/drbd/drbd_req.c 	if (rw == WRITE) {
rw               1404 drivers/block/drbd/drbd_req.c 			&device->pending_master_completion[rw == WRITE]);
rw               1409 drivers/block/drbd/drbd_req.c 			&device->pending_completion[rw == WRITE]);
rw               1453 drivers/block/drbd/drbd_req.c 		const int rw = bio_data_dir(req->master_bio);
rw               1455 drivers/block/drbd/drbd_req.c 		if (rw == WRITE /* rw != WRITE should not even end up here! */
rw                514 drivers/block/loop.c 		     loff_t pos, bool rw)
rw                561 drivers/block/loop.c 	iov_iter_bvec(&iter, rw, bvec, nr_bvec, blk_rq_bytes(rq));
rw                572 drivers/block/loop.c 	if (rw == WRITE)
rw               1277 drivers/block/xen-blkback/blkback.c 	       req->u.indirect.nr_segments : req->u.rw.nr_segments;
rw               1292 drivers/block/xen-blkback/blkback.c 	pending_req->id        = req->u.rw.id;
rw               1298 drivers/block/xen-blkback/blkback.c 		preq.dev               = req->u.rw.handle;
rw               1299 drivers/block/xen-blkback/blkback.c 		preq.sector_number     = req->u.rw.sector_number;
rw               1301 drivers/block/xen-blkback/blkback.c 			pages[i]->gref = req->u.rw.seg[i].gref;
rw               1302 drivers/block/xen-blkback/blkback.c 			seg[i].nsec = req->u.rw.seg[i].last_sect -
rw               1303 drivers/block/xen-blkback/blkback.c 				req->u.rw.seg[i].first_sect + 1;
rw               1304 drivers/block/xen-blkback/blkback.c 			seg[i].offset = (req->u.rw.seg[i].first_sect << 9);
rw               1305 drivers/block/xen-blkback/blkback.c 			if ((req->u.rw.seg[i].last_sect >= (XEN_PAGE_SIZE >> 9)) ||
rw               1306 drivers/block/xen-blkback/blkback.c 			    (req->u.rw.seg[i].last_sect <
rw               1307 drivers/block/xen-blkback/blkback.c 			     req->u.rw.seg[i].first_sect))
rw               1421 drivers/block/xen-blkback/blkback.c 	make_response(ring, req->u.rw.id, req_operation, BLKIF_RSP_ERROR);
rw                124 drivers/block/xen-blkback/common.h 		struct blkif_x86_32_request_rw rw;
rw                180 drivers/block/xen-blkback/common.h 		struct blkif_x86_64_request_rw rw;
rw                405 drivers/block/xen-blkback/common.h 		dst->u.rw.nr_segments = src->u.rw.nr_segments;
rw                406 drivers/block/xen-blkback/common.h 		dst->u.rw.handle = src->u.rw.handle;
rw                407 drivers/block/xen-blkback/common.h 		dst->u.rw.id = src->u.rw.id;
rw                408 drivers/block/xen-blkback/common.h 		dst->u.rw.sector_number = src->u.rw.sector_number;
rw                410 drivers/block/xen-blkback/common.h 		if (n > dst->u.rw.nr_segments)
rw                411 drivers/block/xen-blkback/common.h 			n = dst->u.rw.nr_segments;
rw                413 drivers/block/xen-blkback/common.h 			dst->u.rw.seg[i] = src->u.rw.seg[i];
rw                453 drivers/block/xen-blkback/common.h 		dst->u.rw.nr_segments = src->u.rw.nr_segments;
rw                454 drivers/block/xen-blkback/common.h 		dst->u.rw.handle = src->u.rw.handle;
rw                455 drivers/block/xen-blkback/common.h 		dst->u.rw.id = src->u.rw.id;
rw                456 drivers/block/xen-blkback/common.h 		dst->u.rw.sector_number = src->u.rw.sector_number;
rw                458 drivers/block/xen-blkback/common.h 		if (n > dst->u.rw.nr_segments)
rw                459 drivers/block/xen-blkback/common.h 			n = dst->u.rw.nr_segments;
rw                461 drivers/block/xen-blkback/common.h 			dst->u.rw.seg[i] = src->u.rw.seg[i];
rw                271 drivers/block/xen-blkfront.c 	rinfo->shadow_free = rinfo->shadow[free].req.u.rw.id;
rw                272 drivers/block/xen-blkfront.c 	rinfo->shadow[free].req.u.rw.id = 0x0fffffee; /* debug */
rw                279 drivers/block/xen-blkfront.c 	if (rinfo->shadow[id].req.u.rw.id != id)
rw                283 drivers/block/xen-blkfront.c 	rinfo->shadow[id].req.u.rw.id  = rinfo->shadow_free;
rw                539 drivers/block/xen-blkfront.c 	(*ring_req)->u.rw.id = id;
rw                657 drivers/block/xen-blkfront.c 		ring_req->u.rw.seg[grant_idx] =
rw                676 drivers/block/xen-blkfront.c 	uint16_t nr_segments = first->u.rw.nr_segments;
rw                682 drivers/block/xen-blkfront.c 	first->u.rw.nr_segments = BLKIF_MAX_SEGMENTS_PER_REQUEST;
rw                684 drivers/block/xen-blkfront.c 	second->u.rw.nr_segments = nr_segments - BLKIF_MAX_SEGMENTS_PER_REQUEST;
rw                685 drivers/block/xen-blkfront.c 	second->u.rw.sector_number = first->u.rw.sector_number +
rw                688 drivers/block/xen-blkfront.c 	second->u.rw.handle = first->u.rw.handle;
rw                767 drivers/block/xen-blkfront.c 		ring_req->u.rw.sector_number = (blkif_sector_t)blk_rq_pos(req);
rw                768 drivers/block/xen-blkfront.c 		ring_req->u.rw.handle = info->handle;
rw                788 drivers/block/xen-blkfront.c 		ring_req->u.rw.nr_segments = num_grant;
rw               1290 drivers/block/xen-blkfront.c 		       rinfo->shadow[i].req.u.rw.nr_segments;
rw               1433 drivers/block/xen-blkfront.c 		s->req.u.indirect.nr_segments : s->req.u.rw.nr_segments;
rw               1453 drivers/block/xen-blkfront.c 		num_grant += s2->req.u.rw.nr_segments;
rw               1625 drivers/block/xen-blkfront.c 				     rinfo->shadow[id].req.u.rw.nr_segments == 0)) {
rw               1880 drivers/block/xen-blkfront.c 			rinfo->shadow[j].req.u.rw.id = j + 1;
rw               1881 drivers/block/xen-blkfront.c 		rinfo->shadow[BLK_RING_SIZE(info)-1].req.u.rw.id = 0x0fffffff;
rw               2799 drivers/edac/amd64_edac.c 		u8 rw;
rw               2804 drivers/edac/amd64_edac.c 		rw = dram_rw(pvt, range);
rw               2805 drivers/edac/amd64_edac.c 		if (!rw)
rw               2815 drivers/edac/amd64_edac.c 			 (rw & 0x1) ? "R" : "-",
rw               2816 drivers/edac/amd64_edac.c 			 (rw & 0x2) ? "W" : "-",
rw                743 drivers/gpu/drm/amd/powerplay/inc/amdgpu_smu.h #define smu_thermal_temperature_range_update(smu, range, rw) \
rw                744 drivers/gpu/drm/amd/powerplay/inc/amdgpu_smu.h 	((smu)->ppt_funcs->thermal_temperature_range_update? (smu)->ppt_funcs->thermal_temperature_range_update((smu), (range), (rw)) : 0)
rw                231 drivers/gpu/drm/amd/powerplay/inc/hwmgr.h 	int (*smc_table_manager)(struct pp_hwmgr *hwmgr, uint8_t *table, uint16_t table_id, bool rw); /*rw: true for read, false for write */
rw                115 drivers/gpu/drm/amd/powerplay/inc/smumgr.h extern int smum_smc_table_manager(struct pp_hwmgr *hwmgr, uint8_t *table, uint16_t table_id, bool rw);
rw                286 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c static int smu10_smc_table_manager(struct pp_hwmgr *hwmgr, uint8_t *table, uint16_t table_id, bool rw)
rw                290 drivers/gpu/drm/amd/powerplay/smumgr/smu10_smumgr.c 	if (rw)
rw                213 drivers/gpu/drm/amd/powerplay/smumgr/smumgr.c int smum_smc_table_manager(struct pp_hwmgr *hwmgr, uint8_t *table, uint16_t table_id, bool rw)
rw                216 drivers/gpu/drm/amd/powerplay/smumgr/smumgr.c 		return hwmgr->smumgr_funcs->smc_table_manager(hwmgr, table, table_id, rw);
rw                342 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 				    uint16_t table_id, bool rw)
rw                346 drivers/gpu/drm/amd/powerplay/smumgr/vega10_smumgr.c 	if (rw)
rw                380 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 				    uint16_t table_id, bool rw)
rw                384 drivers/gpu/drm/amd/powerplay/smumgr/vega12_smumgr.c 	if (rw)
rw                590 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 				    uint16_t table_id, bool rw)
rw                594 drivers/gpu/drm/amd/powerplay/smumgr/vega20_smumgr.c 	if (rw)
rw                 66 drivers/gpu/drm/i915/gvt/kvmgt.c 	size_t (*rw)(struct intel_vgpu *vgpu, char *buf,
rw                433 drivers/gpu/drm/i915/gvt/kvmgt.c 	.rw = intel_vgpu_reg_rw_opregion,
rw                538 drivers/gpu/drm/i915/gvt/kvmgt.c 	.rw = intel_vgpu_reg_rw_edid,
rw                974 drivers/gpu/drm/i915/gvt/kvmgt.c 		return vgpu->vdev.region[index].ops->rw(vgpu, buf, count,
rw                 13 drivers/gpu/drm/nouveau/nvkm/subdev/bios/priv.h 	bool rw;
rw                 90 drivers/gpu/drm/nouveau/nvkm/subdev/bios/shadow.c 				if (mthd->func->rw)
rw                163 drivers/gpu/drm/nouveau/nvkm/subdev/bios/shadow.c 	.rw = false,
rw                101 drivers/gpu/drm/nouveau/nvkm/subdev/bios/shadowacpi.c 	.rw = false,
rw                110 drivers/gpu/drm/nouveau/nvkm/subdev/bios/shadowacpi.c 	.rw = false,
rw                 76 drivers/gpu/drm/nouveau/nvkm/subdev/bios/shadowof.c 	.rw = false,
rw                 88 drivers/gpu/drm/nouveau/nvkm/subdev/bios/shadowpci.c 	.rw = true,
rw                120 drivers/gpu/drm/nouveau/nvkm/subdev/bios/shadowpci.c 	.rw = true,
rw                119 drivers/gpu/drm/nouveau/nvkm/subdev/bios/shadowramin.c 	.rw = true,
rw                 63 drivers/gpu/drm/nouveau/nvkm/subdev/bios/shadowrom.c 	.rw = false,
rw                 49 drivers/gpu/drm/vmwgfx/ttm_lock.c 	lock->rw = 0;
rw                 56 drivers/gpu/drm/vmwgfx/ttm_lock.c 	if (--lock->rw == 0)
rw                 66 drivers/gpu/drm/vmwgfx/ttm_lock.c 	if (lock->rw >= 0 && lock->flags == 0) {
rw                 67 drivers/gpu/drm/vmwgfx/ttm_lock.c 		++lock->rw;
rw                 93 drivers/gpu/drm/vmwgfx/ttm_lock.c 	if (lock->rw >= 0 && lock->flags == 0) {
rw                 94 drivers/gpu/drm/vmwgfx/ttm_lock.c 		++lock->rw;
rw                127 drivers/gpu/drm/vmwgfx/ttm_lock.c 	lock->rw = 0;
rw                137 drivers/gpu/drm/vmwgfx/ttm_lock.c 	if (lock->rw == 0 && ((lock->flags & ~TTM_WRITE_LOCK_PENDING) == 0)) {
rw                138 drivers/gpu/drm/vmwgfx/ttm_lock.c 		lock->rw = -1;
rw                180 drivers/gpu/drm/vmwgfx/ttm_lock.c 	if (lock->rw == 0) {
rw                 72 drivers/gpu/drm/vmwgfx/ttm_lock.h 	int32_t rw;
rw                311 drivers/hid/wacom_wac.c 	int rw = 0;
rw                370 drivers/hid/wacom_wac.c 				rw = (data[7] & 0x04) - (data[7] & 0x03);
rw                373 drivers/hid/wacom_wac.c 				rw = 44 - (data[6] >> 2);
rw                374 drivers/hid/wacom_wac.c 				rw = clamp_val(rw, 0, 31);
rw                375 drivers/hid/wacom_wac.c 				input_report_abs(input, ABS_DISTANCE, rw);
rw                380 drivers/hid/wacom_wac.c 					rw = (data[6] & 0x01) ? -1 :
rw                383 drivers/hid/wacom_wac.c 					rw = 0;
rw                387 drivers/hid/wacom_wac.c 				rw = -(signed char)data[6];
rw                389 drivers/hid/wacom_wac.c 			input_report_rel(input, REL_WHEEL, rw);
rw                407 drivers/hid/wacom_wac.c 			rw = ((data[7] & 0x18) >> 3) - ((data[7] & 0x20) >> 3);
rw                408 drivers/hid/wacom_wac.c 			input_report_rel(pad_input, REL_WHEEL, rw);
rw                447 drivers/hid/wacom_wac.c 		rw = (data[7] >> 2 & 0x07);
rw                448 drivers/hid/wacom_wac.c 		battery_capacity = batcap_gr[rw];
rw                449 drivers/hid/wacom_wac.c 		ps_connected = rw == 7;
rw                910 drivers/hwmon/ibmaem.c 				const struct aem_rw_sensor_template *rw)
rw                934 drivers/hwmon/ibmaem.c 	while (rw->label) {
rw                936 drivers/hwmon/ibmaem.c 		sensors->dev_attr.attr.name = rw->label;
rw                938 drivers/hwmon/ibmaem.c 		sensors->dev_attr.show = rw->show;
rw                939 drivers/hwmon/ibmaem.c 		sensors->dev_attr.store = rw->set;
rw                940 drivers/hwmon/ibmaem.c 		sensors->index = rw->index;
rw                948 drivers/hwmon/ibmaem.c 		rw++;
rw                187 drivers/i2c/busses/i2c-ali1563.c 			 union i2c_smbus_data *data, u8 rw)
rw                195 drivers/i2c/busses/i2c-ali1563.c 	if (rw == I2C_SMBUS_WRITE) {
rw                209 drivers/i2c/busses/i2c-ali1563.c 		if (rw == I2C_SMBUS_WRITE) {
rw                234 drivers/i2c/busses/i2c-ali1563.c 			  unsigned short flags, char rw, u8 cmd,
rw                273 drivers/i2c/busses/i2c-ali1563.c 	outb_p(((addr & 0x7f) << 1) | (rw & 0x01), SMB_HST_ADD);
rw                281 drivers/i2c/busses/i2c-ali1563.c 		if (rw == I2C_SMBUS_WRITE)
rw                287 drivers/i2c/busses/i2c-ali1563.c 		if (rw == I2C_SMBUS_WRITE)
rw                292 drivers/i2c/busses/i2c-ali1563.c 		if (rw == I2C_SMBUS_WRITE) {
rw                299 drivers/i2c/busses/i2c-ali1563.c 		error = ali1563_block(a, data, rw);
rw                307 drivers/i2c/busses/i2c-ali1563.c 	if ((rw == I2C_SMBUS_WRITE) || (size == HST_CNTL2_QUICK))
rw                114 drivers/i2c/busses/i2c-cbus-gpio.c static int cbus_transfer(struct cbus_host *host, char rw, unsigned dev,
rw                133 drivers/i2c/busses/i2c-cbus-gpio.c 	cbus_send_bit(host, rw == I2C_SMBUS_READ);
rw                138 drivers/i2c/busses/i2c-cbus-gpio.c 	if (rw == I2C_SMBUS_WRITE) {
rw                104 drivers/i2c/busses/i2c-ismt.c #define ISMT_DESC_ADDR_RW(addr, rw) (((addr) << 1) | (rw))
rw                203 drivers/i2c/busses/i2c-sprd.c static void sprd_i2c_opt_mode(struct sprd_i2c *i2c_dev, int rw)
rw                207 drivers/i2c/busses/i2c-sprd.c 	writel(cmd | rw << 3, i2c_dev->base + I2C_CTL);
rw                270 drivers/i2c/busses/scx200_acb.c 				 char rw, u8 command, int size,
rw                287 drivers/i2c/busses/scx200_acb.c 		buffer = rw ? &data->byte : &command;
rw                314 drivers/i2c/busses/scx200_acb.c 		size, address, command, len, rw);
rw                316 drivers/i2c/busses/scx200_acb.c 	if (!len && rw == I2C_SMBUS_READ) {
rw                323 drivers/i2c/busses/scx200_acb.c 	iface->address_byte = (address << 1) | rw;
rw                347 drivers/i2c/busses/scx200_acb.c 	if (rc == 0 && size == I2C_SMBUS_WORD_DATA && rw == I2C_SMBUS_READ)
rw                371 drivers/ide/ide-atapi.c 		      int ireason, int rw)
rw                375 drivers/ide/ide-atapi.c 	debug_log("ireason: 0x%x, rw: 0x%x\n", ireason, rw);
rw                377 drivers/ide/ide-atapi.c 	if (ireason == (!rw << 1))
rw                379 drivers/ide/ide-atapi.c 	else if (ireason == (rw << 1)) {
rw                384 drivers/ide/ide-atapi.c 			ide_pad_transfer(drive, rw, len);
rw                385 drivers/ide/ide-atapi.c 	} else if (!rw && ireason == ATAPI_COD) {
rw                192 drivers/ide/ide-dma-sff.c 	u8 rw = (cmd->tf_flags & IDE_TFLAG_WRITE) ? 0 : ATA_DMA_WR;
rw                210 drivers/ide/ide-dma-sff.c 		writeb(rw, (void __iomem *)(hwif->dma_base + ATA_DMA_CMD));
rw                212 drivers/ide/ide-dma-sff.c 		outb(rw, hwif->dma_base + ATA_DMA_CMD);
rw                195 drivers/ide/trm290.c 	unsigned int count, rw = (cmd->tf_flags & IDE_TFLAG_WRITE) ? 1 : 2;
rw                202 drivers/ide/trm290.c 	outl(hwif->dmatable_dma | rw, hwif->dma_base);
rw                291 drivers/ide/tx4939ide.c 	u8 rw = (cmd->tf_flags & IDE_TFLAG_WRITE) ? 0 : ATA_DMA_WR;
rw                301 drivers/ide/tx4939ide.c 	tx4939ide_writeb(rw, base, TX4939IDE_DMA_Cmd);
rw                706 drivers/infiniband/hw/mlx5/mr.c 	MLX5_SET(mkc, mkc, rw, !!(acc & IB_ACCESS_REMOTE_WRITE));
rw               1093 drivers/infiniband/hw/mlx5/mr.c 	MLX5_SET(mkc, mkc, rw, !!(access_flags & IB_ACCESS_REMOTE_WRITE));
rw               1173 drivers/infiniband/hw/mlx5/mr.c 	MLX5_SET(mkc, mkc, rw, !!(acc & IB_ACCESS_REMOTE_WRITE));
rw               2419 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 	struct ocrdma_sge *rw;
rw               2432 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		rw = (struct ocrdma_sge *)(hdr + 1);
rw               2434 drivers/infiniband/hw/ocrdma/ocrdma_verbs.c 		ibwc->byte_len = rw->len;
rw               1491 drivers/infiniband/ulp/isert/ib_isert.c 	if (!cmd->rw.nr_ops)
rw               1495 drivers/infiniband/ulp/isert/ib_isert.c 		rdma_rw_ctx_destroy_signature(&cmd->rw, conn->qp,
rw               1500 drivers/infiniband/ulp/isert/ib_isert.c 		rdma_rw_ctx_destroy(&cmd->rw, conn->qp, conn->cm_id->port_num,
rw               1504 drivers/infiniband/ulp/isert/ib_isert.c 	cmd->rw.nr_ops = 0;
rw               1672 drivers/infiniband/ulp/isert/ib_isert.c 	ret = isert_check_pi_status(cmd, isert_cmd->rw.reg->mr);
rw               1718 drivers/infiniband/ulp/isert/ib_isert.c 		ret = isert_check_pi_status(se_cmd, isert_cmd->rw.reg->mr);
rw               2149 drivers/infiniband/ulp/isert/ib_isert.c 		ret = rdma_rw_ctx_signature_init(&cmd->rw, conn->qp, port_num,
rw               2154 drivers/infiniband/ulp/isert/ib_isert.c 		ret = rdma_rw_ctx_init(&cmd->rw, conn->qp, port_num,
rw               2167 drivers/infiniband/ulp/isert/ib_isert.c 	ret = rdma_rw_ctx_post(&cmd->rw, conn->qp, port_num, cqe, chain_wr);
rw                126 drivers/infiniband/ulp/isert/ib_isert.h 	struct rdma_rw_ctx	rw;
rw                918 drivers/infiniband/ulp/srpt/ib_srpt.c 		ret = rdma_rw_ctx_init(&ctx->rw, ch->qp, ch->sport->port,
rw                947 drivers/infiniband/ulp/srpt/ib_srpt.c 		rdma_rw_ctx_destroy(&ctx->rw, ch->qp, ch->sport->port,
rw                965 drivers/infiniband/ulp/srpt/ib_srpt.c 		rdma_rw_ctx_destroy(&ctx->rw, ch->qp, ch->sport->port,
rw               2754 drivers/infiniband/ulp/srpt/ib_srpt.c 		first_wr = rdma_rw_ctx_wrs(&ctx->rw, ch->qp, ch->sport->port,
rw               2830 drivers/infiniband/ulp/srpt/ib_srpt.c 			first_wr = rdma_rw_ctx_wrs(&ctx->rw, ch->qp,
rw                195 drivers/infiniband/ulp/srpt/ib_srpt.h 	struct rdma_rw_ctx	rw;
rw                 32 drivers/input/mouse/cyapa.h #define SMBUS_ENCODE_RW(cmd, rw) ((cmd) | ((rw) & 0x01))
rw               1182 drivers/md/bcache/request.c 	int rw = bio_data_dir(bio);
rw               1229 drivers/md/bcache/request.c 			if (rw)
rw                554 drivers/md/dm-bufio.c static void use_dmio(struct dm_buffer *b, int rw, sector_t sector,
rw                559 drivers/md/dm-bufio.c 		.bi_op = rw,
rw                592 drivers/md/dm-bufio.c static void use_bio(struct dm_buffer *b, int rw, sector_t sector,
rw                606 drivers/md/dm-bufio.c 		use_dmio(b, rw, sector, n_sectors, offset);
rw                612 drivers/md/dm-bufio.c 	bio_set_op_attrs(bio, rw, 0);
rw                634 drivers/md/dm-bufio.c static void submit_io(struct dm_buffer *b, int rw, void (*end_io)(struct dm_buffer *, blk_status_t))
rw                648 drivers/md/dm-bufio.c 	if (rw != REQ_OP_WRITE) {
rw                667 drivers/md/dm-bufio.c 		use_bio(b, rw, sector, n_sectors, offset);
rw                669 drivers/md/dm-bufio.c 		use_dmio(b, rw, sector, n_sectors, offset);
rw               1421 drivers/md/dm-crypt.c 	unsigned rw = bio_data_dir(clone);
rw               1427 drivers/md/dm-crypt.c 	if (rw == WRITE)
rw               1433 drivers/md/dm-crypt.c 	if (rw == READ && !error) {
rw                355 drivers/md/dm-kcopyd.c 	int rw;
rw                421 drivers/md/dm-kcopyd.c 		if (job->rw == READ || !test_bit(DM_KCOPYD_WRITE_SEQ, &job->flags)) {
rw                523 drivers/md/dm-kcopyd.c 		if (op_is_write(job->rw))
rw                535 drivers/md/dm-kcopyd.c 	if (op_is_write(job->rw))
rw                539 drivers/md/dm-kcopyd.c 		job->rw = WRITE;
rw                554 drivers/md/dm-kcopyd.c 		.bi_op = job->rw,
rw                576 drivers/md/dm-kcopyd.c 	if (job->rw == READ)
rw                619 drivers/md/dm-kcopyd.c 			if (op_is_write(job->rw))
rw                823 drivers/md/dm-kcopyd.c 		job->rw = READ;
rw                832 drivers/md/dm-kcopyd.c 		job->rw = REQ_OP_WRITE_ZEROES;
rw                835 drivers/md/dm-kcopyd.c 				job->rw = WRITE;
rw                121 drivers/md/dm-raid1.c static void queue_bio(struct mirror_set *ms, struct bio *bio, int rw)
rw                127 drivers/md/dm-raid1.c 	bl = (rw == WRITE) ? &ms->writes : &ms->reads;
rw               1186 drivers/md/dm-raid1.c 	int r, rw = bio_data_dir(bio);
rw               1195 drivers/md/dm-raid1.c 	if (rw == WRITE) {
rw               1198 drivers/md/dm-raid1.c 		queue_bio(ms, bio, rw);
rw               1213 drivers/md/dm-raid1.c 		queue_bio(ms, bio, rw);
rw               1236 drivers/md/dm-raid1.c 	int rw = bio_data_dir(bio);
rw               1246 drivers/md/dm-raid1.c 	if (rw == WRITE) {
rw               1288 drivers/md/dm-raid1.c 			queue_bio(ms, bio, rw);
rw               2021 drivers/md/dm-thin.c 	int rw = bio_data_dir(bio);
rw               2028 drivers/md/dm-thin.c 		if (lookup_result.shared && (rw == WRITE) && bio->bi_iter.bi_size) {
rw               2043 drivers/md/dm-thin.c 		if (rw != READ) {
rw               1071 drivers/md/dm-writecache.c 	int rw = bio_data_dir(bio);
rw               1081 drivers/md/dm-writecache.c 		if (rw == READ) {
rw                374 drivers/md/md.c 	const int rw = bio_data_dir(bio);
rw                379 drivers/md/md.c 	if (unlikely(test_bit(MD_BROKEN, &mddev->flags)) && (rw == WRITE)) {
rw                390 drivers/md/md.c 	if (mddev->ro == 1 && unlikely(rw == WRITE)) {
rw               1954 drivers/md/raid1.c 			    int sectors, struct page *page, int rw)
rw               1956 drivers/md/raid1.c 	if (sync_page_io(rdev, sector, sectors << 9, page, rw, 0, false))
rw               1959 drivers/md/raid1.c 	if (rw == WRITE) {
rw               2299 drivers/md/raid10.c 			    int sectors, struct page *page, int rw)
rw               2305 drivers/md/raid10.c 	    && (rw == READ || test_bit(WriteErrorSeen, &rdev->flags)))
rw               2307 drivers/md/raid10.c 	if (sync_page_io(rdev, sector, sectors << 9, page, rw, 0, false))
rw               2310 drivers/md/raid10.c 	if (rw == WRITE) {
rw               5584 drivers/md/raid5.c 	const int rw = bio_data_dir(bi);
rw               5613 drivers/md/raid5.c 	if (rw == READ && mddev->degraded == 0 &&
rw               5712 drivers/md/raid5.c 			    !add_stripe_bio(sh, bi, dd_idx, rw, previous)) {
rw               5745 drivers/md/raid5.c 	if (rw == WRITE)
rw                214 drivers/media/dvb-frontends/dib8000.c 	u16 rw[2];
rw                221 drivers/media/dvb-frontends/dib8000.c 	rw[0] = __dib8000_read_word(state, reg + 0);
rw                222 drivers/media/dvb-frontends/dib8000.c 	rw[1] = __dib8000_read_word(state, reg + 1);
rw                226 drivers/media/dvb-frontends/dib8000.c 	return ((rw[0] << 16) | (rw[1]));
rw                103 drivers/media/pci/cx23885/altera-ci.c 	int (*fpga_rw) (void *dev, int flag, int data, int rw);
rw                 27 drivers/media/pci/cx23885/altera-ci.h 	int (*fpga_rw) (void *dev, int ad_rg, int val, int rw);
rw                293 drivers/media/platform/marvell-ccic/cafe-driver.c 		unsigned short flags, char rw, u8 command,
rw                308 drivers/media/platform/marvell-ccic/cafe-driver.c 	if (rw == I2C_SMBUS_WRITE)
rw                310 drivers/media/platform/marvell-ccic/cafe-driver.c 	else if (rw == I2C_SMBUS_READ)
rw                156 drivers/media/usb/uvc/uvc_v4l2.c 	u16 rw, rh;
rw                190 drivers/media/usb/uvc/uvc_v4l2.c 	rw = fmt->fmt.pix.width;
rw                198 drivers/media/usb/uvc/uvc_v4l2.c 		d = min(w, rw) * min(h, rh);
rw                199 drivers/media/usb/uvc/uvc_v4l2.c 		d = w*h + rw*rh - 2*d;
rw                155 drivers/media/v4l2-core/videobuf-dma-sg.c 	int err, rw = 0;
rw                161 drivers/media/v4l2-core/videobuf-dma-sg.c 		rw = READ;
rw                164 drivers/media/v4l2-core/videobuf-dma-sg.c 		rw = WRITE;
rw                180 drivers/media/v4l2-core/videobuf-dma-sg.c 	if (rw == READ)
rw               2121 drivers/mmc/core/mmc_test.c 				     struct mmc_test_multiple_rw *rw)
rw               2128 drivers/mmc/core/mmc_test.c 	if (rw->do_nonblock_req &&
rw               2134 drivers/mmc/core/mmc_test.c 	for (i = 0 ; i < rw->len && ret == 0; i++) {
rw               2135 drivers/mmc/core/mmc_test.c 		ret = mmc_test_rw_multiple(test, rw, rw->bs[i], rw->size, 0);
rw               2143 drivers/mmc/core/mmc_test.c 				       struct mmc_test_multiple_rw *rw)
rw               2148 drivers/mmc/core/mmc_test.c 	for (i = 0 ; i < rw->len && ret == 0; i++) {
rw               2149 drivers/mmc/core/mmc_test.c 		ret = mmc_test_rw_multiple(test, rw, 512 * 1024, rw->size,
rw               2150 drivers/mmc/core/mmc_test.c 					   rw->sg_len[i]);
rw                520 drivers/mmc/host/cavium.c 	int count, rw;
rw                527 drivers/mmc/host/cavium.c 	rw = (data->flags & MMC_DATA_WRITE) ? 1 : 0;
rw                529 drivers/mmc/host/cavium.c 		  FIELD_PREP(MIO_EMM_DMA_CFG_RW, rw);
rw                542 drivers/mmc/host/cavium.c 		 (rw) ? "W" : "R", sg_dma_len(&data->sg[0]), count);
rw                557 drivers/mmc/host/cavium.c 	int count, i, rw;
rw                581 drivers/mmc/host/cavium.c 		rw = (data->flags & MMC_DATA_WRITE) ? 1 : 0;
rw                582 drivers/mmc/host/cavium.c 		fifo_cmd = FIELD_PREP(MIO_EMM_DMA_FIFO_CMD_RW, rw);
rw                599 drivers/mmc/host/cavium.c 			 (rw) ? "W" : "R", sg_dma_len(sg), i, count);
rw               1050 drivers/mmc/host/s3cmci.c 	int rw = (data->flags & MMC_DATA_WRITE) ? 1 : 0;
rw               1057 drivers/mmc/host/s3cmci.c 	host->pio_active = rw ? XFER_WRITE : XFER_READ;
rw               1059 drivers/mmc/host/s3cmci.c 	if (rw) {
rw               1072 drivers/mmc/host/s3cmci.c 	int rw = data->flags & MMC_DATA_WRITE;
rw               1086 drivers/mmc/host/s3cmci.c 	if (!rw)
rw                170 drivers/mtd/nand/raw/meson_nand.c 		struct nand_rw_cmd rw;
rw                580 drivers/mtd/nand/raw/meson_nand.c 	u32 *addrs = nfc->cmdfifo.rw.addrs;
rw                588 drivers/mtd/nand/raw/meson_nand.c 	nfc->cmdfifo.rw.cmd0 = cs | NFC_CMD_CLE | cmd0;
rw                616 drivers/mtd/nand/raw/meson_nand.c 		nfc->cmdfifo.rw.cmd1 = cs | NFC_CMD_CLE | NAND_CMD_READSTART;
rw                617 drivers/mtd/nand/raw/meson_nand.c 		writel(nfc->cmdfifo.rw.cmd1, nfc->reg_base + NFC_REG_CMD);
rw                131 drivers/net/dsa/sja1105/sja1105.h 				sja1105_spi_rw_mode_t rw, u64 reg_addr,
rw                134 drivers/net/dsa/sja1105/sja1105.h 			 sja1105_spi_rw_mode_t rw, u64 reg_addr,
rw                137 drivers/net/dsa/sja1105/sja1105.h 				     sja1105_spi_rw_mode_t rw, u64 base_addr,
rw                 70 drivers/net/dsa/sja1105/sja1105_spi.c 				sja1105_spi_rw_mode_t rw, u64 reg_addr,
rw                 82 drivers/net/dsa/sja1105/sja1105_spi.c 	msg.access = rw;
rw                 84 drivers/net/dsa/sja1105/sja1105_spi.c 	if (rw == SPI_READ)
rw                 89 drivers/net/dsa/sja1105/sja1105_spi.c 	if (rw == SPI_WRITE)
rw                 97 drivers/net/dsa/sja1105/sja1105_spi.c 	if (rw == SPI_READ)
rw                116 drivers/net/dsa/sja1105/sja1105_spi.c 			 sja1105_spi_rw_mode_t rw, u64 reg_addr,
rw                125 drivers/net/dsa/sja1105/sja1105_spi.c 	if (rw == SPI_WRITE)
rw                129 drivers/net/dsa/sja1105/sja1105_spi.c 	rc = sja1105_spi_send_packed_buf(priv, rw, reg_addr, packed_buf,
rw                132 drivers/net/dsa/sja1105/sja1105_spi.c 	if (rw == SPI_READ)
rw                144 drivers/net/dsa/sja1105/sja1105_spi.c 				     sja1105_spi_rw_mode_t rw, u64 base_addr,
rw                161 drivers/net/dsa/sja1105/sja1105_spi.c 		rc = sja1105_spi_send_packed_buf(priv, rw, chunk.spi_address,
rw               1774 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h 		       u32 *val, int rw, bool sleep_ok);
rw               5311 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 			    unsigned int rw, bool sleep_ok)
rw               5321 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 						(rw ? FW_CMD_READ_F :
rw               5327 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 		c.u.addrval.val  = rw ? 0 : cpu_to_be32(vals[i]);
rw               5333 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 		if (rw)
rw               5354 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 			      u32 *buff, u32 nregs, u32 start_index, int rw,
rw               5375 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 		rc = t4_tp_fw_ldst_rw(adap, cmd, buff, nregs, start_index, rw,
rw               5381 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 		if (rw)
rw               7458 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 		       u32 *val, int rw, bool sleep_ok)
rw               7476 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c 		if (rw)
rw                767 drivers/net/ethernet/cortina/gemini.c 	union dma_rwptr rw;
rw                775 drivers/net/ethernet/cortina/gemini.c 	rw.bits32 = readl(ptr_reg);
rw                776 drivers/net/ethernet/cortina/gemini.c 	r = rw.bits.rptr;
rw                777 drivers/net/ethernet/cortina/gemini.c 	w = rw.bits.wptr;
rw                880 drivers/net/ethernet/cortina/gemini.c 	union dma_rwptr rw;
rw                888 drivers/net/ethernet/cortina/gemini.c 	rw.bits32 = readl(geth->base + GLOBAL_SWFQ_RWPTR_REG);
rw                889 drivers/net/ethernet/cortina/gemini.c 	pn = (refill ? rw.bits.wptr : rw.bits.rptr) >> fpp_order;
rw                890 drivers/net/ethernet/cortina/gemini.c 	epn = (rw.bits.rptr >> fpp_order) - 1;
rw               1237 drivers/net/ethernet/cortina/gemini.c 	union dma_rwptr rw;
rw               1248 drivers/net/ethernet/cortina/gemini.c 	rw.bits32 = readl(ptr_reg);
rw               1249 drivers/net/ethernet/cortina/gemini.c 	r = rw.bits.rptr;
rw               1250 drivers/net/ethernet/cortina/gemini.c 	w = rw.bits.wptr;
rw               1411 drivers/net/ethernet/cortina/gemini.c 	union dma_rwptr rw;
rw               1415 drivers/net/ethernet/cortina/gemini.c 	rw.bits32 = readl(ptr_reg);
rw               1419 drivers/net/ethernet/cortina/gemini.c 	r = rw.bits.rptr;
rw               1420 drivers/net/ethernet/cortina/gemini.c 	w = rw.bits.wptr;
rw                974 drivers/net/ethernet/dec/tulip/de4x5.c static void    mii_ta(u_long rw, u_long ioaddr);
rw               4871 drivers/net/ethernet/dec/tulip/de4x5.c mii_ta(u_long rw, u_long ioaddr)
rw               4873 drivers/net/ethernet/dec/tulip/de4x5.c     if (rw == MII_STWR) {
rw                298 drivers/net/ethernet/freescale/fman/fman_keygen.c 	u32 rw = (u32)(write ? FM_KG_KGAR_WRITE : FM_KG_KGAR_READ);
rw                301 drivers/net/ethernet/freescale/fman/fman_keygen.c 			rw |
rw                319 drivers/net/ethernet/freescale/fman/fman_keygen.c 	u32 rw = write ? (u32)FM_KG_KGAR_WRITE : (u32)FM_KG_KGAR_READ;
rw                322 drivers/net/ethernet/freescale/fman/fman_keygen.c 			rw |
rw                363 drivers/net/ethernet/freescale/fman/fman_keygen.c 	u32 rw = write ? (u32)FM_KG_KGAR_WRITE : (u32)FM_KG_KGAR_READ;
rw                366 drivers/net/ethernet/freescale/fman/fman_keygen.c 			rw |
rw                 85 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_icm_pool.c 		MLX5_SET(mkc, mkc, rw, 1);
rw                806 drivers/net/ethernet/mellanox/mlx5/core/steering/dr_send.c 	MLX5_SET(mkc, mkc, rw, 1);
rw                158 drivers/net/ethernet/seeq/sgiseeq.c 		sregs->rw.eth_addr[i] = dev->dev_addr[i];
rw                301 drivers/net/ethernet/seeq/sgiseeq.c 		sregs->rw.wregs.control = sp->control;
rw                302 drivers/net/ethernet/seeq/sgiseeq.c 		sregs->rw.wregs.frame_gap = 0;
rw                419 drivers/net/ethernet/seeq/sgiseeq.c 		sregs->rw.wregs.control = sp->control & ~(SEEQ_CTRL_XCNT);
rw                420 drivers/net/ethernet/seeq/sgiseeq.c 		sregs->rw.wregs.control = sp->control;
rw                783 drivers/net/ethernet/seeq/sgiseeq.c 	sp->is_edlc = !(sp->sregs->rw.rregs.collision_tx[0] & 0xff);
rw                 29 drivers/net/ethernet/seeq/sgiseeq.h 	} rw;
rw                109 drivers/net/wireless/ath/ath6kl/sdio.c static inline void ath6kl_sdio_set_cmd53_arg(u32 *arg, u8 rw, u8 func,
rw                113 drivers/net/wireless/ath/ath6kl/sdio.c 	*arg = (((rw & 1) << 31) |
rw                267 drivers/net/wireless/ath/ath6kl/sdio.c 	u8 opcode, rw;
rw                292 drivers/net/wireless/ath/ath6kl/sdio.c 	rw = (scat_req->req & HIF_WRITE) ? CMD53_ARG_WRITE : CMD53_ARG_READ;
rw                304 drivers/net/wireless/ath/ath6kl/sdio.c 	ath6kl_sdio_set_cmd53_arg(&cmd.arg, rw, ar_sdio->func->num,
rw                 42 drivers/net/wireless/rsi/rsi_91x_sdio.c static u32 rsi_sdio_set_cmd52_arg(bool rw,
rw                 48 drivers/net/wireless/rsi/rsi_91x_sdio.c 	return ((rw & 1) << 31) | ((func & 0x7) << 28) |
rw                 32 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c static int bits(u32 rw, int from, int to)
rw                 34 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 	rw &= ~(0xffffffffU << (to+1));
rw                 35 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 	rw >>= from;
rw                 36 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 	return rw;
rw                 39 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c static int bit(u32 rw, int bit)
rw                 41 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 	return bits(rw, bit, bit);
rw                 44 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c static void dump_regwrite(u32 rw)
rw                 46 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 	int reg = bits(rw, 18, 22);
rw                 47 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 	int rw_flag = bits(rw, 23, 23);
rw                 48 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 	PDEBUG("rf2959 %#010x reg %d rw %d", rw, reg, rw_flag);
rw                 54 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 		       bits(rw, 14, 15), bit(rw, 3), bit(rw, 2), bit(rw, 1),
rw                 55 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 		       bit(rw, 0));
rw                 61 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 		       bit(rw, 17), bit(rw, 16), bit(rw, 15), bit(rw, 14),
rw                 62 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 		       bit(rw, 13), bit(rw, 12), bit(rw, 11), bit(rw, 10),
rw                 63 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 		       bits(rw, 7, 9), bits(rw, 4, 6), bits(rw, 0, 3));
rw                 67 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 		       bits(rw, 6, 17), bits(rw, 0, 5));
rw                 70 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 		PDEBUG("reg3 IFPLL3 num %d", bits(rw, 0, 17));
rw                 74 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 		       bits(rw, 8, 16), bits(rw, 4, 7), bits(rw, 0, 3));
rw                 80 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 		       bit(rw, 17), bit(rw, 16), bit(rw, 15), bit(rw, 14),
rw                 81 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 		       bit(rw, 13), bit(rw, 12), bit(rw, 11), bit(rw, 10),
rw                 82 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 		       bits(rw, 7, 9), bits(rw, 4, 6), bits(rw, 0,3));
rw                 86 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 		       bits(rw, 6, 17), bits(rw, 0, 5));
rw                 89 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 		PDEBUG("reg7 RFPLL3 num2 %d", bits(rw, 0, 17));
rw                 93 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 		       bits(rw, 8, 16), bits(rw, 4, 7), bits(rw, 0, 3));
rw                 97 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 		       bits(rw, 13, 17), bits(rw, 8, 12), bits(rw, 3, 7),
rw                 98 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 		       bits(rw, 0, 2));
rw                104 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 		       bit(rw, 17), bits(rw, 15, 16), bits(rw, 10, 14),
rw                105 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 		       bits(rw, 7, 9), bits(rw, 4, 6), bit(rw, 3), bit(rw, 2),
rw                106 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 		       bit(rw, 1), bit(rw, 0));
rw                111 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 			bits(rw, 15, 17), bits(rw, 9, 14), bits(rw, 3, 8),
rw                112 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 			bits(rw, 0, 2));
rw                116 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 		       bits(rw, 12, 17), bits(rw, 6, 11), bits(rw, 0, 5));
rw                122 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 		       bit(rw, 17), bit(rw, 16), bit(rw, 15),
rw                123 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 		       bits(rw, 8, 9), bits(rw, 5, 7), bits(rw, 3, 4),
rw                124 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 		       bits(rw, 0, 2));
rw                129 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 		       bits(rw, 13, 17), bits(rw, 9, 12), bits(rw, 4, 8),
rw                130 drivers/net/wireless/zydas/zd1211rw/zd_rf_rf2959.c 		       bits(rw, 0, 3));
rw               1912 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 		struct reg_data *rw  = &req->reg_writes[i];
rw               1913 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 		rw->addr = cpu_to_le16((u16)ioreqs[i].addr);
rw               1914 drivers/net/wireless/zydas/zd1211rw/zd_usb.c 		rw->value = cpu_to_le16(ioreqs[i].value);
rw                 64 drivers/nvdimm/blk.c 		struct bio_integrity_payload *bip, u64 lba, int rw)
rw                 94 drivers/nvdimm/blk.c 				cur_len, rw);
rw                110 drivers/nvdimm/blk.c 		struct bio_integrity_payload *bip, u64 lba, int rw)
rw                118 drivers/nvdimm/blk.c 		unsigned int len, unsigned int off, int rw, sector_t sector)
rw                147 drivers/nvdimm/blk.c 		err = ndbr->do_io(ndbr, dev_offset, iobuf + off, cur_len, rw);
rw                153 drivers/nvdimm/blk.c 			err = nd_blk_rw_integrity(nsblk, bip, lba, rw);
rw                172 drivers/nvdimm/blk.c 	int err = 0, rw;
rw                180 drivers/nvdimm/blk.c 	rw = bio_data_dir(bio);
rw                187 drivers/nvdimm/blk.c 				bvec.bv_offset, rw, iter.bi_sector);
rw                191 drivers/nvdimm/blk.c 					(rw == READ) ? "READ" : "WRITE",
rw                205 drivers/nvdimm/blk.c 		resource_size_t offset, void *iobuf, size_t n, int rw,
rw                222 drivers/nvdimm/blk.c 	return ndbr->do_io(ndbr, dev_offset, iobuf, n, rw);
rw               1142 drivers/nvdimm/btt.c 			struct arena_info *arena, u32 postmap, int rw)
rw               1167 drivers/nvdimm/btt.c 		if (rw)
rw               1191 drivers/nvdimm/btt.c 			struct arena_info *arena, u32 postmap, int rw)
rw                252 drivers/nvdimm/claim.c 		resource_size_t offset, void *buf, size_t size, int rw,
rw                268 drivers/nvdimm/claim.c 	if (rw == READ) {
rw                165 drivers/nvdimm/nd.h 			void *iobuf, u64 len, int rw);
rw                699 drivers/nvme/host/core.c 	cmnd->rw.opcode = (rq_data_dir(req) ? nvme_cmd_write : nvme_cmd_read);
rw                700 drivers/nvme/host/core.c 	cmnd->rw.nsid = cpu_to_le32(ns->head->ns_id);
rw                701 drivers/nvme/host/core.c 	cmnd->rw.slba = cpu_to_le64(nvme_block_nr(ns, blk_rq_pos(req)));
rw                702 drivers/nvme/host/core.c 	cmnd->rw.length = cpu_to_le16((blk_rq_bytes(req) >> ns->lba_shift) - 1);
rw                728 drivers/nvme/host/core.c 			cmnd->rw.reftag = cpu_to_le32(t10_pi_ref_tag(req));
rw                733 drivers/nvme/host/core.c 	cmnd->rw.control = cpu_to_le16(control);
rw                734 drivers/nvme/host/core.c 	cmnd->rw.dsmgmt = cpu_to_le32(dsmgmt);
rw               1303 drivers/nvme/host/core.c 	c.rw.opcode = io.opcode;
rw               1304 drivers/nvme/host/core.c 	c.rw.flags = io.flags;
rw               1305 drivers/nvme/host/core.c 	c.rw.nsid = cpu_to_le32(ns->head->ns_id);
rw               1306 drivers/nvme/host/core.c 	c.rw.slba = cpu_to_le64(io.slba);
rw               1307 drivers/nvme/host/core.c 	c.rw.length = cpu_to_le16(io.nblocks);
rw               1308 drivers/nvme/host/core.c 	c.rw.control = cpu_to_le16(io.control);
rw               1309 drivers/nvme/host/core.c 	c.rw.dsmgmt = cpu_to_le32(io.dsmgmt);
rw               1310 drivers/nvme/host/core.c 	c.rw.reftag = cpu_to_le32(io.reftag);
rw               1311 drivers/nvme/host/core.c 	c.rw.apptag = cpu_to_le16(io.apptag);
rw               1312 drivers/nvme/host/core.c 	c.rw.appmask = cpu_to_le16(io.appmask);
rw               2260 drivers/nvme/host/fc.c 	sqe->rw.dptr.sgl.type = (NVME_TRANSPORT_SGL_DATA_DESC << 4) |
rw               2262 drivers/nvme/host/fc.c 	sqe->rw.dptr.sgl.length = cpu_to_le32(data_len);
rw               2263 drivers/nvme/host/fc.c 	sqe->rw.dptr.sgl.addr = 0;
rw                808 drivers/nvme/host/pci.c 							     &cmnd->rw, &bv);
rw                813 drivers/nvme/host/pci.c 							     &cmnd->rw, &bv);
rw                837 drivers/nvme/host/pci.c 		ret = nvme_pci_setup_sgls(dev, req, &cmnd->rw, nr_mapped);
rw                839 drivers/nvme/host/pci.c 		ret = nvme_pci_setup_prps(dev, req, &cmnd->rw);
rw                855 drivers/nvme/host/pci.c 	cmnd->rw.metadata = cpu_to_le64(iod->meta_dma);
rw                601 drivers/nvme/target/admin-cmd.c 	req->ns = nvmet_find_namespace(req->sq->ctrl, req->cmd->rw.nsid);
rw                835 drivers/nvme/target/core.c 	req->ns = nvmet_find_namespace(req->sq->ctrl, cmd->rw.nsid);
rw                124 drivers/nvme/target/io-cmd-bdev.c 		req->error_slba = le64_to_cpu(req->cmd->rw.slba);
rw                158 drivers/nvme/target/io-cmd-bdev.c 	if (req->cmd->rw.opcode == nvme_cmd_write) {
rw                161 drivers/nvme/target/io-cmd-bdev.c 		if (req->cmd->rw.control & cpu_to_le16(NVME_RW_FUA))
rw                170 drivers/nvme/target/io-cmd-bdev.c 	sector = le64_to_cpu(req->cmd->rw.slba);
rw                 96 drivers/nvme/target/io-cmd-file.c 	int rw;
rw                 98 drivers/nvme/target/io-cmd-file.c 	if (req->cmd->rw.opcode == nvme_cmd_write) {
rw                 99 drivers/nvme/target/io-cmd-file.c 		if (req->cmd->rw.control & cpu_to_le16(NVME_RW_FUA))
rw                102 drivers/nvme/target/io-cmd-file.c 		rw = WRITE;
rw                105 drivers/nvme/target/io-cmd-file.c 		rw = READ;
rw                108 drivers/nvme/target/io-cmd-file.c 	iov_iter_bvec(&iter, rw, req->f.bvec, nr_segs, count);
rw                148 drivers/nvme/target/io-cmd-file.c 	pos = le64_to_cpu(req->cmd->rw.slba) << req->ns->blksize_shift;
rw                494 drivers/nvme/target/nvmet.h 	return ((u32)le16_to_cpu(req->cmd->rw.length) + 1) <<
rw                 57 drivers/nvme/target/rdma.c 	struct rdma_rw_ctx	rw;
rw                506 drivers/nvme/target/rdma.c 		rdma_rw_ctx_destroy(&rsp->rw, queue->cm_id->qp,
rw                565 drivers/nvme/target/rdma.c 		first_wr = rdma_rw_ctx_wrs(&rsp->rw, cm_id->qp,
rw                590 drivers/nvme/target/rdma.c 	rdma_rw_ctx_destroy(&rsp->rw, queue->cm_id->qp,
rw                678 drivers/nvme/target/rdma.c 	ret = rdma_rw_ctx_init(&rsp->rw, cm_id->qp, cm_id->port_num,
rw                745 drivers/nvme/target/rdma.c 		if (rdma_rw_ctx_post(&rsp->rw, queue->cm_id->qp,
rw                 56 drivers/nvme/target/trace.h 		ns = nvmet_find_namespace(ctrl, req->cmd->rw.nsid);
rw                 32 drivers/pci/pci-bridge-emul.c 	u32 rw;
rw                 44 drivers/pci/pci-bridge-emul.c 		.rw = (PCI_COMMAND_IO | PCI_COMMAND_MEMORY |
rw                 89 drivers/pci/pci-bridge-emul.c 		.rw = GENMASK(24, 0),
rw                 96 drivers/pci/pci-bridge-emul.c 		.rw = (GENMASK(15, 12) | GENMASK(7, 4)),
rw                115 drivers/pci/pci-bridge-emul.c 		.rw = GENMASK(31, 20) | GENMASK(15, 4),
rw                123 drivers/pci/pci-bridge-emul.c 		.rw = GENMASK(31, 20) | GENMASK(15, 4),
rw                130 drivers/pci/pci-bridge-emul.c 		.rw = ~0,
rw                134 drivers/pci/pci-bridge-emul.c 		.rw = ~0,
rw                138 drivers/pci/pci-bridge-emul.c 		.rw = ~0,
rw                147 drivers/pci/pci-bridge-emul.c 		.rw = GENMASK(31, 11) | BIT(0),
rw                158 drivers/pci/pci-bridge-emul.c 		.rw = (GENMASK(7, 0) |
rw                191 drivers/pci/pci-bridge-emul.c 		.rw = GENMASK(15, 0),
rw                215 drivers/pci/pci-bridge-emul.c 		.rw = GENMASK(11, 3) | GENMASK(1, 0),
rw                233 drivers/pci/pci-bridge-emul.c 		.rw = GENMASK(12, 0),
rw                249 drivers/pci/pci-bridge-emul.c 		.rw = (PCI_EXP_RTCTL_SECEE | PCI_EXP_RTCTL_SENFEE |
rw                301 drivers/pci/pci-bridge-emul.c 		bridge->pci_regs_behavior[PCI_PREF_MEMORY_BASE / 4].rw = 0;
rw                426 drivers/pci/pci-bridge-emul.c 	new = old & (~mask | ~behavior[reg / 4].rw);
rw                429 drivers/pci/pci-bridge-emul.c 	new |= (value << shift) & (behavior[reg / 4].rw & mask);
rw                733 drivers/pinctrl/sh-pfc/core.c 	unsigned int i, n, rw, fw;
rw                740 drivers/pinctrl/sh-pfc/core.c 	for (i = 0, n = 0, rw = 0; (fw = cfg_reg->var_field_width[i]); i++) {
rw                743 drivers/pinctrl/sh-pfc/core.c 				drvname, cfg_reg->reg, rw, rw + fw - 1);
rw                747 drivers/pinctrl/sh-pfc/core.c 		rw += fw;
rw                750 drivers/pinctrl/sh-pfc/core.c 	if (rw != cfg_reg->reg_width) {
rw                752 drivers/pinctrl/sh-pfc/core.c 		       drvname, cfg_reg->reg, rw, cfg_reg->reg_width);
rw                160 drivers/regulator/tps6586x-regulator.c 	TPS6586X_REGULATOR(_id, rw, _pname, vdata, vreg, shift, nbits,	\
rw                 88 drivers/s390/block/dasd_fba.c define_extent(struct ccw1 * ccw, struct DE_fba_data *data, int rw,
rw                 96 drivers/s390/block/dasd_fba.c 	if (rw == WRITE)
rw                 98 drivers/s390/block/dasd_fba.c 	else if (rw == READ)
rw                108 drivers/s390/block/dasd_fba.c locate_record(struct ccw1 * ccw, struct LO_fba_data *data, int rw,
rw                116 drivers/s390/block/dasd_fba.c 	if (rw == WRITE)
rw                118 drivers/s390/block/dasd_fba.c 	else if (rw == READ)
rw               1716 drivers/scsi/sg.c 	int rw = hp->dxfer_direction == SG_DXFER_TO_DEV ? WRITE : READ;
rw               1804 drivers/scsi/sg.c 		res = import_iovec(rw, hp->dxferp, iov_count, 0, &iov, &i);
rw                845 drivers/scsi/st.c 	if (STps->rw == ST_WRITING)	/* Writing */
rw                941 drivers/scsi/st.c 		STps->rw = ST_IDLE;
rw               1077 drivers/scsi/st.c 			STps->rw = ST_IDLE;
rw               1306 drivers/scsi/st.c 		STps->rw = ST_IDLE;
rw               1352 drivers/scsi/st.c 	if (STps->rw == ST_WRITING && !STp->pos_unknown) {
rw               1372 drivers/scsi/st.c 	if (STps->rw == ST_WRITING && !STp->pos_unknown) {
rw               1421 drivers/scsi/st.c 		if (!STm->sysv || STps->rw != ST_READING) {
rw               1659 drivers/scsi/st.c 	if (STps->rw == ST_READING) {
rw               1663 drivers/scsi/st.c 		STps->rw = ST_WRITING;
rw               1664 drivers/scsi/st.c 	} else if (STps->rw != ST_WRITING &&
rw               1720 drivers/scsi/st.c 	STps->rw = ST_WRITING;
rw               2124 drivers/scsi/st.c 	if (STps->rw == ST_WRITING) {
rw               2128 drivers/scsi/st.c 		STps->rw = ST_READING;
rw               2165 drivers/scsi/st.c 	STps->rw = ST_READING;
rw               3001 drivers/scsi/st.c 			STps->rw = ST_IDLE;  /* prevent automatic WEOF at close */
rw               3248 drivers/scsi/st.c 				STps->rw = ST_IDLE;
rw               3600 drivers/scsi/st.c 			if (STps->rw == ST_WRITING &&
rw               3611 drivers/scsi/st.c 				STps->rw = ST_IDLE;
rw               3637 drivers/scsi/st.c 			STps->rw = ST_IDLE;	/* Prevent automatic WEOF and fsf */
rw               3684 drivers/scsi/st.c 				STp->ps[i].rw = ST_IDLE;
rw               3698 drivers/scsi/st.c 				STp->ps[0].rw = ST_IDLE;
rw               3760 drivers/scsi/st.c 			if (STps->rw == ST_WRITING)
rw               3763 drivers/scsi/st.c 			else if (STps->rw == ST_READING)
rw               4360 drivers/scsi/st.c 		STps->rw = ST_IDLE;
rw               4899 drivers/scsi/st.c 			      size_t count, int rw)
rw               4926 drivers/scsi/st.c 	res = get_user_pages_fast(uaddr, nr_pages, rw == READ ? FOLL_WRITE : 0,
rw                 87 drivers/scsi/st.h 	unsigned char rw;
rw               2156 drivers/scsi/ufs/ufshcd.c 		u32 rw;
rw               2157 drivers/scsi/ufs/ufshcd.c 		rw = set & INTERRUPT_MASK_RW_VER_10;
rw               2158 drivers/scsi/ufs/ufshcd.c 		set = rw | ((set ^ intrs) & intrs);
rw               2176 drivers/scsi/ufs/ufshcd.c 		u32 rw;
rw               2177 drivers/scsi/ufs/ufshcd.c 		rw = (set & INTERRUPT_MASK_RW_VER_10) &
rw               2179 drivers/scsi/ufs/ufshcd.c 		set = rw | ((set & intrs) & ~INTERRUPT_MASK_RW_VER_10);
rw                204 drivers/spi/spi-omap2-mcspi.c 	u32 l, rw;
rw                209 drivers/spi/spi-omap2-mcspi.c 		rw = OMAP2_MCSPI_CHCONF_DMAR;
rw                211 drivers/spi/spi-omap2-mcspi.c 		rw = OMAP2_MCSPI_CHCONF_DMAW;
rw                214 drivers/spi/spi-omap2-mcspi.c 		l |= rw;
rw                216 drivers/spi/spi-omap2-mcspi.c 		l &= ~rw;
rw               3238 drivers/staging/exfat/exfat_super.c 	int rw;
rw               3240 drivers/staging/exfat/exfat_super.c 	rw = iov_iter_rw(iter);
rw               3242 drivers/staging/exfat/exfat_super.c 	if (rw == WRITE) {
rw               3248 drivers/staging/exfat/exfat_super.c 	if ((ret < 0) && (rw & WRITE))
rw                857 drivers/target/target_core_pscsi.c 	int rw = (cmd->data_direction == DMA_TO_DEVICE);
rw                893 drivers/target/target_core_pscsi.c 				if (rw)
rw                898 drivers/target/target_core_pscsi.c 					(rw) ? "rw" : "r", nr_vecs);
rw                425 drivers/tty/serial/sunsab.c 		writeb(up->cached_mode, &up->regs->rw.mode);
rw                426 drivers/tty/serial/sunsab.c 		writeb(up->cached_pvr, &up->regs->rw.pvr);
rw                430 drivers/tty/serial/sunsab.c 		tmp = readb(&up->regs->rw.ccr2);
rw                433 drivers/tty/serial/sunsab.c 		writeb(tmp, &up->regs->rw.ccr2);
rw                572 drivers/tty/serial/sunsab.c 	tmp = readb(&up->regs->rw.ccr0);
rw                574 drivers/tty/serial/sunsab.c 	writeb(tmp, &up->regs->rw.ccr0);
rw                611 drivers/tty/serial/sunsab.c 	up->cached_dafo = readb(&up->regs->rw.dafo);
rw                613 drivers/tty/serial/sunsab.c 	writeb(up->cached_dafo, &up->regs->rw.dafo);
rw                617 drivers/tty/serial/sunsab.c 	writeb(up->cached_mode, &up->regs->rw.mode);
rw                631 drivers/tty/serial/sunsab.c 	tmp = readb(&up->regs->rw.ccr0);
rw                633 drivers/tty/serial/sunsab.c 	writeb(tmp, &up->regs->rw.ccr0);
rw               1009 drivers/tty/serial/sunsab.c 	up->cached_mode = readb(&up->regs->rw.mode);
rw               1011 drivers/tty/serial/sunsab.c 	writeb(up->cached_mode, &up->regs->rw.mode);
rw               1013 drivers/tty/serial/sunsab.c 	writeb(up->cached_mode, &up->regs->rw.mode);
rw                116 drivers/tty/serial/sunsab.h 	__volatile__ struct sab82532_async_rw_regs	rw;
rw                 49 drivers/usb/host/isp1362.h #define ISP1362_REG(name, addr, width, rw) \
rw                 50 drivers/usb/host/isp1362.h static isp1362_reg_t ISP1362_REG_##name = ((addr) | (width) | (rw))
rw                 58 drivers/usb/host/isp1362.h #define ISP1362_REG(name, addr, width, rw) \
rw               1170 drivers/vfio/pci/vfio_pci.c 		return vdev->region[index].ops->rw(vdev, buf,
rw                 51 drivers/vfio/pci/vfio_pci_igd.c 	.rw		= vfio_pci_igd_rw,
rw                212 drivers/vfio/pci/vfio_pci_igd.c 	.rw		= vfio_pci_igd_cfg_rw,
rw                188 drivers/vfio/pci/vfio_pci_nvlink2.c 	.rw = vfio_pci_nvgpu_rw,
rw                385 drivers/vfio/pci/vfio_pci_nvlink2.c 	.rw = vfio_pci_npu2_rw,
rw                 55 drivers/vfio/pci/vfio_pci_private.h 	size_t	(*rw)(struct vfio_pci_device *vdev, char __user *buf,
rw                100 drivers/video/backlight/ili922x.c #define START_BYTE(id, rs, rw)	\
rw                101 drivers/video/backlight/ili922x.c 	(0x70 | (((id) & 0x01) << 2) | (((rs) & 0x01) << 1) | ((rw) & 0x01))
rw                210 drivers/video/fbdev/core/bitblit.c 	unsigned int rw = info->var.xres - (vc->vc_cols*cw);
rw                212 drivers/video/fbdev/core/bitblit.c 	unsigned int rs = info->var.xres - rw;
rw                219 drivers/video/fbdev/core/bitblit.c 	if (rw && !bottom_only) {
rw                222 drivers/video/fbdev/core/bitblit.c 		region.width = rw;
rw                196 drivers/video/fbdev/core/fbcon_ccw.c 	unsigned int rw = info->var.yres - (vc->vc_cols*cw);
rw                204 drivers/video/fbdev/core/fbcon_ccw.c 	if (rw && !bottom_only) {
rw                207 drivers/video/fbdev/core/fbcon_ccw.c 		region.height = rw;
rw                179 drivers/video/fbdev/core/fbcon_cw.c 	unsigned int rw = info->var.yres - (vc->vc_cols*cw);
rw                181 drivers/video/fbdev/core/fbcon_cw.c 	unsigned int rs = info->var.yres - rw;
rw                187 drivers/video/fbdev/core/fbcon_cw.c 	if (rw && !bottom_only) {
rw                190 drivers/video/fbdev/core/fbcon_cw.c 		region.height = rw;
rw                227 drivers/video/fbdev/core/fbcon_ud.c 	unsigned int rw = info->var.xres - (vc->vc_cols*cw);
rw                234 drivers/video/fbdev/core/fbcon_ud.c 	if (rw && !bottom_only) {
rw                237 drivers/video/fbdev/core/fbcon_ud.c 		region.width  = rw;
rw                147 drivers/w1/w1.c static BIN_ATTR_RW(rw, PAGE_SIZE);
rw                201 fs/aio.c       		struct kiocb		rw;
rw                564 fs/aio.c       	struct aio_kiocb *req = container_of(iocb, struct aio_kiocb, rw);
rw                624 fs/aio.c       		req->ki_cancel(&req->rw);
rw               1424 fs/aio.c       	struct aio_kiocb *iocb = container_of(kiocb, struct aio_kiocb, rw);
rw               1481 fs/aio.c       static ssize_t aio_setup_rw(int rw, const struct iocb *iocb,
rw               1489 fs/aio.c       		ssize_t ret = import_single_range(rw, buf, len, *iovec, iter);
rw               1495 fs/aio.c       		return compat_import_iovec(rw, buf, len, UIO_FASTIOV, iovec,
rw               1498 fs/aio.c       	return import_iovec(rw, buf, len, UIO_FASTIOV, iovec, iter);
rw               1664 fs/aio.c       	struct aio_kiocb *aiocb = container_of(iocb, struct aio_kiocb, rw);
rw               1837 fs/aio.c       		return aio_read(&req->rw, iocb, false, compat);
rw               1839 fs/aio.c       		return aio_write(&req->rw, iocb, false, compat);
rw               1841 fs/aio.c       		return aio_read(&req->rw, iocb, true, compat);
rw               1843 fs/aio.c       		return aio_write(&req->rw, iocb, true, compat);
rw               2031 fs/aio.c       			ret = kiocb->ki_cancel(&kiocb->rw);
rw                 28 fs/btrfs/locking.h static inline void btrfs_tree_unlock_rw(struct extent_buffer *eb, int rw)
rw                 30 fs/btrfs/locking.h 	if (rw == BTRFS_WRITE_LOCK || rw == BTRFS_WRITE_LOCK_BLOCKING)
rw                 32 fs/btrfs/locking.h 	else if (rw == BTRFS_READ_LOCK_BLOCKING)
rw                 34 fs/btrfs/locking.h 	else if (rw == BTRFS_READ_LOCK)
rw               4358 fs/btrfs/relocation.c 	int rw = 0;
rw               4384 fs/btrfs/relocation.c 	rw = 1;
rw               4460 fs/btrfs/relocation.c 	if (err && rw)
rw                579 fs/cifs/cifsproto.h int setup_aio_ctx_iter(struct cifs_aio_ctx *ctx, struct iov_iter *iter, int rw);
rw                839 fs/cifs/misc.c setup_aio_ctx_iter(struct cifs_aio_ctx *ctx, struct iov_iter *iter, int rw)
rw                924 fs/cifs/misc.c 	iov_iter_bvec(&ctx->iter, rw, ctx->bv, npages, ctx->len);
rw                152 fs/crypto/crypto.c int fscrypt_crypt_block(const struct inode *inode, fscrypt_direction_t rw,
rw                185 fs/crypto/crypto.c 	if (rw == FS_DECRYPT)
rw                192 fs/crypto/crypto.c 			    (rw == FS_DECRYPT ? "De" : "En"), lblk_num, res);
rw                236 fs/crypto/fscrypt_private.h 			       fscrypt_direction_t rw, u64 lblk_num,
rw               2819 fs/f2fs/data.c 	int rw = iov_iter_rw(iter);
rw               2834 fs/f2fs/data.c 	trace_f2fs_direct_IO_enter(inode, offset, count, rw);
rw               2836 fs/f2fs/data.c 	if (rw == WRITE && whint_mode == WHINT_MODE_OFF)
rw               2840 fs/f2fs/data.c 		if (!down_read_trylock(&fi->i_gc_rwsem[rw])) {
rw               2846 fs/f2fs/data.c 			up_read(&fi->i_gc_rwsem[rw]);
rw               2852 fs/f2fs/data.c 		down_read(&fi->i_gc_rwsem[rw]);
rw               2858 fs/f2fs/data.c 			iter, rw == WRITE ? get_data_block_dio_write :
rw               2865 fs/f2fs/data.c 	up_read(&fi->i_gc_rwsem[rw]);
rw               2867 fs/f2fs/data.c 	if (rw == WRITE) {
rw               2881 fs/f2fs/data.c 	trace_f2fs_direct_IO_exit(inode, offset, count, rw, err);
rw               1072 fs/f2fs/f2fs.h #define is_read_io(rw) ((rw) == READ)
rw               3708 fs/f2fs/f2fs.h 	int rw = iov_iter_rw(iter);
rw               3710 fs/f2fs/f2fs.h 	return (test_opt(sbi, LFS) && (rw == WRITE) &&
rw               3718 fs/f2fs/f2fs.h 	int rw = iov_iter_rw(iter);
rw               3730 fs/f2fs/f2fs.h 	if (test_opt(sbi, LFS) && (rw == WRITE)) {
rw                308 fs/io_uring.c  		struct kiocb		rw;
rw                495 fs/io_uring.c  	int rw = 0;
rw                501 fs/io_uring.c  			rw = !(req->rw.ki_flags & IOCB_DIRECT);
rw                506 fs/io_uring.c  	queue_work(ctx->sqo_wq[rw], &req->work);
rw                817 fs/io_uring.c  		struct kiocb *kiocb = &req->rw;
rw                957 fs/io_uring.c  	struct io_kiocb *req = container_of(kiocb, struct io_kiocb, rw);
rw                970 fs/io_uring.c  	struct io_kiocb *req = container_of(kiocb, struct io_kiocb, rw);
rw               1004 fs/io_uring.c  		if (list_req->rw.ki_filp != req->rw.ki_filp)
rw               1080 fs/io_uring.c  	struct kiocb *kiocb = &req->rw;
rw               1165 fs/io_uring.c  static int io_import_fixed(struct io_ring_ctx *ctx, int rw,
rw               1199 fs/io_uring.c  	iov_iter_bvec(iter, rw, imu->bvec, imu->nr_bvecs, offset + len);
rw               1239 fs/io_uring.c  static ssize_t io_import_iovec(struct io_ring_ctx *ctx, int rw,
rw               1259 fs/io_uring.c  		ssize_t ret = io_import_fixed(ctx, rw, sqe, iter);
rw               1269 fs/io_uring.c  		return compat_import_iovec(rw, buf, sqe_len, UIO_FASTIOV,
rw               1273 fs/io_uring.c  	return import_iovec(rw, buf, sqe_len, UIO_FASTIOV, iovec, iter);
rw               1304 fs/io_uring.c  static void io_async_list_note(int rw, struct io_kiocb *req, size_t len)
rw               1306 fs/io_uring.c  	struct async_list *async_list = &req->ctx->pending_async[rw];
rw               1307 fs/io_uring.c  	struct kiocb *kiocb = &req->rw;
rw               1339 fs/io_uring.c  static ssize_t loop_rw_iter(int rw, struct file *file, struct kiocb *kiocb,
rw               1368 fs/io_uring.c  		if (rw == READ) {
rw               1397 fs/io_uring.c  	struct kiocb *kiocb = &req->rw;
rw               1462 fs/io_uring.c  	struct kiocb *kiocb = &req->rw;
rw               1592 fs/io_uring.c  	ret = vfs_fsync_range(req->rw.ki_filp, sqe_off,
rw               1640 fs/io_uring.c  	ret = sync_file_range(req->rw.ki_filp, sqe_off, sqe_len, flags);
rw               2211 fs/io_uring.c  		req->rw.ki_flags &= ~IOCB_NOWAIT;
rw                146 fs/nfs/blocklayout/blocklayout.c do_add_page_to_bio(struct bio *bio, int npg, int rw, sector_t isect,
rw                156 fs/nfs/blocklayout/blocklayout.c 		npg, rw, (unsigned long long)isect, offset, *len);
rw                183 fs/nfs/blocklayout/blocklayout.c 		bio_set_op_attrs(bio, rw, 0);
rw                192 fs/nfs/blocklayout/blocklayout.c static void bl_mark_devices_unavailable(struct nfs_pgio_header *header, bool rw)
rw                203 fs/nfs/blocklayout/blocklayout.c 		if (!ext_tree_lookup(bl, isect, &be, rw))
rw                182 fs/nfs/blocklayout/blocklayout.h int ext_tree_remove(struct pnfs_block_layout *bl, bool rw, sector_t start,
rw                187 fs/nfs/blocklayout/blocklayout.h 		struct pnfs_block_extent *ret, bool rw);
rw                346 fs/nfs/blocklayout/extent_tree.c 	    struct pnfs_block_extent *ret, bool rw)
rw                351 fs/nfs/blocklayout/extent_tree.c 	if (!rw)
rw                360 fs/nfs/blocklayout/extent_tree.c int ext_tree_remove(struct pnfs_block_layout *bl, bool rw,
rw                368 fs/nfs/blocklayout/extent_tree.c 	if (rw) {
rw                416 fs/ocfs2/refcounttree.c 				      struct ocfs2_refcount_tree *tree, int rw)
rw                420 fs/ocfs2/refcounttree.c 	ret = ocfs2_refcount_lock(tree, rw);
rw                426 fs/ocfs2/refcounttree.c 	if (rw)
rw                444 fs/ocfs2/refcounttree.c 			     u64 ref_blkno, int rw,
rw                462 fs/ocfs2/refcounttree.c 	ret = __ocfs2_lock_refcount_tree(osb, tree, rw);
rw                473 fs/ocfs2/refcounttree.c 		ocfs2_unlock_refcount_tree(osb, tree, rw);
rw                494 fs/ocfs2/refcounttree.c 		ocfs2_unlock_refcount_tree(osb, tree, rw);
rw                517 fs/ocfs2/refcounttree.c 				struct ocfs2_refcount_tree *tree, int rw)
rw                519 fs/ocfs2/refcounttree.c 	if (rw)
rw                524 fs/ocfs2/refcounttree.c 	ocfs2_refcount_unlock(tree, rw);
rw                 29 fs/ocfs2/refcounttree.h int ocfs2_lock_refcount_tree(struct ocfs2_super *osb, u64 ref_blkno, int rw,
rw                 34 fs/ocfs2/refcounttree.h 				int rw);
rw                405 include/linux/genhd.h 			int rw);
rw                407 include/linux/genhd.h 			int rw);
rw                960 include/linux/libata.h 			unsigned char *buf, unsigned int buflen, int rw);
rw               1844 include/linux/libata.h 			unsigned char *buf, unsigned int buflen, int rw);
rw               1846 include/linux/libata.h 			unsigned char *buf, unsigned int buflen, int rw);
rw                152 include/linux/libnvdimm.h 			void *iobuf, u64 len, int rw);
rw               3247 include/linux/mlx5/mlx5_ifc.h 	u8         rw[0x1];
rw                108 include/linux/mmiotrace.h extern void mmio_trace_rw(struct mmiotrace_rw *rw);
rw                 54 include/linux/nd.h 			void *buf, size_t size, int rw, unsigned long flags);
rw               1215 include/linux/nvme.h 		struct nvme_rw_command rw;
rw                 55 include/net/llc_conn.h 	u8		    rw;			/* rx window size; max = 127 */
rw                354 include/net/llc_pdu.h 	u8 rw;		/* sender receive window */
rw                376 include/net/llc_pdu.h 	xid_info->rw	 = rx_window << 1;	/* size of receive window */
rw                401 include/net/llc_pdu.h 	xid_info->rw	 = rx_window << 1;
rw                387 include/trace/events/block.h 	TP_PROTO(struct request_queue *q, struct bio *bio, int rw),
rw                389 include/trace/events/block.h 	TP_ARGS(q, bio, rw),
rw                425 include/trace/events/block.h 	TP_PROTO(struct request_queue *q, struct bio *bio, int rw),
rw                427 include/trace/events/block.h 	TP_ARGS(q, bio, rw)
rw                443 include/trace/events/block.h 	TP_PROTO(struct request_queue *q, struct bio *bio, int rw),
rw                445 include/trace/events/block.h 	TP_ARGS(q, bio, rw)
rw               1310 include/trace/events/ext4.h 	TP_PROTO(struct inode *inode, loff_t offset, unsigned long len, int rw),
rw               1312 include/trace/events/ext4.h 	TP_ARGS(inode, offset, len, rw),
rw               1319 include/trace/events/ext4.h 		__field(	int,	rw			)
rw               1327 include/trace/events/ext4.h 		__entry->rw	= rw;
rw               1333 include/trace/events/ext4.h 		  __entry->pos, __entry->len, __entry->rw)
rw               1338 include/trace/events/ext4.h 		 int rw, int ret),
rw               1340 include/trace/events/ext4.h 	TP_ARGS(inode, offset, len, rw, ret),
rw               1347 include/trace/events/ext4.h 		__field(	int,	rw			)
rw               1356 include/trace/events/ext4.h 		__entry->rw	= rw;
rw               1364 include/trace/events/ext4.h 		  __entry->rw, __entry->ret)
rw                914 include/trace/events/f2fs.h 	TP_PROTO(struct inode *inode, loff_t offset, unsigned long len, int rw),
rw                916 include/trace/events/f2fs.h 	TP_ARGS(inode, offset, len, rw),
rw                923 include/trace/events/f2fs.h 		__field(int,	rw)
rw                931 include/trace/events/f2fs.h 		__entry->rw	= rw;
rw                938 include/trace/events/f2fs.h 		__entry->rw)
rw                944 include/trace/events/f2fs.h 		 int rw, int ret),
rw                946 include/trace/events/f2fs.h 	TP_ARGS(inode, offset, len, rw, ret),
rw                953 include/trace/events/f2fs.h 		__field(int,	rw)
rw                962 include/trace/events/f2fs.h 		__entry->rw	= rw;
rw                971 include/trace/events/f2fs.h 		__entry->rw,
rw                248 include/xen/interface/io/blkif.h 		struct blkif_request_rw rw;
rw               1020 include/xen/interface/io/sndif.h 		struct xensnd_rw_req rw;
rw                572 kernel/futex.c get_futex_key(u32 __user *uaddr, int fshared, union futex_key *key, enum futex_access rw)
rw                618 kernel/futex.c 	if (err == -EFAULT && rw == FUTEX_READ) {
rw                206 kernel/trace/blktrace.c #define MASK_TC_BIT(rw, __name) ((rw & REQ_ ## __name) << \
rw                922 kernel/trace/blktrace.c 				struct bio *bio, int rw)
rw                932 kernel/trace/blktrace.c 			__blk_add_trace(bt, 0, 0, rw, 0, BLK_TA_GETRQ, 0, 0,
rw                941 kernel/trace/blktrace.c 				  struct bio *bio, int rw)
rw                951 kernel/trace/blktrace.c 			__blk_add_trace(bt, 0, 0, rw, 0, BLK_TA_SLEEPRQ,
rw                282 kernel/trace/trace_entries.h 		__field_struct(	struct mmiotrace_rw,	rw	)
rw                283 kernel/trace/trace_entries.h 		__field_desc(	resource_size_t, rw,	phys	)
rw                284 kernel/trace/trace_entries.h 		__field_desc(	unsigned long,	rw,	value	)
rw                285 kernel/trace/trace_entries.h 		__field_desc(	unsigned long,	rw,	pc	)
rw                286 kernel/trace/trace_entries.h 		__field_desc(	int, 		rw,	map_id	)
rw                287 kernel/trace/trace_entries.h 		__field_desc(	unsigned char,	rw,	opcode	)
rw                288 kernel/trace/trace_entries.h 		__field_desc(	unsigned char,	rw,	width	)
rw                171 kernel/trace/trace_mmiotrace.c 	struct mmiotrace_rw *rw;
rw                178 kernel/trace/trace_mmiotrace.c 	rw = &field->rw;
rw                180 kernel/trace/trace_mmiotrace.c 	switch (rw->opcode) {
rw                184 kernel/trace/trace_mmiotrace.c 			rw->width, secs, usec_rem, rw->map_id,
rw                185 kernel/trace/trace_mmiotrace.c 			(unsigned long long)rw->phys,
rw                186 kernel/trace/trace_mmiotrace.c 			rw->value, rw->pc, 0);
rw                191 kernel/trace/trace_mmiotrace.c 			rw->width, secs, usec_rem, rw->map_id,
rw                192 kernel/trace/trace_mmiotrace.c 			(unsigned long long)rw->phys,
rw                193 kernel/trace/trace_mmiotrace.c 			rw->value, rw->pc, 0);
rw                199 kernel/trace/trace_mmiotrace.c 			secs, usec_rem, rw->map_id,
rw                200 kernel/trace/trace_mmiotrace.c 			(unsigned long long)rw->phys,
rw                201 kernel/trace/trace_mmiotrace.c 			(rw->value >> 16) & 0xff, (rw->value >> 8) & 0xff,
rw                202 kernel/trace/trace_mmiotrace.c 			(rw->value >> 0) & 0xff, rw->pc, 0);
rw                297 kernel/trace/trace_mmiotrace.c 				struct mmiotrace_rw *rw)
rw                312 kernel/trace/trace_mmiotrace.c 	entry->rw			= *rw;
rw                318 kernel/trace/trace_mmiotrace.c void mmio_trace_rw(struct mmiotrace_rw *rw)
rw                322 kernel/trace/trace_mmiotrace.c 	__trace_mmiotrace_rw(tr, data, rw);
rw               1683 lib/iov_iter.c int import_single_range(int rw, void __user *buf, size_t len,
rw               1693 lib/iov_iter.c 	iov_iter_init(i, rw, iov, 1, len);
rw               1106 net/llc/af_llc.c 		llc->rw = opt;
rw               1166 net/llc/af_llc.c 		val = llc->rw;				break;
rw                 59 net/llc/llc_c_ev.c static u16 llc_util_ns_inside_rx_window(u8 ns, u8 vr, u8 rw)
rw                 62 net/llc/llc_c_ev.c 				     (vr + rw - 1) % LLC_2_SEQ_NBR_MODULO);
rw                205 net/llc/llc_c_ev.c 	       !llc_util_ns_inside_rx_window(ns, vr, llc_sk(sk)->rw) ? 0 : 1;
rw                217 net/llc/llc_c_ev.c 	       !llc_util_ns_inside_rx_window(ns, vr, llc_sk(sk)->rw) ? 0 : 1;
rw                228 net/llc/llc_c_ev.c 		 llc_util_ns_inside_rx_window(ns, vr, llc_sk(sk)->rw) ? 0 : 1;
rw                272 net/llc/llc_c_ev.c 	       !llc_util_ns_inside_rx_window(ns, vr, llc_sk(sk)->rw) ? 0 : 1;
rw                284 net/llc/llc_c_ev.c 	       !llc_util_ns_inside_rx_window(ns, vr, llc_sk(sk)->rw) ? 0 : 1;
rw                295 net/llc/llc_c_ev.c 	       !llc_util_ns_inside_rx_window(ns, vr, llc_sk(sk)->rw) ? 0 : 1;
rw                306 net/llc/llc_c_ev.c 		 llc_util_ns_inside_rx_window(ns, vr, llc_sk(sk)->rw) ? 0 : 1;
rw                901 net/llc/llc_conn.c 	llc->rw = 128; /* rx win size (opt and equal to
rw                192 net/llc/llc_proc.c 		   llc->retry_count, llc->k, llc->rw, llc->p_flag, llc->f_flag,
rw                 70 net/nfc/llcp.h 	u8 rw;
rw                115 net/nfc/llcp.h 	u8 rw;
rw                394 net/nfc/llcp_commands.c 	u8 *rw_tlv = NULL, rw_tlv_length, rw;
rw                420 net/nfc/llcp_commands.c 	rw = sock->rw > LLCP_MAX_RW ? local->rw : sock->rw;
rw                430 net/nfc/llcp_commands.c 	rw_tlv = nfc_llcp_build_tlv(LLCP_TLV_RW, &rw, 0, &rw_tlv_length);
rw                469 net/nfc/llcp_commands.c 	u8 *rw_tlv = NULL, rw_tlv_length, rw;
rw                483 net/nfc/llcp_commands.c 	rw = sock->rw > LLCP_MAX_RW ? local->rw : sock->rw;
rw                493 net/nfc/llcp_commands.c 	rw_tlv = nfc_llcp_build_tlv(LLCP_TLV_RW, &rw, 0, &rw_tlv_length);
rw                951 net/nfc/llcp_core.c 	new_sock->rw = sock->rw;
rw               1595 net/nfc/llcp_core.c 	local->rw = LLCP_MAX_RW;
rw                254 net/nfc/llcp_sock.c 		llcp_sock->rw = (u8) opt;
rw                288 net/nfc/llcp_sock.c 		 llcp_sock->rw, llcp_sock->miux);
rw                301 net/nfc/llcp_sock.c 	u8 rw;
rw                321 net/nfc/llcp_sock.c 		rw = llcp_sock->rw > LLCP_MAX_RW ? local->rw : llcp_sock->rw;
rw                322 net/nfc/llcp_sock.c 		if (put_user(rw, (u32 __user *) optval))
rw                969 net/nfc/llcp_sock.c 	llcp_sock->rw = LLCP_MAX_RW + 1;
rw               1006 net/nfc/netlink.c 	    nla_put_u8(msg, NFC_ATTR_LLC_PARAM_RW, local->rw) ||
rw               1071 net/nfc/netlink.c 	u8 rw = 0;
rw               1083 net/nfc/netlink.c 		rw = nla_get_u8(info->attrs[NFC_ATTR_LLC_PARAM_RW]);
rw               1085 net/nfc/netlink.c 		if (rw > LLCP_MAX_RW)
rw               1120 net/nfc/netlink.c 		local->rw = rw;
rw                145 sound/xen/xen_snd_front.c 	req->op.rw.length = count;
rw                146 sound/xen/xen_snd_front.c 	req->op.rw.offset = pos;
rw                168 sound/xen/xen_snd_front.c 	req->op.rw.length = count;
rw                169 sound/xen/xen_snd_front.c 	req->op.rw.offset = pos;
rw               2645 tools/testing/nvdimm/test/nfit.c 		void *iobuf, u64 len, int rw)
rw               2653 tools/testing/nvdimm/test/nfit.c 	if (rw)