rm                465 arch/arm/net/bpf_jit_32.c static inline void emit_udivmod(u8 rd, u8 rm, u8 rn, struct jit_ctx *ctx, u8 op)
rm                472 arch/arm/net/bpf_jit_32.c 			emit(ARM_UDIV(rd, rm, rn), ctx);
rm                474 arch/arm/net/bpf_jit_32.c 			emit(ARM_UDIV(ARM_IP, rm, rn), ctx);
rm                475 arch/arm/net/bpf_jit_32.c 			emit(ARM_MLS(rd, rn, ARM_IP, rm), ctx);
rm                493 arch/arm/net/bpf_jit_32.c 	if (rm != ARM_R0) {
rm                495 arch/arm/net/bpf_jit_32.c 		emit(ARM_MOV_R(ARM_R0, rm), ctx);
rm                510 arch/arm/net/bpf_jit_32.c 	if (rm != ARM_R0)
rm               1063 arch/arm/net/bpf_jit_32.c 	s8 rm = src;
rm               1068 arch/arm/net/bpf_jit_32.c 		rm = tmp[0];
rm               1070 arch/arm/net/bpf_jit_32.c 	} else if (rd[1] == rm) {
rm               1071 arch/arm/net/bpf_jit_32.c 		emit(ARM_MOV_R(tmp[0], rm), ctx);
rm               1072 arch/arm/net/bpf_jit_32.c 		rm = tmp[0];
rm               1077 arch/arm/net/bpf_jit_32.c 		emit(ARM_LDRB_I(rd[1], rm, off), ctx);
rm               1083 arch/arm/net/bpf_jit_32.c 		emit(ARM_LDRH_I(rd[1], rm, off), ctx);
rm               1089 arch/arm/net/bpf_jit_32.c 		emit(ARM_LDR_I(rd[1], rm, off), ctx);
rm               1095 arch/arm/net/bpf_jit_32.c 		emit(ARM_LDR_I(rd[1], rm, off), ctx);
rm               1096 arch/arm/net/bpf_jit_32.c 		emit(ARM_LDR_I(rd[0], rm, off + 4), ctx);
rm               1103 arch/arm/net/bpf_jit_32.c static inline void emit_ar_r(const u8 rd, const u8 rt, const u8 rm,
rm               1110 arch/arm/net/bpf_jit_32.c 			emit(ARM_AND_R(ARM_LR, rd, rm), ctx);
rm               1123 arch/arm/net/bpf_jit_32.c 			emit(ARM_CMP_R(rd, rm), ctx);
rm               1134 arch/arm/net/bpf_jit_32.c 			emit(ARM_SBCS_R(ARM_IP, rm, rd), ctx);
rm               1140 arch/arm/net/bpf_jit_32.c 			emit(ARM_SBCS_R(ARM_IP, rd, rm), ctx);
rm               1357 arch/arm/net/bpf_jit_32.c 	s8 rd_lo, rt, rm, rn;
rm               1670 arch/arm/net/bpf_jit_32.c 		rm = arm_bpf_get_reg32(src_hi, tmp2[0], ctx);
rm               1708 arch/arm/net/bpf_jit_32.c 		rm = tmp2[0];
rm               1717 arch/arm/net/bpf_jit_32.c 		emit_ar_r(rd[0], rd[1], rm, rn, ctx, BPF_OP(code),
rm                159 arch/arm/net/bpf_jit_32.h #define _AL3_R(op, rd, rn, rm)	((op ## _R) | (rd) << 12 | (rn) << 16 | (rm))
rm                165 arch/arm/net/bpf_jit_32.h #define ARM_ADD_R(rd, rn, rm)	_AL3_R(ARM_INST_ADD, rd, rn, rm)
rm                166 arch/arm/net/bpf_jit_32.h #define ARM_ADDS_R(rd, rn, rm)	_AL3_R(ARM_INST_ADDS, rd, rn, rm)
rm                169 arch/arm/net/bpf_jit_32.h #define ARM_ADC_R(rd, rn, rm)	_AL3_R(ARM_INST_ADC, rd, rn, rm)
rm                172 arch/arm/net/bpf_jit_32.h #define ARM_AND_R(rd, rn, rm)	_AL3_R(ARM_INST_AND, rd, rn, rm)
rm                173 arch/arm/net/bpf_jit_32.h #define ARM_ANDS_R(rd, rn, rm)	_AL3_R(ARM_INST_ANDS, rd, rn, rm)
rm                176 arch/arm/net/bpf_jit_32.h #define ARM_BIC_R(rd, rn, rm)	_AL3_R(ARM_INST_BIC, rd, rn, rm)
rm                180 arch/arm/net/bpf_jit_32.h #define ARM_BX(rm)		(ARM_INST_BX | (rm))
rm                181 arch/arm/net/bpf_jit_32.h #define ARM_BLX_R(rm)		(ARM_INST_BLX_R | (rm))
rm                183 arch/arm/net/bpf_jit_32.h #define ARM_CMP_R(rn, rm)	_AL3_R(ARM_INST_CMP, 0, rn, rm)
rm                186 arch/arm/net/bpf_jit_32.h #define ARM_EOR_R(rd, rn, rm)	_AL3_R(ARM_INST_EOR, rd, rn, rm)
rm                189 arch/arm/net/bpf_jit_32.h #define ARM_LDR_R(rt, rn, rm)	(ARM_INST_LDR_R | ARM_INST_LDST__U \
rm                191 arch/arm/net/bpf_jit_32.h 				 | (rm))
rm                192 arch/arm/net/bpf_jit_32.h #define ARM_LDR_R_SI(rt, rn, rm, type, imm) \
rm                195 arch/arm/net/bpf_jit_32.h 				 | (imm) << 7 | (type) << 5 | (rm))
rm                196 arch/arm/net/bpf_jit_32.h #define ARM_LDRB_R(rt, rn, rm)	(ARM_INST_LDRB_R | ARM_INST_LDST__U \
rm                198 arch/arm/net/bpf_jit_32.h 				 | (rm))
rm                199 arch/arm/net/bpf_jit_32.h #define ARM_LDRH_R(rt, rn, rm)	(ARM_INST_LDRH_R | ARM_INST_LDST__U \
rm                201 arch/arm/net/bpf_jit_32.h 				 | (rm))
rm                206 arch/arm/net/bpf_jit_32.h #define ARM_LSL_R(rd, rn, rm)	(_AL3_R(ARM_INST_LSL, rd, 0, rn) | (rm) << 8)
rm                209 arch/arm/net/bpf_jit_32.h #define ARM_LSR_R(rd, rn, rm)	(_AL3_R(ARM_INST_LSR, rd, 0, rn) | (rm) << 8)
rm                211 arch/arm/net/bpf_jit_32.h #define ARM_ASR_R(rd, rn, rm)   (_AL3_R(ARM_INST_ASR, rd, 0, rn) | (rm) << 8)
rm                214 arch/arm/net/bpf_jit_32.h #define ARM_MOV_R(rd, rm)	_AL3_R(ARM_INST_MOV, rd, 0, rm)
rm                215 arch/arm/net/bpf_jit_32.h #define ARM_MOVS_R(rd, rm)	_AL3_R(ARM_INST_MOVS, rd, 0, rm)
rm                217 arch/arm/net/bpf_jit_32.h #define ARM_MOV_SR(rd, rm, type, rs)	\
rm                218 arch/arm/net/bpf_jit_32.h 	(_AL3_SR(ARM_MOV_R(rd, rm)) | (type) << 5 | (rs) << 8)
rm                219 arch/arm/net/bpf_jit_32.h #define ARM_MOV_SI(rd, rm, type, imm6)	\
rm                220 arch/arm/net/bpf_jit_32.h 	(ARM_MOV_R(rd, rm) | (type) << 5 | (imm6) << 7)
rm                228 arch/arm/net/bpf_jit_32.h #define ARM_MUL(rd, rm, rn)	(ARM_INST_MUL | (rd) << 16 | (rm) << 8 | (rn))
rm                233 arch/arm/net/bpf_jit_32.h #define ARM_ORR_R(rd, rn, rm)	_AL3_R(ARM_INST_ORR, rd, rn, rm)
rm                235 arch/arm/net/bpf_jit_32.h #define ARM_ORR_SR(rd, rn, rm, type, rs)	\
rm                236 arch/arm/net/bpf_jit_32.h 	(_AL3_SR(ARM_ORR_R(rd, rn, rm)) | (type) << 5 | (rs) << 8)
rm                237 arch/arm/net/bpf_jit_32.h #define ARM_ORRS_R(rd, rn, rm)	_AL3_R(ARM_INST_ORRS, rd, rn, rm)
rm                238 arch/arm/net/bpf_jit_32.h #define ARM_ORRS_SR(rd, rn, rm, type, rs)	\
rm                239 arch/arm/net/bpf_jit_32.h 	(_AL3_SR(ARM_ORRS_R(rd, rn, rm)) | (type) << 5 | (rs) << 8)
rm                240 arch/arm/net/bpf_jit_32.h #define ARM_ORR_SI(rd, rn, rm, type, imm6)	\
rm                241 arch/arm/net/bpf_jit_32.h 	(ARM_ORR_R(rd, rn, rm) | (type) << 5 | (imm6) << 7)
rm                242 arch/arm/net/bpf_jit_32.h #define ARM_ORRS_SI(rd, rn, rm, type, imm6)	\
rm                243 arch/arm/net/bpf_jit_32.h 	(ARM_ORRS_R(rd, rn, rm) | (type) << 5 | (imm6) << 7)
rm                245 arch/arm/net/bpf_jit_32.h #define ARM_REV(rd, rm)		(ARM_INST_REV | (rd) << 12 | (rm))
rm                246 arch/arm/net/bpf_jit_32.h #define ARM_REV16(rd, rm)	(ARM_INST_REV16 | (rd) << 12 | (rm))
rm                252 arch/arm/net/bpf_jit_32.h #define ARM_SUB_R(rd, rn, rm)	_AL3_R(ARM_INST_SUB, rd, rn, rm)
rm                253 arch/arm/net/bpf_jit_32.h #define ARM_SUBS_R(rd, rn, rm)	_AL3_R(ARM_INST_SUBS, rd, rn, rm)
rm                254 arch/arm/net/bpf_jit_32.h #define ARM_RSB_R(rd, rn, rm)	_AL3_R(ARM_INST_RSB, rd, rn, rm)
rm                255 arch/arm/net/bpf_jit_32.h #define ARM_SBC_R(rd, rn, rm)	_AL3_R(ARM_INST_SBC, rd, rn, rm)
rm                256 arch/arm/net/bpf_jit_32.h #define ARM_SBCS_R(rd, rn, rm)	_AL3_R(ARM_INST_SBCS, rd, rn, rm)
rm                261 arch/arm/net/bpf_jit_32.h #define ARM_TST_R(rn, rm)	_AL3_R(ARM_INST_TST, 0, rn, rm)
rm                264 arch/arm/net/bpf_jit_32.h #define ARM_UDIV(rd, rn, rm)	(ARM_INST_UDIV | (rd) << 16 | (rn) | (rm) << 8)
rm                266 arch/arm/net/bpf_jit_32.h #define ARM_UMULL(rd_lo, rd_hi, rn, rm)	(ARM_INST_UMULL | (rd_hi) << 16 \
rm                267 arch/arm/net/bpf_jit_32.h 					 | (rd_lo) << 12 | (rm) << 8 | rn)
rm                269 arch/arm/net/bpf_jit_32.h #define ARM_MLS(rd, rn, rm, ra)	(ARM_INST_MLS | (rd) << 16 | (rn) | (rm) << 8 \
rm                271 arch/arm/net/bpf_jit_32.h #define ARM_UXTH(rd, rm)	(ARM_INST_UXTH | (rd) << 12 | (rm))
rm                 79 arch/arm/probes/decode-arm.c 	int rm = insn & 0xf;
rm                 80 arch/arm/probes/decode-arm.c 	long rmv = regs->uregs[rm];
rm                 75 arch/arm/probes/kprobes/actions-arm.c 	int rm = insn & 0xf;
rm                 81 arch/arm/probes/kprobes/actions-arm.c 	register unsigned long rmv asm("r3") = regs->uregs[rm];
rm                104 arch/arm/probes/kprobes/actions-arm.c 	int rm = insn & 0xf;
rm                109 arch/arm/probes/kprobes/actions-arm.c 	register unsigned long rmv asm("r3") = regs->uregs[rm];
rm                135 arch/arm/probes/kprobes/actions-arm.c 	int rm = insn & 0xf;
rm                141 arch/arm/probes/kprobes/actions-arm.c 	register unsigned long rmv asm("r3") = regs->uregs[rm];
rm                161 arch/arm/probes/kprobes/actions-arm.c 	int rm = insn & 0xf;
rm                167 arch/arm/probes/kprobes/actions-arm.c 	register unsigned long rmv asm("r3") = (rm == 15) ? pc
rm                168 arch/arm/probes/kprobes/actions-arm.c 							  : regs->uregs[rm];
rm                195 arch/arm/probes/kprobes/actions-arm.c 	int rm = insn & 0xf;
rm                199 arch/arm/probes/kprobes/actions-arm.c 	register unsigned long rmv asm("r3") = regs->uregs[rm];
rm                223 arch/arm/probes/kprobes/actions-arm.c 	int rm = insn & 0xf;
rm                228 arch/arm/probes/kprobes/actions-arm.c 	register unsigned long rmv asm("r3") = regs->uregs[rm];
rm                251 arch/arm/probes/kprobes/actions-arm.c 	int rm = insn & 0xf;
rm                254 arch/arm/probes/kprobes/actions-arm.c 	register unsigned long rmv asm("r3") = regs->uregs[rm];
rm                274 arch/arm/probes/kprobes/actions-arm.c 	int rm = (insn >> 8) & 0xf;
rm                279 arch/arm/probes/kprobes/actions-arm.c 	register unsigned long rmv asm("r1") = regs->uregs[rm];
rm                 29 arch/arm/probes/kprobes/actions-thumb.c 	int rm = insn & 0xf;
rm                 32 arch/arm/probes/kprobes/actions-thumb.c 	unsigned long rmv = regs->uregs[rm];
rm                190 arch/arm/probes/kprobes/actions-thumb.c 	int rm = insn & 0xf;
rm                194 arch/arm/probes/kprobes/actions-thumb.c 	register unsigned long rmv asm("r3") = regs->uregs[rm];
rm                216 arch/arm/probes/kprobes/actions-thumb.c 	int rm = insn & 0xf;
rm                220 arch/arm/probes/kprobes/actions-thumb.c 	register unsigned long rmv asm("r3") = regs->uregs[rm];
rm                285 arch/arm/probes/kprobes/actions-thumb.c 	int rm = insn & 0xf;
rm                290 arch/arm/probes/kprobes/actions-thumb.c 	register unsigned long rmv asm("r3") = regs->uregs[rm];
rm                310 arch/arm/probes/kprobes/actions-thumb.c 	int rm = (insn >> 3) & 0xf;
rm                311 arch/arm/probes/kprobes/actions-thumb.c 	unsigned long rmv = (rm == 15) ? pc : regs->uregs[rm];
rm                486 arch/arm/probes/kprobes/actions-thumb.c 	int rm = (insn >> 3) & 0xf;
rm                493 arch/arm/probes/kprobes/actions-thumb.c 	rmv = (rm == 15) ? pc : regs->uregs[rm];
rm               1974 arch/mips/math-emu/cp1emu.c 			oldrm = ieee754_csr.rm;
rm               1976 arch/mips/math-emu/cp1emu.c 			ieee754_csr.rm = MIPSInst_FUNC(ir);
rm               1978 arch/mips/math-emu/cp1emu.c 			ieee754_csr.rm = oldrm;
rm               2020 arch/mips/math-emu/cp1emu.c 			oldrm = ieee754_csr.rm;
rm               2022 arch/mips/math-emu/cp1emu.c 			ieee754_csr.rm = MIPSInst_FUNC(ir);
rm               2024 arch/mips/math-emu/cp1emu.c 			ieee754_csr.rm = oldrm;
rm               2325 arch/mips/math-emu/cp1emu.c 			oldrm = ieee754_csr.rm;
rm               2327 arch/mips/math-emu/cp1emu.c 			ieee754_csr.rm = MIPSInst_FUNC(ir);
rm               2329 arch/mips/math-emu/cp1emu.c 			ieee754_csr.rm = oldrm;
rm               2371 arch/mips/math-emu/cp1emu.c 			oldrm = ieee754_csr.rm;
rm               2373 arch/mips/math-emu/cp1emu.c 			ieee754_csr.rm = MIPSInst_FUNC(ir);
rm               2375 arch/mips/math-emu/cp1emu.c 			ieee754_csr.rm = oldrm;
rm                 83 arch/mips/math-emu/dp_add.c 			return ieee754dp_zero(ieee754_csr.rm == FPU_CSR_RD);
rm                154 arch/mips/math-emu/dp_add.c 			return ieee754dp_zero(ieee754_csr.rm == FPU_CSR_RD);
rm                 14 arch/mips/math-emu/dp_div.c 	u64 rm;
rm                116 arch/mips/math-emu/dp_div.c 	rm = 0;
rm                122 arch/mips/math-emu/dp_div.c 			rm |= bm;
rm                129 arch/mips/math-emu/dp_div.c 	rm <<= 1;
rm                131 arch/mips/math-emu/dp_div.c 		rm |= 1;	/* have remainder, set sticky */
rm                133 arch/mips/math-emu/dp_div.c 	assert(rm);
rm                138 arch/mips/math-emu/dp_div.c 	while ((rm >> (DP_FBITS + 3)) == 0) {
rm                139 arch/mips/math-emu/dp_div.c 		rm <<= 1;
rm                143 arch/mips/math-emu/dp_div.c 	return ieee754dp_format(xs == ys ? 0 : 1, re, rm);
rm                150 arch/mips/math-emu/dp_maddf.c 			return ieee754dp_zero(ieee754_csr.rm == FPU_CSR_RD);
rm                291 arch/mips/math-emu/dp_maddf.c 			return ieee754dp_zero(ieee754_csr.rm == FPU_CSR_RD);
rm                 16 arch/mips/math-emu/dp_mul.c 	u64 rm;
rm                144 arch/mips/math-emu/dp_mul.c 	rm = hrm | (lrm != 0);
rm                149 arch/mips/math-emu/dp_mul.c 	if ((s64) rm < 0) {
rm                150 arch/mips/math-emu/dp_mul.c 		rm = (rm >> (64 - (DP_FBITS + 1 + 3))) |
rm                151 arch/mips/math-emu/dp_mul.c 		     ((rm << (DP_FBITS + 1 + 3)) != 0);
rm                154 arch/mips/math-emu/dp_mul.c 		rm = (rm >> (64 - (DP_FBITS + 1 + 3 + 1))) |
rm                155 arch/mips/math-emu/dp_mul.c 		     ((rm << (DP_FBITS + 1 + 3 + 1)) != 0);
rm                157 arch/mips/math-emu/dp_mul.c 	assert(rm & (DP_HIDDEN_BIT << 3));
rm                159 arch/mips/math-emu/dp_mul.c 	return ieee754dp_format(rs, re, rm);
rm                 54 arch/mips/math-emu/dp_rint.c 	switch (ieee754_csr.rm) {
rm                 22 arch/mips/math-emu/dp_simple.c 		oldrm = ieee754_csr.rm;
rm                 23 arch/mips/math-emu/dp_simple.c 		ieee754_csr.rm = FPU_CSR_RD;
rm                 25 arch/mips/math-emu/dp_simple.c 		ieee754_csr.rm = oldrm;
rm                 40 arch/mips/math-emu/dp_simple.c 		oldrm = ieee754_csr.rm;
rm                 41 arch/mips/math-emu/dp_simple.c 		ieee754_csr.rm = FPU_CSR_RD;
rm                 46 arch/mips/math-emu/dp_simple.c 		ieee754_csr.rm = oldrm;
rm                 70 arch/mips/math-emu/dp_sqrt.c 	ieee754_csr.rm = FPU_CSR_RN;
rm                114 arch/mips/math-emu/dp_sqrt.c 	ieee754_csr.rm = FPU_CSR_RZ;
rm                130 arch/mips/math-emu/dp_sqrt.c 		switch (oldcsr.rm) {
rm                 83 arch/mips/math-emu/dp_sub.c 			return ieee754dp_zero(ieee754_csr.rm == FPU_CSR_RD);
rm                158 arch/mips/math-emu/dp_sub.c 			if (ieee754_csr.rm == FPU_CSR_RD)
rm                 67 arch/mips/math-emu/dp_tint.c 		switch (ieee754_csr.rm) {
rm                 72 arch/mips/math-emu/dp_tlong.c 		switch (ieee754_csr.rm) {
rm                151 arch/mips/math-emu/ieee754.h 	__BITFIELD_FIELD(unsigned rm:2,		/* current rounding mode */
rm                158 arch/mips/math-emu/ieee754.h 	return (ieee754_csr.rm);
rm                161 arch/mips/math-emu/ieee754.h static inline unsigned int ieee754_setrm(unsigned int rm)
rm                163 arch/mips/math-emu/ieee754.h 	return (ieee754_csr.rm = rm);
rm                 61 arch/mips/math-emu/ieee754dp.c 		switch (ieee754_csr.rm) {
rm                102 arch/mips/math-emu/ieee754dp.c 			switch(ieee754_csr.rm) {
rm                166 arch/mips/math-emu/ieee754dp.c 		switch (ieee754_csr.rm) {
rm                 61 arch/mips/math-emu/ieee754sp.c 		switch (ieee754_csr.rm) {
rm                102 arch/mips/math-emu/ieee754sp.c 			switch(ieee754_csr.rm) {
rm                165 arch/mips/math-emu/ieee754sp.c 		switch (ieee754_csr.rm) {
rm                 83 arch/mips/math-emu/sp_add.c 			return ieee754sp_zero(ieee754_csr.rm == FPU_CSR_RD);
rm                153 arch/mips/math-emu/sp_add.c 			return ieee754sp_zero(ieee754_csr.rm == FPU_CSR_RD);
rm                 14 arch/mips/math-emu/sp_div.c 	unsigned int rm;
rm                116 arch/mips/math-emu/sp_div.c 	rm = 0;
rm                122 arch/mips/math-emu/sp_div.c 			rm |= bm;
rm                129 arch/mips/math-emu/sp_div.c 	rm <<= 1;
rm                131 arch/mips/math-emu/sp_div.c 		rm |= 1;	/* have remainder, set sticky */
rm                133 arch/mips/math-emu/sp_div.c 	assert(rm);
rm                137 arch/mips/math-emu/sp_div.c 	while ((rm >> (SP_FBITS + 3)) == 0) {
rm                138 arch/mips/math-emu/sp_div.c 		rm <<= 1;
rm                142 arch/mips/math-emu/sp_div.c 	return ieee754sp_format(xs == ys ? 0 : 1, re, rm);
rm                 22 arch/mips/math-emu/sp_fdp.c 	u32 rm;
rm                 58 arch/mips/math-emu/sp_fdp.c 		if ((ieee754_csr.rm == FPU_CSR_RU && !xs) ||
rm                 59 arch/mips/math-emu/sp_fdp.c 				(ieee754_csr.rm == FPU_CSR_RD && xs))
rm                 70 arch/mips/math-emu/sp_fdp.c 	rm = (xm >> (DP_FBITS - (SP_FBITS + 3))) |
rm                 73 arch/mips/math-emu/sp_fdp.c 	return ieee754sp_format(xs, xe, rm);
rm                 20 arch/mips/math-emu/sp_maddf.c 	unsigned int rm;
rm                119 arch/mips/math-emu/sp_maddf.c 			return ieee754sp_zero(ieee754_csr.rm == FPU_CSR_RD);
rm                185 arch/mips/math-emu/sp_maddf.c 		rm = XSPSRS64(rm64, (62 - 26));
rm                186 arch/mips/math-emu/sp_maddf.c 		return ieee754sp_format(rs, re, rm);
rm                231 arch/mips/math-emu/sp_maddf.c 			return ieee754sp_zero(ieee754_csr.rm == FPU_CSR_RD);
rm                 16 arch/mips/math-emu/sp_mul.c 	unsigned int rm;
rm                139 arch/mips/math-emu/sp_mul.c 	rm = hrm | (lrm != 0);
rm                144 arch/mips/math-emu/sp_mul.c 	if ((int) rm < 0) {
rm                145 arch/mips/math-emu/sp_mul.c 		rm = (rm >> (32 - (SP_FBITS + 1 + 3))) |
rm                146 arch/mips/math-emu/sp_mul.c 		    ((rm << (SP_FBITS + 1 + 3)) != 0);
rm                149 arch/mips/math-emu/sp_mul.c 		rm = (rm >> (32 - (SP_FBITS + 1 + 3 + 1))) |
rm                150 arch/mips/math-emu/sp_mul.c 		     ((rm << (SP_FBITS + 1 + 3 + 1)) != 0);
rm                152 arch/mips/math-emu/sp_mul.c 	assert(rm & (SP_HIDDEN_BIT << 3));
rm                154 arch/mips/math-emu/sp_mul.c 	return ieee754sp_format(rs, re, rm);
rm                 55 arch/mips/math-emu/sp_rint.c 	switch (ieee754_csr.rm) {
rm                 22 arch/mips/math-emu/sp_simple.c 		oldrm = ieee754_csr.rm;
rm                 23 arch/mips/math-emu/sp_simple.c 		ieee754_csr.rm = FPU_CSR_RD;
rm                 25 arch/mips/math-emu/sp_simple.c 		ieee754_csr.rm = oldrm;
rm                 40 arch/mips/math-emu/sp_simple.c 		oldrm = ieee754_csr.rm;
rm                 41 arch/mips/math-emu/sp_simple.c 		ieee754_csr.rm = FPU_CSR_RD;
rm                 46 arch/mips/math-emu/sp_simple.c 		ieee754_csr.rm = oldrm;
rm                 90 arch/mips/math-emu/sp_sqrt.c 		switch (ieee754_csr.rm) {
rm                 83 arch/mips/math-emu/sp_sub.c 			return ieee754sp_zero(ieee754_csr.rm == FPU_CSR_RD);
rm                155 arch/mips/math-emu/sp_sub.c 			if (ieee754_csr.rm == FPU_CSR_RD)
rm                 72 arch/mips/math-emu/sp_tint.c 		switch (ieee754_csr.rm) {
rm                 68 arch/mips/math-emu/sp_tlong.c 		switch (ieee754_csr.rm) {
rm               1902 arch/powerpc/platforms/powernv/pci-ioda.c 		unsigned long index, unsigned long npages, bool rm)
rm               1909 arch/powerpc/platforms/powernv/pci-ioda.c 	__be64 __iomem *invalidate = pnv_ioda_get_inval_reg(pe->phb, rm);
rm               1924 arch/powerpc/platforms/powernv/pci-ioda.c 		if (rm)
rm               1985 arch/powerpc/platforms/powernv/pci-ioda.c static void pnv_pci_phb3_tce_invalidate_entire(struct pnv_phb *phb, bool rm)
rm               1987 arch/powerpc/platforms/powernv/pci-ioda.c 	__be64 __iomem *invalidate = pnv_ioda_get_inval_reg(phb, rm);
rm               1991 arch/powerpc/platforms/powernv/pci-ioda.c 	if (rm)
rm               2007 arch/powerpc/platforms/powernv/pci-ioda.c static void pnv_pci_phb3_tce_invalidate(struct pnv_ioda_pe *pe, bool rm,
rm               2011 arch/powerpc/platforms/powernv/pci-ioda.c 	__be64 __iomem *invalidate = pnv_ioda_get_inval_reg(pe->phb, rm);
rm               2026 arch/powerpc/platforms/powernv/pci-ioda.c 		if (rm)
rm               2046 arch/powerpc/platforms/powernv/pci-ioda.c 		unsigned long index, unsigned long npages, bool rm)
rm               2067 arch/powerpc/platforms/powernv/pci-ioda.c 			pnv_pci_phb3_tce_invalidate_entire(phb, rm);
rm               2071 arch/powerpc/platforms/powernv/pci-ioda.c 			pnv_pci_phb3_tce_invalidate(pe, rm, shift,
rm               2081 arch/powerpc/platforms/powernv/pci-ioda.c void pnv_pci_ioda2_tce_invalidate_entire(struct pnv_phb *phb, bool rm)
rm               2084 arch/powerpc/platforms/powernv/pci-ioda.c 		pnv_pci_phb3_tce_invalidate_entire(phb, rm);
rm                214 arch/powerpc/platforms/powernv/pci.h extern void pnv_pci_ioda2_tce_invalidate_entire(struct pnv_phb *phb, bool rm);
rm                101 arch/powerpc/platforms/ps3/mm.c 	struct mem_region rm;
rm                110 arch/powerpc/platforms/ps3/mm.c 	DBG("%s:%d: map.rm.size   = %llxh\n", func, line, m->rm.size);
rm                128 arch/powerpc/platforms/ps3/mm.c 	return (phys_addr < map.rm.size || phys_addr >= map.total)
rm                230 arch/powerpc/platforms/ps3/mm.c 	r->offset = r->base - map.rm.size;
rm                282 arch/powerpc/platforms/ps3/mm.c 	if (result || r->base < map.rm.size) {
rm                289 arch/powerpc/platforms/ps3/mm.c 	r->offset = r->base - map.rm.size;
rm                318 arch/powerpc/platforms/ps3/mm.c 		map.total = map.rm.size;
rm                336 arch/powerpc/platforms/ps3/mm.c 	if (lpar_addr >= map.rm.size)
rm                984 arch/powerpc/platforms/ps3/mm.c 	if (r->offset < map.rm.size) {
rm                986 arch/powerpc/platforms/ps3/mm.c 		virt_addr = map.rm.base + r->offset;
rm                987 arch/powerpc/platforms/ps3/mm.c 		len = map.rm.size - r->offset;
rm                996 arch/powerpc/platforms/ps3/mm.c 	if (r->offset + r->len > map.rm.size) {
rm                998 arch/powerpc/platforms/ps3/mm.c 		virt_addr = map.rm.size;
rm               1000 arch/powerpc/platforms/ps3/mm.c 		if (r->offset >= map.rm.size)
rm               1001 arch/powerpc/platforms/ps3/mm.c 			virt_addr += r->offset - map.rm.size;
rm               1003 arch/powerpc/platforms/ps3/mm.c 			len -= map.rm.size - r->offset;
rm               1026 arch/powerpc/platforms/ps3/mm.c 	if (r->offset < map.rm.size) {
rm               1028 arch/powerpc/platforms/ps3/mm.c 		lpar_addr = map.rm.base + r->offset;
rm               1029 arch/powerpc/platforms/ps3/mm.c 		len = map.rm.size - r->offset;
rm               1037 arch/powerpc/platforms/ps3/mm.c 	if (r->offset + r->len > map.rm.size) {
rm               1041 arch/powerpc/platforms/ps3/mm.c 		if (r->offset >= map.rm.size)
rm               1042 arch/powerpc/platforms/ps3/mm.c 			lpar_addr += r->offset - map.rm.size;
rm               1044 arch/powerpc/platforms/ps3/mm.c 			len -= map.rm.size - r->offset;
rm               1126 arch/powerpc/platforms/ps3/mm.c 	if (r->offset >= map.rm.size)
rm               1192 arch/powerpc/platforms/ps3/mm.c 	result = ps3_repository_read_mm_info(&map.rm.base, &map.rm.size,
rm               1198 arch/powerpc/platforms/ps3/mm.c 	map.rm.offset = map.rm.base;
rm               1203 arch/powerpc/platforms/ps3/mm.c 	BUG_ON(map.rm.base);
rm               1204 arch/powerpc/platforms/ps3/mm.c 	BUG_ON(!map.rm.size);
rm               1209 arch/powerpc/platforms/ps3/mm.c 		result = ps3_mm_region_create(&map.r1, map.total - map.rm.size);
rm               1216 arch/powerpc/platforms/ps3/mm.c 	map.total = map.rm.size + map.r1.size;
rm               1222 arch/powerpc/platforms/ps3/mm.c 			__func__, __LINE__, map.rm.size,
rm               1223 arch/powerpc/platforms/ps3/mm.c 			map.total - map.rm.size);
rm               1224 arch/powerpc/platforms/ps3/mm.c 		memblock_add(map.rm.size, map.total - map.rm.size);
rm                302 arch/sh/kernel/disassemble.c 		int rm = 0;
rm                364 arch/sh/kernel/disassemble.c 				rm = nibs[n];
rm                368 arch/sh/kernel/disassemble.c 				rm = (nibs[n] & 0x3);
rm                408 arch/sh/kernel/disassemble.c 				printk("r%d", rm);
rm                411 arch/sh/kernel/disassemble.c 				printk("@r%d+", rm);
rm                414 arch/sh/kernel/disassemble.c 				printk("@-r%d", rm);
rm                417 arch/sh/kernel/disassemble.c 				printk("@r%d", rm);
rm                420 arch/sh/kernel/disassemble.c 				printk("@(%d,r%d)", imm, rm);
rm                435 arch/sh/kernel/disassemble.c 				printk("@(r0,r%d)", rm);
rm                482 arch/sh/kernel/disassemble.c 				printk("fr%d", rm);
rm                494 arch/sh/kernel/disassemble.c 				if (rm & 1) {
rm                495 arch/sh/kernel/disassemble.c 					printk("xd%d", rm & ~1);
rm                500 arch/sh/kernel/disassemble.c 				printk("dr%d", rm);
rm                517 arch/sh/kernel/disassemble.c 				printk("fv%d", rm*4);
rm                 89 arch/sh/kernel/traps_32.c 	unsigned long *rm, *rn;
rm                 97 arch/sh/kernel/traps_32.c 	rm = &regs->regs[index];
rm                113 arch/sh/kernel/traps_32.c 			srcu = (unsigned char __user *)*rm;
rm                127 arch/sh/kernel/traps_32.c 			src = (unsigned char *)rm;
rm                141 arch/sh/kernel/traps_32.c 		src = (unsigned char*) rm;
rm                153 arch/sh/kernel/traps_32.c 		src = (unsigned char*) rm;
rm                164 arch/sh/kernel/traps_32.c 		srcu = (unsigned char __user *)*rm;
rm                175 arch/sh/kernel/traps_32.c 		srcu = (unsigned char __user *)*rm;
rm                177 arch/sh/kernel/traps_32.c 			*rm += count;
rm                197 arch/sh/kernel/traps_32.c 			dstu = (unsigned char __user *)*rm; /* called Rn in the spec */
rm                206 arch/sh/kernel/traps_32.c 			srcu = (unsigned char __user *)*rm;
rm                306 arch/sh/kernel/traps_32.c 	u_int rm;
rm                316 arch/sh/kernel/traps_32.c 	rm = regs->regs[index];
rm                344 arch/sh/kernel/traps_32.c 				regs->pc += rm + 4;
rm                351 arch/sh/kernel/traps_32.c 				regs->pc += rm + 4;
rm                371 arch/sh/kernel/traps_32.c 				regs->pc = rm;
rm                378 arch/sh/kernel/traps_32.c 				regs->pc = rm;
rm                752 arch/x86/events/intel/rapl.c 	struct rapl_model *rm;
rm                759 arch/x86/events/intel/rapl.c 	rm = (struct rapl_model *) id->driver_data;
rm                761 arch/x86/events/intel/rapl.c 					false, (void *) &rm->events);
rm                763 arch/x86/events/intel/rapl.c 	ret = rapl_check_hw_unit(rm->apply_quirk);
rm                 93 arch/x86/include/asm/fpu/types.h 	u8			rm;
rm                346 arch/x86/lib/insn.c 	insn_byte_t mod, rm, base;
rm                371 arch/x86/lib/insn.c 		rm = X86_MODRM_RM(insn->modrm.value);
rm                379 arch/x86/lib/insn.c 			if ((mod == 0 && rm == 6) || mod == 2) {
rm                385 arch/x86/lib/insn.c 			if ((mod == 0 && rm == 5) || mod == 2 ||
rm                111 arch/x86/math-emu/fpu_proto.h extern int FPU_sub(int flags, int rm, int control_w);
rm                 92 arch/x86/math-emu/fpu_system.h #define FPU_rm			(*(u_char *)&(I387->soft.rm))
rm                214 arch/x86/math-emu/get_address.c 	unsigned rm = FPU_modrm & 7;
rm                229 arch/x86/math-emu/get_address.c 	if (rm == 4 && mod != 3) {
rm                232 arch/x86/math-emu/get_address.c 		cpu_reg_ptr = &REG_(rm);
rm                235 arch/x86/math-emu/get_address.c 			if (rm == 5) {
rm                299 arch/x86/math-emu/get_address.c 	unsigned rm = FPU_modrm & 7;
rm                315 arch/x86/math-emu/get_address.c 		if (rm == 6) {
rm                347 arch/x86/math-emu/get_address.c 	switch (rm) {
rm                129 arch/x86/math-emu/reg_add_sub.c int FPU_sub(int flags, int rm, int control_w)
rm                141 arch/x86/math-emu/reg_add_sub.c 		b = (FPU_REG *) rm;
rm                144 arch/x86/math-emu/reg_add_sub.c 		b = &st(rm);
rm                145 arch/x86/math-emu/reg_add_sub.c 		tagb = FPU_gettagi(rm);
rm                148 arch/x86/math-emu/reg_add_sub.c 			deststnr = rm;
rm                 28 arch/x86/math-emu/reg_divide.c int FPU_div(int flags, int rm, int control_w)
rm                 37 arch/x86/math-emu/reg_divide.c 		deststnr = rm;
rm                 46 arch/x86/math-emu/reg_divide.c 			a = (FPU_REG *) rm;
rm                 49 arch/x86/math-emu/reg_divide.c 			a = &st(rm);
rm                 51 arch/x86/math-emu/reg_divide.c 			taga = FPU_gettagi(rm);
rm                 58 arch/x86/math-emu/reg_divide.c 			b = (FPU_REG *) rm;
rm                 61 arch/x86/math-emu/reg_divide.c 			b = &st(rm);
rm                 63 arch/x86/math-emu/reg_divide.c 			tagb = FPU_gettagi(rm);
rm                133 arch/x86/math-emu/reg_divide.c 			return real_2op_NaN((FPU_REG *) rm, flags & 0x0f, 0,
rm                141 arch/x86/math-emu/reg_divide.c 			return real_2op_NaN(st0_ptr, tag, rm,
rm                142 arch/x86/math-emu/reg_divide.c 					    (flags & REV) ? st0_ptr : &st(rm));
rm                145 arch/x86/math-emu/reg_divide.c 			tag = FPU_gettagi(rm);
rm                147 arch/x86/math-emu/reg_divide.c 				tag = FPU_Special(&st(rm));
rm                148 arch/x86/math-emu/reg_divide.c 			return real_2op_NaN(&st(rm), tag, 0,
rm                149 arch/x86/math-emu/reg_divide.c 					    (flags & REV) ? st0_ptr : &st(rm));
rm                335 drivers/gpu/drm/armada/armada_crtc.c 	uint32_t lm, rm, tm, bm, val, sclk;
rm                341 drivers/gpu/drm/armada/armada_crtc.c 	rm = adj->crtc_hsync_start - adj->crtc_hdisplay;
rm                348 drivers/gpu/drm/armada/armada_crtc.c 	DRM_DEBUG_KMS("lm %d rm %d tm %d bm %d\n", lm, rm, tm, bm);
rm                379 drivers/gpu/drm/armada/armada_crtc.c 	armada_reg_queue_set(regs, i, (lm << 16) | rm, LCD_SPU_H_PORCH);
rm               1810 drivers/gpu/drm/i915/gvt/handlers.c #define MMIO_F(reg, s, f, am, rm, d, r, w) do { \
rm               1812 drivers/gpu/drm/i915/gvt/handlers.c 		f, s, am, rm, d, r, w); \
rm               1832 drivers/gpu/drm/i915/gvt/handlers.c #define MMIO_RO(reg, d, f, rm, r, w) \
rm               1833 drivers/gpu/drm/i915/gvt/handlers.c 	MMIO_F(reg, 4, F_RO | f, 0, rm, d, r, w)
rm               1835 drivers/gpu/drm/i915/gvt/handlers.c #define MMIO_RING_F(prefix, s, f, am, rm, d, r, w) do { \
rm               1836 drivers/gpu/drm/i915/gvt/handlers.c 	MMIO_F(prefix(RENDER_RING_BASE), s, f, am, rm, d, r, w); \
rm               1837 drivers/gpu/drm/i915/gvt/handlers.c 	MMIO_F(prefix(BLT_RING_BASE), s, f, am, rm, d, r, w); \
rm               1838 drivers/gpu/drm/i915/gvt/handlers.c 	MMIO_F(prefix(GEN6_BSD_RING_BASE), s, f, am, rm, d, r, w); \
rm               1839 drivers/gpu/drm/i915/gvt/handlers.c 	MMIO_F(prefix(VEBOX_RING_BASE), s, f, am, rm, d, r, w); \
rm               1841 drivers/gpu/drm/i915/gvt/handlers.c 		MMIO_F(prefix(GEN8_BSD2_RING_BASE), s, f, am, rm, d, r, w); \
rm               1856 drivers/gpu/drm/i915/gvt/handlers.c #define MMIO_RING_RO(prefix, d, f, rm, r, w) \
rm               1857 drivers/gpu/drm/i915/gvt/handlers.c 	MMIO_RING_F(prefix, 4, F_RO | f, 0, rm, d, r, w)
rm                610 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c 			ret = dpu_rm_reserve(&dpu_kms->rm, drm_enc, crtc_state,
rm                995 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c 	ret = dpu_rm_reserve(&dpu_kms->rm, drm_enc, drm_crtc->state,
rm               1006 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c 		if (!dpu_rm_get_hw(&dpu_kms->rm, &hw_iter))
rm               1013 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c 		if (!dpu_rm_get_hw(&dpu_kms->rm, &hw_iter))
rm               1021 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c 		if (!dpu_rm_get_hw(&dpu_kms->rm, &hw_iter))
rm               1062 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c 				if (!dpu_rm_get_hw(&dpu_kms->rm, &hw_iter))
rm               1086 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c 	dpu_rm_release(&dpu_kms->rm, drm_enc);
rm               1241 drivers/gpu/drm/msm/disp/dpu1/dpu_encoder.c 	dpu_rm_release(&dpu_kms->rm, drm_enc);
rm                608 drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c 		dpu_rm_destroy(&dpu_kms->rm);
rm                880 drivers/gpu/drm/msm/disp/dpu1/dpu_kms.c 	rc = dpu_rm_init(&dpu_kms->rm, dpu_kms->catalog, dpu_kms->mmio);
rm                114 drivers/gpu/drm/msm/disp/dpu1/dpu_kms.h 	struct dpu_rm rm;
rm                 53 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c static bool _dpu_rm_get_hw_locked(struct dpu_rm *rm, struct dpu_rm_hw_iter *i)
rm                 57 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 	if (!rm || !i || i->type >= DPU_HW_BLK_MAX) {
rm                 63 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 	blk_list = &rm->hw_blks[i->type];
rm                 86 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c bool dpu_rm_get_hw(struct dpu_rm *rm, struct dpu_rm_hw_iter *i)
rm                 90 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 	mutex_lock(&rm->rm_lock);
rm                 91 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 	ret = _dpu_rm_get_hw_locked(rm, i);
rm                 92 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 	mutex_unlock(&rm->rm_lock);
rm                123 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c int dpu_rm_destroy(struct dpu_rm *rm)
rm                129 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 		list_for_each_entry_safe(hw_cur, hw_nxt, &rm->hw_blks[type],
rm                137 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 	mutex_destroy(&rm->rm_lock);
rm                143 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 		struct dpu_rm *rm,
rm                191 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 	list_add_tail(&blk->list, &rm->hw_blks[type]);
rm                196 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c int dpu_rm_init(struct dpu_rm *rm,
rm                203 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 	if (!rm || !cat || !mmio) {
rm                209 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 	memset(rm, 0, sizeof(*rm));
rm                211 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 	mutex_init(&rm->rm_lock);
rm                214 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 		INIT_LIST_HEAD(&rm->hw_blks[type]);
rm                225 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 		rc = _dpu_rm_hw_blk_create(rm, cat, mmio, DPU_HW_BLK_LM,
rm                232 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 		if (!rm->lm_max_width) {
rm                233 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 			rm->lm_max_width = lm->sblk->maxwidth;
rm                234 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 		} else if (rm->lm_max_width != lm->sblk->maxwidth) {
rm                240 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 			if (rm->lm_max_width > lm->sblk->maxwidth)
rm                241 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 				rm->lm_max_width = lm->sblk->maxwidth;
rm                246 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 		rc = _dpu_rm_hw_blk_create(rm, cat, mmio, DPU_HW_BLK_PINGPONG,
rm                260 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 		rc = _dpu_rm_hw_blk_create(rm, cat, mmio, DPU_HW_BLK_INTF,
rm                269 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 		rc = _dpu_rm_hw_blk_create(rm, cat, mmio, DPU_HW_BLK_CTL,
rm                280 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 	dpu_rm_destroy(rm);
rm                306 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 		struct dpu_rm *rm,
rm                340 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 	while (_dpu_rm_get_hw_locked(rm, &iter)) {
rm                361 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c static int _dpu_rm_reserve_lms(struct dpu_rm *rm, uint32_t enc_id,
rm                379 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 			_dpu_rm_get_hw_locked(rm, &iter_i)) {
rm                387 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 				rm, enc_id, reqs, lm[lm_count],
rm                397 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 				_dpu_rm_get_hw_locked(rm, &iter_j)) {
rm                402 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 					rm, enc_id, reqs, iter_j.blk,
rm                430 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 		struct dpu_rm *rm,
rm                447 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 	while (_dpu_rm_get_hw_locked(rm, &iter)) {
rm                481 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 		struct dpu_rm *rm,
rm                491 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 	while (_dpu_rm_get_hw_locked(rm, &iter)) {
rm                515 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 		struct dpu_rm *rm,
rm                526 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 		ret = _dpu_rm_reserve_intf(rm, enc_id, id,
rm                536 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 		struct dpu_rm *rm,
rm                543 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 	ret = _dpu_rm_reserve_lms(rm, enc->base.id, reqs);
rm                549 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 	ret = _dpu_rm_reserve_ctls(rm, enc->base.id, &reqs->topology);
rm                555 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 	ret = _dpu_rm_reserve_intf_related_hw(rm, enc->base.id, &reqs->hw_res);
rm                563 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 		struct dpu_rm *rm,
rm                580 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c static void _dpu_rm_release_reservation(struct dpu_rm *rm, uint32_t enc_id)
rm                586 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 		list_for_each_entry(blk, &rm->hw_blks[type], list) {
rm                596 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c void dpu_rm_release(struct dpu_rm *rm, struct drm_encoder *enc)
rm                598 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 	mutex_lock(&rm->rm_lock);
rm                600 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 	_dpu_rm_release_reservation(rm, enc->base.id);
rm                602 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 	mutex_unlock(&rm->rm_lock);
rm                606 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 		struct dpu_rm *rm,
rm                622 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 	mutex_lock(&rm->rm_lock);
rm                624 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 	ret = _dpu_rm_populate_requirements(rm, enc, crtc_state, &reqs,
rm                631 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 	ret = _dpu_rm_make_reservation(rm, enc, crtc_state, &reqs);
rm                634 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 		_dpu_rm_release_reservation(rm, enc->base.id);
rm                639 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 		_dpu_rm_release_reservation(rm, enc->base.id);
rm                643 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.c 	mutex_unlock(&rm->rm_lock);
rm                 55 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h int dpu_rm_init(struct dpu_rm *rm,
rm                 64 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h int dpu_rm_destroy(struct dpu_rm *rm);
rm                 80 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h int dpu_rm_reserve(struct dpu_rm *rm,
rm                 93 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h void dpu_rm_release(struct dpu_rm *rm, struct drm_encoder *enc);
rm                119 drivers/gpu/drm/msm/disp/dpu1/dpu_rm.h bool dpu_rm_get_hw(struct dpu_rm *rm, struct dpu_rm_hw_iter *iter);
rm                100 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv04.c 			u32 rm = ((1ULL << c->bits) - 1) << c->regs;
rm                102 drivers/gpu/drm/nouveau/nvkm/engine/fifo/dmanv04.c 			u32 rv = (nvkm_rd32(device, c->regp) &  rm) >> c->regs;
rm                799 drivers/hwmon/asc7621.c #define PREAD(name, n, pri, rm, rl, m, s, r) \
rm                801 drivers/hwmon/asc7621.c 	  .priority = pri, .msb[0] = rm, .lsb[0] = rl, .mask[0] = m, \
rm                804 drivers/hwmon/asc7621.c #define PWRITE(name, n, pri, rm, rl, m, s, r) \
rm                806 drivers/hwmon/asc7621.c 	  .priority = pri, .msb[0] = rm, .lsb[0] = rl, .mask[0] = m, \
rm                813 drivers/hwmon/asc7621.c #define PWRITEM(name, n, pri, rm, rl, m, s, r) \
rm                815 drivers/hwmon/asc7621.c 	  .priority = pri, .msb = rm, .lsb = rl, .mask = m, .shift = s,}
rm                912 drivers/isdn/hardware/mISDN/hfcmulti.c hfcmulti_resync(struct hfc_multi *locked, struct hfc_multi *newmaster, int rm)
rm               1002 drivers/isdn/hardware/mISDN/hfcmulti.c 			if (!rm)
rm               1014 drivers/isdn/hardware/mISDN/hfcmulti.c plxsd_checksync(struct hfc_multi *hc, int rm)
rm               1022 drivers/isdn/hardware/mISDN/hfcmulti.c 			hfcmulti_resync(hc, hc, rm);
rm               1030 drivers/isdn/hardware/mISDN/hfcmulti.c 			hfcmulti_resync(hc, NULL, rm);
rm                 52 drivers/macintosh/rack-meter.c 	struct rackmeter	*rm;
rm                 95 drivers/macintosh/rack-meter.c static void rackmeter_setup_i2s(struct rackmeter *rm)
rm                 97 drivers/macintosh/rack-meter.c 	struct macio_chip *macio = rm->mdev->bus->chip;
rm                106 drivers/macintosh/rack-meter.c 	pmac_call_feature(PMAC_FTR_SOUND_CHIP_ENABLE, rm->i2s, 0, 1);
rm                121 drivers/macintosh/rack-meter.c 	out_le32(rm->i2s_regs + 0x10, 0x01fa0000);
rm                122 drivers/macintosh/rack-meter.c 	(void)in_le32(rm->i2s_regs + 0x10);
rm                132 drivers/macintosh/rack-meter.c static void rackmeter_set_default_pattern(struct rackmeter *rm)
rm                138 drivers/macintosh/rack-meter.c 			rm->ubuf[i] = (i & 1) * 255;
rm                140 drivers/macintosh/rack-meter.c 			rm->ubuf[i] = ((~i) & 1) * 255;
rm                144 drivers/macintosh/rack-meter.c static void rackmeter_do_pause(struct rackmeter *rm, int pause)
rm                146 drivers/macintosh/rack-meter.c 	struct rackmeter_dma *rdma = rm->dma_buf_v;
rm                150 drivers/macintosh/rack-meter.c 	rm->paused = pause;
rm                152 drivers/macintosh/rack-meter.c 		DBDMA_DO_STOP(rm->dma_regs);
rm                158 drivers/macintosh/rack-meter.c 	rm->dma_buf_v->mark = 0;
rm                161 drivers/macintosh/rack-meter.c 	out_le32(&rm->dma_regs->cmdptr_hi, 0);
rm                162 drivers/macintosh/rack-meter.c 	out_le32(&rm->dma_regs->cmdptr, rm->dma_buf_p);
rm                163 drivers/macintosh/rack-meter.c 	out_le32(&rm->dma_regs->control, (RUN << 16) | RUN);
rm                166 drivers/macintosh/rack-meter.c static void rackmeter_setup_dbdma(struct rackmeter *rm)
rm                168 drivers/macintosh/rack-meter.c 	struct rackmeter_dma *db = rm->dma_buf_v;
rm                172 drivers/macintosh/rack-meter.c 	DBDMA_DO_RESET(rm->dma_regs);
rm                185 drivers/macintosh/rack-meter.c 	cmd->phy_addr = cpu_to_le32(rm->dma_buf_p +
rm                192 drivers/macintosh/rack-meter.c 	cmd->phy_addr = cpu_to_le32(rm->dma_buf_p +
rm                198 drivers/macintosh/rack-meter.c 	cmd->phy_addr = cpu_to_le32(rm->dma_buf_p +
rm                205 drivers/macintosh/rack-meter.c 	cmd->phy_addr = cpu_to_le32(rm->dma_buf_p +
rm                207 drivers/macintosh/rack-meter.c 	cmd->cmd_dep = cpu_to_le32(rm->dma_buf_p);
rm                209 drivers/macintosh/rack-meter.c 	rackmeter_do_pause(rm, 0);
rm                216 drivers/macintosh/rack-meter.c 	struct rackmeter *rm = rcpu->rm;
rm                240 drivers/macintosh/rack-meter.c 		rm->ubuf[i + offset] = ub;
rm                246 drivers/macintosh/rack-meter.c 	pause = (rm->cpu[0].zero && rm->cpu[1].zero);
rm                247 drivers/macintosh/rack-meter.c 	if (pause != rm->paused) {
rm                248 drivers/macintosh/rack-meter.c 		mutex_lock(&rm->sem);
rm                249 drivers/macintosh/rack-meter.c 		pause = (rm->cpu[0].zero && rm->cpu[1].zero);
rm                250 drivers/macintosh/rack-meter.c 		rackmeter_do_pause(rm, pause);
rm                251 drivers/macintosh/rack-meter.c 		mutex_unlock(&rm->sem);
rm                257 drivers/macintosh/rack-meter.c static void rackmeter_init_cpu_sniffer(struct rackmeter *rm)
rm                267 drivers/macintosh/rack-meter.c 	rm->cpu[0].rm = rm;
rm                268 drivers/macintosh/rack-meter.c 	INIT_DELAYED_WORK(&rm->cpu[0].sniffer, rackmeter_do_timer);
rm                269 drivers/macintosh/rack-meter.c 	rm->cpu[1].rm = rm;
rm                270 drivers/macintosh/rack-meter.c 	INIT_DELAYED_WORK(&rm->cpu[1].sniffer, rackmeter_do_timer);
rm                277 drivers/macintosh/rack-meter.c 		rcpu = &rm->cpu[cpu];
rm                280 drivers/macintosh/rack-meter.c 		schedule_delayed_work_on(cpu, &rm->cpu[cpu].sniffer,
rm                285 drivers/macintosh/rack-meter.c static void rackmeter_stop_cpu_sniffer(struct rackmeter *rm)
rm                287 drivers/macintosh/rack-meter.c 	cancel_delayed_work_sync(&rm->cpu[0].sniffer);
rm                288 drivers/macintosh/rack-meter.c 	cancel_delayed_work_sync(&rm->cpu[1].sniffer);
rm                291 drivers/macintosh/rack-meter.c static int rackmeter_setup(struct rackmeter *rm)
rm                294 drivers/macintosh/rack-meter.c 	rackmeter_setup_i2s(rm);
rm                297 drivers/macintosh/rack-meter.c 	rackmeter_set_default_pattern(rm);
rm                300 drivers/macintosh/rack-meter.c 	rackmeter_setup_dbdma(rm);
rm                303 drivers/macintosh/rack-meter.c 	rackmeter_init_cpu_sniffer(rm);
rm                311 drivers/macintosh/rack-meter.c static u32 rackmeter_calc_sample(struct rackmeter *rm, unsigned int index)
rm                318 drivers/macintosh/rack-meter.c 		sample |= ((rm->ubuf[led] >= 0x80) << 15);
rm                325 drivers/macintosh/rack-meter.c 	struct rackmeter *rm = arg;
rm                326 drivers/macintosh/rack-meter.c 	struct rackmeter_dma *db = rm->dma_buf_v;
rm                334 drivers/macintosh/rack-meter.c 	(void)in_le32(&rm->dma_regs->status);
rm                345 drivers/macintosh/rack-meter.c 		if (++rm->stale_irq > 3) {
rm                348 drivers/macintosh/rack-meter.c 			DBDMA_DO_RESET(rm->dma_regs);
rm                360 drivers/macintosh/rack-meter.c 		buf[i] = rackmeter_calc_sample(rm, i);
rm                370 drivers/macintosh/rack-meter.c 	struct rackmeter *rm = NULL;
rm                399 drivers/macintosh/rack-meter.c 	rm = kzalloc(sizeof(*rm), GFP_KERNEL);
rm                400 drivers/macintosh/rack-meter.c 	if (rm == NULL) {
rm                405 drivers/macintosh/rack-meter.c 	rm->mdev = mdev;
rm                406 drivers/macintosh/rack-meter.c 	rm->i2s = i2s;
rm                407 drivers/macintosh/rack-meter.c 	mutex_init(&rm->sem);
rm                408 drivers/macintosh/rack-meter.c 	dev_set_drvdata(&mdev->ofdev.dev, rm);
rm                426 drivers/macintosh/rack-meter.c 	rm->irq = macio_irq(mdev, 1);
rm                428 drivers/macintosh/rack-meter.c 	rm->irq = irq_of_parse_and_map(i2s, 1);
rm                429 drivers/macintosh/rack-meter.c 	if (!rm->irq ||
rm                442 drivers/macintosh/rack-meter.c 	pr_debug("  irq %d\n", rm->irq);
rm                444 drivers/macintosh/rack-meter.c 	rm->ubuf = (u8 *)__get_free_page(GFP_KERNEL);
rm                445 drivers/macintosh/rack-meter.c 	if (rm->ubuf == NULL) {
rm                452 drivers/macintosh/rack-meter.c 	rm->dma_buf_v = dma_alloc_coherent(&macio_get_pci_dev(mdev)->dev,
rm                454 drivers/macintosh/rack-meter.c 					   &rm->dma_buf_p, GFP_KERNEL);
rm                455 drivers/macintosh/rack-meter.c 	if (rm->dma_buf_v == NULL) {
rm                462 drivers/macintosh/rack-meter.c 	rm->i2s_regs = ioremap(macio_resource_start(mdev, 0), 0x1000);
rm                464 drivers/macintosh/rack-meter.c 	rm->i2s_regs = ioremap(ri2s.start, 0x1000);
rm                466 drivers/macintosh/rack-meter.c 	if (rm->i2s_regs == NULL) {
rm                473 drivers/macintosh/rack-meter.c 	rm->dma_regs = ioremap(macio_resource_start(mdev, 1), 0x100);
rm                475 drivers/macintosh/rack-meter.c 	rm->dma_regs = ioremap(rdma.start, 0x100);
rm                477 drivers/macintosh/rack-meter.c 	if (rm->dma_regs == NULL) {
rm                484 drivers/macintosh/rack-meter.c 	rc = rackmeter_setup(rm);
rm                492 drivers/macintosh/rack-meter.c 	rc = request_irq(rm->irq, rackmeter_irq, 0, "rackmeter", rm);
rm                502 drivers/macintosh/rack-meter.c 	DBDMA_DO_RESET(rm->dma_regs);
rm                504 drivers/macintosh/rack-meter.c 	iounmap(rm->dma_regs);
rm                506 drivers/macintosh/rack-meter.c 	iounmap(rm->i2s_regs);
rm                510 drivers/macintosh/rack-meter.c 			  rm->dma_buf_v, rm->dma_buf_p);
rm                512 drivers/macintosh/rack-meter.c 	free_page((unsigned long)rm->ubuf);
rm                518 drivers/macintosh/rack-meter.c 	kfree(rm);
rm                528 drivers/macintosh/rack-meter.c 	struct rackmeter *rm = dev_get_drvdata(&mdev->ofdev.dev);
rm                531 drivers/macintosh/rack-meter.c 	rackmeter_stop_cpu_sniffer(rm);
rm                537 drivers/macintosh/rack-meter.c 	DBDMA_DO_RESET(rm->dma_regs);
rm                540 drivers/macintosh/rack-meter.c 	free_irq(rm->irq, rm);
rm                543 drivers/macintosh/rack-meter.c 	iounmap(rm->dma_regs);
rm                544 drivers/macintosh/rack-meter.c 	iounmap(rm->i2s_regs);
rm                549 drivers/macintosh/rack-meter.c 			  rm->dma_buf_v, rm->dma_buf_p);
rm                552 drivers/macintosh/rack-meter.c 	free_page((unsigned long)rm->ubuf);
rm                560 drivers/macintosh/rack-meter.c 	kfree(rm);
rm                567 drivers/macintosh/rack-meter.c 	struct rackmeter *rm = dev_get_drvdata(&mdev->ofdev.dev);
rm                569 drivers/macintosh/rack-meter.c 	if (rm == NULL)
rm                573 drivers/macintosh/rack-meter.c 	rackmeter_stop_cpu_sniffer(rm);
rm                576 drivers/macintosh/rack-meter.c 	DBDMA_DO_RESET(rm->dma_regs);
rm               2252 drivers/media/pci/ddbridge/ddbridge-core.c 	const struct ddb_regmap *rm = io_regmap(io, 0);
rm               2262 drivers/media/pci/ddbridge/ddbridge-core.c 		dma->regs = rm->odma->base + rm->odma->size * nr;
rm               2263 drivers/media/pci/ddbridge/ddbridge-core.c 		dma->bufregs = rm->odma_buf->base + rm->odma_buf->size * nr;
rm               2269 drivers/media/pci/ddbridge/ddbridge-core.c 		dma->regs = rm->idma->base + rm->idma->size * nr;
rm               2270 drivers/media/pci/ddbridge/ddbridge-core.c 		dma->bufregs = rm->idma_buf->base + rm->idma_buf->size * nr;
rm               2284 drivers/media/pci/ddbridge/ddbridge-core.c 	const struct ddb_regmap *rm;
rm               2289 drivers/media/pci/ddbridge/ddbridge-core.c 	rm = io_regmap(input, 1);
rm               2291 drivers/media/pci/ddbridge/ddbridge-core.c 		(rm->input->base + rm->input->size * nr);
rm               2315 drivers/media/pci/ddbridge/ddbridge-core.c 	const struct ddb_regmap *rm;
rm               2320 drivers/media/pci/ddbridge/ddbridge-core.c 	rm = io_regmap(output, 1);
rm               2322 drivers/media/pci/ddbridge/ddbridge-core.c 		(rm->output->base + rm->output->size * nr);
rm               2370 drivers/media/pci/ddbridge/ddbridge-core.c 	const struct ddb_regmap *rm;
rm               2376 drivers/media/pci/ddbridge/ddbridge-core.c 		rm = info->regmap;
rm               2377 drivers/media/pci/ddbridge/ddbridge-core.c 		if (!rm)
rm                518 drivers/mtd/nand/raw/mpc5121_nfc.c 	struct mpc512x_reset_module *rm;
rm                534 drivers/mtd/nand/raw/mpc5121_nfc.c 	rm = of_iomap(rmnode, 0);
rm                535 drivers/mtd/nand/raw/mpc5121_nfc.c 	if (!rm) {
rm                541 drivers/mtd/nand/raw/mpc5121_nfc.c 	rcwh = in_be32(&rm->rcwhr);
rm                586 drivers/mtd/nand/raw/mpc5121_nfc.c 	iounmap(rm);
rm                470 drivers/net/can/usb/peak_usb/pcan_usb_fd.c 	struct pucan_rx_msg *rm = (struct pucan_rx_msg *)rx_msg;
rm                471 drivers/net/can/usb/peak_usb/pcan_usb_fd.c 	struct peak_usb_device *dev = usb_if->dev[pucan_msg_get_channel(rm)];
rm                475 drivers/net/can/usb/peak_usb/pcan_usb_fd.c 	const u16 rx_msg_flags = le16_to_cpu(rm->flags);
rm                489 drivers/net/can/usb/peak_usb/pcan_usb_fd.c 		cfd->len = can_dlc2len(get_canfd_dlc(pucan_msg_get_dlc(rm)));
rm                496 drivers/net/can/usb/peak_usb/pcan_usb_fd.c 		cfd->len = get_can_dlc(pucan_msg_get_dlc(rm));
rm                499 drivers/net/can/usb/peak_usb/pcan_usb_fd.c 	cfd->can_id = le32_to_cpu(rm->can_id);
rm                507 drivers/net/can/usb/peak_usb/pcan_usb_fd.c 		memcpy(cfd->data, rm->d, cfd->len);
rm                509 drivers/net/can/usb/peak_usb/pcan_usb_fd.c 	peak_usb_netif_rx(skb, &usb_if->time_ref, le32_to_cpu(rm->ts_low));
rm               1891 drivers/net/ethernet/apm/xgene/xgene_enet_main.c 		pdata->rm = RM3;
rm               1899 drivers/net/ethernet/apm/xgene/xgene_enet_main.c 		pdata->rm = RM1;
rm               1908 drivers/net/ethernet/apm/xgene/xgene_enet_main.c 		pdata->rm = RM0;
rm               1966 drivers/net/ethernet/apm/xgene/xgene_enet_main.c 		pdata->rm = RM0;
rm                213 drivers/net/ethernet/apm/xgene/xgene_enet_main.h 	enum xgene_enet_rm rm;
rm                259 drivers/net/ethernet/apm/xgene/xgene_enet_main.h 	return ((u16)pdata->rm << 10) | ring->num;
rm                 75 drivers/net/ethernet/chelsio/cxgb/common.h #define t1_rx_mode_promisc(rm)	(rm->dev->flags & IFF_PROMISC)
rm                 76 drivers/net/ethernet/chelsio/cxgb/common.h #define t1_rx_mode_allmulti(rm)	(rm->dev->flags & IFF_ALLMULTI)
rm                 77 drivers/net/ethernet/chelsio/cxgb/common.h #define t1_rx_mode_mc_cnt(rm)	(netdev_mc_count(rm->dev))
rm                 78 drivers/net/ethernet/chelsio/cxgb/common.h #define t1_get_netdev(rm)	(rm->dev)
rm                119 drivers/net/ethernet/chelsio/cxgb/cxgb2.c 	struct t1_rx_mode rm;
rm                121 drivers/net/ethernet/chelsio/cxgb/cxgb2.c 	rm.dev = dev;
rm                122 drivers/net/ethernet/chelsio/cxgb/cxgb2.c 	mac->ops->set_rx_mode(mac, &rm);
rm                111 drivers/net/ethernet/chelsio/cxgb/gmac.h 	int (*set_rx_mode)(struct cmac *, struct t1_rx_mode *rm);
rm                345 drivers/net/ethernet/chelsio/cxgb/pm3393.c static int pm3393_set_rx_mode(struct cmac *cmac, struct t1_rx_mode *rm)
rm                360 drivers/net/ethernet/chelsio/cxgb/pm3393.c 	if (t1_rx_mode_promisc(rm)) {
rm                364 drivers/net/ethernet/chelsio/cxgb/pm3393.c 	if (t1_rx_mode_allmulti(rm)) {
rm                371 drivers/net/ethernet/chelsio/cxgb/pm3393.c 	} else if (t1_rx_mode_mc_cnt(rm)) {
rm                377 drivers/net/ethernet/chelsio/cxgb/pm3393.c 		netdev_for_each_mc_addr(ha, t1_get_netdev(rm)) {
rm                433 drivers/net/ethernet/chelsio/cxgb/vsc7326.c static int mac_set_rx_mode(struct cmac *mac, struct t1_rx_mode *rm)
rm                441 drivers/net/ethernet/chelsio/cxgb/vsc7326.c 	if (t1_rx_mode_promisc(rm))
rm                139 drivers/parport/parport_gsc.h 	const unsigned char rm = (PARPORT_CONTROL_STROBE |
rm                144 drivers/parport/parport_gsc.h 	return priv->ctr & rm; /* Use soft copy */
rm                936 drivers/parport/parport_ip32.c 	const unsigned int rm =
rm                938 drivers/parport/parport_ip32.c 	return __parport_ip32_read_control(p) & rm;
rm               1050 drivers/pinctrl/pinctrl-st.c 	struct regmap *rm = info->regmap;
rm               1065 drivers/pinctrl/pinctrl-st.c 	rt_p->clk1notclk0 = devm_regmap_field_alloc(dev, rm, clk1notclk0);
rm               1066 drivers/pinctrl/pinctrl-st.c 	rt_p->delay_0	= devm_regmap_field_alloc(dev, rm, delay_0);
rm               1067 drivers/pinctrl/pinctrl-st.c 	rt_p->delay_1 = devm_regmap_field_alloc(dev, rm, delay_1);
rm               1068 drivers/pinctrl/pinctrl-st.c 	rt_p->invertclk = devm_regmap_field_alloc(dev, rm, invertclk);
rm               1069 drivers/pinctrl/pinctrl-st.c 	rt_p->retime = devm_regmap_field_alloc(dev, rm, retime);
rm               1070 drivers/pinctrl/pinctrl-st.c 	rt_p->clknotdata = devm_regmap_field_alloc(dev, rm, clknotdata);
rm               1071 drivers/pinctrl/pinctrl-st.c 	rt_p->double_edge = devm_regmap_field_alloc(dev, rm, double_edge);
rm               1086 drivers/pinctrl/pinctrl-st.c 	struct regmap *rm = info->regmap;
rm               1097 drivers/pinctrl/pinctrl-st.c 			rt_d->rt[j] = devm_regmap_field_alloc(dev, rm, reg);
rm               1258 drivers/pinctrl/stm32/pinctrl-stm32.c 	struct regmap *rm;
rm               1266 drivers/pinctrl/stm32/pinctrl-stm32.c 	rm = pctl->regmap;
rm               1288 drivers/pinctrl/stm32/pinctrl-stm32.c 		pctl->irqmux[i] = devm_regmap_field_alloc(dev, rm, mux);
rm               2130 drivers/pinctrl/tegra/pinctrl-tegra20.c 	MUX_PG(rm,     I2C1,      RSVD2,     RSVD3,     RSVD4,         0x14, 25, 0x80, 14, 0xa4, 0),
rm                131 drivers/s390/net/ctcm_mpc.c 	__u32	ct, sw, rm, dup;
rm                141 drivers/s390/net/ctcm_mpc.c 	rm  = 16;
rm                169 drivers/s390/net/ctcm_mpc.c 		rm--;
rm                187 drivers/s390/net/ctcm_mpc.c 		rm = 16;
rm                191 drivers/s390/net/ctcm_mpc.c 		for ( ; rm > 0; rm--, sw++) {
rm                847 drivers/video/console/sticore.c 		u32 *rm;
rm                865 drivers/video/console/sticore.c 			rm = (u32*) &sti->rm_entry;
rm                866 drivers/video/console/sticore.c 			*rm++ = gsc_readl(address+rm_offset+0x00);
rm                867 drivers/video/console/sticore.c 			*rm++ = gsc_readl(address+rm_offset+0x04);
rm                868 drivers/video/console/sticore.c 			*rm++ = gsc_readl(address+rm_offset+0x08);
rm                869 drivers/video/console/sticore.c 			*rm++ = gsc_readl(address+rm_offset+0x0c);
rm                288 drivers/video/fbdev/pm3fb.c 		unsigned long rm = 1 | (3 << 7);
rm                294 drivers/video/fbdev/pm3fb.c 			rm |= 3 << 15;
rm                301 drivers/video/fbdev/pm3fb.c 			rm |= 2 << 15;
rm                313 drivers/video/fbdev/pm3fb.c 		PM3_WRITE_REG(par, PM3RasterizerMode, rm);
rm                295 fs/gfs2/log.c  	int a, b, rm;
rm                302 fs/gfs2/log.c  		rm = (wrap) ? (a || b) : (a && b);
rm                303 fs/gfs2/log.c  		if (!rm)
rm                175 fs/ocfs2/journal.c 	struct ocfs2_recovery_map *rm;
rm                182 fs/ocfs2/journal.c 	rm = kzalloc(sizeof(struct ocfs2_recovery_map) +
rm                185 fs/ocfs2/journal.c 	if (!rm) {
rm                190 fs/ocfs2/journal.c 	rm->rm_entries = (unsigned int *)((char *)rm +
rm                192 fs/ocfs2/journal.c 	osb->recovery_map = rm;
rm                208 fs/ocfs2/journal.c 	struct ocfs2_recovery_map *rm;
rm                227 fs/ocfs2/journal.c 	rm = osb->recovery_map;
rm                230 fs/ocfs2/journal.c 	kfree(rm);
rm                237 fs/ocfs2/journal.c 	struct ocfs2_recovery_map *rm = osb->recovery_map;
rm                241 fs/ocfs2/journal.c 	for (i = 0; i < rm->rm_used; i++) {
rm                242 fs/ocfs2/journal.c 		if (rm->rm_entries[i] == node_num)
rm                253 fs/ocfs2/journal.c 	struct ocfs2_recovery_map *rm = osb->recovery_map;
rm                262 fs/ocfs2/journal.c 	BUG_ON(rm->rm_used >= osb->max_slots);
rm                264 fs/ocfs2/journal.c 	rm->rm_entries[rm->rm_used] = node_num;
rm                265 fs/ocfs2/journal.c 	rm->rm_used++;
rm                275 fs/ocfs2/journal.c 	struct ocfs2_recovery_map *rm = osb->recovery_map;
rm                279 fs/ocfs2/journal.c 	for (i = 0; i < rm->rm_used; i++) {
rm                280 fs/ocfs2/journal.c 		if (rm->rm_entries[i] == node_num)
rm                284 fs/ocfs2/journal.c 	if (i < rm->rm_used) {
rm                286 fs/ocfs2/journal.c 		memmove(&(rm->rm_entries[i]), &(rm->rm_entries[i + 1]),
rm                287 fs/ocfs2/journal.c 			(rm->rm_used - i - 1) * sizeof(unsigned int));
rm                288 fs/ocfs2/journal.c 		rm->rm_used--;
rm               1127 fs/ocfs2/journal.c 	struct ocfs2_recovery_map *rm = osb->recovery_map;
rm               1130 fs/ocfs2/journal.c 	empty = (rm->rm_used == 0);
rm               1369 fs/ocfs2/journal.c 	struct ocfs2_recovery_map *rm = osb->recovery_map;
rm               1408 fs/ocfs2/journal.c 	while (rm->rm_used) {
rm               1411 fs/ocfs2/journal.c 		node_num = rm->rm_entries[0];
rm                218 fs/ocfs2/super.c 	struct ocfs2_recovery_map *rm = osb->recovery_map;
rm                271 fs/ocfs2/super.c 	if (rm->rm_used == 0)
rm                274 fs/ocfs2/super.c 		for (i = 0; i < rm->rm_used; i++)
rm                276 fs/ocfs2/super.c 					rm->rm_entries[i]);
rm                 68 fs/xfs/scrub/refcount.c 	struct xfs_rmap_irec	rm;
rm                132 fs/xfs/scrub/refcount.c 		memcpy(&frag->rm, rec, sizeof(frag->rm));
rm                178 fs/xfs/scrub/refcount.c 		if (frag->rm.rm_startblock < bno)
rm                180 fs/xfs/scrub/refcount.c 		bno = frag->rm.rm_startblock;
rm                188 fs/xfs/scrub/refcount.c 		if (frag->rm.rm_startblock > refchk->bno)
rm                190 fs/xfs/scrub/refcount.c 		bno = frag->rm.rm_startblock + frag->rm.rm_blockcount;
rm                211 fs/xfs/scrub/refcount.c 			bno = frag->rm.rm_startblock + frag->rm.rm_blockcount;
rm                224 fs/xfs/scrub/refcount.c 			bno = frag->rm.rm_startblock + frag->rm.rm_blockcount;
rm                225 fs/xfs/scrub/refcount.c 			if (frag->rm.rm_startblock != rbno)
rm                221 include/linux/math64.h 	} rl, rm, rn, rh, a0, b0;
rm                228 include/linux/math64.h 	rm.ll = mul_u32_u32(a0.l.low, b0.l.high);
rm                237 include/linux/math64.h 	rl.l.high = c = (u64)rl.l.high + rm.l.low + rn.l.low;
rm                238 include/linux/math64.h 	rh.l.low = c = (c >> 32) + rm.l.high + rn.l.high + rh.l.low;
rm                176 include/linux/parport_pc.h 	const unsigned char rm = (PARPORT_CONTROL_STROBE |
rm                181 include/linux/parport_pc.h 	return priv->ctr & rm; /* Use soft copy */
rm                 27 include/linux/transport_class.h #define DECLARE_TRANSPORT_CLASS(cls, nm, su, rm, cfg)			\
rm                 33 include/linux/transport_class.h 	.remove = rm,							\
rm                388 net/ipv4/ipconfig.c 		struct rtentry rm;
rm                391 net/ipv4/ipconfig.c 		memset(&rm, 0, sizeof(rm));
rm                396 net/ipv4/ipconfig.c 		set_sockaddr((struct sockaddr_in *) &rm.rt_dst, 0, 0);
rm                397 net/ipv4/ipconfig.c 		set_sockaddr((struct sockaddr_in *) &rm.rt_genmask, 0, 0);
rm                398 net/ipv4/ipconfig.c 		set_sockaddr((struct sockaddr_in *) &rm.rt_gateway, ic_gateway, 0);
rm                399 net/ipv4/ipconfig.c 		rm.rt_flags = RTF_UP | RTF_GATEWAY;
rm                400 net/ipv4/ipconfig.c 		if ((err = ip_rt_ioctl(&init_net, SIOCADDRT, &rm)) < 0) {
rm                421 net/rds/cong.c 	struct rds_message *rm;
rm                423 net/rds/cong.c 	rm = rds_message_map_pages(map->m_page_addrs, RDS_CONG_MAP_BYTES);
rm                424 net/rds/cong.c 	if (!IS_ERR(rm))
rm                425 net/rds/cong.c 		rm->m_inc.i_hdr.h_flags = RDS_FLAG_CONG_BITMAP;
rm                427 net/rds/cong.c 	return rm;
rm                431 net/rds/connection.c 	struct rds_message *rm, *rtmp;
rm                444 net/rds/connection.c 	list_for_each_entry_safe(rm, rtmp,
rm                447 net/rds/connection.c 		list_del_init(&rm->m_conn_item);
rm                448 net/rds/connection.c 		BUG_ON(!list_empty(&rm->m_sock_item));
rm                449 net/rds/connection.c 		rds_message_put(rm);
rm                530 net/rds/connection.c 	struct rds_message *rm;
rm                565 net/rds/connection.c 				list_for_each_entry(rm, list, m_conn_item) {
rm                568 net/rds/connection.c 						__rds_inc_msg_cp(&rm->m_inc,
rm                431 net/rds/ib.h   int rds_ib_xmit(struct rds_connection *conn, struct rds_message *rm,
rm               1088 net/rds/ib_cm.c 		struct rds_message *rm;
rm               1090 net/rds/ib_cm.c 		rm = container_of(ic->i_data_op, struct rds_message, data);
rm               1091 net/rds/ib_cm.c 		rds_message_put(rm);
rm                 47 net/rds/ib_send.c static void rds_ib_send_complete(struct rds_message *rm,
rm                 49 net/rds/ib_send.c 				 void (*complete)(struct rds_message *rm, int status))
rm                 69 net/rds/ib_send.c 	complete(rm, notify_status);
rm                153 net/rds/ib_send.c 	struct rds_message *rm = NULL;
rm                159 net/rds/ib_send.c 			rm = container_of(send->s_op, struct rds_message, data);
rm                166 net/rds/ib_send.c 			rm = container_of(send->s_op, struct rds_message, rdma);
rm                173 net/rds/ib_send.c 			rm = container_of(send->s_op, struct rds_message, atomic);
rm                186 net/rds/ib_send.c 	return rm;
rm                243 net/rds/ib_send.c 	struct rds_message *rm = NULL;
rm                274 net/rds/ib_send.c 		rm = rds_ib_send_unmap_op(ic, send, wc->status);
rm                280 net/rds/ib_send.c 			if (send->s_op == rm->m_final_op) {
rm                284 net/rds/ib_send.c 				rds_message_unmapped(rm);
rm                286 net/rds/ib_send.c 			rds_message_put(rm);
rm                483 net/rds/ib_send.c int rds_ib_xmit(struct rds_connection *conn, struct rds_message *rm,
rm                510 net/rds/ib_send.c 	    && rm->m_inc.i_hdr.h_flags & RDS_FLAG_CONG_BITMAP) {
rm                512 net/rds/ib_send.c 		scat = &rm->data.op_sg[sg];
rm                518 net/rds/ib_send.c 	if (be32_to_cpu(rm->m_inc.i_hdr.h_len) == 0)
rm                521 net/rds/ib_send.c 		i = DIV_ROUND_UP(be32_to_cpu(rm->m_inc.i_hdr.h_len), RDS_FRAG_SIZE);
rm                549 net/rds/ib_send.c 		if (rm->data.op_nents) {
rm                550 net/rds/ib_send.c 			rm->data.op_count = ib_dma_map_sg(dev,
rm                551 net/rds/ib_send.c 							  rm->data.op_sg,
rm                552 net/rds/ib_send.c 							  rm->data.op_nents,
rm                554 net/rds/ib_send.c 			rdsdebug("ic %p mapping rm %p: %d\n", ic, rm, rm->data.op_count);
rm                555 net/rds/ib_send.c 			if (rm->data.op_count == 0) {
rm                562 net/rds/ib_send.c 			rm->data.op_count = 0;
rm                565 net/rds/ib_send.c 		rds_message_addref(rm);
rm                566 net/rds/ib_send.c 		rm->data.op_dmasg = 0;
rm                567 net/rds/ib_send.c 		rm->data.op_dmaoff = 0;
rm                568 net/rds/ib_send.c 		ic->i_data_op = &rm->data;
rm                571 net/rds/ib_send.c 		if (test_bit(RDS_MSG_ACK_REQUIRED, &rm->m_flags))
rm                572 net/rds/ib_send.c 			rm->m_inc.i_hdr.h_flags |= RDS_FLAG_ACK_REQUIRED;
rm                573 net/rds/ib_send.c 		if (test_bit(RDS_MSG_RETRANSMITTED, &rm->m_flags))
rm                574 net/rds/ib_send.c 			rm->m_inc.i_hdr.h_flags |= RDS_FLAG_RETRANSMITTED;
rm                578 net/rds/ib_send.c 		if (rm->rdma.op_active) {
rm                581 net/rds/ib_send.c 			ext_hdr.h_rdma_rkey = cpu_to_be32(rm->rdma.op_rkey);
rm                582 net/rds/ib_send.c 			rds_message_add_extension(&rm->m_inc.i_hdr,
rm                585 net/rds/ib_send.c 		if (rm->m_rdma_cookie) {
rm                586 net/rds/ib_send.c 			rds_message_add_rdma_dest_extension(&rm->m_inc.i_hdr,
rm                587 net/rds/ib_send.c 					rds_rdma_cookie_key(rm->m_rdma_cookie),
rm                588 net/rds/ib_send.c 					rds_rdma_cookie_offset(rm->m_rdma_cookie));
rm                595 net/rds/ib_send.c 		rm->m_inc.i_hdr.h_ack = cpu_to_be64(rds_ib_piggyb_ack(ic));
rm                596 net/rds/ib_send.c 		rds_message_make_checksum(&rm->m_inc.i_hdr);
rm                614 net/rds/ib_send.c 	if (rm->rdma.op_active && rm->rdma.op_fence)
rm                621 net/rds/ib_send.c 	scat = &ic->i_data_op->op_sg[rm->data.op_dmasg];
rm                638 net/rds/ib_send.c 		memcpy(&ic->i_send_hdrs[pos], &rm->m_inc.i_hdr, sizeof(struct rds_header));
rm                642 net/rds/ib_send.c 		    && scat != &rm->data.op_sg[rm->data.op_count]) {
rm                644 net/rds/ib_send.c 				  sg_dma_len(scat) - rm->data.op_dmaoff);
rm                648 net/rds/ib_send.c 			send->s_sge[1].addr += rm->data.op_dmaoff;
rm                652 net/rds/ib_send.c 			rm->data.op_dmaoff += len;
rm                653 net/rds/ib_send.c 			if (rm->data.op_dmaoff == sg_dma_len(scat)) {
rm                655 net/rds/ib_send.c 				rm->data.op_dmasg++;
rm                656 net/rds/ib_send.c 				rm->data.op_dmaoff = 0;
rm                695 net/rds/ib_send.c 		 && scat != &rm->data.op_sg[rm->data.op_count]);
rm                703 net/rds/ib_send.c 	if (scat == &rm->data.op_sg[rm->data.op_count]) {
rm                 75 net/rds/loop.c static int rds_loop_xmit(struct rds_connection *conn, struct rds_message *rm,
rm                 79 net/rds/loop.c 	struct scatterlist *sgp = &rm->data.op_sg[sg];
rm                 81 net/rds/loop.c 			be32_to_cpu(rm->m_inc.i_hdr.h_len);
rm                 84 net/rds/loop.c 	if (rm->m_inc.i_hdr.h_flags & RDS_FLAG_CONG_BITMAP) {
rm                 92 net/rds/loop.c 	rds_inc_init(&rm->m_inc, conn, &conn->c_laddr);
rm                 94 net/rds/loop.c 	rds_message_addref(rm);
rm                 96 net/rds/loop.c 	rds_recv_incoming(conn, &conn->c_laddr, &conn->c_faddr, &rm->m_inc,
rm                 99 net/rds/loop.c 	rds_send_drop_acked(conn, be64_to_cpu(rm->m_inc.i_hdr.h_sequence),
rm                102 net/rds/loop.c 	rds_inc_put(&rm->m_inc);
rm                113 net/rds/loop.c 	struct rds_message *rm = container_of(inc, struct rds_message, m_inc);
rm                115 net/rds/loop.c 	rds_message_put(rm);
rm                 51 net/rds/message.c void rds_message_addref(struct rds_message *rm)
rm                 53 net/rds/message.c 	rdsdebug("addref rm %p ref %d\n", rm, refcount_read(&rm->m_refcount));
rm                 54 net/rds/message.c 	refcount_inc(&rm->m_refcount);
rm                130 net/rds/message.c static void rds_message_purge(struct rds_message *rm)
rm                135 net/rds/message.c 	if (unlikely(test_bit(RDS_MSG_PAGEVEC, &rm->m_flags)))
rm                138 net/rds/message.c 	spin_lock_irqsave(&rm->m_rs_lock, flags);
rm                139 net/rds/message.c 	if (rm->m_rs) {
rm                140 net/rds/message.c 		struct rds_sock *rs = rm->m_rs;
rm                142 net/rds/message.c 		if (rm->data.op_mmp_znotifier) {
rm                144 net/rds/message.c 			rds_rm_zerocopy_callback(rs, rm->data.op_mmp_znotifier);
rm                146 net/rds/message.c 			rm->data.op_mmp_znotifier = NULL;
rm                149 net/rds/message.c 		rm->m_rs = NULL;
rm                151 net/rds/message.c 	spin_unlock_irqrestore(&rm->m_rs_lock, flags);
rm                153 net/rds/message.c 	for (i = 0; i < rm->data.op_nents; i++) {
rm                156 net/rds/message.c 			__free_page(sg_page(&rm->data.op_sg[i]));
rm                158 net/rds/message.c 			put_page(sg_page(&rm->data.op_sg[i]));
rm                160 net/rds/message.c 	rm->data.op_nents = 0;
rm                162 net/rds/message.c 	if (rm->rdma.op_active)
rm                163 net/rds/message.c 		rds_rdma_free_op(&rm->rdma);
rm                164 net/rds/message.c 	if (rm->rdma.op_rdma_mr)
rm                165 net/rds/message.c 		rds_mr_put(rm->rdma.op_rdma_mr);
rm                167 net/rds/message.c 	if (rm->atomic.op_active)
rm                168 net/rds/message.c 		rds_atomic_free_op(&rm->atomic);
rm                169 net/rds/message.c 	if (rm->atomic.op_rdma_mr)
rm                170 net/rds/message.c 		rds_mr_put(rm->atomic.op_rdma_mr);
rm                173 net/rds/message.c void rds_message_put(struct rds_message *rm)
rm                175 net/rds/message.c 	rdsdebug("put rm %p ref %d\n", rm, refcount_read(&rm->m_refcount));
rm                176 net/rds/message.c 	WARN(!refcount_read(&rm->m_refcount), "danger refcount zero on %p\n", rm);
rm                177 net/rds/message.c 	if (refcount_dec_and_test(&rm->m_refcount)) {
rm                178 net/rds/message.c 		BUG_ON(!list_empty(&rm->m_sock_item));
rm                179 net/rds/message.c 		BUG_ON(!list_empty(&rm->m_conn_item));
rm                180 net/rds/message.c 		rds_message_purge(rm);
rm                182 net/rds/message.c 		kfree(rm);
rm                286 net/rds/message.c 	struct rds_message *rm;
rm                291 net/rds/message.c 	rm = kzalloc(sizeof(struct rds_message) + extra_len, gfp);
rm                292 net/rds/message.c 	if (!rm)
rm                295 net/rds/message.c 	rm->m_used_sgs = 0;
rm                296 net/rds/message.c 	rm->m_total_sgs = extra_len / sizeof(struct scatterlist);
rm                298 net/rds/message.c 	refcount_set(&rm->m_refcount, 1);
rm                299 net/rds/message.c 	INIT_LIST_HEAD(&rm->m_sock_item);
rm                300 net/rds/message.c 	INIT_LIST_HEAD(&rm->m_conn_item);
rm                301 net/rds/message.c 	spin_lock_init(&rm->m_rs_lock);
rm                302 net/rds/message.c 	init_waitqueue_head(&rm->m_flush_wait);
rm                305 net/rds/message.c 	return rm;
rm                311 net/rds/message.c struct scatterlist *rds_message_alloc_sgs(struct rds_message *rm, int nents)
rm                313 net/rds/message.c 	struct scatterlist *sg_first = (struct scatterlist *) &rm[1];
rm                321 net/rds/message.c 	if (rm->m_used_sgs + nents > rm->m_total_sgs) {
rm                323 net/rds/message.c 			rm->m_total_sgs, rm->m_used_sgs, nents);
rm                327 net/rds/message.c 	sg_ret = &sg_first[rm->m_used_sgs];
rm                329 net/rds/message.c 	rm->m_used_sgs += nents;
rm                336 net/rds/message.c 	struct rds_message *rm;
rm                341 net/rds/message.c 	rm = rds_message_alloc(extra_bytes, GFP_NOWAIT);
rm                342 net/rds/message.c 	if (!rm)
rm                345 net/rds/message.c 	set_bit(RDS_MSG_PAGEVEC, &rm->m_flags);
rm                346 net/rds/message.c 	rm->m_inc.i_hdr.h_len = cpu_to_be32(total_len);
rm                347 net/rds/message.c 	rm->data.op_nents = DIV_ROUND_UP(total_len, PAGE_SIZE);
rm                348 net/rds/message.c 	rm->data.op_sg = rds_message_alloc_sgs(rm, num_sgs);
rm                349 net/rds/message.c 	if (IS_ERR(rm->data.op_sg)) {
rm                350 net/rds/message.c 		rds_message_put(rm);
rm                351 net/rds/message.c 		return ERR_CAST(rm->data.op_sg);
rm                354 net/rds/message.c 	for (i = 0; i < rm->data.op_nents; ++i) {
rm                355 net/rds/message.c 		sg_set_page(&rm->data.op_sg[i],
rm                360 net/rds/message.c 	return rm;
rm                363 net/rds/message.c static int rds_message_zcopy_from_user(struct rds_message *rm, struct iov_iter *from)
rm                371 net/rds/message.c 	rm->m_inc.i_hdr.h_len = cpu_to_be32(iov_iter_count(from));
rm                376 net/rds/message.c 	sg = rm->data.op_sg;
rm                382 net/rds/message.c 	rm->data.op_mmp_znotifier = &info->znotif;
rm                383 net/rds/message.c 	if (mm_account_pinned_pages(&rm->data.op_mmp_znotifier->z_mmp,
rm                399 net/rds/message.c 			for (i = 0; i < rm->data.op_nents; i++)
rm                400 net/rds/message.c 				put_page(sg_page(&rm->data.op_sg[i]));
rm                401 net/rds/message.c 			mmp = &rm->data.op_mmp_znotifier->z_mmp;
rm                410 net/rds/message.c 		rm->data.op_nents++;
rm                417 net/rds/message.c 	rm->data.op_mmp_znotifier = NULL;
rm                421 net/rds/message.c int rds_message_copy_from_user(struct rds_message *rm, struct iov_iter *from,
rm                429 net/rds/message.c 	rm->m_inc.i_hdr.h_len = cpu_to_be32(iov_iter_count(from));
rm                432 net/rds/message.c 	sg = rm->data.op_sg;
rm                436 net/rds/message.c 		return rds_message_zcopy_from_user(rm, from);
rm                444 net/rds/message.c 			rm->data.op_nents++;
rm                468 net/rds/message.c 	struct rds_message *rm;
rm                476 net/rds/message.c 	rm = container_of(inc, struct rds_message, m_inc);
rm                477 net/rds/message.c 	len = be32_to_cpu(rm->m_inc.i_hdr.h_len);
rm                479 net/rds/message.c 	sg = rm->data.op_sg;
rm                510 net/rds/message.c void rds_message_wait(struct rds_message *rm)
rm                512 net/rds/message.c 	wait_event_interruptible(rm->m_flush_wait,
rm                513 net/rds/message.c 			!test_bit(RDS_MSG_MAPPED, &rm->m_flags));
rm                516 net/rds/message.c void rds_message_unmapped(struct rds_message *rm)
rm                518 net/rds/message.c 	clear_bit(RDS_MSG_MAPPED, &rm->m_flags);
rm                519 net/rds/message.c 	wake_up_interruptible(&rm->m_flush_wait);
rm                572 net/rds/rdma.c int rds_cmsg_rdma_args(struct rds_sock *rs, struct rds_message *rm,
rm                577 net/rds/rdma.c 	struct rm_rdma_op *op = &rm->rdma;
rm                586 net/rds/rdma.c 	    || rm->rdma.op_active)
rm                627 net/rds/rdma.c 	op->op_sg = rds_message_alloc_sgs(rm, nr_pages);
rm                730 net/rds/rdma.c int rds_cmsg_rdma_dest(struct rds_sock *rs, struct rds_message *rm,
rm                739 net/rds/rdma.c 	    rm->m_rdma_cookie != 0)
rm                742 net/rds/rdma.c 	memcpy(&rm->m_rdma_cookie, CMSG_DATA(cmsg), sizeof(rm->m_rdma_cookie));
rm                749 net/rds/rdma.c 	r_key = rds_rdma_cookie_key(rm->m_rdma_cookie);
rm                761 net/rds/rdma.c 		rm->rdma.op_rdma_mr = mr;
rm                772 net/rds/rdma.c int rds_cmsg_rdma_map(struct rds_sock *rs, struct rds_message *rm,
rm                776 net/rds/rdma.c 	    rm->m_rdma_cookie != 0)
rm                779 net/rds/rdma.c 	return __rds_rdma_map(rs, CMSG_DATA(cmsg), &rm->m_rdma_cookie,
rm                780 net/rds/rdma.c 			      &rm->rdma.op_rdma_mr, rm->m_conn_path);
rm                786 net/rds/rdma.c int rds_cmsg_atomic(struct rds_sock *rs, struct rds_message *rm,
rm                794 net/rds/rdma.c 	 || rm->atomic.op_active)
rm                802 net/rds/rdma.c 		rm->atomic.op_type = RDS_ATOMIC_TYPE_FADD;
rm                803 net/rds/rdma.c 		rm->atomic.op_m_fadd.add = args->fadd.add;
rm                804 net/rds/rdma.c 		rm->atomic.op_m_fadd.nocarry_mask = 0;
rm                807 net/rds/rdma.c 		rm->atomic.op_type = RDS_ATOMIC_TYPE_FADD;
rm                808 net/rds/rdma.c 		rm->atomic.op_m_fadd.add = args->m_fadd.add;
rm                809 net/rds/rdma.c 		rm->atomic.op_m_fadd.nocarry_mask = args->m_fadd.nocarry_mask;
rm                812 net/rds/rdma.c 		rm->atomic.op_type = RDS_ATOMIC_TYPE_CSWP;
rm                813 net/rds/rdma.c 		rm->atomic.op_m_cswp.compare = args->cswp.compare;
rm                814 net/rds/rdma.c 		rm->atomic.op_m_cswp.swap = args->cswp.swap;
rm                815 net/rds/rdma.c 		rm->atomic.op_m_cswp.compare_mask = ~0;
rm                816 net/rds/rdma.c 		rm->atomic.op_m_cswp.swap_mask = ~0;
rm                819 net/rds/rdma.c 		rm->atomic.op_type = RDS_ATOMIC_TYPE_CSWP;
rm                820 net/rds/rdma.c 		rm->atomic.op_m_cswp.compare = args->m_cswp.compare;
rm                821 net/rds/rdma.c 		rm->atomic.op_m_cswp.swap = args->m_cswp.swap;
rm                822 net/rds/rdma.c 		rm->atomic.op_m_cswp.compare_mask = args->m_cswp.compare_mask;
rm                823 net/rds/rdma.c 		rm->atomic.op_m_cswp.swap_mask = args->m_cswp.swap_mask;
rm                829 net/rds/rdma.c 	rm->atomic.op_notify = !!(args->flags & RDS_RDMA_NOTIFY_ME);
rm                830 net/rds/rdma.c 	rm->atomic.op_silent = !!(args->flags & RDS_RDMA_SILENT);
rm                831 net/rds/rdma.c 	rm->atomic.op_active = 1;
rm                832 net/rds/rdma.c 	rm->atomic.op_recverr = rs->rs_recverr;
rm                833 net/rds/rdma.c 	rm->atomic.op_sg = rds_message_alloc_sgs(rm, 1);
rm                834 net/rds/rdma.c 	if (IS_ERR(rm->atomic.op_sg)) {
rm                835 net/rds/rdma.c 		ret = PTR_ERR(rm->atomic.op_sg);
rm                850 net/rds/rdma.c 	sg_set_page(rm->atomic.op_sg, page, 8, offset_in_page(args->local_addr));
rm                852 net/rds/rdma.c 	if (rm->atomic.op_notify || rm->atomic.op_recverr) {
rm                858 net/rds/rdma.c 		rm->atomic.op_notifier = kmalloc(sizeof(*rm->atomic.op_notifier), GFP_KERNEL);
rm                859 net/rds/rdma.c 		if (!rm->atomic.op_notifier) {
rm                864 net/rds/rdma.c 		rm->atomic.op_notifier->n_user_token = args->user_token;
rm                865 net/rds/rdma.c 		rm->atomic.op_notifier->n_status = RDS_RDMA_SUCCESS;
rm                868 net/rds/rdma.c 	rm->atomic.op_rkey = rds_rdma_cookie_key(args->cookie);
rm                869 net/rds/rdma.c 	rm->atomic.op_remote_addr = args->remote_addr + rds_rdma_cookie_offset(args->cookie);
rm                875 net/rds/rdma.c 	rm->atomic.op_active = 0;
rm                876 net/rds/rdma.c 	kfree(rm->atomic.op_notifier);
rm                557 net/rds/rds.h  	int (*xmit)(struct rds_connection *conn, struct rds_message *rm,
rm                852 net/rds/rds.h  struct scatterlist *rds_message_alloc_sgs(struct rds_message *rm, int nents);
rm                853 net/rds/rds.h  int rds_message_copy_from_user(struct rds_message *rm, struct iov_iter *from,
rm                865 net/rds/rds.h  void rds_message_addref(struct rds_message *rm);
rm                866 net/rds/rds.h  void rds_message_put(struct rds_message *rm);
rm                867 net/rds/rds.h  void rds_message_wait(struct rds_message *rm);
rm                868 net/rds/rds.h  void rds_message_unmapped(struct rds_message *rm);
rm                915 net/rds/rds.h  typedef int (*is_acked_func)(struct rds_message *rm, uint64_t ack);
rm                931 net/rds/rds.h  int rds_cmsg_rdma_dest(struct rds_sock *rs, struct rds_message *rm,
rm                933 net/rds/rds.h  int rds_cmsg_rdma_args(struct rds_sock *rs, struct rds_message *rm,
rm                936 net/rds/rds.h  int rds_cmsg_rdma_map(struct rds_sock *rs, struct rds_message *rm,
rm                940 net/rds/rds.h  void rds_rdma_send_complete(struct rds_message *rm, int wc_status);
rm                941 net/rds/rds.h  void rds_atomic_send_complete(struct rds_message *rm, int wc_status);
rm                942 net/rds/rds.h  int rds_cmsg_atomic(struct rds_sock *rs, struct rds_message *rm,
rm                136 net/rds/recv.c 	struct rds_message *rm, *tmp;
rm                150 net/rds/recv.c 				list_for_each_entry_safe(rm, tmp,
rm                153 net/rds/recv.c 					set_bit(RDS_MSG_FLUSH, &rm->m_flags);
rm                 67 net/rds/send.c 	struct rds_message *rm, *tmp;
rm                 71 net/rds/send.c 		rm = cp->cp_xmit_rm;
rm                 77 net/rds/send.c 		rds_message_unmapped(rm);
rm                 78 net/rds/send.c 		rds_message_put(rm);
rm                 95 net/rds/send.c 	list_for_each_entry_safe(rm, tmp, &cp->cp_retrans, m_conn_item) {
rm                 96 net/rds/send.c 		set_bit(RDS_MSG_ACK_REQUIRED, &rm->m_flags);
rm                 97 net/rds/send.c 		set_bit(RDS_MSG_RETRANSMITTED, &rm->m_flags);
rm                140 net/rds/send.c 	struct rds_message *rm;
rm                202 net/rds/send.c 		rm = cp->cp_xmit_rm;
rm                204 net/rds/send.c 		if (!rm) {
rm                219 net/rds/send.c 		if (!rm && test_and_clear_bit(0, &conn->c_map_queued)) {
rm                220 net/rds/send.c 			rm = rds_cong_update_alloc(conn);
rm                221 net/rds/send.c 			if (IS_ERR(rm)) {
rm                222 net/rds/send.c 				ret = PTR_ERR(rm);
rm                225 net/rds/send.c 			rm->data.op_active = 1;
rm                226 net/rds/send.c 			rm->m_inc.i_conn_path = cp;
rm                227 net/rds/send.c 			rm->m_inc.i_conn = cp->cp_conn;
rm                229 net/rds/send.c 			cp->cp_xmit_rm = rm;
rm                239 net/rds/send.c 		if (!rm) {
rm                255 net/rds/send.c 				rm = list_entry(cp->cp_send_queue.next,
rm                258 net/rds/send.c 				rds_message_addref(rm);
rm                264 net/rds/send.c 				list_move_tail(&rm->m_conn_item,
rm                270 net/rds/send.c 			if (!rm)
rm                280 net/rds/send.c 			if (test_bit(RDS_MSG_FLUSH, &rm->m_flags) ||
rm                281 net/rds/send.c 			    (rm->rdma.op_active &&
rm                282 net/rds/send.c 			    test_bit(RDS_MSG_RETRANSMITTED, &rm->m_flags))) {
rm                284 net/rds/send.c 				if (test_and_clear_bit(RDS_MSG_ON_CONN, &rm->m_flags))
rm                285 net/rds/send.c 					list_move(&rm->m_conn_item, &to_be_dropped);
rm                291 net/rds/send.c 			len = ntohl(rm->m_inc.i_hdr.h_len);
rm                294 net/rds/send.c 				set_bit(RDS_MSG_ACK_REQUIRED, &rm->m_flags);
rm                306 net/rds/send.c 			cp->cp_xmit_rm = rm;
rm                310 net/rds/send.c 		if (rm->rdma.op_active && !cp->cp_xmit_rdma_sent) {
rm                311 net/rds/send.c 			rm->m_final_op = &rm->rdma;
rm                315 net/rds/send.c 			set_bit(RDS_MSG_MAPPED, &rm->m_flags);
rm                316 net/rds/send.c 			ret = conn->c_trans->xmit_rdma(conn, &rm->rdma);
rm                318 net/rds/send.c 				clear_bit(RDS_MSG_MAPPED, &rm->m_flags);
rm                319 net/rds/send.c 				wake_up_interruptible(&rm->m_flush_wait);
rm                326 net/rds/send.c 		if (rm->atomic.op_active && !cp->cp_xmit_atomic_sent) {
rm                327 net/rds/send.c 			rm->m_final_op = &rm->atomic;
rm                331 net/rds/send.c 			set_bit(RDS_MSG_MAPPED, &rm->m_flags);
rm                332 net/rds/send.c 			ret = conn->c_trans->xmit_atomic(conn, &rm->atomic);
rm                334 net/rds/send.c 				clear_bit(RDS_MSG_MAPPED, &rm->m_flags);
rm                335 net/rds/send.c 				wake_up_interruptible(&rm->m_flush_wait);
rm                349 net/rds/send.c 		if (rm->data.op_nents == 0) {
rm                353 net/rds/send.c 			ops_present = (rm->atomic.op_active || rm->rdma.op_active);
rm                354 net/rds/send.c 			if (rm->atomic.op_active && !rm->atomic.op_silent)
rm                356 net/rds/send.c 			if (rm->rdma.op_active && !rm->rdma.op_silent)
rm                360 net/rds/send.c 			    && !rm->m_rdma_cookie)
rm                361 net/rds/send.c 				rm->data.op_active = 0;
rm                364 net/rds/send.c 		if (rm->data.op_active && !cp->cp_xmit_data_sent) {
rm                365 net/rds/send.c 			rm->m_final_op = &rm->data;
rm                367 net/rds/send.c 			ret = conn->c_trans->xmit(conn, rm,
rm                382 net/rds/send.c 			sg = &rm->data.op_sg[cp->cp_xmit_sg];
rm                393 net/rds/send.c 					       rm->data.op_nents);
rm                398 net/rds/send.c 			    (cp->cp_xmit_sg == rm->data.op_nents))
rm                407 net/rds/send.c 		if (!rm->data.op_active || cp->cp_xmit_data_sent) {
rm                416 net/rds/send.c 			rds_message_put(rm);
rm                428 net/rds/send.c 		list_for_each_entry(rm, &to_be_dropped, m_conn_item)
rm                429 net/rds/send.c 			rds_message_put(rm);
rm                473 net/rds/send.c static void rds_send_sndbuf_remove(struct rds_sock *rs, struct rds_message *rm)
rm                475 net/rds/send.c 	u32 len = be32_to_cpu(rm->m_inc.i_hdr.h_len);
rm                486 net/rds/send.c static inline int rds_send_is_acked(struct rds_message *rm, u64 ack,
rm                490 net/rds/send.c 		return is_acked(rm, ack);
rm                491 net/rds/send.c 	return be64_to_cpu(rm->m_inc.i_hdr.h_sequence) <= ack;
rm                500 net/rds/send.c void rds_rdma_send_complete(struct rds_message *rm, int status)
rm                507 net/rds/send.c 	spin_lock_irqsave(&rm->m_rs_lock, flags);
rm                509 net/rds/send.c 	ro = &rm->rdma;
rm                510 net/rds/send.c 	if (test_bit(RDS_MSG_ON_SOCK, &rm->m_flags) &&
rm                513 net/rds/send.c 		rs = rm->m_rs;
rm                524 net/rds/send.c 	spin_unlock_irqrestore(&rm->m_rs_lock, flags);
rm                536 net/rds/send.c void rds_atomic_send_complete(struct rds_message *rm, int status)
rm                543 net/rds/send.c 	spin_lock_irqsave(&rm->m_rs_lock, flags);
rm                545 net/rds/send.c 	ao = &rm->atomic;
rm                546 net/rds/send.c 	if (test_bit(RDS_MSG_ON_SOCK, &rm->m_flags)
rm                549 net/rds/send.c 		rs = rm->m_rs;
rm                560 net/rds/send.c 	spin_unlock_irqrestore(&rm->m_rs_lock, flags);
rm                575 net/rds/send.c __rds_send_complete(struct rds_sock *rs, struct rds_message *rm, int status)
rm                580 net/rds/send.c 	ro = &rm->rdma;
rm                587 net/rds/send.c 	ao = &rm->atomic;
rm                609 net/rds/send.c 	struct rds_message *rm;
rm                614 net/rds/send.c 		rm = list_entry(messages->next, struct rds_message,
rm                616 net/rds/send.c 		list_del_init(&rm->m_conn_item);
rm                628 net/rds/send.c 		spin_lock_irqsave(&rm->m_rs_lock, flags);
rm                629 net/rds/send.c 		if (!test_bit(RDS_MSG_ON_SOCK, &rm->m_flags))
rm                632 net/rds/send.c 		if (rs != rm->m_rs) {
rm                637 net/rds/send.c 			rs = rm->m_rs;
rm                645 net/rds/send.c 		if (test_and_clear_bit(RDS_MSG_ON_SOCK, &rm->m_flags)) {
rm                646 net/rds/send.c 			struct rm_rdma_op *ro = &rm->rdma;
rm                649 net/rds/send.c 			list_del_init(&rm->m_sock_item);
rm                650 net/rds/send.c 			rds_send_sndbuf_remove(rs, rm);
rm                659 net/rds/send.c 				rm->rdma.op_notifier = NULL;
rm                666 net/rds/send.c 		spin_unlock_irqrestore(&rm->m_rs_lock, flags);
rm                667 net/rds/send.c 		rds_message_put(rm);
rm                669 net/rds/send.c 			rds_message_put(rm);
rm                689 net/rds/send.c 	struct rds_message *rm, *tmp;
rm                695 net/rds/send.c 	list_for_each_entry_safe(rm, tmp, &cp->cp_retrans, m_conn_item) {
rm                696 net/rds/send.c 		if (!rds_send_is_acked(rm, ack, is_acked))
rm                699 net/rds/send.c 		list_move(&rm->m_conn_item, &list);
rm                700 net/rds/send.c 		clear_bit(RDS_MSG_ON_CONN, &rm->m_flags);
rm                724 net/rds/send.c 	struct rds_message *rm, *tmp;
rm                733 net/rds/send.c 	list_for_each_entry_safe(rm, tmp, &rs->rs_send_queue, m_sock_item) {
rm                735 net/rds/send.c 		    (!ipv6_addr_equal(&dest->sin6_addr, &rm->m_daddr) ||
rm                736 net/rds/send.c 		     dest->sin6_port != rm->m_inc.i_hdr.h_dport))
rm                739 net/rds/send.c 		list_move(&rm->m_sock_item, &list);
rm                740 net/rds/send.c 		rds_send_sndbuf_remove(rs, rm);
rm                741 net/rds/send.c 		clear_bit(RDS_MSG_ON_SOCK, &rm->m_flags);
rm                753 net/rds/send.c 	list_for_each_entry(rm, &list, m_sock_item) {
rm                755 net/rds/send.c 		conn = rm->m_inc.i_conn;
rm                757 net/rds/send.c 			cp = rm->m_inc.i_conn_path;
rm                767 net/rds/send.c 		if (!test_and_clear_bit(RDS_MSG_ON_CONN, &rm->m_flags)) {
rm                771 net/rds/send.c 		list_del_init(&rm->m_conn_item);
rm                778 net/rds/send.c 		spin_lock_irqsave(&rm->m_rs_lock, flags);
rm                781 net/rds/send.c 		__rds_send_complete(rs, rm, RDS_RDMA_CANCELED);
rm                784 net/rds/send.c 		spin_unlock_irqrestore(&rm->m_rs_lock, flags);
rm                786 net/rds/send.c 		rds_message_put(rm);
rm                792 net/rds/send.c 		rm = list_entry(list.next, struct rds_message, m_sock_item);
rm                793 net/rds/send.c 		list_del_init(&rm->m_sock_item);
rm                794 net/rds/send.c 		rds_message_wait(rm);
rm                801 net/rds/send.c 		spin_lock_irqsave(&rm->m_rs_lock, flags);
rm                804 net/rds/send.c 		__rds_send_complete(rs, rm, RDS_RDMA_CANCELED);
rm                807 net/rds/send.c 		spin_unlock_irqrestore(&rm->m_rs_lock, flags);
rm                809 net/rds/send.c 		rds_message_put(rm);
rm                820 net/rds/send.c 			     struct rds_message *rm, __be16 sport,
rm                829 net/rds/send.c 	len = be32_to_cpu(rm->m_inc.i_hdr.h_len);
rm                852 net/rds/send.c 			set_bit(RDS_MSG_ACK_REQUIRED, &rm->m_flags);
rm                854 net/rds/send.c 		list_add_tail(&rm->m_sock_item, &rs->rs_send_queue);
rm                855 net/rds/send.c 		set_bit(RDS_MSG_ON_SOCK, &rm->m_flags);
rm                856 net/rds/send.c 		rds_message_addref(rm);
rm                858 net/rds/send.c 		rm->m_rs = rs;
rm                862 net/rds/send.c 		rds_message_populate_header(&rm->m_inc.i_hdr, sport, dport, 0);
rm                863 net/rds/send.c 		rm->m_inc.i_conn = conn;
rm                864 net/rds/send.c 		rm->m_inc.i_conn_path = cp;
rm                865 net/rds/send.c 		rds_message_addref(rm);
rm                868 net/rds/send.c 		rm->m_inc.i_hdr.h_sequence = cpu_to_be64(cp->cp_next_tx_seq++);
rm                869 net/rds/send.c 		list_add_tail(&rm->m_conn_item, &cp->cp_send_queue);
rm                870 net/rds/send.c 		set_bit(RDS_MSG_ON_CONN, &rm->m_flags);
rm                874 net/rds/send.c 			 rm, len, rs, rs->rs_snd_bytes,
rm                875 net/rds/send.c 			 (unsigned long long)be64_to_cpu(rm->m_inc.i_hdr.h_sequence));
rm                971 net/rds/send.c static int rds_cmsg_zcopy(struct rds_sock *rs, struct rds_message *rm,
rm                977 net/rds/send.c 	    !rm->data.op_mmp_znotifier)
rm                980 net/rds/send.c 	rm->data.op_mmp_znotifier->z_cookie = *cookie;
rm                984 net/rds/send.c static int rds_cmsg_send(struct rds_sock *rs, struct rds_message *rm,
rm               1005 net/rds/send.c 			ret = rds_cmsg_rdma_args(rs, rm, cmsg, &vct->vec[ind]);
rm               1010 net/rds/send.c 			ret = rds_cmsg_rdma_dest(rs, rm, cmsg);
rm               1014 net/rds/send.c 			ret = rds_cmsg_rdma_map(rs, rm, cmsg);
rm               1027 net/rds/send.c 			ret = rds_cmsg_atomic(rs, rm, cmsg);
rm               1031 net/rds/send.c 			ret = rds_cmsg_zcopy(rs, rm, cmsg);
rm               1108 net/rds/send.c 	struct rds_message *rm = NULL;
rm               1269 net/rds/send.c 	rm = rds_message_alloc(ret, GFP_KERNEL);
rm               1270 net/rds/send.c 	if (!rm) {
rm               1277 net/rds/send.c 		rm->data.op_sg = rds_message_alloc_sgs(rm, num_sgs);
rm               1278 net/rds/send.c 		if (IS_ERR(rm->data.op_sg)) {
rm               1279 net/rds/send.c 			ret = PTR_ERR(rm->data.op_sg);
rm               1282 net/rds/send.c 		ret = rds_message_copy_from_user(rm, &msg->msg_iter, zcopy);
rm               1286 net/rds/send.c 	rm->data.op_active = 1;
rm               1288 net/rds/send.c 	rm->m_daddr = daddr;
rm               1313 net/rds/send.c 	rm->m_conn_path = cpath;
rm               1316 net/rds/send.c 	ret = rds_cmsg_send(rs, rm, msg, &allocated_mr, &vct);
rm               1324 net/rds/send.c 	if (rm->rdma.op_active && !conn->c_trans->xmit_rdma) {
rm               1326 net/rds/send.c 			       &rm->rdma, conn->c_trans->xmit_rdma);
rm               1331 net/rds/send.c 	if (rm->atomic.op_active && !conn->c_trans->xmit_atomic) {
rm               1333 net/rds/send.c 			       &rm->atomic, conn->c_trans->xmit_atomic);
rm               1350 net/rds/send.c 	while (!rds_send_queue_rm(rs, conn, cpath, rm, rs->rs_bound_port,
rm               1360 net/rds/send.c 					rds_send_queue_rm(rs, conn, cpath, rm,
rm               1393 net/rds/send.c 	rds_message_put(rm);
rm               1410 net/rds/send.c 		rds_rdma_unuse(rs, rds_rdma_cookie_key(rm->m_rdma_cookie), 1);
rm               1412 net/rds/send.c 	if (rm)
rm               1413 net/rds/send.c 		rds_message_put(rm);
rm               1429 net/rds/send.c 	struct rds_message *rm;
rm               1433 net/rds/send.c 	rm = rds_message_alloc(0, GFP_ATOMIC);
rm               1434 net/rds/send.c 	if (!rm) {
rm               1439 net/rds/send.c 	rm->m_daddr = cp->cp_conn->c_faddr;
rm               1440 net/rds/send.c 	rm->data.op_active = 1;
rm               1449 net/rds/send.c 	list_add_tail(&rm->m_conn_item, &cp->cp_send_queue);
rm               1450 net/rds/send.c 	set_bit(RDS_MSG_ON_CONN, &rm->m_flags);
rm               1451 net/rds/send.c 	rds_message_addref(rm);
rm               1452 net/rds/send.c 	rm->m_inc.i_conn = cp->cp_conn;
rm               1453 net/rds/send.c 	rm->m_inc.i_conn_path = cp;
rm               1455 net/rds/send.c 	rds_message_populate_header(&rm->m_inc.i_hdr, sport, dport,
rm               1457 net/rds/send.c 	rm->m_inc.i_hdr.h_flags |= h_flags;
rm               1465 net/rds/send.c 		rds_message_add_extension(&rm->m_inc.i_hdr,
rm               1468 net/rds/send.c 		rds_message_add_extension(&rm->m_inc.i_hdr,
rm               1484 net/rds/send.c 	rds_message_put(rm);
rm               1488 net/rds/send.c 	if (rm)
rm               1489 net/rds/send.c 		rds_message_put(rm);
rm                 89 net/rds/tcp.h  int rds_tcp_xmit(struct rds_connection *conn, struct rds_message *rm,
rm                 75 net/rds/tcp_send.c int rds_tcp_xmit(struct rds_connection *conn, struct rds_message *rm,
rm                 78 net/rds/tcp_send.c 	struct rds_conn_path *cp = rm->m_inc.i_conn_path;
rm                 90 net/rds/tcp_send.c 		rm->m_ack_seq = tc->t_last_sent_nxt +
rm                 92 net/rds/tcp_send.c 				be32_to_cpu(rm->m_inc.i_hdr.h_len) - 1;
rm                 94 net/rds/tcp_send.c 		set_bit(RDS_MSG_HAS_ACK_SEQ, &rm->m_flags);
rm                 95 net/rds/tcp_send.c 		tc->t_last_expected_una = rm->m_ack_seq + 1;
rm                 97 net/rds/tcp_send.c 		if (test_bit(RDS_MSG_RETRANSMITTED, &rm->m_flags))
rm                 98 net/rds/tcp_send.c 			rm->m_inc.i_hdr.h_flags |= RDS_FLAG_RETRANSMITTED;
rm                101 net/rds/tcp_send.c 			 rm, rds_tcp_write_seq(tc),
rm                102 net/rds/tcp_send.c 			 (unsigned long long)rm->m_ack_seq);
rm                110 net/rds/tcp_send.c 				      (void *)&rm->m_inc.i_hdr + hdr_off,
rm                111 net/rds/tcp_send.c 				      sizeof(rm->m_inc.i_hdr) - hdr_off);
rm                119 net/rds/tcp_send.c 	more = rm->data.op_nents > 1 ? (MSG_MORE | MSG_SENDPAGE_NOTLAST) : 0;
rm                120 net/rds/tcp_send.c 	while (sg < rm->data.op_nents) {
rm                124 net/rds/tcp_send.c 						sg_page(&rm->data.op_sg[sg]),
rm                125 net/rds/tcp_send.c 						rm->data.op_sg[sg].offset + off,
rm                126 net/rds/tcp_send.c 						rm->data.op_sg[sg].length - off,
rm                128 net/rds/tcp_send.c 		rdsdebug("tcp sendpage %p:%u:%u ret %d\n", (void *)sg_page(&rm->data.op_sg[sg]),
rm                129 net/rds/tcp_send.c 			 rm->data.op_sg[sg].offset + off, rm->data.op_sg[sg].length - off,
rm                136 net/rds/tcp_send.c 		if (off == rm->data.op_sg[sg].length) {
rm                140 net/rds/tcp_send.c 		if (sg == rm->data.op_nents - 1)
rm                176 net/rds/tcp_send.c static int rds_tcp_is_acked(struct rds_message *rm, uint64_t ack)
rm                178 net/rds/tcp_send.c 	if (!test_bit(RDS_MSG_HAS_ACK_SEQ, &rm->m_flags))
rm                180 net/rds/tcp_send.c 	return (__s32)((u32)rm->m_ack_seq - (u32)ack) < 0;
rm                813 net/sunrpc/xprtsock.c static int xs_sendpages(struct socket *sock, struct sockaddr *addr, int addrlen, struct xdr_buf *xdr, unsigned int base, rpc_fraghdr rm, int *sent_p)
rm                820 net/sunrpc/xprtsock.c 	unsigned int rmsize = rm ? sizeof(rm) : 0;
rm                835 net/sunrpc/xprtsock.c 			err = xs_send_rm_and_kvec(sock, &msg, rm,
rm                960 net/sunrpc/xprtsock.c 	rpc_fraghdr rm = xs_stream_record_marker(xdr);
rm                961 net/sunrpc/xprtsock.c 	unsigned int msglen = rm ? req->rq_slen + sizeof(rm) : req->rq_slen;
rm                976 net/sunrpc/xprtsock.c 			      transport->xmit.offset, rm, &sent);
rm               1107 net/sunrpc/xprtsock.c 	rpc_fraghdr rm = xs_stream_record_marker(xdr);
rm               1108 net/sunrpc/xprtsock.c 	unsigned int msglen = rm ? req->rq_slen + sizeof(rm) : req->rq_slen;
rm               1134 net/sunrpc/xprtsock.c 				      transport->xmit.offset, rm, &sent);
rm                588 net/tipc/group.c 	struct tipc_member *m, *rm, *pm;
rm                614 net/tipc/group.c 			rm = list_first_entry(active, struct tipc_member, list);
rm                615 net/tipc/group.c 			rm->state = MBR_RECLAIMING;
rm                616 net/tipc/group.c 			list_del_init(&rm->list);
rm                617 net/tipc/group.c 			tipc_group_proto_xmit(grp, rm, GRP_RECLAIM_MSG, xmitq);
rm                255 net/wireless/core.h 		struct cfg80211_roam_info rm;
rm                987 net/wireless/sme.c 		ev->rm.req_ie = next;
rm                988 net/wireless/sme.c 		ev->rm.req_ie_len = info->req_ie_len;
rm                989 net/wireless/sme.c 		memcpy((void *)ev->rm.req_ie, info->req_ie, info->req_ie_len);
rm                993 net/wireless/sme.c 		ev->rm.resp_ie = next;
rm                994 net/wireless/sme.c 		ev->rm.resp_ie_len = info->resp_ie_len;
rm                995 net/wireless/sme.c 		memcpy((void *)ev->rm.resp_ie, info->resp_ie,
rm               1000 net/wireless/sme.c 		ev->rm.fils.kek = next;
rm               1001 net/wireless/sme.c 		ev->rm.fils.kek_len = info->fils.kek_len;
rm               1002 net/wireless/sme.c 		memcpy((void *)ev->rm.fils.kek, info->fils.kek,
rm               1007 net/wireless/sme.c 		ev->rm.fils.pmk = next;
rm               1008 net/wireless/sme.c 		ev->rm.fils.pmk_len = info->fils.pmk_len;
rm               1009 net/wireless/sme.c 		memcpy((void *)ev->rm.fils.pmk, info->fils.pmk,
rm               1014 net/wireless/sme.c 		ev->rm.fils.pmkid = next;
rm               1015 net/wireless/sme.c 		memcpy((void *)ev->rm.fils.pmkid, info->fils.pmkid,
rm               1019 net/wireless/sme.c 	ev->rm.fils.update_erp_next_seq_num = info->fils.update_erp_next_seq_num;
rm               1021 net/wireless/sme.c 		ev->rm.fils.erp_next_seq_num = info->fils.erp_next_seq_num;
rm               1022 net/wireless/sme.c 	ev->rm.bss = info->bss;
rm                876 net/wireless/util.c 			__cfg80211_roamed(wdev, &ev->rm);
rm               1051 net/xfrm/xfrm_policy.c static void xfrm_policy_inexact_gc_tree(struct rb_root *r, bool rm)
rm               1059 net/xfrm/xfrm_policy.c 		xfrm_policy_inexact_gc_tree(&node->root, rm);
rm               1063 net/xfrm/xfrm_policy.c 			WARN_ON_ONCE(rm);
rm                298 sound/mips/ad1843.c 	int lg, rg, lm, rm;
rm                308 sound/mips/ad1843.c 		ad1843_read_multi(ad1843, 2, gp->lmute, &lm, gp->rmute, &rm);
rm                311 sound/mips/ad1843.c 		if (rm)
rm                331 sound/mips/ad1843.c 	int rm = (rg == 0) ? 1 : 0;
rm                338 sound/mips/ad1843.c 		ad1843_write_multi(ad1843, 2, gp->lmute, lm, gp->rmute, rm);
rm               5083 sound/soc/codecs/wcd9335.c 	struct regmap *rm = wcd->regmap;
rm               5086 sound/soc/codecs/wcd9335.c 	regmap_read(rm, WCD9335_CHIP_TIER_CTRL_EFUSE_VAL_OUT0, &val);
rm               5087 sound/soc/codecs/wcd9335.c 	regmap_read(rm, WCD9335_CHIP_TIER_CTRL_CHIP_ID_BYTE0, &byte0);
rm               5097 sound/soc/codecs/wcd9335.c 		regmap_write(rm, WCD9335_CODEC_RPM_RST_CTL, 0x01);
rm               5098 sound/soc/codecs/wcd9335.c 		regmap_write(rm, WCD9335_SIDO_SIDO_TEST_2, 0x00);
rm               5099 sound/soc/codecs/wcd9335.c 		regmap_write(rm, WCD9335_SIDO_SIDO_CCL_8, 0x6F);
rm               5100 sound/soc/codecs/wcd9335.c 		regmap_write(rm, WCD9335_BIAS_VBG_FINE_ADJ, 0x65);
rm               5101 sound/soc/codecs/wcd9335.c 		regmap_write(rm, WCD9335_CODEC_RPM_PWR_CDC_DIG_HM_CTL, 0x5);
rm               5102 sound/soc/codecs/wcd9335.c 		regmap_write(rm, WCD9335_CODEC_RPM_PWR_CDC_DIG_HM_CTL, 0x7);
rm               5103 sound/soc/codecs/wcd9335.c 		regmap_write(rm, WCD9335_CODEC_RPM_PWR_CDC_DIG_HM_CTL, 0x3);
rm               5104 sound/soc/codecs/wcd9335.c 		regmap_write(rm, WCD9335_CODEC_RPM_RST_CTL, 0x3);
rm                346 tools/arch/x86/lib/insn.c 	insn_byte_t mod, rm, base;
rm                371 tools/arch/x86/lib/insn.c 		rm = X86_MODRM_RM(insn->modrm.value);
rm                379 tools/arch/x86/lib/insn.c 			if ((mod == 0 && rm == 6) || mod == 2) {
rm                385 tools/arch/x86/lib/insn.c 			if ((mod == 0 && rm == 5) || mod == 2 ||
rm                116 tools/perf/arch/powerpc/util/perf_regs.c 	regmatch_t rm[5];
rm                129 tools/perf/arch/powerpc/util/perf_regs.c 	if (!regexec(&sdt_op_regex1, old_op, 3, rm, 0)) {
rm                133 tools/perf/arch/powerpc/util/perf_regs.c 		new_len += (int)(rm[2].rm_eo - rm[2].rm_so);
rm                140 tools/perf/arch/powerpc/util/perf_regs.c 			(int)(rm[2].rm_eo - rm[2].rm_so), old_op + rm[2].rm_so);
rm                141 tools/perf/arch/powerpc/util/perf_regs.c 	} else if (!regexec(&sdt_op_regex2, old_op, 5, rm, 0)) {
rm                146 tools/perf/arch/powerpc/util/perf_regs.c 		prefix = (rm[1].rm_so == -1) ? '+' : '-';
rm                149 tools/perf/arch/powerpc/util/perf_regs.c 		new_len += (int)(rm[2].rm_eo - rm[2].rm_so);
rm                150 tools/perf/arch/powerpc/util/perf_regs.c 		new_len += (int)(rm[4].rm_eo - rm[4].rm_so);
rm                157 tools/perf/arch/powerpc/util/perf_regs.c 			(int)(rm[2].rm_eo - rm[2].rm_so), old_op + rm[2].rm_so,
rm                158 tools/perf/arch/powerpc/util/perf_regs.c 			(int)(rm[4].rm_eo - rm[4].rm_so), old_op + rm[4].rm_so);
rm                204 tools/perf/arch/x86/util/perf_regs.c 	regmatch_t rm[6];
rm                223 tools/perf/arch/x86/util/perf_regs.c 	    regexec(&sdt_op_regex, old_op, 6, rm, 0)   ||
rm                224 tools/perf/arch/x86/util/perf_regs.c 	    rm[4].rm_eo - rm[4].rm_so > SDT_REG_NAME_SIZE) {
rm                241 tools/perf/arch/x86/util/perf_regs.c 	if (rm[3].rm_so != rm[3].rm_eo) {
rm                242 tools/perf/arch/x86/util/perf_regs.c 		if (rm[1].rm_so != rm[1].rm_eo)
rm                243 tools/perf/arch/x86/util/perf_regs.c 			prefix[0] = *(old_op + rm[1].rm_so);
rm                244 tools/perf/arch/x86/util/perf_regs.c 		else if (rm[2].rm_so != rm[2].rm_eo)
rm                251 tools/perf/arch/x86/util/perf_regs.c 	sdt_rename_register(old_op + rm[4].rm_so, rm[4].rm_eo - rm[4].rm_so,
rm                256 tools/perf/arch/x86/util/perf_regs.c 		  (rm[2].rm_eo - rm[2].rm_so) +
rm                257 tools/perf/arch/x86/util/perf_regs.c 		  (rm[3].rm_eo - rm[3].rm_so) +
rm                259 tools/perf/arch/x86/util/perf_regs.c 		  (rm[5].rm_eo - rm[5].rm_so) +
rm                268 tools/perf/arch/x86/util/perf_regs.c 		  (int)(rm[2].rm_eo - rm[2].rm_so), old_op + rm[2].rm_so,
rm                269 tools/perf/arch/x86/util/perf_regs.c 		  (int)(rm[3].rm_eo - rm[3].rm_so), old_op + rm[3].rm_so,
rm                271 tools/perf/arch/x86/util/perf_regs.c 		  (int)(rm[5].rm_eo - rm[5].rm_so), old_op + rm[5].rm_so);