rs                 50 arch/arm/mach-omap1/reset.c 	u16 rs;
rs                 52 arch/arm/mach-omap1/reset.c 	rs = __raw_readw(OMAP1_IO_ADDRESS(ARM_SYSST));
rs                 54 arch/arm/mach-omap1/reset.c 	if (rs & (1 << ARM_SYSST_POR_SHIFT))
rs                 56 arch/arm/mach-omap1/reset.c 	if (rs & (1 << ARM_SYSST_EXT_RST_SHIFT))
rs                 58 arch/arm/mach-omap1/reset.c 	if (rs & (1 << ARM_SYSST_ARM_WDRST_SHIFT))
rs                 60 arch/arm/mach-omap1/reset.c 	if (rs & (1 << ARM_SYSST_GLOB_SWRST_SHIFT))
rs                722 arch/arm/net/bpf_jit_32.c 		const s8 *rs;
rs                724 arch/arm/net/bpf_jit_32.c 		rs = arm_bpf_get_reg64(src, tmp2, ctx);
rs                727 arch/arm/net/bpf_jit_32.c 		emit_alu_r(rd[1], rs[1], true, false, op, ctx);
rs                728 arch/arm/net/bpf_jit_32.c 		emit_alu_r(rd[0], rs[0], true, true, op, ctx);
rs                730 arch/arm/net/bpf_jit_32.c 		s8 rs;
rs                732 arch/arm/net/bpf_jit_32.c 		rs = arm_bpf_get_reg32(src_lo, tmp2[1], ctx);
rs                735 arch/arm/net/bpf_jit_32.c 		emit_alu_r(rd[1], rs, true, false, op, ctx);
rs               1356 arch/arm/net/bpf_jit_32.c 	const s8 *rd, *rs;
rs               1633 arch/arm/net/bpf_jit_32.c 		rs = arm_bpf_get_reg64(src, tmp2, ctx);
rs               1634 arch/arm/net/bpf_jit_32.c 		emit_str_r(dst_lo, rs, off, ctx, BPF_SIZE(code));
rs                217 arch/arm/net/bpf_jit_32.h #define ARM_MOV_SR(rd, rm, type, rs)	\
rs                218 arch/arm/net/bpf_jit_32.h 	(_AL3_SR(ARM_MOV_R(rd, rm)) | (type) << 5 | (rs) << 8)
rs                235 arch/arm/net/bpf_jit_32.h #define ARM_ORR_SR(rd, rn, rm, type, rs)	\
rs                236 arch/arm/net/bpf_jit_32.h 	(_AL3_SR(ARM_ORR_R(rd, rn, rm)) | (type) << 5 | (rs) << 8)
rs                238 arch/arm/net/bpf_jit_32.h #define ARM_ORRS_SR(rd, rn, rm, type, rs)	\
rs                239 arch/arm/net/bpf_jit_32.h 	(_AL3_SR(ARM_ORRS_R(rd, rn, rm)) | (type) << 5 | (rs) << 8)
rs                162 arch/arm/probes/kprobes/actions-arm.c 	int rs = (insn >> 8) & 0xf;
rs                169 arch/arm/probes/kprobes/actions-arm.c 	register unsigned long rsv asm("r1") = regs->uregs[rs];
rs                224 arch/arm/probes/kprobes/actions-arm.c 	int rs = (insn >> 8) & 0xf;
rs                229 arch/arm/probes/kprobes/actions-arm.c 	register unsigned long rsv asm("r1") = regs->uregs[rs];
rs                210 arch/arm64/kernel/traps.c 	static DEFINE_RATELIMIT_STATE(rs, DEFAULT_RATELIMIT_INTERVAL,
rs                219 arch/arm64/kernel/traps.c 	    !__ratelimit(&rs))
rs                442 arch/ia64/include/asm/pal.h 			rs		: 1,	/* RSE valid */
rs                698 arch/ia64/include/asm/pal.h #define pmci_proc_rse_valid			pme_processor.rs
rs                134 arch/ia64/include/asm/processor.h 		__u64 rs : 1;
rs                618 arch/ia64/kernel/unwind.c 	struct unw_reg_state *rs;
rs                620 arch/ia64/kernel/unwind.c 	rs = alloc_reg_state();
rs                621 arch/ia64/kernel/unwind.c 	if (!rs) {
rs                625 arch/ia64/kernel/unwind.c 	memcpy(rs, &sr->curr, sizeof(*rs));
rs                626 arch/ia64/kernel/unwind.c 	sr->curr.next = rs;
rs                632 arch/ia64/kernel/unwind.c 	struct unw_reg_state *rs = sr->curr.next;
rs                634 arch/ia64/kernel/unwind.c 	if (!rs) {
rs                638 arch/ia64/kernel/unwind.c 	memcpy(&sr->curr, rs, sizeof(*rs));
rs                639 arch/ia64/kernel/unwind.c 	free_reg_state(rs);
rs                644 arch/ia64/kernel/unwind.c dup_state_stack (struct unw_reg_state *rs)
rs                648 arch/ia64/kernel/unwind.c 	while (rs) {
rs                654 arch/ia64/kernel/unwind.c 		memcpy(copy, rs, sizeof(*copy));
rs                659 arch/ia64/kernel/unwind.c 		rs = rs->next;
rs                667 arch/ia64/kernel/unwind.c free_state_stack (struct unw_reg_state *rs)
rs                671 arch/ia64/kernel/unwind.c 	for (p = rs->next; p != NULL; p = next) {
rs                675 arch/ia64/kernel/unwind.c 	rs->next = NULL;
rs                557 arch/ia64/mm/discontig.c 	unsigned long rs, re, end = start + len;
rs                576 arch/ia64/mm/discontig.c 		rs = max(start, node_memblk[i].start_paddr);
rs                580 arch/ia64/mm/discontig.c 		if (rs < re)
rs                581 arch/ia64/mm/discontig.c 			(*func)(rs, re - rs, node_memblk[i].nid);
rs                186 arch/microblaze/include/asm/page.h #define tophys(rd, rs)	addik rd, rs, 0
rs                187 arch/microblaze/include/asm/page.h #define tovirt(rd, rs)	addik rd, rs, 0
rs                193 arch/microblaze/include/asm/page.h #define tophys(rd, rs) \
rs                194 arch/microblaze/include/asm/page.h 	addik rd, rs, (CONFIG_KERNEL_BASE_ADDR - CONFIG_KERNEL_START)
rs                195 arch/microblaze/include/asm/page.h #define tovirt(rd, rs) \
rs                196 arch/microblaze/include/asm/page.h 	addik rd, rs, (CONFIG_KERNEL_START - CONFIG_KERNEL_BASE_ADDR)
rs                 32 arch/mips/crypto/crc32-mips.c _ASM_MACRO_3R(OP, rt, rs, rt2,						  \
rs                101 arch/mips/dec/kn01-berr.c 			vaddr = regs->regs[insn.i_format.rs] +
rs                204 arch/mips/include/asm/asmmacro.h 	.macro	_EXT	rd, rs, p, s
rs                208 arch/mips/include/asm/asmmacro.h 	.macro	_EXT	rd, rs, p, s
rs                251 arch/mips/include/asm/asmmacro.h 	.macro	_ctcmsa	cd, rs
rs                382 arch/mips/include/asm/asmmacro.h 	.macro	_ctcmsa	cd, rs
rs                 16 arch/mips/include/asm/ginvt.h _ASM_MACRO_1R1I(ginvt, rs, type,
rs               1544 arch/mips/include/asm/mipsregs.h _ASM_MACRO_2R_1S(mfhc0, rt, rs, sel,
rs               1979 arch/mips/include/asm/mipsregs.h _ASM_MACRO_2R_1S(mfgc0, rt, rs, sel,
rs               1982 arch/mips/include/asm/mipsregs.h _ASM_MACRO_2R_1S(dmfgc0, rt, rs, sel,
rs                168 arch/mips/include/asm/msa.h _ASM_MACRO_2R(ctcmsa, cd, rs,
rs                130 arch/mips/include/asm/octeon/cvmx-asm.h 	asm ("pop %[rd],%[rs]" : [rd] "=d" (result) : [rs] "d" (input))
rs                132 arch/mips/include/asm/octeon/cvmx-asm.h 	asm ("dpop %[rd],%[rs]" : [rd] "=d" (result) : [rs] "d" (input))
rs                274 arch/mips/include/asm/octeon/cvmx-pip.h 		uint64_t rs:1;
rs                351 arch/mips/include/asm/octeon/cvmx-sriox-defs.h 		uint64_t rs:1;
rs                357 arch/mips/include/asm/octeon/cvmx-sriox-defs.h 		uint64_t rs:1;
rs                200 arch/mips/include/asm/uasm.h void UASM_i_LA_mostly(u32 **buf, unsigned int rs, long addr);
rs                201 arch/mips/include/asm/uasm.h void UASM_i_LA(u32 **buf, unsigned int rs, long addr);
rs                211 arch/mips/include/asm/uasm.h # define UASM_i_ADDIU(buf, rs, rt, val) uasm_i_daddiu(buf, rs, rt, val)
rs                212 arch/mips/include/asm/uasm.h # define UASM_i_ADDU(buf, rs, rt, rd) uasm_i_daddu(buf, rs, rt, rd)
rs                213 arch/mips/include/asm/uasm.h # define UASM_i_LL(buf, rs, rt, off) uasm_i_lld(buf, rs, rt, off)
rs                214 arch/mips/include/asm/uasm.h # define UASM_i_LW(buf, rs, rt, off) uasm_i_ld(buf, rs, rt, off)
rs                215 arch/mips/include/asm/uasm.h # define UASM_i_LWX(buf, rs, rt, rd) uasm_i_ldx(buf, rs, rt, rd)
rs                218 arch/mips/include/asm/uasm.h # define UASM_i_ROTR(buf, rs, rt, sh) uasm_i_drotr(buf, rs, rt, sh)
rs                219 arch/mips/include/asm/uasm.h # define UASM_i_SC(buf, rs, rt, off) uasm_i_scd(buf, rs, rt, off)
rs                220 arch/mips/include/asm/uasm.h # define UASM_i_SLL(buf, rs, rt, sh) uasm_i_dsll(buf, rs, rt, sh)
rs                221 arch/mips/include/asm/uasm.h # define UASM_i_SRA(buf, rs, rt, sh) uasm_i_dsra(buf, rs, rt, sh)
rs                222 arch/mips/include/asm/uasm.h # define UASM_i_SRL(buf, rs, rt, sh) uasm_i_dsrl(buf, rs, rt, sh)
rs                223 arch/mips/include/asm/uasm.h # define UASM_i_SRL_SAFE(buf, rs, rt, sh) uasm_i_dsrl_safe(buf, rs, rt, sh)
rs                224 arch/mips/include/asm/uasm.h # define UASM_i_SUBU(buf, rs, rt, rd) uasm_i_dsubu(buf, rs, rt, rd)
rs                225 arch/mips/include/asm/uasm.h # define UASM_i_SW(buf, rs, rt, off) uasm_i_sd(buf, rs, rt, off)
rs                227 arch/mips/include/asm/uasm.h # define UASM_i_ADDIU(buf, rs, rt, val) uasm_i_addiu(buf, rs, rt, val)
rs                228 arch/mips/include/asm/uasm.h # define UASM_i_ADDU(buf, rs, rt, rd) uasm_i_addu(buf, rs, rt, rd)
rs                229 arch/mips/include/asm/uasm.h # define UASM_i_LL(buf, rs, rt, off) uasm_i_ll(buf, rs, rt, off)
rs                230 arch/mips/include/asm/uasm.h # define UASM_i_LW(buf, rs, rt, off) uasm_i_lw(buf, rs, rt, off)
rs                231 arch/mips/include/asm/uasm.h # define UASM_i_LWX(buf, rs, rt, rd) uasm_i_lwx(buf, rs, rt, rd)
rs                234 arch/mips/include/asm/uasm.h # define UASM_i_ROTR(buf, rs, rt, sh) uasm_i_rotr(buf, rs, rt, sh)
rs                235 arch/mips/include/asm/uasm.h # define UASM_i_SC(buf, rs, rt, off) uasm_i_sc(buf, rs, rt, off)
rs                236 arch/mips/include/asm/uasm.h # define UASM_i_SLL(buf, rs, rt, sh) uasm_i_sll(buf, rs, rt, sh)
rs                237 arch/mips/include/asm/uasm.h # define UASM_i_SRA(buf, rs, rt, sh) uasm_i_sra(buf, rs, rt, sh)
rs                238 arch/mips/include/asm/uasm.h # define UASM_i_SRL(buf, rs, rt, sh) uasm_i_srl(buf, rs, rt, sh)
rs                239 arch/mips/include/asm/uasm.h # define UASM_i_SRL_SAFE(buf, rs, rt, sh) uasm_i_srl(buf, rs, rt, sh)
rs                240 arch/mips/include/asm/uasm.h # define UASM_i_SUBU(buf, rs, rt, rd) uasm_i_subu(buf, rs, rt, rd)
rs                241 arch/mips/include/asm/uasm.h # define UASM_i_SW(buf, rs, rt, off) uasm_i_sw(buf, rs, rt, off)
rs                245 arch/mips/include/asm/uasm.h #define uasm_i_beqz(buf, rs, off) uasm_i_beq(buf, rs, 0, off)
rs                246 arch/mips/include/asm/uasm.h #define uasm_i_beqzl(buf, rs, off) uasm_i_beql(buf, rs, 0, off)
rs                247 arch/mips/include/asm/uasm.h #define uasm_i_bnez(buf, rs, off) uasm_i_bne(buf, rs, 0, off)
rs                248 arch/mips/include/asm/uasm.h #define uasm_i_bnezl(buf, rs, off) uasm_i_bnel(buf, rs, 0, off)
rs                634 arch/mips/include/uapi/asm/inst.h 	__BITFIELD_FIELD(unsigned int rs : 5,
rs                642 arch/mips/include/uapi/asm/inst.h 	__BITFIELD_FIELD(unsigned int rs : 5,
rs                650 arch/mips/include/uapi/asm/inst.h 	__BITFIELD_FIELD(unsigned int rs : 5,
rs                659 arch/mips/include/uapi/asm/inst.h 	__BITFIELD_FIELD(unsigned int rs : 5,
rs                669 arch/mips/include/uapi/asm/inst.h 	__BITFIELD_FIELD(unsigned int rs : 5,
rs                679 arch/mips/include/uapi/asm/inst.h 	__BITFIELD_FIELD(unsigned int rs : 5,
rs                699 arch/mips/include/uapi/asm/inst.h 	__BITFIELD_FIELD(unsigned int rs : 5,
rs                738 arch/mips/include/uapi/asm/inst.h 	__BITFIELD_FIELD(unsigned int rs : 5,
rs                760 arch/mips/include/uapi/asm/inst.h 	__BITFIELD_FIELD(unsigned int rs : 5,
rs                779 arch/mips/include/uapi/asm/inst.h 	__BITFIELD_FIELD(unsigned int rs:5,
rs                909 arch/mips/include/uapi/asm/inst.h 	__BITFIELD_FIELD(unsigned int rs : 5,
rs                934 arch/mips/include/uapi/asm/inst.h 	__BITFIELD_FIELD(unsigned int rs : 3,
rs                951 arch/mips/include/uapi/asm/inst.h 	__BITFIELD_FIELD(unsigned int rs : 3,
rs                 81 arch/mips/kernel/branch.c 				*contpc = regs->regs[insn.mm_i_format.rs];
rs                 95 arch/mips/kernel/branch.c 			if ((long)regs->regs[insn.mm_i_format.rs] < 0)
rs                111 arch/mips/kernel/branch.c 			if ((long)regs->regs[insn.mm_i_format.rs] >= 0)
rs                121 arch/mips/kernel/branch.c 			if ((long)regs->regs[insn.mm_i_format.rs] <= 0)
rs                131 arch/mips/kernel/branch.c 			if ((long)regs->regs[insn.mm_i_format.rs] <= 0)
rs                160 arch/mips/kernel/branch.c 			bit = (insn.mm_i_format.rs >> 2);
rs                183 arch/mips/kernel/branch.c 			*contpc = regs->regs[insn.mm_i_format.rs];
rs                188 arch/mips/kernel/branch.c 		if ((long)regs->regs[reg16to32map[insn.mm_b1_format.rs]] == 0)
rs                197 arch/mips/kernel/branch.c 		if ((long)regs->regs[reg16to32map[insn.mm_b1_format.rs]] != 0)
rs                210 arch/mips/kernel/branch.c 		if (regs->regs[insn.mm_i_format.rs] ==
rs                221 arch/mips/kernel/branch.c 		if (regs->regs[insn.mm_i_format.rs] !=
rs                439 arch/mips/kernel/branch.c 			regs->cp0_epc = regs->regs[insn.r_format.rs];
rs                456 arch/mips/kernel/branch.c 			if ((long)regs->regs[insn.i_format.rs] < 0) {
rs                470 arch/mips/kernel/branch.c 			if ((long)regs->regs[insn.i_format.rs] >= 0) {
rs                481 arch/mips/kernel/branch.c 			if (NO_R6EMU && (insn.i_format.rs ||
rs                491 arch/mips/kernel/branch.c 			if (!insn.i_format.rs) {
rs                502 arch/mips/kernel/branch.c 			if ((long)regs->regs[insn.i_format.rs] < 0) {
rs                513 arch/mips/kernel/branch.c 			if (NO_R6EMU && (insn.i_format.rs ||
rs                523 arch/mips/kernel/branch.c 			if (!insn.i_format.rs) {
rs                534 arch/mips/kernel/branch.c 			if ((long)regs->regs[insn.i_format.rs] >= 0) {
rs                583 arch/mips/kernel/branch.c 		if (regs->regs[insn.i_format.rs] ==
rs                598 arch/mips/kernel/branch.c 		if (regs->regs[insn.i_format.rs] !=
rs                628 arch/mips/kernel/branch.c 			    ((!insn.i_format.rs && insn.i_format.rt) ||
rs                629 arch/mips/kernel/branch.c 			     (insn.i_format.rs == insn.i_format.rt)))
rs                635 arch/mips/kernel/branch.c 		if ((long)regs->regs[insn.i_format.rs] <= 0) {
rs                664 arch/mips/kernel/branch.c 			    ((!insn.i_format.rs && insn.i_format.rt) ||
rs                665 arch/mips/kernel/branch.c 			    (insn.i_format.rs == insn.i_format.rt)))
rs                672 arch/mips/kernel/branch.c 		if ((long)regs->regs[insn.i_format.rs] > 0) {
rs                689 arch/mips/kernel/branch.c 		    ((insn.i_format.rs == bc1eqz_op) ||
rs                690 arch/mips/kernel/branch.c 		     (insn.i_format.rs == bc1nez_op))) {
rs                695 arch/mips/kernel/branch.c 			if (insn.i_format.rs == bc1eqz_op)
rs                750 arch/mips/kernel/branch.c 		if ((regs->regs[insn.i_format.rs] & (1ull<<insn.i_format.rt))
rs                758 arch/mips/kernel/branch.c 		if ((regs->regs[insn.i_format.rs] &
rs                766 arch/mips/kernel/branch.c 		if (regs->regs[insn.i_format.rs] & (1ull<<insn.i_format.rt))
rs                773 arch/mips/kernel/branch.c 		if (regs->regs[insn.i_format.rs] &
rs                805 arch/mips/kernel/branch.c 		if (!insn.i_format.rs) {
rs                821 arch/mips/kernel/branch.c 		if (insn.i_format.rt && !insn.i_format.rs)
rs                405 arch/mips/kernel/mips-r2-to-r6-emul.c 	s32 rt, rs;
rs                408 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = regs->regs[MIPSInst_RS(ir)];
rs                409 arch/mips/kernel/mips-r2-to-r6-emul.c 	res = (s64)rt * (s64)rs;
rs                411 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = res;
rs                412 arch/mips/kernel/mips-r2-to-r6-emul.c 	regs->lo = (s64)rs;
rs                432 arch/mips/kernel/mips-r2-to-r6-emul.c 	u32 rt, rs;
rs                435 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = regs->regs[MIPSInst_RS(ir)];
rs                436 arch/mips/kernel/mips-r2-to-r6-emul.c 	res = (u64)rt * (u64)rs;
rs                455 arch/mips/kernel/mips-r2-to-r6-emul.c 	s32 rt, rs;
rs                458 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = regs->regs[MIPSInst_RS(ir)];
rs                460 arch/mips/kernel/mips-r2-to-r6-emul.c 	regs->lo = (s64)(rs / rt);
rs                461 arch/mips/kernel/mips-r2-to-r6-emul.c 	regs->hi = (s64)(rs % rt);
rs                477 arch/mips/kernel/mips-r2-to-r6-emul.c 	u32 rt, rs;
rs                480 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = regs->regs[MIPSInst_RS(ir)];
rs                482 arch/mips/kernel/mips-r2-to-r6-emul.c 	regs->lo = (s64)(rs / rt);
rs                483 arch/mips/kernel/mips-r2-to-r6-emul.c 	regs->hi = (s64)(rs % rt);
rs                500 arch/mips/kernel/mips-r2-to-r6-emul.c 	s64 rt, rs;
rs                506 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = regs->regs[MIPSInst_RS(ir)];
rs                507 arch/mips/kernel/mips-r2-to-r6-emul.c 	res = rt * rs;
rs                513 arch/mips/kernel/mips-r2-to-r6-emul.c 		: "r"(rt), "r"(rs));
rs                532 arch/mips/kernel/mips-r2-to-r6-emul.c 	u64 rt, rs;
rs                538 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = regs->regs[MIPSInst_RS(ir)];
rs                539 arch/mips/kernel/mips-r2-to-r6-emul.c 	res = rt * rs;
rs                545 arch/mips/kernel/mips-r2-to-r6-emul.c 		: "r"(rt), "r"(rs));
rs                563 arch/mips/kernel/mips-r2-to-r6-emul.c 	s64 rt, rs;
rs                569 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = regs->regs[MIPSInst_RS(ir)];
rs                571 arch/mips/kernel/mips-r2-to-r6-emul.c 	regs->lo = rs / rt;
rs                572 arch/mips/kernel/mips-r2-to-r6-emul.c 	regs->hi = rs % rt;
rs                588 arch/mips/kernel/mips-r2-to-r6-emul.c 	u64 rt, rs;
rs                594 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = regs->regs[MIPSInst_RS(ir)];
rs                596 arch/mips/kernel/mips-r2-to-r6-emul.c 	regs->lo = rs / rt;
rs                597 arch/mips/kernel/mips-r2-to-r6-emul.c 	regs->hi = rs % rt;
rs                636 arch/mips/kernel/mips-r2-to-r6-emul.c 	s32 rt, rs;
rs                639 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = regs->regs[MIPSInst_RS(ir)];
rs                640 arch/mips/kernel/mips-r2-to-r6-emul.c 	res = (s64)rt * (s64)rs;
rs                642 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = regs->lo;
rs                643 arch/mips/kernel/mips-r2-to-r6-emul.c 	res += ((((s64)rt) << 32) | (u32)rs);
rs                647 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = res >> 32;
rs                648 arch/mips/kernel/mips-r2-to-r6-emul.c 	regs->hi = (s64)rs;
rs                665 arch/mips/kernel/mips-r2-to-r6-emul.c 	u32 rt, rs;
rs                668 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = regs->regs[MIPSInst_RS(ir)];
rs                669 arch/mips/kernel/mips-r2-to-r6-emul.c 	res = (u64)rt * (u64)rs;
rs                671 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = regs->lo;
rs                672 arch/mips/kernel/mips-r2-to-r6-emul.c 	res += ((((s64)rt) << 32) | (u32)rs);
rs                676 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = res >> 32;
rs                677 arch/mips/kernel/mips-r2-to-r6-emul.c 	regs->hi = (s64)(s32)rs;
rs                694 arch/mips/kernel/mips-r2-to-r6-emul.c 	s32 rt, rs;
rs                697 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = regs->regs[MIPSInst_RS(ir)];
rs                698 arch/mips/kernel/mips-r2-to-r6-emul.c 	res = (s64)rt * (s64)rs;
rs                700 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = regs->lo;
rs                701 arch/mips/kernel/mips-r2-to-r6-emul.c 	res = ((((s64)rt) << 32) | (u32)rs) - res;
rs                705 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = res >> 32;
rs                706 arch/mips/kernel/mips-r2-to-r6-emul.c 	regs->hi = (s64)rs;
rs                723 arch/mips/kernel/mips-r2-to-r6-emul.c 	u32 rt, rs;
rs                726 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = regs->regs[MIPSInst_RS(ir)];
rs                727 arch/mips/kernel/mips-r2-to-r6-emul.c 	res = (u64)rt * (u64)rs;
rs                729 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = regs->lo;
rs                730 arch/mips/kernel/mips-r2-to-r6-emul.c 	res = ((((s64)rt) << 32) | (u32)rs) - res;
rs                734 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = res >> 32;
rs                735 arch/mips/kernel/mips-r2-to-r6-emul.c 	regs->hi = (s64)(s32)rs;
rs                752 arch/mips/kernel/mips-r2-to-r6-emul.c 	s32 rt, rs;
rs                757 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = regs->regs[MIPSInst_RS(ir)];
rs                758 arch/mips/kernel/mips-r2-to-r6-emul.c 	res = (s64)rt * (s64)rs;
rs                760 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = res;
rs                761 arch/mips/kernel/mips-r2-to-r6-emul.c 	regs->regs[MIPSInst_RD(ir)] = (s64)rs;
rs                778 arch/mips/kernel/mips-r2-to-r6-emul.c 	u32 rs;
rs                783 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = regs->regs[MIPSInst_RS(ir)];
rs                784 arch/mips/kernel/mips-r2-to-r6-emul.c 	__asm__ __volatile__("clz %0, %1" : "=r"(res) : "r"(rs));
rs                803 arch/mips/kernel/mips-r2-to-r6-emul.c 	u32 rs;
rs                808 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = regs->regs[MIPSInst_RS(ir)];
rs                809 arch/mips/kernel/mips-r2-to-r6-emul.c 	__asm__ __volatile__("clo %0, %1" : "=r"(res) : "r"(rs));
rs                827 arch/mips/kernel/mips-r2-to-r6-emul.c 	u64 rs;
rs                835 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = regs->regs[MIPSInst_RS(ir)];
rs                836 arch/mips/kernel/mips-r2-to-r6-emul.c 	__asm__ __volatile__("dclz %0, %1" : "=r"(res) : "r"(rs));
rs                854 arch/mips/kernel/mips-r2-to-r6-emul.c 	u64 rs;
rs                862 arch/mips/kernel/mips-r2-to-r6-emul.c 	rs = regs->regs[MIPSInst_RS(ir)];
rs                863 arch/mips/kernel/mips-r2-to-r6-emul.c 	__asm__ __volatile__("dclo %0, %1" : "=r"(res) : "r"(rs));
rs                911 arch/mips/kernel/mips-r2-to-r6-emul.c 	unsigned long cpc, epc, nepc, r31, res, rs, rt;
rs                941 arch/mips/kernel/mips-r2-to-r6-emul.c 		rs = MIPSInst_RS(inst);
rs                944 arch/mips/kernel/mips-r2-to-r6-emul.c 			if ((long)regs->regs[rs] >= MIPSInst_SIMM(inst))
rs                951 arch/mips/kernel/mips-r2-to-r6-emul.c 			if (regs->regs[rs] >= MIPSInst_UIMM(inst))
rs                958 arch/mips/kernel/mips-r2-to-r6-emul.c 			if ((long)regs->regs[rs] < MIPSInst_SIMM(inst))
rs                965 arch/mips/kernel/mips-r2-to-r6-emul.c 			if (regs->regs[rs] < MIPSInst_UIMM(inst))
rs                972 arch/mips/kernel/mips-r2-to-r6-emul.c 			if (regs->regs[rs] == MIPSInst_SIMM(inst))
rs                979 arch/mips/kernel/mips-r2-to-r6-emul.c 			if (regs->regs[rs] != MIPSInst_SIMM(inst))
rs               1267 arch/mips/kernel/mips-r2-to-r6-emul.c 			: "+&r"(rt), "=&r"(rs),
rs               1342 arch/mips/kernel/mips-r2-to-r6-emul.c 			: "+&r"(rt), "=&r"(rs),
rs               1413 arch/mips/kernel/mips-r2-to-r6-emul.c 			: "+&r"(rt), "=&r"(rs),
rs               1483 arch/mips/kernel/mips-r2-to-r6-emul.c 			: "+&r"(rt), "=&r"(rs),
rs               1602 arch/mips/kernel/mips-r2-to-r6-emul.c 			: "+&r"(rt), "=&r"(rs),
rs               1721 arch/mips/kernel/mips-r2-to-r6-emul.c 			: "+&r"(rt), "=&r"(rs),
rs               1840 arch/mips/kernel/mips-r2-to-r6-emul.c 			: "+&r"(rt), "=&r"(rs),
rs               1958 arch/mips/kernel/mips-r2-to-r6-emul.c 			: "+&r"(rt), "=&r"(rs),
rs                244 arch/mips/kernel/process.c 		if (ip->i_format.rs != 29)
rs                274 arch/mips/kernel/process.c 		ip->i_format.rs == 29 && ip->i_format.rt == 31) {
rs                353 arch/mips/kernel/process.c 	    ip->mm_i_format.rt == 29 && ip->mm_i_format.rs == 29) {
rs                359 arch/mips/kernel/process.c 	if (ip->i_format.rs != 29 || ip->i_format.rt != 29)
rs                 87 arch/mips/kvm/dyntrans.c 	synci_inst.i_format.rs = inst.i_format.rs;
rs                 69 arch/mips/kvm/emulate.c 			nextpc = arch->gprs[insn.r_format.rs];
rs                 85 arch/mips/kvm/emulate.c 			if ((long)arch->gprs[insn.i_format.rs] < 0)
rs                 94 arch/mips/kvm/emulate.c 			if ((long)arch->gprs[insn.i_format.rs] >= 0)
rs                104 arch/mips/kvm/emulate.c 			if ((long)arch->gprs[insn.i_format.rs] < 0)
rs                114 arch/mips/kvm/emulate.c 			if ((long)arch->gprs[insn.i_format.rs] >= 0)
rs                155 arch/mips/kvm/emulate.c 		if (arch->gprs[insn.i_format.rs] ==
rs                165 arch/mips/kvm/emulate.c 		if (arch->gprs[insn.i_format.rs] !=
rs                179 arch/mips/kvm/emulate.c 		if ((long)arch->gprs[insn.i_format.rs] <= 0)
rs                192 arch/mips/kvm/emulate.c 		if ((long)arch->gprs[insn.i_format.rs] > 0)
rs                215 arch/mips/kvm/emulate.c 		if (insn.i_format.rs != 0 || insn.i_format.rt != 0)
rs                221 arch/mips/kvm/emulate.c 		if (insn.i_format.rs != 0)
rs               1314 arch/mips/kvm/emulate.c 		switch (inst.c0r_format.rs) {
rs               1577 arch/mips/kvm/emulate.c 				vcpu->arch.pc, inst.c0r_format.rs);
rs               1818 arch/mips/kvm/emulate.c 	base = inst.i_format.rs;
rs               2515 arch/mips/kvm/emulate.c 	    inst.r_format.rs == 0 &&
rs                933 arch/mips/kvm/vz.c 		switch (inst.c0r_format.rs) {
rs                984 arch/mips/kvm/vz.c 				if (inst.c0r_format.rs == mfc_op)
rs                989 arch/mips/kvm/vz.c 			trace_kvm_hwr(vcpu, (inst.c0r_format.rs == mfc_op) ?
rs               1000 arch/mips/kvm/vz.c 			trace_kvm_hwr(vcpu, (inst.c0r_format.rs == mtc_op) ?
rs               1026 arch/mips/kvm/vz.c 				val = mips_process_maar(inst.c0r_format.rs,
rs               1083 arch/mips/kvm/vz.c 	base = inst.i_format.rs;
rs               1170 arch/mips/kvm/vz.c 			if (inst.r_format.rs || (inst.r_format.re >> 3))
rs               1229 arch/mips/kvm/vz.c 	    inst.c0r_format.rs == mtc_op &&
rs                 67 arch/mips/lasat/picvue.c 	data &= ~picvue->rs;
rs                 71 arch/mips/lasat/picvue.c 	data |= picvue->rs;
rs                 97 arch/mips/lasat/picvue.c 		data |= picvue->rs;
rs                 99 arch/mips/lasat/picvue.c 		data &= ~picvue->rs;
rs                104 arch/mips/lasat/picvue.c 		data &= ~picvue->rs;
rs                106 arch/mips/lasat/picvue.c 		data |= picvue->rs;
rs                 14 arch/mips/lasat/picvue.h 	u32 rs;
rs                 91 arch/mips/math-emu/cp1emu.c 		mips32_insn.mm_i_format.rt = insn.mm_i_format.rs;
rs                 92 arch/mips/math-emu/cp1emu.c 		mips32_insn.mm_i_format.rs = insn.mm_i_format.rt;
rs                 96 arch/mips/math-emu/cp1emu.c 		mips32_insn.mm_i_format.rt = insn.mm_i_format.rs;
rs                 97 arch/mips/math-emu/cp1emu.c 		mips32_insn.mm_i_format.rs = insn.mm_i_format.rt;
rs                101 arch/mips/math-emu/cp1emu.c 		mips32_insn.mm_i_format.rt = insn.mm_i_format.rs;
rs                102 arch/mips/math-emu/cp1emu.c 		mips32_insn.mm_i_format.rs = insn.mm_i_format.rt;
rs                106 arch/mips/math-emu/cp1emu.c 		mips32_insn.mm_i_format.rt = insn.mm_i_format.rs;
rs                107 arch/mips/math-emu/cp1emu.c 		mips32_insn.mm_i_format.rs = insn.mm_i_format.rt;
rs                168 arch/mips/math-emu/cp1emu.c 				mips32_insn.r_format.rs =
rs                254 arch/mips/math-emu/cp1emu.c 				mips32_insn.r_format.rs = insn.mm_fp4_format.fs;
rs                447 arch/mips/math-emu/cp1emu.c 			*contpc = regs->regs[insn.r_format.rs];
rs                455 arch/mips/math-emu/cp1emu.c 			if (NO_R6EMU && (insn.i_format.rs ||
rs                468 arch/mips/math-emu/cp1emu.c 			if ((long)regs->regs[insn.i_format.rs] < 0)
rs                479 arch/mips/math-emu/cp1emu.c 			if (NO_R6EMU && (insn.i_format.rs ||
rs                492 arch/mips/math-emu/cp1emu.c 			if ((long)regs->regs[insn.i_format.rs] >= 0)
rs                524 arch/mips/math-emu/cp1emu.c 		if (regs->regs[insn.i_format.rs] ==
rs                539 arch/mips/math-emu/cp1emu.c 		if (regs->regs[insn.i_format.rs] !=
rs                569 arch/mips/math-emu/cp1emu.c 			    ((!insn.i_format.rs && insn.i_format.rt) ||
rs                570 arch/mips/math-emu/cp1emu.c 			     (insn.i_format.rs == insn.i_format.rt)))
rs                578 arch/mips/math-emu/cp1emu.c 		if ((long)regs->regs[insn.i_format.rs] <= 0)
rs                607 arch/mips/math-emu/cp1emu.c 			    ((!insn.i_format.rs && insn.i_format.rt) ||
rs                608 arch/mips/math-emu/cp1emu.c 			     (insn.i_format.rs == insn.i_format.rt)))
rs                617 arch/mips/math-emu/cp1emu.c 		if ((long)regs->regs[insn.i_format.rs] > 0)
rs                630 arch/mips/math-emu/cp1emu.c 		if (insn.i_format.rt && !insn.i_format.rs)
rs                638 arch/mips/math-emu/cp1emu.c 		if ((regs->regs[insn.i_format.rs] & (1ull<<insn.i_format.rt)) == 0)
rs                644 arch/mips/math-emu/cp1emu.c 		if ((regs->regs[insn.i_format.rs] & (1ull<<(insn.i_format.rt + 32))) == 0)
rs                650 arch/mips/math-emu/cp1emu.c 		if (regs->regs[insn.i_format.rs] & (1ull<<insn.i_format.rt))
rs                656 arch/mips/math-emu/cp1emu.c 		if (regs->regs[insn.i_format.rs] & (1ull<<(insn.i_format.rt + 32)))
rs                692 arch/mips/math-emu/cp1emu.c 		if (!insn.i_format.rs)
rs                703 arch/mips/math-emu/cp1emu.c 		    ((insn.i_format.rs == bc1eqz_op) ||
rs                704 arch/mips/math-emu/cp1emu.c 		     (insn.i_format.rs == bc1nez_op))) {
rs                708 arch/mips/math-emu/cp1emu.c 			switch (insn.i_format.rs) {
rs                731 arch/mips/math-emu/cp1emu.c 		if (insn.i_format.rs == bc_op) {
rs                 44 arch/mips/math-emu/dp_maddf.c 	int rs;
rs                190 arch/mips/math-emu/dp_maddf.c 	rs = xs ^ ys;
rs                192 arch/mips/math-emu/dp_maddf.c 		rs ^= 1;
rs                242 arch/mips/math-emu/dp_maddf.c 		return ieee754dp_format(rs, re, lrm);
rs                270 arch/mips/math-emu/dp_maddf.c 	if (zs == rs) {
rs                288 arch/mips/math-emu/dp_maddf.c 			zs = rs;
rs                 15 arch/mips/math-emu/dp_mul.c 	int rs;
rs                110 arch/mips/math-emu/dp_mul.c 	rs = xs ^ ys;
rs                159 arch/mips/math-emu/dp_mul.c 	return ieee754dp_format(rs, re, rm);
rs                235 arch/mips/math-emu/dsemul.c 			unsigned int rs;
rs                238 arch/mips/math-emu/dsemul.c 			rs = (((insn.mm_a_format.rs + 0xe) & 0xf) + 2);
rs                241 arch/mips/math-emu/dsemul.c 			regs->regs[rs] = (long)v;
rs                 37 arch/mips/math-emu/ieee754dp.h #define XDPSRS(v,rs)	\
rs                 38 arch/mips/math-emu/ieee754dp.h 	((rs > (DP_FBITS+3))?1:((v) >> (rs)) | ((v) << (64-(rs)) != 0))
rs                 37 arch/mips/math-emu/ieee754sp.h #define XSPSRS64(v, rs)						\
rs                 38 arch/mips/math-emu/ieee754sp.h 	(((rs) >= 64) ? ((v) != 0) : ((v) >> (rs)) | ((v) << (64-(rs)) != 0))
rs                 41 arch/mips/math-emu/ieee754sp.h #define XSPSRS(v, rs)						\
rs                 42 arch/mips/math-emu/ieee754sp.h 	((rs > (SP_FBITS+3))?1:((v) >> (rs)) | ((v) << (32-(rs)) != 0))
rs                 19 arch/mips/math-emu/sp_maddf.c 	int rs;
rs                161 arch/mips/math-emu/sp_maddf.c 	rs = xs ^ ys;
rs                163 arch/mips/math-emu/sp_maddf.c 		rs ^= 1;
rs                186 arch/mips/math-emu/sp_maddf.c 		return ieee754sp_format(rs, re, rm);
rs                213 arch/mips/math-emu/sp_maddf.c 	if (zs == rs) {
rs                228 arch/mips/math-emu/sp_maddf.c 			zs = rs;
rs                 15 arch/mips/math-emu/sp_mul.c 	int rs;
rs                110 arch/mips/math-emu/sp_mul.c 	rs = xs ^ ys;
rs                154 arch/mips/math-emu/sp_mul.c 	return ieee754sp_format(rs, re, rm);
rs                457 arch/mips/mm/uasm.c void UASM_i_LA_mostly(u32 **buf, unsigned int rs, long addr)
rs                460 arch/mips/mm/uasm.c 		uasm_i_lui(buf, rs, uasm_rel_highest(addr));
rs                462 arch/mips/mm/uasm.c 			uasm_i_daddiu(buf, rs, rs, uasm_rel_higher(addr));
rs                464 arch/mips/mm/uasm.c 			uasm_i_dsll(buf, rs, rs, 16);
rs                465 arch/mips/mm/uasm.c 			uasm_i_daddiu(buf, rs, rs,
rs                467 arch/mips/mm/uasm.c 			uasm_i_dsll(buf, rs, rs, 16);
rs                469 arch/mips/mm/uasm.c 			uasm_i_dsll32(buf, rs, rs, 0);
rs                471 arch/mips/mm/uasm.c 		uasm_i_lui(buf, rs, uasm_rel_hi(addr));
rs                475 arch/mips/mm/uasm.c void UASM_i_LA(u32 **buf, unsigned int rs, long addr)
rs                477 arch/mips/mm/uasm.c 	UASM_i_LA_mostly(buf, rs, addr);
rs                480 arch/mips/mm/uasm.c 			uasm_i_daddiu(buf, rs, rs,
rs                483 arch/mips/mm/uasm.c 			uasm_i_addiu(buf, rs, rs,
rs                 36 arch/mips/oprofile/backtrace.c 		&& ip->i_format.rs == 29 && ip->i_format.rt == 31;
rs                 42 arch/mips/oprofile/backtrace.c 	if (ip->i_format.rs != 29 || ip->i_format.rt != 29)
rs                 56 arch/mips/oprofile/backtrace.c 	if (ip->r_format.func == jr_op && ip->r_format.rs == 31)
rs                 44 arch/mips/txx9/generic/mem_tx4927.c 	unsigned int rs = 0;
rs                 59 arch/mips/txx9/generic/mem_tx4927.c 		rs = 2048 << sdccr_rs;
rs                 64 arch/mips/txx9/generic/mem_tx4927.c 	return rs * cs * mw * bs;
rs                393 arch/powerpc/include/asm/kvm_host.h 	void (*slbmte)(struct kvm_vcpu *vcpu, u64 rb, u64 rs);
rs                 78 arch/powerpc/include/asm/kvm_ppc.h 		unsigned int rs, unsigned int bytes, int is_default_endian);
rs                 83 arch/powerpc/include/asm/kvm_ppc.h 				int rs, unsigned int bytes,
rs                492 arch/powerpc/include/asm/ppc-opcode.h #define	PPC_TLBIE_5(rb,rs,ric,prs,r) \
rs                494 arch/powerpc/include/asm/ppc-opcode.h 					___PPC_RB(rb) | ___PPC_RS(rs) | \
rs                497 arch/powerpc/include/asm/ppc-opcode.h #define	PPC_TLBIEL(rb,rs,ric,prs,r) \
rs                499 arch/powerpc/include/asm/ppc-opcode.h 					___PPC_RB(rb) | ___PPC_RS(rs) | \
rs                483 arch/powerpc/include/asm/ppc_asm.h #define tophys(rd,rs)				\
rs                484 arch/powerpc/include/asm/ppc_asm.h 	addis	rd,rs,0
rs                486 arch/powerpc/include/asm/ppc_asm.h #define tovirt(rd,rs)				\
rs                487 arch/powerpc/include/asm/ppc_asm.h 	addis	rd,rs,0
rs                493 arch/powerpc/include/asm/ppc_asm.h #define tophys(rd,rs)                           \
rs                494 arch/powerpc/include/asm/ppc_asm.h 	clrldi	rd,rs,2
rs                496 arch/powerpc/include/asm/ppc_asm.h #define tovirt(rd,rs)                           \
rs                497 arch/powerpc/include/asm/ppc_asm.h 	rotldi	rd,rs,16;			\
rs                504 arch/powerpc/include/asm/ppc_asm.h #define tophys(rd, rs)	addis	rd, rs, -PAGE_OFFSET@h
rs                505 arch/powerpc/include/asm/ppc_asm.h #define tovirt(rd, rs)	addis	rd, rs, PAGE_OFFSET@h
rs                191 arch/powerpc/include/asm/trace.h 		unsigned long rs, unsigned long ric, unsigned long prs,
rs                193 arch/powerpc/include/asm/trace.h 	TP_ARGS(lpid, local, rb, rs, ric, prs, r),
rs                198 arch/powerpc/include/asm/trace.h 		__field(unsigned long, rs)
rs                208 arch/powerpc/include/asm/trace.h 		__entry->rs = rs;
rs                216 arch/powerpc/include/asm/trace.h 		__entry->rb, __entry->rs, __entry->ric, __entry->prs,
rs                175 arch/powerpc/kernel/isa-bridge.c 	int rlen, i, rs;
rs                197 arch/powerpc/kernel/isa-bridge.c 	rs = na + ns + pna;
rs                201 arch/powerpc/kernel/isa-bridge.c 	if (ranges == NULL || rlen < rs) {
rs                208 arch/powerpc/kernel/isa-bridge.c 	for (i = 0; (i + rs - 1) < rlen; i += rs) {
rs                 78 arch/powerpc/kernel/legacy_serial.c 	const __be32 *clk, *spd, *rs;
rs                 92 arch/powerpc/kernel/legacy_serial.c 	rs = of_get_property(np, "reg-shift", NULL);
rs                 93 arch/powerpc/kernel/legacy_serial.c 	if (rs && *rs)
rs                 94 arch/powerpc/kernel/legacy_serial.c 		shift = be32_to_cpup(rs);
rs                313 arch/powerpc/kernel/traps.c 	static DEFINE_RATELIMIT_STATE(rs, DEFAULT_RATELIMIT_INTERVAL,
rs                322 arch/powerpc/kernel/traps.c 	if (!__ratelimit(&rs))
rs               1275 arch/powerpc/kernel/traps.c 	u32 ra,rs;
rs               1279 arch/powerpc/kernel/traps.c 	rs = (instword >> 21) & 0x1f;
rs               1281 arch/powerpc/kernel/traps.c 	tmp = regs->gpr[rs];
rs                378 arch/powerpc/kvm/book3s_64_mmu.c static void kvmppc_mmu_book3s_64_slbmte(struct kvm_vcpu *vcpu, u64 rs, u64 rb)
rs                384 arch/powerpc/kvm/book3s_64_mmu.c 	dprintk("KVM MMU: slbmte(0x%llx, 0x%llx)\n", rs, rb);
rs                395 arch/powerpc/kvm/book3s_64_mmu.c 	slbe->large = (rs & SLB_VSID_L) ? 1 : 0;
rs                396 arch/powerpc/kvm/book3s_64_mmu.c 	slbe->tb    = (rs & SLB_VSID_B_1T) ? 1 : 0;
rs                398 arch/powerpc/kvm/book3s_64_mmu.c 	slbe->vsid  = (rs & ~SLB_VSID_B) >> (kvmppc_slb_sid_shift(slbe) - 16);
rs                400 arch/powerpc/kvm/book3s_64_mmu.c 	slbe->Ks    = (rs & SLB_VSID_KS) ? 1 : 0;
rs                401 arch/powerpc/kvm/book3s_64_mmu.c 	slbe->Kp    = (rs & SLB_VSID_KP) ? 1 : 0;
rs                402 arch/powerpc/kvm/book3s_64_mmu.c 	slbe->nx    = (rs & SLB_VSID_N) ? 1 : 0;
rs                403 arch/powerpc/kvm/book3s_64_mmu.c 	slbe->class = (rs & SLB_VSID_C) ? 1 : 0;
rs                408 arch/powerpc/kvm/book3s_64_mmu.c 			switch (rs & SLB_VSID_LP) {
rs                421 arch/powerpc/kvm/book3s_64_mmu.c 	slbe->origv = rs;
rs                507 arch/powerpc/kvm/book3s_64_mmu.c 	u64 rb = 0, rs = 0;
rs                537 arch/powerpc/kvm/book3s_64_mmu.c 	rs |= (value & 0xfffffff) << 12;
rs                539 arch/powerpc/kvm/book3s_64_mmu.c 	rs |= ((value >> 28) & 0x7) << 9;
rs                541 arch/powerpc/kvm/book3s_64_mmu.c 	kvmppc_mmu_book3s_64_slbmte(vcpu, rs, rb);
rs                243 arch/powerpc/kvm/book3s_emulate.c 	int rs = get_rs(inst);
rs                303 arch/powerpc/kvm/book3s_emulate.c 			ulong rs_val = kvmppc_get_gpr(vcpu, rs);
rs                314 arch/powerpc/kvm/book3s_emulate.c 			kvmppc_set_msr(vcpu, kvmppc_get_gpr(vcpu, rs));
rs                343 arch/powerpc/kvm/book3s_emulate.c 				kvmppc_get_gpr(vcpu, rs));
rs                348 arch/powerpc/kvm/book3s_emulate.c 				kvmppc_get_gpr(vcpu, rs));
rs                393 arch/powerpc/kvm/book3s_emulate.c 					kvmppc_get_gpr(vcpu, rs),
rs                 57 arch/powerpc/kvm/book3s_hv_ras.c 		unsigned long rs = be64_to_cpu(slb->save_area[i].vsid);
rs                 60 arch/powerpc/kvm/book3s_hv_ras.c 		asm volatile("slbmte %0,%1" : : "r" (rs), "r" (rb));
rs                442 arch/powerpc/kvm/book3s_hv_rm_mmu.c 		unsigned long rb,rs,prs,r,ric;
rs                445 arch/powerpc/kvm/book3s_hv_rm_mmu.c 		rs = 0;  /* lpid = 0 */
rs                457 arch/powerpc/kvm/book3s_hv_rm_mmu.c 			       "i"(ric), "r"(rs) : "memory");
rs                 45 arch/powerpc/kvm/book3s_hv_tm.c 	int ra, rs;
rs                 96 arch/powerpc/kvm/book3s_hv_tm.c 		rs = (instr >> 21) & 0x1f;
rs                 97 arch/powerpc/kvm/book3s_hv_tm.c 		newmsr = kvmppc_get_gpr(vcpu, rs);
rs                 24 arch/powerpc/kvm/book3s_hv_tm_builtin.c 	int rs;
rs                 64 arch/powerpc/kvm/book3s_hv_tm_builtin.c 		rs = (instr >> 21) & 0x1f;
rs                 65 arch/powerpc/kvm/book3s_hv_tm_builtin.c 		newmsr = kvmppc_get_gpr(vcpu, rs);
rs                173 arch/powerpc/kvm/book3s_paired_singles.c 				   int rs, ulong addr, int ls_type)
rs                191 arch/powerpc/kvm/book3s_paired_singles.c 		emulated = kvmppc_handle_load(run, vcpu, KVM_MMIO_REG_FPR | rs,
rs                201 arch/powerpc/kvm/book3s_paired_singles.c 		kvm_cvt_fd((u32*)tmp, &VCPU_FPR(vcpu, rs));
rs                202 arch/powerpc/kvm/book3s_paired_singles.c 		vcpu->arch.qpr[rs] = *((u32*)tmp);
rs                205 arch/powerpc/kvm/book3s_paired_singles.c 		VCPU_FPR(vcpu, rs) = *((u64*)tmp);
rs                217 arch/powerpc/kvm/book3s_paired_singles.c 				    int rs, ulong addr, int ls_type)
rs                227 arch/powerpc/kvm/book3s_paired_singles.c 		kvm_cvt_df(&VCPU_FPR(vcpu, rs), (u32*)tmp);
rs                232 arch/powerpc/kvm/book3s_paired_singles.c 		*((u32*)tmp) = VCPU_FPR(vcpu, rs);
rs                233 arch/powerpc/kvm/book3s_paired_singles.c 		val = VCPU_FPR(vcpu, rs) & 0xffffffff;
rs                237 arch/powerpc/kvm/book3s_paired_singles.c 		*((u64*)tmp) = VCPU_FPR(vcpu, rs);
rs                238 arch/powerpc/kvm/book3s_paired_singles.c 		val = VCPU_FPR(vcpu, rs);
rs                263 arch/powerpc/kvm/book3s_paired_singles.c 				   int rs, ulong addr, bool w, int i)
rs                282 arch/powerpc/kvm/book3s_paired_singles.c 		emulated = kvmppc_handle_load(run, vcpu, KVM_MMIO_REG_FPR | rs,
rs                284 arch/powerpc/kvm/book3s_paired_singles.c 		vcpu->arch.qpr[rs] = tmp[1];
rs                287 arch/powerpc/kvm/book3s_paired_singles.c 		emulated = kvmppc_handle_load(run, vcpu, KVM_MMIO_REG_FQPR | rs,
rs                295 arch/powerpc/kvm/book3s_paired_singles.c 	kvm_cvt_fd(&tmp[0], &VCPU_FPR(vcpu, rs));
rs                296 arch/powerpc/kvm/book3s_paired_singles.c 	vcpu->arch.qpr[rs] = tmp[1];
rs                306 arch/powerpc/kvm/book3s_paired_singles.c 				    int rs, ulong addr, bool w, int i)
rs                313 arch/powerpc/kvm/book3s_paired_singles.c 	kvm_cvt_df(&VCPU_FPR(vcpu, rs), &tmp[0]);
rs                314 arch/powerpc/kvm/book3s_paired_singles.c 	tmp[1] = vcpu->arch.qpr[rs];
rs               1483 arch/powerpc/kvm/book3s_pr.c 			u64 rs = sregs->u.s.ppc64.slb[i].slbv;
rs               1486 arch/powerpc/kvm/book3s_pr.c 				vcpu->arch.mmu.slbmte(vcpu, rs, rb);
rs                 46 arch/powerpc/kvm/booke_emulate.c 	int rs = get_rs(inst);
rs                 86 arch/powerpc/kvm/booke_emulate.c 			kvmppc_set_msr(vcpu, kvmppc_get_gpr(vcpu, rs));
rs                 91 arch/powerpc/kvm/booke_emulate.c 					| (kvmppc_get_gpr(vcpu, rs) & MSR_EE);
rs                 75 arch/powerpc/kvm/emulate.c static int kvmppc_emulate_mtspr(struct kvm_vcpu *vcpu, int sprn, int rs)
rs                 78 arch/powerpc/kvm/emulate.c 	ulong spr_val = kvmppc_get_gpr(vcpu, rs);
rs                197 arch/powerpc/kvm/emulate.c 	int rs, rt, sprn;
rs                210 arch/powerpc/kvm/emulate.c 	rs = get_rs(inst);
rs                251 arch/powerpc/kvm/emulate.c 			emulated = kvmppc_emulate_mtspr(vcpu, sprn, rs);
rs               1372 arch/powerpc/kvm/powerpc.c static inline int kvmppc_get_vsr_data(struct kvm_vcpu *vcpu, int rs, u64 *val)
rs               1390 arch/powerpc/kvm/powerpc.c 		if (rs < 32) {
rs               1391 arch/powerpc/kvm/powerpc.c 			*val = VCPU_VSX_FPR(vcpu, rs, vsx_offset);
rs               1393 arch/powerpc/kvm/powerpc.c 			reg.vval = VCPU_VSX_VR(vcpu, rs - 32);
rs               1407 arch/powerpc/kvm/powerpc.c 		if (rs < 32) {
rs               1410 arch/powerpc/kvm/powerpc.c 			reg.vsxval[0] = VCPU_VSX_FPR(vcpu, rs, dword_offset);
rs               1413 arch/powerpc/kvm/powerpc.c 			reg.vval = VCPU_VSX_VR(vcpu, rs - 32);
rs               1427 arch/powerpc/kvm/powerpc.c 			int rs, unsigned int bytes, int is_default_endian)
rs               1432 arch/powerpc/kvm/powerpc.c 	vcpu->arch.io_gpr = rs;
rs               1439 arch/powerpc/kvm/powerpc.c 		if (kvmppc_get_vsr_data(vcpu, rs, &val) == -1)
rs               1589 arch/powerpc/kvm/powerpc.c 		unsigned int rs, unsigned int bytes, int is_default_endian)
rs               1592 arch/powerpc/kvm/powerpc.c 	unsigned int index = rs & KVM_MMIO_REG_MASK;
rs               1598 arch/powerpc/kvm/powerpc.c 	vcpu->arch.io_gpr = rs;
rs                 64 arch/powerpc/mm/book3s64/hash_native.c 	unsigned long rs;
rs                 68 arch/powerpc/mm/book3s64/hash_native.c 	rs = ((unsigned long)pid << PPC_BITLSHIFT(31));
rs                 71 arch/powerpc/mm/book3s64/hash_native.c 		     : : "r"(rb), "r"(rs), "i"(ric), "i"(prs), "r"(r)
rs                206 arch/powerpc/mm/book3s64/hash_native.c 		unsigned long rb,rs,prs,r,ric;
rs                209 arch/powerpc/mm/book3s64/hash_native.c 		rs = 0;  /* lpid = 0 */
rs                221 arch/powerpc/mm/book3s64/hash_native.c 			       "i"(ric), "r"(rs) : "memory");
rs                 33 arch/powerpc/mm/book3s64/radix_tlb.c 	unsigned long rs;
rs                 36 arch/powerpc/mm/book3s64/radix_tlb.c 	rs = ((unsigned long)pid << PPC_BITLSHIFT(31));
rs                 39 arch/powerpc/mm/book3s64/radix_tlb.c 		     : : "r"(rb), "r"(rs), "i"(ric), "i"(prs)
rs                 96 arch/powerpc/mm/book3s64/radix_tlb.c 	unsigned long rb,rs,prs,r;
rs                100 arch/powerpc/mm/book3s64/radix_tlb.c 	rs = ((unsigned long)pid) << PPC_BITLSHIFT(31);
rs                105 arch/powerpc/mm/book3s64/radix_tlb.c 		     : : "r"(rb), "i"(r), "i"(prs), "i"(ric), "r"(rs) : "memory");
rs                106 arch/powerpc/mm/book3s64/radix_tlb.c 	trace_tlbie(0, 1, rb, rs, ric, prs, r);
rs                111 arch/powerpc/mm/book3s64/radix_tlb.c 	unsigned long rb,rs,prs,r;
rs                114 arch/powerpc/mm/book3s64/radix_tlb.c 	rs = pid << PPC_BITLSHIFT(31);
rs                119 arch/powerpc/mm/book3s64/radix_tlb.c 		     : : "r"(rb), "i"(r), "i"(prs), "i"(ric), "r"(rs) : "memory");
rs                120 arch/powerpc/mm/book3s64/radix_tlb.c 	trace_tlbie(0, 0, rb, rs, ric, prs, r);
rs                125 arch/powerpc/mm/book3s64/radix_tlb.c 	unsigned long rb,rs,prs,r;
rs                128 arch/powerpc/mm/book3s64/radix_tlb.c 	rs = lpid;
rs                133 arch/powerpc/mm/book3s64/radix_tlb.c 		     : : "r"(rb), "i"(r), "i"(prs), "i"(ric), "r"(rs) : "memory");
rs                134 arch/powerpc/mm/book3s64/radix_tlb.c 	trace_tlbie(lpid, 0, rb, rs, ric, prs, r);
rs                139 arch/powerpc/mm/book3s64/radix_tlb.c 	unsigned long rb,rs,prs,r;
rs                142 arch/powerpc/mm/book3s64/radix_tlb.c 	rs = lpid;
rs                147 arch/powerpc/mm/book3s64/radix_tlb.c 		     : : "r"(rb), "i"(r), "i"(prs), "i"(ric), "r"(rs) : "memory");
rs                148 arch/powerpc/mm/book3s64/radix_tlb.c 	trace_tlbie(lpid, 0, rb, rs, ric, prs, r);
rs                154 arch/powerpc/mm/book3s64/radix_tlb.c 	unsigned long rb,rs,prs,r;
rs                158 arch/powerpc/mm/book3s64/radix_tlb.c 	rs = pid << PPC_BITLSHIFT(31);
rs                163 arch/powerpc/mm/book3s64/radix_tlb.c 		     : : "r"(rb), "i"(r), "i"(prs), "i"(ric), "r"(rs) : "memory");
rs                164 arch/powerpc/mm/book3s64/radix_tlb.c 	trace_tlbie(0, 1, rb, rs, ric, prs, r);
rs                170 arch/powerpc/mm/book3s64/radix_tlb.c 	unsigned long rb,rs,prs,r;
rs                174 arch/powerpc/mm/book3s64/radix_tlb.c 	rs = pid << PPC_BITLSHIFT(31);
rs                179 arch/powerpc/mm/book3s64/radix_tlb.c 		     : : "r"(rb), "i"(r), "i"(prs), "i"(ric), "r"(rs) : "memory");
rs                180 arch/powerpc/mm/book3s64/radix_tlb.c 	trace_tlbie(0, 0, rb, rs, ric, prs, r);
rs                186 arch/powerpc/mm/book3s64/radix_tlb.c 	unsigned long rb,rs,prs,r;
rs                190 arch/powerpc/mm/book3s64/radix_tlb.c 	rs = lpid;
rs                195 arch/powerpc/mm/book3s64/radix_tlb.c 		     : : "r"(rb), "i"(r), "i"(prs), "i"(ric), "r"(rs) : "memory");
rs                196 arch/powerpc/mm/book3s64/radix_tlb.c 	trace_tlbie(lpid, 0, rb, rs, ric, prs, r);
rs               1194 arch/powerpc/mm/book3s64/radix_tlb.c 	unsigned long rb,prs,r,rs;
rs               1200 arch/powerpc/mm/book3s64/radix_tlb.c 	rs = 1 & ((1UL << 32) - 1); /* any LPID value to flush guest mappings */
rs               1207 arch/powerpc/mm/book3s64/radix_tlb.c 		     : : "r"(rb), "i"(r), "i"(1), "i"(ric), "r"(rs) : "memory");
rs                675 arch/riscv/net/bpf_jit_comp.c static void init_regs(u8 *rd, u8 *rs, const struct bpf_insn *insn,
rs                693 arch/riscv/net/bpf_jit_comp.c 		*rs = bpf_to_rv_reg(insn->src_reg, ctx);
rs                703 arch/riscv/net/bpf_jit_comp.c static void emit_zext_32_rd_rs(u8 *rd, u8 *rs, struct rv_jit_context *ctx)
rs                707 arch/riscv/net/bpf_jit_comp.c 	emit(rv_addi(RV_REG_T1, *rs, 0), ctx);
rs                710 arch/riscv/net/bpf_jit_comp.c 	*rs = RV_REG_T1;
rs                713 arch/riscv/net/bpf_jit_comp.c static void emit_sext_32_rd_rs(u8 *rd, u8 *rs, struct rv_jit_context *ctx)
rs                716 arch/riscv/net/bpf_jit_comp.c 	emit(rv_addiw(RV_REG_T1, *rs, 0), ctx);
rs                718 arch/riscv/net/bpf_jit_comp.c 	*rs = RV_REG_T1;
rs                742 arch/riscv/net/bpf_jit_comp.c 	u8 rd = -1, rs = -1, code = insn->code;
rs                746 arch/riscv/net/bpf_jit_comp.c 	init_regs(&rd, &rs, insn, ctx);
rs                757 arch/riscv/net/bpf_jit_comp.c 		emit(is64 ? rv_addi(rd, rs, 0) : rv_addiw(rd, rs, 0), ctx);
rs                765 arch/riscv/net/bpf_jit_comp.c 		emit(is64 ? rv_add(rd, rd, rs) : rv_addw(rd, rd, rs), ctx);
rs                771 arch/riscv/net/bpf_jit_comp.c 		emit(is64 ? rv_sub(rd, rd, rs) : rv_subw(rd, rd, rs), ctx);
rs                777 arch/riscv/net/bpf_jit_comp.c 		emit(rv_and(rd, rd, rs), ctx);
rs                783 arch/riscv/net/bpf_jit_comp.c 		emit(rv_or(rd, rd, rs), ctx);
rs                789 arch/riscv/net/bpf_jit_comp.c 		emit(rv_xor(rd, rd, rs), ctx);
rs                795 arch/riscv/net/bpf_jit_comp.c 		emit(is64 ? rv_mul(rd, rd, rs) : rv_mulw(rd, rd, rs), ctx);
rs                801 arch/riscv/net/bpf_jit_comp.c 		emit(is64 ? rv_divu(rd, rd, rs) : rv_divuw(rd, rd, rs), ctx);
rs                807 arch/riscv/net/bpf_jit_comp.c 		emit(is64 ? rv_remu(rd, rd, rs) : rv_remuw(rd, rd, rs), ctx);
rs                813 arch/riscv/net/bpf_jit_comp.c 		emit(is64 ? rv_sll(rd, rd, rs) : rv_sllw(rd, rd, rs), ctx);
rs                819 arch/riscv/net/bpf_jit_comp.c 		emit(is64 ? rv_srl(rd, rd, rs) : rv_srlw(rd, rd, rs), ctx);
rs                825 arch/riscv/net/bpf_jit_comp.c 		emit(is64 ? rv_sra(rd, rd, rs) : rv_sraw(rd, rd, rs), ctx);
rs               1025 arch/riscv/net/bpf_jit_comp.c 			emit_zext_32_rd_rs(&rd, &rs, ctx);
rs               1026 arch/riscv/net/bpf_jit_comp.c 		emit(rv_beq(rd, rs, rvoff >> 1), ctx);
rs               1033 arch/riscv/net/bpf_jit_comp.c 			emit_zext_32_rd_rs(&rd, &rs, ctx);
rs               1034 arch/riscv/net/bpf_jit_comp.c 		emit(rv_bltu(rs, rd, rvoff >> 1), ctx);
rs               1041 arch/riscv/net/bpf_jit_comp.c 			emit_zext_32_rd_rs(&rd, &rs, ctx);
rs               1042 arch/riscv/net/bpf_jit_comp.c 		emit(rv_bltu(rd, rs, rvoff >> 1), ctx);
rs               1049 arch/riscv/net/bpf_jit_comp.c 			emit_zext_32_rd_rs(&rd, &rs, ctx);
rs               1050 arch/riscv/net/bpf_jit_comp.c 		emit(rv_bgeu(rd, rs, rvoff >> 1), ctx);
rs               1057 arch/riscv/net/bpf_jit_comp.c 			emit_zext_32_rd_rs(&rd, &rs, ctx);
rs               1058 arch/riscv/net/bpf_jit_comp.c 		emit(rv_bgeu(rs, rd, rvoff >> 1), ctx);
rs               1065 arch/riscv/net/bpf_jit_comp.c 			emit_zext_32_rd_rs(&rd, &rs, ctx);
rs               1066 arch/riscv/net/bpf_jit_comp.c 		emit(rv_bne(rd, rs, rvoff >> 1), ctx);
rs               1073 arch/riscv/net/bpf_jit_comp.c 			emit_sext_32_rd_rs(&rd, &rs, ctx);
rs               1074 arch/riscv/net/bpf_jit_comp.c 		emit(rv_blt(rs, rd, rvoff >> 1), ctx);
rs               1081 arch/riscv/net/bpf_jit_comp.c 			emit_sext_32_rd_rs(&rd, &rs, ctx);
rs               1082 arch/riscv/net/bpf_jit_comp.c 		emit(rv_blt(rd, rs, rvoff >> 1), ctx);
rs               1089 arch/riscv/net/bpf_jit_comp.c 			emit_sext_32_rd_rs(&rd, &rs, ctx);
rs               1090 arch/riscv/net/bpf_jit_comp.c 		emit(rv_bge(rd, rs, rvoff >> 1), ctx);
rs               1097 arch/riscv/net/bpf_jit_comp.c 			emit_sext_32_rd_rs(&rd, &rs, ctx);
rs               1098 arch/riscv/net/bpf_jit_comp.c 		emit(rv_bge(rs, rd, rvoff >> 1), ctx);
rs               1105 arch/riscv/net/bpf_jit_comp.c 			emit_zext_32_rd_rs(&rd, &rs, ctx);
rs               1106 arch/riscv/net/bpf_jit_comp.c 		emit(rv_and(RV_REG_T1, rd, rs), ctx);
rs               1271 arch/riscv/net/bpf_jit_comp.c 			emit(rv_lbu(rd, off, rs), ctx);
rs               1276 arch/riscv/net/bpf_jit_comp.c 		emit(rv_add(RV_REG_T1, RV_REG_T1, rs), ctx);
rs               1283 arch/riscv/net/bpf_jit_comp.c 			emit(rv_lhu(rd, off, rs), ctx);
rs               1288 arch/riscv/net/bpf_jit_comp.c 		emit(rv_add(RV_REG_T1, RV_REG_T1, rs), ctx);
rs               1295 arch/riscv/net/bpf_jit_comp.c 			emit(rv_lwu(rd, off, rs), ctx);
rs               1300 arch/riscv/net/bpf_jit_comp.c 		emit(rv_add(RV_REG_T1, RV_REG_T1, rs), ctx);
rs               1307 arch/riscv/net/bpf_jit_comp.c 			emit(rv_ld(rd, off, rs), ctx);
rs               1312 arch/riscv/net/bpf_jit_comp.c 		emit(rv_add(RV_REG_T1, RV_REG_T1, rs), ctx);
rs               1366 arch/riscv/net/bpf_jit_comp.c 			emit(rv_sb(rd, off, rs), ctx);
rs               1372 arch/riscv/net/bpf_jit_comp.c 		emit(rv_sb(RV_REG_T1, 0, rs), ctx);
rs               1376 arch/riscv/net/bpf_jit_comp.c 			emit(rv_sh(rd, off, rs), ctx);
rs               1382 arch/riscv/net/bpf_jit_comp.c 		emit(rv_sh(RV_REG_T1, 0, rs), ctx);
rs               1386 arch/riscv/net/bpf_jit_comp.c 			emit(rv_sw(rd, off, rs), ctx);
rs               1392 arch/riscv/net/bpf_jit_comp.c 		emit(rv_sw(RV_REG_T1, 0, rs), ctx);
rs               1396 arch/riscv/net/bpf_jit_comp.c 			emit(rv_sd(rd, off, rs), ctx);
rs               1402 arch/riscv/net/bpf_jit_comp.c 		emit(rv_sd(RV_REG_T1, 0, rs), ctx);
rs               1420 arch/riscv/net/bpf_jit_comp.c 		     rv_amoadd_w(RV_REG_ZERO, rs, rd, 0, 0) :
rs               1421 arch/riscv/net/bpf_jit_comp.c 		     rv_amoadd_d(RV_REG_ZERO, rs, rd, 0, 0), ctx);
rs                321 arch/s390/net/bpf_jit_comp.c static void save_regs(struct bpf_jit *jit, u32 rs, u32 re)
rs                323 arch/s390/net/bpf_jit_comp.c 	u32 off = STK_OFF_R6 + (rs - 6) * 8;
rs                325 arch/s390/net/bpf_jit_comp.c 	if (rs == re)
rs                327 arch/s390/net/bpf_jit_comp.c 		_EMIT6(0xe300f000 | rs << 20 | off, 0x0024);
rs                330 arch/s390/net/bpf_jit_comp.c 		_EMIT6_DISP(0xeb00f000 | rs << 20 | re << 16, 0x0024, off);
rs                336 arch/s390/net/bpf_jit_comp.c static void restore_regs(struct bpf_jit *jit, u32 rs, u32 re, u32 stack_depth)
rs                338 arch/s390/net/bpf_jit_comp.c 	u32 off = STK_OFF_R6 + (rs - 6) * 8;
rs                343 arch/s390/net/bpf_jit_comp.c 	if (rs == re)
rs                345 arch/s390/net/bpf_jit_comp.c 		_EMIT6(0xe300f000 | rs << 20 | off, 0x0004);
rs                348 arch/s390/net/bpf_jit_comp.c 		_EMIT6_DISP(0xeb00f000 | rs << 20 | re << 16, 0x0004, off);
rs                388 arch/s390/net/bpf_jit_comp.c 	int re = 6, rs;
rs                391 arch/s390/net/bpf_jit_comp.c 		rs = get_start(jit, re);
rs                392 arch/s390/net/bpf_jit_comp.c 		if (!rs)
rs                394 arch/s390/net/bpf_jit_comp.c 		re = get_end(jit, rs + 1);
rs                396 arch/s390/net/bpf_jit_comp.c 			save_regs(jit, rs, re);
rs                398 arch/s390/net/bpf_jit_comp.c 			restore_regs(jit, rs, re, stack_depth);
rs                 72 arch/sh/include/uapi/asm/ptrace_32.h 	unsigned long	rs;
rs                858 arch/x86/kernel/ptrace.c #define SEG32(rs)							\
rs                859 arch/x86/kernel/ptrace.c 	case offsetof(struct user32, regs.rs):				\
rs                861 arch/x86/kernel/ptrace.c 				       offsetof(struct user_regs_struct, rs), \
rs                932 arch/x86/kernel/ptrace.c #define SEG32(rs)							\
rs                933 arch/x86/kernel/ptrace.c 	case offsetof(struct user32, regs.rs):				\
rs                935 arch/x86/kernel/ptrace.c 				       offsetof(struct user_regs_struct, rs)); \
rs                415 arch/xtensa/include/asm/pgtable.h #define _PGD_INDEX(rt,rs)	extui	rt, rs, PGDIR_SHIFT, 32-PGDIR_SHIFT
rs                416 arch/xtensa/include/asm/pgtable.h #define _PTE_INDEX(rt,rs)	extui	rt, rs, PAGE_SHIFT, PTRS_PER_PTE_SHIFT
rs                 37 block/partitions/atari.c 	struct rootsector *rs;
rs                 53 block/partitions/atari.c 	rs = read_part_sector(state, 0, &sect);
rs                 54 block/partitions/atari.c 	if (!rs)
rs                 59 block/partitions/atari.c 	if (!VALID_PARTITION(&rs->part[0], hd_size) &&
rs                 60 block/partitions/atari.c 	    !VALID_PARTITION(&rs->part[1], hd_size) &&
rs                 61 block/partitions/atari.c 	    !VALID_PARTITION(&rs->part[2], hd_size) &&
rs                 62 block/partitions/atari.c 	    !VALID_PARTITION(&rs->part[3], hd_size)) {
rs                 72 block/partitions/atari.c 	pi = &rs->part[0];
rs                 74 block/partitions/atari.c 	for (slot = 1; pi < &rs->part[4] && slot < state->limit; slot++, pi++) {
rs                135 block/partitions/atari.c 		pi = &rs->icdpart[0];
rs                139 block/partitions/atari.c 			for (; pi < &rs->icdpart[8] && slot < state->limit; slot++, pi++) {
rs                594 crypto/serpent_generic.c 	u32 rs[4], rd[4];
rs                596 crypto/serpent_generic.c 	rs[0] = swab32(s[3]);
rs                597 crypto/serpent_generic.c 	rs[1] = swab32(s[2]);
rs                598 crypto/serpent_generic.c 	rs[2] = swab32(s[1]);
rs                599 crypto/serpent_generic.c 	rs[3] = swab32(s[0]);
rs                601 crypto/serpent_generic.c 	serpent_encrypt(tfm, (u8 *)rd, (u8 *)rs);
rs                614 crypto/serpent_generic.c 	u32 rs[4], rd[4];
rs                616 crypto/serpent_generic.c 	rs[0] = swab32(s[3]);
rs                617 crypto/serpent_generic.c 	rs[1] = swab32(s[2]);
rs                618 crypto/serpent_generic.c 	rs[2] = swab32(s[1]);
rs                619 crypto/serpent_generic.c 	rs[3] = swab32(s[0]);
rs                621 crypto/serpent_generic.c 	serpent_decrypt(tfm, (u8 *)rd, (u8 *)rs);
rs                 62 drivers/auxdisplay/hd44780.c static void hd44780_write_gpio8(struct hd44780 *hd, u8 val, unsigned int rs)
rs                 68 drivers/auxdisplay/hd44780.c 	__assign_bit(8, values, rs);
rs                 78 drivers/auxdisplay/hd44780.c static void hd44780_write_gpio4(struct hd44780 *hd, u8 val, unsigned int rs)
rs                 85 drivers/auxdisplay/hd44780.c 	__assign_bit(4, values, rs);
rs                211 drivers/auxdisplay/panel.c 		int rs;
rs                920 drivers/auxdisplay/panel.c 		lcd.pins.rs = PIN_AUTOLF;
rs                945 drivers/auxdisplay/panel.c 		lcd.pins.rs = PIN_SELECP;
rs                965 drivers/auxdisplay/panel.c 		lcd.pins.rs = PIN_SELECP;
rs                990 drivers/auxdisplay/panel.c 		lcd.pins.rs = lcd_rs_pin;
rs               1023 drivers/auxdisplay/panel.c 		if (lcd.pins.rs == PIN_NOT_SET)
rs               1024 drivers/auxdisplay/panel.c 			lcd.pins.rs = DEFAULT_LCD_PIN_RS;
rs               1036 drivers/auxdisplay/panel.c 	if (lcd.pins.rs == PIN_NOT_SET)
rs               1037 drivers/auxdisplay/panel.c 		lcd.pins.rs = PIN_NONE;
rs               1057 drivers/auxdisplay/panel.c 	pin_to_bits(lcd.pins.rs, lcd_bits[LCD_PORT_D][LCD_BIT_RS],
rs               1740 drivers/auxdisplay/panel.c 		lcd.pins.rs = lcd_rs_pin;
rs                928 drivers/block/drbd/drbd_nl.c drbd_determine_dev_size(struct drbd_device *device, enum dds_flags flags, struct resize_parms *rs) __must_hold(local)
rs                971 drivers/block/drbd/drbd_nl.c 	if (rs) {
rs                973 drivers/block/drbd/drbd_nl.c 		md->al_stripes = rs->al_stripes;
rs                974 drivers/block/drbd/drbd_nl.c 		md->al_stripe_size_4k = rs->al_stripe_size / 4;
rs                975 drivers/block/drbd/drbd_nl.c 		md->al_size_4k = (u64)rs->al_stripes * rs->al_stripe_size / 4;
rs                986 drivers/block/drbd/drbd_nl.c 		if (rs && u_size == 0) {
rs               1028 drivers/block/drbd/drbd_nl.c 	if (la_size_changed || md_moved || rs) {
rs               1061 drivers/block/drbd/drbd_nl.c 		if (rs)
rs               2841 drivers/block/drbd/drbd_nl.c 	struct resize_parms rs;
rs               2863 drivers/block/drbd/drbd_nl.c 	memset(&rs, 0, sizeof(struct resize_parms));
rs               2864 drivers/block/drbd/drbd_nl.c 	rs.al_stripes = device->ldev->md.al_stripes;
rs               2865 drivers/block/drbd/drbd_nl.c 	rs.al_stripe_size = device->ldev->md.al_stripe_size_4k * 4;
rs               2867 drivers/block/drbd/drbd_nl.c 		err = resize_parms_from_attrs(&rs, info);
rs               2886 drivers/block/drbd/drbd_nl.c 	if (rs.no_resync && first_peer_device(device)->connection->agreed_pro_version < 93) {
rs               2894 drivers/block/drbd/drbd_nl.c 	if (u_size != (sector_t)rs.resize_size) {
rs               2902 drivers/block/drbd/drbd_nl.c 	if (device->ldev->md.al_stripes != rs.al_stripes ||
rs               2903 drivers/block/drbd/drbd_nl.c 	    device->ldev->md.al_stripe_size_4k != rs.al_stripe_size / 4) {
rs               2904 drivers/block/drbd/drbd_nl.c 		u32 al_size_k = rs.al_stripes * rs.al_stripe_size;
rs               2916 drivers/block/drbd/drbd_nl.c 		if (device->state.conn != C_CONNECTED && !rs.resize_force) {
rs               2931 drivers/block/drbd/drbd_nl.c 		new_disk_conf->disk_size = (sector_t)rs.resize_size;
rs               2939 drivers/block/drbd/drbd_nl.c 	ddsf = (rs.resize_force ? DDSF_FORCED : 0) | (rs.no_resync ? DDSF_NO_RESYNC : 0);
rs               2940 drivers/block/drbd/drbd_nl.c 	dd = drbd_determine_dev_size(device, ddsf, change_al_layout ? &rs : NULL);
rs                 37 drivers/clk/bcm/clk-cygnus.c #define RESET_VAL(o, rs, prs) { .offset = o, .reset_shift = rs, \
rs                 29 drivers/clk/bcm/clk-ns2.c #define RESET_VAL(o, rs, prs) { .offset = o, .reset_shift = rs, \
rs                 29 drivers/clk/bcm/clk-nsp.c #define RESET_VAL(o, rs, prs) { .offset = o, .reset_shift = rs, \
rs                 21 drivers/clk/bcm/clk-sr.c #define RESET_VAL(o, rs, prs) { .offset = o, .reset_shift = rs, \
rs                 24 drivers/gpu/drm/msm/disp/mdp4/mdp4_irq.c 	static DEFINE_RATELIMIT_STATE(rs, 5*HZ, 1);
rs                 29 drivers/gpu/drm/msm/disp/mdp4/mdp4_irq.c 	if (dumpstate && __ratelimit(&rs)) {
rs                 26 drivers/gpu/drm/msm/disp/mdp5/mdp5_irq.c 	static DEFINE_RATELIMIT_STATE(rs, 5*HZ, 1);
rs                 31 drivers/gpu/drm/msm/disp/mdp5/mdp5_irq.c 	if (dumpstate && __ratelimit(&rs)) {
rs                325 drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c 	uint64_t rs[] = {mode->hdisplay * id3,
rs                328 drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c 	do_div(rs[0], overscan * tv_norm->tv_enc_mode.hdisplay);
rs                329 drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c 	do_div(rs[1], overscan * tv_norm->tv_enc_mode.vdisplay);
rs                332 drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c 		rs[k] = max((int64_t)rs[k], id2);
rs                341 drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c 					   p->ki3r*i*i*i) * rs[k]
rs                345 drivers/gpu/drm/nouveau/dispnv04/tvmodesnv17.c 					   p->ki3rf*i*i*i) * flicker * rs[k];
rs                384 drivers/gpu/drm/tiny/ili9225.c 	struct gpio_desc *rs;
rs                408 drivers/gpu/drm/tiny/ili9225.c 	rs = devm_gpiod_get(dev, "rs", GPIOD_OUT_LOW);
rs                409 drivers/gpu/drm/tiny/ili9225.c 	if (IS_ERR(rs)) {
rs                411 drivers/gpu/drm/tiny/ili9225.c 		return PTR_ERR(rs);
rs                416 drivers/gpu/drm/tiny/ili9225.c 	ret = mipi_dbi_spi_init(spi, dbi, rs);
rs                333 drivers/hid/intel-ish-hid/ishtp/hbm.c 	struct hbm_client_connect_response *rs)
rs                340 drivers/hid/intel-ish-hid/ishtp/hbm.c 		if (!rs->status && ishtp_hbm_cl_addr_equal(cl, rs)) {
rs                378 drivers/hid/intel-ish-hid/ishtp/hbm.c 	struct hbm_client_connect_response *rs)
rs                385 drivers/hid/intel-ish-hid/ishtp/hbm.c 		if (ishtp_hbm_cl_addr_equal(cl, rs)) {
rs                386 drivers/hid/intel-ish-hid/ishtp/hbm.c 			if (!rs->status) {
rs                206 drivers/iommu/amd_iommu.c 	ratelimit_default_init(&dev_data->rs);
rs                545 drivers/iommu/amd_iommu.c 	if (dev_data && __ratelimit(&dev_data->rs)) {
rs                656 drivers/iommu/amd_iommu_types.h 	struct ratelimit_state rs;        /* Ratelimit IOPF messages */
rs               1714 drivers/iommu/dmar.c 	static DEFINE_RATELIMIT_STATE(rs,
rs               1720 drivers/iommu/dmar.c 	if (fault_status && __ratelimit(&rs))
rs               1731 drivers/iommu/dmar.c 		bool ratelimited = !__ratelimit(&rs);
rs                296 drivers/md/dm-clone-target.c 			     unsigned long *rs, unsigned long *nr_regions)
rs                300 drivers/md/dm-clone-target.c 	*rs = dm_sector_div_up(bio->bi_iter.bi_sector, clone->region_size);
rs                303 drivers/md/dm-clone-target.c 	if (*rs >= end)
rs                306 drivers/md/dm-clone-target.c 		*nr_regions = end - *rs;
rs                464 drivers/md/dm-clone-target.c 	unsigned long rs, nr_regions;
rs                473 drivers/md/dm-clone-target.c 		bio_region_range(clone, bio, &rs, &nr_regions);
rs                474 drivers/md/dm-clone-target.c 		trim_bio(bio, region_to_sector(clone, rs),
rs                483 drivers/md/dm-clone-target.c 	unsigned long rs, nr_regions;
rs                485 drivers/md/dm-clone-target.c 	bio_region_range(clone, bio, &rs, &nr_regions);
rs                491 drivers/md/dm-clone-target.c 	if (WARN_ON(rs >= clone->nr_regions || (rs + nr_regions) < rs ||
rs                492 drivers/md/dm-clone-target.c 		    (rs + nr_regions) > clone->nr_regions)) {
rs                494 drivers/md/dm-clone-target.c 		      clone_device_name(clone), rs, nr_regions,
rs                506 drivers/md/dm-clone-target.c 	if (dm_clone_is_range_hydrated(clone->cmd, rs, nr_regions)) {
rs               1187 drivers/md/dm-clone-target.c 	unsigned long rs, nr_regions;
rs               1203 drivers/md/dm-clone-target.c 		bio_region_range(clone, bio, &rs, &nr_regions);
rs               1209 drivers/md/dm-clone-target.c 		r = dm_clone_cond_set_range(clone->cmd, rs, nr_regions);
rs                254 drivers/md/dm-raid.c static void rs_config_backup(struct raid_set *rs, struct rs_layout *l)
rs                256 drivers/md/dm-raid.c 	struct mddev *mddev = &rs->md;
rs                263 drivers/md/dm-raid.c static void rs_config_restore(struct raid_set *rs, struct rs_layout *l)
rs                265 drivers/md/dm-raid.c 	struct mddev *mddev = &rs->md;
rs                394 drivers/md/dm-raid.c static bool rs_is_raid0(struct raid_set *rs)
rs                396 drivers/md/dm-raid.c 	return !rs->md.level;
rs                400 drivers/md/dm-raid.c static bool rs_is_raid1(struct raid_set *rs)
rs                402 drivers/md/dm-raid.c 	return rs->md.level == 1;
rs                406 drivers/md/dm-raid.c static bool rs_is_raid10(struct raid_set *rs)
rs                408 drivers/md/dm-raid.c 	return rs->md.level == 10;
rs                412 drivers/md/dm-raid.c static bool rs_is_raid6(struct raid_set *rs)
rs                414 drivers/md/dm-raid.c 	return rs->md.level == 6;
rs                418 drivers/md/dm-raid.c static bool rs_is_raid456(struct raid_set *rs)
rs                420 drivers/md/dm-raid.c 	return __within_range(rs->md.level, 4, 6);
rs                425 drivers/md/dm-raid.c static bool rs_is_reshapable(struct raid_set *rs)
rs                427 drivers/md/dm-raid.c 	return rs_is_raid456(rs) ||
rs                428 drivers/md/dm-raid.c 	       (rs_is_raid10(rs) && !__is_raid10_far(rs->md.new_layout));
rs                432 drivers/md/dm-raid.c static bool rs_is_recovering(struct raid_set *rs)
rs                434 drivers/md/dm-raid.c 	return rs->md.recovery_cp < rs->md.dev_sectors;
rs                438 drivers/md/dm-raid.c static bool rs_is_reshaping(struct raid_set *rs)
rs                440 drivers/md/dm-raid.c 	return rs->md.reshape_position != MaxSector;
rs                485 drivers/md/dm-raid.c static unsigned long __valid_flags(struct raid_set *rs)
rs                487 drivers/md/dm-raid.c 	if (rt_is_raid0(rs->raid_type))
rs                489 drivers/md/dm-raid.c 	else if (rt_is_raid1(rs->raid_type))
rs                491 drivers/md/dm-raid.c 	else if (rt_is_raid10(rs->raid_type))
rs                493 drivers/md/dm-raid.c 	else if (rt_is_raid45(rs->raid_type))
rs                495 drivers/md/dm-raid.c 	else if (rt_is_raid6(rs->raid_type))
rs                506 drivers/md/dm-raid.c static int rs_check_for_valid_flags(struct raid_set *rs)
rs                508 drivers/md/dm-raid.c 	if (rs->ctr_flags & ~__valid_flags(rs)) {
rs                509 drivers/md/dm-raid.c 		rs->ti->error = "Invalid flags combination";
rs                593 drivers/md/dm-raid.c static int raid10_format_to_md_layout(struct raid_set *rs,
rs                614 drivers/md/dm-raid.c 		if (!test_bit(__CTR_FLAG_RAID10_USE_NEAR_SETS, &rs->ctr_flags))
rs                620 drivers/md/dm-raid.c 		if (!test_bit(__CTR_FLAG_RAID10_USE_NEAR_SETS, &rs->ctr_flags))
rs                678 drivers/md/dm-raid.c static void rs_set_rdev_sectors(struct raid_set *rs)
rs                680 drivers/md/dm-raid.c 	struct mddev *mddev = &rs->md;
rs                695 drivers/md/dm-raid.c static void rs_set_capacity(struct raid_set *rs)
rs                697 drivers/md/dm-raid.c 	struct gendisk *gendisk = dm_disk(dm_table_get_md(rs->ti->table));
rs                699 drivers/md/dm-raid.c 	set_capacity(gendisk, rs->md.array_sectors);
rs                707 drivers/md/dm-raid.c static void rs_set_cur(struct raid_set *rs)
rs                709 drivers/md/dm-raid.c 	struct mddev *mddev = &rs->md;
rs                720 drivers/md/dm-raid.c static void rs_set_new(struct raid_set *rs)
rs                722 drivers/md/dm-raid.c 	struct mddev *mddev = &rs->md;
rs                727 drivers/md/dm-raid.c 	mddev->raid_disks = rs->raid_disks;
rs                735 drivers/md/dm-raid.c 	struct raid_set *rs;
rs                742 drivers/md/dm-raid.c 	rs = kzalloc(struct_size(rs, dev, raid_devs), GFP_KERNEL);
rs                743 drivers/md/dm-raid.c 	if (!rs) {
rs                748 drivers/md/dm-raid.c 	mddev_init(&rs->md);
rs                750 drivers/md/dm-raid.c 	rs->raid_disks = raid_devs;
rs                751 drivers/md/dm-raid.c 	rs->delta_disks = 0;
rs                753 drivers/md/dm-raid.c 	rs->ti = ti;
rs                754 drivers/md/dm-raid.c 	rs->raid_type = raid_type;
rs                755 drivers/md/dm-raid.c 	rs->stripe_cache_entries = 256;
rs                756 drivers/md/dm-raid.c 	rs->md.raid_disks = raid_devs;
rs                757 drivers/md/dm-raid.c 	rs->md.level = raid_type->level;
rs                758 drivers/md/dm-raid.c 	rs->md.new_level = rs->md.level;
rs                759 drivers/md/dm-raid.c 	rs->md.layout = raid_type->algorithm;
rs                760 drivers/md/dm-raid.c 	rs->md.new_layout = rs->md.layout;
rs                761 drivers/md/dm-raid.c 	rs->md.delta_disks = 0;
rs                762 drivers/md/dm-raid.c 	rs->md.recovery_cp = MaxSector;
rs                765 drivers/md/dm-raid.c 		md_rdev_init(&rs->dev[i].rdev);
rs                776 drivers/md/dm-raid.c 	return rs;
rs                780 drivers/md/dm-raid.c static void raid_set_free(struct raid_set *rs)
rs                784 drivers/md/dm-raid.c 	if (rs->journal_dev.dev) {
rs                785 drivers/md/dm-raid.c 		md_rdev_clear(&rs->journal_dev.rdev);
rs                786 drivers/md/dm-raid.c 		dm_put_device(rs->ti, rs->journal_dev.dev);
rs                789 drivers/md/dm-raid.c 	for (i = 0; i < rs->raid_disks; i++) {
rs                790 drivers/md/dm-raid.c 		if (rs->dev[i].meta_dev)
rs                791 drivers/md/dm-raid.c 			dm_put_device(rs->ti, rs->dev[i].meta_dev);
rs                792 drivers/md/dm-raid.c 		md_rdev_clear(&rs->dev[i].rdev);
rs                793 drivers/md/dm-raid.c 		if (rs->dev[i].data_dev)
rs                794 drivers/md/dm-raid.c 			dm_put_device(rs->ti, rs->dev[i].data_dev);
rs                797 drivers/md/dm-raid.c 	kfree(rs);
rs                816 drivers/md/dm-raid.c static int parse_dev_params(struct raid_set *rs, struct dm_arg_set *as)
rs                829 drivers/md/dm-raid.c 	for (i = 0; i < rs->raid_disks; i++) {
rs                830 drivers/md/dm-raid.c 		rs->dev[i].rdev.raid_disk = i;
rs                832 drivers/md/dm-raid.c 		rs->dev[i].meta_dev = NULL;
rs                833 drivers/md/dm-raid.c 		rs->dev[i].data_dev = NULL;
rs                839 drivers/md/dm-raid.c 		rs->dev[i].rdev.data_offset = 0;
rs                840 drivers/md/dm-raid.c 		rs->dev[i].rdev.new_data_offset = 0;
rs                841 drivers/md/dm-raid.c 		rs->dev[i].rdev.mddev = &rs->md;
rs                848 drivers/md/dm-raid.c 			r = dm_get_device(rs->ti, arg, dm_table_get_mode(rs->ti->table),
rs                849 drivers/md/dm-raid.c 					  &rs->dev[i].meta_dev);
rs                851 drivers/md/dm-raid.c 				rs->ti->error = "RAID metadata device lookup failure";
rs                855 drivers/md/dm-raid.c 			rs->dev[i].rdev.sb_page = alloc_page(GFP_KERNEL);
rs                856 drivers/md/dm-raid.c 			if (!rs->dev[i].rdev.sb_page) {
rs                857 drivers/md/dm-raid.c 				rs->ti->error = "Failed to allocate superblock page";
rs                867 drivers/md/dm-raid.c 			if (!test_bit(In_sync, &rs->dev[i].rdev.flags) &&
rs                868 drivers/md/dm-raid.c 			    (!rs->dev[i].rdev.recovery_offset)) {
rs                869 drivers/md/dm-raid.c 				rs->ti->error = "Drive designated for rebuild not specified";
rs                873 drivers/md/dm-raid.c 			if (rs->dev[i].meta_dev) {
rs                874 drivers/md/dm-raid.c 				rs->ti->error = "No data device supplied with metadata device";
rs                881 drivers/md/dm-raid.c 		r = dm_get_device(rs->ti, arg, dm_table_get_mode(rs->ti->table),
rs                882 drivers/md/dm-raid.c 				  &rs->dev[i].data_dev);
rs                884 drivers/md/dm-raid.c 			rs->ti->error = "RAID device lookup failure";
rs                888 drivers/md/dm-raid.c 		if (rs->dev[i].meta_dev) {
rs                890 drivers/md/dm-raid.c 			rs->dev[i].rdev.meta_bdev = rs->dev[i].meta_dev->bdev;
rs                892 drivers/md/dm-raid.c 		rs->dev[i].rdev.bdev = rs->dev[i].data_dev->bdev;
rs                893 drivers/md/dm-raid.c 		list_add_tail(&rs->dev[i].rdev.same_set, &rs->md.disks);
rs                894 drivers/md/dm-raid.c 		if (!test_bit(In_sync, &rs->dev[i].rdev.flags))
rs                898 drivers/md/dm-raid.c 	if (rs->journal_dev.dev)
rs                899 drivers/md/dm-raid.c 		list_add_tail(&rs->journal_dev.rdev.same_set, &rs->md.disks);
rs                902 drivers/md/dm-raid.c 		rs->md.external = 0;
rs                903 drivers/md/dm-raid.c 		rs->md.persistent = 1;
rs                904 drivers/md/dm-raid.c 		rs->md.major_version = 2;
rs                905 drivers/md/dm-raid.c 	} else if (rebuild && !rs->md.recovery_cp) {
rs                917 drivers/md/dm-raid.c 		rs->ti->error = "Unable to rebuild drive while array is not in-sync";
rs                934 drivers/md/dm-raid.c static int validate_region_size(struct raid_set *rs, unsigned long region_size)
rs                936 drivers/md/dm-raid.c 	unsigned long min_region_size = rs->ti->len / (1 << 21);
rs                938 drivers/md/dm-raid.c 	if (rs_is_raid0(rs))
rs                958 drivers/md/dm-raid.c 		if (region_size > rs->ti->len) {
rs                959 drivers/md/dm-raid.c 			rs->ti->error = "Supplied region size is too large";
rs                966 drivers/md/dm-raid.c 			rs->ti->error = "Supplied region size is too small";
rs                971 drivers/md/dm-raid.c 			rs->ti->error = "Region size is not a power of 2";
rs                975 drivers/md/dm-raid.c 		if (region_size < rs->md.chunk_sectors) {
rs                976 drivers/md/dm-raid.c 			rs->ti->error = "Region size is smaller than the chunk size";
rs                984 drivers/md/dm-raid.c 	rs->md.bitmap_info.chunksize = to_bytes(region_size);
rs                998 drivers/md/dm-raid.c static int validate_raid_redundancy(struct raid_set *rs)
rs               1004 drivers/md/dm-raid.c 	for (i = 0; i < rs->md.raid_disks; i++)
rs               1005 drivers/md/dm-raid.c 		if (!test_bit(In_sync, &rs->dev[i].rdev.flags) ||
rs               1006 drivers/md/dm-raid.c 		    !rs->dev[i].rdev.sb_page)
rs               1009 drivers/md/dm-raid.c 	switch (rs->md.level) {
rs               1013 drivers/md/dm-raid.c 		if (rebuild_cnt >= rs->md.raid_disks)
rs               1019 drivers/md/dm-raid.c 		if (rebuild_cnt > rs->raid_type->parity_devs)
rs               1023 drivers/md/dm-raid.c 		copies = raid10_md_layout_to_copies(rs->md.new_layout);
rs               1046 drivers/md/dm-raid.c 		if (__is_raid10_near(rs->md.new_layout)) {
rs               1047 drivers/md/dm-raid.c 			for (i = 0; i < rs->md.raid_disks; i++) {
rs               1050 drivers/md/dm-raid.c 				if ((!rs->dev[i].rdev.sb_page ||
rs               1051 drivers/md/dm-raid.c 				    !test_bit(In_sync, &rs->dev[i].rdev.flags)) &&
rs               1070 drivers/md/dm-raid.c 		group_size = (rs->md.raid_disks / copies);
rs               1071 drivers/md/dm-raid.c 		last_group_start = (rs->md.raid_disks / group_size) - 1;
rs               1073 drivers/md/dm-raid.c 		for (i = 0; i < rs->md.raid_disks; i++) {
rs               1076 drivers/md/dm-raid.c 			if ((!rs->dev[i].rdev.sb_page ||
rs               1077 drivers/md/dm-raid.c 			     !test_bit(In_sync, &rs->dev[i].rdev.flags)) &&
rs               1118 drivers/md/dm-raid.c static int parse_raid_params(struct raid_set *rs, struct dm_arg_set *as,
rs               1128 drivers/md/dm-raid.c 	struct raid_type *rt = rs->raid_type;
rs               1134 drivers/md/dm-raid.c 		rs->ti->error = "Bad numerical argument given for chunk_size";
rs               1147 drivers/md/dm-raid.c 		rs->ti->error = "Chunk size must be a power of 2";
rs               1150 drivers/md/dm-raid.c 		rs->ti->error = "Chunk size value is too small";
rs               1154 drivers/md/dm-raid.c 	rs->md.new_chunk_sectors = rs->md.chunk_sectors = value;
rs               1173 drivers/md/dm-raid.c 	for (i = 0; i < rs->raid_disks; i++) {
rs               1174 drivers/md/dm-raid.c 		set_bit(In_sync, &rs->dev[i].rdev.flags);
rs               1175 drivers/md/dm-raid.c 		rs->dev[i].rdev.recovery_offset = MaxSector;
rs               1184 drivers/md/dm-raid.c 			rs->ti->error = "Not enough raid parameters given";
rs               1189 drivers/md/dm-raid.c 			if (test_and_set_bit(__CTR_FLAG_NOSYNC, &rs->ctr_flags)) {
rs               1190 drivers/md/dm-raid.c 				rs->ti->error = "Only one 'nosync' argument allowed";
rs               1196 drivers/md/dm-raid.c 			if (test_and_set_bit(__CTR_FLAG_SYNC, &rs->ctr_flags)) {
rs               1197 drivers/md/dm-raid.c 				rs->ti->error = "Only one 'sync' argument allowed";
rs               1203 drivers/md/dm-raid.c 			if (test_and_set_bit(__CTR_FLAG_RAID10_USE_NEAR_SETS, &rs->ctr_flags)) {
rs               1204 drivers/md/dm-raid.c 				rs->ti->error = "Only one 'raid10_use_new_sets' argument allowed";
rs               1213 drivers/md/dm-raid.c 			rs->ti->error = "Wrong number of raid parameters given";
rs               1222 drivers/md/dm-raid.c 			if (test_and_set_bit(__CTR_FLAG_RAID10_FORMAT, &rs->ctr_flags)) {
rs               1223 drivers/md/dm-raid.c 				rs->ti->error = "Only one 'raid10_format' argument pair allowed";
rs               1227 drivers/md/dm-raid.c 				rs->ti->error = "'raid10_format' is an invalid parameter for this RAID type";
rs               1232 drivers/md/dm-raid.c 				rs->ti->error = "Invalid 'raid10_format' value given";
rs               1243 drivers/md/dm-raid.c 			if (test_and_set_bit(__CTR_FLAG_JOURNAL_DEV, &rs->ctr_flags)) {
rs               1244 drivers/md/dm-raid.c 				rs->ti->error = "Only one raid4/5/6 set journaling device allowed";
rs               1248 drivers/md/dm-raid.c 				rs->ti->error = "'journal_dev' is an invalid parameter for this RAID type";
rs               1251 drivers/md/dm-raid.c 			r = dm_get_device(rs->ti, arg, dm_table_get_mode(rs->ti->table),
rs               1252 drivers/md/dm-raid.c 					  &rs->journal_dev.dev);
rs               1254 drivers/md/dm-raid.c 				rs->ti->error = "raid4/5/6 journal device lookup failure";
rs               1257 drivers/md/dm-raid.c 			jdev = &rs->journal_dev.rdev;
rs               1259 drivers/md/dm-raid.c 			jdev->mddev = &rs->md;
rs               1260 drivers/md/dm-raid.c 			jdev->bdev = rs->journal_dev.dev->bdev;
rs               1263 drivers/md/dm-raid.c 				rs->ti->error = "No space for raid4/5/6 journal";
rs               1266 drivers/md/dm-raid.c 			rs->journal_dev.mode = R5C_JOURNAL_MODE_WRITE_THROUGH;
rs               1275 drivers/md/dm-raid.c 			if (!test_bit(__CTR_FLAG_JOURNAL_DEV, &rs->ctr_flags)) {
rs               1276 drivers/md/dm-raid.c 				rs->ti->error = "raid4/5/6 'journal_mode' is invalid without 'journal_dev'";
rs               1279 drivers/md/dm-raid.c 			if (test_and_set_bit(__CTR_FLAG_JOURNAL_MODE, &rs->ctr_flags)) {
rs               1280 drivers/md/dm-raid.c 				rs->ti->error = "Only one raid4/5/6 'journal_mode' argument allowed";
rs               1285 drivers/md/dm-raid.c 				rs->ti->error = "Invalid 'journal_mode' argument";
rs               1288 drivers/md/dm-raid.c 			rs->journal_dev.mode = r;
rs               1296 drivers/md/dm-raid.c 			rs->ti->error = "Bad numerical argument given in raid params";
rs               1306 drivers/md/dm-raid.c 			if (!__within_range(value, 0, rs->raid_disks - 1)) {
rs               1307 drivers/md/dm-raid.c 				rs->ti->error = "Invalid rebuild index given";
rs               1311 drivers/md/dm-raid.c 			if (test_and_set_bit(value, (void *) rs->rebuild_disks)) {
rs               1312 drivers/md/dm-raid.c 				rs->ti->error = "rebuild for this index already given";
rs               1316 drivers/md/dm-raid.c 			rd = rs->dev + value;
rs               1320 drivers/md/dm-raid.c 			set_bit(__CTR_FLAG_REBUILD, &rs->ctr_flags);
rs               1323 drivers/md/dm-raid.c 				rs->ti->error = "write_mostly option is only valid for RAID1";
rs               1327 drivers/md/dm-raid.c 			if (!__within_range(value, 0, rs->md.raid_disks - 1)) {
rs               1328 drivers/md/dm-raid.c 				rs->ti->error = "Invalid write_mostly index given";
rs               1333 drivers/md/dm-raid.c 			set_bit(WriteMostly, &rs->dev[value].rdev.flags);
rs               1334 drivers/md/dm-raid.c 			set_bit(__CTR_FLAG_WRITE_MOSTLY, &rs->ctr_flags);
rs               1337 drivers/md/dm-raid.c 				rs->ti->error = "max_write_behind option is only valid for RAID1";
rs               1341 drivers/md/dm-raid.c 			if (test_and_set_bit(__CTR_FLAG_MAX_WRITE_BEHIND, &rs->ctr_flags)) {
rs               1342 drivers/md/dm-raid.c 				rs->ti->error = "Only one max_write_behind argument pair allowed";
rs               1351 drivers/md/dm-raid.c 				rs->ti->error = "Max write-behind limit out of range";
rs               1355 drivers/md/dm-raid.c 			rs->md.bitmap_info.max_write_behind = value / 2;
rs               1357 drivers/md/dm-raid.c 			if (test_and_set_bit(__CTR_FLAG_DAEMON_SLEEP, &rs->ctr_flags)) {
rs               1358 drivers/md/dm-raid.c 				rs->ti->error = "Only one daemon_sleep argument pair allowed";
rs               1362 drivers/md/dm-raid.c 				rs->ti->error = "daemon sleep period out of range";
rs               1365 drivers/md/dm-raid.c 			rs->md.bitmap_info.daemon_sleep = value;
rs               1368 drivers/md/dm-raid.c 			if (test_and_set_bit(__CTR_FLAG_DATA_OFFSET, &rs->ctr_flags)) {
rs               1369 drivers/md/dm-raid.c 				rs->ti->error = "Only one data_offset argument pair allowed";
rs               1375 drivers/md/dm-raid.c 				rs->ti->error = "Bogus data_offset value";
rs               1378 drivers/md/dm-raid.c 			rs->data_offset = value;
rs               1381 drivers/md/dm-raid.c 			if (test_and_set_bit(__CTR_FLAG_DELTA_DISKS, &rs->ctr_flags)) {
rs               1382 drivers/md/dm-raid.c 				rs->ti->error = "Only one delta_disks argument pair allowed";
rs               1387 drivers/md/dm-raid.c 				rs->ti->error = "Too many delta_disk requested";
rs               1391 drivers/md/dm-raid.c 			rs->delta_disks = value;
rs               1393 drivers/md/dm-raid.c 			if (test_and_set_bit(__CTR_FLAG_STRIPE_CACHE, &rs->ctr_flags)) {
rs               1394 drivers/md/dm-raid.c 				rs->ti->error = "Only one stripe_cache argument pair allowed";
rs               1399 drivers/md/dm-raid.c 				rs->ti->error = "Inappropriate argument: stripe_cache";
rs               1404 drivers/md/dm-raid.c 				rs->ti->error = "Bogus stripe cache entries value";
rs               1407 drivers/md/dm-raid.c 			rs->stripe_cache_entries = value;
rs               1409 drivers/md/dm-raid.c 			if (test_and_set_bit(__CTR_FLAG_MIN_RECOVERY_RATE, &rs->ctr_flags)) {
rs               1410 drivers/md/dm-raid.c 				rs->ti->error = "Only one min_recovery_rate argument pair allowed";
rs               1415 drivers/md/dm-raid.c 				rs->ti->error = "min_recovery_rate out of range";
rs               1418 drivers/md/dm-raid.c 			rs->md.sync_speed_min = value;
rs               1420 drivers/md/dm-raid.c 			if (test_and_set_bit(__CTR_FLAG_MAX_RECOVERY_RATE, &rs->ctr_flags)) {
rs               1421 drivers/md/dm-raid.c 				rs->ti->error = "Only one max_recovery_rate argument pair allowed";
rs               1426 drivers/md/dm-raid.c 				rs->ti->error = "max_recovery_rate out of range";
rs               1429 drivers/md/dm-raid.c 			rs->md.sync_speed_max = value;
rs               1431 drivers/md/dm-raid.c 			if (test_and_set_bit(__CTR_FLAG_REGION_SIZE, &rs->ctr_flags)) {
rs               1432 drivers/md/dm-raid.c 				rs->ti->error = "Only one region_size argument pair allowed";
rs               1437 drivers/md/dm-raid.c 			rs->requested_bitmap_chunk_sectors = value;
rs               1439 drivers/md/dm-raid.c 			if (test_and_set_bit(__CTR_FLAG_RAID10_COPIES, &rs->ctr_flags)) {
rs               1440 drivers/md/dm-raid.c 				rs->ti->error = "Only one raid10_copies argument pair allowed";
rs               1444 drivers/md/dm-raid.c 			if (!__within_range(value, 2, rs->md.raid_disks)) {
rs               1445 drivers/md/dm-raid.c 				rs->ti->error = "Bad value for 'raid10_copies'";
rs               1452 drivers/md/dm-raid.c 			rs->ti->error = "Unable to parse RAID parameter";
rs               1457 drivers/md/dm-raid.c 	if (test_bit(__CTR_FLAG_SYNC, &rs->ctr_flags) &&
rs               1458 drivers/md/dm-raid.c 	    test_bit(__CTR_FLAG_NOSYNC, &rs->ctr_flags)) {
rs               1459 drivers/md/dm-raid.c 		rs->ti->error = "sync and nosync are mutually exclusive";
rs               1463 drivers/md/dm-raid.c 	if (test_bit(__CTR_FLAG_REBUILD, &rs->ctr_flags) &&
rs               1464 drivers/md/dm-raid.c 	    (test_bit(__CTR_FLAG_SYNC, &rs->ctr_flags) ||
rs               1465 drivers/md/dm-raid.c 	     test_bit(__CTR_FLAG_NOSYNC, &rs->ctr_flags))) {
rs               1466 drivers/md/dm-raid.c 		rs->ti->error = "sync/nosync and rebuild are mutually exclusive";
rs               1470 drivers/md/dm-raid.c 	if (write_mostly >= rs->md.raid_disks) {
rs               1471 drivers/md/dm-raid.c 		rs->ti->error = "Can't set all raid1 devices to write_mostly";
rs               1475 drivers/md/dm-raid.c 	if (rs->md.sync_speed_max &&
rs               1476 drivers/md/dm-raid.c 	    rs->md.sync_speed_min > rs->md.sync_speed_max) {
rs               1477 drivers/md/dm-raid.c 		rs->ti->error = "Bogus recovery rates";
rs               1481 drivers/md/dm-raid.c 	if (validate_region_size(rs, region_size))
rs               1484 drivers/md/dm-raid.c 	if (rs->md.chunk_sectors)
rs               1485 drivers/md/dm-raid.c 		max_io_len = rs->md.chunk_sectors;
rs               1489 drivers/md/dm-raid.c 	if (dm_set_target_max_io_len(rs->ti, max_io_len))
rs               1493 drivers/md/dm-raid.c 		if (raid10_copies > rs->md.raid_disks) {
rs               1494 drivers/md/dm-raid.c 			rs->ti->error = "Not enough devices to satisfy specification";
rs               1498 drivers/md/dm-raid.c 		rs->md.new_layout = raid10_format_to_md_layout(rs, raid10_format, raid10_copies);
rs               1499 drivers/md/dm-raid.c 		if (rs->md.new_layout < 0) {
rs               1500 drivers/md/dm-raid.c 			rs->ti->error = "Error getting raid10 format";
rs               1501 drivers/md/dm-raid.c 			return rs->md.new_layout;
rs               1504 drivers/md/dm-raid.c 		rt = get_raid_type_by_ll(10, rs->md.new_layout);
rs               1506 drivers/md/dm-raid.c 			rs->ti->error = "Failed to recognize new raid10 layout";
rs               1512 drivers/md/dm-raid.c 		    test_bit(__CTR_FLAG_RAID10_USE_NEAR_SETS, &rs->ctr_flags)) {
rs               1513 drivers/md/dm-raid.c 			rs->ti->error = "RAID10 format 'near' and 'raid10_use_near_sets' are incompatible";
rs               1518 drivers/md/dm-raid.c 	rs->raid10_copies = raid10_copies;
rs               1521 drivers/md/dm-raid.c 	rs->md.persistent = 0;
rs               1522 drivers/md/dm-raid.c 	rs->md.external = 1;
rs               1525 drivers/md/dm-raid.c 	return rs_check_for_valid_flags(rs);
rs               1529 drivers/md/dm-raid.c static int rs_set_raid456_stripe_cache(struct raid_set *rs)
rs               1533 drivers/md/dm-raid.c 	struct mddev *mddev = &rs->md;
rs               1535 drivers/md/dm-raid.c 	uint32_t nr_stripes = rs->stripe_cache_entries;
rs               1537 drivers/md/dm-raid.c 	if (!rt_is_raid456(rs->raid_type)) {
rs               1538 drivers/md/dm-raid.c 		rs->ti->error = "Inappropriate raid level; cannot change stripe_cache size";
rs               1550 drivers/md/dm-raid.c 		rs->ti->error = "Cannot change stripe_cache size on inactive RAID set";
rs               1558 drivers/md/dm-raid.c 			rs->ti->error = "Failed to set raid4/5/6 stripe cache size";
rs               1569 drivers/md/dm-raid.c static unsigned int mddev_data_stripes(struct raid_set *rs)
rs               1571 drivers/md/dm-raid.c 	return rs->md.raid_disks - rs->raid_type->parity_devs;
rs               1575 drivers/md/dm-raid.c static unsigned int rs_data_stripes(struct raid_set *rs)
rs               1577 drivers/md/dm-raid.c 	return rs->raid_disks - rs->raid_type->parity_devs;
rs               1584 drivers/md/dm-raid.c static sector_t __rdev_sectors(struct raid_set *rs)
rs               1588 drivers/md/dm-raid.c 	for (i = 0; i < rs->md.raid_disks; i++) {
rs               1589 drivers/md/dm-raid.c 		struct md_rdev *rdev = &rs->dev[i].rdev;
rs               1600 drivers/md/dm-raid.c static int _check_data_dev_sectors(struct raid_set *rs)
rs               1605 drivers/md/dm-raid.c 	rdev_for_each(rdev, &rs->md)
rs               1608 drivers/md/dm-raid.c 			if (ds < rs->md.dev_sectors) {
rs               1609 drivers/md/dm-raid.c 				rs->ti->error = "Component device(s) too small";
rs               1618 drivers/md/dm-raid.c static int rs_set_dev_and_array_sectors(struct raid_set *rs, bool use_mddev)
rs               1622 drivers/md/dm-raid.c 	struct mddev *mddev = &rs->md;
rs               1624 drivers/md/dm-raid.c 	sector_t array_sectors = rs->ti->len, dev_sectors = rs->ti->len;
rs               1628 drivers/md/dm-raid.c 		data_stripes = mddev_data_stripes(rs);
rs               1630 drivers/md/dm-raid.c 		delta_disks = rs->delta_disks;
rs               1631 drivers/md/dm-raid.c 		data_stripes = rs_data_stripes(rs);
rs               1635 drivers/md/dm-raid.c 	if (rt_is_raid1(rs->raid_type))
rs               1637 drivers/md/dm-raid.c 	else if (rt_is_raid10(rs->raid_type)) {
rs               1638 drivers/md/dm-raid.c 		if (rs->raid10_copies < 2 ||
rs               1640 drivers/md/dm-raid.c 			rs->ti->error = "Bogus raid10 data copies or delta disks";
rs               1644 drivers/md/dm-raid.c 		dev_sectors *= rs->raid10_copies;
rs               1649 drivers/md/dm-raid.c 		if (sector_div(array_sectors, rs->raid10_copies))
rs               1666 drivers/md/dm-raid.c 	return _check_data_dev_sectors(rs);
rs               1668 drivers/md/dm-raid.c 	rs->ti->error = "Target length not divisible by number of data devices";
rs               1673 drivers/md/dm-raid.c static void __rs_setup_recovery(struct raid_set *rs, sector_t dev_sectors)
rs               1676 drivers/md/dm-raid.c 	if (rs_is_raid0(rs))
rs               1677 drivers/md/dm-raid.c 		rs->md.recovery_cp = MaxSector;
rs               1683 drivers/md/dm-raid.c 	else if (rs_is_raid6(rs))
rs               1684 drivers/md/dm-raid.c 		rs->md.recovery_cp = dev_sectors;
rs               1690 drivers/md/dm-raid.c 		rs->md.recovery_cp = test_bit(__CTR_FLAG_NOSYNC, &rs->ctr_flags)
rs               1695 drivers/md/dm-raid.c static void rs_setup_recovery(struct raid_set *rs, sector_t dev_sectors)
rs               1699 drivers/md/dm-raid.c 		__rs_setup_recovery(rs, 0);
rs               1702 drivers/md/dm-raid.c 		__rs_setup_recovery(rs, MaxSector);
rs               1703 drivers/md/dm-raid.c 	else if (__rdev_sectors(rs) < dev_sectors)
rs               1705 drivers/md/dm-raid.c 		__rs_setup_recovery(rs, __rdev_sectors(rs));
rs               1707 drivers/md/dm-raid.c 		__rs_setup_recovery(rs, MaxSector);
rs               1712 drivers/md/dm-raid.c 	struct raid_set *rs = container_of(ws, struct raid_set, md.event_work);
rs               1715 drivers/md/dm-raid.c 	if (!rs_is_reshaping(rs)) {
rs               1716 drivers/md/dm-raid.c 		if (rs_is_raid10(rs))
rs               1717 drivers/md/dm-raid.c 			rs_set_rdev_sectors(rs);
rs               1718 drivers/md/dm-raid.c 		rs_set_capacity(rs);
rs               1720 drivers/md/dm-raid.c 	dm_table_event(rs->ti->table);
rs               1725 drivers/md/dm-raid.c 	struct raid_set *rs = container_of(cb, struct raid_set, callbacks);
rs               1727 drivers/md/dm-raid.c 	return mddev_congested(&rs->md, bits);
rs               1736 drivers/md/dm-raid.c static int rs_check_takeover(struct raid_set *rs)
rs               1738 drivers/md/dm-raid.c 	struct mddev *mddev = &rs->md;
rs               1741 drivers/md/dm-raid.c 	if (rs->md.degraded) {
rs               1742 drivers/md/dm-raid.c 		rs->ti->error = "Can't takeover degraded raid set";
rs               1746 drivers/md/dm-raid.c 	if (rs_is_reshaping(rs)) {
rs               1747 drivers/md/dm-raid.c 		rs->ti->error = "Can't takeover reshaping raid set";
rs               1760 drivers/md/dm-raid.c 		    !(rs->raid_disks % mddev->raid_disks))
rs               1885 drivers/md/dm-raid.c 	rs->ti->error = "takeover not possible";
rs               1890 drivers/md/dm-raid.c static bool rs_takeover_requested(struct raid_set *rs)
rs               1892 drivers/md/dm-raid.c 	return rs->md.new_level != rs->md.level;
rs               1896 drivers/md/dm-raid.c static bool rs_reshape_requested(struct raid_set *rs)
rs               1899 drivers/md/dm-raid.c 	struct mddev *mddev = &rs->md;
rs               1901 drivers/md/dm-raid.c 	if (rs_takeover_requested(rs))
rs               1904 drivers/md/dm-raid.c 	if (rs_is_raid0(rs))
rs               1909 drivers/md/dm-raid.c 		 rs->delta_disks;
rs               1912 drivers/md/dm-raid.c 	if (rs_is_raid1(rs)) {
rs               1913 drivers/md/dm-raid.c 		if (rs->delta_disks)
rs               1914 drivers/md/dm-raid.c 			return !!rs->delta_disks;
rs               1917 drivers/md/dm-raid.c 		       mddev->raid_disks != rs->raid_disks;
rs               1920 drivers/md/dm-raid.c 	if (rs_is_raid10(rs))
rs               1923 drivers/md/dm-raid.c 		       rs->delta_disks >= 0;
rs               2027 drivers/md/dm-raid.c static int rs_check_reshape(struct raid_set *rs)
rs               2029 drivers/md/dm-raid.c 	struct mddev *mddev = &rs->md;
rs               2032 drivers/md/dm-raid.c 		rs->ti->error = "Reshape not supported";
rs               2034 drivers/md/dm-raid.c 		rs->ti->error = "Can't reshape degraded raid set";
rs               2035 drivers/md/dm-raid.c 	else if (rs_is_recovering(rs))
rs               2036 drivers/md/dm-raid.c 		rs->ti->error = "Convert request on recovering raid set prohibited";
rs               2037 drivers/md/dm-raid.c 	else if (rs_is_reshaping(rs))
rs               2038 drivers/md/dm-raid.c 		rs->ti->error = "raid set already reshaping!";
rs               2039 drivers/md/dm-raid.c 	else if (!(rs_is_raid1(rs) || rs_is_raid10(rs) || rs_is_raid456(rs)))
rs               2040 drivers/md/dm-raid.c 		rs->ti->error = "Reshaping only supported for raid1/4/5/6/10";
rs               2102 drivers/md/dm-raid.c 	struct raid_set *rs = container_of(mddev, struct raid_set, md);
rs               2114 drivers/md/dm-raid.c 	for (i = 0; i < rs->raid_disks; i++)
rs               2115 drivers/md/dm-raid.c 		if (!rs->dev[i].data_dev || test_bit(Faulty, &rs->dev[i].rdev.flags)) {
rs               2223 drivers/md/dm-raid.c static int super_init_validation(struct raid_set *rs, struct md_rdev *rdev)
rs               2227 drivers/md/dm-raid.c 	struct mddev *mddev = &rs->md;
rs               2264 drivers/md/dm-raid.c 			if (test_bit(__CTR_FLAG_DELTA_DISKS, &rs->ctr_flags)) {
rs               2276 drivers/md/dm-raid.c 			rs->raid_type = get_raid_type_by_ll(mddev->level, mddev->layout);
rs               2286 drivers/md/dm-raid.c 		if (rs_takeover_requested(rs)) {
rs               2293 drivers/md/dm-raid.c 		} else if (rs_reshape_requested(rs)) {
rs               2306 drivers/md/dm-raid.c 			if (rs->delta_disks)
rs               2308 drivers/md/dm-raid.c 				      mddev->raid_disks, mddev->raid_disks + rs->delta_disks);
rs               2309 drivers/md/dm-raid.c 			if (rs_is_raid10(rs)) {
rs               2323 drivers/md/dm-raid.c 	if (!test_bit(__CTR_FLAG_NOSYNC, &rs->ctr_flags))
rs               2361 drivers/md/dm-raid.c 	if (new_devs == rs->raid_disks || !rebuilds) {
rs               2363 drivers/md/dm-raid.c 		if (new_devs == 1 && !rs->delta_disks)
rs               2365 drivers/md/dm-raid.c 		if (new_devs == rs->raid_disks) {
rs               2369 drivers/md/dm-raid.c 			   new_devs != rs->delta_disks) {
rs               2384 drivers/md/dm-raid.c 		} else if (!test_bit(__CTR_FLAG_REBUILD, &rs->ctr_flags) && rs_is_recovering(rs)) {
rs               2388 drivers/md/dm-raid.c 		} else if (rs_is_reshaping(rs)) {
rs               2417 drivers/md/dm-raid.c 				if (rs_is_raid10(rs) && __is_raid10_near(mddev->layout)) {
rs               2419 drivers/md/dm-raid.c 					    rs->raid_disks % rs->raid10_copies) {
rs               2420 drivers/md/dm-raid.c 						rs->ti->error =
rs               2427 drivers/md/dm-raid.c 				} else if (!(rs_is_raid10(rs) && rt_is_raid0(rs->raid_type)) &&
rs               2428 drivers/md/dm-raid.c 					   !(rs_is_raid0(rs) && rt_is_raid10(rs->raid_type)) &&
rs               2429 drivers/md/dm-raid.c 					   !rt_is_raid1(rs->raid_type)) {
rs               2430 drivers/md/dm-raid.c 					rs->ti->error = "Cannot change device positions in raid set";
rs               2449 drivers/md/dm-raid.c static int super_validate(struct raid_set *rs, struct md_rdev *rdev)
rs               2451 drivers/md/dm-raid.c 	struct mddev *mddev = &rs->md;
rs               2454 drivers/md/dm-raid.c 	if (rs_is_raid0(rs) || !rdev->sb_page || rdev->raid_disk < 0)
rs               2463 drivers/md/dm-raid.c 	if (!mddev->events && super_init_validation(rs, rdev))
rs               2468 drivers/md/dm-raid.c 		rs->ti->error = "Unable to assemble array: Unknown flag(s) in compatible feature flags";
rs               2473 drivers/md/dm-raid.c 		rs->ti->error = "Unable to assemble array: No incompatible feature flags supported yet";
rs               2478 drivers/md/dm-raid.c 	mddev->bitmap_info.offset = (rt_is_raid0(rs->raid_type) || rs->journal_dev.dev) ? 0 : to_sector(4096);
rs               2497 drivers/md/dm-raid.c 		else if (!rs_is_reshaping(rs))
rs               2520 drivers/md/dm-raid.c static int analyse_superblocks(struct dm_target *ti, struct raid_set *rs)
rs               2524 drivers/md/dm-raid.c 	struct mddev *mddev = &rs->md;
rs               2551 drivers/md/dm-raid.c 		if (test_bit(__CTR_FLAG_SYNC, &rs->ctr_flags))
rs               2568 drivers/md/dm-raid.c 			if (rs_is_raid0(rs))
rs               2591 drivers/md/dm-raid.c 	rs->ti->error = "Unable to assemble array: Invalid superblocks";
rs               2592 drivers/md/dm-raid.c 	if (super_validate(rs, freshest))
rs               2595 drivers/md/dm-raid.c 	if (validate_raid_redundancy(rs)) {
rs               2596 drivers/md/dm-raid.c 		rs->ti->error = "Insufficient redundancy to activate array";
rs               2603 drivers/md/dm-raid.c 		    super_validate(rs, rdev))
rs               2616 drivers/md/dm-raid.c static int rs_adjust_data_offsets(struct raid_set *rs)
rs               2622 drivers/md/dm-raid.c 	if (!test_bit(__CTR_FLAG_DATA_OFFSET, &rs->ctr_flags)) {
rs               2623 drivers/md/dm-raid.c 		if (!rs_is_reshapable(rs))
rs               2630 drivers/md/dm-raid.c 	rdev = &rs->dev[0].rdev;
rs               2632 drivers/md/dm-raid.c 	if (rs->delta_disks < 0) {
rs               2642 drivers/md/dm-raid.c 		new_data_offset = rs->data_offset;
rs               2644 drivers/md/dm-raid.c 	} else if (rs->delta_disks > 0) {
rs               2653 drivers/md/dm-raid.c 		data_offset = rs->data_offset;
rs               2675 drivers/md/dm-raid.c 		data_offset = rs->data_offset ? rdev->data_offset : 0;
rs               2676 drivers/md/dm-raid.c 		new_data_offset = data_offset ? 0 : rs->data_offset;
rs               2677 drivers/md/dm-raid.c 		set_bit(RT_FLAG_UPDATE_SBS, &rs->runtime_flags);
rs               2683 drivers/md/dm-raid.c 	if (rs->data_offset &&
rs               2684 drivers/md/dm-raid.c 	    to_sector(i_size_read(rdev->bdev->bd_inode)) - rs->md.dev_sectors < MIN_FREE_RESHAPE_SPACE) {
rs               2685 drivers/md/dm-raid.c 		rs->ti->error = data_offset ? "No space for forward reshape" :
rs               2694 drivers/md/dm-raid.c 	if (rs->md.recovery_cp < rs->md.dev_sectors)
rs               2695 drivers/md/dm-raid.c 		rs->md.recovery_cp += rs->dev[0].rdev.data_offset;
rs               2698 drivers/md/dm-raid.c 	rdev_for_each(rdev, &rs->md) {
rs               2709 drivers/md/dm-raid.c static void __reorder_raid_disk_indexes(struct raid_set *rs)
rs               2714 drivers/md/dm-raid.c 	rdev_for_each(rdev, &rs->md) {
rs               2725 drivers/md/dm-raid.c static int rs_setup_takeover(struct raid_set *rs)
rs               2727 drivers/md/dm-raid.c 	struct mddev *mddev = &rs->md;
rs               2729 drivers/md/dm-raid.c 	unsigned int d = mddev->raid_disks = rs->raid_disks;
rs               2730 drivers/md/dm-raid.c 	sector_t new_data_offset = rs->dev[0].rdev.data_offset ? 0 : rs->data_offset;
rs               2732 drivers/md/dm-raid.c 	if (rt_is_raid10(rs->raid_type)) {
rs               2733 drivers/md/dm-raid.c 		if (rs_is_raid0(rs)) {
rs               2735 drivers/md/dm-raid.c 			__reorder_raid_disk_indexes(rs);
rs               2738 drivers/md/dm-raid.c 			mddev->layout = raid10_format_to_md_layout(rs, ALGORITHM_RAID10_FAR,
rs               2739 drivers/md/dm-raid.c 								   rs->raid10_copies);
rs               2740 drivers/md/dm-raid.c 		} else if (rs_is_raid1(rs))
rs               2742 drivers/md/dm-raid.c 			mddev->layout = raid10_format_to_md_layout(rs, ALGORITHM_RAID10_NEAR,
rs               2743 drivers/md/dm-raid.c 								   rs->raid_disks);
rs               2753 drivers/md/dm-raid.c 		rdev = &rs->dev[d].rdev;
rs               2755 drivers/md/dm-raid.c 		if (test_bit(d, (void *) rs->rebuild_disks)) {
rs               2770 drivers/md/dm-raid.c static int rs_prepare_reshape(struct raid_set *rs)
rs               2773 drivers/md/dm-raid.c 	struct mddev *mddev = &rs->md;
rs               2775 drivers/md/dm-raid.c 	if (rs_is_raid10(rs)) {
rs               2776 drivers/md/dm-raid.c 		if (rs->raid_disks != mddev->raid_disks &&
rs               2778 drivers/md/dm-raid.c 		    rs->raid10_copies &&
rs               2779 drivers/md/dm-raid.c 		    rs->raid10_copies != __raid10_near_copies(mddev->layout)) {
rs               2786 drivers/md/dm-raid.c 			if (rs->raid_disks % rs->raid10_copies) {
rs               2787 drivers/md/dm-raid.c 				rs->ti->error = "Can't reshape raid10 mirror groups";
rs               2792 drivers/md/dm-raid.c 			__reorder_raid_disk_indexes(rs);
rs               2793 drivers/md/dm-raid.c 			mddev->layout = raid10_format_to_md_layout(rs, ALGORITHM_RAID10_NEAR,
rs               2794 drivers/md/dm-raid.c 								   rs->raid10_copies);
rs               2800 drivers/md/dm-raid.c 	} else if (rs_is_raid456(rs))
rs               2803 drivers/md/dm-raid.c 	else if (rs_is_raid1(rs)) {
rs               2804 drivers/md/dm-raid.c 		if (rs->delta_disks) {
rs               2806 drivers/md/dm-raid.c 			mddev->degraded = rs->delta_disks < 0 ? -rs->delta_disks : rs->delta_disks;
rs               2810 drivers/md/dm-raid.c 			mddev->raid_disks = rs->raid_disks;
rs               2814 drivers/md/dm-raid.c 		rs->ti->error = "Called with bogus raid type";
rs               2819 drivers/md/dm-raid.c 		set_bit(RT_FLAG_RESHAPE_RS, &rs->runtime_flags);
rs               2820 drivers/md/dm-raid.c 		set_bit(RT_FLAG_UPDATE_SBS, &rs->runtime_flags);
rs               2821 drivers/md/dm-raid.c 	} else if (mddev->raid_disks < rs->raid_disks)
rs               2823 drivers/md/dm-raid.c 		set_bit(RT_FLAG_UPDATE_SBS, &rs->runtime_flags);
rs               2829 drivers/md/dm-raid.c static sector_t _get_reshape_sectors(struct raid_set *rs)
rs               2834 drivers/md/dm-raid.c 	rdev_for_each(rdev, &rs->md)
rs               2842 drivers/md/dm-raid.c 	return max(reshape_sectors, (sector_t) rs->data_offset);
rs               2852 drivers/md/dm-raid.c static int rs_setup_reshape(struct raid_set *rs)
rs               2856 drivers/md/dm-raid.c 	sector_t reshape_sectors = _get_reshape_sectors(rs);
rs               2857 drivers/md/dm-raid.c 	struct mddev *mddev = &rs->md;
rs               2860 drivers/md/dm-raid.c 	mddev->delta_disks = rs->delta_disks;
rs               2866 drivers/md/dm-raid.c 		DMINFO("Ignoring invalid layout change with delta_disks=%d", rs->delta_disks);
rs               2893 drivers/md/dm-raid.c 	if (rs->delta_disks > 0) {
rs               2895 drivers/md/dm-raid.c 		for (d = cur_raid_devs; d < rs->raid_disks; d++) {
rs               2896 drivers/md/dm-raid.c 			rdev = &rs->dev[d].rdev;
rs               2907 drivers/md/dm-raid.c 			rdev->recovery_offset = rs_is_raid1(rs) ? 0 : MaxSector;
rs               2913 drivers/md/dm-raid.c 	} else if (rs->delta_disks < 0) {
rs               2914 drivers/md/dm-raid.c 		r = rs_set_dev_and_array_sectors(rs, true);
rs               2940 drivers/md/dm-raid.c 		mddev->reshape_backwards = rs->dev[0].rdev.data_offset ? 0 : 1;
rs               2948 drivers/md/dm-raid.c 		rdev_for_each(rdev, &rs->md)
rs               2959 drivers/md/dm-raid.c static void configure_discard_support(struct raid_set *rs)
rs               2963 drivers/md/dm-raid.c 	struct dm_target *ti = rs->ti;
rs               2968 drivers/md/dm-raid.c 	raid456 = rs_is_raid456(rs);
rs               2970 drivers/md/dm-raid.c 	for (i = 0; i < rs->raid_disks; i++) {
rs               2973 drivers/md/dm-raid.c 		if (!rs->dev[i].rdev.bdev)
rs               2976 drivers/md/dm-raid.c 		q = bdev_get_queue(rs->dev[i].rdev.bdev);
rs               3012 drivers/md/dm-raid.c 	struct raid_set *rs = NULL;
rs               3050 drivers/md/dm-raid.c 	rs = raid_set_alloc(ti, rt, num_raid_devs);
rs               3051 drivers/md/dm-raid.c 	if (IS_ERR(rs))
rs               3052 drivers/md/dm-raid.c 		return PTR_ERR(rs);
rs               3054 drivers/md/dm-raid.c 	r = parse_raid_params(rs, &as, num_raid_params);
rs               3058 drivers/md/dm-raid.c 	r = parse_dev_params(rs, &as);
rs               3062 drivers/md/dm-raid.c 	rs->md.sync_super = super_sync;
rs               3070 drivers/md/dm-raid.c 	r = rs_set_dev_and_array_sectors(rs, false);
rs               3074 drivers/md/dm-raid.c 	calculated_dev_sectors = rs->md.dev_sectors;
rs               3081 drivers/md/dm-raid.c 	rs_config_backup(rs, &rs_layout);
rs               3083 drivers/md/dm-raid.c 	r = analyse_superblocks(ti, rs);
rs               3087 drivers/md/dm-raid.c 	rdev_sectors = __rdev_sectors(rs);
rs               3095 drivers/md/dm-raid.c 	reshape_sectors = _get_reshape_sectors(rs);
rs               3099 drivers/md/dm-raid.c 	INIT_WORK(&rs->md.event_work, do_table_event);
rs               3100 drivers/md/dm-raid.c 	ti->private = rs;
rs               3104 drivers/md/dm-raid.c 	rs_config_restore(rs, &rs_layout);
rs               3112 drivers/md/dm-raid.c 	if (test_bit(MD_ARRAY_FIRST_USE, &rs->md.flags)) {
rs               3114 drivers/md/dm-raid.c 		if (rs_is_raid6(rs) &&
rs               3115 drivers/md/dm-raid.c 		    test_bit(__CTR_FLAG_NOSYNC, &rs->ctr_flags)) {
rs               3120 drivers/md/dm-raid.c 		rs_setup_recovery(rs, 0);
rs               3121 drivers/md/dm-raid.c 		set_bit(RT_FLAG_UPDATE_SBS, &rs->runtime_flags);
rs               3122 drivers/md/dm-raid.c 		rs_set_new(rs);
rs               3123 drivers/md/dm-raid.c 	} else if (rs_is_recovering(rs)) {
rs               3125 drivers/md/dm-raid.c 		if (test_bit(__CTR_FLAG_REBUILD, &rs->ctr_flags)) {
rs               3126 drivers/md/dm-raid.c 			set_bit(RT_FLAG_UPDATE_SBS, &rs->runtime_flags);
rs               3127 drivers/md/dm-raid.c 			rs_setup_recovery(rs, MaxSector);
rs               3131 drivers/md/dm-raid.c 	} else if (rs_is_reshaping(rs)) {
rs               3139 drivers/md/dm-raid.c 	} else if (rs_takeover_requested(rs)) {
rs               3140 drivers/md/dm-raid.c 		if (rs_is_reshaping(rs)) {
rs               3147 drivers/md/dm-raid.c 		if (test_bit(__CTR_FLAG_JOURNAL_DEV, &rs->ctr_flags)) {
rs               3161 drivers/md/dm-raid.c 		r = rs_check_takeover(rs);
rs               3165 drivers/md/dm-raid.c 		r = rs_setup_takeover(rs);
rs               3169 drivers/md/dm-raid.c 		set_bit(RT_FLAG_UPDATE_SBS, &rs->runtime_flags);
rs               3171 drivers/md/dm-raid.c 		rs_setup_recovery(rs, MaxSector);
rs               3172 drivers/md/dm-raid.c 		rs_set_new(rs);
rs               3173 drivers/md/dm-raid.c 	} else if (rs_reshape_requested(rs)) {
rs               3180 drivers/md/dm-raid.c 		if (test_bit(__CTR_FLAG_JOURNAL_DEV, &rs->ctr_flags)) {
rs               3187 drivers/md/dm-raid.c 		if (reshape_sectors || rs_is_raid1(rs)) {
rs               3195 drivers/md/dm-raid.c 			r = rs_prepare_reshape(rs);
rs               3200 drivers/md/dm-raid.c 			rs_setup_recovery(rs, MaxSector);
rs               3202 drivers/md/dm-raid.c 		rs_set_cur(rs);
rs               3205 drivers/md/dm-raid.c 		if (test_bit(__CTR_FLAG_REBUILD, &rs->ctr_flags)) {
rs               3206 drivers/md/dm-raid.c 			rs_setup_recovery(rs, MaxSector);
rs               3207 drivers/md/dm-raid.c 			set_bit(RT_FLAG_UPDATE_SBS, &rs->runtime_flags);
rs               3209 drivers/md/dm-raid.c 			rs_setup_recovery(rs, test_bit(__CTR_FLAG_SYNC, &rs->ctr_flags) ?
rs               3211 drivers/md/dm-raid.c 		rs_set_cur(rs);
rs               3215 drivers/md/dm-raid.c 	r = rs_adjust_data_offsets(rs);
rs               3220 drivers/md/dm-raid.c 	rs->md.ro = 1;
rs               3221 drivers/md/dm-raid.c 	rs->md.in_sync = 1;
rs               3224 drivers/md/dm-raid.c 	set_bit(MD_RECOVERY_FROZEN, &rs->md.recovery);
rs               3227 drivers/md/dm-raid.c 	mddev_lock_nointr(&rs->md);
rs               3228 drivers/md/dm-raid.c 	r = md_run(&rs->md);
rs               3229 drivers/md/dm-raid.c 	rs->md.in_sync = 0; /* Assume already marked dirty */
rs               3232 drivers/md/dm-raid.c 		mddev_unlock(&rs->md);
rs               3236 drivers/md/dm-raid.c 	r = md_start(&rs->md);
rs               3240 drivers/md/dm-raid.c 		mddev_unlock(&rs->md);
rs               3244 drivers/md/dm-raid.c 	rs->callbacks.congested_fn = raid_is_congested;
rs               3245 drivers/md/dm-raid.c 	dm_table_add_target_callbacks(ti->table, &rs->callbacks);
rs               3248 drivers/md/dm-raid.c 	if (test_bit(__CTR_FLAG_JOURNAL_MODE, &rs->ctr_flags)) {
rs               3249 drivers/md/dm-raid.c 		r = r5c_journal_mode_set(&rs->md, rs->journal_dev.mode);
rs               3252 drivers/md/dm-raid.c 			mddev_unlock(&rs->md);
rs               3257 drivers/md/dm-raid.c 	mddev_suspend(&rs->md);
rs               3258 drivers/md/dm-raid.c 	set_bit(RT_FLAG_RS_SUSPENDED, &rs->runtime_flags);
rs               3261 drivers/md/dm-raid.c 	if (rs_is_raid456(rs)) {
rs               3262 drivers/md/dm-raid.c 		r = rs_set_raid456_stripe_cache(rs);
rs               3268 drivers/md/dm-raid.c 	if (test_bit(RT_FLAG_RESHAPE_RS, &rs->runtime_flags)) {
rs               3269 drivers/md/dm-raid.c 		r = rs_check_reshape(rs);
rs               3274 drivers/md/dm-raid.c 		rs_config_restore(rs, &rs_layout);
rs               3276 drivers/md/dm-raid.c 		if (rs->md.pers->start_reshape) {
rs               3277 drivers/md/dm-raid.c 			r = rs->md.pers->check_reshape(&rs->md);
rs               3286 drivers/md/dm-raid.c 	configure_discard_support(rs);
rs               3288 drivers/md/dm-raid.c 	mddev_unlock(&rs->md);
rs               3295 drivers/md/dm-raid.c 	md_stop(&rs->md);
rs               3297 drivers/md/dm-raid.c 	raid_set_free(rs);
rs               3304 drivers/md/dm-raid.c 	struct raid_set *rs = ti->private;
rs               3306 drivers/md/dm-raid.c 	list_del_init(&rs->callbacks.list);
rs               3307 drivers/md/dm-raid.c 	md_stop(&rs->md);
rs               3308 drivers/md/dm-raid.c 	raid_set_free(rs);
rs               3313 drivers/md/dm-raid.c 	struct raid_set *rs = ti->private;
rs               3314 drivers/md/dm-raid.c 	struct mddev *mddev = &rs->md;
rs               3391 drivers/md/dm-raid.c static const char *__raid_dev_status(struct raid_set *rs, struct md_rdev *rdev)
rs               3398 drivers/md/dm-raid.c 		return (rs->journal_dev.mode == R5C_JOURNAL_MODE_WRITE_THROUGH) ? "A" : "a";
rs               3399 drivers/md/dm-raid.c 	else if (test_bit(RT_FLAG_RS_RESYNCING, &rs->runtime_flags) ||
rs               3400 drivers/md/dm-raid.c 		 (!test_bit(RT_FLAG_RS_IN_SYNC, &rs->runtime_flags) &&
rs               3408 drivers/md/dm-raid.c static sector_t rs_get_progress(struct raid_set *rs, unsigned long recovery,
rs               3413 drivers/md/dm-raid.c 	struct mddev *mddev = &rs->md;
rs               3415 drivers/md/dm-raid.c 	clear_bit(RT_FLAG_RS_IN_SYNC, &rs->runtime_flags);
rs               3416 drivers/md/dm-raid.c 	clear_bit(RT_FLAG_RS_RESYNCING, &rs->runtime_flags);
rs               3418 drivers/md/dm-raid.c 	if (rs_is_raid0(rs)) {
rs               3420 drivers/md/dm-raid.c 		set_bit(RT_FLAG_RS_IN_SYNC, &rs->runtime_flags);
rs               3436 drivers/md/dm-raid.c 				set_bit(RT_FLAG_RS_IN_SYNC, &rs->runtime_flags);
rs               3450 drivers/md/dm-raid.c 			set_bit(RT_FLAG_RS_RESYNCING, &rs->runtime_flags);
rs               3457 drivers/md/dm-raid.c 			set_bit(RT_FLAG_RS_RESYNCING, &rs->runtime_flags);
rs               3465 drivers/md/dm-raid.c 			set_bit(RT_FLAG_RS_IN_SYNC, &rs->runtime_flags);
rs               3475 drivers/md/dm-raid.c 				set_bit(RT_FLAG_RS_RESYNCING, &rs->runtime_flags);
rs               3483 drivers/md/dm-raid.c 			set_bit(RT_FLAG_RS_IN_SYNC, &rs->runtime_flags);
rs               3487 drivers/md/dm-raid.c 					clear_bit(RT_FLAG_RS_IN_SYNC, &rs->runtime_flags);
rs               3505 drivers/md/dm-raid.c 	struct raid_set *rs = ti->private;
rs               3506 drivers/md/dm-raid.c 	struct mddev *mddev = &rs->md;
rs               3529 drivers/md/dm-raid.c 		recovery = rs->md.recovery;
rs               3531 drivers/md/dm-raid.c 		resync_max_sectors = test_bit(RT_FLAG_RS_PRERESUMED, &rs->runtime_flags) ?
rs               3533 drivers/md/dm-raid.c 		progress = rs_get_progress(rs, recovery, resync_max_sectors);
rs               3536 drivers/md/dm-raid.c 		sync_action = sync_str(decipher_sync_action(&rs->md, recovery));
rs               3539 drivers/md/dm-raid.c 		for (i = 0; i < rs->raid_disks; i++)
rs               3540 drivers/md/dm-raid.c 			DMEMIT(__raid_dev_status(rs, &rs->dev[i].rdev));
rs               3585 drivers/md/dm-raid.c 		DMEMIT(" %llu", (unsigned long long) rs->dev[0].rdev.data_offset);
rs               3590 drivers/md/dm-raid.c 		DMEMIT(" %s", test_bit(__CTR_FLAG_JOURNAL_DEV, &rs->ctr_flags) ?
rs               3591 drivers/md/dm-raid.c 			      __raid_dev_status(rs, &rs->journal_dev.rdev) : "-");
rs               3598 drivers/md/dm-raid.c 		for (i = 0; i < rs->raid_disks; i++)
rs               3599 drivers/md/dm-raid.c 			if (test_bit(WriteMostly, &rs->dev[i].rdev.flags))
rs               3601 drivers/md/dm-raid.c 		rebuild_disks = memweight(rs->rebuild_disks, DISKS_ARRAY_ELEMS * sizeof(*rs->rebuild_disks));
rs               3604 drivers/md/dm-raid.c 				  hweight32(rs->ctr_flags & CTR_FLAG_OPTIONS_NO_ARGS) +
rs               3605 drivers/md/dm-raid.c 				  hweight32(rs->ctr_flags & CTR_FLAG_OPTIONS_ONE_ARG) * 2 +
rs               3606 drivers/md/dm-raid.c 				  (test_bit(__CTR_FLAG_JOURNAL_DEV, &rs->ctr_flags) ? 2 : 0) +
rs               3607 drivers/md/dm-raid.c 				  (test_bit(__CTR_FLAG_JOURNAL_MODE, &rs->ctr_flags) ? 2 : 0);
rs               3611 drivers/md/dm-raid.c 		DMEMIT("%s %u %u", rs->raid_type->name, raid_param_cnt, mddev->new_chunk_sectors);
rs               3612 drivers/md/dm-raid.c 		if (test_bit(__CTR_FLAG_SYNC, &rs->ctr_flags))
rs               3614 drivers/md/dm-raid.c 		if (test_bit(__CTR_FLAG_NOSYNC, &rs->ctr_flags))
rs               3617 drivers/md/dm-raid.c 			for (i = 0; i < rs->raid_disks; i++)
rs               3618 drivers/md/dm-raid.c 				if (test_bit(rs->dev[i].rdev.raid_disk, (void *) rs->rebuild_disks))
rs               3620 drivers/md/dm-raid.c 							 rs->dev[i].rdev.raid_disk);
rs               3621 drivers/md/dm-raid.c 		if (test_bit(__CTR_FLAG_DAEMON_SLEEP, &rs->ctr_flags))
rs               3624 drivers/md/dm-raid.c 		if (test_bit(__CTR_FLAG_MIN_RECOVERY_RATE, &rs->ctr_flags))
rs               3627 drivers/md/dm-raid.c 		if (test_bit(__CTR_FLAG_MAX_RECOVERY_RATE, &rs->ctr_flags))
rs               3631 drivers/md/dm-raid.c 			for (i = 0; i < rs->raid_disks; i++)
rs               3632 drivers/md/dm-raid.c 				if (test_bit(WriteMostly, &rs->dev[i].rdev.flags))
rs               3634 drivers/md/dm-raid.c 					       rs->dev[i].rdev.raid_disk);
rs               3635 drivers/md/dm-raid.c 		if (test_bit(__CTR_FLAG_MAX_WRITE_BEHIND, &rs->ctr_flags))
rs               3638 drivers/md/dm-raid.c 		if (test_bit(__CTR_FLAG_STRIPE_CACHE, &rs->ctr_flags))
rs               3641 drivers/md/dm-raid.c 		if (test_bit(__CTR_FLAG_REGION_SIZE, &rs->ctr_flags))
rs               3644 drivers/md/dm-raid.c 		if (test_bit(__CTR_FLAG_RAID10_COPIES, &rs->ctr_flags))
rs               3647 drivers/md/dm-raid.c 		if (test_bit(__CTR_FLAG_RAID10_FORMAT, &rs->ctr_flags))
rs               3650 drivers/md/dm-raid.c 		if (test_bit(__CTR_FLAG_DELTA_DISKS, &rs->ctr_flags))
rs               3652 drivers/md/dm-raid.c 					 max(rs->delta_disks, mddev->delta_disks));
rs               3653 drivers/md/dm-raid.c 		if (test_bit(__CTR_FLAG_DATA_OFFSET, &rs->ctr_flags))
rs               3655 drivers/md/dm-raid.c 					   (unsigned long long) rs->data_offset);
rs               3656 drivers/md/dm-raid.c 		if (test_bit(__CTR_FLAG_JOURNAL_DEV, &rs->ctr_flags))
rs               3658 drivers/md/dm-raid.c 					__get_dev_name(rs->journal_dev.dev));
rs               3659 drivers/md/dm-raid.c 		if (test_bit(__CTR_FLAG_JOURNAL_MODE, &rs->ctr_flags))
rs               3661 drivers/md/dm-raid.c 					 md_journal_mode_to_dm_raid(rs->journal_dev.mode));
rs               3662 drivers/md/dm-raid.c 		DMEMIT(" %d", rs->raid_disks);
rs               3663 drivers/md/dm-raid.c 		for (i = 0; i < rs->raid_disks; i++)
rs               3664 drivers/md/dm-raid.c 			DMEMIT(" %s %s", __get_dev_name(rs->dev[i].meta_dev),
rs               3665 drivers/md/dm-raid.c 					 __get_dev_name(rs->dev[i].data_dev));
rs               3672 drivers/md/dm-raid.c 	struct raid_set *rs = ti->private;
rs               3673 drivers/md/dm-raid.c 	struct mddev *mddev = &rs->md;
rs               3723 drivers/md/dm-raid.c 	struct raid_set *rs = ti->private;
rs               3727 drivers/md/dm-raid.c 	for (i = 0; !r && i < rs->md.raid_disks; i++)
rs               3728 drivers/md/dm-raid.c 		if (rs->dev[i].data_dev)
rs               3730 drivers/md/dm-raid.c 				 rs->dev[i].data_dev,
rs               3732 drivers/md/dm-raid.c 				 rs->md.dev_sectors,
rs               3740 drivers/md/dm-raid.c 	struct raid_set *rs = ti->private;
rs               3741 drivers/md/dm-raid.c 	unsigned int chunk_size_bytes = to_bytes(rs->md.chunk_sectors);
rs               3744 drivers/md/dm-raid.c 	blk_limits_io_opt(limits, chunk_size_bytes * mddev_data_stripes(rs));
rs               3750 drivers/md/dm-raid.c 	if (rs_is_raid1(rs) || rs_is_raid10(rs)) {
rs               3752 drivers/md/dm-raid.c 		limits->max_discard_sectors = rs->md.chunk_sectors;
rs               3758 drivers/md/dm-raid.c 	struct raid_set *rs = ti->private;
rs               3760 drivers/md/dm-raid.c 	if (!test_and_set_bit(RT_FLAG_RS_SUSPENDED, &rs->runtime_flags)) {
rs               3762 drivers/md/dm-raid.c 		if (!test_bit(MD_RECOVERY_FROZEN, &rs->md.recovery))
rs               3763 drivers/md/dm-raid.c 			md_stop_writes(&rs->md);
rs               3765 drivers/md/dm-raid.c 		mddev_lock_nointr(&rs->md);
rs               3766 drivers/md/dm-raid.c 		mddev_suspend(&rs->md);
rs               3767 drivers/md/dm-raid.c 		mddev_unlock(&rs->md);
rs               3771 drivers/md/dm-raid.c static void attempt_restore_of_faulty_devices(struct raid_set *rs)
rs               3778 drivers/md/dm-raid.c 	struct mddev *mddev = &rs->md;
rs               3788 drivers/md/dm-raid.c 		r = &rs->dev[i].rdev;
rs               3797 drivers/md/dm-raid.c 			       rs->raid_type->name, i);
rs               3839 drivers/md/dm-raid.c 		rdev_for_each(r, &rs->md) {
rs               3854 drivers/md/dm-raid.c static int __load_dirty_region_bitmap(struct raid_set *rs)
rs               3859 drivers/md/dm-raid.c 	if (!rs_is_raid0(rs) &&
rs               3860 drivers/md/dm-raid.c 	    !test_and_set_bit(RT_FLAG_RS_BITMAP_LOADED, &rs->runtime_flags)) {
rs               3861 drivers/md/dm-raid.c 		r = md_bitmap_load(&rs->md);
rs               3870 drivers/md/dm-raid.c static void rs_update_sbs(struct raid_set *rs)
rs               3872 drivers/md/dm-raid.c 	struct mddev *mddev = &rs->md;
rs               3888 drivers/md/dm-raid.c static int rs_start_reshape(struct raid_set *rs)
rs               3891 drivers/md/dm-raid.c 	struct mddev *mddev = &rs->md;
rs               3897 drivers/md/dm-raid.c 	r = rs_setup_reshape(rs);
rs               3908 drivers/md/dm-raid.c 		rs->ti->error = "pers->check_reshape() failed";
rs               3919 drivers/md/dm-raid.c 			rs->ti->error = "pers->start_reshape() failed";
rs               3929 drivers/md/dm-raid.c 	rs_update_sbs(rs);
rs               3937 drivers/md/dm-raid.c 	struct raid_set *rs = ti->private;
rs               3938 drivers/md/dm-raid.c 	struct mddev *mddev = &rs->md;
rs               3941 drivers/md/dm-raid.c 	if (test_and_set_bit(RT_FLAG_RS_PRERESUMED, &rs->runtime_flags))
rs               3950 drivers/md/dm-raid.c 	if (test_bit(RT_FLAG_UPDATE_SBS, &rs->runtime_flags))
rs               3951 drivers/md/dm-raid.c 		rs_update_sbs(rs);
rs               3954 drivers/md/dm-raid.c 	r = __load_dirty_region_bitmap(rs);
rs               3959 drivers/md/dm-raid.c 	if (test_bit(RT_FLAG_RS_BITMAP_LOADED, &rs->runtime_flags) && mddev->bitmap &&
rs               3960 drivers/md/dm-raid.c 	    mddev->bitmap_info.chunksize != to_bytes(rs->requested_bitmap_chunk_sectors)) {
rs               3962 drivers/md/dm-raid.c 				     to_bytes(rs->requested_bitmap_chunk_sectors), 0);
rs               3976 drivers/md/dm-raid.c 	if (test_bit(RT_FLAG_RESHAPE_RS, &rs->runtime_flags)) {
rs               3978 drivers/md/dm-raid.c 		rs_set_rdev_sectors(rs);
rs               3980 drivers/md/dm-raid.c 		r = rs_start_reshape(rs);
rs               3992 drivers/md/dm-raid.c 	struct raid_set *rs = ti->private;
rs               3993 drivers/md/dm-raid.c 	struct mddev *mddev = &rs->md;
rs               3995 drivers/md/dm-raid.c 	if (test_and_set_bit(RT_FLAG_RS_RESUMED, &rs->runtime_flags)) {
rs               4001 drivers/md/dm-raid.c 		attempt_restore_of_faulty_devices(rs);
rs               4004 drivers/md/dm-raid.c 	if (test_and_clear_bit(RT_FLAG_RS_SUSPENDED, &rs->runtime_flags)) {
rs               4007 drivers/md/dm-raid.c 			rs_set_capacity(rs);
rs                 53 drivers/md/dm-verity-fec.c 	return decode_rs8(fio->rs, data, par, v->fec->rsn, NULL, neras,
rs                310 drivers/md/dm-verity-fec.c 	if (!fio->rs)
rs                311 drivers/md/dm-verity-fec.c 		fio->rs = mempool_alloc(&v->fec->rs_pool, GFP_NOIO);
rs                495 drivers/md/dm-verity-fec.c 	mempool_free(fio->rs, &f->rs_pool);
rs                516 drivers/md/dm-verity-fec.c 	fio->rs = NULL;
rs                578 drivers/md/dm-verity-fec.c 	struct rs_control *rs = (struct rs_control *)element;
rs                580 drivers/md/dm-verity-fec.c 	if (rs)
rs                581 drivers/md/dm-verity-fec.c 		free_rs(rs);
rs                 54 drivers/md/dm-verity-fec.h 	struct rs_control *rs;	/* Reed-Solomon state */
rs                 74 drivers/media/usb/gspca/stv06xx/stv06xx_hdcs.c 		u16 rs;
rs                185 drivers/media/usb/gspca/stv06xx/stv06xx_hdcs.c 	rp = hdcs->exp.rs + cp;
rs                383 drivers/media/usb/gspca/stv06xx/stv06xx_hdcs.c 	hdcs->exp.rs = 186;
rs                447 drivers/media/usb/gspca/stv06xx/stv06xx_hdcs.c 	hdcs->exp.rs = 155;
rs                497 drivers/misc/mei/hbm.c 	struct hbm_notification_response *rs =
rs                500 drivers/misc/mei/hbm.c 	return mei_cl_notify_req2fop(rs->start);
rs                515 drivers/misc/mei/hbm.c 	struct hbm_notification_response *rs =
rs                518 drivers/misc/mei/hbm.c 	cl_dbg(dev, cl, "hbm: notify start response status=%d\n", rs->status);
rs                520 drivers/misc/mei/hbm.c 	if (rs->status == MEI_HBMS_SUCCESS ||
rs                521 drivers/misc/mei/hbm.c 	    rs->status == MEI_HBMS_ALREADY_STARTED) {
rs                541 drivers/misc/mei/hbm.c 	struct hbm_notification_response *rs =
rs                544 drivers/misc/mei/hbm.c 	cl_dbg(dev, cl, "hbm: notify stop response status=%d\n", rs->status);
rs                546 drivers/misc/mei/hbm.c 	if (rs->status == MEI_HBMS_SUCCESS ||
rs                547 drivers/misc/mei/hbm.c 	    rs->status == MEI_HBMS_NOT_STARTED) {
rs                792 drivers/misc/mei/hbm.c 	struct hbm_client_connect_response *rs =
rs                795 drivers/misc/mei/hbm.c 	cl_dbg(dev, cl, "hbm: disconnect response status=%d\n", rs->status);
rs                797 drivers/misc/mei/hbm.c 	if (rs->status == MEI_CL_DISCONN_SUCCESS)
rs                829 drivers/misc/mei/hbm.c 	struct hbm_client_connect_response *rs =
rs                833 drivers/misc/mei/hbm.c 			mei_cl_conn_status_str(rs->status));
rs                835 drivers/misc/mei/hbm.c 	if (rs->status == MEI_CL_CONN_SUCCESS)
rs                839 drivers/misc/mei/hbm.c 		if (rs->status == MEI_CL_CONN_NOT_FOUND) {
rs                845 drivers/misc/mei/hbm.c 	cl->status = mei_cl_conn_status_to_errno(rs->status);
rs                857 drivers/misc/mei/hbm.c 			   struct mei_hbm_cl_cmd *rs,
rs                871 drivers/misc/mei/hbm.c 		if (mei_hbm_cl_addr_equal(cl, rs)) {
rs                882 drivers/misc/mei/hbm.c 		mei_hbm_cl_connect_res(dev, cl, rs);
rs                885 drivers/misc/mei/hbm.c 		mei_hbm_cl_disconnect_res(dev, cl, rs);
rs                888 drivers/misc/mei/hbm.c 		mei_hbm_cl_notify_start_res(dev, cl, rs);
rs                891 drivers/misc/mei/hbm.c 		mei_hbm_cl_notify_stop_res(dev, cl, rs);
rs                 89 drivers/misc/phantom.c 	struct phm_regs rs;
rs                129 drivers/misc/phantom.c 		if (copy_from_user(&rs, argp, sizeof(rs)))
rs                132 drivers/misc/phantom.c 		pr_debug("phantom: SRS %u regs %x\n", rs.count, rs.mask);
rs                135 drivers/misc/phantom.c 			memcpy(&dev->oregs, &rs, sizeof(rs));
rs                137 drivers/misc/phantom.c 			u32 m = min(rs.count, 8U);
rs                139 drivers/misc/phantom.c 				if (rs.mask & BIT(i))
rs                140 drivers/misc/phantom.c 					iowrite32(rs.values[i], dev->oaddr + i);
rs                162 drivers/misc/phantom.c 		if (copy_from_user(&rs, argp, sizeof(rs)))
rs                165 drivers/misc/phantom.c 		m = min(rs.count, 8U);
rs                167 drivers/misc/phantom.c 		pr_debug("phantom: GRS %u regs %x\n", rs.count, rs.mask);
rs                170 drivers/misc/phantom.c 			if (rs.mask & BIT(i))
rs                171 drivers/misc/phantom.c 				rs.values[i] = ioread32(dev->iaddr + i);
rs                175 drivers/misc/phantom.c 		if (copy_to_user(argp, &rs, sizeof(rs)))
rs                 64 drivers/mtd/nand/raw/cafe_nand.c 	struct rs_control *rs;
rs                393 drivers/mtd/nand/raw/cafe_nand.c 			syn[i] = cafe->rs->codec->index_of[tmp & 0xfff];
rs                394 drivers/mtd/nand/raw/cafe_nand.c 			syn[i+1] = cafe->rs->codec->index_of[(tmp >> 16) & 0xfff];
rs                397 drivers/mtd/nand/raw/cafe_nand.c 		n = decode_rs16(cafe->rs, NULL, NULL, 1367, syn, 0, pos, 0,
rs                701 drivers/mtd/nand/raw/cafe_nand.c 	cafe->rs = init_rs_non_canonical(12, &cafe_mul, 0, 1, 8);
rs                702 drivers/mtd/nand/raw/cafe_nand.c 	if (!cafe->rs) {
rs                822 drivers/mtd/nand/raw/cafe_nand.c 	free_rs(cafe->rs);
rs                137 drivers/mtd/nand/raw/diskonchip.c static int doc_ecc_decode(struct rs_control *rs, uint8_t *data, uint8_t *ecc)
rs                142 drivers/mtd/nand/raw/diskonchip.c 	struct rs_codec *cd = rs->codec;
rs                174 drivers/mtd/nand/raw/diskonchip.c 	nerr = decode_rs16(rs, NULL, NULL, 1019, syn, 0, errpos, 0, errval);
rs                509 drivers/net/ethernet/cavium/liquidio/liquidio_common.h 	u64 rs:1;
rs                524 drivers/net/ethernet/cavium/liquidio/liquidio_common.h 	u64 rs:1;
rs                662 drivers/net/ethernet/cavium/liquidio/request_manager.c 			ih2->rs = 1;
rs               4345 drivers/net/ethernet/intel/ixgbe/ixgbe_common.c 	u8 rs, eeprom_data;
rs               4350 drivers/net/ethernet/intel/ixgbe/ixgbe_common.c 		rs = IXGBE_SFF_SOFT_RS_SELECT_10G;
rs               4353 drivers/net/ethernet/intel/ixgbe/ixgbe_common.c 		rs = IXGBE_SFF_SOFT_RS_SELECT_1G;
rs               4369 drivers/net/ethernet/intel/ixgbe/ixgbe_common.c 	eeprom_data = (eeprom_data & ~IXGBE_SFF_SOFT_RS_SELECT_MASK) | rs;
rs               4388 drivers/net/ethernet/intel/ixgbe/ixgbe_common.c 	eeprom_data = (eeprom_data & ~IXGBE_SFF_SOFT_RS_SELECT_MASK) | rs;
rs                397 drivers/net/ethernet/sfc/mcdi_port.c 	bool rs = caps & (1 << MC_CMD_PHY_CAP_RS_FEC_LBN),
rs                404 drivers/net/ethernet/sfc/mcdi_port.c 	if (!baser && !rs)
rs                408 drivers/net/ethernet/sfc/mcdi_port.c 	       (baser == baser_req && rs == rs_req ? 0 : ETHTOOL_FEC_AUTO);
rs                280 drivers/net/ethernet/stmicro/stmmac/dwmac-sti.c 		const char *rs;
rs                284 drivers/net/ethernet/stmicro/stmmac/dwmac-sti.c 		err = of_property_read_string(np, "st,tx-retime-src", &rs);
rs                288 drivers/net/ethernet/stmicro/stmmac/dwmac-sti.c 			if (!strcasecmp(rs, "clk_125"))
rs                290 drivers/net/ethernet/stmicro/stmmac/dwmac-sti.c 			else if (!strcasecmp(rs, "txclk"))
rs               1183 drivers/net/wireless/ath/ar5523/ar5523.c 				  struct ar5523_cmd_rateset *rs,
rs               1207 drivers/net/wireless/ath/ar5523/ar5523.c 			rs->set[i] = band->bitrates[bit].hw_value;
rs               1209 drivers/net/wireless/ath/ar5523/ar5523.c 				rs->set[i] |= 0x80;
rs               1216 drivers/net/wireless/ath/ar5523/ar5523.c 	rs->length = i;
rs               3958 drivers/net/wireless/ath/ath10k/wmi.c 	struct radar_detector_specs rs;
rs               4026 drivers/net/wireless/ath/ath10k/wmi.c 	if (!ar->dfs_detector->add_pulse(ar->dfs_detector, &pe, &rs)) {
rs               4046 drivers/net/wireless/ath/ath10k/wmi.c 		radar_info->pri_min = rs.pri_min;
rs               4047 drivers/net/wireless/ath/ath10k/wmi.c 		radar_info->pri_max = rs.pri_max;
rs               4048 drivers/net/wireless/ath/ath10k/wmi.c 		radar_info->width_min = rs.width_min;
rs               4049 drivers/net/wireless/ath/ath10k/wmi.c 		radar_info->width_max = rs.width_max;
rs               1212 drivers/net/wireless/ath/ath5k/base.c 		   struct ath5k_rx_status *rs)
rs               1218 drivers/net/wireless/ath/ath5k/base.c 	if (!(rs->rs_status & AR5K_RXERR_DECRYPT) &&
rs               1219 drivers/net/wireless/ath/ath5k/base.c 			rs->rs_keyix != AR5K_RXKEYIX_INVALID)
rs               1227 drivers/net/wireless/ath/ath5k/base.c 	    !(rs->rs_status & AR5K_RXERR_DECRYPT) &&
rs               1372 drivers/net/wireless/ath/ath5k/base.c 		    struct ath5k_rx_status *rs)
rs               1382 drivers/net/wireless/ath/ath5k/base.c 	if (unlikely(rs->rs_status & AR5K_RXERR_MIC))
rs               1384 drivers/net/wireless/ath/ath5k/base.c 	if (unlikely(rs->rs_status & AR5K_RXERR_CRC))
rs               1397 drivers/net/wireless/ath/ath5k/base.c 	rxs->mactime = ath5k_extend_tsf(ah, rs->rs_tstamp);
rs               1403 drivers/net/wireless/ath/ath5k/base.c 	rxs->signal = ah->ah_noise_floor + rs->rs_rssi;
rs               1405 drivers/net/wireless/ath/ath5k/base.c 	rxs->antenna = rs->rs_antenna;
rs               1407 drivers/net/wireless/ath/ath5k/base.c 	if (rs->rs_antenna > 0 && rs->rs_antenna < 5)
rs               1408 drivers/net/wireless/ath/ath5k/base.c 		ah->stats.antenna_rx[rs->rs_antenna]++;
rs               1412 drivers/net/wireless/ath/ath5k/base.c 	rxs->rate_idx = ath5k_hw_to_driver_rix(ah, rs->rs_rate);
rs               1413 drivers/net/wireless/ath/ath5k/base.c 	rxs->flag |= ath5k_rx_decrypted(ah, skb, rs);
rs               1425 drivers/net/wireless/ath/ath5k/base.c 	if (rs->rs_rate ==
rs               1432 drivers/net/wireless/ath/ath5k/base.c 		ewma_beacon_rssi_add(&ah->ah_beacon_rssi_avg, rs->rs_rssi);
rs               1448 drivers/net/wireless/ath/ath5k/base.c ath5k_receive_frame_ok(struct ath5k_hw *ah, struct ath5k_rx_status *rs)
rs               1451 drivers/net/wireless/ath/ath5k/base.c 	ah->stats.rx_bytes_count += rs->rs_datalen;
rs               1453 drivers/net/wireless/ath/ath5k/base.c 	if (unlikely(rs->rs_status)) {
rs               1456 drivers/net/wireless/ath/ath5k/base.c 		if (rs->rs_status & AR5K_RXERR_CRC)
rs               1458 drivers/net/wireless/ath/ath5k/base.c 		if (rs->rs_status & AR5K_RXERR_FIFO)
rs               1460 drivers/net/wireless/ath/ath5k/base.c 		if (rs->rs_status & AR5K_RXERR_PHY) {
rs               1462 drivers/net/wireless/ath/ath5k/base.c 			if (rs->rs_phyerr > 0 && rs->rs_phyerr < 32)
rs               1463 drivers/net/wireless/ath/ath5k/base.c 				ah->stats.rxerr_phy_code[rs->rs_phyerr]++;
rs               1471 drivers/net/wireless/ath/ath5k/base.c 			if (rs->rs_phyerr == AR5K_RX_PHY_ERROR_OFDM_RESTART ||
rs               1472 drivers/net/wireless/ath/ath5k/base.c 			    rs->rs_phyerr == AR5K_RX_PHY_ERROR_CCK_RESTART) {
rs               1473 drivers/net/wireless/ath/ath5k/base.c 				rs->rs_status |= AR5K_RXERR_CRC;
rs               1474 drivers/net/wireless/ath/ath5k/base.c 				rs->rs_status &= ~AR5K_RXERR_PHY;
rs               1479 drivers/net/wireless/ath/ath5k/base.c 		if (rs->rs_status & AR5K_RXERR_DECRYPT) {
rs               1491 drivers/net/wireless/ath/ath5k/base.c 			if (rs->rs_keyix == AR5K_RXKEYIX_INVALID &&
rs               1492 drivers/net/wireless/ath/ath5k/base.c 			    !(rs->rs_status & AR5K_RXERR_CRC))
rs               1495 drivers/net/wireless/ath/ath5k/base.c 		if (rs->rs_status & AR5K_RXERR_MIC) {
rs               1508 drivers/net/wireless/ath/ath5k/base.c 		if (rs->rs_status & ~filters)
rs               1512 drivers/net/wireless/ath/ath5k/base.c 	if (unlikely(rs->rs_more)) {
rs               1541 drivers/net/wireless/ath/ath5k/base.c 	struct ath5k_rx_status rs = {};
rs               1565 drivers/net/wireless/ath/ath5k/base.c 		ret = ah->ah_proc_rx_desc(ah, ds, &rs);
rs               1574 drivers/net/wireless/ath/ath5k/base.c 		if (ath5k_receive_frame_ok(ah, &rs)) {
rs               1588 drivers/net/wireless/ath/ath5k/base.c 			skb_put(skb, rs.rs_datalen);
rs               1590 drivers/net/wireless/ath/ath5k/base.c 			ath5k_receive_frame(ah, skb, &rs);
rs               1068 drivers/net/wireless/ath/ath5k/debug.c 		       struct ath5k_rx_status *rs)
rs               1078 drivers/net/wireless/ath/ath5k/debug.c 		!done ? ' ' : (rs->rs_status == 0) ? '*' : '!');
rs               1086 drivers/net/wireless/ath/ath5k/debug.c 	struct ath5k_rx_status rs = {};
rs               1098 drivers/net/wireless/ath/ath5k/debug.c 		status = ah->ah_proc_rx_desc(ah, ds, &rs);
rs               1100 drivers/net/wireless/ath/ath5k/debug.c 			ath5k_debug_printrxbuf(bf, status == 0, &rs);
rs                603 drivers/net/wireless/ath/ath5k/desc.c 				struct ath5k_rx_status *rs)
rs                614 drivers/net/wireless/ath/ath5k/desc.c 	memset(rs, 0, sizeof(struct ath5k_rx_status));
rs                619 drivers/net/wireless/ath/ath5k/desc.c 	rs->rs_datalen = rx_status->rx_status_0 &
rs                621 drivers/net/wireless/ath/ath5k/desc.c 	rs->rs_rssi = AR5K_REG_MS(rx_status->rx_status_0,
rs                623 drivers/net/wireless/ath/ath5k/desc.c 	rs->rs_rate = AR5K_REG_MS(rx_status->rx_status_0,
rs                625 drivers/net/wireless/ath/ath5k/desc.c 	rs->rs_more = !!(rx_status->rx_status_0 &
rs                632 drivers/net/wireless/ath/ath5k/desc.c 	rs->rs_tstamp = AR5K_REG_MS(rx_status->rx_status_1,
rs                636 drivers/net/wireless/ath/ath5k/desc.c 		rs->rs_antenna = AR5K_REG_MS(rx_status->rx_status_0,
rs                639 drivers/net/wireless/ath/ath5k/desc.c 		rs->rs_antenna = (rx_status->rx_status_0 &
rs                647 drivers/net/wireless/ath/ath5k/desc.c 		rs->rs_keyix = AR5K_REG_MS(rx_status->rx_status_1,
rs                650 drivers/net/wireless/ath/ath5k/desc.c 		rs->rs_keyix = AR5K_RXKEYIX_INVALID;
rs                659 drivers/net/wireless/ath/ath5k/desc.c 			rs->rs_status |= AR5K_RXERR_CRC;
rs                665 drivers/net/wireless/ath/ath5k/desc.c 			rs->rs_status |= AR5K_RXERR_FIFO;
rs                669 drivers/net/wireless/ath/ath5k/desc.c 			rs->rs_status |= AR5K_RXERR_PHY;
rs                670 drivers/net/wireless/ath/ath5k/desc.c 			rs->rs_phyerr = AR5K_REG_MS(rx_status->rx_status_1,
rs                676 drivers/net/wireless/ath/ath5k/desc.c 			rs->rs_status |= AR5K_RXERR_DECRYPT;
rs                697 drivers/net/wireless/ath/ath5k/desc.c 				struct ath5k_rx_status *rs)
rs                709 drivers/net/wireless/ath/ath5k/desc.c 	memset(rs, 0, sizeof(struct ath5k_rx_status));
rs                715 drivers/net/wireless/ath/ath5k/desc.c 	rs->rs_datalen = rxstat0 & AR5K_5212_RX_DESC_STATUS0_DATA_LEN;
rs                716 drivers/net/wireless/ath/ath5k/desc.c 	rs->rs_rssi = AR5K_REG_MS(rxstat0,
rs                718 drivers/net/wireless/ath/ath5k/desc.c 	rs->rs_rate = AR5K_REG_MS(rxstat0,
rs                720 drivers/net/wireless/ath/ath5k/desc.c 	rs->rs_antenna = AR5K_REG_MS(rxstat0,
rs                722 drivers/net/wireless/ath/ath5k/desc.c 	rs->rs_more = !!(rxstat0 & AR5K_5212_RX_DESC_STATUS0_MORE);
rs                723 drivers/net/wireless/ath/ath5k/desc.c 	rs->rs_tstamp = AR5K_REG_MS(rxstat1,
rs                730 drivers/net/wireless/ath/ath5k/desc.c 		rs->rs_keyix = AR5K_REG_MS(rxstat1,
rs                733 drivers/net/wireless/ath/ath5k/desc.c 		rs->rs_keyix = AR5K_RXKEYIX_INVALID;
rs                740 drivers/net/wireless/ath/ath5k/desc.c 			rs->rs_status |= AR5K_RXERR_CRC;
rs                743 drivers/net/wireless/ath/ath5k/desc.c 			rs->rs_status |= AR5K_RXERR_PHY;
rs                744 drivers/net/wireless/ath/ath5k/desc.c 			rs->rs_phyerr = AR5K_REG_MS(rxstat1,
rs                747 drivers/net/wireless/ath/ath5k/desc.c 				ath5k_ani_phy_error_report(ah, rs->rs_phyerr);
rs                751 drivers/net/wireless/ath/ath5k/desc.c 			rs->rs_status |= AR5K_RXERR_DECRYPT;
rs                754 drivers/net/wireless/ath/ath5k/desc.c 			rs->rs_status |= AR5K_RXERR_MIC;
rs                721 drivers/net/wireless/ath/ath9k/antenna.c void ath_ant_comb_scan(struct ath_softc *sc, struct ath_rx_status *rs)
rs                727 drivers/net/wireless/ath/ath9k/antenna.c 	int main_rssi = rs->rs_rssi_ctl[0];
rs                728 drivers/net/wireless/ath/ath9k/antenna.c 	int alt_rssi = rs->rs_rssi_ctl[1];
rs                732 drivers/net/wireless/ath/ath9k/antenna.c 	rx_ant_conf = (rs->rs_rssi_ctl[2] >> ATH_ANT_RX_CURRENT_SHIFT) &
rs                734 drivers/net/wireless/ath/ath9k/antenna.c 	main_ant_conf = (rs->rs_rssi_ctl[2] >> ATH_ANT_RX_MAIN_SHIFT) &
rs                769 drivers/net/wireless/ath/ath9k/antenna.c 	     rs->rs_moreaggr) && !short_scan)
rs                944 drivers/net/wireless/ath/ath9k/ath9k.h void ath_ant_comb_scan(struct ath_softc *sc, struct ath_rx_status *rs);
rs                 91 drivers/net/wireless/ath/ath9k/common-debug.c 			     struct ath_rx_status *rs)
rs                 97 drivers/net/wireless/ath/ath9k/common-debug.c 	rxstats->rx_bytes_all += rs->rs_datalen;
rs                 99 drivers/net/wireless/ath/ath9k/common-debug.c 	if (rs->rs_status & ATH9K_RXERR_CRC)
rs                101 drivers/net/wireless/ath/ath9k/common-debug.c 	if (rs->rs_status & ATH9K_RXERR_DECRYPT)
rs                103 drivers/net/wireless/ath/ath9k/common-debug.c 	if (rs->rs_status & ATH9K_RXERR_MIC)
rs                105 drivers/net/wireless/ath/ath9k/common-debug.c 	if (rs->rs_status & ATH9K_RX_DELIM_CRC_PRE)
rs                107 drivers/net/wireless/ath/ath9k/common-debug.c 	if (rs->rs_status & ATH9K_RX_DELIM_CRC_POST)
rs                109 drivers/net/wireless/ath/ath9k/common-debug.c 	if (rs->rs_status & ATH9K_RX_DECRYPT_BUSY)
rs                112 drivers/net/wireless/ath/ath9k/common-debug.c 	if (rs->rs_status & ATH9K_RXERR_PHY) {
rs                114 drivers/net/wireless/ath/ath9k/common-debug.c 		if (rs->rs_phyerr < ATH9K_PHYERR_MAX)
rs                115 drivers/net/wireless/ath/ath9k/common-debug.c 			RX_PHY_ERR_INC(rs->rs_phyerr);
rs                 73 drivers/net/wireless/ath/ath9k/common-debug.h 			     struct ath_rx_status *rs);
rs                 90 drivers/net/wireless/ath/ath9k/common-debug.h 					   struct ath_rx_status *rs)
rs                126 drivers/net/wireless/ath/ath9k/common-spectral.c typedef int (ath_cmn_fft_sample_handler) (struct ath_rx_status *rs,
rs                131 drivers/net/wireless/ath/ath9k/common-spectral.c ath_cmn_process_ht20_fft(struct ath_rx_status *rs,
rs                151 drivers/net/wireless/ath/ath9k/common-spectral.c 	fft_sample_20.rssi = fix_rssi_inv_only(rs->rs_rssi_ctl[0]);
rs                235 drivers/net/wireless/ath/ath9k/common-spectral.c ath_cmn_process_ht20_40_fft(struct ath_rx_status *rs,
rs                268 drivers/net/wireless/ath/ath9k/common-spectral.c 		lower_rssi = fix_rssi_inv_only(rs->rs_rssi_ctl[0]);
rs                269 drivers/net/wireless/ath/ath9k/common-spectral.c 		upper_rssi = fix_rssi_inv_only(rs->rs_rssi_ext[0]);
rs                274 drivers/net/wireless/ath/ath9k/common-spectral.c 		lower_rssi = fix_rssi_inv_only(rs->rs_rssi_ext[0]);
rs                275 drivers/net/wireless/ath/ath9k/common-spectral.c 		upper_rssi = fix_rssi_inv_only(rs->rs_rssi_ctl[0]);
rs                469 drivers/net/wireless/ath/ath9k/common-spectral.c 		    struct ath_rx_status *rs, u64 tsf)
rs                477 drivers/net/wireless/ath/ath9k/common-spectral.c 	int len = rs->rs_datalen;
rs                492 drivers/net/wireless/ath/ath9k/common-spectral.c 	if (rs->rs_phyerr != ATH9K_PHYERR_RADAR &&
rs                493 drivers/net/wireless/ath/ath9k/common-spectral.c 	    rs->rs_phyerr != ATH9K_PHYERR_FALSE_RADAR_EXT &&
rs                494 drivers/net/wireless/ath/ath9k/common-spectral.c 	    rs->rs_phyerr != ATH9K_PHYERR_SPECTRAL)
rs                623 drivers/net/wireless/ath/ath9k/common-spectral.c 				ret = fft_handler(rs, spec_priv, sample_buf,
rs                641 drivers/net/wireless/ath/ath9k/common-spectral.c 				ret = fft_handler(rs, spec_priv, sample_start,
rs                181 drivers/net/wireless/ath/ath9k/common-spectral.h 		    struct ath_rx_status *rs, u64 tsf);
rs                199 drivers/net/wireless/ath/ath9k/common-spectral.h 				      struct ath_rx_status *rs, u64 tsf)
rs                819 drivers/net/wireless/ath/ath9k/debug.c void ath_debug_stat_rx(struct ath_softc *sc, struct ath_rx_status *rs)
rs                821 drivers/net/wireless/ath/ath9k/debug.c 	ath9k_cmn_debug_stat_rx(&sc->debug.stats.rxstats, rs);
rs                261 drivers/net/wireless/ath/ath9k/debug.h void ath_debug_stat_rx(struct ath_softc *sc, struct ath_rx_status *rs);
rs                301 drivers/net/wireless/ath/ath9k/debug.h 				     struct ath_rx_status *rs)
rs                320 drivers/net/wireless/ath/ath9k/debug.h 			  struct ath_rx_status *rs,
rs                324 drivers/net/wireless/ath/ath9k/debug.h 					struct ath_rx_status *rs,
rs                 92 drivers/net/wireless/ath/ath9k/debug_sta.c 			  struct ath_rx_status *rs,
rs                115 drivers/net/wireless/ath/ath9k/debug_sta.c 	if (IS_HT_RATE(rs->rs_rate)) {
rs                132 drivers/net/wireless/ath/ath9k/debug_sta.c 	if (IS_CCK_RATE(rs->rs_rate)) {
rs                141 drivers/net/wireless/ath/ath9k/debug_sta.c 	if (IS_OFDM_RATE(rs->rs_rate)) {
rs                290 drivers/net/wireless/ath/ath9k/dfs.c 			      struct ath_rx_status *rs, u64 mactime)
rs                300 drivers/net/wireless/ath/ath9k/dfs.c 	if ((rs->rs_phyerr != ATH9K_PHYERR_RADAR) &&
rs                301 drivers/net/wireless/ath/ath9k/dfs.c 	    (rs->rs_phyerr != ATH9K_PHYERR_FALSE_RADAR_EXT)) {
rs                304 drivers/net/wireless/ath/ath9k/dfs.c 			rs->rs_phyerr);
rs                309 drivers/net/wireless/ath/ath9k/dfs.c 	datalen = rs->rs_datalen;
rs                315 drivers/net/wireless/ath/ath9k/dfs.c 	ard.rssi = rs->rs_rssi_ctl[0];
rs                316 drivers/net/wireless/ath/ath9k/dfs.c 	ard.ext_rssi = rs->rs_rssi_ext[0];
rs                 38 drivers/net/wireless/ath/ath9k/dfs.h 			      struct ath_rx_status *rs, u64 mactime);
rs                 42 drivers/net/wireless/ath/ath9k/dfs.h 			 struct ath_rx_status *rs, u64 mactime) { }
rs                338 drivers/net/wireless/ath/ath9k/htc.h 			   struct ath_rx_status *rs);
rs                384 drivers/net/wireless/ath/ath9k/htc.h 					 struct ath_rx_status *rs)
rs                246 drivers/net/wireless/ath/ath9k/htc_drv_debug.c 			     struct ath_rx_status *rs)
rs                248 drivers/net/wireless/ath/ath9k/htc_drv_debug.c 	ath9k_cmn_debug_stat_rx(&priv->debug.rx_stats, rs);
rs                525 drivers/net/wireless/ath/ath9k/mac.c 			struct ath_rx_status *rs)
rs                536 drivers/net/wireless/ath/ath9k/mac.c 	rs->rs_status = 0;
rs                537 drivers/net/wireless/ath/ath9k/mac.c 	rs->rs_flags = 0;
rs                538 drivers/net/wireless/ath/ath9k/mac.c 	rs->enc_flags = 0;
rs                539 drivers/net/wireless/ath/ath9k/mac.c 	rs->bw = RATE_INFO_BW_20;
rs                541 drivers/net/wireless/ath/ath9k/mac.c 	rs->rs_datalen = ads.ds_rxstatus1 & AR_DataLen;
rs                542 drivers/net/wireless/ath/ath9k/mac.c 	rs->rs_tstamp = ads.AR_RcvTimestamp;
rs                545 drivers/net/wireless/ath/ath9k/mac.c 		rs->rs_rssi = ATH9K_RSSI_BAD;
rs                546 drivers/net/wireless/ath/ath9k/mac.c 		rs->rs_rssi_ctl[0] = ATH9K_RSSI_BAD;
rs                547 drivers/net/wireless/ath/ath9k/mac.c 		rs->rs_rssi_ctl[1] = ATH9K_RSSI_BAD;
rs                548 drivers/net/wireless/ath/ath9k/mac.c 		rs->rs_rssi_ctl[2] = ATH9K_RSSI_BAD;
rs                549 drivers/net/wireless/ath/ath9k/mac.c 		rs->rs_rssi_ext[0] = ATH9K_RSSI_BAD;
rs                550 drivers/net/wireless/ath/ath9k/mac.c 		rs->rs_rssi_ext[1] = ATH9K_RSSI_BAD;
rs                551 drivers/net/wireless/ath/ath9k/mac.c 		rs->rs_rssi_ext[2] = ATH9K_RSSI_BAD;
rs                553 drivers/net/wireless/ath/ath9k/mac.c 		rs->rs_rssi = MS(ads.ds_rxstatus4, AR_RxRSSICombined);
rs                554 drivers/net/wireless/ath/ath9k/mac.c 		rs->rs_rssi_ctl[0] = MS(ads.ds_rxstatus0,
rs                556 drivers/net/wireless/ath/ath9k/mac.c 		rs->rs_rssi_ctl[1] = MS(ads.ds_rxstatus0,
rs                558 drivers/net/wireless/ath/ath9k/mac.c 		rs->rs_rssi_ctl[2] = MS(ads.ds_rxstatus0,
rs                560 drivers/net/wireless/ath/ath9k/mac.c 		rs->rs_rssi_ext[0] = MS(ads.ds_rxstatus4,
rs                562 drivers/net/wireless/ath/ath9k/mac.c 		rs->rs_rssi_ext[1] = MS(ads.ds_rxstatus4,
rs                564 drivers/net/wireless/ath/ath9k/mac.c 		rs->rs_rssi_ext[2] = MS(ads.ds_rxstatus4,
rs                568 drivers/net/wireless/ath/ath9k/mac.c 		rs->rs_keyix = MS(ads.ds_rxstatus8, AR_KeyIdx);
rs                570 drivers/net/wireless/ath/ath9k/mac.c 		rs->rs_keyix = ATH9K_RXKEYIX_INVALID;
rs                572 drivers/net/wireless/ath/ath9k/mac.c 	rs->rs_rate = MS(ads.ds_rxstatus0, AR_RxRate);
rs                573 drivers/net/wireless/ath/ath9k/mac.c 	rs->rs_more = (ads.ds_rxstatus1 & AR_RxMore) ? 1 : 0;
rs                575 drivers/net/wireless/ath/ath9k/mac.c 	rs->rs_firstaggr = (ads.ds_rxstatus8 & AR_RxFirstAggr) ? 1 : 0;
rs                576 drivers/net/wireless/ath/ath9k/mac.c 	rs->rs_isaggr = (ads.ds_rxstatus8 & AR_RxAggr) ? 1 : 0;
rs                577 drivers/net/wireless/ath/ath9k/mac.c 	rs->rs_moreaggr = (ads.ds_rxstatus8 & AR_RxMoreAggr) ? 1 : 0;
rs                578 drivers/net/wireless/ath/ath9k/mac.c 	rs->rs_antenna = MS(ads.ds_rxstatus3, AR_RxAntenna);
rs                581 drivers/net/wireless/ath/ath9k/mac.c 	rs->enc_flags |=
rs                583 drivers/net/wireless/ath/ath9k/mac.c 	rs->bw = (ads.ds_rxstatus3 & AR_2040) ? RATE_INFO_BW_40 :
rs                586 drivers/net/wireless/ath/ath9k/mac.c 		rs->enc_flags |=
rs                592 drivers/net/wireless/ath/ath9k/mac.c 		rs->rs_flags |= ATH9K_RX_DELIM_CRC_PRE;
rs                594 drivers/net/wireless/ath/ath9k/mac.c 		rs->rs_flags |= ATH9K_RX_DELIM_CRC_POST;
rs                596 drivers/net/wireless/ath/ath9k/mac.c 		rs->rs_flags |= ATH9K_RX_DECRYPT_BUSY;
rs                606 drivers/net/wireless/ath/ath9k/mac.c 			rs->rs_status |= ATH9K_RXERR_PHY;
rs                608 drivers/net/wireless/ath/ath9k/mac.c 			rs->rs_phyerr = phyerr;
rs                610 drivers/net/wireless/ath/ath9k/mac.c 			rs->rs_status |= ATH9K_RXERR_CRC;
rs                612 drivers/net/wireless/ath/ath9k/mac.c 			rs->rs_status |= ATH9K_RXERR_DECRYPT;
rs                614 drivers/net/wireless/ath/ath9k/mac.c 			rs->rs_status |= ATH9K_RXERR_MIC;
rs                618 drivers/net/wireless/ath/ath9k/mac.c 			rs->rs_status |= ATH9K_RXERR_CORRUPT_DESC;
rs                621 drivers/net/wireless/ath/ath9k/mac.c 		if (rs->rs_rate >= 0x90)
rs                622 drivers/net/wireless/ath/ath9k/mac.c 			rs->rs_status |= ATH9K_RXERR_CORRUPT_DESC;
rs                626 drivers/net/wireless/ath/ath9k/mac.c 		rs->rs_status |= ATH9K_RXERR_KEYMISS;
rs                732 drivers/net/wireless/ath/ath9k/mac.h 			struct ath_rx_status *rs);
rs                628 drivers/net/wireless/ath/ath9k/recv.c 				 struct ath_rx_status *rs,
rs                648 drivers/net/wireless/ath/ath9k/recv.c 	ret = ath9k_hw_process_rxdesc_edma(ah, rs, skb->data);
rs                680 drivers/net/wireless/ath/ath9k/recv.c 						struct ath_rx_status *rs,
rs                685 drivers/net/wireless/ath/ath9k/recv.c 	while (ath_edma_get_buffers(sc, qtype, rs, &bf)) {
rs                695 drivers/net/wireless/ath/ath9k/recv.c 					   struct ath_rx_status *rs)
rs                725 drivers/net/wireless/ath/ath9k/recv.c 	ret = ath9k_hw_rxprocdesc(ah, ds, rs);
rs                759 drivers/net/wireless/ath/ath9k/recv.c 		ret = ath9k_hw_rxprocdesc(ah, ds, rs);
rs                765 drivers/net/wireless/ath/ath9k/recv.c 			rs->rs_datalen = 0;
rs                766 drivers/net/wireless/ath/ath9k/recv.c 			rs->rs_more = true;
rs                786 drivers/net/wireless/ath/ath9k/recv.c static void ath9k_process_tsf(struct ath_rx_status *rs,
rs                792 drivers/net/wireless/ath/ath9k/recv.c 	rxs->mactime = (tsf & ~0xffffffffULL) | rs->rs_tstamp;
rs                793 drivers/net/wireless/ath/ath9k/recv.c 	if (rs->rs_tstamp > tsf_lower &&
rs                794 drivers/net/wireless/ath/ath9k/recv.c 	    unlikely(rs->rs_tstamp - tsf_lower > 0x10000000))
rs                797 drivers/net/wireless/ath/ath9k/recv.c 	if (rs->rs_tstamp < tsf_lower &&
rs                798 drivers/net/wireless/ath/ath9k/recv.c 	    unlikely(tsf_lower - rs->rs_tstamp > 0x10000000))
rs                963 drivers/net/wireless/ath/ath9k/recv.c 				struct ath_rx_status *rs)
rs                976 drivers/net/wireless/ath/ath9k/recv.c 	if (sc->rx.defant != rs->rs_antenna) {
rs                978 drivers/net/wireless/ath/ath9k/recv.c 			ath_setdefantenna(sc, rs->rs_antenna);
rs                985 drivers/net/wireless/ath/ath9k/recv.c 			ath_ant_comb_scan(sc, rs);
rs                987 drivers/net/wireless/ath/ath9k/recv.c 		ath_ant_comb_scan(sc, rs);
rs                992 drivers/net/wireless/ath/ath9k/recv.c 	struct ath_rx_status *rs, struct ieee80211_rx_status *rxs)
rs                994 drivers/net/wireless/ath/ath9k/recv.c 	if (rs->rs_isaggr) {
rs                999 drivers/net/wireless/ath/ath9k/recv.c 		if (!rs->rs_moreaggr) {
rs               1004 drivers/net/wireless/ath/ath9k/recv.c 		if (rs->rs_flags & ATH9K_RX_DELIM_CRC_PRE)
rs               1010 drivers/net/wireless/ath/ath9k/recv.c 				 struct ath_rx_status *rs,
rs               1021 drivers/net/wireless/ath/ath9k/recv.c 	u16 len = rs->rs_datalen;
rs               1048 drivers/net/wireless/ath/ath9k/recv.c 		phy = IS_CCK_RATE(rs->rs_rate) ? WLAN_RC_PHY_CCK : WLAN_RC_PHY_OFDM;
rs               1068 drivers/net/wireless/ath/ath9k/recv.c 	struct ath_rx_status rs;
rs               1090 drivers/net/wireless/ath/ath9k/recv.c 		memset(&rs, 0, sizeof(rs));
rs               1092 drivers/net/wireless/ath/ath9k/recv.c 			bf = ath_edma_get_next_rx_buf(sc, &rs, qtype);
rs               1094 drivers/net/wireless/ath/ath9k/recv.c 			bf = ath_get_next_rx_buf(sc, &rs);
rs               1115 drivers/net/wireless/ath/ath9k/recv.c 		retval = ath9k_rx_skb_preprocess(sc, hdr_skb, &rs, rxs,
rs               1148 drivers/net/wireless/ath/ath9k/recv.c 		skb_put(skb, rs.rs_datalen + ah->caps.rx_status_len);
rs               1152 drivers/net/wireless/ath/ath9k/recv.c 		if (!rs.rs_more)
rs               1153 drivers/net/wireless/ath/ath9k/recv.c 			ath9k_cmn_rx_skb_postprocess(common, hdr_skb, &rs,
rs               1156 drivers/net/wireless/ath/ath9k/recv.c 		if (rs.rs_more) {
rs               1199 drivers/net/wireless/ath/ath9k/recv.c 			ath_rx_ps(sc, skb, rs.is_mybeacon);
rs               1202 drivers/net/wireless/ath/ath9k/recv.c 		ath9k_antenna_check(sc, &rs);
rs               1203 drivers/net/wireless/ath/ath9k/recv.c 		ath9k_apply_ampdu_details(sc, &rs, rxs);
rs               1204 drivers/net/wireless/ath/ath9k/recv.c 		ath_debug_rate_stats(sc, &rs, skb);
rs               1205 drivers/net/wireless/ath/ath9k/recv.c 		ath_rx_count_airtime(sc, &rs, skb);
rs               1209 drivers/net/wireless/ath/ath9k/recv.c 			ath_dynack_sample_ack_ts(sc->sc_ah, skb, rs.rs_tstamp);
rs                212 drivers/net/wireless/ath/dfs_pattern_detector.c 		const struct radar_detector_specs *rs = &dpd->radar_spec[i];
rs                213 drivers/net/wireless/ath/dfs_pattern_detector.c 		struct pri_detector *de = pri_detector_init(rs);
rs                272 drivers/net/wireless/ath/dfs_pattern_detector.c 	      struct radar_detector_specs *rs)
rs                298 drivers/net/wireless/ath/dfs_pattern_detector.c 			if (rs != NULL)
rs                299 drivers/net/wireless/ath/dfs_pattern_detector.c 				memcpy(rs, pd->rs, sizeof(*rs));
rs                303 drivers/net/wireless/ath/dfs_pattern_detector.c 				event->freq, pd->rs->type_id,
rs                101 drivers/net/wireless/ath/dfs_pattern_detector.h 			  struct radar_detector_specs *rs);
rs                236 drivers/net/wireless/ath/dfs_pri_detector.c 		if (delta_ts < pde->rs->pri_min)
rs                240 drivers/net/wireless/ath/dfs_pri_detector.c 		if (delta_ts > pde->rs->pri_max)
rs                249 drivers/net/wireless/ath/dfs_pri_detector.c 		ps.pri = GET_PRI_TO_USE(pde->rs->pri_min,
rs                250 drivers/net/wireless/ath/dfs_pri_detector.c 			pde->rs->pri_max, ts - p->ts);
rs                251 drivers/net/wireless/ath/dfs_pri_detector.c 		ps.dur = ps.pri * (pde->rs->ppb - 1)
rs                252 drivers/net/wireless/ath/dfs_pri_detector.c 				+ 2 * pde->rs->max_pri_tolerance;
rs                265 drivers/net/wireless/ath/dfs_pri_detector.c 						  pde->rs->max_pri_tolerance);
rs                322 drivers/net/wireless/ath/dfs_pri_detector.c 					  pde->rs->max_pri_tolerance);
rs                350 drivers/net/wireless/ath/dfs_pri_detector.c 		if ((ps->count >= pde->rs->ppb_thresh) &&
rs                351 drivers/net/wireless/ath/dfs_pri_detector.c 		    (ps->count * pde->rs->num_pri >= ps->count_falses))
rs                388 drivers/net/wireless/ath/dfs_pri_detector.c 	const struct radar_detector_specs *rs = de->rs;
rs                391 drivers/net/wireless/ath/dfs_pri_detector.c 	if ((rs->width_min > event->width) || (rs->width_max < event->width))
rs                394 drivers/net/wireless/ath/dfs_pri_detector.c 	if ((ts - de->last_ts) < rs->max_pri_tolerance)
rs                398 drivers/net/wireless/ath/dfs_pri_detector.c 	if (rs->chirp && rs->chirp != event->chirp)
rs                418 drivers/net/wireless/ath/dfs_pri_detector.c struct pri_detector *pri_detector_init(const struct radar_detector_specs *rs)
rs                431 drivers/net/wireless/ath/dfs_pri_detector.c 	de->window_size = rs->pri_max * rs->ppb * rs->num_pri;
rs                432 drivers/net/wireless/ath/dfs_pri_detector.c 	de->max_count = rs->ppb * 2;
rs                433 drivers/net/wireless/ath/dfs_pri_detector.c 	de->rs = rs;
rs                 65 drivers/net/wireless/ath/dfs_pri_detector.h 	const struct radar_detector_specs *rs;
rs                 76 drivers/net/wireless/ath/dfs_pri_detector.h struct pri_detector *pri_detector_init(const struct radar_detector_specs *rs);
rs                262 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c static void brcms_set_basic_rate(struct brcm_rateset *rs, u16 rate, bool is_br)
rs                266 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 	for (i = 0; i < rs->count; i++) {
rs                267 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 		if (rate != (rs->rates[i] & 0x7f))
rs                271 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 			rs->rates[i] |= BRCMS_RATE_FLAG;
rs                273 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 			rs->rates[i] &= BRCMS_RATE_MASK;
rs                632 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 		struct brcm_rateset rs;
rs                637 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 		brcms_c_get_current_rateset(wl->wlc, &rs);
rs                647 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 			brcms_set_basic_rate(&rs, rate, br_mask & 1);
rs                653 drivers/net/wireless/broadcom/brcm80211/brcmsmac/mac80211_if.c 		error = brcms_c_set_rateset(wl->wlc, &rs);
rs               3416 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c brcms_default_rateset(struct brcms_c_info *wlc, struct brcms_c_rateset *rs)
rs               3418 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 	brcms_c_rateset_default(rs, NULL, wlc->band->phytype,
rs               3677 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 	struct brcms_c_rateset rs;
rs               3683 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 	brcms_c_rateset_copy(rs_dflt, &rs);
rs               3684 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 	brcms_c_rateset_mcs_upd(&rs, wlc->stf->txstreams);
rs               3687 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 	for (i = 0; i < rs.count; i++) {
rs               3688 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 		rate = rs.rates[i] & BRCMS_RATE_MASK;
rs               3698 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 			basic_rate = rs.rates[0] & BRCMS_RATE_MASK;
rs               5247 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 	struct brcms_c_rateset rs;
rs               5274 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 	memset(&rs, 0, sizeof(rs));
rs               5279 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 		brcms_c_rateset_copy(&gphy_legacy_rates, &rs);
rs               5311 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 	if (!rs.count)
rs               5312 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 		brcms_c_rateset_copy(&cck_ofdm_rates, &rs);
rs               5315 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 		for (i = 0; i < rs.count; i++) {
rs               5316 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 			if (rs.rates[i] == BRCM_RATE_6M
rs               5317 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 			    || rs.rates[i] == BRCM_RATE_12M
rs               5318 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 			    || rs.rates[i] == BRCM_RATE_24M)
rs               5319 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 				rs.rates[i] |= BRCMS_RATE_FLAG;
rs               5324 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 	wlc->default_bss->rateset.count = rs.count;
rs               5325 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 	memcpy(wlc->default_bss->rateset.rates, rs.rates,
rs               5362 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 	struct brcms_c_rateset rs, new;
rs               5365 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 	memcpy(&rs, rs_arg, sizeof(struct brcms_c_rateset));
rs               5368 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 	if ((rs.count == 0) || (rs.count > BRCMS_NUMRATES))
rs               5373 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 	memcpy(&new, &rs, sizeof(struct brcms_c_rateset));
rs               5382 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 		memcpy(&new, &rs, sizeof(struct brcms_c_rateset));
rs               5472 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 	struct brcms_c_rateset *rs;
rs               5475 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 		rs = &wlc->bsscfg->current_bss->rateset;
rs               5477 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 		rs = &wlc->default_bss->rateset;
rs               5480 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 	currs->count = rs->count;
rs               5481 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 	memcpy(&currs->rates, &rs->rates, rs->count);
rs               5484 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c int brcms_c_set_rateset(struct brcms_c_info *wlc, struct brcm_rateset *rs)
rs               5489 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 	if (rs->count > BRCMS_NUMRATES)
rs               5495 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 	internal_rs.count = rs->count;
rs               5496 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 	memcpy(&internal_rs.rates, &rs->rates, internal_rs.count);
rs               7243 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 	struct brcms_c_rateset rs;
rs               7254 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 	brcms_c_rateset_copy(rs_dflt, &rs);
rs               7255 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 	brcms_c_rateset_mcs_upd(&rs, wlc->stf->txstreams);
rs               7261 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 	for (i = 0; i < rs.count; i++) {
rs               7262 drivers/net/wireless/broadcom/brcm80211/brcmsmac/main.c 		rate = rs.rates[i] & BRCMS_RATE_MASK;
rs                316 drivers/net/wireless/broadcom/brcm80211/brcmsmac/pub.h int brcms_c_set_rateset(struct brcms_c_info *wlc, struct brcm_rateset *rs);
rs                272 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c static bool brcms_c_rateset_valid(struct brcms_c_rateset *rs, bool check_brate)
rs                276 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c 	if (!rs->count)
rs                283 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c 	for (idx = 0; idx < rs->count; idx++) {
rs                284 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c 		if (rs->rates[idx] & BRCMS_RATE_FLAG)
rs                290 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c void brcms_c_rateset_mcs_upd(struct brcms_c_rateset *rs, u8 txstreams)
rs                294 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c 		rs->mcs[i] = 0;
rs                302 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c brcms_c_rate_hwrs_filter_sort_validate(struct brcms_c_rateset *rs,
rs                312 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c 	count = rs->count;
rs                316 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c 		r = (int)rs->rates[i] & BRCMS_RATE_MASK;
rs                319 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c 		rateset[r] = rs->rates[i];	/* preserve basic bit! */
rs                327 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c 			rs->rates[count++] = rateset[r];
rs                330 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c 	rs->count = count;
rs                334 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c 		rs->mcs[i] = (rs->mcs[i] & hw_rs->mcs[i]);
rs                336 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.c 	if (brcms_c_rateset_valid(rs, check_brate))
rs                219 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.h bool brcms_c_rate_hwrs_filter_sort_validate(struct brcms_c_rateset *rs,
rs                240 drivers/net/wireless/broadcom/brcm80211/brcmsmac/rate.h void brcms_c_rateset_mcs_upd(struct brcms_c_rateset *rs, u8 txstreams);
rs                951 drivers/net/wireless/mediatek/mt76/mt7603/mac.c 	struct mt7603_rate_set *rs;
rs               1007 drivers/net/wireless/mediatek/mt76/mt7603/mac.c 	rs = &sta->rateset[rs_idx];
rs               1009 drivers/net/wireless/mediatek/mt76/mt7603/mac.c 	if (!first_idx && rs->probe_rate.idx >= 0) {
rs               1010 drivers/net/wireless/mediatek/mt76/mt7603/mac.c 		info->status.rates[0] = rs->probe_rate;
rs               1020 drivers/net/wireless/mediatek/mt76/mt7603/mac.c 		info->status.rates[0] = rs->rates[first_idx / 2];
rs               1028 drivers/net/wireless/mediatek/mt76/mt7603/mac.c 		cur_rate = &rs->rates[idx / 2];
rs                839 drivers/net/wireless/mediatek/mt76/mt7615/mac.c 	struct mt7615_rate_set *rs;
rs                890 drivers/net/wireless/mediatek/mt76/mt7615/mac.c 	rs = &sta->rateset[rs_idx];
rs                892 drivers/net/wireless/mediatek/mt76/mt7615/mac.c 	if (!first_idx && rs->probe_rate.idx >= 0) {
rs                893 drivers/net/wireless/mediatek/mt76/mt7615/mac.c 		info->status.rates[0] = rs->probe_rate;
rs                902 drivers/net/wireless/mediatek/mt76/mt7615/mac.c 		info->status.rates[0] = rs->rates[first_idx / 2];
rs                910 drivers/net/wireless/mediatek/mt76/mt7615/mac.c 		cur_rate = &rs->rates[idx / 2];
rs               1262 drivers/net/wireless/realtek/rtw88/phy.c 				       u8 bw, u8 rs, u8 ch, s8 pwr_limit)
rs               1274 drivers/net/wireless/realtek/rtw88/phy.c 	    rs >= RTW_RATE_SECTION_MAX || ch_idx < 0) {
rs               1277 drivers/net/wireless/realtek/rtw88/phy.c 		     regd, band, bw, rs, ch_idx, pwr_limit);
rs               1282 drivers/net/wireless/realtek/rtw88/phy.c 		hal->tx_pwr_limit_2g[regd][bw][rs][ch_idx] = pwr_limit;
rs               1283 drivers/net/wireless/realtek/rtw88/phy.c 		ww = hal->tx_pwr_limit_2g[RTW_REGD_WW][bw][rs][ch_idx];
rs               1285 drivers/net/wireless/realtek/rtw88/phy.c 		hal->tx_pwr_limit_2g[RTW_REGD_WW][bw][rs][ch_idx] = ww;
rs               1287 drivers/net/wireless/realtek/rtw88/phy.c 		hal->tx_pwr_limit_5g[regd][bw][rs][ch_idx] = pwr_limit;
rs               1288 drivers/net/wireless/realtek/rtw88/phy.c 		ww = hal->tx_pwr_limit_5g[RTW_REGD_WW][bw][rs][ch_idx];
rs               1290 drivers/net/wireless/realtek/rtw88/phy.c 		hal->tx_pwr_limit_5g[RTW_REGD_WW][bw][rs][ch_idx] = ww;
rs               1367 drivers/net/wireless/realtek/rtw88/phy.c 					   p->bw, p->rs, p->ch, p->txpwr_lmt);
rs               1682 drivers/net/wireless/realtek/rtw88/phy.c 	u8 rs;
rs               1691 drivers/net/wireless/realtek/rtw88/phy.c 		rs = RTW_RATE_SECTION_CCK;
rs               1693 drivers/net/wireless/realtek/rtw88/phy.c 		rs = RTW_RATE_SECTION_OFDM;
rs               1695 drivers/net/wireless/realtek/rtw88/phy.c 		rs = RTW_RATE_SECTION_HT_1S;
rs               1697 drivers/net/wireless/realtek/rtw88/phy.c 		rs = RTW_RATE_SECTION_HT_2S;
rs               1699 drivers/net/wireless/realtek/rtw88/phy.c 		rs = RTW_RATE_SECTION_VHT_1S;
rs               1701 drivers/net/wireless/realtek/rtw88/phy.c 		rs = RTW_RATE_SECTION_VHT_2S;
rs               1706 drivers/net/wireless/realtek/rtw88/phy.c 	if (rs == RTW_RATE_SECTION_CCK || rs == RTW_RATE_SECTION_OFDM)
rs               1710 drivers/net/wireless/realtek/rtw88/phy.c 	if (rs == RTW_RATE_SECTION_HT_1S || rs == RTW_RATE_SECTION_HT_2S)
rs               1722 drivers/net/wireless/realtek/rtw88/phy.c 			hal->tx_pwr_limit_2g[regd][cur_bw][rs][ch_idx] :
rs               1723 drivers/net/wireless/realtek/rtw88/phy.c 			hal->tx_pwr_limit_5g[regd][cur_bw][rs][ch_idx];
rs               1794 drivers/net/wireless/realtek/rtw88/phy.c 					     u8 ch, u8 path, u8 rs)
rs               1805 drivers/net/wireless/realtek/rtw88/phy.c 	if (rs >= RTW_RATE_SECTION_MAX)
rs               1808 drivers/net/wireless/realtek/rtw88/phy.c 	rates = rtw_rate_section[rs];
rs               1809 drivers/net/wireless/realtek/rtw88/phy.c 	size = rtw_rate_size[rs];
rs               1828 drivers/net/wireless/realtek/rtw88/phy.c 	u8 rs;
rs               1832 drivers/net/wireless/realtek/rtw88/phy.c 		rs = RTW_RATE_SECTION_CCK;
rs               1834 drivers/net/wireless/realtek/rtw88/phy.c 		rs = RTW_RATE_SECTION_OFDM;
rs               1836 drivers/net/wireless/realtek/rtw88/phy.c 	for (; rs < RTW_RATE_SECTION_MAX; rs++)
rs               1837 drivers/net/wireless/realtek/rtw88/phy.c 		rtw_phy_set_tx_power_index_by_rs(rtwdev, ch, path, rs);
rs               1857 drivers/net/wireless/realtek/rtw88/phy.c 					u8 rs, u8 size, u8 *rates)
rs               1863 drivers/net/wireless/realtek/rtw88/phy.c 	if (rs >= RTW_RATE_SECTION_VHT_1S)
rs               1869 drivers/net/wireless/realtek/rtw88/phy.c 	hal->tx_pwr_by_rate_base_2g[path][rs] = base_2g;
rs               1870 drivers/net/wireless/realtek/rtw88/phy.c 	hal->tx_pwr_by_rate_base_5g[path][rs] = base_5g;
rs               1905 drivers/net/wireless/realtek/rtw88/phy.c __rtw_phy_tx_power_limit_config(struct rtw_hal *hal, u8 regd, u8 bw, u8 rs)
rs               1911 drivers/net/wireless/realtek/rtw88/phy.c 		base = hal->tx_pwr_by_rate_base_2g[0][rs];
rs               1912 drivers/net/wireless/realtek/rtw88/phy.c 		hal->tx_pwr_limit_2g[regd][bw][rs][ch] -= base;
rs               1916 drivers/net/wireless/realtek/rtw88/phy.c 		base = hal->tx_pwr_by_rate_base_5g[0][rs];
rs               1917 drivers/net/wireless/realtek/rtw88/phy.c 		hal->tx_pwr_limit_5g[regd][bw][rs][ch] -= base;
rs               1923 drivers/net/wireless/realtek/rtw88/phy.c 	u8 regd, bw, rs;
rs               1930 drivers/net/wireless/realtek/rtw88/phy.c 			for (rs = 0; rs < RTW_RATE_SECTION_MAX; rs++)
rs               1931 drivers/net/wireless/realtek/rtw88/phy.c 				__rtw_phy_tx_power_limit_config(hal, regd, bw, rs);
rs               1935 drivers/net/wireless/realtek/rtw88/phy.c 					u8 regd, u8 bw, u8 rs)
rs               1943 drivers/net/wireless/realtek/rtw88/phy.c 		hal->tx_pwr_limit_2g[regd][bw][rs][ch] = max_power_index;
rs               1947 drivers/net/wireless/realtek/rtw88/phy.c 		hal->tx_pwr_limit_5g[regd][bw][rs][ch] = max_power_index;
rs               1953 drivers/net/wireless/realtek/rtw88/phy.c 	u8 regd, path, rate, rs, bw;
rs               1966 drivers/net/wireless/realtek/rtw88/phy.c 			for (rs = 0; rs < RTW_RATE_SECTION_MAX; rs++)
rs               1968 drivers/net/wireless/realtek/rtw88/phy.c 							    rs);
rs                 52 drivers/net/wireless/realtek/rtw88/phy.h 	u8 rs;
rs                867 drivers/net/wireless/realtek/rtw88/rtw8822b.c rtw8822b_set_tx_power_index_by_rate(struct rtw_dev *rtwdev, u8 path, u8 rs)
rs                875 drivers/net/wireless/realtek/rtw88/rtw8822b.c 	for (j = 0; j < rtw_rate_size[rs]; j++) {
rs                876 drivers/net/wireless/realtek/rtw88/rtw8822b.c 		rate = rtw_rate_section[rs][j];
rs                892 drivers/net/wireless/realtek/rtw88/rtw8822b.c 	int rs, path;
rs                895 drivers/net/wireless/realtek/rtw88/rtw8822b.c 		for (rs = 0; rs < RTW_RATE_SECTION_MAX; rs++)
rs                896 drivers/net/wireless/realtek/rtw88/rtw8822b.c 			rtw8822b_set_tx_power_index_by_rate(rtwdev, path, rs);
rs               1780 drivers/net/wireless/realtek/rtw88/rtw8822c.c 	u8 rs, rate, j;
rs               1790 drivers/net/wireless/realtek/rtw88/rtw8822c.c 	for (rs = 0; rs < RTW_RATE_SECTION_MAX; rs++) {
rs               1791 drivers/net/wireless/realtek/rtw88/rtw8822c.c 		for (j = 0; j < rtw_rate_size[rs]; j++) {
rs               1792 drivers/net/wireless/realtek/rtw88/rtw8822c.c 			rate = rtw_rate_section[rs][j];
rs               1795 drivers/net/wireless/realtek/rtw88/rtw8822c.c 			if (rs == 0) {
rs                 58 drivers/pci/hotplug/ibmphp_res.c 	struct resource_node *rs;
rs                 65 drivers/pci/hotplug/ibmphp_res.c 	rs = kzalloc(sizeof(struct resource_node), GFP_KERNEL);
rs                 66 drivers/pci/hotplug/ibmphp_res.c 	if (!rs)
rs                 69 drivers/pci/hotplug/ibmphp_res.c 	rs->busno = curr->bus_num;
rs                 70 drivers/pci/hotplug/ibmphp_res.c 	rs->devfunc = curr->dev_fun;
rs                 71 drivers/pci/hotplug/ibmphp_res.c 	rs->start = curr->start_addr;
rs                 72 drivers/pci/hotplug/ibmphp_res.c 	rs->end = curr->end_addr;
rs                 73 drivers/pci/hotplug/ibmphp_res.c 	rs->len = curr->end_addr - curr->start_addr + 1;
rs                 74 drivers/pci/hotplug/ibmphp_res.c 	return rs;
rs                179 drivers/platform/chrome/wilco_ec/debugfs.c 	struct hi_gpio_status_response rs;
rs                191 drivers/platform/chrome/wilco_ec/debugfs.c 	msg.response_data = &rs;
rs                192 drivers/platform/chrome/wilco_ec/debugfs.c 	msg.response_size = sizeof(rs);
rs                196 drivers/platform/chrome/wilco_ec/debugfs.c 	if (rs.status)
rs                199 drivers/platform/chrome/wilco_ec/debugfs.c 	*val = rs.val;
rs                119 drivers/platform/chrome/wilco_ec/mailbox.c 	struct wilco_ec_response *rs;
rs                151 drivers/platform/chrome/wilco_ec/mailbox.c 	rs = ec->data_buffer;
rs                153 drivers/platform/chrome/wilco_ec/mailbox.c 					    sizeof(*rs) + EC_MAILBOX_DATA_SIZE,
rs                154 drivers/platform/chrome/wilco_ec/mailbox.c 					    (u8 *)rs);
rs                156 drivers/platform/chrome/wilco_ec/mailbox.c 		dev_dbg(ec->dev, "bad packet checksum 0x%02x\n", rs->checksum);
rs                160 drivers/platform/chrome/wilco_ec/mailbox.c 	if (rs->result) {
rs                161 drivers/platform/chrome/wilco_ec/mailbox.c 		dev_dbg(ec->dev, "EC reported failure: 0x%02x\n", rs->result);
rs                165 drivers/platform/chrome/wilco_ec/mailbox.c 	if (rs->data_size != EC_MAILBOX_DATA_SIZE) {
rs                167 drivers/platform/chrome/wilco_ec/mailbox.c 			rs->data_size, EC_MAILBOX_DATA_SIZE);
rs                171 drivers/platform/chrome/wilco_ec/mailbox.c 	if (rs->data_size < msg->response_size) {
rs                173 drivers/platform/chrome/wilco_ec/mailbox.c 			rs->data_size, msg->response_size);
rs                177 drivers/platform/chrome/wilco_ec/mailbox.c 	memcpy(msg->response_data, rs->data, msg->response_size);
rs                179 drivers/platform/chrome/wilco_ec/mailbox.c 	return rs->data_size;
rs                 33 drivers/platform/chrome/wilco_ec/properties.c 			     struct ec_property_response *rs)
rs                 42 drivers/platform/chrome/wilco_ec/properties.c 	ec_msg.response_data = rs;
rs                 43 drivers/platform/chrome/wilco_ec/properties.c 	ec_msg.response_size = sizeof(*rs);
rs                 48 drivers/platform/chrome/wilco_ec/properties.c 	if (rs->op != rq->op)
rs                 50 drivers/platform/chrome/wilco_ec/properties.c 	if (memcmp(rq->property_id, rs->property_id, sizeof(rs->property_id)))
rs                 60 drivers/platform/chrome/wilco_ec/properties.c 	struct ec_property_response rs;
rs                 67 drivers/platform/chrome/wilco_ec/properties.c 	ret = send_property_msg(ec, &rq, &rs);
rs                 71 drivers/platform/chrome/wilco_ec/properties.c 	prop_msg->length = rs.length;
rs                 72 drivers/platform/chrome/wilco_ec/properties.c 	memcpy(prop_msg->data, rs.data, rs.length);
rs                 82 drivers/platform/chrome/wilco_ec/properties.c 	struct ec_property_response rs;
rs                 91 drivers/platform/chrome/wilco_ec/properties.c 	ret = send_property_msg(ec, &rq, &rs);
rs                 94 drivers/platform/chrome/wilco_ec/properties.c 	if (rs.length != prop_msg->length)
rs                289 drivers/s390/cio/chsc.c 	u8  rs;				/* reporting source */
rs                393 drivers/s390/cio/chsc.c 		      sei_area->rs, sei_area->rsid, sei_area->ccdf[0]);
rs                410 drivers/s390/cio/chsc.c 			sei_area->rs, sei_area->rsid, lir->ic, iuparams,
rs                416 drivers/s390/cio/chsc.c 		       sei_area->rs, sei_area->rsid, lir->ic, iuparams,
rs                432 drivers/s390/cio/chsc.c 		      "rs_id=%04x)\n", sei_area->rs, sei_area->rsid);
rs                433 drivers/s390/cio/chsc.c 	if (sei_area->rs != 4)
rs                472 drivers/s390/cio/chsc.c 	if (sei_area->rs != 0)
rs                508 drivers/s390/cio/chsc.c 	if (sei_area->rs != 0)
rs                537 drivers/s390/cio/chsc.c 	if (sei_area->rs != 7)
rs                551 drivers/s390/cio/chsc.c 	if (sei_area->rs != 7)
rs                563 drivers/s390/cio/chsc.c 	if (sei_area->rs != 5)
rs                 67 drivers/spi/spi-mt7621.c static inline u32 mt7621_spi_read(struct mt7621_spi *rs, u32 reg)
rs                 69 drivers/spi/spi-mt7621.c 	return ioread32(rs->base + reg);
rs                 72 drivers/spi/spi-mt7621.c static inline void mt7621_spi_write(struct mt7621_spi *rs, u32 reg, u32 val)
rs                 74 drivers/spi/spi-mt7621.c 	iowrite32(val, rs->base + reg);
rs                 79 drivers/spi/spi-mt7621.c 	struct mt7621_spi *rs = spidev_to_mt7621_spi(spi);
rs                 89 drivers/spi/spi-mt7621.c 	master = mt7621_spi_read(rs, MT7621_SPI_MASTER);
rs                 92 drivers/spi/spi-mt7621.c 	mt7621_spi_write(rs, MT7621_SPI_MASTER, master);
rs                 94 drivers/spi/spi-mt7621.c 	rs->pending_write = 0;
rs                 98 drivers/spi/spi-mt7621.c 	mt7621_spi_write(rs, MT7621_SPI_POLAR, polar);
rs                103 drivers/spi/spi-mt7621.c 	struct mt7621_spi *rs = spidev_to_mt7621_spi(spi);
rs                109 drivers/spi/spi-mt7621.c 	rate = DIV_ROUND_UP(rs->sys_freq, speed);
rs                118 drivers/spi/spi-mt7621.c 	reg = mt7621_spi_read(rs, MT7621_SPI_MASTER);
rs                121 drivers/spi/spi-mt7621.c 	rs->speed = speed;
rs                134 drivers/spi/spi-mt7621.c 	mt7621_spi_write(rs, MT7621_SPI_MASTER, reg);
rs                139 drivers/spi/spi-mt7621.c static inline int mt7621_spi_wait_till_ready(struct mt7621_spi *rs)
rs                146 drivers/spi/spi-mt7621.c 		status = mt7621_spi_read(rs, MT7621_SPI_TRANS);
rs                156 drivers/spi/spi-mt7621.c static void mt7621_spi_read_half_duplex(struct mt7621_spi *rs,
rs                166 drivers/spi/spi-mt7621.c 	tx_len = rs->pending_write;
rs                167 drivers/spi/spi-mt7621.c 	rs->pending_write = 0;
rs                177 drivers/spi/spi-mt7621.c 		mt7621_spi_write(rs, MT7621_SPI_MOREBUF, val);
rs                181 drivers/spi/spi-mt7621.c 		val = mt7621_spi_read(rs, MT7621_SPI_TRANS);
rs                183 drivers/spi/spi-mt7621.c 		mt7621_spi_write(rs, MT7621_SPI_TRANS, val);
rs                185 drivers/spi/spi-mt7621.c 		mt7621_spi_wait_till_ready(rs);
rs                189 drivers/spi/spi-mt7621.c 				val = mt7621_spi_read(rs, MT7621_SPI_DATA0 + i);
rs                198 drivers/spi/spi-mt7621.c static inline void mt7621_spi_flush(struct mt7621_spi *rs)
rs                200 drivers/spi/spi-mt7621.c 	mt7621_spi_read_half_duplex(rs, 0, NULL);
rs                203 drivers/spi/spi-mt7621.c static void mt7621_spi_write_half_duplex(struct mt7621_spi *rs,
rs                206 drivers/spi/spi-mt7621.c 	int len = rs->pending_write;
rs                210 drivers/spi/spi-mt7621.c 		val = mt7621_spi_read(rs, MT7621_SPI_OPCODE + (len & ~3));
rs                219 drivers/spi/spi-mt7621.c 			rs->pending_write = len;
rs                220 drivers/spi/spi-mt7621.c 			mt7621_spi_flush(rs);
rs                230 drivers/spi/spi-mt7621.c 			mt7621_spi_write(rs, MT7621_SPI_OPCODE + len - 4, val);
rs                241 drivers/spi/spi-mt7621.c 		mt7621_spi_write(rs, MT7621_SPI_OPCODE + (len & ~3), val);
rs                244 drivers/spi/spi-mt7621.c 	rs->pending_write = len;
rs                250 drivers/spi/spi-mt7621.c 	struct mt7621_spi *rs = spi_controller_get_devdata(master);
rs                256 drivers/spi/spi-mt7621.c 	mt7621_spi_wait_till_ready(rs);
rs                283 drivers/spi/spi-mt7621.c 			mt7621_spi_read_half_duplex(rs, t->len, t->rx_buf);
rs                285 drivers/spi/spi-mt7621.c 			mt7621_spi_write_half_duplex(rs, t->len, t->tx_buf);
rs                291 drivers/spi/spi-mt7621.c 	mt7621_spi_flush(rs);
rs                303 drivers/spi/spi-mt7621.c 	struct mt7621_spi *rs = spidev_to_mt7621_spi(spi);
rs                306 drivers/spi/spi-mt7621.c 	    (spi->max_speed_hz > (rs->sys_freq / 2)))
rs                307 drivers/spi/spi-mt7621.c 		spi->max_speed_hz = rs->sys_freq / 2;
rs                309 drivers/spi/spi-mt7621.c 	if (spi->max_speed_hz < (rs->sys_freq / 4097)) {
rs                328 drivers/spi/spi-mt7621.c 	struct mt7621_spi *rs;
rs                353 drivers/spi/spi-mt7621.c 	master = spi_alloc_master(&pdev->dev, sizeof(*rs));
rs                369 drivers/spi/spi-mt7621.c 	rs = spi_controller_get_devdata(master);
rs                370 drivers/spi/spi-mt7621.c 	rs->base = base;
rs                371 drivers/spi/spi-mt7621.c 	rs->clk = clk;
rs                372 drivers/spi/spi-mt7621.c 	rs->master = master;
rs                373 drivers/spi/spi-mt7621.c 	rs->sys_freq = clk_get_rate(rs->clk);
rs                374 drivers/spi/spi-mt7621.c 	rs->pending_write = 0;
rs                375 drivers/spi/spi-mt7621.c 	dev_info(&pdev->dev, "sys_freq: %u\n", rs->sys_freq);
rs                389 drivers/spi/spi-mt7621.c 	struct mt7621_spi *rs;
rs                392 drivers/spi/spi-mt7621.c 	rs = spi_controller_get_devdata(master);
rs                394 drivers/spi/spi-mt7621.c 	clk_disable_unprepare(rs->clk);
rs                188 drivers/spi/spi-rockchip.c static inline void spi_enable_chip(struct rockchip_spi *rs, bool enable)
rs                190 drivers/spi/spi-rockchip.c 	writel_relaxed((enable ? 1U : 0U), rs->regs + ROCKCHIP_SPI_SSIENR);
rs                193 drivers/spi/spi-rockchip.c static inline void wait_for_idle(struct rockchip_spi *rs)
rs                198 drivers/spi/spi-rockchip.c 		if (!(readl_relaxed(rs->regs + ROCKCHIP_SPI_SR) & SR_BUSY))
rs                202 drivers/spi/spi-rockchip.c 	dev_warn(rs->dev, "spi controller is in busy state!\n");
rs                205 drivers/spi/spi-rockchip.c static u32 get_fifo_len(struct rockchip_spi *rs)
rs                210 drivers/spi/spi-rockchip.c 		writel_relaxed(fifo, rs->regs + ROCKCHIP_SPI_TXFTLR);
rs                211 drivers/spi/spi-rockchip.c 		if (fifo != readl_relaxed(rs->regs + ROCKCHIP_SPI_TXFTLR))
rs                215 drivers/spi/spi-rockchip.c 	writel_relaxed(0, rs->regs + ROCKCHIP_SPI_TXFTLR);
rs                223 drivers/spi/spi-rockchip.c 	struct rockchip_spi *rs = spi_master_get_devdata(master);
rs                227 drivers/spi/spi-rockchip.c 	if (cs_asserted == rs->cs_asserted[spi->chip_select])
rs                232 drivers/spi/spi-rockchip.c 		pm_runtime_get_sync(rs->dev);
rs                234 drivers/spi/spi-rockchip.c 		ROCKCHIP_SPI_SET_BITS(rs->regs + ROCKCHIP_SPI_SER,
rs                237 drivers/spi/spi-rockchip.c 		ROCKCHIP_SPI_CLR_BITS(rs->regs + ROCKCHIP_SPI_SER,
rs                241 drivers/spi/spi-rockchip.c 		pm_runtime_put(rs->dev);
rs                244 drivers/spi/spi-rockchip.c 	rs->cs_asserted[spi->chip_select] = cs_asserted;
rs                250 drivers/spi/spi-rockchip.c 	struct rockchip_spi *rs = spi_master_get_devdata(master);
rs                255 drivers/spi/spi-rockchip.c 	spi_enable_chip(rs, false);
rs                258 drivers/spi/spi-rockchip.c 	writel_relaxed(0, rs->regs + ROCKCHIP_SPI_IMR);
rs                260 drivers/spi/spi-rockchip.c 	if (atomic_read(&rs->state) & TXDMA)
rs                263 drivers/spi/spi-rockchip.c 	if (atomic_read(&rs->state) & RXDMA)
rs                267 drivers/spi/spi-rockchip.c static void rockchip_spi_pio_writer(struct rockchip_spi *rs)
rs                269 drivers/spi/spi-rockchip.c 	u32 tx_free = rs->fifo_len - readl_relaxed(rs->regs + ROCKCHIP_SPI_TXFLR);
rs                270 drivers/spi/spi-rockchip.c 	u32 words = min(rs->tx_left, tx_free);
rs                272 drivers/spi/spi-rockchip.c 	rs->tx_left -= words;
rs                276 drivers/spi/spi-rockchip.c 		if (rs->n_bytes == 1)
rs                277 drivers/spi/spi-rockchip.c 			txw = *(u8 *)rs->tx;
rs                279 drivers/spi/spi-rockchip.c 			txw = *(u16 *)rs->tx;
rs                281 drivers/spi/spi-rockchip.c 		writel_relaxed(txw, rs->regs + ROCKCHIP_SPI_TXDR);
rs                282 drivers/spi/spi-rockchip.c 		rs->tx += rs->n_bytes;
rs                286 drivers/spi/spi-rockchip.c static void rockchip_spi_pio_reader(struct rockchip_spi *rs)
rs                288 drivers/spi/spi-rockchip.c 	u32 words = readl_relaxed(rs->regs + ROCKCHIP_SPI_RXFLR);
rs                289 drivers/spi/spi-rockchip.c 	u32 rx_left = rs->rx_left - words;
rs                297 drivers/spi/spi-rockchip.c 		u32 ftl = readl_relaxed(rs->regs + ROCKCHIP_SPI_RXFTLR) + 1;
rs                301 drivers/spi/spi-rockchip.c 			words = rs->rx_left - rx_left;
rs                305 drivers/spi/spi-rockchip.c 	rs->rx_left = rx_left;
rs                307 drivers/spi/spi-rockchip.c 		u32 rxw = readl_relaxed(rs->regs + ROCKCHIP_SPI_RXDR);
rs                309 drivers/spi/spi-rockchip.c 		if (!rs->rx)
rs                312 drivers/spi/spi-rockchip.c 		if (rs->n_bytes == 1)
rs                313 drivers/spi/spi-rockchip.c 			*(u8 *)rs->rx = (u8)rxw;
rs                315 drivers/spi/spi-rockchip.c 			*(u16 *)rs->rx = (u16)rxw;
rs                316 drivers/spi/spi-rockchip.c 		rs->rx += rs->n_bytes;
rs                323 drivers/spi/spi-rockchip.c 	struct rockchip_spi *rs = spi_master_get_devdata(master);
rs                325 drivers/spi/spi-rockchip.c 	if (rs->tx_left)
rs                326 drivers/spi/spi-rockchip.c 		rockchip_spi_pio_writer(rs);
rs                328 drivers/spi/spi-rockchip.c 	rockchip_spi_pio_reader(rs);
rs                329 drivers/spi/spi-rockchip.c 	if (!rs->rx_left) {
rs                330 drivers/spi/spi-rockchip.c 		spi_enable_chip(rs, false);
rs                331 drivers/spi/spi-rockchip.c 		writel_relaxed(0, rs->regs + ROCKCHIP_SPI_IMR);
rs                338 drivers/spi/spi-rockchip.c static int rockchip_spi_prepare_irq(struct rockchip_spi *rs,
rs                341 drivers/spi/spi-rockchip.c 	rs->tx = xfer->tx_buf;
rs                342 drivers/spi/spi-rockchip.c 	rs->rx = xfer->rx_buf;
rs                343 drivers/spi/spi-rockchip.c 	rs->tx_left = rs->tx ? xfer->len / rs->n_bytes : 0;
rs                344 drivers/spi/spi-rockchip.c 	rs->rx_left = xfer->len / rs->n_bytes;
rs                346 drivers/spi/spi-rockchip.c 	writel_relaxed(INT_RF_FULL, rs->regs + ROCKCHIP_SPI_IMR);
rs                347 drivers/spi/spi-rockchip.c 	spi_enable_chip(rs, true);
rs                349 drivers/spi/spi-rockchip.c 	if (rs->tx_left)
rs                350 drivers/spi/spi-rockchip.c 		rockchip_spi_pio_writer(rs);
rs                359 drivers/spi/spi-rockchip.c 	struct rockchip_spi *rs = spi_master_get_devdata(master);
rs                360 drivers/spi/spi-rockchip.c 	int state = atomic_fetch_andnot(RXDMA, &rs->state);
rs                365 drivers/spi/spi-rockchip.c 	spi_enable_chip(rs, false);
rs                372 drivers/spi/spi-rockchip.c 	struct rockchip_spi *rs = spi_master_get_devdata(master);
rs                373 drivers/spi/spi-rockchip.c 	int state = atomic_fetch_andnot(TXDMA, &rs->state);
rs                379 drivers/spi/spi-rockchip.c 	wait_for_idle(rs);
rs                381 drivers/spi/spi-rockchip.c 	spi_enable_chip(rs, false);
rs                385 drivers/spi/spi-rockchip.c static int rockchip_spi_prepare_dma(struct rockchip_spi *rs,
rs                390 drivers/spi/spi-rockchip.c 	atomic_set(&rs->state, 0);
rs                396 drivers/spi/spi-rockchip.c 			.src_addr = rs->dma_addr_rx,
rs                397 drivers/spi/spi-rockchip.c 			.src_addr_width = rs->n_bytes,
rs                418 drivers/spi/spi-rockchip.c 			.dst_addr = rs->dma_addr_tx,
rs                419 drivers/spi/spi-rockchip.c 			.dst_addr_width = rs->n_bytes,
rs                420 drivers/spi/spi-rockchip.c 			.dst_maxburst = rs->fifo_len / 4,
rs                441 drivers/spi/spi-rockchip.c 		atomic_or(RXDMA, &rs->state);
rs                446 drivers/spi/spi-rockchip.c 	spi_enable_chip(rs, true);
rs                449 drivers/spi/spi-rockchip.c 		atomic_or(TXDMA, &rs->state);
rs                458 drivers/spi/spi-rockchip.c static void rockchip_spi_config(struct rockchip_spi *rs,
rs                469 drivers/spi/spi-rockchip.c 	cr0 |= rs->rsd << CR0_RSD_OFFSET;
rs                509 drivers/spi/spi-rockchip.c 	writel_relaxed(cr0, rs->regs + ROCKCHIP_SPI_CTRLR0);
rs                510 drivers/spi/spi-rockchip.c 	writel_relaxed(cr1, rs->regs + ROCKCHIP_SPI_CTRLR1);
rs                516 drivers/spi/spi-rockchip.c 	if (xfer->len < rs->fifo_len)
rs                517 drivers/spi/spi-rockchip.c 		writel_relaxed(xfer->len - 1, rs->regs + ROCKCHIP_SPI_RXFTLR);
rs                519 drivers/spi/spi-rockchip.c 		writel_relaxed(rs->fifo_len / 2 - 1, rs->regs + ROCKCHIP_SPI_RXFTLR);
rs                521 drivers/spi/spi-rockchip.c 	writel_relaxed(rs->fifo_len / 2, rs->regs + ROCKCHIP_SPI_DMATDLR);
rs                522 drivers/spi/spi-rockchip.c 	writel_relaxed(0, rs->regs + ROCKCHIP_SPI_DMARDLR);
rs                523 drivers/spi/spi-rockchip.c 	writel_relaxed(dmacr, rs->regs + ROCKCHIP_SPI_DMACR);
rs                529 drivers/spi/spi-rockchip.c 	writel_relaxed(2 * DIV_ROUND_UP(rs->freq, 2 * xfer->speed_hz),
rs                530 drivers/spi/spi-rockchip.c 			rs->regs + ROCKCHIP_SPI_BAUDR);
rs                543 drivers/spi/spi-rockchip.c 	struct rockchip_spi *rs = spi_master_get_devdata(master);
rs                546 drivers/spi/spi-rockchip.c 	WARN_ON(readl_relaxed(rs->regs + ROCKCHIP_SPI_SSIENR) &&
rs                547 drivers/spi/spi-rockchip.c 		(readl_relaxed(rs->regs + ROCKCHIP_SPI_SR) & SR_BUSY));
rs                550 drivers/spi/spi-rockchip.c 		dev_err(rs->dev, "No buffer for transfer\n");
rs                555 drivers/spi/spi-rockchip.c 		dev_err(rs->dev, "Transfer is too long (%d)\n", xfer->len);
rs                559 drivers/spi/spi-rockchip.c 	rs->n_bytes = xfer->bits_per_word <= 8 ? 1 : 2;
rs                563 drivers/spi/spi-rockchip.c 	rockchip_spi_config(rs, spi, xfer, use_dma);
rs                566 drivers/spi/spi-rockchip.c 		return rockchip_spi_prepare_dma(rs, master, xfer);
rs                568 drivers/spi/spi-rockchip.c 	return rockchip_spi_prepare_irq(rs, xfer);
rs                575 drivers/spi/spi-rockchip.c 	struct rockchip_spi *rs = spi_master_get_devdata(master);
rs                582 drivers/spi/spi-rockchip.c 	return xfer->len / bytes_per_word >= rs->fifo_len;
rs                588 drivers/spi/spi-rockchip.c 	struct rockchip_spi *rs;
rs                599 drivers/spi/spi-rockchip.c 	rs = spi_master_get_devdata(master);
rs                603 drivers/spi/spi-rockchip.c 	rs->regs = devm_ioremap_resource(&pdev->dev, mem);
rs                604 drivers/spi/spi-rockchip.c 	if (IS_ERR(rs->regs)) {
rs                605 drivers/spi/spi-rockchip.c 		ret =  PTR_ERR(rs->regs);
rs                609 drivers/spi/spi-rockchip.c 	rs->apb_pclk = devm_clk_get(&pdev->dev, "apb_pclk");
rs                610 drivers/spi/spi-rockchip.c 	if (IS_ERR(rs->apb_pclk)) {
rs                612 drivers/spi/spi-rockchip.c 		ret = PTR_ERR(rs->apb_pclk);
rs                616 drivers/spi/spi-rockchip.c 	rs->spiclk = devm_clk_get(&pdev->dev, "spiclk");
rs                617 drivers/spi/spi-rockchip.c 	if (IS_ERR(rs->spiclk)) {
rs                619 drivers/spi/spi-rockchip.c 		ret = PTR_ERR(rs->spiclk);
rs                623 drivers/spi/spi-rockchip.c 	ret = clk_prepare_enable(rs->apb_pclk);
rs                629 drivers/spi/spi-rockchip.c 	ret = clk_prepare_enable(rs->spiclk);
rs                635 drivers/spi/spi-rockchip.c 	spi_enable_chip(rs, false);
rs                646 drivers/spi/spi-rockchip.c 	rs->dev = &pdev->dev;
rs                647 drivers/spi/spi-rockchip.c 	rs->freq = clk_get_rate(rs->spiclk);
rs                652 drivers/spi/spi-rockchip.c 		u32 rsd = DIV_ROUND_CLOSEST(rsd_nsecs * (rs->freq >> 8),
rs                655 drivers/spi/spi-rockchip.c 			dev_warn(rs->dev, "%u Hz are too slow to express %u ns delay\n",
rs                656 drivers/spi/spi-rockchip.c 					rs->freq, rsd_nsecs);
rs                659 drivers/spi/spi-rockchip.c 			dev_warn(rs->dev, "%u Hz are too fast to express %u ns delay, clamping at %u ns\n",
rs                660 drivers/spi/spi-rockchip.c 					rs->freq, rsd_nsecs,
rs                661 drivers/spi/spi-rockchip.c 					CR0_RSD_MAX * 1000000000U / rs->freq);
rs                663 drivers/spi/spi-rockchip.c 		rs->rsd = rsd;
rs                666 drivers/spi/spi-rockchip.c 	rs->fifo_len = get_fifo_len(rs);
rs                667 drivers/spi/spi-rockchip.c 	if (!rs->fifo_len) {
rs                682 drivers/spi/spi-rockchip.c 	master->min_speed_hz = rs->freq / BAUDR_SCKDV_MAX;
rs                683 drivers/spi/spi-rockchip.c 	master->max_speed_hz = min(rs->freq / BAUDR_SCKDV_MIN, MAX_SCLK_OUT);
rs                691 drivers/spi/spi-rockchip.c 	master->dma_tx = dma_request_chan(rs->dev, "tx");
rs                698 drivers/spi/spi-rockchip.c 		dev_warn(rs->dev, "Failed to request TX DMA channel\n");
rs                702 drivers/spi/spi-rockchip.c 	master->dma_rx = dma_request_chan(rs->dev, "rx");
rs                708 drivers/spi/spi-rockchip.c 		dev_warn(rs->dev, "Failed to request RX DMA channel\n");
rs                713 drivers/spi/spi-rockchip.c 		rs->dma_addr_tx = mem->start + ROCKCHIP_SPI_TXDR;
rs                714 drivers/spi/spi-rockchip.c 		rs->dma_addr_rx = mem->start + ROCKCHIP_SPI_RXDR;
rs                735 drivers/spi/spi-rockchip.c 	clk_disable_unprepare(rs->spiclk);
rs                737 drivers/spi/spi-rockchip.c 	clk_disable_unprepare(rs->apb_pclk);
rs                747 drivers/spi/spi-rockchip.c 	struct rockchip_spi *rs = spi_master_get_devdata(master);
rs                751 drivers/spi/spi-rockchip.c 	clk_disable_unprepare(rs->spiclk);
rs                752 drivers/spi/spi-rockchip.c 	clk_disable_unprepare(rs->apb_pclk);
rs                791 drivers/spi/spi-rockchip.c 	struct rockchip_spi *rs = spi_master_get_devdata(master);
rs                801 drivers/spi/spi-rockchip.c 		clk_disable_unprepare(rs->spiclk);
rs                802 drivers/spi/spi-rockchip.c 		clk_disable_unprepare(rs->apb_pclk);
rs                813 drivers/spi/spi-rockchip.c 	struct rockchip_spi *rs = spi_master_get_devdata(master);
rs                815 drivers/spi/spi-rockchip.c 	clk_disable_unprepare(rs->spiclk);
rs                816 drivers/spi/spi-rockchip.c 	clk_disable_unprepare(rs->apb_pclk);
rs                825 drivers/spi/spi-rockchip.c 	struct rockchip_spi *rs = spi_master_get_devdata(master);
rs                827 drivers/spi/spi-rockchip.c 	ret = clk_prepare_enable(rs->apb_pclk);
rs                831 drivers/spi/spi-rockchip.c 	ret = clk_prepare_enable(rs->spiclk);
rs                833 drivers/spi/spi-rockchip.c 		clk_disable_unprepare(rs->apb_pclk);
rs                285 drivers/staging/comedi/drivers/ni_routes.c 		struct ni_route_set *rs = &valid_routes->routes[n];
rs                288 drivers/staging/comedi/drivers/ni_routes.c 		rs->n_src = 0;
rs                289 drivers/staging/comedi/drivers/ni_routes.c 		while (rs->src[rs->n_src])
rs                290 drivers/staging/comedi/drivers/ni_routes.c 			++rs->n_src;
rs               2441 drivers/tty/cyclades.c 		u32 rs;
rs               2447 drivers/tty/cyclades.c 		rs = readl(&ch_ctrl->rs_control);
rs               2449 drivers/tty/cyclades.c 			rs |= C_RS_RTS;
rs               2451 drivers/tty/cyclades.c 			rs &= ~C_RS_RTS;
rs               2453 drivers/tty/cyclades.c 			rs |= C_RS_DTR;
rs               2459 drivers/tty/cyclades.c 			rs &= ~C_RS_DTR;
rs               2465 drivers/tty/cyclades.c 		cy_writel(&ch_ctrl->rs_control, rs);
rs               3012 drivers/tty/cyclades.c 	u32 rs;
rs               3014 drivers/tty/cyclades.c 	rs = readl(&ch_ctrl->rs_control);
rs               3016 drivers/tty/cyclades.c 		rs |= C_RS_RTS | C_RS_DTR;
rs               3018 drivers/tty/cyclades.c 		rs &= ~(C_RS_RTS | C_RS_DTR);
rs               3019 drivers/tty/cyclades.c 	cy_writel(&ch_ctrl->rs_control, rs);
rs                100 drivers/video/backlight/ili922x.c #define START_BYTE(id, rs, rw)	\
rs                101 drivers/video/backlight/ili922x.c 	(0x70 | (((id) & 0x01) << 2) | (((rs) & 0x01) << 1) | ((rw) & 0x01))
rs                147 drivers/video/backlight/ili922x.c static int ili922x_read_status(struct spi_device *spi, u16 *rs)
rs                180 drivers/video/backlight/ili922x.c 	*rs = (rbuf[2] << 8) + rbuf[3];
rs                212 drivers/video/fbdev/core/bitblit.c 	unsigned int rs = info->var.xres - rw;
rs                220 drivers/video/fbdev/core/bitblit.c 		region.dx = info->var.xoffset + rs;
rs                230 drivers/video/fbdev/core/bitblit.c 		region.width = rs;
rs                181 drivers/video/fbdev/core/fbcon_cw.c 	unsigned int rs = info->var.yres - rw;
rs                189 drivers/video/fbdev/core/fbcon_cw.c 		region.dy = info->var.yoffset + rs;
rs                266 drivers/watchdog/omap_wdt.c 		u32 rs = pdata->read_reset_sources();
rs                267 drivers/watchdog/omap_wdt.c 		if (rs & (1 << OMAP_MPU_WD_RST_SRC_ID_SHIFT))
rs                241 fs/9p/vfs_super.c 	struct p9_rstatfs rs;
rs                252 fs/9p/vfs_super.c 		res = p9_client_statfs(fid, &rs);
rs                254 fs/9p/vfs_super.c 			buf->f_type = rs.type;
rs                255 fs/9p/vfs_super.c 			buf->f_bsize = rs.bsize;
rs                256 fs/9p/vfs_super.c 			buf->f_blocks = rs.blocks;
rs                257 fs/9p/vfs_super.c 			buf->f_bfree = rs.bfree;
rs                258 fs/9p/vfs_super.c 			buf->f_bavail = rs.bavail;
rs                259 fs/9p/vfs_super.c 			buf->f_files = rs.files;
rs                260 fs/9p/vfs_super.c 			buf->f_ffree = rs.ffree;
rs                261 fs/9p/vfs_super.c 			buf->f_fsid.val[0] = rs.fsid & 0xFFFFFFFFUL;
rs                262 fs/9p/vfs_super.c 			buf->f_fsid.val[1] = (rs.fsid >> 32) & 0xFFFFFFFFUL;
rs                263 fs/9p/vfs_super.c 			buf->f_namelen = rs.namelen;
rs                305 fs/ceph/inode.c 	struct ceph_frag_tree_split *rs = (struct ceph_frag_tree_split*)r;
rs                307 fs/ceph/inode.c 				 le32_to_cpu(rs->frag));
rs                457 fs/crypto/crypto.c 	static DEFINE_RATELIMIT_STATE(rs, DEFAULT_RATELIMIT_INTERVAL,
rs                462 fs/crypto/crypto.c 	if (!__ratelimit(&rs))
rs                 72 fs/dlm/rcom.c  static void set_rcom_status(struct dlm_ls *ls, struct rcom_status *rs,
rs                 75 fs/dlm/rcom.c  	rs->rs_flags = cpu_to_le32(flags);
rs                196 fs/dlm/rcom.c  	struct rcom_status *rs;
rs                208 fs/dlm/rcom.c  	rs = (struct rcom_status *)rc_in->rc_buf;
rs                210 fs/dlm/rcom.c  	if (!(le32_to_cpu(rs->rs_flags) & DLM_RSF_NEED_SLOTS)) {
rs                353 fs/f2fs/segment.h 					struct f2fs_sit_entry *rs)
rs                355 fs/f2fs/segment.h 	se->valid_blocks = GET_SIT_VBLOCKS(rs);
rs                356 fs/f2fs/segment.h 	se->ckpt_valid_blocks = GET_SIT_VBLOCKS(rs);
rs                357 fs/f2fs/segment.h 	memcpy(se->cur_valid_map, rs->valid_map, SIT_VBLOCK_MAP_SIZE);
rs                358 fs/f2fs/segment.h 	memcpy(se->ckpt_valid_map, rs->valid_map, SIT_VBLOCK_MAP_SIZE);
rs                360 fs/f2fs/segment.h 	memcpy(se->cur_valid_map_mir, rs->valid_map, SIT_VBLOCK_MAP_SIZE);
rs                362 fs/f2fs/segment.h 	se->type = GET_SIT_TYPE(rs);
rs                363 fs/f2fs/segment.h 	se->mtime = le64_to_cpu(rs->mtime);
rs                367 fs/f2fs/segment.h 					struct f2fs_sit_entry *rs)
rs                371 fs/f2fs/segment.h 	rs->vblocks = cpu_to_le16(raw_vblocks);
rs                372 fs/f2fs/segment.h 	memcpy(rs->valid_map, se->cur_valid_map, SIT_VBLOCK_MAP_SIZE);
rs                373 fs/f2fs/segment.h 	rs->mtime = cpu_to_le64(se->mtime);
rs                381 fs/f2fs/segment.h 	struct f2fs_sit_entry *rs;
rs                389 fs/f2fs/segment.h 		rs = &raw_sit->entries[i];
rs                391 fs/f2fs/segment.h 		__seg_info_to_raw_sit(se, rs);
rs                396 fs/f2fs/segment.h 					struct f2fs_sit_entry *rs)
rs                398 fs/f2fs/segment.h 	__seg_info_to_raw_sit(se, rs);
rs                400 fs/f2fs/segment.h 	memcpy(se->ckpt_valid_map, rs->valid_map, SIT_VBLOCK_MAP_SIZE);
rs                196 fs/gfs2/rgrp.c static inline int rs_cmp(u64 blk, u32 len, struct gfs2_blkreserv *rs)
rs                198 fs/gfs2/rgrp.c 	u64 startblk = gfs2_rbm_to_block(&rs->rs_rbm);
rs                200 fs/gfs2/rgrp.c 	if (blk >= startblk + rs->rs_free)
rs                613 fs/gfs2/rgrp.c static void dump_rs(struct seq_file *seq, const struct gfs2_blkreserv *rs,
rs                616 fs/gfs2/rgrp.c 	struct gfs2_inode *ip = container_of(rs, struct gfs2_inode, i_res);
rs                620 fs/gfs2/rgrp.c 		       (unsigned long long)gfs2_rbm_to_block(&rs->rs_rbm),
rs                621 fs/gfs2/rgrp.c 		       rs->rs_rbm.offset, rs->rs_free);
rs                629 fs/gfs2/rgrp.c static void __rs_deltree(struct gfs2_blkreserv *rs)
rs                633 fs/gfs2/rgrp.c 	if (!gfs2_rs_active(rs))
rs                636 fs/gfs2/rgrp.c 	rgd = rs->rs_rbm.rgd;
rs                637 fs/gfs2/rgrp.c 	trace_gfs2_rs(rs, TRACE_RS_TREEDEL);
rs                638 fs/gfs2/rgrp.c 	rb_erase(&rs->rs_node, &rgd->rd_rstree);
rs                639 fs/gfs2/rgrp.c 	RB_CLEAR_NODE(&rs->rs_node);
rs                641 fs/gfs2/rgrp.c 	if (rs->rs_free) {
rs                642 fs/gfs2/rgrp.c 		u64 last_block = gfs2_rbm_to_block(&rs->rs_rbm) +
rs                643 fs/gfs2/rgrp.c 				 rs->rs_free - 1;
rs                644 fs/gfs2/rgrp.c 		struct gfs2_rbm last_rbm = { .rgd = rs->rs_rbm.rgd, };
rs                648 fs/gfs2/rgrp.c 		BUG_ON(rs->rs_rbm.rgd->rd_reserved < rs->rs_free);
rs                649 fs/gfs2/rgrp.c 		rs->rs_rbm.rgd->rd_reserved -= rs->rs_free;
rs                654 fs/gfs2/rgrp.c 		rgd->rd_extfail_pt += rs->rs_free;
rs                655 fs/gfs2/rgrp.c 		rs->rs_free = 0;
rs                658 fs/gfs2/rgrp.c 		start = rbm_bi(&rs->rs_rbm);
rs                671 fs/gfs2/rgrp.c void gfs2_rs_deltree(struct gfs2_blkreserv *rs)
rs                675 fs/gfs2/rgrp.c 	rgd = rs->rs_rbm.rgd;
rs                678 fs/gfs2/rgrp.c 		__rs_deltree(rs);
rs                679 fs/gfs2/rgrp.c 		BUG_ON(rs->rs_free);
rs                710 fs/gfs2/rgrp.c 	struct gfs2_blkreserv *rs;
rs                714 fs/gfs2/rgrp.c 		rs = rb_entry(n, struct gfs2_blkreserv, rs_node);
rs                715 fs/gfs2/rgrp.c 		__rs_deltree(rs);
rs               1493 fs/gfs2/rgrp.c 	struct gfs2_blkreserv *rs = &ip->i_res;
rs               1494 fs/gfs2/rgrp.c 	struct gfs2_rgrpd *rgd = rs->rs_rbm.rgd;
rs               1495 fs/gfs2/rgrp.c 	u64 fsblock = gfs2_rbm_to_block(&rs->rs_rbm);
rs               1497 fs/gfs2/rgrp.c 	BUG_ON(gfs2_rs_active(rs));
rs               1506 fs/gfs2/rgrp.c 		rc = rs_cmp(fsblock, rs->rs_free, cur);
rs               1518 fs/gfs2/rgrp.c 	rb_link_node(&rs->rs_node, parent, newn);
rs               1519 fs/gfs2/rgrp.c 	rb_insert_color(&rs->rs_node, &rgd->rd_rstree);
rs               1522 fs/gfs2/rgrp.c 	rgd->rd_reserved += rs->rs_free; /* blocks reserved */
rs               1524 fs/gfs2/rgrp.c 	trace_gfs2_rs(rs, TRACE_RS_INSERT);
rs               1539 fs/gfs2/rgrp.c static inline u32 rgd_free(struct gfs2_rgrpd *rgd, struct gfs2_blkreserv *rs)
rs               1543 fs/gfs2/rgrp.c 	if (WARN_ON_ONCE(rgd->rd_reserved < rs->rs_free))
rs               1545 fs/gfs2/rgrp.c 	tot_reserved = rgd->rd_reserved - rs->rs_free;
rs               1568 fs/gfs2/rgrp.c 	struct gfs2_blkreserv *rs = &ip->i_res;
rs               1570 fs/gfs2/rgrp.c 	u32 free_blocks = rgd_free(rgd, rs);
rs               1594 fs/gfs2/rgrp.c 		rs->rs_rbm = rbm;
rs               1595 fs/gfs2/rgrp.c 		rs->rs_free = extlen;
rs               1620 fs/gfs2/rgrp.c 	struct gfs2_blkreserv *rs;
rs               1627 fs/gfs2/rgrp.c 		rs = rb_entry(n, struct gfs2_blkreserv, rs_node);
rs               1628 fs/gfs2/rgrp.c 		rc = rs_cmp(block, length, rs);
rs               1638 fs/gfs2/rgrp.c 		while ((rs_cmp(block, length, rs) == 0) && (&ip->i_res != rs)) {
rs               1639 fs/gfs2/rgrp.c 			block = gfs2_rbm_to_block(&rs->rs_rbm) + rs->rs_free;
rs               1643 fs/gfs2/rgrp.c 			rs = rb_entry(n, struct gfs2_blkreserv, rs_node);
rs               1970 fs/gfs2/rgrp.c static bool gfs2_rgrp_used_recently(const struct gfs2_blkreserv *rs,
rs               1976 fs/gfs2/rgrp.c                             rs->rs_rbm.rgd->rd_gl->gl_dstamp));
rs               2043 fs/gfs2/rgrp.c 	struct gfs2_blkreserv *rs = &ip->i_res;
rs               2053 fs/gfs2/rgrp.c 	if (gfs2_rs_active(rs)) {
rs               2054 fs/gfs2/rgrp.c 		begin = rs->rs_rbm.rgd;
rs               2055 fs/gfs2/rgrp.c 	} else if (rs->rs_rbm.rgd &&
rs               2056 fs/gfs2/rgrp.c 		   rgrp_contains_block(rs->rs_rbm.rgd, ip->i_goal)) {
rs               2057 fs/gfs2/rgrp.c 		begin = rs->rs_rbm.rgd;
rs               2060 fs/gfs2/rgrp.c 		rs->rs_rbm.rgd = begin = gfs2_blk2rgrpd(sdp, ip->i_goal, 1);
rs               2064 fs/gfs2/rgrp.c 	if (rs->rs_rbm.rgd == NULL)
rs               2070 fs/gfs2/rgrp.c 		if (!gfs2_glock_is_locked_by_me(rs->rs_rbm.rgd->rd_gl)) {
rs               2074 fs/gfs2/rgrp.c 			if (!gfs2_rs_active(rs)) {
rs               2076 fs/gfs2/rgrp.c 				    !fast_to_acquire(rs->rs_rbm.rgd))
rs               2079 fs/gfs2/rgrp.c 				    gfs2_rgrp_used_recently(rs, 1000) &&
rs               2080 fs/gfs2/rgrp.c 				    gfs2_rgrp_congested(rs->rs_rbm.rgd, loops))
rs               2083 fs/gfs2/rgrp.c 			error = gfs2_glock_nq_init(rs->rs_rbm.rgd->rd_gl,
rs               2088 fs/gfs2/rgrp.c 			if (!gfs2_rs_active(rs) && (loops < 2) &&
rs               2089 fs/gfs2/rgrp.c 			    gfs2_rgrp_congested(rs->rs_rbm.rgd, loops))
rs               2092 fs/gfs2/rgrp.c 				error = update_rgrp_lvb(rs->rs_rbm.rgd);
rs               2101 fs/gfs2/rgrp.c 		if ((rs->rs_rbm.rgd->rd_flags & (GFS2_RGF_NOALLOC |
rs               2103 fs/gfs2/rgrp.c 		    (loops == 0 && ap->target > rs->rs_rbm.rgd->rd_extfail_pt))
rs               2107 fs/gfs2/rgrp.c 			gfs2_rgrp_bh_get(rs->rs_rbm.rgd);
rs               2110 fs/gfs2/rgrp.c 		if (!gfs2_rs_active(rs))
rs               2111 fs/gfs2/rgrp.c 			rg_mblk_search(rs->rs_rbm.rgd, ip, ap);
rs               2114 fs/gfs2/rgrp.c 		if (!gfs2_rs_active(rs) && (loops < 1))
rs               2118 fs/gfs2/rgrp.c 		free_blocks = rgd_free(rs->rs_rbm.rgd, rs);
rs               2127 fs/gfs2/rgrp.c 		if (rs->rs_rbm.rgd->rd_flags & GFS2_RDF_CHECK)
rs               2128 fs/gfs2/rgrp.c 			try_rgrp_unlink(rs->rs_rbm.rgd, &last_unlinked,
rs               2132 fs/gfs2/rgrp.c 		if (gfs2_rs_active(rs))
rs               2133 fs/gfs2/rgrp.c 			gfs2_rs_deltree(rs);
rs               2140 fs/gfs2/rgrp.c 		if (gfs2_select_rgrp(&rs->rs_rbm.rgd, begin))
rs               2312 fs/gfs2/rgrp.c 	struct gfs2_blkreserv *rs = &ip->i_res;
rs               2319 fs/gfs2/rgrp.c 	if (gfs2_rs_active(rs)) {
rs               2320 fs/gfs2/rgrp.c 		if (gfs2_rbm_eq(&rs->rs_rbm, rbm)) {
rs               2322 fs/gfs2/rgrp.c 			ret = gfs2_rbm_from_block(&rs->rs_rbm, block + len);
rs               2323 fs/gfs2/rgrp.c 			rlen = min(rs->rs_free, len);
rs               2324 fs/gfs2/rgrp.c 			rs->rs_free -= rlen;
rs               2326 fs/gfs2/rgrp.c 			trace_gfs2_rs(rs, TRACE_RS_CLAIM);
rs               2327 fs/gfs2/rgrp.c 			if (rs->rs_free && !ret)
rs               2333 fs/gfs2/rgrp.c 		__rs_deltree(rs);
rs                 49 fs/gfs2/rgrp.h extern void gfs2_rs_deltree(struct gfs2_blkreserv *rs);
rs                 80 fs/gfs2/rgrp.h static inline bool gfs2_rs_active(const struct gfs2_blkreserv *rs)
rs                 82 fs/gfs2/rgrp.h 	return rs && !RB_EMPTY_NODE(&rs->rs_node);
rs                589 fs/gfs2/trace_gfs2.h 	TP_PROTO(const struct gfs2_blkreserv *rs, u8 func),
rs                591 fs/gfs2/trace_gfs2.h 	TP_ARGS(rs, func),
rs                605 fs/gfs2/trace_gfs2.h 		__entry->dev		= rs->rs_rbm.rgd->rd_sbd->sd_vfs->s_dev;
rs                606 fs/gfs2/trace_gfs2.h 		__entry->rd_addr	= rs->rs_rbm.rgd->rd_addr;
rs                607 fs/gfs2/trace_gfs2.h 		__entry->rd_free_clone	= rs->rs_rbm.rgd->rd_free_clone;
rs                608 fs/gfs2/trace_gfs2.h 		__entry->rd_reserved	= rs->rs_rbm.rgd->rd_reserved;
rs                609 fs/gfs2/trace_gfs2.h 		__entry->inum		= container_of(rs, struct gfs2_inode,
rs                611 fs/gfs2/trace_gfs2.h 		__entry->start		= gfs2_rbm_to_block(&rs->rs_rbm);
rs                612 fs/gfs2/trace_gfs2.h 		__entry->free		= rs->rs_free;
rs                 54 fs/isofs/rock.c 			struct inode *inode, struct rock_state *rs)
rs                 56 fs/isofs/rock.c 	rs->len = sizeof(struct iso_directory_record) + de->name_len[0];
rs                 57 fs/isofs/rock.c 	if (rs->len & 1)
rs                 58 fs/isofs/rock.c 		(rs->len)++;
rs                 59 fs/isofs/rock.c 	rs->chr = (unsigned char *)de + rs->len;
rs                 60 fs/isofs/rock.c 	rs->len = *((unsigned char *)de) - rs->len;
rs                 61 fs/isofs/rock.c 	if (rs->len < 0)
rs                 62 fs/isofs/rock.c 		rs->len = 0;
rs                 65 fs/isofs/rock.c 		rs->len -= ISOFS_SB(inode->i_sb)->s_rock_offset;
rs                 66 fs/isofs/rock.c 		rs->chr += ISOFS_SB(inode->i_sb)->s_rock_offset;
rs                 67 fs/isofs/rock.c 		if (rs->len < 0)
rs                 68 fs/isofs/rock.c 			rs->len = 0;
rs                 72 fs/isofs/rock.c static void init_rock_state(struct rock_state *rs, struct inode *inode)
rs                 74 fs/isofs/rock.c 	memset(rs, 0, sizeof(*rs));
rs                 75 fs/isofs/rock.c 	rs->inode = inode;
rs                 85 fs/isofs/rock.c static int rock_continue(struct rock_state *rs)
rs                 88 fs/isofs/rock.c 	int blocksize = 1 << rs->inode->i_blkbits;
rs                 91 fs/isofs/rock.c 	kfree(rs->buffer);
rs                 92 fs/isofs/rock.c 	rs->buffer = NULL;
rs                 94 fs/isofs/rock.c 	if ((unsigned)rs->cont_offset > blocksize - min_de_size ||
rs                 95 fs/isofs/rock.c 	    (unsigned)rs->cont_size > blocksize ||
rs                 96 fs/isofs/rock.c 	    (unsigned)(rs->cont_offset + rs->cont_size) > blocksize) {
rs                 99 fs/isofs/rock.c 			rs->cont_extent, rs->cont_offset, rs->cont_size);
rs                104 fs/isofs/rock.c 	if (rs->cont_extent) {
rs                107 fs/isofs/rock.c 		rs->buffer = kmalloc(rs->cont_size, GFP_KERNEL);
rs                108 fs/isofs/rock.c 		if (!rs->buffer) {
rs                113 fs/isofs/rock.c 		if (++rs->cont_loops >= RR_MAX_CE_ENTRIES)
rs                115 fs/isofs/rock.c 		bh = sb_bread(rs->inode->i_sb, rs->cont_extent);
rs                117 fs/isofs/rock.c 			memcpy(rs->buffer, bh->b_data + rs->cont_offset,
rs                118 fs/isofs/rock.c 					rs->cont_size);
rs                120 fs/isofs/rock.c 			rs->chr = rs->buffer;
rs                121 fs/isofs/rock.c 			rs->len = rs->cont_size;
rs                122 fs/isofs/rock.c 			rs->cont_extent = 0;
rs                123 fs/isofs/rock.c 			rs->cont_size = 0;
rs                124 fs/isofs/rock.c 			rs->cont_offset = 0;
rs                130 fs/isofs/rock.c 	kfree(rs->buffer);
rs                131 fs/isofs/rock.c 	rs->buffer = NULL;
rs                139 fs/isofs/rock.c static int rock_check_overflow(struct rock_state *rs, int sig)
rs                185 fs/isofs/rock.c 	if (len > rs->len) {
rs                189 fs/isofs/rock.c 				sig, len, rs->len);
rs                201 fs/isofs/rock.c 	struct rock_state rs;
rs                214 fs/isofs/rock.c 	init_rock_state(&rs, inode);
rs                215 fs/isofs/rock.c 	setup_rock_ridge(de, inode, &rs);
rs                218 fs/isofs/rock.c 	while (rs.len > 2) { /* There may be one byte for padding somewhere */
rs                219 fs/isofs/rock.c 		rr = (struct rock_ridge *)rs.chr;
rs                227 fs/isofs/rock.c 		sig = isonum_721(rs.chr);
rs                228 fs/isofs/rock.c 		if (rock_check_overflow(&rs, sig))
rs                230 fs/isofs/rock.c 		rs.chr += rr->len;
rs                231 fs/isofs/rock.c 		rs.len -= rr->len;
rs                236 fs/isofs/rock.c 		if (rs.len < 0)
rs                249 fs/isofs/rock.c 			rs.cont_extent = isonum_733(rr->u.CE.extent);
rs                250 fs/isofs/rock.c 			rs.cont_offset = isonum_733(rr->u.CE.offset);
rs                251 fs/isofs/rock.c 			rs.cont_size = isonum_733(rr->u.CE.size);
rs                286 fs/isofs/rock.c 			kfree(rs.buffer);
rs                292 fs/isofs/rock.c 	ret = rock_continue(&rs);
rs                298 fs/isofs/rock.c 	kfree(rs.buffer);
rs                318 fs/isofs/rock.c 	struct rock_state rs;
rs                324 fs/isofs/rock.c 	init_rock_state(&rs, inode);
rs                325 fs/isofs/rock.c 	setup_rock_ridge(de, inode, &rs);
rs                327 fs/isofs/rock.c 		rs.chr += 14;
rs                328 fs/isofs/rock.c 		rs.len -= 14;
rs                329 fs/isofs/rock.c 		if (rs.len < 0)
rs                330 fs/isofs/rock.c 			rs.len = 0;
rs                334 fs/isofs/rock.c 	while (rs.len > 2) { /* There may be one byte for padding somewhere */
rs                335 fs/isofs/rock.c 		rr = (struct rock_ridge *)rs.chr;
rs                343 fs/isofs/rock.c 		sig = isonum_721(rs.chr);
rs                344 fs/isofs/rock.c 		if (rock_check_overflow(&rs, sig))
rs                346 fs/isofs/rock.c 		rs.chr += rr->len;
rs                347 fs/isofs/rock.c 		rs.len -= rr->len;
rs                352 fs/isofs/rock.c 		if (rs.len < 0)
rs                368 fs/isofs/rock.c 			rs.cont_extent = isonum_733(rr->u.CE.extent);
rs                369 fs/isofs/rock.c 			rs.cont_offset = isonum_733(rr->u.CE.offset);
rs                370 fs/isofs/rock.c 			rs.cont_size = isonum_733(rr->u.CE.size);
rs                590 fs/isofs/rock.c 	ret = rock_continue(&rs);
rs                596 fs/isofs/rock.c 	kfree(rs.buffer);
rs                707 fs/isofs/rock.c 	struct rock_state rs;
rs                713 fs/isofs/rock.c 	init_rock_state(&rs, inode);
rs                735 fs/isofs/rock.c 	setup_rock_ridge(raw_de, inode, &rs);
rs                738 fs/isofs/rock.c 	while (rs.len > 2) { /* There may be one byte for padding somewhere */
rs                739 fs/isofs/rock.c 		rr = (struct rock_ridge *)rs.chr;
rs                742 fs/isofs/rock.c 		sig = isonum_721(rs.chr);
rs                743 fs/isofs/rock.c 		if (rock_check_overflow(&rs, sig))
rs                745 fs/isofs/rock.c 		rs.chr += rr->len;
rs                746 fs/isofs/rock.c 		rs.len -= rr->len;
rs                747 fs/isofs/rock.c 		if (rs.len < 0)
rs                767 fs/isofs/rock.c 			rs.cont_extent = isonum_733(rr->u.CE.extent);
rs                768 fs/isofs/rock.c 			rs.cont_offset = isonum_733(rr->u.CE.offset);
rs                769 fs/isofs/rock.c 			rs.cont_size = isonum_733(rr->u.CE.size);
rs                774 fs/isofs/rock.c 	ret = rock_continue(&rs);
rs                790 fs/isofs/rock.c 	kfree(rs.buffer);
rs                 88 fs/jffs2/compr_rubin.c static void init_rubin(struct rubin_state *rs, int div, int *bits)
rs                 92 fs/jffs2/compr_rubin.c 	rs->q = 0;
rs                 93 fs/jffs2/compr_rubin.c 	rs->p = (long) (2 * UPPER_BIT_RUBIN);
rs                 94 fs/jffs2/compr_rubin.c 	rs->bit_number = (long) 0;
rs                 95 fs/jffs2/compr_rubin.c 	rs->bit_divider = div;
rs                 98 fs/jffs2/compr_rubin.c 		rs->bits[c] = bits[c];
rs                102 fs/jffs2/compr_rubin.c static int encode(struct rubin_state *rs, long A, long B, int symbol)
rs                108 fs/jffs2/compr_rubin.c 	while ((rs->q >= UPPER_BIT_RUBIN) ||
rs                109 fs/jffs2/compr_rubin.c 	       ((rs->p + rs->q) <= UPPER_BIT_RUBIN)) {
rs                110 fs/jffs2/compr_rubin.c 		rs->bit_number++;
rs                112 fs/jffs2/compr_rubin.c 		ret = pushbit(&rs->pp, (rs->q & UPPER_BIT_RUBIN) ? 1 : 0, 0);
rs                115 fs/jffs2/compr_rubin.c 		rs->q &= LOWER_BITS_RUBIN;
rs                116 fs/jffs2/compr_rubin.c 		rs->q <<= 1;
rs                117 fs/jffs2/compr_rubin.c 		rs->p <<= 1;
rs                119 fs/jffs2/compr_rubin.c 	i0 = A * rs->p / (A + B);
rs                123 fs/jffs2/compr_rubin.c 	if (i0 >= rs->p)
rs                124 fs/jffs2/compr_rubin.c 		i0 = rs->p - 1;
rs                126 fs/jffs2/compr_rubin.c 	i1 = rs->p - i0;
rs                129 fs/jffs2/compr_rubin.c 		rs->p = i0;
rs                131 fs/jffs2/compr_rubin.c 		rs->p = i1;
rs                132 fs/jffs2/compr_rubin.c 		rs->q += i0;
rs                138 fs/jffs2/compr_rubin.c static void end_rubin(struct rubin_state *rs)
rs                144 fs/jffs2/compr_rubin.c 		pushbit(&rs->pp, (UPPER_BIT_RUBIN & rs->q) ? 1 : 0, 1);
rs                145 fs/jffs2/compr_rubin.c 		rs->q &= LOWER_BITS_RUBIN;
rs                146 fs/jffs2/compr_rubin.c 		rs->q <<= 1;
rs                151 fs/jffs2/compr_rubin.c static void init_decode(struct rubin_state *rs, int div, int *bits)
rs                153 fs/jffs2/compr_rubin.c 	init_rubin(rs, div, bits);
rs                156 fs/jffs2/compr_rubin.c 	rs->rec_q = 0;
rs                158 fs/jffs2/compr_rubin.c 	for (rs->bit_number = 0; rs->bit_number++ < RUBIN_REG_SIZE;
rs                159 fs/jffs2/compr_rubin.c 	     rs->rec_q = rs->rec_q * 2 + (long) (pullbit(&rs->pp)))
rs                163 fs/jffs2/compr_rubin.c static void __do_decode(struct rubin_state *rs, unsigned long p,
rs                182 fs/jffs2/compr_rubin.c 	rs->p = p;
rs                183 fs/jffs2/compr_rubin.c 	rs->q = q;
rs                185 fs/jffs2/compr_rubin.c 	rs->bit_number += bits;
rs                190 fs/jffs2/compr_rubin.c 	rec_q = rs->rec_q;
rs                192 fs/jffs2/compr_rubin.c 		c = pullbit(&rs->pp);
rs                197 fs/jffs2/compr_rubin.c 	rs->rec_q = rec_q;
rs                200 fs/jffs2/compr_rubin.c static int decode(struct rubin_state *rs, long A, long B)
rs                202 fs/jffs2/compr_rubin.c 	unsigned long p = rs->p, q = rs->q;
rs                207 fs/jffs2/compr_rubin.c 		__do_decode(rs, p, q);
rs                209 fs/jffs2/compr_rubin.c 	i0 = A * rs->p / (A + B);
rs                213 fs/jffs2/compr_rubin.c 	if (i0 >= rs->p)
rs                214 fs/jffs2/compr_rubin.c 		i0 = rs->p - 1;
rs                216 fs/jffs2/compr_rubin.c 	threshold = rs->q + i0;
rs                217 fs/jffs2/compr_rubin.c 	symbol = rs->rec_q >= threshold;
rs                218 fs/jffs2/compr_rubin.c 	if (rs->rec_q >= threshold) {
rs                219 fs/jffs2/compr_rubin.c 		rs->q += i0;
rs                220 fs/jffs2/compr_rubin.c 		i0 = rs->p - i0;
rs                223 fs/jffs2/compr_rubin.c 	rs->p = i0;
rs                230 fs/jffs2/compr_rubin.c static int out_byte(struct rubin_state *rs, unsigned char byte)
rs                234 fs/jffs2/compr_rubin.c 	rs_copy = *rs;
rs                237 fs/jffs2/compr_rubin.c 		ret = encode(rs, rs->bit_divider-rs->bits[i],
rs                238 fs/jffs2/compr_rubin.c 			     rs->bits[i], byte & 1);
rs                241 fs/jffs2/compr_rubin.c 			*rs = rs_copy;
rs                249 fs/jffs2/compr_rubin.c static int in_byte(struct rubin_state *rs)
rs                251 fs/jffs2/compr_rubin.c 	int i, result = 0, bit_divider = rs->bit_divider;
rs                254 fs/jffs2/compr_rubin.c 		result |= decode(rs, bit_divider - rs->bits[i],
rs                255 fs/jffs2/compr_rubin.c 				 rs->bits[i]) << i;
rs                268 fs/jffs2/compr_rubin.c 	struct rubin_state rs;
rs                270 fs/jffs2/compr_rubin.c 	init_pushpull(&rs.pp, cpage_out, *dstlen * 8, 0, 32);
rs                272 fs/jffs2/compr_rubin.c 	init_rubin(&rs, bit_divider, bits);
rs                274 fs/jffs2/compr_rubin.c 	while (pos < (*sourcelen) && !out_byte(&rs, data_in[pos]))
rs                277 fs/jffs2/compr_rubin.c 	end_rubin(&rs);
rs                287 fs/jffs2/compr_rubin.c 	outpos = (pushedbits(&rs.pp)+7)/8;
rs                374 fs/jffs2/compr_rubin.c 	struct rubin_state rs;
rs                376 fs/jffs2/compr_rubin.c 	init_pushpull(&rs.pp, cdata_in, srclen, 0, 0);
rs                377 fs/jffs2/compr_rubin.c 	init_decode(&rs, bit_divider, bits);
rs                380 fs/jffs2/compr_rubin.c 		page_out[outpos++] = in_byte(&rs);
rs                421 fs/reiserfs/bitmap.c 	struct reiserfs_super_block *rs;
rs                428 fs/reiserfs/bitmap.c 	rs = SB_DISK_SUPER_BLOCK(s);
rs                457 fs/reiserfs/bitmap.c 	set_sb_free_blocks(rs, sb_free_blocks(rs) + 1);
rs               2717 fs/reiserfs/journal.c 	struct reiserfs_super_block *rs;
rs               2772 fs/reiserfs/journal.c 	rs = SB_DISK_SUPER_BLOCK(sb);
rs               2786 fs/reiserfs/journal.c 	if (is_reiserfs_jr(rs)
rs               2788 fs/reiserfs/journal.c 		sb_jp_journal_magic(rs))) {
rs               2794 fs/reiserfs/journal.c 				 sb_jp_journal_magic(rs));
rs                 11 fs/reiserfs/objectid.c #define objectid_map(s,rs) (old_format_only (s) ? \
rs                 12 fs/reiserfs/objectid.c                          (__le32 *)((struct reiserfs_super_block_v1 *)(rs) + 1) :\
rs                 13 fs/reiserfs/objectid.c 			 (__le32 *)((rs) + 1))
rs                 52 fs/reiserfs/objectid.c 	struct reiserfs_super_block *rs = SB_DISK_SUPER_BLOCK(s);
rs                 53 fs/reiserfs/objectid.c 	__le32 *map = objectid_map(s, rs);
rs                 86 fs/reiserfs/objectid.c 	if (sb_oid_cursize(rs) > 2 && map[1] == map[2]) {
rs                 88 fs/reiserfs/objectid.c 			(sb_oid_cursize(rs) - 3) * sizeof(__u32));
rs                 89 fs/reiserfs/objectid.c 		set_sb_oid_cursize(rs, sb_oid_cursize(rs) - 2);
rs                101 fs/reiserfs/objectid.c 	struct reiserfs_super_block *rs = SB_DISK_SUPER_BLOCK(s);
rs                102 fs/reiserfs/objectid.c 	__le32 *map = objectid_map(s, rs);
rs                119 fs/reiserfs/objectid.c 	while (i < sb_oid_cursize(rs)) {
rs                131 fs/reiserfs/objectid.c 					(sb_oid_cursize(rs) - i -
rs                133 fs/reiserfs/objectid.c 				set_sb_oid_cursize(rs, sb_oid_cursize(rs) - 2);
rs                135 fs/reiserfs/objectid.c 				RFALSE(sb_oid_cursize(rs) < 2 ||
rs                136 fs/reiserfs/objectid.c 				       sb_oid_cursize(rs) > sb_oid_maxsize(rs),
rs                138 fs/reiserfs/objectid.c 				       sb_oid_cursize(rs), sb_oid_maxsize(rs));
rs                159 fs/reiserfs/objectid.c 			if (sb_oid_cursize(rs) == sb_oid_maxsize(rs)) {
rs                166 fs/reiserfs/objectid.c 				(sb_oid_cursize(rs) - i - 1) * sizeof(__u32));
rs                169 fs/reiserfs/objectid.c 			set_sb_oid_cursize(rs, sb_oid_cursize(rs) + 2);
rs                548 fs/reiserfs/prints.c 	struct reiserfs_super_block *rs =
rs                553 fs/reiserfs/prints.c 	if (is_reiserfs_3_5(rs)) {
rs                555 fs/reiserfs/prints.c 	} else if (is_reiserfs_3_6(rs)) {
rs                557 fs/reiserfs/prints.c 	} else if (is_reiserfs_jr(rs)) {
rs                558 fs/reiserfs/prints.c 		version = ((sb_version(rs) == REISERFS_VERSION_2) ?
rs                567 fs/reiserfs/prints.c 	printk("Block count %u\n", sb_block_count(rs));
rs                568 fs/reiserfs/prints.c 	printk("Blocksize %d\n", sb_blocksize(rs));
rs                569 fs/reiserfs/prints.c 	printk("Free blocks %u\n", sb_free_blocks(rs));
rs                576 fs/reiserfs/prints.c 	data_blocks = sb_block_count(rs) - skipped - 1 - sb_bmap_nr(rs) -
rs                577 fs/reiserfs/prints.c 	    (!is_reiserfs_jr(rs) ? sb_jp_journal_size(rs) +
rs                578 fs/reiserfs/prints.c 	     1 : sb_reserved_for_journal(rs)) - sb_free_blocks(rs);
rs                581 fs/reiserfs/prints.c 	     "1 super block, %d data blocks\n", skipped, sb_bmap_nr(rs),
rs                582 fs/reiserfs/prints.c 	     (!is_reiserfs_jr(rs) ? (sb_jp_journal_size(rs) + 1) :
rs                583 fs/reiserfs/prints.c 	      sb_reserved_for_journal(rs)), data_blocks);
rs                584 fs/reiserfs/prints.c 	printk("Root block %u\n", sb_root_block(rs));
rs                585 fs/reiserfs/prints.c 	printk("Journal block (first) %d\n", sb_jp_journal_1st_block(rs));
rs                586 fs/reiserfs/prints.c 	printk("Journal dev %d\n", sb_jp_journal_dev(rs));
rs                587 fs/reiserfs/prints.c 	printk("Journal orig size %d\n", sb_jp_journal_size(rs));
rs                588 fs/reiserfs/prints.c 	printk("FS state %d\n", sb_fs_state(rs));
rs                590 fs/reiserfs/prints.c 	       reiserfs_hashname(sb_hash_function_code(rs)));
rs                592 fs/reiserfs/prints.c 	printk("Tree height %d\n", sb_tree_height(rs));
rs                 57 fs/reiserfs/procfs.c #define DF( x ) D2C( rs -> s_v1.x )
rs                 58 fs/reiserfs/procfs.c #define DFL( x ) D4C( rs -> s_v1.x )
rs                 60 fs/reiserfs/procfs.c #define objectid_map( s, rs ) (old_format_only (s) ?				\
rs                 61 fs/reiserfs/procfs.c                          (__le32 *)((struct reiserfs_super_block_v1 *)rs + 1) :	\
rs                 62 fs/reiserfs/procfs.c 			 (__le32 *)(rs + 1))
rs                 63 fs/reiserfs/procfs.c #define MAP( i ) D4C( objectid_map( sb, rs )[ i ] )
rs                 65 fs/reiserfs/procfs.c #define DJF( x ) le32_to_cpu( rs -> x )
rs                229 fs/reiserfs/procfs.c 	struct reiserfs_super_block *rs = sb_info->s_rs;
rs                255 fs/reiserfs/procfs.c 		   rs->s_v1.s_magic,
rs                273 fs/reiserfs/procfs.c 	struct reiserfs_super_block *rs = sb_info->s_rs;
rs                274 fs/reiserfs/procfs.c 	unsigned int mapsize = le16_to_cpu(rs->s_v1.s_oid_cursize);
rs                296 fs/reiserfs/procfs.c 		   mapsize, le16_to_cpu(rs->s_v1.s_oid_maxsize), total_used);
rs                311 fs/reiserfs/procfs.c 	struct reiserfs_super_block *rs = r->s_rs;
rs                312 fs/reiserfs/procfs.c 	struct journal_params *jp = &rs->s_v1.s_journal;
rs               1104 fs/reiserfs/reiserfs.h int is_reiserfs_3_5(struct reiserfs_super_block *rs);
rs               1105 fs/reiserfs/reiserfs.h int is_reiserfs_3_6(struct reiserfs_super_block *rs);
rs               1106 fs/reiserfs/reiserfs.h int is_reiserfs_jr(struct reiserfs_super_block *rs);
rs                 40 fs/reiserfs/super.c int is_reiserfs_3_5(struct reiserfs_super_block *rs)
rs                 42 fs/reiserfs/super.c 	return !strncmp(rs->s_v1.s_magic, reiserfs_3_5_magic_string,
rs                 46 fs/reiserfs/super.c int is_reiserfs_3_6(struct reiserfs_super_block *rs)
rs                 48 fs/reiserfs/super.c 	return !strncmp(rs->s_v1.s_magic, reiserfs_3_6_magic_string,
rs                 52 fs/reiserfs/super.c int is_reiserfs_jr(struct reiserfs_super_block *rs)
rs                 54 fs/reiserfs/super.c 	return !strncmp(rs->s_v1.s_magic, reiserfs_jr_magic_string,
rs                 58 fs/reiserfs/super.c static int is_any_reiserfs_magic_string(struct reiserfs_super_block *rs)
rs                 60 fs/reiserfs/super.c 	return (is_reiserfs_3_5(rs) || is_reiserfs_3_6(rs) ||
rs                 61 fs/reiserfs/super.c 		is_reiserfs_jr(rs));
rs               1397 fs/reiserfs/super.c 	struct reiserfs_super_block *rs = SB_DISK_SUPER_BLOCK(s);
rs               1406 fs/reiserfs/super.c 		if (!(le32_to_cpu(rs->s_flags) & reiserfs_attrs_cleared)) {
rs               1433 fs/reiserfs/super.c 	struct reiserfs_super_block *rs;
rs               1459 fs/reiserfs/super.c 	rs = SB_DISK_SUPER_BLOCK(s);
rs               1532 fs/reiserfs/super.c 		if (sb_umount_state(rs) == REISERFS_VALID_FS
rs               1545 fs/reiserfs/super.c 		set_sb_umount_state(rs, REISERFS_SB(s)->s_mount_state);
rs               1562 fs/reiserfs/super.c 		REISERFS_SB(s)->s_mount_state = sb_umount_state(rs);
rs               1572 fs/reiserfs/super.c 		REISERFS_SB(s)->s_mount_state = sb_umount_state(rs);
rs               1574 fs/reiserfs/super.c 		set_sb_umount_state(rs, REISERFS_ERROR_FS);
rs               1576 fs/reiserfs/super.c 			set_sb_mnt_count(rs, sb_mnt_count(rs) + 1);
rs               1609 fs/reiserfs/super.c 	struct reiserfs_super_block *rs;
rs               1621 fs/reiserfs/super.c 	rs = (struct reiserfs_super_block *)bh->b_data;
rs               1622 fs/reiserfs/super.c 	if (!is_any_reiserfs_magic_string(rs)) {
rs               1629 fs/reiserfs/super.c 	fs_blocksize = sb_blocksize(rs);
rs               1642 fs/reiserfs/super.c 	rs = (struct reiserfs_super_block *)bh->b_data;
rs               1643 fs/reiserfs/super.c 	if (sb_blocksize(rs) != s->s_blocksize) {
rs               1653 fs/reiserfs/super.c 	if (rs->s_v1.s_root_block == cpu_to_le32(-1)) {
rs               1664 fs/reiserfs/super.c 	SB_DISK_SUPER_BLOCK(s) = rs;
rs               1670 fs/reiserfs/super.c 	if (is_reiserfs_jr(rs)) {
rs               1671 fs/reiserfs/super.c 		if (sb_version(rs) == REISERFS_VERSION_2)
rs               1674 fs/reiserfs/super.c 		else if (sb_version(rs) == REISERFS_VERSION_1)
rs               1680 fs/reiserfs/super.c 					 "non-standard magic", sb_version(rs));
rs               1690 fs/reiserfs/super.c 			      is_reiserfs_3_5(rs) ? "3.5" : "3.6");
rs               1905 fs/reiserfs/super.c 	struct reiserfs_super_block *rs;
rs               1983 fs/reiserfs/super.c 	rs = SB_DISK_SUPER_BLOCK(s);
rs               1991 fs/reiserfs/super.c 	    sb_block_count(rs) * sb_blocksize(rs)) {
rs               2096 fs/reiserfs/super.c 	if (is_reiserfs_3_5(rs)
rs               2097 fs/reiserfs/super.c 	    || (is_reiserfs_jr(rs) && SB_VERSION(s) == REISERFS_VERSION_1))
rs               2114 fs/reiserfs/super.c 		set_sb_umount_state(rs, REISERFS_ERROR_FS);
rs               2115 fs/reiserfs/super.c 		set_sb_fs_state(rs, 0);
rs               2124 fs/reiserfs/super.c 		    sb_bmap_nr(rs) != 0) {
rs               2128 fs/reiserfs/super.c 					"Clearing to zero.", sb_bmap_nr(rs),
rs               2131 fs/reiserfs/super.c 			set_sb_bmap_nr(rs, 0);
rs               2145 fs/reiserfs/super.c 				if (is_reiserfs_3_5(rs))
rs               2151 fs/reiserfs/super.c 					memcpy(rs->s_v1.s_magic,
rs               2156 fs/reiserfs/super.c 				set_sb_version(rs, REISERFS_VERSION_2);
rs               2164 fs/reiserfs/super.c 			set_sb_mnt_count(rs, sb_mnt_count(rs) + 1);
rs               2205 fs/reiserfs/super.c 	set_sb_hash_function_code(rs, function2code(sbi->s_hash_function));
rs               2253 fs/reiserfs/super.c 	struct reiserfs_super_block *rs = SB_DISK_SUPER_BLOCK(dentry->d_sb);
rs               2256 fs/reiserfs/super.c 	buf->f_bfree = sb_free_blocks(rs);
rs               2258 fs/reiserfs/super.c 	buf->f_blocks = sb_block_count(rs) - sb_bmap_nr(rs) - 1;
rs               2262 fs/reiserfs/super.c 	buf->f_fsid.val[0] = (u32)crc32_le(0, rs->s_uuid, sizeof(rs->s_uuid)/2);
rs               2263 fs/reiserfs/super.c 	buf->f_fsid.val[1] = (u32)crc32_le(0, rs->s_uuid + sizeof(rs->s_uuid)/2,
rs               2264 fs/reiserfs/super.c 				sizeof(rs->s_uuid)/2);
rs                 15 fs/verity/init.c 	static DEFINE_RATELIMIT_STATE(rs, DEFAULT_RATELIMIT_INTERVAL,
rs                 20 fs/verity/init.c 	if (!__ratelimit(&rs))
rs                 40 include/linux/ratelimit.h static inline void ratelimit_state_init(struct ratelimit_state *rs,
rs                 43 include/linux/ratelimit.h 	memset(rs, 0, sizeof(*rs));
rs                 45 include/linux/ratelimit.h 	raw_spin_lock_init(&rs->lock);
rs                 46 include/linux/ratelimit.h 	rs->interval	= interval;
rs                 47 include/linux/ratelimit.h 	rs->burst	= burst;
rs                 50 include/linux/ratelimit.h static inline void ratelimit_default_init(struct ratelimit_state *rs)
rs                 52 include/linux/ratelimit.h 	return ratelimit_state_init(rs, DEFAULT_RATELIMIT_INTERVAL,
rs                 56 include/linux/ratelimit.h static inline void ratelimit_state_exit(struct ratelimit_state *rs)
rs                 58 include/linux/ratelimit.h 	if (!(rs->flags & RATELIMIT_MSG_ON_RELEASE))
rs                 61 include/linux/ratelimit.h 	if (rs->missed) {
rs                 63 include/linux/ratelimit.h 			current->comm, rs->missed);
rs                 64 include/linux/ratelimit.h 		rs->missed = 0;
rs                 69 include/linux/ratelimit.h ratelimit_set_flags(struct ratelimit_state *rs, unsigned long flags)
rs                 71 include/linux/ratelimit.h 	rs->flags = flags;
rs                 76 include/linux/ratelimit.h extern int ___ratelimit(struct ratelimit_state *rs, const char *func);
rs                 62 include/linux/rslib.h int encode_rs8(struct rs_control *rs, uint8_t *data, int len, uint16_t *par,
rs                 66 include/linux/rslib.h int decode_rs8(struct rs_control *rs, uint8_t *data, uint16_t *par, int len,
rs                 73 include/linux/rslib.h int encode_rs16(struct rs_control *rs, uint16_t *data, int len, uint16_t *par,
rs                 77 include/linux/rslib.h int decode_rs16(struct rs_control *rs, uint16_t *data, uint16_t *par, int len,
rs                108 include/linux/rslib.h void free_rs(struct rs_control *rs);
rs                122 include/linux/rslib.h static inline int rs_modnn(struct rs_codec *rs, int x)
rs                124 include/linux/rslib.h 	while (x >= rs->nn) {
rs                125 include/linux/rslib.h 		x -= rs->nn;
rs                126 include/linux/rslib.h 		x = (x >> rs->mm) + (x & rs->nn);
rs                122 include/math-emu/double.h #define FP_FROM_INT_D(X,r,rs,rt)	_FP_FROM_INT(D,2,X,r,rs,rt)
rs                197 include/math-emu/double.h #define FP_FROM_INT_D(X,r,rs,rt)	_FP_FROM_INT(D,1,X,r,rs,rt)
rs                128 include/math-emu/quad.h #define FP_FROM_INT_Q(X,r,rs,rt)	_FP_FROM_INT(Q,4,X,r,rs,rt)
rs                201 include/math-emu/quad.h #define FP_FROM_INT_Q(X,r,rs,rt)	_FP_FROM_INT(Q,2,X,r,rs,rt)
rs                111 include/math-emu/single.h #define FP_FROM_INT_S(X,r,rs,rt)	_FP_FROM_INT(S,1,X,r,rs,rt)
rs               1068 include/net/tcp.h 	void (*cong_control)(struct sock *sk, const struct rate_sample *rs);
rs               1136 include/net/tcp.h 			    struct rate_sample *rs);
rs               1138 include/net/tcp.h 		  bool is_sack_reneg, struct rate_sample *rs);
rs               2035 include/net/tcp.h extern void tcp_rack_update_reo_wnd(struct sock *sk, struct rate_sample *rs);
rs                799 kernel/printk/printk.c 	struct ratelimit_state rs;
rs                836 kernel/printk/printk.c 		if (!___ratelimit(&user->rs, current->comm))
rs               1027 kernel/printk/printk.c 	ratelimit_default_init(&user->rs);
rs               1028 kernel/printk/printk.c 	ratelimit_set_flags(&user->rs, RATELIMIT_MSG_ON_RELEASE);
rs               1048 kernel/printk/printk.c 	ratelimit_state_exit(&user->rs);
rs                183 kernel/rcu/srcutiny.c 	struct rcu_synchronize rs;
rs                185 kernel/rcu/srcutiny.c 	init_rcu_head_on_stack(&rs.head);
rs                186 kernel/rcu/srcutiny.c 	init_completion(&rs.completion);
rs                187 kernel/rcu/srcutiny.c 	call_srcu(ssp, &rs.head, wakeme_after_rcu);
rs                188 kernel/rcu/srcutiny.c 	wait_for_completion(&rs.completion);
rs                189 kernel/rcu/srcutiny.c 	destroy_rcu_head_on_stack(&rs.head);
rs                 65 lib/crypto/des.c static const u8 rs[256] = {
rs                634 lib/crypto/des.c 	pe[15 * 2 + 0] = DES_PC2(a, b, c, d); d = rs[d];
rs                635 lib/crypto/des.c 	pe[14 * 2 + 0] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b];
rs                636 lib/crypto/des.c 	pe[13 * 2 + 0] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d];
rs                637 lib/crypto/des.c 	pe[12 * 2 + 0] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b];
rs                638 lib/crypto/des.c 	pe[11 * 2 + 0] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d];
rs                639 lib/crypto/des.c 	pe[10 * 2 + 0] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b];
rs                640 lib/crypto/des.c 	pe[ 9 * 2 + 0] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d];
rs                641 lib/crypto/des.c 	pe[ 8 * 2 + 0] = DES_PC2(d, a, b, c); c = rs[c];
rs                642 lib/crypto/des.c 	pe[ 7 * 2 + 0] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a];
rs                643 lib/crypto/des.c 	pe[ 6 * 2 + 0] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c];
rs                644 lib/crypto/des.c 	pe[ 5 * 2 + 0] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a];
rs                645 lib/crypto/des.c 	pe[ 4 * 2 + 0] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c];
rs                646 lib/crypto/des.c 	pe[ 3 * 2 + 0] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a];
rs                647 lib/crypto/des.c 	pe[ 2 * 2 + 0] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c];
rs                648 lib/crypto/des.c 	pe[ 1 * 2 + 0] = DES_PC2(c, d, a, b); b = rs[b];
rs                652 lib/crypto/des.c 	w  = (a ^ c) | (b ^ d) | (rs[a] ^ c) | (b ^ rs[d]);
rs                663 lib/crypto/des.c 	w |= (a ^ c) | (b ^ d) | (rs[a] ^ c) | (b ^ rs[d]);
rs                665 lib/crypto/des.c 	pe[15 * 2 + 1] = DES_PC2(a, b, c, d); d = rs[d];
rs                666 lib/crypto/des.c 	pe[14 * 2 + 1] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b];
rs                667 lib/crypto/des.c 	pe[13 * 2 + 1] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d];
rs                668 lib/crypto/des.c 	pe[12 * 2 + 1] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b];
rs                669 lib/crypto/des.c 	pe[11 * 2 + 1] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d];
rs                670 lib/crypto/des.c 	pe[10 * 2 + 1] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b];
rs                671 lib/crypto/des.c 	pe[ 9 * 2 + 1] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d];
rs                672 lib/crypto/des.c 	pe[ 8 * 2 + 1] = DES_PC2(d, a, b, c); c = rs[c];
rs                673 lib/crypto/des.c 	pe[ 7 * 2 + 1] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a];
rs                674 lib/crypto/des.c 	pe[ 6 * 2 + 1] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c];
rs                675 lib/crypto/des.c 	pe[ 5 * 2 + 1] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a];
rs                676 lib/crypto/des.c 	pe[ 4 * 2 + 1] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c];
rs                677 lib/crypto/des.c 	pe[ 3 * 2 + 1] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a];
rs                678 lib/crypto/des.c 	pe[ 2 * 2 + 1] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c];
rs                679 lib/crypto/des.c 	pe[ 1 * 2 + 1] = DES_PC2(c, d, a, b); b = rs[b];
rs                725 lib/crypto/des.c 	pe[ 0 * 2] = DES_PC2(a, b, c, d); d = rs[d];
rs                726 lib/crypto/des.c 	pe[ 1 * 2] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b];
rs                727 lib/crypto/des.c 	pe[ 2 * 2] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d];
rs                728 lib/crypto/des.c 	pe[ 3 * 2] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b];
rs                729 lib/crypto/des.c 	pe[ 4 * 2] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d];
rs                730 lib/crypto/des.c 	pe[ 5 * 2] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b];
rs                731 lib/crypto/des.c 	pe[ 6 * 2] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d];
rs                732 lib/crypto/des.c 	pe[ 7 * 2] = DES_PC2(d, a, b, c); c = rs[c];
rs                733 lib/crypto/des.c 	pe[ 8 * 2] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a];
rs                734 lib/crypto/des.c 	pe[ 9 * 2] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c];
rs                735 lib/crypto/des.c 	pe[10 * 2] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a];
rs                736 lib/crypto/des.c 	pe[11 * 2] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c];
rs                737 lib/crypto/des.c 	pe[12 * 2] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a];
rs                738 lib/crypto/des.c 	pe[13 * 2] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c];
rs                739 lib/crypto/des.c 	pe[14 * 2] = DES_PC2(c, d, a, b); b = rs[b];
rs                750 lib/crypto/des.c 	pe[ 0 * 2 + 1] = DES_PC2(a, b, c, d); d = rs[d];
rs                751 lib/crypto/des.c 	pe[ 1 * 2 + 1] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b];
rs                752 lib/crypto/des.c 	pe[ 2 * 2 + 1] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d];
rs                753 lib/crypto/des.c 	pe[ 3 * 2 + 1] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b];
rs                754 lib/crypto/des.c 	pe[ 4 * 2 + 1] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d];
rs                755 lib/crypto/des.c 	pe[ 5 * 2 + 1] = DES_PC2(d, a, b, c); c = rs[c]; b = rs[b];
rs                756 lib/crypto/des.c 	pe[ 6 * 2 + 1] = DES_PC2(b, c, d, a); a = rs[a]; d = rs[d];
rs                757 lib/crypto/des.c 	pe[ 7 * 2 + 1] = DES_PC2(d, a, b, c); c = rs[c];
rs                758 lib/crypto/des.c 	pe[ 8 * 2 + 1] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a];
rs                759 lib/crypto/des.c 	pe[ 9 * 2 + 1] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c];
rs                760 lib/crypto/des.c 	pe[10 * 2 + 1] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a];
rs                761 lib/crypto/des.c 	pe[11 * 2 + 1] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c];
rs                762 lib/crypto/des.c 	pe[12 * 2 + 1] = DES_PC2(c, d, a, b); b = rs[b]; a = rs[a];
rs                763 lib/crypto/des.c 	pe[13 * 2 + 1] = DES_PC2(a, b, c, d); d = rs[d]; c = rs[c];
rs                764 lib/crypto/des.c 	pe[14 * 2 + 1] = DES_PC2(c, d, a, b); b = rs[b];
rs                 27 lib/ratelimit.c int ___ratelimit(struct ratelimit_state *rs, const char *func)
rs                 32 lib/ratelimit.c 	if (!rs->interval)
rs                 41 lib/ratelimit.c 	if (!raw_spin_trylock_irqsave(&rs->lock, flags))
rs                 44 lib/ratelimit.c 	if (!rs->begin)
rs                 45 lib/ratelimit.c 		rs->begin = jiffies;
rs                 47 lib/ratelimit.c 	if (time_is_before_jiffies(rs->begin + rs->interval)) {
rs                 48 lib/ratelimit.c 		if (rs->missed) {
rs                 49 lib/ratelimit.c 			if (!(rs->flags & RATELIMIT_MSG_ON_RELEASE)) {
rs                 52 lib/ratelimit.c 						func, rs->missed);
rs                 53 lib/ratelimit.c 				rs->missed = 0;
rs                 56 lib/ratelimit.c 		rs->begin   = jiffies;
rs                 57 lib/ratelimit.c 		rs->printed = 0;
rs                 59 lib/ratelimit.c 	if (rs->burst && rs->burst > rs->printed) {
rs                 60 lib/ratelimit.c 		rs->printed++;
rs                 63 lib/ratelimit.c 		rs->missed++;
rs                 66 lib/ratelimit.c 	raw_spin_unlock_irqrestore(&rs->lock, flags);
rs                 13 lib/reed_solomon/decode_rs.c 	struct rs_codec *rs = rsc->codec;
rs                 16 lib/reed_solomon/decode_rs.c 	int nn = rs->nn;
rs                 17 lib/reed_solomon/decode_rs.c 	int nroots = rs->nroots;
rs                 18 lib/reed_solomon/decode_rs.c 	int fcr = rs->fcr;
rs                 19 lib/reed_solomon/decode_rs.c 	int prim = rs->prim;
rs                 20 lib/reed_solomon/decode_rs.c 	int iprim = rs->iprim;
rs                 21 lib/reed_solomon/decode_rs.c 	uint16_t *alpha_to = rs->alpha_to;
rs                 22 lib/reed_solomon/decode_rs.c 	uint16_t *index_of = rs->index_of;
rs                 26 lib/reed_solomon/decode_rs.c 	uint16_t msk = (uint16_t) rs->nn;
rs                 72 lib/reed_solomon/decode_rs.c 					alpha_to[rs_modnn(rs, index_of[syn[i]] +
rs                 84 lib/reed_solomon/decode_rs.c 					alpha_to[rs_modnn(rs, index_of[syn[i]] +
rs                111 lib/reed_solomon/decode_rs.c 		lambda[1] = alpha_to[rs_modnn(rs,
rs                114 lib/reed_solomon/decode_rs.c 			u = rs_modnn(rs, prim * (nn - 1 - (eras_pos[i] + pad)));
rs                119 lib/reed_solomon/decode_rs.c 						alpha_to[rs_modnn(rs, u + tmp)];
rs                140 lib/reed_solomon/decode_rs.c 					alpha_to[rs_modnn(rs,
rs                156 lib/reed_solomon/decode_rs.c 						alpha_to[rs_modnn(rs, discr_r +
rs                169 lib/reed_solomon/decode_rs.c 						rs_modnn(rs, index_of[lambda[i]]
rs                200 lib/reed_solomon/decode_rs.c 	for (i = 1, k = iprim - 1; i <= nn; i++, k = rs_modnn(rs, k + iprim)) {
rs                204 lib/reed_solomon/decode_rs.c 				reg[j] = rs_modnn(rs, reg[j] + j);
rs                242 lib/reed_solomon/decode_rs.c 				    alpha_to[rs_modnn(rs, s[i - j] + lambda[j])];
rs                257 lib/reed_solomon/decode_rs.c 				num1 ^= alpha_to[rs_modnn(rs, omega[i] +
rs                267 lib/reed_solomon/decode_rs.c 		num2 = alpha_to[rs_modnn(rs, root[j] * (fcr - 1) + nn)];
rs                274 lib/reed_solomon/decode_rs.c 				den ^= alpha_to[rs_modnn(rs, lambda[i + 1] +
rs                279 lib/reed_solomon/decode_rs.c 		b[j] = alpha_to[rs_modnn(rs, index_of[num1] +
rs                296 lib/reed_solomon/decode_rs.c 			tmp ^= alpha_to[rs_modnn(rs, index_of[b[j]] + k)];
rs                 13 lib/reed_solomon/encode_rs.c 	struct rs_codec *rs = rsc->codec;
rs                 15 lib/reed_solomon/encode_rs.c 	int nn = rs->nn;
rs                 16 lib/reed_solomon/encode_rs.c 	int nroots = rs->nroots;
rs                 17 lib/reed_solomon/encode_rs.c 	uint16_t *alpha_to = rs->alpha_to;
rs                 18 lib/reed_solomon/encode_rs.c 	uint16_t *index_of = rs->index_of;
rs                 19 lib/reed_solomon/encode_rs.c 	uint16_t *genpoly = rs->genpoly;
rs                 21 lib/reed_solomon/encode_rs.c 	uint16_t msk = (uint16_t) rs->nn;
rs                 33 lib/reed_solomon/encode_rs.c 				par[j] ^= alpha_to[rs_modnn(rs, fb +
rs                 40 lib/reed_solomon/encode_rs.c 			par[nroots - 1] = alpha_to[rs_modnn(rs,
rs                 74 lib/reed_solomon/reed_solomon.c 	struct rs_codec *rs;
rs                 76 lib/reed_solomon/reed_solomon.c 	rs = kzalloc(sizeof(*rs), gfp);
rs                 77 lib/reed_solomon/reed_solomon.c 	if (!rs)
rs                 80 lib/reed_solomon/reed_solomon.c 	INIT_LIST_HEAD(&rs->list);
rs                 82 lib/reed_solomon/reed_solomon.c 	rs->mm = symsize;
rs                 83 lib/reed_solomon/reed_solomon.c 	rs->nn = (1 << symsize) - 1;
rs                 84 lib/reed_solomon/reed_solomon.c 	rs->fcr = fcr;
rs                 85 lib/reed_solomon/reed_solomon.c 	rs->prim = prim;
rs                 86 lib/reed_solomon/reed_solomon.c 	rs->nroots = nroots;
rs                 87 lib/reed_solomon/reed_solomon.c 	rs->gfpoly = gfpoly;
rs                 88 lib/reed_solomon/reed_solomon.c 	rs->gffunc = gffunc;
rs                 91 lib/reed_solomon/reed_solomon.c 	rs->alpha_to = kmalloc_array(rs->nn + 1, sizeof(uint16_t), gfp);
rs                 92 lib/reed_solomon/reed_solomon.c 	if (rs->alpha_to == NULL)
rs                 95 lib/reed_solomon/reed_solomon.c 	rs->index_of = kmalloc_array(rs->nn + 1, sizeof(uint16_t), gfp);
rs                 96 lib/reed_solomon/reed_solomon.c 	if (rs->index_of == NULL)
rs                 99 lib/reed_solomon/reed_solomon.c 	rs->genpoly = kmalloc_array(rs->nroots + 1, sizeof(uint16_t), gfp);
rs                100 lib/reed_solomon/reed_solomon.c 	if(rs->genpoly == NULL)
rs                104 lib/reed_solomon/reed_solomon.c 	rs->index_of[0] = rs->nn;	/* log(zero) = -inf */
rs                105 lib/reed_solomon/reed_solomon.c 	rs->alpha_to[rs->nn] = 0;	/* alpha**-inf = 0 */
rs                108 lib/reed_solomon/reed_solomon.c 		for (i = 0; i < rs->nn; i++) {
rs                109 lib/reed_solomon/reed_solomon.c 			rs->index_of[sr] = i;
rs                110 lib/reed_solomon/reed_solomon.c 			rs->alpha_to[i] = sr;
rs                114 lib/reed_solomon/reed_solomon.c 			sr &= rs->nn;
rs                118 lib/reed_solomon/reed_solomon.c 		for (i = 0; i < rs->nn; i++) {
rs                119 lib/reed_solomon/reed_solomon.c 			rs->index_of[sr] = i;
rs                120 lib/reed_solomon/reed_solomon.c 			rs->alpha_to[i] = sr;
rs                125 lib/reed_solomon/reed_solomon.c 	if(sr != rs->alpha_to[0])
rs                129 lib/reed_solomon/reed_solomon.c 	for(iprim = 1; (iprim % prim) != 0; iprim += rs->nn);
rs                131 lib/reed_solomon/reed_solomon.c 	rs->iprim = iprim / prim;
rs                134 lib/reed_solomon/reed_solomon.c 	rs->genpoly[0] = 1;
rs                136 lib/reed_solomon/reed_solomon.c 		rs->genpoly[i + 1] = 1;
rs                139 lib/reed_solomon/reed_solomon.c 			if (rs->genpoly[j] != 0) {
rs                140 lib/reed_solomon/reed_solomon.c 				rs->genpoly[j] = rs->genpoly[j -1] ^
rs                141 lib/reed_solomon/reed_solomon.c 					rs->alpha_to[rs_modnn(rs,
rs                142 lib/reed_solomon/reed_solomon.c 					rs->index_of[rs->genpoly[j]] + root)];
rs                144 lib/reed_solomon/reed_solomon.c 				rs->genpoly[j] = rs->genpoly[j - 1];
rs                147 lib/reed_solomon/reed_solomon.c 		rs->genpoly[0] =
rs                148 lib/reed_solomon/reed_solomon.c 			rs->alpha_to[rs_modnn(rs,
rs                149 lib/reed_solomon/reed_solomon.c 				rs->index_of[rs->genpoly[0]] + root)];
rs                153 lib/reed_solomon/reed_solomon.c 		rs->genpoly[i] = rs->index_of[rs->genpoly[i]];
rs                155 lib/reed_solomon/reed_solomon.c 	rs->users = 1;
rs                156 lib/reed_solomon/reed_solomon.c 	list_add(&rs->list, &codec_list);
rs                157 lib/reed_solomon/reed_solomon.c 	return rs;
rs                160 lib/reed_solomon/reed_solomon.c 	kfree(rs->genpoly);
rs                161 lib/reed_solomon/reed_solomon.c 	kfree(rs->index_of);
rs                162 lib/reed_solomon/reed_solomon.c 	kfree(rs->alpha_to);
rs                163 lib/reed_solomon/reed_solomon.c 	kfree(rs);
rs                176 lib/reed_solomon/reed_solomon.c void free_rs(struct rs_control *rs)
rs                180 lib/reed_solomon/reed_solomon.c 	if (!rs)
rs                183 lib/reed_solomon/reed_solomon.c 	cd = rs->codec;
rs                194 lib/reed_solomon/reed_solomon.c 	kfree(rs);
rs                218 lib/reed_solomon/reed_solomon.c 	struct rs_control *rs;
rs                237 lib/reed_solomon/reed_solomon.c 	rs = kzalloc(sizeof(*rs) + bsize, gfp);
rs                238 lib/reed_solomon/reed_solomon.c 	if (!rs)
rs                261 lib/reed_solomon/reed_solomon.c 		rs->codec = cd;
rs                266 lib/reed_solomon/reed_solomon.c 	rs->codec = codec_init(symsize, gfpoly, gffunc, fcr, prim, nroots, gfp);
rs                267 lib/reed_solomon/reed_solomon.c 	if (!rs->codec) {
rs                268 lib/reed_solomon/reed_solomon.c 		kfree(rs);
rs                269 lib/reed_solomon/reed_solomon.c 		rs = NULL;
rs                273 lib/reed_solomon/reed_solomon.c 	return rs;
rs                108 lib/reed_solomon/test_rslib.c static struct wspace *alloc_ws(struct rs_codec *rs)
rs                110 lib/reed_solomon/test_rslib.c 	int nroots = rs->nroots;
rs                112 lib/reed_solomon/test_rslib.c 	int nn = rs->nn;
rs                151 lib/reed_solomon/test_rslib.c static int get_rcw_we(struct rs_control *rs, struct wspace *ws,
rs                154 lib/reed_solomon/test_rslib.c 	int nroots = rs->codec->nroots;
rs                158 lib/reed_solomon/test_rslib.c 	int nn = rs->codec->nn;
rs                170 lib/reed_solomon/test_rslib.c 	encode_rs16(rs, c, dlen, c + dlen, 0);
rs                232 lib/reed_solomon/test_rslib.c 	struct rs_codec *rs = rsc->codec;
rs                233 lib/reed_solomon/test_rslib.c 	uint16_t *alpha_to = rs->alpha_to;
rs                234 lib/reed_solomon/test_rslib.c 	uint16_t *index_of = rs->index_of;
rs                235 lib/reed_solomon/test_rslib.c 	int nroots = rs->nroots;
rs                236 lib/reed_solomon/test_rslib.c 	int prim = rs->prim;
rs                237 lib/reed_solomon/test_rslib.c 	int fcr = rs->fcr;
rs                248 lib/reed_solomon/test_rslib.c 					alpha_to[rs_modnn(rs, index_of[syn[i]]
rs                256 lib/reed_solomon/test_rslib.c 		syn[i] = rs->index_of[syn[i]];
rs                260 lib/reed_solomon/test_rslib.c static void test_uc(struct rs_control *rs, int len, int errs,
rs                264 lib/reed_solomon/test_rslib.c 	int dlen = len - rs->codec->nroots;
rs                275 lib/reed_solomon/test_rslib.c 		nerrs = get_rcw_we(rs, ws, len, errs, eras);
rs                279 lib/reed_solomon/test_rslib.c 			derrs = decode_rs16(rs, r, r + dlen, dlen,
rs                284 lib/reed_solomon/test_rslib.c 			compute_syndrome(rs, r, len, s);
rs                285 lib/reed_solomon/test_rslib.c 			derrs = decode_rs16(rs, NULL, NULL, dlen,
rs                290 lib/reed_solomon/test_rslib.c 			derrs = decode_rs16(rs, r, r + dlen, dlen,
rs                313 lib/reed_solomon/test_rslib.c static int ex_rs_helper(struct rs_control *rs, struct wspace *ws,
rs                323 lib/reed_solomon/test_rslib.c 	int nroots = rs->codec->nroots;
rs                331 lib/reed_solomon/test_rslib.c 			test_uc(rs, len, errs, eras, trials, &stat, ws, method);
rs                349 lib/reed_solomon/test_rslib.c static int exercise_rs(struct rs_control *rs, struct wspace *ws,
rs                360 lib/reed_solomon/test_rslib.c 		retval |= ex_rs_helper(rs, ws, len, trials, i);
rs                366 lib/reed_solomon/test_rslib.c static void test_bc(struct rs_control *rs, int len, int errs,
rs                370 lib/reed_solomon/test_rslib.c 	int nroots = rs->codec->nroots;
rs                378 lib/reed_solomon/test_rslib.c 		get_rcw_we(rs, ws, len, errs, eras);
rs                379 lib/reed_solomon/test_rslib.c 		derrs = decode_rs16(rs, r, r + dlen, dlen,
rs                396 lib/reed_solomon/test_rslib.c 			encode_rs16(rs, r, dlen, corr, 0);
rs                407 lib/reed_solomon/test_rslib.c static int exercise_rs_bc(struct rs_control *rs, struct wspace *ws,
rs                411 lib/reed_solomon/test_rslib.c 	int nroots = rs->codec->nroots;
rs                424 lib/reed_solomon/test_rslib.c 			test_bc(rs, len, errs, eras, trials, &stat, ws);
rs                 74 lib/test_hexdump.c 	int gs = groupsize, rs = rowsize;
rs                 78 lib/test_hexdump.c 	if (rs != 16 && rs != 32)
rs                 79 lib/test_hexdump.c 		rs = 16;
rs                 81 lib/test_hexdump.c 	if (l > rs)
rs                 82 lib/test_hexdump.c 		l = rs;
rs                114 lib/test_hexdump.c 		} while (p < test + rs * 2 + rs / gs + 1);
rs                166 lib/test_hexdump.c 	int rs = rowsize, gs = groupsize;
rs                174 lib/test_hexdump.c 	r = hex_dump_to_buffer(data_b, len, rs, gs, buf, buflen, ascii);
rs                180 lib/test_hexdump.c 	ae = rs * 2 /* hex */ + rs / gs /* spaces */ + 1 /* space */ + len /* ascii */;
rs                190 lib/test_hexdump.c 		test_hexdump_prepare_test(len, rs, gs, test, sizeof(test), ascii);
rs                211 lib/test_hexdump.c 	int rs = (get_random_int() % 2 + 1) * 16;
rs                215 lib/test_hexdump.c 		size_t len = get_random_int() % rs + gs;
rs                217 lib/test_hexdump.c 		test_hexdump_overflow(buflen, rounddown(len, gs), rs, gs, ascii);
rs               1515 mm/memory-failure.c #define unpoison_pr_info(fmt, pfn, rs)			\
rs               1517 mm/memory-failure.c 	if (__ratelimit(rs))				\
rs                273 mm/percpu.c    static void pcpu_next_unpop(unsigned long *bitmap, int *rs, int *re, int end)
rs                275 mm/percpu.c    	*rs = find_next_zero_bit(bitmap, end, *rs);
rs                276 mm/percpu.c    	*re = find_next_bit(bitmap, end, *rs + 1);
rs                279 mm/percpu.c    static void pcpu_next_pop(unsigned long *bitmap, int *rs, int *re, int end)
rs                281 mm/percpu.c    	*rs = find_next_bit(bitmap, end, *rs);
rs                282 mm/percpu.c    	*re = find_next_zero_bit(bitmap, end, *rs + 1);
rs                290 mm/percpu.c    #define pcpu_for_each_unpop_region(bitmap, rs, re, start, end)		     \
rs                291 mm/percpu.c    	for ((rs) = (start), pcpu_next_unpop((bitmap), &(rs), &(re), (end)); \
rs                292 mm/percpu.c    	     (rs) < (re);						     \
rs                293 mm/percpu.c    	     (rs) = (re) + 1, pcpu_next_unpop((bitmap), &(rs), &(re), (end)))
rs                295 mm/percpu.c    #define pcpu_for_each_pop_region(bitmap, rs, re, start, end)		     \
rs                296 mm/percpu.c    	for ((rs) = (start), pcpu_next_pop((bitmap), &(rs), &(re), (end));   \
rs                297 mm/percpu.c    	     (rs) < (re);						     \
rs                298 mm/percpu.c    	     (rs) = (re) + 1, pcpu_next_pop((bitmap), &(rs), &(re), (end)))
rs                752 mm/percpu.c    	int rs, re, start;	/* region start, region end */
rs                768 mm/percpu.c    	pcpu_for_each_unpop_region(alloc_map, rs, re, start,
rs                770 mm/percpu.c    		pcpu_block_update(block, rs, re);
rs               1044 mm/percpu.c    	int page_start, page_end, rs, re;
rs               1049 mm/percpu.c    	rs = page_start;
rs               1050 mm/percpu.c    	pcpu_next_unpop(chunk->populated, &rs, &re, page_end);
rs               1051 mm/percpu.c    	if (rs >= page_end)
rs               1705 mm/percpu.c    		int page_start, page_end, rs, re;
rs               1710 mm/percpu.c    		pcpu_for_each_unpop_region(chunk->populated, rs, re,
rs               1714 mm/percpu.c    			ret = pcpu_populate_chunk(chunk, rs, re, pcpu_gfp);
rs               1722 mm/percpu.c    			pcpu_chunk_populated(chunk, rs, re);
rs               1861 mm/percpu.c    		int rs, re;
rs               1863 mm/percpu.c    		pcpu_for_each_pop_region(chunk->populated, rs, re, 0,
rs               1865 mm/percpu.c    			pcpu_depopulate_chunk(chunk, rs, re);
rs               1867 mm/percpu.c    			pcpu_chunk_depopulated(chunk, rs, re);
rs               1896 mm/percpu.c    		int nr_unpop = 0, rs, re;
rs               1913 mm/percpu.c    		pcpu_for_each_unpop_region(chunk->populated, rs, re, 0,
rs               1915 mm/percpu.c    			int nr = min(re - rs, nr_to_pop);
rs               1917 mm/percpu.c    			ret = pcpu_populate_chunk(chunk, rs, rs + nr, gfp);
rs               1921 mm/percpu.c    				pcpu_chunk_populated(chunk, rs, rs + nr);
rs                175 mm/vmscan.c    				   struct reclaim_state *rs)
rs                178 mm/vmscan.c    	WARN_ON_ONCE(rs && task->reclaim_state);
rs                181 mm/vmscan.c    	WARN_ON_ONCE(!rs && !task->reclaim_state);
rs                183 mm/vmscan.c    	task->reclaim_state = rs;
rs                502 net/decnet/dn_neigh.c 	unsigned char *rs;
rs                519 net/decnet/dn_neigh.c 		s->rs = dn_find_slot(s->ptr, s->n, dn->priority);
rs                522 net/decnet/dn_neigh.c 	if (s->rs == NULL)
rs                525 net/decnet/dn_neigh.c 	dn_dn2eth(s->rs, dn->addr);
rs                526 net/decnet/dn_neigh.c 	s->rs += 6;
rs                527 net/decnet/dn_neigh.c 	*(s->rs) = neigh->nud_state & NUD_CONNECTED ? 0x80 : 0x0;
rs                528 net/decnet/dn_neigh.c 	*(s->rs) |= dn->priority;
rs                529 net/decnet/dn_neigh.c 	s->rs++;
rs                540 net/decnet/dn_neigh.c 	state.rs = ptr;
rs                478 net/ipv4/tcp_bbr.c 	struct sock *sk, const struct rate_sample *rs, u32 acked, u32 *new_cwnd)
rs                489 net/ipv4/tcp_bbr.c 	if (rs->losses > 0)
rs                490 net/ipv4/tcp_bbr.c 		cwnd = max_t(s32, cwnd - rs->losses, 1);
rs                516 net/ipv4/tcp_bbr.c static void bbr_set_cwnd(struct sock *sk, const struct rate_sample *rs,
rs                526 net/ipv4/tcp_bbr.c 	if (bbr_set_cwnd_to_recover_or_restore(sk, rs, acked, &cwnd))
rs                552 net/ipv4/tcp_bbr.c 				    const struct rate_sample *rs)
rs                567 net/ipv4/tcp_bbr.c 	inflight = bbr_packets_in_net_at_edt(sk, rs->prior_in_flight);
rs                577 net/ipv4/tcp_bbr.c 			(rs->losses ||  /* perhaps pacing_gain*BDP won't fit */
rs                599 net/ipv4/tcp_bbr.c 				   const struct rate_sample *rs)
rs                603 net/ipv4/tcp_bbr.c 	if (bbr->mode == BBR_PROBE_BW && bbr_is_next_cycle_phase(sk, rs))
rs                685 net/ipv4/tcp_bbr.c static void bbr_lt_bw_sampling(struct sock *sk, const struct rate_sample *rs)
rs                707 net/ipv4/tcp_bbr.c 		if (!rs->losses)
rs                714 net/ipv4/tcp_bbr.c 	if (rs->is_app_limited) {
rs                732 net/ipv4/tcp_bbr.c 	if (!rs->losses)
rs                758 net/ipv4/tcp_bbr.c static void bbr_update_bw(struct sock *sk, const struct rate_sample *rs)
rs                765 net/ipv4/tcp_bbr.c 	if (rs->delivered < 0 || rs->interval_us <= 0)
rs                769 net/ipv4/tcp_bbr.c 	if (!before(rs->prior_delivered, bbr->next_rtt_delivered)) {
rs                776 net/ipv4/tcp_bbr.c 	bbr_lt_bw_sampling(sk, rs);
rs                782 net/ipv4/tcp_bbr.c 	bw = div64_long((u64)rs->delivered * BW_UNIT, rs->interval_us);
rs                795 net/ipv4/tcp_bbr.c 	if (!rs->is_app_limited || bw >= bbr_max_bw(sk)) {
rs                815 net/ipv4/tcp_bbr.c 				       const struct rate_sample *rs)
rs                821 net/ipv4/tcp_bbr.c 	if (!bbr_extra_acked_gain || rs->acked_sacked <= 0 ||
rs                822 net/ipv4/tcp_bbr.c 	    rs->delivered < 0 || rs->interval_us <= 0)
rs                846 net/ipv4/tcp_bbr.c 	    (bbr->ack_epoch_acked + rs->acked_sacked >=
rs                855 net/ipv4/tcp_bbr.c 				     bbr->ack_epoch_acked + rs->acked_sacked);
rs                871 net/ipv4/tcp_bbr.c 				      const struct rate_sample *rs)
rs                876 net/ipv4/tcp_bbr.c 	if (bbr_full_bw_reached(sk) || !bbr->round_start || rs->is_app_limited)
rs                890 net/ipv4/tcp_bbr.c static void bbr_check_drain(struct sock *sk, const struct rate_sample *rs)
rs                938 net/ipv4/tcp_bbr.c static void bbr_update_min_rtt(struct sock *sk, const struct rate_sample *rs)
rs                947 net/ipv4/tcp_bbr.c 	if (rs->rtt_us >= 0 &&
rs                948 net/ipv4/tcp_bbr.c 	    (rs->rtt_us <= bbr->min_rtt_us ||
rs                949 net/ipv4/tcp_bbr.c 	     (filter_expired && !rs->is_ack_delayed))) {
rs                950 net/ipv4/tcp_bbr.c 		bbr->min_rtt_us = rs->rtt_us;
rs                980 net/ipv4/tcp_bbr.c 	if (rs->delivered > 0)
rs               1013 net/ipv4/tcp_bbr.c static void bbr_update_model(struct sock *sk, const struct rate_sample *rs)
rs               1015 net/ipv4/tcp_bbr.c 	bbr_update_bw(sk, rs);
rs               1016 net/ipv4/tcp_bbr.c 	bbr_update_ack_aggregation(sk, rs);
rs               1017 net/ipv4/tcp_bbr.c 	bbr_update_cycle_phase(sk, rs);
rs               1018 net/ipv4/tcp_bbr.c 	bbr_check_full_bw_reached(sk, rs);
rs               1019 net/ipv4/tcp_bbr.c 	bbr_check_drain(sk, rs);
rs               1020 net/ipv4/tcp_bbr.c 	bbr_update_min_rtt(sk, rs);
rs               1024 net/ipv4/tcp_bbr.c static void bbr_main(struct sock *sk, const struct rate_sample *rs)
rs               1029 net/ipv4/tcp_bbr.c 	bbr_update_model(sk, rs);
rs               1033 net/ipv4/tcp_bbr.c 	bbr_set_cwnd(sk, rs, rs->acked_sacked, bw, bbr->cwnd_gain);
rs               1131 net/ipv4/tcp_bbr.c 		struct rate_sample rs = { .losses = 1 };
rs               1136 net/ipv4/tcp_bbr.c 		bbr_lt_bw_sampling(sk, &rs);
rs               2923 net/ipv4/tcp_input.c 			       long ca_rtt_us, struct rate_sample *rs)
rs               2950 net/ipv4/tcp_input.c 	rs->rtt_us = ca_rtt_us; /* RTT of last (S)ACKed packet (or -1) */
rs               2970 net/ipv4/tcp_input.c 	struct rate_sample rs;
rs               2976 net/ipv4/tcp_input.c 	tcp_ack_update_rtt(sk, FLAG_SYN_ACKED, rtt_us, -1L, rtt_us, &rs);
rs               3324 net/ipv4/tcp_input.c 			     int flag, const struct rate_sample *rs)
rs               3329 net/ipv4/tcp_input.c 		icsk->icsk_ca_ops->cong_control(sk, rs);
rs               3590 net/ipv4/tcp_input.c 	struct rate_sample rs = { .prior_delivered = 0 };
rs               3603 net/ipv4/tcp_input.c 	sack_state.rate = &rs;
rs               3639 net/ipv4/tcp_input.c 	rs.prior_in_flight = tcp_packets_in_flight(tp);
rs               3697 net/ipv4/tcp_input.c 	tcp_rack_update_reo_wnd(sk, &rs);
rs               3721 net/ipv4/tcp_input.c 	rs.is_ack_delayed = !!(flag & FLAG_ACK_MAYBE_DELAYED);
rs                 79 net/ipv4/tcp_rate.c 			    struct rate_sample *rs)
rs                 87 net/ipv4/tcp_rate.c 	if (!rs->prior_delivered ||
rs                 88 net/ipv4/tcp_rate.c 	    after(scb->tx.delivered, rs->prior_delivered)) {
rs                 89 net/ipv4/tcp_rate.c 		rs->prior_delivered  = scb->tx.delivered;
rs                 90 net/ipv4/tcp_rate.c 		rs->prior_mstamp     = scb->tx.delivered_mstamp;
rs                 91 net/ipv4/tcp_rate.c 		rs->is_app_limited   = scb->tx.is_app_limited;
rs                 92 net/ipv4/tcp_rate.c 		rs->is_retrans	     = scb->sacked & TCPCB_RETRANS;
rs                 97 net/ipv4/tcp_rate.c 		rs->interval_us = tcp_stamp_us_delta(tp->first_tx_mstamp,
rs                111 net/ipv4/tcp_rate.c 		  bool is_sack_reneg, struct rate_sample *rs)
rs                127 net/ipv4/tcp_rate.c 	rs->acked_sacked = delivered;	/* freshly ACKed or SACKed */
rs                128 net/ipv4/tcp_rate.c 	rs->losses = lost;		/* freshly marked lost */
rs                134 net/ipv4/tcp_rate.c 	if (!rs->prior_mstamp || is_sack_reneg) {
rs                135 net/ipv4/tcp_rate.c 		rs->delivered = -1;
rs                136 net/ipv4/tcp_rate.c 		rs->interval_us = -1;
rs                139 net/ipv4/tcp_rate.c 	rs->delivered   = tp->delivered - rs->prior_delivered;
rs                146 net/ipv4/tcp_rate.c 	snd_us = rs->interval_us;				/* send phase */
rs                148 net/ipv4/tcp_rate.c 				    rs->prior_mstamp); /* ack phase */
rs                149 net/ipv4/tcp_rate.c 	rs->interval_us = max(snd_us, ack_us);
rs                152 net/ipv4/tcp_rate.c 	rs->snd_interval_us = snd_us;
rs                153 net/ipv4/tcp_rate.c 	rs->rcv_interval_us = ack_us;
rs                162 net/ipv4/tcp_rate.c 	if (unlikely(rs->interval_us < tcp_min_rtt(tp))) {
rs                163 net/ipv4/tcp_rate.c 		if (!rs->is_retrans)
rs                165 net/ipv4/tcp_rate.c 				 rs->interval_us, rs->delivered,
rs                168 net/ipv4/tcp_rate.c 		rs->interval_us = -1;
rs                173 net/ipv4/tcp_rate.c 	if (!rs->is_app_limited ||
rs                174 net/ipv4/tcp_rate.c 	    ((u64)rs->delivered * tp->rate_interval_us >=
rs                175 net/ipv4/tcp_rate.c 	     (u64)tp->rate_delivered * rs->interval_us)) {
rs                176 net/ipv4/tcp_rate.c 		tp->rate_delivered = rs->delivered;
rs                177 net/ipv4/tcp_rate.c 		tp->rate_interval_us = rs->interval_us;
rs                178 net/ipv4/tcp_rate.c 		tp->rate_app_limited = rs->is_app_limited;
rs                202 net/ipv4/tcp_recovery.c void tcp_rack_update_reo_wnd(struct sock *sk, struct rate_sample *rs)
rs                207 net/ipv4/tcp_recovery.c 	    !rs->prior_delivered)
rs                211 net/ipv4/tcp_recovery.c 	if (before(rs->prior_delivered, tp->rack.last_delivered))
rs                 62 net/rds/af_rds.c 	struct rds_sock *rs;
rs                 67 net/rds/af_rds.c 	rs = rds_sk_to_rs(sk);
rs                 73 net/rds/af_rds.c 	rds_clear_recv_queue(rs);
rs                 74 net/rds/af_rds.c 	rds_cong_remove_socket(rs);
rs                 76 net/rds/af_rds.c 	rds_remove_bound(rs);
rs                 78 net/rds/af_rds.c 	rds_send_drop_to(rs, NULL);
rs                 79 net/rds/af_rds.c 	rds_rdma_drop_keys(rs);
rs                 80 net/rds/af_rds.c 	rds_notify_queue_get(rs, NULL);
rs                 81 net/rds/af_rds.c 	rds_notify_msg_zcopy_purge(&rs->rs_zcookie_queue);
rs                 84 net/rds/af_rds.c 	list_del_init(&rs->rs_item);
rs                 88 net/rds/af_rds.c 	rds_trans_put(rs->rs_transport);
rs                105 net/rds/af_rds.c void rds_wake_sk_sleep(struct rds_sock *rs)
rs                109 net/rds/af_rds.c 	read_lock_irqsave(&rs->rs_recv_lock, flags);
rs                110 net/rds/af_rds.c 	__rds_wake_sk_sleep(rds_rs_to_sk(rs));
rs                111 net/rds/af_rds.c 	read_unlock_irqrestore(&rs->rs_recv_lock, flags);
rs                117 net/rds/af_rds.c 	struct rds_sock *rs = rds_sk_to_rs(sock->sk);
rs                124 net/rds/af_rds.c 		if (ipv6_addr_any(&rs->rs_conn_addr))
rs                127 net/rds/af_rds.c 		if (ipv6_addr_v4mapped(&rs->rs_conn_addr)) {
rs                131 net/rds/af_rds.c 			sin->sin_port = rs->rs_conn_port;
rs                132 net/rds/af_rds.c 			sin->sin_addr.s_addr = rs->rs_conn_addr_v4;
rs                137 net/rds/af_rds.c 			sin6->sin6_port = rs->rs_conn_port;
rs                138 net/rds/af_rds.c 			sin6->sin6_addr = rs->rs_conn_addr;
rs                141 net/rds/af_rds.c 			sin6->sin6_scope_id = rs->rs_bound_scope_id;
rs                151 net/rds/af_rds.c 		if (ipv6_addr_any(&rs->rs_bound_addr)) {
rs                152 net/rds/af_rds.c 			if (ipv6_addr_any(&rs->rs_conn_addr)) {
rs                160 net/rds/af_rds.c 			if (!(ipv6_addr_type(&rs->rs_conn_addr) &
rs                174 net/rds/af_rds.c 		if (ipv6_addr_v4mapped(&rs->rs_bound_addr)) {
rs                178 net/rds/af_rds.c 			sin->sin_port = rs->rs_bound_port;
rs                179 net/rds/af_rds.c 			sin->sin_addr.s_addr = rs->rs_bound_addr_v4;
rs                184 net/rds/af_rds.c 			sin6->sin6_port = rs->rs_bound_port;
rs                185 net/rds/af_rds.c 			sin6->sin6_addr = rs->rs_bound_addr;
rs                187 net/rds/af_rds.c 			sin6->sin6_scope_id = rs->rs_bound_scope_id;
rs                216 net/rds/af_rds.c 	struct rds_sock *rs = rds_sk_to_rs(sk);
rs                222 net/rds/af_rds.c 	if (rs->rs_seen_congestion)
rs                225 net/rds/af_rds.c 	read_lock_irqsave(&rs->rs_recv_lock, flags);
rs                226 net/rds/af_rds.c 	if (!rs->rs_cong_monitor) {
rs                230 net/rds/af_rds.c 		if (rds_cong_updated_since(&rs->rs_cong_track))
rs                233 net/rds/af_rds.c 		spin_lock(&rs->rs_lock);
rs                234 net/rds/af_rds.c 		if (rs->rs_cong_notify)
rs                236 net/rds/af_rds.c 		spin_unlock(&rs->rs_lock);
rs                238 net/rds/af_rds.c 	if (!list_empty(&rs->rs_recv_queue) ||
rs                239 net/rds/af_rds.c 	    !list_empty(&rs->rs_notify_queue) ||
rs                240 net/rds/af_rds.c 	    !list_empty(&rs->rs_zcookie_queue.zcookie_head))
rs                242 net/rds/af_rds.c 	if (rs->rs_snd_bytes < rds_sk_sndbuf(rs))
rs                246 net/rds/af_rds.c 	read_unlock_irqrestore(&rs->rs_recv_lock, flags);
rs                250 net/rds/af_rds.c 		rs->rs_seen_congestion = 0;
rs                257 net/rds/af_rds.c 	struct rds_sock *rs = rds_sk_to_rs(sock->sk);
rs                265 net/rds/af_rds.c 		if (rs->rs_transport &&
rs                266 net/rds/af_rds.c 		    rs->rs_transport->get_tos_map)
rs                267 net/rds/af_rds.c 			tos = rs->rs_transport->get_tos_map(utos);
rs                272 net/rds/af_rds.c 		if (rs->rs_tos || rs->rs_conn) {
rs                276 net/rds/af_rds.c 		rs->rs_tos = tos;
rs                281 net/rds/af_rds.c 		tos = rs->rs_tos;
rs                293 net/rds/af_rds.c static int rds_cancel_sent_to(struct rds_sock *rs, char __user *optval,
rs                301 net/rds/af_rds.c 	if (ipv6_addr_any(&rs->rs_bound_addr)) {
rs                325 net/rds/af_rds.c 	rds_send_drop_to(rs, &sin6);
rs                343 net/rds/af_rds.c static int rds_cong_monitor(struct rds_sock *rs, char __user *optval,
rs                348 net/rds/af_rds.c 	ret = rds_set_bool_option(&rs->rs_cong_monitor, optval, optlen);
rs                350 net/rds/af_rds.c 		if (rs->rs_cong_monitor) {
rs                351 net/rds/af_rds.c 			rds_cong_add_socket(rs);
rs                353 net/rds/af_rds.c 			rds_cong_remove_socket(rs);
rs                354 net/rds/af_rds.c 			rs->rs_cong_mask = 0;
rs                355 net/rds/af_rds.c 			rs->rs_cong_notify = 0;
rs                361 net/rds/af_rds.c static int rds_set_transport(struct rds_sock *rs, char __user *optval,
rs                366 net/rds/af_rds.c 	if (rs->rs_transport)
rs                378 net/rds/af_rds.c 	rs->rs_transport = rds_trans_get(t_type);
rs                380 net/rds/af_rds.c 	return rs->rs_transport ? 0 : -ENOPROTOOPT;
rs                407 net/rds/af_rds.c static int rds_recv_track_latency(struct rds_sock *rs, char __user *optval,
rs                422 net/rds/af_rds.c 	rs->rs_rx_traces = trace.rx_traces;
rs                423 net/rds/af_rds.c 	for (i = 0; i < rs->rs_rx_traces; i++) {
rs                425 net/rds/af_rds.c 			rs->rs_rx_traces = 0;
rs                428 net/rds/af_rds.c 		rs->rs_rx_trace[i] = trace.rx_trace_pos[i];
rs                437 net/rds/af_rds.c 	struct rds_sock *rs = rds_sk_to_rs(sock->sk);
rs                447 net/rds/af_rds.c 		ret = rds_cancel_sent_to(rs, optval, optlen);
rs                450 net/rds/af_rds.c 		ret = rds_get_mr(rs, optval, optlen);
rs                453 net/rds/af_rds.c 		ret = rds_get_mr_for_dest(rs, optval, optlen);
rs                456 net/rds/af_rds.c 		ret = rds_free_mr(rs, optval, optlen);
rs                459 net/rds/af_rds.c 		ret = rds_set_bool_option(&rs->rs_recverr, optval, optlen);
rs                462 net/rds/af_rds.c 		ret = rds_cong_monitor(rs, optval, optlen);
rs                466 net/rds/af_rds.c 		ret = rds_set_transport(rs, optval, optlen);
rs                476 net/rds/af_rds.c 		ret = rds_recv_track_latency(rs, optval, optlen);
rs                488 net/rds/af_rds.c 	struct rds_sock *rs = rds_sk_to_rs(sock->sk);
rs                510 net/rds/af_rds.c 		if (put_user(rs->rs_recverr, (int __user *) optval) ||
rs                521 net/rds/af_rds.c 		trans = (rs->rs_transport ? rs->rs_transport->t_type :
rs                543 net/rds/af_rds.c 	struct rds_sock *rs = rds_sk_to_rs(sk);
rs                567 net/rds/af_rds.c 		ipv6_addr_set_v4mapped(sin->sin_addr.s_addr, &rs->rs_conn_addr);
rs                568 net/rds/af_rds.c 		rs->rs_conn_port = sin->sin_port;
rs                607 net/rds/af_rds.c 			    (!ipv6_addr_any(&rs->rs_bound_addr) &&
rs                608 net/rds/af_rds.c 			     rs->rs_bound_scope_id &&
rs                609 net/rds/af_rds.c 			     sin6->sin6_scope_id != rs->rs_bound_scope_id)) {
rs                617 net/rds/af_rds.c 			rs->rs_bound_scope_id = sin6->sin6_scope_id;
rs                619 net/rds/af_rds.c 		rs->rs_conn_addr = sin6->sin6_addr;
rs                620 net/rds/af_rds.c 		rs->rs_conn_port = sin6->sin6_port;
rs                663 net/rds/af_rds.c 	struct rds_sock *rs = rds_sk_to_rs(sk);
rs                665 net/rds/af_rds.c 	WARN_ON((&rs->rs_item != rs->rs_item.next ||
rs                666 net/rds/af_rds.c 		 &rs->rs_item != rs->rs_item.prev));
rs                671 net/rds/af_rds.c 	struct rds_sock *rs;
rs                678 net/rds/af_rds.c 	rs = rds_sk_to_rs(sk);
rs                679 net/rds/af_rds.c 	spin_lock_init(&rs->rs_lock);
rs                680 net/rds/af_rds.c 	rwlock_init(&rs->rs_recv_lock);
rs                681 net/rds/af_rds.c 	INIT_LIST_HEAD(&rs->rs_send_queue);
rs                682 net/rds/af_rds.c 	INIT_LIST_HEAD(&rs->rs_recv_queue);
rs                683 net/rds/af_rds.c 	INIT_LIST_HEAD(&rs->rs_notify_queue);
rs                684 net/rds/af_rds.c 	INIT_LIST_HEAD(&rs->rs_cong_list);
rs                685 net/rds/af_rds.c 	rds_message_zcopy_queue_init(&rs->rs_zcookie_queue);
rs                686 net/rds/af_rds.c 	spin_lock_init(&rs->rs_rdma_lock);
rs                687 net/rds/af_rds.c 	rs->rs_rdma_keys = RB_ROOT;
rs                688 net/rds/af_rds.c 	rs->rs_rx_traces = 0;
rs                689 net/rds/af_rds.c 	rs->rs_tos = 0;
rs                690 net/rds/af_rds.c 	rs->rs_conn = NULL;
rs                693 net/rds/af_rds.c 	list_add_tail(&rs->rs_item, &rds_sock_list);
rs                715 net/rds/af_rds.c void rds_sock_addref(struct rds_sock *rs)
rs                717 net/rds/af_rds.c 	sock_hold(rds_rs_to_sk(rs));
rs                720 net/rds/af_rds.c void rds_sock_put(struct rds_sock *rs)
rs                722 net/rds/af_rds.c 	sock_put(rds_rs_to_sk(rs));
rs                735 net/rds/af_rds.c 	struct rds_sock *rs;
rs                743 net/rds/af_rds.c 	list_for_each_entry(rs, &rds_sock_list, rs_item) {
rs                745 net/rds/af_rds.c 		if (!ipv6_addr_v4mapped(&rs->rs_bound_addr))
rs                748 net/rds/af_rds.c 		read_lock(&rs->rs_recv_lock);
rs                751 net/rds/af_rds.c 		list_for_each_entry(inc, &rs->rs_recv_queue, i_item) {
rs                756 net/rds/af_rds.c 						  rs->rs_bound_addr_v4,
rs                760 net/rds/af_rds.c 		read_unlock(&rs->rs_recv_lock);
rs                776 net/rds/af_rds.c 	struct rds_sock *rs;
rs                782 net/rds/af_rds.c 	list_for_each_entry(rs, &rds_sock_list, rs_item) {
rs                783 net/rds/af_rds.c 		read_lock(&rs->rs_recv_lock);
rs                785 net/rds/af_rds.c 		list_for_each_entry(inc, &rs->rs_recv_queue, i_item) {
rs                789 net/rds/af_rds.c 						   &rs->rs_bound_addr, 1);
rs                792 net/rds/af_rds.c 		read_unlock(&rs->rs_recv_lock);
rs                808 net/rds/af_rds.c 	struct rds_sock *rs;
rs                819 net/rds/af_rds.c 	list_for_each_entry(rs, &rds_sock_list, rs_item) {
rs                821 net/rds/af_rds.c 		if (!ipv6_addr_v4mapped(&rs->rs_bound_addr))
rs                823 net/rds/af_rds.c 		sinfo.sndbuf = rds_sk_sndbuf(rs);
rs                824 net/rds/af_rds.c 		sinfo.rcvbuf = rds_sk_rcvbuf(rs);
rs                825 net/rds/af_rds.c 		sinfo.bound_addr = rs->rs_bound_addr_v4;
rs                826 net/rds/af_rds.c 		sinfo.connected_addr = rs->rs_conn_addr_v4;
rs                827 net/rds/af_rds.c 		sinfo.bound_port = rs->rs_bound_port;
rs                828 net/rds/af_rds.c 		sinfo.connected_port = rs->rs_conn_port;
rs                829 net/rds/af_rds.c 		sinfo.inum = sock_i_ino(rds_rs_to_sk(rs));
rs                848 net/rds/af_rds.c 	struct rds_sock *rs;
rs                857 net/rds/af_rds.c 	list_for_each_entry(rs, &rds_sock_list, rs_item) {
rs                858 net/rds/af_rds.c 		sinfo6.sndbuf = rds_sk_sndbuf(rs);
rs                859 net/rds/af_rds.c 		sinfo6.rcvbuf = rds_sk_rcvbuf(rs);
rs                860 net/rds/af_rds.c 		sinfo6.bound_addr = rs->rs_bound_addr;
rs                861 net/rds/af_rds.c 		sinfo6.connected_addr = rs->rs_conn_addr;
rs                862 net/rds/af_rds.c 		sinfo6.bound_port = rs->rs_bound_port;
rs                863 net/rds/af_rds.c 		sinfo6.connected_port = rs->rs_conn_port;
rs                864 net/rds/af_rds.c 		sinfo6.inum = sock_i_ino(rds_rs_to_sk(rs));
rs                 76 net/rds/bind.c 	struct rds_sock *rs;
rs                 80 net/rds/bind.c 	rs = rhashtable_lookup(&bind_hash_table, key, ht_parms);
rs                 81 net/rds/bind.c 	if (rs && (sock_flag(rds_rs_to_sk(rs), SOCK_DEAD) ||
rs                 82 net/rds/bind.c 		   !refcount_inc_not_zero(&rds_rs_to_sk(rs)->sk_refcnt)))
rs                 83 net/rds/bind.c 		rs = NULL;
rs                 87 net/rds/bind.c 	rdsdebug("returning rs %p for %pI6c:%u\n", rs, addr,
rs                 90 net/rds/bind.c 	return rs;
rs                 94 net/rds/bind.c static int rds_add_bound(struct rds_sock *rs, const struct in6_addr *addr,
rs                122 net/rds/bind.c 		memcpy(rs->rs_bound_key, key, sizeof(rs->rs_bound_key));
rs                123 net/rds/bind.c 		rs->rs_bound_addr = *addr;
rs                124 net/rds/bind.c 		net_get_random_once(&rs->rs_hash_initval,
rs                125 net/rds/bind.c 				    sizeof(rs->rs_hash_initval));
rs                126 net/rds/bind.c 		rs->rs_bound_port = cpu_to_be16(rover);
rs                127 net/rds/bind.c 		rs->rs_bound_node.next = NULL;
rs                128 net/rds/bind.c 		rds_sock_addref(rs);
rs                130 net/rds/bind.c 					    &rs->rs_bound_node, ht_parms)) {
rs                131 net/rds/bind.c 			*port = rs->rs_bound_port;
rs                132 net/rds/bind.c 			rs->rs_bound_scope_id = scope_id;
rs                135 net/rds/bind.c 				 rs, addr, (int)ntohs(*port));
rs                138 net/rds/bind.c 			rs->rs_bound_addr = in6addr_any;
rs                139 net/rds/bind.c 			rds_sock_put(rs);
rs                148 net/rds/bind.c void rds_remove_bound(struct rds_sock *rs)
rs                151 net/rds/bind.c 	if (ipv6_addr_any(&rs->rs_bound_addr))
rs                155 net/rds/bind.c 		 rs, &rs->rs_bound_addr,
rs                156 net/rds/bind.c 		 ntohs(rs->rs_bound_port));
rs                158 net/rds/bind.c 	rhashtable_remove_fast(&bind_hash_table, &rs->rs_bound_node, ht_parms);
rs                159 net/rds/bind.c 	rds_sock_put(rs);
rs                160 net/rds/bind.c 	rs->rs_bound_addr = in6addr_any;
rs                166 net/rds/bind.c 	struct rds_sock *rs = rds_sk_to_rs(sk);
rs                227 net/rds/bind.c 	if (!ipv6_addr_any(&rs->rs_bound_addr)) {
rs                235 net/rds/bind.c 	if (!ipv6_addr_any(&rs->rs_conn_addr) && scope_id &&
rs                236 net/rds/bind.c 	    rs->rs_bound_scope_id &&
rs                237 net/rds/bind.c 	    scope_id != rs->rs_bound_scope_id) {
rs                245 net/rds/bind.c 	if (rs->rs_transport) {
rs                246 net/rds/bind.c 		trans = rs->rs_transport;
rs                262 net/rds/bind.c 		rs->rs_transport = trans;
rs                266 net/rds/bind.c 	ret = rds_add_bound(rs, binding_addr, &port, scope_id);
rs                268 net/rds/bind.c 		rs->rs_transport = NULL;
rs                266 net/rds/cong.c 		struct rds_sock *rs;
rs                269 net/rds/cong.c 		list_for_each_entry(rs, &rds_cong_monitor, rs_cong_list) {
rs                270 net/rds/cong.c 			spin_lock(&rs->rs_lock);
rs                271 net/rds/cong.c 			rs->rs_cong_notify |= (rs->rs_cong_mask & portmask);
rs                272 net/rds/cong.c 			rs->rs_cong_mask &= ~portmask;
rs                273 net/rds/cong.c 			spin_unlock(&rs->rs_lock);
rs                274 net/rds/cong.c 			if (rs->rs_cong_notify)
rs                275 net/rds/cong.c 				rds_wake_sk_sleep(rs);
rs                338 net/rds/cong.c void rds_cong_add_socket(struct rds_sock *rs)
rs                343 net/rds/cong.c 	if (list_empty(&rs->rs_cong_list))
rs                344 net/rds/cong.c 		list_add(&rs->rs_cong_list, &rds_cong_monitor);
rs                348 net/rds/cong.c void rds_cong_remove_socket(struct rds_sock *rs)
rs                354 net/rds/cong.c 	list_del_init(&rs->rs_cong_list);
rs                359 net/rds/cong.c 	map = rds_cong_tree_walk(&rs->rs_bound_addr, NULL);
rs                362 net/rds/cong.c 	if (map && rds_cong_test_bit(map, rs->rs_bound_port)) {
rs                363 net/rds/cong.c 		rds_cong_clear_bit(map, rs->rs_bound_port);
rs                369 net/rds/cong.c 		  struct rds_sock *rs)
rs                374 net/rds/cong.c 		if (rs && rs->rs_cong_monitor) {
rs                379 net/rds/cong.c 			spin_lock_irqsave(&rs->rs_lock, flags);
rs                380 net/rds/cong.c 			rs->rs_cong_mask |= RDS_CONG_MONITOR_MASK(ntohs(port));
rs                381 net/rds/cong.c 			spin_unlock_irqrestore(&rs->rs_lock, flags);
rs                124 net/rds/ib_mr.h 		    struct rds_sock *rs, u32 *key_ret,
rs                530 net/rds/ib_rdma.c 		    struct rds_sock *rs, u32 *key_ret,
rs                538 net/rds/ib_rdma.c 	rds_ibdev = rds_ib_get_device(rs->rs_bound_addr.s6_addr32[3]);
rs                 92 net/rds/message.c static void rds_rm_zerocopy_callback(struct rds_sock *rs,
rs                103 net/rds/message.c 	q = &rs->rs_zcookie_queue;
rs                140 net/rds/message.c 		struct rds_sock *rs = rm->m_rs;
rs                144 net/rds/message.c 			rds_rm_zerocopy_callback(rs, rm->data.op_mmp_znotifier);
rs                145 net/rds/message.c 			rds_wake_sk_sleep(rs);
rs                148 net/rds/message.c 		sock_put(rds_rs_to_sk(rs));
rs                 97 net/rds/rdma.c 	struct rds_sock *rs = mr->r_sock;
rs                107 net/rds/rdma.c 	spin_lock_irqsave(&rs->rs_rdma_lock, flags);
rs                109 net/rds/rdma.c 		rb_erase(&mr->r_rb_node, &rs->rs_rdma_keys);
rs                112 net/rds/rdma.c 	spin_unlock_irqrestore(&rs->rs_rdma_lock, flags);
rs                128 net/rds/rdma.c void rds_rdma_drop_keys(struct rds_sock *rs)
rs                135 net/rds/rdma.c 	spin_lock_irqsave(&rs->rs_rdma_lock, flags);
rs                136 net/rds/rdma.c 	while ((node = rb_first(&rs->rs_rdma_keys))) {
rs                138 net/rds/rdma.c 		if (mr->r_trans == rs->rs_transport)
rs                140 net/rds/rdma.c 		rb_erase(&mr->r_rb_node, &rs->rs_rdma_keys);
rs                142 net/rds/rdma.c 		spin_unlock_irqrestore(&rs->rs_rdma_lock, flags);
rs                145 net/rds/rdma.c 		spin_lock_irqsave(&rs->rs_rdma_lock, flags);
rs                147 net/rds/rdma.c 	spin_unlock_irqrestore(&rs->rs_rdma_lock, flags);
rs                149 net/rds/rdma.c 	if (rs->rs_transport && rs->rs_transport->flush_mrs)
rs                150 net/rds/rdma.c 		rs->rs_transport->flush_mrs();
rs                173 net/rds/rdma.c static int __rds_rdma_map(struct rds_sock *rs, struct rds_get_mr_args *args,
rs                188 net/rds/rdma.c 	if (ipv6_addr_any(&rs->rs_bound_addr) || !rs->rs_transport) {
rs                193 net/rds/rdma.c 	if (!rs->rs_transport->get_mr) {
rs                230 net/rds/rdma.c 	mr->r_trans = rs->rs_transport;
rs                231 net/rds/rdma.c 	mr->r_sock = rs;
rs                273 net/rds/rdma.c 	trans_private = rs->rs_transport->get_mr(sg, nents, rs,
rs                305 net/rds/rdma.c 	spin_lock_irqsave(&rs->rs_rdma_lock, flags);
rs                306 net/rds/rdma.c 	found = rds_mr_tree_walk(&rs->rs_rdma_keys, mr->r_key, mr);
rs                307 net/rds/rdma.c 	spin_unlock_irqrestore(&rs->rs_rdma_lock, flags);
rs                325 net/rds/rdma.c int rds_get_mr(struct rds_sock *rs, char __user *optval, int optlen)
rs                336 net/rds/rdma.c 	return __rds_rdma_map(rs, &args, NULL, NULL, NULL);
rs                339 net/rds/rdma.c int rds_get_mr_for_dest(struct rds_sock *rs, char __user *optval, int optlen)
rs                360 net/rds/rdma.c 	return __rds_rdma_map(rs, &new_args, NULL, NULL, NULL);
rs                366 net/rds/rdma.c int rds_free_mr(struct rds_sock *rs, char __user *optval, int optlen)
rs                381 net/rds/rdma.c 		if (!rs->rs_transport || !rs->rs_transport->flush_mrs)
rs                383 net/rds/rdma.c 		rs->rs_transport->flush_mrs();
rs                391 net/rds/rdma.c 	spin_lock_irqsave(&rs->rs_rdma_lock, flags);
rs                392 net/rds/rdma.c 	mr = rds_mr_tree_walk(&rs->rs_rdma_keys, rds_rdma_cookie_key(args.cookie), NULL);
rs                394 net/rds/rdma.c 		rb_erase(&mr->r_rb_node, &rs->rs_rdma_keys);
rs                399 net/rds/rdma.c 	spin_unlock_irqrestore(&rs->rs_rdma_lock, flags);
rs                419 net/rds/rdma.c void rds_rdma_unuse(struct rds_sock *rs, u32 r_key, int force)
rs                425 net/rds/rdma.c 	spin_lock_irqsave(&rs->rs_rdma_lock, flags);
rs                426 net/rds/rdma.c 	mr = rds_mr_tree_walk(&rs->rs_rdma_keys, r_key, NULL);
rs                430 net/rds/rdma.c 		spin_unlock_irqrestore(&rs->rs_rdma_lock, flags);
rs                435 net/rds/rdma.c 		rb_erase(&mr->r_rb_node, &rs->rs_rdma_keys);
rs                439 net/rds/rdma.c 	spin_unlock_irqrestore(&rs->rs_rdma_lock, flags);
rs                572 net/rds/rdma.c int rds_cmsg_rdma_args(struct rds_sock *rs, struct rds_message *rm,
rs                591 net/rds/rdma.c 	if (ipv6_addr_any(&rs->rs_bound_addr)) {
rs                625 net/rds/rdma.c 	op->op_recverr = rs->rs_recverr;
rs                670 net/rds/rdma.c 		rs->rs_user_addr = iov->addr;
rs                671 net/rds/rdma.c 		rs->rs_user_bytes = iov->bytes;
rs                730 net/rds/rdma.c int rds_cmsg_rdma_dest(struct rds_sock *rs, struct rds_message *rm,
rs                751 net/rds/rdma.c 	spin_lock_irqsave(&rs->rs_rdma_lock, flags);
rs                752 net/rds/rdma.c 	mr = rds_mr_tree_walk(&rs->rs_rdma_keys, r_key, NULL);
rs                757 net/rds/rdma.c 	spin_unlock_irqrestore(&rs->rs_rdma_lock, flags);
rs                772 net/rds/rdma.c int rds_cmsg_rdma_map(struct rds_sock *rs, struct rds_message *rm,
rs                779 net/rds/rdma.c 	return __rds_rdma_map(rs, CMSG_DATA(cmsg), &rm->m_rdma_cookie,
rs                786 net/rds/rdma.c int rds_cmsg_atomic(struct rds_sock *rs, struct rds_message *rm,
rs                832 net/rds/rdma.c 	rm->atomic.op_recverr = rs->rs_recverr;
rs                 97 net/rds/rds.h  #define	RDS_MPATH_HASH(rs, n) (jhash_1word((rs)->rs_bound_port, \
rs                 98 net/rds/rds.h  			       (rs)->rs_hash_initval) & ((n) - 1))
rs                575 net/rds/rds.h  			struct rds_sock *rs, u32 *key_ret,
rs                670 net/rds/rds.h  static inline struct sock *rds_rs_to_sk(struct rds_sock *rs)
rs                672 net/rds/rds.h  	return &rs->rs_sk;
rs                680 net/rds/rds.h  static inline int rds_sk_sndbuf(struct rds_sock *rs)
rs                682 net/rds/rds.h  	return rds_rs_to_sk(rs)->sk_sndbuf / 2;
rs                684 net/rds/rds.h  static inline int rds_sk_rcvbuf(struct rds_sock *rs)
rs                686 net/rds/rds.h  	return rds_rs_to_sk(rs)->sk_rcvbuf / 2;
rs                729 net/rds/rds.h  void rds_sock_addref(struct rds_sock *rs);
rs                730 net/rds/rds.h  void rds_sock_put(struct rds_sock *rs);
rs                731 net/rds/rds.h  void rds_wake_sk_sleep(struct rds_sock *rs);
rs                744 net/rds/rds.h  void rds_remove_bound(struct rds_sock *rs);
rs                756 net/rds/rds.h  int rds_cong_wait(struct rds_cong_map *map, __be16 port, int nonblock, struct rds_sock *rs);
rs                899 net/rds/rds.h  void rds_clear_recv_queue(struct rds_sock *rs);
rs                900 net/rds/rds.h  int rds_notify_queue_get(struct rds_sock *rs, struct msghdr *msg);
rs                914 net/rds/rds.h  void rds_send_drop_to(struct rds_sock *rs, struct sockaddr_in6 *dest);
rs                924 net/rds/rds.h  void rds_rdma_unuse(struct rds_sock *rs, u32 r_key, int force);
rs                925 net/rds/rds.h  int rds_get_mr(struct rds_sock *rs, char __user *optval, int optlen);
rs                926 net/rds/rds.h  int rds_get_mr_for_dest(struct rds_sock *rs, char __user *optval, int optlen);
rs                927 net/rds/rds.h  int rds_free_mr(struct rds_sock *rs, char __user *optval, int optlen);
rs                928 net/rds/rds.h  void rds_rdma_drop_keys(struct rds_sock *rs);
rs                931 net/rds/rds.h  int rds_cmsg_rdma_dest(struct rds_sock *rs, struct rds_message *rm,
rs                933 net/rds/rds.h  int rds_cmsg_rdma_args(struct rds_sock *rs, struct rds_message *rm,
rs                936 net/rds/rds.h  int rds_cmsg_rdma_map(struct rds_sock *rs, struct rds_message *rm,
rs                942 net/rds/rds.h  int rds_cmsg_atomic(struct rds_sock *rs, struct rds_message *rm,
rs                 87 net/rds/recv.c static void rds_recv_rcvbuf_delta(struct rds_sock *rs, struct sock *sk,
rs                 96 net/rds/recv.c 	rs->rs_rcv_bytes += delta;
rs                103 net/rds/recv.c 	if (rs->rs_transport->t_type == RDS_TRANS_LOOP)
rs                106 net/rds/recv.c 	now_congested = rs->rs_rcv_bytes > rds_sk_rcvbuf(rs);
rs                110 net/rds/recv.c 	  rs, &rs->rs_bound_addr,
rs                111 net/rds/recv.c 	  ntohs(rs->rs_bound_port), rs->rs_rcv_bytes,
rs                112 net/rds/recv.c 	  rds_sk_rcvbuf(rs), now_congested, delta);
rs                115 net/rds/recv.c 	if (!rs->rs_congested && now_congested) {
rs                116 net/rds/recv.c 		rs->rs_congested = 1;
rs                123 net/rds/recv.c 	else if (rs->rs_congested && (rs->rs_rcv_bytes < (rds_sk_rcvbuf(rs)/2))) {
rs                124 net/rds/recv.c 		rs->rs_congested = 0;
rs                165 net/rds/recv.c static void rds_recv_incoming_exthdrs(struct rds_incoming *inc, struct rds_sock *rs)
rs                183 net/rds/recv.c 			rds_rdma_unuse(rs, be32_to_cpu(buffer.rdma.h_rdma_rkey), 0);
rs                286 net/rds/recv.c 	struct rds_sock *rs = NULL;
rs                362 net/rds/recv.c 	rs = rds_find_bound(daddr, inc->i_hdr.h_dport, conn->c_bound_if);
rs                363 net/rds/recv.c 	if (!rs) {
rs                369 net/rds/recv.c 	rds_recv_incoming_exthdrs(inc, rs);
rs                372 net/rds/recv.c 	sk = rds_rs_to_sk(rs);
rs                375 net/rds/recv.c 	write_lock_irqsave(&rs->rs_recv_lock, flags);
rs                377 net/rds/recv.c 		rdsdebug("adding inc %p to rs %p's recv queue\n", inc, rs);
rs                379 net/rds/recv.c 		rds_recv_rcvbuf_delta(rs, sk, inc->i_conn->c_lcong,
rs                386 net/rds/recv.c 		list_add_tail(&inc->i_item, &rs->rs_recv_queue);
rs                391 net/rds/recv.c 	write_unlock_irqrestore(&rs->rs_recv_lock, flags);
rs                394 net/rds/recv.c 	if (rs)
rs                395 net/rds/recv.c 		rds_sock_put(rs);
rs                403 net/rds/recv.c static int rds_next_incoming(struct rds_sock *rs, struct rds_incoming **inc)
rs                408 net/rds/recv.c 		read_lock_irqsave(&rs->rs_recv_lock, flags);
rs                409 net/rds/recv.c 		if (!list_empty(&rs->rs_recv_queue)) {
rs                410 net/rds/recv.c 			*inc = list_entry(rs->rs_recv_queue.next,
rs                415 net/rds/recv.c 		read_unlock_irqrestore(&rs->rs_recv_lock, flags);
rs                421 net/rds/recv.c static int rds_still_queued(struct rds_sock *rs, struct rds_incoming *inc,
rs                424 net/rds/recv.c 	struct sock *sk = rds_rs_to_sk(rs);
rs                428 net/rds/recv.c 	write_lock_irqsave(&rs->rs_recv_lock, flags);
rs                433 net/rds/recv.c 			rds_recv_rcvbuf_delta(rs, sk, inc->i_conn->c_lcong,
rs                440 net/rds/recv.c 	write_unlock_irqrestore(&rs->rs_recv_lock, flags);
rs                442 net/rds/recv.c 	rdsdebug("inc %p rs %p still %d dropped %d\n", inc, rs, ret, drop);
rs                450 net/rds/recv.c int rds_notify_queue_get(struct rds_sock *rs, struct msghdr *msghdr)
rs                473 net/rds/recv.c 	spin_lock_irqsave(&rs->rs_lock, flags);
rs                474 net/rds/recv.c 	while (!list_empty(&rs->rs_notify_queue) && count < max_messages) {
rs                475 net/rds/recv.c 		notifier = list_entry(rs->rs_notify_queue.next,
rs                480 net/rds/recv.c 	spin_unlock_irqrestore(&rs->rs_lock, flags);
rs                506 net/rds/recv.c 		spin_lock_irqsave(&rs->rs_lock, flags);
rs                507 net/rds/recv.c 		list_splice(&copy, &rs->rs_notify_queue);
rs                508 net/rds/recv.c 		spin_unlock_irqrestore(&rs->rs_lock, flags);
rs                517 net/rds/recv.c static int rds_notify_cong(struct rds_sock *rs, struct msghdr *msghdr)
rs                519 net/rds/recv.c 	uint64_t notify = rs->rs_cong_notify;
rs                528 net/rds/recv.c 	spin_lock_irqsave(&rs->rs_lock, flags);
rs                529 net/rds/recv.c 	rs->rs_cong_notify &= ~notify;
rs                530 net/rds/recv.c 	spin_unlock_irqrestore(&rs->rs_lock, flags);
rs                539 net/rds/recv.c 			 struct rds_sock *rs)
rs                552 net/rds/recv.c 	    sock_flag(rds_rs_to_sk(rs), SOCK_RCVTSTAMP)) {
rs                556 net/rds/recv.c 		if (!sock_flag(rds_rs_to_sk(rs), SOCK_TSTAMP_NEW)) {
rs                573 net/rds/recv.c 	if (rs->rs_rx_traces) {
rs                579 net/rds/recv.c 		t.rx_traces =  rs->rs_rx_traces;
rs                580 net/rds/recv.c 		for (i = 0; i < rs->rs_rx_traces; i++) {
rs                581 net/rds/recv.c 			j = rs->rs_rx_trace[i];
rs                597 net/rds/recv.c static bool rds_recvmsg_zcookie(struct rds_sock *rs, struct msghdr *msg)
rs                599 net/rds/recv.c 	struct rds_msg_zcopy_queue *q = &rs->rs_zcookie_queue;
rs                607 net/rds/recv.c 	if (!sock_flag(rds_rs_to_sk(rs), SOCK_ZEROCOPY) ||
rs                636 net/rds/recv.c 	struct rds_sock *rs = rds_sk_to_rs(sk);
rs                655 net/rds/recv.c 		if (!list_empty(&rs->rs_notify_queue)) {
rs                656 net/rds/recv.c 			ret = rds_notify_queue_get(rs, msg);
rs                660 net/rds/recv.c 		if (rs->rs_cong_notify) {
rs                661 net/rds/recv.c 			ret = rds_notify_cong(rs, msg);
rs                665 net/rds/recv.c 		if (!rds_next_incoming(rs, &inc)) {
rs                667 net/rds/recv.c 				bool reaped = rds_recvmsg_zcookie(rs, msg);
rs                674 net/rds/recv.c 					(!list_empty(&rs->rs_notify_queue) ||
rs                675 net/rds/recv.c 					 rs->rs_cong_notify ||
rs                676 net/rds/recv.c 					 rds_next_incoming(rs, &inc)), timeo);
rs                700 net/rds/recv.c 		if (!rds_still_queued(rs, inc, !(msg_flags & MSG_PEEK))) {
rs                714 net/rds/recv.c 		if (rds_cmsg_recv(inc, msg, rs)) {
rs                718 net/rds/recv.c 		rds_recvmsg_zcookie(rs, msg);
rs                739 net/rds/recv.c 				sin6->sin6_scope_id = rs->rs_bound_scope_id;
rs                758 net/rds/recv.c void rds_clear_recv_queue(struct rds_sock *rs)
rs                760 net/rds/recv.c 	struct sock *sk = rds_rs_to_sk(rs);
rs                764 net/rds/recv.c 	write_lock_irqsave(&rs->rs_recv_lock, flags);
rs                765 net/rds/recv.c 	list_for_each_entry_safe(inc, tmp, &rs->rs_recv_queue, i_item) {
rs                766 net/rds/recv.c 		rds_recv_rcvbuf_delta(rs, sk, inc->i_conn->c_lcong,
rs                772 net/rds/recv.c 	write_unlock_irqrestore(&rs->rs_recv_lock, flags);
rs                473 net/rds/send.c static void rds_send_sndbuf_remove(struct rds_sock *rs, struct rds_message *rm)
rs                477 net/rds/send.c 	assert_spin_locked(&rs->rs_lock);
rs                479 net/rds/send.c 	BUG_ON(rs->rs_snd_bytes < len);
rs                480 net/rds/send.c 	rs->rs_snd_bytes -= len;
rs                482 net/rds/send.c 	if (rs->rs_snd_bytes == 0)
rs                502 net/rds/send.c 	struct rds_sock *rs = NULL;
rs                513 net/rds/send.c 		rs = rm->m_rs;
rs                514 net/rds/send.c 		sock_hold(rds_rs_to_sk(rs));
rs                517 net/rds/send.c 		spin_lock(&rs->rs_lock);
rs                518 net/rds/send.c 		list_add_tail(&notifier->n_list, &rs->rs_notify_queue);
rs                519 net/rds/send.c 		spin_unlock(&rs->rs_lock);
rs                526 net/rds/send.c 	if (rs) {
rs                527 net/rds/send.c 		rds_wake_sk_sleep(rs);
rs                528 net/rds/send.c 		sock_put(rds_rs_to_sk(rs));
rs                538 net/rds/send.c 	struct rds_sock *rs = NULL;
rs                549 net/rds/send.c 		rs = rm->m_rs;
rs                550 net/rds/send.c 		sock_hold(rds_rs_to_sk(rs));
rs                553 net/rds/send.c 		spin_lock(&rs->rs_lock);
rs                554 net/rds/send.c 		list_add_tail(&notifier->n_list, &rs->rs_notify_queue);
rs                555 net/rds/send.c 		spin_unlock(&rs->rs_lock);
rs                562 net/rds/send.c 	if (rs) {
rs                563 net/rds/send.c 		rds_wake_sk_sleep(rs);
rs                564 net/rds/send.c 		sock_put(rds_rs_to_sk(rs));
rs                575 net/rds/send.c __rds_send_complete(struct rds_sock *rs, struct rds_message *rm, int status)
rs                583 net/rds/send.c 		list_add_tail(&ro->op_notifier->n_list, &rs->rs_notify_queue);
rs                590 net/rds/send.c 		list_add_tail(&ao->op_notifier->n_list, &rs->rs_notify_queue);
rs                608 net/rds/send.c 	struct rds_sock *rs = NULL;
rs                632 net/rds/send.c 		if (rs != rm->m_rs) {
rs                633 net/rds/send.c 			if (rs) {
rs                634 net/rds/send.c 				rds_wake_sk_sleep(rs);
rs                635 net/rds/send.c 				sock_put(rds_rs_to_sk(rs));
rs                637 net/rds/send.c 			rs = rm->m_rs;
rs                638 net/rds/send.c 			if (rs)
rs                639 net/rds/send.c 				sock_hold(rds_rs_to_sk(rs));
rs                641 net/rds/send.c 		if (!rs)
rs                643 net/rds/send.c 		spin_lock(&rs->rs_lock);
rs                650 net/rds/send.c 			rds_send_sndbuf_remove(rs, rm);
rs                656 net/rds/send.c 						&rs->rs_notify_queue);
rs                663 net/rds/send.c 		spin_unlock(&rs->rs_lock);
rs                672 net/rds/send.c 	if (rs) {
rs                673 net/rds/send.c 		rds_wake_sk_sleep(rs);
rs                674 net/rds/send.c 		sock_put(rds_rs_to_sk(rs));
rs                722 net/rds/send.c void rds_send_drop_to(struct rds_sock *rs, struct sockaddr_in6 *dest)
rs                731 net/rds/send.c 	spin_lock_irqsave(&rs->rs_lock, flags);
rs                733 net/rds/send.c 	list_for_each_entry_safe(rm, tmp, &rs->rs_send_queue, m_sock_item) {
rs                740 net/rds/send.c 		rds_send_sndbuf_remove(rs, rm);
rs                747 net/rds/send.c 	spin_unlock_irqrestore(&rs->rs_lock, flags);
rs                780 net/rds/send.c 		spin_lock(&rs->rs_lock);
rs                781 net/rds/send.c 		__rds_send_complete(rs, rm, RDS_RDMA_CANCELED);
rs                782 net/rds/send.c 		spin_unlock(&rs->rs_lock);
rs                789 net/rds/send.c 	rds_wake_sk_sleep(rs);
rs                803 net/rds/send.c 		spin_lock(&rs->rs_lock);
rs                804 net/rds/send.c 		__rds_send_complete(rs, rm, RDS_RDMA_CANCELED);
rs                805 net/rds/send.c 		spin_unlock(&rs->rs_lock);
rs                818 net/rds/send.c static int rds_send_queue_rm(struct rds_sock *rs, struct rds_connection *conn,
rs                833 net/rds/send.c 	spin_lock_irqsave(&rs->rs_lock, flags);
rs                843 net/rds/send.c 	if (rs->rs_snd_bytes < rds_sk_sndbuf(rs)) {
rs                844 net/rds/send.c 		rs->rs_snd_bytes += len;
rs                851 net/rds/send.c 		if (rs->rs_snd_bytes >= rds_sk_sndbuf(rs) / 2)
rs                854 net/rds/send.c 		list_add_tail(&rm->m_sock_item, &rs->rs_send_queue);
rs                857 net/rds/send.c 		sock_hold(rds_rs_to_sk(rs));
rs                858 net/rds/send.c 		rm->m_rs = rs;
rs                874 net/rds/send.c 			 rm, len, rs, rs->rs_snd_bytes,
rs                880 net/rds/send.c 	spin_unlock_irqrestore(&rs->rs_lock, flags);
rs                971 net/rds/send.c static int rds_cmsg_zcopy(struct rds_sock *rs, struct rds_message *rm,
rs                984 net/rds/send.c static int rds_cmsg_send(struct rds_sock *rs, struct rds_message *rm,
rs               1005 net/rds/send.c 			ret = rds_cmsg_rdma_args(rs, rm, cmsg, &vct->vec[ind]);
rs               1010 net/rds/send.c 			ret = rds_cmsg_rdma_dest(rs, rm, cmsg);
rs               1014 net/rds/send.c 			ret = rds_cmsg_rdma_map(rs, rm, cmsg);
rs               1027 net/rds/send.c 			ret = rds_cmsg_atomic(rs, rm, cmsg);
rs               1031 net/rds/send.c 			ret = rds_cmsg_zcopy(rs, rm, cmsg);
rs               1045 net/rds/send.c static int rds_send_mprds_hash(struct rds_sock *rs,
rs               1051 net/rds/send.c 		hash = RDS_MPATH_HASH(rs, RDS_MPATH_WORKERS);
rs               1053 net/rds/send.c 		hash = RDS_MPATH_HASH(rs, conn->c_npaths);
rs               1104 net/rds/send.c 	struct rds_sock *rs = rds_sk_to_rs(sk);
rs               1119 net/rds/send.c 		      sock_flag(rds_rs_to_sk(rs), SOCK_ZEROCOPY));
rs               1204 net/rds/send.c 		daddr = rs->rs_conn_addr;
rs               1205 net/rds/send.c 		dport = rs->rs_conn_port;
rs               1206 net/rds/send.c 		scope_id = rs->rs_bound_scope_id;
rs               1211 net/rds/send.c 	if (ipv6_addr_any(&rs->rs_bound_addr) || ipv6_addr_any(&daddr)) {
rs               1221 net/rds/send.c 		    ipv6_addr_v4mapped(&rs->rs_bound_addr)) {
rs               1230 net/rds/send.c 		if (scope_id != rs->rs_bound_scope_id) {
rs               1232 net/rds/send.c 				scope_id = rs->rs_bound_scope_id;
rs               1233 net/rds/send.c 			} else if (rs->rs_bound_scope_id) {
rs               1252 net/rds/send.c 	if (payload_len > rds_sk_sndbuf(rs)) {
rs               1258 net/rds/send.c 		if (rs->rs_transport->t_type != RDS_TRANS_TCP) {
rs               1292 net/rds/send.c 	if (rs->rs_conn && ipv6_addr_equal(&rs->rs_conn->c_faddr, &daddr) &&
rs               1293 net/rds/send.c 	    rs->rs_tos == rs->rs_conn->c_tos) {
rs               1294 net/rds/send.c 		conn = rs->rs_conn;
rs               1297 net/rds/send.c 						&rs->rs_bound_addr, &daddr,
rs               1298 net/rds/send.c 						rs->rs_transport, rs->rs_tos,
rs               1305 net/rds/send.c 		rs->rs_conn = conn;
rs               1309 net/rds/send.c 		cpath = &conn->c_path[rds_send_mprds_hash(rs, conn, nonblock)];
rs               1316 net/rds/send.c 	ret = rds_cmsg_send(rs, rm, msg, &allocated_mr, &vct);
rs               1345 net/rds/send.c 	ret = rds_cong_wait(conn->c_fcong, dport, nonblock, rs);
rs               1347 net/rds/send.c 		rs->rs_seen_congestion = 1;
rs               1350 net/rds/send.c 	while (!rds_send_queue_rm(rs, conn, cpath, rm, rs->rs_bound_port,
rs               1360 net/rds/send.c 					rds_send_queue_rm(rs, conn, cpath, rm,
rs               1361 net/rds/send.c 							  rs->rs_bound_port,
rs               1410 net/rds/send.c 		rds_rdma_unuse(rs, rds_rdma_cookie_key(rm->m_rdma_cookie), 1);
rs                124 net/xfrm/xfrm_user.c 	struct xfrm_replay_state_esn *rs;
rs                129 net/xfrm/xfrm_user.c 	rs = nla_data(rt);
rs                131 net/xfrm/xfrm_user.c 	if (rs->bmp_len > XFRMA_REPLAY_ESN_MAX / sizeof(rs->bmp[0]) / 8)
rs                134 net/xfrm/xfrm_user.c 	if (nla_len(rt) < (int)xfrm_replay_state_esn_len(rs) &&
rs                135 net/xfrm/xfrm_user.c 	    nla_len(rt) != sizeof(*rs))
rs                251 tools/perf/builtin-stat.c 			       int thread, struct timespec *rs)
rs                254 tools/perf/builtin-stat.c 		u64 val = rs->tv_nsec + rs->tv_sec*1000000000ULL;
rs                268 tools/perf/builtin-stat.c static int read_counter(struct evsel *counter, struct timespec *rs)
rs                295 tools/perf/builtin-stat.c 			    read_single_counter(counter, cpu, thread, rs)) {
rs                324 tools/perf/builtin-stat.c static void read_counters(struct timespec *rs)
rs                330 tools/perf/builtin-stat.c 		ret = read_counter(counter, rs);
rs                341 tools/perf/builtin-stat.c 	struct timespec ts, rs;
rs                344 tools/perf/builtin-stat.c 	diff_timespec(&rs, &ts, &ref_time);
rs                346 tools/perf/builtin-stat.c 	read_counters(&rs);
rs                349 tools/perf/builtin-stat.c 		if (WRITE_STAT_ROUND_EVENT(rs.tv_sec * NSEC_PER_SEC + rs.tv_nsec, INTERVAL))
rs                355 tools/perf/builtin-stat.c 	print_counters(&rs, 0, NULL);
rs                483 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define tophys(rd,rs)				\
rs                484 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h 	addis	rd,rs,0
rs                486 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define tovirt(rd,rs)				\
rs                487 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h 	addis	rd,rs,0
rs                493 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define tophys(rd,rs)                           \
rs                494 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h 	clrldi	rd,rs,2
rs                496 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define tovirt(rd,rs)                           \
rs                497 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h 	rotldi	rd,rs,16;			\
rs                504 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define tophys(rd, rs)	addis	rd, rs, -PAGE_OFFSET@h
rs                505 tools/testing/selftests/powerpc/primitives/asm/ppc_asm.h #define tovirt(rd, rs)	addis	rd, rs, PAGE_OFFSET@h
rs                346 tools/testing/selftests/x86/protection_keys.c 	int r, rs;
rs                358 tools/testing/selftests/x86/protection_keys.c 	rs = sigprocmask(SIG_SETMASK, 0, &newact.sa_mask);
rs                359 tools/testing/selftests/x86/protection_keys.c 	pkey_assert(rs == 0);