cur                97 arch/alpha/kernel/srmcons.c 	char *cur;
cur               100 arch/alpha/kernel/srmcons.c 	for (cur = (char *)buf; remaining > 0; ) {
cur               107 arch/alpha/kernel/srmcons.c 			if (cur[c] == '\n')
cur               111 arch/alpha/kernel/srmcons.c 			result.as_long = callback_puts(0, cur, c);
cur               114 arch/alpha/kernel/srmcons.c 			cur += result.bits.c;
cur               247 arch/arc/kernel/kprobes.c 	struct kprobe *cur = kprobe_running();
cur               250 arch/arc/kernel/kprobes.c 	if (!cur)
cur               253 arch/arc/kernel/kprobes.c 	resume_execution(cur, addr, regs);
cur               256 arch/arc/kernel/kprobes.c 	arch_arm_kprobe(cur);
cur               265 arch/arc/kernel/kprobes.c 	if ((kcb->kprobe_status != KPROBE_REENTER) && cur->post_handler) {
cur               267 arch/arc/kernel/kprobes.c 		cur->post_handler(cur, regs, 0);
cur               291 arch/arc/kernel/kprobes.c 	struct kprobe *cur = kprobe_running();
cur               303 arch/arc/kernel/kprobes.c 		resume_execution(cur, (unsigned long)cur->addr, regs);
cur               324 arch/arc/kernel/kprobes.c 		kprobes_inc_nmissed_count(cur);
cur               333 arch/arc/kernel/kprobes.c 		if (cur->fault_handler && cur->fault_handler(cur, regs, trapnr))
cur               462 arch/arc/kernel/unwind.c 	const u8 *cur = *pcur;
cur               466 arch/arc/kernel/unwind.c 	for (shift = 0, value = 0; cur < end; shift += 7) {
cur               468 arch/arc/kernel/unwind.c 		    && (*cur & 0x7fU) >= (1U << (8 * sizeof(value) - shift))) {
cur               469 arch/arc/kernel/unwind.c 			cur = end + 1;
cur               472 arch/arc/kernel/unwind.c 		value |= (uleb128_t) (*cur & 0x7f) << shift;
cur               473 arch/arc/kernel/unwind.c 		if (!(*cur++ & 0x80))
cur               476 arch/arc/kernel/unwind.c 	*pcur = cur;
cur               483 arch/arc/kernel/unwind.c 	const u8 *cur = *pcur;
cur               487 arch/arc/kernel/unwind.c 	for (shift = 0, value = 0; cur < end; shift += 7) {
cur               489 arch/arc/kernel/unwind.c 		    && (*cur & 0x7fU) >= (1U << (8 * sizeof(value) - shift))) {
cur               490 arch/arc/kernel/unwind.c 			cur = end + 1;
cur               493 arch/arc/kernel/unwind.c 		value |= (sleb128_t) (*cur & 0x7f) << shift;
cur               494 arch/arc/kernel/unwind.c 		if (!(*cur & 0x80)) {
cur               495 arch/arc/kernel/unwind.c 			value |= -(*cur++ & 0x40) << shift;
cur               499 arch/arc/kernel/unwind.c 	*pcur = cur;
cur               965 arch/arc/kernel/unwind.c 					const u8 *cur =
cur               968 arch/arc/kernel/unwind.c 					startLoc = read_pointer(&cur,
cur               969 arch/arc/kernel/unwind.c 								cur + tableSize,
cur               974 arch/arc/kernel/unwind.c 						ptr = cur - tableSize;
cur                44 arch/arm/mach-integrator/impd1.c 	u32 cur;
cur                47 arch/arm/mach-integrator/impd1.c 	cur = readl(impd1->base + IMPD1_CTRL) & ~mask;
cur                48 arch/arm/mach-integrator/impd1.c 	writel(cur | val, impd1->base + IMPD1_CTRL);
cur               100 arch/arm/mach-rpc/dma.c 	unsigned int status, cur, end;
cur               113 arch/arm/mach-rpc/dma.c 			cur = CURA;
cur               116 arch/arm/mach-rpc/dma.c 			cur = CURB;
cur               119 arch/arm/mach-rpc/dma.c 		writel(idma->cur_addr, base + cur);
cur               267 arch/arm/mach-s3c24xx/mach-h1940.c 	{ .volt = 4070, .cur = 162, .level = 100},
cur               268 arch/arm/mach-s3c24xx/mach-h1940.c 	{ .volt = 4040, .cur = 165, .level = 95},
cur               269 arch/arm/mach-s3c24xx/mach-h1940.c 	{ .volt = 4016, .cur = 164, .level = 90},
cur               270 arch/arm/mach-s3c24xx/mach-h1940.c 	{ .volt = 3996, .cur = 166, .level = 85},
cur               271 arch/arm/mach-s3c24xx/mach-h1940.c 	{ .volt = 3971, .cur = 168, .level = 80},
cur               272 arch/arm/mach-s3c24xx/mach-h1940.c 	{ .volt = 3951, .cur = 168, .level = 75},
cur               273 arch/arm/mach-s3c24xx/mach-h1940.c 	{ .volt = 3931, .cur = 170, .level = 70},
cur               274 arch/arm/mach-s3c24xx/mach-h1940.c 	{ .volt = 3903, .cur = 172, .level = 65},
cur               275 arch/arm/mach-s3c24xx/mach-h1940.c 	{ .volt = 3886, .cur = 172, .level = 60},
cur               276 arch/arm/mach-s3c24xx/mach-h1940.c 	{ .volt = 3858, .cur = 176, .level = 55},
cur               277 arch/arm/mach-s3c24xx/mach-h1940.c 	{ .volt = 3842, .cur = 176, .level = 50},
cur               278 arch/arm/mach-s3c24xx/mach-h1940.c 	{ .volt = 3818, .cur = 176, .level = 45},
cur               279 arch/arm/mach-s3c24xx/mach-h1940.c 	{ .volt = 3789, .cur = 180, .level = 40},
cur               280 arch/arm/mach-s3c24xx/mach-h1940.c 	{ .volt = 3769, .cur = 180, .level = 35},
cur               281 arch/arm/mach-s3c24xx/mach-h1940.c 	{ .volt = 3749, .cur = 184, .level = 30},
cur               282 arch/arm/mach-s3c24xx/mach-h1940.c 	{ .volt = 3732, .cur = 184, .level = 25},
cur               283 arch/arm/mach-s3c24xx/mach-h1940.c 	{ .volt = 3716, .cur = 184, .level = 20},
cur               284 arch/arm/mach-s3c24xx/mach-h1940.c 	{ .volt = 3708, .cur = 184, .level = 15},
cur               285 arch/arm/mach-s3c24xx/mach-h1940.c 	{ .volt = 3716, .cur = 96, .level = 10},
cur               286 arch/arm/mach-s3c24xx/mach-h1940.c 	{ .volt = 3700, .cur = 96, .level = 5},
cur               287 arch/arm/mach-s3c24xx/mach-h1940.c 	{ .volt = 3684, .cur = 96, .level = 0},
cur               291 arch/arm/mach-s3c24xx/mach-h1940.c 	{ .volt = 4130, .cur = 0, .level = 100},
cur               292 arch/arm/mach-s3c24xx/mach-h1940.c 	{ .volt = 3982, .cur = 0, .level = 50},
cur               293 arch/arm/mach-s3c24xx/mach-h1940.c 	{ .volt = 3854, .cur = 0, .level = 10},
cur               294 arch/arm/mach-s3c24xx/mach-h1940.c 	{ .volt = 3841, .cur = 0, .level = 0},
cur               166 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 4100, .cur = 156, .level = 100},
cur               167 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 4050, .cur = 156, .level = 95},
cur               168 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 4025, .cur = 141, .level = 90},
cur               169 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3995, .cur = 144, .level = 85},
cur               170 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3957, .cur = 162, .level = 80},
cur               171 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3931, .cur = 147, .level = 75},
cur               172 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3902, .cur = 147, .level = 70},
cur               173 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3863, .cur = 153, .level = 65},
cur               174 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3838, .cur = 150, .level = 60},
cur               175 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3800, .cur = 153, .level = 55},
cur               176 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3765, .cur = 153, .level = 50},
cur               177 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3748, .cur = 172, .level = 45},
cur               178 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3740, .cur = 153, .level = 40},
cur               179 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3714, .cur = 175, .level = 35},
cur               180 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3710, .cur = 156, .level = 30},
cur               181 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3963, .cur = 156, .level = 25},
cur               182 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3672, .cur = 178, .level = 20},
cur               183 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3651, .cur = 178, .level = 15},
cur               184 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3629, .cur = 178, .level = 10},
cur               185 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3612, .cur = 162, .level = 5},
cur               186 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3605, .cur = 162, .level = 0},
cur               190 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 4200, .cur = 0, .level = 100},
cur               191 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 4190, .cur = 0, .level = 99},
cur               192 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 4178, .cur = 0, .level = 95},
cur               193 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 4110, .cur = 0, .level = 70},
cur               194 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 4076, .cur = 0, .level = 65},
cur               195 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 4046, .cur = 0, .level = 60},
cur               196 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 4021, .cur = 0, .level = 55},
cur               197 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3999, .cur = 0, .level = 50},
cur               198 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3982, .cur = 0, .level = 45},
cur               199 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3965, .cur = 0, .level = 40},
cur               200 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3957, .cur = 0, .level = 35},
cur               201 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3948, .cur = 0, .level = 30},
cur               202 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3936, .cur = 0, .level = 25},
cur               203 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3927, .cur = 0, .level = 20},
cur               204 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3906, .cur = 0, .level = 15},
cur               205 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3880, .cur = 0, .level = 10},
cur               206 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3829, .cur = 0, .level = 5},
cur               207 arch/arm/mach-s3c24xx/mach-rx1950.c 	{ .volt = 3820, .cur = 0, .level = 0},
cur                66 arch/arm/plat-samsung/adc.c 	struct s3c_adc_client	*cur;
cur               138 arch/arm/plat-samsung/adc.c 		adc->cur = next;
cur               171 arch/arm/plat-samsung/adc.c 	if (!adc->cur)
cur               257 arch/arm/plat-samsung/adc.c 	if (adc_dev->cur == client)
cur               258 arch/arm/plat-samsung/adc.c 		adc_dev->cur = NULL;
cur               272 arch/arm/plat-samsung/adc.c 	if (adc_dev->cur == NULL)
cur               283 arch/arm/plat-samsung/adc.c 	struct s3c_adc_client *client = adc->cur;
cur               318 arch/arm/plat-samsung/adc.c 		adc->cur = NULL;
cur               236 arch/arm/probes/kprobes/core.c 	struct kprobe *p, *cur;
cur               240 arch/arm/probes/kprobes/core.c 	cur = kprobe_running();
cur               265 arch/arm/probes/kprobes/core.c 		} else if (cur) {
cur               331 arch/arm/probes/kprobes/core.c 	struct kprobe *cur = kprobe_running();
cur               344 arch/arm/probes/kprobes/core.c 		regs->ARM_pc = (long)cur->addr;
cur               359 arch/arm/probes/kprobes/core.c 		kprobes_inc_nmissed_count(cur);
cur               368 arch/arm/probes/kprobes/core.c 		if (cur->fault_handler && cur->fault_handler(cur, regs, fsr))
cur               131 arch/arm64/kernel/alternative.c 	u64 cur, d_size, ctr_el0;
cur               136 arch/arm64/kernel/alternative.c 	cur = start & ~(d_size - 1);
cur               143 arch/arm64/kernel/alternative.c 		asm volatile("dc civac, %0" : : "r" (cur) : "memory");
cur               144 arch/arm64/kernel/alternative.c 	} while (cur += d_size, cur < end);
cur               472 arch/arm64/kernel/cpufeature.c 				s64 cur)
cur               481 arch/arm64/kernel/cpufeature.c 		ret = new < cur ? new : cur;
cur               484 arch/arm64/kernel/cpufeature.c 		if (!cur || !new)
cur               488 arch/arm64/kernel/cpufeature.c 		ret = new > cur ? new : cur;
cur               266 arch/arm64/kernel/probes/kprobes.c 	struct kprobe *cur = kprobe_running();
cur               268 arch/arm64/kernel/probes/kprobes.c 	if (!cur)
cur               272 arch/arm64/kernel/probes/kprobes.c 	if (cur->ainsn.api.restore != 0)
cur               273 arch/arm64/kernel/probes/kprobes.c 		instruction_pointer_set(regs, cur->ainsn.api.restore);
cur               282 arch/arm64/kernel/probes/kprobes.c 	if (cur->post_handler)	{
cur               286 arch/arm64/kernel/probes/kprobes.c 		cur->post_handler(cur, regs, 0);
cur               294 arch/arm64/kernel/probes/kprobes.c 	struct kprobe *cur = kprobe_running();
cur               307 arch/arm64/kernel/probes/kprobes.c 		instruction_pointer_set(regs, (unsigned long) cur->addr);
cur               326 arch/arm64/kernel/probes/kprobes.c 		kprobes_inc_nmissed_count(cur);
cur               335 arch/arm64/kernel/probes/kprobes.c 		if (cur->fault_handler && cur->fault_handler(cur, regs, fsr))
cur                14 arch/csky/kernel/cpu-probe.c 	unsigned int cur, next, i;
cur                20 arch/csky/kernel/cpu-probe.c 	cur  = mfcr("cr13");
cur                22 arch/csky/kernel/cpu-probe.c 		seq_printf(m, "product info[%d] : 0x%08x\n", i, cur);
cur                27 arch/csky/kernel/cpu-probe.c 		if (cur == next)
cur                30 arch/csky/kernel/cpu-probe.c 		cur = next;
cur               868 arch/ia64/kernel/kprobes.c 	struct kprobe *cur = kprobe_running();
cur               871 arch/ia64/kernel/kprobes.c 	if (!cur)
cur               874 arch/ia64/kernel/kprobes.c 	if ((kcb->kprobe_status != KPROBE_REENTER) && cur->post_handler) {
cur               876 arch/ia64/kernel/kprobes.c 		cur->post_handler(cur, regs, 0);
cur               879 arch/ia64/kernel/kprobes.c 	resume_execution(cur, regs);
cur               895 arch/ia64/kernel/kprobes.c 	struct kprobe *cur = kprobe_running();
cur               909 arch/ia64/kernel/kprobes.c 		regs->cr_iip = ((unsigned long)cur->addr) & ~0xFULL;
cur               910 arch/ia64/kernel/kprobes.c 		ia64_psr(regs)->ri = ((unsigned long)cur->addr) & 0xf;
cur               924 arch/ia64/kernel/kprobes.c 		kprobes_inc_nmissed_count(cur);
cur               933 arch/ia64/kernel/kprobes.c 		if (cur->fault_handler && cur->fault_handler(cur, regs, trapnr))
cur               107 arch/ia64/kernel/perfmon_default_smpl.c 	void *cur, *last;
cur               119 arch/ia64/kernel/perfmon_default_smpl.c 	cur         = buf+hdr->hdr_cur_offs;
cur               127 arch/ia64/kernel/perfmon_default_smpl.c 	if ((last - cur) < PFM_DEFAULT_MAX_ENTRY_SIZE) goto full;
cur               131 arch/ia64/kernel/perfmon_default_smpl.c 	ent = (pfm_default_smpl_entry_t *)cur;
cur               145 arch/ia64/kernel/perfmon_default_smpl.c 			cur, last,
cur               146 arch/ia64/kernel/perfmon_default_smpl.c 			last-cur,
cur               188 arch/ia64/kernel/perfmon_default_smpl.c 	cur               += entry_size;
cur               193 arch/ia64/kernel/perfmon_default_smpl.c 	if ((last - cur) < PFM_DEFAULT_MAX_ENTRY_SIZE) goto full;
cur               205 arch/ia64/kernel/perfmon_default_smpl.c 	DPRINT_ovfl(("sampling buffer full free=%lu, count=%lu, ovfl_notify=%d\n", last-cur, hdr->hdr_count, ovfl_notify));
cur                88 arch/m68k/sun3/sun3dvma.c 	struct list_head *cur;
cur                92 arch/m68k/sun3/sun3dvma.c 	list_for_each(cur, holes) {
cur                93 arch/m68k/sun3/sun3dvma.c 		hole = list_entry(cur, struct hole, list);
cur               111 arch/m68k/sun3/sun3dvma.c 	struct list_head *cur;
cur               114 arch/m68k/sun3/sun3dvma.c 	list_for_each(cur, &hole_list) {
cur               115 arch/m68k/sun3/sun3dvma.c 		hole = list_entry(cur, struct hole, list);
cur               155 arch/m68k/sun3/sun3dvma.c 	struct list_head *cur;
cur               167 arch/m68k/sun3/sun3dvma.c 	list_for_each(cur, &hole_list) {
cur               170 arch/m68k/sun3/sun3dvma.c 		hole = list_entry(cur, struct hole, list);
cur               208 arch/m68k/sun3/sun3dvma.c 	struct list_head *cur;
cur               222 arch/m68k/sun3/sun3dvma.c 	list_for_each(cur, &hole_list) {
cur               223 arch/m68k/sun3/sun3dvma.c 		hole = list_entry(cur, struct hole, list);
cur               244 arch/m68k/sun3/sun3dvma.c 	list_add(&(hole->list), cur);
cur                71 arch/mips/boot/elf2ecoff.c 	int remaining, cur, count;
cur                81 arch/mips/boot/elf2ecoff.c 		cur = remaining;
cur                82 arch/mips/boot/elf2ecoff.c 		if (cur > sizeof ibuf)
cur                83 arch/mips/boot/elf2ecoff.c 			cur = sizeof ibuf;
cur                84 arch/mips/boot/elf2ecoff.c 		remaining -= cur;
cur                85 arch/mips/boot/elf2ecoff.c 		if ((count = read(in, ibuf, cur)) != cur) {
cur                91 arch/mips/boot/elf2ecoff.c 		if ((count = write(out, ibuf, cur)) != cur) {
cur               146 arch/mips/cavium-octeon/csrc-octeon.c 	u64 cur, end, inc;
cur               148 arch/mips/cavium-octeon/csrc-octeon.c 	cur = read_c0_cvmcount();
cur               151 arch/mips/cavium-octeon/csrc-octeon.c 	end = cur + inc;
cur               153 arch/mips/cavium-octeon/csrc-octeon.c 	while (end > cur)
cur               154 arch/mips/cavium-octeon/csrc-octeon.c 		cur = read_c0_cvmcount();
cur               160 arch/mips/cavium-octeon/csrc-octeon.c 	u64 cur, end, inc;
cur               162 arch/mips/cavium-octeon/csrc-octeon.c 	cur = read_c0_cvmcount();
cur               165 arch/mips/cavium-octeon/csrc-octeon.c 	end = cur + inc;
cur               167 arch/mips/cavium-octeon/csrc-octeon.c 	while (end > cur)
cur               168 arch/mips/cavium-octeon/csrc-octeon.c 		cur = read_c0_cvmcount();
cur               174 arch/mips/cavium-octeon/csrc-octeon.c 	u64 cur, end;
cur               176 arch/mips/cavium-octeon/csrc-octeon.c 	cur = read_c0_cvmcount();
cur               177 arch/mips/cavium-octeon/csrc-octeon.c 	end = cur + loops;
cur               179 arch/mips/cavium-octeon/csrc-octeon.c 	while (end > cur)
cur               180 arch/mips/cavium-octeon/csrc-octeon.c 		cur = read_c0_cvmcount();
cur               195 arch/mips/cavium-octeon/csrc-octeon.c 	u64 cur, end;
cur               197 arch/mips/cavium-octeon/csrc-octeon.c 	cur = read_c0_cvmcount();
cur               207 arch/mips/cavium-octeon/csrc-octeon.c 		end = cur + end;
cur               209 arch/mips/cavium-octeon/csrc-octeon.c 		end = cur + count;
cur               211 arch/mips/cavium-octeon/csrc-octeon.c 	while (end > cur)
cur               212 arch/mips/cavium-octeon/csrc-octeon.c 		cur = read_c0_cvmcount();
cur               178 arch/mips/include/asm/sgiarcs.h 	struct linux_bigint   cur;
cur               374 arch/mips/kernel/kprobes.c 	struct kprobe *cur = kprobe_running();
cur               377 arch/mips/kernel/kprobes.c 	if (!cur)
cur               380 arch/mips/kernel/kprobes.c 	if ((kcb->kprobe_status != KPROBE_REENTER) && cur->post_handler) {
cur               382 arch/mips/kernel/kprobes.c 		cur->post_handler(cur, regs, 0);
cur               385 arch/mips/kernel/kprobes.c 	resume_execution(cur, regs, kcb);
cur               403 arch/mips/kernel/kprobes.c 	struct kprobe *cur = kprobe_running();
cur               406 arch/mips/kernel/kprobes.c 	if (cur->fault_handler && cur->fault_handler(cur, regs, trapnr))
cur               410 arch/mips/kernel/kprobes.c 		resume_execution(cur, regs, kcb);
cur               889 arch/mips/kvm/trap_emul.c 	unsigned int cur, change;
cur               974 arch/mips/kvm/trap_emul.c 		cur = kvm_read_c0_guest_config1(cop0);
cur               975 arch/mips/kvm/trap_emul.c 		change = (cur ^ v) & kvm_mips_config1_wrmask(vcpu);
cur               977 arch/mips/kvm/trap_emul.c 			v = cur ^ change;
cur               985 arch/mips/kvm/trap_emul.c 		cur = kvm_read_c0_guest_config3(cop0);
cur               986 arch/mips/kvm/trap_emul.c 		change = (cur ^ v) & kvm_mips_config3_wrmask(vcpu);
cur               988 arch/mips/kvm/trap_emul.c 			v = cur ^ change;
cur               993 arch/mips/kvm/trap_emul.c 		cur = kvm_read_c0_guest_config4(cop0);
cur               994 arch/mips/kvm/trap_emul.c 		change = (cur ^ v) & kvm_mips_config4_wrmask(vcpu);
cur               996 arch/mips/kvm/trap_emul.c 			v = cur ^ change;
cur              1001 arch/mips/kvm/trap_emul.c 		cur = kvm_read_c0_guest_config5(cop0);
cur              1002 arch/mips/kvm/trap_emul.c 		change = (cur ^ v) & kvm_mips_config5_wrmask(vcpu);
cur              1004 arch/mips/kvm/trap_emul.c 			v = cur ^ change;
cur              2050 arch/mips/kvm/vz.c 	unsigned int cur, change;
cur              2194 arch/mips/kvm/vz.c 		cur = read_gc0_config();
cur              2195 arch/mips/kvm/vz.c 		change = (cur ^ v) & kvm_vz_config_user_wrmask(vcpu);
cur              2197 arch/mips/kvm/vz.c 			v = cur ^ change;
cur              2204 arch/mips/kvm/vz.c 		cur = read_gc0_config1();
cur              2205 arch/mips/kvm/vz.c 		change = (cur ^ v) & kvm_vz_config1_user_wrmask(vcpu);
cur              2207 arch/mips/kvm/vz.c 			v = cur ^ change;
cur              2214 arch/mips/kvm/vz.c 		cur = read_gc0_config2();
cur              2215 arch/mips/kvm/vz.c 		change = (cur ^ v) & kvm_vz_config2_user_wrmask(vcpu);
cur              2217 arch/mips/kvm/vz.c 			v = cur ^ change;
cur              2224 arch/mips/kvm/vz.c 		cur = read_gc0_config3();
cur              2225 arch/mips/kvm/vz.c 		change = (cur ^ v) & kvm_vz_config3_user_wrmask(vcpu);
cur              2227 arch/mips/kvm/vz.c 			v = cur ^ change;
cur              2234 arch/mips/kvm/vz.c 		cur = read_gc0_config4();
cur              2235 arch/mips/kvm/vz.c 		change = (cur ^ v) & kvm_vz_config4_user_wrmask(vcpu);
cur              2237 arch/mips/kvm/vz.c 			v = cur ^ change;
cur              2244 arch/mips/kvm/vz.c 		cur = read_gc0_config5();
cur              2245 arch/mips/kvm/vz.c 		change = (cur ^ v) & kvm_vz_config5_user_wrmask(vcpu);
cur              2247 arch/mips/kvm/vz.c 			v = cur ^ change;
cur              1761 arch/mips/mm/tlbex.c 	int cur = pte;
cur              1769 arch/mips/mm/tlbex.c 				uasm_i_srl(p, t, cur, _PAGE_PRESENT_SHIFT);
cur              1770 arch/mips/mm/tlbex.c 				cur = t;
cur              1772 arch/mips/mm/tlbex.c 			uasm_i_andi(p, t, cur, 1);
cur              1780 arch/mips/mm/tlbex.c 			uasm_i_srl(p, t, cur, _PAGE_PRESENT_SHIFT);
cur              1781 arch/mips/mm/tlbex.c 			cur = t;
cur              1783 arch/mips/mm/tlbex.c 		uasm_i_andi(p, t, cur,
cur              1813 arch/mips/mm/tlbex.c 	int cur = pte;
cur              1816 arch/mips/mm/tlbex.c 		uasm_i_srl(p, t, cur, _PAGE_PRESENT_SHIFT);
cur              1817 arch/mips/mm/tlbex.c 		cur = t;
cur              1819 arch/mips/mm/tlbex.c 	uasm_i_andi(p, t, cur,
cur               482 arch/powerpc/kernel/kprobes.c 	struct kprobe *cur = kprobe_running();
cur               485 arch/powerpc/kernel/kprobes.c 	if (!cur || user_mode(regs))
cur               489 arch/powerpc/kernel/kprobes.c 	if (((unsigned long)cur->ainsn.insn + 4) != regs->nip)
cur               492 arch/powerpc/kernel/kprobes.c 	if ((kcb->kprobe_status != KPROBE_REENTER) && cur->post_handler) {
cur               494 arch/powerpc/kernel/kprobes.c 		cur->post_handler(cur, regs, 0);
cur               498 arch/powerpc/kernel/kprobes.c 	regs->nip = (unsigned long)cur->addr + 4;
cur               524 arch/powerpc/kernel/kprobes.c 	struct kprobe *cur = kprobe_running();
cur               538 arch/powerpc/kernel/kprobes.c 		regs->nip = (unsigned long)cur->addr;
cur               554 arch/powerpc/kernel/kprobes.c 		kprobes_inc_nmissed_count(cur);
cur               563 arch/powerpc/kernel/kprobes.c 		if (cur->fault_handler && cur->fault_handler(cur, regs, trapnr))
cur              3787 arch/powerpc/kvm/book3s_hv.c 	ktime_t cur, start_poll, start_wait;
cur              3793 arch/powerpc/kvm/book3s_hv.c 	cur = start_poll = ktime_get();
cur              3806 arch/powerpc/kvm/book3s_hv.c 			cur = ktime_get();
cur              3807 arch/powerpc/kvm/book3s_hv.c 		} while (single_task_running() && ktime_before(cur, stop));
cur              3841 arch/powerpc/kvm/book3s_hv.c 	cur = ktime_get();
cur              3844 arch/powerpc/kvm/book3s_hv.c 	block_ns = ktime_to_ns(cur) - ktime_to_ns(start_poll);
cur              3849 arch/powerpc/kvm/book3s_hv.c 			ktime_to_ns(cur) - ktime_to_ns(start_wait);
cur              3859 arch/powerpc/kvm/book3s_hv.c 				ktime_to_ns(cur) -
cur               258 arch/powerpc/kvm/book3s_xive.h 	u32 cur;
cur               262 arch/powerpc/kvm/book3s_xive.h 	cur = be32_to_cpup(qpage + *idx);
cur               263 arch/powerpc/kvm/book3s_xive.h 	if ((cur >> 31) == *toggle)
cur               268 arch/powerpc/kvm/book3s_xive.h 	return cur & 0x7fffffff;
cur                88 arch/powerpc/platforms/cell/cpufreq_spudemand.c 	if (!policy->cur) {
cur               667 arch/powerpc/platforms/powermac/pci.c 	int i, cur = -1;
cur               683 arch/powerpc/platforms/powermac/pci.c 			if (++cur >= 3) {
cur               687 arch/powerpc/platforms/powermac/pci.c 			hose->mem_resources[cur].flags = IORESOURCE_MEM;
cur               688 arch/powerpc/platforms/powermac/pci.c 			hose->mem_resources[cur].name = hose->dn->full_name;
cur               689 arch/powerpc/platforms/powermac/pci.c 			hose->mem_resources[cur].start = base;
cur               690 arch/powerpc/platforms/powermac/pci.c 			hose->mem_resources[cur].end = end;
cur               691 arch/powerpc/platforms/powermac/pci.c 			hose->mem_offset[cur] = 0;
cur               692 arch/powerpc/platforms/powermac/pci.c 			DBG("  %d: 0x%08lx-0x%08lx\n", cur, base, end);
cur               695 arch/powerpc/platforms/powermac/pci.c 			hose->mem_resources[cur].end = end;
cur               166 arch/powerpc/platforms/powernv/opal-powercap.c 		u32 cur, min, max;
cur               180 arch/powerpc/platforms/powernv/opal-powercap.c 		if (!of_property_read_u32(node, "powercap-current", &cur)) {
cur               214 arch/powerpc/platforms/powernv/opal-powercap.c 			powercap_add_attr(cur, "powercap-current",
cur                83 arch/powerpc/sysdev/xive/common.c 	u32 cur;
cur                87 arch/powerpc/sysdev/xive/common.c 	cur = be32_to_cpup(q->qpage + q->idx);
cur                90 arch/powerpc/sysdev/xive/common.c 	if ((cur >> 31) == q->toggle)
cur               103 arch/powerpc/sysdev/xive/common.c 	return cur & 0x7fffffff;
cur               564 arch/s390/kvm/vsie.c 	struct vsie_page *cur;
cur               583 arch/s390/kvm/vsie.c 		cur = page_to_virt(page);
cur               584 arch/s390/kvm/vsie.c 		if (READ_ONCE(cur->gmap) != gmap)
cur               586 arch/s390/kvm/vsie.c 		prefix = cur->scb_s.prefix << GUEST_PREFIX_SHIFT;
cur               588 arch/s390/kvm/vsie.c 		prefix += cur->scb_s.mso;
cur               590 arch/s390/kvm/vsie.c 			prefix_unmapped_sync(cur);
cur                73 arch/s390/mm/dump_pagetables.c 	unsigned int prot, cur;
cur                82 arch/s390/mm/dump_pagetables.c 	cur = st->current_prot;
cur                90 arch/s390/mm/dump_pagetables.c 	} else if (prot != cur || level != st->level ||
cur               241 arch/s390/numa/mode_emu.c 	struct toptree *cur, *tmp;
cur               243 arch/s390/numa/mode_emu.c 	toptree_for_each_safe(cur, tmp, phys, level) {
cur               246 arch/s390/numa/mode_emu.c 			if (cores_free == toptree_count(cur, CORE))
cur               247 arch/s390/numa/mode_emu.c 				toptree_move(cur, node);
cur               249 arch/s390/numa/mode_emu.c 			if (cores_free >= toptree_count(cur, CORE))
cur               250 arch/s390/numa/mode_emu.c 				toptree_move(cur, node);
cur               282 arch/s390/numa/toptree.c static struct toptree *toptree_next_sibling(struct toptree *cur)
cur               284 arch/s390/numa/toptree.c 	if (cur->parent == NULL)
cur               287 arch/s390/numa/toptree.c 	if (cur == list_last_entry(&cur->parent->children,
cur               290 arch/s390/numa/toptree.c 	return (struct toptree *) list_next_entry(cur, sibling);
cur               304 arch/s390/numa/toptree.c struct toptree *toptree_next(struct toptree *cur, struct toptree *context,
cur               309 arch/s390/numa/toptree.c 	if (!cur)
cur               315 arch/s390/numa/toptree.c 	tmp = toptree_next_sibling(cur);
cur               319 arch/s390/numa/toptree.c 	cur_context = cur;
cur               345 arch/s390/numa/toptree.c 	struct toptree *cur;
cur               348 arch/s390/numa/toptree.c 	toptree_for_each(cur, context, level)
cur                33 arch/s390/numa/toptree.h struct toptree *toptree_next(struct toptree *cur, struct toptree *context,
cur               365 arch/sh/kernel/kprobes.c 	struct kprobe *cur = kprobe_running();
cur               370 arch/sh/kernel/kprobes.c 	if (!cur)
cur               373 arch/sh/kernel/kprobes.c 	if ((kcb->kprobe_status != KPROBE_REENTER) && cur->post_handler) {
cur               375 arch/sh/kernel/kprobes.c 		cur->post_handler(cur, regs, 0);
cur               414 arch/sh/kernel/kprobes.c 	struct kprobe *cur = kprobe_running();
cur               428 arch/sh/kernel/kprobes.c 		regs->pc = (unsigned long)cur->addr;
cur               442 arch/sh/kernel/kprobes.c 		kprobes_inc_nmissed_count(cur);
cur               451 arch/sh/kernel/kprobes.c 		if (cur->fault_handler && cur->fault_handler(cur, regs, trapnr))
cur               295 arch/sparc/kernel/kprobes.c 	struct kprobe *cur = kprobe_running();
cur               298 arch/sparc/kernel/kprobes.c 	if (!cur)
cur               301 arch/sparc/kernel/kprobes.c 	if ((kcb->kprobe_status != KPROBE_REENTER) && cur->post_handler) {
cur               303 arch/sparc/kernel/kprobes.c 		cur->post_handler(cur, regs, 0);
cur               306 arch/sparc/kernel/kprobes.c 	resume_execution(cur, regs, kcb);
cur               322 arch/sparc/kernel/kprobes.c 	struct kprobe *cur = kprobe_running();
cur               336 arch/sparc/kernel/kprobes.c 		regs->tpc = (unsigned long)cur->addr;
cur               353 arch/sparc/kernel/kprobes.c 		kprobes_inc_nmissed_count(cur);
cur               362 arch/sparc/kernel/kprobes.c 		if (cur->fault_handler && cur->fault_handler(cur, regs, trapnr))
cur               662 arch/x86/events/intel/pt.c 		buf->first = buf->last = buf->cur = topa;
cur               784 arch/x86/events/intel/pt.c 	if (buf->cur_idx == buf->cur->last) {
cur               785 arch/x86/events/intel/pt.c 		if (buf->cur == buf->last)
cur               786 arch/x86/events/intel/pt.c 			buf->cur = buf->first;
cur               788 arch/x86/events/intel/pt.c 			buf->cur = list_entry(buf->cur->list.next, struct topa,
cur               806 arch/x86/events/intel/pt.c 	base = buf->cur->offset + buf->output_off;
cur               810 arch/x86/events/intel/pt.c 		base += TOPA_ENTRY_SIZE(buf->cur, topa_idx);
cur               830 arch/x86/events/intel/pt.c 	return phys_to_virt(TOPA_ENTRY(buf->cur, buf->cur_idx)->base << TOPA_SHIFT);
cur               839 arch/x86/events/intel/pt.c 	return TOPA_ENTRY_SIZE(buf->cur, buf->cur_idx);
cur               911 arch/x86/events/intel/pt.c 	buf->cur = &tp->topa;
cur              1118 arch/x86/events/intel/pt.c 	buf->cur = &cur_tp->topa;
cur              1119 arch/x86/events/intel/pt.c 	buf->cur_idx = te - TOPA_ENTRY(buf->cur, 0);
cur              1382 arch/x86/events/intel/pt.c 		pt_config_buffer(topa_to_page(buf->cur)->table, buf->cur_idx,
cur              1447 arch/x86/events/intel/pt.c 	pt_config_buffer(topa_to_page(buf->cur)->table, buf->cur_idx,
cur                76 arch/x86/events/intel/pt.h 	struct topa		*first, *last, *cur;
cur                30 arch/x86/hyperv/mmu.c 	unsigned long cur = start, diff;
cur                33 arch/x86/hyperv/mmu.c 		diff = end > cur ? end - cur : 0;
cur                35 arch/x86/hyperv/mmu.c 		gva_list[gva_n] = cur & PAGE_MASK;
cur                42 arch/x86/hyperv/mmu.c 			cur += HV_TLB_FLUSH_UNIT;
cur                45 arch/x86/hyperv/mmu.c 			cur = end;
cur                50 arch/x86/hyperv/mmu.c 	} while (cur < end);
cur                63 arch/x86/hyperv/nested.c 	u64 cur = start_gfn;
cur                79 arch/x86/hyperv/nested.c 		flush->gpa_list[gpa_n].page.basepfn = cur;
cur                82 arch/x86/hyperv/nested.c 		cur += additional_pages + 1;
cur                66 arch/x86/ia32/ia32_signal.c 	unsigned int cur = get_user_seg(seg);	\
cur                67 arch/x86/ia32/ia32_signal.c 	if (pre != cur)				\
cur               976 arch/x86/kernel/kprobes/core.c 	struct kprobe *cur = kprobe_running();
cur               979 arch/x86/kernel/kprobes/core.c 	if (!cur)
cur               982 arch/x86/kernel/kprobes/core.c 	resume_execution(cur, regs, kcb);
cur               985 arch/x86/kernel/kprobes/core.c 	if ((kcb->kprobe_status != KPROBE_REENTER) && cur->post_handler) {
cur               987 arch/x86/kernel/kprobes/core.c 		cur->post_handler(cur, regs, 0);
cur              1011 arch/x86/kernel/kprobes/core.c 	struct kprobe *cur = kprobe_running();
cur              1014 arch/x86/kernel/kprobes/core.c 	if (unlikely(regs->ip == (unsigned long)cur->ainsn.insn)) {
cur              1025 arch/x86/kernel/kprobes/core.c 		regs->ip = (unsigned long)cur->addr;
cur              1050 arch/x86/kernel/kprobes/core.c 		kprobes_inc_nmissed_count(cur);
cur              1059 arch/x86/kernel/kprobes/core.c 		if (cur->fault_handler && cur->fault_handler(cur, regs, trapnr))
cur                84 arch/x86/kernel/reboot_fixups_32.c 	const struct device_fixup *cur;
cur                94 arch/x86/kernel/reboot_fixups_32.c 		cur = &(fixups_table[i]);
cur                95 arch/x86/kernel/reboot_fixups_32.c 		dev = pci_get_device(cur->vendor, cur->device, NULL);
cur                99 arch/x86/kernel/reboot_fixups_32.c 		cur->reboot_fixup(dev);
cur                80 arch/x86/kernel/tsc_sync.c static void tsc_sanitize_first_cpu(struct tsc_adjust *cur, s64 bootval,
cur               110 arch/x86/kernel/tsc_sync.c 	cur->adjusted = bootval;
cur               116 arch/x86/kernel/tsc_sync.c 	struct tsc_adjust *cur = this_cpu_ptr(&tsc_adjust);
cur               127 arch/x86/kernel/tsc_sync.c 	cur->bootval = bootval;
cur               128 arch/x86/kernel/tsc_sync.c 	cur->nextcheck = jiffies + HZ;
cur               129 arch/x86/kernel/tsc_sync.c 	tsc_sanitize_first_cpu(cur, bootval, smp_processor_id(), bootcpu);
cur               140 arch/x86/kernel/tsc_sync.c 	struct tsc_adjust *ref, *cur = this_cpu_ptr(&tsc_adjust);
cur               149 arch/x86/kernel/tsc_sync.c 	cur->bootval = bootval;
cur               150 arch/x86/kernel/tsc_sync.c 	cur->nextcheck = jiffies + HZ;
cur               151 arch/x86/kernel/tsc_sync.c 	cur->warned = false;
cur               158 arch/x86/kernel/tsc_sync.c 		cur->adjusted = bootval;
cur               171 arch/x86/kernel/tsc_sync.c 		tsc_sanitize_first_cpu(cur, bootval, smp_processor_id(),
cur               191 arch/x86/kernel/tsc_sync.c 		cur->adjusted = ref->adjusted;
cur               402 arch/x86/kernel/tsc_sync.c 	struct tsc_adjust *cur = this_cpu_ptr(&tsc_adjust);
cur               484 arch/x86/kernel/tsc_sync.c 	cur->adjusted += cur_max_warp;
cur               487 arch/x86/kernel/tsc_sync.c 		cpu, cur_max_warp, cur->adjusted);
cur               489 arch/x86/kernel/tsc_sync.c 	wrmsrl(MSR_IA32_TSC_ADJUST, cur->adjusted);
cur               347 arch/x86/kvm/mtrr.c 	struct kvm_mtrr_range *tmp, *cur;
cur               352 arch/x86/kvm/mtrr.c 	cur = &mtrr_state->var_ranges[index];
cur               355 arch/x86/kvm/mtrr.c 	if (var_mtrr_range_is_valid(cur))
cur               363 arch/x86/kvm/mtrr.c 		cur->base = data;
cur               365 arch/x86/kvm/mtrr.c 		cur->mask = data | (-1LL << cpuid_maxphyaddr(vcpu));
cur               368 arch/x86/kvm/mtrr.c 	if (var_mtrr_range_is_valid(cur)) {
cur               370 arch/x86/kvm/mtrr.c 			if (cur->base >= tmp->base)
cur               372 arch/x86/kvm/mtrr.c 		list_add_tail(&cur->node, &tmp->node);
cur              5218 arch/x86/kvm/svm.c 	struct amd_svm_iommu_ir *cur;
cur              5221 arch/x86/kvm/svm.c 	list_for_each_entry(cur, &svm->ir_list, node) {
cur              5222 arch/x86/kvm/svm.c 		if (cur->data != pi->ir_data)
cur              5224 arch/x86/kvm/svm.c 		list_del(&cur->node);
cur              5225 arch/x86/kvm/svm.c 		kfree(cur);
cur               271 arch/x86/mm/dump_pagetables.c 	pgprotval_t prot, cur, eff;
cur               280 arch/x86/mm/dump_pagetables.c 	cur = pgprot_val(st->current_prot);
cur               292 arch/x86/mm/dump_pagetables.c 	} else if (prot != cur || new_eff != eff || level != st->level ||
cur              1658 block/blk-cgroup.c 		u64 cur = atomic64_read(&blkg->delay_nsec);
cur              1675 block/blk-cgroup.c 		if (unlikely(cur < sub)) {
cur              1680 block/blk-cgroup.c 			blkg->last_delay = cur - sub;
cur                11 block/blk-rq-qos.c 	unsigned int cur = atomic_read(v);
cur                16 block/blk-rq-qos.c 		if (cur >= below)
cur                18 block/blk-rq-qos.c 		old = atomic_cmpxchg(v, cur, cur + 1);
cur                19 block/blk-rq-qos.c 		if (old == cur)
cur                21 block/blk-rq-qos.c 		cur = old;
cur               111 block/blk-rq-qos.h 	struct rq_qos **cur;
cur               113 block/blk-rq-qos.h 	for (cur = &q->rq_qos; *cur; cur = &(*cur)->next) {
cur               114 block/blk-rq-qos.h 		if (*cur == rqos) {
cur               115 block/blk-rq-qos.h 			*cur = rqos->next;
cur                86 block/blk-wbt.c 		const unsigned long cur = jiffies;
cur                88 block/blk-wbt.c 		if (cur != *var)
cur                89 block/blk-wbt.c 			*var = cur;
cur               497 crypto/af_alg.c 	if (!sg || sgl->cur >= MAX_SGL_ENTS) {
cur               505 crypto/af_alg.c 		sgl->cur = 0;
cur               541 crypto/af_alg.c 		for (i = 0; i < sgl->cur; i++) {
cur               597 crypto/af_alg.c 		for (i = 0; i < sgl->cur; i++) {
cur               872 crypto/af_alg.c 			sg = sgl->sg + sgl->cur - 1;
cur               908 crypto/af_alg.c 		if (sgl->cur)
cur               909 crypto/af_alg.c 			sg_unmark_end(sg + sgl->cur - 1);
cur               912 crypto/af_alg.c 			unsigned int i = sgl->cur;
cur               935 crypto/af_alg.c 			sgl->cur++;
cur               936 crypto/af_alg.c 		} while (len && sgl->cur < MAX_SGL_ENTS);
cur               939 crypto/af_alg.c 			sg_mark_end(sg + sgl->cur - 1);
cur               993 crypto/af_alg.c 	if (sgl->cur)
cur               994 crypto/af_alg.c 		sg_unmark_end(sgl->sg + sgl->cur - 1);
cur               996 crypto/af_alg.c 	sg_mark_end(sgl->sg + sgl->cur);
cur               999 crypto/af_alg.c 	sg_set_page(sgl->sg + sgl->cur, page, size, offset);
cur              1000 crypto/af_alg.c 	sgl->cur++;
cur               183 crypto/algif_aead.c 		for (i = 0; i < tsgl->cur; i++) {
cur               357 crypto/asymmetric_keys/asym_tpm.c 	uint8_t *cur = buf;
cur               363 crypto/asymmetric_keys/asym_tpm.c 	cur = encode_tag_length(cur, 0x30, n_len + e_len);
cur               365 crypto/asymmetric_keys/asym_tpm.c 	cur = encode_tag_length(cur, 0x02, len + 1);
cur               366 crypto/asymmetric_keys/asym_tpm.c 	cur[0] = 0x00;
cur               367 crypto/asymmetric_keys/asym_tpm.c 	memcpy(cur + 1, pub_key, len);
cur               368 crypto/asymmetric_keys/asym_tpm.c 	cur += len + 1;
cur               369 crypto/asymmetric_keys/asym_tpm.c 	cur = encode_tag_length(cur, 0x02, sizeof(e));
cur               370 crypto/asymmetric_keys/asym_tpm.c 	memcpy(cur, e, sizeof(e));
cur               371 crypto/asymmetric_keys/asym_tpm.c 	cur += sizeof(e);
cur               373 crypto/asymmetric_keys/asym_tpm.c 	memset(cur, 0, SETKEY_PARAMS_SIZE);
cur               375 crypto/asymmetric_keys/asym_tpm.c 	return cur - buf;
cur               832 crypto/asymmetric_keys/asym_tpm.c 	const void *cur = tk->blob;
cur               842 crypto/asymmetric_keys/asym_tpm.c 	if (get_unaligned_be16(cur + 4) != 0x0015)
cur               846 crypto/asymmetric_keys/asym_tpm.c 	cur += 11;
cur               853 crypto/asymmetric_keys/asym_tpm.c 	if (get_unaligned_be32(cur) != 0x00000001)
cur               857 crypto/asymmetric_keys/asym_tpm.c 	if (get_unaligned_be16(cur + 4) != 0x0002)
cur               861 crypto/asymmetric_keys/asym_tpm.c 	if (get_unaligned_be16(cur + 6) != 0x0003)
cur               864 crypto/asymmetric_keys/asym_tpm.c 	sz = get_unaligned_be32(cur + 8);
cur               870 crypto/asymmetric_keys/asym_tpm.c 	cur += 12;
cur               873 crypto/asymmetric_keys/asym_tpm.c 	key_len = get_unaligned_be32(cur);
cur               886 crypto/asymmetric_keys/asym_tpm.c 	cur += sz;
cur               892 crypto/asymmetric_keys/asym_tpm.c 	sz = get_unaligned_be32(cur);
cur               897 crypto/asymmetric_keys/asym_tpm.c 	cur += 4 + sz;
cur               901 crypto/asymmetric_keys/asym_tpm.c 	sz = get_unaligned_be32(cur);
cur               905 crypto/asymmetric_keys/asym_tpm.c 	pub_key = cur + 4;
cur               376 crypto/tcrypt.c 				struct test_mb_aead_data *cur = &data[j];
cur               378 crypto/tcrypt.c 				assoc = cur->axbuf[0];
cur               381 crypto/tcrypt.c 				sg_init_aead(cur->sg, cur->xbuf,
cur               385 crypto/tcrypt.c 				sg_init_aead(cur->sgout, cur->xoutbuf,
cur               389 crypto/tcrypt.c 				aead_request_set_ad(cur->req, aad_size);
cur               393 crypto/tcrypt.c 					aead_request_set_crypt(cur->req,
cur               394 crypto/tcrypt.c 							       cur->sgout,
cur               395 crypto/tcrypt.c 							       cur->sg,
cur               397 crypto/tcrypt.c 					ret = crypto_aead_encrypt(cur->req);
cur               398 crypto/tcrypt.c 					ret = do_one_aead_op(cur->req, ret);
cur               407 crypto/tcrypt.c 				aead_request_set_crypt(cur->req, cur->sg,
cur               408 crypto/tcrypt.c 						       cur->sgout, *b_size +
cur              1354 crypto/tcrypt.c 				struct test_mb_skcipher_data *cur = &data[j];
cur              1359 crypto/tcrypt.c 				sg_init_table(cur->sg, pages);
cur              1362 crypto/tcrypt.c 					sg_set_buf(cur->sg + p, cur->xbuf[p],
cur              1364 crypto/tcrypt.c 					memset(cur->xbuf[p], 0xff, PAGE_SIZE);
cur              1369 crypto/tcrypt.c 				sg_set_buf(cur->sg + p, cur->xbuf[p], k);
cur              1370 crypto/tcrypt.c 				memset(cur->xbuf[p], 0xff, k);
cur              1372 crypto/tcrypt.c 				skcipher_request_set_crypt(cur->req, cur->sg,
cur              1373 crypto/tcrypt.c 							   cur->sg, *b_size,
cur               204 drivers/android/binder.c 	unsigned int cur = atomic_inc_return(&log->cur);
cur               206 drivers/android/binder.c 	if (cur >= ARRAY_SIZE(log->entry))
cur               208 drivers/android/binder.c 	e = &log->entry[cur % ARRAY_SIZE(log->entry)];
cur              6048 drivers/android/binder.c 	unsigned int log_cur = atomic_read(&log->cur);
cur              6050 drivers/android/binder.c 	unsigned int cur;
cur              6054 drivers/android/binder.c 	cur = count < ARRAY_SIZE(log->entry) && !log->full ?
cur              6059 drivers/android/binder.c 		unsigned int index = cur++ % ARRAY_SIZE(log->entry);
cur              6118 drivers/android/binder.c 	atomic_set(&binder_transaction_log.cur, ~0U);
cur              6119 drivers/android/binder.c 	atomic_set(&binder_transaction_log_failed.cur, ~0U);
cur               139 drivers/android/binder_internal.h 	atomic_t cur;
cur              3803 drivers/ata/libata-core.c 	u32 last, cur;
cur              3810 drivers/ata/libata-core.c 	if ((rc = sata_scr_read(link, SCR_STATUS, &cur)))
cur              3812 drivers/ata/libata-core.c 	cur &= 0xf;
cur              3814 drivers/ata/libata-core.c 	last = cur;
cur              3819 drivers/ata/libata-core.c 		if ((rc = sata_scr_read(link, SCR_STATUS, &cur)))
cur              3821 drivers/ata/libata-core.c 		cur &= 0xf;
cur              3824 drivers/ata/libata-core.c 		if (cur == last) {
cur              3825 drivers/ata/libata-core.c 			if (cur == 1 && time_before(jiffies, deadline))
cur              3834 drivers/ata/libata-core.c 		last = cur;
cur              6328 drivers/ata/libata-core.c 	const struct ata_port_operations *cur;
cur              6338 drivers/ata/libata-core.c 	for (cur = ops->inherits; cur; cur = cur->inherits) {
cur              6339 drivers/ata/libata-core.c 		void **inherit = (void **)cur;
cur              6895 drivers/ata/libata-core.c static int __init ata_parse_force_one(char **cur,
cur              6954 drivers/ata/libata-core.c 	char *start = *cur, *p = *cur;
cur              6964 drivers/ata/libata-core.c 		*cur = p;
cur              6966 drivers/ata/libata-core.c 		*cur = p + 1;
cur              7033 drivers/ata/libata-core.c 	char *p, *cur, *next;
cur              7048 drivers/ata/libata-core.c 	for (cur = ata_force_param_buf; *cur != '\0'; cur = next) {
cur              7052 drivers/ata/libata-core.c 		next = cur;
cur              7056 drivers/ata/libata-core.c 			       cur, reason);
cur               299 drivers/ata/libata-eh.c 		const u8 *cur;
cur               301 drivers/ata/libata-eh.c 		for (cur = ata_eh_cmd_timeout_table[i].commands; *cur; cur++)
cur               302 drivers/ata/libata-eh.c 			if (*cur == cmd)
cur               426 drivers/base/devres.c 	struct list_head *cur;
cur               431 drivers/base/devres.c 	cur = first;
cur               432 drivers/base/devres.c 	while (cur != end) {
cur               436 drivers/base/devres.c 		node = list_entry(cur, struct devres_node, entry);
cur               437 drivers/base/devres.c 		cur = cur->next;
cur               462 drivers/base/devres.c 	cur = first;
cur               463 drivers/base/devres.c 	while (cur != end) {
cur               467 drivers/base/devres.c 		node = list_entry(cur, struct devres_node, entry);
cur               468 drivers/base/devres.c 		cur = cur->next;
cur               705 drivers/base/regmap/regcache.c 					 unsigned int base, unsigned int cur)
cur               713 drivers/base/regmap/regcache.c 	count = (cur - base) / map->reg_stride;
cur               716 drivers/base/regmap/regcache.c 		count * val_bytes, count, base, cur - map->reg_stride);
cur               723 drivers/base/regmap/regcache.c 			base, cur - map->reg_stride, ret);
cur              1184 drivers/block/drbd/drbd_main.c 	len = bs.cur.b - p->code + !!bs.cur.bit;
cur              1200 drivers/block/drbd/drbd_main.c 	dcbp_set_pad_bits(p, (8 - bs.cur.bit) & 0x7);
cur              4769 drivers/block/drbd/drbd_receiver.c 				(unsigned int)(bs.cur.b - p->code),
cur               199 drivers/block/drbd/drbd_vli.h static inline void bitstream_cursor_reset(struct bitstream_cursor *cur, void *s)
cur               201 drivers/block/drbd/drbd_vli.h 	cur->b = s;
cur               202 drivers/block/drbd/drbd_vli.h 	cur->bit = 0;
cur               207 drivers/block/drbd/drbd_vli.h static inline void bitstream_cursor_advance(struct bitstream_cursor *cur, unsigned int bits)
cur               209 drivers/block/drbd/drbd_vli.h 	bits += cur->bit;
cur               210 drivers/block/drbd/drbd_vli.h 	cur->b = cur->b + (bits >> 3);
cur               211 drivers/block/drbd/drbd_vli.h 	cur->bit = bits & 7;
cur               216 drivers/block/drbd/drbd_vli.h 	struct bitstream_cursor cur;
cur               231 drivers/block/drbd/drbd_vli.h 	bitstream_cursor_reset(&bs->cur, bs->buf);
cur               236 drivers/block/drbd/drbd_vli.h 	bitstream_cursor_reset(&bs->cur, bs->buf);
cur               250 drivers/block/drbd/drbd_vli.h 	unsigned char *b = bs->cur.b;
cur               256 drivers/block/drbd/drbd_vli.h 	if ((bs->cur.b + ((bs->cur.bit + bits -1) >> 3)) - bs->buf >= bs->buf_len)
cur               263 drivers/block/drbd/drbd_vli.h 	*b++ |= (val & 0xff) << bs->cur.bit;
cur               265 drivers/block/drbd/drbd_vli.h 	for (tmp = 8 - bs->cur.bit; tmp < bits; tmp += 8)
cur               268 drivers/block/drbd/drbd_vli.h 	bitstream_cursor_advance(&bs->cur, bits);
cur               289 drivers/block/drbd/drbd_vli.h 	if (bs->cur.b + ((bs->cur.bit + bs->pad_bits + bits -1) >> 3) - bs->buf >= bs->buf_len)
cur               290 drivers/block/drbd/drbd_vli.h 		bits = ((bs->buf_len - (bs->cur.b - bs->buf)) << 3)
cur               291 drivers/block/drbd/drbd_vli.h 			- bs->cur.bit - bs->pad_bits;
cur               300 drivers/block/drbd/drbd_vli.h 	n = (bs->cur.bit + bits + 7) >> 3;
cur               304 drivers/block/drbd/drbd_vli.h 		memcpy(&val, bs->cur.b+1, n - 1);
cur               305 drivers/block/drbd/drbd_vli.h 		val = le64_to_cpu(val) << (8 - bs->cur.bit);
cur               309 drivers/block/drbd/drbd_vli.h 	val |= bs->cur.b[0] >> bs->cur.bit;
cur               314 drivers/block/drbd/drbd_vli.h 	bitstream_cursor_advance(&bs->cur, bits);
cur              2855 drivers/block/mtip32xx/mtip32xx.c 				attr242.cur, le32_to_cpu(attr242.data));
cur               166 drivers/block/mtip32xx/mtip32xx.h 	u8 cur;
cur                25 drivers/char/agp/isoch.c 	struct agp_3_5_dev *cur, *n = list_entry(new, struct agp_3_5_dev, list);
cur                29 drivers/char/agp/isoch.c 		cur = list_entry(pos, struct agp_3_5_dev, list);
cur                30 drivers/char/agp/isoch.c 		if (cur->maxbw > n->maxbw)
cur                38 drivers/char/agp/isoch.c 	struct agp_3_5_dev *cur;
cur                46 drivers/char/agp/isoch.c 		cur = list_entry(pos, struct agp_3_5_dev, list);
cur                47 drivers/char/agp/isoch.c 		dev = cur->dev;
cur                49 drivers/char/agp/isoch.c 		pci_read_config_dword(dev, cur->capndx+AGPNISTAT, &nistat);
cur                50 drivers/char/agp/isoch.c 		cur->maxbw = (nistat >> 16) & 0xff;
cur                82 drivers/char/agp/isoch.c 	struct agp_3_5_dev *cur;
cur               138 drivers/char/agp/isoch.c 		cur = list_entry(pos, struct agp_3_5_dev, list);
cur               139 drivers/char/agp/isoch.c 		dev = cur->dev;
cur               141 drivers/char/agp/isoch.c 		mcapndx = cur->capndx;
cur               143 drivers/char/agp/isoch.c 		pci_read_config_dword(dev, cur->capndx+AGPNISTAT, &mnistat);
cur               148 drivers/char/agp/isoch.c 		master[cdev].dev   = cur;
cur               251 drivers/char/agp/isoch.c 		cur = master[cdev].dev;
cur               252 drivers/char/agp/isoch.c 		dev = cur->dev;
cur               254 drivers/char/agp/isoch.c 		mcapndx = cur->capndx;
cur               259 drivers/char/agp/isoch.c 		pci_read_config_word(dev, cur->capndx+AGPNICMD, &mnicmd);
cur               260 drivers/char/agp/isoch.c 		pci_read_config_dword(dev, cur->capndx+AGPCMD, &mcmd);
cur               270 drivers/char/agp/isoch.c 		pci_write_config_dword(dev, cur->capndx+AGPCMD, mcmd);
cur               271 drivers/char/agp/isoch.c 		pci_write_config_word(dev, cur->capndx+AGPNICMD, mnicmd);
cur               291 drivers/char/agp/isoch.c 	struct agp_3_5_dev *cur;
cur               305 drivers/char/agp/isoch.c 		cur = list_entry(pos, struct agp_3_5_dev, list);
cur               307 drivers/char/agp/isoch.c 		pci_read_config_dword(cur->dev, cur->capndx+AGPCMD, &mcmd);
cur               310 drivers/char/agp/isoch.c 		pci_write_config_dword(cur->dev, cur->capndx+AGPCMD, mcmd);
cur               326 drivers/char/agp/isoch.c 	struct agp_3_5_dev *dev_list, *cur;
cur               372 drivers/char/agp/isoch.c 				if ((cur = kmalloc(sizeof(*cur), GFP_KERNEL)) == NULL) {
cur               376 drivers/char/agp/isoch.c 				cur->dev = dev;
cur               378 drivers/char/agp/isoch.c 				pos = &cur->list;
cur               395 drivers/char/agp/isoch.c 		cur = list_entry(pos, struct agp_3_5_dev, list);
cur               396 drivers/char/agp/isoch.c 		dev = cur->dev;
cur               429 drivers/char/agp/isoch.c 		cur->capndx = mcapndx;
cur               431 drivers/char/agp/isoch.c 		pci_read_config_dword(dev, cur->capndx+AGPSTAT, &mstatus);
cur               463 drivers/char/agp/isoch.c 		cur = list_entry(pos, struct agp_3_5_dev, list);
cur               466 drivers/char/agp/isoch.c 		kfree(cur);
cur               155 drivers/char/bsr.c 	struct bsr_dev *cur, *n;
cur               157 drivers/char/bsr.c 	list_for_each_entry_safe(cur, n, &bsr_devs, bsr_list) {
cur               158 drivers/char/bsr.c 		if (cur->bsr_device) {
cur               159 drivers/char/bsr.c 			cdev_del(&cur->bsr_cdev);
cur               160 drivers/char/bsr.c 			device_del(cur->bsr_device);
cur               162 drivers/char/bsr.c 		list_del(&cur->bsr_list);
cur               163 drivers/char/bsr.c 		kfree(cur);
cur               187 drivers/char/bsr.c 		struct bsr_dev *cur = kzalloc(sizeof(struct bsr_dev),
cur               192 drivers/char/bsr.c 		if (!cur) {
cur               201 drivers/char/bsr.c 			kfree(cur);
cur               205 drivers/char/bsr.c 		cur->bsr_minor  = i + total_bsr_devs;
cur               206 drivers/char/bsr.c 		cur->bsr_addr   = res.start;
cur               207 drivers/char/bsr.c 		cur->bsr_len    = resource_size(&res);
cur               208 drivers/char/bsr.c 		cur->bsr_bytes  = bsr_bytes[i];
cur               209 drivers/char/bsr.c 		cur->bsr_stride = bsr_stride[i];
cur               210 drivers/char/bsr.c 		cur->bsr_dev    = MKDEV(bsr_major, i + total_bsr_devs);
cur               214 drivers/char/bsr.c 		if (cur->bsr_len > 4096 && cur->bsr_len < PAGE_SIZE)
cur               215 drivers/char/bsr.c 			cur->bsr_len = 4096;
cur               217 drivers/char/bsr.c 		switch(cur->bsr_bytes) {
cur               219 drivers/char/bsr.c 			cur->bsr_type = BSR_8;
cur               222 drivers/char/bsr.c 			cur->bsr_type = BSR_16;
cur               225 drivers/char/bsr.c 			cur->bsr_type = BSR_64;
cur               228 drivers/char/bsr.c 			cur->bsr_type = BSR_128;
cur               231 drivers/char/bsr.c 			cur->bsr_type = BSR_4096;
cur               234 drivers/char/bsr.c 			cur->bsr_type = BSR_UNKNOWN;
cur               237 drivers/char/bsr.c 		cur->bsr_num = bsr_types[cur->bsr_type];
cur               238 drivers/char/bsr.c 		snprintf(cur->bsr_name, 32, "bsr%d_%d",
cur               239 drivers/char/bsr.c 			 cur->bsr_bytes, cur->bsr_num);
cur               241 drivers/char/bsr.c 		cdev_init(&cur->bsr_cdev, &bsr_fops);
cur               242 drivers/char/bsr.c 		result = cdev_add(&cur->bsr_cdev, cur->bsr_dev, 1);
cur               244 drivers/char/bsr.c 			kfree(cur);
cur               248 drivers/char/bsr.c 		cur->bsr_device = device_create(bsr_class, NULL, cur->bsr_dev,
cur               249 drivers/char/bsr.c 						cur, "%s", cur->bsr_name);
cur               250 drivers/char/bsr.c 		if (IS_ERR(cur->bsr_device)) {
cur               252 drivers/char/bsr.c 			       cur->bsr_name);
cur               253 drivers/char/bsr.c 			cdev_del(&cur->bsr_cdev);
cur               254 drivers/char/bsr.c 			kfree(cur);
cur               258 drivers/char/bsr.c 		bsr_types[cur->bsr_type] = cur->bsr_num + 1;
cur               259 drivers/char/bsr.c 		list_add_tail(&cur->bsr_list, &bsr_devs);
cur                80 drivers/clk/clk-conf.c 	const __be32 *cur;
cur                85 drivers/clk/clk-conf.c 	of_property_for_each_u32(node, "assigned-clock-rates", prop, cur, rate) {
cur              4826 drivers/clk/clk.c 	const __be32 *cur;
cur              4832 drivers/clk/clk.c 	of_property_for_each_u32(np, "clock-critical", prop, cur, idx)
cur               156 drivers/clk/rockchip/clk-pll.c 	struct rockchip_pll_rate_table cur;
cur               159 drivers/clk/rockchip/clk-pll.c 	rockchip_rk3036_pll_get_params(pll, &cur);
cur               161 drivers/clk/rockchip/clk-pll.c 	rate64 *= cur.fbdiv;
cur               162 drivers/clk/rockchip/clk-pll.c 	do_div(rate64, cur.refdiv);
cur               164 drivers/clk/rockchip/clk-pll.c 	if (cur.dsmpd == 0) {
cur               166 drivers/clk/rockchip/clk-pll.c 		u64 frac_rate64 = prate * cur.frac;
cur               168 drivers/clk/rockchip/clk-pll.c 		do_div(frac_rate64, cur.refdiv);
cur               172 drivers/clk/rockchip/clk-pll.c 	do_div(rate64, cur.postdiv1);
cur               173 drivers/clk/rockchip/clk-pll.c 	do_div(rate64, cur.postdiv2);
cur               183 drivers/clk/rockchip/clk-pll.c 	struct rockchip_pll_rate_table cur;
cur               193 drivers/clk/rockchip/clk-pll.c 	rockchip_rk3036_pll_get_params(pll, &cur);
cur               194 drivers/clk/rockchip/clk-pll.c 	cur.rate = 0;
cur               228 drivers/clk/rockchip/clk-pll.c 		rockchip_rk3036_pll_set_params(pll, &cur);
cur               289 drivers/clk/rockchip/clk-pll.c 	struct rockchip_pll_rate_table cur;
cur               302 drivers/clk/rockchip/clk-pll.c 	rockchip_rk3036_pll_get_params(pll, &cur);
cur               307 drivers/clk/rockchip/clk-pll.c 		 cur.fbdiv, cur.postdiv1, cur.refdiv, cur.postdiv2,
cur               308 drivers/clk/rockchip/clk-pll.c 		 cur.dsmpd, cur.frac);
cur               313 drivers/clk/rockchip/clk-pll.c 	if (rate->fbdiv != cur.fbdiv || rate->postdiv1 != cur.postdiv1 ||
cur               314 drivers/clk/rockchip/clk-pll.c 		rate->refdiv != cur.refdiv || rate->postdiv2 != cur.postdiv2 ||
cur               315 drivers/clk/rockchip/clk-pll.c 		rate->dsmpd != cur.dsmpd ||
cur               316 drivers/clk/rockchip/clk-pll.c 		(!cur.dsmpd && (rate->frac != cur.frac))) {
cur               391 drivers/clk/rockchip/clk-pll.c 	struct rockchip_pll_rate_table cur;
cur               402 drivers/clk/rockchip/clk-pll.c 	rockchip_rk3066_pll_get_params(pll, &cur);
cur               404 drivers/clk/rockchip/clk-pll.c 	rate64 *= cur.nf;
cur               405 drivers/clk/rockchip/clk-pll.c 	do_div(rate64, cur.nr);
cur               406 drivers/clk/rockchip/clk-pll.c 	do_div(rate64, cur.no);
cur               416 drivers/clk/rockchip/clk-pll.c 	struct rockchip_pll_rate_table cur;
cur               424 drivers/clk/rockchip/clk-pll.c 	rockchip_rk3066_pll_get_params(pll, &cur);
cur               425 drivers/clk/rockchip/clk-pll.c 	cur.rate = 0;
cur               461 drivers/clk/rockchip/clk-pll.c 		rockchip_rk3066_pll_set_params(pll, &cur);
cur               522 drivers/clk/rockchip/clk-pll.c 	struct rockchip_pll_rate_table cur;
cur               535 drivers/clk/rockchip/clk-pll.c 	rockchip_rk3066_pll_get_params(pll, &cur);
cur               538 drivers/clk/rockchip/clk-pll.c 		 __func__, clk_hw_get_name(hw), drate, rate->nr, cur.nr,
cur               539 drivers/clk/rockchip/clk-pll.c 		 rate->no, cur.no, rate->nf, cur.nf, rate->nb, cur.nb);
cur               540 drivers/clk/rockchip/clk-pll.c 	if (rate->nr != cur.nr || rate->no != cur.no || rate->nf != cur.nf
cur               541 drivers/clk/rockchip/clk-pll.c 						     || rate->nb != cur.nb) {
cur               633 drivers/clk/rockchip/clk-pll.c 	struct rockchip_pll_rate_table cur;
cur               636 drivers/clk/rockchip/clk-pll.c 	rockchip_rk3399_pll_get_params(pll, &cur);
cur               638 drivers/clk/rockchip/clk-pll.c 	rate64 *= cur.fbdiv;
cur               639 drivers/clk/rockchip/clk-pll.c 	do_div(rate64, cur.refdiv);
cur               641 drivers/clk/rockchip/clk-pll.c 	if (cur.dsmpd == 0) {
cur               643 drivers/clk/rockchip/clk-pll.c 		u64 frac_rate64 = prate * cur.frac;
cur               645 drivers/clk/rockchip/clk-pll.c 		do_div(frac_rate64, cur.refdiv);
cur               649 drivers/clk/rockchip/clk-pll.c 	do_div(rate64, cur.postdiv1);
cur               650 drivers/clk/rockchip/clk-pll.c 	do_div(rate64, cur.postdiv2);
cur               660 drivers/clk/rockchip/clk-pll.c 	struct rockchip_pll_rate_table cur;
cur               670 drivers/clk/rockchip/clk-pll.c 	rockchip_rk3399_pll_get_params(pll, &cur);
cur               671 drivers/clk/rockchip/clk-pll.c 	cur.rate = 0;
cur               707 drivers/clk/rockchip/clk-pll.c 		rockchip_rk3399_pll_set_params(pll, &cur);
cur               768 drivers/clk/rockchip/clk-pll.c 	struct rockchip_pll_rate_table cur;
cur               781 drivers/clk/rockchip/clk-pll.c 	rockchip_rk3399_pll_get_params(pll, &cur);
cur               786 drivers/clk/rockchip/clk-pll.c 		 cur.fbdiv, cur.postdiv1, cur.refdiv, cur.postdiv2,
cur               787 drivers/clk/rockchip/clk-pll.c 		 cur.dsmpd, cur.frac);
cur               792 drivers/clk/rockchip/clk-pll.c 	if (rate->fbdiv != cur.fbdiv || rate->postdiv1 != cur.postdiv1 ||
cur               793 drivers/clk/rockchip/clk-pll.c 		rate->refdiv != cur.refdiv || rate->postdiv2 != cur.postdiv2 ||
cur               794 drivers/clk/rockchip/clk-pll.c 		rate->dsmpd != cur.dsmpd ||
cur               795 drivers/clk/rockchip/clk-pll.c 		(!cur.dsmpd && (rate->frac != cur.frac))) {
cur               423 drivers/clocksource/samsung_pwm_timer.c 	const __be32 *cur;
cur               431 drivers/clocksource/samsung_pwm_timer.c 	of_property_for_each_u32(np, "samsung,pwm-outputs", prop, cur, val) {
cur               795 drivers/cpufreq/acpi-cpufreq.c 		policy->cur = acpi_cpufreq_guess_freq(data, policy->cpu);
cur                62 drivers/cpufreq/amd_freq_sensitivity.c 		freq_next = policy->cur;
cur                71 drivers/cpufreq/amd_freq_sensitivity.c 		freq_next = policy->cur;
cur                82 drivers/cpufreq/amd_freq_sensitivity.c 		if (data->freq_prev == policy->cur)
cur                83 drivers/cpufreq/amd_freq_sensitivity.c 			freq_next = policy->cur;
cur                85 drivers/cpufreq/amd_freq_sensitivity.c 		if (freq_next > policy->cur)
cur                86 drivers/cpufreq/amd_freq_sensitivity.c 			freq_next = policy->cur;
cur                87 drivers/cpufreq/amd_freq_sensitivity.c 		else if (freq_next < policy->cur)
cur                93 drivers/cpufreq/amd_freq_sensitivity.c 							   policy->cur - 1);
cur               600 drivers/cpufreq/brcmstb-avs-cpufreq.c 			policy->cur = freq_table[pstate].frequency;
cur               206 drivers/cpufreq/cppc_cpufreq.c 	freqs.old = policy->cur;
cur               343 drivers/cpufreq/cppc_cpufreq.c 	policy->cur = cppc_cpufreq_perf_to_khz(cpu,
cur               361 drivers/cpufreq/cpufreq.c 		if (policy->cur && policy->cur != freqs->old) {
cur               363 drivers/cpufreq/cpufreq.c 				 freqs->old, policy->cur);
cur               364 drivers/cpufreq/cpufreq.c 			freqs->old = policy->cur;
cur               385 drivers/cpufreq/cpufreq.c 		policy->cur = freqs->new;
cur               700 drivers/cpufreq/cpufreq.c 		ret = sprintf(buf, "%u\n", policy->cur);
cur              1411 drivers/cpufreq/cpufreq.c 		policy->cur = cpufreq_driver->get(policy->cpu);
cur              1412 drivers/cpufreq/cpufreq.c 		if (!policy->cur) {
cur              1439 drivers/cpufreq/cpufreq.c 		ret = cpufreq_frequency_table_get_index(policy, policy->cur);
cur              1443 drivers/cpufreq/cpufreq.c 				__func__, policy->cpu, policy->cur);
cur              1444 drivers/cpufreq/cpufreq.c 			ret = __cpufreq_driver_target(policy, policy->cur - 1,
cur              1454 drivers/cpufreq/cpufreq.c 				__func__, policy->cpu, policy->cur);
cur              1645 drivers/cpufreq/cpufreq.c 		 policy->cur, new_freq);
cur              1647 drivers/cpufreq/cpufreq.c 	freqs.old = policy->cur;
cur              1669 drivers/cpufreq/cpufreq.c 	if (policy->cur != new_freq) {
cur              1703 drivers/cpufreq/cpufreq.c 		ret_freq = policy->cur;
cur              2058 drivers/cpufreq/cpufreq.c 	struct cpufreq_freqs freqs = {.old = policy->cur, .flags = 0};
cur              2064 drivers/cpufreq/cpufreq.c 	if (newfreq == policy->cur)
cur              2135 drivers/cpufreq/cpufreq.c 	if (target_freq == policy->cur)
cur              2139 drivers/cpufreq/cpufreq.c 	policy->restore_freq = policy->cur;
cur                81 drivers/cpufreq/cpufreq_conservative.c 		requested_freq = policy->cur;
cur               311 drivers/cpufreq/cpufreq_conservative.c 	dbs_info->requested_freq = policy->cur;
cur               513 drivers/cpufreq/cpufreq_governor.c 	if (!policy->cur)
cur               122 drivers/cpufreq/cpufreq_ondemand.c 	else if (policy->cur == policy->max)
cur               147 drivers/cpufreq/cpufreq_ondemand.c 		if (policy->cur < policy->max)
cur               210 drivers/cpufreq/cpufreq_stats.c 	stats->last_index = freq_table_get_index(stats, policy->cur);
cur                49 drivers/cpufreq/cpufreq_userspace.c 	return sprintf(buf, "%u\n", policy->cur);
cur                76 drivers/cpufreq/cpufreq_userspace.c 	BUG_ON(!policy->cur);
cur                81 drivers/cpufreq/cpufreq_userspace.c 	*setspeed = policy->cur;
cur               105 drivers/cpufreq/cpufreq_userspace.c 		 policy->cpu, policy->min, policy->max, policy->cur, *setspeed);
cur                42 drivers/cpufreq/davinci-cpufreq.c 	old_freq = policy->cur;
cur              2330 drivers/cpufreq/intel_pstate.c 	freqs.old = policy->cur;
cur              2393 drivers/cpufreq/intel_pstate.c 	policy->cur = policy->cpuinfo.min_freq;
cur               963 drivers/cpufreq/longhaul.c 			freqs.old = policy->cur;
cur                55 drivers/cpufreq/loongson1-cpufreq.c 	old_freq = policy->cur;
cur                49 drivers/cpufreq/omap-cpufreq.c 	old_freq = policy->cur;
cur               194 drivers/cpufreq/pasemi-cpufreq.c 	policy->cur = pas_freqs[cur_astate].frequency;
cur               195 drivers/cpufreq/pasemi-cpufreq.c 	ppc_proc_freq = policy->cur * 1000ul;
cur               215 drivers/cpufreq/pcc-cpufreq.c 	freqs.old = policy->cur;
cur               230 drivers/cpufreq/powernow-k6.c 			freqs.old = policy->cur;
cur               981 drivers/cpufreq/powernow-k8.c 	pol->cur = find_khz_freq_from_fid(data->currfid);
cur               922 drivers/cpufreq/powernv-cpufreq.c 		index = cpufreq_table_find_index_c(&policy, policy.cur);
cur               106 drivers/cpufreq/ppc_cbe_cpufreq.c 	policy->cur = cbe_freqs[cur_pmode].frequency;
cur               203 drivers/cpufreq/pxa2xx-cpufreq.c 			 policy->cur / 1000,  new_freq_cpu / 1000);
cur               205 drivers/cpufreq/pxa2xx-cpufreq.c 	if (vcc_core && new_freq_cpu > policy->cur) {
cur               222 drivers/cpufreq/pxa2xx-cpufreq.c 	if (vcc_core && new_freq_cpu < policy->cur)
cur               242 drivers/cpufreq/s5pv210-cpufreq.c 	old_freq = policy->cur;
cur               166 drivers/cpufreq/sa1100-cpufreq.c 	unsigned int cur = sa11x0_getspeed(0);
cur               171 drivers/cpufreq/sa1100-cpufreq.c 	if (new_freq > cur)
cur               172 drivers/cpufreq/sa1100-cpufreq.c 		sa1100_update_dram_timings(cur, new_freq);
cur               176 drivers/cpufreq/sa1100-cpufreq.c 	if (new_freq < cur)
cur               177 drivers/cpufreq/sa1100-cpufreq.c 		sa1100_update_dram_timings(cur, new_freq);
cur               295 drivers/cpufreq/sparc-us2e-cpufreq.c 	policy->cur = clock_tick;
cur               139 drivers/cpufreq/sparc-us3-cpufreq.c 	policy->cur = clock_tick;
cur                50 drivers/cpufreq/tegra20-cpufreq.c 	if (freq_table[index].frequency == ifreq || policy->cur == ifreq)
cur                41 drivers/cpufreq/unicore2-cpufreq.c 	freqs.old = policy->cur;
cur               169 drivers/crypto/n2_core.c static void *spu_queue_next(struct spu_queue *q, void *cur)
cur               171 drivers/crypto/n2_core.c 	return q->q + spu_next_offset(q, cur - q->q);
cur               280 drivers/crypto/picoxcell_crypto.c 	struct scatterlist *cur;
cur               299 drivers/crypto/picoxcell_crypto.c 	for_each_sg(payload, cur, mapped_ents, i)
cur               300 drivers/crypto/picoxcell_crypto.c 		ddt_set(&ddt[i], sg_dma_address(cur), sg_dma_len(cur));
cur               318 drivers/crypto/picoxcell_crypto.c 	struct scatterlist *cur;
cur               381 drivers/crypto/picoxcell_crypto.c 	for_each_sg(areq->src, cur, src_ents, i)
cur               382 drivers/crypto/picoxcell_crypto.c 		ddt_set(src_ddt++, sg_dma_address(cur), sg_dma_len(cur));
cur               386 drivers/crypto/picoxcell_crypto.c 	for_each_sg(areq->dst, cur, dst_ents, i) {
cur               387 drivers/crypto/picoxcell_crypto.c 		unsigned len = sg_dma_len(cur);
cur               394 drivers/crypto/picoxcell_crypto.c 		ddt_set(dst_ddt++, sg_dma_address(cur) + total, len - total);
cur               477 drivers/crypto/qat/qat_common/qat_uclo.c 				 char *chunk_id, void *cur)
cur               485 drivers/crypto/qat/qat_common/qat_uclo.c 		if ((cur < (void *)&chunk_hdr[i]) &&
cur               131 drivers/dma-buf/dma-fence.c 	struct dma_fence_cb *cur, *tmp;
cur               147 drivers/dma-buf/dma-fence.c 	list_for_each_entry_safe(cur, tmp, &cb_list, node) {
cur               148 drivers/dma-buf/dma-fence.c 		INIT_LIST_HEAD(&cur->node);
cur               149 drivers/dma-buf/dma-fence.c 		cur->func(fence, cur);
cur                85 drivers/firmware/google/memconsole-x86-legacy.c 	size_t length, cur;
cur               101 drivers/firmware/google/memconsole-x86-legacy.c 	for (cur = 0; cur < length; cur++) {
cur               102 drivers/firmware/google/memconsole-x86-legacy.c 		struct biosmemcon_ebda *hdr = phys_to_virt(address + cur);
cur              4669 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c 	unsigned cur;
cur              4673 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c 	cur = (ring->wptr - 1) & ring->buf_mask;
cur              4674 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c 	if (likely(cur > offset))
cur              4675 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c 		ring->ring[offset] = cur - offset;
cur              4677 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c 		ring->ring[offset] = (ring->buf_mask + 1) - offset + cur;
cur              6473 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c 	unsigned cur;
cur              6478 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c 	cur = (ring->wptr & ring->buf_mask) - 1;
cur              6479 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c 	if (likely(cur > offset))
cur              6480 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c 		ring->ring[offset] = cur - offset;
cur              6482 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c 		ring->ring[offset] = (ring->ring_size >> 2) - offset + cur;
cur              5394 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c 	unsigned cur;
cur              5398 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c 	cur = (ring->wptr & ring->buf_mask) - 1;
cur              5399 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c 	if (likely(cur > offset))
cur              5400 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c 		ring->ring[offset] = cur - offset;
cur              5402 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c 		ring->ring[offset] = (ring->ring_size>>2) - offset + cur;
cur               249 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 	unsigned cur;
cur               254 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 	cur = (ring->wptr - 1) & ring->buf_mask;
cur               255 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 	if (cur > offset)
cur               256 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 		ring->ring[offset] = cur - offset;
cur               258 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c 		ring->ring[offset] = (ring->buf_mask + 1) - offset + cur;
cur               809 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c 	struct device_process_node *cur, *next;
cur               817 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c 	list_for_each_entry_safe(cur, next, &dqm->queues, list) {
cur               818 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c 		if (qpd == cur->qpd) {
cur               819 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c 			list_del(&cur->list);
cur               820 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c 			kfree(cur);
cur              1499 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c 	struct device_process_node *cur, *next_dpn;
cur              1515 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c 	list_for_each_entry_safe(cur, next_dpn, &dqm->queues, list) {
cur              1516 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c 		if (qpd == cur->qpd) {
cur              1517 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c 			list_del(&cur->list);
cur              1518 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c 			kfree(cur);
cur              1575 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c 	struct device_process_node *cur, *next_dpn;
cur              1610 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c 	list_for_each_entry_safe(cur, next_dpn, &dqm->queues, list) {
cur              1611 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c 		if (qpd == cur->qpd) {
cur              1612 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c 			list_del(&cur->list);
cur              1613 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c 			kfree(cur);
cur               127 drivers/gpu/drm/amd/amdkfd/kfd_packet_manager.c 	struct device_process_node *cur;
cur               147 drivers/gpu/drm/amd/amdkfd/kfd_packet_manager.c 	list_for_each_entry(cur, queues, list) {
cur               148 drivers/gpu/drm/amd/amdkfd/kfd_packet_manager.c 		qpd = cur->qpd;
cur              3679 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c 	struct dm_crtc_state *cur = to_dm_crtc_state(state);
cur              3682 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c 	if (cur->stream)
cur              3683 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c 		dc_stream_release(cur->stream);
cur              3711 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c 	struct dm_crtc_state *state, *cur;
cur              3713 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c 	cur = to_dm_crtc_state(crtc->state);
cur              3724 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c 	if (cur->stream) {
cur              3725 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c 		state->stream = cur->stream;
cur              3729 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c 	state->active_planes = cur->active_planes;
cur              3730 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c 	state->interrupts_enabled = cur->interrupts_enabled;
cur              3731 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c 	state->vrr_params = cur->vrr_params;
cur              3732 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c 	state->vrr_infopacket = cur->vrr_infopacket;
cur              3733 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c 	state->abm_level = cur->abm_level;
cur              3734 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c 	state->vrr_supported = cur->vrr_supported;
cur              3735 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c 	state->freesync_config = cur->freesync_config;
cur              3736 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c 	state->crc_src = cur->crc_src;
cur              3737 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c 	state->cm_has_degamma = cur->cm_has_degamma;
cur              3738 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c 	state->cm_is_degamma_srgb = cur->cm_is_degamma_srgb;
cur              1548 drivers/gpu/drm/amd/display/dc/core/dc_link_dp.c 	struct dc_link_settings *cur = &cur_link_setting;
cur              1588 drivers/gpu/drm/amd/display/dc/core/dc_link_dp.c 		if (cur->link_rate == LINK_RATE_LOW)
cur              1595 drivers/gpu/drm/amd/display/dc/core/dc_link_dp.c 				cur);
cur              1603 drivers/gpu/drm/amd/display/dc/core/dc_link_dp.c 							cur,
cur              1612 drivers/gpu/drm/amd/display/dc/core/dc_link_dp.c 			link->verified_link_cap = *cur;
cur              1626 drivers/gpu/drm/amd/display/dc/core/dc_link_dp.c 			initial_link_settings, cur, status));
cur               423 drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_crtc.c 	struct atmel_hlcdc_crtc_state *state, *cur;
cur               433 drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_crtc.c 	cur = drm_crtc_state_to_atmel_hlcdc_crtc_state(crtc->state);
cur               434 drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_crtc.c 	state->output_mode = cur->output_mode;
cur               228 drivers/gpu/drm/drm_syncobj.c 	struct syncobj_wait_entry *cur, *tmp;
cur               242 drivers/gpu/drm/drm_syncobj.c 	list_for_each_entry_safe(cur, tmp, &syncobj->cb_list, node)
cur               243 drivers/gpu/drm/drm_syncobj.c 		syncobj_wait_syncobj_func(syncobj, cur);
cur               263 drivers/gpu/drm/drm_syncobj.c 	struct syncobj_wait_entry *cur, *tmp;
cur               275 drivers/gpu/drm/drm_syncobj.c 		list_for_each_entry_safe(cur, tmp, &syncobj->cb_list, node)
cur               276 drivers/gpu/drm/drm_syncobj.c 			syncobj_wait_syncobj_func(syncobj, cur);
cur              2066 drivers/gpu/drm/gma500/cdv_intel_dp.c 		struct edp_power_seq cur;
cur              2085 drivers/gpu/drm/gma500/cdv_intel_dp.c                 cur.t1_t3 = (pp_on & PANEL_POWER_UP_DELAY_MASK) >>
cur              2088 drivers/gpu/drm/gma500/cdv_intel_dp.c                 cur.t8 = (pp_on & PANEL_LIGHT_ON_DELAY_MASK) >>
cur              2091 drivers/gpu/drm/gma500/cdv_intel_dp.c                 cur.t9 = (pp_off & PANEL_LIGHT_OFF_DELAY_MASK) >>
cur              2094 drivers/gpu/drm/gma500/cdv_intel_dp.c                 cur.t10 = (pp_off & PANEL_POWER_DOWN_DELAY_MASK) >>
cur              2097 drivers/gpu/drm/gma500/cdv_intel_dp.c                 cur.t11_t12 = ((pp_div & PANEL_POWER_CYCLE_DELAY_MASK) >>
cur              2101 drivers/gpu/drm/gma500/cdv_intel_dp.c                               cur.t1_t3, cur.t8, cur.t9, cur.t10, cur.t11_t12);
cur              2104 drivers/gpu/drm/gma500/cdv_intel_dp.c 		intel_dp->panel_power_up_delay = cur.t1_t3 / 10;
cur              2105 drivers/gpu/drm/gma500/cdv_intel_dp.c                 intel_dp->backlight_on_delay = cur.t8 / 10;
cur              2106 drivers/gpu/drm/gma500/cdv_intel_dp.c                 intel_dp->backlight_off_delay = cur.t9 / 10;
cur              2107 drivers/gpu/drm/gma500/cdv_intel_dp.c                 intel_dp->panel_power_down_delay = cur.t10 / 10;
cur              2108 drivers/gpu/drm/gma500/cdv_intel_dp.c                 intel_dp->panel_power_cycle_delay = (cur.t11_t12 - 1) * 100;
cur              11470 drivers/gpu/drm/i915/display/intel_display.c static bool intel_wm_need_update(const struct intel_plane_state *cur,
cur              11474 drivers/gpu/drm/i915/display/intel_display.c 	if (new->base.visible != cur->base.visible)
cur              11477 drivers/gpu/drm/i915/display/intel_display.c 	if (!cur->base.fb || !new->base.fb)
cur              11480 drivers/gpu/drm/i915/display/intel_display.c 	if (cur->base.fb->modifier != new->base.fb->modifier ||
cur              11481 drivers/gpu/drm/i915/display/intel_display.c 	    cur->base.rotation != new->base.rotation ||
cur              11482 drivers/gpu/drm/i915/display/intel_display.c 	    drm_rect_width(&new->base.src) != drm_rect_width(&cur->base.src) ||
cur              11483 drivers/gpu/drm/i915/display/intel_display.c 	    drm_rect_height(&new->base.src) != drm_rect_height(&cur->base.src) ||
cur              11484 drivers/gpu/drm/i915/display/intel_display.c 	    drm_rect_width(&new->base.dst) != drm_rect_width(&cur->base.dst) ||
cur              11485 drivers/gpu/drm/i915/display/intel_display.c 	    drm_rect_height(&new->base.dst) != drm_rect_height(&cur->base.dst))
cur              6440 drivers/gpu/drm/i915/display/intel_dp.c 	struct edp_power_seq cur, vbt, spec,
cur              6449 drivers/gpu/drm/i915/display/intel_dp.c 	intel_pps_readout_hw_state(intel_dp, &cur);
cur              6451 drivers/gpu/drm/i915/display/intel_dp.c 	intel_pps_dump_state("cur", &cur);
cur              6486 drivers/gpu/drm/i915/display/intel_dp.c #define assign_final(field)	final->field = (max(cur.field, vbt.field) == 0 ? \
cur              6488 drivers/gpu/drm/i915/display/intel_dp.c 				       max(cur.field, vbt.field))
cur               120 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c 	struct dma_fence_cb *cur, *tmp;
cur               125 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c 	list_for_each_entry_safe(cur, tmp, list, node) {
cur               126 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c 		INIT_LIST_HEAD(&cur->node);
cur               127 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c 		cur->func(fence, cur);
cur               632 drivers/gpu/drm/i915/gt/intel_lrc.c 	struct intel_engine_cs *cur, *old;
cur               638 drivers/gpu/drm/i915/gt/intel_lrc.c 		cur = ptr_unmask_bits(old, 2) ? ptr_dec(old) : NULL;
cur               639 drivers/gpu/drm/i915/gt/intel_lrc.c 	while (!try_cmpxchg(&ce->inflight, &old, cur));
cur               640 drivers/gpu/drm/i915/gt/intel_lrc.c 	if (!cur)
cur               951 drivers/gpu/drm/i915/gt/intel_workarounds.c wa_verify(const struct i915_wa *wa, u32 cur, const char *name, const char *from)
cur               953 drivers/gpu/drm/i915/gt/intel_workarounds.c 	if ((cur ^ wa->val) & wa->read) {
cur               956 drivers/gpu/drm/i915/gt/intel_workarounds.c 			  cur, cur & wa->read,
cur                71 drivers/gpu/drm/i915/i915_gpu_error.c 		__sg_set_buf(e->cur++, e->buf, e->bytes, e->iter);
cur                77 drivers/gpu/drm/i915/i915_gpu_error.c 	if (e->cur == e->end) {
cur                86 drivers/gpu/drm/i915/i915_gpu_error.c 		if (e->cur) {
cur                87 drivers/gpu/drm/i915/i915_gpu_error.c 			e->cur->offset = 0;
cur                88 drivers/gpu/drm/i915/i915_gpu_error.c 			e->cur->length = 0;
cur                89 drivers/gpu/drm/i915/i915_gpu_error.c 			e->cur->page_link =
cur                95 drivers/gpu/drm/i915/i915_gpu_error.c 		e->cur = sgl;
cur               807 drivers/gpu/drm/i915/i915_gpu_error.c 		__sg_set_buf(m.cur++, m.buf, m.bytes, m.iter);
cur               811 drivers/gpu/drm/i915/i915_gpu_error.c 	if (m.cur) {
cur               812 drivers/gpu/drm/i915/i915_gpu_error.c 		GEM_BUG_ON(m.end < m.cur);
cur               813 drivers/gpu/drm/i915/i915_gpu_error.c 		sg_mark_end(m.cur - 1);
cur               815 drivers/gpu/drm/i915/i915_gpu_error.c 	GEM_BUG_ON(m.sgl && !m.cur);
cur               186 drivers/gpu/drm/i915/i915_gpu_error.h 	struct scatterlist *sgl, *cur, *end;
cur               164 drivers/gpu/drm/i915/i915_pmu.c 	sample->cur += val;
cur               221 drivers/gpu/drm/i915/i915_pmu.c 	sample->cur += mul_u32_u32(val, mul);
cur               471 drivers/gpu/drm/i915/i915_pmu.c 		if (val >= pmu->sample[__I915_SAMPLE_RC6_ESTIMATED].cur) {
cur               472 drivers/gpu/drm/i915/i915_pmu.c 			pmu->sample[__I915_SAMPLE_RC6_ESTIMATED].cur = 0;
cur               473 drivers/gpu/drm/i915/i915_pmu.c 			pmu->sample[__I915_SAMPLE_RC6].cur = val;
cur               475 drivers/gpu/drm/i915/i915_pmu.c 			val = pmu->sample[__I915_SAMPLE_RC6_ESTIMATED].cur;
cur               505 drivers/gpu/drm/i915/i915_pmu.c 			if (!pmu->sample[__I915_SAMPLE_RC6_ESTIMATED].cur)
cur               509 drivers/gpu/drm/i915/i915_pmu.c 			val += pmu->sample[__I915_SAMPLE_RC6].cur;
cur               511 drivers/gpu/drm/i915/i915_pmu.c 			pmu->sample[__I915_SAMPLE_RC6_ESTIMATED].cur = val;
cur               512 drivers/gpu/drm/i915/i915_pmu.c 		} else if (pmu->sample[__I915_SAMPLE_RC6_ESTIMATED].cur) {
cur               513 drivers/gpu/drm/i915/i915_pmu.c 			val = pmu->sample[__I915_SAMPLE_RC6_ESTIMATED].cur;
cur               515 drivers/gpu/drm/i915/i915_pmu.c 			val = pmu->sample[__I915_SAMPLE_RC6].cur;
cur               548 drivers/gpu/drm/i915/i915_pmu.c 			val = engine->pmu.sample[sample].cur;
cur               554 drivers/gpu/drm/i915/i915_pmu.c 			   div_u64(pmu->sample[__I915_SAMPLE_FREQ_ACT].cur,
cur               559 drivers/gpu/drm/i915/i915_pmu.c 			   div_u64(pmu->sample[__I915_SAMPLE_FREQ_REQ].cur,
cur                37 drivers/gpu/drm/i915/i915_pmu.h 	u64 cur;
cur              6108 drivers/gpu/drm/i915/i915_reg.h #define HSW_WM_LP_VAL(lat, fbc, pri, cur) \
cur              6110 drivers/gpu/drm/i915/i915_reg.h 	 ((fbc) << WM1_LP_FBC_SHIFT) | ((pri) << WM1_LP_SR_SHIFT) | (cur))
cur              2497 drivers/gpu/drm/i915/intel_pm.c 	u16 cur;
cur              2706 drivers/gpu/drm/i915/intel_pm.c 	max->cur = ilk_cursor_wm_max(dev_priv, level, config);
cur              2716 drivers/gpu/drm/i915/intel_pm.c 	max->cur = ilk_cursor_wm_reg_max(dev_priv, level);
cur              2732 drivers/gpu/drm/i915/intel_pm.c 			 result->cur_val <= max->cur;
cur              2748 drivers/gpu/drm/i915/intel_pm.c 		if (result->cur_val > max->cur)
cur              2750 drivers/gpu/drm/i915/intel_pm.c 				      level, result->cur_val, max->cur);
cur              2754 drivers/gpu/drm/i915/intel_pm.c 		result->cur_val = min_t(u32, result->cur_val, max->cur);
cur               197 drivers/gpu/drm/mediatek/mtk_drm_ddp.c static unsigned int mtk_ddp_mout_en(enum mtk_ddp_comp_id cur,
cur               203 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	if (cur == DDP_COMPONENT_OVL0 && next == DDP_COMPONENT_COLOR0) {
cur               206 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_OVL0 && next == DDP_COMPONENT_RDMA0) {
cur               209 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_OD0 && next == DDP_COMPONENT_RDMA0) {
cur               212 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_UFOE && next == DDP_COMPONENT_DSI0) {
cur               215 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_OVL1 && next == DDP_COMPONENT_COLOR1) {
cur               218 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_GAMMA && next == DDP_COMPONENT_RDMA1) {
cur               221 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_OD1 && next == DDP_COMPONENT_RDMA1) {
cur               224 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA0 && next == DDP_COMPONENT_DPI0) {
cur               227 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA0 && next == DDP_COMPONENT_DPI1) {
cur               230 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA0 && next == DDP_COMPONENT_DSI1) {
cur               233 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA0 && next == DDP_COMPONENT_DSI2) {
cur               236 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA0 && next == DDP_COMPONENT_DSI3) {
cur               239 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI1) {
cur               242 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI2) {
cur               245 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI3) {
cur               248 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DPI0) {
cur               251 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DPI1) {
cur               254 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DPI0) {
cur               257 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DPI1) {
cur               260 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI1) {
cur               263 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI2) {
cur               266 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI3) {
cur               276 drivers/gpu/drm/mediatek/mtk_drm_ddp.c static unsigned int mtk_ddp_sel_in(enum mtk_ddp_comp_id cur,
cur               282 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	if (cur == DDP_COMPONENT_OVL0 && next == DDP_COMPONENT_COLOR0) {
cur               285 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DPI0) {
cur               288 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DPI1) {
cur               291 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI0) {
cur               294 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI1) {
cur               297 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI2) {
cur               300 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI3) {
cur               303 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DPI0) {
cur               306 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DPI1) {
cur               309 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI0) {
cur               312 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI1) {
cur               315 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI2) {
cur               318 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI3) {
cur               321 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_OVL1 && next == DDP_COMPONENT_COLOR1) {
cur               324 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_BLS && next == DDP_COMPONENT_DSI0) {
cur               335 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 			     enum mtk_ddp_comp_id cur,
cur               338 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	if (cur == DDP_COMPONENT_BLS && next == DDP_COMPONENT_DSI0) {
cur               341 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	} else if (cur == DDP_COMPONENT_BLS && next == DDP_COMPONENT_DPI0) {
cur               352 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 			      enum mtk_ddp_comp_id cur,
cur               357 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	value = mtk_ddp_mout_en(cur, next, &addr);
cur               363 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	mtk_ddp_sout_sel(config_regs, cur, next);
cur               365 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	value = mtk_ddp_sel_in(cur, next, &addr);
cur               373 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 				   enum mtk_ddp_comp_id cur,
cur               378 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	value = mtk_ddp_mout_en(cur, next, &addr);
cur               384 drivers/gpu/drm/mediatek/mtk_drm_ddp.c 	value = mtk_ddp_sel_in(cur, next, &addr);
cur                16 drivers/gpu/drm/mediatek/mtk_drm_ddp.h 			      enum mtk_ddp_comp_id cur,
cur                19 drivers/gpu/drm/mediatek/mtk_drm_ddp.h 				   enum mtk_ddp_comp_id cur,
cur                31 drivers/gpu/drm/msm/adreno/a5xx_gpu.c 	ring->cur = ring->next;
cur                15 drivers/gpu/drm/msm/adreno/a5xx_preempt.c 	enum preempt_state cur = atomic_cmpxchg(&a5xx_gpu->preempt_state,
cur                18 drivers/gpu/drm/msm/adreno/a5xx_preempt.c 	return (cur == old);
cur                60 drivers/gpu/drm/msm/adreno/a6xx_gpu.c 	ring->cur = ring->next;
cur               349 drivers/gpu/drm/msm/adreno/adreno_gpu.c 		ring->cur = ring->start;
cur               492 drivers/gpu/drm/msm/adreno/adreno_gpu.c 	ring->cur = ring->next;
cur               368 drivers/gpu/drm/msm/adreno/adreno_gpu.h 	return (ring->cur - ring->start) % (MSM_GPU_RINGBUFFER_SZ >> 2);
cur                54 drivers/gpu/drm/msm/disp/mdp5/mdp5_mixer.c 		struct mdp5_hw_mixer *cur = mdp5_kms->hwmixers[i];
cur                63 drivers/gpu/drm/msm/disp/mdp5/mdp5_mixer.c 		if (new_state->hwmixer_to_crtc[cur->idx] &&
cur                64 drivers/gpu/drm/msm/disp/mdp5/mdp5_mixer.c 		    new_state->hwmixer_to_crtc[cur->idx] != crtc)
cur                68 drivers/gpu/drm/msm/disp/mdp5/mdp5_mixer.c 		if (caps & ~cur->caps)
cur                74 drivers/gpu/drm/msm/disp/mdp5/mdp5_mixer.c 			pair_idx = get_right_pair_idx(mdp5_kms, cur->lm);
cur                97 drivers/gpu/drm/msm/disp/mdp5/mdp5_mixer.c 		if (!(*mixer) || cur->caps & MDP_LM_CAP_PAIR)
cur                98 drivers/gpu/drm/msm/disp/mdp5/mdp5_mixer.c 			*mixer = cur;
cur                31 drivers/gpu/drm/msm/disp/mdp5/mdp5_pipe.c 		struct mdp5_hw_pipe *cur = mdp5_kms->hwpipes[i];
cur                40 drivers/gpu/drm/msm/disp/mdp5/mdp5_pipe.c 		if (new_state->hwpipe_to_plane[cur->idx] ||
cur                41 drivers/gpu/drm/msm/disp/mdp5/mdp5_pipe.c 				old_state->hwpipe_to_plane[cur->idx])
cur                45 drivers/gpu/drm/msm/disp/mdp5/mdp5_pipe.c 		if (caps & ~cur->caps)
cur                52 drivers/gpu/drm/msm/disp/mdp5/mdp5_pipe.c 		if (cur->caps & MDP_PIPE_CAP_CURSOR &&
cur                59 drivers/gpu/drm/msm/disp/mdp5/mdp5_pipe.c 		if (!(*hwpipe) || (hweight_long(cur->caps & ~caps) <
cur                70 drivers/gpu/drm/msm/disp/mdp5/mdp5_pipe.c 					if (r_cur->caps != cur->caps)
cur                74 drivers/gpu/drm/msm/disp/mdp5/mdp5_pipe.c 					if (cur->pipe > r_cur->pipe)
cur                84 drivers/gpu/drm/msm/disp/mdp5/mdp5_pipe.c 				*hwpipe = cur;
cur                42 drivers/gpu/drm/msm/msm_ringbuffer.c 	ring->cur   = ring->start;
cur                40 drivers/gpu/drm/msm/msm_ringbuffer.h 	uint32_t *start, *end, *cur, *next;
cur                40 drivers/gpu/drm/nouveau/dispnv04/overlay.c 	struct nouveau_bo *cur;
cur               125 drivers/gpu/drm/nouveau/dispnv04/overlay.c 	struct nouveau_bo *cur = nv_plane->cur;
cur               147 drivers/gpu/drm/nouveau/dispnv04/overlay.c 	nv_plane->cur = nv_fb->nvbo;
cur               183 drivers/gpu/drm/nouveau/dispnv04/overlay.c 	if (cur)
cur               184 drivers/gpu/drm/nouveau/dispnv04/overlay.c 		nouveau_bo_unpin(cur);
cur               198 drivers/gpu/drm/nouveau/dispnv04/overlay.c 	if (nv_plane->cur) {
cur               199 drivers/gpu/drm/nouveau/dispnv04/overlay.c 		nouveau_bo_unpin(nv_plane->cur);
cur               200 drivers/gpu/drm/nouveau/dispnv04/overlay.c 		nv_plane->cur = NULL;
cur               229 drivers/gpu/drm/nouveau/dispnv04/overlay.c 	if (plane->cur) {
cur               372 drivers/gpu/drm/nouveau/dispnv04/overlay.c 	struct nouveau_bo *cur = nv_plane->cur;
cur               391 drivers/gpu/drm/nouveau/dispnv04/overlay.c 	nv_plane->cur = nv_fb->nvbo;
cur               432 drivers/gpu/drm/nouveau/dispnv04/overlay.c 	if (cur)
cur               433 drivers/gpu/drm/nouveau/dispnv04/overlay.c 		nouveau_bo_unpin(cur);
cur               450 drivers/gpu/drm/nouveau/dispnv04/overlay.c 	if (nv_plane->cur) {
cur               451 drivers/gpu/drm/nouveau/dispnv04/overlay.c 		nouveau_bo_unpin(nv_plane->cur);
cur               452 drivers/gpu/drm/nouveau/dispnv04/overlay.c 		nv_plane->cur = NULL;
cur               442 drivers/gpu/drm/nouveau/nouveau_chan.c 	chan->dma.cur = chan->dma.put;
cur               443 drivers/gpu/drm/nouveau/nouveau_chan.c 	chan->dma.free = chan->dma.max - chan->dma.cur;
cur                34 drivers/gpu/drm/nouveau/nouveau_chan.h 		int cur;
cur               147 drivers/gpu/drm/nouveau/nouveau_debugfs.c 	char buf[32] = {}, *tmp, *cur = buf;
cur               162 drivers/gpu/drm/nouveau/nouveau_debugfs.c 	if (!strncasecmp(cur, "dc:", 3)) {
cur               164 drivers/gpu/drm/nouveau/nouveau_debugfs.c 		cur += 3;
cur               166 drivers/gpu/drm/nouveau/nouveau_debugfs.c 	if (!strncasecmp(cur, "ac:", 3)) {
cur               168 drivers/gpu/drm/nouveau/nouveau_debugfs.c 		cur += 3;
cur               171 drivers/gpu/drm/nouveau/nouveau_debugfs.c 	if (!strcasecmp(cur, "none"))
cur               174 drivers/gpu/drm/nouveau/nouveau_debugfs.c 	if (!strcasecmp(cur, "auto"))
cur               177 drivers/gpu/drm/nouveau/nouveau_debugfs.c 		ret = kstrtol(cur, 16, &value);
cur                38 drivers/gpu/drm/nouveau/nouveau_dma.c 	mem = &mem[chan->dma.cur];
cur                43 drivers/gpu/drm/nouveau/nouveau_dma.c 	chan->dma.cur += nr_dwords;
cur               153 drivers/gpu/drm/nouveau/nouveau_dma.c 		if (get <= chan->dma.cur) {
cur               154 drivers/gpu/drm/nouveau/nouveau_dma.c 			chan->dma.free = chan->dma.max - chan->dma.cur;
cur               167 drivers/gpu/drm/nouveau/nouveau_dma.c 			chan->dma.cur = 0;
cur               171 drivers/gpu/drm/nouveau/nouveau_dma.c 		chan->dma.free = get - chan->dma.cur - 1;
cur               203 drivers/gpu/drm/nouveau/nouveau_dma.c 		if (get <= chan->dma.cur) {
cur               217 drivers/gpu/drm/nouveau/nouveau_dma.c 			chan->dma.free = chan->dma.max - chan->dma.cur;
cur               244 drivers/gpu/drm/nouveau/nouveau_dma.c 			chan->dma.cur  =
cur               254 drivers/gpu/drm/nouveau/nouveau_dma.c 		chan->dma.free = get - chan->dma.cur - 1;
cur               102 drivers/gpu/drm/nouveau/nouveau_dma.h 	nouveau_bo_wr32(chan->push.buffer, chan->dma.cur++, data);
cur               147 drivers/gpu/drm/nouveau/nouveau_dma.h 	if (chan->dma.cur == chan->dma.put)
cur               153 drivers/gpu/drm/nouveau/nouveau_dma.h 			      (chan->dma.cur - chan->dma.put) << 2);
cur               155 drivers/gpu/drm/nouveau/nouveau_dma.h 		WRITE_PUT(chan->dma.cur);
cur               158 drivers/gpu/drm/nouveau/nouveau_dma.h 	chan->dma.put = chan->dma.cur;
cur               164 drivers/gpu/drm/nouveau/nouveau_dma.h 	chan->dma.cur = chan->dma.put;
cur               816 drivers/gpu/drm/nouveau/nouveau_gem.c 			cmd = chan->push.addr + ((chan->dma.cur + 2) << 2);
cur               868 drivers/gpu/drm/nouveau/nouveau_gem.c 			      (chan->push.addr + ((chan->dma.cur + 2) << 2));
cur                56 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c 	struct nvkm_memory *cur;
cur                61 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c 	cur = fifo->runlist.mem[fifo->runlist.active];
cur                64 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c 	nvkm_kmap(cur);
cur                66 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c 		nvkm_wo32(cur, (nr * 8) + 0, chan->base.chid);
cur                67 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c 		nvkm_wo32(cur, (nr * 8) + 4, 0x00000004);
cur                70 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c 	nvkm_done(cur);
cur                72 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c 	switch (nvkm_memory_target(cur)) {
cur                81 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c 	nvkm_wr32(device, 0x002270, (nvkm_memory_addr(cur) >> 12) |
cur                33 drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c 	struct nvkm_memory *cur;
cur                36 drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c 	cur = fifo->runlist[fifo->cur_runlist];
cur                39 drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c 	nvkm_kmap(cur);
cur                42 drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c 			nvkm_wo32(cur, p++ * 4, i);
cur                44 drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c 	nvkm_done(cur);
cur                46 drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c 	nvkm_wr32(device, 0x0032f4, nvkm_memory_addr(cur) >> 12);
cur               129 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c 	u32 cycles, cur, prev;
cur               149 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c 		cur = nvkm_gpio_get(gpio, 0, therm->fan->tach.func,
cur               151 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c 		if (prev != cur) {
cur               155 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c 			prev = cur;
cur               103 drivers/gpu/drm/nouveau/nvkm/subdev/therm/g84.c 	int temp, cur;
cur               118 drivers/gpu/drm/nouveau/nvkm/subdev/therm/g84.c 	cur = therm->func->temp_get(therm);
cur               119 drivers/gpu/drm/nouveau/nvkm/subdev/therm/g84.c 	if (new_state == NVKM_THERM_THRS_LOWER && cur > thrs->temp)
cur               122 drivers/gpu/drm/nouveau/nvkm/subdev/therm/g84.c 		cur < thrs->temp - thrs->hysteresis)
cur                64 drivers/gpu/drm/qxl/qxl_release.c 	unsigned long cur, end = jiffies + timeout;
cur               111 drivers/gpu/drm/qxl/qxl_release.c 	cur = jiffies;
cur               112 drivers/gpu/drm/qxl/qxl_release.c 	if (time_after(cur, end))
cur               114 drivers/gpu/drm/qxl/qxl_release.c 	return end - cur;
cur               194 drivers/gpu/drm/radeon/rv6xx_dpm.c 						      struct rv6xx_sclk_stepping *cur,
cur               199 drivers/gpu/drm/radeon/rv6xx_dpm.c 	next.post_divider = cur->post_divider;
cur               202 drivers/gpu/drm/radeon/rv6xx_dpm.c 		next.vco_frequency = (cur->vco_frequency * (100 + step_size)) / 100;
cur               204 drivers/gpu/drm/radeon/rv6xx_dpm.c 		next.vco_frequency = (cur->vco_frequency * 100 + 99 + step_size) / (100 + step_size);
cur               210 drivers/gpu/drm/radeon/rv6xx_dpm.c 				    struct rv6xx_sclk_stepping *cur,
cur               213 drivers/gpu/drm/radeon/rv6xx_dpm.c 	return (cur->post_divider > target->post_divider) &&
cur               214 drivers/gpu/drm/radeon/rv6xx_dpm.c 		((cur->vco_frequency * target->post_divider) <=
cur               215 drivers/gpu/drm/radeon/rv6xx_dpm.c 		 (target->vco_frequency * (cur->post_divider - 1)));
cur               219 drivers/gpu/drm/radeon/rv6xx_dpm.c 							   struct rv6xx_sclk_stepping *cur,
cur               222 drivers/gpu/drm/radeon/rv6xx_dpm.c 	struct rv6xx_sclk_stepping next = *cur;
cur               231 drivers/gpu/drm/radeon/rv6xx_dpm.c 					  struct rv6xx_sclk_stepping *cur,
cur               235 drivers/gpu/drm/radeon/rv6xx_dpm.c 	return (increasing_vco && (cur->vco_frequency >= target->vco_frequency)) ||
cur               236 drivers/gpu/drm/radeon/rv6xx_dpm.c 		(!increasing_vco && (cur->vco_frequency <= target->vco_frequency));
cur               243 drivers/gpu/drm/radeon/rv6xx_dpm.c 	struct rv6xx_sclk_stepping cur;
cur               248 drivers/gpu/drm/radeon/rv6xx_dpm.c 	rv6xx_convert_clock_to_stepping(rdev, low, &cur);
cur               251 drivers/gpu/drm/radeon/rv6xx_dpm.c 	rv6xx_output_stepping(rdev, step_index++, &cur);
cur               253 drivers/gpu/drm/radeon/rv6xx_dpm.c 	increasing_vco = (target.vco_frequency >= cur.vco_frequency);
cur               255 drivers/gpu/drm/radeon/rv6xx_dpm.c 	if (target.post_divider > cur.post_divider)
cur               256 drivers/gpu/drm/radeon/rv6xx_dpm.c 		cur.post_divider = target.post_divider;
cur               261 drivers/gpu/drm/radeon/rv6xx_dpm.c 		if (rv6xx_can_step_post_div(rdev, &cur, &target))
cur               262 drivers/gpu/drm/radeon/rv6xx_dpm.c 			next = rv6xx_next_post_div_step(rdev, &cur, &target);
cur               264 drivers/gpu/drm/radeon/rv6xx_dpm.c 			next = rv6xx_next_vco_step(rdev, &cur, increasing_vco, R600_VCOSTEPPCT_DFLT);
cur               271 drivers/gpu/drm/radeon/rv6xx_dpm.c 			if (!rv6xx_reached_stepping_target(rdev, &tiny, &cur, !increasing_vco))
cur               289 drivers/gpu/drm/radeon/rv6xx_dpm.c 		cur = next;
cur               374 drivers/gpu/drm/savage/savage_bci.c 	unsigned int cur = dev_priv->current_dma_page;
cur               376 drivers/gpu/drm/savage/savage_bci.c 	    dev_priv->dma_pages[cur].used;
cur               383 drivers/gpu/drm/savage/savage_bci.c 		  cur, dev_priv->dma_pages[cur].used, n, rest, nr_pages);
cur               385 drivers/gpu/drm/savage/savage_bci.c 	if (cur + nr_pages < dev_priv->nr_dma_pages) {
cur               387 drivers/gpu/drm/savage/savage_bci.c 		    cur * SAVAGE_DMA_PAGE_SIZE + dev_priv->dma_pages[cur].used;
cur               390 drivers/gpu/drm/savage/savage_bci.c 		dev_priv->dma_pages[cur].used += rest;
cur               392 drivers/gpu/drm/savage/savage_bci.c 		cur++;
cur               397 drivers/gpu/drm/savage/savage_bci.c 		for (i = cur; i < dev_priv->nr_dma_pages; ++i) {
cur               403 drivers/gpu/drm/savage/savage_bci.c 		dev_priv->first_dma_page = cur = 0;
cur               405 drivers/gpu/drm/savage/savage_bci.c 	for (i = cur; nr_pages > 0; ++i, --nr_pages) {
cur               431 drivers/gpu/drm/savage/savage_bci.c 	unsigned int cur = dev_priv->current_dma_page;
cur               437 drivers/gpu/drm/savage/savage_bci.c 	if (first == cur &&
cur               438 drivers/gpu/drm/savage/savage_bci.c 	    dev_priv->dma_pages[cur].used == dev_priv->dma_pages[cur].flushed)
cur               443 drivers/gpu/drm/savage/savage_bci.c 	pad = -dev_priv->dma_pages[cur].used & 1;
cur               444 drivers/gpu/drm/savage/savage_bci.c 	align = -(dev_priv->dma_pages[cur].used + pad) & 7;
cur               448 drivers/gpu/drm/savage/savage_bci.c 		  first, cur, dev_priv->dma_pages[first].flushed,
cur               449 drivers/gpu/drm/savage/savage_bci.c 		  dev_priv->dma_pages[cur].used, pad, align);
cur               454 drivers/gpu/drm/savage/savage_bci.c 		    cur * SAVAGE_DMA_PAGE_SIZE + dev_priv->dma_pages[cur].used;
cur               455 drivers/gpu/drm/savage/savage_bci.c 		dev_priv->dma_pages[cur].used += pad;
cur               468 drivers/gpu/drm/savage/savage_bci.c 	len = (cur - first) * SAVAGE_DMA_PAGE_SIZE +
cur               469 drivers/gpu/drm/savage/savage_bci.c 	    dev_priv->dma_pages[cur].used - dev_priv->dma_pages[first].flushed;
cur               480 drivers/gpu/drm/savage/savage_bci.c 	dev_priv->dma_pages[cur].used += align;
cur               485 drivers/gpu/drm/savage/savage_bci.c 	for (i = first; i < cur; ++i) {
cur               491 drivers/gpu/drm/savage/savage_bci.c 	if (dev_priv->dma_pages[cur].used == SAVAGE_DMA_PAGE_SIZE) {
cur               492 drivers/gpu/drm/savage/savage_bci.c 		SET_AGE(&dev_priv->dma_pages[cur].age, event, wrap);
cur               493 drivers/gpu/drm/savage/savage_bci.c 		dev_priv->dma_pages[cur].used = 0;
cur               494 drivers/gpu/drm/savage/savage_bci.c 		dev_priv->dma_pages[cur].flushed = 0;
cur               496 drivers/gpu/drm/savage/savage_bci.c 		cur++;
cur               497 drivers/gpu/drm/savage/savage_bci.c 		if (cur == dev_priv->nr_dma_pages)
cur               498 drivers/gpu/drm/savage/savage_bci.c 			cur = 0;
cur               499 drivers/gpu/drm/savage/savage_bci.c 		dev_priv->first_dma_page = dev_priv->current_dma_page = cur;
cur               501 drivers/gpu/drm/savage/savage_bci.c 		dev_priv->first_dma_page = cur;
cur               502 drivers/gpu/drm/savage/savage_bci.c 		dev_priv->dma_pages[cur].flushed = dev_priv->dma_pages[i].used;
cur               506 drivers/gpu/drm/savage/savage_bci.c 	DRM_DEBUG("first=cur=%u, cur->used=%u, cur->flushed=%u\n", cur,
cur               507 drivers/gpu/drm/savage/savage_bci.c 		  dev_priv->dma_pages[cur].used,
cur               508 drivers/gpu/drm/savage/savage_bci.c 		  dev_priv->dma_pages[cur].flushed);
cur               518 drivers/gpu/drm/savage/savage_drv.h 	unsigned int cur = dev_priv->current_dma_page;			\
cur               520 drivers/gpu/drm/savage/savage_drv.h 		dev_priv->dma_pages[cur].used;				\
cur               525 drivers/gpu/drm/savage/savage_drv.h 			cur * SAVAGE_DMA_PAGE_SIZE +			\
cur               526 drivers/gpu/drm/savage/savage_drv.h 			dev_priv->dma_pages[cur].used;			\
cur               527 drivers/gpu/drm/savage/savage_drv.h 		if (dev_priv->dma_pages[cur].used == 0)			\
cur               528 drivers/gpu/drm/savage/savage_drv.h 			savage_dma_wait(dev_priv, cur);			\
cur               529 drivers/gpu/drm/savage/savage_drv.h 		dev_priv->dma_pages[cur].used += (n);			\
cur               542 drivers/gpu/drm/savage/savage_drv.h 	unsigned int cur = dev_priv->current_dma_page;			\
cur               544 drivers/gpu/drm/savage/savage_drv.h 			cur * SAVAGE_DMA_PAGE_SIZE +			\
cur               545 drivers/gpu/drm/savage/savage_drv.h 			dev_priv->dma_pages[cur].used;			\
cur               316 drivers/gpu/drm/via/via_dmablit.c 	int cur;
cur               333 drivers/gpu/drm/via/via_dmablit.c 	cur = blitq->cur;
cur               336 drivers/gpu/drm/via/via_dmablit.c 		blitq->blits[cur]->aborted = blitq->aborting;
cur               338 drivers/gpu/drm/via/via_dmablit.c 		wake_up(blitq->blit_queue + cur);
cur               340 drivers/gpu/drm/via/via_dmablit.c 		cur++;
cur               341 drivers/gpu/drm/via/via_dmablit.c 		if (cur >= VIA_NUM_BLIT_SLOTS)
cur               342 drivers/gpu/drm/via/via_dmablit.c 			cur = 0;
cur               343 drivers/gpu/drm/via/via_dmablit.c 		blitq->cur = cur;
cur               368 drivers/gpu/drm/via/via_dmablit.c 			via_fire_dmablit(dev, blitq->blits[cur], engine);
cur               370 drivers/gpu/drm/via/via_dmablit.c 			blitq->cur = cur;
cur               411 drivers/gpu/drm/via/via_dmablit.c 		slot = handle - blitq->done_blit_handle + blitq->cur - 1;
cur               507 drivers/gpu/drm/via/via_dmablit.c 	while (blitq->serviced != blitq->cur) {
cur               553 drivers/gpu/drm/via/via_dmablit.c 		blitq->cur = 0;
cur                67 drivers/gpu/drm/via/via_dmablit.h 	unsigned cur;
cur               121 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	struct vmw_cmdbuf_header *cur;
cur               662 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	struct vmw_cmdbuf_header *cur = man->cur;
cur               666 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	if (!cur)
cur               671 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 		__vmw_cmdbuf_header_free(cur);
cur               675 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	man->cur->cb_header->length = man->cur_pos;
cur               676 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	vmw_cmdbuf_ctx_add(man, man->cur, SVGA_CB_CONTEXT_0);
cur               679 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	man->cur = NULL;
cur              1003 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	struct vmw_cmdbuf_header *cur;
cur              1009 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	cur = man->cur;
cur              1010 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	if (cur && (size + man->cur_pos > cur->size ||
cur              1011 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 		    ((cur->cb_header->flags & SVGA_CB_FLAG_DX_CONTEXT) &&
cur              1012 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 		     ctx_id != cur->cb_header->dxContext)))
cur              1015 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	if (!man->cur) {
cur              1018 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 				       interruptible, &man->cur);
cur              1024 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 		cur = man->cur;
cur              1028 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 		cur->cb_header->flags |= SVGA_CB_FLAG_DX_CONTEXT;
cur              1029 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 		cur->cb_header->dxContext = ctx_id;
cur              1032 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	cur->reserved = size;
cur              1034 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	return (void *) (man->cur->cmd + man->cur_pos);
cur              1047 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	struct vmw_cmdbuf_header *cur = man->cur;
cur              1051 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	WARN_ON(size > cur->reserved);
cur              1054 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 		cur->cb_header->flags &= ~SVGA_CB_FLAG_DX_CONTEXT;
cur              1112 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c 	man->cur = header;
cur               118 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 	struct vmw_ctx_binding_state *cur;
cur               176 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 			vmw_binding_state_commit(entry->cur, entry->staged);
cur               241 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 	node->cur = vmw_context_binding_state(res);
cur               752 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c 		ret = vmw_binding_rebind_all(val->cur);
cur               473 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 		dma_addr_t cur = vmw_piter_dma_addr(&iter);
cur               475 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 		if (cur != old + PAGE_SIZE)
cur               477 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c 		old = cur;
cur               239 drivers/hid/wacom.h struct wacom_led *wacom_led_next(struct wacom *wacom, struct wacom_led *cur);
cur              1472 drivers/hid/wacom_sys.c struct wacom_led *wacom_led_next(struct wacom *wacom, struct wacom_led *cur)
cur              1477 drivers/hid/wacom_sys.c 	if (!wacom || !cur)
cur              1480 drivers/hid/wacom_sys.c 	group = cur->group;
cur              1481 drivers/hid/wacom_sys.c 	next = cur->id;
cur              1485 drivers/hid/wacom_sys.c 		if (!next_led || next_led == cur)
cur              3951 drivers/hid/wacom_wac.c 	int cur;
cur              3958 drivers/hid/wacom_wac.c 	cur = wacom->led.groups[group].select;
cur              3960 drivers/hid/wacom_wac.c 	led = wacom_led_find(wacom, group, cur);
cur              3963 drivers/hid/wacom_wac.c 			cur, group);
cur              1427 drivers/hv/channel_mgmt.c 	struct list_head *cur, *tmp;
cur              1433 drivers/hv/channel_mgmt.c 	list_for_each_safe(cur, tmp, &primary_channel->sc_list) {
cur              1434 drivers/hv/channel_mgmt.c 		cur_channel = list_entry(cur, struct vmbus_channel, sc_list);
cur               316 drivers/hv/connection.c 	struct list_head *cur, *tmp;
cur               329 drivers/hv/connection.c 			list_for_each_safe(cur, tmp, &channel->sc_list) {
cur               330 drivers/hv/connection.c 				cur_sc = list_entry(cur, struct vmbus_channel,
cur               523 drivers/hv/vmbus_drv.c 	struct list_head *cur;
cur               533 drivers/hv/vmbus_drv.c 	list_for_each(cur, &channel->sc_list) {
cur               537 drivers/hv/vmbus_drv.c 		cur_sc = list_entry(cur, struct vmbus_channel, sc_list);
cur               243 drivers/hwmon/lm77.c 	int i, cur, conf, hyst, crit, min, max;
cur               264 drivers/hwmon/lm77.c 	cur = i2c_smbus_read_word_data(client, 0);
cur               280 drivers/hwmon/lm77.c 	if (((cur & 0x00f0) != 0xf0 && (cur & 0x00f0) != 0x0)
cur               292 drivers/hwmon/lm77.c 	cur = i2c_smbus_read_word_data(client, 0);
cur               293 drivers/hwmon/lm77.c 	if (i2c_smbus_read_word_data(client, 6) != cur
cur               294 drivers/hwmon/lm77.c 	 || i2c_smbus_read_word_data(client, 7) != cur)
cur               552 drivers/hwmon/lm80.c 	int i, cur, man_id, dev_id;
cur               579 drivers/hwmon/lm80.c 			cur = i2c_smbus_read_byte_data(client, i);
cur               580 drivers/hwmon/lm80.c 			if ((i2c_smbus_read_byte_data(client, i + 0x40) != cur)
cur               581 drivers/hwmon/lm80.c 			 || (i2c_smbus_read_byte_data(client, i + 0x80) != cur)
cur               582 drivers/hwmon/lm80.c 			 || (i2c_smbus_read_byte_data(client, i + 0xc0) != cur))
cur               221 drivers/hwmon/lm93.c 		u8 cur;
cur              1057 drivers/hwmon/lm93.c 		data->block4[i].cur =
cur              2114 drivers/hwmon/lm93.c 	return sprintf(buf, "%d\n", data->block4[nr].cur);
cur               414 drivers/hwtracing/coresight/coresight-etb10.c 	buf->cur = head / PAGE_SIZE;
cur               429 drivers/hwtracing/coresight/coresight-etb10.c 	int i, cur;
cur               526 drivers/hwtracing/coresight/coresight-etb10.c 	cur = buf->cur;
cur               531 drivers/hwtracing/coresight/coresight-etb10.c 		buf_ptr = buf->data_pages[cur] + offset;
cur               545 drivers/hwtracing/coresight/coresight-etb10.c 			cur++;
cur               547 drivers/hwtracing/coresight/coresight-etb10.c 			cur &= buf->nr_pages - 1;
cur                95 drivers/hwtracing/coresight/coresight-priv.h 	unsigned int		cur;
cur               430 drivers/hwtracing/coresight/coresight-tmc-etf.c 	buf->cur = head / PAGE_SIZE;
cur               445 drivers/hwtracing/coresight/coresight-tmc-etf.c 	int i, cur;
cur               520 drivers/hwtracing/coresight/coresight-tmc-etf.c 	cur = buf->cur;
cur               526 drivers/hwtracing/coresight/coresight-tmc-etf.c 		buf_ptr = buf->data_pages[cur] + offset;
cur               537 drivers/hwtracing/coresight/coresight-tmc-etf.c 			cur++;
cur               539 drivers/hwtracing/coresight/coresight-tmc-etf.c 			cur &= buf->nr_pages - 1;
cur                83 drivers/i2c/busses/i2c-qcom-geni.c 	struct i2c_msg *cur;
cur               192 drivers/i2c/busses/i2c-qcom-geni.c 	if (gi2c->cur)
cur               194 drivers/i2c/busses/i2c-qcom-geni.c 			gi2c->cur->len, gi2c->cur->addr, gi2c->cur->flags);
cur               213 drivers/i2c/busses/i2c-qcom-geni.c 	struct i2c_msg *cur;
cur               222 drivers/i2c/busses/i2c-qcom-geni.c 	cur = gi2c->cur;
cur               224 drivers/i2c/busses/i2c-qcom-geni.c 	if (!cur ||
cur               248 drivers/i2c/busses/i2c-qcom-geni.c 	} else if (cur->flags & I2C_M_RD &&
cur               255 drivers/i2c/busses/i2c-qcom-geni.c 			while (gi2c->cur_rd < cur->len && p < sizeof(val)) {
cur               256 drivers/i2c/busses/i2c-qcom-geni.c 				cur->buf[gi2c->cur_rd++] = val & 0xff;
cur               260 drivers/i2c/busses/i2c-qcom-geni.c 			if (gi2c->cur_rd == cur->len)
cur               263 drivers/i2c/busses/i2c-qcom-geni.c 	} else if (!(cur->flags & I2C_M_RD) &&
cur               270 drivers/i2c/busses/i2c-qcom-geni.c 			while (gi2c->cur_wr < cur->len && p < sizeof(val)) {
cur               271 drivers/i2c/busses/i2c-qcom-geni.c 				temp = cur->buf[gi2c->cur_wr++];
cur               277 drivers/i2c/busses/i2c-qcom-geni.c 			if (gi2c->cur_wr == cur->len) {
cur               311 drivers/i2c/busses/i2c-qcom-geni.c 	gi2c->cur = NULL;
cur               462 drivers/i2c/busses/i2c-qcom-geni.c 		gi2c->cur = &msgs[i];
cur               476 drivers/i2c/busses/i2c-qcom-geni.c 	gi2c->cur = NULL;
cur               132 drivers/i2c/muxes/i2c-demux-pinctrl.c 	int ret, cur = priv->cur_chan;
cur               134 drivers/i2c/muxes/i2c-demux-pinctrl.c 	if (cur < 0)
cur               138 drivers/i2c/muxes/i2c-demux-pinctrl.c 	i2c_put_adapter(priv->chan[cur].parent_adap);
cur               140 drivers/i2c/muxes/i2c-demux-pinctrl.c 	ret = of_changeset_revert(&priv->chan[cur].chgset);
cur               142 drivers/i2c/muxes/i2c-demux-pinctrl.c 	priv->chan[cur].parent_adap = NULL;
cur               234 drivers/i3c/master/dw-i3c-master.c 		struct dw_i3c_xfer *cur;
cur               380 drivers/i3c/master/dw-i3c-master.c 	struct dw_i3c_xfer *xfer = master->xferqueue.cur;
cur               413 drivers/i3c/master/dw-i3c-master.c 	if (master->xferqueue.cur) {
cur               416 drivers/i3c/master/dw-i3c-master.c 		master->xferqueue.cur = xfer;
cur               425 drivers/i3c/master/dw-i3c-master.c 	if (master->xferqueue.cur == xfer) {
cur               428 drivers/i3c/master/dw-i3c-master.c 		master->xferqueue.cur = NULL;
cur               453 drivers/i3c/master/dw-i3c-master.c 	struct dw_i3c_xfer *xfer = master->xferqueue.cur;
cur               514 drivers/i3c/master/dw-i3c-master.c 	master->xferqueue.cur = xfer;
cur               403 drivers/i3c/master/i3c-master-cdns.c 		struct cdns_i3c_xfer *cur;
cur               522 drivers/i3c/master/i3c-master-cdns.c 	struct cdns_i3c_xfer *xfer = master->xferqueue.cur;
cur               552 drivers/i3c/master/i3c-master-cdns.c 	struct cdns_i3c_xfer *xfer = master->xferqueue.cur;
cur               619 drivers/i3c/master/i3c-master-cdns.c 	master->xferqueue.cur = xfer;
cur               630 drivers/i3c/master/i3c-master-cdns.c 	if (master->xferqueue.cur) {
cur               633 drivers/i3c/master/i3c-master-cdns.c 		master->xferqueue.cur = xfer;
cur               645 drivers/i3c/master/i3c-master-cdns.c 	if (master->xferqueue.cur == xfer) {
cur               653 drivers/i3c/master/i3c-master-cdns.c 		master->xferqueue.cur = NULL;
cur               203 drivers/ide/ide-tape.c 	void *cur;
cur               874 drivers/ide/ide-tape.c 	tape->cur = tape->buf;
cur               933 drivers/ide/ide-tape.c 		memset(tape->cur, 0, aligned - tape->valid);
cur               967 drivers/ide/ide-tape.c 	tape->cur = tape->buf;
cur              1168 drivers/ide/ide-tape.c 		if (copy_to_user(buf + done, tape->cur, todo))
cur              1171 drivers/ide/ide-tape.c 		tape->cur += todo;
cur              1216 drivers/ide/ide-tape.c 		if (copy_from_user(tape->cur, buf + done, todo))
cur              1219 drivers/ide/ide-tape.c 		tape->cur += todo;
cur              1663 drivers/iio/adc/stm32-adc.c 	const __be32 *cur;
cur              1709 drivers/iio/adc/stm32-adc.c 	of_property_for_each_u32(node, "st,adc-channels", prop, cur, val) {
cur               581 drivers/iio/adc/ti_am335x_adc.c 	const __be32 *cur;
cur               585 drivers/iio/adc/ti_am335x_adc.c 	of_property_for_each_u32(node, "ti,adc-channels", prop, cur, val) {
cur               370 drivers/iio/trigger/stm32-timer-trigger.c 	const char * const *cur = priv->triggers;
cur               372 drivers/iio/trigger/stm32-timer-trigger.c 	while (cur && *cur) {
cur               374 drivers/iio/trigger/stm32-timer-trigger.c 		bool cur_is_trgo = stm32_timer_is_trgo_name(*cur);
cur               375 drivers/iio/trigger/stm32-timer-trigger.c 		bool cur_is_trgo2 = stm32_timer_is_trgo2_name(*cur);
cur               378 drivers/iio/trigger/stm32-timer-trigger.c 			cur++;
cur               382 drivers/iio/trigger/stm32-timer-trigger.c 		trig = devm_iio_trigger_alloc(priv->dev, "%s", *cur);
cur               401 drivers/iio/trigger/stm32-timer-trigger.c 		cur++;
cur               481 drivers/iio/trigger/stm32-timer-trigger.c 	const char * const *cur = priv->valids;
cur               487 drivers/iio/trigger/stm32-timer-trigger.c 	while (cur && *cur) {
cur               488 drivers/iio/trigger/stm32-timer-trigger.c 		if (!strncmp(trig->name, *cur, strlen(trig->name))) {
cur               494 drivers/iio/trigger/stm32-timer-trigger.c 		cur++;
cur              2232 drivers/infiniband/core/device.c 	struct ib_port_data *cur;
cur              2235 drivers/infiniband/core/device.c 	hash_for_each_possible_rcu (ndev_hash, cur, ndev_hash_link,
cur              2237 drivers/infiniband/core/device.c 		if (rcu_access_pointer(cur->netdev) == ndev &&
cur              2239 drivers/infiniband/core/device.c 		     cur->ib_dev->ops.driver_id == driver_id) &&
cur              2240 drivers/infiniband/core/device.c 		    ib_device_try_get(cur->ib_dev)) {
cur              2241 drivers/infiniband/core/device.c 			res = cur->ib_dev;
cur               126 drivers/infiniband/core/uverbs_cmd.c 	const void __user *cur;
cur               141 drivers/infiniband/core/uverbs_cmd.c 	iter->cur = attrs->ucore.inbuf + req_len;
cur               149 drivers/infiniband/core/uverbs_cmd.c 	if (iter->cur + len > iter->end)
cur               152 drivers/infiniband/core/uverbs_cmd.c 	if (copy_from_user(val, iter->cur, len))
cur               155 drivers/infiniband/core/uverbs_cmd.c 	iter->cur += len;
cur               162 drivers/infiniband/core/uverbs_cmd.c 	const void __user *res = iter->cur;
cur               164 drivers/infiniband/core/uverbs_cmd.c 	if (iter->cur + len > iter->end)
cur               166 drivers/infiniband/core/uverbs_cmd.c 	iter->cur += len;
cur               172 drivers/infiniband/core/uverbs_cmd.c 	if (!ib_is_buffer_cleared(iter->cur, iter->end - iter->cur))
cur               476 drivers/infiniband/hw/cxgb4/device.c 			dev->rdev.stats.pd.total, dev->rdev.stats.pd.cur,
cur               479 drivers/infiniband/hw/cxgb4/device.c 			dev->rdev.stats.qid.total, dev->rdev.stats.qid.cur,
cur               482 drivers/infiniband/hw/cxgb4/device.c 		   dev->rdev.stats.srqt.total, dev->rdev.stats.srqt.cur,
cur               485 drivers/infiniband/hw/cxgb4/device.c 			dev->rdev.stats.stag.total, dev->rdev.stats.stag.cur,
cur               488 drivers/infiniband/hw/cxgb4/device.c 			dev->rdev.stats.pbl.total, dev->rdev.stats.pbl.cur,
cur               491 drivers/infiniband/hw/cxgb4/device.c 			dev->rdev.stats.rqt.total, dev->rdev.stats.rqt.cur,
cur               494 drivers/infiniband/hw/cxgb4/device.c 			dev->rdev.stats.ocqp.total, dev->rdev.stats.ocqp.cur,
cur               761 drivers/infiniband/hw/cxgb4/device.c 			rdev->stats.qid.cur -= rdev->qpmask + 1;
cur               122 drivers/infiniband/hw/cxgb4/iw_cxgb4.h 	u64 cur;
cur               302 drivers/infiniband/hw/cxgb4/mem.c 		rdev->stats.stag.cur += 32;
cur               303 drivers/infiniband/hw/cxgb4/mem.c 		if (rdev->stats.stag.cur > rdev->stats.stag.max)
cur               304 drivers/infiniband/hw/cxgb4/mem.c 			rdev->stats.stag.max = rdev->stats.stag.cur;
cur               339 drivers/infiniband/hw/cxgb4/mem.c 		rdev->stats.stag.cur -= 32;
cur               203 drivers/infiniband/hw/cxgb4/provider.c 	rhp->rdev.stats.pd.cur--;
cur               231 drivers/infiniband/hw/cxgb4/provider.c 	rhp->rdev.stats.pd.cur++;
cur               232 drivers/infiniband/hw/cxgb4/provider.c 	if (rhp->rdev.stats.pd.cur > rhp->rdev.stats.pd.max)
cur               233 drivers/infiniband/hw/cxgb4/provider.c 		rhp->rdev.stats.pd.max = rhp->rdev.stats.pd.cur;
cur               126 drivers/infiniband/hw/cxgb4/resource.c 		rdev->stats.qid.cur += rdev->qpmask + 1;
cur               157 drivers/infiniband/hw/cxgb4/resource.c 	if (rdev->stats.qid.cur > rdev->stats.qid.max)
cur               158 drivers/infiniband/hw/cxgb4/resource.c 		rdev->stats.qid.max = rdev->stats.qid.cur;
cur               200 drivers/infiniband/hw/cxgb4/resource.c 		rdev->stats.qid.cur += rdev->qpmask + 1;
cur               231 drivers/infiniband/hw/cxgb4/resource.c 	if (rdev->stats.qid.cur > rdev->stats.qid.max)
cur               232 drivers/infiniband/hw/cxgb4/resource.c 		rdev->stats.qid.max = rdev->stats.qid.cur;
cur               271 drivers/infiniband/hw/cxgb4/resource.c 		rdev->stats.pbl.cur += roundup(size, 1 << MIN_PBL_SHIFT);
cur               272 drivers/infiniband/hw/cxgb4/resource.c 		if (rdev->stats.pbl.cur > rdev->stats.pbl.max)
cur               273 drivers/infiniband/hw/cxgb4/resource.c 			rdev->stats.pbl.max = rdev->stats.pbl.cur;
cur               294 drivers/infiniband/hw/cxgb4/resource.c 	rdev->stats.pbl.cur -= roundup(size, 1 << MIN_PBL_SHIFT);
cur               353 drivers/infiniband/hw/cxgb4/resource.c 		rdev->stats.rqt.cur += roundup(size << 6, 1 << MIN_RQT_SHIFT);
cur               354 drivers/infiniband/hw/cxgb4/resource.c 		if (rdev->stats.rqt.cur > rdev->stats.rqt.max)
cur               355 drivers/infiniband/hw/cxgb4/resource.c 			rdev->stats.rqt.max = rdev->stats.rqt.cur;
cur               376 drivers/infiniband/hw/cxgb4/resource.c 	rdev->stats.rqt.cur -= roundup(size << 6, 1 << MIN_RQT_SHIFT);
cur               438 drivers/infiniband/hw/cxgb4/resource.c 	rdev->stats.srqt.cur++;
cur               439 drivers/infiniband/hw/cxgb4/resource.c 	if (rdev->stats.srqt.cur > rdev->stats.srqt.max)
cur               440 drivers/infiniband/hw/cxgb4/resource.c 		rdev->stats.srqt.max = rdev->stats.srqt.cur;
cur               449 drivers/infiniband/hw/cxgb4/resource.c 	rdev->stats.srqt.cur--;
cur               464 drivers/infiniband/hw/cxgb4/resource.c 		rdev->stats.ocqp.cur += roundup(size, 1 << MIN_OCQP_SHIFT);
cur               465 drivers/infiniband/hw/cxgb4/resource.c 		if (rdev->stats.ocqp.cur > rdev->stats.ocqp.max)
cur               466 drivers/infiniband/hw/cxgb4/resource.c 			rdev->stats.ocqp.max = rdev->stats.ocqp.cur;
cur               476 drivers/infiniband/hw/cxgb4/resource.c 	rdev->stats.ocqp.cur -= roundup(size, 1 << MIN_OCQP_SHIFT);
cur              1295 drivers/infiniband/hw/hns/hns_roce_hem.c 	struct roce_hem_item *cur, *pre;
cur              1326 drivers/infiniband/hw/hns/hns_roce_hem.c 		cur = hem_list_search_item(&mid_bt[level], offset);
cur              1327 drivers/infiniband/hw/hns/hns_roce_hem.c 		if (cur) {
cur              1328 drivers/infiniband/hw/hns/hns_roce_hem.c 			hem_ptrs[level] = cur;
cur              1340 drivers/infiniband/hw/hns/hns_roce_hem.c 		cur = hem_list_alloc_item(hr_dev, start_aligned, end, unit,
cur              1342 drivers/infiniband/hw/hns/hns_roce_hem.c 		if (!cur) {
cur              1346 drivers/infiniband/hw/hns/hns_roce_hem.c 		hem_ptrs[level] = cur;
cur              1347 drivers/infiniband/hw/hns/hns_roce_hem.c 		list_add(&cur->list, &temp_list[level]);
cur              1349 drivers/infiniband/hw/hns/hns_roce_hem.c 			list_add(&cur->sibling, &temp_list[0]);
cur              1354 drivers/infiniband/hw/hns/hns_roce_hem.c 			step = (cur->start - pre->start) / step * BA_BYTE_LEN;
cur              1356 drivers/infiniband/hw/hns/hns_roce_hem.c 					 cur->dma_addr);
cur              1291 drivers/infiniband/hw/hns/hns_roce_qp.c 	u32 cur;
cur              1293 drivers/infiniband/hw/hns/hns_roce_qp.c 	cur = hr_wq->head - hr_wq->tail;
cur              1294 drivers/infiniband/hw/hns/hns_roce_qp.c 	if (likely(cur + nreq < hr_wq->max_post))
cur              1299 drivers/infiniband/hw/hns/hns_roce_qp.c 	cur = hr_wq->head - hr_wq->tail;
cur              1302 drivers/infiniband/hw/hns/hns_roce_qp.c 	return cur + nreq >= hr_wq->max_post;
cur               616 drivers/infiniband/hw/mlx4/cq.c 	unsigned cur;
cur               620 drivers/infiniband/hw/mlx4/cq.c 	cur = wq->head - wq->tail;
cur               622 drivers/infiniband/hw/mlx4/cq.c 	if (cur == 0)
cur               625 drivers/infiniband/hw/mlx4/cq.c 	for (i = 0;  i < cur && *npolled < num_entries; i++) {
cur              3291 drivers/infiniband/hw/mlx4/qp.c 	unsigned cur;
cur              3294 drivers/infiniband/hw/mlx4/qp.c 	cur = wq->head - wq->tail;
cur              3295 drivers/infiniband/hw/mlx4/qp.c 	if (likely(cur + nreq < wq->max_post))
cur              3300 drivers/infiniband/hw/mlx4/qp.c 	cur = wq->head - wq->tail;
cur              3303 drivers/infiniband/hw/mlx4/qp.c 	return cur + nreq >= wq->max_post;
cur               390 drivers/infiniband/hw/mlx5/cq.c 	unsigned int cur;
cur               395 drivers/infiniband/hw/mlx5/cq.c 	cur = wq->head - wq->tail;
cur               398 drivers/infiniband/hw/mlx5/cq.c 	if (cur == 0)
cur               401 drivers/infiniband/hw/mlx5/cq.c 	for (i = 0;  i < cur && np < num_entries; i++) {
cur               135 drivers/infiniband/hw/mlx5/mem.c 	u64 cur = 0;
cur               175 drivers/infiniband/hw/mlx5/mem.c 				cur = base + (k << PAGE_SHIFT);
cur               176 drivers/infiniband/hw/mlx5/mem.c 				cur |= access_flags;
cur               179 drivers/infiniband/hw/mlx5/mem.c 				pas[idx] = cpu_to_be64(cur);
cur               680 drivers/infiniband/hw/mlx5/mlx5_ib.h 	u32                     cur;
cur               121 drivers/infiniband/hw/mlx5/mr.c 	ent->cur++;
cur               215 drivers/infiniband/hw/mlx5/mr.c 		ent->cur--;
cur               312 drivers/infiniband/hw/mlx5/mr.c 	if (ent->cur < ent->limit) {
cur               313 drivers/infiniband/hw/mlx5/mr.c 		err = add_keys(dev, c, 2 * ent->limit - ent->cur);
cur               347 drivers/infiniband/hw/mlx5/mr.c 		if (cache->ent[i].cur < cache->ent[i].limit)
cur               365 drivers/infiniband/hw/mlx5/mr.c 	if (ent->cur < 2 * ent->limit && !dev->fill_delay) {
cur               367 drivers/infiniband/hw/mlx5/mr.c 		if (ent->cur < 2 * ent->limit) {
cur               382 drivers/infiniband/hw/mlx5/mr.c 	} else if (ent->cur > 2 * ent->limit) {
cur               398 drivers/infiniband/hw/mlx5/mr.c 			if (ent->cur > ent->limit)
cur               449 drivers/infiniband/hw/mlx5/mr.c 			ent->cur--;
cur               451 drivers/infiniband/hw/mlx5/mr.c 			if (ent->cur < ent->limit)
cur               484 drivers/infiniband/hw/mlx5/mr.c 			ent->cur--;
cur               486 drivers/infiniband/hw/mlx5/mr.c 			if (ent->cur < ent->limit)
cur               518 drivers/infiniband/hw/mlx5/mr.c 		if (ent->cur < ent->limit)
cur               526 drivers/infiniband/hw/mlx5/mr.c 	ent->cur++;
cur               527 drivers/infiniband/hw/mlx5/mr.c 	if (ent->cur > 2 * ent->limit)
cur               552 drivers/infiniband/hw/mlx5/mr.c 		ent->cur--;
cur               595 drivers/infiniband/hw/mlx5/mr.c 		debugfs_create_u32("cur", 0400, dir, &ent->cur);
cur              4073 drivers/infiniband/hw/mlx5/qp.c 	unsigned cur;
cur              4075 drivers/infiniband/hw/mlx5/qp.c 	cur = wq->head - wq->tail;
cur              4076 drivers/infiniband/hw/mlx5/qp.c 	if (likely(cur + nreq < wq->max_post))
cur              4081 drivers/infiniband/hw/mlx5/qp.c 	cur = wq->head - wq->tail;
cur              4084 drivers/infiniband/hw/mlx5/qp.c 	return cur + nreq >= wq->max_post;
cur              1727 drivers/infiniband/hw/mthca/mthca_cmd.c int mthca_MODIFY_QP(struct mthca_dev *dev, enum ib_qp_state cur,
cur              1775 drivers/infiniband/hw/mthca/mthca_cmd.c 	if (op[cur][next] == CMD_ERR2RST_QPEE) {
cur              1790 drivers/infiniband/hw/mthca/mthca_cmd.c 				    op[cur][next], CMD_TIME_CLASS_C);
cur              1824 drivers/infiniband/hw/mthca/mthca_cmd.c 				op_mod, op[cur][next], CMD_TIME_CLASS_C);
cur               308 drivers/infiniband/hw/mthca/mthca_cmd.h int mthca_MODIFY_QP(struct mthca_dev *dev, enum ib_qp_state cur,
cur              1569 drivers/infiniband/hw/mthca/mthca_qp.c 	unsigned cur;
cur              1572 drivers/infiniband/hw/mthca/mthca_qp.c 	cur = wq->head - wq->tail;
cur              1573 drivers/infiniband/hw/mthca/mthca_qp.c 	if (likely(cur + nreq < wq->max))
cur              1578 drivers/infiniband/hw/mthca/mthca_qp.c 	cur = wq->head - wq->tail;
cur              1581 drivers/infiniband/hw/mthca/mthca_qp.c 	return cur + nreq >= wq->max;
cur               906 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	struct list_head *cur;
cur               910 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 	list_for_each(cur, head) {
cur               912 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 			qp = list_entry(cur, struct ocrdma_qp, sq_entry);
cur               914 drivers/infiniband/hw/ocrdma/ocrdma_hw.c 			qp = list_entry(cur, struct ocrdma_qp, rq_entry);
cur              2957 drivers/infiniband/hw/qedr/verbs.c 			u32 cur;
cur              2968 drivers/infiniband/hw/qedr/verbs.c 			cur = min_t(u32, len, seg_siz);
cur              2969 drivers/infiniband/hw/qedr/verbs.c 			memcpy(seg_prt, src, cur);
cur              2972 drivers/infiniband/hw/qedr/verbs.c 			seg_prt += cur;
cur              2973 drivers/infiniband/hw/qedr/verbs.c 			seg_siz -= cur;
cur              2976 drivers/infiniband/hw/qedr/verbs.c 			src += cur;
cur              2977 drivers/infiniband/hw/qedr/verbs.c 			len -= cur;
cur               202 drivers/iommu/arm-smmu.c 		it.cur = &pci_sid;
cur               738 drivers/iommu/dma-iommu.c 	struct scatterlist *s, *cur = sg;
cur               768 drivers/iommu/dma-iommu.c 				cur = sg_next(cur);
cur               772 drivers/iommu/dma-iommu.c 			sg_dma_address(cur) = dma_addr + s_iova_off;
cur               775 drivers/iommu/dma-iommu.c 		sg_dma_len(cur) = cur_len;
cur               465 drivers/iommu/virtio-iommu.c 	size_t cur = 0;
cur               496 drivers/iommu/virtio-iommu.c 	       cur < viommu->probe_size) {
cur               510 drivers/iommu/virtio-iommu.c 		cur += len;
cur               511 drivers/iommu/virtio-iommu.c 		if (cur >= viommu->probe_size)
cur               514 drivers/iommu/virtio-iommu.c 		prop = (void *)probe->properties + cur;
cur               246 drivers/leds/leds-powernv.c 	const char *cur = NULL;
cur               256 drivers/leds/leds-powernv.c 		while ((cur = of_prop_next_string(p, cur)) != NULL) {
cur               267 drivers/leds/leds-powernv.c 			rc = powernv_led_create(dev, powernv_led, cur);
cur               952 drivers/lightnvm/pblk-core.c 				  struct pblk_line *cur)
cur              1002 drivers/lightnvm/pblk-core.c 	if (cur) {
cur              1003 drivers/lightnvm/pblk-core.c 		memcpy(line->lun_bitmap, cur->lun_bitmap, lm->lun_bitmap_len);
cur              1004 drivers/lightnvm/pblk-core.c 		smeta_buf->prev_id = cpu_to_le32(cur->id);
cur              1005 drivers/lightnvm/pblk-core.c 		cur->emeta->buf->next_id = cpu_to_le32(line->id);
cur              1546 drivers/lightnvm/pblk-core.c 	struct pblk_line *cur, *new = NULL;
cur              1554 drivers/lightnvm/pblk-core.c 	cur = l_mg->data_line;
cur              1577 drivers/lightnvm/pblk-core.c 	if (!pblk_line_init_metadata(pblk, new, cur)) {
cur               705 drivers/md/bcache/journal.c 	j->cur = (j->cur == j->w)
cur               716 drivers/md/bcache/journal.c 	j->cur->data->seq	= ++j->seq;
cur               717 drivers/md/bcache/journal.c 	j->cur->dirty		= false;
cur               718 drivers/md/bcache/journal.c 	j->cur->need_write	= false;
cur               719 drivers/md/bcache/journal.c 	j->cur->data->keys	= 0;
cur               738 drivers/md/bcache/journal.c 	struct journal_write *w = (j->cur == j->w)
cur               760 drivers/md/bcache/journal.c 	struct journal_write *w = c->journal.cur;
cur               849 drivers/md/bcache/journal.c 	struct journal_write *w = c->journal.cur;
cur               874 drivers/md/bcache/journal.c 		struct journal_write *w = c->journal.cur;
cur               922 drivers/md/bcache/journal.c 	if (c->journal.cur->dirty)
cur               121 drivers/md/bcache/journal.h 	struct journal_write	w[2], *cur;
cur              1645 drivers/md/bcache/super.c 	if (c->journal.cur) {
cur              1320 drivers/md/dm-ioctl.c static bool is_valid_type(enum dm_queue_mode cur, enum dm_queue_mode new)
cur              1322 drivers/md/dm-ioctl.c 	if (cur == new ||
cur              1323 drivers/md/dm-ioctl.c 	    (cur == DM_TYPE_BIO_BASED && new == DM_TYPE_DAX_BIO_BASED))
cur              1318 drivers/media/common/cx2341x.c 	return ctrl && ctrl->val != ctrl->cur.val;
cur               312 drivers/media/common/saa7146/saa7146_hlp.c 	int cur = 0, next = 0;
cur               328 drivers/media/common/saa7146/saa7146_hlp.c 	for( cur = 0, next = 1; next < *count; next++) {
cur               329 drivers/media/common/saa7146/saa7146_hlp.c 		if( values[cur] != values[next])
cur               330 drivers/media/common/saa7146/saa7146_hlp.c 			values[++cur] = values[next];
cur               333 drivers/media/common/saa7146/saa7146_hlp.c 	*count = cur + 1;
cur               150 drivers/media/i2c/adp1653.c 	ctrl->cur.val = 0;
cur               153 drivers/media/i2c/adp1653.c 		ctrl->cur.val |= V4L2_FLASH_FAULT_SHORT_CIRCUIT;
cur               155 drivers/media/i2c/adp1653.c 		ctrl->cur.val |= V4L2_FLASH_FAULT_OVER_TEMPERATURE;
cur               157 drivers/media/i2c/adp1653.c 		ctrl->cur.val |= V4L2_FLASH_FAULT_TIMEOUT;
cur               159 drivers/media/i2c/adp1653.c 		ctrl->cur.val |= V4L2_FLASH_FAULT_OVER_VOLTAGE;
cur              1162 drivers/media/i2c/imx274.c 	v_pos = imx274->ctrls.vflip->cur.val ?
cur               179 drivers/media/i2c/lm3560.c 		ctrl->cur.val = fault;
cur               191 drivers/media/i2c/m5mols/m5mols_controls.c 	if ((ctrl->val ^ ctrl->cur.val) & V4L2_LOCK_EXPOSURE) {
cur               200 drivers/media/i2c/m5mols/m5mols_controls.c 	if (((ctrl->val ^ ctrl->cur.val) & V4L2_LOCK_WHITE_BALANCE)
cur               213 drivers/media/i2c/m5mols/m5mols_controls.c 	if ((ctrl->val ^ ctrl->cur.val) & V4L2_LOCK_FOCUS)
cur               791 drivers/media/i2c/max2175.c 	max2175_set_hsls(ctx, ctx->hsls->cur.val);
cur               794 drivers/media/i2c/max2175.c 	max2175_i2s_enable(ctx, ctx->i2s_en->cur.val);
cur              1012 drivers/media/i2c/max2175.c 		max2175_tune_rf_freq(ctx, ctx->freq, ctx->hsls->cur.val);
cur              1016 drivers/media/i2c/max2175.c 				     ctx->hsls->cur.val);
cur              1088 drivers/media/i2c/max2175.c 	ctx->rx_mode->cur.val = rx_mode;
cur              1091 drivers/media/i2c/max2175.c 	return max2175_tune_rf_freq(ctx, freq, ctx->hsls->cur.val);
cur              1112 drivers/media/i2c/max2175.c 	    max2175_freq_rx_mode_valid(ctx, ctx->rx_mode->cur.val, freq))
cur              1113 drivers/media/i2c/max2175.c 		ret = max2175_tune_rf_freq(ctx, freq, ctx->hsls->cur.val);
cur              1119 drivers/media/i2c/max2175.c 		ret, ctx->freq, ctx->mode_resolved, ctx->rx_mode->cur.val);
cur               400 drivers/media/i2c/msp3400-driver.c 	state->volume->val = state->volume->cur.val;
cur               401 drivers/media/i2c/msp3400-driver.c 	state->muted->val = state->muted->cur.val;
cur               657 drivers/media/i2c/mt9p031.c 	if (mt9p031->blc_auto->cur.val != 0) {
cur               664 drivers/media/i2c/mt9p031.c 	if (mt9p031->blc_offset->cur.val != 0) {
cur               666 drivers/media/i2c/mt9p031.c 				    mt9p031->blc_offset->cur.val);
cur               552 drivers/media/i2c/mt9t001.c 			if (gain->val != gain->cur.val)
cur               566 drivers/media/i2c/mt9t001.c 			if (gain->val == gain->cur.val)
cur               190 drivers/media/i2c/s5c73m3/s5c73m3-ctrls.c 	if ((ctrl->val ^ ctrl->cur.val) & V4L2_LOCK_EXPOSURE) {
cur               197 drivers/media/i2c/s5c73m3/s5c73m3-ctrls.c 	if (((ctrl->val ^ ctrl->cur.val) & V4L2_LOCK_WHITE_BALANCE)
cur               205 drivers/media/i2c/s5c73m3/s5c73m3-ctrls.c 	if ((ctrl->val ^ ctrl->cur.val) & V4L2_LOCK_FOCUS)
cur               251 drivers/media/pci/cx18/cx18-av-core.c 	state->volume->cur.val = state->volume->default_value = default_volume;
cur               745 drivers/media/pci/cx18/cx18-driver.c 	cx->temporal_strength = cx->cxhdl.video_temporal_filter->cur.val;
cur               746 drivers/media/pci/cx18/cx18-driver.c 	cx->spatial_strength = cx->cxhdl.video_spatial_filter->cur.val;
cur               747 drivers/media/pci/cx18/cx18-driver.c 	cx->filter_mode = cx->cxhdl.video_spatial_filter_mode->cur.val |
cur               748 drivers/media/pci/cx18/cx18-driver.c 		(cx->cxhdl.video_temporal_filter_mode->cur.val << 1) |
cur               749 drivers/media/pci/cx18/cx18-driver.c 		(cx->cxhdl.video_median_filter_type->cur.val << 2);
cur               168 drivers/media/pci/mantis/mantis_core.c 	u32 cur;
cur               170 drivers/media/pci/mantis/mantis_core.c 	cur = mmread(MANTIS_GPIF_ADDR);
cur               172 drivers/media/pci/mantis/mantis_core.c 		mantis->gpio_status = cur | (1 << bitpos);
cur               174 drivers/media/pci/mantis/mantis_core.c 		mantis->gpio_status = cur & (~(1 << bitpos));
cur                69 drivers/media/pci/mantis/mantis_ioc.c 	u32 cur;
cur                72 drivers/media/pci/mantis/mantis_ioc.c 	cur = mmread(MANTIS_GPIF_ADDR);
cur                74 drivers/media/pci/mantis/mantis_ioc.c 		mantis->gpio_status = cur | (1 << bitpos);
cur                76 drivers/media/pci/mantis/mantis_ioc.c 		mantis->gpio_status = cur & (~(1 << bitpos));
cur               643 drivers/media/platform/exynos4-is/fimc-core.c 		fimc_set_color_effect(ctx, ctrls->colorfx->cur.val);
cur               668 drivers/media/platform/exynos4-is/fimc-core.c 	if (ctrl->cur.val > ctrl->maximum)
cur               669 drivers/media/platform/exynos4-is/fimc-core.c 		ctrl->cur.val = ctrl->maximum;
cur               810 drivers/media/platform/qcom/camss/camss-csid.c 		if (csid->testgen_mode->cur.val == 0) {
cur               860 drivers/media/platform/qcom/camss/camss-csid.c 		if (csid->testgen_mode->cur.val == 0) {
cur              1250 drivers/media/platform/qcom/camss/camss-csid.c 		if (csid->testgen_mode->cur.val != 0)
cur                19 drivers/media/platform/sti/delta/delta-mjpeg-hdr.c 	char *cur = str;
cur                25 drivers/media/platform/sti/delta/delta-mjpeg-hdr.c 	snprintf(cur, left, "[MJPEG header]\n"
cur               813 drivers/media/platform/ti-vpe/vpe.c 	const struct vpe_dei_regs *cur = &dei_regs;
cur               815 drivers/media/platform/ti-vpe/vpe.c 	dei_mmr[2]  = cur->mdt_spacial_freq_thr_reg;
cur               816 drivers/media/platform/ti-vpe/vpe.c 	dei_mmr[3]  = cur->edi_config_reg;
cur               817 drivers/media/platform/ti-vpe/vpe.c 	dei_mmr[4]  = cur->edi_lut_reg0;
cur               818 drivers/media/platform/ti-vpe/vpe.c 	dei_mmr[5]  = cur->edi_lut_reg1;
cur               819 drivers/media/platform/ti-vpe/vpe.c 	dei_mmr[6]  = cur->edi_lut_reg2;
cur               820 drivers/media/platform/ti-vpe/vpe.c 	dei_mmr[7]  = cur->edi_lut_reg3;
cur               615 drivers/media/platform/vicodec/codec-fwht.c decide_blocktype(const u8 *cur, const u8 *reference, s16 *deltablock,
cur               625 drivers/media/platform/vicodec/codec-fwht.c 	fill_encoder_block(cur, tmp, stride, input_step);
cur               500 drivers/media/platform/vivid/vivid-kthread-cap.c 			dev->brightness->cur.val,
cur               501 drivers/media/platform/vivid/vivid-kthread-cap.c 			dev->contrast->cur.val,
cur               502 drivers/media/platform/vivid/vivid-kthread-cap.c 			dev->saturation->cur.val,
cur               503 drivers/media/platform/vivid/vivid-kthread-cap.c 			dev->hue->cur.val);
cur               507 drivers/media/platform/vivid/vivid-kthread-cap.c 			dev->autogain->cur.val, gain, dev->alpha->cur.val);
cur               513 drivers/media/platform/vivid/vivid-kthread-cap.c 			dev->volume->cur.val, dev->mute->cur.val);
cur               518 drivers/media/platform/vivid/vivid-kthread-cap.c 			dev->int32->cur.val,
cur               520 drivers/media/platform/vivid/vivid-kthread-cap.c 			dev->bitmask->cur.val);
cur               523 drivers/media/platform/vivid/vivid-kthread-cap.c 			dev->boolean->cur.val,
cur               524 drivers/media/platform/vivid/vivid-kthread-cap.c 			dev->menu->qmenu[dev->menu->cur.val],
cur               528 drivers/media/platform/vivid/vivid-kthread-cap.c 			dev->int_menu->qmenu_int[dev->int_menu->cur.val],
cur               529 drivers/media/platform/vivid/vivid-kthread-cap.c 			dev->int_menu->cur.val);
cur                70 drivers/media/platform/vivid/vivid-radio-common.c 		rds->picode = dev->radio_tx_rds_pi->cur.val;
cur                71 drivers/media/platform/vivid/vivid-radio-common.c 		rds->pty = dev->radio_tx_rds_pty->cur.val;
cur                72 drivers/media/platform/vivid/vivid-radio-common.c 		rds->mono_stereo = dev->radio_tx_rds_mono_stereo->cur.val;
cur                73 drivers/media/platform/vivid/vivid-radio-common.c 		rds->art_head = dev->radio_tx_rds_art_head->cur.val;
cur                74 drivers/media/platform/vivid/vivid-radio-common.c 		rds->compressed = dev->radio_tx_rds_compressed->cur.val;
cur                75 drivers/media/platform/vivid/vivid-radio-common.c 		rds->dyn_pty = dev->radio_tx_rds_dyn_pty->cur.val;
cur                76 drivers/media/platform/vivid/vivid-radio-common.c 		rds->ta = dev->radio_tx_rds_ta->cur.val;
cur                77 drivers/media/platform/vivid/vivid-radio-common.c 		rds->tp = dev->radio_tx_rds_tp->cur.val;
cur                78 drivers/media/platform/vivid/vivid-radio-common.c 		rds->ms = dev->radio_tx_rds_ms->cur.val;
cur               159 drivers/media/platform/vsp1/vsp1_hgo.c 	hgo->max_rgb = hgo->ctrls.max_rgb->cur.val;
cur               161 drivers/media/platform/vsp1/vsp1_hgo.c 		hgo->num_bins = hgo_num_bins[hgo->ctrls.num_bins->cur.val];
cur               219 drivers/media/platform/xilinx/xilinx-tpg.c 			 XTPG_PATTERN_MASK, xtpg->pattern->cur.val);
cur               225 drivers/media/platform/xilinx/xilinx-tpg.c 	passthrough = xtpg->pattern->cur.val == 0;
cur               281 drivers/media/radio/radio-isa.c 	ops->s_mute_volume(isa, true, isa->volume ? isa->volume->cur.val : 0);
cur               158 drivers/media/radio/radio-sf16fmr2.c 		balance = fmr2->balance->cur.val;
cur               162 drivers/media/radio/radio-sf16fmr2.c 		volume = fmr2->volume->cur.val;
cur               382 drivers/media/tuners/e4000.c 		dev->lna_gain_auto->cur.val, dev->lna_gain_auto->val,
cur               383 drivers/media/tuners/e4000.c 		dev->lna_gain->cur.val, dev->lna_gain->val);
cur               385 drivers/media/tuners/e4000.c 	if (dev->lna_gain_auto->val && dev->if_gain_auto->cur.val)
cur               389 drivers/media/tuners/e4000.c 	else if (dev->if_gain_auto->cur.val)
cur               418 drivers/media/tuners/e4000.c 		dev->mixer_gain_auto->cur.val, dev->mixer_gain_auto->val,
cur               419 drivers/media/tuners/e4000.c 		dev->mixer_gain->cur.val, dev->mixer_gain->val);
cur               451 drivers/media/tuners/e4000.c 		dev->if_gain_auto->cur.val, dev->if_gain_auto->val,
cur               452 drivers/media/tuners/e4000.c 		dev->if_gain->cur.val, dev->if_gain->val);
cur               454 drivers/media/tuners/e4000.c 	if (dev->if_gain_auto->val && dev->lna_gain_auto->cur.val)
cur               456 drivers/media/tuners/e4000.c 	else if (dev->lna_gain_auto->cur.val)
cur               474 drivers/media/tuners/fc2580.c 		ctrl->id, ctrl->name, ctrl->cur.val, ctrl->val);
cur               267 drivers/media/tuners/msi001.c 	ret = msi001_set_gain(dev, dev->lna_gain->cur.val,
cur               268 drivers/media/tuners/msi001.c 			      dev->mixer_gain->cur.val, dev->if_gain->cur.val);
cur               397 drivers/media/tuners/msi001.c 				      dev->mixer_gain->cur.val,
cur               398 drivers/media/tuners/msi001.c 				      dev->if_gain->cur.val);
cur               401 drivers/media/tuners/msi001.c 		ret = msi001_set_gain(dev, dev->lna_gain->cur.val,
cur               403 drivers/media/tuners/msi001.c 				      dev->if_gain->cur.val);
cur               406 drivers/media/tuners/msi001.c 		ret = msi001_set_gain(dev, dev->lna_gain->cur.val,
cur               407 drivers/media/tuners/msi001.c 				      dev->mixer_gain->cur.val,
cur               868 drivers/media/usb/airspy/airspy.c 			s->lna_gain_auto->cur.val, s->lna_gain_auto->val,
cur               869 drivers/media/usb/airspy/airspy.c 			s->lna_gain->cur.val, s->lna_gain->val);
cur               895 drivers/media/usb/airspy/airspy.c 			s->mixer_gain_auto->cur.val, s->mixer_gain_auto->val,
cur               896 drivers/media/usb/airspy/airspy.c 			s->mixer_gain->cur.val, s->mixer_gain->val);
cur               921 drivers/media/usb/airspy/airspy.c 	dev_dbg(s->dev, "val=%d->%d\n", s->if_gain->cur.val, s->if_gain->val);
cur               878 drivers/media/usb/gspca/conex.c 		setbrightness(gspca_dev, ctrl->val, sd->sat->cur.val);
cur               881 drivers/media/usb/gspca/conex.c 		setcontrast(gspca_dev, ctrl->val, sd->sat->cur.val);
cur               884 drivers/media/usb/gspca/conex.c 		setbrightness(gspca_dev, sd->brightness->cur.val, ctrl->val);
cur               885 drivers/media/usb/gspca/conex.c 		setcontrast(gspca_dev, sd->contrast->cur.val, ctrl->val);
cur              2213 drivers/media/usb/gspca/sn9c20x.c 			s32 curqual = sd->jpegqual->cur.val;
cur              2221 drivers/media/usb/gspca/sn9c20x.c 				sd->jpegqual->cur.val = new_qual;
cur              3970 drivers/media/usb/gspca/topro.c 		s32 old = gspca_dev->gain->cur.val ?
cur              3971 drivers/media/usb/gspca/topro.c 					gspca_dev->gain->cur.val : 1;
cur               311 drivers/media/usb/hackrf/hackrf.c 		dev->rx_bandwidth->cur.val = uitmp;
cur               327 drivers/media/usb/hackrf/hackrf.c 		dev->tx_bandwidth->cur.val = uitmp;
cur               376 drivers/media/usb/hackrf/hackrf.c 			dev->rx_rf_gain->cur.val, dev->rx_rf_gain->val);
cur               388 drivers/media/usb/hackrf/hackrf.c 			dev->tx_rf_gain->cur.val, dev->tx_rf_gain->val);
cur               400 drivers/media/usb/hackrf/hackrf.c 			dev->rx_lna_gain->cur.val, dev->rx_lna_gain->val);
cur               411 drivers/media/usb/hackrf/hackrf.c 			dev->rx_if_gain->cur.val, dev->rx_if_gain->val);
cur               422 drivers/media/usb/hackrf/hackrf.c 			dev->tx_lna_gain->cur.val, dev->tx_lna_gain->val);
cur                47 drivers/media/v4l2-core/v4l2-ctrls.c 	return master->is_auto && master->cur.val == master->manual_mode_value;
cur              2496 drivers/media/v4l2-core/v4l2-ctrls.c 	ctrl->cur.val = ctrl->val = def;
cur              2504 drivers/media/v4l2-core/v4l2-ctrls.c 		ctrl->p_cur.p = &ctrl->cur.val;
cur               130 drivers/mfd/ti_am335x_tscadc.c 	const __be32            *cur;
cur               147 drivers/mfd/ti_am335x_tscadc.c 	of_property_for_each_u32(node, "ti,adc-channels", prop, cur, val) {
cur               220 drivers/misc/cxl/guest.c 	struct irq_avail *cur;
cur               223 drivers/misc/cxl/guest.c 		cur = &adapter->guest->irq_avail[i];
cur               224 drivers/misc/cxl/guest.c 		n = bitmap_find_next_zero_area(cur->bitmap, cur->range,
cur               226 drivers/misc/cxl/guest.c 		if (n < cur->range) {
cur               227 drivers/misc/cxl/guest.c 			bitmap_set(cur->bitmap, n, len);
cur               228 drivers/misc/cxl/guest.c 			*irq = cur->offset + n;
cur               241 drivers/misc/cxl/guest.c 	struct irq_avail *cur;
cur               247 drivers/misc/cxl/guest.c 		cur = &adapter->guest->irq_avail[i];
cur               248 drivers/misc/cxl/guest.c 		if (irq >= cur->offset &&
cur               249 drivers/misc/cxl/guest.c 			(irq + len) <= (cur->offset + cur->range)) {
cur               250 drivers/misc/cxl/guest.c 			n = irq - cur->offset;
cur               251 drivers/misc/cxl/guest.c 			bitmap_clear(cur->bitmap, n, len);
cur              1043 drivers/misc/cxl/guest.c 	struct irq_avail *cur;
cur              1049 drivers/misc/cxl/guest.c 				cur = &adapter->guest->irq_avail[i];
cur              1050 drivers/misc/cxl/guest.c 				kfree(cur->bitmap);
cur               287 drivers/misc/cxl/of.c 	struct irq_avail *cur;
cur               308 drivers/misc/cxl/of.c 		cur = &adapter->guest->irq_avail[i];
cur               309 drivers/misc/cxl/of.c 		cur->offset = be32_to_cpu(ranges[i * 2]);
cur               310 drivers/misc/cxl/of.c 		cur->range  = be32_to_cpu(ranges[i * 2 + 1]);
cur               311 drivers/misc/cxl/of.c 		cur->bitmap = kcalloc(BITS_TO_LONGS(cur->range),
cur               312 drivers/misc/cxl/of.c 				sizeof(*cur->bitmap), GFP_KERNEL);
cur               313 drivers/misc/cxl/of.c 		if (cur->bitmap == NULL)
cur               315 drivers/misc/cxl/of.c 		if (cur->offset < adapter->guest->irq_base_offset)
cur               316 drivers/misc/cxl/of.c 			adapter->guest->irq_base_offset = cur->offset;
cur               319 drivers/misc/cxl/of.c 				cur->offset, cur->offset + cur->range - 1,
cur               320 drivers/misc/cxl/of.c 				cur->range);
cur               328 drivers/misc/cxl/of.c 		cur = &adapter->guest->irq_avail[i];
cur               329 drivers/misc/cxl/of.c 		kfree(cur->bitmap);
cur               183 drivers/misc/ibmvmc.c 	queue->cur = 0;
cur               211 drivers/misc/ibmvmc.c 	crq = &queue->msgs[queue->cur];
cur               213 drivers/misc/ibmvmc.c 		if (++queue->cur == queue->size)
cur               214 drivers/misc/ibmvmc.c 			queue->cur = 0;
cur              2150 drivers/misc/ibmvmc.c 	queue->cur = 0;
cur               154 drivers/misc/ibmvmc.h 	int size, cur;
cur               442 drivers/misc/lkdtm/core.c 		struct crashpoint *cur = &crashpoints[i];
cur               444 drivers/misc/lkdtm/core.c 		debugfs_create_file(cur->name, 0644, lkdtm_debugfs_root, cur,
cur               445 drivers/misc/lkdtm/core.c 				    &cur->fops);
cur                15 drivers/misc/ocxl/pasid.c 	struct id_range *cur;
cur                18 drivers/misc/ocxl/pasid.c 	list_for_each_entry(cur, head, list) {
cur                19 drivers/misc/ocxl/pasid.c 		pr_debug("Range %d->%d\n", cur->start, cur->end);
cur                28 drivers/misc/ocxl/pasid.c 	struct id_range *cur, *new;
cur                37 drivers/misc/ocxl/pasid.c 	list_for_each_entry(cur, head, list) {
cur                38 drivers/misc/ocxl/pasid.c 		if ((cur->start - last_end) > size)
cur                40 drivers/misc/ocxl/pasid.c 		last_end = cur->end;
cur                41 drivers/misc/ocxl/pasid.c 		pos = &cur->list;
cur                65 drivers/misc/ocxl/pasid.c 	struct id_range *cur, *tmp;
cur                67 drivers/misc/ocxl/pasid.c 	list_for_each_entry_safe(cur, tmp, head, list) {
cur                68 drivers/misc/ocxl/pasid.c 		if (cur->start == start && cur->end == (start + size - 1)) {
cur                70 drivers/misc/ocxl/pasid.c 			list_del(&cur->list);
cur                71 drivers/misc/ocxl/pasid.c 			kfree(cur);
cur                49 drivers/misc/vmw_vmci/vmci_event.c 		struct vmci_subscription *cur, *p2;
cur                50 drivers/misc/vmw_vmci/vmci_event.c 		list_for_each_entry_safe(cur, p2, &subscriber_array[e], node) {
cur                58 drivers/misc/vmw_vmci/vmci_event.c 			list_del(&cur->node);
cur                59 drivers/misc/vmw_vmci/vmci_event.c 			kfree(cur);
cur                72 drivers/misc/vmw_vmci/vmci_event.c 		struct vmci_subscription *cur;
cur                73 drivers/misc/vmw_vmci/vmci_event.c 		list_for_each_entry(cur, &subscriber_array[e], node) {
cur                74 drivers/misc/vmw_vmci/vmci_event.c 			if (cur->id == sub_id)
cur                75 drivers/misc/vmw_vmci/vmci_event.c 				return cur;
cur                87 drivers/misc/vmw_vmci/vmci_event.c 	struct vmci_subscription *cur;
cur                92 drivers/misc/vmw_vmci/vmci_event.c 	list_for_each_entry_rcu(cur, subscriber_list, node) {
cur                93 drivers/misc/vmw_vmci/vmci_event.c 		cur->callback(cur->id, &event_msg->event_data,
cur                94 drivers/misc/vmw_vmci/vmci_event.c 			      cur->callback_data);
cur               625 drivers/mmc/host/mmci.c 	struct dma_chan	*cur;
cur               733 drivers/mmc/host/mmci.c 	dmaengine_terminate_all(dmae->cur);
cur               735 drivers/mmc/host/mmci.c 	dmae->cur = NULL;
cur               783 drivers/mmc/host/mmci.c 	dmae->cur = NULL;
cur               864 drivers/mmc/host/mmci.c 	if (dmae->cur && dmae->desc_current)
cur               868 drivers/mmc/host/mmci.c 	return _mmci_dmae_prep_data(host, data, &dmae->cur,
cur               878 drivers/mmc/host/mmci.c 	dma_async_issue_pending(dmae->cur);
cur               896 drivers/mmc/host/mmci.c 	dmae->cur = next->chan;
cur               924 drivers/mmc/host/mmci.c 		if (dmae->cur == next->chan) {
cur               926 drivers/mmc/host/mmci.c 			dmae->cur = NULL;
cur               206 drivers/mtd/mtdswap.c 	struct swap_eb *cur;
cur               211 drivers/mtd/mtdswap.c 		cur = rb_entry(parent, struct swap_eb, rb);
cur               212 drivers/mtd/mtdswap.c 		if (eb->erase_count > cur->erase_count)
cur               130 drivers/mtd/nand/raw/nand_ecc.c 	uint32_t cur;		/* current value in buffer */
cur               160 drivers/mtd/nand/raw/nand_ecc.c 		cur = *bp++;
cur               161 drivers/mtd/nand/raw/nand_ecc.c 		tmppar = cur;
cur               162 drivers/mtd/nand/raw/nand_ecc.c 		rp4 ^= cur;
cur               163 drivers/mtd/nand/raw/nand_ecc.c 		cur = *bp++;
cur               164 drivers/mtd/nand/raw/nand_ecc.c 		tmppar ^= cur;
cur               166 drivers/mtd/nand/raw/nand_ecc.c 		cur = *bp++;
cur               167 drivers/mtd/nand/raw/nand_ecc.c 		tmppar ^= cur;
cur               168 drivers/mtd/nand/raw/nand_ecc.c 		rp4 ^= cur;
cur               169 drivers/mtd/nand/raw/nand_ecc.c 		cur = *bp++;
cur               170 drivers/mtd/nand/raw/nand_ecc.c 		tmppar ^= cur;
cur               173 drivers/mtd/nand/raw/nand_ecc.c 		cur = *bp++;
cur               174 drivers/mtd/nand/raw/nand_ecc.c 		tmppar ^= cur;
cur               175 drivers/mtd/nand/raw/nand_ecc.c 		rp4 ^= cur;
cur               176 drivers/mtd/nand/raw/nand_ecc.c 		rp6 ^= cur;
cur               177 drivers/mtd/nand/raw/nand_ecc.c 		cur = *bp++;
cur               178 drivers/mtd/nand/raw/nand_ecc.c 		tmppar ^= cur;
cur               179 drivers/mtd/nand/raw/nand_ecc.c 		rp6 ^= cur;
cur               180 drivers/mtd/nand/raw/nand_ecc.c 		cur = *bp++;
cur               181 drivers/mtd/nand/raw/nand_ecc.c 		tmppar ^= cur;
cur               182 drivers/mtd/nand/raw/nand_ecc.c 		rp4 ^= cur;
cur               183 drivers/mtd/nand/raw/nand_ecc.c 		cur = *bp++;
cur               184 drivers/mtd/nand/raw/nand_ecc.c 		tmppar ^= cur;
cur               187 drivers/mtd/nand/raw/nand_ecc.c 		cur = *bp++;
cur               188 drivers/mtd/nand/raw/nand_ecc.c 		tmppar ^= cur;
cur               189 drivers/mtd/nand/raw/nand_ecc.c 		rp4 ^= cur;
cur               190 drivers/mtd/nand/raw/nand_ecc.c 		rp6 ^= cur;
cur               191 drivers/mtd/nand/raw/nand_ecc.c 		rp8 ^= cur;
cur               192 drivers/mtd/nand/raw/nand_ecc.c 		cur = *bp++;
cur               193 drivers/mtd/nand/raw/nand_ecc.c 		tmppar ^= cur;
cur               194 drivers/mtd/nand/raw/nand_ecc.c 		rp6 ^= cur;
cur               195 drivers/mtd/nand/raw/nand_ecc.c 		rp8 ^= cur;
cur               196 drivers/mtd/nand/raw/nand_ecc.c 		cur = *bp++;
cur               197 drivers/mtd/nand/raw/nand_ecc.c 		tmppar ^= cur;
cur               198 drivers/mtd/nand/raw/nand_ecc.c 		rp4 ^= cur;
cur               199 drivers/mtd/nand/raw/nand_ecc.c 		rp8 ^= cur;
cur               200 drivers/mtd/nand/raw/nand_ecc.c 		cur = *bp++;
cur               201 drivers/mtd/nand/raw/nand_ecc.c 		tmppar ^= cur;
cur               202 drivers/mtd/nand/raw/nand_ecc.c 		rp8 ^= cur;
cur               204 drivers/mtd/nand/raw/nand_ecc.c 		cur = *bp++;
cur               205 drivers/mtd/nand/raw/nand_ecc.c 		tmppar ^= cur;
cur               206 drivers/mtd/nand/raw/nand_ecc.c 		rp4 ^= cur;
cur               207 drivers/mtd/nand/raw/nand_ecc.c 		rp6 ^= cur;
cur               208 drivers/mtd/nand/raw/nand_ecc.c 		cur = *bp++;
cur               209 drivers/mtd/nand/raw/nand_ecc.c 		tmppar ^= cur;
cur               210 drivers/mtd/nand/raw/nand_ecc.c 		rp6 ^= cur;
cur               211 drivers/mtd/nand/raw/nand_ecc.c 		cur = *bp++;
cur               212 drivers/mtd/nand/raw/nand_ecc.c 		tmppar ^= cur;
cur               213 drivers/mtd/nand/raw/nand_ecc.c 		rp4 ^= cur;
cur               214 drivers/mtd/nand/raw/nand_ecc.c 		cur = *bp++;
cur               215 drivers/mtd/nand/raw/nand_ecc.c 		tmppar ^= cur;
cur               409 drivers/mtd/nand/raw/s3c2410.c 	unsigned long cur;
cur               417 drivers/mtd/nand/raw/s3c2410.c 	cur = readl(info->sel_reg);
cur               420 drivers/mtd/nand/raw/s3c2410.c 		cur |= info->sel_bit;
cur               432 drivers/mtd/nand/raw/s3c2410.c 		cur &= ~info->sel_bit;
cur               435 drivers/mtd/nand/raw/s3c2410.c 	writel(cur, info->sel_reg);
cur              3444 drivers/net/bonding/bond_main.c 	int cur = 0, max = 0;
cur              3458 drivers/net/bonding/bond_main.c 			dev_stack[cur] = now;
cur              3459 drivers/net/bonding/bond_main.c 			iter_stack[cur++] = iter;
cur              3460 drivers/net/bonding/bond_main.c 			if (max <= cur)
cur              3461 drivers/net/bonding/bond_main.c 				max = cur;
cur              3466 drivers/net/bonding/bond_main.c 			if (!cur)
cur              3468 drivers/net/bonding/bond_main.c 			next = dev_stack[--cur];
cur              3469 drivers/net/bonding/bond_main.c 			niter = iter_stack[cur];
cur               417 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 		ring->cur = 0;
cur               461 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 		ring->cur = 0;
cur               535 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 	DBGPR("-->xgbe_map_tx_skb: cur = %d\n", ring->cur);
cur               538 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 	start_index = ring->cur;
cur               539 drivers/net/ethernet/amd/xgbe/xgbe-desc.c 	cur_index = ring->cur;
cur              1413 drivers/net/ethernet/amd/xgbe/xgbe-dev.c 	int start_index = ring->cur;
cur              1491 drivers/net/ethernet/amd/xgbe/xgbe-dev.c 	unsigned int start_index = ring->cur;
cur              1652 drivers/net/ethernet/amd/xgbe/xgbe-dev.c 	rdata = XGBE_GET_DESC_DATA(ring, ring->cur);
cur              1677 drivers/net/ethernet/amd/xgbe/xgbe-dev.c 	int start_index = ring->cur;
cur              1678 drivers/net/ethernet/amd/xgbe/xgbe-dev.c 	int cur_index = ring->cur;
cur              1889 drivers/net/ethernet/amd/xgbe/xgbe-dev.c 	ring->cur = cur_index + 1;
cur              1899 drivers/net/ethernet/amd/xgbe/xgbe-dev.c 	      (ring->cur - 1) & (ring->rdesc_count - 1));
cur              1914 drivers/net/ethernet/amd/xgbe/xgbe-dev.c 	DBGPR("-->xgbe_dev_read: cur = %d\n", ring->cur);
cur              1916 drivers/net/ethernet/amd/xgbe/xgbe-dev.c 	rdata = XGBE_GET_DESC_DATA(ring, ring->cur);
cur              1927 drivers/net/ethernet/amd/xgbe/xgbe-dev.c 		xgbe_dump_rx_desc(pdata, ring, ring->cur);
cur              2059 drivers/net/ethernet/amd/xgbe/xgbe-dev.c 	      ring->cur & (ring->rdesc_count - 1), ring->cur);
cur               265 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 	return (ring->rdesc_count - (ring->cur - ring->dirty));
cur               270 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 	return (ring->cur - ring->dirty);
cur              2524 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 	while (ring->dirty != ring->cur) {
cur              2624 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 	unsigned int cur;
cur              2632 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 	cur = ring->cur;
cur              2640 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 	       (ring->dirty != cur)) {
cur              2710 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 	rdata = XGBE_GET_DESC_DATA(ring, ring->cur);
cur              2713 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 		DBGPR("  cur = %d\n", ring->cur);
cur              2728 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 		rdata = XGBE_GET_DESC_DATA(ring, ring->cur);
cur              2737 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 		ring->cur++;
cur              2861 drivers/net/ethernet/amd/xgbe/xgbe-drv.c 		rdata = XGBE_GET_DESC_DATA(ring, ring->cur);
cur               476 drivers/net/ethernet/amd/xgbe/xgbe.h 	unsigned int cur;
cur               818 drivers/net/ethernet/apple/bmac.c 	unsigned int counter, cur = curval, next = nxtval;
cur               827 drivers/net/ethernet/apple/bmac.c 		if ((cur & 0x80000000) == 0) high_crc_set = 0;
cur               830 drivers/net/ethernet/apple/bmac.c 		cur = cur << 1;
cur               838 drivers/net/ethernet/apple/bmac.c 		if (high_crc_set ^ low_data_set) cur = cur ^ CRC32_POLY_BE;
cur               840 drivers/net/ethernet/apple/bmac.c 	return cur;
cur                78 drivers/net/ethernet/atheros/alx/main.c 	u16 cur, next, count = 0;
cur                80 drivers/net/ethernet/atheros/alx/main.c 	next = cur = rxq->write_idx;
cur                83 drivers/net/ethernet/atheros/alx/main.c 	cur_buf = &rxq->bufs[cur];
cur                86 drivers/net/ethernet/atheros/alx/main.c 		struct alx_rfd *rfd = &rxq->rfd[cur];
cur               125 drivers/net/ethernet/atheros/alx/main.c 		cur = next;
cur               128 drivers/net/ethernet/atheros/alx/main.c 		cur_buf = &rxq->bufs[cur];
cur               135 drivers/net/ethernet/atheros/alx/main.c 		rxq->write_idx = cur;
cur               136 drivers/net/ethernet/atheros/alx/main.c 		alx_write_mem16(&alx->hw, ALX_RFD_PIDX, cur);
cur               619 drivers/net/ethernet/broadcom/b44.c 	u32 cur, cons;
cur               622 drivers/net/ethernet/broadcom/b44.c 	cur  = br32(bp, B44_DMATX_STAT) & DMATX_STAT_CDMASK;
cur               623 drivers/net/ethernet/broadcom/b44.c 	cur /= sizeof(struct dma_desc);
cur               626 drivers/net/ethernet/broadcom/b44.c 	for (cons = bp->tx_cons; cons != cur; cons = NEXT_TX(cons)) {
cur              2507 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 	u32 cur;
cur              2512 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 	cur = SHMEM2_RD(bp, os_driver_state[BP_FW_MB_IDX(bp)]);
cur              2514 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c 	   cur, state);
cur              3017 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c 	u64 cur[BNX2X_MCAST_VEC_SZ], req[BNX2X_MCAST_VEC_SZ];
cur              3024 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c 	memcpy(cur, o->registry.aprox_match.vec,
cur              3049 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c 		bool b_current = !!BIT_VEC64_TEST_BIT(cur, i);
cur              1376 drivers/net/ethernet/fealnx.c 	struct fealnx_desc *cur;
cur              1386 drivers/net/ethernet/fealnx.c 		cur = &np->tx_ring[i];
cur              1387 drivers/net/ethernet/fealnx.c 		if (cur->skbuff) {
cur              1388 drivers/net/ethernet/fealnx.c 			pci_unmap_single(np->pci_dev, cur->buffer,
cur              1389 drivers/net/ethernet/fealnx.c 				cur->skbuff->len, PCI_DMA_TODEVICE);
cur              1390 drivers/net/ethernet/fealnx.c 			dev_kfree_skb_any(cur->skbuff);
cur              1391 drivers/net/ethernet/fealnx.c 			cur->skbuff = NULL;
cur              1393 drivers/net/ethernet/fealnx.c 		cur->status = 0;
cur              1394 drivers/net/ethernet/fealnx.c 		cur->control = 0;	/* needed? */
cur              1396 drivers/net/ethernet/fealnx.c 		cur->next_desc = np->tx_ring_dma +
cur              1398 drivers/net/ethernet/fealnx.c 		cur->next_desc_logical = &np->tx_ring[i + 1];
cur              1410 drivers/net/ethernet/fealnx.c 	struct fealnx_desc *cur = np->cur_rx;
cur              1416 drivers/net/ethernet/fealnx.c 		if (cur->skbuff)
cur              1417 drivers/net/ethernet/fealnx.c 			cur->status = RXOWN;
cur              1418 drivers/net/ethernet/fealnx.c 		cur = cur->next_desc_logical;
cur              1642 drivers/net/ethernet/fealnx.c 					struct fealnx_desc *cur;
cur              1645 drivers/net/ethernet/fealnx.c 					cur = np->cur_rx;
cur              1648 drivers/net/ethernet/fealnx.c 						if ((!(cur->status & RXOWN)) &&
cur              1649 drivers/net/ethernet/fealnx.c 						    (cur->status & RXLSD))
cur              1652 drivers/net/ethernet/fealnx.c 						cur = cur->next_desc_logical;
cur               465 drivers/net/ethernet/freescale/fec.h 	struct bufdesc	*cur;
cur               310 drivers/net/ethernet/freescale/fec_main.c 			(const char *)txq->bd.cur) >> txq->bd.dsize_log2) - 1;
cur               350 drivers/net/ethernet/freescale/fec_main.c 			bdp == txq->bd.cur ? 'S' : ' ',
cur               389 drivers/net/ethernet/freescale/fec_main.c 	struct bufdesc *bdp = txq->bd.cur;
cur               462 drivers/net/ethernet/freescale/fec_main.c 	bdp = txq->bd.cur;
cur               500 drivers/net/ethernet/freescale/fec_main.c 	bdp = txq->bd.cur;
cur               590 drivers/net/ethernet/freescale/fec_main.c 	txq->bd.cur = bdp;
cur               719 drivers/net/ethernet/freescale/fec_main.c 	struct bufdesc *bdp = txq->bd.cur;
cur               780 drivers/net/ethernet/freescale/fec_main.c 	txq->bd.cur = bdp;
cur               855 drivers/net/ethernet/freescale/fec_main.c 		rxq->bd.cur = rxq->bd.base;
cur               862 drivers/net/ethernet/freescale/fec_main.c 		txq->bd.cur = bdp;
cur              1265 drivers/net/ethernet/freescale/fec_main.c 	while (bdp != READ_ONCE(txq->bd.cur)) {
cur              1342 drivers/net/ethernet/freescale/fec_main.c 	if (bdp != txq->bd.cur &&
cur              1439 drivers/net/ethernet/freescale/fec_main.c 	bdp = rxq->bd.cur;
cur              1598 drivers/net/ethernet/freescale/fec_main.c 	rxq->bd.cur = bdp;
cur              3279 drivers/net/ethernet/freescale/fec_main.c 		rxq->bd.cur = cbd_base;
cur              3295 drivers/net/ethernet/freescale/fec_main.c 		txq->bd.cur = cbd_base;
cur              1332 drivers/net/ethernet/ibm/ibmvnic.c 	u8 *data, *cur;
cur              1336 drivers/net/ethernet/ibm/ibmvnic.c 		cur = hdr_data + len - tmp_len;
cur              1339 drivers/net/ethernet/ibm/ibmvnic.c 		if (cur != hdr_data) {
cur              1356 drivers/net/ethernet/ibm/ibmvnic.c 		memcpy(data, cur, tmp);
cur              1476 drivers/net/ethernet/ibm/ibmvnic.c 		int cur, i;
cur              1480 drivers/net/ethernet/ibm/ibmvnic.c 		cur = skb_headlen(skb);
cur              1486 drivers/net/ethernet/ibm/ibmvnic.c 			memcpy(dst + cur,
cur              1489 drivers/net/ethernet/ibm/ibmvnic.c 			cur += skb_frag_size(frag);
cur              2687 drivers/net/ethernet/ibm/ibmvnic.c 	scrq->cur = 0;
cur              3292 drivers/net/ethernet/ibm/ibmvnic.c 	union sub_crq *entry = &scrq->msgs[scrq->cur];
cur              3307 drivers/net/ethernet/ibm/ibmvnic.c 	entry = &scrq->msgs[scrq->cur];
cur              3309 drivers/net/ethernet/ibm/ibmvnic.c 		if (++scrq->cur == scrq->size)
cur              3310 drivers/net/ethernet/ibm/ibmvnic.c 			scrq->cur = 0;
cur              3324 drivers/net/ethernet/ibm/ibmvnic.c 	crq = &queue->msgs[queue->cur];
cur              3326 drivers/net/ethernet/ibm/ibmvnic.c 		if (++queue->cur == queue->size)
cur              3327 drivers/net/ethernet/ibm/ibmvnic.c 			queue->cur = 0;
cur              4699 drivers/net/ethernet/ibm/ibmvnic.c 	crq->cur = 0;
cur              4796 drivers/net/ethernet/ibm/ibmvnic.c 	crq->cur = 0;
cur               846 drivers/net/ethernet/ibm/ibmvnic.h 	int size, cur;
cur               866 drivers/net/ethernet/ibm/ibmvnic.h 	int size, cur;
cur               525 drivers/net/ethernet/intel/ice/ice_common.c 				hw->fw_log.evnts[m].cur = flgs;
cur               613 drivers/net/ethernet/intel/ice/ice_common.c 			if (hw->fw_log.evnts[i].cfg == hw->fw_log.evnts[i].cur)
cur               671 drivers/net/ethernet/intel/ice/ice_common.c 				hw->fw_log.evnts[i].cur = 0;
cur               677 drivers/net/ethernet/intel/ice/ice_common.c 			hw->fw_log.evnts[m].cur = hw->fw_log.evnts[m].cfg;
cur               385 drivers/net/ethernet/intel/ice/ice_type.h 	u8 cur : 4;	/* Current/active event enables */
cur               359 drivers/net/ethernet/jme.c 	dpi->cur		= PCC_P1;
cur              1120 drivers/net/ethernet/jme.c 	if (likely(atmp == dpi->cur)) {
cur              1147 drivers/net/ethernet/jme.c 	if (unlikely(dpi->attempt != dpi->cur && dpi->cnt > 5)) {
cur              1148 drivers/net/ethernet/jme.c 		if (dpi->attempt < dpi->cur)
cur              1151 drivers/net/ethernet/jme.c 		dpi->cur = dpi->attempt;
cur              2432 drivers/net/ethernet/jme.c 	switch (jme->dpi.cur) {
cur              2465 drivers/net/ethernet/jme.c 		dpi->cur		= PCC_P1;
cur              3048 drivers/net/ethernet/jme.c 	jme->dpi.cur = PCC_P1;
cur               146 drivers/net/ethernet/jme.h 	unsigned char	cur;
cur               904 drivers/net/ethernet/mellanox/mlx4/mcg.c 	struct mlx4_spec_list *cur;
cur               913 drivers/net/ethernet/mellanox/mlx4/mcg.c 	list_for_each_entry(cur, &rule->list, list) {
cur               914 drivers/net/ethernet/mellanox/mlx4/mcg.c 		switch (cur->id) {
cur               917 drivers/net/ethernet/mellanox/mlx4/mcg.c 					"dmac = %pM ", &cur->eth.dst_mac);
cur               918 drivers/net/ethernet/mellanox/mlx4/mcg.c 			if (cur->eth.ether_type)
cur               921 drivers/net/ethernet/mellanox/mlx4/mcg.c 						be16_to_cpu(cur->eth.ether_type));
cur               922 drivers/net/ethernet/mellanox/mlx4/mcg.c 			if (cur->eth.vlan_id)
cur               925 drivers/net/ethernet/mellanox/mlx4/mcg.c 						be16_to_cpu(cur->eth.vlan_id));
cur               929 drivers/net/ethernet/mellanox/mlx4/mcg.c 			if (cur->ipv4.src_ip)
cur               932 drivers/net/ethernet/mellanox/mlx4/mcg.c 						&cur->ipv4.src_ip);
cur               933 drivers/net/ethernet/mellanox/mlx4/mcg.c 			if (cur->ipv4.dst_ip)
cur               936 drivers/net/ethernet/mellanox/mlx4/mcg.c 						&cur->ipv4.dst_ip);
cur               941 drivers/net/ethernet/mellanox/mlx4/mcg.c 			if (cur->tcp_udp.src_port)
cur               944 drivers/net/ethernet/mellanox/mlx4/mcg.c 						be16_to_cpu(cur->tcp_udp.src_port));
cur               945 drivers/net/ethernet/mellanox/mlx4/mcg.c 			if (cur->tcp_udp.dst_port)
cur               948 drivers/net/ethernet/mellanox/mlx4/mcg.c 						be16_to_cpu(cur->tcp_udp.dst_port));
cur               953 drivers/net/ethernet/mellanox/mlx4/mcg.c 					"dst-gid = %pI6\n", cur->ib.dst_gid);
cur               956 drivers/net/ethernet/mellanox/mlx4/mcg.c 					cur->ib.dst_gid_msk);
cur               961 drivers/net/ethernet/mellanox/mlx4/mcg.c 					"VNID = %d ", be32_to_cpu(cur->vxlan.vni));
cur               981 drivers/net/ethernet/mellanox/mlx4/mcg.c 	struct mlx4_spec_list *cur;
cur               999 drivers/net/ethernet/mellanox/mlx4/mcg.c 	list_for_each_entry(cur, &rule->list, list) {
cur              1000 drivers/net/ethernet/mellanox/mlx4/mcg.c 		ret = parse_trans_rule(dev, cur, mailbox->buf + size);
cur                24 drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c 	u32 cur;
cur               269 drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c 	mlxsw_reg_sbpm_unpack(sbpm_pl, &pm->occ.cur, &pm->occ.max);
cur              1217 drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c 						  &cm->occ.cur, &cm->occ.max);
cur              1231 drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c 						  &cm->occ.cur, &cm->occ.max);
cur              1379 drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c 	*p_cur = mlxsw_sp_cells_bytes(mlxsw_sp, pm->occ.cur);
cur              1398 drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c 	*p_cur = mlxsw_sp_cells_bytes(mlxsw_sp, cm->occ.cur);
cur              1001 drivers/net/ethernet/micrel/ksz884x.c 	struct ksz_desc *cur;
cur              1829 drivers/net/ethernet/micrel/ksz884x.c 	u32 cur;
cur              1849 drivers/net/ethernet/micrel/ksz884x.c 		cur = *last;
cur              1850 drivers/net/ethernet/micrel/ksz884x.c 		if (data != cur) {
cur              1852 drivers/net/ethernet/micrel/ksz884x.c 			if (data < cur)
cur              1854 drivers/net/ethernet/micrel/ksz884x.c 			data -= cur;
cur              3831 drivers/net/ethernet/micrel/ksz884x.c 	struct ksz_desc *cur = desc_info->ring;
cur              3835 drivers/net/ethernet/micrel/ksz884x.c 		cur->phw = desc++;
cur              3837 drivers/net/ethernet/micrel/ksz884x.c 		previous = cur++;
cur              3847 drivers/net/ethernet/micrel/ksz884x.c 	desc_info->cur = desc_info->ring;
cur              3867 drivers/net/ethernet/micrel/ksz884x.c 	info->cur = info->ring;
cur              3977 drivers/net/ethernet/micrel/ksz884x.c 	get_tx_pkt(&hw->tx_desc_info, &hw->tx_desc_info.cur);
cur              3978 drivers/net/ethernet/micrel/ksz884x.c 	hw->tx_desc_info.cur->sw.buf.tx.first_seg = 1;
cur              4002 drivers/net/ethernet/micrel/ksz884x.c 	struct ksz_desc *cur = hw->tx_desc_info.cur;
cur              4004 drivers/net/ethernet/micrel/ksz884x.c 	cur->sw.buf.tx.last_seg = 1;
cur              4008 drivers/net/ethernet/micrel/ksz884x.c 		cur->sw.buf.tx.intr = 1;
cur              4014 drivers/net/ethernet/micrel/ksz884x.c 	cur->sw.buf.tx.dest_port = hw->dst_ports;
cur              4016 drivers/net/ethernet/micrel/ksz884x.c 	release_desc(cur);
cur              4650 drivers/net/ethernet/micrel/ksz884x.c 	first = info->cur;
cur              4696 drivers/net/ethernet/micrel/ksz884x.c 		info->cur = desc;
cur               640 drivers/net/ethernet/natsemi/natsemi.c static int netdev_get_wol(struct net_device *dev, u32 *supported, u32 *cur);
cur              2736 drivers/net/ethernet/natsemi/natsemi.c static int netdev_get_wol(struct net_device *dev, u32 *supported, u32 *cur)
cur              2749 drivers/net/ethernet/natsemi/natsemi.c 	*cur = 0;
cur              2753 drivers/net/ethernet/natsemi/natsemi.c 		*cur |= WAKE_PHY;
cur              2755 drivers/net/ethernet/natsemi/natsemi.c 		*cur |= WAKE_UCAST;
cur              2757 drivers/net/ethernet/natsemi/natsemi.c 		*cur |= WAKE_MCAST;
cur              2759 drivers/net/ethernet/natsemi/natsemi.c 		*cur |= WAKE_BCAST;
cur              2761 drivers/net/ethernet/natsemi/natsemi.c 		*cur |= WAKE_ARP;
cur              2763 drivers/net/ethernet/natsemi/natsemi.c 		*cur |= WAKE_MAGIC;
cur              2766 drivers/net/ethernet/natsemi/natsemi.c 		*cur |= WAKE_MAGICSECURE;
cur               297 drivers/net/ethernet/pensando/ionic/ionic_dev.c 	struct ionic_cq_info *cur;
cur               315 drivers/net/ethernet/pensando/ionic/ionic_dev.c 	cur = cq->info;
cur               319 drivers/net/ethernet/pensando/ionic/ionic_dev.c 			cur->next = cq->info;
cur               320 drivers/net/ethernet/pensando/ionic/ionic_dev.c 			cur->last = true;
cur               322 drivers/net/ethernet/pensando/ionic/ionic_dev.c 			cur->next = cur + 1;
cur               324 drivers/net/ethernet/pensando/ionic/ionic_dev.c 		cur->index = i;
cur               325 drivers/net/ethernet/pensando/ionic/ionic_dev.c 		cur++;
cur               333 drivers/net/ethernet/pensando/ionic/ionic_dev.c 	struct ionic_cq_info *cur;
cur               339 drivers/net/ethernet/pensando/ionic/ionic_dev.c 	for (i = 0, cur = cq->info; i < cq->num_descs; i++, cur++)
cur               340 drivers/net/ethernet/pensando/ionic/ionic_dev.c 		cur->cq_desc = base + (i * cq->desc_size);
cur               378 drivers/net/ethernet/pensando/ionic/ionic_dev.c 	struct ionic_desc_info *cur;
cur               401 drivers/net/ethernet/pensando/ionic/ionic_dev.c 	cur = q->info;
cur               405 drivers/net/ethernet/pensando/ionic/ionic_dev.c 			cur->next = q->info;
cur               407 drivers/net/ethernet/pensando/ionic/ionic_dev.c 			cur->next = cur + 1;
cur               408 drivers/net/ethernet/pensando/ionic/ionic_dev.c 		cur->index = i;
cur               409 drivers/net/ethernet/pensando/ionic/ionic_dev.c 		cur->left = num_descs - i;
cur               410 drivers/net/ethernet/pensando/ionic/ionic_dev.c 		cur++;
cur               418 drivers/net/ethernet/pensando/ionic/ionic_dev.c 	struct ionic_desc_info *cur;
cur               424 drivers/net/ethernet/pensando/ionic/ionic_dev.c 	for (i = 0, cur = q->info; i < q->num_descs; i++, cur++)
cur               425 drivers/net/ethernet/pensando/ionic/ionic_dev.c 		cur->desc = base + (i * q->desc_size);
cur               430 drivers/net/ethernet/pensando/ionic/ionic_dev.c 	struct ionic_desc_info *cur;
cur               436 drivers/net/ethernet/pensando/ionic/ionic_dev.c 	for (i = 0, cur = q->info; i < q->num_descs; i++, cur++)
cur               437 drivers/net/ethernet/pensando/ionic/ionic_dev.c 		cur->sg_desc = base + (i * q->sg_desc_size);
cur              1791 drivers/net/ethernet/pensando/ionic/ionic_lif.c 	struct list_head *cur, *tmp;
cur              1794 drivers/net/ethernet/pensando/ionic/ionic_lif.c 	list_for_each_safe(cur, tmp, &ionic->lifs) {
cur              1795 drivers/net/ethernet/pensando/ionic/ionic_lif.c 		lif = list_entry(cur, struct ionic_lif, list);
cur              1820 drivers/net/ethernet/pensando/ionic/ionic_lif.c 	struct list_head *cur, *tmp;
cur              1823 drivers/net/ethernet/pensando/ionic/ionic_lif.c 	list_for_each_safe(cur, tmp, &ionic->lifs) {
cur              1824 drivers/net/ethernet/pensando/ionic/ionic_lif.c 		lif = list_entry(cur, struct ionic_lif, list);
cur              2060 drivers/net/ethernet/pensando/ionic/ionic_lif.c 	struct list_head *cur, *tmp;
cur              2064 drivers/net/ethernet/pensando/ionic/ionic_lif.c 	list_for_each_safe(cur, tmp, &ionic->lifs) {
cur              2065 drivers/net/ethernet/pensando/ionic/ionic_lif.c 		lif = list_entry(cur, struct ionic_lif, list);
cur               287 drivers/net/ethernet/pensando/ionic/ionic_txrx.c 	struct ionic_desc_info *cur;
cur               290 drivers/net/ethernet/pensando/ionic/ionic_txrx.c 	for (cur = q->tail; cur != q->head; cur = cur->next) {
cur               291 drivers/net/ethernet/pensando/ionic/ionic_txrx.c 		desc = cur->desc;
cur               294 drivers/net/ethernet/pensando/ionic/ionic_txrx.c 		dev_kfree_skb(cur->cb_arg);
cur               295 drivers/net/ethernet/pensando/ionic/ionic_txrx.c 		cur->cb_arg = NULL;
cur               625 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 	nx_mac_list_t *cur;
cur               629 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 		cur = list_entry(head, nx_mac_list_t, list);
cur               631 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 		if (ether_addr_equal(addr, cur->mac_addr)) {
cur               637 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 	cur = kzalloc(sizeof(nx_mac_list_t), GFP_ATOMIC);
cur               638 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 	if (cur == NULL)
cur               641 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 	memcpy(cur->mac_addr, addr, ETH_ALEN);
cur               642 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 	list_add_tail(&cur->list, &adapter->mac_list);
cur               644 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 				cur->mac_addr, NETXEN_MAC_ADD);
cur               657 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 	nx_mac_list_t *cur;
cur               687 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 		cur = list_entry(head->next, nx_mac_list_t, list);
cur               690 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 				cur->mac_addr, NETXEN_MAC_DEL);
cur               691 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 		list_del(&cur->list);
cur               692 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 		kfree(cur);
cur               717 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 	nx_mac_list_t *cur;
cur               721 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 		cur = list_entry(head->next, nx_mac_list_t, list);
cur               723 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 				cur->mac_addr, NETXEN_MAC_DEL);
cur               724 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 		list_del(&cur->list);
cur               725 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c 		kfree(cur);
cur              1651 drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c 	struct list_head *cur;
cur              1713 drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c 			list_for_each(cur, &sds_ring->free_list[ring]) {
cur              1714 drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c 				rxbuf = list_entry(cur,
cur              3177 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c 	struct nx_ip_list  *cur, *tmp_cur;
cur              3179 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c 	list_for_each_entry_safe(cur, tmp_cur, &adapter->ip_list, list) {
cur              3181 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c 			if (cur->master) {
cur              3182 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c 				netxen_config_ipaddr(adapter, cur->ip_addr,
cur              3184 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c 				list_del(&cur->list);
cur              3185 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c 				kfree(cur);
cur              3188 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c 			netxen_config_ipaddr(adapter, cur->ip_addr, NX_IP_DOWN);
cur              3189 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c 			list_del(&cur->list);
cur              3190 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c 			kfree(cur);
cur              3200 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c 	struct nx_ip_list *cur, *tmp_cur;
cur              3212 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c 			cur = list_entry(head, struct nx_ip_list, list);
cur              3214 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c 			if (cur->ip_addr == ifa->ifa_address)
cur              3218 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c 		cur = kzalloc(sizeof(struct nx_ip_list), GFP_ATOMIC);
cur              3219 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c 		if (cur == NULL)
cur              3223 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c 		cur->master = !!netif_is_bond_master(dev);
cur              3224 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c 		cur->ip_addr = ifa->ifa_address;
cur              3225 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c 		list_add_tail(&cur->list, &adapter->ip_list);
cur              3230 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c 		list_for_each_entry_safe(cur, tmp_cur,
cur              3232 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c 			if (cur->ip_addr == ifa->ifa_address) {
cur              3233 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c 				list_del(&cur->list);
cur              3234 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c 				kfree(cur);
cur               146 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	u32 cur, prev;
cur               147 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	cur = adapter->ahw->idc.curr_state;
cur               152 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 		 adapter->ahw->idc.name[cur],
cur               984 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	u32 cur, prev, next;
cur               986 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	cur = adapter->ahw->idc.curr_state;
cur               994 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 			__func__, cur, prev, state);
cur               998 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 	if ((cur == QLC_83XX_IDC_DEV_UNKNOWN) &&
cur              1004 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 				__func__, cur, prev, next);
cur              1015 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c 				__func__, cur, prev, next);
cur               463 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	struct qlcnic_mac_vlan_list *cur;
cur               469 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		cur = list_entry(head, struct qlcnic_mac_vlan_list, list);
cur               470 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		if (ether_addr_equal(addr, cur->mac_addr)) {
cur               471 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 			err = qlcnic_sre_macaddr_change(adapter, cur->mac_addr,
cur               475 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 			list_del(&cur->list);
cur               476 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 			kfree(cur);
cur               486 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	struct qlcnic_mac_vlan_list *cur;
cur               491 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		cur = list_entry(head, struct qlcnic_mac_vlan_list, list);
cur               492 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		if (ether_addr_equal(addr, cur->mac_addr) &&
cur               493 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		    cur->vlan_id == vlan)
cur               497 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	cur = kzalloc(sizeof(*cur), GFP_ATOMIC);
cur               498 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	if (cur == NULL)
cur               501 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	memcpy(cur->mac_addr, addr, ETH_ALEN);
cur               504 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 				cur->mac_addr, vlan, QLCNIC_MAC_ADD)) {
cur               505 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		kfree(cur);
cur               509 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	cur->vlan_id = vlan;
cur               510 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	cur->mac_type = mac_type;
cur               512 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	list_add_tail(&cur->list, &adapter->mac_list);
cur               518 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	struct qlcnic_mac_vlan_list *cur;
cur               522 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		cur = list_entry(head, struct qlcnic_mac_vlan_list, list);
cur               523 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		if (cur->mac_type != QLCNIC_MULTICAST_MAC)
cur               526 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		qlcnic_sre_macaddr_change(adapter, cur->mac_addr,
cur               527 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 					  cur->vlan_id, QLCNIC_MAC_DEL);
cur               528 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		list_del(&cur->list);
cur               529 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		kfree(cur);
cur               623 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 	struct qlcnic_mac_vlan_list *cur;
cur               626 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		cur = list_entry(head->next, struct qlcnic_mac_vlan_list, list);
cur               628 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 				cur->mac_addr, 0, QLCNIC_MAC_DEL);
cur               629 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		list_del(&cur->list);
cur               630 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c 		kfree(cur);
cur              1361 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c 	struct list_head *cur;
cur              1415 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c 			list_for_each(cur, &sds_ring->free_list[ring]) {
cur              1416 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c 				rxbuf = list_entry(cur, struct qlcnic_rx_buffer,
cur              1885 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c 	struct list_head *cur;
cur              1931 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c 			list_for_each(cur, &sds_ring->free_list[ring]) {
cur              1932 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c 				rxbuf = list_entry(cur, struct qlcnic_rx_buffer,
cur               322 drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c 	struct qlcnic_mac_vlan_list *cur;
cur               326 drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c 		cur = list_entry(head, struct qlcnic_mac_vlan_list, list);
cur               327 drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c 		if (ether_addr_equal_unaligned(adapter->mac_addr, cur->mac_addr)) {
cur               328 drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c 			qlcnic_sre_macaddr_change(adapter, cur->mac_addr,
cur               330 drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c 			list_del(&cur->list);
cur               331 drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c 			kfree(cur);
cur              2100 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 	struct qlcnic_mac_vlan_list *cur;
cur              2103 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		cur = list_entry(head->next, struct qlcnic_mac_vlan_list, list);
cur              2104 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		qlcnic_sre_macaddr_change(adapter, cur->mac_addr, cur->vlan_id,
cur              2106 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		list_del(&cur->list);
cur              2107 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c 		kfree(cur);
cur               515 drivers/net/ethernet/sis/sis190.c 	u32 cur;
cur               517 drivers/net/ethernet/sis/sis190.c 	for (cur = start; cur < end; cur++) {
cur               518 drivers/net/ethernet/sis/sis190.c 		unsigned int i = cur % NUM_RX_DESC;
cur               528 drivers/net/ethernet/sis/sis190.c 	return cur - start;
cur              1245 drivers/net/ethernet/sis/sis190.c 	struct sis190_phy *cur, *next;
cur              1247 drivers/net/ethernet/sis/sis190.c 	list_for_each_entry_safe(cur, next, first_phy, list) {
cur              1248 drivers/net/ethernet/sis/sis190.c 		kfree(cur);
cur               120 drivers/net/ethernet/smsc/smc911x.c 		int cur;
cur               122 drivers/net/ethernet/smsc/smc911x.c 		for (cur = 0; cur < 8; cur++) {
cur               169 drivers/net/ethernet/smsc/smc91x.c 		int cur;
cur               171 drivers/net/ethernet/smsc/smc91x.c 		for (cur = 0; cur < 8; cur++) {
cur              8567 drivers/net/ethernet/sun/niu.c 	if (p->cur[type] >= NIU_MAX_PORTS) {
cur              8571 drivers/net/ethernet/sun/niu.c 	idx = p->cur[type];
cur              8574 drivers/net/ethernet/sun/niu.c 	p->cur[type] = idx + 1;
cur              8582 drivers/net/ethernet/sun/niu.c 	for (i = 0; i < p->cur[PHY_TYPE_PMA_PMD]; i++) {
cur              8586 drivers/net/ethernet/sun/niu.c 	for (i = 0; i < p->cur[PHY_TYPE_PCS]; i++) {
cur              8614 drivers/net/ethernet/sun/niu.c 	if (p->cur[PHY_TYPE_MII])
cur              8617 drivers/net/ethernet/sun/niu.c 	return p->cur[PHY_TYPE_MII];
cur              3041 drivers/net/ethernet/sun/niu.h 	u8	cur[PHY_TYPE_MAX];
cur               450 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 		ring->cur = 0;
cur               490 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 		ring->cur = 0;
cur               512 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 	start_index = ring->cur;
cur               513 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c 	cur_index = ring->cur;
cur               688 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c 	desc_data = XLGMAC_GET_DESC_DATA(ring, ring->cur);
cur               711 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c 	int start_index = ring->cur;
cur               712 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c 	int cur_index = ring->cur;
cur               997 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c 	ring->cur = cur_index + 1;
cur              1007 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c 		  (ring->cur - 1) & (ring->dma_desc_count - 1));
cur              1060 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c 	int start_index = ring->cur;
cur              1143 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c 	unsigned int start_index = ring->cur;
cur              2648 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c 	desc_data = XLGMAC_GET_DESC_DATA(ring, ring->cur);
cur              2662 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c 		xlgmac_dump_rx_desc(pdata, ring, ring->cur);
cur              2811 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c 		  ring->cur & (ring->dma_desc_count - 1), ring->cur);
cur                30 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 	return (ring->dma_desc_count - (ring->cur - ring->dirty));
cur                35 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 	return (ring->cur - ring->dirty);
cur               962 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 	while (ring->dirty != ring->cur) {
cur              1046 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 	unsigned int cur;
cur              1055 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 	cur = ring->cur;
cur              1063 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 	       (ring->dirty != cur)) {
cur              1134 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 	desc_data = XLGMAC_GET_DESC_DATA(ring, ring->cur);
cur              1150 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 		desc_data = XLGMAC_GET_DESC_DATA(ring, ring->cur);
cur              1159 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 		ring->cur++;
cur              1268 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c 		desc_data = XLGMAC_GET_DESC_DATA(ring, ring->cur);
cur               339 drivers/net/ethernet/synopsys/dwc-xlgmac.h 	unsigned int cur;
cur                95 drivers/net/ethernet/xilinx/ll_temac_main.c 	ktime_t cur = ktime_get();
cur                97 drivers/net/ethernet/xilinx/ll_temac_main.c 	return hard_acs_rdy(lp) || ktime_after(cur, timeout);
cur                87 drivers/net/phy/mdio-mux.c 		  int (*switch_fn)(int cur, int desired, void *data),
cur               822 drivers/net/team/team.c 						      struct team_port *cur)
cur               824 drivers/net/team/team.c 	if (port->priority < cur->priority)
cur               826 drivers/net/team/team.c 	if (port->priority > cur->priority)
cur               828 drivers/net/team/team.c 	if (port->index < cur->index)
cur               836 drivers/net/team/team.c 	struct team_port *cur;
cur               844 drivers/net/team/team.c 	list_for_each_entry(cur, qom_list, qom_list) {
cur               845 drivers/net/team/team.c 		if (team_queue_override_port_has_gt_prio_than(port, cur))
cur               847 drivers/net/team/team.c 		node = &cur->qom_list;
cur               916 drivers/net/team/team.c 	struct team_port *cur;
cur               918 drivers/net/team/team.c 	list_for_each_entry(cur, &team->port_list, list)
cur               919 drivers/net/team/team.c 		if (cur == port)
cur                17 drivers/net/team/team_mode_broadcast.c 	struct team_port *cur;
cur                23 drivers/net/team/team_mode_broadcast.c 	list_for_each_entry_rcu(cur, &team->port_list, list) {
cur                24 drivers/net/team/team_mode_broadcast.c 		if (team_port_txable(cur)) {
cur                34 drivers/net/team/team_mode_broadcast.c 			last = cur;
cur              1113 drivers/net/wireless/ath/ath6kl/txrx.c 	u16 idx, st, cur, end;
cur              1136 drivers/net/wireless/ath/ath6kl/txrx.c 	cur = seq_no;
cur              1139 drivers/net/wireless/ath/ath6kl/txrx.c 	if (((st < end) && (cur < st || cur > end)) ||
cur              1140 drivers/net/wireless/ath/ath6kl/txrx.c 	    ((st > end) && (cur > end) && (cur < st))) {
cur              1145 drivers/net/wireless/ath/ath6kl/txrx.c 		     (cur < end || cur > extended_end)) ||
cur              1146 drivers/net/wireless/ath/ath6kl/txrx.c 		    ((end > extended_end) && (cur > extended_end) &&
cur              1147 drivers/net/wireless/ath/ath6kl/txrx.c 		     (cur < end))) {
cur              1150 drivers/net/wireless/ath/ath6kl/txrx.c 			if (cur >= rxtid->hold_q_sz - 1)
cur              1151 drivers/net/wireless/ath/ath6kl/txrx.c 				rxtid->seq_next = cur - (rxtid->hold_q_sz - 1);
cur              1154 drivers/net/wireless/ath/ath6kl/txrx.c 						  (rxtid->hold_q_sz - 2 - cur);
cur              1161 drivers/net/wireless/ath/ath6kl/txrx.c 			if (cur >= rxtid->hold_q_sz - 1)
cur              1162 drivers/net/wireless/ath/ath6kl/txrx.c 				st = cur - (rxtid->hold_q_sz - 1);
cur              1165 drivers/net/wireless/ath/ath6kl/txrx.c 					(rxtid->hold_q_sz - 2 - cur);
cur               336 drivers/net/wireless/ath/ath9k/channel.c 	struct ath_chanctx *prev, *cur;
cur               343 drivers/net/wireless/ath/ath9k/channel.c 	cur = sc->cur_chan;
cur               344 drivers/net/wireless/ath/ath9k/channel.c 	prev = ath_chanctx_get_next(sc, cur);
cur               350 drivers/net/wireless/ath/ath9k/channel.c 	cur_tsf = (u32) cur->tsf_val +
cur               351 drivers/net/wireless/ath/ath9k/channel.c 		  ath9k_hw_get_tsf_offset(&cur->tsf_ts, &ts);
cur              1852 drivers/net/wireless/ath/ath9k/hw.c u32 ath9k_hw_get_tsf_offset(struct timespec64 *last, struct timespec64 *cur)
cur              1857 drivers/net/wireless/ath/ath9k/hw.c 	if (!cur) {
cur              1859 drivers/net/wireless/ath/ath9k/hw.c 		cur = &ts;
cur              1862 drivers/net/wireless/ath/ath9k/hw.c 	usec = cur->tv_sec * 1000000ULL + cur->tv_nsec / 1000;
cur              1064 drivers/net/wireless/ath/ath9k/hw.h u32 ath9k_hw_get_tsf_offset(struct timespec64 *last, struct timespec64 *cur);
cur               241 drivers/net/wireless/ath/wcn36xx/dxe.c 	struct wcn36xx_dxe_ctl *cur = ch->head_blk_ctl;
cur               247 drivers/net/wireless/ath/wcn36xx/dxe.c 			cur->bd_phy_addr = bd_phy_addr;
cur               248 drivers/net/wireless/ath/wcn36xx/dxe.c 			cur->bd_cpu_addr = bd_cpu_addr;
cur               252 drivers/net/wireless/ath/wcn36xx/dxe.c 			cur->bd_phy_addr = 0;
cur               253 drivers/net/wireless/ath/wcn36xx/dxe.c 			cur->bd_cpu_addr = NULL;
cur               255 drivers/net/wireless/ath/wcn36xx/dxe.c 		cur = cur->next;
cur               319 drivers/net/wireless/ath/wcn36xx/dxe.c 	struct wcn36xx_dxe_ctl *cur = wcn_ch->head_blk_ctl;
cur               323 drivers/net/wireless/ath/wcn36xx/dxe.c 		kfree_skb(cur->skb);
cur               324 drivers/net/wireless/ath/wcn36xx/dxe.c 		cur = cur->next;
cur               773 drivers/net/wireless/broadcom/b43/debugfs.c 	struct b43_txstatus *cur;
cur               783 drivers/net/wireless/broadcom/b43/debugfs.c 	cur = &(log->log[i]);
cur               784 drivers/net/wireless/broadcom/b43/debugfs.c 	memcpy(cur, status, sizeof(*cur));
cur               445 drivers/net/wireless/broadcom/b43legacy/debugfs.c 	struct b43legacy_txstatus *cur;
cur               457 drivers/net/wireless/broadcom/b43legacy/debugfs.c 	cur = &(log->log[i]);
cur               458 drivers/net/wireless/broadcom/b43legacy/debugfs.c 	memcpy(cur, status, sizeof(*cur));
cur               524 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c 	struct brcmf_fw_item *cur;
cur               534 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c 	cur = &fwctx->req->items[fwctx->curpos];
cur               544 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c 		else if (!(cur->flags & BRCMF_FW_REQF_OPTIONAL))
cur               559 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c 	if (!nvram && !(cur->flags & BRCMF_FW_REQF_OPTIONAL))
cur               563 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c 	cur->nv_data.data = nvram;
cur               564 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c 	cur->nv_data.len = nvram_length;
cur               574 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c 	struct brcmf_fw_item *cur = &fwctx->req->items[fwctx->curpos];
cur               577 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c 	brcmf_dbg(TRACE, "firmware %s %sfound\n", cur->path, fw ? "" : "not ");
cur               579 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c 	switch (cur->type) {
cur               585 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c 			cur->binary = fw;
cur               591 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c 		brcmf_err("unknown fw type: %d\n", cur->type);
cur               596 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c 	return (cur->flags & BRCMF_FW_REQF_OPTIONAL) ? 0 : ret;
cur               602 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c 	struct brcmf_fw_item *cur = &fwctx->req->items[fwctx->curpos];
cur               606 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c 	if (cur->type == BRCMF_FW_TYPE_NVRAM && fwctx->req->board_type) {
cur               609 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c 		strlcpy(alt_path, cur->path, BRCMF_FW_NAME_LEN);
cur               621 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c 	return request_firmware(fw, cur->path, fwctx->dev);
cur              1257 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 	struct sk_buff *cur, *next;
cur              1259 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 	skb_queue_walk_safe(&bus->glom, cur, next) {
cur              1260 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 		skb_unlink(cur, &bus->glom);
cur              1261 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c 		brcmu_pkt_buf_free_skb(cur);
cur               963 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 			uint cur;
cur               964 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 			cur =
cur               971 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c 				      di->rxin, di->rxout, cur);
cur              1303 drivers/net/wireless/cisco/airo.c static void age_mic_context(miccntx *cur, miccntx *old, u8 *key, int key_len,
cur              1309 drivers/net/wireless/cisco/airo.c 	if (cur->valid && (memcmp(cur->key, key, key_len) == 0))
cur              1313 drivers/net/wireless/cisco/airo.c 	memcpy(old, cur, sizeof(*cur));
cur              1316 drivers/net/wireless/cisco/airo.c 	memcpy(cur->key, key, key_len);
cur              1317 drivers/net/wireless/cisco/airo.c 	cur->window  = 33; /* Window always points to the middle */
cur              1318 drivers/net/wireless/cisco/airo.c 	cur->rx      = 0;  /* Rx Sequence numbers */
cur              1319 drivers/net/wireless/cisco/airo.c 	cur->tx      = 0;  /* Tx sequence numbers */
cur              1320 drivers/net/wireless/cisco/airo.c 	cur->valid   = 1;  /* Key is now valid */
cur              1323 drivers/net/wireless/cisco/airo.c 	emmh32_setseed(&cur->seed, key, key_len, tfm);
cur               287 drivers/net/wireless/intel/iwlwifi/dvm/rx.c static void accum_stats(__le32 *prev, __le32 *cur, __le32 *delta,
cur               294 drivers/net/wireless/intel/iwlwifi/dvm/rx.c 	     i++, prev++, cur++, delta++, max_delta++, accum++) {
cur               295 drivers/net/wireless/intel/iwlwifi/dvm/rx.c 		if (le32_to_cpu(*cur) > le32_to_cpu(*prev)) {
cur               297 drivers/net/wireless/intel/iwlwifi/dvm/rx.c 				le32_to_cpu(*cur) - le32_to_cpu(*prev));
cur              2010 drivers/net/wireless/intel/iwlwifi/mvm/rs.c 	u8 cur = lq_sta->lq.reduced_tpc;
cur              2020 drivers/net/wireless/intel/iwlwifi/mvm/rs.c 		return cur != lq_sta->pers.dbg_fixed_txp_reduction;
cur              2036 drivers/net/wireless/intel/iwlwifi/mvm/rs.c 		return cur != TPC_NO_REDUCTION;
cur              2039 drivers/net/wireless/intel/iwlwifi/mvm/rs.c 	rs_get_adjacent_txp(mvm, cur, &weak, &strong);
cur              2043 drivers/net/wireless/intel/iwlwifi/mvm/rs.c 	sr = window[cur].success_ratio;
cur              2044 drivers/net/wireless/intel/iwlwifi/mvm/rs.c 	current_tpt = window[cur].average_tpt;
cur              2052 drivers/net/wireless/intel/iwlwifi/mvm/rs.c 		       cur, current_tpt, sr, weak, strong,
cur               912 drivers/net/wireless/marvell/libertas/if_sdio.c 	struct if_sdio_packet *packet, *cur;
cur               955 drivers/net/wireless/marvell/libertas/if_sdio.c 		cur = card->packets;
cur               956 drivers/net/wireless/marvell/libertas/if_sdio.c 		while (cur->next)
cur               957 drivers/net/wireless/marvell/libertas/if_sdio.c 			cur = cur->next;
cur               958 drivers/net/wireless/marvell/libertas/if_sdio.c 		cur->next = packet;
cur               530 drivers/net/wireless/mediatek/mt76/dma.c 	int qid, done = 0, cur;
cur               538 drivers/net/wireless/mediatek/mt76/dma.c 		cur = mt76_dma_rx_process(dev, &dev->q_rx[qid], budget - done);
cur               540 drivers/net/wireless/mediatek/mt76/dma.c 		done += cur;
cur               541 drivers/net/wireless/mediatek/mt76/dma.c 	} while (cur && done < budget);
cur                21 drivers/net/wireless/mediatek/mt76/mt76x2/pci_mcu.c 	__le32 *cur;
cur                58 drivers/net/wireless/mediatek/mt76/mt76x2/pci_mcu.c 	cur = (__le32 *)(fw->data + sizeof(*hdr));
cur                60 drivers/net/wireless/mediatek/mt76/mt76x2/pci_mcu.c 	mt76_wr_copy(dev, MT_MCU_ROM_PATCH_ADDR, cur, len);
cur                86 drivers/net/wireless/mediatek/mt76/mt76x2/pci_mcu.c 	__le32 *cur;
cur               113 drivers/net/wireless/mediatek/mt76/mt76x2/pci_mcu.c 	cur = (__le32 *)(fw->data + sizeof(*hdr));
cur               117 drivers/net/wireless/mediatek/mt76/mt76x2/pci_mcu.c 	mt76_wr_copy(dev, MT_MCU_ILM_ADDR, cur, len);
cur               119 drivers/net/wireless/mediatek/mt76/mt76x2/pci_mcu.c 	cur += len / sizeof(*cur);
cur               128 drivers/net/wireless/mediatek/mt76/mt76x2/pci_mcu.c 	mt76_wr_copy(dev, offset, cur, len);
cur                12 drivers/net/wireless/mediatek/mt76/util.c 	u32 cur;
cur                16 drivers/net/wireless/mediatek/mt76/util.c 		cur = dev->bus->rr(dev, offset) & mask;
cur                17 drivers/net/wireless/mediatek/mt76/util.c 		if (cur == val)
cur                30 drivers/net/wireless/mediatek/mt76/util.c 	u32 cur;
cur                34 drivers/net/wireless/mediatek/mt76/util.c 		cur = dev->bus->rr(dev, offset) & mask;
cur                35 drivers/net/wireless/mediatek/mt76/util.c 		if (cur == val)
cur                47 drivers/net/wireless/mediatek/mt76/util.c 	int i, idx = 0, cur;
cur                55 drivers/net/wireless/mediatek/mt76/util.c 		cur = i * BITS_PER_LONG + idx;
cur                56 drivers/net/wireless/mediatek/mt76/util.c 		if (cur >= size)
cur                60 drivers/net/wireless/mediatek/mt76/util.c 		return cur;
cur                31 drivers/net/wireless/mediatek/mt7601u/core.c 	u32 cur;
cur                38 drivers/net/wireless/mediatek/mt7601u/core.c 		cur = mt7601u_rr(dev, offset) & mask;
cur                39 drivers/net/wireless/mediatek/mt7601u/core.c 		if (cur == val)
cur                53 drivers/net/wireless/mediatek/mt7601u/core.c 	u32 cur;
cur                60 drivers/net/wireless/mediatek/mt7601u/core.c 		cur = mt7601u_rr(dev, offset) & mask;
cur                61 drivers/net/wireless/mediatek/mt7601u/core.c 		if (cur == val)
cur               200 drivers/net/wireless/realtek/rtw88/tx.c 	struct sk_buff *cur, *tmp;
cur               211 drivers/net/wireless/realtek/rtw88/tx.c 	skb_queue_walk_safe(&tx_report->queue, cur, tmp) {
cur               212 drivers/net/wireless/realtek/rtw88/tx.c 		n = (u8 *)IEEE80211_SKB_CB(cur)->status.status_driver_data;
cur               214 drivers/net/wireless/realtek/rtw88/tx.c 			__skb_unlink(cur, &tx_report->queue);
cur               215 drivers/net/wireless/realtek/rtw88/tx.c 			rtw_tx_report_tx_status(rtwdev, cur, st == 0);
cur              1087 drivers/nvme/host/core.c 		struct nvme_ns_id_desc *cur = data + pos;
cur              1089 drivers/nvme/host/core.c 		if (cur->nidl == 0)
cur              1092 drivers/nvme/host/core.c 		switch (cur->nidt) {
cur              1094 drivers/nvme/host/core.c 			if (cur->nidl != NVME_NIDT_EUI64_LEN) {
cur              1097 drivers/nvme/host/core.c 					 cur->nidl);
cur              1101 drivers/nvme/host/core.c 			memcpy(ids->eui64, data + pos + sizeof(*cur), len);
cur              1104 drivers/nvme/host/core.c 			if (cur->nidl != NVME_NIDT_NGUID_LEN) {
cur              1107 drivers/nvme/host/core.c 					 cur->nidl);
cur              1111 drivers/nvme/host/core.c 			memcpy(ids->nguid, data + pos + sizeof(*cur), len);
cur              1114 drivers/nvme/host/core.c 			if (cur->nidl != NVME_NIDT_UUID_LEN) {
cur              1117 drivers/nvme/host/core.c 					 cur->nidl);
cur              1121 drivers/nvme/host/core.c 			uuid_copy(&ids->uuid, data + pos + sizeof(*cur));
cur              1125 drivers/nvme/host/core.c 			len = cur->nidl;
cur              1129 drivers/nvme/host/core.c 		len += sizeof(*cur);
cur              1305 drivers/of/base.c 	it->cur = list;
cur              1320 drivers/of/base.c 	if (!it->cur || it->phandle_end >= it->list_end)
cur              1323 drivers/of/base.c 	it->cur = it->phandle_end;
cur              1326 drivers/of/base.c 	it->phandle = be32_to_cpup(it->cur++);
cur              1368 drivers/of/base.c 		if (it->cur + count > it->list_end) {
cur              1376 drivers/of/base.c 	it->phandle_end = it->cur + count;
cur              1403 drivers/of/base.c 		args[i] = be32_to_cpup(it->cur++);
cur              1585 drivers/of/base.c 	struct device_node *cur, *new = NULL;
cur              1620 drivers/of/base.c 	cur = out_args->np;
cur              1621 drivers/of/base.c 	ret = of_property_read_u32(cur, cells_name, &list_size);
cur              1630 drivers/of/base.c 	while (cur) {
cur              1632 drivers/of/base.c 		map = of_get_property(cur, map_name, &map_len);
cur              1640 drivers/of/base.c 		mask = of_get_property(cur, mask_name, NULL);
cur              1681 drivers/of/base.c 		pass = of_get_property(cur, pass_name, NULL);
cur              1704 drivers/of/base.c 		of_node_put(cur);
cur              1705 drivers/of/base.c 		cur = new;
cur              1708 drivers/of/base.c 	of_node_put(cur);
cur               115 drivers/of/fdt.c 	int cur;
cur               119 drivers/of/fdt.c 	for (cur = fdt_first_property_offset(blob, offset);
cur               120 drivers/of/fdt.c 	     cur >= 0;
cur               121 drivers/of/fdt.c 	     cur = fdt_next_property_offset(blob, cur)) {
cur               126 drivers/of/fdt.c 		val = fdt_getprop_by_offset(blob, cur, &pname, &sz);
cur               128 drivers/of/fdt.c 			pr_warn("Cannot locate property at 0x%x\n", cur);
cur               133 drivers/of/fdt.c 			pr_warn("Cannot find property name at 0x%x\n", cur);
cur               483 drivers/of/property.c const __be32 *of_prop_next_u32(struct property *prop, const __be32 *cur,
cur               486 drivers/of/property.c 	const void *curv = cur;
cur               491 drivers/of/property.c 	if (!cur) {
cur               496 drivers/of/property.c 	curv += sizeof(*cur);
cur               506 drivers/of/property.c const char *of_prop_next_string(struct property *prop, const char *cur)
cur               508 drivers/of/property.c 	const void *curv = cur;
cur               513 drivers/of/property.c 	if (!cur)
cur               516 drivers/of/property.c 	curv += strlen(cur) + 1;
cur                77 drivers/of/resolver.c 	char *value, *cur, *end, *node_path, *prop_name, *s;
cur                87 drivers/of/resolver.c 	for (cur = value; cur < end; cur += len + 1) {
cur                88 drivers/of/resolver.c 		len = strlen(cur);
cur                90 drivers/of/resolver.c 		node_path = cur;
cur                91 drivers/of/resolver.c 		s = strchr(cur, ':');
cur               438 drivers/parisc/eisa.c 	char *cur = str;
cur               442 drivers/parisc/eisa.c 	while (cur != NULL) {
cur               445 drivers/parisc/eisa.c 		val = (int) simple_strtoul(cur, &pe, 0);
cur               456 drivers/parisc/eisa.c 		if ((cur = strchr(cur, ','))) {
cur               457 drivers/parisc/eisa.c 			cur++;
cur               179 drivers/parisc/led.c 	char *cur, lbuf[32];
cur               192 drivers/parisc/led.c 	cur = lbuf;
cur               197 drivers/parisc/led.c 		d = *cur++ - '0';
cur               201 drivers/parisc/led.c 		if (*cur++ != ' ') goto parse_error;
cur               203 drivers/parisc/led.c 		d = *cur++ - '0';
cur               207 drivers/parisc/led.c 		if (*cur++ != ' ') goto parse_error;
cur               209 drivers/parisc/led.c 		d = *cur++ - '0';
cur               215 drivers/parisc/led.c 		if (*cur && cur[strlen(cur)-1] == '\n')
cur               216 drivers/parisc/led.c 			cur[strlen(cur)-1] = 0;
cur               217 drivers/parisc/led.c 		if (*cur == 0) 
cur               218 drivers/parisc/led.c 			cur = lcd_text_default;
cur               219 drivers/parisc/led.c 		lcd_print(cur);
cur               335 drivers/pci/controller/pci-mvebu.c 				  struct mvebu_pcie_window *cur)
cur               337 drivers/pci/controller/pci-mvebu.c 	if (desired->base == cur->base && desired->remap == cur->remap &&
cur               338 drivers/pci/controller/pci-mvebu.c 	    desired->size == cur->size)
cur               341 drivers/pci/controller/pci-mvebu.c 	if (cur->size != 0) {
cur               342 drivers/pci/controller/pci-mvebu.c 		mvebu_pcie_del_windows(port, cur->base, cur->size);
cur               343 drivers/pci/controller/pci-mvebu.c 		cur->size = 0;
cur               344 drivers/pci/controller/pci-mvebu.c 		cur->base = 0;
cur               358 drivers/pci/controller/pci-mvebu.c 	*cur = *desired;
cur               278 drivers/pinctrl/bcm/pinctrl-bcm2835.c 	enum bcm2835_fsel cur = (val >> FSEL_SHIFT(pin)) & BCM2835_FSEL_MASK;
cur               281 drivers/pinctrl/bcm/pinctrl-bcm2835.c 			bcm2835_functions[cur]);
cur               283 drivers/pinctrl/bcm/pinctrl-bcm2835.c 	if (cur == fsel)
cur               286 drivers/pinctrl/bcm/pinctrl-bcm2835.c 	if (cur != BCM2835_FSEL_GPIO_IN && fsel != BCM2835_FSEL_GPIO_IN) {
cur               124 drivers/pinctrl/pinctrl-single.c 	int cur;
cur               664 drivers/pinctrl/pinctrl-single.c 	i = pcs->pins.cur;
cur               685 drivers/pinctrl/pinctrl-single.c 	pcs->pins.cur++;
cur               323 drivers/platform/x86/eeepc-laptop.c 	int cur;
cur               328 drivers/platform/x86/eeepc-laptop.c 	c->cur = get_acpi(eeepc, CM_ASL_CPUFV);
cur               329 drivers/platform/x86/eeepc-laptop.c 	if (c->cur < 0)
cur               332 drivers/platform/x86/eeepc-laptop.c 	c->num = (c->cur >> 8) & 0xff;
cur               333 drivers/platform/x86/eeepc-laptop.c 	c->cur &= 0xff;
cur               365 drivers/platform/x86/eeepc-laptop.c 	return sprintf(buf, "%#x\n", (c.num << 8) | c.cur);
cur               197 drivers/platform/x86/msi-wmi.c 			ktime_t cur = ktime_get_real();
cur               198 drivers/platform/x86/msi-wmi.c 			ktime_t diff = ktime_sub(cur, last_pressed);
cur               207 drivers/platform/x86/msi-wmi.c 			last_pressed = cur;
cur                72 drivers/pnp/core.c 		struct pnp_protocol *cur = to_pnp_protocol(pos);
cur                73 drivers/pnp/core.c 		if (cur->number == nodenum) {
cur               122 drivers/power/supply/apm_power.c 	union power_supply_propval cur;
cur               182 drivers/power/supply/apm_power.c 	if (_MPSY_PROP(cur_avg_prop, &cur)) {
cur               184 drivers/power/supply/apm_power.c 		if (_MPSY_PROP(cur_now_prop, &cur))
cur               189 drivers/power/supply/apm_power.c 		return ((cur.intval - full.intval) * 60L) / I.intval;
cur               191 drivers/power/supply/apm_power.c 		return -((cur.intval - empty.intval) * 60L) / I.intval;
cur               218 drivers/power/supply/apm_power.c 	union power_supply_propval empty, full, cur;
cur               257 drivers/power/supply/apm_power.c 	if (_MPSY_PROP(avg_prop, &cur)) {
cur               259 drivers/power/supply/apm_power.c 		if (_MPSY_PROP(now_prop, &cur))
cur               270 drivers/power/supply/apm_power.c 		ret =  ((cur.intval - empty.intval) * 100L) /
cur               389 drivers/power/supply/axp288_fuel_gauge.c static int fuel_gauge_get_current(struct axp288_fg_info *info, int *cur)
cur               399 drivers/power/supply/axp288_fuel_gauge.c 		*cur = -1 * discharge;
cur               403 drivers/power/supply/axp288_fuel_gauge.c 	return iio_read_channel_raw(info->iio_channel[BAT_CHRG_CURR], cur);
cur               182 drivers/power/supply/s3c_adc_battery.c 	if (full_volt < calc_full_volt(lut->volt, lut->cur,
cur               189 drivers/power/supply/s3c_adc_battery.c 			lut_volt1 = calc_full_volt(lut[0].volt, lut[0].cur,
cur               191 drivers/power/supply/s3c_adc_battery.c 			lut_volt2 = calc_full_volt(lut[1].volt, lut[1].cur,
cur               110 drivers/power/supply/sc2731_charger.c static int sc2731_charger_set_current(struct sc2731_charger_info *info, u32 cur)
cur               115 drivers/power/supply/sc2731_charger.c 	if (cur > SC2731_CURRENT_LIMIT_2000)
cur               116 drivers/power/supply/sc2731_charger.c 		cur = SC2731_CURRENT_LIMIT_2000;
cur               117 drivers/power/supply/sc2731_charger.c 	else if (cur < SC2731_CURRENT_PRECHG)
cur               118 drivers/power/supply/sc2731_charger.c 		cur = SC2731_CURRENT_PRECHG;
cur               121 drivers/power/supply/sc2731_charger.c 	val = (cur - SC2731_CURRENT_PRECHG) / SC2731_CURRENT_STEP;
cur               150 drivers/power/supply/sc2731_charger.c 				      u32 *cur)
cur               160 drivers/power/supply/sc2731_charger.c 	*cur = val * SC2731_CURRENT_STEP + SC2731_CURRENT_PRECHG;
cur               166 drivers/power/supply/sc2731_charger.c 					    u32 *cur)
cur               179 drivers/power/supply/sc2731_charger.c 		*cur = SC2731_CURRENT_LIMIT_100;
cur               183 drivers/power/supply/sc2731_charger.c 		*cur = SC2731_CURRENT_LIMIT_2000;
cur               187 drivers/power/supply/sc2731_charger.c 		*cur = SC2731_CURRENT_LIMIT_900;
cur               191 drivers/power/supply/sc2731_charger.c 		*cur = SC2731_CURRENT_LIMIT_500;
cur               244 drivers/power/supply/sc2731_charger.c 	u32 cur;
cur               260 drivers/power/supply/sc2731_charger.c 			ret = sc2731_charger_get_current(info, &cur);
cur               264 drivers/power/supply/sc2731_charger.c 			val->intval = cur * 1000;
cur               272 drivers/power/supply/sc2731_charger.c 			ret = sc2731_charger_get_current_limit(info, &cur);
cur               276 drivers/power/supply/sc2731_charger.c 			val->intval = cur * 1000;
cur               274 drivers/power/supply/sc27xx_fuel_gauge.c 	int volt, cur, oci, ocv, ret;
cur               295 drivers/power/supply/sc27xx_fuel_gauge.c 			  &cur);
cur               299 drivers/power/supply/sc27xx_fuel_gauge.c 	cur <<= 1;
cur               300 drivers/power/supply/sc27xx_fuel_gauge.c 	oci = sc27xx_fgu_adc_to_current(data, cur - SC27XX_FGU_CUR_BASIC_ADC);
cur               420 drivers/power/supply/sc27xx_fuel_gauge.c 	int ret, cur;
cur               422 drivers/power/supply/sc27xx_fuel_gauge.c 	ret = regmap_read(data->regmap, data->base + SC27XX_FGU_CURRENT, &cur);
cur               430 drivers/power/supply/sc27xx_fuel_gauge.c 	*val = sc27xx_fgu_adc_to_current(data, cur - SC27XX_FGU_CUR_BASIC_ADC);
cur               437 drivers/power/supply/sc27xx_fuel_gauge.c 	int vol, cur, ret;
cur               443 drivers/power/supply/sc27xx_fuel_gauge.c 	ret = sc27xx_fgu_get_current(data, &cur);
cur               448 drivers/power/supply/sc27xx_fuel_gauge.c 	*val = vol * 1000 - cur * data->internal_resist;
cur               241 drivers/power/supply/twl4030_charger.c 	int cur;
cur               252 drivers/power/supply/twl4030_charger.c 		cur = bci->ac_cur;
cur               255 drivers/power/supply/twl4030_charger.c 		cur = bci->usb_cur;
cur               257 drivers/power/supply/twl4030_charger.c 		if (cur > bci->usb_cur_target) {
cur               258 drivers/power/supply/twl4030_charger.c 			cur = bci->usb_cur_target;
cur               259 drivers/power/supply/twl4030_charger.c 			bci->usb_cur = cur;
cur               261 drivers/power/supply/twl4030_charger.c 		if (cur < bci->usb_cur_target)
cur               272 drivers/power/supply/twl4030_charger.c 	if (cur > 852000)
cur               353 drivers/power/supply/twl4030_charger.c 	reg = ua2regval(cur, cgain);
cur               481 drivers/pwm/pwm-samsung.c 	const __be32 *cur;
cur               490 drivers/pwm/pwm-samsung.c 	of_property_for_each_u32(np, "samsung,pwm-outputs", prop, cur, val) {
cur                97 drivers/pwm/pwm-sti.c 	struct pwm_device *cur;
cur               169 drivers/pwm/pwm-sti.c 	struct pwm_device *cur = pc->cur;
cur               176 drivers/pwm/pwm-sti.c 		period_same = (period_ns == pwm_get_period(cur));
cur               191 drivers/pwm/pwm-sti.c 	    ((ncfg == 1) && (pwm->hwpwm == cur->hwpwm)) ||
cur               192 drivers/pwm/pwm-sti.c 	    ((ncfg == 1) && (pwm->hwpwm != cur->hwpwm) && period_same) ||
cur               236 drivers/pwm/pwm-sti.c 		pc->cur = pwm;
cur               382 drivers/s390/block/dasd_devmap.c 	char *cur;
cur               386 drivers/s390/block/dasd_devmap.c 		cur = dasd[i];
cur               387 drivers/s390/block/dasd_devmap.c 		if (!cur)
cur               389 drivers/s390/block/dasd_devmap.c 		if (*cur == '\0')
cur               392 drivers/s390/block/dasd_devmap.c 		rc = dasd_parse_keyword(cur);
cur               394 drivers/s390/block/dasd_devmap.c 			rc = dasd_parse_range(cur);
cur              3234 drivers/scsi/gdth.c     char *cur = str;
cur              3239 drivers/scsi/gdth.c     while (cur && isdigit(*cur) && i < MAXHA) {
cur              3240 drivers/scsi/gdth.c         ints[i++] = simple_strtoul(cur, NULL, 0);
cur              3241 drivers/scsi/gdth.c         if ((cur = strchr(cur, ',')) != NULL) cur++;
cur              3245 drivers/scsi/gdth.c     internal_setup(cur, ints);
cur               583 drivers/scsi/ibmvscsi/ibmvfc.c 		vhost->async_crq.cur = 0;
cur               715 drivers/scsi/ibmvscsi/ibmvfc.c 	crq->cur = 0;
cur              3110 drivers/scsi/ibmvscsi/ibmvfc.c 	crq = &async_crq->msgs[async_crq->cur];
cur              3112 drivers/scsi/ibmvscsi/ibmvfc.c 		if (++async_crq->cur == async_crq->size)
cur              3113 drivers/scsi/ibmvscsi/ibmvfc.c 			async_crq->cur = 0;
cur              3133 drivers/scsi/ibmvscsi/ibmvfc.c 	crq = &queue->msgs[queue->cur];
cur              3135 drivers/scsi/ibmvscsi/ibmvfc.c 		if (++queue->cur == queue->size)
cur              3136 drivers/scsi/ibmvscsi/ibmvfc.c 			queue->cur = 0;
cur              4551 drivers/scsi/ibmvscsi/ibmvfc.c 	crq->cur = 0;
cur               552 drivers/scsi/ibmvscsi/ibmvfc.h 	int size, cur;
cur               577 drivers/scsi/ibmvscsi/ibmvfc.h 	int size, cur;
cur               172 drivers/scsi/ibmvscsi/ibmvscsi.c 	crq = &queue->msgs[queue->cur];
cur               174 drivers/scsi/ibmvscsi/ibmvscsi.c 		if (++queue->cur == queue->size)
cur               175 drivers/scsi/ibmvscsi/ibmvscsi.c 			queue->cur = 0;
cur               299 drivers/scsi/ibmvscsi/ibmvscsi.c 	queue->cur = 0;
cur               366 drivers/scsi/ibmvscsi/ibmvscsi.c 	queue->cur = 0;
cur                45 drivers/scsi/ibmvscsi/ibmvscsi.h 	int size, cur;
cur               638 drivers/scsi/ncr53c8xx.c 	char *cur = setup_token;
cur               642 drivers/scsi/ncr53c8xx.c 	while (cur != NULL && (pc = strchr(cur, ':')) != NULL) {
cur               645 drivers/scsi/ncr53c8xx.c 		if (!strncmp(p, cur, pc - cur))
cur               647 drivers/scsi/ncr53c8xx.c 		cur = pc;
cur               655 drivers/scsi/ncr53c8xx.c 	char *cur = str;
cur               660 drivers/scsi/ncr53c8xx.c 	while (cur != NULL && (pc = strchr(cur, ':')) != NULL) {
cur               674 drivers/scsi/ncr53c8xx.c 		switch (get_setup_token(cur)) {
cur               763 drivers/scsi/ncr53c8xx.c 			printk("sym53c8xx_setup: unexpected boot option '%.*s' ignored\n", (int)(pc-cur+1), cur);
cur               767 drivers/scsi/ncr53c8xx.c 		if ((cur = strchr(cur, ARG_SEP)) != NULL)
cur               768 drivers/scsi/ncr53c8xx.c 			++cur;
cur               188 drivers/scsi/snic/snic_disc.c 	struct list_head *cur, *nxt;
cur               191 drivers/scsi/snic/snic_disc.c 	list_for_each_safe(cur, nxt, &snic->disc.tgt_list) {
cur               192 drivers/scsi/snic/snic_disc.c 		tgt = list_entry(cur, struct snic_tgt, list);
cur               545 drivers/scsi/snic/snic_disc.c 	struct list_head *cur, *nxt;
cur               553 drivers/scsi/snic/snic_disc.c 	list_for_each_safe(cur, nxt, &snic->disc.tgt_list) {
cur               554 drivers/scsi/snic/snic_disc.c 		tgt = list_entry(cur, struct snic_tgt, list);
cur               401 drivers/scsi/snic/snic_io.c 	struct list_head *cur, *nxt;
cur               405 drivers/scsi/snic/snic_io.c 	list_for_each_safe(cur, nxt, &snic->spl_cmd_list) {
cur               406 drivers/scsi/snic/snic_io.c 		rqi = list_entry(cur, struct snic_req_info, list);
cur               351 drivers/scsi/sym53c8xx_2/sym_fw.c 	u32 *end, *cur;
cur               354 drivers/scsi/sym53c8xx_2/sym_fw.c 	cur = start;
cur               357 drivers/scsi/sym53c8xx_2/sym_fw.c 	while (cur < end) {
cur               359 drivers/scsi/sym53c8xx_2/sym_fw.c 		opcode = *cur;
cur               369 drivers/scsi/sym53c8xx_2/sym_fw.c 				sym_name(np), (int) (cur-start));
cur               370 drivers/scsi/sym53c8xx_2/sym_fw.c 			++cur;
cur               379 drivers/scsi/sym53c8xx_2/sym_fw.c 			*cur++ = 0;
cur               384 drivers/scsi/sym53c8xx_2/sym_fw.c 			printf ("%d:  <%x>\n", (int) (cur-start),
cur               408 drivers/scsi/sym53c8xx_2/sym_fw.c 			tmp1 = cur[1];
cur               409 drivers/scsi/sym53c8xx_2/sym_fw.c 			tmp2 = cur[2];
cur               412 drivers/scsi/sym53c8xx_2/sym_fw.c 					sym_name(np), (int) (cur-start));
cur               485 drivers/scsi/sym53c8xx_2/sym_fw.c 		*cur++ = cpu_to_scr(opcode);
cur               492 drivers/scsi/sym53c8xx_2/sym_fw.c 			*cur = cpu_to_scr(*cur);
cur               493 drivers/scsi/sym53c8xx_2/sym_fw.c 			++cur;
cur               501 drivers/scsi/sym53c8xx_2/sym_fw.c 			old = *cur;
cur               534 drivers/scsi/sym53c8xx_2/sym_fw.c 			*cur++ = cpu_to_scr(new);
cur               374 drivers/staging/isdn/gigaset/common.c 	struct at_state_t *cur, *next;
cur               376 drivers/staging/isdn/gigaset/common.c 	list_for_each_entry_safe(cur, next, &cs->temp_at_states, list) {
cur               377 drivers/staging/isdn/gigaset/common.c 		list_del(&cur->list);
cur               378 drivers/staging/isdn/gigaset/common.c 		clear_at_state(cur);
cur               379 drivers/staging/isdn/gigaset/common.c 		kfree(cur);
cur                28 drivers/staging/kpc2000/kpc_dma/dma.c 	struct kpc_dma_descriptor *cur;
cur                39 drivers/staging/kpc2000/kpc_dma/dma.c 	cur = eng->desc_completed;
cur                41 drivers/staging/kpc2000/kpc_dma/dma.c 		cur = cur->Next;
cur                42 drivers/staging/kpc2000/kpc_dma/dma.c 		dev_dbg(&eng->pldev->dev, "Handling completed descriptor %p (acd = %p)\n", cur, cur->acd);
cur                43 drivers/staging/kpc2000/kpc_dma/dma.c 		BUG_ON(cur == eng->desc_next); // Ordering failure.
cur                45 drivers/staging/kpc2000/kpc_dma/dma.c 		if (cur->DescControlFlags & DMA_DESC_CTL_SOP) {
cur                50 drivers/staging/kpc2000/kpc_dma/dma.c 		eng->accumulated_bytes += cur->DescByteCount;
cur                51 drivers/staging/kpc2000/kpc_dma/dma.c 		if (cur->DescStatusFlags & DMA_DESC_STS_ERROR)
cur                54 drivers/staging/kpc2000/kpc_dma/dma.c 		if (cur->DescStatusFlags & DMA_DESC_STS_SHORT)
cur                57 drivers/staging/kpc2000/kpc_dma/dma.c 		if (cur->DescControlFlags & DMA_DESC_CTL_EOP) {
cur                58 drivers/staging/kpc2000/kpc_dma/dma.c 			if (cur->acd)
cur                59 drivers/staging/kpc2000/kpc_dma/dma.c 				transfer_complete_cb(cur->acd, eng->accumulated_bytes, eng->accumulated_flags | ACD_FLAG_DONE);
cur                62 drivers/staging/kpc2000/kpc_dma/dma.c 		eng->desc_completed = cur;
cur                63 drivers/staging/kpc2000/kpc_dma/dma.c 	} while (cur->MyDMAAddr != GetEngineCompletePtr(eng));
cur                88 drivers/staging/kpc2000/kpc_dma/dma.c 	struct kpc_dma_descriptor *cur;
cur               119 drivers/staging/kpc2000/kpc_dma/dma.c 	cur = eng->desc_pool_first;
cur               128 drivers/staging/kpc2000/kpc_dma/dma.c 		cur->DescNextDescPtr = next_handle;
cur               129 drivers/staging/kpc2000/kpc_dma/dma.c 		cur->Next = next;
cur               130 drivers/staging/kpc2000/kpc_dma/dma.c 		cur = next;
cur               135 drivers/staging/kpc2000/kpc_dma/dma.c 	cur->Next = eng->desc_pool_first;
cur               136 drivers/staging/kpc2000/kpc_dma/dma.c 	cur->DescNextDescPtr = eng->desc_pool_first->MyDMAAddr;
cur               138 drivers/staging/kpc2000/kpc_dma/dma.c 	eng->desc_pool_last = cur;
cur               209 drivers/staging/kpc2000/kpc_dma/dma.c 	struct kpc_dma_descriptor *cur;
cur               215 drivers/staging/kpc2000/kpc_dma/dma.c 	cur = eng->desc_pool_first;
cur               219 drivers/staging/kpc2000/kpc_dma/dma.c 		struct kpc_dma_descriptor *next = cur->Next;
cur               220 drivers/staging/kpc2000/kpc_dma/dma.c 		dma_addr_t next_handle = cur->DescNextDescPtr;
cur               222 drivers/staging/kpc2000/kpc_dma/dma.c 		dma_pool_free(eng->desc_pool, cur, cur_handle);
cur               224 drivers/staging/kpc2000/kpc_dma/dma.c 		cur = next;
cur               236 drivers/staging/kpc2000/kpc_dma/dma.c 	struct kpc_dma_descriptor *cur = eng->desc_next;
cur               238 drivers/staging/kpc2000/kpc_dma/dma.c 	while (cur != eng->desc_completed) {
cur               239 drivers/staging/kpc2000/kpc_dma/dma.c 		BUG_ON(cur == NULL);
cur               241 drivers/staging/kpc2000/kpc_dma/dma.c 		cur = cur->Next;
cur               271 drivers/staging/kpc2000/kpc_dma/fileops.c 	struct kpc_dma_descriptor *cur;
cur               279 drivers/staging/kpc2000/kpc_dma/fileops.c 	cur = eng->desc_completed->Next;
cur               280 drivers/staging/kpc2000/kpc_dma/fileops.c 	while (cur != eng->desc_next) {
cur               281 drivers/staging/kpc2000/kpc_dma/fileops.c 		dev_dbg(&eng->pldev->dev, "Aborting descriptor %p (acd = %p)\n", cur, cur->acd);
cur               282 drivers/staging/kpc2000/kpc_dma/fileops.c 		if (cur->DescControlFlags & DMA_DESC_CTL_EOP) {
cur               283 drivers/staging/kpc2000/kpc_dma/fileops.c 			if (cur->acd)
cur               284 drivers/staging/kpc2000/kpc_dma/fileops.c 				transfer_complete_cb(cur->acd, 0, ACD_FLAG_ABORT);
cur               287 drivers/staging/kpc2000/kpc_dma/fileops.c 		clear_desc(cur);
cur               288 drivers/staging/kpc2000/kpc_dma/fileops.c 		eng->desc_completed = cur;
cur               290 drivers/staging/kpc2000/kpc_dma/fileops.c 		cur = cur->Next;
cur                99 drivers/staging/media/imx/imx-media-fim.c 		fim->enabled = en->cur.val;
cur               100 drivers/staging/media/imx/imx-media-fim.c 		fim->icap_flags = icap_edge->cur.val;
cur               101 drivers/staging/media/imx/imx-media-fim.c 		fim->icap_channel = icap_chan->cur.val;
cur               102 drivers/staging/media/imx/imx-media-fim.c 		fim->num_avg = num->cur.val;
cur               103 drivers/staging/media/imx/imx-media-fim.c 		fim->num_skip = skip->cur.val;
cur               104 drivers/staging/media/imx/imx-media-fim.c 		fim->tolerance_min = tol_min->cur.val;
cur               105 drivers/staging/media/imx/imx-media-fim.c 		fim->tolerance_max = tol_max->cur.val;
cur              2513 drivers/staging/rtl8188eu/core/rtw_mlme_ext.c 	struct wlan_bssid_ex *cur = &pmlmeinfo->network;
cur              2529 drivers/staging/rtl8188eu/core/rtw_mlme_ext.c 		if ((ielen != 0 && memcmp((void *)(p+2), (void *)cur->ssid.ssid, cur->ssid.ssid_length)) ||
cur              2908 drivers/staging/rtl8188eu/core/rtw_mlme_ext.c 	struct wlan_bssid_ex *cur = &pmlmeinfo->network;
cur              2978 drivers/staging/rtl8188eu/core/rtw_mlme_ext.c 		if (memcmp((void *)(p+2), cur->ssid.ssid, cur->ssid.ssid_length))
cur              2981 drivers/staging/rtl8188eu/core/rtw_mlme_ext.c 		if (ie_len != cur->ssid.ssid_length)
cur               605 drivers/staging/rtl8723bs/core/rtw_mlme_ext.c 	struct wlan_bssid_ex	*cur = &pmlmeinfo->network;
cur               739 drivers/staging/rtl8723bs/core/rtw_mlme_ext.c 		if ((ielen != 0 && false == !memcmp((void *)(p+2), (void *)cur->Ssid.Ssid, cur->Ssid.SsidLength))
cur              1187 drivers/staging/rtl8723bs/core/rtw_mlme_ext.c 	struct wlan_bssid_ex	*cur = &(pmlmeinfo->network);
cur              1265 drivers/staging/rtl8723bs/core/rtw_mlme_ext.c 		if (memcmp((void *)(p+2), cur->Ssid.Ssid, cur->Ssid.SsidLength))
cur              1268 drivers/staging/rtl8723bs/core/rtw_mlme_ext.c 		if (ie_len != cur->Ssid.SsidLength)
cur               532 drivers/tty/vt/keyboard.c 	int i, cur = fg_console;
cur               536 drivers/tty/vt/keyboard.c 		cur = want_console;
cur               538 drivers/tty/vt/keyboard.c 	for (i = cur - 1; i != cur; i--) {
cur               549 drivers/tty/vt/keyboard.c 	int i, cur = fg_console;
cur               553 drivers/tty/vt/keyboard.c 		cur = want_console;
cur               555 drivers/tty/vt/keyboard.c 	for (i = cur+1; i != cur; i++) {
cur              1209 drivers/usb/gadget/udc/bcm63xx_udc.c 	struct bcm63xx_req *breq = our_req(req), *cur;
cur              1219 drivers/usb/gadget/udc/bcm63xx_udc.c 	cur = list_first_entry(&bep->queue, struct bcm63xx_req, queue);
cur              1222 drivers/usb/gadget/udc/bcm63xx_udc.c 	if (breq == cur) {
cur               385 drivers/usb/musb/tusb6010_omap.c 		int cur = (reg & (0xf << (i * 5))) >> (i * 5);
cur               386 drivers/usb/musb/tusb6010_omap.c 		if (cur == 0) {
cur               376 drivers/usb/typec/altmodes/displayport.c 	u8 cur;
cur               382 drivers/usb/typec/altmodes/displayport.c 	cur = DP_CONF_CURRENTLY(dp->data.conf);
cur               384 drivers/usb/typec/altmodes/displayport.c 	len = sprintf(buf, "%s ", cur ? "USB" : "[USB]");
cur               387 drivers/usb/typec/altmodes/displayport.c 		if (i == cur)
cur               469 drivers/usb/typec/altmodes/displayport.c 	u8 cur;
cur               474 drivers/usb/typec/altmodes/displayport.c 	cur = get_count_order(DP_CONF_GET_PIN_ASSIGN(dp->data.conf));
cur               483 drivers/usb/typec/altmodes/displayport.c 			if (i == cur)
cur                52 drivers/usb/typec/ucsi/displayport.c 	u8 cur = 0;
cur                67 drivers/usb/typec/ucsi/displayport.c 	ret = ucsi_send_command(dp->con->ucsi, &ctrl, &cur, sizeof(cur));
cur                73 drivers/usb/typec/ucsi/displayport.c 		cur = 0xff;
cur                76 drivers/usb/typec/ucsi/displayport.c 	if (cur != 0xff) {
cur                78 drivers/usb/typec/ucsi/displayport.c 		if (dp->con->port_altmode[cur] == alt)
cur               227 drivers/usb/typec/ucsi/ucsi.c 	u8 cur;
cur               231 drivers/usb/typec/ucsi/ucsi.c 	ret = ucsi_run_command(con->ucsi, &ctrl, &cur, sizeof(cur));
cur               238 drivers/usb/typec/ucsi/ucsi.c 		cur = 0xff;
cur               241 drivers/usb/typec/ucsi/ucsi.c 	if (cur < UCSI_MAX_ALTMODES)
cur               242 drivers/usb/typec/ucsi/ucsi.c 		altmode = typec_altmode_get_partner(con->port_altmode[cur]);
cur               558 drivers/vfio/pci/vfio_pci.c 	int cur;
cur               567 drivers/vfio/pci/vfio_pci.c 	if (fill->cur == fill->max)
cur               574 drivers/vfio/pci/vfio_pci.c 	fill->devices[fill->cur].group_id = iommu_group_id(iommu_group);
cur               575 drivers/vfio/pci/vfio_pci.c 	fill->devices[fill->cur].segment = pci_domain_nr(pdev->bus);
cur               576 drivers/vfio/pci/vfio_pci.c 	fill->devices[fill->cur].bus = pdev->bus->number;
cur               577 drivers/vfio/pci/vfio_pci.c 	fill->devices[fill->cur].devfn = pdev->devfn;
cur               578 drivers/vfio/pci/vfio_pci.c 	fill->cur++;
cur              1000 drivers/vfio/pci/vfio_pci.c 			hdr.count = fill.cur;
cur                64 drivers/video/backlight/cr_bllcd.c 	u32 cur = inl(addr);
cur                76 drivers/video/backlight/cr_bllcd.c 		cur &= ~CRVML_BACKLIGHT_OFF;
cur                77 drivers/video/backlight/cr_bllcd.c 		outl(cur, addr);
cur                79 drivers/video/backlight/cr_bllcd.c 		cur |= CRVML_BACKLIGHT_OFF;
cur                80 drivers/video/backlight/cr_bllcd.c 		outl(cur, addr);
cur                89 drivers/video/backlight/cr_bllcd.c 	u32 cur = inl(addr);
cur                92 drivers/video/backlight/cr_bllcd.c 	if (cur & CRVML_BACKLIGHT_OFF)
cur               108 drivers/video/backlight/cr_bllcd.c 	u32 cur = inl(addr);
cur               110 drivers/video/backlight/cr_bllcd.c 	if (!(cur & CRVML_PANEL_ON)) {
cur               112 drivers/video/backlight/cr_bllcd.c 		if (cur & 0x00000001) {
cur               113 drivers/video/backlight/cr_bllcd.c 			cur &= ~CRVML_LVDS_ON;
cur               114 drivers/video/backlight/cr_bllcd.c 			outl(cur, addr);
cur               118 drivers/video/backlight/cr_bllcd.c 		cur |= CRVML_PANEL_ON;
cur               119 drivers/video/backlight/cr_bllcd.c 		outl(cur, addr);
cur               124 drivers/video/backlight/cr_bllcd.c 	if (!(cur & CRVML_LVDS_ON)) {
cur               126 drivers/video/backlight/cr_bllcd.c 		outl(cur | CRVML_LVDS_ON, addr);
cur               133 drivers/video/backlight/cr_bllcd.c 	u32 cur = inl(addr);
cur               136 drivers/video/backlight/cr_bllcd.c 	if (cur & CRVML_LVDS_ON) {
cur               137 drivers/video/backlight/cr_bllcd.c 		cur &= ~CRVML_LVDS_ON;
cur               138 drivers/video/backlight/cr_bllcd.c 		outl(cur, addr);
cur               140 drivers/video/backlight/cr_bllcd.c 	if (cur & CRVML_PANEL_ON) {
cur               142 drivers/video/backlight/cr_bllcd.c 		outl(cur & ~CRVML_PANEL_ON, addr);
cur               176 drivers/video/console/vgacon.c 	int cur;
cur               197 drivers/video/console/vgacon.c 	scrollback->cur  = 0;
cur               268 drivers/video/console/vgacon.c 		vgacon_scrollback_cur->cur = vgacon_scrollback_cur->cnt;
cur               282 drivers/video/console/vgacon.c 		vgacon_scrollback_cur->cur = vgacon_scrollback_cur->cnt;
cur               306 drivers/video/console/vgacon.c 	start = vgacon_scrollback_cur->cur + lines;
cur               321 drivers/video/console/vgacon.c 	vgacon_scrollback_cur->cur = start;
cur               937 drivers/video/fbdev/broadsheetfb.c 	struct page *cur;
cur               947 drivers/video/fbdev/broadsheetfb.c 	list_for_each_entry(cur, &fbdefio->pagelist, lru) {
cur               950 drivers/video/fbdev/broadsheetfb.c 			y1 = (cur->index << PAGE_SHIFT) / xres;
cur               952 drivers/video/fbdev/broadsheetfb.c 		} else if ((prev_index + 1) == cur->index) {
cur               959 drivers/video/fbdev/broadsheetfb.c 			y1 = (cur->index << PAGE_SHIFT) / xres;
cur               962 drivers/video/fbdev/broadsheetfb.c 		prev_index = cur->index;
cur                98 drivers/video/fbdev/core/fb_defio.c 	struct page *cur;
cur               127 drivers/video/fbdev/core/fb_defio.c 	list_for_each_entry(cur, &fbdefio->pagelist, lru) {
cur               133 drivers/video/fbdev/core/fb_defio.c 		if (unlikely(cur == page))
cur               135 drivers/video/fbdev/core/fb_defio.c 		else if (cur->index > page->index)
cur               139 drivers/video/fbdev/core/fb_defio.c 	list_add_tail(&page->lru, &cur->lru);
cur               182 drivers/video/fbdev/core/fb_defio.c 	struct page *cur;
cur               187 drivers/video/fbdev/core/fb_defio.c 	list_for_each_entry(cur, &fbdefio->pagelist, lru) {
cur               188 drivers/video/fbdev/core/fb_defio.c 		lock_page(cur);
cur               189 drivers/video/fbdev/core/fb_defio.c 		page_mkclean(cur);
cur               190 drivers/video/fbdev/core/fb_defio.c 		unlock_page(cur);
cur               472 drivers/video/fbdev/metronomefb.c 	struct page *cur;
cur               477 drivers/video/fbdev/metronomefb.c 	list_for_each_entry(cur, &fbdefio->pagelist, lru) {
cur               479 drivers/video/fbdev/metronomefb.c 					(cur->index << PAGE_SHIFT));
cur               480 drivers/video/fbdev/metronomefb.c 		par->metromem_img_csum -= par->csum_table[cur->index];
cur               481 drivers/video/fbdev/metronomefb.c 		par->csum_table[cur->index] = cksum;
cur                75 drivers/video/fbdev/nvidia/nv_hw.c 	int cur = par->CurrentState->cursor1;
cur                85 drivers/video/fbdev/nvidia/nv_hw.c 	return (cur & 0x01);
cur               173 drivers/video/fbdev/riva/riva_hw.c   char cur;
cur               251 drivers/video/fbdev/riva/riva_hw.c     int last, next, cur;
cur               260 drivers/video/fbdev/riva/riva_hw.c     cur = ainfo->cur;
cur               328 drivers/video/fbdev/riva/riva_hw.c         last = cur;
cur               329 drivers/video/fbdev/riva/riva_hw.c         cur = next;
cur               331 drivers/video/fbdev/riva/riva_hw.c         switch (cur)
cur               334 drivers/video/fbdev/riva/riva_hw.c                 if (last==cur)    misses = 0;
cur               338 drivers/video/fbdev/riva/riva_hw.c                 if (last!=cur)
cur               350 drivers/video/fbdev/riva/riva_hw.c                 if (last==cur)    misses = 0;
cur               354 drivers/video/fbdev/riva/riva_hw.c                 if (last!=cur)
cur               366 drivers/video/fbdev/riva/riva_hw.c                 if (last==cur)    misses = 0;
cur               455 drivers/video/fbdev/riva/riva_hw.c         ainfo->cur = ENGINE;
cur               467 drivers/video/fbdev/riva/riva_hw.c         ainfo->cur = MPORT;
cur               482 drivers/video/fbdev/riva/riva_hw.c         ainfo->cur = GRAPHICS;
cur               494 drivers/video/fbdev/riva/riva_hw.c         ainfo->cur = VIDEO;
cur               958 drivers/video/fbdev/smscufx.c 	struct page *cur;
cur               969 drivers/video/fbdev/smscufx.c 	list_for_each_entry(cur, &fbdefio->pagelist, lru) {
cur               974 drivers/video/fbdev/smscufx.c 		const int y = (cur->index << PAGE_SHIFT) / (width * 2);
cur               784 drivers/video/fbdev/udlfb.c 	struct page *cur;
cur               811 drivers/video/fbdev/udlfb.c 	list_for_each_entry(cur, &fbdefio->pagelist, lru) {
cur               814 drivers/video/fbdev/udlfb.c 				  &cmd, cur->index << PAGE_SHIFT,
cur               887 drivers/video/fbdev/vermilion/vermilion.c 	u32 cur = VML_READ32(par, VML_PIPEACONF);
cur               894 drivers/video/fbdev/vermilion/vermilion.c 		VML_WRITE32(par, VML_PIPEACONF, cur & ~VML_PIPE_FORCE_BORDER);
cur               901 drivers/video/fbdev/vermilion/vermilion.c 		VML_WRITE32(par, VML_PIPEACONF, cur | VML_PIPE_FORCE_BORDER);
cur              1376 drivers/video/fbdev/via/hw.c 	struct via_pll_config cur, up, down, best = {0, 1, 0};
cur              1381 drivers/video/fbdev/via/hw.c 		cur.rshift = limits[i].rshift;
cur              1382 drivers/video/fbdev/via/hw.c 		cur.divisor = limits[i].divisor;
cur              1383 drivers/video/fbdev/via/hw.c 		cur.multiplier = clk / ((f0 / cur.divisor)>>cur.rshift);
cur              1384 drivers/video/fbdev/via/hw.c 		f = abs(get_pll_output_frequency(f0, cur) - clk);
cur              1385 drivers/video/fbdev/via/hw.c 		up = down = cur;
cur              1389 drivers/video/fbdev/via/hw.c 			cur = up;
cur              1391 drivers/video/fbdev/via/hw.c 			cur = down;
cur              1393 drivers/video/fbdev/via/hw.c 		if (cur.multiplier < limits[i].multiplier_min)
cur              1394 drivers/video/fbdev/via/hw.c 			cur.multiplier = limits[i].multiplier_min;
cur              1395 drivers/video/fbdev/via/hw.c 		else if (cur.multiplier > limits[i].multiplier_max)
cur              1396 drivers/video/fbdev/via/hw.c 			cur.multiplier = limits[i].multiplier_max;
cur              1398 drivers/video/fbdev/via/hw.c 		f = abs(get_pll_output_frequency(f0, cur) - clk);
cur              1400 drivers/video/fbdev/via/hw.c 			best = cur;
cur                58 drivers/watchdog/bcm2835_wdt.c 	uint32_t cur;
cur                60 drivers/watchdog/bcm2835_wdt.c 	cur = readl(wdt->base + PM_RSTC);
cur                62 drivers/watchdog/bcm2835_wdt.c 	return !!(cur & PM_RSTC_WRCFG_FULL_RESET);
cur                68 drivers/watchdog/bcm2835_wdt.c 	uint32_t cur;
cur                75 drivers/watchdog/bcm2835_wdt.c 	cur = readl_relaxed(wdt->base + PM_RSTC);
cur                76 drivers/watchdog/bcm2835_wdt.c 	writel_relaxed(PM_PASSWORD | (cur & PM_RSTC_WRCFG_CLR) |
cur               112 drivers/watchdog/bcm_kona_wdt.c 		int ctl, cur, ctl_sec, cur_sec, res;
cur               116 drivers/watchdog/bcm_kona_wdt.c 		cur = cur_val & SECWDOG_COUNT_MASK;
cur               118 drivers/watchdog/bcm_kona_wdt.c 		cur_sec = TICKS_TO_SECS(cur, wdt);
cur               126 drivers/watchdog/bcm_kona_wdt.c 			   cur_sec, cur, cur,
cur              1389 drivers/xen/grant-table.c 	unsigned int cur, extra;
cur              1392 drivers/xen/grant-table.c 	cur = nr_grant_frames;
cur              1395 drivers/xen/grant-table.c 	if (cur + extra > gnttab_max_grant_frames()) {
cur              1399 drivers/xen/grant-table.c 				    cur, extra, gnttab_max_grant_frames(),
cur              1404 drivers/xen/grant-table.c 	rc = gnttab_map(cur, cur + extra - 1);
cur               104 fs/afs/server_list.c 	struct afs_server *cur;
cur               118 fs/afs/server_list.c 	cur = old->servers[old->preferred].server;
cur               120 fs/afs/server_list.c 		if (new->servers[j].server == cur) {
cur              2004 fs/btrfs/backref.c 	u32 cur;
cur              2039 fs/btrfs/backref.c 		for (cur = 0; cur < btrfs_item_size(eb, item); cur += len) {
cur              2044 fs/btrfs/backref.c 				cur, found_key.objectid,
cur              1203 fs/btrfs/check-integrity.c 	size_t cur;
cur              1214 fs/btrfs/check-integrity.c 		cur = min(len, ((size_t)PAGE_SIZE - pgoff));
cur              1217 fs/btrfs/check-integrity.c 		memcpy(dst, kaddr + pgoff, cur);
cur              1219 fs/btrfs/check-integrity.c 		dst += cur;
cur              1220 fs/btrfs/check-integrity.c 		len -= cur;
cur               404 fs/btrfs/ctree.c 	struct tree_mod_elem *cur;
cur               413 fs/btrfs/ctree.c 		cur = rb_entry(*new, struct tree_mod_elem, node);
cur               415 fs/btrfs/ctree.c 		if (cur->logical < tm->logical)
cur               417 fs/btrfs/ctree.c 		else if (cur->logical > tm->logical)
cur               419 fs/btrfs/ctree.c 		else if (cur->seq < tm->seq)
cur               421 fs/btrfs/ctree.c 		else if (cur->seq > tm->seq)
cur               686 fs/btrfs/ctree.c 	struct tree_mod_elem *cur = NULL;
cur               693 fs/btrfs/ctree.c 		cur = rb_entry(node, struct tree_mod_elem, node);
cur               694 fs/btrfs/ctree.c 		if (cur->logical < start) {
cur               696 fs/btrfs/ctree.c 		} else if (cur->logical > start) {
cur               698 fs/btrfs/ctree.c 		} else if (cur->seq < min_seq) {
cur               703 fs/btrfs/ctree.c 				BUG_ON(found->seq > cur->seq);
cur               704 fs/btrfs/ctree.c 			found = cur;
cur               706 fs/btrfs/ctree.c 		} else if (cur->seq > min_seq) {
cur               709 fs/btrfs/ctree.c 				BUG_ON(found->seq < cur->seq);
cur               710 fs/btrfs/ctree.c 			found = cur;
cur               713 fs/btrfs/ctree.c 			found = cur;
cur              1590 fs/btrfs/ctree.c 	struct extent_buffer *cur;
cur              1648 fs/btrfs/ctree.c 		cur = find_extent_buffer(fs_info, blocknr);
cur              1649 fs/btrfs/ctree.c 		if (cur)
cur              1650 fs/btrfs/ctree.c 			uptodate = btrfs_buffer_uptodate(cur, gen, 0);
cur              1653 fs/btrfs/ctree.c 		if (!cur || !uptodate) {
cur              1654 fs/btrfs/ctree.c 			if (!cur) {
cur              1655 fs/btrfs/ctree.c 				cur = read_tree_block(fs_info, blocknr, gen,
cur              1658 fs/btrfs/ctree.c 				if (IS_ERR(cur)) {
cur              1659 fs/btrfs/ctree.c 					return PTR_ERR(cur);
cur              1660 fs/btrfs/ctree.c 				} else if (!extent_buffer_uptodate(cur)) {
cur              1661 fs/btrfs/ctree.c 					free_extent_buffer(cur);
cur              1665 fs/btrfs/ctree.c 				err = btrfs_read_buffer(cur, gen,
cur              1668 fs/btrfs/ctree.c 					free_extent_buffer(cur);
cur              1676 fs/btrfs/ctree.c 		btrfs_tree_lock(cur);
cur              1677 fs/btrfs/ctree.c 		btrfs_set_lock_blocking_write(cur);
cur              1678 fs/btrfs/ctree.c 		err = __btrfs_cow_block(trans, root, cur, parent, i,
cur              1679 fs/btrfs/ctree.c 					&cur, search_start,
cur              1683 fs/btrfs/ctree.c 			btrfs_tree_unlock(cur);
cur              1684 fs/btrfs/ctree.c 			free_extent_buffer(cur);
cur              1687 fs/btrfs/ctree.c 		search_start = cur->start;
cur              1688 fs/btrfs/ctree.c 		last_block = cur->start;
cur              1690 fs/btrfs/ctree.c 		btrfs_tree_unlock(cur);
cur              1691 fs/btrfs/ctree.c 		free_extent_buffer(cur);
cur              5157 fs/btrfs/ctree.c 	struct extent_buffer *cur;
cur              5168 fs/btrfs/ctree.c 	cur = btrfs_read_lock_root_node(root);
cur              5169 fs/btrfs/ctree.c 	level = btrfs_header_level(cur);
cur              5171 fs/btrfs/ctree.c 	path->nodes[level] = cur;
cur              5174 fs/btrfs/ctree.c 	if (btrfs_header_generation(cur) < min_trans) {
cur              5179 fs/btrfs/ctree.c 		nritems = btrfs_header_nritems(cur);
cur              5180 fs/btrfs/ctree.c 		level = btrfs_header_level(cur);
cur              5181 fs/btrfs/ctree.c 		sret = btrfs_bin_search(cur, min_key, level, &slot);
cur              5193 fs/btrfs/ctree.c 			btrfs_item_key_to_cpu(cur, &found_key, slot);
cur              5205 fs/btrfs/ctree.c 			gen = btrfs_node_ptr_generation(cur, slot);
cur              5230 fs/btrfs/ctree.c 		btrfs_node_key_to_cpu(cur, &found_key, slot);
cur              5237 fs/btrfs/ctree.c 		cur = btrfs_read_node_slot(cur, slot);
cur              5238 fs/btrfs/ctree.c 		if (IS_ERR(cur)) {
cur              5239 fs/btrfs/ctree.c 			ret = PTR_ERR(cur);
cur              5243 fs/btrfs/ctree.c 		btrfs_tree_read_lock(cur);
cur              5246 fs/btrfs/ctree.c 		path->nodes[level - 1] = cur;
cur               378 fs/btrfs/dir-item.c 	u32 cur = 0;
cur               386 fs/btrfs/dir-item.c 	while (cur < total_len) {
cur               396 fs/btrfs/dir-item.c 		cur += this_len;
cur              1725 fs/btrfs/disk-io.c 	struct btrfs_transaction *cur;
cur              1737 fs/btrfs/disk-io.c 		cur = fs_info->running_transaction;
cur              1738 fs/btrfs/disk-io.c 		if (!cur) {
cur              1744 fs/btrfs/disk-io.c 		if (cur->state < TRANS_STATE_BLOCKED &&
cur              1746 fs/btrfs/disk-io.c 		    (now < cur->start_time ||
cur              1747 fs/btrfs/disk-io.c 		     now - cur->start_time < fs_info->commit_interval)) {
cur              1752 fs/btrfs/disk-io.c 		transid = cur->transid;
cur              1793 fs/btrfs/disk-io.c 	u64 cur;
cur              1800 fs/btrfs/disk-io.c 		cur = btrfs_backup_tree_root_gen(root_backup);
cur              1801 fs/btrfs/disk-io.c 		if (cur == newest_gen)
cur              1808 fs/btrfs/disk-io.c 		cur = btrfs_backup_tree_root_gen(root_backup);
cur              1809 fs/btrfs/disk-io.c 		if (cur == newest_gen)
cur              3066 fs/btrfs/extent_io.c 	u64 cur = start;
cur              3103 fs/btrfs/extent_io.c 	while (cur <= end) {
cur              3107 fs/btrfs/extent_io.c 		if (cur >= last_byte) {
cur              3116 fs/btrfs/extent_io.c 			set_extent_uptodate(tree, cur, cur + iosize - 1,
cur              3118 fs/btrfs/extent_io.c 			unlock_extent_cached(tree, cur,
cur              3119 fs/btrfs/extent_io.c 					     cur + iosize - 1, &cached);
cur              3122 fs/btrfs/extent_io.c 		em = __get_extent_map(inode, page, pg_offset, cur,
cur              3123 fs/btrfs/extent_io.c 				      end - cur + 1, get_extent, em_cached);
cur              3126 fs/btrfs/extent_io.c 			unlock_extent(tree, cur, end);
cur              3129 fs/btrfs/extent_io.c 		extent_offset = cur - em->start;
cur              3130 fs/btrfs/extent_io.c 		BUG_ON(extent_map_end(em) <= cur);
cur              3131 fs/btrfs/extent_io.c 		BUG_ON(end < cur);
cur              3139 fs/btrfs/extent_io.c 		iosize = min(extent_map_end(em) - cur, end - cur + 1);
cur              3209 fs/btrfs/extent_io.c 			set_extent_uptodate(tree, cur, cur + iosize - 1,
cur              3211 fs/btrfs/extent_io.c 			unlock_extent_cached(tree, cur,
cur              3212 fs/btrfs/extent_io.c 					     cur + iosize - 1, &cached);
cur              3213 fs/btrfs/extent_io.c 			cur = cur + iosize;
cur              3218 fs/btrfs/extent_io.c 		if (test_range_bit(tree, cur, cur_end,
cur              3221 fs/btrfs/extent_io.c 			unlock_extent(tree, cur, cur + iosize - 1);
cur              3222 fs/btrfs/extent_io.c 			cur = cur + iosize;
cur              3231 fs/btrfs/extent_io.c 			unlock_extent(tree, cur, cur + iosize - 1);
cur              3232 fs/btrfs/extent_io.c 			cur = cur + iosize;
cur              3249 fs/btrfs/extent_io.c 			unlock_extent(tree, cur, cur + iosize - 1);
cur              3252 fs/btrfs/extent_io.c 		cur = cur + iosize;
cur              3422 fs/btrfs/extent_io.c 	u64 cur = start;
cur              3461 fs/btrfs/extent_io.c 	while (cur <= end) {
cur              3465 fs/btrfs/extent_io.c 		if (cur >= i_size) {
cur              3466 fs/btrfs/extent_io.c 			btrfs_writepage_endio_finish_ordered(page, cur,
cur              3470 fs/btrfs/extent_io.c 		em = btrfs_get_extent(BTRFS_I(inode), page, pg_offset, cur,
cur              3471 fs/btrfs/extent_io.c 				     end - cur + 1, 1);
cur              3478 fs/btrfs/extent_io.c 		extent_offset = cur - em->start;
cur              3480 fs/btrfs/extent_io.c 		BUG_ON(em_end <= cur);
cur              3481 fs/btrfs/extent_io.c 		BUG_ON(end < cur);
cur              3482 fs/btrfs/extent_io.c 		iosize = min(em_end - cur, end - cur + 1);
cur              3502 fs/btrfs/extent_io.c 				btrfs_writepage_endio_finish_ordered(page, cur,
cur              3503 fs/btrfs/extent_io.c 							    cur + iosize - 1,
cur              3513 fs/btrfs/extent_io.c 			cur += iosize;
cur              3518 fs/btrfs/extent_io.c 		btrfs_set_range_writeback(tree, cur, cur + iosize - 1);
cur              3522 fs/btrfs/extent_io.c 			       page->index, cur, end);
cur              3536 fs/btrfs/extent_io.c 		cur = cur + iosize;
cur              5558 fs/btrfs/extent_io.c 	size_t cur;
cur              5578 fs/btrfs/extent_io.c 		cur = min(len, (PAGE_SIZE - offset));
cur              5580 fs/btrfs/extent_io.c 		memcpy(dst, kaddr + offset, cur);
cur              5582 fs/btrfs/extent_io.c 		dst += cur;
cur              5583 fs/btrfs/extent_io.c 		len -= cur;
cur              5593 fs/btrfs/extent_io.c 	size_t cur;
cur              5610 fs/btrfs/extent_io.c 		cur = min(len, (PAGE_SIZE - offset));
cur              5612 fs/btrfs/extent_io.c 		if (copy_to_user(dst, kaddr + offset, cur)) {
cur              5617 fs/btrfs/extent_io.c 		dst += cur;
cur              5618 fs/btrfs/extent_io.c 		len -= cur;
cur              5671 fs/btrfs/extent_io.c 	size_t cur;
cur              5688 fs/btrfs/extent_io.c 		cur = min(len, (PAGE_SIZE - offset));
cur              5691 fs/btrfs/extent_io.c 		ret = memcmp(ptr, kaddr + offset, cur);
cur              5695 fs/btrfs/extent_io.c 		ptr += cur;
cur              5696 fs/btrfs/extent_io.c 		len -= cur;
cur              5727 fs/btrfs/extent_io.c 	size_t cur;
cur              5744 fs/btrfs/extent_io.c 		cur = min(len, PAGE_SIZE - offset);
cur              5746 fs/btrfs/extent_io.c 		memcpy(kaddr + offset, src, cur);
cur              5748 fs/btrfs/extent_io.c 		src += cur;
cur              5749 fs/btrfs/extent_io.c 		len -= cur;
cur              5758 fs/btrfs/extent_io.c 	size_t cur;
cur              5774 fs/btrfs/extent_io.c 		cur = min(len, PAGE_SIZE - offset);
cur              5776 fs/btrfs/extent_io.c 		memset(kaddr + offset, 0, cur);
cur              5778 fs/btrfs/extent_io.c 		len -= cur;
cur              5803 fs/btrfs/extent_io.c 	size_t cur;
cur              5818 fs/btrfs/extent_io.c 		cur = min(len, (unsigned long)(PAGE_SIZE - offset));
cur              5821 fs/btrfs/extent_io.c 		read_extent_buffer(src, kaddr + offset, src_offset, cur);
cur              5823 fs/btrfs/extent_io.c 		src_offset += cur;
cur              5824 fs/btrfs/extent_io.c 		len -= cur;
cur              5999 fs/btrfs/extent_io.c 	size_t cur;
cur              6026 fs/btrfs/extent_io.c 		cur = min(len, (unsigned long)(PAGE_SIZE -
cur              6028 fs/btrfs/extent_io.c 		cur = min_t(unsigned long, cur,
cur              6032 fs/btrfs/extent_io.c 			   dst_off_in_page, src_off_in_page, cur);
cur              6034 fs/btrfs/extent_io.c 		src_offset += cur;
cur              6035 fs/btrfs/extent_io.c 		dst_offset += cur;
cur              6036 fs/btrfs/extent_io.c 		len -= cur;
cur              6044 fs/btrfs/extent_io.c 	size_t cur;
cur              6076 fs/btrfs/extent_io.c 		cur = min_t(unsigned long, len, src_off_in_page + 1);
cur              6077 fs/btrfs/extent_io.c 		cur = min(cur, dst_off_in_page + 1);
cur              6079 fs/btrfs/extent_io.c 			   dst_off_in_page - cur + 1,
cur              6080 fs/btrfs/extent_io.c 			   src_off_in_page - cur + 1, cur);
cur              6082 fs/btrfs/extent_io.c 		dst_end -= cur;
cur              6083 fs/btrfs/extent_io.c 		src_end -= cur;
cur              6084 fs/btrfs/extent_io.c 		len -= cur;
cur               496 fs/btrfs/extent_map.c 			    struct extent_map *cur,
cur               500 fs/btrfs/extent_map.c 	WARN_ON(test_bit(EXTENT_FLAG_PINNED, &cur->flags));
cur               501 fs/btrfs/extent_map.c 	ASSERT(extent_map_in_tree(cur));
cur               502 fs/btrfs/extent_map.c 	if (!test_bit(EXTENT_FLAG_LOGGING, &cur->flags))
cur               503 fs/btrfs/extent_map.c 		list_del_init(&cur->list);
cur               504 fs/btrfs/extent_map.c 	rb_replace_node_cached(&cur->rb_node, &new->rb_node, &tree->map);
cur               505 fs/btrfs/extent_map.c 	RB_CLEAR_NODE(&cur->rb_node);
cur                91 fs/btrfs/extent_map.h 			    struct extent_map *cur,
cur               339 fs/btrfs/free-space-cache.c 	if (io_ctl->cur) {
cur               340 fs/btrfs/free-space-cache.c 		io_ctl->cur = NULL;
cur               349 fs/btrfs/free-space-cache.c 	io_ctl->cur = page_address(io_ctl->page);
cur               350 fs/btrfs/free-space-cache.c 	io_ctl->orig = io_ctl->cur;
cur               353 fs/btrfs/free-space-cache.c 		clear_page(io_ctl->cur);
cur               422 fs/btrfs/free-space-cache.c 		io_ctl->cur += (sizeof(u32) * io_ctl->num_pages);
cur               425 fs/btrfs/free-space-cache.c 		io_ctl->cur += sizeof(u64);
cur               429 fs/btrfs/free-space-cache.c 	val = io_ctl->cur;
cur               431 fs/btrfs/free-space-cache.c 	io_ctl->cur += sizeof(u64);
cur               443 fs/btrfs/free-space-cache.c 		io_ctl->cur += sizeof(u32) * io_ctl->num_pages;
cur               447 fs/btrfs/free-space-cache.c 		io_ctl->cur += sizeof(u64);
cur               451 fs/btrfs/free-space-cache.c 	gen = io_ctl->cur;
cur               459 fs/btrfs/free-space-cache.c 	io_ctl->cur += sizeof(u64);
cur               521 fs/btrfs/free-space-cache.c 	if (!io_ctl->cur)
cur               524 fs/btrfs/free-space-cache.c 	entry = io_ctl->cur;
cur               529 fs/btrfs/free-space-cache.c 	io_ctl->cur += sizeof(struct btrfs_free_space_entry);
cur               548 fs/btrfs/free-space-cache.c 	if (!io_ctl->cur)
cur               555 fs/btrfs/free-space-cache.c 	if (io_ctl->cur != io_ctl->orig) {
cur               562 fs/btrfs/free-space-cache.c 	copy_page(io_ctl->cur, bitmap);
cur               575 fs/btrfs/free-space-cache.c 	if (io_ctl->cur != io_ctl->orig)
cur               592 fs/btrfs/free-space-cache.c 	if (!io_ctl->cur) {
cur               598 fs/btrfs/free-space-cache.c 	e = io_ctl->cur;
cur               602 fs/btrfs/free-space-cache.c 	io_ctl->cur += sizeof(struct btrfs_free_space_entry);
cur               622 fs/btrfs/free-space-cache.c 	copy_page(entry->bitmap, io_ctl->cur);
cur                40 fs/btrfs/free-space-cache.h 	void *cur, *orig;
cur               434 fs/btrfs/ordered-data.c 	struct list_head *cur;
cur               447 fs/btrfs/ordered-data.c 			cur = entry->list.next;
cur               448 fs/btrfs/ordered-data.c 			sum = list_entry(cur, struct btrfs_ordered_sum, list);
cur               133 fs/btrfs/props.c 		u32 total_len, cur, this_len;
cur               160 fs/btrfs/props.c 		cur = 0;
cur               163 fs/btrfs/props.c 		while (cur < total_len) {
cur               208 fs/btrfs/props.c 			cur += this_len;
cur              3847 fs/btrfs/qgroup.c 	struct rb_node **cur;
cur              3897 fs/btrfs/qgroup.c 	cur = &blocks->blocks[level].rb_node;
cur              3898 fs/btrfs/qgroup.c 	while (*cur) {
cur              3901 fs/btrfs/qgroup.c 		parent = *cur;
cur              3906 fs/btrfs/qgroup.c 			cur = &(*cur)->rb_left;
cur              3908 fs/btrfs/qgroup.c 			cur = &(*cur)->rb_right;
cur              3929 fs/btrfs/qgroup.c 	rb_link_node(&block->node, parent, cur);
cur               205 fs/btrfs/raid56.c 	struct btrfs_stripe_hash *cur;
cur               232 fs/btrfs/raid56.c 		cur = h + i;
cur               233 fs/btrfs/raid56.c 		INIT_LIST_HEAD(&cur->hash_list);
cur               234 fs/btrfs/raid56.c 		spin_lock_init(&cur->lock);
cur               555 fs/btrfs/raid56.c 			  struct btrfs_raid_bio *cur)
cur               558 fs/btrfs/raid56.c 	    test_bit(RBIO_RMW_LOCKED_BIT, &cur->flags))
cur               569 fs/btrfs/raid56.c 	    test_bit(RBIO_CACHE_BIT, &cur->flags))
cur               573 fs/btrfs/raid56.c 	    cur->bbio->raid_map[0])
cur               577 fs/btrfs/raid56.c 	if (last->operation != cur->operation)
cur               596 fs/btrfs/raid56.c 		int cur_fa = cur->faila;
cur               597 fs/btrfs/raid56.c 		int cur_fb = cur->failb;
cur               604 fs/btrfs/raid56.c 		if (cur->faila >= cur->failb) {
cur               605 fs/btrfs/raid56.c 			cur_fa = cur->failb;
cur               606 fs/btrfs/raid56.c 			cur_fb = cur->faila;
cur               676 fs/btrfs/raid56.c 	struct btrfs_raid_bio *cur;
cur               684 fs/btrfs/raid56.c 	list_for_each_entry(cur, &h->hash_list, hash_list) {
cur               685 fs/btrfs/raid56.c 		if (cur->bbio->raid_map[0] == rbio->bbio->raid_map[0]) {
cur               686 fs/btrfs/raid56.c 			spin_lock(&cur->bio_list_lock);
cur               689 fs/btrfs/raid56.c 			if (bio_list_empty(&cur->bio_list) &&
cur               690 fs/btrfs/raid56.c 			    list_empty(&cur->plug_list) &&
cur               691 fs/btrfs/raid56.c 			    test_bit(RBIO_CACHE_BIT, &cur->flags) &&
cur               692 fs/btrfs/raid56.c 			    !test_bit(RBIO_RMW_LOCKED_BIT, &cur->flags)) {
cur               693 fs/btrfs/raid56.c 				list_del_init(&cur->hash_list);
cur               694 fs/btrfs/raid56.c 				refcount_dec(&cur->refs);
cur               696 fs/btrfs/raid56.c 				steal_rbio(cur, rbio);
cur               697 fs/btrfs/raid56.c 				cache_drop = cur;
cur               698 fs/btrfs/raid56.c 				spin_unlock(&cur->bio_list_lock);
cur               704 fs/btrfs/raid56.c 			if (rbio_can_merge(cur, rbio)) {
cur               705 fs/btrfs/raid56.c 				merge_rbio(cur, rbio);
cur               706 fs/btrfs/raid56.c 				spin_unlock(&cur->bio_list_lock);
cur               721 fs/btrfs/raid56.c 			list_for_each_entry(pending, &cur->plug_list,
cur               725 fs/btrfs/raid56.c 					spin_unlock(&cur->bio_list_lock);
cur               736 fs/btrfs/raid56.c 			list_add_tail(&rbio->plug_list, &cur->plug_list);
cur               737 fs/btrfs/raid56.c 			spin_unlock(&cur->bio_list_lock);
cur               857 fs/btrfs/raid56.c static void rbio_endio_bio_list(struct bio *cur, blk_status_t err)
cur               861 fs/btrfs/raid56.c 	while (cur) {
cur               862 fs/btrfs/raid56.c 		next = cur->bi_next;
cur               863 fs/btrfs/raid56.c 		cur->bi_next = NULL;
cur               864 fs/btrfs/raid56.c 		cur->bi_status = err;
cur               865 fs/btrfs/raid56.c 		bio_endio(cur);
cur               866 fs/btrfs/raid56.c 		cur = next;
cur               876 fs/btrfs/raid56.c 	struct bio *cur = bio_list_get(&rbio->bio_list);
cur               894 fs/btrfs/raid56.c 	rbio_endio_bio_list(cur, err);
cur              1690 fs/btrfs/raid56.c 	struct btrfs_raid_bio *cur;
cur              1700 fs/btrfs/raid56.c 		cur = list_entry(plug->rbio_list.next,
cur              1702 fs/btrfs/raid56.c 		list_del_init(&cur->plug_list);
cur              1704 fs/btrfs/raid56.c 		if (rbio_is_full(cur)) {
cur              1708 fs/btrfs/raid56.c 			ret = full_stripe_write(cur);
cur              1713 fs/btrfs/raid56.c 			if (rbio_can_merge(last, cur)) {
cur              1714 fs/btrfs/raid56.c 				merge_rbio(last, cur);
cur              1715 fs/btrfs/raid56.c 				__free_raid_bio(cur);
cur              1721 fs/btrfs/raid56.c 		last = cur;
cur               691 fs/btrfs/relocation.c 	struct backref_node *cur;
cur               726 fs/btrfs/relocation.c 	cur = node;
cur               730 fs/btrfs/relocation.c 	key.objectid = cur->bytenr;
cur               747 fs/btrfs/relocation.c 	WARN_ON(cur->checked);
cur               748 fs/btrfs/relocation.c 	if (!list_empty(&cur->upper)) {
cur               753 fs/btrfs/relocation.c 		ASSERT(list_is_singular(&cur->upper));
cur               754 fs/btrfs/relocation.c 		edge = list_entry(cur->upper.next, struct backref_edge,
cur               785 fs/btrfs/relocation.c 			if (key.objectid != cur->bytenr) {
cur               837 fs/btrfs/relocation.c 				root = find_reloc_root(rc, cur->bytenr);
cur               839 fs/btrfs/relocation.c 				cur->root = root;
cur               857 fs/btrfs/relocation.c 				upper->level = cur->level + 1;
cur               869 fs/btrfs/relocation.c 			list_add_tail(&edge->list[LOWER], &cur->upper);
cur               870 fs/btrfs/relocation.c 			edge->node[LOWER] = cur;
cur               896 fs/btrfs/relocation.c 			cur->cowonly = 1;
cur               898 fs/btrfs/relocation.c 		if (btrfs_root_level(&root->root_item) == cur->level) {
cur               901 fs/btrfs/relocation.c 			       cur->bytenr);
cur               903 fs/btrfs/relocation.c 				list_add(&cur->list, &useless);
cur               905 fs/btrfs/relocation.c 				cur->root = root;
cur               909 fs/btrfs/relocation.c 		level = cur->level + 1;
cur               926 fs/btrfs/relocation.c 		    cur->bytenr) {
cur               929 fs/btrfs/relocation.c 				  cur->bytenr, level - 1,
cur               936 fs/btrfs/relocation.c 		lower = cur;
cur              1029 fs/btrfs/relocation.c 	cur->checked = 1;
cur              1036 fs/btrfs/relocation.c 		cur = edge->node[UPPER];
cur               899 fs/btrfs/send.c 	u32 cur = 0;
cur               935 fs/btrfs/send.c 	while (cur < total) {
cur               939 fs/btrfs/send.c 			iref = (struct btrfs_inode_ref *)(ptr + cur);
cur               945 fs/btrfs/send.c 			extref = (struct btrfs_inode_extref *)(ptr + cur);
cur               984 fs/btrfs/send.c 		cur += elem_size + name_len;
cur              1021 fs/btrfs/send.c 	u32 cur;
cur              1045 fs/btrfs/send.c 	cur = 0;
cur              1050 fs/btrfs/send.c 	while (cur < total) {
cur              1103 fs/btrfs/send.c 		cur += len;
cur              2079 fs/btrfs/send.c 	struct name_cache_entry *cur;
cur              2085 fs/btrfs/send.c 	list_for_each_entry(cur, nce_head, radix_list) {
cur              2086 fs/btrfs/send.c 		if (cur->ino == ino && cur->gen == gen)
cur              2087 fs/btrfs/send.c 			return cur;
cur              2816 fs/btrfs/send.c 	struct recorded_ref *cur;
cur              2819 fs/btrfs/send.c 		cur = list_entry(head->next, struct recorded_ref, list);
cur              2820 fs/btrfs/send.c 		fs_path_free(cur->full_path);
cur              2821 fs/btrfs/send.c 		list_del(&cur->list);
cur              2822 fs/btrfs/send.c 		kfree(cur);
cur              3096 fs/btrfs/send.c 	struct recorded_ref *cur;
cur              3123 fs/btrfs/send.c 	list_for_each_entry(cur, deleted_refs, list) {
cur              3124 fs/btrfs/send.c 		ret = dup_ref(cur, &pm->update_refs);
cur              3128 fs/btrfs/send.c 	list_for_each_entry(cur, new_refs, list) {
cur              3129 fs/btrfs/send.c 		ret = dup_ref(cur, &pm->update_refs);
cur              3219 fs/btrfs/send.c 	struct recorded_ref *cur;
cur              3326 fs/btrfs/send.c 	list_for_each_entry(cur, &pm->update_refs, list) {
cur              3330 fs/btrfs/send.c 		ret = get_inode_info(sctx->send_root, cur->dir, NULL,
cur              3339 fs/btrfs/send.c 		ret = send_utimes(sctx, cur->dir, cur->dir_gen);
cur              3812 fs/btrfs/send.c 	struct recorded_ref *cur;
cur              3873 fs/btrfs/send.c 	list_for_each_entry(cur, &sctx->new_refs, list) {
cur              3881 fs/btrfs/send.c 		ret = get_cur_inode_state(sctx, cur->dir, cur->dir_gen);
cur              3891 fs/btrfs/send.c 				if (cur == cur2)
cur              3893 fs/btrfs/send.c 				if (cur2->dir == cur->dir) {
cur              3904 fs/btrfs/send.c 				ret = did_create_dir(sctx, cur->dir);
cur              3908 fs/btrfs/send.c 				ret = send_create_inode(sctx, cur->dir);
cur              3920 fs/btrfs/send.c 		ret = will_overwrite_ref(sctx, cur->dir, cur->dir_gen,
cur              3921 fs/btrfs/send.c 				cur->name, cur->name_len,
cur              3927 fs/btrfs/send.c 					   ow_inode, cur->dir, cur->name,
cur              3928 fs/btrfs/send.c 					   cur->name_len);
cur              3936 fs/btrfs/send.c 						cur->full_path);
cur              3991 fs/btrfs/send.c 				ret = send_unlink(sctx, cur->full_path);
cur              3998 fs/btrfs/send.c 			ret = wait_for_dest_dir_move(sctx, cur, is_orphan);
cur              4009 fs/btrfs/send.c 			ret = wait_for_parent_move(sctx, cur, is_orphan);
cur              4024 fs/btrfs/send.c 			ret = send_rename(sctx, valid_path, cur->full_path);
cur              4028 fs/btrfs/send.c 			ret = fs_path_copy(valid_path, cur->full_path);
cur              4039 fs/btrfs/send.c 						  cur->full_path);
cur              4042 fs/btrfs/send.c 							   cur->full_path);
cur              4054 fs/btrfs/send.c 					ret = update_ref_path(sctx, cur);
cur              4058 fs/btrfs/send.c 				ret = send_link(sctx, cur->full_path,
cur              4064 fs/btrfs/send.c 		ret = dup_ref(cur, &check_dirs);
cur              4092 fs/btrfs/send.c 		list_for_each_entry(cur, &sctx->deleted_refs, list) {
cur              4093 fs/btrfs/send.c 			ret = dup_ref(cur, &check_dirs);
cur              4102 fs/btrfs/send.c 		cur = list_entry(sctx->deleted_refs.next, struct recorded_ref,
cur              4104 fs/btrfs/send.c 		ret = dup_ref(cur, &check_dirs);
cur              4113 fs/btrfs/send.c 		list_for_each_entry(cur, &sctx->deleted_refs, list) {
cur              4114 fs/btrfs/send.c 			ret = did_overwrite_ref(sctx, cur->dir, cur->dir_gen,
cur              4116 fs/btrfs/send.c 					cur->name, cur->name_len);
cur              4128 fs/btrfs/send.c 					ret = update_ref_path(sctx, cur);
cur              4132 fs/btrfs/send.c 				ret = send_unlink(sctx, cur->full_path);
cur              4136 fs/btrfs/send.c 			ret = dup_ref(cur, &check_dirs);
cur              4161 fs/btrfs/send.c 	list_for_each_entry(cur, &check_dirs, list) {
cur              4167 fs/btrfs/send.c 		if (cur->dir > sctx->cur_ino)
cur              4170 fs/btrfs/send.c 		ret = get_cur_inode_state(sctx, cur->dir, cur->dir_gen);
cur              4177 fs/btrfs/send.c 			ret = send_utimes(sctx, cur->dir, cur->dir_gen);
cur              4181 fs/btrfs/send.c 			   cur->dir != last_dir_ino_rm) {
cur              4182 fs/btrfs/send.c 			ret = can_rmdir(sctx, cur->dir, cur->dir_gen,
cur              4187 fs/btrfs/send.c 				ret = get_cur_path(sctx, cur->dir,
cur              4188 fs/btrfs/send.c 						   cur->dir_gen, valid_path);
cur              4194 fs/btrfs/send.c 				last_dir_ino_rm = cur->dir;
cur               321 fs/btrfs/tree-checker.c 	u32 cur = 0;
cur               324 fs/btrfs/tree-checker.c 	while (cur < item_size) {
cur               333 fs/btrfs/tree-checker.c 		if (cur + sizeof(*di) > item_size) {
cur               336 fs/btrfs/tree-checker.c 				cur + sizeof(*di), item_size);
cur               394 fs/btrfs/tree-checker.c 		if (cur + total_size > item_size) {
cur               397 fs/btrfs/tree-checker.c 				cur + total_size, item_size);
cur               419 fs/btrfs/tree-checker.c 		cur += total_size;
cur              2348 fs/btrfs/tree-log.c 		u32 cur;
cur              2358 fs/btrfs/tree-log.c 		cur = 0;
cur              2359 fs/btrfs/tree-log.c 		while (cur < total_size) {
cur              2400 fs/btrfs/tree-log.c 			cur += this_len;
cur              2690 fs/btrfs/tree-log.c 	struct extent_buffer *cur;
cur              2703 fs/btrfs/tree-log.c 		cur = path->nodes[*level];
cur              2705 fs/btrfs/tree-log.c 		WARN_ON(btrfs_header_level(cur) != *level);
cur              2708 fs/btrfs/tree-log.c 		    btrfs_header_nritems(cur))
cur              2711 fs/btrfs/tree-log.c 		bytenr = btrfs_node_blockptr(cur, path->slots[*level]);
cur              2712 fs/btrfs/tree-log.c 		ptr_gen = btrfs_node_ptr_generation(cur, path->slots[*level]);
cur              2713 fs/btrfs/tree-log.c 		btrfs_node_key_to_cpu(cur, &first_key, path->slots[*level]);
cur               148 fs/btrfs/ulist.c 	struct ulist_node *cur = NULL;
cur               152 fs/btrfs/ulist.c 		cur = rb_entry(parent, struct ulist_node, rb_node);
cur               154 fs/btrfs/ulist.c 		if (cur->val < ins->val)
cur               156 fs/btrfs/ulist.c 		else if (cur->val > ins->val)
cur               536 fs/btrfs/volumes.c 	struct bio *cur;
cur               616 fs/btrfs/volumes.c 		cur = pending;
cur               618 fs/btrfs/volumes.c 		cur->bi_next = NULL;
cur               620 fs/btrfs/volumes.c 		BUG_ON(atomic_read(&cur->__bi_cnt) == 0);
cur               638 fs/btrfs/volumes.c 		btrfsic_submit_bio(cur);
cur              2939 fs/btrfs/volumes.c 	u32 cur;
cur              2946 fs/btrfs/volumes.c 	cur = 0;
cur              2948 fs/btrfs/volumes.c 	while (cur < array_size) {
cur              2964 fs/btrfs/volumes.c 			memmove(ptr, ptr + len, array_size - (cur + len));
cur              2969 fs/btrfs/volumes.c 			cur += len;
cur               283 fs/btrfs/xattr.c 		u32 cur;
cur               314 fs/btrfs/xattr.c 		cur = 0;
cur               315 fs/btrfs/xattr.c 		while (cur < item_size) {
cur               340 fs/btrfs/xattr.c 			cur += this_len;
cur               454 fs/cifs/cifsfs.c cifs_show_nls(struct seq_file *s, struct nls_table *cur)
cur               460 fs/cifs/cifsfs.c 	if (def != cur)
cur               461 fs/cifs/cifsfs.c 		seq_printf(s, ",iocharset=%s", cur->charset);
cur               270 fs/cifs/file.c 	struct cifs_fid_locks *cur;
cur               274 fs/cifs/file.c 	list_for_each_entry(cur, &cinode->llist, llist) {
cur               275 fs/cifs/file.c 		if (!list_empty(&cur->locks)) {
cur              1012 fs/cifs/file.c 	struct cifs_fid_locks *cur;
cur              1015 fs/cifs/file.c 	list_for_each_entry(cur, &cinode->llist, llist) {
cur              1016 fs/cifs/file.c 		rc = cifs_find_fid_lock_conflict(cur, offset, length, type,
cur              1193 fs/cifs/file.c 	LOCKING_ANDX_RANGE *buf, *cur;
cur              1226 fs/cifs/file.c 		cur = buf;
cur              1231 fs/cifs/file.c 			cur->Pid = cpu_to_le16(li->pid);
cur              1232 fs/cifs/file.c 			cur->LengthLow = cpu_to_le32((u32)li->length);
cur              1233 fs/cifs/file.c 			cur->LengthHigh = cpu_to_le32((u32)(li->length>>32));
cur              1234 fs/cifs/file.c 			cur->OffsetLow = cpu_to_le32((u32)li->offset);
cur              1235 fs/cifs/file.c 			cur->OffsetHigh = cpu_to_le32((u32)(li->offset>>32));
cur              1243 fs/cifs/file.c 				cur = buf;
cur              1246 fs/cifs/file.c 				cur++;
cur              1538 fs/cifs/file.c 	LOCKING_ANDX_RANGE *buf, *cur;
cur              1567 fs/cifs/file.c 		cur = buf;
cur              1588 fs/cifs/file.c 			cur->Pid = cpu_to_le16(li->pid);
cur              1589 fs/cifs/file.c 			cur->LengthLow = cpu_to_le32((u32)li->length);
cur              1590 fs/cifs/file.c 			cur->LengthHigh = cpu_to_le32((u32)(li->length>>32));
cur              1591 fs/cifs/file.c 			cur->OffsetLow = cpu_to_le32((u32)li->offset);
cur              1592 fs/cifs/file.c 			cur->OffsetHigh = cpu_to_le32((u32)(li->offset>>32));
cur              1618 fs/cifs/file.c 				cur = buf;
cur              1621 fs/cifs/file.c 				cur++;
cur               122 fs/cifs/smb2file.c 	struct smb2_lock_element *buf, *cur;
cur               146 fs/cifs/smb2file.c 	cur = buf;
cur               166 fs/cifs/smb2file.c 		cur->Length = cpu_to_le64(li->length);
cur               167 fs/cifs/smb2file.c 		cur->Offset = cpu_to_le64(li->offset);
cur               168 fs/cifs/smb2file.c 		cur->Flags = cpu_to_le32(SMB2_LOCKFLAG_UNLOCK);
cur               194 fs/cifs/smb2file.c 			cur = buf;
cur               197 fs/cifs/smb2file.c 			cur++;
cur               223 fs/cifs/smb2file.c 	struct smb2_lock_element *cur = buf;
cur               227 fs/cifs/smb2file.c 		cur->Length = cpu_to_le64(li->length);
cur               228 fs/cifs/smb2file.c 		cur->Offset = cpu_to_le64(li->offset);
cur               229 fs/cifs/smb2file.c 		cur->Flags = cpu_to_le32(li->type |
cur               238 fs/cifs/smb2file.c 			cur = buf;
cur               241 fs/cifs/smb2file.c 			cur++;
cur                49 fs/configfs/symlink.c 		int cur = strlen(config_item_name(p));
cur                52 fs/configfs/symlink.c 		length -= cur;
cur                53 fs/configfs/symlink.c 		memcpy(buffer + length, config_item_name(p), cur);
cur               121 fs/coredump.c  	int cur = cn->used;
cur               134 fs/coredump.c  		if ((cn->used - cur == 1 && cn->corename[cur] == '.') ||
cur               135 fs/coredump.c  				(cn->used - cur == 2 && cn->corename[cur] == '.'
cur               136 fs/coredump.c  				&& cn->corename[cur+1] == '.'))
cur               137 fs/coredump.c  			cn->corename[cur] = '!';
cur               145 fs/coredump.c  		if (cn->used == cur)
cur               149 fs/coredump.c  	for (; cur < cn->used; ++cur) {
cur               150 fs/coredump.c  		if (cn->corename[cur] == '/')
cur               151 fs/coredump.c  			cn->corename[cur] = '!';
cur               477 fs/dlm/lock.c  		struct dlm_rsb *cur = rb_entry(*newn, struct dlm_rsb,
cur               481 fs/dlm/lock.c  		rc = rsb_cmp(cur, rsb->res_name, rsb->res_length);
cur               489 fs/dlm/lock.c  			dlm_dump_rsb(cur);
cur               204 fs/efs/inode.c 	int cur, last, first = 1;
cur               229 fs/efs/inode.c 			cur = (last + dirext) % in->numextents;
cur               230 fs/efs/inode.c 			if ((result = efs_extent_check(&in->extents[cur], block, sb))) {
cur               231 fs/efs/inode.c 				in->lastextent = cur;
cur               246 fs/efs/inode.c 		cur = (last + indext) % indexts;
cur               256 fs/efs/inode.c 		for(dirext = 0; cur < ibase && dirext < direxts; dirext++) {
cur               264 fs/efs/inode.c 			       cur, block);
cur               271 fs/efs/inode.c 			(cur - ibase) /
cur               273 fs/efs/inode.c 		ioffset = (cur - ibase) %
cur               297 fs/efs/inode.c 			       cur, iblock);
cur               304 fs/efs/inode.c 			in->lastextent = cur;
cur               198 fs/erofs/decompressor.c 	const char *cur = dst - pageofs_out;
cur               200 fs/erofs/decompressor.c 	while (cur < end) {
cur               206 fs/erofs/decompressor.c 			if (cur >= dst) {
cur               207 fs/erofs/decompressor.c 				memcpy(buf, cur, min_t(uint, PAGE_SIZE,
cur               208 fs/erofs/decompressor.c 						       end - cur));
cur               210 fs/erofs/decompressor.c 				memcpy(buf + pageofs_out, cur + pageofs_out,
cur               211 fs/erofs/decompressor.c 				       min_t(uint, righthalf, end - cur));
cur               215 fs/erofs/decompressor.c 		cur += PAGE_SIZE;
cur               582 fs/erofs/zdata.c 	unsigned int cur, end, spiltted, index;
cur               591 fs/erofs/zdata.c 	cur = end - 1;
cur               594 fs/erofs/zdata.c 	if (offset + cur >= map->m_la &&
cur               595 fs/erofs/zdata.c 	    offset + cur < map->m_la + map->m_llen) {
cur               603 fs/erofs/zdata.c 	erofs_dbg("%s: [out-of-range] pos %llu", __func__, offset + cur);
cur               608 fs/erofs/zdata.c 	map->m_la = offset + cur;
cur               641 fs/erofs/zdata.c 	cur = end - min_t(unsigned int, offset + end - map->m_la, end);
cur               643 fs/erofs/zdata.c 		zero_user_segment(page, cur, end);
cur               648 fs/erofs/zdata.c 	page_type = cur ? Z_EROFS_VLE_PAGE_TYPE_HEAD :
cur               653 fs/erofs/zdata.c 	if (cur)
cur               682 fs/erofs/zdata.c 	map->m_llen = offset + cur - map->m_la;
cur               684 fs/erofs/zdata.c 	end = cur;
cur               785 fs/ext2/balloc.c 	ext2_fsblk_t cur;
cur               790 fs/ext2/balloc.c 	cur = start_block;
cur               796 fs/ext2/balloc.c 		if (cur <= rsv->rsv_end)
cur               797 fs/ext2/balloc.c 			cur = rsv->rsv_end + 1;
cur               808 fs/ext2/balloc.c 		if (cur > last_block)
cur               822 fs/ext2/balloc.c 		if (cur + size <= rsv->rsv_start) {
cur               851 fs/ext2/balloc.c 	my_rsv->rsv_start = cur;
cur               852 fs/ext2/balloc.c 	my_rsv->rsv_end = cur + size - 1;
cur              1197 fs/ext4/ext4.h extern void ext4_set_bits(void *bm, int cur, int len);
cur               557 fs/ext4/mballoc.c 	struct list_head *cur;
cur               629 fs/ext4/mballoc.c 	list_for_each(cur, &grp->bb_prealloc_list) {
cur               632 fs/ext4/mballoc.c 		pa = list_entry(cur, struct ext4_prealloc_space, pa_group_list);
cur              1276 fs/ext4/mballoc.c static void mb_clear_bits(void *bm, int cur, int len)
cur              1280 fs/ext4/mballoc.c 	len = cur + len;
cur              1281 fs/ext4/mballoc.c 	while (cur < len) {
cur              1282 fs/ext4/mballoc.c 		if ((cur & 31) == 0 && (len - cur) >= 32) {
cur              1284 fs/ext4/mballoc.c 			addr = bm + (cur >> 3);
cur              1286 fs/ext4/mballoc.c 			cur += 32;
cur              1289 fs/ext4/mballoc.c 		mb_clear_bit(cur, bm);
cur              1290 fs/ext4/mballoc.c 		cur++;
cur              1297 fs/ext4/mballoc.c static int mb_test_and_clear_bits(void *bm, int cur, int len)
cur              1302 fs/ext4/mballoc.c 	len = cur + len;
cur              1303 fs/ext4/mballoc.c 	while (cur < len) {
cur              1304 fs/ext4/mballoc.c 		if ((cur & 31) == 0 && (len - cur) >= 32) {
cur              1306 fs/ext4/mballoc.c 			addr = bm + (cur >> 3);
cur              1308 fs/ext4/mballoc.c 				zero_bit = cur + mb_find_next_zero_bit(addr, 32, 0);
cur              1310 fs/ext4/mballoc.c 			cur += 32;
cur              1313 fs/ext4/mballoc.c 		if (!mb_test_and_clear_bit(cur, bm) && zero_bit == -1)
cur              1314 fs/ext4/mballoc.c 			zero_bit = cur;
cur              1315 fs/ext4/mballoc.c 		cur++;
cur              1321 fs/ext4/mballoc.c void ext4_set_bits(void *bm, int cur, int len)
cur              1325 fs/ext4/mballoc.c 	len = cur + len;
cur              1326 fs/ext4/mballoc.c 	while (cur < len) {
cur              1327 fs/ext4/mballoc.c 		if ((cur & 31) == 0 && (len - cur) >= 32) {
cur              1329 fs/ext4/mballoc.c 			addr = bm + (cur >> 3);
cur              1331 fs/ext4/mballoc.c 			cur += 32;
cur              1334 fs/ext4/mballoc.c 		mb_set_bit(cur, bm);
cur              1335 fs/ext4/mballoc.c 		cur++;
cur              1561 fs/ext4/mballoc.c 	int cur;
cur              1616 fs/ext4/mballoc.c 		cur = (start >> ord) & ~1U;
cur              1618 fs/ext4/mballoc.c 		mb_clear_bit(cur, buddy);
cur              1619 fs/ext4/mballoc.c 		mb_clear_bit(cur + 1, buddy);
cur              2701 fs/ext4/mballoc.c 	struct list_head *cur, *tmp;
cur              2704 fs/ext4/mballoc.c 	list_for_each_safe(cur, tmp, &grp->bb_prealloc_list) {
cur              2705 fs/ext4/mballoc.c 		pa = list_entry(cur, struct ext4_prealloc_space, pa_group_list);
cur              3542 fs/ext4/mballoc.c 	struct list_head *cur;
cur              3556 fs/ext4/mballoc.c 	list_for_each(cur, &grp->bb_prealloc_list) {
cur              3557 fs/ext4/mballoc.c 		pa = list_entry(cur, struct ext4_prealloc_space, pa_group_list);
cur              4151 fs/ext4/mballoc.c 		struct list_head *cur;
cur              4153 fs/ext4/mballoc.c 		list_for_each(cur, &grp->bb_prealloc_list) {
cur              4154 fs/ext4/mballoc.c 			pa = list_entry(cur, struct ext4_prealloc_space,
cur               168 fs/ext4/page-io.c 	struct list_head *cur, *before, *after;
cur               176 fs/ext4/page-io.c 		cur = &io->list;
cur               177 fs/ext4/page-io.c 		before = cur->prev;
cur               179 fs/ext4/page-io.c 		after = cur->next;
cur              3137 fs/ext4/super.c 	unsigned long next_wakeup, cur;
cur              3194 fs/ext4/super.c 		cur = jiffies;
cur              3195 fs/ext4/super.c 		if ((time_after_eq(cur, next_wakeup)) ||
cur              3201 fs/ext4/super.c 		schedule_timeout_interruptible(next_wakeup - cur);
cur              3006 fs/f2fs/data.c 		struct inmem_pages *cur;
cur              3007 fs/f2fs/data.c 		list_for_each_entry(cur, &fi->inmem_pages, list)
cur              3008 fs/f2fs/data.c 			if (cur->page == page) {
cur              3009 fs/f2fs/data.c 				cur->page = newpage;
cur               172 fs/f2fs/extent_cache.c 	struct rb_node *cur = rb_first_cached(root), *next;
cur               175 fs/f2fs/extent_cache.c 	if (!cur)
cur               178 fs/f2fs/extent_cache.c 	while (cur) {
cur               179 fs/f2fs/extent_cache.c 		next = rb_next(cur);
cur               183 fs/f2fs/extent_cache.c 		cur_re = rb_entry(cur, struct rb_entry, rb_node);
cur               193 fs/f2fs/extent_cache.c 		cur = next;
cur               749 fs/f2fs/f2fs.h static inline bool __is_discard_back_mergeable(struct discard_info *cur,
cur               752 fs/f2fs/f2fs.h 	return __is_discard_mergeable(back, cur, max_len);
cur               755 fs/f2fs/f2fs.h static inline bool __is_discard_front_mergeable(struct discard_info *cur,
cur               758 fs/f2fs/f2fs.h 	return __is_discard_mergeable(cur, front, max_len);
cur               768 fs/f2fs/f2fs.h static inline bool __is_back_mergeable(struct extent_info *cur,
cur               771 fs/f2fs/f2fs.h 	return __is_extent_mergeable(back, cur);
cur               774 fs/f2fs/f2fs.h static inline bool __is_front_mergeable(struct extent_info *cur,
cur               777 fs/f2fs/f2fs.h 	return __is_extent_mergeable(cur, front);
cur              3381 fs/f2fs/f2fs.h 		int cur = atomic_read(&F2FS_I_SB(inode)->aw_cnt);	\
cur              3383 fs/f2fs/f2fs.h 		if (cur > max)						\
cur              3384 fs/f2fs/f2fs.h 			atomic_set(&F2FS_I_SB(inode)->max_aw_cnt, cur);	\
cur              3392 fs/f2fs/f2fs.h 		int cur = atomic_read(&F2FS_I_SB(inode)->vw_cnt);	\
cur              3394 fs/f2fs/f2fs.h 		if (cur > max)						\
cur              3395 fs/f2fs/f2fs.h 			atomic_set(&F2FS_I_SB(inode)->max_vw_cnt, cur);	\
cur              2742 fs/f2fs/node.c 	struct nat_entry_set *cur;
cur              2747 fs/f2fs/node.c 	list_for_each_entry(cur, head, set_list) {
cur              2748 fs/f2fs/node.c 		if (cur->entry_cnt >= nes->entry_cnt) {
cur              2749 fs/f2fs/node.c 			list_add(&nes->set_list, cur->set_list.prev);
cur              2798 fs/f2fs/node.c 	struct nat_entry *ne, *cur;
cur              2822 fs/f2fs/node.c 	list_for_each_entry_safe(ne, cur, &set->entry_list, list) {
cur               215 fs/f2fs/segment.c 	struct inmem_pages *cur, *tmp;
cur               218 fs/f2fs/segment.c 	list_for_each_entry_safe(cur, tmp, head, list) {
cur               219 fs/f2fs/segment.c 		struct page *page = cur->page;
cur               262 fs/f2fs/segment.c 			if (cur->old_addr == NEW_ADDR) {
cur               267 fs/f2fs/segment.c 					cur->old_addr, ni.version, true, true);
cur               279 fs/f2fs/segment.c 		list_del(&cur->list);
cur               280 fs/f2fs/segment.c 		kmem_cache_free(inmem_entry_slab, cur);
cur               354 fs/f2fs/segment.c 	struct inmem_pages *cur = NULL;
cur               359 fs/f2fs/segment.c 	list_for_each_entry(cur, head, list) {
cur               360 fs/f2fs/segment.c 		if (cur->page == page)
cur               364 fs/f2fs/segment.c 	f2fs_bug_on(sbi, list_empty(head) || cur->page != page);
cur               365 fs/f2fs/segment.c 	list_del(&cur->list);
cur               369 fs/f2fs/segment.c 	kmem_cache_free(inmem_entry_slab, cur);
cur               382 fs/f2fs/segment.c 	struct inmem_pages *cur, *tmp;
cur               397 fs/f2fs/segment.c 	list_for_each_entry_safe(cur, tmp, &fi->inmem_pages, list) {
cur               398 fs/f2fs/segment.c 		struct page *page = cur->page;
cur               427 fs/f2fs/segment.c 			cur->old_addr = fio.old_blkaddr;
cur               431 fs/f2fs/segment.c 		list_move_tail(&cur->list, &revoke_list);
cur               640 fs/gfs2/dir.c  	struct gfs2_dirent *cur = *dent, *tmp;
cur               644 fs/gfs2/dir.c  	ret = dirent_check_reclen(dip, cur, bh_end);
cur               648 fs/gfs2/dir.c  	tmp = (void *)cur + ret;
cur               673 fs/gfs2/dir.c  		       struct gfs2_dirent *prev, struct gfs2_dirent *cur)
cur               677 fs/gfs2/dir.c  	if (gfs2_dirent_sentinel(cur)) {
cur               689 fs/gfs2/dir.c  		cur->de_inum.no_addr = 0;
cur               690 fs/gfs2/dir.c  		cur->de_inum.no_formal_ino = 0;
cur               697 fs/gfs2/dir.c  	cur_rec_len = be16_to_cpu(cur->de_rec_len);
cur               699 fs/gfs2/dir.c  	if ((char *)prev + prev_rec_len != (char *)cur)
cur               701 fs/gfs2/dir.c  	if ((char *)cur + cur_rec_len > bh->b_data + bh->b_size)
cur               511 fs/gfs2/rgrp.c 	struct gfs2_rgrpd *cur;
cur               516 fs/gfs2/rgrp.c 		cur = rb_entry(n, struct gfs2_rgrpd, rd_node);
cur               518 fs/gfs2/rgrp.c 		if (blk < cur->rd_addr)
cur               520 fs/gfs2/rgrp.c 		else if (blk >= cur->rd_data0 + cur->rd_data)
cur               525 fs/gfs2/rgrp.c 				if (blk < cur->rd_addr)
cur               527 fs/gfs2/rgrp.c 				if (blk >= cur->rd_data0 + cur->rd_data)
cur               530 fs/gfs2/rgrp.c 			return cur;
cur               872 fs/gfs2/rgrp.c 		struct gfs2_rgrpd *cur = rb_entry(*newn, struct gfs2_rgrpd,
cur               876 fs/gfs2/rgrp.c 		if (rgd->rd_addr < cur->rd_addr)
cur               878 fs/gfs2/rgrp.c 		else if (rgd->rd_addr > cur->rd_addr)
cur              1502 fs/gfs2/rgrp.c 		struct gfs2_blkreserv *cur =
cur              1506 fs/gfs2/rgrp.c 		rc = rs_cmp(fsblock, rs->rs_free, cur);
cur               242 fs/nfsd/nfs4state.c 	struct nfsd4_blocked_lock *cur, *found = NULL;
cur               245 fs/nfsd/nfs4state.c 	list_for_each_entry(cur, &lo->lo_blocked, nbl_list) {
cur               246 fs/nfsd/nfs4state.c 		if (fh_match(fh, &cur->nbl_fh)) {
cur               247 fs/nfsd/nfs4state.c 			list_del_init(&cur->nbl_list);
cur               248 fs/nfsd/nfs4state.c 			list_del_init(&cur->nbl_lru);
cur               249 fs/nfsd/nfs4state.c 			found = cur;
cur              2148 fs/nfsd/nfs4xdr.c 	struct path cur = *path;
cur              2156 fs/nfsd/nfs4xdr.c 	path_get(&cur);
cur              2161 fs/nfsd/nfs4xdr.c 		if (path_equal(&cur, root))
cur              2163 fs/nfsd/nfs4xdr.c 		if (cur.dentry == cur.mnt->mnt_root) {
cur              2164 fs/nfsd/nfs4xdr.c 			if (follow_up(&cur))
cur              2177 fs/nfsd/nfs4xdr.c 		components[ncomponents++] = cur.dentry;
cur              2178 fs/nfsd/nfs4xdr.c 		cur.dentry = dget_parent(cur.dentry);
cur              2210 fs/nfsd/nfs4xdr.c 	path_put(&cur);
cur               883 fs/pipe.c      	int cur = *cnt;	
cur               885 fs/pipe.c      	while (cur == *cnt) {
cur               890 fs/pipe.c      	return cur == *cnt ? -ERESTARTSYS : 0;
cur              1397 fs/reiserfs/bitmap.c 	unsigned long *cur = (unsigned long *)(bh->b_data + bh->b_size);
cur              1406 fs/reiserfs/bitmap.c 	while (--cur >= (unsigned long *)bh->b_data) {
cur              1408 fs/reiserfs/bitmap.c 		if (*cur == 0)
cur              1410 fs/reiserfs/bitmap.c 		else if (*cur != ~0L)	/* A mix, investigate */
cur              1411 fs/reiserfs/bitmap.c 			info->free_count += BITS_PER_LONG - hweight_long(*cur);
cur               131 fs/reiserfs/ibalance.c 	struct buffer_head *cur = cur_bi->bi_bh;
cur               142 fs/reiserfs/ibalance.c 	blkh = B_BLK_HEAD(cur);
cur               146 fs/reiserfs/ibalance.c 	RFALSE(B_FREE_SPACE(cur) < count * (KEY_SIZE + DC_SIZE),
cur               148 fs/reiserfs/ibalance.c 	       B_FREE_SPACE(cur), count * (KEY_SIZE + DC_SIZE));
cur               151 fs/reiserfs/ibalance.c 	dc = B_N_CHILD(cur, to + 1);
cur               164 fs/reiserfs/ibalance.c 	ih = internal_key(cur, ((to == -1) ? 0 : to));
cur               180 fs/reiserfs/ibalance.c 	do_balance_mark_internal_dirty(cur_bi->tb, cur, 0);
cur               183 fs/reiserfs/ibalance.c 	check_internal(cur);
cur               209 fs/reiserfs/ibalance.c 	struct buffer_head *cur = cur_bi->bi_bh;
cur               215 fs/reiserfs/ibalance.c 	RFALSE(cur == NULL, "buffer is 0");
cur               218 fs/reiserfs/ibalance.c 	RFALSE(first_p < 0 || first_p + del_num > B_NR_ITEMS(cur) + 1
cur               223 fs/reiserfs/ibalance.c 	       B_NR_ITEMS(cur) + 1, first_i);
cur               227 fs/reiserfs/ibalance.c 	blkh = B_BLK_HEAD(cur);
cur               237 fs/reiserfs/ibalance.c 	RFALSE(first_i + del_num > B_NR_ITEMS(cur),
cur               240 fs/reiserfs/ibalance.c 	       first_i, del_num, first_i + del_num, cur, cur);
cur               243 fs/reiserfs/ibalance.c 	dc = B_N_CHILD(cur, first_p);
cur               246 fs/reiserfs/ibalance.c 	key = internal_key(cur, first_i);
cur               257 fs/reiserfs/ibalance.c 	do_balance_mark_internal_dirty(cur_bi->tb, cur, 0);
cur               259 fs/reiserfs/ibalance.c 	check_internal(cur);
cur              1845 fs/reiserfs/journal.c 	struct reiserfs_journal_cnode *cur;
cur              1852 fs/reiserfs/journal.c 	cur = *head;
cur              1853 fs/reiserfs/journal.c 	while (cur) {
cur              1854 fs/reiserfs/journal.c 		if (cur->blocknr == block && cur->sb == sb
cur              1855 fs/reiserfs/journal.c 		    && (jl == NULL || jl == cur->jlist)
cur              1856 fs/reiserfs/journal.c 		    && (!test_bit(BLOCK_FREED, &cur->state) || remove_freed)) {
cur              1857 fs/reiserfs/journal.c 			if (cur->hnext) {
cur              1858 fs/reiserfs/journal.c 				cur->hnext->hprev = cur->hprev;
cur              1860 fs/reiserfs/journal.c 			if (cur->hprev) {
cur              1861 fs/reiserfs/journal.c 				cur->hprev->hnext = cur->hnext;
cur              1863 fs/reiserfs/journal.c 				*head = cur->hnext;
cur              1865 fs/reiserfs/journal.c 			cur->blocknr = 0;
cur              1866 fs/reiserfs/journal.c 			cur->sb = NULL;
cur              1867 fs/reiserfs/journal.c 			cur->state = 0;
cur              1872 fs/reiserfs/journal.c 			if (cur->bh && cur->jlist)
cur              1873 fs/reiserfs/journal.c 				atomic_dec(&cur->jlist->j_nonzerolen);
cur              1874 fs/reiserfs/journal.c 			cur->bh = NULL;
cur              1875 fs/reiserfs/journal.c 			cur->jlist = NULL;
cur              1877 fs/reiserfs/journal.c 		cur = cur->hnext;
cur              3476 fs/reiserfs/journal.c 	struct reiserfs_journal_cnode *cur = cn->hprev;
cur              3484 fs/reiserfs/journal.c 	while (cur && can_dirty) {
cur              3485 fs/reiserfs/journal.c 		if (cur->jlist && cur->bh && cur->blocknr && cur->sb == sb &&
cur              3486 fs/reiserfs/journal.c 		    cur->blocknr == blocknr) {
cur              3489 fs/reiserfs/journal.c 		cur = cur->hprev;
cur              3495 fs/reiserfs/journal.c 	cur = cn->hnext;
cur              3496 fs/reiserfs/journal.c 	while (cur && can_dirty) {
cur              3497 fs/reiserfs/journal.c 		if (cur->jlist && cur->jlist->j_len > 0 &&
cur              3498 fs/reiserfs/journal.c 		    atomic_read(&cur->jlist->j_commit_left) > 0 && cur->bh &&
cur              3499 fs/reiserfs/journal.c 		    cur->blocknr && cur->sb == sb && cur->blocknr == blocknr) {
cur              3502 fs/reiserfs/journal.c 		cur = cur->hnext;
cur              2063 fs/reiserfs/reiserfs.h #define B_CHILD_SIZE(cur) (MAX_CHILD_SIZE(cur)-(B_FREE_SPACE(cur)))
cur               979 fs/reiserfs/super.c static int reiserfs_getopt(struct super_block *s, char **cur, opt_desc_t * opts,
cur               993 fs/reiserfs/super.c 	p = *cur;
cur               996 fs/reiserfs/super.c 	*cur = strchr(p, ',');
cur               997 fs/reiserfs/super.c 	if (*cur) {
cur               998 fs/reiserfs/super.c 		*(*cur) = '\0';
cur               999 fs/reiserfs/super.c 		(*cur)++;
cur              2300 fs/ubifs/debug.c 	struct list_head *cur;
cur              2306 fs/ubifs/debug.c 	for (cur = head->next; cur->next != head; cur = cur->next) {
cur              2311 fs/ubifs/debug.c 		sa = container_of(cur, struct ubifs_scan_node, list);
cur              2312 fs/ubifs/debug.c 		sb = container_of(cur->next, struct ubifs_scan_node, list);
cur              2367 fs/ubifs/debug.c 	struct list_head *cur;
cur              2373 fs/ubifs/debug.c 	for (cur = head->next; cur->next != head; cur = cur->next) {
cur              2378 fs/ubifs/debug.c 		sa = container_of(cur, struct ubifs_scan_node, list);
cur              2379 fs/ubifs/debug.c 		sb = container_of(cur->next, struct ubifs_scan_node, list);
cur               108 fs/unicode/utf8-core.c 	struct utf8cursor cur;
cur               111 fs/unicode/utf8-core.c 	if (utf8ncursor(&cur, data, str->name, str->len) < 0)
cur               115 fs/unicode/utf8-core.c 		int c = utf8byte(&cur);
cur               132 fs/unicode/utf8-core.c 	struct utf8cursor cur;
cur               135 fs/unicode/utf8-core.c 	if (utf8ncursor(&cur, data, str->name, str->len) < 0)
cur               139 fs/unicode/utf8-core.c 		int c = utf8byte(&cur);
cur              1306 fs/userfaultfd.c 	struct vm_area_struct *vma, *prev, *cur;
cur              1381 fs/userfaultfd.c 	for (cur = vma; cur && cur->vm_start < end; cur = cur->vm_next) {
cur              1384 fs/userfaultfd.c 		BUG_ON(!!cur->vm_userfaultfd_ctx.ctx ^
cur              1385 fs/userfaultfd.c 		       !!(cur->vm_flags & (VM_UFFD_MISSING | VM_UFFD_WP)));
cur              1389 fs/userfaultfd.c 		if (!vma_can_userfault(cur))
cur              1401 fs/userfaultfd.c 		if (unlikely(!(cur->vm_flags & VM_MAYWRITE)))
cur              1408 fs/userfaultfd.c 		if (is_vm_hugetlb_page(cur) && end <= cur->vm_end &&
cur              1409 fs/userfaultfd.c 		    end > cur->vm_start) {
cur              1410 fs/userfaultfd.c 			unsigned long vma_hpagesize = vma_kernel_pagesize(cur);
cur              1425 fs/userfaultfd.c 		if (cur->vm_userfaultfd_ctx.ctx &&
cur              1426 fs/userfaultfd.c 		    cur->vm_userfaultfd_ctx.ctx != ctx)
cur              1432 fs/userfaultfd.c 		if (is_vm_hugetlb_page(cur))
cur              1518 fs/userfaultfd.c 	struct vm_area_struct *vma, *prev, *cur;
cur              1570 fs/userfaultfd.c 	for (cur = vma; cur && cur->vm_start < end; cur = cur->vm_next) {
cur              1573 fs/userfaultfd.c 		BUG_ON(!!cur->vm_userfaultfd_ctx.ctx ^
cur              1574 fs/userfaultfd.c 		       !!(cur->vm_flags & (VM_UFFD_MISSING | VM_UFFD_WP)));
cur              1583 fs/userfaultfd.c 		if (!vma_can_userfault(cur))
cur               144 fs/xfs/libxfs/xfs_alloc.c 	struct xfs_btree_cur	*cur,	/* btree cursor */
cur               149 fs/xfs/libxfs/xfs_alloc.c 	cur->bc_rec.a.ar_startblock = bno;
cur               150 fs/xfs/libxfs/xfs_alloc.c 	cur->bc_rec.a.ar_blockcount = len;
cur               151 fs/xfs/libxfs/xfs_alloc.c 	return xfs_btree_lookup(cur, XFS_LOOKUP_EQ, stat);
cur               160 fs/xfs/libxfs/xfs_alloc.c 	struct xfs_btree_cur	*cur,	/* btree cursor */
cur               165 fs/xfs/libxfs/xfs_alloc.c 	cur->bc_rec.a.ar_startblock = bno;
cur               166 fs/xfs/libxfs/xfs_alloc.c 	cur->bc_rec.a.ar_blockcount = len;
cur               167 fs/xfs/libxfs/xfs_alloc.c 	return xfs_btree_lookup(cur, XFS_LOOKUP_GE, stat);
cur               176 fs/xfs/libxfs/xfs_alloc.c 	struct xfs_btree_cur	*cur,	/* btree cursor */
cur               181 fs/xfs/libxfs/xfs_alloc.c 	cur->bc_rec.a.ar_startblock = bno;
cur               182 fs/xfs/libxfs/xfs_alloc.c 	cur->bc_rec.a.ar_blockcount = len;
cur               183 fs/xfs/libxfs/xfs_alloc.c 	return xfs_btree_lookup(cur, XFS_LOOKUP_LE, stat);
cur               193 fs/xfs/libxfs/xfs_alloc.c 	struct xfs_btree_cur	*cur,	/* btree cursor */
cur               201 fs/xfs/libxfs/xfs_alloc.c 	return xfs_btree_update(cur, &rec);
cur               209 fs/xfs/libxfs/xfs_alloc.c 	struct xfs_btree_cur	*cur,	/* btree cursor */
cur               214 fs/xfs/libxfs/xfs_alloc.c 	struct xfs_mount	*mp = cur->bc_mp;
cur               215 fs/xfs/libxfs/xfs_alloc.c 	xfs_agnumber_t		agno = cur->bc_private.a.agno;
cur               219 fs/xfs/libxfs/xfs_alloc.c 	error = xfs_btree_get_rec(cur, &rec, stat);
cur               242 fs/xfs/libxfs/xfs_alloc.c 		cur->bc_btnum == XFS_BTNUM_BNO ? "Block" : "Size", agno);
cur              3070 fs/xfs/libxfs/xfs_alloc.c 	struct xfs_btree_cur		*cur,
cur              3079 fs/xfs/libxfs/xfs_alloc.c 	return query->fn(cur, &irec, query->priv);
cur              3085 fs/xfs/libxfs/xfs_alloc.c 	struct xfs_btree_cur			*cur,
cur              3095 fs/xfs/libxfs/xfs_alloc.c 	ASSERT(cur->bc_btnum == XFS_BTNUM_BNO);
cur              3100 fs/xfs/libxfs/xfs_alloc.c 	return xfs_btree_query_range(cur, &low_brec, &high_brec,
cur              3107 fs/xfs/libxfs/xfs_alloc.c 	struct xfs_btree_cur			*cur,
cur              3113 fs/xfs/libxfs/xfs_alloc.c 	ASSERT(cur->bc_btnum == XFS_BTNUM_BNO);
cur              3116 fs/xfs/libxfs/xfs_alloc.c 	return xfs_btree_query_all(cur, xfs_alloc_query_range_helper, &query);
cur              3122 fs/xfs/libxfs/xfs_alloc.c 	struct xfs_btree_cur	*cur,
cur              3135 fs/xfs/libxfs/xfs_alloc.c 	return xfs_btree_has_record(cur, &low, &high, exists);
cur               202 fs/xfs/libxfs/xfs_alloc.h 	struct xfs_btree_cur	*cur,	/* btree cursor */
cur               209 fs/xfs/libxfs/xfs_alloc.h 	struct xfs_btree_cur	*cur,	/* btree cursor */
cur               216 fs/xfs/libxfs/xfs_alloc.h 	struct xfs_btree_cur	*cur,	/* btree cursor */
cur               234 fs/xfs/libxfs/xfs_alloc.h 	struct xfs_btree_cur		*cur,
cur               238 fs/xfs/libxfs/xfs_alloc.h int xfs_alloc_query_range(struct xfs_btree_cur *cur,
cur               242 fs/xfs/libxfs/xfs_alloc.h int xfs_alloc_query_all(struct xfs_btree_cur *cur, xfs_alloc_query_range_fn fn,
cur               245 fs/xfs/libxfs/xfs_alloc.h int xfs_alloc_has_record(struct xfs_btree_cur *cur, xfs_agblock_t bno,
cur                25 fs/xfs/libxfs/xfs_alloc_btree.c 	struct xfs_btree_cur	*cur)
cur                27 fs/xfs/libxfs/xfs_alloc_btree.c 	return xfs_allocbt_init_cursor(cur->bc_mp, cur->bc_tp,
cur                28 fs/xfs/libxfs/xfs_alloc_btree.c 			cur->bc_private.a.agbp, cur->bc_private.a.agno,
cur                29 fs/xfs/libxfs/xfs_alloc_btree.c 			cur->bc_btnum);
cur                34 fs/xfs/libxfs/xfs_alloc_btree.c 	struct xfs_btree_cur	*cur,
cur                38 fs/xfs/libxfs/xfs_alloc_btree.c 	struct xfs_buf		*agbp = cur->bc_private.a.agbp;
cur                41 fs/xfs/libxfs/xfs_alloc_btree.c 	int			btnum = cur->bc_btnum;
cur                42 fs/xfs/libxfs/xfs_alloc_btree.c 	struct xfs_perag	*pag = xfs_perag_get(cur->bc_mp, seqno);
cur                51 fs/xfs/libxfs/xfs_alloc_btree.c 	xfs_alloc_log_agf(cur->bc_tp, agbp, XFS_AGF_ROOTS | XFS_AGF_LEVELS);
cur                56 fs/xfs/libxfs/xfs_alloc_btree.c 	struct xfs_btree_cur	*cur,
cur                65 fs/xfs/libxfs/xfs_alloc_btree.c 	error = xfs_alloc_get_freelist(cur->bc_tp, cur->bc_private.a.agbp,
cur                75 fs/xfs/libxfs/xfs_alloc_btree.c 	xfs_extent_busy_reuse(cur->bc_mp, cur->bc_private.a.agno, bno, 1, false);
cur                77 fs/xfs/libxfs/xfs_alloc_btree.c 	xfs_trans_agbtree_delta(cur->bc_tp, 1);
cur                86 fs/xfs/libxfs/xfs_alloc_btree.c 	struct xfs_btree_cur	*cur,
cur                89 fs/xfs/libxfs/xfs_alloc_btree.c 	struct xfs_buf		*agbp = cur->bc_private.a.agbp;
cur                94 fs/xfs/libxfs/xfs_alloc_btree.c 	bno = xfs_daddr_to_agbno(cur->bc_mp, XFS_BUF_ADDR(bp));
cur                95 fs/xfs/libxfs/xfs_alloc_btree.c 	error = xfs_alloc_put_freelist(cur->bc_tp, agbp, NULL, bno, 1);
cur                99 fs/xfs/libxfs/xfs_alloc_btree.c 	xfs_extent_busy_insert(cur->bc_tp, be32_to_cpu(agf->agf_seqno), bno, 1,
cur               101 fs/xfs/libxfs/xfs_alloc_btree.c 	xfs_trans_agbtree_delta(cur->bc_tp, -1);
cur               110 fs/xfs/libxfs/xfs_alloc_btree.c 	struct xfs_btree_cur	*cur,
cur               116 fs/xfs/libxfs/xfs_alloc_btree.c 	struct xfs_agf		*agf = XFS_BUF_TO_AGF(cur->bc_private.a.agbp);
cur               122 fs/xfs/libxfs/xfs_alloc_btree.c 	ASSERT(cur->bc_btnum == XFS_BTNUM_CNT);
cur               149 fs/xfs/libxfs/xfs_alloc_btree.c 			rrp = XFS_ALLOC_REC_ADDR(cur->bc_mp, block, numrecs);
cur               162 fs/xfs/libxfs/xfs_alloc_btree.c 	pag = xfs_perag_get(cur->bc_mp, seqno);
cur               165 fs/xfs/libxfs/xfs_alloc_btree.c 	xfs_alloc_log_agf(cur->bc_tp, cur->bc_private.a.agbp, XFS_AGF_LONGEST);
cur               170 fs/xfs/libxfs/xfs_alloc_btree.c 	struct xfs_btree_cur	*cur,
cur               173 fs/xfs/libxfs/xfs_alloc_btree.c 	return cur->bc_mp->m_alloc_mnr[level != 0];
cur               178 fs/xfs/libxfs/xfs_alloc_btree.c 	struct xfs_btree_cur	*cur,
cur               181 fs/xfs/libxfs/xfs_alloc_btree.c 	return cur->bc_mp->m_alloc_mxr[level != 0];
cur               217 fs/xfs/libxfs/xfs_alloc_btree.c 	struct xfs_btree_cur	*cur,
cur               220 fs/xfs/libxfs/xfs_alloc_btree.c 	rec->alloc.ar_startblock = cpu_to_be32(cur->bc_rec.a.ar_startblock);
cur               221 fs/xfs/libxfs/xfs_alloc_btree.c 	rec->alloc.ar_blockcount = cpu_to_be32(cur->bc_rec.a.ar_blockcount);
cur               226 fs/xfs/libxfs/xfs_alloc_btree.c 	struct xfs_btree_cur	*cur,
cur               229 fs/xfs/libxfs/xfs_alloc_btree.c 	struct xfs_agf		*agf = XFS_BUF_TO_AGF(cur->bc_private.a.agbp);
cur               231 fs/xfs/libxfs/xfs_alloc_btree.c 	ASSERT(cur->bc_private.a.agno == be32_to_cpu(agf->agf_seqno));
cur               233 fs/xfs/libxfs/xfs_alloc_btree.c 	ptr->s = agf->agf_roots[cur->bc_btnum];
cur               238 fs/xfs/libxfs/xfs_alloc_btree.c 	struct xfs_btree_cur	*cur,
cur               241 fs/xfs/libxfs/xfs_alloc_btree.c 	xfs_alloc_rec_incore_t	*rec = &cur->bc_rec.a;
cur               249 fs/xfs/libxfs/xfs_alloc_btree.c 	struct xfs_btree_cur	*cur,
cur               252 fs/xfs/libxfs/xfs_alloc_btree.c 	xfs_alloc_rec_incore_t	*rec = &cur->bc_rec.a;
cur               265 fs/xfs/libxfs/xfs_alloc_btree.c 	struct xfs_btree_cur	*cur,
cur               275 fs/xfs/libxfs/xfs_alloc_btree.c 	struct xfs_btree_cur	*cur,
cur               385 fs/xfs/libxfs/xfs_alloc_btree.c 	struct xfs_btree_cur	*cur,
cur               395 fs/xfs/libxfs/xfs_alloc_btree.c 	struct xfs_btree_cur	*cur,
cur               406 fs/xfs/libxfs/xfs_alloc_btree.c 	struct xfs_btree_cur	*cur,
cur               419 fs/xfs/libxfs/xfs_alloc_btree.c 	struct xfs_btree_cur	*cur,
cur               486 fs/xfs/libxfs/xfs_alloc_btree.c 	struct xfs_btree_cur	*cur;
cur               490 fs/xfs/libxfs/xfs_alloc_btree.c 	cur = kmem_zone_zalloc(xfs_btree_cur_zone, KM_NOFS);
cur               492 fs/xfs/libxfs/xfs_alloc_btree.c 	cur->bc_tp = tp;
cur               493 fs/xfs/libxfs/xfs_alloc_btree.c 	cur->bc_mp = mp;
cur               494 fs/xfs/libxfs/xfs_alloc_btree.c 	cur->bc_btnum = btnum;
cur               495 fs/xfs/libxfs/xfs_alloc_btree.c 	cur->bc_blocklog = mp->m_sb.sb_blocklog;
cur               498 fs/xfs/libxfs/xfs_alloc_btree.c 		cur->bc_statoff = XFS_STATS_CALC_INDEX(xs_abtc_2);
cur               499 fs/xfs/libxfs/xfs_alloc_btree.c 		cur->bc_ops = &xfs_cntbt_ops;
cur               500 fs/xfs/libxfs/xfs_alloc_btree.c 		cur->bc_nlevels = be32_to_cpu(agf->agf_levels[XFS_BTNUM_CNT]);
cur               501 fs/xfs/libxfs/xfs_alloc_btree.c 		cur->bc_flags = XFS_BTREE_LASTREC_UPDATE;
cur               503 fs/xfs/libxfs/xfs_alloc_btree.c 		cur->bc_statoff = XFS_STATS_CALC_INDEX(xs_abtb_2);
cur               504 fs/xfs/libxfs/xfs_alloc_btree.c 		cur->bc_ops = &xfs_bnobt_ops;
cur               505 fs/xfs/libxfs/xfs_alloc_btree.c 		cur->bc_nlevels = be32_to_cpu(agf->agf_levels[XFS_BTNUM_BNO]);
cur               508 fs/xfs/libxfs/xfs_alloc_btree.c 	cur->bc_private.a.agbp = agbp;
cur               509 fs/xfs/libxfs/xfs_alloc_btree.c 	cur->bc_private.a.agno = agno;
cur               512 fs/xfs/libxfs/xfs_alloc_btree.c 		cur->bc_flags |= XFS_BTREE_CRC_BLOCKS;
cur               514 fs/xfs/libxfs/xfs_alloc_btree.c 	return cur;
cur                98 fs/xfs/libxfs/xfs_bmap.c 	struct xfs_btree_cur	*cur,
cur               102 fs/xfs/libxfs/xfs_bmap.c 	cur->bc_rec.b = *irec;
cur               103 fs/xfs/libxfs/xfs_bmap.c 	return xfs_btree_lookup(cur, XFS_LOOKUP_EQ, stat);
cur               108 fs/xfs/libxfs/xfs_bmap.c 	struct xfs_btree_cur	*cur,
cur               111 fs/xfs/libxfs/xfs_bmap.c 	cur->bc_rec.b.br_startoff = 0;
cur               112 fs/xfs/libxfs/xfs_bmap.c 	cur->bc_rec.b.br_startblock = 0;
cur               113 fs/xfs/libxfs/xfs_bmap.c 	cur->bc_rec.b.br_blockcount = 0;
cur               114 fs/xfs/libxfs/xfs_bmap.c 	return xfs_btree_lookup(cur, XFS_LOOKUP_GE, stat);
cur               145 fs/xfs/libxfs/xfs_bmap.c 	struct xfs_btree_cur	*cur,
cur               151 fs/xfs/libxfs/xfs_bmap.c 	return xfs_btree_update(cur, &rec);
cur               229 fs/xfs/libxfs/xfs_bmap.c 	struct xfs_btree_cur	*cur,
cur               235 fs/xfs/libxfs/xfs_bmap.c 	if (!cur)
cur               239 fs/xfs/libxfs/xfs_bmap.c 		if (!cur->bc_bufs[i])
cur               241 fs/xfs/libxfs/xfs_bmap.c 		if (XFS_BUF_ADDR(cur->bc_bufs[i]) == bno)
cur               242 fs/xfs/libxfs/xfs_bmap.c 			return cur->bc_bufs[i];
cur               246 fs/xfs/libxfs/xfs_bmap.c 	list_for_each_entry(lip, &cur->bc_tp->t_items, li_trans) {
cur               315 fs/xfs/libxfs/xfs_bmap.c 	xfs_btree_cur_t		*cur,	/* btree cursor or null */
cur               365 fs/xfs/libxfs/xfs_bmap.c 		bp = xfs_bmap_get_bp(cur, XFS_FSB_TO_DADDR(mp, bno));
cur               449 fs/xfs/libxfs/xfs_bmap.c 		bp = xfs_bmap_get_bp(cur, XFS_FSB_TO_DADDR(mp, bno));
cur               517 fs/xfs/libxfs/xfs_bmap.c #define xfs_bmap_check_leaf_extents(cur, ip, whichfork)		do { } while (0)
cur               585 fs/xfs/libxfs/xfs_bmap.c 	struct xfs_btree_cur	*cur,	/* btree cursor */
cur               603 fs/xfs/libxfs/xfs_bmap.c 	ASSERT(cur);
cur               614 fs/xfs/libxfs/xfs_bmap.c 	XFS_WANT_CORRUPTED_RETURN(cur->bc_mp,
cur               615 fs/xfs/libxfs/xfs_bmap.c 			xfs_btree_check_lptr(cur, cbno, 1));
cur               622 fs/xfs/libxfs/xfs_bmap.c 	if ((error = xfs_btree_check_block(cur, cblock, 0, cbp)))
cur               625 fs/xfs/libxfs/xfs_bmap.c 	xfs_bmap_add_free(cur->bc_tp, cbno, 1, &oinfo);
cur               629 fs/xfs/libxfs/xfs_bmap.c 	if (cur->bc_bufs[0] == cbp)
cur               630 fs/xfs/libxfs/xfs_bmap.c 		cur->bc_bufs[0] = NULL;
cur               657 fs/xfs/libxfs/xfs_bmap.c 	struct xfs_btree_cur	*cur;		/* bmap btree cursor */
cur               689 fs/xfs/libxfs/xfs_bmap.c 	cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork);
cur               690 fs/xfs/libxfs/xfs_bmap.c 	cur->bc_private.b.flags = wasdel ? XFS_BTCUR_BPRV_WASDEL : 0;
cur               727 fs/xfs/libxfs/xfs_bmap.c 	cur->bc_private.b.allocated++;
cur               761 fs/xfs/libxfs/xfs_bmap.c 	pp = XFS_BMBT_PTR_ADDR(mp, block, 1, xfs_bmbt_get_maxrecs(cur,
cur               769 fs/xfs/libxfs/xfs_bmap.c 	xfs_btree_log_block(cur, abp, XFS_BB_ALL_BITS);
cur               770 fs/xfs/libxfs/xfs_bmap.c 	xfs_btree_log_recs(cur, abp, 1, be16_to_cpu(ablock->bb_numrecs));
cur               772 fs/xfs/libxfs/xfs_bmap.c 	*curp = cur;
cur               782 fs/xfs/libxfs/xfs_bmap.c 	xfs_btree_del_cursor(cur, XFS_BTREE_ERROR);
cur               925 fs/xfs/libxfs/xfs_bmap.c 	xfs_btree_cur_t		*cur;		/* btree cursor */
cur               934 fs/xfs/libxfs/xfs_bmap.c 		cur = xfs_bmbt_init_cursor(mp, tp, ip, XFS_DATA_FORK);
cur               935 fs/xfs/libxfs/xfs_bmap.c 		error = xfs_bmbt_lookup_first(cur, &stat);
cur               940 fs/xfs/libxfs/xfs_bmap.c 		if ((error = xfs_btree_new_iroot(cur, flags, &stat)))
cur               943 fs/xfs/libxfs/xfs_bmap.c 			xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR);
cur               946 fs/xfs/libxfs/xfs_bmap.c 		cur->bc_private.b.allocated = 0;
cur               947 fs/xfs/libxfs/xfs_bmap.c 		xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR);
cur               951 fs/xfs/libxfs/xfs_bmap.c 	xfs_btree_del_cursor(cur, XFS_BTREE_ERROR);
cur               964 fs/xfs/libxfs/xfs_bmap.c 	xfs_btree_cur_t		*cur;		/* bmap btree cursor */
cur               969 fs/xfs/libxfs/xfs_bmap.c 	cur = NULL;
cur               970 fs/xfs/libxfs/xfs_bmap.c 	error = xfs_bmap_extents_to_btree(tp, ip, &cur, 0, flags,
cur               972 fs/xfs/libxfs/xfs_bmap.c 	if (cur) {
cur               973 fs/xfs/libxfs/xfs_bmap.c 		cur->bc_private.b.allocated = 0;
cur               974 fs/xfs/libxfs/xfs_bmap.c 		xfs_btree_del_cursor(cur, error);
cur              1557 fs/xfs/libxfs/xfs_bmap.c 	ASSERT(!bma->cur ||
cur              1558 fs/xfs/libxfs/xfs_bmap.c 	       (bma->cur->bc_private.b.flags & XFS_BTCUR_BPRV_WASDEL));
cur              1648 fs/xfs/libxfs/xfs_bmap.c 		if (bma->cur == NULL)
cur              1652 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_lookup_eq(bma->cur, &RIGHT, &i);
cur              1656 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_btree_delete(bma->cur, &i);
cur              1660 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_btree_decrement(bma->cur, 0, &i);
cur              1664 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_update(bma->cur, &LEFT);
cur              1682 fs/xfs/libxfs/xfs_bmap.c 		if (bma->cur == NULL)
cur              1686 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_lookup_eq(bma->cur, &old, &i);
cur              1690 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_update(bma->cur, &LEFT);
cur              1712 fs/xfs/libxfs/xfs_bmap.c 		if (bma->cur == NULL)
cur              1716 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_lookup_eq(bma->cur, &RIGHT, &i);
cur              1720 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_update(bma->cur, &PREV);
cur              1737 fs/xfs/libxfs/xfs_bmap.c 		if (bma->cur == NULL)
cur              1741 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_lookup_eq(bma->cur, new, &i);
cur              1745 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_btree_insert(bma->cur, &i);
cur              1772 fs/xfs/libxfs/xfs_bmap.c 		if (bma->cur == NULL)
cur              1776 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_lookup_eq(bma->cur, &old, &i);
cur              1780 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_update(bma->cur, &LEFT);
cur              1793 fs/xfs/libxfs/xfs_bmap.c 		if (bma->cur == NULL)
cur              1797 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_lookup_eq(bma->cur, new, &i);
cur              1801 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_btree_insert(bma->cur, &i);
cur              1809 fs/xfs/libxfs/xfs_bmap.c 					&bma->cur, 1, &tmp_rval, whichfork);
cur              1818 fs/xfs/libxfs/xfs_bmap.c 			(bma->cur ? bma->cur->bc_private.b.allocated : 0));
cur              1838 fs/xfs/libxfs/xfs_bmap.c 		if (bma->cur == NULL)
cur              1842 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_lookup_eq(bma->cur, &old, &i);
cur              1846 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_update(bma->cur, &RIGHT);
cur              1870 fs/xfs/libxfs/xfs_bmap.c 		if (bma->cur == NULL)
cur              1874 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_lookup_eq(bma->cur, new, &i);
cur              1878 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_btree_insert(bma->cur, &i);
cur              1886 fs/xfs/libxfs/xfs_bmap.c 				&bma->cur, 1, &tmp_rval, whichfork);
cur              1895 fs/xfs/libxfs/xfs_bmap.c 			(bma->cur ? bma->cur->bc_private.b.allocated : 0));
cur              1950 fs/xfs/libxfs/xfs_bmap.c 		if (bma->cur == NULL)
cur              1954 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_lookup_eq(bma->cur, new, &i);
cur              1958 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_btree_insert(bma->cur, &i);
cur              1966 fs/xfs/libxfs/xfs_bmap.c 					&bma->cur, 1, &tmp_rval, whichfork);
cur              1997 fs/xfs/libxfs/xfs_bmap.c 		ASSERT(bma->cur == NULL);
cur              1999 fs/xfs/libxfs/xfs_bmap.c 				&bma->cur, da_old > 0, &tmp_logflags,
cur              2009 fs/xfs/libxfs/xfs_bmap.c 	if (bma->cur) {
cur              2010 fs/xfs/libxfs/xfs_bmap.c 		da_new += bma->cur->bc_private.b.allocated;
cur              2011 fs/xfs/libxfs/xfs_bmap.c 		bma->cur->bc_private.b.allocated = 0;
cur              2021 fs/xfs/libxfs/xfs_bmap.c 	xfs_bmap_check_leaf_extents(bma->cur, bma->ip, whichfork);
cur              2044 fs/xfs/libxfs/xfs_bmap.c 	xfs_btree_cur_t		*cur;	/* btree cursor */
cur              2058 fs/xfs/libxfs/xfs_bmap.c 	cur = *curp;
cur              2148 fs/xfs/libxfs/xfs_bmap.c 		if (cur == NULL)
cur              2152 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_lookup_eq(cur, &RIGHT, &i);
cur              2156 fs/xfs/libxfs/xfs_bmap.c 			if ((error = xfs_btree_delete(cur, &i)))
cur              2159 fs/xfs/libxfs/xfs_bmap.c 			if ((error = xfs_btree_decrement(cur, 0, &i)))
cur              2162 fs/xfs/libxfs/xfs_bmap.c 			if ((error = xfs_btree_delete(cur, &i)))
cur              2165 fs/xfs/libxfs/xfs_bmap.c 			if ((error = xfs_btree_decrement(cur, 0, &i)))
cur              2168 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_update(cur, &LEFT);
cur              2186 fs/xfs/libxfs/xfs_bmap.c 		if (cur == NULL)
cur              2190 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_lookup_eq(cur, &PREV, &i);
cur              2194 fs/xfs/libxfs/xfs_bmap.c 			if ((error = xfs_btree_delete(cur, &i)))
cur              2197 fs/xfs/libxfs/xfs_bmap.c 			if ((error = xfs_btree_decrement(cur, 0, &i)))
cur              2200 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_update(cur, &LEFT);
cur              2221 fs/xfs/libxfs/xfs_bmap.c 		if (cur == NULL)
cur              2225 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_lookup_eq(cur, &RIGHT, &i);
cur              2229 fs/xfs/libxfs/xfs_bmap.c 			if ((error = xfs_btree_delete(cur, &i)))
cur              2232 fs/xfs/libxfs/xfs_bmap.c 			if ((error = xfs_btree_decrement(cur, 0, &i)))
cur              2235 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_update(cur, &PREV);
cur              2250 fs/xfs/libxfs/xfs_bmap.c 		if (cur == NULL)
cur              2254 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_lookup_eq(cur, new, &i);
cur              2258 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_update(cur, &PREV);
cur              2280 fs/xfs/libxfs/xfs_bmap.c 		if (cur == NULL)
cur              2284 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_lookup_eq(cur, &old, &i);
cur              2288 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_update(cur, &PREV);
cur              2291 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_btree_decrement(cur, 0, &i);
cur              2294 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_update(cur, &LEFT);
cur              2314 fs/xfs/libxfs/xfs_bmap.c 		if (cur == NULL)
cur              2318 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_lookup_eq(cur, &old, &i);
cur              2322 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_update(cur, &PREV);
cur              2325 fs/xfs/libxfs/xfs_bmap.c 			cur->bc_rec.b = *new;
cur              2326 fs/xfs/libxfs/xfs_bmap.c 			if ((error = xfs_btree_insert(cur, &i)))
cur              2348 fs/xfs/libxfs/xfs_bmap.c 		if (cur == NULL)
cur              2352 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_lookup_eq(cur, &old, &i);
cur              2356 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_update(cur, &PREV);
cur              2359 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_btree_increment(cur, 0, &i);
cur              2362 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_update(cur, &RIGHT);
cur              2382 fs/xfs/libxfs/xfs_bmap.c 		if (cur == NULL)
cur              2386 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_lookup_eq(cur, &old, &i);
cur              2390 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_update(cur, &PREV);
cur              2393 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_lookup_eq(cur, new, &i);
cur              2397 fs/xfs/libxfs/xfs_bmap.c 			if ((error = xfs_btree_insert(cur, &i)))
cur              2426 fs/xfs/libxfs/xfs_bmap.c 		if (cur == NULL)
cur              2430 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_lookup_eq(cur, &old, &i);
cur              2435 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_update(cur, &r[1]);
cur              2439 fs/xfs/libxfs/xfs_bmap.c 			cur->bc_rec.b = PREV;
cur              2440 fs/xfs/libxfs/xfs_bmap.c 			if ((error = xfs_btree_insert(cur, &i)))
cur              2448 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_lookup_eq(cur, new, &i);
cur              2453 fs/xfs/libxfs/xfs_bmap.c 			if ((error = xfs_btree_insert(cur, &i)))
cur              2479 fs/xfs/libxfs/xfs_bmap.c 		ASSERT(cur == NULL);
cur              2480 fs/xfs/libxfs/xfs_bmap.c 		error = xfs_bmap_extents_to_btree(tp, ip, &cur, 0,
cur              2488 fs/xfs/libxfs/xfs_bmap.c 	if (cur) {
cur              2489 fs/xfs/libxfs/xfs_bmap.c 		cur->bc_private.b.allocated = 0;
cur              2490 fs/xfs/libxfs/xfs_bmap.c 		*curp = cur;
cur              2658 fs/xfs/libxfs/xfs_bmap.c 	struct xfs_btree_cur	*cur = *curp;
cur              2668 fs/xfs/libxfs/xfs_bmap.c 	ASSERT(!cur || !(cur->bc_private.b.flags & XFS_BTCUR_BPRV_WASDEL));
cur              2731 fs/xfs/libxfs/xfs_bmap.c 		if (cur == NULL) {
cur              2735 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_lookup_eq(cur, &right, &i);
cur              2739 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_btree_delete(cur, &i);
cur              2743 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_btree_decrement(cur, 0, &i);
cur              2747 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_update(cur, &left);
cur              2765 fs/xfs/libxfs/xfs_bmap.c 		if (cur == NULL) {
cur              2769 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_lookup_eq(cur, &old, &i);
cur              2773 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_update(cur, &left);
cur              2792 fs/xfs/libxfs/xfs_bmap.c 		if (cur == NULL) {
cur              2796 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_lookup_eq(cur, &old, &i);
cur              2800 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_update(cur, &right);
cur              2815 fs/xfs/libxfs/xfs_bmap.c 		if (cur == NULL) {
cur              2819 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_lookup_eq(cur, new, &i);
cur              2823 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_btree_insert(cur, &i);
cur              2839 fs/xfs/libxfs/xfs_bmap.c 		ASSERT(cur == NULL);
cur              2843 fs/xfs/libxfs/xfs_bmap.c 		cur = *curp;
cur              2849 fs/xfs/libxfs/xfs_bmap.c 	if (cur)
cur              2850 fs/xfs/libxfs/xfs_bmap.c 		cur->bc_private.b.allocated = 0;
cur              2852 fs/xfs/libxfs/xfs_bmap.c 	xfs_bmap_check_leaf_extents(cur, ip, whichfork);
cur              4076 fs/xfs/libxfs/xfs_bmap.c 	if ((ifp->if_flags & XFS_IFBROOT) && !bma->cur)
cur              4077 fs/xfs/libxfs/xfs_bmap.c 		bma->cur = xfs_bmbt_init_cursor(mp, bma->tp, bma->ip, whichfork);
cur              4084 fs/xfs/libxfs/xfs_bmap.c 	if (bma->cur)
cur              4085 fs/xfs/libxfs/xfs_bmap.c 		bma->cur->bc_private.b.flags =
cur              4110 fs/xfs/libxfs/xfs_bmap.c 				whichfork, &bma->icur, &bma->cur, &bma->got,
cur              4159 fs/xfs/libxfs/xfs_bmap.c 	if ((ifp->if_flags & XFS_IFBROOT) && !bma->cur) {
cur              4160 fs/xfs/libxfs/xfs_bmap.c 		bma->cur = xfs_bmbt_init_cursor(bma->ip->i_mount, bma->tp,
cur              4178 fs/xfs/libxfs/xfs_bmap.c 			&bma->icur, &bma->cur, mval, &tmp_logflags);
cur              4245 fs/xfs/libxfs/xfs_bmap.c 	if (bma->cur)
cur              4246 fs/xfs/libxfs/xfs_bmap.c 		xfs_btree_del_cursor(bma->cur, error);
cur              4432 fs/xfs/libxfs/xfs_bmap.c 	error = xfs_bmap_btree_to_extents(tp, ip, bma.cur, &bma.logflags,
cur              4537 fs/xfs/libxfs/xfs_bmap.c 	error = xfs_bmap_btree_to_extents(tp, ip, bma.cur, &bma.logflags,
cur              4566 fs/xfs/libxfs/xfs_bmap.c 	struct xfs_btree_cur	*cur = NULL;
cur              4608 fs/xfs/libxfs/xfs_bmap.c 		cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork);
cur              4609 fs/xfs/libxfs/xfs_bmap.c 		cur->bc_private.b.flags = 0;
cur              4621 fs/xfs/libxfs/xfs_bmap.c 			&cur, &got, &logflags, flags);
cur              4625 fs/xfs/libxfs/xfs_bmap.c 	error = xfs_bmap_btree_to_extents(tp, ip, cur, &logflags, whichfork);
cur              4635 fs/xfs/libxfs/xfs_bmap.c 	if (cur)
cur              4636 fs/xfs/libxfs/xfs_bmap.c 		xfs_btree_del_cursor(cur, error);
cur              4934 fs/xfs/libxfs/xfs_bmap.c 	xfs_btree_cur_t		*cur,	/* if null, not a btree */
cur              5012 fs/xfs/libxfs/xfs_bmap.c 	if (cur) {
cur              5013 fs/xfs/libxfs/xfs_bmap.c 		error = xfs_bmbt_lookup_eq(cur, &got, &i);
cur              5034 fs/xfs/libxfs/xfs_bmap.c 		if (!cur) {
cur              5038 fs/xfs/libxfs/xfs_bmap.c 		if ((error = xfs_btree_delete(cur, &i)))
cur              5050 fs/xfs/libxfs/xfs_bmap.c 		if (!cur) {
cur              5054 fs/xfs/libxfs/xfs_bmap.c 		error = xfs_bmbt_update(cur, &got);
cur              5064 fs/xfs/libxfs/xfs_bmap.c 		if (!cur) {
cur              5068 fs/xfs/libxfs/xfs_bmap.c 		error = xfs_bmbt_update(cur, &got);
cur              5087 fs/xfs/libxfs/xfs_bmap.c 		if (cur) {
cur              5088 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmbt_update(cur, &got);
cur              5091 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_btree_increment(cur, 0, &i);
cur              5094 fs/xfs/libxfs/xfs_bmap.c 			cur->bc_rec.b = new;
cur              5095 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_btree_insert(cur, &i);
cur              5108 fs/xfs/libxfs/xfs_bmap.c 				error = xfs_bmbt_lookup_eq(cur, &got, &i);
cur              5116 fs/xfs/libxfs/xfs_bmap.c 				error = xfs_bmbt_update(cur, &old);
cur              5186 fs/xfs/libxfs/xfs_bmap.c 	struct xfs_btree_cur	*cur;		/* bmap btree cursor */
cur              5257 fs/xfs/libxfs/xfs_bmap.c 		cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork);
cur              5258 fs/xfs/libxfs/xfs_bmap.c 		cur->bc_private.b.flags = 0;
cur              5260 fs/xfs/libxfs/xfs_bmap.c 		cur = NULL;
cur              5371 fs/xfs/libxfs/xfs_bmap.c 					whichfork, &icur, &cur, &del,
cur              5427 fs/xfs/libxfs/xfs_bmap.c 						ip, whichfork, &icur, &cur,
cur              5436 fs/xfs/libxfs/xfs_bmap.c 						ip, whichfork, &icur, &cur,
cur              5449 fs/xfs/libxfs/xfs_bmap.c 			error = xfs_bmap_del_extent_real(ip, tp, &icur, cur,
cur              5483 fs/xfs/libxfs/xfs_bmap.c 		ASSERT(cur == NULL);
cur              5484 fs/xfs/libxfs/xfs_bmap.c 		error = xfs_bmap_extents_to_btree(tp, ip, &cur, 0,
cur              5488 fs/xfs/libxfs/xfs_bmap.c 		error = xfs_bmap_btree_to_extents(tp, ip, cur, &logflags,
cur              5509 fs/xfs/libxfs/xfs_bmap.c 	if (cur) {
cur              5511 fs/xfs/libxfs/xfs_bmap.c 			cur->bc_private.b.allocated = 0;
cur              5512 fs/xfs/libxfs/xfs_bmap.c 		xfs_btree_del_cursor(cur, error);
cur              5580 fs/xfs/libxfs/xfs_bmap.c 	struct xfs_btree_cur		*cur,
cur              5604 fs/xfs/libxfs/xfs_bmap.c 	if (!cur) {
cur              5610 fs/xfs/libxfs/xfs_bmap.c 	error = xfs_bmbt_lookup_eq(cur, got, &i);
cur              5615 fs/xfs/libxfs/xfs_bmap.c 	error = xfs_btree_delete(cur, &i);
cur              5621 fs/xfs/libxfs/xfs_bmap.c 	error = xfs_bmbt_lookup_eq(cur, left, &i);
cur              5626 fs/xfs/libxfs/xfs_bmap.c 	error = xfs_bmbt_update(cur, &new);
cur              5631 fs/xfs/libxfs/xfs_bmap.c 	error = xfs_bmap_btree_to_extents(tp, ip, cur, logflags, whichfork);
cur              5656 fs/xfs/libxfs/xfs_bmap.c 	struct xfs_btree_cur	*cur,
cur              5668 fs/xfs/libxfs/xfs_bmap.c 	if (cur) {
cur              5669 fs/xfs/libxfs/xfs_bmap.c 		error = xfs_bmbt_lookup_eq(cur, &prev, &i);
cur              5674 fs/xfs/libxfs/xfs_bmap.c 		error = xfs_bmbt_update(cur, got);
cur              5701 fs/xfs/libxfs/xfs_bmap.c 	struct xfs_btree_cur	*cur = NULL;
cur              5728 fs/xfs/libxfs/xfs_bmap.c 		cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork);
cur              5729 fs/xfs/libxfs/xfs_bmap.c 		cur->bc_private.b.flags = 0;
cur              5749 fs/xfs/libxfs/xfs_bmap.c 					cur, &logflags);
cur              5762 fs/xfs/libxfs/xfs_bmap.c 			cur, &logflags, new_startoff);
cur              5774 fs/xfs/libxfs/xfs_bmap.c 	if (cur)
cur              5775 fs/xfs/libxfs/xfs_bmap.c 		xfs_btree_del_cursor(cur, error);
cur              5819 fs/xfs/libxfs/xfs_bmap.c 	struct xfs_btree_cur	*cur = NULL;
cur              5846 fs/xfs/libxfs/xfs_bmap.c 		cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork);
cur              5847 fs/xfs/libxfs/xfs_bmap.c 		cur->bc_private.b.flags = 0;
cur              5889 fs/xfs/libxfs/xfs_bmap.c 			cur, &logflags, new_startoff);
cur              5901 fs/xfs/libxfs/xfs_bmap.c 	if (cur)
cur              5902 fs/xfs/libxfs/xfs_bmap.c 		xfs_btree_del_cursor(cur, error);
cur              5921 fs/xfs/libxfs/xfs_bmap.c 	struct xfs_btree_cur		*cur = NULL;
cur              5966 fs/xfs/libxfs/xfs_bmap.c 		cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork);
cur              5967 fs/xfs/libxfs/xfs_bmap.c 		cur->bc_private.b.flags = 0;
cur              5968 fs/xfs/libxfs/xfs_bmap.c 		error = xfs_bmbt_lookup_eq(cur, &got, &i);
cur              5979 fs/xfs/libxfs/xfs_bmap.c 	if (cur) {
cur              5980 fs/xfs/libxfs/xfs_bmap.c 		error = xfs_bmbt_update(cur, &got);
cur              5992 fs/xfs/libxfs/xfs_bmap.c 	if (cur) {
cur              5993 fs/xfs/libxfs/xfs_bmap.c 		error = xfs_bmbt_lookup_eq(cur, &new, &i);
cur              5997 fs/xfs/libxfs/xfs_bmap.c 		error = xfs_btree_insert(cur, &i);
cur              6009 fs/xfs/libxfs/xfs_bmap.c 		ASSERT(cur == NULL);
cur              6010 fs/xfs/libxfs/xfs_bmap.c 		error = xfs_bmap_extents_to_btree(tp, ip, &cur, 0,
cur              6016 fs/xfs/libxfs/xfs_bmap.c 	if (cur) {
cur              6017 fs/xfs/libxfs/xfs_bmap.c 		cur->bc_private.b.allocated = 0;
cur              6018 fs/xfs/libxfs/xfs_bmap.c 		xfs_btree_del_cursor(cur, error);
cur                31 fs/xfs/libxfs/xfs_bmap.h 	struct xfs_btree_cur	*cur;	/* btree cursor */
cur               211 fs/xfs/libxfs/xfs_bmap.h 		struct xfs_iext_cursor *cur, struct xfs_bmbt_irec *got,
cur               214 fs/xfs/libxfs/xfs_bmap.h 		struct xfs_iext_cursor *cur, struct xfs_bmbt_irec *got,
cur               228 fs/xfs/libxfs/xfs_bmap.h 		struct xfs_bmbt_irec *got, struct xfs_iext_cursor *cur,
cur               164 fs/xfs/libxfs/xfs_bmap_btree.c 	struct xfs_btree_cur	*cur)
cur               168 fs/xfs/libxfs/xfs_bmap_btree.c 	new = xfs_bmbt_init_cursor(cur->bc_mp, cur->bc_tp,
cur               169 fs/xfs/libxfs/xfs_bmap_btree.c 			cur->bc_private.b.ip, cur->bc_private.b.whichfork);
cur               175 fs/xfs/libxfs/xfs_bmap_btree.c 	new->bc_private.b.flags = cur->bc_private.b.flags;
cur               196 fs/xfs/libxfs/xfs_bmap_btree.c 	struct xfs_btree_cur	*cur,
cur               205 fs/xfs/libxfs/xfs_bmap_btree.c 	args.tp = cur->bc_tp;
cur               206 fs/xfs/libxfs/xfs_bmap_btree.c 	args.mp = cur->bc_mp;
cur               207 fs/xfs/libxfs/xfs_bmap_btree.c 	args.fsbno = cur->bc_tp->t_firstblock;
cur               208 fs/xfs/libxfs/xfs_bmap_btree.c 	xfs_rmap_ino_bmbt_owner(&args.oinfo, cur->bc_private.b.ip->i_ino,
cur               209 fs/xfs/libxfs/xfs_bmap_btree.c 			cur->bc_private.b.whichfork);
cur               226 fs/xfs/libxfs/xfs_bmap_btree.c 	} else if (cur->bc_tp->t_flags & XFS_TRANS_LOWMODE) {
cur               233 fs/xfs/libxfs/xfs_bmap_btree.c 	args.wasdel = cur->bc_private.b.flags & XFS_BTCUR_BPRV_WASDEL;
cur               253 fs/xfs/libxfs/xfs_bmap_btree.c 		cur->bc_tp->t_flags |= XFS_TRANS_LOWMODE;
cur               261 fs/xfs/libxfs/xfs_bmap_btree.c 	cur->bc_tp->t_firstblock = args.fsbno;
cur               262 fs/xfs/libxfs/xfs_bmap_btree.c 	cur->bc_private.b.allocated++;
cur               263 fs/xfs/libxfs/xfs_bmap_btree.c 	cur->bc_private.b.ip->i_d.di_nblocks++;
cur               264 fs/xfs/libxfs/xfs_bmap_btree.c 	xfs_trans_log_inode(args.tp, cur->bc_private.b.ip, XFS_ILOG_CORE);
cur               265 fs/xfs/libxfs/xfs_bmap_btree.c 	xfs_trans_mod_dquot_byino(args.tp, cur->bc_private.b.ip,
cur               279 fs/xfs/libxfs/xfs_bmap_btree.c 	struct xfs_btree_cur	*cur,
cur               282 fs/xfs/libxfs/xfs_bmap_btree.c 	struct xfs_mount	*mp = cur->bc_mp;
cur               283 fs/xfs/libxfs/xfs_bmap_btree.c 	struct xfs_inode	*ip = cur->bc_private.b.ip;
cur               284 fs/xfs/libxfs/xfs_bmap_btree.c 	struct xfs_trans	*tp = cur->bc_tp;
cur               288 fs/xfs/libxfs/xfs_bmap_btree.c 	xfs_rmap_ino_bmbt_owner(&oinfo, ip->i_ino, cur->bc_private.b.whichfork);
cur               289 fs/xfs/libxfs/xfs_bmap_btree.c 	xfs_bmap_add_free(cur->bc_tp, fsbno, 1, &oinfo);
cur               299 fs/xfs/libxfs/xfs_bmap_btree.c 	struct xfs_btree_cur	*cur,
cur               302 fs/xfs/libxfs/xfs_bmap_btree.c 	if (level == cur->bc_nlevels - 1) {
cur               305 fs/xfs/libxfs/xfs_bmap_btree.c 		ifp = XFS_IFORK_PTR(cur->bc_private.b.ip,
cur               306 fs/xfs/libxfs/xfs_bmap_btree.c 				    cur->bc_private.b.whichfork);
cur               308 fs/xfs/libxfs/xfs_bmap_btree.c 		return xfs_bmbt_maxrecs(cur->bc_mp,
cur               312 fs/xfs/libxfs/xfs_bmap_btree.c 	return cur->bc_mp->m_bmap_dmnr[level != 0];
cur               317 fs/xfs/libxfs/xfs_bmap_btree.c 	struct xfs_btree_cur	*cur,
cur               320 fs/xfs/libxfs/xfs_bmap_btree.c 	if (level == cur->bc_nlevels - 1) {
cur               323 fs/xfs/libxfs/xfs_bmap_btree.c 		ifp = XFS_IFORK_PTR(cur->bc_private.b.ip,
cur               324 fs/xfs/libxfs/xfs_bmap_btree.c 				    cur->bc_private.b.whichfork);
cur               326 fs/xfs/libxfs/xfs_bmap_btree.c 		return xfs_bmbt_maxrecs(cur->bc_mp,
cur               330 fs/xfs/libxfs/xfs_bmap_btree.c 	return cur->bc_mp->m_bmap_dmxr[level != 0];
cur               345 fs/xfs/libxfs/xfs_bmap_btree.c 	struct xfs_btree_cur	*cur,
cur               348 fs/xfs/libxfs/xfs_bmap_btree.c 	if (level != cur->bc_nlevels - 1)
cur               349 fs/xfs/libxfs/xfs_bmap_btree.c 		return cur->bc_mp->m_bmap_dmxr[level != 0];
cur               350 fs/xfs/libxfs/xfs_bmap_btree.c 	return xfs_bmdr_maxrecs(cur->bc_private.b.forksize, level == 0);
cur               374 fs/xfs/libxfs/xfs_bmap_btree.c 	struct xfs_btree_cur	*cur,
cur               377 fs/xfs/libxfs/xfs_bmap_btree.c 	xfs_bmbt_disk_set_all(&rec->bmbt, &cur->bc_rec.b);
cur               382 fs/xfs/libxfs/xfs_bmap_btree.c 	struct xfs_btree_cur	*cur,
cur               390 fs/xfs/libxfs/xfs_bmap_btree.c 	struct xfs_btree_cur	*cur,
cur               394 fs/xfs/libxfs/xfs_bmap_btree.c 				      cur->bc_rec.b.br_startoff;
cur               399 fs/xfs/libxfs/xfs_bmap_btree.c 	struct xfs_btree_cur	*cur,
cur               500 fs/xfs/libxfs/xfs_bmap_btree.c 	struct xfs_btree_cur	*cur,
cur               510 fs/xfs/libxfs/xfs_bmap_btree.c 	struct xfs_btree_cur	*cur,
cur               552 fs/xfs/libxfs/xfs_bmap_btree.c 	struct xfs_btree_cur	*cur;
cur               555 fs/xfs/libxfs/xfs_bmap_btree.c 	cur = kmem_zone_zalloc(xfs_btree_cur_zone, KM_NOFS);
cur               557 fs/xfs/libxfs/xfs_bmap_btree.c 	cur->bc_tp = tp;
cur               558 fs/xfs/libxfs/xfs_bmap_btree.c 	cur->bc_mp = mp;
cur               559 fs/xfs/libxfs/xfs_bmap_btree.c 	cur->bc_nlevels = be16_to_cpu(ifp->if_broot->bb_level) + 1;
cur               560 fs/xfs/libxfs/xfs_bmap_btree.c 	cur->bc_btnum = XFS_BTNUM_BMAP;
cur               561 fs/xfs/libxfs/xfs_bmap_btree.c 	cur->bc_blocklog = mp->m_sb.sb_blocklog;
cur               562 fs/xfs/libxfs/xfs_bmap_btree.c 	cur->bc_statoff = XFS_STATS_CALC_INDEX(xs_bmbt_2);
cur               564 fs/xfs/libxfs/xfs_bmap_btree.c 	cur->bc_ops = &xfs_bmbt_ops;
cur               565 fs/xfs/libxfs/xfs_bmap_btree.c 	cur->bc_flags = XFS_BTREE_LONG_PTRS | XFS_BTREE_ROOT_IN_INODE;
cur               567 fs/xfs/libxfs/xfs_bmap_btree.c 		cur->bc_flags |= XFS_BTREE_CRC_BLOCKS;
cur               569 fs/xfs/libxfs/xfs_bmap_btree.c 	cur->bc_private.b.forksize = XFS_IFORK_SIZE(ip, whichfork);
cur               570 fs/xfs/libxfs/xfs_bmap_btree.c 	cur->bc_private.b.ip = ip;
cur               571 fs/xfs/libxfs/xfs_bmap_btree.c 	cur->bc_private.b.allocated = 0;
cur               572 fs/xfs/libxfs/xfs_bmap_btree.c 	cur->bc_private.b.flags = 0;
cur               573 fs/xfs/libxfs/xfs_bmap_btree.c 	cur->bc_private.b.whichfork = whichfork;
cur               575 fs/xfs/libxfs/xfs_bmap_btree.c 	return cur;
cur               634 fs/xfs/libxfs/xfs_bmap_btree.c 	struct xfs_btree_cur	*cur;
cur               644 fs/xfs/libxfs/xfs_bmap_btree.c 	cur = xfs_bmbt_init_cursor(ip->i_mount, tp, ip, whichfork);
cur               645 fs/xfs/libxfs/xfs_bmap_btree.c 	if (!cur)
cur               647 fs/xfs/libxfs/xfs_bmap_btree.c 	cur->bc_private.b.flags |= XFS_BTCUR_BPRV_INVALID_OWNER;
cur               649 fs/xfs/libxfs/xfs_bmap_btree.c 	error = xfs_btree_change_owner(cur, new_owner, buffer_list);
cur               650 fs/xfs/libxfs/xfs_bmap_btree.c 	xfs_btree_del_cursor(cur, error);
cur                58 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur                63 fs/xfs/libxfs/xfs_btree.c 	struct xfs_mount	*mp = cur->bc_mp;
cur                64 fs/xfs/libxfs/xfs_btree.c 	xfs_btnum_t		btnum = cur->bc_btnum;
cur                82 fs/xfs/libxfs/xfs_btree.c 	    cur->bc_ops->get_maxrecs(cur, level))
cur                85 fs/xfs/libxfs/xfs_btree.c 	    !xfs_btree_check_lptr(cur, be64_to_cpu(block->bb_u.l.bb_leftsib),
cur                89 fs/xfs/libxfs/xfs_btree.c 	    !xfs_btree_check_lptr(cur, be64_to_cpu(block->bb_u.l.bb_rightsib),
cur                99 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur               104 fs/xfs/libxfs/xfs_btree.c 	struct xfs_mount	*mp = cur->bc_mp;
cur               107 fs/xfs/libxfs/xfs_btree.c 	fa = __xfs_btree_check_lblock(cur, block, level, bp);
cur               124 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur               129 fs/xfs/libxfs/xfs_btree.c 	struct xfs_mount	*mp = cur->bc_mp;
cur               130 fs/xfs/libxfs/xfs_btree.c 	xfs_btnum_t		btnum = cur->bc_btnum;
cur               146 fs/xfs/libxfs/xfs_btree.c 	    cur->bc_ops->get_maxrecs(cur, level))
cur               149 fs/xfs/libxfs/xfs_btree.c 	    !xfs_btree_check_sptr(cur, be32_to_cpu(block->bb_u.s.bb_leftsib),
cur               153 fs/xfs/libxfs/xfs_btree.c 	    !xfs_btree_check_sptr(cur, be32_to_cpu(block->bb_u.s.bb_rightsib),
cur               163 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur               168 fs/xfs/libxfs/xfs_btree.c 	struct xfs_mount	*mp = cur->bc_mp;
cur               171 fs/xfs/libxfs/xfs_btree.c 	fa = __xfs_btree_check_sblock(cur, block, level, bp);
cur               187 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,	/* btree cursor */
cur               192 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_flags & XFS_BTREE_LONG_PTRS)
cur               193 fs/xfs/libxfs/xfs_btree.c 		return xfs_btree_check_lblock(cur, block, level, bp);
cur               195 fs/xfs/libxfs/xfs_btree.c 		return xfs_btree_check_sblock(cur, block, level, bp);
cur               201 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur               207 fs/xfs/libxfs/xfs_btree.c 	return xfs_verify_fsbno(cur->bc_mp, fsbno);
cur               213 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur               219 fs/xfs/libxfs/xfs_btree.c 	return xfs_verify_agbno(cur->bc_mp, cur->bc_private.a.agno, agbno);
cur               228 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur               233 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_flags & XFS_BTREE_LONG_PTRS) {
cur               234 fs/xfs/libxfs/xfs_btree.c 		if (xfs_btree_check_lptr(cur, be64_to_cpu((&ptr->l)[index]),
cur               237 fs/xfs/libxfs/xfs_btree.c 		xfs_err(cur->bc_mp,
cur               239 fs/xfs/libxfs/xfs_btree.c 				cur->bc_private.b.ip->i_ino,
cur               240 fs/xfs/libxfs/xfs_btree.c 				cur->bc_private.b.whichfork, cur->bc_btnum,
cur               243 fs/xfs/libxfs/xfs_btree.c 		if (xfs_btree_check_sptr(cur, be32_to_cpu((&ptr->s)[index]),
cur               246 fs/xfs/libxfs/xfs_btree.c 		xfs_err(cur->bc_mp,
cur               248 fs/xfs/libxfs/xfs_btree.c 				cur->bc_private.a.agno, cur->bc_btnum,
cur               339 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur               344 fs/xfs/libxfs/xfs_btree.c 	error = cur->bc_ops->free_block(cur, bp);
cur               346 fs/xfs/libxfs/xfs_btree.c 		xfs_trans_binval(cur->bc_tp, bp);
cur               347 fs/xfs/libxfs/xfs_btree.c 		XFS_BTREE_STATS_INC(cur, free);
cur               357 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_cur_t	*cur,		/* btree cursor */
cur               372 fs/xfs/libxfs/xfs_btree.c 	for (i = 0; i < cur->bc_nlevels; i++) {
cur               373 fs/xfs/libxfs/xfs_btree.c 		if (cur->bc_bufs[i])
cur               374 fs/xfs/libxfs/xfs_btree.c 			xfs_trans_brelse(cur->bc_tp, cur->bc_bufs[i]);
cur               382 fs/xfs/libxfs/xfs_btree.c 	ASSERT(cur->bc_btnum != XFS_BTNUM_BMAP ||
cur               383 fs/xfs/libxfs/xfs_btree.c 	       cur->bc_private.b.allocated == 0);
cur               387 fs/xfs/libxfs/xfs_btree.c 	kmem_zone_free(xfs_btree_cur_zone, cur);
cur               396 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_cur_t	*cur,		/* input cursor */
cur               406 fs/xfs/libxfs/xfs_btree.c 	tp = cur->bc_tp;
cur               407 fs/xfs/libxfs/xfs_btree.c 	mp = cur->bc_mp;
cur               412 fs/xfs/libxfs/xfs_btree.c 	new = cur->bc_ops->dup_cursor(cur);
cur               417 fs/xfs/libxfs/xfs_btree.c 	new->bc_rec = cur->bc_rec;
cur               423 fs/xfs/libxfs/xfs_btree.c 		new->bc_ptrs[i] = cur->bc_ptrs[i];
cur               424 fs/xfs/libxfs/xfs_btree.c 		new->bc_ra[i] = cur->bc_ra[i];
cur               425 fs/xfs/libxfs/xfs_btree.c 		bp = cur->bc_bufs[i];
cur               430 fs/xfs/libxfs/xfs_btree.c 						   cur->bc_ops->buf_ops);
cur               520 fs/xfs/libxfs/xfs_btree.c static inline size_t xfs_btree_block_len(struct xfs_btree_cur *cur)
cur               522 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_flags & XFS_BTREE_LONG_PTRS) {
cur               523 fs/xfs/libxfs/xfs_btree.c 		if (cur->bc_flags & XFS_BTREE_CRC_BLOCKS)
cur               527 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_flags & XFS_BTREE_CRC_BLOCKS)
cur               535 fs/xfs/libxfs/xfs_btree.c static inline size_t xfs_btree_ptr_len(struct xfs_btree_cur *cur)
cur               537 fs/xfs/libxfs/xfs_btree.c 	return (cur->bc_flags & XFS_BTREE_LONG_PTRS) ?
cur               546 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur               549 fs/xfs/libxfs/xfs_btree.c 	return xfs_btree_block_len(cur) +
cur               550 fs/xfs/libxfs/xfs_btree.c 		(n - 1) * cur->bc_ops->rec_len;
cur               558 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur               561 fs/xfs/libxfs/xfs_btree.c 	return xfs_btree_block_len(cur) +
cur               562 fs/xfs/libxfs/xfs_btree.c 		(n - 1) * cur->bc_ops->key_len;
cur               570 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur               573 fs/xfs/libxfs/xfs_btree.c 	return xfs_btree_block_len(cur) +
cur               574 fs/xfs/libxfs/xfs_btree.c 		(n - 1) * cur->bc_ops->key_len + (cur->bc_ops->key_len / 2);
cur               582 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur               586 fs/xfs/libxfs/xfs_btree.c 	return xfs_btree_block_len(cur) +
cur               587 fs/xfs/libxfs/xfs_btree.c 		cur->bc_ops->get_maxrecs(cur, level) * cur->bc_ops->key_len +
cur               588 fs/xfs/libxfs/xfs_btree.c 		(n - 1) * xfs_btree_ptr_len(cur);
cur               596 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur               601 fs/xfs/libxfs/xfs_btree.c 		((char *)block + xfs_btree_rec_offset(cur, n));
cur               609 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur               614 fs/xfs/libxfs/xfs_btree.c 		((char *)block + xfs_btree_key_offset(cur, n));
cur               622 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur               627 fs/xfs/libxfs/xfs_btree.c 		((char *)block + xfs_btree_high_key_offset(cur, n));
cur               635 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur               644 fs/xfs/libxfs/xfs_btree.c 		((char *)block + xfs_btree_ptr_offset(cur, n, level));
cur               655 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur)
cur               659 fs/xfs/libxfs/xfs_btree.c 	ifp = XFS_IFORK_PTR(cur->bc_private.b.ip, cur->bc_private.b.whichfork);
cur               669 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,	/* btree cursor */
cur               673 fs/xfs/libxfs/xfs_btree.c 	if ((cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) &&
cur               674 fs/xfs/libxfs/xfs_btree.c 	    (level == cur->bc_nlevels - 1)) {
cur               676 fs/xfs/libxfs/xfs_btree.c 		return xfs_btree_get_iroot(cur);
cur               679 fs/xfs/libxfs/xfs_btree.c 	*bpp = cur->bc_bufs[level];
cur               724 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_cur_t		*cur,	/* btree cursor */
cur               730 fs/xfs/libxfs/xfs_btree.c 	block = xfs_btree_get_block(cur, level, &bp);
cur               731 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_check_block(cur, block, level, bp);
cur               732 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_flags & XFS_BTREE_LONG_PTRS)
cur               744 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_cur_t		*cur,	/* btree cursor */
cur               753 fs/xfs/libxfs/xfs_btree.c 	block = xfs_btree_get_block(cur, level, &bp);
cur               754 fs/xfs/libxfs/xfs_btree.c 	if (xfs_btree_check_block(cur, block, level, bp))
cur               764 fs/xfs/libxfs/xfs_btree.c 	cur->bc_ptrs[level] = 1;
cur               774 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_cur_t		*cur,	/* btree cursor */
cur               783 fs/xfs/libxfs/xfs_btree.c 	block = xfs_btree_get_block(cur, level, &bp);
cur               784 fs/xfs/libxfs/xfs_btree.c 	if (xfs_btree_check_block(cur, block, level, bp))
cur               794 fs/xfs/libxfs/xfs_btree.c 	cur->bc_ptrs[level] = be16_to_cpu(block->bb_numrecs);
cur               906 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur               915 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_reada_bufl(cur->bc_mp, left, 1,
cur               916 fs/xfs/libxfs/xfs_btree.c 				     cur->bc_ops->buf_ops);
cur               921 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_reada_bufl(cur->bc_mp, right, 1,
cur               922 fs/xfs/libxfs/xfs_btree.c 				     cur->bc_ops->buf_ops);
cur               931 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur               941 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_reada_bufs(cur->bc_mp, cur->bc_private.a.agno,
cur               942 fs/xfs/libxfs/xfs_btree.c 				     left, 1, cur->bc_ops->buf_ops);
cur               947 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_reada_bufs(cur->bc_mp, cur->bc_private.a.agno,
cur               948 fs/xfs/libxfs/xfs_btree.c 				     right, 1, cur->bc_ops->buf_ops);
cur               961 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,		/* btree cursor */
cur               971 fs/xfs/libxfs/xfs_btree.c 	if ((cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) &&
cur               972 fs/xfs/libxfs/xfs_btree.c 	    (lev == cur->bc_nlevels - 1))
cur               975 fs/xfs/libxfs/xfs_btree.c 	if ((cur->bc_ra[lev] | lr) == cur->bc_ra[lev])
cur               978 fs/xfs/libxfs/xfs_btree.c 	cur->bc_ra[lev] |= lr;
cur               979 fs/xfs/libxfs/xfs_btree.c 	block = XFS_BUF_TO_BLOCK(cur->bc_bufs[lev]);
cur               981 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_flags & XFS_BTREE_LONG_PTRS)
cur               982 fs/xfs/libxfs/xfs_btree.c 		return xfs_btree_readahead_lblock(cur, lr, block);
cur               983 fs/xfs/libxfs/xfs_btree.c 	return xfs_btree_readahead_sblock(cur, lr, block);
cur               988 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur               996 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_check_ptr(cur, ptr, 0, 1);
cur              1000 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_flags & XFS_BTREE_LONG_PTRS) {
cur              1002 fs/xfs/libxfs/xfs_btree.c 		*daddr = XFS_FSB_TO_DADDR(cur->bc_mp, fsbno);
cur              1005 fs/xfs/libxfs/xfs_btree.c 		*daddr = XFS_AGB_TO_DADDR(cur->bc_mp, cur->bc_private.a.agno,
cur              1020 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              1026 fs/xfs/libxfs/xfs_btree.c 	if (xfs_btree_ptr_to_daddr(cur, ptr, &daddr))
cur              1028 fs/xfs/libxfs/xfs_btree.c 	xfs_buf_readahead(cur->bc_mp->m_ddev_targp, daddr,
cur              1029 fs/xfs/libxfs/xfs_btree.c 			  cur->bc_mp->m_bsize * count, cur->bc_ops->buf_ops);
cur              1038 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_cur_t		*cur,	/* btree cursor */
cur              1044 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_bufs[lev])
cur              1045 fs/xfs/libxfs/xfs_btree.c 		xfs_trans_brelse(cur->bc_tp, cur->bc_bufs[lev]);
cur              1046 fs/xfs/libxfs/xfs_btree.c 	cur->bc_bufs[lev] = bp;
cur              1047 fs/xfs/libxfs/xfs_btree.c 	cur->bc_ra[lev] = 0;
cur              1050 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_flags & XFS_BTREE_LONG_PTRS) {
cur              1052 fs/xfs/libxfs/xfs_btree.c 			cur->bc_ra[lev] |= XFS_BTCUR_LEFTRA;
cur              1054 fs/xfs/libxfs/xfs_btree.c 			cur->bc_ra[lev] |= XFS_BTCUR_RIGHTRA;
cur              1057 fs/xfs/libxfs/xfs_btree.c 			cur->bc_ra[lev] |= XFS_BTCUR_LEFTRA;
cur              1059 fs/xfs/libxfs/xfs_btree.c 			cur->bc_ra[lev] |= XFS_BTCUR_RIGHTRA;
cur              1065 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              1068 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_flags & XFS_BTREE_LONG_PTRS)
cur              1076 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              1079 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_flags & XFS_BTREE_LONG_PTRS)
cur              1090 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              1097 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_flags & XFS_BTREE_LONG_PTRS) {
cur              1112 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              1119 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_flags & XFS_BTREE_LONG_PTRS) {
cur              1190 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              1203 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_flags & XFS_BTREE_LONG_PTRS)
cur              1204 fs/xfs/libxfs/xfs_btree.c 		owner = cur->bc_private.b.ip->i_ino;
cur              1206 fs/xfs/libxfs/xfs_btree.c 		owner = cur->bc_private.a.agno;
cur              1208 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_init_block_int(cur->bc_mp, XFS_BUF_TO_BLOCK(bp), bp->b_bn,
cur              1209 fs/xfs/libxfs/xfs_btree.c 				 cur->bc_btnum, level, numrecs,
cur              1210 fs/xfs/libxfs/xfs_btree.c 				 owner, cur->bc_flags);
cur              1220 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              1228 fs/xfs/libxfs/xfs_btree.c 	if (!(cur->bc_flags & XFS_BTREE_LASTREC_UPDATE))
cur              1231 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_get_sibling(cur, block, &ptr, XFS_BB_RIGHTSIB);
cur              1232 fs/xfs/libxfs/xfs_btree.c 	if (!xfs_btree_ptr_is_null(cur, &ptr))
cur              1239 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              1243 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_flags & XFS_BTREE_LONG_PTRS)
cur              1244 fs/xfs/libxfs/xfs_btree.c 		ptr->l = cpu_to_be64(XFS_DADDR_TO_FSB(cur->bc_mp,
cur              1247 fs/xfs/libxfs/xfs_btree.c 		ptr->s = cpu_to_be32(xfs_daddr_to_agbno(cur->bc_mp,
cur              1254 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              1257 fs/xfs/libxfs/xfs_btree.c 	switch (cur->bc_btnum) {
cur              1282 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              1287 fs/xfs/libxfs/xfs_btree.c 	struct xfs_mount	*mp = cur->bc_mp;
cur              1291 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_ptr_to_daddr(cur, ptr, &d);
cur              1294 fs/xfs/libxfs/xfs_btree.c 	*bpp = xfs_trans_get_buf(cur->bc_tp, mp->m_ddev_targp, d,
cur              1300 fs/xfs/libxfs/xfs_btree.c 	(*bpp)->b_ops = cur->bc_ops->buf_ops;
cur              1311 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              1317 fs/xfs/libxfs/xfs_btree.c 	struct xfs_mount	*mp = cur->bc_mp;
cur              1324 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_ptr_to_daddr(cur, ptr, &d);
cur              1327 fs/xfs/libxfs/xfs_btree.c 	error = xfs_trans_read_buf(mp, cur->bc_tp, mp->m_ddev_targp, d,
cur              1329 fs/xfs/libxfs/xfs_btree.c 				   cur->bc_ops->buf_ops);
cur              1333 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_set_refs(cur, *bpp);
cur              1343 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              1349 fs/xfs/libxfs/xfs_btree.c 	memcpy(dst_key, src_key, numkeys * cur->bc_ops->key_len);
cur              1357 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              1363 fs/xfs/libxfs/xfs_btree.c 	memcpy(dst_rec, src_rec, numrecs * cur->bc_ops->rec_len);
cur              1371 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              1377 fs/xfs/libxfs/xfs_btree.c 	memcpy(dst_ptr, src_ptr, numptrs * xfs_btree_ptr_len(cur));
cur              1385 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              1395 fs/xfs/libxfs/xfs_btree.c 	dst_key = (char *)key + (dir * cur->bc_ops->key_len);
cur              1396 fs/xfs/libxfs/xfs_btree.c 	memmove(dst_key, key, numkeys * cur->bc_ops->key_len);
cur              1404 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              1414 fs/xfs/libxfs/xfs_btree.c 	dst_rec = (char *)rec + (dir * cur->bc_ops->rec_len);
cur              1415 fs/xfs/libxfs/xfs_btree.c 	memmove(dst_rec, rec, numrecs * cur->bc_ops->rec_len);
cur              1423 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              1433 fs/xfs/libxfs/xfs_btree.c 	dst_ptr = (char *)ptr + (dir * xfs_btree_ptr_len(cur));
cur              1434 fs/xfs/libxfs/xfs_btree.c 	memmove(dst_ptr, ptr, numptrs * xfs_btree_ptr_len(cur));
cur              1442 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              1449 fs/xfs/libxfs/xfs_btree.c 		xfs_trans_buf_set_type(cur->bc_tp, bp, XFS_BLFT_BTREE_BUF);
cur              1450 fs/xfs/libxfs/xfs_btree.c 		xfs_trans_log_buf(cur->bc_tp, bp,
cur              1451 fs/xfs/libxfs/xfs_btree.c 				  xfs_btree_key_offset(cur, first),
cur              1452 fs/xfs/libxfs/xfs_btree.c 				  xfs_btree_key_offset(cur, last + 1) - 1);
cur              1454 fs/xfs/libxfs/xfs_btree.c 		xfs_trans_log_inode(cur->bc_tp, cur->bc_private.b.ip,
cur              1455 fs/xfs/libxfs/xfs_btree.c 				xfs_ilog_fbroot(cur->bc_private.b.whichfork));
cur              1464 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              1470 fs/xfs/libxfs/xfs_btree.c 	xfs_trans_buf_set_type(cur->bc_tp, bp, XFS_BLFT_BTREE_BUF);
cur              1471 fs/xfs/libxfs/xfs_btree.c 	xfs_trans_log_buf(cur->bc_tp, bp,
cur              1472 fs/xfs/libxfs/xfs_btree.c 			  xfs_btree_rec_offset(cur, first),
cur              1473 fs/xfs/libxfs/xfs_btree.c 			  xfs_btree_rec_offset(cur, last + 1) - 1);
cur              1482 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,	/* btree cursor */
cur              1492 fs/xfs/libxfs/xfs_btree.c 		xfs_trans_buf_set_type(cur->bc_tp, bp, XFS_BLFT_BTREE_BUF);
cur              1493 fs/xfs/libxfs/xfs_btree.c 		xfs_trans_log_buf(cur->bc_tp, bp,
cur              1494 fs/xfs/libxfs/xfs_btree.c 				xfs_btree_ptr_offset(cur, first, level),
cur              1495 fs/xfs/libxfs/xfs_btree.c 				xfs_btree_ptr_offset(cur, last + 1, level) - 1);
cur              1497 fs/xfs/libxfs/xfs_btree.c 		xfs_trans_log_inode(cur->bc_tp, cur->bc_private.b.ip,
cur              1498 fs/xfs/libxfs/xfs_btree.c 			xfs_ilog_fbroot(cur->bc_private.b.whichfork));
cur              1508 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,	/* btree cursor */
cur              1545 fs/xfs/libxfs/xfs_btree.c 		if (cur->bc_flags & XFS_BTREE_CRC_BLOCKS) {
cur              1560 fs/xfs/libxfs/xfs_btree.c 				  (cur->bc_flags & XFS_BTREE_LONG_PTRS) ?
cur              1563 fs/xfs/libxfs/xfs_btree.c 		xfs_trans_buf_set_type(cur->bc_tp, bp, XFS_BLFT_BTREE_BUF);
cur              1564 fs/xfs/libxfs/xfs_btree.c 		xfs_trans_log_buf(cur->bc_tp, bp, first, last);
cur              1566 fs/xfs/libxfs/xfs_btree.c 		xfs_trans_log_inode(cur->bc_tp, cur->bc_private.b.ip,
cur              1567 fs/xfs/libxfs/xfs_btree.c 			xfs_ilog_fbroot(cur->bc_private.b.whichfork));
cur              1577 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              1587 fs/xfs/libxfs/xfs_btree.c 	ASSERT(level < cur->bc_nlevels);
cur              1590 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_readahead(cur, level, XFS_BTCUR_RIGHTRA);
cur              1593 fs/xfs/libxfs/xfs_btree.c 	block = xfs_btree_get_block(cur, level, &bp);
cur              1596 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_check_block(cur, block, level, bp);
cur              1602 fs/xfs/libxfs/xfs_btree.c 	if (++cur->bc_ptrs[level] <= xfs_btree_get_numrecs(block))
cur              1606 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_get_sibling(cur, block, &ptr, XFS_BB_RIGHTSIB);
cur              1607 fs/xfs/libxfs/xfs_btree.c 	if (xfs_btree_ptr_is_null(cur, &ptr))
cur              1610 fs/xfs/libxfs/xfs_btree.c 	XFS_BTREE_STATS_INC(cur, increment);
cur              1616 fs/xfs/libxfs/xfs_btree.c 	for (lev = level + 1; lev < cur->bc_nlevels; lev++) {
cur              1617 fs/xfs/libxfs/xfs_btree.c 		block = xfs_btree_get_block(cur, lev, &bp);
cur              1620 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_check_block(cur, block, lev, bp);
cur              1625 fs/xfs/libxfs/xfs_btree.c 		if (++cur->bc_ptrs[lev] <= xfs_btree_get_numrecs(block))
cur              1629 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_readahead(cur, lev, XFS_BTCUR_RIGHTRA);
cur              1636 fs/xfs/libxfs/xfs_btree.c 	if (lev == cur->bc_nlevels) {
cur              1637 fs/xfs/libxfs/xfs_btree.c 		if (cur->bc_flags & XFS_BTREE_ROOT_IN_INODE)
cur              1643 fs/xfs/libxfs/xfs_btree.c 	ASSERT(lev < cur->bc_nlevels);
cur              1649 fs/xfs/libxfs/xfs_btree.c 	for (block = xfs_btree_get_block(cur, lev, &bp); lev > level; ) {
cur              1652 fs/xfs/libxfs/xfs_btree.c 		ptrp = xfs_btree_ptr_addr(cur, cur->bc_ptrs[lev], block);
cur              1654 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_read_buf_block(cur, ptrp, 0, &block, &bp);
cur              1658 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_setbuf(cur, lev, bp);
cur              1659 fs/xfs/libxfs/xfs_btree.c 		cur->bc_ptrs[lev] = 1;
cur              1679 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              1689 fs/xfs/libxfs/xfs_btree.c 	ASSERT(level < cur->bc_nlevels);
cur              1692 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_readahead(cur, level, XFS_BTCUR_LEFTRA);
cur              1695 fs/xfs/libxfs/xfs_btree.c 	if (--cur->bc_ptrs[level] > 0)
cur              1699 fs/xfs/libxfs/xfs_btree.c 	block = xfs_btree_get_block(cur, level, &bp);
cur              1702 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_check_block(cur, block, level, bp);
cur              1708 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_get_sibling(cur, block, &ptr, XFS_BB_LEFTSIB);
cur              1709 fs/xfs/libxfs/xfs_btree.c 	if (xfs_btree_ptr_is_null(cur, &ptr))
cur              1712 fs/xfs/libxfs/xfs_btree.c 	XFS_BTREE_STATS_INC(cur, decrement);
cur              1718 fs/xfs/libxfs/xfs_btree.c 	for (lev = level + 1; lev < cur->bc_nlevels; lev++) {
cur              1719 fs/xfs/libxfs/xfs_btree.c 		if (--cur->bc_ptrs[lev] > 0)
cur              1722 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_readahead(cur, lev, XFS_BTCUR_LEFTRA);
cur              1729 fs/xfs/libxfs/xfs_btree.c 	if (lev == cur->bc_nlevels) {
cur              1730 fs/xfs/libxfs/xfs_btree.c 		if (cur->bc_flags & XFS_BTREE_ROOT_IN_INODE)
cur              1736 fs/xfs/libxfs/xfs_btree.c 	ASSERT(lev < cur->bc_nlevels);
cur              1742 fs/xfs/libxfs/xfs_btree.c 	for (block = xfs_btree_get_block(cur, lev, &bp); lev > level; ) {
cur              1745 fs/xfs/libxfs/xfs_btree.c 		ptrp = xfs_btree_ptr_addr(cur, cur->bc_ptrs[lev], block);
cur              1747 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_read_buf_block(cur, ptrp, 0, &block, &bp);
cur              1750 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_setbuf(cur, lev, bp);
cur              1751 fs/xfs/libxfs/xfs_btree.c 		cur->bc_ptrs[lev] = xfs_btree_get_numrecs(block);
cur              1767 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,	/* btree cursor */
cur              1777 fs/xfs/libxfs/xfs_btree.c 	if ((cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) &&
cur              1778 fs/xfs/libxfs/xfs_btree.c 	    (level == cur->bc_nlevels - 1)) {
cur              1779 fs/xfs/libxfs/xfs_btree.c 		*blkp = xfs_btree_get_iroot(cur);
cur              1789 fs/xfs/libxfs/xfs_btree.c 	bp = cur->bc_bufs[level];
cur              1790 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_ptr_to_daddr(cur, pp, &daddr);
cur              1798 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_read_buf_block(cur, pp, 0, blkp, &bp);
cur              1803 fs/xfs/libxfs/xfs_btree.c 	if (xfs_sb_version_hascrc(&cur->bc_mp->m_sb) &&
cur              1804 fs/xfs/libxfs/xfs_btree.c 	    !(cur->bc_private.b.flags & XFS_BTCUR_BPRV_INVALID_OWNER) &&
cur              1805 fs/xfs/libxfs/xfs_btree.c 	    (cur->bc_flags & XFS_BTREE_LONG_PTRS) &&
cur              1807 fs/xfs/libxfs/xfs_btree.c 			cur->bc_private.b.ip->i_ino)
cur              1818 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_setbuf(cur, level, bp);
cur              1823 fs/xfs/libxfs/xfs_btree.c 	xfs_trans_brelse(cur->bc_tp, bp);
cur              1834 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              1841 fs/xfs/libxfs/xfs_btree.c 		cur->bc_ops->init_key_from_rec(kp,
cur              1842 fs/xfs/libxfs/xfs_btree.c 				xfs_btree_rec_addr(cur, keyno, block));
cur              1846 fs/xfs/libxfs/xfs_btree.c 	return xfs_btree_key_addr(cur, keyno, block);
cur              1855 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,	/* btree cursor */
cur              1867 fs/xfs/libxfs/xfs_btree.c 	XFS_BTREE_STATS_INC(cur, lookup);
cur              1870 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_nlevels == 0)
cur              1877 fs/xfs/libxfs/xfs_btree.c 	cur->bc_ops->init_ptr_from_cur(cur, &ptr);
cur              1886 fs/xfs/libxfs/xfs_btree.c 	for (level = cur->bc_nlevels - 1, diff = 1; level >= 0; level--) {
cur              1888 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_lookup_get_block(cur, level, pp, &block);
cur              1909 fs/xfs/libxfs/xfs_btree.c 				if (level != 0 || cur->bc_nlevels != 1) {
cur              1912 fs/xfs/libxfs/xfs_btree.c 							cur->bc_mp, block,
cur              1917 fs/xfs/libxfs/xfs_btree.c 				cur->bc_ptrs[0] = dir != XFS_LOOKUP_LE;
cur              1927 fs/xfs/libxfs/xfs_btree.c 				XFS_BTREE_STATS_INC(cur, compare);
cur              1933 fs/xfs/libxfs/xfs_btree.c 				kp = xfs_lookup_get_search_key(cur, level,
cur              1942 fs/xfs/libxfs/xfs_btree.c 				diff = cur->bc_ops->key_diff(cur, kp);
cur              1963 fs/xfs/libxfs/xfs_btree.c 			pp = xfs_btree_ptr_addr(cur, keyno, block);
cur              1965 fs/xfs/libxfs/xfs_btree.c 			error = xfs_btree_debug_check_ptr(cur, pp, 0, level);
cur              1969 fs/xfs/libxfs/xfs_btree.c 			cur->bc_ptrs[level] = keyno;
cur              1980 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_get_sibling(cur, block, &ptr, XFS_BB_RIGHTSIB);
cur              1983 fs/xfs/libxfs/xfs_btree.c 		    !xfs_btree_ptr_is_null(cur, &ptr)) {
cur              1986 fs/xfs/libxfs/xfs_btree.c 			cur->bc_ptrs[0] = keyno;
cur              1987 fs/xfs/libxfs/xfs_btree.c 			error = xfs_btree_increment(cur, 0, &i);
cur              1990 fs/xfs/libxfs/xfs_btree.c 			XFS_WANT_CORRUPTED_RETURN(cur->bc_mp, i == 1);
cur              1996 fs/xfs/libxfs/xfs_btree.c 	cur->bc_ptrs[0] = keyno;
cur              2014 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              2017 fs/xfs/libxfs/xfs_btree.c 	ASSERT(cur->bc_flags & XFS_BTREE_OVERLAPPING);
cur              2019 fs/xfs/libxfs/xfs_btree.c 			(cur->bc_ops->key_len / 2));
cur              2025 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              2035 fs/xfs/libxfs/xfs_btree.c 	rec = xfs_btree_rec_addr(cur, 1, block);
cur              2036 fs/xfs/libxfs/xfs_btree.c 	cur->bc_ops->init_key_from_rec(key, rec);
cur              2038 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_flags & XFS_BTREE_OVERLAPPING) {
cur              2040 fs/xfs/libxfs/xfs_btree.c 		cur->bc_ops->init_high_key_from_rec(&max_hkey, rec);
cur              2042 fs/xfs/libxfs/xfs_btree.c 			rec = xfs_btree_rec_addr(cur, n, block);
cur              2043 fs/xfs/libxfs/xfs_btree.c 			cur->bc_ops->init_high_key_from_rec(&hkey, rec);
cur              2044 fs/xfs/libxfs/xfs_btree.c 			if (cur->bc_ops->diff_two_keys(cur, &hkey, &max_hkey)
cur              2049 fs/xfs/libxfs/xfs_btree.c 		high = xfs_btree_high_key_from_key(cur, key);
cur              2050 fs/xfs/libxfs/xfs_btree.c 		memcpy(high, &max_hkey, cur->bc_ops->key_len / 2);
cur              2057 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              2066 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_flags & XFS_BTREE_OVERLAPPING) {
cur              2067 fs/xfs/libxfs/xfs_btree.c 		memcpy(key, xfs_btree_key_addr(cur, 1, block),
cur              2068 fs/xfs/libxfs/xfs_btree.c 				cur->bc_ops->key_len / 2);
cur              2070 fs/xfs/libxfs/xfs_btree.c 		max_hkey = xfs_btree_high_key_addr(cur, 1, block);
cur              2072 fs/xfs/libxfs/xfs_btree.c 			hkey = xfs_btree_high_key_addr(cur, n, block);
cur              2073 fs/xfs/libxfs/xfs_btree.c 			if (cur->bc_ops->diff_two_keys(cur, hkey, max_hkey) > 0)
cur              2077 fs/xfs/libxfs/xfs_btree.c 		high = xfs_btree_high_key_from_key(cur, key);
cur              2078 fs/xfs/libxfs/xfs_btree.c 		memcpy(high, max_hkey, cur->bc_ops->key_len / 2);
cur              2080 fs/xfs/libxfs/xfs_btree.c 		memcpy(key, xfs_btree_key_addr(cur, 1, block),
cur              2081 fs/xfs/libxfs/xfs_btree.c 				cur->bc_ops->key_len);
cur              2088 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              2093 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_get_leaf_keys(cur, block, key);
cur              2095 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_get_node_keys(cur, block, key);
cur              2107 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              2110 fs/xfs/libxfs/xfs_btree.c 	return (cur->bc_flags & XFS_BTREE_OVERLAPPING) || ptr == 1;
cur              2120 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              2134 fs/xfs/libxfs/xfs_btree.c 	ASSERT(cur->bc_flags & XFS_BTREE_OVERLAPPING);
cur              2137 fs/xfs/libxfs/xfs_btree.c 	if (level + 1 >= cur->bc_nlevels)
cur              2140 fs/xfs/libxfs/xfs_btree.c 	trace_xfs_btree_updkeys(cur, level, bp0);
cur              2143 fs/xfs/libxfs/xfs_btree.c 	hkey = xfs_btree_high_key_from_key(cur, lkey);
cur              2144 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_get_keys(cur, block, lkey);
cur              2145 fs/xfs/libxfs/xfs_btree.c 	for (level++; level < cur->bc_nlevels; level++) {
cur              2149 fs/xfs/libxfs/xfs_btree.c 		block = xfs_btree_get_block(cur, level, &bp);
cur              2150 fs/xfs/libxfs/xfs_btree.c 		trace_xfs_btree_updkeys(cur, level, bp);
cur              2152 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_check_block(cur, block, level, bp);
cur              2156 fs/xfs/libxfs/xfs_btree.c 		ptr = cur->bc_ptrs[level];
cur              2157 fs/xfs/libxfs/xfs_btree.c 		nlkey = xfs_btree_key_addr(cur, ptr, block);
cur              2158 fs/xfs/libxfs/xfs_btree.c 		nhkey = xfs_btree_high_key_addr(cur, ptr, block);
cur              2160 fs/xfs/libxfs/xfs_btree.c 		    !(cur->bc_ops->diff_two_keys(cur, nlkey, lkey) != 0 ||
cur              2161 fs/xfs/libxfs/xfs_btree.c 		      cur->bc_ops->diff_two_keys(cur, nhkey, hkey) != 0))
cur              2163 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_copy_keys(cur, nlkey, lkey, 1);
cur              2164 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_log_keys(cur, bp, ptr, ptr);
cur              2165 fs/xfs/libxfs/xfs_btree.c 		if (level + 1 >= cur->bc_nlevels)
cur              2167 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_get_node_keys(cur, block, lkey);
cur              2176 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              2182 fs/xfs/libxfs/xfs_btree.c 	block = xfs_btree_get_block(cur, level, &bp);
cur              2183 fs/xfs/libxfs/xfs_btree.c 	return __xfs_btree_updkeys(cur, level, block, bp, true);
cur              2191 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              2202 fs/xfs/libxfs/xfs_btree.c 	block = xfs_btree_get_block(cur, level, &bp);
cur              2203 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_flags & XFS_BTREE_OVERLAPPING)
cur              2204 fs/xfs/libxfs/xfs_btree.c 		return __xfs_btree_updkeys(cur, level, block, bp, false);
cur              2212 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_get_keys(cur, block, &key);
cur              2213 fs/xfs/libxfs/xfs_btree.c 	for (level++, ptr = 1; ptr == 1 && level < cur->bc_nlevels; level++) {
cur              2217 fs/xfs/libxfs/xfs_btree.c 		block = xfs_btree_get_block(cur, level, &bp);
cur              2219 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_check_block(cur, block, level, bp);
cur              2223 fs/xfs/libxfs/xfs_btree.c 		ptr = cur->bc_ptrs[level];
cur              2224 fs/xfs/libxfs/xfs_btree.c 		kp = xfs_btree_key_addr(cur, ptr, block);
cur              2225 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_copy_keys(cur, kp, &key, 1);
cur              2226 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_log_keys(cur, bp, ptr, ptr);
cur              2239 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              2249 fs/xfs/libxfs/xfs_btree.c 	block = xfs_btree_get_block(cur, 0, &bp);
cur              2252 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_check_block(cur, block, 0, bp);
cur              2257 fs/xfs/libxfs/xfs_btree.c 	ptr = cur->bc_ptrs[0];
cur              2258 fs/xfs/libxfs/xfs_btree.c 	rp = xfs_btree_rec_addr(cur, ptr, block);
cur              2261 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_copy_recs(cur, rp, rec, 1);
cur              2262 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_log_recs(cur, bp, ptr, ptr);
cur              2268 fs/xfs/libxfs/xfs_btree.c 	if (xfs_btree_is_lastrec(cur, block, 0)) {
cur              2269 fs/xfs/libxfs/xfs_btree.c 		cur->bc_ops->update_lastrec(cur, block, rec,
cur              2274 fs/xfs/libxfs/xfs_btree.c 	if (xfs_btree_needs_key_update(cur, ptr)) {
cur              2275 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_update_keys(cur, 0);
cur              2292 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              2310 fs/xfs/libxfs/xfs_btree.c 	if ((cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) &&
cur              2311 fs/xfs/libxfs/xfs_btree.c 	    level == cur->bc_nlevels - 1)
cur              2315 fs/xfs/libxfs/xfs_btree.c 	right = xfs_btree_get_block(cur, level, &rbp);
cur              2318 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_check_block(cur, right, level, rbp);
cur              2324 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_get_sibling(cur, right, &lptr, XFS_BB_LEFTSIB);
cur              2325 fs/xfs/libxfs/xfs_btree.c 	if (xfs_btree_ptr_is_null(cur, &lptr))
cur              2332 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_ptrs[level] <= 1)
cur              2336 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_read_buf_block(cur, &lptr, 0, &left, &lbp);
cur              2342 fs/xfs/libxfs/xfs_btree.c 	if (lrecs == cur->bc_ops->get_maxrecs(cur, level))
cur              2355 fs/xfs/libxfs/xfs_btree.c 	XFS_BTREE_STATS_INC(cur, lshift);
cur              2356 fs/xfs/libxfs/xfs_btree.c 	XFS_BTREE_STATS_ADD(cur, moves, 1);
cur              2367 fs/xfs/libxfs/xfs_btree.c 		lkp = xfs_btree_key_addr(cur, lrecs, left);
cur              2368 fs/xfs/libxfs/xfs_btree.c 		rkp = xfs_btree_key_addr(cur, 1, right);
cur              2370 fs/xfs/libxfs/xfs_btree.c 		lpp = xfs_btree_ptr_addr(cur, lrecs, left);
cur              2371 fs/xfs/libxfs/xfs_btree.c 		rpp = xfs_btree_ptr_addr(cur, 1, right);
cur              2373 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_debug_check_ptr(cur, rpp, 0, level);
cur              2377 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_copy_keys(cur, lkp, rkp, 1);
cur              2378 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_copy_ptrs(cur, lpp, rpp, 1);
cur              2380 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_log_keys(cur, lbp, lrecs, lrecs);
cur              2381 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_log_ptrs(cur, lbp, lrecs, lrecs);
cur              2383 fs/xfs/libxfs/xfs_btree.c 		ASSERT(cur->bc_ops->keys_inorder(cur,
cur              2384 fs/xfs/libxfs/xfs_btree.c 			xfs_btree_key_addr(cur, lrecs - 1, left), lkp));
cur              2389 fs/xfs/libxfs/xfs_btree.c 		lrp = xfs_btree_rec_addr(cur, lrecs, left);
cur              2390 fs/xfs/libxfs/xfs_btree.c 		rrp = xfs_btree_rec_addr(cur, 1, right);
cur              2392 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_copy_recs(cur, lrp, rrp, 1);
cur              2393 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_log_recs(cur, lbp, lrecs, lrecs);
cur              2395 fs/xfs/libxfs/xfs_btree.c 		ASSERT(cur->bc_ops->recs_inorder(cur,
cur              2396 fs/xfs/libxfs/xfs_btree.c 			xfs_btree_rec_addr(cur, lrecs - 1, left), lrp));
cur              2400 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_log_block(cur, lbp, XFS_BB_NUMRECS);
cur              2403 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_log_block(cur, rbp, XFS_BB_NUMRECS);
cur              2408 fs/xfs/libxfs/xfs_btree.c 	XFS_BTREE_STATS_ADD(cur, moves, rrecs - 1);
cur              2414 fs/xfs/libxfs/xfs_btree.c 			error = xfs_btree_debug_check_ptr(cur, rpp, i + 1, level);
cur              2419 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_shift_keys(cur,
cur              2420 fs/xfs/libxfs/xfs_btree.c 				xfs_btree_key_addr(cur, 2, right),
cur              2422 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_shift_ptrs(cur,
cur              2423 fs/xfs/libxfs/xfs_btree.c 				xfs_btree_ptr_addr(cur, 2, right),
cur              2426 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_log_keys(cur, rbp, 1, rrecs);
cur              2427 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_log_ptrs(cur, rbp, 1, rrecs);
cur              2430 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_shift_recs(cur,
cur              2431 fs/xfs/libxfs/xfs_btree.c 			xfs_btree_rec_addr(cur, 2, right),
cur              2433 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_log_recs(cur, rbp, 1, rrecs);
cur              2440 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_flags & XFS_BTREE_OVERLAPPING) {
cur              2441 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_dup_cursor(cur, &tcur);
cur              2460 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_update_keys(cur, level);
cur              2465 fs/xfs/libxfs/xfs_btree.c 	cur->bc_ptrs[level]--;
cur              2488 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              2504 fs/xfs/libxfs/xfs_btree.c 	if ((cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) &&
cur              2505 fs/xfs/libxfs/xfs_btree.c 	    (level == cur->bc_nlevels - 1))
cur              2509 fs/xfs/libxfs/xfs_btree.c 	left = xfs_btree_get_block(cur, level, &lbp);
cur              2512 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_check_block(cur, left, level, lbp);
cur              2518 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_get_sibling(cur, left, &rptr, XFS_BB_RIGHTSIB);
cur              2519 fs/xfs/libxfs/xfs_btree.c 	if (xfs_btree_ptr_is_null(cur, &rptr))
cur              2527 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_ptrs[level] >= lrecs)
cur              2531 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_read_buf_block(cur, &rptr, 0, &right, &rbp);
cur              2537 fs/xfs/libxfs/xfs_btree.c 	if (rrecs == cur->bc_ops->get_maxrecs(cur, level))
cur              2540 fs/xfs/libxfs/xfs_btree.c 	XFS_BTREE_STATS_INC(cur, rshift);
cur              2541 fs/xfs/libxfs/xfs_btree.c 	XFS_BTREE_STATS_ADD(cur, moves, rrecs);
cur              2553 fs/xfs/libxfs/xfs_btree.c 		lkp = xfs_btree_key_addr(cur, lrecs, left);
cur              2554 fs/xfs/libxfs/xfs_btree.c 		lpp = xfs_btree_ptr_addr(cur, lrecs, left);
cur              2555 fs/xfs/libxfs/xfs_btree.c 		rkp = xfs_btree_key_addr(cur, 1, right);
cur              2556 fs/xfs/libxfs/xfs_btree.c 		rpp = xfs_btree_ptr_addr(cur, 1, right);
cur              2559 fs/xfs/libxfs/xfs_btree.c 			error = xfs_btree_debug_check_ptr(cur, rpp, i, level);
cur              2564 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_shift_keys(cur, rkp, 1, rrecs);
cur              2565 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_shift_ptrs(cur, rpp, 1, rrecs);
cur              2567 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_debug_check_ptr(cur, lpp, 0, level);
cur              2572 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_copy_keys(cur, rkp, lkp, 1);
cur              2573 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_copy_ptrs(cur, rpp, lpp, 1);
cur              2575 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_log_keys(cur, rbp, 1, rrecs + 1);
cur              2576 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_log_ptrs(cur, rbp, 1, rrecs + 1);
cur              2578 fs/xfs/libxfs/xfs_btree.c 		ASSERT(cur->bc_ops->keys_inorder(cur, rkp,
cur              2579 fs/xfs/libxfs/xfs_btree.c 			xfs_btree_key_addr(cur, 2, right)));
cur              2585 fs/xfs/libxfs/xfs_btree.c 		lrp = xfs_btree_rec_addr(cur, lrecs, left);
cur              2586 fs/xfs/libxfs/xfs_btree.c 		rrp = xfs_btree_rec_addr(cur, 1, right);
cur              2588 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_shift_recs(cur, rrp, 1, rrecs);
cur              2591 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_copy_recs(cur, rrp, lrp, 1);
cur              2592 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_log_recs(cur, rbp, 1, rrecs + 1);
cur              2599 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_log_block(cur, lbp, XFS_BB_NUMRECS);
cur              2602 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_log_block(cur, rbp, XFS_BB_NUMRECS);
cur              2608 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_dup_cursor(cur, &tcur);
cur              2619 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_flags & XFS_BTREE_OVERLAPPING) {
cur              2620 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_update_keys(cur, level);
cur              2654 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              2676 fs/xfs/libxfs/xfs_btree.c 	XFS_BTREE_STATS_INC(cur, split);
cur              2679 fs/xfs/libxfs/xfs_btree.c 	left = xfs_btree_get_block(cur, level, &lbp);
cur              2682 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_check_block(cur, left, level, lbp);
cur              2687 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_buf_to_ptr(cur, lbp, &lptr);
cur              2690 fs/xfs/libxfs/xfs_btree.c 	error = cur->bc_ops->alloc_block(cur, &lptr, &rptr, stat);
cur              2695 fs/xfs/libxfs/xfs_btree.c 	XFS_BTREE_STATS_INC(cur, alloc);
cur              2698 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_get_buf_block(cur, &rptr, &right, &rbp);
cur              2703 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_init_block_cur(cur, rbp, xfs_btree_get_level(left), 0);
cur              2712 fs/xfs/libxfs/xfs_btree.c 	if ((lrecs & 1) && cur->bc_ptrs[level] <= rrecs + 1)
cur              2716 fs/xfs/libxfs/xfs_btree.c 	XFS_BTREE_STATS_ADD(cur, moves, rrecs);
cur              2735 fs/xfs/libxfs/xfs_btree.c 		lkp = xfs_btree_key_addr(cur, src_index, left);
cur              2736 fs/xfs/libxfs/xfs_btree.c 		lpp = xfs_btree_ptr_addr(cur, src_index, left);
cur              2737 fs/xfs/libxfs/xfs_btree.c 		rkp = xfs_btree_key_addr(cur, 1, right);
cur              2738 fs/xfs/libxfs/xfs_btree.c 		rpp = xfs_btree_ptr_addr(cur, 1, right);
cur              2741 fs/xfs/libxfs/xfs_btree.c 			error = xfs_btree_debug_check_ptr(cur, lpp, i, level);
cur              2747 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_copy_keys(cur, rkp, lkp, rrecs);
cur              2748 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_copy_ptrs(cur, rpp, lpp, rrecs);
cur              2750 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_log_keys(cur, rbp, 1, rrecs);
cur              2751 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_log_ptrs(cur, rbp, 1, rrecs);
cur              2754 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_get_node_keys(cur, right, key);
cur              2760 fs/xfs/libxfs/xfs_btree.c 		lrp = xfs_btree_rec_addr(cur, src_index, left);
cur              2761 fs/xfs/libxfs/xfs_btree.c 		rrp = xfs_btree_rec_addr(cur, 1, right);
cur              2764 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_copy_recs(cur, rrp, lrp, rrecs);
cur              2765 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_log_recs(cur, rbp, 1, rrecs);
cur              2768 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_get_leaf_keys(cur, right, key);
cur              2775 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_get_sibling(cur, left, &rrptr, XFS_BB_RIGHTSIB);
cur              2776 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_set_sibling(cur, right, &rrptr, XFS_BB_RIGHTSIB);
cur              2777 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_set_sibling(cur, right, &lptr, XFS_BB_LEFTSIB);
cur              2778 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_set_sibling(cur, left, &rptr, XFS_BB_RIGHTSIB);
cur              2780 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_log_block(cur, rbp, XFS_BB_ALL_BITS);
cur              2781 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_log_block(cur, lbp, XFS_BB_NUMRECS | XFS_BB_RIGHTSIB);
cur              2787 fs/xfs/libxfs/xfs_btree.c 	if (!xfs_btree_ptr_is_null(cur, &rrptr)) {
cur              2788 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_read_buf_block(cur, &rrptr,
cur              2792 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_set_sibling(cur, rrblock, &rptr, XFS_BB_LEFTSIB);
cur              2793 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_log_block(cur, rrbp, XFS_BB_LEFTSIB);
cur              2797 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_flags & XFS_BTREE_OVERLAPPING) {
cur              2798 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_update_keys(cur, level);
cur              2808 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_ptrs[level] > lrecs + 1) {
cur              2809 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_setbuf(cur, level, rbp);
cur              2810 fs/xfs/libxfs/xfs_btree.c 		cur->bc_ptrs[level] -= lrecs;
cur              2816 fs/xfs/libxfs/xfs_btree.c 	if (level + 1 < cur->bc_nlevels) {
cur              2817 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_dup_cursor(cur, curp);
cur              2834 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur;
cur              2869 fs/xfs/libxfs/xfs_btree.c 	args->result = __xfs_btree_split(args->cur, args->level, args->ptrp,
cur              2883 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              2893 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_btnum != XFS_BTNUM_BMAP)
cur              2894 fs/xfs/libxfs/xfs_btree.c 		return __xfs_btree_split(cur, level, ptrp, key, curp, stat);
cur              2896 fs/xfs/libxfs/xfs_btree.c 	args.cur = cur;
cur              2918 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,		/* btree cursor */
cur              2934 fs/xfs/libxfs/xfs_btree.c 	XFS_BTREE_STATS_INC(cur, newroot);
cur              2936 fs/xfs/libxfs/xfs_btree.c 	ASSERT(cur->bc_flags & XFS_BTREE_ROOT_IN_INODE);
cur              2938 fs/xfs/libxfs/xfs_btree.c 	level = cur->bc_nlevels - 1;
cur              2940 fs/xfs/libxfs/xfs_btree.c 	block = xfs_btree_get_iroot(cur);
cur              2941 fs/xfs/libxfs/xfs_btree.c 	pp = xfs_btree_ptr_addr(cur, 1, block);
cur              2944 fs/xfs/libxfs/xfs_btree.c 	error = cur->bc_ops->alloc_block(cur, pp, &nptr, stat);
cur              2950 fs/xfs/libxfs/xfs_btree.c 	XFS_BTREE_STATS_INC(cur, alloc);
cur              2953 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_get_buf_block(cur, &nptr, &cblock, &cbp);
cur              2961 fs/xfs/libxfs/xfs_btree.c 	memcpy(cblock, block, xfs_btree_block_len(cur));
cur              2962 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_flags & XFS_BTREE_CRC_BLOCKS) {
cur              2963 fs/xfs/libxfs/xfs_btree.c 		if (cur->bc_flags & XFS_BTREE_LONG_PTRS)
cur              2971 fs/xfs/libxfs/xfs_btree.c 	cur->bc_nlevels++;
cur              2972 fs/xfs/libxfs/xfs_btree.c 	cur->bc_ptrs[level + 1] = 1;
cur              2974 fs/xfs/libxfs/xfs_btree.c 	kp = xfs_btree_key_addr(cur, 1, block);
cur              2975 fs/xfs/libxfs/xfs_btree.c 	ckp = xfs_btree_key_addr(cur, 1, cblock);
cur              2976 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_copy_keys(cur, ckp, kp, xfs_btree_get_numrecs(cblock));
cur              2978 fs/xfs/libxfs/xfs_btree.c 	cpp = xfs_btree_ptr_addr(cur, 1, cblock);
cur              2980 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_debug_check_ptr(cur, pp, i, level);
cur              2985 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_copy_ptrs(cur, cpp, pp, xfs_btree_get_numrecs(cblock));
cur              2987 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_debug_check_ptr(cur, &nptr, 0, level);
cur              2991 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_copy_ptrs(cur, pp, &nptr, 1);
cur              2993 fs/xfs/libxfs/xfs_btree.c 	xfs_iroot_realloc(cur->bc_private.b.ip,
cur              2995 fs/xfs/libxfs/xfs_btree.c 			  cur->bc_private.b.whichfork);
cur              2997 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_setbuf(cur, level, cbp);
cur              3003 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_log_block(cur, cbp, XFS_BB_ALL_BITS);
cur              3004 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_log_keys(cur, cbp, 1, be16_to_cpu(cblock->bb_numrecs));
cur              3005 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_log_ptrs(cur, cbp, 1, be16_to_cpu(cblock->bb_numrecs));
cur              3008 fs/xfs/libxfs/xfs_btree.c 		XFS_ILOG_CORE | xfs_ilog_fbroot(cur->bc_private.b.whichfork);
cur              3020 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,	/* btree cursor */
cur              3036 fs/xfs/libxfs/xfs_btree.c 	XFS_BTREE_STATS_INC(cur, newroot);
cur              3039 fs/xfs/libxfs/xfs_btree.c 	cur->bc_ops->init_ptr_from_cur(cur, &rptr);
cur              3042 fs/xfs/libxfs/xfs_btree.c 	error = cur->bc_ops->alloc_block(cur, &rptr, &lptr, stat);
cur              3047 fs/xfs/libxfs/xfs_btree.c 	XFS_BTREE_STATS_INC(cur, alloc);
cur              3050 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_get_buf_block(cur, &lptr, &new, &nbp);
cur              3055 fs/xfs/libxfs/xfs_btree.c 	cur->bc_ops->set_root(cur, &lptr, 1);
cur              3063 fs/xfs/libxfs/xfs_btree.c 	block = xfs_btree_get_block(cur, cur->bc_nlevels - 1, &bp);
cur              3066 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_check_block(cur, block, cur->bc_nlevels - 1, bp);
cur              3071 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_get_sibling(cur, block, &rptr, XFS_BB_RIGHTSIB);
cur              3072 fs/xfs/libxfs/xfs_btree.c 	if (!xfs_btree_ptr_is_null(cur, &rptr)) {
cur              3075 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_buf_to_ptr(cur, lbp, &lptr);
cur              3077 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_read_buf_block(cur, &rptr, 0, &right, &rbp);
cur              3085 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_buf_to_ptr(cur, rbp, &rptr);
cur              3087 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_get_sibling(cur, right, &lptr, XFS_BB_LEFTSIB);
cur              3088 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_read_buf_block(cur, &lptr, 0, &left, &lbp);
cur              3096 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_init_block_cur(cur, nbp, cur->bc_nlevels, 2);
cur              3097 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_log_block(cur, nbp, XFS_BB_ALL_BITS);
cur              3098 fs/xfs/libxfs/xfs_btree.c 	ASSERT(!xfs_btree_ptr_is_null(cur, &lptr) &&
cur              3099 fs/xfs/libxfs/xfs_btree.c 			!xfs_btree_ptr_is_null(cur, &rptr));
cur              3107 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_get_node_keys(cur, left,
cur              3108 fs/xfs/libxfs/xfs_btree.c 				xfs_btree_key_addr(cur, 1, new));
cur              3109 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_get_node_keys(cur, right,
cur              3110 fs/xfs/libxfs/xfs_btree.c 				xfs_btree_key_addr(cur, 2, new));
cur              3117 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_get_leaf_keys(cur, left,
cur              3118 fs/xfs/libxfs/xfs_btree.c 			xfs_btree_key_addr(cur, 1, new));
cur              3119 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_get_leaf_keys(cur, right,
cur              3120 fs/xfs/libxfs/xfs_btree.c 			xfs_btree_key_addr(cur, 2, new));
cur              3122 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_log_keys(cur, nbp, 1, 2);
cur              3125 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_copy_ptrs(cur,
cur              3126 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_ptr_addr(cur, 1, new), &lptr, 1);
cur              3127 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_copy_ptrs(cur,
cur              3128 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_ptr_addr(cur, 2, new), &rptr, 1);
cur              3129 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_log_ptrs(cur, nbp, 1, 2);
cur              3132 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_setbuf(cur, cur->bc_nlevels, nbp);
cur              3133 fs/xfs/libxfs/xfs_btree.c 	cur->bc_ptrs[cur->bc_nlevels] = nptr;
cur              3134 fs/xfs/libxfs/xfs_btree.c 	cur->bc_nlevels++;
cur              3146 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,	/* btree cursor */
cur              3158 fs/xfs/libxfs/xfs_btree.c 	if ((cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) &&
cur              3159 fs/xfs/libxfs/xfs_btree.c 	    level == cur->bc_nlevels - 1) {
cur              3160 fs/xfs/libxfs/xfs_btree.c 		struct xfs_inode *ip = cur->bc_private.b.ip;
cur              3162 fs/xfs/libxfs/xfs_btree.c 		if (numrecs < cur->bc_ops->get_dmaxrecs(cur, level)) {
cur              3164 fs/xfs/libxfs/xfs_btree.c 			xfs_iroot_realloc(ip, 1, cur->bc_private.b.whichfork);
cur              3170 fs/xfs/libxfs/xfs_btree.c 			error = xfs_btree_new_iroot(cur, &logflags, stat);
cur              3174 fs/xfs/libxfs/xfs_btree.c 			xfs_trans_log_inode(cur->bc_tp, ip, logflags);
cur              3181 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_rshift(cur, level, stat);
cur              3186 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_lshift(cur, level, stat);
cur              3191 fs/xfs/libxfs/xfs_btree.c 		*oindex = *index = cur->bc_ptrs[level];
cur              3201 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_split(cur, level, nptr, key, ncur, stat);
cur              3206 fs/xfs/libxfs/xfs_btree.c 	*index = cur->bc_ptrs[level];
cur              3216 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,	/* btree cursor */
cur              3244 fs/xfs/libxfs/xfs_btree.c 	if (!(cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) &&
cur              3245 fs/xfs/libxfs/xfs_btree.c 	    (level >= cur->bc_nlevels)) {
cur              3246 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_new_root(cur, stat);
cur              3247 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_set_ptr_null(cur, ptrp);
cur              3253 fs/xfs/libxfs/xfs_btree.c 	ptr = cur->bc_ptrs[level];
cur              3261 fs/xfs/libxfs/xfs_btree.c 	XFS_BTREE_STATS_INC(cur, insrec);
cur              3264 fs/xfs/libxfs/xfs_btree.c 	block = xfs_btree_get_block(cur, level, &bp);
cur              3269 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_check_block(cur, block, level, bp);
cur              3276 fs/xfs/libxfs/xfs_btree.c 			ASSERT(cur->bc_ops->recs_inorder(cur, rec,
cur              3277 fs/xfs/libxfs/xfs_btree.c 				xfs_btree_rec_addr(cur, ptr, block)));
cur              3279 fs/xfs/libxfs/xfs_btree.c 			ASSERT(cur->bc_ops->keys_inorder(cur, key,
cur              3280 fs/xfs/libxfs/xfs_btree.c 				xfs_btree_key_addr(cur, ptr, block)));
cur              3289 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_set_ptr_null(cur, &nptr);
cur              3290 fs/xfs/libxfs/xfs_btree.c 	if (numrecs == cur->bc_ops->get_maxrecs(cur, level)) {
cur              3291 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_make_block_unfull(cur, level, numrecs,
cur              3301 fs/xfs/libxfs/xfs_btree.c 	block = xfs_btree_get_block(cur, level, &bp);
cur              3305 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_check_block(cur, block, level, bp);
cur              3314 fs/xfs/libxfs/xfs_btree.c 	XFS_BTREE_STATS_ADD(cur, moves, numrecs - ptr + 1);
cur              3321 fs/xfs/libxfs/xfs_btree.c 		kp = xfs_btree_key_addr(cur, ptr, block);
cur              3322 fs/xfs/libxfs/xfs_btree.c 		pp = xfs_btree_ptr_addr(cur, ptr, block);
cur              3325 fs/xfs/libxfs/xfs_btree.c 			error = xfs_btree_debug_check_ptr(cur, pp, i, level);
cur              3330 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_shift_keys(cur, kp, 1, numrecs - ptr + 1);
cur              3331 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_shift_ptrs(cur, pp, 1, numrecs - ptr + 1);
cur              3333 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_debug_check_ptr(cur, ptrp, 0, level);
cur              3338 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_copy_keys(cur, kp, key, 1);
cur              3339 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_copy_ptrs(cur, pp, ptrp, 1);
cur              3342 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_log_ptrs(cur, bp, ptr, numrecs);
cur              3343 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_log_keys(cur, bp, ptr, numrecs);
cur              3346 fs/xfs/libxfs/xfs_btree.c 			ASSERT(cur->bc_ops->keys_inorder(cur, kp,
cur              3347 fs/xfs/libxfs/xfs_btree.c 				xfs_btree_key_addr(cur, ptr + 1, block)));
cur              3354 fs/xfs/libxfs/xfs_btree.c 		rp = xfs_btree_rec_addr(cur, ptr, block);
cur              3356 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_shift_recs(cur, rp, 1, numrecs - ptr + 1);
cur              3359 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_copy_recs(cur, rp, rec, 1);
cur              3361 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_log_recs(cur, bp, ptr, numrecs);
cur              3364 fs/xfs/libxfs/xfs_btree.c 			ASSERT(cur->bc_ops->recs_inorder(cur, rp,
cur              3365 fs/xfs/libxfs/xfs_btree.c 				xfs_btree_rec_addr(cur, ptr + 1, block)));
cur              3371 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_log_block(cur, bp, XFS_BB_NUMRECS);
cur              3382 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_get_keys(cur, block, lkey);
cur              3383 fs/xfs/libxfs/xfs_btree.c 	} else if (xfs_btree_needs_key_update(cur, optr)) {
cur              3384 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_update_keys(cur, level);
cur              3393 fs/xfs/libxfs/xfs_btree.c 	if (xfs_btree_is_lastrec(cur, block, level)) {
cur              3394 fs/xfs/libxfs/xfs_btree.c 		cur->bc_ops->update_lastrec(cur, block, rec,
cur              3403 fs/xfs/libxfs/xfs_btree.c 	if (!xfs_btree_ptr_is_null(cur, &nptr)) {
cur              3404 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_copy_keys(cur, key, lkey, 1);
cur              3424 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              3439 fs/xfs/libxfs/xfs_btree.c 	pcur = cur;
cur              3442 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_set_ptr_null(cur, &nptr);
cur              3445 fs/xfs/libxfs/xfs_btree.c 	cur->bc_ops->init_rec_from_cur(cur, &rec);
cur              3446 fs/xfs/libxfs/xfs_btree.c 	cur->bc_ops->init_key_from_rec(key, &rec);
cur              3461 fs/xfs/libxfs/xfs_btree.c 			if (pcur != cur)
cur              3466 fs/xfs/libxfs/xfs_btree.c 		XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, i == 1, error0);
cur              3474 fs/xfs/libxfs/xfs_btree.c 		if (pcur != cur &&
cur              3475 fs/xfs/libxfs/xfs_btree.c 		    (ncur || xfs_btree_ptr_is_null(cur, &nptr))) {
cur              3477 fs/xfs/libxfs/xfs_btree.c 			if (cur->bc_ops->update_cursor)
cur              3478 fs/xfs/libxfs/xfs_btree.c 				cur->bc_ops->update_cursor(pcur, cur);
cur              3479 fs/xfs/libxfs/xfs_btree.c 			cur->bc_nlevels = pcur->bc_nlevels;
cur              3487 fs/xfs/libxfs/xfs_btree.c 	} while (!xfs_btree_ptr_is_null(cur, &nptr));
cur              3505 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur)
cur              3507 fs/xfs/libxfs/xfs_btree.c 	int			whichfork = cur->bc_private.b.whichfork;
cur              3508 fs/xfs/libxfs/xfs_btree.c 	struct xfs_inode	*ip = cur->bc_private.b.ip;
cur              3526 fs/xfs/libxfs/xfs_btree.c 	ASSERT(cur->bc_flags & XFS_BTREE_ROOT_IN_INODE);
cur              3527 fs/xfs/libxfs/xfs_btree.c 	ASSERT(cur->bc_nlevels > 1);
cur              3533 fs/xfs/libxfs/xfs_btree.c 	level = cur->bc_nlevels - 1;
cur              3540 fs/xfs/libxfs/xfs_btree.c 	block = xfs_btree_get_iroot(cur);
cur              3544 fs/xfs/libxfs/xfs_btree.c 	cblock = xfs_btree_get_block(cur, level - 1, &cbp);
cur              3552 fs/xfs/libxfs/xfs_btree.c 	if (numrecs > cur->bc_ops->get_dmaxrecs(cur, level))
cur              3555 fs/xfs/libxfs/xfs_btree.c 	XFS_BTREE_STATS_INC(cur, killroot);
cur              3558 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_get_sibling(cur, block, &ptr, XFS_BB_LEFTSIB);
cur              3559 fs/xfs/libxfs/xfs_btree.c 	ASSERT(xfs_btree_ptr_is_null(cur, &ptr));
cur              3560 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_get_sibling(cur, block, &ptr, XFS_BB_RIGHTSIB);
cur              3561 fs/xfs/libxfs/xfs_btree.c 	ASSERT(xfs_btree_ptr_is_null(cur, &ptr));
cur              3564 fs/xfs/libxfs/xfs_btree.c 	index = numrecs - cur->bc_ops->get_maxrecs(cur, level);
cur              3566 fs/xfs/libxfs/xfs_btree.c 		xfs_iroot_realloc(cur->bc_private.b.ip, index,
cur              3567 fs/xfs/libxfs/xfs_btree.c 				  cur->bc_private.b.whichfork);
cur              3574 fs/xfs/libxfs/xfs_btree.c 	kp = xfs_btree_key_addr(cur, 1, block);
cur              3575 fs/xfs/libxfs/xfs_btree.c 	ckp = xfs_btree_key_addr(cur, 1, cblock);
cur              3576 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_copy_keys(cur, kp, ckp, numrecs);
cur              3578 fs/xfs/libxfs/xfs_btree.c 	pp = xfs_btree_ptr_addr(cur, 1, block);
cur              3579 fs/xfs/libxfs/xfs_btree.c 	cpp = xfs_btree_ptr_addr(cur, 1, cblock);
cur              3582 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_debug_check_ptr(cur, cpp, i, level - 1);
cur              3587 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_copy_ptrs(cur, pp, cpp, numrecs);
cur              3589 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_free_block(cur, cbp);
cur              3593 fs/xfs/libxfs/xfs_btree.c 	cur->bc_bufs[level - 1] = NULL;
cur              3595 fs/xfs/libxfs/xfs_btree.c 	xfs_trans_log_inode(cur->bc_tp, ip,
cur              3596 fs/xfs/libxfs/xfs_btree.c 		XFS_ILOG_CORE | xfs_ilog_fbroot(cur->bc_private.b.whichfork));
cur              3597 fs/xfs/libxfs/xfs_btree.c 	cur->bc_nlevels--;
cur              3607 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              3614 fs/xfs/libxfs/xfs_btree.c 	XFS_BTREE_STATS_INC(cur, killroot);
cur              3620 fs/xfs/libxfs/xfs_btree.c 	cur->bc_ops->set_root(cur, newroot, -1);
cur              3622 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_free_block(cur, bp);
cur              3626 fs/xfs/libxfs/xfs_btree.c 	cur->bc_bufs[level] = NULL;
cur              3627 fs/xfs/libxfs/xfs_btree.c 	cur->bc_ra[level] = 0;
cur              3628 fs/xfs/libxfs/xfs_btree.c 	cur->bc_nlevels--;
cur              3635 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              3643 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_decrement(cur, level, &i);
cur              3660 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,		/* btree cursor */
cur              3686 fs/xfs/libxfs/xfs_btree.c 	ptr = cur->bc_ptrs[level];
cur              3693 fs/xfs/libxfs/xfs_btree.c 	block = xfs_btree_get_block(cur, level, &bp);
cur              3697 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_check_block(cur, block, level, bp);
cur              3708 fs/xfs/libxfs/xfs_btree.c 	XFS_BTREE_STATS_INC(cur, delrec);
cur              3709 fs/xfs/libxfs/xfs_btree.c 	XFS_BTREE_STATS_ADD(cur, moves, numrecs - ptr);
cur              3717 fs/xfs/libxfs/xfs_btree.c 		lkp = xfs_btree_key_addr(cur, ptr + 1, block);
cur              3718 fs/xfs/libxfs/xfs_btree.c 		lpp = xfs_btree_ptr_addr(cur, ptr + 1, block);
cur              3721 fs/xfs/libxfs/xfs_btree.c 			error = xfs_btree_debug_check_ptr(cur, lpp, i, level);
cur              3727 fs/xfs/libxfs/xfs_btree.c 			xfs_btree_shift_keys(cur, lkp, -1, numrecs - ptr);
cur              3728 fs/xfs/libxfs/xfs_btree.c 			xfs_btree_shift_ptrs(cur, lpp, -1, numrecs - ptr);
cur              3729 fs/xfs/libxfs/xfs_btree.c 			xfs_btree_log_keys(cur, bp, ptr, numrecs - 1);
cur              3730 fs/xfs/libxfs/xfs_btree.c 			xfs_btree_log_ptrs(cur, bp, ptr, numrecs - 1);
cur              3735 fs/xfs/libxfs/xfs_btree.c 			xfs_btree_shift_recs(cur,
cur              3736 fs/xfs/libxfs/xfs_btree.c 				xfs_btree_rec_addr(cur, ptr + 1, block),
cur              3738 fs/xfs/libxfs/xfs_btree.c 			xfs_btree_log_recs(cur, bp, ptr, numrecs - 1);
cur              3746 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_log_block(cur, bp, XFS_BB_NUMRECS);
cur              3752 fs/xfs/libxfs/xfs_btree.c 	if (xfs_btree_is_lastrec(cur, block, level)) {
cur              3753 fs/xfs/libxfs/xfs_btree.c 		cur->bc_ops->update_lastrec(cur, block, NULL,
cur              3762 fs/xfs/libxfs/xfs_btree.c 	if (level == cur->bc_nlevels - 1) {
cur              3763 fs/xfs/libxfs/xfs_btree.c 		if (cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) {
cur              3764 fs/xfs/libxfs/xfs_btree.c 			xfs_iroot_realloc(cur->bc_private.b.ip, -1,
cur              3765 fs/xfs/libxfs/xfs_btree.c 					  cur->bc_private.b.whichfork);
cur              3767 fs/xfs/libxfs/xfs_btree.c 			error = xfs_btree_kill_iroot(cur);
cur              3771 fs/xfs/libxfs/xfs_btree.c 			error = xfs_btree_dec_cursor(cur, level, stat);
cur              3789 fs/xfs/libxfs/xfs_btree.c 			pp = xfs_btree_ptr_addr(cur, 1, block);
cur              3790 fs/xfs/libxfs/xfs_btree.c 			error = xfs_btree_kill_root(cur, bp, level, pp);
cur              3794 fs/xfs/libxfs/xfs_btree.c 			error = xfs_btree_dec_cursor(cur, level, stat);
cur              3806 fs/xfs/libxfs/xfs_btree.c 	if (xfs_btree_needs_key_update(cur, ptr)) {
cur              3807 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_update_keys(cur, level);
cur              3816 fs/xfs/libxfs/xfs_btree.c 	if (numrecs >= cur->bc_ops->get_minrecs(cur, level)) {
cur              3817 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_dec_cursor(cur, level, stat);
cur              3828 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_get_sibling(cur, block, &rptr, XFS_BB_RIGHTSIB);
cur              3829 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_get_sibling(cur, block, &lptr, XFS_BB_LEFTSIB);
cur              3831 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) {
cur              3837 fs/xfs/libxfs/xfs_btree.c 		if (xfs_btree_ptr_is_null(cur, &rptr) &&
cur              3838 fs/xfs/libxfs/xfs_btree.c 		    xfs_btree_ptr_is_null(cur, &lptr) &&
cur              3839 fs/xfs/libxfs/xfs_btree.c 		    level == cur->bc_nlevels - 2) {
cur              3840 fs/xfs/libxfs/xfs_btree.c 			error = xfs_btree_kill_iroot(cur);
cur              3842 fs/xfs/libxfs/xfs_btree.c 				error = xfs_btree_dec_cursor(cur, level, stat);
cur              3849 fs/xfs/libxfs/xfs_btree.c 	ASSERT(!xfs_btree_ptr_is_null(cur, &rptr) ||
cur              3850 fs/xfs/libxfs/xfs_btree.c 	       !xfs_btree_ptr_is_null(cur, &lptr));
cur              3856 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_dup_cursor(cur, &tcur);
cur              3864 fs/xfs/libxfs/xfs_btree.c 	if (!xfs_btree_ptr_is_null(cur, &rptr)) {
cur              3870 fs/xfs/libxfs/xfs_btree.c 		XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, i == 1, error0);
cur              3875 fs/xfs/libxfs/xfs_btree.c 		XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, i == 1, error0);
cur              3878 fs/xfs/libxfs/xfs_btree.c 		XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, i == 1, error0);
cur              3896 fs/xfs/libxfs/xfs_btree.c 		    cur->bc_ops->get_minrecs(tcur, level)) {
cur              3902 fs/xfs/libxfs/xfs_btree.c 				       cur->bc_ops->get_minrecs(tcur, level));
cur              3907 fs/xfs/libxfs/xfs_btree.c 				error = xfs_btree_dec_cursor(cur, level, stat);
cur              3920 fs/xfs/libxfs/xfs_btree.c 		if (!xfs_btree_ptr_is_null(cur, &lptr)) {
cur              3922 fs/xfs/libxfs/xfs_btree.c 			XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, i == 1, error0);
cur              3927 fs/xfs/libxfs/xfs_btree.c 			XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, i == 1, error0);
cur              3935 fs/xfs/libxfs/xfs_btree.c 	if (!xfs_btree_ptr_is_null(cur, &lptr)) {
cur              3941 fs/xfs/libxfs/xfs_btree.c 		XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, i == 1, error0);
cur              3947 fs/xfs/libxfs/xfs_btree.c 		XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, i == 1, error0);
cur              3952 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_check_block(cur, left, level, lbp);
cur              3965 fs/xfs/libxfs/xfs_btree.c 		    cur->bc_ops->get_minrecs(tcur, level)) {
cur              3971 fs/xfs/libxfs/xfs_btree.c 				       cur->bc_ops->get_minrecs(tcur, level));
cur              3975 fs/xfs/libxfs/xfs_btree.c 					cur->bc_ptrs[0]++;
cur              3994 fs/xfs/libxfs/xfs_btree.c 	ASSERT(!xfs_btree_ptr_is_null(cur, &cptr));
cur              3996 fs/xfs/libxfs/xfs_btree.c 	if (!xfs_btree_ptr_is_null(cur, &lptr) &&
cur              3998 fs/xfs/libxfs/xfs_btree.c 			cur->bc_ops->get_maxrecs(cur, level)) {
cur              4006 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_read_buf_block(cur, &lptr, 0, &left, &lbp);
cur              4013 fs/xfs/libxfs/xfs_btree.c 	} else if (!xfs_btree_ptr_is_null(cur, &rptr) &&
cur              4015 fs/xfs/libxfs/xfs_btree.c 			cur->bc_ops->get_maxrecs(cur, level)) {
cur              4023 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_read_buf_block(cur, &rptr, 0, &right, &rbp);
cur              4032 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_dec_cursor(cur, level, stat);
cur              4045 fs/xfs/libxfs/xfs_btree.c 	XFS_BTREE_STATS_ADD(cur, moves, rrecs);
cur              4053 fs/xfs/libxfs/xfs_btree.c 		lkp = xfs_btree_key_addr(cur, lrecs + 1, left);
cur              4054 fs/xfs/libxfs/xfs_btree.c 		lpp = xfs_btree_ptr_addr(cur, lrecs + 1, left);
cur              4055 fs/xfs/libxfs/xfs_btree.c 		rkp = xfs_btree_key_addr(cur, 1, right);
cur              4056 fs/xfs/libxfs/xfs_btree.c 		rpp = xfs_btree_ptr_addr(cur, 1, right);
cur              4059 fs/xfs/libxfs/xfs_btree.c 			error = xfs_btree_debug_check_ptr(cur, rpp, i, level);
cur              4064 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_copy_keys(cur, lkp, rkp, rrecs);
cur              4065 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_copy_ptrs(cur, lpp, rpp, rrecs);
cur              4067 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_log_keys(cur, lbp, lrecs + 1, lrecs + rrecs);
cur              4068 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_log_ptrs(cur, lbp, lrecs + 1, lrecs + rrecs);
cur              4074 fs/xfs/libxfs/xfs_btree.c 		lrp = xfs_btree_rec_addr(cur, lrecs + 1, left);
cur              4075 fs/xfs/libxfs/xfs_btree.c 		rrp = xfs_btree_rec_addr(cur, 1, right);
cur              4077 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_copy_recs(cur, lrp, rrp, rrecs);
cur              4078 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_log_recs(cur, lbp, lrecs + 1, lrecs + rrecs);
cur              4081 fs/xfs/libxfs/xfs_btree.c 	XFS_BTREE_STATS_INC(cur, join);
cur              4088 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_get_sibling(cur, right, &cptr, XFS_BB_RIGHTSIB),
cur              4089 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_set_sibling(cur, left, &cptr, XFS_BB_RIGHTSIB);
cur              4090 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_log_block(cur, lbp, XFS_BB_NUMRECS | XFS_BB_RIGHTSIB);
cur              4093 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_get_sibling(cur, left, &cptr, XFS_BB_RIGHTSIB);
cur              4094 fs/xfs/libxfs/xfs_btree.c 	if (!xfs_btree_ptr_is_null(cur, &cptr)) {
cur              4095 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_read_buf_block(cur, &cptr, 0, &rrblock, &rrbp);
cur              4098 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_set_sibling(cur, rrblock, &lptr, XFS_BB_LEFTSIB);
cur              4099 fs/xfs/libxfs/xfs_btree.c 		xfs_btree_log_block(cur, rrbp, XFS_BB_LEFTSIB);
cur              4103 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_free_block(cur, rbp);
cur              4112 fs/xfs/libxfs/xfs_btree.c 		cur->bc_bufs[level] = lbp;
cur              4113 fs/xfs/libxfs/xfs_btree.c 		cur->bc_ptrs[level] += lrecs;
cur              4114 fs/xfs/libxfs/xfs_btree.c 		cur->bc_ra[level] = 0;
cur              4120 fs/xfs/libxfs/xfs_btree.c 	else if ((cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) ||
cur              4121 fs/xfs/libxfs/xfs_btree.c 		   (level + 1 < cur->bc_nlevels)) {
cur              4122 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_increment(cur, level + 1, &i);
cur              4134 fs/xfs/libxfs/xfs_btree.c 		cur->bc_ptrs[level]--;
cur              4163 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              4178 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_delrec(cur, level, &i);
cur              4189 fs/xfs/libxfs/xfs_btree.c 	if (joined && (cur->bc_flags & XFS_BTREE_OVERLAPPING)) {
cur              4190 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_updkeys_force(cur, 0);
cur              4196 fs/xfs/libxfs/xfs_btree.c 		for (level = 1; level < cur->bc_nlevels; level++) {
cur              4197 fs/xfs/libxfs/xfs_btree.c 			if (cur->bc_ptrs[level] == 0) {
cur              4198 fs/xfs/libxfs/xfs_btree.c 				error = xfs_btree_decrement(cur, level, &i);
cur              4217 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,	/* btree cursor */
cur              4228 fs/xfs/libxfs/xfs_btree.c 	ptr = cur->bc_ptrs[0];
cur              4229 fs/xfs/libxfs/xfs_btree.c 	block = xfs_btree_get_block(cur, 0, &bp);
cur              4232 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_check_block(cur, block, 0, bp);
cur              4248 fs/xfs/libxfs/xfs_btree.c 	*recp = xfs_btree_rec_addr(cur, ptr, block);
cur              4256 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur		*cur,
cur              4267 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_readahead(cur, level, XFS_BTCUR_RIGHTRA);
cur              4268 fs/xfs/libxfs/xfs_btree.c 	block = xfs_btree_get_block(cur, level, &bp);
cur              4271 fs/xfs/libxfs/xfs_btree.c 	error = fn(cur, level, data);
cur              4276 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_get_sibling(cur, block, &rptr, XFS_BB_RIGHTSIB);
cur              4277 fs/xfs/libxfs/xfs_btree.c 	if (xfs_btree_ptr_is_null(cur, &rptr))
cur              4280 fs/xfs/libxfs/xfs_btree.c 	return xfs_btree_lookup_get_block(cur, level, &rptr, &block);
cur              4287 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur		*cur,
cur              4296 fs/xfs/libxfs/xfs_btree.c 	cur->bc_ops->init_ptr_from_cur(cur, &lptr);
cur              4299 fs/xfs/libxfs/xfs_btree.c 	for (level = cur->bc_nlevels - 1; level >= 0; level--) {
cur              4301 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_lookup_get_block(cur, level, &lptr, &block);
cur              4309 fs/xfs/libxfs/xfs_btree.c 			ptr = xfs_btree_ptr_addr(cur, 1, block);
cur              4310 fs/xfs/libxfs/xfs_btree.c 			xfs_btree_readahead_ptr(cur, ptr, 1);
cur              4313 fs/xfs/libxfs/xfs_btree.c 			xfs_btree_copy_ptrs(cur, &lptr, ptr, 1);
cur              4318 fs/xfs/libxfs/xfs_btree.c 			error = xfs_btree_visit_block(cur, level, fn, data);
cur              4359 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              4368 fs/xfs/libxfs/xfs_btree.c 	block = xfs_btree_get_block(cur, level, &bp);
cur              4369 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_flags & XFS_BTREE_LONG_PTRS) {
cur              4387 fs/xfs/libxfs/xfs_btree.c 		ASSERT(cur->bc_flags & XFS_BTREE_ROOT_IN_INODE);
cur              4388 fs/xfs/libxfs/xfs_btree.c 		ASSERT(level == cur->bc_nlevels - 1);
cur              4392 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_tp) {
cur              4393 fs/xfs/libxfs/xfs_btree.c 		if (!xfs_trans_ordered_buf(cur->bc_tp, bp)) {
cur              4394 fs/xfs/libxfs/xfs_btree.c 			xfs_btree_log_block(cur, bp, XFS_BB_OWNER);
cur              4406 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              4415 fs/xfs/libxfs/xfs_btree.c 	return xfs_btree_visit_blocks(cur, xfs_btree_block_change_owner,
cur              4545 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur		*cur,
cur              4558 fs/xfs/libxfs/xfs_btree.c 	ASSERT(cur->bc_ops->init_high_key_from_rec);
cur              4559 fs/xfs/libxfs/xfs_btree.c 	ASSERT(cur->bc_ops->diff_two_keys);
cur              4566 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_lookup(cur, XFS_LOOKUP_LE, &stat);
cur              4572 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_increment(cur, 0, &stat);
cur              4579 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_get_rec(cur, &recp, &stat);
cur              4585 fs/xfs/libxfs/xfs_btree.c 			cur->bc_ops->init_high_key_from_rec(&rec_key, recp);
cur              4587 fs/xfs/libxfs/xfs_btree.c 			diff = cur->bc_ops->diff_two_keys(cur, low_key,
cur              4594 fs/xfs/libxfs/xfs_btree.c 		cur->bc_ops->init_key_from_rec(&rec_key, recp);
cur              4595 fs/xfs/libxfs/xfs_btree.c 		diff = cur->bc_ops->diff_two_keys(cur, &rec_key, high_key);
cur              4600 fs/xfs/libxfs/xfs_btree.c 		error = fn(cur, recp, priv);
cur              4606 fs/xfs/libxfs/xfs_btree.c 		error = xfs_btree_increment(cur, 0, &stat);
cur              4636 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur		*cur,
cur              4658 fs/xfs/libxfs/xfs_btree.c 	level = cur->bc_nlevels - 1;
cur              4659 fs/xfs/libxfs/xfs_btree.c 	cur->bc_ops->init_ptr_from_cur(cur, &ptr);
cur              4660 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_lookup_get_block(cur, level, &ptr, &block);
cur              4663 fs/xfs/libxfs/xfs_btree.c 	xfs_btree_get_block(cur, level, &bp);
cur              4664 fs/xfs/libxfs/xfs_btree.c 	trace_xfs_btree_overlapped_query_range(cur, level, bp);
cur              4666 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_check_block(cur, block, level, bp);
cur              4670 fs/xfs/libxfs/xfs_btree.c 	cur->bc_ptrs[level] = 1;
cur              4672 fs/xfs/libxfs/xfs_btree.c 	while (level < cur->bc_nlevels) {
cur              4673 fs/xfs/libxfs/xfs_btree.c 		block = xfs_btree_get_block(cur, level, &bp);
cur              4676 fs/xfs/libxfs/xfs_btree.c 		if (cur->bc_ptrs[level] > be16_to_cpu(block->bb_numrecs)) {
cur              4678 fs/xfs/libxfs/xfs_btree.c 			if (level < cur->bc_nlevels - 1)
cur              4679 fs/xfs/libxfs/xfs_btree.c 				cur->bc_ptrs[level + 1]++;
cur              4686 fs/xfs/libxfs/xfs_btree.c 			recp = xfs_btree_rec_addr(cur, cur->bc_ptrs[0], block);
cur              4688 fs/xfs/libxfs/xfs_btree.c 			cur->bc_ops->init_high_key_from_rec(&rec_hkey, recp);
cur              4689 fs/xfs/libxfs/xfs_btree.c 			ldiff = cur->bc_ops->diff_two_keys(cur, &rec_hkey,
cur              4692 fs/xfs/libxfs/xfs_btree.c 			cur->bc_ops->init_key_from_rec(&rec_key, recp);
cur              4693 fs/xfs/libxfs/xfs_btree.c 			hdiff = cur->bc_ops->diff_two_keys(cur, high_key,
cur              4702 fs/xfs/libxfs/xfs_btree.c 				error = fn(cur, recp, priv);
cur              4709 fs/xfs/libxfs/xfs_btree.c 			cur->bc_ptrs[level]++;
cur              4714 fs/xfs/libxfs/xfs_btree.c 		lkp = xfs_btree_key_addr(cur, cur->bc_ptrs[level], block);
cur              4715 fs/xfs/libxfs/xfs_btree.c 		hkp = xfs_btree_high_key_addr(cur, cur->bc_ptrs[level], block);
cur              4716 fs/xfs/libxfs/xfs_btree.c 		pp = xfs_btree_ptr_addr(cur, cur->bc_ptrs[level], block);
cur              4718 fs/xfs/libxfs/xfs_btree.c 		ldiff = cur->bc_ops->diff_two_keys(cur, hkp, low_key);
cur              4719 fs/xfs/libxfs/xfs_btree.c 		hdiff = cur->bc_ops->diff_two_keys(cur, high_key, lkp);
cur              4728 fs/xfs/libxfs/xfs_btree.c 			error = xfs_btree_lookup_get_block(cur, level, pp,
cur              4732 fs/xfs/libxfs/xfs_btree.c 			xfs_btree_get_block(cur, level, &bp);
cur              4733 fs/xfs/libxfs/xfs_btree.c 			trace_xfs_btree_overlapped_query_range(cur, level, bp);
cur              4735 fs/xfs/libxfs/xfs_btree.c 			error = xfs_btree_check_block(cur, block, level, bp);
cur              4739 fs/xfs/libxfs/xfs_btree.c 			cur->bc_ptrs[level] = 1;
cur              4745 fs/xfs/libxfs/xfs_btree.c 		cur->bc_ptrs[level]++;
cur              4756 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_bufs[0] == NULL) {
cur              4757 fs/xfs/libxfs/xfs_btree.c 		for (i = 0; i < cur->bc_nlevels; i++) {
cur              4758 fs/xfs/libxfs/xfs_btree.c 			if (cur->bc_bufs[i]) {
cur              4759 fs/xfs/libxfs/xfs_btree.c 				xfs_trans_brelse(cur->bc_tp, cur->bc_bufs[i]);
cur              4760 fs/xfs/libxfs/xfs_btree.c 				cur->bc_bufs[i] = NULL;
cur              4761 fs/xfs/libxfs/xfs_btree.c 				cur->bc_ptrs[i] = 0;
cur              4762 fs/xfs/libxfs/xfs_btree.c 				cur->bc_ra[i] = 0;
cur              4778 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur		*cur,
cur              4789 fs/xfs/libxfs/xfs_btree.c 	cur->bc_rec = *high_rec;
cur              4790 fs/xfs/libxfs/xfs_btree.c 	cur->bc_ops->init_rec_from_cur(cur, &rec);
cur              4791 fs/xfs/libxfs/xfs_btree.c 	cur->bc_ops->init_key_from_rec(&high_key, &rec);
cur              4793 fs/xfs/libxfs/xfs_btree.c 	cur->bc_rec = *low_rec;
cur              4794 fs/xfs/libxfs/xfs_btree.c 	cur->bc_ops->init_rec_from_cur(cur, &rec);
cur              4795 fs/xfs/libxfs/xfs_btree.c 	cur->bc_ops->init_key_from_rec(&low_key, &rec);
cur              4798 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_ops->diff_two_keys(cur, &low_key, &high_key) > 0)
cur              4801 fs/xfs/libxfs/xfs_btree.c 	if (!(cur->bc_flags & XFS_BTREE_OVERLAPPING))
cur              4802 fs/xfs/libxfs/xfs_btree.c 		return xfs_btree_simple_query_range(cur, &low_key,
cur              4804 fs/xfs/libxfs/xfs_btree.c 	return xfs_btree_overlapped_query_range(cur, &low_key, &high_key,
cur              4811 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur		*cur,
cur              4818 fs/xfs/libxfs/xfs_btree.c 	memset(&cur->bc_rec, 0, sizeof(cur->bc_rec));
cur              4822 fs/xfs/libxfs/xfs_btree.c 	return xfs_btree_simple_query_range(cur, &low_key, &high_key, fn, priv);
cur              4850 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              4863 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              4867 fs/xfs/libxfs/xfs_btree.c 	return xfs_btree_visit_blocks(cur, xfs_btree_count_blocks_helper,
cur              4874 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur		*cur,
cur              4878 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_flags & XFS_BTREE_LONG_PTRS)
cur              4886 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur		*cur,
cur              4896 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur,
cur              4903 fs/xfs/libxfs/xfs_btree.c 	error = xfs_btree_query_range(cur, low, high,
cur              4916 fs/xfs/libxfs/xfs_btree.c 	struct xfs_btree_cur	*cur)
cur              4921 fs/xfs/libxfs/xfs_btree.c 	block = xfs_btree_get_block(cur, 0, &bp);
cur              4924 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_ptrs[0] < xfs_btree_get_numrecs(block))
cur              4928 fs/xfs/libxfs/xfs_btree.c 	if (cur->bc_flags & XFS_BTREE_LONG_PTRS)
cur                88 fs/xfs/libxfs/xfs_btree.h #define XFS_BTREE_STATS_INC(cur, stat)	\
cur                89 fs/xfs/libxfs/xfs_btree.h 	XFS_STATS_INC_OFF((cur)->bc_mp, (cur)->bc_statoff + __XBTS_ ## stat)
cur                90 fs/xfs/libxfs/xfs_btree.h #define XFS_BTREE_STATS_ADD(cur, stat, val)	\
cur                91 fs/xfs/libxfs/xfs_btree.h 	XFS_STATS_ADD_OFF((cur)->bc_mp, (cur)->bc_statoff + __XBTS_ ## stat, val)
cur               106 fs/xfs/libxfs/xfs_btree.h 	void	(*set_root)(struct xfs_btree_cur *cur,
cur               110 fs/xfs/libxfs/xfs_btree.h 	int	(*alloc_block)(struct xfs_btree_cur *cur,
cur               114 fs/xfs/libxfs/xfs_btree.h 	int	(*free_block)(struct xfs_btree_cur *cur, struct xfs_buf *bp);
cur               117 fs/xfs/libxfs/xfs_btree.h 	void	(*update_lastrec)(struct xfs_btree_cur *cur,
cur               123 fs/xfs/libxfs/xfs_btree.h 	int	(*get_minrecs)(struct xfs_btree_cur *cur, int level);
cur               124 fs/xfs/libxfs/xfs_btree.h 	int	(*get_maxrecs)(struct xfs_btree_cur *cur, int level);
cur               127 fs/xfs/libxfs/xfs_btree.h 	int	(*get_dmaxrecs)(struct xfs_btree_cur *cur, int level);
cur               132 fs/xfs/libxfs/xfs_btree.h 	void	(*init_rec_from_cur)(struct xfs_btree_cur *cur,
cur               134 fs/xfs/libxfs/xfs_btree.h 	void	(*init_ptr_from_cur)(struct xfs_btree_cur *cur,
cur               140 fs/xfs/libxfs/xfs_btree.h 	int64_t (*key_diff)(struct xfs_btree_cur *cur,
cur               147 fs/xfs/libxfs/xfs_btree.h 	int64_t (*diff_two_keys)(struct xfs_btree_cur *cur,
cur               154 fs/xfs/libxfs/xfs_btree.h 	int	(*keys_inorder)(struct xfs_btree_cur *cur,
cur               159 fs/xfs/libxfs/xfs_btree.h 	int	(*recs_inorder)(struct xfs_btree_cur *cur,
cur               246 fs/xfs/libxfs/xfs_btree.h xfs_failaddr_t __xfs_btree_check_lblock(struct xfs_btree_cur *cur,
cur               248 fs/xfs/libxfs/xfs_btree.h xfs_failaddr_t __xfs_btree_check_sblock(struct xfs_btree_cur *cur,
cur               256 fs/xfs/libxfs/xfs_btree.h 	struct xfs_btree_cur	*cur,	/* btree cursor */
cur               266 fs/xfs/libxfs/xfs_btree.h 	struct xfs_btree_cur	*cur,	/* btree cursor */
cur               275 fs/xfs/libxfs/xfs_btree.h 	struct xfs_btree_cur	*cur,	/* btree cursor */
cur               284 fs/xfs/libxfs/xfs_btree.h 	xfs_btree_cur_t		*cur,	/* btree cursor */
cur               293 fs/xfs/libxfs/xfs_btree.h 	xfs_btree_cur_t		*cur,	/* input cursor */
cur               322 fs/xfs/libxfs/xfs_btree.h 	xfs_btree_cur_t		*cur,	/* btree cursor */
cur               407 fs/xfs/libxfs/xfs_btree.h int xfs_btree_change_owner(struct xfs_btree_cur *cur, uint64_t new_owner,
cur               474 fs/xfs/libxfs/xfs_btree.h typedef int (*xfs_btree_query_range_fn)(struct xfs_btree_cur *cur,
cur               477 fs/xfs/libxfs/xfs_btree.h int xfs_btree_query_range(struct xfs_btree_cur *cur,
cur               480 fs/xfs/libxfs/xfs_btree.h int xfs_btree_query_all(struct xfs_btree_cur *cur, xfs_btree_query_range_fn fn,
cur               483 fs/xfs/libxfs/xfs_btree.h typedef int (*xfs_btree_visit_blocks_fn)(struct xfs_btree_cur *cur, int level,
cur               485 fs/xfs/libxfs/xfs_btree.h int xfs_btree_visit_blocks(struct xfs_btree_cur *cur,
cur               488 fs/xfs/libxfs/xfs_btree.h int xfs_btree_count_blocks(struct xfs_btree_cur *cur, xfs_extlen_t *blocks);
cur               490 fs/xfs/libxfs/xfs_btree.h union xfs_btree_rec *xfs_btree_rec_addr(struct xfs_btree_cur *cur, int n,
cur               492 fs/xfs/libxfs/xfs_btree.h union xfs_btree_key *xfs_btree_key_addr(struct xfs_btree_cur *cur, int n,
cur               494 fs/xfs/libxfs/xfs_btree.h union xfs_btree_key *xfs_btree_high_key_addr(struct xfs_btree_cur *cur, int n,
cur               496 fs/xfs/libxfs/xfs_btree.h union xfs_btree_ptr *xfs_btree_ptr_addr(struct xfs_btree_cur *cur, int n,
cur               498 fs/xfs/libxfs/xfs_btree.h int xfs_btree_lookup_get_block(struct xfs_btree_cur *cur, int level,
cur               500 fs/xfs/libxfs/xfs_btree.h struct xfs_btree_block *xfs_btree_get_block(struct xfs_btree_cur *cur,
cur               502 fs/xfs/libxfs/xfs_btree.h bool xfs_btree_ptr_is_null(struct xfs_btree_cur *cur, union xfs_btree_ptr *ptr);
cur               503 fs/xfs/libxfs/xfs_btree.h int64_t xfs_btree_diff_two_ptrs(struct xfs_btree_cur *cur,
cur               506 fs/xfs/libxfs/xfs_btree.h void xfs_btree_get_sibling(struct xfs_btree_cur *cur,
cur               509 fs/xfs/libxfs/xfs_btree.h void xfs_btree_get_keys(struct xfs_btree_cur *cur,
cur               511 fs/xfs/libxfs/xfs_btree.h union xfs_btree_key *xfs_btree_high_key_from_key(struct xfs_btree_cur *cur,
cur               513 fs/xfs/libxfs/xfs_btree.h int xfs_btree_has_record(struct xfs_btree_cur *cur, union xfs_btree_irec *low,
cur               515 fs/xfs/libxfs/xfs_btree.h bool xfs_btree_has_more_records(struct xfs_btree_cur *cur);
cur                36 fs/xfs/libxfs/xfs_ialloc.c 	struct xfs_btree_cur	*cur,	/* btree cursor */
cur                41 fs/xfs/libxfs/xfs_ialloc.c 	cur->bc_rec.i.ir_startino = ino;
cur                42 fs/xfs/libxfs/xfs_ialloc.c 	cur->bc_rec.i.ir_holemask = 0;
cur                43 fs/xfs/libxfs/xfs_ialloc.c 	cur->bc_rec.i.ir_count = 0;
cur                44 fs/xfs/libxfs/xfs_ialloc.c 	cur->bc_rec.i.ir_freecount = 0;
cur                45 fs/xfs/libxfs/xfs_ialloc.c 	cur->bc_rec.i.ir_free = 0;
cur                46 fs/xfs/libxfs/xfs_ialloc.c 	return xfs_btree_lookup(cur, dir, stat);
cur                55 fs/xfs/libxfs/xfs_ialloc.c 	struct xfs_btree_cur	*cur,	/* btree cursor */
cur                61 fs/xfs/libxfs/xfs_ialloc.c 	if (xfs_sb_version_hassparseinodes(&cur->bc_mp->m_sb)) {
cur                70 fs/xfs/libxfs/xfs_ialloc.c 	return xfs_btree_update(cur, &rec);
cur               103 fs/xfs/libxfs/xfs_ialloc.c 	struct xfs_btree_cur		*cur,
cur               107 fs/xfs/libxfs/xfs_ialloc.c 	struct xfs_mount		*mp = cur->bc_mp;
cur               108 fs/xfs/libxfs/xfs_ialloc.c 	xfs_agnumber_t			agno = cur->bc_private.a.agno;
cur               113 fs/xfs/libxfs/xfs_ialloc.c 	error = xfs_btree_get_rec(cur, &rec, stat);
cur               140 fs/xfs/libxfs/xfs_ialloc.c 		cur->bc_btnum == XFS_BTNUM_INO ? "Used" : "Free", agno);
cur               153 fs/xfs/libxfs/xfs_ialloc.c 	struct xfs_btree_cur	*cur,
cur               160 fs/xfs/libxfs/xfs_ialloc.c 	cur->bc_rec.i.ir_holemask = holemask;
cur               161 fs/xfs/libxfs/xfs_ialloc.c 	cur->bc_rec.i.ir_count = count;
cur               162 fs/xfs/libxfs/xfs_ialloc.c 	cur->bc_rec.i.ir_freecount = freecount;
cur               163 fs/xfs/libxfs/xfs_ialloc.c 	cur->bc_rec.i.ir_free = free;
cur               164 fs/xfs/libxfs/xfs_ialloc.c 	return xfs_btree_insert(cur, stat);
cur               179 fs/xfs/libxfs/xfs_ialloc.c 	struct xfs_btree_cur	*cur;
cur               186 fs/xfs/libxfs/xfs_ialloc.c 	cur = xfs_inobt_init_cursor(mp, tp, agbp, agno, btnum);
cur               191 fs/xfs/libxfs/xfs_ialloc.c 		error = xfs_inobt_lookup(cur, thisino, XFS_LOOKUP_EQ, &i);
cur               193 fs/xfs/libxfs/xfs_ialloc.c 			xfs_btree_del_cursor(cur, XFS_BTREE_ERROR);
cur               198 fs/xfs/libxfs/xfs_ialloc.c 		error = xfs_inobt_insert_rec(cur, XFS_INOBT_HOLEMASK_FULL,
cur               203 fs/xfs/libxfs/xfs_ialloc.c 			xfs_btree_del_cursor(cur, XFS_BTREE_ERROR);
cur               209 fs/xfs/libxfs/xfs_ialloc.c 	xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR);
cur               220 fs/xfs/libxfs/xfs_ialloc.c 	struct xfs_btree_cur	*cur,
cur               223 fs/xfs/libxfs/xfs_ialloc.c 	if (cur->bc_nlevels == 1) {
cur               229 fs/xfs/libxfs/xfs_ialloc.c 		error = xfs_inobt_lookup(cur, 0, XFS_LOOKUP_GE, &i);
cur               234 fs/xfs/libxfs/xfs_ialloc.c 			error = xfs_inobt_get_rec(cur, &rec, &i);
cur               240 fs/xfs/libxfs/xfs_ialloc.c 				error = xfs_btree_increment(cur, 0, &i);
cur               246 fs/xfs/libxfs/xfs_ialloc.c 		if (!XFS_FORCED_SHUTDOWN(cur->bc_mp))
cur               252 fs/xfs/libxfs/xfs_ialloc.c #define xfs_check_agi_freecount(cur, agi)	0
cur               527 fs/xfs/libxfs/xfs_ialloc.c 	struct xfs_btree_cur		*cur;
cur               534 fs/xfs/libxfs/xfs_ialloc.c 	cur = xfs_inobt_init_cursor(mp, tp, agbp, agno, btnum);
cur               537 fs/xfs/libxfs/xfs_ialloc.c 	error = xfs_inobt_lookup(cur, nrec->ir_startino, XFS_LOOKUP_EQ, &i);
cur               542 fs/xfs/libxfs/xfs_ialloc.c 		error = xfs_inobt_insert_rec(cur, nrec->ir_holemask,
cur               557 fs/xfs/libxfs/xfs_ialloc.c 		error = xfs_inobt_get_rec(cur, &rec, &i);
cur               587 fs/xfs/libxfs/xfs_ialloc.c 	error = xfs_inobt_update(cur, nrec);
cur               592 fs/xfs/libxfs/xfs_ialloc.c 	xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR);
cur               595 fs/xfs/libxfs/xfs_ialloc.c 	xfs_btree_del_cursor(cur, XFS_BTREE_ERROR);
cur              1040 fs/xfs/libxfs/xfs_ialloc.c 	struct xfs_btree_cur	*cur,
cur              1049 fs/xfs/libxfs/xfs_ialloc.c 		error = xfs_btree_decrement(cur, 0, &i);
cur              1051 fs/xfs/libxfs/xfs_ialloc.c 		error = xfs_btree_increment(cur, 0, &i);
cur              1057 fs/xfs/libxfs/xfs_ialloc.c 		error = xfs_inobt_get_rec(cur, rec, &i);
cur              1060 fs/xfs/libxfs/xfs_ialloc.c 		XFS_WANT_CORRUPTED_RETURN(cur->bc_mp, i == 1);
cur              1068 fs/xfs/libxfs/xfs_ialloc.c 	struct xfs_btree_cur	*cur,
cur              1076 fs/xfs/libxfs/xfs_ialloc.c 	error = xfs_inobt_lookup(cur, agino, XFS_LOOKUP_EQ, &i);
cur              1081 fs/xfs/libxfs/xfs_ialloc.c 		error = xfs_inobt_get_rec(cur, rec, &i);
cur              1084 fs/xfs/libxfs/xfs_ialloc.c 		XFS_WANT_CORRUPTED_RETURN(cur->bc_mp, i == 1);
cur              1127 fs/xfs/libxfs/xfs_ialloc.c 	struct xfs_btree_cur	*cur, *tcur;
cur              1142 fs/xfs/libxfs/xfs_ialloc.c 	cur = xfs_inobt_init_cursor(mp, tp, agbp, agno, XFS_BTNUM_INO);
cur              1150 fs/xfs/libxfs/xfs_ialloc.c 	error = xfs_check_agi_freecount(cur, agi);
cur              1161 fs/xfs/libxfs/xfs_ialloc.c 		error = xfs_inobt_lookup(cur, pagino, XFS_LOOKUP_LE, &i);
cur              1166 fs/xfs/libxfs/xfs_ialloc.c 		error = xfs_inobt_get_rec(cur, &rec, &j);
cur              1185 fs/xfs/libxfs/xfs_ialloc.c 		error = xfs_btree_dup_cursor(cur, &tcur);
cur              1201 fs/xfs/libxfs/xfs_ialloc.c 			error = xfs_ialloc_get_rec(cur, pag->pagl_rightrec,
cur              1212 fs/xfs/libxfs/xfs_ialloc.c 			error = xfs_ialloc_next_rec(cur, &rec, &doneright, 0);
cur              1234 fs/xfs/libxfs/xfs_ialloc.c 				xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR);
cur              1235 fs/xfs/libxfs/xfs_ialloc.c 				cur = tcur;
cur              1259 fs/xfs/libxfs/xfs_ialloc.c 				error = xfs_ialloc_next_rec(cur, &rec,
cur              1288 fs/xfs/libxfs/xfs_ialloc.c 			xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR);
cur              1298 fs/xfs/libxfs/xfs_ialloc.c 		error = xfs_inobt_lookup(cur, be32_to_cpu(agi->agi_newino),
cur              1304 fs/xfs/libxfs/xfs_ialloc.c 			error = xfs_inobt_get_rec(cur, &rec, &j);
cur              1321 fs/xfs/libxfs/xfs_ialloc.c 	error = xfs_inobt_lookup(cur, 0, XFS_LOOKUP_GE, &i);
cur              1327 fs/xfs/libxfs/xfs_ialloc.c 		error = xfs_inobt_get_rec(cur, &rec, &i);
cur              1333 fs/xfs/libxfs/xfs_ialloc.c 		error = xfs_btree_increment(cur, 0, &i);
cur              1348 fs/xfs/libxfs/xfs_ialloc.c 	error = xfs_inobt_update(cur, &rec);
cur              1355 fs/xfs/libxfs/xfs_ialloc.c 	error = xfs_check_agi_freecount(cur, agi);
cur              1359 fs/xfs/libxfs/xfs_ialloc.c 	xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR);
cur              1367 fs/xfs/libxfs/xfs_ialloc.c 	xfs_btree_del_cursor(cur, XFS_BTREE_ERROR);
cur              1460 fs/xfs/libxfs/xfs_ialloc.c 	struct xfs_btree_cur		*cur,
cur              1467 fs/xfs/libxfs/xfs_ialloc.c 		error = xfs_inobt_lookup(cur, be32_to_cpu(agi->agi_newino),
cur              1472 fs/xfs/libxfs/xfs_ialloc.c 			error = xfs_inobt_get_rec(cur, rec, &i);
cur              1475 fs/xfs/libxfs/xfs_ialloc.c 			XFS_WANT_CORRUPTED_RETURN(cur->bc_mp, i == 1);
cur              1483 fs/xfs/libxfs/xfs_ialloc.c 	error = xfs_inobt_lookup(cur, 0, XFS_LOOKUP_GE, &i);
cur              1486 fs/xfs/libxfs/xfs_ialloc.c 	XFS_WANT_CORRUPTED_RETURN(cur->bc_mp, i == 1);
cur              1488 fs/xfs/libxfs/xfs_ialloc.c 	error = xfs_inobt_get_rec(cur, rec, &i);
cur              1491 fs/xfs/libxfs/xfs_ialloc.c 	XFS_WANT_CORRUPTED_RETURN(cur->bc_mp, i == 1);
cur              1502 fs/xfs/libxfs/xfs_ialloc.c 	struct xfs_btree_cur		*cur,	/* inobt cursor */
cur              1510 fs/xfs/libxfs/xfs_ialloc.c 	error = xfs_inobt_lookup(cur, frec->ir_startino, XFS_LOOKUP_EQ, &i);
cur              1513 fs/xfs/libxfs/xfs_ialloc.c 	XFS_WANT_CORRUPTED_RETURN(cur->bc_mp, i == 1);
cur              1515 fs/xfs/libxfs/xfs_ialloc.c 	error = xfs_inobt_get_rec(cur, &rec, &i);
cur              1518 fs/xfs/libxfs/xfs_ialloc.c 	XFS_WANT_CORRUPTED_RETURN(cur->bc_mp, i == 1);
cur              1519 fs/xfs/libxfs/xfs_ialloc.c 	ASSERT((XFS_AGINO_TO_OFFSET(cur->bc_mp, rec.ir_startino) %
cur              1525 fs/xfs/libxfs/xfs_ialloc.c 	XFS_WANT_CORRUPTED_RETURN(cur->bc_mp, (rec.ir_free == frec->ir_free) &&
cur              1528 fs/xfs/libxfs/xfs_ialloc.c 	return xfs_inobt_update(cur, &rec);
cur              1551 fs/xfs/libxfs/xfs_ialloc.c 	struct xfs_btree_cur		*cur;	/* finobt cursor */
cur              1571 fs/xfs/libxfs/xfs_ialloc.c 	cur = xfs_inobt_init_cursor(mp, tp, agbp, agno, XFS_BTNUM_FINO);
cur              1573 fs/xfs/libxfs/xfs_ialloc.c 	error = xfs_check_agi_freecount(cur, agi);
cur              1583 fs/xfs/libxfs/xfs_ialloc.c 		error = xfs_dialloc_ag_finobt_near(pagino, &cur, &rec);
cur              1585 fs/xfs/libxfs/xfs_ialloc.c 		error = xfs_dialloc_ag_finobt_newino(agi, cur, &rec);
cur              1602 fs/xfs/libxfs/xfs_ialloc.c 		error = xfs_inobt_update(cur, &rec);
cur              1604 fs/xfs/libxfs/xfs_ialloc.c 		error = xfs_btree_delete(cur, &i);
cur              1637 fs/xfs/libxfs/xfs_ialloc.c 	error = xfs_check_agi_freecount(cur, agi);
cur              1642 fs/xfs/libxfs/xfs_ialloc.c 	xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR);
cur              1650 fs/xfs/libxfs/xfs_ialloc.c 	xfs_btree_del_cursor(cur, XFS_BTREE_ERROR);
cur              1909 fs/xfs/libxfs/xfs_ialloc.c 	struct xfs_btree_cur		*cur;
cur              1922 fs/xfs/libxfs/xfs_ialloc.c 	cur = xfs_inobt_init_cursor(mp, tp, agbp, agno, XFS_BTNUM_INO);
cur              1924 fs/xfs/libxfs/xfs_ialloc.c 	error = xfs_check_agi_freecount(cur, agi);
cur              1931 fs/xfs/libxfs/xfs_ialloc.c 	if ((error = xfs_inobt_lookup(cur, agino, XFS_LOOKUP_LE, &i))) {
cur              1937 fs/xfs/libxfs/xfs_ialloc.c 	error = xfs_inobt_get_rec(cur, &rec, &i);
cur              1984 fs/xfs/libxfs/xfs_ialloc.c 		if ((error = xfs_btree_delete(cur, &i))) {
cur              1994 fs/xfs/libxfs/xfs_ialloc.c 		error = xfs_inobt_update(cur, &rec);
cur              2012 fs/xfs/libxfs/xfs_ialloc.c 	error = xfs_check_agi_freecount(cur, agi);
cur              2017 fs/xfs/libxfs/xfs_ialloc.c 	xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR);
cur              2021 fs/xfs/libxfs/xfs_ialloc.c 	xfs_btree_del_cursor(cur, XFS_BTREE_ERROR);
cur              2038 fs/xfs/libxfs/xfs_ialloc.c 	struct xfs_btree_cur		*cur;
cur              2044 fs/xfs/libxfs/xfs_ialloc.c 	cur = xfs_inobt_init_cursor(mp, tp, agbp, agno, XFS_BTNUM_FINO);
cur              2046 fs/xfs/libxfs/xfs_ialloc.c 	error = xfs_inobt_lookup(cur, ibtrec->ir_startino, XFS_LOOKUP_EQ, &i);
cur              2057 fs/xfs/libxfs/xfs_ialloc.c 		error = xfs_inobt_insert_rec(cur, ibtrec->ir_holemask,
cur              2075 fs/xfs/libxfs/xfs_ialloc.c 	error = xfs_inobt_get_rec(cur, &rec, &i);
cur              2102 fs/xfs/libxfs/xfs_ialloc.c 		error = xfs_btree_delete(cur, &i);
cur              2107 fs/xfs/libxfs/xfs_ialloc.c 		error = xfs_inobt_update(cur, &rec);
cur              2113 fs/xfs/libxfs/xfs_ialloc.c 	error = xfs_check_agi_freecount(cur, agi);
cur              2117 fs/xfs/libxfs/xfs_ialloc.c 	xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR);
cur              2121 fs/xfs/libxfs/xfs_ialloc.c 	xfs_btree_del_cursor(cur, XFS_BTREE_ERROR);
cur              2217 fs/xfs/libxfs/xfs_ialloc.c 	struct xfs_btree_cur	*cur;
cur              2236 fs/xfs/libxfs/xfs_ialloc.c 	cur = xfs_inobt_init_cursor(mp, tp, agbp, agno, XFS_BTNUM_INO);
cur              2237 fs/xfs/libxfs/xfs_ialloc.c 	error = xfs_inobt_lookup(cur, agino, XFS_LOOKUP_LE, &i);
cur              2240 fs/xfs/libxfs/xfs_ialloc.c 			error = xfs_inobt_get_rec(cur, &rec, &i);
cur              2246 fs/xfs/libxfs/xfs_ialloc.c 	xfs_btree_del_cursor(cur, error);
cur              2650 fs/xfs/libxfs/xfs_ialloc.c 	struct xfs_btree_cur	*cur,
cur              2663 fs/xfs/libxfs/xfs_ialloc.c 	error = xfs_inobt_lookup(cur, low, XFS_LOOKUP_LE, &has_record);
cur              2665 fs/xfs/libxfs/xfs_ialloc.c 		error = xfs_inobt_get_rec(cur, &irec, &has_record);
cur              2682 fs/xfs/libxfs/xfs_ialloc.c 		error = xfs_btree_increment(cur, 0, &has_record);
cur              2690 fs/xfs/libxfs/xfs_ialloc.c 	struct xfs_btree_cur	*cur,
cur              2698 fs/xfs/libxfs/xfs_ialloc.c 	low = XFS_AGB_TO_AGINO(cur->bc_mp, bno);
cur              2699 fs/xfs/libxfs/xfs_ialloc.c 	high = XFS_AGB_TO_AGINO(cur->bc_mp, bno + len) - 1;
cur              2701 fs/xfs/libxfs/xfs_ialloc.c 	return xfs_ialloc_has_inode_record(cur, low, high, exists);
cur              2712 fs/xfs/libxfs/xfs_ialloc.c 	struct xfs_btree_cur		*cur,
cur              2719 fs/xfs/libxfs/xfs_ialloc.c 	xfs_inobt_btrec_to_irec(cur->bc_mp, rec, &irec);
cur              2729 fs/xfs/libxfs/xfs_ialloc.c 	struct xfs_btree_cur		*cur,
cur              2736 fs/xfs/libxfs/xfs_ialloc.c 	ASSERT(cur->bc_btnum == XFS_BTNUM_INO);
cur              2737 fs/xfs/libxfs/xfs_ialloc.c 	error = xfs_btree_query_all(cur, xfs_ialloc_count_inodes_rec, &ci);
cur               120 fs/xfs/libxfs/xfs_ialloc.h int xfs_inobt_lookup(struct xfs_btree_cur *cur, xfs_agino_t ino,
cur               126 fs/xfs/libxfs/xfs_ialloc.h int xfs_inobt_get_rec(struct xfs_btree_cur *cur,
cur               143 fs/xfs/libxfs/xfs_ialloc.h int xfs_ialloc_has_inodes_at_extent(struct xfs_btree_cur *cur,
cur               145 fs/xfs/libxfs/xfs_ialloc.h int xfs_ialloc_has_inode_record(struct xfs_btree_cur *cur, xfs_agino_t low,
cur               147 fs/xfs/libxfs/xfs_ialloc.h int xfs_ialloc_count_inodes(struct xfs_btree_cur *cur, xfs_agino_t *count,
cur               149 fs/xfs/libxfs/xfs_ialloc.h int xfs_inobt_insert_rec(struct xfs_btree_cur *cur, uint16_t holemask,
cur                26 fs/xfs/libxfs/xfs_ialloc_btree.c 	struct xfs_btree_cur	*cur,
cur                29 fs/xfs/libxfs/xfs_ialloc_btree.c 	return M_IGEO(cur->bc_mp)->inobt_mnr[level != 0];
cur                34 fs/xfs/libxfs/xfs_ialloc_btree.c 	struct xfs_btree_cur	*cur)
cur                36 fs/xfs/libxfs/xfs_ialloc_btree.c 	return xfs_inobt_init_cursor(cur->bc_mp, cur->bc_tp,
cur                37 fs/xfs/libxfs/xfs_ialloc_btree.c 			cur->bc_private.a.agbp, cur->bc_private.a.agno,
cur                38 fs/xfs/libxfs/xfs_ialloc_btree.c 			cur->bc_btnum);
cur                43 fs/xfs/libxfs/xfs_ialloc_btree.c 	struct xfs_btree_cur	*cur,
cur                47 fs/xfs/libxfs/xfs_ialloc_btree.c 	struct xfs_buf		*agbp = cur->bc_private.a.agbp;
cur                52 fs/xfs/libxfs/xfs_ialloc_btree.c 	xfs_ialloc_log_agi(cur->bc_tp, agbp, XFS_AGI_ROOT | XFS_AGI_LEVEL);
cur                57 fs/xfs/libxfs/xfs_ialloc_btree.c 	struct xfs_btree_cur	*cur,
cur                61 fs/xfs/libxfs/xfs_ialloc_btree.c 	struct xfs_buf		*agbp = cur->bc_private.a.agbp;
cur                66 fs/xfs/libxfs/xfs_ialloc_btree.c 	xfs_ialloc_log_agi(cur->bc_tp, agbp,
cur                72 fs/xfs/libxfs/xfs_ialloc_btree.c 	struct xfs_btree_cur	*cur,
cur                83 fs/xfs/libxfs/xfs_ialloc_btree.c 	args.tp = cur->bc_tp;
cur                84 fs/xfs/libxfs/xfs_ialloc_btree.c 	args.mp = cur->bc_mp;
cur                86 fs/xfs/libxfs/xfs_ialloc_btree.c 	args.fsbno = XFS_AGB_TO_FSB(args.mp, cur->bc_private.a.agno, sbno);
cur               110 fs/xfs/libxfs/xfs_ialloc_btree.c 	struct xfs_btree_cur	*cur,
cur               115 fs/xfs/libxfs/xfs_ialloc_btree.c 	return __xfs_inobt_alloc_block(cur, start, new, stat, XFS_AG_RESV_NONE);
cur               120 fs/xfs/libxfs/xfs_ialloc_btree.c 	struct xfs_btree_cur	*cur,
cur               125 fs/xfs/libxfs/xfs_ialloc_btree.c 	if (cur->bc_mp->m_finobt_nores)
cur               126 fs/xfs/libxfs/xfs_ialloc_btree.c 		return xfs_inobt_alloc_block(cur, start, new, stat);
cur               127 fs/xfs/libxfs/xfs_ialloc_btree.c 	return __xfs_inobt_alloc_block(cur, start, new, stat,
cur               133 fs/xfs/libxfs/xfs_ialloc_btree.c 	struct xfs_btree_cur	*cur,
cur               137 fs/xfs/libxfs/xfs_ialloc_btree.c 	return xfs_free_extent(cur->bc_tp,
cur               138 fs/xfs/libxfs/xfs_ialloc_btree.c 			XFS_DADDR_TO_FSB(cur->bc_mp, XFS_BUF_ADDR(bp)), 1,
cur               144 fs/xfs/libxfs/xfs_ialloc_btree.c 	struct xfs_btree_cur	*cur,
cur               147 fs/xfs/libxfs/xfs_ialloc_btree.c 	return __xfs_inobt_free_block(cur, bp, XFS_AG_RESV_NONE);
cur               152 fs/xfs/libxfs/xfs_ialloc_btree.c 	struct xfs_btree_cur	*cur,
cur               155 fs/xfs/libxfs/xfs_ialloc_btree.c 	if (cur->bc_mp->m_finobt_nores)
cur               156 fs/xfs/libxfs/xfs_ialloc_btree.c 		return xfs_inobt_free_block(cur, bp);
cur               157 fs/xfs/libxfs/xfs_ialloc_btree.c 	return __xfs_inobt_free_block(cur, bp, XFS_AG_RESV_METADATA);
cur               162 fs/xfs/libxfs/xfs_ialloc_btree.c 	struct xfs_btree_cur	*cur,
cur               165 fs/xfs/libxfs/xfs_ialloc_btree.c 	return M_IGEO(cur->bc_mp)->inobt_mxr[level != 0];
cur               190 fs/xfs/libxfs/xfs_ialloc_btree.c 	struct xfs_btree_cur	*cur,
cur               193 fs/xfs/libxfs/xfs_ialloc_btree.c 	rec->inobt.ir_startino = cpu_to_be32(cur->bc_rec.i.ir_startino);
cur               194 fs/xfs/libxfs/xfs_ialloc_btree.c 	if (xfs_sb_version_hassparseinodes(&cur->bc_mp->m_sb)) {
cur               196 fs/xfs/libxfs/xfs_ialloc_btree.c 					cpu_to_be16(cur->bc_rec.i.ir_holemask);
cur               197 fs/xfs/libxfs/xfs_ialloc_btree.c 		rec->inobt.ir_u.sp.ir_count = cur->bc_rec.i.ir_count;
cur               198 fs/xfs/libxfs/xfs_ialloc_btree.c 		rec->inobt.ir_u.sp.ir_freecount = cur->bc_rec.i.ir_freecount;
cur               202 fs/xfs/libxfs/xfs_ialloc_btree.c 					cpu_to_be32(cur->bc_rec.i.ir_freecount);
cur               204 fs/xfs/libxfs/xfs_ialloc_btree.c 	rec->inobt.ir_free = cpu_to_be64(cur->bc_rec.i.ir_free);
cur               212 fs/xfs/libxfs/xfs_ialloc_btree.c 	struct xfs_btree_cur	*cur,
cur               215 fs/xfs/libxfs/xfs_ialloc_btree.c 	struct xfs_agi		*agi = XFS_BUF_TO_AGI(cur->bc_private.a.agbp);
cur               217 fs/xfs/libxfs/xfs_ialloc_btree.c 	ASSERT(cur->bc_private.a.agno == be32_to_cpu(agi->agi_seqno));
cur               224 fs/xfs/libxfs/xfs_ialloc_btree.c 	struct xfs_btree_cur	*cur,
cur               227 fs/xfs/libxfs/xfs_ialloc_btree.c 	struct xfs_agi		*agi = XFS_BUF_TO_AGI(cur->bc_private.a.agbp);
cur               229 fs/xfs/libxfs/xfs_ialloc_btree.c 	ASSERT(cur->bc_private.a.agno == be32_to_cpu(agi->agi_seqno));
cur               235 fs/xfs/libxfs/xfs_ialloc_btree.c 	struct xfs_btree_cur	*cur,
cur               239 fs/xfs/libxfs/xfs_ialloc_btree.c 			  cur->bc_rec.i.ir_startino;
cur               244 fs/xfs/libxfs/xfs_ialloc_btree.c 	struct xfs_btree_cur	*cur,
cur               342 fs/xfs/libxfs/xfs_ialloc_btree.c 	struct xfs_btree_cur	*cur,
cur               352 fs/xfs/libxfs/xfs_ialloc_btree.c 	struct xfs_btree_cur	*cur,
cur               414 fs/xfs/libxfs/xfs_ialloc_btree.c 	struct xfs_btree_cur	*cur;
cur               416 fs/xfs/libxfs/xfs_ialloc_btree.c 	cur = kmem_zone_zalloc(xfs_btree_cur_zone, KM_NOFS);
cur               418 fs/xfs/libxfs/xfs_ialloc_btree.c 	cur->bc_tp = tp;
cur               419 fs/xfs/libxfs/xfs_ialloc_btree.c 	cur->bc_mp = mp;
cur               420 fs/xfs/libxfs/xfs_ialloc_btree.c 	cur->bc_btnum = btnum;
cur               422 fs/xfs/libxfs/xfs_ialloc_btree.c 		cur->bc_nlevels = be32_to_cpu(agi->agi_level);
cur               423 fs/xfs/libxfs/xfs_ialloc_btree.c 		cur->bc_ops = &xfs_inobt_ops;
cur               424 fs/xfs/libxfs/xfs_ialloc_btree.c 		cur->bc_statoff = XFS_STATS_CALC_INDEX(xs_ibt_2);
cur               426 fs/xfs/libxfs/xfs_ialloc_btree.c 		cur->bc_nlevels = be32_to_cpu(agi->agi_free_level);
cur               427 fs/xfs/libxfs/xfs_ialloc_btree.c 		cur->bc_ops = &xfs_finobt_ops;
cur               428 fs/xfs/libxfs/xfs_ialloc_btree.c 		cur->bc_statoff = XFS_STATS_CALC_INDEX(xs_fibt_2);
cur               431 fs/xfs/libxfs/xfs_ialloc_btree.c 	cur->bc_blocklog = mp->m_sb.sb_blocklog;
cur               434 fs/xfs/libxfs/xfs_ialloc_btree.c 		cur->bc_flags |= XFS_BTREE_CRC_BLOCKS;
cur               436 fs/xfs/libxfs/xfs_ialloc_btree.c 	cur->bc_private.a.agbp = agbp;
cur               437 fs/xfs/libxfs/xfs_ialloc_btree.c 	cur->bc_private.a.agno = agno;
cur               439 fs/xfs/libxfs/xfs_ialloc_btree.c 	return cur;
cur               575 fs/xfs/libxfs/xfs_ialloc_btree.c 	struct xfs_btree_cur	*cur;
cur               585 fs/xfs/libxfs/xfs_ialloc_btree.c 	cur = xfs_inobt_init_cursor(mp, tp, *agi_bpp, agno, which);
cur               586 fs/xfs/libxfs/xfs_ialloc_btree.c 	if (!cur) {
cur               591 fs/xfs/libxfs/xfs_ialloc_btree.c 	*curpp = cur;
cur               604 fs/xfs/libxfs/xfs_ialloc_btree.c 	struct xfs_btree_cur	*cur = NULL;
cur               607 fs/xfs/libxfs/xfs_ialloc_btree.c 	error = xfs_inobt_cur(mp, tp, agno, btnum, &cur, &agbp);
cur               611 fs/xfs/libxfs/xfs_ialloc_btree.c 	error = xfs_btree_count_blocks(cur, tree_blocks);
cur               612 fs/xfs/libxfs/xfs_ialloc_btree.c 	xfs_btree_del_cursor(cur, error);
cur               140 fs/xfs/libxfs/xfs_iext_tree.c static inline struct xfs_iext_rec *cur_rec(struct xfs_iext_cursor *cur)
cur               142 fs/xfs/libxfs/xfs_iext_tree.c 	return &cur->leaf->recs[cur->pos];
cur               146 fs/xfs/libxfs/xfs_iext_tree.c 		struct xfs_iext_cursor *cur)
cur               148 fs/xfs/libxfs/xfs_iext_tree.c 	if (!cur->leaf)
cur               150 fs/xfs/libxfs/xfs_iext_tree.c 	if (cur->pos < 0 || cur->pos >= xfs_iext_max_recs(ifp))
cur               152 fs/xfs/libxfs/xfs_iext_tree.c 	if (xfs_iext_rec_is_empty(cur_rec(cur)))
cur               199 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_cursor	*cur)
cur               201 fs/xfs/libxfs/xfs_iext_tree.c 	cur->pos = 0;
cur               202 fs/xfs/libxfs/xfs_iext_tree.c 	cur->leaf = xfs_iext_find_first_leaf(ifp);
cur               208 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_cursor	*cur)
cur               212 fs/xfs/libxfs/xfs_iext_tree.c 	cur->leaf = xfs_iext_find_last_leaf(ifp);
cur               213 fs/xfs/libxfs/xfs_iext_tree.c 	if (!cur->leaf) {
cur               214 fs/xfs/libxfs/xfs_iext_tree.c 		cur->pos = 0;
cur               219 fs/xfs/libxfs/xfs_iext_tree.c 		if (xfs_iext_rec_is_empty(&cur->leaf->recs[i]))
cur               222 fs/xfs/libxfs/xfs_iext_tree.c 	cur->pos = i - 1;
cur               228 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_cursor	*cur)
cur               230 fs/xfs/libxfs/xfs_iext_tree.c 	if (!cur->leaf) {
cur               231 fs/xfs/libxfs/xfs_iext_tree.c 		ASSERT(cur->pos <= 0 || cur->pos >= RECS_PER_LEAF);
cur               232 fs/xfs/libxfs/xfs_iext_tree.c 		xfs_iext_first(ifp, cur);
cur               236 fs/xfs/libxfs/xfs_iext_tree.c 	ASSERT(cur->pos >= 0);
cur               237 fs/xfs/libxfs/xfs_iext_tree.c 	ASSERT(cur->pos < xfs_iext_max_recs(ifp));
cur               239 fs/xfs/libxfs/xfs_iext_tree.c 	cur->pos++;
cur               240 fs/xfs/libxfs/xfs_iext_tree.c 	if (ifp->if_height > 1 && !xfs_iext_valid(ifp, cur) &&
cur               241 fs/xfs/libxfs/xfs_iext_tree.c 	    cur->leaf->next) {
cur               242 fs/xfs/libxfs/xfs_iext_tree.c 		cur->leaf = cur->leaf->next;
cur               243 fs/xfs/libxfs/xfs_iext_tree.c 		cur->pos = 0;
cur               250 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_cursor	*cur)
cur               252 fs/xfs/libxfs/xfs_iext_tree.c 	if (!cur->leaf) {
cur               253 fs/xfs/libxfs/xfs_iext_tree.c 		ASSERT(cur->pos <= 0 || cur->pos >= RECS_PER_LEAF);
cur               254 fs/xfs/libxfs/xfs_iext_tree.c 		xfs_iext_last(ifp, cur);
cur               258 fs/xfs/libxfs/xfs_iext_tree.c 	ASSERT(cur->pos >= 0);
cur               259 fs/xfs/libxfs/xfs_iext_tree.c 	ASSERT(cur->pos <= RECS_PER_LEAF);
cur               263 fs/xfs/libxfs/xfs_iext_tree.c 		cur->pos--;
cur               264 fs/xfs/libxfs/xfs_iext_tree.c 		if (xfs_iext_valid(ifp, cur))
cur               266 fs/xfs/libxfs/xfs_iext_tree.c 	} while (cur->pos > 0);
cur               268 fs/xfs/libxfs/xfs_iext_tree.c 	if (ifp->if_height > 1 && cur->leaf->prev) {
cur               269 fs/xfs/libxfs/xfs_iext_tree.c 		cur->leaf = cur->leaf->prev;
cur               270 fs/xfs/libxfs/xfs_iext_tree.c 		cur->pos = RECS_PER_LEAF;
cur               541 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_cursor	*cur,
cur               544 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_leaf	*leaf = cur->leaf;
cur               551 fs/xfs/libxfs/xfs_iext_tree.c 	if (cur->pos == RECS_PER_LEAF) {
cur               552 fs/xfs/libxfs/xfs_iext_tree.c 		cur->leaf = new;
cur               553 fs/xfs/libxfs/xfs_iext_tree.c 		cur->pos = 0;
cur               563 fs/xfs/libxfs/xfs_iext_tree.c 	if (cur->pos >= nr_keep) {
cur               564 fs/xfs/libxfs/xfs_iext_tree.c 		cur->leaf = new;
cur               565 fs/xfs/libxfs/xfs_iext_tree.c 		cur->pos -= nr_keep;
cur               582 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_cursor	*cur)
cur               590 fs/xfs/libxfs/xfs_iext_tree.c 	cur->leaf = ifp->if_u1.if_root;
cur               591 fs/xfs/libxfs/xfs_iext_tree.c 	cur->pos = 0;
cur               597 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_cursor	*cur)
cur               609 fs/xfs/libxfs/xfs_iext_tree.c 	cur->leaf = new;
cur               627 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_cursor	*cur,
cur               639 fs/xfs/libxfs/xfs_iext_tree.c 		xfs_iext_alloc_root(ifp, cur);
cur               641 fs/xfs/libxfs/xfs_iext_tree.c 		xfs_iext_realloc_root(ifp, cur);
cur               643 fs/xfs/libxfs/xfs_iext_tree.c 	nr_entries = xfs_iext_leaf_nr_entries(ifp, cur->leaf, cur->pos);
cur               645 fs/xfs/libxfs/xfs_iext_tree.c 	ASSERT(cur->pos >= nr_entries ||
cur               646 fs/xfs/libxfs/xfs_iext_tree.c 	       xfs_iext_rec_cmp(cur_rec(cur), irec->br_startoff) != 0);
cur               649 fs/xfs/libxfs/xfs_iext_tree.c 		new = xfs_iext_split_leaf(cur, &nr_entries);
cur               655 fs/xfs/libxfs/xfs_iext_tree.c 	if (cur->leaf != new && cur->pos == 0 && nr_entries > 0) {
cur               656 fs/xfs/libxfs/xfs_iext_tree.c 		xfs_iext_update_node(ifp, xfs_iext_leaf_key(cur->leaf, 0),
cur               657 fs/xfs/libxfs/xfs_iext_tree.c 				offset, 1, cur->leaf);
cur               660 fs/xfs/libxfs/xfs_iext_tree.c 	for (i = nr_entries; i > cur->pos; i--)
cur               661 fs/xfs/libxfs/xfs_iext_tree.c 		cur->leaf->recs[i] = cur->leaf->recs[i - 1];
cur               662 fs/xfs/libxfs/xfs_iext_tree.c 	xfs_iext_set(cur_rec(cur), irec);
cur               665 fs/xfs/libxfs/xfs_iext_tree.c 	trace_xfs_iext_insert(ip, cur, state, _RET_IP_);
cur               790 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_cursor	*cur,
cur               810 fs/xfs/libxfs/xfs_iext_tree.c 			if (cur->leaf == leaf) {
cur               811 fs/xfs/libxfs/xfs_iext_tree.c 				cur->leaf = leaf->prev;
cur               812 fs/xfs/libxfs/xfs_iext_tree.c 				cur->pos += nr_prev;
cur               832 fs/xfs/libxfs/xfs_iext_tree.c 			if (cur->leaf == leaf->next) {
cur               833 fs/xfs/libxfs/xfs_iext_tree.c 				cur->leaf = leaf;
cur               834 fs/xfs/libxfs/xfs_iext_tree.c 				cur->pos += nr_entries;
cur               864 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_cursor	*cur,
cur               868 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_leaf	*leaf = cur->leaf;
cur               872 fs/xfs/libxfs/xfs_iext_tree.c 	trace_xfs_iext_remove(ip, cur, state, _RET_IP_);
cur               876 fs/xfs/libxfs/xfs_iext_tree.c 	ASSERT(xfs_iext_valid(ifp, cur));
cur               880 fs/xfs/libxfs/xfs_iext_tree.c 	nr_entries = xfs_iext_leaf_nr_entries(ifp, leaf, cur->pos) - 1;
cur               881 fs/xfs/libxfs/xfs_iext_tree.c 	for (i = cur->pos; i < nr_entries; i++)
cur               886 fs/xfs/libxfs/xfs_iext_tree.c 	if (cur->pos == 0 && nr_entries > 0) {
cur               890 fs/xfs/libxfs/xfs_iext_tree.c 	} else if (cur->pos == nr_entries) {
cur               892 fs/xfs/libxfs/xfs_iext_tree.c 			cur->leaf = leaf->next;
cur               894 fs/xfs/libxfs/xfs_iext_tree.c 			cur->leaf = NULL;
cur               895 fs/xfs/libxfs/xfs_iext_tree.c 		cur->pos = 0;
cur               902 fs/xfs/libxfs/xfs_iext_tree.c 		xfs_iext_rebalance_leaf(ifp, cur, leaf, offset, nr_entries);
cur               923 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_cursor	*cur,
cur               928 fs/xfs/libxfs/xfs_iext_tree.c 	cur->leaf = xfs_iext_find_level(ifp, offset, 1);
cur               929 fs/xfs/libxfs/xfs_iext_tree.c 	if (!cur->leaf) {
cur               930 fs/xfs/libxfs/xfs_iext_tree.c 		cur->pos = 0;
cur               934 fs/xfs/libxfs/xfs_iext_tree.c 	for (cur->pos = 0; cur->pos < xfs_iext_max_recs(ifp); cur->pos++) {
cur               935 fs/xfs/libxfs/xfs_iext_tree.c 		struct xfs_iext_rec *rec = cur_rec(cur);
cur               944 fs/xfs/libxfs/xfs_iext_tree.c 	if (ifp->if_height == 1 || !cur->leaf->next)
cur               946 fs/xfs/libxfs/xfs_iext_tree.c 	cur->leaf = cur->leaf->next;
cur               947 fs/xfs/libxfs/xfs_iext_tree.c 	cur->pos = 0;
cur               948 fs/xfs/libxfs/xfs_iext_tree.c 	if (!xfs_iext_valid(ifp, cur))
cur               951 fs/xfs/libxfs/xfs_iext_tree.c 	xfs_iext_get(gotp, cur_rec(cur));
cur               964 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_cursor	*cur,
cur               968 fs/xfs/libxfs/xfs_iext_tree.c 	if (xfs_iext_lookup_extent(ip, ifp, *end - 1, cur, gotp) &&
cur               971 fs/xfs/libxfs/xfs_iext_tree.c 	if (!xfs_iext_prev_extent(ifp, cur, gotp))
cur               981 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_cursor	*cur,
cur               988 fs/xfs/libxfs/xfs_iext_tree.c 	if (cur->pos == 0) {
cur               991 fs/xfs/libxfs/xfs_iext_tree.c 		xfs_iext_get(&old, cur_rec(cur));
cur               994 fs/xfs/libxfs/xfs_iext_tree.c 					new->br_startoff, 1, cur->leaf);
cur               998 fs/xfs/libxfs/xfs_iext_tree.c 	trace_xfs_bmap_pre_update(ip, cur, state, _RET_IP_);
cur               999 fs/xfs/libxfs/xfs_iext_tree.c 	xfs_iext_set(cur_rec(cur), new);
cur              1000 fs/xfs/libxfs/xfs_iext_tree.c 	trace_xfs_bmap_post_update(ip, cur, state, _RET_IP_);
cur              1010 fs/xfs/libxfs/xfs_iext_tree.c 	struct xfs_iext_cursor	*cur,
cur              1013 fs/xfs/libxfs/xfs_iext_tree.c 	if (!xfs_iext_valid(ifp, cur))
cur              1015 fs/xfs/libxfs/xfs_iext_tree.c 	xfs_iext_get(gotp, cur_rec(cur));
cur               104 fs/xfs/libxfs/xfs_inode_fork.h void		xfs_iext_insert(struct xfs_inode *, struct xfs_iext_cursor *cur,
cur               112 fs/xfs/libxfs/xfs_inode_fork.h 			struct xfs_iext_cursor *cur,
cur               116 fs/xfs/libxfs/xfs_inode_fork.h 			struct xfs_iext_cursor *cur,
cur               119 fs/xfs/libxfs/xfs_inode_fork.h 			struct xfs_iext_cursor *cur,
cur               122 fs/xfs/libxfs/xfs_inode_fork.h 			struct xfs_iext_cursor *cur,
cur               131 fs/xfs/libxfs/xfs_inode_fork.h 		struct xfs_iext_cursor *cur, struct xfs_bmbt_irec *gotp)
cur               133 fs/xfs/libxfs/xfs_inode_fork.h 	xfs_iext_next(ifp, cur);
cur               134 fs/xfs/libxfs/xfs_inode_fork.h 	return xfs_iext_get_extent(ifp, cur, gotp);
cur               138 fs/xfs/libxfs/xfs_inode_fork.h 		struct xfs_iext_cursor *cur, struct xfs_bmbt_irec *gotp)
cur               140 fs/xfs/libxfs/xfs_inode_fork.h 	xfs_iext_prev(ifp, cur);
cur               141 fs/xfs/libxfs/xfs_inode_fork.h 	return xfs_iext_get_extent(ifp, cur, gotp);
cur               148 fs/xfs/libxfs/xfs_inode_fork.h 		struct xfs_iext_cursor *cur, struct xfs_bmbt_irec *gotp)
cur               150 fs/xfs/libxfs/xfs_inode_fork.h 	struct xfs_iext_cursor ncur = *cur;
cur               160 fs/xfs/libxfs/xfs_inode_fork.h 		struct xfs_iext_cursor *cur, struct xfs_bmbt_irec *gotp)
cur               162 fs/xfs/libxfs/xfs_inode_fork.h 	struct xfs_iext_cursor ncur = *cur;
cur                45 fs/xfs/libxfs/xfs_refcount.c 	struct xfs_btree_cur	*cur,
cur                49 fs/xfs/libxfs/xfs_refcount.c 	trace_xfs_refcount_lookup(cur->bc_mp, cur->bc_private.a.agno, bno,
cur                51 fs/xfs/libxfs/xfs_refcount.c 	cur->bc_rec.rc.rc_startblock = bno;
cur                52 fs/xfs/libxfs/xfs_refcount.c 	cur->bc_rec.rc.rc_blockcount = 0;
cur                53 fs/xfs/libxfs/xfs_refcount.c 	return xfs_btree_lookup(cur, XFS_LOOKUP_LE, stat);
cur                62 fs/xfs/libxfs/xfs_refcount.c 	struct xfs_btree_cur	*cur,
cur                66 fs/xfs/libxfs/xfs_refcount.c 	trace_xfs_refcount_lookup(cur->bc_mp, cur->bc_private.a.agno, bno,
cur                68 fs/xfs/libxfs/xfs_refcount.c 	cur->bc_rec.rc.rc_startblock = bno;
cur                69 fs/xfs/libxfs/xfs_refcount.c 	cur->bc_rec.rc.rc_blockcount = 0;
cur                70 fs/xfs/libxfs/xfs_refcount.c 	return xfs_btree_lookup(cur, XFS_LOOKUP_GE, stat);
cur                79 fs/xfs/libxfs/xfs_refcount.c 	struct xfs_btree_cur	*cur,
cur                83 fs/xfs/libxfs/xfs_refcount.c 	trace_xfs_refcount_lookup(cur->bc_mp, cur->bc_private.a.agno, bno,
cur                85 fs/xfs/libxfs/xfs_refcount.c 	cur->bc_rec.rc.rc_startblock = bno;
cur                86 fs/xfs/libxfs/xfs_refcount.c 	cur->bc_rec.rc.rc_blockcount = 0;
cur                87 fs/xfs/libxfs/xfs_refcount.c 	return xfs_btree_lookup(cur, XFS_LOOKUP_EQ, stat);
cur               106 fs/xfs/libxfs/xfs_refcount.c 	struct xfs_btree_cur		*cur,
cur               110 fs/xfs/libxfs/xfs_refcount.c 	struct xfs_mount		*mp = cur->bc_mp;
cur               111 fs/xfs/libxfs/xfs_refcount.c 	xfs_agnumber_t			agno = cur->bc_private.a.agno;
cur               116 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_btree_get_rec(cur, &rec, stat);
cur               122 fs/xfs/libxfs/xfs_refcount.c 	agno = cur->bc_private.a.agno;
cur               147 fs/xfs/libxfs/xfs_refcount.c 	trace_xfs_refcount_get(cur->bc_mp, cur->bc_private.a.agno, irec);
cur               166 fs/xfs/libxfs/xfs_refcount.c 	struct xfs_btree_cur		*cur,
cur               172 fs/xfs/libxfs/xfs_refcount.c 	trace_xfs_refcount_update(cur->bc_mp, cur->bc_private.a.agno, irec);
cur               176 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_btree_update(cur, &rec);
cur               178 fs/xfs/libxfs/xfs_refcount.c 		trace_xfs_refcount_update_error(cur->bc_mp,
cur               179 fs/xfs/libxfs/xfs_refcount.c 				cur->bc_private.a.agno, error, _RET_IP_);
cur               190 fs/xfs/libxfs/xfs_refcount.c 	struct xfs_btree_cur		*cur,
cur               196 fs/xfs/libxfs/xfs_refcount.c 	trace_xfs_refcount_insert(cur->bc_mp, cur->bc_private.a.agno, irec);
cur               197 fs/xfs/libxfs/xfs_refcount.c 	cur->bc_rec.rc.rc_startblock = irec->rc_startblock;
cur               198 fs/xfs/libxfs/xfs_refcount.c 	cur->bc_rec.rc.rc_blockcount = irec->rc_blockcount;
cur               199 fs/xfs/libxfs/xfs_refcount.c 	cur->bc_rec.rc.rc_refcount = irec->rc_refcount;
cur               200 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_btree_insert(cur, i);
cur               203 fs/xfs/libxfs/xfs_refcount.c 	XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, *i == 1, out_error);
cur               207 fs/xfs/libxfs/xfs_refcount.c 		trace_xfs_refcount_insert_error(cur->bc_mp,
cur               208 fs/xfs/libxfs/xfs_refcount.c 				cur->bc_private.a.agno, error, _RET_IP_);
cur               220 fs/xfs/libxfs/xfs_refcount.c 	struct xfs_btree_cur	*cur,
cur               227 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_get_rec(cur, &irec, &found_rec);
cur               230 fs/xfs/libxfs/xfs_refcount.c 	XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, out_error);
cur               231 fs/xfs/libxfs/xfs_refcount.c 	trace_xfs_refcount_delete(cur->bc_mp, cur->bc_private.a.agno, &irec);
cur               232 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_btree_delete(cur, i);
cur               233 fs/xfs/libxfs/xfs_refcount.c 	XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, *i == 1, out_error);
cur               236 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_lookup_ge(cur, irec.rc_startblock, &found_rec);
cur               239 fs/xfs/libxfs/xfs_refcount.c 		trace_xfs_refcount_delete_error(cur->bc_mp,
cur               240 fs/xfs/libxfs/xfs_refcount.c 				cur->bc_private.a.agno, error, _RET_IP_);
cur               334 fs/xfs/libxfs/xfs_refcount.c 	struct xfs_btree_cur		*cur,
cur               343 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_lookup_le(cur, agbno, &found_rec);
cur               349 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_get_rec(cur, &rcext, &found_rec);
cur               352 fs/xfs/libxfs/xfs_refcount.c 	XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, out_error);
cur               357 fs/xfs/libxfs/xfs_refcount.c 	trace_xfs_refcount_split_extent(cur->bc_mp, cur->bc_private.a.agno,
cur               364 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_update(cur, &tmp);
cur               371 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_insert(cur, &tmp, &found_rec);
cur               374 fs/xfs/libxfs/xfs_refcount.c 	XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, out_error);
cur               378 fs/xfs/libxfs/xfs_refcount.c 	trace_xfs_refcount_split_extent_error(cur->bc_mp,
cur               379 fs/xfs/libxfs/xfs_refcount.c 			cur->bc_private.a.agno, error, _RET_IP_);
cur               388 fs/xfs/libxfs/xfs_refcount.c 	struct xfs_btree_cur		*cur,
cur               398 fs/xfs/libxfs/xfs_refcount.c 	trace_xfs_refcount_merge_center_extents(cur->bc_mp,
cur               399 fs/xfs/libxfs/xfs_refcount.c 			cur->bc_private.a.agno, left, center, right);
cur               409 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_lookup_ge(cur, center->rc_startblock,
cur               413 fs/xfs/libxfs/xfs_refcount.c 	XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, out_error);
cur               415 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_delete(cur, &found_rec);
cur               418 fs/xfs/libxfs/xfs_refcount.c 	XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, out_error);
cur               421 fs/xfs/libxfs/xfs_refcount.c 		error = xfs_refcount_delete(cur, &found_rec);
cur               424 fs/xfs/libxfs/xfs_refcount.c 		XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1,
cur               429 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_lookup_le(cur, left->rc_startblock,
cur               433 fs/xfs/libxfs/xfs_refcount.c 	XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, out_error);
cur               436 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_update(cur, left);
cur               444 fs/xfs/libxfs/xfs_refcount.c 	trace_xfs_refcount_merge_center_extents_error(cur->bc_mp,
cur               445 fs/xfs/libxfs/xfs_refcount.c 			cur->bc_private.a.agno, error, _RET_IP_);
cur               454 fs/xfs/libxfs/xfs_refcount.c 	struct xfs_btree_cur		*cur,
cur               463 fs/xfs/libxfs/xfs_refcount.c 	trace_xfs_refcount_merge_left_extent(cur->bc_mp,
cur               464 fs/xfs/libxfs/xfs_refcount.c 			cur->bc_private.a.agno, left, cleft);
cur               468 fs/xfs/libxfs/xfs_refcount.c 		error = xfs_refcount_lookup_le(cur, cleft->rc_startblock,
cur               472 fs/xfs/libxfs/xfs_refcount.c 		XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1,
cur               475 fs/xfs/libxfs/xfs_refcount.c 		error = xfs_refcount_delete(cur, &found_rec);
cur               478 fs/xfs/libxfs/xfs_refcount.c 		XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1,
cur               483 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_lookup_le(cur, left->rc_startblock,
cur               487 fs/xfs/libxfs/xfs_refcount.c 	XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, out_error);
cur               490 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_update(cur, left);
cur               499 fs/xfs/libxfs/xfs_refcount.c 	trace_xfs_refcount_merge_left_extent_error(cur->bc_mp,
cur               500 fs/xfs/libxfs/xfs_refcount.c 			cur->bc_private.a.agno, error, _RET_IP_);
cur               509 fs/xfs/libxfs/xfs_refcount.c 	struct xfs_btree_cur		*cur,
cur               517 fs/xfs/libxfs/xfs_refcount.c 	trace_xfs_refcount_merge_right_extent(cur->bc_mp,
cur               518 fs/xfs/libxfs/xfs_refcount.c 			cur->bc_private.a.agno, cright, right);
cur               525 fs/xfs/libxfs/xfs_refcount.c 		error = xfs_refcount_lookup_le(cur, cright->rc_startblock,
cur               529 fs/xfs/libxfs/xfs_refcount.c 		XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1,
cur               532 fs/xfs/libxfs/xfs_refcount.c 		error = xfs_refcount_delete(cur, &found_rec);
cur               535 fs/xfs/libxfs/xfs_refcount.c 		XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1,
cur               540 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_lookup_le(cur, right->rc_startblock,
cur               544 fs/xfs/libxfs/xfs_refcount.c 	XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, out_error);
cur               548 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_update(cur, right);
cur               556 fs/xfs/libxfs/xfs_refcount.c 	trace_xfs_refcount_merge_right_extent_error(cur->bc_mp,
cur               557 fs/xfs/libxfs/xfs_refcount.c 			cur->bc_private.a.agno, error, _RET_IP_);
cur               569 fs/xfs/libxfs/xfs_refcount.c 	struct xfs_btree_cur		*cur,
cur               581 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_lookup_le(cur, agbno - 1, &found_rec);
cur               587 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_get_rec(cur, &tmp, &found_rec);
cur               590 fs/xfs/libxfs/xfs_refcount.c 	XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, out_error);
cur               601 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_btree_increment(cur, 0, &found_rec);
cur               605 fs/xfs/libxfs/xfs_refcount.c 		error = xfs_refcount_get_rec(cur, &tmp, &found_rec);
cur               608 fs/xfs/libxfs/xfs_refcount.c 		XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1,
cur               637 fs/xfs/libxfs/xfs_refcount.c 	trace_xfs_refcount_find_left_extent(cur->bc_mp, cur->bc_private.a.agno,
cur               642 fs/xfs/libxfs/xfs_refcount.c 	trace_xfs_refcount_find_left_extent_error(cur->bc_mp,
cur               643 fs/xfs/libxfs/xfs_refcount.c 			cur->bc_private.a.agno, error, _RET_IP_);
cur               653 fs/xfs/libxfs/xfs_refcount.c 	struct xfs_btree_cur		*cur,
cur               665 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_lookup_ge(cur, agbno + aglen, &found_rec);
cur               671 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_get_rec(cur, &tmp, &found_rec);
cur               674 fs/xfs/libxfs/xfs_refcount.c 	XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, out_error);
cur               685 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_btree_decrement(cur, 0, &found_rec);
cur               689 fs/xfs/libxfs/xfs_refcount.c 		error = xfs_refcount_get_rec(cur, &tmp, &found_rec);
cur               692 fs/xfs/libxfs/xfs_refcount.c 		XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1,
cur               721 fs/xfs/libxfs/xfs_refcount.c 	trace_xfs_refcount_find_right_extent(cur->bc_mp, cur->bc_private.a.agno,
cur               726 fs/xfs/libxfs/xfs_refcount.c 	trace_xfs_refcount_find_right_extent_error(cur->bc_mp,
cur               727 fs/xfs/libxfs/xfs_refcount.c 			cur->bc_private.a.agno, error, _RET_IP_);
cur               744 fs/xfs/libxfs/xfs_refcount.c 	struct xfs_btree_cur	*cur,
cur               763 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_find_left_extents(cur, &left, &cleft, *agbno,
cur               767 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_find_right_extents(cur, &right, &cright, *agbno,
cur               788 fs/xfs/libxfs/xfs_refcount.c 		return xfs_refcount_merge_center_extents(cur, &left, &cleft,
cur               798 fs/xfs/libxfs/xfs_refcount.c 		error = xfs_refcount_merge_left_extent(cur, &left, &cleft,
cur               817 fs/xfs/libxfs/xfs_refcount.c 		return xfs_refcount_merge_right_extent(cur, &right, &cright,
cur               832 fs/xfs/libxfs/xfs_refcount.c 	struct xfs_btree_cur		*cur)
cur               836 fs/xfs/libxfs/xfs_refcount.c 	overhead = cur->bc_private.a.priv.refc.shape_changes *
cur               837 fs/xfs/libxfs/xfs_refcount.c 			xfs_allocfree_log_count(cur->bc_mp, 1);
cur               838 fs/xfs/libxfs/xfs_refcount.c 	overhead *= cur->bc_mp->m_sb.sb_blocksize;
cur               844 fs/xfs/libxfs/xfs_refcount.c 	if (cur->bc_private.a.priv.refc.nr_ops > 2 &&
cur               845 fs/xfs/libxfs/xfs_refcount.c 	    XFS_TEST_ERROR(false, cur->bc_mp,
cur               849 fs/xfs/libxfs/xfs_refcount.c 	if (cur->bc_private.a.priv.refc.nr_ops == 0)
cur               851 fs/xfs/libxfs/xfs_refcount.c 	else if (overhead > cur->bc_tp->t_log_res)
cur               853 fs/xfs/libxfs/xfs_refcount.c 	return  cur->bc_tp->t_log_res - overhead >
cur               854 fs/xfs/libxfs/xfs_refcount.c 		cur->bc_private.a.priv.refc.nr_ops * XFS_REFCOUNT_ITEM_OVERHEAD;
cur               865 fs/xfs/libxfs/xfs_refcount.c 	struct xfs_btree_cur	*cur,
cur               880 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_lookup_ge(cur, *agbno, &found_rec);
cur               884 fs/xfs/libxfs/xfs_refcount.c 	while (*aglen > 0 && xfs_refcount_still_have_space(cur)) {
cur               885 fs/xfs/libxfs/xfs_refcount.c 		error = xfs_refcount_get_rec(cur, &ext, &found_rec);
cur               889 fs/xfs/libxfs/xfs_refcount.c 			ext.rc_startblock = cur->bc_mp->m_sb.sb_agblocks;
cur               904 fs/xfs/libxfs/xfs_refcount.c 			trace_xfs_refcount_modify_extent(cur->bc_mp,
cur               905 fs/xfs/libxfs/xfs_refcount.c 					cur->bc_private.a.agno, &tmp);
cur               912 fs/xfs/libxfs/xfs_refcount.c 				error = xfs_refcount_insert(cur, &tmp,
cur               916 fs/xfs/libxfs/xfs_refcount.c 				XFS_WANT_CORRUPTED_GOTO(cur->bc_mp,
cur               918 fs/xfs/libxfs/xfs_refcount.c 				cur->bc_private.a.priv.refc.nr_ops++;
cur               920 fs/xfs/libxfs/xfs_refcount.c 				fsbno = XFS_AGB_TO_FSB(cur->bc_mp,
cur               921 fs/xfs/libxfs/xfs_refcount.c 						cur->bc_private.a.agno,
cur               923 fs/xfs/libxfs/xfs_refcount.c 				xfs_bmap_add_free(cur->bc_tp, fsbno,
cur               930 fs/xfs/libxfs/xfs_refcount.c 			error = xfs_refcount_lookup_ge(cur, *agbno,
cur               937 fs/xfs/libxfs/xfs_refcount.c 		if (*aglen == 0 || !xfs_refcount_still_have_space(cur))
cur               947 fs/xfs/libxfs/xfs_refcount.c 		trace_xfs_refcount_modify_extent(cur->bc_mp,
cur               948 fs/xfs/libxfs/xfs_refcount.c 				cur->bc_private.a.agno, &ext);
cur               950 fs/xfs/libxfs/xfs_refcount.c 			error = xfs_refcount_update(cur, &ext);
cur               953 fs/xfs/libxfs/xfs_refcount.c 			cur->bc_private.a.priv.refc.nr_ops++;
cur               955 fs/xfs/libxfs/xfs_refcount.c 			error = xfs_refcount_delete(cur, &found_rec);
cur               958 fs/xfs/libxfs/xfs_refcount.c 			XFS_WANT_CORRUPTED_GOTO(cur->bc_mp,
cur               960 fs/xfs/libxfs/xfs_refcount.c 			cur->bc_private.a.priv.refc.nr_ops++;
cur               963 fs/xfs/libxfs/xfs_refcount.c 			fsbno = XFS_AGB_TO_FSB(cur->bc_mp,
cur               964 fs/xfs/libxfs/xfs_refcount.c 					cur->bc_private.a.agno,
cur               966 fs/xfs/libxfs/xfs_refcount.c 			xfs_bmap_add_free(cur->bc_tp, fsbno, ext.rc_blockcount,
cur               971 fs/xfs/libxfs/xfs_refcount.c 		error = xfs_btree_increment(cur, 0, &found_rec);
cur               982 fs/xfs/libxfs/xfs_refcount.c 	trace_xfs_refcount_modify_extent_error(cur->bc_mp,
cur               983 fs/xfs/libxfs/xfs_refcount.c 			cur->bc_private.a.agno, error, _RET_IP_);
cur               990 fs/xfs/libxfs/xfs_refcount.c 	struct xfs_btree_cur	*cur,
cur              1005 fs/xfs/libxfs/xfs_refcount.c 		trace_xfs_refcount_increase(cur->bc_mp, cur->bc_private.a.agno,
cur              1008 fs/xfs/libxfs/xfs_refcount.c 		trace_xfs_refcount_decrease(cur->bc_mp, cur->bc_private.a.agno,
cur              1014 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_split_extent(cur, agbno, &shape_changed);
cur              1020 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_split_extent(cur, agbno + aglen, &shape_changed);
cur              1029 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_merge_extents(cur, new_agbno, new_aglen, adj,
cur              1036 fs/xfs/libxfs/xfs_refcount.c 		cur->bc_private.a.priv.refc.shape_changes++;
cur              1039 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_adjust_extents(cur, new_agbno, new_aglen,
cur              1047 fs/xfs/libxfs/xfs_refcount.c 	trace_xfs_refcount_adjust_error(cur->bc_mp, cur->bc_private.a.agno,
cur              1241 fs/xfs/libxfs/xfs_refcount.c 	struct xfs_btree_cur		*cur,
cur              1253 fs/xfs/libxfs/xfs_refcount.c 	trace_xfs_refcount_find_shared(cur->bc_mp, cur->bc_private.a.agno,
cur              1261 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_lookup_le(cur, agbno, &have);
cur              1266 fs/xfs/libxfs/xfs_refcount.c 		error = xfs_btree_increment(cur, 0, &have);
cur              1272 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_get_rec(cur, &tmp, &i);
cur              1275 fs/xfs/libxfs/xfs_refcount.c 	XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, i == 1, out_error);
cur              1279 fs/xfs/libxfs/xfs_refcount.c 		error = xfs_btree_increment(cur, 0, &have);
cur              1284 fs/xfs/libxfs/xfs_refcount.c 		error = xfs_refcount_get_rec(cur, &tmp, &i);
cur              1287 fs/xfs/libxfs/xfs_refcount.c 		XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, i == 1, out_error);
cur              1307 fs/xfs/libxfs/xfs_refcount.c 		error = xfs_btree_increment(cur, 0, &have);
cur              1312 fs/xfs/libxfs/xfs_refcount.c 		error = xfs_refcount_get_rec(cur, &tmp, &i);
cur              1315 fs/xfs/libxfs/xfs_refcount.c 		XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, i == 1, out_error);
cur              1323 fs/xfs/libxfs/xfs_refcount.c 	trace_xfs_refcount_find_shared_result(cur->bc_mp,
cur              1324 fs/xfs/libxfs/xfs_refcount.c 			cur->bc_private.a.agno, *fbno, *flen);
cur              1328 fs/xfs/libxfs/xfs_refcount.c 		trace_xfs_refcount_find_shared_error(cur->bc_mp,
cur              1329 fs/xfs/libxfs/xfs_refcount.c 				cur->bc_private.a.agno, error, _RET_IP_);
cur              1387 fs/xfs/libxfs/xfs_refcount.c 	struct xfs_btree_cur	*cur,
cur              1400 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_lookup_ge(cur, agbno, &found_rec);
cur              1403 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_get_rec(cur, &ext, &found_rec);
cur              1407 fs/xfs/libxfs/xfs_refcount.c 		ext.rc_startblock = cur->bc_mp->m_sb.sb_agblocks +
cur              1416 fs/xfs/libxfs/xfs_refcount.c 		XFS_WANT_CORRUPTED_GOTO(cur->bc_mp,
cur              1422 fs/xfs/libxfs/xfs_refcount.c 		trace_xfs_refcount_modify_extent(cur->bc_mp,
cur              1423 fs/xfs/libxfs/xfs_refcount.c 				cur->bc_private.a.agno, &tmp);
cur              1425 fs/xfs/libxfs/xfs_refcount.c 		error = xfs_refcount_insert(cur, &tmp,
cur              1429 fs/xfs/libxfs/xfs_refcount.c 		XFS_WANT_CORRUPTED_GOTO(cur->bc_mp,
cur              1434 fs/xfs/libxfs/xfs_refcount.c 		XFS_WANT_CORRUPTED_GOTO(cur->bc_mp,
cur              1436 fs/xfs/libxfs/xfs_refcount.c 		XFS_WANT_CORRUPTED_GOTO(cur->bc_mp,
cur              1438 fs/xfs/libxfs/xfs_refcount.c 		XFS_WANT_CORRUPTED_GOTO(cur->bc_mp,
cur              1442 fs/xfs/libxfs/xfs_refcount.c 		trace_xfs_refcount_modify_extent(cur->bc_mp,
cur              1443 fs/xfs/libxfs/xfs_refcount.c 				cur->bc_private.a.agno, &ext);
cur              1444 fs/xfs/libxfs/xfs_refcount.c 		error = xfs_refcount_delete(cur, &found_rec);
cur              1447 fs/xfs/libxfs/xfs_refcount.c 		XFS_WANT_CORRUPTED_GOTO(cur->bc_mp,
cur              1456 fs/xfs/libxfs/xfs_refcount.c 	trace_xfs_refcount_modify_extent_error(cur->bc_mp,
cur              1457 fs/xfs/libxfs/xfs_refcount.c 			cur->bc_private.a.agno, error, _RET_IP_);
cur              1466 fs/xfs/libxfs/xfs_refcount.c 	struct xfs_btree_cur	*cur,
cur              1479 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_split_extent(cur, agbno, &shape_changed);
cur              1483 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_split_extent(cur, agbno + aglen, &shape_changed);
cur              1490 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_merge_extents(cur, &agbno, &aglen, adj,
cur              1496 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_refcount_adjust_cow_extents(cur, agbno, aglen, adj);
cur              1503 fs/xfs/libxfs/xfs_refcount.c 	trace_xfs_refcount_adjust_cow_error(cur->bc_mp, cur->bc_private.a.agno,
cur              1587 fs/xfs/libxfs/xfs_refcount.c 	struct xfs_btree_cur 		*cur,
cur              1611 fs/xfs/libxfs/xfs_refcount.c 	struct xfs_btree_cur		*cur;
cur              1647 fs/xfs/libxfs/xfs_refcount.c 	cur = xfs_refcountbt_init_cursor(mp, tp, agbp, agno);
cur              1654 fs/xfs/libxfs/xfs_refcount.c 	error = xfs_btree_query_range(cur, &low, &high,
cur              1656 fs/xfs/libxfs/xfs_refcount.c 	xfs_btree_del_cursor(cur, error);
cur              1703 fs/xfs/libxfs/xfs_refcount.c 	struct xfs_btree_cur	*cur,
cur              1716 fs/xfs/libxfs/xfs_refcount.c 	return xfs_btree_has_record(cur, &low, &high, exists);
cur                 9 fs/xfs/libxfs/xfs_refcount.h extern int xfs_refcount_lookup_le(struct xfs_btree_cur *cur,
cur                11 fs/xfs/libxfs/xfs_refcount.h extern int xfs_refcount_lookup_ge(struct xfs_btree_cur *cur,
cur                13 fs/xfs/libxfs/xfs_refcount.h extern int xfs_refcount_lookup_eq(struct xfs_btree_cur *cur,
cur                15 fs/xfs/libxfs/xfs_refcount.h extern int xfs_refcount_get_rec(struct xfs_btree_cur *cur,
cur                44 fs/xfs/libxfs/xfs_refcount.h extern int xfs_refcount_find_shared(struct xfs_btree_cur *cur,
cur                71 fs/xfs/libxfs/xfs_refcount.h extern int xfs_refcount_has_record(struct xfs_btree_cur *cur,
cur                76 fs/xfs/libxfs/xfs_refcount.h extern int xfs_refcount_insert(struct xfs_btree_cur *cur,
cur                25 fs/xfs/libxfs/xfs_refcount_btree.c 	struct xfs_btree_cur	*cur)
cur                27 fs/xfs/libxfs/xfs_refcount_btree.c 	return xfs_refcountbt_init_cursor(cur->bc_mp, cur->bc_tp,
cur                28 fs/xfs/libxfs/xfs_refcount_btree.c 			cur->bc_private.a.agbp, cur->bc_private.a.agno);
cur                33 fs/xfs/libxfs/xfs_refcount_btree.c 	struct xfs_btree_cur	*cur,
cur                37 fs/xfs/libxfs/xfs_refcount_btree.c 	struct xfs_buf		*agbp = cur->bc_private.a.agbp;
cur                40 fs/xfs/libxfs/xfs_refcount_btree.c 	struct xfs_perag	*pag = xfs_perag_get(cur->bc_mp, seqno);
cur                49 fs/xfs/libxfs/xfs_refcount_btree.c 	xfs_alloc_log_agf(cur->bc_tp, agbp,
cur                55 fs/xfs/libxfs/xfs_refcount_btree.c 	struct xfs_btree_cur	*cur,
cur                60 fs/xfs/libxfs/xfs_refcount_btree.c 	struct xfs_buf		*agbp = cur->bc_private.a.agbp;
cur                66 fs/xfs/libxfs/xfs_refcount_btree.c 	args.tp = cur->bc_tp;
cur                67 fs/xfs/libxfs/xfs_refcount_btree.c 	args.mp = cur->bc_mp;
cur                69 fs/xfs/libxfs/xfs_refcount_btree.c 	args.fsbno = XFS_AGB_TO_FSB(cur->bc_mp, cur->bc_private.a.agno,
cur                78 fs/xfs/libxfs/xfs_refcount_btree.c 	trace_xfs_refcountbt_alloc_block(cur->bc_mp, cur->bc_private.a.agno,
cur                84 fs/xfs/libxfs/xfs_refcount_btree.c 	ASSERT(args.agno == cur->bc_private.a.agno);
cur                89 fs/xfs/libxfs/xfs_refcount_btree.c 	xfs_alloc_log_agf(cur->bc_tp, agbp, XFS_AGF_REFCOUNT_BLOCKS);
cur               100 fs/xfs/libxfs/xfs_refcount_btree.c 	struct xfs_btree_cur	*cur,
cur               103 fs/xfs/libxfs/xfs_refcount_btree.c 	struct xfs_mount	*mp = cur->bc_mp;
cur               104 fs/xfs/libxfs/xfs_refcount_btree.c 	struct xfs_buf		*agbp = cur->bc_private.a.agbp;
cur               109 fs/xfs/libxfs/xfs_refcount_btree.c 	trace_xfs_refcountbt_free_block(cur->bc_mp, cur->bc_private.a.agno,
cur               110 fs/xfs/libxfs/xfs_refcount_btree.c 			XFS_FSB_TO_AGBNO(cur->bc_mp, fsbno), 1);
cur               112 fs/xfs/libxfs/xfs_refcount_btree.c 	xfs_alloc_log_agf(cur->bc_tp, agbp, XFS_AGF_REFCOUNT_BLOCKS);
cur               113 fs/xfs/libxfs/xfs_refcount_btree.c 	error = xfs_free_extent(cur->bc_tp, fsbno, 1, &XFS_RMAP_OINFO_REFC,
cur               123 fs/xfs/libxfs/xfs_refcount_btree.c 	struct xfs_btree_cur	*cur,
cur               126 fs/xfs/libxfs/xfs_refcount_btree.c 	return cur->bc_mp->m_refc_mnr[level != 0];
cur               131 fs/xfs/libxfs/xfs_refcount_btree.c 	struct xfs_btree_cur	*cur,
cur               134 fs/xfs/libxfs/xfs_refcount_btree.c 	return cur->bc_mp->m_refc_mxr[level != 0];
cur               159 fs/xfs/libxfs/xfs_refcount_btree.c 	struct xfs_btree_cur	*cur,
cur               162 fs/xfs/libxfs/xfs_refcount_btree.c 	rec->refc.rc_startblock = cpu_to_be32(cur->bc_rec.rc.rc_startblock);
cur               163 fs/xfs/libxfs/xfs_refcount_btree.c 	rec->refc.rc_blockcount = cpu_to_be32(cur->bc_rec.rc.rc_blockcount);
cur               164 fs/xfs/libxfs/xfs_refcount_btree.c 	rec->refc.rc_refcount = cpu_to_be32(cur->bc_rec.rc.rc_refcount);
cur               169 fs/xfs/libxfs/xfs_refcount_btree.c 	struct xfs_btree_cur	*cur,
cur               172 fs/xfs/libxfs/xfs_refcount_btree.c 	struct xfs_agf		*agf = XFS_BUF_TO_AGF(cur->bc_private.a.agbp);
cur               174 fs/xfs/libxfs/xfs_refcount_btree.c 	ASSERT(cur->bc_private.a.agno == be32_to_cpu(agf->agf_seqno));
cur               181 fs/xfs/libxfs/xfs_refcount_btree.c 	struct xfs_btree_cur	*cur,
cur               184 fs/xfs/libxfs/xfs_refcount_btree.c 	struct xfs_refcount_irec	*rec = &cur->bc_rec.rc;
cur               192 fs/xfs/libxfs/xfs_refcount_btree.c 	struct xfs_btree_cur	*cur,
cur               273 fs/xfs/libxfs/xfs_refcount_btree.c 	struct xfs_btree_cur	*cur,
cur               283 fs/xfs/libxfs/xfs_refcount_btree.c 	struct xfs_btree_cur	*cur,
cur               324 fs/xfs/libxfs/xfs_refcount_btree.c 	struct xfs_btree_cur	*cur;
cur               328 fs/xfs/libxfs/xfs_refcount_btree.c 	cur = kmem_zone_zalloc(xfs_btree_cur_zone, KM_NOFS);
cur               330 fs/xfs/libxfs/xfs_refcount_btree.c 	cur->bc_tp = tp;
cur               331 fs/xfs/libxfs/xfs_refcount_btree.c 	cur->bc_mp = mp;
cur               332 fs/xfs/libxfs/xfs_refcount_btree.c 	cur->bc_btnum = XFS_BTNUM_REFC;
cur               333 fs/xfs/libxfs/xfs_refcount_btree.c 	cur->bc_blocklog = mp->m_sb.sb_blocklog;
cur               334 fs/xfs/libxfs/xfs_refcount_btree.c 	cur->bc_ops = &xfs_refcountbt_ops;
cur               335 fs/xfs/libxfs/xfs_refcount_btree.c 	cur->bc_statoff = XFS_STATS_CALC_INDEX(xs_refcbt_2);
cur               337 fs/xfs/libxfs/xfs_refcount_btree.c 	cur->bc_nlevels = be32_to_cpu(agf->agf_refcount_level);
cur               339 fs/xfs/libxfs/xfs_refcount_btree.c 	cur->bc_private.a.agbp = agbp;
cur               340 fs/xfs/libxfs/xfs_refcount_btree.c 	cur->bc_private.a.agno = agno;
cur               341 fs/xfs/libxfs/xfs_refcount_btree.c 	cur->bc_flags |= XFS_BTREE_CRC_BLOCKS;
cur               343 fs/xfs/libxfs/xfs_refcount_btree.c 	cur->bc_private.a.priv.refc.nr_ops = 0;
cur               344 fs/xfs/libxfs/xfs_refcount_btree.c 	cur->bc_private.a.priv.refc.shape_changes = 0;
cur               346 fs/xfs/libxfs/xfs_refcount_btree.c 	return cur;
cur                31 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_btree_cur	*cur,
cur                39 fs/xfs/libxfs/xfs_rmap.c 	cur->bc_rec.r.rm_startblock = bno;
cur                40 fs/xfs/libxfs/xfs_rmap.c 	cur->bc_rec.r.rm_blockcount = len;
cur                41 fs/xfs/libxfs/xfs_rmap.c 	cur->bc_rec.r.rm_owner = owner;
cur                42 fs/xfs/libxfs/xfs_rmap.c 	cur->bc_rec.r.rm_offset = offset;
cur                43 fs/xfs/libxfs/xfs_rmap.c 	cur->bc_rec.r.rm_flags = flags;
cur                44 fs/xfs/libxfs/xfs_rmap.c 	return xfs_btree_lookup(cur, XFS_LOOKUP_LE, stat);
cur                53 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_btree_cur	*cur,
cur                61 fs/xfs/libxfs/xfs_rmap.c 	cur->bc_rec.r.rm_startblock = bno;
cur                62 fs/xfs/libxfs/xfs_rmap.c 	cur->bc_rec.r.rm_blockcount = len;
cur                63 fs/xfs/libxfs/xfs_rmap.c 	cur->bc_rec.r.rm_owner = owner;
cur                64 fs/xfs/libxfs/xfs_rmap.c 	cur->bc_rec.r.rm_offset = offset;
cur                65 fs/xfs/libxfs/xfs_rmap.c 	cur->bc_rec.r.rm_flags = flags;
cur                66 fs/xfs/libxfs/xfs_rmap.c 	return xfs_btree_lookup(cur, XFS_LOOKUP_EQ, stat);
cur                76 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_btree_cur	*cur,
cur                82 fs/xfs/libxfs/xfs_rmap.c 	trace_xfs_rmap_update(cur->bc_mp, cur->bc_private.a.agno,
cur                91 fs/xfs/libxfs/xfs_rmap.c 	error = xfs_btree_update(cur, &rec);
cur                93 fs/xfs/libxfs/xfs_rmap.c 		trace_xfs_rmap_update_error(cur->bc_mp,
cur                94 fs/xfs/libxfs/xfs_rmap.c 				cur->bc_private.a.agno, error, _RET_IP_);
cur               183 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_btree_cur	*cur,
cur               187 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_mount	*mp = cur->bc_mp;
cur               188 fs/xfs/libxfs/xfs_rmap.c 	xfs_agnumber_t		agno = cur->bc_private.a.agno;
cur               192 fs/xfs/libxfs/xfs_rmap.c 	error = xfs_btree_get_rec(cur, &rec, stat);
cur               244 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_btree_cur	*cur,
cur               250 fs/xfs/libxfs/xfs_rmap.c 	trace_xfs_rmap_find_left_neighbor_candidate(cur->bc_mp,
cur               251 fs/xfs/libxfs/xfs_rmap.c 			cur->bc_private.a.agno, rec->rm_startblock,
cur               274 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_btree_cur	*cur,
cur               302 fs/xfs/libxfs/xfs_rmap.c 	trace_xfs_rmap_find_left_neighbor_query(cur->bc_mp,
cur               303 fs/xfs/libxfs/xfs_rmap.c 			cur->bc_private.a.agno, bno, 0, owner, offset, flags);
cur               305 fs/xfs/libxfs/xfs_rmap.c 	error = xfs_rmap_query_range(cur, &info.high, &info.high,
cur               310 fs/xfs/libxfs/xfs_rmap.c 		trace_xfs_rmap_find_left_neighbor_result(cur->bc_mp,
cur               311 fs/xfs/libxfs/xfs_rmap.c 				cur->bc_private.a.agno, irec->rm_startblock,
cur               320 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_btree_cur	*cur,
cur               326 fs/xfs/libxfs/xfs_rmap.c 	trace_xfs_rmap_lookup_le_range_candidate(cur->bc_mp,
cur               327 fs/xfs/libxfs/xfs_rmap.c 			cur->bc_private.a.agno, rec->rm_startblock,
cur               352 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_btree_cur	*cur,
cur               375 fs/xfs/libxfs/xfs_rmap.c 	trace_xfs_rmap_lookup_le_range(cur->bc_mp,
cur               376 fs/xfs/libxfs/xfs_rmap.c 			cur->bc_private.a.agno, bno, 0, owner, offset, flags);
cur               377 fs/xfs/libxfs/xfs_rmap.c 	error = xfs_rmap_query_range(cur, &info.high, &info.high,
cur               382 fs/xfs/libxfs/xfs_rmap.c 		trace_xfs_rmap_lookup_le_range_result(cur->bc_mp,
cur               383 fs/xfs/libxfs/xfs_rmap.c 				cur->bc_private.a.agno, irec->rm_startblock,
cur               453 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_btree_cur		*cur,
cur               459 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_mount		*mp = cur->bc_mp;
cur               474 fs/xfs/libxfs/xfs_rmap.c 	trace_xfs_rmap_unmap(mp, cur->bc_private.a.agno, bno, len,
cur               482 fs/xfs/libxfs/xfs_rmap.c 	error = xfs_rmap_lookup_le(cur, bno, len, owner, offset, flags, &i);
cur               487 fs/xfs/libxfs/xfs_rmap.c 	error = xfs_rmap_get_rec(cur, &ltrec, &i);
cur               491 fs/xfs/libxfs/xfs_rmap.c 	trace_xfs_rmap_lookup_le_range_result(cur->bc_mp,
cur               492 fs/xfs/libxfs/xfs_rmap.c 			cur->bc_private.a.agno, ltrec.rm_startblock,
cur               521 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_btree_increment(cur, 0, &i);
cur               526 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_get_rec(cur, &rtrec, &i);
cur               547 fs/xfs/libxfs/xfs_rmap.c 		trace_xfs_rmap_delete(mp, cur->bc_private.a.agno,
cur               551 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_btree_delete(cur, &i);
cur               570 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &ltrec);
cur               585 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &ltrec);
cur               605 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &ltrec);
cur               609 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_btree_increment(cur, 0, &i);
cur               613 fs/xfs/libxfs/xfs_rmap.c 		cur->bc_rec.r.rm_startblock = bno + len;
cur               614 fs/xfs/libxfs/xfs_rmap.c 		cur->bc_rec.r.rm_blockcount = orig_len - len -
cur               616 fs/xfs/libxfs/xfs_rmap.c 		cur->bc_rec.r.rm_owner = ltrec.rm_owner;
cur               618 fs/xfs/libxfs/xfs_rmap.c 			cur->bc_rec.r.rm_offset = 0;
cur               620 fs/xfs/libxfs/xfs_rmap.c 			cur->bc_rec.r.rm_offset = offset + len;
cur               621 fs/xfs/libxfs/xfs_rmap.c 		cur->bc_rec.r.rm_flags = flags;
cur               622 fs/xfs/libxfs/xfs_rmap.c 		trace_xfs_rmap_insert(mp, cur->bc_private.a.agno,
cur               623 fs/xfs/libxfs/xfs_rmap.c 				cur->bc_rec.r.rm_startblock,
cur               624 fs/xfs/libxfs/xfs_rmap.c 				cur->bc_rec.r.rm_blockcount,
cur               625 fs/xfs/libxfs/xfs_rmap.c 				cur->bc_rec.r.rm_owner,
cur               626 fs/xfs/libxfs/xfs_rmap.c 				cur->bc_rec.r.rm_offset,
cur               627 fs/xfs/libxfs/xfs_rmap.c 				cur->bc_rec.r.rm_flags);
cur               628 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_btree_insert(cur, &i);
cur               634 fs/xfs/libxfs/xfs_rmap.c 	trace_xfs_rmap_unmap_done(mp, cur->bc_private.a.agno, bno, len,
cur               638 fs/xfs/libxfs/xfs_rmap.c 		trace_xfs_rmap_unmap_error(mp, cur->bc_private.a.agno,
cur               656 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_btree_cur		*cur;
cur               662 fs/xfs/libxfs/xfs_rmap.c 	cur = xfs_rmapbt_init_cursor(mp, tp, agbp, agno);
cur               664 fs/xfs/libxfs/xfs_rmap.c 	error = xfs_rmap_unmap(cur, bno, len, false, oinfo);
cur               666 fs/xfs/libxfs/xfs_rmap.c 	xfs_btree_del_cursor(cur, error);
cur               705 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_btree_cur		*cur,
cur               711 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_mount		*mp = cur->bc_mp;
cur               729 fs/xfs/libxfs/xfs_rmap.c 	trace_xfs_rmap_map(mp, cur->bc_private.a.agno, bno, len,
cur               738 fs/xfs/libxfs/xfs_rmap.c 	error = xfs_rmap_lookup_le(cur, bno, len, owner, offset, flags,
cur               743 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_get_rec(cur, &ltrec, &have_lt);
cur               747 fs/xfs/libxfs/xfs_rmap.c 		trace_xfs_rmap_lookup_le_range_result(cur->bc_mp,
cur               748 fs/xfs/libxfs/xfs_rmap.c 				cur->bc_private.a.agno, ltrec.rm_startblock,
cur               765 fs/xfs/libxfs/xfs_rmap.c 	error = xfs_btree_increment(cur, 0, &have_gt);
cur               769 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_get_rec(cur, &gtrec, &have_gt);
cur               775 fs/xfs/libxfs/xfs_rmap.c 		trace_xfs_rmap_find_right_neighbor_result(cur->bc_mp,
cur               776 fs/xfs/libxfs/xfs_rmap.c 			cur->bc_private.a.agno, gtrec.rm_startblock,
cur               815 fs/xfs/libxfs/xfs_rmap.c 			trace_xfs_rmap_delete(mp, cur->bc_private.a.agno,
cur               821 fs/xfs/libxfs/xfs_rmap.c 			error = xfs_btree_delete(cur, &i);
cur               828 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_btree_decrement(cur, 0, &have_gt);
cur               831 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &ltrec);
cur               850 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &gtrec);
cur               858 fs/xfs/libxfs/xfs_rmap.c 		cur->bc_rec.r.rm_startblock = bno;
cur               859 fs/xfs/libxfs/xfs_rmap.c 		cur->bc_rec.r.rm_blockcount = len;
cur               860 fs/xfs/libxfs/xfs_rmap.c 		cur->bc_rec.r.rm_owner = owner;
cur               861 fs/xfs/libxfs/xfs_rmap.c 		cur->bc_rec.r.rm_offset = offset;
cur               862 fs/xfs/libxfs/xfs_rmap.c 		cur->bc_rec.r.rm_flags = flags;
cur               863 fs/xfs/libxfs/xfs_rmap.c 		trace_xfs_rmap_insert(mp, cur->bc_private.a.agno, bno, len,
cur               865 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_btree_insert(cur, &i);
cur               871 fs/xfs/libxfs/xfs_rmap.c 	trace_xfs_rmap_map_done(mp, cur->bc_private.a.agno, bno, len,
cur               875 fs/xfs/libxfs/xfs_rmap.c 		trace_xfs_rmap_map_error(mp, cur->bc_private.a.agno,
cur               893 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_btree_cur		*cur;
cur               899 fs/xfs/libxfs/xfs_rmap.c 	cur = xfs_rmapbt_init_cursor(mp, tp, agbp, agno);
cur               900 fs/xfs/libxfs/xfs_rmap.c 	error = xfs_rmap_map(cur, bno, len, false, oinfo);
cur               902 fs/xfs/libxfs/xfs_rmap.c 	xfs_btree_del_cursor(cur, error);
cur               924 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_btree_cur		*cur,
cur               930 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_mount		*mp = cur->bc_mp;
cur               949 fs/xfs/libxfs/xfs_rmap.c 	trace_xfs_rmap_convert(mp, cur->bc_private.a.agno, bno, len,
cur               957 fs/xfs/libxfs/xfs_rmap.c 	error = xfs_rmap_lookup_le(cur, bno, len, owner, offset, oldext, &i);
cur               962 fs/xfs/libxfs/xfs_rmap.c 	error = xfs_rmap_get_rec(cur, &PREV, &i);
cur               966 fs/xfs/libxfs/xfs_rmap.c 	trace_xfs_rmap_lookup_le_range_result(cur->bc_mp,
cur               967 fs/xfs/libxfs/xfs_rmap.c 			cur->bc_private.a.agno, PREV.rm_startblock,
cur               990 fs/xfs/libxfs/xfs_rmap.c 	error = xfs_btree_decrement(cur, 0, &i);
cur               995 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_get_rec(cur, &LEFT, &i);
cur              1002 fs/xfs/libxfs/xfs_rmap.c 		trace_xfs_rmap_find_left_neighbor_result(cur->bc_mp,
cur              1003 fs/xfs/libxfs/xfs_rmap.c 				cur->bc_private.a.agno, LEFT.rm_startblock,
cur              1017 fs/xfs/libxfs/xfs_rmap.c 	error = xfs_btree_increment(cur, 0, &i);
cur              1021 fs/xfs/libxfs/xfs_rmap.c 	error = xfs_btree_increment(cur, 0, &i);
cur              1026 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_get_rec(cur, &RIGHT, &i);
cur              1032 fs/xfs/libxfs/xfs_rmap.c 		trace_xfs_rmap_find_right_neighbor_result(cur->bc_mp,
cur              1033 fs/xfs/libxfs/xfs_rmap.c 				cur->bc_private.a.agno, RIGHT.rm_startblock,
cur              1051 fs/xfs/libxfs/xfs_rmap.c 	trace_xfs_rmap_convert_state(mp, cur->bc_private.a.agno, state,
cur              1055 fs/xfs/libxfs/xfs_rmap.c 	error = xfs_rmap_lookup_le(cur, bno, len, owner, offset, oldext, &i);
cur              1071 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_btree_increment(cur, 0, &i);
cur              1075 fs/xfs/libxfs/xfs_rmap.c 		trace_xfs_rmap_delete(mp, cur->bc_private.a.agno,
cur              1079 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_btree_delete(cur, &i);
cur              1083 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_btree_decrement(cur, 0, &i);
cur              1087 fs/xfs/libxfs/xfs_rmap.c 		trace_xfs_rmap_delete(mp, cur->bc_private.a.agno,
cur              1091 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_btree_delete(cur, &i);
cur              1095 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_btree_decrement(cur, 0, &i);
cur              1101 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &NEW);
cur              1111 fs/xfs/libxfs/xfs_rmap.c 		trace_xfs_rmap_delete(mp, cur->bc_private.a.agno,
cur              1115 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_btree_delete(cur, &i);
cur              1119 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_btree_decrement(cur, 0, &i);
cur              1125 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &NEW);
cur              1135 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_btree_increment(cur, 0, &i);
cur              1139 fs/xfs/libxfs/xfs_rmap.c 		trace_xfs_rmap_delete(mp, cur->bc_private.a.agno,
cur              1143 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_btree_delete(cur, &i);
cur              1147 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_btree_decrement(cur, 0, &i);
cur              1154 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &NEW);
cur              1167 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &NEW);
cur              1181 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &NEW);
cur              1184 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_btree_decrement(cur, 0, &i);
cur              1189 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &NEW);
cur              1203 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &NEW);
cur              1211 fs/xfs/libxfs/xfs_rmap.c 		cur->bc_rec.r = NEW;
cur              1212 fs/xfs/libxfs/xfs_rmap.c 		trace_xfs_rmap_insert(mp, cur->bc_private.a.agno, bno,
cur              1214 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_btree_insert(cur, &i);
cur              1227 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &NEW);
cur              1230 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_btree_increment(cur, 0, &i);
cur              1237 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &NEW);
cur              1249 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &NEW);
cur              1252 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_lookup_eq(cur, bno, len, owner, offset,
cur              1262 fs/xfs/libxfs/xfs_rmap.c 		cur->bc_rec.r = NEW;
cur              1263 fs/xfs/libxfs/xfs_rmap.c 		trace_xfs_rmap_insert(mp, cur->bc_private.a.agno, bno,
cur              1265 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_btree_insert(cur, &i);
cur              1284 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &NEW);
cur              1290 fs/xfs/libxfs/xfs_rmap.c 		cur->bc_rec.r = NEW;
cur              1291 fs/xfs/libxfs/xfs_rmap.c 		trace_xfs_rmap_insert(mp, cur->bc_private.a.agno,
cur              1295 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_btree_insert(cur, &i);
cur              1304 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_lookup_eq(cur, bno, len, owner, offset,
cur              1310 fs/xfs/libxfs/xfs_rmap.c 		cur->bc_rec.r.rm_flags &= ~XFS_RMAP_UNWRITTEN;
cur              1311 fs/xfs/libxfs/xfs_rmap.c 		cur->bc_rec.r.rm_flags |= newext;
cur              1312 fs/xfs/libxfs/xfs_rmap.c 		trace_xfs_rmap_insert(mp, cur->bc_private.a.agno, bno, len,
cur              1314 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_btree_insert(cur, &i);
cur              1333 fs/xfs/libxfs/xfs_rmap.c 	trace_xfs_rmap_convert_done(mp, cur->bc_private.a.agno, bno, len,
cur              1337 fs/xfs/libxfs/xfs_rmap.c 		trace_xfs_rmap_convert_error(cur->bc_mp,
cur              1338 fs/xfs/libxfs/xfs_rmap.c 				cur->bc_private.a.agno, error, _RET_IP_);
cur              1349 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_btree_cur		*cur,
cur              1355 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_mount		*mp = cur->bc_mp;
cur              1374 fs/xfs/libxfs/xfs_rmap.c 	trace_xfs_rmap_convert(mp, cur->bc_private.a.agno, bno, len,
cur              1382 fs/xfs/libxfs/xfs_rmap.c 	error = xfs_rmap_lookup_le_range(cur, bno, owner, offset, flags,
cur              1403 fs/xfs/libxfs/xfs_rmap.c 	error = xfs_rmap_find_left_neighbor(cur, bno, owner, offset, newext,
cur              1417 fs/xfs/libxfs/xfs_rmap.c 	error = xfs_rmap_lookup_eq(cur, bno + len, len, owner, offset + len,
cur              1423 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_get_rec(cur, &RIGHT, &i);
cur              1429 fs/xfs/libxfs/xfs_rmap.c 		trace_xfs_rmap_find_right_neighbor_result(cur->bc_mp,
cur              1430 fs/xfs/libxfs/xfs_rmap.c 				cur->bc_private.a.agno, RIGHT.rm_startblock,
cur              1446 fs/xfs/libxfs/xfs_rmap.c 	trace_xfs_rmap_convert_state(mp, cur->bc_private.a.agno, state,
cur              1459 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_delete(cur, RIGHT.rm_startblock,
cur              1464 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_delete(cur, PREV.rm_startblock,
cur              1470 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_lookup_eq(cur, NEW.rm_startblock,
cur              1477 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &NEW);
cur              1487 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_delete(cur, PREV.rm_startblock,
cur              1493 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_lookup_eq(cur, NEW.rm_startblock,
cur              1500 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &NEW);
cur              1510 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_delete(cur, RIGHT.rm_startblock,
cur              1516 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_lookup_eq(cur, NEW.rm_startblock,
cur              1524 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &NEW);
cur              1536 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_lookup_eq(cur, NEW.rm_startblock,
cur              1543 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &NEW);
cur              1554 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_delete(cur, NEW.rm_startblock,
cur              1562 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_insert(cur, NEW.rm_startblock,
cur              1568 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_lookup_eq(cur, NEW.rm_startblock,
cur              1575 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &NEW);
cur              1586 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_delete(cur, NEW.rm_startblock,
cur              1594 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_insert(cur, NEW.rm_startblock,
cur              1599 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_insert(cur, bno, len, owner, offset, newext);
cur              1610 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_lookup_eq(cur, NEW.rm_startblock,
cur              1617 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &NEW);
cur              1621 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_delete(cur, NEW.rm_startblock,
cur              1629 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_insert(cur, NEW.rm_startblock,
cur              1642 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_lookup_eq(cur, NEW.rm_startblock,
cur              1649 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &NEW);
cur              1652 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_insert(cur, bno, len, owner, offset, newext);
cur              1670 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_insert(cur, NEW.rm_startblock,
cur              1677 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_lookup_eq(cur, NEW.rm_startblock,
cur              1684 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &NEW);
cur              1693 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_insert(cur, NEW.rm_startblock,
cur              1713 fs/xfs/libxfs/xfs_rmap.c 	trace_xfs_rmap_convert_done(mp, cur->bc_private.a.agno, bno, len,
cur              1717 fs/xfs/libxfs/xfs_rmap.c 		trace_xfs_rmap_convert_error(cur->bc_mp,
cur              1718 fs/xfs/libxfs/xfs_rmap.c 				cur->bc_private.a.agno, error, _RET_IP_);
cur              1738 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_btree_cur		*cur,
cur              1744 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_mount		*mp = cur->bc_mp;
cur              1756 fs/xfs/libxfs/xfs_rmap.c 	trace_xfs_rmap_unmap(mp, cur->bc_private.a.agno, bno, len,
cur              1764 fs/xfs/libxfs/xfs_rmap.c 	error = xfs_rmap_lookup_le_range(cur, bno, owner, offset, flags,
cur              1790 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_delete(cur, ltrec.rm_startblock,
cur              1808 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_delete(cur, ltrec.rm_startblock,
cur              1818 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_insert(cur, ltrec.rm_startblock,
cur              1834 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_lookup_eq(cur, ltrec.rm_startblock,
cur              1841 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &ltrec);
cur              1860 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_lookup_eq(cur, ltrec.rm_startblock,
cur              1867 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &ltrec);
cur              1872 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_insert(cur, bno + len,
cur              1880 fs/xfs/libxfs/xfs_rmap.c 	trace_xfs_rmap_unmap_done(mp, cur->bc_private.a.agno, bno, len,
cur              1884 fs/xfs/libxfs/xfs_rmap.c 		trace_xfs_rmap_unmap_error(cur->bc_mp,
cur              1885 fs/xfs/libxfs/xfs_rmap.c 				cur->bc_private.a.agno, error, _RET_IP_);
cur              1900 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_btree_cur		*cur,
cur              1906 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_mount		*mp = cur->bc_mp;
cur              1920 fs/xfs/libxfs/xfs_rmap.c 	trace_xfs_rmap_map(mp, cur->bc_private.a.agno, bno, len,
cur              1924 fs/xfs/libxfs/xfs_rmap.c 	error = xfs_rmap_find_left_neighbor(cur, bno, owner, offset, flags,
cur              1933 fs/xfs/libxfs/xfs_rmap.c 	error = xfs_rmap_lookup_eq(cur, bno + len, len, owner, offset + len,
cur              1938 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_get_rec(cur, &gtrec, &have_gt);
cur              1942 fs/xfs/libxfs/xfs_rmap.c 		trace_xfs_rmap_find_right_neighbor_result(cur->bc_mp,
cur              1943 fs/xfs/libxfs/xfs_rmap.c 			cur->bc_private.a.agno, gtrec.rm_startblock,
cur              1977 fs/xfs/libxfs/xfs_rmap.c 			error = xfs_rmap_delete(cur, gtrec.rm_startblock,
cur              1985 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_lookup_eq(cur, ltrec.rm_startblock,
cur              1992 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_update(cur, &ltrec);
cur              2008 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_delete(cur, gtrec.rm_startblock,
cur              2018 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_insert(cur, gtrec.rm_startblock,
cur              2028 fs/xfs/libxfs/xfs_rmap.c 		error = xfs_rmap_insert(cur, bno, len, owner, offset, flags);
cur              2033 fs/xfs/libxfs/xfs_rmap.c 	trace_xfs_rmap_map_done(mp, cur->bc_private.a.agno, bno, len,
cur              2037 fs/xfs/libxfs/xfs_rmap.c 		trace_xfs_rmap_map_error(cur->bc_mp,
cur              2038 fs/xfs/libxfs/xfs_rmap.c 				cur->bc_private.a.agno, error, _RET_IP_);
cur              2045 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_btree_cur	*cur,
cur              2059 fs/xfs/libxfs/xfs_rmap.c 		return xfs_rmap_map(cur, rmap->rm_startblock,
cur              2064 fs/xfs/libxfs/xfs_rmap.c 	return xfs_rmap_map_shared(cur, rmap->rm_startblock,
cur              2078 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_btree_cur	*cur,
cur              2089 fs/xfs/libxfs/xfs_rmap.c 	return query->fn(cur, &irec, query->priv);
cur              2095 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_btree_cur			*cur,
cur              2109 fs/xfs/libxfs/xfs_rmap.c 	return xfs_btree_query_range(cur, &low_brec, &high_brec,
cur              2116 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_btree_cur			*cur,
cur              2124 fs/xfs/libxfs/xfs_rmap.c 	return xfs_btree_query_all(cur, xfs_rmap_query_range_helper, &query);
cur              2428 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_btree_cur	*cur,
cur              2441 fs/xfs/libxfs/xfs_rmap.c 	return xfs_btree_has_record(cur, &low, &high, exists);
cur              2453 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_btree_cur		*cur,
cur              2470 fs/xfs/libxfs/xfs_rmap.c 	error = xfs_rmap_lookup_le(cur, bno, len, owner, offset, flags,
cur              2479 fs/xfs/libxfs/xfs_rmap.c 	error = xfs_rmap_get_rec(cur, &irec, &has_record);
cur              2502 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_btree_cur		*cur,
cur              2521 fs/xfs/libxfs/xfs_rmap.c 	struct xfs_btree_cur		*cur,
cur              2539 fs/xfs/libxfs/xfs_rmap.c 	error = xfs_rmap_query_range(cur, &low, &high,
cur               122 fs/xfs/libxfs/xfs_rmap.h int xfs_rmap_lookup_le(struct xfs_btree_cur *cur, xfs_agblock_t bno,
cur               125 fs/xfs/libxfs/xfs_rmap.h int xfs_rmap_lookup_eq(struct xfs_btree_cur *cur, xfs_agblock_t bno,
cur               131 fs/xfs/libxfs/xfs_rmap.h int xfs_rmap_get_rec(struct xfs_btree_cur *cur, struct xfs_rmap_irec *irec,
cur               135 fs/xfs/libxfs/xfs_rmap.h 	struct xfs_btree_cur	*cur,
cur               139 fs/xfs/libxfs/xfs_rmap.h int xfs_rmap_query_range(struct xfs_btree_cur *cur,
cur               142 fs/xfs/libxfs/xfs_rmap.h int xfs_rmap_query_all(struct xfs_btree_cur *cur, xfs_rmap_query_range_fn fn,
cur               184 fs/xfs/libxfs/xfs_rmap.h int xfs_rmap_find_left_neighbor(struct xfs_btree_cur *cur, xfs_agblock_t bno,
cur               187 fs/xfs/libxfs/xfs_rmap.h int xfs_rmap_lookup_le_range(struct xfs_btree_cur *cur, xfs_agblock_t bno,
cur               195 fs/xfs/libxfs/xfs_rmap.h int xfs_rmap_has_record(struct xfs_btree_cur *cur, xfs_agblock_t bno,
cur               197 fs/xfs/libxfs/xfs_rmap.h int xfs_rmap_record_exists(struct xfs_btree_cur *cur, xfs_agblock_t bno,
cur               200 fs/xfs/libxfs/xfs_rmap.h int xfs_rmap_has_other_keys(struct xfs_btree_cur *cur, xfs_agblock_t bno,
cur               203 fs/xfs/libxfs/xfs_rmap.h int xfs_rmap_map_raw(struct xfs_btree_cur *cur, struct xfs_rmap_irec *rmap);
cur                51 fs/xfs/libxfs/xfs_rmap_btree.c 	struct xfs_btree_cur	*cur)
cur                53 fs/xfs/libxfs/xfs_rmap_btree.c 	return xfs_rmapbt_init_cursor(cur->bc_mp, cur->bc_tp,
cur                54 fs/xfs/libxfs/xfs_rmap_btree.c 			cur->bc_private.a.agbp, cur->bc_private.a.agno);
cur                59 fs/xfs/libxfs/xfs_rmap_btree.c 	struct xfs_btree_cur	*cur,
cur                63 fs/xfs/libxfs/xfs_rmap_btree.c 	struct xfs_buf		*agbp = cur->bc_private.a.agbp;
cur                66 fs/xfs/libxfs/xfs_rmap_btree.c 	int			btnum = cur->bc_btnum;
cur                67 fs/xfs/libxfs/xfs_rmap_btree.c 	struct xfs_perag	*pag = xfs_perag_get(cur->bc_mp, seqno);
cur                76 fs/xfs/libxfs/xfs_rmap_btree.c 	xfs_alloc_log_agf(cur->bc_tp, agbp, XFS_AGF_ROOTS | XFS_AGF_LEVELS);
cur                81 fs/xfs/libxfs/xfs_rmap_btree.c 	struct xfs_btree_cur	*cur,
cur                86 fs/xfs/libxfs/xfs_rmap_btree.c 	struct xfs_buf		*agbp = cur->bc_private.a.agbp;
cur                92 fs/xfs/libxfs/xfs_rmap_btree.c 	error = xfs_alloc_get_freelist(cur->bc_tp, cur->bc_private.a.agbp,
cur                97 fs/xfs/libxfs/xfs_rmap_btree.c 	trace_xfs_rmapbt_alloc_block(cur->bc_mp, cur->bc_private.a.agno,
cur               104 fs/xfs/libxfs/xfs_rmap_btree.c 	xfs_extent_busy_reuse(cur->bc_mp, cur->bc_private.a.agno, bno, 1,
cur               107 fs/xfs/libxfs/xfs_rmap_btree.c 	xfs_trans_agbtree_delta(cur->bc_tp, 1);
cur               110 fs/xfs/libxfs/xfs_rmap_btree.c 	xfs_alloc_log_agf(cur->bc_tp, agbp, XFS_AGF_RMAP_BLOCKS);
cur               112 fs/xfs/libxfs/xfs_rmap_btree.c 	xfs_ag_resv_rmapbt_alloc(cur->bc_mp, cur->bc_private.a.agno);
cur               120 fs/xfs/libxfs/xfs_rmap_btree.c 	struct xfs_btree_cur	*cur,
cur               123 fs/xfs/libxfs/xfs_rmap_btree.c 	struct xfs_buf		*agbp = cur->bc_private.a.agbp;
cur               128 fs/xfs/libxfs/xfs_rmap_btree.c 	bno = xfs_daddr_to_agbno(cur->bc_mp, XFS_BUF_ADDR(bp));
cur               129 fs/xfs/libxfs/xfs_rmap_btree.c 	trace_xfs_rmapbt_free_block(cur->bc_mp, cur->bc_private.a.agno,
cur               132 fs/xfs/libxfs/xfs_rmap_btree.c 	xfs_alloc_log_agf(cur->bc_tp, agbp, XFS_AGF_RMAP_BLOCKS);
cur               133 fs/xfs/libxfs/xfs_rmap_btree.c 	error = xfs_alloc_put_freelist(cur->bc_tp, agbp, NULL, bno, 1);
cur               137 fs/xfs/libxfs/xfs_rmap_btree.c 	xfs_extent_busy_insert(cur->bc_tp, be32_to_cpu(agf->agf_seqno), bno, 1,
cur               139 fs/xfs/libxfs/xfs_rmap_btree.c 	xfs_trans_agbtree_delta(cur->bc_tp, -1);
cur               141 fs/xfs/libxfs/xfs_rmap_btree.c 	xfs_ag_resv_rmapbt_free(cur->bc_mp, cur->bc_private.a.agno);
cur               148 fs/xfs/libxfs/xfs_rmap_btree.c 	struct xfs_btree_cur	*cur,
cur               151 fs/xfs/libxfs/xfs_rmap_btree.c 	return cur->bc_mp->m_rmap_mnr[level != 0];
cur               156 fs/xfs/libxfs/xfs_rmap_btree.c 	struct xfs_btree_cur	*cur,
cur               159 fs/xfs/libxfs/xfs_rmap_btree.c 	return cur->bc_mp->m_rmap_mxr[level != 0];
cur               203 fs/xfs/libxfs/xfs_rmap_btree.c 	struct xfs_btree_cur	*cur,
cur               206 fs/xfs/libxfs/xfs_rmap_btree.c 	rec->rmap.rm_startblock = cpu_to_be32(cur->bc_rec.r.rm_startblock);
cur               207 fs/xfs/libxfs/xfs_rmap_btree.c 	rec->rmap.rm_blockcount = cpu_to_be32(cur->bc_rec.r.rm_blockcount);
cur               208 fs/xfs/libxfs/xfs_rmap_btree.c 	rec->rmap.rm_owner = cpu_to_be64(cur->bc_rec.r.rm_owner);
cur               210 fs/xfs/libxfs/xfs_rmap_btree.c 			xfs_rmap_irec_offset_pack(&cur->bc_rec.r));
cur               215 fs/xfs/libxfs/xfs_rmap_btree.c 	struct xfs_btree_cur	*cur,
cur               218 fs/xfs/libxfs/xfs_rmap_btree.c 	struct xfs_agf		*agf = XFS_BUF_TO_AGF(cur->bc_private.a.agbp);
cur               220 fs/xfs/libxfs/xfs_rmap_btree.c 	ASSERT(cur->bc_private.a.agno == be32_to_cpu(agf->agf_seqno));
cur               222 fs/xfs/libxfs/xfs_rmap_btree.c 	ptr->s = agf->agf_roots[cur->bc_btnum];
cur               227 fs/xfs/libxfs/xfs_rmap_btree.c 	struct xfs_btree_cur	*cur,
cur               230 fs/xfs/libxfs/xfs_rmap_btree.c 	struct xfs_rmap_irec	*rec = &cur->bc_rec.r;
cur               257 fs/xfs/libxfs/xfs_rmap_btree.c 	struct xfs_btree_cur	*cur,
cur               372 fs/xfs/libxfs/xfs_rmap_btree.c 	struct xfs_btree_cur	*cur,
cur               402 fs/xfs/libxfs/xfs_rmap_btree.c 	struct xfs_btree_cur	*cur,
cur               462 fs/xfs/libxfs/xfs_rmap_btree.c 	struct xfs_btree_cur	*cur;
cur               464 fs/xfs/libxfs/xfs_rmap_btree.c 	cur = kmem_zone_zalloc(xfs_btree_cur_zone, KM_NOFS);
cur               465 fs/xfs/libxfs/xfs_rmap_btree.c 	cur->bc_tp = tp;
cur               466 fs/xfs/libxfs/xfs_rmap_btree.c 	cur->bc_mp = mp;
cur               468 fs/xfs/libxfs/xfs_rmap_btree.c 	cur->bc_btnum = XFS_BTNUM_RMAP;
cur               469 fs/xfs/libxfs/xfs_rmap_btree.c 	cur->bc_flags = XFS_BTREE_CRC_BLOCKS | XFS_BTREE_OVERLAPPING;
cur               470 fs/xfs/libxfs/xfs_rmap_btree.c 	cur->bc_blocklog = mp->m_sb.sb_blocklog;
cur               471 fs/xfs/libxfs/xfs_rmap_btree.c 	cur->bc_ops = &xfs_rmapbt_ops;
cur               472 fs/xfs/libxfs/xfs_rmap_btree.c 	cur->bc_nlevels = be32_to_cpu(agf->agf_levels[XFS_BTNUM_RMAP]);
cur               473 fs/xfs/libxfs/xfs_rmap_btree.c 	cur->bc_statoff = XFS_STATS_CALC_INDEX(xs_rmap_2);
cur               475 fs/xfs/libxfs/xfs_rmap_btree.c 	cur->bc_private.a.agbp = agbp;
cur               476 fs/xfs/libxfs/xfs_rmap_btree.c 	cur->bc_private.a.agno = agno;
cur               478 fs/xfs/libxfs/xfs_rmap_btree.c 	return cur;
cur               346 fs/xfs/scrub/agheader.c 	struct xfs_btree_cur		*cur,
cur                71 fs/xfs/scrub/agheader_repair.c 	struct xfs_btree_cur		*cur,
cur               240 fs/xfs/scrub/agheader_repair.c 	struct xfs_btree_cur	*cur = NULL;
cur               248 fs/xfs/scrub/agheader_repair.c 	cur = xfs_allocbt_init_cursor(mp, sc->tp, agf_bp, sc->sa.agno,
cur               250 fs/xfs/scrub/agheader_repair.c 	error = xfs_alloc_query_all(cur, xrep_agf_walk_allocbt, &raa);
cur               253 fs/xfs/scrub/agheader_repair.c 	error = xfs_btree_count_blocks(cur, &blocks);
cur               256 fs/xfs/scrub/agheader_repair.c 	xfs_btree_del_cursor(cur, error);
cur               262 fs/xfs/scrub/agheader_repair.c 	cur = xfs_allocbt_init_cursor(mp, sc->tp, agf_bp, sc->sa.agno,
cur               264 fs/xfs/scrub/agheader_repair.c 	error = xfs_btree_count_blocks(cur, &blocks);
cur               267 fs/xfs/scrub/agheader_repair.c 	xfs_btree_del_cursor(cur, error);
cur               271 fs/xfs/scrub/agheader_repair.c 	cur = xfs_rmapbt_init_cursor(mp, sc->tp, agf_bp, sc->sa.agno);
cur               272 fs/xfs/scrub/agheader_repair.c 	error = xfs_btree_count_blocks(cur, &blocks);
cur               275 fs/xfs/scrub/agheader_repair.c 	xfs_btree_del_cursor(cur, error);
cur               283 fs/xfs/scrub/agheader_repair.c 		cur = xfs_refcountbt_init_cursor(mp, sc->tp, agf_bp,
cur               285 fs/xfs/scrub/agheader_repair.c 		error = xfs_btree_count_blocks(cur, &blocks);
cur               288 fs/xfs/scrub/agheader_repair.c 		xfs_btree_del_cursor(cur, error);
cur               294 fs/xfs/scrub/agheader_repair.c 	xfs_btree_del_cursor(cur, error);
cur               443 fs/xfs/scrub/agheader_repair.c 	struct xfs_btree_cur	*cur,
cur               456 fs/xfs/scrub/agheader_repair.c 		fsb = XFS_AGB_TO_FSB(cur->bc_mp, cur->bc_private.a.agno,
cur               463 fs/xfs/scrub/agheader_repair.c 	return xfs_bitmap_set_btcur_path(&ra->agmetablocks, cur);
cur               484 fs/xfs/scrub/agheader_repair.c 	struct xfs_btree_cur	*cur;
cur               494 fs/xfs/scrub/agheader_repair.c 	cur = xfs_rmapbt_init_cursor(mp, sc->tp, agf_bp, sc->sa.agno);
cur               495 fs/xfs/scrub/agheader_repair.c 	error = xfs_rmap_query_all(cur, xrep_agfl_walk_rmap, &ra);
cur               498 fs/xfs/scrub/agheader_repair.c 	xfs_btree_del_cursor(cur, error);
cur               501 fs/xfs/scrub/agheader_repair.c 	cur = xfs_allocbt_init_cursor(mp, sc->tp, agf_bp, sc->sa.agno,
cur               503 fs/xfs/scrub/agheader_repair.c 	error = xfs_bitmap_set_btblocks(&ra.agmetablocks, cur);
cur               506 fs/xfs/scrub/agheader_repair.c 	xfs_btree_del_cursor(cur, error);
cur               509 fs/xfs/scrub/agheader_repair.c 	cur = xfs_allocbt_init_cursor(mp, sc->tp, agf_bp, sc->sa.agno,
cur               511 fs/xfs/scrub/agheader_repair.c 	error = xfs_bitmap_set_btblocks(&ra.agmetablocks, cur);
cur               515 fs/xfs/scrub/agheader_repair.c 	xfs_btree_del_cursor(cur, error);
cur               542 fs/xfs/scrub/agheader_repair.c 	xfs_btree_del_cursor(cur, error);
cur               813 fs/xfs/scrub/agheader_repair.c 	struct xfs_btree_cur	*cur;
cur               820 fs/xfs/scrub/agheader_repair.c 	cur = xfs_inobt_init_cursor(mp, sc->tp, agi_bp, sc->sa.agno,
cur               822 fs/xfs/scrub/agheader_repair.c 	error = xfs_ialloc_count_inodes(cur, &count, &freecount);
cur               825 fs/xfs/scrub/agheader_repair.c 	xfs_btree_del_cursor(cur, error);
cur               831 fs/xfs/scrub/agheader_repair.c 	xfs_btree_del_cursor(cur, error);
cur                96 fs/xfs/scrub/alloc.c 	struct xfs_mount	*mp = bs->cur->bc_mp;
cur                97 fs/xfs/scrub/alloc.c 	xfs_agnumber_t		agno = bs->cur->bc_private.a.agno;
cur               107 fs/xfs/scrub/alloc.c 		xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               120 fs/xfs/scrub/alloc.c 	struct xfs_btree_cur	*cur;
cur               122 fs/xfs/scrub/alloc.c 	cur = which == XFS_BTNUM_BNO ? sc->sa.bno_cur : sc->sa.cnt_cur;
cur               123 fs/xfs/scrub/alloc.c 	return xchk_btree(sc, cur, xchk_allocbt_rec, &XFS_RMAP_OINFO_AG, NULL);
cur               252 fs/xfs/scrub/bitmap.c 	struct xfs_btree_cur	*cur)
cur               259 fs/xfs/scrub/bitmap.c 	for (i = 0; i < cur->bc_nlevels && cur->bc_ptrs[i] == 1; i++) {
cur               260 fs/xfs/scrub/bitmap.c 		xfs_btree_get_block(cur, i, &bp);
cur               263 fs/xfs/scrub/bitmap.c 		fsb = XFS_DADDR_TO_FSB(cur->bc_mp, bp->b_bn);
cur               275 fs/xfs/scrub/bitmap.c 	struct xfs_btree_cur	*cur,
cur               283 fs/xfs/scrub/bitmap.c 	xfs_btree_get_block(cur, level, &bp);
cur               287 fs/xfs/scrub/bitmap.c 	fsbno = XFS_DADDR_TO_FSB(cur->bc_mp, bp->b_bn);
cur               295 fs/xfs/scrub/bitmap.c 	struct xfs_btree_cur	*cur)
cur               297 fs/xfs/scrub/bitmap.c 	return xfs_btree_visit_blocks(cur, xfs_bitmap_collect_btblock, bitmap);
cur                32 fs/xfs/scrub/bitmap.h 		struct xfs_btree_cur *cur);
cur                34 fs/xfs/scrub/bitmap.h 		struct xfs_btree_cur *cur);
cur               377 fs/xfs/scrub/bmap.c 	struct xfs_inode	*ip = bs->cur->bc_private.b.ip;
cur               388 fs/xfs/scrub/bmap.c 	if (xfs_sb_version_hascrc(&bs->cur->bc_mp->m_sb) &&
cur               389 fs/xfs/scrub/bmap.c 	    bs->cur->bc_ptrs[0] == 1) {
cur               390 fs/xfs/scrub/bmap.c 		for (i = 0; i < bs->cur->bc_nlevels - 1; i++) {
cur               391 fs/xfs/scrub/bmap.c 			block = xfs_btree_get_block(bs->cur, i, &bp);
cur               432 fs/xfs/scrub/bmap.c 	struct xfs_btree_cur	*cur;
cur               444 fs/xfs/scrub/bmap.c 	cur = xfs_bmbt_init_cursor(mp, sc->tp, ip, whichfork);
cur               446 fs/xfs/scrub/bmap.c 	error = xchk_btree(sc, cur, xchk_bmapbt_rec, &oinfo, info);
cur               447 fs/xfs/scrub/bmap.c 	xfs_btree_del_cursor(cur, error);
cur               461 fs/xfs/scrub/bmap.c 	struct xfs_btree_cur		*cur,
cur               504 fs/xfs/scrub/bmap.c 				cur->bc_private.a.agno, rec->rm_startblock))
cur               537 fs/xfs/scrub/bmap.c 	struct xfs_btree_cur		*cur;
cur               545 fs/xfs/scrub/bmap.c 	cur = xfs_rmapbt_init_cursor(sc->mp, sc->tp, agf, agno);
cur               546 fs/xfs/scrub/bmap.c 	if (!cur) {
cur               553 fs/xfs/scrub/bmap.c 	error = xfs_rmap_query_all(cur, xchk_bmap_check_rmap, &sbcri);
cur               557 fs/xfs/scrub/bmap.c 	xfs_btree_del_cursor(cur, error);
cur                27 fs/xfs/scrub/btree.c 	struct xfs_btree_cur	*cur,
cur                48 fs/xfs/scrub/btree.c 		if (cur->bc_flags & XFS_BTREE_ROOT_IN_INODE)
cur                49 fs/xfs/scrub/btree.c 			trace_xchk_ifork_btree_op_error(sc, cur, level,
cur                52 fs/xfs/scrub/btree.c 			trace_xchk_btree_op_error(sc, cur, level,
cur                62 fs/xfs/scrub/btree.c 	struct xfs_btree_cur	*cur,
cur                66 fs/xfs/scrub/btree.c 	return __xchk_btree_process_error(sc, cur, level, error,
cur                73 fs/xfs/scrub/btree.c 	struct xfs_btree_cur	*cur,
cur                77 fs/xfs/scrub/btree.c 	return __xchk_btree_process_error(sc, cur, level, error,
cur                85 fs/xfs/scrub/btree.c 	struct xfs_btree_cur	*cur,
cur                92 fs/xfs/scrub/btree.c 	if (cur->bc_flags & XFS_BTREE_ROOT_IN_INODE)
cur                93 fs/xfs/scrub/btree.c 		trace_xchk_ifork_btree_error(sc, cur, level,
cur                96 fs/xfs/scrub/btree.c 		trace_xchk_btree_error(sc, cur, level,
cur               103 fs/xfs/scrub/btree.c 	struct xfs_btree_cur	*cur,
cur               106 fs/xfs/scrub/btree.c 	__xchk_btree_set_corrupt(sc, cur, level, XFS_SCRUB_OFLAG_CORRUPT,
cur               113 fs/xfs/scrub/btree.c 	struct xfs_btree_cur	*cur,
cur               116 fs/xfs/scrub/btree.c 	__xchk_btree_set_corrupt(sc, cur, level, XFS_SCRUB_OFLAG_XCORRUPT,
cur               128 fs/xfs/scrub/btree.c 	struct xfs_btree_cur	*cur = bs->cur;
cur               137 fs/xfs/scrub/btree.c 	block = xfs_btree_get_block(cur, 0, &bp);
cur               138 fs/xfs/scrub/btree.c 	rec = xfs_btree_rec_addr(cur, cur->bc_ptrs[0], block);
cur               140 fs/xfs/scrub/btree.c 	trace_xchk_btree_rec(bs->sc, cur, 0);
cur               143 fs/xfs/scrub/btree.c 	if (!bs->firstrec && !cur->bc_ops->recs_inorder(cur, &bs->lastrec, rec))
cur               144 fs/xfs/scrub/btree.c 		xchk_btree_set_corrupt(bs->sc, cur, 0);
cur               146 fs/xfs/scrub/btree.c 	memcpy(&bs->lastrec, rec, cur->bc_ops->rec_len);
cur               148 fs/xfs/scrub/btree.c 	if (cur->bc_nlevels == 1)
cur               152 fs/xfs/scrub/btree.c 	cur->bc_ops->init_key_from_rec(&key, rec);
cur               153 fs/xfs/scrub/btree.c 	keyblock = xfs_btree_get_block(cur, 1, &bp);
cur               154 fs/xfs/scrub/btree.c 	keyp = xfs_btree_key_addr(cur, cur->bc_ptrs[1], keyblock);
cur               155 fs/xfs/scrub/btree.c 	if (cur->bc_ops->diff_two_keys(cur, &key, keyp) < 0)
cur               156 fs/xfs/scrub/btree.c 		xchk_btree_set_corrupt(bs->sc, cur, 1);
cur               158 fs/xfs/scrub/btree.c 	if (!(cur->bc_flags & XFS_BTREE_OVERLAPPING))
cur               162 fs/xfs/scrub/btree.c 	cur->bc_ops->init_high_key_from_rec(&hkey, rec);
cur               163 fs/xfs/scrub/btree.c 	keyp = xfs_btree_high_key_addr(cur, cur->bc_ptrs[1], keyblock);
cur               164 fs/xfs/scrub/btree.c 	if (cur->bc_ops->diff_two_keys(cur, keyp, &hkey) < 0)
cur               165 fs/xfs/scrub/btree.c 		xchk_btree_set_corrupt(bs->sc, cur, 1);
cur               177 fs/xfs/scrub/btree.c 	struct xfs_btree_cur	*cur = bs->cur;
cur               184 fs/xfs/scrub/btree.c 	block = xfs_btree_get_block(cur, level, &bp);
cur               185 fs/xfs/scrub/btree.c 	key = xfs_btree_key_addr(cur, cur->bc_ptrs[level], block);
cur               187 fs/xfs/scrub/btree.c 	trace_xchk_btree_key(bs->sc, cur, level);
cur               191 fs/xfs/scrub/btree.c 	    !cur->bc_ops->keys_inorder(cur, &bs->lastkey[level], key))
cur               192 fs/xfs/scrub/btree.c 		xchk_btree_set_corrupt(bs->sc, cur, level);
cur               194 fs/xfs/scrub/btree.c 	memcpy(&bs->lastkey[level], key, cur->bc_ops->key_len);
cur               196 fs/xfs/scrub/btree.c 	if (level + 1 >= cur->bc_nlevels)
cur               200 fs/xfs/scrub/btree.c 	keyblock = xfs_btree_get_block(cur, level + 1, &bp);
cur               201 fs/xfs/scrub/btree.c 	keyp = xfs_btree_key_addr(cur, cur->bc_ptrs[level + 1], keyblock);
cur               202 fs/xfs/scrub/btree.c 	if (cur->bc_ops->diff_two_keys(cur, key, keyp) < 0)
cur               203 fs/xfs/scrub/btree.c 		xchk_btree_set_corrupt(bs->sc, cur, level);
cur               205 fs/xfs/scrub/btree.c 	if (!(cur->bc_flags & XFS_BTREE_OVERLAPPING))
cur               209 fs/xfs/scrub/btree.c 	key = xfs_btree_high_key_addr(cur, cur->bc_ptrs[level], block);
cur               210 fs/xfs/scrub/btree.c 	keyp = xfs_btree_high_key_addr(cur, cur->bc_ptrs[level + 1], keyblock);
cur               211 fs/xfs/scrub/btree.c 	if (cur->bc_ops->diff_two_keys(cur, keyp, key) < 0)
cur               212 fs/xfs/scrub/btree.c 		xchk_btree_set_corrupt(bs->sc, cur, level);
cur               228 fs/xfs/scrub/btree.c 	if ((bs->cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) &&
cur               229 fs/xfs/scrub/btree.c 	    level == bs->cur->bc_nlevels)
cur               233 fs/xfs/scrub/btree.c 	if (bs->cur->bc_flags & XFS_BTREE_LONG_PTRS)
cur               234 fs/xfs/scrub/btree.c 		res = xfs_btree_check_lptr(bs->cur, be64_to_cpu(ptr->l), level);
cur               236 fs/xfs/scrub/btree.c 		res = xfs_btree_check_sptr(bs->cur, be32_to_cpu(ptr->s), level);
cur               238 fs/xfs/scrub/btree.c 		xchk_btree_set_corrupt(bs->sc, bs->cur, level);
cur               251 fs/xfs/scrub/btree.c 	struct xfs_btree_cur	*cur = bs->cur;
cur               259 fs/xfs/scrub/btree.c 	error = xfs_btree_dup_cursor(cur, &ncur);
cur               260 fs/xfs/scrub/btree.c 	if (!xchk_btree_process_error(bs->sc, cur, level + 1, &error) ||
cur               268 fs/xfs/scrub/btree.c 	if (xfs_btree_ptr_is_null(cur, sibling)) {
cur               274 fs/xfs/scrub/btree.c 			xchk_btree_set_corrupt(bs->sc, cur, level);
cur               284 fs/xfs/scrub/btree.c 	if (!xchk_btree_process_error(bs->sc, cur, level + 1, &error))
cur               287 fs/xfs/scrub/btree.c 		xchk_btree_set_corrupt(bs->sc, cur, level + 1);
cur               299 fs/xfs/scrub/btree.c 	if (xfs_btree_diff_two_ptrs(cur, pp, sibling))
cur               300 fs/xfs/scrub/btree.c 		xchk_btree_set_corrupt(bs->sc, cur, level);
cur               312 fs/xfs/scrub/btree.c 	struct xfs_btree_cur	*cur = bs->cur;
cur               318 fs/xfs/scrub/btree.c 	xfs_btree_get_sibling(cur, block, &leftsib, XFS_BB_LEFTSIB);
cur               319 fs/xfs/scrub/btree.c 	xfs_btree_get_sibling(cur, block, &rightsib, XFS_BB_RIGHTSIB);
cur               323 fs/xfs/scrub/btree.c 	if (level == cur->bc_nlevels - 1) {
cur               324 fs/xfs/scrub/btree.c 		if (!xfs_btree_ptr_is_null(cur, &leftsib) ||
cur               325 fs/xfs/scrub/btree.c 		    !xfs_btree_ptr_is_null(cur, &rightsib))
cur               326 fs/xfs/scrub/btree.c 			xchk_btree_set_corrupt(bs->sc, cur, level);
cur               367 fs/xfs/scrub/btree.c 	if (!bs->cur)
cur               370 fs/xfs/scrub/btree.c 	btnum = bs->cur->bc_btnum;
cur               371 fs/xfs/scrub/btree.c 	agno = xfs_daddr_to_agno(bs->cur->bc_mp, daddr);
cur               372 fs/xfs/scrub/btree.c 	agbno = xfs_daddr_to_agbno(bs->cur->bc_mp, daddr);
cur               374 fs/xfs/scrub/btree.c 	init_sa = bs->cur->bc_flags & XFS_BTREE_LONG_PTRS;
cur               377 fs/xfs/scrub/btree.c 		if (!xchk_btree_xref_process_error(bs->sc, bs->cur,
cur               389 fs/xfs/scrub/btree.c 		bs->cur = NULL;
cur               393 fs/xfs/scrub/btree.c 		bs->cur = NULL;
cur               408 fs/xfs/scrub/btree.c 	struct xfs_btree_cur	*cur = bs->cur;
cur               418 fs/xfs/scrub/btree.c 		if (!(cur->bc_flags & XFS_BTREE_ROOT_IN_INODE))
cur               419 fs/xfs/scrub/btree.c 			xchk_btree_set_corrupt(bs->sc, bs->cur, level);
cur               431 fs/xfs/scrub/btree.c 	if (cur->bc_btnum == XFS_BTNUM_BNO || cur->bc_btnum == XFS_BTNUM_RMAP) {
cur               461 fs/xfs/scrub/btree.c 	if (numrecs >= bs->cur->bc_ops->get_minrecs(bs->cur, level))
cur               474 fs/xfs/scrub/btree.c 	ok_level = bs->cur->bc_nlevels - 1;
cur               475 fs/xfs/scrub/btree.c 	if (bs->cur->bc_flags & XFS_BTREE_ROOT_IN_INODE)
cur               480 fs/xfs/scrub/btree.c 	xchk_btree_set_corrupt(bs->sc, bs->cur, level);
cur               501 fs/xfs/scrub/btree.c 	error = xfs_btree_lookup_get_block(bs->cur, level, pp, pblock);
cur               502 fs/xfs/scrub/btree.c 	if (!xchk_btree_process_error(bs->sc, bs->cur, level, &error) ||
cur               506 fs/xfs/scrub/btree.c 	xfs_btree_get_block(bs->cur, level, pbp);
cur               507 fs/xfs/scrub/btree.c 	if (bs->cur->bc_flags & XFS_BTREE_LONG_PTRS)
cur               508 fs/xfs/scrub/btree.c 		failed_at = __xfs_btree_check_lblock(bs->cur, *pblock,
cur               511 fs/xfs/scrub/btree.c 		failed_at = __xfs_btree_check_sblock(bs->cur, *pblock,
cur               514 fs/xfs/scrub/btree.c 		xchk_btree_set_corrupt(bs->sc, bs->cur, level);
cur               548 fs/xfs/scrub/btree.c 	struct xfs_btree_cur	*cur = bs->cur;
cur               555 fs/xfs/scrub/btree.c 	if (level >= cur->bc_nlevels - 1)
cur               559 fs/xfs/scrub/btree.c 	xfs_btree_get_keys(cur, block, &block_keys);
cur               562 fs/xfs/scrub/btree.c 	parent_block = xfs_btree_get_block(cur, level + 1, &bp);
cur               563 fs/xfs/scrub/btree.c 	parent_keys = xfs_btree_key_addr(cur, cur->bc_ptrs[level + 1],
cur               566 fs/xfs/scrub/btree.c 	if (cur->bc_ops->diff_two_keys(cur, &block_keys, parent_keys) != 0)
cur               567 fs/xfs/scrub/btree.c 		xchk_btree_set_corrupt(bs->sc, cur, 1);
cur               569 fs/xfs/scrub/btree.c 	if (!(cur->bc_flags & XFS_BTREE_OVERLAPPING))
cur               573 fs/xfs/scrub/btree.c 	high_bk = xfs_btree_high_key_from_key(cur, &block_keys);
cur               574 fs/xfs/scrub/btree.c 	high_pk = xfs_btree_high_key_addr(cur, cur->bc_ptrs[level + 1],
cur               577 fs/xfs/scrub/btree.c 	if (cur->bc_ops->diff_two_keys(cur, high_bk, high_pk) != 0)
cur               578 fs/xfs/scrub/btree.c 		xchk_btree_set_corrupt(bs->sc, cur, 1);
cur               589 fs/xfs/scrub/btree.c 	struct xfs_btree_cur		*cur,
cur               595 fs/xfs/scrub/btree.c 		.cur			= cur,
cur               619 fs/xfs/scrub/btree.c 	if (cur->bc_nlevels > XFS_BTREE_MAXLEVELS) {
cur               620 fs/xfs/scrub/btree.c 		xchk_btree_set_corrupt(sc, cur, 0);
cur               628 fs/xfs/scrub/btree.c 	level = cur->bc_nlevels - 1;
cur               629 fs/xfs/scrub/btree.c 	cur->bc_ops->init_ptr_from_cur(cur, &ptr);
cur               630 fs/xfs/scrub/btree.c 	if (!xchk_btree_ptr_ok(&bs, cur->bc_nlevels, &ptr))
cur               636 fs/xfs/scrub/btree.c 	cur->bc_ptrs[level] = 1;
cur               638 fs/xfs/scrub/btree.c 	while (level < cur->bc_nlevels) {
cur               639 fs/xfs/scrub/btree.c 		block = xfs_btree_get_block(cur, level, &bp);
cur               643 fs/xfs/scrub/btree.c 			if (cur->bc_ptrs[level] >
cur               646 fs/xfs/scrub/btree.c 				if (level < cur->bc_nlevels - 1)
cur               647 fs/xfs/scrub/btree.c 					cur->bc_ptrs[level + 1]++;
cur               656 fs/xfs/scrub/btree.c 			recp = xfs_btree_rec_addr(cur, cur->bc_ptrs[0], block);
cur               664 fs/xfs/scrub/btree.c 			cur->bc_ptrs[level]++;
cur               669 fs/xfs/scrub/btree.c 		if (cur->bc_ptrs[level] > be16_to_cpu(block->bb_numrecs)) {
cur               671 fs/xfs/scrub/btree.c 			if (level < cur->bc_nlevels - 1)
cur               672 fs/xfs/scrub/btree.c 				cur->bc_ptrs[level + 1]++;
cur               681 fs/xfs/scrub/btree.c 		pp = xfs_btree_ptr_addr(cur, cur->bc_ptrs[level], block);
cur               683 fs/xfs/scrub/btree.c 			cur->bc_ptrs[level]++;
cur               691 fs/xfs/scrub/btree.c 		cur->bc_ptrs[level] = 1;
cur               697 fs/xfs/scrub/btree.c 		if (!error && bs.cur)
cur                13 fs/xfs/scrub/btree.h 		struct xfs_btree_cur *cur, int level, int *error);
cur                17 fs/xfs/scrub/btree.h 		struct xfs_btree_cur *cur, int level, int *error);
cur                21 fs/xfs/scrub/btree.h 		struct xfs_btree_cur *cur, int level);
cur                25 fs/xfs/scrub/btree.h 		struct xfs_btree_cur *cur, int level);
cur                35 fs/xfs/scrub/btree.h 	struct xfs_btree_cur		*cur;
cur                47 fs/xfs/scrub/btree.h int xchk_btree(struct xfs_scrub *sc, struct xfs_btree_cur *cur,
cur               324 fs/xfs/scrub/common.c 	struct xfs_btree_cur		*cur,
cur               351 fs/xfs/scrub/common.c 	struct xfs_btree_cur		*cur,
cur               361 fs/xfs/scrub/common.c 	return xfs_rmap_query_all(cur, xchk_count_rmap_ownedby_irec,
cur               127 fs/xfs/scrub/common.h int xchk_count_rmap_ownedby_ag(struct xfs_scrub *sc, struct xfs_btree_cur *cur,
cur               106 fs/xfs/scrub/ialloc.c 	struct xfs_mount		*mp = bs->cur->bc_mp;
cur               107 fs/xfs/scrub/ialloc.c 	xfs_agnumber_t			agno = bs->cur->bc_private.a.agno;
cur               114 fs/xfs/scrub/ialloc.c 		xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               151 fs/xfs/scrub/ialloc.c 	struct xfs_mount		*mp = bs->cur->bc_mp;
cur               167 fs/xfs/scrub/ialloc.c 	fsino = XFS_AGINO_TO_INO(mp, bs->cur->bc_private.a.agno, agino);
cur               172 fs/xfs/scrub/ialloc.c 		xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               176 fs/xfs/scrub/ialloc.c 	error = xfs_icache_inode_is_allocated(mp, bs->cur->bc_tp, fsino,
cur               195 fs/xfs/scrub/ialloc.c 		xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               214 fs/xfs/scrub/ialloc.c 	struct xfs_mount		*mp = bs->cur->bc_mp;
cur               218 fs/xfs/scrub/ialloc.c 	xfs_agnumber_t			agno = bs->cur->bc_private.a.agno;
cur               253 fs/xfs/scrub/ialloc.c 		xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               265 fs/xfs/scrub/ialloc.c 		xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               281 fs/xfs/scrub/ialloc.c 	error = xfs_imap_to_bp(mp, bs->cur->bc_tp, &imap, &dip, &cluster_bp,
cur               283 fs/xfs/scrub/ialloc.c 	if (!xchk_btree_xref_process_error(bs->sc, bs->cur, 0, &error))
cur               291 fs/xfs/scrub/ialloc.c 			xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               303 fs/xfs/scrub/ialloc.c 	xfs_trans_brelse(bs->cur->bc_tp, cluster_bp);
cur               365 fs/xfs/scrub/ialloc.c 	if (bs->cur->bc_btnum == XFS_BTNUM_FINO) {
cur               371 fs/xfs/scrub/ialloc.c 			xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               382 fs/xfs/scrub/ialloc.c 			xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               398 fs/xfs/scrub/ialloc.c 		xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               403 fs/xfs/scrub/ialloc.c 		xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               425 fs/xfs/scrub/ialloc.c 	struct xfs_mount		*mp = bs->cur->bc_mp;
cur               429 fs/xfs/scrub/ialloc.c 	xfs_agnumber_t			agno = bs->cur->bc_private.a.agno;
cur               442 fs/xfs/scrub/ialloc.c 		xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               447 fs/xfs/scrub/ialloc.c 		xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               453 fs/xfs/scrub/ialloc.c 		xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               468 fs/xfs/scrub/ialloc.c 			xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               483 fs/xfs/scrub/ialloc.c 		xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               496 fs/xfs/scrub/ialloc.c 		xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               578 fs/xfs/scrub/ialloc.c 	struct xfs_btree_cur	*cur;
cur               586 fs/xfs/scrub/ialloc.c 	cur = which == XFS_BTNUM_INO ? sc->sa.ino_cur : sc->sa.fino_cur;
cur               587 fs/xfs/scrub/ialloc.c 	error = xchk_btree(sc, cur, xchk_iallocbt_rec, &XFS_RMAP_OINFO_INOBT,
cur                93 fs/xfs/scrub/refcount.c 	struct xfs_btree_cur		*cur,
cur               111 fs/xfs/scrub/refcount.c 		xchk_btree_xref_set_corrupt(refchk->sc, cur, 0);
cur               337 fs/xfs/scrub/refcount.c 	struct xfs_mount	*mp = bs->cur->bc_mp;
cur               339 fs/xfs/scrub/refcount.c 	xfs_agnumber_t		agno = bs->cur->bc_private.a.agno;
cur               352 fs/xfs/scrub/refcount.c 		xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               361 fs/xfs/scrub/refcount.c 		xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               364 fs/xfs/scrub/refcount.c 		xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               526 fs/xfs/scrub/repair.c 	struct xfs_btree_cur		*cur;
cur               550 fs/xfs/scrub/repair.c 	cur = xfs_rmapbt_init_cursor(sc->mp, sc->tp, agf_bp, agno);
cur               553 fs/xfs/scrub/repair.c 	error = xfs_rmap_has_other_keys(cur, agbno, 1, oinfo, &has_other_rmap);
cur               554 fs/xfs/scrub/repair.c 	xfs_btree_del_cursor(cur, error);
cur               828 fs/xfs/scrub/repair.c 	struct xfs_btree_cur		*cur,
cur               872 fs/xfs/scrub/repair.c 	struct xfs_btree_cur		*cur;
cur               889 fs/xfs/scrub/repair.c 	cur = xfs_rmapbt_init_cursor(mp, sc->tp, agf_bp, sc->sa.agno);
cur               890 fs/xfs/scrub/repair.c 	error = xfs_rmap_query_all(cur, xrep_findroot_rmap, &ri);
cur               891 fs/xfs/scrub/repair.c 	xfs_btree_del_cursor(cur, error);
cur                93 fs/xfs/scrub/rmap.c 	struct xfs_mount	*mp = bs->cur->bc_mp;
cur                95 fs/xfs/scrub/rmap.c 	xfs_agnumber_t		agno = bs->cur->bc_private.a.agno;
cur               103 fs/xfs/scrub/rmap.c 	if (!xchk_btree_process_error(bs->sc, bs->cur, 0, &error))
cur               108 fs/xfs/scrub/rmap.c 		xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               118 fs/xfs/scrub/rmap.c 			xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               127 fs/xfs/scrub/rmap.c 			xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               137 fs/xfs/scrub/rmap.c 		xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               140 fs/xfs/scrub/rmap.c 		xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               143 fs/xfs/scrub/rmap.c 		xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               146 fs/xfs/scrub/rmap.c 		xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               150 fs/xfs/scrub/rmap.c 			xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur               155 fs/xfs/scrub/rmap.c 			xchk_btree_set_corrupt(bs->sc, bs->cur, 0);
cur                20 fs/xfs/scrub/trace.c 	struct xfs_btree_cur	*cur,
cur                23 fs/xfs/scrub/trace.c 	if (level < cur->bc_nlevels && cur->bc_bufs[level])
cur                24 fs/xfs/scrub/trace.c 		return XFS_DADDR_TO_FSB(cur->bc_mp, cur->bc_bufs[level]->b_bn);
cur                25 fs/xfs/scrub/trace.c 	else if (level == cur->bc_nlevels - 1 &&
cur                26 fs/xfs/scrub/trace.c 		 cur->bc_flags & XFS_BTREE_LONG_PTRS)
cur                27 fs/xfs/scrub/trace.c 		return XFS_INO_TO_FSB(cur->bc_mp, cur->bc_private.b.ip->i_ino);
cur                28 fs/xfs/scrub/trace.c 	else if (!(cur->bc_flags & XFS_BTREE_LONG_PTRS))
cur                29 fs/xfs/scrub/trace.c 		return XFS_AGB_TO_FSB(cur->bc_mp, cur->bc_private.a.agno, 0);
cur               322 fs/xfs/scrub/trace.h 	TP_PROTO(struct xfs_scrub *sc, struct xfs_btree_cur *cur,
cur               324 fs/xfs/scrub/trace.h 	TP_ARGS(sc, cur, level, error, ret_ip),
cur               337 fs/xfs/scrub/trace.h 		xfs_fsblock_t fsbno = xchk_btree_cur_fsbno(cur, level);
cur               341 fs/xfs/scrub/trace.h 		__entry->btnum = cur->bc_btnum;
cur               343 fs/xfs/scrub/trace.h 		__entry->agno = XFS_FSB_TO_AGNO(cur->bc_mp, fsbno);
cur               344 fs/xfs/scrub/trace.h 		__entry->bno = XFS_FSB_TO_AGBNO(cur->bc_mp, fsbno);
cur               345 fs/xfs/scrub/trace.h 		__entry->ptr = cur->bc_ptrs[level];
cur               362 fs/xfs/scrub/trace.h 	TP_PROTO(struct xfs_scrub *sc, struct xfs_btree_cur *cur,
cur               364 fs/xfs/scrub/trace.h 	TP_ARGS(sc, cur, level, error, ret_ip),
cur               379 fs/xfs/scrub/trace.h 		xfs_fsblock_t fsbno = xchk_btree_cur_fsbno(cur, level);
cur               382 fs/xfs/scrub/trace.h 		__entry->whichfork = cur->bc_private.b.whichfork;
cur               384 fs/xfs/scrub/trace.h 		__entry->btnum = cur->bc_btnum;
cur               386 fs/xfs/scrub/trace.h 		__entry->ptr = cur->bc_ptrs[level];
cur               387 fs/xfs/scrub/trace.h 		__entry->agno = XFS_FSB_TO_AGNO(cur->bc_mp, fsbno);
cur               388 fs/xfs/scrub/trace.h 		__entry->bno = XFS_FSB_TO_AGBNO(cur->bc_mp, fsbno);
cur               407 fs/xfs/scrub/trace.h 	TP_PROTO(struct xfs_scrub *sc, struct xfs_btree_cur *cur,
cur               409 fs/xfs/scrub/trace.h 	TP_ARGS(sc, cur, level, ret_ip),
cur               421 fs/xfs/scrub/trace.h 		xfs_fsblock_t fsbno = xchk_btree_cur_fsbno(cur, level);
cur               424 fs/xfs/scrub/trace.h 		__entry->btnum = cur->bc_btnum;
cur               426 fs/xfs/scrub/trace.h 		__entry->agno = XFS_FSB_TO_AGNO(cur->bc_mp, fsbno);
cur               427 fs/xfs/scrub/trace.h 		__entry->bno = XFS_FSB_TO_AGBNO(cur->bc_mp, fsbno);
cur               428 fs/xfs/scrub/trace.h 		__entry->ptr = cur->bc_ptrs[level];
cur               443 fs/xfs/scrub/trace.h 	TP_PROTO(struct xfs_scrub *sc, struct xfs_btree_cur *cur,
cur               445 fs/xfs/scrub/trace.h 	TP_ARGS(sc, cur, level, ret_ip),
cur               459 fs/xfs/scrub/trace.h 		xfs_fsblock_t fsbno = xchk_btree_cur_fsbno(cur, level);
cur               462 fs/xfs/scrub/trace.h 		__entry->whichfork = cur->bc_private.b.whichfork;
cur               464 fs/xfs/scrub/trace.h 		__entry->btnum = cur->bc_btnum;
cur               466 fs/xfs/scrub/trace.h 		__entry->agno = XFS_FSB_TO_AGNO(cur->bc_mp, fsbno);
cur               467 fs/xfs/scrub/trace.h 		__entry->bno = XFS_FSB_TO_AGBNO(cur->bc_mp, fsbno);
cur               468 fs/xfs/scrub/trace.h 		__entry->ptr = cur->bc_ptrs[level];
cur               485 fs/xfs/scrub/trace.h 	TP_PROTO(struct xfs_scrub *sc, struct xfs_btree_cur *cur,
cur               487 fs/xfs/scrub/trace.h 	TP_ARGS(sc, cur, level),
cur               499 fs/xfs/scrub/trace.h 		xfs_fsblock_t fsbno = xchk_btree_cur_fsbno(cur, level);
cur               503 fs/xfs/scrub/trace.h 		__entry->btnum = cur->bc_btnum;
cur               504 fs/xfs/scrub/trace.h 		__entry->agno = XFS_FSB_TO_AGNO(cur->bc_mp, fsbno);
cur               505 fs/xfs/scrub/trace.h 		__entry->bno = XFS_FSB_TO_AGBNO(cur->bc_mp, fsbno);
cur               507 fs/xfs/scrub/trace.h 		__entry->nlevels = cur->bc_nlevels;
cur               508 fs/xfs/scrub/trace.h 		__entry->ptr = cur->bc_ptrs[level];
cur               522 fs/xfs/scrub/trace.h 	TP_PROTO(struct xfs_scrub *sc, struct xfs_btree_cur *cur, \
cur               524 fs/xfs/scrub/trace.h 	TP_ARGS(sc, cur, level))
cur                31 fs/xfs/xfs_discard.c 	struct xfs_btree_cur	*cur;
cur                50 fs/xfs/xfs_discard.c 	cur = xfs_allocbt_init_cursor(mp, NULL, agbp, agno, XFS_BTNUM_CNT);
cur                55 fs/xfs/xfs_discard.c 	error = xfs_alloc_lookup_ge(cur, 0,
cur                70 fs/xfs/xfs_discard.c 		error = xfs_alloc_get_rec(cur, &fbno, &flen, &i);
cur               118 fs/xfs/xfs_discard.c 		error = xfs_btree_decrement(cur, 0, &i);
cur               129 fs/xfs/xfs_discard.c 	xfs_btree_del_cursor(cur, error);
cur               613 fs/xfs/xfs_dquot.c 	struct xfs_iext_cursor	cur;
cur               637 fs/xfs/xfs_dquot.c 	if (xfs_iext_lookup_extent(quotip, &quotip->i_df, start, &cur, &got)) {
cur               198 fs/xfs/xfs_fsmap.c 	struct xfs_btree_cur		*cur;
cur               212 fs/xfs/xfs_fsmap.c 	cur = xfs_refcountbt_init_cursor(mp, tp, info->agf_bp,
cur               215 fs/xfs/xfs_fsmap.c 	error = xfs_refcount_find_shared(cur, rec->rm_startblock,
cur               218 fs/xfs/xfs_fsmap.c 	xfs_btree_del_cursor(cur, error);
cur               337 fs/xfs/xfs_fsmap.c 	struct xfs_btree_cur		*cur,
cur               341 fs/xfs/xfs_fsmap.c 	struct xfs_mount		*mp = cur->bc_mp;
cur               346 fs/xfs/xfs_fsmap.c 	fsb = XFS_AGB_TO_FSB(mp, cur->bc_private.a.agno, rec->rm_startblock);
cur               349 fs/xfs/xfs_fsmap.c 	return xfs_getfsmap_helper(cur->bc_tp, info, rec, rec_daddr);
cur               355 fs/xfs/xfs_fsmap.c 	struct xfs_btree_cur		*cur,
cur               359 fs/xfs/xfs_fsmap.c 	struct xfs_mount		*mp = cur->bc_mp;
cur               364 fs/xfs/xfs_fsmap.c 	rec_daddr = XFS_AGB_TO_DADDR(mp, cur->bc_private.a.agno,
cur               373 fs/xfs/xfs_fsmap.c 	return xfs_getfsmap_helper(cur->bc_tp, info, &irec, rec_daddr);
cur               385 fs/xfs/xfs_iwalk.c 	struct xfs_btree_cur		*cur = NULL;
cur               394 fs/xfs/xfs_iwalk.c 	error = xfs_iwalk_ag_start(iwag, agno, agino, &cur, &agi_bp, &has_more);
cur               405 fs/xfs/xfs_iwalk.c 		error = xfs_inobt_get_rec(cur, irec, &has_more);
cur               411 fs/xfs/xfs_iwalk.c 			error = xfs_btree_increment(cur, 0, &has_more);
cur               429 fs/xfs/xfs_iwalk.c 			error = xfs_btree_increment(cur, 0, &has_more);
cur               442 fs/xfs/xfs_iwalk.c 		error = xfs_iwalk_run_callbacks(iwag, agno, &cur, &agi_bp,
cur               450 fs/xfs/xfs_iwalk.c 	error = xfs_iwalk_run_callbacks(iwag, agno, &cur, &agi_bp, &has_more);
cur               453 fs/xfs/xfs_iwalk.c 	xfs_iwalk_del_inobt(tp, &cur, &agi_bp, error);
cur              3388 fs/xfs/xfs_log_recover.c 	struct xfs_ail_cursor	cur;
cur              3403 fs/xfs/xfs_log_recover.c 	lip = xfs_trans_ail_cursor_first(ailp, &cur, 0);
cur              3418 fs/xfs/xfs_log_recover.c 		lip = xfs_trans_ail_cursor_next(ailp, &cur);
cur              3421 fs/xfs/xfs_log_recover.c 	xfs_trans_ail_cursor_done(&cur);
cur              3484 fs/xfs/xfs_log_recover.c 	struct xfs_ail_cursor		cur;
cur              3496 fs/xfs/xfs_log_recover.c 	lip = xfs_trans_ail_cursor_first(ailp, &cur, 0);
cur              3511 fs/xfs/xfs_log_recover.c 		lip = xfs_trans_ail_cursor_next(ailp, &cur);
cur              3514 fs/xfs/xfs_log_recover.c 	xfs_trans_ail_cursor_done(&cur);
cur              3600 fs/xfs/xfs_log_recover.c 	struct xfs_ail_cursor		cur;
cur              3613 fs/xfs/xfs_log_recover.c 	lip = xfs_trans_ail_cursor_first(ailp, &cur, 0);
cur              3628 fs/xfs/xfs_log_recover.c 		lip = xfs_trans_ail_cursor_next(ailp, &cur);
cur              3631 fs/xfs/xfs_log_recover.c 	xfs_trans_ail_cursor_done(&cur);
cur              3719 fs/xfs/xfs_log_recover.c 	struct xfs_ail_cursor		cur;
cur              3732 fs/xfs/xfs_log_recover.c 	lip = xfs_trans_ail_cursor_first(ailp, &cur, 0);
cur              3747 fs/xfs/xfs_log_recover.c 		lip = xfs_trans_ail_cursor_next(ailp, &cur);
cur              3750 fs/xfs/xfs_log_recover.c 	xfs_trans_ail_cursor_done(&cur);
cur              4795 fs/xfs/xfs_log_recover.c 	struct xfs_ail_cursor	cur;
cur              4818 fs/xfs/xfs_log_recover.c 	lip = xfs_trans_ail_cursor_first(ailp, &cur, 0);
cur              4829 fs/xfs/xfs_log_recover.c 			for (; lip; lip = xfs_trans_ail_cursor_next(ailp, &cur))
cur              4864 fs/xfs/xfs_log_recover.c 		lip = xfs_trans_ail_cursor_next(ailp, &cur);
cur              4867 fs/xfs/xfs_log_recover.c 	xfs_trans_ail_cursor_done(&cur);
cur              4885 fs/xfs/xfs_log_recover.c 	struct xfs_ail_cursor	cur;
cur              4890 fs/xfs/xfs_log_recover.c 	lip = xfs_trans_ail_cursor_first(ailp, &cur, 0);
cur              4898 fs/xfs/xfs_log_recover.c 			for (; lip; lip = xfs_trans_ail_cursor_next(ailp, &cur))
cur              4919 fs/xfs/xfs_log_recover.c 		lip = xfs_trans_ail_cursor_next(ailp, &cur);
cur              4922 fs/xfs/xfs_log_recover.c 	xfs_trans_ail_cursor_done(&cur);
cur               140 fs/xfs/xfs_reflink.c 	struct xfs_btree_cur	*cur;
cur               149 fs/xfs/xfs_reflink.c 	cur = xfs_refcountbt_init_cursor(mp, tp, agbp, agno);
cur               151 fs/xfs/xfs_reflink.c 	error = xfs_refcount_find_shared(cur, agbno, aglen, fbno, flen,
cur               154 fs/xfs/xfs_reflink.c 	xfs_btree_del_cursor(cur, error);
cur               215 fs/xfs/xfs_trace.h 	TP_PROTO(struct xfs_inode *ip, struct xfs_iext_cursor *cur, int state,
cur               217 fs/xfs/xfs_trace.h 	TP_ARGS(ip, cur, state, caller_ip),
cur               235 fs/xfs/xfs_trace.h 		xfs_iext_get_extent(ifp, cur, &r);
cur               238 fs/xfs/xfs_trace.h 		__entry->leaf = cur->leaf;
cur               239 fs/xfs/xfs_trace.h 		__entry->pos = cur->pos;
cur               263 fs/xfs/xfs_trace.h 	TP_PROTO(struct xfs_inode *ip, struct xfs_iext_cursor *cur, int state, \
cur               265 fs/xfs/xfs_trace.h 	TP_ARGS(ip, cur, state, caller_ip))
cur              2205 fs/xfs/xfs_trace.h 	TP_PROTO(struct xfs_btree_cur *cur, int level, struct xfs_buf *bp),
cur              2206 fs/xfs/xfs_trace.h 	TP_ARGS(cur, level, bp),
cur              2216 fs/xfs/xfs_trace.h 		__entry->dev = cur->bc_mp->m_super->s_dev;
cur              2217 fs/xfs/xfs_trace.h 		__entry->btnum = cur->bc_btnum;
cur              2219 fs/xfs/xfs_trace.h 		__entry->nlevels = cur->bc_nlevels;
cur              2220 fs/xfs/xfs_trace.h 		__entry->ptr = cur->bc_ptrs[level];
cur              2234 fs/xfs/xfs_trace.h 	TP_PROTO(struct xfs_btree_cur *cur, int level, struct xfs_buf *bp), \
cur              2235 fs/xfs/xfs_trace.h 	TP_ARGS(cur, level, bp))
cur               788 fs/xfs/xfs_trans.c 	struct xfs_ail_cursor	*cur,
cur               797 fs/xfs/xfs_trans.c 	xfs_trans_ail_update_bulk(ailp, cur, log_items, nr_items, commit_lsn);
cur               837 fs/xfs/xfs_trans.c 	struct xfs_ail_cursor	cur;
cur               841 fs/xfs/xfs_trans.c 	xfs_trans_ail_cursor_last(ailp, &cur, commit_lsn);
cur               899 fs/xfs/xfs_trans.c 			xfs_log_item_batch_insert(ailp, &cur, log_items,
cur               907 fs/xfs/xfs_trans.c 		xfs_log_item_batch_insert(ailp, &cur, log_items, i, commit_lsn);
cur               910 fs/xfs/xfs_trans.c 	xfs_trans_ail_cursor_done(&cur);
cur               156 fs/xfs/xfs_trans_ail.c 	struct xfs_ail_cursor	*cur)
cur               158 fs/xfs/xfs_trans_ail.c 	cur->item = NULL;
cur               159 fs/xfs/xfs_trans_ail.c 	list_add_tail(&cur->list, &ailp->ail_cursors);
cur               169 fs/xfs/xfs_trans_ail.c 	struct xfs_ail_cursor	*cur)
cur               171 fs/xfs/xfs_trans_ail.c 	struct xfs_log_item	*lip = cur->item;
cur               176 fs/xfs/xfs_trans_ail.c 		cur->item = xfs_ail_next(ailp, lip);
cur               186 fs/xfs/xfs_trans_ail.c 	struct xfs_ail_cursor	*cur)
cur               188 fs/xfs/xfs_trans_ail.c 	cur->item = NULL;
cur               189 fs/xfs/xfs_trans_ail.c 	list_del_init(&cur->list);
cur               205 fs/xfs/xfs_trans_ail.c 	struct xfs_ail_cursor	*cur;
cur               207 fs/xfs/xfs_trans_ail.c 	list_for_each_entry(cur, &ailp->ail_cursors, list) {
cur               208 fs/xfs/xfs_trans_ail.c 		if (cur->item == lip)
cur               209 fs/xfs/xfs_trans_ail.c 			cur->item = (struct xfs_log_item *)
cur               210 fs/xfs/xfs_trans_ail.c 					((uintptr_t)cur->item | 1);
cur               223 fs/xfs/xfs_trans_ail.c 	struct xfs_ail_cursor	*cur,
cur               228 fs/xfs/xfs_trans_ail.c 	xfs_trans_ail_cursor_init(ailp, cur);
cur               243 fs/xfs/xfs_trans_ail.c 		cur->item = xfs_ail_next(ailp, lip);
cur               270 fs/xfs/xfs_trans_ail.c 	struct xfs_ail_cursor	*cur,
cur               273 fs/xfs/xfs_trans_ail.c 	xfs_trans_ail_cursor_init(ailp, cur);
cur               274 fs/xfs/xfs_trans_ail.c 	cur->item = __xfs_trans_ail_cursor_last(ailp, lsn);
cur               275 fs/xfs/xfs_trans_ail.c 	return cur->item;
cur               287 fs/xfs/xfs_trans_ail.c 	struct xfs_ail_cursor	*cur,
cur               300 fs/xfs/xfs_trans_ail.c 	lip = cur ? cur->item : NULL;
cur               311 fs/xfs/xfs_trans_ail.c 	if (cur)
cur               312 fs/xfs/xfs_trans_ail.c 		cur->item = list_entry(list->prev, struct xfs_log_item, li_ail);
cur               367 fs/xfs/xfs_trans_ail.c 	struct xfs_ail_cursor	cur;
cur               397 fs/xfs/xfs_trans_ail.c 	lip = xfs_trans_ail_cursor_first(ailp, &cur, ailp->ail_last_pushed_lsn);
cur               403 fs/xfs/xfs_trans_ail.c 		xfs_trans_ail_cursor_done(&cur);
cur               483 fs/xfs/xfs_trans_ail.c 		lip = xfs_trans_ail_cursor_next(ailp, &cur);
cur               488 fs/xfs/xfs_trans_ail.c 	xfs_trans_ail_cursor_done(&cur);
cur               708 fs/xfs/xfs_trans_ail.c 	struct xfs_ail_cursor	*cur,
cur               740 fs/xfs/xfs_trans_ail.c 		xfs_ail_splice(ailp, cur, &tmp, lsn);
cur                70 fs/xfs/xfs_trans_priv.h 				struct xfs_ail_cursor *cur,
cur               120 fs/xfs/xfs_trans_priv.h 					struct xfs_ail_cursor *cur,
cur               123 fs/xfs/xfs_trans_priv.h 					struct xfs_ail_cursor *cur,
cur               126 fs/xfs/xfs_trans_priv.h 					struct xfs_ail_cursor *cur);
cur               127 fs/xfs/xfs_trans_priv.h void			xfs_trans_ail_cursor_done(struct xfs_ail_cursor *cur);
cur                68 include/crypto/if_alg.h 	unsigned int cur;		/* Last processed SG entry */
cur               786 include/linux/blk-cgroup.h 		int cur = atomic_cmpxchg(&blkg->use_delay, old, old - 1);
cur               787 include/linux/blk-cgroup.h 		if (cur == old)
cur               789 include/linux/blk-cgroup.h 		old = cur;
cur               806 include/linux/blk-cgroup.h 		int cur = atomic_cmpxchg(&blkg->use_delay, old, 0);
cur               807 include/linux/blk-cgroup.h 		if (cur == old) {
cur               811 include/linux/blk-cgroup.h 		old = cur;
cur               403 include/linux/bpf_verifier.h 	struct bpf_verifier_state *cur = env->cur_state;
cur               405 include/linux/bpf_verifier.h 	return cur->frame[cur->curframe];
cur                97 include/linux/bvec.h 		const struct bio_vec *cur = bv + iter->bi_idx;
cur                99 include/linux/bvec.h 				    cur->bv_len - iter->bi_bvec_done);
cur               105 include/linux/bvec.h 		if (iter->bi_bvec_done == cur->bv_len) {
cur               199 include/linux/ceph/libceph.h 		type *cur = rb_entry(*n, type, nodefld);		\
cur               203 include/linux/ceph/libceph.h 		cmp = cmpexp(keyexp(t->keyfld), keyexp(cur->keyfld));	\
cur               233 include/linux/ceph/libceph.h 		type *cur = rb_entry(n, type, nodefld);			\
cur               236 include/linux/ceph/libceph.h 		cmp = cmpexp(key, keyexp(cur->keyfld));			\
cur               242 include/linux/ceph/libceph.h 			return cur;					\
cur                66 include/linux/cpufreq.h 	unsigned int		cur;    /* in kHz, only needed if cpufreq
cur               583 include/linux/cpufreq.h 	if (policy->max < policy->cur)
cur               585 include/linux/cpufreq.h 	else if (policy->min > policy->cur)
cur               236 include/linux/device.h bus_find_next_device(struct bus_type *bus,struct device *cur)
cur               238 include/linux/device.h 	return bus_find_device(bus, cur, NULL, device_match_any);
cur               287 include/linux/if_team.h 	struct team_port *cur;
cur               291 include/linux/if_team.h 	cur = port;
cur               292 include/linux/if_team.h 	list_for_each_entry_continue_rcu(cur, &team->port_list, list)
cur               293 include/linux/if_team.h 		if (team_port_txable(cur))
cur               294 include/linux/if_team.h 			return cur;
cur               295 include/linux/if_team.h 	list_for_each_entry_rcu(cur, &team->port_list, list) {
cur               296 include/linux/if_team.h 		if (cur == port)
cur               298 include/linux/if_team.h 		if (team_port_txable(cur))
cur               299 include/linux/if_team.h 			return cur;
cur               126 include/linux/iversion.h 	u64 cur, old;
cur               128 include/linux/iversion.h 	cur = inode_peek_iversion_raw(inode);
cur               130 include/linux/iversion.h 		if (cur > val)
cur               132 include/linux/iversion.h 		old = atomic64_cmpxchg(&inode->i_version, cur, val);
cur               133 include/linux/iversion.h 		if (likely(old == cur))
cur               135 include/linux/iversion.h 		cur = old;
cur               200 include/linux/iversion.h 	u64 cur, old, new;
cur               213 include/linux/iversion.h 	cur = inode_peek_iversion_raw(inode);
cur               216 include/linux/iversion.h 		if (!force && !(cur & I_VERSION_QUERIED))
cur               220 include/linux/iversion.h 		new = (cur & ~I_VERSION_QUERIED) + I_VERSION_INCREMENT;
cur               222 include/linux/iversion.h 		old = atomic64_cmpxchg(&inode->i_version, cur, new);
cur               223 include/linux/iversion.h 		if (likely(old == cur))
cur               225 include/linux/iversion.h 		cur = old;
cur               307 include/linux/iversion.h 	u64 cur, old, new;
cur               309 include/linux/iversion.h 	cur = inode_peek_iversion_raw(inode);
cur               312 include/linux/iversion.h 		if (cur & I_VERSION_QUERIED) {
cur               322 include/linux/iversion.h 		new = cur | I_VERSION_QUERIED;
cur               323 include/linux/iversion.h 		old = atomic64_cmpxchg(&inode->i_version, cur, new);
cur               324 include/linux/iversion.h 		if (likely(old == cur))
cur               326 include/linux/iversion.h 		cur = old;
cur               328 include/linux/iversion.h 	return cur >> I_VERSION_QUERIED_SHIFT;
cur               267 include/linux/lockd/lockd.h typedef int	  (*nlm_host_match_fn_t)(void *cur, struct nlm_host *ref);
cur                25 include/linux/mdio-mux.h 		  int (*switch_fn) (int cur, int desired, void *data),
cur                91 include/linux/of.h 	const __be32 *cur;
cur               540 include/linux/of.h const __be32 *of_prop_next_u32(struct property *prop, const __be32 *cur,
cur               549 include/linux/of.h const char *of_prop_next_string(struct property *prop, const char *cur);
cur               925 include/linux/of.h 		const __be32 *cur, u32 *pu)
cur               931 include/linux/of.h 		const char *cur)
cur                 7 include/linux/s3c_adc_battery.h 	int cur; /* mA */
cur              1571 include/linux/sched.h extern int cpuset_cpumask_can_shrink(const struct cpumask *cur, const struct cpumask *trial);
cur                18 include/linux/swap_slots.h 	int		cur;
cur               172 include/linux/usb/pd_vdo.h #define VDO_CABLE(hw, fw, cbl, gdr, lat, term, tx1d, tx2d, rx1d, rx2d, cur,\
cur               177 include/linux/usb/pd_vdo.h 	 | ((cur) & 0x3) << 5 | (vps) << 4 | (sopp) << 3		\
cur               255 include/media/v4l2-ctrls.h 	} cur;
cur               170 kernel/bpf/bpf_lru_list.c 	struct list_head *cur, *last, *next = inactive;
cur               181 kernel/bpf/bpf_lru_list.c 	cur = l->next_inactive_rotation;
cur               183 kernel/bpf/bpf_lru_list.c 		if (cur == inactive) {
cur               184 kernel/bpf/bpf_lru_list.c 			cur = cur->prev;
cur               188 kernel/bpf/bpf_lru_list.c 		node = list_entry(cur, struct bpf_lru_node, list);
cur               189 kernel/bpf/bpf_lru_list.c 		next = cur->prev;
cur               192 kernel/bpf/bpf_lru_list.c 		if (cur == last)
cur               194 kernel/bpf/bpf_lru_list.c 		cur = next;
cur              2970 kernel/bpf/btf.c 	void *cur, *end;
cur              2973 kernel/bpf/btf.c 	cur = btf->nohdr_data + hdr->type_off;
cur              2974 kernel/bpf/btf.c 	end = cur + hdr->type_len;
cur              2977 kernel/bpf/btf.c 	while (cur < end) {
cur              2978 kernel/bpf/btf.c 		struct btf_type *t = cur;
cur              2981 kernel/bpf/btf.c 		meta_size = btf_check_meta(env, t, end - cur);
cur              2986 kernel/bpf/btf.c 		cur += meta_size;
cur               427 kernel/bpf/verifier.c 	struct bpf_verifier_state *cur = env->cur_state;
cur               429 kernel/bpf/verifier.c 	return cur->frame[reg->frameno];
cur               778 kernel/bpf/verifier.c 	struct bpf_verifier_state *cur = env->cur_state;
cur               785 kernel/bpf/verifier.c 	if (cur) {
cur               786 kernel/bpf/verifier.c 		err = copy_verifier_state(cur, &head->st);
cur               806 kernel/bpf/verifier.c 	struct bpf_verifier_state *cur = env->cur_state;
cur               819 kernel/bpf/verifier.c 	err = copy_verifier_state(&elem->st, cur);
cur              1405 kernel/bpf/verifier.c 			    struct bpf_verifier_state *cur)
cur              1407 kernel/bpf/verifier.c 	u32 cnt = cur->jmp_history_cnt;
cur              1411 kernel/bpf/verifier.c 	p = krealloc(cur->jmp_history, cnt * sizeof(*p), GFP_USER);
cur              1416 kernel/bpf/verifier.c 	cur->jmp_history = p;
cur              1417 kernel/bpf/verifier.c 	cur->jmp_history_cnt = cnt;
cur              1897 kernel/bpf/verifier.c 	struct bpf_func_state *cur; /* state of the current function */
cur              1916 kernel/bpf/verifier.c 	cur = env->cur_state->frame[env->cur_state->curframe];
cur              1918 kernel/bpf/verifier.c 		reg = &cur->regs[value_regno];
cur              1942 kernel/bpf/verifier.c 		if (state != cur && reg->type == PTR_TO_STACK) {
cur              3144 kernel/bpf/verifier.c 	struct bpf_verifier_state *cur = env->cur_state;
cur              3186 kernel/bpf/verifier.c 		if (cur->active_spin_lock) {
cur              3191 kernel/bpf/verifier.c 		cur->active_spin_lock = reg->id;
cur              3193 kernel/bpf/verifier.c 		if (!cur->active_spin_lock) {
cur              3197 kernel/bpf/verifier.c 		if (cur->active_spin_lock != reg->id) {
cur              3201 kernel/bpf/verifier.c 		cur->active_spin_lock = 0;
cur              6290 kernel/bpf/verifier.c 	struct bpf_verifier_state *cur = env->cur_state;
cur              6291 kernel/bpf/verifier.c 	struct bpf_func_state *state = cur->frame[cur->curframe];
cur              6754 kernel/bpf/verifier.c 			 struct bpf_reg_state *cur)
cur              6756 kernel/bpf/verifier.c 	return old->umin_value <= cur->umin_value &&
cur              6757 kernel/bpf/verifier.c 	       old->umax_value >= cur->umax_value &&
cur              6758 kernel/bpf/verifier.c 	       old->smin_value <= cur->smin_value &&
cur              6759 kernel/bpf/verifier.c 	       old->smax_value >= cur->smax_value;
cur              6766 kernel/bpf/verifier.c 	u32 cur;
cur              6787 kernel/bpf/verifier.c 			idmap[i].cur = cur_id;
cur              6791 kernel/bpf/verifier.c 			return idmap[i].cur == cur_id;
cur              6873 kernel/bpf/verifier.c 			      struct bpf_verifier_state *cur)
cur              6883 kernel/bpf/verifier.c 		    sl->state.curframe != cur->curframe)
cur              6885 kernel/bpf/verifier.c 		for (i = 0; i <= cur->curframe; i++)
cur              6886 kernel/bpf/verifier.c 			if (sl->state.frame[i]->callsite != cur->frame[i]->callsite)
cur              7012 kernel/bpf/verifier.c 		      struct bpf_func_state *cur,
cur              7036 kernel/bpf/verifier.c 		if (i >= cur->allocated_stack)
cur              7044 kernel/bpf/verifier.c 		    cur->stack[spi].slot_type[i % BPF_REG_SIZE] == STACK_ZERO)
cur              7047 kernel/bpf/verifier.c 		    cur->stack[spi].slot_type[i % BPF_REG_SIZE])
cur              7059 kernel/bpf/verifier.c 			     &cur->stack[spi].spilled_ptr,
cur              7076 kernel/bpf/verifier.c static bool refsafe(struct bpf_func_state *old, struct bpf_func_state *cur)
cur              7078 kernel/bpf/verifier.c 	if (old->acquired_refs != cur->acquired_refs)
cur              7080 kernel/bpf/verifier.c 	return !memcmp(old->refs, cur->refs,
cur              7111 kernel/bpf/verifier.c 			      struct bpf_func_state *cur)
cur              7123 kernel/bpf/verifier.c 		if (!regsafe(&old->regs[i], &cur->regs[i], idmap))
cur              7127 kernel/bpf/verifier.c 	if (!stacksafe(old, cur, idmap))
cur              7130 kernel/bpf/verifier.c 	if (!refsafe(old, cur))
cur              7140 kernel/bpf/verifier.c 			 struct bpf_verifier_state *cur)
cur              7144 kernel/bpf/verifier.c 	if (old->curframe != cur->curframe)
cur              7150 kernel/bpf/verifier.c 	if (old->speculative && !cur->speculative)
cur              7153 kernel/bpf/verifier.c 	if (old->active_spin_lock != cur->active_spin_lock)
cur              7160 kernel/bpf/verifier.c 		if (old->frame[i]->callsite != cur->frame[i]->callsite)
cur              7162 kernel/bpf/verifier.c 		if (!func_states_equal(old->frame[i], cur->frame[i]))
cur              7289 kernel/bpf/verifier.c 				 struct bpf_verifier_state *cur)
cur              7292 kernel/bpf/verifier.c 	int i, fr = cur->curframe;
cur              7298 kernel/bpf/verifier.c 	fcur = cur->frame[fr];
cur              7311 kernel/bpf/verifier.c 	struct bpf_verifier_state *cur = env->cur_state, *new;
cur              7315 kernel/bpf/verifier.c 	cur->last_insn_idx = env->prev_insn_idx;
cur              7337 kernel/bpf/verifier.c 	clean_live_states(env, insn_idx, cur);
cur              7344 kernel/bpf/verifier.c 			if (states_maybe_looping(&sl->state, cur) &&
cur              7345 kernel/bpf/verifier.c 			    states_equal(env, &sl->state, cur)) {
cur              7367 kernel/bpf/verifier.c 		if (states_equal(env, &sl->state, cur)) {
cur              7379 kernel/bpf/verifier.c 			err = propagate_liveness(env, &sl->state, cur);
cur              7386 kernel/bpf/verifier.c 			err = err ? : push_jmp_history(env, cur);
cur              7440 kernel/bpf/verifier.c 		return push_jmp_history(env, cur);
cur              7443 kernel/bpf/verifier.c 		return push_jmp_history(env, cur);
cur              7464 kernel/bpf/verifier.c 	err = copy_verifier_state(new, cur);
cur              7474 kernel/bpf/verifier.c 	cur->parent = new;
cur              7475 kernel/bpf/verifier.c 	cur->first_insn_idx = insn_idx;
cur              7476 kernel/bpf/verifier.c 	clear_jmp_history(cur);
cur              7492 kernel/bpf/verifier.c 	for (j = 0; j <= cur->curframe; j++) {
cur              7493 kernel/bpf/verifier.c 		for (i = j < cur->curframe ? BPF_REG_6 : 0; i < BPF_REG_FP; i++)
cur              7494 kernel/bpf/verifier.c 			cur->frame[j]->regs[i].parent = &new->frame[j]->regs[i];
cur              7496 kernel/bpf/verifier.c 			cur->frame[j]->regs[i].live = REG_LIVE_NONE;
cur              7500 kernel/bpf/verifier.c 	for (j = 0; j <= cur->curframe; j++) {
cur              7501 kernel/bpf/verifier.c 		struct bpf_func_state *frame = cur->frame[j];
cur               563 kernel/cgroup/cpuset.c static int validate_change(struct cpuset *cur, struct cpuset *trial)
cur               573 kernel/cgroup/cpuset.c 	cpuset_for_each_child(c, css, cur)
cur               579 kernel/cgroup/cpuset.c 	if (cur == &top_cpuset)
cur               582 kernel/cgroup/cpuset.c 	par = parent_cs(cur);
cur               596 kernel/cgroup/cpuset.c 		    c != cur &&
cur               600 kernel/cgroup/cpuset.c 		    c != cur &&
cur               610 kernel/cgroup/cpuset.c 	if ((cgroup_is_populated(cur->css.cgroup) || cur->attach_in_progress)) {
cur               611 kernel/cgroup/cpuset.c 		if (!cpumask_empty(cur->cpus_allowed) &&
cur               614 kernel/cgroup/cpuset.c 		if (!nodes_empty(cur->mems_allowed) &&
cur               624 kernel/cgroup/cpuset.c 	if (is_cpu_exclusive(cur) &&
cur               625 kernel/cgroup/cpuset.c 	    !cpuset_cpumask_can_shrink(cur->cpus_allowed,
cur              2112 kernel/cpu.c   	ssize_t cur, res = 0;
cur              2120 kernel/cpu.c   			cur = sprintf(buf, "%3d: %s\n", i, sp->name);
cur              2121 kernel/cpu.c   			buf += cur;
cur              2122 kernel/cpu.c   			res += cur;
cur                41 kernel/crash_core.c 	char *cur = cmdline, *tmp;
cur                48 kernel/crash_core.c 		start = memparse(cur, &tmp);
cur                49 kernel/crash_core.c 		if (cur == tmp) {
cur                53 kernel/crash_core.c 		cur = tmp;
cur                54 kernel/crash_core.c 		if (*cur != '-') {
cur                58 kernel/crash_core.c 		cur++;
cur                61 kernel/crash_core.c 		if (*cur != ':') {
cur                62 kernel/crash_core.c 			end = memparse(cur, &tmp);
cur                63 kernel/crash_core.c 			if (cur == tmp) {
cur                67 kernel/crash_core.c 			cur = tmp;
cur                74 kernel/crash_core.c 		if (*cur != ':') {
cur                78 kernel/crash_core.c 		cur++;
cur                80 kernel/crash_core.c 		size = memparse(cur, &tmp);
cur                81 kernel/crash_core.c 		if (cur == tmp) {
cur                85 kernel/crash_core.c 		cur = tmp;
cur                96 kernel/crash_core.c 	} while (*cur++ == ',');
cur                99 kernel/crash_core.c 		while (*cur && *cur != ' ' && *cur != '@')
cur               100 kernel/crash_core.c 			cur++;
cur               101 kernel/crash_core.c 		if (*cur == '@') {
cur               102 kernel/crash_core.c 			cur++;
cur               103 kernel/crash_core.c 			*crash_base = memparse(cur, &tmp);
cur               104 kernel/crash_core.c 			if (cur == tmp) {
cur               126 kernel/crash_core.c 	char *cur = cmdline;
cur               128 kernel/crash_core.c 	*crash_size = memparse(cmdline, &cur);
cur               129 kernel/crash_core.c 	if (cmdline == cur) {
cur               134 kernel/crash_core.c 	if (*cur == '@')
cur               135 kernel/crash_core.c 		*crash_base = memparse(cur+1, &cur);
cur               136 kernel/crash_core.c 	else if (*cur != ' ' && *cur != '\0') {
cur               137 kernel/crash_core.c 		pr_warn("crashkernel: unrecognized char: %c\n", *cur);
cur               164 kernel/crash_core.c 	char *cur = cmdline;
cur               166 kernel/crash_core.c 	*crash_size = memparse(cmdline, &cur);
cur               167 kernel/crash_core.c 	if (cmdline == cur) {
cur               173 kernel/crash_core.c 	if (strncmp(cur, suffix, strlen(suffix))) {
cur               174 kernel/crash_core.c 		pr_warn("crashkernel: unrecognized char: %c\n", *cur);
cur               177 kernel/crash_core.c 	cur += strlen(suffix);
cur               178 kernel/crash_core.c 	if (*cur != ' ' && *cur != '\0') {
cur               179 kernel/crash_core.c 		pr_warn("crashkernel: unrecognized char: %c\n", *cur);
cur              1150 kernel/kprobes.c 	struct kprobe *cur = __this_cpu_read(kprobe_instance);
cur              1156 kernel/kprobes.c 	if (cur && cur->fault_handler) {
cur              1157 kernel/kprobes.c 		if (cur->fault_handler(cur, regs, trapnr))
cur               446 kernel/locking/mutex.c 	struct mutex_waiter *cur;
cur               450 kernel/locking/mutex.c 	list_for_each_entry(cur, &lock->wait_list, list) {
cur               451 kernel/locking/mutex.c 		if (!cur->ww_ctx)
cur               454 kernel/locking/mutex.c 		if (__ww_mutex_die(lock, cur, ww_ctx) ||
cur               455 kernel/locking/mutex.c 		    __ww_mutex_wound(lock, cur->ww_ctx, ww_ctx))
cur               810 kernel/locking/mutex.c 	struct mutex_waiter *cur;
cur               829 kernel/locking/mutex.c 	cur = waiter;
cur               830 kernel/locking/mutex.c 	list_for_each_entry_continue_reverse(cur, &lock->wait_list, list) {
cur               831 kernel/locking/mutex.c 		if (!cur->ww_ctx)
cur               856 kernel/locking/mutex.c 	struct mutex_waiter *cur;
cur               875 kernel/locking/mutex.c 	list_for_each_entry_reverse(cur, &lock->wait_list, list) {
cur               876 kernel/locking/mutex.c 		if (!cur->ww_ctx)
cur               879 kernel/locking/mutex.c 		if (__ww_ctx_stamp_after(ww_ctx, cur->ww_ctx)) {
cur               895 kernel/locking/mutex.c 		pos = &cur->list;
cur               898 kernel/locking/mutex.c 		__ww_mutex_die(lock, cur, ww_ctx);
cur               331 kernel/padata.c 	struct padata_priv *cur;
cur               335 kernel/padata.c 	list_for_each_entry_reverse(cur, &pqueue->reorder.list, list)
cur               336 kernel/padata.c 		if (cur->seq_nr < padata->seq_nr)
cur               338 kernel/padata.c 	list_add(&padata->list, &cur->list);
cur               132 kernel/power/power.h 	unsigned int	cur;	/* number of the block of PAGE_SIZE bytes the
cur               382 kernel/power/snapshot.c 	struct bm_position cur;	/* most recently used bit position */
cur               550 kernel/power/snapshot.c 	bm->cur.zone = list_entry(bm->zones.next, struct mem_zone_bm_rtree,
cur               552 kernel/power/snapshot.c 	bm->cur.node = list_entry(bm->cur.zone->leaves.next,
cur               554 kernel/power/snapshot.c 	bm->cur.node_pfn = 0;
cur               555 kernel/power/snapshot.c 	bm->cur.node_bit = 0;
cur               595 kernel/power/snapshot.c 		struct mem_extent *ext, *cur, *aux;
cur               626 kernel/power/snapshot.c 		cur = ext;
cur               627 kernel/power/snapshot.c 		list_for_each_entry_safe_continue(cur, aux, list, hook) {
cur               628 kernel/power/snapshot.c 			if (zone_end < cur->start)
cur               630 kernel/power/snapshot.c 			if (zone_end < cur->end)
cur               631 kernel/power/snapshot.c 				ext->end = cur->end;
cur               632 kernel/power/snapshot.c 			list_del(&cur->hook);
cur               633 kernel/power/snapshot.c 			kfree(cur);
cur               714 kernel/power/snapshot.c 	zone = bm->cur.zone;
cur               743 kernel/power/snapshot.c 	node = bm->cur.node;
cur               744 kernel/power/snapshot.c 	if (zone == bm->cur.zone &&
cur               745 kernel/power/snapshot.c 	    ((pfn - zone->start_pfn) & ~BM_BLOCK_MASK) == bm->cur.node_pfn)
cur               762 kernel/power/snapshot.c 	bm->cur.zone = zone;
cur               763 kernel/power/snapshot.c 	bm->cur.node = node;
cur               764 kernel/power/snapshot.c 	bm->cur.node_pfn = (pfn - zone->start_pfn) & ~BM_BLOCK_MASK;
cur               812 kernel/power/snapshot.c 	bit = max(bm->cur.node_bit - 1, 0);
cur               813 kernel/power/snapshot.c 	clear_bit(bit, bm->cur.node->data);
cur               847 kernel/power/snapshot.c 	if (!list_is_last(&bm->cur.node->list, &bm->cur.zone->leaves)) {
cur               848 kernel/power/snapshot.c 		bm->cur.node = list_entry(bm->cur.node->list.next,
cur               850 kernel/power/snapshot.c 		bm->cur.node_pfn += BM_BITS_PER_BLOCK;
cur               851 kernel/power/snapshot.c 		bm->cur.node_bit  = 0;
cur               857 kernel/power/snapshot.c 	if (!list_is_last(&bm->cur.zone->list, &bm->zones)) {
cur               858 kernel/power/snapshot.c 		bm->cur.zone = list_entry(bm->cur.zone->list.next,
cur               860 kernel/power/snapshot.c 		bm->cur.node = list_entry(bm->cur.zone->leaves.next,
cur               862 kernel/power/snapshot.c 		bm->cur.node_pfn = 0;
cur               863 kernel/power/snapshot.c 		bm->cur.node_bit = 0;
cur               888 kernel/power/snapshot.c 		pages	  = bm->cur.zone->end_pfn - bm->cur.zone->start_pfn;
cur               889 kernel/power/snapshot.c 		bits      = min(pages - bm->cur.node_pfn, BM_BITS_PER_BLOCK);
cur               890 kernel/power/snapshot.c 		bit	  = find_next_bit(bm->cur.node->data, bits,
cur               891 kernel/power/snapshot.c 					  bm->cur.node_bit);
cur               893 kernel/power/snapshot.c 			pfn = bm->cur.zone->start_pfn + bm->cur.node_pfn + bit;
cur               894 kernel/power/snapshot.c 			bm->cur.node_bit = bit + 1;
cur              2097 kernel/power/snapshot.c 	if (handle->cur > nr_meta_pages + nr_copy_pages)
cur              2106 kernel/power/snapshot.c 	if (!handle->cur) {
cur              2115 kernel/power/snapshot.c 	} else if (handle->cur <= nr_meta_pages) {
cur              2138 kernel/power/snapshot.c 	handle->cur++;
cur              2597 kernel/power/snapshot.c 	if (handle->cur > 1 && handle->cur > nr_meta_pages + nr_copy_pages)
cur              2602 kernel/power/snapshot.c 	if (!handle->cur) {
cur              2611 kernel/power/snapshot.c 	} else if (handle->cur == 1) {
cur              2628 kernel/power/snapshot.c 	} else if (handle->cur <= nr_meta_pages + 1) {
cur              2633 kernel/power/snapshot.c 		if (handle->cur == nr_meta_pages + 1) {
cur              2657 kernel/power/snapshot.c 	handle->cur++;
cur              2677 kernel/power/snapshot.c 	if (handle->cur > 1 && handle->cur > nr_meta_pages + nr_copy_pages) {
cur              2686 kernel/power/snapshot.c 			handle->cur <= nr_meta_pages + nr_copy_pages);
cur                96 kernel/power/swap.c 	struct swap_map_page *cur;
cur               404 kernel/power/swap.c 	if (handle->cur)
cur               405 kernel/power/swap.c 		free_page((unsigned long)handle->cur);
cur               406 kernel/power/swap.c 	handle->cur = NULL;
cur               419 kernel/power/swap.c 	handle->cur = (struct swap_map_page *)get_zeroed_page(GFP_KERNEL);
cur               420 kernel/power/swap.c 	if (!handle->cur) {
cur               446 kernel/power/swap.c 	if (!handle->cur)
cur               452 kernel/power/swap.c 	handle->cur->entries[handle->k++] = offset;
cur               457 kernel/power/swap.c 		handle->cur->next_swap = offset;
cur               458 kernel/power/swap.c 		error = write_page(handle->cur, handle->cur_swap, hb);
cur               461 kernel/power/swap.c 		clear_page(handle->cur);
cur               482 kernel/power/swap.c 	if (handle->cur && handle->cur_swap)
cur               483 kernel/power/swap.c 		return write_page(handle->cur, handle->cur_swap, NULL);
cur               958 kernel/power/swap.c 	handle->cur = NULL;
cur               973 kernel/power/swap.c 	handle->cur = NULL;
cur              1003 kernel/power/swap.c 	handle->cur = handle->maps->map;
cur              1014 kernel/power/swap.c 	if (!handle->cur)
cur              1016 kernel/power/swap.c 	offset = handle->cur->entries[handle->k];
cur              1031 kernel/power/swap.c 			handle->cur = handle->maps->map;
cur              1299 kernel/power/swap.c 				if (handle->cur &&
cur              1300 kernel/power/swap.c 				    handle->cur->entries[handle->k]) {
cur              1834 kernel/sched/core.c int migrate_swap(struct task_struct *cur, struct task_struct *p,
cur              1841 kernel/sched/core.c 		.src_task = cur,
cur              1863 kernel/sched/core.c 	trace_sched_swap_numa(cur, arg.src_cpu, p, arg.dst_cpu);
cur              6080 kernel/sched/core.c int cpuset_cpumask_can_shrink(const struct cpumask *cur,
cur              6085 kernel/sched/core.c 	if (!cpumask_weight(cur))
cur              6088 kernel/sched/core.c 	ret = dl_cpuset_cpumask_can_shrink(cur, trial);
cur               127 kernel/sched/cpufreq_schedutil.c 	policy->cur = next_freq;
cur               174 kernel/sched/cpufreq_schedutil.c 				policy->cpuinfo.max_freq : policy->cur;
cur              2749 kernel/sched/deadline.c int dl_cpuset_cpumask_can_shrink(const struct cpumask *cur,
cur              2757 kernel/sched/deadline.c 	cur_dl_b = dl_bw_of(cpumask_any(cur));
cur              1595 kernel/sched/fair.c 	struct task_struct *cur;
cur              1605 kernel/sched/fair.c 	cur = rcu_dereference(dst_rq->curr);
cur              1606 kernel/sched/fair.c 	if (cur && ((cur->flags & PF_EXITING) || is_idle_task(cur)))
cur              1607 kernel/sched/fair.c 		cur = NULL;
cur              1613 kernel/sched/fair.c 	if (cur == env->p)
cur              1616 kernel/sched/fair.c 	if (!cur) {
cur              1631 kernel/sched/fair.c 	if (!cpumask_test_cpu(env->src_cpu, cur->cpus_ptr))
cur              1638 kernel/sched/fair.c 	cur_ng = rcu_dereference(cur->numa_group);
cur              1640 kernel/sched/fair.c 		imp = taskimp + task_weight(cur, env->src_nid, dist) -
cur              1641 kernel/sched/fair.c 		      task_weight(cur, env->dst_nid, dist);
cur              1654 kernel/sched/fair.c 			imp += group_weight(cur, env->src_nid, dist) -
cur              1655 kernel/sched/fair.c 			       group_weight(cur, env->dst_nid, dist);
cur              1657 kernel/sched/fair.c 			imp += task_weight(cur, env->src_nid, dist) -
cur              1658 kernel/sched/fair.c 			       task_weight(cur, env->dst_nid, dist);
cur              1663 kernel/sched/fair.c 		cur = NULL;
cur              1679 kernel/sched/fair.c 	load = task_h_load(env->p) - task_h_load(cur);
cur              1694 kernel/sched/fair.c 	if (!cur) {
cur              1705 kernel/sched/fair.c 	task_numa_assign(env, cur, imp);
cur               324 kernel/sched/sched.h extern int  dl_cpuset_cpumask_can_shrink(const struct cpumask *cur, const struct cpumask *trial);
cur              2179 kernel/sched/topology.c static int dattrs_equal(struct sched_domain_attr *cur, int idx_cur,
cur              2185 kernel/sched/topology.c 	if (!new && !cur)
cur              2190 kernel/sched/topology.c 	return !memcmp(cur ? (cur + idx_cur) : &tmp,
cur              1014 kernel/seccomp.c 	struct seccomp_knotif *knotif = NULL, *cur;
cur              1032 kernel/seccomp.c 	list_for_each_entry(cur, &filter->notif->notifications, list) {
cur              1033 kernel/seccomp.c 		if (cur->state == SECCOMP_NOTIFY_INIT) {
cur              1034 kernel/seccomp.c 			knotif = cur;
cur              1070 kernel/seccomp.c 		list_for_each_entry(cur, &filter->notif->notifications, list) {
cur              1071 kernel/seccomp.c 			if (cur->id == unotif.id) {
cur              1072 kernel/seccomp.c 				knotif = cur;
cur              1091 kernel/seccomp.c 	struct seccomp_knotif *knotif = NULL, *cur;
cur              1104 kernel/seccomp.c 	list_for_each_entry(cur, &filter->notif->notifications, list) {
cur              1105 kernel/seccomp.c 		if (cur->id == resp.id) {
cur              1106 kernel/seccomp.c 			knotif = cur;
cur              1183 kernel/seccomp.c 	struct seccomp_knotif *cur;
cur              1190 kernel/seccomp.c 	list_for_each_entry(cur, &filter->notif->notifications, list) {
cur              1191 kernel/seccomp.c 		if (cur->state == SECCOMP_NOTIFY_INIT)
cur              1193 kernel/seccomp.c 		if (cur->state == SECCOMP_NOTIFY_SENT)
cur              1214 kernel/seccomp.c 	struct seccomp_filter *cur;
cur              1216 kernel/seccomp.c 	for (cur = current->seccomp.filter; cur; cur = cur->prev) {
cur              1217 kernel/seccomp.c 		if (cur->notif)
cur              1617 kernel/seccomp.c 	const struct seccomp_log_name *cur;
cur              1620 kernel/seccomp.c 	for (cur = seccomp_log_names; cur->name && size; cur++) {
cur              1623 kernel/seccomp.c 		if (!(actions_logged & cur->log))
cur              1636 kernel/seccomp.c 		ret = strscpy(names, cur->name, size);
cur              1650 kernel/seccomp.c 	const struct seccomp_log_name *cur;
cur              1652 kernel/seccomp.c 	for (cur = seccomp_log_names; cur->name; cur++) {
cur              1653 kernel/seccomp.c 		if (!strcmp(cur->name, name)) {
cur              1654 kernel/seccomp.c 			*action_logged = cur->log;
cur               215 kernel/smpboot.c 	struct smp_hotplug_thread *cur;
cur               219 kernel/smpboot.c 	list_for_each_entry(cur, &hotplug_threads, list) {
cur               220 kernel/smpboot.c 		ret = __smpboot_create_thread(cur, cpu);
cur               238 kernel/smpboot.c 	struct smp_hotplug_thread *cur;
cur               241 kernel/smpboot.c 	list_for_each_entry(cur, &hotplug_threads, list)
cur               242 kernel/smpboot.c 		smpboot_unpark_thread(cur, cpu);
cur               257 kernel/smpboot.c 	struct smp_hotplug_thread *cur;
cur               260 kernel/smpboot.c 	list_for_each_entry_reverse(cur, &hotplug_threads, list)
cur               261 kernel/smpboot.c 		smpboot_park_thread(cur, cpu);
cur                96 kernel/time/tick-broadcast.c 	struct clock_event_device *cur = tick_broadcast_device.evtdev;
cur                98 kernel/time/tick-broadcast.c 	if (!tick_check_broadcast_device(cur, dev))
cur               104 kernel/time/tick-broadcast.c 	clockevents_exchange_device(cur, dev);
cur               105 kernel/time/tick-broadcast.c 	if (cur)
cur               106 kernel/time/tick-broadcast.c 		cur->event_handler = clockevents_handle_noop;
cur                52 lib/cmdline.c  	char *cur = *str;
cur                54 lib/cmdline.c  	if (!cur || !(*cur))
cur                56 lib/cmdline.c  	*pint = simple_strtol(cur, str, 0);
cur                57 lib/cmdline.c  	if (cur == *str)
cur                60 lib/errseq.c   	errseq_t cur, old;
cur                89 lib/errseq.c   			cur = new;
cur                94 lib/errseq.c   		cur = cmpxchg(eseq, old, new);
cur               100 lib/errseq.c   		if (likely(cur == old || cur == new))
cur               104 lib/errseq.c   		old = cur;
cur               106 lib/errseq.c   	return cur;
cur               146 lib/errseq.c   	errseq_t cur = READ_ONCE(*eseq);
cur               148 lib/errseq.c   	if (likely(cur == since))
cur               150 lib/errseq.c   	return -(cur & MAX_ERRNO);
cur               153 lib/kobject.c  		int cur = strlen(kobject_name(parent));
cur               155 lib/kobject.c  		length -= cur;
cur               156 lib/kobject.c  		memcpy(path + length, kobject_name(parent), cur);
cur               173 lib/rbtree_test.c 	struct test_node *cur, *n;
cur               175 lib/rbtree_test.c 	rbtree_postorder_for_each_entry_safe(cur, n, &root.rb_root, rb)
cur                74 lib/test_list_sort.c 	struct list_head *cur;
cur               100 lib/test_list_sort.c 	for (cur = head.next; cur->next != &head; cur = cur->next) {
cur               104 lib/test_list_sort.c 		if (cur->next->prev != cur) {
cur               109 lib/test_list_sort.c 		cmp_result = cmp(NULL, cur, cur->next);
cur               115 lib/test_list_sort.c 		el = container_of(cur, struct debug_el, list);
cur               116 lib/test_list_sort.c 		el1 = container_of(cur->next, struct debug_el, list);
cur               129 lib/test_list_sort.c 	if (head.prev != cur) {
cur               137 lib/ts_fsm.c   	struct ts_fsm_token *cur = NULL, *next;
cur               166 lib/ts_fsm.c   		cur = &fsm->tokens[tok_idx];
cur               173 lib/ts_fsm.c   		switch (cur->recur) {
cur               178 lib/ts_fsm.c   			if (!match_token(cur, data[block_idx]))
cur               184 lib/ts_fsm.c   			    !match_token(cur, data[block_idx]))
cur               192 lib/ts_fsm.c   			if (!match_token(cur, data[block_idx]))
cur               206 lib/ts_fsm.c   				if (!match_token(cur, data[block_idx]))
cur               229 lib/ts_fsm.c   				if (!match_token(cur, data[block_idx]))
cur              1171 lib/vsprintf.c 	int cur, rbot, rtop;
cur              1177 lib/vsprintf.c 	rbot = cur = find_first_bit(bitmap, nr_bits);
cur              1178 lib/vsprintf.c 	while (cur < nr_bits) {
cur              1179 lib/vsprintf.c 		rtop = cur;
cur              1180 lib/vsprintf.c 		cur = find_next_bit(bitmap, nr_bits, cur + 1);
cur              1181 lib/vsprintf.c 		if (cur < nr_bits && cur <= rtop + 1)
cur              1200 lib/vsprintf.c 		rbot = cur;
cur               440 lib/zstd/zstd_opt.h 		U32 cur, match_num, last_pos, litlen, price;
cur               457 lib/zstd/zstd_opt.h 						cur = 0;
cur               482 lib/zstd/zstd_opt.h 			cur = 0;
cur               515 lib/zstd/zstd_opt.h 		for (cur = 1; cur <= last_pos; cur++) {
cur               516 lib/zstd/zstd_opt.h 			inr = ip + cur;
cur               518 lib/zstd/zstd_opt.h 			if (opt[cur - 1].mlen == 1) {
cur               519 lib/zstd/zstd_opt.h 				litlen = opt[cur - 1].litlen + 1;
cur               520 lib/zstd/zstd_opt.h 				if (cur > litlen) {
cur               521 lib/zstd/zstd_opt.h 					price = opt[cur - litlen].price + ZSTD_getLiteralPrice(seqStorePtr, litlen, inr - litlen);
cur               526 lib/zstd/zstd_opt.h 				price = opt[cur - 1].price + ZSTD_getLiteralPrice(seqStorePtr, litlen, inr - 1);
cur               529 lib/zstd/zstd_opt.h 			if (cur > last_pos || price <= opt[cur].price)
cur               530 lib/zstd/zstd_opt.h 				SET_PRICE(cur, 1, 0, litlen, price);
cur               532 lib/zstd/zstd_opt.h 			if (cur == last_pos)
cur               538 lib/zstd/zstd_opt.h 			mlen = opt[cur].mlen;
cur               539 lib/zstd/zstd_opt.h 			if (opt[cur].off > ZSTD_REP_MOVE_OPT) {
cur               540 lib/zstd/zstd_opt.h 				opt[cur].rep[2] = opt[cur - mlen].rep[1];
cur               541 lib/zstd/zstd_opt.h 				opt[cur].rep[1] = opt[cur - mlen].rep[0];
cur               542 lib/zstd/zstd_opt.h 				opt[cur].rep[0] = opt[cur].off - ZSTD_REP_MOVE_OPT;
cur               544 lib/zstd/zstd_opt.h 				opt[cur].rep[2] = (opt[cur].off > 1) ? opt[cur - mlen].rep[1] : opt[cur - mlen].rep[2];
cur               545 lib/zstd/zstd_opt.h 				opt[cur].rep[1] = (opt[cur].off > 0) ? opt[cur - mlen].rep[0] : opt[cur - mlen].rep[1];
cur               546 lib/zstd/zstd_opt.h 				opt[cur].rep[0] =
cur               547 lib/zstd/zstd_opt.h 				    ((opt[cur].off == ZSTD_REP_MOVE_OPT) && (mlen != 1)) ? (opt[cur - mlen].rep[0] - 1) : (opt[cur - mlen].rep[opt[cur].off]);
cur               553 lib/zstd/zstd_opt.h 				for (i = (opt[cur].mlen != 1); i < last_i; i++) { /* check rep */
cur               554 lib/zstd/zstd_opt.h 					const S32 repCur = (i == ZSTD_REP_MOVE_OPT) ? (opt[cur].rep[0] - 1) : opt[cur].rep[i];
cur               559 lib/zstd/zstd_opt.h 						if (mlen > sufficient_len || cur + mlen >= ZSTD_OPT_NUM) {
cur               562 lib/zstd/zstd_opt.h 							last_pos = cur + 1;
cur               566 lib/zstd/zstd_opt.h 						best_off = i - (opt[cur].mlen != 1);
cur               571 lib/zstd/zstd_opt.h 							if (opt[cur].mlen == 1) {
cur               572 lib/zstd/zstd_opt.h 								litlen = opt[cur].litlen;
cur               573 lib/zstd/zstd_opt.h 								if (cur > litlen) {
cur               574 lib/zstd/zstd_opt.h 									price = opt[cur - litlen].price + ZSTD_getPrice(seqStorePtr, litlen, inr - litlen,
cur               580 lib/zstd/zstd_opt.h 								price = opt[cur].price + ZSTD_getPrice(seqStorePtr, 0, NULL, best_off, mlen - MINMATCH, ultra);
cur               583 lib/zstd/zstd_opt.h 							if (cur + mlen > last_pos || price <= opt[cur + mlen].price)
cur               584 lib/zstd/zstd_opt.h 								SET_PRICE(cur + mlen, mlen, i, litlen, price);
cur               593 lib/zstd/zstd_opt.h 			if (match_num > 0 && (matches[match_num - 1].len > sufficient_len || cur + matches[match_num - 1].len >= ZSTD_OPT_NUM)) {
cur               596 lib/zstd/zstd_opt.h 				last_pos = cur + 1;
cur               606 lib/zstd/zstd_opt.h 					if (opt[cur].mlen == 1) {
cur               607 lib/zstd/zstd_opt.h 						litlen = opt[cur].litlen;
cur               608 lib/zstd/zstd_opt.h 						if (cur > litlen)
cur               609 lib/zstd/zstd_opt.h 							price = opt[cur - litlen].price + ZSTD_getPrice(seqStorePtr, litlen, ip + cur - litlen,
cur               615 lib/zstd/zstd_opt.h 						price = opt[cur].price + ZSTD_getPrice(seqStorePtr, 0, NULL, matches[u].off - 1, mlen - MINMATCH, ultra);
cur               618 lib/zstd/zstd_opt.h 					if (cur + mlen > last_pos || (price < opt[cur + mlen].price))
cur               619 lib/zstd/zstd_opt.h 						SET_PRICE(cur + mlen, mlen, matches[u].off, litlen, price);
cur               628 lib/zstd/zstd_opt.h 		cur = last_pos - best_mlen;
cur               635 lib/zstd/zstd_opt.h 			mlen = opt[cur].mlen;
cur               636 lib/zstd/zstd_opt.h 			offset = opt[cur].off;
cur               637 lib/zstd/zstd_opt.h 			opt[cur].mlen = best_mlen;
cur               638 lib/zstd/zstd_opt.h 			opt[cur].off = best_off;
cur               641 lib/zstd/zstd_opt.h 			if (mlen > cur)
cur               643 lib/zstd/zstd_opt.h 			cur -= mlen;
cur               650 lib/zstd/zstd_opt.h 		for (cur = 0; cur < last_pos;) {
cur               651 lib/zstd/zstd_opt.h 			mlen = opt[cur].mlen;
cur               654 lib/zstd/zstd_opt.h 				cur++;
cur               657 lib/zstd/zstd_opt.h 			offset = opt[cur].off;
cur               658 lib/zstd/zstd_opt.h 			cur += mlen;
cur               738 lib/zstd/zstd_opt.h 		U32 cur, match_num, last_pos, litlen, price;
cur               763 lib/zstd/zstd_opt.h 						cur = 0;
cur               797 lib/zstd/zstd_opt.h 			cur = 0;
cur               823 lib/zstd/zstd_opt.h 		for (cur = 1; cur <= last_pos; cur++) {
cur               824 lib/zstd/zstd_opt.h 			inr = ip + cur;
cur               826 lib/zstd/zstd_opt.h 			if (opt[cur - 1].mlen == 1) {
cur               827 lib/zstd/zstd_opt.h 				litlen = opt[cur - 1].litlen + 1;
cur               828 lib/zstd/zstd_opt.h 				if (cur > litlen) {
cur               829 lib/zstd/zstd_opt.h 					price = opt[cur - litlen].price + ZSTD_getLiteralPrice(seqStorePtr, litlen, inr - litlen);
cur               834 lib/zstd/zstd_opt.h 				price = opt[cur - 1].price + ZSTD_getLiteralPrice(seqStorePtr, litlen, inr - 1);
cur               837 lib/zstd/zstd_opt.h 			if (cur > last_pos || price <= opt[cur].price)
cur               838 lib/zstd/zstd_opt.h 				SET_PRICE(cur, 1, 0, litlen, price);
cur               840 lib/zstd/zstd_opt.h 			if (cur == last_pos)
cur               846 lib/zstd/zstd_opt.h 			mlen = opt[cur].mlen;
cur               847 lib/zstd/zstd_opt.h 			if (opt[cur].off > ZSTD_REP_MOVE_OPT) {
cur               848 lib/zstd/zstd_opt.h 				opt[cur].rep[2] = opt[cur - mlen].rep[1];
cur               849 lib/zstd/zstd_opt.h 				opt[cur].rep[1] = opt[cur - mlen].rep[0];
cur               850 lib/zstd/zstd_opt.h 				opt[cur].rep[0] = opt[cur].off - ZSTD_REP_MOVE_OPT;
cur               852 lib/zstd/zstd_opt.h 				opt[cur].rep[2] = (opt[cur].off > 1) ? opt[cur - mlen].rep[1] : opt[cur - mlen].rep[2];
cur               853 lib/zstd/zstd_opt.h 				opt[cur].rep[1] = (opt[cur].off > 0) ? opt[cur - mlen].rep[0] : opt[cur - mlen].rep[1];
cur               854 lib/zstd/zstd_opt.h 				opt[cur].rep[0] =
cur               855 lib/zstd/zstd_opt.h 				    ((opt[cur].off == ZSTD_REP_MOVE_OPT) && (mlen != 1)) ? (opt[cur - mlen].rep[0] - 1) : (opt[cur - mlen].rep[opt[cur].off]);
cur               862 lib/zstd/zstd_opt.h 					const S32 repCur = (i == ZSTD_REP_MOVE_OPT) ? (opt[cur].rep[0] - 1) : opt[cur].rep[i];
cur               863 lib/zstd/zstd_opt.h 					const U32 repIndex = (U32)(curr + cur - repCur);
cur               866 lib/zstd/zstd_opt.h 					if ((repCur > 0 && repCur <= (S32)(curr + cur)) &&
cur               873 lib/zstd/zstd_opt.h 						if (mlen > sufficient_len || cur + mlen >= ZSTD_OPT_NUM) {
cur               876 lib/zstd/zstd_opt.h 							last_pos = cur + 1;
cur               880 lib/zstd/zstd_opt.h 						best_off = i - (opt[cur].mlen != 1);
cur               885 lib/zstd/zstd_opt.h 							if (opt[cur].mlen == 1) {
cur               886 lib/zstd/zstd_opt.h 								litlen = opt[cur].litlen;
cur               887 lib/zstd/zstd_opt.h 								if (cur > litlen) {
cur               888 lib/zstd/zstd_opt.h 									price = opt[cur - litlen].price + ZSTD_getPrice(seqStorePtr, litlen, inr - litlen,
cur               894 lib/zstd/zstd_opt.h 								price = opt[cur].price + ZSTD_getPrice(seqStorePtr, 0, NULL, best_off, mlen - MINMATCH, ultra);
cur               897 lib/zstd/zstd_opt.h 							if (cur + mlen > last_pos || price <= opt[cur + mlen].price)
cur               898 lib/zstd/zstd_opt.h 								SET_PRICE(cur + mlen, mlen, i, litlen, price);
cur               907 lib/zstd/zstd_opt.h 			if (match_num > 0 && (matches[match_num - 1].len > sufficient_len || cur + matches[match_num - 1].len >= ZSTD_OPT_NUM)) {
cur               910 lib/zstd/zstd_opt.h 				last_pos = cur + 1;
cur               920 lib/zstd/zstd_opt.h 					if (opt[cur].mlen == 1) {
cur               921 lib/zstd/zstd_opt.h 						litlen = opt[cur].litlen;
cur               922 lib/zstd/zstd_opt.h 						if (cur > litlen)
cur               923 lib/zstd/zstd_opt.h 							price = opt[cur - litlen].price + ZSTD_getPrice(seqStorePtr, litlen, ip + cur - litlen,
cur               929 lib/zstd/zstd_opt.h 						price = opt[cur].price + ZSTD_getPrice(seqStorePtr, 0, NULL, matches[u].off - 1, mlen - MINMATCH, ultra);
cur               932 lib/zstd/zstd_opt.h 					if (cur + mlen > last_pos || (price < opt[cur + mlen].price))
cur               933 lib/zstd/zstd_opt.h 						SET_PRICE(cur + mlen, mlen, matches[u].off, litlen, price);
cur               942 lib/zstd/zstd_opt.h 		cur = last_pos - best_mlen;
cur               949 lib/zstd/zstd_opt.h 			mlen = opt[cur].mlen;
cur               950 lib/zstd/zstd_opt.h 			offset = opt[cur].off;
cur               951 lib/zstd/zstd_opt.h 			opt[cur].mlen = best_mlen;
cur               952 lib/zstd/zstd_opt.h 			opt[cur].off = best_off;
cur               955 lib/zstd/zstd_opt.h 			if (mlen > cur)
cur               957 lib/zstd/zstd_opt.h 			cur -= mlen;
cur               964 lib/zstd/zstd_opt.h 		for (cur = 0; cur < last_pos;) {
cur               965 lib/zstd/zstd_opt.h 			mlen = opt[cur].mlen;
cur               968 lib/zstd/zstd_opt.h 				cur++;
cur               971 lib/zstd/zstd_opt.h 			offset = opt[cur].off;
cur               972 lib/zstd/zstd_opt.h 			cur += mlen;
cur               278 mm/readahead.c 	unsigned long cur = ra->size;
cur               280 mm/readahead.c 	if (cur < max / 16)
cur               281 mm/readahead.c 		return 4 * cur;
cur               282 mm/readahead.c 	if (cur <= max / 2)
cur               283 mm/readahead.c 		return 2 * cur;
cur               240 mm/slob.c      	slob_t *prev, *cur, *aligned = NULL;
cur               244 mm/slob.c      	for (prev = NULL, cur = sp->freelist; ; prev = cur, cur = slob_next(cur)) {
cur               245 mm/slob.c      		slobidx_t avail = slob_units(cur);
cur               256 mm/slob.c      				(ALIGN((unsigned long)cur + align_offset, align)
cur               258 mm/slob.c      			delta = aligned - cur;
cur               264 mm/slob.c      				next = slob_next(cur);
cur               266 mm/slob.c      				set_slob(cur, delta, aligned);
cur               267 mm/slob.c      				prev = cur;
cur               268 mm/slob.c      				cur = aligned;
cur               269 mm/slob.c      				avail = slob_units(cur);
cur               272 mm/slob.c      			next = slob_next(cur);
cur               280 mm/slob.c      					set_slob(prev, slob_units(prev), cur + units);
cur               282 mm/slob.c      					sp->freelist = cur + units;
cur               283 mm/slob.c      				set_slob(cur + units, avail - units, next);
cur               291 mm/slob.c      			return cur;
cur               293 mm/slob.c      		if (slob_last(cur))
cur              1574 mm/slub.c      	void *cur;
cur              1588 mm/slub.c      	cur = next_freelist_entry(s, page, &pos, start, page_limit,
cur              1590 mm/slub.c      	cur = setup_object(s, page, cur);
cur              1591 mm/slub.c      	page->freelist = cur;
cur              1597 mm/slub.c      		set_freepointer(s, cur, next);
cur              1598 mm/slub.c      		cur = next;
cur              1600 mm/slub.c      	set_freepointer(s, cur, NULL);
cur               148 mm/swap_slots.c 	cache->cur = 0;
cur               179 mm/swap_slots.c 		swapcache_free_entries(cache->slots + cache->cur, cache->nr);
cur               180 mm/swap_slots.c 		cache->cur = 0;
cur               270 mm/swap_slots.c 	cache->cur = 0;
cur               339 mm/swap_slots.c 				pentry = &cache->slots[cache->cur++];
cur               274 net/bridge/br.c 	bool cur = !!br_opt_get(br, opt);
cur               277 net/bridge/br.c 		 opt, cur, on);
cur               279 net/bridge/br.c 	if (cur == on)
cur              1844 net/ceph/osd_client.c 		struct ceph_osd_backoff *cur =
cur              1848 net/ceph/osd_client.c 		cmp = hoid_compare(hoid, cur->begin);
cur              1852 net/ceph/osd_client.c 			if (hoid_compare(hoid, cur->end) < 0)
cur              1853 net/ceph/osd_client.c 				return cur;
cur              1857 net/ceph/osd_client.c 			return cur;
cur              6622 net/core/dev.c 	int ret, cur = 0;
cur              6645 net/core/dev.c 			dev_stack[cur] = now;
cur              6646 net/core/dev.c 			iter_stack[cur++] = iter;
cur              6651 net/core/dev.c 			if (!cur)
cur              6653 net/core/dev.c 			next = dev_stack[--cur];
cur              6654 net/core/dev.c 			niter = iter_stack[cur];
cur              6671 net/core/dev.c 	int ret, cur = 0;
cur              6691 net/core/dev.c 			dev_stack[cur] = now;
cur              6692 net/core/dev.c 			iter_stack[cur++] = iter;
cur              6697 net/core/dev.c 			if (!cur)
cur              6699 net/core/dev.c 			next = dev_stack[--cur];
cur              6700 net/core/dev.c 			niter = iter_stack[cur];
cur              6840 net/core/dev.c 	int ret, cur = 0;
cur              6860 net/core/dev.c 			dev_stack[cur] = now;
cur              6861 net/core/dev.c 			iter_stack[cur++] = iter;
cur              6866 net/core/dev.c 			if (!cur)
cur              6868 net/core/dev.c 			next = dev_stack[--cur];
cur              6869 net/core/dev.c 			niter = iter_stack[cur];
cur              6887 net/core/dev.c 	int ret, cur = 0;
cur              6910 net/core/dev.c 			dev_stack[cur] = now;
cur              6911 net/core/dev.c 			iter_stack[cur++] = iter;
cur              6916 net/core/dev.c 			if (!cur)
cur              6918 net/core/dev.c 			next = dev_stack[--cur];
cur              6919 net/core/dev.c 			niter = iter_stack[cur];
cur              7003 net/core/dev.c 	int ret, cur = 0;
cur              7023 net/core/dev.c 			dev_stack[cur] = now;
cur              7024 net/core/dev.c 			iter_stack[cur++] = iter;
cur              7029 net/core/dev.c 			if (!cur)
cur              7031 net/core/dev.c 			next = dev_stack[--cur];
cur              7032 net/core/dev.c 			niter = iter_stack[cur];
cur              1141 net/core/devlink.c 		u32 cur;
cur              1145 net/core/devlink.c 						pool_index, &cur, &max);
cur              1149 net/core/devlink.c 			if (nla_put_u32(msg, DEVLINK_ATTR_SB_OCC_CUR, cur))
cur              1339 net/core/devlink.c 		u32 cur;
cur              1345 net/core/devlink.c 						   &cur, &max);
cur              1349 net/core/devlink.c 			if (nla_put_u32(msg, DEVLINK_ATTR_SB_OCC_CUR, cur))
cur               504 net/core/netpoll.c 	char *cur=opt, *delim;
cur               508 net/core/netpoll.c 	if (*cur != '@') {
cur               509 net/core/netpoll.c 		if ((delim = strchr(cur, '@')) == NULL)
cur               512 net/core/netpoll.c 		if (kstrtou16(cur, 10, &np->local_port))
cur               514 net/core/netpoll.c 		cur = delim;
cur               516 net/core/netpoll.c 	cur++;
cur               518 net/core/netpoll.c 	if (*cur != '/') {
cur               520 net/core/netpoll.c 		if ((delim = strchr(cur, '/')) == NULL)
cur               523 net/core/netpoll.c 		ipv6 = netpoll_parse_ip_addr(cur, &np->local_ip);
cur               528 net/core/netpoll.c 		cur = delim;
cur               530 net/core/netpoll.c 	cur++;
cur               532 net/core/netpoll.c 	if (*cur != ',') {
cur               534 net/core/netpoll.c 		if ((delim = strchr(cur, ',')) == NULL)
cur               537 net/core/netpoll.c 		strlcpy(np->dev_name, cur, sizeof(np->dev_name));
cur               538 net/core/netpoll.c 		cur = delim;
cur               540 net/core/netpoll.c 	cur++;
cur               542 net/core/netpoll.c 	if (*cur != '@') {
cur               544 net/core/netpoll.c 		if ((delim = strchr(cur, '@')) == NULL)
cur               547 net/core/netpoll.c 		if (*cur == ' ' || *cur == '\t')
cur               549 net/core/netpoll.c 		if (kstrtou16(cur, 10, &np->remote_port))
cur               551 net/core/netpoll.c 		cur = delim;
cur               553 net/core/netpoll.c 	cur++;
cur               556 net/core/netpoll.c 	if ((delim = strchr(cur, '/')) == NULL)
cur               559 net/core/netpoll.c 	ipv6 = netpoll_parse_ip_addr(cur, &np->remote_ip);
cur               566 net/core/netpoll.c 	cur = delim + 1;
cur               568 net/core/netpoll.c 	if (*cur != 0) {
cur               570 net/core/netpoll.c 		if (!mac_pton(cur, np->remote_mac))
cur               579 net/core/netpoll.c 	np_info(np, "couldn't parse config at '%s'!\n", cur);
cur              3232 net/core/pktgen.c 	struct pktgen_dev *cur;
cur              3237 net/core/pktgen.c 		cur = list_entry(q, struct pktgen_dev, list);
cur              3239 net/core/pktgen.c 		if (!cur->removal_mark)
cur              3242 net/core/pktgen.c 		kfree_skb(cur->skb);
cur              3243 net/core/pktgen.c 		cur->skb = NULL;
cur              3245 net/core/pktgen.c 		pktgen_remove_device(t, cur);
cur              3254 net/core/pktgen.c 	struct pktgen_dev *cur;
cur              3261 net/core/pktgen.c 		cur = list_entry(q, struct pktgen_dev, list);
cur              3263 net/core/pktgen.c 		kfree_skb(cur->skb);
cur              3264 net/core/pktgen.c 		cur->skb = NULL;
cur              3266 net/core/pktgen.c 		pktgen_remove_device(t, cur);
cur               121 net/core/sysctl_net_core.c 	struct sd_flow_limit *cur;
cur               135 net/core/sysctl_net_core.c 		len = sizeof(*cur) + netdev_flow_limit_table_len;
cur               138 net/core/sysctl_net_core.c 			cur = rcu_dereference_protected(sd->flow_limit,
cur               140 net/core/sysctl_net_core.c 			if (cur && !cpumask_test_cpu(i, mask)) {
cur               143 net/core/sysctl_net_core.c 				kfree(cur);
cur               144 net/core/sysctl_net_core.c 			} else if (!cur && cpumask_test_cpu(i, mask)) {
cur               145 net/core/sysctl_net_core.c 				cur = kzalloc_node(len, GFP_KERNEL,
cur               147 net/core/sysctl_net_core.c 				if (!cur) {
cur               152 net/core/sysctl_net_core.c 				cur->num_buckets = netdev_flow_limit_table_len;
cur               153 net/core/sysctl_net_core.c 				rcu_assign_pointer(sd->flow_limit, cur);
cur                30 net/dccp/ackvec.c 	struct dccp_ackvec_record *cur, *next;
cur                32 net/dccp/ackvec.c 	list_for_each_entry_safe(cur, next, &av->av_records, avr_node)
cur                33 net/dccp/ackvec.c 		kmem_cache_free(dccp_ackvec_record_slab, cur);
cur               364 net/dccp/ackvec.c 	struct dccp_ackvec_parsed *cur, *next;
cur               366 net/dccp/ackvec.c 	list_for_each_entry_safe(cur, next, parsed_chunks, node)
cur               367 net/dccp/ackvec.c 		kfree(cur);
cur                86 net/dccp/ccids/lib/loss_interval.c 	struct tfrc_loss_interval *cur = tfrc_lh_peek(lh);
cur                90 net/dccp/ccids/lib/loss_interval.c 	if (cur == NULL)			/* not initialised */
cur                93 net/dccp/ccids/lib/loss_interval.c 	len = dccp_delta_seqno(cur->li_seqno, DCCP_SKB_CB(skb)->dccpd_seq) + 1;
cur                95 net/dccp/ccids/lib/loss_interval.c 	if (len - (s64)cur->li_length <= 0)	/* duplicate or reordered */
cur                98 net/dccp/ccids/lib/loss_interval.c 	if (SUB16(dccp_hdr(skb)->dccph_ccval, cur->li_ccval) > 4)
cur               107 net/dccp/ccids/lib/loss_interval.c 		cur->li_is_closed = 1;
cur               112 net/dccp/ccids/lib/loss_interval.c 	cur->li_length = len;
cur               119 net/dccp/ccids/lib/loss_interval.c static inline u8 tfrc_lh_is_new_loss(struct tfrc_loss_interval *cur,
cur               122 net/dccp/ccids/lib/loss_interval.c 	return	dccp_delta_seqno(cur->li_seqno, new_loss->tfrchrx_seqno) > 0 &&
cur               123 net/dccp/ccids/lib/loss_interval.c 		(cur->li_is_closed || SUB16(new_loss->tfrchrx_ccval, cur->li_ccval) > 4);
cur               138 net/dccp/ccids/lib/loss_interval.c 	struct tfrc_loss_interval *cur = tfrc_lh_peek(lh), *new;
cur               140 net/dccp/ccids/lib/loss_interval.c 	if (cur != NULL && !tfrc_lh_is_new_loss(cur, tfrc_rx_hist_loss_prev(rh)))
cur               156 net/dccp/ccids/lib/loss_interval.c 		cur->li_length = dccp_delta_seqno(cur->li_seqno, new->li_seqno);
cur              1500 net/dccp/feat.c 	struct dccp_feat_entry *cur, *next;
cur              1506 net/dccp/feat.c 	list_for_each_entry(cur, fn_list, node) {
cur              1512 net/dccp/feat.c 		if (cur->empty_confirm)
cur              1515 net/dccp/feat.c 		idx = dccp_feat_index(cur->feat_num);
cur              1517 net/dccp/feat.c 			DCCP_BUG("Unknown feature %u", cur->feat_num);
cur              1520 net/dccp/feat.c 		if (cur->state != FEAT_STABLE) {
cur              1522 net/dccp/feat.c 				  cur->is_local ? "local" : "remote",
cur              1523 net/dccp/feat.c 				  dccp_feat_fname(cur->feat_num),
cur              1524 net/dccp/feat.c 				  dccp_feat_sname[cur->state]);
cur              1527 net/dccp/feat.c 		fvals[idx][cur->is_local] = &cur->val;
cur              1544 net/dccp/feat.c 	list_for_each_entry_safe(cur, next, fn_list, node)
cur              1545 net/dccp/feat.c 		if (!cur->needs_confirm)
cur              1546 net/dccp/feat.c 			dccp_feat_list_pop(cur);
cur              2146 net/ipv4/tcp_ipv4.c static void *listening_get_next(struct seq_file *seq, void *cur)
cur              2153 net/ipv4/tcp_ipv4.c 	struct sock *sk = cur;
cur              2239 net/ipv4/tcp_ipv4.c static void *established_get_next(struct seq_file *seq, void *cur)
cur              2242 net/ipv4/tcp_ipv4.c 	struct sock *sk = cur;
cur              1916 net/ipv6/ip6_fib.c 		struct fib6_info *cur = rcu_dereference_protected(*rtp,
cur              1918 net/ipv6/ip6_fib.c 		if (rt == cur) {
cur              1922 net/ipv6/ip6_fib.c 		rtp_next = &cur->fib6_next;
cur               181 net/ipv6/ndisc.c static struct nd_opt_hdr *ndisc_next_option(struct nd_opt_hdr *cur,
cur               185 net/ipv6/ndisc.c 	if (!cur || !end || cur >= end)
cur               187 net/ipv6/ndisc.c 	type = cur->nd_opt_type;
cur               189 net/ipv6/ndisc.c 		cur = ((void *)cur) + (cur->nd_opt_len << 3);
cur               190 net/ipv6/ndisc.c 	} while (cur < end && cur->nd_opt_type != type);
cur               191 net/ipv6/ndisc.c 	return cur <= end && cur->nd_opt_type == type ? cur : NULL;
cur               204 net/ipv6/ndisc.c 					     struct nd_opt_hdr *cur,
cur               207 net/ipv6/ndisc.c 	if (!cur || !end || cur >= end)
cur               210 net/ipv6/ndisc.c 		cur = ((void *)cur) + (cur->nd_opt_len << 3);
cur               211 net/ipv6/ndisc.c 	} while (cur < end && !ndisc_is_useropt(dev, cur));
cur               212 net/ipv6/ndisc.c 	return cur <= end && ndisc_is_useropt(dev, cur) ? cur : NULL;
cur              3663 net/mac80211/util.c 		s32 cur;
cur              3671 net/mac80211/util.c 		cur = data->desc[i].start - tsf;
cur              3672 net/mac80211/util.c 		if (cur > *offset)
cur              3675 net/mac80211/util.c 		cur = data->desc[i].start + data->desc[i].duration - tsf;
cur              3676 net/mac80211/util.c 		if (cur > *offset)
cur              3677 net/mac80211/util.c 			*offset = cur;
cur                95 net/netfilter/nf_conntrack_h323_asn1.c 	unsigned char *cur;
cur               100 net/netfilter/nf_conntrack_h323_asn1.c #define INC_BIT(bs) if((++(bs)->bit)>7){(bs)->cur++;(bs)->bit=0;}
cur               101 net/netfilter/nf_conntrack_h323_asn1.c #define INC_BITS(bs,b) if(((bs)->bit+=(b))>7){(bs)->cur+=(bs)->bit>>3;(bs)->bit&=7;}
cur               102 net/netfilter/nf_conntrack_h323_asn1.c #define BYTE_ALIGN(bs) if((bs)->bit){(bs)->cur++;(bs)->bit=0;}
cur               154 net/netfilter/nf_conntrack_h323_asn1.c 	v = *bs->cur++;
cur               159 net/netfilter/nf_conntrack_h323_asn1.c 		v += *bs->cur++;
cur               172 net/netfilter/nf_conntrack_h323_asn1.c 	if (bs->cur + bytes > bs->end)
cur               180 net/netfilter/nf_conntrack_h323_asn1.c 	unsigned int b = (*bs->cur) & (0x80 >> bs->bit);
cur               192 net/netfilter/nf_conntrack_h323_asn1.c 	v = (*bs->cur) & (0xffU >> bs->bit);
cur               199 net/netfilter/nf_conntrack_h323_asn1.c 		bs->cur++;
cur               204 net/netfilter/nf_conntrack_h323_asn1.c 		v += *(++bs->cur);
cur               223 net/netfilter/nf_conntrack_h323_asn1.c 		v = (unsigned int)(*bs->cur) << (bs->bit + 24);
cur               226 net/netfilter/nf_conntrack_h323_asn1.c 		v = (unsigned int)(*bs->cur++) << (bs->bit + 24);
cur               231 net/netfilter/nf_conntrack_h323_asn1.c 			v |= (unsigned int)(*bs->cur++) << shift;
cur               234 net/netfilter/nf_conntrack_h323_asn1.c 			v |= (unsigned int)(*bs->cur) << shift;
cur               238 net/netfilter/nf_conntrack_h323_asn1.c 			v |= (*bs->cur) >> (8 - bs->bit);
cur               258 net/netfilter/nf_conntrack_h323_asn1.c 		v |= *bs->cur++;
cur               262 net/netfilter/nf_conntrack_h323_asn1.c 		v |= *bs->cur++;
cur               266 net/netfilter/nf_conntrack_h323_asn1.c 		v |= *bs->cur++;
cur               270 net/netfilter/nf_conntrack_h323_asn1.c 		v |= *bs->cur++;
cur               306 net/netfilter/nf_conntrack_h323_asn1.c 	len = *bs->cur++;
cur               307 net/netfilter/nf_conntrack_h323_asn1.c 	bs->cur += len;
cur               324 net/netfilter/nf_conntrack_h323_asn1.c 		bs->cur++;
cur               328 net/netfilter/nf_conntrack_h323_asn1.c 		bs->cur += 2;
cur               340 net/netfilter/nf_conntrack_h323_asn1.c 		bs->cur += len;
cur               347 net/netfilter/nf_conntrack_h323_asn1.c 		bs->cur += len;
cur               392 net/netfilter/nf_conntrack_h323_asn1.c 		len = (*bs->cur++) << 8;
cur               393 net/netfilter/nf_conntrack_h323_asn1.c 		len += (*bs->cur++) + f->lb;
cur               405 net/netfilter/nf_conntrack_h323_asn1.c 	bs->cur += len >> 3;
cur               448 net/netfilter/nf_conntrack_h323_asn1.c 					     bs->cur[0], bs->cur[1],
cur               449 net/netfilter/nf_conntrack_h323_asn1.c 					     bs->cur[2], bs->cur[3],
cur               450 net/netfilter/nf_conntrack_h323_asn1.c 					     bs->cur[4] * 256 + bs->cur[5]));
cur               452 net/netfilter/nf_conntrack_h323_asn1.c 				    bs->cur - bs->buf;
cur               461 net/netfilter/nf_conntrack_h323_asn1.c 		len = (*bs->cur++) + f->lb;
cur               477 net/netfilter/nf_conntrack_h323_asn1.c 	bs->cur += len;
cur               498 net/netfilter/nf_conntrack_h323_asn1.c 		len = (*bs->cur++) + f->lb;
cur               508 net/netfilter/nf_conntrack_h323_asn1.c 	bs->cur += len << 1;
cur               563 net/netfilter/nf_conntrack_h323_asn1.c 				bs->cur += len;
cur               566 net/netfilter/nf_conntrack_h323_asn1.c 			beg = bs->cur;
cur               574 net/netfilter/nf_conntrack_h323_asn1.c 			bs->cur = beg + len;
cur               607 net/netfilter/nf_conntrack_h323_asn1.c 			bs->cur += len;
cur               628 net/netfilter/nf_conntrack_h323_asn1.c 			bs->cur += len;
cur               631 net/netfilter/nf_conntrack_h323_asn1.c 		beg = bs->cur;
cur               638 net/netfilter/nf_conntrack_h323_asn1.c 		bs->cur = beg + len;
cur               663 net/netfilter/nf_conntrack_h323_asn1.c 		count = *bs->cur++;
cur               669 net/netfilter/nf_conntrack_h323_asn1.c 		count = *bs->cur++;
cur               671 net/netfilter/nf_conntrack_h323_asn1.c 		count += *bs->cur++;
cur               709 net/netfilter/nf_conntrack_h323_asn1.c 				bs->cur += len;
cur               712 net/netfilter/nf_conntrack_h323_asn1.c 			beg = bs->cur;
cur               722 net/netfilter/nf_conntrack_h323_asn1.c 			bs->cur = beg + len;
cur               782 net/netfilter/nf_conntrack_h323_asn1.c 		bs->cur += len;
cur               803 net/netfilter/nf_conntrack_h323_asn1.c 			bs->cur += len;
cur               806 net/netfilter/nf_conntrack_h323_asn1.c 		beg = bs->cur;
cur               812 net/netfilter/nf_conntrack_h323_asn1.c 		bs->cur = beg + len;
cur               829 net/netfilter/nf_conntrack_h323_asn1.c 	bs.buf = bs.beg = bs.cur = buf;
cur               846 net/netfilter/nf_conntrack_h323_asn1.c 	bs.beg = bs.cur = beg;
cur               863 net/netfilter/nf_conntrack_h323_asn1.c 	bs.buf = bs.beg = bs.cur = buf;
cur               137 net/netfilter/nf_conntrack_helper.c 	struct nf_conntrack_nat_helper *cur;
cur               140 net/netfilter/nf_conntrack_helper.c 	list_for_each_entry_rcu(cur, &nf_ct_nat_helpers, list) {
cur               141 net/netfilter/nf_conntrack_helper.c 		if (!strcmp(cur->mod_name, mod_name)) {
cur               146 net/netfilter/nf_conntrack_helper.c 	return found ? cur : NULL;
cur               339 net/netfilter/nf_conntrack_helper.c 	struct nf_ct_helper_expectfn *cur;
cur               342 net/netfilter/nf_conntrack_helper.c 	list_for_each_entry_rcu(cur, &nf_ct_helper_expectfn_list, head) {
cur               343 net/netfilter/nf_conntrack_helper.c 		if (!strcmp(cur->name, name)) {
cur               348 net/netfilter/nf_conntrack_helper.c 	return found ? cur : NULL;
cur               356 net/netfilter/nf_conntrack_helper.c 	struct nf_ct_helper_expectfn *cur;
cur               359 net/netfilter/nf_conntrack_helper.c 	list_for_each_entry_rcu(cur, &nf_ct_helper_expectfn_list, head) {
cur               360 net/netfilter/nf_conntrack_helper.c 		if (cur->expectfn == symbol) {
cur               365 net/netfilter/nf_conntrack_helper.c 	return found ? cur : NULL;
cur               400 net/netfilter/nf_conntrack_helper.c 	struct nf_conntrack_helper *cur;
cur               412 net/netfilter/nf_conntrack_helper.c 		hlist_for_each_entry(cur, &nf_ct_helper_hash[i], hnode) {
cur               413 net/netfilter/nf_conntrack_helper.c 			if (!strcmp(cur->name, me->name) &&
cur               414 net/netfilter/nf_conntrack_helper.c 			    (cur->tuple.src.l3num == NFPROTO_UNSPEC ||
cur               415 net/netfilter/nf_conntrack_helper.c 			     cur->tuple.src.l3num == me->tuple.src.l3num) &&
cur               416 net/netfilter/nf_conntrack_helper.c 			    cur->tuple.dst.protonum == me->tuple.dst.protonum) {
cur               425 net/netfilter/nf_conntrack_helper.c 		hlist_for_each_entry(cur, &nf_ct_helper_hash[h], hnode) {
cur               426 net/netfilter/nf_conntrack_helper.c 			if (nf_ct_tuple_src_mask_cmp(&cur->tuple, &me->tuple,
cur               191 net/netfilter/nfnetlink_acct.c 	struct nf_acct *cur, *last;
cur               202 net/netfilter/nfnetlink_acct.c 	list_for_each_entry_rcu(cur, &net->nfnl_acct_list, head) {
cur               204 net/netfilter/nfnetlink_acct.c 			if (cur != last)
cur               210 net/netfilter/nfnetlink_acct.c 		if (filter && (cur->flags & filter->mask) != filter->value)
cur               216 net/netfilter/nfnetlink_acct.c 				       NFNL_MSG_ACCT_NEW, cur) < 0) {
cur               217 net/netfilter/nfnetlink_acct.c 			cb->args[1] = (unsigned long)cur;
cur               273 net/netfilter/nfnetlink_acct.c 	struct nf_acct *cur;
cur               291 net/netfilter/nfnetlink_acct.c 	list_for_each_entry(cur, &net->nfnl_acct_list, head) {
cur               294 net/netfilter/nfnetlink_acct.c 		if (strncmp(cur->name, acct_name, NFACCT_NAME_MAX)!= 0)
cur               306 net/netfilter/nfnetlink_acct.c 					 NFNL_MSG_ACCT_NEW, cur);
cur               323 net/netfilter/nfnetlink_acct.c static int nfnl_acct_try_del(struct nf_acct *cur)
cur               330 net/netfilter/nfnetlink_acct.c 	if (refcount_dec_if_one(&cur->refcnt)) {
cur               332 net/netfilter/nfnetlink_acct.c 		list_del_rcu(&cur->head);
cur               333 net/netfilter/nfnetlink_acct.c 		kfree_rcu(cur, rcu_head);
cur               345 net/netfilter/nfnetlink_acct.c 	struct nf_acct *cur, *tmp;
cur               350 net/netfilter/nfnetlink_acct.c 		list_for_each_entry_safe(cur, tmp, &net->nfnl_acct_list, head)
cur               351 net/netfilter/nfnetlink_acct.c 			nfnl_acct_try_del(cur);
cur               357 net/netfilter/nfnetlink_acct.c 	list_for_each_entry(cur, &net->nfnl_acct_list, head) {
cur               358 net/netfilter/nfnetlink_acct.c 		if (strncmp(cur->name, acct_name, NFACCT_NAME_MAX) != 0)
cur               361 net/netfilter/nfnetlink_acct.c 		ret = nfnl_acct_try_del(cur);
cur               405 net/netfilter/nfnetlink_acct.c 	struct nf_acct *cur, *acct = NULL;
cur               408 net/netfilter/nfnetlink_acct.c 	list_for_each_entry_rcu(cur, &net->nfnl_acct_list, head) {
cur               409 net/netfilter/nfnetlink_acct.c 		if (strncmp(cur->name, acct_name, NFACCT_NAME_MAX)!= 0)
cur               415 net/netfilter/nfnetlink_acct.c 		if (!refcount_inc_not_zero(&cur->refcnt)) {
cur               420 net/netfilter/nfnetlink_acct.c 		acct = cur;
cur               498 net/netfilter/nfnetlink_acct.c 	struct nf_acct *cur, *tmp;
cur               500 net/netfilter/nfnetlink_acct.c 	list_for_each_entry_safe(cur, tmp, &net->nfnl_acct_list, head) {
cur               501 net/netfilter/nfnetlink_acct.c 		list_del_rcu(&cur->head);
cur               503 net/netfilter/nfnetlink_acct.c 		if (refcount_dec_and_test(&cur->refcnt))
cur               504 net/netfilter/nfnetlink_acct.c 			kfree_rcu(cur, rcu_head);
cur               417 net/netfilter/nfnetlink_cthelper.c 	struct nf_conntrack_helper *cur, *helper = NULL;
cur               435 net/netfilter/nfnetlink_cthelper.c 		cur = &nlcth->helper;
cur               437 net/netfilter/nfnetlink_cthelper.c 		if (strncmp(cur->name, helper_name, NF_CT_HELPER_NAME_LEN))
cur               440 net/netfilter/nfnetlink_cthelper.c 		if ((tuple.src.l3num != cur->tuple.src.l3num ||
cur               441 net/netfilter/nfnetlink_cthelper.c 		     tuple.dst.protonum != cur->tuple.dst.protonum))
cur               447 net/netfilter/nfnetlink_cthelper.c 		helper = cur;
cur               578 net/netfilter/nfnetlink_cthelper.c 	struct nf_conntrack_helper *cur, *last;
cur               584 net/netfilter/nfnetlink_cthelper.c 		hlist_for_each_entry_rcu(cur,
cur               588 net/netfilter/nfnetlink_cthelper.c 			if (!(cur->flags & NF_CT_HELPER_F_USERSPACE))
cur               592 net/netfilter/nfnetlink_cthelper.c 				if (cur != last)
cur               600 net/netfilter/nfnetlink_cthelper.c 					    NFNL_MSG_CTHELPER_NEW, cur) < 0) {
cur               601 net/netfilter/nfnetlink_cthelper.c 				cb->args[1] = (unsigned long)cur;
cur               621 net/netfilter/nfnetlink_cthelper.c 	struct nf_conntrack_helper *cur;
cur               650 net/netfilter/nfnetlink_cthelper.c 		cur = &nlcth->helper;
cur               652 net/netfilter/nfnetlink_cthelper.c 		    strncmp(cur->name, helper_name, NF_CT_HELPER_NAME_LEN))
cur               656 net/netfilter/nfnetlink_cthelper.c 		    (tuple.src.l3num != cur->tuple.src.l3num ||
cur               657 net/netfilter/nfnetlink_cthelper.c 		     tuple.dst.protonum != cur->tuple.dst.protonum))
cur               669 net/netfilter/nfnetlink_cthelper.c 					      NFNL_MSG_CTHELPER_NEW, cur);
cur               692 net/netfilter/nfnetlink_cthelper.c 	struct nf_conntrack_helper *cur;
cur               714 net/netfilter/nfnetlink_cthelper.c 		cur = &nlcth->helper;
cur               718 net/netfilter/nfnetlink_cthelper.c 		    strncmp(cur->name, helper_name, NF_CT_HELPER_NAME_LEN))
cur               722 net/netfilter/nfnetlink_cthelper.c 		    (tuple.src.l3num != cur->tuple.src.l3num ||
cur               723 net/netfilter/nfnetlink_cthelper.c 		     tuple.dst.protonum != cur->tuple.dst.protonum))
cur               726 net/netfilter/nfnetlink_cthelper.c 		if (refcount_dec_if_one(&cur->refcnt)) {
cur               728 net/netfilter/nfnetlink_cthelper.c 			nf_conntrack_helper_unregister(cur);
cur               729 net/netfilter/nfnetlink_cthelper.c 			kfree(cur->expect_policy);
cur               787 net/netfilter/nfnetlink_cthelper.c 	struct nf_conntrack_helper *cur;
cur               793 net/netfilter/nfnetlink_cthelper.c 		cur = &nlcth->helper;
cur               795 net/netfilter/nfnetlink_cthelper.c 		nf_conntrack_helper_unregister(cur);
cur               796 net/netfilter/nfnetlink_cthelper.c 		kfree(cur->expect_policy);
cur               210 net/netfilter/nfnetlink_cttimeout.c 	struct ctnl_timeout *cur, *last;
cur               220 net/netfilter/nfnetlink_cttimeout.c 	list_for_each_entry_rcu(cur, &net->nfct_timeout_list, head) {
cur               222 net/netfilter/nfnetlink_cttimeout.c 			if (cur != last)
cur               230 net/netfilter/nfnetlink_cttimeout.c 					   IPCTNL_MSG_TIMEOUT_NEW, cur) < 0) {
cur               231 net/netfilter/nfnetlink_cttimeout.c 			cb->args[1] = (unsigned long)cur;
cur               249 net/netfilter/nfnetlink_cttimeout.c 	struct ctnl_timeout *cur;
cur               262 net/netfilter/nfnetlink_cttimeout.c 	list_for_each_entry(cur, &net->nfct_timeout_list, head) {
cur               265 net/netfilter/nfnetlink_cttimeout.c 		if (strncmp(cur->name, name, CTNL_TIMEOUT_NAME_MAX) != 0)
cur               277 net/netfilter/nfnetlink_cttimeout.c 					     IPCTNL_MSG_TIMEOUT_NEW, cur);
cur               318 net/netfilter/nfnetlink_cttimeout.c 	struct ctnl_timeout *cur, *tmp;
cur               323 net/netfilter/nfnetlink_cttimeout.c 		list_for_each_entry_safe(cur, tmp, &net->nfct_timeout_list,
cur               325 net/netfilter/nfnetlink_cttimeout.c 			ctnl_timeout_try_del(net, cur);
cur               331 net/netfilter/nfnetlink_cttimeout.c 	list_for_each_entry(cur, &net->nfct_timeout_list, head) {
cur               332 net/netfilter/nfnetlink_cttimeout.c 		if (strncmp(cur->name, name, CTNL_TIMEOUT_NAME_MAX) != 0)
cur               335 net/netfilter/nfnetlink_cttimeout.c 		ret = ctnl_timeout_try_del(net, cur);
cur               583 net/netfilter/nfnetlink_cttimeout.c 	struct ctnl_timeout *cur, *tmp;
cur               588 net/netfilter/nfnetlink_cttimeout.c 	list_for_each_entry_safe(cur, tmp, &net->nfct_timeout_list, head) {
cur               589 net/netfilter/nfnetlink_cttimeout.c 		list_del_rcu(&cur->head);
cur               591 net/netfilter/nfnetlink_cttimeout.c 		if (refcount_dec_and_test(&cur->refcnt))
cur               592 net/netfilter/nfnetlink_cttimeout.c 			kfree_rcu(cur, rcu_head);
cur                54 net/netfilter/x_tables.c 	unsigned int cur; /* number of used slots in compat_tab[] */
cur               652 net/netfilter/x_tables.c 	if (xp->cur >= xp->number)
cur               655 net/netfilter/x_tables.c 	if (xp->cur)
cur               656 net/netfilter/x_tables.c 		delta += xp->compat_tab[xp->cur - 1].delta;
cur               657 net/netfilter/x_tables.c 	xp->compat_tab[xp->cur].offset = offset;
cur               658 net/netfilter/x_tables.c 	xp->compat_tab[xp->cur].delta = delta;
cur               659 net/netfilter/x_tables.c 	xp->cur++;
cur               672 net/netfilter/x_tables.c 		xt[af].cur = 0;
cur               680 net/netfilter/x_tables.c 	int mid, left = 0, right = xt[af].cur - 1;
cur               716 net/netfilter/x_tables.c 	xt[af].cur = 0;
cur               314 net/rds/ib_rdma.c 	struct llist_node *cur = NULL;
cur               318 net/rds/ib_rdma.c 		cur = &ibmr->llnode;
cur               319 net/rds/ib_rdma.c 		*next = cur;
cur               320 net/rds/ib_rdma.c 		next = &cur->next;
cur               323 net/rds/ib_rdma.c 	*nodes_tail = cur;
cur               110 net/rfkill/core.c 	bool cur, sav;
cur               372 net/rfkill/core.c 		rfkill_global_states[type].cur = blocked;
cur               377 net/rfkill/core.c 		rfkill_global_states[i].cur = blocked;
cur               453 net/rfkill/core.c 		rfkill_global_states[i].sav = rfkill_global_states[i].cur;
cur               454 net/rfkill/core.c 		rfkill_global_states[i].cur = true;
cur               521 net/rfkill/core.c 	return rfkill_global_states[type].cur;
cur               875 net/rfkill/core.c 	bool cur;
cur               880 net/rfkill/core.c 		cur = !!(rfkill->state & RFKILL_BLOCK_SW);
cur               881 net/rfkill/core.c 		rfkill_set_block(rfkill, cur);
cur               992 net/rfkill/core.c 	bool cur;
cur               997 net/rfkill/core.c 	cur = rfkill_global_states[rfkill->type].cur;
cur               998 net/rfkill/core.c 	rfkill_set_block(rfkill, cur);
cur                74 net/smc/smc_core.c 		struct smc_connection *cur = rb_entry(*link,
cur                78 net/smc/smc_core.c 		if (cur->alert_token_local > token)
cur               265 net/smc/smc_core.h 		struct smc_connection *cur = rb_entry(node,
cur               268 net/smc/smc_core.h 		if (cur->alert_token_local > token) {
cur               271 net/smc/smc_core.h 			if (cur->alert_token_local < token) {
cur               274 net/smc/smc_core.h 				res = cur;
cur               987 net/sunrpc/xdr.c 	unsigned int cur = xdr_stream_pos(xdr);
cur               995 net/sunrpc/xdr.c 	if (iov->iov_len > cur) {
cur               996 net/sunrpc/xdr.c 		offset = iov->iov_len - cur;
cur               999 net/sunrpc/xdr.c 		xdr->nwords = XDR_QUADLEN(buf->len - cur);
cur              1013 net/sunrpc/xdr.c 		xdr->nwords = XDR_QUADLEN(buf->len - cur);
cur              1165 net/sunrpc/xdr.c 	size_t cur;
cur              1169 net/sunrpc/xdr.c 		cur = min_t(size_t, buf->tail[0].iov_len, trim);
cur              1170 net/sunrpc/xdr.c 		buf->tail[0].iov_len -= cur;
cur              1171 net/sunrpc/xdr.c 		trim -= cur;
cur              1177 net/sunrpc/xdr.c 		cur = min_t(unsigned int, buf->page_len, trim);
cur              1178 net/sunrpc/xdr.c 		buf->page_len -= cur;
cur              1179 net/sunrpc/xdr.c 		trim -= cur;
cur              1185 net/sunrpc/xdr.c 		cur = min_t(size_t, buf->head[0].iov_len, trim);
cur              1186 net/sunrpc/xdr.c 		buf->head[0].iov_len -= cur;
cur              1187 net/sunrpc/xdr.c 		trim -= cur;
cur               228 net/sunrpc/xprtmultipath.c 		const struct rpc_xprt *cur)
cur               234 net/sunrpc/xprtmultipath.c 		if (cur == pos)
cur               278 net/sunrpc/xprtmultipath.c 		const struct rpc_xprt *cur)
cur               284 net/sunrpc/xprtmultipath.c 		if (cur == prev)
cur               319 net/sunrpc/xprtmultipath.c 		const struct rpc_xprt *cur)
cur               323 net/sunrpc/xprtmultipath.c 	ret = xprt_switch_find_next_entry(head, cur);
cur               331 net/sunrpc/xprtmultipath.c 		const struct rpc_xprt *cur)
cur               340 net/sunrpc/xprtmultipath.c 		xprt = __xprt_switch_find_next_entry_roundrobin(head, cur);
cur               349 net/sunrpc/xprtmultipath.c 		cur = xprt;
cur               363 net/sunrpc/xprtmultipath.c 		const struct rpc_xprt *cur)
cur               365 net/sunrpc/xprtmultipath.c 	return xprt_switch_find_next_entry(&xps->xps_xprt_list, cur);
cur               360 net/tipc/monitor.c 	struct tipc_peer *cur, *prev, *p;
cur               374 net/tipc/monitor.c 	list_for_each_entry(cur, &self->list, list) {
cur               375 net/tipc/monitor.c 		if ((addr > prev->addr) && (addr < cur->addr))
cur               377 net/tipc/monitor.c 		if (((addr < cur->addr) || (addr > prev->addr)) &&
cur               378 net/tipc/monitor.c 		    (prev->addr > cur->addr))
cur               380 net/tipc/monitor.c 		prev = cur;
cur               382 net/tipc/monitor.c 	list_add_tail(&p->list, &cur->list);
cur               553 scripts/genksyms/genksyms.c 		struct string_list *cur;
cur               556 scripts/genksyms/genksyms.c 		cur = *(b++);
cur               557 scripts/genksyms/genksyms.c 		switch (cur->tag) {
cur               560 scripts/genksyms/genksyms.c 				fprintf(debugfile, "%s ", cur->string);
cur               561 scripts/genksyms/genksyms.c 			crc = partial_crc32(cur->string, crc);
cur               567 scripts/genksyms/genksyms.c 			subsym = find_symbol(cur->string, cur->tag, 0);
cur               571 scripts/genksyms/genksyms.c 					fprintf(debugfile, "%s ", cur->string);
cur               572 scripts/genksyms/genksyms.c 				crc = partial_crc32(cur->string, crc);
cur               584 scripts/genksyms/genksyms.c 			subsym = find_symbol(cur->string, cur->tag, 0);
cur               589 scripts/genksyms/genksyms.c 					       symbol_types[cur->tag].name,
cur               590 scripts/genksyms/genksyms.c 					       cur->string);
cur               592 scripts/genksyms/genksyms.c 						(symbol_types[cur->tag].name),
cur               593 scripts/genksyms/genksyms.c 						mk_node(cur->string),
cur               598 scripts/genksyms/genksyms.c 				    add_symbol(cur->string, cur->tag, n, 0);
cur               603 scripts/genksyms/genksyms.c 						symbol_types[cur->tag].name,
cur               604 scripts/genksyms/genksyms.c 						cur->string);
cur               607 scripts/genksyms/genksyms.c 				crc = partial_crc32(symbol_types[cur->tag].name,
cur               610 scripts/genksyms/genksyms.c 				crc = partial_crc32(cur->string, crc);
cur               594 scripts/kconfig/nconf.c 	ITEM *cur;
cur               597 scripts/kconfig/nconf.c 	cur = current_item(curses_menu);
cur               598 scripts/kconfig/nconf.c 	if (cur == NULL)
cur               600 scripts/kconfig/nconf.c 	mcur = (struct mitem *) item_userptr(cur);
cur               611 scripts/kconfig/nconf.c 	ITEM *cur;
cur               614 scripts/kconfig/nconf.c 	cur = current_item(curses_menu);
cur               615 scripts/kconfig/nconf.c 	if (!cur)
cur               617 scripts/kconfig/nconf.c 	mcur = (struct mitem *) item_userptr(cur);
cur               192 security/loadpin/loadpin.c 	char *cur;
cur               205 security/loadpin/loadpin.c 		cur = exclude_read_files[i];
cur               206 security/loadpin/loadpin.c 		if (!cur)
cur               208 security/loadpin/loadpin.c 		if (*cur == '\0')
cur               212 security/loadpin/loadpin.c 			if (strcmp(cur, kernel_read_file_str[j]) == 0) {
cur                70 security/selinux/ss/avtab.c 		  struct avtab_node *prev, struct avtab_node *cur,
cur               109 security/selinux/ss/avtab.c 	struct avtab_node *prev, *cur, *newnode;
cur               116 security/selinux/ss/avtab.c 	for (prev = NULL, cur = h->htable[hvalue];
cur               117 security/selinux/ss/avtab.c 	     cur;
cur               118 security/selinux/ss/avtab.c 	     prev = cur, cur = cur->next) {
cur               119 security/selinux/ss/avtab.c 		if (key->source_type == cur->key.source_type &&
cur               120 security/selinux/ss/avtab.c 		    key->target_type == cur->key.target_type &&
cur               121 security/selinux/ss/avtab.c 		    key->target_class == cur->key.target_class &&
cur               122 security/selinux/ss/avtab.c 		    (specified & cur->key.specified)) {
cur               128 security/selinux/ss/avtab.c 		if (key->source_type < cur->key.source_type)
cur               130 security/selinux/ss/avtab.c 		if (key->source_type == cur->key.source_type &&
cur               131 security/selinux/ss/avtab.c 		    key->target_type < cur->key.target_type)
cur               133 security/selinux/ss/avtab.c 		if (key->source_type == cur->key.source_type &&
cur               134 security/selinux/ss/avtab.c 		    key->target_type == cur->key.target_type &&
cur               135 security/selinux/ss/avtab.c 		    key->target_class < cur->key.target_class)
cur               139 security/selinux/ss/avtab.c 	newnode = avtab_insert_node(h, hvalue, prev, cur, key, datum);
cur               154 security/selinux/ss/avtab.c 	struct avtab_node *prev, *cur;
cur               160 security/selinux/ss/avtab.c 	for (prev = NULL, cur = h->htable[hvalue];
cur               161 security/selinux/ss/avtab.c 	     cur;
cur               162 security/selinux/ss/avtab.c 	     prev = cur, cur = cur->next) {
cur               163 security/selinux/ss/avtab.c 		if (key->source_type == cur->key.source_type &&
cur               164 security/selinux/ss/avtab.c 		    key->target_type == cur->key.target_type &&
cur               165 security/selinux/ss/avtab.c 		    key->target_class == cur->key.target_class &&
cur               166 security/selinux/ss/avtab.c 		    (specified & cur->key.specified))
cur               168 security/selinux/ss/avtab.c 		if (key->source_type < cur->key.source_type)
cur               170 security/selinux/ss/avtab.c 		if (key->source_type == cur->key.source_type &&
cur               171 security/selinux/ss/avtab.c 		    key->target_type < cur->key.target_type)
cur               173 security/selinux/ss/avtab.c 		if (key->source_type == cur->key.source_type &&
cur               174 security/selinux/ss/avtab.c 		    key->target_type == cur->key.target_type &&
cur               175 security/selinux/ss/avtab.c 		    key->target_class < cur->key.target_class)
cur               178 security/selinux/ss/avtab.c 	return avtab_insert_node(h, hvalue, prev, cur, key, datum);
cur               184 security/selinux/ss/avtab.c 	struct avtab_node *cur;
cur               191 security/selinux/ss/avtab.c 	for (cur = h->htable[hvalue]; cur;
cur               192 security/selinux/ss/avtab.c 	     cur = cur->next) {
cur               193 security/selinux/ss/avtab.c 		if (key->source_type == cur->key.source_type &&
cur               194 security/selinux/ss/avtab.c 		    key->target_type == cur->key.target_type &&
cur               195 security/selinux/ss/avtab.c 		    key->target_class == cur->key.target_class &&
cur               196 security/selinux/ss/avtab.c 		    (specified & cur->key.specified))
cur               197 security/selinux/ss/avtab.c 			return &cur->datum;
cur               199 security/selinux/ss/avtab.c 		if (key->source_type < cur->key.source_type)
cur               201 security/selinux/ss/avtab.c 		if (key->source_type == cur->key.source_type &&
cur               202 security/selinux/ss/avtab.c 		    key->target_type < cur->key.target_type)
cur               204 security/selinux/ss/avtab.c 		if (key->source_type == cur->key.source_type &&
cur               205 security/selinux/ss/avtab.c 		    key->target_type == cur->key.target_type &&
cur               206 security/selinux/ss/avtab.c 		    key->target_class < cur->key.target_class)
cur               220 security/selinux/ss/avtab.c 	struct avtab_node *cur;
cur               227 security/selinux/ss/avtab.c 	for (cur = h->htable[hvalue]; cur;
cur               228 security/selinux/ss/avtab.c 	     cur = cur->next) {
cur               229 security/selinux/ss/avtab.c 		if (key->source_type == cur->key.source_type &&
cur               230 security/selinux/ss/avtab.c 		    key->target_type == cur->key.target_type &&
cur               231 security/selinux/ss/avtab.c 		    key->target_class == cur->key.target_class &&
cur               232 security/selinux/ss/avtab.c 		    (specified & cur->key.specified))
cur               233 security/selinux/ss/avtab.c 			return cur;
cur               235 security/selinux/ss/avtab.c 		if (key->source_type < cur->key.source_type)
cur               237 security/selinux/ss/avtab.c 		if (key->source_type == cur->key.source_type &&
cur               238 security/selinux/ss/avtab.c 		    key->target_type < cur->key.target_type)
cur               240 security/selinux/ss/avtab.c 		if (key->source_type == cur->key.source_type &&
cur               241 security/selinux/ss/avtab.c 		    key->target_type == cur->key.target_type &&
cur               242 security/selinux/ss/avtab.c 		    key->target_class < cur->key.target_class)
cur               251 security/selinux/ss/avtab.c 	struct avtab_node *cur;
cur               257 security/selinux/ss/avtab.c 	for (cur = node->next; cur; cur = cur->next) {
cur               258 security/selinux/ss/avtab.c 		if (node->key.source_type == cur->key.source_type &&
cur               259 security/selinux/ss/avtab.c 		    node->key.target_type == cur->key.target_type &&
cur               260 security/selinux/ss/avtab.c 		    node->key.target_class == cur->key.target_class &&
cur               261 security/selinux/ss/avtab.c 		    (specified & cur->key.specified))
cur               262 security/selinux/ss/avtab.c 			return cur;
cur               264 security/selinux/ss/avtab.c 		if (node->key.source_type < cur->key.source_type)
cur               266 security/selinux/ss/avtab.c 		if (node->key.source_type == cur->key.source_type &&
cur               267 security/selinux/ss/avtab.c 		    node->key.target_type < cur->key.target_type)
cur               269 security/selinux/ss/avtab.c 		if (node->key.source_type == cur->key.source_type &&
cur               270 security/selinux/ss/avtab.c 		    node->key.target_type == cur->key.target_type &&
cur               271 security/selinux/ss/avtab.c 		    node->key.target_class < cur->key.target_class)
cur               280 security/selinux/ss/avtab.c 	struct avtab_node *cur, *temp;
cur               286 security/selinux/ss/avtab.c 		cur = h->htable[i];
cur               287 security/selinux/ss/avtab.c 		while (cur) {
cur               288 security/selinux/ss/avtab.c 			temp = cur;
cur               289 security/selinux/ss/avtab.c 			cur = cur->next;
cur               348 security/selinux/ss/avtab.c 	struct avtab_node *cur;
cur               354 security/selinux/ss/avtab.c 		cur = h->htable[i];
cur               355 security/selinux/ss/avtab.c 		if (cur) {
cur               358 security/selinux/ss/avtab.c 			while (cur) {
cur               360 security/selinux/ss/avtab.c 				cur = cur->next;
cur               599 security/selinux/ss/avtab.c int avtab_write_item(struct policydb *p, struct avtab_node *cur, void *fp)
cur               602 security/selinux/ss/avtab.c 	__le32 buf32[ARRAY_SIZE(cur->datum.u.xperms->perms.p)];
cur               606 security/selinux/ss/avtab.c 	buf16[0] = cpu_to_le16(cur->key.source_type);
cur               607 security/selinux/ss/avtab.c 	buf16[1] = cpu_to_le16(cur->key.target_type);
cur               608 security/selinux/ss/avtab.c 	buf16[2] = cpu_to_le16(cur->key.target_class);
cur               609 security/selinux/ss/avtab.c 	buf16[3] = cpu_to_le16(cur->key.specified);
cur               614 security/selinux/ss/avtab.c 	if (cur->key.specified & AVTAB_XPERMS) {
cur               615 security/selinux/ss/avtab.c 		rc = put_entry(&cur->datum.u.xperms->specified, sizeof(u8), 1, fp);
cur               618 security/selinux/ss/avtab.c 		rc = put_entry(&cur->datum.u.xperms->driver, sizeof(u8), 1, fp);
cur               621 security/selinux/ss/avtab.c 		for (i = 0; i < ARRAY_SIZE(cur->datum.u.xperms->perms.p); i++)
cur               622 security/selinux/ss/avtab.c 			buf32[i] = cpu_to_le32(cur->datum.u.xperms->perms.p[i]);
cur               624 security/selinux/ss/avtab.c 				ARRAY_SIZE(cur->datum.u.xperms->perms.p), fp);
cur               626 security/selinux/ss/avtab.c 		buf32[0] = cpu_to_le32(cur->datum.u.data);
cur               638 security/selinux/ss/avtab.c 	struct avtab_node *cur;
cur               647 security/selinux/ss/avtab.c 		for (cur = a->htable[i]; cur;
cur               648 security/selinux/ss/avtab.c 		     cur = cur->next) {
cur               649 security/selinux/ss/avtab.c 			rc = avtab_write_item(p, cur, fp);
cur               103 security/selinux/ss/avtab.h int avtab_write_item(struct policydb *p, struct avtab_node *cur, void *fp);
cur                27 security/selinux/ss/conditional.c 	struct cond_expr *cur;
cur                31 security/selinux/ss/conditional.c 	for (cur = expr; cur; cur = cur->next) {
cur                32 security/selinux/ss/conditional.c 		switch (cur->expr_type) {
cur                37 security/selinux/ss/conditional.c 			s[sp] = p->bool_val_to_struct[cur->bool - 1]->state;
cur                91 security/selinux/ss/conditional.c 	struct cond_av_list *cur;
cur                99 security/selinux/ss/conditional.c 		for (cur = node->true_list; cur; cur = cur->next) {
cur               101 security/selinux/ss/conditional.c 				cur->node->key.specified &= ~AVTAB_ENABLED;
cur               103 security/selinux/ss/conditional.c 				cur->node->key.specified |= AVTAB_ENABLED;
cur               106 security/selinux/ss/conditional.c 		for (cur = node->false_list; cur; cur = cur->next) {
cur               109 security/selinux/ss/conditional.c 				cur->node->key.specified &= ~AVTAB_ENABLED;
cur               111 security/selinux/ss/conditional.c 				cur->node->key.specified |= AVTAB_ENABLED;
cur               133 security/selinux/ss/conditional.c 	struct cond_av_list *cur, *next;
cur               134 security/selinux/ss/conditional.c 	for (cur = list; cur; cur = next) {
cur               135 security/selinux/ss/conditional.c 		next = cur->next;
cur               137 security/selinux/ss/conditional.c 		kfree(cur);
cur               156 security/selinux/ss/conditional.c 	struct cond_node *next, *cur;
cur               161 security/selinux/ss/conditional.c 	for (cur = list; cur; cur = next) {
cur               162 security/selinux/ss/conditional.c 		next = cur->next;
cur               163 security/selinux/ss/conditional.c 		cond_node_destroy(cur);
cur               272 security/selinux/ss/conditional.c 	struct cond_av_list *other = data->other, *list, *cur;
cur               303 security/selinux/ss/conditional.c 				for (cur = other; cur; cur = cur->next) {
cur               304 security/selinux/ss/conditional.c 					if (cur->node == node_ptr) {
cur               591 security/selinux/ss/conditional.c 	struct cond_node *cur;
cur               597 security/selinux/ss/conditional.c 	for (cur = list; cur != NULL; cur = cur->next)
cur               604 security/selinux/ss/conditional.c 	for (cur = list; cur != NULL; cur = cur->next) {
cur               605 security/selinux/ss/conditional.c 		rc = cond_write_node(p, cur, fp);
cur                45 security/selinux/ss/hashtab.c 	struct hashtab_node *prev, *cur, *newnode;
cur                54 security/selinux/ss/hashtab.c 	cur = h->htable[hvalue];
cur                55 security/selinux/ss/hashtab.c 	while (cur && h->keycmp(h, key, cur->key) > 0) {
cur                56 security/selinux/ss/hashtab.c 		prev = cur;
cur                57 security/selinux/ss/hashtab.c 		cur = cur->next;
cur                60 security/selinux/ss/hashtab.c 	if (cur && (h->keycmp(h, key, cur->key) == 0))
cur                83 security/selinux/ss/hashtab.c 	struct hashtab_node *cur;
cur                89 security/selinux/ss/hashtab.c 	cur = h->htable[hvalue];
cur                90 security/selinux/ss/hashtab.c 	while (cur && h->keycmp(h, key, cur->key) > 0)
cur                91 security/selinux/ss/hashtab.c 		cur = cur->next;
cur                93 security/selinux/ss/hashtab.c 	if (!cur || (h->keycmp(h, key, cur->key) != 0))
cur                96 security/selinux/ss/hashtab.c 	return cur->datum;
cur               102 security/selinux/ss/hashtab.c 	struct hashtab_node *cur, *temp;
cur               108 security/selinux/ss/hashtab.c 		cur = h->htable[i];
cur               109 security/selinux/ss/hashtab.c 		while (cur) {
cur               110 security/selinux/ss/hashtab.c 			temp = cur;
cur               111 security/selinux/ss/hashtab.c 			cur = cur->next;
cur               129 security/selinux/ss/hashtab.c 	struct hashtab_node *cur;
cur               135 security/selinux/ss/hashtab.c 		cur = h->htable[i];
cur               136 security/selinux/ss/hashtab.c 		while (cur) {
cur               137 security/selinux/ss/hashtab.c 			ret = apply(cur->key, cur->datum, args);
cur               140 security/selinux/ss/hashtab.c 			cur = cur->next;
cur               150 security/selinux/ss/hashtab.c 	struct hashtab_node *cur;
cur               155 security/selinux/ss/hashtab.c 		cur = h->htable[i];
cur               156 security/selinux/ss/hashtab.c 		if (cur) {
cur               159 security/selinux/ss/hashtab.c 			while (cur) {
cur               161 security/selinux/ss/hashtab.c 				cur = cur->next;
cur              2858 security/selinux/ss/services.c 	struct cond_node *cur;
cur              2886 security/selinux/ss/services.c 	for (cur = policydb->cond_list; cur; cur = cur->next) {
cur              2887 security/selinux/ss/services.c 		rc = evaluate_cond_node(policydb, cur);
cur              2933 security/selinux/ss/services.c 	struct cond_node *cur;
cur              2943 security/selinux/ss/services.c 	for (cur = policydb->cond_list; cur; cur = cur->next) {
cur              2944 security/selinux/ss/services.c 		rc = evaluate_cond_node(policydb, cur);
cur               204 sound/core/seq/seq_clientmgr.c 	res->cur += num;
cur               205 sound/core/seq/seq_clientmgr.c 	if (res->cur > res->peak)
cur               206 sound/core/seq/seq_clientmgr.c 		res->peak = res->cur;
cur               211 sound/core/seq/seq_clientmgr.c 	res->cur -= num;
cur              1180 sound/core/seq/seq_clientmgr.c 	info->cur_clients = client_usage.cur;
cur              2452 sound/core/seq/seq_clientmgr.c 	snd_iprintf(buffer, "  cur  clients : %d\n", client_usage.cur);
cur                63 sound/core/seq/seq_clientmgr.h 	int cur;
cur               134 sound/core/seq/seq_prioq.c 	struct snd_seq_event_cell *cur, *prev;
cur               165 sound/core/seq/seq_prioq.c 	cur = f->head;		/* cursor */
cur               168 sound/core/seq/seq_prioq.c 	while (cur != NULL) {
cur               170 sound/core/seq/seq_prioq.c 		int rel = compare_timestamp_rel(&cell->event, &cur->event);
cur               179 sound/core/seq/seq_prioq.c 		prev = cur;
cur               180 sound/core/seq/seq_prioq.c 		cur = cur->next;
cur               191 sound/core/seq/seq_prioq.c 	cell->next = cur;
cur               193 sound/core/seq/seq_prioq.c 	if (f->head == cur) /* this is the first cell, set head to it */
cur               195 sound/core/seq/seq_prioq.c 	if (cur == NULL) /* reached end of the list */
cur              1166 sound/isa/es18xx.c 		unsigned char cur = inb(chip->port + ES18XX_PM);
cur              1168 sound/isa/es18xx.c 		if ((cur & mask) == val)
cur              1170 sound/isa/es18xx.c 		outb((cur & ~mask) | val, chip->port + ES18XX_PM);
cur              4120 sound/pci/hda/hda_generic.c static bool dyn_adc_pcm_resetup(struct hda_codec *codec, int cur);
cur              5609 sound/pci/hda/hda_generic.c static bool dyn_adc_pcm_resetup(struct hda_codec *codec, int cur)
cur              5612 sound/pci/hda/hda_generic.c 	hda_nid_t new_adc = spec->adc_nids[spec->dyn_adc_idx[cur]];
cur              1260 sound/pci/pcxhr/pcxhr.c 			int cur = rmh.stat[0];
cur              1271 sound/pci/pcxhr/pcxhr.c 				cur = 100 - (100 * cur) / ref;
cur              1272 sound/pci/pcxhr/pcxhr.c 				snd_iprintf(buffer, "cpu load    %d%%\n", cur);
cur               700 sound/soc/bcm/cygnus-pcm.c 	unsigned int res = 0, cur = 0, base = 0;
cur               711 sound/soc/bcm/cygnus-pcm.c 		cur = readl(aio->cygaud->audio + p_rbuf->rdaddr);
cur               713 sound/soc/bcm/cygnus-pcm.c 		cur = readl(aio->cygaud->audio + p_rbuf->wraddr);
cur               721 sound/soc/bcm/cygnus-pcm.c 	res = (cur & 0x7fffffff) - (base & 0x7fffffff);
cur              2789 sound/soc/codecs/arizona.c 	const __be32 *cur;
cur              2796 sound/soc/codecs/arizona.c 	of_property_for_each_u32(np, "wlf,inmode", prop, cur, val) {
cur              2805 sound/soc/codecs/arizona.c 	of_property_for_each_u32(np, "wlf,dmic-ref", prop, cur, val) {
cur              2814 sound/soc/codecs/arizona.c 	of_property_for_each_u32(np, "wlf,out-mono", prop, cur, val) {
cur              2823 sound/soc/codecs/arizona.c 	of_property_for_each_u32(np, "wlf,max-channels-clocked", prop, cur, val) {
cur              2832 sound/soc/codecs/arizona.c 	of_property_for_each_u32(np, "wlf,out-volume-limit", prop, cur, val) {
cur                63 sound/soc/codecs/cs47l92.c 	unsigned int ep_sel, mux, change, cur;
cur                82 sound/soc/codecs/cs47l92.c 	ret = regmap_read(madera->regmap, MADERA_OUTPUT_ENABLES_1, &cur);
cur               108 sound/soc/codecs/cs47l92.c 				 MADERA_OUT3L_ENA | MADERA_OUT3R_ENA, cur);
cur               113 sound/soc/codecs/cs47l92.c 		if (cur & (MADERA_OUT3L_ENA | MADERA_OUT3R_ENA))
cur               963 sound/soc/codecs/madera.c 	unsigned int cur, new;
cur               975 sound/soc/codecs/madera.c 	ret = regmap_read(dsp->regmap,  dsp->base, &cur);
cur               982 sound/soc/codecs/madera.c 	cur &= MADERA_DSP_RATE_MASK;
cur               986 sound/soc/codecs/madera.c 	if (new == cur) {
cur              2957 sound/soc/codecs/madera.c 	unsigned int reg, cur, tar;
cur              3004 sound/soc/codecs/madera.c 			  base + MADERA_AIF_RATE_CTRL, &cur);
cur              3010 sound/soc/codecs/madera.c 	if ((cur & MADERA_AIF1_RATE_MASK) == (tar & MADERA_AIF1_RATE_MASK))
cur               746 sound/soc/uniphier/aio-core.c 	int cur, diff, slope = 0, fs;
cur               751 sound/soc/uniphier/aio-core.c 	cur = aio_port_get_volume(sub);
cur               752 sound/soc/uniphier/aio-core.c 	diff = abs(vol - cur);
cur               763 sound/soc/uniphier/aio-core.c 	if (cur < vol)
cur                51 sound/synth/emux/soundfont.c static void add_preset(struct snd_sf_list *sflist, struct snd_sf_zone *cur);
cur              1138 sound/synth/emux/soundfont.c 	struct snd_sf_zone *cur;
cur              1145 sound/synth/emux/soundfont.c 		for (cur = sf->zones; cur; cur = cur->next) {
cur              1146 sound/synth/emux/soundfont.c 			if (! cur->mapped && cur->sample == NULL) {
cur              1148 sound/synth/emux/soundfont.c 				cur->sample = set_sample(sf, &cur->v);
cur              1149 sound/synth/emux/soundfont.c 				if (cur->sample == NULL)
cur              1153 sound/synth/emux/soundfont.c 			add_preset(sflist, cur);
cur              1163 sound/synth/emux/soundfont.c add_preset(struct snd_sf_list *sflist, struct snd_sf_zone *cur)
cur              1168 sound/synth/emux/soundfont.c 	zone = search_first_zone(sflist, cur->bank, cur->instr, cur->v.low);
cur              1169 sound/synth/emux/soundfont.c 	if (zone && zone->v.sf_id != cur->v.sf_id) {
cur              1174 sound/synth/emux/soundfont.c 			if (p->counter > cur->counter)
cur              1184 sound/synth/emux/soundfont.c 	if ((index = get_index(cur->bank, cur->instr, cur->v.low)) < 0)
cur              1186 sound/synth/emux/soundfont.c 	cur->next_zone = zone; /* zone link */
cur              1187 sound/synth/emux/soundfont.c 	cur->next_instr = sflist->presets[index]; /* preset table link */
cur              1188 sound/synth/emux/soundfont.c 	sflist->presets[index] = cur;
cur               262 sound/usb/clock.c 		int ret, i, cur;
cur               280 sound/usb/clock.c 		cur = ret;
cur               291 sound/usb/clock.c 			if (i == cur)
cur               356 sound/usb/clock.c 		int ret, i, cur;
cur               374 sound/usb/clock.c 		cur = ret;
cur               385 sound/usb/clock.c 			if (i == cur)
cur               131 tools/bpf/bpftool/cfg.c static bool cfg_partition_funcs(struct cfg *cfg, struct bpf_insn *cur,
cur               136 tools/bpf/bpftool/cfg.c 	func = cfg_append_func(cfg, cur);
cur               140 tools/bpf/bpftool/cfg.c 	for (; cur < end; cur++) {
cur               141 tools/bpf/bpftool/cfg.c 		if (cur->code != (BPF_JMP | BPF_CALL))
cur               143 tools/bpf/bpftool/cfg.c 		if (cur->src_reg != BPF_PSEUDO_CALL)
cur               145 tools/bpf/bpftool/cfg.c 		func = cfg_append_func(cfg, cur + cur->off + 1);
cur               167 tools/bpf/bpftool/cfg.c 	struct bpf_insn *cur, *end;
cur               170 tools/bpf/bpftool/cfg.c 	cur = func->start;
cur               173 tools/bpf/bpftool/cfg.c 	bb = func_append_bb(func, cur);
cur               177 tools/bpf/bpftool/cfg.c 	for (; cur <= end; cur++) {
cur               178 tools/bpf/bpftool/cfg.c 		if (is_jmp_insn(cur->code)) {
cur               179 tools/bpf/bpftool/cfg.c 			u8 opcode = BPF_OP(cur->code);
cur               184 tools/bpf/bpftool/cfg.c 			bb = func_append_bb(func, cur + cur->off + 1);
cur               189 tools/bpf/bpftool/cfg.c 				bb = func_append_bb(func, cur + 1);
cur               365 tools/bpf/bpftool/xlated_dumper.c 	struct bpf_insn *cur = insn_start;
cur               367 tools/bpf/bpftool/xlated_dumper.c 	for (; cur <= insn_end; cur++) {
cur               368 tools/bpf/bpftool/xlated_dumper.c 		printf("% 4d: ", (int)(cur - insn_start + start_idx));
cur               369 tools/bpf/bpftool/xlated_dumper.c 		print_bpf_insn(&cbs, cur, true);
cur               370 tools/bpf/bpftool/xlated_dumper.c 		if (cur != insn_end)
cur                35 tools/lib/bitmap.c 	int cur, rbot, rtop;
cur                39 tools/lib/bitmap.c 	rbot = cur = find_first_bit(bitmap, nbits);
cur                40 tools/lib/bitmap.c 	while (cur < nbits) {
cur                41 tools/lib/bitmap.c 		rtop = cur;
cur                42 tools/lib/bitmap.c 		cur = find_next_bit(bitmap, nbits, cur + 1);
cur                43 tools/lib/bitmap.c 		if (cur < nbits && cur <= rtop + 1)
cur                55 tools/lib/bitmap.c 		rbot = cur;
cur                91 tools/lib/bpf/hashmap.c 	struct hashmap_entry *cur, *tmp;
cur               105 tools/lib/bpf/hashmap.c 	hashmap__for_each_entry_safe(map, cur, tmp, bkt) {
cur               106 tools/lib/bpf/hashmap.c 		h = hash_bits(map->hash_fn(cur->key, map->ctx), new_cap_bits);
cur               107 tools/lib/bpf/hashmap.c 		hashmap_add_entry(&new_buckets[h], cur);
cur               123 tools/lib/bpf/hashmap.c 	struct hashmap_entry *cur, **prev_ptr;
cur               128 tools/lib/bpf/hashmap.c 	for (prev_ptr = &map->buckets[hash], cur = *prev_ptr;
cur               129 tools/lib/bpf/hashmap.c 	     cur;
cur               130 tools/lib/bpf/hashmap.c 	     prev_ptr = &cur->next, cur = cur->next) {
cur               131 tools/lib/bpf/hashmap.c 		if (map->equal_fn(cur->key, key, map->ctx)) {
cur               134 tools/lib/bpf/hashmap.c 			*entry = cur;
cur               138 tools/lib/bpf/hashmap.h #define hashmap__for_each_entry(map, cur, bkt)				    \
cur               140 tools/lib/bpf/hashmap.h 		for (cur = map->buckets[bkt]; cur; cur = cur->next)
cur               150 tools/lib/bpf/hashmap.h #define hashmap__for_each_entry_safe(map, cur, tmp, bkt)		    \
cur               152 tools/lib/bpf/hashmap.h 		for (cur = map->buckets[bkt];				    \
cur               153 tools/lib/bpf/hashmap.h 		     cur && ({tmp = cur->next; true; });		    \
cur               154 tools/lib/bpf/hashmap.h 		     cur = tmp)
cur               162 tools/lib/bpf/hashmap.h #define hashmap__for_each_key_entry(map, cur, _key)			    \
cur               163 tools/lib/bpf/hashmap.h 	for (cur = ({ size_t bkt = hash_bits(map->hash_fn((_key), map->ctx),\
cur               166 tools/lib/bpf/hashmap.h 	     cur;							    \
cur               167 tools/lib/bpf/hashmap.h 	     cur = cur->next)						    \
cur               168 tools/lib/bpf/hashmap.h 		if (map->equal_fn(cur->key, (_key), map->ctx))
cur               170 tools/lib/bpf/hashmap.h #define hashmap__for_each_key_entry_safe(map, cur, tmp, _key)		    \
cur               171 tools/lib/bpf/hashmap.h 	for (cur = ({ size_t bkt = hash_bits(map->hash_fn((_key), map->ctx),\
cur               173 tools/lib/bpf/hashmap.h 		     cur = map->buckets ? map->buckets[bkt] : NULL; });	    \
cur               174 tools/lib/bpf/hashmap.h 	     cur && ({ tmp = cur->next; true; });			    \
cur               175 tools/lib/bpf/hashmap.h 	     cur = tmp)							    \
cur               176 tools/lib/bpf/hashmap.h 		if (map->equal_fn(cur->key, (_key), map->ctx))
cur               703 tools/perf/util/cs-etm-decoder/cs-etm-decoder.c 	ocsd_datapath_resp_t cur = OCSD_RESP_CONT;
cur               710 tools/perf/util/cs-etm-decoder/cs-etm-decoder.c 			cur = ocsd_dt_process_data(decoder->dcd_tree,
cur               717 tools/perf/util/cs-etm-decoder/cs-etm-decoder.c 			cur = ocsd_dt_process_data(decoder->dcd_tree,
cur               734 tools/perf/util/cs-etm-decoder/cs-etm-decoder.c 		if (OCSD_DATA_RESP_IS_WAIT(cur))
cur               737 tools/perf/util/cs-etm-decoder/cs-etm-decoder.c 		prev_return = cur;
cur               740 tools/perf/util/cs-etm-decoder/cs-etm-decoder.c 	decoder->prev_return = cur;
cur               193 tools/perf/util/genelf_debug.c 		ubyte cur = data & 0x7F;
cur               196 tools/perf/util/genelf_debug.c 			cur |= 0x80;
cur               197 tools/perf/util/genelf_debug.c 		buffer_ext_add(be, &cur, 1);
cur               207 tools/perf/util/genelf_debug.c 		ubyte cur = data & 0x7F;
cur               211 tools/perf/util/genelf_debug.c 		if ((data == 0 && !(cur & 0x40)) ||
cur               212 tools/perf/util/genelf_debug.c 		    (data == -1l && (cur & 0x40)))
cur               215 tools/perf/util/genelf_debug.c 			cur |= 0x80;
cur               216 tools/perf/util/genelf_debug.c 		buffer_ext_add(be, &cur, 1);
cur                84 tools/perf/util/strfilter.c 	struct strfilter_node root, *cur, *last_op;
cur                91 tools/perf/util/strfilter.c 	last_op = cur = &root;
cur                97 tools/perf/util/strfilter.c 			if (!cur->r || !last_op->r)
cur                99 tools/perf/util/strfilter.c 			cur = strfilter_node__alloc(OP_and, last_op->r, NULL);
cur               100 tools/perf/util/strfilter.c 			if (!cur)
cur               102 tools/perf/util/strfilter.c 			last_op->r = cur;
cur               103 tools/perf/util/strfilter.c 			last_op = cur;
cur               106 tools/perf/util/strfilter.c 			if (!cur->r || !root.r)
cur               108 tools/perf/util/strfilter.c 			cur = strfilter_node__alloc(OP_or, root.r, NULL);
cur               109 tools/perf/util/strfilter.c 			if (!cur)
cur               111 tools/perf/util/strfilter.c 			root.r = cur;
cur               112 tools/perf/util/strfilter.c 			last_op = cur;
cur               115 tools/perf/util/strfilter.c 			if (cur->r)
cur               117 tools/perf/util/strfilter.c 			cur->r = strfilter_node__alloc(OP_not, NULL, NULL);
cur               118 tools/perf/util/strfilter.c 			if (!cur->r)
cur               120 tools/perf/util/strfilter.c 			cur = cur->r;
cur               123 tools/perf/util/strfilter.c 			if (cur->r)
cur               125 tools/perf/util/strfilter.c 			cur->r = strfilter_node__new(s + 1, &s);
cur               128 tools/perf/util/strfilter.c 			if (!cur->r || *s != ')')
cur               133 tools/perf/util/strfilter.c 			if (cur->r)
cur               135 tools/perf/util/strfilter.c 			cur->r = strfilter_node__alloc(NULL, NULL, NULL);
cur               136 tools/perf/util/strfilter.c 			if (!cur->r)
cur               138 tools/perf/util/strfilter.c 			cur->r->p = strndup(s, e - s);
cur               139 tools/perf/util/strfilter.c 			if (!cur->r->p)
cur               144 tools/perf/util/strfilter.c 	if (!cur->r)
cur               114 tools/perf/util/unwind-libunwind-local.c 	u8 *cur = *p;
cur               122 tools/perf/util/unwind-libunwind-local.c 		*val = dw_read(cur, unsigned long, end);
cur               132 tools/perf/util/unwind-libunwind-local.c 		*val = (unsigned long) cur;
cur               143 tools/perf/util/unwind-libunwind-local.c 		*val += dw_read(cur, s32, end);
cur               146 tools/perf/util/unwind-libunwind-local.c 		*val += dw_read(cur, u32, end);
cur               149 tools/perf/util/unwind-libunwind-local.c 		*val += dw_read(cur, s64, end);
cur               152 tools/perf/util/unwind-libunwind-local.c 		*val += dw_read(cur, u64, end);
cur               159 tools/perf/util/unwind-libunwind-local.c 	*p = cur;
cur               279 tools/power/cpupower/utils/helpers/bitmask.c 	unsigned int cur, rbot, rtop;
cur               283 tools/power/cpupower/utils/helpers/bitmask.c 	rbot = cur = bitmask_first(bmp);
cur               284 tools/power/cpupower/utils/helpers/bitmask.c 	while (cur < bmp->size) {
cur               285 tools/power/cpupower/utils/helpers/bitmask.c 		rtop = cur;
cur               286 tools/power/cpupower/utils/helpers/bitmask.c 		cur = bitmask_next(bmp, cur+1);
cur               287 tools/power/cpupower/utils/helpers/bitmask.c 		if (cur >= bmp->size || cur > rtop + 1) {
cur               289 tools/power/cpupower/utils/helpers/bitmask.c 			rbot = cur;
cur               160 tools/testing/radix-tree/main.c 	unsigned long start, end, count = 0, tagged, cur, tmp;
cur               167 tools/testing/radix-tree/main.c 		cur = start;
cur               169 tools/testing/radix-tree/main.c 		end = cur;
cur               173 tools/testing/radix-tree/main.c 	cur = rand();
cur               174 tools/testing/radix-tree/main.c 	if (cur & 1) {
cur               176 tools/testing/radix-tree/main.c 		if (cur & 2) {
cur               182 tools/testing/radix-tree/main.c 	if (cur & 4) {
cur               184 tools/testing/radix-tree/main.c 		if (cur & 8)
cur               187 tools/testing/radix-tree/main.c 	if (cur & 16) {
cur               189 tools/testing/radix-tree/main.c 		if (cur & 32) {
cur               195 tools/testing/radix-tree/main.c 	if (cur & 64) {
cur               197 tools/testing/radix-tree/main.c 		if (cur & 128)
cur              6709 tools/testing/selftests/bpf/test_btf.c 	const char *cur = strs;
cur              6712 tools/testing/selftests/bpf/test_btf.c 	while (cur < strs + len) {
cur              6713 tools/testing/selftests/bpf/test_btf.c 		fprintf(stderr, "string #%d: '%s'\n", i, cur);
cur              6714 tools/testing/selftests/bpf/test_btf.c 		cur += strlen(cur) + 1;
cur                78 tools/testing/selftests/gpio/gpio-mockup-chardev.c 	char *cur;
cur                92 tools/testing/selftests/gpio/gpio-mockup-chardev.c 		cur = strstr(line, consumer);
cur                93 tools/testing/selftests/gpio/gpio-mockup-chardev.c 		if (cur == NULL)
cur                96 tools/testing/selftests/gpio/gpio-mockup-chardev.c 		cur = strchr(line, ')');
cur                97 tools/testing/selftests/gpio/gpio-mockup-chardev.c 		if (!cur)
cur               100 tools/testing/selftests/gpio/gpio-mockup-chardev.c 		cur += 2;
cur               101 tools/testing/selftests/gpio/gpio-mockup-chardev.c 		if (!strncmp(cur, "out", 3)) {
cur               103 tools/testing/selftests/gpio/gpio-mockup-chardev.c 			cur += 4;
cur               104 tools/testing/selftests/gpio/gpio-mockup-chardev.c 		} else if (!strncmp(cur, "in", 2)) {
cur               106 tools/testing/selftests/gpio/gpio-mockup-chardev.c 			cur += 4;
cur               109 tools/testing/selftests/gpio/gpio-mockup-chardev.c 		if (!strncmp(cur, "hi", 2))
cur               111 tools/testing/selftests/gpio/gpio-mockup-chardev.c 		else if (!strncmp(cur, "lo", 2))
cur               191 tools/testing/selftests/net/udpgso_bench_rx.c 	char cur = data[0];
cur               195 tools/testing/selftests/net/udpgso_bench_rx.c 	if (cur < 'a' || cur > 'z')
cur               199 tools/testing/selftests/net/udpgso_bench_rx.c 		if (cur == 'z')
cur               200 tools/testing/selftests/net/udpgso_bench_rx.c 			cur = 'a';
cur               202 tools/testing/selftests/net/udpgso_bench_rx.c 			cur++;
cur               204 tools/testing/selftests/net/udpgso_bench_rx.c 		if (data[i] != cur)
cur               208 tools/testing/selftests/net/udpgso_bench_rx.c 			      sanitized_char(cur), cur);
cur               107 tools/testing/selftests/networking/timestamping/txtimestamp.c static void validate_timestamp(struct timespec *cur, int min_delay)
cur               112 tools/testing/selftests/networking/timestamping/txtimestamp.c 	cur64 = timespec_to_us64(cur);
cur               122 tools/testing/selftests/networking/timestamping/txtimestamp.c static void __print_timestamp(const char *name, struct timespec *cur,
cur               125 tools/testing/selftests/networking/timestamping/txtimestamp.c 	if (!(cur->tv_sec | cur->tv_nsec))
cur               129 tools/testing/selftests/networking/timestamping/txtimestamp.c 			name, cur->tv_sec, cur->tv_nsec / 1000,
cur               132 tools/testing/selftests/networking/timestamping/txtimestamp.c 	if (cur != &ts_usr)
cur               134 tools/testing/selftests/networking/timestamping/txtimestamp.c 			timespec_to_us64(cur) - timespec_to_us64(&ts_usr));
cur               664 tools/testing/selftests/networking/timestamping/txtimestamp.c 	struct addrinfo *addrs, *cur;
cur               671 tools/testing/selftests/networking/timestamping/txtimestamp.c 	cur = addrs;
cur               672 tools/testing/selftests/networking/timestamping/txtimestamp.c 	while (cur && !have_ipv4 && !have_ipv6) {
cur               673 tools/testing/selftests/networking/timestamping/txtimestamp.c 		if (!have_ipv4 && cur->ai_family == AF_INET) {
cur               674 tools/testing/selftests/networking/timestamping/txtimestamp.c 			memcpy(&daddr, cur->ai_addr, sizeof(daddr));
cur               678 tools/testing/selftests/networking/timestamping/txtimestamp.c 		else if (!have_ipv6 && cur->ai_family == AF_INET6) {
cur               679 tools/testing/selftests/networking/timestamping/txtimestamp.c 			memcpy(&daddr6, cur->ai_addr, sizeof(daddr6));
cur               683 tools/testing/selftests/networking/timestamping/txtimestamp.c 		cur = cur->ai_next;
cur               143 tools/testing/selftests/vm/mlock-random-test.c 	struct rlimit cur;
cur               146 tools/testing/selftests/vm/mlock-random-test.c 	getrlimit(RLIMIT_MEMLOCK, &cur);
cur               147 tools/testing/selftests/vm/mlock-random-test.c 	if (cur.rlim_cur < alloc_size) {
cur               149 tools/testing/selftests/vm/mlock-random-test.c 				alloc_size, (unsigned int)cur.rlim_cur);
cur               220 tools/testing/selftests/vm/mlock-random-test.c 	struct rlimit cur;
cur               222 tools/testing/selftests/vm/mlock-random-test.c 	getrlimit(RLIMIT_MEMLOCK, &cur);
cur               223 tools/testing/selftests/vm/mlock-random-test.c 	if (cur.rlim_cur >= alloc_size) {
cur               225 tools/testing/selftests/vm/mlock-random-test.c 				alloc_size, (unsigned int)cur.rlim_cur);
cur               233 tools/testing/selftests/vm/mlock-random-test.c 		int lock_size = (rand() % (alloc_size - cur.rlim_cur))
cur               234 tools/testing/selftests/vm/mlock-random-test.c 			+ cur.rlim_cur;
cur              1123 virt/kvm/arm/vgic/vgic-its.c 	struct its_device *cur, *temp;
cur              1125 virt/kvm/arm/vgic/vgic-its.c 	list_for_each_entry_safe(cur, temp, &its->device_list, dev_list)
cur              1126 virt/kvm/arm/vgic/vgic-its.c 		vgic_its_free_device(kvm, cur);
cur              1132 virt/kvm/arm/vgic/vgic-its.c 	struct its_collection *cur, *temp;
cur              1134 virt/kvm/arm/vgic/vgic-its.c 	list_for_each_entry_safe(cur, temp, &its->collection_list, coll_list)
cur              1135 virt/kvm/arm/vgic/vgic-its.c 		vgic_its_free_collection(its, cur->collection_id);
cur              2471 virt/kvm/kvm_main.c 	ktime_t start, cur;
cur              2478 virt/kvm/kvm_main.c 	start = cur = ktime_get();
cur              2494 virt/kvm/kvm_main.c 			cur = ktime_get();
cur              2495 virt/kvm/kvm_main.c 		} while (single_task_running() && ktime_before(cur, stop));
cur              2509 virt/kvm/kvm_main.c 	cur = ktime_get();
cur              2512 virt/kvm/kvm_main.c 	block_ns = ktime_to_ns(cur) - ktime_to_ns(start);