cur 97 arch/alpha/kernel/srmcons.c char *cur; cur 100 arch/alpha/kernel/srmcons.c for (cur = (char *)buf; remaining > 0; ) { cur 107 arch/alpha/kernel/srmcons.c if (cur[c] == '\n') cur 111 arch/alpha/kernel/srmcons.c result.as_long = callback_puts(0, cur, c); cur 114 arch/alpha/kernel/srmcons.c cur += result.bits.c; cur 247 arch/arc/kernel/kprobes.c struct kprobe *cur = kprobe_running(); cur 250 arch/arc/kernel/kprobes.c if (!cur) cur 253 arch/arc/kernel/kprobes.c resume_execution(cur, addr, regs); cur 256 arch/arc/kernel/kprobes.c arch_arm_kprobe(cur); cur 265 arch/arc/kernel/kprobes.c if ((kcb->kprobe_status != KPROBE_REENTER) && cur->post_handler) { cur 267 arch/arc/kernel/kprobes.c cur->post_handler(cur, regs, 0); cur 291 arch/arc/kernel/kprobes.c struct kprobe *cur = kprobe_running(); cur 303 arch/arc/kernel/kprobes.c resume_execution(cur, (unsigned long)cur->addr, regs); cur 324 arch/arc/kernel/kprobes.c kprobes_inc_nmissed_count(cur); cur 333 arch/arc/kernel/kprobes.c if (cur->fault_handler && cur->fault_handler(cur, regs, trapnr)) cur 462 arch/arc/kernel/unwind.c const u8 *cur = *pcur; cur 466 arch/arc/kernel/unwind.c for (shift = 0, value = 0; cur < end; shift += 7) { cur 468 arch/arc/kernel/unwind.c && (*cur & 0x7fU) >= (1U << (8 * sizeof(value) - shift))) { cur 469 arch/arc/kernel/unwind.c cur = end + 1; cur 472 arch/arc/kernel/unwind.c value |= (uleb128_t) (*cur & 0x7f) << shift; cur 473 arch/arc/kernel/unwind.c if (!(*cur++ & 0x80)) cur 476 arch/arc/kernel/unwind.c *pcur = cur; cur 483 arch/arc/kernel/unwind.c const u8 *cur = *pcur; cur 487 arch/arc/kernel/unwind.c for (shift = 0, value = 0; cur < end; shift += 7) { cur 489 arch/arc/kernel/unwind.c && (*cur & 0x7fU) >= (1U << (8 * sizeof(value) - shift))) { cur 490 arch/arc/kernel/unwind.c cur = end + 1; cur 493 arch/arc/kernel/unwind.c value |= (sleb128_t) (*cur & 0x7f) << shift; cur 494 arch/arc/kernel/unwind.c if (!(*cur & 0x80)) { cur 495 arch/arc/kernel/unwind.c value |= -(*cur++ & 0x40) << shift; cur 499 arch/arc/kernel/unwind.c *pcur = cur; cur 965 arch/arc/kernel/unwind.c const u8 *cur = cur 968 arch/arc/kernel/unwind.c startLoc = read_pointer(&cur, cur 969 arch/arc/kernel/unwind.c cur + tableSize, cur 974 arch/arc/kernel/unwind.c ptr = cur - tableSize; cur 44 arch/arm/mach-integrator/impd1.c u32 cur; cur 47 arch/arm/mach-integrator/impd1.c cur = readl(impd1->base + IMPD1_CTRL) & ~mask; cur 48 arch/arm/mach-integrator/impd1.c writel(cur | val, impd1->base + IMPD1_CTRL); cur 100 arch/arm/mach-rpc/dma.c unsigned int status, cur, end; cur 113 arch/arm/mach-rpc/dma.c cur = CURA; cur 116 arch/arm/mach-rpc/dma.c cur = CURB; cur 119 arch/arm/mach-rpc/dma.c writel(idma->cur_addr, base + cur); cur 267 arch/arm/mach-s3c24xx/mach-h1940.c { .volt = 4070, .cur = 162, .level = 100}, cur 268 arch/arm/mach-s3c24xx/mach-h1940.c { .volt = 4040, .cur = 165, .level = 95}, cur 269 arch/arm/mach-s3c24xx/mach-h1940.c { .volt = 4016, .cur = 164, .level = 90}, cur 270 arch/arm/mach-s3c24xx/mach-h1940.c { .volt = 3996, .cur = 166, .level = 85}, cur 271 arch/arm/mach-s3c24xx/mach-h1940.c { .volt = 3971, .cur = 168, .level = 80}, cur 272 arch/arm/mach-s3c24xx/mach-h1940.c { .volt = 3951, .cur = 168, .level = 75}, cur 273 arch/arm/mach-s3c24xx/mach-h1940.c { .volt = 3931, .cur = 170, .level = 70}, cur 274 arch/arm/mach-s3c24xx/mach-h1940.c { .volt = 3903, .cur = 172, .level = 65}, cur 275 arch/arm/mach-s3c24xx/mach-h1940.c { .volt = 3886, .cur = 172, .level = 60}, cur 276 arch/arm/mach-s3c24xx/mach-h1940.c { .volt = 3858, .cur = 176, .level = 55}, cur 277 arch/arm/mach-s3c24xx/mach-h1940.c { .volt = 3842, .cur = 176, .level = 50}, cur 278 arch/arm/mach-s3c24xx/mach-h1940.c { .volt = 3818, .cur = 176, .level = 45}, cur 279 arch/arm/mach-s3c24xx/mach-h1940.c { .volt = 3789, .cur = 180, .level = 40}, cur 280 arch/arm/mach-s3c24xx/mach-h1940.c { .volt = 3769, .cur = 180, .level = 35}, cur 281 arch/arm/mach-s3c24xx/mach-h1940.c { .volt = 3749, .cur = 184, .level = 30}, cur 282 arch/arm/mach-s3c24xx/mach-h1940.c { .volt = 3732, .cur = 184, .level = 25}, cur 283 arch/arm/mach-s3c24xx/mach-h1940.c { .volt = 3716, .cur = 184, .level = 20}, cur 284 arch/arm/mach-s3c24xx/mach-h1940.c { .volt = 3708, .cur = 184, .level = 15}, cur 285 arch/arm/mach-s3c24xx/mach-h1940.c { .volt = 3716, .cur = 96, .level = 10}, cur 286 arch/arm/mach-s3c24xx/mach-h1940.c { .volt = 3700, .cur = 96, .level = 5}, cur 287 arch/arm/mach-s3c24xx/mach-h1940.c { .volt = 3684, .cur = 96, .level = 0}, cur 291 arch/arm/mach-s3c24xx/mach-h1940.c { .volt = 4130, .cur = 0, .level = 100}, cur 292 arch/arm/mach-s3c24xx/mach-h1940.c { .volt = 3982, .cur = 0, .level = 50}, cur 293 arch/arm/mach-s3c24xx/mach-h1940.c { .volt = 3854, .cur = 0, .level = 10}, cur 294 arch/arm/mach-s3c24xx/mach-h1940.c { .volt = 3841, .cur = 0, .level = 0}, cur 166 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 4100, .cur = 156, .level = 100}, cur 167 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 4050, .cur = 156, .level = 95}, cur 168 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 4025, .cur = 141, .level = 90}, cur 169 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3995, .cur = 144, .level = 85}, cur 170 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3957, .cur = 162, .level = 80}, cur 171 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3931, .cur = 147, .level = 75}, cur 172 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3902, .cur = 147, .level = 70}, cur 173 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3863, .cur = 153, .level = 65}, cur 174 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3838, .cur = 150, .level = 60}, cur 175 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3800, .cur = 153, .level = 55}, cur 176 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3765, .cur = 153, .level = 50}, cur 177 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3748, .cur = 172, .level = 45}, cur 178 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3740, .cur = 153, .level = 40}, cur 179 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3714, .cur = 175, .level = 35}, cur 180 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3710, .cur = 156, .level = 30}, cur 181 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3963, .cur = 156, .level = 25}, cur 182 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3672, .cur = 178, .level = 20}, cur 183 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3651, .cur = 178, .level = 15}, cur 184 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3629, .cur = 178, .level = 10}, cur 185 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3612, .cur = 162, .level = 5}, cur 186 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3605, .cur = 162, .level = 0}, cur 190 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 4200, .cur = 0, .level = 100}, cur 191 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 4190, .cur = 0, .level = 99}, cur 192 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 4178, .cur = 0, .level = 95}, cur 193 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 4110, .cur = 0, .level = 70}, cur 194 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 4076, .cur = 0, .level = 65}, cur 195 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 4046, .cur = 0, .level = 60}, cur 196 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 4021, .cur = 0, .level = 55}, cur 197 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3999, .cur = 0, .level = 50}, cur 198 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3982, .cur = 0, .level = 45}, cur 199 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3965, .cur = 0, .level = 40}, cur 200 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3957, .cur = 0, .level = 35}, cur 201 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3948, .cur = 0, .level = 30}, cur 202 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3936, .cur = 0, .level = 25}, cur 203 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3927, .cur = 0, .level = 20}, cur 204 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3906, .cur = 0, .level = 15}, cur 205 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3880, .cur = 0, .level = 10}, cur 206 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3829, .cur = 0, .level = 5}, cur 207 arch/arm/mach-s3c24xx/mach-rx1950.c { .volt = 3820, .cur = 0, .level = 0}, cur 66 arch/arm/plat-samsung/adc.c struct s3c_adc_client *cur; cur 138 arch/arm/plat-samsung/adc.c adc->cur = next; cur 171 arch/arm/plat-samsung/adc.c if (!adc->cur) cur 257 arch/arm/plat-samsung/adc.c if (adc_dev->cur == client) cur 258 arch/arm/plat-samsung/adc.c adc_dev->cur = NULL; cur 272 arch/arm/plat-samsung/adc.c if (adc_dev->cur == NULL) cur 283 arch/arm/plat-samsung/adc.c struct s3c_adc_client *client = adc->cur; cur 318 arch/arm/plat-samsung/adc.c adc->cur = NULL; cur 236 arch/arm/probes/kprobes/core.c struct kprobe *p, *cur; cur 240 arch/arm/probes/kprobes/core.c cur = kprobe_running(); cur 265 arch/arm/probes/kprobes/core.c } else if (cur) { cur 331 arch/arm/probes/kprobes/core.c struct kprobe *cur = kprobe_running(); cur 344 arch/arm/probes/kprobes/core.c regs->ARM_pc = (long)cur->addr; cur 359 arch/arm/probes/kprobes/core.c kprobes_inc_nmissed_count(cur); cur 368 arch/arm/probes/kprobes/core.c if (cur->fault_handler && cur->fault_handler(cur, regs, fsr)) cur 131 arch/arm64/kernel/alternative.c u64 cur, d_size, ctr_el0; cur 136 arch/arm64/kernel/alternative.c cur = start & ~(d_size - 1); cur 143 arch/arm64/kernel/alternative.c asm volatile("dc civac, %0" : : "r" (cur) : "memory"); cur 144 arch/arm64/kernel/alternative.c } while (cur += d_size, cur < end); cur 472 arch/arm64/kernel/cpufeature.c s64 cur) cur 481 arch/arm64/kernel/cpufeature.c ret = new < cur ? new : cur; cur 484 arch/arm64/kernel/cpufeature.c if (!cur || !new) cur 488 arch/arm64/kernel/cpufeature.c ret = new > cur ? new : cur; cur 266 arch/arm64/kernel/probes/kprobes.c struct kprobe *cur = kprobe_running(); cur 268 arch/arm64/kernel/probes/kprobes.c if (!cur) cur 272 arch/arm64/kernel/probes/kprobes.c if (cur->ainsn.api.restore != 0) cur 273 arch/arm64/kernel/probes/kprobes.c instruction_pointer_set(regs, cur->ainsn.api.restore); cur 282 arch/arm64/kernel/probes/kprobes.c if (cur->post_handler) { cur 286 arch/arm64/kernel/probes/kprobes.c cur->post_handler(cur, regs, 0); cur 294 arch/arm64/kernel/probes/kprobes.c struct kprobe *cur = kprobe_running(); cur 307 arch/arm64/kernel/probes/kprobes.c instruction_pointer_set(regs, (unsigned long) cur->addr); cur 326 arch/arm64/kernel/probes/kprobes.c kprobes_inc_nmissed_count(cur); cur 335 arch/arm64/kernel/probes/kprobes.c if (cur->fault_handler && cur->fault_handler(cur, regs, fsr)) cur 14 arch/csky/kernel/cpu-probe.c unsigned int cur, next, i; cur 20 arch/csky/kernel/cpu-probe.c cur = mfcr("cr13"); cur 22 arch/csky/kernel/cpu-probe.c seq_printf(m, "product info[%d] : 0x%08x\n", i, cur); cur 27 arch/csky/kernel/cpu-probe.c if (cur == next) cur 30 arch/csky/kernel/cpu-probe.c cur = next; cur 868 arch/ia64/kernel/kprobes.c struct kprobe *cur = kprobe_running(); cur 871 arch/ia64/kernel/kprobes.c if (!cur) cur 874 arch/ia64/kernel/kprobes.c if ((kcb->kprobe_status != KPROBE_REENTER) && cur->post_handler) { cur 876 arch/ia64/kernel/kprobes.c cur->post_handler(cur, regs, 0); cur 879 arch/ia64/kernel/kprobes.c resume_execution(cur, regs); cur 895 arch/ia64/kernel/kprobes.c struct kprobe *cur = kprobe_running(); cur 909 arch/ia64/kernel/kprobes.c regs->cr_iip = ((unsigned long)cur->addr) & ~0xFULL; cur 910 arch/ia64/kernel/kprobes.c ia64_psr(regs)->ri = ((unsigned long)cur->addr) & 0xf; cur 924 arch/ia64/kernel/kprobes.c kprobes_inc_nmissed_count(cur); cur 933 arch/ia64/kernel/kprobes.c if (cur->fault_handler && cur->fault_handler(cur, regs, trapnr)) cur 107 arch/ia64/kernel/perfmon_default_smpl.c void *cur, *last; cur 119 arch/ia64/kernel/perfmon_default_smpl.c cur = buf+hdr->hdr_cur_offs; cur 127 arch/ia64/kernel/perfmon_default_smpl.c if ((last - cur) < PFM_DEFAULT_MAX_ENTRY_SIZE) goto full; cur 131 arch/ia64/kernel/perfmon_default_smpl.c ent = (pfm_default_smpl_entry_t *)cur; cur 145 arch/ia64/kernel/perfmon_default_smpl.c cur, last, cur 146 arch/ia64/kernel/perfmon_default_smpl.c last-cur, cur 188 arch/ia64/kernel/perfmon_default_smpl.c cur += entry_size; cur 193 arch/ia64/kernel/perfmon_default_smpl.c if ((last - cur) < PFM_DEFAULT_MAX_ENTRY_SIZE) goto full; cur 205 arch/ia64/kernel/perfmon_default_smpl.c DPRINT_ovfl(("sampling buffer full free=%lu, count=%lu, ovfl_notify=%d\n", last-cur, hdr->hdr_count, ovfl_notify)); cur 88 arch/m68k/sun3/sun3dvma.c struct list_head *cur; cur 92 arch/m68k/sun3/sun3dvma.c list_for_each(cur, holes) { cur 93 arch/m68k/sun3/sun3dvma.c hole = list_entry(cur, struct hole, list); cur 111 arch/m68k/sun3/sun3dvma.c struct list_head *cur; cur 114 arch/m68k/sun3/sun3dvma.c list_for_each(cur, &hole_list) { cur 115 arch/m68k/sun3/sun3dvma.c hole = list_entry(cur, struct hole, list); cur 155 arch/m68k/sun3/sun3dvma.c struct list_head *cur; cur 167 arch/m68k/sun3/sun3dvma.c list_for_each(cur, &hole_list) { cur 170 arch/m68k/sun3/sun3dvma.c hole = list_entry(cur, struct hole, list); cur 208 arch/m68k/sun3/sun3dvma.c struct list_head *cur; cur 222 arch/m68k/sun3/sun3dvma.c list_for_each(cur, &hole_list) { cur 223 arch/m68k/sun3/sun3dvma.c hole = list_entry(cur, struct hole, list); cur 244 arch/m68k/sun3/sun3dvma.c list_add(&(hole->list), cur); cur 71 arch/mips/boot/elf2ecoff.c int remaining, cur, count; cur 81 arch/mips/boot/elf2ecoff.c cur = remaining; cur 82 arch/mips/boot/elf2ecoff.c if (cur > sizeof ibuf) cur 83 arch/mips/boot/elf2ecoff.c cur = sizeof ibuf; cur 84 arch/mips/boot/elf2ecoff.c remaining -= cur; cur 85 arch/mips/boot/elf2ecoff.c if ((count = read(in, ibuf, cur)) != cur) { cur 91 arch/mips/boot/elf2ecoff.c if ((count = write(out, ibuf, cur)) != cur) { cur 146 arch/mips/cavium-octeon/csrc-octeon.c u64 cur, end, inc; cur 148 arch/mips/cavium-octeon/csrc-octeon.c cur = read_c0_cvmcount(); cur 151 arch/mips/cavium-octeon/csrc-octeon.c end = cur + inc; cur 153 arch/mips/cavium-octeon/csrc-octeon.c while (end > cur) cur 154 arch/mips/cavium-octeon/csrc-octeon.c cur = read_c0_cvmcount(); cur 160 arch/mips/cavium-octeon/csrc-octeon.c u64 cur, end, inc; cur 162 arch/mips/cavium-octeon/csrc-octeon.c cur = read_c0_cvmcount(); cur 165 arch/mips/cavium-octeon/csrc-octeon.c end = cur + inc; cur 167 arch/mips/cavium-octeon/csrc-octeon.c while (end > cur) cur 168 arch/mips/cavium-octeon/csrc-octeon.c cur = read_c0_cvmcount(); cur 174 arch/mips/cavium-octeon/csrc-octeon.c u64 cur, end; cur 176 arch/mips/cavium-octeon/csrc-octeon.c cur = read_c0_cvmcount(); cur 177 arch/mips/cavium-octeon/csrc-octeon.c end = cur + loops; cur 179 arch/mips/cavium-octeon/csrc-octeon.c while (end > cur) cur 180 arch/mips/cavium-octeon/csrc-octeon.c cur = read_c0_cvmcount(); cur 195 arch/mips/cavium-octeon/csrc-octeon.c u64 cur, end; cur 197 arch/mips/cavium-octeon/csrc-octeon.c cur = read_c0_cvmcount(); cur 207 arch/mips/cavium-octeon/csrc-octeon.c end = cur + end; cur 209 arch/mips/cavium-octeon/csrc-octeon.c end = cur + count; cur 211 arch/mips/cavium-octeon/csrc-octeon.c while (end > cur) cur 212 arch/mips/cavium-octeon/csrc-octeon.c cur = read_c0_cvmcount(); cur 178 arch/mips/include/asm/sgiarcs.h struct linux_bigint cur; cur 374 arch/mips/kernel/kprobes.c struct kprobe *cur = kprobe_running(); cur 377 arch/mips/kernel/kprobes.c if (!cur) cur 380 arch/mips/kernel/kprobes.c if ((kcb->kprobe_status != KPROBE_REENTER) && cur->post_handler) { cur 382 arch/mips/kernel/kprobes.c cur->post_handler(cur, regs, 0); cur 385 arch/mips/kernel/kprobes.c resume_execution(cur, regs, kcb); cur 403 arch/mips/kernel/kprobes.c struct kprobe *cur = kprobe_running(); cur 406 arch/mips/kernel/kprobes.c if (cur->fault_handler && cur->fault_handler(cur, regs, trapnr)) cur 410 arch/mips/kernel/kprobes.c resume_execution(cur, regs, kcb); cur 889 arch/mips/kvm/trap_emul.c unsigned int cur, change; cur 974 arch/mips/kvm/trap_emul.c cur = kvm_read_c0_guest_config1(cop0); cur 975 arch/mips/kvm/trap_emul.c change = (cur ^ v) & kvm_mips_config1_wrmask(vcpu); cur 977 arch/mips/kvm/trap_emul.c v = cur ^ change; cur 985 arch/mips/kvm/trap_emul.c cur = kvm_read_c0_guest_config3(cop0); cur 986 arch/mips/kvm/trap_emul.c change = (cur ^ v) & kvm_mips_config3_wrmask(vcpu); cur 988 arch/mips/kvm/trap_emul.c v = cur ^ change; cur 993 arch/mips/kvm/trap_emul.c cur = kvm_read_c0_guest_config4(cop0); cur 994 arch/mips/kvm/trap_emul.c change = (cur ^ v) & kvm_mips_config4_wrmask(vcpu); cur 996 arch/mips/kvm/trap_emul.c v = cur ^ change; cur 1001 arch/mips/kvm/trap_emul.c cur = kvm_read_c0_guest_config5(cop0); cur 1002 arch/mips/kvm/trap_emul.c change = (cur ^ v) & kvm_mips_config5_wrmask(vcpu); cur 1004 arch/mips/kvm/trap_emul.c v = cur ^ change; cur 2050 arch/mips/kvm/vz.c unsigned int cur, change; cur 2194 arch/mips/kvm/vz.c cur = read_gc0_config(); cur 2195 arch/mips/kvm/vz.c change = (cur ^ v) & kvm_vz_config_user_wrmask(vcpu); cur 2197 arch/mips/kvm/vz.c v = cur ^ change; cur 2204 arch/mips/kvm/vz.c cur = read_gc0_config1(); cur 2205 arch/mips/kvm/vz.c change = (cur ^ v) & kvm_vz_config1_user_wrmask(vcpu); cur 2207 arch/mips/kvm/vz.c v = cur ^ change; cur 2214 arch/mips/kvm/vz.c cur = read_gc0_config2(); cur 2215 arch/mips/kvm/vz.c change = (cur ^ v) & kvm_vz_config2_user_wrmask(vcpu); cur 2217 arch/mips/kvm/vz.c v = cur ^ change; cur 2224 arch/mips/kvm/vz.c cur = read_gc0_config3(); cur 2225 arch/mips/kvm/vz.c change = (cur ^ v) & kvm_vz_config3_user_wrmask(vcpu); cur 2227 arch/mips/kvm/vz.c v = cur ^ change; cur 2234 arch/mips/kvm/vz.c cur = read_gc0_config4(); cur 2235 arch/mips/kvm/vz.c change = (cur ^ v) & kvm_vz_config4_user_wrmask(vcpu); cur 2237 arch/mips/kvm/vz.c v = cur ^ change; cur 2244 arch/mips/kvm/vz.c cur = read_gc0_config5(); cur 2245 arch/mips/kvm/vz.c change = (cur ^ v) & kvm_vz_config5_user_wrmask(vcpu); cur 2247 arch/mips/kvm/vz.c v = cur ^ change; cur 1761 arch/mips/mm/tlbex.c int cur = pte; cur 1769 arch/mips/mm/tlbex.c uasm_i_srl(p, t, cur, _PAGE_PRESENT_SHIFT); cur 1770 arch/mips/mm/tlbex.c cur = t; cur 1772 arch/mips/mm/tlbex.c uasm_i_andi(p, t, cur, 1); cur 1780 arch/mips/mm/tlbex.c uasm_i_srl(p, t, cur, _PAGE_PRESENT_SHIFT); cur 1781 arch/mips/mm/tlbex.c cur = t; cur 1783 arch/mips/mm/tlbex.c uasm_i_andi(p, t, cur, cur 1813 arch/mips/mm/tlbex.c int cur = pte; cur 1816 arch/mips/mm/tlbex.c uasm_i_srl(p, t, cur, _PAGE_PRESENT_SHIFT); cur 1817 arch/mips/mm/tlbex.c cur = t; cur 1819 arch/mips/mm/tlbex.c uasm_i_andi(p, t, cur, cur 482 arch/powerpc/kernel/kprobes.c struct kprobe *cur = kprobe_running(); cur 485 arch/powerpc/kernel/kprobes.c if (!cur || user_mode(regs)) cur 489 arch/powerpc/kernel/kprobes.c if (((unsigned long)cur->ainsn.insn + 4) != regs->nip) cur 492 arch/powerpc/kernel/kprobes.c if ((kcb->kprobe_status != KPROBE_REENTER) && cur->post_handler) { cur 494 arch/powerpc/kernel/kprobes.c cur->post_handler(cur, regs, 0); cur 498 arch/powerpc/kernel/kprobes.c regs->nip = (unsigned long)cur->addr + 4; cur 524 arch/powerpc/kernel/kprobes.c struct kprobe *cur = kprobe_running(); cur 538 arch/powerpc/kernel/kprobes.c regs->nip = (unsigned long)cur->addr; cur 554 arch/powerpc/kernel/kprobes.c kprobes_inc_nmissed_count(cur); cur 563 arch/powerpc/kernel/kprobes.c if (cur->fault_handler && cur->fault_handler(cur, regs, trapnr)) cur 3787 arch/powerpc/kvm/book3s_hv.c ktime_t cur, start_poll, start_wait; cur 3793 arch/powerpc/kvm/book3s_hv.c cur = start_poll = ktime_get(); cur 3806 arch/powerpc/kvm/book3s_hv.c cur = ktime_get(); cur 3807 arch/powerpc/kvm/book3s_hv.c } while (single_task_running() && ktime_before(cur, stop)); cur 3841 arch/powerpc/kvm/book3s_hv.c cur = ktime_get(); cur 3844 arch/powerpc/kvm/book3s_hv.c block_ns = ktime_to_ns(cur) - ktime_to_ns(start_poll); cur 3849 arch/powerpc/kvm/book3s_hv.c ktime_to_ns(cur) - ktime_to_ns(start_wait); cur 3859 arch/powerpc/kvm/book3s_hv.c ktime_to_ns(cur) - cur 258 arch/powerpc/kvm/book3s_xive.h u32 cur; cur 262 arch/powerpc/kvm/book3s_xive.h cur = be32_to_cpup(qpage + *idx); cur 263 arch/powerpc/kvm/book3s_xive.h if ((cur >> 31) == *toggle) cur 268 arch/powerpc/kvm/book3s_xive.h return cur & 0x7fffffff; cur 88 arch/powerpc/platforms/cell/cpufreq_spudemand.c if (!policy->cur) { cur 667 arch/powerpc/platforms/powermac/pci.c int i, cur = -1; cur 683 arch/powerpc/platforms/powermac/pci.c if (++cur >= 3) { cur 687 arch/powerpc/platforms/powermac/pci.c hose->mem_resources[cur].flags = IORESOURCE_MEM; cur 688 arch/powerpc/platforms/powermac/pci.c hose->mem_resources[cur].name = hose->dn->full_name; cur 689 arch/powerpc/platforms/powermac/pci.c hose->mem_resources[cur].start = base; cur 690 arch/powerpc/platforms/powermac/pci.c hose->mem_resources[cur].end = end; cur 691 arch/powerpc/platforms/powermac/pci.c hose->mem_offset[cur] = 0; cur 692 arch/powerpc/platforms/powermac/pci.c DBG(" %d: 0x%08lx-0x%08lx\n", cur, base, end); cur 695 arch/powerpc/platforms/powermac/pci.c hose->mem_resources[cur].end = end; cur 166 arch/powerpc/platforms/powernv/opal-powercap.c u32 cur, min, max; cur 180 arch/powerpc/platforms/powernv/opal-powercap.c if (!of_property_read_u32(node, "powercap-current", &cur)) { cur 214 arch/powerpc/platforms/powernv/opal-powercap.c powercap_add_attr(cur, "powercap-current", cur 83 arch/powerpc/sysdev/xive/common.c u32 cur; cur 87 arch/powerpc/sysdev/xive/common.c cur = be32_to_cpup(q->qpage + q->idx); cur 90 arch/powerpc/sysdev/xive/common.c if ((cur >> 31) == q->toggle) cur 103 arch/powerpc/sysdev/xive/common.c return cur & 0x7fffffff; cur 564 arch/s390/kvm/vsie.c struct vsie_page *cur; cur 583 arch/s390/kvm/vsie.c cur = page_to_virt(page); cur 584 arch/s390/kvm/vsie.c if (READ_ONCE(cur->gmap) != gmap) cur 586 arch/s390/kvm/vsie.c prefix = cur->scb_s.prefix << GUEST_PREFIX_SHIFT; cur 588 arch/s390/kvm/vsie.c prefix += cur->scb_s.mso; cur 590 arch/s390/kvm/vsie.c prefix_unmapped_sync(cur); cur 73 arch/s390/mm/dump_pagetables.c unsigned int prot, cur; cur 82 arch/s390/mm/dump_pagetables.c cur = st->current_prot; cur 90 arch/s390/mm/dump_pagetables.c } else if (prot != cur || level != st->level || cur 241 arch/s390/numa/mode_emu.c struct toptree *cur, *tmp; cur 243 arch/s390/numa/mode_emu.c toptree_for_each_safe(cur, tmp, phys, level) { cur 246 arch/s390/numa/mode_emu.c if (cores_free == toptree_count(cur, CORE)) cur 247 arch/s390/numa/mode_emu.c toptree_move(cur, node); cur 249 arch/s390/numa/mode_emu.c if (cores_free >= toptree_count(cur, CORE)) cur 250 arch/s390/numa/mode_emu.c toptree_move(cur, node); cur 282 arch/s390/numa/toptree.c static struct toptree *toptree_next_sibling(struct toptree *cur) cur 284 arch/s390/numa/toptree.c if (cur->parent == NULL) cur 287 arch/s390/numa/toptree.c if (cur == list_last_entry(&cur->parent->children, cur 290 arch/s390/numa/toptree.c return (struct toptree *) list_next_entry(cur, sibling); cur 304 arch/s390/numa/toptree.c struct toptree *toptree_next(struct toptree *cur, struct toptree *context, cur 309 arch/s390/numa/toptree.c if (!cur) cur 315 arch/s390/numa/toptree.c tmp = toptree_next_sibling(cur); cur 319 arch/s390/numa/toptree.c cur_context = cur; cur 345 arch/s390/numa/toptree.c struct toptree *cur; cur 348 arch/s390/numa/toptree.c toptree_for_each(cur, context, level) cur 33 arch/s390/numa/toptree.h struct toptree *toptree_next(struct toptree *cur, struct toptree *context, cur 365 arch/sh/kernel/kprobes.c struct kprobe *cur = kprobe_running(); cur 370 arch/sh/kernel/kprobes.c if (!cur) cur 373 arch/sh/kernel/kprobes.c if ((kcb->kprobe_status != KPROBE_REENTER) && cur->post_handler) { cur 375 arch/sh/kernel/kprobes.c cur->post_handler(cur, regs, 0); cur 414 arch/sh/kernel/kprobes.c struct kprobe *cur = kprobe_running(); cur 428 arch/sh/kernel/kprobes.c regs->pc = (unsigned long)cur->addr; cur 442 arch/sh/kernel/kprobes.c kprobes_inc_nmissed_count(cur); cur 451 arch/sh/kernel/kprobes.c if (cur->fault_handler && cur->fault_handler(cur, regs, trapnr)) cur 295 arch/sparc/kernel/kprobes.c struct kprobe *cur = kprobe_running(); cur 298 arch/sparc/kernel/kprobes.c if (!cur) cur 301 arch/sparc/kernel/kprobes.c if ((kcb->kprobe_status != KPROBE_REENTER) && cur->post_handler) { cur 303 arch/sparc/kernel/kprobes.c cur->post_handler(cur, regs, 0); cur 306 arch/sparc/kernel/kprobes.c resume_execution(cur, regs, kcb); cur 322 arch/sparc/kernel/kprobes.c struct kprobe *cur = kprobe_running(); cur 336 arch/sparc/kernel/kprobes.c regs->tpc = (unsigned long)cur->addr; cur 353 arch/sparc/kernel/kprobes.c kprobes_inc_nmissed_count(cur); cur 362 arch/sparc/kernel/kprobes.c if (cur->fault_handler && cur->fault_handler(cur, regs, trapnr)) cur 662 arch/x86/events/intel/pt.c buf->first = buf->last = buf->cur = topa; cur 784 arch/x86/events/intel/pt.c if (buf->cur_idx == buf->cur->last) { cur 785 arch/x86/events/intel/pt.c if (buf->cur == buf->last) cur 786 arch/x86/events/intel/pt.c buf->cur = buf->first; cur 788 arch/x86/events/intel/pt.c buf->cur = list_entry(buf->cur->list.next, struct topa, cur 806 arch/x86/events/intel/pt.c base = buf->cur->offset + buf->output_off; cur 810 arch/x86/events/intel/pt.c base += TOPA_ENTRY_SIZE(buf->cur, topa_idx); cur 830 arch/x86/events/intel/pt.c return phys_to_virt(TOPA_ENTRY(buf->cur, buf->cur_idx)->base << TOPA_SHIFT); cur 839 arch/x86/events/intel/pt.c return TOPA_ENTRY_SIZE(buf->cur, buf->cur_idx); cur 911 arch/x86/events/intel/pt.c buf->cur = &tp->topa; cur 1118 arch/x86/events/intel/pt.c buf->cur = &cur_tp->topa; cur 1119 arch/x86/events/intel/pt.c buf->cur_idx = te - TOPA_ENTRY(buf->cur, 0); cur 1382 arch/x86/events/intel/pt.c pt_config_buffer(topa_to_page(buf->cur)->table, buf->cur_idx, cur 1447 arch/x86/events/intel/pt.c pt_config_buffer(topa_to_page(buf->cur)->table, buf->cur_idx, cur 76 arch/x86/events/intel/pt.h struct topa *first, *last, *cur; cur 30 arch/x86/hyperv/mmu.c unsigned long cur = start, diff; cur 33 arch/x86/hyperv/mmu.c diff = end > cur ? end - cur : 0; cur 35 arch/x86/hyperv/mmu.c gva_list[gva_n] = cur & PAGE_MASK; cur 42 arch/x86/hyperv/mmu.c cur += HV_TLB_FLUSH_UNIT; cur 45 arch/x86/hyperv/mmu.c cur = end; cur 50 arch/x86/hyperv/mmu.c } while (cur < end); cur 63 arch/x86/hyperv/nested.c u64 cur = start_gfn; cur 79 arch/x86/hyperv/nested.c flush->gpa_list[gpa_n].page.basepfn = cur; cur 82 arch/x86/hyperv/nested.c cur += additional_pages + 1; cur 66 arch/x86/ia32/ia32_signal.c unsigned int cur = get_user_seg(seg); \ cur 67 arch/x86/ia32/ia32_signal.c if (pre != cur) \ cur 976 arch/x86/kernel/kprobes/core.c struct kprobe *cur = kprobe_running(); cur 979 arch/x86/kernel/kprobes/core.c if (!cur) cur 982 arch/x86/kernel/kprobes/core.c resume_execution(cur, regs, kcb); cur 985 arch/x86/kernel/kprobes/core.c if ((kcb->kprobe_status != KPROBE_REENTER) && cur->post_handler) { cur 987 arch/x86/kernel/kprobes/core.c cur->post_handler(cur, regs, 0); cur 1011 arch/x86/kernel/kprobes/core.c struct kprobe *cur = kprobe_running(); cur 1014 arch/x86/kernel/kprobes/core.c if (unlikely(regs->ip == (unsigned long)cur->ainsn.insn)) { cur 1025 arch/x86/kernel/kprobes/core.c regs->ip = (unsigned long)cur->addr; cur 1050 arch/x86/kernel/kprobes/core.c kprobes_inc_nmissed_count(cur); cur 1059 arch/x86/kernel/kprobes/core.c if (cur->fault_handler && cur->fault_handler(cur, regs, trapnr)) cur 84 arch/x86/kernel/reboot_fixups_32.c const struct device_fixup *cur; cur 94 arch/x86/kernel/reboot_fixups_32.c cur = &(fixups_table[i]); cur 95 arch/x86/kernel/reboot_fixups_32.c dev = pci_get_device(cur->vendor, cur->device, NULL); cur 99 arch/x86/kernel/reboot_fixups_32.c cur->reboot_fixup(dev); cur 80 arch/x86/kernel/tsc_sync.c static void tsc_sanitize_first_cpu(struct tsc_adjust *cur, s64 bootval, cur 110 arch/x86/kernel/tsc_sync.c cur->adjusted = bootval; cur 116 arch/x86/kernel/tsc_sync.c struct tsc_adjust *cur = this_cpu_ptr(&tsc_adjust); cur 127 arch/x86/kernel/tsc_sync.c cur->bootval = bootval; cur 128 arch/x86/kernel/tsc_sync.c cur->nextcheck = jiffies + HZ; cur 129 arch/x86/kernel/tsc_sync.c tsc_sanitize_first_cpu(cur, bootval, smp_processor_id(), bootcpu); cur 140 arch/x86/kernel/tsc_sync.c struct tsc_adjust *ref, *cur = this_cpu_ptr(&tsc_adjust); cur 149 arch/x86/kernel/tsc_sync.c cur->bootval = bootval; cur 150 arch/x86/kernel/tsc_sync.c cur->nextcheck = jiffies + HZ; cur 151 arch/x86/kernel/tsc_sync.c cur->warned = false; cur 158 arch/x86/kernel/tsc_sync.c cur->adjusted = bootval; cur 171 arch/x86/kernel/tsc_sync.c tsc_sanitize_first_cpu(cur, bootval, smp_processor_id(), cur 191 arch/x86/kernel/tsc_sync.c cur->adjusted = ref->adjusted; cur 402 arch/x86/kernel/tsc_sync.c struct tsc_adjust *cur = this_cpu_ptr(&tsc_adjust); cur 484 arch/x86/kernel/tsc_sync.c cur->adjusted += cur_max_warp; cur 487 arch/x86/kernel/tsc_sync.c cpu, cur_max_warp, cur->adjusted); cur 489 arch/x86/kernel/tsc_sync.c wrmsrl(MSR_IA32_TSC_ADJUST, cur->adjusted); cur 347 arch/x86/kvm/mtrr.c struct kvm_mtrr_range *tmp, *cur; cur 352 arch/x86/kvm/mtrr.c cur = &mtrr_state->var_ranges[index]; cur 355 arch/x86/kvm/mtrr.c if (var_mtrr_range_is_valid(cur)) cur 363 arch/x86/kvm/mtrr.c cur->base = data; cur 365 arch/x86/kvm/mtrr.c cur->mask = data | (-1LL << cpuid_maxphyaddr(vcpu)); cur 368 arch/x86/kvm/mtrr.c if (var_mtrr_range_is_valid(cur)) { cur 370 arch/x86/kvm/mtrr.c if (cur->base >= tmp->base) cur 372 arch/x86/kvm/mtrr.c list_add_tail(&cur->node, &tmp->node); cur 5218 arch/x86/kvm/svm.c struct amd_svm_iommu_ir *cur; cur 5221 arch/x86/kvm/svm.c list_for_each_entry(cur, &svm->ir_list, node) { cur 5222 arch/x86/kvm/svm.c if (cur->data != pi->ir_data) cur 5224 arch/x86/kvm/svm.c list_del(&cur->node); cur 5225 arch/x86/kvm/svm.c kfree(cur); cur 271 arch/x86/mm/dump_pagetables.c pgprotval_t prot, cur, eff; cur 280 arch/x86/mm/dump_pagetables.c cur = pgprot_val(st->current_prot); cur 292 arch/x86/mm/dump_pagetables.c } else if (prot != cur || new_eff != eff || level != st->level || cur 1658 block/blk-cgroup.c u64 cur = atomic64_read(&blkg->delay_nsec); cur 1675 block/blk-cgroup.c if (unlikely(cur < sub)) { cur 1680 block/blk-cgroup.c blkg->last_delay = cur - sub; cur 11 block/blk-rq-qos.c unsigned int cur = atomic_read(v); cur 16 block/blk-rq-qos.c if (cur >= below) cur 18 block/blk-rq-qos.c old = atomic_cmpxchg(v, cur, cur + 1); cur 19 block/blk-rq-qos.c if (old == cur) cur 21 block/blk-rq-qos.c cur = old; cur 111 block/blk-rq-qos.h struct rq_qos **cur; cur 113 block/blk-rq-qos.h for (cur = &q->rq_qos; *cur; cur = &(*cur)->next) { cur 114 block/blk-rq-qos.h if (*cur == rqos) { cur 115 block/blk-rq-qos.h *cur = rqos->next; cur 86 block/blk-wbt.c const unsigned long cur = jiffies; cur 88 block/blk-wbt.c if (cur != *var) cur 89 block/blk-wbt.c *var = cur; cur 497 crypto/af_alg.c if (!sg || sgl->cur >= MAX_SGL_ENTS) { cur 505 crypto/af_alg.c sgl->cur = 0; cur 541 crypto/af_alg.c for (i = 0; i < sgl->cur; i++) { cur 597 crypto/af_alg.c for (i = 0; i < sgl->cur; i++) { cur 872 crypto/af_alg.c sg = sgl->sg + sgl->cur - 1; cur 908 crypto/af_alg.c if (sgl->cur) cur 909 crypto/af_alg.c sg_unmark_end(sg + sgl->cur - 1); cur 912 crypto/af_alg.c unsigned int i = sgl->cur; cur 935 crypto/af_alg.c sgl->cur++; cur 936 crypto/af_alg.c } while (len && sgl->cur < MAX_SGL_ENTS); cur 939 crypto/af_alg.c sg_mark_end(sg + sgl->cur - 1); cur 993 crypto/af_alg.c if (sgl->cur) cur 994 crypto/af_alg.c sg_unmark_end(sgl->sg + sgl->cur - 1); cur 996 crypto/af_alg.c sg_mark_end(sgl->sg + sgl->cur); cur 999 crypto/af_alg.c sg_set_page(sgl->sg + sgl->cur, page, size, offset); cur 1000 crypto/af_alg.c sgl->cur++; cur 183 crypto/algif_aead.c for (i = 0; i < tsgl->cur; i++) { cur 357 crypto/asymmetric_keys/asym_tpm.c uint8_t *cur = buf; cur 363 crypto/asymmetric_keys/asym_tpm.c cur = encode_tag_length(cur, 0x30, n_len + e_len); cur 365 crypto/asymmetric_keys/asym_tpm.c cur = encode_tag_length(cur, 0x02, len + 1); cur 366 crypto/asymmetric_keys/asym_tpm.c cur[0] = 0x00; cur 367 crypto/asymmetric_keys/asym_tpm.c memcpy(cur + 1, pub_key, len); cur 368 crypto/asymmetric_keys/asym_tpm.c cur += len + 1; cur 369 crypto/asymmetric_keys/asym_tpm.c cur = encode_tag_length(cur, 0x02, sizeof(e)); cur 370 crypto/asymmetric_keys/asym_tpm.c memcpy(cur, e, sizeof(e)); cur 371 crypto/asymmetric_keys/asym_tpm.c cur += sizeof(e); cur 373 crypto/asymmetric_keys/asym_tpm.c memset(cur, 0, SETKEY_PARAMS_SIZE); cur 375 crypto/asymmetric_keys/asym_tpm.c return cur - buf; cur 832 crypto/asymmetric_keys/asym_tpm.c const void *cur = tk->blob; cur 842 crypto/asymmetric_keys/asym_tpm.c if (get_unaligned_be16(cur + 4) != 0x0015) cur 846 crypto/asymmetric_keys/asym_tpm.c cur += 11; cur 853 crypto/asymmetric_keys/asym_tpm.c if (get_unaligned_be32(cur) != 0x00000001) cur 857 crypto/asymmetric_keys/asym_tpm.c if (get_unaligned_be16(cur + 4) != 0x0002) cur 861 crypto/asymmetric_keys/asym_tpm.c if (get_unaligned_be16(cur + 6) != 0x0003) cur 864 crypto/asymmetric_keys/asym_tpm.c sz = get_unaligned_be32(cur + 8); cur 870 crypto/asymmetric_keys/asym_tpm.c cur += 12; cur 873 crypto/asymmetric_keys/asym_tpm.c key_len = get_unaligned_be32(cur); cur 886 crypto/asymmetric_keys/asym_tpm.c cur += sz; cur 892 crypto/asymmetric_keys/asym_tpm.c sz = get_unaligned_be32(cur); cur 897 crypto/asymmetric_keys/asym_tpm.c cur += 4 + sz; cur 901 crypto/asymmetric_keys/asym_tpm.c sz = get_unaligned_be32(cur); cur 905 crypto/asymmetric_keys/asym_tpm.c pub_key = cur + 4; cur 376 crypto/tcrypt.c struct test_mb_aead_data *cur = &data[j]; cur 378 crypto/tcrypt.c assoc = cur->axbuf[0]; cur 381 crypto/tcrypt.c sg_init_aead(cur->sg, cur->xbuf, cur 385 crypto/tcrypt.c sg_init_aead(cur->sgout, cur->xoutbuf, cur 389 crypto/tcrypt.c aead_request_set_ad(cur->req, aad_size); cur 393 crypto/tcrypt.c aead_request_set_crypt(cur->req, cur 394 crypto/tcrypt.c cur->sgout, cur 395 crypto/tcrypt.c cur->sg, cur 397 crypto/tcrypt.c ret = crypto_aead_encrypt(cur->req); cur 398 crypto/tcrypt.c ret = do_one_aead_op(cur->req, ret); cur 407 crypto/tcrypt.c aead_request_set_crypt(cur->req, cur->sg, cur 408 crypto/tcrypt.c cur->sgout, *b_size + cur 1354 crypto/tcrypt.c struct test_mb_skcipher_data *cur = &data[j]; cur 1359 crypto/tcrypt.c sg_init_table(cur->sg, pages); cur 1362 crypto/tcrypt.c sg_set_buf(cur->sg + p, cur->xbuf[p], cur 1364 crypto/tcrypt.c memset(cur->xbuf[p], 0xff, PAGE_SIZE); cur 1369 crypto/tcrypt.c sg_set_buf(cur->sg + p, cur->xbuf[p], k); cur 1370 crypto/tcrypt.c memset(cur->xbuf[p], 0xff, k); cur 1372 crypto/tcrypt.c skcipher_request_set_crypt(cur->req, cur->sg, cur 1373 crypto/tcrypt.c cur->sg, *b_size, cur 204 drivers/android/binder.c unsigned int cur = atomic_inc_return(&log->cur); cur 206 drivers/android/binder.c if (cur >= ARRAY_SIZE(log->entry)) cur 208 drivers/android/binder.c e = &log->entry[cur % ARRAY_SIZE(log->entry)]; cur 6048 drivers/android/binder.c unsigned int log_cur = atomic_read(&log->cur); cur 6050 drivers/android/binder.c unsigned int cur; cur 6054 drivers/android/binder.c cur = count < ARRAY_SIZE(log->entry) && !log->full ? cur 6059 drivers/android/binder.c unsigned int index = cur++ % ARRAY_SIZE(log->entry); cur 6118 drivers/android/binder.c atomic_set(&binder_transaction_log.cur, ~0U); cur 6119 drivers/android/binder.c atomic_set(&binder_transaction_log_failed.cur, ~0U); cur 139 drivers/android/binder_internal.h atomic_t cur; cur 3803 drivers/ata/libata-core.c u32 last, cur; cur 3810 drivers/ata/libata-core.c if ((rc = sata_scr_read(link, SCR_STATUS, &cur))) cur 3812 drivers/ata/libata-core.c cur &= 0xf; cur 3814 drivers/ata/libata-core.c last = cur; cur 3819 drivers/ata/libata-core.c if ((rc = sata_scr_read(link, SCR_STATUS, &cur))) cur 3821 drivers/ata/libata-core.c cur &= 0xf; cur 3824 drivers/ata/libata-core.c if (cur == last) { cur 3825 drivers/ata/libata-core.c if (cur == 1 && time_before(jiffies, deadline)) cur 3834 drivers/ata/libata-core.c last = cur; cur 6328 drivers/ata/libata-core.c const struct ata_port_operations *cur; cur 6338 drivers/ata/libata-core.c for (cur = ops->inherits; cur; cur = cur->inherits) { cur 6339 drivers/ata/libata-core.c void **inherit = (void **)cur; cur 6895 drivers/ata/libata-core.c static int __init ata_parse_force_one(char **cur, cur 6954 drivers/ata/libata-core.c char *start = *cur, *p = *cur; cur 6964 drivers/ata/libata-core.c *cur = p; cur 6966 drivers/ata/libata-core.c *cur = p + 1; cur 7033 drivers/ata/libata-core.c char *p, *cur, *next; cur 7048 drivers/ata/libata-core.c for (cur = ata_force_param_buf; *cur != '\0'; cur = next) { cur 7052 drivers/ata/libata-core.c next = cur; cur 7056 drivers/ata/libata-core.c cur, reason); cur 299 drivers/ata/libata-eh.c const u8 *cur; cur 301 drivers/ata/libata-eh.c for (cur = ata_eh_cmd_timeout_table[i].commands; *cur; cur++) cur 302 drivers/ata/libata-eh.c if (*cur == cmd) cur 426 drivers/base/devres.c struct list_head *cur; cur 431 drivers/base/devres.c cur = first; cur 432 drivers/base/devres.c while (cur != end) { cur 436 drivers/base/devres.c node = list_entry(cur, struct devres_node, entry); cur 437 drivers/base/devres.c cur = cur->next; cur 462 drivers/base/devres.c cur = first; cur 463 drivers/base/devres.c while (cur != end) { cur 467 drivers/base/devres.c node = list_entry(cur, struct devres_node, entry); cur 468 drivers/base/devres.c cur = cur->next; cur 705 drivers/base/regmap/regcache.c unsigned int base, unsigned int cur) cur 713 drivers/base/regmap/regcache.c count = (cur - base) / map->reg_stride; cur 716 drivers/base/regmap/regcache.c count * val_bytes, count, base, cur - map->reg_stride); cur 723 drivers/base/regmap/regcache.c base, cur - map->reg_stride, ret); cur 1184 drivers/block/drbd/drbd_main.c len = bs.cur.b - p->code + !!bs.cur.bit; cur 1200 drivers/block/drbd/drbd_main.c dcbp_set_pad_bits(p, (8 - bs.cur.bit) & 0x7); cur 4769 drivers/block/drbd/drbd_receiver.c (unsigned int)(bs.cur.b - p->code), cur 199 drivers/block/drbd/drbd_vli.h static inline void bitstream_cursor_reset(struct bitstream_cursor *cur, void *s) cur 201 drivers/block/drbd/drbd_vli.h cur->b = s; cur 202 drivers/block/drbd/drbd_vli.h cur->bit = 0; cur 207 drivers/block/drbd/drbd_vli.h static inline void bitstream_cursor_advance(struct bitstream_cursor *cur, unsigned int bits) cur 209 drivers/block/drbd/drbd_vli.h bits += cur->bit; cur 210 drivers/block/drbd/drbd_vli.h cur->b = cur->b + (bits >> 3); cur 211 drivers/block/drbd/drbd_vli.h cur->bit = bits & 7; cur 216 drivers/block/drbd/drbd_vli.h struct bitstream_cursor cur; cur 231 drivers/block/drbd/drbd_vli.h bitstream_cursor_reset(&bs->cur, bs->buf); cur 236 drivers/block/drbd/drbd_vli.h bitstream_cursor_reset(&bs->cur, bs->buf); cur 250 drivers/block/drbd/drbd_vli.h unsigned char *b = bs->cur.b; cur 256 drivers/block/drbd/drbd_vli.h if ((bs->cur.b + ((bs->cur.bit + bits -1) >> 3)) - bs->buf >= bs->buf_len) cur 263 drivers/block/drbd/drbd_vli.h *b++ |= (val & 0xff) << bs->cur.bit; cur 265 drivers/block/drbd/drbd_vli.h for (tmp = 8 - bs->cur.bit; tmp < bits; tmp += 8) cur 268 drivers/block/drbd/drbd_vli.h bitstream_cursor_advance(&bs->cur, bits); cur 289 drivers/block/drbd/drbd_vli.h if (bs->cur.b + ((bs->cur.bit + bs->pad_bits + bits -1) >> 3) - bs->buf >= bs->buf_len) cur 290 drivers/block/drbd/drbd_vli.h bits = ((bs->buf_len - (bs->cur.b - bs->buf)) << 3) cur 291 drivers/block/drbd/drbd_vli.h - bs->cur.bit - bs->pad_bits; cur 300 drivers/block/drbd/drbd_vli.h n = (bs->cur.bit + bits + 7) >> 3; cur 304 drivers/block/drbd/drbd_vli.h memcpy(&val, bs->cur.b+1, n - 1); cur 305 drivers/block/drbd/drbd_vli.h val = le64_to_cpu(val) << (8 - bs->cur.bit); cur 309 drivers/block/drbd/drbd_vli.h val |= bs->cur.b[0] >> bs->cur.bit; cur 314 drivers/block/drbd/drbd_vli.h bitstream_cursor_advance(&bs->cur, bits); cur 2855 drivers/block/mtip32xx/mtip32xx.c attr242.cur, le32_to_cpu(attr242.data)); cur 166 drivers/block/mtip32xx/mtip32xx.h u8 cur; cur 25 drivers/char/agp/isoch.c struct agp_3_5_dev *cur, *n = list_entry(new, struct agp_3_5_dev, list); cur 29 drivers/char/agp/isoch.c cur = list_entry(pos, struct agp_3_5_dev, list); cur 30 drivers/char/agp/isoch.c if (cur->maxbw > n->maxbw) cur 38 drivers/char/agp/isoch.c struct agp_3_5_dev *cur; cur 46 drivers/char/agp/isoch.c cur = list_entry(pos, struct agp_3_5_dev, list); cur 47 drivers/char/agp/isoch.c dev = cur->dev; cur 49 drivers/char/agp/isoch.c pci_read_config_dword(dev, cur->capndx+AGPNISTAT, &nistat); cur 50 drivers/char/agp/isoch.c cur->maxbw = (nistat >> 16) & 0xff; cur 82 drivers/char/agp/isoch.c struct agp_3_5_dev *cur; cur 138 drivers/char/agp/isoch.c cur = list_entry(pos, struct agp_3_5_dev, list); cur 139 drivers/char/agp/isoch.c dev = cur->dev; cur 141 drivers/char/agp/isoch.c mcapndx = cur->capndx; cur 143 drivers/char/agp/isoch.c pci_read_config_dword(dev, cur->capndx+AGPNISTAT, &mnistat); cur 148 drivers/char/agp/isoch.c master[cdev].dev = cur; cur 251 drivers/char/agp/isoch.c cur = master[cdev].dev; cur 252 drivers/char/agp/isoch.c dev = cur->dev; cur 254 drivers/char/agp/isoch.c mcapndx = cur->capndx; cur 259 drivers/char/agp/isoch.c pci_read_config_word(dev, cur->capndx+AGPNICMD, &mnicmd); cur 260 drivers/char/agp/isoch.c pci_read_config_dword(dev, cur->capndx+AGPCMD, &mcmd); cur 270 drivers/char/agp/isoch.c pci_write_config_dword(dev, cur->capndx+AGPCMD, mcmd); cur 271 drivers/char/agp/isoch.c pci_write_config_word(dev, cur->capndx+AGPNICMD, mnicmd); cur 291 drivers/char/agp/isoch.c struct agp_3_5_dev *cur; cur 305 drivers/char/agp/isoch.c cur = list_entry(pos, struct agp_3_5_dev, list); cur 307 drivers/char/agp/isoch.c pci_read_config_dword(cur->dev, cur->capndx+AGPCMD, &mcmd); cur 310 drivers/char/agp/isoch.c pci_write_config_dword(cur->dev, cur->capndx+AGPCMD, mcmd); cur 326 drivers/char/agp/isoch.c struct agp_3_5_dev *dev_list, *cur; cur 372 drivers/char/agp/isoch.c if ((cur = kmalloc(sizeof(*cur), GFP_KERNEL)) == NULL) { cur 376 drivers/char/agp/isoch.c cur->dev = dev; cur 378 drivers/char/agp/isoch.c pos = &cur->list; cur 395 drivers/char/agp/isoch.c cur = list_entry(pos, struct agp_3_5_dev, list); cur 396 drivers/char/agp/isoch.c dev = cur->dev; cur 429 drivers/char/agp/isoch.c cur->capndx = mcapndx; cur 431 drivers/char/agp/isoch.c pci_read_config_dword(dev, cur->capndx+AGPSTAT, &mstatus); cur 463 drivers/char/agp/isoch.c cur = list_entry(pos, struct agp_3_5_dev, list); cur 466 drivers/char/agp/isoch.c kfree(cur); cur 155 drivers/char/bsr.c struct bsr_dev *cur, *n; cur 157 drivers/char/bsr.c list_for_each_entry_safe(cur, n, &bsr_devs, bsr_list) { cur 158 drivers/char/bsr.c if (cur->bsr_device) { cur 159 drivers/char/bsr.c cdev_del(&cur->bsr_cdev); cur 160 drivers/char/bsr.c device_del(cur->bsr_device); cur 162 drivers/char/bsr.c list_del(&cur->bsr_list); cur 163 drivers/char/bsr.c kfree(cur); cur 187 drivers/char/bsr.c struct bsr_dev *cur = kzalloc(sizeof(struct bsr_dev), cur 192 drivers/char/bsr.c if (!cur) { cur 201 drivers/char/bsr.c kfree(cur); cur 205 drivers/char/bsr.c cur->bsr_minor = i + total_bsr_devs; cur 206 drivers/char/bsr.c cur->bsr_addr = res.start; cur 207 drivers/char/bsr.c cur->bsr_len = resource_size(&res); cur 208 drivers/char/bsr.c cur->bsr_bytes = bsr_bytes[i]; cur 209 drivers/char/bsr.c cur->bsr_stride = bsr_stride[i]; cur 210 drivers/char/bsr.c cur->bsr_dev = MKDEV(bsr_major, i + total_bsr_devs); cur 214 drivers/char/bsr.c if (cur->bsr_len > 4096 && cur->bsr_len < PAGE_SIZE) cur 215 drivers/char/bsr.c cur->bsr_len = 4096; cur 217 drivers/char/bsr.c switch(cur->bsr_bytes) { cur 219 drivers/char/bsr.c cur->bsr_type = BSR_8; cur 222 drivers/char/bsr.c cur->bsr_type = BSR_16; cur 225 drivers/char/bsr.c cur->bsr_type = BSR_64; cur 228 drivers/char/bsr.c cur->bsr_type = BSR_128; cur 231 drivers/char/bsr.c cur->bsr_type = BSR_4096; cur 234 drivers/char/bsr.c cur->bsr_type = BSR_UNKNOWN; cur 237 drivers/char/bsr.c cur->bsr_num = bsr_types[cur->bsr_type]; cur 238 drivers/char/bsr.c snprintf(cur->bsr_name, 32, "bsr%d_%d", cur 239 drivers/char/bsr.c cur->bsr_bytes, cur->bsr_num); cur 241 drivers/char/bsr.c cdev_init(&cur->bsr_cdev, &bsr_fops); cur 242 drivers/char/bsr.c result = cdev_add(&cur->bsr_cdev, cur->bsr_dev, 1); cur 244 drivers/char/bsr.c kfree(cur); cur 248 drivers/char/bsr.c cur->bsr_device = device_create(bsr_class, NULL, cur->bsr_dev, cur 249 drivers/char/bsr.c cur, "%s", cur->bsr_name); cur 250 drivers/char/bsr.c if (IS_ERR(cur->bsr_device)) { cur 252 drivers/char/bsr.c cur->bsr_name); cur 253 drivers/char/bsr.c cdev_del(&cur->bsr_cdev); cur 254 drivers/char/bsr.c kfree(cur); cur 258 drivers/char/bsr.c bsr_types[cur->bsr_type] = cur->bsr_num + 1; cur 259 drivers/char/bsr.c list_add_tail(&cur->bsr_list, &bsr_devs); cur 80 drivers/clk/clk-conf.c const __be32 *cur; cur 85 drivers/clk/clk-conf.c of_property_for_each_u32(node, "assigned-clock-rates", prop, cur, rate) { cur 4826 drivers/clk/clk.c const __be32 *cur; cur 4832 drivers/clk/clk.c of_property_for_each_u32(np, "clock-critical", prop, cur, idx) cur 156 drivers/clk/rockchip/clk-pll.c struct rockchip_pll_rate_table cur; cur 159 drivers/clk/rockchip/clk-pll.c rockchip_rk3036_pll_get_params(pll, &cur); cur 161 drivers/clk/rockchip/clk-pll.c rate64 *= cur.fbdiv; cur 162 drivers/clk/rockchip/clk-pll.c do_div(rate64, cur.refdiv); cur 164 drivers/clk/rockchip/clk-pll.c if (cur.dsmpd == 0) { cur 166 drivers/clk/rockchip/clk-pll.c u64 frac_rate64 = prate * cur.frac; cur 168 drivers/clk/rockchip/clk-pll.c do_div(frac_rate64, cur.refdiv); cur 172 drivers/clk/rockchip/clk-pll.c do_div(rate64, cur.postdiv1); cur 173 drivers/clk/rockchip/clk-pll.c do_div(rate64, cur.postdiv2); cur 183 drivers/clk/rockchip/clk-pll.c struct rockchip_pll_rate_table cur; cur 193 drivers/clk/rockchip/clk-pll.c rockchip_rk3036_pll_get_params(pll, &cur); cur 194 drivers/clk/rockchip/clk-pll.c cur.rate = 0; cur 228 drivers/clk/rockchip/clk-pll.c rockchip_rk3036_pll_set_params(pll, &cur); cur 289 drivers/clk/rockchip/clk-pll.c struct rockchip_pll_rate_table cur; cur 302 drivers/clk/rockchip/clk-pll.c rockchip_rk3036_pll_get_params(pll, &cur); cur 307 drivers/clk/rockchip/clk-pll.c cur.fbdiv, cur.postdiv1, cur.refdiv, cur.postdiv2, cur 308 drivers/clk/rockchip/clk-pll.c cur.dsmpd, cur.frac); cur 313 drivers/clk/rockchip/clk-pll.c if (rate->fbdiv != cur.fbdiv || rate->postdiv1 != cur.postdiv1 || cur 314 drivers/clk/rockchip/clk-pll.c rate->refdiv != cur.refdiv || rate->postdiv2 != cur.postdiv2 || cur 315 drivers/clk/rockchip/clk-pll.c rate->dsmpd != cur.dsmpd || cur 316 drivers/clk/rockchip/clk-pll.c (!cur.dsmpd && (rate->frac != cur.frac))) { cur 391 drivers/clk/rockchip/clk-pll.c struct rockchip_pll_rate_table cur; cur 402 drivers/clk/rockchip/clk-pll.c rockchip_rk3066_pll_get_params(pll, &cur); cur 404 drivers/clk/rockchip/clk-pll.c rate64 *= cur.nf; cur 405 drivers/clk/rockchip/clk-pll.c do_div(rate64, cur.nr); cur 406 drivers/clk/rockchip/clk-pll.c do_div(rate64, cur.no); cur 416 drivers/clk/rockchip/clk-pll.c struct rockchip_pll_rate_table cur; cur 424 drivers/clk/rockchip/clk-pll.c rockchip_rk3066_pll_get_params(pll, &cur); cur 425 drivers/clk/rockchip/clk-pll.c cur.rate = 0; cur 461 drivers/clk/rockchip/clk-pll.c rockchip_rk3066_pll_set_params(pll, &cur); cur 522 drivers/clk/rockchip/clk-pll.c struct rockchip_pll_rate_table cur; cur 535 drivers/clk/rockchip/clk-pll.c rockchip_rk3066_pll_get_params(pll, &cur); cur 538 drivers/clk/rockchip/clk-pll.c __func__, clk_hw_get_name(hw), drate, rate->nr, cur.nr, cur 539 drivers/clk/rockchip/clk-pll.c rate->no, cur.no, rate->nf, cur.nf, rate->nb, cur.nb); cur 540 drivers/clk/rockchip/clk-pll.c if (rate->nr != cur.nr || rate->no != cur.no || rate->nf != cur.nf cur 541 drivers/clk/rockchip/clk-pll.c || rate->nb != cur.nb) { cur 633 drivers/clk/rockchip/clk-pll.c struct rockchip_pll_rate_table cur; cur 636 drivers/clk/rockchip/clk-pll.c rockchip_rk3399_pll_get_params(pll, &cur); cur 638 drivers/clk/rockchip/clk-pll.c rate64 *= cur.fbdiv; cur 639 drivers/clk/rockchip/clk-pll.c do_div(rate64, cur.refdiv); cur 641 drivers/clk/rockchip/clk-pll.c if (cur.dsmpd == 0) { cur 643 drivers/clk/rockchip/clk-pll.c u64 frac_rate64 = prate * cur.frac; cur 645 drivers/clk/rockchip/clk-pll.c do_div(frac_rate64, cur.refdiv); cur 649 drivers/clk/rockchip/clk-pll.c do_div(rate64, cur.postdiv1); cur 650 drivers/clk/rockchip/clk-pll.c do_div(rate64, cur.postdiv2); cur 660 drivers/clk/rockchip/clk-pll.c struct rockchip_pll_rate_table cur; cur 670 drivers/clk/rockchip/clk-pll.c rockchip_rk3399_pll_get_params(pll, &cur); cur 671 drivers/clk/rockchip/clk-pll.c cur.rate = 0; cur 707 drivers/clk/rockchip/clk-pll.c rockchip_rk3399_pll_set_params(pll, &cur); cur 768 drivers/clk/rockchip/clk-pll.c struct rockchip_pll_rate_table cur; cur 781 drivers/clk/rockchip/clk-pll.c rockchip_rk3399_pll_get_params(pll, &cur); cur 786 drivers/clk/rockchip/clk-pll.c cur.fbdiv, cur.postdiv1, cur.refdiv, cur.postdiv2, cur 787 drivers/clk/rockchip/clk-pll.c cur.dsmpd, cur.frac); cur 792 drivers/clk/rockchip/clk-pll.c if (rate->fbdiv != cur.fbdiv || rate->postdiv1 != cur.postdiv1 || cur 793 drivers/clk/rockchip/clk-pll.c rate->refdiv != cur.refdiv || rate->postdiv2 != cur.postdiv2 || cur 794 drivers/clk/rockchip/clk-pll.c rate->dsmpd != cur.dsmpd || cur 795 drivers/clk/rockchip/clk-pll.c (!cur.dsmpd && (rate->frac != cur.frac))) { cur 423 drivers/clocksource/samsung_pwm_timer.c const __be32 *cur; cur 431 drivers/clocksource/samsung_pwm_timer.c of_property_for_each_u32(np, "samsung,pwm-outputs", prop, cur, val) { cur 795 drivers/cpufreq/acpi-cpufreq.c policy->cur = acpi_cpufreq_guess_freq(data, policy->cpu); cur 62 drivers/cpufreq/amd_freq_sensitivity.c freq_next = policy->cur; cur 71 drivers/cpufreq/amd_freq_sensitivity.c freq_next = policy->cur; cur 82 drivers/cpufreq/amd_freq_sensitivity.c if (data->freq_prev == policy->cur) cur 83 drivers/cpufreq/amd_freq_sensitivity.c freq_next = policy->cur; cur 85 drivers/cpufreq/amd_freq_sensitivity.c if (freq_next > policy->cur) cur 86 drivers/cpufreq/amd_freq_sensitivity.c freq_next = policy->cur; cur 87 drivers/cpufreq/amd_freq_sensitivity.c else if (freq_next < policy->cur) cur 93 drivers/cpufreq/amd_freq_sensitivity.c policy->cur - 1); cur 600 drivers/cpufreq/brcmstb-avs-cpufreq.c policy->cur = freq_table[pstate].frequency; cur 206 drivers/cpufreq/cppc_cpufreq.c freqs.old = policy->cur; cur 343 drivers/cpufreq/cppc_cpufreq.c policy->cur = cppc_cpufreq_perf_to_khz(cpu, cur 361 drivers/cpufreq/cpufreq.c if (policy->cur && policy->cur != freqs->old) { cur 363 drivers/cpufreq/cpufreq.c freqs->old, policy->cur); cur 364 drivers/cpufreq/cpufreq.c freqs->old = policy->cur; cur 385 drivers/cpufreq/cpufreq.c policy->cur = freqs->new; cur 700 drivers/cpufreq/cpufreq.c ret = sprintf(buf, "%u\n", policy->cur); cur 1411 drivers/cpufreq/cpufreq.c policy->cur = cpufreq_driver->get(policy->cpu); cur 1412 drivers/cpufreq/cpufreq.c if (!policy->cur) { cur 1439 drivers/cpufreq/cpufreq.c ret = cpufreq_frequency_table_get_index(policy, policy->cur); cur 1443 drivers/cpufreq/cpufreq.c __func__, policy->cpu, policy->cur); cur 1444 drivers/cpufreq/cpufreq.c ret = __cpufreq_driver_target(policy, policy->cur - 1, cur 1454 drivers/cpufreq/cpufreq.c __func__, policy->cpu, policy->cur); cur 1645 drivers/cpufreq/cpufreq.c policy->cur, new_freq); cur 1647 drivers/cpufreq/cpufreq.c freqs.old = policy->cur; cur 1669 drivers/cpufreq/cpufreq.c if (policy->cur != new_freq) { cur 1703 drivers/cpufreq/cpufreq.c ret_freq = policy->cur; cur 2058 drivers/cpufreq/cpufreq.c struct cpufreq_freqs freqs = {.old = policy->cur, .flags = 0}; cur 2064 drivers/cpufreq/cpufreq.c if (newfreq == policy->cur) cur 2135 drivers/cpufreq/cpufreq.c if (target_freq == policy->cur) cur 2139 drivers/cpufreq/cpufreq.c policy->restore_freq = policy->cur; cur 81 drivers/cpufreq/cpufreq_conservative.c requested_freq = policy->cur; cur 311 drivers/cpufreq/cpufreq_conservative.c dbs_info->requested_freq = policy->cur; cur 513 drivers/cpufreq/cpufreq_governor.c if (!policy->cur) cur 122 drivers/cpufreq/cpufreq_ondemand.c else if (policy->cur == policy->max) cur 147 drivers/cpufreq/cpufreq_ondemand.c if (policy->cur < policy->max) cur 210 drivers/cpufreq/cpufreq_stats.c stats->last_index = freq_table_get_index(stats, policy->cur); cur 49 drivers/cpufreq/cpufreq_userspace.c return sprintf(buf, "%u\n", policy->cur); cur 76 drivers/cpufreq/cpufreq_userspace.c BUG_ON(!policy->cur); cur 81 drivers/cpufreq/cpufreq_userspace.c *setspeed = policy->cur; cur 105 drivers/cpufreq/cpufreq_userspace.c policy->cpu, policy->min, policy->max, policy->cur, *setspeed); cur 42 drivers/cpufreq/davinci-cpufreq.c old_freq = policy->cur; cur 2330 drivers/cpufreq/intel_pstate.c freqs.old = policy->cur; cur 2393 drivers/cpufreq/intel_pstate.c policy->cur = policy->cpuinfo.min_freq; cur 963 drivers/cpufreq/longhaul.c freqs.old = policy->cur; cur 55 drivers/cpufreq/loongson1-cpufreq.c old_freq = policy->cur; cur 49 drivers/cpufreq/omap-cpufreq.c old_freq = policy->cur; cur 194 drivers/cpufreq/pasemi-cpufreq.c policy->cur = pas_freqs[cur_astate].frequency; cur 195 drivers/cpufreq/pasemi-cpufreq.c ppc_proc_freq = policy->cur * 1000ul; cur 215 drivers/cpufreq/pcc-cpufreq.c freqs.old = policy->cur; cur 230 drivers/cpufreq/powernow-k6.c freqs.old = policy->cur; cur 981 drivers/cpufreq/powernow-k8.c pol->cur = find_khz_freq_from_fid(data->currfid); cur 922 drivers/cpufreq/powernv-cpufreq.c index = cpufreq_table_find_index_c(&policy, policy.cur); cur 106 drivers/cpufreq/ppc_cbe_cpufreq.c policy->cur = cbe_freqs[cur_pmode].frequency; cur 203 drivers/cpufreq/pxa2xx-cpufreq.c policy->cur / 1000, new_freq_cpu / 1000); cur 205 drivers/cpufreq/pxa2xx-cpufreq.c if (vcc_core && new_freq_cpu > policy->cur) { cur 222 drivers/cpufreq/pxa2xx-cpufreq.c if (vcc_core && new_freq_cpu < policy->cur) cur 242 drivers/cpufreq/s5pv210-cpufreq.c old_freq = policy->cur; cur 166 drivers/cpufreq/sa1100-cpufreq.c unsigned int cur = sa11x0_getspeed(0); cur 171 drivers/cpufreq/sa1100-cpufreq.c if (new_freq > cur) cur 172 drivers/cpufreq/sa1100-cpufreq.c sa1100_update_dram_timings(cur, new_freq); cur 176 drivers/cpufreq/sa1100-cpufreq.c if (new_freq < cur) cur 177 drivers/cpufreq/sa1100-cpufreq.c sa1100_update_dram_timings(cur, new_freq); cur 295 drivers/cpufreq/sparc-us2e-cpufreq.c policy->cur = clock_tick; cur 139 drivers/cpufreq/sparc-us3-cpufreq.c policy->cur = clock_tick; cur 50 drivers/cpufreq/tegra20-cpufreq.c if (freq_table[index].frequency == ifreq || policy->cur == ifreq) cur 41 drivers/cpufreq/unicore2-cpufreq.c freqs.old = policy->cur; cur 169 drivers/crypto/n2_core.c static void *spu_queue_next(struct spu_queue *q, void *cur) cur 171 drivers/crypto/n2_core.c return q->q + spu_next_offset(q, cur - q->q); cur 280 drivers/crypto/picoxcell_crypto.c struct scatterlist *cur; cur 299 drivers/crypto/picoxcell_crypto.c for_each_sg(payload, cur, mapped_ents, i) cur 300 drivers/crypto/picoxcell_crypto.c ddt_set(&ddt[i], sg_dma_address(cur), sg_dma_len(cur)); cur 318 drivers/crypto/picoxcell_crypto.c struct scatterlist *cur; cur 381 drivers/crypto/picoxcell_crypto.c for_each_sg(areq->src, cur, src_ents, i) cur 382 drivers/crypto/picoxcell_crypto.c ddt_set(src_ddt++, sg_dma_address(cur), sg_dma_len(cur)); cur 386 drivers/crypto/picoxcell_crypto.c for_each_sg(areq->dst, cur, dst_ents, i) { cur 387 drivers/crypto/picoxcell_crypto.c unsigned len = sg_dma_len(cur); cur 394 drivers/crypto/picoxcell_crypto.c ddt_set(dst_ddt++, sg_dma_address(cur) + total, len - total); cur 477 drivers/crypto/qat/qat_common/qat_uclo.c char *chunk_id, void *cur) cur 485 drivers/crypto/qat/qat_common/qat_uclo.c if ((cur < (void *)&chunk_hdr[i]) && cur 131 drivers/dma-buf/dma-fence.c struct dma_fence_cb *cur, *tmp; cur 147 drivers/dma-buf/dma-fence.c list_for_each_entry_safe(cur, tmp, &cb_list, node) { cur 148 drivers/dma-buf/dma-fence.c INIT_LIST_HEAD(&cur->node); cur 149 drivers/dma-buf/dma-fence.c cur->func(fence, cur); cur 85 drivers/firmware/google/memconsole-x86-legacy.c size_t length, cur; cur 101 drivers/firmware/google/memconsole-x86-legacy.c for (cur = 0; cur < length; cur++) { cur 102 drivers/firmware/google/memconsole-x86-legacy.c struct biosmemcon_ebda *hdr = phys_to_virt(address + cur); cur 4669 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c unsigned cur; cur 4673 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c cur = (ring->wptr - 1) & ring->buf_mask; cur 4674 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c if (likely(cur > offset)) cur 4675 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c ring->ring[offset] = cur - offset; cur 4677 drivers/gpu/drm/amd/amdgpu/gfx_v10_0.c ring->ring[offset] = (ring->buf_mask + 1) - offset + cur; cur 6473 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c unsigned cur; cur 6478 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c cur = (ring->wptr & ring->buf_mask) - 1; cur 6479 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c if (likely(cur > offset)) cur 6480 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c ring->ring[offset] = cur - offset; cur 6482 drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c ring->ring[offset] = (ring->ring_size >> 2) - offset + cur; cur 5394 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c unsigned cur; cur 5398 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c cur = (ring->wptr & ring->buf_mask) - 1; cur 5399 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c if (likely(cur > offset)) cur 5400 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c ring->ring[offset] = cur - offset; cur 5402 drivers/gpu/drm/amd/amdgpu/gfx_v9_0.c ring->ring[offset] = (ring->ring_size>>2) - offset + cur; cur 249 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c unsigned cur; cur 254 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c cur = (ring->wptr - 1) & ring->buf_mask; cur 255 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c if (cur > offset) cur 256 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c ring->ring[offset] = cur - offset; cur 258 drivers/gpu/drm/amd/amdgpu/sdma_v5_0.c ring->ring[offset] = (ring->buf_mask + 1) - offset + cur; cur 809 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c struct device_process_node *cur, *next; cur 817 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c list_for_each_entry_safe(cur, next, &dqm->queues, list) { cur 818 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c if (qpd == cur->qpd) { cur 819 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c list_del(&cur->list); cur 820 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c kfree(cur); cur 1499 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c struct device_process_node *cur, *next_dpn; cur 1515 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c list_for_each_entry_safe(cur, next_dpn, &dqm->queues, list) { cur 1516 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c if (qpd == cur->qpd) { cur 1517 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c list_del(&cur->list); cur 1518 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c kfree(cur); cur 1575 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c struct device_process_node *cur, *next_dpn; cur 1610 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c list_for_each_entry_safe(cur, next_dpn, &dqm->queues, list) { cur 1611 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c if (qpd == cur->qpd) { cur 1612 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c list_del(&cur->list); cur 1613 drivers/gpu/drm/amd/amdkfd/kfd_device_queue_manager.c kfree(cur); cur 127 drivers/gpu/drm/amd/amdkfd/kfd_packet_manager.c struct device_process_node *cur; cur 147 drivers/gpu/drm/amd/amdkfd/kfd_packet_manager.c list_for_each_entry(cur, queues, list) { cur 148 drivers/gpu/drm/amd/amdkfd/kfd_packet_manager.c qpd = cur->qpd; cur 3679 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c struct dm_crtc_state *cur = to_dm_crtc_state(state); cur 3682 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c if (cur->stream) cur 3683 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c dc_stream_release(cur->stream); cur 3711 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c struct dm_crtc_state *state, *cur; cur 3713 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c cur = to_dm_crtc_state(crtc->state); cur 3724 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c if (cur->stream) { cur 3725 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c state->stream = cur->stream; cur 3729 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c state->active_planes = cur->active_planes; cur 3730 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c state->interrupts_enabled = cur->interrupts_enabled; cur 3731 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c state->vrr_params = cur->vrr_params; cur 3732 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c state->vrr_infopacket = cur->vrr_infopacket; cur 3733 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c state->abm_level = cur->abm_level; cur 3734 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c state->vrr_supported = cur->vrr_supported; cur 3735 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c state->freesync_config = cur->freesync_config; cur 3736 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c state->crc_src = cur->crc_src; cur 3737 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c state->cm_has_degamma = cur->cm_has_degamma; cur 3738 drivers/gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c state->cm_is_degamma_srgb = cur->cm_is_degamma_srgb; cur 1548 drivers/gpu/drm/amd/display/dc/core/dc_link_dp.c struct dc_link_settings *cur = &cur_link_setting; cur 1588 drivers/gpu/drm/amd/display/dc/core/dc_link_dp.c if (cur->link_rate == LINK_RATE_LOW) cur 1595 drivers/gpu/drm/amd/display/dc/core/dc_link_dp.c cur); cur 1603 drivers/gpu/drm/amd/display/dc/core/dc_link_dp.c cur, cur 1612 drivers/gpu/drm/amd/display/dc/core/dc_link_dp.c link->verified_link_cap = *cur; cur 1626 drivers/gpu/drm/amd/display/dc/core/dc_link_dp.c initial_link_settings, cur, status)); cur 423 drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_crtc.c struct atmel_hlcdc_crtc_state *state, *cur; cur 433 drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_crtc.c cur = drm_crtc_state_to_atmel_hlcdc_crtc_state(crtc->state); cur 434 drivers/gpu/drm/atmel-hlcdc/atmel_hlcdc_crtc.c state->output_mode = cur->output_mode; cur 228 drivers/gpu/drm/drm_syncobj.c struct syncobj_wait_entry *cur, *tmp; cur 242 drivers/gpu/drm/drm_syncobj.c list_for_each_entry_safe(cur, tmp, &syncobj->cb_list, node) cur 243 drivers/gpu/drm/drm_syncobj.c syncobj_wait_syncobj_func(syncobj, cur); cur 263 drivers/gpu/drm/drm_syncobj.c struct syncobj_wait_entry *cur, *tmp; cur 275 drivers/gpu/drm/drm_syncobj.c list_for_each_entry_safe(cur, tmp, &syncobj->cb_list, node) cur 276 drivers/gpu/drm/drm_syncobj.c syncobj_wait_syncobj_func(syncobj, cur); cur 2066 drivers/gpu/drm/gma500/cdv_intel_dp.c struct edp_power_seq cur; cur 2085 drivers/gpu/drm/gma500/cdv_intel_dp.c cur.t1_t3 = (pp_on & PANEL_POWER_UP_DELAY_MASK) >> cur 2088 drivers/gpu/drm/gma500/cdv_intel_dp.c cur.t8 = (pp_on & PANEL_LIGHT_ON_DELAY_MASK) >> cur 2091 drivers/gpu/drm/gma500/cdv_intel_dp.c cur.t9 = (pp_off & PANEL_LIGHT_OFF_DELAY_MASK) >> cur 2094 drivers/gpu/drm/gma500/cdv_intel_dp.c cur.t10 = (pp_off & PANEL_POWER_DOWN_DELAY_MASK) >> cur 2097 drivers/gpu/drm/gma500/cdv_intel_dp.c cur.t11_t12 = ((pp_div & PANEL_POWER_CYCLE_DELAY_MASK) >> cur 2101 drivers/gpu/drm/gma500/cdv_intel_dp.c cur.t1_t3, cur.t8, cur.t9, cur.t10, cur.t11_t12); cur 2104 drivers/gpu/drm/gma500/cdv_intel_dp.c intel_dp->panel_power_up_delay = cur.t1_t3 / 10; cur 2105 drivers/gpu/drm/gma500/cdv_intel_dp.c intel_dp->backlight_on_delay = cur.t8 / 10; cur 2106 drivers/gpu/drm/gma500/cdv_intel_dp.c intel_dp->backlight_off_delay = cur.t9 / 10; cur 2107 drivers/gpu/drm/gma500/cdv_intel_dp.c intel_dp->panel_power_down_delay = cur.t10 / 10; cur 2108 drivers/gpu/drm/gma500/cdv_intel_dp.c intel_dp->panel_power_cycle_delay = (cur.t11_t12 - 1) * 100; cur 11470 drivers/gpu/drm/i915/display/intel_display.c static bool intel_wm_need_update(const struct intel_plane_state *cur, cur 11474 drivers/gpu/drm/i915/display/intel_display.c if (new->base.visible != cur->base.visible) cur 11477 drivers/gpu/drm/i915/display/intel_display.c if (!cur->base.fb || !new->base.fb) cur 11480 drivers/gpu/drm/i915/display/intel_display.c if (cur->base.fb->modifier != new->base.fb->modifier || cur 11481 drivers/gpu/drm/i915/display/intel_display.c cur->base.rotation != new->base.rotation || cur 11482 drivers/gpu/drm/i915/display/intel_display.c drm_rect_width(&new->base.src) != drm_rect_width(&cur->base.src) || cur 11483 drivers/gpu/drm/i915/display/intel_display.c drm_rect_height(&new->base.src) != drm_rect_height(&cur->base.src) || cur 11484 drivers/gpu/drm/i915/display/intel_display.c drm_rect_width(&new->base.dst) != drm_rect_width(&cur->base.dst) || cur 11485 drivers/gpu/drm/i915/display/intel_display.c drm_rect_height(&new->base.dst) != drm_rect_height(&cur->base.dst)) cur 6440 drivers/gpu/drm/i915/display/intel_dp.c struct edp_power_seq cur, vbt, spec, cur 6449 drivers/gpu/drm/i915/display/intel_dp.c intel_pps_readout_hw_state(intel_dp, &cur); cur 6451 drivers/gpu/drm/i915/display/intel_dp.c intel_pps_dump_state("cur", &cur); cur 6486 drivers/gpu/drm/i915/display/intel_dp.c #define assign_final(field) final->field = (max(cur.field, vbt.field) == 0 ? \ cur 6488 drivers/gpu/drm/i915/display/intel_dp.c max(cur.field, vbt.field)) cur 120 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c struct dma_fence_cb *cur, *tmp; cur 125 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c list_for_each_entry_safe(cur, tmp, list, node) { cur 126 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c INIT_LIST_HEAD(&cur->node); cur 127 drivers/gpu/drm/i915/gt/intel_breadcrumbs.c cur->func(fence, cur); cur 632 drivers/gpu/drm/i915/gt/intel_lrc.c struct intel_engine_cs *cur, *old; cur 638 drivers/gpu/drm/i915/gt/intel_lrc.c cur = ptr_unmask_bits(old, 2) ? ptr_dec(old) : NULL; cur 639 drivers/gpu/drm/i915/gt/intel_lrc.c while (!try_cmpxchg(&ce->inflight, &old, cur)); cur 640 drivers/gpu/drm/i915/gt/intel_lrc.c if (!cur) cur 951 drivers/gpu/drm/i915/gt/intel_workarounds.c wa_verify(const struct i915_wa *wa, u32 cur, const char *name, const char *from) cur 953 drivers/gpu/drm/i915/gt/intel_workarounds.c if ((cur ^ wa->val) & wa->read) { cur 956 drivers/gpu/drm/i915/gt/intel_workarounds.c cur, cur & wa->read, cur 71 drivers/gpu/drm/i915/i915_gpu_error.c __sg_set_buf(e->cur++, e->buf, e->bytes, e->iter); cur 77 drivers/gpu/drm/i915/i915_gpu_error.c if (e->cur == e->end) { cur 86 drivers/gpu/drm/i915/i915_gpu_error.c if (e->cur) { cur 87 drivers/gpu/drm/i915/i915_gpu_error.c e->cur->offset = 0; cur 88 drivers/gpu/drm/i915/i915_gpu_error.c e->cur->length = 0; cur 89 drivers/gpu/drm/i915/i915_gpu_error.c e->cur->page_link = cur 95 drivers/gpu/drm/i915/i915_gpu_error.c e->cur = sgl; cur 807 drivers/gpu/drm/i915/i915_gpu_error.c __sg_set_buf(m.cur++, m.buf, m.bytes, m.iter); cur 811 drivers/gpu/drm/i915/i915_gpu_error.c if (m.cur) { cur 812 drivers/gpu/drm/i915/i915_gpu_error.c GEM_BUG_ON(m.end < m.cur); cur 813 drivers/gpu/drm/i915/i915_gpu_error.c sg_mark_end(m.cur - 1); cur 815 drivers/gpu/drm/i915/i915_gpu_error.c GEM_BUG_ON(m.sgl && !m.cur); cur 186 drivers/gpu/drm/i915/i915_gpu_error.h struct scatterlist *sgl, *cur, *end; cur 164 drivers/gpu/drm/i915/i915_pmu.c sample->cur += val; cur 221 drivers/gpu/drm/i915/i915_pmu.c sample->cur += mul_u32_u32(val, mul); cur 471 drivers/gpu/drm/i915/i915_pmu.c if (val >= pmu->sample[__I915_SAMPLE_RC6_ESTIMATED].cur) { cur 472 drivers/gpu/drm/i915/i915_pmu.c pmu->sample[__I915_SAMPLE_RC6_ESTIMATED].cur = 0; cur 473 drivers/gpu/drm/i915/i915_pmu.c pmu->sample[__I915_SAMPLE_RC6].cur = val; cur 475 drivers/gpu/drm/i915/i915_pmu.c val = pmu->sample[__I915_SAMPLE_RC6_ESTIMATED].cur; cur 505 drivers/gpu/drm/i915/i915_pmu.c if (!pmu->sample[__I915_SAMPLE_RC6_ESTIMATED].cur) cur 509 drivers/gpu/drm/i915/i915_pmu.c val += pmu->sample[__I915_SAMPLE_RC6].cur; cur 511 drivers/gpu/drm/i915/i915_pmu.c pmu->sample[__I915_SAMPLE_RC6_ESTIMATED].cur = val; cur 512 drivers/gpu/drm/i915/i915_pmu.c } else if (pmu->sample[__I915_SAMPLE_RC6_ESTIMATED].cur) { cur 513 drivers/gpu/drm/i915/i915_pmu.c val = pmu->sample[__I915_SAMPLE_RC6_ESTIMATED].cur; cur 515 drivers/gpu/drm/i915/i915_pmu.c val = pmu->sample[__I915_SAMPLE_RC6].cur; cur 548 drivers/gpu/drm/i915/i915_pmu.c val = engine->pmu.sample[sample].cur; cur 554 drivers/gpu/drm/i915/i915_pmu.c div_u64(pmu->sample[__I915_SAMPLE_FREQ_ACT].cur, cur 559 drivers/gpu/drm/i915/i915_pmu.c div_u64(pmu->sample[__I915_SAMPLE_FREQ_REQ].cur, cur 37 drivers/gpu/drm/i915/i915_pmu.h u64 cur; cur 6108 drivers/gpu/drm/i915/i915_reg.h #define HSW_WM_LP_VAL(lat, fbc, pri, cur) \ cur 6110 drivers/gpu/drm/i915/i915_reg.h ((fbc) << WM1_LP_FBC_SHIFT) | ((pri) << WM1_LP_SR_SHIFT) | (cur)) cur 2497 drivers/gpu/drm/i915/intel_pm.c u16 cur; cur 2706 drivers/gpu/drm/i915/intel_pm.c max->cur = ilk_cursor_wm_max(dev_priv, level, config); cur 2716 drivers/gpu/drm/i915/intel_pm.c max->cur = ilk_cursor_wm_reg_max(dev_priv, level); cur 2732 drivers/gpu/drm/i915/intel_pm.c result->cur_val <= max->cur; cur 2748 drivers/gpu/drm/i915/intel_pm.c if (result->cur_val > max->cur) cur 2750 drivers/gpu/drm/i915/intel_pm.c level, result->cur_val, max->cur); cur 2754 drivers/gpu/drm/i915/intel_pm.c result->cur_val = min_t(u32, result->cur_val, max->cur); cur 197 drivers/gpu/drm/mediatek/mtk_drm_ddp.c static unsigned int mtk_ddp_mout_en(enum mtk_ddp_comp_id cur, cur 203 drivers/gpu/drm/mediatek/mtk_drm_ddp.c if (cur == DDP_COMPONENT_OVL0 && next == DDP_COMPONENT_COLOR0) { cur 206 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_OVL0 && next == DDP_COMPONENT_RDMA0) { cur 209 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_OD0 && next == DDP_COMPONENT_RDMA0) { cur 212 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_UFOE && next == DDP_COMPONENT_DSI0) { cur 215 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_OVL1 && next == DDP_COMPONENT_COLOR1) { cur 218 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_GAMMA && next == DDP_COMPONENT_RDMA1) { cur 221 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_OD1 && next == DDP_COMPONENT_RDMA1) { cur 224 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA0 && next == DDP_COMPONENT_DPI0) { cur 227 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA0 && next == DDP_COMPONENT_DPI1) { cur 230 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA0 && next == DDP_COMPONENT_DSI1) { cur 233 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA0 && next == DDP_COMPONENT_DSI2) { cur 236 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA0 && next == DDP_COMPONENT_DSI3) { cur 239 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI1) { cur 242 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI2) { cur 245 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI3) { cur 248 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DPI0) { cur 251 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DPI1) { cur 254 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DPI0) { cur 257 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DPI1) { cur 260 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI1) { cur 263 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI2) { cur 266 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI3) { cur 276 drivers/gpu/drm/mediatek/mtk_drm_ddp.c static unsigned int mtk_ddp_sel_in(enum mtk_ddp_comp_id cur, cur 282 drivers/gpu/drm/mediatek/mtk_drm_ddp.c if (cur == DDP_COMPONENT_OVL0 && next == DDP_COMPONENT_COLOR0) { cur 285 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DPI0) { cur 288 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DPI1) { cur 291 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI0) { cur 294 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI1) { cur 297 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI2) { cur 300 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA1 && next == DDP_COMPONENT_DSI3) { cur 303 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DPI0) { cur 306 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DPI1) { cur 309 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI0) { cur 312 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI1) { cur 315 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI2) { cur 318 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_RDMA2 && next == DDP_COMPONENT_DSI3) { cur 321 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_OVL1 && next == DDP_COMPONENT_COLOR1) { cur 324 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_BLS && next == DDP_COMPONENT_DSI0) { cur 335 drivers/gpu/drm/mediatek/mtk_drm_ddp.c enum mtk_ddp_comp_id cur, cur 338 drivers/gpu/drm/mediatek/mtk_drm_ddp.c if (cur == DDP_COMPONENT_BLS && next == DDP_COMPONENT_DSI0) { cur 341 drivers/gpu/drm/mediatek/mtk_drm_ddp.c } else if (cur == DDP_COMPONENT_BLS && next == DDP_COMPONENT_DPI0) { cur 352 drivers/gpu/drm/mediatek/mtk_drm_ddp.c enum mtk_ddp_comp_id cur, cur 357 drivers/gpu/drm/mediatek/mtk_drm_ddp.c value = mtk_ddp_mout_en(cur, next, &addr); cur 363 drivers/gpu/drm/mediatek/mtk_drm_ddp.c mtk_ddp_sout_sel(config_regs, cur, next); cur 365 drivers/gpu/drm/mediatek/mtk_drm_ddp.c value = mtk_ddp_sel_in(cur, next, &addr); cur 373 drivers/gpu/drm/mediatek/mtk_drm_ddp.c enum mtk_ddp_comp_id cur, cur 378 drivers/gpu/drm/mediatek/mtk_drm_ddp.c value = mtk_ddp_mout_en(cur, next, &addr); cur 384 drivers/gpu/drm/mediatek/mtk_drm_ddp.c value = mtk_ddp_sel_in(cur, next, &addr); cur 16 drivers/gpu/drm/mediatek/mtk_drm_ddp.h enum mtk_ddp_comp_id cur, cur 19 drivers/gpu/drm/mediatek/mtk_drm_ddp.h enum mtk_ddp_comp_id cur, cur 31 drivers/gpu/drm/msm/adreno/a5xx_gpu.c ring->cur = ring->next; cur 15 drivers/gpu/drm/msm/adreno/a5xx_preempt.c enum preempt_state cur = atomic_cmpxchg(&a5xx_gpu->preempt_state, cur 18 drivers/gpu/drm/msm/adreno/a5xx_preempt.c return (cur == old); cur 60 drivers/gpu/drm/msm/adreno/a6xx_gpu.c ring->cur = ring->next; cur 349 drivers/gpu/drm/msm/adreno/adreno_gpu.c ring->cur = ring->start; cur 492 drivers/gpu/drm/msm/adreno/adreno_gpu.c ring->cur = ring->next; cur 368 drivers/gpu/drm/msm/adreno/adreno_gpu.h return (ring->cur - ring->start) % (MSM_GPU_RINGBUFFER_SZ >> 2); cur 54 drivers/gpu/drm/msm/disp/mdp5/mdp5_mixer.c struct mdp5_hw_mixer *cur = mdp5_kms->hwmixers[i]; cur 63 drivers/gpu/drm/msm/disp/mdp5/mdp5_mixer.c if (new_state->hwmixer_to_crtc[cur->idx] && cur 64 drivers/gpu/drm/msm/disp/mdp5/mdp5_mixer.c new_state->hwmixer_to_crtc[cur->idx] != crtc) cur 68 drivers/gpu/drm/msm/disp/mdp5/mdp5_mixer.c if (caps & ~cur->caps) cur 74 drivers/gpu/drm/msm/disp/mdp5/mdp5_mixer.c pair_idx = get_right_pair_idx(mdp5_kms, cur->lm); cur 97 drivers/gpu/drm/msm/disp/mdp5/mdp5_mixer.c if (!(*mixer) || cur->caps & MDP_LM_CAP_PAIR) cur 98 drivers/gpu/drm/msm/disp/mdp5/mdp5_mixer.c *mixer = cur; cur 31 drivers/gpu/drm/msm/disp/mdp5/mdp5_pipe.c struct mdp5_hw_pipe *cur = mdp5_kms->hwpipes[i]; cur 40 drivers/gpu/drm/msm/disp/mdp5/mdp5_pipe.c if (new_state->hwpipe_to_plane[cur->idx] || cur 41 drivers/gpu/drm/msm/disp/mdp5/mdp5_pipe.c old_state->hwpipe_to_plane[cur->idx]) cur 45 drivers/gpu/drm/msm/disp/mdp5/mdp5_pipe.c if (caps & ~cur->caps) cur 52 drivers/gpu/drm/msm/disp/mdp5/mdp5_pipe.c if (cur->caps & MDP_PIPE_CAP_CURSOR && cur 59 drivers/gpu/drm/msm/disp/mdp5/mdp5_pipe.c if (!(*hwpipe) || (hweight_long(cur->caps & ~caps) < cur 70 drivers/gpu/drm/msm/disp/mdp5/mdp5_pipe.c if (r_cur->caps != cur->caps) cur 74 drivers/gpu/drm/msm/disp/mdp5/mdp5_pipe.c if (cur->pipe > r_cur->pipe) cur 84 drivers/gpu/drm/msm/disp/mdp5/mdp5_pipe.c *hwpipe = cur; cur 42 drivers/gpu/drm/msm/msm_ringbuffer.c ring->cur = ring->start; cur 40 drivers/gpu/drm/msm/msm_ringbuffer.h uint32_t *start, *end, *cur, *next; cur 40 drivers/gpu/drm/nouveau/dispnv04/overlay.c struct nouveau_bo *cur; cur 125 drivers/gpu/drm/nouveau/dispnv04/overlay.c struct nouveau_bo *cur = nv_plane->cur; cur 147 drivers/gpu/drm/nouveau/dispnv04/overlay.c nv_plane->cur = nv_fb->nvbo; cur 183 drivers/gpu/drm/nouveau/dispnv04/overlay.c if (cur) cur 184 drivers/gpu/drm/nouveau/dispnv04/overlay.c nouveau_bo_unpin(cur); cur 198 drivers/gpu/drm/nouveau/dispnv04/overlay.c if (nv_plane->cur) { cur 199 drivers/gpu/drm/nouveau/dispnv04/overlay.c nouveau_bo_unpin(nv_plane->cur); cur 200 drivers/gpu/drm/nouveau/dispnv04/overlay.c nv_plane->cur = NULL; cur 229 drivers/gpu/drm/nouveau/dispnv04/overlay.c if (plane->cur) { cur 372 drivers/gpu/drm/nouveau/dispnv04/overlay.c struct nouveau_bo *cur = nv_plane->cur; cur 391 drivers/gpu/drm/nouveau/dispnv04/overlay.c nv_plane->cur = nv_fb->nvbo; cur 432 drivers/gpu/drm/nouveau/dispnv04/overlay.c if (cur) cur 433 drivers/gpu/drm/nouveau/dispnv04/overlay.c nouveau_bo_unpin(cur); cur 450 drivers/gpu/drm/nouveau/dispnv04/overlay.c if (nv_plane->cur) { cur 451 drivers/gpu/drm/nouveau/dispnv04/overlay.c nouveau_bo_unpin(nv_plane->cur); cur 452 drivers/gpu/drm/nouveau/dispnv04/overlay.c nv_plane->cur = NULL; cur 442 drivers/gpu/drm/nouveau/nouveau_chan.c chan->dma.cur = chan->dma.put; cur 443 drivers/gpu/drm/nouveau/nouveau_chan.c chan->dma.free = chan->dma.max - chan->dma.cur; cur 34 drivers/gpu/drm/nouveau/nouveau_chan.h int cur; cur 147 drivers/gpu/drm/nouveau/nouveau_debugfs.c char buf[32] = {}, *tmp, *cur = buf; cur 162 drivers/gpu/drm/nouveau/nouveau_debugfs.c if (!strncasecmp(cur, "dc:", 3)) { cur 164 drivers/gpu/drm/nouveau/nouveau_debugfs.c cur += 3; cur 166 drivers/gpu/drm/nouveau/nouveau_debugfs.c if (!strncasecmp(cur, "ac:", 3)) { cur 168 drivers/gpu/drm/nouveau/nouveau_debugfs.c cur += 3; cur 171 drivers/gpu/drm/nouveau/nouveau_debugfs.c if (!strcasecmp(cur, "none")) cur 174 drivers/gpu/drm/nouveau/nouveau_debugfs.c if (!strcasecmp(cur, "auto")) cur 177 drivers/gpu/drm/nouveau/nouveau_debugfs.c ret = kstrtol(cur, 16, &value); cur 38 drivers/gpu/drm/nouveau/nouveau_dma.c mem = &mem[chan->dma.cur]; cur 43 drivers/gpu/drm/nouveau/nouveau_dma.c chan->dma.cur += nr_dwords; cur 153 drivers/gpu/drm/nouveau/nouveau_dma.c if (get <= chan->dma.cur) { cur 154 drivers/gpu/drm/nouveau/nouveau_dma.c chan->dma.free = chan->dma.max - chan->dma.cur; cur 167 drivers/gpu/drm/nouveau/nouveau_dma.c chan->dma.cur = 0; cur 171 drivers/gpu/drm/nouveau/nouveau_dma.c chan->dma.free = get - chan->dma.cur - 1; cur 203 drivers/gpu/drm/nouveau/nouveau_dma.c if (get <= chan->dma.cur) { cur 217 drivers/gpu/drm/nouveau/nouveau_dma.c chan->dma.free = chan->dma.max - chan->dma.cur; cur 244 drivers/gpu/drm/nouveau/nouveau_dma.c chan->dma.cur = cur 254 drivers/gpu/drm/nouveau/nouveau_dma.c chan->dma.free = get - chan->dma.cur - 1; cur 102 drivers/gpu/drm/nouveau/nouveau_dma.h nouveau_bo_wr32(chan->push.buffer, chan->dma.cur++, data); cur 147 drivers/gpu/drm/nouveau/nouveau_dma.h if (chan->dma.cur == chan->dma.put) cur 153 drivers/gpu/drm/nouveau/nouveau_dma.h (chan->dma.cur - chan->dma.put) << 2); cur 155 drivers/gpu/drm/nouveau/nouveau_dma.h WRITE_PUT(chan->dma.cur); cur 158 drivers/gpu/drm/nouveau/nouveau_dma.h chan->dma.put = chan->dma.cur; cur 164 drivers/gpu/drm/nouveau/nouveau_dma.h chan->dma.cur = chan->dma.put; cur 816 drivers/gpu/drm/nouveau/nouveau_gem.c cmd = chan->push.addr + ((chan->dma.cur + 2) << 2); cur 868 drivers/gpu/drm/nouveau/nouveau_gem.c (chan->push.addr + ((chan->dma.cur + 2) << 2)); cur 56 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c struct nvkm_memory *cur; cur 61 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c cur = fifo->runlist.mem[fifo->runlist.active]; cur 64 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c nvkm_kmap(cur); cur 66 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c nvkm_wo32(cur, (nr * 8) + 0, chan->base.chid); cur 67 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c nvkm_wo32(cur, (nr * 8) + 4, 0x00000004); cur 70 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c nvkm_done(cur); cur 72 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c switch (nvkm_memory_target(cur)) { cur 81 drivers/gpu/drm/nouveau/nvkm/engine/fifo/gf100.c nvkm_wr32(device, 0x002270, (nvkm_memory_addr(cur) >> 12) | cur 33 drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c struct nvkm_memory *cur; cur 36 drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c cur = fifo->runlist[fifo->cur_runlist]; cur 39 drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c nvkm_kmap(cur); cur 42 drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c nvkm_wo32(cur, p++ * 4, i); cur 44 drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c nvkm_done(cur); cur 46 drivers/gpu/drm/nouveau/nvkm/engine/fifo/nv50.c nvkm_wr32(device, 0x0032f4, nvkm_memory_addr(cur) >> 12); cur 129 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c u32 cycles, cur, prev; cur 149 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c cur = nvkm_gpio_get(gpio, 0, therm->fan->tach.func, cur 151 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c if (prev != cur) { cur 155 drivers/gpu/drm/nouveau/nvkm/subdev/therm/fan.c prev = cur; cur 103 drivers/gpu/drm/nouveau/nvkm/subdev/therm/g84.c int temp, cur; cur 118 drivers/gpu/drm/nouveau/nvkm/subdev/therm/g84.c cur = therm->func->temp_get(therm); cur 119 drivers/gpu/drm/nouveau/nvkm/subdev/therm/g84.c if (new_state == NVKM_THERM_THRS_LOWER && cur > thrs->temp) cur 122 drivers/gpu/drm/nouveau/nvkm/subdev/therm/g84.c cur < thrs->temp - thrs->hysteresis) cur 64 drivers/gpu/drm/qxl/qxl_release.c unsigned long cur, end = jiffies + timeout; cur 111 drivers/gpu/drm/qxl/qxl_release.c cur = jiffies; cur 112 drivers/gpu/drm/qxl/qxl_release.c if (time_after(cur, end)) cur 114 drivers/gpu/drm/qxl/qxl_release.c return end - cur; cur 194 drivers/gpu/drm/radeon/rv6xx_dpm.c struct rv6xx_sclk_stepping *cur, cur 199 drivers/gpu/drm/radeon/rv6xx_dpm.c next.post_divider = cur->post_divider; cur 202 drivers/gpu/drm/radeon/rv6xx_dpm.c next.vco_frequency = (cur->vco_frequency * (100 + step_size)) / 100; cur 204 drivers/gpu/drm/radeon/rv6xx_dpm.c next.vco_frequency = (cur->vco_frequency * 100 + 99 + step_size) / (100 + step_size); cur 210 drivers/gpu/drm/radeon/rv6xx_dpm.c struct rv6xx_sclk_stepping *cur, cur 213 drivers/gpu/drm/radeon/rv6xx_dpm.c return (cur->post_divider > target->post_divider) && cur 214 drivers/gpu/drm/radeon/rv6xx_dpm.c ((cur->vco_frequency * target->post_divider) <= cur 215 drivers/gpu/drm/radeon/rv6xx_dpm.c (target->vco_frequency * (cur->post_divider - 1))); cur 219 drivers/gpu/drm/radeon/rv6xx_dpm.c struct rv6xx_sclk_stepping *cur, cur 222 drivers/gpu/drm/radeon/rv6xx_dpm.c struct rv6xx_sclk_stepping next = *cur; cur 231 drivers/gpu/drm/radeon/rv6xx_dpm.c struct rv6xx_sclk_stepping *cur, cur 235 drivers/gpu/drm/radeon/rv6xx_dpm.c return (increasing_vco && (cur->vco_frequency >= target->vco_frequency)) || cur 236 drivers/gpu/drm/radeon/rv6xx_dpm.c (!increasing_vco && (cur->vco_frequency <= target->vco_frequency)); cur 243 drivers/gpu/drm/radeon/rv6xx_dpm.c struct rv6xx_sclk_stepping cur; cur 248 drivers/gpu/drm/radeon/rv6xx_dpm.c rv6xx_convert_clock_to_stepping(rdev, low, &cur); cur 251 drivers/gpu/drm/radeon/rv6xx_dpm.c rv6xx_output_stepping(rdev, step_index++, &cur); cur 253 drivers/gpu/drm/radeon/rv6xx_dpm.c increasing_vco = (target.vco_frequency >= cur.vco_frequency); cur 255 drivers/gpu/drm/radeon/rv6xx_dpm.c if (target.post_divider > cur.post_divider) cur 256 drivers/gpu/drm/radeon/rv6xx_dpm.c cur.post_divider = target.post_divider; cur 261 drivers/gpu/drm/radeon/rv6xx_dpm.c if (rv6xx_can_step_post_div(rdev, &cur, &target)) cur 262 drivers/gpu/drm/radeon/rv6xx_dpm.c next = rv6xx_next_post_div_step(rdev, &cur, &target); cur 264 drivers/gpu/drm/radeon/rv6xx_dpm.c next = rv6xx_next_vco_step(rdev, &cur, increasing_vco, R600_VCOSTEPPCT_DFLT); cur 271 drivers/gpu/drm/radeon/rv6xx_dpm.c if (!rv6xx_reached_stepping_target(rdev, &tiny, &cur, !increasing_vco)) cur 289 drivers/gpu/drm/radeon/rv6xx_dpm.c cur = next; cur 374 drivers/gpu/drm/savage/savage_bci.c unsigned int cur = dev_priv->current_dma_page; cur 376 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[cur].used; cur 383 drivers/gpu/drm/savage/savage_bci.c cur, dev_priv->dma_pages[cur].used, n, rest, nr_pages); cur 385 drivers/gpu/drm/savage/savage_bci.c if (cur + nr_pages < dev_priv->nr_dma_pages) { cur 387 drivers/gpu/drm/savage/savage_bci.c cur * SAVAGE_DMA_PAGE_SIZE + dev_priv->dma_pages[cur].used; cur 390 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[cur].used += rest; cur 392 drivers/gpu/drm/savage/savage_bci.c cur++; cur 397 drivers/gpu/drm/savage/savage_bci.c for (i = cur; i < dev_priv->nr_dma_pages; ++i) { cur 403 drivers/gpu/drm/savage/savage_bci.c dev_priv->first_dma_page = cur = 0; cur 405 drivers/gpu/drm/savage/savage_bci.c for (i = cur; nr_pages > 0; ++i, --nr_pages) { cur 431 drivers/gpu/drm/savage/savage_bci.c unsigned int cur = dev_priv->current_dma_page; cur 437 drivers/gpu/drm/savage/savage_bci.c if (first == cur && cur 438 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[cur].used == dev_priv->dma_pages[cur].flushed) cur 443 drivers/gpu/drm/savage/savage_bci.c pad = -dev_priv->dma_pages[cur].used & 1; cur 444 drivers/gpu/drm/savage/savage_bci.c align = -(dev_priv->dma_pages[cur].used + pad) & 7; cur 448 drivers/gpu/drm/savage/savage_bci.c first, cur, dev_priv->dma_pages[first].flushed, cur 449 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[cur].used, pad, align); cur 454 drivers/gpu/drm/savage/savage_bci.c cur * SAVAGE_DMA_PAGE_SIZE + dev_priv->dma_pages[cur].used; cur 455 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[cur].used += pad; cur 468 drivers/gpu/drm/savage/savage_bci.c len = (cur - first) * SAVAGE_DMA_PAGE_SIZE + cur 469 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[cur].used - dev_priv->dma_pages[first].flushed; cur 480 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[cur].used += align; cur 485 drivers/gpu/drm/savage/savage_bci.c for (i = first; i < cur; ++i) { cur 491 drivers/gpu/drm/savage/savage_bci.c if (dev_priv->dma_pages[cur].used == SAVAGE_DMA_PAGE_SIZE) { cur 492 drivers/gpu/drm/savage/savage_bci.c SET_AGE(&dev_priv->dma_pages[cur].age, event, wrap); cur 493 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[cur].used = 0; cur 494 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[cur].flushed = 0; cur 496 drivers/gpu/drm/savage/savage_bci.c cur++; cur 497 drivers/gpu/drm/savage/savage_bci.c if (cur == dev_priv->nr_dma_pages) cur 498 drivers/gpu/drm/savage/savage_bci.c cur = 0; cur 499 drivers/gpu/drm/savage/savage_bci.c dev_priv->first_dma_page = dev_priv->current_dma_page = cur; cur 501 drivers/gpu/drm/savage/savage_bci.c dev_priv->first_dma_page = cur; cur 502 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[cur].flushed = dev_priv->dma_pages[i].used; cur 506 drivers/gpu/drm/savage/savage_bci.c DRM_DEBUG("first=cur=%u, cur->used=%u, cur->flushed=%u\n", cur, cur 507 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[cur].used, cur 508 drivers/gpu/drm/savage/savage_bci.c dev_priv->dma_pages[cur].flushed); cur 518 drivers/gpu/drm/savage/savage_drv.h unsigned int cur = dev_priv->current_dma_page; \ cur 520 drivers/gpu/drm/savage/savage_drv.h dev_priv->dma_pages[cur].used; \ cur 525 drivers/gpu/drm/savage/savage_drv.h cur * SAVAGE_DMA_PAGE_SIZE + \ cur 526 drivers/gpu/drm/savage/savage_drv.h dev_priv->dma_pages[cur].used; \ cur 527 drivers/gpu/drm/savage/savage_drv.h if (dev_priv->dma_pages[cur].used == 0) \ cur 528 drivers/gpu/drm/savage/savage_drv.h savage_dma_wait(dev_priv, cur); \ cur 529 drivers/gpu/drm/savage/savage_drv.h dev_priv->dma_pages[cur].used += (n); \ cur 542 drivers/gpu/drm/savage/savage_drv.h unsigned int cur = dev_priv->current_dma_page; \ cur 544 drivers/gpu/drm/savage/savage_drv.h cur * SAVAGE_DMA_PAGE_SIZE + \ cur 545 drivers/gpu/drm/savage/savage_drv.h dev_priv->dma_pages[cur].used; \ cur 316 drivers/gpu/drm/via/via_dmablit.c int cur; cur 333 drivers/gpu/drm/via/via_dmablit.c cur = blitq->cur; cur 336 drivers/gpu/drm/via/via_dmablit.c blitq->blits[cur]->aborted = blitq->aborting; cur 338 drivers/gpu/drm/via/via_dmablit.c wake_up(blitq->blit_queue + cur); cur 340 drivers/gpu/drm/via/via_dmablit.c cur++; cur 341 drivers/gpu/drm/via/via_dmablit.c if (cur >= VIA_NUM_BLIT_SLOTS) cur 342 drivers/gpu/drm/via/via_dmablit.c cur = 0; cur 343 drivers/gpu/drm/via/via_dmablit.c blitq->cur = cur; cur 368 drivers/gpu/drm/via/via_dmablit.c via_fire_dmablit(dev, blitq->blits[cur], engine); cur 370 drivers/gpu/drm/via/via_dmablit.c blitq->cur = cur; cur 411 drivers/gpu/drm/via/via_dmablit.c slot = handle - blitq->done_blit_handle + blitq->cur - 1; cur 507 drivers/gpu/drm/via/via_dmablit.c while (blitq->serviced != blitq->cur) { cur 553 drivers/gpu/drm/via/via_dmablit.c blitq->cur = 0; cur 67 drivers/gpu/drm/via/via_dmablit.h unsigned cur; cur 121 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c struct vmw_cmdbuf_header *cur; cur 662 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c struct vmw_cmdbuf_header *cur = man->cur; cur 666 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c if (!cur) cur 671 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c __vmw_cmdbuf_header_free(cur); cur 675 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c man->cur->cb_header->length = man->cur_pos; cur 676 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c vmw_cmdbuf_ctx_add(man, man->cur, SVGA_CB_CONTEXT_0); cur 679 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c man->cur = NULL; cur 1003 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c struct vmw_cmdbuf_header *cur; cur 1009 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c cur = man->cur; cur 1010 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c if (cur && (size + man->cur_pos > cur->size || cur 1011 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c ((cur->cb_header->flags & SVGA_CB_FLAG_DX_CONTEXT) && cur 1012 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c ctx_id != cur->cb_header->dxContext))) cur 1015 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c if (!man->cur) { cur 1018 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c interruptible, &man->cur); cur 1024 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c cur = man->cur; cur 1028 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c cur->cb_header->flags |= SVGA_CB_FLAG_DX_CONTEXT; cur 1029 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c cur->cb_header->dxContext = ctx_id; cur 1032 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c cur->reserved = size; cur 1034 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c return (void *) (man->cur->cmd + man->cur_pos); cur 1047 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c struct vmw_cmdbuf_header *cur = man->cur; cur 1051 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c WARN_ON(size > cur->reserved); cur 1054 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c cur->cb_header->flags &= ~SVGA_CB_FLAG_DX_CONTEXT; cur 1112 drivers/gpu/drm/vmwgfx/vmwgfx_cmdbuf.c man->cur = header; cur 118 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c struct vmw_ctx_binding_state *cur; cur 176 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c vmw_binding_state_commit(entry->cur, entry->staged); cur 241 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c node->cur = vmw_context_binding_state(res); cur 752 drivers/gpu/drm/vmwgfx/vmwgfx_execbuf.c ret = vmw_binding_rebind_all(val->cur); cur 473 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c dma_addr_t cur = vmw_piter_dma_addr(&iter); cur 475 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c if (cur != old + PAGE_SIZE) cur 477 drivers/gpu/drm/vmwgfx/vmwgfx_ttm_buffer.c old = cur; cur 239 drivers/hid/wacom.h struct wacom_led *wacom_led_next(struct wacom *wacom, struct wacom_led *cur); cur 1472 drivers/hid/wacom_sys.c struct wacom_led *wacom_led_next(struct wacom *wacom, struct wacom_led *cur) cur 1477 drivers/hid/wacom_sys.c if (!wacom || !cur) cur 1480 drivers/hid/wacom_sys.c group = cur->group; cur 1481 drivers/hid/wacom_sys.c next = cur->id; cur 1485 drivers/hid/wacom_sys.c if (!next_led || next_led == cur) cur 3951 drivers/hid/wacom_wac.c int cur; cur 3958 drivers/hid/wacom_wac.c cur = wacom->led.groups[group].select; cur 3960 drivers/hid/wacom_wac.c led = wacom_led_find(wacom, group, cur); cur 3963 drivers/hid/wacom_wac.c cur, group); cur 1427 drivers/hv/channel_mgmt.c struct list_head *cur, *tmp; cur 1433 drivers/hv/channel_mgmt.c list_for_each_safe(cur, tmp, &primary_channel->sc_list) { cur 1434 drivers/hv/channel_mgmt.c cur_channel = list_entry(cur, struct vmbus_channel, sc_list); cur 316 drivers/hv/connection.c struct list_head *cur, *tmp; cur 329 drivers/hv/connection.c list_for_each_safe(cur, tmp, &channel->sc_list) { cur 330 drivers/hv/connection.c cur_sc = list_entry(cur, struct vmbus_channel, cur 523 drivers/hv/vmbus_drv.c struct list_head *cur; cur 533 drivers/hv/vmbus_drv.c list_for_each(cur, &channel->sc_list) { cur 537 drivers/hv/vmbus_drv.c cur_sc = list_entry(cur, struct vmbus_channel, sc_list); cur 243 drivers/hwmon/lm77.c int i, cur, conf, hyst, crit, min, max; cur 264 drivers/hwmon/lm77.c cur = i2c_smbus_read_word_data(client, 0); cur 280 drivers/hwmon/lm77.c if (((cur & 0x00f0) != 0xf0 && (cur & 0x00f0) != 0x0) cur 292 drivers/hwmon/lm77.c cur = i2c_smbus_read_word_data(client, 0); cur 293 drivers/hwmon/lm77.c if (i2c_smbus_read_word_data(client, 6) != cur cur 294 drivers/hwmon/lm77.c || i2c_smbus_read_word_data(client, 7) != cur) cur 552 drivers/hwmon/lm80.c int i, cur, man_id, dev_id; cur 579 drivers/hwmon/lm80.c cur = i2c_smbus_read_byte_data(client, i); cur 580 drivers/hwmon/lm80.c if ((i2c_smbus_read_byte_data(client, i + 0x40) != cur) cur 581 drivers/hwmon/lm80.c || (i2c_smbus_read_byte_data(client, i + 0x80) != cur) cur 582 drivers/hwmon/lm80.c || (i2c_smbus_read_byte_data(client, i + 0xc0) != cur)) cur 221 drivers/hwmon/lm93.c u8 cur; cur 1057 drivers/hwmon/lm93.c data->block4[i].cur = cur 2114 drivers/hwmon/lm93.c return sprintf(buf, "%d\n", data->block4[nr].cur); cur 414 drivers/hwtracing/coresight/coresight-etb10.c buf->cur = head / PAGE_SIZE; cur 429 drivers/hwtracing/coresight/coresight-etb10.c int i, cur; cur 526 drivers/hwtracing/coresight/coresight-etb10.c cur = buf->cur; cur 531 drivers/hwtracing/coresight/coresight-etb10.c buf_ptr = buf->data_pages[cur] + offset; cur 545 drivers/hwtracing/coresight/coresight-etb10.c cur++; cur 547 drivers/hwtracing/coresight/coresight-etb10.c cur &= buf->nr_pages - 1; cur 95 drivers/hwtracing/coresight/coresight-priv.h unsigned int cur; cur 430 drivers/hwtracing/coresight/coresight-tmc-etf.c buf->cur = head / PAGE_SIZE; cur 445 drivers/hwtracing/coresight/coresight-tmc-etf.c int i, cur; cur 520 drivers/hwtracing/coresight/coresight-tmc-etf.c cur = buf->cur; cur 526 drivers/hwtracing/coresight/coresight-tmc-etf.c buf_ptr = buf->data_pages[cur] + offset; cur 537 drivers/hwtracing/coresight/coresight-tmc-etf.c cur++; cur 539 drivers/hwtracing/coresight/coresight-tmc-etf.c cur &= buf->nr_pages - 1; cur 83 drivers/i2c/busses/i2c-qcom-geni.c struct i2c_msg *cur; cur 192 drivers/i2c/busses/i2c-qcom-geni.c if (gi2c->cur) cur 194 drivers/i2c/busses/i2c-qcom-geni.c gi2c->cur->len, gi2c->cur->addr, gi2c->cur->flags); cur 213 drivers/i2c/busses/i2c-qcom-geni.c struct i2c_msg *cur; cur 222 drivers/i2c/busses/i2c-qcom-geni.c cur = gi2c->cur; cur 224 drivers/i2c/busses/i2c-qcom-geni.c if (!cur || cur 248 drivers/i2c/busses/i2c-qcom-geni.c } else if (cur->flags & I2C_M_RD && cur 255 drivers/i2c/busses/i2c-qcom-geni.c while (gi2c->cur_rd < cur->len && p < sizeof(val)) { cur 256 drivers/i2c/busses/i2c-qcom-geni.c cur->buf[gi2c->cur_rd++] = val & 0xff; cur 260 drivers/i2c/busses/i2c-qcom-geni.c if (gi2c->cur_rd == cur->len) cur 263 drivers/i2c/busses/i2c-qcom-geni.c } else if (!(cur->flags & I2C_M_RD) && cur 270 drivers/i2c/busses/i2c-qcom-geni.c while (gi2c->cur_wr < cur->len && p < sizeof(val)) { cur 271 drivers/i2c/busses/i2c-qcom-geni.c temp = cur->buf[gi2c->cur_wr++]; cur 277 drivers/i2c/busses/i2c-qcom-geni.c if (gi2c->cur_wr == cur->len) { cur 311 drivers/i2c/busses/i2c-qcom-geni.c gi2c->cur = NULL; cur 462 drivers/i2c/busses/i2c-qcom-geni.c gi2c->cur = &msgs[i]; cur 476 drivers/i2c/busses/i2c-qcom-geni.c gi2c->cur = NULL; cur 132 drivers/i2c/muxes/i2c-demux-pinctrl.c int ret, cur = priv->cur_chan; cur 134 drivers/i2c/muxes/i2c-demux-pinctrl.c if (cur < 0) cur 138 drivers/i2c/muxes/i2c-demux-pinctrl.c i2c_put_adapter(priv->chan[cur].parent_adap); cur 140 drivers/i2c/muxes/i2c-demux-pinctrl.c ret = of_changeset_revert(&priv->chan[cur].chgset); cur 142 drivers/i2c/muxes/i2c-demux-pinctrl.c priv->chan[cur].parent_adap = NULL; cur 234 drivers/i3c/master/dw-i3c-master.c struct dw_i3c_xfer *cur; cur 380 drivers/i3c/master/dw-i3c-master.c struct dw_i3c_xfer *xfer = master->xferqueue.cur; cur 413 drivers/i3c/master/dw-i3c-master.c if (master->xferqueue.cur) { cur 416 drivers/i3c/master/dw-i3c-master.c master->xferqueue.cur = xfer; cur 425 drivers/i3c/master/dw-i3c-master.c if (master->xferqueue.cur == xfer) { cur 428 drivers/i3c/master/dw-i3c-master.c master->xferqueue.cur = NULL; cur 453 drivers/i3c/master/dw-i3c-master.c struct dw_i3c_xfer *xfer = master->xferqueue.cur; cur 514 drivers/i3c/master/dw-i3c-master.c master->xferqueue.cur = xfer; cur 403 drivers/i3c/master/i3c-master-cdns.c struct cdns_i3c_xfer *cur; cur 522 drivers/i3c/master/i3c-master-cdns.c struct cdns_i3c_xfer *xfer = master->xferqueue.cur; cur 552 drivers/i3c/master/i3c-master-cdns.c struct cdns_i3c_xfer *xfer = master->xferqueue.cur; cur 619 drivers/i3c/master/i3c-master-cdns.c master->xferqueue.cur = xfer; cur 630 drivers/i3c/master/i3c-master-cdns.c if (master->xferqueue.cur) { cur 633 drivers/i3c/master/i3c-master-cdns.c master->xferqueue.cur = xfer; cur 645 drivers/i3c/master/i3c-master-cdns.c if (master->xferqueue.cur == xfer) { cur 653 drivers/i3c/master/i3c-master-cdns.c master->xferqueue.cur = NULL; cur 203 drivers/ide/ide-tape.c void *cur; cur 874 drivers/ide/ide-tape.c tape->cur = tape->buf; cur 933 drivers/ide/ide-tape.c memset(tape->cur, 0, aligned - tape->valid); cur 967 drivers/ide/ide-tape.c tape->cur = tape->buf; cur 1168 drivers/ide/ide-tape.c if (copy_to_user(buf + done, tape->cur, todo)) cur 1171 drivers/ide/ide-tape.c tape->cur += todo; cur 1216 drivers/ide/ide-tape.c if (copy_from_user(tape->cur, buf + done, todo)) cur 1219 drivers/ide/ide-tape.c tape->cur += todo; cur 1663 drivers/iio/adc/stm32-adc.c const __be32 *cur; cur 1709 drivers/iio/adc/stm32-adc.c of_property_for_each_u32(node, "st,adc-channels", prop, cur, val) { cur 581 drivers/iio/adc/ti_am335x_adc.c const __be32 *cur; cur 585 drivers/iio/adc/ti_am335x_adc.c of_property_for_each_u32(node, "ti,adc-channels", prop, cur, val) { cur 370 drivers/iio/trigger/stm32-timer-trigger.c const char * const *cur = priv->triggers; cur 372 drivers/iio/trigger/stm32-timer-trigger.c while (cur && *cur) { cur 374 drivers/iio/trigger/stm32-timer-trigger.c bool cur_is_trgo = stm32_timer_is_trgo_name(*cur); cur 375 drivers/iio/trigger/stm32-timer-trigger.c bool cur_is_trgo2 = stm32_timer_is_trgo2_name(*cur); cur 378 drivers/iio/trigger/stm32-timer-trigger.c cur++; cur 382 drivers/iio/trigger/stm32-timer-trigger.c trig = devm_iio_trigger_alloc(priv->dev, "%s", *cur); cur 401 drivers/iio/trigger/stm32-timer-trigger.c cur++; cur 481 drivers/iio/trigger/stm32-timer-trigger.c const char * const *cur = priv->valids; cur 487 drivers/iio/trigger/stm32-timer-trigger.c while (cur && *cur) { cur 488 drivers/iio/trigger/stm32-timer-trigger.c if (!strncmp(trig->name, *cur, strlen(trig->name))) { cur 494 drivers/iio/trigger/stm32-timer-trigger.c cur++; cur 2232 drivers/infiniband/core/device.c struct ib_port_data *cur; cur 2235 drivers/infiniband/core/device.c hash_for_each_possible_rcu (ndev_hash, cur, ndev_hash_link, cur 2237 drivers/infiniband/core/device.c if (rcu_access_pointer(cur->netdev) == ndev && cur 2239 drivers/infiniband/core/device.c cur->ib_dev->ops.driver_id == driver_id) && cur 2240 drivers/infiniband/core/device.c ib_device_try_get(cur->ib_dev)) { cur 2241 drivers/infiniband/core/device.c res = cur->ib_dev; cur 126 drivers/infiniband/core/uverbs_cmd.c const void __user *cur; cur 141 drivers/infiniband/core/uverbs_cmd.c iter->cur = attrs->ucore.inbuf + req_len; cur 149 drivers/infiniband/core/uverbs_cmd.c if (iter->cur + len > iter->end) cur 152 drivers/infiniband/core/uverbs_cmd.c if (copy_from_user(val, iter->cur, len)) cur 155 drivers/infiniband/core/uverbs_cmd.c iter->cur += len; cur 162 drivers/infiniband/core/uverbs_cmd.c const void __user *res = iter->cur; cur 164 drivers/infiniband/core/uverbs_cmd.c if (iter->cur + len > iter->end) cur 166 drivers/infiniband/core/uverbs_cmd.c iter->cur += len; cur 172 drivers/infiniband/core/uverbs_cmd.c if (!ib_is_buffer_cleared(iter->cur, iter->end - iter->cur)) cur 476 drivers/infiniband/hw/cxgb4/device.c dev->rdev.stats.pd.total, dev->rdev.stats.pd.cur, cur 479 drivers/infiniband/hw/cxgb4/device.c dev->rdev.stats.qid.total, dev->rdev.stats.qid.cur, cur 482 drivers/infiniband/hw/cxgb4/device.c dev->rdev.stats.srqt.total, dev->rdev.stats.srqt.cur, cur 485 drivers/infiniband/hw/cxgb4/device.c dev->rdev.stats.stag.total, dev->rdev.stats.stag.cur, cur 488 drivers/infiniband/hw/cxgb4/device.c dev->rdev.stats.pbl.total, dev->rdev.stats.pbl.cur, cur 491 drivers/infiniband/hw/cxgb4/device.c dev->rdev.stats.rqt.total, dev->rdev.stats.rqt.cur, cur 494 drivers/infiniband/hw/cxgb4/device.c dev->rdev.stats.ocqp.total, dev->rdev.stats.ocqp.cur, cur 761 drivers/infiniband/hw/cxgb4/device.c rdev->stats.qid.cur -= rdev->qpmask + 1; cur 122 drivers/infiniband/hw/cxgb4/iw_cxgb4.h u64 cur; cur 302 drivers/infiniband/hw/cxgb4/mem.c rdev->stats.stag.cur += 32; cur 303 drivers/infiniband/hw/cxgb4/mem.c if (rdev->stats.stag.cur > rdev->stats.stag.max) cur 304 drivers/infiniband/hw/cxgb4/mem.c rdev->stats.stag.max = rdev->stats.stag.cur; cur 339 drivers/infiniband/hw/cxgb4/mem.c rdev->stats.stag.cur -= 32; cur 203 drivers/infiniband/hw/cxgb4/provider.c rhp->rdev.stats.pd.cur--; cur 231 drivers/infiniband/hw/cxgb4/provider.c rhp->rdev.stats.pd.cur++; cur 232 drivers/infiniband/hw/cxgb4/provider.c if (rhp->rdev.stats.pd.cur > rhp->rdev.stats.pd.max) cur 233 drivers/infiniband/hw/cxgb4/provider.c rhp->rdev.stats.pd.max = rhp->rdev.stats.pd.cur; cur 126 drivers/infiniband/hw/cxgb4/resource.c rdev->stats.qid.cur += rdev->qpmask + 1; cur 157 drivers/infiniband/hw/cxgb4/resource.c if (rdev->stats.qid.cur > rdev->stats.qid.max) cur 158 drivers/infiniband/hw/cxgb4/resource.c rdev->stats.qid.max = rdev->stats.qid.cur; cur 200 drivers/infiniband/hw/cxgb4/resource.c rdev->stats.qid.cur += rdev->qpmask + 1; cur 231 drivers/infiniband/hw/cxgb4/resource.c if (rdev->stats.qid.cur > rdev->stats.qid.max) cur 232 drivers/infiniband/hw/cxgb4/resource.c rdev->stats.qid.max = rdev->stats.qid.cur; cur 271 drivers/infiniband/hw/cxgb4/resource.c rdev->stats.pbl.cur += roundup(size, 1 << MIN_PBL_SHIFT); cur 272 drivers/infiniband/hw/cxgb4/resource.c if (rdev->stats.pbl.cur > rdev->stats.pbl.max) cur 273 drivers/infiniband/hw/cxgb4/resource.c rdev->stats.pbl.max = rdev->stats.pbl.cur; cur 294 drivers/infiniband/hw/cxgb4/resource.c rdev->stats.pbl.cur -= roundup(size, 1 << MIN_PBL_SHIFT); cur 353 drivers/infiniband/hw/cxgb4/resource.c rdev->stats.rqt.cur += roundup(size << 6, 1 << MIN_RQT_SHIFT); cur 354 drivers/infiniband/hw/cxgb4/resource.c if (rdev->stats.rqt.cur > rdev->stats.rqt.max) cur 355 drivers/infiniband/hw/cxgb4/resource.c rdev->stats.rqt.max = rdev->stats.rqt.cur; cur 376 drivers/infiniband/hw/cxgb4/resource.c rdev->stats.rqt.cur -= roundup(size << 6, 1 << MIN_RQT_SHIFT); cur 438 drivers/infiniband/hw/cxgb4/resource.c rdev->stats.srqt.cur++; cur 439 drivers/infiniband/hw/cxgb4/resource.c if (rdev->stats.srqt.cur > rdev->stats.srqt.max) cur 440 drivers/infiniband/hw/cxgb4/resource.c rdev->stats.srqt.max = rdev->stats.srqt.cur; cur 449 drivers/infiniband/hw/cxgb4/resource.c rdev->stats.srqt.cur--; cur 464 drivers/infiniband/hw/cxgb4/resource.c rdev->stats.ocqp.cur += roundup(size, 1 << MIN_OCQP_SHIFT); cur 465 drivers/infiniband/hw/cxgb4/resource.c if (rdev->stats.ocqp.cur > rdev->stats.ocqp.max) cur 466 drivers/infiniband/hw/cxgb4/resource.c rdev->stats.ocqp.max = rdev->stats.ocqp.cur; cur 476 drivers/infiniband/hw/cxgb4/resource.c rdev->stats.ocqp.cur -= roundup(size, 1 << MIN_OCQP_SHIFT); cur 1295 drivers/infiniband/hw/hns/hns_roce_hem.c struct roce_hem_item *cur, *pre; cur 1326 drivers/infiniband/hw/hns/hns_roce_hem.c cur = hem_list_search_item(&mid_bt[level], offset); cur 1327 drivers/infiniband/hw/hns/hns_roce_hem.c if (cur) { cur 1328 drivers/infiniband/hw/hns/hns_roce_hem.c hem_ptrs[level] = cur; cur 1340 drivers/infiniband/hw/hns/hns_roce_hem.c cur = hem_list_alloc_item(hr_dev, start_aligned, end, unit, cur 1342 drivers/infiniband/hw/hns/hns_roce_hem.c if (!cur) { cur 1346 drivers/infiniband/hw/hns/hns_roce_hem.c hem_ptrs[level] = cur; cur 1347 drivers/infiniband/hw/hns/hns_roce_hem.c list_add(&cur->list, &temp_list[level]); cur 1349 drivers/infiniband/hw/hns/hns_roce_hem.c list_add(&cur->sibling, &temp_list[0]); cur 1354 drivers/infiniband/hw/hns/hns_roce_hem.c step = (cur->start - pre->start) / step * BA_BYTE_LEN; cur 1356 drivers/infiniband/hw/hns/hns_roce_hem.c cur->dma_addr); cur 1291 drivers/infiniband/hw/hns/hns_roce_qp.c u32 cur; cur 1293 drivers/infiniband/hw/hns/hns_roce_qp.c cur = hr_wq->head - hr_wq->tail; cur 1294 drivers/infiniband/hw/hns/hns_roce_qp.c if (likely(cur + nreq < hr_wq->max_post)) cur 1299 drivers/infiniband/hw/hns/hns_roce_qp.c cur = hr_wq->head - hr_wq->tail; cur 1302 drivers/infiniband/hw/hns/hns_roce_qp.c return cur + nreq >= hr_wq->max_post; cur 616 drivers/infiniband/hw/mlx4/cq.c unsigned cur; cur 620 drivers/infiniband/hw/mlx4/cq.c cur = wq->head - wq->tail; cur 622 drivers/infiniband/hw/mlx4/cq.c if (cur == 0) cur 625 drivers/infiniband/hw/mlx4/cq.c for (i = 0; i < cur && *npolled < num_entries; i++) { cur 3291 drivers/infiniband/hw/mlx4/qp.c unsigned cur; cur 3294 drivers/infiniband/hw/mlx4/qp.c cur = wq->head - wq->tail; cur 3295 drivers/infiniband/hw/mlx4/qp.c if (likely(cur + nreq < wq->max_post)) cur 3300 drivers/infiniband/hw/mlx4/qp.c cur = wq->head - wq->tail; cur 3303 drivers/infiniband/hw/mlx4/qp.c return cur + nreq >= wq->max_post; cur 390 drivers/infiniband/hw/mlx5/cq.c unsigned int cur; cur 395 drivers/infiniband/hw/mlx5/cq.c cur = wq->head - wq->tail; cur 398 drivers/infiniband/hw/mlx5/cq.c if (cur == 0) cur 401 drivers/infiniband/hw/mlx5/cq.c for (i = 0; i < cur && np < num_entries; i++) { cur 135 drivers/infiniband/hw/mlx5/mem.c u64 cur = 0; cur 175 drivers/infiniband/hw/mlx5/mem.c cur = base + (k << PAGE_SHIFT); cur 176 drivers/infiniband/hw/mlx5/mem.c cur |= access_flags; cur 179 drivers/infiniband/hw/mlx5/mem.c pas[idx] = cpu_to_be64(cur); cur 680 drivers/infiniband/hw/mlx5/mlx5_ib.h u32 cur; cur 121 drivers/infiniband/hw/mlx5/mr.c ent->cur++; cur 215 drivers/infiniband/hw/mlx5/mr.c ent->cur--; cur 312 drivers/infiniband/hw/mlx5/mr.c if (ent->cur < ent->limit) { cur 313 drivers/infiniband/hw/mlx5/mr.c err = add_keys(dev, c, 2 * ent->limit - ent->cur); cur 347 drivers/infiniband/hw/mlx5/mr.c if (cache->ent[i].cur < cache->ent[i].limit) cur 365 drivers/infiniband/hw/mlx5/mr.c if (ent->cur < 2 * ent->limit && !dev->fill_delay) { cur 367 drivers/infiniband/hw/mlx5/mr.c if (ent->cur < 2 * ent->limit) { cur 382 drivers/infiniband/hw/mlx5/mr.c } else if (ent->cur > 2 * ent->limit) { cur 398 drivers/infiniband/hw/mlx5/mr.c if (ent->cur > ent->limit) cur 449 drivers/infiniband/hw/mlx5/mr.c ent->cur--; cur 451 drivers/infiniband/hw/mlx5/mr.c if (ent->cur < ent->limit) cur 484 drivers/infiniband/hw/mlx5/mr.c ent->cur--; cur 486 drivers/infiniband/hw/mlx5/mr.c if (ent->cur < ent->limit) cur 518 drivers/infiniband/hw/mlx5/mr.c if (ent->cur < ent->limit) cur 526 drivers/infiniband/hw/mlx5/mr.c ent->cur++; cur 527 drivers/infiniband/hw/mlx5/mr.c if (ent->cur > 2 * ent->limit) cur 552 drivers/infiniband/hw/mlx5/mr.c ent->cur--; cur 595 drivers/infiniband/hw/mlx5/mr.c debugfs_create_u32("cur", 0400, dir, &ent->cur); cur 4073 drivers/infiniband/hw/mlx5/qp.c unsigned cur; cur 4075 drivers/infiniband/hw/mlx5/qp.c cur = wq->head - wq->tail; cur 4076 drivers/infiniband/hw/mlx5/qp.c if (likely(cur + nreq < wq->max_post)) cur 4081 drivers/infiniband/hw/mlx5/qp.c cur = wq->head - wq->tail; cur 4084 drivers/infiniband/hw/mlx5/qp.c return cur + nreq >= wq->max_post; cur 1727 drivers/infiniband/hw/mthca/mthca_cmd.c int mthca_MODIFY_QP(struct mthca_dev *dev, enum ib_qp_state cur, cur 1775 drivers/infiniband/hw/mthca/mthca_cmd.c if (op[cur][next] == CMD_ERR2RST_QPEE) { cur 1790 drivers/infiniband/hw/mthca/mthca_cmd.c op[cur][next], CMD_TIME_CLASS_C); cur 1824 drivers/infiniband/hw/mthca/mthca_cmd.c op_mod, op[cur][next], CMD_TIME_CLASS_C); cur 308 drivers/infiniband/hw/mthca/mthca_cmd.h int mthca_MODIFY_QP(struct mthca_dev *dev, enum ib_qp_state cur, cur 1569 drivers/infiniband/hw/mthca/mthca_qp.c unsigned cur; cur 1572 drivers/infiniband/hw/mthca/mthca_qp.c cur = wq->head - wq->tail; cur 1573 drivers/infiniband/hw/mthca/mthca_qp.c if (likely(cur + nreq < wq->max)) cur 1578 drivers/infiniband/hw/mthca/mthca_qp.c cur = wq->head - wq->tail; cur 1581 drivers/infiniband/hw/mthca/mthca_qp.c return cur + nreq >= wq->max; cur 906 drivers/infiniband/hw/ocrdma/ocrdma_hw.c struct list_head *cur; cur 910 drivers/infiniband/hw/ocrdma/ocrdma_hw.c list_for_each(cur, head) { cur 912 drivers/infiniband/hw/ocrdma/ocrdma_hw.c qp = list_entry(cur, struct ocrdma_qp, sq_entry); cur 914 drivers/infiniband/hw/ocrdma/ocrdma_hw.c qp = list_entry(cur, struct ocrdma_qp, rq_entry); cur 2957 drivers/infiniband/hw/qedr/verbs.c u32 cur; cur 2968 drivers/infiniband/hw/qedr/verbs.c cur = min_t(u32, len, seg_siz); cur 2969 drivers/infiniband/hw/qedr/verbs.c memcpy(seg_prt, src, cur); cur 2972 drivers/infiniband/hw/qedr/verbs.c seg_prt += cur; cur 2973 drivers/infiniband/hw/qedr/verbs.c seg_siz -= cur; cur 2976 drivers/infiniband/hw/qedr/verbs.c src += cur; cur 2977 drivers/infiniband/hw/qedr/verbs.c len -= cur; cur 202 drivers/iommu/arm-smmu.c it.cur = &pci_sid; cur 738 drivers/iommu/dma-iommu.c struct scatterlist *s, *cur = sg; cur 768 drivers/iommu/dma-iommu.c cur = sg_next(cur); cur 772 drivers/iommu/dma-iommu.c sg_dma_address(cur) = dma_addr + s_iova_off; cur 775 drivers/iommu/dma-iommu.c sg_dma_len(cur) = cur_len; cur 465 drivers/iommu/virtio-iommu.c size_t cur = 0; cur 496 drivers/iommu/virtio-iommu.c cur < viommu->probe_size) { cur 510 drivers/iommu/virtio-iommu.c cur += len; cur 511 drivers/iommu/virtio-iommu.c if (cur >= viommu->probe_size) cur 514 drivers/iommu/virtio-iommu.c prop = (void *)probe->properties + cur; cur 246 drivers/leds/leds-powernv.c const char *cur = NULL; cur 256 drivers/leds/leds-powernv.c while ((cur = of_prop_next_string(p, cur)) != NULL) { cur 267 drivers/leds/leds-powernv.c rc = powernv_led_create(dev, powernv_led, cur); cur 952 drivers/lightnvm/pblk-core.c struct pblk_line *cur) cur 1002 drivers/lightnvm/pblk-core.c if (cur) { cur 1003 drivers/lightnvm/pblk-core.c memcpy(line->lun_bitmap, cur->lun_bitmap, lm->lun_bitmap_len); cur 1004 drivers/lightnvm/pblk-core.c smeta_buf->prev_id = cpu_to_le32(cur->id); cur 1005 drivers/lightnvm/pblk-core.c cur->emeta->buf->next_id = cpu_to_le32(line->id); cur 1546 drivers/lightnvm/pblk-core.c struct pblk_line *cur, *new = NULL; cur 1554 drivers/lightnvm/pblk-core.c cur = l_mg->data_line; cur 1577 drivers/lightnvm/pblk-core.c if (!pblk_line_init_metadata(pblk, new, cur)) { cur 705 drivers/md/bcache/journal.c j->cur = (j->cur == j->w) cur 716 drivers/md/bcache/journal.c j->cur->data->seq = ++j->seq; cur 717 drivers/md/bcache/journal.c j->cur->dirty = false; cur 718 drivers/md/bcache/journal.c j->cur->need_write = false; cur 719 drivers/md/bcache/journal.c j->cur->data->keys = 0; cur 738 drivers/md/bcache/journal.c struct journal_write *w = (j->cur == j->w) cur 760 drivers/md/bcache/journal.c struct journal_write *w = c->journal.cur; cur 849 drivers/md/bcache/journal.c struct journal_write *w = c->journal.cur; cur 874 drivers/md/bcache/journal.c struct journal_write *w = c->journal.cur; cur 922 drivers/md/bcache/journal.c if (c->journal.cur->dirty) cur 121 drivers/md/bcache/journal.h struct journal_write w[2], *cur; cur 1645 drivers/md/bcache/super.c if (c->journal.cur) { cur 1320 drivers/md/dm-ioctl.c static bool is_valid_type(enum dm_queue_mode cur, enum dm_queue_mode new) cur 1322 drivers/md/dm-ioctl.c if (cur == new || cur 1323 drivers/md/dm-ioctl.c (cur == DM_TYPE_BIO_BASED && new == DM_TYPE_DAX_BIO_BASED)) cur 1318 drivers/media/common/cx2341x.c return ctrl && ctrl->val != ctrl->cur.val; cur 312 drivers/media/common/saa7146/saa7146_hlp.c int cur = 0, next = 0; cur 328 drivers/media/common/saa7146/saa7146_hlp.c for( cur = 0, next = 1; next < *count; next++) { cur 329 drivers/media/common/saa7146/saa7146_hlp.c if( values[cur] != values[next]) cur 330 drivers/media/common/saa7146/saa7146_hlp.c values[++cur] = values[next]; cur 333 drivers/media/common/saa7146/saa7146_hlp.c *count = cur + 1; cur 150 drivers/media/i2c/adp1653.c ctrl->cur.val = 0; cur 153 drivers/media/i2c/adp1653.c ctrl->cur.val |= V4L2_FLASH_FAULT_SHORT_CIRCUIT; cur 155 drivers/media/i2c/adp1653.c ctrl->cur.val |= V4L2_FLASH_FAULT_OVER_TEMPERATURE; cur 157 drivers/media/i2c/adp1653.c ctrl->cur.val |= V4L2_FLASH_FAULT_TIMEOUT; cur 159 drivers/media/i2c/adp1653.c ctrl->cur.val |= V4L2_FLASH_FAULT_OVER_VOLTAGE; cur 1162 drivers/media/i2c/imx274.c v_pos = imx274->ctrls.vflip->cur.val ? cur 179 drivers/media/i2c/lm3560.c ctrl->cur.val = fault; cur 191 drivers/media/i2c/m5mols/m5mols_controls.c if ((ctrl->val ^ ctrl->cur.val) & V4L2_LOCK_EXPOSURE) { cur 200 drivers/media/i2c/m5mols/m5mols_controls.c if (((ctrl->val ^ ctrl->cur.val) & V4L2_LOCK_WHITE_BALANCE) cur 213 drivers/media/i2c/m5mols/m5mols_controls.c if ((ctrl->val ^ ctrl->cur.val) & V4L2_LOCK_FOCUS) cur 791 drivers/media/i2c/max2175.c max2175_set_hsls(ctx, ctx->hsls->cur.val); cur 794 drivers/media/i2c/max2175.c max2175_i2s_enable(ctx, ctx->i2s_en->cur.val); cur 1012 drivers/media/i2c/max2175.c max2175_tune_rf_freq(ctx, ctx->freq, ctx->hsls->cur.val); cur 1016 drivers/media/i2c/max2175.c ctx->hsls->cur.val); cur 1088 drivers/media/i2c/max2175.c ctx->rx_mode->cur.val = rx_mode; cur 1091 drivers/media/i2c/max2175.c return max2175_tune_rf_freq(ctx, freq, ctx->hsls->cur.val); cur 1112 drivers/media/i2c/max2175.c max2175_freq_rx_mode_valid(ctx, ctx->rx_mode->cur.val, freq)) cur 1113 drivers/media/i2c/max2175.c ret = max2175_tune_rf_freq(ctx, freq, ctx->hsls->cur.val); cur 1119 drivers/media/i2c/max2175.c ret, ctx->freq, ctx->mode_resolved, ctx->rx_mode->cur.val); cur 400 drivers/media/i2c/msp3400-driver.c state->volume->val = state->volume->cur.val; cur 401 drivers/media/i2c/msp3400-driver.c state->muted->val = state->muted->cur.val; cur 657 drivers/media/i2c/mt9p031.c if (mt9p031->blc_auto->cur.val != 0) { cur 664 drivers/media/i2c/mt9p031.c if (mt9p031->blc_offset->cur.val != 0) { cur 666 drivers/media/i2c/mt9p031.c mt9p031->blc_offset->cur.val); cur 552 drivers/media/i2c/mt9t001.c if (gain->val != gain->cur.val) cur 566 drivers/media/i2c/mt9t001.c if (gain->val == gain->cur.val) cur 190 drivers/media/i2c/s5c73m3/s5c73m3-ctrls.c if ((ctrl->val ^ ctrl->cur.val) & V4L2_LOCK_EXPOSURE) { cur 197 drivers/media/i2c/s5c73m3/s5c73m3-ctrls.c if (((ctrl->val ^ ctrl->cur.val) & V4L2_LOCK_WHITE_BALANCE) cur 205 drivers/media/i2c/s5c73m3/s5c73m3-ctrls.c if ((ctrl->val ^ ctrl->cur.val) & V4L2_LOCK_FOCUS) cur 251 drivers/media/pci/cx18/cx18-av-core.c state->volume->cur.val = state->volume->default_value = default_volume; cur 745 drivers/media/pci/cx18/cx18-driver.c cx->temporal_strength = cx->cxhdl.video_temporal_filter->cur.val; cur 746 drivers/media/pci/cx18/cx18-driver.c cx->spatial_strength = cx->cxhdl.video_spatial_filter->cur.val; cur 747 drivers/media/pci/cx18/cx18-driver.c cx->filter_mode = cx->cxhdl.video_spatial_filter_mode->cur.val | cur 748 drivers/media/pci/cx18/cx18-driver.c (cx->cxhdl.video_temporal_filter_mode->cur.val << 1) | cur 749 drivers/media/pci/cx18/cx18-driver.c (cx->cxhdl.video_median_filter_type->cur.val << 2); cur 168 drivers/media/pci/mantis/mantis_core.c u32 cur; cur 170 drivers/media/pci/mantis/mantis_core.c cur = mmread(MANTIS_GPIF_ADDR); cur 172 drivers/media/pci/mantis/mantis_core.c mantis->gpio_status = cur | (1 << bitpos); cur 174 drivers/media/pci/mantis/mantis_core.c mantis->gpio_status = cur & (~(1 << bitpos)); cur 69 drivers/media/pci/mantis/mantis_ioc.c u32 cur; cur 72 drivers/media/pci/mantis/mantis_ioc.c cur = mmread(MANTIS_GPIF_ADDR); cur 74 drivers/media/pci/mantis/mantis_ioc.c mantis->gpio_status = cur | (1 << bitpos); cur 76 drivers/media/pci/mantis/mantis_ioc.c mantis->gpio_status = cur & (~(1 << bitpos)); cur 643 drivers/media/platform/exynos4-is/fimc-core.c fimc_set_color_effect(ctx, ctrls->colorfx->cur.val); cur 668 drivers/media/platform/exynos4-is/fimc-core.c if (ctrl->cur.val > ctrl->maximum) cur 669 drivers/media/platform/exynos4-is/fimc-core.c ctrl->cur.val = ctrl->maximum; cur 810 drivers/media/platform/qcom/camss/camss-csid.c if (csid->testgen_mode->cur.val == 0) { cur 860 drivers/media/platform/qcom/camss/camss-csid.c if (csid->testgen_mode->cur.val == 0) { cur 1250 drivers/media/platform/qcom/camss/camss-csid.c if (csid->testgen_mode->cur.val != 0) cur 19 drivers/media/platform/sti/delta/delta-mjpeg-hdr.c char *cur = str; cur 25 drivers/media/platform/sti/delta/delta-mjpeg-hdr.c snprintf(cur, left, "[MJPEG header]\n" cur 813 drivers/media/platform/ti-vpe/vpe.c const struct vpe_dei_regs *cur = &dei_regs; cur 815 drivers/media/platform/ti-vpe/vpe.c dei_mmr[2] = cur->mdt_spacial_freq_thr_reg; cur 816 drivers/media/platform/ti-vpe/vpe.c dei_mmr[3] = cur->edi_config_reg; cur 817 drivers/media/platform/ti-vpe/vpe.c dei_mmr[4] = cur->edi_lut_reg0; cur 818 drivers/media/platform/ti-vpe/vpe.c dei_mmr[5] = cur->edi_lut_reg1; cur 819 drivers/media/platform/ti-vpe/vpe.c dei_mmr[6] = cur->edi_lut_reg2; cur 820 drivers/media/platform/ti-vpe/vpe.c dei_mmr[7] = cur->edi_lut_reg3; cur 615 drivers/media/platform/vicodec/codec-fwht.c decide_blocktype(const u8 *cur, const u8 *reference, s16 *deltablock, cur 625 drivers/media/platform/vicodec/codec-fwht.c fill_encoder_block(cur, tmp, stride, input_step); cur 500 drivers/media/platform/vivid/vivid-kthread-cap.c dev->brightness->cur.val, cur 501 drivers/media/platform/vivid/vivid-kthread-cap.c dev->contrast->cur.val, cur 502 drivers/media/platform/vivid/vivid-kthread-cap.c dev->saturation->cur.val, cur 503 drivers/media/platform/vivid/vivid-kthread-cap.c dev->hue->cur.val); cur 507 drivers/media/platform/vivid/vivid-kthread-cap.c dev->autogain->cur.val, gain, dev->alpha->cur.val); cur 513 drivers/media/platform/vivid/vivid-kthread-cap.c dev->volume->cur.val, dev->mute->cur.val); cur 518 drivers/media/platform/vivid/vivid-kthread-cap.c dev->int32->cur.val, cur 520 drivers/media/platform/vivid/vivid-kthread-cap.c dev->bitmask->cur.val); cur 523 drivers/media/platform/vivid/vivid-kthread-cap.c dev->boolean->cur.val, cur 524 drivers/media/platform/vivid/vivid-kthread-cap.c dev->menu->qmenu[dev->menu->cur.val], cur 528 drivers/media/platform/vivid/vivid-kthread-cap.c dev->int_menu->qmenu_int[dev->int_menu->cur.val], cur 529 drivers/media/platform/vivid/vivid-kthread-cap.c dev->int_menu->cur.val); cur 70 drivers/media/platform/vivid/vivid-radio-common.c rds->picode = dev->radio_tx_rds_pi->cur.val; cur 71 drivers/media/platform/vivid/vivid-radio-common.c rds->pty = dev->radio_tx_rds_pty->cur.val; cur 72 drivers/media/platform/vivid/vivid-radio-common.c rds->mono_stereo = dev->radio_tx_rds_mono_stereo->cur.val; cur 73 drivers/media/platform/vivid/vivid-radio-common.c rds->art_head = dev->radio_tx_rds_art_head->cur.val; cur 74 drivers/media/platform/vivid/vivid-radio-common.c rds->compressed = dev->radio_tx_rds_compressed->cur.val; cur 75 drivers/media/platform/vivid/vivid-radio-common.c rds->dyn_pty = dev->radio_tx_rds_dyn_pty->cur.val; cur 76 drivers/media/platform/vivid/vivid-radio-common.c rds->ta = dev->radio_tx_rds_ta->cur.val; cur 77 drivers/media/platform/vivid/vivid-radio-common.c rds->tp = dev->radio_tx_rds_tp->cur.val; cur 78 drivers/media/platform/vivid/vivid-radio-common.c rds->ms = dev->radio_tx_rds_ms->cur.val; cur 159 drivers/media/platform/vsp1/vsp1_hgo.c hgo->max_rgb = hgo->ctrls.max_rgb->cur.val; cur 161 drivers/media/platform/vsp1/vsp1_hgo.c hgo->num_bins = hgo_num_bins[hgo->ctrls.num_bins->cur.val]; cur 219 drivers/media/platform/xilinx/xilinx-tpg.c XTPG_PATTERN_MASK, xtpg->pattern->cur.val); cur 225 drivers/media/platform/xilinx/xilinx-tpg.c passthrough = xtpg->pattern->cur.val == 0; cur 281 drivers/media/radio/radio-isa.c ops->s_mute_volume(isa, true, isa->volume ? isa->volume->cur.val : 0); cur 158 drivers/media/radio/radio-sf16fmr2.c balance = fmr2->balance->cur.val; cur 162 drivers/media/radio/radio-sf16fmr2.c volume = fmr2->volume->cur.val; cur 382 drivers/media/tuners/e4000.c dev->lna_gain_auto->cur.val, dev->lna_gain_auto->val, cur 383 drivers/media/tuners/e4000.c dev->lna_gain->cur.val, dev->lna_gain->val); cur 385 drivers/media/tuners/e4000.c if (dev->lna_gain_auto->val && dev->if_gain_auto->cur.val) cur 389 drivers/media/tuners/e4000.c else if (dev->if_gain_auto->cur.val) cur 418 drivers/media/tuners/e4000.c dev->mixer_gain_auto->cur.val, dev->mixer_gain_auto->val, cur 419 drivers/media/tuners/e4000.c dev->mixer_gain->cur.val, dev->mixer_gain->val); cur 451 drivers/media/tuners/e4000.c dev->if_gain_auto->cur.val, dev->if_gain_auto->val, cur 452 drivers/media/tuners/e4000.c dev->if_gain->cur.val, dev->if_gain->val); cur 454 drivers/media/tuners/e4000.c if (dev->if_gain_auto->val && dev->lna_gain_auto->cur.val) cur 456 drivers/media/tuners/e4000.c else if (dev->lna_gain_auto->cur.val) cur 474 drivers/media/tuners/fc2580.c ctrl->id, ctrl->name, ctrl->cur.val, ctrl->val); cur 267 drivers/media/tuners/msi001.c ret = msi001_set_gain(dev, dev->lna_gain->cur.val, cur 268 drivers/media/tuners/msi001.c dev->mixer_gain->cur.val, dev->if_gain->cur.val); cur 397 drivers/media/tuners/msi001.c dev->mixer_gain->cur.val, cur 398 drivers/media/tuners/msi001.c dev->if_gain->cur.val); cur 401 drivers/media/tuners/msi001.c ret = msi001_set_gain(dev, dev->lna_gain->cur.val, cur 403 drivers/media/tuners/msi001.c dev->if_gain->cur.val); cur 406 drivers/media/tuners/msi001.c ret = msi001_set_gain(dev, dev->lna_gain->cur.val, cur 407 drivers/media/tuners/msi001.c dev->mixer_gain->cur.val, cur 868 drivers/media/usb/airspy/airspy.c s->lna_gain_auto->cur.val, s->lna_gain_auto->val, cur 869 drivers/media/usb/airspy/airspy.c s->lna_gain->cur.val, s->lna_gain->val); cur 895 drivers/media/usb/airspy/airspy.c s->mixer_gain_auto->cur.val, s->mixer_gain_auto->val, cur 896 drivers/media/usb/airspy/airspy.c s->mixer_gain->cur.val, s->mixer_gain->val); cur 921 drivers/media/usb/airspy/airspy.c dev_dbg(s->dev, "val=%d->%d\n", s->if_gain->cur.val, s->if_gain->val); cur 878 drivers/media/usb/gspca/conex.c setbrightness(gspca_dev, ctrl->val, sd->sat->cur.val); cur 881 drivers/media/usb/gspca/conex.c setcontrast(gspca_dev, ctrl->val, sd->sat->cur.val); cur 884 drivers/media/usb/gspca/conex.c setbrightness(gspca_dev, sd->brightness->cur.val, ctrl->val); cur 885 drivers/media/usb/gspca/conex.c setcontrast(gspca_dev, sd->contrast->cur.val, ctrl->val); cur 2213 drivers/media/usb/gspca/sn9c20x.c s32 curqual = sd->jpegqual->cur.val; cur 2221 drivers/media/usb/gspca/sn9c20x.c sd->jpegqual->cur.val = new_qual; cur 3970 drivers/media/usb/gspca/topro.c s32 old = gspca_dev->gain->cur.val ? cur 3971 drivers/media/usb/gspca/topro.c gspca_dev->gain->cur.val : 1; cur 311 drivers/media/usb/hackrf/hackrf.c dev->rx_bandwidth->cur.val = uitmp; cur 327 drivers/media/usb/hackrf/hackrf.c dev->tx_bandwidth->cur.val = uitmp; cur 376 drivers/media/usb/hackrf/hackrf.c dev->rx_rf_gain->cur.val, dev->rx_rf_gain->val); cur 388 drivers/media/usb/hackrf/hackrf.c dev->tx_rf_gain->cur.val, dev->tx_rf_gain->val); cur 400 drivers/media/usb/hackrf/hackrf.c dev->rx_lna_gain->cur.val, dev->rx_lna_gain->val); cur 411 drivers/media/usb/hackrf/hackrf.c dev->rx_if_gain->cur.val, dev->rx_if_gain->val); cur 422 drivers/media/usb/hackrf/hackrf.c dev->tx_lna_gain->cur.val, dev->tx_lna_gain->val); cur 47 drivers/media/v4l2-core/v4l2-ctrls.c return master->is_auto && master->cur.val == master->manual_mode_value; cur 2496 drivers/media/v4l2-core/v4l2-ctrls.c ctrl->cur.val = ctrl->val = def; cur 2504 drivers/media/v4l2-core/v4l2-ctrls.c ctrl->p_cur.p = &ctrl->cur.val; cur 130 drivers/mfd/ti_am335x_tscadc.c const __be32 *cur; cur 147 drivers/mfd/ti_am335x_tscadc.c of_property_for_each_u32(node, "ti,adc-channels", prop, cur, val) { cur 220 drivers/misc/cxl/guest.c struct irq_avail *cur; cur 223 drivers/misc/cxl/guest.c cur = &adapter->guest->irq_avail[i]; cur 224 drivers/misc/cxl/guest.c n = bitmap_find_next_zero_area(cur->bitmap, cur->range, cur 226 drivers/misc/cxl/guest.c if (n < cur->range) { cur 227 drivers/misc/cxl/guest.c bitmap_set(cur->bitmap, n, len); cur 228 drivers/misc/cxl/guest.c *irq = cur->offset + n; cur 241 drivers/misc/cxl/guest.c struct irq_avail *cur; cur 247 drivers/misc/cxl/guest.c cur = &adapter->guest->irq_avail[i]; cur 248 drivers/misc/cxl/guest.c if (irq >= cur->offset && cur 249 drivers/misc/cxl/guest.c (irq + len) <= (cur->offset + cur->range)) { cur 250 drivers/misc/cxl/guest.c n = irq - cur->offset; cur 251 drivers/misc/cxl/guest.c bitmap_clear(cur->bitmap, n, len); cur 1043 drivers/misc/cxl/guest.c struct irq_avail *cur; cur 1049 drivers/misc/cxl/guest.c cur = &adapter->guest->irq_avail[i]; cur 1050 drivers/misc/cxl/guest.c kfree(cur->bitmap); cur 287 drivers/misc/cxl/of.c struct irq_avail *cur; cur 308 drivers/misc/cxl/of.c cur = &adapter->guest->irq_avail[i]; cur 309 drivers/misc/cxl/of.c cur->offset = be32_to_cpu(ranges[i * 2]); cur 310 drivers/misc/cxl/of.c cur->range = be32_to_cpu(ranges[i * 2 + 1]); cur 311 drivers/misc/cxl/of.c cur->bitmap = kcalloc(BITS_TO_LONGS(cur->range), cur 312 drivers/misc/cxl/of.c sizeof(*cur->bitmap), GFP_KERNEL); cur 313 drivers/misc/cxl/of.c if (cur->bitmap == NULL) cur 315 drivers/misc/cxl/of.c if (cur->offset < adapter->guest->irq_base_offset) cur 316 drivers/misc/cxl/of.c adapter->guest->irq_base_offset = cur->offset; cur 319 drivers/misc/cxl/of.c cur->offset, cur->offset + cur->range - 1, cur 320 drivers/misc/cxl/of.c cur->range); cur 328 drivers/misc/cxl/of.c cur = &adapter->guest->irq_avail[i]; cur 329 drivers/misc/cxl/of.c kfree(cur->bitmap); cur 183 drivers/misc/ibmvmc.c queue->cur = 0; cur 211 drivers/misc/ibmvmc.c crq = &queue->msgs[queue->cur]; cur 213 drivers/misc/ibmvmc.c if (++queue->cur == queue->size) cur 214 drivers/misc/ibmvmc.c queue->cur = 0; cur 2150 drivers/misc/ibmvmc.c queue->cur = 0; cur 154 drivers/misc/ibmvmc.h int size, cur; cur 442 drivers/misc/lkdtm/core.c struct crashpoint *cur = &crashpoints[i]; cur 444 drivers/misc/lkdtm/core.c debugfs_create_file(cur->name, 0644, lkdtm_debugfs_root, cur, cur 445 drivers/misc/lkdtm/core.c &cur->fops); cur 15 drivers/misc/ocxl/pasid.c struct id_range *cur; cur 18 drivers/misc/ocxl/pasid.c list_for_each_entry(cur, head, list) { cur 19 drivers/misc/ocxl/pasid.c pr_debug("Range %d->%d\n", cur->start, cur->end); cur 28 drivers/misc/ocxl/pasid.c struct id_range *cur, *new; cur 37 drivers/misc/ocxl/pasid.c list_for_each_entry(cur, head, list) { cur 38 drivers/misc/ocxl/pasid.c if ((cur->start - last_end) > size) cur 40 drivers/misc/ocxl/pasid.c last_end = cur->end; cur 41 drivers/misc/ocxl/pasid.c pos = &cur->list; cur 65 drivers/misc/ocxl/pasid.c struct id_range *cur, *tmp; cur 67 drivers/misc/ocxl/pasid.c list_for_each_entry_safe(cur, tmp, head, list) { cur 68 drivers/misc/ocxl/pasid.c if (cur->start == start && cur->end == (start + size - 1)) { cur 70 drivers/misc/ocxl/pasid.c list_del(&cur->list); cur 71 drivers/misc/ocxl/pasid.c kfree(cur); cur 49 drivers/misc/vmw_vmci/vmci_event.c struct vmci_subscription *cur, *p2; cur 50 drivers/misc/vmw_vmci/vmci_event.c list_for_each_entry_safe(cur, p2, &subscriber_array[e], node) { cur 58 drivers/misc/vmw_vmci/vmci_event.c list_del(&cur->node); cur 59 drivers/misc/vmw_vmci/vmci_event.c kfree(cur); cur 72 drivers/misc/vmw_vmci/vmci_event.c struct vmci_subscription *cur; cur 73 drivers/misc/vmw_vmci/vmci_event.c list_for_each_entry(cur, &subscriber_array[e], node) { cur 74 drivers/misc/vmw_vmci/vmci_event.c if (cur->id == sub_id) cur 75 drivers/misc/vmw_vmci/vmci_event.c return cur; cur 87 drivers/misc/vmw_vmci/vmci_event.c struct vmci_subscription *cur; cur 92 drivers/misc/vmw_vmci/vmci_event.c list_for_each_entry_rcu(cur, subscriber_list, node) { cur 93 drivers/misc/vmw_vmci/vmci_event.c cur->callback(cur->id, &event_msg->event_data, cur 94 drivers/misc/vmw_vmci/vmci_event.c cur->callback_data); cur 625 drivers/mmc/host/mmci.c struct dma_chan *cur; cur 733 drivers/mmc/host/mmci.c dmaengine_terminate_all(dmae->cur); cur 735 drivers/mmc/host/mmci.c dmae->cur = NULL; cur 783 drivers/mmc/host/mmci.c dmae->cur = NULL; cur 864 drivers/mmc/host/mmci.c if (dmae->cur && dmae->desc_current) cur 868 drivers/mmc/host/mmci.c return _mmci_dmae_prep_data(host, data, &dmae->cur, cur 878 drivers/mmc/host/mmci.c dma_async_issue_pending(dmae->cur); cur 896 drivers/mmc/host/mmci.c dmae->cur = next->chan; cur 924 drivers/mmc/host/mmci.c if (dmae->cur == next->chan) { cur 926 drivers/mmc/host/mmci.c dmae->cur = NULL; cur 206 drivers/mtd/mtdswap.c struct swap_eb *cur; cur 211 drivers/mtd/mtdswap.c cur = rb_entry(parent, struct swap_eb, rb); cur 212 drivers/mtd/mtdswap.c if (eb->erase_count > cur->erase_count) cur 130 drivers/mtd/nand/raw/nand_ecc.c uint32_t cur; /* current value in buffer */ cur 160 drivers/mtd/nand/raw/nand_ecc.c cur = *bp++; cur 161 drivers/mtd/nand/raw/nand_ecc.c tmppar = cur; cur 162 drivers/mtd/nand/raw/nand_ecc.c rp4 ^= cur; cur 163 drivers/mtd/nand/raw/nand_ecc.c cur = *bp++; cur 164 drivers/mtd/nand/raw/nand_ecc.c tmppar ^= cur; cur 166 drivers/mtd/nand/raw/nand_ecc.c cur = *bp++; cur 167 drivers/mtd/nand/raw/nand_ecc.c tmppar ^= cur; cur 168 drivers/mtd/nand/raw/nand_ecc.c rp4 ^= cur; cur 169 drivers/mtd/nand/raw/nand_ecc.c cur = *bp++; cur 170 drivers/mtd/nand/raw/nand_ecc.c tmppar ^= cur; cur 173 drivers/mtd/nand/raw/nand_ecc.c cur = *bp++; cur 174 drivers/mtd/nand/raw/nand_ecc.c tmppar ^= cur; cur 175 drivers/mtd/nand/raw/nand_ecc.c rp4 ^= cur; cur 176 drivers/mtd/nand/raw/nand_ecc.c rp6 ^= cur; cur 177 drivers/mtd/nand/raw/nand_ecc.c cur = *bp++; cur 178 drivers/mtd/nand/raw/nand_ecc.c tmppar ^= cur; cur 179 drivers/mtd/nand/raw/nand_ecc.c rp6 ^= cur; cur 180 drivers/mtd/nand/raw/nand_ecc.c cur = *bp++; cur 181 drivers/mtd/nand/raw/nand_ecc.c tmppar ^= cur; cur 182 drivers/mtd/nand/raw/nand_ecc.c rp4 ^= cur; cur 183 drivers/mtd/nand/raw/nand_ecc.c cur = *bp++; cur 184 drivers/mtd/nand/raw/nand_ecc.c tmppar ^= cur; cur 187 drivers/mtd/nand/raw/nand_ecc.c cur = *bp++; cur 188 drivers/mtd/nand/raw/nand_ecc.c tmppar ^= cur; cur 189 drivers/mtd/nand/raw/nand_ecc.c rp4 ^= cur; cur 190 drivers/mtd/nand/raw/nand_ecc.c rp6 ^= cur; cur 191 drivers/mtd/nand/raw/nand_ecc.c rp8 ^= cur; cur 192 drivers/mtd/nand/raw/nand_ecc.c cur = *bp++; cur 193 drivers/mtd/nand/raw/nand_ecc.c tmppar ^= cur; cur 194 drivers/mtd/nand/raw/nand_ecc.c rp6 ^= cur; cur 195 drivers/mtd/nand/raw/nand_ecc.c rp8 ^= cur; cur 196 drivers/mtd/nand/raw/nand_ecc.c cur = *bp++; cur 197 drivers/mtd/nand/raw/nand_ecc.c tmppar ^= cur; cur 198 drivers/mtd/nand/raw/nand_ecc.c rp4 ^= cur; cur 199 drivers/mtd/nand/raw/nand_ecc.c rp8 ^= cur; cur 200 drivers/mtd/nand/raw/nand_ecc.c cur = *bp++; cur 201 drivers/mtd/nand/raw/nand_ecc.c tmppar ^= cur; cur 202 drivers/mtd/nand/raw/nand_ecc.c rp8 ^= cur; cur 204 drivers/mtd/nand/raw/nand_ecc.c cur = *bp++; cur 205 drivers/mtd/nand/raw/nand_ecc.c tmppar ^= cur; cur 206 drivers/mtd/nand/raw/nand_ecc.c rp4 ^= cur; cur 207 drivers/mtd/nand/raw/nand_ecc.c rp6 ^= cur; cur 208 drivers/mtd/nand/raw/nand_ecc.c cur = *bp++; cur 209 drivers/mtd/nand/raw/nand_ecc.c tmppar ^= cur; cur 210 drivers/mtd/nand/raw/nand_ecc.c rp6 ^= cur; cur 211 drivers/mtd/nand/raw/nand_ecc.c cur = *bp++; cur 212 drivers/mtd/nand/raw/nand_ecc.c tmppar ^= cur; cur 213 drivers/mtd/nand/raw/nand_ecc.c rp4 ^= cur; cur 214 drivers/mtd/nand/raw/nand_ecc.c cur = *bp++; cur 215 drivers/mtd/nand/raw/nand_ecc.c tmppar ^= cur; cur 409 drivers/mtd/nand/raw/s3c2410.c unsigned long cur; cur 417 drivers/mtd/nand/raw/s3c2410.c cur = readl(info->sel_reg); cur 420 drivers/mtd/nand/raw/s3c2410.c cur |= info->sel_bit; cur 432 drivers/mtd/nand/raw/s3c2410.c cur &= ~info->sel_bit; cur 435 drivers/mtd/nand/raw/s3c2410.c writel(cur, info->sel_reg); cur 3444 drivers/net/bonding/bond_main.c int cur = 0, max = 0; cur 3458 drivers/net/bonding/bond_main.c dev_stack[cur] = now; cur 3459 drivers/net/bonding/bond_main.c iter_stack[cur++] = iter; cur 3460 drivers/net/bonding/bond_main.c if (max <= cur) cur 3461 drivers/net/bonding/bond_main.c max = cur; cur 3466 drivers/net/bonding/bond_main.c if (!cur) cur 3468 drivers/net/bonding/bond_main.c next = dev_stack[--cur]; cur 3469 drivers/net/bonding/bond_main.c niter = iter_stack[cur]; cur 417 drivers/net/ethernet/amd/xgbe/xgbe-desc.c ring->cur = 0; cur 461 drivers/net/ethernet/amd/xgbe/xgbe-desc.c ring->cur = 0; cur 535 drivers/net/ethernet/amd/xgbe/xgbe-desc.c DBGPR("-->xgbe_map_tx_skb: cur = %d\n", ring->cur); cur 538 drivers/net/ethernet/amd/xgbe/xgbe-desc.c start_index = ring->cur; cur 539 drivers/net/ethernet/amd/xgbe/xgbe-desc.c cur_index = ring->cur; cur 1413 drivers/net/ethernet/amd/xgbe/xgbe-dev.c int start_index = ring->cur; cur 1491 drivers/net/ethernet/amd/xgbe/xgbe-dev.c unsigned int start_index = ring->cur; cur 1652 drivers/net/ethernet/amd/xgbe/xgbe-dev.c rdata = XGBE_GET_DESC_DATA(ring, ring->cur); cur 1677 drivers/net/ethernet/amd/xgbe/xgbe-dev.c int start_index = ring->cur; cur 1678 drivers/net/ethernet/amd/xgbe/xgbe-dev.c int cur_index = ring->cur; cur 1889 drivers/net/ethernet/amd/xgbe/xgbe-dev.c ring->cur = cur_index + 1; cur 1899 drivers/net/ethernet/amd/xgbe/xgbe-dev.c (ring->cur - 1) & (ring->rdesc_count - 1)); cur 1914 drivers/net/ethernet/amd/xgbe/xgbe-dev.c DBGPR("-->xgbe_dev_read: cur = %d\n", ring->cur); cur 1916 drivers/net/ethernet/amd/xgbe/xgbe-dev.c rdata = XGBE_GET_DESC_DATA(ring, ring->cur); cur 1927 drivers/net/ethernet/amd/xgbe/xgbe-dev.c xgbe_dump_rx_desc(pdata, ring, ring->cur); cur 2059 drivers/net/ethernet/amd/xgbe/xgbe-dev.c ring->cur & (ring->rdesc_count - 1), ring->cur); cur 265 drivers/net/ethernet/amd/xgbe/xgbe-drv.c return (ring->rdesc_count - (ring->cur - ring->dirty)); cur 270 drivers/net/ethernet/amd/xgbe/xgbe-drv.c return (ring->cur - ring->dirty); cur 2524 drivers/net/ethernet/amd/xgbe/xgbe-drv.c while (ring->dirty != ring->cur) { cur 2624 drivers/net/ethernet/amd/xgbe/xgbe-drv.c unsigned int cur; cur 2632 drivers/net/ethernet/amd/xgbe/xgbe-drv.c cur = ring->cur; cur 2640 drivers/net/ethernet/amd/xgbe/xgbe-drv.c (ring->dirty != cur)) { cur 2710 drivers/net/ethernet/amd/xgbe/xgbe-drv.c rdata = XGBE_GET_DESC_DATA(ring, ring->cur); cur 2713 drivers/net/ethernet/amd/xgbe/xgbe-drv.c DBGPR(" cur = %d\n", ring->cur); cur 2728 drivers/net/ethernet/amd/xgbe/xgbe-drv.c rdata = XGBE_GET_DESC_DATA(ring, ring->cur); cur 2737 drivers/net/ethernet/amd/xgbe/xgbe-drv.c ring->cur++; cur 2861 drivers/net/ethernet/amd/xgbe/xgbe-drv.c rdata = XGBE_GET_DESC_DATA(ring, ring->cur); cur 476 drivers/net/ethernet/amd/xgbe/xgbe.h unsigned int cur; cur 818 drivers/net/ethernet/apple/bmac.c unsigned int counter, cur = curval, next = nxtval; cur 827 drivers/net/ethernet/apple/bmac.c if ((cur & 0x80000000) == 0) high_crc_set = 0; cur 830 drivers/net/ethernet/apple/bmac.c cur = cur << 1; cur 838 drivers/net/ethernet/apple/bmac.c if (high_crc_set ^ low_data_set) cur = cur ^ CRC32_POLY_BE; cur 840 drivers/net/ethernet/apple/bmac.c return cur; cur 78 drivers/net/ethernet/atheros/alx/main.c u16 cur, next, count = 0; cur 80 drivers/net/ethernet/atheros/alx/main.c next = cur = rxq->write_idx; cur 83 drivers/net/ethernet/atheros/alx/main.c cur_buf = &rxq->bufs[cur]; cur 86 drivers/net/ethernet/atheros/alx/main.c struct alx_rfd *rfd = &rxq->rfd[cur]; cur 125 drivers/net/ethernet/atheros/alx/main.c cur = next; cur 128 drivers/net/ethernet/atheros/alx/main.c cur_buf = &rxq->bufs[cur]; cur 135 drivers/net/ethernet/atheros/alx/main.c rxq->write_idx = cur; cur 136 drivers/net/ethernet/atheros/alx/main.c alx_write_mem16(&alx->hw, ALX_RFD_PIDX, cur); cur 619 drivers/net/ethernet/broadcom/b44.c u32 cur, cons; cur 622 drivers/net/ethernet/broadcom/b44.c cur = br32(bp, B44_DMATX_STAT) & DMATX_STAT_CDMASK; cur 623 drivers/net/ethernet/broadcom/b44.c cur /= sizeof(struct dma_desc); cur 626 drivers/net/ethernet/broadcom/b44.c for (cons = bp->tx_cons; cons != cur; cons = NEXT_TX(cons)) { cur 2507 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c u32 cur; cur 2512 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c cur = SHMEM2_RD(bp, os_driver_state[BP_FW_MB_IDX(bp)]); cur 2514 drivers/net/ethernet/broadcom/bnx2x/bnx2x_cmn.c cur, state); cur 3017 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c u64 cur[BNX2X_MCAST_VEC_SZ], req[BNX2X_MCAST_VEC_SZ]; cur 3024 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c memcpy(cur, o->registry.aprox_match.vec, cur 3049 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sp.c bool b_current = !!BIT_VEC64_TEST_BIT(cur, i); cur 1376 drivers/net/ethernet/fealnx.c struct fealnx_desc *cur; cur 1386 drivers/net/ethernet/fealnx.c cur = &np->tx_ring[i]; cur 1387 drivers/net/ethernet/fealnx.c if (cur->skbuff) { cur 1388 drivers/net/ethernet/fealnx.c pci_unmap_single(np->pci_dev, cur->buffer, cur 1389 drivers/net/ethernet/fealnx.c cur->skbuff->len, PCI_DMA_TODEVICE); cur 1390 drivers/net/ethernet/fealnx.c dev_kfree_skb_any(cur->skbuff); cur 1391 drivers/net/ethernet/fealnx.c cur->skbuff = NULL; cur 1393 drivers/net/ethernet/fealnx.c cur->status = 0; cur 1394 drivers/net/ethernet/fealnx.c cur->control = 0; /* needed? */ cur 1396 drivers/net/ethernet/fealnx.c cur->next_desc = np->tx_ring_dma + cur 1398 drivers/net/ethernet/fealnx.c cur->next_desc_logical = &np->tx_ring[i + 1]; cur 1410 drivers/net/ethernet/fealnx.c struct fealnx_desc *cur = np->cur_rx; cur 1416 drivers/net/ethernet/fealnx.c if (cur->skbuff) cur 1417 drivers/net/ethernet/fealnx.c cur->status = RXOWN; cur 1418 drivers/net/ethernet/fealnx.c cur = cur->next_desc_logical; cur 1642 drivers/net/ethernet/fealnx.c struct fealnx_desc *cur; cur 1645 drivers/net/ethernet/fealnx.c cur = np->cur_rx; cur 1648 drivers/net/ethernet/fealnx.c if ((!(cur->status & RXOWN)) && cur 1649 drivers/net/ethernet/fealnx.c (cur->status & RXLSD)) cur 1652 drivers/net/ethernet/fealnx.c cur = cur->next_desc_logical; cur 465 drivers/net/ethernet/freescale/fec.h struct bufdesc *cur; cur 310 drivers/net/ethernet/freescale/fec_main.c (const char *)txq->bd.cur) >> txq->bd.dsize_log2) - 1; cur 350 drivers/net/ethernet/freescale/fec_main.c bdp == txq->bd.cur ? 'S' : ' ', cur 389 drivers/net/ethernet/freescale/fec_main.c struct bufdesc *bdp = txq->bd.cur; cur 462 drivers/net/ethernet/freescale/fec_main.c bdp = txq->bd.cur; cur 500 drivers/net/ethernet/freescale/fec_main.c bdp = txq->bd.cur; cur 590 drivers/net/ethernet/freescale/fec_main.c txq->bd.cur = bdp; cur 719 drivers/net/ethernet/freescale/fec_main.c struct bufdesc *bdp = txq->bd.cur; cur 780 drivers/net/ethernet/freescale/fec_main.c txq->bd.cur = bdp; cur 855 drivers/net/ethernet/freescale/fec_main.c rxq->bd.cur = rxq->bd.base; cur 862 drivers/net/ethernet/freescale/fec_main.c txq->bd.cur = bdp; cur 1265 drivers/net/ethernet/freescale/fec_main.c while (bdp != READ_ONCE(txq->bd.cur)) { cur 1342 drivers/net/ethernet/freescale/fec_main.c if (bdp != txq->bd.cur && cur 1439 drivers/net/ethernet/freescale/fec_main.c bdp = rxq->bd.cur; cur 1598 drivers/net/ethernet/freescale/fec_main.c rxq->bd.cur = bdp; cur 3279 drivers/net/ethernet/freescale/fec_main.c rxq->bd.cur = cbd_base; cur 3295 drivers/net/ethernet/freescale/fec_main.c txq->bd.cur = cbd_base; cur 1332 drivers/net/ethernet/ibm/ibmvnic.c u8 *data, *cur; cur 1336 drivers/net/ethernet/ibm/ibmvnic.c cur = hdr_data + len - tmp_len; cur 1339 drivers/net/ethernet/ibm/ibmvnic.c if (cur != hdr_data) { cur 1356 drivers/net/ethernet/ibm/ibmvnic.c memcpy(data, cur, tmp); cur 1476 drivers/net/ethernet/ibm/ibmvnic.c int cur, i; cur 1480 drivers/net/ethernet/ibm/ibmvnic.c cur = skb_headlen(skb); cur 1486 drivers/net/ethernet/ibm/ibmvnic.c memcpy(dst + cur, cur 1489 drivers/net/ethernet/ibm/ibmvnic.c cur += skb_frag_size(frag); cur 2687 drivers/net/ethernet/ibm/ibmvnic.c scrq->cur = 0; cur 3292 drivers/net/ethernet/ibm/ibmvnic.c union sub_crq *entry = &scrq->msgs[scrq->cur]; cur 3307 drivers/net/ethernet/ibm/ibmvnic.c entry = &scrq->msgs[scrq->cur]; cur 3309 drivers/net/ethernet/ibm/ibmvnic.c if (++scrq->cur == scrq->size) cur 3310 drivers/net/ethernet/ibm/ibmvnic.c scrq->cur = 0; cur 3324 drivers/net/ethernet/ibm/ibmvnic.c crq = &queue->msgs[queue->cur]; cur 3326 drivers/net/ethernet/ibm/ibmvnic.c if (++queue->cur == queue->size) cur 3327 drivers/net/ethernet/ibm/ibmvnic.c queue->cur = 0; cur 4699 drivers/net/ethernet/ibm/ibmvnic.c crq->cur = 0; cur 4796 drivers/net/ethernet/ibm/ibmvnic.c crq->cur = 0; cur 846 drivers/net/ethernet/ibm/ibmvnic.h int size, cur; cur 866 drivers/net/ethernet/ibm/ibmvnic.h int size, cur; cur 525 drivers/net/ethernet/intel/ice/ice_common.c hw->fw_log.evnts[m].cur = flgs; cur 613 drivers/net/ethernet/intel/ice/ice_common.c if (hw->fw_log.evnts[i].cfg == hw->fw_log.evnts[i].cur) cur 671 drivers/net/ethernet/intel/ice/ice_common.c hw->fw_log.evnts[i].cur = 0; cur 677 drivers/net/ethernet/intel/ice/ice_common.c hw->fw_log.evnts[m].cur = hw->fw_log.evnts[m].cfg; cur 385 drivers/net/ethernet/intel/ice/ice_type.h u8 cur : 4; /* Current/active event enables */ cur 359 drivers/net/ethernet/jme.c dpi->cur = PCC_P1; cur 1120 drivers/net/ethernet/jme.c if (likely(atmp == dpi->cur)) { cur 1147 drivers/net/ethernet/jme.c if (unlikely(dpi->attempt != dpi->cur && dpi->cnt > 5)) { cur 1148 drivers/net/ethernet/jme.c if (dpi->attempt < dpi->cur) cur 1151 drivers/net/ethernet/jme.c dpi->cur = dpi->attempt; cur 2432 drivers/net/ethernet/jme.c switch (jme->dpi.cur) { cur 2465 drivers/net/ethernet/jme.c dpi->cur = PCC_P1; cur 3048 drivers/net/ethernet/jme.c jme->dpi.cur = PCC_P1; cur 146 drivers/net/ethernet/jme.h unsigned char cur; cur 904 drivers/net/ethernet/mellanox/mlx4/mcg.c struct mlx4_spec_list *cur; cur 913 drivers/net/ethernet/mellanox/mlx4/mcg.c list_for_each_entry(cur, &rule->list, list) { cur 914 drivers/net/ethernet/mellanox/mlx4/mcg.c switch (cur->id) { cur 917 drivers/net/ethernet/mellanox/mlx4/mcg.c "dmac = %pM ", &cur->eth.dst_mac); cur 918 drivers/net/ethernet/mellanox/mlx4/mcg.c if (cur->eth.ether_type) cur 921 drivers/net/ethernet/mellanox/mlx4/mcg.c be16_to_cpu(cur->eth.ether_type)); cur 922 drivers/net/ethernet/mellanox/mlx4/mcg.c if (cur->eth.vlan_id) cur 925 drivers/net/ethernet/mellanox/mlx4/mcg.c be16_to_cpu(cur->eth.vlan_id)); cur 929 drivers/net/ethernet/mellanox/mlx4/mcg.c if (cur->ipv4.src_ip) cur 932 drivers/net/ethernet/mellanox/mlx4/mcg.c &cur->ipv4.src_ip); cur 933 drivers/net/ethernet/mellanox/mlx4/mcg.c if (cur->ipv4.dst_ip) cur 936 drivers/net/ethernet/mellanox/mlx4/mcg.c &cur->ipv4.dst_ip); cur 941 drivers/net/ethernet/mellanox/mlx4/mcg.c if (cur->tcp_udp.src_port) cur 944 drivers/net/ethernet/mellanox/mlx4/mcg.c be16_to_cpu(cur->tcp_udp.src_port)); cur 945 drivers/net/ethernet/mellanox/mlx4/mcg.c if (cur->tcp_udp.dst_port) cur 948 drivers/net/ethernet/mellanox/mlx4/mcg.c be16_to_cpu(cur->tcp_udp.dst_port)); cur 953 drivers/net/ethernet/mellanox/mlx4/mcg.c "dst-gid = %pI6\n", cur->ib.dst_gid); cur 956 drivers/net/ethernet/mellanox/mlx4/mcg.c cur->ib.dst_gid_msk); cur 961 drivers/net/ethernet/mellanox/mlx4/mcg.c "VNID = %d ", be32_to_cpu(cur->vxlan.vni)); cur 981 drivers/net/ethernet/mellanox/mlx4/mcg.c struct mlx4_spec_list *cur; cur 999 drivers/net/ethernet/mellanox/mlx4/mcg.c list_for_each_entry(cur, &rule->list, list) { cur 1000 drivers/net/ethernet/mellanox/mlx4/mcg.c ret = parse_trans_rule(dev, cur, mailbox->buf + size); cur 24 drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c u32 cur; cur 269 drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c mlxsw_reg_sbpm_unpack(sbpm_pl, &pm->occ.cur, &pm->occ.max); cur 1217 drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c &cm->occ.cur, &cm->occ.max); cur 1231 drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c &cm->occ.cur, &cm->occ.max); cur 1379 drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c *p_cur = mlxsw_sp_cells_bytes(mlxsw_sp, pm->occ.cur); cur 1398 drivers/net/ethernet/mellanox/mlxsw/spectrum_buffers.c *p_cur = mlxsw_sp_cells_bytes(mlxsw_sp, cm->occ.cur); cur 1001 drivers/net/ethernet/micrel/ksz884x.c struct ksz_desc *cur; cur 1829 drivers/net/ethernet/micrel/ksz884x.c u32 cur; cur 1849 drivers/net/ethernet/micrel/ksz884x.c cur = *last; cur 1850 drivers/net/ethernet/micrel/ksz884x.c if (data != cur) { cur 1852 drivers/net/ethernet/micrel/ksz884x.c if (data < cur) cur 1854 drivers/net/ethernet/micrel/ksz884x.c data -= cur; cur 3831 drivers/net/ethernet/micrel/ksz884x.c struct ksz_desc *cur = desc_info->ring; cur 3835 drivers/net/ethernet/micrel/ksz884x.c cur->phw = desc++; cur 3837 drivers/net/ethernet/micrel/ksz884x.c previous = cur++; cur 3847 drivers/net/ethernet/micrel/ksz884x.c desc_info->cur = desc_info->ring; cur 3867 drivers/net/ethernet/micrel/ksz884x.c info->cur = info->ring; cur 3977 drivers/net/ethernet/micrel/ksz884x.c get_tx_pkt(&hw->tx_desc_info, &hw->tx_desc_info.cur); cur 3978 drivers/net/ethernet/micrel/ksz884x.c hw->tx_desc_info.cur->sw.buf.tx.first_seg = 1; cur 4002 drivers/net/ethernet/micrel/ksz884x.c struct ksz_desc *cur = hw->tx_desc_info.cur; cur 4004 drivers/net/ethernet/micrel/ksz884x.c cur->sw.buf.tx.last_seg = 1; cur 4008 drivers/net/ethernet/micrel/ksz884x.c cur->sw.buf.tx.intr = 1; cur 4014 drivers/net/ethernet/micrel/ksz884x.c cur->sw.buf.tx.dest_port = hw->dst_ports; cur 4016 drivers/net/ethernet/micrel/ksz884x.c release_desc(cur); cur 4650 drivers/net/ethernet/micrel/ksz884x.c first = info->cur; cur 4696 drivers/net/ethernet/micrel/ksz884x.c info->cur = desc; cur 640 drivers/net/ethernet/natsemi/natsemi.c static int netdev_get_wol(struct net_device *dev, u32 *supported, u32 *cur); cur 2736 drivers/net/ethernet/natsemi/natsemi.c static int netdev_get_wol(struct net_device *dev, u32 *supported, u32 *cur) cur 2749 drivers/net/ethernet/natsemi/natsemi.c *cur = 0; cur 2753 drivers/net/ethernet/natsemi/natsemi.c *cur |= WAKE_PHY; cur 2755 drivers/net/ethernet/natsemi/natsemi.c *cur |= WAKE_UCAST; cur 2757 drivers/net/ethernet/natsemi/natsemi.c *cur |= WAKE_MCAST; cur 2759 drivers/net/ethernet/natsemi/natsemi.c *cur |= WAKE_BCAST; cur 2761 drivers/net/ethernet/natsemi/natsemi.c *cur |= WAKE_ARP; cur 2763 drivers/net/ethernet/natsemi/natsemi.c *cur |= WAKE_MAGIC; cur 2766 drivers/net/ethernet/natsemi/natsemi.c *cur |= WAKE_MAGICSECURE; cur 297 drivers/net/ethernet/pensando/ionic/ionic_dev.c struct ionic_cq_info *cur; cur 315 drivers/net/ethernet/pensando/ionic/ionic_dev.c cur = cq->info; cur 319 drivers/net/ethernet/pensando/ionic/ionic_dev.c cur->next = cq->info; cur 320 drivers/net/ethernet/pensando/ionic/ionic_dev.c cur->last = true; cur 322 drivers/net/ethernet/pensando/ionic/ionic_dev.c cur->next = cur + 1; cur 324 drivers/net/ethernet/pensando/ionic/ionic_dev.c cur->index = i; cur 325 drivers/net/ethernet/pensando/ionic/ionic_dev.c cur++; cur 333 drivers/net/ethernet/pensando/ionic/ionic_dev.c struct ionic_cq_info *cur; cur 339 drivers/net/ethernet/pensando/ionic/ionic_dev.c for (i = 0, cur = cq->info; i < cq->num_descs; i++, cur++) cur 340 drivers/net/ethernet/pensando/ionic/ionic_dev.c cur->cq_desc = base + (i * cq->desc_size); cur 378 drivers/net/ethernet/pensando/ionic/ionic_dev.c struct ionic_desc_info *cur; cur 401 drivers/net/ethernet/pensando/ionic/ionic_dev.c cur = q->info; cur 405 drivers/net/ethernet/pensando/ionic/ionic_dev.c cur->next = q->info; cur 407 drivers/net/ethernet/pensando/ionic/ionic_dev.c cur->next = cur + 1; cur 408 drivers/net/ethernet/pensando/ionic/ionic_dev.c cur->index = i; cur 409 drivers/net/ethernet/pensando/ionic/ionic_dev.c cur->left = num_descs - i; cur 410 drivers/net/ethernet/pensando/ionic/ionic_dev.c cur++; cur 418 drivers/net/ethernet/pensando/ionic/ionic_dev.c struct ionic_desc_info *cur; cur 424 drivers/net/ethernet/pensando/ionic/ionic_dev.c for (i = 0, cur = q->info; i < q->num_descs; i++, cur++) cur 425 drivers/net/ethernet/pensando/ionic/ionic_dev.c cur->desc = base + (i * q->desc_size); cur 430 drivers/net/ethernet/pensando/ionic/ionic_dev.c struct ionic_desc_info *cur; cur 436 drivers/net/ethernet/pensando/ionic/ionic_dev.c for (i = 0, cur = q->info; i < q->num_descs; i++, cur++) cur 437 drivers/net/ethernet/pensando/ionic/ionic_dev.c cur->sg_desc = base + (i * q->sg_desc_size); cur 1791 drivers/net/ethernet/pensando/ionic/ionic_lif.c struct list_head *cur, *tmp; cur 1794 drivers/net/ethernet/pensando/ionic/ionic_lif.c list_for_each_safe(cur, tmp, &ionic->lifs) { cur 1795 drivers/net/ethernet/pensando/ionic/ionic_lif.c lif = list_entry(cur, struct ionic_lif, list); cur 1820 drivers/net/ethernet/pensando/ionic/ionic_lif.c struct list_head *cur, *tmp; cur 1823 drivers/net/ethernet/pensando/ionic/ionic_lif.c list_for_each_safe(cur, tmp, &ionic->lifs) { cur 1824 drivers/net/ethernet/pensando/ionic/ionic_lif.c lif = list_entry(cur, struct ionic_lif, list); cur 2060 drivers/net/ethernet/pensando/ionic/ionic_lif.c struct list_head *cur, *tmp; cur 2064 drivers/net/ethernet/pensando/ionic/ionic_lif.c list_for_each_safe(cur, tmp, &ionic->lifs) { cur 2065 drivers/net/ethernet/pensando/ionic/ionic_lif.c lif = list_entry(cur, struct ionic_lif, list); cur 287 drivers/net/ethernet/pensando/ionic/ionic_txrx.c struct ionic_desc_info *cur; cur 290 drivers/net/ethernet/pensando/ionic/ionic_txrx.c for (cur = q->tail; cur != q->head; cur = cur->next) { cur 291 drivers/net/ethernet/pensando/ionic/ionic_txrx.c desc = cur->desc; cur 294 drivers/net/ethernet/pensando/ionic/ionic_txrx.c dev_kfree_skb(cur->cb_arg); cur 295 drivers/net/ethernet/pensando/ionic/ionic_txrx.c cur->cb_arg = NULL; cur 625 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c nx_mac_list_t *cur; cur 629 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c cur = list_entry(head, nx_mac_list_t, list); cur 631 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c if (ether_addr_equal(addr, cur->mac_addr)) { cur 637 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c cur = kzalloc(sizeof(nx_mac_list_t), GFP_ATOMIC); cur 638 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c if (cur == NULL) cur 641 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c memcpy(cur->mac_addr, addr, ETH_ALEN); cur 642 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c list_add_tail(&cur->list, &adapter->mac_list); cur 644 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c cur->mac_addr, NETXEN_MAC_ADD); cur 657 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c nx_mac_list_t *cur; cur 687 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c cur = list_entry(head->next, nx_mac_list_t, list); cur 690 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c cur->mac_addr, NETXEN_MAC_DEL); cur 691 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c list_del(&cur->list); cur 692 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c kfree(cur); cur 717 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c nx_mac_list_t *cur; cur 721 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c cur = list_entry(head->next, nx_mac_list_t, list); cur 723 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c cur->mac_addr, NETXEN_MAC_DEL); cur 724 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c list_del(&cur->list); cur 725 drivers/net/ethernet/qlogic/netxen/netxen_nic_hw.c kfree(cur); cur 1651 drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c struct list_head *cur; cur 1713 drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c list_for_each(cur, &sds_ring->free_list[ring]) { cur 1714 drivers/net/ethernet/qlogic/netxen/netxen_nic_init.c rxbuf = list_entry(cur, cur 3177 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c struct nx_ip_list *cur, *tmp_cur; cur 3179 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c list_for_each_entry_safe(cur, tmp_cur, &adapter->ip_list, list) { cur 3181 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c if (cur->master) { cur 3182 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c netxen_config_ipaddr(adapter, cur->ip_addr, cur 3184 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c list_del(&cur->list); cur 3185 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c kfree(cur); cur 3188 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c netxen_config_ipaddr(adapter, cur->ip_addr, NX_IP_DOWN); cur 3189 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c list_del(&cur->list); cur 3190 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c kfree(cur); cur 3200 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c struct nx_ip_list *cur, *tmp_cur; cur 3212 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c cur = list_entry(head, struct nx_ip_list, list); cur 3214 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c if (cur->ip_addr == ifa->ifa_address) cur 3218 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c cur = kzalloc(sizeof(struct nx_ip_list), GFP_ATOMIC); cur 3219 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c if (cur == NULL) cur 3223 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c cur->master = !!netif_is_bond_master(dev); cur 3224 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c cur->ip_addr = ifa->ifa_address; cur 3225 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c list_add_tail(&cur->list, &adapter->ip_list); cur 3230 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c list_for_each_entry_safe(cur, tmp_cur, cur 3232 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c if (cur->ip_addr == ifa->ifa_address) { cur 3233 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c list_del(&cur->list); cur 3234 drivers/net/ethernet/qlogic/netxen/netxen_nic_main.c kfree(cur); cur 146 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c u32 cur, prev; cur 147 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c cur = adapter->ahw->idc.curr_state; cur 152 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c adapter->ahw->idc.name[cur], cur 984 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c u32 cur, prev, next; cur 986 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c cur = adapter->ahw->idc.curr_state; cur 994 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c __func__, cur, prev, state); cur 998 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c if ((cur == QLC_83XX_IDC_DEV_UNKNOWN) && cur 1004 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c __func__, cur, prev, next); cur 1015 drivers/net/ethernet/qlogic/qlcnic/qlcnic_83xx_init.c __func__, cur, prev, next); cur 463 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c struct qlcnic_mac_vlan_list *cur; cur 469 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c cur = list_entry(head, struct qlcnic_mac_vlan_list, list); cur 470 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c if (ether_addr_equal(addr, cur->mac_addr)) { cur 471 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c err = qlcnic_sre_macaddr_change(adapter, cur->mac_addr, cur 475 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c list_del(&cur->list); cur 476 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c kfree(cur); cur 486 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c struct qlcnic_mac_vlan_list *cur; cur 491 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c cur = list_entry(head, struct qlcnic_mac_vlan_list, list); cur 492 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c if (ether_addr_equal(addr, cur->mac_addr) && cur 493 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c cur->vlan_id == vlan) cur 497 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c cur = kzalloc(sizeof(*cur), GFP_ATOMIC); cur 498 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c if (cur == NULL) cur 501 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c memcpy(cur->mac_addr, addr, ETH_ALEN); cur 504 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c cur->mac_addr, vlan, QLCNIC_MAC_ADD)) { cur 505 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c kfree(cur); cur 509 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c cur->vlan_id = vlan; cur 510 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c cur->mac_type = mac_type; cur 512 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c list_add_tail(&cur->list, &adapter->mac_list); cur 518 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c struct qlcnic_mac_vlan_list *cur; cur 522 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c cur = list_entry(head, struct qlcnic_mac_vlan_list, list); cur 523 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c if (cur->mac_type != QLCNIC_MULTICAST_MAC) cur 526 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c qlcnic_sre_macaddr_change(adapter, cur->mac_addr, cur 527 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c cur->vlan_id, QLCNIC_MAC_DEL); cur 528 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c list_del(&cur->list); cur 529 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c kfree(cur); cur 623 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c struct qlcnic_mac_vlan_list *cur; cur 626 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c cur = list_entry(head->next, struct qlcnic_mac_vlan_list, list); cur 628 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c cur->mac_addr, 0, QLCNIC_MAC_DEL); cur 629 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c list_del(&cur->list); cur 630 drivers/net/ethernet/qlogic/qlcnic/qlcnic_hw.c kfree(cur); cur 1361 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c struct list_head *cur; cur 1415 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c list_for_each(cur, &sds_ring->free_list[ring]) { cur 1416 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c rxbuf = list_entry(cur, struct qlcnic_rx_buffer, cur 1885 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c struct list_head *cur; cur 1931 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c list_for_each(cur, &sds_ring->free_list[ring]) { cur 1932 drivers/net/ethernet/qlogic/qlcnic/qlcnic_io.c rxbuf = list_entry(cur, struct qlcnic_rx_buffer, cur 322 drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c struct qlcnic_mac_vlan_list *cur; cur 326 drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c cur = list_entry(head, struct qlcnic_mac_vlan_list, list); cur 327 drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c if (ether_addr_equal_unaligned(adapter->mac_addr, cur->mac_addr)) { cur 328 drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c qlcnic_sre_macaddr_change(adapter, cur->mac_addr, cur 330 drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c list_del(&cur->list); cur 331 drivers/net/ethernet/qlogic/qlcnic/qlcnic_main.c kfree(cur); cur 2100 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c struct qlcnic_mac_vlan_list *cur; cur 2103 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c cur = list_entry(head->next, struct qlcnic_mac_vlan_list, list); cur 2104 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c qlcnic_sre_macaddr_change(adapter, cur->mac_addr, cur->vlan_id, cur 2106 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c list_del(&cur->list); cur 2107 drivers/net/ethernet/qlogic/qlcnic/qlcnic_sriov_common.c kfree(cur); cur 515 drivers/net/ethernet/sis/sis190.c u32 cur; cur 517 drivers/net/ethernet/sis/sis190.c for (cur = start; cur < end; cur++) { cur 518 drivers/net/ethernet/sis/sis190.c unsigned int i = cur % NUM_RX_DESC; cur 528 drivers/net/ethernet/sis/sis190.c return cur - start; cur 1245 drivers/net/ethernet/sis/sis190.c struct sis190_phy *cur, *next; cur 1247 drivers/net/ethernet/sis/sis190.c list_for_each_entry_safe(cur, next, first_phy, list) { cur 1248 drivers/net/ethernet/sis/sis190.c kfree(cur); cur 120 drivers/net/ethernet/smsc/smc911x.c int cur; cur 122 drivers/net/ethernet/smsc/smc911x.c for (cur = 0; cur < 8; cur++) { cur 169 drivers/net/ethernet/smsc/smc91x.c int cur; cur 171 drivers/net/ethernet/smsc/smc91x.c for (cur = 0; cur < 8; cur++) { cur 8567 drivers/net/ethernet/sun/niu.c if (p->cur[type] >= NIU_MAX_PORTS) { cur 8571 drivers/net/ethernet/sun/niu.c idx = p->cur[type]; cur 8574 drivers/net/ethernet/sun/niu.c p->cur[type] = idx + 1; cur 8582 drivers/net/ethernet/sun/niu.c for (i = 0; i < p->cur[PHY_TYPE_PMA_PMD]; i++) { cur 8586 drivers/net/ethernet/sun/niu.c for (i = 0; i < p->cur[PHY_TYPE_PCS]; i++) { cur 8614 drivers/net/ethernet/sun/niu.c if (p->cur[PHY_TYPE_MII]) cur 8617 drivers/net/ethernet/sun/niu.c return p->cur[PHY_TYPE_MII]; cur 3041 drivers/net/ethernet/sun/niu.h u8 cur[PHY_TYPE_MAX]; cur 450 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c ring->cur = 0; cur 490 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c ring->cur = 0; cur 512 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c start_index = ring->cur; cur 513 drivers/net/ethernet/synopsys/dwc-xlgmac-desc.c cur_index = ring->cur; cur 688 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c desc_data = XLGMAC_GET_DESC_DATA(ring, ring->cur); cur 711 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c int start_index = ring->cur; cur 712 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c int cur_index = ring->cur; cur 997 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c ring->cur = cur_index + 1; cur 1007 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c (ring->cur - 1) & (ring->dma_desc_count - 1)); cur 1060 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c int start_index = ring->cur; cur 1143 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c unsigned int start_index = ring->cur; cur 2648 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c desc_data = XLGMAC_GET_DESC_DATA(ring, ring->cur); cur 2662 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c xlgmac_dump_rx_desc(pdata, ring, ring->cur); cur 2811 drivers/net/ethernet/synopsys/dwc-xlgmac-hw.c ring->cur & (ring->dma_desc_count - 1), ring->cur); cur 30 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c return (ring->dma_desc_count - (ring->cur - ring->dirty)); cur 35 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c return (ring->cur - ring->dirty); cur 962 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c while (ring->dirty != ring->cur) { cur 1046 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c unsigned int cur; cur 1055 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c cur = ring->cur; cur 1063 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c (ring->dirty != cur)) { cur 1134 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c desc_data = XLGMAC_GET_DESC_DATA(ring, ring->cur); cur 1150 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c desc_data = XLGMAC_GET_DESC_DATA(ring, ring->cur); cur 1159 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c ring->cur++; cur 1268 drivers/net/ethernet/synopsys/dwc-xlgmac-net.c desc_data = XLGMAC_GET_DESC_DATA(ring, ring->cur); cur 339 drivers/net/ethernet/synopsys/dwc-xlgmac.h unsigned int cur; cur 95 drivers/net/ethernet/xilinx/ll_temac_main.c ktime_t cur = ktime_get(); cur 97 drivers/net/ethernet/xilinx/ll_temac_main.c return hard_acs_rdy(lp) || ktime_after(cur, timeout); cur 87 drivers/net/phy/mdio-mux.c int (*switch_fn)(int cur, int desired, void *data), cur 822 drivers/net/team/team.c struct team_port *cur) cur 824 drivers/net/team/team.c if (port->priority < cur->priority) cur 826 drivers/net/team/team.c if (port->priority > cur->priority) cur 828 drivers/net/team/team.c if (port->index < cur->index) cur 836 drivers/net/team/team.c struct team_port *cur; cur 844 drivers/net/team/team.c list_for_each_entry(cur, qom_list, qom_list) { cur 845 drivers/net/team/team.c if (team_queue_override_port_has_gt_prio_than(port, cur)) cur 847 drivers/net/team/team.c node = &cur->qom_list; cur 916 drivers/net/team/team.c struct team_port *cur; cur 918 drivers/net/team/team.c list_for_each_entry(cur, &team->port_list, list) cur 919 drivers/net/team/team.c if (cur == port) cur 17 drivers/net/team/team_mode_broadcast.c struct team_port *cur; cur 23 drivers/net/team/team_mode_broadcast.c list_for_each_entry_rcu(cur, &team->port_list, list) { cur 24 drivers/net/team/team_mode_broadcast.c if (team_port_txable(cur)) { cur 34 drivers/net/team/team_mode_broadcast.c last = cur; cur 1113 drivers/net/wireless/ath/ath6kl/txrx.c u16 idx, st, cur, end; cur 1136 drivers/net/wireless/ath/ath6kl/txrx.c cur = seq_no; cur 1139 drivers/net/wireless/ath/ath6kl/txrx.c if (((st < end) && (cur < st || cur > end)) || cur 1140 drivers/net/wireless/ath/ath6kl/txrx.c ((st > end) && (cur > end) && (cur < st))) { cur 1145 drivers/net/wireless/ath/ath6kl/txrx.c (cur < end || cur > extended_end)) || cur 1146 drivers/net/wireless/ath/ath6kl/txrx.c ((end > extended_end) && (cur > extended_end) && cur 1147 drivers/net/wireless/ath/ath6kl/txrx.c (cur < end))) { cur 1150 drivers/net/wireless/ath/ath6kl/txrx.c if (cur >= rxtid->hold_q_sz - 1) cur 1151 drivers/net/wireless/ath/ath6kl/txrx.c rxtid->seq_next = cur - (rxtid->hold_q_sz - 1); cur 1154 drivers/net/wireless/ath/ath6kl/txrx.c (rxtid->hold_q_sz - 2 - cur); cur 1161 drivers/net/wireless/ath/ath6kl/txrx.c if (cur >= rxtid->hold_q_sz - 1) cur 1162 drivers/net/wireless/ath/ath6kl/txrx.c st = cur - (rxtid->hold_q_sz - 1); cur 1165 drivers/net/wireless/ath/ath6kl/txrx.c (rxtid->hold_q_sz - 2 - cur); cur 336 drivers/net/wireless/ath/ath9k/channel.c struct ath_chanctx *prev, *cur; cur 343 drivers/net/wireless/ath/ath9k/channel.c cur = sc->cur_chan; cur 344 drivers/net/wireless/ath/ath9k/channel.c prev = ath_chanctx_get_next(sc, cur); cur 350 drivers/net/wireless/ath/ath9k/channel.c cur_tsf = (u32) cur->tsf_val + cur 351 drivers/net/wireless/ath/ath9k/channel.c ath9k_hw_get_tsf_offset(&cur->tsf_ts, &ts); cur 1852 drivers/net/wireless/ath/ath9k/hw.c u32 ath9k_hw_get_tsf_offset(struct timespec64 *last, struct timespec64 *cur) cur 1857 drivers/net/wireless/ath/ath9k/hw.c if (!cur) { cur 1859 drivers/net/wireless/ath/ath9k/hw.c cur = &ts; cur 1862 drivers/net/wireless/ath/ath9k/hw.c usec = cur->tv_sec * 1000000ULL + cur->tv_nsec / 1000; cur 1064 drivers/net/wireless/ath/ath9k/hw.h u32 ath9k_hw_get_tsf_offset(struct timespec64 *last, struct timespec64 *cur); cur 241 drivers/net/wireless/ath/wcn36xx/dxe.c struct wcn36xx_dxe_ctl *cur = ch->head_blk_ctl; cur 247 drivers/net/wireless/ath/wcn36xx/dxe.c cur->bd_phy_addr = bd_phy_addr; cur 248 drivers/net/wireless/ath/wcn36xx/dxe.c cur->bd_cpu_addr = bd_cpu_addr; cur 252 drivers/net/wireless/ath/wcn36xx/dxe.c cur->bd_phy_addr = 0; cur 253 drivers/net/wireless/ath/wcn36xx/dxe.c cur->bd_cpu_addr = NULL; cur 255 drivers/net/wireless/ath/wcn36xx/dxe.c cur = cur->next; cur 319 drivers/net/wireless/ath/wcn36xx/dxe.c struct wcn36xx_dxe_ctl *cur = wcn_ch->head_blk_ctl; cur 323 drivers/net/wireless/ath/wcn36xx/dxe.c kfree_skb(cur->skb); cur 324 drivers/net/wireless/ath/wcn36xx/dxe.c cur = cur->next; cur 773 drivers/net/wireless/broadcom/b43/debugfs.c struct b43_txstatus *cur; cur 783 drivers/net/wireless/broadcom/b43/debugfs.c cur = &(log->log[i]); cur 784 drivers/net/wireless/broadcom/b43/debugfs.c memcpy(cur, status, sizeof(*cur)); cur 445 drivers/net/wireless/broadcom/b43legacy/debugfs.c struct b43legacy_txstatus *cur; cur 457 drivers/net/wireless/broadcom/b43legacy/debugfs.c cur = &(log->log[i]); cur 458 drivers/net/wireless/broadcom/b43legacy/debugfs.c memcpy(cur, status, sizeof(*cur)); cur 524 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c struct brcmf_fw_item *cur; cur 534 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c cur = &fwctx->req->items[fwctx->curpos]; cur 544 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c else if (!(cur->flags & BRCMF_FW_REQF_OPTIONAL)) cur 559 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c if (!nvram && !(cur->flags & BRCMF_FW_REQF_OPTIONAL)) cur 563 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c cur->nv_data.data = nvram; cur 564 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c cur->nv_data.len = nvram_length; cur 574 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c struct brcmf_fw_item *cur = &fwctx->req->items[fwctx->curpos]; cur 577 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c brcmf_dbg(TRACE, "firmware %s %sfound\n", cur->path, fw ? "" : "not "); cur 579 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c switch (cur->type) { cur 585 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c cur->binary = fw; cur 591 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c brcmf_err("unknown fw type: %d\n", cur->type); cur 596 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c return (cur->flags & BRCMF_FW_REQF_OPTIONAL) ? 0 : ret; cur 602 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c struct brcmf_fw_item *cur = &fwctx->req->items[fwctx->curpos]; cur 606 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c if (cur->type == BRCMF_FW_TYPE_NVRAM && fwctx->req->board_type) { cur 609 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c strlcpy(alt_path, cur->path, BRCMF_FW_NAME_LEN); cur 621 drivers/net/wireless/broadcom/brcm80211/brcmfmac/firmware.c return request_firmware(fw, cur->path, fwctx->dev); cur 1257 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c struct sk_buff *cur, *next; cur 1259 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c skb_queue_walk_safe(&bus->glom, cur, next) { cur 1260 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c skb_unlink(cur, &bus->glom); cur 1261 drivers/net/wireless/broadcom/brcm80211/brcmfmac/sdio.c brcmu_pkt_buf_free_skb(cur); cur 963 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c uint cur; cur 964 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c cur = cur 971 drivers/net/wireless/broadcom/brcm80211/brcmsmac/dma.c di->rxin, di->rxout, cur); cur 1303 drivers/net/wireless/cisco/airo.c static void age_mic_context(miccntx *cur, miccntx *old, u8 *key, int key_len, cur 1309 drivers/net/wireless/cisco/airo.c if (cur->valid && (memcmp(cur->key, key, key_len) == 0)) cur 1313 drivers/net/wireless/cisco/airo.c memcpy(old, cur, sizeof(*cur)); cur 1316 drivers/net/wireless/cisco/airo.c memcpy(cur->key, key, key_len); cur 1317 drivers/net/wireless/cisco/airo.c cur->window = 33; /* Window always points to the middle */ cur 1318 drivers/net/wireless/cisco/airo.c cur->rx = 0; /* Rx Sequence numbers */ cur 1319 drivers/net/wireless/cisco/airo.c cur->tx = 0; /* Tx sequence numbers */ cur 1320 drivers/net/wireless/cisco/airo.c cur->valid = 1; /* Key is now valid */ cur 1323 drivers/net/wireless/cisco/airo.c emmh32_setseed(&cur->seed, key, key_len, tfm); cur 287 drivers/net/wireless/intel/iwlwifi/dvm/rx.c static void accum_stats(__le32 *prev, __le32 *cur, __le32 *delta, cur 294 drivers/net/wireless/intel/iwlwifi/dvm/rx.c i++, prev++, cur++, delta++, max_delta++, accum++) { cur 295 drivers/net/wireless/intel/iwlwifi/dvm/rx.c if (le32_to_cpu(*cur) > le32_to_cpu(*prev)) { cur 297 drivers/net/wireless/intel/iwlwifi/dvm/rx.c le32_to_cpu(*cur) - le32_to_cpu(*prev)); cur 2010 drivers/net/wireless/intel/iwlwifi/mvm/rs.c u8 cur = lq_sta->lq.reduced_tpc; cur 2020 drivers/net/wireless/intel/iwlwifi/mvm/rs.c return cur != lq_sta->pers.dbg_fixed_txp_reduction; cur 2036 drivers/net/wireless/intel/iwlwifi/mvm/rs.c return cur != TPC_NO_REDUCTION; cur 2039 drivers/net/wireless/intel/iwlwifi/mvm/rs.c rs_get_adjacent_txp(mvm, cur, &weak, &strong); cur 2043 drivers/net/wireless/intel/iwlwifi/mvm/rs.c sr = window[cur].success_ratio; cur 2044 drivers/net/wireless/intel/iwlwifi/mvm/rs.c current_tpt = window[cur].average_tpt; cur 2052 drivers/net/wireless/intel/iwlwifi/mvm/rs.c cur, current_tpt, sr, weak, strong, cur 912 drivers/net/wireless/marvell/libertas/if_sdio.c struct if_sdio_packet *packet, *cur; cur 955 drivers/net/wireless/marvell/libertas/if_sdio.c cur = card->packets; cur 956 drivers/net/wireless/marvell/libertas/if_sdio.c while (cur->next) cur 957 drivers/net/wireless/marvell/libertas/if_sdio.c cur = cur->next; cur 958 drivers/net/wireless/marvell/libertas/if_sdio.c cur->next = packet; cur 530 drivers/net/wireless/mediatek/mt76/dma.c int qid, done = 0, cur; cur 538 drivers/net/wireless/mediatek/mt76/dma.c cur = mt76_dma_rx_process(dev, &dev->q_rx[qid], budget - done); cur 540 drivers/net/wireless/mediatek/mt76/dma.c done += cur; cur 541 drivers/net/wireless/mediatek/mt76/dma.c } while (cur && done < budget); cur 21 drivers/net/wireless/mediatek/mt76/mt76x2/pci_mcu.c __le32 *cur; cur 58 drivers/net/wireless/mediatek/mt76/mt76x2/pci_mcu.c cur = (__le32 *)(fw->data + sizeof(*hdr)); cur 60 drivers/net/wireless/mediatek/mt76/mt76x2/pci_mcu.c mt76_wr_copy(dev, MT_MCU_ROM_PATCH_ADDR, cur, len); cur 86 drivers/net/wireless/mediatek/mt76/mt76x2/pci_mcu.c __le32 *cur; cur 113 drivers/net/wireless/mediatek/mt76/mt76x2/pci_mcu.c cur = (__le32 *)(fw->data + sizeof(*hdr)); cur 117 drivers/net/wireless/mediatek/mt76/mt76x2/pci_mcu.c mt76_wr_copy(dev, MT_MCU_ILM_ADDR, cur, len); cur 119 drivers/net/wireless/mediatek/mt76/mt76x2/pci_mcu.c cur += len / sizeof(*cur); cur 128 drivers/net/wireless/mediatek/mt76/mt76x2/pci_mcu.c mt76_wr_copy(dev, offset, cur, len); cur 12 drivers/net/wireless/mediatek/mt76/util.c u32 cur; cur 16 drivers/net/wireless/mediatek/mt76/util.c cur = dev->bus->rr(dev, offset) & mask; cur 17 drivers/net/wireless/mediatek/mt76/util.c if (cur == val) cur 30 drivers/net/wireless/mediatek/mt76/util.c u32 cur; cur 34 drivers/net/wireless/mediatek/mt76/util.c cur = dev->bus->rr(dev, offset) & mask; cur 35 drivers/net/wireless/mediatek/mt76/util.c if (cur == val) cur 47 drivers/net/wireless/mediatek/mt76/util.c int i, idx = 0, cur; cur 55 drivers/net/wireless/mediatek/mt76/util.c cur = i * BITS_PER_LONG + idx; cur 56 drivers/net/wireless/mediatek/mt76/util.c if (cur >= size) cur 60 drivers/net/wireless/mediatek/mt76/util.c return cur; cur 31 drivers/net/wireless/mediatek/mt7601u/core.c u32 cur; cur 38 drivers/net/wireless/mediatek/mt7601u/core.c cur = mt7601u_rr(dev, offset) & mask; cur 39 drivers/net/wireless/mediatek/mt7601u/core.c if (cur == val) cur 53 drivers/net/wireless/mediatek/mt7601u/core.c u32 cur; cur 60 drivers/net/wireless/mediatek/mt7601u/core.c cur = mt7601u_rr(dev, offset) & mask; cur 61 drivers/net/wireless/mediatek/mt7601u/core.c if (cur == val) cur 200 drivers/net/wireless/realtek/rtw88/tx.c struct sk_buff *cur, *tmp; cur 211 drivers/net/wireless/realtek/rtw88/tx.c skb_queue_walk_safe(&tx_report->queue, cur, tmp) { cur 212 drivers/net/wireless/realtek/rtw88/tx.c n = (u8 *)IEEE80211_SKB_CB(cur)->status.status_driver_data; cur 214 drivers/net/wireless/realtek/rtw88/tx.c __skb_unlink(cur, &tx_report->queue); cur 215 drivers/net/wireless/realtek/rtw88/tx.c rtw_tx_report_tx_status(rtwdev, cur, st == 0); cur 1087 drivers/nvme/host/core.c struct nvme_ns_id_desc *cur = data + pos; cur 1089 drivers/nvme/host/core.c if (cur->nidl == 0) cur 1092 drivers/nvme/host/core.c switch (cur->nidt) { cur 1094 drivers/nvme/host/core.c if (cur->nidl != NVME_NIDT_EUI64_LEN) { cur 1097 drivers/nvme/host/core.c cur->nidl); cur 1101 drivers/nvme/host/core.c memcpy(ids->eui64, data + pos + sizeof(*cur), len); cur 1104 drivers/nvme/host/core.c if (cur->nidl != NVME_NIDT_NGUID_LEN) { cur 1107 drivers/nvme/host/core.c cur->nidl); cur 1111 drivers/nvme/host/core.c memcpy(ids->nguid, data + pos + sizeof(*cur), len); cur 1114 drivers/nvme/host/core.c if (cur->nidl != NVME_NIDT_UUID_LEN) { cur 1117 drivers/nvme/host/core.c cur->nidl); cur 1121 drivers/nvme/host/core.c uuid_copy(&ids->uuid, data + pos + sizeof(*cur)); cur 1125 drivers/nvme/host/core.c len = cur->nidl; cur 1129 drivers/nvme/host/core.c len += sizeof(*cur); cur 1305 drivers/of/base.c it->cur = list; cur 1320 drivers/of/base.c if (!it->cur || it->phandle_end >= it->list_end) cur 1323 drivers/of/base.c it->cur = it->phandle_end; cur 1326 drivers/of/base.c it->phandle = be32_to_cpup(it->cur++); cur 1368 drivers/of/base.c if (it->cur + count > it->list_end) { cur 1376 drivers/of/base.c it->phandle_end = it->cur + count; cur 1403 drivers/of/base.c args[i] = be32_to_cpup(it->cur++); cur 1585 drivers/of/base.c struct device_node *cur, *new = NULL; cur 1620 drivers/of/base.c cur = out_args->np; cur 1621 drivers/of/base.c ret = of_property_read_u32(cur, cells_name, &list_size); cur 1630 drivers/of/base.c while (cur) { cur 1632 drivers/of/base.c map = of_get_property(cur, map_name, &map_len); cur 1640 drivers/of/base.c mask = of_get_property(cur, mask_name, NULL); cur 1681 drivers/of/base.c pass = of_get_property(cur, pass_name, NULL); cur 1704 drivers/of/base.c of_node_put(cur); cur 1705 drivers/of/base.c cur = new; cur 1708 drivers/of/base.c of_node_put(cur); cur 115 drivers/of/fdt.c int cur; cur 119 drivers/of/fdt.c for (cur = fdt_first_property_offset(blob, offset); cur 120 drivers/of/fdt.c cur >= 0; cur 121 drivers/of/fdt.c cur = fdt_next_property_offset(blob, cur)) { cur 126 drivers/of/fdt.c val = fdt_getprop_by_offset(blob, cur, &pname, &sz); cur 128 drivers/of/fdt.c pr_warn("Cannot locate property at 0x%x\n", cur); cur 133 drivers/of/fdt.c pr_warn("Cannot find property name at 0x%x\n", cur); cur 483 drivers/of/property.c const __be32 *of_prop_next_u32(struct property *prop, const __be32 *cur, cur 486 drivers/of/property.c const void *curv = cur; cur 491 drivers/of/property.c if (!cur) { cur 496 drivers/of/property.c curv += sizeof(*cur); cur 506 drivers/of/property.c const char *of_prop_next_string(struct property *prop, const char *cur) cur 508 drivers/of/property.c const void *curv = cur; cur 513 drivers/of/property.c if (!cur) cur 516 drivers/of/property.c curv += strlen(cur) + 1; cur 77 drivers/of/resolver.c char *value, *cur, *end, *node_path, *prop_name, *s; cur 87 drivers/of/resolver.c for (cur = value; cur < end; cur += len + 1) { cur 88 drivers/of/resolver.c len = strlen(cur); cur 90 drivers/of/resolver.c node_path = cur; cur 91 drivers/of/resolver.c s = strchr(cur, ':'); cur 438 drivers/parisc/eisa.c char *cur = str; cur 442 drivers/parisc/eisa.c while (cur != NULL) { cur 445 drivers/parisc/eisa.c val = (int) simple_strtoul(cur, &pe, 0); cur 456 drivers/parisc/eisa.c if ((cur = strchr(cur, ','))) { cur 457 drivers/parisc/eisa.c cur++; cur 179 drivers/parisc/led.c char *cur, lbuf[32]; cur 192 drivers/parisc/led.c cur = lbuf; cur 197 drivers/parisc/led.c d = *cur++ - '0'; cur 201 drivers/parisc/led.c if (*cur++ != ' ') goto parse_error; cur 203 drivers/parisc/led.c d = *cur++ - '0'; cur 207 drivers/parisc/led.c if (*cur++ != ' ') goto parse_error; cur 209 drivers/parisc/led.c d = *cur++ - '0'; cur 215 drivers/parisc/led.c if (*cur && cur[strlen(cur)-1] == '\n') cur 216 drivers/parisc/led.c cur[strlen(cur)-1] = 0; cur 217 drivers/parisc/led.c if (*cur == 0) cur 218 drivers/parisc/led.c cur = lcd_text_default; cur 219 drivers/parisc/led.c lcd_print(cur); cur 335 drivers/pci/controller/pci-mvebu.c struct mvebu_pcie_window *cur) cur 337 drivers/pci/controller/pci-mvebu.c if (desired->base == cur->base && desired->remap == cur->remap && cur 338 drivers/pci/controller/pci-mvebu.c desired->size == cur->size) cur 341 drivers/pci/controller/pci-mvebu.c if (cur->size != 0) { cur 342 drivers/pci/controller/pci-mvebu.c mvebu_pcie_del_windows(port, cur->base, cur->size); cur 343 drivers/pci/controller/pci-mvebu.c cur->size = 0; cur 344 drivers/pci/controller/pci-mvebu.c cur->base = 0; cur 358 drivers/pci/controller/pci-mvebu.c *cur = *desired; cur 278 drivers/pinctrl/bcm/pinctrl-bcm2835.c enum bcm2835_fsel cur = (val >> FSEL_SHIFT(pin)) & BCM2835_FSEL_MASK; cur 281 drivers/pinctrl/bcm/pinctrl-bcm2835.c bcm2835_functions[cur]); cur 283 drivers/pinctrl/bcm/pinctrl-bcm2835.c if (cur == fsel) cur 286 drivers/pinctrl/bcm/pinctrl-bcm2835.c if (cur != BCM2835_FSEL_GPIO_IN && fsel != BCM2835_FSEL_GPIO_IN) { cur 124 drivers/pinctrl/pinctrl-single.c int cur; cur 664 drivers/pinctrl/pinctrl-single.c i = pcs->pins.cur; cur 685 drivers/pinctrl/pinctrl-single.c pcs->pins.cur++; cur 323 drivers/platform/x86/eeepc-laptop.c int cur; cur 328 drivers/platform/x86/eeepc-laptop.c c->cur = get_acpi(eeepc, CM_ASL_CPUFV); cur 329 drivers/platform/x86/eeepc-laptop.c if (c->cur < 0) cur 332 drivers/platform/x86/eeepc-laptop.c c->num = (c->cur >> 8) & 0xff; cur 333 drivers/platform/x86/eeepc-laptop.c c->cur &= 0xff; cur 365 drivers/platform/x86/eeepc-laptop.c return sprintf(buf, "%#x\n", (c.num << 8) | c.cur); cur 197 drivers/platform/x86/msi-wmi.c ktime_t cur = ktime_get_real(); cur 198 drivers/platform/x86/msi-wmi.c ktime_t diff = ktime_sub(cur, last_pressed); cur 207 drivers/platform/x86/msi-wmi.c last_pressed = cur; cur 72 drivers/pnp/core.c struct pnp_protocol *cur = to_pnp_protocol(pos); cur 73 drivers/pnp/core.c if (cur->number == nodenum) { cur 122 drivers/power/supply/apm_power.c union power_supply_propval cur; cur 182 drivers/power/supply/apm_power.c if (_MPSY_PROP(cur_avg_prop, &cur)) { cur 184 drivers/power/supply/apm_power.c if (_MPSY_PROP(cur_now_prop, &cur)) cur 189 drivers/power/supply/apm_power.c return ((cur.intval - full.intval) * 60L) / I.intval; cur 191 drivers/power/supply/apm_power.c return -((cur.intval - empty.intval) * 60L) / I.intval; cur 218 drivers/power/supply/apm_power.c union power_supply_propval empty, full, cur; cur 257 drivers/power/supply/apm_power.c if (_MPSY_PROP(avg_prop, &cur)) { cur 259 drivers/power/supply/apm_power.c if (_MPSY_PROP(now_prop, &cur)) cur 270 drivers/power/supply/apm_power.c ret = ((cur.intval - empty.intval) * 100L) / cur 389 drivers/power/supply/axp288_fuel_gauge.c static int fuel_gauge_get_current(struct axp288_fg_info *info, int *cur) cur 399 drivers/power/supply/axp288_fuel_gauge.c *cur = -1 * discharge; cur 403 drivers/power/supply/axp288_fuel_gauge.c return iio_read_channel_raw(info->iio_channel[BAT_CHRG_CURR], cur); cur 182 drivers/power/supply/s3c_adc_battery.c if (full_volt < calc_full_volt(lut->volt, lut->cur, cur 189 drivers/power/supply/s3c_adc_battery.c lut_volt1 = calc_full_volt(lut[0].volt, lut[0].cur, cur 191 drivers/power/supply/s3c_adc_battery.c lut_volt2 = calc_full_volt(lut[1].volt, lut[1].cur, cur 110 drivers/power/supply/sc2731_charger.c static int sc2731_charger_set_current(struct sc2731_charger_info *info, u32 cur) cur 115 drivers/power/supply/sc2731_charger.c if (cur > SC2731_CURRENT_LIMIT_2000) cur 116 drivers/power/supply/sc2731_charger.c cur = SC2731_CURRENT_LIMIT_2000; cur 117 drivers/power/supply/sc2731_charger.c else if (cur < SC2731_CURRENT_PRECHG) cur 118 drivers/power/supply/sc2731_charger.c cur = SC2731_CURRENT_PRECHG; cur 121 drivers/power/supply/sc2731_charger.c val = (cur - SC2731_CURRENT_PRECHG) / SC2731_CURRENT_STEP; cur 150 drivers/power/supply/sc2731_charger.c u32 *cur) cur 160 drivers/power/supply/sc2731_charger.c *cur = val * SC2731_CURRENT_STEP + SC2731_CURRENT_PRECHG; cur 166 drivers/power/supply/sc2731_charger.c u32 *cur) cur 179 drivers/power/supply/sc2731_charger.c *cur = SC2731_CURRENT_LIMIT_100; cur 183 drivers/power/supply/sc2731_charger.c *cur = SC2731_CURRENT_LIMIT_2000; cur 187 drivers/power/supply/sc2731_charger.c *cur = SC2731_CURRENT_LIMIT_900; cur 191 drivers/power/supply/sc2731_charger.c *cur = SC2731_CURRENT_LIMIT_500; cur 244 drivers/power/supply/sc2731_charger.c u32 cur; cur 260 drivers/power/supply/sc2731_charger.c ret = sc2731_charger_get_current(info, &cur); cur 264 drivers/power/supply/sc2731_charger.c val->intval = cur * 1000; cur 272 drivers/power/supply/sc2731_charger.c ret = sc2731_charger_get_current_limit(info, &cur); cur 276 drivers/power/supply/sc2731_charger.c val->intval = cur * 1000; cur 274 drivers/power/supply/sc27xx_fuel_gauge.c int volt, cur, oci, ocv, ret; cur 295 drivers/power/supply/sc27xx_fuel_gauge.c &cur); cur 299 drivers/power/supply/sc27xx_fuel_gauge.c cur <<= 1; cur 300 drivers/power/supply/sc27xx_fuel_gauge.c oci = sc27xx_fgu_adc_to_current(data, cur - SC27XX_FGU_CUR_BASIC_ADC); cur 420 drivers/power/supply/sc27xx_fuel_gauge.c int ret, cur; cur 422 drivers/power/supply/sc27xx_fuel_gauge.c ret = regmap_read(data->regmap, data->base + SC27XX_FGU_CURRENT, &cur); cur 430 drivers/power/supply/sc27xx_fuel_gauge.c *val = sc27xx_fgu_adc_to_current(data, cur - SC27XX_FGU_CUR_BASIC_ADC); cur 437 drivers/power/supply/sc27xx_fuel_gauge.c int vol, cur, ret; cur 443 drivers/power/supply/sc27xx_fuel_gauge.c ret = sc27xx_fgu_get_current(data, &cur); cur 448 drivers/power/supply/sc27xx_fuel_gauge.c *val = vol * 1000 - cur * data->internal_resist; cur 241 drivers/power/supply/twl4030_charger.c int cur; cur 252 drivers/power/supply/twl4030_charger.c cur = bci->ac_cur; cur 255 drivers/power/supply/twl4030_charger.c cur = bci->usb_cur; cur 257 drivers/power/supply/twl4030_charger.c if (cur > bci->usb_cur_target) { cur 258 drivers/power/supply/twl4030_charger.c cur = bci->usb_cur_target; cur 259 drivers/power/supply/twl4030_charger.c bci->usb_cur = cur; cur 261 drivers/power/supply/twl4030_charger.c if (cur < bci->usb_cur_target) cur 272 drivers/power/supply/twl4030_charger.c if (cur > 852000) cur 353 drivers/power/supply/twl4030_charger.c reg = ua2regval(cur, cgain); cur 481 drivers/pwm/pwm-samsung.c const __be32 *cur; cur 490 drivers/pwm/pwm-samsung.c of_property_for_each_u32(np, "samsung,pwm-outputs", prop, cur, val) { cur 97 drivers/pwm/pwm-sti.c struct pwm_device *cur; cur 169 drivers/pwm/pwm-sti.c struct pwm_device *cur = pc->cur; cur 176 drivers/pwm/pwm-sti.c period_same = (period_ns == pwm_get_period(cur)); cur 191 drivers/pwm/pwm-sti.c ((ncfg == 1) && (pwm->hwpwm == cur->hwpwm)) || cur 192 drivers/pwm/pwm-sti.c ((ncfg == 1) && (pwm->hwpwm != cur->hwpwm) && period_same) || cur 236 drivers/pwm/pwm-sti.c pc->cur = pwm; cur 382 drivers/s390/block/dasd_devmap.c char *cur; cur 386 drivers/s390/block/dasd_devmap.c cur = dasd[i]; cur 387 drivers/s390/block/dasd_devmap.c if (!cur) cur 389 drivers/s390/block/dasd_devmap.c if (*cur == '\0') cur 392 drivers/s390/block/dasd_devmap.c rc = dasd_parse_keyword(cur); cur 394 drivers/s390/block/dasd_devmap.c rc = dasd_parse_range(cur); cur 3234 drivers/scsi/gdth.c char *cur = str; cur 3239 drivers/scsi/gdth.c while (cur && isdigit(*cur) && i < MAXHA) { cur 3240 drivers/scsi/gdth.c ints[i++] = simple_strtoul(cur, NULL, 0); cur 3241 drivers/scsi/gdth.c if ((cur = strchr(cur, ',')) != NULL) cur++; cur 3245 drivers/scsi/gdth.c internal_setup(cur, ints); cur 583 drivers/scsi/ibmvscsi/ibmvfc.c vhost->async_crq.cur = 0; cur 715 drivers/scsi/ibmvscsi/ibmvfc.c crq->cur = 0; cur 3110 drivers/scsi/ibmvscsi/ibmvfc.c crq = &async_crq->msgs[async_crq->cur]; cur 3112 drivers/scsi/ibmvscsi/ibmvfc.c if (++async_crq->cur == async_crq->size) cur 3113 drivers/scsi/ibmvscsi/ibmvfc.c async_crq->cur = 0; cur 3133 drivers/scsi/ibmvscsi/ibmvfc.c crq = &queue->msgs[queue->cur]; cur 3135 drivers/scsi/ibmvscsi/ibmvfc.c if (++queue->cur == queue->size) cur 3136 drivers/scsi/ibmvscsi/ibmvfc.c queue->cur = 0; cur 4551 drivers/scsi/ibmvscsi/ibmvfc.c crq->cur = 0; cur 552 drivers/scsi/ibmvscsi/ibmvfc.h int size, cur; cur 577 drivers/scsi/ibmvscsi/ibmvfc.h int size, cur; cur 172 drivers/scsi/ibmvscsi/ibmvscsi.c crq = &queue->msgs[queue->cur]; cur 174 drivers/scsi/ibmvscsi/ibmvscsi.c if (++queue->cur == queue->size) cur 175 drivers/scsi/ibmvscsi/ibmvscsi.c queue->cur = 0; cur 299 drivers/scsi/ibmvscsi/ibmvscsi.c queue->cur = 0; cur 366 drivers/scsi/ibmvscsi/ibmvscsi.c queue->cur = 0; cur 45 drivers/scsi/ibmvscsi/ibmvscsi.h int size, cur; cur 638 drivers/scsi/ncr53c8xx.c char *cur = setup_token; cur 642 drivers/scsi/ncr53c8xx.c while (cur != NULL && (pc = strchr(cur, ':')) != NULL) { cur 645 drivers/scsi/ncr53c8xx.c if (!strncmp(p, cur, pc - cur)) cur 647 drivers/scsi/ncr53c8xx.c cur = pc; cur 655 drivers/scsi/ncr53c8xx.c char *cur = str; cur 660 drivers/scsi/ncr53c8xx.c while (cur != NULL && (pc = strchr(cur, ':')) != NULL) { cur 674 drivers/scsi/ncr53c8xx.c switch (get_setup_token(cur)) { cur 763 drivers/scsi/ncr53c8xx.c printk("sym53c8xx_setup: unexpected boot option '%.*s' ignored\n", (int)(pc-cur+1), cur); cur 767 drivers/scsi/ncr53c8xx.c if ((cur = strchr(cur, ARG_SEP)) != NULL) cur 768 drivers/scsi/ncr53c8xx.c ++cur; cur 188 drivers/scsi/snic/snic_disc.c struct list_head *cur, *nxt; cur 191 drivers/scsi/snic/snic_disc.c list_for_each_safe(cur, nxt, &snic->disc.tgt_list) { cur 192 drivers/scsi/snic/snic_disc.c tgt = list_entry(cur, struct snic_tgt, list); cur 545 drivers/scsi/snic/snic_disc.c struct list_head *cur, *nxt; cur 553 drivers/scsi/snic/snic_disc.c list_for_each_safe(cur, nxt, &snic->disc.tgt_list) { cur 554 drivers/scsi/snic/snic_disc.c tgt = list_entry(cur, struct snic_tgt, list); cur 401 drivers/scsi/snic/snic_io.c struct list_head *cur, *nxt; cur 405 drivers/scsi/snic/snic_io.c list_for_each_safe(cur, nxt, &snic->spl_cmd_list) { cur 406 drivers/scsi/snic/snic_io.c rqi = list_entry(cur, struct snic_req_info, list); cur 351 drivers/scsi/sym53c8xx_2/sym_fw.c u32 *end, *cur; cur 354 drivers/scsi/sym53c8xx_2/sym_fw.c cur = start; cur 357 drivers/scsi/sym53c8xx_2/sym_fw.c while (cur < end) { cur 359 drivers/scsi/sym53c8xx_2/sym_fw.c opcode = *cur; cur 369 drivers/scsi/sym53c8xx_2/sym_fw.c sym_name(np), (int) (cur-start)); cur 370 drivers/scsi/sym53c8xx_2/sym_fw.c ++cur; cur 379 drivers/scsi/sym53c8xx_2/sym_fw.c *cur++ = 0; cur 384 drivers/scsi/sym53c8xx_2/sym_fw.c printf ("%d: <%x>\n", (int) (cur-start), cur 408 drivers/scsi/sym53c8xx_2/sym_fw.c tmp1 = cur[1]; cur 409 drivers/scsi/sym53c8xx_2/sym_fw.c tmp2 = cur[2]; cur 412 drivers/scsi/sym53c8xx_2/sym_fw.c sym_name(np), (int) (cur-start)); cur 485 drivers/scsi/sym53c8xx_2/sym_fw.c *cur++ = cpu_to_scr(opcode); cur 492 drivers/scsi/sym53c8xx_2/sym_fw.c *cur = cpu_to_scr(*cur); cur 493 drivers/scsi/sym53c8xx_2/sym_fw.c ++cur; cur 501 drivers/scsi/sym53c8xx_2/sym_fw.c old = *cur; cur 534 drivers/scsi/sym53c8xx_2/sym_fw.c *cur++ = cpu_to_scr(new); cur 374 drivers/staging/isdn/gigaset/common.c struct at_state_t *cur, *next; cur 376 drivers/staging/isdn/gigaset/common.c list_for_each_entry_safe(cur, next, &cs->temp_at_states, list) { cur 377 drivers/staging/isdn/gigaset/common.c list_del(&cur->list); cur 378 drivers/staging/isdn/gigaset/common.c clear_at_state(cur); cur 379 drivers/staging/isdn/gigaset/common.c kfree(cur); cur 28 drivers/staging/kpc2000/kpc_dma/dma.c struct kpc_dma_descriptor *cur; cur 39 drivers/staging/kpc2000/kpc_dma/dma.c cur = eng->desc_completed; cur 41 drivers/staging/kpc2000/kpc_dma/dma.c cur = cur->Next; cur 42 drivers/staging/kpc2000/kpc_dma/dma.c dev_dbg(&eng->pldev->dev, "Handling completed descriptor %p (acd = %p)\n", cur, cur->acd); cur 43 drivers/staging/kpc2000/kpc_dma/dma.c BUG_ON(cur == eng->desc_next); // Ordering failure. cur 45 drivers/staging/kpc2000/kpc_dma/dma.c if (cur->DescControlFlags & DMA_DESC_CTL_SOP) { cur 50 drivers/staging/kpc2000/kpc_dma/dma.c eng->accumulated_bytes += cur->DescByteCount; cur 51 drivers/staging/kpc2000/kpc_dma/dma.c if (cur->DescStatusFlags & DMA_DESC_STS_ERROR) cur 54 drivers/staging/kpc2000/kpc_dma/dma.c if (cur->DescStatusFlags & DMA_DESC_STS_SHORT) cur 57 drivers/staging/kpc2000/kpc_dma/dma.c if (cur->DescControlFlags & DMA_DESC_CTL_EOP) { cur 58 drivers/staging/kpc2000/kpc_dma/dma.c if (cur->acd) cur 59 drivers/staging/kpc2000/kpc_dma/dma.c transfer_complete_cb(cur->acd, eng->accumulated_bytes, eng->accumulated_flags | ACD_FLAG_DONE); cur 62 drivers/staging/kpc2000/kpc_dma/dma.c eng->desc_completed = cur; cur 63 drivers/staging/kpc2000/kpc_dma/dma.c } while (cur->MyDMAAddr != GetEngineCompletePtr(eng)); cur 88 drivers/staging/kpc2000/kpc_dma/dma.c struct kpc_dma_descriptor *cur; cur 119 drivers/staging/kpc2000/kpc_dma/dma.c cur = eng->desc_pool_first; cur 128 drivers/staging/kpc2000/kpc_dma/dma.c cur->DescNextDescPtr = next_handle; cur 129 drivers/staging/kpc2000/kpc_dma/dma.c cur->Next = next; cur 130 drivers/staging/kpc2000/kpc_dma/dma.c cur = next; cur 135 drivers/staging/kpc2000/kpc_dma/dma.c cur->Next = eng->desc_pool_first; cur 136 drivers/staging/kpc2000/kpc_dma/dma.c cur->DescNextDescPtr = eng->desc_pool_first->MyDMAAddr; cur 138 drivers/staging/kpc2000/kpc_dma/dma.c eng->desc_pool_last = cur; cur 209 drivers/staging/kpc2000/kpc_dma/dma.c struct kpc_dma_descriptor *cur; cur 215 drivers/staging/kpc2000/kpc_dma/dma.c cur = eng->desc_pool_first; cur 219 drivers/staging/kpc2000/kpc_dma/dma.c struct kpc_dma_descriptor *next = cur->Next; cur 220 drivers/staging/kpc2000/kpc_dma/dma.c dma_addr_t next_handle = cur->DescNextDescPtr; cur 222 drivers/staging/kpc2000/kpc_dma/dma.c dma_pool_free(eng->desc_pool, cur, cur_handle); cur 224 drivers/staging/kpc2000/kpc_dma/dma.c cur = next; cur 236 drivers/staging/kpc2000/kpc_dma/dma.c struct kpc_dma_descriptor *cur = eng->desc_next; cur 238 drivers/staging/kpc2000/kpc_dma/dma.c while (cur != eng->desc_completed) { cur 239 drivers/staging/kpc2000/kpc_dma/dma.c BUG_ON(cur == NULL); cur 241 drivers/staging/kpc2000/kpc_dma/dma.c cur = cur->Next; cur 271 drivers/staging/kpc2000/kpc_dma/fileops.c struct kpc_dma_descriptor *cur; cur 279 drivers/staging/kpc2000/kpc_dma/fileops.c cur = eng->desc_completed->Next; cur 280 drivers/staging/kpc2000/kpc_dma/fileops.c while (cur != eng->desc_next) { cur 281 drivers/staging/kpc2000/kpc_dma/fileops.c dev_dbg(&eng->pldev->dev, "Aborting descriptor %p (acd = %p)\n", cur, cur->acd); cur 282 drivers/staging/kpc2000/kpc_dma/fileops.c if (cur->DescControlFlags & DMA_DESC_CTL_EOP) { cur 283 drivers/staging/kpc2000/kpc_dma/fileops.c if (cur->acd) cur 284 drivers/staging/kpc2000/kpc_dma/fileops.c transfer_complete_cb(cur->acd, 0, ACD_FLAG_ABORT); cur 287 drivers/staging/kpc2000/kpc_dma/fileops.c clear_desc(cur); cur 288 drivers/staging/kpc2000/kpc_dma/fileops.c eng->desc_completed = cur; cur 290 drivers/staging/kpc2000/kpc_dma/fileops.c cur = cur->Next; cur 99 drivers/staging/media/imx/imx-media-fim.c fim->enabled = en->cur.val; cur 100 drivers/staging/media/imx/imx-media-fim.c fim->icap_flags = icap_edge->cur.val; cur 101 drivers/staging/media/imx/imx-media-fim.c fim->icap_channel = icap_chan->cur.val; cur 102 drivers/staging/media/imx/imx-media-fim.c fim->num_avg = num->cur.val; cur 103 drivers/staging/media/imx/imx-media-fim.c fim->num_skip = skip->cur.val; cur 104 drivers/staging/media/imx/imx-media-fim.c fim->tolerance_min = tol_min->cur.val; cur 105 drivers/staging/media/imx/imx-media-fim.c fim->tolerance_max = tol_max->cur.val; cur 2513 drivers/staging/rtl8188eu/core/rtw_mlme_ext.c struct wlan_bssid_ex *cur = &pmlmeinfo->network; cur 2529 drivers/staging/rtl8188eu/core/rtw_mlme_ext.c if ((ielen != 0 && memcmp((void *)(p+2), (void *)cur->ssid.ssid, cur->ssid.ssid_length)) || cur 2908 drivers/staging/rtl8188eu/core/rtw_mlme_ext.c struct wlan_bssid_ex *cur = &pmlmeinfo->network; cur 2978 drivers/staging/rtl8188eu/core/rtw_mlme_ext.c if (memcmp((void *)(p+2), cur->ssid.ssid, cur->ssid.ssid_length)) cur 2981 drivers/staging/rtl8188eu/core/rtw_mlme_ext.c if (ie_len != cur->ssid.ssid_length) cur 605 drivers/staging/rtl8723bs/core/rtw_mlme_ext.c struct wlan_bssid_ex *cur = &pmlmeinfo->network; cur 739 drivers/staging/rtl8723bs/core/rtw_mlme_ext.c if ((ielen != 0 && false == !memcmp((void *)(p+2), (void *)cur->Ssid.Ssid, cur->Ssid.SsidLength)) cur 1187 drivers/staging/rtl8723bs/core/rtw_mlme_ext.c struct wlan_bssid_ex *cur = &(pmlmeinfo->network); cur 1265 drivers/staging/rtl8723bs/core/rtw_mlme_ext.c if (memcmp((void *)(p+2), cur->Ssid.Ssid, cur->Ssid.SsidLength)) cur 1268 drivers/staging/rtl8723bs/core/rtw_mlme_ext.c if (ie_len != cur->Ssid.SsidLength) cur 532 drivers/tty/vt/keyboard.c int i, cur = fg_console; cur 536 drivers/tty/vt/keyboard.c cur = want_console; cur 538 drivers/tty/vt/keyboard.c for (i = cur - 1; i != cur; i--) { cur 549 drivers/tty/vt/keyboard.c int i, cur = fg_console; cur 553 drivers/tty/vt/keyboard.c cur = want_console; cur 555 drivers/tty/vt/keyboard.c for (i = cur+1; i != cur; i++) { cur 1209 drivers/usb/gadget/udc/bcm63xx_udc.c struct bcm63xx_req *breq = our_req(req), *cur; cur 1219 drivers/usb/gadget/udc/bcm63xx_udc.c cur = list_first_entry(&bep->queue, struct bcm63xx_req, queue); cur 1222 drivers/usb/gadget/udc/bcm63xx_udc.c if (breq == cur) { cur 385 drivers/usb/musb/tusb6010_omap.c int cur = (reg & (0xf << (i * 5))) >> (i * 5); cur 386 drivers/usb/musb/tusb6010_omap.c if (cur == 0) { cur 376 drivers/usb/typec/altmodes/displayport.c u8 cur; cur 382 drivers/usb/typec/altmodes/displayport.c cur = DP_CONF_CURRENTLY(dp->data.conf); cur 384 drivers/usb/typec/altmodes/displayport.c len = sprintf(buf, "%s ", cur ? "USB" : "[USB]"); cur 387 drivers/usb/typec/altmodes/displayport.c if (i == cur) cur 469 drivers/usb/typec/altmodes/displayport.c u8 cur; cur 474 drivers/usb/typec/altmodes/displayport.c cur = get_count_order(DP_CONF_GET_PIN_ASSIGN(dp->data.conf)); cur 483 drivers/usb/typec/altmodes/displayport.c if (i == cur) cur 52 drivers/usb/typec/ucsi/displayport.c u8 cur = 0; cur 67 drivers/usb/typec/ucsi/displayport.c ret = ucsi_send_command(dp->con->ucsi, &ctrl, &cur, sizeof(cur)); cur 73 drivers/usb/typec/ucsi/displayport.c cur = 0xff; cur 76 drivers/usb/typec/ucsi/displayport.c if (cur != 0xff) { cur 78 drivers/usb/typec/ucsi/displayport.c if (dp->con->port_altmode[cur] == alt) cur 227 drivers/usb/typec/ucsi/ucsi.c u8 cur; cur 231 drivers/usb/typec/ucsi/ucsi.c ret = ucsi_run_command(con->ucsi, &ctrl, &cur, sizeof(cur)); cur 238 drivers/usb/typec/ucsi/ucsi.c cur = 0xff; cur 241 drivers/usb/typec/ucsi/ucsi.c if (cur < UCSI_MAX_ALTMODES) cur 242 drivers/usb/typec/ucsi/ucsi.c altmode = typec_altmode_get_partner(con->port_altmode[cur]); cur 558 drivers/vfio/pci/vfio_pci.c int cur; cur 567 drivers/vfio/pci/vfio_pci.c if (fill->cur == fill->max) cur 574 drivers/vfio/pci/vfio_pci.c fill->devices[fill->cur].group_id = iommu_group_id(iommu_group); cur 575 drivers/vfio/pci/vfio_pci.c fill->devices[fill->cur].segment = pci_domain_nr(pdev->bus); cur 576 drivers/vfio/pci/vfio_pci.c fill->devices[fill->cur].bus = pdev->bus->number; cur 577 drivers/vfio/pci/vfio_pci.c fill->devices[fill->cur].devfn = pdev->devfn; cur 578 drivers/vfio/pci/vfio_pci.c fill->cur++; cur 1000 drivers/vfio/pci/vfio_pci.c hdr.count = fill.cur; cur 64 drivers/video/backlight/cr_bllcd.c u32 cur = inl(addr); cur 76 drivers/video/backlight/cr_bllcd.c cur &= ~CRVML_BACKLIGHT_OFF; cur 77 drivers/video/backlight/cr_bllcd.c outl(cur, addr); cur 79 drivers/video/backlight/cr_bllcd.c cur |= CRVML_BACKLIGHT_OFF; cur 80 drivers/video/backlight/cr_bllcd.c outl(cur, addr); cur 89 drivers/video/backlight/cr_bllcd.c u32 cur = inl(addr); cur 92 drivers/video/backlight/cr_bllcd.c if (cur & CRVML_BACKLIGHT_OFF) cur 108 drivers/video/backlight/cr_bllcd.c u32 cur = inl(addr); cur 110 drivers/video/backlight/cr_bllcd.c if (!(cur & CRVML_PANEL_ON)) { cur 112 drivers/video/backlight/cr_bllcd.c if (cur & 0x00000001) { cur 113 drivers/video/backlight/cr_bllcd.c cur &= ~CRVML_LVDS_ON; cur 114 drivers/video/backlight/cr_bllcd.c outl(cur, addr); cur 118 drivers/video/backlight/cr_bllcd.c cur |= CRVML_PANEL_ON; cur 119 drivers/video/backlight/cr_bllcd.c outl(cur, addr); cur 124 drivers/video/backlight/cr_bllcd.c if (!(cur & CRVML_LVDS_ON)) { cur 126 drivers/video/backlight/cr_bllcd.c outl(cur | CRVML_LVDS_ON, addr); cur 133 drivers/video/backlight/cr_bllcd.c u32 cur = inl(addr); cur 136 drivers/video/backlight/cr_bllcd.c if (cur & CRVML_LVDS_ON) { cur 137 drivers/video/backlight/cr_bllcd.c cur &= ~CRVML_LVDS_ON; cur 138 drivers/video/backlight/cr_bllcd.c outl(cur, addr); cur 140 drivers/video/backlight/cr_bllcd.c if (cur & CRVML_PANEL_ON) { cur 142 drivers/video/backlight/cr_bllcd.c outl(cur & ~CRVML_PANEL_ON, addr); cur 176 drivers/video/console/vgacon.c int cur; cur 197 drivers/video/console/vgacon.c scrollback->cur = 0; cur 268 drivers/video/console/vgacon.c vgacon_scrollback_cur->cur = vgacon_scrollback_cur->cnt; cur 282 drivers/video/console/vgacon.c vgacon_scrollback_cur->cur = vgacon_scrollback_cur->cnt; cur 306 drivers/video/console/vgacon.c start = vgacon_scrollback_cur->cur + lines; cur 321 drivers/video/console/vgacon.c vgacon_scrollback_cur->cur = start; cur 937 drivers/video/fbdev/broadsheetfb.c struct page *cur; cur 947 drivers/video/fbdev/broadsheetfb.c list_for_each_entry(cur, &fbdefio->pagelist, lru) { cur 950 drivers/video/fbdev/broadsheetfb.c y1 = (cur->index << PAGE_SHIFT) / xres; cur 952 drivers/video/fbdev/broadsheetfb.c } else if ((prev_index + 1) == cur->index) { cur 959 drivers/video/fbdev/broadsheetfb.c y1 = (cur->index << PAGE_SHIFT) / xres; cur 962 drivers/video/fbdev/broadsheetfb.c prev_index = cur->index; cur 98 drivers/video/fbdev/core/fb_defio.c struct page *cur; cur 127 drivers/video/fbdev/core/fb_defio.c list_for_each_entry(cur, &fbdefio->pagelist, lru) { cur 133 drivers/video/fbdev/core/fb_defio.c if (unlikely(cur == page)) cur 135 drivers/video/fbdev/core/fb_defio.c else if (cur->index > page->index) cur 139 drivers/video/fbdev/core/fb_defio.c list_add_tail(&page->lru, &cur->lru); cur 182 drivers/video/fbdev/core/fb_defio.c struct page *cur; cur 187 drivers/video/fbdev/core/fb_defio.c list_for_each_entry(cur, &fbdefio->pagelist, lru) { cur 188 drivers/video/fbdev/core/fb_defio.c lock_page(cur); cur 189 drivers/video/fbdev/core/fb_defio.c page_mkclean(cur); cur 190 drivers/video/fbdev/core/fb_defio.c unlock_page(cur); cur 472 drivers/video/fbdev/metronomefb.c struct page *cur; cur 477 drivers/video/fbdev/metronomefb.c list_for_each_entry(cur, &fbdefio->pagelist, lru) { cur 479 drivers/video/fbdev/metronomefb.c (cur->index << PAGE_SHIFT)); cur 480 drivers/video/fbdev/metronomefb.c par->metromem_img_csum -= par->csum_table[cur->index]; cur 481 drivers/video/fbdev/metronomefb.c par->csum_table[cur->index] = cksum; cur 75 drivers/video/fbdev/nvidia/nv_hw.c int cur = par->CurrentState->cursor1; cur 85 drivers/video/fbdev/nvidia/nv_hw.c return (cur & 0x01); cur 173 drivers/video/fbdev/riva/riva_hw.c char cur; cur 251 drivers/video/fbdev/riva/riva_hw.c int last, next, cur; cur 260 drivers/video/fbdev/riva/riva_hw.c cur = ainfo->cur; cur 328 drivers/video/fbdev/riva/riva_hw.c last = cur; cur 329 drivers/video/fbdev/riva/riva_hw.c cur = next; cur 331 drivers/video/fbdev/riva/riva_hw.c switch (cur) cur 334 drivers/video/fbdev/riva/riva_hw.c if (last==cur) misses = 0; cur 338 drivers/video/fbdev/riva/riva_hw.c if (last!=cur) cur 350 drivers/video/fbdev/riva/riva_hw.c if (last==cur) misses = 0; cur 354 drivers/video/fbdev/riva/riva_hw.c if (last!=cur) cur 366 drivers/video/fbdev/riva/riva_hw.c if (last==cur) misses = 0; cur 455 drivers/video/fbdev/riva/riva_hw.c ainfo->cur = ENGINE; cur 467 drivers/video/fbdev/riva/riva_hw.c ainfo->cur = MPORT; cur 482 drivers/video/fbdev/riva/riva_hw.c ainfo->cur = GRAPHICS; cur 494 drivers/video/fbdev/riva/riva_hw.c ainfo->cur = VIDEO; cur 958 drivers/video/fbdev/smscufx.c struct page *cur; cur 969 drivers/video/fbdev/smscufx.c list_for_each_entry(cur, &fbdefio->pagelist, lru) { cur 974 drivers/video/fbdev/smscufx.c const int y = (cur->index << PAGE_SHIFT) / (width * 2); cur 784 drivers/video/fbdev/udlfb.c struct page *cur; cur 811 drivers/video/fbdev/udlfb.c list_for_each_entry(cur, &fbdefio->pagelist, lru) { cur 814 drivers/video/fbdev/udlfb.c &cmd, cur->index << PAGE_SHIFT, cur 887 drivers/video/fbdev/vermilion/vermilion.c u32 cur = VML_READ32(par, VML_PIPEACONF); cur 894 drivers/video/fbdev/vermilion/vermilion.c VML_WRITE32(par, VML_PIPEACONF, cur & ~VML_PIPE_FORCE_BORDER); cur 901 drivers/video/fbdev/vermilion/vermilion.c VML_WRITE32(par, VML_PIPEACONF, cur | VML_PIPE_FORCE_BORDER); cur 1376 drivers/video/fbdev/via/hw.c struct via_pll_config cur, up, down, best = {0, 1, 0}; cur 1381 drivers/video/fbdev/via/hw.c cur.rshift = limits[i].rshift; cur 1382 drivers/video/fbdev/via/hw.c cur.divisor = limits[i].divisor; cur 1383 drivers/video/fbdev/via/hw.c cur.multiplier = clk / ((f0 / cur.divisor)>>cur.rshift); cur 1384 drivers/video/fbdev/via/hw.c f = abs(get_pll_output_frequency(f0, cur) - clk); cur 1385 drivers/video/fbdev/via/hw.c up = down = cur; cur 1389 drivers/video/fbdev/via/hw.c cur = up; cur 1391 drivers/video/fbdev/via/hw.c cur = down; cur 1393 drivers/video/fbdev/via/hw.c if (cur.multiplier < limits[i].multiplier_min) cur 1394 drivers/video/fbdev/via/hw.c cur.multiplier = limits[i].multiplier_min; cur 1395 drivers/video/fbdev/via/hw.c else if (cur.multiplier > limits[i].multiplier_max) cur 1396 drivers/video/fbdev/via/hw.c cur.multiplier = limits[i].multiplier_max; cur 1398 drivers/video/fbdev/via/hw.c f = abs(get_pll_output_frequency(f0, cur) - clk); cur 1400 drivers/video/fbdev/via/hw.c best = cur; cur 58 drivers/watchdog/bcm2835_wdt.c uint32_t cur; cur 60 drivers/watchdog/bcm2835_wdt.c cur = readl(wdt->base + PM_RSTC); cur 62 drivers/watchdog/bcm2835_wdt.c return !!(cur & PM_RSTC_WRCFG_FULL_RESET); cur 68 drivers/watchdog/bcm2835_wdt.c uint32_t cur; cur 75 drivers/watchdog/bcm2835_wdt.c cur = readl_relaxed(wdt->base + PM_RSTC); cur 76 drivers/watchdog/bcm2835_wdt.c writel_relaxed(PM_PASSWORD | (cur & PM_RSTC_WRCFG_CLR) | cur 112 drivers/watchdog/bcm_kona_wdt.c int ctl, cur, ctl_sec, cur_sec, res; cur 116 drivers/watchdog/bcm_kona_wdt.c cur = cur_val & SECWDOG_COUNT_MASK; cur 118 drivers/watchdog/bcm_kona_wdt.c cur_sec = TICKS_TO_SECS(cur, wdt); cur 126 drivers/watchdog/bcm_kona_wdt.c cur_sec, cur, cur, cur 1389 drivers/xen/grant-table.c unsigned int cur, extra; cur 1392 drivers/xen/grant-table.c cur = nr_grant_frames; cur 1395 drivers/xen/grant-table.c if (cur + extra > gnttab_max_grant_frames()) { cur 1399 drivers/xen/grant-table.c cur, extra, gnttab_max_grant_frames(), cur 1404 drivers/xen/grant-table.c rc = gnttab_map(cur, cur + extra - 1); cur 104 fs/afs/server_list.c struct afs_server *cur; cur 118 fs/afs/server_list.c cur = old->servers[old->preferred].server; cur 120 fs/afs/server_list.c if (new->servers[j].server == cur) { cur 2004 fs/btrfs/backref.c u32 cur; cur 2039 fs/btrfs/backref.c for (cur = 0; cur < btrfs_item_size(eb, item); cur += len) { cur 2044 fs/btrfs/backref.c cur, found_key.objectid, cur 1203 fs/btrfs/check-integrity.c size_t cur; cur 1214 fs/btrfs/check-integrity.c cur = min(len, ((size_t)PAGE_SIZE - pgoff)); cur 1217 fs/btrfs/check-integrity.c memcpy(dst, kaddr + pgoff, cur); cur 1219 fs/btrfs/check-integrity.c dst += cur; cur 1220 fs/btrfs/check-integrity.c len -= cur; cur 404 fs/btrfs/ctree.c struct tree_mod_elem *cur; cur 413 fs/btrfs/ctree.c cur = rb_entry(*new, struct tree_mod_elem, node); cur 415 fs/btrfs/ctree.c if (cur->logical < tm->logical) cur 417 fs/btrfs/ctree.c else if (cur->logical > tm->logical) cur 419 fs/btrfs/ctree.c else if (cur->seq < tm->seq) cur 421 fs/btrfs/ctree.c else if (cur->seq > tm->seq) cur 686 fs/btrfs/ctree.c struct tree_mod_elem *cur = NULL; cur 693 fs/btrfs/ctree.c cur = rb_entry(node, struct tree_mod_elem, node); cur 694 fs/btrfs/ctree.c if (cur->logical < start) { cur 696 fs/btrfs/ctree.c } else if (cur->logical > start) { cur 698 fs/btrfs/ctree.c } else if (cur->seq < min_seq) { cur 703 fs/btrfs/ctree.c BUG_ON(found->seq > cur->seq); cur 704 fs/btrfs/ctree.c found = cur; cur 706 fs/btrfs/ctree.c } else if (cur->seq > min_seq) { cur 709 fs/btrfs/ctree.c BUG_ON(found->seq < cur->seq); cur 710 fs/btrfs/ctree.c found = cur; cur 713 fs/btrfs/ctree.c found = cur; cur 1590 fs/btrfs/ctree.c struct extent_buffer *cur; cur 1648 fs/btrfs/ctree.c cur = find_extent_buffer(fs_info, blocknr); cur 1649 fs/btrfs/ctree.c if (cur) cur 1650 fs/btrfs/ctree.c uptodate = btrfs_buffer_uptodate(cur, gen, 0); cur 1653 fs/btrfs/ctree.c if (!cur || !uptodate) { cur 1654 fs/btrfs/ctree.c if (!cur) { cur 1655 fs/btrfs/ctree.c cur = read_tree_block(fs_info, blocknr, gen, cur 1658 fs/btrfs/ctree.c if (IS_ERR(cur)) { cur 1659 fs/btrfs/ctree.c return PTR_ERR(cur); cur 1660 fs/btrfs/ctree.c } else if (!extent_buffer_uptodate(cur)) { cur 1661 fs/btrfs/ctree.c free_extent_buffer(cur); cur 1665 fs/btrfs/ctree.c err = btrfs_read_buffer(cur, gen, cur 1668 fs/btrfs/ctree.c free_extent_buffer(cur); cur 1676 fs/btrfs/ctree.c btrfs_tree_lock(cur); cur 1677 fs/btrfs/ctree.c btrfs_set_lock_blocking_write(cur); cur 1678 fs/btrfs/ctree.c err = __btrfs_cow_block(trans, root, cur, parent, i, cur 1679 fs/btrfs/ctree.c &cur, search_start, cur 1683 fs/btrfs/ctree.c btrfs_tree_unlock(cur); cur 1684 fs/btrfs/ctree.c free_extent_buffer(cur); cur 1687 fs/btrfs/ctree.c search_start = cur->start; cur 1688 fs/btrfs/ctree.c last_block = cur->start; cur 1690 fs/btrfs/ctree.c btrfs_tree_unlock(cur); cur 1691 fs/btrfs/ctree.c free_extent_buffer(cur); cur 5157 fs/btrfs/ctree.c struct extent_buffer *cur; cur 5168 fs/btrfs/ctree.c cur = btrfs_read_lock_root_node(root); cur 5169 fs/btrfs/ctree.c level = btrfs_header_level(cur); cur 5171 fs/btrfs/ctree.c path->nodes[level] = cur; cur 5174 fs/btrfs/ctree.c if (btrfs_header_generation(cur) < min_trans) { cur 5179 fs/btrfs/ctree.c nritems = btrfs_header_nritems(cur); cur 5180 fs/btrfs/ctree.c level = btrfs_header_level(cur); cur 5181 fs/btrfs/ctree.c sret = btrfs_bin_search(cur, min_key, level, &slot); cur 5193 fs/btrfs/ctree.c btrfs_item_key_to_cpu(cur, &found_key, slot); cur 5205 fs/btrfs/ctree.c gen = btrfs_node_ptr_generation(cur, slot); cur 5230 fs/btrfs/ctree.c btrfs_node_key_to_cpu(cur, &found_key, slot); cur 5237 fs/btrfs/ctree.c cur = btrfs_read_node_slot(cur, slot); cur 5238 fs/btrfs/ctree.c if (IS_ERR(cur)) { cur 5239 fs/btrfs/ctree.c ret = PTR_ERR(cur); cur 5243 fs/btrfs/ctree.c btrfs_tree_read_lock(cur); cur 5246 fs/btrfs/ctree.c path->nodes[level - 1] = cur; cur 378 fs/btrfs/dir-item.c u32 cur = 0; cur 386 fs/btrfs/dir-item.c while (cur < total_len) { cur 396 fs/btrfs/dir-item.c cur += this_len; cur 1725 fs/btrfs/disk-io.c struct btrfs_transaction *cur; cur 1737 fs/btrfs/disk-io.c cur = fs_info->running_transaction; cur 1738 fs/btrfs/disk-io.c if (!cur) { cur 1744 fs/btrfs/disk-io.c if (cur->state < TRANS_STATE_BLOCKED && cur 1746 fs/btrfs/disk-io.c (now < cur->start_time || cur 1747 fs/btrfs/disk-io.c now - cur->start_time < fs_info->commit_interval)) { cur 1752 fs/btrfs/disk-io.c transid = cur->transid; cur 1793 fs/btrfs/disk-io.c u64 cur; cur 1800 fs/btrfs/disk-io.c cur = btrfs_backup_tree_root_gen(root_backup); cur 1801 fs/btrfs/disk-io.c if (cur == newest_gen) cur 1808 fs/btrfs/disk-io.c cur = btrfs_backup_tree_root_gen(root_backup); cur 1809 fs/btrfs/disk-io.c if (cur == newest_gen) cur 3066 fs/btrfs/extent_io.c u64 cur = start; cur 3103 fs/btrfs/extent_io.c while (cur <= end) { cur 3107 fs/btrfs/extent_io.c if (cur >= last_byte) { cur 3116 fs/btrfs/extent_io.c set_extent_uptodate(tree, cur, cur + iosize - 1, cur 3118 fs/btrfs/extent_io.c unlock_extent_cached(tree, cur, cur 3119 fs/btrfs/extent_io.c cur + iosize - 1, &cached); cur 3122 fs/btrfs/extent_io.c em = __get_extent_map(inode, page, pg_offset, cur, cur 3123 fs/btrfs/extent_io.c end - cur + 1, get_extent, em_cached); cur 3126 fs/btrfs/extent_io.c unlock_extent(tree, cur, end); cur 3129 fs/btrfs/extent_io.c extent_offset = cur - em->start; cur 3130 fs/btrfs/extent_io.c BUG_ON(extent_map_end(em) <= cur); cur 3131 fs/btrfs/extent_io.c BUG_ON(end < cur); cur 3139 fs/btrfs/extent_io.c iosize = min(extent_map_end(em) - cur, end - cur + 1); cur 3209 fs/btrfs/extent_io.c set_extent_uptodate(tree, cur, cur + iosize - 1, cur 3211 fs/btrfs/extent_io.c unlock_extent_cached(tree, cur, cur 3212 fs/btrfs/extent_io.c cur + iosize - 1, &cached); cur 3213 fs/btrfs/extent_io.c cur = cur + iosize; cur 3218 fs/btrfs/extent_io.c if (test_range_bit(tree, cur, cur_end, cur 3221 fs/btrfs/extent_io.c unlock_extent(tree, cur, cur + iosize - 1); cur 3222 fs/btrfs/extent_io.c cur = cur + iosize; cur 3231 fs/btrfs/extent_io.c unlock_extent(tree, cur, cur + iosize - 1); cur 3232 fs/btrfs/extent_io.c cur = cur + iosize; cur 3249 fs/btrfs/extent_io.c unlock_extent(tree, cur, cur + iosize - 1); cur 3252 fs/btrfs/extent_io.c cur = cur + iosize; cur 3422 fs/btrfs/extent_io.c u64 cur = start; cur 3461 fs/btrfs/extent_io.c while (cur <= end) { cur 3465 fs/btrfs/extent_io.c if (cur >= i_size) { cur 3466 fs/btrfs/extent_io.c btrfs_writepage_endio_finish_ordered(page, cur, cur 3470 fs/btrfs/extent_io.c em = btrfs_get_extent(BTRFS_I(inode), page, pg_offset, cur, cur 3471 fs/btrfs/extent_io.c end - cur + 1, 1); cur 3478 fs/btrfs/extent_io.c extent_offset = cur - em->start; cur 3480 fs/btrfs/extent_io.c BUG_ON(em_end <= cur); cur 3481 fs/btrfs/extent_io.c BUG_ON(end < cur); cur 3482 fs/btrfs/extent_io.c iosize = min(em_end - cur, end - cur + 1); cur 3502 fs/btrfs/extent_io.c btrfs_writepage_endio_finish_ordered(page, cur, cur 3503 fs/btrfs/extent_io.c cur + iosize - 1, cur 3513 fs/btrfs/extent_io.c cur += iosize; cur 3518 fs/btrfs/extent_io.c btrfs_set_range_writeback(tree, cur, cur + iosize - 1); cur 3522 fs/btrfs/extent_io.c page->index, cur, end); cur 3536 fs/btrfs/extent_io.c cur = cur + iosize; cur 5558 fs/btrfs/extent_io.c size_t cur; cur 5578 fs/btrfs/extent_io.c cur = min(len, (PAGE_SIZE - offset)); cur 5580 fs/btrfs/extent_io.c memcpy(dst, kaddr + offset, cur); cur 5582 fs/btrfs/extent_io.c dst += cur; cur 5583 fs/btrfs/extent_io.c len -= cur; cur 5593 fs/btrfs/extent_io.c size_t cur; cur 5610 fs/btrfs/extent_io.c cur = min(len, (PAGE_SIZE - offset)); cur 5612 fs/btrfs/extent_io.c if (copy_to_user(dst, kaddr + offset, cur)) { cur 5617 fs/btrfs/extent_io.c dst += cur; cur 5618 fs/btrfs/extent_io.c len -= cur; cur 5671 fs/btrfs/extent_io.c size_t cur; cur 5688 fs/btrfs/extent_io.c cur = min(len, (PAGE_SIZE - offset)); cur 5691 fs/btrfs/extent_io.c ret = memcmp(ptr, kaddr + offset, cur); cur 5695 fs/btrfs/extent_io.c ptr += cur; cur 5696 fs/btrfs/extent_io.c len -= cur; cur 5727 fs/btrfs/extent_io.c size_t cur; cur 5744 fs/btrfs/extent_io.c cur = min(len, PAGE_SIZE - offset); cur 5746 fs/btrfs/extent_io.c memcpy(kaddr + offset, src, cur); cur 5748 fs/btrfs/extent_io.c src += cur; cur 5749 fs/btrfs/extent_io.c len -= cur; cur 5758 fs/btrfs/extent_io.c size_t cur; cur 5774 fs/btrfs/extent_io.c cur = min(len, PAGE_SIZE - offset); cur 5776 fs/btrfs/extent_io.c memset(kaddr + offset, 0, cur); cur 5778 fs/btrfs/extent_io.c len -= cur; cur 5803 fs/btrfs/extent_io.c size_t cur; cur 5818 fs/btrfs/extent_io.c cur = min(len, (unsigned long)(PAGE_SIZE - offset)); cur 5821 fs/btrfs/extent_io.c read_extent_buffer(src, kaddr + offset, src_offset, cur); cur 5823 fs/btrfs/extent_io.c src_offset += cur; cur 5824 fs/btrfs/extent_io.c len -= cur; cur 5999 fs/btrfs/extent_io.c size_t cur; cur 6026 fs/btrfs/extent_io.c cur = min(len, (unsigned long)(PAGE_SIZE - cur 6028 fs/btrfs/extent_io.c cur = min_t(unsigned long, cur, cur 6032 fs/btrfs/extent_io.c dst_off_in_page, src_off_in_page, cur); cur 6034 fs/btrfs/extent_io.c src_offset += cur; cur 6035 fs/btrfs/extent_io.c dst_offset += cur; cur 6036 fs/btrfs/extent_io.c len -= cur; cur 6044 fs/btrfs/extent_io.c size_t cur; cur 6076 fs/btrfs/extent_io.c cur = min_t(unsigned long, len, src_off_in_page + 1); cur 6077 fs/btrfs/extent_io.c cur = min(cur, dst_off_in_page + 1); cur 6079 fs/btrfs/extent_io.c dst_off_in_page - cur + 1, cur 6080 fs/btrfs/extent_io.c src_off_in_page - cur + 1, cur); cur 6082 fs/btrfs/extent_io.c dst_end -= cur; cur 6083 fs/btrfs/extent_io.c src_end -= cur; cur 6084 fs/btrfs/extent_io.c len -= cur; cur 496 fs/btrfs/extent_map.c struct extent_map *cur, cur 500 fs/btrfs/extent_map.c WARN_ON(test_bit(EXTENT_FLAG_PINNED, &cur->flags)); cur 501 fs/btrfs/extent_map.c ASSERT(extent_map_in_tree(cur)); cur 502 fs/btrfs/extent_map.c if (!test_bit(EXTENT_FLAG_LOGGING, &cur->flags)) cur 503 fs/btrfs/extent_map.c list_del_init(&cur->list); cur 504 fs/btrfs/extent_map.c rb_replace_node_cached(&cur->rb_node, &new->rb_node, &tree->map); cur 505 fs/btrfs/extent_map.c RB_CLEAR_NODE(&cur->rb_node); cur 91 fs/btrfs/extent_map.h struct extent_map *cur, cur 339 fs/btrfs/free-space-cache.c if (io_ctl->cur) { cur 340 fs/btrfs/free-space-cache.c io_ctl->cur = NULL; cur 349 fs/btrfs/free-space-cache.c io_ctl->cur = page_address(io_ctl->page); cur 350 fs/btrfs/free-space-cache.c io_ctl->orig = io_ctl->cur; cur 353 fs/btrfs/free-space-cache.c clear_page(io_ctl->cur); cur 422 fs/btrfs/free-space-cache.c io_ctl->cur += (sizeof(u32) * io_ctl->num_pages); cur 425 fs/btrfs/free-space-cache.c io_ctl->cur += sizeof(u64); cur 429 fs/btrfs/free-space-cache.c val = io_ctl->cur; cur 431 fs/btrfs/free-space-cache.c io_ctl->cur += sizeof(u64); cur 443 fs/btrfs/free-space-cache.c io_ctl->cur += sizeof(u32) * io_ctl->num_pages; cur 447 fs/btrfs/free-space-cache.c io_ctl->cur += sizeof(u64); cur 451 fs/btrfs/free-space-cache.c gen = io_ctl->cur; cur 459 fs/btrfs/free-space-cache.c io_ctl->cur += sizeof(u64); cur 521 fs/btrfs/free-space-cache.c if (!io_ctl->cur) cur 524 fs/btrfs/free-space-cache.c entry = io_ctl->cur; cur 529 fs/btrfs/free-space-cache.c io_ctl->cur += sizeof(struct btrfs_free_space_entry); cur 548 fs/btrfs/free-space-cache.c if (!io_ctl->cur) cur 555 fs/btrfs/free-space-cache.c if (io_ctl->cur != io_ctl->orig) { cur 562 fs/btrfs/free-space-cache.c copy_page(io_ctl->cur, bitmap); cur 575 fs/btrfs/free-space-cache.c if (io_ctl->cur != io_ctl->orig) cur 592 fs/btrfs/free-space-cache.c if (!io_ctl->cur) { cur 598 fs/btrfs/free-space-cache.c e = io_ctl->cur; cur 602 fs/btrfs/free-space-cache.c io_ctl->cur += sizeof(struct btrfs_free_space_entry); cur 622 fs/btrfs/free-space-cache.c copy_page(entry->bitmap, io_ctl->cur); cur 40 fs/btrfs/free-space-cache.h void *cur, *orig; cur 434 fs/btrfs/ordered-data.c struct list_head *cur; cur 447 fs/btrfs/ordered-data.c cur = entry->list.next; cur 448 fs/btrfs/ordered-data.c sum = list_entry(cur, struct btrfs_ordered_sum, list); cur 133 fs/btrfs/props.c u32 total_len, cur, this_len; cur 160 fs/btrfs/props.c cur = 0; cur 163 fs/btrfs/props.c while (cur < total_len) { cur 208 fs/btrfs/props.c cur += this_len; cur 3847 fs/btrfs/qgroup.c struct rb_node **cur; cur 3897 fs/btrfs/qgroup.c cur = &blocks->blocks[level].rb_node; cur 3898 fs/btrfs/qgroup.c while (*cur) { cur 3901 fs/btrfs/qgroup.c parent = *cur; cur 3906 fs/btrfs/qgroup.c cur = &(*cur)->rb_left; cur 3908 fs/btrfs/qgroup.c cur = &(*cur)->rb_right; cur 3929 fs/btrfs/qgroup.c rb_link_node(&block->node, parent, cur); cur 205 fs/btrfs/raid56.c struct btrfs_stripe_hash *cur; cur 232 fs/btrfs/raid56.c cur = h + i; cur 233 fs/btrfs/raid56.c INIT_LIST_HEAD(&cur->hash_list); cur 234 fs/btrfs/raid56.c spin_lock_init(&cur->lock); cur 555 fs/btrfs/raid56.c struct btrfs_raid_bio *cur) cur 558 fs/btrfs/raid56.c test_bit(RBIO_RMW_LOCKED_BIT, &cur->flags)) cur 569 fs/btrfs/raid56.c test_bit(RBIO_CACHE_BIT, &cur->flags)) cur 573 fs/btrfs/raid56.c cur->bbio->raid_map[0]) cur 577 fs/btrfs/raid56.c if (last->operation != cur->operation) cur 596 fs/btrfs/raid56.c int cur_fa = cur->faila; cur 597 fs/btrfs/raid56.c int cur_fb = cur->failb; cur 604 fs/btrfs/raid56.c if (cur->faila >= cur->failb) { cur 605 fs/btrfs/raid56.c cur_fa = cur->failb; cur 606 fs/btrfs/raid56.c cur_fb = cur->faila; cur 676 fs/btrfs/raid56.c struct btrfs_raid_bio *cur; cur 684 fs/btrfs/raid56.c list_for_each_entry(cur, &h->hash_list, hash_list) { cur 685 fs/btrfs/raid56.c if (cur->bbio->raid_map[0] == rbio->bbio->raid_map[0]) { cur 686 fs/btrfs/raid56.c spin_lock(&cur->bio_list_lock); cur 689 fs/btrfs/raid56.c if (bio_list_empty(&cur->bio_list) && cur 690 fs/btrfs/raid56.c list_empty(&cur->plug_list) && cur 691 fs/btrfs/raid56.c test_bit(RBIO_CACHE_BIT, &cur->flags) && cur 692 fs/btrfs/raid56.c !test_bit(RBIO_RMW_LOCKED_BIT, &cur->flags)) { cur 693 fs/btrfs/raid56.c list_del_init(&cur->hash_list); cur 694 fs/btrfs/raid56.c refcount_dec(&cur->refs); cur 696 fs/btrfs/raid56.c steal_rbio(cur, rbio); cur 697 fs/btrfs/raid56.c cache_drop = cur; cur 698 fs/btrfs/raid56.c spin_unlock(&cur->bio_list_lock); cur 704 fs/btrfs/raid56.c if (rbio_can_merge(cur, rbio)) { cur 705 fs/btrfs/raid56.c merge_rbio(cur, rbio); cur 706 fs/btrfs/raid56.c spin_unlock(&cur->bio_list_lock); cur 721 fs/btrfs/raid56.c list_for_each_entry(pending, &cur->plug_list, cur 725 fs/btrfs/raid56.c spin_unlock(&cur->bio_list_lock); cur 736 fs/btrfs/raid56.c list_add_tail(&rbio->plug_list, &cur->plug_list); cur 737 fs/btrfs/raid56.c spin_unlock(&cur->bio_list_lock); cur 857 fs/btrfs/raid56.c static void rbio_endio_bio_list(struct bio *cur, blk_status_t err) cur 861 fs/btrfs/raid56.c while (cur) { cur 862 fs/btrfs/raid56.c next = cur->bi_next; cur 863 fs/btrfs/raid56.c cur->bi_next = NULL; cur 864 fs/btrfs/raid56.c cur->bi_status = err; cur 865 fs/btrfs/raid56.c bio_endio(cur); cur 866 fs/btrfs/raid56.c cur = next; cur 876 fs/btrfs/raid56.c struct bio *cur = bio_list_get(&rbio->bio_list); cur 894 fs/btrfs/raid56.c rbio_endio_bio_list(cur, err); cur 1690 fs/btrfs/raid56.c struct btrfs_raid_bio *cur; cur 1700 fs/btrfs/raid56.c cur = list_entry(plug->rbio_list.next, cur 1702 fs/btrfs/raid56.c list_del_init(&cur->plug_list); cur 1704 fs/btrfs/raid56.c if (rbio_is_full(cur)) { cur 1708 fs/btrfs/raid56.c ret = full_stripe_write(cur); cur 1713 fs/btrfs/raid56.c if (rbio_can_merge(last, cur)) { cur 1714 fs/btrfs/raid56.c merge_rbio(last, cur); cur 1715 fs/btrfs/raid56.c __free_raid_bio(cur); cur 1721 fs/btrfs/raid56.c last = cur; cur 691 fs/btrfs/relocation.c struct backref_node *cur; cur 726 fs/btrfs/relocation.c cur = node; cur 730 fs/btrfs/relocation.c key.objectid = cur->bytenr; cur 747 fs/btrfs/relocation.c WARN_ON(cur->checked); cur 748 fs/btrfs/relocation.c if (!list_empty(&cur->upper)) { cur 753 fs/btrfs/relocation.c ASSERT(list_is_singular(&cur->upper)); cur 754 fs/btrfs/relocation.c edge = list_entry(cur->upper.next, struct backref_edge, cur 785 fs/btrfs/relocation.c if (key.objectid != cur->bytenr) { cur 837 fs/btrfs/relocation.c root = find_reloc_root(rc, cur->bytenr); cur 839 fs/btrfs/relocation.c cur->root = root; cur 857 fs/btrfs/relocation.c upper->level = cur->level + 1; cur 869 fs/btrfs/relocation.c list_add_tail(&edge->list[LOWER], &cur->upper); cur 870 fs/btrfs/relocation.c edge->node[LOWER] = cur; cur 896 fs/btrfs/relocation.c cur->cowonly = 1; cur 898 fs/btrfs/relocation.c if (btrfs_root_level(&root->root_item) == cur->level) { cur 901 fs/btrfs/relocation.c cur->bytenr); cur 903 fs/btrfs/relocation.c list_add(&cur->list, &useless); cur 905 fs/btrfs/relocation.c cur->root = root; cur 909 fs/btrfs/relocation.c level = cur->level + 1; cur 926 fs/btrfs/relocation.c cur->bytenr) { cur 929 fs/btrfs/relocation.c cur->bytenr, level - 1, cur 936 fs/btrfs/relocation.c lower = cur; cur 1029 fs/btrfs/relocation.c cur->checked = 1; cur 1036 fs/btrfs/relocation.c cur = edge->node[UPPER]; cur 899 fs/btrfs/send.c u32 cur = 0; cur 935 fs/btrfs/send.c while (cur < total) { cur 939 fs/btrfs/send.c iref = (struct btrfs_inode_ref *)(ptr + cur); cur 945 fs/btrfs/send.c extref = (struct btrfs_inode_extref *)(ptr + cur); cur 984 fs/btrfs/send.c cur += elem_size + name_len; cur 1021 fs/btrfs/send.c u32 cur; cur 1045 fs/btrfs/send.c cur = 0; cur 1050 fs/btrfs/send.c while (cur < total) { cur 1103 fs/btrfs/send.c cur += len; cur 2079 fs/btrfs/send.c struct name_cache_entry *cur; cur 2085 fs/btrfs/send.c list_for_each_entry(cur, nce_head, radix_list) { cur 2086 fs/btrfs/send.c if (cur->ino == ino && cur->gen == gen) cur 2087 fs/btrfs/send.c return cur; cur 2816 fs/btrfs/send.c struct recorded_ref *cur; cur 2819 fs/btrfs/send.c cur = list_entry(head->next, struct recorded_ref, list); cur 2820 fs/btrfs/send.c fs_path_free(cur->full_path); cur 2821 fs/btrfs/send.c list_del(&cur->list); cur 2822 fs/btrfs/send.c kfree(cur); cur 3096 fs/btrfs/send.c struct recorded_ref *cur; cur 3123 fs/btrfs/send.c list_for_each_entry(cur, deleted_refs, list) { cur 3124 fs/btrfs/send.c ret = dup_ref(cur, &pm->update_refs); cur 3128 fs/btrfs/send.c list_for_each_entry(cur, new_refs, list) { cur 3129 fs/btrfs/send.c ret = dup_ref(cur, &pm->update_refs); cur 3219 fs/btrfs/send.c struct recorded_ref *cur; cur 3326 fs/btrfs/send.c list_for_each_entry(cur, &pm->update_refs, list) { cur 3330 fs/btrfs/send.c ret = get_inode_info(sctx->send_root, cur->dir, NULL, cur 3339 fs/btrfs/send.c ret = send_utimes(sctx, cur->dir, cur->dir_gen); cur 3812 fs/btrfs/send.c struct recorded_ref *cur; cur 3873 fs/btrfs/send.c list_for_each_entry(cur, &sctx->new_refs, list) { cur 3881 fs/btrfs/send.c ret = get_cur_inode_state(sctx, cur->dir, cur->dir_gen); cur 3891 fs/btrfs/send.c if (cur == cur2) cur 3893 fs/btrfs/send.c if (cur2->dir == cur->dir) { cur 3904 fs/btrfs/send.c ret = did_create_dir(sctx, cur->dir); cur 3908 fs/btrfs/send.c ret = send_create_inode(sctx, cur->dir); cur 3920 fs/btrfs/send.c ret = will_overwrite_ref(sctx, cur->dir, cur->dir_gen, cur 3921 fs/btrfs/send.c cur->name, cur->name_len, cur 3927 fs/btrfs/send.c ow_inode, cur->dir, cur->name, cur 3928 fs/btrfs/send.c cur->name_len); cur 3936 fs/btrfs/send.c cur->full_path); cur 3991 fs/btrfs/send.c ret = send_unlink(sctx, cur->full_path); cur 3998 fs/btrfs/send.c ret = wait_for_dest_dir_move(sctx, cur, is_orphan); cur 4009 fs/btrfs/send.c ret = wait_for_parent_move(sctx, cur, is_orphan); cur 4024 fs/btrfs/send.c ret = send_rename(sctx, valid_path, cur->full_path); cur 4028 fs/btrfs/send.c ret = fs_path_copy(valid_path, cur->full_path); cur 4039 fs/btrfs/send.c cur->full_path); cur 4042 fs/btrfs/send.c cur->full_path); cur 4054 fs/btrfs/send.c ret = update_ref_path(sctx, cur); cur 4058 fs/btrfs/send.c ret = send_link(sctx, cur->full_path, cur 4064 fs/btrfs/send.c ret = dup_ref(cur, &check_dirs); cur 4092 fs/btrfs/send.c list_for_each_entry(cur, &sctx->deleted_refs, list) { cur 4093 fs/btrfs/send.c ret = dup_ref(cur, &check_dirs); cur 4102 fs/btrfs/send.c cur = list_entry(sctx->deleted_refs.next, struct recorded_ref, cur 4104 fs/btrfs/send.c ret = dup_ref(cur, &check_dirs); cur 4113 fs/btrfs/send.c list_for_each_entry(cur, &sctx->deleted_refs, list) { cur 4114 fs/btrfs/send.c ret = did_overwrite_ref(sctx, cur->dir, cur->dir_gen, cur 4116 fs/btrfs/send.c cur->name, cur->name_len); cur 4128 fs/btrfs/send.c ret = update_ref_path(sctx, cur); cur 4132 fs/btrfs/send.c ret = send_unlink(sctx, cur->full_path); cur 4136 fs/btrfs/send.c ret = dup_ref(cur, &check_dirs); cur 4161 fs/btrfs/send.c list_for_each_entry(cur, &check_dirs, list) { cur 4167 fs/btrfs/send.c if (cur->dir > sctx->cur_ino) cur 4170 fs/btrfs/send.c ret = get_cur_inode_state(sctx, cur->dir, cur->dir_gen); cur 4177 fs/btrfs/send.c ret = send_utimes(sctx, cur->dir, cur->dir_gen); cur 4181 fs/btrfs/send.c cur->dir != last_dir_ino_rm) { cur 4182 fs/btrfs/send.c ret = can_rmdir(sctx, cur->dir, cur->dir_gen, cur 4187 fs/btrfs/send.c ret = get_cur_path(sctx, cur->dir, cur 4188 fs/btrfs/send.c cur->dir_gen, valid_path); cur 4194 fs/btrfs/send.c last_dir_ino_rm = cur->dir; cur 321 fs/btrfs/tree-checker.c u32 cur = 0; cur 324 fs/btrfs/tree-checker.c while (cur < item_size) { cur 333 fs/btrfs/tree-checker.c if (cur + sizeof(*di) > item_size) { cur 336 fs/btrfs/tree-checker.c cur + sizeof(*di), item_size); cur 394 fs/btrfs/tree-checker.c if (cur + total_size > item_size) { cur 397 fs/btrfs/tree-checker.c cur + total_size, item_size); cur 419 fs/btrfs/tree-checker.c cur += total_size; cur 2348 fs/btrfs/tree-log.c u32 cur; cur 2358 fs/btrfs/tree-log.c cur = 0; cur 2359 fs/btrfs/tree-log.c while (cur < total_size) { cur 2400 fs/btrfs/tree-log.c cur += this_len; cur 2690 fs/btrfs/tree-log.c struct extent_buffer *cur; cur 2703 fs/btrfs/tree-log.c cur = path->nodes[*level]; cur 2705 fs/btrfs/tree-log.c WARN_ON(btrfs_header_level(cur) != *level); cur 2708 fs/btrfs/tree-log.c btrfs_header_nritems(cur)) cur 2711 fs/btrfs/tree-log.c bytenr = btrfs_node_blockptr(cur, path->slots[*level]); cur 2712 fs/btrfs/tree-log.c ptr_gen = btrfs_node_ptr_generation(cur, path->slots[*level]); cur 2713 fs/btrfs/tree-log.c btrfs_node_key_to_cpu(cur, &first_key, path->slots[*level]); cur 148 fs/btrfs/ulist.c struct ulist_node *cur = NULL; cur 152 fs/btrfs/ulist.c cur = rb_entry(parent, struct ulist_node, rb_node); cur 154 fs/btrfs/ulist.c if (cur->val < ins->val) cur 156 fs/btrfs/ulist.c else if (cur->val > ins->val) cur 536 fs/btrfs/volumes.c struct bio *cur; cur 616 fs/btrfs/volumes.c cur = pending; cur 618 fs/btrfs/volumes.c cur->bi_next = NULL; cur 620 fs/btrfs/volumes.c BUG_ON(atomic_read(&cur->__bi_cnt) == 0); cur 638 fs/btrfs/volumes.c btrfsic_submit_bio(cur); cur 2939 fs/btrfs/volumes.c u32 cur; cur 2946 fs/btrfs/volumes.c cur = 0; cur 2948 fs/btrfs/volumes.c while (cur < array_size) { cur 2964 fs/btrfs/volumes.c memmove(ptr, ptr + len, array_size - (cur + len)); cur 2969 fs/btrfs/volumes.c cur += len; cur 283 fs/btrfs/xattr.c u32 cur; cur 314 fs/btrfs/xattr.c cur = 0; cur 315 fs/btrfs/xattr.c while (cur < item_size) { cur 340 fs/btrfs/xattr.c cur += this_len; cur 454 fs/cifs/cifsfs.c cifs_show_nls(struct seq_file *s, struct nls_table *cur) cur 460 fs/cifs/cifsfs.c if (def != cur) cur 461 fs/cifs/cifsfs.c seq_printf(s, ",iocharset=%s", cur->charset); cur 270 fs/cifs/file.c struct cifs_fid_locks *cur; cur 274 fs/cifs/file.c list_for_each_entry(cur, &cinode->llist, llist) { cur 275 fs/cifs/file.c if (!list_empty(&cur->locks)) { cur 1012 fs/cifs/file.c struct cifs_fid_locks *cur; cur 1015 fs/cifs/file.c list_for_each_entry(cur, &cinode->llist, llist) { cur 1016 fs/cifs/file.c rc = cifs_find_fid_lock_conflict(cur, offset, length, type, cur 1193 fs/cifs/file.c LOCKING_ANDX_RANGE *buf, *cur; cur 1226 fs/cifs/file.c cur = buf; cur 1231 fs/cifs/file.c cur->Pid = cpu_to_le16(li->pid); cur 1232 fs/cifs/file.c cur->LengthLow = cpu_to_le32((u32)li->length); cur 1233 fs/cifs/file.c cur->LengthHigh = cpu_to_le32((u32)(li->length>>32)); cur 1234 fs/cifs/file.c cur->OffsetLow = cpu_to_le32((u32)li->offset); cur 1235 fs/cifs/file.c cur->OffsetHigh = cpu_to_le32((u32)(li->offset>>32)); cur 1243 fs/cifs/file.c cur = buf; cur 1246 fs/cifs/file.c cur++; cur 1538 fs/cifs/file.c LOCKING_ANDX_RANGE *buf, *cur; cur 1567 fs/cifs/file.c cur = buf; cur 1588 fs/cifs/file.c cur->Pid = cpu_to_le16(li->pid); cur 1589 fs/cifs/file.c cur->LengthLow = cpu_to_le32((u32)li->length); cur 1590 fs/cifs/file.c cur->LengthHigh = cpu_to_le32((u32)(li->length>>32)); cur 1591 fs/cifs/file.c cur->OffsetLow = cpu_to_le32((u32)li->offset); cur 1592 fs/cifs/file.c cur->OffsetHigh = cpu_to_le32((u32)(li->offset>>32)); cur 1618 fs/cifs/file.c cur = buf; cur 1621 fs/cifs/file.c cur++; cur 122 fs/cifs/smb2file.c struct smb2_lock_element *buf, *cur; cur 146 fs/cifs/smb2file.c cur = buf; cur 166 fs/cifs/smb2file.c cur->Length = cpu_to_le64(li->length); cur 167 fs/cifs/smb2file.c cur->Offset = cpu_to_le64(li->offset); cur 168 fs/cifs/smb2file.c cur->Flags = cpu_to_le32(SMB2_LOCKFLAG_UNLOCK); cur 194 fs/cifs/smb2file.c cur = buf; cur 197 fs/cifs/smb2file.c cur++; cur 223 fs/cifs/smb2file.c struct smb2_lock_element *cur = buf; cur 227 fs/cifs/smb2file.c cur->Length = cpu_to_le64(li->length); cur 228 fs/cifs/smb2file.c cur->Offset = cpu_to_le64(li->offset); cur 229 fs/cifs/smb2file.c cur->Flags = cpu_to_le32(li->type | cur 238 fs/cifs/smb2file.c cur = buf; cur 241 fs/cifs/smb2file.c cur++; cur 49 fs/configfs/symlink.c int cur = strlen(config_item_name(p)); cur 52 fs/configfs/symlink.c length -= cur; cur 53 fs/configfs/symlink.c memcpy(buffer + length, config_item_name(p), cur); cur 121 fs/coredump.c int cur = cn->used; cur 134 fs/coredump.c if ((cn->used - cur == 1 && cn->corename[cur] == '.') || cur 135 fs/coredump.c (cn->used - cur == 2 && cn->corename[cur] == '.' cur 136 fs/coredump.c && cn->corename[cur+1] == '.')) cur 137 fs/coredump.c cn->corename[cur] = '!'; cur 145 fs/coredump.c if (cn->used == cur) cur 149 fs/coredump.c for (; cur < cn->used; ++cur) { cur 150 fs/coredump.c if (cn->corename[cur] == '/') cur 151 fs/coredump.c cn->corename[cur] = '!'; cur 477 fs/dlm/lock.c struct dlm_rsb *cur = rb_entry(*newn, struct dlm_rsb, cur 481 fs/dlm/lock.c rc = rsb_cmp(cur, rsb->res_name, rsb->res_length); cur 489 fs/dlm/lock.c dlm_dump_rsb(cur); cur 204 fs/efs/inode.c int cur, last, first = 1; cur 229 fs/efs/inode.c cur = (last + dirext) % in->numextents; cur 230 fs/efs/inode.c if ((result = efs_extent_check(&in->extents[cur], block, sb))) { cur 231 fs/efs/inode.c in->lastextent = cur; cur 246 fs/efs/inode.c cur = (last + indext) % indexts; cur 256 fs/efs/inode.c for(dirext = 0; cur < ibase && dirext < direxts; dirext++) { cur 264 fs/efs/inode.c cur, block); cur 271 fs/efs/inode.c (cur - ibase) / cur 273 fs/efs/inode.c ioffset = (cur - ibase) % cur 297 fs/efs/inode.c cur, iblock); cur 304 fs/efs/inode.c in->lastextent = cur; cur 198 fs/erofs/decompressor.c const char *cur = dst - pageofs_out; cur 200 fs/erofs/decompressor.c while (cur < end) { cur 206 fs/erofs/decompressor.c if (cur >= dst) { cur 207 fs/erofs/decompressor.c memcpy(buf, cur, min_t(uint, PAGE_SIZE, cur 208 fs/erofs/decompressor.c end - cur)); cur 210 fs/erofs/decompressor.c memcpy(buf + pageofs_out, cur + pageofs_out, cur 211 fs/erofs/decompressor.c min_t(uint, righthalf, end - cur)); cur 215 fs/erofs/decompressor.c cur += PAGE_SIZE; cur 582 fs/erofs/zdata.c unsigned int cur, end, spiltted, index; cur 591 fs/erofs/zdata.c cur = end - 1; cur 594 fs/erofs/zdata.c if (offset + cur >= map->m_la && cur 595 fs/erofs/zdata.c offset + cur < map->m_la + map->m_llen) { cur 603 fs/erofs/zdata.c erofs_dbg("%s: [out-of-range] pos %llu", __func__, offset + cur); cur 608 fs/erofs/zdata.c map->m_la = offset + cur; cur 641 fs/erofs/zdata.c cur = end - min_t(unsigned int, offset + end - map->m_la, end); cur 643 fs/erofs/zdata.c zero_user_segment(page, cur, end); cur 648 fs/erofs/zdata.c page_type = cur ? Z_EROFS_VLE_PAGE_TYPE_HEAD : cur 653 fs/erofs/zdata.c if (cur) cur 682 fs/erofs/zdata.c map->m_llen = offset + cur - map->m_la; cur 684 fs/erofs/zdata.c end = cur; cur 785 fs/ext2/balloc.c ext2_fsblk_t cur; cur 790 fs/ext2/balloc.c cur = start_block; cur 796 fs/ext2/balloc.c if (cur <= rsv->rsv_end) cur 797 fs/ext2/balloc.c cur = rsv->rsv_end + 1; cur 808 fs/ext2/balloc.c if (cur > last_block) cur 822 fs/ext2/balloc.c if (cur + size <= rsv->rsv_start) { cur 851 fs/ext2/balloc.c my_rsv->rsv_start = cur; cur 852 fs/ext2/balloc.c my_rsv->rsv_end = cur + size - 1; cur 1197 fs/ext4/ext4.h extern void ext4_set_bits(void *bm, int cur, int len); cur 557 fs/ext4/mballoc.c struct list_head *cur; cur 629 fs/ext4/mballoc.c list_for_each(cur, &grp->bb_prealloc_list) { cur 632 fs/ext4/mballoc.c pa = list_entry(cur, struct ext4_prealloc_space, pa_group_list); cur 1276 fs/ext4/mballoc.c static void mb_clear_bits(void *bm, int cur, int len) cur 1280 fs/ext4/mballoc.c len = cur + len; cur 1281 fs/ext4/mballoc.c while (cur < len) { cur 1282 fs/ext4/mballoc.c if ((cur & 31) == 0 && (len - cur) >= 32) { cur 1284 fs/ext4/mballoc.c addr = bm + (cur >> 3); cur 1286 fs/ext4/mballoc.c cur += 32; cur 1289 fs/ext4/mballoc.c mb_clear_bit(cur, bm); cur 1290 fs/ext4/mballoc.c cur++; cur 1297 fs/ext4/mballoc.c static int mb_test_and_clear_bits(void *bm, int cur, int len) cur 1302 fs/ext4/mballoc.c len = cur + len; cur 1303 fs/ext4/mballoc.c while (cur < len) { cur 1304 fs/ext4/mballoc.c if ((cur & 31) == 0 && (len - cur) >= 32) { cur 1306 fs/ext4/mballoc.c addr = bm + (cur >> 3); cur 1308 fs/ext4/mballoc.c zero_bit = cur + mb_find_next_zero_bit(addr, 32, 0); cur 1310 fs/ext4/mballoc.c cur += 32; cur 1313 fs/ext4/mballoc.c if (!mb_test_and_clear_bit(cur, bm) && zero_bit == -1) cur 1314 fs/ext4/mballoc.c zero_bit = cur; cur 1315 fs/ext4/mballoc.c cur++; cur 1321 fs/ext4/mballoc.c void ext4_set_bits(void *bm, int cur, int len) cur 1325 fs/ext4/mballoc.c len = cur + len; cur 1326 fs/ext4/mballoc.c while (cur < len) { cur 1327 fs/ext4/mballoc.c if ((cur & 31) == 0 && (len - cur) >= 32) { cur 1329 fs/ext4/mballoc.c addr = bm + (cur >> 3); cur 1331 fs/ext4/mballoc.c cur += 32; cur 1334 fs/ext4/mballoc.c mb_set_bit(cur, bm); cur 1335 fs/ext4/mballoc.c cur++; cur 1561 fs/ext4/mballoc.c int cur; cur 1616 fs/ext4/mballoc.c cur = (start >> ord) & ~1U; cur 1618 fs/ext4/mballoc.c mb_clear_bit(cur, buddy); cur 1619 fs/ext4/mballoc.c mb_clear_bit(cur + 1, buddy); cur 2701 fs/ext4/mballoc.c struct list_head *cur, *tmp; cur 2704 fs/ext4/mballoc.c list_for_each_safe(cur, tmp, &grp->bb_prealloc_list) { cur 2705 fs/ext4/mballoc.c pa = list_entry(cur, struct ext4_prealloc_space, pa_group_list); cur 3542 fs/ext4/mballoc.c struct list_head *cur; cur 3556 fs/ext4/mballoc.c list_for_each(cur, &grp->bb_prealloc_list) { cur 3557 fs/ext4/mballoc.c pa = list_entry(cur, struct ext4_prealloc_space, pa_group_list); cur 4151 fs/ext4/mballoc.c struct list_head *cur; cur 4153 fs/ext4/mballoc.c list_for_each(cur, &grp->bb_prealloc_list) { cur 4154 fs/ext4/mballoc.c pa = list_entry(cur, struct ext4_prealloc_space, cur 168 fs/ext4/page-io.c struct list_head *cur, *before, *after; cur 176 fs/ext4/page-io.c cur = &io->list; cur 177 fs/ext4/page-io.c before = cur->prev; cur 179 fs/ext4/page-io.c after = cur->next; cur 3137 fs/ext4/super.c unsigned long next_wakeup, cur; cur 3194 fs/ext4/super.c cur = jiffies; cur 3195 fs/ext4/super.c if ((time_after_eq(cur, next_wakeup)) || cur 3201 fs/ext4/super.c schedule_timeout_interruptible(next_wakeup - cur); cur 3006 fs/f2fs/data.c struct inmem_pages *cur; cur 3007 fs/f2fs/data.c list_for_each_entry(cur, &fi->inmem_pages, list) cur 3008 fs/f2fs/data.c if (cur->page == page) { cur 3009 fs/f2fs/data.c cur->page = newpage; cur 172 fs/f2fs/extent_cache.c struct rb_node *cur = rb_first_cached(root), *next; cur 175 fs/f2fs/extent_cache.c if (!cur) cur 178 fs/f2fs/extent_cache.c while (cur) { cur 179 fs/f2fs/extent_cache.c next = rb_next(cur); cur 183 fs/f2fs/extent_cache.c cur_re = rb_entry(cur, struct rb_entry, rb_node); cur 193 fs/f2fs/extent_cache.c cur = next; cur 749 fs/f2fs/f2fs.h static inline bool __is_discard_back_mergeable(struct discard_info *cur, cur 752 fs/f2fs/f2fs.h return __is_discard_mergeable(back, cur, max_len); cur 755 fs/f2fs/f2fs.h static inline bool __is_discard_front_mergeable(struct discard_info *cur, cur 758 fs/f2fs/f2fs.h return __is_discard_mergeable(cur, front, max_len); cur 768 fs/f2fs/f2fs.h static inline bool __is_back_mergeable(struct extent_info *cur, cur 771 fs/f2fs/f2fs.h return __is_extent_mergeable(back, cur); cur 774 fs/f2fs/f2fs.h static inline bool __is_front_mergeable(struct extent_info *cur, cur 777 fs/f2fs/f2fs.h return __is_extent_mergeable(cur, front); cur 3381 fs/f2fs/f2fs.h int cur = atomic_read(&F2FS_I_SB(inode)->aw_cnt); \ cur 3383 fs/f2fs/f2fs.h if (cur > max) \ cur 3384 fs/f2fs/f2fs.h atomic_set(&F2FS_I_SB(inode)->max_aw_cnt, cur); \ cur 3392 fs/f2fs/f2fs.h int cur = atomic_read(&F2FS_I_SB(inode)->vw_cnt); \ cur 3394 fs/f2fs/f2fs.h if (cur > max) \ cur 3395 fs/f2fs/f2fs.h atomic_set(&F2FS_I_SB(inode)->max_vw_cnt, cur); \ cur 2742 fs/f2fs/node.c struct nat_entry_set *cur; cur 2747 fs/f2fs/node.c list_for_each_entry(cur, head, set_list) { cur 2748 fs/f2fs/node.c if (cur->entry_cnt >= nes->entry_cnt) { cur 2749 fs/f2fs/node.c list_add(&nes->set_list, cur->set_list.prev); cur 2798 fs/f2fs/node.c struct nat_entry *ne, *cur; cur 2822 fs/f2fs/node.c list_for_each_entry_safe(ne, cur, &set->entry_list, list) { cur 215 fs/f2fs/segment.c struct inmem_pages *cur, *tmp; cur 218 fs/f2fs/segment.c list_for_each_entry_safe(cur, tmp, head, list) { cur 219 fs/f2fs/segment.c struct page *page = cur->page; cur 262 fs/f2fs/segment.c if (cur->old_addr == NEW_ADDR) { cur 267 fs/f2fs/segment.c cur->old_addr, ni.version, true, true); cur 279 fs/f2fs/segment.c list_del(&cur->list); cur 280 fs/f2fs/segment.c kmem_cache_free(inmem_entry_slab, cur); cur 354 fs/f2fs/segment.c struct inmem_pages *cur = NULL; cur 359 fs/f2fs/segment.c list_for_each_entry(cur, head, list) { cur 360 fs/f2fs/segment.c if (cur->page == page) cur 364 fs/f2fs/segment.c f2fs_bug_on(sbi, list_empty(head) || cur->page != page); cur 365 fs/f2fs/segment.c list_del(&cur->list); cur 369 fs/f2fs/segment.c kmem_cache_free(inmem_entry_slab, cur); cur 382 fs/f2fs/segment.c struct inmem_pages *cur, *tmp; cur 397 fs/f2fs/segment.c list_for_each_entry_safe(cur, tmp, &fi->inmem_pages, list) { cur 398 fs/f2fs/segment.c struct page *page = cur->page; cur 427 fs/f2fs/segment.c cur->old_addr = fio.old_blkaddr; cur 431 fs/f2fs/segment.c list_move_tail(&cur->list, &revoke_list); cur 640 fs/gfs2/dir.c struct gfs2_dirent *cur = *dent, *tmp; cur 644 fs/gfs2/dir.c ret = dirent_check_reclen(dip, cur, bh_end); cur 648 fs/gfs2/dir.c tmp = (void *)cur + ret; cur 673 fs/gfs2/dir.c struct gfs2_dirent *prev, struct gfs2_dirent *cur) cur 677 fs/gfs2/dir.c if (gfs2_dirent_sentinel(cur)) { cur 689 fs/gfs2/dir.c cur->de_inum.no_addr = 0; cur 690 fs/gfs2/dir.c cur->de_inum.no_formal_ino = 0; cur 697 fs/gfs2/dir.c cur_rec_len = be16_to_cpu(cur->de_rec_len); cur 699 fs/gfs2/dir.c if ((char *)prev + prev_rec_len != (char *)cur) cur 701 fs/gfs2/dir.c if ((char *)cur + cur_rec_len > bh->b_data + bh->b_size) cur 511 fs/gfs2/rgrp.c struct gfs2_rgrpd *cur; cur 516 fs/gfs2/rgrp.c cur = rb_entry(n, struct gfs2_rgrpd, rd_node); cur 518 fs/gfs2/rgrp.c if (blk < cur->rd_addr) cur 520 fs/gfs2/rgrp.c else if (blk >= cur->rd_data0 + cur->rd_data) cur 525 fs/gfs2/rgrp.c if (blk < cur->rd_addr) cur 527 fs/gfs2/rgrp.c if (blk >= cur->rd_data0 + cur->rd_data) cur 530 fs/gfs2/rgrp.c return cur; cur 872 fs/gfs2/rgrp.c struct gfs2_rgrpd *cur = rb_entry(*newn, struct gfs2_rgrpd, cur 876 fs/gfs2/rgrp.c if (rgd->rd_addr < cur->rd_addr) cur 878 fs/gfs2/rgrp.c else if (rgd->rd_addr > cur->rd_addr) cur 1502 fs/gfs2/rgrp.c struct gfs2_blkreserv *cur = cur 1506 fs/gfs2/rgrp.c rc = rs_cmp(fsblock, rs->rs_free, cur); cur 242 fs/nfsd/nfs4state.c struct nfsd4_blocked_lock *cur, *found = NULL; cur 245 fs/nfsd/nfs4state.c list_for_each_entry(cur, &lo->lo_blocked, nbl_list) { cur 246 fs/nfsd/nfs4state.c if (fh_match(fh, &cur->nbl_fh)) { cur 247 fs/nfsd/nfs4state.c list_del_init(&cur->nbl_list); cur 248 fs/nfsd/nfs4state.c list_del_init(&cur->nbl_lru); cur 249 fs/nfsd/nfs4state.c found = cur; cur 2148 fs/nfsd/nfs4xdr.c struct path cur = *path; cur 2156 fs/nfsd/nfs4xdr.c path_get(&cur); cur 2161 fs/nfsd/nfs4xdr.c if (path_equal(&cur, root)) cur 2163 fs/nfsd/nfs4xdr.c if (cur.dentry == cur.mnt->mnt_root) { cur 2164 fs/nfsd/nfs4xdr.c if (follow_up(&cur)) cur 2177 fs/nfsd/nfs4xdr.c components[ncomponents++] = cur.dentry; cur 2178 fs/nfsd/nfs4xdr.c cur.dentry = dget_parent(cur.dentry); cur 2210 fs/nfsd/nfs4xdr.c path_put(&cur); cur 883 fs/pipe.c int cur = *cnt; cur 885 fs/pipe.c while (cur == *cnt) { cur 890 fs/pipe.c return cur == *cnt ? -ERESTARTSYS : 0; cur 1397 fs/reiserfs/bitmap.c unsigned long *cur = (unsigned long *)(bh->b_data + bh->b_size); cur 1406 fs/reiserfs/bitmap.c while (--cur >= (unsigned long *)bh->b_data) { cur 1408 fs/reiserfs/bitmap.c if (*cur == 0) cur 1410 fs/reiserfs/bitmap.c else if (*cur != ~0L) /* A mix, investigate */ cur 1411 fs/reiserfs/bitmap.c info->free_count += BITS_PER_LONG - hweight_long(*cur); cur 131 fs/reiserfs/ibalance.c struct buffer_head *cur = cur_bi->bi_bh; cur 142 fs/reiserfs/ibalance.c blkh = B_BLK_HEAD(cur); cur 146 fs/reiserfs/ibalance.c RFALSE(B_FREE_SPACE(cur) < count * (KEY_SIZE + DC_SIZE), cur 148 fs/reiserfs/ibalance.c B_FREE_SPACE(cur), count * (KEY_SIZE + DC_SIZE)); cur 151 fs/reiserfs/ibalance.c dc = B_N_CHILD(cur, to + 1); cur 164 fs/reiserfs/ibalance.c ih = internal_key(cur, ((to == -1) ? 0 : to)); cur 180 fs/reiserfs/ibalance.c do_balance_mark_internal_dirty(cur_bi->tb, cur, 0); cur 183 fs/reiserfs/ibalance.c check_internal(cur); cur 209 fs/reiserfs/ibalance.c struct buffer_head *cur = cur_bi->bi_bh; cur 215 fs/reiserfs/ibalance.c RFALSE(cur == NULL, "buffer is 0"); cur 218 fs/reiserfs/ibalance.c RFALSE(first_p < 0 || first_p + del_num > B_NR_ITEMS(cur) + 1 cur 223 fs/reiserfs/ibalance.c B_NR_ITEMS(cur) + 1, first_i); cur 227 fs/reiserfs/ibalance.c blkh = B_BLK_HEAD(cur); cur 237 fs/reiserfs/ibalance.c RFALSE(first_i + del_num > B_NR_ITEMS(cur), cur 240 fs/reiserfs/ibalance.c first_i, del_num, first_i + del_num, cur, cur); cur 243 fs/reiserfs/ibalance.c dc = B_N_CHILD(cur, first_p); cur 246 fs/reiserfs/ibalance.c key = internal_key(cur, first_i); cur 257 fs/reiserfs/ibalance.c do_balance_mark_internal_dirty(cur_bi->tb, cur, 0); cur 259 fs/reiserfs/ibalance.c check_internal(cur); cur 1845 fs/reiserfs/journal.c struct reiserfs_journal_cnode *cur; cur 1852 fs/reiserfs/journal.c cur = *head; cur 1853 fs/reiserfs/journal.c while (cur) { cur 1854 fs/reiserfs/journal.c if (cur->blocknr == block && cur->sb == sb cur 1855 fs/reiserfs/journal.c && (jl == NULL || jl == cur->jlist) cur 1856 fs/reiserfs/journal.c && (!test_bit(BLOCK_FREED, &cur->state) || remove_freed)) { cur 1857 fs/reiserfs/journal.c if (cur->hnext) { cur 1858 fs/reiserfs/journal.c cur->hnext->hprev = cur->hprev; cur 1860 fs/reiserfs/journal.c if (cur->hprev) { cur 1861 fs/reiserfs/journal.c cur->hprev->hnext = cur->hnext; cur 1863 fs/reiserfs/journal.c *head = cur->hnext; cur 1865 fs/reiserfs/journal.c cur->blocknr = 0; cur 1866 fs/reiserfs/journal.c cur->sb = NULL; cur 1867 fs/reiserfs/journal.c cur->state = 0; cur 1872 fs/reiserfs/journal.c if (cur->bh && cur->jlist) cur 1873 fs/reiserfs/journal.c atomic_dec(&cur->jlist->j_nonzerolen); cur 1874 fs/reiserfs/journal.c cur->bh = NULL; cur 1875 fs/reiserfs/journal.c cur->jlist = NULL; cur 1877 fs/reiserfs/journal.c cur = cur->hnext; cur 3476 fs/reiserfs/journal.c struct reiserfs_journal_cnode *cur = cn->hprev; cur 3484 fs/reiserfs/journal.c while (cur && can_dirty) { cur 3485 fs/reiserfs/journal.c if (cur->jlist && cur->bh && cur->blocknr && cur->sb == sb && cur 3486 fs/reiserfs/journal.c cur->blocknr == blocknr) { cur 3489 fs/reiserfs/journal.c cur = cur->hprev; cur 3495 fs/reiserfs/journal.c cur = cn->hnext; cur 3496 fs/reiserfs/journal.c while (cur && can_dirty) { cur 3497 fs/reiserfs/journal.c if (cur->jlist && cur->jlist->j_len > 0 && cur 3498 fs/reiserfs/journal.c atomic_read(&cur->jlist->j_commit_left) > 0 && cur->bh && cur 3499 fs/reiserfs/journal.c cur->blocknr && cur->sb == sb && cur->blocknr == blocknr) { cur 3502 fs/reiserfs/journal.c cur = cur->hnext; cur 2063 fs/reiserfs/reiserfs.h #define B_CHILD_SIZE(cur) (MAX_CHILD_SIZE(cur)-(B_FREE_SPACE(cur))) cur 979 fs/reiserfs/super.c static int reiserfs_getopt(struct super_block *s, char **cur, opt_desc_t * opts, cur 993 fs/reiserfs/super.c p = *cur; cur 996 fs/reiserfs/super.c *cur = strchr(p, ','); cur 997 fs/reiserfs/super.c if (*cur) { cur 998 fs/reiserfs/super.c *(*cur) = '\0'; cur 999 fs/reiserfs/super.c (*cur)++; cur 2300 fs/ubifs/debug.c struct list_head *cur; cur 2306 fs/ubifs/debug.c for (cur = head->next; cur->next != head; cur = cur->next) { cur 2311 fs/ubifs/debug.c sa = container_of(cur, struct ubifs_scan_node, list); cur 2312 fs/ubifs/debug.c sb = container_of(cur->next, struct ubifs_scan_node, list); cur 2367 fs/ubifs/debug.c struct list_head *cur; cur 2373 fs/ubifs/debug.c for (cur = head->next; cur->next != head; cur = cur->next) { cur 2378 fs/ubifs/debug.c sa = container_of(cur, struct ubifs_scan_node, list); cur 2379 fs/ubifs/debug.c sb = container_of(cur->next, struct ubifs_scan_node, list); cur 108 fs/unicode/utf8-core.c struct utf8cursor cur; cur 111 fs/unicode/utf8-core.c if (utf8ncursor(&cur, data, str->name, str->len) < 0) cur 115 fs/unicode/utf8-core.c int c = utf8byte(&cur); cur 132 fs/unicode/utf8-core.c struct utf8cursor cur; cur 135 fs/unicode/utf8-core.c if (utf8ncursor(&cur, data, str->name, str->len) < 0) cur 139 fs/unicode/utf8-core.c int c = utf8byte(&cur); cur 1306 fs/userfaultfd.c struct vm_area_struct *vma, *prev, *cur; cur 1381 fs/userfaultfd.c for (cur = vma; cur && cur->vm_start < end; cur = cur->vm_next) { cur 1384 fs/userfaultfd.c BUG_ON(!!cur->vm_userfaultfd_ctx.ctx ^ cur 1385 fs/userfaultfd.c !!(cur->vm_flags & (VM_UFFD_MISSING | VM_UFFD_WP))); cur 1389 fs/userfaultfd.c if (!vma_can_userfault(cur)) cur 1401 fs/userfaultfd.c if (unlikely(!(cur->vm_flags & VM_MAYWRITE))) cur 1408 fs/userfaultfd.c if (is_vm_hugetlb_page(cur) && end <= cur->vm_end && cur 1409 fs/userfaultfd.c end > cur->vm_start) { cur 1410 fs/userfaultfd.c unsigned long vma_hpagesize = vma_kernel_pagesize(cur); cur 1425 fs/userfaultfd.c if (cur->vm_userfaultfd_ctx.ctx && cur 1426 fs/userfaultfd.c cur->vm_userfaultfd_ctx.ctx != ctx) cur 1432 fs/userfaultfd.c if (is_vm_hugetlb_page(cur)) cur 1518 fs/userfaultfd.c struct vm_area_struct *vma, *prev, *cur; cur 1570 fs/userfaultfd.c for (cur = vma; cur && cur->vm_start < end; cur = cur->vm_next) { cur 1573 fs/userfaultfd.c BUG_ON(!!cur->vm_userfaultfd_ctx.ctx ^ cur 1574 fs/userfaultfd.c !!(cur->vm_flags & (VM_UFFD_MISSING | VM_UFFD_WP))); cur 1583 fs/userfaultfd.c if (!vma_can_userfault(cur)) cur 144 fs/xfs/libxfs/xfs_alloc.c struct xfs_btree_cur *cur, /* btree cursor */ cur 149 fs/xfs/libxfs/xfs_alloc.c cur->bc_rec.a.ar_startblock = bno; cur 150 fs/xfs/libxfs/xfs_alloc.c cur->bc_rec.a.ar_blockcount = len; cur 151 fs/xfs/libxfs/xfs_alloc.c return xfs_btree_lookup(cur, XFS_LOOKUP_EQ, stat); cur 160 fs/xfs/libxfs/xfs_alloc.c struct xfs_btree_cur *cur, /* btree cursor */ cur 165 fs/xfs/libxfs/xfs_alloc.c cur->bc_rec.a.ar_startblock = bno; cur 166 fs/xfs/libxfs/xfs_alloc.c cur->bc_rec.a.ar_blockcount = len; cur 167 fs/xfs/libxfs/xfs_alloc.c return xfs_btree_lookup(cur, XFS_LOOKUP_GE, stat); cur 176 fs/xfs/libxfs/xfs_alloc.c struct xfs_btree_cur *cur, /* btree cursor */ cur 181 fs/xfs/libxfs/xfs_alloc.c cur->bc_rec.a.ar_startblock = bno; cur 182 fs/xfs/libxfs/xfs_alloc.c cur->bc_rec.a.ar_blockcount = len; cur 183 fs/xfs/libxfs/xfs_alloc.c return xfs_btree_lookup(cur, XFS_LOOKUP_LE, stat); cur 193 fs/xfs/libxfs/xfs_alloc.c struct xfs_btree_cur *cur, /* btree cursor */ cur 201 fs/xfs/libxfs/xfs_alloc.c return xfs_btree_update(cur, &rec); cur 209 fs/xfs/libxfs/xfs_alloc.c struct xfs_btree_cur *cur, /* btree cursor */ cur 214 fs/xfs/libxfs/xfs_alloc.c struct xfs_mount *mp = cur->bc_mp; cur 215 fs/xfs/libxfs/xfs_alloc.c xfs_agnumber_t agno = cur->bc_private.a.agno; cur 219 fs/xfs/libxfs/xfs_alloc.c error = xfs_btree_get_rec(cur, &rec, stat); cur 242 fs/xfs/libxfs/xfs_alloc.c cur->bc_btnum == XFS_BTNUM_BNO ? "Block" : "Size", agno); cur 3070 fs/xfs/libxfs/xfs_alloc.c struct xfs_btree_cur *cur, cur 3079 fs/xfs/libxfs/xfs_alloc.c return query->fn(cur, &irec, query->priv); cur 3085 fs/xfs/libxfs/xfs_alloc.c struct xfs_btree_cur *cur, cur 3095 fs/xfs/libxfs/xfs_alloc.c ASSERT(cur->bc_btnum == XFS_BTNUM_BNO); cur 3100 fs/xfs/libxfs/xfs_alloc.c return xfs_btree_query_range(cur, &low_brec, &high_brec, cur 3107 fs/xfs/libxfs/xfs_alloc.c struct xfs_btree_cur *cur, cur 3113 fs/xfs/libxfs/xfs_alloc.c ASSERT(cur->bc_btnum == XFS_BTNUM_BNO); cur 3116 fs/xfs/libxfs/xfs_alloc.c return xfs_btree_query_all(cur, xfs_alloc_query_range_helper, &query); cur 3122 fs/xfs/libxfs/xfs_alloc.c struct xfs_btree_cur *cur, cur 3135 fs/xfs/libxfs/xfs_alloc.c return xfs_btree_has_record(cur, &low, &high, exists); cur 202 fs/xfs/libxfs/xfs_alloc.h struct xfs_btree_cur *cur, /* btree cursor */ cur 209 fs/xfs/libxfs/xfs_alloc.h struct xfs_btree_cur *cur, /* btree cursor */ cur 216 fs/xfs/libxfs/xfs_alloc.h struct xfs_btree_cur *cur, /* btree cursor */ cur 234 fs/xfs/libxfs/xfs_alloc.h struct xfs_btree_cur *cur, cur 238 fs/xfs/libxfs/xfs_alloc.h int xfs_alloc_query_range(struct xfs_btree_cur *cur, cur 242 fs/xfs/libxfs/xfs_alloc.h int xfs_alloc_query_all(struct xfs_btree_cur *cur, xfs_alloc_query_range_fn fn, cur 245 fs/xfs/libxfs/xfs_alloc.h int xfs_alloc_has_record(struct xfs_btree_cur *cur, xfs_agblock_t bno, cur 25 fs/xfs/libxfs/xfs_alloc_btree.c struct xfs_btree_cur *cur) cur 27 fs/xfs/libxfs/xfs_alloc_btree.c return xfs_allocbt_init_cursor(cur->bc_mp, cur->bc_tp, cur 28 fs/xfs/libxfs/xfs_alloc_btree.c cur->bc_private.a.agbp, cur->bc_private.a.agno, cur 29 fs/xfs/libxfs/xfs_alloc_btree.c cur->bc_btnum); cur 34 fs/xfs/libxfs/xfs_alloc_btree.c struct xfs_btree_cur *cur, cur 38 fs/xfs/libxfs/xfs_alloc_btree.c struct xfs_buf *agbp = cur->bc_private.a.agbp; cur 41 fs/xfs/libxfs/xfs_alloc_btree.c int btnum = cur->bc_btnum; cur 42 fs/xfs/libxfs/xfs_alloc_btree.c struct xfs_perag *pag = xfs_perag_get(cur->bc_mp, seqno); cur 51 fs/xfs/libxfs/xfs_alloc_btree.c xfs_alloc_log_agf(cur->bc_tp, agbp, XFS_AGF_ROOTS | XFS_AGF_LEVELS); cur 56 fs/xfs/libxfs/xfs_alloc_btree.c struct xfs_btree_cur *cur, cur 65 fs/xfs/libxfs/xfs_alloc_btree.c error = xfs_alloc_get_freelist(cur->bc_tp, cur->bc_private.a.agbp, cur 75 fs/xfs/libxfs/xfs_alloc_btree.c xfs_extent_busy_reuse(cur->bc_mp, cur->bc_private.a.agno, bno, 1, false); cur 77 fs/xfs/libxfs/xfs_alloc_btree.c xfs_trans_agbtree_delta(cur->bc_tp, 1); cur 86 fs/xfs/libxfs/xfs_alloc_btree.c struct xfs_btree_cur *cur, cur 89 fs/xfs/libxfs/xfs_alloc_btree.c struct xfs_buf *agbp = cur->bc_private.a.agbp; cur 94 fs/xfs/libxfs/xfs_alloc_btree.c bno = xfs_daddr_to_agbno(cur->bc_mp, XFS_BUF_ADDR(bp)); cur 95 fs/xfs/libxfs/xfs_alloc_btree.c error = xfs_alloc_put_freelist(cur->bc_tp, agbp, NULL, bno, 1); cur 99 fs/xfs/libxfs/xfs_alloc_btree.c xfs_extent_busy_insert(cur->bc_tp, be32_to_cpu(agf->agf_seqno), bno, 1, cur 101 fs/xfs/libxfs/xfs_alloc_btree.c xfs_trans_agbtree_delta(cur->bc_tp, -1); cur 110 fs/xfs/libxfs/xfs_alloc_btree.c struct xfs_btree_cur *cur, cur 116 fs/xfs/libxfs/xfs_alloc_btree.c struct xfs_agf *agf = XFS_BUF_TO_AGF(cur->bc_private.a.agbp); cur 122 fs/xfs/libxfs/xfs_alloc_btree.c ASSERT(cur->bc_btnum == XFS_BTNUM_CNT); cur 149 fs/xfs/libxfs/xfs_alloc_btree.c rrp = XFS_ALLOC_REC_ADDR(cur->bc_mp, block, numrecs); cur 162 fs/xfs/libxfs/xfs_alloc_btree.c pag = xfs_perag_get(cur->bc_mp, seqno); cur 165 fs/xfs/libxfs/xfs_alloc_btree.c xfs_alloc_log_agf(cur->bc_tp, cur->bc_private.a.agbp, XFS_AGF_LONGEST); cur 170 fs/xfs/libxfs/xfs_alloc_btree.c struct xfs_btree_cur *cur, cur 173 fs/xfs/libxfs/xfs_alloc_btree.c return cur->bc_mp->m_alloc_mnr[level != 0]; cur 178 fs/xfs/libxfs/xfs_alloc_btree.c struct xfs_btree_cur *cur, cur 181 fs/xfs/libxfs/xfs_alloc_btree.c return cur->bc_mp->m_alloc_mxr[level != 0]; cur 217 fs/xfs/libxfs/xfs_alloc_btree.c struct xfs_btree_cur *cur, cur 220 fs/xfs/libxfs/xfs_alloc_btree.c rec->alloc.ar_startblock = cpu_to_be32(cur->bc_rec.a.ar_startblock); cur 221 fs/xfs/libxfs/xfs_alloc_btree.c rec->alloc.ar_blockcount = cpu_to_be32(cur->bc_rec.a.ar_blockcount); cur 226 fs/xfs/libxfs/xfs_alloc_btree.c struct xfs_btree_cur *cur, cur 229 fs/xfs/libxfs/xfs_alloc_btree.c struct xfs_agf *agf = XFS_BUF_TO_AGF(cur->bc_private.a.agbp); cur 231 fs/xfs/libxfs/xfs_alloc_btree.c ASSERT(cur->bc_private.a.agno == be32_to_cpu(agf->agf_seqno)); cur 233 fs/xfs/libxfs/xfs_alloc_btree.c ptr->s = agf->agf_roots[cur->bc_btnum]; cur 238 fs/xfs/libxfs/xfs_alloc_btree.c struct xfs_btree_cur *cur, cur 241 fs/xfs/libxfs/xfs_alloc_btree.c xfs_alloc_rec_incore_t *rec = &cur->bc_rec.a; cur 249 fs/xfs/libxfs/xfs_alloc_btree.c struct xfs_btree_cur *cur, cur 252 fs/xfs/libxfs/xfs_alloc_btree.c xfs_alloc_rec_incore_t *rec = &cur->bc_rec.a; cur 265 fs/xfs/libxfs/xfs_alloc_btree.c struct xfs_btree_cur *cur, cur 275 fs/xfs/libxfs/xfs_alloc_btree.c struct xfs_btree_cur *cur, cur 385 fs/xfs/libxfs/xfs_alloc_btree.c struct xfs_btree_cur *cur, cur 395 fs/xfs/libxfs/xfs_alloc_btree.c struct xfs_btree_cur *cur, cur 406 fs/xfs/libxfs/xfs_alloc_btree.c struct xfs_btree_cur *cur, cur 419 fs/xfs/libxfs/xfs_alloc_btree.c struct xfs_btree_cur *cur, cur 486 fs/xfs/libxfs/xfs_alloc_btree.c struct xfs_btree_cur *cur; cur 490 fs/xfs/libxfs/xfs_alloc_btree.c cur = kmem_zone_zalloc(xfs_btree_cur_zone, KM_NOFS); cur 492 fs/xfs/libxfs/xfs_alloc_btree.c cur->bc_tp = tp; cur 493 fs/xfs/libxfs/xfs_alloc_btree.c cur->bc_mp = mp; cur 494 fs/xfs/libxfs/xfs_alloc_btree.c cur->bc_btnum = btnum; cur 495 fs/xfs/libxfs/xfs_alloc_btree.c cur->bc_blocklog = mp->m_sb.sb_blocklog; cur 498 fs/xfs/libxfs/xfs_alloc_btree.c cur->bc_statoff = XFS_STATS_CALC_INDEX(xs_abtc_2); cur 499 fs/xfs/libxfs/xfs_alloc_btree.c cur->bc_ops = &xfs_cntbt_ops; cur 500 fs/xfs/libxfs/xfs_alloc_btree.c cur->bc_nlevels = be32_to_cpu(agf->agf_levels[XFS_BTNUM_CNT]); cur 501 fs/xfs/libxfs/xfs_alloc_btree.c cur->bc_flags = XFS_BTREE_LASTREC_UPDATE; cur 503 fs/xfs/libxfs/xfs_alloc_btree.c cur->bc_statoff = XFS_STATS_CALC_INDEX(xs_abtb_2); cur 504 fs/xfs/libxfs/xfs_alloc_btree.c cur->bc_ops = &xfs_bnobt_ops; cur 505 fs/xfs/libxfs/xfs_alloc_btree.c cur->bc_nlevels = be32_to_cpu(agf->agf_levels[XFS_BTNUM_BNO]); cur 508 fs/xfs/libxfs/xfs_alloc_btree.c cur->bc_private.a.agbp = agbp; cur 509 fs/xfs/libxfs/xfs_alloc_btree.c cur->bc_private.a.agno = agno; cur 512 fs/xfs/libxfs/xfs_alloc_btree.c cur->bc_flags |= XFS_BTREE_CRC_BLOCKS; cur 514 fs/xfs/libxfs/xfs_alloc_btree.c return cur; cur 98 fs/xfs/libxfs/xfs_bmap.c struct xfs_btree_cur *cur, cur 102 fs/xfs/libxfs/xfs_bmap.c cur->bc_rec.b = *irec; cur 103 fs/xfs/libxfs/xfs_bmap.c return xfs_btree_lookup(cur, XFS_LOOKUP_EQ, stat); cur 108 fs/xfs/libxfs/xfs_bmap.c struct xfs_btree_cur *cur, cur 111 fs/xfs/libxfs/xfs_bmap.c cur->bc_rec.b.br_startoff = 0; cur 112 fs/xfs/libxfs/xfs_bmap.c cur->bc_rec.b.br_startblock = 0; cur 113 fs/xfs/libxfs/xfs_bmap.c cur->bc_rec.b.br_blockcount = 0; cur 114 fs/xfs/libxfs/xfs_bmap.c return xfs_btree_lookup(cur, XFS_LOOKUP_GE, stat); cur 145 fs/xfs/libxfs/xfs_bmap.c struct xfs_btree_cur *cur, cur 151 fs/xfs/libxfs/xfs_bmap.c return xfs_btree_update(cur, &rec); cur 229 fs/xfs/libxfs/xfs_bmap.c struct xfs_btree_cur *cur, cur 235 fs/xfs/libxfs/xfs_bmap.c if (!cur) cur 239 fs/xfs/libxfs/xfs_bmap.c if (!cur->bc_bufs[i]) cur 241 fs/xfs/libxfs/xfs_bmap.c if (XFS_BUF_ADDR(cur->bc_bufs[i]) == bno) cur 242 fs/xfs/libxfs/xfs_bmap.c return cur->bc_bufs[i]; cur 246 fs/xfs/libxfs/xfs_bmap.c list_for_each_entry(lip, &cur->bc_tp->t_items, li_trans) { cur 315 fs/xfs/libxfs/xfs_bmap.c xfs_btree_cur_t *cur, /* btree cursor or null */ cur 365 fs/xfs/libxfs/xfs_bmap.c bp = xfs_bmap_get_bp(cur, XFS_FSB_TO_DADDR(mp, bno)); cur 449 fs/xfs/libxfs/xfs_bmap.c bp = xfs_bmap_get_bp(cur, XFS_FSB_TO_DADDR(mp, bno)); cur 517 fs/xfs/libxfs/xfs_bmap.c #define xfs_bmap_check_leaf_extents(cur, ip, whichfork) do { } while (0) cur 585 fs/xfs/libxfs/xfs_bmap.c struct xfs_btree_cur *cur, /* btree cursor */ cur 603 fs/xfs/libxfs/xfs_bmap.c ASSERT(cur); cur 614 fs/xfs/libxfs/xfs_bmap.c XFS_WANT_CORRUPTED_RETURN(cur->bc_mp, cur 615 fs/xfs/libxfs/xfs_bmap.c xfs_btree_check_lptr(cur, cbno, 1)); cur 622 fs/xfs/libxfs/xfs_bmap.c if ((error = xfs_btree_check_block(cur, cblock, 0, cbp))) cur 625 fs/xfs/libxfs/xfs_bmap.c xfs_bmap_add_free(cur->bc_tp, cbno, 1, &oinfo); cur 629 fs/xfs/libxfs/xfs_bmap.c if (cur->bc_bufs[0] == cbp) cur 630 fs/xfs/libxfs/xfs_bmap.c cur->bc_bufs[0] = NULL; cur 657 fs/xfs/libxfs/xfs_bmap.c struct xfs_btree_cur *cur; /* bmap btree cursor */ cur 689 fs/xfs/libxfs/xfs_bmap.c cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork); cur 690 fs/xfs/libxfs/xfs_bmap.c cur->bc_private.b.flags = wasdel ? XFS_BTCUR_BPRV_WASDEL : 0; cur 727 fs/xfs/libxfs/xfs_bmap.c cur->bc_private.b.allocated++; cur 761 fs/xfs/libxfs/xfs_bmap.c pp = XFS_BMBT_PTR_ADDR(mp, block, 1, xfs_bmbt_get_maxrecs(cur, cur 769 fs/xfs/libxfs/xfs_bmap.c xfs_btree_log_block(cur, abp, XFS_BB_ALL_BITS); cur 770 fs/xfs/libxfs/xfs_bmap.c xfs_btree_log_recs(cur, abp, 1, be16_to_cpu(ablock->bb_numrecs)); cur 772 fs/xfs/libxfs/xfs_bmap.c *curp = cur; cur 782 fs/xfs/libxfs/xfs_bmap.c xfs_btree_del_cursor(cur, XFS_BTREE_ERROR); cur 925 fs/xfs/libxfs/xfs_bmap.c xfs_btree_cur_t *cur; /* btree cursor */ cur 934 fs/xfs/libxfs/xfs_bmap.c cur = xfs_bmbt_init_cursor(mp, tp, ip, XFS_DATA_FORK); cur 935 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_first(cur, &stat); cur 940 fs/xfs/libxfs/xfs_bmap.c if ((error = xfs_btree_new_iroot(cur, flags, &stat))) cur 943 fs/xfs/libxfs/xfs_bmap.c xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR); cur 946 fs/xfs/libxfs/xfs_bmap.c cur->bc_private.b.allocated = 0; cur 947 fs/xfs/libxfs/xfs_bmap.c xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR); cur 951 fs/xfs/libxfs/xfs_bmap.c xfs_btree_del_cursor(cur, XFS_BTREE_ERROR); cur 964 fs/xfs/libxfs/xfs_bmap.c xfs_btree_cur_t *cur; /* bmap btree cursor */ cur 969 fs/xfs/libxfs/xfs_bmap.c cur = NULL; cur 970 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmap_extents_to_btree(tp, ip, &cur, 0, flags, cur 972 fs/xfs/libxfs/xfs_bmap.c if (cur) { cur 973 fs/xfs/libxfs/xfs_bmap.c cur->bc_private.b.allocated = 0; cur 974 fs/xfs/libxfs/xfs_bmap.c xfs_btree_del_cursor(cur, error); cur 1557 fs/xfs/libxfs/xfs_bmap.c ASSERT(!bma->cur || cur 1558 fs/xfs/libxfs/xfs_bmap.c (bma->cur->bc_private.b.flags & XFS_BTCUR_BPRV_WASDEL)); cur 1648 fs/xfs/libxfs/xfs_bmap.c if (bma->cur == NULL) cur 1652 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(bma->cur, &RIGHT, &i); cur 1656 fs/xfs/libxfs/xfs_bmap.c error = xfs_btree_delete(bma->cur, &i); cur 1660 fs/xfs/libxfs/xfs_bmap.c error = xfs_btree_decrement(bma->cur, 0, &i); cur 1664 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(bma->cur, &LEFT); cur 1682 fs/xfs/libxfs/xfs_bmap.c if (bma->cur == NULL) cur 1686 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(bma->cur, &old, &i); cur 1690 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(bma->cur, &LEFT); cur 1712 fs/xfs/libxfs/xfs_bmap.c if (bma->cur == NULL) cur 1716 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(bma->cur, &RIGHT, &i); cur 1720 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(bma->cur, &PREV); cur 1737 fs/xfs/libxfs/xfs_bmap.c if (bma->cur == NULL) cur 1741 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(bma->cur, new, &i); cur 1745 fs/xfs/libxfs/xfs_bmap.c error = xfs_btree_insert(bma->cur, &i); cur 1772 fs/xfs/libxfs/xfs_bmap.c if (bma->cur == NULL) cur 1776 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(bma->cur, &old, &i); cur 1780 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(bma->cur, &LEFT); cur 1793 fs/xfs/libxfs/xfs_bmap.c if (bma->cur == NULL) cur 1797 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(bma->cur, new, &i); cur 1801 fs/xfs/libxfs/xfs_bmap.c error = xfs_btree_insert(bma->cur, &i); cur 1809 fs/xfs/libxfs/xfs_bmap.c &bma->cur, 1, &tmp_rval, whichfork); cur 1818 fs/xfs/libxfs/xfs_bmap.c (bma->cur ? bma->cur->bc_private.b.allocated : 0)); cur 1838 fs/xfs/libxfs/xfs_bmap.c if (bma->cur == NULL) cur 1842 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(bma->cur, &old, &i); cur 1846 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(bma->cur, &RIGHT); cur 1870 fs/xfs/libxfs/xfs_bmap.c if (bma->cur == NULL) cur 1874 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(bma->cur, new, &i); cur 1878 fs/xfs/libxfs/xfs_bmap.c error = xfs_btree_insert(bma->cur, &i); cur 1886 fs/xfs/libxfs/xfs_bmap.c &bma->cur, 1, &tmp_rval, whichfork); cur 1895 fs/xfs/libxfs/xfs_bmap.c (bma->cur ? bma->cur->bc_private.b.allocated : 0)); cur 1950 fs/xfs/libxfs/xfs_bmap.c if (bma->cur == NULL) cur 1954 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(bma->cur, new, &i); cur 1958 fs/xfs/libxfs/xfs_bmap.c error = xfs_btree_insert(bma->cur, &i); cur 1966 fs/xfs/libxfs/xfs_bmap.c &bma->cur, 1, &tmp_rval, whichfork); cur 1997 fs/xfs/libxfs/xfs_bmap.c ASSERT(bma->cur == NULL); cur 1999 fs/xfs/libxfs/xfs_bmap.c &bma->cur, da_old > 0, &tmp_logflags, cur 2009 fs/xfs/libxfs/xfs_bmap.c if (bma->cur) { cur 2010 fs/xfs/libxfs/xfs_bmap.c da_new += bma->cur->bc_private.b.allocated; cur 2011 fs/xfs/libxfs/xfs_bmap.c bma->cur->bc_private.b.allocated = 0; cur 2021 fs/xfs/libxfs/xfs_bmap.c xfs_bmap_check_leaf_extents(bma->cur, bma->ip, whichfork); cur 2044 fs/xfs/libxfs/xfs_bmap.c xfs_btree_cur_t *cur; /* btree cursor */ cur 2058 fs/xfs/libxfs/xfs_bmap.c cur = *curp; cur 2148 fs/xfs/libxfs/xfs_bmap.c if (cur == NULL) cur 2152 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, &RIGHT, &i); cur 2156 fs/xfs/libxfs/xfs_bmap.c if ((error = xfs_btree_delete(cur, &i))) cur 2159 fs/xfs/libxfs/xfs_bmap.c if ((error = xfs_btree_decrement(cur, 0, &i))) cur 2162 fs/xfs/libxfs/xfs_bmap.c if ((error = xfs_btree_delete(cur, &i))) cur 2165 fs/xfs/libxfs/xfs_bmap.c if ((error = xfs_btree_decrement(cur, 0, &i))) cur 2168 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(cur, &LEFT); cur 2186 fs/xfs/libxfs/xfs_bmap.c if (cur == NULL) cur 2190 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, &PREV, &i); cur 2194 fs/xfs/libxfs/xfs_bmap.c if ((error = xfs_btree_delete(cur, &i))) cur 2197 fs/xfs/libxfs/xfs_bmap.c if ((error = xfs_btree_decrement(cur, 0, &i))) cur 2200 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(cur, &LEFT); cur 2221 fs/xfs/libxfs/xfs_bmap.c if (cur == NULL) cur 2225 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, &RIGHT, &i); cur 2229 fs/xfs/libxfs/xfs_bmap.c if ((error = xfs_btree_delete(cur, &i))) cur 2232 fs/xfs/libxfs/xfs_bmap.c if ((error = xfs_btree_decrement(cur, 0, &i))) cur 2235 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(cur, &PREV); cur 2250 fs/xfs/libxfs/xfs_bmap.c if (cur == NULL) cur 2254 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, new, &i); cur 2258 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(cur, &PREV); cur 2280 fs/xfs/libxfs/xfs_bmap.c if (cur == NULL) cur 2284 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, &old, &i); cur 2288 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(cur, &PREV); cur 2291 fs/xfs/libxfs/xfs_bmap.c error = xfs_btree_decrement(cur, 0, &i); cur 2294 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(cur, &LEFT); cur 2314 fs/xfs/libxfs/xfs_bmap.c if (cur == NULL) cur 2318 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, &old, &i); cur 2322 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(cur, &PREV); cur 2325 fs/xfs/libxfs/xfs_bmap.c cur->bc_rec.b = *new; cur 2326 fs/xfs/libxfs/xfs_bmap.c if ((error = xfs_btree_insert(cur, &i))) cur 2348 fs/xfs/libxfs/xfs_bmap.c if (cur == NULL) cur 2352 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, &old, &i); cur 2356 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(cur, &PREV); cur 2359 fs/xfs/libxfs/xfs_bmap.c error = xfs_btree_increment(cur, 0, &i); cur 2362 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(cur, &RIGHT); cur 2382 fs/xfs/libxfs/xfs_bmap.c if (cur == NULL) cur 2386 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, &old, &i); cur 2390 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(cur, &PREV); cur 2393 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, new, &i); cur 2397 fs/xfs/libxfs/xfs_bmap.c if ((error = xfs_btree_insert(cur, &i))) cur 2426 fs/xfs/libxfs/xfs_bmap.c if (cur == NULL) cur 2430 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, &old, &i); cur 2435 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(cur, &r[1]); cur 2439 fs/xfs/libxfs/xfs_bmap.c cur->bc_rec.b = PREV; cur 2440 fs/xfs/libxfs/xfs_bmap.c if ((error = xfs_btree_insert(cur, &i))) cur 2448 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, new, &i); cur 2453 fs/xfs/libxfs/xfs_bmap.c if ((error = xfs_btree_insert(cur, &i))) cur 2479 fs/xfs/libxfs/xfs_bmap.c ASSERT(cur == NULL); cur 2480 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmap_extents_to_btree(tp, ip, &cur, 0, cur 2488 fs/xfs/libxfs/xfs_bmap.c if (cur) { cur 2489 fs/xfs/libxfs/xfs_bmap.c cur->bc_private.b.allocated = 0; cur 2490 fs/xfs/libxfs/xfs_bmap.c *curp = cur; cur 2658 fs/xfs/libxfs/xfs_bmap.c struct xfs_btree_cur *cur = *curp; cur 2668 fs/xfs/libxfs/xfs_bmap.c ASSERT(!cur || !(cur->bc_private.b.flags & XFS_BTCUR_BPRV_WASDEL)); cur 2731 fs/xfs/libxfs/xfs_bmap.c if (cur == NULL) { cur 2735 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, &right, &i); cur 2739 fs/xfs/libxfs/xfs_bmap.c error = xfs_btree_delete(cur, &i); cur 2743 fs/xfs/libxfs/xfs_bmap.c error = xfs_btree_decrement(cur, 0, &i); cur 2747 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(cur, &left); cur 2765 fs/xfs/libxfs/xfs_bmap.c if (cur == NULL) { cur 2769 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, &old, &i); cur 2773 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(cur, &left); cur 2792 fs/xfs/libxfs/xfs_bmap.c if (cur == NULL) { cur 2796 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, &old, &i); cur 2800 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(cur, &right); cur 2815 fs/xfs/libxfs/xfs_bmap.c if (cur == NULL) { cur 2819 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, new, &i); cur 2823 fs/xfs/libxfs/xfs_bmap.c error = xfs_btree_insert(cur, &i); cur 2839 fs/xfs/libxfs/xfs_bmap.c ASSERT(cur == NULL); cur 2843 fs/xfs/libxfs/xfs_bmap.c cur = *curp; cur 2849 fs/xfs/libxfs/xfs_bmap.c if (cur) cur 2850 fs/xfs/libxfs/xfs_bmap.c cur->bc_private.b.allocated = 0; cur 2852 fs/xfs/libxfs/xfs_bmap.c xfs_bmap_check_leaf_extents(cur, ip, whichfork); cur 4076 fs/xfs/libxfs/xfs_bmap.c if ((ifp->if_flags & XFS_IFBROOT) && !bma->cur) cur 4077 fs/xfs/libxfs/xfs_bmap.c bma->cur = xfs_bmbt_init_cursor(mp, bma->tp, bma->ip, whichfork); cur 4084 fs/xfs/libxfs/xfs_bmap.c if (bma->cur) cur 4085 fs/xfs/libxfs/xfs_bmap.c bma->cur->bc_private.b.flags = cur 4110 fs/xfs/libxfs/xfs_bmap.c whichfork, &bma->icur, &bma->cur, &bma->got, cur 4159 fs/xfs/libxfs/xfs_bmap.c if ((ifp->if_flags & XFS_IFBROOT) && !bma->cur) { cur 4160 fs/xfs/libxfs/xfs_bmap.c bma->cur = xfs_bmbt_init_cursor(bma->ip->i_mount, bma->tp, cur 4178 fs/xfs/libxfs/xfs_bmap.c &bma->icur, &bma->cur, mval, &tmp_logflags); cur 4245 fs/xfs/libxfs/xfs_bmap.c if (bma->cur) cur 4246 fs/xfs/libxfs/xfs_bmap.c xfs_btree_del_cursor(bma->cur, error); cur 4432 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmap_btree_to_extents(tp, ip, bma.cur, &bma.logflags, cur 4537 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmap_btree_to_extents(tp, ip, bma.cur, &bma.logflags, cur 4566 fs/xfs/libxfs/xfs_bmap.c struct xfs_btree_cur *cur = NULL; cur 4608 fs/xfs/libxfs/xfs_bmap.c cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork); cur 4609 fs/xfs/libxfs/xfs_bmap.c cur->bc_private.b.flags = 0; cur 4621 fs/xfs/libxfs/xfs_bmap.c &cur, &got, &logflags, flags); cur 4625 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmap_btree_to_extents(tp, ip, cur, &logflags, whichfork); cur 4635 fs/xfs/libxfs/xfs_bmap.c if (cur) cur 4636 fs/xfs/libxfs/xfs_bmap.c xfs_btree_del_cursor(cur, error); cur 4934 fs/xfs/libxfs/xfs_bmap.c xfs_btree_cur_t *cur, /* if null, not a btree */ cur 5012 fs/xfs/libxfs/xfs_bmap.c if (cur) { cur 5013 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, &got, &i); cur 5034 fs/xfs/libxfs/xfs_bmap.c if (!cur) { cur 5038 fs/xfs/libxfs/xfs_bmap.c if ((error = xfs_btree_delete(cur, &i))) cur 5050 fs/xfs/libxfs/xfs_bmap.c if (!cur) { cur 5054 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(cur, &got); cur 5064 fs/xfs/libxfs/xfs_bmap.c if (!cur) { cur 5068 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(cur, &got); cur 5087 fs/xfs/libxfs/xfs_bmap.c if (cur) { cur 5088 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(cur, &got); cur 5091 fs/xfs/libxfs/xfs_bmap.c error = xfs_btree_increment(cur, 0, &i); cur 5094 fs/xfs/libxfs/xfs_bmap.c cur->bc_rec.b = new; cur 5095 fs/xfs/libxfs/xfs_bmap.c error = xfs_btree_insert(cur, &i); cur 5108 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, &got, &i); cur 5116 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(cur, &old); cur 5186 fs/xfs/libxfs/xfs_bmap.c struct xfs_btree_cur *cur; /* bmap btree cursor */ cur 5257 fs/xfs/libxfs/xfs_bmap.c cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork); cur 5258 fs/xfs/libxfs/xfs_bmap.c cur->bc_private.b.flags = 0; cur 5260 fs/xfs/libxfs/xfs_bmap.c cur = NULL; cur 5371 fs/xfs/libxfs/xfs_bmap.c whichfork, &icur, &cur, &del, cur 5427 fs/xfs/libxfs/xfs_bmap.c ip, whichfork, &icur, &cur, cur 5436 fs/xfs/libxfs/xfs_bmap.c ip, whichfork, &icur, &cur, cur 5449 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmap_del_extent_real(ip, tp, &icur, cur, cur 5483 fs/xfs/libxfs/xfs_bmap.c ASSERT(cur == NULL); cur 5484 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmap_extents_to_btree(tp, ip, &cur, 0, cur 5488 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmap_btree_to_extents(tp, ip, cur, &logflags, cur 5509 fs/xfs/libxfs/xfs_bmap.c if (cur) { cur 5511 fs/xfs/libxfs/xfs_bmap.c cur->bc_private.b.allocated = 0; cur 5512 fs/xfs/libxfs/xfs_bmap.c xfs_btree_del_cursor(cur, error); cur 5580 fs/xfs/libxfs/xfs_bmap.c struct xfs_btree_cur *cur, cur 5604 fs/xfs/libxfs/xfs_bmap.c if (!cur) { cur 5610 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, got, &i); cur 5615 fs/xfs/libxfs/xfs_bmap.c error = xfs_btree_delete(cur, &i); cur 5621 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, left, &i); cur 5626 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(cur, &new); cur 5631 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmap_btree_to_extents(tp, ip, cur, logflags, whichfork); cur 5656 fs/xfs/libxfs/xfs_bmap.c struct xfs_btree_cur *cur, cur 5668 fs/xfs/libxfs/xfs_bmap.c if (cur) { cur 5669 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, &prev, &i); cur 5674 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(cur, got); cur 5701 fs/xfs/libxfs/xfs_bmap.c struct xfs_btree_cur *cur = NULL; cur 5728 fs/xfs/libxfs/xfs_bmap.c cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork); cur 5729 fs/xfs/libxfs/xfs_bmap.c cur->bc_private.b.flags = 0; cur 5749 fs/xfs/libxfs/xfs_bmap.c cur, &logflags); cur 5762 fs/xfs/libxfs/xfs_bmap.c cur, &logflags, new_startoff); cur 5774 fs/xfs/libxfs/xfs_bmap.c if (cur) cur 5775 fs/xfs/libxfs/xfs_bmap.c xfs_btree_del_cursor(cur, error); cur 5819 fs/xfs/libxfs/xfs_bmap.c struct xfs_btree_cur *cur = NULL; cur 5846 fs/xfs/libxfs/xfs_bmap.c cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork); cur 5847 fs/xfs/libxfs/xfs_bmap.c cur->bc_private.b.flags = 0; cur 5889 fs/xfs/libxfs/xfs_bmap.c cur, &logflags, new_startoff); cur 5901 fs/xfs/libxfs/xfs_bmap.c if (cur) cur 5902 fs/xfs/libxfs/xfs_bmap.c xfs_btree_del_cursor(cur, error); cur 5921 fs/xfs/libxfs/xfs_bmap.c struct xfs_btree_cur *cur = NULL; cur 5966 fs/xfs/libxfs/xfs_bmap.c cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork); cur 5967 fs/xfs/libxfs/xfs_bmap.c cur->bc_private.b.flags = 0; cur 5968 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, &got, &i); cur 5979 fs/xfs/libxfs/xfs_bmap.c if (cur) { cur 5980 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_update(cur, &got); cur 5992 fs/xfs/libxfs/xfs_bmap.c if (cur) { cur 5993 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmbt_lookup_eq(cur, &new, &i); cur 5997 fs/xfs/libxfs/xfs_bmap.c error = xfs_btree_insert(cur, &i); cur 6009 fs/xfs/libxfs/xfs_bmap.c ASSERT(cur == NULL); cur 6010 fs/xfs/libxfs/xfs_bmap.c error = xfs_bmap_extents_to_btree(tp, ip, &cur, 0, cur 6016 fs/xfs/libxfs/xfs_bmap.c if (cur) { cur 6017 fs/xfs/libxfs/xfs_bmap.c cur->bc_private.b.allocated = 0; cur 6018 fs/xfs/libxfs/xfs_bmap.c xfs_btree_del_cursor(cur, error); cur 31 fs/xfs/libxfs/xfs_bmap.h struct xfs_btree_cur *cur; /* btree cursor */ cur 211 fs/xfs/libxfs/xfs_bmap.h struct xfs_iext_cursor *cur, struct xfs_bmbt_irec *got, cur 214 fs/xfs/libxfs/xfs_bmap.h struct xfs_iext_cursor *cur, struct xfs_bmbt_irec *got, cur 228 fs/xfs/libxfs/xfs_bmap.h struct xfs_bmbt_irec *got, struct xfs_iext_cursor *cur, cur 164 fs/xfs/libxfs/xfs_bmap_btree.c struct xfs_btree_cur *cur) cur 168 fs/xfs/libxfs/xfs_bmap_btree.c new = xfs_bmbt_init_cursor(cur->bc_mp, cur->bc_tp, cur 169 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_private.b.ip, cur->bc_private.b.whichfork); cur 175 fs/xfs/libxfs/xfs_bmap_btree.c new->bc_private.b.flags = cur->bc_private.b.flags; cur 196 fs/xfs/libxfs/xfs_bmap_btree.c struct xfs_btree_cur *cur, cur 205 fs/xfs/libxfs/xfs_bmap_btree.c args.tp = cur->bc_tp; cur 206 fs/xfs/libxfs/xfs_bmap_btree.c args.mp = cur->bc_mp; cur 207 fs/xfs/libxfs/xfs_bmap_btree.c args.fsbno = cur->bc_tp->t_firstblock; cur 208 fs/xfs/libxfs/xfs_bmap_btree.c xfs_rmap_ino_bmbt_owner(&args.oinfo, cur->bc_private.b.ip->i_ino, cur 209 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_private.b.whichfork); cur 226 fs/xfs/libxfs/xfs_bmap_btree.c } else if (cur->bc_tp->t_flags & XFS_TRANS_LOWMODE) { cur 233 fs/xfs/libxfs/xfs_bmap_btree.c args.wasdel = cur->bc_private.b.flags & XFS_BTCUR_BPRV_WASDEL; cur 253 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_tp->t_flags |= XFS_TRANS_LOWMODE; cur 261 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_tp->t_firstblock = args.fsbno; cur 262 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_private.b.allocated++; cur 263 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_private.b.ip->i_d.di_nblocks++; cur 264 fs/xfs/libxfs/xfs_bmap_btree.c xfs_trans_log_inode(args.tp, cur->bc_private.b.ip, XFS_ILOG_CORE); cur 265 fs/xfs/libxfs/xfs_bmap_btree.c xfs_trans_mod_dquot_byino(args.tp, cur->bc_private.b.ip, cur 279 fs/xfs/libxfs/xfs_bmap_btree.c struct xfs_btree_cur *cur, cur 282 fs/xfs/libxfs/xfs_bmap_btree.c struct xfs_mount *mp = cur->bc_mp; cur 283 fs/xfs/libxfs/xfs_bmap_btree.c struct xfs_inode *ip = cur->bc_private.b.ip; cur 284 fs/xfs/libxfs/xfs_bmap_btree.c struct xfs_trans *tp = cur->bc_tp; cur 288 fs/xfs/libxfs/xfs_bmap_btree.c xfs_rmap_ino_bmbt_owner(&oinfo, ip->i_ino, cur->bc_private.b.whichfork); cur 289 fs/xfs/libxfs/xfs_bmap_btree.c xfs_bmap_add_free(cur->bc_tp, fsbno, 1, &oinfo); cur 299 fs/xfs/libxfs/xfs_bmap_btree.c struct xfs_btree_cur *cur, cur 302 fs/xfs/libxfs/xfs_bmap_btree.c if (level == cur->bc_nlevels - 1) { cur 305 fs/xfs/libxfs/xfs_bmap_btree.c ifp = XFS_IFORK_PTR(cur->bc_private.b.ip, cur 306 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_private.b.whichfork); cur 308 fs/xfs/libxfs/xfs_bmap_btree.c return xfs_bmbt_maxrecs(cur->bc_mp, cur 312 fs/xfs/libxfs/xfs_bmap_btree.c return cur->bc_mp->m_bmap_dmnr[level != 0]; cur 317 fs/xfs/libxfs/xfs_bmap_btree.c struct xfs_btree_cur *cur, cur 320 fs/xfs/libxfs/xfs_bmap_btree.c if (level == cur->bc_nlevels - 1) { cur 323 fs/xfs/libxfs/xfs_bmap_btree.c ifp = XFS_IFORK_PTR(cur->bc_private.b.ip, cur 324 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_private.b.whichfork); cur 326 fs/xfs/libxfs/xfs_bmap_btree.c return xfs_bmbt_maxrecs(cur->bc_mp, cur 330 fs/xfs/libxfs/xfs_bmap_btree.c return cur->bc_mp->m_bmap_dmxr[level != 0]; cur 345 fs/xfs/libxfs/xfs_bmap_btree.c struct xfs_btree_cur *cur, cur 348 fs/xfs/libxfs/xfs_bmap_btree.c if (level != cur->bc_nlevels - 1) cur 349 fs/xfs/libxfs/xfs_bmap_btree.c return cur->bc_mp->m_bmap_dmxr[level != 0]; cur 350 fs/xfs/libxfs/xfs_bmap_btree.c return xfs_bmdr_maxrecs(cur->bc_private.b.forksize, level == 0); cur 374 fs/xfs/libxfs/xfs_bmap_btree.c struct xfs_btree_cur *cur, cur 377 fs/xfs/libxfs/xfs_bmap_btree.c xfs_bmbt_disk_set_all(&rec->bmbt, &cur->bc_rec.b); cur 382 fs/xfs/libxfs/xfs_bmap_btree.c struct xfs_btree_cur *cur, cur 390 fs/xfs/libxfs/xfs_bmap_btree.c struct xfs_btree_cur *cur, cur 394 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_rec.b.br_startoff; cur 399 fs/xfs/libxfs/xfs_bmap_btree.c struct xfs_btree_cur *cur, cur 500 fs/xfs/libxfs/xfs_bmap_btree.c struct xfs_btree_cur *cur, cur 510 fs/xfs/libxfs/xfs_bmap_btree.c struct xfs_btree_cur *cur, cur 552 fs/xfs/libxfs/xfs_bmap_btree.c struct xfs_btree_cur *cur; cur 555 fs/xfs/libxfs/xfs_bmap_btree.c cur = kmem_zone_zalloc(xfs_btree_cur_zone, KM_NOFS); cur 557 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_tp = tp; cur 558 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_mp = mp; cur 559 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_nlevels = be16_to_cpu(ifp->if_broot->bb_level) + 1; cur 560 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_btnum = XFS_BTNUM_BMAP; cur 561 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_blocklog = mp->m_sb.sb_blocklog; cur 562 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_statoff = XFS_STATS_CALC_INDEX(xs_bmbt_2); cur 564 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_ops = &xfs_bmbt_ops; cur 565 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_flags = XFS_BTREE_LONG_PTRS | XFS_BTREE_ROOT_IN_INODE; cur 567 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_flags |= XFS_BTREE_CRC_BLOCKS; cur 569 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_private.b.forksize = XFS_IFORK_SIZE(ip, whichfork); cur 570 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_private.b.ip = ip; cur 571 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_private.b.allocated = 0; cur 572 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_private.b.flags = 0; cur 573 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_private.b.whichfork = whichfork; cur 575 fs/xfs/libxfs/xfs_bmap_btree.c return cur; cur 634 fs/xfs/libxfs/xfs_bmap_btree.c struct xfs_btree_cur *cur; cur 644 fs/xfs/libxfs/xfs_bmap_btree.c cur = xfs_bmbt_init_cursor(ip->i_mount, tp, ip, whichfork); cur 645 fs/xfs/libxfs/xfs_bmap_btree.c if (!cur) cur 647 fs/xfs/libxfs/xfs_bmap_btree.c cur->bc_private.b.flags |= XFS_BTCUR_BPRV_INVALID_OWNER; cur 649 fs/xfs/libxfs/xfs_bmap_btree.c error = xfs_btree_change_owner(cur, new_owner, buffer_list); cur 650 fs/xfs/libxfs/xfs_bmap_btree.c xfs_btree_del_cursor(cur, error); cur 58 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 63 fs/xfs/libxfs/xfs_btree.c struct xfs_mount *mp = cur->bc_mp; cur 64 fs/xfs/libxfs/xfs_btree.c xfs_btnum_t btnum = cur->bc_btnum; cur 82 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->get_maxrecs(cur, level)) cur 85 fs/xfs/libxfs/xfs_btree.c !xfs_btree_check_lptr(cur, be64_to_cpu(block->bb_u.l.bb_leftsib), cur 89 fs/xfs/libxfs/xfs_btree.c !xfs_btree_check_lptr(cur, be64_to_cpu(block->bb_u.l.bb_rightsib), cur 99 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 104 fs/xfs/libxfs/xfs_btree.c struct xfs_mount *mp = cur->bc_mp; cur 107 fs/xfs/libxfs/xfs_btree.c fa = __xfs_btree_check_lblock(cur, block, level, bp); cur 124 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 129 fs/xfs/libxfs/xfs_btree.c struct xfs_mount *mp = cur->bc_mp; cur 130 fs/xfs/libxfs/xfs_btree.c xfs_btnum_t btnum = cur->bc_btnum; cur 146 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->get_maxrecs(cur, level)) cur 149 fs/xfs/libxfs/xfs_btree.c !xfs_btree_check_sptr(cur, be32_to_cpu(block->bb_u.s.bb_leftsib), cur 153 fs/xfs/libxfs/xfs_btree.c !xfs_btree_check_sptr(cur, be32_to_cpu(block->bb_u.s.bb_rightsib), cur 163 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 168 fs/xfs/libxfs/xfs_btree.c struct xfs_mount *mp = cur->bc_mp; cur 171 fs/xfs/libxfs/xfs_btree.c fa = __xfs_btree_check_sblock(cur, block, level, bp); cur 187 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, /* btree cursor */ cur 192 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_LONG_PTRS) cur 193 fs/xfs/libxfs/xfs_btree.c return xfs_btree_check_lblock(cur, block, level, bp); cur 195 fs/xfs/libxfs/xfs_btree.c return xfs_btree_check_sblock(cur, block, level, bp); cur 201 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 207 fs/xfs/libxfs/xfs_btree.c return xfs_verify_fsbno(cur->bc_mp, fsbno); cur 213 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 219 fs/xfs/libxfs/xfs_btree.c return xfs_verify_agbno(cur->bc_mp, cur->bc_private.a.agno, agbno); cur 228 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 233 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_LONG_PTRS) { cur 234 fs/xfs/libxfs/xfs_btree.c if (xfs_btree_check_lptr(cur, be64_to_cpu((&ptr->l)[index]), cur 237 fs/xfs/libxfs/xfs_btree.c xfs_err(cur->bc_mp, cur 239 fs/xfs/libxfs/xfs_btree.c cur->bc_private.b.ip->i_ino, cur 240 fs/xfs/libxfs/xfs_btree.c cur->bc_private.b.whichfork, cur->bc_btnum, cur 243 fs/xfs/libxfs/xfs_btree.c if (xfs_btree_check_sptr(cur, be32_to_cpu((&ptr->s)[index]), cur 246 fs/xfs/libxfs/xfs_btree.c xfs_err(cur->bc_mp, cur 248 fs/xfs/libxfs/xfs_btree.c cur->bc_private.a.agno, cur->bc_btnum, cur 339 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 344 fs/xfs/libxfs/xfs_btree.c error = cur->bc_ops->free_block(cur, bp); cur 346 fs/xfs/libxfs/xfs_btree.c xfs_trans_binval(cur->bc_tp, bp); cur 347 fs/xfs/libxfs/xfs_btree.c XFS_BTREE_STATS_INC(cur, free); cur 357 fs/xfs/libxfs/xfs_btree.c xfs_btree_cur_t *cur, /* btree cursor */ cur 372 fs/xfs/libxfs/xfs_btree.c for (i = 0; i < cur->bc_nlevels; i++) { cur 373 fs/xfs/libxfs/xfs_btree.c if (cur->bc_bufs[i]) cur 374 fs/xfs/libxfs/xfs_btree.c xfs_trans_brelse(cur->bc_tp, cur->bc_bufs[i]); cur 382 fs/xfs/libxfs/xfs_btree.c ASSERT(cur->bc_btnum != XFS_BTNUM_BMAP || cur 383 fs/xfs/libxfs/xfs_btree.c cur->bc_private.b.allocated == 0); cur 387 fs/xfs/libxfs/xfs_btree.c kmem_zone_free(xfs_btree_cur_zone, cur); cur 396 fs/xfs/libxfs/xfs_btree.c xfs_btree_cur_t *cur, /* input cursor */ cur 406 fs/xfs/libxfs/xfs_btree.c tp = cur->bc_tp; cur 407 fs/xfs/libxfs/xfs_btree.c mp = cur->bc_mp; cur 412 fs/xfs/libxfs/xfs_btree.c new = cur->bc_ops->dup_cursor(cur); cur 417 fs/xfs/libxfs/xfs_btree.c new->bc_rec = cur->bc_rec; cur 423 fs/xfs/libxfs/xfs_btree.c new->bc_ptrs[i] = cur->bc_ptrs[i]; cur 424 fs/xfs/libxfs/xfs_btree.c new->bc_ra[i] = cur->bc_ra[i]; cur 425 fs/xfs/libxfs/xfs_btree.c bp = cur->bc_bufs[i]; cur 430 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->buf_ops); cur 520 fs/xfs/libxfs/xfs_btree.c static inline size_t xfs_btree_block_len(struct xfs_btree_cur *cur) cur 522 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_LONG_PTRS) { cur 523 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_CRC_BLOCKS) cur 527 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_CRC_BLOCKS) cur 535 fs/xfs/libxfs/xfs_btree.c static inline size_t xfs_btree_ptr_len(struct xfs_btree_cur *cur) cur 537 fs/xfs/libxfs/xfs_btree.c return (cur->bc_flags & XFS_BTREE_LONG_PTRS) ? cur 546 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 549 fs/xfs/libxfs/xfs_btree.c return xfs_btree_block_len(cur) + cur 550 fs/xfs/libxfs/xfs_btree.c (n - 1) * cur->bc_ops->rec_len; cur 558 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 561 fs/xfs/libxfs/xfs_btree.c return xfs_btree_block_len(cur) + cur 562 fs/xfs/libxfs/xfs_btree.c (n - 1) * cur->bc_ops->key_len; cur 570 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 573 fs/xfs/libxfs/xfs_btree.c return xfs_btree_block_len(cur) + cur 574 fs/xfs/libxfs/xfs_btree.c (n - 1) * cur->bc_ops->key_len + (cur->bc_ops->key_len / 2); cur 582 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 586 fs/xfs/libxfs/xfs_btree.c return xfs_btree_block_len(cur) + cur 587 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->get_maxrecs(cur, level) * cur->bc_ops->key_len + cur 588 fs/xfs/libxfs/xfs_btree.c (n - 1) * xfs_btree_ptr_len(cur); cur 596 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 601 fs/xfs/libxfs/xfs_btree.c ((char *)block + xfs_btree_rec_offset(cur, n)); cur 609 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 614 fs/xfs/libxfs/xfs_btree.c ((char *)block + xfs_btree_key_offset(cur, n)); cur 622 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 627 fs/xfs/libxfs/xfs_btree.c ((char *)block + xfs_btree_high_key_offset(cur, n)); cur 635 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 644 fs/xfs/libxfs/xfs_btree.c ((char *)block + xfs_btree_ptr_offset(cur, n, level)); cur 655 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur) cur 659 fs/xfs/libxfs/xfs_btree.c ifp = XFS_IFORK_PTR(cur->bc_private.b.ip, cur->bc_private.b.whichfork); cur 669 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, /* btree cursor */ cur 673 fs/xfs/libxfs/xfs_btree.c if ((cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) && cur 674 fs/xfs/libxfs/xfs_btree.c (level == cur->bc_nlevels - 1)) { cur 676 fs/xfs/libxfs/xfs_btree.c return xfs_btree_get_iroot(cur); cur 679 fs/xfs/libxfs/xfs_btree.c *bpp = cur->bc_bufs[level]; cur 724 fs/xfs/libxfs/xfs_btree.c xfs_btree_cur_t *cur, /* btree cursor */ cur 730 fs/xfs/libxfs/xfs_btree.c block = xfs_btree_get_block(cur, level, &bp); cur 731 fs/xfs/libxfs/xfs_btree.c xfs_btree_check_block(cur, block, level, bp); cur 732 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_LONG_PTRS) cur 744 fs/xfs/libxfs/xfs_btree.c xfs_btree_cur_t *cur, /* btree cursor */ cur 753 fs/xfs/libxfs/xfs_btree.c block = xfs_btree_get_block(cur, level, &bp); cur 754 fs/xfs/libxfs/xfs_btree.c if (xfs_btree_check_block(cur, block, level, bp)) cur 764 fs/xfs/libxfs/xfs_btree.c cur->bc_ptrs[level] = 1; cur 774 fs/xfs/libxfs/xfs_btree.c xfs_btree_cur_t *cur, /* btree cursor */ cur 783 fs/xfs/libxfs/xfs_btree.c block = xfs_btree_get_block(cur, level, &bp); cur 784 fs/xfs/libxfs/xfs_btree.c if (xfs_btree_check_block(cur, block, level, bp)) cur 794 fs/xfs/libxfs/xfs_btree.c cur->bc_ptrs[level] = be16_to_cpu(block->bb_numrecs); cur 906 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 915 fs/xfs/libxfs/xfs_btree.c xfs_btree_reada_bufl(cur->bc_mp, left, 1, cur 916 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->buf_ops); cur 921 fs/xfs/libxfs/xfs_btree.c xfs_btree_reada_bufl(cur->bc_mp, right, 1, cur 922 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->buf_ops); cur 931 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 941 fs/xfs/libxfs/xfs_btree.c xfs_btree_reada_bufs(cur->bc_mp, cur->bc_private.a.agno, cur 942 fs/xfs/libxfs/xfs_btree.c left, 1, cur->bc_ops->buf_ops); cur 947 fs/xfs/libxfs/xfs_btree.c xfs_btree_reada_bufs(cur->bc_mp, cur->bc_private.a.agno, cur 948 fs/xfs/libxfs/xfs_btree.c right, 1, cur->bc_ops->buf_ops); cur 961 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, /* btree cursor */ cur 971 fs/xfs/libxfs/xfs_btree.c if ((cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) && cur 972 fs/xfs/libxfs/xfs_btree.c (lev == cur->bc_nlevels - 1)) cur 975 fs/xfs/libxfs/xfs_btree.c if ((cur->bc_ra[lev] | lr) == cur->bc_ra[lev]) cur 978 fs/xfs/libxfs/xfs_btree.c cur->bc_ra[lev] |= lr; cur 979 fs/xfs/libxfs/xfs_btree.c block = XFS_BUF_TO_BLOCK(cur->bc_bufs[lev]); cur 981 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_LONG_PTRS) cur 982 fs/xfs/libxfs/xfs_btree.c return xfs_btree_readahead_lblock(cur, lr, block); cur 983 fs/xfs/libxfs/xfs_btree.c return xfs_btree_readahead_sblock(cur, lr, block); cur 988 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 996 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_check_ptr(cur, ptr, 0, 1); cur 1000 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_LONG_PTRS) { cur 1002 fs/xfs/libxfs/xfs_btree.c *daddr = XFS_FSB_TO_DADDR(cur->bc_mp, fsbno); cur 1005 fs/xfs/libxfs/xfs_btree.c *daddr = XFS_AGB_TO_DADDR(cur->bc_mp, cur->bc_private.a.agno, cur 1020 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 1026 fs/xfs/libxfs/xfs_btree.c if (xfs_btree_ptr_to_daddr(cur, ptr, &daddr)) cur 1028 fs/xfs/libxfs/xfs_btree.c xfs_buf_readahead(cur->bc_mp->m_ddev_targp, daddr, cur 1029 fs/xfs/libxfs/xfs_btree.c cur->bc_mp->m_bsize * count, cur->bc_ops->buf_ops); cur 1038 fs/xfs/libxfs/xfs_btree.c xfs_btree_cur_t *cur, /* btree cursor */ cur 1044 fs/xfs/libxfs/xfs_btree.c if (cur->bc_bufs[lev]) cur 1045 fs/xfs/libxfs/xfs_btree.c xfs_trans_brelse(cur->bc_tp, cur->bc_bufs[lev]); cur 1046 fs/xfs/libxfs/xfs_btree.c cur->bc_bufs[lev] = bp; cur 1047 fs/xfs/libxfs/xfs_btree.c cur->bc_ra[lev] = 0; cur 1050 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_LONG_PTRS) { cur 1052 fs/xfs/libxfs/xfs_btree.c cur->bc_ra[lev] |= XFS_BTCUR_LEFTRA; cur 1054 fs/xfs/libxfs/xfs_btree.c cur->bc_ra[lev] |= XFS_BTCUR_RIGHTRA; cur 1057 fs/xfs/libxfs/xfs_btree.c cur->bc_ra[lev] |= XFS_BTCUR_LEFTRA; cur 1059 fs/xfs/libxfs/xfs_btree.c cur->bc_ra[lev] |= XFS_BTCUR_RIGHTRA; cur 1065 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 1068 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_LONG_PTRS) cur 1076 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 1079 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_LONG_PTRS) cur 1090 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 1097 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_LONG_PTRS) { cur 1112 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 1119 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_LONG_PTRS) { cur 1190 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 1203 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_LONG_PTRS) cur 1204 fs/xfs/libxfs/xfs_btree.c owner = cur->bc_private.b.ip->i_ino; cur 1206 fs/xfs/libxfs/xfs_btree.c owner = cur->bc_private.a.agno; cur 1208 fs/xfs/libxfs/xfs_btree.c xfs_btree_init_block_int(cur->bc_mp, XFS_BUF_TO_BLOCK(bp), bp->b_bn, cur 1209 fs/xfs/libxfs/xfs_btree.c cur->bc_btnum, level, numrecs, cur 1210 fs/xfs/libxfs/xfs_btree.c owner, cur->bc_flags); cur 1220 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 1228 fs/xfs/libxfs/xfs_btree.c if (!(cur->bc_flags & XFS_BTREE_LASTREC_UPDATE)) cur 1231 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_sibling(cur, block, &ptr, XFS_BB_RIGHTSIB); cur 1232 fs/xfs/libxfs/xfs_btree.c if (!xfs_btree_ptr_is_null(cur, &ptr)) cur 1239 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 1243 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_LONG_PTRS) cur 1244 fs/xfs/libxfs/xfs_btree.c ptr->l = cpu_to_be64(XFS_DADDR_TO_FSB(cur->bc_mp, cur 1247 fs/xfs/libxfs/xfs_btree.c ptr->s = cpu_to_be32(xfs_daddr_to_agbno(cur->bc_mp, cur 1254 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 1257 fs/xfs/libxfs/xfs_btree.c switch (cur->bc_btnum) { cur 1282 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 1287 fs/xfs/libxfs/xfs_btree.c struct xfs_mount *mp = cur->bc_mp; cur 1291 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_ptr_to_daddr(cur, ptr, &d); cur 1294 fs/xfs/libxfs/xfs_btree.c *bpp = xfs_trans_get_buf(cur->bc_tp, mp->m_ddev_targp, d, cur 1300 fs/xfs/libxfs/xfs_btree.c (*bpp)->b_ops = cur->bc_ops->buf_ops; cur 1311 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 1317 fs/xfs/libxfs/xfs_btree.c struct xfs_mount *mp = cur->bc_mp; cur 1324 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_ptr_to_daddr(cur, ptr, &d); cur 1327 fs/xfs/libxfs/xfs_btree.c error = xfs_trans_read_buf(mp, cur->bc_tp, mp->m_ddev_targp, d, cur 1329 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->buf_ops); cur 1333 fs/xfs/libxfs/xfs_btree.c xfs_btree_set_refs(cur, *bpp); cur 1343 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 1349 fs/xfs/libxfs/xfs_btree.c memcpy(dst_key, src_key, numkeys * cur->bc_ops->key_len); cur 1357 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 1363 fs/xfs/libxfs/xfs_btree.c memcpy(dst_rec, src_rec, numrecs * cur->bc_ops->rec_len); cur 1371 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 1377 fs/xfs/libxfs/xfs_btree.c memcpy(dst_ptr, src_ptr, numptrs * xfs_btree_ptr_len(cur)); cur 1385 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 1395 fs/xfs/libxfs/xfs_btree.c dst_key = (char *)key + (dir * cur->bc_ops->key_len); cur 1396 fs/xfs/libxfs/xfs_btree.c memmove(dst_key, key, numkeys * cur->bc_ops->key_len); cur 1404 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 1414 fs/xfs/libxfs/xfs_btree.c dst_rec = (char *)rec + (dir * cur->bc_ops->rec_len); cur 1415 fs/xfs/libxfs/xfs_btree.c memmove(dst_rec, rec, numrecs * cur->bc_ops->rec_len); cur 1423 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 1433 fs/xfs/libxfs/xfs_btree.c dst_ptr = (char *)ptr + (dir * xfs_btree_ptr_len(cur)); cur 1434 fs/xfs/libxfs/xfs_btree.c memmove(dst_ptr, ptr, numptrs * xfs_btree_ptr_len(cur)); cur 1442 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 1449 fs/xfs/libxfs/xfs_btree.c xfs_trans_buf_set_type(cur->bc_tp, bp, XFS_BLFT_BTREE_BUF); cur 1450 fs/xfs/libxfs/xfs_btree.c xfs_trans_log_buf(cur->bc_tp, bp, cur 1451 fs/xfs/libxfs/xfs_btree.c xfs_btree_key_offset(cur, first), cur 1452 fs/xfs/libxfs/xfs_btree.c xfs_btree_key_offset(cur, last + 1) - 1); cur 1454 fs/xfs/libxfs/xfs_btree.c xfs_trans_log_inode(cur->bc_tp, cur->bc_private.b.ip, cur 1455 fs/xfs/libxfs/xfs_btree.c xfs_ilog_fbroot(cur->bc_private.b.whichfork)); cur 1464 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 1470 fs/xfs/libxfs/xfs_btree.c xfs_trans_buf_set_type(cur->bc_tp, bp, XFS_BLFT_BTREE_BUF); cur 1471 fs/xfs/libxfs/xfs_btree.c xfs_trans_log_buf(cur->bc_tp, bp, cur 1472 fs/xfs/libxfs/xfs_btree.c xfs_btree_rec_offset(cur, first), cur 1473 fs/xfs/libxfs/xfs_btree.c xfs_btree_rec_offset(cur, last + 1) - 1); cur 1482 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, /* btree cursor */ cur 1492 fs/xfs/libxfs/xfs_btree.c xfs_trans_buf_set_type(cur->bc_tp, bp, XFS_BLFT_BTREE_BUF); cur 1493 fs/xfs/libxfs/xfs_btree.c xfs_trans_log_buf(cur->bc_tp, bp, cur 1494 fs/xfs/libxfs/xfs_btree.c xfs_btree_ptr_offset(cur, first, level), cur 1495 fs/xfs/libxfs/xfs_btree.c xfs_btree_ptr_offset(cur, last + 1, level) - 1); cur 1497 fs/xfs/libxfs/xfs_btree.c xfs_trans_log_inode(cur->bc_tp, cur->bc_private.b.ip, cur 1498 fs/xfs/libxfs/xfs_btree.c xfs_ilog_fbroot(cur->bc_private.b.whichfork)); cur 1508 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, /* btree cursor */ cur 1545 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_CRC_BLOCKS) { cur 1560 fs/xfs/libxfs/xfs_btree.c (cur->bc_flags & XFS_BTREE_LONG_PTRS) ? cur 1563 fs/xfs/libxfs/xfs_btree.c xfs_trans_buf_set_type(cur->bc_tp, bp, XFS_BLFT_BTREE_BUF); cur 1564 fs/xfs/libxfs/xfs_btree.c xfs_trans_log_buf(cur->bc_tp, bp, first, last); cur 1566 fs/xfs/libxfs/xfs_btree.c xfs_trans_log_inode(cur->bc_tp, cur->bc_private.b.ip, cur 1567 fs/xfs/libxfs/xfs_btree.c xfs_ilog_fbroot(cur->bc_private.b.whichfork)); cur 1577 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 1587 fs/xfs/libxfs/xfs_btree.c ASSERT(level < cur->bc_nlevels); cur 1590 fs/xfs/libxfs/xfs_btree.c xfs_btree_readahead(cur, level, XFS_BTCUR_RIGHTRA); cur 1593 fs/xfs/libxfs/xfs_btree.c block = xfs_btree_get_block(cur, level, &bp); cur 1596 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_check_block(cur, block, level, bp); cur 1602 fs/xfs/libxfs/xfs_btree.c if (++cur->bc_ptrs[level] <= xfs_btree_get_numrecs(block)) cur 1606 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_sibling(cur, block, &ptr, XFS_BB_RIGHTSIB); cur 1607 fs/xfs/libxfs/xfs_btree.c if (xfs_btree_ptr_is_null(cur, &ptr)) cur 1610 fs/xfs/libxfs/xfs_btree.c XFS_BTREE_STATS_INC(cur, increment); cur 1616 fs/xfs/libxfs/xfs_btree.c for (lev = level + 1; lev < cur->bc_nlevels; lev++) { cur 1617 fs/xfs/libxfs/xfs_btree.c block = xfs_btree_get_block(cur, lev, &bp); cur 1620 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_check_block(cur, block, lev, bp); cur 1625 fs/xfs/libxfs/xfs_btree.c if (++cur->bc_ptrs[lev] <= xfs_btree_get_numrecs(block)) cur 1629 fs/xfs/libxfs/xfs_btree.c xfs_btree_readahead(cur, lev, XFS_BTCUR_RIGHTRA); cur 1636 fs/xfs/libxfs/xfs_btree.c if (lev == cur->bc_nlevels) { cur 1637 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) cur 1643 fs/xfs/libxfs/xfs_btree.c ASSERT(lev < cur->bc_nlevels); cur 1649 fs/xfs/libxfs/xfs_btree.c for (block = xfs_btree_get_block(cur, lev, &bp); lev > level; ) { cur 1652 fs/xfs/libxfs/xfs_btree.c ptrp = xfs_btree_ptr_addr(cur, cur->bc_ptrs[lev], block); cur 1654 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_read_buf_block(cur, ptrp, 0, &block, &bp); cur 1658 fs/xfs/libxfs/xfs_btree.c xfs_btree_setbuf(cur, lev, bp); cur 1659 fs/xfs/libxfs/xfs_btree.c cur->bc_ptrs[lev] = 1; cur 1679 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 1689 fs/xfs/libxfs/xfs_btree.c ASSERT(level < cur->bc_nlevels); cur 1692 fs/xfs/libxfs/xfs_btree.c xfs_btree_readahead(cur, level, XFS_BTCUR_LEFTRA); cur 1695 fs/xfs/libxfs/xfs_btree.c if (--cur->bc_ptrs[level] > 0) cur 1699 fs/xfs/libxfs/xfs_btree.c block = xfs_btree_get_block(cur, level, &bp); cur 1702 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_check_block(cur, block, level, bp); cur 1708 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_sibling(cur, block, &ptr, XFS_BB_LEFTSIB); cur 1709 fs/xfs/libxfs/xfs_btree.c if (xfs_btree_ptr_is_null(cur, &ptr)) cur 1712 fs/xfs/libxfs/xfs_btree.c XFS_BTREE_STATS_INC(cur, decrement); cur 1718 fs/xfs/libxfs/xfs_btree.c for (lev = level + 1; lev < cur->bc_nlevels; lev++) { cur 1719 fs/xfs/libxfs/xfs_btree.c if (--cur->bc_ptrs[lev] > 0) cur 1722 fs/xfs/libxfs/xfs_btree.c xfs_btree_readahead(cur, lev, XFS_BTCUR_LEFTRA); cur 1729 fs/xfs/libxfs/xfs_btree.c if (lev == cur->bc_nlevels) { cur 1730 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) cur 1736 fs/xfs/libxfs/xfs_btree.c ASSERT(lev < cur->bc_nlevels); cur 1742 fs/xfs/libxfs/xfs_btree.c for (block = xfs_btree_get_block(cur, lev, &bp); lev > level; ) { cur 1745 fs/xfs/libxfs/xfs_btree.c ptrp = xfs_btree_ptr_addr(cur, cur->bc_ptrs[lev], block); cur 1747 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_read_buf_block(cur, ptrp, 0, &block, &bp); cur 1750 fs/xfs/libxfs/xfs_btree.c xfs_btree_setbuf(cur, lev, bp); cur 1751 fs/xfs/libxfs/xfs_btree.c cur->bc_ptrs[lev] = xfs_btree_get_numrecs(block); cur 1767 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, /* btree cursor */ cur 1777 fs/xfs/libxfs/xfs_btree.c if ((cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) && cur 1778 fs/xfs/libxfs/xfs_btree.c (level == cur->bc_nlevels - 1)) { cur 1779 fs/xfs/libxfs/xfs_btree.c *blkp = xfs_btree_get_iroot(cur); cur 1789 fs/xfs/libxfs/xfs_btree.c bp = cur->bc_bufs[level]; cur 1790 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_ptr_to_daddr(cur, pp, &daddr); cur 1798 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_read_buf_block(cur, pp, 0, blkp, &bp); cur 1803 fs/xfs/libxfs/xfs_btree.c if (xfs_sb_version_hascrc(&cur->bc_mp->m_sb) && cur 1804 fs/xfs/libxfs/xfs_btree.c !(cur->bc_private.b.flags & XFS_BTCUR_BPRV_INVALID_OWNER) && cur 1805 fs/xfs/libxfs/xfs_btree.c (cur->bc_flags & XFS_BTREE_LONG_PTRS) && cur 1807 fs/xfs/libxfs/xfs_btree.c cur->bc_private.b.ip->i_ino) cur 1818 fs/xfs/libxfs/xfs_btree.c xfs_btree_setbuf(cur, level, bp); cur 1823 fs/xfs/libxfs/xfs_btree.c xfs_trans_brelse(cur->bc_tp, bp); cur 1834 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 1841 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->init_key_from_rec(kp, cur 1842 fs/xfs/libxfs/xfs_btree.c xfs_btree_rec_addr(cur, keyno, block)); cur 1846 fs/xfs/libxfs/xfs_btree.c return xfs_btree_key_addr(cur, keyno, block); cur 1855 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, /* btree cursor */ cur 1867 fs/xfs/libxfs/xfs_btree.c XFS_BTREE_STATS_INC(cur, lookup); cur 1870 fs/xfs/libxfs/xfs_btree.c if (cur->bc_nlevels == 0) cur 1877 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->init_ptr_from_cur(cur, &ptr); cur 1886 fs/xfs/libxfs/xfs_btree.c for (level = cur->bc_nlevels - 1, diff = 1; level >= 0; level--) { cur 1888 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_lookup_get_block(cur, level, pp, &block); cur 1909 fs/xfs/libxfs/xfs_btree.c if (level != 0 || cur->bc_nlevels != 1) { cur 1912 fs/xfs/libxfs/xfs_btree.c cur->bc_mp, block, cur 1917 fs/xfs/libxfs/xfs_btree.c cur->bc_ptrs[0] = dir != XFS_LOOKUP_LE; cur 1927 fs/xfs/libxfs/xfs_btree.c XFS_BTREE_STATS_INC(cur, compare); cur 1933 fs/xfs/libxfs/xfs_btree.c kp = xfs_lookup_get_search_key(cur, level, cur 1942 fs/xfs/libxfs/xfs_btree.c diff = cur->bc_ops->key_diff(cur, kp); cur 1963 fs/xfs/libxfs/xfs_btree.c pp = xfs_btree_ptr_addr(cur, keyno, block); cur 1965 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_debug_check_ptr(cur, pp, 0, level); cur 1969 fs/xfs/libxfs/xfs_btree.c cur->bc_ptrs[level] = keyno; cur 1980 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_sibling(cur, block, &ptr, XFS_BB_RIGHTSIB); cur 1983 fs/xfs/libxfs/xfs_btree.c !xfs_btree_ptr_is_null(cur, &ptr)) { cur 1986 fs/xfs/libxfs/xfs_btree.c cur->bc_ptrs[0] = keyno; cur 1987 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_increment(cur, 0, &i); cur 1990 fs/xfs/libxfs/xfs_btree.c XFS_WANT_CORRUPTED_RETURN(cur->bc_mp, i == 1); cur 1996 fs/xfs/libxfs/xfs_btree.c cur->bc_ptrs[0] = keyno; cur 2014 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 2017 fs/xfs/libxfs/xfs_btree.c ASSERT(cur->bc_flags & XFS_BTREE_OVERLAPPING); cur 2019 fs/xfs/libxfs/xfs_btree.c (cur->bc_ops->key_len / 2)); cur 2025 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 2035 fs/xfs/libxfs/xfs_btree.c rec = xfs_btree_rec_addr(cur, 1, block); cur 2036 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->init_key_from_rec(key, rec); cur 2038 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_OVERLAPPING) { cur 2040 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->init_high_key_from_rec(&max_hkey, rec); cur 2042 fs/xfs/libxfs/xfs_btree.c rec = xfs_btree_rec_addr(cur, n, block); cur 2043 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->init_high_key_from_rec(&hkey, rec); cur 2044 fs/xfs/libxfs/xfs_btree.c if (cur->bc_ops->diff_two_keys(cur, &hkey, &max_hkey) cur 2049 fs/xfs/libxfs/xfs_btree.c high = xfs_btree_high_key_from_key(cur, key); cur 2050 fs/xfs/libxfs/xfs_btree.c memcpy(high, &max_hkey, cur->bc_ops->key_len / 2); cur 2057 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 2066 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_OVERLAPPING) { cur 2067 fs/xfs/libxfs/xfs_btree.c memcpy(key, xfs_btree_key_addr(cur, 1, block), cur 2068 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->key_len / 2); cur 2070 fs/xfs/libxfs/xfs_btree.c max_hkey = xfs_btree_high_key_addr(cur, 1, block); cur 2072 fs/xfs/libxfs/xfs_btree.c hkey = xfs_btree_high_key_addr(cur, n, block); cur 2073 fs/xfs/libxfs/xfs_btree.c if (cur->bc_ops->diff_two_keys(cur, hkey, max_hkey) > 0) cur 2077 fs/xfs/libxfs/xfs_btree.c high = xfs_btree_high_key_from_key(cur, key); cur 2078 fs/xfs/libxfs/xfs_btree.c memcpy(high, max_hkey, cur->bc_ops->key_len / 2); cur 2080 fs/xfs/libxfs/xfs_btree.c memcpy(key, xfs_btree_key_addr(cur, 1, block), cur 2081 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->key_len); cur 2088 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 2093 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_leaf_keys(cur, block, key); cur 2095 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_node_keys(cur, block, key); cur 2107 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 2110 fs/xfs/libxfs/xfs_btree.c return (cur->bc_flags & XFS_BTREE_OVERLAPPING) || ptr == 1; cur 2120 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 2134 fs/xfs/libxfs/xfs_btree.c ASSERT(cur->bc_flags & XFS_BTREE_OVERLAPPING); cur 2137 fs/xfs/libxfs/xfs_btree.c if (level + 1 >= cur->bc_nlevels) cur 2140 fs/xfs/libxfs/xfs_btree.c trace_xfs_btree_updkeys(cur, level, bp0); cur 2143 fs/xfs/libxfs/xfs_btree.c hkey = xfs_btree_high_key_from_key(cur, lkey); cur 2144 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_keys(cur, block, lkey); cur 2145 fs/xfs/libxfs/xfs_btree.c for (level++; level < cur->bc_nlevels; level++) { cur 2149 fs/xfs/libxfs/xfs_btree.c block = xfs_btree_get_block(cur, level, &bp); cur 2150 fs/xfs/libxfs/xfs_btree.c trace_xfs_btree_updkeys(cur, level, bp); cur 2152 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_check_block(cur, block, level, bp); cur 2156 fs/xfs/libxfs/xfs_btree.c ptr = cur->bc_ptrs[level]; cur 2157 fs/xfs/libxfs/xfs_btree.c nlkey = xfs_btree_key_addr(cur, ptr, block); cur 2158 fs/xfs/libxfs/xfs_btree.c nhkey = xfs_btree_high_key_addr(cur, ptr, block); cur 2160 fs/xfs/libxfs/xfs_btree.c !(cur->bc_ops->diff_two_keys(cur, nlkey, lkey) != 0 || cur 2161 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->diff_two_keys(cur, nhkey, hkey) != 0)) cur 2163 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_keys(cur, nlkey, lkey, 1); cur 2164 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_keys(cur, bp, ptr, ptr); cur 2165 fs/xfs/libxfs/xfs_btree.c if (level + 1 >= cur->bc_nlevels) cur 2167 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_node_keys(cur, block, lkey); cur 2176 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 2182 fs/xfs/libxfs/xfs_btree.c block = xfs_btree_get_block(cur, level, &bp); cur 2183 fs/xfs/libxfs/xfs_btree.c return __xfs_btree_updkeys(cur, level, block, bp, true); cur 2191 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 2202 fs/xfs/libxfs/xfs_btree.c block = xfs_btree_get_block(cur, level, &bp); cur 2203 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_OVERLAPPING) cur 2204 fs/xfs/libxfs/xfs_btree.c return __xfs_btree_updkeys(cur, level, block, bp, false); cur 2212 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_keys(cur, block, &key); cur 2213 fs/xfs/libxfs/xfs_btree.c for (level++, ptr = 1; ptr == 1 && level < cur->bc_nlevels; level++) { cur 2217 fs/xfs/libxfs/xfs_btree.c block = xfs_btree_get_block(cur, level, &bp); cur 2219 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_check_block(cur, block, level, bp); cur 2223 fs/xfs/libxfs/xfs_btree.c ptr = cur->bc_ptrs[level]; cur 2224 fs/xfs/libxfs/xfs_btree.c kp = xfs_btree_key_addr(cur, ptr, block); cur 2225 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_keys(cur, kp, &key, 1); cur 2226 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_keys(cur, bp, ptr, ptr); cur 2239 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 2249 fs/xfs/libxfs/xfs_btree.c block = xfs_btree_get_block(cur, 0, &bp); cur 2252 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_check_block(cur, block, 0, bp); cur 2257 fs/xfs/libxfs/xfs_btree.c ptr = cur->bc_ptrs[0]; cur 2258 fs/xfs/libxfs/xfs_btree.c rp = xfs_btree_rec_addr(cur, ptr, block); cur 2261 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_recs(cur, rp, rec, 1); cur 2262 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_recs(cur, bp, ptr, ptr); cur 2268 fs/xfs/libxfs/xfs_btree.c if (xfs_btree_is_lastrec(cur, block, 0)) { cur 2269 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->update_lastrec(cur, block, rec, cur 2274 fs/xfs/libxfs/xfs_btree.c if (xfs_btree_needs_key_update(cur, ptr)) { cur 2275 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_update_keys(cur, 0); cur 2292 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 2310 fs/xfs/libxfs/xfs_btree.c if ((cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) && cur 2311 fs/xfs/libxfs/xfs_btree.c level == cur->bc_nlevels - 1) cur 2315 fs/xfs/libxfs/xfs_btree.c right = xfs_btree_get_block(cur, level, &rbp); cur 2318 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_check_block(cur, right, level, rbp); cur 2324 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_sibling(cur, right, &lptr, XFS_BB_LEFTSIB); cur 2325 fs/xfs/libxfs/xfs_btree.c if (xfs_btree_ptr_is_null(cur, &lptr)) cur 2332 fs/xfs/libxfs/xfs_btree.c if (cur->bc_ptrs[level] <= 1) cur 2336 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_read_buf_block(cur, &lptr, 0, &left, &lbp); cur 2342 fs/xfs/libxfs/xfs_btree.c if (lrecs == cur->bc_ops->get_maxrecs(cur, level)) cur 2355 fs/xfs/libxfs/xfs_btree.c XFS_BTREE_STATS_INC(cur, lshift); cur 2356 fs/xfs/libxfs/xfs_btree.c XFS_BTREE_STATS_ADD(cur, moves, 1); cur 2367 fs/xfs/libxfs/xfs_btree.c lkp = xfs_btree_key_addr(cur, lrecs, left); cur 2368 fs/xfs/libxfs/xfs_btree.c rkp = xfs_btree_key_addr(cur, 1, right); cur 2370 fs/xfs/libxfs/xfs_btree.c lpp = xfs_btree_ptr_addr(cur, lrecs, left); cur 2371 fs/xfs/libxfs/xfs_btree.c rpp = xfs_btree_ptr_addr(cur, 1, right); cur 2373 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_debug_check_ptr(cur, rpp, 0, level); cur 2377 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_keys(cur, lkp, rkp, 1); cur 2378 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_ptrs(cur, lpp, rpp, 1); cur 2380 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_keys(cur, lbp, lrecs, lrecs); cur 2381 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_ptrs(cur, lbp, lrecs, lrecs); cur 2383 fs/xfs/libxfs/xfs_btree.c ASSERT(cur->bc_ops->keys_inorder(cur, cur 2384 fs/xfs/libxfs/xfs_btree.c xfs_btree_key_addr(cur, lrecs - 1, left), lkp)); cur 2389 fs/xfs/libxfs/xfs_btree.c lrp = xfs_btree_rec_addr(cur, lrecs, left); cur 2390 fs/xfs/libxfs/xfs_btree.c rrp = xfs_btree_rec_addr(cur, 1, right); cur 2392 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_recs(cur, lrp, rrp, 1); cur 2393 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_recs(cur, lbp, lrecs, lrecs); cur 2395 fs/xfs/libxfs/xfs_btree.c ASSERT(cur->bc_ops->recs_inorder(cur, cur 2396 fs/xfs/libxfs/xfs_btree.c xfs_btree_rec_addr(cur, lrecs - 1, left), lrp)); cur 2400 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_block(cur, lbp, XFS_BB_NUMRECS); cur 2403 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_block(cur, rbp, XFS_BB_NUMRECS); cur 2408 fs/xfs/libxfs/xfs_btree.c XFS_BTREE_STATS_ADD(cur, moves, rrecs - 1); cur 2414 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_debug_check_ptr(cur, rpp, i + 1, level); cur 2419 fs/xfs/libxfs/xfs_btree.c xfs_btree_shift_keys(cur, cur 2420 fs/xfs/libxfs/xfs_btree.c xfs_btree_key_addr(cur, 2, right), cur 2422 fs/xfs/libxfs/xfs_btree.c xfs_btree_shift_ptrs(cur, cur 2423 fs/xfs/libxfs/xfs_btree.c xfs_btree_ptr_addr(cur, 2, right), cur 2426 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_keys(cur, rbp, 1, rrecs); cur 2427 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_ptrs(cur, rbp, 1, rrecs); cur 2430 fs/xfs/libxfs/xfs_btree.c xfs_btree_shift_recs(cur, cur 2431 fs/xfs/libxfs/xfs_btree.c xfs_btree_rec_addr(cur, 2, right), cur 2433 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_recs(cur, rbp, 1, rrecs); cur 2440 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_OVERLAPPING) { cur 2441 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_dup_cursor(cur, &tcur); cur 2460 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_update_keys(cur, level); cur 2465 fs/xfs/libxfs/xfs_btree.c cur->bc_ptrs[level]--; cur 2488 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 2504 fs/xfs/libxfs/xfs_btree.c if ((cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) && cur 2505 fs/xfs/libxfs/xfs_btree.c (level == cur->bc_nlevels - 1)) cur 2509 fs/xfs/libxfs/xfs_btree.c left = xfs_btree_get_block(cur, level, &lbp); cur 2512 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_check_block(cur, left, level, lbp); cur 2518 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_sibling(cur, left, &rptr, XFS_BB_RIGHTSIB); cur 2519 fs/xfs/libxfs/xfs_btree.c if (xfs_btree_ptr_is_null(cur, &rptr)) cur 2527 fs/xfs/libxfs/xfs_btree.c if (cur->bc_ptrs[level] >= lrecs) cur 2531 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_read_buf_block(cur, &rptr, 0, &right, &rbp); cur 2537 fs/xfs/libxfs/xfs_btree.c if (rrecs == cur->bc_ops->get_maxrecs(cur, level)) cur 2540 fs/xfs/libxfs/xfs_btree.c XFS_BTREE_STATS_INC(cur, rshift); cur 2541 fs/xfs/libxfs/xfs_btree.c XFS_BTREE_STATS_ADD(cur, moves, rrecs); cur 2553 fs/xfs/libxfs/xfs_btree.c lkp = xfs_btree_key_addr(cur, lrecs, left); cur 2554 fs/xfs/libxfs/xfs_btree.c lpp = xfs_btree_ptr_addr(cur, lrecs, left); cur 2555 fs/xfs/libxfs/xfs_btree.c rkp = xfs_btree_key_addr(cur, 1, right); cur 2556 fs/xfs/libxfs/xfs_btree.c rpp = xfs_btree_ptr_addr(cur, 1, right); cur 2559 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_debug_check_ptr(cur, rpp, i, level); cur 2564 fs/xfs/libxfs/xfs_btree.c xfs_btree_shift_keys(cur, rkp, 1, rrecs); cur 2565 fs/xfs/libxfs/xfs_btree.c xfs_btree_shift_ptrs(cur, rpp, 1, rrecs); cur 2567 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_debug_check_ptr(cur, lpp, 0, level); cur 2572 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_keys(cur, rkp, lkp, 1); cur 2573 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_ptrs(cur, rpp, lpp, 1); cur 2575 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_keys(cur, rbp, 1, rrecs + 1); cur 2576 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_ptrs(cur, rbp, 1, rrecs + 1); cur 2578 fs/xfs/libxfs/xfs_btree.c ASSERT(cur->bc_ops->keys_inorder(cur, rkp, cur 2579 fs/xfs/libxfs/xfs_btree.c xfs_btree_key_addr(cur, 2, right))); cur 2585 fs/xfs/libxfs/xfs_btree.c lrp = xfs_btree_rec_addr(cur, lrecs, left); cur 2586 fs/xfs/libxfs/xfs_btree.c rrp = xfs_btree_rec_addr(cur, 1, right); cur 2588 fs/xfs/libxfs/xfs_btree.c xfs_btree_shift_recs(cur, rrp, 1, rrecs); cur 2591 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_recs(cur, rrp, lrp, 1); cur 2592 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_recs(cur, rbp, 1, rrecs + 1); cur 2599 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_block(cur, lbp, XFS_BB_NUMRECS); cur 2602 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_block(cur, rbp, XFS_BB_NUMRECS); cur 2608 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_dup_cursor(cur, &tcur); cur 2619 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_OVERLAPPING) { cur 2620 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_update_keys(cur, level); cur 2654 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 2676 fs/xfs/libxfs/xfs_btree.c XFS_BTREE_STATS_INC(cur, split); cur 2679 fs/xfs/libxfs/xfs_btree.c left = xfs_btree_get_block(cur, level, &lbp); cur 2682 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_check_block(cur, left, level, lbp); cur 2687 fs/xfs/libxfs/xfs_btree.c xfs_btree_buf_to_ptr(cur, lbp, &lptr); cur 2690 fs/xfs/libxfs/xfs_btree.c error = cur->bc_ops->alloc_block(cur, &lptr, &rptr, stat); cur 2695 fs/xfs/libxfs/xfs_btree.c XFS_BTREE_STATS_INC(cur, alloc); cur 2698 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_get_buf_block(cur, &rptr, &right, &rbp); cur 2703 fs/xfs/libxfs/xfs_btree.c xfs_btree_init_block_cur(cur, rbp, xfs_btree_get_level(left), 0); cur 2712 fs/xfs/libxfs/xfs_btree.c if ((lrecs & 1) && cur->bc_ptrs[level] <= rrecs + 1) cur 2716 fs/xfs/libxfs/xfs_btree.c XFS_BTREE_STATS_ADD(cur, moves, rrecs); cur 2735 fs/xfs/libxfs/xfs_btree.c lkp = xfs_btree_key_addr(cur, src_index, left); cur 2736 fs/xfs/libxfs/xfs_btree.c lpp = xfs_btree_ptr_addr(cur, src_index, left); cur 2737 fs/xfs/libxfs/xfs_btree.c rkp = xfs_btree_key_addr(cur, 1, right); cur 2738 fs/xfs/libxfs/xfs_btree.c rpp = xfs_btree_ptr_addr(cur, 1, right); cur 2741 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_debug_check_ptr(cur, lpp, i, level); cur 2747 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_keys(cur, rkp, lkp, rrecs); cur 2748 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_ptrs(cur, rpp, lpp, rrecs); cur 2750 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_keys(cur, rbp, 1, rrecs); cur 2751 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_ptrs(cur, rbp, 1, rrecs); cur 2754 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_node_keys(cur, right, key); cur 2760 fs/xfs/libxfs/xfs_btree.c lrp = xfs_btree_rec_addr(cur, src_index, left); cur 2761 fs/xfs/libxfs/xfs_btree.c rrp = xfs_btree_rec_addr(cur, 1, right); cur 2764 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_recs(cur, rrp, lrp, rrecs); cur 2765 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_recs(cur, rbp, 1, rrecs); cur 2768 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_leaf_keys(cur, right, key); cur 2775 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_sibling(cur, left, &rrptr, XFS_BB_RIGHTSIB); cur 2776 fs/xfs/libxfs/xfs_btree.c xfs_btree_set_sibling(cur, right, &rrptr, XFS_BB_RIGHTSIB); cur 2777 fs/xfs/libxfs/xfs_btree.c xfs_btree_set_sibling(cur, right, &lptr, XFS_BB_LEFTSIB); cur 2778 fs/xfs/libxfs/xfs_btree.c xfs_btree_set_sibling(cur, left, &rptr, XFS_BB_RIGHTSIB); cur 2780 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_block(cur, rbp, XFS_BB_ALL_BITS); cur 2781 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_block(cur, lbp, XFS_BB_NUMRECS | XFS_BB_RIGHTSIB); cur 2787 fs/xfs/libxfs/xfs_btree.c if (!xfs_btree_ptr_is_null(cur, &rrptr)) { cur 2788 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_read_buf_block(cur, &rrptr, cur 2792 fs/xfs/libxfs/xfs_btree.c xfs_btree_set_sibling(cur, rrblock, &rptr, XFS_BB_LEFTSIB); cur 2793 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_block(cur, rrbp, XFS_BB_LEFTSIB); cur 2797 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_OVERLAPPING) { cur 2798 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_update_keys(cur, level); cur 2808 fs/xfs/libxfs/xfs_btree.c if (cur->bc_ptrs[level] > lrecs + 1) { cur 2809 fs/xfs/libxfs/xfs_btree.c xfs_btree_setbuf(cur, level, rbp); cur 2810 fs/xfs/libxfs/xfs_btree.c cur->bc_ptrs[level] -= lrecs; cur 2816 fs/xfs/libxfs/xfs_btree.c if (level + 1 < cur->bc_nlevels) { cur 2817 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_dup_cursor(cur, curp); cur 2834 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur; cur 2869 fs/xfs/libxfs/xfs_btree.c args->result = __xfs_btree_split(args->cur, args->level, args->ptrp, cur 2883 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 2893 fs/xfs/libxfs/xfs_btree.c if (cur->bc_btnum != XFS_BTNUM_BMAP) cur 2894 fs/xfs/libxfs/xfs_btree.c return __xfs_btree_split(cur, level, ptrp, key, curp, stat); cur 2896 fs/xfs/libxfs/xfs_btree.c args.cur = cur; cur 2918 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, /* btree cursor */ cur 2934 fs/xfs/libxfs/xfs_btree.c XFS_BTREE_STATS_INC(cur, newroot); cur 2936 fs/xfs/libxfs/xfs_btree.c ASSERT(cur->bc_flags & XFS_BTREE_ROOT_IN_INODE); cur 2938 fs/xfs/libxfs/xfs_btree.c level = cur->bc_nlevels - 1; cur 2940 fs/xfs/libxfs/xfs_btree.c block = xfs_btree_get_iroot(cur); cur 2941 fs/xfs/libxfs/xfs_btree.c pp = xfs_btree_ptr_addr(cur, 1, block); cur 2944 fs/xfs/libxfs/xfs_btree.c error = cur->bc_ops->alloc_block(cur, pp, &nptr, stat); cur 2950 fs/xfs/libxfs/xfs_btree.c XFS_BTREE_STATS_INC(cur, alloc); cur 2953 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_get_buf_block(cur, &nptr, &cblock, &cbp); cur 2961 fs/xfs/libxfs/xfs_btree.c memcpy(cblock, block, xfs_btree_block_len(cur)); cur 2962 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_CRC_BLOCKS) { cur 2963 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_LONG_PTRS) cur 2971 fs/xfs/libxfs/xfs_btree.c cur->bc_nlevels++; cur 2972 fs/xfs/libxfs/xfs_btree.c cur->bc_ptrs[level + 1] = 1; cur 2974 fs/xfs/libxfs/xfs_btree.c kp = xfs_btree_key_addr(cur, 1, block); cur 2975 fs/xfs/libxfs/xfs_btree.c ckp = xfs_btree_key_addr(cur, 1, cblock); cur 2976 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_keys(cur, ckp, kp, xfs_btree_get_numrecs(cblock)); cur 2978 fs/xfs/libxfs/xfs_btree.c cpp = xfs_btree_ptr_addr(cur, 1, cblock); cur 2980 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_debug_check_ptr(cur, pp, i, level); cur 2985 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_ptrs(cur, cpp, pp, xfs_btree_get_numrecs(cblock)); cur 2987 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_debug_check_ptr(cur, &nptr, 0, level); cur 2991 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_ptrs(cur, pp, &nptr, 1); cur 2993 fs/xfs/libxfs/xfs_btree.c xfs_iroot_realloc(cur->bc_private.b.ip, cur 2995 fs/xfs/libxfs/xfs_btree.c cur->bc_private.b.whichfork); cur 2997 fs/xfs/libxfs/xfs_btree.c xfs_btree_setbuf(cur, level, cbp); cur 3003 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_block(cur, cbp, XFS_BB_ALL_BITS); cur 3004 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_keys(cur, cbp, 1, be16_to_cpu(cblock->bb_numrecs)); cur 3005 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_ptrs(cur, cbp, 1, be16_to_cpu(cblock->bb_numrecs)); cur 3008 fs/xfs/libxfs/xfs_btree.c XFS_ILOG_CORE | xfs_ilog_fbroot(cur->bc_private.b.whichfork); cur 3020 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, /* btree cursor */ cur 3036 fs/xfs/libxfs/xfs_btree.c XFS_BTREE_STATS_INC(cur, newroot); cur 3039 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->init_ptr_from_cur(cur, &rptr); cur 3042 fs/xfs/libxfs/xfs_btree.c error = cur->bc_ops->alloc_block(cur, &rptr, &lptr, stat); cur 3047 fs/xfs/libxfs/xfs_btree.c XFS_BTREE_STATS_INC(cur, alloc); cur 3050 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_get_buf_block(cur, &lptr, &new, &nbp); cur 3055 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->set_root(cur, &lptr, 1); cur 3063 fs/xfs/libxfs/xfs_btree.c block = xfs_btree_get_block(cur, cur->bc_nlevels - 1, &bp); cur 3066 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_check_block(cur, block, cur->bc_nlevels - 1, bp); cur 3071 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_sibling(cur, block, &rptr, XFS_BB_RIGHTSIB); cur 3072 fs/xfs/libxfs/xfs_btree.c if (!xfs_btree_ptr_is_null(cur, &rptr)) { cur 3075 fs/xfs/libxfs/xfs_btree.c xfs_btree_buf_to_ptr(cur, lbp, &lptr); cur 3077 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_read_buf_block(cur, &rptr, 0, &right, &rbp); cur 3085 fs/xfs/libxfs/xfs_btree.c xfs_btree_buf_to_ptr(cur, rbp, &rptr); cur 3087 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_sibling(cur, right, &lptr, XFS_BB_LEFTSIB); cur 3088 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_read_buf_block(cur, &lptr, 0, &left, &lbp); cur 3096 fs/xfs/libxfs/xfs_btree.c xfs_btree_init_block_cur(cur, nbp, cur->bc_nlevels, 2); cur 3097 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_block(cur, nbp, XFS_BB_ALL_BITS); cur 3098 fs/xfs/libxfs/xfs_btree.c ASSERT(!xfs_btree_ptr_is_null(cur, &lptr) && cur 3099 fs/xfs/libxfs/xfs_btree.c !xfs_btree_ptr_is_null(cur, &rptr)); cur 3107 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_node_keys(cur, left, cur 3108 fs/xfs/libxfs/xfs_btree.c xfs_btree_key_addr(cur, 1, new)); cur 3109 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_node_keys(cur, right, cur 3110 fs/xfs/libxfs/xfs_btree.c xfs_btree_key_addr(cur, 2, new)); cur 3117 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_leaf_keys(cur, left, cur 3118 fs/xfs/libxfs/xfs_btree.c xfs_btree_key_addr(cur, 1, new)); cur 3119 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_leaf_keys(cur, right, cur 3120 fs/xfs/libxfs/xfs_btree.c xfs_btree_key_addr(cur, 2, new)); cur 3122 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_keys(cur, nbp, 1, 2); cur 3125 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_ptrs(cur, cur 3126 fs/xfs/libxfs/xfs_btree.c xfs_btree_ptr_addr(cur, 1, new), &lptr, 1); cur 3127 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_ptrs(cur, cur 3128 fs/xfs/libxfs/xfs_btree.c xfs_btree_ptr_addr(cur, 2, new), &rptr, 1); cur 3129 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_ptrs(cur, nbp, 1, 2); cur 3132 fs/xfs/libxfs/xfs_btree.c xfs_btree_setbuf(cur, cur->bc_nlevels, nbp); cur 3133 fs/xfs/libxfs/xfs_btree.c cur->bc_ptrs[cur->bc_nlevels] = nptr; cur 3134 fs/xfs/libxfs/xfs_btree.c cur->bc_nlevels++; cur 3146 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, /* btree cursor */ cur 3158 fs/xfs/libxfs/xfs_btree.c if ((cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) && cur 3159 fs/xfs/libxfs/xfs_btree.c level == cur->bc_nlevels - 1) { cur 3160 fs/xfs/libxfs/xfs_btree.c struct xfs_inode *ip = cur->bc_private.b.ip; cur 3162 fs/xfs/libxfs/xfs_btree.c if (numrecs < cur->bc_ops->get_dmaxrecs(cur, level)) { cur 3164 fs/xfs/libxfs/xfs_btree.c xfs_iroot_realloc(ip, 1, cur->bc_private.b.whichfork); cur 3170 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_new_iroot(cur, &logflags, stat); cur 3174 fs/xfs/libxfs/xfs_btree.c xfs_trans_log_inode(cur->bc_tp, ip, logflags); cur 3181 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_rshift(cur, level, stat); cur 3186 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_lshift(cur, level, stat); cur 3191 fs/xfs/libxfs/xfs_btree.c *oindex = *index = cur->bc_ptrs[level]; cur 3201 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_split(cur, level, nptr, key, ncur, stat); cur 3206 fs/xfs/libxfs/xfs_btree.c *index = cur->bc_ptrs[level]; cur 3216 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, /* btree cursor */ cur 3244 fs/xfs/libxfs/xfs_btree.c if (!(cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) && cur 3245 fs/xfs/libxfs/xfs_btree.c (level >= cur->bc_nlevels)) { cur 3246 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_new_root(cur, stat); cur 3247 fs/xfs/libxfs/xfs_btree.c xfs_btree_set_ptr_null(cur, ptrp); cur 3253 fs/xfs/libxfs/xfs_btree.c ptr = cur->bc_ptrs[level]; cur 3261 fs/xfs/libxfs/xfs_btree.c XFS_BTREE_STATS_INC(cur, insrec); cur 3264 fs/xfs/libxfs/xfs_btree.c block = xfs_btree_get_block(cur, level, &bp); cur 3269 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_check_block(cur, block, level, bp); cur 3276 fs/xfs/libxfs/xfs_btree.c ASSERT(cur->bc_ops->recs_inorder(cur, rec, cur 3277 fs/xfs/libxfs/xfs_btree.c xfs_btree_rec_addr(cur, ptr, block))); cur 3279 fs/xfs/libxfs/xfs_btree.c ASSERT(cur->bc_ops->keys_inorder(cur, key, cur 3280 fs/xfs/libxfs/xfs_btree.c xfs_btree_key_addr(cur, ptr, block))); cur 3289 fs/xfs/libxfs/xfs_btree.c xfs_btree_set_ptr_null(cur, &nptr); cur 3290 fs/xfs/libxfs/xfs_btree.c if (numrecs == cur->bc_ops->get_maxrecs(cur, level)) { cur 3291 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_make_block_unfull(cur, level, numrecs, cur 3301 fs/xfs/libxfs/xfs_btree.c block = xfs_btree_get_block(cur, level, &bp); cur 3305 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_check_block(cur, block, level, bp); cur 3314 fs/xfs/libxfs/xfs_btree.c XFS_BTREE_STATS_ADD(cur, moves, numrecs - ptr + 1); cur 3321 fs/xfs/libxfs/xfs_btree.c kp = xfs_btree_key_addr(cur, ptr, block); cur 3322 fs/xfs/libxfs/xfs_btree.c pp = xfs_btree_ptr_addr(cur, ptr, block); cur 3325 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_debug_check_ptr(cur, pp, i, level); cur 3330 fs/xfs/libxfs/xfs_btree.c xfs_btree_shift_keys(cur, kp, 1, numrecs - ptr + 1); cur 3331 fs/xfs/libxfs/xfs_btree.c xfs_btree_shift_ptrs(cur, pp, 1, numrecs - ptr + 1); cur 3333 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_debug_check_ptr(cur, ptrp, 0, level); cur 3338 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_keys(cur, kp, key, 1); cur 3339 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_ptrs(cur, pp, ptrp, 1); cur 3342 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_ptrs(cur, bp, ptr, numrecs); cur 3343 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_keys(cur, bp, ptr, numrecs); cur 3346 fs/xfs/libxfs/xfs_btree.c ASSERT(cur->bc_ops->keys_inorder(cur, kp, cur 3347 fs/xfs/libxfs/xfs_btree.c xfs_btree_key_addr(cur, ptr + 1, block))); cur 3354 fs/xfs/libxfs/xfs_btree.c rp = xfs_btree_rec_addr(cur, ptr, block); cur 3356 fs/xfs/libxfs/xfs_btree.c xfs_btree_shift_recs(cur, rp, 1, numrecs - ptr + 1); cur 3359 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_recs(cur, rp, rec, 1); cur 3361 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_recs(cur, bp, ptr, numrecs); cur 3364 fs/xfs/libxfs/xfs_btree.c ASSERT(cur->bc_ops->recs_inorder(cur, rp, cur 3365 fs/xfs/libxfs/xfs_btree.c xfs_btree_rec_addr(cur, ptr + 1, block))); cur 3371 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_block(cur, bp, XFS_BB_NUMRECS); cur 3382 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_keys(cur, block, lkey); cur 3383 fs/xfs/libxfs/xfs_btree.c } else if (xfs_btree_needs_key_update(cur, optr)) { cur 3384 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_update_keys(cur, level); cur 3393 fs/xfs/libxfs/xfs_btree.c if (xfs_btree_is_lastrec(cur, block, level)) { cur 3394 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->update_lastrec(cur, block, rec, cur 3403 fs/xfs/libxfs/xfs_btree.c if (!xfs_btree_ptr_is_null(cur, &nptr)) { cur 3404 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_keys(cur, key, lkey, 1); cur 3424 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 3439 fs/xfs/libxfs/xfs_btree.c pcur = cur; cur 3442 fs/xfs/libxfs/xfs_btree.c xfs_btree_set_ptr_null(cur, &nptr); cur 3445 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->init_rec_from_cur(cur, &rec); cur 3446 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->init_key_from_rec(key, &rec); cur 3461 fs/xfs/libxfs/xfs_btree.c if (pcur != cur) cur 3466 fs/xfs/libxfs/xfs_btree.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, i == 1, error0); cur 3474 fs/xfs/libxfs/xfs_btree.c if (pcur != cur && cur 3475 fs/xfs/libxfs/xfs_btree.c (ncur || xfs_btree_ptr_is_null(cur, &nptr))) { cur 3477 fs/xfs/libxfs/xfs_btree.c if (cur->bc_ops->update_cursor) cur 3478 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->update_cursor(pcur, cur); cur 3479 fs/xfs/libxfs/xfs_btree.c cur->bc_nlevels = pcur->bc_nlevels; cur 3487 fs/xfs/libxfs/xfs_btree.c } while (!xfs_btree_ptr_is_null(cur, &nptr)); cur 3505 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur) cur 3507 fs/xfs/libxfs/xfs_btree.c int whichfork = cur->bc_private.b.whichfork; cur 3508 fs/xfs/libxfs/xfs_btree.c struct xfs_inode *ip = cur->bc_private.b.ip; cur 3526 fs/xfs/libxfs/xfs_btree.c ASSERT(cur->bc_flags & XFS_BTREE_ROOT_IN_INODE); cur 3527 fs/xfs/libxfs/xfs_btree.c ASSERT(cur->bc_nlevels > 1); cur 3533 fs/xfs/libxfs/xfs_btree.c level = cur->bc_nlevels - 1; cur 3540 fs/xfs/libxfs/xfs_btree.c block = xfs_btree_get_iroot(cur); cur 3544 fs/xfs/libxfs/xfs_btree.c cblock = xfs_btree_get_block(cur, level - 1, &cbp); cur 3552 fs/xfs/libxfs/xfs_btree.c if (numrecs > cur->bc_ops->get_dmaxrecs(cur, level)) cur 3555 fs/xfs/libxfs/xfs_btree.c XFS_BTREE_STATS_INC(cur, killroot); cur 3558 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_sibling(cur, block, &ptr, XFS_BB_LEFTSIB); cur 3559 fs/xfs/libxfs/xfs_btree.c ASSERT(xfs_btree_ptr_is_null(cur, &ptr)); cur 3560 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_sibling(cur, block, &ptr, XFS_BB_RIGHTSIB); cur 3561 fs/xfs/libxfs/xfs_btree.c ASSERT(xfs_btree_ptr_is_null(cur, &ptr)); cur 3564 fs/xfs/libxfs/xfs_btree.c index = numrecs - cur->bc_ops->get_maxrecs(cur, level); cur 3566 fs/xfs/libxfs/xfs_btree.c xfs_iroot_realloc(cur->bc_private.b.ip, index, cur 3567 fs/xfs/libxfs/xfs_btree.c cur->bc_private.b.whichfork); cur 3574 fs/xfs/libxfs/xfs_btree.c kp = xfs_btree_key_addr(cur, 1, block); cur 3575 fs/xfs/libxfs/xfs_btree.c ckp = xfs_btree_key_addr(cur, 1, cblock); cur 3576 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_keys(cur, kp, ckp, numrecs); cur 3578 fs/xfs/libxfs/xfs_btree.c pp = xfs_btree_ptr_addr(cur, 1, block); cur 3579 fs/xfs/libxfs/xfs_btree.c cpp = xfs_btree_ptr_addr(cur, 1, cblock); cur 3582 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_debug_check_ptr(cur, cpp, i, level - 1); cur 3587 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_ptrs(cur, pp, cpp, numrecs); cur 3589 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_free_block(cur, cbp); cur 3593 fs/xfs/libxfs/xfs_btree.c cur->bc_bufs[level - 1] = NULL; cur 3595 fs/xfs/libxfs/xfs_btree.c xfs_trans_log_inode(cur->bc_tp, ip, cur 3596 fs/xfs/libxfs/xfs_btree.c XFS_ILOG_CORE | xfs_ilog_fbroot(cur->bc_private.b.whichfork)); cur 3597 fs/xfs/libxfs/xfs_btree.c cur->bc_nlevels--; cur 3607 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 3614 fs/xfs/libxfs/xfs_btree.c XFS_BTREE_STATS_INC(cur, killroot); cur 3620 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->set_root(cur, newroot, -1); cur 3622 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_free_block(cur, bp); cur 3626 fs/xfs/libxfs/xfs_btree.c cur->bc_bufs[level] = NULL; cur 3627 fs/xfs/libxfs/xfs_btree.c cur->bc_ra[level] = 0; cur 3628 fs/xfs/libxfs/xfs_btree.c cur->bc_nlevels--; cur 3635 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 3643 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_decrement(cur, level, &i); cur 3660 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, /* btree cursor */ cur 3686 fs/xfs/libxfs/xfs_btree.c ptr = cur->bc_ptrs[level]; cur 3693 fs/xfs/libxfs/xfs_btree.c block = xfs_btree_get_block(cur, level, &bp); cur 3697 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_check_block(cur, block, level, bp); cur 3708 fs/xfs/libxfs/xfs_btree.c XFS_BTREE_STATS_INC(cur, delrec); cur 3709 fs/xfs/libxfs/xfs_btree.c XFS_BTREE_STATS_ADD(cur, moves, numrecs - ptr); cur 3717 fs/xfs/libxfs/xfs_btree.c lkp = xfs_btree_key_addr(cur, ptr + 1, block); cur 3718 fs/xfs/libxfs/xfs_btree.c lpp = xfs_btree_ptr_addr(cur, ptr + 1, block); cur 3721 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_debug_check_ptr(cur, lpp, i, level); cur 3727 fs/xfs/libxfs/xfs_btree.c xfs_btree_shift_keys(cur, lkp, -1, numrecs - ptr); cur 3728 fs/xfs/libxfs/xfs_btree.c xfs_btree_shift_ptrs(cur, lpp, -1, numrecs - ptr); cur 3729 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_keys(cur, bp, ptr, numrecs - 1); cur 3730 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_ptrs(cur, bp, ptr, numrecs - 1); cur 3735 fs/xfs/libxfs/xfs_btree.c xfs_btree_shift_recs(cur, cur 3736 fs/xfs/libxfs/xfs_btree.c xfs_btree_rec_addr(cur, ptr + 1, block), cur 3738 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_recs(cur, bp, ptr, numrecs - 1); cur 3746 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_block(cur, bp, XFS_BB_NUMRECS); cur 3752 fs/xfs/libxfs/xfs_btree.c if (xfs_btree_is_lastrec(cur, block, level)) { cur 3753 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->update_lastrec(cur, block, NULL, cur 3762 fs/xfs/libxfs/xfs_btree.c if (level == cur->bc_nlevels - 1) { cur 3763 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) { cur 3764 fs/xfs/libxfs/xfs_btree.c xfs_iroot_realloc(cur->bc_private.b.ip, -1, cur 3765 fs/xfs/libxfs/xfs_btree.c cur->bc_private.b.whichfork); cur 3767 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_kill_iroot(cur); cur 3771 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_dec_cursor(cur, level, stat); cur 3789 fs/xfs/libxfs/xfs_btree.c pp = xfs_btree_ptr_addr(cur, 1, block); cur 3790 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_kill_root(cur, bp, level, pp); cur 3794 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_dec_cursor(cur, level, stat); cur 3806 fs/xfs/libxfs/xfs_btree.c if (xfs_btree_needs_key_update(cur, ptr)) { cur 3807 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_update_keys(cur, level); cur 3816 fs/xfs/libxfs/xfs_btree.c if (numrecs >= cur->bc_ops->get_minrecs(cur, level)) { cur 3817 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_dec_cursor(cur, level, stat); cur 3828 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_sibling(cur, block, &rptr, XFS_BB_RIGHTSIB); cur 3829 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_sibling(cur, block, &lptr, XFS_BB_LEFTSIB); cur 3831 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) { cur 3837 fs/xfs/libxfs/xfs_btree.c if (xfs_btree_ptr_is_null(cur, &rptr) && cur 3838 fs/xfs/libxfs/xfs_btree.c xfs_btree_ptr_is_null(cur, &lptr) && cur 3839 fs/xfs/libxfs/xfs_btree.c level == cur->bc_nlevels - 2) { cur 3840 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_kill_iroot(cur); cur 3842 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_dec_cursor(cur, level, stat); cur 3849 fs/xfs/libxfs/xfs_btree.c ASSERT(!xfs_btree_ptr_is_null(cur, &rptr) || cur 3850 fs/xfs/libxfs/xfs_btree.c !xfs_btree_ptr_is_null(cur, &lptr)); cur 3856 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_dup_cursor(cur, &tcur); cur 3864 fs/xfs/libxfs/xfs_btree.c if (!xfs_btree_ptr_is_null(cur, &rptr)) { cur 3870 fs/xfs/libxfs/xfs_btree.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, i == 1, error0); cur 3875 fs/xfs/libxfs/xfs_btree.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, i == 1, error0); cur 3878 fs/xfs/libxfs/xfs_btree.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, i == 1, error0); cur 3896 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->get_minrecs(tcur, level)) { cur 3902 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->get_minrecs(tcur, level)); cur 3907 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_dec_cursor(cur, level, stat); cur 3920 fs/xfs/libxfs/xfs_btree.c if (!xfs_btree_ptr_is_null(cur, &lptr)) { cur 3922 fs/xfs/libxfs/xfs_btree.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, i == 1, error0); cur 3927 fs/xfs/libxfs/xfs_btree.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, i == 1, error0); cur 3935 fs/xfs/libxfs/xfs_btree.c if (!xfs_btree_ptr_is_null(cur, &lptr)) { cur 3941 fs/xfs/libxfs/xfs_btree.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, i == 1, error0); cur 3947 fs/xfs/libxfs/xfs_btree.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, i == 1, error0); cur 3952 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_check_block(cur, left, level, lbp); cur 3965 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->get_minrecs(tcur, level)) { cur 3971 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->get_minrecs(tcur, level)); cur 3975 fs/xfs/libxfs/xfs_btree.c cur->bc_ptrs[0]++; cur 3994 fs/xfs/libxfs/xfs_btree.c ASSERT(!xfs_btree_ptr_is_null(cur, &cptr)); cur 3996 fs/xfs/libxfs/xfs_btree.c if (!xfs_btree_ptr_is_null(cur, &lptr) && cur 3998 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->get_maxrecs(cur, level)) { cur 4006 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_read_buf_block(cur, &lptr, 0, &left, &lbp); cur 4013 fs/xfs/libxfs/xfs_btree.c } else if (!xfs_btree_ptr_is_null(cur, &rptr) && cur 4015 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->get_maxrecs(cur, level)) { cur 4023 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_read_buf_block(cur, &rptr, 0, &right, &rbp); cur 4032 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_dec_cursor(cur, level, stat); cur 4045 fs/xfs/libxfs/xfs_btree.c XFS_BTREE_STATS_ADD(cur, moves, rrecs); cur 4053 fs/xfs/libxfs/xfs_btree.c lkp = xfs_btree_key_addr(cur, lrecs + 1, left); cur 4054 fs/xfs/libxfs/xfs_btree.c lpp = xfs_btree_ptr_addr(cur, lrecs + 1, left); cur 4055 fs/xfs/libxfs/xfs_btree.c rkp = xfs_btree_key_addr(cur, 1, right); cur 4056 fs/xfs/libxfs/xfs_btree.c rpp = xfs_btree_ptr_addr(cur, 1, right); cur 4059 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_debug_check_ptr(cur, rpp, i, level); cur 4064 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_keys(cur, lkp, rkp, rrecs); cur 4065 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_ptrs(cur, lpp, rpp, rrecs); cur 4067 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_keys(cur, lbp, lrecs + 1, lrecs + rrecs); cur 4068 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_ptrs(cur, lbp, lrecs + 1, lrecs + rrecs); cur 4074 fs/xfs/libxfs/xfs_btree.c lrp = xfs_btree_rec_addr(cur, lrecs + 1, left); cur 4075 fs/xfs/libxfs/xfs_btree.c rrp = xfs_btree_rec_addr(cur, 1, right); cur 4077 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_recs(cur, lrp, rrp, rrecs); cur 4078 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_recs(cur, lbp, lrecs + 1, lrecs + rrecs); cur 4081 fs/xfs/libxfs/xfs_btree.c XFS_BTREE_STATS_INC(cur, join); cur 4088 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_sibling(cur, right, &cptr, XFS_BB_RIGHTSIB), cur 4089 fs/xfs/libxfs/xfs_btree.c xfs_btree_set_sibling(cur, left, &cptr, XFS_BB_RIGHTSIB); cur 4090 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_block(cur, lbp, XFS_BB_NUMRECS | XFS_BB_RIGHTSIB); cur 4093 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_sibling(cur, left, &cptr, XFS_BB_RIGHTSIB); cur 4094 fs/xfs/libxfs/xfs_btree.c if (!xfs_btree_ptr_is_null(cur, &cptr)) { cur 4095 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_read_buf_block(cur, &cptr, 0, &rrblock, &rrbp); cur 4098 fs/xfs/libxfs/xfs_btree.c xfs_btree_set_sibling(cur, rrblock, &lptr, XFS_BB_LEFTSIB); cur 4099 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_block(cur, rrbp, XFS_BB_LEFTSIB); cur 4103 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_free_block(cur, rbp); cur 4112 fs/xfs/libxfs/xfs_btree.c cur->bc_bufs[level] = lbp; cur 4113 fs/xfs/libxfs/xfs_btree.c cur->bc_ptrs[level] += lrecs; cur 4114 fs/xfs/libxfs/xfs_btree.c cur->bc_ra[level] = 0; cur 4120 fs/xfs/libxfs/xfs_btree.c else if ((cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) || cur 4121 fs/xfs/libxfs/xfs_btree.c (level + 1 < cur->bc_nlevels)) { cur 4122 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_increment(cur, level + 1, &i); cur 4134 fs/xfs/libxfs/xfs_btree.c cur->bc_ptrs[level]--; cur 4163 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 4178 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_delrec(cur, level, &i); cur 4189 fs/xfs/libxfs/xfs_btree.c if (joined && (cur->bc_flags & XFS_BTREE_OVERLAPPING)) { cur 4190 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_updkeys_force(cur, 0); cur 4196 fs/xfs/libxfs/xfs_btree.c for (level = 1; level < cur->bc_nlevels; level++) { cur 4197 fs/xfs/libxfs/xfs_btree.c if (cur->bc_ptrs[level] == 0) { cur 4198 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_decrement(cur, level, &i); cur 4217 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, /* btree cursor */ cur 4228 fs/xfs/libxfs/xfs_btree.c ptr = cur->bc_ptrs[0]; cur 4229 fs/xfs/libxfs/xfs_btree.c block = xfs_btree_get_block(cur, 0, &bp); cur 4232 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_check_block(cur, block, 0, bp); cur 4248 fs/xfs/libxfs/xfs_btree.c *recp = xfs_btree_rec_addr(cur, ptr, block); cur 4256 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 4267 fs/xfs/libxfs/xfs_btree.c xfs_btree_readahead(cur, level, XFS_BTCUR_RIGHTRA); cur 4268 fs/xfs/libxfs/xfs_btree.c block = xfs_btree_get_block(cur, level, &bp); cur 4271 fs/xfs/libxfs/xfs_btree.c error = fn(cur, level, data); cur 4276 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_sibling(cur, block, &rptr, XFS_BB_RIGHTSIB); cur 4277 fs/xfs/libxfs/xfs_btree.c if (xfs_btree_ptr_is_null(cur, &rptr)) cur 4280 fs/xfs/libxfs/xfs_btree.c return xfs_btree_lookup_get_block(cur, level, &rptr, &block); cur 4287 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 4296 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->init_ptr_from_cur(cur, &lptr); cur 4299 fs/xfs/libxfs/xfs_btree.c for (level = cur->bc_nlevels - 1; level >= 0; level--) { cur 4301 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_lookup_get_block(cur, level, &lptr, &block); cur 4309 fs/xfs/libxfs/xfs_btree.c ptr = xfs_btree_ptr_addr(cur, 1, block); cur 4310 fs/xfs/libxfs/xfs_btree.c xfs_btree_readahead_ptr(cur, ptr, 1); cur 4313 fs/xfs/libxfs/xfs_btree.c xfs_btree_copy_ptrs(cur, &lptr, ptr, 1); cur 4318 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_visit_block(cur, level, fn, data); cur 4359 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 4368 fs/xfs/libxfs/xfs_btree.c block = xfs_btree_get_block(cur, level, &bp); cur 4369 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_LONG_PTRS) { cur 4387 fs/xfs/libxfs/xfs_btree.c ASSERT(cur->bc_flags & XFS_BTREE_ROOT_IN_INODE); cur 4388 fs/xfs/libxfs/xfs_btree.c ASSERT(level == cur->bc_nlevels - 1); cur 4392 fs/xfs/libxfs/xfs_btree.c if (cur->bc_tp) { cur 4393 fs/xfs/libxfs/xfs_btree.c if (!xfs_trans_ordered_buf(cur->bc_tp, bp)) { cur 4394 fs/xfs/libxfs/xfs_btree.c xfs_btree_log_block(cur, bp, XFS_BB_OWNER); cur 4406 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 4415 fs/xfs/libxfs/xfs_btree.c return xfs_btree_visit_blocks(cur, xfs_btree_block_change_owner, cur 4545 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 4558 fs/xfs/libxfs/xfs_btree.c ASSERT(cur->bc_ops->init_high_key_from_rec); cur 4559 fs/xfs/libxfs/xfs_btree.c ASSERT(cur->bc_ops->diff_two_keys); cur 4566 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_lookup(cur, XFS_LOOKUP_LE, &stat); cur 4572 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_increment(cur, 0, &stat); cur 4579 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_get_rec(cur, &recp, &stat); cur 4585 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->init_high_key_from_rec(&rec_key, recp); cur 4587 fs/xfs/libxfs/xfs_btree.c diff = cur->bc_ops->diff_two_keys(cur, low_key, cur 4594 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->init_key_from_rec(&rec_key, recp); cur 4595 fs/xfs/libxfs/xfs_btree.c diff = cur->bc_ops->diff_two_keys(cur, &rec_key, high_key); cur 4600 fs/xfs/libxfs/xfs_btree.c error = fn(cur, recp, priv); cur 4606 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_increment(cur, 0, &stat); cur 4636 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 4658 fs/xfs/libxfs/xfs_btree.c level = cur->bc_nlevels - 1; cur 4659 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->init_ptr_from_cur(cur, &ptr); cur 4660 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_lookup_get_block(cur, level, &ptr, &block); cur 4663 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_block(cur, level, &bp); cur 4664 fs/xfs/libxfs/xfs_btree.c trace_xfs_btree_overlapped_query_range(cur, level, bp); cur 4666 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_check_block(cur, block, level, bp); cur 4670 fs/xfs/libxfs/xfs_btree.c cur->bc_ptrs[level] = 1; cur 4672 fs/xfs/libxfs/xfs_btree.c while (level < cur->bc_nlevels) { cur 4673 fs/xfs/libxfs/xfs_btree.c block = xfs_btree_get_block(cur, level, &bp); cur 4676 fs/xfs/libxfs/xfs_btree.c if (cur->bc_ptrs[level] > be16_to_cpu(block->bb_numrecs)) { cur 4678 fs/xfs/libxfs/xfs_btree.c if (level < cur->bc_nlevels - 1) cur 4679 fs/xfs/libxfs/xfs_btree.c cur->bc_ptrs[level + 1]++; cur 4686 fs/xfs/libxfs/xfs_btree.c recp = xfs_btree_rec_addr(cur, cur->bc_ptrs[0], block); cur 4688 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->init_high_key_from_rec(&rec_hkey, recp); cur 4689 fs/xfs/libxfs/xfs_btree.c ldiff = cur->bc_ops->diff_two_keys(cur, &rec_hkey, cur 4692 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->init_key_from_rec(&rec_key, recp); cur 4693 fs/xfs/libxfs/xfs_btree.c hdiff = cur->bc_ops->diff_two_keys(cur, high_key, cur 4702 fs/xfs/libxfs/xfs_btree.c error = fn(cur, recp, priv); cur 4709 fs/xfs/libxfs/xfs_btree.c cur->bc_ptrs[level]++; cur 4714 fs/xfs/libxfs/xfs_btree.c lkp = xfs_btree_key_addr(cur, cur->bc_ptrs[level], block); cur 4715 fs/xfs/libxfs/xfs_btree.c hkp = xfs_btree_high_key_addr(cur, cur->bc_ptrs[level], block); cur 4716 fs/xfs/libxfs/xfs_btree.c pp = xfs_btree_ptr_addr(cur, cur->bc_ptrs[level], block); cur 4718 fs/xfs/libxfs/xfs_btree.c ldiff = cur->bc_ops->diff_two_keys(cur, hkp, low_key); cur 4719 fs/xfs/libxfs/xfs_btree.c hdiff = cur->bc_ops->diff_two_keys(cur, high_key, lkp); cur 4728 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_lookup_get_block(cur, level, pp, cur 4732 fs/xfs/libxfs/xfs_btree.c xfs_btree_get_block(cur, level, &bp); cur 4733 fs/xfs/libxfs/xfs_btree.c trace_xfs_btree_overlapped_query_range(cur, level, bp); cur 4735 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_check_block(cur, block, level, bp); cur 4739 fs/xfs/libxfs/xfs_btree.c cur->bc_ptrs[level] = 1; cur 4745 fs/xfs/libxfs/xfs_btree.c cur->bc_ptrs[level]++; cur 4756 fs/xfs/libxfs/xfs_btree.c if (cur->bc_bufs[0] == NULL) { cur 4757 fs/xfs/libxfs/xfs_btree.c for (i = 0; i < cur->bc_nlevels; i++) { cur 4758 fs/xfs/libxfs/xfs_btree.c if (cur->bc_bufs[i]) { cur 4759 fs/xfs/libxfs/xfs_btree.c xfs_trans_brelse(cur->bc_tp, cur->bc_bufs[i]); cur 4760 fs/xfs/libxfs/xfs_btree.c cur->bc_bufs[i] = NULL; cur 4761 fs/xfs/libxfs/xfs_btree.c cur->bc_ptrs[i] = 0; cur 4762 fs/xfs/libxfs/xfs_btree.c cur->bc_ra[i] = 0; cur 4778 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 4789 fs/xfs/libxfs/xfs_btree.c cur->bc_rec = *high_rec; cur 4790 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->init_rec_from_cur(cur, &rec); cur 4791 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->init_key_from_rec(&high_key, &rec); cur 4793 fs/xfs/libxfs/xfs_btree.c cur->bc_rec = *low_rec; cur 4794 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->init_rec_from_cur(cur, &rec); cur 4795 fs/xfs/libxfs/xfs_btree.c cur->bc_ops->init_key_from_rec(&low_key, &rec); cur 4798 fs/xfs/libxfs/xfs_btree.c if (cur->bc_ops->diff_two_keys(cur, &low_key, &high_key) > 0) cur 4801 fs/xfs/libxfs/xfs_btree.c if (!(cur->bc_flags & XFS_BTREE_OVERLAPPING)) cur 4802 fs/xfs/libxfs/xfs_btree.c return xfs_btree_simple_query_range(cur, &low_key, cur 4804 fs/xfs/libxfs/xfs_btree.c return xfs_btree_overlapped_query_range(cur, &low_key, &high_key, cur 4811 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 4818 fs/xfs/libxfs/xfs_btree.c memset(&cur->bc_rec, 0, sizeof(cur->bc_rec)); cur 4822 fs/xfs/libxfs/xfs_btree.c return xfs_btree_simple_query_range(cur, &low_key, &high_key, fn, priv); cur 4850 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 4863 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 4867 fs/xfs/libxfs/xfs_btree.c return xfs_btree_visit_blocks(cur, xfs_btree_count_blocks_helper, cur 4874 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 4878 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_LONG_PTRS) cur 4886 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 4896 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur, cur 4903 fs/xfs/libxfs/xfs_btree.c error = xfs_btree_query_range(cur, low, high, cur 4916 fs/xfs/libxfs/xfs_btree.c struct xfs_btree_cur *cur) cur 4921 fs/xfs/libxfs/xfs_btree.c block = xfs_btree_get_block(cur, 0, &bp); cur 4924 fs/xfs/libxfs/xfs_btree.c if (cur->bc_ptrs[0] < xfs_btree_get_numrecs(block)) cur 4928 fs/xfs/libxfs/xfs_btree.c if (cur->bc_flags & XFS_BTREE_LONG_PTRS) cur 88 fs/xfs/libxfs/xfs_btree.h #define XFS_BTREE_STATS_INC(cur, stat) \ cur 89 fs/xfs/libxfs/xfs_btree.h XFS_STATS_INC_OFF((cur)->bc_mp, (cur)->bc_statoff + __XBTS_ ## stat) cur 90 fs/xfs/libxfs/xfs_btree.h #define XFS_BTREE_STATS_ADD(cur, stat, val) \ cur 91 fs/xfs/libxfs/xfs_btree.h XFS_STATS_ADD_OFF((cur)->bc_mp, (cur)->bc_statoff + __XBTS_ ## stat, val) cur 106 fs/xfs/libxfs/xfs_btree.h void (*set_root)(struct xfs_btree_cur *cur, cur 110 fs/xfs/libxfs/xfs_btree.h int (*alloc_block)(struct xfs_btree_cur *cur, cur 114 fs/xfs/libxfs/xfs_btree.h int (*free_block)(struct xfs_btree_cur *cur, struct xfs_buf *bp); cur 117 fs/xfs/libxfs/xfs_btree.h void (*update_lastrec)(struct xfs_btree_cur *cur, cur 123 fs/xfs/libxfs/xfs_btree.h int (*get_minrecs)(struct xfs_btree_cur *cur, int level); cur 124 fs/xfs/libxfs/xfs_btree.h int (*get_maxrecs)(struct xfs_btree_cur *cur, int level); cur 127 fs/xfs/libxfs/xfs_btree.h int (*get_dmaxrecs)(struct xfs_btree_cur *cur, int level); cur 132 fs/xfs/libxfs/xfs_btree.h void (*init_rec_from_cur)(struct xfs_btree_cur *cur, cur 134 fs/xfs/libxfs/xfs_btree.h void (*init_ptr_from_cur)(struct xfs_btree_cur *cur, cur 140 fs/xfs/libxfs/xfs_btree.h int64_t (*key_diff)(struct xfs_btree_cur *cur, cur 147 fs/xfs/libxfs/xfs_btree.h int64_t (*diff_two_keys)(struct xfs_btree_cur *cur, cur 154 fs/xfs/libxfs/xfs_btree.h int (*keys_inorder)(struct xfs_btree_cur *cur, cur 159 fs/xfs/libxfs/xfs_btree.h int (*recs_inorder)(struct xfs_btree_cur *cur, cur 246 fs/xfs/libxfs/xfs_btree.h xfs_failaddr_t __xfs_btree_check_lblock(struct xfs_btree_cur *cur, cur 248 fs/xfs/libxfs/xfs_btree.h xfs_failaddr_t __xfs_btree_check_sblock(struct xfs_btree_cur *cur, cur 256 fs/xfs/libxfs/xfs_btree.h struct xfs_btree_cur *cur, /* btree cursor */ cur 266 fs/xfs/libxfs/xfs_btree.h struct xfs_btree_cur *cur, /* btree cursor */ cur 275 fs/xfs/libxfs/xfs_btree.h struct xfs_btree_cur *cur, /* btree cursor */ cur 284 fs/xfs/libxfs/xfs_btree.h xfs_btree_cur_t *cur, /* btree cursor */ cur 293 fs/xfs/libxfs/xfs_btree.h xfs_btree_cur_t *cur, /* input cursor */ cur 322 fs/xfs/libxfs/xfs_btree.h xfs_btree_cur_t *cur, /* btree cursor */ cur 407 fs/xfs/libxfs/xfs_btree.h int xfs_btree_change_owner(struct xfs_btree_cur *cur, uint64_t new_owner, cur 474 fs/xfs/libxfs/xfs_btree.h typedef int (*xfs_btree_query_range_fn)(struct xfs_btree_cur *cur, cur 477 fs/xfs/libxfs/xfs_btree.h int xfs_btree_query_range(struct xfs_btree_cur *cur, cur 480 fs/xfs/libxfs/xfs_btree.h int xfs_btree_query_all(struct xfs_btree_cur *cur, xfs_btree_query_range_fn fn, cur 483 fs/xfs/libxfs/xfs_btree.h typedef int (*xfs_btree_visit_blocks_fn)(struct xfs_btree_cur *cur, int level, cur 485 fs/xfs/libxfs/xfs_btree.h int xfs_btree_visit_blocks(struct xfs_btree_cur *cur, cur 488 fs/xfs/libxfs/xfs_btree.h int xfs_btree_count_blocks(struct xfs_btree_cur *cur, xfs_extlen_t *blocks); cur 490 fs/xfs/libxfs/xfs_btree.h union xfs_btree_rec *xfs_btree_rec_addr(struct xfs_btree_cur *cur, int n, cur 492 fs/xfs/libxfs/xfs_btree.h union xfs_btree_key *xfs_btree_key_addr(struct xfs_btree_cur *cur, int n, cur 494 fs/xfs/libxfs/xfs_btree.h union xfs_btree_key *xfs_btree_high_key_addr(struct xfs_btree_cur *cur, int n, cur 496 fs/xfs/libxfs/xfs_btree.h union xfs_btree_ptr *xfs_btree_ptr_addr(struct xfs_btree_cur *cur, int n, cur 498 fs/xfs/libxfs/xfs_btree.h int xfs_btree_lookup_get_block(struct xfs_btree_cur *cur, int level, cur 500 fs/xfs/libxfs/xfs_btree.h struct xfs_btree_block *xfs_btree_get_block(struct xfs_btree_cur *cur, cur 502 fs/xfs/libxfs/xfs_btree.h bool xfs_btree_ptr_is_null(struct xfs_btree_cur *cur, union xfs_btree_ptr *ptr); cur 503 fs/xfs/libxfs/xfs_btree.h int64_t xfs_btree_diff_two_ptrs(struct xfs_btree_cur *cur, cur 506 fs/xfs/libxfs/xfs_btree.h void xfs_btree_get_sibling(struct xfs_btree_cur *cur, cur 509 fs/xfs/libxfs/xfs_btree.h void xfs_btree_get_keys(struct xfs_btree_cur *cur, cur 511 fs/xfs/libxfs/xfs_btree.h union xfs_btree_key *xfs_btree_high_key_from_key(struct xfs_btree_cur *cur, cur 513 fs/xfs/libxfs/xfs_btree.h int xfs_btree_has_record(struct xfs_btree_cur *cur, union xfs_btree_irec *low, cur 515 fs/xfs/libxfs/xfs_btree.h bool xfs_btree_has_more_records(struct xfs_btree_cur *cur); cur 36 fs/xfs/libxfs/xfs_ialloc.c struct xfs_btree_cur *cur, /* btree cursor */ cur 41 fs/xfs/libxfs/xfs_ialloc.c cur->bc_rec.i.ir_startino = ino; cur 42 fs/xfs/libxfs/xfs_ialloc.c cur->bc_rec.i.ir_holemask = 0; cur 43 fs/xfs/libxfs/xfs_ialloc.c cur->bc_rec.i.ir_count = 0; cur 44 fs/xfs/libxfs/xfs_ialloc.c cur->bc_rec.i.ir_freecount = 0; cur 45 fs/xfs/libxfs/xfs_ialloc.c cur->bc_rec.i.ir_free = 0; cur 46 fs/xfs/libxfs/xfs_ialloc.c return xfs_btree_lookup(cur, dir, stat); cur 55 fs/xfs/libxfs/xfs_ialloc.c struct xfs_btree_cur *cur, /* btree cursor */ cur 61 fs/xfs/libxfs/xfs_ialloc.c if (xfs_sb_version_hassparseinodes(&cur->bc_mp->m_sb)) { cur 70 fs/xfs/libxfs/xfs_ialloc.c return xfs_btree_update(cur, &rec); cur 103 fs/xfs/libxfs/xfs_ialloc.c struct xfs_btree_cur *cur, cur 107 fs/xfs/libxfs/xfs_ialloc.c struct xfs_mount *mp = cur->bc_mp; cur 108 fs/xfs/libxfs/xfs_ialloc.c xfs_agnumber_t agno = cur->bc_private.a.agno; cur 113 fs/xfs/libxfs/xfs_ialloc.c error = xfs_btree_get_rec(cur, &rec, stat); cur 140 fs/xfs/libxfs/xfs_ialloc.c cur->bc_btnum == XFS_BTNUM_INO ? "Used" : "Free", agno); cur 153 fs/xfs/libxfs/xfs_ialloc.c struct xfs_btree_cur *cur, cur 160 fs/xfs/libxfs/xfs_ialloc.c cur->bc_rec.i.ir_holemask = holemask; cur 161 fs/xfs/libxfs/xfs_ialloc.c cur->bc_rec.i.ir_count = count; cur 162 fs/xfs/libxfs/xfs_ialloc.c cur->bc_rec.i.ir_freecount = freecount; cur 163 fs/xfs/libxfs/xfs_ialloc.c cur->bc_rec.i.ir_free = free; cur 164 fs/xfs/libxfs/xfs_ialloc.c return xfs_btree_insert(cur, stat); cur 179 fs/xfs/libxfs/xfs_ialloc.c struct xfs_btree_cur *cur; cur 186 fs/xfs/libxfs/xfs_ialloc.c cur = xfs_inobt_init_cursor(mp, tp, agbp, agno, btnum); cur 191 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_lookup(cur, thisino, XFS_LOOKUP_EQ, &i); cur 193 fs/xfs/libxfs/xfs_ialloc.c xfs_btree_del_cursor(cur, XFS_BTREE_ERROR); cur 198 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_insert_rec(cur, XFS_INOBT_HOLEMASK_FULL, cur 203 fs/xfs/libxfs/xfs_ialloc.c xfs_btree_del_cursor(cur, XFS_BTREE_ERROR); cur 209 fs/xfs/libxfs/xfs_ialloc.c xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR); cur 220 fs/xfs/libxfs/xfs_ialloc.c struct xfs_btree_cur *cur, cur 223 fs/xfs/libxfs/xfs_ialloc.c if (cur->bc_nlevels == 1) { cur 229 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_lookup(cur, 0, XFS_LOOKUP_GE, &i); cur 234 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_get_rec(cur, &rec, &i); cur 240 fs/xfs/libxfs/xfs_ialloc.c error = xfs_btree_increment(cur, 0, &i); cur 246 fs/xfs/libxfs/xfs_ialloc.c if (!XFS_FORCED_SHUTDOWN(cur->bc_mp)) cur 252 fs/xfs/libxfs/xfs_ialloc.c #define xfs_check_agi_freecount(cur, agi) 0 cur 527 fs/xfs/libxfs/xfs_ialloc.c struct xfs_btree_cur *cur; cur 534 fs/xfs/libxfs/xfs_ialloc.c cur = xfs_inobt_init_cursor(mp, tp, agbp, agno, btnum); cur 537 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_lookup(cur, nrec->ir_startino, XFS_LOOKUP_EQ, &i); cur 542 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_insert_rec(cur, nrec->ir_holemask, cur 557 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_get_rec(cur, &rec, &i); cur 587 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_update(cur, nrec); cur 592 fs/xfs/libxfs/xfs_ialloc.c xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR); cur 595 fs/xfs/libxfs/xfs_ialloc.c xfs_btree_del_cursor(cur, XFS_BTREE_ERROR); cur 1040 fs/xfs/libxfs/xfs_ialloc.c struct xfs_btree_cur *cur, cur 1049 fs/xfs/libxfs/xfs_ialloc.c error = xfs_btree_decrement(cur, 0, &i); cur 1051 fs/xfs/libxfs/xfs_ialloc.c error = xfs_btree_increment(cur, 0, &i); cur 1057 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_get_rec(cur, rec, &i); cur 1060 fs/xfs/libxfs/xfs_ialloc.c XFS_WANT_CORRUPTED_RETURN(cur->bc_mp, i == 1); cur 1068 fs/xfs/libxfs/xfs_ialloc.c struct xfs_btree_cur *cur, cur 1076 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_lookup(cur, agino, XFS_LOOKUP_EQ, &i); cur 1081 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_get_rec(cur, rec, &i); cur 1084 fs/xfs/libxfs/xfs_ialloc.c XFS_WANT_CORRUPTED_RETURN(cur->bc_mp, i == 1); cur 1127 fs/xfs/libxfs/xfs_ialloc.c struct xfs_btree_cur *cur, *tcur; cur 1142 fs/xfs/libxfs/xfs_ialloc.c cur = xfs_inobt_init_cursor(mp, tp, agbp, agno, XFS_BTNUM_INO); cur 1150 fs/xfs/libxfs/xfs_ialloc.c error = xfs_check_agi_freecount(cur, agi); cur 1161 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_lookup(cur, pagino, XFS_LOOKUP_LE, &i); cur 1166 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_get_rec(cur, &rec, &j); cur 1185 fs/xfs/libxfs/xfs_ialloc.c error = xfs_btree_dup_cursor(cur, &tcur); cur 1201 fs/xfs/libxfs/xfs_ialloc.c error = xfs_ialloc_get_rec(cur, pag->pagl_rightrec, cur 1212 fs/xfs/libxfs/xfs_ialloc.c error = xfs_ialloc_next_rec(cur, &rec, &doneright, 0); cur 1234 fs/xfs/libxfs/xfs_ialloc.c xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR); cur 1235 fs/xfs/libxfs/xfs_ialloc.c cur = tcur; cur 1259 fs/xfs/libxfs/xfs_ialloc.c error = xfs_ialloc_next_rec(cur, &rec, cur 1288 fs/xfs/libxfs/xfs_ialloc.c xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR); cur 1298 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_lookup(cur, be32_to_cpu(agi->agi_newino), cur 1304 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_get_rec(cur, &rec, &j); cur 1321 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_lookup(cur, 0, XFS_LOOKUP_GE, &i); cur 1327 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_get_rec(cur, &rec, &i); cur 1333 fs/xfs/libxfs/xfs_ialloc.c error = xfs_btree_increment(cur, 0, &i); cur 1348 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_update(cur, &rec); cur 1355 fs/xfs/libxfs/xfs_ialloc.c error = xfs_check_agi_freecount(cur, agi); cur 1359 fs/xfs/libxfs/xfs_ialloc.c xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR); cur 1367 fs/xfs/libxfs/xfs_ialloc.c xfs_btree_del_cursor(cur, XFS_BTREE_ERROR); cur 1460 fs/xfs/libxfs/xfs_ialloc.c struct xfs_btree_cur *cur, cur 1467 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_lookup(cur, be32_to_cpu(agi->agi_newino), cur 1472 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_get_rec(cur, rec, &i); cur 1475 fs/xfs/libxfs/xfs_ialloc.c XFS_WANT_CORRUPTED_RETURN(cur->bc_mp, i == 1); cur 1483 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_lookup(cur, 0, XFS_LOOKUP_GE, &i); cur 1486 fs/xfs/libxfs/xfs_ialloc.c XFS_WANT_CORRUPTED_RETURN(cur->bc_mp, i == 1); cur 1488 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_get_rec(cur, rec, &i); cur 1491 fs/xfs/libxfs/xfs_ialloc.c XFS_WANT_CORRUPTED_RETURN(cur->bc_mp, i == 1); cur 1502 fs/xfs/libxfs/xfs_ialloc.c struct xfs_btree_cur *cur, /* inobt cursor */ cur 1510 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_lookup(cur, frec->ir_startino, XFS_LOOKUP_EQ, &i); cur 1513 fs/xfs/libxfs/xfs_ialloc.c XFS_WANT_CORRUPTED_RETURN(cur->bc_mp, i == 1); cur 1515 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_get_rec(cur, &rec, &i); cur 1518 fs/xfs/libxfs/xfs_ialloc.c XFS_WANT_CORRUPTED_RETURN(cur->bc_mp, i == 1); cur 1519 fs/xfs/libxfs/xfs_ialloc.c ASSERT((XFS_AGINO_TO_OFFSET(cur->bc_mp, rec.ir_startino) % cur 1525 fs/xfs/libxfs/xfs_ialloc.c XFS_WANT_CORRUPTED_RETURN(cur->bc_mp, (rec.ir_free == frec->ir_free) && cur 1528 fs/xfs/libxfs/xfs_ialloc.c return xfs_inobt_update(cur, &rec); cur 1551 fs/xfs/libxfs/xfs_ialloc.c struct xfs_btree_cur *cur; /* finobt cursor */ cur 1571 fs/xfs/libxfs/xfs_ialloc.c cur = xfs_inobt_init_cursor(mp, tp, agbp, agno, XFS_BTNUM_FINO); cur 1573 fs/xfs/libxfs/xfs_ialloc.c error = xfs_check_agi_freecount(cur, agi); cur 1583 fs/xfs/libxfs/xfs_ialloc.c error = xfs_dialloc_ag_finobt_near(pagino, &cur, &rec); cur 1585 fs/xfs/libxfs/xfs_ialloc.c error = xfs_dialloc_ag_finobt_newino(agi, cur, &rec); cur 1602 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_update(cur, &rec); cur 1604 fs/xfs/libxfs/xfs_ialloc.c error = xfs_btree_delete(cur, &i); cur 1637 fs/xfs/libxfs/xfs_ialloc.c error = xfs_check_agi_freecount(cur, agi); cur 1642 fs/xfs/libxfs/xfs_ialloc.c xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR); cur 1650 fs/xfs/libxfs/xfs_ialloc.c xfs_btree_del_cursor(cur, XFS_BTREE_ERROR); cur 1909 fs/xfs/libxfs/xfs_ialloc.c struct xfs_btree_cur *cur; cur 1922 fs/xfs/libxfs/xfs_ialloc.c cur = xfs_inobt_init_cursor(mp, tp, agbp, agno, XFS_BTNUM_INO); cur 1924 fs/xfs/libxfs/xfs_ialloc.c error = xfs_check_agi_freecount(cur, agi); cur 1931 fs/xfs/libxfs/xfs_ialloc.c if ((error = xfs_inobt_lookup(cur, agino, XFS_LOOKUP_LE, &i))) { cur 1937 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_get_rec(cur, &rec, &i); cur 1984 fs/xfs/libxfs/xfs_ialloc.c if ((error = xfs_btree_delete(cur, &i))) { cur 1994 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_update(cur, &rec); cur 2012 fs/xfs/libxfs/xfs_ialloc.c error = xfs_check_agi_freecount(cur, agi); cur 2017 fs/xfs/libxfs/xfs_ialloc.c xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR); cur 2021 fs/xfs/libxfs/xfs_ialloc.c xfs_btree_del_cursor(cur, XFS_BTREE_ERROR); cur 2038 fs/xfs/libxfs/xfs_ialloc.c struct xfs_btree_cur *cur; cur 2044 fs/xfs/libxfs/xfs_ialloc.c cur = xfs_inobt_init_cursor(mp, tp, agbp, agno, XFS_BTNUM_FINO); cur 2046 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_lookup(cur, ibtrec->ir_startino, XFS_LOOKUP_EQ, &i); cur 2057 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_insert_rec(cur, ibtrec->ir_holemask, cur 2075 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_get_rec(cur, &rec, &i); cur 2102 fs/xfs/libxfs/xfs_ialloc.c error = xfs_btree_delete(cur, &i); cur 2107 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_update(cur, &rec); cur 2113 fs/xfs/libxfs/xfs_ialloc.c error = xfs_check_agi_freecount(cur, agi); cur 2117 fs/xfs/libxfs/xfs_ialloc.c xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR); cur 2121 fs/xfs/libxfs/xfs_ialloc.c xfs_btree_del_cursor(cur, XFS_BTREE_ERROR); cur 2217 fs/xfs/libxfs/xfs_ialloc.c struct xfs_btree_cur *cur; cur 2236 fs/xfs/libxfs/xfs_ialloc.c cur = xfs_inobt_init_cursor(mp, tp, agbp, agno, XFS_BTNUM_INO); cur 2237 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_lookup(cur, agino, XFS_LOOKUP_LE, &i); cur 2240 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_get_rec(cur, &rec, &i); cur 2246 fs/xfs/libxfs/xfs_ialloc.c xfs_btree_del_cursor(cur, error); cur 2650 fs/xfs/libxfs/xfs_ialloc.c struct xfs_btree_cur *cur, cur 2663 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_lookup(cur, low, XFS_LOOKUP_LE, &has_record); cur 2665 fs/xfs/libxfs/xfs_ialloc.c error = xfs_inobt_get_rec(cur, &irec, &has_record); cur 2682 fs/xfs/libxfs/xfs_ialloc.c error = xfs_btree_increment(cur, 0, &has_record); cur 2690 fs/xfs/libxfs/xfs_ialloc.c struct xfs_btree_cur *cur, cur 2698 fs/xfs/libxfs/xfs_ialloc.c low = XFS_AGB_TO_AGINO(cur->bc_mp, bno); cur 2699 fs/xfs/libxfs/xfs_ialloc.c high = XFS_AGB_TO_AGINO(cur->bc_mp, bno + len) - 1; cur 2701 fs/xfs/libxfs/xfs_ialloc.c return xfs_ialloc_has_inode_record(cur, low, high, exists); cur 2712 fs/xfs/libxfs/xfs_ialloc.c struct xfs_btree_cur *cur, cur 2719 fs/xfs/libxfs/xfs_ialloc.c xfs_inobt_btrec_to_irec(cur->bc_mp, rec, &irec); cur 2729 fs/xfs/libxfs/xfs_ialloc.c struct xfs_btree_cur *cur, cur 2736 fs/xfs/libxfs/xfs_ialloc.c ASSERT(cur->bc_btnum == XFS_BTNUM_INO); cur 2737 fs/xfs/libxfs/xfs_ialloc.c error = xfs_btree_query_all(cur, xfs_ialloc_count_inodes_rec, &ci); cur 120 fs/xfs/libxfs/xfs_ialloc.h int xfs_inobt_lookup(struct xfs_btree_cur *cur, xfs_agino_t ino, cur 126 fs/xfs/libxfs/xfs_ialloc.h int xfs_inobt_get_rec(struct xfs_btree_cur *cur, cur 143 fs/xfs/libxfs/xfs_ialloc.h int xfs_ialloc_has_inodes_at_extent(struct xfs_btree_cur *cur, cur 145 fs/xfs/libxfs/xfs_ialloc.h int xfs_ialloc_has_inode_record(struct xfs_btree_cur *cur, xfs_agino_t low, cur 147 fs/xfs/libxfs/xfs_ialloc.h int xfs_ialloc_count_inodes(struct xfs_btree_cur *cur, xfs_agino_t *count, cur 149 fs/xfs/libxfs/xfs_ialloc.h int xfs_inobt_insert_rec(struct xfs_btree_cur *cur, uint16_t holemask, cur 26 fs/xfs/libxfs/xfs_ialloc_btree.c struct xfs_btree_cur *cur, cur 29 fs/xfs/libxfs/xfs_ialloc_btree.c return M_IGEO(cur->bc_mp)->inobt_mnr[level != 0]; cur 34 fs/xfs/libxfs/xfs_ialloc_btree.c struct xfs_btree_cur *cur) cur 36 fs/xfs/libxfs/xfs_ialloc_btree.c return xfs_inobt_init_cursor(cur->bc_mp, cur->bc_tp, cur 37 fs/xfs/libxfs/xfs_ialloc_btree.c cur->bc_private.a.agbp, cur->bc_private.a.agno, cur 38 fs/xfs/libxfs/xfs_ialloc_btree.c cur->bc_btnum); cur 43 fs/xfs/libxfs/xfs_ialloc_btree.c struct xfs_btree_cur *cur, cur 47 fs/xfs/libxfs/xfs_ialloc_btree.c struct xfs_buf *agbp = cur->bc_private.a.agbp; cur 52 fs/xfs/libxfs/xfs_ialloc_btree.c xfs_ialloc_log_agi(cur->bc_tp, agbp, XFS_AGI_ROOT | XFS_AGI_LEVEL); cur 57 fs/xfs/libxfs/xfs_ialloc_btree.c struct xfs_btree_cur *cur, cur 61 fs/xfs/libxfs/xfs_ialloc_btree.c struct xfs_buf *agbp = cur->bc_private.a.agbp; cur 66 fs/xfs/libxfs/xfs_ialloc_btree.c xfs_ialloc_log_agi(cur->bc_tp, agbp, cur 72 fs/xfs/libxfs/xfs_ialloc_btree.c struct xfs_btree_cur *cur, cur 83 fs/xfs/libxfs/xfs_ialloc_btree.c args.tp = cur->bc_tp; cur 84 fs/xfs/libxfs/xfs_ialloc_btree.c args.mp = cur->bc_mp; cur 86 fs/xfs/libxfs/xfs_ialloc_btree.c args.fsbno = XFS_AGB_TO_FSB(args.mp, cur->bc_private.a.agno, sbno); cur 110 fs/xfs/libxfs/xfs_ialloc_btree.c struct xfs_btree_cur *cur, cur 115 fs/xfs/libxfs/xfs_ialloc_btree.c return __xfs_inobt_alloc_block(cur, start, new, stat, XFS_AG_RESV_NONE); cur 120 fs/xfs/libxfs/xfs_ialloc_btree.c struct xfs_btree_cur *cur, cur 125 fs/xfs/libxfs/xfs_ialloc_btree.c if (cur->bc_mp->m_finobt_nores) cur 126 fs/xfs/libxfs/xfs_ialloc_btree.c return xfs_inobt_alloc_block(cur, start, new, stat); cur 127 fs/xfs/libxfs/xfs_ialloc_btree.c return __xfs_inobt_alloc_block(cur, start, new, stat, cur 133 fs/xfs/libxfs/xfs_ialloc_btree.c struct xfs_btree_cur *cur, cur 137 fs/xfs/libxfs/xfs_ialloc_btree.c return xfs_free_extent(cur->bc_tp, cur 138 fs/xfs/libxfs/xfs_ialloc_btree.c XFS_DADDR_TO_FSB(cur->bc_mp, XFS_BUF_ADDR(bp)), 1, cur 144 fs/xfs/libxfs/xfs_ialloc_btree.c struct xfs_btree_cur *cur, cur 147 fs/xfs/libxfs/xfs_ialloc_btree.c return __xfs_inobt_free_block(cur, bp, XFS_AG_RESV_NONE); cur 152 fs/xfs/libxfs/xfs_ialloc_btree.c struct xfs_btree_cur *cur, cur 155 fs/xfs/libxfs/xfs_ialloc_btree.c if (cur->bc_mp->m_finobt_nores) cur 156 fs/xfs/libxfs/xfs_ialloc_btree.c return xfs_inobt_free_block(cur, bp); cur 157 fs/xfs/libxfs/xfs_ialloc_btree.c return __xfs_inobt_free_block(cur, bp, XFS_AG_RESV_METADATA); cur 162 fs/xfs/libxfs/xfs_ialloc_btree.c struct xfs_btree_cur *cur, cur 165 fs/xfs/libxfs/xfs_ialloc_btree.c return M_IGEO(cur->bc_mp)->inobt_mxr[level != 0]; cur 190 fs/xfs/libxfs/xfs_ialloc_btree.c struct xfs_btree_cur *cur, cur 193 fs/xfs/libxfs/xfs_ialloc_btree.c rec->inobt.ir_startino = cpu_to_be32(cur->bc_rec.i.ir_startino); cur 194 fs/xfs/libxfs/xfs_ialloc_btree.c if (xfs_sb_version_hassparseinodes(&cur->bc_mp->m_sb)) { cur 196 fs/xfs/libxfs/xfs_ialloc_btree.c cpu_to_be16(cur->bc_rec.i.ir_holemask); cur 197 fs/xfs/libxfs/xfs_ialloc_btree.c rec->inobt.ir_u.sp.ir_count = cur->bc_rec.i.ir_count; cur 198 fs/xfs/libxfs/xfs_ialloc_btree.c rec->inobt.ir_u.sp.ir_freecount = cur->bc_rec.i.ir_freecount; cur 202 fs/xfs/libxfs/xfs_ialloc_btree.c cpu_to_be32(cur->bc_rec.i.ir_freecount); cur 204 fs/xfs/libxfs/xfs_ialloc_btree.c rec->inobt.ir_free = cpu_to_be64(cur->bc_rec.i.ir_free); cur 212 fs/xfs/libxfs/xfs_ialloc_btree.c struct xfs_btree_cur *cur, cur 215 fs/xfs/libxfs/xfs_ialloc_btree.c struct xfs_agi *agi = XFS_BUF_TO_AGI(cur->bc_private.a.agbp); cur 217 fs/xfs/libxfs/xfs_ialloc_btree.c ASSERT(cur->bc_private.a.agno == be32_to_cpu(agi->agi_seqno)); cur 224 fs/xfs/libxfs/xfs_ialloc_btree.c struct xfs_btree_cur *cur, cur 227 fs/xfs/libxfs/xfs_ialloc_btree.c struct xfs_agi *agi = XFS_BUF_TO_AGI(cur->bc_private.a.agbp); cur 229 fs/xfs/libxfs/xfs_ialloc_btree.c ASSERT(cur->bc_private.a.agno == be32_to_cpu(agi->agi_seqno)); cur 235 fs/xfs/libxfs/xfs_ialloc_btree.c struct xfs_btree_cur *cur, cur 239 fs/xfs/libxfs/xfs_ialloc_btree.c cur->bc_rec.i.ir_startino; cur 244 fs/xfs/libxfs/xfs_ialloc_btree.c struct xfs_btree_cur *cur, cur 342 fs/xfs/libxfs/xfs_ialloc_btree.c struct xfs_btree_cur *cur, cur 352 fs/xfs/libxfs/xfs_ialloc_btree.c struct xfs_btree_cur *cur, cur 414 fs/xfs/libxfs/xfs_ialloc_btree.c struct xfs_btree_cur *cur; cur 416 fs/xfs/libxfs/xfs_ialloc_btree.c cur = kmem_zone_zalloc(xfs_btree_cur_zone, KM_NOFS); cur 418 fs/xfs/libxfs/xfs_ialloc_btree.c cur->bc_tp = tp; cur 419 fs/xfs/libxfs/xfs_ialloc_btree.c cur->bc_mp = mp; cur 420 fs/xfs/libxfs/xfs_ialloc_btree.c cur->bc_btnum = btnum; cur 422 fs/xfs/libxfs/xfs_ialloc_btree.c cur->bc_nlevels = be32_to_cpu(agi->agi_level); cur 423 fs/xfs/libxfs/xfs_ialloc_btree.c cur->bc_ops = &xfs_inobt_ops; cur 424 fs/xfs/libxfs/xfs_ialloc_btree.c cur->bc_statoff = XFS_STATS_CALC_INDEX(xs_ibt_2); cur 426 fs/xfs/libxfs/xfs_ialloc_btree.c cur->bc_nlevels = be32_to_cpu(agi->agi_free_level); cur 427 fs/xfs/libxfs/xfs_ialloc_btree.c cur->bc_ops = &xfs_finobt_ops; cur 428 fs/xfs/libxfs/xfs_ialloc_btree.c cur->bc_statoff = XFS_STATS_CALC_INDEX(xs_fibt_2); cur 431 fs/xfs/libxfs/xfs_ialloc_btree.c cur->bc_blocklog = mp->m_sb.sb_blocklog; cur 434 fs/xfs/libxfs/xfs_ialloc_btree.c cur->bc_flags |= XFS_BTREE_CRC_BLOCKS; cur 436 fs/xfs/libxfs/xfs_ialloc_btree.c cur->bc_private.a.agbp = agbp; cur 437 fs/xfs/libxfs/xfs_ialloc_btree.c cur->bc_private.a.agno = agno; cur 439 fs/xfs/libxfs/xfs_ialloc_btree.c return cur; cur 575 fs/xfs/libxfs/xfs_ialloc_btree.c struct xfs_btree_cur *cur; cur 585 fs/xfs/libxfs/xfs_ialloc_btree.c cur = xfs_inobt_init_cursor(mp, tp, *agi_bpp, agno, which); cur 586 fs/xfs/libxfs/xfs_ialloc_btree.c if (!cur) { cur 591 fs/xfs/libxfs/xfs_ialloc_btree.c *curpp = cur; cur 604 fs/xfs/libxfs/xfs_ialloc_btree.c struct xfs_btree_cur *cur = NULL; cur 607 fs/xfs/libxfs/xfs_ialloc_btree.c error = xfs_inobt_cur(mp, tp, agno, btnum, &cur, &agbp); cur 611 fs/xfs/libxfs/xfs_ialloc_btree.c error = xfs_btree_count_blocks(cur, tree_blocks); cur 612 fs/xfs/libxfs/xfs_ialloc_btree.c xfs_btree_del_cursor(cur, error); cur 140 fs/xfs/libxfs/xfs_iext_tree.c static inline struct xfs_iext_rec *cur_rec(struct xfs_iext_cursor *cur) cur 142 fs/xfs/libxfs/xfs_iext_tree.c return &cur->leaf->recs[cur->pos]; cur 146 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_cursor *cur) cur 148 fs/xfs/libxfs/xfs_iext_tree.c if (!cur->leaf) cur 150 fs/xfs/libxfs/xfs_iext_tree.c if (cur->pos < 0 || cur->pos >= xfs_iext_max_recs(ifp)) cur 152 fs/xfs/libxfs/xfs_iext_tree.c if (xfs_iext_rec_is_empty(cur_rec(cur))) cur 199 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_cursor *cur) cur 201 fs/xfs/libxfs/xfs_iext_tree.c cur->pos = 0; cur 202 fs/xfs/libxfs/xfs_iext_tree.c cur->leaf = xfs_iext_find_first_leaf(ifp); cur 208 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_cursor *cur) cur 212 fs/xfs/libxfs/xfs_iext_tree.c cur->leaf = xfs_iext_find_last_leaf(ifp); cur 213 fs/xfs/libxfs/xfs_iext_tree.c if (!cur->leaf) { cur 214 fs/xfs/libxfs/xfs_iext_tree.c cur->pos = 0; cur 219 fs/xfs/libxfs/xfs_iext_tree.c if (xfs_iext_rec_is_empty(&cur->leaf->recs[i])) cur 222 fs/xfs/libxfs/xfs_iext_tree.c cur->pos = i - 1; cur 228 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_cursor *cur) cur 230 fs/xfs/libxfs/xfs_iext_tree.c if (!cur->leaf) { cur 231 fs/xfs/libxfs/xfs_iext_tree.c ASSERT(cur->pos <= 0 || cur->pos >= RECS_PER_LEAF); cur 232 fs/xfs/libxfs/xfs_iext_tree.c xfs_iext_first(ifp, cur); cur 236 fs/xfs/libxfs/xfs_iext_tree.c ASSERT(cur->pos >= 0); cur 237 fs/xfs/libxfs/xfs_iext_tree.c ASSERT(cur->pos < xfs_iext_max_recs(ifp)); cur 239 fs/xfs/libxfs/xfs_iext_tree.c cur->pos++; cur 240 fs/xfs/libxfs/xfs_iext_tree.c if (ifp->if_height > 1 && !xfs_iext_valid(ifp, cur) && cur 241 fs/xfs/libxfs/xfs_iext_tree.c cur->leaf->next) { cur 242 fs/xfs/libxfs/xfs_iext_tree.c cur->leaf = cur->leaf->next; cur 243 fs/xfs/libxfs/xfs_iext_tree.c cur->pos = 0; cur 250 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_cursor *cur) cur 252 fs/xfs/libxfs/xfs_iext_tree.c if (!cur->leaf) { cur 253 fs/xfs/libxfs/xfs_iext_tree.c ASSERT(cur->pos <= 0 || cur->pos >= RECS_PER_LEAF); cur 254 fs/xfs/libxfs/xfs_iext_tree.c xfs_iext_last(ifp, cur); cur 258 fs/xfs/libxfs/xfs_iext_tree.c ASSERT(cur->pos >= 0); cur 259 fs/xfs/libxfs/xfs_iext_tree.c ASSERT(cur->pos <= RECS_PER_LEAF); cur 263 fs/xfs/libxfs/xfs_iext_tree.c cur->pos--; cur 264 fs/xfs/libxfs/xfs_iext_tree.c if (xfs_iext_valid(ifp, cur)) cur 266 fs/xfs/libxfs/xfs_iext_tree.c } while (cur->pos > 0); cur 268 fs/xfs/libxfs/xfs_iext_tree.c if (ifp->if_height > 1 && cur->leaf->prev) { cur 269 fs/xfs/libxfs/xfs_iext_tree.c cur->leaf = cur->leaf->prev; cur 270 fs/xfs/libxfs/xfs_iext_tree.c cur->pos = RECS_PER_LEAF; cur 541 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_cursor *cur, cur 544 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_leaf *leaf = cur->leaf; cur 551 fs/xfs/libxfs/xfs_iext_tree.c if (cur->pos == RECS_PER_LEAF) { cur 552 fs/xfs/libxfs/xfs_iext_tree.c cur->leaf = new; cur 553 fs/xfs/libxfs/xfs_iext_tree.c cur->pos = 0; cur 563 fs/xfs/libxfs/xfs_iext_tree.c if (cur->pos >= nr_keep) { cur 564 fs/xfs/libxfs/xfs_iext_tree.c cur->leaf = new; cur 565 fs/xfs/libxfs/xfs_iext_tree.c cur->pos -= nr_keep; cur 582 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_cursor *cur) cur 590 fs/xfs/libxfs/xfs_iext_tree.c cur->leaf = ifp->if_u1.if_root; cur 591 fs/xfs/libxfs/xfs_iext_tree.c cur->pos = 0; cur 597 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_cursor *cur) cur 609 fs/xfs/libxfs/xfs_iext_tree.c cur->leaf = new; cur 627 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_cursor *cur, cur 639 fs/xfs/libxfs/xfs_iext_tree.c xfs_iext_alloc_root(ifp, cur); cur 641 fs/xfs/libxfs/xfs_iext_tree.c xfs_iext_realloc_root(ifp, cur); cur 643 fs/xfs/libxfs/xfs_iext_tree.c nr_entries = xfs_iext_leaf_nr_entries(ifp, cur->leaf, cur->pos); cur 645 fs/xfs/libxfs/xfs_iext_tree.c ASSERT(cur->pos >= nr_entries || cur 646 fs/xfs/libxfs/xfs_iext_tree.c xfs_iext_rec_cmp(cur_rec(cur), irec->br_startoff) != 0); cur 649 fs/xfs/libxfs/xfs_iext_tree.c new = xfs_iext_split_leaf(cur, &nr_entries); cur 655 fs/xfs/libxfs/xfs_iext_tree.c if (cur->leaf != new && cur->pos == 0 && nr_entries > 0) { cur 656 fs/xfs/libxfs/xfs_iext_tree.c xfs_iext_update_node(ifp, xfs_iext_leaf_key(cur->leaf, 0), cur 657 fs/xfs/libxfs/xfs_iext_tree.c offset, 1, cur->leaf); cur 660 fs/xfs/libxfs/xfs_iext_tree.c for (i = nr_entries; i > cur->pos; i--) cur 661 fs/xfs/libxfs/xfs_iext_tree.c cur->leaf->recs[i] = cur->leaf->recs[i - 1]; cur 662 fs/xfs/libxfs/xfs_iext_tree.c xfs_iext_set(cur_rec(cur), irec); cur 665 fs/xfs/libxfs/xfs_iext_tree.c trace_xfs_iext_insert(ip, cur, state, _RET_IP_); cur 790 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_cursor *cur, cur 810 fs/xfs/libxfs/xfs_iext_tree.c if (cur->leaf == leaf) { cur 811 fs/xfs/libxfs/xfs_iext_tree.c cur->leaf = leaf->prev; cur 812 fs/xfs/libxfs/xfs_iext_tree.c cur->pos += nr_prev; cur 832 fs/xfs/libxfs/xfs_iext_tree.c if (cur->leaf == leaf->next) { cur 833 fs/xfs/libxfs/xfs_iext_tree.c cur->leaf = leaf; cur 834 fs/xfs/libxfs/xfs_iext_tree.c cur->pos += nr_entries; cur 864 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_cursor *cur, cur 868 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_leaf *leaf = cur->leaf; cur 872 fs/xfs/libxfs/xfs_iext_tree.c trace_xfs_iext_remove(ip, cur, state, _RET_IP_); cur 876 fs/xfs/libxfs/xfs_iext_tree.c ASSERT(xfs_iext_valid(ifp, cur)); cur 880 fs/xfs/libxfs/xfs_iext_tree.c nr_entries = xfs_iext_leaf_nr_entries(ifp, leaf, cur->pos) - 1; cur 881 fs/xfs/libxfs/xfs_iext_tree.c for (i = cur->pos; i < nr_entries; i++) cur 886 fs/xfs/libxfs/xfs_iext_tree.c if (cur->pos == 0 && nr_entries > 0) { cur 890 fs/xfs/libxfs/xfs_iext_tree.c } else if (cur->pos == nr_entries) { cur 892 fs/xfs/libxfs/xfs_iext_tree.c cur->leaf = leaf->next; cur 894 fs/xfs/libxfs/xfs_iext_tree.c cur->leaf = NULL; cur 895 fs/xfs/libxfs/xfs_iext_tree.c cur->pos = 0; cur 902 fs/xfs/libxfs/xfs_iext_tree.c xfs_iext_rebalance_leaf(ifp, cur, leaf, offset, nr_entries); cur 923 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_cursor *cur, cur 928 fs/xfs/libxfs/xfs_iext_tree.c cur->leaf = xfs_iext_find_level(ifp, offset, 1); cur 929 fs/xfs/libxfs/xfs_iext_tree.c if (!cur->leaf) { cur 930 fs/xfs/libxfs/xfs_iext_tree.c cur->pos = 0; cur 934 fs/xfs/libxfs/xfs_iext_tree.c for (cur->pos = 0; cur->pos < xfs_iext_max_recs(ifp); cur->pos++) { cur 935 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_rec *rec = cur_rec(cur); cur 944 fs/xfs/libxfs/xfs_iext_tree.c if (ifp->if_height == 1 || !cur->leaf->next) cur 946 fs/xfs/libxfs/xfs_iext_tree.c cur->leaf = cur->leaf->next; cur 947 fs/xfs/libxfs/xfs_iext_tree.c cur->pos = 0; cur 948 fs/xfs/libxfs/xfs_iext_tree.c if (!xfs_iext_valid(ifp, cur)) cur 951 fs/xfs/libxfs/xfs_iext_tree.c xfs_iext_get(gotp, cur_rec(cur)); cur 964 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_cursor *cur, cur 968 fs/xfs/libxfs/xfs_iext_tree.c if (xfs_iext_lookup_extent(ip, ifp, *end - 1, cur, gotp) && cur 971 fs/xfs/libxfs/xfs_iext_tree.c if (!xfs_iext_prev_extent(ifp, cur, gotp)) cur 981 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_cursor *cur, cur 988 fs/xfs/libxfs/xfs_iext_tree.c if (cur->pos == 0) { cur 991 fs/xfs/libxfs/xfs_iext_tree.c xfs_iext_get(&old, cur_rec(cur)); cur 994 fs/xfs/libxfs/xfs_iext_tree.c new->br_startoff, 1, cur->leaf); cur 998 fs/xfs/libxfs/xfs_iext_tree.c trace_xfs_bmap_pre_update(ip, cur, state, _RET_IP_); cur 999 fs/xfs/libxfs/xfs_iext_tree.c xfs_iext_set(cur_rec(cur), new); cur 1000 fs/xfs/libxfs/xfs_iext_tree.c trace_xfs_bmap_post_update(ip, cur, state, _RET_IP_); cur 1010 fs/xfs/libxfs/xfs_iext_tree.c struct xfs_iext_cursor *cur, cur 1013 fs/xfs/libxfs/xfs_iext_tree.c if (!xfs_iext_valid(ifp, cur)) cur 1015 fs/xfs/libxfs/xfs_iext_tree.c xfs_iext_get(gotp, cur_rec(cur)); cur 104 fs/xfs/libxfs/xfs_inode_fork.h void xfs_iext_insert(struct xfs_inode *, struct xfs_iext_cursor *cur, cur 112 fs/xfs/libxfs/xfs_inode_fork.h struct xfs_iext_cursor *cur, cur 116 fs/xfs/libxfs/xfs_inode_fork.h struct xfs_iext_cursor *cur, cur 119 fs/xfs/libxfs/xfs_inode_fork.h struct xfs_iext_cursor *cur, cur 122 fs/xfs/libxfs/xfs_inode_fork.h struct xfs_iext_cursor *cur, cur 131 fs/xfs/libxfs/xfs_inode_fork.h struct xfs_iext_cursor *cur, struct xfs_bmbt_irec *gotp) cur 133 fs/xfs/libxfs/xfs_inode_fork.h xfs_iext_next(ifp, cur); cur 134 fs/xfs/libxfs/xfs_inode_fork.h return xfs_iext_get_extent(ifp, cur, gotp); cur 138 fs/xfs/libxfs/xfs_inode_fork.h struct xfs_iext_cursor *cur, struct xfs_bmbt_irec *gotp) cur 140 fs/xfs/libxfs/xfs_inode_fork.h xfs_iext_prev(ifp, cur); cur 141 fs/xfs/libxfs/xfs_inode_fork.h return xfs_iext_get_extent(ifp, cur, gotp); cur 148 fs/xfs/libxfs/xfs_inode_fork.h struct xfs_iext_cursor *cur, struct xfs_bmbt_irec *gotp) cur 150 fs/xfs/libxfs/xfs_inode_fork.h struct xfs_iext_cursor ncur = *cur; cur 160 fs/xfs/libxfs/xfs_inode_fork.h struct xfs_iext_cursor *cur, struct xfs_bmbt_irec *gotp) cur 162 fs/xfs/libxfs/xfs_inode_fork.h struct xfs_iext_cursor ncur = *cur; cur 45 fs/xfs/libxfs/xfs_refcount.c struct xfs_btree_cur *cur, cur 49 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_lookup(cur->bc_mp, cur->bc_private.a.agno, bno, cur 51 fs/xfs/libxfs/xfs_refcount.c cur->bc_rec.rc.rc_startblock = bno; cur 52 fs/xfs/libxfs/xfs_refcount.c cur->bc_rec.rc.rc_blockcount = 0; cur 53 fs/xfs/libxfs/xfs_refcount.c return xfs_btree_lookup(cur, XFS_LOOKUP_LE, stat); cur 62 fs/xfs/libxfs/xfs_refcount.c struct xfs_btree_cur *cur, cur 66 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_lookup(cur->bc_mp, cur->bc_private.a.agno, bno, cur 68 fs/xfs/libxfs/xfs_refcount.c cur->bc_rec.rc.rc_startblock = bno; cur 69 fs/xfs/libxfs/xfs_refcount.c cur->bc_rec.rc.rc_blockcount = 0; cur 70 fs/xfs/libxfs/xfs_refcount.c return xfs_btree_lookup(cur, XFS_LOOKUP_GE, stat); cur 79 fs/xfs/libxfs/xfs_refcount.c struct xfs_btree_cur *cur, cur 83 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_lookup(cur->bc_mp, cur->bc_private.a.agno, bno, cur 85 fs/xfs/libxfs/xfs_refcount.c cur->bc_rec.rc.rc_startblock = bno; cur 86 fs/xfs/libxfs/xfs_refcount.c cur->bc_rec.rc.rc_blockcount = 0; cur 87 fs/xfs/libxfs/xfs_refcount.c return xfs_btree_lookup(cur, XFS_LOOKUP_EQ, stat); cur 106 fs/xfs/libxfs/xfs_refcount.c struct xfs_btree_cur *cur, cur 110 fs/xfs/libxfs/xfs_refcount.c struct xfs_mount *mp = cur->bc_mp; cur 111 fs/xfs/libxfs/xfs_refcount.c xfs_agnumber_t agno = cur->bc_private.a.agno; cur 116 fs/xfs/libxfs/xfs_refcount.c error = xfs_btree_get_rec(cur, &rec, stat); cur 122 fs/xfs/libxfs/xfs_refcount.c agno = cur->bc_private.a.agno; cur 147 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_get(cur->bc_mp, cur->bc_private.a.agno, irec); cur 166 fs/xfs/libxfs/xfs_refcount.c struct xfs_btree_cur *cur, cur 172 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_update(cur->bc_mp, cur->bc_private.a.agno, irec); cur 176 fs/xfs/libxfs/xfs_refcount.c error = xfs_btree_update(cur, &rec); cur 178 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_update_error(cur->bc_mp, cur 179 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.agno, error, _RET_IP_); cur 190 fs/xfs/libxfs/xfs_refcount.c struct xfs_btree_cur *cur, cur 196 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_insert(cur->bc_mp, cur->bc_private.a.agno, irec); cur 197 fs/xfs/libxfs/xfs_refcount.c cur->bc_rec.rc.rc_startblock = irec->rc_startblock; cur 198 fs/xfs/libxfs/xfs_refcount.c cur->bc_rec.rc.rc_blockcount = irec->rc_blockcount; cur 199 fs/xfs/libxfs/xfs_refcount.c cur->bc_rec.rc.rc_refcount = irec->rc_refcount; cur 200 fs/xfs/libxfs/xfs_refcount.c error = xfs_btree_insert(cur, i); cur 203 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, *i == 1, out_error); cur 207 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_insert_error(cur->bc_mp, cur 208 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.agno, error, _RET_IP_); cur 220 fs/xfs/libxfs/xfs_refcount.c struct xfs_btree_cur *cur, cur 227 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_get_rec(cur, &irec, &found_rec); cur 230 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, out_error); cur 231 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_delete(cur->bc_mp, cur->bc_private.a.agno, &irec); cur 232 fs/xfs/libxfs/xfs_refcount.c error = xfs_btree_delete(cur, i); cur 233 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, *i == 1, out_error); cur 236 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_lookup_ge(cur, irec.rc_startblock, &found_rec); cur 239 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_delete_error(cur->bc_mp, cur 240 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.agno, error, _RET_IP_); cur 334 fs/xfs/libxfs/xfs_refcount.c struct xfs_btree_cur *cur, cur 343 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_lookup_le(cur, agbno, &found_rec); cur 349 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_get_rec(cur, &rcext, &found_rec); cur 352 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, out_error); cur 357 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_split_extent(cur->bc_mp, cur->bc_private.a.agno, cur 364 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_update(cur, &tmp); cur 371 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_insert(cur, &tmp, &found_rec); cur 374 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, out_error); cur 378 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_split_extent_error(cur->bc_mp, cur 379 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.agno, error, _RET_IP_); cur 388 fs/xfs/libxfs/xfs_refcount.c struct xfs_btree_cur *cur, cur 398 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_merge_center_extents(cur->bc_mp, cur 399 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.agno, left, center, right); cur 409 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_lookup_ge(cur, center->rc_startblock, cur 413 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, out_error); cur 415 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_delete(cur, &found_rec); cur 418 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, out_error); cur 421 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_delete(cur, &found_rec); cur 424 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, cur 429 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_lookup_le(cur, left->rc_startblock, cur 433 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, out_error); cur 436 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_update(cur, left); cur 444 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_merge_center_extents_error(cur->bc_mp, cur 445 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.agno, error, _RET_IP_); cur 454 fs/xfs/libxfs/xfs_refcount.c struct xfs_btree_cur *cur, cur 463 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_merge_left_extent(cur->bc_mp, cur 464 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.agno, left, cleft); cur 468 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_lookup_le(cur, cleft->rc_startblock, cur 472 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, cur 475 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_delete(cur, &found_rec); cur 478 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, cur 483 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_lookup_le(cur, left->rc_startblock, cur 487 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, out_error); cur 490 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_update(cur, left); cur 499 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_merge_left_extent_error(cur->bc_mp, cur 500 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.agno, error, _RET_IP_); cur 509 fs/xfs/libxfs/xfs_refcount.c struct xfs_btree_cur *cur, cur 517 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_merge_right_extent(cur->bc_mp, cur 518 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.agno, cright, right); cur 525 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_lookup_le(cur, cright->rc_startblock, cur 529 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, cur 532 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_delete(cur, &found_rec); cur 535 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, cur 540 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_lookup_le(cur, right->rc_startblock, cur 544 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, out_error); cur 548 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_update(cur, right); cur 556 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_merge_right_extent_error(cur->bc_mp, cur 557 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.agno, error, _RET_IP_); cur 569 fs/xfs/libxfs/xfs_refcount.c struct xfs_btree_cur *cur, cur 581 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_lookup_le(cur, agbno - 1, &found_rec); cur 587 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_get_rec(cur, &tmp, &found_rec); cur 590 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, out_error); cur 601 fs/xfs/libxfs/xfs_refcount.c error = xfs_btree_increment(cur, 0, &found_rec); cur 605 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_get_rec(cur, &tmp, &found_rec); cur 608 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, cur 637 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_find_left_extent(cur->bc_mp, cur->bc_private.a.agno, cur 642 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_find_left_extent_error(cur->bc_mp, cur 643 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.agno, error, _RET_IP_); cur 653 fs/xfs/libxfs/xfs_refcount.c struct xfs_btree_cur *cur, cur 665 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_lookup_ge(cur, agbno + aglen, &found_rec); cur 671 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_get_rec(cur, &tmp, &found_rec); cur 674 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, out_error); cur 685 fs/xfs/libxfs/xfs_refcount.c error = xfs_btree_decrement(cur, 0, &found_rec); cur 689 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_get_rec(cur, &tmp, &found_rec); cur 692 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, found_rec == 1, cur 721 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_find_right_extent(cur->bc_mp, cur->bc_private.a.agno, cur 726 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_find_right_extent_error(cur->bc_mp, cur 727 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.agno, error, _RET_IP_); cur 744 fs/xfs/libxfs/xfs_refcount.c struct xfs_btree_cur *cur, cur 763 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_find_left_extents(cur, &left, &cleft, *agbno, cur 767 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_find_right_extents(cur, &right, &cright, *agbno, cur 788 fs/xfs/libxfs/xfs_refcount.c return xfs_refcount_merge_center_extents(cur, &left, &cleft, cur 798 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_merge_left_extent(cur, &left, &cleft, cur 817 fs/xfs/libxfs/xfs_refcount.c return xfs_refcount_merge_right_extent(cur, &right, &cright, cur 832 fs/xfs/libxfs/xfs_refcount.c struct xfs_btree_cur *cur) cur 836 fs/xfs/libxfs/xfs_refcount.c overhead = cur->bc_private.a.priv.refc.shape_changes * cur 837 fs/xfs/libxfs/xfs_refcount.c xfs_allocfree_log_count(cur->bc_mp, 1); cur 838 fs/xfs/libxfs/xfs_refcount.c overhead *= cur->bc_mp->m_sb.sb_blocksize; cur 844 fs/xfs/libxfs/xfs_refcount.c if (cur->bc_private.a.priv.refc.nr_ops > 2 && cur 845 fs/xfs/libxfs/xfs_refcount.c XFS_TEST_ERROR(false, cur->bc_mp, cur 849 fs/xfs/libxfs/xfs_refcount.c if (cur->bc_private.a.priv.refc.nr_ops == 0) cur 851 fs/xfs/libxfs/xfs_refcount.c else if (overhead > cur->bc_tp->t_log_res) cur 853 fs/xfs/libxfs/xfs_refcount.c return cur->bc_tp->t_log_res - overhead > cur 854 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.priv.refc.nr_ops * XFS_REFCOUNT_ITEM_OVERHEAD; cur 865 fs/xfs/libxfs/xfs_refcount.c struct xfs_btree_cur *cur, cur 880 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_lookup_ge(cur, *agbno, &found_rec); cur 884 fs/xfs/libxfs/xfs_refcount.c while (*aglen > 0 && xfs_refcount_still_have_space(cur)) { cur 885 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_get_rec(cur, &ext, &found_rec); cur 889 fs/xfs/libxfs/xfs_refcount.c ext.rc_startblock = cur->bc_mp->m_sb.sb_agblocks; cur 904 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_modify_extent(cur->bc_mp, cur 905 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.agno, &tmp); cur 912 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_insert(cur, &tmp, cur 916 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, cur 918 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.priv.refc.nr_ops++; cur 920 fs/xfs/libxfs/xfs_refcount.c fsbno = XFS_AGB_TO_FSB(cur->bc_mp, cur 921 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.agno, cur 923 fs/xfs/libxfs/xfs_refcount.c xfs_bmap_add_free(cur->bc_tp, fsbno, cur 930 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_lookup_ge(cur, *agbno, cur 937 fs/xfs/libxfs/xfs_refcount.c if (*aglen == 0 || !xfs_refcount_still_have_space(cur)) cur 947 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_modify_extent(cur->bc_mp, cur 948 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.agno, &ext); cur 950 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_update(cur, &ext); cur 953 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.priv.refc.nr_ops++; cur 955 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_delete(cur, &found_rec); cur 958 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, cur 960 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.priv.refc.nr_ops++; cur 963 fs/xfs/libxfs/xfs_refcount.c fsbno = XFS_AGB_TO_FSB(cur->bc_mp, cur 964 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.agno, cur 966 fs/xfs/libxfs/xfs_refcount.c xfs_bmap_add_free(cur->bc_tp, fsbno, ext.rc_blockcount, cur 971 fs/xfs/libxfs/xfs_refcount.c error = xfs_btree_increment(cur, 0, &found_rec); cur 982 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_modify_extent_error(cur->bc_mp, cur 983 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.agno, error, _RET_IP_); cur 990 fs/xfs/libxfs/xfs_refcount.c struct xfs_btree_cur *cur, cur 1005 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_increase(cur->bc_mp, cur->bc_private.a.agno, cur 1008 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_decrease(cur->bc_mp, cur->bc_private.a.agno, cur 1014 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_split_extent(cur, agbno, &shape_changed); cur 1020 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_split_extent(cur, agbno + aglen, &shape_changed); cur 1029 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_merge_extents(cur, new_agbno, new_aglen, adj, cur 1036 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.priv.refc.shape_changes++; cur 1039 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_adjust_extents(cur, new_agbno, new_aglen, cur 1047 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_adjust_error(cur->bc_mp, cur->bc_private.a.agno, cur 1241 fs/xfs/libxfs/xfs_refcount.c struct xfs_btree_cur *cur, cur 1253 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_find_shared(cur->bc_mp, cur->bc_private.a.agno, cur 1261 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_lookup_le(cur, agbno, &have); cur 1266 fs/xfs/libxfs/xfs_refcount.c error = xfs_btree_increment(cur, 0, &have); cur 1272 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_get_rec(cur, &tmp, &i); cur 1275 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, i == 1, out_error); cur 1279 fs/xfs/libxfs/xfs_refcount.c error = xfs_btree_increment(cur, 0, &have); cur 1284 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_get_rec(cur, &tmp, &i); cur 1287 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, i == 1, out_error); cur 1307 fs/xfs/libxfs/xfs_refcount.c error = xfs_btree_increment(cur, 0, &have); cur 1312 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_get_rec(cur, &tmp, &i); cur 1315 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, i == 1, out_error); cur 1323 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_find_shared_result(cur->bc_mp, cur 1324 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.agno, *fbno, *flen); cur 1328 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_find_shared_error(cur->bc_mp, cur 1329 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.agno, error, _RET_IP_); cur 1387 fs/xfs/libxfs/xfs_refcount.c struct xfs_btree_cur *cur, cur 1400 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_lookup_ge(cur, agbno, &found_rec); cur 1403 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_get_rec(cur, &ext, &found_rec); cur 1407 fs/xfs/libxfs/xfs_refcount.c ext.rc_startblock = cur->bc_mp->m_sb.sb_agblocks + cur 1416 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, cur 1422 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_modify_extent(cur->bc_mp, cur 1423 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.agno, &tmp); cur 1425 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_insert(cur, &tmp, cur 1429 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, cur 1434 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, cur 1436 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, cur 1438 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, cur 1442 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_modify_extent(cur->bc_mp, cur 1443 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.agno, &ext); cur 1444 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_delete(cur, &found_rec); cur 1447 fs/xfs/libxfs/xfs_refcount.c XFS_WANT_CORRUPTED_GOTO(cur->bc_mp, cur 1456 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_modify_extent_error(cur->bc_mp, cur 1457 fs/xfs/libxfs/xfs_refcount.c cur->bc_private.a.agno, error, _RET_IP_); cur 1466 fs/xfs/libxfs/xfs_refcount.c struct xfs_btree_cur *cur, cur 1479 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_split_extent(cur, agbno, &shape_changed); cur 1483 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_split_extent(cur, agbno + aglen, &shape_changed); cur 1490 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_merge_extents(cur, &agbno, &aglen, adj, cur 1496 fs/xfs/libxfs/xfs_refcount.c error = xfs_refcount_adjust_cow_extents(cur, agbno, aglen, adj); cur 1503 fs/xfs/libxfs/xfs_refcount.c trace_xfs_refcount_adjust_cow_error(cur->bc_mp, cur->bc_private.a.agno, cur 1587 fs/xfs/libxfs/xfs_refcount.c struct xfs_btree_cur *cur, cur 1611 fs/xfs/libxfs/xfs_refcount.c struct xfs_btree_cur *cur; cur 1647 fs/xfs/libxfs/xfs_refcount.c cur = xfs_refcountbt_init_cursor(mp, tp, agbp, agno); cur 1654 fs/xfs/libxfs/xfs_refcount.c error = xfs_btree_query_range(cur, &low, &high, cur 1656 fs/xfs/libxfs/xfs_refcount.c xfs_btree_del_cursor(cur, error); cur 1703 fs/xfs/libxfs/xfs_refcount.c struct xfs_btree_cur *cur, cur 1716 fs/xfs/libxfs/xfs_refcount.c return xfs_btree_has_record(cur, &low, &high, exists); cur 9 fs/xfs/libxfs/xfs_refcount.h extern int xfs_refcount_lookup_le(struct xfs_btree_cur *cur, cur 11 fs/xfs/libxfs/xfs_refcount.h extern int xfs_refcount_lookup_ge(struct xfs_btree_cur *cur, cur 13 fs/xfs/libxfs/xfs_refcount.h extern int xfs_refcount_lookup_eq(struct xfs_btree_cur *cur, cur 15 fs/xfs/libxfs/xfs_refcount.h extern int xfs_refcount_get_rec(struct xfs_btree_cur *cur, cur 44 fs/xfs/libxfs/xfs_refcount.h extern int xfs_refcount_find_shared(struct xfs_btree_cur *cur, cur 71 fs/xfs/libxfs/xfs_refcount.h extern int xfs_refcount_has_record(struct xfs_btree_cur *cur, cur 76 fs/xfs/libxfs/xfs_refcount.h extern int xfs_refcount_insert(struct xfs_btree_cur *cur, cur 25 fs/xfs/libxfs/xfs_refcount_btree.c struct xfs_btree_cur *cur) cur 27 fs/xfs/libxfs/xfs_refcount_btree.c return xfs_refcountbt_init_cursor(cur->bc_mp, cur->bc_tp, cur 28 fs/xfs/libxfs/xfs_refcount_btree.c cur->bc_private.a.agbp, cur->bc_private.a.agno); cur 33 fs/xfs/libxfs/xfs_refcount_btree.c struct xfs_btree_cur *cur, cur 37 fs/xfs/libxfs/xfs_refcount_btree.c struct xfs_buf *agbp = cur->bc_private.a.agbp; cur 40 fs/xfs/libxfs/xfs_refcount_btree.c struct xfs_perag *pag = xfs_perag_get(cur->bc_mp, seqno); cur 49 fs/xfs/libxfs/xfs_refcount_btree.c xfs_alloc_log_agf(cur->bc_tp, agbp, cur 55 fs/xfs/libxfs/xfs_refcount_btree.c struct xfs_btree_cur *cur, cur 60 fs/xfs/libxfs/xfs_refcount_btree.c struct xfs_buf *agbp = cur->bc_private.a.agbp; cur 66 fs/xfs/libxfs/xfs_refcount_btree.c args.tp = cur->bc_tp; cur 67 fs/xfs/libxfs/xfs_refcount_btree.c args.mp = cur->bc_mp; cur 69 fs/xfs/libxfs/xfs_refcount_btree.c args.fsbno = XFS_AGB_TO_FSB(cur->bc_mp, cur->bc_private.a.agno, cur 78 fs/xfs/libxfs/xfs_refcount_btree.c trace_xfs_refcountbt_alloc_block(cur->bc_mp, cur->bc_private.a.agno, cur 84 fs/xfs/libxfs/xfs_refcount_btree.c ASSERT(args.agno == cur->bc_private.a.agno); cur 89 fs/xfs/libxfs/xfs_refcount_btree.c xfs_alloc_log_agf(cur->bc_tp, agbp, XFS_AGF_REFCOUNT_BLOCKS); cur 100 fs/xfs/libxfs/xfs_refcount_btree.c struct xfs_btree_cur *cur, cur 103 fs/xfs/libxfs/xfs_refcount_btree.c struct xfs_mount *mp = cur->bc_mp; cur 104 fs/xfs/libxfs/xfs_refcount_btree.c struct xfs_buf *agbp = cur->bc_private.a.agbp; cur 109 fs/xfs/libxfs/xfs_refcount_btree.c trace_xfs_refcountbt_free_block(cur->bc_mp, cur->bc_private.a.agno, cur 110 fs/xfs/libxfs/xfs_refcount_btree.c XFS_FSB_TO_AGBNO(cur->bc_mp, fsbno), 1); cur 112 fs/xfs/libxfs/xfs_refcount_btree.c xfs_alloc_log_agf(cur->bc_tp, agbp, XFS_AGF_REFCOUNT_BLOCKS); cur 113 fs/xfs/libxfs/xfs_refcount_btree.c error = xfs_free_extent(cur->bc_tp, fsbno, 1, &XFS_RMAP_OINFO_REFC, cur 123 fs/xfs/libxfs/xfs_refcount_btree.c struct xfs_btree_cur *cur, cur 126 fs/xfs/libxfs/xfs_refcount_btree.c return cur->bc_mp->m_refc_mnr[level != 0]; cur 131 fs/xfs/libxfs/xfs_refcount_btree.c struct xfs_btree_cur *cur, cur 134 fs/xfs/libxfs/xfs_refcount_btree.c return cur->bc_mp->m_refc_mxr[level != 0]; cur 159 fs/xfs/libxfs/xfs_refcount_btree.c struct xfs_btree_cur *cur, cur 162 fs/xfs/libxfs/xfs_refcount_btree.c rec->refc.rc_startblock = cpu_to_be32(cur->bc_rec.rc.rc_startblock); cur 163 fs/xfs/libxfs/xfs_refcount_btree.c rec->refc.rc_blockcount = cpu_to_be32(cur->bc_rec.rc.rc_blockcount); cur 164 fs/xfs/libxfs/xfs_refcount_btree.c rec->refc.rc_refcount = cpu_to_be32(cur->bc_rec.rc.rc_refcount); cur 169 fs/xfs/libxfs/xfs_refcount_btree.c struct xfs_btree_cur *cur, cur 172 fs/xfs/libxfs/xfs_refcount_btree.c struct xfs_agf *agf = XFS_BUF_TO_AGF(cur->bc_private.a.agbp); cur 174 fs/xfs/libxfs/xfs_refcount_btree.c ASSERT(cur->bc_private.a.agno == be32_to_cpu(agf->agf_seqno)); cur 181 fs/xfs/libxfs/xfs_refcount_btree.c struct xfs_btree_cur *cur, cur 184 fs/xfs/libxfs/xfs_refcount_btree.c struct xfs_refcount_irec *rec = &cur->bc_rec.rc; cur 192 fs/xfs/libxfs/xfs_refcount_btree.c struct xfs_btree_cur *cur, cur 273 fs/xfs/libxfs/xfs_refcount_btree.c struct xfs_btree_cur *cur, cur 283 fs/xfs/libxfs/xfs_refcount_btree.c struct xfs_btree_cur *cur, cur 324 fs/xfs/libxfs/xfs_refcount_btree.c struct xfs_btree_cur *cur; cur 328 fs/xfs/libxfs/xfs_refcount_btree.c cur = kmem_zone_zalloc(xfs_btree_cur_zone, KM_NOFS); cur 330 fs/xfs/libxfs/xfs_refcount_btree.c cur->bc_tp = tp; cur 331 fs/xfs/libxfs/xfs_refcount_btree.c cur->bc_mp = mp; cur 332 fs/xfs/libxfs/xfs_refcount_btree.c cur->bc_btnum = XFS_BTNUM_REFC; cur 333 fs/xfs/libxfs/xfs_refcount_btree.c cur->bc_blocklog = mp->m_sb.sb_blocklog; cur 334 fs/xfs/libxfs/xfs_refcount_btree.c cur->bc_ops = &xfs_refcountbt_ops; cur 335 fs/xfs/libxfs/xfs_refcount_btree.c cur->bc_statoff = XFS_STATS_CALC_INDEX(xs_refcbt_2); cur 337 fs/xfs/libxfs/xfs_refcount_btree.c cur->bc_nlevels = be32_to_cpu(agf->agf_refcount_level); cur 339 fs/xfs/libxfs/xfs_refcount_btree.c cur->bc_private.a.agbp = agbp; cur 340 fs/xfs/libxfs/xfs_refcount_btree.c cur->bc_private.a.agno = agno; cur 341 fs/xfs/libxfs/xfs_refcount_btree.c cur->bc_flags |= XFS_BTREE_CRC_BLOCKS; cur 343 fs/xfs/libxfs/xfs_refcount_btree.c cur->bc_private.a.priv.refc.nr_ops = 0; cur 344 fs/xfs/libxfs/xfs_refcount_btree.c cur->bc_private.a.priv.refc.shape_changes = 0; cur 346 fs/xfs/libxfs/xfs_refcount_btree.c return cur; cur 31 fs/xfs/libxfs/xfs_rmap.c struct xfs_btree_cur *cur, cur 39 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_startblock = bno; cur 40 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_blockcount = len; cur 41 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_owner = owner; cur 42 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_offset = offset; cur 43 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_flags = flags; cur 44 fs/xfs/libxfs/xfs_rmap.c return xfs_btree_lookup(cur, XFS_LOOKUP_LE, stat); cur 53 fs/xfs/libxfs/xfs_rmap.c struct xfs_btree_cur *cur, cur 61 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_startblock = bno; cur 62 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_blockcount = len; cur 63 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_owner = owner; cur 64 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_offset = offset; cur 65 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_flags = flags; cur 66 fs/xfs/libxfs/xfs_rmap.c return xfs_btree_lookup(cur, XFS_LOOKUP_EQ, stat); cur 76 fs/xfs/libxfs/xfs_rmap.c struct xfs_btree_cur *cur, cur 82 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_update(cur->bc_mp, cur->bc_private.a.agno, cur 91 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_update(cur, &rec); cur 93 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_update_error(cur->bc_mp, cur 94 fs/xfs/libxfs/xfs_rmap.c cur->bc_private.a.agno, error, _RET_IP_); cur 183 fs/xfs/libxfs/xfs_rmap.c struct xfs_btree_cur *cur, cur 187 fs/xfs/libxfs/xfs_rmap.c struct xfs_mount *mp = cur->bc_mp; cur 188 fs/xfs/libxfs/xfs_rmap.c xfs_agnumber_t agno = cur->bc_private.a.agno; cur 192 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_get_rec(cur, &rec, stat); cur 244 fs/xfs/libxfs/xfs_rmap.c struct xfs_btree_cur *cur, cur 250 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_find_left_neighbor_candidate(cur->bc_mp, cur 251 fs/xfs/libxfs/xfs_rmap.c cur->bc_private.a.agno, rec->rm_startblock, cur 274 fs/xfs/libxfs/xfs_rmap.c struct xfs_btree_cur *cur, cur 302 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_find_left_neighbor_query(cur->bc_mp, cur 303 fs/xfs/libxfs/xfs_rmap.c cur->bc_private.a.agno, bno, 0, owner, offset, flags); cur 305 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_query_range(cur, &info.high, &info.high, cur 310 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_find_left_neighbor_result(cur->bc_mp, cur 311 fs/xfs/libxfs/xfs_rmap.c cur->bc_private.a.agno, irec->rm_startblock, cur 320 fs/xfs/libxfs/xfs_rmap.c struct xfs_btree_cur *cur, cur 326 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_lookup_le_range_candidate(cur->bc_mp, cur 327 fs/xfs/libxfs/xfs_rmap.c cur->bc_private.a.agno, rec->rm_startblock, cur 352 fs/xfs/libxfs/xfs_rmap.c struct xfs_btree_cur *cur, cur 375 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_lookup_le_range(cur->bc_mp, cur 376 fs/xfs/libxfs/xfs_rmap.c cur->bc_private.a.agno, bno, 0, owner, offset, flags); cur 377 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_query_range(cur, &info.high, &info.high, cur 382 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_lookup_le_range_result(cur->bc_mp, cur 383 fs/xfs/libxfs/xfs_rmap.c cur->bc_private.a.agno, irec->rm_startblock, cur 453 fs/xfs/libxfs/xfs_rmap.c struct xfs_btree_cur *cur, cur 459 fs/xfs/libxfs/xfs_rmap.c struct xfs_mount *mp = cur->bc_mp; cur 474 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_unmap(mp, cur->bc_private.a.agno, bno, len, cur 482 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_lookup_le(cur, bno, len, owner, offset, flags, &i); cur 487 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_get_rec(cur, <rec, &i); cur 491 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_lookup_le_range_result(cur->bc_mp, cur 492 fs/xfs/libxfs/xfs_rmap.c cur->bc_private.a.agno, ltrec.rm_startblock, cur 521 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_increment(cur, 0, &i); cur 526 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_get_rec(cur, &rtrec, &i); cur 547 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_delete(mp, cur->bc_private.a.agno, cur 551 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_delete(cur, &i); cur 570 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, <rec); cur 585 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, <rec); cur 605 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, <rec); cur 609 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_increment(cur, 0, &i); cur 613 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_startblock = bno + len; cur 614 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_blockcount = orig_len - len - cur 616 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_owner = ltrec.rm_owner; cur 618 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_offset = 0; cur 620 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_offset = offset + len; cur 621 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_flags = flags; cur 622 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_insert(mp, cur->bc_private.a.agno, cur 623 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_startblock, cur 624 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_blockcount, cur 625 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_owner, cur 626 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_offset, cur 627 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_flags); cur 628 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_insert(cur, &i); cur 634 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_unmap_done(mp, cur->bc_private.a.agno, bno, len, cur 638 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_unmap_error(mp, cur->bc_private.a.agno, cur 656 fs/xfs/libxfs/xfs_rmap.c struct xfs_btree_cur *cur; cur 662 fs/xfs/libxfs/xfs_rmap.c cur = xfs_rmapbt_init_cursor(mp, tp, agbp, agno); cur 664 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_unmap(cur, bno, len, false, oinfo); cur 666 fs/xfs/libxfs/xfs_rmap.c xfs_btree_del_cursor(cur, error); cur 705 fs/xfs/libxfs/xfs_rmap.c struct xfs_btree_cur *cur, cur 711 fs/xfs/libxfs/xfs_rmap.c struct xfs_mount *mp = cur->bc_mp; cur 729 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_map(mp, cur->bc_private.a.agno, bno, len, cur 738 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_lookup_le(cur, bno, len, owner, offset, flags, cur 743 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_get_rec(cur, <rec, &have_lt); cur 747 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_lookup_le_range_result(cur->bc_mp, cur 748 fs/xfs/libxfs/xfs_rmap.c cur->bc_private.a.agno, ltrec.rm_startblock, cur 765 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_increment(cur, 0, &have_gt); cur 769 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_get_rec(cur, >rec, &have_gt); cur 775 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_find_right_neighbor_result(cur->bc_mp, cur 776 fs/xfs/libxfs/xfs_rmap.c cur->bc_private.a.agno, gtrec.rm_startblock, cur 815 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_delete(mp, cur->bc_private.a.agno, cur 821 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_delete(cur, &i); cur 828 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_decrement(cur, 0, &have_gt); cur 831 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, <rec); cur 850 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, >rec); cur 858 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_startblock = bno; cur 859 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_blockcount = len; cur 860 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_owner = owner; cur 861 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_offset = offset; cur 862 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_flags = flags; cur 863 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_insert(mp, cur->bc_private.a.agno, bno, len, cur 865 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_insert(cur, &i); cur 871 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_map_done(mp, cur->bc_private.a.agno, bno, len, cur 875 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_map_error(mp, cur->bc_private.a.agno, cur 893 fs/xfs/libxfs/xfs_rmap.c struct xfs_btree_cur *cur; cur 899 fs/xfs/libxfs/xfs_rmap.c cur = xfs_rmapbt_init_cursor(mp, tp, agbp, agno); cur 900 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_map(cur, bno, len, false, oinfo); cur 902 fs/xfs/libxfs/xfs_rmap.c xfs_btree_del_cursor(cur, error); cur 924 fs/xfs/libxfs/xfs_rmap.c struct xfs_btree_cur *cur, cur 930 fs/xfs/libxfs/xfs_rmap.c struct xfs_mount *mp = cur->bc_mp; cur 949 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_convert(mp, cur->bc_private.a.agno, bno, len, cur 957 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_lookup_le(cur, bno, len, owner, offset, oldext, &i); cur 962 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_get_rec(cur, &PREV, &i); cur 966 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_lookup_le_range_result(cur->bc_mp, cur 967 fs/xfs/libxfs/xfs_rmap.c cur->bc_private.a.agno, PREV.rm_startblock, cur 990 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_decrement(cur, 0, &i); cur 995 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_get_rec(cur, &LEFT, &i); cur 1002 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_find_left_neighbor_result(cur->bc_mp, cur 1003 fs/xfs/libxfs/xfs_rmap.c cur->bc_private.a.agno, LEFT.rm_startblock, cur 1017 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_increment(cur, 0, &i); cur 1021 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_increment(cur, 0, &i); cur 1026 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_get_rec(cur, &RIGHT, &i); cur 1032 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_find_right_neighbor_result(cur->bc_mp, cur 1033 fs/xfs/libxfs/xfs_rmap.c cur->bc_private.a.agno, RIGHT.rm_startblock, cur 1051 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_convert_state(mp, cur->bc_private.a.agno, state, cur 1055 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_lookup_le(cur, bno, len, owner, offset, oldext, &i); cur 1071 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_increment(cur, 0, &i); cur 1075 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_delete(mp, cur->bc_private.a.agno, cur 1079 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_delete(cur, &i); cur 1083 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_decrement(cur, 0, &i); cur 1087 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_delete(mp, cur->bc_private.a.agno, cur 1091 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_delete(cur, &i); cur 1095 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_decrement(cur, 0, &i); cur 1101 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, &NEW); cur 1111 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_delete(mp, cur->bc_private.a.agno, cur 1115 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_delete(cur, &i); cur 1119 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_decrement(cur, 0, &i); cur 1125 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, &NEW); cur 1135 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_increment(cur, 0, &i); cur 1139 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_delete(mp, cur->bc_private.a.agno, cur 1143 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_delete(cur, &i); cur 1147 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_decrement(cur, 0, &i); cur 1154 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, &NEW); cur 1167 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, &NEW); cur 1181 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, &NEW); cur 1184 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_decrement(cur, 0, &i); cur 1189 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, &NEW); cur 1203 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, &NEW); cur 1211 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r = NEW; cur 1212 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_insert(mp, cur->bc_private.a.agno, bno, cur 1214 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_insert(cur, &i); cur 1227 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, &NEW); cur 1230 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_increment(cur, 0, &i); cur 1237 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, &NEW); cur 1249 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, &NEW); cur 1252 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_lookup_eq(cur, bno, len, owner, offset, cur 1262 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r = NEW; cur 1263 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_insert(mp, cur->bc_private.a.agno, bno, cur 1265 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_insert(cur, &i); cur 1284 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, &NEW); cur 1290 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r = NEW; cur 1291 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_insert(mp, cur->bc_private.a.agno, cur 1295 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_insert(cur, &i); cur 1304 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_lookup_eq(cur, bno, len, owner, offset, cur 1310 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_flags &= ~XFS_RMAP_UNWRITTEN; cur 1311 fs/xfs/libxfs/xfs_rmap.c cur->bc_rec.r.rm_flags |= newext; cur 1312 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_insert(mp, cur->bc_private.a.agno, bno, len, cur 1314 fs/xfs/libxfs/xfs_rmap.c error = xfs_btree_insert(cur, &i); cur 1333 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_convert_done(mp, cur->bc_private.a.agno, bno, len, cur 1337 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_convert_error(cur->bc_mp, cur 1338 fs/xfs/libxfs/xfs_rmap.c cur->bc_private.a.agno, error, _RET_IP_); cur 1349 fs/xfs/libxfs/xfs_rmap.c struct xfs_btree_cur *cur, cur 1355 fs/xfs/libxfs/xfs_rmap.c struct xfs_mount *mp = cur->bc_mp; cur 1374 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_convert(mp, cur->bc_private.a.agno, bno, len, cur 1382 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_lookup_le_range(cur, bno, owner, offset, flags, cur 1403 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_find_left_neighbor(cur, bno, owner, offset, newext, cur 1417 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_lookup_eq(cur, bno + len, len, owner, offset + len, cur 1423 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_get_rec(cur, &RIGHT, &i); cur 1429 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_find_right_neighbor_result(cur->bc_mp, cur 1430 fs/xfs/libxfs/xfs_rmap.c cur->bc_private.a.agno, RIGHT.rm_startblock, cur 1446 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_convert_state(mp, cur->bc_private.a.agno, state, cur 1459 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_delete(cur, RIGHT.rm_startblock, cur 1464 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_delete(cur, PREV.rm_startblock, cur 1470 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_lookup_eq(cur, NEW.rm_startblock, cur 1477 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, &NEW); cur 1487 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_delete(cur, PREV.rm_startblock, cur 1493 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_lookup_eq(cur, NEW.rm_startblock, cur 1500 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, &NEW); cur 1510 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_delete(cur, RIGHT.rm_startblock, cur 1516 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_lookup_eq(cur, NEW.rm_startblock, cur 1524 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, &NEW); cur 1536 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_lookup_eq(cur, NEW.rm_startblock, cur 1543 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, &NEW); cur 1554 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_delete(cur, NEW.rm_startblock, cur 1562 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_insert(cur, NEW.rm_startblock, cur 1568 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_lookup_eq(cur, NEW.rm_startblock, cur 1575 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, &NEW); cur 1586 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_delete(cur, NEW.rm_startblock, cur 1594 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_insert(cur, NEW.rm_startblock, cur 1599 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_insert(cur, bno, len, owner, offset, newext); cur 1610 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_lookup_eq(cur, NEW.rm_startblock, cur 1617 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, &NEW); cur 1621 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_delete(cur, NEW.rm_startblock, cur 1629 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_insert(cur, NEW.rm_startblock, cur 1642 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_lookup_eq(cur, NEW.rm_startblock, cur 1649 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, &NEW); cur 1652 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_insert(cur, bno, len, owner, offset, newext); cur 1670 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_insert(cur, NEW.rm_startblock, cur 1677 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_lookup_eq(cur, NEW.rm_startblock, cur 1684 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, &NEW); cur 1693 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_insert(cur, NEW.rm_startblock, cur 1713 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_convert_done(mp, cur->bc_private.a.agno, bno, len, cur 1717 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_convert_error(cur->bc_mp, cur 1718 fs/xfs/libxfs/xfs_rmap.c cur->bc_private.a.agno, error, _RET_IP_); cur 1738 fs/xfs/libxfs/xfs_rmap.c struct xfs_btree_cur *cur, cur 1744 fs/xfs/libxfs/xfs_rmap.c struct xfs_mount *mp = cur->bc_mp; cur 1756 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_unmap(mp, cur->bc_private.a.agno, bno, len, cur 1764 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_lookup_le_range(cur, bno, owner, offset, flags, cur 1790 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_delete(cur, ltrec.rm_startblock, cur 1808 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_delete(cur, ltrec.rm_startblock, cur 1818 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_insert(cur, ltrec.rm_startblock, cur 1834 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_lookup_eq(cur, ltrec.rm_startblock, cur 1841 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, <rec); cur 1860 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_lookup_eq(cur, ltrec.rm_startblock, cur 1867 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, <rec); cur 1872 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_insert(cur, bno + len, cur 1880 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_unmap_done(mp, cur->bc_private.a.agno, bno, len, cur 1884 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_unmap_error(cur->bc_mp, cur 1885 fs/xfs/libxfs/xfs_rmap.c cur->bc_private.a.agno, error, _RET_IP_); cur 1900 fs/xfs/libxfs/xfs_rmap.c struct xfs_btree_cur *cur, cur 1906 fs/xfs/libxfs/xfs_rmap.c struct xfs_mount *mp = cur->bc_mp; cur 1920 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_map(mp, cur->bc_private.a.agno, bno, len, cur 1924 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_find_left_neighbor(cur, bno, owner, offset, flags, cur 1933 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_lookup_eq(cur, bno + len, len, owner, offset + len, cur 1938 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_get_rec(cur, >rec, &have_gt); cur 1942 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_find_right_neighbor_result(cur->bc_mp, cur 1943 fs/xfs/libxfs/xfs_rmap.c cur->bc_private.a.agno, gtrec.rm_startblock, cur 1977 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_delete(cur, gtrec.rm_startblock, cur 1985 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_lookup_eq(cur, ltrec.rm_startblock, cur 1992 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_update(cur, <rec); cur 2008 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_delete(cur, gtrec.rm_startblock, cur 2018 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_insert(cur, gtrec.rm_startblock, cur 2028 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_insert(cur, bno, len, owner, offset, flags); cur 2033 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_map_done(mp, cur->bc_private.a.agno, bno, len, cur 2037 fs/xfs/libxfs/xfs_rmap.c trace_xfs_rmap_map_error(cur->bc_mp, cur 2038 fs/xfs/libxfs/xfs_rmap.c cur->bc_private.a.agno, error, _RET_IP_); cur 2045 fs/xfs/libxfs/xfs_rmap.c struct xfs_btree_cur *cur, cur 2059 fs/xfs/libxfs/xfs_rmap.c return xfs_rmap_map(cur, rmap->rm_startblock, cur 2064 fs/xfs/libxfs/xfs_rmap.c return xfs_rmap_map_shared(cur, rmap->rm_startblock, cur 2078 fs/xfs/libxfs/xfs_rmap.c struct xfs_btree_cur *cur, cur 2089 fs/xfs/libxfs/xfs_rmap.c return query->fn(cur, &irec, query->priv); cur 2095 fs/xfs/libxfs/xfs_rmap.c struct xfs_btree_cur *cur, cur 2109 fs/xfs/libxfs/xfs_rmap.c return xfs_btree_query_range(cur, &low_brec, &high_brec, cur 2116 fs/xfs/libxfs/xfs_rmap.c struct xfs_btree_cur *cur, cur 2124 fs/xfs/libxfs/xfs_rmap.c return xfs_btree_query_all(cur, xfs_rmap_query_range_helper, &query); cur 2428 fs/xfs/libxfs/xfs_rmap.c struct xfs_btree_cur *cur, cur 2441 fs/xfs/libxfs/xfs_rmap.c return xfs_btree_has_record(cur, &low, &high, exists); cur 2453 fs/xfs/libxfs/xfs_rmap.c struct xfs_btree_cur *cur, cur 2470 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_lookup_le(cur, bno, len, owner, offset, flags, cur 2479 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_get_rec(cur, &irec, &has_record); cur 2502 fs/xfs/libxfs/xfs_rmap.c struct xfs_btree_cur *cur, cur 2521 fs/xfs/libxfs/xfs_rmap.c struct xfs_btree_cur *cur, cur 2539 fs/xfs/libxfs/xfs_rmap.c error = xfs_rmap_query_range(cur, &low, &high, cur 122 fs/xfs/libxfs/xfs_rmap.h int xfs_rmap_lookup_le(struct xfs_btree_cur *cur, xfs_agblock_t bno, cur 125 fs/xfs/libxfs/xfs_rmap.h int xfs_rmap_lookup_eq(struct xfs_btree_cur *cur, xfs_agblock_t bno, cur 131 fs/xfs/libxfs/xfs_rmap.h int xfs_rmap_get_rec(struct xfs_btree_cur *cur, struct xfs_rmap_irec *irec, cur 135 fs/xfs/libxfs/xfs_rmap.h struct xfs_btree_cur *cur, cur 139 fs/xfs/libxfs/xfs_rmap.h int xfs_rmap_query_range(struct xfs_btree_cur *cur, cur 142 fs/xfs/libxfs/xfs_rmap.h int xfs_rmap_query_all(struct xfs_btree_cur *cur, xfs_rmap_query_range_fn fn, cur 184 fs/xfs/libxfs/xfs_rmap.h int xfs_rmap_find_left_neighbor(struct xfs_btree_cur *cur, xfs_agblock_t bno, cur 187 fs/xfs/libxfs/xfs_rmap.h int xfs_rmap_lookup_le_range(struct xfs_btree_cur *cur, xfs_agblock_t bno, cur 195 fs/xfs/libxfs/xfs_rmap.h int xfs_rmap_has_record(struct xfs_btree_cur *cur, xfs_agblock_t bno, cur 197 fs/xfs/libxfs/xfs_rmap.h int xfs_rmap_record_exists(struct xfs_btree_cur *cur, xfs_agblock_t bno, cur 200 fs/xfs/libxfs/xfs_rmap.h int xfs_rmap_has_other_keys(struct xfs_btree_cur *cur, xfs_agblock_t bno, cur 203 fs/xfs/libxfs/xfs_rmap.h int xfs_rmap_map_raw(struct xfs_btree_cur *cur, struct xfs_rmap_irec *rmap); cur 51 fs/xfs/libxfs/xfs_rmap_btree.c struct xfs_btree_cur *cur) cur 53 fs/xfs/libxfs/xfs_rmap_btree.c return xfs_rmapbt_init_cursor(cur->bc_mp, cur->bc_tp, cur 54 fs/xfs/libxfs/xfs_rmap_btree.c cur->bc_private.a.agbp, cur->bc_private.a.agno); cur 59 fs/xfs/libxfs/xfs_rmap_btree.c struct xfs_btree_cur *cur, cur 63 fs/xfs/libxfs/xfs_rmap_btree.c struct xfs_buf *agbp = cur->bc_private.a.agbp; cur 66 fs/xfs/libxfs/xfs_rmap_btree.c int btnum = cur->bc_btnum; cur 67 fs/xfs/libxfs/xfs_rmap_btree.c struct xfs_perag *pag = xfs_perag_get(cur->bc_mp, seqno); cur 76 fs/xfs/libxfs/xfs_rmap_btree.c xfs_alloc_log_agf(cur->bc_tp, agbp, XFS_AGF_ROOTS | XFS_AGF_LEVELS); cur 81 fs/xfs/libxfs/xfs_rmap_btree.c struct xfs_btree_cur *cur, cur 86 fs/xfs/libxfs/xfs_rmap_btree.c struct xfs_buf *agbp = cur->bc_private.a.agbp; cur 92 fs/xfs/libxfs/xfs_rmap_btree.c error = xfs_alloc_get_freelist(cur->bc_tp, cur->bc_private.a.agbp, cur 97 fs/xfs/libxfs/xfs_rmap_btree.c trace_xfs_rmapbt_alloc_block(cur->bc_mp, cur->bc_private.a.agno, cur 104 fs/xfs/libxfs/xfs_rmap_btree.c xfs_extent_busy_reuse(cur->bc_mp, cur->bc_private.a.agno, bno, 1, cur 107 fs/xfs/libxfs/xfs_rmap_btree.c xfs_trans_agbtree_delta(cur->bc_tp, 1); cur 110 fs/xfs/libxfs/xfs_rmap_btree.c xfs_alloc_log_agf(cur->bc_tp, agbp, XFS_AGF_RMAP_BLOCKS); cur 112 fs/xfs/libxfs/xfs_rmap_btree.c xfs_ag_resv_rmapbt_alloc(cur->bc_mp, cur->bc_private.a.agno); cur 120 fs/xfs/libxfs/xfs_rmap_btree.c struct xfs_btree_cur *cur, cur 123 fs/xfs/libxfs/xfs_rmap_btree.c struct xfs_buf *agbp = cur->bc_private.a.agbp; cur 128 fs/xfs/libxfs/xfs_rmap_btree.c bno = xfs_daddr_to_agbno(cur->bc_mp, XFS_BUF_ADDR(bp)); cur 129 fs/xfs/libxfs/xfs_rmap_btree.c trace_xfs_rmapbt_free_block(cur->bc_mp, cur->bc_private.a.agno, cur 132 fs/xfs/libxfs/xfs_rmap_btree.c xfs_alloc_log_agf(cur->bc_tp, agbp, XFS_AGF_RMAP_BLOCKS); cur 133 fs/xfs/libxfs/xfs_rmap_btree.c error = xfs_alloc_put_freelist(cur->bc_tp, agbp, NULL, bno, 1); cur 137 fs/xfs/libxfs/xfs_rmap_btree.c xfs_extent_busy_insert(cur->bc_tp, be32_to_cpu(agf->agf_seqno), bno, 1, cur 139 fs/xfs/libxfs/xfs_rmap_btree.c xfs_trans_agbtree_delta(cur->bc_tp, -1); cur 141 fs/xfs/libxfs/xfs_rmap_btree.c xfs_ag_resv_rmapbt_free(cur->bc_mp, cur->bc_private.a.agno); cur 148 fs/xfs/libxfs/xfs_rmap_btree.c struct xfs_btree_cur *cur, cur 151 fs/xfs/libxfs/xfs_rmap_btree.c return cur->bc_mp->m_rmap_mnr[level != 0]; cur 156 fs/xfs/libxfs/xfs_rmap_btree.c struct xfs_btree_cur *cur, cur 159 fs/xfs/libxfs/xfs_rmap_btree.c return cur->bc_mp->m_rmap_mxr[level != 0]; cur 203 fs/xfs/libxfs/xfs_rmap_btree.c struct xfs_btree_cur *cur, cur 206 fs/xfs/libxfs/xfs_rmap_btree.c rec->rmap.rm_startblock = cpu_to_be32(cur->bc_rec.r.rm_startblock); cur 207 fs/xfs/libxfs/xfs_rmap_btree.c rec->rmap.rm_blockcount = cpu_to_be32(cur->bc_rec.r.rm_blockcount); cur 208 fs/xfs/libxfs/xfs_rmap_btree.c rec->rmap.rm_owner = cpu_to_be64(cur->bc_rec.r.rm_owner); cur 210 fs/xfs/libxfs/xfs_rmap_btree.c xfs_rmap_irec_offset_pack(&cur->bc_rec.r)); cur 215 fs/xfs/libxfs/xfs_rmap_btree.c struct xfs_btree_cur *cur, cur 218 fs/xfs/libxfs/xfs_rmap_btree.c struct xfs_agf *agf = XFS_BUF_TO_AGF(cur->bc_private.a.agbp); cur 220 fs/xfs/libxfs/xfs_rmap_btree.c ASSERT(cur->bc_private.a.agno == be32_to_cpu(agf->agf_seqno)); cur 222 fs/xfs/libxfs/xfs_rmap_btree.c ptr->s = agf->agf_roots[cur->bc_btnum]; cur 227 fs/xfs/libxfs/xfs_rmap_btree.c struct xfs_btree_cur *cur, cur 230 fs/xfs/libxfs/xfs_rmap_btree.c struct xfs_rmap_irec *rec = &cur->bc_rec.r; cur 257 fs/xfs/libxfs/xfs_rmap_btree.c struct xfs_btree_cur *cur, cur 372 fs/xfs/libxfs/xfs_rmap_btree.c struct xfs_btree_cur *cur, cur 402 fs/xfs/libxfs/xfs_rmap_btree.c struct xfs_btree_cur *cur, cur 462 fs/xfs/libxfs/xfs_rmap_btree.c struct xfs_btree_cur *cur; cur 464 fs/xfs/libxfs/xfs_rmap_btree.c cur = kmem_zone_zalloc(xfs_btree_cur_zone, KM_NOFS); cur 465 fs/xfs/libxfs/xfs_rmap_btree.c cur->bc_tp = tp; cur 466 fs/xfs/libxfs/xfs_rmap_btree.c cur->bc_mp = mp; cur 468 fs/xfs/libxfs/xfs_rmap_btree.c cur->bc_btnum = XFS_BTNUM_RMAP; cur 469 fs/xfs/libxfs/xfs_rmap_btree.c cur->bc_flags = XFS_BTREE_CRC_BLOCKS | XFS_BTREE_OVERLAPPING; cur 470 fs/xfs/libxfs/xfs_rmap_btree.c cur->bc_blocklog = mp->m_sb.sb_blocklog; cur 471 fs/xfs/libxfs/xfs_rmap_btree.c cur->bc_ops = &xfs_rmapbt_ops; cur 472 fs/xfs/libxfs/xfs_rmap_btree.c cur->bc_nlevels = be32_to_cpu(agf->agf_levels[XFS_BTNUM_RMAP]); cur 473 fs/xfs/libxfs/xfs_rmap_btree.c cur->bc_statoff = XFS_STATS_CALC_INDEX(xs_rmap_2); cur 475 fs/xfs/libxfs/xfs_rmap_btree.c cur->bc_private.a.agbp = agbp; cur 476 fs/xfs/libxfs/xfs_rmap_btree.c cur->bc_private.a.agno = agno; cur 478 fs/xfs/libxfs/xfs_rmap_btree.c return cur; cur 346 fs/xfs/scrub/agheader.c struct xfs_btree_cur *cur, cur 71 fs/xfs/scrub/agheader_repair.c struct xfs_btree_cur *cur, cur 240 fs/xfs/scrub/agheader_repair.c struct xfs_btree_cur *cur = NULL; cur 248 fs/xfs/scrub/agheader_repair.c cur = xfs_allocbt_init_cursor(mp, sc->tp, agf_bp, sc->sa.agno, cur 250 fs/xfs/scrub/agheader_repair.c error = xfs_alloc_query_all(cur, xrep_agf_walk_allocbt, &raa); cur 253 fs/xfs/scrub/agheader_repair.c error = xfs_btree_count_blocks(cur, &blocks); cur 256 fs/xfs/scrub/agheader_repair.c xfs_btree_del_cursor(cur, error); cur 262 fs/xfs/scrub/agheader_repair.c cur = xfs_allocbt_init_cursor(mp, sc->tp, agf_bp, sc->sa.agno, cur 264 fs/xfs/scrub/agheader_repair.c error = xfs_btree_count_blocks(cur, &blocks); cur 267 fs/xfs/scrub/agheader_repair.c xfs_btree_del_cursor(cur, error); cur 271 fs/xfs/scrub/agheader_repair.c cur = xfs_rmapbt_init_cursor(mp, sc->tp, agf_bp, sc->sa.agno); cur 272 fs/xfs/scrub/agheader_repair.c error = xfs_btree_count_blocks(cur, &blocks); cur 275 fs/xfs/scrub/agheader_repair.c xfs_btree_del_cursor(cur, error); cur 283 fs/xfs/scrub/agheader_repair.c cur = xfs_refcountbt_init_cursor(mp, sc->tp, agf_bp, cur 285 fs/xfs/scrub/agheader_repair.c error = xfs_btree_count_blocks(cur, &blocks); cur 288 fs/xfs/scrub/agheader_repair.c xfs_btree_del_cursor(cur, error); cur 294 fs/xfs/scrub/agheader_repair.c xfs_btree_del_cursor(cur, error); cur 443 fs/xfs/scrub/agheader_repair.c struct xfs_btree_cur *cur, cur 456 fs/xfs/scrub/agheader_repair.c fsb = XFS_AGB_TO_FSB(cur->bc_mp, cur->bc_private.a.agno, cur 463 fs/xfs/scrub/agheader_repair.c return xfs_bitmap_set_btcur_path(&ra->agmetablocks, cur); cur 484 fs/xfs/scrub/agheader_repair.c struct xfs_btree_cur *cur; cur 494 fs/xfs/scrub/agheader_repair.c cur = xfs_rmapbt_init_cursor(mp, sc->tp, agf_bp, sc->sa.agno); cur 495 fs/xfs/scrub/agheader_repair.c error = xfs_rmap_query_all(cur, xrep_agfl_walk_rmap, &ra); cur 498 fs/xfs/scrub/agheader_repair.c xfs_btree_del_cursor(cur, error); cur 501 fs/xfs/scrub/agheader_repair.c cur = xfs_allocbt_init_cursor(mp, sc->tp, agf_bp, sc->sa.agno, cur 503 fs/xfs/scrub/agheader_repair.c error = xfs_bitmap_set_btblocks(&ra.agmetablocks, cur); cur 506 fs/xfs/scrub/agheader_repair.c xfs_btree_del_cursor(cur, error); cur 509 fs/xfs/scrub/agheader_repair.c cur = xfs_allocbt_init_cursor(mp, sc->tp, agf_bp, sc->sa.agno, cur 511 fs/xfs/scrub/agheader_repair.c error = xfs_bitmap_set_btblocks(&ra.agmetablocks, cur); cur 515 fs/xfs/scrub/agheader_repair.c xfs_btree_del_cursor(cur, error); cur 542 fs/xfs/scrub/agheader_repair.c xfs_btree_del_cursor(cur, error); cur 813 fs/xfs/scrub/agheader_repair.c struct xfs_btree_cur *cur; cur 820 fs/xfs/scrub/agheader_repair.c cur = xfs_inobt_init_cursor(mp, sc->tp, agi_bp, sc->sa.agno, cur 822 fs/xfs/scrub/agheader_repair.c error = xfs_ialloc_count_inodes(cur, &count, &freecount); cur 825 fs/xfs/scrub/agheader_repair.c xfs_btree_del_cursor(cur, error); cur 831 fs/xfs/scrub/agheader_repair.c xfs_btree_del_cursor(cur, error); cur 96 fs/xfs/scrub/alloc.c struct xfs_mount *mp = bs->cur->bc_mp; cur 97 fs/xfs/scrub/alloc.c xfs_agnumber_t agno = bs->cur->bc_private.a.agno; cur 107 fs/xfs/scrub/alloc.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 120 fs/xfs/scrub/alloc.c struct xfs_btree_cur *cur; cur 122 fs/xfs/scrub/alloc.c cur = which == XFS_BTNUM_BNO ? sc->sa.bno_cur : sc->sa.cnt_cur; cur 123 fs/xfs/scrub/alloc.c return xchk_btree(sc, cur, xchk_allocbt_rec, &XFS_RMAP_OINFO_AG, NULL); cur 252 fs/xfs/scrub/bitmap.c struct xfs_btree_cur *cur) cur 259 fs/xfs/scrub/bitmap.c for (i = 0; i < cur->bc_nlevels && cur->bc_ptrs[i] == 1; i++) { cur 260 fs/xfs/scrub/bitmap.c xfs_btree_get_block(cur, i, &bp); cur 263 fs/xfs/scrub/bitmap.c fsb = XFS_DADDR_TO_FSB(cur->bc_mp, bp->b_bn); cur 275 fs/xfs/scrub/bitmap.c struct xfs_btree_cur *cur, cur 283 fs/xfs/scrub/bitmap.c xfs_btree_get_block(cur, level, &bp); cur 287 fs/xfs/scrub/bitmap.c fsbno = XFS_DADDR_TO_FSB(cur->bc_mp, bp->b_bn); cur 295 fs/xfs/scrub/bitmap.c struct xfs_btree_cur *cur) cur 297 fs/xfs/scrub/bitmap.c return xfs_btree_visit_blocks(cur, xfs_bitmap_collect_btblock, bitmap); cur 32 fs/xfs/scrub/bitmap.h struct xfs_btree_cur *cur); cur 34 fs/xfs/scrub/bitmap.h struct xfs_btree_cur *cur); cur 377 fs/xfs/scrub/bmap.c struct xfs_inode *ip = bs->cur->bc_private.b.ip; cur 388 fs/xfs/scrub/bmap.c if (xfs_sb_version_hascrc(&bs->cur->bc_mp->m_sb) && cur 389 fs/xfs/scrub/bmap.c bs->cur->bc_ptrs[0] == 1) { cur 390 fs/xfs/scrub/bmap.c for (i = 0; i < bs->cur->bc_nlevels - 1; i++) { cur 391 fs/xfs/scrub/bmap.c block = xfs_btree_get_block(bs->cur, i, &bp); cur 432 fs/xfs/scrub/bmap.c struct xfs_btree_cur *cur; cur 444 fs/xfs/scrub/bmap.c cur = xfs_bmbt_init_cursor(mp, sc->tp, ip, whichfork); cur 446 fs/xfs/scrub/bmap.c error = xchk_btree(sc, cur, xchk_bmapbt_rec, &oinfo, info); cur 447 fs/xfs/scrub/bmap.c xfs_btree_del_cursor(cur, error); cur 461 fs/xfs/scrub/bmap.c struct xfs_btree_cur *cur, cur 504 fs/xfs/scrub/bmap.c cur->bc_private.a.agno, rec->rm_startblock)) cur 537 fs/xfs/scrub/bmap.c struct xfs_btree_cur *cur; cur 545 fs/xfs/scrub/bmap.c cur = xfs_rmapbt_init_cursor(sc->mp, sc->tp, agf, agno); cur 546 fs/xfs/scrub/bmap.c if (!cur) { cur 553 fs/xfs/scrub/bmap.c error = xfs_rmap_query_all(cur, xchk_bmap_check_rmap, &sbcri); cur 557 fs/xfs/scrub/bmap.c xfs_btree_del_cursor(cur, error); cur 27 fs/xfs/scrub/btree.c struct xfs_btree_cur *cur, cur 48 fs/xfs/scrub/btree.c if (cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) cur 49 fs/xfs/scrub/btree.c trace_xchk_ifork_btree_op_error(sc, cur, level, cur 52 fs/xfs/scrub/btree.c trace_xchk_btree_op_error(sc, cur, level, cur 62 fs/xfs/scrub/btree.c struct xfs_btree_cur *cur, cur 66 fs/xfs/scrub/btree.c return __xchk_btree_process_error(sc, cur, level, error, cur 73 fs/xfs/scrub/btree.c struct xfs_btree_cur *cur, cur 77 fs/xfs/scrub/btree.c return __xchk_btree_process_error(sc, cur, level, error, cur 85 fs/xfs/scrub/btree.c struct xfs_btree_cur *cur, cur 92 fs/xfs/scrub/btree.c if (cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) cur 93 fs/xfs/scrub/btree.c trace_xchk_ifork_btree_error(sc, cur, level, cur 96 fs/xfs/scrub/btree.c trace_xchk_btree_error(sc, cur, level, cur 103 fs/xfs/scrub/btree.c struct xfs_btree_cur *cur, cur 106 fs/xfs/scrub/btree.c __xchk_btree_set_corrupt(sc, cur, level, XFS_SCRUB_OFLAG_CORRUPT, cur 113 fs/xfs/scrub/btree.c struct xfs_btree_cur *cur, cur 116 fs/xfs/scrub/btree.c __xchk_btree_set_corrupt(sc, cur, level, XFS_SCRUB_OFLAG_XCORRUPT, cur 128 fs/xfs/scrub/btree.c struct xfs_btree_cur *cur = bs->cur; cur 137 fs/xfs/scrub/btree.c block = xfs_btree_get_block(cur, 0, &bp); cur 138 fs/xfs/scrub/btree.c rec = xfs_btree_rec_addr(cur, cur->bc_ptrs[0], block); cur 140 fs/xfs/scrub/btree.c trace_xchk_btree_rec(bs->sc, cur, 0); cur 143 fs/xfs/scrub/btree.c if (!bs->firstrec && !cur->bc_ops->recs_inorder(cur, &bs->lastrec, rec)) cur 144 fs/xfs/scrub/btree.c xchk_btree_set_corrupt(bs->sc, cur, 0); cur 146 fs/xfs/scrub/btree.c memcpy(&bs->lastrec, rec, cur->bc_ops->rec_len); cur 148 fs/xfs/scrub/btree.c if (cur->bc_nlevels == 1) cur 152 fs/xfs/scrub/btree.c cur->bc_ops->init_key_from_rec(&key, rec); cur 153 fs/xfs/scrub/btree.c keyblock = xfs_btree_get_block(cur, 1, &bp); cur 154 fs/xfs/scrub/btree.c keyp = xfs_btree_key_addr(cur, cur->bc_ptrs[1], keyblock); cur 155 fs/xfs/scrub/btree.c if (cur->bc_ops->diff_two_keys(cur, &key, keyp) < 0) cur 156 fs/xfs/scrub/btree.c xchk_btree_set_corrupt(bs->sc, cur, 1); cur 158 fs/xfs/scrub/btree.c if (!(cur->bc_flags & XFS_BTREE_OVERLAPPING)) cur 162 fs/xfs/scrub/btree.c cur->bc_ops->init_high_key_from_rec(&hkey, rec); cur 163 fs/xfs/scrub/btree.c keyp = xfs_btree_high_key_addr(cur, cur->bc_ptrs[1], keyblock); cur 164 fs/xfs/scrub/btree.c if (cur->bc_ops->diff_two_keys(cur, keyp, &hkey) < 0) cur 165 fs/xfs/scrub/btree.c xchk_btree_set_corrupt(bs->sc, cur, 1); cur 177 fs/xfs/scrub/btree.c struct xfs_btree_cur *cur = bs->cur; cur 184 fs/xfs/scrub/btree.c block = xfs_btree_get_block(cur, level, &bp); cur 185 fs/xfs/scrub/btree.c key = xfs_btree_key_addr(cur, cur->bc_ptrs[level], block); cur 187 fs/xfs/scrub/btree.c trace_xchk_btree_key(bs->sc, cur, level); cur 191 fs/xfs/scrub/btree.c !cur->bc_ops->keys_inorder(cur, &bs->lastkey[level], key)) cur 192 fs/xfs/scrub/btree.c xchk_btree_set_corrupt(bs->sc, cur, level); cur 194 fs/xfs/scrub/btree.c memcpy(&bs->lastkey[level], key, cur->bc_ops->key_len); cur 196 fs/xfs/scrub/btree.c if (level + 1 >= cur->bc_nlevels) cur 200 fs/xfs/scrub/btree.c keyblock = xfs_btree_get_block(cur, level + 1, &bp); cur 201 fs/xfs/scrub/btree.c keyp = xfs_btree_key_addr(cur, cur->bc_ptrs[level + 1], keyblock); cur 202 fs/xfs/scrub/btree.c if (cur->bc_ops->diff_two_keys(cur, key, keyp) < 0) cur 203 fs/xfs/scrub/btree.c xchk_btree_set_corrupt(bs->sc, cur, level); cur 205 fs/xfs/scrub/btree.c if (!(cur->bc_flags & XFS_BTREE_OVERLAPPING)) cur 209 fs/xfs/scrub/btree.c key = xfs_btree_high_key_addr(cur, cur->bc_ptrs[level], block); cur 210 fs/xfs/scrub/btree.c keyp = xfs_btree_high_key_addr(cur, cur->bc_ptrs[level + 1], keyblock); cur 211 fs/xfs/scrub/btree.c if (cur->bc_ops->diff_two_keys(cur, keyp, key) < 0) cur 212 fs/xfs/scrub/btree.c xchk_btree_set_corrupt(bs->sc, cur, level); cur 228 fs/xfs/scrub/btree.c if ((bs->cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) && cur 229 fs/xfs/scrub/btree.c level == bs->cur->bc_nlevels) cur 233 fs/xfs/scrub/btree.c if (bs->cur->bc_flags & XFS_BTREE_LONG_PTRS) cur 234 fs/xfs/scrub/btree.c res = xfs_btree_check_lptr(bs->cur, be64_to_cpu(ptr->l), level); cur 236 fs/xfs/scrub/btree.c res = xfs_btree_check_sptr(bs->cur, be32_to_cpu(ptr->s), level); cur 238 fs/xfs/scrub/btree.c xchk_btree_set_corrupt(bs->sc, bs->cur, level); cur 251 fs/xfs/scrub/btree.c struct xfs_btree_cur *cur = bs->cur; cur 259 fs/xfs/scrub/btree.c error = xfs_btree_dup_cursor(cur, &ncur); cur 260 fs/xfs/scrub/btree.c if (!xchk_btree_process_error(bs->sc, cur, level + 1, &error) || cur 268 fs/xfs/scrub/btree.c if (xfs_btree_ptr_is_null(cur, sibling)) { cur 274 fs/xfs/scrub/btree.c xchk_btree_set_corrupt(bs->sc, cur, level); cur 284 fs/xfs/scrub/btree.c if (!xchk_btree_process_error(bs->sc, cur, level + 1, &error)) cur 287 fs/xfs/scrub/btree.c xchk_btree_set_corrupt(bs->sc, cur, level + 1); cur 299 fs/xfs/scrub/btree.c if (xfs_btree_diff_two_ptrs(cur, pp, sibling)) cur 300 fs/xfs/scrub/btree.c xchk_btree_set_corrupt(bs->sc, cur, level); cur 312 fs/xfs/scrub/btree.c struct xfs_btree_cur *cur = bs->cur; cur 318 fs/xfs/scrub/btree.c xfs_btree_get_sibling(cur, block, &leftsib, XFS_BB_LEFTSIB); cur 319 fs/xfs/scrub/btree.c xfs_btree_get_sibling(cur, block, &rightsib, XFS_BB_RIGHTSIB); cur 323 fs/xfs/scrub/btree.c if (level == cur->bc_nlevels - 1) { cur 324 fs/xfs/scrub/btree.c if (!xfs_btree_ptr_is_null(cur, &leftsib) || cur 325 fs/xfs/scrub/btree.c !xfs_btree_ptr_is_null(cur, &rightsib)) cur 326 fs/xfs/scrub/btree.c xchk_btree_set_corrupt(bs->sc, cur, level); cur 367 fs/xfs/scrub/btree.c if (!bs->cur) cur 370 fs/xfs/scrub/btree.c btnum = bs->cur->bc_btnum; cur 371 fs/xfs/scrub/btree.c agno = xfs_daddr_to_agno(bs->cur->bc_mp, daddr); cur 372 fs/xfs/scrub/btree.c agbno = xfs_daddr_to_agbno(bs->cur->bc_mp, daddr); cur 374 fs/xfs/scrub/btree.c init_sa = bs->cur->bc_flags & XFS_BTREE_LONG_PTRS; cur 377 fs/xfs/scrub/btree.c if (!xchk_btree_xref_process_error(bs->sc, bs->cur, cur 389 fs/xfs/scrub/btree.c bs->cur = NULL; cur 393 fs/xfs/scrub/btree.c bs->cur = NULL; cur 408 fs/xfs/scrub/btree.c struct xfs_btree_cur *cur = bs->cur; cur 418 fs/xfs/scrub/btree.c if (!(cur->bc_flags & XFS_BTREE_ROOT_IN_INODE)) cur 419 fs/xfs/scrub/btree.c xchk_btree_set_corrupt(bs->sc, bs->cur, level); cur 431 fs/xfs/scrub/btree.c if (cur->bc_btnum == XFS_BTNUM_BNO || cur->bc_btnum == XFS_BTNUM_RMAP) { cur 461 fs/xfs/scrub/btree.c if (numrecs >= bs->cur->bc_ops->get_minrecs(bs->cur, level)) cur 474 fs/xfs/scrub/btree.c ok_level = bs->cur->bc_nlevels - 1; cur 475 fs/xfs/scrub/btree.c if (bs->cur->bc_flags & XFS_BTREE_ROOT_IN_INODE) cur 480 fs/xfs/scrub/btree.c xchk_btree_set_corrupt(bs->sc, bs->cur, level); cur 501 fs/xfs/scrub/btree.c error = xfs_btree_lookup_get_block(bs->cur, level, pp, pblock); cur 502 fs/xfs/scrub/btree.c if (!xchk_btree_process_error(bs->sc, bs->cur, level, &error) || cur 506 fs/xfs/scrub/btree.c xfs_btree_get_block(bs->cur, level, pbp); cur 507 fs/xfs/scrub/btree.c if (bs->cur->bc_flags & XFS_BTREE_LONG_PTRS) cur 508 fs/xfs/scrub/btree.c failed_at = __xfs_btree_check_lblock(bs->cur, *pblock, cur 511 fs/xfs/scrub/btree.c failed_at = __xfs_btree_check_sblock(bs->cur, *pblock, cur 514 fs/xfs/scrub/btree.c xchk_btree_set_corrupt(bs->sc, bs->cur, level); cur 548 fs/xfs/scrub/btree.c struct xfs_btree_cur *cur = bs->cur; cur 555 fs/xfs/scrub/btree.c if (level >= cur->bc_nlevels - 1) cur 559 fs/xfs/scrub/btree.c xfs_btree_get_keys(cur, block, &block_keys); cur 562 fs/xfs/scrub/btree.c parent_block = xfs_btree_get_block(cur, level + 1, &bp); cur 563 fs/xfs/scrub/btree.c parent_keys = xfs_btree_key_addr(cur, cur->bc_ptrs[level + 1], cur 566 fs/xfs/scrub/btree.c if (cur->bc_ops->diff_two_keys(cur, &block_keys, parent_keys) != 0) cur 567 fs/xfs/scrub/btree.c xchk_btree_set_corrupt(bs->sc, cur, 1); cur 569 fs/xfs/scrub/btree.c if (!(cur->bc_flags & XFS_BTREE_OVERLAPPING)) cur 573 fs/xfs/scrub/btree.c high_bk = xfs_btree_high_key_from_key(cur, &block_keys); cur 574 fs/xfs/scrub/btree.c high_pk = xfs_btree_high_key_addr(cur, cur->bc_ptrs[level + 1], cur 577 fs/xfs/scrub/btree.c if (cur->bc_ops->diff_two_keys(cur, high_bk, high_pk) != 0) cur 578 fs/xfs/scrub/btree.c xchk_btree_set_corrupt(bs->sc, cur, 1); cur 589 fs/xfs/scrub/btree.c struct xfs_btree_cur *cur, cur 595 fs/xfs/scrub/btree.c .cur = cur, cur 619 fs/xfs/scrub/btree.c if (cur->bc_nlevels > XFS_BTREE_MAXLEVELS) { cur 620 fs/xfs/scrub/btree.c xchk_btree_set_corrupt(sc, cur, 0); cur 628 fs/xfs/scrub/btree.c level = cur->bc_nlevels - 1; cur 629 fs/xfs/scrub/btree.c cur->bc_ops->init_ptr_from_cur(cur, &ptr); cur 630 fs/xfs/scrub/btree.c if (!xchk_btree_ptr_ok(&bs, cur->bc_nlevels, &ptr)) cur 636 fs/xfs/scrub/btree.c cur->bc_ptrs[level] = 1; cur 638 fs/xfs/scrub/btree.c while (level < cur->bc_nlevels) { cur 639 fs/xfs/scrub/btree.c block = xfs_btree_get_block(cur, level, &bp); cur 643 fs/xfs/scrub/btree.c if (cur->bc_ptrs[level] > cur 646 fs/xfs/scrub/btree.c if (level < cur->bc_nlevels - 1) cur 647 fs/xfs/scrub/btree.c cur->bc_ptrs[level + 1]++; cur 656 fs/xfs/scrub/btree.c recp = xfs_btree_rec_addr(cur, cur->bc_ptrs[0], block); cur 664 fs/xfs/scrub/btree.c cur->bc_ptrs[level]++; cur 669 fs/xfs/scrub/btree.c if (cur->bc_ptrs[level] > be16_to_cpu(block->bb_numrecs)) { cur 671 fs/xfs/scrub/btree.c if (level < cur->bc_nlevels - 1) cur 672 fs/xfs/scrub/btree.c cur->bc_ptrs[level + 1]++; cur 681 fs/xfs/scrub/btree.c pp = xfs_btree_ptr_addr(cur, cur->bc_ptrs[level], block); cur 683 fs/xfs/scrub/btree.c cur->bc_ptrs[level]++; cur 691 fs/xfs/scrub/btree.c cur->bc_ptrs[level] = 1; cur 697 fs/xfs/scrub/btree.c if (!error && bs.cur) cur 13 fs/xfs/scrub/btree.h struct xfs_btree_cur *cur, int level, int *error); cur 17 fs/xfs/scrub/btree.h struct xfs_btree_cur *cur, int level, int *error); cur 21 fs/xfs/scrub/btree.h struct xfs_btree_cur *cur, int level); cur 25 fs/xfs/scrub/btree.h struct xfs_btree_cur *cur, int level); cur 35 fs/xfs/scrub/btree.h struct xfs_btree_cur *cur; cur 47 fs/xfs/scrub/btree.h int xchk_btree(struct xfs_scrub *sc, struct xfs_btree_cur *cur, cur 324 fs/xfs/scrub/common.c struct xfs_btree_cur *cur, cur 351 fs/xfs/scrub/common.c struct xfs_btree_cur *cur, cur 361 fs/xfs/scrub/common.c return xfs_rmap_query_all(cur, xchk_count_rmap_ownedby_irec, cur 127 fs/xfs/scrub/common.h int xchk_count_rmap_ownedby_ag(struct xfs_scrub *sc, struct xfs_btree_cur *cur, cur 106 fs/xfs/scrub/ialloc.c struct xfs_mount *mp = bs->cur->bc_mp; cur 107 fs/xfs/scrub/ialloc.c xfs_agnumber_t agno = bs->cur->bc_private.a.agno; cur 114 fs/xfs/scrub/ialloc.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 151 fs/xfs/scrub/ialloc.c struct xfs_mount *mp = bs->cur->bc_mp; cur 167 fs/xfs/scrub/ialloc.c fsino = XFS_AGINO_TO_INO(mp, bs->cur->bc_private.a.agno, agino); cur 172 fs/xfs/scrub/ialloc.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 176 fs/xfs/scrub/ialloc.c error = xfs_icache_inode_is_allocated(mp, bs->cur->bc_tp, fsino, cur 195 fs/xfs/scrub/ialloc.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 214 fs/xfs/scrub/ialloc.c struct xfs_mount *mp = bs->cur->bc_mp; cur 218 fs/xfs/scrub/ialloc.c xfs_agnumber_t agno = bs->cur->bc_private.a.agno; cur 253 fs/xfs/scrub/ialloc.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 265 fs/xfs/scrub/ialloc.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 281 fs/xfs/scrub/ialloc.c error = xfs_imap_to_bp(mp, bs->cur->bc_tp, &imap, &dip, &cluster_bp, cur 283 fs/xfs/scrub/ialloc.c if (!xchk_btree_xref_process_error(bs->sc, bs->cur, 0, &error)) cur 291 fs/xfs/scrub/ialloc.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 303 fs/xfs/scrub/ialloc.c xfs_trans_brelse(bs->cur->bc_tp, cluster_bp); cur 365 fs/xfs/scrub/ialloc.c if (bs->cur->bc_btnum == XFS_BTNUM_FINO) { cur 371 fs/xfs/scrub/ialloc.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 382 fs/xfs/scrub/ialloc.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 398 fs/xfs/scrub/ialloc.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 403 fs/xfs/scrub/ialloc.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 425 fs/xfs/scrub/ialloc.c struct xfs_mount *mp = bs->cur->bc_mp; cur 429 fs/xfs/scrub/ialloc.c xfs_agnumber_t agno = bs->cur->bc_private.a.agno; cur 442 fs/xfs/scrub/ialloc.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 447 fs/xfs/scrub/ialloc.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 453 fs/xfs/scrub/ialloc.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 468 fs/xfs/scrub/ialloc.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 483 fs/xfs/scrub/ialloc.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 496 fs/xfs/scrub/ialloc.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 578 fs/xfs/scrub/ialloc.c struct xfs_btree_cur *cur; cur 586 fs/xfs/scrub/ialloc.c cur = which == XFS_BTNUM_INO ? sc->sa.ino_cur : sc->sa.fino_cur; cur 587 fs/xfs/scrub/ialloc.c error = xchk_btree(sc, cur, xchk_iallocbt_rec, &XFS_RMAP_OINFO_INOBT, cur 93 fs/xfs/scrub/refcount.c struct xfs_btree_cur *cur, cur 111 fs/xfs/scrub/refcount.c xchk_btree_xref_set_corrupt(refchk->sc, cur, 0); cur 337 fs/xfs/scrub/refcount.c struct xfs_mount *mp = bs->cur->bc_mp; cur 339 fs/xfs/scrub/refcount.c xfs_agnumber_t agno = bs->cur->bc_private.a.agno; cur 352 fs/xfs/scrub/refcount.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 361 fs/xfs/scrub/refcount.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 364 fs/xfs/scrub/refcount.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 526 fs/xfs/scrub/repair.c struct xfs_btree_cur *cur; cur 550 fs/xfs/scrub/repair.c cur = xfs_rmapbt_init_cursor(sc->mp, sc->tp, agf_bp, agno); cur 553 fs/xfs/scrub/repair.c error = xfs_rmap_has_other_keys(cur, agbno, 1, oinfo, &has_other_rmap); cur 554 fs/xfs/scrub/repair.c xfs_btree_del_cursor(cur, error); cur 828 fs/xfs/scrub/repair.c struct xfs_btree_cur *cur, cur 872 fs/xfs/scrub/repair.c struct xfs_btree_cur *cur; cur 889 fs/xfs/scrub/repair.c cur = xfs_rmapbt_init_cursor(mp, sc->tp, agf_bp, sc->sa.agno); cur 890 fs/xfs/scrub/repair.c error = xfs_rmap_query_all(cur, xrep_findroot_rmap, &ri); cur 891 fs/xfs/scrub/repair.c xfs_btree_del_cursor(cur, error); cur 93 fs/xfs/scrub/rmap.c struct xfs_mount *mp = bs->cur->bc_mp; cur 95 fs/xfs/scrub/rmap.c xfs_agnumber_t agno = bs->cur->bc_private.a.agno; cur 103 fs/xfs/scrub/rmap.c if (!xchk_btree_process_error(bs->sc, bs->cur, 0, &error)) cur 108 fs/xfs/scrub/rmap.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 118 fs/xfs/scrub/rmap.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 127 fs/xfs/scrub/rmap.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 137 fs/xfs/scrub/rmap.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 140 fs/xfs/scrub/rmap.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 143 fs/xfs/scrub/rmap.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 146 fs/xfs/scrub/rmap.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 150 fs/xfs/scrub/rmap.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 155 fs/xfs/scrub/rmap.c xchk_btree_set_corrupt(bs->sc, bs->cur, 0); cur 20 fs/xfs/scrub/trace.c struct xfs_btree_cur *cur, cur 23 fs/xfs/scrub/trace.c if (level < cur->bc_nlevels && cur->bc_bufs[level]) cur 24 fs/xfs/scrub/trace.c return XFS_DADDR_TO_FSB(cur->bc_mp, cur->bc_bufs[level]->b_bn); cur 25 fs/xfs/scrub/trace.c else if (level == cur->bc_nlevels - 1 && cur 26 fs/xfs/scrub/trace.c cur->bc_flags & XFS_BTREE_LONG_PTRS) cur 27 fs/xfs/scrub/trace.c return XFS_INO_TO_FSB(cur->bc_mp, cur->bc_private.b.ip->i_ino); cur 28 fs/xfs/scrub/trace.c else if (!(cur->bc_flags & XFS_BTREE_LONG_PTRS)) cur 29 fs/xfs/scrub/trace.c return XFS_AGB_TO_FSB(cur->bc_mp, cur->bc_private.a.agno, 0); cur 322 fs/xfs/scrub/trace.h TP_PROTO(struct xfs_scrub *sc, struct xfs_btree_cur *cur, cur 324 fs/xfs/scrub/trace.h TP_ARGS(sc, cur, level, error, ret_ip), cur 337 fs/xfs/scrub/trace.h xfs_fsblock_t fsbno = xchk_btree_cur_fsbno(cur, level); cur 341 fs/xfs/scrub/trace.h __entry->btnum = cur->bc_btnum; cur 343 fs/xfs/scrub/trace.h __entry->agno = XFS_FSB_TO_AGNO(cur->bc_mp, fsbno); cur 344 fs/xfs/scrub/trace.h __entry->bno = XFS_FSB_TO_AGBNO(cur->bc_mp, fsbno); cur 345 fs/xfs/scrub/trace.h __entry->ptr = cur->bc_ptrs[level]; cur 362 fs/xfs/scrub/trace.h TP_PROTO(struct xfs_scrub *sc, struct xfs_btree_cur *cur, cur 364 fs/xfs/scrub/trace.h TP_ARGS(sc, cur, level, error, ret_ip), cur 379 fs/xfs/scrub/trace.h xfs_fsblock_t fsbno = xchk_btree_cur_fsbno(cur, level); cur 382 fs/xfs/scrub/trace.h __entry->whichfork = cur->bc_private.b.whichfork; cur 384 fs/xfs/scrub/trace.h __entry->btnum = cur->bc_btnum; cur 386 fs/xfs/scrub/trace.h __entry->ptr = cur->bc_ptrs[level]; cur 387 fs/xfs/scrub/trace.h __entry->agno = XFS_FSB_TO_AGNO(cur->bc_mp, fsbno); cur 388 fs/xfs/scrub/trace.h __entry->bno = XFS_FSB_TO_AGBNO(cur->bc_mp, fsbno); cur 407 fs/xfs/scrub/trace.h TP_PROTO(struct xfs_scrub *sc, struct xfs_btree_cur *cur, cur 409 fs/xfs/scrub/trace.h TP_ARGS(sc, cur, level, ret_ip), cur 421 fs/xfs/scrub/trace.h xfs_fsblock_t fsbno = xchk_btree_cur_fsbno(cur, level); cur 424 fs/xfs/scrub/trace.h __entry->btnum = cur->bc_btnum; cur 426 fs/xfs/scrub/trace.h __entry->agno = XFS_FSB_TO_AGNO(cur->bc_mp, fsbno); cur 427 fs/xfs/scrub/trace.h __entry->bno = XFS_FSB_TO_AGBNO(cur->bc_mp, fsbno); cur 428 fs/xfs/scrub/trace.h __entry->ptr = cur->bc_ptrs[level]; cur 443 fs/xfs/scrub/trace.h TP_PROTO(struct xfs_scrub *sc, struct xfs_btree_cur *cur, cur 445 fs/xfs/scrub/trace.h TP_ARGS(sc, cur, level, ret_ip), cur 459 fs/xfs/scrub/trace.h xfs_fsblock_t fsbno = xchk_btree_cur_fsbno(cur, level); cur 462 fs/xfs/scrub/trace.h __entry->whichfork = cur->bc_private.b.whichfork; cur 464 fs/xfs/scrub/trace.h __entry->btnum = cur->bc_btnum; cur 466 fs/xfs/scrub/trace.h __entry->agno = XFS_FSB_TO_AGNO(cur->bc_mp, fsbno); cur 467 fs/xfs/scrub/trace.h __entry->bno = XFS_FSB_TO_AGBNO(cur->bc_mp, fsbno); cur 468 fs/xfs/scrub/trace.h __entry->ptr = cur->bc_ptrs[level]; cur 485 fs/xfs/scrub/trace.h TP_PROTO(struct xfs_scrub *sc, struct xfs_btree_cur *cur, cur 487 fs/xfs/scrub/trace.h TP_ARGS(sc, cur, level), cur 499 fs/xfs/scrub/trace.h xfs_fsblock_t fsbno = xchk_btree_cur_fsbno(cur, level); cur 503 fs/xfs/scrub/trace.h __entry->btnum = cur->bc_btnum; cur 504 fs/xfs/scrub/trace.h __entry->agno = XFS_FSB_TO_AGNO(cur->bc_mp, fsbno); cur 505 fs/xfs/scrub/trace.h __entry->bno = XFS_FSB_TO_AGBNO(cur->bc_mp, fsbno); cur 507 fs/xfs/scrub/trace.h __entry->nlevels = cur->bc_nlevels; cur 508 fs/xfs/scrub/trace.h __entry->ptr = cur->bc_ptrs[level]; cur 522 fs/xfs/scrub/trace.h TP_PROTO(struct xfs_scrub *sc, struct xfs_btree_cur *cur, \ cur 524 fs/xfs/scrub/trace.h TP_ARGS(sc, cur, level)) cur 31 fs/xfs/xfs_discard.c struct xfs_btree_cur *cur; cur 50 fs/xfs/xfs_discard.c cur = xfs_allocbt_init_cursor(mp, NULL, agbp, agno, XFS_BTNUM_CNT); cur 55 fs/xfs/xfs_discard.c error = xfs_alloc_lookup_ge(cur, 0, cur 70 fs/xfs/xfs_discard.c error = xfs_alloc_get_rec(cur, &fbno, &flen, &i); cur 118 fs/xfs/xfs_discard.c error = xfs_btree_decrement(cur, 0, &i); cur 129 fs/xfs/xfs_discard.c xfs_btree_del_cursor(cur, error); cur 613 fs/xfs/xfs_dquot.c struct xfs_iext_cursor cur; cur 637 fs/xfs/xfs_dquot.c if (xfs_iext_lookup_extent(quotip, "ip->i_df, start, &cur, &got)) { cur 198 fs/xfs/xfs_fsmap.c struct xfs_btree_cur *cur; cur 212 fs/xfs/xfs_fsmap.c cur = xfs_refcountbt_init_cursor(mp, tp, info->agf_bp, cur 215 fs/xfs/xfs_fsmap.c error = xfs_refcount_find_shared(cur, rec->rm_startblock, cur 218 fs/xfs/xfs_fsmap.c xfs_btree_del_cursor(cur, error); cur 337 fs/xfs/xfs_fsmap.c struct xfs_btree_cur *cur, cur 341 fs/xfs/xfs_fsmap.c struct xfs_mount *mp = cur->bc_mp; cur 346 fs/xfs/xfs_fsmap.c fsb = XFS_AGB_TO_FSB(mp, cur->bc_private.a.agno, rec->rm_startblock); cur 349 fs/xfs/xfs_fsmap.c return xfs_getfsmap_helper(cur->bc_tp, info, rec, rec_daddr); cur 355 fs/xfs/xfs_fsmap.c struct xfs_btree_cur *cur, cur 359 fs/xfs/xfs_fsmap.c struct xfs_mount *mp = cur->bc_mp; cur 364 fs/xfs/xfs_fsmap.c rec_daddr = XFS_AGB_TO_DADDR(mp, cur->bc_private.a.agno, cur 373 fs/xfs/xfs_fsmap.c return xfs_getfsmap_helper(cur->bc_tp, info, &irec, rec_daddr); cur 385 fs/xfs/xfs_iwalk.c struct xfs_btree_cur *cur = NULL; cur 394 fs/xfs/xfs_iwalk.c error = xfs_iwalk_ag_start(iwag, agno, agino, &cur, &agi_bp, &has_more); cur 405 fs/xfs/xfs_iwalk.c error = xfs_inobt_get_rec(cur, irec, &has_more); cur 411 fs/xfs/xfs_iwalk.c error = xfs_btree_increment(cur, 0, &has_more); cur 429 fs/xfs/xfs_iwalk.c error = xfs_btree_increment(cur, 0, &has_more); cur 442 fs/xfs/xfs_iwalk.c error = xfs_iwalk_run_callbacks(iwag, agno, &cur, &agi_bp, cur 450 fs/xfs/xfs_iwalk.c error = xfs_iwalk_run_callbacks(iwag, agno, &cur, &agi_bp, &has_more); cur 453 fs/xfs/xfs_iwalk.c xfs_iwalk_del_inobt(tp, &cur, &agi_bp, error); cur 3388 fs/xfs/xfs_log_recover.c struct xfs_ail_cursor cur; cur 3403 fs/xfs/xfs_log_recover.c lip = xfs_trans_ail_cursor_first(ailp, &cur, 0); cur 3418 fs/xfs/xfs_log_recover.c lip = xfs_trans_ail_cursor_next(ailp, &cur); cur 3421 fs/xfs/xfs_log_recover.c xfs_trans_ail_cursor_done(&cur); cur 3484 fs/xfs/xfs_log_recover.c struct xfs_ail_cursor cur; cur 3496 fs/xfs/xfs_log_recover.c lip = xfs_trans_ail_cursor_first(ailp, &cur, 0); cur 3511 fs/xfs/xfs_log_recover.c lip = xfs_trans_ail_cursor_next(ailp, &cur); cur 3514 fs/xfs/xfs_log_recover.c xfs_trans_ail_cursor_done(&cur); cur 3600 fs/xfs/xfs_log_recover.c struct xfs_ail_cursor cur; cur 3613 fs/xfs/xfs_log_recover.c lip = xfs_trans_ail_cursor_first(ailp, &cur, 0); cur 3628 fs/xfs/xfs_log_recover.c lip = xfs_trans_ail_cursor_next(ailp, &cur); cur 3631 fs/xfs/xfs_log_recover.c xfs_trans_ail_cursor_done(&cur); cur 3719 fs/xfs/xfs_log_recover.c struct xfs_ail_cursor cur; cur 3732 fs/xfs/xfs_log_recover.c lip = xfs_trans_ail_cursor_first(ailp, &cur, 0); cur 3747 fs/xfs/xfs_log_recover.c lip = xfs_trans_ail_cursor_next(ailp, &cur); cur 3750 fs/xfs/xfs_log_recover.c xfs_trans_ail_cursor_done(&cur); cur 4795 fs/xfs/xfs_log_recover.c struct xfs_ail_cursor cur; cur 4818 fs/xfs/xfs_log_recover.c lip = xfs_trans_ail_cursor_first(ailp, &cur, 0); cur 4829 fs/xfs/xfs_log_recover.c for (; lip; lip = xfs_trans_ail_cursor_next(ailp, &cur)) cur 4864 fs/xfs/xfs_log_recover.c lip = xfs_trans_ail_cursor_next(ailp, &cur); cur 4867 fs/xfs/xfs_log_recover.c xfs_trans_ail_cursor_done(&cur); cur 4885 fs/xfs/xfs_log_recover.c struct xfs_ail_cursor cur; cur 4890 fs/xfs/xfs_log_recover.c lip = xfs_trans_ail_cursor_first(ailp, &cur, 0); cur 4898 fs/xfs/xfs_log_recover.c for (; lip; lip = xfs_trans_ail_cursor_next(ailp, &cur)) cur 4919 fs/xfs/xfs_log_recover.c lip = xfs_trans_ail_cursor_next(ailp, &cur); cur 4922 fs/xfs/xfs_log_recover.c xfs_trans_ail_cursor_done(&cur); cur 140 fs/xfs/xfs_reflink.c struct xfs_btree_cur *cur; cur 149 fs/xfs/xfs_reflink.c cur = xfs_refcountbt_init_cursor(mp, tp, agbp, agno); cur 151 fs/xfs/xfs_reflink.c error = xfs_refcount_find_shared(cur, agbno, aglen, fbno, flen, cur 154 fs/xfs/xfs_reflink.c xfs_btree_del_cursor(cur, error); cur 215 fs/xfs/xfs_trace.h TP_PROTO(struct xfs_inode *ip, struct xfs_iext_cursor *cur, int state, cur 217 fs/xfs/xfs_trace.h TP_ARGS(ip, cur, state, caller_ip), cur 235 fs/xfs/xfs_trace.h xfs_iext_get_extent(ifp, cur, &r); cur 238 fs/xfs/xfs_trace.h __entry->leaf = cur->leaf; cur 239 fs/xfs/xfs_trace.h __entry->pos = cur->pos; cur 263 fs/xfs/xfs_trace.h TP_PROTO(struct xfs_inode *ip, struct xfs_iext_cursor *cur, int state, \ cur 265 fs/xfs/xfs_trace.h TP_ARGS(ip, cur, state, caller_ip)) cur 2205 fs/xfs/xfs_trace.h TP_PROTO(struct xfs_btree_cur *cur, int level, struct xfs_buf *bp), cur 2206 fs/xfs/xfs_trace.h TP_ARGS(cur, level, bp), cur 2216 fs/xfs/xfs_trace.h __entry->dev = cur->bc_mp->m_super->s_dev; cur 2217 fs/xfs/xfs_trace.h __entry->btnum = cur->bc_btnum; cur 2219 fs/xfs/xfs_trace.h __entry->nlevels = cur->bc_nlevels; cur 2220 fs/xfs/xfs_trace.h __entry->ptr = cur->bc_ptrs[level]; cur 2234 fs/xfs/xfs_trace.h TP_PROTO(struct xfs_btree_cur *cur, int level, struct xfs_buf *bp), \ cur 2235 fs/xfs/xfs_trace.h TP_ARGS(cur, level, bp)) cur 788 fs/xfs/xfs_trans.c struct xfs_ail_cursor *cur, cur 797 fs/xfs/xfs_trans.c xfs_trans_ail_update_bulk(ailp, cur, log_items, nr_items, commit_lsn); cur 837 fs/xfs/xfs_trans.c struct xfs_ail_cursor cur; cur 841 fs/xfs/xfs_trans.c xfs_trans_ail_cursor_last(ailp, &cur, commit_lsn); cur 899 fs/xfs/xfs_trans.c xfs_log_item_batch_insert(ailp, &cur, log_items, cur 907 fs/xfs/xfs_trans.c xfs_log_item_batch_insert(ailp, &cur, log_items, i, commit_lsn); cur 910 fs/xfs/xfs_trans.c xfs_trans_ail_cursor_done(&cur); cur 156 fs/xfs/xfs_trans_ail.c struct xfs_ail_cursor *cur) cur 158 fs/xfs/xfs_trans_ail.c cur->item = NULL; cur 159 fs/xfs/xfs_trans_ail.c list_add_tail(&cur->list, &ailp->ail_cursors); cur 169 fs/xfs/xfs_trans_ail.c struct xfs_ail_cursor *cur) cur 171 fs/xfs/xfs_trans_ail.c struct xfs_log_item *lip = cur->item; cur 176 fs/xfs/xfs_trans_ail.c cur->item = xfs_ail_next(ailp, lip); cur 186 fs/xfs/xfs_trans_ail.c struct xfs_ail_cursor *cur) cur 188 fs/xfs/xfs_trans_ail.c cur->item = NULL; cur 189 fs/xfs/xfs_trans_ail.c list_del_init(&cur->list); cur 205 fs/xfs/xfs_trans_ail.c struct xfs_ail_cursor *cur; cur 207 fs/xfs/xfs_trans_ail.c list_for_each_entry(cur, &ailp->ail_cursors, list) { cur 208 fs/xfs/xfs_trans_ail.c if (cur->item == lip) cur 209 fs/xfs/xfs_trans_ail.c cur->item = (struct xfs_log_item *) cur 210 fs/xfs/xfs_trans_ail.c ((uintptr_t)cur->item | 1); cur 223 fs/xfs/xfs_trans_ail.c struct xfs_ail_cursor *cur, cur 228 fs/xfs/xfs_trans_ail.c xfs_trans_ail_cursor_init(ailp, cur); cur 243 fs/xfs/xfs_trans_ail.c cur->item = xfs_ail_next(ailp, lip); cur 270 fs/xfs/xfs_trans_ail.c struct xfs_ail_cursor *cur, cur 273 fs/xfs/xfs_trans_ail.c xfs_trans_ail_cursor_init(ailp, cur); cur 274 fs/xfs/xfs_trans_ail.c cur->item = __xfs_trans_ail_cursor_last(ailp, lsn); cur 275 fs/xfs/xfs_trans_ail.c return cur->item; cur 287 fs/xfs/xfs_trans_ail.c struct xfs_ail_cursor *cur, cur 300 fs/xfs/xfs_trans_ail.c lip = cur ? cur->item : NULL; cur 311 fs/xfs/xfs_trans_ail.c if (cur) cur 312 fs/xfs/xfs_trans_ail.c cur->item = list_entry(list->prev, struct xfs_log_item, li_ail); cur 367 fs/xfs/xfs_trans_ail.c struct xfs_ail_cursor cur; cur 397 fs/xfs/xfs_trans_ail.c lip = xfs_trans_ail_cursor_first(ailp, &cur, ailp->ail_last_pushed_lsn); cur 403 fs/xfs/xfs_trans_ail.c xfs_trans_ail_cursor_done(&cur); cur 483 fs/xfs/xfs_trans_ail.c lip = xfs_trans_ail_cursor_next(ailp, &cur); cur 488 fs/xfs/xfs_trans_ail.c xfs_trans_ail_cursor_done(&cur); cur 708 fs/xfs/xfs_trans_ail.c struct xfs_ail_cursor *cur, cur 740 fs/xfs/xfs_trans_ail.c xfs_ail_splice(ailp, cur, &tmp, lsn); cur 70 fs/xfs/xfs_trans_priv.h struct xfs_ail_cursor *cur, cur 120 fs/xfs/xfs_trans_priv.h struct xfs_ail_cursor *cur, cur 123 fs/xfs/xfs_trans_priv.h struct xfs_ail_cursor *cur, cur 126 fs/xfs/xfs_trans_priv.h struct xfs_ail_cursor *cur); cur 127 fs/xfs/xfs_trans_priv.h void xfs_trans_ail_cursor_done(struct xfs_ail_cursor *cur); cur 68 include/crypto/if_alg.h unsigned int cur; /* Last processed SG entry */ cur 786 include/linux/blk-cgroup.h int cur = atomic_cmpxchg(&blkg->use_delay, old, old - 1); cur 787 include/linux/blk-cgroup.h if (cur == old) cur 789 include/linux/blk-cgroup.h old = cur; cur 806 include/linux/blk-cgroup.h int cur = atomic_cmpxchg(&blkg->use_delay, old, 0); cur 807 include/linux/blk-cgroup.h if (cur == old) { cur 811 include/linux/blk-cgroup.h old = cur; cur 403 include/linux/bpf_verifier.h struct bpf_verifier_state *cur = env->cur_state; cur 405 include/linux/bpf_verifier.h return cur->frame[cur->curframe]; cur 97 include/linux/bvec.h const struct bio_vec *cur = bv + iter->bi_idx; cur 99 include/linux/bvec.h cur->bv_len - iter->bi_bvec_done); cur 105 include/linux/bvec.h if (iter->bi_bvec_done == cur->bv_len) { cur 199 include/linux/ceph/libceph.h type *cur = rb_entry(*n, type, nodefld); \ cur 203 include/linux/ceph/libceph.h cmp = cmpexp(keyexp(t->keyfld), keyexp(cur->keyfld)); \ cur 233 include/linux/ceph/libceph.h type *cur = rb_entry(n, type, nodefld); \ cur 236 include/linux/ceph/libceph.h cmp = cmpexp(key, keyexp(cur->keyfld)); \ cur 242 include/linux/ceph/libceph.h return cur; \ cur 66 include/linux/cpufreq.h unsigned int cur; /* in kHz, only needed if cpufreq cur 583 include/linux/cpufreq.h if (policy->max < policy->cur) cur 585 include/linux/cpufreq.h else if (policy->min > policy->cur) cur 236 include/linux/device.h bus_find_next_device(struct bus_type *bus,struct device *cur) cur 238 include/linux/device.h return bus_find_device(bus, cur, NULL, device_match_any); cur 287 include/linux/if_team.h struct team_port *cur; cur 291 include/linux/if_team.h cur = port; cur 292 include/linux/if_team.h list_for_each_entry_continue_rcu(cur, &team->port_list, list) cur 293 include/linux/if_team.h if (team_port_txable(cur)) cur 294 include/linux/if_team.h return cur; cur 295 include/linux/if_team.h list_for_each_entry_rcu(cur, &team->port_list, list) { cur 296 include/linux/if_team.h if (cur == port) cur 298 include/linux/if_team.h if (team_port_txable(cur)) cur 299 include/linux/if_team.h return cur; cur 126 include/linux/iversion.h u64 cur, old; cur 128 include/linux/iversion.h cur = inode_peek_iversion_raw(inode); cur 130 include/linux/iversion.h if (cur > val) cur 132 include/linux/iversion.h old = atomic64_cmpxchg(&inode->i_version, cur, val); cur 133 include/linux/iversion.h if (likely(old == cur)) cur 135 include/linux/iversion.h cur = old; cur 200 include/linux/iversion.h u64 cur, old, new; cur 213 include/linux/iversion.h cur = inode_peek_iversion_raw(inode); cur 216 include/linux/iversion.h if (!force && !(cur & I_VERSION_QUERIED)) cur 220 include/linux/iversion.h new = (cur & ~I_VERSION_QUERIED) + I_VERSION_INCREMENT; cur 222 include/linux/iversion.h old = atomic64_cmpxchg(&inode->i_version, cur, new); cur 223 include/linux/iversion.h if (likely(old == cur)) cur 225 include/linux/iversion.h cur = old; cur 307 include/linux/iversion.h u64 cur, old, new; cur 309 include/linux/iversion.h cur = inode_peek_iversion_raw(inode); cur 312 include/linux/iversion.h if (cur & I_VERSION_QUERIED) { cur 322 include/linux/iversion.h new = cur | I_VERSION_QUERIED; cur 323 include/linux/iversion.h old = atomic64_cmpxchg(&inode->i_version, cur, new); cur 324 include/linux/iversion.h if (likely(old == cur)) cur 326 include/linux/iversion.h cur = old; cur 328 include/linux/iversion.h return cur >> I_VERSION_QUERIED_SHIFT; cur 267 include/linux/lockd/lockd.h typedef int (*nlm_host_match_fn_t)(void *cur, struct nlm_host *ref); cur 25 include/linux/mdio-mux.h int (*switch_fn) (int cur, int desired, void *data), cur 91 include/linux/of.h const __be32 *cur; cur 540 include/linux/of.h const __be32 *of_prop_next_u32(struct property *prop, const __be32 *cur, cur 549 include/linux/of.h const char *of_prop_next_string(struct property *prop, const char *cur); cur 925 include/linux/of.h const __be32 *cur, u32 *pu) cur 931 include/linux/of.h const char *cur) cur 7 include/linux/s3c_adc_battery.h int cur; /* mA */ cur 1571 include/linux/sched.h extern int cpuset_cpumask_can_shrink(const struct cpumask *cur, const struct cpumask *trial); cur 18 include/linux/swap_slots.h int cur; cur 172 include/linux/usb/pd_vdo.h #define VDO_CABLE(hw, fw, cbl, gdr, lat, term, tx1d, tx2d, rx1d, rx2d, cur,\ cur 177 include/linux/usb/pd_vdo.h | ((cur) & 0x3) << 5 | (vps) << 4 | (sopp) << 3 \ cur 255 include/media/v4l2-ctrls.h } cur; cur 170 kernel/bpf/bpf_lru_list.c struct list_head *cur, *last, *next = inactive; cur 181 kernel/bpf/bpf_lru_list.c cur = l->next_inactive_rotation; cur 183 kernel/bpf/bpf_lru_list.c if (cur == inactive) { cur 184 kernel/bpf/bpf_lru_list.c cur = cur->prev; cur 188 kernel/bpf/bpf_lru_list.c node = list_entry(cur, struct bpf_lru_node, list); cur 189 kernel/bpf/bpf_lru_list.c next = cur->prev; cur 192 kernel/bpf/bpf_lru_list.c if (cur == last) cur 194 kernel/bpf/bpf_lru_list.c cur = next; cur 2970 kernel/bpf/btf.c void *cur, *end; cur 2973 kernel/bpf/btf.c cur = btf->nohdr_data + hdr->type_off; cur 2974 kernel/bpf/btf.c end = cur + hdr->type_len; cur 2977 kernel/bpf/btf.c while (cur < end) { cur 2978 kernel/bpf/btf.c struct btf_type *t = cur; cur 2981 kernel/bpf/btf.c meta_size = btf_check_meta(env, t, end - cur); cur 2986 kernel/bpf/btf.c cur += meta_size; cur 427 kernel/bpf/verifier.c struct bpf_verifier_state *cur = env->cur_state; cur 429 kernel/bpf/verifier.c return cur->frame[reg->frameno]; cur 778 kernel/bpf/verifier.c struct bpf_verifier_state *cur = env->cur_state; cur 785 kernel/bpf/verifier.c if (cur) { cur 786 kernel/bpf/verifier.c err = copy_verifier_state(cur, &head->st); cur 806 kernel/bpf/verifier.c struct bpf_verifier_state *cur = env->cur_state; cur 819 kernel/bpf/verifier.c err = copy_verifier_state(&elem->st, cur); cur 1405 kernel/bpf/verifier.c struct bpf_verifier_state *cur) cur 1407 kernel/bpf/verifier.c u32 cnt = cur->jmp_history_cnt; cur 1411 kernel/bpf/verifier.c p = krealloc(cur->jmp_history, cnt * sizeof(*p), GFP_USER); cur 1416 kernel/bpf/verifier.c cur->jmp_history = p; cur 1417 kernel/bpf/verifier.c cur->jmp_history_cnt = cnt; cur 1897 kernel/bpf/verifier.c struct bpf_func_state *cur; /* state of the current function */ cur 1916 kernel/bpf/verifier.c cur = env->cur_state->frame[env->cur_state->curframe]; cur 1918 kernel/bpf/verifier.c reg = &cur->regs[value_regno]; cur 1942 kernel/bpf/verifier.c if (state != cur && reg->type == PTR_TO_STACK) { cur 3144 kernel/bpf/verifier.c struct bpf_verifier_state *cur = env->cur_state; cur 3186 kernel/bpf/verifier.c if (cur->active_spin_lock) { cur 3191 kernel/bpf/verifier.c cur->active_spin_lock = reg->id; cur 3193 kernel/bpf/verifier.c if (!cur->active_spin_lock) { cur 3197 kernel/bpf/verifier.c if (cur->active_spin_lock != reg->id) { cur 3201 kernel/bpf/verifier.c cur->active_spin_lock = 0; cur 6290 kernel/bpf/verifier.c struct bpf_verifier_state *cur = env->cur_state; cur 6291 kernel/bpf/verifier.c struct bpf_func_state *state = cur->frame[cur->curframe]; cur 6754 kernel/bpf/verifier.c struct bpf_reg_state *cur) cur 6756 kernel/bpf/verifier.c return old->umin_value <= cur->umin_value && cur 6757 kernel/bpf/verifier.c old->umax_value >= cur->umax_value && cur 6758 kernel/bpf/verifier.c old->smin_value <= cur->smin_value && cur 6759 kernel/bpf/verifier.c old->smax_value >= cur->smax_value; cur 6766 kernel/bpf/verifier.c u32 cur; cur 6787 kernel/bpf/verifier.c idmap[i].cur = cur_id; cur 6791 kernel/bpf/verifier.c return idmap[i].cur == cur_id; cur 6873 kernel/bpf/verifier.c struct bpf_verifier_state *cur) cur 6883 kernel/bpf/verifier.c sl->state.curframe != cur->curframe) cur 6885 kernel/bpf/verifier.c for (i = 0; i <= cur->curframe; i++) cur 6886 kernel/bpf/verifier.c if (sl->state.frame[i]->callsite != cur->frame[i]->callsite) cur 7012 kernel/bpf/verifier.c struct bpf_func_state *cur, cur 7036 kernel/bpf/verifier.c if (i >= cur->allocated_stack) cur 7044 kernel/bpf/verifier.c cur->stack[spi].slot_type[i % BPF_REG_SIZE] == STACK_ZERO) cur 7047 kernel/bpf/verifier.c cur->stack[spi].slot_type[i % BPF_REG_SIZE]) cur 7059 kernel/bpf/verifier.c &cur->stack[spi].spilled_ptr, cur 7076 kernel/bpf/verifier.c static bool refsafe(struct bpf_func_state *old, struct bpf_func_state *cur) cur 7078 kernel/bpf/verifier.c if (old->acquired_refs != cur->acquired_refs) cur 7080 kernel/bpf/verifier.c return !memcmp(old->refs, cur->refs, cur 7111 kernel/bpf/verifier.c struct bpf_func_state *cur) cur 7123 kernel/bpf/verifier.c if (!regsafe(&old->regs[i], &cur->regs[i], idmap)) cur 7127 kernel/bpf/verifier.c if (!stacksafe(old, cur, idmap)) cur 7130 kernel/bpf/verifier.c if (!refsafe(old, cur)) cur 7140 kernel/bpf/verifier.c struct bpf_verifier_state *cur) cur 7144 kernel/bpf/verifier.c if (old->curframe != cur->curframe) cur 7150 kernel/bpf/verifier.c if (old->speculative && !cur->speculative) cur 7153 kernel/bpf/verifier.c if (old->active_spin_lock != cur->active_spin_lock) cur 7160 kernel/bpf/verifier.c if (old->frame[i]->callsite != cur->frame[i]->callsite) cur 7162 kernel/bpf/verifier.c if (!func_states_equal(old->frame[i], cur->frame[i])) cur 7289 kernel/bpf/verifier.c struct bpf_verifier_state *cur) cur 7292 kernel/bpf/verifier.c int i, fr = cur->curframe; cur 7298 kernel/bpf/verifier.c fcur = cur->frame[fr]; cur 7311 kernel/bpf/verifier.c struct bpf_verifier_state *cur = env->cur_state, *new; cur 7315 kernel/bpf/verifier.c cur->last_insn_idx = env->prev_insn_idx; cur 7337 kernel/bpf/verifier.c clean_live_states(env, insn_idx, cur); cur 7344 kernel/bpf/verifier.c if (states_maybe_looping(&sl->state, cur) && cur 7345 kernel/bpf/verifier.c states_equal(env, &sl->state, cur)) { cur 7367 kernel/bpf/verifier.c if (states_equal(env, &sl->state, cur)) { cur 7379 kernel/bpf/verifier.c err = propagate_liveness(env, &sl->state, cur); cur 7386 kernel/bpf/verifier.c err = err ? : push_jmp_history(env, cur); cur 7440 kernel/bpf/verifier.c return push_jmp_history(env, cur); cur 7443 kernel/bpf/verifier.c return push_jmp_history(env, cur); cur 7464 kernel/bpf/verifier.c err = copy_verifier_state(new, cur); cur 7474 kernel/bpf/verifier.c cur->parent = new; cur 7475 kernel/bpf/verifier.c cur->first_insn_idx = insn_idx; cur 7476 kernel/bpf/verifier.c clear_jmp_history(cur); cur 7492 kernel/bpf/verifier.c for (j = 0; j <= cur->curframe; j++) { cur 7493 kernel/bpf/verifier.c for (i = j < cur->curframe ? BPF_REG_6 : 0; i < BPF_REG_FP; i++) cur 7494 kernel/bpf/verifier.c cur->frame[j]->regs[i].parent = &new->frame[j]->regs[i]; cur 7496 kernel/bpf/verifier.c cur->frame[j]->regs[i].live = REG_LIVE_NONE; cur 7500 kernel/bpf/verifier.c for (j = 0; j <= cur->curframe; j++) { cur 7501 kernel/bpf/verifier.c struct bpf_func_state *frame = cur->frame[j]; cur 563 kernel/cgroup/cpuset.c static int validate_change(struct cpuset *cur, struct cpuset *trial) cur 573 kernel/cgroup/cpuset.c cpuset_for_each_child(c, css, cur) cur 579 kernel/cgroup/cpuset.c if (cur == &top_cpuset) cur 582 kernel/cgroup/cpuset.c par = parent_cs(cur); cur 596 kernel/cgroup/cpuset.c c != cur && cur 600 kernel/cgroup/cpuset.c c != cur && cur 610 kernel/cgroup/cpuset.c if ((cgroup_is_populated(cur->css.cgroup) || cur->attach_in_progress)) { cur 611 kernel/cgroup/cpuset.c if (!cpumask_empty(cur->cpus_allowed) && cur 614 kernel/cgroup/cpuset.c if (!nodes_empty(cur->mems_allowed) && cur 624 kernel/cgroup/cpuset.c if (is_cpu_exclusive(cur) && cur 625 kernel/cgroup/cpuset.c !cpuset_cpumask_can_shrink(cur->cpus_allowed, cur 2112 kernel/cpu.c ssize_t cur, res = 0; cur 2120 kernel/cpu.c cur = sprintf(buf, "%3d: %s\n", i, sp->name); cur 2121 kernel/cpu.c buf += cur; cur 2122 kernel/cpu.c res += cur; cur 41 kernel/crash_core.c char *cur = cmdline, *tmp; cur 48 kernel/crash_core.c start = memparse(cur, &tmp); cur 49 kernel/crash_core.c if (cur == tmp) { cur 53 kernel/crash_core.c cur = tmp; cur 54 kernel/crash_core.c if (*cur != '-') { cur 58 kernel/crash_core.c cur++; cur 61 kernel/crash_core.c if (*cur != ':') { cur 62 kernel/crash_core.c end = memparse(cur, &tmp); cur 63 kernel/crash_core.c if (cur == tmp) { cur 67 kernel/crash_core.c cur = tmp; cur 74 kernel/crash_core.c if (*cur != ':') { cur 78 kernel/crash_core.c cur++; cur 80 kernel/crash_core.c size = memparse(cur, &tmp); cur 81 kernel/crash_core.c if (cur == tmp) { cur 85 kernel/crash_core.c cur = tmp; cur 96 kernel/crash_core.c } while (*cur++ == ','); cur 99 kernel/crash_core.c while (*cur && *cur != ' ' && *cur != '@') cur 100 kernel/crash_core.c cur++; cur 101 kernel/crash_core.c if (*cur == '@') { cur 102 kernel/crash_core.c cur++; cur 103 kernel/crash_core.c *crash_base = memparse(cur, &tmp); cur 104 kernel/crash_core.c if (cur == tmp) { cur 126 kernel/crash_core.c char *cur = cmdline; cur 128 kernel/crash_core.c *crash_size = memparse(cmdline, &cur); cur 129 kernel/crash_core.c if (cmdline == cur) { cur 134 kernel/crash_core.c if (*cur == '@') cur 135 kernel/crash_core.c *crash_base = memparse(cur+1, &cur); cur 136 kernel/crash_core.c else if (*cur != ' ' && *cur != '\0') { cur 137 kernel/crash_core.c pr_warn("crashkernel: unrecognized char: %c\n", *cur); cur 164 kernel/crash_core.c char *cur = cmdline; cur 166 kernel/crash_core.c *crash_size = memparse(cmdline, &cur); cur 167 kernel/crash_core.c if (cmdline == cur) { cur 173 kernel/crash_core.c if (strncmp(cur, suffix, strlen(suffix))) { cur 174 kernel/crash_core.c pr_warn("crashkernel: unrecognized char: %c\n", *cur); cur 177 kernel/crash_core.c cur += strlen(suffix); cur 178 kernel/crash_core.c if (*cur != ' ' && *cur != '\0') { cur 179 kernel/crash_core.c pr_warn("crashkernel: unrecognized char: %c\n", *cur); cur 1150 kernel/kprobes.c struct kprobe *cur = __this_cpu_read(kprobe_instance); cur 1156 kernel/kprobes.c if (cur && cur->fault_handler) { cur 1157 kernel/kprobes.c if (cur->fault_handler(cur, regs, trapnr)) cur 446 kernel/locking/mutex.c struct mutex_waiter *cur; cur 450 kernel/locking/mutex.c list_for_each_entry(cur, &lock->wait_list, list) { cur 451 kernel/locking/mutex.c if (!cur->ww_ctx) cur 454 kernel/locking/mutex.c if (__ww_mutex_die(lock, cur, ww_ctx) || cur 455 kernel/locking/mutex.c __ww_mutex_wound(lock, cur->ww_ctx, ww_ctx)) cur 810 kernel/locking/mutex.c struct mutex_waiter *cur; cur 829 kernel/locking/mutex.c cur = waiter; cur 830 kernel/locking/mutex.c list_for_each_entry_continue_reverse(cur, &lock->wait_list, list) { cur 831 kernel/locking/mutex.c if (!cur->ww_ctx) cur 856 kernel/locking/mutex.c struct mutex_waiter *cur; cur 875 kernel/locking/mutex.c list_for_each_entry_reverse(cur, &lock->wait_list, list) { cur 876 kernel/locking/mutex.c if (!cur->ww_ctx) cur 879 kernel/locking/mutex.c if (__ww_ctx_stamp_after(ww_ctx, cur->ww_ctx)) { cur 895 kernel/locking/mutex.c pos = &cur->list; cur 898 kernel/locking/mutex.c __ww_mutex_die(lock, cur, ww_ctx); cur 331 kernel/padata.c struct padata_priv *cur; cur 335 kernel/padata.c list_for_each_entry_reverse(cur, &pqueue->reorder.list, list) cur 336 kernel/padata.c if (cur->seq_nr < padata->seq_nr) cur 338 kernel/padata.c list_add(&padata->list, &cur->list); cur 132 kernel/power/power.h unsigned int cur; /* number of the block of PAGE_SIZE bytes the cur 382 kernel/power/snapshot.c struct bm_position cur; /* most recently used bit position */ cur 550 kernel/power/snapshot.c bm->cur.zone = list_entry(bm->zones.next, struct mem_zone_bm_rtree, cur 552 kernel/power/snapshot.c bm->cur.node = list_entry(bm->cur.zone->leaves.next, cur 554 kernel/power/snapshot.c bm->cur.node_pfn = 0; cur 555 kernel/power/snapshot.c bm->cur.node_bit = 0; cur 595 kernel/power/snapshot.c struct mem_extent *ext, *cur, *aux; cur 626 kernel/power/snapshot.c cur = ext; cur 627 kernel/power/snapshot.c list_for_each_entry_safe_continue(cur, aux, list, hook) { cur 628 kernel/power/snapshot.c if (zone_end < cur->start) cur 630 kernel/power/snapshot.c if (zone_end < cur->end) cur 631 kernel/power/snapshot.c ext->end = cur->end; cur 632 kernel/power/snapshot.c list_del(&cur->hook); cur 633 kernel/power/snapshot.c kfree(cur); cur 714 kernel/power/snapshot.c zone = bm->cur.zone; cur 743 kernel/power/snapshot.c node = bm->cur.node; cur 744 kernel/power/snapshot.c if (zone == bm->cur.zone && cur 745 kernel/power/snapshot.c ((pfn - zone->start_pfn) & ~BM_BLOCK_MASK) == bm->cur.node_pfn) cur 762 kernel/power/snapshot.c bm->cur.zone = zone; cur 763 kernel/power/snapshot.c bm->cur.node = node; cur 764 kernel/power/snapshot.c bm->cur.node_pfn = (pfn - zone->start_pfn) & ~BM_BLOCK_MASK; cur 812 kernel/power/snapshot.c bit = max(bm->cur.node_bit - 1, 0); cur 813 kernel/power/snapshot.c clear_bit(bit, bm->cur.node->data); cur 847 kernel/power/snapshot.c if (!list_is_last(&bm->cur.node->list, &bm->cur.zone->leaves)) { cur 848 kernel/power/snapshot.c bm->cur.node = list_entry(bm->cur.node->list.next, cur 850 kernel/power/snapshot.c bm->cur.node_pfn += BM_BITS_PER_BLOCK; cur 851 kernel/power/snapshot.c bm->cur.node_bit = 0; cur 857 kernel/power/snapshot.c if (!list_is_last(&bm->cur.zone->list, &bm->zones)) { cur 858 kernel/power/snapshot.c bm->cur.zone = list_entry(bm->cur.zone->list.next, cur 860 kernel/power/snapshot.c bm->cur.node = list_entry(bm->cur.zone->leaves.next, cur 862 kernel/power/snapshot.c bm->cur.node_pfn = 0; cur 863 kernel/power/snapshot.c bm->cur.node_bit = 0; cur 888 kernel/power/snapshot.c pages = bm->cur.zone->end_pfn - bm->cur.zone->start_pfn; cur 889 kernel/power/snapshot.c bits = min(pages - bm->cur.node_pfn, BM_BITS_PER_BLOCK); cur 890 kernel/power/snapshot.c bit = find_next_bit(bm->cur.node->data, bits, cur 891 kernel/power/snapshot.c bm->cur.node_bit); cur 893 kernel/power/snapshot.c pfn = bm->cur.zone->start_pfn + bm->cur.node_pfn + bit; cur 894 kernel/power/snapshot.c bm->cur.node_bit = bit + 1; cur 2097 kernel/power/snapshot.c if (handle->cur > nr_meta_pages + nr_copy_pages) cur 2106 kernel/power/snapshot.c if (!handle->cur) { cur 2115 kernel/power/snapshot.c } else if (handle->cur <= nr_meta_pages) { cur 2138 kernel/power/snapshot.c handle->cur++; cur 2597 kernel/power/snapshot.c if (handle->cur > 1 && handle->cur > nr_meta_pages + nr_copy_pages) cur 2602 kernel/power/snapshot.c if (!handle->cur) { cur 2611 kernel/power/snapshot.c } else if (handle->cur == 1) { cur 2628 kernel/power/snapshot.c } else if (handle->cur <= nr_meta_pages + 1) { cur 2633 kernel/power/snapshot.c if (handle->cur == nr_meta_pages + 1) { cur 2657 kernel/power/snapshot.c handle->cur++; cur 2677 kernel/power/snapshot.c if (handle->cur > 1 && handle->cur > nr_meta_pages + nr_copy_pages) { cur 2686 kernel/power/snapshot.c handle->cur <= nr_meta_pages + nr_copy_pages); cur 96 kernel/power/swap.c struct swap_map_page *cur; cur 404 kernel/power/swap.c if (handle->cur) cur 405 kernel/power/swap.c free_page((unsigned long)handle->cur); cur 406 kernel/power/swap.c handle->cur = NULL; cur 419 kernel/power/swap.c handle->cur = (struct swap_map_page *)get_zeroed_page(GFP_KERNEL); cur 420 kernel/power/swap.c if (!handle->cur) { cur 446 kernel/power/swap.c if (!handle->cur) cur 452 kernel/power/swap.c handle->cur->entries[handle->k++] = offset; cur 457 kernel/power/swap.c handle->cur->next_swap = offset; cur 458 kernel/power/swap.c error = write_page(handle->cur, handle->cur_swap, hb); cur 461 kernel/power/swap.c clear_page(handle->cur); cur 482 kernel/power/swap.c if (handle->cur && handle->cur_swap) cur 483 kernel/power/swap.c return write_page(handle->cur, handle->cur_swap, NULL); cur 958 kernel/power/swap.c handle->cur = NULL; cur 973 kernel/power/swap.c handle->cur = NULL; cur 1003 kernel/power/swap.c handle->cur = handle->maps->map; cur 1014 kernel/power/swap.c if (!handle->cur) cur 1016 kernel/power/swap.c offset = handle->cur->entries[handle->k]; cur 1031 kernel/power/swap.c handle->cur = handle->maps->map; cur 1299 kernel/power/swap.c if (handle->cur && cur 1300 kernel/power/swap.c handle->cur->entries[handle->k]) { cur 1834 kernel/sched/core.c int migrate_swap(struct task_struct *cur, struct task_struct *p, cur 1841 kernel/sched/core.c .src_task = cur, cur 1863 kernel/sched/core.c trace_sched_swap_numa(cur, arg.src_cpu, p, arg.dst_cpu); cur 6080 kernel/sched/core.c int cpuset_cpumask_can_shrink(const struct cpumask *cur, cur 6085 kernel/sched/core.c if (!cpumask_weight(cur)) cur 6088 kernel/sched/core.c ret = dl_cpuset_cpumask_can_shrink(cur, trial); cur 127 kernel/sched/cpufreq_schedutil.c policy->cur = next_freq; cur 174 kernel/sched/cpufreq_schedutil.c policy->cpuinfo.max_freq : policy->cur; cur 2749 kernel/sched/deadline.c int dl_cpuset_cpumask_can_shrink(const struct cpumask *cur, cur 2757 kernel/sched/deadline.c cur_dl_b = dl_bw_of(cpumask_any(cur)); cur 1595 kernel/sched/fair.c struct task_struct *cur; cur 1605 kernel/sched/fair.c cur = rcu_dereference(dst_rq->curr); cur 1606 kernel/sched/fair.c if (cur && ((cur->flags & PF_EXITING) || is_idle_task(cur))) cur 1607 kernel/sched/fair.c cur = NULL; cur 1613 kernel/sched/fair.c if (cur == env->p) cur 1616 kernel/sched/fair.c if (!cur) { cur 1631 kernel/sched/fair.c if (!cpumask_test_cpu(env->src_cpu, cur->cpus_ptr)) cur 1638 kernel/sched/fair.c cur_ng = rcu_dereference(cur->numa_group); cur 1640 kernel/sched/fair.c imp = taskimp + task_weight(cur, env->src_nid, dist) - cur 1641 kernel/sched/fair.c task_weight(cur, env->dst_nid, dist); cur 1654 kernel/sched/fair.c imp += group_weight(cur, env->src_nid, dist) - cur 1655 kernel/sched/fair.c group_weight(cur, env->dst_nid, dist); cur 1657 kernel/sched/fair.c imp += task_weight(cur, env->src_nid, dist) - cur 1658 kernel/sched/fair.c task_weight(cur, env->dst_nid, dist); cur 1663 kernel/sched/fair.c cur = NULL; cur 1679 kernel/sched/fair.c load = task_h_load(env->p) - task_h_load(cur); cur 1694 kernel/sched/fair.c if (!cur) { cur 1705 kernel/sched/fair.c task_numa_assign(env, cur, imp); cur 324 kernel/sched/sched.h extern int dl_cpuset_cpumask_can_shrink(const struct cpumask *cur, const struct cpumask *trial); cur 2179 kernel/sched/topology.c static int dattrs_equal(struct sched_domain_attr *cur, int idx_cur, cur 2185 kernel/sched/topology.c if (!new && !cur) cur 2190 kernel/sched/topology.c return !memcmp(cur ? (cur + idx_cur) : &tmp, cur 1014 kernel/seccomp.c struct seccomp_knotif *knotif = NULL, *cur; cur 1032 kernel/seccomp.c list_for_each_entry(cur, &filter->notif->notifications, list) { cur 1033 kernel/seccomp.c if (cur->state == SECCOMP_NOTIFY_INIT) { cur 1034 kernel/seccomp.c knotif = cur; cur 1070 kernel/seccomp.c list_for_each_entry(cur, &filter->notif->notifications, list) { cur 1071 kernel/seccomp.c if (cur->id == unotif.id) { cur 1072 kernel/seccomp.c knotif = cur; cur 1091 kernel/seccomp.c struct seccomp_knotif *knotif = NULL, *cur; cur 1104 kernel/seccomp.c list_for_each_entry(cur, &filter->notif->notifications, list) { cur 1105 kernel/seccomp.c if (cur->id == resp.id) { cur 1106 kernel/seccomp.c knotif = cur; cur 1183 kernel/seccomp.c struct seccomp_knotif *cur; cur 1190 kernel/seccomp.c list_for_each_entry(cur, &filter->notif->notifications, list) { cur 1191 kernel/seccomp.c if (cur->state == SECCOMP_NOTIFY_INIT) cur 1193 kernel/seccomp.c if (cur->state == SECCOMP_NOTIFY_SENT) cur 1214 kernel/seccomp.c struct seccomp_filter *cur; cur 1216 kernel/seccomp.c for (cur = current->seccomp.filter; cur; cur = cur->prev) { cur 1217 kernel/seccomp.c if (cur->notif) cur 1617 kernel/seccomp.c const struct seccomp_log_name *cur; cur 1620 kernel/seccomp.c for (cur = seccomp_log_names; cur->name && size; cur++) { cur 1623 kernel/seccomp.c if (!(actions_logged & cur->log)) cur 1636 kernel/seccomp.c ret = strscpy(names, cur->name, size); cur 1650 kernel/seccomp.c const struct seccomp_log_name *cur; cur 1652 kernel/seccomp.c for (cur = seccomp_log_names; cur->name; cur++) { cur 1653 kernel/seccomp.c if (!strcmp(cur->name, name)) { cur 1654 kernel/seccomp.c *action_logged = cur->log; cur 215 kernel/smpboot.c struct smp_hotplug_thread *cur; cur 219 kernel/smpboot.c list_for_each_entry(cur, &hotplug_threads, list) { cur 220 kernel/smpboot.c ret = __smpboot_create_thread(cur, cpu); cur 238 kernel/smpboot.c struct smp_hotplug_thread *cur; cur 241 kernel/smpboot.c list_for_each_entry(cur, &hotplug_threads, list) cur 242 kernel/smpboot.c smpboot_unpark_thread(cur, cpu); cur 257 kernel/smpboot.c struct smp_hotplug_thread *cur; cur 260 kernel/smpboot.c list_for_each_entry_reverse(cur, &hotplug_threads, list) cur 261 kernel/smpboot.c smpboot_park_thread(cur, cpu); cur 96 kernel/time/tick-broadcast.c struct clock_event_device *cur = tick_broadcast_device.evtdev; cur 98 kernel/time/tick-broadcast.c if (!tick_check_broadcast_device(cur, dev)) cur 104 kernel/time/tick-broadcast.c clockevents_exchange_device(cur, dev); cur 105 kernel/time/tick-broadcast.c if (cur) cur 106 kernel/time/tick-broadcast.c cur->event_handler = clockevents_handle_noop; cur 52 lib/cmdline.c char *cur = *str; cur 54 lib/cmdline.c if (!cur || !(*cur)) cur 56 lib/cmdline.c *pint = simple_strtol(cur, str, 0); cur 57 lib/cmdline.c if (cur == *str) cur 60 lib/errseq.c errseq_t cur, old; cur 89 lib/errseq.c cur = new; cur 94 lib/errseq.c cur = cmpxchg(eseq, old, new); cur 100 lib/errseq.c if (likely(cur == old || cur == new)) cur 104 lib/errseq.c old = cur; cur 106 lib/errseq.c return cur; cur 146 lib/errseq.c errseq_t cur = READ_ONCE(*eseq); cur 148 lib/errseq.c if (likely(cur == since)) cur 150 lib/errseq.c return -(cur & MAX_ERRNO); cur 153 lib/kobject.c int cur = strlen(kobject_name(parent)); cur 155 lib/kobject.c length -= cur; cur 156 lib/kobject.c memcpy(path + length, kobject_name(parent), cur); cur 173 lib/rbtree_test.c struct test_node *cur, *n; cur 175 lib/rbtree_test.c rbtree_postorder_for_each_entry_safe(cur, n, &root.rb_root, rb) cur 74 lib/test_list_sort.c struct list_head *cur; cur 100 lib/test_list_sort.c for (cur = head.next; cur->next != &head; cur = cur->next) { cur 104 lib/test_list_sort.c if (cur->next->prev != cur) { cur 109 lib/test_list_sort.c cmp_result = cmp(NULL, cur, cur->next); cur 115 lib/test_list_sort.c el = container_of(cur, struct debug_el, list); cur 116 lib/test_list_sort.c el1 = container_of(cur->next, struct debug_el, list); cur 129 lib/test_list_sort.c if (head.prev != cur) { cur 137 lib/ts_fsm.c struct ts_fsm_token *cur = NULL, *next; cur 166 lib/ts_fsm.c cur = &fsm->tokens[tok_idx]; cur 173 lib/ts_fsm.c switch (cur->recur) { cur 178 lib/ts_fsm.c if (!match_token(cur, data[block_idx])) cur 184 lib/ts_fsm.c !match_token(cur, data[block_idx])) cur 192 lib/ts_fsm.c if (!match_token(cur, data[block_idx])) cur 206 lib/ts_fsm.c if (!match_token(cur, data[block_idx])) cur 229 lib/ts_fsm.c if (!match_token(cur, data[block_idx])) cur 1171 lib/vsprintf.c int cur, rbot, rtop; cur 1177 lib/vsprintf.c rbot = cur = find_first_bit(bitmap, nr_bits); cur 1178 lib/vsprintf.c while (cur < nr_bits) { cur 1179 lib/vsprintf.c rtop = cur; cur 1180 lib/vsprintf.c cur = find_next_bit(bitmap, nr_bits, cur + 1); cur 1181 lib/vsprintf.c if (cur < nr_bits && cur <= rtop + 1) cur 1200 lib/vsprintf.c rbot = cur; cur 440 lib/zstd/zstd_opt.h U32 cur, match_num, last_pos, litlen, price; cur 457 lib/zstd/zstd_opt.h cur = 0; cur 482 lib/zstd/zstd_opt.h cur = 0; cur 515 lib/zstd/zstd_opt.h for (cur = 1; cur <= last_pos; cur++) { cur 516 lib/zstd/zstd_opt.h inr = ip + cur; cur 518 lib/zstd/zstd_opt.h if (opt[cur - 1].mlen == 1) { cur 519 lib/zstd/zstd_opt.h litlen = opt[cur - 1].litlen + 1; cur 520 lib/zstd/zstd_opt.h if (cur > litlen) { cur 521 lib/zstd/zstd_opt.h price = opt[cur - litlen].price + ZSTD_getLiteralPrice(seqStorePtr, litlen, inr - litlen); cur 526 lib/zstd/zstd_opt.h price = opt[cur - 1].price + ZSTD_getLiteralPrice(seqStorePtr, litlen, inr - 1); cur 529 lib/zstd/zstd_opt.h if (cur > last_pos || price <= opt[cur].price) cur 530 lib/zstd/zstd_opt.h SET_PRICE(cur, 1, 0, litlen, price); cur 532 lib/zstd/zstd_opt.h if (cur == last_pos) cur 538 lib/zstd/zstd_opt.h mlen = opt[cur].mlen; cur 539 lib/zstd/zstd_opt.h if (opt[cur].off > ZSTD_REP_MOVE_OPT) { cur 540 lib/zstd/zstd_opt.h opt[cur].rep[2] = opt[cur - mlen].rep[1]; cur 541 lib/zstd/zstd_opt.h opt[cur].rep[1] = opt[cur - mlen].rep[0]; cur 542 lib/zstd/zstd_opt.h opt[cur].rep[0] = opt[cur].off - ZSTD_REP_MOVE_OPT; cur 544 lib/zstd/zstd_opt.h opt[cur].rep[2] = (opt[cur].off > 1) ? opt[cur - mlen].rep[1] : opt[cur - mlen].rep[2]; cur 545 lib/zstd/zstd_opt.h opt[cur].rep[1] = (opt[cur].off > 0) ? opt[cur - mlen].rep[0] : opt[cur - mlen].rep[1]; cur 546 lib/zstd/zstd_opt.h opt[cur].rep[0] = cur 547 lib/zstd/zstd_opt.h ((opt[cur].off == ZSTD_REP_MOVE_OPT) && (mlen != 1)) ? (opt[cur - mlen].rep[0] - 1) : (opt[cur - mlen].rep[opt[cur].off]); cur 553 lib/zstd/zstd_opt.h for (i = (opt[cur].mlen != 1); i < last_i; i++) { /* check rep */ cur 554 lib/zstd/zstd_opt.h const S32 repCur = (i == ZSTD_REP_MOVE_OPT) ? (opt[cur].rep[0] - 1) : opt[cur].rep[i]; cur 559 lib/zstd/zstd_opt.h if (mlen > sufficient_len || cur + mlen >= ZSTD_OPT_NUM) { cur 562 lib/zstd/zstd_opt.h last_pos = cur + 1; cur 566 lib/zstd/zstd_opt.h best_off = i - (opt[cur].mlen != 1); cur 571 lib/zstd/zstd_opt.h if (opt[cur].mlen == 1) { cur 572 lib/zstd/zstd_opt.h litlen = opt[cur].litlen; cur 573 lib/zstd/zstd_opt.h if (cur > litlen) { cur 574 lib/zstd/zstd_opt.h price = opt[cur - litlen].price + ZSTD_getPrice(seqStorePtr, litlen, inr - litlen, cur 580 lib/zstd/zstd_opt.h price = opt[cur].price + ZSTD_getPrice(seqStorePtr, 0, NULL, best_off, mlen - MINMATCH, ultra); cur 583 lib/zstd/zstd_opt.h if (cur + mlen > last_pos || price <= opt[cur + mlen].price) cur 584 lib/zstd/zstd_opt.h SET_PRICE(cur + mlen, mlen, i, litlen, price); cur 593 lib/zstd/zstd_opt.h if (match_num > 0 && (matches[match_num - 1].len > sufficient_len || cur + matches[match_num - 1].len >= ZSTD_OPT_NUM)) { cur 596 lib/zstd/zstd_opt.h last_pos = cur + 1; cur 606 lib/zstd/zstd_opt.h if (opt[cur].mlen == 1) { cur 607 lib/zstd/zstd_opt.h litlen = opt[cur].litlen; cur 608 lib/zstd/zstd_opt.h if (cur > litlen) cur 609 lib/zstd/zstd_opt.h price = opt[cur - litlen].price + ZSTD_getPrice(seqStorePtr, litlen, ip + cur - litlen, cur 615 lib/zstd/zstd_opt.h price = opt[cur].price + ZSTD_getPrice(seqStorePtr, 0, NULL, matches[u].off - 1, mlen - MINMATCH, ultra); cur 618 lib/zstd/zstd_opt.h if (cur + mlen > last_pos || (price < opt[cur + mlen].price)) cur 619 lib/zstd/zstd_opt.h SET_PRICE(cur + mlen, mlen, matches[u].off, litlen, price); cur 628 lib/zstd/zstd_opt.h cur = last_pos - best_mlen; cur 635 lib/zstd/zstd_opt.h mlen = opt[cur].mlen; cur 636 lib/zstd/zstd_opt.h offset = opt[cur].off; cur 637 lib/zstd/zstd_opt.h opt[cur].mlen = best_mlen; cur 638 lib/zstd/zstd_opt.h opt[cur].off = best_off; cur 641 lib/zstd/zstd_opt.h if (mlen > cur) cur 643 lib/zstd/zstd_opt.h cur -= mlen; cur 650 lib/zstd/zstd_opt.h for (cur = 0; cur < last_pos;) { cur 651 lib/zstd/zstd_opt.h mlen = opt[cur].mlen; cur 654 lib/zstd/zstd_opt.h cur++; cur 657 lib/zstd/zstd_opt.h offset = opt[cur].off; cur 658 lib/zstd/zstd_opt.h cur += mlen; cur 738 lib/zstd/zstd_opt.h U32 cur, match_num, last_pos, litlen, price; cur 763 lib/zstd/zstd_opt.h cur = 0; cur 797 lib/zstd/zstd_opt.h cur = 0; cur 823 lib/zstd/zstd_opt.h for (cur = 1; cur <= last_pos; cur++) { cur 824 lib/zstd/zstd_opt.h inr = ip + cur; cur 826 lib/zstd/zstd_opt.h if (opt[cur - 1].mlen == 1) { cur 827 lib/zstd/zstd_opt.h litlen = opt[cur - 1].litlen + 1; cur 828 lib/zstd/zstd_opt.h if (cur > litlen) { cur 829 lib/zstd/zstd_opt.h price = opt[cur - litlen].price + ZSTD_getLiteralPrice(seqStorePtr, litlen, inr - litlen); cur 834 lib/zstd/zstd_opt.h price = opt[cur - 1].price + ZSTD_getLiteralPrice(seqStorePtr, litlen, inr - 1); cur 837 lib/zstd/zstd_opt.h if (cur > last_pos || price <= opt[cur].price) cur 838 lib/zstd/zstd_opt.h SET_PRICE(cur, 1, 0, litlen, price); cur 840 lib/zstd/zstd_opt.h if (cur == last_pos) cur 846 lib/zstd/zstd_opt.h mlen = opt[cur].mlen; cur 847 lib/zstd/zstd_opt.h if (opt[cur].off > ZSTD_REP_MOVE_OPT) { cur 848 lib/zstd/zstd_opt.h opt[cur].rep[2] = opt[cur - mlen].rep[1]; cur 849 lib/zstd/zstd_opt.h opt[cur].rep[1] = opt[cur - mlen].rep[0]; cur 850 lib/zstd/zstd_opt.h opt[cur].rep[0] = opt[cur].off - ZSTD_REP_MOVE_OPT; cur 852 lib/zstd/zstd_opt.h opt[cur].rep[2] = (opt[cur].off > 1) ? opt[cur - mlen].rep[1] : opt[cur - mlen].rep[2]; cur 853 lib/zstd/zstd_opt.h opt[cur].rep[1] = (opt[cur].off > 0) ? opt[cur - mlen].rep[0] : opt[cur - mlen].rep[1]; cur 854 lib/zstd/zstd_opt.h opt[cur].rep[0] = cur 855 lib/zstd/zstd_opt.h ((opt[cur].off == ZSTD_REP_MOVE_OPT) && (mlen != 1)) ? (opt[cur - mlen].rep[0] - 1) : (opt[cur - mlen].rep[opt[cur].off]); cur 862 lib/zstd/zstd_opt.h const S32 repCur = (i == ZSTD_REP_MOVE_OPT) ? (opt[cur].rep[0] - 1) : opt[cur].rep[i]; cur 863 lib/zstd/zstd_opt.h const U32 repIndex = (U32)(curr + cur - repCur); cur 866 lib/zstd/zstd_opt.h if ((repCur > 0 && repCur <= (S32)(curr + cur)) && cur 873 lib/zstd/zstd_opt.h if (mlen > sufficient_len || cur + mlen >= ZSTD_OPT_NUM) { cur 876 lib/zstd/zstd_opt.h last_pos = cur + 1; cur 880 lib/zstd/zstd_opt.h best_off = i - (opt[cur].mlen != 1); cur 885 lib/zstd/zstd_opt.h if (opt[cur].mlen == 1) { cur 886 lib/zstd/zstd_opt.h litlen = opt[cur].litlen; cur 887 lib/zstd/zstd_opt.h if (cur > litlen) { cur 888 lib/zstd/zstd_opt.h price = opt[cur - litlen].price + ZSTD_getPrice(seqStorePtr, litlen, inr - litlen, cur 894 lib/zstd/zstd_opt.h price = opt[cur].price + ZSTD_getPrice(seqStorePtr, 0, NULL, best_off, mlen - MINMATCH, ultra); cur 897 lib/zstd/zstd_opt.h if (cur + mlen > last_pos || price <= opt[cur + mlen].price) cur 898 lib/zstd/zstd_opt.h SET_PRICE(cur + mlen, mlen, i, litlen, price); cur 907 lib/zstd/zstd_opt.h if (match_num > 0 && (matches[match_num - 1].len > sufficient_len || cur + matches[match_num - 1].len >= ZSTD_OPT_NUM)) { cur 910 lib/zstd/zstd_opt.h last_pos = cur + 1; cur 920 lib/zstd/zstd_opt.h if (opt[cur].mlen == 1) { cur 921 lib/zstd/zstd_opt.h litlen = opt[cur].litlen; cur 922 lib/zstd/zstd_opt.h if (cur > litlen) cur 923 lib/zstd/zstd_opt.h price = opt[cur - litlen].price + ZSTD_getPrice(seqStorePtr, litlen, ip + cur - litlen, cur 929 lib/zstd/zstd_opt.h price = opt[cur].price + ZSTD_getPrice(seqStorePtr, 0, NULL, matches[u].off - 1, mlen - MINMATCH, ultra); cur 932 lib/zstd/zstd_opt.h if (cur + mlen > last_pos || (price < opt[cur + mlen].price)) cur 933 lib/zstd/zstd_opt.h SET_PRICE(cur + mlen, mlen, matches[u].off, litlen, price); cur 942 lib/zstd/zstd_opt.h cur = last_pos - best_mlen; cur 949 lib/zstd/zstd_opt.h mlen = opt[cur].mlen; cur 950 lib/zstd/zstd_opt.h offset = opt[cur].off; cur 951 lib/zstd/zstd_opt.h opt[cur].mlen = best_mlen; cur 952 lib/zstd/zstd_opt.h opt[cur].off = best_off; cur 955 lib/zstd/zstd_opt.h if (mlen > cur) cur 957 lib/zstd/zstd_opt.h cur -= mlen; cur 964 lib/zstd/zstd_opt.h for (cur = 0; cur < last_pos;) { cur 965 lib/zstd/zstd_opt.h mlen = opt[cur].mlen; cur 968 lib/zstd/zstd_opt.h cur++; cur 971 lib/zstd/zstd_opt.h offset = opt[cur].off; cur 972 lib/zstd/zstd_opt.h cur += mlen; cur 278 mm/readahead.c unsigned long cur = ra->size; cur 280 mm/readahead.c if (cur < max / 16) cur 281 mm/readahead.c return 4 * cur; cur 282 mm/readahead.c if (cur <= max / 2) cur 283 mm/readahead.c return 2 * cur; cur 240 mm/slob.c slob_t *prev, *cur, *aligned = NULL; cur 244 mm/slob.c for (prev = NULL, cur = sp->freelist; ; prev = cur, cur = slob_next(cur)) { cur 245 mm/slob.c slobidx_t avail = slob_units(cur); cur 256 mm/slob.c (ALIGN((unsigned long)cur + align_offset, align) cur 258 mm/slob.c delta = aligned - cur; cur 264 mm/slob.c next = slob_next(cur); cur 266 mm/slob.c set_slob(cur, delta, aligned); cur 267 mm/slob.c prev = cur; cur 268 mm/slob.c cur = aligned; cur 269 mm/slob.c avail = slob_units(cur); cur 272 mm/slob.c next = slob_next(cur); cur 280 mm/slob.c set_slob(prev, slob_units(prev), cur + units); cur 282 mm/slob.c sp->freelist = cur + units; cur 283 mm/slob.c set_slob(cur + units, avail - units, next); cur 291 mm/slob.c return cur; cur 293 mm/slob.c if (slob_last(cur)) cur 1574 mm/slub.c void *cur; cur 1588 mm/slub.c cur = next_freelist_entry(s, page, &pos, start, page_limit, cur 1590 mm/slub.c cur = setup_object(s, page, cur); cur 1591 mm/slub.c page->freelist = cur; cur 1597 mm/slub.c set_freepointer(s, cur, next); cur 1598 mm/slub.c cur = next; cur 1600 mm/slub.c set_freepointer(s, cur, NULL); cur 148 mm/swap_slots.c cache->cur = 0; cur 179 mm/swap_slots.c swapcache_free_entries(cache->slots + cache->cur, cache->nr); cur 180 mm/swap_slots.c cache->cur = 0; cur 270 mm/swap_slots.c cache->cur = 0; cur 339 mm/swap_slots.c pentry = &cache->slots[cache->cur++]; cur 274 net/bridge/br.c bool cur = !!br_opt_get(br, opt); cur 277 net/bridge/br.c opt, cur, on); cur 279 net/bridge/br.c if (cur == on) cur 1844 net/ceph/osd_client.c struct ceph_osd_backoff *cur = cur 1848 net/ceph/osd_client.c cmp = hoid_compare(hoid, cur->begin); cur 1852 net/ceph/osd_client.c if (hoid_compare(hoid, cur->end) < 0) cur 1853 net/ceph/osd_client.c return cur; cur 1857 net/ceph/osd_client.c return cur; cur 6622 net/core/dev.c int ret, cur = 0; cur 6645 net/core/dev.c dev_stack[cur] = now; cur 6646 net/core/dev.c iter_stack[cur++] = iter; cur 6651 net/core/dev.c if (!cur) cur 6653 net/core/dev.c next = dev_stack[--cur]; cur 6654 net/core/dev.c niter = iter_stack[cur]; cur 6671 net/core/dev.c int ret, cur = 0; cur 6691 net/core/dev.c dev_stack[cur] = now; cur 6692 net/core/dev.c iter_stack[cur++] = iter; cur 6697 net/core/dev.c if (!cur) cur 6699 net/core/dev.c next = dev_stack[--cur]; cur 6700 net/core/dev.c niter = iter_stack[cur]; cur 6840 net/core/dev.c int ret, cur = 0; cur 6860 net/core/dev.c dev_stack[cur] = now; cur 6861 net/core/dev.c iter_stack[cur++] = iter; cur 6866 net/core/dev.c if (!cur) cur 6868 net/core/dev.c next = dev_stack[--cur]; cur 6869 net/core/dev.c niter = iter_stack[cur]; cur 6887 net/core/dev.c int ret, cur = 0; cur 6910 net/core/dev.c dev_stack[cur] = now; cur 6911 net/core/dev.c iter_stack[cur++] = iter; cur 6916 net/core/dev.c if (!cur) cur 6918 net/core/dev.c next = dev_stack[--cur]; cur 6919 net/core/dev.c niter = iter_stack[cur]; cur 7003 net/core/dev.c int ret, cur = 0; cur 7023 net/core/dev.c dev_stack[cur] = now; cur 7024 net/core/dev.c iter_stack[cur++] = iter; cur 7029 net/core/dev.c if (!cur) cur 7031 net/core/dev.c next = dev_stack[--cur]; cur 7032 net/core/dev.c niter = iter_stack[cur]; cur 1141 net/core/devlink.c u32 cur; cur 1145 net/core/devlink.c pool_index, &cur, &max); cur 1149 net/core/devlink.c if (nla_put_u32(msg, DEVLINK_ATTR_SB_OCC_CUR, cur)) cur 1339 net/core/devlink.c u32 cur; cur 1345 net/core/devlink.c &cur, &max); cur 1349 net/core/devlink.c if (nla_put_u32(msg, DEVLINK_ATTR_SB_OCC_CUR, cur)) cur 504 net/core/netpoll.c char *cur=opt, *delim; cur 508 net/core/netpoll.c if (*cur != '@') { cur 509 net/core/netpoll.c if ((delim = strchr(cur, '@')) == NULL) cur 512 net/core/netpoll.c if (kstrtou16(cur, 10, &np->local_port)) cur 514 net/core/netpoll.c cur = delim; cur 516 net/core/netpoll.c cur++; cur 518 net/core/netpoll.c if (*cur != '/') { cur 520 net/core/netpoll.c if ((delim = strchr(cur, '/')) == NULL) cur 523 net/core/netpoll.c ipv6 = netpoll_parse_ip_addr(cur, &np->local_ip); cur 528 net/core/netpoll.c cur = delim; cur 530 net/core/netpoll.c cur++; cur 532 net/core/netpoll.c if (*cur != ',') { cur 534 net/core/netpoll.c if ((delim = strchr(cur, ',')) == NULL) cur 537 net/core/netpoll.c strlcpy(np->dev_name, cur, sizeof(np->dev_name)); cur 538 net/core/netpoll.c cur = delim; cur 540 net/core/netpoll.c cur++; cur 542 net/core/netpoll.c if (*cur != '@') { cur 544 net/core/netpoll.c if ((delim = strchr(cur, '@')) == NULL) cur 547 net/core/netpoll.c if (*cur == ' ' || *cur == '\t') cur 549 net/core/netpoll.c if (kstrtou16(cur, 10, &np->remote_port)) cur 551 net/core/netpoll.c cur = delim; cur 553 net/core/netpoll.c cur++; cur 556 net/core/netpoll.c if ((delim = strchr(cur, '/')) == NULL) cur 559 net/core/netpoll.c ipv6 = netpoll_parse_ip_addr(cur, &np->remote_ip); cur 566 net/core/netpoll.c cur = delim + 1; cur 568 net/core/netpoll.c if (*cur != 0) { cur 570 net/core/netpoll.c if (!mac_pton(cur, np->remote_mac)) cur 579 net/core/netpoll.c np_info(np, "couldn't parse config at '%s'!\n", cur); cur 3232 net/core/pktgen.c struct pktgen_dev *cur; cur 3237 net/core/pktgen.c cur = list_entry(q, struct pktgen_dev, list); cur 3239 net/core/pktgen.c if (!cur->removal_mark) cur 3242 net/core/pktgen.c kfree_skb(cur->skb); cur 3243 net/core/pktgen.c cur->skb = NULL; cur 3245 net/core/pktgen.c pktgen_remove_device(t, cur); cur 3254 net/core/pktgen.c struct pktgen_dev *cur; cur 3261 net/core/pktgen.c cur = list_entry(q, struct pktgen_dev, list); cur 3263 net/core/pktgen.c kfree_skb(cur->skb); cur 3264 net/core/pktgen.c cur->skb = NULL; cur 3266 net/core/pktgen.c pktgen_remove_device(t, cur); cur 121 net/core/sysctl_net_core.c struct sd_flow_limit *cur; cur 135 net/core/sysctl_net_core.c len = sizeof(*cur) + netdev_flow_limit_table_len; cur 138 net/core/sysctl_net_core.c cur = rcu_dereference_protected(sd->flow_limit, cur 140 net/core/sysctl_net_core.c if (cur && !cpumask_test_cpu(i, mask)) { cur 143 net/core/sysctl_net_core.c kfree(cur); cur 144 net/core/sysctl_net_core.c } else if (!cur && cpumask_test_cpu(i, mask)) { cur 145 net/core/sysctl_net_core.c cur = kzalloc_node(len, GFP_KERNEL, cur 147 net/core/sysctl_net_core.c if (!cur) { cur 152 net/core/sysctl_net_core.c cur->num_buckets = netdev_flow_limit_table_len; cur 153 net/core/sysctl_net_core.c rcu_assign_pointer(sd->flow_limit, cur); cur 30 net/dccp/ackvec.c struct dccp_ackvec_record *cur, *next; cur 32 net/dccp/ackvec.c list_for_each_entry_safe(cur, next, &av->av_records, avr_node) cur 33 net/dccp/ackvec.c kmem_cache_free(dccp_ackvec_record_slab, cur); cur 364 net/dccp/ackvec.c struct dccp_ackvec_parsed *cur, *next; cur 366 net/dccp/ackvec.c list_for_each_entry_safe(cur, next, parsed_chunks, node) cur 367 net/dccp/ackvec.c kfree(cur); cur 86 net/dccp/ccids/lib/loss_interval.c struct tfrc_loss_interval *cur = tfrc_lh_peek(lh); cur 90 net/dccp/ccids/lib/loss_interval.c if (cur == NULL) /* not initialised */ cur 93 net/dccp/ccids/lib/loss_interval.c len = dccp_delta_seqno(cur->li_seqno, DCCP_SKB_CB(skb)->dccpd_seq) + 1; cur 95 net/dccp/ccids/lib/loss_interval.c if (len - (s64)cur->li_length <= 0) /* duplicate or reordered */ cur 98 net/dccp/ccids/lib/loss_interval.c if (SUB16(dccp_hdr(skb)->dccph_ccval, cur->li_ccval) > 4) cur 107 net/dccp/ccids/lib/loss_interval.c cur->li_is_closed = 1; cur 112 net/dccp/ccids/lib/loss_interval.c cur->li_length = len; cur 119 net/dccp/ccids/lib/loss_interval.c static inline u8 tfrc_lh_is_new_loss(struct tfrc_loss_interval *cur, cur 122 net/dccp/ccids/lib/loss_interval.c return dccp_delta_seqno(cur->li_seqno, new_loss->tfrchrx_seqno) > 0 && cur 123 net/dccp/ccids/lib/loss_interval.c (cur->li_is_closed || SUB16(new_loss->tfrchrx_ccval, cur->li_ccval) > 4); cur 138 net/dccp/ccids/lib/loss_interval.c struct tfrc_loss_interval *cur = tfrc_lh_peek(lh), *new; cur 140 net/dccp/ccids/lib/loss_interval.c if (cur != NULL && !tfrc_lh_is_new_loss(cur, tfrc_rx_hist_loss_prev(rh))) cur 156 net/dccp/ccids/lib/loss_interval.c cur->li_length = dccp_delta_seqno(cur->li_seqno, new->li_seqno); cur 1500 net/dccp/feat.c struct dccp_feat_entry *cur, *next; cur 1506 net/dccp/feat.c list_for_each_entry(cur, fn_list, node) { cur 1512 net/dccp/feat.c if (cur->empty_confirm) cur 1515 net/dccp/feat.c idx = dccp_feat_index(cur->feat_num); cur 1517 net/dccp/feat.c DCCP_BUG("Unknown feature %u", cur->feat_num); cur 1520 net/dccp/feat.c if (cur->state != FEAT_STABLE) { cur 1522 net/dccp/feat.c cur->is_local ? "local" : "remote", cur 1523 net/dccp/feat.c dccp_feat_fname(cur->feat_num), cur 1524 net/dccp/feat.c dccp_feat_sname[cur->state]); cur 1527 net/dccp/feat.c fvals[idx][cur->is_local] = &cur->val; cur 1544 net/dccp/feat.c list_for_each_entry_safe(cur, next, fn_list, node) cur 1545 net/dccp/feat.c if (!cur->needs_confirm) cur 1546 net/dccp/feat.c dccp_feat_list_pop(cur); cur 2146 net/ipv4/tcp_ipv4.c static void *listening_get_next(struct seq_file *seq, void *cur) cur 2153 net/ipv4/tcp_ipv4.c struct sock *sk = cur; cur 2239 net/ipv4/tcp_ipv4.c static void *established_get_next(struct seq_file *seq, void *cur) cur 2242 net/ipv4/tcp_ipv4.c struct sock *sk = cur; cur 1916 net/ipv6/ip6_fib.c struct fib6_info *cur = rcu_dereference_protected(*rtp, cur 1918 net/ipv6/ip6_fib.c if (rt == cur) { cur 1922 net/ipv6/ip6_fib.c rtp_next = &cur->fib6_next; cur 181 net/ipv6/ndisc.c static struct nd_opt_hdr *ndisc_next_option(struct nd_opt_hdr *cur, cur 185 net/ipv6/ndisc.c if (!cur || !end || cur >= end) cur 187 net/ipv6/ndisc.c type = cur->nd_opt_type; cur 189 net/ipv6/ndisc.c cur = ((void *)cur) + (cur->nd_opt_len << 3); cur 190 net/ipv6/ndisc.c } while (cur < end && cur->nd_opt_type != type); cur 191 net/ipv6/ndisc.c return cur <= end && cur->nd_opt_type == type ? cur : NULL; cur 204 net/ipv6/ndisc.c struct nd_opt_hdr *cur, cur 207 net/ipv6/ndisc.c if (!cur || !end || cur >= end) cur 210 net/ipv6/ndisc.c cur = ((void *)cur) + (cur->nd_opt_len << 3); cur 211 net/ipv6/ndisc.c } while (cur < end && !ndisc_is_useropt(dev, cur)); cur 212 net/ipv6/ndisc.c return cur <= end && ndisc_is_useropt(dev, cur) ? cur : NULL; cur 3663 net/mac80211/util.c s32 cur; cur 3671 net/mac80211/util.c cur = data->desc[i].start - tsf; cur 3672 net/mac80211/util.c if (cur > *offset) cur 3675 net/mac80211/util.c cur = data->desc[i].start + data->desc[i].duration - tsf; cur 3676 net/mac80211/util.c if (cur > *offset) cur 3677 net/mac80211/util.c *offset = cur; cur 95 net/netfilter/nf_conntrack_h323_asn1.c unsigned char *cur; cur 100 net/netfilter/nf_conntrack_h323_asn1.c #define INC_BIT(bs) if((++(bs)->bit)>7){(bs)->cur++;(bs)->bit=0;} cur 101 net/netfilter/nf_conntrack_h323_asn1.c #define INC_BITS(bs,b) if(((bs)->bit+=(b))>7){(bs)->cur+=(bs)->bit>>3;(bs)->bit&=7;} cur 102 net/netfilter/nf_conntrack_h323_asn1.c #define BYTE_ALIGN(bs) if((bs)->bit){(bs)->cur++;(bs)->bit=0;} cur 154 net/netfilter/nf_conntrack_h323_asn1.c v = *bs->cur++; cur 159 net/netfilter/nf_conntrack_h323_asn1.c v += *bs->cur++; cur 172 net/netfilter/nf_conntrack_h323_asn1.c if (bs->cur + bytes > bs->end) cur 180 net/netfilter/nf_conntrack_h323_asn1.c unsigned int b = (*bs->cur) & (0x80 >> bs->bit); cur 192 net/netfilter/nf_conntrack_h323_asn1.c v = (*bs->cur) & (0xffU >> bs->bit); cur 199 net/netfilter/nf_conntrack_h323_asn1.c bs->cur++; cur 204 net/netfilter/nf_conntrack_h323_asn1.c v += *(++bs->cur); cur 223 net/netfilter/nf_conntrack_h323_asn1.c v = (unsigned int)(*bs->cur) << (bs->bit + 24); cur 226 net/netfilter/nf_conntrack_h323_asn1.c v = (unsigned int)(*bs->cur++) << (bs->bit + 24); cur 231 net/netfilter/nf_conntrack_h323_asn1.c v |= (unsigned int)(*bs->cur++) << shift; cur 234 net/netfilter/nf_conntrack_h323_asn1.c v |= (unsigned int)(*bs->cur) << shift; cur 238 net/netfilter/nf_conntrack_h323_asn1.c v |= (*bs->cur) >> (8 - bs->bit); cur 258 net/netfilter/nf_conntrack_h323_asn1.c v |= *bs->cur++; cur 262 net/netfilter/nf_conntrack_h323_asn1.c v |= *bs->cur++; cur 266 net/netfilter/nf_conntrack_h323_asn1.c v |= *bs->cur++; cur 270 net/netfilter/nf_conntrack_h323_asn1.c v |= *bs->cur++; cur 306 net/netfilter/nf_conntrack_h323_asn1.c len = *bs->cur++; cur 307 net/netfilter/nf_conntrack_h323_asn1.c bs->cur += len; cur 324 net/netfilter/nf_conntrack_h323_asn1.c bs->cur++; cur 328 net/netfilter/nf_conntrack_h323_asn1.c bs->cur += 2; cur 340 net/netfilter/nf_conntrack_h323_asn1.c bs->cur += len; cur 347 net/netfilter/nf_conntrack_h323_asn1.c bs->cur += len; cur 392 net/netfilter/nf_conntrack_h323_asn1.c len = (*bs->cur++) << 8; cur 393 net/netfilter/nf_conntrack_h323_asn1.c len += (*bs->cur++) + f->lb; cur 405 net/netfilter/nf_conntrack_h323_asn1.c bs->cur += len >> 3; cur 448 net/netfilter/nf_conntrack_h323_asn1.c bs->cur[0], bs->cur[1], cur 449 net/netfilter/nf_conntrack_h323_asn1.c bs->cur[2], bs->cur[3], cur 450 net/netfilter/nf_conntrack_h323_asn1.c bs->cur[4] * 256 + bs->cur[5])); cur 452 net/netfilter/nf_conntrack_h323_asn1.c bs->cur - bs->buf; cur 461 net/netfilter/nf_conntrack_h323_asn1.c len = (*bs->cur++) + f->lb; cur 477 net/netfilter/nf_conntrack_h323_asn1.c bs->cur += len; cur 498 net/netfilter/nf_conntrack_h323_asn1.c len = (*bs->cur++) + f->lb; cur 508 net/netfilter/nf_conntrack_h323_asn1.c bs->cur += len << 1; cur 563 net/netfilter/nf_conntrack_h323_asn1.c bs->cur += len; cur 566 net/netfilter/nf_conntrack_h323_asn1.c beg = bs->cur; cur 574 net/netfilter/nf_conntrack_h323_asn1.c bs->cur = beg + len; cur 607 net/netfilter/nf_conntrack_h323_asn1.c bs->cur += len; cur 628 net/netfilter/nf_conntrack_h323_asn1.c bs->cur += len; cur 631 net/netfilter/nf_conntrack_h323_asn1.c beg = bs->cur; cur 638 net/netfilter/nf_conntrack_h323_asn1.c bs->cur = beg + len; cur 663 net/netfilter/nf_conntrack_h323_asn1.c count = *bs->cur++; cur 669 net/netfilter/nf_conntrack_h323_asn1.c count = *bs->cur++; cur 671 net/netfilter/nf_conntrack_h323_asn1.c count += *bs->cur++; cur 709 net/netfilter/nf_conntrack_h323_asn1.c bs->cur += len; cur 712 net/netfilter/nf_conntrack_h323_asn1.c beg = bs->cur; cur 722 net/netfilter/nf_conntrack_h323_asn1.c bs->cur = beg + len; cur 782 net/netfilter/nf_conntrack_h323_asn1.c bs->cur += len; cur 803 net/netfilter/nf_conntrack_h323_asn1.c bs->cur += len; cur 806 net/netfilter/nf_conntrack_h323_asn1.c beg = bs->cur; cur 812 net/netfilter/nf_conntrack_h323_asn1.c bs->cur = beg + len; cur 829 net/netfilter/nf_conntrack_h323_asn1.c bs.buf = bs.beg = bs.cur = buf; cur 846 net/netfilter/nf_conntrack_h323_asn1.c bs.beg = bs.cur = beg; cur 863 net/netfilter/nf_conntrack_h323_asn1.c bs.buf = bs.beg = bs.cur = buf; cur 137 net/netfilter/nf_conntrack_helper.c struct nf_conntrack_nat_helper *cur; cur 140 net/netfilter/nf_conntrack_helper.c list_for_each_entry_rcu(cur, &nf_ct_nat_helpers, list) { cur 141 net/netfilter/nf_conntrack_helper.c if (!strcmp(cur->mod_name, mod_name)) { cur 146 net/netfilter/nf_conntrack_helper.c return found ? cur : NULL; cur 339 net/netfilter/nf_conntrack_helper.c struct nf_ct_helper_expectfn *cur; cur 342 net/netfilter/nf_conntrack_helper.c list_for_each_entry_rcu(cur, &nf_ct_helper_expectfn_list, head) { cur 343 net/netfilter/nf_conntrack_helper.c if (!strcmp(cur->name, name)) { cur 348 net/netfilter/nf_conntrack_helper.c return found ? cur : NULL; cur 356 net/netfilter/nf_conntrack_helper.c struct nf_ct_helper_expectfn *cur; cur 359 net/netfilter/nf_conntrack_helper.c list_for_each_entry_rcu(cur, &nf_ct_helper_expectfn_list, head) { cur 360 net/netfilter/nf_conntrack_helper.c if (cur->expectfn == symbol) { cur 365 net/netfilter/nf_conntrack_helper.c return found ? cur : NULL; cur 400 net/netfilter/nf_conntrack_helper.c struct nf_conntrack_helper *cur; cur 412 net/netfilter/nf_conntrack_helper.c hlist_for_each_entry(cur, &nf_ct_helper_hash[i], hnode) { cur 413 net/netfilter/nf_conntrack_helper.c if (!strcmp(cur->name, me->name) && cur 414 net/netfilter/nf_conntrack_helper.c (cur->tuple.src.l3num == NFPROTO_UNSPEC || cur 415 net/netfilter/nf_conntrack_helper.c cur->tuple.src.l3num == me->tuple.src.l3num) && cur 416 net/netfilter/nf_conntrack_helper.c cur->tuple.dst.protonum == me->tuple.dst.protonum) { cur 425 net/netfilter/nf_conntrack_helper.c hlist_for_each_entry(cur, &nf_ct_helper_hash[h], hnode) { cur 426 net/netfilter/nf_conntrack_helper.c if (nf_ct_tuple_src_mask_cmp(&cur->tuple, &me->tuple, cur 191 net/netfilter/nfnetlink_acct.c struct nf_acct *cur, *last; cur 202 net/netfilter/nfnetlink_acct.c list_for_each_entry_rcu(cur, &net->nfnl_acct_list, head) { cur 204 net/netfilter/nfnetlink_acct.c if (cur != last) cur 210 net/netfilter/nfnetlink_acct.c if (filter && (cur->flags & filter->mask) != filter->value) cur 216 net/netfilter/nfnetlink_acct.c NFNL_MSG_ACCT_NEW, cur) < 0) { cur 217 net/netfilter/nfnetlink_acct.c cb->args[1] = (unsigned long)cur; cur 273 net/netfilter/nfnetlink_acct.c struct nf_acct *cur; cur 291 net/netfilter/nfnetlink_acct.c list_for_each_entry(cur, &net->nfnl_acct_list, head) { cur 294 net/netfilter/nfnetlink_acct.c if (strncmp(cur->name, acct_name, NFACCT_NAME_MAX)!= 0) cur 306 net/netfilter/nfnetlink_acct.c NFNL_MSG_ACCT_NEW, cur); cur 323 net/netfilter/nfnetlink_acct.c static int nfnl_acct_try_del(struct nf_acct *cur) cur 330 net/netfilter/nfnetlink_acct.c if (refcount_dec_if_one(&cur->refcnt)) { cur 332 net/netfilter/nfnetlink_acct.c list_del_rcu(&cur->head); cur 333 net/netfilter/nfnetlink_acct.c kfree_rcu(cur, rcu_head); cur 345 net/netfilter/nfnetlink_acct.c struct nf_acct *cur, *tmp; cur 350 net/netfilter/nfnetlink_acct.c list_for_each_entry_safe(cur, tmp, &net->nfnl_acct_list, head) cur 351 net/netfilter/nfnetlink_acct.c nfnl_acct_try_del(cur); cur 357 net/netfilter/nfnetlink_acct.c list_for_each_entry(cur, &net->nfnl_acct_list, head) { cur 358 net/netfilter/nfnetlink_acct.c if (strncmp(cur->name, acct_name, NFACCT_NAME_MAX) != 0) cur 361 net/netfilter/nfnetlink_acct.c ret = nfnl_acct_try_del(cur); cur 405 net/netfilter/nfnetlink_acct.c struct nf_acct *cur, *acct = NULL; cur 408 net/netfilter/nfnetlink_acct.c list_for_each_entry_rcu(cur, &net->nfnl_acct_list, head) { cur 409 net/netfilter/nfnetlink_acct.c if (strncmp(cur->name, acct_name, NFACCT_NAME_MAX)!= 0) cur 415 net/netfilter/nfnetlink_acct.c if (!refcount_inc_not_zero(&cur->refcnt)) { cur 420 net/netfilter/nfnetlink_acct.c acct = cur; cur 498 net/netfilter/nfnetlink_acct.c struct nf_acct *cur, *tmp; cur 500 net/netfilter/nfnetlink_acct.c list_for_each_entry_safe(cur, tmp, &net->nfnl_acct_list, head) { cur 501 net/netfilter/nfnetlink_acct.c list_del_rcu(&cur->head); cur 503 net/netfilter/nfnetlink_acct.c if (refcount_dec_and_test(&cur->refcnt)) cur 504 net/netfilter/nfnetlink_acct.c kfree_rcu(cur, rcu_head); cur 417 net/netfilter/nfnetlink_cthelper.c struct nf_conntrack_helper *cur, *helper = NULL; cur 435 net/netfilter/nfnetlink_cthelper.c cur = &nlcth->helper; cur 437 net/netfilter/nfnetlink_cthelper.c if (strncmp(cur->name, helper_name, NF_CT_HELPER_NAME_LEN)) cur 440 net/netfilter/nfnetlink_cthelper.c if ((tuple.src.l3num != cur->tuple.src.l3num || cur 441 net/netfilter/nfnetlink_cthelper.c tuple.dst.protonum != cur->tuple.dst.protonum)) cur 447 net/netfilter/nfnetlink_cthelper.c helper = cur; cur 578 net/netfilter/nfnetlink_cthelper.c struct nf_conntrack_helper *cur, *last; cur 584 net/netfilter/nfnetlink_cthelper.c hlist_for_each_entry_rcu(cur, cur 588 net/netfilter/nfnetlink_cthelper.c if (!(cur->flags & NF_CT_HELPER_F_USERSPACE)) cur 592 net/netfilter/nfnetlink_cthelper.c if (cur != last) cur 600 net/netfilter/nfnetlink_cthelper.c NFNL_MSG_CTHELPER_NEW, cur) < 0) { cur 601 net/netfilter/nfnetlink_cthelper.c cb->args[1] = (unsigned long)cur; cur 621 net/netfilter/nfnetlink_cthelper.c struct nf_conntrack_helper *cur; cur 650 net/netfilter/nfnetlink_cthelper.c cur = &nlcth->helper; cur 652 net/netfilter/nfnetlink_cthelper.c strncmp(cur->name, helper_name, NF_CT_HELPER_NAME_LEN)) cur 656 net/netfilter/nfnetlink_cthelper.c (tuple.src.l3num != cur->tuple.src.l3num || cur 657 net/netfilter/nfnetlink_cthelper.c tuple.dst.protonum != cur->tuple.dst.protonum)) cur 669 net/netfilter/nfnetlink_cthelper.c NFNL_MSG_CTHELPER_NEW, cur); cur 692 net/netfilter/nfnetlink_cthelper.c struct nf_conntrack_helper *cur; cur 714 net/netfilter/nfnetlink_cthelper.c cur = &nlcth->helper; cur 718 net/netfilter/nfnetlink_cthelper.c strncmp(cur->name, helper_name, NF_CT_HELPER_NAME_LEN)) cur 722 net/netfilter/nfnetlink_cthelper.c (tuple.src.l3num != cur->tuple.src.l3num || cur 723 net/netfilter/nfnetlink_cthelper.c tuple.dst.protonum != cur->tuple.dst.protonum)) cur 726 net/netfilter/nfnetlink_cthelper.c if (refcount_dec_if_one(&cur->refcnt)) { cur 728 net/netfilter/nfnetlink_cthelper.c nf_conntrack_helper_unregister(cur); cur 729 net/netfilter/nfnetlink_cthelper.c kfree(cur->expect_policy); cur 787 net/netfilter/nfnetlink_cthelper.c struct nf_conntrack_helper *cur; cur 793 net/netfilter/nfnetlink_cthelper.c cur = &nlcth->helper; cur 795 net/netfilter/nfnetlink_cthelper.c nf_conntrack_helper_unregister(cur); cur 796 net/netfilter/nfnetlink_cthelper.c kfree(cur->expect_policy); cur 210 net/netfilter/nfnetlink_cttimeout.c struct ctnl_timeout *cur, *last; cur 220 net/netfilter/nfnetlink_cttimeout.c list_for_each_entry_rcu(cur, &net->nfct_timeout_list, head) { cur 222 net/netfilter/nfnetlink_cttimeout.c if (cur != last) cur 230 net/netfilter/nfnetlink_cttimeout.c IPCTNL_MSG_TIMEOUT_NEW, cur) < 0) { cur 231 net/netfilter/nfnetlink_cttimeout.c cb->args[1] = (unsigned long)cur; cur 249 net/netfilter/nfnetlink_cttimeout.c struct ctnl_timeout *cur; cur 262 net/netfilter/nfnetlink_cttimeout.c list_for_each_entry(cur, &net->nfct_timeout_list, head) { cur 265 net/netfilter/nfnetlink_cttimeout.c if (strncmp(cur->name, name, CTNL_TIMEOUT_NAME_MAX) != 0) cur 277 net/netfilter/nfnetlink_cttimeout.c IPCTNL_MSG_TIMEOUT_NEW, cur); cur 318 net/netfilter/nfnetlink_cttimeout.c struct ctnl_timeout *cur, *tmp; cur 323 net/netfilter/nfnetlink_cttimeout.c list_for_each_entry_safe(cur, tmp, &net->nfct_timeout_list, cur 325 net/netfilter/nfnetlink_cttimeout.c ctnl_timeout_try_del(net, cur); cur 331 net/netfilter/nfnetlink_cttimeout.c list_for_each_entry(cur, &net->nfct_timeout_list, head) { cur 332 net/netfilter/nfnetlink_cttimeout.c if (strncmp(cur->name, name, CTNL_TIMEOUT_NAME_MAX) != 0) cur 335 net/netfilter/nfnetlink_cttimeout.c ret = ctnl_timeout_try_del(net, cur); cur 583 net/netfilter/nfnetlink_cttimeout.c struct ctnl_timeout *cur, *tmp; cur 588 net/netfilter/nfnetlink_cttimeout.c list_for_each_entry_safe(cur, tmp, &net->nfct_timeout_list, head) { cur 589 net/netfilter/nfnetlink_cttimeout.c list_del_rcu(&cur->head); cur 591 net/netfilter/nfnetlink_cttimeout.c if (refcount_dec_and_test(&cur->refcnt)) cur 592 net/netfilter/nfnetlink_cttimeout.c kfree_rcu(cur, rcu_head); cur 54 net/netfilter/x_tables.c unsigned int cur; /* number of used slots in compat_tab[] */ cur 652 net/netfilter/x_tables.c if (xp->cur >= xp->number) cur 655 net/netfilter/x_tables.c if (xp->cur) cur 656 net/netfilter/x_tables.c delta += xp->compat_tab[xp->cur - 1].delta; cur 657 net/netfilter/x_tables.c xp->compat_tab[xp->cur].offset = offset; cur 658 net/netfilter/x_tables.c xp->compat_tab[xp->cur].delta = delta; cur 659 net/netfilter/x_tables.c xp->cur++; cur 672 net/netfilter/x_tables.c xt[af].cur = 0; cur 680 net/netfilter/x_tables.c int mid, left = 0, right = xt[af].cur - 1; cur 716 net/netfilter/x_tables.c xt[af].cur = 0; cur 314 net/rds/ib_rdma.c struct llist_node *cur = NULL; cur 318 net/rds/ib_rdma.c cur = &ibmr->llnode; cur 319 net/rds/ib_rdma.c *next = cur; cur 320 net/rds/ib_rdma.c next = &cur->next; cur 323 net/rds/ib_rdma.c *nodes_tail = cur; cur 110 net/rfkill/core.c bool cur, sav; cur 372 net/rfkill/core.c rfkill_global_states[type].cur = blocked; cur 377 net/rfkill/core.c rfkill_global_states[i].cur = blocked; cur 453 net/rfkill/core.c rfkill_global_states[i].sav = rfkill_global_states[i].cur; cur 454 net/rfkill/core.c rfkill_global_states[i].cur = true; cur 521 net/rfkill/core.c return rfkill_global_states[type].cur; cur 875 net/rfkill/core.c bool cur; cur 880 net/rfkill/core.c cur = !!(rfkill->state & RFKILL_BLOCK_SW); cur 881 net/rfkill/core.c rfkill_set_block(rfkill, cur); cur 992 net/rfkill/core.c bool cur; cur 997 net/rfkill/core.c cur = rfkill_global_states[rfkill->type].cur; cur 998 net/rfkill/core.c rfkill_set_block(rfkill, cur); cur 74 net/smc/smc_core.c struct smc_connection *cur = rb_entry(*link, cur 78 net/smc/smc_core.c if (cur->alert_token_local > token) cur 265 net/smc/smc_core.h struct smc_connection *cur = rb_entry(node, cur 268 net/smc/smc_core.h if (cur->alert_token_local > token) { cur 271 net/smc/smc_core.h if (cur->alert_token_local < token) { cur 274 net/smc/smc_core.h res = cur; cur 987 net/sunrpc/xdr.c unsigned int cur = xdr_stream_pos(xdr); cur 995 net/sunrpc/xdr.c if (iov->iov_len > cur) { cur 996 net/sunrpc/xdr.c offset = iov->iov_len - cur; cur 999 net/sunrpc/xdr.c xdr->nwords = XDR_QUADLEN(buf->len - cur); cur 1013 net/sunrpc/xdr.c xdr->nwords = XDR_QUADLEN(buf->len - cur); cur 1165 net/sunrpc/xdr.c size_t cur; cur 1169 net/sunrpc/xdr.c cur = min_t(size_t, buf->tail[0].iov_len, trim); cur 1170 net/sunrpc/xdr.c buf->tail[0].iov_len -= cur; cur 1171 net/sunrpc/xdr.c trim -= cur; cur 1177 net/sunrpc/xdr.c cur = min_t(unsigned int, buf->page_len, trim); cur 1178 net/sunrpc/xdr.c buf->page_len -= cur; cur 1179 net/sunrpc/xdr.c trim -= cur; cur 1185 net/sunrpc/xdr.c cur = min_t(size_t, buf->head[0].iov_len, trim); cur 1186 net/sunrpc/xdr.c buf->head[0].iov_len -= cur; cur 1187 net/sunrpc/xdr.c trim -= cur; cur 228 net/sunrpc/xprtmultipath.c const struct rpc_xprt *cur) cur 234 net/sunrpc/xprtmultipath.c if (cur == pos) cur 278 net/sunrpc/xprtmultipath.c const struct rpc_xprt *cur) cur 284 net/sunrpc/xprtmultipath.c if (cur == prev) cur 319 net/sunrpc/xprtmultipath.c const struct rpc_xprt *cur) cur 323 net/sunrpc/xprtmultipath.c ret = xprt_switch_find_next_entry(head, cur); cur 331 net/sunrpc/xprtmultipath.c const struct rpc_xprt *cur) cur 340 net/sunrpc/xprtmultipath.c xprt = __xprt_switch_find_next_entry_roundrobin(head, cur); cur 349 net/sunrpc/xprtmultipath.c cur = xprt; cur 363 net/sunrpc/xprtmultipath.c const struct rpc_xprt *cur) cur 365 net/sunrpc/xprtmultipath.c return xprt_switch_find_next_entry(&xps->xps_xprt_list, cur); cur 360 net/tipc/monitor.c struct tipc_peer *cur, *prev, *p; cur 374 net/tipc/monitor.c list_for_each_entry(cur, &self->list, list) { cur 375 net/tipc/monitor.c if ((addr > prev->addr) && (addr < cur->addr)) cur 377 net/tipc/monitor.c if (((addr < cur->addr) || (addr > prev->addr)) && cur 378 net/tipc/monitor.c (prev->addr > cur->addr)) cur 380 net/tipc/monitor.c prev = cur; cur 382 net/tipc/monitor.c list_add_tail(&p->list, &cur->list); cur 553 scripts/genksyms/genksyms.c struct string_list *cur; cur 556 scripts/genksyms/genksyms.c cur = *(b++); cur 557 scripts/genksyms/genksyms.c switch (cur->tag) { cur 560 scripts/genksyms/genksyms.c fprintf(debugfile, "%s ", cur->string); cur 561 scripts/genksyms/genksyms.c crc = partial_crc32(cur->string, crc); cur 567 scripts/genksyms/genksyms.c subsym = find_symbol(cur->string, cur->tag, 0); cur 571 scripts/genksyms/genksyms.c fprintf(debugfile, "%s ", cur->string); cur 572 scripts/genksyms/genksyms.c crc = partial_crc32(cur->string, crc); cur 584 scripts/genksyms/genksyms.c subsym = find_symbol(cur->string, cur->tag, 0); cur 589 scripts/genksyms/genksyms.c symbol_types[cur->tag].name, cur 590 scripts/genksyms/genksyms.c cur->string); cur 592 scripts/genksyms/genksyms.c (symbol_types[cur->tag].name), cur 593 scripts/genksyms/genksyms.c mk_node(cur->string), cur 598 scripts/genksyms/genksyms.c add_symbol(cur->string, cur->tag, n, 0); cur 603 scripts/genksyms/genksyms.c symbol_types[cur->tag].name, cur 604 scripts/genksyms/genksyms.c cur->string); cur 607 scripts/genksyms/genksyms.c crc = partial_crc32(symbol_types[cur->tag].name, cur 610 scripts/genksyms/genksyms.c crc = partial_crc32(cur->string, crc); cur 594 scripts/kconfig/nconf.c ITEM *cur; cur 597 scripts/kconfig/nconf.c cur = current_item(curses_menu); cur 598 scripts/kconfig/nconf.c if (cur == NULL) cur 600 scripts/kconfig/nconf.c mcur = (struct mitem *) item_userptr(cur); cur 611 scripts/kconfig/nconf.c ITEM *cur; cur 614 scripts/kconfig/nconf.c cur = current_item(curses_menu); cur 615 scripts/kconfig/nconf.c if (!cur) cur 617 scripts/kconfig/nconf.c mcur = (struct mitem *) item_userptr(cur); cur 192 security/loadpin/loadpin.c char *cur; cur 205 security/loadpin/loadpin.c cur = exclude_read_files[i]; cur 206 security/loadpin/loadpin.c if (!cur) cur 208 security/loadpin/loadpin.c if (*cur == '\0') cur 212 security/loadpin/loadpin.c if (strcmp(cur, kernel_read_file_str[j]) == 0) { cur 70 security/selinux/ss/avtab.c struct avtab_node *prev, struct avtab_node *cur, cur 109 security/selinux/ss/avtab.c struct avtab_node *prev, *cur, *newnode; cur 116 security/selinux/ss/avtab.c for (prev = NULL, cur = h->htable[hvalue]; cur 117 security/selinux/ss/avtab.c cur; cur 118 security/selinux/ss/avtab.c prev = cur, cur = cur->next) { cur 119 security/selinux/ss/avtab.c if (key->source_type == cur->key.source_type && cur 120 security/selinux/ss/avtab.c key->target_type == cur->key.target_type && cur 121 security/selinux/ss/avtab.c key->target_class == cur->key.target_class && cur 122 security/selinux/ss/avtab.c (specified & cur->key.specified)) { cur 128 security/selinux/ss/avtab.c if (key->source_type < cur->key.source_type) cur 130 security/selinux/ss/avtab.c if (key->source_type == cur->key.source_type && cur 131 security/selinux/ss/avtab.c key->target_type < cur->key.target_type) cur 133 security/selinux/ss/avtab.c if (key->source_type == cur->key.source_type && cur 134 security/selinux/ss/avtab.c key->target_type == cur->key.target_type && cur 135 security/selinux/ss/avtab.c key->target_class < cur->key.target_class) cur 139 security/selinux/ss/avtab.c newnode = avtab_insert_node(h, hvalue, prev, cur, key, datum); cur 154 security/selinux/ss/avtab.c struct avtab_node *prev, *cur; cur 160 security/selinux/ss/avtab.c for (prev = NULL, cur = h->htable[hvalue]; cur 161 security/selinux/ss/avtab.c cur; cur 162 security/selinux/ss/avtab.c prev = cur, cur = cur->next) { cur 163 security/selinux/ss/avtab.c if (key->source_type == cur->key.source_type && cur 164 security/selinux/ss/avtab.c key->target_type == cur->key.target_type && cur 165 security/selinux/ss/avtab.c key->target_class == cur->key.target_class && cur 166 security/selinux/ss/avtab.c (specified & cur->key.specified)) cur 168 security/selinux/ss/avtab.c if (key->source_type < cur->key.source_type) cur 170 security/selinux/ss/avtab.c if (key->source_type == cur->key.source_type && cur 171 security/selinux/ss/avtab.c key->target_type < cur->key.target_type) cur 173 security/selinux/ss/avtab.c if (key->source_type == cur->key.source_type && cur 174 security/selinux/ss/avtab.c key->target_type == cur->key.target_type && cur 175 security/selinux/ss/avtab.c key->target_class < cur->key.target_class) cur 178 security/selinux/ss/avtab.c return avtab_insert_node(h, hvalue, prev, cur, key, datum); cur 184 security/selinux/ss/avtab.c struct avtab_node *cur; cur 191 security/selinux/ss/avtab.c for (cur = h->htable[hvalue]; cur; cur 192 security/selinux/ss/avtab.c cur = cur->next) { cur 193 security/selinux/ss/avtab.c if (key->source_type == cur->key.source_type && cur 194 security/selinux/ss/avtab.c key->target_type == cur->key.target_type && cur 195 security/selinux/ss/avtab.c key->target_class == cur->key.target_class && cur 196 security/selinux/ss/avtab.c (specified & cur->key.specified)) cur 197 security/selinux/ss/avtab.c return &cur->datum; cur 199 security/selinux/ss/avtab.c if (key->source_type < cur->key.source_type) cur 201 security/selinux/ss/avtab.c if (key->source_type == cur->key.source_type && cur 202 security/selinux/ss/avtab.c key->target_type < cur->key.target_type) cur 204 security/selinux/ss/avtab.c if (key->source_type == cur->key.source_type && cur 205 security/selinux/ss/avtab.c key->target_type == cur->key.target_type && cur 206 security/selinux/ss/avtab.c key->target_class < cur->key.target_class) cur 220 security/selinux/ss/avtab.c struct avtab_node *cur; cur 227 security/selinux/ss/avtab.c for (cur = h->htable[hvalue]; cur; cur 228 security/selinux/ss/avtab.c cur = cur->next) { cur 229 security/selinux/ss/avtab.c if (key->source_type == cur->key.source_type && cur 230 security/selinux/ss/avtab.c key->target_type == cur->key.target_type && cur 231 security/selinux/ss/avtab.c key->target_class == cur->key.target_class && cur 232 security/selinux/ss/avtab.c (specified & cur->key.specified)) cur 233 security/selinux/ss/avtab.c return cur; cur 235 security/selinux/ss/avtab.c if (key->source_type < cur->key.source_type) cur 237 security/selinux/ss/avtab.c if (key->source_type == cur->key.source_type && cur 238 security/selinux/ss/avtab.c key->target_type < cur->key.target_type) cur 240 security/selinux/ss/avtab.c if (key->source_type == cur->key.source_type && cur 241 security/selinux/ss/avtab.c key->target_type == cur->key.target_type && cur 242 security/selinux/ss/avtab.c key->target_class < cur->key.target_class) cur 251 security/selinux/ss/avtab.c struct avtab_node *cur; cur 257 security/selinux/ss/avtab.c for (cur = node->next; cur; cur = cur->next) { cur 258 security/selinux/ss/avtab.c if (node->key.source_type == cur->key.source_type && cur 259 security/selinux/ss/avtab.c node->key.target_type == cur->key.target_type && cur 260 security/selinux/ss/avtab.c node->key.target_class == cur->key.target_class && cur 261 security/selinux/ss/avtab.c (specified & cur->key.specified)) cur 262 security/selinux/ss/avtab.c return cur; cur 264 security/selinux/ss/avtab.c if (node->key.source_type < cur->key.source_type) cur 266 security/selinux/ss/avtab.c if (node->key.source_type == cur->key.source_type && cur 267 security/selinux/ss/avtab.c node->key.target_type < cur->key.target_type) cur 269 security/selinux/ss/avtab.c if (node->key.source_type == cur->key.source_type && cur 270 security/selinux/ss/avtab.c node->key.target_type == cur->key.target_type && cur 271 security/selinux/ss/avtab.c node->key.target_class < cur->key.target_class) cur 280 security/selinux/ss/avtab.c struct avtab_node *cur, *temp; cur 286 security/selinux/ss/avtab.c cur = h->htable[i]; cur 287 security/selinux/ss/avtab.c while (cur) { cur 288 security/selinux/ss/avtab.c temp = cur; cur 289 security/selinux/ss/avtab.c cur = cur->next; cur 348 security/selinux/ss/avtab.c struct avtab_node *cur; cur 354 security/selinux/ss/avtab.c cur = h->htable[i]; cur 355 security/selinux/ss/avtab.c if (cur) { cur 358 security/selinux/ss/avtab.c while (cur) { cur 360 security/selinux/ss/avtab.c cur = cur->next; cur 599 security/selinux/ss/avtab.c int avtab_write_item(struct policydb *p, struct avtab_node *cur, void *fp) cur 602 security/selinux/ss/avtab.c __le32 buf32[ARRAY_SIZE(cur->datum.u.xperms->perms.p)]; cur 606 security/selinux/ss/avtab.c buf16[0] = cpu_to_le16(cur->key.source_type); cur 607 security/selinux/ss/avtab.c buf16[1] = cpu_to_le16(cur->key.target_type); cur 608 security/selinux/ss/avtab.c buf16[2] = cpu_to_le16(cur->key.target_class); cur 609 security/selinux/ss/avtab.c buf16[3] = cpu_to_le16(cur->key.specified); cur 614 security/selinux/ss/avtab.c if (cur->key.specified & AVTAB_XPERMS) { cur 615 security/selinux/ss/avtab.c rc = put_entry(&cur->datum.u.xperms->specified, sizeof(u8), 1, fp); cur 618 security/selinux/ss/avtab.c rc = put_entry(&cur->datum.u.xperms->driver, sizeof(u8), 1, fp); cur 621 security/selinux/ss/avtab.c for (i = 0; i < ARRAY_SIZE(cur->datum.u.xperms->perms.p); i++) cur 622 security/selinux/ss/avtab.c buf32[i] = cpu_to_le32(cur->datum.u.xperms->perms.p[i]); cur 624 security/selinux/ss/avtab.c ARRAY_SIZE(cur->datum.u.xperms->perms.p), fp); cur 626 security/selinux/ss/avtab.c buf32[0] = cpu_to_le32(cur->datum.u.data); cur 638 security/selinux/ss/avtab.c struct avtab_node *cur; cur 647 security/selinux/ss/avtab.c for (cur = a->htable[i]; cur; cur 648 security/selinux/ss/avtab.c cur = cur->next) { cur 649 security/selinux/ss/avtab.c rc = avtab_write_item(p, cur, fp); cur 103 security/selinux/ss/avtab.h int avtab_write_item(struct policydb *p, struct avtab_node *cur, void *fp); cur 27 security/selinux/ss/conditional.c struct cond_expr *cur; cur 31 security/selinux/ss/conditional.c for (cur = expr; cur; cur = cur->next) { cur 32 security/selinux/ss/conditional.c switch (cur->expr_type) { cur 37 security/selinux/ss/conditional.c s[sp] = p->bool_val_to_struct[cur->bool - 1]->state; cur 91 security/selinux/ss/conditional.c struct cond_av_list *cur; cur 99 security/selinux/ss/conditional.c for (cur = node->true_list; cur; cur = cur->next) { cur 101 security/selinux/ss/conditional.c cur->node->key.specified &= ~AVTAB_ENABLED; cur 103 security/selinux/ss/conditional.c cur->node->key.specified |= AVTAB_ENABLED; cur 106 security/selinux/ss/conditional.c for (cur = node->false_list; cur; cur = cur->next) { cur 109 security/selinux/ss/conditional.c cur->node->key.specified &= ~AVTAB_ENABLED; cur 111 security/selinux/ss/conditional.c cur->node->key.specified |= AVTAB_ENABLED; cur 133 security/selinux/ss/conditional.c struct cond_av_list *cur, *next; cur 134 security/selinux/ss/conditional.c for (cur = list; cur; cur = next) { cur 135 security/selinux/ss/conditional.c next = cur->next; cur 137 security/selinux/ss/conditional.c kfree(cur); cur 156 security/selinux/ss/conditional.c struct cond_node *next, *cur; cur 161 security/selinux/ss/conditional.c for (cur = list; cur; cur = next) { cur 162 security/selinux/ss/conditional.c next = cur->next; cur 163 security/selinux/ss/conditional.c cond_node_destroy(cur); cur 272 security/selinux/ss/conditional.c struct cond_av_list *other = data->other, *list, *cur; cur 303 security/selinux/ss/conditional.c for (cur = other; cur; cur = cur->next) { cur 304 security/selinux/ss/conditional.c if (cur->node == node_ptr) { cur 591 security/selinux/ss/conditional.c struct cond_node *cur; cur 597 security/selinux/ss/conditional.c for (cur = list; cur != NULL; cur = cur->next) cur 604 security/selinux/ss/conditional.c for (cur = list; cur != NULL; cur = cur->next) { cur 605 security/selinux/ss/conditional.c rc = cond_write_node(p, cur, fp); cur 45 security/selinux/ss/hashtab.c struct hashtab_node *prev, *cur, *newnode; cur 54 security/selinux/ss/hashtab.c cur = h->htable[hvalue]; cur 55 security/selinux/ss/hashtab.c while (cur && h->keycmp(h, key, cur->key) > 0) { cur 56 security/selinux/ss/hashtab.c prev = cur; cur 57 security/selinux/ss/hashtab.c cur = cur->next; cur 60 security/selinux/ss/hashtab.c if (cur && (h->keycmp(h, key, cur->key) == 0)) cur 83 security/selinux/ss/hashtab.c struct hashtab_node *cur; cur 89 security/selinux/ss/hashtab.c cur = h->htable[hvalue]; cur 90 security/selinux/ss/hashtab.c while (cur && h->keycmp(h, key, cur->key) > 0) cur 91 security/selinux/ss/hashtab.c cur = cur->next; cur 93 security/selinux/ss/hashtab.c if (!cur || (h->keycmp(h, key, cur->key) != 0)) cur 96 security/selinux/ss/hashtab.c return cur->datum; cur 102 security/selinux/ss/hashtab.c struct hashtab_node *cur, *temp; cur 108 security/selinux/ss/hashtab.c cur = h->htable[i]; cur 109 security/selinux/ss/hashtab.c while (cur) { cur 110 security/selinux/ss/hashtab.c temp = cur; cur 111 security/selinux/ss/hashtab.c cur = cur->next; cur 129 security/selinux/ss/hashtab.c struct hashtab_node *cur; cur 135 security/selinux/ss/hashtab.c cur = h->htable[i]; cur 136 security/selinux/ss/hashtab.c while (cur) { cur 137 security/selinux/ss/hashtab.c ret = apply(cur->key, cur->datum, args); cur 140 security/selinux/ss/hashtab.c cur = cur->next; cur 150 security/selinux/ss/hashtab.c struct hashtab_node *cur; cur 155 security/selinux/ss/hashtab.c cur = h->htable[i]; cur 156 security/selinux/ss/hashtab.c if (cur) { cur 159 security/selinux/ss/hashtab.c while (cur) { cur 161 security/selinux/ss/hashtab.c cur = cur->next; cur 2858 security/selinux/ss/services.c struct cond_node *cur; cur 2886 security/selinux/ss/services.c for (cur = policydb->cond_list; cur; cur = cur->next) { cur 2887 security/selinux/ss/services.c rc = evaluate_cond_node(policydb, cur); cur 2933 security/selinux/ss/services.c struct cond_node *cur; cur 2943 security/selinux/ss/services.c for (cur = policydb->cond_list; cur; cur = cur->next) { cur 2944 security/selinux/ss/services.c rc = evaluate_cond_node(policydb, cur); cur 204 sound/core/seq/seq_clientmgr.c res->cur += num; cur 205 sound/core/seq/seq_clientmgr.c if (res->cur > res->peak) cur 206 sound/core/seq/seq_clientmgr.c res->peak = res->cur; cur 211 sound/core/seq/seq_clientmgr.c res->cur -= num; cur 1180 sound/core/seq/seq_clientmgr.c info->cur_clients = client_usage.cur; cur 2452 sound/core/seq/seq_clientmgr.c snd_iprintf(buffer, " cur clients : %d\n", client_usage.cur); cur 63 sound/core/seq/seq_clientmgr.h int cur; cur 134 sound/core/seq/seq_prioq.c struct snd_seq_event_cell *cur, *prev; cur 165 sound/core/seq/seq_prioq.c cur = f->head; /* cursor */ cur 168 sound/core/seq/seq_prioq.c while (cur != NULL) { cur 170 sound/core/seq/seq_prioq.c int rel = compare_timestamp_rel(&cell->event, &cur->event); cur 179 sound/core/seq/seq_prioq.c prev = cur; cur 180 sound/core/seq/seq_prioq.c cur = cur->next; cur 191 sound/core/seq/seq_prioq.c cell->next = cur; cur 193 sound/core/seq/seq_prioq.c if (f->head == cur) /* this is the first cell, set head to it */ cur 195 sound/core/seq/seq_prioq.c if (cur == NULL) /* reached end of the list */ cur 1166 sound/isa/es18xx.c unsigned char cur = inb(chip->port + ES18XX_PM); cur 1168 sound/isa/es18xx.c if ((cur & mask) == val) cur 1170 sound/isa/es18xx.c outb((cur & ~mask) | val, chip->port + ES18XX_PM); cur 4120 sound/pci/hda/hda_generic.c static bool dyn_adc_pcm_resetup(struct hda_codec *codec, int cur); cur 5609 sound/pci/hda/hda_generic.c static bool dyn_adc_pcm_resetup(struct hda_codec *codec, int cur) cur 5612 sound/pci/hda/hda_generic.c hda_nid_t new_adc = spec->adc_nids[spec->dyn_adc_idx[cur]]; cur 1260 sound/pci/pcxhr/pcxhr.c int cur = rmh.stat[0]; cur 1271 sound/pci/pcxhr/pcxhr.c cur = 100 - (100 * cur) / ref; cur 1272 sound/pci/pcxhr/pcxhr.c snd_iprintf(buffer, "cpu load %d%%\n", cur); cur 700 sound/soc/bcm/cygnus-pcm.c unsigned int res = 0, cur = 0, base = 0; cur 711 sound/soc/bcm/cygnus-pcm.c cur = readl(aio->cygaud->audio + p_rbuf->rdaddr); cur 713 sound/soc/bcm/cygnus-pcm.c cur = readl(aio->cygaud->audio + p_rbuf->wraddr); cur 721 sound/soc/bcm/cygnus-pcm.c res = (cur & 0x7fffffff) - (base & 0x7fffffff); cur 2789 sound/soc/codecs/arizona.c const __be32 *cur; cur 2796 sound/soc/codecs/arizona.c of_property_for_each_u32(np, "wlf,inmode", prop, cur, val) { cur 2805 sound/soc/codecs/arizona.c of_property_for_each_u32(np, "wlf,dmic-ref", prop, cur, val) { cur 2814 sound/soc/codecs/arizona.c of_property_for_each_u32(np, "wlf,out-mono", prop, cur, val) { cur 2823 sound/soc/codecs/arizona.c of_property_for_each_u32(np, "wlf,max-channels-clocked", prop, cur, val) { cur 2832 sound/soc/codecs/arizona.c of_property_for_each_u32(np, "wlf,out-volume-limit", prop, cur, val) { cur 63 sound/soc/codecs/cs47l92.c unsigned int ep_sel, mux, change, cur; cur 82 sound/soc/codecs/cs47l92.c ret = regmap_read(madera->regmap, MADERA_OUTPUT_ENABLES_1, &cur); cur 108 sound/soc/codecs/cs47l92.c MADERA_OUT3L_ENA | MADERA_OUT3R_ENA, cur); cur 113 sound/soc/codecs/cs47l92.c if (cur & (MADERA_OUT3L_ENA | MADERA_OUT3R_ENA)) cur 963 sound/soc/codecs/madera.c unsigned int cur, new; cur 975 sound/soc/codecs/madera.c ret = regmap_read(dsp->regmap, dsp->base, &cur); cur 982 sound/soc/codecs/madera.c cur &= MADERA_DSP_RATE_MASK; cur 986 sound/soc/codecs/madera.c if (new == cur) { cur 2957 sound/soc/codecs/madera.c unsigned int reg, cur, tar; cur 3004 sound/soc/codecs/madera.c base + MADERA_AIF_RATE_CTRL, &cur); cur 3010 sound/soc/codecs/madera.c if ((cur & MADERA_AIF1_RATE_MASK) == (tar & MADERA_AIF1_RATE_MASK)) cur 746 sound/soc/uniphier/aio-core.c int cur, diff, slope = 0, fs; cur 751 sound/soc/uniphier/aio-core.c cur = aio_port_get_volume(sub); cur 752 sound/soc/uniphier/aio-core.c diff = abs(vol - cur); cur 763 sound/soc/uniphier/aio-core.c if (cur < vol) cur 51 sound/synth/emux/soundfont.c static void add_preset(struct snd_sf_list *sflist, struct snd_sf_zone *cur); cur 1138 sound/synth/emux/soundfont.c struct snd_sf_zone *cur; cur 1145 sound/synth/emux/soundfont.c for (cur = sf->zones; cur; cur = cur->next) { cur 1146 sound/synth/emux/soundfont.c if (! cur->mapped && cur->sample == NULL) { cur 1148 sound/synth/emux/soundfont.c cur->sample = set_sample(sf, &cur->v); cur 1149 sound/synth/emux/soundfont.c if (cur->sample == NULL) cur 1153 sound/synth/emux/soundfont.c add_preset(sflist, cur); cur 1163 sound/synth/emux/soundfont.c add_preset(struct snd_sf_list *sflist, struct snd_sf_zone *cur) cur 1168 sound/synth/emux/soundfont.c zone = search_first_zone(sflist, cur->bank, cur->instr, cur->v.low); cur 1169 sound/synth/emux/soundfont.c if (zone && zone->v.sf_id != cur->v.sf_id) { cur 1174 sound/synth/emux/soundfont.c if (p->counter > cur->counter) cur 1184 sound/synth/emux/soundfont.c if ((index = get_index(cur->bank, cur->instr, cur->v.low)) < 0) cur 1186 sound/synth/emux/soundfont.c cur->next_zone = zone; /* zone link */ cur 1187 sound/synth/emux/soundfont.c cur->next_instr = sflist->presets[index]; /* preset table link */ cur 1188 sound/synth/emux/soundfont.c sflist->presets[index] = cur; cur 262 sound/usb/clock.c int ret, i, cur; cur 280 sound/usb/clock.c cur = ret; cur 291 sound/usb/clock.c if (i == cur) cur 356 sound/usb/clock.c int ret, i, cur; cur 374 sound/usb/clock.c cur = ret; cur 385 sound/usb/clock.c if (i == cur) cur 131 tools/bpf/bpftool/cfg.c static bool cfg_partition_funcs(struct cfg *cfg, struct bpf_insn *cur, cur 136 tools/bpf/bpftool/cfg.c func = cfg_append_func(cfg, cur); cur 140 tools/bpf/bpftool/cfg.c for (; cur < end; cur++) { cur 141 tools/bpf/bpftool/cfg.c if (cur->code != (BPF_JMP | BPF_CALL)) cur 143 tools/bpf/bpftool/cfg.c if (cur->src_reg != BPF_PSEUDO_CALL) cur 145 tools/bpf/bpftool/cfg.c func = cfg_append_func(cfg, cur + cur->off + 1); cur 167 tools/bpf/bpftool/cfg.c struct bpf_insn *cur, *end; cur 170 tools/bpf/bpftool/cfg.c cur = func->start; cur 173 tools/bpf/bpftool/cfg.c bb = func_append_bb(func, cur); cur 177 tools/bpf/bpftool/cfg.c for (; cur <= end; cur++) { cur 178 tools/bpf/bpftool/cfg.c if (is_jmp_insn(cur->code)) { cur 179 tools/bpf/bpftool/cfg.c u8 opcode = BPF_OP(cur->code); cur 184 tools/bpf/bpftool/cfg.c bb = func_append_bb(func, cur + cur->off + 1); cur 189 tools/bpf/bpftool/cfg.c bb = func_append_bb(func, cur + 1); cur 365 tools/bpf/bpftool/xlated_dumper.c struct bpf_insn *cur = insn_start; cur 367 tools/bpf/bpftool/xlated_dumper.c for (; cur <= insn_end; cur++) { cur 368 tools/bpf/bpftool/xlated_dumper.c printf("% 4d: ", (int)(cur - insn_start + start_idx)); cur 369 tools/bpf/bpftool/xlated_dumper.c print_bpf_insn(&cbs, cur, true); cur 370 tools/bpf/bpftool/xlated_dumper.c if (cur != insn_end) cur 35 tools/lib/bitmap.c int cur, rbot, rtop; cur 39 tools/lib/bitmap.c rbot = cur = find_first_bit(bitmap, nbits); cur 40 tools/lib/bitmap.c while (cur < nbits) { cur 41 tools/lib/bitmap.c rtop = cur; cur 42 tools/lib/bitmap.c cur = find_next_bit(bitmap, nbits, cur + 1); cur 43 tools/lib/bitmap.c if (cur < nbits && cur <= rtop + 1) cur 55 tools/lib/bitmap.c rbot = cur; cur 91 tools/lib/bpf/hashmap.c struct hashmap_entry *cur, *tmp; cur 105 tools/lib/bpf/hashmap.c hashmap__for_each_entry_safe(map, cur, tmp, bkt) { cur 106 tools/lib/bpf/hashmap.c h = hash_bits(map->hash_fn(cur->key, map->ctx), new_cap_bits); cur 107 tools/lib/bpf/hashmap.c hashmap_add_entry(&new_buckets[h], cur); cur 123 tools/lib/bpf/hashmap.c struct hashmap_entry *cur, **prev_ptr; cur 128 tools/lib/bpf/hashmap.c for (prev_ptr = &map->buckets[hash], cur = *prev_ptr; cur 129 tools/lib/bpf/hashmap.c cur; cur 130 tools/lib/bpf/hashmap.c prev_ptr = &cur->next, cur = cur->next) { cur 131 tools/lib/bpf/hashmap.c if (map->equal_fn(cur->key, key, map->ctx)) { cur 134 tools/lib/bpf/hashmap.c *entry = cur; cur 138 tools/lib/bpf/hashmap.h #define hashmap__for_each_entry(map, cur, bkt) \ cur 140 tools/lib/bpf/hashmap.h for (cur = map->buckets[bkt]; cur; cur = cur->next) cur 150 tools/lib/bpf/hashmap.h #define hashmap__for_each_entry_safe(map, cur, tmp, bkt) \ cur 152 tools/lib/bpf/hashmap.h for (cur = map->buckets[bkt]; \ cur 153 tools/lib/bpf/hashmap.h cur && ({tmp = cur->next; true; }); \ cur 154 tools/lib/bpf/hashmap.h cur = tmp) cur 162 tools/lib/bpf/hashmap.h #define hashmap__for_each_key_entry(map, cur, _key) \ cur 163 tools/lib/bpf/hashmap.h for (cur = ({ size_t bkt = hash_bits(map->hash_fn((_key), map->ctx),\ cur 166 tools/lib/bpf/hashmap.h cur; \ cur 167 tools/lib/bpf/hashmap.h cur = cur->next) \ cur 168 tools/lib/bpf/hashmap.h if (map->equal_fn(cur->key, (_key), map->ctx)) cur 170 tools/lib/bpf/hashmap.h #define hashmap__for_each_key_entry_safe(map, cur, tmp, _key) \ cur 171 tools/lib/bpf/hashmap.h for (cur = ({ size_t bkt = hash_bits(map->hash_fn((_key), map->ctx),\ cur 173 tools/lib/bpf/hashmap.h cur = map->buckets ? map->buckets[bkt] : NULL; }); \ cur 174 tools/lib/bpf/hashmap.h cur && ({ tmp = cur->next; true; }); \ cur 175 tools/lib/bpf/hashmap.h cur = tmp) \ cur 176 tools/lib/bpf/hashmap.h if (map->equal_fn(cur->key, (_key), map->ctx)) cur 703 tools/perf/util/cs-etm-decoder/cs-etm-decoder.c ocsd_datapath_resp_t cur = OCSD_RESP_CONT; cur 710 tools/perf/util/cs-etm-decoder/cs-etm-decoder.c cur = ocsd_dt_process_data(decoder->dcd_tree, cur 717 tools/perf/util/cs-etm-decoder/cs-etm-decoder.c cur = ocsd_dt_process_data(decoder->dcd_tree, cur 734 tools/perf/util/cs-etm-decoder/cs-etm-decoder.c if (OCSD_DATA_RESP_IS_WAIT(cur)) cur 737 tools/perf/util/cs-etm-decoder/cs-etm-decoder.c prev_return = cur; cur 740 tools/perf/util/cs-etm-decoder/cs-etm-decoder.c decoder->prev_return = cur; cur 193 tools/perf/util/genelf_debug.c ubyte cur = data & 0x7F; cur 196 tools/perf/util/genelf_debug.c cur |= 0x80; cur 197 tools/perf/util/genelf_debug.c buffer_ext_add(be, &cur, 1); cur 207 tools/perf/util/genelf_debug.c ubyte cur = data & 0x7F; cur 211 tools/perf/util/genelf_debug.c if ((data == 0 && !(cur & 0x40)) || cur 212 tools/perf/util/genelf_debug.c (data == -1l && (cur & 0x40))) cur 215 tools/perf/util/genelf_debug.c cur |= 0x80; cur 216 tools/perf/util/genelf_debug.c buffer_ext_add(be, &cur, 1); cur 84 tools/perf/util/strfilter.c struct strfilter_node root, *cur, *last_op; cur 91 tools/perf/util/strfilter.c last_op = cur = &root; cur 97 tools/perf/util/strfilter.c if (!cur->r || !last_op->r) cur 99 tools/perf/util/strfilter.c cur = strfilter_node__alloc(OP_and, last_op->r, NULL); cur 100 tools/perf/util/strfilter.c if (!cur) cur 102 tools/perf/util/strfilter.c last_op->r = cur; cur 103 tools/perf/util/strfilter.c last_op = cur; cur 106 tools/perf/util/strfilter.c if (!cur->r || !root.r) cur 108 tools/perf/util/strfilter.c cur = strfilter_node__alloc(OP_or, root.r, NULL); cur 109 tools/perf/util/strfilter.c if (!cur) cur 111 tools/perf/util/strfilter.c root.r = cur; cur 112 tools/perf/util/strfilter.c last_op = cur; cur 115 tools/perf/util/strfilter.c if (cur->r) cur 117 tools/perf/util/strfilter.c cur->r = strfilter_node__alloc(OP_not, NULL, NULL); cur 118 tools/perf/util/strfilter.c if (!cur->r) cur 120 tools/perf/util/strfilter.c cur = cur->r; cur 123 tools/perf/util/strfilter.c if (cur->r) cur 125 tools/perf/util/strfilter.c cur->r = strfilter_node__new(s + 1, &s); cur 128 tools/perf/util/strfilter.c if (!cur->r || *s != ')') cur 133 tools/perf/util/strfilter.c if (cur->r) cur 135 tools/perf/util/strfilter.c cur->r = strfilter_node__alloc(NULL, NULL, NULL); cur 136 tools/perf/util/strfilter.c if (!cur->r) cur 138 tools/perf/util/strfilter.c cur->r->p = strndup(s, e - s); cur 139 tools/perf/util/strfilter.c if (!cur->r->p) cur 144 tools/perf/util/strfilter.c if (!cur->r) cur 114 tools/perf/util/unwind-libunwind-local.c u8 *cur = *p; cur 122 tools/perf/util/unwind-libunwind-local.c *val = dw_read(cur, unsigned long, end); cur 132 tools/perf/util/unwind-libunwind-local.c *val = (unsigned long) cur; cur 143 tools/perf/util/unwind-libunwind-local.c *val += dw_read(cur, s32, end); cur 146 tools/perf/util/unwind-libunwind-local.c *val += dw_read(cur, u32, end); cur 149 tools/perf/util/unwind-libunwind-local.c *val += dw_read(cur, s64, end); cur 152 tools/perf/util/unwind-libunwind-local.c *val += dw_read(cur, u64, end); cur 159 tools/perf/util/unwind-libunwind-local.c *p = cur; cur 279 tools/power/cpupower/utils/helpers/bitmask.c unsigned int cur, rbot, rtop; cur 283 tools/power/cpupower/utils/helpers/bitmask.c rbot = cur = bitmask_first(bmp); cur 284 tools/power/cpupower/utils/helpers/bitmask.c while (cur < bmp->size) { cur 285 tools/power/cpupower/utils/helpers/bitmask.c rtop = cur; cur 286 tools/power/cpupower/utils/helpers/bitmask.c cur = bitmask_next(bmp, cur+1); cur 287 tools/power/cpupower/utils/helpers/bitmask.c if (cur >= bmp->size || cur > rtop + 1) { cur 289 tools/power/cpupower/utils/helpers/bitmask.c rbot = cur; cur 160 tools/testing/radix-tree/main.c unsigned long start, end, count = 0, tagged, cur, tmp; cur 167 tools/testing/radix-tree/main.c cur = start; cur 169 tools/testing/radix-tree/main.c end = cur; cur 173 tools/testing/radix-tree/main.c cur = rand(); cur 174 tools/testing/radix-tree/main.c if (cur & 1) { cur 176 tools/testing/radix-tree/main.c if (cur & 2) { cur 182 tools/testing/radix-tree/main.c if (cur & 4) { cur 184 tools/testing/radix-tree/main.c if (cur & 8) cur 187 tools/testing/radix-tree/main.c if (cur & 16) { cur 189 tools/testing/radix-tree/main.c if (cur & 32) { cur 195 tools/testing/radix-tree/main.c if (cur & 64) { cur 197 tools/testing/radix-tree/main.c if (cur & 128) cur 6709 tools/testing/selftests/bpf/test_btf.c const char *cur = strs; cur 6712 tools/testing/selftests/bpf/test_btf.c while (cur < strs + len) { cur 6713 tools/testing/selftests/bpf/test_btf.c fprintf(stderr, "string #%d: '%s'\n", i, cur); cur 6714 tools/testing/selftests/bpf/test_btf.c cur += strlen(cur) + 1; cur 78 tools/testing/selftests/gpio/gpio-mockup-chardev.c char *cur; cur 92 tools/testing/selftests/gpio/gpio-mockup-chardev.c cur = strstr(line, consumer); cur 93 tools/testing/selftests/gpio/gpio-mockup-chardev.c if (cur == NULL) cur 96 tools/testing/selftests/gpio/gpio-mockup-chardev.c cur = strchr(line, ')'); cur 97 tools/testing/selftests/gpio/gpio-mockup-chardev.c if (!cur) cur 100 tools/testing/selftests/gpio/gpio-mockup-chardev.c cur += 2; cur 101 tools/testing/selftests/gpio/gpio-mockup-chardev.c if (!strncmp(cur, "out", 3)) { cur 103 tools/testing/selftests/gpio/gpio-mockup-chardev.c cur += 4; cur 104 tools/testing/selftests/gpio/gpio-mockup-chardev.c } else if (!strncmp(cur, "in", 2)) { cur 106 tools/testing/selftests/gpio/gpio-mockup-chardev.c cur += 4; cur 109 tools/testing/selftests/gpio/gpio-mockup-chardev.c if (!strncmp(cur, "hi", 2)) cur 111 tools/testing/selftests/gpio/gpio-mockup-chardev.c else if (!strncmp(cur, "lo", 2)) cur 191 tools/testing/selftests/net/udpgso_bench_rx.c char cur = data[0]; cur 195 tools/testing/selftests/net/udpgso_bench_rx.c if (cur < 'a' || cur > 'z') cur 199 tools/testing/selftests/net/udpgso_bench_rx.c if (cur == 'z') cur 200 tools/testing/selftests/net/udpgso_bench_rx.c cur = 'a'; cur 202 tools/testing/selftests/net/udpgso_bench_rx.c cur++; cur 204 tools/testing/selftests/net/udpgso_bench_rx.c if (data[i] != cur) cur 208 tools/testing/selftests/net/udpgso_bench_rx.c sanitized_char(cur), cur); cur 107 tools/testing/selftests/networking/timestamping/txtimestamp.c static void validate_timestamp(struct timespec *cur, int min_delay) cur 112 tools/testing/selftests/networking/timestamping/txtimestamp.c cur64 = timespec_to_us64(cur); cur 122 tools/testing/selftests/networking/timestamping/txtimestamp.c static void __print_timestamp(const char *name, struct timespec *cur, cur 125 tools/testing/selftests/networking/timestamping/txtimestamp.c if (!(cur->tv_sec | cur->tv_nsec)) cur 129 tools/testing/selftests/networking/timestamping/txtimestamp.c name, cur->tv_sec, cur->tv_nsec / 1000, cur 132 tools/testing/selftests/networking/timestamping/txtimestamp.c if (cur != &ts_usr) cur 134 tools/testing/selftests/networking/timestamping/txtimestamp.c timespec_to_us64(cur) - timespec_to_us64(&ts_usr)); cur 664 tools/testing/selftests/networking/timestamping/txtimestamp.c struct addrinfo *addrs, *cur; cur 671 tools/testing/selftests/networking/timestamping/txtimestamp.c cur = addrs; cur 672 tools/testing/selftests/networking/timestamping/txtimestamp.c while (cur && !have_ipv4 && !have_ipv6) { cur 673 tools/testing/selftests/networking/timestamping/txtimestamp.c if (!have_ipv4 && cur->ai_family == AF_INET) { cur 674 tools/testing/selftests/networking/timestamping/txtimestamp.c memcpy(&daddr, cur->ai_addr, sizeof(daddr)); cur 678 tools/testing/selftests/networking/timestamping/txtimestamp.c else if (!have_ipv6 && cur->ai_family == AF_INET6) { cur 679 tools/testing/selftests/networking/timestamping/txtimestamp.c memcpy(&daddr6, cur->ai_addr, sizeof(daddr6)); cur 683 tools/testing/selftests/networking/timestamping/txtimestamp.c cur = cur->ai_next; cur 143 tools/testing/selftests/vm/mlock-random-test.c struct rlimit cur; cur 146 tools/testing/selftests/vm/mlock-random-test.c getrlimit(RLIMIT_MEMLOCK, &cur); cur 147 tools/testing/selftests/vm/mlock-random-test.c if (cur.rlim_cur < alloc_size) { cur 149 tools/testing/selftests/vm/mlock-random-test.c alloc_size, (unsigned int)cur.rlim_cur); cur 220 tools/testing/selftests/vm/mlock-random-test.c struct rlimit cur; cur 222 tools/testing/selftests/vm/mlock-random-test.c getrlimit(RLIMIT_MEMLOCK, &cur); cur 223 tools/testing/selftests/vm/mlock-random-test.c if (cur.rlim_cur >= alloc_size) { cur 225 tools/testing/selftests/vm/mlock-random-test.c alloc_size, (unsigned int)cur.rlim_cur); cur 233 tools/testing/selftests/vm/mlock-random-test.c int lock_size = (rand() % (alloc_size - cur.rlim_cur)) cur 234 tools/testing/selftests/vm/mlock-random-test.c + cur.rlim_cur; cur 1123 virt/kvm/arm/vgic/vgic-its.c struct its_device *cur, *temp; cur 1125 virt/kvm/arm/vgic/vgic-its.c list_for_each_entry_safe(cur, temp, &its->device_list, dev_list) cur 1126 virt/kvm/arm/vgic/vgic-its.c vgic_its_free_device(kvm, cur); cur 1132 virt/kvm/arm/vgic/vgic-its.c struct its_collection *cur, *temp; cur 1134 virt/kvm/arm/vgic/vgic-its.c list_for_each_entry_safe(cur, temp, &its->collection_list, coll_list) cur 1135 virt/kvm/arm/vgic/vgic-its.c vgic_its_free_collection(its, cur->collection_id); cur 2471 virt/kvm/kvm_main.c ktime_t start, cur; cur 2478 virt/kvm/kvm_main.c start = cur = ktime_get(); cur 2494 virt/kvm/kvm_main.c cur = ktime_get(); cur 2495 virt/kvm/kvm_main.c } while (single_task_running() && ktime_before(cur, stop)); cur 2509 virt/kvm/kvm_main.c cur = ktime_get(); cur 2512 virt/kvm/kvm_main.c block_ns = ktime_to_ns(cur) - ktime_to_ns(start);